Thu, 22 Sep 2011 10:57:37 -0700
6484982: G1: process references during evacuation pauses
Summary: G1 now uses two reference processors - one is used by concurrent marking and the other is used by STW GCs (both full and incremental evacuation pauses). In an evacuation pause, the reference processor is embedded into the closures used to scan objects. Doing so causes causes reference objects to be 'discovered' by the reference processor. At the end of the evacuation pause, these discovered reference objects are processed - preserving (and copying) referent objects (and their reachable graphs) as appropriate.
Reviewed-by: ysr, jwilhelm, brutisso, stefank, tonyp
1 /*
2 * Copyright (c) 2001, 2011, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
25 #include "precompiled.hpp"
26 #include "gc_implementation/parallelScavenge/parallelScavengeHeap.hpp"
27 #include "gc_implementation/parallelScavenge/psMarkSweepDecorator.hpp"
28 #include "gc_implementation/parallelScavenge/psParallelCompact.hpp"
29 #include "gc_implementation/parallelScavenge/psPermGen.hpp"
30 #include "gc_implementation/shared/gcUtil.hpp"
31 #include "gc_implementation/shared/markSweep.inline.hpp"
32 #include "oops/markOop.inline.hpp"
34 PSPermGen::PSPermGen(ReservedSpace rs, size_t alignment,
35 size_t initial_size, size_t min_size, size_t max_size,
36 const char* gen_name, int level) :
37 PSOldGen(rs, alignment, initial_size, min_size, max_size, gen_name, level),
38 _last_used(0)
39 {
40 assert(object_mark_sweep() != NULL, "Sanity");
42 object_mark_sweep()->set_allowed_dead_ratio(PermMarkSweepDeadRatio);
43 _avg_size = new AdaptivePaddedAverage(AdaptivePermSizeWeight,
44 PermGenPadding);
45 }
47 HeapWord* PSPermGen::allocate_permanent(size_t size) {
48 assert_locked_or_safepoint(Heap_lock);
49 HeapWord* obj = allocate_noexpand(size);
51 if (obj == NULL) {
52 obj = expand_and_allocate(size);
53 }
55 return obj;
56 }
58 void PSPermGen::compute_new_size(size_t used_before_collection) {
59 // Update our padded average of objects allocated in perm
60 // gen between collections.
61 assert(used_before_collection >= _last_used,
62 "negative allocation amount since last GC?");
64 const size_t alloc_since_last_gc = used_before_collection - _last_used;
65 _avg_size->sample(alloc_since_last_gc);
67 const size_t current_live = used_in_bytes();
68 // Stash away the current amount live for the next call to this method.
69 _last_used = current_live;
71 // We have different alignment constraints than the rest of the heap.
72 const size_t alignment = MAX2(MinPermHeapExpansion,
73 virtual_space()->alignment());
75 // Compute the desired size:
76 // The free space is the newly computed padded average,
77 // so the desired size is what's live + the free space.
78 size_t desired_size = current_live + (size_t)_avg_size->padded_average();
79 desired_size = align_size_up(desired_size, alignment);
81 // ...and no larger or smaller than our max and min allowed.
82 desired_size = MAX2(MIN2(desired_size, _max_gen_size), _min_gen_size);
83 assert(desired_size <= _max_gen_size, "just checking");
85 const size_t size_before = _virtual_space->committed_size();
87 if (desired_size == size_before) {
88 // no change, we're done
89 return;
90 }
92 {
93 // We'll be growing or shrinking the heap: in either case,
94 // we need to hold a lock.
95 MutexLocker x(ExpandHeap_lock);
96 if (desired_size > size_before) {
97 const size_t change_bytes = desired_size - size_before;
98 const size_t aligned_change_bytes =
99 align_size_up(change_bytes, alignment);
100 expand_by(aligned_change_bytes);
101 } else {
102 // Shrinking
103 const size_t change_bytes =
104 size_before - desired_size;
105 const size_t aligned_change_bytes = align_size_down(change_bytes, alignment);
106 shrink(aligned_change_bytes);
107 }
108 }
110 // While this code isn't controlled by AdaptiveSizePolicy, it's
111 // convenient to see all resizing decsions under the same flag.
112 if (PrintAdaptiveSizePolicy) {
113 ParallelScavengeHeap* heap = (ParallelScavengeHeap*)Universe::heap();
114 assert(heap->kind() == CollectedHeap::ParallelScavengeHeap, "Sanity");
116 gclog_or_tty->print_cr("AdaptiveSizePolicy::perm generation size: "
117 "collection: %d "
118 "(" SIZE_FORMAT ") -> (" SIZE_FORMAT ") ",
119 heap->total_collections(),
120 size_before, _virtual_space->committed_size());
121 }
122 }
124 void PSPermGen::precompact() {
125 // Reset start array first.
126 _start_array.reset();
127 object_mark_sweep()->precompact();
128 }