Tue, 18 Sep 2012 23:35:42 -0700
7045397: NPG: Add freelists to class loader arenas.
Reviewed-by: coleenp, stefank, jprovino, ohair
1 /*
2 * Copyright (c) 2005, 2012, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
25 #include "precompiled.hpp"
26 #include "gc_implementation/concurrentMarkSweep/concurrentMarkSweepGeneration.inline.hpp"
27 #include "gc_implementation/concurrentMarkSweep/concurrentMarkSweepThread.hpp"
28 #include "gc_implementation/concurrentMarkSweep/vmCMSOperations.hpp"
29 #include "gc_implementation/shared/isGCActiveMark.hpp"
30 #include "memory/gcLocker.inline.hpp"
31 #include "runtime/interfaceSupport.hpp"
32 #include "utilities/dtrace.hpp"
35 #ifndef USDT2
36 HS_DTRACE_PROBE_DECL(hs_private, cms__initmark__begin);
37 HS_DTRACE_PROBE_DECL(hs_private, cms__initmark__end);
39 HS_DTRACE_PROBE_DECL(hs_private, cms__remark__begin);
40 HS_DTRACE_PROBE_DECL(hs_private, cms__remark__end);
41 #endif /* !USDT2 */
43 //////////////////////////////////////////////////////////
44 // Methods in abstract class VM_CMS_Operation
45 //////////////////////////////////////////////////////////
46 void VM_CMS_Operation::acquire_pending_list_lock() {
47 // The caller may block while communicating
48 // with the SLT thread in order to acquire/release the PLL.
49 ConcurrentMarkSweepThread::slt()->
50 manipulatePLL(SurrogateLockerThread::acquirePLL);
51 }
53 void VM_CMS_Operation::release_and_notify_pending_list_lock() {
54 // The caller may block while communicating
55 // with the SLT thread in order to acquire/release the PLL.
56 ConcurrentMarkSweepThread::slt()->
57 manipulatePLL(SurrogateLockerThread::releaseAndNotifyPLL);
58 }
60 void VM_CMS_Operation::verify_before_gc() {
61 if (VerifyBeforeGC &&
62 GenCollectedHeap::heap()->total_collections() >= VerifyGCStartAt) {
63 HandleMark hm;
64 FreelistLocker x(_collector);
65 MutexLockerEx y(_collector->bitMapLock(), Mutex::_no_safepoint_check_flag);
66 Universe::heap()->prepare_for_verify();
67 Universe::verify();
68 }
69 }
71 void VM_CMS_Operation::verify_after_gc() {
72 if (VerifyAfterGC &&
73 GenCollectedHeap::heap()->total_collections() >= VerifyGCStartAt) {
74 HandleMark hm;
75 FreelistLocker x(_collector);
76 MutexLockerEx y(_collector->bitMapLock(), Mutex::_no_safepoint_check_flag);
77 Universe::verify();
78 }
79 }
81 bool VM_CMS_Operation::lost_race() const {
82 if (CMSCollector::abstract_state() == CMSCollector::Idling) {
83 // We lost a race to a foreground collection
84 // -- there's nothing to do
85 return true;
86 }
87 assert(CMSCollector::abstract_state() == legal_state(),
88 "Inconsistent collector state?");
89 return false;
90 }
92 bool VM_CMS_Operation::doit_prologue() {
93 assert(Thread::current()->is_ConcurrentGC_thread(), "just checking");
94 assert(!CMSCollector::foregroundGCShouldWait(), "Possible deadlock");
95 assert(!ConcurrentMarkSweepThread::cms_thread_has_cms_token(),
96 "Possible deadlock");
98 if (needs_pll()) {
99 acquire_pending_list_lock();
100 }
101 // Get the Heap_lock after the pending_list_lock.
102 Heap_lock->lock();
103 if (lost_race()) {
104 assert(_prologue_succeeded == false, "Initialized in c'tor");
105 Heap_lock->unlock();
106 if (needs_pll()) {
107 release_and_notify_pending_list_lock();
108 }
109 } else {
110 _prologue_succeeded = true;
111 }
112 return _prologue_succeeded;
113 }
115 void VM_CMS_Operation::doit_epilogue() {
116 assert(Thread::current()->is_ConcurrentGC_thread(), "just checking");
117 assert(!CMSCollector::foregroundGCShouldWait(), "Possible deadlock");
118 assert(!ConcurrentMarkSweepThread::cms_thread_has_cms_token(),
119 "Possible deadlock");
121 // Release the Heap_lock first.
122 Heap_lock->unlock();
123 if (needs_pll()) {
124 release_and_notify_pending_list_lock();
125 }
126 }
128 //////////////////////////////////////////////////////////
129 // Methods in class VM_CMS_Initial_Mark
130 //////////////////////////////////////////////////////////
131 void VM_CMS_Initial_Mark::doit() {
132 if (lost_race()) {
133 // Nothing to do.
134 return;
135 }
136 #ifndef USDT2
137 HS_DTRACE_PROBE(hs_private, cms__initmark__begin);
138 #else /* USDT2 */
139 HS_PRIVATE_CMS_INITMARK_BEGIN(
140 );
141 #endif /* USDT2 */
143 GenCollectedHeap* gch = GenCollectedHeap::heap();
144 GCCauseSetter gccs(gch, GCCause::_cms_initial_mark);
146 VM_CMS_Operation::verify_before_gc();
148 IsGCActiveMark x; // stop-world GC active
149 _collector->do_CMS_operation(CMSCollector::CMS_op_checkpointRootsInitial, gch->gc_cause());
151 VM_CMS_Operation::verify_after_gc();
152 #ifndef USDT2
153 HS_DTRACE_PROBE(hs_private, cms__initmark__end);
154 #else /* USDT2 */
155 HS_PRIVATE_CMS_INITMARK_END(
156 );
157 #endif /* USDT2 */
158 }
160 //////////////////////////////////////////////////////////
161 // Methods in class VM_CMS_Final_Remark_Operation
162 //////////////////////////////////////////////////////////
163 void VM_CMS_Final_Remark::doit() {
164 if (lost_race()) {
165 // Nothing to do.
166 return;
167 }
168 #ifndef USDT2
169 HS_DTRACE_PROBE(hs_private, cms__remark__begin);
170 #else /* USDT2 */
171 HS_PRIVATE_CMS_REMARK_BEGIN(
172 );
173 #endif /* USDT2 */
175 GenCollectedHeap* gch = GenCollectedHeap::heap();
176 GCCauseSetter gccs(gch, GCCause::_cms_final_remark);
178 VM_CMS_Operation::verify_before_gc();
180 IsGCActiveMark x; // stop-world GC active
181 _collector->do_CMS_operation(CMSCollector::CMS_op_checkpointRootsFinal, gch->gc_cause());
183 VM_CMS_Operation::verify_after_gc();
184 #ifndef USDT2
185 HS_DTRACE_PROBE(hs_private, cms__remark__end);
186 #else /* USDT2 */
187 HS_PRIVATE_CMS_REMARK_END(
188 );
189 #endif /* USDT2 */
190 }
192 // VM operation to invoke a concurrent collection of a
193 // GenCollectedHeap heap.
194 void VM_GenCollectFullConcurrent::doit() {
195 assert(Thread::current()->is_VM_thread(), "Should be VM thread");
196 assert(GCLockerInvokesConcurrent || ExplicitGCInvokesConcurrent, "Unexpected");
198 GenCollectedHeap* gch = GenCollectedHeap::heap();
199 if (_gc_count_before == gch->total_collections()) {
200 // The "full" of do_full_collection call below "forces"
201 // a collection; the second arg, 0, below ensures that
202 // only the young gen is collected. XXX In the future,
203 // we'll probably need to have something in this interface
204 // to say do this only if we are sure we will not bail
205 // out to a full collection in this attempt, but that's
206 // for the future.
207 assert(SafepointSynchronize::is_at_safepoint(),
208 "We can only be executing this arm of if at a safepoint");
209 GCCauseSetter gccs(gch, _gc_cause);
210 gch->do_full_collection(gch->must_clear_all_soft_refs(),
211 0 /* collect only youngest gen */);
212 } // Else no need for a foreground young gc
213 assert((_gc_count_before < gch->total_collections()) ||
214 (GC_locker::is_active() /* gc may have been skipped */
215 && (_gc_count_before == gch->total_collections())),
216 "total_collections() should be monotonically increasing");
218 MutexLockerEx x(FullGCCount_lock, Mutex::_no_safepoint_check_flag);
219 assert(_full_gc_count_before <= gch->total_full_collections(), "Error");
220 if (gch->total_full_collections() == _full_gc_count_before) {
221 // Disable iCMS until the full collection is done, and
222 // remember that we did so.
223 CMSCollector::disable_icms();
224 _disabled_icms = true;
225 // In case CMS thread was in icms_wait(), wake it up.
226 CMSCollector::start_icms();
227 // Nudge the CMS thread to start a concurrent collection.
228 CMSCollector::request_full_gc(_full_gc_count_before);
229 } else {
230 assert(_full_gc_count_before < gch->total_full_collections(), "Error");
231 FullGCCount_lock->notify_all(); // Inform the Java thread its work is done
232 }
233 }
235 bool VM_GenCollectFullConcurrent::evaluate_at_safepoint() const {
236 Thread* thr = Thread::current();
237 assert(thr != NULL, "Unexpected tid");
238 if (!thr->is_Java_thread()) {
239 assert(thr->is_VM_thread(), "Expected to be evaluated by VM thread");
240 GenCollectedHeap* gch = GenCollectedHeap::heap();
241 if (_gc_count_before != gch->total_collections()) {
242 // No need to do a young gc, we'll just nudge the CMS thread
243 // in the doit() method above, to be executed soon.
244 assert(_gc_count_before < gch->total_collections(),
245 "total_collections() should be monotnically increasing");
246 return false; // no need for foreground young gc
247 }
248 }
249 return true; // may still need foreground young gc
250 }
253 void VM_GenCollectFullConcurrent::doit_epilogue() {
254 Thread* thr = Thread::current();
255 assert(thr->is_Java_thread(), "just checking");
256 JavaThread* jt = (JavaThread*)thr;
257 // Release the Heap_lock first.
258 Heap_lock->unlock();
259 release_and_notify_pending_list_lock();
261 // It is fine to test whether completed collections has
262 // exceeded our request count without locking because
263 // the completion count is monotonically increasing;
264 // this will break for very long-running apps when the
265 // count overflows and wraps around. XXX fix me !!!
266 // e.g. at the rate of 1 full gc per ms, this could
267 // overflow in about 1000 years.
268 GenCollectedHeap* gch = GenCollectedHeap::heap();
269 if (_gc_cause != GCCause::_gc_locker &&
270 gch->total_full_collections_completed() <= _full_gc_count_before) {
271 // maybe we should change the condition to test _gc_cause ==
272 // GCCause::_java_lang_system_gc, instead of
273 // _gc_cause != GCCause::_gc_locker
274 assert(_gc_cause == GCCause::_java_lang_system_gc,
275 "the only way to get here if this was a System.gc()-induced GC");
276 assert(ExplicitGCInvokesConcurrent, "Error");
277 // Now, wait for witnessing concurrent gc cycle to complete,
278 // but do so in native mode, because we want to lock the
279 // FullGCEvent_lock, which may be needed by the VM thread
280 // or by the CMS thread, so we do not want to be suspended
281 // while holding that lock.
282 ThreadToNativeFromVM native(jt);
283 MutexLockerEx ml(FullGCCount_lock, Mutex::_no_safepoint_check_flag);
284 // Either a concurrent or a stop-world full gc is sufficient
285 // witness to our request.
286 while (gch->total_full_collections_completed() <= _full_gc_count_before) {
287 FullGCCount_lock->wait(Mutex::_no_safepoint_check_flag);
288 }
289 }
290 // Enable iCMS back if we disabled it earlier.
291 if (_disabled_icms) {
292 CMSCollector::enable_icms();
293 }
294 }