Mon, 16 Sep 2013 15:35:04 -0700
7164841: Improvements to the GC log file rotation
Summary: made changes to easily identify current log file in rotation. Parameterize the input with %t for time replacement in file name.
Reviewed-by: ccheung, tschatzl, tamao, zgu
Contributed-by: yumin.qi@oracle.com
1 /*
2 * Copyright (c) 1997, 2013, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
25 #include "precompiled.hpp"
26 #include "code/codeCache.hpp"
27 #include "code/compiledIC.hpp"
28 #include "code/icBuffer.hpp"
29 #include "code/nmethod.hpp"
30 #include "compiler/compileBroker.hpp"
31 #include "memory/resourceArea.hpp"
32 #include "oops/method.hpp"
33 #include "runtime/atomic.hpp"
34 #include "runtime/compilationPolicy.hpp"
35 #include "runtime/mutexLocker.hpp"
36 #include "runtime/os.hpp"
37 #include "runtime/sweeper.hpp"
38 #include "runtime/vm_operations.hpp"
39 #include "trace/tracing.hpp"
40 #include "utilities/events.hpp"
41 #include "utilities/xmlstream.hpp"
43 #ifdef ASSERT
45 #define SWEEP(nm) record_sweep(nm, __LINE__)
46 // Sweeper logging code
47 class SweeperRecord {
48 public:
49 int traversal;
50 int invocation;
51 int compile_id;
52 long traversal_mark;
53 int state;
54 const char* kind;
55 address vep;
56 address uep;
57 int line;
59 void print() {
60 tty->print_cr("traversal = %d invocation = %d compile_id = %d %s uep = " PTR_FORMAT " vep = "
61 PTR_FORMAT " state = %d traversal_mark %d line = %d",
62 traversal,
63 invocation,
64 compile_id,
65 kind == NULL ? "" : kind,
66 uep,
67 vep,
68 state,
69 traversal_mark,
70 line);
71 }
72 };
74 static int _sweep_index = 0;
75 static SweeperRecord* _records = NULL;
77 void NMethodSweeper::report_events(int id, address entry) {
78 if (_records != NULL) {
79 for (int i = _sweep_index; i < SweeperLogEntries; i++) {
80 if (_records[i].uep == entry ||
81 _records[i].vep == entry ||
82 _records[i].compile_id == id) {
83 _records[i].print();
84 }
85 }
86 for (int i = 0; i < _sweep_index; i++) {
87 if (_records[i].uep == entry ||
88 _records[i].vep == entry ||
89 _records[i].compile_id == id) {
90 _records[i].print();
91 }
92 }
93 }
94 }
96 void NMethodSweeper::report_events() {
97 if (_records != NULL) {
98 for (int i = _sweep_index; i < SweeperLogEntries; i++) {
99 // skip empty records
100 if (_records[i].vep == NULL) continue;
101 _records[i].print();
102 }
103 for (int i = 0; i < _sweep_index; i++) {
104 // skip empty records
105 if (_records[i].vep == NULL) continue;
106 _records[i].print();
107 }
108 }
109 }
111 void NMethodSweeper::record_sweep(nmethod* nm, int line) {
112 if (_records != NULL) {
113 _records[_sweep_index].traversal = _traversals;
114 _records[_sweep_index].traversal_mark = nm->_stack_traversal_mark;
115 _records[_sweep_index].invocation = _invocations;
116 _records[_sweep_index].compile_id = nm->compile_id();
117 _records[_sweep_index].kind = nm->compile_kind();
118 _records[_sweep_index].state = nm->_state;
119 _records[_sweep_index].vep = nm->verified_entry_point();
120 _records[_sweep_index].uep = nm->entry_point();
121 _records[_sweep_index].line = line;
123 _sweep_index = (_sweep_index + 1) % SweeperLogEntries;
124 }
125 }
126 #else
127 #define SWEEP(nm)
128 #endif
131 long NMethodSweeper::_traversals = 0; // No. of stack traversals performed
132 nmethod* NMethodSweeper::_current = NULL; // Current nmethod
133 int NMethodSweeper::_seen = 0 ; // No. of nmethods we have currently processed in current pass of CodeCache
134 int NMethodSweeper::_flushed_count = 0; // Nof. nmethods flushed in current sweep
135 int NMethodSweeper::_zombified_count = 0; // Nof. nmethods made zombie in current sweep
136 int NMethodSweeper::_marked_count = 0; // Nof. nmethods marked for reclaim in current sweep
138 volatile int NMethodSweeper::_invocations = 0; // No. of invocations left until we are completed with this pass
139 volatile int NMethodSweeper::_sweep_started = 0; // Whether a sweep is in progress.
141 jint NMethodSweeper::_locked_seen = 0;
142 jint NMethodSweeper::_not_entrant_seen_on_stack = 0;
143 bool NMethodSweeper::_resweep = false;
144 jint NMethodSweeper::_flush_token = 0;
145 jlong NMethodSweeper::_last_full_flush_time = 0;
146 int NMethodSweeper::_highest_marked = 0;
147 int NMethodSweeper::_dead_compile_ids = 0;
148 long NMethodSweeper::_last_flush_traversal_id = 0;
150 int NMethodSweeper::_number_of_flushes = 0; // Total of full traversals caused by full cache
151 int NMethodSweeper::_total_nof_methods_reclaimed = 0;
152 jlong NMethodSweeper::_total_time_sweeping = 0;
153 jlong NMethodSweeper::_total_time_this_sweep = 0;
154 jlong NMethodSweeper::_peak_sweep_time = 0;
155 jlong NMethodSweeper::_peak_sweep_fraction_time = 0;
156 jlong NMethodSweeper::_total_disconnect_time = 0;
157 jlong NMethodSweeper::_peak_disconnect_time = 0;
159 class MarkActivationClosure: public CodeBlobClosure {
160 public:
161 virtual void do_code_blob(CodeBlob* cb) {
162 // If we see an activation belonging to a non_entrant nmethod, we mark it.
163 if (cb->is_nmethod() && ((nmethod*)cb)->is_not_entrant()) {
164 ((nmethod*)cb)->mark_as_seen_on_stack();
165 }
166 }
167 };
168 static MarkActivationClosure mark_activation_closure;
170 bool NMethodSweeper::sweep_in_progress() {
171 return (_current != NULL);
172 }
174 void NMethodSweeper::scan_stacks() {
175 assert(SafepointSynchronize::is_at_safepoint(), "must be executed at a safepoint");
176 if (!MethodFlushing) return;
178 // No need to synchronize access, since this is always executed at a
179 // safepoint.
181 // Make sure CompiledIC_lock in unlocked, since we might update some
182 // inline caches. If it is, we just bail-out and try later.
183 if (CompiledIC_lock->is_locked() || Patching_lock->is_locked()) return;
185 // Check for restart
186 assert(CodeCache::find_blob_unsafe(_current) == _current, "Sweeper nmethod cached state invalid");
187 if (!sweep_in_progress() && _resweep) {
188 _seen = 0;
189 _invocations = NmethodSweepFraction;
190 _current = CodeCache::first_nmethod();
191 _traversals += 1;
192 _total_time_this_sweep = 0;
194 if (PrintMethodFlushing) {
195 tty->print_cr("### Sweep: stack traversal %d", _traversals);
196 }
197 Threads::nmethods_do(&mark_activation_closure);
199 // reset the flags since we started a scan from the beginning.
200 _resweep = false;
201 _locked_seen = 0;
202 _not_entrant_seen_on_stack = 0;
203 }
205 if (UseCodeCacheFlushing) {
206 // only allow new flushes after the interval is complete.
207 jlong now = os::javaTimeMillis();
208 jlong max_interval = (jlong)MinCodeCacheFlushingInterval * (jlong)1000;
209 jlong curr_interval = now - _last_full_flush_time;
210 if (curr_interval > max_interval) {
211 _flush_token = 0;
212 }
214 if (!CodeCache::needs_flushing() && !CompileBroker::should_compile_new_jobs()) {
215 CompileBroker::set_should_compile_new_jobs(CompileBroker::run_compilation);
216 log_sweep("restart_compiler");
217 }
218 }
219 }
221 void NMethodSweeper::possibly_sweep() {
222 assert(JavaThread::current()->thread_state() == _thread_in_vm, "must run in vm mode");
223 if (!MethodFlushing || !sweep_in_progress()) return;
225 if (_invocations > 0) {
226 // Only one thread at a time will sweep
227 jint old = Atomic::cmpxchg( 1, &_sweep_started, 0 );
228 if (old != 0) {
229 return;
230 }
231 #ifdef ASSERT
232 if (LogSweeper && _records == NULL) {
233 // Create the ring buffer for the logging code
234 _records = NEW_C_HEAP_ARRAY(SweeperRecord, SweeperLogEntries, mtGC);
235 memset(_records, 0, sizeof(SweeperRecord) * SweeperLogEntries);
236 }
237 #endif
238 if (_invocations > 0) {
239 sweep_code_cache();
240 _invocations--;
241 }
242 _sweep_started = 0;
243 }
244 }
246 void NMethodSweeper::sweep_code_cache() {
248 jlong sweep_start_counter = os::elapsed_counter();
250 _flushed_count = 0;
251 _zombified_count = 0;
252 _marked_count = 0;
254 if (PrintMethodFlushing && Verbose) {
255 tty->print_cr("### Sweep at %d out of %d. Invocations left: %d", _seen, CodeCache::nof_nmethods(), _invocations);
256 }
258 if (!CompileBroker::should_compile_new_jobs()) {
259 // If we have turned off compilations we might as well do full sweeps
260 // in order to reach the clean state faster. Otherwise the sleeping compiler
261 // threads will slow down sweeping. After a few iterations the cache
262 // will be clean and sweeping stops (_resweep will not be set)
263 _invocations = 1;
264 }
266 // We want to visit all nmethods after NmethodSweepFraction
267 // invocations so divide the remaining number of nmethods by the
268 // remaining number of invocations. This is only an estimate since
269 // the number of nmethods changes during the sweep so the final
270 // stage must iterate until it there are no more nmethods.
271 int todo = (CodeCache::nof_nmethods() - _seen) / _invocations;
273 assert(!SafepointSynchronize::is_at_safepoint(), "should not be in safepoint when we get here");
274 assert(!CodeCache_lock->owned_by_self(), "just checking");
276 {
277 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
279 // The last invocation iterates until there are no more nmethods
280 for (int i = 0; (i < todo || _invocations == 1) && _current != NULL; i++) {
281 if (SafepointSynchronize::is_synchronizing()) { // Safepoint request
282 if (PrintMethodFlushing && Verbose) {
283 tty->print_cr("### Sweep at %d out of %d, invocation: %d, yielding to safepoint", _seen, CodeCache::nof_nmethods(), _invocations);
284 }
285 MutexUnlockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
287 assert(Thread::current()->is_Java_thread(), "should be java thread");
288 JavaThread* thread = (JavaThread*)Thread::current();
289 ThreadBlockInVM tbivm(thread);
290 thread->java_suspend_self();
291 }
292 // Since we will give up the CodeCache_lock, always skip ahead
293 // to the next nmethod. Other blobs can be deleted by other
294 // threads but nmethods are only reclaimed by the sweeper.
295 nmethod* next = CodeCache::next_nmethod(_current);
297 // Now ready to process nmethod and give up CodeCache_lock
298 {
299 MutexUnlockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
300 process_nmethod(_current);
301 }
302 _seen++;
303 _current = next;
304 }
305 }
307 assert(_invocations > 1 || _current == NULL, "must have scanned the whole cache");
309 if (!sweep_in_progress() && !_resweep && (_locked_seen || _not_entrant_seen_on_stack)) {
310 // we've completed a scan without making progress but there were
311 // nmethods we were unable to process either because they were
312 // locked or were still on stack. We don't have to aggresively
313 // clean them up so just stop scanning. We could scan once more
314 // but that complicates the control logic and it's unlikely to
315 // matter much.
316 if (PrintMethodFlushing) {
317 tty->print_cr("### Couldn't make progress on some nmethods so stopping sweep");
318 }
319 }
321 jlong sweep_end_counter = os::elapsed_counter();
322 jlong sweep_time = sweep_end_counter - sweep_start_counter;
323 _total_time_sweeping += sweep_time;
324 _total_time_this_sweep += sweep_time;
325 _peak_sweep_fraction_time = MAX2(sweep_time, _peak_sweep_fraction_time);
326 _total_nof_methods_reclaimed += _flushed_count;
328 EventSweepCodeCache event(UNTIMED);
329 if (event.should_commit()) {
330 event.set_starttime(sweep_start_counter);
331 event.set_endtime(sweep_end_counter);
332 event.set_sweepIndex(_traversals);
333 event.set_sweepFractionIndex(NmethodSweepFraction - _invocations + 1);
334 event.set_sweptCount(todo);
335 event.set_flushedCount(_flushed_count);
336 event.set_markedCount(_marked_count);
337 event.set_zombifiedCount(_zombified_count);
338 event.commit();
339 }
341 #ifdef ASSERT
342 if(PrintMethodFlushing) {
343 tty->print_cr("### sweeper: sweep time(%d): " INT64_FORMAT, _invocations, (jlong)sweep_time);
344 }
345 #endif
347 if (_invocations == 1) {
348 _peak_sweep_time = MAX2(_peak_sweep_time, _total_time_this_sweep);
349 log_sweep("finished");
350 }
352 // Sweeper is the only case where memory is released,
353 // check here if it is time to restart the compiler.
354 if (UseCodeCacheFlushing && !CompileBroker::should_compile_new_jobs() && !CodeCache::needs_flushing()) {
355 CompileBroker::set_should_compile_new_jobs(CompileBroker::run_compilation);
356 log_sweep("restart_compiler");
357 }
358 }
360 class NMethodMarker: public StackObj {
361 private:
362 CompilerThread* _thread;
363 public:
364 NMethodMarker(nmethod* nm) {
365 _thread = CompilerThread::current();
366 if (!nm->is_zombie() && !nm->is_unloaded()) {
367 // Only expose live nmethods for scanning
368 _thread->set_scanned_nmethod(nm);
369 }
370 }
371 ~NMethodMarker() {
372 _thread->set_scanned_nmethod(NULL);
373 }
374 };
376 void NMethodSweeper::release_nmethod(nmethod *nm) {
377 // Clean up any CompiledICHolders
378 {
379 ResourceMark rm;
380 MutexLocker ml_patch(CompiledIC_lock);
381 RelocIterator iter(nm);
382 while (iter.next()) {
383 if (iter.type() == relocInfo::virtual_call_type) {
384 CompiledIC::cleanup_call_site(iter.virtual_call_reloc());
385 }
386 }
387 }
389 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
390 nm->flush();
391 }
393 void NMethodSweeper::process_nmethod(nmethod *nm) {
394 assert(!CodeCache_lock->owned_by_self(), "just checking");
396 // Make sure this nmethod doesn't get unloaded during the scan,
397 // since the locks acquired below might safepoint.
398 NMethodMarker nmm(nm);
400 SWEEP(nm);
402 // Skip methods that are currently referenced by the VM
403 if (nm->is_locked_by_vm()) {
404 // But still remember to clean-up inline caches for alive nmethods
405 if (nm->is_alive()) {
406 // Clean-up all inline caches that points to zombie/non-reentrant methods
407 MutexLocker cl(CompiledIC_lock);
408 nm->cleanup_inline_caches();
409 SWEEP(nm);
410 } else {
411 _locked_seen++;
412 SWEEP(nm);
413 }
414 return;
415 }
417 if (nm->is_zombie()) {
418 // If it is first time, we see nmethod then we mark it. Otherwise,
419 // we reclame it. When we have seen a zombie method twice, we know that
420 // there are no inline caches that refer to it.
421 if (nm->is_marked_for_reclamation()) {
422 assert(!nm->is_locked_by_vm(), "must not flush locked nmethods");
423 if (PrintMethodFlushing && Verbose) {
424 tty->print_cr("### Nmethod %3d/" PTR_FORMAT " (marked for reclamation) being flushed", nm->compile_id(), nm);
425 }
426 release_nmethod(nm);
427 _flushed_count++;
428 } else {
429 if (PrintMethodFlushing && Verbose) {
430 tty->print_cr("### Nmethod %3d/" PTR_FORMAT " (zombie) being marked for reclamation", nm->compile_id(), nm);
431 }
432 nm->mark_for_reclamation();
433 _resweep = true;
434 _marked_count++;
435 SWEEP(nm);
436 }
437 } else if (nm->is_not_entrant()) {
438 // If there is no current activations of this method on the
439 // stack we can safely convert it to a zombie method
440 if (nm->can_not_entrant_be_converted()) {
441 if (PrintMethodFlushing && Verbose) {
442 tty->print_cr("### Nmethod %3d/" PTR_FORMAT " (not entrant) being made zombie", nm->compile_id(), nm);
443 }
444 nm->make_zombie();
445 _resweep = true;
446 _zombified_count++;
447 SWEEP(nm);
448 } else {
449 // Still alive, clean up its inline caches
450 MutexLocker cl(CompiledIC_lock);
451 nm->cleanup_inline_caches();
452 // we coudn't transition this nmethod so don't immediately
453 // request a rescan. If this method stays on the stack for a
454 // long time we don't want to keep rescanning the code cache.
455 _not_entrant_seen_on_stack++;
456 SWEEP(nm);
457 }
458 } else if (nm->is_unloaded()) {
459 // Unloaded code, just make it a zombie
460 if (PrintMethodFlushing && Verbose)
461 tty->print_cr("### Nmethod %3d/" PTR_FORMAT " (unloaded) being made zombie", nm->compile_id(), nm);
463 if (nm->is_osr_method()) {
464 SWEEP(nm);
465 // No inline caches will ever point to osr methods, so we can just remove it
466 release_nmethod(nm);
467 _flushed_count++;
468 } else {
469 nm->make_zombie();
470 _resweep = true;
471 _zombified_count++;
472 SWEEP(nm);
473 }
474 } else {
475 assert(nm->is_alive(), "should be alive");
477 if (UseCodeCacheFlushing) {
478 if (nm->is_speculatively_disconnected() && !nm->is_locked_by_vm() && !nm->is_osr_method() &&
479 (_traversals > _last_flush_traversal_id + 2) && (nm->compile_id() < _highest_marked)) {
480 // This method has not been called since the forced cleanup happened
481 nm->make_not_entrant();
482 }
483 }
485 // Clean-up all inline caches that points to zombie/non-reentrant methods
486 MutexLocker cl(CompiledIC_lock);
487 nm->cleanup_inline_caches();
488 SWEEP(nm);
489 }
490 }
492 // Code cache unloading: when compilers notice the code cache is getting full,
493 // they will call a vm op that comes here. This code attempts to speculatively
494 // unload the oldest half of the nmethods (based on the compile job id) by
495 // saving the old code in a list in the CodeCache. Then
496 // execution resumes. If a method so marked is not called by the second sweeper
497 // stack traversal after the current one, the nmethod will be marked non-entrant and
498 // got rid of by normal sweeping. If the method is called, the Method*'s
499 // _code field is restored and the Method*/nmethod
500 // go back to their normal state.
501 void NMethodSweeper::handle_full_code_cache(bool is_full) {
503 if (is_full) {
504 // Since code cache is full, immediately stop new compiles
505 if (CompileBroker::set_should_compile_new_jobs(CompileBroker::stop_compilation)) {
506 log_sweep("disable_compiler");
507 }
508 }
510 // Make sure only one thread can flush
511 // The token is reset after CodeCacheMinimumFlushInterval in scan stacks,
512 // no need to check the timeout here.
513 jint old = Atomic::cmpxchg( 1, &_flush_token, 0 );
514 if (old != 0) {
515 return;
516 }
518 VM_HandleFullCodeCache op(is_full);
519 VMThread::execute(&op);
521 // resweep again as soon as possible
522 _resweep = true;
523 }
525 void NMethodSweeper::speculative_disconnect_nmethods(bool is_full) {
526 // If there was a race in detecting full code cache, only run
527 // one vm op for it or keep the compiler shut off
529 jlong disconnect_start_counter = os::elapsed_counter();
531 // Traverse the code cache trying to dump the oldest nmethods
532 int curr_max_comp_id = CompileBroker::get_compilation_id();
533 int flush_target = ((curr_max_comp_id - _dead_compile_ids) / CodeCacheFlushingFraction) + _dead_compile_ids;
535 log_sweep("start_cleaning");
537 nmethod* nm = CodeCache::alive_nmethod(CodeCache::first());
538 jint disconnected = 0;
539 jint made_not_entrant = 0;
540 jint nmethod_count = 0;
542 while ((nm != NULL)){
543 int curr_comp_id = nm->compile_id();
545 // OSR methods cannot be flushed like this. Also, don't flush native methods
546 // since they are part of the JDK in most cases
547 if (!nm->is_osr_method() && !nm->is_locked_by_vm() && !nm->is_native_method()) {
549 // only count methods that can be speculatively disconnected
550 nmethod_count++;
552 if (nm->is_in_use() && (curr_comp_id < flush_target)) {
553 if ((nm->method()->code() == nm)) {
554 // This method has not been previously considered for
555 // unloading or it was restored already
556 CodeCache::speculatively_disconnect(nm);
557 disconnected++;
558 } else if (nm->is_speculatively_disconnected()) {
559 // This method was previously considered for preemptive unloading and was not called since then
560 CompilationPolicy::policy()->delay_compilation(nm->method());
561 nm->make_not_entrant();
562 made_not_entrant++;
563 }
565 if (curr_comp_id > _highest_marked) {
566 _highest_marked = curr_comp_id;
567 }
568 }
569 }
570 nm = CodeCache::alive_nmethod(CodeCache::next(nm));
571 }
573 // remember how many compile_ids wheren't seen last flush.
574 _dead_compile_ids = curr_max_comp_id - nmethod_count;
576 log_sweep("stop_cleaning",
577 "disconnected='" UINT32_FORMAT "' made_not_entrant='" UINT32_FORMAT "'",
578 disconnected, made_not_entrant);
580 // Shut off compiler. Sweeper will start over with a new stack scan and
581 // traversal cycle and turn it back on if it clears enough space.
582 if (is_full) {
583 _last_full_flush_time = os::javaTimeMillis();
584 }
586 jlong disconnect_end_counter = os::elapsed_counter();
587 jlong disconnect_time = disconnect_end_counter - disconnect_start_counter;
588 _total_disconnect_time += disconnect_time;
589 _peak_disconnect_time = MAX2(disconnect_time, _peak_disconnect_time);
591 EventCleanCodeCache event(UNTIMED);
592 if (event.should_commit()) {
593 event.set_starttime(disconnect_start_counter);
594 event.set_endtime(disconnect_end_counter);
595 event.set_disconnectedCount(disconnected);
596 event.set_madeNonEntrantCount(made_not_entrant);
597 event.commit();
598 }
599 _number_of_flushes++;
601 // After two more traversals the sweeper will get rid of unrestored nmethods
602 _last_flush_traversal_id = _traversals;
603 _resweep = true;
604 #ifdef ASSERT
606 if(PrintMethodFlushing && Verbose) {
607 tty->print_cr("### sweeper: unload time: " INT64_FORMAT, (jlong)disconnect_time);
608 }
609 #endif
610 }
613 // Print out some state information about the current sweep and the
614 // state of the code cache if it's requested.
615 void NMethodSweeper::log_sweep(const char* msg, const char* format, ...) {
616 if (PrintMethodFlushing) {
617 stringStream s;
618 // Dump code cache state into a buffer before locking the tty,
619 // because log_state() will use locks causing lock conflicts.
620 CodeCache::log_state(&s);
622 ttyLocker ttyl;
623 tty->print("### sweeper: %s ", msg);
624 if (format != NULL) {
625 va_list ap;
626 va_start(ap, format);
627 tty->vprint(format, ap);
628 va_end(ap);
629 }
630 tty->print_cr(s.as_string());
631 }
633 if (LogCompilation && (xtty != NULL)) {
634 stringStream s;
635 // Dump code cache state into a buffer before locking the tty,
636 // because log_state() will use locks causing lock conflicts.
637 CodeCache::log_state(&s);
639 ttyLocker ttyl;
640 xtty->begin_elem("sweeper state='%s' traversals='" INTX_FORMAT "' ", msg, (intx)traversal_count());
641 if (format != NULL) {
642 va_list ap;
643 va_start(ap, format);
644 xtty->vprint(format, ap);
645 va_end(ap);
646 }
647 xtty->print(s.as_string());
648 xtty->stamp();
649 xtty->end_elem();
650 }
651 }