src/share/vm/runtime/sweeper.cpp

Sat, 23 Nov 2013 12:25:13 +0100

author
mgronlun
date
Sat, 23 Nov 2013 12:25:13 +0100
changeset 6131
86e6d691f2e1
parent 6114
938e1e64e28f
child 6205
908afcc9d1cb
child 6493
3205e78d8193
permissions
-rw-r--r--

8028128: Add a type safe alternative for working with counter based data
Reviewed-by: dholmes, egahlin

     1 /*
     2  * Copyright (c) 1997, 2013, Oracle and/or its affiliates. All rights reserved.
     3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
     4  *
     5  * This code is free software; you can redistribute it and/or modify it
     6  * under the terms of the GNU General Public License version 2 only, as
     7  * published by the Free Software Foundation.
     8  *
     9  * This code is distributed in the hope that it will be useful, but WITHOUT
    10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
    11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
    12  * version 2 for more details (a copy is included in the LICENSE file that
    13  * accompanied this code).
    14  *
    15  * You should have received a copy of the GNU General Public License version
    16  * 2 along with this work; if not, write to the Free Software Foundation,
    17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
    18  *
    19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
    20  * or visit www.oracle.com if you need additional information or have any
    21  * questions.
    22  *
    23  */
    25 #include "precompiled.hpp"
    26 #include "code/codeCache.hpp"
    27 #include "code/compiledIC.hpp"
    28 #include "code/icBuffer.hpp"
    29 #include "code/nmethod.hpp"
    30 #include "compiler/compileBroker.hpp"
    31 #include "memory/resourceArea.hpp"
    32 #include "oops/method.hpp"
    33 #include "runtime/atomic.hpp"
    34 #include "runtime/compilationPolicy.hpp"
    35 #include "runtime/mutexLocker.hpp"
    36 #include "runtime/os.hpp"
    37 #include "runtime/sweeper.hpp"
    38 #include "runtime/vm_operations.hpp"
    39 #include "trace/tracing.hpp"
    40 #include "utilities/events.hpp"
    41 #include "utilities/ticks.inline.hpp"
    42 #include "utilities/xmlstream.hpp"
    44 #ifdef ASSERT
    46 #define SWEEP(nm) record_sweep(nm, __LINE__)
    47 // Sweeper logging code
    48 class SweeperRecord {
    49  public:
    50   int traversal;
    51   int invocation;
    52   int compile_id;
    53   long traversal_mark;
    54   int state;
    55   const char* kind;
    56   address vep;
    57   address uep;
    58   int line;
    60   void print() {
    61       tty->print_cr("traversal = %d invocation = %d compile_id = %d %s uep = " PTR_FORMAT " vep = "
    62                     PTR_FORMAT " state = %d traversal_mark %d line = %d",
    63                     traversal,
    64                     invocation,
    65                     compile_id,
    66                     kind == NULL ? "" : kind,
    67                     uep,
    68                     vep,
    69                     state,
    70                     traversal_mark,
    71                     line);
    72   }
    73 };
    75 static int _sweep_index = 0;
    76 static SweeperRecord* _records = NULL;
    78 void NMethodSweeper::report_events(int id, address entry) {
    79   if (_records != NULL) {
    80     for (int i = _sweep_index; i < SweeperLogEntries; i++) {
    81       if (_records[i].uep == entry ||
    82           _records[i].vep == entry ||
    83           _records[i].compile_id == id) {
    84         _records[i].print();
    85       }
    86     }
    87     for (int i = 0; i < _sweep_index; i++) {
    88       if (_records[i].uep == entry ||
    89           _records[i].vep == entry ||
    90           _records[i].compile_id == id) {
    91         _records[i].print();
    92       }
    93     }
    94   }
    95 }
    97 void NMethodSweeper::report_events() {
    98   if (_records != NULL) {
    99     for (int i = _sweep_index; i < SweeperLogEntries; i++) {
   100       // skip empty records
   101       if (_records[i].vep == NULL) continue;
   102       _records[i].print();
   103     }
   104     for (int i = 0; i < _sweep_index; i++) {
   105       // skip empty records
   106       if (_records[i].vep == NULL) continue;
   107       _records[i].print();
   108     }
   109   }
   110 }
   112 void NMethodSweeper::record_sweep(nmethod* nm, int line) {
   113   if (_records != NULL) {
   114     _records[_sweep_index].traversal = _traversals;
   115     _records[_sweep_index].traversal_mark = nm->_stack_traversal_mark;
   116     _records[_sweep_index].invocation = _sweep_fractions_left;
   117     _records[_sweep_index].compile_id = nm->compile_id();
   118     _records[_sweep_index].kind = nm->compile_kind();
   119     _records[_sweep_index].state = nm->_state;
   120     _records[_sweep_index].vep = nm->verified_entry_point();
   121     _records[_sweep_index].uep = nm->entry_point();
   122     _records[_sweep_index].line = line;
   123     _sweep_index = (_sweep_index + 1) % SweeperLogEntries;
   124   }
   125 }
   126 #else
   127 #define SWEEP(nm)
   128 #endif
   130 nmethod* NMethodSweeper::_current                      = NULL; // Current nmethod
   131 long     NMethodSweeper::_traversals                   = 0;    // Stack scan count, also sweep ID.
   132 long     NMethodSweeper::_time_counter                 = 0;    // Virtual time used to periodically invoke sweeper
   133 long     NMethodSweeper::_last_sweep                   = 0;    // Value of _time_counter when the last sweep happened
   134 int      NMethodSweeper::_seen                         = 0;    // Nof. nmethod we have currently processed in current pass of CodeCache
   135 int      NMethodSweeper::_flushed_count                = 0;    // Nof. nmethods flushed in current sweep
   136 int      NMethodSweeper::_zombified_count              = 0;    // Nof. nmethods made zombie in current sweep
   137 int      NMethodSweeper::_marked_for_reclamation_count = 0;    // Nof. nmethods marked for reclaim in current sweep
   139 volatile bool NMethodSweeper::_should_sweep            = true; // Indicates if we should invoke the sweeper
   140 volatile int  NMethodSweeper::_sweep_fractions_left    = 0;    // Nof. invocations left until we are completed with this pass
   141 volatile int  NMethodSweeper::_sweep_started           = 0;    // Flag to control conc sweeper
   142 volatile int  NMethodSweeper::_bytes_changed           = 0;    // Counts the total nmethod size if the nmethod changed from:
   143                                                                //   1) alive       -> not_entrant
   144                                                                //   2) not_entrant -> zombie
   145                                                                //   3) zombie      -> marked_for_reclamation
   147 int   NMethodSweeper::_total_nof_methods_reclaimed     = 0;    // Accumulated nof methods flushed
   148 Tickspan NMethodSweeper::_total_time_sweeping;                 // Accumulated time sweeping
   149 Tickspan NMethodSweeper::_total_time_this_sweep;               // Total time this sweep
   150 Tickspan NMethodSweeper::_peak_sweep_time;                     // Peak time for a full sweep
   151 Tickspan NMethodSweeper::_peak_sweep_fraction_time;            // Peak time sweeping one fraction
   152 int   NMethodSweeper::_hotness_counter_reset_val       = 0;
   155 class MarkActivationClosure: public CodeBlobClosure {
   156 public:
   157   virtual void do_code_blob(CodeBlob* cb) {
   158     if (cb->is_nmethod()) {
   159       nmethod* nm = (nmethod*)cb;
   160       nm->set_hotness_counter(NMethodSweeper::hotness_counter_reset_val());
   161       // If we see an activation belonging to a non_entrant nmethod, we mark it.
   162       if (nm->is_not_entrant()) {
   163         nm->mark_as_seen_on_stack();
   164       }
   165     }
   166   }
   167 };
   168 static MarkActivationClosure mark_activation_closure;
   170 class SetHotnessClosure: public CodeBlobClosure {
   171 public:
   172   virtual void do_code_blob(CodeBlob* cb) {
   173     if (cb->is_nmethod()) {
   174       nmethod* nm = (nmethod*)cb;
   175       nm->set_hotness_counter(NMethodSweeper::hotness_counter_reset_val());
   176     }
   177   }
   178 };
   179 static SetHotnessClosure set_hotness_closure;
   182 int NMethodSweeper::hotness_counter_reset_val() {
   183   if (_hotness_counter_reset_val == 0) {
   184     _hotness_counter_reset_val = (ReservedCodeCacheSize < M) ? 1 : (ReservedCodeCacheSize / M) * 2;
   185   }
   186   return _hotness_counter_reset_val;
   187 }
   188 bool NMethodSweeper::sweep_in_progress() {
   189   return (_current != NULL);
   190 }
   192 // Scans the stacks of all Java threads and marks activations of not-entrant methods.
   193 // No need to synchronize access, since 'mark_active_nmethods' is always executed at a
   194 // safepoint.
   195 void NMethodSweeper::mark_active_nmethods() {
   196   assert(SafepointSynchronize::is_at_safepoint(), "must be executed at a safepoint");
   197   // If we do not want to reclaim not-entrant or zombie methods there is no need
   198   // to scan stacks
   199   if (!MethodFlushing) {
   200     return;
   201   }
   203   // Increase time so that we can estimate when to invoke the sweeper again.
   204   _time_counter++;
   206   // Check for restart
   207   assert(CodeCache::find_blob_unsafe(_current) == _current, "Sweeper nmethod cached state invalid");
   208   if (!sweep_in_progress()) {
   209     _seen = 0;
   210     _sweep_fractions_left = NmethodSweepFraction;
   211     _current = CodeCache::first_nmethod();
   212     _traversals += 1;
   213     _total_time_this_sweep = Tickspan();
   215     if (PrintMethodFlushing) {
   216       tty->print_cr("### Sweep: stack traversal %d", _traversals);
   217     }
   218     Threads::nmethods_do(&mark_activation_closure);
   220   } else {
   221     // Only set hotness counter
   222     Threads::nmethods_do(&set_hotness_closure);
   223   }
   225   OrderAccess::storestore();
   226 }
   227 /**
   228  * This function invokes the sweeper if at least one of the three conditions is met:
   229  *    (1) The code cache is getting full
   230  *    (2) There are sufficient state changes in/since the last sweep.
   231  *    (3) We have not been sweeping for 'some time'
   232  */
   233 void NMethodSweeper::possibly_sweep() {
   234   assert(JavaThread::current()->thread_state() == _thread_in_vm, "must run in vm mode");
   235   // Only compiler threads are allowed to sweep
   236   if (!MethodFlushing || !sweep_in_progress() || !Thread::current()->is_Compiler_thread()) {
   237     return;
   238   }
   240   // If there was no state change while nmethod sweeping, 'should_sweep' will be false.
   241   // This is one of the two places where should_sweep can be set to true. The general
   242   // idea is as follows: If there is enough free space in the code cache, there is no
   243   // need to invoke the sweeper. The following formula (which determines whether to invoke
   244   // the sweeper or not) depends on the assumption that for larger ReservedCodeCacheSizes
   245   // we need less frequent sweeps than for smaller ReservedCodecCacheSizes. Furthermore,
   246   // the formula considers how much space in the code cache is currently used. Here are
   247   // some examples that will (hopefully) help in understanding.
   248   //
   249   // Small ReservedCodeCacheSizes:  (e.g., < 16M) We invoke the sweeper every time, since
   250   //                                              the result of the division is 0. This
   251   //                                              keeps the used code cache size small
   252   //                                              (important for embedded Java)
   253   // Large ReservedCodeCacheSize :  (e.g., 256M + code cache is 10% full). The formula
   254   //                                              computes: (256 / 16) - 1 = 15
   255   //                                              As a result, we invoke the sweeper after
   256   //                                              15 invocations of 'mark_active_nmethods.
   257   // Large ReservedCodeCacheSize:   (e.g., 256M + code Cache is 90% full). The formula
   258   //                                              computes: (256 / 16) - 10 = 6.
   259   if (!_should_sweep) {
   260     int time_since_last_sweep = _time_counter - _last_sweep;
   261     double wait_until_next_sweep = (ReservedCodeCacheSize / (16 * M)) - time_since_last_sweep -
   262                                 CodeCache::reverse_free_ratio();
   264     if ((wait_until_next_sweep <= 0.0) || !CompileBroker::should_compile_new_jobs()) {
   265       _should_sweep = true;
   266     }
   267   }
   269   if (_should_sweep && _sweep_fractions_left > 0) {
   270     // Only one thread at a time will sweep
   271     jint old = Atomic::cmpxchg( 1, &_sweep_started, 0 );
   272     if (old != 0) {
   273       return;
   274     }
   275 #ifdef ASSERT
   276     if (LogSweeper && _records == NULL) {
   277       // Create the ring buffer for the logging code
   278       _records = NEW_C_HEAP_ARRAY(SweeperRecord, SweeperLogEntries, mtGC);
   279       memset(_records, 0, sizeof(SweeperRecord) * SweeperLogEntries);
   280     }
   281 #endif
   283     if (_sweep_fractions_left > 0) {
   284       sweep_code_cache();
   285       _sweep_fractions_left--;
   286     }
   288     // We are done with sweeping the code cache once.
   289     if (_sweep_fractions_left == 0) {
   290       _last_sweep = _time_counter;
   291       // Reset flag; temporarily disables sweeper
   292       _should_sweep = false;
   293       // If there was enough state change, 'possibly_enable_sweeper()'
   294       // sets '_should_sweep' to true
   295       possibly_enable_sweeper();
   296       // Reset _bytes_changed only if there was enough state change. _bytes_changed
   297       // can further increase by calls to 'report_state_change'.
   298       if (_should_sweep) {
   299         _bytes_changed = 0;
   300       }
   301     }
   302     _sweep_started = 0;
   303   }
   304 }
   306 void NMethodSweeper::sweep_code_cache() {
   307   Ticks sweep_start_counter = Ticks::now();
   309   _flushed_count                = 0;
   310   _zombified_count              = 0;
   311   _marked_for_reclamation_count = 0;
   313   if (PrintMethodFlushing && Verbose) {
   314     tty->print_cr("### Sweep at %d out of %d. Invocations left: %d", _seen, CodeCache::nof_nmethods(), _sweep_fractions_left);
   315   }
   317   if (!CompileBroker::should_compile_new_jobs()) {
   318     // If we have turned off compilations we might as well do full sweeps
   319     // in order to reach the clean state faster. Otherwise the sleeping compiler
   320     // threads will slow down sweeping.
   321     _sweep_fractions_left = 1;
   322   }
   324   // We want to visit all nmethods after NmethodSweepFraction
   325   // invocations so divide the remaining number of nmethods by the
   326   // remaining number of invocations.  This is only an estimate since
   327   // the number of nmethods changes during the sweep so the final
   328   // stage must iterate until it there are no more nmethods.
   329   int todo = (CodeCache::nof_nmethods() - _seen) / _sweep_fractions_left;
   330   int swept_count = 0;
   333   assert(!SafepointSynchronize::is_at_safepoint(), "should not be in safepoint when we get here");
   334   assert(!CodeCache_lock->owned_by_self(), "just checking");
   336   int freed_memory = 0;
   337   {
   338     MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
   340     // The last invocation iterates until there are no more nmethods
   341     for (int i = 0; (i < todo || _sweep_fractions_left == 1) && _current != NULL; i++) {
   342       swept_count++;
   343       if (SafepointSynchronize::is_synchronizing()) { // Safepoint request
   344         if (PrintMethodFlushing && Verbose) {
   345           tty->print_cr("### Sweep at %d out of %d, invocation: %d, yielding to safepoint", _seen, CodeCache::nof_nmethods(), _sweep_fractions_left);
   346         }
   347         MutexUnlockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
   349         assert(Thread::current()->is_Java_thread(), "should be java thread");
   350         JavaThread* thread = (JavaThread*)Thread::current();
   351         ThreadBlockInVM tbivm(thread);
   352         thread->java_suspend_self();
   353       }
   354       // Since we will give up the CodeCache_lock, always skip ahead
   355       // to the next nmethod.  Other blobs can be deleted by other
   356       // threads but nmethods are only reclaimed by the sweeper.
   357       nmethod* next = CodeCache::next_nmethod(_current);
   359       // Now ready to process nmethod and give up CodeCache_lock
   360       {
   361         MutexUnlockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
   362         freed_memory += process_nmethod(_current);
   363       }
   364       _seen++;
   365       _current = next;
   366     }
   367   }
   369   assert(_sweep_fractions_left > 1 || _current == NULL, "must have scanned the whole cache");
   371   const Ticks sweep_end_counter = Ticks::now();
   372   const Tickspan sweep_time = sweep_end_counter - sweep_start_counter;
   373   _total_time_sweeping  += sweep_time;
   374   _total_time_this_sweep += sweep_time;
   375   _peak_sweep_fraction_time = MAX2(sweep_time, _peak_sweep_fraction_time);
   376   _total_nof_methods_reclaimed += _flushed_count;
   378   EventSweepCodeCache event(UNTIMED);
   379   if (event.should_commit()) {
   380     event.set_starttime(sweep_start_counter);
   381     event.set_endtime(sweep_end_counter);
   382     event.set_sweepIndex(_traversals);
   383     event.set_sweepFractionIndex(NmethodSweepFraction - _sweep_fractions_left + 1);
   384     event.set_sweptCount(swept_count);
   385     event.set_flushedCount(_flushed_count);
   386     event.set_markedCount(_marked_for_reclamation_count);
   387     event.set_zombifiedCount(_zombified_count);
   388     event.commit();
   389   }
   391 #ifdef ASSERT
   392   if(PrintMethodFlushing) {
   393     tty->print_cr("### sweeper:      sweep time(%d): "
   394       INT64_FORMAT, _sweep_fractions_left, (jlong)sweep_time.value());
   395   }
   396 #endif
   398   if (_sweep_fractions_left == 1) {
   399     _peak_sweep_time = MAX2(_peak_sweep_time, _total_time_this_sweep);
   400     log_sweep("finished");
   401   }
   403   // Sweeper is the only case where memory is released, check here if it
   404   // is time to restart the compiler. Only checking if there is a certain
   405   // amount of free memory in the code cache might lead to re-enabling
   406   // compilation although no memory has been released. For example, there are
   407   // cases when compilation was disabled although there is 4MB (or more) free
   408   // memory in the code cache. The reason is code cache fragmentation. Therefore,
   409   // it only makes sense to re-enable compilation if we have actually freed memory.
   410   // Note that typically several kB are released for sweeping 16MB of the code
   411   // cache. As a result, 'freed_memory' > 0 to restart the compiler.
   412   if (!CompileBroker::should_compile_new_jobs() && (freed_memory > 0)) {
   413     CompileBroker::set_should_compile_new_jobs(CompileBroker::run_compilation);
   414     log_sweep("restart_compiler");
   415   }
   416 }
   418 /**
   419  * This function updates the sweeper statistics that keep track of nmethods
   420  * state changes. If there is 'enough' state change, the sweeper is invoked
   421  * as soon as possible. There can be data races on _bytes_changed. The data
   422  * races are benign, since it does not matter if we loose a couple of bytes.
   423  * In the worst case we call the sweeper a little later. Also, we are guaranteed
   424  * to invoke the sweeper if the code cache gets full.
   425  */
   426 void NMethodSweeper::report_state_change(nmethod* nm) {
   427   _bytes_changed += nm->total_size();
   428   possibly_enable_sweeper();
   429 }
   431 /**
   432  * Function determines if there was 'enough' state change in the code cache to invoke
   433  * the sweeper again. Currently, we determine 'enough' as more than 1% state change in
   434  * the code cache since the last sweep.
   435  */
   436 void NMethodSweeper::possibly_enable_sweeper() {
   437   double percent_changed = ((double)_bytes_changed / (double)ReservedCodeCacheSize) * 100;
   438   if (percent_changed > 1.0) {
   439     _should_sweep = true;
   440   }
   441 }
   443 class NMethodMarker: public StackObj {
   444  private:
   445   CompilerThread* _thread;
   446  public:
   447   NMethodMarker(nmethod* nm) {
   448     _thread = CompilerThread::current();
   449     if (!nm->is_zombie() && !nm->is_unloaded()) {
   450       // Only expose live nmethods for scanning
   451       _thread->set_scanned_nmethod(nm);
   452     }
   453   }
   454   ~NMethodMarker() {
   455     _thread->set_scanned_nmethod(NULL);
   456   }
   457 };
   459 void NMethodSweeper::release_nmethod(nmethod *nm) {
   460   // Clean up any CompiledICHolders
   461   {
   462     ResourceMark rm;
   463     MutexLocker ml_patch(CompiledIC_lock);
   464     RelocIterator iter(nm);
   465     while (iter.next()) {
   466       if (iter.type() == relocInfo::virtual_call_type) {
   467         CompiledIC::cleanup_call_site(iter.virtual_call_reloc());
   468       }
   469     }
   470   }
   472   MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
   473   nm->flush();
   474 }
   476 int NMethodSweeper::process_nmethod(nmethod *nm) {
   477   assert(!CodeCache_lock->owned_by_self(), "just checking");
   479   int freed_memory = 0;
   480   // Make sure this nmethod doesn't get unloaded during the scan,
   481   // since safepoints may happen during acquired below locks.
   482   NMethodMarker nmm(nm);
   483   SWEEP(nm);
   485   // Skip methods that are currently referenced by the VM
   486   if (nm->is_locked_by_vm()) {
   487     // But still remember to clean-up inline caches for alive nmethods
   488     if (nm->is_alive()) {
   489       // Clean inline caches that point to zombie/non-entrant methods
   490       MutexLocker cl(CompiledIC_lock);
   491       nm->cleanup_inline_caches();
   492       SWEEP(nm);
   493     }
   494     return freed_memory;
   495   }
   497   if (nm->is_zombie()) {
   498     // If it is the first time we see nmethod then we mark it. Otherwise,
   499     // we reclaim it. When we have seen a zombie method twice, we know that
   500     // there are no inline caches that refer to it.
   501     if (nm->is_marked_for_reclamation()) {
   502       assert(!nm->is_locked_by_vm(), "must not flush locked nmethods");
   503       if (PrintMethodFlushing && Verbose) {
   504         tty->print_cr("### Nmethod %3d/" PTR_FORMAT " (marked for reclamation) being flushed", nm->compile_id(), nm);
   505       }
   506       freed_memory = nm->total_size();
   507       release_nmethod(nm);
   508       _flushed_count++;
   509     } else {
   510       if (PrintMethodFlushing && Verbose) {
   511         tty->print_cr("### Nmethod %3d/" PTR_FORMAT " (zombie) being marked for reclamation", nm->compile_id(), nm);
   512       }
   513       nm->mark_for_reclamation();
   514       // Keep track of code cache state change
   515       _bytes_changed += nm->total_size();
   516       _marked_for_reclamation_count++;
   517       SWEEP(nm);
   518     }
   519   } else if (nm->is_not_entrant()) {
   520     // If there are no current activations of this method on the
   521     // stack we can safely convert it to a zombie method
   522     if (nm->can_not_entrant_be_converted()) {
   523       if (PrintMethodFlushing && Verbose) {
   524         tty->print_cr("### Nmethod %3d/" PTR_FORMAT " (not entrant) being made zombie", nm->compile_id(), nm);
   525       }
   526       // Code cache state change is tracked in make_zombie()
   527       nm->make_zombie();
   528       _zombified_count++;
   529       SWEEP(nm);
   530     } else {
   531       // Still alive, clean up its inline caches
   532       MutexLocker cl(CompiledIC_lock);
   533       nm->cleanup_inline_caches();
   534       SWEEP(nm);
   535     }
   536   } else if (nm->is_unloaded()) {
   537     // Unloaded code, just make it a zombie
   538     if (PrintMethodFlushing && Verbose) {
   539       tty->print_cr("### Nmethod %3d/" PTR_FORMAT " (unloaded) being made zombie", nm->compile_id(), nm);
   540     }
   541     if (nm->is_osr_method()) {
   542       SWEEP(nm);
   543       // No inline caches will ever point to osr methods, so we can just remove it
   544       freed_memory = nm->total_size();
   545       release_nmethod(nm);
   546       _flushed_count++;
   547     } else {
   548       // Code cache state change is tracked in make_zombie()
   549       nm->make_zombie();
   550       _zombified_count++;
   551       SWEEP(nm);
   552     }
   553   } else {
   554     if (UseCodeCacheFlushing) {
   555       if (!nm->is_locked_by_vm() && !nm->is_osr_method() && !nm->is_native_method()) {
   556         // Do not make native methods and OSR-methods not-entrant
   557         nm->dec_hotness_counter();
   558         // Get the initial value of the hotness counter. This value depends on the
   559         // ReservedCodeCacheSize
   560         int reset_val = hotness_counter_reset_val();
   561         int time_since_reset = reset_val - nm->hotness_counter();
   562         double threshold = -reset_val + (CodeCache::reverse_free_ratio() * NmethodSweepActivity);
   563         // The less free space in the code cache we have - the bigger reverse_free_ratio() is.
   564         // I.e., 'threshold' increases with lower available space in the code cache and a higher
   565         // NmethodSweepActivity. If the current hotness counter - which decreases from its initial
   566         // value until it is reset by stack walking - is smaller than the computed threshold, the
   567         // corresponding nmethod is considered for removal.
   568         if ((NmethodSweepActivity > 0) && (nm->hotness_counter() < threshold) && (time_since_reset > 10)) {
   569           // A method is marked as not-entrant if the method is
   570           // 1) 'old enough': nm->hotness_counter() < threshold
   571           // 2) The method was in_use for a minimum amount of time: (time_since_reset > 10)
   572           //    The second condition is necessary if we are dealing with very small code cache
   573           //    sizes (e.g., <10m) and the code cache size is too small to hold all hot methods.
   574           //    The second condition ensures that methods are not immediately made not-entrant
   575           //    after compilation.
   576           nm->make_not_entrant();
   577           // Code cache state change is tracked in make_not_entrant()
   578           if (PrintMethodFlushing && Verbose) {
   579             tty->print_cr("### Nmethod %d/" PTR_FORMAT "made not-entrant: hotness counter %d/%d threshold %f",
   580                           nm->compile_id(), nm, nm->hotness_counter(), reset_val, threshold);
   581           }
   582         }
   583       }
   584     }
   585     // Clean-up all inline caches that point to zombie/non-reentrant methods
   586     MutexLocker cl(CompiledIC_lock);
   587     nm->cleanup_inline_caches();
   588     SWEEP(nm);
   589   }
   590   return freed_memory;
   591 }
   593 // Print out some state information about the current sweep and the
   594 // state of the code cache if it's requested.
   595 void NMethodSweeper::log_sweep(const char* msg, const char* format, ...) {
   596   if (PrintMethodFlushing) {
   597     stringStream s;
   598     // Dump code cache state into a buffer before locking the tty,
   599     // because log_state() will use locks causing lock conflicts.
   600     CodeCache::log_state(&s);
   602     ttyLocker ttyl;
   603     tty->print("### sweeper: %s ", msg);
   604     if (format != NULL) {
   605       va_list ap;
   606       va_start(ap, format);
   607       tty->vprint(format, ap);
   608       va_end(ap);
   609     }
   610     tty->print_cr(s.as_string());
   611   }
   613   if (LogCompilation && (xtty != NULL)) {
   614     stringStream s;
   615     // Dump code cache state into a buffer before locking the tty,
   616     // because log_state() will use locks causing lock conflicts.
   617     CodeCache::log_state(&s);
   619     ttyLocker ttyl;
   620     xtty->begin_elem("sweeper state='%s' traversals='" INTX_FORMAT "' ", msg, (intx)traversal_count());
   621     if (format != NULL) {
   622       va_list ap;
   623       va_start(ap, format);
   624       xtty->vprint(format, ap);
   625       va_end(ap);
   626     }
   627     xtty->print(s.as_string());
   628     xtty->stamp();
   629     xtty->end_elem();
   630   }
   631 }

mercurial