src/share/vm/gc_implementation/g1/g1GCPhaseTimes.cpp

Tue, 31 Mar 2015 11:36:37 +0200

author
tschatzl
date
Tue, 31 Mar 2015 11:36:37 +0200
changeset 7673
c04f46b4abe4
parent 7660
3ca53859c3c7
child 7828
cbc7c4c9e11c
permissions
-rw-r--r--

8068036: assert(is_available(index)) failed in G1 cset
Summary: Some verification code iterated over the heap using the region mapping array. This is not allowed. Changed to use the regular iteration method with closure.
Reviewed-by: jwilhelm, brutisso

     1 /*
     2  * Copyright (c) 2013, 2014 Oracle and/or its affiliates. All rights reserved.
     3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
     4  *
     5  * This code is free software; you can redistribute it and/or modify it
     6  * under the terms of the GNU General Public License version 2 only, as
     7  * published by the Free Software Foundation.
     8  *
     9  * This code is distributed in the hope that it will be useful, but WITHOUT
    10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
    11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
    12  * version 2 for more details (a copy is included in the LICENSE file that
    13  * accompanied this code).
    14  *
    15  * You should have received a copy of the GNU General Public License version
    16  * 2 along with this work; if not, write to the Free Software Foundation,
    17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
    18  *
    19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
    20  * or visit www.oracle.com if you need additional information or have any
    21  * questions.
    22  *
    23  */
    25 #include "precompiled.hpp"
    26 #include "gc_implementation/g1/g1CollectedHeap.inline.hpp"
    27 #include "gc_implementation/g1/g1GCPhaseTimes.hpp"
    28 #include "gc_implementation/g1/g1Log.hpp"
    29 #include "gc_implementation/g1/g1StringDedup.hpp"
    30 #include "memory/allocation.hpp"
    31 #include "runtime/os.hpp"
    33 // Helper class for avoiding interleaved logging
    34 class LineBuffer: public StackObj {
    36 private:
    37   static const int BUFFER_LEN = 1024;
    38   static const int INDENT_CHARS = 3;
    39   char _buffer[BUFFER_LEN];
    40   int _indent_level;
    41   int _cur;
    43   void vappend(const char* format, va_list ap)  ATTRIBUTE_PRINTF(2, 0) {
    44     int res = vsnprintf(&_buffer[_cur], BUFFER_LEN - _cur, format, ap);
    45     if (res != -1) {
    46       _cur += res;
    47     } else {
    48       DEBUG_ONLY(warning("buffer too small in LineBuffer");)
    49       _buffer[BUFFER_LEN -1] = 0;
    50       _cur = BUFFER_LEN; // vsnprintf above should not add to _buffer if we are called again
    51     }
    52   }
    54 public:
    55   explicit LineBuffer(int indent_level): _indent_level(indent_level), _cur(0) {
    56     for (; (_cur < BUFFER_LEN && _cur < (_indent_level * INDENT_CHARS)); _cur++) {
    57       _buffer[_cur] = ' ';
    58     }
    59   }
    61 #ifndef PRODUCT
    62   ~LineBuffer() {
    63     assert(_cur == _indent_level * INDENT_CHARS, "pending data in buffer - append_and_print_cr() not called?");
    64   }
    65 #endif
    67   void append(const char* format, ...)  ATTRIBUTE_PRINTF(2, 3) {
    68     va_list ap;
    69     va_start(ap, format);
    70     vappend(format, ap);
    71     va_end(ap);
    72   }
    74   void print_cr() {
    75     gclog_or_tty->print_cr("%s", _buffer);
    76     _cur = _indent_level * INDENT_CHARS;
    77   }
    79   void append_and_print_cr(const char* format, ...)  ATTRIBUTE_PRINTF(2, 3) {
    80     va_list ap;
    81     va_start(ap, format);
    82     vappend(format, ap);
    83     va_end(ap);
    84     print_cr();
    85   }
    86 };
    88 template <class T>
    89 class WorkerDataArray  : public CHeapObj<mtGC> {
    90   friend class G1GCParPhasePrinter;
    91   T*          _data;
    92   uint        _length;
    93   const char* _title;
    94   bool        _print_sum;
    95   int         _log_level;
    96   uint        _indent_level;
    97   bool        _enabled;
    99   WorkerDataArray<size_t>* _thread_work_items;
   101   NOT_PRODUCT(T uninitialized();)
   103   // We are caching the sum and average to only have to calculate them once.
   104   // This is not done in an MT-safe way. It is intended to allow single
   105   // threaded code to call sum() and average() multiple times in any order
   106   // without having to worry about the cost.
   107   bool   _has_new_data;
   108   T      _sum;
   109   T      _min;
   110   T      _max;
   111   double _average;
   113  public:
   114   WorkerDataArray(uint length, const char* title, bool print_sum, int log_level, uint indent_level) :
   115     _title(title), _length(0), _print_sum(print_sum), _log_level(log_level), _indent_level(indent_level),
   116     _has_new_data(true), _thread_work_items(NULL), _enabled(true) {
   117     assert(length > 0, "Must have some workers to store data for");
   118     _length = length;
   119     _data = NEW_C_HEAP_ARRAY(T, _length, mtGC);
   120   }
   122   ~WorkerDataArray() {
   123     FREE_C_HEAP_ARRAY(T, _data, mtGC);
   124   }
   126   void link_thread_work_items(WorkerDataArray<size_t>* thread_work_items) {
   127     _thread_work_items = thread_work_items;
   128   }
   130   WorkerDataArray<size_t>* thread_work_items() { return _thread_work_items; }
   132   void set(uint worker_i, T value) {
   133     assert(worker_i < _length, err_msg("Worker %d is greater than max: %d", worker_i, _length));
   134     assert(_data[worker_i] == WorkerDataArray<T>::uninitialized(), err_msg("Overwriting data for worker %d in %s", worker_i, _title));
   135     _data[worker_i] = value;
   136     _has_new_data = true;
   137   }
   139   void set_thread_work_item(uint worker_i, size_t value) {
   140     assert(_thread_work_items != NULL, "No sub count");
   141     _thread_work_items->set(worker_i, value);
   142   }
   144   T get(uint worker_i) {
   145     assert(worker_i < _length, err_msg("Worker %d is greater than max: %d", worker_i, _length));
   146     assert(_data[worker_i] != WorkerDataArray<T>::uninitialized(), err_msg("No data added for worker %d", worker_i));
   147     return _data[worker_i];
   148   }
   150   void add(uint worker_i, T value) {
   151     assert(worker_i < _length, err_msg("Worker %d is greater than max: %d", worker_i, _length));
   152     assert(_data[worker_i] != WorkerDataArray<T>::uninitialized(), err_msg("No data to add to for worker %d", worker_i));
   153     _data[worker_i] += value;
   154     _has_new_data = true;
   155   }
   157   double average(){
   158     calculate_totals();
   159     return _average;
   160   }
   162   T sum() {
   163     calculate_totals();
   164     return _sum;
   165   }
   167   T minimum() {
   168     calculate_totals();
   169     return _min;
   170   }
   172   T maximum() {
   173     calculate_totals();
   174     return _max;
   175   }
   177   void reset() PRODUCT_RETURN;
   178   void verify() PRODUCT_RETURN;
   180   void set_enabled(bool enabled) { _enabled = enabled; }
   182   int log_level() { return _log_level;  }
   184  private:
   186   void calculate_totals(){
   187     if (!_has_new_data) {
   188       return;
   189     }
   191     _sum = (T)0;
   192     _min = _data[0];
   193     _max = _min;
   194     for (uint i = 0; i < _length; ++i) {
   195       T val = _data[i];
   196       _sum += val;
   197       _min = MIN2(_min, val);
   198       _max = MAX2(_max, val);
   199     }
   200     _average = (double)_sum / (double)_length;
   201     _has_new_data = false;
   202   }
   203 };
   206 #ifndef PRODUCT
   208 template <>
   209 size_t WorkerDataArray<size_t>::uninitialized() {
   210   return (size_t)-1;
   211 }
   213 template <>
   214 double WorkerDataArray<double>::uninitialized() {
   215   return -1.0;
   216 }
   218 template <class T>
   219 void WorkerDataArray<T>::reset() {
   220   for (uint i = 0; i < _length; i++) {
   221     _data[i] = WorkerDataArray<T>::uninitialized();
   222   }
   223   if (_thread_work_items != NULL) {
   224     _thread_work_items->reset();
   225   }
   226 }
   228 template <class T>
   229 void WorkerDataArray<T>::verify() {
   230   if (!_enabled) {
   231     return;
   232   }
   234   for (uint i = 0; i < _length; i++) {
   235     assert(_data[i] != WorkerDataArray<T>::uninitialized(),
   236         err_msg("Invalid data for worker %u in '%s'", i, _title));
   237   }
   238   if (_thread_work_items != NULL) {
   239     _thread_work_items->verify();
   240   }
   241 }
   243 #endif
   245 G1GCPhaseTimes::G1GCPhaseTimes(uint max_gc_threads) :
   246   _max_gc_threads(max_gc_threads)
   247 {
   248   assert(max_gc_threads > 0, "Must have some GC threads");
   250   _gc_par_phases[GCWorkerStart] = new WorkerDataArray<double>(max_gc_threads, "GC Worker Start (ms)", false, G1Log::LevelFiner, 2);
   251   _gc_par_phases[ExtRootScan] = new WorkerDataArray<double>(max_gc_threads, "Ext Root Scanning (ms)", true, G1Log::LevelFiner, 2);
   253   // Root scanning phases
   254   _gc_par_phases[ThreadRoots] = new WorkerDataArray<double>(max_gc_threads, "Thread Roots (ms)", true, G1Log::LevelFinest, 3);
   255   _gc_par_phases[StringTableRoots] = new WorkerDataArray<double>(max_gc_threads, "StringTable Roots (ms)", true, G1Log::LevelFinest, 3);
   256   _gc_par_phases[UniverseRoots] = new WorkerDataArray<double>(max_gc_threads, "Universe Roots (ms)", true, G1Log::LevelFinest, 3);
   257   _gc_par_phases[JNIRoots] = new WorkerDataArray<double>(max_gc_threads, "JNI Handles Roots (ms)", true, G1Log::LevelFinest, 3);
   258   _gc_par_phases[ObjectSynchronizerRoots] = new WorkerDataArray<double>(max_gc_threads, "ObjectSynchronizer Roots (ms)", true, G1Log::LevelFinest, 3);
   259   _gc_par_phases[FlatProfilerRoots] = new WorkerDataArray<double>(max_gc_threads, "FlatProfiler Roots (ms)", true, G1Log::LevelFinest, 3);
   260   _gc_par_phases[ManagementRoots] = new WorkerDataArray<double>(max_gc_threads, "Management Roots (ms)", true, G1Log::LevelFinest, 3);
   261   _gc_par_phases[SystemDictionaryRoots] = new WorkerDataArray<double>(max_gc_threads, "SystemDictionary Roots (ms)", true, G1Log::LevelFinest, 3);
   262   _gc_par_phases[CLDGRoots] = new WorkerDataArray<double>(max_gc_threads, "CLDG Roots (ms)", true, G1Log::LevelFinest, 3);
   263   _gc_par_phases[JVMTIRoots] = new WorkerDataArray<double>(max_gc_threads, "JVMTI Roots (ms)", true, G1Log::LevelFinest, 3);
   264   _gc_par_phases[CodeCacheRoots] = new WorkerDataArray<double>(max_gc_threads, "CodeCache Roots (ms)", true, G1Log::LevelFinest, 3);
   265   _gc_par_phases[CMRefRoots] = new WorkerDataArray<double>(max_gc_threads, "CM RefProcessor Roots (ms)", true, G1Log::LevelFinest, 3);
   266   _gc_par_phases[WaitForStrongCLD] = new WorkerDataArray<double>(max_gc_threads, "Wait For Strong CLD (ms)", true, G1Log::LevelFinest, 3);
   267   _gc_par_phases[WeakCLDRoots] = new WorkerDataArray<double>(max_gc_threads, "Weak CLD Roots (ms)", true, G1Log::LevelFinest, 3);
   268   _gc_par_phases[SATBFiltering] = new WorkerDataArray<double>(max_gc_threads, "SATB Filtering (ms)", true, G1Log::LevelFinest, 3);
   270   _gc_par_phases[UpdateRS] = new WorkerDataArray<double>(max_gc_threads, "Update RS (ms)", true, G1Log::LevelFiner, 2);
   271   _gc_par_phases[ScanRS] = new WorkerDataArray<double>(max_gc_threads, "Scan RS (ms)", true, G1Log::LevelFiner, 2);
   272   _gc_par_phases[CodeRoots] = new WorkerDataArray<double>(max_gc_threads, "Code Root Scanning (ms)", true, G1Log::LevelFiner, 2);
   273   _gc_par_phases[ObjCopy] = new WorkerDataArray<double>(max_gc_threads, "Object Copy (ms)", true, G1Log::LevelFiner, 2);
   274   _gc_par_phases[Termination] = new WorkerDataArray<double>(max_gc_threads, "Termination (ms)", true, G1Log::LevelFiner, 2);
   275   _gc_par_phases[GCWorkerTotal] = new WorkerDataArray<double>(max_gc_threads, "GC Worker Total (ms)", true, G1Log::LevelFiner, 2);
   276   _gc_par_phases[GCWorkerEnd] = new WorkerDataArray<double>(max_gc_threads, "GC Worker End (ms)", false, G1Log::LevelFiner, 2);
   277   _gc_par_phases[Other] = new WorkerDataArray<double>(max_gc_threads, "GC Worker Other (ms)", true, G1Log::LevelFiner, 2);
   279   _update_rs_processed_buffers = new WorkerDataArray<size_t>(max_gc_threads, "Processed Buffers", true, G1Log::LevelFiner, 3);
   280   _gc_par_phases[UpdateRS]->link_thread_work_items(_update_rs_processed_buffers);
   282   _termination_attempts = new WorkerDataArray<size_t>(max_gc_threads, "Termination Attempts", true, G1Log::LevelFinest, 3);
   283   _gc_par_phases[Termination]->link_thread_work_items(_termination_attempts);
   285   _gc_par_phases[StringDedupQueueFixup] = new WorkerDataArray<double>(max_gc_threads, "Queue Fixup (ms)", true, G1Log::LevelFiner, 2);
   286   _gc_par_phases[StringDedupTableFixup] = new WorkerDataArray<double>(max_gc_threads, "Table Fixup (ms)", true, G1Log::LevelFiner, 2);
   288   _gc_par_phases[RedirtyCards] = new WorkerDataArray<double>(max_gc_threads, "Parallel Redirty", true, G1Log::LevelFinest, 3);
   289   _redirtied_cards = new WorkerDataArray<size_t>(max_gc_threads, "Redirtied Cards", true, G1Log::LevelFinest, 3);
   290   _gc_par_phases[RedirtyCards]->link_thread_work_items(_redirtied_cards);
   291 }
   293 void G1GCPhaseTimes::note_gc_start(uint active_gc_threads, bool mark_in_progress) {
   294   assert(active_gc_threads > 0, "The number of threads must be > 0");
   295   assert(active_gc_threads <= _max_gc_threads, "The number of active threads must be <= the max number of threads");
   296   _active_gc_threads = active_gc_threads;
   298   for (int i = 0; i < GCParPhasesSentinel; i++) {
   299     _gc_par_phases[i]->reset();
   300   }
   302   _gc_par_phases[StringDedupQueueFixup]->set_enabled(G1StringDedup::is_enabled());
   303   _gc_par_phases[StringDedupTableFixup]->set_enabled(G1StringDedup::is_enabled());
   304 }
   306 void G1GCPhaseTimes::note_gc_end() {
   307   for (uint i = 0; i < _active_gc_threads; i++) {
   308     double worker_time = _gc_par_phases[GCWorkerEnd]->get(i) - _gc_par_phases[GCWorkerStart]->get(i);
   309     record_time_secs(GCWorkerTotal, i , worker_time);
   311     double worker_known_time =
   312         _gc_par_phases[ExtRootScan]->get(i) +
   313         _gc_par_phases[SATBFiltering]->get(i) +
   314         _gc_par_phases[UpdateRS]->get(i) +
   315         _gc_par_phases[ScanRS]->get(i) +
   316         _gc_par_phases[CodeRoots]->get(i) +
   317         _gc_par_phases[ObjCopy]->get(i) +
   318         _gc_par_phases[Termination]->get(i);
   320     record_time_secs(Other, i, worker_time - worker_known_time);
   321   }
   323   for (int i = 0; i < GCParPhasesSentinel; i++) {
   324     _gc_par_phases[i]->verify();
   325   }
   326 }
   328 void G1GCPhaseTimes::print_stats(int level, const char* str, double value) {
   329   LineBuffer(level).append_and_print_cr("[%s: %.1lf ms]", str, value);
   330 }
   332 void G1GCPhaseTimes::print_stats(int level, const char* str, size_t value) {
   333   LineBuffer(level).append_and_print_cr("[%s: "SIZE_FORMAT"]", str, value);
   334 }
   336 void G1GCPhaseTimes::print_stats(int level, const char* str, double value, uint workers) {
   337   LineBuffer(level).append_and_print_cr("[%s: %.1lf ms, GC Workers: %u]", str, value, workers);
   338 }
   340 double G1GCPhaseTimes::accounted_time_ms() {
   341     // Subtract the root region scanning wait time. It's initialized to
   342     // zero at the start of the pause.
   343     double misc_time_ms = _root_region_scan_wait_time_ms;
   345     misc_time_ms += _cur_collection_par_time_ms;
   347     // Now subtract the time taken to fix up roots in generated code
   348     misc_time_ms += _cur_collection_code_root_fixup_time_ms;
   350     // Strong code root purge time
   351     misc_time_ms += _cur_strong_code_root_purge_time_ms;
   353     if (G1StringDedup::is_enabled()) {
   354       // String dedup fixup time
   355       misc_time_ms += _cur_string_dedup_fixup_time_ms;
   356     }
   358     // Subtract the time taken to clean the card table from the
   359     // current value of "other time"
   360     misc_time_ms += _cur_clear_ct_time_ms;
   362     return misc_time_ms;
   363 }
   365 // record the time a phase took in seconds
   366 void G1GCPhaseTimes::record_time_secs(GCParPhases phase, uint worker_i, double secs) {
   367   _gc_par_phases[phase]->set(worker_i, secs);
   368 }
   370 // add a number of seconds to a phase
   371 void G1GCPhaseTimes::add_time_secs(GCParPhases phase, uint worker_i, double secs) {
   372   _gc_par_phases[phase]->add(worker_i, secs);
   373 }
   375 void G1GCPhaseTimes::record_thread_work_item(GCParPhases phase, uint worker_i, size_t count) {
   376   _gc_par_phases[phase]->set_thread_work_item(worker_i, count);
   377 }
   379 // return the average time for a phase in milliseconds
   380 double G1GCPhaseTimes::average_time_ms(GCParPhases phase) {
   381   return _gc_par_phases[phase]->average() * 1000.0;
   382 }
   384 double G1GCPhaseTimes::get_time_ms(GCParPhases phase, uint worker_i) {
   385   return _gc_par_phases[phase]->get(worker_i) * 1000.0;
   386 }
   388 double G1GCPhaseTimes::sum_time_ms(GCParPhases phase) {
   389   return _gc_par_phases[phase]->sum() * 1000.0;
   390 }
   392 double G1GCPhaseTimes::min_time_ms(GCParPhases phase) {
   393   return _gc_par_phases[phase]->minimum() * 1000.0;
   394 }
   396 double G1GCPhaseTimes::max_time_ms(GCParPhases phase) {
   397   return _gc_par_phases[phase]->maximum() * 1000.0;
   398 }
   400 size_t G1GCPhaseTimes::get_thread_work_item(GCParPhases phase, uint worker_i) {
   401   assert(_gc_par_phases[phase]->thread_work_items() != NULL, "No sub count");
   402   return _gc_par_phases[phase]->thread_work_items()->get(worker_i);
   403 }
   405 size_t G1GCPhaseTimes::sum_thread_work_items(GCParPhases phase) {
   406   assert(_gc_par_phases[phase]->thread_work_items() != NULL, "No sub count");
   407   return _gc_par_phases[phase]->thread_work_items()->sum();
   408 }
   410 double G1GCPhaseTimes::average_thread_work_items(GCParPhases phase) {
   411   assert(_gc_par_phases[phase]->thread_work_items() != NULL, "No sub count");
   412   return _gc_par_phases[phase]->thread_work_items()->average();
   413 }
   415 size_t G1GCPhaseTimes::min_thread_work_items(GCParPhases phase) {
   416   assert(_gc_par_phases[phase]->thread_work_items() != NULL, "No sub count");
   417   return _gc_par_phases[phase]->thread_work_items()->minimum();
   418 }
   420 size_t G1GCPhaseTimes::max_thread_work_items(GCParPhases phase) {
   421   assert(_gc_par_phases[phase]->thread_work_items() != NULL, "No sub count");
   422   return _gc_par_phases[phase]->thread_work_items()->maximum();
   423 }
   425 class G1GCParPhasePrinter : public StackObj {
   426   G1GCPhaseTimes* _phase_times;
   427  public:
   428   G1GCParPhasePrinter(G1GCPhaseTimes* phase_times) : _phase_times(phase_times) {}
   430   void print(G1GCPhaseTimes::GCParPhases phase_id) {
   431     WorkerDataArray<double>* phase = _phase_times->_gc_par_phases[phase_id];
   433     if (phase->_log_level > G1Log::level() || !phase->_enabled) {
   434       return;
   435     }
   437     if (phase->_length == 1) {
   438       print_single_length(phase_id, phase);
   439     } else {
   440       print_multi_length(phase_id, phase);
   441     }
   442   }
   444  private:
   446   void print_single_length(G1GCPhaseTimes::GCParPhases phase_id, WorkerDataArray<double>* phase) {
   447     // No need for min, max, average and sum for only one worker
   448     LineBuffer buf(phase->_indent_level);
   449     buf.append_and_print_cr("[%s:  %.1lf]", phase->_title, _phase_times->get_time_ms(phase_id, 0));
   451     if (phase->_thread_work_items != NULL) {
   452       LineBuffer buf2(phase->_thread_work_items->_indent_level);
   453       buf2.append_and_print_cr("[%s:  "SIZE_FORMAT"]", phase->_thread_work_items->_title, _phase_times->sum_thread_work_items(phase_id));
   454     }
   455   }
   457   void print_time_values(LineBuffer& buf, G1GCPhaseTimes::GCParPhases phase_id, WorkerDataArray<double>* phase) {
   458     for (uint i = 0; i < phase->_length; ++i) {
   459       buf.append("  %.1lf", _phase_times->get_time_ms(phase_id, i));
   460     }
   461     buf.print_cr();
   462   }
   464   void print_count_values(LineBuffer& buf, G1GCPhaseTimes::GCParPhases phase_id, WorkerDataArray<size_t>* thread_work_items) {
   465     for (uint i = 0; i < thread_work_items->_length; ++i) {
   466       buf.append("  " SIZE_FORMAT, _phase_times->get_thread_work_item(phase_id, i));
   467     }
   468     buf.print_cr();
   469   }
   471   void print_thread_work_items(G1GCPhaseTimes::GCParPhases phase_id, WorkerDataArray<size_t>* thread_work_items) {
   472     LineBuffer buf(thread_work_items->_indent_level);
   473     buf.append("[%s:", thread_work_items->_title);
   475     if (G1Log::finest()) {
   476       print_count_values(buf, phase_id, thread_work_items);
   477     }
   479     assert(thread_work_items->_print_sum, err_msg("%s does not have print sum true even though it is a count", thread_work_items->_title));
   481     buf.append_and_print_cr(" Min: " SIZE_FORMAT ", Avg: %.1lf, Max: " SIZE_FORMAT ", Diff: " SIZE_FORMAT ", Sum: " SIZE_FORMAT "]",
   482         _phase_times->min_thread_work_items(phase_id), _phase_times->average_thread_work_items(phase_id), _phase_times->max_thread_work_items(phase_id),
   483         _phase_times->max_thread_work_items(phase_id) - _phase_times->min_thread_work_items(phase_id), _phase_times->sum_thread_work_items(phase_id));
   484   }
   486   void print_multi_length(G1GCPhaseTimes::GCParPhases phase_id, WorkerDataArray<double>* phase) {
   487     LineBuffer buf(phase->_indent_level);
   488     buf.append("[%s:", phase->_title);
   490     if (G1Log::finest()) {
   491       print_time_values(buf, phase_id, phase);
   492     }
   494     buf.append(" Min: %.1lf, Avg: %.1lf, Max: %.1lf, Diff: %.1lf",
   495         _phase_times->min_time_ms(phase_id), _phase_times->average_time_ms(phase_id), _phase_times->max_time_ms(phase_id),
   496         _phase_times->max_time_ms(phase_id) - _phase_times->min_time_ms(phase_id));
   498     if (phase->_print_sum) {
   499       // for things like the start and end times the sum is not
   500       // that relevant
   501       buf.append(", Sum: %.1lf", _phase_times->sum_time_ms(phase_id));
   502     }
   504     buf.append_and_print_cr("]");
   506     if (phase->_thread_work_items != NULL) {
   507       print_thread_work_items(phase_id, phase->_thread_work_items);
   508     }
   509   }
   510 };
   512 void G1GCPhaseTimes::print(double pause_time_sec) {
   513   G1GCParPhasePrinter par_phase_printer(this);
   515   if (_root_region_scan_wait_time_ms > 0.0) {
   516     print_stats(1, "Root Region Scan Waiting", _root_region_scan_wait_time_ms);
   517   }
   519   print_stats(1, "Parallel Time", _cur_collection_par_time_ms, _active_gc_threads);
   520   for (int i = 0; i <= GCMainParPhasesLast; i++) {
   521     par_phase_printer.print((GCParPhases) i);
   522   }
   524   print_stats(1, "Code Root Fixup", _cur_collection_code_root_fixup_time_ms);
   525   print_stats(1, "Code Root Purge", _cur_strong_code_root_purge_time_ms);
   526   if (G1StringDedup::is_enabled()) {
   527     print_stats(1, "String Dedup Fixup", _cur_string_dedup_fixup_time_ms, _active_gc_threads);
   528     for (int i = StringDedupPhasesFirst; i <= StringDedupPhasesLast; i++) {
   529       par_phase_printer.print((GCParPhases) i);
   530     }
   531   }
   532   print_stats(1, "Clear CT", _cur_clear_ct_time_ms);
   533   double misc_time_ms = pause_time_sec * MILLIUNITS - accounted_time_ms();
   534   print_stats(1, "Other", misc_time_ms);
   535   if (_cur_verify_before_time_ms > 0.0) {
   536     print_stats(2, "Verify Before", _cur_verify_before_time_ms);
   537   }
   538   if (G1CollectedHeap::heap()->evacuation_failed()) {
   539     double evac_fail_handling = _cur_evac_fail_recalc_used + _cur_evac_fail_remove_self_forwards +
   540       _cur_evac_fail_restore_remsets;
   541     print_stats(2, "Evacuation Failure", evac_fail_handling);
   542     if (G1Log::finest()) {
   543       print_stats(3, "Recalculate Used", _cur_evac_fail_recalc_used);
   544       print_stats(3, "Remove Self Forwards", _cur_evac_fail_remove_self_forwards);
   545       print_stats(3, "Restore RemSet", _cur_evac_fail_restore_remsets);
   546     }
   547   }
   548   print_stats(2, "Choose CSet",
   549     (_recorded_young_cset_choice_time_ms +
   550     _recorded_non_young_cset_choice_time_ms));
   551   print_stats(2, "Ref Proc", _cur_ref_proc_time_ms);
   552   print_stats(2, "Ref Enq", _cur_ref_enq_time_ms);
   553   print_stats(2, "Redirty Cards", _recorded_redirty_logged_cards_time_ms);
   554   par_phase_printer.print(RedirtyCards);
   555   if (G1ReclaimDeadHumongousObjectsAtYoungGC) {
   556     print_stats(2, "Humongous Reclaim", _cur_fast_reclaim_humongous_time_ms);
   557     if (G1Log::finest()) {
   558       print_stats(3, "Humongous Total", _cur_fast_reclaim_humongous_total);
   559       print_stats(3, "Humongous Candidate", _cur_fast_reclaim_humongous_candidates);
   560       print_stats(3, "Humongous Reclaimed", _cur_fast_reclaim_humongous_reclaimed);
   561     }
   562   }
   563   print_stats(2, "Free CSet",
   564     (_recorded_young_free_cset_time_ms +
   565     _recorded_non_young_free_cset_time_ms));
   566   if (G1Log::finest()) {
   567     print_stats(3, "Young Free CSet", _recorded_young_free_cset_time_ms);
   568     print_stats(3, "Non-Young Free CSet", _recorded_non_young_free_cset_time_ms);
   569   }
   570   if (_cur_verify_after_time_ms > 0.0) {
   571     print_stats(2, "Verify After", _cur_verify_after_time_ms);
   572   }
   573 }
   575 G1GCParPhaseTimesTracker::G1GCParPhaseTimesTracker(G1GCPhaseTimes* phase_times, G1GCPhaseTimes::GCParPhases phase, uint worker_id) :
   576     _phase_times(phase_times), _phase(phase), _worker_id(worker_id) {
   577   if (_phase_times != NULL) {
   578     _start_time = os::elapsedTime();
   579   }
   580 }
   582 G1GCParPhaseTimesTracker::~G1GCParPhaseTimesTracker() {
   583   if (_phase_times != NULL) {
   584     _phase_times->record_time_secs(_phase, _worker_id, os::elapsedTime() - _start_time);
   585   }
   586 }

mercurial