src/share/vm/gc_implementation/g1/concurrentMark.cpp

Sat, 06 Oct 2012 01:17:44 -0700

author
johnc
date
Sat, 06 Oct 2012 01:17:44 -0700
changeset 4173
8a5ea0a9ccc4
parent 4123
988bf00cc564
child 4176
4202510ee0fe
permissions
-rw-r--r--

7127708: G1: change task num types from int to uint in concurrent mark
Summary: Change the type of various task num fields, parameters etc to unsigned and rename them to be more consistent with the other collectors. Code changes were also reviewed by Vitaly Davidovich.
Reviewed-by: johnc
Contributed-by: Kaushik Srenevasan <kaushik@twitter.com>

ysr@777 1 /*
johnc@3412 2 * Copyright (c) 2001, 2012, Oracle and/or its affiliates. All rights reserved.
ysr@777 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
ysr@777 4 *
ysr@777 5 * This code is free software; you can redistribute it and/or modify it
ysr@777 6 * under the terms of the GNU General Public License version 2 only, as
ysr@777 7 * published by the Free Software Foundation.
ysr@777 8 *
ysr@777 9 * This code is distributed in the hope that it will be useful, but WITHOUT
ysr@777 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
ysr@777 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
ysr@777 12 * version 2 for more details (a copy is included in the LICENSE file that
ysr@777 13 * accompanied this code).
ysr@777 14 *
ysr@777 15 * You should have received a copy of the GNU General Public License version
ysr@777 16 * 2 along with this work; if not, write to the Free Software Foundation,
ysr@777 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
ysr@777 18 *
trims@1907 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
trims@1907 20 * or visit www.oracle.com if you need additional information or have any
trims@1907 21 * questions.
ysr@777 22 *
ysr@777 23 */
ysr@777 24
stefank@2314 25 #include "precompiled.hpp"
stefank@2314 26 #include "classfile/symbolTable.hpp"
tonyp@2968 27 #include "gc_implementation/g1/concurrentMark.inline.hpp"
stefank@2314 28 #include "gc_implementation/g1/concurrentMarkThread.inline.hpp"
stefank@2314 29 #include "gc_implementation/g1/g1CollectedHeap.inline.hpp"
stefank@2314 30 #include "gc_implementation/g1/g1CollectorPolicy.hpp"
tonyp@3114 31 #include "gc_implementation/g1/g1ErgoVerbose.hpp"
brutisso@3710 32 #include "gc_implementation/g1/g1Log.hpp"
tonyp@2968 33 #include "gc_implementation/g1/g1OopClosures.inline.hpp"
stefank@2314 34 #include "gc_implementation/g1/g1RemSet.hpp"
tonyp@3416 35 #include "gc_implementation/g1/heapRegion.inline.hpp"
stefank@2314 36 #include "gc_implementation/g1/heapRegionRemSet.hpp"
stefank@2314 37 #include "gc_implementation/g1/heapRegionSeq.inline.hpp"
kamg@2445 38 #include "gc_implementation/shared/vmGCOperations.hpp"
stefank@2314 39 #include "memory/genOopClosures.inline.hpp"
stefank@2314 40 #include "memory/referencePolicy.hpp"
stefank@2314 41 #include "memory/resourceArea.hpp"
stefank@2314 42 #include "oops/oop.inline.hpp"
stefank@2314 43 #include "runtime/handles.inline.hpp"
stefank@2314 44 #include "runtime/java.hpp"
zgu@3900 45 #include "services/memTracker.hpp"
ysr@777 46
brutisso@3455 47 // Concurrent marking bit map wrapper
ysr@777 48
johnc@3292 49 CMBitMapRO::CMBitMapRO(ReservedSpace rs, int shifter) :
ysr@777 50 _bm((uintptr_t*)NULL,0),
ysr@777 51 _shifter(shifter) {
ysr@777 52 _bmStartWord = (HeapWord*)(rs.base());
ysr@777 53 _bmWordSize = rs.size()/HeapWordSize; // rs.size() is in bytes
ysr@777 54 ReservedSpace brs(ReservedSpace::allocation_align_size_up(
ysr@777 55 (_bmWordSize >> (_shifter + LogBitsPerByte)) + 1));
ysr@777 56
zgu@3900 57 MemTracker::record_virtual_memory_type((address)brs.base(), mtGC);
zgu@3900 58
brutisso@3455 59 guarantee(brs.is_reserved(), "couldn't allocate concurrent marking bit map");
ysr@777 60 // For now we'll just commit all of the bit map up fromt.
ysr@777 61 // Later on we'll try to be more parsimonious with swap.
ysr@777 62 guarantee(_virtual_space.initialize(brs, brs.size()),
brutisso@3455 63 "couldn't reseve backing store for concurrent marking bit map");
ysr@777 64 assert(_virtual_space.committed_size() == brs.size(),
brutisso@3455 65 "didn't reserve backing store for all of concurrent marking bit map?");
ysr@777 66 _bm.set_map((uintptr_t*)_virtual_space.low());
ysr@777 67 assert(_virtual_space.committed_size() << (_shifter + LogBitsPerByte) >=
ysr@777 68 _bmWordSize, "inconsistency in bit map sizing");
ysr@777 69 _bm.set_size(_bmWordSize >> _shifter);
ysr@777 70 }
ysr@777 71
ysr@777 72 HeapWord* CMBitMapRO::getNextMarkedWordAddress(HeapWord* addr,
ysr@777 73 HeapWord* limit) const {
ysr@777 74 // First we must round addr *up* to a possible object boundary.
ysr@777 75 addr = (HeapWord*)align_size_up((intptr_t)addr,
ysr@777 76 HeapWordSize << _shifter);
ysr@777 77 size_t addrOffset = heapWordToOffset(addr);
tonyp@2973 78 if (limit == NULL) {
tonyp@2973 79 limit = _bmStartWord + _bmWordSize;
tonyp@2973 80 }
ysr@777 81 size_t limitOffset = heapWordToOffset(limit);
ysr@777 82 size_t nextOffset = _bm.get_next_one_offset(addrOffset, limitOffset);
ysr@777 83 HeapWord* nextAddr = offsetToHeapWord(nextOffset);
ysr@777 84 assert(nextAddr >= addr, "get_next_one postcondition");
ysr@777 85 assert(nextAddr == limit || isMarked(nextAddr),
ysr@777 86 "get_next_one postcondition");
ysr@777 87 return nextAddr;
ysr@777 88 }
ysr@777 89
ysr@777 90 HeapWord* CMBitMapRO::getNextUnmarkedWordAddress(HeapWord* addr,
ysr@777 91 HeapWord* limit) const {
ysr@777 92 size_t addrOffset = heapWordToOffset(addr);
tonyp@2973 93 if (limit == NULL) {
tonyp@2973 94 limit = _bmStartWord + _bmWordSize;
tonyp@2973 95 }
ysr@777 96 size_t limitOffset = heapWordToOffset(limit);
ysr@777 97 size_t nextOffset = _bm.get_next_zero_offset(addrOffset, limitOffset);
ysr@777 98 HeapWord* nextAddr = offsetToHeapWord(nextOffset);
ysr@777 99 assert(nextAddr >= addr, "get_next_one postcondition");
ysr@777 100 assert(nextAddr == limit || !isMarked(nextAddr),
ysr@777 101 "get_next_one postcondition");
ysr@777 102 return nextAddr;
ysr@777 103 }
ysr@777 104
ysr@777 105 int CMBitMapRO::heapWordDiffToOffsetDiff(size_t diff) const {
ysr@777 106 assert((diff & ((1 << _shifter) - 1)) == 0, "argument check");
ysr@777 107 return (int) (diff >> _shifter);
ysr@777 108 }
ysr@777 109
ysr@777 110 #ifndef PRODUCT
ysr@777 111 bool CMBitMapRO::covers(ReservedSpace rs) const {
ysr@777 112 // assert(_bm.map() == _virtual_space.low(), "map inconsistency");
brutisso@4061 113 assert(((size_t)_bm.size() * ((size_t)1 << _shifter)) == _bmWordSize,
ysr@777 114 "size inconsistency");
ysr@777 115 return _bmStartWord == (HeapWord*)(rs.base()) &&
ysr@777 116 _bmWordSize == rs.size()>>LogHeapWordSize;
ysr@777 117 }
ysr@777 118 #endif
ysr@777 119
ysr@777 120 void CMBitMap::clearAll() {
ysr@777 121 _bm.clear();
ysr@777 122 return;
ysr@777 123 }
ysr@777 124
ysr@777 125 void CMBitMap::markRange(MemRegion mr) {
ysr@777 126 mr.intersection(MemRegion(_bmStartWord, _bmWordSize));
ysr@777 127 assert(!mr.is_empty(), "unexpected empty region");
ysr@777 128 assert((offsetToHeapWord(heapWordToOffset(mr.end())) ==
ysr@777 129 ((HeapWord *) mr.end())),
ysr@777 130 "markRange memory region end is not card aligned");
ysr@777 131 // convert address range into offset range
ysr@777 132 _bm.at_put_range(heapWordToOffset(mr.start()),
ysr@777 133 heapWordToOffset(mr.end()), true);
ysr@777 134 }
ysr@777 135
ysr@777 136 void CMBitMap::clearRange(MemRegion mr) {
ysr@777 137 mr.intersection(MemRegion(_bmStartWord, _bmWordSize));
ysr@777 138 assert(!mr.is_empty(), "unexpected empty region");
ysr@777 139 // convert address range into offset range
ysr@777 140 _bm.at_put_range(heapWordToOffset(mr.start()),
ysr@777 141 heapWordToOffset(mr.end()), false);
ysr@777 142 }
ysr@777 143
ysr@777 144 MemRegion CMBitMap::getAndClearMarkedRegion(HeapWord* addr,
ysr@777 145 HeapWord* end_addr) {
ysr@777 146 HeapWord* start = getNextMarkedWordAddress(addr);
ysr@777 147 start = MIN2(start, end_addr);
ysr@777 148 HeapWord* end = getNextUnmarkedWordAddress(start);
ysr@777 149 end = MIN2(end, end_addr);
ysr@777 150 assert(start <= end, "Consistency check");
ysr@777 151 MemRegion mr(start, end);
ysr@777 152 if (!mr.is_empty()) {
ysr@777 153 clearRange(mr);
ysr@777 154 }
ysr@777 155 return mr;
ysr@777 156 }
ysr@777 157
ysr@777 158 CMMarkStack::CMMarkStack(ConcurrentMark* cm) :
ysr@777 159 _base(NULL), _cm(cm)
ysr@777 160 #ifdef ASSERT
ysr@777 161 , _drain_in_progress(false)
ysr@777 162 , _drain_in_progress_yields(false)
ysr@777 163 #endif
ysr@777 164 {}
ysr@777 165
ysr@777 166 void CMMarkStack::allocate(size_t size) {
zgu@3900 167 _base = NEW_C_HEAP_ARRAY(oop, size, mtGC);
tonyp@2973 168 if (_base == NULL) {
tonyp@3416 169 vm_exit_during_initialization("Failed to allocate CM region mark stack");
tonyp@2973 170 }
ysr@777 171 _index = 0;
ysr@777 172 _capacity = (jint) size;
tonyp@3416 173 _saved_index = -1;
ysr@777 174 NOT_PRODUCT(_max_depth = 0);
ysr@777 175 }
ysr@777 176
ysr@777 177 CMMarkStack::~CMMarkStack() {
tonyp@2973 178 if (_base != NULL) {
zgu@3900 179 FREE_C_HEAP_ARRAY(oop, _base, mtGC);
tonyp@2973 180 }
ysr@777 181 }
ysr@777 182
ysr@777 183 void CMMarkStack::par_push(oop ptr) {
ysr@777 184 while (true) {
ysr@777 185 if (isFull()) {
ysr@777 186 _overflow = true;
ysr@777 187 return;
ysr@777 188 }
ysr@777 189 // Otherwise...
ysr@777 190 jint index = _index;
ysr@777 191 jint next_index = index+1;
ysr@777 192 jint res = Atomic::cmpxchg(next_index, &_index, index);
ysr@777 193 if (res == index) {
ysr@777 194 _base[index] = ptr;
ysr@777 195 // Note that we don't maintain this atomically. We could, but it
ysr@777 196 // doesn't seem necessary.
ysr@777 197 NOT_PRODUCT(_max_depth = MAX2(_max_depth, next_index));
ysr@777 198 return;
ysr@777 199 }
ysr@777 200 // Otherwise, we need to try again.
ysr@777 201 }
ysr@777 202 }
ysr@777 203
ysr@777 204 void CMMarkStack::par_adjoin_arr(oop* ptr_arr, int n) {
ysr@777 205 while (true) {
ysr@777 206 if (isFull()) {
ysr@777 207 _overflow = true;
ysr@777 208 return;
ysr@777 209 }
ysr@777 210 // Otherwise...
ysr@777 211 jint index = _index;
ysr@777 212 jint next_index = index + n;
ysr@777 213 if (next_index > _capacity) {
ysr@777 214 _overflow = true;
ysr@777 215 return;
ysr@777 216 }
ysr@777 217 jint res = Atomic::cmpxchg(next_index, &_index, index);
ysr@777 218 if (res == index) {
ysr@777 219 for (int i = 0; i < n; i++) {
ysr@777 220 int ind = index + i;
ysr@777 221 assert(ind < _capacity, "By overflow test above.");
ysr@777 222 _base[ind] = ptr_arr[i];
ysr@777 223 }
ysr@777 224 NOT_PRODUCT(_max_depth = MAX2(_max_depth, next_index));
ysr@777 225 return;
ysr@777 226 }
ysr@777 227 // Otherwise, we need to try again.
ysr@777 228 }
ysr@777 229 }
ysr@777 230
ysr@777 231
ysr@777 232 void CMMarkStack::par_push_arr(oop* ptr_arr, int n) {
ysr@777 233 MutexLockerEx x(ParGCRareEvent_lock, Mutex::_no_safepoint_check_flag);
ysr@777 234 jint start = _index;
ysr@777 235 jint next_index = start + n;
ysr@777 236 if (next_index > _capacity) {
ysr@777 237 _overflow = true;
ysr@777 238 return;
ysr@777 239 }
ysr@777 240 // Otherwise.
ysr@777 241 _index = next_index;
ysr@777 242 for (int i = 0; i < n; i++) {
ysr@777 243 int ind = start + i;
tonyp@1458 244 assert(ind < _capacity, "By overflow test above.");
ysr@777 245 _base[ind] = ptr_arr[i];
ysr@777 246 }
ysr@777 247 }
ysr@777 248
ysr@777 249
ysr@777 250 bool CMMarkStack::par_pop_arr(oop* ptr_arr, int max, int* n) {
ysr@777 251 MutexLockerEx x(ParGCRareEvent_lock, Mutex::_no_safepoint_check_flag);
ysr@777 252 jint index = _index;
ysr@777 253 if (index == 0) {
ysr@777 254 *n = 0;
ysr@777 255 return false;
ysr@777 256 } else {
ysr@777 257 int k = MIN2(max, index);
ysr@777 258 jint new_ind = index - k;
ysr@777 259 for (int j = 0; j < k; j++) {
ysr@777 260 ptr_arr[j] = _base[new_ind + j];
ysr@777 261 }
ysr@777 262 _index = new_ind;
ysr@777 263 *n = k;
ysr@777 264 return true;
ysr@777 265 }
ysr@777 266 }
ysr@777 267
ysr@777 268 template<class OopClosureClass>
ysr@777 269 bool CMMarkStack::drain(OopClosureClass* cl, CMBitMap* bm, bool yield_after) {
ysr@777 270 assert(!_drain_in_progress || !_drain_in_progress_yields || yield_after
ysr@777 271 || SafepointSynchronize::is_at_safepoint(),
ysr@777 272 "Drain recursion must be yield-safe.");
ysr@777 273 bool res = true;
ysr@777 274 debug_only(_drain_in_progress = true);
ysr@777 275 debug_only(_drain_in_progress_yields = yield_after);
ysr@777 276 while (!isEmpty()) {
ysr@777 277 oop newOop = pop();
ysr@777 278 assert(G1CollectedHeap::heap()->is_in_reserved(newOop), "Bad pop");
ysr@777 279 assert(newOop->is_oop(), "Expected an oop");
ysr@777 280 assert(bm == NULL || bm->isMarked((HeapWord*)newOop),
ysr@777 281 "only grey objects on this stack");
ysr@777 282 newOop->oop_iterate(cl);
ysr@777 283 if (yield_after && _cm->do_yield_check()) {
tonyp@2973 284 res = false;
tonyp@2973 285 break;
ysr@777 286 }
ysr@777 287 }
ysr@777 288 debug_only(_drain_in_progress = false);
ysr@777 289 return res;
ysr@777 290 }
ysr@777 291
tonyp@3416 292 void CMMarkStack::note_start_of_gc() {
tonyp@3416 293 assert(_saved_index == -1,
tonyp@3416 294 "note_start_of_gc()/end_of_gc() bracketed incorrectly");
tonyp@3416 295 _saved_index = _index;
tonyp@3416 296 }
tonyp@3416 297
tonyp@3416 298 void CMMarkStack::note_end_of_gc() {
tonyp@3416 299 // This is intentionally a guarantee, instead of an assert. If we
tonyp@3416 300 // accidentally add something to the mark stack during GC, it
tonyp@3416 301 // will be a correctness issue so it's better if we crash. we'll
tonyp@3416 302 // only check this once per GC anyway, so it won't be a performance
tonyp@3416 303 // issue in any way.
tonyp@3416 304 guarantee(_saved_index == _index,
tonyp@3416 305 err_msg("saved index: %d index: %d", _saved_index, _index));
tonyp@3416 306 _saved_index = -1;
tonyp@3416 307 }
tonyp@3416 308
ysr@777 309 void CMMarkStack::oops_do(OopClosure* f) {
tonyp@3416 310 assert(_saved_index == _index,
tonyp@3416 311 err_msg("saved index: %d index: %d", _saved_index, _index));
tonyp@3416 312 for (int i = 0; i < _index; i += 1) {
ysr@777 313 f->do_oop(&_base[i]);
ysr@777 314 }
ysr@777 315 }
ysr@777 316
ysr@777 317 bool ConcurrentMark::not_yet_marked(oop obj) const {
coleenp@4037 318 return _g1h->is_obj_ill(obj);
ysr@777 319 }
ysr@777 320
tonyp@3464 321 CMRootRegions::CMRootRegions() :
tonyp@3464 322 _young_list(NULL), _cm(NULL), _scan_in_progress(false),
tonyp@3464 323 _should_abort(false), _next_survivor(NULL) { }
tonyp@3464 324
tonyp@3464 325 void CMRootRegions::init(G1CollectedHeap* g1h, ConcurrentMark* cm) {
tonyp@3464 326 _young_list = g1h->young_list();
tonyp@3464 327 _cm = cm;
tonyp@3464 328 }
tonyp@3464 329
tonyp@3464 330 void CMRootRegions::prepare_for_scan() {
tonyp@3464 331 assert(!scan_in_progress(), "pre-condition");
tonyp@3464 332
tonyp@3464 333 // Currently, only survivors can be root regions.
tonyp@3464 334 assert(_next_survivor == NULL, "pre-condition");
tonyp@3464 335 _next_survivor = _young_list->first_survivor_region();
tonyp@3464 336 _scan_in_progress = (_next_survivor != NULL);
tonyp@3464 337 _should_abort = false;
tonyp@3464 338 }
tonyp@3464 339
tonyp@3464 340 HeapRegion* CMRootRegions::claim_next() {
tonyp@3464 341 if (_should_abort) {
tonyp@3464 342 // If someone has set the should_abort flag, we return NULL to
tonyp@3464 343 // force the caller to bail out of their loop.
tonyp@3464 344 return NULL;
tonyp@3464 345 }
tonyp@3464 346
tonyp@3464 347 // Currently, only survivors can be root regions.
tonyp@3464 348 HeapRegion* res = _next_survivor;
tonyp@3464 349 if (res != NULL) {
tonyp@3464 350 MutexLockerEx x(RootRegionScan_lock, Mutex::_no_safepoint_check_flag);
tonyp@3464 351 // Read it again in case it changed while we were waiting for the lock.
tonyp@3464 352 res = _next_survivor;
tonyp@3464 353 if (res != NULL) {
tonyp@3464 354 if (res == _young_list->last_survivor_region()) {
tonyp@3464 355 // We just claimed the last survivor so store NULL to indicate
tonyp@3464 356 // that we're done.
tonyp@3464 357 _next_survivor = NULL;
tonyp@3464 358 } else {
tonyp@3464 359 _next_survivor = res->get_next_young_region();
tonyp@3464 360 }
tonyp@3464 361 } else {
tonyp@3464 362 // Someone else claimed the last survivor while we were trying
tonyp@3464 363 // to take the lock so nothing else to do.
tonyp@3464 364 }
tonyp@3464 365 }
tonyp@3464 366 assert(res == NULL || res->is_survivor(), "post-condition");
tonyp@3464 367
tonyp@3464 368 return res;
tonyp@3464 369 }
tonyp@3464 370
tonyp@3464 371 void CMRootRegions::scan_finished() {
tonyp@3464 372 assert(scan_in_progress(), "pre-condition");
tonyp@3464 373
tonyp@3464 374 // Currently, only survivors can be root regions.
tonyp@3464 375 if (!_should_abort) {
tonyp@3464 376 assert(_next_survivor == NULL, "we should have claimed all survivors");
tonyp@3464 377 }
tonyp@3464 378 _next_survivor = NULL;
tonyp@3464 379
tonyp@3464 380 {
tonyp@3464 381 MutexLockerEx x(RootRegionScan_lock, Mutex::_no_safepoint_check_flag);
tonyp@3464 382 _scan_in_progress = false;
tonyp@3464 383 RootRegionScan_lock->notify_all();
tonyp@3464 384 }
tonyp@3464 385 }
tonyp@3464 386
tonyp@3464 387 bool CMRootRegions::wait_until_scan_finished() {
tonyp@3464 388 if (!scan_in_progress()) return false;
tonyp@3464 389
tonyp@3464 390 {
tonyp@3464 391 MutexLockerEx x(RootRegionScan_lock, Mutex::_no_safepoint_check_flag);
tonyp@3464 392 while (scan_in_progress()) {
tonyp@3464 393 RootRegionScan_lock->wait(Mutex::_no_safepoint_check_flag);
tonyp@3464 394 }
tonyp@3464 395 }
tonyp@3464 396 return true;
tonyp@3464 397 }
tonyp@3464 398
ysr@777 399 #ifdef _MSC_VER // the use of 'this' below gets a warning, make it go away
ysr@777 400 #pragma warning( disable:4355 ) // 'this' : used in base member initializer list
ysr@777 401 #endif // _MSC_VER
ysr@777 402
jmasa@3357 403 uint ConcurrentMark::scale_parallel_threads(uint n_par_threads) {
jmasa@3357 404 return MAX2((n_par_threads + 2) / 4, 1U);
jmasa@3294 405 }
jmasa@3294 406
tonyp@3713 407 ConcurrentMark::ConcurrentMark(ReservedSpace rs, uint max_regions) :
ysr@777 408 _markBitMap1(rs, MinObjAlignment - 1),
ysr@777 409 _markBitMap2(rs, MinObjAlignment - 1),
ysr@777 410
ysr@777 411 _parallel_marking_threads(0),
jmasa@3294 412 _max_parallel_marking_threads(0),
ysr@777 413 _sleep_factor(0.0),
ysr@777 414 _marking_task_overhead(1.0),
ysr@777 415 _cleanup_sleep_factor(0.0),
ysr@777 416 _cleanup_task_overhead(1.0),
tonyp@2472 417 _cleanup_list("Cleanup List"),
tonyp@3713 418 _region_bm((BitMap::idx_t) max_regions, false /* in_resource_area*/),
ysr@777 419 _card_bm((rs.size() + CardTableModRefBS::card_size - 1) >>
ysr@777 420 CardTableModRefBS::card_shift,
ysr@777 421 false /* in_resource_area*/),
johnc@3463 422
ysr@777 423 _prevMarkBitMap(&_markBitMap1),
ysr@777 424 _nextMarkBitMap(&_markBitMap2),
ysr@777 425
ysr@777 426 _markStack(this),
ysr@777 427 // _finger set in set_non_marking_state
ysr@777 428
johnc@4173 429 _max_worker_id(MAX2((uint)ParallelGCThreads, 1U)),
ysr@777 430 // _active_tasks set in set_non_marking_state
ysr@777 431 // _tasks set inside the constructor
johnc@4173 432 _task_queues(new CMTaskQueueSet((int) _max_worker_id)),
johnc@4173 433 _terminator(ParallelTaskTerminator((int) _max_worker_id, _task_queues)),
ysr@777 434
ysr@777 435 _has_overflown(false),
ysr@777 436 _concurrent(false),
tonyp@1054 437 _has_aborted(false),
tonyp@1054 438 _restart_for_overflow(false),
tonyp@1054 439 _concurrent_marking_in_progress(false),
ysr@777 440
ysr@777 441 // _verbose_level set below
ysr@777 442
ysr@777 443 _init_times(),
ysr@777 444 _remark_times(), _remark_mark_times(), _remark_weak_ref_times(),
ysr@777 445 _cleanup_times(),
ysr@777 446 _total_counting_time(0.0),
ysr@777 447 _total_rs_scrub_time(0.0),
johnc@3463 448
johnc@3463 449 _parallel_workers(NULL),
johnc@3463 450
johnc@3463 451 _count_card_bitmaps(NULL),
johnc@3463 452 _count_marked_bytes(NULL) {
tonyp@2973 453 CMVerboseLevel verbose_level = (CMVerboseLevel) G1MarkingVerboseLevel;
tonyp@2973 454 if (verbose_level < no_verbose) {
ysr@777 455 verbose_level = no_verbose;
tonyp@2973 456 }
tonyp@2973 457 if (verbose_level > high_verbose) {
ysr@777 458 verbose_level = high_verbose;
tonyp@2973 459 }
ysr@777 460 _verbose_level = verbose_level;
ysr@777 461
tonyp@2973 462 if (verbose_low()) {
ysr@777 463 gclog_or_tty->print_cr("[global] init, heap start = "PTR_FORMAT", "
ysr@777 464 "heap end = "PTR_FORMAT, _heap_start, _heap_end);
tonyp@2973 465 }
ysr@777 466
jmasa@1719 467 _markStack.allocate(MarkStackSize);
ysr@777 468
ysr@777 469 // Create & start a ConcurrentMark thread.
ysr@1280 470 _cmThread = new ConcurrentMarkThread(this);
ysr@1280 471 assert(cmThread() != NULL, "CM Thread should have been created");
ysr@1280 472 assert(cmThread()->cm() != NULL, "CM Thread should refer to this cm");
ysr@1280 473
ysr@777 474 _g1h = G1CollectedHeap::heap();
ysr@777 475 assert(CGC_lock != NULL, "Where's the CGC_lock?");
ysr@777 476 assert(_markBitMap1.covers(rs), "_markBitMap1 inconsistency");
ysr@777 477 assert(_markBitMap2.covers(rs), "_markBitMap2 inconsistency");
ysr@777 478
ysr@777 479 SATBMarkQueueSet& satb_qs = JavaThread::satb_mark_queue_set();
tonyp@1717 480 satb_qs.set_buffer_size(G1SATBBufferSize);
ysr@777 481
tonyp@3464 482 _root_regions.init(_g1h, this);
tonyp@3464 483
johnc@4173 484 _tasks = NEW_C_HEAP_ARRAY(CMTask*, _max_worker_id, mtGC);
johnc@4173 485 _accum_task_vtime = NEW_C_HEAP_ARRAY(double, _max_worker_id, mtGC);
johnc@4173 486
johnc@4173 487 _count_card_bitmaps = NEW_C_HEAP_ARRAY(BitMap, _max_worker_id, mtGC);
johnc@4173 488 _count_marked_bytes = NEW_C_HEAP_ARRAY(size_t*, _max_worker_id, mtGC);
johnc@3463 489
johnc@3463 490 BitMap::idx_t card_bm_size = _card_bm.size();
johnc@3463 491
ysr@777 492 // so that the assertion in MarkingTaskQueue::task_queue doesn't fail
johnc@4173 493 _active_tasks = _max_worker_id;
johnc@4173 494 for (uint i = 0; i < _max_worker_id; ++i) {
ysr@777 495 CMTaskQueue* task_queue = new CMTaskQueue();
ysr@777 496 task_queue->initialize();
ysr@777 497 _task_queues->register_queue(i, task_queue);
ysr@777 498
johnc@3463 499 _count_card_bitmaps[i] = BitMap(card_bm_size, false);
zgu@3900 500 _count_marked_bytes[i] = NEW_C_HEAP_ARRAY(size_t, (size_t) max_regions, mtGC);
johnc@3463 501
johnc@3463 502 _tasks[i] = new CMTask(i, this,
johnc@3463 503 _count_marked_bytes[i],
johnc@3463 504 &_count_card_bitmaps[i],
johnc@3463 505 task_queue, _task_queues);
johnc@3463 506
ysr@777 507 _accum_task_vtime[i] = 0.0;
ysr@777 508 }
ysr@777 509
johnc@3463 510 // Calculate the card number for the bottom of the heap. Used
johnc@3463 511 // in biasing indexes into the accounting card bitmaps.
johnc@3463 512 _heap_bottom_card_num =
johnc@3463 513 intptr_t(uintptr_t(_g1h->reserved_region().start()) >>
johnc@3463 514 CardTableModRefBS::card_shift);
johnc@3463 515
johnc@3463 516 // Clear all the liveness counting data
johnc@3463 517 clear_all_count_data();
johnc@3463 518
jmasa@1719 519 if (ConcGCThreads > ParallelGCThreads) {
jmasa@1719 520 vm_exit_during_initialization("Can't have more ConcGCThreads "
ysr@777 521 "than ParallelGCThreads.");
ysr@777 522 }
ysr@777 523 if (ParallelGCThreads == 0) {
ysr@777 524 // if we are not running with any parallel GC threads we will not
ysr@777 525 // spawn any marking threads either
jmasa@3294 526 _parallel_marking_threads = 0;
jmasa@3294 527 _max_parallel_marking_threads = 0;
jmasa@3294 528 _sleep_factor = 0.0;
jmasa@3294 529 _marking_task_overhead = 1.0;
ysr@777 530 } else {
jmasa@1719 531 if (ConcGCThreads > 0) {
jmasa@1719 532 // notice that ConcGCThreads overwrites G1MarkingOverheadPercent
ysr@777 533 // if both are set
ysr@777 534
jmasa@3357 535 _parallel_marking_threads = (uint) ConcGCThreads;
jmasa@3294 536 _max_parallel_marking_threads = _parallel_marking_threads;
ysr@777 537 _sleep_factor = 0.0;
ysr@777 538 _marking_task_overhead = 1.0;
johnc@1186 539 } else if (G1MarkingOverheadPercent > 0) {
ysr@777 540 // we will calculate the number of parallel marking threads
ysr@777 541 // based on a target overhead with respect to the soft real-time
ysr@777 542 // goal
ysr@777 543
johnc@1186 544 double marking_overhead = (double) G1MarkingOverheadPercent / 100.0;
ysr@777 545 double overall_cm_overhead =
johnc@1186 546 (double) MaxGCPauseMillis * marking_overhead /
johnc@1186 547 (double) GCPauseIntervalMillis;
ysr@777 548 double cpu_ratio = 1.0 / (double) os::processor_count();
ysr@777 549 double marking_thread_num = ceil(overall_cm_overhead / cpu_ratio);
ysr@777 550 double marking_task_overhead =
ysr@777 551 overall_cm_overhead / marking_thread_num *
ysr@777 552 (double) os::processor_count();
ysr@777 553 double sleep_factor =
ysr@777 554 (1.0 - marking_task_overhead) / marking_task_overhead;
ysr@777 555
jmasa@3357 556 _parallel_marking_threads = (uint) marking_thread_num;
jmasa@3294 557 _max_parallel_marking_threads = _parallel_marking_threads;
ysr@777 558 _sleep_factor = sleep_factor;
ysr@777 559 _marking_task_overhead = marking_task_overhead;
ysr@777 560 } else {
jmasa@3357 561 _parallel_marking_threads = scale_parallel_threads((uint)ParallelGCThreads);
jmasa@3294 562 _max_parallel_marking_threads = _parallel_marking_threads;
ysr@777 563 _sleep_factor = 0.0;
ysr@777 564 _marking_task_overhead = 1.0;
ysr@777 565 }
ysr@777 566
tonyp@2973 567 if (parallel_marking_threads() > 1) {
ysr@777 568 _cleanup_task_overhead = 1.0;
tonyp@2973 569 } else {
ysr@777 570 _cleanup_task_overhead = marking_task_overhead();
tonyp@2973 571 }
ysr@777 572 _cleanup_sleep_factor =
ysr@777 573 (1.0 - cleanup_task_overhead()) / cleanup_task_overhead();
ysr@777 574
ysr@777 575 #if 0
ysr@777 576 gclog_or_tty->print_cr("Marking Threads %d", parallel_marking_threads());
ysr@777 577 gclog_or_tty->print_cr("CM Marking Task Overhead %1.4lf", marking_task_overhead());
ysr@777 578 gclog_or_tty->print_cr("CM Sleep Factor %1.4lf", sleep_factor());
ysr@777 579 gclog_or_tty->print_cr("CL Marking Task Overhead %1.4lf", cleanup_task_overhead());
ysr@777 580 gclog_or_tty->print_cr("CL Sleep Factor %1.4lf", cleanup_sleep_factor());
ysr@777 581 #endif
ysr@777 582
tonyp@1458 583 guarantee(parallel_marking_threads() > 0, "peace of mind");
jmasa@2188 584 _parallel_workers = new FlexibleWorkGang("G1 Parallel Marking Threads",
jmasa@3357 585 _max_parallel_marking_threads, false, true);
jmasa@2188 586 if (_parallel_workers == NULL) {
ysr@777 587 vm_exit_during_initialization("Failed necessary allocation.");
jmasa@2188 588 } else {
jmasa@2188 589 _parallel_workers->initialize_workers();
jmasa@2188 590 }
ysr@777 591 }
ysr@777 592
ysr@777 593 // so that the call below can read a sensible value
ysr@777 594 _heap_start = (HeapWord*) rs.base();
ysr@777 595 set_non_marking_state();
ysr@777 596 }
ysr@777 597
ysr@777 598 void ConcurrentMark::update_g1_committed(bool force) {
ysr@777 599 // If concurrent marking is not in progress, then we do not need to
tonyp@3691 600 // update _heap_end.
tonyp@2973 601 if (!concurrent_marking_in_progress() && !force) return;
ysr@777 602
ysr@777 603 MemRegion committed = _g1h->g1_committed();
tonyp@1458 604 assert(committed.start() == _heap_start, "start shouldn't change");
ysr@777 605 HeapWord* new_end = committed.end();
ysr@777 606 if (new_end > _heap_end) {
ysr@777 607 // The heap has been expanded.
ysr@777 608
ysr@777 609 _heap_end = new_end;
ysr@777 610 }
ysr@777 611 // Notice that the heap can also shrink. However, this only happens
ysr@777 612 // during a Full GC (at least currently) and the entire marking
ysr@777 613 // phase will bail out and the task will not be restarted. So, let's
ysr@777 614 // do nothing.
ysr@777 615 }
ysr@777 616
ysr@777 617 void ConcurrentMark::reset() {
ysr@777 618 // Starting values for these two. This should be called in a STW
ysr@777 619 // phase. CM will be notified of any future g1_committed expansions
ysr@777 620 // will be at the end of evacuation pauses, when tasks are
ysr@777 621 // inactive.
ysr@777 622 MemRegion committed = _g1h->g1_committed();
ysr@777 623 _heap_start = committed.start();
ysr@777 624 _heap_end = committed.end();
ysr@777 625
tonyp@1458 626 // Separated the asserts so that we know which one fires.
tonyp@1458 627 assert(_heap_start != NULL, "heap bounds should look ok");
tonyp@1458 628 assert(_heap_end != NULL, "heap bounds should look ok");
tonyp@1458 629 assert(_heap_start < _heap_end, "heap bounds should look ok");
ysr@777 630
ysr@777 631 // reset all the marking data structures and any necessary flags
ysr@777 632 clear_marking_state();
ysr@777 633
tonyp@2973 634 if (verbose_low()) {
ysr@777 635 gclog_or_tty->print_cr("[global] resetting");
tonyp@2973 636 }
ysr@777 637
ysr@777 638 // We do reset all of them, since different phases will use
ysr@777 639 // different number of active threads. So, it's easiest to have all
ysr@777 640 // of them ready.
johnc@4173 641 for (uint i = 0; i < _max_worker_id; ++i) {
ysr@777 642 _tasks[i]->reset(_nextMarkBitMap);
johnc@2190 643 }
ysr@777 644
ysr@777 645 // we need this to make sure that the flag is on during the evac
ysr@777 646 // pause with initial mark piggy-backed
ysr@777 647 set_concurrent_marking_in_progress();
ysr@777 648 }
ysr@777 649
jmasa@3357 650 void ConcurrentMark::set_phase(uint active_tasks, bool concurrent) {
johnc@4173 651 assert(active_tasks <= _max_worker_id, "we should not have more");
ysr@777 652
ysr@777 653 _active_tasks = active_tasks;
ysr@777 654 // Need to update the three data structures below according to the
ysr@777 655 // number of active threads for this phase.
ysr@777 656 _terminator = ParallelTaskTerminator((int) active_tasks, _task_queues);
ysr@777 657 _first_overflow_barrier_sync.set_n_workers((int) active_tasks);
ysr@777 658 _second_overflow_barrier_sync.set_n_workers((int) active_tasks);
ysr@777 659
ysr@777 660 _concurrent = concurrent;
ysr@777 661 // We propagate this to all tasks, not just the active ones.
johnc@4173 662 for (uint i = 0; i < _max_worker_id; ++i)
ysr@777 663 _tasks[i]->set_concurrent(concurrent);
ysr@777 664
ysr@777 665 if (concurrent) {
ysr@777 666 set_concurrent_marking_in_progress();
ysr@777 667 } else {
ysr@777 668 // We currently assume that the concurrent flag has been set to
ysr@777 669 // false before we start remark. At this point we should also be
ysr@777 670 // in a STW phase.
tonyp@1458 671 assert(!concurrent_marking_in_progress(), "invariant");
tonyp@1458 672 assert(_finger == _heap_end, "only way to get here");
ysr@777 673 update_g1_committed(true);
ysr@777 674 }
ysr@777 675 }
ysr@777 676
ysr@777 677 void ConcurrentMark::set_non_marking_state() {
ysr@777 678 // We set the global marking state to some default values when we're
ysr@777 679 // not doing marking.
ysr@777 680 clear_marking_state();
ysr@777 681 _active_tasks = 0;
ysr@777 682 clear_concurrent_marking_in_progress();
ysr@777 683 }
ysr@777 684
ysr@777 685 ConcurrentMark::~ConcurrentMark() {
stefank@3364 686 // The ConcurrentMark instance is never freed.
stefank@3364 687 ShouldNotReachHere();
ysr@777 688 }
ysr@777 689
ysr@777 690 void ConcurrentMark::clearNextBitmap() {
tonyp@1794 691 G1CollectedHeap* g1h = G1CollectedHeap::heap();
tonyp@1794 692 G1CollectorPolicy* g1p = g1h->g1_policy();
tonyp@1794 693
tonyp@1794 694 // Make sure that the concurrent mark thread looks to still be in
tonyp@1794 695 // the current cycle.
tonyp@1794 696 guarantee(cmThread()->during_cycle(), "invariant");
tonyp@1794 697
tonyp@1794 698 // We are finishing up the current cycle by clearing the next
tonyp@1794 699 // marking bitmap and getting it ready for the next cycle. During
tonyp@1794 700 // this time no other cycle can start. So, let's make sure that this
tonyp@1794 701 // is the case.
tonyp@1794 702 guarantee(!g1h->mark_in_progress(), "invariant");
tonyp@1794 703
tonyp@1794 704 // clear the mark bitmap (no grey objects to start with).
tonyp@1794 705 // We need to do this in chunks and offer to yield in between
tonyp@1794 706 // each chunk.
tonyp@1794 707 HeapWord* start = _nextMarkBitMap->startWord();
tonyp@1794 708 HeapWord* end = _nextMarkBitMap->endWord();
tonyp@1794 709 HeapWord* cur = start;
tonyp@1794 710 size_t chunkSize = M;
tonyp@1794 711 while (cur < end) {
tonyp@1794 712 HeapWord* next = cur + chunkSize;
tonyp@2973 713 if (next > end) {
tonyp@1794 714 next = end;
tonyp@2973 715 }
tonyp@1794 716 MemRegion mr(cur,next);
tonyp@1794 717 _nextMarkBitMap->clearRange(mr);
tonyp@1794 718 cur = next;
tonyp@1794 719 do_yield_check();
tonyp@1794 720
tonyp@1794 721 // Repeat the asserts from above. We'll do them as asserts here to
tonyp@1794 722 // minimize their overhead on the product. However, we'll have
tonyp@1794 723 // them as guarantees at the beginning / end of the bitmap
tonyp@1794 724 // clearing to get some checking in the product.
tonyp@1794 725 assert(cmThread()->during_cycle(), "invariant");
tonyp@1794 726 assert(!g1h->mark_in_progress(), "invariant");
tonyp@1794 727 }
tonyp@1794 728
johnc@3463 729 // Clear the liveness counting data
johnc@3463 730 clear_all_count_data();
johnc@3463 731
tonyp@1794 732 // Repeat the asserts from above.
tonyp@1794 733 guarantee(cmThread()->during_cycle(), "invariant");
tonyp@1794 734 guarantee(!g1h->mark_in_progress(), "invariant");
ysr@777 735 }
ysr@777 736
ysr@777 737 class NoteStartOfMarkHRClosure: public HeapRegionClosure {
ysr@777 738 public:
ysr@777 739 bool doHeapRegion(HeapRegion* r) {
ysr@777 740 if (!r->continuesHumongous()) {
tonyp@3416 741 r->note_start_of_marking();
ysr@777 742 }
ysr@777 743 return false;
ysr@777 744 }
ysr@777 745 };
ysr@777 746
ysr@777 747 void ConcurrentMark::checkpointRootsInitialPre() {
ysr@777 748 G1CollectedHeap* g1h = G1CollectedHeap::heap();
ysr@777 749 G1CollectorPolicy* g1p = g1h->g1_policy();
ysr@777 750
ysr@777 751 _has_aborted = false;
ysr@777 752
jcoomes@1902 753 #ifndef PRODUCT
tonyp@1479 754 if (G1PrintReachableAtInitialMark) {
tonyp@1823 755 print_reachable("at-cycle-start",
johnc@2969 756 VerifyOption_G1UsePrevMarking, true /* all */);
tonyp@1479 757 }
jcoomes@1902 758 #endif
ysr@777 759
ysr@777 760 // Initialise marking structures. This has to be done in a STW phase.
ysr@777 761 reset();
tonyp@3416 762
tonyp@3416 763 // For each region note start of marking.
tonyp@3416 764 NoteStartOfMarkHRClosure startcl;
tonyp@3416 765 g1h->heap_region_iterate(&startcl);
ysr@777 766 }
ysr@777 767
ysr@777 768
ysr@777 769 void ConcurrentMark::checkpointRootsInitialPost() {
ysr@777 770 G1CollectedHeap* g1h = G1CollectedHeap::heap();
ysr@777 771
tonyp@2848 772 // If we force an overflow during remark, the remark operation will
tonyp@2848 773 // actually abort and we'll restart concurrent marking. If we always
tonyp@2848 774 // force an oveflow during remark we'll never actually complete the
tonyp@2848 775 // marking phase. So, we initilize this here, at the start of the
tonyp@2848 776 // cycle, so that at the remaining overflow number will decrease at
tonyp@2848 777 // every remark and we'll eventually not need to cause one.
tonyp@2848 778 force_overflow_stw()->init();
tonyp@2848 779
johnc@3175 780 // Start Concurrent Marking weak-reference discovery.
johnc@3175 781 ReferenceProcessor* rp = g1h->ref_processor_cm();
johnc@3175 782 // enable ("weak") refs discovery
johnc@3175 783 rp->enable_discovery(true /*verify_disabled*/, true /*verify_no_refs*/);
ysr@892 784 rp->setup_policy(false); // snapshot the soft ref policy to be used in this cycle
ysr@777 785
ysr@777 786 SATBMarkQueueSet& satb_mq_set = JavaThread::satb_mark_queue_set();
tonyp@1752 787 // This is the start of the marking cycle, we're expected all
tonyp@1752 788 // threads to have SATB queues with active set to false.
tonyp@1752 789 satb_mq_set.set_active_all_threads(true, /* new active value */
tonyp@1752 790 false /* expected_active */);
ysr@777 791
tonyp@3464 792 _root_regions.prepare_for_scan();
tonyp@3464 793
ysr@777 794 // update_g1_committed() will be called at the end of an evac pause
ysr@777 795 // when marking is on. So, it's also called at the end of the
ysr@777 796 // initial-mark pause to update the heap end, if the heap expands
ysr@777 797 // during it. No need to call it here.
ysr@777 798 }
ysr@777 799
ysr@777 800 /*
tonyp@2848 801 * Notice that in the next two methods, we actually leave the STS
tonyp@2848 802 * during the barrier sync and join it immediately afterwards. If we
tonyp@2848 803 * do not do this, the following deadlock can occur: one thread could
tonyp@2848 804 * be in the barrier sync code, waiting for the other thread to also
tonyp@2848 805 * sync up, whereas another one could be trying to yield, while also
tonyp@2848 806 * waiting for the other threads to sync up too.
tonyp@2848 807 *
tonyp@2848 808 * Note, however, that this code is also used during remark and in
tonyp@2848 809 * this case we should not attempt to leave / enter the STS, otherwise
tonyp@2848 810 * we'll either hit an asseert (debug / fastdebug) or deadlock
tonyp@2848 811 * (product). So we should only leave / enter the STS if we are
tonyp@2848 812 * operating concurrently.
tonyp@2848 813 *
tonyp@2848 814 * Because the thread that does the sync barrier has left the STS, it
tonyp@2848 815 * is possible to be suspended for a Full GC or an evacuation pause
tonyp@2848 816 * could occur. This is actually safe, since the entering the sync
tonyp@2848 817 * barrier is one of the last things do_marking_step() does, and it
tonyp@2848 818 * doesn't manipulate any data structures afterwards.
tonyp@2848 819 */
ysr@777 820
johnc@4173 821 void ConcurrentMark::enter_first_sync_barrier(uint worker_id) {
tonyp@2973 822 if (verbose_low()) {
johnc@4173 823 gclog_or_tty->print_cr("[%u] entering first barrier", worker_id);
tonyp@2973 824 }
ysr@777 825
tonyp@2848 826 if (concurrent()) {
tonyp@2848 827 ConcurrentGCThread::stsLeave();
tonyp@2848 828 }
ysr@777 829 _first_overflow_barrier_sync.enter();
tonyp@2848 830 if (concurrent()) {
tonyp@2848 831 ConcurrentGCThread::stsJoin();
tonyp@2848 832 }
ysr@777 833 // at this point everyone should have synced up and not be doing any
ysr@777 834 // more work
ysr@777 835
tonyp@2973 836 if (verbose_low()) {
johnc@4173 837 gclog_or_tty->print_cr("[%u] leaving first barrier", worker_id);
tonyp@2973 838 }
ysr@777 839
johnc@4173 840 // let the task associated with with worker 0 do this
johnc@4173 841 if (worker_id == 0) {
ysr@777 842 // task 0 is responsible for clearing the global data structures
tonyp@2848 843 // We should be here because of an overflow. During STW we should
tonyp@2848 844 // not clear the overflow flag since we rely on it being true when
tonyp@2848 845 // we exit this method to abort the pause and restart concurent
tonyp@2848 846 // marking.
tonyp@2848 847 clear_marking_state(concurrent() /* clear_overflow */);
tonyp@2848 848 force_overflow()->update();
ysr@777 849
brutisso@3710 850 if (G1Log::fine()) {
ysr@777 851 gclog_or_tty->date_stamp(PrintGCDateStamps);
ysr@777 852 gclog_or_tty->stamp(PrintGCTimeStamps);
ysr@777 853 gclog_or_tty->print_cr("[GC concurrent-mark-reset-for-overflow]");
ysr@777 854 }
ysr@777 855 }
ysr@777 856
ysr@777 857 // after this, each task should reset its own data structures then
ysr@777 858 // then go into the second barrier
ysr@777 859 }
ysr@777 860
johnc@4173 861 void ConcurrentMark::enter_second_sync_barrier(uint worker_id) {
tonyp@2973 862 if (verbose_low()) {
johnc@4173 863 gclog_or_tty->print_cr("[%u] entering second barrier", worker_id);
tonyp@2973 864 }
ysr@777 865
tonyp@2848 866 if (concurrent()) {
tonyp@2848 867 ConcurrentGCThread::stsLeave();
tonyp@2848 868 }
ysr@777 869 _second_overflow_barrier_sync.enter();
tonyp@2848 870 if (concurrent()) {
tonyp@2848 871 ConcurrentGCThread::stsJoin();
tonyp@2848 872 }
ysr@777 873 // at this point everything should be re-initialised and ready to go
ysr@777 874
tonyp@2973 875 if (verbose_low()) {
johnc@4173 876 gclog_or_tty->print_cr("[%u] leaving second barrier", worker_id);
tonyp@2973 877 }
ysr@777 878 }
ysr@777 879
tonyp@2848 880 #ifndef PRODUCT
tonyp@2848 881 void ForceOverflowSettings::init() {
tonyp@2848 882 _num_remaining = G1ConcMarkForceOverflow;
tonyp@2848 883 _force = false;
tonyp@2848 884 update();
tonyp@2848 885 }
tonyp@2848 886
tonyp@2848 887 void ForceOverflowSettings::update() {
tonyp@2848 888 if (_num_remaining > 0) {
tonyp@2848 889 _num_remaining -= 1;
tonyp@2848 890 _force = true;
tonyp@2848 891 } else {
tonyp@2848 892 _force = false;
tonyp@2848 893 }
tonyp@2848 894 }
tonyp@2848 895
tonyp@2848 896 bool ForceOverflowSettings::should_force() {
tonyp@2848 897 if (_force) {
tonyp@2848 898 _force = false;
tonyp@2848 899 return true;
tonyp@2848 900 } else {
tonyp@2848 901 return false;
tonyp@2848 902 }
tonyp@2848 903 }
tonyp@2848 904 #endif // !PRODUCT
tonyp@2848 905
ysr@777 906 class CMConcurrentMarkingTask: public AbstractGangTask {
ysr@777 907 private:
ysr@777 908 ConcurrentMark* _cm;
ysr@777 909 ConcurrentMarkThread* _cmt;
ysr@777 910
ysr@777 911 public:
jmasa@3357 912 void work(uint worker_id) {
tonyp@1458 913 assert(Thread::current()->is_ConcurrentGC_thread(),
tonyp@1458 914 "this should only be done by a conc GC thread");
johnc@2316 915 ResourceMark rm;
ysr@777 916
ysr@777 917 double start_vtime = os::elapsedVTime();
ysr@777 918
ysr@777 919 ConcurrentGCThread::stsJoin();
ysr@777 920
jmasa@3357 921 assert(worker_id < _cm->active_tasks(), "invariant");
jmasa@3357 922 CMTask* the_task = _cm->task(worker_id);
ysr@777 923 the_task->record_start_time();
ysr@777 924 if (!_cm->has_aborted()) {
ysr@777 925 do {
ysr@777 926 double start_vtime_sec = os::elapsedVTime();
ysr@777 927 double start_time_sec = os::elapsedTime();
johnc@2494 928 double mark_step_duration_ms = G1ConcMarkStepDurationMillis;
johnc@2494 929
johnc@2494 930 the_task->do_marking_step(mark_step_duration_ms,
johnc@2494 931 true /* do_stealing */,
johnc@2494 932 true /* do_termination */);
johnc@2494 933
ysr@777 934 double end_time_sec = os::elapsedTime();
ysr@777 935 double end_vtime_sec = os::elapsedVTime();
ysr@777 936 double elapsed_vtime_sec = end_vtime_sec - start_vtime_sec;
ysr@777 937 double elapsed_time_sec = end_time_sec - start_time_sec;
ysr@777 938 _cm->clear_has_overflown();
ysr@777 939
jmasa@3357 940 bool ret = _cm->do_yield_check(worker_id);
ysr@777 941
ysr@777 942 jlong sleep_time_ms;
ysr@777 943 if (!_cm->has_aborted() && the_task->has_aborted()) {
ysr@777 944 sleep_time_ms =
ysr@777 945 (jlong) (elapsed_vtime_sec * _cm->sleep_factor() * 1000.0);
ysr@777 946 ConcurrentGCThread::stsLeave();
ysr@777 947 os::sleep(Thread::current(), sleep_time_ms, false);
ysr@777 948 ConcurrentGCThread::stsJoin();
ysr@777 949 }
ysr@777 950 double end_time2_sec = os::elapsedTime();
ysr@777 951 double elapsed_time2_sec = end_time2_sec - start_time_sec;
ysr@777 952
ysr@777 953 #if 0
ysr@777 954 gclog_or_tty->print_cr("CM: elapsed %1.4lf ms, sleep %1.4lf ms, "
ysr@777 955 "overhead %1.4lf",
ysr@777 956 elapsed_vtime_sec * 1000.0, (double) sleep_time_ms,
ysr@777 957 the_task->conc_overhead(os::elapsedTime()) * 8.0);
ysr@777 958 gclog_or_tty->print_cr("elapsed time %1.4lf ms, time 2: %1.4lf ms",
ysr@777 959 elapsed_time_sec * 1000.0, elapsed_time2_sec * 1000.0);
ysr@777 960 #endif
ysr@777 961 } while (!_cm->has_aborted() && the_task->has_aborted());
ysr@777 962 }
ysr@777 963 the_task->record_end_time();
tonyp@1458 964 guarantee(!the_task->has_aborted() || _cm->has_aborted(), "invariant");
ysr@777 965
ysr@777 966 ConcurrentGCThread::stsLeave();
ysr@777 967
ysr@777 968 double end_vtime = os::elapsedVTime();
jmasa@3357 969 _cm->update_accum_task_vtime(worker_id, end_vtime - start_vtime);
ysr@777 970 }
ysr@777 971
ysr@777 972 CMConcurrentMarkingTask(ConcurrentMark* cm,
ysr@777 973 ConcurrentMarkThread* cmt) :
ysr@777 974 AbstractGangTask("Concurrent Mark"), _cm(cm), _cmt(cmt) { }
ysr@777 975
ysr@777 976 ~CMConcurrentMarkingTask() { }
ysr@777 977 };
ysr@777 978
jmasa@3294 979 // Calculates the number of active workers for a concurrent
jmasa@3294 980 // phase.
jmasa@3357 981 uint ConcurrentMark::calc_parallel_marking_threads() {
johnc@3338 982 if (G1CollectedHeap::use_parallel_gc_threads()) {
jmasa@3357 983 uint n_conc_workers = 0;
jmasa@3294 984 if (!UseDynamicNumberOfGCThreads ||
jmasa@3294 985 (!FLAG_IS_DEFAULT(ConcGCThreads) &&
jmasa@3294 986 !ForceDynamicNumberOfGCThreads)) {
jmasa@3294 987 n_conc_workers = max_parallel_marking_threads();
jmasa@3294 988 } else {
jmasa@3294 989 n_conc_workers =
jmasa@3294 990 AdaptiveSizePolicy::calc_default_active_workers(
jmasa@3294 991 max_parallel_marking_threads(),
jmasa@3294 992 1, /* Minimum workers */
jmasa@3294 993 parallel_marking_threads(),
jmasa@3294 994 Threads::number_of_non_daemon_threads());
jmasa@3294 995 // Don't scale down "n_conc_workers" by scale_parallel_threads() because
jmasa@3294 996 // that scaling has already gone into "_max_parallel_marking_threads".
jmasa@3294 997 }
johnc@3338 998 assert(n_conc_workers > 0, "Always need at least 1");
johnc@3338 999 return n_conc_workers;
jmasa@3294 1000 }
johnc@3338 1001 // If we are not running with any parallel GC threads we will not
johnc@3338 1002 // have spawned any marking threads either. Hence the number of
johnc@3338 1003 // concurrent workers should be 0.
johnc@3338 1004 return 0;
jmasa@3294 1005 }
jmasa@3294 1006
tonyp@3464 1007 void ConcurrentMark::scanRootRegion(HeapRegion* hr, uint worker_id) {
tonyp@3464 1008 // Currently, only survivors can be root regions.
tonyp@3464 1009 assert(hr->next_top_at_mark_start() == hr->bottom(), "invariant");
tonyp@3464 1010 G1RootRegionScanClosure cl(_g1h, this, worker_id);
tonyp@3464 1011
tonyp@3464 1012 const uintx interval = PrefetchScanIntervalInBytes;
tonyp@3464 1013 HeapWord* curr = hr->bottom();
tonyp@3464 1014 const HeapWord* end = hr->top();
tonyp@3464 1015 while (curr < end) {
tonyp@3464 1016 Prefetch::read(curr, interval);
tonyp@3464 1017 oop obj = oop(curr);
tonyp@3464 1018 int size = obj->oop_iterate(&cl);
tonyp@3464 1019 assert(size == obj->size(), "sanity");
tonyp@3464 1020 curr += size;
tonyp@3464 1021 }
tonyp@3464 1022 }
tonyp@3464 1023
tonyp@3464 1024 class CMRootRegionScanTask : public AbstractGangTask {
tonyp@3464 1025 private:
tonyp@3464 1026 ConcurrentMark* _cm;
tonyp@3464 1027
tonyp@3464 1028 public:
tonyp@3464 1029 CMRootRegionScanTask(ConcurrentMark* cm) :
tonyp@3464 1030 AbstractGangTask("Root Region Scan"), _cm(cm) { }
tonyp@3464 1031
tonyp@3464 1032 void work(uint worker_id) {
tonyp@3464 1033 assert(Thread::current()->is_ConcurrentGC_thread(),
tonyp@3464 1034 "this should only be done by a conc GC thread");
tonyp@3464 1035
tonyp@3464 1036 CMRootRegions* root_regions = _cm->root_regions();
tonyp@3464 1037 HeapRegion* hr = root_regions->claim_next();
tonyp@3464 1038 while (hr != NULL) {
tonyp@3464 1039 _cm->scanRootRegion(hr, worker_id);
tonyp@3464 1040 hr = root_regions->claim_next();
tonyp@3464 1041 }
tonyp@3464 1042 }
tonyp@3464 1043 };
tonyp@3464 1044
tonyp@3464 1045 void ConcurrentMark::scanRootRegions() {
tonyp@3464 1046 // scan_in_progress() will have been set to true only if there was
tonyp@3464 1047 // at least one root region to scan. So, if it's false, we
tonyp@3464 1048 // should not attempt to do any further work.
tonyp@3464 1049 if (root_regions()->scan_in_progress()) {
tonyp@3464 1050 _parallel_marking_threads = calc_parallel_marking_threads();
tonyp@3464 1051 assert(parallel_marking_threads() <= max_parallel_marking_threads(),
tonyp@3464 1052 "Maximum number of marking threads exceeded");
tonyp@3464 1053 uint active_workers = MAX2(1U, parallel_marking_threads());
tonyp@3464 1054
tonyp@3464 1055 CMRootRegionScanTask task(this);
tonyp@3464 1056 if (parallel_marking_threads() > 0) {
tonyp@3464 1057 _parallel_workers->set_active_workers((int) active_workers);
tonyp@3464 1058 _parallel_workers->run_task(&task);
tonyp@3464 1059 } else {
tonyp@3464 1060 task.work(0);
tonyp@3464 1061 }
tonyp@3464 1062
tonyp@3464 1063 // It's possible that has_aborted() is true here without actually
tonyp@3464 1064 // aborting the survivor scan earlier. This is OK as it's
tonyp@3464 1065 // mainly used for sanity checking.
tonyp@3464 1066 root_regions()->scan_finished();
tonyp@3464 1067 }
tonyp@3464 1068 }
tonyp@3464 1069
ysr@777 1070 void ConcurrentMark::markFromRoots() {
ysr@777 1071 // we might be tempted to assert that:
ysr@777 1072 // assert(asynch == !SafepointSynchronize::is_at_safepoint(),
ysr@777 1073 // "inconsistent argument?");
ysr@777 1074 // However that wouldn't be right, because it's possible that
ysr@777 1075 // a safepoint is indeed in progress as a younger generation
ysr@777 1076 // stop-the-world GC happens even as we mark in this generation.
ysr@777 1077
ysr@777 1078 _restart_for_overflow = false;
tonyp@2848 1079 force_overflow_conc()->init();
jmasa@3294 1080
jmasa@3294 1081 // _g1h has _n_par_threads
jmasa@3294 1082 _parallel_marking_threads = calc_parallel_marking_threads();
jmasa@3294 1083 assert(parallel_marking_threads() <= max_parallel_marking_threads(),
jmasa@3294 1084 "Maximum number of marking threads exceeded");
johnc@3338 1085
jmasa@3357 1086 uint active_workers = MAX2(1U, parallel_marking_threads());
johnc@3338 1087
johnc@3338 1088 // Parallel task terminator is set in "set_phase()"
johnc@3338 1089 set_phase(active_workers, true /* concurrent */);
ysr@777 1090
ysr@777 1091 CMConcurrentMarkingTask markingTask(this, cmThread());
tonyp@2973 1092 if (parallel_marking_threads() > 0) {
johnc@3338 1093 _parallel_workers->set_active_workers((int)active_workers);
johnc@3338 1094 // Don't set _n_par_threads because it affects MT in proceess_strong_roots()
johnc@3338 1095 // and the decisions on that MT processing is made elsewhere.
johnc@3338 1096 assert(_parallel_workers->active_workers() > 0, "Should have been set");
ysr@777 1097 _parallel_workers->run_task(&markingTask);
tonyp@2973 1098 } else {
ysr@777 1099 markingTask.work(0);
tonyp@2973 1100 }
ysr@777 1101 print_stats();
ysr@777 1102 }
ysr@777 1103
ysr@777 1104 void ConcurrentMark::checkpointRootsFinal(bool clear_all_soft_refs) {
ysr@777 1105 // world is stopped at this checkpoint
ysr@777 1106 assert(SafepointSynchronize::is_at_safepoint(),
ysr@777 1107 "world should be stopped");
johnc@3175 1108
ysr@777 1109 G1CollectedHeap* g1h = G1CollectedHeap::heap();
ysr@777 1110
ysr@777 1111 // If a full collection has happened, we shouldn't do this.
ysr@777 1112 if (has_aborted()) {
ysr@777 1113 g1h->set_marking_complete(); // So bitmap clearing isn't confused
ysr@777 1114 return;
ysr@777 1115 }
ysr@777 1116
kamg@2445 1117 SvcGCMarker sgcm(SvcGCMarker::OTHER);
kamg@2445 1118
ysr@1280 1119 if (VerifyDuringGC) {
ysr@1280 1120 HandleMark hm; // handle scope
ysr@1280 1121 gclog_or_tty->print(" VerifyDuringGC:(before)");
ysr@1280 1122 Universe::heap()->prepare_for_verify();
brutisso@3711 1123 Universe::verify(/* silent */ false,
johnc@2969 1124 /* option */ VerifyOption_G1UsePrevMarking);
ysr@1280 1125 }
ysr@1280 1126
ysr@777 1127 G1CollectorPolicy* g1p = g1h->g1_policy();
ysr@777 1128 g1p->record_concurrent_mark_remark_start();
ysr@777 1129
ysr@777 1130 double start = os::elapsedTime();
ysr@777 1131
ysr@777 1132 checkpointRootsFinalWork();
ysr@777 1133
ysr@777 1134 double mark_work_end = os::elapsedTime();
ysr@777 1135
ysr@777 1136 weakRefsWork(clear_all_soft_refs);
ysr@777 1137
ysr@777 1138 if (has_overflown()) {
ysr@777 1139 // Oops. We overflowed. Restart concurrent marking.
ysr@777 1140 _restart_for_overflow = true;
ysr@777 1141 // Clear the flag. We do not need it any more.
ysr@777 1142 clear_has_overflown();
tonyp@2973 1143 if (G1TraceMarkStackOverflow) {
ysr@777 1144 gclog_or_tty->print_cr("\nRemark led to restart for overflow.");
tonyp@2973 1145 }
ysr@777 1146 } else {
johnc@3463 1147 // Aggregate the per-task counting data that we have accumulated
johnc@3463 1148 // while marking.
johnc@3463 1149 aggregate_count_data();
johnc@3463 1150
tonyp@2469 1151 SATBMarkQueueSet& satb_mq_set = JavaThread::satb_mark_queue_set();
ysr@777 1152 // We're done with marking.
tonyp@1752 1153 // This is the end of the marking cycle, we're expected all
tonyp@1752 1154 // threads to have SATB queues with active set to true.
tonyp@2469 1155 satb_mq_set.set_active_all_threads(false, /* new active value */
tonyp@2469 1156 true /* expected_active */);
tonyp@1246 1157
tonyp@1246 1158 if (VerifyDuringGC) {
ysr@1280 1159 HandleMark hm; // handle scope
ysr@1280 1160 gclog_or_tty->print(" VerifyDuringGC:(after)");
ysr@1280 1161 Universe::heap()->prepare_for_verify();
brutisso@3711 1162 Universe::verify(/* silent */ false,
johnc@2969 1163 /* option */ VerifyOption_G1UseNextMarking);
tonyp@1246 1164 }
johnc@2494 1165 assert(!restart_for_overflow(), "sanity");
johnc@2494 1166 }
johnc@2494 1167
johnc@2494 1168 // Reset the marking state if marking completed
johnc@2494 1169 if (!restart_for_overflow()) {
johnc@2494 1170 set_non_marking_state();
ysr@777 1171 }
ysr@777 1172
ysr@777 1173 #if VERIFY_OBJS_PROCESSED
ysr@777 1174 _scan_obj_cl.objs_processed = 0;
ysr@777 1175 ThreadLocalObjQueue::objs_enqueued = 0;
ysr@777 1176 #endif
ysr@777 1177
ysr@777 1178 // Statistics
ysr@777 1179 double now = os::elapsedTime();
ysr@777 1180 _remark_mark_times.add((mark_work_end - start) * 1000.0);
ysr@777 1181 _remark_weak_ref_times.add((now - mark_work_end) * 1000.0);
ysr@777 1182 _remark_times.add((now - start) * 1000.0);
ysr@777 1183
ysr@777 1184 g1p->record_concurrent_mark_remark_end();
ysr@777 1185 }
ysr@777 1186
johnc@3731 1187 // Base class of the closures that finalize and verify the
johnc@3731 1188 // liveness counting data.
johnc@3731 1189 class CMCountDataClosureBase: public HeapRegionClosure {
johnc@3731 1190 protected:
johnc@4123 1191 G1CollectedHeap* _g1h;
ysr@777 1192 ConcurrentMark* _cm;
johnc@4123 1193 CardTableModRefBS* _ct_bs;
johnc@4123 1194
johnc@3463 1195 BitMap* _region_bm;
johnc@3463 1196 BitMap* _card_bm;
johnc@3463 1197
johnc@4123 1198 // Takes a region that's not empty (i.e., it has at least one
tonyp@1264 1199 // live object in it and sets its corresponding bit on the region
tonyp@1264 1200 // bitmap to 1. If the region is "starts humongous" it will also set
tonyp@1264 1201 // to 1 the bits on the region bitmap that correspond to its
tonyp@1264 1202 // associated "continues humongous" regions.
tonyp@1264 1203 void set_bit_for_region(HeapRegion* hr) {
tonyp@1264 1204 assert(!hr->continuesHumongous(), "should have filtered those out");
tonyp@1264 1205
tonyp@3713 1206 BitMap::idx_t index = (BitMap::idx_t) hr->hrs_index();
tonyp@1264 1207 if (!hr->startsHumongous()) {
tonyp@1264 1208 // Normal (non-humongous) case: just set the bit.
tonyp@3713 1209 _region_bm->par_at_put(index, true);
tonyp@1264 1210 } else {
tonyp@1264 1211 // Starts humongous case: calculate how many regions are part of
johnc@3463 1212 // this humongous region and then set the bit range.
tonyp@3957 1213 BitMap::idx_t end_index = (BitMap::idx_t) hr->last_hc_index();
tonyp@3713 1214 _region_bm->par_at_put_range(index, end_index, true);
tonyp@1264 1215 }
tonyp@1264 1216 }
tonyp@1264 1217
johnc@3731 1218 public:
johnc@4123 1219 CMCountDataClosureBase(G1CollectedHeap* g1h,
johnc@3731 1220 BitMap* region_bm, BitMap* card_bm):
johnc@4123 1221 _g1h(g1h), _cm(g1h->concurrent_mark()),
johnc@4123 1222 _ct_bs((CardTableModRefBS*) (g1h->barrier_set())),
johnc@4123 1223 _region_bm(region_bm), _card_bm(card_bm) { }
johnc@3731 1224 };
johnc@3731 1225
johnc@3731 1226 // Closure that calculates the # live objects per region. Used
johnc@3731 1227 // for verification purposes during the cleanup pause.
johnc@3731 1228 class CalcLiveObjectsClosure: public CMCountDataClosureBase {
johnc@3731 1229 CMBitMapRO* _bm;
johnc@3731 1230 size_t _region_marked_bytes;
johnc@3731 1231
johnc@3731 1232 public:
johnc@4123 1233 CalcLiveObjectsClosure(CMBitMapRO *bm, G1CollectedHeap* g1h,
johnc@3731 1234 BitMap* region_bm, BitMap* card_bm) :
johnc@4123 1235 CMCountDataClosureBase(g1h, region_bm, card_bm),
johnc@3731 1236 _bm(bm), _region_marked_bytes(0) { }
johnc@3731 1237
ysr@777 1238 bool doHeapRegion(HeapRegion* hr) {
ysr@777 1239
iveresov@1074 1240 if (hr->continuesHumongous()) {
tonyp@1264 1241 // We will ignore these here and process them when their
tonyp@1264 1242 // associated "starts humongous" region is processed (see
tonyp@1264 1243 // set_bit_for_heap_region()). Note that we cannot rely on their
tonyp@1264 1244 // associated "starts humongous" region to have their bit set to
tonyp@1264 1245 // 1 since, due to the region chunking in the parallel region
tonyp@1264 1246 // iteration, a "continues humongous" region might be visited
tonyp@1264 1247 // before its associated "starts humongous".
iveresov@1074 1248 return false;
iveresov@1074 1249 }
ysr@777 1250
johnc@4123 1251 HeapWord* ntams = hr->next_top_at_mark_start();
johnc@4123 1252 HeapWord* start = hr->bottom();
johnc@4123 1253
johnc@4123 1254 assert(start <= hr->end() && start <= ntams && ntams <= hr->end(),
johnc@3463 1255 err_msg("Preconditions not met - "
johnc@4123 1256 "start: "PTR_FORMAT", ntams: "PTR_FORMAT", end: "PTR_FORMAT,
johnc@4123 1257 start, ntams, hr->end()));
johnc@3463 1258
ysr@777 1259 // Find the first marked object at or after "start".
johnc@4123 1260 start = _bm->getNextMarkedWordAddress(start, ntams);
johnc@3463 1261
ysr@777 1262 size_t marked_bytes = 0;
ysr@777 1263
johnc@4123 1264 while (start < ntams) {
ysr@777 1265 oop obj = oop(start);
ysr@777 1266 int obj_sz = obj->size();
johnc@4123 1267 HeapWord* obj_end = start + obj_sz;
johnc@3731 1268
johnc@3731 1269 BitMap::idx_t start_idx = _cm->card_bitmap_index_for(start);
johnc@4123 1270 BitMap::idx_t end_idx = _cm->card_bitmap_index_for(obj_end);
johnc@4123 1271
johnc@4123 1272 // Note: if we're looking at the last region in heap - obj_end
johnc@4123 1273 // could be actually just beyond the end of the heap; end_idx
johnc@4123 1274 // will then correspond to a (non-existent) card that is also
johnc@4123 1275 // just beyond the heap.
johnc@4123 1276 if (_g1h->is_in_g1_reserved(obj_end) && !_ct_bs->is_card_aligned(obj_end)) {
johnc@4123 1277 // end of object is not card aligned - increment to cover
johnc@4123 1278 // all the cards spanned by the object
johnc@4123 1279 end_idx += 1;
johnc@4123 1280 }
johnc@4123 1281
johnc@4123 1282 // Set the bits in the card BM for the cards spanned by this object.
johnc@4123 1283 _cm->set_card_bitmap_range(_card_bm, start_idx, end_idx, true /* is_par */);
johnc@3731 1284
johnc@3731 1285 // Add the size of this object to the number of marked bytes.
apetrusenko@1465 1286 marked_bytes += (size_t)obj_sz * HeapWordSize;
johnc@3463 1287
ysr@777 1288 // Find the next marked object after this one.
johnc@4123 1289 start = _bm->getNextMarkedWordAddress(obj_end, ntams);
tonyp@2973 1290 }
johnc@3463 1291
johnc@3463 1292 // Mark the allocated-since-marking portion...
johnc@3463 1293 HeapWord* top = hr->top();
johnc@4123 1294 if (ntams < top) {
johnc@4123 1295 BitMap::idx_t start_idx = _cm->card_bitmap_index_for(ntams);
johnc@4123 1296 BitMap::idx_t end_idx = _cm->card_bitmap_index_for(top);
johnc@4123 1297
johnc@4123 1298 // Note: if we're looking at the last region in heap - top
johnc@4123 1299 // could be actually just beyond the end of the heap; end_idx
johnc@4123 1300 // will then correspond to a (non-existent) card that is also
johnc@4123 1301 // just beyond the heap.
johnc@4123 1302 if (_g1h->is_in_g1_reserved(top) && !_ct_bs->is_card_aligned(top)) {
johnc@4123 1303 // end of object is not card aligned - increment to cover
johnc@4123 1304 // all the cards spanned by the object
johnc@4123 1305 end_idx += 1;
johnc@4123 1306 }
johnc@4123 1307 _cm->set_card_bitmap_range(_card_bm, start_idx, end_idx, true /* is_par */);
johnc@3463 1308
johnc@3463 1309 // This definitely means the region has live objects.
johnc@3463 1310 set_bit_for_region(hr);
ysr@777 1311 }
ysr@777 1312
ysr@777 1313 // Update the live region bitmap.
ysr@777 1314 if (marked_bytes > 0) {
tonyp@1264 1315 set_bit_for_region(hr);
ysr@777 1316 }
johnc@3463 1317
johnc@3463 1318 // Set the marked bytes for the current region so that
johnc@3463 1319 // it can be queried by a calling verificiation routine
johnc@3463 1320 _region_marked_bytes = marked_bytes;
johnc@3463 1321
johnc@3463 1322 return false;
johnc@3463 1323 }
johnc@3463 1324
johnc@3463 1325 size_t region_marked_bytes() const { return _region_marked_bytes; }
johnc@3463 1326 };
johnc@3463 1327
johnc@3463 1328 // Heap region closure used for verifying the counting data
johnc@3463 1329 // that was accumulated concurrently and aggregated during
johnc@3463 1330 // the remark pause. This closure is applied to the heap
johnc@3463 1331 // regions during the STW cleanup pause.
johnc@3463 1332
johnc@3463 1333 class VerifyLiveObjectDataHRClosure: public HeapRegionClosure {
johnc@4123 1334 G1CollectedHeap* _g1h;
johnc@3463 1335 ConcurrentMark* _cm;
johnc@3463 1336 CalcLiveObjectsClosure _calc_cl;
johnc@3463 1337 BitMap* _region_bm; // Region BM to be verified
johnc@3463 1338 BitMap* _card_bm; // Card BM to be verified
johnc@3463 1339 bool _verbose; // verbose output?
johnc@3463 1340
johnc@3463 1341 BitMap* _exp_region_bm; // Expected Region BM values
johnc@3463 1342 BitMap* _exp_card_bm; // Expected card BM values
johnc@3463 1343
johnc@3463 1344 int _failures;
johnc@3463 1345
johnc@3463 1346 public:
johnc@4123 1347 VerifyLiveObjectDataHRClosure(G1CollectedHeap* g1h,
johnc@3463 1348 BitMap* region_bm,
johnc@3463 1349 BitMap* card_bm,
johnc@3463 1350 BitMap* exp_region_bm,
johnc@3463 1351 BitMap* exp_card_bm,
johnc@3463 1352 bool verbose) :
johnc@4123 1353 _g1h(g1h), _cm(g1h->concurrent_mark()),
johnc@4123 1354 _calc_cl(_cm->nextMarkBitMap(), g1h, exp_region_bm, exp_card_bm),
johnc@3463 1355 _region_bm(region_bm), _card_bm(card_bm), _verbose(verbose),
johnc@3463 1356 _exp_region_bm(exp_region_bm), _exp_card_bm(exp_card_bm),
johnc@3463 1357 _failures(0) { }
johnc@3463 1358
johnc@3463 1359 int failures() const { return _failures; }
johnc@3463 1360
johnc@3463 1361 bool doHeapRegion(HeapRegion* hr) {
johnc@3463 1362 if (hr->continuesHumongous()) {
johnc@3463 1363 // We will ignore these here and process them when their
johnc@3463 1364 // associated "starts humongous" region is processed (see
johnc@3463 1365 // set_bit_for_heap_region()). Note that we cannot rely on their
johnc@3463 1366 // associated "starts humongous" region to have their bit set to
johnc@3463 1367 // 1 since, due to the region chunking in the parallel region
johnc@3463 1368 // iteration, a "continues humongous" region might be visited
johnc@3463 1369 // before its associated "starts humongous".
johnc@3463 1370 return false;
johnc@3463 1371 }
johnc@3463 1372
johnc@3463 1373 int failures = 0;
johnc@3463 1374
johnc@3463 1375 // Call the CalcLiveObjectsClosure to walk the marking bitmap for
johnc@3463 1376 // this region and set the corresponding bits in the expected region
johnc@3463 1377 // and card bitmaps.
johnc@3463 1378 bool res = _calc_cl.doHeapRegion(hr);
johnc@3463 1379 assert(res == false, "should be continuing");
johnc@3463 1380
johnc@3463 1381 MutexLockerEx x((_verbose ? ParGCRareEvent_lock : NULL),
johnc@3463 1382 Mutex::_no_safepoint_check_flag);
johnc@3463 1383
johnc@3463 1384 // Verify the marked bytes for this region.
johnc@3463 1385 size_t exp_marked_bytes = _calc_cl.region_marked_bytes();
johnc@3463 1386 size_t act_marked_bytes = hr->next_marked_bytes();
johnc@3463 1387
johnc@3463 1388 // We're not OK if expected marked bytes > actual marked bytes. It means
johnc@3463 1389 // we have missed accounting some objects during the actual marking.
johnc@3463 1390 if (exp_marked_bytes > act_marked_bytes) {
johnc@3463 1391 if (_verbose) {
tonyp@3713 1392 gclog_or_tty->print_cr("Region %u: marked bytes mismatch: "
johnc@3463 1393 "expected: " SIZE_FORMAT ", actual: " SIZE_FORMAT,
johnc@3463 1394 hr->hrs_index(), exp_marked_bytes, act_marked_bytes);
johnc@3463 1395 }
johnc@3463 1396 failures += 1;
johnc@3463 1397 }
johnc@3463 1398
johnc@3463 1399 // Verify the bit, for this region, in the actual and expected
johnc@3463 1400 // (which was just calculated) region bit maps.
johnc@3463 1401 // We're not OK if the bit in the calculated expected region
johnc@3463 1402 // bitmap is set and the bit in the actual region bitmap is not.
tonyp@3713 1403 BitMap::idx_t index = (BitMap::idx_t) hr->hrs_index();
johnc@3463 1404
johnc@3463 1405 bool expected = _exp_region_bm->at(index);
johnc@3463 1406 bool actual = _region_bm->at(index);
johnc@3463 1407 if (expected && !actual) {
johnc@3463 1408 if (_verbose) {
tonyp@3713 1409 gclog_or_tty->print_cr("Region %u: region bitmap mismatch: "
tonyp@3713 1410 "expected: %s, actual: %s",
tonyp@3713 1411 hr->hrs_index(),
tonyp@3713 1412 BOOL_TO_STR(expected), BOOL_TO_STR(actual));
johnc@3463 1413 }
johnc@3463 1414 failures += 1;
johnc@3463 1415 }
johnc@3463 1416
johnc@3463 1417 // Verify that the card bit maps for the cards spanned by the current
johnc@3463 1418 // region match. We have an error if we have a set bit in the expected
johnc@3463 1419 // bit map and the corresponding bit in the actual bitmap is not set.
johnc@3463 1420
johnc@3463 1421 BitMap::idx_t start_idx = _cm->card_bitmap_index_for(hr->bottom());
johnc@3463 1422 BitMap::idx_t end_idx = _cm->card_bitmap_index_for(hr->top());
johnc@3463 1423
johnc@3463 1424 for (BitMap::idx_t i = start_idx; i < end_idx; i+=1) {
johnc@3463 1425 expected = _exp_card_bm->at(i);
johnc@3463 1426 actual = _card_bm->at(i);
johnc@3463 1427
johnc@3463 1428 if (expected && !actual) {
johnc@3463 1429 if (_verbose) {
tonyp@3713 1430 gclog_or_tty->print_cr("Region %u: card bitmap mismatch at " SIZE_FORMAT ": "
tonyp@3713 1431 "expected: %s, actual: %s",
tonyp@3713 1432 hr->hrs_index(), i,
tonyp@3713 1433 BOOL_TO_STR(expected), BOOL_TO_STR(actual));
ysr@777 1434 }
johnc@3463 1435 failures += 1;
ysr@777 1436 }
ysr@777 1437 }
ysr@777 1438
johnc@3463 1439 if (failures > 0 && _verbose) {
johnc@3463 1440 gclog_or_tty->print_cr("Region " HR_FORMAT ", ntams: " PTR_FORMAT ", "
johnc@3463 1441 "marked_bytes: calc/actual " SIZE_FORMAT "/" SIZE_FORMAT,
johnc@3463 1442 HR_FORMAT_PARAMS(hr), hr->next_top_at_mark_start(),
johnc@3463 1443 _calc_cl.region_marked_bytes(), hr->next_marked_bytes());
johnc@3463 1444 }
johnc@3463 1445
johnc@3463 1446 _failures += failures;
johnc@3463 1447
johnc@3463 1448 // We could stop iteration over the heap when we
johnc@3731 1449 // find the first violating region by returning true.
ysr@777 1450 return false;
ysr@777 1451 }
ysr@777 1452 };
ysr@777 1453
ysr@777 1454
johnc@3463 1455 class G1ParVerifyFinalCountTask: public AbstractGangTask {
johnc@3463 1456 protected:
johnc@3463 1457 G1CollectedHeap* _g1h;
johnc@3463 1458 ConcurrentMark* _cm;
johnc@3463 1459 BitMap* _actual_region_bm;
johnc@3463 1460 BitMap* _actual_card_bm;
johnc@3463 1461
johnc@3463 1462 uint _n_workers;
johnc@3463 1463
johnc@3463 1464 BitMap* _expected_region_bm;
johnc@3463 1465 BitMap* _expected_card_bm;
johnc@3463 1466
johnc@3463 1467 int _failures;
johnc@3463 1468 bool _verbose;
johnc@3463 1469
johnc@3463 1470 public:
johnc@3463 1471 G1ParVerifyFinalCountTask(G1CollectedHeap* g1h,
johnc@3463 1472 BitMap* region_bm, BitMap* card_bm,
johnc@3463 1473 BitMap* expected_region_bm, BitMap* expected_card_bm)
johnc@3463 1474 : AbstractGangTask("G1 verify final counting"),
johnc@3463 1475 _g1h(g1h), _cm(_g1h->concurrent_mark()),
johnc@3463 1476 _actual_region_bm(region_bm), _actual_card_bm(card_bm),
johnc@3463 1477 _expected_region_bm(expected_region_bm), _expected_card_bm(expected_card_bm),
johnc@3463 1478 _failures(0), _verbose(false),
johnc@3463 1479 _n_workers(0) {
johnc@3463 1480 assert(VerifyDuringGC, "don't call this otherwise");
johnc@3463 1481
johnc@3463 1482 // Use the value already set as the number of active threads
johnc@3463 1483 // in the call to run_task().
johnc@3463 1484 if (G1CollectedHeap::use_parallel_gc_threads()) {
johnc@3463 1485 assert( _g1h->workers()->active_workers() > 0,
johnc@3463 1486 "Should have been previously set");
johnc@3463 1487 _n_workers = _g1h->workers()->active_workers();
johnc@3463 1488 } else {
johnc@3463 1489 _n_workers = 1;
johnc@3463 1490 }
johnc@3463 1491
johnc@3463 1492 assert(_expected_card_bm->size() == _actual_card_bm->size(), "sanity");
johnc@3463 1493 assert(_expected_region_bm->size() == _actual_region_bm->size(), "sanity");
johnc@3463 1494
johnc@3463 1495 _verbose = _cm->verbose_medium();
johnc@3463 1496 }
johnc@3463 1497
johnc@3463 1498 void work(uint worker_id) {
johnc@3463 1499 assert(worker_id < _n_workers, "invariant");
johnc@3463 1500
johnc@4123 1501 VerifyLiveObjectDataHRClosure verify_cl(_g1h,
johnc@3463 1502 _actual_region_bm, _actual_card_bm,
johnc@3463 1503 _expected_region_bm,
johnc@3463 1504 _expected_card_bm,
johnc@3463 1505 _verbose);
johnc@3463 1506
johnc@3463 1507 if (G1CollectedHeap::use_parallel_gc_threads()) {
johnc@3463 1508 _g1h->heap_region_par_iterate_chunked(&verify_cl,
johnc@3463 1509 worker_id,
johnc@3463 1510 _n_workers,
johnc@3463 1511 HeapRegion::VerifyCountClaimValue);
johnc@3463 1512 } else {
johnc@3463 1513 _g1h->heap_region_iterate(&verify_cl);
johnc@3463 1514 }
johnc@3463 1515
johnc@3463 1516 Atomic::add(verify_cl.failures(), &_failures);
johnc@3463 1517 }
johnc@3463 1518
johnc@3463 1519 int failures() const { return _failures; }
johnc@3463 1520 };
johnc@3463 1521
johnc@3731 1522 // Closure that finalizes the liveness counting data.
johnc@3731 1523 // Used during the cleanup pause.
johnc@3731 1524 // Sets the bits corresponding to the interval [NTAMS, top]
johnc@3731 1525 // (which contains the implicitly live objects) in the
johnc@3731 1526 // card liveness bitmap. Also sets the bit for each region,
johnc@3731 1527 // containing live data, in the region liveness bitmap.
johnc@3731 1528
johnc@3731 1529 class FinalCountDataUpdateClosure: public CMCountDataClosureBase {
johnc@3463 1530 public:
johnc@4123 1531 FinalCountDataUpdateClosure(G1CollectedHeap* g1h,
johnc@3463 1532 BitMap* region_bm,
johnc@3463 1533 BitMap* card_bm) :
johnc@4123 1534 CMCountDataClosureBase(g1h, region_bm, card_bm) { }
johnc@3463 1535
johnc@3463 1536 bool doHeapRegion(HeapRegion* hr) {
johnc@3463 1537
johnc@3463 1538 if (hr->continuesHumongous()) {
johnc@3463 1539 // We will ignore these here and process them when their
johnc@3463 1540 // associated "starts humongous" region is processed (see
johnc@3463 1541 // set_bit_for_heap_region()). Note that we cannot rely on their
johnc@3463 1542 // associated "starts humongous" region to have their bit set to
johnc@3463 1543 // 1 since, due to the region chunking in the parallel region
johnc@3463 1544 // iteration, a "continues humongous" region might be visited
johnc@3463 1545 // before its associated "starts humongous".
johnc@3463 1546 return false;
johnc@3463 1547 }
johnc@3463 1548
johnc@3463 1549 HeapWord* ntams = hr->next_top_at_mark_start();
johnc@3463 1550 HeapWord* top = hr->top();
johnc@3463 1551
johnc@3731 1552 assert(hr->bottom() <= ntams && ntams <= hr->end(), "Preconditions.");
johnc@3463 1553
johnc@3463 1554 // Mark the allocated-since-marking portion...
johnc@3463 1555 if (ntams < top) {
johnc@3463 1556 // This definitely means the region has live objects.
johnc@3463 1557 set_bit_for_region(hr);
johnc@4123 1558
johnc@4123 1559 // Now set the bits in the card bitmap for [ntams, top)
johnc@4123 1560 BitMap::idx_t start_idx = _cm->card_bitmap_index_for(ntams);
johnc@4123 1561 BitMap::idx_t end_idx = _cm->card_bitmap_index_for(top);
johnc@4123 1562
johnc@4123 1563 // Note: if we're looking at the last region in heap - top
johnc@4123 1564 // could be actually just beyond the end of the heap; end_idx
johnc@4123 1565 // will then correspond to a (non-existent) card that is also
johnc@4123 1566 // just beyond the heap.
johnc@4123 1567 if (_g1h->is_in_g1_reserved(top) && !_ct_bs->is_card_aligned(top)) {
johnc@4123 1568 // end of object is not card aligned - increment to cover
johnc@4123 1569 // all the cards spanned by the object
johnc@4123 1570 end_idx += 1;
johnc@4123 1571 }
johnc@4123 1572
johnc@4123 1573 assert(end_idx <= _card_bm->size(),
johnc@4123 1574 err_msg("oob: end_idx= "SIZE_FORMAT", bitmap size= "SIZE_FORMAT,
johnc@4123 1575 end_idx, _card_bm->size()));
johnc@4123 1576 assert(start_idx < _card_bm->size(),
johnc@4123 1577 err_msg("oob: start_idx= "SIZE_FORMAT", bitmap size= "SIZE_FORMAT,
johnc@4123 1578 start_idx, _card_bm->size()));
johnc@4123 1579
johnc@4123 1580 _cm->set_card_bitmap_range(_card_bm, start_idx, end_idx, true /* is_par */);
coleenp@4037 1581 }
johnc@3463 1582
johnc@3463 1583 // Set the bit for the region if it contains live data
johnc@3463 1584 if (hr->next_marked_bytes() > 0) {
johnc@3463 1585 set_bit_for_region(hr);
johnc@3463 1586 }
johnc@3463 1587
johnc@3463 1588 return false;
johnc@3463 1589 }
johnc@3463 1590 };
ysr@777 1591
ysr@777 1592 class G1ParFinalCountTask: public AbstractGangTask {
ysr@777 1593 protected:
ysr@777 1594 G1CollectedHeap* _g1h;
johnc@3463 1595 ConcurrentMark* _cm;
johnc@3463 1596 BitMap* _actual_region_bm;
johnc@3463 1597 BitMap* _actual_card_bm;
johnc@3463 1598
jmasa@3357 1599 uint _n_workers;
johnc@3463 1600
ysr@777 1601 public:
johnc@3463 1602 G1ParFinalCountTask(G1CollectedHeap* g1h, BitMap* region_bm, BitMap* card_bm)
johnc@3463 1603 : AbstractGangTask("G1 final counting"),
johnc@3463 1604 _g1h(g1h), _cm(_g1h->concurrent_mark()),
johnc@3463 1605 _actual_region_bm(region_bm), _actual_card_bm(card_bm),
johnc@3463 1606 _n_workers(0) {
jmasa@3294 1607 // Use the value already set as the number of active threads
tonyp@3714 1608 // in the call to run_task().
jmasa@3294 1609 if (G1CollectedHeap::use_parallel_gc_threads()) {
jmasa@3294 1610 assert( _g1h->workers()->active_workers() > 0,
jmasa@3294 1611 "Should have been previously set");
jmasa@3294 1612 _n_workers = _g1h->workers()->active_workers();
tonyp@2973 1613 } else {
ysr@777 1614 _n_workers = 1;
tonyp@2973 1615 }
ysr@777 1616 }
ysr@777 1617
jmasa@3357 1618 void work(uint worker_id) {
johnc@3463 1619 assert(worker_id < _n_workers, "invariant");
johnc@3463 1620
johnc@4123 1621 FinalCountDataUpdateClosure final_update_cl(_g1h,
johnc@3463 1622 _actual_region_bm,
johnc@3463 1623 _actual_card_bm);
johnc@3463 1624
jmasa@2188 1625 if (G1CollectedHeap::use_parallel_gc_threads()) {
johnc@3463 1626 _g1h->heap_region_par_iterate_chunked(&final_update_cl,
johnc@3463 1627 worker_id,
johnc@3463 1628 _n_workers,
tonyp@790 1629 HeapRegion::FinalCountClaimValue);
ysr@777 1630 } else {
johnc@3463 1631 _g1h->heap_region_iterate(&final_update_cl);
ysr@777 1632 }
ysr@777 1633 }
ysr@777 1634 };
ysr@777 1635
ysr@777 1636 class G1ParNoteEndTask;
ysr@777 1637
ysr@777 1638 class G1NoteEndOfConcMarkClosure : public HeapRegionClosure {
ysr@777 1639 G1CollectedHeap* _g1;
ysr@777 1640 int _worker_num;
ysr@777 1641 size_t _max_live_bytes;
tonyp@3713 1642 uint _regions_claimed;
ysr@777 1643 size_t _freed_bytes;
tonyp@2493 1644 FreeRegionList* _local_cleanup_list;
tonyp@3268 1645 OldRegionSet* _old_proxy_set;
tonyp@2493 1646 HumongousRegionSet* _humongous_proxy_set;
tonyp@2493 1647 HRRSCleanupTask* _hrrs_cleanup_task;
ysr@777 1648 double _claimed_region_time;
ysr@777 1649 double _max_region_time;
ysr@777 1650
ysr@777 1651 public:
ysr@777 1652 G1NoteEndOfConcMarkClosure(G1CollectedHeap* g1,
tonyp@2493 1653 int worker_num,
tonyp@2493 1654 FreeRegionList* local_cleanup_list,
tonyp@3268 1655 OldRegionSet* old_proxy_set,
tonyp@2493 1656 HumongousRegionSet* humongous_proxy_set,
johnc@3292 1657 HRRSCleanupTask* hrrs_cleanup_task) :
johnc@3292 1658 _g1(g1), _worker_num(worker_num),
johnc@3292 1659 _max_live_bytes(0), _regions_claimed(0),
johnc@3292 1660 _freed_bytes(0),
johnc@3292 1661 _claimed_region_time(0.0), _max_region_time(0.0),
johnc@3292 1662 _local_cleanup_list(local_cleanup_list),
johnc@3292 1663 _old_proxy_set(old_proxy_set),
johnc@3292 1664 _humongous_proxy_set(humongous_proxy_set),
johnc@3292 1665 _hrrs_cleanup_task(hrrs_cleanup_task) { }
johnc@3292 1666
ysr@777 1667 size_t freed_bytes() { return _freed_bytes; }
ysr@777 1668
johnc@3292 1669 bool doHeapRegion(HeapRegion *hr) {
tonyp@3957 1670 if (hr->continuesHumongous()) {
tonyp@3957 1671 return false;
tonyp@3957 1672 }
johnc@3292 1673 // We use a claim value of zero here because all regions
johnc@3292 1674 // were claimed with value 1 in the FinalCount task.
tonyp@3957 1675 _g1->reset_gc_time_stamps(hr);
tonyp@3957 1676 double start = os::elapsedTime();
tonyp@3957 1677 _regions_claimed++;
tonyp@3957 1678 hr->note_end_of_marking();
tonyp@3957 1679 _max_live_bytes += hr->max_live_bytes();
tonyp@3957 1680 _g1->free_region_if_empty(hr,
tonyp@3957 1681 &_freed_bytes,
tonyp@3957 1682 _local_cleanup_list,
tonyp@3957 1683 _old_proxy_set,
tonyp@3957 1684 _humongous_proxy_set,
tonyp@3957 1685 _hrrs_cleanup_task,
tonyp@3957 1686 true /* par */);
tonyp@3957 1687 double region_time = (os::elapsedTime() - start);
tonyp@3957 1688 _claimed_region_time += region_time;
tonyp@3957 1689 if (region_time > _max_region_time) {
tonyp@3957 1690 _max_region_time = region_time;
johnc@3292 1691 }
johnc@3292 1692 return false;
johnc@3292 1693 }
ysr@777 1694
ysr@777 1695 size_t max_live_bytes() { return _max_live_bytes; }
tonyp@3713 1696 uint regions_claimed() { return _regions_claimed; }
ysr@777 1697 double claimed_region_time_sec() { return _claimed_region_time; }
ysr@777 1698 double max_region_time_sec() { return _max_region_time; }
ysr@777 1699 };
ysr@777 1700
ysr@777 1701 class G1ParNoteEndTask: public AbstractGangTask {
ysr@777 1702 friend class G1NoteEndOfConcMarkClosure;
tonyp@2472 1703
ysr@777 1704 protected:
ysr@777 1705 G1CollectedHeap* _g1h;
ysr@777 1706 size_t _max_live_bytes;
ysr@777 1707 size_t _freed_bytes;
tonyp@2472 1708 FreeRegionList* _cleanup_list;
tonyp@2472 1709
ysr@777 1710 public:
ysr@777 1711 G1ParNoteEndTask(G1CollectedHeap* g1h,
tonyp@2472 1712 FreeRegionList* cleanup_list) :
ysr@777 1713 AbstractGangTask("G1 note end"), _g1h(g1h),
tonyp@2472 1714 _max_live_bytes(0), _freed_bytes(0), _cleanup_list(cleanup_list) { }
ysr@777 1715
jmasa@3357 1716 void work(uint worker_id) {
ysr@777 1717 double start = os::elapsedTime();
tonyp@2493 1718 FreeRegionList local_cleanup_list("Local Cleanup List");
tonyp@3268 1719 OldRegionSet old_proxy_set("Local Cleanup Old Proxy Set");
tonyp@2493 1720 HumongousRegionSet humongous_proxy_set("Local Cleanup Humongous Proxy Set");
tonyp@2493 1721 HRRSCleanupTask hrrs_cleanup_task;
jmasa@3357 1722 G1NoteEndOfConcMarkClosure g1_note_end(_g1h, worker_id, &local_cleanup_list,
tonyp@3268 1723 &old_proxy_set,
tonyp@2493 1724 &humongous_proxy_set,
tonyp@2493 1725 &hrrs_cleanup_task);
jmasa@2188 1726 if (G1CollectedHeap::use_parallel_gc_threads()) {
jmasa@3357 1727 _g1h->heap_region_par_iterate_chunked(&g1_note_end, worker_id,
jmasa@3294 1728 _g1h->workers()->active_workers(),
tonyp@790 1729 HeapRegion::NoteEndClaimValue);
ysr@777 1730 } else {
ysr@777 1731 _g1h->heap_region_iterate(&g1_note_end);
ysr@777 1732 }
ysr@777 1733 assert(g1_note_end.complete(), "Shouldn't have yielded!");
ysr@777 1734
tonyp@2472 1735 // Now update the lists
tonyp@2472 1736 _g1h->update_sets_after_freeing_regions(g1_note_end.freed_bytes(),
tonyp@2472 1737 NULL /* free_list */,
tonyp@3268 1738 &old_proxy_set,
tonyp@2493 1739 &humongous_proxy_set,
tonyp@2472 1740 true /* par */);
ysr@777 1741 {
ysr@777 1742 MutexLockerEx x(ParGCRareEvent_lock, Mutex::_no_safepoint_check_flag);
ysr@777 1743 _max_live_bytes += g1_note_end.max_live_bytes();
ysr@777 1744 _freed_bytes += g1_note_end.freed_bytes();
tonyp@2472 1745
tonyp@2975 1746 // If we iterate over the global cleanup list at the end of
tonyp@2975 1747 // cleanup to do this printing we will not guarantee to only
tonyp@2975 1748 // generate output for the newly-reclaimed regions (the list
tonyp@2975 1749 // might not be empty at the beginning of cleanup; we might
tonyp@2975 1750 // still be working on its previous contents). So we do the
tonyp@2975 1751 // printing here, before we append the new regions to the global
tonyp@2975 1752 // cleanup list.
tonyp@2975 1753
tonyp@2975 1754 G1HRPrinter* hr_printer = _g1h->hr_printer();
tonyp@2975 1755 if (hr_printer->is_active()) {
tonyp@2975 1756 HeapRegionLinkedListIterator iter(&local_cleanup_list);
tonyp@2975 1757 while (iter.more_available()) {
tonyp@2975 1758 HeapRegion* hr = iter.get_next();
tonyp@2975 1759 hr_printer->cleanup(hr);
tonyp@2975 1760 }
tonyp@2975 1761 }
tonyp@2975 1762
tonyp@2493 1763 _cleanup_list->add_as_tail(&local_cleanup_list);
tonyp@2493 1764 assert(local_cleanup_list.is_empty(), "post-condition");
tonyp@2493 1765
tonyp@2493 1766 HeapRegionRemSet::finish_cleanup_task(&hrrs_cleanup_task);
ysr@777 1767 }
ysr@777 1768 }
ysr@777 1769 size_t max_live_bytes() { return _max_live_bytes; }
ysr@777 1770 size_t freed_bytes() { return _freed_bytes; }
ysr@777 1771 };
ysr@777 1772
ysr@777 1773 class G1ParScrubRemSetTask: public AbstractGangTask {
ysr@777 1774 protected:
ysr@777 1775 G1RemSet* _g1rs;
ysr@777 1776 BitMap* _region_bm;
ysr@777 1777 BitMap* _card_bm;
ysr@777 1778 public:
ysr@777 1779 G1ParScrubRemSetTask(G1CollectedHeap* g1h,
ysr@777 1780 BitMap* region_bm, BitMap* card_bm) :
ysr@777 1781 AbstractGangTask("G1 ScrubRS"), _g1rs(g1h->g1_rem_set()),
johnc@3463 1782 _region_bm(region_bm), _card_bm(card_bm) { }
ysr@777 1783
jmasa@3357 1784 void work(uint worker_id) {
jmasa@2188 1785 if (G1CollectedHeap::use_parallel_gc_threads()) {
jmasa@3357 1786 _g1rs->scrub_par(_region_bm, _card_bm, worker_id,
tonyp@790 1787 HeapRegion::ScrubRemSetClaimValue);
ysr@777 1788 } else {
ysr@777 1789 _g1rs->scrub(_region_bm, _card_bm);
ysr@777 1790 }
ysr@777 1791 }
ysr@777 1792
ysr@777 1793 };
ysr@777 1794
ysr@777 1795 void ConcurrentMark::cleanup() {
ysr@777 1796 // world is stopped at this checkpoint
ysr@777 1797 assert(SafepointSynchronize::is_at_safepoint(),
ysr@777 1798 "world should be stopped");
ysr@777 1799 G1CollectedHeap* g1h = G1CollectedHeap::heap();
ysr@777 1800
ysr@777 1801 // If a full collection has happened, we shouldn't do this.
ysr@777 1802 if (has_aborted()) {
ysr@777 1803 g1h->set_marking_complete(); // So bitmap clearing isn't confused
ysr@777 1804 return;
ysr@777 1805 }
ysr@777 1806
tonyp@3268 1807 HRSPhaseSetter x(HRSPhaseCleanup);
tonyp@2472 1808 g1h->verify_region_sets_optional();
tonyp@2472 1809
ysr@1280 1810 if (VerifyDuringGC) {
ysr@1280 1811 HandleMark hm; // handle scope
ysr@1280 1812 gclog_or_tty->print(" VerifyDuringGC:(before)");
ysr@1280 1813 Universe::heap()->prepare_for_verify();
brutisso@3711 1814 Universe::verify(/* silent */ false,
johnc@2969 1815 /* option */ VerifyOption_G1UsePrevMarking);
ysr@1280 1816 }
ysr@1280 1817
ysr@777 1818 G1CollectorPolicy* g1p = G1CollectedHeap::heap()->g1_policy();
ysr@777 1819 g1p->record_concurrent_mark_cleanup_start();
ysr@777 1820
ysr@777 1821 double start = os::elapsedTime();
ysr@777 1822
tonyp@2493 1823 HeapRegionRemSet::reset_for_cleanup_tasks();
tonyp@2493 1824
jmasa@3357 1825 uint n_workers;
jmasa@3294 1826
ysr@777 1827 // Do counting once more with the world stopped for good measure.
johnc@3463 1828 G1ParFinalCountTask g1_par_count_task(g1h, &_region_bm, &_card_bm);
johnc@3463 1829
jmasa@2188 1830 if (G1CollectedHeap::use_parallel_gc_threads()) {
johnc@3463 1831 assert(g1h->check_heap_region_claim_values(HeapRegion::InitialClaimValue),
tonyp@790 1832 "sanity check");
tonyp@790 1833
johnc@3338 1834 g1h->set_par_threads();
johnc@3338 1835 n_workers = g1h->n_par_threads();
jmasa@3357 1836 assert(g1h->n_par_threads() == n_workers,
johnc@3338 1837 "Should not have been reset");
ysr@777 1838 g1h->workers()->run_task(&g1_par_count_task);
jmasa@3294 1839 // Done with the parallel phase so reset to 0.
ysr@777 1840 g1h->set_par_threads(0);
tonyp@790 1841
johnc@3463 1842 assert(g1h->check_heap_region_claim_values(HeapRegion::FinalCountClaimValue),
tonyp@790 1843 "sanity check");
ysr@777 1844 } else {
johnc@3338 1845 n_workers = 1;
ysr@777 1846 g1_par_count_task.work(0);
ysr@777 1847 }
ysr@777 1848
johnc@3463 1849 if (VerifyDuringGC) {
johnc@3463 1850 // Verify that the counting data accumulated during marking matches
johnc@3463 1851 // that calculated by walking the marking bitmap.
johnc@3463 1852
johnc@3463 1853 // Bitmaps to hold expected values
johnc@3463 1854 BitMap expected_region_bm(_region_bm.size(), false);
johnc@3463 1855 BitMap expected_card_bm(_card_bm.size(), false);
johnc@3463 1856
johnc@3463 1857 G1ParVerifyFinalCountTask g1_par_verify_task(g1h,
johnc@3463 1858 &_region_bm,
johnc@3463 1859 &_card_bm,
johnc@3463 1860 &expected_region_bm,
johnc@3463 1861 &expected_card_bm);
johnc@3463 1862
johnc@3463 1863 if (G1CollectedHeap::use_parallel_gc_threads()) {
johnc@3463 1864 g1h->set_par_threads((int)n_workers);
johnc@3463 1865 g1h->workers()->run_task(&g1_par_verify_task);
johnc@3463 1866 // Done with the parallel phase so reset to 0.
johnc@3463 1867 g1h->set_par_threads(0);
johnc@3463 1868
johnc@3463 1869 assert(g1h->check_heap_region_claim_values(HeapRegion::VerifyCountClaimValue),
johnc@3463 1870 "sanity check");
johnc@3463 1871 } else {
johnc@3463 1872 g1_par_verify_task.work(0);
johnc@3463 1873 }
johnc@3463 1874
johnc@3463 1875 guarantee(g1_par_verify_task.failures() == 0, "Unexpected accounting failures");
johnc@3463 1876 }
johnc@3463 1877
ysr@777 1878 size_t start_used_bytes = g1h->used();
ysr@777 1879 g1h->set_marking_complete();
ysr@777 1880
ysr@777 1881 double count_end = os::elapsedTime();
ysr@777 1882 double this_final_counting_time = (count_end - start);
ysr@777 1883 _total_counting_time += this_final_counting_time;
ysr@777 1884
tonyp@2717 1885 if (G1PrintRegionLivenessInfo) {
tonyp@2717 1886 G1PrintRegionLivenessInfoClosure cl(gclog_or_tty, "Post-Marking");
tonyp@2717 1887 _g1h->heap_region_iterate(&cl);
tonyp@2717 1888 }
tonyp@2717 1889
ysr@777 1890 // Install newly created mark bitMap as "prev".
ysr@777 1891 swapMarkBitMaps();
ysr@777 1892
ysr@777 1893 g1h->reset_gc_time_stamp();
ysr@777 1894
ysr@777 1895 // Note end of marking in all heap regions.
tonyp@2472 1896 G1ParNoteEndTask g1_par_note_end_task(g1h, &_cleanup_list);
jmasa@2188 1897 if (G1CollectedHeap::use_parallel_gc_threads()) {
jmasa@3294 1898 g1h->set_par_threads((int)n_workers);
ysr@777 1899 g1h->workers()->run_task(&g1_par_note_end_task);
ysr@777 1900 g1h->set_par_threads(0);
tonyp@790 1901
tonyp@790 1902 assert(g1h->check_heap_region_claim_values(HeapRegion::NoteEndClaimValue),
tonyp@790 1903 "sanity check");
ysr@777 1904 } else {
ysr@777 1905 g1_par_note_end_task.work(0);
ysr@777 1906 }
tonyp@3957 1907 g1h->check_gc_time_stamps();
tonyp@2472 1908
tonyp@2472 1909 if (!cleanup_list_is_empty()) {
tonyp@2472 1910 // The cleanup list is not empty, so we'll have to process it
tonyp@2472 1911 // concurrently. Notify anyone else that might be wanting free
tonyp@2472 1912 // regions that there will be more free regions coming soon.
tonyp@2472 1913 g1h->set_free_regions_coming();
tonyp@2472 1914 }
ysr@777 1915
ysr@777 1916 // call below, since it affects the metric by which we sort the heap
ysr@777 1917 // regions.
ysr@777 1918 if (G1ScrubRemSets) {
ysr@777 1919 double rs_scrub_start = os::elapsedTime();
ysr@777 1920 G1ParScrubRemSetTask g1_par_scrub_rs_task(g1h, &_region_bm, &_card_bm);
jmasa@2188 1921 if (G1CollectedHeap::use_parallel_gc_threads()) {
jmasa@3294 1922 g1h->set_par_threads((int)n_workers);
ysr@777 1923 g1h->workers()->run_task(&g1_par_scrub_rs_task);
ysr@777 1924 g1h->set_par_threads(0);
tonyp@790 1925
tonyp@790 1926 assert(g1h->check_heap_region_claim_values(
tonyp@790 1927 HeapRegion::ScrubRemSetClaimValue),
tonyp@790 1928 "sanity check");
ysr@777 1929 } else {
ysr@777 1930 g1_par_scrub_rs_task.work(0);
ysr@777 1931 }
ysr@777 1932
ysr@777 1933 double rs_scrub_end = os::elapsedTime();
ysr@777 1934 double this_rs_scrub_time = (rs_scrub_end - rs_scrub_start);
ysr@777 1935 _total_rs_scrub_time += this_rs_scrub_time;
ysr@777 1936 }
ysr@777 1937
ysr@777 1938 // this will also free any regions totally full of garbage objects,
ysr@777 1939 // and sort the regions.
jmasa@3294 1940 g1h->g1_policy()->record_concurrent_mark_cleanup_end((int)n_workers);
ysr@777 1941
ysr@777 1942 // Statistics.
ysr@777 1943 double end = os::elapsedTime();
ysr@777 1944 _cleanup_times.add((end - start) * 1000.0);
ysr@777 1945
brutisso@3710 1946 if (G1Log::fine()) {
ysr@777 1947 g1h->print_size_transition(gclog_or_tty,
ysr@777 1948 start_used_bytes,
ysr@777 1949 g1h->used(),
ysr@777 1950 g1h->capacity());
ysr@777 1951 }
ysr@777 1952
johnc@3175 1953 // Clean up will have freed any regions completely full of garbage.
johnc@3175 1954 // Update the soft reference policy with the new heap occupancy.
johnc@3175 1955 Universe::update_heap_info_at_gc();
johnc@3175 1956
ysr@777 1957 // We need to make this be a "collection" so any collection pause that
ysr@777 1958 // races with it goes around and waits for completeCleanup to finish.
ysr@777 1959 g1h->increment_total_collections();
ysr@777 1960
tonyp@3457 1961 // We reclaimed old regions so we should calculate the sizes to make
tonyp@3457 1962 // sure we update the old gen/space data.
tonyp@3457 1963 g1h->g1mm()->update_sizes();
tonyp@3457 1964
johnc@1186 1965 if (VerifyDuringGC) {
ysr@1280 1966 HandleMark hm; // handle scope
ysr@1280 1967 gclog_or_tty->print(" VerifyDuringGC:(after)");
ysr@1280 1968 Universe::heap()->prepare_for_verify();
brutisso@3711 1969 Universe::verify(/* silent */ false,
johnc@2969 1970 /* option */ VerifyOption_G1UsePrevMarking);
ysr@777 1971 }
tonyp@2472 1972
tonyp@2472 1973 g1h->verify_region_sets_optional();
ysr@777 1974 }
ysr@777 1975
ysr@777 1976 void ConcurrentMark::completeCleanup() {
ysr@777 1977 if (has_aborted()) return;
ysr@777 1978
tonyp@2472 1979 G1CollectedHeap* g1h = G1CollectedHeap::heap();
tonyp@2472 1980
tonyp@2472 1981 _cleanup_list.verify_optional();
tonyp@2643 1982 FreeRegionList tmp_free_list("Tmp Free List");
tonyp@2472 1983
tonyp@2472 1984 if (G1ConcRegionFreeingVerbose) {
tonyp@2472 1985 gclog_or_tty->print_cr("G1ConcRegionFreeing [complete cleanup] : "
tonyp@3713 1986 "cleanup list has %u entries",
tonyp@2472 1987 _cleanup_list.length());
tonyp@2472 1988 }
tonyp@2472 1989
tonyp@2472 1990 // Noone else should be accessing the _cleanup_list at this point,
tonyp@2472 1991 // so it's not necessary to take any locks
tonyp@2472 1992 while (!_cleanup_list.is_empty()) {
tonyp@2472 1993 HeapRegion* hr = _cleanup_list.remove_head();
tonyp@2472 1994 assert(hr != NULL, "the list was not empty");
tonyp@2849 1995 hr->par_clear();
tonyp@2643 1996 tmp_free_list.add_as_tail(hr);
tonyp@2472 1997
tonyp@2472 1998 // Instead of adding one region at a time to the secondary_free_list,
tonyp@2472 1999 // we accumulate them in the local list and move them a few at a
tonyp@2472 2000 // time. This also cuts down on the number of notify_all() calls
tonyp@2472 2001 // we do during this process. We'll also append the local list when
tonyp@2472 2002 // _cleanup_list is empty (which means we just removed the last
tonyp@2472 2003 // region from the _cleanup_list).
tonyp@2643 2004 if ((tmp_free_list.length() % G1SecondaryFreeListAppendLength == 0) ||
tonyp@2472 2005 _cleanup_list.is_empty()) {
tonyp@2472 2006 if (G1ConcRegionFreeingVerbose) {
tonyp@2472 2007 gclog_or_tty->print_cr("G1ConcRegionFreeing [complete cleanup] : "
tonyp@3713 2008 "appending %u entries to the secondary_free_list, "
tonyp@3713 2009 "cleanup list still has %u entries",
tonyp@2643 2010 tmp_free_list.length(),
tonyp@2472 2011 _cleanup_list.length());
ysr@777 2012 }
tonyp@2472 2013
tonyp@2472 2014 {
tonyp@2472 2015 MutexLockerEx x(SecondaryFreeList_lock, Mutex::_no_safepoint_check_flag);
tonyp@2643 2016 g1h->secondary_free_list_add_as_tail(&tmp_free_list);
tonyp@2472 2017 SecondaryFreeList_lock->notify_all();
tonyp@2472 2018 }
tonyp@2472 2019
tonyp@2472 2020 if (G1StressConcRegionFreeing) {
tonyp@2472 2021 for (uintx i = 0; i < G1StressConcRegionFreeingDelayMillis; ++i) {
tonyp@2472 2022 os::sleep(Thread::current(), (jlong) 1, false);
tonyp@2472 2023 }
tonyp@2472 2024 }
ysr@777 2025 }
ysr@777 2026 }
tonyp@2643 2027 assert(tmp_free_list.is_empty(), "post-condition");
ysr@777 2028 }
ysr@777 2029
johnc@2494 2030 // Support closures for reference procssing in G1
johnc@2494 2031
johnc@2379 2032 bool G1CMIsAliveClosure::do_object_b(oop obj) {
johnc@2379 2033 HeapWord* addr = (HeapWord*)obj;
johnc@2379 2034 return addr != NULL &&
johnc@2379 2035 (!_g1->is_in_g1_reserved(addr) || !_g1->is_obj_ill(obj));
johnc@2379 2036 }
ysr@777 2037
coleenp@4037 2038 class G1CMKeepAliveClosure: public ExtendedOopClosure {
ysr@777 2039 G1CollectedHeap* _g1;
ysr@777 2040 ConcurrentMark* _cm;
ysr@777 2041 public:
johnc@3463 2042 G1CMKeepAliveClosure(G1CollectedHeap* g1, ConcurrentMark* cm) :
johnc@3463 2043 _g1(g1), _cm(cm) {
johnc@3463 2044 assert(Thread::current()->is_VM_thread(), "otherwise fix worker id");
johnc@3463 2045 }
ysr@777 2046
ysr@1280 2047 virtual void do_oop(narrowOop* p) { do_oop_work(p); }
ysr@1280 2048 virtual void do_oop( oop* p) { do_oop_work(p); }
ysr@1280 2049
ysr@1280 2050 template <class T> void do_oop_work(T* p) {
johnc@2494 2051 oop obj = oopDesc::load_decode_heap_oop(p);
johnc@2494 2052 HeapWord* addr = (HeapWord*)obj;
johnc@2494 2053
tonyp@2973 2054 if (_cm->verbose_high()) {
johnc@2494 2055 gclog_or_tty->print_cr("\t[0] we're looking at location "
tonyp@2973 2056 "*"PTR_FORMAT" = "PTR_FORMAT,
tonyp@2973 2057 p, (void*) obj);
tonyp@2973 2058 }
johnc@2494 2059
johnc@2494 2060 if (_g1->is_in_g1_reserved(addr) && _g1->is_obj_ill(obj)) {
johnc@3463 2061 _cm->mark_and_count(obj);
johnc@2494 2062 _cm->mark_stack_push(obj);
ysr@777 2063 }
ysr@777 2064 }
ysr@777 2065 };
ysr@777 2066
ysr@777 2067 class G1CMDrainMarkingStackClosure: public VoidClosure {
johnc@3463 2068 ConcurrentMark* _cm;
ysr@777 2069 CMMarkStack* _markStack;
ysr@777 2070 G1CMKeepAliveClosure* _oopClosure;
ysr@777 2071 public:
johnc@3463 2072 G1CMDrainMarkingStackClosure(ConcurrentMark* cm, CMMarkStack* markStack,
ysr@777 2073 G1CMKeepAliveClosure* oopClosure) :
johnc@3463 2074 _cm(cm),
ysr@777 2075 _markStack(markStack),
johnc@3463 2076 _oopClosure(oopClosure) { }
ysr@777 2077
ysr@777 2078 void do_void() {
coleenp@4037 2079 _markStack->drain(_oopClosure, _cm->nextMarkBitMap(), false);
ysr@777 2080 }
ysr@777 2081 };
ysr@777 2082
johnc@2494 2083 // 'Keep Alive' closure used by parallel reference processing.
johnc@2494 2084 // An instance of this closure is used in the parallel reference processing
johnc@2494 2085 // code rather than an instance of G1CMKeepAliveClosure. We could have used
johnc@2494 2086 // the G1CMKeepAliveClosure as it is MT-safe. Also reference objects are
johnc@2494 2087 // placed on to discovered ref lists once so we can mark and push with no
johnc@2494 2088 // need to check whether the object has already been marked. Using the
johnc@2494 2089 // G1CMKeepAliveClosure would mean, however, having all the worker threads
johnc@2494 2090 // operating on the global mark stack. This means that an individual
johnc@2494 2091 // worker would be doing lock-free pushes while it processes its own
johnc@2494 2092 // discovered ref list followed by drain call. If the discovered ref lists
johnc@2494 2093 // are unbalanced then this could cause interference with the other
johnc@2494 2094 // workers. Using a CMTask (and its embedded local data structures)
johnc@2494 2095 // avoids that potential interference.
johnc@2494 2096 class G1CMParKeepAliveAndDrainClosure: public OopClosure {
johnc@2494 2097 ConcurrentMark* _cm;
johnc@2494 2098 CMTask* _task;
johnc@2494 2099 int _ref_counter_limit;
johnc@2494 2100 int _ref_counter;
johnc@2494 2101 public:
johnc@3292 2102 G1CMParKeepAliveAndDrainClosure(ConcurrentMark* cm, CMTask* task) :
johnc@3292 2103 _cm(cm), _task(task),
johnc@3292 2104 _ref_counter_limit(G1RefProcDrainInterval) {
johnc@2494 2105 assert(_ref_counter_limit > 0, "sanity");
johnc@2494 2106 _ref_counter = _ref_counter_limit;
johnc@2494 2107 }
johnc@2494 2108
johnc@2494 2109 virtual void do_oop(narrowOop* p) { do_oop_work(p); }
johnc@2494 2110 virtual void do_oop( oop* p) { do_oop_work(p); }
johnc@2494 2111
johnc@2494 2112 template <class T> void do_oop_work(T* p) {
johnc@2494 2113 if (!_cm->has_overflown()) {
johnc@2494 2114 oop obj = oopDesc::load_decode_heap_oop(p);
tonyp@2973 2115 if (_cm->verbose_high()) {
johnc@4173 2116 gclog_or_tty->print_cr("\t[%u] we're looking at location "
johnc@2494 2117 "*"PTR_FORMAT" = "PTR_FORMAT,
johnc@4173 2118 _task->worker_id(), p, (void*) obj);
tonyp@2973 2119 }
johnc@2494 2120
johnc@2494 2121 _task->deal_with_reference(obj);
johnc@2494 2122 _ref_counter--;
johnc@2494 2123
johnc@2494 2124 if (_ref_counter == 0) {
johnc@2494 2125 // We have dealt with _ref_counter_limit references, pushing them and objects
johnc@2494 2126 // reachable from them on to the local stack (and possibly the global stack).
johnc@2494 2127 // Call do_marking_step() to process these entries. We call the routine in a
johnc@2494 2128 // loop, which we'll exit if there's nothing more to do (i.e. we're done
johnc@2494 2129 // with the entries that we've pushed as a result of the deal_with_reference
johnc@2494 2130 // calls above) or we overflow.
johnc@2494 2131 // Note: CMTask::do_marking_step() can set the CMTask::has_aborted() flag
johnc@2494 2132 // while there may still be some work to do. (See the comment at the
johnc@2494 2133 // beginning of CMTask::do_marking_step() for those conditions - one of which
johnc@2494 2134 // is reaching the specified time target.) It is only when
johnc@2494 2135 // CMTask::do_marking_step() returns without setting the has_aborted() flag
johnc@2494 2136 // that the marking has completed.
johnc@2494 2137 do {
johnc@2494 2138 double mark_step_duration_ms = G1ConcMarkStepDurationMillis;
johnc@2494 2139 _task->do_marking_step(mark_step_duration_ms,
johnc@2494 2140 false /* do_stealing */,
johnc@2494 2141 false /* do_termination */);
johnc@2494 2142 } while (_task->has_aborted() && !_cm->has_overflown());
johnc@2494 2143 _ref_counter = _ref_counter_limit;
johnc@2494 2144 }
johnc@2494 2145 } else {
tonyp@2973 2146 if (_cm->verbose_high()) {
johnc@4173 2147 gclog_or_tty->print_cr("\t[%u] CM Overflow", _task->worker_id());
tonyp@2973 2148 }
johnc@2494 2149 }
johnc@2494 2150 }
johnc@2494 2151 };
johnc@2494 2152
johnc@2494 2153 class G1CMParDrainMarkingStackClosure: public VoidClosure {
johnc@2494 2154 ConcurrentMark* _cm;
johnc@2494 2155 CMTask* _task;
johnc@2494 2156 public:
johnc@2494 2157 G1CMParDrainMarkingStackClosure(ConcurrentMark* cm, CMTask* task) :
johnc@3463 2158 _cm(cm), _task(task) { }
johnc@2494 2159
johnc@2494 2160 void do_void() {
johnc@2494 2161 do {
tonyp@2973 2162 if (_cm->verbose_high()) {
johnc@4173 2163 gclog_or_tty->print_cr("\t[%u] Drain: Calling do marking_step",
johnc@4173 2164 _task->worker_id());
tonyp@2973 2165 }
johnc@2494 2166
johnc@2494 2167 // We call CMTask::do_marking_step() to completely drain the local and
johnc@2494 2168 // global marking stacks. The routine is called in a loop, which we'll
johnc@2494 2169 // exit if there's nothing more to do (i.e. we'completely drained the
johnc@2494 2170 // entries that were pushed as a result of applying the
johnc@2494 2171 // G1CMParKeepAliveAndDrainClosure to the entries on the discovered ref
johnc@2494 2172 // lists above) or we overflow the global marking stack.
johnc@2494 2173 // Note: CMTask::do_marking_step() can set the CMTask::has_aborted() flag
johnc@2494 2174 // while there may still be some work to do. (See the comment at the
johnc@2494 2175 // beginning of CMTask::do_marking_step() for those conditions - one of which
johnc@2494 2176 // is reaching the specified time target.) It is only when
johnc@2494 2177 // CMTask::do_marking_step() returns without setting the has_aborted() flag
johnc@2494 2178 // that the marking has completed.
johnc@2494 2179
johnc@2494 2180 _task->do_marking_step(1000000000.0 /* something very large */,
johnc@2494 2181 true /* do_stealing */,
johnc@2494 2182 true /* do_termination */);
johnc@2494 2183 } while (_task->has_aborted() && !_cm->has_overflown());
johnc@2494 2184 }
johnc@2494 2185 };
johnc@2494 2186
johnc@3175 2187 // Implementation of AbstractRefProcTaskExecutor for parallel
johnc@3175 2188 // reference processing at the end of G1 concurrent marking
johnc@3175 2189
johnc@3175 2190 class G1CMRefProcTaskExecutor: public AbstractRefProcTaskExecutor {
johnc@2494 2191 private:
johnc@2494 2192 G1CollectedHeap* _g1h;
johnc@2494 2193 ConcurrentMark* _cm;
johnc@2494 2194 WorkGang* _workers;
johnc@2494 2195 int _active_workers;
johnc@2494 2196
johnc@2494 2197 public:
johnc@3175 2198 G1CMRefProcTaskExecutor(G1CollectedHeap* g1h,
johnc@2494 2199 ConcurrentMark* cm,
johnc@2494 2200 WorkGang* workers,
johnc@2494 2201 int n_workers) :
johnc@3292 2202 _g1h(g1h), _cm(cm),
johnc@3292 2203 _workers(workers), _active_workers(n_workers) { }
johnc@2494 2204
johnc@2494 2205 // Executes the given task using concurrent marking worker threads.
johnc@2494 2206 virtual void execute(ProcessTask& task);
johnc@2494 2207 virtual void execute(EnqueueTask& task);
johnc@2494 2208 };
johnc@2494 2209
johnc@3175 2210 class G1CMRefProcTaskProxy: public AbstractGangTask {
johnc@2494 2211 typedef AbstractRefProcTaskExecutor::ProcessTask ProcessTask;
johnc@2494 2212 ProcessTask& _proc_task;
johnc@2494 2213 G1CollectedHeap* _g1h;
johnc@2494 2214 ConcurrentMark* _cm;
johnc@2494 2215
johnc@2494 2216 public:
johnc@3175 2217 G1CMRefProcTaskProxy(ProcessTask& proc_task,
johnc@2494 2218 G1CollectedHeap* g1h,
johnc@3292 2219 ConcurrentMark* cm) :
johnc@2494 2220 AbstractGangTask("Process reference objects in parallel"),
johnc@3292 2221 _proc_task(proc_task), _g1h(g1h), _cm(cm) { }
johnc@2494 2222
jmasa@3357 2223 virtual void work(uint worker_id) {
jmasa@3357 2224 CMTask* marking_task = _cm->task(worker_id);
johnc@2494 2225 G1CMIsAliveClosure g1_is_alive(_g1h);
johnc@3292 2226 G1CMParKeepAliveAndDrainClosure g1_par_keep_alive(_cm, marking_task);
johnc@2494 2227 G1CMParDrainMarkingStackClosure g1_par_drain(_cm, marking_task);
johnc@2494 2228
jmasa@3357 2229 _proc_task.work(worker_id, g1_is_alive, g1_par_keep_alive, g1_par_drain);
johnc@2494 2230 }
johnc@2494 2231 };
johnc@2494 2232
johnc@3175 2233 void G1CMRefProcTaskExecutor::execute(ProcessTask& proc_task) {
johnc@2494 2234 assert(_workers != NULL, "Need parallel worker threads.");
johnc@2494 2235
johnc@3292 2236 G1CMRefProcTaskProxy proc_task_proxy(proc_task, _g1h, _cm);
johnc@2494 2237
johnc@2494 2238 // We need to reset the phase for each task execution so that
johnc@2494 2239 // the termination protocol of CMTask::do_marking_step works.
johnc@2494 2240 _cm->set_phase(_active_workers, false /* concurrent */);
johnc@2494 2241 _g1h->set_par_threads(_active_workers);
johnc@2494 2242 _workers->run_task(&proc_task_proxy);
johnc@2494 2243 _g1h->set_par_threads(0);
johnc@2494 2244 }
johnc@2494 2245
johnc@3175 2246 class G1CMRefEnqueueTaskProxy: public AbstractGangTask {
johnc@2494 2247 typedef AbstractRefProcTaskExecutor::EnqueueTask EnqueueTask;
johnc@2494 2248 EnqueueTask& _enq_task;
johnc@2494 2249
johnc@2494 2250 public:
johnc@3175 2251 G1CMRefEnqueueTaskProxy(EnqueueTask& enq_task) :
johnc@2494 2252 AbstractGangTask("Enqueue reference objects in parallel"),
johnc@3292 2253 _enq_task(enq_task) { }
johnc@2494 2254
jmasa@3357 2255 virtual void work(uint worker_id) {
jmasa@3357 2256 _enq_task.work(worker_id);
johnc@2494 2257 }
johnc@2494 2258 };
johnc@2494 2259
johnc@3175 2260 void G1CMRefProcTaskExecutor::execute(EnqueueTask& enq_task) {
johnc@2494 2261 assert(_workers != NULL, "Need parallel worker threads.");
johnc@2494 2262
johnc@3175 2263 G1CMRefEnqueueTaskProxy enq_task_proxy(enq_task);
johnc@2494 2264
johnc@2494 2265 _g1h->set_par_threads(_active_workers);
johnc@2494 2266 _workers->run_task(&enq_task_proxy);
johnc@2494 2267 _g1h->set_par_threads(0);
johnc@2494 2268 }
johnc@2494 2269
ysr@777 2270 void ConcurrentMark::weakRefsWork(bool clear_all_soft_refs) {
ysr@777 2271 ResourceMark rm;
ysr@777 2272 HandleMark hm;
johnc@3171 2273
johnc@3171 2274 G1CollectedHeap* g1h = G1CollectedHeap::heap();
johnc@3171 2275
johnc@3171 2276 // Is alive closure.
johnc@3171 2277 G1CMIsAliveClosure g1_is_alive(g1h);
johnc@3171 2278
johnc@3171 2279 // Inner scope to exclude the cleaning of the string and symbol
johnc@3171 2280 // tables from the displayed time.
johnc@3171 2281 {
brutisso@3710 2282 if (G1Log::finer()) {
johnc@3171 2283 gclog_or_tty->put(' ');
johnc@3171 2284 }
brutisso@3710 2285 TraceTime t("GC ref-proc", G1Log::finer(), false, gclog_or_tty);
johnc@3171 2286
johnc@3175 2287 ReferenceProcessor* rp = g1h->ref_processor_cm();
johnc@3171 2288
johnc@3171 2289 // See the comment in G1CollectedHeap::ref_processing_init()
johnc@3171 2290 // about how reference processing currently works in G1.
johnc@3171 2291
johnc@3171 2292 // Process weak references.
johnc@3171 2293 rp->setup_policy(clear_all_soft_refs);
johnc@3171 2294 assert(_markStack.isEmpty(), "mark stack should be empty");
johnc@3171 2295
johnc@3463 2296 G1CMKeepAliveClosure g1_keep_alive(g1h, this);
johnc@3171 2297 G1CMDrainMarkingStackClosure
johnc@3463 2298 g1_drain_mark_stack(this, &_markStack, &g1_keep_alive);
johnc@3171 2299
johnc@3171 2300 // We use the work gang from the G1CollectedHeap and we utilize all
johnc@3171 2301 // the worker threads.
jmasa@3357 2302 uint active_workers = g1h->workers() ? g1h->workers()->active_workers() : 1U;
johnc@4173 2303 active_workers = MAX2(MIN2(active_workers, _max_worker_id), 1U);
johnc@3171 2304
johnc@3292 2305 G1CMRefProcTaskExecutor par_task_executor(g1h, this,
johnc@3175 2306 g1h->workers(), active_workers);
johnc@3171 2307
johnc@3171 2308 if (rp->processing_is_mt()) {
johnc@3171 2309 // Set the degree of MT here. If the discovery is done MT, there
johnc@3171 2310 // may have been a different number of threads doing the discovery
johnc@3171 2311 // and a different number of discovered lists may have Ref objects.
johnc@3171 2312 // That is OK as long as the Reference lists are balanced (see
johnc@3171 2313 // balance_all_queues() and balance_queues()).
johnc@3171 2314 rp->set_active_mt_degree(active_workers);
johnc@3171 2315
johnc@3171 2316 rp->process_discovered_references(&g1_is_alive,
johnc@2494 2317 &g1_keep_alive,
johnc@2494 2318 &g1_drain_mark_stack,
johnc@2494 2319 &par_task_executor);
johnc@2494 2320
johnc@3171 2321 // The work routines of the parallel keep_alive and drain_marking_stack
johnc@3171 2322 // will set the has_overflown flag if we overflow the global marking
johnc@3171 2323 // stack.
johnc@3171 2324 } else {
johnc@3171 2325 rp->process_discovered_references(&g1_is_alive,
johnc@3171 2326 &g1_keep_alive,
johnc@3171 2327 &g1_drain_mark_stack,
johnc@3171 2328 NULL);
johnc@3171 2329 }
johnc@3171 2330
johnc@3171 2331 assert(_markStack.overflow() || _markStack.isEmpty(),
johnc@3171 2332 "mark stack should be empty (unless it overflowed)");
johnc@3171 2333 if (_markStack.overflow()) {
johnc@3171 2334 // Should have been done already when we tried to push an
johnc@3171 2335 // entry on to the global mark stack. But let's do it again.
johnc@3171 2336 set_has_overflown();
johnc@3171 2337 }
johnc@3171 2338
johnc@3171 2339 if (rp->processing_is_mt()) {
johnc@3171 2340 assert(rp->num_q() == active_workers, "why not");
johnc@3171 2341 rp->enqueue_discovered_references(&par_task_executor);
johnc@3171 2342 } else {
johnc@3171 2343 rp->enqueue_discovered_references();
johnc@3171 2344 }
johnc@3171 2345
johnc@3171 2346 rp->verify_no_references_recorded();
johnc@3175 2347 assert(!rp->discovery_enabled(), "Post condition");
johnc@2494 2348 }
johnc@2494 2349
coleenp@2497 2350 // Now clean up stale oops in StringTable
johnc@2379 2351 StringTable::unlink(&g1_is_alive);
coleenp@2497 2352 // Clean up unreferenced symbols in symbol table.
coleenp@2497 2353 SymbolTable::unlink();
ysr@777 2354 }
ysr@777 2355
ysr@777 2356 void ConcurrentMark::swapMarkBitMaps() {
ysr@777 2357 CMBitMapRO* temp = _prevMarkBitMap;
ysr@777 2358 _prevMarkBitMap = (CMBitMapRO*)_nextMarkBitMap;
ysr@777 2359 _nextMarkBitMap = (CMBitMap*) temp;
ysr@777 2360 }
ysr@777 2361
ysr@777 2362 class CMRemarkTask: public AbstractGangTask {
ysr@777 2363 private:
ysr@777 2364 ConcurrentMark *_cm;
ysr@777 2365
ysr@777 2366 public:
jmasa@3357 2367 void work(uint worker_id) {
ysr@777 2368 // Since all available tasks are actually started, we should
ysr@777 2369 // only proceed if we're supposed to be actived.
jmasa@3357 2370 if (worker_id < _cm->active_tasks()) {
jmasa@3357 2371 CMTask* task = _cm->task(worker_id);
ysr@777 2372 task->record_start_time();
ysr@777 2373 do {
johnc@2494 2374 task->do_marking_step(1000000000.0 /* something very large */,
johnc@2494 2375 true /* do_stealing */,
johnc@2494 2376 true /* do_termination */);
ysr@777 2377 } while (task->has_aborted() && !_cm->has_overflown());
ysr@777 2378 // If we overflow, then we do not want to restart. We instead
ysr@777 2379 // want to abort remark and do concurrent marking again.
ysr@777 2380 task->record_end_time();
ysr@777 2381 }
ysr@777 2382 }
ysr@777 2383
johnc@3338 2384 CMRemarkTask(ConcurrentMark* cm, int active_workers) :
jmasa@3294 2385 AbstractGangTask("Par Remark"), _cm(cm) {
johnc@3338 2386 _cm->terminator()->reset_for_reuse(active_workers);
jmasa@3294 2387 }
ysr@777 2388 };
ysr@777 2389
ysr@777 2390 void ConcurrentMark::checkpointRootsFinalWork() {
ysr@777 2391 ResourceMark rm;
ysr@777 2392 HandleMark hm;
ysr@777 2393 G1CollectedHeap* g1h = G1CollectedHeap::heap();
ysr@777 2394
ysr@777 2395 g1h->ensure_parsability(false);
ysr@777 2396
jmasa@2188 2397 if (G1CollectedHeap::use_parallel_gc_threads()) {
jrose@1424 2398 G1CollectedHeap::StrongRootsScope srs(g1h);
jmasa@3294 2399 // this is remark, so we'll use up all active threads
jmasa@3357 2400 uint active_workers = g1h->workers()->active_workers();
jmasa@3294 2401 if (active_workers == 0) {
jmasa@3294 2402 assert(active_workers > 0, "Should have been set earlier");
jmasa@3357 2403 active_workers = (uint) ParallelGCThreads;
jmasa@3294 2404 g1h->workers()->set_active_workers(active_workers);
jmasa@3294 2405 }
johnc@2494 2406 set_phase(active_workers, false /* concurrent */);
jmasa@3294 2407 // Leave _parallel_marking_threads at it's
jmasa@3294 2408 // value originally calculated in the ConcurrentMark
jmasa@3294 2409 // constructor and pass values of the active workers
jmasa@3294 2410 // through the gang in the task.
ysr@777 2411
johnc@3338 2412 CMRemarkTask remarkTask(this, active_workers);
jmasa@3294 2413 g1h->set_par_threads(active_workers);
ysr@777 2414 g1h->workers()->run_task(&remarkTask);
ysr@777 2415 g1h->set_par_threads(0);
ysr@777 2416 } else {
jrose@1424 2417 G1CollectedHeap::StrongRootsScope srs(g1h);
ysr@777 2418 // this is remark, so we'll use up all available threads
jmasa@3357 2419 uint active_workers = 1;
johnc@2494 2420 set_phase(active_workers, false /* concurrent */);
ysr@777 2421
johnc@3338 2422 CMRemarkTask remarkTask(this, active_workers);
ysr@777 2423 // We will start all available threads, even if we decide that the
ysr@777 2424 // active_workers will be fewer. The extra ones will just bail out
ysr@777 2425 // immediately.
ysr@777 2426 remarkTask.work(0);
ysr@777 2427 }
tonyp@1458 2428 SATBMarkQueueSet& satb_mq_set = JavaThread::satb_mark_queue_set();
tonyp@1458 2429 guarantee(satb_mq_set.completed_buffers_num() == 0, "invariant");
ysr@777 2430
ysr@777 2431 print_stats();
ysr@777 2432
ysr@777 2433 #if VERIFY_OBJS_PROCESSED
ysr@777 2434 if (_scan_obj_cl.objs_processed != ThreadLocalObjQueue::objs_enqueued) {
ysr@777 2435 gclog_or_tty->print_cr("Processed = %d, enqueued = %d.",
ysr@777 2436 _scan_obj_cl.objs_processed,
ysr@777 2437 ThreadLocalObjQueue::objs_enqueued);
ysr@777 2438 guarantee(_scan_obj_cl.objs_processed ==
ysr@777 2439 ThreadLocalObjQueue::objs_enqueued,
ysr@777 2440 "Different number of objs processed and enqueued.");
ysr@777 2441 }
ysr@777 2442 #endif
ysr@777 2443 }
ysr@777 2444
tonyp@1479 2445 #ifndef PRODUCT
tonyp@1479 2446
tonyp@1823 2447 class PrintReachableOopClosure: public OopClosure {
ysr@777 2448 private:
ysr@777 2449 G1CollectedHeap* _g1h;
ysr@777 2450 outputStream* _out;
johnc@2969 2451 VerifyOption _vo;
tonyp@1823 2452 bool _all;
ysr@777 2453
ysr@777 2454 public:
johnc@2969 2455 PrintReachableOopClosure(outputStream* out,
johnc@2969 2456 VerifyOption vo,
tonyp@1823 2457 bool all) :
tonyp@1479 2458 _g1h(G1CollectedHeap::heap()),
johnc@2969 2459 _out(out), _vo(vo), _all(all) { }
ysr@777 2460
ysr@1280 2461 void do_oop(narrowOop* p) { do_oop_work(p); }
ysr@1280 2462 void do_oop( oop* p) { do_oop_work(p); }
ysr@1280 2463
ysr@1280 2464 template <class T> void do_oop_work(T* p) {
ysr@1280 2465 oop obj = oopDesc::load_decode_heap_oop(p);
ysr@777 2466 const char* str = NULL;
ysr@777 2467 const char* str2 = "";
ysr@777 2468
tonyp@1823 2469 if (obj == NULL) {
tonyp@1823 2470 str = "";
tonyp@1823 2471 } else if (!_g1h->is_in_g1_reserved(obj)) {
tonyp@1823 2472 str = " O";
tonyp@1823 2473 } else {
ysr@777 2474 HeapRegion* hr = _g1h->heap_region_containing(obj);
tonyp@1458 2475 guarantee(hr != NULL, "invariant");
tonyp@3957 2476 bool over_tams = _g1h->allocated_since_marking(obj, hr, _vo);
tonyp@3957 2477 bool marked = _g1h->is_marked(obj, _vo);
tonyp@1479 2478
tonyp@1479 2479 if (over_tams) {
tonyp@1823 2480 str = " >";
tonyp@1823 2481 if (marked) {
ysr@777 2482 str2 = " AND MARKED";
tonyp@1479 2483 }
tonyp@1823 2484 } else if (marked) {
tonyp@1823 2485 str = " M";
tonyp@1479 2486 } else {
tonyp@1823 2487 str = " NOT";
tonyp@1479 2488 }
ysr@777 2489 }
ysr@777 2490
tonyp@1823 2491 _out->print_cr(" "PTR_FORMAT": "PTR_FORMAT"%s%s",
ysr@777 2492 p, (void*) obj, str, str2);
ysr@777 2493 }
ysr@777 2494 };
ysr@777 2495
tonyp@1823 2496 class PrintReachableObjectClosure : public ObjectClosure {
ysr@777 2497 private:
johnc@2969 2498 G1CollectedHeap* _g1h;
johnc@2969 2499 outputStream* _out;
johnc@2969 2500 VerifyOption _vo;
johnc@2969 2501 bool _all;
johnc@2969 2502 HeapRegion* _hr;
ysr@777 2503
ysr@777 2504 public:
johnc@2969 2505 PrintReachableObjectClosure(outputStream* out,
johnc@2969 2506 VerifyOption vo,
tonyp@1823 2507 bool all,
tonyp@1823 2508 HeapRegion* hr) :
johnc@2969 2509 _g1h(G1CollectedHeap::heap()),
johnc@2969 2510 _out(out), _vo(vo), _all(all), _hr(hr) { }
tonyp@1823 2511
tonyp@1823 2512 void do_object(oop o) {
tonyp@3957 2513 bool over_tams = _g1h->allocated_since_marking(o, _hr, _vo);
tonyp@3957 2514 bool marked = _g1h->is_marked(o, _vo);
tonyp@1823 2515 bool print_it = _all || over_tams || marked;
tonyp@1823 2516
tonyp@1823 2517 if (print_it) {
tonyp@1823 2518 _out->print_cr(" "PTR_FORMAT"%s",
tonyp@1823 2519 o, (over_tams) ? " >" : (marked) ? " M" : "");
johnc@2969 2520 PrintReachableOopClosure oopCl(_out, _vo, _all);
coleenp@4037 2521 o->oop_iterate_no_header(&oopCl);
tonyp@1823 2522 }
ysr@777 2523 }
ysr@777 2524 };
ysr@777 2525
tonyp@1823 2526 class PrintReachableRegionClosure : public HeapRegionClosure {
ysr@777 2527 private:
tonyp@3957 2528 G1CollectedHeap* _g1h;
tonyp@3957 2529 outputStream* _out;
tonyp@3957 2530 VerifyOption _vo;
tonyp@3957 2531 bool _all;
ysr@777 2532
ysr@777 2533 public:
ysr@777 2534 bool doHeapRegion(HeapRegion* hr) {
ysr@777 2535 HeapWord* b = hr->bottom();
ysr@777 2536 HeapWord* e = hr->end();
ysr@777 2537 HeapWord* t = hr->top();
tonyp@3957 2538 HeapWord* p = _g1h->top_at_mark_start(hr, _vo);
ysr@777 2539 _out->print_cr("** ["PTR_FORMAT", "PTR_FORMAT"] top: "PTR_FORMAT" "
tonyp@1479 2540 "TAMS: "PTR_FORMAT, b, e, t, p);
tonyp@1823 2541 _out->cr();
tonyp@1823 2542
tonyp@1823 2543 HeapWord* from = b;
tonyp@1823 2544 HeapWord* to = t;
tonyp@1823 2545
tonyp@1823 2546 if (to > from) {
tonyp@1823 2547 _out->print_cr("Objects in ["PTR_FORMAT", "PTR_FORMAT"]", from, to);
tonyp@1823 2548 _out->cr();
johnc@2969 2549 PrintReachableObjectClosure ocl(_out, _vo, _all, hr);
tonyp@1823 2550 hr->object_iterate_mem_careful(MemRegion(from, to), &ocl);
tonyp@1823 2551 _out->cr();
tonyp@1823 2552 }
ysr@777 2553
ysr@777 2554 return false;
ysr@777 2555 }
ysr@777 2556
johnc@2969 2557 PrintReachableRegionClosure(outputStream* out,
johnc@2969 2558 VerifyOption vo,
tonyp@1823 2559 bool all) :
tonyp@3957 2560 _g1h(G1CollectedHeap::heap()), _out(out), _vo(vo), _all(all) { }
ysr@777 2561 };
ysr@777 2562
tonyp@1823 2563 void ConcurrentMark::print_reachable(const char* str,
johnc@2969 2564 VerifyOption vo,
tonyp@1823 2565 bool all) {
tonyp@1823 2566 gclog_or_tty->cr();
tonyp@1823 2567 gclog_or_tty->print_cr("== Doing heap dump... ");
tonyp@1479 2568
tonyp@1479 2569 if (G1PrintReachableBaseFile == NULL) {
tonyp@1479 2570 gclog_or_tty->print_cr(" #### error: no base file defined");
tonyp@1479 2571 return;
tonyp@1479 2572 }
tonyp@1479 2573
tonyp@1479 2574 if (strlen(G1PrintReachableBaseFile) + 1 + strlen(str) >
tonyp@1479 2575 (JVM_MAXPATHLEN - 1)) {
tonyp@1479 2576 gclog_or_tty->print_cr(" #### error: file name too long");
tonyp@1479 2577 return;
tonyp@1479 2578 }
tonyp@1479 2579
tonyp@1479 2580 char file_name[JVM_MAXPATHLEN];
tonyp@1479 2581 sprintf(file_name, "%s.%s", G1PrintReachableBaseFile, str);
tonyp@1479 2582 gclog_or_tty->print_cr(" dumping to file %s", file_name);
tonyp@1479 2583
tonyp@1479 2584 fileStream fout(file_name);
tonyp@1479 2585 if (!fout.is_open()) {
tonyp@1479 2586 gclog_or_tty->print_cr(" #### error: could not open file");
tonyp@1479 2587 return;
tonyp@1479 2588 }
tonyp@1479 2589
tonyp@1479 2590 outputStream* out = &fout;
tonyp@3957 2591 out->print_cr("-- USING %s", _g1h->top_at_mark_start_str(vo));
tonyp@1479 2592 out->cr();
tonyp@1479 2593
tonyp@1823 2594 out->print_cr("--- ITERATING OVER REGIONS");
tonyp@1479 2595 out->cr();
johnc@2969 2596 PrintReachableRegionClosure rcl(out, vo, all);
ysr@777 2597 _g1h->heap_region_iterate(&rcl);
tonyp@1479 2598 out->cr();
tonyp@1479 2599
tonyp@1479 2600 gclog_or_tty->print_cr(" done");
tonyp@1823 2601 gclog_or_tty->flush();
ysr@777 2602 }
ysr@777 2603
tonyp@1479 2604 #endif // PRODUCT
tonyp@1479 2605
tonyp@3416 2606 void ConcurrentMark::clearRangePrevBitmap(MemRegion mr) {
ysr@777 2607 // Note we are overriding the read-only view of the prev map here, via
ysr@777 2608 // the cast.
ysr@777 2609 ((CMBitMap*)_prevMarkBitMap)->clearRange(mr);
tonyp@3416 2610 }
tonyp@3416 2611
tonyp@3416 2612 void ConcurrentMark::clearRangeNextBitmap(MemRegion mr) {
ysr@777 2613 _nextMarkBitMap->clearRange(mr);
ysr@777 2614 }
ysr@777 2615
tonyp@3416 2616 void ConcurrentMark::clearRangeBothBitmaps(MemRegion mr) {
tonyp@3416 2617 clearRangePrevBitmap(mr);
tonyp@3416 2618 clearRangeNextBitmap(mr);
tonyp@3416 2619 }
tonyp@3416 2620
ysr@777 2621 HeapRegion*
johnc@4173 2622 ConcurrentMark::claim_region(uint worker_id) {
ysr@777 2623 // "checkpoint" the finger
ysr@777 2624 HeapWord* finger = _finger;
ysr@777 2625
ysr@777 2626 // _heap_end will not change underneath our feet; it only changes at
ysr@777 2627 // yield points.
ysr@777 2628 while (finger < _heap_end) {
tonyp@1458 2629 assert(_g1h->is_in_g1_reserved(finger), "invariant");
ysr@777 2630
tonyp@2968 2631 // Note on how this code handles humongous regions. In the
tonyp@2968 2632 // normal case the finger will reach the start of a "starts
tonyp@2968 2633 // humongous" (SH) region. Its end will either be the end of the
tonyp@2968 2634 // last "continues humongous" (CH) region in the sequence, or the
tonyp@2968 2635 // standard end of the SH region (if the SH is the only region in
tonyp@2968 2636 // the sequence). That way claim_region() will skip over the CH
tonyp@2968 2637 // regions. However, there is a subtle race between a CM thread
tonyp@2968 2638 // executing this method and a mutator thread doing a humongous
tonyp@2968 2639 // object allocation. The two are not mutually exclusive as the CM
tonyp@2968 2640 // thread does not need to hold the Heap_lock when it gets
tonyp@2968 2641 // here. So there is a chance that claim_region() will come across
tonyp@2968 2642 // a free region that's in the progress of becoming a SH or a CH
tonyp@2968 2643 // region. In the former case, it will either
tonyp@2968 2644 // a) Miss the update to the region's end, in which case it will
tonyp@2968 2645 // visit every subsequent CH region, will find their bitmaps
tonyp@2968 2646 // empty, and do nothing, or
tonyp@2968 2647 // b) Will observe the update of the region's end (in which case
tonyp@2968 2648 // it will skip the subsequent CH regions).
tonyp@2968 2649 // If it comes across a region that suddenly becomes CH, the
tonyp@2968 2650 // scenario will be similar to b). So, the race between
tonyp@2968 2651 // claim_region() and a humongous object allocation might force us
tonyp@2968 2652 // to do a bit of unnecessary work (due to some unnecessary bitmap
tonyp@2968 2653 // iterations) but it should not introduce and correctness issues.
tonyp@2968 2654 HeapRegion* curr_region = _g1h->heap_region_containing_raw(finger);
ysr@777 2655 HeapWord* bottom = curr_region->bottom();
ysr@777 2656 HeapWord* end = curr_region->end();
ysr@777 2657 HeapWord* limit = curr_region->next_top_at_mark_start();
ysr@777 2658
tonyp@2968 2659 if (verbose_low()) {
johnc@4173 2660 gclog_or_tty->print_cr("[%u] curr_region = "PTR_FORMAT" "
ysr@777 2661 "["PTR_FORMAT", "PTR_FORMAT"), "
ysr@777 2662 "limit = "PTR_FORMAT,
johnc@4173 2663 worker_id, curr_region, bottom, end, limit);
tonyp@2968 2664 }
tonyp@2968 2665
tonyp@2968 2666 // Is the gap between reading the finger and doing the CAS too long?
tonyp@2968 2667 HeapWord* res = (HeapWord*) Atomic::cmpxchg_ptr(end, &_finger, finger);
ysr@777 2668 if (res == finger) {
ysr@777 2669 // we succeeded
ysr@777 2670
ysr@777 2671 // notice that _finger == end cannot be guaranteed here since,
ysr@777 2672 // someone else might have moved the finger even further
tonyp@1458 2673 assert(_finger >= end, "the finger should have moved forward");
ysr@777 2674
tonyp@2973 2675 if (verbose_low()) {
johnc@4173 2676 gclog_or_tty->print_cr("[%u] we were successful with region = "
johnc@4173 2677 PTR_FORMAT, worker_id, curr_region);
tonyp@2973 2678 }
ysr@777 2679
ysr@777 2680 if (limit > bottom) {
tonyp@2973 2681 if (verbose_low()) {
johnc@4173 2682 gclog_or_tty->print_cr("[%u] region "PTR_FORMAT" is not empty, "
johnc@4173 2683 "returning it ", worker_id, curr_region);
tonyp@2973 2684 }
ysr@777 2685 return curr_region;
ysr@777 2686 } else {
tonyp@1458 2687 assert(limit == bottom,
tonyp@1458 2688 "the region limit should be at bottom");
tonyp@2973 2689 if (verbose_low()) {
johnc@4173 2690 gclog_or_tty->print_cr("[%u] region "PTR_FORMAT" is empty, "
johnc@4173 2691 "returning NULL", worker_id, curr_region);
tonyp@2973 2692 }
ysr@777 2693 // we return NULL and the caller should try calling
ysr@777 2694 // claim_region() again.
ysr@777 2695 return NULL;
ysr@777 2696 }
ysr@777 2697 } else {
tonyp@1458 2698 assert(_finger > finger, "the finger should have moved forward");
tonyp@2973 2699 if (verbose_low()) {
johnc@4173 2700 gclog_or_tty->print_cr("[%u] somebody else moved the finger, "
ysr@777 2701 "global finger = "PTR_FORMAT", "
ysr@777 2702 "our finger = "PTR_FORMAT,
johnc@4173 2703 worker_id, _finger, finger);
tonyp@2973 2704 }
ysr@777 2705
ysr@777 2706 // read it again
ysr@777 2707 finger = _finger;
ysr@777 2708 }
ysr@777 2709 }
ysr@777 2710
ysr@777 2711 return NULL;
ysr@777 2712 }
ysr@777 2713
tonyp@3416 2714 #ifndef PRODUCT
tonyp@3416 2715 enum VerifyNoCSetOopsPhase {
tonyp@3416 2716 VerifyNoCSetOopsStack,
tonyp@3416 2717 VerifyNoCSetOopsQueues,
tonyp@3416 2718 VerifyNoCSetOopsSATBCompleted,
tonyp@3416 2719 VerifyNoCSetOopsSATBThread
tonyp@3416 2720 };
tonyp@3416 2721
tonyp@3416 2722 class VerifyNoCSetOopsClosure : public OopClosure, public ObjectClosure {
tonyp@3416 2723 private:
tonyp@3416 2724 G1CollectedHeap* _g1h;
tonyp@3416 2725 VerifyNoCSetOopsPhase _phase;
tonyp@3416 2726 int _info;
tonyp@3416 2727
tonyp@3416 2728 const char* phase_str() {
tonyp@3416 2729 switch (_phase) {
tonyp@3416 2730 case VerifyNoCSetOopsStack: return "Stack";
tonyp@3416 2731 case VerifyNoCSetOopsQueues: return "Queue";
tonyp@3416 2732 case VerifyNoCSetOopsSATBCompleted: return "Completed SATB Buffers";
tonyp@3416 2733 case VerifyNoCSetOopsSATBThread: return "Thread SATB Buffers";
tonyp@3416 2734 default: ShouldNotReachHere();
tonyp@3416 2735 }
tonyp@3416 2736 return NULL;
ysr@777 2737 }
johnc@2190 2738
tonyp@3416 2739 void do_object_work(oop obj) {
tonyp@3416 2740 guarantee(!_g1h->obj_in_cs(obj),
tonyp@3416 2741 err_msg("obj: "PTR_FORMAT" in CSet, phase: %s, info: %d",
tonyp@3416 2742 (void*) obj, phase_str(), _info));
johnc@2190 2743 }
johnc@2190 2744
tonyp@3416 2745 public:
tonyp@3416 2746 VerifyNoCSetOopsClosure() : _g1h(G1CollectedHeap::heap()) { }
tonyp@3416 2747
tonyp@3416 2748 void set_phase(VerifyNoCSetOopsPhase phase, int info = -1) {
tonyp@3416 2749 _phase = phase;
tonyp@3416 2750 _info = info;
tonyp@3416 2751 }
tonyp@3416 2752
tonyp@3416 2753 virtual void do_oop(oop* p) {
tonyp@3416 2754 oop obj = oopDesc::load_decode_heap_oop(p);
tonyp@3416 2755 do_object_work(obj);
tonyp@3416 2756 }
tonyp@3416 2757
tonyp@3416 2758 virtual void do_oop(narrowOop* p) {
tonyp@3416 2759 // We should not come across narrow oops while scanning marking
tonyp@3416 2760 // stacks and SATB buffers.
tonyp@3416 2761 ShouldNotReachHere();
tonyp@3416 2762 }
tonyp@3416 2763
tonyp@3416 2764 virtual void do_object(oop obj) {
tonyp@3416 2765 do_object_work(obj);
tonyp@3416 2766 }
tonyp@3416 2767 };
tonyp@3416 2768
tonyp@3416 2769 void ConcurrentMark::verify_no_cset_oops(bool verify_stacks,
tonyp@3416 2770 bool verify_enqueued_buffers,
tonyp@3416 2771 bool verify_thread_buffers,
tonyp@3416 2772 bool verify_fingers) {
tonyp@3416 2773 assert(SafepointSynchronize::is_at_safepoint(), "should be at a safepoint");
tonyp@3416 2774 if (!G1CollectedHeap::heap()->mark_in_progress()) {
tonyp@3416 2775 return;
tonyp@3416 2776 }
tonyp@3416 2777
tonyp@3416 2778 VerifyNoCSetOopsClosure cl;
tonyp@3416 2779
tonyp@3416 2780 if (verify_stacks) {
tonyp@3416 2781 // Verify entries on the global mark stack
tonyp@3416 2782 cl.set_phase(VerifyNoCSetOopsStack);
tonyp@3416 2783 _markStack.oops_do(&cl);
tonyp@3416 2784
tonyp@3416 2785 // Verify entries on the task queues
johnc@4173 2786 for (uint i = 0; i < _max_worker_id; i += 1) {
tonyp@3416 2787 cl.set_phase(VerifyNoCSetOopsQueues, i);
tonyp@3416 2788 OopTaskQueue* queue = _task_queues->queue(i);
tonyp@3416 2789 queue->oops_do(&cl);
tonyp@3416 2790 }
tonyp@3416 2791 }
tonyp@3416 2792
tonyp@3416 2793 SATBMarkQueueSet& satb_qs = JavaThread::satb_mark_queue_set();
tonyp@3416 2794
tonyp@3416 2795 // Verify entries on the enqueued SATB buffers
tonyp@3416 2796 if (verify_enqueued_buffers) {
tonyp@3416 2797 cl.set_phase(VerifyNoCSetOopsSATBCompleted);
tonyp@3416 2798 satb_qs.iterate_completed_buffers_read_only(&cl);
tonyp@3416 2799 }
tonyp@3416 2800
tonyp@3416 2801 // Verify entries on the per-thread SATB buffers
tonyp@3416 2802 if (verify_thread_buffers) {
tonyp@3416 2803 cl.set_phase(VerifyNoCSetOopsSATBThread);
tonyp@3416 2804 satb_qs.iterate_thread_buffers_read_only(&cl);
tonyp@3416 2805 }
tonyp@3416 2806
tonyp@3416 2807 if (verify_fingers) {
tonyp@3416 2808 // Verify the global finger
tonyp@3416 2809 HeapWord* global_finger = finger();
tonyp@3416 2810 if (global_finger != NULL && global_finger < _heap_end) {
tonyp@3416 2811 // The global finger always points to a heap region boundary. We
tonyp@3416 2812 // use heap_region_containing_raw() to get the containing region
tonyp@3416 2813 // given that the global finger could be pointing to a free region
tonyp@3416 2814 // which subsequently becomes continues humongous. If that
tonyp@3416 2815 // happens, heap_region_containing() will return the bottom of the
tonyp@3416 2816 // corresponding starts humongous region and the check below will
tonyp@3416 2817 // not hold any more.
tonyp@3416 2818 HeapRegion* global_hr = _g1h->heap_region_containing_raw(global_finger);
tonyp@3416 2819 guarantee(global_finger == global_hr->bottom(),
tonyp@3416 2820 err_msg("global finger: "PTR_FORMAT" region: "HR_FORMAT,
tonyp@3416 2821 global_finger, HR_FORMAT_PARAMS(global_hr)));
tonyp@3416 2822 }
tonyp@3416 2823
tonyp@3416 2824 // Verify the task fingers
johnc@4173 2825 assert(parallel_marking_threads() <= _max_worker_id, "sanity");
tonyp@3416 2826 for (int i = 0; i < (int) parallel_marking_threads(); i += 1) {
tonyp@3416 2827 CMTask* task = _tasks[i];
tonyp@3416 2828 HeapWord* task_finger = task->finger();
tonyp@3416 2829 if (task_finger != NULL && task_finger < _heap_end) {
tonyp@3416 2830 // See above note on the global finger verification.
tonyp@3416 2831 HeapRegion* task_hr = _g1h->heap_region_containing_raw(task_finger);
tonyp@3416 2832 guarantee(task_finger == task_hr->bottom() ||
tonyp@3416 2833 !task_hr->in_collection_set(),
tonyp@3416 2834 err_msg("task finger: "PTR_FORMAT" region: "HR_FORMAT,
tonyp@3416 2835 task_finger, HR_FORMAT_PARAMS(task_hr)));
tonyp@3416 2836 }
tonyp@3416 2837 }
tonyp@3416 2838 }
ysr@777 2839 }
tonyp@3416 2840 #endif // PRODUCT
ysr@777 2841
tonyp@2848 2842 void ConcurrentMark::clear_marking_state(bool clear_overflow) {
ysr@777 2843 _markStack.setEmpty();
ysr@777 2844 _markStack.clear_overflow();
tonyp@2848 2845 if (clear_overflow) {
tonyp@2848 2846 clear_has_overflown();
tonyp@2848 2847 } else {
tonyp@2848 2848 assert(has_overflown(), "pre-condition");
tonyp@2848 2849 }
ysr@777 2850 _finger = _heap_start;
ysr@777 2851
johnc@4173 2852 for (uint i = 0; i < _max_worker_id; ++i) {
ysr@777 2853 OopTaskQueue* queue = _task_queues->queue(i);
ysr@777 2854 queue->set_empty();
ysr@777 2855 }
ysr@777 2856 }
ysr@777 2857
johnc@3463 2858 // Aggregate the counting data that was constructed concurrently
johnc@3463 2859 // with marking.
johnc@3463 2860 class AggregateCountDataHRClosure: public HeapRegionClosure {
johnc@4123 2861 G1CollectedHeap* _g1h;
johnc@3463 2862 ConcurrentMark* _cm;
johnc@4123 2863 CardTableModRefBS* _ct_bs;
johnc@3463 2864 BitMap* _cm_card_bm;
johnc@4173 2865 uint _max_worker_id;
johnc@3463 2866
johnc@3463 2867 public:
johnc@4123 2868 AggregateCountDataHRClosure(G1CollectedHeap* g1h,
johnc@3463 2869 BitMap* cm_card_bm,
johnc@4173 2870 uint max_worker_id) :
johnc@4123 2871 _g1h(g1h), _cm(g1h->concurrent_mark()),
johnc@4123 2872 _ct_bs((CardTableModRefBS*) (g1h->barrier_set())),
johnc@4173 2873 _cm_card_bm(cm_card_bm), _max_worker_id(max_worker_id) { }
johnc@3463 2874
johnc@3463 2875 bool doHeapRegion(HeapRegion* hr) {
johnc@3463 2876 if (hr->continuesHumongous()) {
johnc@3463 2877 // We will ignore these here and process them when their
johnc@3463 2878 // associated "starts humongous" region is processed.
johnc@3463 2879 // Note that we cannot rely on their associated
johnc@3463 2880 // "starts humongous" region to have their bit set to 1
johnc@3463 2881 // since, due to the region chunking in the parallel region
johnc@3463 2882 // iteration, a "continues humongous" region might be visited
johnc@3463 2883 // before its associated "starts humongous".
johnc@3463 2884 return false;
johnc@3463 2885 }
johnc@3463 2886
johnc@3463 2887 HeapWord* start = hr->bottom();
johnc@3463 2888 HeapWord* limit = hr->next_top_at_mark_start();
johnc@3463 2889 HeapWord* end = hr->end();
johnc@3463 2890
johnc@3463 2891 assert(start <= limit && limit <= hr->top() && hr->top() <= hr->end(),
johnc@3463 2892 err_msg("Preconditions not met - "
johnc@3463 2893 "start: "PTR_FORMAT", limit: "PTR_FORMAT", "
johnc@3463 2894 "top: "PTR_FORMAT", end: "PTR_FORMAT,
johnc@3463 2895 start, limit, hr->top(), hr->end()));
johnc@3463 2896
johnc@3463 2897 assert(hr->next_marked_bytes() == 0, "Precondition");
johnc@3463 2898
johnc@3463 2899 if (start == limit) {
johnc@3463 2900 // NTAMS of this region has not been set so nothing to do.
johnc@3463 2901 return false;
johnc@3463 2902 }
johnc@3463 2903
johnc@4123 2904 // 'start' should be in the heap.
johnc@4123 2905 assert(_g1h->is_in_g1_reserved(start) && _ct_bs->is_card_aligned(start), "sanity");
johnc@4123 2906 // 'end' *may* be just beyone the end of the heap (if hr is the last region)
johnc@4123 2907 assert(!_g1h->is_in_g1_reserved(end) || _ct_bs->is_card_aligned(end), "sanity");
johnc@3463 2908
johnc@3463 2909 BitMap::idx_t start_idx = _cm->card_bitmap_index_for(start);
johnc@3463 2910 BitMap::idx_t limit_idx = _cm->card_bitmap_index_for(limit);
johnc@3463 2911 BitMap::idx_t end_idx = _cm->card_bitmap_index_for(end);
johnc@3463 2912
johnc@4123 2913 // If ntams is not card aligned then we bump card bitmap index
johnc@4123 2914 // for limit so that we get the all the cards spanned by
johnc@4123 2915 // the object ending at ntams.
johnc@4123 2916 // Note: if this is the last region in the heap then ntams
johnc@4123 2917 // could be actually just beyond the end of the the heap;
johnc@4123 2918 // limit_idx will then correspond to a (non-existent) card
johnc@4123 2919 // that is also outside the heap.
johnc@4123 2920 if (_g1h->is_in_g1_reserved(limit) && !_ct_bs->is_card_aligned(limit)) {
johnc@3463 2921 limit_idx += 1;
johnc@3463 2922 }
johnc@3463 2923
johnc@3463 2924 assert(limit_idx <= end_idx, "or else use atomics");
johnc@3463 2925
johnc@3463 2926 // Aggregate the "stripe" in the count data associated with hr.
tonyp@3713 2927 uint hrs_index = hr->hrs_index();
johnc@3463 2928 size_t marked_bytes = 0;
johnc@3463 2929
johnc@4173 2930 for (uint i = 0; i < _max_worker_id; i += 1) {
johnc@3463 2931 size_t* marked_bytes_array = _cm->count_marked_bytes_array_for(i);
johnc@3463 2932 BitMap* task_card_bm = _cm->count_card_bitmap_for(i);
johnc@3463 2933
johnc@3463 2934 // Fetch the marked_bytes in this region for task i and
johnc@3463 2935 // add it to the running total for this region.
johnc@3463 2936 marked_bytes += marked_bytes_array[hrs_index];
johnc@3463 2937
johnc@4173 2938 // Now union the bitmaps[0,max_worker_id)[start_idx..limit_idx)
johnc@3463 2939 // into the global card bitmap.
johnc@3463 2940 BitMap::idx_t scan_idx = task_card_bm->get_next_one_offset(start_idx, limit_idx);
johnc@3463 2941
johnc@3463 2942 while (scan_idx < limit_idx) {
johnc@3463 2943 assert(task_card_bm->at(scan_idx) == true, "should be");
johnc@3463 2944 _cm_card_bm->set_bit(scan_idx);
johnc@3463 2945 assert(_cm_card_bm->at(scan_idx) == true, "should be");
johnc@3463 2946
johnc@3463 2947 // BitMap::get_next_one_offset() can handle the case when
johnc@3463 2948 // its left_offset parameter is greater than its right_offset
johnc@4123 2949 // parameter. It does, however, have an early exit if
johnc@3463 2950 // left_offset == right_offset. So let's limit the value
johnc@3463 2951 // passed in for left offset here.
johnc@3463 2952 BitMap::idx_t next_idx = MIN2(scan_idx + 1, limit_idx);
johnc@3463 2953 scan_idx = task_card_bm->get_next_one_offset(next_idx, limit_idx);
johnc@3463 2954 }
johnc@3463 2955 }
johnc@3463 2956
johnc@3463 2957 // Update the marked bytes for this region.
johnc@3463 2958 hr->add_to_marked_bytes(marked_bytes);
johnc@3463 2959
johnc@3463 2960 // Next heap region
johnc@3463 2961 return false;
johnc@3463 2962 }
johnc@3463 2963 };
johnc@3463 2964
johnc@3463 2965 class G1AggregateCountDataTask: public AbstractGangTask {
johnc@3463 2966 protected:
johnc@3463 2967 G1CollectedHeap* _g1h;
johnc@3463 2968 ConcurrentMark* _cm;
johnc@3463 2969 BitMap* _cm_card_bm;
johnc@4173 2970 uint _max_worker_id;
johnc@3463 2971 int _active_workers;
johnc@3463 2972
johnc@3463 2973 public:
johnc@3463 2974 G1AggregateCountDataTask(G1CollectedHeap* g1h,
johnc@3463 2975 ConcurrentMark* cm,
johnc@3463 2976 BitMap* cm_card_bm,
johnc@4173 2977 uint max_worker_id,
johnc@3463 2978 int n_workers) :
johnc@3463 2979 AbstractGangTask("Count Aggregation"),
johnc@3463 2980 _g1h(g1h), _cm(cm), _cm_card_bm(cm_card_bm),
johnc@4173 2981 _max_worker_id(max_worker_id),
johnc@3463 2982 _active_workers(n_workers) { }
johnc@3463 2983
johnc@3463 2984 void work(uint worker_id) {
johnc@4173 2985 AggregateCountDataHRClosure cl(_g1h, _cm_card_bm, _max_worker_id);
johnc@3463 2986
johnc@3463 2987 if (G1CollectedHeap::use_parallel_gc_threads()) {
johnc@3463 2988 _g1h->heap_region_par_iterate_chunked(&cl, worker_id,
johnc@3463 2989 _active_workers,
johnc@3463 2990 HeapRegion::AggregateCountClaimValue);
johnc@3463 2991 } else {
johnc@3463 2992 _g1h->heap_region_iterate(&cl);
johnc@3463 2993 }
johnc@3463 2994 }
johnc@3463 2995 };
johnc@3463 2996
johnc@3463 2997
johnc@3463 2998 void ConcurrentMark::aggregate_count_data() {
johnc@3463 2999 int n_workers = (G1CollectedHeap::use_parallel_gc_threads() ?
johnc@3463 3000 _g1h->workers()->active_workers() :
johnc@3463 3001 1);
johnc@3463 3002
johnc@3463 3003 G1AggregateCountDataTask g1_par_agg_task(_g1h, this, &_card_bm,
johnc@4173 3004 _max_worker_id, n_workers);
johnc@3463 3005
johnc@3463 3006 if (G1CollectedHeap::use_parallel_gc_threads()) {
johnc@3463 3007 assert(_g1h->check_heap_region_claim_values(HeapRegion::InitialClaimValue),
johnc@3463 3008 "sanity check");
johnc@3463 3009 _g1h->set_par_threads(n_workers);
johnc@3463 3010 _g1h->workers()->run_task(&g1_par_agg_task);
johnc@3463 3011 _g1h->set_par_threads(0);
johnc@3463 3012
johnc@3463 3013 assert(_g1h->check_heap_region_claim_values(HeapRegion::AggregateCountClaimValue),
johnc@3463 3014 "sanity check");
johnc@3463 3015 _g1h->reset_heap_region_claim_values();
johnc@3463 3016 } else {
johnc@3463 3017 g1_par_agg_task.work(0);
johnc@3463 3018 }
johnc@3463 3019 }
johnc@3463 3020
johnc@3463 3021 // Clear the per-worker arrays used to store the per-region counting data
johnc@3463 3022 void ConcurrentMark::clear_all_count_data() {
johnc@3463 3023 // Clear the global card bitmap - it will be filled during
johnc@3463 3024 // liveness count aggregation (during remark) and the
johnc@3463 3025 // final counting task.
johnc@3463 3026 _card_bm.clear();
johnc@3463 3027
johnc@3463 3028 // Clear the global region bitmap - it will be filled as part
johnc@3463 3029 // of the final counting task.
johnc@3463 3030 _region_bm.clear();
johnc@3463 3031
tonyp@3713 3032 uint max_regions = _g1h->max_regions();
johnc@4173 3033 assert(_max_worker_id > 0, "uninitialized");
johnc@4173 3034
johnc@4173 3035 for (uint i = 0; i < _max_worker_id; i += 1) {
johnc@3463 3036 BitMap* task_card_bm = count_card_bitmap_for(i);
johnc@3463 3037 size_t* marked_bytes_array = count_marked_bytes_array_for(i);
johnc@3463 3038
johnc@3463 3039 assert(task_card_bm->size() == _card_bm.size(), "size mismatch");
johnc@3463 3040 assert(marked_bytes_array != NULL, "uninitialized");
johnc@3463 3041
tonyp@3713 3042 memset(marked_bytes_array, 0, (size_t) max_regions * sizeof(size_t));
johnc@3463 3043 task_card_bm->clear();
johnc@3463 3044 }
johnc@3463 3045 }
johnc@3463 3046
ysr@777 3047 void ConcurrentMark::print_stats() {
ysr@777 3048 if (verbose_stats()) {
ysr@777 3049 gclog_or_tty->print_cr("---------------------------------------------------------------------");
ysr@777 3050 for (size_t i = 0; i < _active_tasks; ++i) {
ysr@777 3051 _tasks[i]->print_stats();
ysr@777 3052 gclog_or_tty->print_cr("---------------------------------------------------------------------");
ysr@777 3053 }
ysr@777 3054 }
ysr@777 3055 }
ysr@777 3056
ysr@777 3057 // abandon current marking iteration due to a Full GC
ysr@777 3058 void ConcurrentMark::abort() {
ysr@777 3059 // Clear all marks to force marking thread to do nothing
ysr@777 3060 _nextMarkBitMap->clearAll();
johnc@3463 3061 // Clear the liveness counting data
johnc@3463 3062 clear_all_count_data();
ysr@777 3063 // Empty mark stack
ysr@777 3064 clear_marking_state();
johnc@4173 3065 for (uint i = 0; i < _max_worker_id; ++i) {
ysr@777 3066 _tasks[i]->clear_region_fields();
johnc@2190 3067 }
ysr@777 3068 _has_aborted = true;
ysr@777 3069
ysr@777 3070 SATBMarkQueueSet& satb_mq_set = JavaThread::satb_mark_queue_set();
ysr@777 3071 satb_mq_set.abandon_partial_marking();
tonyp@1752 3072 // This can be called either during or outside marking, we'll read
tonyp@1752 3073 // the expected_active value from the SATB queue set.
tonyp@1752 3074 satb_mq_set.set_active_all_threads(
tonyp@1752 3075 false, /* new active value */
tonyp@1752 3076 satb_mq_set.is_active() /* expected_active */);
ysr@777 3077 }
ysr@777 3078
ysr@777 3079 static void print_ms_time_info(const char* prefix, const char* name,
ysr@777 3080 NumberSeq& ns) {
ysr@777 3081 gclog_or_tty->print_cr("%s%5d %12s: total time = %8.2f s (avg = %8.2f ms).",
ysr@777 3082 prefix, ns.num(), name, ns.sum()/1000.0, ns.avg());
ysr@777 3083 if (ns.num() > 0) {
ysr@777 3084 gclog_or_tty->print_cr("%s [std. dev = %8.2f ms, max = %8.2f ms]",
ysr@777 3085 prefix, ns.sd(), ns.maximum());
ysr@777 3086 }
ysr@777 3087 }
ysr@777 3088
ysr@777 3089 void ConcurrentMark::print_summary_info() {
ysr@777 3090 gclog_or_tty->print_cr(" Concurrent marking:");
ysr@777 3091 print_ms_time_info(" ", "init marks", _init_times);
ysr@777 3092 print_ms_time_info(" ", "remarks", _remark_times);
ysr@777 3093 {
ysr@777 3094 print_ms_time_info(" ", "final marks", _remark_mark_times);
ysr@777 3095 print_ms_time_info(" ", "weak refs", _remark_weak_ref_times);
ysr@777 3096
ysr@777 3097 }
ysr@777 3098 print_ms_time_info(" ", "cleanups", _cleanup_times);
ysr@777 3099 gclog_or_tty->print_cr(" Final counting total time = %8.2f s (avg = %8.2f ms).",
ysr@777 3100 _total_counting_time,
ysr@777 3101 (_cleanup_times.num() > 0 ? _total_counting_time * 1000.0 /
ysr@777 3102 (double)_cleanup_times.num()
ysr@777 3103 : 0.0));
ysr@777 3104 if (G1ScrubRemSets) {
ysr@777 3105 gclog_or_tty->print_cr(" RS scrub total time = %8.2f s (avg = %8.2f ms).",
ysr@777 3106 _total_rs_scrub_time,
ysr@777 3107 (_cleanup_times.num() > 0 ? _total_rs_scrub_time * 1000.0 /
ysr@777 3108 (double)_cleanup_times.num()
ysr@777 3109 : 0.0));
ysr@777 3110 }
ysr@777 3111 gclog_or_tty->print_cr(" Total stop_world time = %8.2f s.",
ysr@777 3112 (_init_times.sum() + _remark_times.sum() +
ysr@777 3113 _cleanup_times.sum())/1000.0);
ysr@777 3114 gclog_or_tty->print_cr(" Total concurrent time = %8.2f s "
johnc@3463 3115 "(%8.2f s marking).",
ysr@777 3116 cmThread()->vtime_accum(),
johnc@3463 3117 cmThread()->vtime_mark_accum());
ysr@777 3118 }
ysr@777 3119
tonyp@1454 3120 void ConcurrentMark::print_worker_threads_on(outputStream* st) const {
tonyp@1454 3121 _parallel_workers->print_worker_threads_on(st);
tonyp@1454 3122 }
tonyp@1454 3123
ysr@777 3124 // We take a break if someone is trying to stop the world.
jmasa@3357 3125 bool ConcurrentMark::do_yield_check(uint worker_id) {
ysr@777 3126 if (should_yield()) {
jmasa@3357 3127 if (worker_id == 0) {
ysr@777 3128 _g1h->g1_policy()->record_concurrent_pause();
tonyp@2973 3129 }
ysr@777 3130 cmThread()->yield();
ysr@777 3131 return true;
ysr@777 3132 } else {
ysr@777 3133 return false;
ysr@777 3134 }
ysr@777 3135 }
ysr@777 3136
ysr@777 3137 bool ConcurrentMark::should_yield() {
ysr@777 3138 return cmThread()->should_yield();
ysr@777 3139 }
ysr@777 3140
ysr@777 3141 bool ConcurrentMark::containing_card_is_marked(void* p) {
ysr@777 3142 size_t offset = pointer_delta(p, _g1h->reserved_region().start(), 1);
ysr@777 3143 return _card_bm.at(offset >> CardTableModRefBS::card_shift);
ysr@777 3144 }
ysr@777 3145
ysr@777 3146 bool ConcurrentMark::containing_cards_are_marked(void* start,
ysr@777 3147 void* last) {
tonyp@2973 3148 return containing_card_is_marked(start) &&
tonyp@2973 3149 containing_card_is_marked(last);
ysr@777 3150 }
ysr@777 3151
ysr@777 3152 #ifndef PRODUCT
ysr@777 3153 // for debugging purposes
ysr@777 3154 void ConcurrentMark::print_finger() {
ysr@777 3155 gclog_or_tty->print_cr("heap ["PTR_FORMAT", "PTR_FORMAT"), global finger = "PTR_FORMAT,
ysr@777 3156 _heap_start, _heap_end, _finger);
johnc@4173 3157 for (uint i = 0; i < _max_worker_id; ++i) {
johnc@4173 3158 gclog_or_tty->print(" %u: "PTR_FORMAT, i, _tasks[i]->finger());
ysr@777 3159 }
ysr@777 3160 gclog_or_tty->print_cr("");
ysr@777 3161 }
ysr@777 3162 #endif
ysr@777 3163
tonyp@2968 3164 void CMTask::scan_object(oop obj) {
tonyp@2968 3165 assert(_nextMarkBitMap->isMarked((HeapWord*) obj), "invariant");
tonyp@2968 3166
tonyp@2968 3167 if (_cm->verbose_high()) {
johnc@4173 3168 gclog_or_tty->print_cr("[%u] we're scanning object "PTR_FORMAT,
johnc@4173 3169 _worker_id, (void*) obj);
tonyp@2968 3170 }
tonyp@2968 3171
tonyp@2968 3172 size_t obj_size = obj->size();
tonyp@2968 3173 _words_scanned += obj_size;
tonyp@2968 3174
tonyp@2968 3175 obj->oop_iterate(_cm_oop_closure);
tonyp@2968 3176 statsOnly( ++_objs_scanned );
tonyp@2968 3177 check_limits();
tonyp@2968 3178 }
tonyp@2968 3179
ysr@777 3180 // Closure for iteration over bitmaps
ysr@777 3181 class CMBitMapClosure : public BitMapClosure {
ysr@777 3182 private:
ysr@777 3183 // the bitmap that is being iterated over
ysr@777 3184 CMBitMap* _nextMarkBitMap;
ysr@777 3185 ConcurrentMark* _cm;
ysr@777 3186 CMTask* _task;
ysr@777 3187
ysr@777 3188 public:
tonyp@3691 3189 CMBitMapClosure(CMTask *task, ConcurrentMark* cm, CMBitMap* nextMarkBitMap) :
tonyp@3691 3190 _task(task), _cm(cm), _nextMarkBitMap(nextMarkBitMap) { }
ysr@777 3191
ysr@777 3192 bool do_bit(size_t offset) {
ysr@777 3193 HeapWord* addr = _nextMarkBitMap->offsetToHeapWord(offset);
tonyp@1458 3194 assert(_nextMarkBitMap->isMarked(addr), "invariant");
tonyp@1458 3195 assert( addr < _cm->finger(), "invariant");
ysr@777 3196
tonyp@3691 3197 statsOnly( _task->increase_objs_found_on_bitmap() );
tonyp@3691 3198 assert(addr >= _task->finger(), "invariant");
tonyp@3691 3199
tonyp@3691 3200 // We move that task's local finger along.
tonyp@3691 3201 _task->move_finger_to(addr);
ysr@777 3202
ysr@777 3203 _task->scan_object(oop(addr));
ysr@777 3204 // we only partially drain the local queue and global stack
ysr@777 3205 _task->drain_local_queue(true);
ysr@777 3206 _task->drain_global_stack(true);
ysr@777 3207
ysr@777 3208 // if the has_aborted flag has been raised, we need to bail out of
ysr@777 3209 // the iteration
ysr@777 3210 return !_task->has_aborted();
ysr@777 3211 }
ysr@777 3212 };
ysr@777 3213
ysr@777 3214 // Closure for iterating over objects, currently only used for
ysr@777 3215 // processing SATB buffers.
ysr@777 3216 class CMObjectClosure : public ObjectClosure {
ysr@777 3217 private:
ysr@777 3218 CMTask* _task;
ysr@777 3219
ysr@777 3220 public:
ysr@777 3221 void do_object(oop obj) {
ysr@777 3222 _task->deal_with_reference(obj);
ysr@777 3223 }
ysr@777 3224
ysr@777 3225 CMObjectClosure(CMTask* task) : _task(task) { }
ysr@777 3226 };
ysr@777 3227
tonyp@2968 3228 G1CMOopClosure::G1CMOopClosure(G1CollectedHeap* g1h,
tonyp@2968 3229 ConcurrentMark* cm,
tonyp@2968 3230 CMTask* task)
tonyp@2968 3231 : _g1h(g1h), _cm(cm), _task(task) {
tonyp@2968 3232 assert(_ref_processor == NULL, "should be initialized to NULL");
tonyp@2968 3233
tonyp@2968 3234 if (G1UseConcMarkReferenceProcessing) {
johnc@3175 3235 _ref_processor = g1h->ref_processor_cm();
tonyp@2968 3236 assert(_ref_processor != NULL, "should not be NULL");
ysr@777 3237 }
tonyp@2968 3238 }
ysr@777 3239
ysr@777 3240 void CMTask::setup_for_region(HeapRegion* hr) {
tonyp@1458 3241 // Separated the asserts so that we know which one fires.
tonyp@1458 3242 assert(hr != NULL,
tonyp@1458 3243 "claim_region() should have filtered out continues humongous regions");
tonyp@1458 3244 assert(!hr->continuesHumongous(),
tonyp@1458 3245 "claim_region() should have filtered out continues humongous regions");
ysr@777 3246
tonyp@2973 3247 if (_cm->verbose_low()) {
johnc@4173 3248 gclog_or_tty->print_cr("[%u] setting up for region "PTR_FORMAT,
johnc@4173 3249 _worker_id, hr);
tonyp@2973 3250 }
ysr@777 3251
ysr@777 3252 _curr_region = hr;
ysr@777 3253 _finger = hr->bottom();
ysr@777 3254 update_region_limit();
ysr@777 3255 }
ysr@777 3256
ysr@777 3257 void CMTask::update_region_limit() {
ysr@777 3258 HeapRegion* hr = _curr_region;
ysr@777 3259 HeapWord* bottom = hr->bottom();
ysr@777 3260 HeapWord* limit = hr->next_top_at_mark_start();
ysr@777 3261
ysr@777 3262 if (limit == bottom) {
tonyp@2973 3263 if (_cm->verbose_low()) {
johnc@4173 3264 gclog_or_tty->print_cr("[%u] found an empty region "
ysr@777 3265 "["PTR_FORMAT", "PTR_FORMAT")",
johnc@4173 3266 _worker_id, bottom, limit);
tonyp@2973 3267 }
ysr@777 3268 // The region was collected underneath our feet.
ysr@777 3269 // We set the finger to bottom to ensure that the bitmap
ysr@777 3270 // iteration that will follow this will not do anything.
ysr@777 3271 // (this is not a condition that holds when we set the region up,
ysr@777 3272 // as the region is not supposed to be empty in the first place)
ysr@777 3273 _finger = bottom;
ysr@777 3274 } else if (limit >= _region_limit) {
tonyp@1458 3275 assert(limit >= _finger, "peace of mind");
ysr@777 3276 } else {
tonyp@1458 3277 assert(limit < _region_limit, "only way to get here");
ysr@777 3278 // This can happen under some pretty unusual circumstances. An
ysr@777 3279 // evacuation pause empties the region underneath our feet (NTAMS
ysr@777 3280 // at bottom). We then do some allocation in the region (NTAMS
ysr@777 3281 // stays at bottom), followed by the region being used as a GC
ysr@777 3282 // alloc region (NTAMS will move to top() and the objects
ysr@777 3283 // originally below it will be grayed). All objects now marked in
ysr@777 3284 // the region are explicitly grayed, if below the global finger,
ysr@777 3285 // and we do not need in fact to scan anything else. So, we simply
ysr@777 3286 // set _finger to be limit to ensure that the bitmap iteration
ysr@777 3287 // doesn't do anything.
ysr@777 3288 _finger = limit;
ysr@777 3289 }
ysr@777 3290
ysr@777 3291 _region_limit = limit;
ysr@777 3292 }
ysr@777 3293
ysr@777 3294 void CMTask::giveup_current_region() {
tonyp@1458 3295 assert(_curr_region != NULL, "invariant");
tonyp@2973 3296 if (_cm->verbose_low()) {
johnc@4173 3297 gclog_or_tty->print_cr("[%u] giving up region "PTR_FORMAT,
johnc@4173 3298 _worker_id, _curr_region);
tonyp@2973 3299 }
ysr@777 3300 clear_region_fields();
ysr@777 3301 }
ysr@777 3302
ysr@777 3303 void CMTask::clear_region_fields() {
ysr@777 3304 // Values for these three fields that indicate that we're not
ysr@777 3305 // holding on to a region.
ysr@777 3306 _curr_region = NULL;
ysr@777 3307 _finger = NULL;
ysr@777 3308 _region_limit = NULL;
ysr@777 3309 }
ysr@777 3310
tonyp@2968 3311 void CMTask::set_cm_oop_closure(G1CMOopClosure* cm_oop_closure) {
tonyp@2968 3312 if (cm_oop_closure == NULL) {
tonyp@2968 3313 assert(_cm_oop_closure != NULL, "invariant");
tonyp@2968 3314 } else {
tonyp@2968 3315 assert(_cm_oop_closure == NULL, "invariant");
tonyp@2968 3316 }
tonyp@2968 3317 _cm_oop_closure = cm_oop_closure;
tonyp@2968 3318 }
tonyp@2968 3319
ysr@777 3320 void CMTask::reset(CMBitMap* nextMarkBitMap) {
tonyp@1458 3321 guarantee(nextMarkBitMap != NULL, "invariant");
ysr@777 3322
tonyp@2973 3323 if (_cm->verbose_low()) {
johnc@4173 3324 gclog_or_tty->print_cr("[%u] resetting", _worker_id);
tonyp@2973 3325 }
ysr@777 3326
ysr@777 3327 _nextMarkBitMap = nextMarkBitMap;
ysr@777 3328 clear_region_fields();
ysr@777 3329
ysr@777 3330 _calls = 0;
ysr@777 3331 _elapsed_time_ms = 0.0;
ysr@777 3332 _termination_time_ms = 0.0;
ysr@777 3333 _termination_start_time_ms = 0.0;
ysr@777 3334
ysr@777 3335 #if _MARKING_STATS_
ysr@777 3336 _local_pushes = 0;
ysr@777 3337 _local_pops = 0;
ysr@777 3338 _local_max_size = 0;
ysr@777 3339 _objs_scanned = 0;
ysr@777 3340 _global_pushes = 0;
ysr@777 3341 _global_pops = 0;
ysr@777 3342 _global_max_size = 0;
ysr@777 3343 _global_transfers_to = 0;
ysr@777 3344 _global_transfers_from = 0;
ysr@777 3345 _regions_claimed = 0;
ysr@777 3346 _objs_found_on_bitmap = 0;
ysr@777 3347 _satb_buffers_processed = 0;
ysr@777 3348 _steal_attempts = 0;
ysr@777 3349 _steals = 0;
ysr@777 3350 _aborted = 0;
ysr@777 3351 _aborted_overflow = 0;
ysr@777 3352 _aborted_cm_aborted = 0;
ysr@777 3353 _aborted_yield = 0;
ysr@777 3354 _aborted_timed_out = 0;
ysr@777 3355 _aborted_satb = 0;
ysr@777 3356 _aborted_termination = 0;
ysr@777 3357 #endif // _MARKING_STATS_
ysr@777 3358 }
ysr@777 3359
ysr@777 3360 bool CMTask::should_exit_termination() {
ysr@777 3361 regular_clock_call();
ysr@777 3362 // This is called when we are in the termination protocol. We should
ysr@777 3363 // quit if, for some reason, this task wants to abort or the global
ysr@777 3364 // stack is not empty (this means that we can get work from it).
ysr@777 3365 return !_cm->mark_stack_empty() || has_aborted();
ysr@777 3366 }
ysr@777 3367
ysr@777 3368 void CMTask::reached_limit() {
tonyp@1458 3369 assert(_words_scanned >= _words_scanned_limit ||
tonyp@1458 3370 _refs_reached >= _refs_reached_limit ,
tonyp@1458 3371 "shouldn't have been called otherwise");
ysr@777 3372 regular_clock_call();
ysr@777 3373 }
ysr@777 3374
ysr@777 3375 void CMTask::regular_clock_call() {
tonyp@2973 3376 if (has_aborted()) return;
ysr@777 3377
ysr@777 3378 // First, we need to recalculate the words scanned and refs reached
ysr@777 3379 // limits for the next clock call.
ysr@777 3380 recalculate_limits();
ysr@777 3381
ysr@777 3382 // During the regular clock call we do the following
ysr@777 3383
ysr@777 3384 // (1) If an overflow has been flagged, then we abort.
ysr@777 3385 if (_cm->has_overflown()) {
ysr@777 3386 set_has_aborted();
ysr@777 3387 return;
ysr@777 3388 }
ysr@777 3389
ysr@777 3390 // If we are not concurrent (i.e. we're doing remark) we don't need
ysr@777 3391 // to check anything else. The other steps are only needed during
ysr@777 3392 // the concurrent marking phase.
tonyp@2973 3393 if (!concurrent()) return;
ysr@777 3394
ysr@777 3395 // (2) If marking has been aborted for Full GC, then we also abort.
ysr@777 3396 if (_cm->has_aborted()) {
ysr@777 3397 set_has_aborted();
ysr@777 3398 statsOnly( ++_aborted_cm_aborted );
ysr@777 3399 return;
ysr@777 3400 }
ysr@777 3401
ysr@777 3402 double curr_time_ms = os::elapsedVTime() * 1000.0;
ysr@777 3403
ysr@777 3404 // (3) If marking stats are enabled, then we update the step history.
ysr@777 3405 #if _MARKING_STATS_
tonyp@2973 3406 if (_words_scanned >= _words_scanned_limit) {
ysr@777 3407 ++_clock_due_to_scanning;
tonyp@2973 3408 }
tonyp@2973 3409 if (_refs_reached >= _refs_reached_limit) {
ysr@777 3410 ++_clock_due_to_marking;
tonyp@2973 3411 }
ysr@777 3412
ysr@777 3413 double last_interval_ms = curr_time_ms - _interval_start_time_ms;
ysr@777 3414 _interval_start_time_ms = curr_time_ms;
ysr@777 3415 _all_clock_intervals_ms.add(last_interval_ms);
ysr@777 3416
ysr@777 3417 if (_cm->verbose_medium()) {
johnc@4173 3418 gclog_or_tty->print_cr("[%u] regular clock, interval = %1.2lfms, "
tonyp@2973 3419 "scanned = %d%s, refs reached = %d%s",
johnc@4173 3420 _worker_id, last_interval_ms,
tonyp@2973 3421 _words_scanned,
tonyp@2973 3422 (_words_scanned >= _words_scanned_limit) ? " (*)" : "",
tonyp@2973 3423 _refs_reached,
tonyp@2973 3424 (_refs_reached >= _refs_reached_limit) ? " (*)" : "");
ysr@777 3425 }
ysr@777 3426 #endif // _MARKING_STATS_
ysr@777 3427
ysr@777 3428 // (4) We check whether we should yield. If we have to, then we abort.
ysr@777 3429 if (_cm->should_yield()) {
ysr@777 3430 // We should yield. To do this we abort the task. The caller is
ysr@777 3431 // responsible for yielding.
ysr@777 3432 set_has_aborted();
ysr@777 3433 statsOnly( ++_aborted_yield );
ysr@777 3434 return;
ysr@777 3435 }
ysr@777 3436
ysr@777 3437 // (5) We check whether we've reached our time quota. If we have,
ysr@777 3438 // then we abort.
ysr@777 3439 double elapsed_time_ms = curr_time_ms - _start_time_ms;
ysr@777 3440 if (elapsed_time_ms > _time_target_ms) {
ysr@777 3441 set_has_aborted();
johnc@2494 3442 _has_timed_out = true;
ysr@777 3443 statsOnly( ++_aborted_timed_out );
ysr@777 3444 return;
ysr@777 3445 }
ysr@777 3446
ysr@777 3447 // (6) Finally, we check whether there are enough completed STAB
ysr@777 3448 // buffers available for processing. If there are, we abort.
ysr@777 3449 SATBMarkQueueSet& satb_mq_set = JavaThread::satb_mark_queue_set();
ysr@777 3450 if (!_draining_satb_buffers && satb_mq_set.process_completed_buffers()) {
tonyp@2973 3451 if (_cm->verbose_low()) {
johnc@4173 3452 gclog_or_tty->print_cr("[%u] aborting to deal with pending SATB buffers",
johnc@4173 3453 _worker_id);
tonyp@2973 3454 }
ysr@777 3455 // we do need to process SATB buffers, we'll abort and restart
ysr@777 3456 // the marking task to do so
ysr@777 3457 set_has_aborted();
ysr@777 3458 statsOnly( ++_aborted_satb );
ysr@777 3459 return;
ysr@777 3460 }
ysr@777 3461 }
ysr@777 3462
ysr@777 3463 void CMTask::recalculate_limits() {
ysr@777 3464 _real_words_scanned_limit = _words_scanned + words_scanned_period;
ysr@777 3465 _words_scanned_limit = _real_words_scanned_limit;
ysr@777 3466
ysr@777 3467 _real_refs_reached_limit = _refs_reached + refs_reached_period;
ysr@777 3468 _refs_reached_limit = _real_refs_reached_limit;
ysr@777 3469 }
ysr@777 3470
ysr@777 3471 void CMTask::decrease_limits() {
ysr@777 3472 // This is called when we believe that we're going to do an infrequent
ysr@777 3473 // operation which will increase the per byte scanned cost (i.e. move
ysr@777 3474 // entries to/from the global stack). It basically tries to decrease the
ysr@777 3475 // scanning limit so that the clock is called earlier.
ysr@777 3476
tonyp@2973 3477 if (_cm->verbose_medium()) {
johnc@4173 3478 gclog_or_tty->print_cr("[%u] decreasing limits", _worker_id);
tonyp@2973 3479 }
ysr@777 3480
ysr@777 3481 _words_scanned_limit = _real_words_scanned_limit -
ysr@777 3482 3 * words_scanned_period / 4;
ysr@777 3483 _refs_reached_limit = _real_refs_reached_limit -
ysr@777 3484 3 * refs_reached_period / 4;
ysr@777 3485 }
ysr@777 3486
ysr@777 3487 void CMTask::move_entries_to_global_stack() {
ysr@777 3488 // local array where we'll store the entries that will be popped
ysr@777 3489 // from the local queue
ysr@777 3490 oop buffer[global_stack_transfer_size];
ysr@777 3491
ysr@777 3492 int n = 0;
ysr@777 3493 oop obj;
ysr@777 3494 while (n < global_stack_transfer_size && _task_queue->pop_local(obj)) {
ysr@777 3495 buffer[n] = obj;
ysr@777 3496 ++n;
ysr@777 3497 }
ysr@777 3498
ysr@777 3499 if (n > 0) {
ysr@777 3500 // we popped at least one entry from the local queue
ysr@777 3501
ysr@777 3502 statsOnly( ++_global_transfers_to; _local_pops += n );
ysr@777 3503
ysr@777 3504 if (!_cm->mark_stack_push(buffer, n)) {
tonyp@2973 3505 if (_cm->verbose_low()) {
johnc@4173 3506 gclog_or_tty->print_cr("[%u] aborting due to global stack overflow",
johnc@4173 3507 _worker_id);
tonyp@2973 3508 }
ysr@777 3509 set_has_aborted();
ysr@777 3510 } else {
ysr@777 3511 // the transfer was successful
ysr@777 3512
tonyp@2973 3513 if (_cm->verbose_medium()) {
johnc@4173 3514 gclog_or_tty->print_cr("[%u] pushed %d entries to the global stack",
johnc@4173 3515 _worker_id, n);
tonyp@2973 3516 }
ysr@777 3517 statsOnly( int tmp_size = _cm->mark_stack_size();
tonyp@2973 3518 if (tmp_size > _global_max_size) {
ysr@777 3519 _global_max_size = tmp_size;
tonyp@2973 3520 }
ysr@777 3521 _global_pushes += n );
ysr@777 3522 }
ysr@777 3523 }
ysr@777 3524
ysr@777 3525 // this operation was quite expensive, so decrease the limits
ysr@777 3526 decrease_limits();
ysr@777 3527 }
ysr@777 3528
ysr@777 3529 void CMTask::get_entries_from_global_stack() {
ysr@777 3530 // local array where we'll store the entries that will be popped
ysr@777 3531 // from the global stack.
ysr@777 3532 oop buffer[global_stack_transfer_size];
ysr@777 3533 int n;
ysr@777 3534 _cm->mark_stack_pop(buffer, global_stack_transfer_size, &n);
tonyp@1458 3535 assert(n <= global_stack_transfer_size,
tonyp@1458 3536 "we should not pop more than the given limit");
ysr@777 3537 if (n > 0) {
ysr@777 3538 // yes, we did actually pop at least one entry
ysr@777 3539
ysr@777 3540 statsOnly( ++_global_transfers_from; _global_pops += n );
tonyp@2973 3541 if (_cm->verbose_medium()) {
johnc@4173 3542 gclog_or_tty->print_cr("[%u] popped %d entries from the global stack",
johnc@4173 3543 _worker_id, n);
tonyp@2973 3544 }
ysr@777 3545 for (int i = 0; i < n; ++i) {
ysr@777 3546 bool success = _task_queue->push(buffer[i]);
ysr@777 3547 // We only call this when the local queue is empty or under a
ysr@777 3548 // given target limit. So, we do not expect this push to fail.
tonyp@1458 3549 assert(success, "invariant");
ysr@777 3550 }
ysr@777 3551
ysr@777 3552 statsOnly( int tmp_size = _task_queue->size();
tonyp@2973 3553 if (tmp_size > _local_max_size) {
ysr@777 3554 _local_max_size = tmp_size;
tonyp@2973 3555 }
ysr@777 3556 _local_pushes += n );
ysr@777 3557 }
ysr@777 3558
ysr@777 3559 // this operation was quite expensive, so decrease the limits
ysr@777 3560 decrease_limits();
ysr@777 3561 }
ysr@777 3562
ysr@777 3563 void CMTask::drain_local_queue(bool partially) {
tonyp@2973 3564 if (has_aborted()) return;
ysr@777 3565
ysr@777 3566 // Decide what the target size is, depending whether we're going to
ysr@777 3567 // drain it partially (so that other tasks can steal if they run out
ysr@777 3568 // of things to do) or totally (at the very end).
ysr@777 3569 size_t target_size;
tonyp@2973 3570 if (partially) {
ysr@777 3571 target_size = MIN2((size_t)_task_queue->max_elems()/3, GCDrainStackTargetSize);
tonyp@2973 3572 } else {
ysr@777 3573 target_size = 0;
tonyp@2973 3574 }
ysr@777 3575
ysr@777 3576 if (_task_queue->size() > target_size) {
tonyp@2973 3577 if (_cm->verbose_high()) {
johnc@4173 3578 gclog_or_tty->print_cr("[%u] draining local queue, target size = %d",
johnc@4173 3579 _worker_id, target_size);
tonyp@2973 3580 }
ysr@777 3581
ysr@777 3582 oop obj;
ysr@777 3583 bool ret = _task_queue->pop_local(obj);
ysr@777 3584 while (ret) {
ysr@777 3585 statsOnly( ++_local_pops );
ysr@777 3586
tonyp@2973 3587 if (_cm->verbose_high()) {
johnc@4173 3588 gclog_or_tty->print_cr("[%u] popped "PTR_FORMAT, _worker_id,
ysr@777 3589 (void*) obj);
tonyp@2973 3590 }
ysr@777 3591
tonyp@1458 3592 assert(_g1h->is_in_g1_reserved((HeapWord*) obj), "invariant" );
tonyp@2643 3593 assert(!_g1h->is_on_master_free_list(
tonyp@2472 3594 _g1h->heap_region_containing((HeapWord*) obj)), "invariant");
ysr@777 3595
ysr@777 3596 scan_object(obj);
ysr@777 3597
tonyp@2973 3598 if (_task_queue->size() <= target_size || has_aborted()) {
ysr@777 3599 ret = false;
tonyp@2973 3600 } else {
ysr@777 3601 ret = _task_queue->pop_local(obj);
tonyp@2973 3602 }
ysr@777 3603 }
ysr@777 3604
tonyp@2973 3605 if (_cm->verbose_high()) {
johnc@4173 3606 gclog_or_tty->print_cr("[%u] drained local queue, size = %d",
johnc@4173 3607 _worker_id, _task_queue->size());
tonyp@2973 3608 }
ysr@777 3609 }
ysr@777 3610 }
ysr@777 3611
ysr@777 3612 void CMTask::drain_global_stack(bool partially) {
tonyp@2973 3613 if (has_aborted()) return;
ysr@777 3614
ysr@777 3615 // We have a policy to drain the local queue before we attempt to
ysr@777 3616 // drain the global stack.
tonyp@1458 3617 assert(partially || _task_queue->size() == 0, "invariant");
ysr@777 3618
ysr@777 3619 // Decide what the target size is, depending whether we're going to
ysr@777 3620 // drain it partially (so that other tasks can steal if they run out
ysr@777 3621 // of things to do) or totally (at the very end). Notice that,
ysr@777 3622 // because we move entries from the global stack in chunks or
ysr@777 3623 // because another task might be doing the same, we might in fact
ysr@777 3624 // drop below the target. But, this is not a problem.
ysr@777 3625 size_t target_size;
tonyp@2973 3626 if (partially) {
ysr@777 3627 target_size = _cm->partial_mark_stack_size_target();
tonyp@2973 3628 } else {
ysr@777 3629 target_size = 0;
tonyp@2973 3630 }
ysr@777 3631
ysr@777 3632 if (_cm->mark_stack_size() > target_size) {
tonyp@2973 3633 if (_cm->verbose_low()) {
johnc@4173 3634 gclog_or_tty->print_cr("[%u] draining global_stack, target size %d",
johnc@4173 3635 _worker_id, target_size);
tonyp@2973 3636 }
ysr@777 3637
ysr@777 3638 while (!has_aborted() && _cm->mark_stack_size() > target_size) {
ysr@777 3639 get_entries_from_global_stack();
ysr@777 3640 drain_local_queue(partially);
ysr@777 3641 }
ysr@777 3642
tonyp@2973 3643 if (_cm->verbose_low()) {
johnc@4173 3644 gclog_or_tty->print_cr("[%u] drained global stack, size = %d",
johnc@4173 3645 _worker_id, _cm->mark_stack_size());
tonyp@2973 3646 }
ysr@777 3647 }
ysr@777 3648 }
ysr@777 3649
ysr@777 3650 // SATB Queue has several assumptions on whether to call the par or
ysr@777 3651 // non-par versions of the methods. this is why some of the code is
ysr@777 3652 // replicated. We should really get rid of the single-threaded version
ysr@777 3653 // of the code to simplify things.
ysr@777 3654 void CMTask::drain_satb_buffers() {
tonyp@2973 3655 if (has_aborted()) return;
ysr@777 3656
ysr@777 3657 // We set this so that the regular clock knows that we're in the
ysr@777 3658 // middle of draining buffers and doesn't set the abort flag when it
ysr@777 3659 // notices that SATB buffers are available for draining. It'd be
ysr@777 3660 // very counter productive if it did that. :-)
ysr@777 3661 _draining_satb_buffers = true;
ysr@777 3662
ysr@777 3663 CMObjectClosure oc(this);
ysr@777 3664 SATBMarkQueueSet& satb_mq_set = JavaThread::satb_mark_queue_set();
tonyp@2973 3665 if (G1CollectedHeap::use_parallel_gc_threads()) {
johnc@4173 3666 satb_mq_set.set_par_closure(_worker_id, &oc);
tonyp@2973 3667 } else {
ysr@777 3668 satb_mq_set.set_closure(&oc);
tonyp@2973 3669 }
ysr@777 3670
ysr@777 3671 // This keeps claiming and applying the closure to completed buffers
ysr@777 3672 // until we run out of buffers or we need to abort.
jmasa@2188 3673 if (G1CollectedHeap::use_parallel_gc_threads()) {
ysr@777 3674 while (!has_aborted() &&
johnc@4173 3675 satb_mq_set.par_apply_closure_to_completed_buffer(_worker_id)) {
tonyp@2973 3676 if (_cm->verbose_medium()) {
johnc@4173 3677 gclog_or_tty->print_cr("[%u] processed an SATB buffer", _worker_id);
tonyp@2973 3678 }
ysr@777 3679 statsOnly( ++_satb_buffers_processed );
ysr@777 3680 regular_clock_call();
ysr@777 3681 }
ysr@777 3682 } else {
ysr@777 3683 while (!has_aborted() &&
ysr@777 3684 satb_mq_set.apply_closure_to_completed_buffer()) {
tonyp@2973 3685 if (_cm->verbose_medium()) {
johnc@4173 3686 gclog_or_tty->print_cr("[%u] processed an SATB buffer", _worker_id);
tonyp@2973 3687 }
ysr@777 3688 statsOnly( ++_satb_buffers_processed );
ysr@777 3689 regular_clock_call();
ysr@777 3690 }
ysr@777 3691 }
ysr@777 3692
ysr@777 3693 if (!concurrent() && !has_aborted()) {
ysr@777 3694 // We should only do this during remark.
tonyp@2973 3695 if (G1CollectedHeap::use_parallel_gc_threads()) {
johnc@4173 3696 satb_mq_set.par_iterate_closure_all_threads(_worker_id);
tonyp@2973 3697 } else {
ysr@777 3698 satb_mq_set.iterate_closure_all_threads();
tonyp@2973 3699 }
ysr@777 3700 }
ysr@777 3701
ysr@777 3702 _draining_satb_buffers = false;
ysr@777 3703
tonyp@1458 3704 assert(has_aborted() ||
tonyp@1458 3705 concurrent() ||
tonyp@1458 3706 satb_mq_set.completed_buffers_num() == 0, "invariant");
ysr@777 3707
tonyp@2973 3708 if (G1CollectedHeap::use_parallel_gc_threads()) {
johnc@4173 3709 satb_mq_set.set_par_closure(_worker_id, NULL);
tonyp@2973 3710 } else {
ysr@777 3711 satb_mq_set.set_closure(NULL);
tonyp@2973 3712 }
ysr@777 3713
ysr@777 3714 // again, this was a potentially expensive operation, decrease the
ysr@777 3715 // limits to get the regular clock call early
ysr@777 3716 decrease_limits();
ysr@777 3717 }
ysr@777 3718
ysr@777 3719 void CMTask::print_stats() {
johnc@4173 3720 gclog_or_tty->print_cr("Marking Stats, task = %u, calls = %d",
johnc@4173 3721 _worker_id, _calls);
ysr@777 3722 gclog_or_tty->print_cr(" Elapsed time = %1.2lfms, Termination time = %1.2lfms",
ysr@777 3723 _elapsed_time_ms, _termination_time_ms);
ysr@777 3724 gclog_or_tty->print_cr(" Step Times (cum): num = %d, avg = %1.2lfms, sd = %1.2lfms",
ysr@777 3725 _step_times_ms.num(), _step_times_ms.avg(),
ysr@777 3726 _step_times_ms.sd());
ysr@777 3727 gclog_or_tty->print_cr(" max = %1.2lfms, total = %1.2lfms",
ysr@777 3728 _step_times_ms.maximum(), _step_times_ms.sum());
ysr@777 3729
ysr@777 3730 #if _MARKING_STATS_
ysr@777 3731 gclog_or_tty->print_cr(" Clock Intervals (cum): num = %d, avg = %1.2lfms, sd = %1.2lfms",
ysr@777 3732 _all_clock_intervals_ms.num(), _all_clock_intervals_ms.avg(),
ysr@777 3733 _all_clock_intervals_ms.sd());
ysr@777 3734 gclog_or_tty->print_cr(" max = %1.2lfms, total = %1.2lfms",
ysr@777 3735 _all_clock_intervals_ms.maximum(),
ysr@777 3736 _all_clock_intervals_ms.sum());
ysr@777 3737 gclog_or_tty->print_cr(" Clock Causes (cum): scanning = %d, marking = %d",
ysr@777 3738 _clock_due_to_scanning, _clock_due_to_marking);
ysr@777 3739 gclog_or_tty->print_cr(" Objects: scanned = %d, found on the bitmap = %d",
ysr@777 3740 _objs_scanned, _objs_found_on_bitmap);
ysr@777 3741 gclog_or_tty->print_cr(" Local Queue: pushes = %d, pops = %d, max size = %d",
ysr@777 3742 _local_pushes, _local_pops, _local_max_size);
ysr@777 3743 gclog_or_tty->print_cr(" Global Stack: pushes = %d, pops = %d, max size = %d",
ysr@777 3744 _global_pushes, _global_pops, _global_max_size);
ysr@777 3745 gclog_or_tty->print_cr(" transfers to = %d, transfers from = %d",
ysr@777 3746 _global_transfers_to,_global_transfers_from);
tonyp@3691 3747 gclog_or_tty->print_cr(" Regions: claimed = %d", _regions_claimed);
ysr@777 3748 gclog_or_tty->print_cr(" SATB buffers: processed = %d", _satb_buffers_processed);
ysr@777 3749 gclog_or_tty->print_cr(" Steals: attempts = %d, successes = %d",
ysr@777 3750 _steal_attempts, _steals);
ysr@777 3751 gclog_or_tty->print_cr(" Aborted: %d, due to", _aborted);
ysr@777 3752 gclog_or_tty->print_cr(" overflow: %d, global abort: %d, yield: %d",
ysr@777 3753 _aborted_overflow, _aborted_cm_aborted, _aborted_yield);
ysr@777 3754 gclog_or_tty->print_cr(" time out: %d, SATB: %d, termination: %d",
ysr@777 3755 _aborted_timed_out, _aborted_satb, _aborted_termination);
ysr@777 3756 #endif // _MARKING_STATS_
ysr@777 3757 }
ysr@777 3758
ysr@777 3759 /*****************************************************************************
ysr@777 3760
ysr@777 3761 The do_marking_step(time_target_ms) method is the building block
ysr@777 3762 of the parallel marking framework. It can be called in parallel
ysr@777 3763 with other invocations of do_marking_step() on different tasks
ysr@777 3764 (but only one per task, obviously) and concurrently with the
ysr@777 3765 mutator threads, or during remark, hence it eliminates the need
ysr@777 3766 for two versions of the code. When called during remark, it will
ysr@777 3767 pick up from where the task left off during the concurrent marking
ysr@777 3768 phase. Interestingly, tasks are also claimable during evacuation
ysr@777 3769 pauses too, since do_marking_step() ensures that it aborts before
ysr@777 3770 it needs to yield.
ysr@777 3771
ysr@777 3772 The data structures that is uses to do marking work are the
ysr@777 3773 following:
ysr@777 3774
ysr@777 3775 (1) Marking Bitmap. If there are gray objects that appear only
ysr@777 3776 on the bitmap (this happens either when dealing with an overflow
ysr@777 3777 or when the initial marking phase has simply marked the roots
ysr@777 3778 and didn't push them on the stack), then tasks claim heap
ysr@777 3779 regions whose bitmap they then scan to find gray objects. A
ysr@777 3780 global finger indicates where the end of the last claimed region
ysr@777 3781 is. A local finger indicates how far into the region a task has
ysr@777 3782 scanned. The two fingers are used to determine how to gray an
ysr@777 3783 object (i.e. whether simply marking it is OK, as it will be
ysr@777 3784 visited by a task in the future, or whether it needs to be also
ysr@777 3785 pushed on a stack).
ysr@777 3786
ysr@777 3787 (2) Local Queue. The local queue of the task which is accessed
ysr@777 3788 reasonably efficiently by the task. Other tasks can steal from
ysr@777 3789 it when they run out of work. Throughout the marking phase, a
ysr@777 3790 task attempts to keep its local queue short but not totally
ysr@777 3791 empty, so that entries are available for stealing by other
ysr@777 3792 tasks. Only when there is no more work, a task will totally
ysr@777 3793 drain its local queue.
ysr@777 3794
ysr@777 3795 (3) Global Mark Stack. This handles local queue overflow. During
ysr@777 3796 marking only sets of entries are moved between it and the local
ysr@777 3797 queues, as access to it requires a mutex and more fine-grain
ysr@777 3798 interaction with it which might cause contention. If it
ysr@777 3799 overflows, then the marking phase should restart and iterate
ysr@777 3800 over the bitmap to identify gray objects. Throughout the marking
ysr@777 3801 phase, tasks attempt to keep the global mark stack at a small
ysr@777 3802 length but not totally empty, so that entries are available for
ysr@777 3803 popping by other tasks. Only when there is no more work, tasks
ysr@777 3804 will totally drain the global mark stack.
ysr@777 3805
tonyp@3691 3806 (4) SATB Buffer Queue. This is where completed SATB buffers are
ysr@777 3807 made available. Buffers are regularly removed from this queue
ysr@777 3808 and scanned for roots, so that the queue doesn't get too
ysr@777 3809 long. During remark, all completed buffers are processed, as
ysr@777 3810 well as the filled in parts of any uncompleted buffers.
ysr@777 3811
ysr@777 3812 The do_marking_step() method tries to abort when the time target
ysr@777 3813 has been reached. There are a few other cases when the
ysr@777 3814 do_marking_step() method also aborts:
ysr@777 3815
ysr@777 3816 (1) When the marking phase has been aborted (after a Full GC).
ysr@777 3817
tonyp@3691 3818 (2) When a global overflow (on the global stack) has been
tonyp@3691 3819 triggered. Before the task aborts, it will actually sync up with
tonyp@3691 3820 the other tasks to ensure that all the marking data structures
tonyp@3691 3821 (local queues, stacks, fingers etc.) are re-initialised so that
tonyp@3691 3822 when do_marking_step() completes, the marking phase can
tonyp@3691 3823 immediately restart.
ysr@777 3824
ysr@777 3825 (3) When enough completed SATB buffers are available. The
ysr@777 3826 do_marking_step() method only tries to drain SATB buffers right
ysr@777 3827 at the beginning. So, if enough buffers are available, the
ysr@777 3828 marking step aborts and the SATB buffers are processed at
ysr@777 3829 the beginning of the next invocation.
ysr@777 3830
ysr@777 3831 (4) To yield. when we have to yield then we abort and yield
ysr@777 3832 right at the end of do_marking_step(). This saves us from a lot
ysr@777 3833 of hassle as, by yielding we might allow a Full GC. If this
ysr@777 3834 happens then objects will be compacted underneath our feet, the
ysr@777 3835 heap might shrink, etc. We save checking for this by just
ysr@777 3836 aborting and doing the yield right at the end.
ysr@777 3837
ysr@777 3838 From the above it follows that the do_marking_step() method should
ysr@777 3839 be called in a loop (or, otherwise, regularly) until it completes.
ysr@777 3840
ysr@777 3841 If a marking step completes without its has_aborted() flag being
ysr@777 3842 true, it means it has completed the current marking phase (and
ysr@777 3843 also all other marking tasks have done so and have all synced up).
ysr@777 3844
ysr@777 3845 A method called regular_clock_call() is invoked "regularly" (in
ysr@777 3846 sub ms intervals) throughout marking. It is this clock method that
ysr@777 3847 checks all the abort conditions which were mentioned above and
ysr@777 3848 decides when the task should abort. A work-based scheme is used to
ysr@777 3849 trigger this clock method: when the number of object words the
ysr@777 3850 marking phase has scanned or the number of references the marking
ysr@777 3851 phase has visited reach a given limit. Additional invocations to
ysr@777 3852 the method clock have been planted in a few other strategic places
ysr@777 3853 too. The initial reason for the clock method was to avoid calling
ysr@777 3854 vtime too regularly, as it is quite expensive. So, once it was in
ysr@777 3855 place, it was natural to piggy-back all the other conditions on it
ysr@777 3856 too and not constantly check them throughout the code.
ysr@777 3857
ysr@777 3858 *****************************************************************************/
ysr@777 3859
johnc@2494 3860 void CMTask::do_marking_step(double time_target_ms,
johnc@2494 3861 bool do_stealing,
johnc@2494 3862 bool do_termination) {
tonyp@1458 3863 assert(time_target_ms >= 1.0, "minimum granularity is 1ms");
tonyp@1458 3864 assert(concurrent() == _cm->concurrent(), "they should be the same");
tonyp@1458 3865
ysr@777 3866 G1CollectorPolicy* g1_policy = _g1h->g1_policy();
tonyp@1458 3867 assert(_task_queues != NULL, "invariant");
tonyp@1458 3868 assert(_task_queue != NULL, "invariant");
johnc@4173 3869 assert(_task_queues->queue(_worker_id) == _task_queue, "invariant");
tonyp@1458 3870
tonyp@1458 3871 assert(!_claimed,
tonyp@1458 3872 "only one thread should claim this task at any one time");
ysr@777 3873
ysr@777 3874 // OK, this doesn't safeguard again all possible scenarios, as it is
ysr@777 3875 // possible for two threads to set the _claimed flag at the same
ysr@777 3876 // time. But it is only for debugging purposes anyway and it will
ysr@777 3877 // catch most problems.
ysr@777 3878 _claimed = true;
ysr@777 3879
ysr@777 3880 _start_time_ms = os::elapsedVTime() * 1000.0;
ysr@777 3881 statsOnly( _interval_start_time_ms = _start_time_ms );
ysr@777 3882
ysr@777 3883 double diff_prediction_ms =
ysr@777 3884 g1_policy->get_new_prediction(&_marking_step_diffs_ms);
ysr@777 3885 _time_target_ms = time_target_ms - diff_prediction_ms;
ysr@777 3886
ysr@777 3887 // set up the variables that are used in the work-based scheme to
ysr@777 3888 // call the regular clock method
ysr@777 3889 _words_scanned = 0;
ysr@777 3890 _refs_reached = 0;
ysr@777 3891 recalculate_limits();
ysr@777 3892
ysr@777 3893 // clear all flags
ysr@777 3894 clear_has_aborted();
johnc@2494 3895 _has_timed_out = false;
ysr@777 3896 _draining_satb_buffers = false;
ysr@777 3897
ysr@777 3898 ++_calls;
ysr@777 3899
tonyp@2973 3900 if (_cm->verbose_low()) {
johnc@4173 3901 gclog_or_tty->print_cr("[%u] >>>>>>>>>> START, call = %d, "
ysr@777 3902 "target = %1.2lfms >>>>>>>>>>",
johnc@4173 3903 _worker_id, _calls, _time_target_ms);
tonyp@2973 3904 }
ysr@777 3905
ysr@777 3906 // Set up the bitmap and oop closures. Anything that uses them is
ysr@777 3907 // eventually called from this method, so it is OK to allocate these
ysr@777 3908 // statically.
ysr@777 3909 CMBitMapClosure bitmap_closure(this, _cm, _nextMarkBitMap);
tonyp@2968 3910 G1CMOopClosure cm_oop_closure(_g1h, _cm, this);
tonyp@2968 3911 set_cm_oop_closure(&cm_oop_closure);
ysr@777 3912
ysr@777 3913 if (_cm->has_overflown()) {
tonyp@3691 3914 // This can happen if the mark stack overflows during a GC pause
tonyp@3691 3915 // and this task, after a yield point, restarts. We have to abort
tonyp@3691 3916 // as we need to get into the overflow protocol which happens
tonyp@3691 3917 // right at the end of this task.
ysr@777 3918 set_has_aborted();
ysr@777 3919 }
ysr@777 3920
ysr@777 3921 // First drain any available SATB buffers. After this, we will not
ysr@777 3922 // look at SATB buffers before the next invocation of this method.
ysr@777 3923 // If enough completed SATB buffers are queued up, the regular clock
ysr@777 3924 // will abort this task so that it restarts.
ysr@777 3925 drain_satb_buffers();
ysr@777 3926 // ...then partially drain the local queue and the global stack
ysr@777 3927 drain_local_queue(true);
ysr@777 3928 drain_global_stack(true);
ysr@777 3929
ysr@777 3930 do {
ysr@777 3931 if (!has_aborted() && _curr_region != NULL) {
ysr@777 3932 // This means that we're already holding on to a region.
tonyp@1458 3933 assert(_finger != NULL, "if region is not NULL, then the finger "
tonyp@1458 3934 "should not be NULL either");
ysr@777 3935
ysr@777 3936 // We might have restarted this task after an evacuation pause
ysr@777 3937 // which might have evacuated the region we're holding on to
ysr@777 3938 // underneath our feet. Let's read its limit again to make sure
ysr@777 3939 // that we do not iterate over a region of the heap that
ysr@777 3940 // contains garbage (update_region_limit() will also move
ysr@777 3941 // _finger to the start of the region if it is found empty).
ysr@777 3942 update_region_limit();
ysr@777 3943 // We will start from _finger not from the start of the region,
ysr@777 3944 // as we might be restarting this task after aborting half-way
ysr@777 3945 // through scanning this region. In this case, _finger points to
ysr@777 3946 // the address where we last found a marked object. If this is a
ysr@777 3947 // fresh region, _finger points to start().
ysr@777 3948 MemRegion mr = MemRegion(_finger, _region_limit);
ysr@777 3949
tonyp@2973 3950 if (_cm->verbose_low()) {
johnc@4173 3951 gclog_or_tty->print_cr("[%u] we're scanning part "
ysr@777 3952 "["PTR_FORMAT", "PTR_FORMAT") "
ysr@777 3953 "of region "PTR_FORMAT,
johnc@4173 3954 _worker_id, _finger, _region_limit, _curr_region);
tonyp@2973 3955 }
ysr@777 3956
ysr@777 3957 // Let's iterate over the bitmap of the part of the
ysr@777 3958 // region that is left.
tonyp@3691 3959 if (mr.is_empty() || _nextMarkBitMap->iterate(&bitmap_closure, mr)) {
ysr@777 3960 // We successfully completed iterating over the region. Now,
ysr@777 3961 // let's give up the region.
ysr@777 3962 giveup_current_region();
ysr@777 3963 regular_clock_call();
ysr@777 3964 } else {
tonyp@1458 3965 assert(has_aborted(), "currently the only way to do so");
ysr@777 3966 // The only way to abort the bitmap iteration is to return
ysr@777 3967 // false from the do_bit() method. However, inside the
ysr@777 3968 // do_bit() method we move the _finger to point to the
ysr@777 3969 // object currently being looked at. So, if we bail out, we
ysr@777 3970 // have definitely set _finger to something non-null.
tonyp@1458 3971 assert(_finger != NULL, "invariant");
ysr@777 3972
ysr@777 3973 // Region iteration was actually aborted. So now _finger
ysr@777 3974 // points to the address of the object we last scanned. If we
ysr@777 3975 // leave it there, when we restart this task, we will rescan
ysr@777 3976 // the object. It is easy to avoid this. We move the finger by
ysr@777 3977 // enough to point to the next possible object header (the
ysr@777 3978 // bitmap knows by how much we need to move it as it knows its
ysr@777 3979 // granularity).
apetrusenko@1749 3980 assert(_finger < _region_limit, "invariant");
apetrusenko@1749 3981 HeapWord* new_finger = _nextMarkBitMap->nextWord(_finger);
apetrusenko@1749 3982 // Check if bitmap iteration was aborted while scanning the last object
apetrusenko@1749 3983 if (new_finger >= _region_limit) {
tonyp@3691 3984 giveup_current_region();
apetrusenko@1749 3985 } else {
tonyp@3691 3986 move_finger_to(new_finger);
apetrusenko@1749 3987 }
ysr@777 3988 }
ysr@777 3989 }
ysr@777 3990 // At this point we have either completed iterating over the
ysr@777 3991 // region we were holding on to, or we have aborted.
ysr@777 3992
ysr@777 3993 // We then partially drain the local queue and the global stack.
ysr@777 3994 // (Do we really need this?)
ysr@777 3995 drain_local_queue(true);
ysr@777 3996 drain_global_stack(true);
ysr@777 3997
ysr@777 3998 // Read the note on the claim_region() method on why it might
ysr@777 3999 // return NULL with potentially more regions available for
ysr@777 4000 // claiming and why we have to check out_of_regions() to determine
ysr@777 4001 // whether we're done or not.
ysr@777 4002 while (!has_aborted() && _curr_region == NULL && !_cm->out_of_regions()) {
ysr@777 4003 // We are going to try to claim a new region. We should have
ysr@777 4004 // given up on the previous one.
tonyp@1458 4005 // Separated the asserts so that we know which one fires.
tonyp@1458 4006 assert(_curr_region == NULL, "invariant");
tonyp@1458 4007 assert(_finger == NULL, "invariant");
tonyp@1458 4008 assert(_region_limit == NULL, "invariant");
tonyp@2973 4009 if (_cm->verbose_low()) {
johnc@4173 4010 gclog_or_tty->print_cr("[%u] trying to claim a new region", _worker_id);
tonyp@2973 4011 }
johnc@4173 4012 HeapRegion* claimed_region = _cm->claim_region(_worker_id);
ysr@777 4013 if (claimed_region != NULL) {
ysr@777 4014 // Yes, we managed to claim one
ysr@777 4015 statsOnly( ++_regions_claimed );
ysr@777 4016
tonyp@2973 4017 if (_cm->verbose_low()) {
johnc@4173 4018 gclog_or_tty->print_cr("[%u] we successfully claimed "
ysr@777 4019 "region "PTR_FORMAT,
johnc@4173 4020 _worker_id, claimed_region);
tonyp@2973 4021 }
ysr@777 4022
ysr@777 4023 setup_for_region(claimed_region);
tonyp@1458 4024 assert(_curr_region == claimed_region, "invariant");
ysr@777 4025 }
ysr@777 4026 // It is important to call the regular clock here. It might take
ysr@777 4027 // a while to claim a region if, for example, we hit a large
ysr@777 4028 // block of empty regions. So we need to call the regular clock
ysr@777 4029 // method once round the loop to make sure it's called
ysr@777 4030 // frequently enough.
ysr@777 4031 regular_clock_call();
ysr@777 4032 }
ysr@777 4033
ysr@777 4034 if (!has_aborted() && _curr_region == NULL) {
tonyp@1458 4035 assert(_cm->out_of_regions(),
tonyp@1458 4036 "at this point we should be out of regions");
ysr@777 4037 }
ysr@777 4038 } while ( _curr_region != NULL && !has_aborted());
ysr@777 4039
ysr@777 4040 if (!has_aborted()) {
ysr@777 4041 // We cannot check whether the global stack is empty, since other
tonyp@3691 4042 // tasks might be pushing objects to it concurrently.
tonyp@1458 4043 assert(_cm->out_of_regions(),
tonyp@1458 4044 "at this point we should be out of regions");
ysr@777 4045
tonyp@2973 4046 if (_cm->verbose_low()) {
johnc@4173 4047 gclog_or_tty->print_cr("[%u] all regions claimed", _worker_id);
tonyp@2973 4048 }
ysr@777 4049
ysr@777 4050 // Try to reduce the number of available SATB buffers so that
ysr@777 4051 // remark has less work to do.
ysr@777 4052 drain_satb_buffers();
ysr@777 4053 }
ysr@777 4054
ysr@777 4055 // Since we've done everything else, we can now totally drain the
ysr@777 4056 // local queue and global stack.
ysr@777 4057 drain_local_queue(false);
ysr@777 4058 drain_global_stack(false);
ysr@777 4059
ysr@777 4060 // Attempt at work stealing from other task's queues.
johnc@2494 4061 if (do_stealing && !has_aborted()) {
ysr@777 4062 // We have not aborted. This means that we have finished all that
ysr@777 4063 // we could. Let's try to do some stealing...
ysr@777 4064
ysr@777 4065 // We cannot check whether the global stack is empty, since other
tonyp@3691 4066 // tasks might be pushing objects to it concurrently.
tonyp@1458 4067 assert(_cm->out_of_regions() && _task_queue->size() == 0,
tonyp@1458 4068 "only way to reach here");
ysr@777 4069
tonyp@2973 4070 if (_cm->verbose_low()) {
johnc@4173 4071 gclog_or_tty->print_cr("[%u] starting to steal", _worker_id);
tonyp@2973 4072 }
ysr@777 4073
ysr@777 4074 while (!has_aborted()) {
ysr@777 4075 oop obj;
ysr@777 4076 statsOnly( ++_steal_attempts );
ysr@777 4077
johnc@4173 4078 if (_cm->try_stealing(_worker_id, &_hash_seed, obj)) {
tonyp@2973 4079 if (_cm->verbose_medium()) {
johnc@4173 4080 gclog_or_tty->print_cr("[%u] stolen "PTR_FORMAT" successfully",
johnc@4173 4081 _worker_id, (void*) obj);
tonyp@2973 4082 }
ysr@777 4083
ysr@777 4084 statsOnly( ++_steals );
ysr@777 4085
tonyp@1458 4086 assert(_nextMarkBitMap->isMarked((HeapWord*) obj),
tonyp@1458 4087 "any stolen object should be marked");
ysr@777 4088 scan_object(obj);
ysr@777 4089
ysr@777 4090 // And since we're towards the end, let's totally drain the
ysr@777 4091 // local queue and global stack.
ysr@777 4092 drain_local_queue(false);
ysr@777 4093 drain_global_stack(false);
ysr@777 4094 } else {
ysr@777 4095 break;
ysr@777 4096 }
ysr@777 4097 }
ysr@777 4098 }
ysr@777 4099
tonyp@2848 4100 // If we are about to wrap up and go into termination, check if we
tonyp@2848 4101 // should raise the overflow flag.
tonyp@2848 4102 if (do_termination && !has_aborted()) {
tonyp@2848 4103 if (_cm->force_overflow()->should_force()) {
tonyp@2848 4104 _cm->set_has_overflown();
tonyp@2848 4105 regular_clock_call();
tonyp@2848 4106 }
tonyp@2848 4107 }
tonyp@2848 4108
ysr@777 4109 // We still haven't aborted. Now, let's try to get into the
ysr@777 4110 // termination protocol.
johnc@2494 4111 if (do_termination && !has_aborted()) {
ysr@777 4112 // We cannot check whether the global stack is empty, since other
tonyp@3691 4113 // tasks might be concurrently pushing objects on it.
tonyp@1458 4114 // Separated the asserts so that we know which one fires.
tonyp@1458 4115 assert(_cm->out_of_regions(), "only way to reach here");
tonyp@1458 4116 assert(_task_queue->size() == 0, "only way to reach here");
ysr@777 4117
tonyp@2973 4118 if (_cm->verbose_low()) {
johnc@4173 4119 gclog_or_tty->print_cr("[%u] starting termination protocol", _worker_id);
tonyp@2973 4120 }
ysr@777 4121
ysr@777 4122 _termination_start_time_ms = os::elapsedVTime() * 1000.0;
ysr@777 4123 // The CMTask class also extends the TerminatorTerminator class,
ysr@777 4124 // hence its should_exit_termination() method will also decide
ysr@777 4125 // whether to exit the termination protocol or not.
ysr@777 4126 bool finished = _cm->terminator()->offer_termination(this);
ysr@777 4127 double termination_end_time_ms = os::elapsedVTime() * 1000.0;
ysr@777 4128 _termination_time_ms +=
ysr@777 4129 termination_end_time_ms - _termination_start_time_ms;
ysr@777 4130
ysr@777 4131 if (finished) {
ysr@777 4132 // We're all done.
ysr@777 4133
johnc@4173 4134 if (_worker_id == 0) {
ysr@777 4135 // let's allow task 0 to do this
ysr@777 4136 if (concurrent()) {
tonyp@1458 4137 assert(_cm->concurrent_marking_in_progress(), "invariant");
ysr@777 4138 // we need to set this to false before the next
ysr@777 4139 // safepoint. This way we ensure that the marking phase
ysr@777 4140 // doesn't observe any more heap expansions.
ysr@777 4141 _cm->clear_concurrent_marking_in_progress();
ysr@777 4142 }
ysr@777 4143 }
ysr@777 4144
ysr@777 4145 // We can now guarantee that the global stack is empty, since
tonyp@1458 4146 // all other tasks have finished. We separated the guarantees so
tonyp@1458 4147 // that, if a condition is false, we can immediately find out
tonyp@1458 4148 // which one.
tonyp@1458 4149 guarantee(_cm->out_of_regions(), "only way to reach here");
tonyp@1458 4150 guarantee(_cm->mark_stack_empty(), "only way to reach here");
tonyp@1458 4151 guarantee(_task_queue->size() == 0, "only way to reach here");
tonyp@1458 4152 guarantee(!_cm->has_overflown(), "only way to reach here");
tonyp@1458 4153 guarantee(!_cm->mark_stack_overflow(), "only way to reach here");
ysr@777 4154
tonyp@2973 4155 if (_cm->verbose_low()) {
johnc@4173 4156 gclog_or_tty->print_cr("[%u] all tasks terminated", _worker_id);
tonyp@2973 4157 }
ysr@777 4158 } else {
ysr@777 4159 // Apparently there's more work to do. Let's abort this task. It
ysr@777 4160 // will restart it and we can hopefully find more things to do.
ysr@777 4161
tonyp@2973 4162 if (_cm->verbose_low()) {
johnc@4173 4163 gclog_or_tty->print_cr("[%u] apparently there is more work to do",
johnc@4173 4164 _worker_id);
tonyp@2973 4165 }
ysr@777 4166
ysr@777 4167 set_has_aborted();
ysr@777 4168 statsOnly( ++_aborted_termination );
ysr@777 4169 }
ysr@777 4170 }
ysr@777 4171
ysr@777 4172 // Mainly for debugging purposes to make sure that a pointer to the
ysr@777 4173 // closure which was statically allocated in this frame doesn't
ysr@777 4174 // escape it by accident.
tonyp@2968 4175 set_cm_oop_closure(NULL);
ysr@777 4176 double end_time_ms = os::elapsedVTime() * 1000.0;
ysr@777 4177 double elapsed_time_ms = end_time_ms - _start_time_ms;
ysr@777 4178 // Update the step history.
ysr@777 4179 _step_times_ms.add(elapsed_time_ms);
ysr@777 4180
ysr@777 4181 if (has_aborted()) {
ysr@777 4182 // The task was aborted for some reason.
ysr@777 4183
ysr@777 4184 statsOnly( ++_aborted );
ysr@777 4185
johnc@2494 4186 if (_has_timed_out) {
ysr@777 4187 double diff_ms = elapsed_time_ms - _time_target_ms;
ysr@777 4188 // Keep statistics of how well we did with respect to hitting
ysr@777 4189 // our target only if we actually timed out (if we aborted for
ysr@777 4190 // other reasons, then the results might get skewed).
ysr@777 4191 _marking_step_diffs_ms.add(diff_ms);
ysr@777 4192 }
ysr@777 4193
ysr@777 4194 if (_cm->has_overflown()) {
ysr@777 4195 // This is the interesting one. We aborted because a global
ysr@777 4196 // overflow was raised. This means we have to restart the
ysr@777 4197 // marking phase and start iterating over regions. However, in
ysr@777 4198 // order to do this we have to make sure that all tasks stop
ysr@777 4199 // what they are doing and re-initialise in a safe manner. We
ysr@777 4200 // will achieve this with the use of two barrier sync points.
ysr@777 4201
tonyp@2973 4202 if (_cm->verbose_low()) {
johnc@4173 4203 gclog_or_tty->print_cr("[%u] detected overflow", _worker_id);
tonyp@2973 4204 }
ysr@777 4205
johnc@4173 4206 _cm->enter_first_sync_barrier(_worker_id);
ysr@777 4207 // When we exit this sync barrier we know that all tasks have
ysr@777 4208 // stopped doing marking work. So, it's now safe to
ysr@777 4209 // re-initialise our data structures. At the end of this method,
ysr@777 4210 // task 0 will clear the global data structures.
ysr@777 4211
ysr@777 4212 statsOnly( ++_aborted_overflow );
ysr@777 4213
ysr@777 4214 // We clear the local state of this task...
ysr@777 4215 clear_region_fields();
ysr@777 4216
ysr@777 4217 // ...and enter the second barrier.
johnc@4173 4218 _cm->enter_second_sync_barrier(_worker_id);
ysr@777 4219 // At this point everything has bee re-initialised and we're
ysr@777 4220 // ready to restart.
ysr@777 4221 }
ysr@777 4222
ysr@777 4223 if (_cm->verbose_low()) {
johnc@4173 4224 gclog_or_tty->print_cr("[%u] <<<<<<<<<< ABORTING, target = %1.2lfms, "
ysr@777 4225 "elapsed = %1.2lfms <<<<<<<<<<",
johnc@4173 4226 _worker_id, _time_target_ms, elapsed_time_ms);
tonyp@2973 4227 if (_cm->has_aborted()) {
johnc@4173 4228 gclog_or_tty->print_cr("[%u] ========== MARKING ABORTED ==========",
johnc@4173 4229 _worker_id);
tonyp@2973 4230 }
ysr@777 4231 }
ysr@777 4232 } else {
tonyp@2973 4233 if (_cm->verbose_low()) {
johnc@4173 4234 gclog_or_tty->print_cr("[%u] <<<<<<<<<< FINISHED, target = %1.2lfms, "
ysr@777 4235 "elapsed = %1.2lfms <<<<<<<<<<",
johnc@4173 4236 _worker_id, _time_target_ms, elapsed_time_ms);
tonyp@2973 4237 }
ysr@777 4238 }
ysr@777 4239
ysr@777 4240 _claimed = false;
ysr@777 4241 }
ysr@777 4242
johnc@4173 4243 CMTask::CMTask(uint worker_id,
ysr@777 4244 ConcurrentMark* cm,
johnc@3463 4245 size_t* marked_bytes,
johnc@3463 4246 BitMap* card_bm,
ysr@777 4247 CMTaskQueue* task_queue,
ysr@777 4248 CMTaskQueueSet* task_queues)
ysr@777 4249 : _g1h(G1CollectedHeap::heap()),
johnc@4173 4250 _worker_id(worker_id), _cm(cm),
ysr@777 4251 _claimed(false),
ysr@777 4252 _nextMarkBitMap(NULL), _hash_seed(17),
ysr@777 4253 _task_queue(task_queue),
ysr@777 4254 _task_queues(task_queues),
tonyp@2968 4255 _cm_oop_closure(NULL),
johnc@3463 4256 _marked_bytes_array(marked_bytes),
johnc@3463 4257 _card_bm(card_bm) {
tonyp@1458 4258 guarantee(task_queue != NULL, "invariant");
tonyp@1458 4259 guarantee(task_queues != NULL, "invariant");
ysr@777 4260
ysr@777 4261 statsOnly( _clock_due_to_scanning = 0;
ysr@777 4262 _clock_due_to_marking = 0 );
ysr@777 4263
ysr@777 4264 _marking_step_diffs_ms.add(0.5);
ysr@777 4265 }
tonyp@2717 4266
tonyp@2717 4267 // These are formatting macros that are used below to ensure
tonyp@2717 4268 // consistent formatting. The *_H_* versions are used to format the
tonyp@2717 4269 // header for a particular value and they should be kept consistent
tonyp@2717 4270 // with the corresponding macro. Also note that most of the macros add
tonyp@2717 4271 // the necessary white space (as a prefix) which makes them a bit
tonyp@2717 4272 // easier to compose.
tonyp@2717 4273
tonyp@2717 4274 // All the output lines are prefixed with this string to be able to
tonyp@2717 4275 // identify them easily in a large log file.
tonyp@2717 4276 #define G1PPRL_LINE_PREFIX "###"
tonyp@2717 4277
tonyp@2717 4278 #define G1PPRL_ADDR_BASE_FORMAT " "PTR_FORMAT"-"PTR_FORMAT
tonyp@2717 4279 #ifdef _LP64
tonyp@2717 4280 #define G1PPRL_ADDR_BASE_H_FORMAT " %37s"
tonyp@2717 4281 #else // _LP64
tonyp@2717 4282 #define G1PPRL_ADDR_BASE_H_FORMAT " %21s"
tonyp@2717 4283 #endif // _LP64
tonyp@2717 4284
tonyp@2717 4285 // For per-region info
tonyp@2717 4286 #define G1PPRL_TYPE_FORMAT " %-4s"
tonyp@2717 4287 #define G1PPRL_TYPE_H_FORMAT " %4s"
tonyp@2717 4288 #define G1PPRL_BYTE_FORMAT " "SIZE_FORMAT_W(9)
tonyp@2717 4289 #define G1PPRL_BYTE_H_FORMAT " %9s"
tonyp@2717 4290 #define G1PPRL_DOUBLE_FORMAT " %14.1f"
tonyp@2717 4291 #define G1PPRL_DOUBLE_H_FORMAT " %14s"
tonyp@2717 4292
tonyp@2717 4293 // For summary info
tonyp@2717 4294 #define G1PPRL_SUM_ADDR_FORMAT(tag) " "tag":"G1PPRL_ADDR_BASE_FORMAT
tonyp@2717 4295 #define G1PPRL_SUM_BYTE_FORMAT(tag) " "tag": "SIZE_FORMAT
tonyp@2717 4296 #define G1PPRL_SUM_MB_FORMAT(tag) " "tag": %1.2f MB"
tonyp@2717 4297 #define G1PPRL_SUM_MB_PERC_FORMAT(tag) G1PPRL_SUM_MB_FORMAT(tag)" / %1.2f %%"
tonyp@2717 4298
tonyp@2717 4299 G1PrintRegionLivenessInfoClosure::
tonyp@2717 4300 G1PrintRegionLivenessInfoClosure(outputStream* out, const char* phase_name)
tonyp@2717 4301 : _out(out),
tonyp@2717 4302 _total_used_bytes(0), _total_capacity_bytes(0),
tonyp@2717 4303 _total_prev_live_bytes(0), _total_next_live_bytes(0),
tonyp@2717 4304 _hum_used_bytes(0), _hum_capacity_bytes(0),
tonyp@2717 4305 _hum_prev_live_bytes(0), _hum_next_live_bytes(0) {
tonyp@2717 4306 G1CollectedHeap* g1h = G1CollectedHeap::heap();
tonyp@2717 4307 MemRegion g1_committed = g1h->g1_committed();
tonyp@2717 4308 MemRegion g1_reserved = g1h->g1_reserved();
tonyp@2717 4309 double now = os::elapsedTime();
tonyp@2717 4310
tonyp@2717 4311 // Print the header of the output.
tonyp@2717 4312 _out->cr();
tonyp@2717 4313 _out->print_cr(G1PPRL_LINE_PREFIX" PHASE %s @ %1.3f", phase_name, now);
tonyp@2717 4314 _out->print_cr(G1PPRL_LINE_PREFIX" HEAP"
tonyp@2717 4315 G1PPRL_SUM_ADDR_FORMAT("committed")
tonyp@2717 4316 G1PPRL_SUM_ADDR_FORMAT("reserved")
tonyp@2717 4317 G1PPRL_SUM_BYTE_FORMAT("region-size"),
tonyp@2717 4318 g1_committed.start(), g1_committed.end(),
tonyp@2717 4319 g1_reserved.start(), g1_reserved.end(),
johnc@3182 4320 HeapRegion::GrainBytes);
tonyp@2717 4321 _out->print_cr(G1PPRL_LINE_PREFIX);
tonyp@2717 4322 _out->print_cr(G1PPRL_LINE_PREFIX
tonyp@2717 4323 G1PPRL_TYPE_H_FORMAT
tonyp@2717 4324 G1PPRL_ADDR_BASE_H_FORMAT
tonyp@2717 4325 G1PPRL_BYTE_H_FORMAT
tonyp@2717 4326 G1PPRL_BYTE_H_FORMAT
tonyp@2717 4327 G1PPRL_BYTE_H_FORMAT
tonyp@2717 4328 G1PPRL_DOUBLE_H_FORMAT,
tonyp@2717 4329 "type", "address-range",
tonyp@2717 4330 "used", "prev-live", "next-live", "gc-eff");
johnc@3173 4331 _out->print_cr(G1PPRL_LINE_PREFIX
johnc@3173 4332 G1PPRL_TYPE_H_FORMAT
johnc@3173 4333 G1PPRL_ADDR_BASE_H_FORMAT
johnc@3173 4334 G1PPRL_BYTE_H_FORMAT
johnc@3173 4335 G1PPRL_BYTE_H_FORMAT
johnc@3173 4336 G1PPRL_BYTE_H_FORMAT
johnc@3173 4337 G1PPRL_DOUBLE_H_FORMAT,
johnc@3173 4338 "", "",
johnc@3173 4339 "(bytes)", "(bytes)", "(bytes)", "(bytes/ms)");
tonyp@2717 4340 }
tonyp@2717 4341
tonyp@2717 4342 // It takes as a parameter a reference to one of the _hum_* fields, it
tonyp@2717 4343 // deduces the corresponding value for a region in a humongous region
tonyp@2717 4344 // series (either the region size, or what's left if the _hum_* field
tonyp@2717 4345 // is < the region size), and updates the _hum_* field accordingly.
tonyp@2717 4346 size_t G1PrintRegionLivenessInfoClosure::get_hum_bytes(size_t* hum_bytes) {
tonyp@2717 4347 size_t bytes = 0;
tonyp@2717 4348 // The > 0 check is to deal with the prev and next live bytes which
tonyp@2717 4349 // could be 0.
tonyp@2717 4350 if (*hum_bytes > 0) {
johnc@3182 4351 bytes = MIN2(HeapRegion::GrainBytes, *hum_bytes);
tonyp@2717 4352 *hum_bytes -= bytes;
tonyp@2717 4353 }
tonyp@2717 4354 return bytes;
tonyp@2717 4355 }
tonyp@2717 4356
tonyp@2717 4357 // It deduces the values for a region in a humongous region series
tonyp@2717 4358 // from the _hum_* fields and updates those accordingly. It assumes
tonyp@2717 4359 // that that _hum_* fields have already been set up from the "starts
tonyp@2717 4360 // humongous" region and we visit the regions in address order.
tonyp@2717 4361 void G1PrintRegionLivenessInfoClosure::get_hum_bytes(size_t* used_bytes,
tonyp@2717 4362 size_t* capacity_bytes,
tonyp@2717 4363 size_t* prev_live_bytes,
tonyp@2717 4364 size_t* next_live_bytes) {
tonyp@2717 4365 assert(_hum_used_bytes > 0 && _hum_capacity_bytes > 0, "pre-condition");
tonyp@2717 4366 *used_bytes = get_hum_bytes(&_hum_used_bytes);
tonyp@2717 4367 *capacity_bytes = get_hum_bytes(&_hum_capacity_bytes);
tonyp@2717 4368 *prev_live_bytes = get_hum_bytes(&_hum_prev_live_bytes);
tonyp@2717 4369 *next_live_bytes = get_hum_bytes(&_hum_next_live_bytes);
tonyp@2717 4370 }
tonyp@2717 4371
tonyp@2717 4372 bool G1PrintRegionLivenessInfoClosure::doHeapRegion(HeapRegion* r) {
tonyp@2717 4373 const char* type = "";
tonyp@2717 4374 HeapWord* bottom = r->bottom();
tonyp@2717 4375 HeapWord* end = r->end();
tonyp@2717 4376 size_t capacity_bytes = r->capacity();
tonyp@2717 4377 size_t used_bytes = r->used();
tonyp@2717 4378 size_t prev_live_bytes = r->live_bytes();
tonyp@2717 4379 size_t next_live_bytes = r->next_live_bytes();
tonyp@2717 4380 double gc_eff = r->gc_efficiency();
tonyp@2717 4381 if (r->used() == 0) {
tonyp@2717 4382 type = "FREE";
tonyp@2717 4383 } else if (r->is_survivor()) {
tonyp@2717 4384 type = "SURV";
tonyp@2717 4385 } else if (r->is_young()) {
tonyp@2717 4386 type = "EDEN";
tonyp@2717 4387 } else if (r->startsHumongous()) {
tonyp@2717 4388 type = "HUMS";
tonyp@2717 4389
tonyp@2717 4390 assert(_hum_used_bytes == 0 && _hum_capacity_bytes == 0 &&
tonyp@2717 4391 _hum_prev_live_bytes == 0 && _hum_next_live_bytes == 0,
tonyp@2717 4392 "they should have been zeroed after the last time we used them");
tonyp@2717 4393 // Set up the _hum_* fields.
tonyp@2717 4394 _hum_capacity_bytes = capacity_bytes;
tonyp@2717 4395 _hum_used_bytes = used_bytes;
tonyp@2717 4396 _hum_prev_live_bytes = prev_live_bytes;
tonyp@2717 4397 _hum_next_live_bytes = next_live_bytes;
tonyp@2717 4398 get_hum_bytes(&used_bytes, &capacity_bytes,
tonyp@2717 4399 &prev_live_bytes, &next_live_bytes);
tonyp@2717 4400 end = bottom + HeapRegion::GrainWords;
tonyp@2717 4401 } else if (r->continuesHumongous()) {
tonyp@2717 4402 type = "HUMC";
tonyp@2717 4403 get_hum_bytes(&used_bytes, &capacity_bytes,
tonyp@2717 4404 &prev_live_bytes, &next_live_bytes);
tonyp@2717 4405 assert(end == bottom + HeapRegion::GrainWords, "invariant");
tonyp@2717 4406 } else {
tonyp@2717 4407 type = "OLD";
tonyp@2717 4408 }
tonyp@2717 4409
tonyp@2717 4410 _total_used_bytes += used_bytes;
tonyp@2717 4411 _total_capacity_bytes += capacity_bytes;
tonyp@2717 4412 _total_prev_live_bytes += prev_live_bytes;
tonyp@2717 4413 _total_next_live_bytes += next_live_bytes;
tonyp@2717 4414
tonyp@2717 4415 // Print a line for this particular region.
tonyp@2717 4416 _out->print_cr(G1PPRL_LINE_PREFIX
tonyp@2717 4417 G1PPRL_TYPE_FORMAT
tonyp@2717 4418 G1PPRL_ADDR_BASE_FORMAT
tonyp@2717 4419 G1PPRL_BYTE_FORMAT
tonyp@2717 4420 G1PPRL_BYTE_FORMAT
tonyp@2717 4421 G1PPRL_BYTE_FORMAT
tonyp@2717 4422 G1PPRL_DOUBLE_FORMAT,
tonyp@2717 4423 type, bottom, end,
tonyp@2717 4424 used_bytes, prev_live_bytes, next_live_bytes, gc_eff);
tonyp@2717 4425
tonyp@2717 4426 return false;
tonyp@2717 4427 }
tonyp@2717 4428
tonyp@2717 4429 G1PrintRegionLivenessInfoClosure::~G1PrintRegionLivenessInfoClosure() {
tonyp@2717 4430 // Print the footer of the output.
tonyp@2717 4431 _out->print_cr(G1PPRL_LINE_PREFIX);
tonyp@2717 4432 _out->print_cr(G1PPRL_LINE_PREFIX
tonyp@2717 4433 " SUMMARY"
tonyp@2717 4434 G1PPRL_SUM_MB_FORMAT("capacity")
tonyp@2717 4435 G1PPRL_SUM_MB_PERC_FORMAT("used")
tonyp@2717 4436 G1PPRL_SUM_MB_PERC_FORMAT("prev-live")
tonyp@2717 4437 G1PPRL_SUM_MB_PERC_FORMAT("next-live"),
tonyp@2717 4438 bytes_to_mb(_total_capacity_bytes),
tonyp@2717 4439 bytes_to_mb(_total_used_bytes),
tonyp@2717 4440 perc(_total_used_bytes, _total_capacity_bytes),
tonyp@2717 4441 bytes_to_mb(_total_prev_live_bytes),
tonyp@2717 4442 perc(_total_prev_live_bytes, _total_capacity_bytes),
tonyp@2717 4443 bytes_to_mb(_total_next_live_bytes),
tonyp@2717 4444 perc(_total_next_live_bytes, _total_capacity_bytes));
tonyp@2717 4445 _out->cr();
tonyp@2717 4446 }

mercurial