src/share/vm/gc_implementation/g1/g1CollectedHeap.inline.hpp

Mon, 03 Aug 2009 12:59:30 -0700

author
johnc
date
Mon, 03 Aug 2009 12:59:30 -0700
changeset 1324
15c5903cf9e1
parent 1187
a2957df801a1
child 1907
c18cbe5936b8
permissions
-rw-r--r--

6865703: G1: Parallelize hot card cache cleanup
Summary: Have the GC worker threads clear the hot card cache in parallel by having each worker thread claim a chunk of the card cache and process the cards in that chunk. The size of the chunks that each thread will claim is determined at VM initialization from the size of the card cache and the number of worker threads.
Reviewed-by: jmasa, tonyp

ysr@777 1 /*
xdono@1014 2 * Copyright 2001-2009 Sun Microsystems, Inc. All Rights Reserved.
ysr@777 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
ysr@777 4 *
ysr@777 5 * This code is free software; you can redistribute it and/or modify it
ysr@777 6 * under the terms of the GNU General Public License version 2 only, as
ysr@777 7 * published by the Free Software Foundation.
ysr@777 8 *
ysr@777 9 * This code is distributed in the hope that it will be useful, but WITHOUT
ysr@777 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
ysr@777 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
ysr@777 12 * version 2 for more details (a copy is included in the LICENSE file that
ysr@777 13 * accompanied this code).
ysr@777 14 *
ysr@777 15 * You should have received a copy of the GNU General Public License version
ysr@777 16 * 2 along with this work; if not, write to the Free Software Foundation,
ysr@777 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
ysr@777 18 *
ysr@777 19 * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
ysr@777 20 * CA 95054 USA or visit www.sun.com if you need additional information or
ysr@777 21 * have any questions.
ysr@777 22 *
ysr@777 23 */
ysr@777 24
ysr@777 25 // Inline functions for G1CollectedHeap
ysr@777 26
ysr@777 27 inline HeapRegion*
ysr@777 28 G1CollectedHeap::heap_region_containing(const void* addr) const {
ysr@777 29 HeapRegion* hr = _hrs->addr_to_region(addr);
ysr@777 30 // hr can be null if addr in perm_gen
ysr@777 31 if (hr != NULL && hr->continuesHumongous()) {
ysr@777 32 hr = hr->humongous_start_region();
ysr@777 33 }
ysr@777 34 return hr;
ysr@777 35 }
ysr@777 36
ysr@777 37 inline HeapRegion*
ysr@777 38 G1CollectedHeap::heap_region_containing_raw(const void* addr) const {
tonyp@961 39 assert(_g1_reserved.contains(addr), "invariant");
johnc@1187 40 size_t index = pointer_delta(addr, _g1_reserved.start(), 1)
johnc@1187 41 >> HeapRegion::LogOfHRGrainBytes;
johnc@1187 42
tonyp@961 43 HeapRegion* res = _hrs->at(index);
tonyp@961 44 assert(res == _hrs->addr_to_region(addr), "sanity");
ysr@777 45 return res;
ysr@777 46 }
ysr@777 47
ysr@777 48 inline bool G1CollectedHeap::obj_in_cs(oop obj) {
ysr@777 49 HeapRegion* r = _hrs->addr_to_region(obj);
ysr@777 50 return r != NULL && r->in_collection_set();
ysr@777 51 }
ysr@777 52
ysr@777 53 inline HeapWord* G1CollectedHeap::attempt_allocation(size_t word_size,
ysr@777 54 bool permit_collection_pause) {
ysr@777 55 HeapWord* res = NULL;
ysr@777 56
ysr@777 57 assert( SafepointSynchronize::is_at_safepoint() ||
ysr@777 58 Heap_lock->owned_by_self(), "pre-condition of the call" );
ysr@777 59
ysr@777 60 if (_cur_alloc_region != NULL) {
ysr@777 61
ysr@777 62 // If this allocation causes a region to become non empty,
ysr@777 63 // then we need to update our free_regions count.
ysr@777 64
ysr@777 65 if (_cur_alloc_region->is_empty()) {
ysr@777 66 res = _cur_alloc_region->allocate(word_size);
ysr@777 67 if (res != NULL)
ysr@777 68 _free_regions--;
ysr@777 69 } else {
ysr@777 70 res = _cur_alloc_region->allocate(word_size);
ysr@777 71 }
ysr@777 72 }
ysr@777 73 if (res != NULL) {
ysr@777 74 if (!SafepointSynchronize::is_at_safepoint()) {
ysr@777 75 assert( Heap_lock->owned_by_self(), "invariant" );
ysr@777 76 Heap_lock->unlock();
ysr@777 77 }
ysr@777 78 return res;
ysr@777 79 }
ysr@777 80 // attempt_allocation_slow will also unlock the heap lock when appropriate.
ysr@777 81 return attempt_allocation_slow(word_size, permit_collection_pause);
ysr@777 82 }
ysr@777 83
ysr@777 84 inline RefToScanQueue* G1CollectedHeap::task_queue(int i) {
ysr@777 85 return _task_queues->queue(i);
ysr@777 86 }
ysr@777 87
ysr@777 88
ysr@777 89 inline bool G1CollectedHeap::isMarkedPrev(oop obj) const {
ysr@777 90 return _cm->prevMarkBitMap()->isMarked((HeapWord *)obj);
ysr@777 91 }
ysr@777 92
ysr@777 93 inline bool G1CollectedHeap::isMarkedNext(oop obj) const {
ysr@777 94 return _cm->nextMarkBitMap()->isMarked((HeapWord *)obj);
ysr@777 95 }

mercurial