src/share/vm/memory/barrierSet.inline.hpp

Thu, 21 Jan 2010 11:33:32 -0800

author
jmasa
date
Thu, 21 Jan 2010 11:33:32 -0800
changeset 1625
4788266644c1
parent 1526
6aa7255741f3
child 1680
6484c4ee11cb
permissions
-rw-r--r--

6895236: CMS: cmsOopClosures.inline.hpp:43 assert(..., "Should remember klasses in this context")
Summary: Adjust assertion checking for ExplicitGCInvokesConcurrentAndUnloadsClasses as a reason for class unloading
Reviewed-by: ysr

     1 /*
     2  * Copyright 2001-2008 Sun Microsystems, Inc.  All Rights Reserved.
     3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
     4  *
     5  * This code is free software; you can redistribute it and/or modify it
     6  * under the terms of the GNU General Public License version 2 only, as
     7  * published by the Free Software Foundation.
     8  *
     9  * This code is distributed in the hope that it will be useful, but WITHOUT
    10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
    11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
    12  * version 2 for more details (a copy is included in the LICENSE file that
    13  * accompanied this code).
    14  *
    15  * You should have received a copy of the GNU General Public License version
    16  * 2 along with this work; if not, write to the Free Software Foundation,
    17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
    18  *
    19  * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
    20  * CA 95054 USA or visit www.sun.com if you need additional information or
    21  * have any questions.
    22  *
    23  */
    25 // Inline functions of BarrierSet, which de-virtualize certain
    26 // performance-critical calls when the barrier is the most common
    27 // card-table kind.
    29 template <class T> void BarrierSet::write_ref_field_pre(T* field, oop new_val) {
    30   if (kind() == CardTableModRef) {
    31     ((CardTableModRefBS*)this)->inline_write_ref_field_pre(field, new_val);
    32   } else {
    33     write_ref_field_pre_work(field, new_val);
    34   }
    35 }
    37 void BarrierSet::write_ref_field(void* field, oop new_val) {
    38   if (kind() == CardTableModRef) {
    39     ((CardTableModRefBS*)this)->inline_write_ref_field(field, new_val);
    40   } else {
    41     write_ref_field_work(field, new_val);
    42   }
    43 }
    45 void BarrierSet::write_ref_array(MemRegion mr) {
    46   assert((HeapWord*)align_size_down((uintptr_t)mr.start(), HeapWordSize) == mr.start() , "Unaligned start");
    47   assert((HeapWord*)align_size_up  ((uintptr_t)mr.end(),   HeapWordSize) == mr.end(),    "Unaligned end"  );
    48   if (kind() == CardTableModRef) {
    49     ((CardTableModRefBS*)this)->inline_write_ref_array(mr);
    50   } else {
    51     write_ref_array_work(mr);
    52   }
    53 }
    55 // count is number of array elements being written
    56 void BarrierSet::write_ref_array(HeapWord* start, size_t count) {
    57   assert(count <= (size_t)max_intx, "count too large");
    58   HeapWord* end = (HeapWord*)((char*)start + (count*heapOopSize));
    59   // In the case of compressed oops, start and end may potentially be misaligned;
    60   // so we need to conservatively align the first downward (this is not
    61   // strictly necessary for current uses, but a case of good hygiene and,
    62   // if you will, aesthetics) and the second upward (this is essential for
    63   // current uses) to a HeapWord boundary, so we mark all cards overlapping
    64   // this write. In the event that this evolves in the future to calling a
    65   // logging barrier of narrow oop granularity, like the pre-barrier for G1
    66   // (mentioned here merely by way of example), we will need to change this
    67   // interface, much like the pre-barrier one above, so it is "exactly precise"
    68   // (if i may be allowed the adverbial redundancy for emphasis) and does not
    69   // include narrow oop slots not included in the original write interval.
    70   HeapWord* aligned_start = (HeapWord*)align_size_down((uintptr_t)start, HeapWordSize);
    71   HeapWord* aligned_end   = (HeapWord*)align_size_up  ((uintptr_t)end,   HeapWordSize);
    72   // If compressed oops were not being used, these should already be aligned
    73   assert(UseCompressedOops || (aligned_start == start && aligned_end == end),
    74          "Expected heap word alignment of start and end");
    75 #if 0
    76   warning("Post:\t" INTPTR_FORMAT "[" SIZE_FORMAT "] : [" INTPTR_FORMAT","INTPTR_FORMAT")\t",
    77                    start,            count,              aligned_start,   aligned_end);
    78 #endif
    79   write_ref_array_work(MemRegion(aligned_start, aligned_end));
    80 }
    83 void BarrierSet::write_region(MemRegion mr) {
    84   if (kind() == CardTableModRef) {
    85     ((CardTableModRefBS*)this)->inline_write_region(mr);
    86   } else {
    87     write_region_work(mr);
    88   }
    89 }

mercurial