src/share/vm/oops/instanceMirrorKlass.cpp

Thu, 07 Apr 2011 09:53:20 -0700

author
johnc
date
Thu, 07 Apr 2011 09:53:20 -0700
changeset 2781
e1162778c1c8
parent 2658
c7f3d0b4570f
child 3205
e5928e7dab26
permissions
-rw-r--r--

7009266: G1: assert(obj->is_oop_or_null(true )) failed: Error
Summary: A referent object that is only weakly reachable at the start of concurrent marking but is re-attached to the strongly reachable object graph during marking may not be marked as live. This can cause the reference object to be processed prematurely and leave dangling pointers to the referent object. Implement a read barrier for the java.lang.ref.Reference::referent field by intrinsifying the Reference.get() method, and intercepting accesses though JNI, reflection, and Unsafe, so that when a non-null referent object is read it is also logged in an SATB buffer.
Reviewed-by: kvn, iveresov, never, tonyp, dholmes

     1 /*
     2  * Copyright (c) 2011, Oracle and/or its affiliates. All rights reserved.
     3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
     4  *
     5  * This code is free software; you can redistribute it and/or modify it
     6  * under the terms of the GNU General Public License version 2 only, as
     7  * published by the Free Software Foundation.
     8  *
     9  * This code is distributed in the hope that it will be useful, but WITHOUT
    10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
    11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
    12  * version 2 for more details (a copy is included in the LICENSE file that
    13  * accompanied this code).
    14  *
    15  * You should have received a copy of the GNU General Public License version
    16  * 2 along with this work; if not, write to the Free Software Foundation,
    17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
    18  *
    19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
    20  * or visit www.oracle.com if you need additional information or have any
    21  * questions.
    22  *
    23  */
    25 #include "precompiled.hpp"
    26 #include "classfile/javaClasses.hpp"
    27 #include "classfile/systemDictionary.hpp"
    28 #include "gc_implementation/shared/markSweep.inline.hpp"
    29 #include "gc_interface/collectedHeap.inline.hpp"
    30 #include "memory/genOopClosures.inline.hpp"
    31 #include "memory/oopFactory.hpp"
    32 #include "memory/permGen.hpp"
    33 #include "oops/instanceKlass.hpp"
    34 #include "oops/instanceMirrorKlass.hpp"
    35 #include "oops/instanceOop.hpp"
    36 #include "oops/oop.inline.hpp"
    37 #include "oops/symbol.hpp"
    38 #include "runtime/handles.inline.hpp"
    39 #ifndef SERIALGC
    40 #include "gc_implementation/g1/g1CollectedHeap.inline.hpp"
    41 #include "gc_implementation/g1/g1OopClosures.inline.hpp"
    42 #include "gc_implementation/g1/g1RemSet.inline.hpp"
    43 #include "gc_implementation/g1/heapRegionSeq.inline.hpp"
    44 #include "gc_implementation/parNew/parOopClosures.inline.hpp"
    45 #include "gc_implementation/parallelScavenge/psPromotionManager.inline.hpp"
    46 #include "gc_implementation/parallelScavenge/psScavenge.inline.hpp"
    47 #include "oops/oop.pcgc.inline.hpp"
    48 #endif
    50 int instanceMirrorKlass::_offset_of_static_fields = 0;
    52 #ifdef ASSERT
    53 template <class T> void assert_is_in(T *p) {
    54   T heap_oop = oopDesc::load_heap_oop(p);
    55   if (!oopDesc::is_null(heap_oop)) {
    56     oop o = oopDesc::decode_heap_oop_not_null(heap_oop);
    57     assert(Universe::heap()->is_in(o), "should be in heap");
    58   }
    59 }
    60 template <class T> void assert_is_in_closed_subset(T *p) {
    61   T heap_oop = oopDesc::load_heap_oop(p);
    62   if (!oopDesc::is_null(heap_oop)) {
    63     oop o = oopDesc::decode_heap_oop_not_null(heap_oop);
    64     assert(Universe::heap()->is_in_closed_subset(o), "should be in closed");
    65   }
    66 }
    67 template <class T> void assert_is_in_reserved(T *p) {
    68   T heap_oop = oopDesc::load_heap_oop(p);
    69   if (!oopDesc::is_null(heap_oop)) {
    70     oop o = oopDesc::decode_heap_oop_not_null(heap_oop);
    71     assert(Universe::heap()->is_in_reserved(o), "should be in reserved");
    72   }
    73 }
    74 template <class T> void assert_nothing(T *p) {}
    76 #else
    77 template <class T> void assert_is_in(T *p) {}
    78 template <class T> void assert_is_in_closed_subset(T *p) {}
    79 template <class T> void assert_is_in_reserved(T *p) {}
    80 template <class T> void assert_nothing(T *p) {}
    81 #endif // ASSERT
    83 #define InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE( \
    84   T, start_p, count, do_oop,                         \
    85   assert_fn)                                         \
    86 {                                                    \
    87   T* p         = (T*)(start_p);                      \
    88   T* const end = p + (count);                        \
    89   while (p < end) {                                  \
    90     (assert_fn)(p);                                  \
    91     do_oop;                                          \
    92     ++p;                                             \
    93   }                                                  \
    94 }
    96 #define InstanceMirrorKlass_SPECIALIZED_BOUNDED_OOP_ITERATE( \
    97   T, start_p, count, low, high,                              \
    98   do_oop, assert_fn)                                         \
    99 {                                                            \
   100   T* const l = (T*)(low);                                    \
   101   T* const h = (T*)(high);                                   \
   102   assert(mask_bits((intptr_t)l, sizeof(T)-1) == 0 &&         \
   103          mask_bits((intptr_t)h, sizeof(T)-1) == 0,           \
   104          "bounded region must be properly aligned");         \
   105   T* p       = (T*)(start_p);                                \
   106   T* end     = p + (count);                                  \
   107   if (p < l) p = l;                                          \
   108   if (end > h) end = h;                                      \
   109   while (p < end) {                                          \
   110     (assert_fn)(p);                                          \
   111     do_oop;                                                  \
   112     ++p;                                                     \
   113   }                                                          \
   114 }
   117 #define InstanceMirrorKlass_OOP_ITERATE(start_p, count,    \
   118                                   do_oop, assert_fn)       \
   119 {                                                          \
   120   if (UseCompressedOops) {                                 \
   121     InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE(narrowOop, \
   122       start_p, count,                                      \
   123       do_oop, assert_fn)                                   \
   124   } else {                                                 \
   125     InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE(oop,       \
   126       start_p, count,                                      \
   127       do_oop, assert_fn)                                   \
   128   }                                                        \
   129 }
   131 // The following macros call specialized macros, passing either oop or
   132 // narrowOop as the specialization type.  These test the UseCompressedOops
   133 // flag.
   134 #define InstanceMirrorKlass_BOUNDED_OOP_ITERATE(start_p, count, low, high, \
   135                                           do_oop, assert_fn)               \
   136 {                                                                          \
   137   if (UseCompressedOops) {                                                 \
   138     InstanceMirrorKlass_SPECIALIZED_BOUNDED_OOP_ITERATE(narrowOop,         \
   139       start_p, count,                                                      \
   140       low, high,                                                           \
   141       do_oop, assert_fn)                                                   \
   142   } else {                                                                 \
   143     InstanceMirrorKlass_SPECIALIZED_BOUNDED_OOP_ITERATE(oop,               \
   144       start_p, count,                                                      \
   145       low, high,                                                           \
   146       do_oop, assert_fn)                                                   \
   147   }                                                                        \
   148 }
   151 void instanceMirrorKlass::oop_follow_contents(oop obj) {
   152   instanceKlass::oop_follow_contents(obj);
   153   InstanceMirrorKlass_OOP_ITERATE(                                                    \
   154     start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),        \
   155     MarkSweep::mark_and_push(p),                                                      \
   156     assert_is_in_closed_subset)
   157 }
   159 #ifndef SERIALGC
   160 void instanceMirrorKlass::oop_follow_contents(ParCompactionManager* cm,
   161                                               oop obj) {
   162   instanceKlass::oop_follow_contents(cm, obj);
   163   InstanceMirrorKlass_OOP_ITERATE(                                                    \
   164     start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),        \
   165     PSParallelCompact::mark_and_push(cm, p),                                          \
   166     assert_is_in)
   167 }
   168 #endif // SERIALGC
   170 int instanceMirrorKlass::oop_adjust_pointers(oop obj) {
   171   int size = oop_size(obj);
   172   instanceKlass::oop_adjust_pointers(obj);
   173   InstanceMirrorKlass_OOP_ITERATE(                                                    \
   174     start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),        \
   175     MarkSweep::adjust_pointer(p),                                                     \
   176     assert_nothing)
   177   return size;
   178 }
   180 #define InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(T, nv_suffix)                \
   181   InstanceMirrorKlass_OOP_ITERATE(                                                    \
   182     start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),        \
   183       (closure)->do_oop##nv_suffix(p),                                                \
   184     assert_is_in_closed_subset)                                                       \
   185   return oop_size(obj);                                                               \
   187 #define InstanceMirrorKlass_BOUNDED_SPECIALIZED_OOP_ITERATE(T, nv_suffix, mr)         \
   188   InstanceMirrorKlass_BOUNDED_OOP_ITERATE(                                            \
   189     start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),        \
   190     mr.start(), mr.end(),                                                             \
   191       (closure)->do_oop##nv_suffix(p),                                                \
   192     assert_is_in_closed_subset)                                                       \
   193   return oop_size(obj);                                                               \
   196 // Macro to define instanceMirrorKlass::oop_oop_iterate for virtual/nonvirtual for
   197 // all closures.  Macros calling macros above for each oop size.
   199 #define InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN(OopClosureType, nv_suffix)           \
   200                                                                                       \
   201 int instanceMirrorKlass::                                                             \
   202 oop_oop_iterate##nv_suffix(oop obj, OopClosureType* closure) {                        \
   203   /* Get size before changing pointers */                                             \
   204   SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk);      \
   205                                                                                       \
   206   instanceKlass::oop_oop_iterate##nv_suffix(obj, closure);                            \
   207                                                                                       \
   208   if (UseCompressedOops) {                                                            \
   209     InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(narrowOop, nv_suffix);           \
   210   } else {                                                                            \
   211     InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(oop, nv_suffix);                 \
   212   }                                                                                   \
   213 }
   215 #ifndef SERIALGC
   216 #define InstanceMirrorKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN(OopClosureType, nv_suffix) \
   217                                                                                       \
   218 int instanceMirrorKlass::                                                             \
   219 oop_oop_iterate_backwards##nv_suffix(oop obj, OopClosureType* closure) {              \
   220   /* Get size before changing pointers */                                             \
   221   SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk);      \
   222                                                                                       \
   223   instanceKlass::oop_oop_iterate_backwards##nv_suffix(obj, closure);                  \
   224                                                                                       \
   225   if (UseCompressedOops) {                                                            \
   226     InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(narrowOop, nv_suffix);           \
   227   } else {                                                                            \
   228     InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(oop, nv_suffix);                 \
   229   }                                                                                   \
   230 }
   231 #endif // !SERIALGC
   234 #define InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN_m(OopClosureType, nv_suffix)         \
   235                                                                                       \
   236 int instanceMirrorKlass::                                                             \
   237 oop_oop_iterate##nv_suffix##_m(oop obj,                                               \
   238                                OopClosureType* closure,                               \
   239                                MemRegion mr) {                                        \
   240   SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk);      \
   241                                                                                       \
   242   instanceKlass::oop_oop_iterate##nv_suffix##_m(obj, closure, mr);                    \
   243   if (UseCompressedOops) {                                                            \
   244     InstanceMirrorKlass_BOUNDED_SPECIALIZED_OOP_ITERATE(narrowOop, nv_suffix, mr);    \
   245   } else {                                                                            \
   246     InstanceMirrorKlass_BOUNDED_SPECIALIZED_OOP_ITERATE(oop, nv_suffix, mr);          \
   247   }                                                                                   \
   248 }
   250 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN)
   251 ALL_OOP_OOP_ITERATE_CLOSURES_2(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN)
   252 #ifndef SERIALGC
   253 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceMirrorKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN)
   254 ALL_OOP_OOP_ITERATE_CLOSURES_2(InstanceMirrorKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN)
   255 #endif // SERIALGC
   256 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN_m)
   257 ALL_OOP_OOP_ITERATE_CLOSURES_2(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN_m)
   259 #ifndef SERIALGC
   260 void instanceMirrorKlass::oop_push_contents(PSPromotionManager* pm, oop obj) {
   261   instanceKlass::oop_push_contents(pm, obj);
   262   InstanceMirrorKlass_OOP_ITERATE(                                            \
   263     start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),\
   264     if (PSScavenge::should_scavenge(p)) {                                     \
   265       pm->claim_or_forward_depth(p);                                          \
   266     },                                                                        \
   267     assert_nothing )
   268 }
   270 int instanceMirrorKlass::oop_update_pointers(ParCompactionManager* cm, oop obj) {
   271   instanceKlass::oop_update_pointers(cm, obj);
   272   InstanceMirrorKlass_OOP_ITERATE(                                            \
   273     start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),\
   274     PSParallelCompact::adjust_pointer(p),                                     \
   275     assert_nothing)
   276   return oop_size(obj);
   277 }
   278 #endif // SERIALGC
   280 int instanceMirrorKlass::instance_size(KlassHandle k) {
   281   if (k() != NULL && k->oop_is_instance()) {
   282     return align_object_size(size_helper() + instanceKlass::cast(k())->static_field_size());
   283   }
   284   return size_helper();
   285 }
   287 instanceOop instanceMirrorKlass::allocate_instance(KlassHandle k, TRAPS) {
   288   // Query before forming handle.
   289   int size = instance_size(k);
   290   KlassHandle h_k(THREAD, as_klassOop());
   291   instanceOop i;
   293   if (JavaObjectsInPerm) {
   294     i = (instanceOop) CollectedHeap::permanent_obj_allocate(h_k, size, CHECK_NULL);
   295   } else {
   296     assert(ScavengeRootsInCode > 0, "must be");
   297     i = (instanceOop) CollectedHeap::obj_allocate(h_k, size, CHECK_NULL);
   298   }
   300   return i;
   301 }
   303 int instanceMirrorKlass::oop_size(oop obj) const {
   304   return java_lang_Class::oop_size(obj);
   305 }
   307 int instanceMirrorKlass::compute_static_oop_field_count(oop obj) {
   308   klassOop k = java_lang_Class::as_klassOop(obj);
   309   if (k != NULL && k->klass_part()->oop_is_instance()) {
   310     return instanceKlass::cast(k)->static_oop_field_count();
   311   }
   312   return 0;
   313 }

mercurial