Thu, 07 Apr 2011 09:53:20 -0700
7009266: G1: assert(obj->is_oop_or_null(true )) failed: Error
Summary: A referent object that is only weakly reachable at the start of concurrent marking but is re-attached to the strongly reachable object graph during marking may not be marked as live. This can cause the reference object to be processed prematurely and leave dangling pointers to the referent object. Implement a read barrier for the java.lang.ref.Reference::referent field by intrinsifying the Reference.get() method, and intercepting accesses though JNI, reflection, and Unsafe, so that when a non-null referent object is read it is also logged in an SATB buffer.
Reviewed-by: kvn, iveresov, never, tonyp, dholmes
1 /*
2 * Copyright (c) 2011, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
25 #include "precompiled.hpp"
26 #include "classfile/javaClasses.hpp"
27 #include "classfile/systemDictionary.hpp"
28 #include "gc_implementation/shared/markSweep.inline.hpp"
29 #include "gc_interface/collectedHeap.inline.hpp"
30 #include "memory/genOopClosures.inline.hpp"
31 #include "memory/oopFactory.hpp"
32 #include "memory/permGen.hpp"
33 #include "oops/instanceKlass.hpp"
34 #include "oops/instanceMirrorKlass.hpp"
35 #include "oops/instanceOop.hpp"
36 #include "oops/oop.inline.hpp"
37 #include "oops/symbol.hpp"
38 #include "runtime/handles.inline.hpp"
39 #ifndef SERIALGC
40 #include "gc_implementation/g1/g1CollectedHeap.inline.hpp"
41 #include "gc_implementation/g1/g1OopClosures.inline.hpp"
42 #include "gc_implementation/g1/g1RemSet.inline.hpp"
43 #include "gc_implementation/g1/heapRegionSeq.inline.hpp"
44 #include "gc_implementation/parNew/parOopClosures.inline.hpp"
45 #include "gc_implementation/parallelScavenge/psPromotionManager.inline.hpp"
46 #include "gc_implementation/parallelScavenge/psScavenge.inline.hpp"
47 #include "oops/oop.pcgc.inline.hpp"
48 #endif
50 int instanceMirrorKlass::_offset_of_static_fields = 0;
52 #ifdef ASSERT
53 template <class T> void assert_is_in(T *p) {
54 T heap_oop = oopDesc::load_heap_oop(p);
55 if (!oopDesc::is_null(heap_oop)) {
56 oop o = oopDesc::decode_heap_oop_not_null(heap_oop);
57 assert(Universe::heap()->is_in(o), "should be in heap");
58 }
59 }
60 template <class T> void assert_is_in_closed_subset(T *p) {
61 T heap_oop = oopDesc::load_heap_oop(p);
62 if (!oopDesc::is_null(heap_oop)) {
63 oop o = oopDesc::decode_heap_oop_not_null(heap_oop);
64 assert(Universe::heap()->is_in_closed_subset(o), "should be in closed");
65 }
66 }
67 template <class T> void assert_is_in_reserved(T *p) {
68 T heap_oop = oopDesc::load_heap_oop(p);
69 if (!oopDesc::is_null(heap_oop)) {
70 oop o = oopDesc::decode_heap_oop_not_null(heap_oop);
71 assert(Universe::heap()->is_in_reserved(o), "should be in reserved");
72 }
73 }
74 template <class T> void assert_nothing(T *p) {}
76 #else
77 template <class T> void assert_is_in(T *p) {}
78 template <class T> void assert_is_in_closed_subset(T *p) {}
79 template <class T> void assert_is_in_reserved(T *p) {}
80 template <class T> void assert_nothing(T *p) {}
81 #endif // ASSERT
83 #define InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE( \
84 T, start_p, count, do_oop, \
85 assert_fn) \
86 { \
87 T* p = (T*)(start_p); \
88 T* const end = p + (count); \
89 while (p < end) { \
90 (assert_fn)(p); \
91 do_oop; \
92 ++p; \
93 } \
94 }
96 #define InstanceMirrorKlass_SPECIALIZED_BOUNDED_OOP_ITERATE( \
97 T, start_p, count, low, high, \
98 do_oop, assert_fn) \
99 { \
100 T* const l = (T*)(low); \
101 T* const h = (T*)(high); \
102 assert(mask_bits((intptr_t)l, sizeof(T)-1) == 0 && \
103 mask_bits((intptr_t)h, sizeof(T)-1) == 0, \
104 "bounded region must be properly aligned"); \
105 T* p = (T*)(start_p); \
106 T* end = p + (count); \
107 if (p < l) p = l; \
108 if (end > h) end = h; \
109 while (p < end) { \
110 (assert_fn)(p); \
111 do_oop; \
112 ++p; \
113 } \
114 }
117 #define InstanceMirrorKlass_OOP_ITERATE(start_p, count, \
118 do_oop, assert_fn) \
119 { \
120 if (UseCompressedOops) { \
121 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE(narrowOop, \
122 start_p, count, \
123 do_oop, assert_fn) \
124 } else { \
125 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE(oop, \
126 start_p, count, \
127 do_oop, assert_fn) \
128 } \
129 }
131 // The following macros call specialized macros, passing either oop or
132 // narrowOop as the specialization type. These test the UseCompressedOops
133 // flag.
134 #define InstanceMirrorKlass_BOUNDED_OOP_ITERATE(start_p, count, low, high, \
135 do_oop, assert_fn) \
136 { \
137 if (UseCompressedOops) { \
138 InstanceMirrorKlass_SPECIALIZED_BOUNDED_OOP_ITERATE(narrowOop, \
139 start_p, count, \
140 low, high, \
141 do_oop, assert_fn) \
142 } else { \
143 InstanceMirrorKlass_SPECIALIZED_BOUNDED_OOP_ITERATE(oop, \
144 start_p, count, \
145 low, high, \
146 do_oop, assert_fn) \
147 } \
148 }
151 void instanceMirrorKlass::oop_follow_contents(oop obj) {
152 instanceKlass::oop_follow_contents(obj);
153 InstanceMirrorKlass_OOP_ITERATE( \
154 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \
155 MarkSweep::mark_and_push(p), \
156 assert_is_in_closed_subset)
157 }
159 #ifndef SERIALGC
160 void instanceMirrorKlass::oop_follow_contents(ParCompactionManager* cm,
161 oop obj) {
162 instanceKlass::oop_follow_contents(cm, obj);
163 InstanceMirrorKlass_OOP_ITERATE( \
164 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \
165 PSParallelCompact::mark_and_push(cm, p), \
166 assert_is_in)
167 }
168 #endif // SERIALGC
170 int instanceMirrorKlass::oop_adjust_pointers(oop obj) {
171 int size = oop_size(obj);
172 instanceKlass::oop_adjust_pointers(obj);
173 InstanceMirrorKlass_OOP_ITERATE( \
174 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \
175 MarkSweep::adjust_pointer(p), \
176 assert_nothing)
177 return size;
178 }
180 #define InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(T, nv_suffix) \
181 InstanceMirrorKlass_OOP_ITERATE( \
182 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \
183 (closure)->do_oop##nv_suffix(p), \
184 assert_is_in_closed_subset) \
185 return oop_size(obj); \
187 #define InstanceMirrorKlass_BOUNDED_SPECIALIZED_OOP_ITERATE(T, nv_suffix, mr) \
188 InstanceMirrorKlass_BOUNDED_OOP_ITERATE( \
189 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj), \
190 mr.start(), mr.end(), \
191 (closure)->do_oop##nv_suffix(p), \
192 assert_is_in_closed_subset) \
193 return oop_size(obj); \
196 // Macro to define instanceMirrorKlass::oop_oop_iterate for virtual/nonvirtual for
197 // all closures. Macros calling macros above for each oop size.
199 #define InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN(OopClosureType, nv_suffix) \
200 \
201 int instanceMirrorKlass:: \
202 oop_oop_iterate##nv_suffix(oop obj, OopClosureType* closure) { \
203 /* Get size before changing pointers */ \
204 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk); \
205 \
206 instanceKlass::oop_oop_iterate##nv_suffix(obj, closure); \
207 \
208 if (UseCompressedOops) { \
209 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(narrowOop, nv_suffix); \
210 } else { \
211 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(oop, nv_suffix); \
212 } \
213 }
215 #ifndef SERIALGC
216 #define InstanceMirrorKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN(OopClosureType, nv_suffix) \
217 \
218 int instanceMirrorKlass:: \
219 oop_oop_iterate_backwards##nv_suffix(oop obj, OopClosureType* closure) { \
220 /* Get size before changing pointers */ \
221 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk); \
222 \
223 instanceKlass::oop_oop_iterate_backwards##nv_suffix(obj, closure); \
224 \
225 if (UseCompressedOops) { \
226 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(narrowOop, nv_suffix); \
227 } else { \
228 InstanceMirrorKlass_SPECIALIZED_OOP_ITERATE_DEFN(oop, nv_suffix); \
229 } \
230 }
231 #endif // !SERIALGC
234 #define InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN_m(OopClosureType, nv_suffix) \
235 \
236 int instanceMirrorKlass:: \
237 oop_oop_iterate##nv_suffix##_m(oop obj, \
238 OopClosureType* closure, \
239 MemRegion mr) { \
240 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk); \
241 \
242 instanceKlass::oop_oop_iterate##nv_suffix##_m(obj, closure, mr); \
243 if (UseCompressedOops) { \
244 InstanceMirrorKlass_BOUNDED_SPECIALIZED_OOP_ITERATE(narrowOop, nv_suffix, mr); \
245 } else { \
246 InstanceMirrorKlass_BOUNDED_SPECIALIZED_OOP_ITERATE(oop, nv_suffix, mr); \
247 } \
248 }
250 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN)
251 ALL_OOP_OOP_ITERATE_CLOSURES_2(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN)
252 #ifndef SERIALGC
253 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceMirrorKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN)
254 ALL_OOP_OOP_ITERATE_CLOSURES_2(InstanceMirrorKlass_OOP_OOP_ITERATE_BACKWARDS_DEFN)
255 #endif // SERIALGC
256 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN_m)
257 ALL_OOP_OOP_ITERATE_CLOSURES_2(InstanceMirrorKlass_OOP_OOP_ITERATE_DEFN_m)
259 #ifndef SERIALGC
260 void instanceMirrorKlass::oop_push_contents(PSPromotionManager* pm, oop obj) {
261 instanceKlass::oop_push_contents(pm, obj);
262 InstanceMirrorKlass_OOP_ITERATE( \
263 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),\
264 if (PSScavenge::should_scavenge(p)) { \
265 pm->claim_or_forward_depth(p); \
266 }, \
267 assert_nothing )
268 }
270 int instanceMirrorKlass::oop_update_pointers(ParCompactionManager* cm, oop obj) {
271 instanceKlass::oop_update_pointers(cm, obj);
272 InstanceMirrorKlass_OOP_ITERATE( \
273 start_of_static_fields(obj), java_lang_Class::static_oop_field_count(obj),\
274 PSParallelCompact::adjust_pointer(p), \
275 assert_nothing)
276 return oop_size(obj);
277 }
278 #endif // SERIALGC
280 int instanceMirrorKlass::instance_size(KlassHandle k) {
281 if (k() != NULL && k->oop_is_instance()) {
282 return align_object_size(size_helper() + instanceKlass::cast(k())->static_field_size());
283 }
284 return size_helper();
285 }
287 instanceOop instanceMirrorKlass::allocate_instance(KlassHandle k, TRAPS) {
288 // Query before forming handle.
289 int size = instance_size(k);
290 KlassHandle h_k(THREAD, as_klassOop());
291 instanceOop i;
293 if (JavaObjectsInPerm) {
294 i = (instanceOop) CollectedHeap::permanent_obj_allocate(h_k, size, CHECK_NULL);
295 } else {
296 assert(ScavengeRootsInCode > 0, "must be");
297 i = (instanceOop) CollectedHeap::obj_allocate(h_k, size, CHECK_NULL);
298 }
300 return i;
301 }
303 int instanceMirrorKlass::oop_size(oop obj) const {
304 return java_lang_Class::oop_size(obj);
305 }
307 int instanceMirrorKlass::compute_static_oop_field_count(oop obj) {
308 klassOop k = java_lang_Class::as_klassOop(obj);
309 if (k != NULL && k->klass_part()->oop_is_instance()) {
310 return instanceKlass::cast(k)->static_oop_field_count();
311 }
312 return 0;
313 }