Wed, 02 Jul 2008 12:55:16 -0700
6719955: Update copyright year
Summary: Update copyright year for files that have been modified in 2008
Reviewed-by: ohair, tbell
1 /*
2 * Copyright 1997-2008 Sun Microsystems, Inc. All Rights Reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
20 * CA 95054 USA or visit www.sun.com if you need additional information or
21 * have any questions.
22 *
23 */
25 # include "incls/_precompiled.incl"
26 # include "incls/_instanceRefKlass.cpp.incl"
28 template <class T>
29 static void specialized_oop_follow_contents(instanceRefKlass* ref, oop obj) {
30 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
31 oop referent = oopDesc::load_decode_heap_oop(referent_addr);
32 debug_only(
33 if(TraceReferenceGC && PrintGCDetails) {
34 gclog_or_tty->print_cr("instanceRefKlass::oop_follow_contents " INTPTR_FORMAT, obj);
35 }
36 )
37 if (referent != NULL) {
38 if (!referent->is_gc_marked() &&
39 MarkSweep::ref_processor()->
40 discover_reference(obj, ref->reference_type())) {
41 // reference already enqueued, referent will be traversed later
42 ref->instanceKlass::oop_follow_contents(obj);
43 debug_only(
44 if(TraceReferenceGC && PrintGCDetails) {
45 gclog_or_tty->print_cr(" Non NULL enqueued " INTPTR_FORMAT, obj);
46 }
47 )
48 return;
49 } else {
50 // treat referent as normal oop
51 debug_only(
52 if(TraceReferenceGC && PrintGCDetails) {
53 gclog_or_tty->print_cr(" Non NULL normal " INTPTR_FORMAT, obj);
54 }
55 )
56 MarkSweep::mark_and_push(referent_addr);
57 }
58 }
59 // treat next as normal oop. next is a link in the pending list.
60 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
61 debug_only(
62 if(TraceReferenceGC && PrintGCDetails) {
63 gclog_or_tty->print_cr(" Process next as normal " INTPTR_FORMAT, next_addr);
64 }
65 )
66 MarkSweep::mark_and_push(next_addr);
67 ref->instanceKlass::oop_follow_contents(obj);
68 }
70 void instanceRefKlass::oop_follow_contents(oop obj) {
71 if (UseCompressedOops) {
72 specialized_oop_follow_contents<narrowOop>(this, obj);
73 } else {
74 specialized_oop_follow_contents<oop>(this, obj);
75 }
76 }
78 #ifndef SERIALGC
79 template <class T>
80 static void specialized_oop_follow_contents(instanceRefKlass* ref,
81 ParCompactionManager* cm,
82 oop obj) {
83 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
84 oop referent = oopDesc::load_decode_heap_oop(referent_addr);
85 debug_only(
86 if(TraceReferenceGC && PrintGCDetails) {
87 gclog_or_tty->print_cr("instanceRefKlass::oop_follow_contents " INTPTR_FORMAT, obj);
88 }
89 )
90 if (referent != NULL) {
91 if (PSParallelCompact::mark_bitmap()->is_unmarked(referent) &&
92 PSParallelCompact::ref_processor()->
93 discover_reference(obj, ref->reference_type())) {
94 // reference already enqueued, referent will be traversed later
95 ref->instanceKlass::oop_follow_contents(cm, obj);
96 debug_only(
97 if(TraceReferenceGC && PrintGCDetails) {
98 gclog_or_tty->print_cr(" Non NULL enqueued " INTPTR_FORMAT, obj);
99 }
100 )
101 return;
102 } else {
103 // treat referent as normal oop
104 debug_only(
105 if(TraceReferenceGC && PrintGCDetails) {
106 gclog_or_tty->print_cr(" Non NULL normal " INTPTR_FORMAT, obj);
107 }
108 )
109 PSParallelCompact::mark_and_push(cm, referent_addr);
110 }
111 }
112 // treat next as normal oop. next is a link in the pending list.
113 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
114 debug_only(
115 if(TraceReferenceGC && PrintGCDetails) {
116 gclog_or_tty->print_cr(" Process next as normal " INTPTR_FORMAT, next_addr);
117 }
118 )
119 PSParallelCompact::mark_and_push(cm, next_addr);
120 ref->instanceKlass::oop_follow_contents(cm, obj);
121 }
123 void instanceRefKlass::oop_follow_contents(ParCompactionManager* cm,
124 oop obj) {
125 if (UseCompressedOops) {
126 specialized_oop_follow_contents<narrowOop>(this, cm, obj);
127 } else {
128 specialized_oop_follow_contents<oop>(this, cm, obj);
129 }
130 }
131 #endif // SERIALGC
133 #ifdef ASSERT
134 template <class T> void trace_reference_gc(const char *s, oop obj,
135 T* referent_addr,
136 T* next_addr,
137 T* discovered_addr) {
138 if(TraceReferenceGC && PrintGCDetails) {
139 gclog_or_tty->print_cr("%s obj " INTPTR_FORMAT, s, (address)obj);
140 gclog_or_tty->print_cr(" referent_addr/* " INTPTR_FORMAT " / "
141 INTPTR_FORMAT, referent_addr,
142 referent_addr ?
143 (address)oopDesc::load_decode_heap_oop(referent_addr) : NULL);
144 gclog_or_tty->print_cr(" next_addr/* " INTPTR_FORMAT " / "
145 INTPTR_FORMAT, next_addr,
146 next_addr ? (address)oopDesc::load_decode_heap_oop(next_addr) : NULL);
147 gclog_or_tty->print_cr(" discovered_addr/* " INTPTR_FORMAT " / "
148 INTPTR_FORMAT, discovered_addr,
149 discovered_addr ?
150 (address)oopDesc::load_decode_heap_oop(discovered_addr) : NULL);
151 }
152 }
153 #endif
155 template <class T> void specialized_oop_adjust_pointers(instanceRefKlass *ref, oop obj) {
156 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
157 MarkSweep::adjust_pointer(referent_addr);
158 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
159 MarkSweep::adjust_pointer(next_addr);
160 T* discovered_addr = (T*)java_lang_ref_Reference::discovered_addr(obj);
161 MarkSweep::adjust_pointer(discovered_addr);
162 debug_only(trace_reference_gc("instanceRefKlass::oop_adjust_pointers", obj,
163 referent_addr, next_addr, discovered_addr);)
164 }
166 int instanceRefKlass::oop_adjust_pointers(oop obj) {
167 int size = size_helper();
168 instanceKlass::oop_adjust_pointers(obj);
170 if (UseCompressedOops) {
171 specialized_oop_adjust_pointers<narrowOop>(this, obj);
172 } else {
173 specialized_oop_adjust_pointers<oop>(this, obj);
174 }
175 return size;
176 }
178 #define InstanceRefKlass_SPECIALIZED_OOP_ITERATE(T, nv_suffix, contains) \
179 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj); \
180 oop referent = oopDesc::load_decode_heap_oop(referent_addr); \
181 if (referent != NULL && contains(referent_addr)) { \
182 ReferenceProcessor* rp = closure->_ref_processor; \
183 if (!referent->is_gc_marked() && (rp != NULL) && \
184 rp->discover_reference(obj, reference_type())) { \
185 return size; \
186 } else { \
187 /* treat referent as normal oop */ \
188 SpecializationStats::record_do_oop_call##nv_suffix(SpecializationStats::irk);\
189 closure->do_oop##nv_suffix(referent_addr); \
190 } \
191 } \
192 /* treat next as normal oop */ \
193 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj); \
194 if (contains(next_addr)) { \
195 SpecializationStats::record_do_oop_call##nv_suffix(SpecializationStats::irk); \
196 closure->do_oop##nv_suffix(next_addr); \
197 } \
198 return size; \
201 template <class T> bool contains(T *t) { return true; }
203 // Macro to define instanceRefKlass::oop_oop_iterate for virtual/nonvirtual for
204 // all closures. Macros calling macros above for each oop size.
206 #define InstanceRefKlass_OOP_OOP_ITERATE_DEFN(OopClosureType, nv_suffix) \
207 \
208 int instanceRefKlass:: \
209 oop_oop_iterate##nv_suffix(oop obj, OopClosureType* closure) { \
210 /* Get size before changing pointers */ \
211 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk);\
212 \
213 int size = instanceKlass::oop_oop_iterate##nv_suffix(obj, closure); \
214 \
215 if (UseCompressedOops) { \
216 InstanceRefKlass_SPECIALIZED_OOP_ITERATE(narrowOop, nv_suffix, contains); \
217 } else { \
218 InstanceRefKlass_SPECIALIZED_OOP_ITERATE(oop, nv_suffix, contains); \
219 } \
220 }
222 #define InstanceRefKlass_OOP_OOP_ITERATE_DEFN_m(OopClosureType, nv_suffix) \
223 \
224 int instanceRefKlass:: \
225 oop_oop_iterate##nv_suffix##_m(oop obj, \
226 OopClosureType* closure, \
227 MemRegion mr) { \
228 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk);\
229 \
230 int size = instanceKlass::oop_oop_iterate##nv_suffix##_m(obj, closure, mr); \
231 if (UseCompressedOops) { \
232 InstanceRefKlass_SPECIALIZED_OOP_ITERATE(narrowOop, nv_suffix, mr.contains); \
233 } else { \
234 InstanceRefKlass_SPECIALIZED_OOP_ITERATE(oop, nv_suffix, mr.contains); \
235 } \
236 }
238 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceRefKlass_OOP_OOP_ITERATE_DEFN)
239 ALL_OOP_OOP_ITERATE_CLOSURES_3(InstanceRefKlass_OOP_OOP_ITERATE_DEFN)
240 ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceRefKlass_OOP_OOP_ITERATE_DEFN_m)
241 ALL_OOP_OOP_ITERATE_CLOSURES_3(InstanceRefKlass_OOP_OOP_ITERATE_DEFN_m)
243 #ifndef SERIALGC
244 template <class T>
245 void specialized_oop_copy_contents(instanceRefKlass *ref,
246 PSPromotionManager* pm, oop obj) {
247 assert(!pm->depth_first(), "invariant");
248 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
249 if (PSScavenge::should_scavenge(referent_addr)) {
250 ReferenceProcessor* rp = PSScavenge::reference_processor();
251 if (rp->discover_reference(obj, ref->reference_type())) {
252 // reference already enqueued, referent and next will be traversed later
253 ref->instanceKlass::oop_copy_contents(pm, obj);
254 return;
255 } else {
256 // treat referent as normal oop
257 pm->claim_or_forward_breadth(referent_addr);
258 }
259 }
260 // treat next as normal oop
261 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
262 if (PSScavenge::should_scavenge(next_addr)) {
263 pm->claim_or_forward_breadth(next_addr);
264 }
265 ref->instanceKlass::oop_copy_contents(pm, obj);
266 }
268 void instanceRefKlass::oop_copy_contents(PSPromotionManager* pm, oop obj) {
269 if (UseCompressedOops) {
270 specialized_oop_copy_contents<narrowOop>(this, pm, obj);
271 } else {
272 specialized_oop_copy_contents<oop>(this, pm, obj);
273 }
274 }
276 template <class T>
277 void specialized_oop_push_contents(instanceRefKlass *ref,
278 PSPromotionManager* pm, oop obj) {
279 assert(pm->depth_first(), "invariant");
280 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
281 if (PSScavenge::should_scavenge(referent_addr)) {
282 ReferenceProcessor* rp = PSScavenge::reference_processor();
283 if (rp->discover_reference(obj, ref->reference_type())) {
284 // reference already enqueued, referent and next will be traversed later
285 ref->instanceKlass::oop_push_contents(pm, obj);
286 return;
287 } else {
288 // treat referent as normal oop
289 pm->claim_or_forward_depth(referent_addr);
290 }
291 }
292 // treat next as normal oop
293 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
294 if (PSScavenge::should_scavenge(next_addr)) {
295 pm->claim_or_forward_depth(next_addr);
296 }
297 ref->instanceKlass::oop_push_contents(pm, obj);
298 }
300 void instanceRefKlass::oop_push_contents(PSPromotionManager* pm, oop obj) {
301 if (UseCompressedOops) {
302 specialized_oop_push_contents<narrowOop>(this, pm, obj);
303 } else {
304 specialized_oop_push_contents<oop>(this, pm, obj);
305 }
306 }
308 template <class T>
309 void specialized_oop_update_pointers(instanceRefKlass *ref,
310 ParCompactionManager* cm, oop obj) {
311 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
312 PSParallelCompact::adjust_pointer(referent_addr);
313 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
314 PSParallelCompact::adjust_pointer(next_addr);
315 T* discovered_addr = (T*)java_lang_ref_Reference::discovered_addr(obj);
316 PSParallelCompact::adjust_pointer(discovered_addr);
317 debug_only(trace_reference_gc("instanceRefKlass::oop_update_ptrs", obj,
318 referent_addr, next_addr, discovered_addr);)
319 }
321 int instanceRefKlass::oop_update_pointers(ParCompactionManager* cm, oop obj) {
322 instanceKlass::oop_update_pointers(cm, obj);
323 if (UseCompressedOops) {
324 specialized_oop_update_pointers<narrowOop>(this, cm, obj);
325 } else {
326 specialized_oop_update_pointers<oop>(this, cm, obj);
327 }
328 return size_helper();
329 }
332 template <class T> void
333 specialized_oop_update_pointers(ParCompactionManager* cm, oop obj,
334 HeapWord* beg_addr, HeapWord* end_addr) {
335 T* p;
336 T* referent_addr = p = (T*)java_lang_ref_Reference::referent_addr(obj);
337 PSParallelCompact::adjust_pointer(p, beg_addr, end_addr);
338 T* next_addr = p = (T*)java_lang_ref_Reference::next_addr(obj);
339 PSParallelCompact::adjust_pointer(p, beg_addr, end_addr);
340 T* discovered_addr = p = (T*)java_lang_ref_Reference::discovered_addr(obj);
341 PSParallelCompact::adjust_pointer(p, beg_addr, end_addr);
342 debug_only(trace_reference_gc("instanceRefKlass::oop_update_ptrs", obj,
343 referent_addr, next_addr, discovered_addr);)
344 }
346 int
347 instanceRefKlass::oop_update_pointers(ParCompactionManager* cm, oop obj,
348 HeapWord* beg_addr, HeapWord* end_addr) {
349 instanceKlass::oop_update_pointers(cm, obj, beg_addr, end_addr);
350 if (UseCompressedOops) {
351 specialized_oop_update_pointers<narrowOop>(cm, obj, beg_addr, end_addr);
352 } else {
353 specialized_oop_update_pointers<oop>(cm, obj, beg_addr, end_addr);
354 }
355 return size_helper();
356 }
357 #endif // SERIALGC
359 void instanceRefKlass::update_nonstatic_oop_maps(klassOop k) {
360 // Clear the nonstatic oop-map entries corresponding to referent
361 // and nextPending field. They are treated specially by the
362 // garbage collector.
363 // The discovered field is used only by the garbage collector
364 // and is also treated specially.
365 instanceKlass* ik = instanceKlass::cast(k);
367 // Check that we have the right class
368 debug_only(static bool first_time = true);
369 assert(k == SystemDictionary::reference_klass() && first_time,
370 "Invalid update of maps");
371 debug_only(first_time = false);
372 assert(ik->nonstatic_oop_map_size() == 1, "just checking");
374 OopMapBlock* map = ik->start_of_nonstatic_oop_maps();
376 // Check that the current map is (2,4) - currently points at field with
377 // offset 2 (words) and has 4 map entries.
378 debug_only(int offset = java_lang_ref_Reference::referent_offset);
379 debug_only(int length = ((java_lang_ref_Reference::discovered_offset -
380 java_lang_ref_Reference::referent_offset)/heapOopSize) + 1);
382 if (UseSharedSpaces) {
383 assert(map->offset() == java_lang_ref_Reference::queue_offset &&
384 map->length() == 1, "just checking");
385 } else {
386 assert(map->offset() == offset && map->length() == length,
387 "just checking");
389 // Update map to (3,1) - point to offset of 3 (words) with 1 map entry.
390 map->set_offset(java_lang_ref_Reference::queue_offset);
391 map->set_length(1);
392 }
393 }
396 // Verification
398 void instanceRefKlass::oop_verify_on(oop obj, outputStream* st) {
399 instanceKlass::oop_verify_on(obj, st);
400 // Verify referent field
401 oop referent = java_lang_ref_Reference::referent(obj);
403 // We should make this general to all heaps
404 GenCollectedHeap* gch = NULL;
405 if (Universe::heap()->kind() == CollectedHeap::GenCollectedHeap)
406 gch = GenCollectedHeap::heap();
408 if (referent != NULL) {
409 guarantee(referent->is_oop(), "referent field heap failed");
410 if (gch != NULL && !gch->is_in_youngest(obj)) {
411 // We do a specific remembered set check here since the referent
412 // field is not part of the oop mask and therefore skipped by the
413 // regular verify code.
414 if (UseCompressedOops) {
415 narrowOop* referent_addr = (narrowOop*)java_lang_ref_Reference::referent_addr(obj);
416 obj->verify_old_oop(referent_addr, true);
417 } else {
418 oop* referent_addr = (oop*)java_lang_ref_Reference::referent_addr(obj);
419 obj->verify_old_oop(referent_addr, true);
420 }
421 }
422 }
423 // Verify next field
424 oop next = java_lang_ref_Reference::next(obj);
425 if (next != NULL) {
426 guarantee(next->is_oop(), "next field verify fa iled");
427 guarantee(next->is_instanceRef(), "next field verify failed");
428 if (gch != NULL && !gch->is_in_youngest(obj)) {
429 // We do a specific remembered set check here since the next field is
430 // not part of the oop mask and therefore skipped by the regular
431 // verify code.
432 if (UseCompressedOops) {
433 narrowOop* next_addr = (narrowOop*)java_lang_ref_Reference::next_addr(obj);
434 obj->verify_old_oop(next_addr, true);
435 } else {
436 oop* next_addr = (oop*)java_lang_ref_Reference::next_addr(obj);
437 obj->verify_old_oop(next_addr, true);
438 }
439 }
440 }
441 }
443 void instanceRefKlass::acquire_pending_list_lock(BasicLock *pending_list_basic_lock) {
444 // we may enter this with pending exception set
445 PRESERVE_EXCEPTION_MARK; // exceptions are never thrown, needed for TRAPS argument
446 Handle h_lock(THREAD, java_lang_ref_Reference::pending_list_lock());
447 ObjectSynchronizer::fast_enter(h_lock, pending_list_basic_lock, false, THREAD);
448 assert(ObjectSynchronizer::current_thread_holds_lock(
449 JavaThread::current(), h_lock),
450 "Locking should have succeeded");
451 if (HAS_PENDING_EXCEPTION) CLEAR_PENDING_EXCEPTION;
452 }
454 void instanceRefKlass::release_and_notify_pending_list_lock(
455 BasicLock *pending_list_basic_lock) {
456 // we may enter this with pending exception set
457 PRESERVE_EXCEPTION_MARK; // exceptions are never thrown, needed for TRAPS argument
458 //
459 Handle h_lock(THREAD, java_lang_ref_Reference::pending_list_lock());
460 assert(ObjectSynchronizer::current_thread_holds_lock(
461 JavaThread::current(), h_lock),
462 "Lock should be held");
463 // Notify waiters on pending lists lock if there is any reference.
464 if (java_lang_ref_Reference::pending_list() != NULL) {
465 ObjectSynchronizer::notifyall(h_lock, THREAD);
466 }
467 ObjectSynchronizer::fast_exit(h_lock(), pending_list_basic_lock, THREAD);
468 if (HAS_PENDING_EXCEPTION) CLEAR_PENDING_EXCEPTION;
469 }