Wed, 02 Nov 2011 08:04:23 +0100
7106751: G1: gc/gctests/nativeGC03 crashes VM with SIGSEGV
Summary: _cset_rs_update_cl[] was indexed with values beyond what it is set up to handle.
Reviewed-by: ysr, jmasa, johnc
1 /*
2 * Copyright (c) 2001, 2011, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
25 #ifndef SHARE_VM_GC_IMPLEMENTATION_G1_G1OOPCLOSURES_HPP
26 #define SHARE_VM_GC_IMPLEMENTATION_G1_G1OOPCLOSURES_HPP
28 class HeapRegion;
29 class G1CollectedHeap;
30 class G1RemSet;
31 class ConcurrentMark;
32 class DirtyCardToOopClosure;
33 class CMBitMap;
34 class CMMarkStack;
35 class G1ParScanThreadState;
36 class CMTask;
37 class ReferenceProcessor;
39 // A class that scans oops in a given heap region (much as OopsInGenClosure
40 // scans oops in a generation.)
41 class OopsInHeapRegionClosure: public OopsInGenClosure {
42 protected:
43 HeapRegion* _from;
44 public:
45 void set_region(HeapRegion* from) { _from = from; }
46 };
48 class G1ParClosureSuper : public OopsInHeapRegionClosure {
49 protected:
50 G1CollectedHeap* _g1;
51 G1RemSet* _g1_rem;
52 ConcurrentMark* _cm;
53 G1ParScanThreadState* _par_scan_state;
54 bool _during_initial_mark;
55 bool _mark_in_progress;
56 public:
57 G1ParClosureSuper(G1CollectedHeap* g1, G1ParScanThreadState* par_scan_state);
58 bool apply_to_weak_ref_discovered_field() { return true; }
59 };
61 class G1ParPushHeapRSClosure : public G1ParClosureSuper {
62 public:
63 G1ParPushHeapRSClosure(G1CollectedHeap* g1,
64 G1ParScanThreadState* par_scan_state):
65 G1ParClosureSuper(g1, par_scan_state) { }
67 template <class T> void do_oop_nv(T* p);
68 virtual void do_oop(oop* p) { do_oop_nv(p); }
69 virtual void do_oop(narrowOop* p) { do_oop_nv(p); }
70 };
72 class G1ParScanClosure : public G1ParClosureSuper {
73 public:
74 G1ParScanClosure(G1CollectedHeap* g1, G1ParScanThreadState* par_scan_state, ReferenceProcessor* rp) :
75 G1ParClosureSuper(g1, par_scan_state)
76 {
77 assert(_ref_processor == NULL, "sanity");
78 _ref_processor = rp;
79 }
81 template <class T> void do_oop_nv(T* p);
82 virtual void do_oop(oop* p) { do_oop_nv(p); }
83 virtual void do_oop(narrowOop* p) { do_oop_nv(p); }
84 };
86 #define G1_PARTIAL_ARRAY_MASK 0x2
88 template <class T> inline bool has_partial_array_mask(T* ref) {
89 return ((uintptr_t)ref & G1_PARTIAL_ARRAY_MASK) == G1_PARTIAL_ARRAY_MASK;
90 }
92 template <class T> inline T* set_partial_array_mask(T obj) {
93 assert(((uintptr_t)obj & G1_PARTIAL_ARRAY_MASK) == 0, "Information loss!");
94 return (T*) ((uintptr_t)obj | G1_PARTIAL_ARRAY_MASK);
95 }
97 template <class T> inline oop clear_partial_array_mask(T* ref) {
98 return oop((intptr_t)ref & ~G1_PARTIAL_ARRAY_MASK);
99 }
101 class G1ParScanPartialArrayClosure : public G1ParClosureSuper {
102 G1ParScanClosure _scanner;
104 public:
105 G1ParScanPartialArrayClosure(G1CollectedHeap* g1, G1ParScanThreadState* par_scan_state, ReferenceProcessor* rp) :
106 G1ParClosureSuper(g1, par_scan_state), _scanner(g1, par_scan_state, rp)
107 {
108 assert(_ref_processor == NULL, "sanity");
109 }
111 G1ParScanClosure* scanner() {
112 return &_scanner;
113 }
115 template <class T> void do_oop_nv(T* p);
116 virtual void do_oop(oop* p) { do_oop_nv(p); }
117 virtual void do_oop(narrowOop* p) { do_oop_nv(p); }
118 };
121 class G1ParCopyHelper : public G1ParClosureSuper {
122 G1ParScanClosure *_scanner;
123 protected:
124 template <class T> void mark_object(T* p);
125 oop copy_to_survivor_space(oop obj, bool should_mark_root,
126 bool should_mark_copy);
127 public:
128 G1ParCopyHelper(G1CollectedHeap* g1, G1ParScanThreadState* par_scan_state,
129 G1ParScanClosure *scanner) :
130 G1ParClosureSuper(g1, par_scan_state), _scanner(scanner) { }
131 };
133 template<bool do_gen_barrier, G1Barrier barrier,
134 bool do_mark_object>
135 class G1ParCopyClosure : public G1ParCopyHelper {
136 G1ParScanClosure _scanner;
138 template <class T> void do_oop_work(T* p);
140 public:
141 G1ParCopyClosure(G1CollectedHeap* g1, G1ParScanThreadState* par_scan_state,
142 ReferenceProcessor* rp) :
143 _scanner(g1, par_scan_state, rp),
144 G1ParCopyHelper(g1, par_scan_state, &_scanner)
145 {
146 assert(_ref_processor == NULL, "sanity");
147 }
149 G1ParScanClosure* scanner() { return &_scanner; }
151 template <class T> void do_oop_nv(T* p) {
152 do_oop_work(p);
153 }
154 virtual void do_oop(oop* p) { do_oop_nv(p); }
155 virtual void do_oop(narrowOop* p) { do_oop_nv(p); }
156 };
158 typedef G1ParCopyClosure<false, G1BarrierNone, false> G1ParScanExtRootClosure;
159 typedef G1ParCopyClosure<true, G1BarrierNone, false> G1ParScanPermClosure;
161 typedef G1ParCopyClosure<false, G1BarrierNone, true> G1ParScanAndMarkExtRootClosure;
162 typedef G1ParCopyClosure<true, G1BarrierNone, true> G1ParScanAndMarkPermClosure;
164 // The following closure types are no longer used but are retained
165 // for historical reasons:
166 // typedef G1ParCopyClosure<false, G1BarrierRS, false> G1ParScanHeapRSClosure;
167 // typedef G1ParCopyClosure<false, G1BarrierRS, true> G1ParScanAndMarkHeapRSClosure;
169 // The following closure type is defined in g1_specialized_oop_closures.hpp:
170 //
171 // typedef G1ParCopyClosure<false, G1BarrierEvac, false> G1ParScanHeapEvacClosure;
173 // We use a separate closure to handle references during evacuation
174 // failure processing.
175 // We could have used another instance of G1ParScanHeapEvacClosure
176 // (since that closure no longer assumes that the references it
177 // handles point into the collection set).
179 typedef G1ParCopyClosure<false, G1BarrierEvac, false> G1ParScanHeapEvacFailureClosure;
181 class FilterIntoCSClosure: public OopClosure {
182 G1CollectedHeap* _g1;
183 OopClosure* _oc;
184 DirtyCardToOopClosure* _dcto_cl;
185 public:
186 FilterIntoCSClosure( DirtyCardToOopClosure* dcto_cl,
187 G1CollectedHeap* g1,
188 OopClosure* oc) :
189 _dcto_cl(dcto_cl), _g1(g1), _oc(oc) { }
191 template <class T> void do_oop_nv(T* p);
192 virtual void do_oop(oop* p) { do_oop_nv(p); }
193 virtual void do_oop(narrowOop* p) { do_oop_nv(p); }
194 bool apply_to_weak_ref_discovered_field() { return true; }
195 bool do_header() { return false; }
196 };
198 class FilterOutOfRegionClosure: public OopClosure {
199 HeapWord* _r_bottom;
200 HeapWord* _r_end;
201 OopClosure* _oc;
202 int _out_of_region;
203 public:
204 FilterOutOfRegionClosure(HeapRegion* r, OopClosure* oc);
205 template <class T> void do_oop_nv(T* p);
206 virtual void do_oop(oop* p) { do_oop_nv(p); }
207 virtual void do_oop(narrowOop* p) { do_oop_nv(p); }
208 bool apply_to_weak_ref_discovered_field() { return true; }
209 bool do_header() { return false; }
210 int out_of_region() { return _out_of_region; }
211 };
213 // Closure for iterating over object fields during concurrent marking
214 class G1CMOopClosure : public OopClosure {
215 G1CollectedHeap* _g1h;
216 ConcurrentMark* _cm;
217 CMTask* _task;
218 public:
219 G1CMOopClosure(G1CollectedHeap* g1h, ConcurrentMark* cm, CMTask* task);
220 template <class T> void do_oop_nv(T* p);
221 virtual void do_oop( oop* p) { do_oop_nv(p); }
222 virtual void do_oop(narrowOop* p) { do_oop_nv(p); }
223 };
225 #endif // SHARE_VM_GC_IMPLEMENTATION_G1_G1OOPCLOSURES_HPP