Thu, 27 Jan 2011 16:11:27 -0800
6990754: Use native memory and reference counting to implement SymbolTable
Summary: move symbols from permgen into C heap and reference count them
Reviewed-by: never, acorn, jmasa, stefank
1 /*
2 * Copyright (c) 1997, 2010, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
25 #ifndef SHARE_VM_MEMORY_ITERATOR_HPP
26 #define SHARE_VM_MEMORY_ITERATOR_HPP
28 #include "memory/allocation.hpp"
29 #include "memory/memRegion.hpp"
30 #include "runtime/prefetch.hpp"
31 #include "utilities/top.hpp"
33 // The following classes are C++ `closures` for iterating over objects, roots and spaces
35 class CodeBlob;
36 class nmethod;
37 class ReferenceProcessor;
38 class DataLayout;
40 // Closure provides abortability.
42 class Closure : public StackObj {
43 protected:
44 bool _abort;
45 void set_abort() { _abort = true; }
46 public:
47 Closure() : _abort(false) {}
48 // A subtype can use this mechanism to indicate to some iterator mapping
49 // functions that the iteration should cease.
50 bool abort() { return _abort; }
51 void clear_abort() { _abort = false; }
52 };
54 // OopClosure is used for iterating through roots (oop*)
56 class OopClosure : public Closure {
57 public:
58 ReferenceProcessor* _ref_processor;
59 OopClosure(ReferenceProcessor* rp) : _ref_processor(rp) { }
60 OopClosure() : _ref_processor(NULL) { }
61 virtual void do_oop(oop* o) = 0;
62 virtual void do_oop_v(oop* o) { do_oop(o); }
63 virtual void do_oop(narrowOop* o) = 0;
64 virtual void do_oop_v(narrowOop* o) { do_oop(o); }
66 // In support of post-processing of weak links of KlassKlass objects;
67 // see KlassKlass::oop_oop_iterate().
69 virtual const bool should_remember_klasses() const {
70 assert(!must_remember_klasses(), "Should have overriden this method.");
71 return false;
72 }
74 virtual void remember_klass(Klass* k) { /* do nothing */ }
76 // In support of post-processing of weak references in
77 // ProfileData (MethodDataOop) objects; see, for example,
78 // VirtualCallData::oop_iterate().
79 virtual const bool should_remember_mdo() const { return false; }
80 virtual void remember_mdo(DataLayout* v) { /* do nothing */ }
82 // The methods below control how object iterations invoking this closure
83 // should be performed:
85 // If "true", invoke on header klass field.
86 bool do_header() { return true; } // Note that this is non-virtual.
87 // Controls how prefetching is done for invocations of this closure.
88 Prefetch::style prefetch_style() { // Note that this is non-virtual.
89 return Prefetch::do_none;
90 }
92 // True iff this closure may be safely applied more than once to an oop
93 // location without an intervening "major reset" (like the end of a GC).
94 virtual bool idempotent() { return false; }
95 virtual bool apply_to_weak_ref_discovered_field() { return false; }
97 #ifdef ASSERT
98 static bool _must_remember_klasses;
99 static bool must_remember_klasses();
100 static void set_must_remember_klasses(bool v);
101 #endif
102 };
104 // ObjectClosure is used for iterating through an object space
106 class ObjectClosure : public Closure {
107 public:
108 // Called for each object.
109 virtual void do_object(oop obj) = 0;
110 };
113 class BoolObjectClosure : public ObjectClosure {
114 public:
115 virtual bool do_object_b(oop obj) = 0;
116 };
118 // Applies an oop closure to all ref fields in objects iterated over in an
119 // object iteration.
120 class ObjectToOopClosure: public ObjectClosure {
121 OopClosure* _cl;
122 public:
123 void do_object(oop obj);
124 ObjectToOopClosure(OopClosure* cl) : _cl(cl) {}
125 };
127 // A version of ObjectClosure with "memory" (see _previous_address below)
128 class UpwardsObjectClosure: public BoolObjectClosure {
129 HeapWord* _previous_address;
130 public:
131 UpwardsObjectClosure() : _previous_address(NULL) { }
132 void set_previous(HeapWord* addr) { _previous_address = addr; }
133 HeapWord* previous() { return _previous_address; }
134 // A return value of "true" can be used by the caller to decide
135 // if this object's end should *NOT* be recorded in
136 // _previous_address above.
137 virtual bool do_object_bm(oop obj, MemRegion mr) = 0;
138 };
140 // A version of ObjectClosure that is expected to be robust
141 // in the face of possibly uninitialized objects.
142 class ObjectClosureCareful : public ObjectClosure {
143 public:
144 virtual size_t do_object_careful_m(oop p, MemRegion mr) = 0;
145 virtual size_t do_object_careful(oop p) = 0;
146 };
148 // The following are used in CompactibleFreeListSpace and
149 // ConcurrentMarkSweepGeneration.
151 // Blk closure (abstract class)
152 class BlkClosure : public StackObj {
153 public:
154 virtual size_t do_blk(HeapWord* addr) = 0;
155 };
157 // A version of BlkClosure that is expected to be robust
158 // in the face of possibly uninitialized objects.
159 class BlkClosureCareful : public BlkClosure {
160 public:
161 size_t do_blk(HeapWord* addr) {
162 guarantee(false, "call do_blk_careful instead");
163 return 0;
164 }
165 virtual size_t do_blk_careful(HeapWord* addr) = 0;
166 };
168 // SpaceClosure is used for iterating over spaces
170 class Space;
171 class CompactibleSpace;
173 class SpaceClosure : public StackObj {
174 public:
175 // Called for each space
176 virtual void do_space(Space* s) = 0;
177 };
179 class CompactibleSpaceClosure : public StackObj {
180 public:
181 // Called for each compactible space
182 virtual void do_space(CompactibleSpace* s) = 0;
183 };
186 // CodeBlobClosure is used for iterating through code blobs
187 // in the code cache or on thread stacks
189 class CodeBlobClosure : public Closure {
190 public:
191 // Called for each code blob.
192 virtual void do_code_blob(CodeBlob* cb) = 0;
193 };
196 class MarkingCodeBlobClosure : public CodeBlobClosure {
197 public:
198 // Called for each code blob, but at most once per unique blob.
199 virtual void do_newly_marked_nmethod(nmethod* nm) = 0;
201 virtual void do_code_blob(CodeBlob* cb);
202 // = { if (!nmethod(cb)->test_set_oops_do_mark()) do_newly_marked_nmethod(cb); }
204 class MarkScope : public StackObj {
205 protected:
206 bool _active;
207 public:
208 MarkScope(bool activate = true);
209 // = { if (active) nmethod::oops_do_marking_prologue(); }
210 ~MarkScope();
211 // = { if (active) nmethod::oops_do_marking_epilogue(); }
212 };
213 };
216 // Applies an oop closure to all ref fields in code blobs
217 // iterated over in an object iteration.
218 class CodeBlobToOopClosure: public MarkingCodeBlobClosure {
219 OopClosure* _cl;
220 bool _do_marking;
221 public:
222 virtual void do_newly_marked_nmethod(nmethod* cb);
223 // = { cb->oops_do(_cl); }
224 virtual void do_code_blob(CodeBlob* cb);
225 // = { if (_do_marking) super::do_code_blob(cb); else cb->oops_do(_cl); }
226 CodeBlobToOopClosure(OopClosure* cl, bool do_marking)
227 : _cl(cl), _do_marking(do_marking) {}
228 };
232 // MonitorClosure is used for iterating over monitors in the monitors cache
234 class ObjectMonitor;
236 class MonitorClosure : public StackObj {
237 public:
238 // called for each monitor in cache
239 virtual void do_monitor(ObjectMonitor* m) = 0;
240 };
242 // A closure that is applied without any arguments.
243 class VoidClosure : public StackObj {
244 public:
245 // I would have liked to declare this a pure virtual, but that breaks
246 // in mysterious ways, for unknown reasons.
247 virtual void do_void();
248 };
251 // YieldClosure is intended for use by iteration loops
252 // to incrementalize their work, allowing interleaving
253 // of an interruptable task so as to allow other
254 // threads to run (which may not otherwise be able to access
255 // exclusive resources, for instance). Additionally, the
256 // closure also allows for aborting an ongoing iteration
257 // by means of checking the return value from the polling
258 // call.
259 class YieldClosure : public StackObj {
260 public:
261 virtual bool should_return() = 0;
262 };
264 // Abstract closure for serializing data (read or write).
266 class SerializeOopClosure : public OopClosure {
267 public:
268 // Return bool indicating whether closure implements read or write.
269 virtual bool reading() const = 0;
271 // Read/write the int pointed to by i.
272 virtual void do_int(int* i) = 0;
274 // Read/write the size_t pointed to by i.
275 virtual void do_size_t(size_t* i) = 0;
277 // Read/write the void pointer pointed to by p.
278 virtual void do_ptr(void** p) = 0;
280 // Read/write the HeapWord pointer pointed to be p.
281 virtual void do_ptr(HeapWord** p) = 0;
283 // Read/write the region specified.
284 virtual void do_region(u_char* start, size_t size) = 0;
286 // Check/write the tag. If reading, then compare the tag against
287 // the passed in value and fail is they don't match. This allows
288 // for verification that sections of the serialized data are of the
289 // correct length.
290 virtual void do_tag(int tag) = 0;
291 };
293 class SymbolClosure : public StackObj {
294 public:
295 virtual void do_symbol(Symbol**) = 0;
297 // Clear LSB in symbol address; it can be set by CPSlot.
298 static Symbol* load_symbol(Symbol** p) {
299 return (Symbol*)(intptr_t(*p) & ~1);
300 }
302 // Store symbol, adjusting new pointer if the original pointer was adjusted
303 // (symbol references in constant pool slots have their LSB set to 1).
304 static void store_symbol(Symbol** p, Symbol* sym) {
305 *p = (Symbol*)(intptr_t(sym) | (intptr_t(*p) & 1));
306 }
307 };
309 #ifdef ASSERT
310 // This class is used to flag phases of a collection that
311 // can unload classes and which should override the
312 // should_remember_klasses() and remember_klass() of OopClosure.
313 // The _must_remember_klasses is set in the contructor and restored
314 // in the destructor. _must_remember_klasses is checked in assertions
315 // in the OopClosure implementations of should_remember_klasses() and
316 // remember_klass() and the expectation is that the OopClosure
317 // implementation should not be in use if _must_remember_klasses is set.
318 // Instances of RememberKlassesChecker can be place in
319 // marking phases of collections which can do class unloading.
320 // RememberKlassesChecker can be passed "false" to turn off checking.
321 // It is used by CMS when CMS yields to a different collector.
322 class RememberKlassesChecker: StackObj {
323 bool _saved_state;
324 bool _do_check;
325 public:
326 RememberKlassesChecker(bool checking_on) : _saved_state(false),
327 _do_check(true) {
328 // The ClassUnloading unloading flag affects the collectors except
329 // for CMS.
330 // CMS unloads classes if CMSClassUnloadingEnabled is true or
331 // if ExplicitGCInvokesConcurrentAndUnloadsClasses is true and
332 // the current collection is an explicit collection. Turning
333 // on the checking in general for
334 // ExplicitGCInvokesConcurrentAndUnloadsClasses and
335 // UseConcMarkSweepGC should not lead to false positives.
336 _do_check =
337 ClassUnloading && !UseConcMarkSweepGC ||
338 CMSClassUnloadingEnabled && UseConcMarkSweepGC ||
339 ExplicitGCInvokesConcurrentAndUnloadsClasses && UseConcMarkSweepGC;
340 if (_do_check) {
341 _saved_state = OopClosure::must_remember_klasses();
342 OopClosure::set_must_remember_klasses(checking_on);
343 }
344 }
345 ~RememberKlassesChecker() {
346 if (_do_check) {
347 OopClosure::set_must_remember_klasses(_saved_state);
348 }
349 }
350 };
351 #endif // ASSERT
353 #endif // SHARE_VM_MEMORY_ITERATOR_HPP