Tue, 23 Nov 2010 13:22:55 -0800
6989984: Use standard include model for Hospot
Summary: Replaced MakeDeps and the includeDB files with more standardized solutions.
Reviewed-by: coleenp, kvn, kamg
1 /*
2 * Copyright (c) 2003, 2010, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
25 #ifndef SHARE_VM_MEMORY_COMPACTINGPERMGENGEN_HPP
26 #define SHARE_VM_MEMORY_COMPACTINGPERMGENGEN_HPP
28 #include "gc_implementation/shared/generationCounters.hpp"
29 #include "memory/space.hpp"
31 // All heaps contains a "permanent generation," containing permanent
32 // (reflective) objects. This is like a regular generation in some ways,
33 // but unlike one in others, and so is split apart.
35 class PermanentGenerationSpec;
37 // This is the "generation" view of a CompactingPermGen.
38 // NOTE: the shared spaces used for CDS are here handled in
39 // a somewhat awkward and potentially buggy fashion, see CR 6801625.
40 // This infelicity should be fixed, see CR 6897789.
41 class CompactingPermGenGen: public OneContigSpaceCardGeneration {
42 friend class VMStructs;
43 // Abstractly, this is a subtype that gets access to protected fields.
44 friend class CompactingPermGen;
46 private:
47 // Shared spaces
48 PermanentGenerationSpec* _spec;
49 size_t _shared_space_size;
50 VirtualSpace _ro_vs;
51 VirtualSpace _rw_vs;
52 VirtualSpace _md_vs;
53 VirtualSpace _mc_vs;
54 BlockOffsetSharedArray* _ro_bts;
55 BlockOffsetSharedArray* _rw_bts;
56 OffsetTableContigSpace* _ro_space;
57 OffsetTableContigSpace* _rw_space;
59 // With shared spaces there is a dichotomy in the use of the
60 // _virtual_space of the generation. There is a portion of the
61 // _virtual_space that is used for the unshared part of the
62 // permanent generation and a portion that is reserved for the shared part.
63 // The _reserved field in the generation represents both the
64 // unshared and shared parts of the generation. The _reserved
65 // variable is initialized for only the unshared part but is
66 // later extended to include the shared part during initialization
67 // if shared spaces are being used.
68 // The reserved size for the _virtual_space for CompactingPermGenGen
69 // is the size of the space for the permanent generation including the
70 // the shared spaces. This can be seen by the use of MaxPermSize
71 // in the allocation of PermanentGenerationSpec. The space for the
72 // shared spaces is committed separately (???).
73 // In general at initialization only a part of the
74 // space for the unshared part of the permanent generation is
75 // committed and more is committed as the permanent generation is
76 // grown. In growing the permanent generation the capacity() and
77 // max_capacity() of the generation are used. For the permanent
78 // generation (implemented with a CompactingPermGenGen) the capacity()
79 // is taken from the capacity of the space (_the_space variable used for the
80 // unshared part of the generation) and the max_capacity() is based
81 // on the size of the _reserved variable (which includes the size of the
82 // shared spaces) minus the size of the shared spaces.
84 // These values are redundant, but are called out separately to avoid
85 // going through heap/space/gen pointers for performance.
86 static HeapWord* unshared_bottom;
87 static HeapWord* unshared_end;
88 static HeapWord* shared_bottom;
89 static HeapWord* readonly_bottom;
90 static HeapWord* readonly_end;
91 static HeapWord* readwrite_bottom;
92 static HeapWord* readwrite_end;
93 static HeapWord* miscdata_bottom;
94 static HeapWord* miscdata_end;
95 static HeapWord* misccode_bottom;
96 static HeapWord* misccode_end;
97 static HeapWord* shared_end;
99 // List of klassOops whose vtbl entries are used to patch others.
100 static void** _vtbl_list;
102 // Performance Counters
103 GenerationCounters* _gen_counters;
104 CSpaceCounters* _space_counters;
106 void initialize_performance_counters();
108 public:
110 enum {
111 vtbl_list_size = 16, // number of entries in the shared space vtable list.
112 num_virtuals = 200 // number of virtual methods in Klass (or
113 // subclass) objects, or greater.
114 };
116 enum {
117 ro = 0, // read-only shared space in the heap
118 rw = 1, // read-write shared space in the heap
119 md = 2, // miscellaneous data for initializing tables, etc.
120 mc = 3, // miscellaneous code - vtable replacement.
121 n_regions = 4
122 };
124 CompactingPermGenGen(ReservedSpace rs, ReservedSpace shared_rs,
125 size_t initial_byte_size, int level, GenRemSet* remset,
126 ContiguousSpace* space,
127 PermanentGenerationSpec* perm_spec);
129 const char* name() const {
130 return "compacting perm gen";
131 }
133 const char* short_name() const {
134 return "Perm";
135 }
137 // Return the maximum capacity for the object space. This
138 // explicitly does not include the shared spaces.
139 size_t max_capacity() const;
141 void update_counters();
143 void compute_new_size() {
144 assert(false, "Should not call this -- handled at PermGen level.");
145 }
147 bool must_be_youngest() const { return false; }
148 bool must_be_oldest() const { return false; }
150 OffsetTableContigSpace* ro_space() const { return _ro_space; }
151 OffsetTableContigSpace* rw_space() const { return _rw_space; }
152 VirtualSpace* md_space() { return &_md_vs; }
153 VirtualSpace* mc_space() { return &_mc_vs; }
154 ContiguousSpace* unshared_space() const { return _the_space; }
156 static bool inline is_shared(const oopDesc* p) {
157 return (HeapWord*)p >= shared_bottom && (HeapWord*)p < shared_end;
158 }
159 // RedefineClasses note: this tester is used to check residence of
160 // the specified oop in the shared readonly space and not whether
161 // the oop is readonly.
162 static bool inline is_shared_readonly(const oopDesc* p) {
163 return (HeapWord*)p >= readonly_bottom && (HeapWord*)p < readonly_end;
164 }
165 // RedefineClasses note: this tester is used to check residence of
166 // the specified oop in the shared readwrite space and not whether
167 // the oop is readwrite.
168 static bool inline is_shared_readwrite(const oopDesc* p) {
169 return (HeapWord*)p >= readwrite_bottom && (HeapWord*)p < readwrite_end;
170 }
172 bool is_in_unshared(const void* p) const {
173 return OneContigSpaceCardGeneration::is_in(p);
174 }
176 bool is_in_shared(const void* p) const {
177 return p >= shared_bottom && p < shared_end;
178 }
180 inline bool is_in(const void* p) const {
181 return is_in_unshared(p) || is_in_shared(p);
182 }
184 inline PermanentGenerationSpec* spec() const { return _spec; }
185 inline void set_spec(PermanentGenerationSpec* spec) { _spec = spec; }
187 void pre_adjust_pointers();
188 void adjust_pointers();
189 void space_iterate(SpaceClosure* blk, bool usedOnly = false);
190 void print_on(outputStream* st) const;
191 void younger_refs_iterate(OopsInGenClosure* blk);
192 void compact();
193 void post_compact();
194 size_t contiguous_available() const;
196 void clear_remembered_set();
197 void invalidate_remembered_set();
199 inline bool block_is_obj(const HeapWord* addr) const {
200 if (addr < the_space()->top()) return true;
201 else if (addr < the_space()->end()) return false;
202 else if (addr < ro_space()->top()) return true;
203 else if (addr < ro_space()->end()) return false;
204 else if (addr < rw_space()->top()) return true;
205 else return false;
206 }
209 inline size_t block_size(const HeapWord* addr) const {
210 if (addr < the_space()->top()) {
211 return oop(addr)->size();
212 }
213 else if (addr < the_space()->end()) {
214 assert(addr == the_space()->top(), "non-block head arg to block_size");
215 return the_space()->end() - the_space()->top();
216 }
218 else if (addr < ro_space()->top()) {
219 return oop(addr)->size();
220 }
221 else if (addr < ro_space()->end()) {
222 assert(addr == ro_space()->top(), "non-block head arg to block_size");
223 return ro_space()->end() - ro_space()->top();
224 }
226 else if (addr < rw_space()->top()) {
227 return oop(addr)->size();
228 }
229 else {
230 assert(addr == rw_space()->top(), "non-block head arg to block_size");
231 return rw_space()->end() - rw_space()->top();
232 }
233 }
235 static void generate_vtable_methods(void** vtbl_list,
236 void** vtable,
237 char** md_top, char* md_end,
238 char** mc_top, char* mc_end);
240 void verify(bool allow_dirty);
242 // Serialization
243 static void initialize_oops() KERNEL_RETURN;
244 static void serialize_oops(SerializeOopClosure* soc);
245 void serialize_bts(SerializeOopClosure* soc);
247 // Initiate dumping of shared file.
248 static jint dump_shared(GrowableArray<oop>* class_promote_order, TRAPS);
250 // JVM/TI RedefineClasses() support:
251 // Remap the shared readonly space to shared readwrite, private if
252 // sharing is enabled. Simply returns true if sharing is not enabled
253 // or if the remapping has already been done by a prior call.
254 static bool remap_shared_readonly_as_readwrite();
255 };
257 #endif // SHARE_VM_MEMORY_COMPACTINGPERMGENGEN_HPP