Thu, 26 Jun 2014 11:36:58 +0200
8047818: G1 HeapRegions can no longer be ContiguousSpaces
Summary: Change parent of G1OffsetTableContigSpace to CompactibleSpace, reimplement missing functionality
Reviewed-by: stefank, jmasa, tschatzl
duke@435 | 1 | /* |
fparain@5409 | 2 | * Copyright (c) 1997, 2013, Oracle and/or its affiliates. All rights reserved. |
duke@435 | 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
duke@435 | 4 | * |
duke@435 | 5 | * This code is free software; you can redistribute it and/or modify it |
duke@435 | 6 | * under the terms of the GNU General Public License version 2 only, as |
duke@435 | 7 | * published by the Free Software Foundation. |
duke@435 | 8 | * |
duke@435 | 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
duke@435 | 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
duke@435 | 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
duke@435 | 12 | * version 2 for more details (a copy is included in the LICENSE file that |
duke@435 | 13 | * accompanied this code). |
duke@435 | 14 | * |
duke@435 | 15 | * You should have received a copy of the GNU General Public License version |
duke@435 | 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
duke@435 | 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
duke@435 | 18 | * |
trims@1907 | 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
trims@1907 | 20 | * or visit www.oracle.com if you need additional information or have any |
trims@1907 | 21 | * questions. |
duke@435 | 22 | * |
duke@435 | 23 | */ |
duke@435 | 24 | |
stefank@2314 | 25 | #ifndef SHARE_VM_MEMORY_RESOURCEAREA_HPP |
stefank@2314 | 26 | #define SHARE_VM_MEMORY_RESOURCEAREA_HPP |
stefank@2314 | 27 | |
stefank@2314 | 28 | #include "memory/allocation.hpp" |
stefank@4299 | 29 | #include "runtime/thread.inline.hpp" |
stefank@2314 | 30 | |
duke@435 | 31 | // The resource area holds temporary data structures in the VM. |
duke@435 | 32 | // The actual allocation areas are thread local. Typical usage: |
duke@435 | 33 | // |
duke@435 | 34 | // ... |
duke@435 | 35 | // { |
duke@435 | 36 | // ResourceMark rm; |
duke@435 | 37 | // int foo[] = NEW_RESOURCE_ARRAY(int, 64); |
duke@435 | 38 | // ... |
duke@435 | 39 | // } |
duke@435 | 40 | // ... |
duke@435 | 41 | |
duke@435 | 42 | //------------------------------ResourceArea----------------------------------- |
duke@435 | 43 | // A ResourceArea is an Arena that supports safe usage of ResourceMark. |
duke@435 | 44 | class ResourceArea: public Arena { |
duke@435 | 45 | friend class ResourceMark; |
duke@435 | 46 | friend class DeoptResourceMark; |
never@3138 | 47 | friend class VMStructs; |
duke@435 | 48 | debug_only(int _nesting;) // current # of nested ResourceMarks |
duke@435 | 49 | debug_only(static int _warned;) // to suppress multiple warnings |
duke@435 | 50 | |
duke@435 | 51 | public: |
duke@435 | 52 | ResourceArea() { |
duke@435 | 53 | debug_only(_nesting = 0;) |
duke@435 | 54 | } |
duke@435 | 55 | |
duke@435 | 56 | ResourceArea(size_t init_size) : Arena(init_size) { |
duke@435 | 57 | debug_only(_nesting = 0;); |
duke@435 | 58 | } |
duke@435 | 59 | |
nloodin@4183 | 60 | char* allocate_bytes(size_t size, AllocFailType alloc_failmode = AllocFailStrategy::EXIT_OOM) { |
duke@435 | 61 | #ifdef ASSERT |
duke@435 | 62 | if (_nesting < 1 && !_warned++) |
duke@435 | 63 | fatal("memory leak: allocating without ResourceMark"); |
duke@435 | 64 | if (UseMallocOnly) { |
duke@435 | 65 | // use malloc, but save pointer in res. area for later freeing |
duke@435 | 66 | char** save = (char**)internal_malloc_4(sizeof(char*)); |
zgu@3900 | 67 | return (*save = (char*)os::malloc(size, mtThread)); |
duke@435 | 68 | } |
duke@435 | 69 | #endif |
nloodin@4183 | 70 | return (char*)Amalloc(size, alloc_failmode); |
duke@435 | 71 | } |
duke@435 | 72 | |
duke@435 | 73 | debug_only(int nesting() const { return _nesting; }); |
duke@435 | 74 | }; |
duke@435 | 75 | |
duke@435 | 76 | |
duke@435 | 77 | //------------------------------ResourceMark----------------------------------- |
duke@435 | 78 | // A resource mark releases all resources allocated after it was constructed |
duke@435 | 79 | // when the destructor is called. Typically used as a local variable. |
duke@435 | 80 | class ResourceMark: public StackObj { |
duke@435 | 81 | protected: |
duke@435 | 82 | ResourceArea *_area; // Resource area to stack allocate |
duke@435 | 83 | Chunk *_chunk; // saved arena chunk |
duke@435 | 84 | char *_hwm, *_max; |
zgu@3900 | 85 | size_t _size_in_bytes; |
fparain@5409 | 86 | #ifdef ASSERT |
fparain@5409 | 87 | Thread* _thread; |
fparain@5409 | 88 | ResourceMark* _previous_resource_mark; |
fparain@5409 | 89 | #endif //ASSERT |
duke@435 | 90 | |
duke@435 | 91 | void initialize(Thread *thread) { |
duke@435 | 92 | _area = thread->resource_area(); |
duke@435 | 93 | _chunk = _area->_chunk; |
duke@435 | 94 | _hwm = _area->_hwm; |
duke@435 | 95 | _max= _area->_max; |
zgu@3900 | 96 | _size_in_bytes = _area->size_in_bytes(); |
duke@435 | 97 | debug_only(_area->_nesting++;) |
duke@435 | 98 | assert( _area->_nesting > 0, "must stack allocate RMs" ); |
fparain@5409 | 99 | #ifdef ASSERT |
fparain@5409 | 100 | _thread = thread; |
fparain@5409 | 101 | _previous_resource_mark = thread->current_resource_mark(); |
fparain@5409 | 102 | thread->set_current_resource_mark(this); |
fparain@5409 | 103 | #endif // ASSERT |
duke@435 | 104 | } |
duke@435 | 105 | public: |
duke@435 | 106 | |
duke@435 | 107 | #ifndef ASSERT |
duke@435 | 108 | ResourceMark(Thread *thread) { |
duke@435 | 109 | assert(thread == Thread::current(), "not the current thread"); |
duke@435 | 110 | initialize(thread); |
duke@435 | 111 | } |
duke@435 | 112 | #else |
duke@435 | 113 | ResourceMark(Thread *thread); |
duke@435 | 114 | #endif // ASSERT |
duke@435 | 115 | |
duke@435 | 116 | ResourceMark() { initialize(Thread::current()); } |
duke@435 | 117 | |
duke@435 | 118 | ResourceMark( ResourceArea *r ) : |
duke@435 | 119 | _area(r), _chunk(r->_chunk), _hwm(r->_hwm), _max(r->_max) { |
zgu@3900 | 120 | _size_in_bytes = r->_size_in_bytes; |
duke@435 | 121 | debug_only(_area->_nesting++;) |
duke@435 | 122 | assert( _area->_nesting > 0, "must stack allocate RMs" ); |
fparain@5409 | 123 | #ifdef ASSERT |
fparain@5409 | 124 | Thread* thread = ThreadLocalStorage::thread(); |
fparain@5409 | 125 | if (thread != NULL) { |
fparain@5409 | 126 | _thread = thread; |
fparain@5409 | 127 | _previous_resource_mark = thread->current_resource_mark(); |
fparain@5409 | 128 | thread->set_current_resource_mark(this); |
fparain@5409 | 129 | } else { |
fparain@5409 | 130 | _thread = NULL; |
fparain@5409 | 131 | _previous_resource_mark = NULL; |
fparain@5409 | 132 | } |
fparain@5409 | 133 | #endif // ASSERT |
duke@435 | 134 | } |
duke@435 | 135 | |
duke@435 | 136 | void reset_to_mark() { |
duke@435 | 137 | if (UseMallocOnly) free_malloced_objects(); |
duke@435 | 138 | |
zgu@4193 | 139 | if( _chunk->next() ) { // Delete later chunks |
zgu@4193 | 140 | // reset arena size before delete chunks. Otherwise, the total |
zgu@4193 | 141 | // arena size could exceed total chunk size |
zgu@4193 | 142 | assert(_area->size_in_bytes() > size_in_bytes(), "Sanity check"); |
zgu@4193 | 143 | _area->set_size_in_bytes(size_in_bytes()); |
duke@435 | 144 | _chunk->next_chop(); |
zgu@4193 | 145 | } else { |
zgu@4193 | 146 | assert(_area->size_in_bytes() == size_in_bytes(), "Sanity check"); |
zgu@4193 | 147 | } |
duke@435 | 148 | _area->_chunk = _chunk; // Roll back arena to saved chunk |
duke@435 | 149 | _area->_hwm = _hwm; |
duke@435 | 150 | _area->_max = _max; |
duke@435 | 151 | |
duke@435 | 152 | // clear out this chunk (to detect allocation bugs) |
duke@435 | 153 | if (ZapResourceArea) memset(_hwm, badResourceValue, _max - _hwm); |
duke@435 | 154 | } |
duke@435 | 155 | |
duke@435 | 156 | ~ResourceMark() { |
duke@435 | 157 | assert( _area->_nesting > 0, "must stack allocate RMs" ); |
duke@435 | 158 | debug_only(_area->_nesting--;) |
duke@435 | 159 | reset_to_mark(); |
fparain@5409 | 160 | #ifdef ASSERT |
fparain@5409 | 161 | if (_thread != NULL) { |
fparain@5409 | 162 | _thread->set_current_resource_mark(_previous_resource_mark); |
fparain@5409 | 163 | } |
fparain@5409 | 164 | #endif // ASSERT |
duke@435 | 165 | } |
duke@435 | 166 | |
duke@435 | 167 | |
duke@435 | 168 | private: |
duke@435 | 169 | void free_malloced_objects() PRODUCT_RETURN; |
zgu@3900 | 170 | size_t size_in_bytes() { return _size_in_bytes; } |
duke@435 | 171 | }; |
duke@435 | 172 | |
duke@435 | 173 | //------------------------------DeoptResourceMark----------------------------------- |
duke@435 | 174 | // A deopt resource mark releases all resources allocated after it was constructed |
duke@435 | 175 | // when the destructor is called. Typically used as a local variable. It differs |
duke@435 | 176 | // from a typical resource more in that it is C-Heap allocated so that deoptimization |
duke@435 | 177 | // can use data structures that are arena based but are not amenable to vanilla |
duke@435 | 178 | // ResourceMarks because deoptimization can not use a stack allocated mark. During |
duke@435 | 179 | // deoptimization we go thru the following steps: |
duke@435 | 180 | // |
duke@435 | 181 | // 0: start in assembly stub and call either uncommon_trap/fetch_unroll_info |
duke@435 | 182 | // 1: create the vframeArray (contains pointers to Resource allocated structures) |
duke@435 | 183 | // This allocates the DeoptResourceMark. |
duke@435 | 184 | // 2: return to assembly stub and remove stub frame and deoptee frame and create |
duke@435 | 185 | // the new skeletal frames. |
duke@435 | 186 | // 3: push new stub frame and call unpack_frames |
duke@435 | 187 | // 4: retrieve information from the vframeArray to populate the skeletal frames |
duke@435 | 188 | // 5: release the DeoptResourceMark |
duke@435 | 189 | // 6: return to stub and eventually to interpreter |
duke@435 | 190 | // |
duke@435 | 191 | // With old style eager deoptimization the vframeArray was created by the vmThread there |
duke@435 | 192 | // was no way for the vframeArray to contain resource allocated objects and so |
duke@435 | 193 | // a complex set of data structures to simulate an array of vframes in CHeap memory |
duke@435 | 194 | // was used. With new style lazy deoptimization the vframeArray is created in the |
duke@435 | 195 | // the thread that will use it and we can use a much simpler scheme for the vframeArray |
duke@435 | 196 | // leveraging existing data structures if we simply create a way to manage this one |
duke@435 | 197 | // special need for a ResourceMark. If ResourceMark simply inherited from CHeapObj |
duke@435 | 198 | // then existing ResourceMarks would work fine since no one use new to allocate them |
duke@435 | 199 | // and they would be stack allocated. This leaves open the possibilty of accidental |
duke@435 | 200 | // misuse so we simple duplicate the ResourceMark functionality here. |
duke@435 | 201 | |
zgu@3900 | 202 | class DeoptResourceMark: public CHeapObj<mtInternal> { |
duke@435 | 203 | protected: |
duke@435 | 204 | ResourceArea *_area; // Resource area to stack allocate |
duke@435 | 205 | Chunk *_chunk; // saved arena chunk |
duke@435 | 206 | char *_hwm, *_max; |
zgu@3900 | 207 | size_t _size_in_bytes; |
duke@435 | 208 | |
duke@435 | 209 | void initialize(Thread *thread) { |
duke@435 | 210 | _area = thread->resource_area(); |
duke@435 | 211 | _chunk = _area->_chunk; |
duke@435 | 212 | _hwm = _area->_hwm; |
duke@435 | 213 | _max= _area->_max; |
zgu@3900 | 214 | _size_in_bytes = _area->size_in_bytes(); |
duke@435 | 215 | debug_only(_area->_nesting++;) |
duke@435 | 216 | assert( _area->_nesting > 0, "must stack allocate RMs" ); |
duke@435 | 217 | } |
duke@435 | 218 | |
duke@435 | 219 | public: |
duke@435 | 220 | |
duke@435 | 221 | #ifndef ASSERT |
duke@435 | 222 | DeoptResourceMark(Thread *thread) { |
duke@435 | 223 | assert(thread == Thread::current(), "not the current thread"); |
duke@435 | 224 | initialize(thread); |
duke@435 | 225 | } |
duke@435 | 226 | #else |
duke@435 | 227 | DeoptResourceMark(Thread *thread); |
duke@435 | 228 | #endif // ASSERT |
duke@435 | 229 | |
duke@435 | 230 | DeoptResourceMark() { initialize(Thread::current()); } |
duke@435 | 231 | |
duke@435 | 232 | DeoptResourceMark( ResourceArea *r ) : |
duke@435 | 233 | _area(r), _chunk(r->_chunk), _hwm(r->_hwm), _max(r->_max) { |
zgu@3900 | 234 | _size_in_bytes = _area->size_in_bytes(); |
duke@435 | 235 | debug_only(_area->_nesting++;) |
duke@435 | 236 | assert( _area->_nesting > 0, "must stack allocate RMs" ); |
duke@435 | 237 | } |
duke@435 | 238 | |
duke@435 | 239 | void reset_to_mark() { |
duke@435 | 240 | if (UseMallocOnly) free_malloced_objects(); |
duke@435 | 241 | |
zgu@4193 | 242 | if( _chunk->next() ) { // Delete later chunks |
zgu@4193 | 243 | // reset arena size before delete chunks. Otherwise, the total |
zgu@4193 | 244 | // arena size could exceed total chunk size |
zgu@4193 | 245 | assert(_area->size_in_bytes() > size_in_bytes(), "Sanity check"); |
zgu@4193 | 246 | _area->set_size_in_bytes(size_in_bytes()); |
duke@435 | 247 | _chunk->next_chop(); |
zgu@4193 | 248 | } else { |
zgu@4193 | 249 | assert(_area->size_in_bytes() == size_in_bytes(), "Sanity check"); |
zgu@4193 | 250 | } |
duke@435 | 251 | _area->_chunk = _chunk; // Roll back arena to saved chunk |
duke@435 | 252 | _area->_hwm = _hwm; |
duke@435 | 253 | _area->_max = _max; |
duke@435 | 254 | |
duke@435 | 255 | // clear out this chunk (to detect allocation bugs) |
duke@435 | 256 | if (ZapResourceArea) memset(_hwm, badResourceValue, _max - _hwm); |
duke@435 | 257 | } |
duke@435 | 258 | |
duke@435 | 259 | ~DeoptResourceMark() { |
duke@435 | 260 | assert( _area->_nesting > 0, "must stack allocate RMs" ); |
duke@435 | 261 | debug_only(_area->_nesting--;) |
duke@435 | 262 | reset_to_mark(); |
duke@435 | 263 | } |
duke@435 | 264 | |
duke@435 | 265 | |
duke@435 | 266 | private: |
duke@435 | 267 | void free_malloced_objects() PRODUCT_RETURN; |
zgu@3900 | 268 | size_t size_in_bytes() { return _size_in_bytes; }; |
duke@435 | 269 | }; |
stefank@2314 | 270 | |
stefank@2314 | 271 | #endif // SHARE_VM_MEMORY_RESOURCEAREA_HPP |