Tue, 09 Oct 2012 10:09:34 -0700
7197424: update copyright year to match last edit in jdk8 hotspot repository
Summary: Update copyright year to 2012 for relevant files
Reviewed-by: dholmes, coleenp
zgu@3900 | 1 | /* |
zgu@3900 | 2 | * Copyright (c) 2012, Oracle and/or its affiliates. All rights reserved. |
zgu@3900 | 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
zgu@3900 | 4 | * |
zgu@3900 | 5 | * This code is free software; you can redistribute it and/or modify it |
zgu@3900 | 6 | * under the terms of the GNU General Public License version 2 only, as |
zgu@3900 | 7 | * published by the Free Software Foundation. |
zgu@3900 | 8 | * |
zgu@3900 | 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
zgu@3900 | 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
zgu@3900 | 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
zgu@3900 | 12 | * version 2 for more details (a copy is included in the LICENSE file that |
zgu@3900 | 13 | * accompanied this code). |
zgu@3900 | 14 | * |
zgu@3900 | 15 | * You should have received a copy of the GNU General Public License version |
zgu@3900 | 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
zgu@3900 | 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
zgu@3900 | 18 | * |
zgu@3900 | 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
zgu@3900 | 20 | * or visit www.oracle.com if you need additional information or have any |
zgu@3900 | 21 | * questions. |
zgu@3900 | 22 | * |
zgu@3900 | 23 | */ |
zgu@3900 | 24 | |
zgu@3900 | 25 | #ifndef SHARE_VM_SERVICES_MEM_PTR_HPP |
zgu@3900 | 26 | #define SHARE_VM_SERVICES_MEM_PTR_HPP |
zgu@3900 | 27 | |
zgu@3900 | 28 | #include "memory/allocation.hpp" |
zgu@3900 | 29 | #include "runtime/atomic.hpp" |
zgu@3900 | 30 | #include "runtime/os.hpp" |
zgu@3900 | 31 | #include "runtime/safepoint.hpp" |
zgu@3900 | 32 | |
zgu@3900 | 33 | /* |
zgu@3900 | 34 | * global sequence generator that generates sequence numbers to serialize |
zgu@3900 | 35 | * memory records. |
zgu@3900 | 36 | */ |
zgu@3900 | 37 | class SequenceGenerator : AllStatic { |
zgu@3900 | 38 | public: |
zgu@3900 | 39 | static jint next(); |
zgu@3900 | 40 | |
zgu@3900 | 41 | // peek last sequence number |
zgu@3900 | 42 | static jint peek() { |
zgu@3900 | 43 | return _seq_number; |
zgu@3900 | 44 | } |
zgu@3900 | 45 | |
zgu@3900 | 46 | // reset sequence number |
zgu@3900 | 47 | static void reset() { |
zgu@3900 | 48 | assert(SafepointSynchronize::is_at_safepoint(), "Safepoint required"); |
zgu@3900 | 49 | _seq_number = 1; |
zgu@3900 | 50 | DEBUG_ONLY(_generation ++;) |
zgu@3900 | 51 | }; |
zgu@3900 | 52 | |
zgu@3900 | 53 | DEBUG_ONLY(static unsigned long current_generation() { return (unsigned long)_generation; }) |
zgu@3994 | 54 | NOT_PRODUCT(static jint max_seq_num() { return _max_seq_number; }) |
zgu@3900 | 55 | |
zgu@3900 | 56 | private: |
zgu@3900 | 57 | static volatile jint _seq_number; |
zgu@3994 | 58 | NOT_PRODUCT(static jint _max_seq_number; ) |
zgu@3900 | 59 | DEBUG_ONLY(static volatile unsigned long _generation; ) |
zgu@3900 | 60 | }; |
zgu@3900 | 61 | |
zgu@3900 | 62 | /* |
zgu@3900 | 63 | * followings are the classes that are used to hold memory activity records in different stages. |
zgu@3900 | 64 | * MemPointer |
zgu@3900 | 65 | * |--------MemPointerRecord |
zgu@3900 | 66 | * | |
zgu@3900 | 67 | * |----MemPointerRecordEx |
zgu@3900 | 68 | * | | |
zgu@3900 | 69 | * | |-------SeqMemPointerRecordEx |
zgu@3900 | 70 | * | |
zgu@3900 | 71 | * |----SeqMemPointerRecord |
zgu@3900 | 72 | * | |
zgu@3900 | 73 | * |----VMMemRegion |
zgu@3900 | 74 | * | |
zgu@3900 | 75 | * |-----VMMemRegionEx |
zgu@3900 | 76 | * |
zgu@3900 | 77 | * |
zgu@3900 | 78 | * prefix 'Seq' - sequenced, the record contains a sequence number |
zgu@3900 | 79 | * surfix 'Ex' - extension, the record contains a caller's pc |
zgu@3900 | 80 | * |
zgu@3900 | 81 | * per-thread recorder : SeqMemPointerRecord(Ex) |
zgu@3900 | 82 | * snapshot staging : SeqMemPointerRecord(Ex) |
zgu@3900 | 83 | * snapshot : MemPointerRecord(Ex) and VMMemRegion(Ex) |
zgu@3900 | 84 | * |
zgu@3900 | 85 | */ |
zgu@3900 | 86 | |
zgu@3900 | 87 | /* |
zgu@3900 | 88 | * class that wraps an address to a memory block, |
zgu@3900 | 89 | * the memory pointer either points to a malloc'd |
zgu@3900 | 90 | * memory block, or a mmap'd memory block |
zgu@3900 | 91 | */ |
zgu@3900 | 92 | class MemPointer : public _ValueObj { |
zgu@3900 | 93 | public: |
zgu@3900 | 94 | MemPointer(): _addr(0) { } |
zgu@3900 | 95 | MemPointer(address addr): _addr(addr) { } |
zgu@3900 | 96 | |
zgu@3900 | 97 | MemPointer(const MemPointer& copy_from) { |
zgu@3900 | 98 | _addr = copy_from.addr(); |
zgu@3900 | 99 | } |
zgu@3900 | 100 | |
zgu@3900 | 101 | inline address addr() const { |
zgu@3900 | 102 | return _addr; |
zgu@3900 | 103 | } |
zgu@3900 | 104 | |
zgu@3900 | 105 | inline operator address() const { |
zgu@3900 | 106 | return addr(); |
zgu@3900 | 107 | } |
zgu@3900 | 108 | |
zgu@3900 | 109 | inline bool operator == (const MemPointer& other) const { |
zgu@3900 | 110 | return addr() == other.addr(); |
zgu@3900 | 111 | } |
zgu@3900 | 112 | |
zgu@3900 | 113 | inline MemPointer& operator = (const MemPointer& other) { |
zgu@3900 | 114 | _addr = other.addr(); |
zgu@3900 | 115 | return *this; |
zgu@3900 | 116 | } |
zgu@3900 | 117 | |
zgu@3900 | 118 | protected: |
zgu@3900 | 119 | inline void set_addr(address addr) { _addr = addr; } |
zgu@3900 | 120 | |
zgu@3900 | 121 | protected: |
zgu@3900 | 122 | // memory address |
zgu@3900 | 123 | address _addr; |
zgu@3900 | 124 | }; |
zgu@3900 | 125 | |
zgu@3900 | 126 | /* MemPointerRecord records an activityand associated |
zgu@3900 | 127 | * attributes on a memory block. |
zgu@3900 | 128 | */ |
zgu@3900 | 129 | class MemPointerRecord : public MemPointer { |
zgu@3900 | 130 | private: |
zgu@3900 | 131 | MEMFLAGS _flags; |
zgu@3900 | 132 | size_t _size; |
zgu@3900 | 133 | |
zgu@3900 | 134 | public: |
zgu@3900 | 135 | /* extension of MemoryType enum |
zgu@3900 | 136 | * see share/vm/memory/allocation.hpp for details. |
zgu@3900 | 137 | * |
zgu@3900 | 138 | * The tag values are associated to sorting orders, so be |
zgu@3900 | 139 | * careful if changes are needed. |
zgu@3900 | 140 | * The allocation records should be sorted ahead of tagging |
zgu@3900 | 141 | * records, which in turn ahead of deallocation records |
zgu@3900 | 142 | */ |
zgu@3900 | 143 | enum MemPointerTags { |
zgu@3900 | 144 | tag_alloc = 0x0001, // malloc or reserve record |
zgu@3900 | 145 | tag_commit = 0x0002, // commit record |
zgu@3900 | 146 | tag_type = 0x0003, // tag virtual memory to a memory type |
zgu@3900 | 147 | tag_uncommit = 0x0004, // uncommit record |
zgu@3900 | 148 | tag_release = 0x0005, // free or release record |
zgu@3900 | 149 | tag_size = 0x0006, // arena size |
zgu@3900 | 150 | tag_masks = 0x0007, // all tag bits |
zgu@3900 | 151 | vmBit = 0x0008 |
zgu@3900 | 152 | }; |
zgu@3900 | 153 | |
zgu@3900 | 154 | /* helper functions to interpret the tagging flags */ |
zgu@3900 | 155 | |
zgu@3900 | 156 | inline static bool is_allocation_record(MEMFLAGS flags) { |
zgu@3900 | 157 | return (flags & tag_masks) == tag_alloc; |
zgu@3900 | 158 | } |
zgu@3900 | 159 | |
zgu@3900 | 160 | inline static bool is_deallocation_record(MEMFLAGS flags) { |
zgu@3900 | 161 | return (flags & tag_masks) == tag_release; |
zgu@3900 | 162 | } |
zgu@3900 | 163 | |
zgu@3900 | 164 | inline static bool is_arena_record(MEMFLAGS flags) { |
zgu@3900 | 165 | return (flags & (otArena | tag_size)) == otArena; |
zgu@3900 | 166 | } |
zgu@3900 | 167 | |
zgu@3900 | 168 | inline static bool is_arena_size_record(MEMFLAGS flags) { |
zgu@3900 | 169 | return (flags & (otArena | tag_size)) == (otArena | tag_size); |
zgu@3900 | 170 | } |
zgu@3900 | 171 | |
zgu@3900 | 172 | inline static bool is_virtual_memory_record(MEMFLAGS flags) { |
zgu@3900 | 173 | return (flags & vmBit) != 0; |
zgu@3900 | 174 | } |
zgu@3900 | 175 | |
zgu@3900 | 176 | inline static bool is_virtual_memory_reserve_record(MEMFLAGS flags) { |
zgu@3900 | 177 | return (flags & 0x0F) == (tag_alloc | vmBit); |
zgu@3900 | 178 | } |
zgu@3900 | 179 | |
zgu@3900 | 180 | inline static bool is_virtual_memory_commit_record(MEMFLAGS flags) { |
zgu@3900 | 181 | return (flags & 0x0F) == (tag_commit | vmBit); |
zgu@3900 | 182 | } |
zgu@3900 | 183 | |
zgu@3900 | 184 | inline static bool is_virtual_memory_uncommit_record(MEMFLAGS flags) { |
zgu@3900 | 185 | return (flags & 0x0F) == (tag_uncommit | vmBit); |
zgu@3900 | 186 | } |
zgu@3900 | 187 | |
zgu@3900 | 188 | inline static bool is_virtual_memory_release_record(MEMFLAGS flags) { |
zgu@3900 | 189 | return (flags & 0x0F) == (tag_release | vmBit); |
zgu@3900 | 190 | } |
zgu@3900 | 191 | |
zgu@3900 | 192 | inline static bool is_virtual_memory_type_record(MEMFLAGS flags) { |
zgu@3900 | 193 | return (flags & 0x0F) == (tag_type | vmBit); |
zgu@3900 | 194 | } |
zgu@3900 | 195 | |
zgu@3900 | 196 | /* tagging flags */ |
zgu@3900 | 197 | inline static MEMFLAGS malloc_tag() { return tag_alloc; } |
zgu@3900 | 198 | inline static MEMFLAGS free_tag() { return tag_release; } |
zgu@3900 | 199 | inline static MEMFLAGS arena_size_tag() { return tag_size | otArena; } |
zgu@3900 | 200 | inline static MEMFLAGS virtual_memory_tag() { return vmBit; } |
zgu@3900 | 201 | inline static MEMFLAGS virtual_memory_reserve_tag() { return (tag_alloc | vmBit); } |
zgu@3900 | 202 | inline static MEMFLAGS virtual_memory_commit_tag() { return (tag_commit | vmBit); } |
zgu@3900 | 203 | inline static MEMFLAGS virtual_memory_uncommit_tag(){ return (tag_uncommit | vmBit); } |
zgu@3900 | 204 | inline static MEMFLAGS virtual_memory_release_tag() { return (tag_release | vmBit); } |
zgu@3900 | 205 | inline static MEMFLAGS virtual_memory_type_tag() { return (tag_type | vmBit); } |
zgu@3900 | 206 | |
zgu@3900 | 207 | public: |
zgu@3900 | 208 | MemPointerRecord(): _size(0), _flags(mtNone) { } |
zgu@3900 | 209 | |
zgu@3900 | 210 | MemPointerRecord(address addr, MEMFLAGS memflags, size_t size = 0): |
zgu@3900 | 211 | MemPointer(addr), _flags(memflags), _size(size) { } |
zgu@3900 | 212 | |
zgu@3900 | 213 | MemPointerRecord(const MemPointerRecord& copy_from): |
zgu@3900 | 214 | MemPointer(copy_from), _flags(copy_from.flags()), |
zgu@3900 | 215 | _size(copy_from.size()) { |
zgu@3900 | 216 | } |
zgu@3900 | 217 | |
zgu@3900 | 218 | /* MemPointerRecord is not sequenced, it always return |
zgu@3900 | 219 | * 0 to indicate non-sequenced |
zgu@3900 | 220 | */ |
zgu@3900 | 221 | virtual jint seq() const { return 0; } |
zgu@3900 | 222 | |
zgu@3900 | 223 | inline size_t size() const { return _size; } |
zgu@3900 | 224 | inline void set_size(size_t size) { _size = size; } |
zgu@3900 | 225 | |
zgu@3900 | 226 | inline MEMFLAGS flags() const { return _flags; } |
zgu@3900 | 227 | inline void set_flags(MEMFLAGS flags) { _flags = flags; } |
zgu@3900 | 228 | |
zgu@3900 | 229 | MemPointerRecord& operator= (const MemPointerRecord& ptr) { |
zgu@3900 | 230 | MemPointer::operator=(ptr); |
zgu@3900 | 231 | _flags = ptr.flags(); |
zgu@3900 | 232 | #ifdef ASSERT |
zgu@3900 | 233 | if (IS_ARENA_OBJ(_flags)) { |
zgu@3900 | 234 | assert(!is_vm_pointer(), "wrong flags"); |
zgu@3900 | 235 | assert((_flags & ot_masks) == otArena, "wrong flags"); |
zgu@3900 | 236 | } |
zgu@3900 | 237 | #endif |
zgu@3900 | 238 | _size = ptr.size(); |
zgu@3900 | 239 | return *this; |
zgu@3900 | 240 | } |
zgu@3900 | 241 | |
zgu@3900 | 242 | // if the pointer represents a malloc-ed memory address |
zgu@3900 | 243 | inline bool is_malloced_pointer() const { |
zgu@3900 | 244 | return !is_vm_pointer(); |
zgu@3900 | 245 | } |
zgu@3900 | 246 | |
zgu@3900 | 247 | // if the pointer represents a virtual memory address |
zgu@3900 | 248 | inline bool is_vm_pointer() const { |
zgu@3900 | 249 | return is_virtual_memory_record(_flags); |
zgu@3900 | 250 | } |
zgu@3900 | 251 | |
zgu@3900 | 252 | // if this record records a 'malloc' or virtual memory |
zgu@3900 | 253 | // 'reserve' call |
zgu@3900 | 254 | inline bool is_allocation_record() const { |
zgu@3900 | 255 | return is_allocation_record(_flags); |
zgu@3900 | 256 | } |
zgu@3900 | 257 | |
zgu@3900 | 258 | // if this record records a size information of an arena |
zgu@3900 | 259 | inline bool is_arena_size_record() const { |
zgu@3900 | 260 | return is_arena_size_record(_flags); |
zgu@3900 | 261 | } |
zgu@3900 | 262 | |
zgu@3900 | 263 | // if this pointer represents an address to an arena object |
zgu@3900 | 264 | inline bool is_arena_record() const { |
zgu@3900 | 265 | return is_arena_record(_flags); |
zgu@3900 | 266 | } |
zgu@3900 | 267 | |
zgu@3900 | 268 | // if this record represents a size information of specific arena |
zgu@3900 | 269 | inline bool is_size_record_of_arena(const MemPointerRecord* arena_rc) { |
zgu@3900 | 270 | assert(is_arena_size_record(), "not size record"); |
zgu@3900 | 271 | assert(arena_rc->is_arena_record(), "not arena record"); |
zgu@3900 | 272 | return (arena_rc->addr() + sizeof(void*)) == addr(); |
zgu@3900 | 273 | } |
zgu@3900 | 274 | |
zgu@3900 | 275 | // if this record records a 'free' or virtual memory 'free' call |
zgu@3900 | 276 | inline bool is_deallocation_record() const { |
zgu@3900 | 277 | return is_deallocation_record(_flags); |
zgu@3900 | 278 | } |
zgu@3900 | 279 | |
zgu@3900 | 280 | // if this record records a virtual memory 'commit' call |
zgu@3900 | 281 | inline bool is_commit_record() const { |
zgu@3900 | 282 | return is_virtual_memory_commit_record(_flags); |
zgu@3900 | 283 | } |
zgu@3900 | 284 | |
zgu@3900 | 285 | // if this record records a virtual memory 'uncommit' call |
zgu@3900 | 286 | inline bool is_uncommit_record() const { |
zgu@3900 | 287 | return is_virtual_memory_uncommit_record(_flags); |
zgu@3900 | 288 | } |
zgu@3900 | 289 | |
zgu@3900 | 290 | // if this record is a tagging record of a virtual memory block |
zgu@3900 | 291 | inline bool is_type_tagging_record() const { |
zgu@3900 | 292 | return is_virtual_memory_type_record(_flags); |
zgu@3900 | 293 | } |
zgu@3900 | 294 | }; |
zgu@3900 | 295 | |
zgu@3900 | 296 | // MemPointerRecordEx also records callsite pc, from where |
zgu@3900 | 297 | // the memory block is allocated |
zgu@3900 | 298 | class MemPointerRecordEx : public MemPointerRecord { |
zgu@3900 | 299 | private: |
zgu@3900 | 300 | address _pc; // callsite pc |
zgu@3900 | 301 | |
zgu@3900 | 302 | public: |
zgu@3900 | 303 | MemPointerRecordEx(): _pc(0) { } |
zgu@3900 | 304 | |
zgu@3900 | 305 | MemPointerRecordEx(address addr, MEMFLAGS memflags, size_t size = 0, address pc = 0): |
zgu@3900 | 306 | MemPointerRecord(addr, memflags, size), _pc(pc) {} |
zgu@3900 | 307 | |
zgu@3900 | 308 | MemPointerRecordEx(const MemPointerRecordEx& copy_from): |
zgu@3900 | 309 | MemPointerRecord(copy_from), _pc(copy_from.pc()) {} |
zgu@3900 | 310 | |
zgu@3900 | 311 | inline address pc() const { return _pc; } |
zgu@3900 | 312 | |
zgu@3900 | 313 | void init(const MemPointerRecordEx* mpe) { |
zgu@3900 | 314 | MemPointerRecord::operator=(*mpe); |
zgu@3900 | 315 | _pc = mpe->pc(); |
zgu@3900 | 316 | } |
zgu@3900 | 317 | |
zgu@3900 | 318 | void init(const MemPointerRecord* mp) { |
zgu@3900 | 319 | MemPointerRecord::operator=(*mp); |
zgu@3900 | 320 | _pc = 0; |
zgu@3900 | 321 | } |
zgu@3900 | 322 | }; |
zgu@3900 | 323 | |
zgu@3900 | 324 | // a virtual memory region |
zgu@3900 | 325 | class VMMemRegion : public MemPointerRecord { |
zgu@3900 | 326 | private: |
zgu@3900 | 327 | // committed size |
zgu@3900 | 328 | size_t _committed_size; |
zgu@3900 | 329 | |
zgu@3900 | 330 | public: |
zgu@3900 | 331 | VMMemRegion(): _committed_size(0) { } |
zgu@3900 | 332 | |
zgu@3900 | 333 | void init(const MemPointerRecord* mp) { |
zgu@3900 | 334 | assert(mp->is_vm_pointer(), "not virtual memory pointer"); |
zgu@3900 | 335 | _addr = mp->addr(); |
zgu@3900 | 336 | if (mp->is_commit_record() || mp->is_uncommit_record()) { |
zgu@3900 | 337 | _committed_size = mp->size(); |
zgu@3900 | 338 | set_size(_committed_size); |
zgu@3900 | 339 | } else { |
zgu@3900 | 340 | set_size(mp->size()); |
zgu@3900 | 341 | _committed_size = 0; |
zgu@3900 | 342 | } |
zgu@3900 | 343 | set_flags(mp->flags()); |
zgu@3900 | 344 | } |
zgu@3900 | 345 | |
zgu@3900 | 346 | VMMemRegion& operator=(const VMMemRegion& other) { |
zgu@3900 | 347 | MemPointerRecord::operator=(other); |
zgu@3900 | 348 | _committed_size = other.committed_size(); |
zgu@3900 | 349 | return *this; |
zgu@3900 | 350 | } |
zgu@3900 | 351 | |
zgu@3900 | 352 | inline bool is_reserve_record() const { |
zgu@3900 | 353 | return is_virtual_memory_reserve_record(flags()); |
zgu@3900 | 354 | } |
zgu@3900 | 355 | |
zgu@3900 | 356 | inline bool is_release_record() const { |
zgu@3900 | 357 | return is_virtual_memory_release_record(flags()); |
zgu@3900 | 358 | } |
zgu@3900 | 359 | |
zgu@3900 | 360 | // resize reserved VM range |
zgu@3900 | 361 | inline void set_reserved_size(size_t new_size) { |
zgu@3900 | 362 | assert(new_size >= committed_size(), "resize"); |
zgu@3900 | 363 | set_size(new_size); |
zgu@3900 | 364 | } |
zgu@3900 | 365 | |
zgu@3900 | 366 | inline void commit(size_t size) { |
zgu@3900 | 367 | _committed_size += size; |
zgu@3900 | 368 | } |
zgu@3900 | 369 | |
zgu@3900 | 370 | inline void uncommit(size_t size) { |
zgu@3900 | 371 | if (_committed_size >= size) { |
zgu@3900 | 372 | _committed_size -= size; |
zgu@3900 | 373 | } else { |
zgu@3900 | 374 | _committed_size = 0; |
zgu@3900 | 375 | } |
zgu@3900 | 376 | } |
zgu@3900 | 377 | |
zgu@3900 | 378 | /* |
zgu@3900 | 379 | * if this virtual memory range covers whole range of |
zgu@3900 | 380 | * the other VMMemRegion |
zgu@3900 | 381 | */ |
zgu@3900 | 382 | bool contains(const VMMemRegion* mr) const; |
zgu@3900 | 383 | |
zgu@3900 | 384 | /* base address of this virtual memory range */ |
zgu@3900 | 385 | inline address base() const { |
zgu@3900 | 386 | return addr(); |
zgu@3900 | 387 | } |
zgu@3900 | 388 | |
zgu@3900 | 389 | /* tag this virtual memory range to the specified memory type */ |
zgu@3900 | 390 | inline void tag(MEMFLAGS f) { |
zgu@3900 | 391 | set_flags(flags() | (f & mt_masks)); |
zgu@3900 | 392 | } |
zgu@3900 | 393 | |
zgu@3900 | 394 | // release part of memory range |
zgu@3900 | 395 | inline void partial_release(address add, size_t sz) { |
zgu@3900 | 396 | assert(add >= addr() && add < addr() + size(), "not valid address"); |
zgu@3900 | 397 | // for now, it can partially release from the both ends, |
zgu@3900 | 398 | // but not in the middle |
zgu@3900 | 399 | assert(add == addr() || (add + sz) == (addr() + size()), |
zgu@3900 | 400 | "release in the middle"); |
zgu@3900 | 401 | if (add == addr()) { |
zgu@3900 | 402 | set_addr(add + sz); |
zgu@3900 | 403 | set_size(size() - sz); |
zgu@3900 | 404 | } else { |
zgu@3900 | 405 | set_size(size() - sz); |
zgu@3900 | 406 | } |
zgu@3900 | 407 | } |
zgu@3900 | 408 | |
zgu@3900 | 409 | // the committed size of the virtual memory block |
zgu@3900 | 410 | inline size_t committed_size() const { |
zgu@3900 | 411 | return _committed_size; |
zgu@3900 | 412 | } |
zgu@3900 | 413 | |
zgu@3900 | 414 | // the reserved size of the virtual memory block |
zgu@3900 | 415 | inline size_t reserved_size() const { |
zgu@3900 | 416 | return size(); |
zgu@3900 | 417 | } |
zgu@3900 | 418 | }; |
zgu@3900 | 419 | |
zgu@3900 | 420 | class VMMemRegionEx : public VMMemRegion { |
zgu@3900 | 421 | private: |
zgu@3900 | 422 | jint _seq; // sequence number |
zgu@3900 | 423 | |
zgu@3900 | 424 | public: |
zgu@3900 | 425 | VMMemRegionEx(): _pc(0) { } |
zgu@3900 | 426 | |
zgu@3900 | 427 | void init(const MemPointerRecordEx* mpe) { |
zgu@3900 | 428 | VMMemRegion::init(mpe); |
zgu@3900 | 429 | _pc = mpe->pc(); |
zgu@3900 | 430 | } |
zgu@3900 | 431 | |
zgu@3900 | 432 | void init(const MemPointerRecord* mpe) { |
zgu@3900 | 433 | VMMemRegion::init(mpe); |
zgu@3900 | 434 | _pc = 0; |
zgu@3900 | 435 | } |
zgu@3900 | 436 | |
zgu@3900 | 437 | VMMemRegionEx& operator=(const VMMemRegionEx& other) { |
zgu@3900 | 438 | VMMemRegion::operator=(other); |
zgu@3900 | 439 | _pc = other.pc(); |
zgu@3900 | 440 | return *this; |
zgu@3900 | 441 | } |
zgu@3900 | 442 | |
zgu@3900 | 443 | inline address pc() const { return _pc; } |
zgu@3900 | 444 | private: |
zgu@3900 | 445 | address _pc; |
zgu@3900 | 446 | }; |
zgu@3900 | 447 | |
zgu@3900 | 448 | /* |
zgu@3900 | 449 | * Sequenced memory record |
zgu@3900 | 450 | */ |
zgu@3900 | 451 | class SeqMemPointerRecord : public MemPointerRecord { |
zgu@3900 | 452 | private: |
zgu@3900 | 453 | jint _seq; // sequence number |
zgu@3900 | 454 | |
zgu@3900 | 455 | public: |
zgu@3900 | 456 | SeqMemPointerRecord(): _seq(0){ } |
zgu@3900 | 457 | |
zgu@3900 | 458 | SeqMemPointerRecord(address addr, MEMFLAGS flags, size_t size) |
zgu@3900 | 459 | : MemPointerRecord(addr, flags, size) { |
zgu@3900 | 460 | _seq = SequenceGenerator::next(); |
zgu@3900 | 461 | } |
zgu@3900 | 462 | |
zgu@3900 | 463 | SeqMemPointerRecord(const SeqMemPointerRecord& copy_from) |
zgu@3900 | 464 | : MemPointerRecord(copy_from) { |
zgu@3900 | 465 | _seq = copy_from.seq(); |
zgu@3900 | 466 | } |
zgu@3900 | 467 | |
zgu@3900 | 468 | SeqMemPointerRecord& operator= (const SeqMemPointerRecord& ptr) { |
zgu@3900 | 469 | MemPointerRecord::operator=(ptr); |
zgu@3900 | 470 | _seq = ptr.seq(); |
zgu@3900 | 471 | return *this; |
zgu@3900 | 472 | } |
zgu@3900 | 473 | |
zgu@3900 | 474 | inline jint seq() const { |
zgu@3900 | 475 | return _seq; |
zgu@3900 | 476 | } |
zgu@3900 | 477 | }; |
zgu@3900 | 478 | |
zgu@3900 | 479 | |
zgu@3900 | 480 | |
zgu@3900 | 481 | class SeqMemPointerRecordEx : public MemPointerRecordEx { |
zgu@3900 | 482 | private: |
zgu@3900 | 483 | jint _seq; // sequence number |
zgu@3900 | 484 | |
zgu@3900 | 485 | public: |
zgu@3900 | 486 | SeqMemPointerRecordEx(): _seq(0) { } |
zgu@3900 | 487 | |
zgu@3900 | 488 | SeqMemPointerRecordEx(address addr, MEMFLAGS flags, size_t size, |
zgu@3900 | 489 | address pc): MemPointerRecordEx(addr, flags, size, pc) { |
zgu@3900 | 490 | _seq = SequenceGenerator::next(); |
zgu@3900 | 491 | } |
zgu@3900 | 492 | |
zgu@3900 | 493 | SeqMemPointerRecordEx(const SeqMemPointerRecordEx& copy_from) |
zgu@3900 | 494 | : MemPointerRecordEx(copy_from) { |
zgu@3900 | 495 | _seq = copy_from.seq(); |
zgu@3900 | 496 | } |
zgu@3900 | 497 | |
zgu@3900 | 498 | SeqMemPointerRecordEx& operator= (const SeqMemPointerRecordEx& ptr) { |
zgu@3900 | 499 | MemPointerRecordEx::operator=(ptr); |
zgu@3900 | 500 | _seq = ptr.seq(); |
zgu@3900 | 501 | return *this; |
zgu@3900 | 502 | } |
zgu@3900 | 503 | |
zgu@3900 | 504 | inline jint seq() const { |
zgu@3900 | 505 | return _seq; |
zgu@3900 | 506 | } |
zgu@3900 | 507 | }; |
zgu@3900 | 508 | |
zgu@3900 | 509 | #endif // SHARE_VM_SERVICES_MEM_PTR_HPP |