Fri, 01 Feb 2013 23:48:08 +0100
8005012: Add WB APIs to better support NMT testing
Summary: Add WB API functions to enable better NMT testing
Reviewed-by: dholmes, zgu
zgu@3900 | 1 | /* |
zgu@3900 | 2 | * Copyright (c) 2012, Oracle and/or its affiliates. All rights reserved. |
zgu@3900 | 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
zgu@3900 | 4 | * |
zgu@3900 | 5 | * This code is free software; you can redistribute it and/or modify it |
zgu@3900 | 6 | * under the terms of the GNU General Public License version 2 only, as |
zgu@3900 | 7 | * published by the Free Software Foundation. |
zgu@3900 | 8 | * |
zgu@3900 | 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
zgu@3900 | 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
zgu@3900 | 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
zgu@3900 | 12 | * version 2 for more details (a copy is included in the LICENSE file that |
zgu@3900 | 13 | * accompanied this code). |
zgu@3900 | 14 | * |
zgu@3900 | 15 | * You should have received a copy of the GNU General Public License version |
zgu@3900 | 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
zgu@3900 | 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
zgu@3900 | 18 | * |
zgu@3900 | 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
zgu@3900 | 20 | * or visit www.oracle.com if you need additional information or have any |
zgu@3900 | 21 | * questions. |
zgu@3900 | 22 | * |
zgu@3900 | 23 | */ |
zgu@3900 | 24 | #include "precompiled.hpp" |
zgu@3900 | 25 | #include "memory/allocation.hpp" |
zgu@3900 | 26 | #include "services/memBaseline.hpp" |
zgu@3900 | 27 | #include "services/memTracker.hpp" |
zgu@3900 | 28 | |
zgu@3900 | 29 | MemType2Name MemBaseline::MemType2NameMap[NUMBER_OF_MEMORY_TYPE] = { |
zgu@3900 | 30 | {mtJavaHeap, "Java Heap"}, |
zgu@3900 | 31 | {mtClass, "Class"}, |
zgu@3900 | 32 | {mtThreadStack,"Thread Stack"}, |
zgu@3900 | 33 | {mtThread, "Thread"}, |
zgu@3900 | 34 | {mtCode, "Code"}, |
zgu@3900 | 35 | {mtGC, "GC"}, |
zgu@3900 | 36 | {mtCompiler, "Compiler"}, |
zgu@3900 | 37 | {mtInternal, "Internal"}, |
zgu@3900 | 38 | {mtOther, "Other"}, |
zgu@3900 | 39 | {mtSymbol, "Symbol"}, |
zgu@3900 | 40 | {mtNMT, "Memory Tracking"}, |
zgu@3900 | 41 | {mtChunk, "Pooled Free Chunks"}, |
zgu@4193 | 42 | {mtClassShared,"Shared spaces for classes"}, |
ctornqvi@4512 | 43 | {mtTest, "Test"}, |
zgu@3900 | 44 | {mtNone, "Unknown"} // It can happen when type tagging records are lagging |
zgu@3900 | 45 | // behind |
zgu@3900 | 46 | }; |
zgu@3900 | 47 | |
zgu@3900 | 48 | MemBaseline::MemBaseline() { |
zgu@3900 | 49 | _baselined = false; |
zgu@3900 | 50 | |
zgu@3900 | 51 | for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) { |
zgu@3900 | 52 | _malloc_data[index].set_type(MemType2NameMap[index]._flag); |
zgu@3900 | 53 | _vm_data[index].set_type(MemType2NameMap[index]._flag); |
zgu@3900 | 54 | _arena_data[index].set_type(MemType2NameMap[index]._flag); |
zgu@3900 | 55 | } |
zgu@3900 | 56 | |
zgu@3900 | 57 | _malloc_cs = NULL; |
zgu@3900 | 58 | _vm_cs = NULL; |
zgu@4193 | 59 | _vm_map = NULL; |
zgu@3900 | 60 | |
zgu@3900 | 61 | _number_of_classes = 0; |
zgu@3900 | 62 | _number_of_threads = 0; |
zgu@3900 | 63 | } |
zgu@3900 | 64 | |
zgu@3900 | 65 | |
zgu@3900 | 66 | void MemBaseline::clear() { |
zgu@3900 | 67 | if (_malloc_cs != NULL) { |
zgu@3900 | 68 | delete _malloc_cs; |
zgu@3900 | 69 | _malloc_cs = NULL; |
zgu@3900 | 70 | } |
zgu@3900 | 71 | |
zgu@3900 | 72 | if (_vm_cs != NULL) { |
zgu@3900 | 73 | delete _vm_cs; |
zgu@3900 | 74 | _vm_cs = NULL; |
zgu@3900 | 75 | } |
zgu@3900 | 76 | |
zgu@4193 | 77 | if (_vm_map != NULL) { |
zgu@4193 | 78 | delete _vm_map; |
zgu@4193 | 79 | _vm_map = NULL; |
zgu@4193 | 80 | } |
zgu@4193 | 81 | |
zgu@3900 | 82 | reset(); |
zgu@3900 | 83 | } |
zgu@3900 | 84 | |
zgu@3900 | 85 | |
zgu@3900 | 86 | void MemBaseline::reset() { |
zgu@3900 | 87 | _baselined = false; |
zgu@3900 | 88 | _total_vm_reserved = 0; |
zgu@3900 | 89 | _total_vm_committed = 0; |
zgu@3900 | 90 | _total_malloced = 0; |
zgu@3900 | 91 | _number_of_classes = 0; |
zgu@3900 | 92 | |
zgu@3900 | 93 | if (_malloc_cs != NULL) _malloc_cs->clear(); |
zgu@3900 | 94 | if (_vm_cs != NULL) _vm_cs->clear(); |
zgu@4193 | 95 | if (_vm_map != NULL) _vm_map->clear(); |
zgu@3900 | 96 | |
zgu@3900 | 97 | for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) { |
zgu@3900 | 98 | _malloc_data[index].clear(); |
zgu@3900 | 99 | _vm_data[index].clear(); |
zgu@3900 | 100 | _arena_data[index].clear(); |
zgu@3900 | 101 | } |
zgu@3900 | 102 | } |
zgu@3900 | 103 | |
zgu@3900 | 104 | MemBaseline::~MemBaseline() { |
zgu@4193 | 105 | clear(); |
zgu@3900 | 106 | } |
zgu@3900 | 107 | |
zgu@3900 | 108 | // baseline malloc'd memory records, generate overall summary and summaries by |
zgu@3900 | 109 | // memory types |
zgu@3900 | 110 | bool MemBaseline::baseline_malloc_summary(const MemPointerArray* malloc_records) { |
zgu@4193 | 111 | MemPointerArrayIteratorImpl malloc_itr((MemPointerArray*)malloc_records); |
zgu@4193 | 112 | MemPointerRecord* malloc_ptr = (MemPointerRecord*)malloc_itr.current(); |
zgu@3900 | 113 | size_t used_arena_size = 0; |
zgu@3900 | 114 | int index; |
zgu@4193 | 115 | while (malloc_ptr != NULL) { |
zgu@4193 | 116 | index = flag2index(FLAGS_TO_MEMORY_TYPE(malloc_ptr->flags())); |
zgu@4193 | 117 | size_t size = malloc_ptr->size(); |
zgu@4274 | 118 | if (malloc_ptr->is_arena_memory_record()) { |
zgu@4274 | 119 | // We do have anonymous arenas, they are either used as value objects, |
zgu@4274 | 120 | // which are embedded inside other objects, or used as stack objects. |
zgu@4274 | 121 | _arena_data[index].inc(size); |
zgu@4274 | 122 | used_arena_size += size; |
zgu@4274 | 123 | } else { |
zgu@4274 | 124 | _total_malloced += size; |
zgu@4274 | 125 | _malloc_data[index].inc(size); |
zgu@4274 | 126 | if (malloc_ptr->is_arena_record()) { |
zgu@4274 | 127 | // see if arena memory record present |
zgu@4274 | 128 | MemPointerRecord* next_malloc_ptr = (MemPointerRecordEx*)malloc_itr.peek_next(); |
zgu@4274 | 129 | if (next_malloc_ptr->is_arena_memory_record()) { |
zgu@4274 | 130 | assert(next_malloc_ptr->is_memory_record_of_arena(malloc_ptr), |
zgu@4274 | 131 | "Arena records do not match"); |
zgu@4274 | 132 | size = next_malloc_ptr->size(); |
zgu@4274 | 133 | _arena_data[index].inc(size); |
zgu@4274 | 134 | used_arena_size += size; |
zgu@4274 | 135 | malloc_itr.next(); |
zgu@4274 | 136 | } |
zgu@3900 | 137 | } |
zgu@3900 | 138 | } |
zgu@4193 | 139 | malloc_ptr = (MemPointerRecordEx*)malloc_itr.next(); |
zgu@3900 | 140 | } |
zgu@3900 | 141 | |
zgu@3900 | 142 | // substract used arena size to get size of arena chunk in free list |
zgu@3900 | 143 | index = flag2index(mtChunk); |
zgu@3900 | 144 | _malloc_data[index].reduce(used_arena_size); |
zgu@3900 | 145 | // we really don't know how many chunks in free list, so just set to |
zgu@3900 | 146 | // 0 |
zgu@3900 | 147 | _malloc_data[index].overwrite_counter(0); |
zgu@3900 | 148 | |
zgu@3900 | 149 | return true; |
zgu@3900 | 150 | } |
zgu@3900 | 151 | |
zgu@3900 | 152 | // baseline mmap'd memory records, generate overall summary and summaries by |
zgu@3900 | 153 | // memory types |
zgu@3900 | 154 | bool MemBaseline::baseline_vm_summary(const MemPointerArray* vm_records) { |
zgu@4193 | 155 | MemPointerArrayIteratorImpl vm_itr((MemPointerArray*)vm_records); |
zgu@4193 | 156 | VMMemRegion* vm_ptr = (VMMemRegion*)vm_itr.current(); |
zgu@3900 | 157 | int index; |
zgu@4193 | 158 | while (vm_ptr != NULL) { |
zgu@4193 | 159 | if (vm_ptr->is_reserved_region()) { |
zgu@4193 | 160 | index = flag2index(FLAGS_TO_MEMORY_TYPE(vm_ptr->flags())); |
zgu@3900 | 161 | // we use the number of thread stack to count threads |
zgu@4193 | 162 | if (IS_MEMORY_TYPE(vm_ptr->flags(), mtThreadStack)) { |
zgu@3900 | 163 | _number_of_threads ++; |
zgu@3900 | 164 | } |
zgu@4193 | 165 | _total_vm_reserved += vm_ptr->size(); |
zgu@4193 | 166 | _vm_data[index].inc(vm_ptr->size(), 0); |
zgu@4193 | 167 | } else { |
zgu@4193 | 168 | _total_vm_committed += vm_ptr->size(); |
zgu@4193 | 169 | _vm_data[index].inc(0, vm_ptr->size()); |
zgu@4193 | 170 | } |
zgu@4193 | 171 | vm_ptr = (VMMemRegion*)vm_itr.next(); |
zgu@3900 | 172 | } |
zgu@3900 | 173 | return true; |
zgu@3900 | 174 | } |
zgu@3900 | 175 | |
zgu@3900 | 176 | // baseline malloc'd memory by callsites, but only the callsites with memory allocation |
zgu@3900 | 177 | // over 1KB are stored. |
zgu@3900 | 178 | bool MemBaseline::baseline_malloc_details(const MemPointerArray* malloc_records) { |
zgu@3900 | 179 | assert(MemTracker::track_callsite(), "detail tracking is off"); |
zgu@3900 | 180 | |
zgu@4193 | 181 | MemPointerArrayIteratorImpl malloc_itr(const_cast<MemPointerArray*>(malloc_records)); |
zgu@4193 | 182 | MemPointerRecordEx* malloc_ptr = (MemPointerRecordEx*)malloc_itr.current(); |
zgu@4193 | 183 | MallocCallsitePointer malloc_callsite; |
zgu@3900 | 184 | |
zgu@4193 | 185 | // initailize malloc callsite array |
zgu@3900 | 186 | if (_malloc_cs == NULL) { |
zgu@3900 | 187 | _malloc_cs = new (std::nothrow) MemPointerArrayImpl<MallocCallsitePointer>(64); |
zgu@3900 | 188 | // out of native memory |
zgu@4193 | 189 | if (_malloc_cs == NULL || _malloc_cs->out_of_memory()) { |
zgu@3900 | 190 | return false; |
zgu@3900 | 191 | } |
zgu@3900 | 192 | } else { |
zgu@3900 | 193 | _malloc_cs->clear(); |
zgu@3900 | 194 | } |
zgu@3900 | 195 | |
zgu@4193 | 196 | MemPointerArray* malloc_data = const_cast<MemPointerArray*>(malloc_records); |
zgu@4193 | 197 | |
zgu@4193 | 198 | // sort into callsite pc order. Details are aggregated by callsites |
zgu@4193 | 199 | malloc_data->sort((FN_SORT)malloc_sort_by_pc); |
zgu@4193 | 200 | bool ret = true; |
zgu@4193 | 201 | |
zgu@3900 | 202 | // baseline memory that is totaled over 1 KB |
zgu@4193 | 203 | while (malloc_ptr != NULL) { |
zgu@4274 | 204 | if (!MemPointerRecord::is_arena_memory_record(malloc_ptr->flags())) { |
zgu@3900 | 205 | // skip thread stacks |
zgu@4193 | 206 | if (!IS_MEMORY_TYPE(malloc_ptr->flags(), mtThreadStack)) { |
zgu@4193 | 207 | if (malloc_callsite.addr() != malloc_ptr->pc()) { |
zgu@4193 | 208 | if ((malloc_callsite.amount()/K) > 0) { |
zgu@4193 | 209 | if (!_malloc_cs->append(&malloc_callsite)) { |
zgu@4193 | 210 | ret = false; |
zgu@4193 | 211 | break; |
zgu@4193 | 212 | } |
zgu@4193 | 213 | } |
zgu@4193 | 214 | malloc_callsite = MallocCallsitePointer(malloc_ptr->pc()); |
zgu@4193 | 215 | } |
zgu@4193 | 216 | malloc_callsite.inc(malloc_ptr->size()); |
zgu@4193 | 217 | } |
zgu@4193 | 218 | } |
zgu@4193 | 219 | malloc_ptr = (MemPointerRecordEx*)malloc_itr.next(); |
zgu@4193 | 220 | } |
zgu@4193 | 221 | |
zgu@4193 | 222 | // restore to address order. Snapshot malloc data is maintained in memory |
zgu@4193 | 223 | // address order. |
zgu@4193 | 224 | malloc_data->sort((FN_SORT)malloc_sort_by_addr); |
zgu@4193 | 225 | |
zgu@4193 | 226 | if (!ret) { |
zgu@3900 | 227 | return false; |
zgu@3900 | 228 | } |
zgu@4193 | 229 | // deal with last record |
zgu@4193 | 230 | if (malloc_callsite.addr() != 0 && (malloc_callsite.amount()/K) > 0) { |
zgu@4193 | 231 | if (!_malloc_cs->append(&malloc_callsite)) { |
zgu@3900 | 232 | return false; |
zgu@3900 | 233 | } |
zgu@3900 | 234 | } |
zgu@3900 | 235 | return true; |
zgu@3900 | 236 | } |
zgu@3900 | 237 | |
zgu@3900 | 238 | // baseline mmap'd memory by callsites |
zgu@3900 | 239 | bool MemBaseline::baseline_vm_details(const MemPointerArray* vm_records) { |
zgu@3900 | 240 | assert(MemTracker::track_callsite(), "detail tracking is off"); |
zgu@3900 | 241 | |
zgu@4193 | 242 | VMCallsitePointer vm_callsite; |
zgu@4193 | 243 | VMCallsitePointer* cur_callsite = NULL; |
zgu@4193 | 244 | MemPointerArrayIteratorImpl vm_itr((MemPointerArray*)vm_records); |
zgu@4193 | 245 | VMMemRegionEx* vm_ptr = (VMMemRegionEx*)vm_itr.current(); |
zgu@3900 | 246 | |
zgu@4193 | 247 | // initialize virtual memory map array |
zgu@4193 | 248 | if (_vm_map == NULL) { |
zgu@4193 | 249 | _vm_map = new (std::nothrow) MemPointerArrayImpl<VMMemRegionEx>(vm_records->length()); |
zgu@4193 | 250 | if (_vm_map == NULL || _vm_map->out_of_memory()) { |
zgu@4193 | 251 | return false; |
zgu@4193 | 252 | } |
zgu@4193 | 253 | } else { |
zgu@4193 | 254 | _vm_map->clear(); |
zgu@4193 | 255 | } |
zgu@4193 | 256 | |
zgu@4193 | 257 | // initialize virtual memory callsite array |
zgu@3900 | 258 | if (_vm_cs == NULL) { |
zgu@3900 | 259 | _vm_cs = new (std::nothrow) MemPointerArrayImpl<VMCallsitePointer>(64); |
zgu@4193 | 260 | if (_vm_cs == NULL || _vm_cs->out_of_memory()) { |
zgu@3900 | 261 | return false; |
zgu@3900 | 262 | } |
zgu@3900 | 263 | } else { |
zgu@3900 | 264 | _vm_cs->clear(); |
zgu@3900 | 265 | } |
zgu@3900 | 266 | |
zgu@4193 | 267 | // consolidate virtual memory data |
zgu@4193 | 268 | VMMemRegionEx* reserved_rec = NULL; |
zgu@4193 | 269 | VMMemRegionEx* committed_rec = NULL; |
zgu@4193 | 270 | |
zgu@4193 | 271 | // vm_ptr is coming in increasing base address order |
zgu@4193 | 272 | while (vm_ptr != NULL) { |
zgu@4193 | 273 | if (vm_ptr->is_reserved_region()) { |
zgu@4193 | 274 | // consolidate reserved memory regions for virtual memory map. |
zgu@4193 | 275 | // The criteria for consolidation is: |
zgu@4193 | 276 | // 1. two adjacent reserved memory regions |
zgu@4193 | 277 | // 2. belong to the same memory type |
zgu@4193 | 278 | // 3. reserved from the same callsite |
zgu@4193 | 279 | if (reserved_rec == NULL || |
zgu@4193 | 280 | reserved_rec->base() + reserved_rec->size() != vm_ptr->addr() || |
zgu@4193 | 281 | FLAGS_TO_MEMORY_TYPE(reserved_rec->flags()) != FLAGS_TO_MEMORY_TYPE(vm_ptr->flags()) || |
zgu@4193 | 282 | reserved_rec->pc() != vm_ptr->pc()) { |
zgu@4193 | 283 | if (!_vm_map->append(vm_ptr)) { |
zgu@3900 | 284 | return false; |
zgu@3900 | 285 | } |
zgu@4193 | 286 | // inserted reserved region, we need the pointer to the element in virtual |
zgu@4193 | 287 | // memory map array. |
zgu@4193 | 288 | reserved_rec = (VMMemRegionEx*)_vm_map->at(_vm_map->length() - 1); |
zgu@4193 | 289 | } else { |
zgu@4193 | 290 | reserved_rec->expand_region(vm_ptr->addr(), vm_ptr->size()); |
zgu@3900 | 291 | } |
zgu@4193 | 292 | |
zgu@4193 | 293 | if (cur_callsite != NULL && !_vm_cs->append(cur_callsite)) { |
zgu@3900 | 294 | return false; |
zgu@3900 | 295 | } |
zgu@4193 | 296 | vm_callsite = VMCallsitePointer(vm_ptr->pc()); |
zgu@4193 | 297 | cur_callsite = &vm_callsite; |
zgu@4193 | 298 | vm_callsite.inc(vm_ptr->size(), 0); |
zgu@4193 | 299 | } else { |
zgu@4193 | 300 | // consolidate committed memory regions for virtual memory map |
zgu@4193 | 301 | // The criterial is: |
zgu@4193 | 302 | // 1. two adjacent committed memory regions |
zgu@4193 | 303 | // 2. committed from the same callsite |
zgu@4193 | 304 | if (committed_rec == NULL || |
zgu@4193 | 305 | committed_rec->base() + committed_rec->size() != vm_ptr->addr() || |
zgu@4193 | 306 | committed_rec->pc() != vm_ptr->pc()) { |
zgu@4193 | 307 | if (!_vm_map->append(vm_ptr)) { |
zgu@4193 | 308 | return false; |
zgu@3900 | 309 | } |
zgu@4193 | 310 | committed_rec = (VMMemRegionEx*)_vm_map->at(_vm_map->length() - 1); |
zgu@4193 | 311 | } else { |
zgu@4193 | 312 | committed_rec->expand_region(vm_ptr->addr(), vm_ptr->size()); |
zgu@4193 | 313 | } |
zgu@4193 | 314 | vm_callsite.inc(0, vm_ptr->size()); |
zgu@4193 | 315 | } |
zgu@4193 | 316 | vm_ptr = (VMMemRegionEx*)vm_itr.next(); |
zgu@4193 | 317 | } |
zgu@4193 | 318 | // deal with last record |
zgu@4193 | 319 | if (cur_callsite != NULL && !_vm_cs->append(cur_callsite)) { |
zgu@4193 | 320 | return false; |
zgu@4193 | 321 | } |
zgu@4193 | 322 | |
zgu@4193 | 323 | // sort it into callsite pc order. Details are aggregated by callsites |
zgu@4193 | 324 | _vm_cs->sort((FN_SORT)bl_vm_sort_by_pc); |
zgu@4193 | 325 | |
zgu@4193 | 326 | // walk the array to consolidate record by pc |
zgu@4193 | 327 | MemPointerArrayIteratorImpl itr(_vm_cs); |
zgu@4193 | 328 | VMCallsitePointer* callsite_rec = (VMCallsitePointer*)itr.current(); |
zgu@4193 | 329 | VMCallsitePointer* next_rec = (VMCallsitePointer*)itr.next(); |
zgu@4193 | 330 | while (next_rec != NULL) { |
zgu@4193 | 331 | assert(callsite_rec != NULL, "Sanity check"); |
zgu@4193 | 332 | if (next_rec->addr() == callsite_rec->addr()) { |
zgu@4193 | 333 | callsite_rec->inc(next_rec->reserved_amount(), next_rec->committed_amount()); |
zgu@4193 | 334 | itr.remove(); |
zgu@4193 | 335 | next_rec = (VMCallsitePointer*)itr.current(); |
zgu@4193 | 336 | } else { |
zgu@4193 | 337 | callsite_rec = next_rec; |
zgu@4193 | 338 | next_rec = (VMCallsitePointer*)itr.next(); |
zgu@4193 | 339 | } |
zgu@4193 | 340 | } |
zgu@4193 | 341 | |
zgu@3900 | 342 | return true; |
zgu@3900 | 343 | } |
zgu@3900 | 344 | |
zgu@3900 | 345 | // baseline a snapshot. If summary_only = false, memory usages aggregated by |
zgu@3900 | 346 | // callsites are also baselined. |
zgu@3900 | 347 | bool MemBaseline::baseline(MemSnapshot& snapshot, bool summary_only) { |
zgu@3900 | 348 | MutexLockerEx snapshot_locker(snapshot._lock, true); |
zgu@3900 | 349 | reset(); |
zgu@3900 | 350 | _baselined = baseline_malloc_summary(snapshot._alloc_ptrs) && |
zgu@3900 | 351 | baseline_vm_summary(snapshot._vm_ptrs); |
zgu@4400 | 352 | _number_of_classes = snapshot.number_of_classes(); |
zgu@3900 | 353 | |
zgu@3900 | 354 | if (!summary_only && MemTracker::track_callsite() && _baselined) { |
zgu@3900 | 355 | _baselined = baseline_malloc_details(snapshot._alloc_ptrs) && |
zgu@3900 | 356 | baseline_vm_details(snapshot._vm_ptrs); |
zgu@3900 | 357 | } |
zgu@3900 | 358 | return _baselined; |
zgu@3900 | 359 | } |
zgu@3900 | 360 | |
zgu@3900 | 361 | |
zgu@3900 | 362 | int MemBaseline::flag2index(MEMFLAGS flag) const { |
zgu@3900 | 363 | for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) { |
zgu@3900 | 364 | if (MemType2NameMap[index]._flag == flag) { |
zgu@3900 | 365 | return index; |
zgu@3900 | 366 | } |
zgu@3900 | 367 | } |
zgu@3900 | 368 | assert(false, "no type"); |
zgu@3900 | 369 | return -1; |
zgu@3900 | 370 | } |
zgu@3900 | 371 | |
zgu@3900 | 372 | const char* MemBaseline::type2name(MEMFLAGS type) { |
zgu@3900 | 373 | for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) { |
zgu@3900 | 374 | if (MemType2NameMap[index]._flag == type) { |
zgu@3900 | 375 | return MemType2NameMap[index]._name; |
zgu@3900 | 376 | } |
zgu@3900 | 377 | } |
zgu@4193 | 378 | assert(false, err_msg("bad type %x", type)); |
zgu@3900 | 379 | return NULL; |
zgu@3900 | 380 | } |
zgu@3900 | 381 | |
zgu@3900 | 382 | |
zgu@3900 | 383 | MemBaseline& MemBaseline::operator=(const MemBaseline& other) { |
zgu@3900 | 384 | _total_malloced = other._total_malloced; |
zgu@3900 | 385 | _total_vm_reserved = other._total_vm_reserved; |
zgu@3900 | 386 | _total_vm_committed = other._total_vm_committed; |
zgu@3900 | 387 | |
zgu@3900 | 388 | _baselined = other._baselined; |
zgu@3900 | 389 | _number_of_classes = other._number_of_classes; |
zgu@3900 | 390 | |
zgu@3900 | 391 | for (int index = 0; index < NUMBER_OF_MEMORY_TYPE; index ++) { |
zgu@3900 | 392 | _malloc_data[index] = other._malloc_data[index]; |
zgu@3900 | 393 | _vm_data[index] = other._vm_data[index]; |
zgu@3900 | 394 | _arena_data[index] = other._arena_data[index]; |
zgu@3900 | 395 | } |
zgu@3900 | 396 | |
zgu@3900 | 397 | if (MemTracker::track_callsite()) { |
zgu@3900 | 398 | assert(_malloc_cs != NULL && _vm_cs != NULL, "out of memory"); |
zgu@3900 | 399 | assert(other._malloc_cs != NULL && other._vm_cs != NULL, |
zgu@3900 | 400 | "not properly baselined"); |
zgu@3900 | 401 | _malloc_cs->clear(); |
zgu@3900 | 402 | _vm_cs->clear(); |
zgu@3900 | 403 | int index; |
zgu@3900 | 404 | for (index = 0; index < other._malloc_cs->length(); index ++) { |
zgu@3900 | 405 | _malloc_cs->append(other._malloc_cs->at(index)); |
zgu@3900 | 406 | } |
zgu@3900 | 407 | |
zgu@3900 | 408 | for (index = 0; index < other._vm_cs->length(); index ++) { |
zgu@3900 | 409 | _vm_cs->append(other._vm_cs->at(index)); |
zgu@3900 | 410 | } |
zgu@3900 | 411 | } |
zgu@3900 | 412 | return *this; |
zgu@3900 | 413 | } |
zgu@3900 | 414 | |
zgu@3900 | 415 | /* compare functions for sorting */ |
zgu@3900 | 416 | |
zgu@3900 | 417 | // sort snapshot malloc'd records in callsite pc order |
zgu@3900 | 418 | int MemBaseline::malloc_sort_by_pc(const void* p1, const void* p2) { |
zgu@3900 | 419 | assert(MemTracker::track_callsite(),"Just check"); |
zgu@3900 | 420 | const MemPointerRecordEx* mp1 = (const MemPointerRecordEx*)p1; |
zgu@3900 | 421 | const MemPointerRecordEx* mp2 = (const MemPointerRecordEx*)p2; |
zgu@3900 | 422 | return UNSIGNED_COMPARE(mp1->pc(), mp2->pc()); |
zgu@3900 | 423 | } |
zgu@3900 | 424 | |
zgu@3900 | 425 | // sort baselined malloc'd records in size order |
zgu@3900 | 426 | int MemBaseline::bl_malloc_sort_by_size(const void* p1, const void* p2) { |
zgu@3900 | 427 | assert(MemTracker::is_on(), "Just check"); |
zgu@3900 | 428 | const MallocCallsitePointer* mp1 = (const MallocCallsitePointer*)p1; |
zgu@3900 | 429 | const MallocCallsitePointer* mp2 = (const MallocCallsitePointer*)p2; |
zgu@3900 | 430 | return UNSIGNED_COMPARE(mp2->amount(), mp1->amount()); |
zgu@3900 | 431 | } |
zgu@3900 | 432 | |
zgu@3900 | 433 | // sort baselined malloc'd records in callsite pc order |
zgu@3900 | 434 | int MemBaseline::bl_malloc_sort_by_pc(const void* p1, const void* p2) { |
zgu@3900 | 435 | assert(MemTracker::is_on(), "Just check"); |
zgu@3900 | 436 | const MallocCallsitePointer* mp1 = (const MallocCallsitePointer*)p1; |
zgu@3900 | 437 | const MallocCallsitePointer* mp2 = (const MallocCallsitePointer*)p2; |
zgu@3900 | 438 | return UNSIGNED_COMPARE(mp1->addr(), mp2->addr()); |
zgu@3900 | 439 | } |
zgu@3900 | 440 | |
zgu@3900 | 441 | |
zgu@3900 | 442 | // sort baselined mmap'd records in size (reserved size) order |
zgu@3900 | 443 | int MemBaseline::bl_vm_sort_by_size(const void* p1, const void* p2) { |
zgu@3900 | 444 | assert(MemTracker::is_on(), "Just check"); |
zgu@3900 | 445 | const VMCallsitePointer* mp1 = (const VMCallsitePointer*)p1; |
zgu@3900 | 446 | const VMCallsitePointer* mp2 = (const VMCallsitePointer*)p2; |
zgu@3900 | 447 | return UNSIGNED_COMPARE(mp2->reserved_amount(), mp1->reserved_amount()); |
zgu@3900 | 448 | } |
zgu@3900 | 449 | |
zgu@3900 | 450 | // sort baselined mmap'd records in callsite pc order |
zgu@3900 | 451 | int MemBaseline::bl_vm_sort_by_pc(const void* p1, const void* p2) { |
zgu@3900 | 452 | assert(MemTracker::is_on(), "Just check"); |
zgu@3900 | 453 | const VMCallsitePointer* mp1 = (const VMCallsitePointer*)p1; |
zgu@3900 | 454 | const VMCallsitePointer* mp2 = (const VMCallsitePointer*)p2; |
zgu@3900 | 455 | return UNSIGNED_COMPARE(mp1->addr(), mp2->addr()); |
zgu@3900 | 456 | } |
zgu@3900 | 457 | |
zgu@3900 | 458 | |
zgu@3900 | 459 | // sort snapshot malloc'd records in memory block address order |
zgu@3900 | 460 | int MemBaseline::malloc_sort_by_addr(const void* p1, const void* p2) { |
zgu@3900 | 461 | assert(MemTracker::is_on(), "Just check"); |
zgu@3900 | 462 | const MemPointerRecord* mp1 = (const MemPointerRecord*)p1; |
zgu@3900 | 463 | const MemPointerRecord* mp2 = (const MemPointerRecord*)p2; |
zgu@3900 | 464 | int delta = UNSIGNED_COMPARE(mp1->addr(), mp2->addr()); |
zgu@3900 | 465 | assert(delta != 0, "dup pointer"); |
zgu@3900 | 466 | return delta; |
zgu@3900 | 467 | } |
zgu@3900 | 468 |