Thu, 13 Jun 2013 22:02:40 -0700
8014431: cleanup warnings indicated by the -Wunused-value compiler option on linux
Reviewed-by: dholmes, coleenp
Contributed-by: jeremymanson@google.com, calvin.cheung@oracle.com
duke@435 | 1 | /* |
mikael@4153 | 2 | * Copyright (c) 1997, 2012, Oracle and/or its affiliates. All rights reserved. |
duke@435 | 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
duke@435 | 4 | * |
duke@435 | 5 | * This code is free software; you can redistribute it and/or modify it |
duke@435 | 6 | * under the terms of the GNU General Public License version 2 only, as |
duke@435 | 7 | * published by the Free Software Foundation. |
duke@435 | 8 | * |
duke@435 | 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
duke@435 | 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
duke@435 | 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
duke@435 | 12 | * version 2 for more details (a copy is included in the LICENSE file that |
duke@435 | 13 | * accompanied this code). |
duke@435 | 14 | * |
duke@435 | 15 | * You should have received a copy of the GNU General Public License version |
duke@435 | 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
duke@435 | 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
duke@435 | 18 | * |
trims@1907 | 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
trims@1907 | 20 | * or visit www.oracle.com if you need additional information or have any |
trims@1907 | 21 | * questions. |
duke@435 | 22 | * |
duke@435 | 23 | */ |
duke@435 | 24 | |
stefank@2314 | 25 | #include "precompiled.hpp" |
stefank@2314 | 26 | #include "memory/heap.hpp" |
stefank@2314 | 27 | #include "oops/oop.inline.hpp" |
stefank@2314 | 28 | #include "runtime/os.hpp" |
zgu@3900 | 29 | #include "services/memTracker.hpp" |
duke@435 | 30 | |
duke@435 | 31 | size_t CodeHeap::header_size() { |
duke@435 | 32 | return sizeof(HeapBlock); |
duke@435 | 33 | } |
duke@435 | 34 | |
duke@435 | 35 | |
duke@435 | 36 | // Implementation of Heap |
duke@435 | 37 | |
duke@435 | 38 | CodeHeap::CodeHeap() { |
duke@435 | 39 | _number_of_committed_segments = 0; |
duke@435 | 40 | _number_of_reserved_segments = 0; |
duke@435 | 41 | _segment_size = 0; |
duke@435 | 42 | _log2_segment_size = 0; |
duke@435 | 43 | _next_segment = 0; |
duke@435 | 44 | _freelist = NULL; |
neliasso@4952 | 45 | _freelist_segments = 0; |
duke@435 | 46 | } |
duke@435 | 47 | |
duke@435 | 48 | |
duke@435 | 49 | void CodeHeap::mark_segmap_as_free(size_t beg, size_t end) { |
duke@435 | 50 | assert(0 <= beg && beg < _number_of_committed_segments, "interval begin out of bounds"); |
duke@435 | 51 | assert(beg < end && end <= _number_of_committed_segments, "interval end out of bounds"); |
duke@435 | 52 | // setup _segmap pointers for faster indexing |
duke@435 | 53 | address p = (address)_segmap.low() + beg; |
duke@435 | 54 | address q = (address)_segmap.low() + end; |
duke@435 | 55 | // initialize interval |
duke@435 | 56 | while (p < q) *p++ = 0xFF; |
duke@435 | 57 | } |
duke@435 | 58 | |
duke@435 | 59 | |
duke@435 | 60 | void CodeHeap::mark_segmap_as_used(size_t beg, size_t end) { |
duke@435 | 61 | assert(0 <= beg && beg < _number_of_committed_segments, "interval begin out of bounds"); |
duke@435 | 62 | assert(beg < end && end <= _number_of_committed_segments, "interval end out of bounds"); |
duke@435 | 63 | // setup _segmap pointers for faster indexing |
duke@435 | 64 | address p = (address)_segmap.low() + beg; |
duke@435 | 65 | address q = (address)_segmap.low() + end; |
duke@435 | 66 | // initialize interval |
duke@435 | 67 | int i = 0; |
duke@435 | 68 | while (p < q) { |
duke@435 | 69 | *p++ = i++; |
duke@435 | 70 | if (i == 0xFF) i = 1; |
duke@435 | 71 | } |
duke@435 | 72 | } |
duke@435 | 73 | |
duke@435 | 74 | |
duke@435 | 75 | static size_t align_to_page_size(size_t size) { |
duke@435 | 76 | const size_t alignment = (size_t)os::vm_page_size(); |
duke@435 | 77 | assert(is_power_of_2(alignment), "no kidding ???"); |
duke@435 | 78 | return (size + alignment - 1) & ~(alignment - 1); |
duke@435 | 79 | } |
duke@435 | 80 | |
duke@435 | 81 | |
duke@435 | 82 | void CodeHeap::on_code_mapping(char* base, size_t size) { |
duke@435 | 83 | #ifdef LINUX |
duke@435 | 84 | extern void linux_wrap_code(char* base, size_t size); |
duke@435 | 85 | linux_wrap_code(base, size); |
duke@435 | 86 | #endif |
duke@435 | 87 | } |
duke@435 | 88 | |
duke@435 | 89 | |
duke@435 | 90 | bool CodeHeap::reserve(size_t reserved_size, size_t committed_size, |
duke@435 | 91 | size_t segment_size) { |
duke@435 | 92 | assert(reserved_size >= committed_size, "reserved < committed"); |
duke@435 | 93 | assert(segment_size >= sizeof(FreeBlock), "segment size is too small"); |
duke@435 | 94 | assert(is_power_of_2(segment_size), "segment_size must be a power of 2"); |
duke@435 | 95 | |
duke@435 | 96 | _segment_size = segment_size; |
duke@435 | 97 | _log2_segment_size = exact_log2(segment_size); |
duke@435 | 98 | |
duke@435 | 99 | // Reserve and initialize space for _memory. |
jcoomes@514 | 100 | const size_t page_size = os::can_execute_large_page_memory() ? |
jcoomes@514 | 101 | os::page_size_for_region(committed_size, reserved_size, 8) : |
jcoomes@514 | 102 | os::vm_page_size(); |
duke@435 | 103 | const size_t granularity = os::vm_allocation_granularity(); |
duke@435 | 104 | const size_t r_align = MAX2(page_size, granularity); |
duke@435 | 105 | const size_t r_size = align_size_up(reserved_size, r_align); |
duke@435 | 106 | const size_t c_size = align_size_up(committed_size, page_size); |
duke@435 | 107 | |
duke@435 | 108 | const size_t rs_align = page_size == (size_t) os::vm_page_size() ? 0 : |
duke@435 | 109 | MAX2(page_size, granularity); |
coleenp@1091 | 110 | ReservedCodeSpace rs(r_size, rs_align, rs_align > 0); |
duke@435 | 111 | os::trace_page_sizes("code heap", committed_size, reserved_size, page_size, |
duke@435 | 112 | rs.base(), rs.size()); |
duke@435 | 113 | if (!_memory.initialize(rs, c_size)) { |
duke@435 | 114 | return false; |
duke@435 | 115 | } |
duke@435 | 116 | |
duke@435 | 117 | on_code_mapping(_memory.low(), _memory.committed_size()); |
neliasso@4952 | 118 | _number_of_committed_segments = size_to_segments(_memory.committed_size()); |
neliasso@4952 | 119 | _number_of_reserved_segments = size_to_segments(_memory.reserved_size()); |
duke@435 | 120 | assert(_number_of_reserved_segments >= _number_of_committed_segments, "just checking"); |
duke@435 | 121 | |
duke@435 | 122 | // reserve space for _segmap |
duke@435 | 123 | if (!_segmap.initialize(align_to_page_size(_number_of_reserved_segments), align_to_page_size(_number_of_committed_segments))) { |
duke@435 | 124 | return false; |
duke@435 | 125 | } |
zgu@3900 | 126 | |
zgu@3900 | 127 | MemTracker::record_virtual_memory_type((address)_segmap.low_boundary(), mtCode); |
zgu@3900 | 128 | |
duke@435 | 129 | assert(_segmap.committed_size() >= (size_t) _number_of_committed_segments, "could not commit enough space for segment map"); |
duke@435 | 130 | assert(_segmap.reserved_size() >= (size_t) _number_of_reserved_segments , "could not reserve enough space for segment map"); |
duke@435 | 131 | assert(_segmap.reserved_size() >= _segmap.committed_size() , "just checking"); |
duke@435 | 132 | |
duke@435 | 133 | // initialize remaining instance variables |
duke@435 | 134 | clear(); |
duke@435 | 135 | return true; |
duke@435 | 136 | } |
duke@435 | 137 | |
duke@435 | 138 | |
duke@435 | 139 | void CodeHeap::release() { |
duke@435 | 140 | Unimplemented(); |
duke@435 | 141 | } |
duke@435 | 142 | |
duke@435 | 143 | |
duke@435 | 144 | bool CodeHeap::expand_by(size_t size) { |
duke@435 | 145 | // expand _memory space |
duke@435 | 146 | size_t dm = align_to_page_size(_memory.committed_size() + size) - _memory.committed_size(); |
duke@435 | 147 | if (dm > 0) { |
duke@435 | 148 | char* base = _memory.low() + _memory.committed_size(); |
duke@435 | 149 | if (!_memory.expand_by(dm)) return false; |
duke@435 | 150 | on_code_mapping(base, dm); |
duke@435 | 151 | size_t i = _number_of_committed_segments; |
neliasso@4952 | 152 | _number_of_committed_segments = size_to_segments(_memory.committed_size()); |
neliasso@4952 | 153 | assert(_number_of_reserved_segments == size_to_segments(_memory.reserved_size()), "number of reserved segments should not change"); |
duke@435 | 154 | assert(_number_of_reserved_segments >= _number_of_committed_segments, "just checking"); |
duke@435 | 155 | // expand _segmap space |
duke@435 | 156 | size_t ds = align_to_page_size(_number_of_committed_segments) - _segmap.committed_size(); |
duke@435 | 157 | if (ds > 0) { |
duke@435 | 158 | if (!_segmap.expand_by(ds)) return false; |
duke@435 | 159 | } |
duke@435 | 160 | assert(_segmap.committed_size() >= (size_t) _number_of_committed_segments, "just checking"); |
duke@435 | 161 | // initialize additional segmap entries |
duke@435 | 162 | mark_segmap_as_free(i, _number_of_committed_segments); |
duke@435 | 163 | } |
duke@435 | 164 | return true; |
duke@435 | 165 | } |
duke@435 | 166 | |
duke@435 | 167 | |
duke@435 | 168 | void CodeHeap::shrink_by(size_t size) { |
duke@435 | 169 | Unimplemented(); |
duke@435 | 170 | } |
duke@435 | 171 | |
duke@435 | 172 | |
duke@435 | 173 | void CodeHeap::clear() { |
duke@435 | 174 | _next_segment = 0; |
duke@435 | 175 | mark_segmap_as_free(0, _number_of_committed_segments); |
duke@435 | 176 | } |
duke@435 | 177 | |
duke@435 | 178 | |
neliasso@4952 | 179 | void* CodeHeap::allocate(size_t instance_size, bool is_critical) { |
neliasso@4952 | 180 | size_t number_of_segments = size_to_segments(instance_size + sizeof(HeapBlock)); |
neliasso@4952 | 181 | assert(segments_to_size(number_of_segments) >= sizeof(FreeBlock), "not enough room for FreeList"); |
duke@435 | 182 | |
duke@435 | 183 | // First check if we can satify request from freelist |
duke@435 | 184 | debug_only(verify()); |
neliasso@4952 | 185 | HeapBlock* block = search_freelist(number_of_segments, is_critical); |
duke@435 | 186 | debug_only(if (VerifyCodeCacheOften) verify()); |
duke@435 | 187 | if (block != NULL) { |
neliasso@4952 | 188 | assert(block->length() >= number_of_segments && block->length() < number_of_segments + CodeCacheMinBlockLength, "sanity check"); |
duke@435 | 189 | assert(!block->free(), "must be marked free"); |
duke@435 | 190 | #ifdef ASSERT |
neliasso@4952 | 191 | memset((void *)block->allocated_space(), badCodeHeapNewVal, instance_size); |
duke@435 | 192 | #endif |
duke@435 | 193 | return block->allocated_space(); |
duke@435 | 194 | } |
duke@435 | 195 | |
neliasso@4952 | 196 | // Ensure minimum size for allocation to the heap. |
neliasso@4952 | 197 | if (number_of_segments < CodeCacheMinBlockLength) { |
neliasso@4952 | 198 | number_of_segments = CodeCacheMinBlockLength; |
duke@435 | 199 | } |
neliasso@4952 | 200 | |
neliasso@4952 | 201 | if (!is_critical) { |
neliasso@4952 | 202 | // Make sure the allocation fits in the unallocated heap without using |
neliasso@4952 | 203 | // the CodeCacheMimimumFreeSpace that is reserved for critical allocations. |
neliasso@4952 | 204 | if (segments_to_size(number_of_segments) > (heap_unallocated_capacity() - CodeCacheMinimumFreeSpace)) { |
neliasso@4952 | 205 | // Fail allocation |
neliasso@4952 | 206 | return NULL; |
neliasso@4952 | 207 | } |
neliasso@4952 | 208 | } |
neliasso@4952 | 209 | |
neliasso@4952 | 210 | if (_next_segment + number_of_segments <= _number_of_committed_segments) { |
neliasso@4952 | 211 | mark_segmap_as_used(_next_segment, _next_segment + number_of_segments); |
duke@435 | 212 | HeapBlock* b = block_at(_next_segment); |
neliasso@4952 | 213 | b->initialize(number_of_segments); |
neliasso@4952 | 214 | _next_segment += number_of_segments; |
duke@435 | 215 | #ifdef ASSERT |
neliasso@4952 | 216 | memset((void *)b->allocated_space(), badCodeHeapNewVal, instance_size); |
duke@435 | 217 | #endif |
duke@435 | 218 | return b->allocated_space(); |
duke@435 | 219 | } else { |
duke@435 | 220 | return NULL; |
duke@435 | 221 | } |
duke@435 | 222 | } |
duke@435 | 223 | |
duke@435 | 224 | |
duke@435 | 225 | void CodeHeap::deallocate(void* p) { |
duke@435 | 226 | assert(p == find_start(p), "illegal deallocation"); |
duke@435 | 227 | // Find start of HeapBlock |
duke@435 | 228 | HeapBlock* b = (((HeapBlock *)p) - 1); |
duke@435 | 229 | assert(b->allocated_space() == p, "sanity check"); |
duke@435 | 230 | #ifdef ASSERT |
duke@435 | 231 | memset((void *)b->allocated_space(), |
duke@435 | 232 | badCodeHeapFreeVal, |
neliasso@4952 | 233 | segments_to_size(b->length()) - sizeof(HeapBlock)); |
duke@435 | 234 | #endif |
duke@435 | 235 | add_to_freelist(b); |
duke@435 | 236 | |
duke@435 | 237 | debug_only(if (VerifyCodeCacheOften) verify()); |
duke@435 | 238 | } |
duke@435 | 239 | |
duke@435 | 240 | |
duke@435 | 241 | void* CodeHeap::find_start(void* p) const { |
duke@435 | 242 | if (!contains(p)) { |
duke@435 | 243 | return NULL; |
duke@435 | 244 | } |
duke@435 | 245 | size_t i = segment_for(p); |
duke@435 | 246 | address b = (address)_segmap.low(); |
duke@435 | 247 | if (b[i] == 0xFF) { |
duke@435 | 248 | return NULL; |
duke@435 | 249 | } |
duke@435 | 250 | while (b[i] > 0) i -= (int)b[i]; |
duke@435 | 251 | HeapBlock* h = block_at(i); |
duke@435 | 252 | if (h->free()) { |
duke@435 | 253 | return NULL; |
duke@435 | 254 | } |
duke@435 | 255 | return h->allocated_space(); |
duke@435 | 256 | } |
duke@435 | 257 | |
duke@435 | 258 | |
duke@435 | 259 | size_t CodeHeap::alignment_unit() const { |
duke@435 | 260 | // this will be a power of two |
duke@435 | 261 | return _segment_size; |
duke@435 | 262 | } |
duke@435 | 263 | |
duke@435 | 264 | |
duke@435 | 265 | size_t CodeHeap::alignment_offset() const { |
duke@435 | 266 | // The lowest address in any allocated block will be |
duke@435 | 267 | // equal to alignment_offset (mod alignment_unit). |
duke@435 | 268 | return sizeof(HeapBlock) & (_segment_size - 1); |
duke@435 | 269 | } |
duke@435 | 270 | |
duke@435 | 271 | // Finds the next free heapblock. If the current one is free, that it returned |
duke@435 | 272 | void* CodeHeap::next_free(HeapBlock *b) const { |
duke@435 | 273 | // Since free blocks are merged, there is max. on free block |
duke@435 | 274 | // between two used ones |
duke@435 | 275 | if (b != NULL && b->free()) b = next_block(b); |
duke@435 | 276 | assert(b == NULL || !b->free(), "must be in use or at end of heap"); |
duke@435 | 277 | return (b == NULL) ? NULL : b->allocated_space(); |
duke@435 | 278 | } |
duke@435 | 279 | |
duke@435 | 280 | // Returns the first used HeapBlock |
duke@435 | 281 | HeapBlock* CodeHeap::first_block() const { |
duke@435 | 282 | if (_next_segment > 0) |
duke@435 | 283 | return block_at(0); |
duke@435 | 284 | return NULL; |
duke@435 | 285 | } |
duke@435 | 286 | |
duke@435 | 287 | HeapBlock *CodeHeap::block_start(void *q) const { |
duke@435 | 288 | HeapBlock* b = (HeapBlock*)find_start(q); |
duke@435 | 289 | if (b == NULL) return NULL; |
duke@435 | 290 | return b - 1; |
duke@435 | 291 | } |
duke@435 | 292 | |
duke@435 | 293 | // Returns the next Heap block an offset into one |
duke@435 | 294 | HeapBlock* CodeHeap::next_block(HeapBlock *b) const { |
duke@435 | 295 | if (b == NULL) return NULL; |
duke@435 | 296 | size_t i = segment_for(b) + b->length(); |
duke@435 | 297 | if (i < _next_segment) |
duke@435 | 298 | return block_at(i); |
duke@435 | 299 | return NULL; |
duke@435 | 300 | } |
duke@435 | 301 | |
duke@435 | 302 | |
duke@435 | 303 | // Returns current capacity |
duke@435 | 304 | size_t CodeHeap::capacity() const { |
duke@435 | 305 | return _memory.committed_size(); |
duke@435 | 306 | } |
duke@435 | 307 | |
duke@435 | 308 | size_t CodeHeap::max_capacity() const { |
duke@435 | 309 | return _memory.reserved_size(); |
duke@435 | 310 | } |
duke@435 | 311 | |
duke@435 | 312 | size_t CodeHeap::allocated_capacity() const { |
neliasso@4952 | 313 | // size of used heap - size on freelist |
neliasso@4952 | 314 | return segments_to_size(_next_segment - _freelist_segments); |
duke@435 | 315 | } |
duke@435 | 316 | |
neliasso@4952 | 317 | // Returns size of the unallocated heap block |
neliasso@4952 | 318 | size_t CodeHeap::heap_unallocated_capacity() const { |
neliasso@4952 | 319 | // Total number of segments - number currently used |
neliasso@4952 | 320 | return segments_to_size(_number_of_reserved_segments - _next_segment); |
kvn@2414 | 321 | } |
kvn@2414 | 322 | |
duke@435 | 323 | // Free list management |
duke@435 | 324 | |
duke@435 | 325 | FreeBlock *CodeHeap::following_block(FreeBlock *b) { |
duke@435 | 326 | return (FreeBlock*)(((address)b) + _segment_size * b->length()); |
duke@435 | 327 | } |
duke@435 | 328 | |
duke@435 | 329 | // Inserts block b after a |
duke@435 | 330 | void CodeHeap::insert_after(FreeBlock* a, FreeBlock* b) { |
duke@435 | 331 | assert(a != NULL && b != NULL, "must be real pointers"); |
duke@435 | 332 | |
duke@435 | 333 | // Link b into the list after a |
duke@435 | 334 | b->set_link(a->link()); |
duke@435 | 335 | a->set_link(b); |
duke@435 | 336 | |
duke@435 | 337 | // See if we can merge blocks |
duke@435 | 338 | merge_right(b); // Try to make b bigger |
duke@435 | 339 | merge_right(a); // Try to make a include b |
duke@435 | 340 | } |
duke@435 | 341 | |
duke@435 | 342 | // Try to merge this block with the following block |
duke@435 | 343 | void CodeHeap::merge_right(FreeBlock *a) { |
duke@435 | 344 | assert(a->free(), "must be a free block"); |
duke@435 | 345 | if (following_block(a) == a->link()) { |
duke@435 | 346 | assert(a->link() != NULL && a->link()->free(), "must be free too"); |
duke@435 | 347 | // Update block a to include the following block |
duke@435 | 348 | a->set_length(a->length() + a->link()->length()); |
duke@435 | 349 | a->set_link(a->link()->link()); |
duke@435 | 350 | // Update find_start map |
duke@435 | 351 | size_t beg = segment_for(a); |
duke@435 | 352 | mark_segmap_as_used(beg, beg + a->length()); |
duke@435 | 353 | } |
duke@435 | 354 | } |
duke@435 | 355 | |
duke@435 | 356 | void CodeHeap::add_to_freelist(HeapBlock *a) { |
duke@435 | 357 | FreeBlock* b = (FreeBlock*)a; |
duke@435 | 358 | assert(b != _freelist, "cannot be removed twice"); |
duke@435 | 359 | |
duke@435 | 360 | // Mark as free and update free space count |
neliasso@4952 | 361 | _freelist_segments += b->length(); |
duke@435 | 362 | b->set_free(); |
duke@435 | 363 | |
duke@435 | 364 | // First element in list? |
duke@435 | 365 | if (_freelist == NULL) { |
duke@435 | 366 | _freelist = b; |
duke@435 | 367 | b->set_link(NULL); |
duke@435 | 368 | return; |
duke@435 | 369 | } |
duke@435 | 370 | |
duke@435 | 371 | // Scan for right place to put into list. List |
duke@435 | 372 | // is sorted by increasing addresseses |
duke@435 | 373 | FreeBlock* prev = NULL; |
duke@435 | 374 | FreeBlock* cur = _freelist; |
duke@435 | 375 | while(cur != NULL && cur < b) { |
duke@435 | 376 | assert(prev == NULL || prev < cur, "must be ordered"); |
duke@435 | 377 | prev = cur; |
duke@435 | 378 | cur = cur->link(); |
duke@435 | 379 | } |
duke@435 | 380 | |
duke@435 | 381 | assert( (prev == NULL && b < _freelist) || |
duke@435 | 382 | (prev < b && (cur == NULL || b < cur)), "list must be ordered"); |
duke@435 | 383 | |
duke@435 | 384 | if (prev == NULL) { |
duke@435 | 385 | // Insert first in list |
duke@435 | 386 | b->set_link(_freelist); |
duke@435 | 387 | _freelist = b; |
duke@435 | 388 | merge_right(_freelist); |
duke@435 | 389 | } else { |
duke@435 | 390 | insert_after(prev, b); |
duke@435 | 391 | } |
duke@435 | 392 | } |
duke@435 | 393 | |
duke@435 | 394 | // Search freelist for an entry on the list with the best fit |
duke@435 | 395 | // Return NULL if no one was found |
neliasso@4952 | 396 | FreeBlock* CodeHeap::search_freelist(size_t length, bool is_critical) { |
duke@435 | 397 | FreeBlock *best_block = NULL; |
duke@435 | 398 | FreeBlock *best_prev = NULL; |
duke@435 | 399 | size_t best_length = 0; |
duke@435 | 400 | |
duke@435 | 401 | // Search for smallest block which is bigger than length |
duke@435 | 402 | FreeBlock *prev = NULL; |
duke@435 | 403 | FreeBlock *cur = _freelist; |
duke@435 | 404 | while(cur != NULL) { |
duke@435 | 405 | size_t l = cur->length(); |
duke@435 | 406 | if (l >= length && (best_block == NULL || best_length > l)) { |
neliasso@4952 | 407 | |
neliasso@4952 | 408 | // Non critical allocations are not allowed to use the last part of the code heap. |
neliasso@4952 | 409 | if (!is_critical) { |
neliasso@4952 | 410 | // Make sure the end of the allocation doesn't cross into the last part of the code heap |
neliasso@4952 | 411 | if (((size_t)cur + length) > ((size_t)high_boundary() - CodeCacheMinimumFreeSpace)) { |
neliasso@4952 | 412 | // the freelist is sorted by address - if one fails, all consecutive will also fail. |
neliasso@4952 | 413 | break; |
neliasso@4952 | 414 | } |
neliasso@4952 | 415 | } |
neliasso@4952 | 416 | |
duke@435 | 417 | // Remember best block, its previous element, and its length |
duke@435 | 418 | best_block = cur; |
duke@435 | 419 | best_prev = prev; |
duke@435 | 420 | best_length = best_block->length(); |
duke@435 | 421 | } |
duke@435 | 422 | |
duke@435 | 423 | // Next element in list |
duke@435 | 424 | prev = cur; |
duke@435 | 425 | cur = cur->link(); |
duke@435 | 426 | } |
duke@435 | 427 | |
duke@435 | 428 | if (best_block == NULL) { |
duke@435 | 429 | // None found |
duke@435 | 430 | return NULL; |
duke@435 | 431 | } |
duke@435 | 432 | |
duke@435 | 433 | assert((best_prev == NULL && _freelist == best_block ) || |
duke@435 | 434 | (best_prev != NULL && best_prev->link() == best_block), "sanity check"); |
duke@435 | 435 | |
duke@435 | 436 | // Exact (or at least good enough) fit. Remove from list. |
duke@435 | 437 | // Don't leave anything on the freelist smaller than CodeCacheMinBlockLength. |
duke@435 | 438 | if (best_length < length + CodeCacheMinBlockLength) { |
duke@435 | 439 | length = best_length; |
duke@435 | 440 | if (best_prev == NULL) { |
duke@435 | 441 | assert(_freelist == best_block, "sanity check"); |
duke@435 | 442 | _freelist = _freelist->link(); |
duke@435 | 443 | } else { |
duke@435 | 444 | // Unmap element |
duke@435 | 445 | best_prev->set_link(best_block->link()); |
duke@435 | 446 | } |
duke@435 | 447 | } else { |
duke@435 | 448 | // Truncate block and return a pointer to the following block |
duke@435 | 449 | best_block->set_length(best_length - length); |
duke@435 | 450 | best_block = following_block(best_block); |
duke@435 | 451 | // Set used bit and length on new block |
duke@435 | 452 | size_t beg = segment_for(best_block); |
duke@435 | 453 | mark_segmap_as_used(beg, beg + length); |
duke@435 | 454 | best_block->set_length(length); |
duke@435 | 455 | } |
duke@435 | 456 | |
duke@435 | 457 | best_block->set_used(); |
neliasso@4952 | 458 | _freelist_segments -= length; |
duke@435 | 459 | return best_block; |
duke@435 | 460 | } |
duke@435 | 461 | |
duke@435 | 462 | //---------------------------------------------------------------------------- |
duke@435 | 463 | // Non-product code |
duke@435 | 464 | |
duke@435 | 465 | #ifndef PRODUCT |
duke@435 | 466 | |
duke@435 | 467 | void CodeHeap::print() { |
duke@435 | 468 | tty->print_cr("The Heap"); |
duke@435 | 469 | } |
duke@435 | 470 | |
duke@435 | 471 | #endif |
duke@435 | 472 | |
duke@435 | 473 | void CodeHeap::verify() { |
duke@435 | 474 | // Count the number of blocks on the freelist, and the amount of space |
duke@435 | 475 | // represented. |
duke@435 | 476 | int count = 0; |
duke@435 | 477 | size_t len = 0; |
duke@435 | 478 | for(FreeBlock* b = _freelist; b != NULL; b = b->link()) { |
duke@435 | 479 | len += b->length(); |
duke@435 | 480 | count++; |
duke@435 | 481 | } |
duke@435 | 482 | |
duke@435 | 483 | // Verify that freelist contains the right amount of free space |
neliasso@4952 | 484 | // guarantee(len == _freelist_segments, "wrong freelist"); |
duke@435 | 485 | |
duke@435 | 486 | // Verify that the number of free blocks is not out of hand. |
duke@435 | 487 | static int free_block_threshold = 10000; |
duke@435 | 488 | if (count > free_block_threshold) { |
duke@435 | 489 | warning("CodeHeap: # of free blocks > %d", free_block_threshold); |
duke@435 | 490 | // Double the warning limit |
duke@435 | 491 | free_block_threshold *= 2; |
duke@435 | 492 | } |
duke@435 | 493 | |
duke@435 | 494 | // Verify that the freelist contains the same number of free blocks that is |
duke@435 | 495 | // found on the full list. |
duke@435 | 496 | for(HeapBlock *h = first_block(); h != NULL; h = next_block(h)) { |
duke@435 | 497 | if (h->free()) count--; |
duke@435 | 498 | } |
phh@1558 | 499 | // guarantee(count == 0, "missing free blocks"); |
duke@435 | 500 | } |