Tue, 24 Jul 2012 10:51:00 -0700
7023639: JSR 292 method handle invocation needs a fast path for compiled code
6984705: JSR 292 method handle creation should not go through JNI
Summary: remove assembly code for JDK 7 chained method handles
Reviewed-by: jrose, twisti, kvn, mhaupt
Contributed-by: John Rose <john.r.rose@oracle.com>, Christian Thalinger <christian.thalinger@oracle.com>, Michael Haupt <michael.haupt@oracle.com>
duke@435 | 1 | /* |
trims@2708 | 2 | * Copyright (c) 1998, 2011, Oracle and/or its affiliates. All rights reserved. |
duke@435 | 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
duke@435 | 4 | * |
duke@435 | 5 | * This code is free software; you can redistribute it and/or modify it |
duke@435 | 6 | * under the terms of the GNU General Public License version 2 only, as |
duke@435 | 7 | * published by the Free Software Foundation. |
duke@435 | 8 | * |
duke@435 | 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
duke@435 | 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
duke@435 | 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
duke@435 | 12 | * version 2 for more details (a copy is included in the LICENSE file that |
duke@435 | 13 | * accompanied this code). |
duke@435 | 14 | * |
duke@435 | 15 | * You should have received a copy of the GNU General Public License version |
duke@435 | 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
duke@435 | 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
duke@435 | 18 | * |
trims@1907 | 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
trims@1907 | 20 | * or visit www.oracle.com if you need additional information or have any |
trims@1907 | 21 | * questions. |
duke@435 | 22 | * |
duke@435 | 23 | */ |
duke@435 | 24 | |
stefank@2314 | 25 | #include "precompiled.hpp" |
stefank@2314 | 26 | #include "code/codeBlob.hpp" |
stefank@2314 | 27 | #include "code/codeCache.hpp" |
stefank@2314 | 28 | #include "code/relocInfo.hpp" |
stefank@2314 | 29 | #include "compiler/disassembler.hpp" |
stefank@2314 | 30 | #include "interpreter/bytecode.hpp" |
stefank@2314 | 31 | #include "memory/allocation.inline.hpp" |
stefank@2314 | 32 | #include "memory/heap.hpp" |
stefank@2314 | 33 | #include "oops/oop.inline.hpp" |
stefank@2314 | 34 | #include "prims/forte.hpp" |
stefank@2314 | 35 | #include "runtime/handles.inline.hpp" |
stefank@2314 | 36 | #include "runtime/interfaceSupport.hpp" |
stefank@2314 | 37 | #include "runtime/mutexLocker.hpp" |
stefank@2314 | 38 | #include "runtime/safepoint.hpp" |
stefank@2314 | 39 | #include "runtime/sharedRuntime.hpp" |
stefank@2314 | 40 | #include "runtime/vframe.hpp" |
stefank@2314 | 41 | #include "services/memoryService.hpp" |
stefank@2314 | 42 | #ifdef TARGET_ARCH_x86 |
stefank@2314 | 43 | # include "nativeInst_x86.hpp" |
stefank@2314 | 44 | #endif |
stefank@2314 | 45 | #ifdef TARGET_ARCH_sparc |
stefank@2314 | 46 | # include "nativeInst_sparc.hpp" |
stefank@2314 | 47 | #endif |
stefank@2314 | 48 | #ifdef TARGET_ARCH_zero |
stefank@2314 | 49 | # include "nativeInst_zero.hpp" |
stefank@2314 | 50 | #endif |
bobv@2508 | 51 | #ifdef TARGET_ARCH_arm |
bobv@2508 | 52 | # include "nativeInst_arm.hpp" |
bobv@2508 | 53 | #endif |
bobv@2508 | 54 | #ifdef TARGET_ARCH_ppc |
bobv@2508 | 55 | # include "nativeInst_ppc.hpp" |
bobv@2508 | 56 | #endif |
stefank@2314 | 57 | #ifdef COMPILER1 |
stefank@2314 | 58 | #include "c1/c1_Runtime1.hpp" |
stefank@2314 | 59 | #endif |
duke@435 | 60 | |
duke@435 | 61 | unsigned int align_code_offset(int offset) { |
duke@435 | 62 | // align the size to CodeEntryAlignment |
duke@435 | 63 | return |
duke@435 | 64 | ((offset + (int)CodeHeap::header_size() + (CodeEntryAlignment-1)) & ~(CodeEntryAlignment-1)) |
duke@435 | 65 | - (int)CodeHeap::header_size(); |
duke@435 | 66 | } |
duke@435 | 67 | |
duke@435 | 68 | |
duke@435 | 69 | // This must be consistent with the CodeBlob constructor's layout actions. |
duke@435 | 70 | unsigned int CodeBlob::allocation_size(CodeBuffer* cb, int header_size) { |
duke@435 | 71 | unsigned int size = header_size; |
duke@435 | 72 | size += round_to(cb->total_relocation_size(), oopSize); |
duke@435 | 73 | // align the size to CodeEntryAlignment |
duke@435 | 74 | size = align_code_offset(size); |
twisti@2103 | 75 | size += round_to(cb->total_content_size(), oopSize); |
duke@435 | 76 | size += round_to(cb->total_oop_size(), oopSize); |
duke@435 | 77 | return size; |
duke@435 | 78 | } |
duke@435 | 79 | |
duke@435 | 80 | |
duke@435 | 81 | // Creates a simple CodeBlob. Sets up the size of the different regions. |
duke@435 | 82 | CodeBlob::CodeBlob(const char* name, int header_size, int size, int frame_complete, int locs_size) { |
twisti@2103 | 83 | assert(size == round_to(size, oopSize), "unaligned size"); |
twisti@2103 | 84 | assert(locs_size == round_to(locs_size, oopSize), "unaligned size"); |
duke@435 | 85 | assert(header_size == round_to(header_size, oopSize), "unaligned size"); |
duke@435 | 86 | assert(!UseRelocIndex, "no space allocated for reloc index yet"); |
duke@435 | 87 | |
duke@435 | 88 | // Note: If UseRelocIndex is enabled, there needs to be (at least) one |
duke@435 | 89 | // extra word for the relocation information, containing the reloc |
duke@435 | 90 | // index table length. Unfortunately, the reloc index table imple- |
duke@435 | 91 | // mentation is not easily understandable and thus it is not clear |
duke@435 | 92 | // what exactly the format is supposed to be. For now, we just turn |
duke@435 | 93 | // off the use of this table (gri 7/6/2000). |
duke@435 | 94 | |
duke@435 | 95 | _name = name; |
duke@435 | 96 | _size = size; |
duke@435 | 97 | _frame_complete_offset = frame_complete; |
duke@435 | 98 | _header_size = header_size; |
duke@435 | 99 | _relocation_size = locs_size; |
twisti@2103 | 100 | _content_offset = align_code_offset(header_size + _relocation_size); |
twisti@2103 | 101 | _code_offset = _content_offset; |
duke@435 | 102 | _data_offset = size; |
duke@435 | 103 | _frame_size = 0; |
duke@435 | 104 | set_oop_maps(NULL); |
duke@435 | 105 | } |
duke@435 | 106 | |
duke@435 | 107 | |
duke@435 | 108 | // Creates a CodeBlob from a CodeBuffer. Sets up the size of the different regions, |
duke@435 | 109 | // and copy code and relocation info. |
duke@435 | 110 | CodeBlob::CodeBlob( |
duke@435 | 111 | const char* name, |
duke@435 | 112 | CodeBuffer* cb, |
duke@435 | 113 | int header_size, |
duke@435 | 114 | int size, |
duke@435 | 115 | int frame_complete, |
duke@435 | 116 | int frame_size, |
duke@435 | 117 | OopMapSet* oop_maps |
duke@435 | 118 | ) { |
twisti@2103 | 119 | assert(size == round_to(size, oopSize), "unaligned size"); |
duke@435 | 120 | assert(header_size == round_to(header_size, oopSize), "unaligned size"); |
duke@435 | 121 | |
duke@435 | 122 | _name = name; |
duke@435 | 123 | _size = size; |
duke@435 | 124 | _frame_complete_offset = frame_complete; |
duke@435 | 125 | _header_size = header_size; |
duke@435 | 126 | _relocation_size = round_to(cb->total_relocation_size(), oopSize); |
twisti@2103 | 127 | _content_offset = align_code_offset(header_size + _relocation_size); |
twisti@2117 | 128 | _code_offset = _content_offset + cb->total_offset_of(cb->insts()); |
twisti@2103 | 129 | _data_offset = _content_offset + round_to(cb->total_content_size(), oopSize); |
duke@435 | 130 | assert(_data_offset <= size, "codeBlob is too small"); |
duke@435 | 131 | |
duke@435 | 132 | cb->copy_code_and_locs_to(this); |
duke@435 | 133 | set_oop_maps(oop_maps); |
duke@435 | 134 | _frame_size = frame_size; |
duke@435 | 135 | #ifdef COMPILER1 |
duke@435 | 136 | // probably wrong for tiered |
duke@435 | 137 | assert(_frame_size >= -1, "must use frame size or -1 for runtime stubs"); |
duke@435 | 138 | #endif // COMPILER1 |
duke@435 | 139 | } |
duke@435 | 140 | |
duke@435 | 141 | |
duke@435 | 142 | void CodeBlob::set_oop_maps(OopMapSet* p) { |
duke@435 | 143 | // Danger Will Robinson! This method allocates a big |
duke@435 | 144 | // chunk of memory, its your job to free it. |
duke@435 | 145 | if (p != NULL) { |
duke@435 | 146 | // We need to allocate a chunk big enough to hold the OopMapSet and all of its OopMaps |
zgu@3900 | 147 | _oop_maps = (OopMapSet* )NEW_C_HEAP_ARRAY(unsigned char, p->heap_size(), mtCode); |
duke@435 | 148 | p->copy_to((address)_oop_maps); |
duke@435 | 149 | } else { |
duke@435 | 150 | _oop_maps = NULL; |
duke@435 | 151 | } |
duke@435 | 152 | } |
duke@435 | 153 | |
duke@435 | 154 | |
never@2895 | 155 | void CodeBlob::trace_new_stub(CodeBlob* stub, const char* name1, const char* name2) { |
never@2895 | 156 | // Do not hold the CodeCache lock during name formatting. |
never@2895 | 157 | assert(!CodeCache_lock->owned_by_self(), "release CodeCache before registering the stub"); |
never@2895 | 158 | |
never@2895 | 159 | if (stub != NULL) { |
never@2895 | 160 | char stub_id[256]; |
never@2895 | 161 | assert(strlen(name1) + strlen(name2) < sizeof(stub_id), ""); |
never@2895 | 162 | jio_snprintf(stub_id, sizeof(stub_id), "%s%s", name1, name2); |
never@2895 | 163 | if (PrintStubCode) { |
never@2895 | 164 | tty->print_cr("Decoding %s " INTPTR_FORMAT, stub_id, (intptr_t) stub); |
never@2895 | 165 | Disassembler::decode(stub->code_begin(), stub->code_end()); |
never@2895 | 166 | } |
never@2895 | 167 | Forte::register_stub(stub_id, stub->code_begin(), stub->code_end()); |
never@2895 | 168 | |
never@2895 | 169 | if (JvmtiExport::should_post_dynamic_code_generated()) { |
never@2895 | 170 | const char* stub_name = name2; |
never@2895 | 171 | if (name2[0] == '\0') stub_name = name1; |
never@2895 | 172 | JvmtiExport::post_dynamic_code_generated(stub_name, stub->code_begin(), stub->code_end()); |
never@2895 | 173 | } |
never@2895 | 174 | } |
never@2895 | 175 | |
never@2895 | 176 | // Track memory usage statistic after releasing CodeCache_lock |
never@2895 | 177 | MemoryService::track_code_cache_memory_usage(); |
never@2895 | 178 | } |
never@2895 | 179 | |
never@2895 | 180 | |
duke@435 | 181 | void CodeBlob::flush() { |
duke@435 | 182 | if (_oop_maps) { |
zgu@3900 | 183 | FREE_C_HEAP_ARRAY(unsigned char, _oop_maps, mtCode); |
duke@435 | 184 | _oop_maps = NULL; |
duke@435 | 185 | } |
duke@435 | 186 | _comments.free(); |
duke@435 | 187 | } |
duke@435 | 188 | |
duke@435 | 189 | |
duke@435 | 190 | OopMap* CodeBlob::oop_map_for_return_address(address return_address) { |
twisti@2103 | 191 | assert(oop_maps() != NULL, "nope"); |
twisti@2103 | 192 | return oop_maps()->find_map_at_offset((intptr_t) return_address - (intptr_t) code_begin()); |
duke@435 | 193 | } |
duke@435 | 194 | |
duke@435 | 195 | |
duke@435 | 196 | //---------------------------------------------------------------------------------------------------- |
duke@435 | 197 | // Implementation of BufferBlob |
duke@435 | 198 | |
duke@435 | 199 | |
duke@435 | 200 | BufferBlob::BufferBlob(const char* name, int size) |
duke@435 | 201 | : CodeBlob(name, sizeof(BufferBlob), size, CodeOffsets::frame_never_safe, /*locs_size:*/ 0) |
duke@435 | 202 | {} |
duke@435 | 203 | |
duke@435 | 204 | BufferBlob* BufferBlob::create(const char* name, int buffer_size) { |
duke@435 | 205 | ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
duke@435 | 206 | |
duke@435 | 207 | BufferBlob* blob = NULL; |
duke@435 | 208 | unsigned int size = sizeof(BufferBlob); |
duke@435 | 209 | // align the size to CodeEntryAlignment |
duke@435 | 210 | size = align_code_offset(size); |
duke@435 | 211 | size += round_to(buffer_size, oopSize); |
duke@435 | 212 | assert(name != NULL, "must provide a name"); |
duke@435 | 213 | { |
duke@435 | 214 | MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
duke@435 | 215 | blob = new (size) BufferBlob(name, size); |
duke@435 | 216 | } |
duke@435 | 217 | // Track memory usage statistic after releasing CodeCache_lock |
duke@435 | 218 | MemoryService::track_code_cache_memory_usage(); |
duke@435 | 219 | |
duke@435 | 220 | return blob; |
duke@435 | 221 | } |
duke@435 | 222 | |
duke@435 | 223 | |
duke@435 | 224 | BufferBlob::BufferBlob(const char* name, int size, CodeBuffer* cb) |
duke@435 | 225 | : CodeBlob(name, cb, sizeof(BufferBlob), size, CodeOffsets::frame_never_safe, 0, NULL) |
duke@435 | 226 | {} |
duke@435 | 227 | |
duke@435 | 228 | BufferBlob* BufferBlob::create(const char* name, CodeBuffer* cb) { |
duke@435 | 229 | ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
duke@435 | 230 | |
duke@435 | 231 | BufferBlob* blob = NULL; |
duke@435 | 232 | unsigned int size = allocation_size(cb, sizeof(BufferBlob)); |
duke@435 | 233 | assert(name != NULL, "must provide a name"); |
duke@435 | 234 | { |
duke@435 | 235 | MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
duke@435 | 236 | blob = new (size) BufferBlob(name, size, cb); |
duke@435 | 237 | } |
duke@435 | 238 | // Track memory usage statistic after releasing CodeCache_lock |
duke@435 | 239 | MemoryService::track_code_cache_memory_usage(); |
duke@435 | 240 | |
duke@435 | 241 | return blob; |
duke@435 | 242 | } |
duke@435 | 243 | |
duke@435 | 244 | |
duke@435 | 245 | void* BufferBlob::operator new(size_t s, unsigned size) { |
duke@435 | 246 | void* p = CodeCache::allocate(size); |
duke@435 | 247 | return p; |
duke@435 | 248 | } |
duke@435 | 249 | |
duke@435 | 250 | |
duke@435 | 251 | void BufferBlob::free( BufferBlob *blob ) { |
duke@435 | 252 | ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
duke@435 | 253 | { |
duke@435 | 254 | MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
duke@435 | 255 | CodeCache::free((CodeBlob*)blob); |
duke@435 | 256 | } |
duke@435 | 257 | // Track memory usage statistic after releasing CodeCache_lock |
duke@435 | 258 | MemoryService::track_code_cache_memory_usage(); |
duke@435 | 259 | } |
duke@435 | 260 | |
twisti@1734 | 261 | |
twisti@1734 | 262 | //---------------------------------------------------------------------------------------------------- |
twisti@1734 | 263 | // Implementation of AdapterBlob |
twisti@1734 | 264 | |
never@2018 | 265 | AdapterBlob::AdapterBlob(int size, CodeBuffer* cb) : |
never@2018 | 266 | BufferBlob("I2C/C2I adapters", size, cb) { |
never@2018 | 267 | CodeCache::commit(this); |
never@2018 | 268 | } |
never@2018 | 269 | |
twisti@1734 | 270 | AdapterBlob* AdapterBlob::create(CodeBuffer* cb) { |
twisti@1734 | 271 | ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
twisti@1734 | 272 | |
twisti@1734 | 273 | AdapterBlob* blob = NULL; |
twisti@1734 | 274 | unsigned int size = allocation_size(cb, sizeof(AdapterBlob)); |
twisti@1734 | 275 | { |
twisti@1734 | 276 | MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
twisti@1734 | 277 | blob = new (size) AdapterBlob(size, cb); |
twisti@1734 | 278 | } |
twisti@1734 | 279 | // Track memory usage statistic after releasing CodeCache_lock |
twisti@1734 | 280 | MemoryService::track_code_cache_memory_usage(); |
twisti@1734 | 281 | |
twisti@1734 | 282 | return blob; |
duke@435 | 283 | } |
duke@435 | 284 | |
twisti@1734 | 285 | |
twisti@1734 | 286 | //---------------------------------------------------------------------------------------------------- |
twisti@1734 | 287 | // Implementation of MethodHandlesAdapterBlob |
twisti@1734 | 288 | |
twisti@1734 | 289 | MethodHandlesAdapterBlob* MethodHandlesAdapterBlob::create(int buffer_size) { |
twisti@1734 | 290 | ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
twisti@1734 | 291 | |
twisti@1734 | 292 | MethodHandlesAdapterBlob* blob = NULL; |
twisti@1734 | 293 | unsigned int size = sizeof(MethodHandlesAdapterBlob); |
twisti@1734 | 294 | // align the size to CodeEntryAlignment |
twisti@1734 | 295 | size = align_code_offset(size); |
twisti@1734 | 296 | size += round_to(buffer_size, oopSize); |
twisti@1734 | 297 | { |
twisti@1734 | 298 | MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
twisti@1734 | 299 | blob = new (size) MethodHandlesAdapterBlob(size); |
twisti@1734 | 300 | } |
twisti@1734 | 301 | // Track memory usage statistic after releasing CodeCache_lock |
twisti@1734 | 302 | MemoryService::track_code_cache_memory_usage(); |
twisti@1734 | 303 | |
twisti@1734 | 304 | return blob; |
twisti@1734 | 305 | } |
twisti@1734 | 306 | |
twisti@1734 | 307 | |
duke@435 | 308 | //---------------------------------------------------------------------------------------------------- |
duke@435 | 309 | // Implementation of RuntimeStub |
duke@435 | 310 | |
duke@435 | 311 | RuntimeStub::RuntimeStub( |
duke@435 | 312 | const char* name, |
duke@435 | 313 | CodeBuffer* cb, |
duke@435 | 314 | int size, |
duke@435 | 315 | int frame_complete, |
duke@435 | 316 | int frame_size, |
duke@435 | 317 | OopMapSet* oop_maps, |
duke@435 | 318 | bool caller_must_gc_arguments |
duke@435 | 319 | ) |
duke@435 | 320 | : CodeBlob(name, cb, sizeof(RuntimeStub), size, frame_complete, frame_size, oop_maps) |
duke@435 | 321 | { |
duke@435 | 322 | _caller_must_gc_arguments = caller_must_gc_arguments; |
duke@435 | 323 | } |
duke@435 | 324 | |
duke@435 | 325 | |
duke@435 | 326 | RuntimeStub* RuntimeStub::new_runtime_stub(const char* stub_name, |
duke@435 | 327 | CodeBuffer* cb, |
duke@435 | 328 | int frame_complete, |
duke@435 | 329 | int frame_size, |
duke@435 | 330 | OopMapSet* oop_maps, |
duke@435 | 331 | bool caller_must_gc_arguments) |
duke@435 | 332 | { |
duke@435 | 333 | RuntimeStub* stub = NULL; |
duke@435 | 334 | ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
duke@435 | 335 | { |
duke@435 | 336 | MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
duke@435 | 337 | unsigned int size = allocation_size(cb, sizeof(RuntimeStub)); |
duke@435 | 338 | stub = new (size) RuntimeStub(stub_name, cb, size, frame_complete, frame_size, oop_maps, caller_must_gc_arguments); |
duke@435 | 339 | } |
duke@435 | 340 | |
never@2895 | 341 | trace_new_stub(stub, "RuntimeStub - ", stub_name); |
duke@435 | 342 | |
duke@435 | 343 | return stub; |
duke@435 | 344 | } |
duke@435 | 345 | |
duke@435 | 346 | |
duke@435 | 347 | void* RuntimeStub::operator new(size_t s, unsigned size) { |
duke@435 | 348 | void* p = CodeCache::allocate(size); |
duke@435 | 349 | if (!p) fatal("Initial size of CodeCache is too small"); |
duke@435 | 350 | return p; |
duke@435 | 351 | } |
duke@435 | 352 | |
never@2895 | 353 | // operator new shared by all singletons: |
never@2895 | 354 | void* SingletonBlob::operator new(size_t s, unsigned size) { |
never@2895 | 355 | void* p = CodeCache::allocate(size); |
never@2895 | 356 | if (!p) fatal("Initial size of CodeCache is too small"); |
never@2895 | 357 | return p; |
never@2895 | 358 | } |
never@2895 | 359 | |
never@2895 | 360 | |
never@2895 | 361 | //---------------------------------------------------------------------------------------------------- |
duke@435 | 362 | // Implementation of DeoptimizationBlob |
duke@435 | 363 | |
duke@435 | 364 | DeoptimizationBlob::DeoptimizationBlob( |
duke@435 | 365 | CodeBuffer* cb, |
duke@435 | 366 | int size, |
duke@435 | 367 | OopMapSet* oop_maps, |
duke@435 | 368 | int unpack_offset, |
duke@435 | 369 | int unpack_with_exception_offset, |
duke@435 | 370 | int unpack_with_reexecution_offset, |
duke@435 | 371 | int frame_size |
duke@435 | 372 | ) |
duke@435 | 373 | : SingletonBlob("DeoptimizationBlob", cb, sizeof(DeoptimizationBlob), size, frame_size, oop_maps) |
duke@435 | 374 | { |
duke@435 | 375 | _unpack_offset = unpack_offset; |
duke@435 | 376 | _unpack_with_exception = unpack_with_exception_offset; |
duke@435 | 377 | _unpack_with_reexecution = unpack_with_reexecution_offset; |
duke@435 | 378 | #ifdef COMPILER1 |
duke@435 | 379 | _unpack_with_exception_in_tls = -1; |
duke@435 | 380 | #endif |
duke@435 | 381 | } |
duke@435 | 382 | |
duke@435 | 383 | |
duke@435 | 384 | DeoptimizationBlob* DeoptimizationBlob::create( |
duke@435 | 385 | CodeBuffer* cb, |
duke@435 | 386 | OopMapSet* oop_maps, |
duke@435 | 387 | int unpack_offset, |
duke@435 | 388 | int unpack_with_exception_offset, |
duke@435 | 389 | int unpack_with_reexecution_offset, |
duke@435 | 390 | int frame_size) |
duke@435 | 391 | { |
duke@435 | 392 | DeoptimizationBlob* blob = NULL; |
duke@435 | 393 | ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
duke@435 | 394 | { |
duke@435 | 395 | MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
duke@435 | 396 | unsigned int size = allocation_size(cb, sizeof(DeoptimizationBlob)); |
duke@435 | 397 | blob = new (size) DeoptimizationBlob(cb, |
duke@435 | 398 | size, |
duke@435 | 399 | oop_maps, |
duke@435 | 400 | unpack_offset, |
duke@435 | 401 | unpack_with_exception_offset, |
duke@435 | 402 | unpack_with_reexecution_offset, |
duke@435 | 403 | frame_size); |
duke@435 | 404 | } |
duke@435 | 405 | |
never@2895 | 406 | trace_new_stub(blob, "DeoptimizationBlob"); |
duke@435 | 407 | |
duke@435 | 408 | return blob; |
duke@435 | 409 | } |
duke@435 | 410 | |
duke@435 | 411 | |
duke@435 | 412 | //---------------------------------------------------------------------------------------------------- |
duke@435 | 413 | // Implementation of UncommonTrapBlob |
duke@435 | 414 | |
duke@435 | 415 | #ifdef COMPILER2 |
duke@435 | 416 | UncommonTrapBlob::UncommonTrapBlob( |
duke@435 | 417 | CodeBuffer* cb, |
duke@435 | 418 | int size, |
duke@435 | 419 | OopMapSet* oop_maps, |
duke@435 | 420 | int frame_size |
duke@435 | 421 | ) |
duke@435 | 422 | : SingletonBlob("UncommonTrapBlob", cb, sizeof(UncommonTrapBlob), size, frame_size, oop_maps) |
duke@435 | 423 | {} |
duke@435 | 424 | |
duke@435 | 425 | |
duke@435 | 426 | UncommonTrapBlob* UncommonTrapBlob::create( |
duke@435 | 427 | CodeBuffer* cb, |
duke@435 | 428 | OopMapSet* oop_maps, |
duke@435 | 429 | int frame_size) |
duke@435 | 430 | { |
duke@435 | 431 | UncommonTrapBlob* blob = NULL; |
duke@435 | 432 | ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
duke@435 | 433 | { |
duke@435 | 434 | MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
duke@435 | 435 | unsigned int size = allocation_size(cb, sizeof(UncommonTrapBlob)); |
duke@435 | 436 | blob = new (size) UncommonTrapBlob(cb, size, oop_maps, frame_size); |
duke@435 | 437 | } |
duke@435 | 438 | |
never@2895 | 439 | trace_new_stub(blob, "UncommonTrapBlob"); |
duke@435 | 440 | |
duke@435 | 441 | return blob; |
duke@435 | 442 | } |
duke@435 | 443 | |
duke@435 | 444 | |
duke@435 | 445 | #endif // COMPILER2 |
duke@435 | 446 | |
duke@435 | 447 | |
duke@435 | 448 | //---------------------------------------------------------------------------------------------------- |
duke@435 | 449 | // Implementation of ExceptionBlob |
duke@435 | 450 | |
duke@435 | 451 | #ifdef COMPILER2 |
duke@435 | 452 | ExceptionBlob::ExceptionBlob( |
duke@435 | 453 | CodeBuffer* cb, |
duke@435 | 454 | int size, |
duke@435 | 455 | OopMapSet* oop_maps, |
duke@435 | 456 | int frame_size |
duke@435 | 457 | ) |
duke@435 | 458 | : SingletonBlob("ExceptionBlob", cb, sizeof(ExceptionBlob), size, frame_size, oop_maps) |
duke@435 | 459 | {} |
duke@435 | 460 | |
duke@435 | 461 | |
duke@435 | 462 | ExceptionBlob* ExceptionBlob::create( |
duke@435 | 463 | CodeBuffer* cb, |
duke@435 | 464 | OopMapSet* oop_maps, |
duke@435 | 465 | int frame_size) |
duke@435 | 466 | { |
duke@435 | 467 | ExceptionBlob* blob = NULL; |
duke@435 | 468 | ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
duke@435 | 469 | { |
duke@435 | 470 | MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
duke@435 | 471 | unsigned int size = allocation_size(cb, sizeof(ExceptionBlob)); |
duke@435 | 472 | blob = new (size) ExceptionBlob(cb, size, oop_maps, frame_size); |
duke@435 | 473 | } |
duke@435 | 474 | |
never@2895 | 475 | trace_new_stub(blob, "ExceptionBlob"); |
duke@435 | 476 | |
duke@435 | 477 | return blob; |
duke@435 | 478 | } |
duke@435 | 479 | |
duke@435 | 480 | |
duke@435 | 481 | #endif // COMPILER2 |
duke@435 | 482 | |
duke@435 | 483 | |
duke@435 | 484 | //---------------------------------------------------------------------------------------------------- |
duke@435 | 485 | // Implementation of SafepointBlob |
duke@435 | 486 | |
duke@435 | 487 | SafepointBlob::SafepointBlob( |
duke@435 | 488 | CodeBuffer* cb, |
duke@435 | 489 | int size, |
duke@435 | 490 | OopMapSet* oop_maps, |
duke@435 | 491 | int frame_size |
duke@435 | 492 | ) |
duke@435 | 493 | : SingletonBlob("SafepointBlob", cb, sizeof(SafepointBlob), size, frame_size, oop_maps) |
duke@435 | 494 | {} |
duke@435 | 495 | |
duke@435 | 496 | |
duke@435 | 497 | SafepointBlob* SafepointBlob::create( |
duke@435 | 498 | CodeBuffer* cb, |
duke@435 | 499 | OopMapSet* oop_maps, |
duke@435 | 500 | int frame_size) |
duke@435 | 501 | { |
duke@435 | 502 | SafepointBlob* blob = NULL; |
duke@435 | 503 | ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
duke@435 | 504 | { |
duke@435 | 505 | MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
duke@435 | 506 | unsigned int size = allocation_size(cb, sizeof(SafepointBlob)); |
duke@435 | 507 | blob = new (size) SafepointBlob(cb, size, oop_maps, frame_size); |
duke@435 | 508 | } |
duke@435 | 509 | |
never@2895 | 510 | trace_new_stub(blob, "SafepointBlob"); |
duke@435 | 511 | |
duke@435 | 512 | return blob; |
duke@435 | 513 | } |
duke@435 | 514 | |
duke@435 | 515 | |
duke@435 | 516 | //---------------------------------------------------------------------------------------------------- |
duke@435 | 517 | // Verification and printing |
duke@435 | 518 | |
duke@435 | 519 | void CodeBlob::verify() { |
duke@435 | 520 | ShouldNotReachHere(); |
duke@435 | 521 | } |
duke@435 | 522 | |
bobv@2036 | 523 | void CodeBlob::print_on(outputStream* st) const { |
bobv@2036 | 524 | st->print_cr("[CodeBlob (" INTPTR_FORMAT ")]", this); |
bobv@2036 | 525 | st->print_cr("Framesize: %d", _frame_size); |
duke@435 | 526 | } |
duke@435 | 527 | |
duke@435 | 528 | void CodeBlob::print_value_on(outputStream* st) const { |
duke@435 | 529 | st->print_cr("[CodeBlob]"); |
duke@435 | 530 | } |
duke@435 | 531 | |
duke@435 | 532 | void BufferBlob::verify() { |
duke@435 | 533 | // unimplemented |
duke@435 | 534 | } |
duke@435 | 535 | |
bobv@2036 | 536 | void BufferBlob::print_on(outputStream* st) const { |
bobv@2036 | 537 | CodeBlob::print_on(st); |
bobv@2036 | 538 | print_value_on(st); |
duke@435 | 539 | } |
duke@435 | 540 | |
duke@435 | 541 | void BufferBlob::print_value_on(outputStream* st) const { |
duke@435 | 542 | st->print_cr("BufferBlob (" INTPTR_FORMAT ") used for %s", this, name()); |
duke@435 | 543 | } |
duke@435 | 544 | |
duke@435 | 545 | void RuntimeStub::verify() { |
duke@435 | 546 | // unimplemented |
duke@435 | 547 | } |
duke@435 | 548 | |
bobv@2036 | 549 | void RuntimeStub::print_on(outputStream* st) const { |
bobv@2036 | 550 | CodeBlob::print_on(st); |
bobv@2036 | 551 | st->print("Runtime Stub (" INTPTR_FORMAT "): ", this); |
bobv@2036 | 552 | st->print_cr(name()); |
bobv@2036 | 553 | Disassembler::decode((CodeBlob*)this, st); |
duke@435 | 554 | } |
duke@435 | 555 | |
duke@435 | 556 | void RuntimeStub::print_value_on(outputStream* st) const { |
duke@435 | 557 | st->print("RuntimeStub (" INTPTR_FORMAT "): ", this); st->print(name()); |
duke@435 | 558 | } |
duke@435 | 559 | |
duke@435 | 560 | void SingletonBlob::verify() { |
duke@435 | 561 | // unimplemented |
duke@435 | 562 | } |
duke@435 | 563 | |
bobv@2036 | 564 | void SingletonBlob::print_on(outputStream* st) const { |
bobv@2036 | 565 | CodeBlob::print_on(st); |
bobv@2036 | 566 | st->print_cr(name()); |
bobv@2036 | 567 | Disassembler::decode((CodeBlob*)this, st); |
duke@435 | 568 | } |
duke@435 | 569 | |
duke@435 | 570 | void SingletonBlob::print_value_on(outputStream* st) const { |
duke@435 | 571 | st->print_cr(name()); |
duke@435 | 572 | } |
duke@435 | 573 | |
duke@435 | 574 | void DeoptimizationBlob::print_value_on(outputStream* st) const { |
duke@435 | 575 | st->print_cr("Deoptimization (frame not available)"); |
duke@435 | 576 | } |