|
1 /* |
|
2 * Copyright (c) 1998, 2014, Oracle and/or its affiliates. All rights reserved. |
|
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
|
4 * |
|
5 * This code is free software; you can redistribute it and/or modify it |
|
6 * under the terms of the GNU General Public License version 2 only, as |
|
7 * published by the Free Software Foundation. |
|
8 * |
|
9 * This code is distributed in the hope that it will be useful, but WITHOUT |
|
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
|
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
|
12 * version 2 for more details (a copy is included in the LICENSE file that |
|
13 * accompanied this code). |
|
14 * |
|
15 * You should have received a copy of the GNU General Public License version |
|
16 * 2 along with this work; if not, write to the Free Software Foundation, |
|
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
|
18 * |
|
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
|
20 * or visit www.oracle.com if you need additional information or have any |
|
21 * questions. |
|
22 * |
|
23 */ |
|
24 |
|
25 #include "precompiled.hpp" |
|
26 #include "code/codeBlob.hpp" |
|
27 #include "code/codeCache.hpp" |
|
28 #include "code/relocInfo.hpp" |
|
29 #include "compiler/disassembler.hpp" |
|
30 #include "interpreter/bytecode.hpp" |
|
31 #include "memory/allocation.inline.hpp" |
|
32 #include "memory/heap.hpp" |
|
33 #include "oops/oop.inline.hpp" |
|
34 #include "prims/forte.hpp" |
|
35 #include "runtime/handles.inline.hpp" |
|
36 #include "runtime/interfaceSupport.hpp" |
|
37 #include "runtime/mutexLocker.hpp" |
|
38 #include "runtime/safepoint.hpp" |
|
39 #include "runtime/sharedRuntime.hpp" |
|
40 #include "runtime/vframe.hpp" |
|
41 #include "services/memoryService.hpp" |
|
42 #ifdef TARGET_ARCH_x86 |
|
43 # include "nativeInst_x86.hpp" |
|
44 #endif |
|
45 #ifdef TARGET_ARCH_sparc |
|
46 # include "nativeInst_sparc.hpp" |
|
47 #endif |
|
48 #ifdef TARGET_ARCH_zero |
|
49 # include "nativeInst_zero.hpp" |
|
50 #endif |
|
51 #ifdef TARGET_ARCH_arm |
|
52 # include "nativeInst_arm.hpp" |
|
53 #endif |
|
54 #ifdef TARGET_ARCH_ppc |
|
55 # include "nativeInst_ppc.hpp" |
|
56 #endif |
|
57 #ifdef COMPILER1 |
|
58 #include "c1/c1_Runtime1.hpp" |
|
59 #endif |
|
60 |
|
61 unsigned int align_code_offset(int offset) { |
|
62 // align the size to CodeEntryAlignment |
|
63 return |
|
64 ((offset + (int)CodeHeap::header_size() + (CodeEntryAlignment-1)) & ~(CodeEntryAlignment-1)) |
|
65 - (int)CodeHeap::header_size(); |
|
66 } |
|
67 |
|
68 |
|
69 // This must be consistent with the CodeBlob constructor's layout actions. |
|
70 unsigned int CodeBlob::allocation_size(CodeBuffer* cb, int header_size) { |
|
71 unsigned int size = header_size; |
|
72 size += round_to(cb->total_relocation_size(), oopSize); |
|
73 // align the size to CodeEntryAlignment |
|
74 size = align_code_offset(size); |
|
75 size += round_to(cb->total_content_size(), oopSize); |
|
76 size += round_to(cb->total_oop_size(), oopSize); |
|
77 size += round_to(cb->total_metadata_size(), oopSize); |
|
78 return size; |
|
79 } |
|
80 |
|
81 |
|
82 // Creates a simple CodeBlob. Sets up the size of the different regions. |
|
83 CodeBlob::CodeBlob(const char* name, int header_size, int size, int frame_complete, int locs_size) { |
|
84 assert(size == round_to(size, oopSize), "unaligned size"); |
|
85 assert(locs_size == round_to(locs_size, oopSize), "unaligned size"); |
|
86 assert(header_size == round_to(header_size, oopSize), "unaligned size"); |
|
87 assert(!UseRelocIndex, "no space allocated for reloc index yet"); |
|
88 |
|
89 // Note: If UseRelocIndex is enabled, there needs to be (at least) one |
|
90 // extra word for the relocation information, containing the reloc |
|
91 // index table length. Unfortunately, the reloc index table imple- |
|
92 // mentation is not easily understandable and thus it is not clear |
|
93 // what exactly the format is supposed to be. For now, we just turn |
|
94 // off the use of this table (gri 7/6/2000). |
|
95 |
|
96 _name = name; |
|
97 _size = size; |
|
98 _frame_complete_offset = frame_complete; |
|
99 _header_size = header_size; |
|
100 _relocation_size = locs_size; |
|
101 _content_offset = align_code_offset(header_size + _relocation_size); |
|
102 _code_offset = _content_offset; |
|
103 _data_offset = size; |
|
104 _frame_size = 0; |
|
105 set_oop_maps(NULL); |
|
106 } |
|
107 |
|
108 |
|
109 // Creates a CodeBlob from a CodeBuffer. Sets up the size of the different regions, |
|
110 // and copy code and relocation info. |
|
111 CodeBlob::CodeBlob( |
|
112 const char* name, |
|
113 CodeBuffer* cb, |
|
114 int header_size, |
|
115 int size, |
|
116 int frame_complete, |
|
117 int frame_size, |
|
118 OopMapSet* oop_maps |
|
119 ) { |
|
120 assert(size == round_to(size, oopSize), "unaligned size"); |
|
121 assert(header_size == round_to(header_size, oopSize), "unaligned size"); |
|
122 |
|
123 _name = name; |
|
124 _size = size; |
|
125 _frame_complete_offset = frame_complete; |
|
126 _header_size = header_size; |
|
127 _relocation_size = round_to(cb->total_relocation_size(), oopSize); |
|
128 _content_offset = align_code_offset(header_size + _relocation_size); |
|
129 _code_offset = _content_offset + cb->total_offset_of(cb->insts()); |
|
130 _data_offset = _content_offset + round_to(cb->total_content_size(), oopSize); |
|
131 assert(_data_offset <= size, "codeBlob is too small"); |
|
132 |
|
133 cb->copy_code_and_locs_to(this); |
|
134 set_oop_maps(oop_maps); |
|
135 _frame_size = frame_size; |
|
136 #ifdef COMPILER1 |
|
137 // probably wrong for tiered |
|
138 assert(_frame_size >= -1, "must use frame size or -1 for runtime stubs"); |
|
139 #endif // COMPILER1 |
|
140 } |
|
141 |
|
142 |
|
143 void CodeBlob::set_oop_maps(OopMapSet* p) { |
|
144 // Danger Will Robinson! This method allocates a big |
|
145 // chunk of memory, its your job to free it. |
|
146 if (p != NULL) { |
|
147 // We need to allocate a chunk big enough to hold the OopMapSet and all of its OopMaps |
|
148 _oop_maps = (OopMapSet* )NEW_C_HEAP_ARRAY(unsigned char, p->heap_size(), mtCode); |
|
149 p->copy_to((address)_oop_maps); |
|
150 } else { |
|
151 _oop_maps = NULL; |
|
152 } |
|
153 } |
|
154 |
|
155 |
|
156 void CodeBlob::trace_new_stub(CodeBlob* stub, const char* name1, const char* name2) { |
|
157 // Do not hold the CodeCache lock during name formatting. |
|
158 assert(!CodeCache_lock->owned_by_self(), "release CodeCache before registering the stub"); |
|
159 |
|
160 if (stub != NULL) { |
|
161 char stub_id[256]; |
|
162 assert(strlen(name1) + strlen(name2) < sizeof(stub_id), ""); |
|
163 jio_snprintf(stub_id, sizeof(stub_id), "%s%s", name1, name2); |
|
164 if (PrintStubCode) { |
|
165 ttyLocker ttyl; |
|
166 tty->print_cr("Decoding %s " INTPTR_FORMAT, stub_id, (intptr_t) stub); |
|
167 Disassembler::decode(stub->code_begin(), stub->code_end()); |
|
168 tty->cr(); |
|
169 } |
|
170 Forte::register_stub(stub_id, stub->code_begin(), stub->code_end()); |
|
171 |
|
172 if (JvmtiExport::should_post_dynamic_code_generated()) { |
|
173 const char* stub_name = name2; |
|
174 if (name2[0] == '\0') stub_name = name1; |
|
175 JvmtiExport::post_dynamic_code_generated(stub_name, stub->code_begin(), stub->code_end()); |
|
176 } |
|
177 } |
|
178 |
|
179 // Track memory usage statistic after releasing CodeCache_lock |
|
180 MemoryService::track_code_cache_memory_usage(); |
|
181 } |
|
182 |
|
183 |
|
184 void CodeBlob::flush() { |
|
185 if (_oop_maps) { |
|
186 FREE_C_HEAP_ARRAY(unsigned char, _oop_maps, mtCode); |
|
187 _oop_maps = NULL; |
|
188 } |
|
189 _strings.free(); |
|
190 } |
|
191 |
|
192 |
|
193 OopMap* CodeBlob::oop_map_for_return_address(address return_address) { |
|
194 assert(oop_maps() != NULL, "nope"); |
|
195 return oop_maps()->find_map_at_offset((intptr_t) return_address - (intptr_t) code_begin()); |
|
196 } |
|
197 |
|
198 |
|
199 //---------------------------------------------------------------------------------------------------- |
|
200 // Implementation of BufferBlob |
|
201 |
|
202 |
|
203 BufferBlob::BufferBlob(const char* name, int size) |
|
204 : CodeBlob(name, sizeof(BufferBlob), size, CodeOffsets::frame_never_safe, /*locs_size:*/ 0) |
|
205 {} |
|
206 |
|
207 BufferBlob* BufferBlob::create(const char* name, int buffer_size) { |
|
208 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
|
209 |
|
210 BufferBlob* blob = NULL; |
|
211 unsigned int size = sizeof(BufferBlob); |
|
212 // align the size to CodeEntryAlignment |
|
213 size = align_code_offset(size); |
|
214 size += round_to(buffer_size, oopSize); |
|
215 assert(name != NULL, "must provide a name"); |
|
216 { |
|
217 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
|
218 blob = new (size) BufferBlob(name, size); |
|
219 } |
|
220 // Track memory usage statistic after releasing CodeCache_lock |
|
221 MemoryService::track_code_cache_memory_usage(); |
|
222 |
|
223 return blob; |
|
224 } |
|
225 |
|
226 |
|
227 BufferBlob::BufferBlob(const char* name, int size, CodeBuffer* cb) |
|
228 : CodeBlob(name, cb, sizeof(BufferBlob), size, CodeOffsets::frame_never_safe, 0, NULL) |
|
229 {} |
|
230 |
|
231 BufferBlob* BufferBlob::create(const char* name, CodeBuffer* cb) { |
|
232 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
|
233 |
|
234 BufferBlob* blob = NULL; |
|
235 unsigned int size = allocation_size(cb, sizeof(BufferBlob)); |
|
236 assert(name != NULL, "must provide a name"); |
|
237 { |
|
238 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
|
239 blob = new (size) BufferBlob(name, size, cb); |
|
240 } |
|
241 // Track memory usage statistic after releasing CodeCache_lock |
|
242 MemoryService::track_code_cache_memory_usage(); |
|
243 |
|
244 return blob; |
|
245 } |
|
246 |
|
247 |
|
248 void* BufferBlob::operator new(size_t s, unsigned size, bool is_critical) throw() { |
|
249 void* p = CodeCache::allocate(size, is_critical); |
|
250 return p; |
|
251 } |
|
252 |
|
253 |
|
254 void BufferBlob::free( BufferBlob *blob ) { |
|
255 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
|
256 { |
|
257 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
|
258 CodeCache::free((CodeBlob*)blob); |
|
259 } |
|
260 // Track memory usage statistic after releasing CodeCache_lock |
|
261 MemoryService::track_code_cache_memory_usage(); |
|
262 } |
|
263 |
|
264 |
|
265 //---------------------------------------------------------------------------------------------------- |
|
266 // Implementation of AdapterBlob |
|
267 |
|
268 AdapterBlob::AdapterBlob(int size, CodeBuffer* cb) : |
|
269 BufferBlob("I2C/C2I adapters", size, cb) { |
|
270 CodeCache::commit(this); |
|
271 } |
|
272 |
|
273 AdapterBlob* AdapterBlob::create(CodeBuffer* cb) { |
|
274 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
|
275 |
|
276 AdapterBlob* blob = NULL; |
|
277 unsigned int size = allocation_size(cb, sizeof(AdapterBlob)); |
|
278 { |
|
279 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
|
280 // The parameter 'true' indicates a critical memory allocation. |
|
281 // This means that CodeCacheMinimumFreeSpace is used, if necessary |
|
282 const bool is_critical = true; |
|
283 blob = new (size, is_critical) AdapterBlob(size, cb); |
|
284 } |
|
285 // Track memory usage statistic after releasing CodeCache_lock |
|
286 MemoryService::track_code_cache_memory_usage(); |
|
287 |
|
288 return blob; |
|
289 } |
|
290 |
|
291 |
|
292 //---------------------------------------------------------------------------------------------------- |
|
293 // Implementation of MethodHandlesAdapterBlob |
|
294 |
|
295 MethodHandlesAdapterBlob* MethodHandlesAdapterBlob::create(int buffer_size) { |
|
296 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
|
297 |
|
298 MethodHandlesAdapterBlob* blob = NULL; |
|
299 unsigned int size = sizeof(MethodHandlesAdapterBlob); |
|
300 // align the size to CodeEntryAlignment |
|
301 size = align_code_offset(size); |
|
302 size += round_to(buffer_size, oopSize); |
|
303 { |
|
304 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
|
305 // The parameter 'true' indicates a critical memory allocation. |
|
306 // This means that CodeCacheMinimumFreeSpace is used, if necessary |
|
307 const bool is_critical = true; |
|
308 blob = new (size, is_critical) MethodHandlesAdapterBlob(size); |
|
309 } |
|
310 // Track memory usage statistic after releasing CodeCache_lock |
|
311 MemoryService::track_code_cache_memory_usage(); |
|
312 |
|
313 return blob; |
|
314 } |
|
315 |
|
316 |
|
317 //---------------------------------------------------------------------------------------------------- |
|
318 // Implementation of RuntimeStub |
|
319 |
|
320 RuntimeStub::RuntimeStub( |
|
321 const char* name, |
|
322 CodeBuffer* cb, |
|
323 int size, |
|
324 int frame_complete, |
|
325 int frame_size, |
|
326 OopMapSet* oop_maps, |
|
327 bool caller_must_gc_arguments |
|
328 ) |
|
329 : CodeBlob(name, cb, sizeof(RuntimeStub), size, frame_complete, frame_size, oop_maps) |
|
330 { |
|
331 _caller_must_gc_arguments = caller_must_gc_arguments; |
|
332 } |
|
333 |
|
334 |
|
335 RuntimeStub* RuntimeStub::new_runtime_stub(const char* stub_name, |
|
336 CodeBuffer* cb, |
|
337 int frame_complete, |
|
338 int frame_size, |
|
339 OopMapSet* oop_maps, |
|
340 bool caller_must_gc_arguments) |
|
341 { |
|
342 RuntimeStub* stub = NULL; |
|
343 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
|
344 { |
|
345 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
|
346 unsigned int size = allocation_size(cb, sizeof(RuntimeStub)); |
|
347 stub = new (size) RuntimeStub(stub_name, cb, size, frame_complete, frame_size, oop_maps, caller_must_gc_arguments); |
|
348 } |
|
349 |
|
350 trace_new_stub(stub, "RuntimeStub - ", stub_name); |
|
351 |
|
352 return stub; |
|
353 } |
|
354 |
|
355 |
|
356 void* RuntimeStub::operator new(size_t s, unsigned size) throw() { |
|
357 void* p = CodeCache::allocate(size, true); |
|
358 if (!p) fatal("Initial size of CodeCache is too small"); |
|
359 return p; |
|
360 } |
|
361 |
|
362 // operator new shared by all singletons: |
|
363 void* SingletonBlob::operator new(size_t s, unsigned size) throw() { |
|
364 void* p = CodeCache::allocate(size, true); |
|
365 if (!p) fatal("Initial size of CodeCache is too small"); |
|
366 return p; |
|
367 } |
|
368 |
|
369 |
|
370 //---------------------------------------------------------------------------------------------------- |
|
371 // Implementation of DeoptimizationBlob |
|
372 |
|
373 DeoptimizationBlob::DeoptimizationBlob( |
|
374 CodeBuffer* cb, |
|
375 int size, |
|
376 OopMapSet* oop_maps, |
|
377 int unpack_offset, |
|
378 int unpack_with_exception_offset, |
|
379 int unpack_with_reexecution_offset, |
|
380 int frame_size |
|
381 ) |
|
382 : SingletonBlob("DeoptimizationBlob", cb, sizeof(DeoptimizationBlob), size, frame_size, oop_maps) |
|
383 { |
|
384 _unpack_offset = unpack_offset; |
|
385 _unpack_with_exception = unpack_with_exception_offset; |
|
386 _unpack_with_reexecution = unpack_with_reexecution_offset; |
|
387 #ifdef COMPILER1 |
|
388 _unpack_with_exception_in_tls = -1; |
|
389 #endif |
|
390 } |
|
391 |
|
392 |
|
393 DeoptimizationBlob* DeoptimizationBlob::create( |
|
394 CodeBuffer* cb, |
|
395 OopMapSet* oop_maps, |
|
396 int unpack_offset, |
|
397 int unpack_with_exception_offset, |
|
398 int unpack_with_reexecution_offset, |
|
399 int frame_size) |
|
400 { |
|
401 DeoptimizationBlob* blob = NULL; |
|
402 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
|
403 { |
|
404 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
|
405 unsigned int size = allocation_size(cb, sizeof(DeoptimizationBlob)); |
|
406 blob = new (size) DeoptimizationBlob(cb, |
|
407 size, |
|
408 oop_maps, |
|
409 unpack_offset, |
|
410 unpack_with_exception_offset, |
|
411 unpack_with_reexecution_offset, |
|
412 frame_size); |
|
413 } |
|
414 |
|
415 trace_new_stub(blob, "DeoptimizationBlob"); |
|
416 |
|
417 return blob; |
|
418 } |
|
419 |
|
420 |
|
421 //---------------------------------------------------------------------------------------------------- |
|
422 // Implementation of UncommonTrapBlob |
|
423 |
|
424 #ifdef COMPILER2 |
|
425 UncommonTrapBlob::UncommonTrapBlob( |
|
426 CodeBuffer* cb, |
|
427 int size, |
|
428 OopMapSet* oop_maps, |
|
429 int frame_size |
|
430 ) |
|
431 : SingletonBlob("UncommonTrapBlob", cb, sizeof(UncommonTrapBlob), size, frame_size, oop_maps) |
|
432 {} |
|
433 |
|
434 |
|
435 UncommonTrapBlob* UncommonTrapBlob::create( |
|
436 CodeBuffer* cb, |
|
437 OopMapSet* oop_maps, |
|
438 int frame_size) |
|
439 { |
|
440 UncommonTrapBlob* blob = NULL; |
|
441 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
|
442 { |
|
443 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
|
444 unsigned int size = allocation_size(cb, sizeof(UncommonTrapBlob)); |
|
445 blob = new (size) UncommonTrapBlob(cb, size, oop_maps, frame_size); |
|
446 } |
|
447 |
|
448 trace_new_stub(blob, "UncommonTrapBlob"); |
|
449 |
|
450 return blob; |
|
451 } |
|
452 |
|
453 |
|
454 #endif // COMPILER2 |
|
455 |
|
456 |
|
457 //---------------------------------------------------------------------------------------------------- |
|
458 // Implementation of ExceptionBlob |
|
459 |
|
460 #ifdef COMPILER2 |
|
461 ExceptionBlob::ExceptionBlob( |
|
462 CodeBuffer* cb, |
|
463 int size, |
|
464 OopMapSet* oop_maps, |
|
465 int frame_size |
|
466 ) |
|
467 : SingletonBlob("ExceptionBlob", cb, sizeof(ExceptionBlob), size, frame_size, oop_maps) |
|
468 {} |
|
469 |
|
470 |
|
471 ExceptionBlob* ExceptionBlob::create( |
|
472 CodeBuffer* cb, |
|
473 OopMapSet* oop_maps, |
|
474 int frame_size) |
|
475 { |
|
476 ExceptionBlob* blob = NULL; |
|
477 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
|
478 { |
|
479 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
|
480 unsigned int size = allocation_size(cb, sizeof(ExceptionBlob)); |
|
481 blob = new (size) ExceptionBlob(cb, size, oop_maps, frame_size); |
|
482 } |
|
483 |
|
484 trace_new_stub(blob, "ExceptionBlob"); |
|
485 |
|
486 return blob; |
|
487 } |
|
488 |
|
489 |
|
490 #endif // COMPILER2 |
|
491 |
|
492 |
|
493 //---------------------------------------------------------------------------------------------------- |
|
494 // Implementation of SafepointBlob |
|
495 |
|
496 SafepointBlob::SafepointBlob( |
|
497 CodeBuffer* cb, |
|
498 int size, |
|
499 OopMapSet* oop_maps, |
|
500 int frame_size |
|
501 ) |
|
502 : SingletonBlob("SafepointBlob", cb, sizeof(SafepointBlob), size, frame_size, oop_maps) |
|
503 {} |
|
504 |
|
505 |
|
506 SafepointBlob* SafepointBlob::create( |
|
507 CodeBuffer* cb, |
|
508 OopMapSet* oop_maps, |
|
509 int frame_size) |
|
510 { |
|
511 SafepointBlob* blob = NULL; |
|
512 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock |
|
513 { |
|
514 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag); |
|
515 unsigned int size = allocation_size(cb, sizeof(SafepointBlob)); |
|
516 blob = new (size) SafepointBlob(cb, size, oop_maps, frame_size); |
|
517 } |
|
518 |
|
519 trace_new_stub(blob, "SafepointBlob"); |
|
520 |
|
521 return blob; |
|
522 } |
|
523 |
|
524 |
|
525 //---------------------------------------------------------------------------------------------------- |
|
526 // Verification and printing |
|
527 |
|
528 void CodeBlob::verify() { |
|
529 ShouldNotReachHere(); |
|
530 } |
|
531 |
|
532 void CodeBlob::print_on(outputStream* st) const { |
|
533 st->print_cr("[CodeBlob (" INTPTR_FORMAT ")]", p2i(this)); |
|
534 st->print_cr("Framesize: %d", _frame_size); |
|
535 } |
|
536 |
|
537 void CodeBlob::print_value_on(outputStream* st) const { |
|
538 st->print_cr("[CodeBlob]"); |
|
539 } |
|
540 |
|
541 void BufferBlob::verify() { |
|
542 // unimplemented |
|
543 } |
|
544 |
|
545 void BufferBlob::print_on(outputStream* st) const { |
|
546 CodeBlob::print_on(st); |
|
547 print_value_on(st); |
|
548 } |
|
549 |
|
550 void BufferBlob::print_value_on(outputStream* st) const { |
|
551 st->print_cr("BufferBlob (" INTPTR_FORMAT ") used for %s", p2i(this), name()); |
|
552 } |
|
553 |
|
554 void RuntimeStub::verify() { |
|
555 // unimplemented |
|
556 } |
|
557 |
|
558 void RuntimeStub::print_on(outputStream* st) const { |
|
559 ttyLocker ttyl; |
|
560 CodeBlob::print_on(st); |
|
561 st->print("Runtime Stub (" INTPTR_FORMAT "): ", p2i(this)); |
|
562 st->print_cr("%s", name()); |
|
563 Disassembler::decode((CodeBlob*)this, st); |
|
564 } |
|
565 |
|
566 void RuntimeStub::print_value_on(outputStream* st) const { |
|
567 st->print("RuntimeStub (" INTPTR_FORMAT "): ", p2i(this)); st->print("%s", name()); |
|
568 } |
|
569 |
|
570 void SingletonBlob::verify() { |
|
571 // unimplemented |
|
572 } |
|
573 |
|
574 void SingletonBlob::print_on(outputStream* st) const { |
|
575 ttyLocker ttyl; |
|
576 CodeBlob::print_on(st); |
|
577 st->print_cr("%s", name()); |
|
578 Disassembler::decode((CodeBlob*)this, st); |
|
579 } |
|
580 |
|
581 void SingletonBlob::print_value_on(outputStream* st) const { |
|
582 st->print_cr("%s", name()); |
|
583 } |
|
584 |
|
585 void DeoptimizationBlob::print_value_on(outputStream* st) const { |
|
586 st->print_cr("Deoptimization (frame not available)"); |
|
587 } |