Fri, 30 Mar 2018 20:09:45 +0000
8199406: Performance drop with Java JDK 1.8.0_162-b32
Summary: Improve the nmethod unloading times by optimizing the search for an itable stub in VtableStubs array
Reviewed-by: kvn, coleenp, tschatzl
1 /*
2 * Copyright (c) 1998, 2018, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
25 #include "precompiled.hpp"
26 #include "code/codeBlob.hpp"
27 #include "code/codeCache.hpp"
28 #include "code/relocInfo.hpp"
29 #include "compiler/disassembler.hpp"
30 #include "interpreter/bytecode.hpp"
31 #include "memory/allocation.inline.hpp"
32 #include "memory/heap.hpp"
33 #include "oops/oop.inline.hpp"
34 #include "prims/forte.hpp"
35 #include "runtime/handles.inline.hpp"
36 #include "runtime/interfaceSupport.hpp"
37 #include "runtime/mutexLocker.hpp"
38 #include "runtime/safepoint.hpp"
39 #include "runtime/sharedRuntime.hpp"
40 #include "runtime/vframe.hpp"
41 #include "services/memoryService.hpp"
42 #ifdef TARGET_ARCH_x86
43 # include "nativeInst_x86.hpp"
44 #endif
45 #ifdef TARGET_ARCH_sparc
46 # include "nativeInst_sparc.hpp"
47 #endif
48 #ifdef TARGET_ARCH_zero
49 # include "nativeInst_zero.hpp"
50 #endif
51 #ifdef TARGET_ARCH_arm
52 # include "nativeInst_arm.hpp"
53 #endif
54 #ifdef TARGET_ARCH_ppc
55 # include "nativeInst_ppc.hpp"
56 #endif
57 #ifdef COMPILER1
58 #include "c1/c1_Runtime1.hpp"
59 #endif
61 unsigned int align_code_offset(int offset) {
62 // align the size to CodeEntryAlignment
63 return
64 ((offset + (int)CodeHeap::header_size() + (CodeEntryAlignment-1)) & ~(CodeEntryAlignment-1))
65 - (int)CodeHeap::header_size();
66 }
69 // This must be consistent with the CodeBlob constructor's layout actions.
70 unsigned int CodeBlob::allocation_size(CodeBuffer* cb, int header_size) {
71 unsigned int size = header_size;
72 size += round_to(cb->total_relocation_size(), oopSize);
73 // align the size to CodeEntryAlignment
74 size = align_code_offset(size);
75 size += round_to(cb->total_content_size(), oopSize);
76 size += round_to(cb->total_oop_size(), oopSize);
77 size += round_to(cb->total_metadata_size(), oopSize);
78 return size;
79 }
82 // Creates a simple CodeBlob. Sets up the size of the different regions.
83 CodeBlob::CodeBlob(const char* name, int header_size, int size, int frame_complete, int locs_size) {
84 assert(size == round_to(size, oopSize), "unaligned size");
85 assert(locs_size == round_to(locs_size, oopSize), "unaligned size");
86 assert(header_size == round_to(header_size, oopSize), "unaligned size");
87 assert(!UseRelocIndex, "no space allocated for reloc index yet");
89 // Note: If UseRelocIndex is enabled, there needs to be (at least) one
90 // extra word for the relocation information, containing the reloc
91 // index table length. Unfortunately, the reloc index table imple-
92 // mentation is not easily understandable and thus it is not clear
93 // what exactly the format is supposed to be. For now, we just turn
94 // off the use of this table (gri 7/6/2000).
96 _name = name;
97 _size = size;
98 _frame_complete_offset = frame_complete;
99 _header_size = header_size;
100 _relocation_size = locs_size;
101 _content_offset = align_code_offset(header_size + _relocation_size);
102 _code_offset = _content_offset;
103 _data_offset = size;
104 _frame_size = 0;
105 set_oop_maps(NULL);
106 }
109 // Creates a CodeBlob from a CodeBuffer. Sets up the size of the different regions,
110 // and copy code and relocation info.
111 CodeBlob::CodeBlob(
112 const char* name,
113 CodeBuffer* cb,
114 int header_size,
115 int size,
116 int frame_complete,
117 int frame_size,
118 OopMapSet* oop_maps
119 ) {
120 assert(size == round_to(size, oopSize), "unaligned size");
121 assert(header_size == round_to(header_size, oopSize), "unaligned size");
123 _name = name;
124 _size = size;
125 _frame_complete_offset = frame_complete;
126 _header_size = header_size;
127 _relocation_size = round_to(cb->total_relocation_size(), oopSize);
128 _content_offset = align_code_offset(header_size + _relocation_size);
129 _code_offset = _content_offset + cb->total_offset_of(cb->insts());
130 _data_offset = _content_offset + round_to(cb->total_content_size(), oopSize);
131 assert(_data_offset <= size, "codeBlob is too small");
133 cb->copy_code_and_locs_to(this);
134 set_oop_maps(oop_maps);
135 _frame_size = frame_size;
136 #ifdef COMPILER1
137 // probably wrong for tiered
138 assert(_frame_size >= -1, "must use frame size or -1 for runtime stubs");
139 #endif // COMPILER1
140 }
143 void CodeBlob::set_oop_maps(OopMapSet* p) {
144 // Danger Will Robinson! This method allocates a big
145 // chunk of memory, its your job to free it.
146 if (p != NULL) {
147 // We need to allocate a chunk big enough to hold the OopMapSet and all of its OopMaps
148 _oop_maps = (OopMapSet* )NEW_C_HEAP_ARRAY(unsigned char, p->heap_size(), mtCode);
149 p->copy_to((address)_oop_maps);
150 } else {
151 _oop_maps = NULL;
152 }
153 }
156 void CodeBlob::trace_new_stub(CodeBlob* stub, const char* name1, const char* name2) {
157 // Do not hold the CodeCache lock during name formatting.
158 assert(!CodeCache_lock->owned_by_self(), "release CodeCache before registering the stub");
160 if (stub != NULL) {
161 char stub_id[256];
162 assert(strlen(name1) + strlen(name2) < sizeof(stub_id), "");
163 jio_snprintf(stub_id, sizeof(stub_id), "%s%s", name1, name2);
164 if (PrintStubCode) {
165 ttyLocker ttyl;
166 tty->print_cr("Decoding %s " INTPTR_FORMAT, stub_id, (intptr_t) stub);
167 Disassembler::decode(stub->code_begin(), stub->code_end());
168 tty->cr();
169 }
170 Forte::register_stub(stub_id, stub->code_begin(), stub->code_end());
172 if (JvmtiExport::should_post_dynamic_code_generated()) {
173 const char* stub_name = name2;
174 if (name2[0] == '\0') stub_name = name1;
175 JvmtiExport::post_dynamic_code_generated(stub_name, stub->code_begin(), stub->code_end());
176 }
177 }
179 // Track memory usage statistic after releasing CodeCache_lock
180 MemoryService::track_code_cache_memory_usage();
181 }
184 void CodeBlob::flush() {
185 if (_oop_maps) {
186 FREE_C_HEAP_ARRAY(unsigned char, _oop_maps, mtCode);
187 _oop_maps = NULL;
188 }
189 _strings.free();
190 }
193 OopMap* CodeBlob::oop_map_for_return_address(address return_address) {
194 assert(oop_maps() != NULL, "nope");
195 return oop_maps()->find_map_at_offset((intptr_t) return_address - (intptr_t) code_begin());
196 }
199 //----------------------------------------------------------------------------------------------------
200 // Implementation of BufferBlob
203 BufferBlob::BufferBlob(const char* name, int size)
204 : CodeBlob(name, sizeof(BufferBlob), size, CodeOffsets::frame_never_safe, /*locs_size:*/ 0)
205 {}
207 BufferBlob* BufferBlob::create(const char* name, int buffer_size) {
208 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock
210 BufferBlob* blob = NULL;
211 unsigned int size = sizeof(BufferBlob);
212 // align the size to CodeEntryAlignment
213 size = align_code_offset(size);
214 size += round_to(buffer_size, oopSize);
215 assert(name != NULL, "must provide a name");
216 {
217 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
218 blob = new (size) BufferBlob(name, size);
219 }
220 // Track memory usage statistic after releasing CodeCache_lock
221 MemoryService::track_code_cache_memory_usage();
223 return blob;
224 }
227 BufferBlob::BufferBlob(const char* name, int size, CodeBuffer* cb)
228 : CodeBlob(name, cb, sizeof(BufferBlob), size, CodeOffsets::frame_never_safe, 0, NULL)
229 {}
231 BufferBlob* BufferBlob::create(const char* name, CodeBuffer* cb) {
232 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock
234 BufferBlob* blob = NULL;
235 unsigned int size = allocation_size(cb, sizeof(BufferBlob));
236 assert(name != NULL, "must provide a name");
237 {
238 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
239 blob = new (size) BufferBlob(name, size, cb);
240 }
241 // Track memory usage statistic after releasing CodeCache_lock
242 MemoryService::track_code_cache_memory_usage();
244 return blob;
245 }
248 void* BufferBlob::operator new(size_t s, unsigned size, bool is_critical) throw() {
249 void* p = CodeCache::allocate(size, is_critical);
250 return p;
251 }
254 void BufferBlob::free( BufferBlob *blob ) {
255 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock
256 blob->flush();
257 {
258 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
259 CodeCache::free((CodeBlob*)blob);
260 }
261 // Track memory usage statistic after releasing CodeCache_lock
262 MemoryService::track_code_cache_memory_usage();
263 }
266 //----------------------------------------------------------------------------------------------------
267 // Implementation of AdapterBlob
269 AdapterBlob::AdapterBlob(int size, CodeBuffer* cb) :
270 BufferBlob("I2C/C2I adapters", size, cb) {
271 CodeCache::commit(this);
272 }
274 AdapterBlob* AdapterBlob::create(CodeBuffer* cb) {
275 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock
277 AdapterBlob* blob = NULL;
278 unsigned int size = allocation_size(cb, sizeof(AdapterBlob));
279 {
280 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
281 // The parameter 'true' indicates a critical memory allocation.
282 // This means that CodeCacheMinimumFreeSpace is used, if necessary
283 const bool is_critical = true;
284 blob = new (size, is_critical) AdapterBlob(size, cb);
285 }
286 // Track memory usage statistic after releasing CodeCache_lock
287 MemoryService::track_code_cache_memory_usage();
289 return blob;
290 }
292 VtableBlob::VtableBlob(const char* name, int size) :
293 BufferBlob(name, size) {
294 }
296 VtableBlob* VtableBlob::create(const char* name, int buffer_size) {
297 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock
299 VtableBlob* blob = NULL;
300 unsigned int size = sizeof(VtableBlob);
301 // align the size to CodeEntryAlignment
302 size = align_code_offset(size);
303 size += round_to(buffer_size, oopSize);
304 assert(name != NULL, "must provide a name");
305 {
306 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
307 blob = new (size) VtableBlob(name, size);
308 }
309 // Track memory usage statistic after releasing CodeCache_lock
310 MemoryService::track_code_cache_memory_usage();
312 return blob;
313 }
315 //----------------------------------------------------------------------------------------------------
316 // Implementation of MethodHandlesAdapterBlob
318 MethodHandlesAdapterBlob* MethodHandlesAdapterBlob::create(int buffer_size) {
319 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock
321 MethodHandlesAdapterBlob* blob = NULL;
322 unsigned int size = sizeof(MethodHandlesAdapterBlob);
323 // align the size to CodeEntryAlignment
324 size = align_code_offset(size);
325 size += round_to(buffer_size, oopSize);
326 {
327 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
328 // The parameter 'true' indicates a critical memory allocation.
329 // This means that CodeCacheMinimumFreeSpace is used, if necessary
330 const bool is_critical = true;
331 blob = new (size, is_critical) MethodHandlesAdapterBlob(size);
332 }
333 // Track memory usage statistic after releasing CodeCache_lock
334 MemoryService::track_code_cache_memory_usage();
336 return blob;
337 }
340 //----------------------------------------------------------------------------------------------------
341 // Implementation of RuntimeStub
343 RuntimeStub::RuntimeStub(
344 const char* name,
345 CodeBuffer* cb,
346 int size,
347 int frame_complete,
348 int frame_size,
349 OopMapSet* oop_maps,
350 bool caller_must_gc_arguments
351 )
352 : CodeBlob(name, cb, sizeof(RuntimeStub), size, frame_complete, frame_size, oop_maps)
353 {
354 _caller_must_gc_arguments = caller_must_gc_arguments;
355 }
358 RuntimeStub* RuntimeStub::new_runtime_stub(const char* stub_name,
359 CodeBuffer* cb,
360 int frame_complete,
361 int frame_size,
362 OopMapSet* oop_maps,
363 bool caller_must_gc_arguments)
364 {
365 RuntimeStub* stub = NULL;
366 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock
367 {
368 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
369 unsigned int size = allocation_size(cb, sizeof(RuntimeStub));
370 stub = new (size) RuntimeStub(stub_name, cb, size, frame_complete, frame_size, oop_maps, caller_must_gc_arguments);
371 }
373 trace_new_stub(stub, "RuntimeStub - ", stub_name);
375 return stub;
376 }
379 void* RuntimeStub::operator new(size_t s, unsigned size) throw() {
380 void* p = CodeCache::allocate(size, true);
381 if (!p) fatal("Initial size of CodeCache is too small");
382 return p;
383 }
385 // operator new shared by all singletons:
386 void* SingletonBlob::operator new(size_t s, unsigned size) throw() {
387 void* p = CodeCache::allocate(size, true);
388 if (!p) fatal("Initial size of CodeCache is too small");
389 return p;
390 }
393 //----------------------------------------------------------------------------------------------------
394 // Implementation of DeoptimizationBlob
396 DeoptimizationBlob::DeoptimizationBlob(
397 CodeBuffer* cb,
398 int size,
399 OopMapSet* oop_maps,
400 int unpack_offset,
401 int unpack_with_exception_offset,
402 int unpack_with_reexecution_offset,
403 int frame_size
404 )
405 : SingletonBlob("DeoptimizationBlob", cb, sizeof(DeoptimizationBlob), size, frame_size, oop_maps)
406 {
407 _unpack_offset = unpack_offset;
408 _unpack_with_exception = unpack_with_exception_offset;
409 _unpack_with_reexecution = unpack_with_reexecution_offset;
410 #ifdef COMPILER1
411 _unpack_with_exception_in_tls = -1;
412 #endif
413 }
416 DeoptimizationBlob* DeoptimizationBlob::create(
417 CodeBuffer* cb,
418 OopMapSet* oop_maps,
419 int unpack_offset,
420 int unpack_with_exception_offset,
421 int unpack_with_reexecution_offset,
422 int frame_size)
423 {
424 DeoptimizationBlob* blob = NULL;
425 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock
426 {
427 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
428 unsigned int size = allocation_size(cb, sizeof(DeoptimizationBlob));
429 blob = new (size) DeoptimizationBlob(cb,
430 size,
431 oop_maps,
432 unpack_offset,
433 unpack_with_exception_offset,
434 unpack_with_reexecution_offset,
435 frame_size);
436 }
438 trace_new_stub(blob, "DeoptimizationBlob");
440 return blob;
441 }
444 //----------------------------------------------------------------------------------------------------
445 // Implementation of UncommonTrapBlob
447 #ifdef COMPILER2
448 UncommonTrapBlob::UncommonTrapBlob(
449 CodeBuffer* cb,
450 int size,
451 OopMapSet* oop_maps,
452 int frame_size
453 )
454 : SingletonBlob("UncommonTrapBlob", cb, sizeof(UncommonTrapBlob), size, frame_size, oop_maps)
455 {}
458 UncommonTrapBlob* UncommonTrapBlob::create(
459 CodeBuffer* cb,
460 OopMapSet* oop_maps,
461 int frame_size)
462 {
463 UncommonTrapBlob* blob = NULL;
464 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock
465 {
466 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
467 unsigned int size = allocation_size(cb, sizeof(UncommonTrapBlob));
468 blob = new (size) UncommonTrapBlob(cb, size, oop_maps, frame_size);
469 }
471 trace_new_stub(blob, "UncommonTrapBlob");
473 return blob;
474 }
477 #endif // COMPILER2
480 //----------------------------------------------------------------------------------------------------
481 // Implementation of ExceptionBlob
483 #ifdef COMPILER2
484 ExceptionBlob::ExceptionBlob(
485 CodeBuffer* cb,
486 int size,
487 OopMapSet* oop_maps,
488 int frame_size
489 )
490 : SingletonBlob("ExceptionBlob", cb, sizeof(ExceptionBlob), size, frame_size, oop_maps)
491 {}
494 ExceptionBlob* ExceptionBlob::create(
495 CodeBuffer* cb,
496 OopMapSet* oop_maps,
497 int frame_size)
498 {
499 ExceptionBlob* blob = NULL;
500 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock
501 {
502 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
503 unsigned int size = allocation_size(cb, sizeof(ExceptionBlob));
504 blob = new (size) ExceptionBlob(cb, size, oop_maps, frame_size);
505 }
507 trace_new_stub(blob, "ExceptionBlob");
509 return blob;
510 }
513 #endif // COMPILER2
516 //----------------------------------------------------------------------------------------------------
517 // Implementation of SafepointBlob
519 SafepointBlob::SafepointBlob(
520 CodeBuffer* cb,
521 int size,
522 OopMapSet* oop_maps,
523 int frame_size
524 )
525 : SingletonBlob("SafepointBlob", cb, sizeof(SafepointBlob), size, frame_size, oop_maps)
526 {}
529 SafepointBlob* SafepointBlob::create(
530 CodeBuffer* cb,
531 OopMapSet* oop_maps,
532 int frame_size)
533 {
534 SafepointBlob* blob = NULL;
535 ThreadInVMfromUnknown __tiv; // get to VM state in case we block on CodeCache_lock
536 {
537 MutexLockerEx mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
538 unsigned int size = allocation_size(cb, sizeof(SafepointBlob));
539 blob = new (size) SafepointBlob(cb, size, oop_maps, frame_size);
540 }
542 trace_new_stub(blob, "SafepointBlob");
544 return blob;
545 }
548 //----------------------------------------------------------------------------------------------------
549 // Verification and printing
551 void CodeBlob::verify() {
552 ShouldNotReachHere();
553 }
555 void CodeBlob::print_on(outputStream* st) const {
556 st->print_cr("[CodeBlob (" INTPTR_FORMAT ")]", p2i(this));
557 st->print_cr("Framesize: %d", _frame_size);
558 }
560 void CodeBlob::print_value_on(outputStream* st) const {
561 st->print_cr("[CodeBlob]");
562 }
564 void BufferBlob::verify() {
565 // unimplemented
566 }
568 void BufferBlob::print_on(outputStream* st) const {
569 CodeBlob::print_on(st);
570 print_value_on(st);
571 }
573 void BufferBlob::print_value_on(outputStream* st) const {
574 st->print_cr("BufferBlob (" INTPTR_FORMAT ") used for %s", p2i(this), name());
575 }
577 void RuntimeStub::verify() {
578 // unimplemented
579 }
581 void RuntimeStub::print_on(outputStream* st) const {
582 ttyLocker ttyl;
583 CodeBlob::print_on(st);
584 st->print("Runtime Stub (" INTPTR_FORMAT "): ", p2i(this));
585 st->print_cr("%s", name());
586 Disassembler::decode((CodeBlob*)this, st);
587 }
589 void RuntimeStub::print_value_on(outputStream* st) const {
590 st->print("RuntimeStub (" INTPTR_FORMAT "): ", p2i(this)); st->print("%s", name());
591 }
593 void SingletonBlob::verify() {
594 // unimplemented
595 }
597 void SingletonBlob::print_on(outputStream* st) const {
598 ttyLocker ttyl;
599 CodeBlob::print_on(st);
600 st->print_cr("%s", name());
601 Disassembler::decode((CodeBlob*)this, st);
602 }
604 void SingletonBlob::print_value_on(outputStream* st) const {
605 st->print_cr("%s", name());
606 }
608 void DeoptimizationBlob::print_value_on(outputStream* st) const {
609 st->print_cr("Deoptimization (frame not available)");
610 }