Thu, 24 May 2018 17:06:56 +0800
Merge
1 /*
2 * Copyright (c) 2003, 2016, Oracle and/or its affiliates. All rights reserved.
3 * Copyright 2007, 2008, 2009, 2010, 2011 Red Hat, Inc.
4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
5 *
6 * This code is free software; you can redistribute it and/or modify it
7 * under the terms of the GNU General Public License version 2 only, as
8 * published by the Free Software Foundation.
9 *
10 * This code is distributed in the hope that it will be useful, but WITHOUT
11 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
12 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
13 * version 2 for more details (a copy is included in the LICENSE file that
14 * accompanied this code).
15 *
16 * You should have received a copy of the GNU General Public License version
17 * 2 along with this work; if not, write to the Free Software Foundation,
18 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
19 *
20 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
21 * or visit www.oracle.com if you need additional information or have any
22 * questions.
23 *
24 */
26 #include "precompiled.hpp"
27 #include "asm/assembler.hpp"
28 #include "interpreter/bytecodeHistogram.hpp"
29 #include "interpreter/cppInterpreter.hpp"
30 #include "interpreter/interpreter.hpp"
31 #include "interpreter/interpreterGenerator.hpp"
32 #include "interpreter/interpreterRuntime.hpp"
33 #include "oops/arrayOop.hpp"
34 #include "oops/methodData.hpp"
35 #include "oops/method.hpp"
36 #include "oops/oop.inline.hpp"
37 #include "prims/jvmtiExport.hpp"
38 #include "prims/jvmtiThreadState.hpp"
39 #include "runtime/arguments.hpp"
40 #include "runtime/deoptimization.hpp"
41 #include "runtime/frame.inline.hpp"
42 #include "runtime/interfaceSupport.hpp"
43 #include "runtime/orderAccess.inline.hpp"
44 #include "runtime/sharedRuntime.hpp"
45 #include "runtime/stubRoutines.hpp"
46 #include "runtime/synchronizer.hpp"
47 #include "runtime/timer.hpp"
48 #include "runtime/vframeArray.hpp"
49 #include "stack_zero.inline.hpp"
50 #include "utilities/debug.hpp"
51 #include "utilities/macros.hpp"
52 #ifdef SHARK
53 #include "shark/shark_globals.hpp"
54 #endif
56 #ifdef CC_INTERP
58 #define fixup_after_potential_safepoint() \
59 method = istate->method()
61 #define CALL_VM_NOCHECK_NOFIX(func) \
62 thread->set_last_Java_frame(); \
63 func; \
64 thread->reset_last_Java_frame();
66 #define CALL_VM_NOCHECK(func) \
67 CALL_VM_NOCHECK_NOFIX(func) \
68 fixup_after_potential_safepoint()
70 int CppInterpreter::normal_entry(Method* method, intptr_t UNUSED, TRAPS) {
71 JavaThread *thread = (JavaThread *) THREAD;
73 // Allocate and initialize our frame.
74 InterpreterFrame *frame = InterpreterFrame::build(method, CHECK_0);
75 thread->push_zero_frame(frame);
77 // Execute those bytecodes!
78 main_loop(0, THREAD);
80 // No deoptimized frames on the stack
81 return 0;
82 }
84 intptr_t narrow(BasicType type, intptr_t result) {
85 // mask integer result to narrower return type.
86 switch (type) {
87 case T_BOOLEAN:
88 return result&1;
89 case T_BYTE:
90 return (intptr_t)(jbyte)result;
91 case T_CHAR:
92 return (intptr_t)(uintptr_t)(jchar)result;
93 case T_SHORT:
94 return (intptr_t)(jshort)result;
95 case T_OBJECT: // nothing to do fall through
96 case T_ARRAY:
97 case T_LONG:
98 case T_INT:
99 case T_FLOAT:
100 case T_DOUBLE:
101 case T_VOID:
102 return result;
103 default : ShouldNotReachHere();
104 }
105 }
108 void CppInterpreter::main_loop(int recurse, TRAPS) {
109 JavaThread *thread = (JavaThread *) THREAD;
110 ZeroStack *stack = thread->zero_stack();
112 // If we are entering from a deopt we may need to call
113 // ourself a few times in order to get to our frame.
114 if (recurse)
115 main_loop(recurse - 1, THREAD);
117 InterpreterFrame *frame = thread->top_zero_frame()->as_interpreter_frame();
118 interpreterState istate = frame->interpreter_state();
119 Method* method = istate->method();
121 intptr_t *result = NULL;
122 int result_slots = 0;
124 while (true) {
125 // We can set up the frame anchor with everything we want at
126 // this point as we are thread_in_Java and no safepoints can
127 // occur until we go to vm mode. We do have to clear flags
128 // on return from vm but that is it.
129 thread->set_last_Java_frame();
131 // Call the interpreter
132 if (JvmtiExport::can_post_interpreter_events())
133 BytecodeInterpreter::runWithChecks(istate);
134 else
135 BytecodeInterpreter::run(istate);
136 fixup_after_potential_safepoint();
138 // Clear the frame anchor
139 thread->reset_last_Java_frame();
141 // Examine the message from the interpreter to decide what to do
142 if (istate->msg() == BytecodeInterpreter::call_method) {
143 Method* callee = istate->callee();
145 // Trim back the stack to put the parameters at the top
146 stack->set_sp(istate->stack() + 1);
148 // Make the call
149 Interpreter::invoke_method(callee, istate->callee_entry_point(), THREAD);
150 fixup_after_potential_safepoint();
152 // Convert the result
153 istate->set_stack(stack->sp() - 1);
155 // Restore the stack
156 stack->set_sp(istate->stack_limit() + 1);
158 // Resume the interpreter
159 istate->set_msg(BytecodeInterpreter::method_resume);
160 }
161 else if (istate->msg() == BytecodeInterpreter::more_monitors) {
162 int monitor_words = frame::interpreter_frame_monitor_size();
164 // Allocate the space
165 stack->overflow_check(monitor_words, THREAD);
166 if (HAS_PENDING_EXCEPTION)
167 break;
168 stack->alloc(monitor_words * wordSize);
170 // Move the expression stack contents
171 for (intptr_t *p = istate->stack() + 1; p < istate->stack_base(); p++)
172 *(p - monitor_words) = *p;
174 // Move the expression stack pointers
175 istate->set_stack_limit(istate->stack_limit() - monitor_words);
176 istate->set_stack(istate->stack() - monitor_words);
177 istate->set_stack_base(istate->stack_base() - monitor_words);
179 // Zero the new monitor so the interpreter can find it.
180 ((BasicObjectLock *) istate->stack_base())->set_obj(NULL);
182 // Resume the interpreter
183 istate->set_msg(BytecodeInterpreter::got_monitors);
184 }
185 else if (istate->msg() == BytecodeInterpreter::return_from_method) {
186 // Copy the result into the caller's frame
187 result_slots = type2size[method->result_type()];
188 assert(result_slots >= 0 && result_slots <= 2, "what?");
189 result = istate->stack() + result_slots;
190 break;
191 }
192 else if (istate->msg() == BytecodeInterpreter::throwing_exception) {
193 assert(HAS_PENDING_EXCEPTION, "should do");
194 break;
195 }
196 else if (istate->msg() == BytecodeInterpreter::do_osr) {
197 // Unwind the current frame
198 thread->pop_zero_frame();
200 // Remove any extension of the previous frame
201 int extra_locals = method->max_locals() - method->size_of_parameters();
202 stack->set_sp(stack->sp() + extra_locals);
204 // Jump into the OSR method
205 Interpreter::invoke_osr(
206 method, istate->osr_entry(), istate->osr_buf(), THREAD);
207 return;
208 }
209 else {
210 ShouldNotReachHere();
211 }
212 }
214 // Unwind the current frame
215 thread->pop_zero_frame();
217 // Pop our local variables
218 stack->set_sp(stack->sp() + method->max_locals());
220 // Push our result
221 for (int i = 0; i < result_slots; i++) {
222 // Adjust result to smaller
223 union {
224 intptr_t res;
225 jint res_jint;
226 };
227 res = result[-i];
228 if (result_slots == 1) {
229 BasicType t = method->result_type();
230 if (is_subword_type(t)) {
231 res_jint = (jint)narrow(t, res_jint);
232 }
233 }
234 stack->push(res);
235 }
236 }
238 int CppInterpreter::native_entry(Method* method, intptr_t UNUSED, TRAPS) {
239 // Make sure method is native and not abstract
240 assert(method->is_native() && !method->is_abstract(), "should be");
242 JavaThread *thread = (JavaThread *) THREAD;
243 ZeroStack *stack = thread->zero_stack();
245 // Allocate and initialize our frame
246 InterpreterFrame *frame = InterpreterFrame::build(method, CHECK_0);
247 thread->push_zero_frame(frame);
248 interpreterState istate = frame->interpreter_state();
249 intptr_t *locals = istate->locals();
251 // Update the invocation counter
252 if ((UseCompiler || CountCompiledCalls) && !method->is_synchronized()) {
253 MethodCounters* mcs = method->method_counters();
254 if (mcs == NULL) {
255 CALL_VM_NOCHECK(mcs = InterpreterRuntime::build_method_counters(thread, method));
256 if (HAS_PENDING_EXCEPTION)
257 goto unwind_and_return;
258 }
259 InvocationCounter *counter = mcs->invocation_counter();
260 counter->increment();
261 if (counter->reached_InvocationLimit(mcs->backedge_counter())) {
262 CALL_VM_NOCHECK(
263 InterpreterRuntime::frequency_counter_overflow(thread, NULL));
264 if (HAS_PENDING_EXCEPTION)
265 goto unwind_and_return;
266 }
267 }
269 // Lock if necessary
270 BasicObjectLock *monitor;
271 monitor = NULL;
272 if (method->is_synchronized()) {
273 monitor = (BasicObjectLock*) istate->stack_base();
274 oop lockee = monitor->obj();
275 markOop disp = lockee->mark()->set_unlocked();
277 monitor->lock()->set_displaced_header(disp);
278 if (Atomic::cmpxchg_ptr(monitor, lockee->mark_addr(), disp) != disp) {
279 if (thread->is_lock_owned((address) disp->clear_lock_bits())) {
280 monitor->lock()->set_displaced_header(NULL);
281 }
282 else {
283 CALL_VM_NOCHECK(InterpreterRuntime::monitorenter(thread, monitor));
284 if (HAS_PENDING_EXCEPTION)
285 goto unwind_and_return;
286 }
287 }
288 }
290 // Get the signature handler
291 InterpreterRuntime::SignatureHandler *handler; {
292 address handlerAddr = method->signature_handler();
293 if (handlerAddr == NULL) {
294 CALL_VM_NOCHECK(InterpreterRuntime::prepare_native_call(thread, method));
295 if (HAS_PENDING_EXCEPTION)
296 goto unlock_unwind_and_return;
298 handlerAddr = method->signature_handler();
299 assert(handlerAddr != NULL, "eh?");
300 }
301 if (handlerAddr == (address) InterpreterRuntime::slow_signature_handler) {
302 CALL_VM_NOCHECK(handlerAddr =
303 InterpreterRuntime::slow_signature_handler(thread, method, NULL,NULL));
304 if (HAS_PENDING_EXCEPTION)
305 goto unlock_unwind_and_return;
306 }
307 handler = \
308 InterpreterRuntime::SignatureHandler::from_handlerAddr(handlerAddr);
309 }
311 // Get the native function entry point
312 address function;
313 function = method->native_function();
314 assert(function != NULL, "should be set if signature handler is");
316 // Build the argument list
317 stack->overflow_check(handler->argument_count() * 2, THREAD);
318 if (HAS_PENDING_EXCEPTION)
319 goto unlock_unwind_and_return;
321 void **arguments;
322 void *mirror; {
323 arguments =
324 (void **) stack->alloc(handler->argument_count() * sizeof(void **));
325 void **dst = arguments;
327 void *env = thread->jni_environment();
328 *(dst++) = &env;
330 if (method->is_static()) {
331 istate->set_oop_temp(
332 method->constants()->pool_holder()->java_mirror());
333 mirror = istate->oop_temp_addr();
334 *(dst++) = &mirror;
335 }
337 intptr_t *src = locals;
338 for (int i = dst - arguments; i < handler->argument_count(); i++) {
339 ffi_type *type = handler->argument_type(i);
340 if (type == &ffi_type_pointer) {
341 if (*src) {
342 stack->push((intptr_t) src);
343 *(dst++) = stack->sp();
344 }
345 else {
346 *(dst++) = src;
347 }
348 src--;
349 }
350 else if (type->size == 4) {
351 *(dst++) = src--;
352 }
353 else if (type->size == 8) {
354 src--;
355 *(dst++) = src--;
356 }
357 else {
358 ShouldNotReachHere();
359 }
360 }
361 }
363 // Set up the Java frame anchor
364 thread->set_last_Java_frame();
366 // Change the thread state to _thread_in_native
367 ThreadStateTransition::transition_from_java(thread, _thread_in_native);
369 // Make the call
370 intptr_t result[4 - LogBytesPerWord];
371 ffi_call(handler->cif(), (void (*)()) function, result, arguments);
373 // Change the thread state back to _thread_in_Java.
374 // ThreadStateTransition::transition_from_native() cannot be used
375 // here because it does not check for asynchronous exceptions.
376 // We have to manage the transition ourself.
377 thread->set_thread_state(_thread_in_native_trans);
379 // Make sure new state is visible in the GC thread
380 if (os::is_MP()) {
381 if (UseMembar) {
382 OrderAccess::fence();
383 }
384 else {
385 InterfaceSupport::serialize_memory(thread);
386 }
387 }
389 // Handle safepoint operations, pending suspend requests,
390 // and pending asynchronous exceptions.
391 if (SafepointSynchronize::do_call_back() ||
392 thread->has_special_condition_for_native_trans()) {
393 JavaThread::check_special_condition_for_native_trans(thread);
394 CHECK_UNHANDLED_OOPS_ONLY(thread->clear_unhandled_oops());
395 }
397 // Finally we can change the thread state to _thread_in_Java.
398 thread->set_thread_state(_thread_in_Java);
399 fixup_after_potential_safepoint();
401 // Clear the frame anchor
402 thread->reset_last_Java_frame();
404 // If the result was an oop then unbox it and store it in
405 // oop_temp where the garbage collector can see it before
406 // we release the handle it might be protected by.
407 if (handler->result_type() == &ffi_type_pointer) {
408 if (result[0])
409 istate->set_oop_temp(*(oop *) result[0]);
410 else
411 istate->set_oop_temp(NULL);
412 }
414 // Reset handle block
415 thread->active_handles()->clear();
417 unlock_unwind_and_return:
419 // Unlock if necessary
420 if (monitor) {
421 BasicLock *lock = monitor->lock();
422 markOop header = lock->displaced_header();
423 oop rcvr = monitor->obj();
424 monitor->set_obj(NULL);
426 if (header != NULL) {
427 if (Atomic::cmpxchg_ptr(header, rcvr->mark_addr(), lock) != lock) {
428 monitor->set_obj(rcvr); {
429 HandleMark hm(thread);
430 CALL_VM_NOCHECK(InterpreterRuntime::monitorexit(thread, monitor));
431 }
432 }
433 }
434 }
436 unwind_and_return:
438 // Unwind the current activation
439 thread->pop_zero_frame();
441 // Pop our parameters
442 stack->set_sp(stack->sp() + method->size_of_parameters());
444 // Push our result
445 if (!HAS_PENDING_EXCEPTION) {
446 BasicType type = method->result_type();
447 stack->set_sp(stack->sp() - type2size[type]);
449 switch (type) {
450 case T_VOID:
451 break;
453 case T_BOOLEAN:
454 #ifndef VM_LITTLE_ENDIAN
455 result[0] <<= (BitsPerWord - BitsPerByte);
456 #endif
457 SET_LOCALS_INT(*(jboolean *) result != 0, 0);
458 break;
460 case T_CHAR:
461 #ifndef VM_LITTLE_ENDIAN
462 result[0] <<= (BitsPerWord - BitsPerShort);
463 #endif
464 SET_LOCALS_INT(*(jchar *) result, 0);
465 break;
467 case T_BYTE:
468 #ifndef VM_LITTLE_ENDIAN
469 result[0] <<= (BitsPerWord - BitsPerByte);
470 #endif
471 SET_LOCALS_INT(*(jbyte *) result, 0);
472 break;
474 case T_SHORT:
475 #ifndef VM_LITTLE_ENDIAN
476 result[0] <<= (BitsPerWord - BitsPerShort);
477 #endif
478 SET_LOCALS_INT(*(jshort *) result, 0);
479 break;
481 case T_INT:
482 #ifndef VM_LITTLE_ENDIAN
483 result[0] <<= (BitsPerWord - BitsPerInt);
484 #endif
485 SET_LOCALS_INT(*(jint *) result, 0);
486 break;
488 case T_LONG:
489 SET_LOCALS_LONG(*(jlong *) result, 0);
490 break;
492 case T_FLOAT:
493 SET_LOCALS_FLOAT(*(jfloat *) result, 0);
494 break;
496 case T_DOUBLE:
497 SET_LOCALS_DOUBLE(*(jdouble *) result, 0);
498 break;
500 case T_OBJECT:
501 case T_ARRAY:
502 SET_LOCALS_OBJECT(istate->oop_temp(), 0);
503 break;
505 default:
506 ShouldNotReachHere();
507 }
508 }
510 // No deoptimized frames on the stack
511 return 0;
512 }
514 int CppInterpreter::accessor_entry(Method* method, intptr_t UNUSED, TRAPS) {
515 JavaThread *thread = (JavaThread *) THREAD;
516 ZeroStack *stack = thread->zero_stack();
517 intptr_t *locals = stack->sp();
519 // Drop into the slow path if we need a safepoint check
520 if (SafepointSynchronize::do_call_back()) {
521 return normal_entry(method, 0, THREAD);
522 }
524 // Load the object pointer and drop into the slow path
525 // if we have a NullPointerException
526 oop object = LOCALS_OBJECT(0);
527 if (object == NULL) {
528 return normal_entry(method, 0, THREAD);
529 }
531 // Read the field index from the bytecode, which looks like this:
532 // 0: aload_0
533 // 1: getfield
534 // 2: index
535 // 3: index
536 // 4: ireturn/areturn
537 // NB this is not raw bytecode: index is in machine order
538 u1 *code = method->code_base();
539 assert(code[0] == Bytecodes::_aload_0 &&
540 code[1] == Bytecodes::_getfield &&
541 (code[4] == Bytecodes::_ireturn ||
542 code[4] == Bytecodes::_areturn), "should do");
543 u2 index = Bytes::get_native_u2(&code[2]);
545 // Get the entry from the constant pool cache, and drop into
546 // the slow path if it has not been resolved
547 ConstantPoolCache* cache = method->constants()->cache();
548 ConstantPoolCacheEntry* entry = cache->entry_at(index);
549 if (!entry->is_resolved(Bytecodes::_getfield)) {
550 return normal_entry(method, 0, THREAD);
551 }
553 // Get the result and push it onto the stack
554 switch (entry->flag_state()) {
555 case ltos:
556 case dtos:
557 stack->overflow_check(1, CHECK_0);
558 stack->alloc(wordSize);
559 break;
560 }
561 if (entry->is_volatile()) {
562 switch (entry->flag_state()) {
563 case ctos:
564 SET_LOCALS_INT(object->char_field_acquire(entry->f2_as_index()), 0);
565 break;
567 case btos:
568 case ztos:
569 SET_LOCALS_INT(object->byte_field_acquire(entry->f2_as_index()), 0);
570 break;
572 case stos:
573 SET_LOCALS_INT(object->short_field_acquire(entry->f2_as_index()), 0);
574 break;
576 case itos:
577 SET_LOCALS_INT(object->int_field_acquire(entry->f2_as_index()), 0);
578 break;
580 case ltos:
581 SET_LOCALS_LONG(object->long_field_acquire(entry->f2_as_index()), 0);
582 break;
584 case ftos:
585 SET_LOCALS_FLOAT(object->float_field_acquire(entry->f2_as_index()), 0);
586 break;
588 case dtos:
589 SET_LOCALS_DOUBLE(object->double_field_acquire(entry->f2_as_index()), 0);
590 break;
592 case atos:
593 SET_LOCALS_OBJECT(object->obj_field_acquire(entry->f2_as_index()), 0);
594 break;
596 default:
597 ShouldNotReachHere();
598 }
599 }
600 else {
601 switch (entry->flag_state()) {
602 case ctos:
603 SET_LOCALS_INT(object->char_field(entry->f2_as_index()), 0);
604 break;
606 case btos:
607 case ztos:
608 SET_LOCALS_INT(object->byte_field(entry->f2_as_index()), 0);
609 break;
611 case stos:
612 SET_LOCALS_INT(object->short_field(entry->f2_as_index()), 0);
613 break;
615 case itos:
616 SET_LOCALS_INT(object->int_field(entry->f2_as_index()), 0);
617 break;
619 case ltos:
620 SET_LOCALS_LONG(object->long_field(entry->f2_as_index()), 0);
621 break;
623 case ftos:
624 SET_LOCALS_FLOAT(object->float_field(entry->f2_as_index()), 0);
625 break;
627 case dtos:
628 SET_LOCALS_DOUBLE(object->double_field(entry->f2_as_index()), 0);
629 break;
631 case atos:
632 SET_LOCALS_OBJECT(object->obj_field(entry->f2_as_index()), 0);
633 break;
635 default:
636 ShouldNotReachHere();
637 }
638 }
640 // No deoptimized frames on the stack
641 return 0;
642 }
644 int CppInterpreter::empty_entry(Method* method, intptr_t UNUSED, TRAPS) {
645 JavaThread *thread = (JavaThread *) THREAD;
646 ZeroStack *stack = thread->zero_stack();
648 // Drop into the slow path if we need a safepoint check
649 if (SafepointSynchronize::do_call_back()) {
650 return normal_entry(method, 0, THREAD);
651 }
653 // Pop our parameters
654 stack->set_sp(stack->sp() + method->size_of_parameters());
656 // No deoptimized frames on the stack
657 return 0;
658 }
660 // The new slots will be inserted before slot insert_before.
661 // Slots < insert_before will have the same slot number after the insert.
662 // Slots >= insert_before will become old_slot + num_slots.
663 void CppInterpreter::insert_vmslots(int insert_before, int num_slots, TRAPS) {
664 JavaThread *thread = (JavaThread *) THREAD;
665 ZeroStack *stack = thread->zero_stack();
667 // Allocate the space
668 stack->overflow_check(num_slots, CHECK);
669 stack->alloc(num_slots * wordSize);
670 intptr_t *vmslots = stack->sp();
672 // Shuffle everything up
673 for (int i = 0; i < insert_before; i++)
674 SET_VMSLOTS_SLOT(VMSLOTS_SLOT(i + num_slots), i);
675 }
677 void CppInterpreter::remove_vmslots(int first_slot, int num_slots, TRAPS) {
678 JavaThread *thread = (JavaThread *) THREAD;
679 ZeroStack *stack = thread->zero_stack();
680 intptr_t *vmslots = stack->sp();
682 // Move everything down
683 for (int i = first_slot - 1; i >= 0; i--)
684 SET_VMSLOTS_SLOT(VMSLOTS_SLOT(i), i + num_slots);
686 // Deallocate the space
687 stack->set_sp(stack->sp() + num_slots);
688 }
690 BasicType CppInterpreter::result_type_of_handle(oop method_handle) {
691 oop method_type = java_lang_invoke_MethodHandle::type(method_handle);
692 oop return_type = java_lang_invoke_MethodType::rtype(method_type);
693 return java_lang_Class::as_BasicType(return_type, (Klass* *) NULL);
694 }
696 intptr_t* CppInterpreter::calculate_unwind_sp(ZeroStack* stack,
697 oop method_handle) {
698 oop method_type = java_lang_invoke_MethodHandle::type(method_handle);
699 int argument_slots = java_lang_invoke_MethodType::ptype_slot_count(method_type);
701 return stack->sp() + argument_slots;
702 }
704 IRT_ENTRY(void, CppInterpreter::throw_exception(JavaThread* thread,
705 Symbol* name,
706 char* message))
707 THROW_MSG(name, message);
708 IRT_END
710 InterpreterFrame *InterpreterFrame::build(Method* const method, TRAPS) {
711 JavaThread *thread = (JavaThread *) THREAD;
712 ZeroStack *stack = thread->zero_stack();
714 // Calculate the size of the frame we'll build, including
715 // any adjustments to the caller's frame that we'll make.
716 int extra_locals = 0;
717 int monitor_words = 0;
718 int stack_words = 0;
720 if (!method->is_native()) {
721 extra_locals = method->max_locals() - method->size_of_parameters();
722 stack_words = method->max_stack();
723 }
724 if (method->is_synchronized()) {
725 monitor_words = frame::interpreter_frame_monitor_size();
726 }
727 stack->overflow_check(
728 extra_locals + header_words + monitor_words + stack_words, CHECK_NULL);
730 // Adjust the caller's stack frame to accomodate any additional
731 // local variables we have contiguously with our parameters.
732 for (int i = 0; i < extra_locals; i++)
733 stack->push(0);
735 intptr_t *locals;
736 if (method->is_native())
737 locals = stack->sp() + (method->size_of_parameters() - 1);
738 else
739 locals = stack->sp() + (method->max_locals() - 1);
741 stack->push(0); // next_frame, filled in later
742 intptr_t *fp = stack->sp();
743 assert(fp - stack->sp() == next_frame_off, "should be");
745 stack->push(INTERPRETER_FRAME);
746 assert(fp - stack->sp() == frame_type_off, "should be");
748 interpreterState istate =
749 (interpreterState) stack->alloc(sizeof(BytecodeInterpreter));
750 assert(fp - stack->sp() == istate_off, "should be");
752 istate->set_locals(locals);
753 istate->set_method(method);
754 istate->set_self_link(istate);
755 istate->set_prev_link(NULL);
756 istate->set_thread(thread);
757 istate->set_bcp(method->is_native() ? NULL : method->code_base());
758 istate->set_constants(method->constants()->cache());
759 istate->set_msg(BytecodeInterpreter::method_entry);
760 istate->set_oop_temp(NULL);
761 istate->set_mdx(NULL);
762 istate->set_callee(NULL);
764 istate->set_monitor_base((BasicObjectLock *) stack->sp());
765 if (method->is_synchronized()) {
766 BasicObjectLock *monitor =
767 (BasicObjectLock *) stack->alloc(monitor_words * wordSize);
768 oop object;
769 if (method->is_static())
770 object = method->constants()->pool_holder()->java_mirror();
771 else
772 object = (oop) (void*)locals[0];
773 monitor->set_obj(object);
774 }
776 istate->set_stack_base(stack->sp());
777 istate->set_stack(stack->sp() - 1);
778 if (stack_words)
779 stack->alloc(stack_words * wordSize);
780 istate->set_stack_limit(stack->sp() - 1);
782 return (InterpreterFrame *) fp;
783 }
785 int AbstractInterpreter::BasicType_as_index(BasicType type) {
786 int i = 0;
787 switch (type) {
788 case T_BOOLEAN: i = 0; break;
789 case T_CHAR : i = 1; break;
790 case T_BYTE : i = 2; break;
791 case T_SHORT : i = 3; break;
792 case T_INT : i = 4; break;
793 case T_LONG : i = 5; break;
794 case T_VOID : i = 6; break;
795 case T_FLOAT : i = 7; break;
796 case T_DOUBLE : i = 8; break;
797 case T_OBJECT : i = 9; break;
798 case T_ARRAY : i = 9; break;
799 default : ShouldNotReachHere();
800 }
801 assert(0 <= i && i < AbstractInterpreter::number_of_result_handlers,
802 "index out of bounds");
803 return i;
804 }
806 address InterpreterGenerator::generate_empty_entry() {
807 if (!UseFastEmptyMethods)
808 return NULL;
810 return generate_entry((address) CppInterpreter::empty_entry);
811 }
813 address InterpreterGenerator::generate_accessor_entry() {
814 if (!UseFastAccessorMethods)
815 return NULL;
817 return generate_entry((address) CppInterpreter::accessor_entry);
818 }
820 address InterpreterGenerator::generate_Reference_get_entry(void) {
821 #if INCLUDE_ALL_GCS
822 if (UseG1GC) {
823 // We need to generate have a routine that generates code to:
824 // * load the value in the referent field
825 // * passes that value to the pre-barrier.
826 //
827 // In the case of G1 this will record the value of the
828 // referent in an SATB buffer if marking is active.
829 // This will cause concurrent marking to mark the referent
830 // field as live.
831 Unimplemented();
832 }
833 #endif // INCLUDE_ALL_GCS
835 // If G1 is not enabled then attempt to go through the accessor entry point
836 // Reference.get is an accessor
837 return generate_accessor_entry();
838 }
840 address InterpreterGenerator::generate_native_entry(bool synchronized) {
841 assert(synchronized == false, "should be");
843 return generate_entry((address) CppInterpreter::native_entry);
844 }
846 address InterpreterGenerator::generate_normal_entry(bool synchronized) {
847 assert(synchronized == false, "should be");
849 return generate_entry((address) CppInterpreter::normal_entry);
850 }
852 address AbstractInterpreterGenerator::generate_method_entry(
853 AbstractInterpreter::MethodKind kind) {
854 address entry_point = NULL;
856 switch (kind) {
857 case Interpreter::zerolocals:
858 case Interpreter::zerolocals_synchronized:
859 break;
861 case Interpreter::native:
862 entry_point = ((InterpreterGenerator*) this)->generate_native_entry(false);
863 break;
865 case Interpreter::native_synchronized:
866 entry_point = ((InterpreterGenerator*) this)->generate_native_entry(false);
867 break;
869 case Interpreter::empty:
870 entry_point = ((InterpreterGenerator*) this)->generate_empty_entry();
871 break;
873 case Interpreter::accessor:
874 entry_point = ((InterpreterGenerator*) this)->generate_accessor_entry();
875 break;
877 case Interpreter::abstract:
878 entry_point = ((InterpreterGenerator*) this)->generate_abstract_entry();
879 break;
881 case Interpreter::java_lang_math_sin:
882 case Interpreter::java_lang_math_cos:
883 case Interpreter::java_lang_math_tan:
884 case Interpreter::java_lang_math_abs:
885 case Interpreter::java_lang_math_log:
886 case Interpreter::java_lang_math_log10:
887 case Interpreter::java_lang_math_sqrt:
888 case Interpreter::java_lang_math_pow:
889 case Interpreter::java_lang_math_exp:
890 entry_point = ((InterpreterGenerator*) this)->generate_math_entry(kind);
891 break;
893 case Interpreter::java_lang_ref_reference_get:
894 entry_point = ((InterpreterGenerator*)this)->generate_Reference_get_entry();
895 break;
897 default:
898 ShouldNotReachHere();
899 }
901 if (entry_point == NULL)
902 entry_point = ((InterpreterGenerator*) this)->generate_normal_entry(false);
904 return entry_point;
905 }
907 InterpreterGenerator::InterpreterGenerator(StubQueue* code)
908 : CppInterpreterGenerator(code) {
909 generate_all();
910 }
912 // Deoptimization helpers
914 InterpreterFrame *InterpreterFrame::build(int size, TRAPS) {
915 ZeroStack *stack = ((JavaThread *) THREAD)->zero_stack();
917 int size_in_words = size >> LogBytesPerWord;
918 assert(size_in_words * wordSize == size, "unaligned");
919 assert(size_in_words >= header_words, "too small");
920 stack->overflow_check(size_in_words, CHECK_NULL);
922 stack->push(0); // next_frame, filled in later
923 intptr_t *fp = stack->sp();
924 assert(fp - stack->sp() == next_frame_off, "should be");
926 stack->push(INTERPRETER_FRAME);
927 assert(fp - stack->sp() == frame_type_off, "should be");
929 interpreterState istate =
930 (interpreterState) stack->alloc(sizeof(BytecodeInterpreter));
931 assert(fp - stack->sp() == istate_off, "should be");
932 istate->set_self_link(NULL); // mark invalid
934 stack->alloc((size_in_words - header_words) * wordSize);
936 return (InterpreterFrame *) fp;
937 }
939 int AbstractInterpreter::size_activation(int max_stack,
940 int tempcount,
941 int extra_args,
942 int moncount,
943 int callee_param_count,
944 int callee_locals,
945 bool is_top_frame) {
946 int header_words = InterpreterFrame::header_words;
947 int monitor_words = moncount * frame::interpreter_frame_monitor_size();
948 int stack_words = is_top_frame ? max_stack : tempcount;
949 int callee_extra_locals = callee_locals - callee_param_count;
951 return header_words + monitor_words + stack_words + callee_extra_locals;
952 }
954 void AbstractInterpreter::layout_activation(Method* method,
955 int tempcount,
956 int popframe_extra_args,
957 int moncount,
958 int caller_actual_parameters,
959 int callee_param_count,
960 int callee_locals,
961 frame* caller,
962 frame* interpreter_frame,
963 bool is_top_frame,
964 bool is_bottom_frame) {
965 assert(popframe_extra_args == 0, "what to do?");
966 assert(!is_top_frame || (!callee_locals && !callee_param_count),
967 "top frame should have no caller");
969 // This code must exactly match what InterpreterFrame::build
970 // does (the full InterpreterFrame::build, that is, not the
971 // one that creates empty frames for the deoptimizer).
972 //
973 // interpreter_frame will be filled in. It's size is determined by
974 // a previous call to the size_activation() method,
975 //
976 // Note that tempcount is the current size of the expression
977 // stack. For top most frames we will allocate a full sized
978 // expression stack and not the trimmed version that non-top
979 // frames have.
981 int monitor_words = moncount * frame::interpreter_frame_monitor_size();
982 intptr_t *locals = interpreter_frame->fp() + method->max_locals();
983 interpreterState istate = interpreter_frame->get_interpreterState();
984 intptr_t *monitor_base = (intptr_t*) istate;
985 intptr_t *stack_base = monitor_base - monitor_words;
986 intptr_t *stack = stack_base - tempcount - 1;
988 BytecodeInterpreter::layout_interpreterState(istate,
989 caller,
990 NULL,
991 method,
992 locals,
993 stack,
994 stack_base,
995 monitor_base,
996 NULL,
997 is_top_frame);
998 }
1000 void BytecodeInterpreter::layout_interpreterState(interpreterState istate,
1001 frame* caller,
1002 frame* current,
1003 Method* method,
1004 intptr_t* locals,
1005 intptr_t* stack,
1006 intptr_t* stack_base,
1007 intptr_t* monitor_base,
1008 intptr_t* frame_bottom,
1009 bool is_top_frame) {
1010 istate->set_locals(locals);
1011 istate->set_method(method);
1012 istate->set_self_link(istate);
1013 istate->set_prev_link(NULL);
1014 // thread will be set by a hacky repurposing of frame::patch_pc()
1015 // bcp will be set by vframeArrayElement::unpack_on_stack()
1016 istate->set_constants(method->constants()->cache());
1017 istate->set_msg(BytecodeInterpreter::method_resume);
1018 istate->set_bcp_advance(0);
1019 istate->set_oop_temp(NULL);
1020 istate->set_mdx(NULL);
1021 if (caller->is_interpreted_frame()) {
1022 interpreterState prev = caller->get_interpreterState();
1023 prev->set_callee(method);
1024 if (*prev->bcp() == Bytecodes::_invokeinterface)
1025 prev->set_bcp_advance(5);
1026 else
1027 prev->set_bcp_advance(3);
1028 }
1029 istate->set_callee(NULL);
1030 istate->set_monitor_base((BasicObjectLock *) monitor_base);
1031 istate->set_stack_base(stack_base);
1032 istate->set_stack(stack);
1033 istate->set_stack_limit(stack_base - method->max_stack() - 1);
1034 }
1036 address CppInterpreter::return_entry(TosState state, int length, Bytecodes::Code code) {
1037 ShouldNotCallThis();
1038 return NULL;
1039 }
1041 address CppInterpreter::deopt_entry(TosState state, int length) {
1042 return NULL;
1043 }
1045 // Helper for (runtime) stack overflow checks
1047 int AbstractInterpreter::size_top_interpreter_activation(Method* method) {
1048 return 0;
1049 }
1051 // Helper for figuring out if frames are interpreter frames
1053 bool CppInterpreter::contains(address pc) {
1054 return false; // make frame::print_value_on work
1055 }
1057 // Result handlers and convertors
1059 address CppInterpreterGenerator::generate_result_handler_for(
1060 BasicType type) {
1061 assembler()->advance(1);
1062 return ShouldNotCallThisStub();
1063 }
1065 address CppInterpreterGenerator::generate_tosca_to_stack_converter(
1066 BasicType type) {
1067 assembler()->advance(1);
1068 return ShouldNotCallThisStub();
1069 }
1071 address CppInterpreterGenerator::generate_stack_to_stack_converter(
1072 BasicType type) {
1073 assembler()->advance(1);
1074 return ShouldNotCallThisStub();
1075 }
1077 address CppInterpreterGenerator::generate_stack_to_native_abi_converter(
1078 BasicType type) {
1079 assembler()->advance(1);
1080 return ShouldNotCallThisStub();
1081 }
1083 #endif // CC_INTERP