Tue, 17 Oct 2017 12:58:25 +0800
merge
aoqi@0 | 1 | /* |
aoqi@0 | 2 | * Copyright (c) 2014, Oracle and/or its affiliates. All rights reserved. |
aoqi@0 | 3 | * Copyright 2013, 2014 SAP AG. All rights reserved. |
aoqi@0 | 4 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
aoqi@0 | 5 | * |
aoqi@0 | 6 | * This code is free software; you can redistribute it and/or modify it |
aoqi@0 | 7 | * under the terms of the GNU General Public License version 2 only, as |
aoqi@0 | 8 | * published by the Free Software Foundation. |
aoqi@0 | 9 | * |
aoqi@0 | 10 | * This code is distributed in the hope that it will be useful, but WITHOUT |
aoqi@0 | 11 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
aoqi@0 | 12 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
aoqi@0 | 13 | * version 2 for more details (a copy is included in the LICENSE file that |
aoqi@0 | 14 | * accompanied this code). |
aoqi@0 | 15 | * |
aoqi@0 | 16 | * You should have received a copy of the GNU General Public License version |
aoqi@0 | 17 | * 2 along with this work; if not, write to the Free Software Foundation, |
aoqi@0 | 18 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
aoqi@0 | 19 | * |
aoqi@0 | 20 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
aoqi@0 | 21 | * or visit www.oracle.com if you need additional information or have any |
aoqi@0 | 22 | * questions. |
aoqi@0 | 23 | * |
aoqi@0 | 24 | */ |
aoqi@0 | 25 | |
aoqi@0 | 26 | #include "precompiled.hpp" |
aoqi@0 | 27 | #ifndef CC_INTERP |
aoqi@0 | 28 | #include "asm/macroAssembler.inline.hpp" |
aoqi@0 | 29 | #include "interpreter/bytecodeHistogram.hpp" |
aoqi@0 | 30 | #include "interpreter/interpreter.hpp" |
aoqi@0 | 31 | #include "interpreter/interpreterGenerator.hpp" |
aoqi@0 | 32 | #include "interpreter/interpreterRuntime.hpp" |
aoqi@0 | 33 | #include "interpreter/templateTable.hpp" |
aoqi@0 | 34 | #include "oops/arrayOop.hpp" |
aoqi@0 | 35 | #include "oops/methodData.hpp" |
aoqi@0 | 36 | #include "oops/method.hpp" |
aoqi@0 | 37 | #include "oops/oop.inline.hpp" |
aoqi@0 | 38 | #include "prims/jvmtiExport.hpp" |
aoqi@0 | 39 | #include "prims/jvmtiThreadState.hpp" |
aoqi@0 | 40 | #include "runtime/arguments.hpp" |
aoqi@0 | 41 | #include "runtime/deoptimization.hpp" |
aoqi@0 | 42 | #include "runtime/frame.inline.hpp" |
aoqi@0 | 43 | #include "runtime/sharedRuntime.hpp" |
aoqi@0 | 44 | #include "runtime/stubRoutines.hpp" |
aoqi@0 | 45 | #include "runtime/synchronizer.hpp" |
aoqi@0 | 46 | #include "runtime/timer.hpp" |
aoqi@0 | 47 | #include "runtime/vframeArray.hpp" |
aoqi@0 | 48 | #include "utilities/debug.hpp" |
aoqi@0 | 49 | #include "utilities/macros.hpp" |
aoqi@0 | 50 | |
aoqi@0 | 51 | #undef __ |
aoqi@0 | 52 | #define __ _masm-> |
aoqi@0 | 53 | |
aoqi@0 | 54 | #ifdef PRODUCT |
aoqi@0 | 55 | #define BLOCK_COMMENT(str) /* nothing */ |
aoqi@0 | 56 | #else |
aoqi@0 | 57 | #define BLOCK_COMMENT(str) __ block_comment(str) |
aoqi@0 | 58 | #endif |
aoqi@0 | 59 | |
aoqi@0 | 60 | #define BIND(label) bind(label); BLOCK_COMMENT(#label ":") |
aoqi@0 | 61 | |
aoqi@0 | 62 | //----------------------------------------------------------------------------- |
aoqi@0 | 63 | |
aoqi@0 | 64 | // Actually we should never reach here since we do stack overflow checks before pushing any frame. |
aoqi@0 | 65 | address TemplateInterpreterGenerator::generate_StackOverflowError_handler() { |
aoqi@0 | 66 | address entry = __ pc(); |
aoqi@0 | 67 | __ unimplemented("generate_StackOverflowError_handler"); |
aoqi@0 | 68 | return entry; |
aoqi@0 | 69 | } |
aoqi@0 | 70 | |
aoqi@0 | 71 | address TemplateInterpreterGenerator::generate_ArrayIndexOutOfBounds_handler(const char* name) { |
aoqi@0 | 72 | address entry = __ pc(); |
aoqi@0 | 73 | __ empty_expression_stack(); |
aoqi@0 | 74 | __ load_const_optimized(R4_ARG2, (address) name); |
aoqi@0 | 75 | // Index is in R17_tos. |
aoqi@0 | 76 | __ mr(R5_ARG3, R17_tos); |
aoqi@0 | 77 | __ call_VM(noreg, CAST_FROM_FN_PTR(address, InterpreterRuntime::throw_ArrayIndexOutOfBoundsException)); |
aoqi@0 | 78 | return entry; |
aoqi@0 | 79 | } |
aoqi@0 | 80 | |
aoqi@0 | 81 | #if 0 |
aoqi@0 | 82 | // Call special ClassCastException constructor taking object to cast |
aoqi@0 | 83 | // and target class as arguments. |
aoqi@0 | 84 | address TemplateInterpreterGenerator::generate_ClassCastException_verbose_handler() { |
aoqi@0 | 85 | address entry = __ pc(); |
aoqi@0 | 86 | |
aoqi@0 | 87 | // Expression stack must be empty before entering the VM if an |
aoqi@0 | 88 | // exception happened. |
aoqi@0 | 89 | __ empty_expression_stack(); |
aoqi@0 | 90 | |
aoqi@0 | 91 | // Thread will be loaded to R3_ARG1. |
aoqi@0 | 92 | // Target class oop is in register R5_ARG3 by convention! |
goetz@7424 | 93 | __ call_VM(noreg, CAST_FROM_FN_PTR(address, InterpreterRuntime::throw_ClassCastException_verbose), R17_tos, R5_ARG3); |
aoqi@0 | 94 | // Above call must not return here since exception pending. |
aoqi@0 | 95 | DEBUG_ONLY(__ should_not_reach_here();) |
aoqi@0 | 96 | return entry; |
aoqi@0 | 97 | } |
aoqi@0 | 98 | #endif |
aoqi@0 | 99 | |
aoqi@0 | 100 | address TemplateInterpreterGenerator::generate_ClassCastException_handler() { |
aoqi@0 | 101 | address entry = __ pc(); |
aoqi@0 | 102 | // Expression stack must be empty before entering the VM if an |
aoqi@0 | 103 | // exception happened. |
aoqi@0 | 104 | __ empty_expression_stack(); |
aoqi@0 | 105 | |
aoqi@0 | 106 | // Load exception object. |
aoqi@0 | 107 | // Thread will be loaded to R3_ARG1. |
aoqi@0 | 108 | __ call_VM(noreg, CAST_FROM_FN_PTR(address, InterpreterRuntime::throw_ClassCastException), R17_tos); |
aoqi@0 | 109 | #ifdef ASSERT |
aoqi@0 | 110 | // Above call must not return here since exception pending. |
aoqi@0 | 111 | __ should_not_reach_here(); |
aoqi@0 | 112 | #endif |
aoqi@0 | 113 | return entry; |
aoqi@0 | 114 | } |
aoqi@0 | 115 | |
aoqi@0 | 116 | address TemplateInterpreterGenerator::generate_exception_handler_common(const char* name, const char* message, bool pass_oop) { |
aoqi@0 | 117 | address entry = __ pc(); |
aoqi@0 | 118 | //__ untested("generate_exception_handler_common"); |
aoqi@0 | 119 | Register Rexception = R17_tos; |
aoqi@0 | 120 | |
aoqi@0 | 121 | // Expression stack must be empty before entering the VM if an exception happened. |
aoqi@0 | 122 | __ empty_expression_stack(); |
aoqi@0 | 123 | |
aoqi@0 | 124 | __ load_const_optimized(R4_ARG2, (address) name, R11_scratch1); |
aoqi@0 | 125 | if (pass_oop) { |
aoqi@0 | 126 | __ mr(R5_ARG3, Rexception); |
aoqi@0 | 127 | __ call_VM(Rexception, CAST_FROM_FN_PTR(address, InterpreterRuntime::create_klass_exception), false); |
aoqi@0 | 128 | } else { |
aoqi@0 | 129 | __ load_const_optimized(R5_ARG3, (address) message, R11_scratch1); |
aoqi@0 | 130 | __ call_VM(Rexception, CAST_FROM_FN_PTR(address, InterpreterRuntime::create_exception), false); |
aoqi@0 | 131 | } |
aoqi@0 | 132 | |
aoqi@0 | 133 | // Throw exception. |
aoqi@0 | 134 | __ mr(R3_ARG1, Rexception); |
aoqi@0 | 135 | __ load_const_optimized(R11_scratch1, Interpreter::throw_exception_entry(), R12_scratch2); |
aoqi@0 | 136 | __ mtctr(R11_scratch1); |
aoqi@0 | 137 | __ bctr(); |
aoqi@0 | 138 | |
aoqi@0 | 139 | return entry; |
aoqi@0 | 140 | } |
aoqi@0 | 141 | |
aoqi@0 | 142 | address TemplateInterpreterGenerator::generate_continuation_for(TosState state) { |
aoqi@0 | 143 | address entry = __ pc(); |
aoqi@0 | 144 | __ unimplemented("generate_continuation_for"); |
aoqi@0 | 145 | return entry; |
aoqi@0 | 146 | } |
aoqi@0 | 147 | |
aoqi@0 | 148 | // This entry is returned to when a call returns to the interpreter. |
aoqi@0 | 149 | // When we arrive here, we expect that the callee stack frame is already popped. |
aoqi@0 | 150 | address TemplateInterpreterGenerator::generate_return_entry_for(TosState state, int step, size_t index_size) { |
aoqi@0 | 151 | address entry = __ pc(); |
aoqi@0 | 152 | |
aoqi@0 | 153 | // Move the value out of the return register back to the TOS cache of current frame. |
aoqi@0 | 154 | switch (state) { |
aoqi@0 | 155 | case ltos: |
aoqi@0 | 156 | case btos: |
aoqi@0 | 157 | case ctos: |
aoqi@0 | 158 | case stos: |
aoqi@0 | 159 | case atos: |
aoqi@0 | 160 | case itos: __ mr(R17_tos, R3_RET); break; // RET -> TOS cache |
aoqi@0 | 161 | case ftos: |
aoqi@0 | 162 | case dtos: __ fmr(F15_ftos, F1_RET); break; // TOS cache -> GR_FRET |
aoqi@0 | 163 | case vtos: break; // Nothing to do, this was a void return. |
aoqi@0 | 164 | default : ShouldNotReachHere(); |
aoqi@0 | 165 | } |
aoqi@0 | 166 | |
aoqi@0 | 167 | __ restore_interpreter_state(R11_scratch1); // Sets R11_scratch1 = fp. |
aoqi@0 | 168 | __ ld(R12_scratch2, _ijava_state_neg(top_frame_sp), R11_scratch1); |
aoqi@0 | 169 | __ resize_frame_absolute(R12_scratch2, R11_scratch1, R0); |
aoqi@0 | 170 | |
aoqi@0 | 171 | // Compiled code destroys templateTableBase, reload. |
aoqi@0 | 172 | __ load_const_optimized(R25_templateTableBase, (address)Interpreter::dispatch_table((TosState)0), R12_scratch2); |
aoqi@0 | 173 | |
goetz@7424 | 174 | if (state == atos) { |
goetz@7424 | 175 | __ profile_return_type(R3_RET, R11_scratch1, R12_scratch2); |
goetz@7424 | 176 | } |
goetz@7424 | 177 | |
aoqi@0 | 178 | const Register cache = R11_scratch1; |
aoqi@0 | 179 | const Register size = R12_scratch2; |
aoqi@0 | 180 | __ get_cache_and_index_at_bcp(cache, 1, index_size); |
aoqi@0 | 181 | |
kvn@7132 | 182 | // Get least significant byte of 64 bit value: |
kvn@7132 | 183 | #if defined(VM_LITTLE_ENDIAN) |
kvn@7132 | 184 | __ lbz(size, in_bytes(ConstantPoolCache::base_offset() + ConstantPoolCacheEntry::flags_offset()), cache); |
kvn@7132 | 185 | #else |
aoqi@0 | 186 | __ lbz(size, in_bytes(ConstantPoolCache::base_offset() + ConstantPoolCacheEntry::flags_offset()) + 7, cache); |
kvn@7132 | 187 | #endif |
aoqi@0 | 188 | __ sldi(size, size, Interpreter::logStackElementSize); |
aoqi@0 | 189 | __ add(R15_esp, R15_esp, size); |
aoqi@0 | 190 | __ dispatch_next(state, step); |
aoqi@0 | 191 | return entry; |
aoqi@0 | 192 | } |
aoqi@0 | 193 | |
aoqi@0 | 194 | address TemplateInterpreterGenerator::generate_deopt_entry_for(TosState state, int step) { |
aoqi@0 | 195 | address entry = __ pc(); |
aoqi@0 | 196 | // If state != vtos, we're returning from a native method, which put it's result |
aoqi@0 | 197 | // into the result register. So move the value out of the return register back |
aoqi@0 | 198 | // to the TOS cache of current frame. |
aoqi@0 | 199 | |
aoqi@0 | 200 | switch (state) { |
aoqi@0 | 201 | case ltos: |
aoqi@0 | 202 | case btos: |
aoqi@0 | 203 | case ctos: |
aoqi@0 | 204 | case stos: |
aoqi@0 | 205 | case atos: |
aoqi@0 | 206 | case itos: __ mr(R17_tos, R3_RET); break; // GR_RET -> TOS cache |
aoqi@0 | 207 | case ftos: |
aoqi@0 | 208 | case dtos: __ fmr(F15_ftos, F1_RET); break; // TOS cache -> GR_FRET |
aoqi@0 | 209 | case vtos: break; // Nothing to do, this was a void return. |
aoqi@0 | 210 | default : ShouldNotReachHere(); |
aoqi@0 | 211 | } |
aoqi@0 | 212 | |
aoqi@0 | 213 | // Load LcpoolCache @@@ should be already set! |
aoqi@0 | 214 | __ get_constant_pool_cache(R27_constPoolCache); |
aoqi@0 | 215 | |
aoqi@0 | 216 | // Handle a pending exception, fall through if none. |
aoqi@0 | 217 | __ check_and_forward_exception(R11_scratch1, R12_scratch2); |
aoqi@0 | 218 | |
aoqi@0 | 219 | // Start executing bytecodes. |
aoqi@0 | 220 | __ dispatch_next(state, step); |
aoqi@0 | 221 | |
aoqi@0 | 222 | return entry; |
aoqi@0 | 223 | } |
aoqi@0 | 224 | |
aoqi@0 | 225 | // A result handler converts the native result into java format. |
aoqi@0 | 226 | // Use the shared code between c++ and template interpreter. |
aoqi@0 | 227 | address TemplateInterpreterGenerator::generate_result_handler_for(BasicType type) { |
aoqi@0 | 228 | return AbstractInterpreterGenerator::generate_result_handler_for(type); |
aoqi@0 | 229 | } |
aoqi@0 | 230 | |
aoqi@0 | 231 | address TemplateInterpreterGenerator::generate_safept_entry_for(TosState state, address runtime_entry) { |
aoqi@0 | 232 | address entry = __ pc(); |
aoqi@0 | 233 | |
aoqi@0 | 234 | __ push(state); |
aoqi@0 | 235 | __ call_VM(noreg, runtime_entry); |
aoqi@0 | 236 | __ dispatch_via(vtos, Interpreter::_normal_table.table_for(vtos)); |
aoqi@0 | 237 | |
aoqi@0 | 238 | return entry; |
aoqi@0 | 239 | } |
aoqi@0 | 240 | |
aoqi@0 | 241 | // Helpers for commoning out cases in the various type of method entries. |
aoqi@0 | 242 | |
aoqi@0 | 243 | // Increment invocation count & check for overflow. |
aoqi@0 | 244 | // |
aoqi@0 | 245 | // Note: checking for negative value instead of overflow |
aoqi@0 | 246 | // so we have a 'sticky' overflow test. |
aoqi@0 | 247 | // |
aoqi@0 | 248 | void TemplateInterpreterGenerator::generate_counter_incr(Label* overflow, Label* profile_method, Label* profile_method_continue) { |
aoqi@0 | 249 | // Note: In tiered we increment either counters in method or in MDO depending if we're profiling or not. |
aoqi@0 | 250 | Register Rscratch1 = R11_scratch1; |
aoqi@0 | 251 | Register Rscratch2 = R12_scratch2; |
aoqi@0 | 252 | Register R3_counters = R3_ARG1; |
aoqi@0 | 253 | Label done; |
aoqi@0 | 254 | |
aoqi@0 | 255 | if (TieredCompilation) { |
aoqi@0 | 256 | const int increment = InvocationCounter::count_increment; |
aoqi@0 | 257 | const int mask = ((1 << Tier0InvokeNotifyFreqLog) - 1) << InvocationCounter::count_shift; |
aoqi@0 | 258 | Label no_mdo; |
aoqi@0 | 259 | if (ProfileInterpreter) { |
aoqi@0 | 260 | const Register Rmdo = Rscratch1; |
aoqi@0 | 261 | // If no method data exists, go to profile_continue. |
aoqi@0 | 262 | __ ld(Rmdo, in_bytes(Method::method_data_offset()), R19_method); |
aoqi@0 | 263 | __ cmpdi(CCR0, Rmdo, 0); |
aoqi@0 | 264 | __ beq(CCR0, no_mdo); |
aoqi@0 | 265 | |
aoqi@0 | 266 | // Increment backedge counter in the MDO. |
aoqi@0 | 267 | const int mdo_bc_offs = in_bytes(MethodData::backedge_counter_offset()) + in_bytes(InvocationCounter::counter_offset()); |
aoqi@0 | 268 | __ lwz(Rscratch2, mdo_bc_offs, Rmdo); |
aoqi@0 | 269 | __ addi(Rscratch2, Rscratch2, increment); |
aoqi@0 | 270 | __ stw(Rscratch2, mdo_bc_offs, Rmdo); |
aoqi@0 | 271 | __ load_const_optimized(Rscratch1, mask, R0); |
aoqi@0 | 272 | __ and_(Rscratch1, Rscratch2, Rscratch1); |
aoqi@0 | 273 | __ bne(CCR0, done); |
aoqi@0 | 274 | __ b(*overflow); |
aoqi@0 | 275 | } |
aoqi@0 | 276 | |
aoqi@0 | 277 | // Increment counter in MethodCounters*. |
aoqi@0 | 278 | const int mo_bc_offs = in_bytes(MethodCounters::backedge_counter_offset()) + in_bytes(InvocationCounter::counter_offset()); |
aoqi@0 | 279 | __ bind(no_mdo); |
aoqi@0 | 280 | __ get_method_counters(R19_method, R3_counters, done); |
aoqi@0 | 281 | __ lwz(Rscratch2, mo_bc_offs, R3_counters); |
aoqi@0 | 282 | __ addi(Rscratch2, Rscratch2, increment); |
aoqi@0 | 283 | __ stw(Rscratch2, mo_bc_offs, R3_counters); |
aoqi@0 | 284 | __ load_const_optimized(Rscratch1, mask, R0); |
aoqi@0 | 285 | __ and_(Rscratch1, Rscratch2, Rscratch1); |
aoqi@0 | 286 | __ beq(CCR0, *overflow); |
aoqi@0 | 287 | |
aoqi@0 | 288 | __ bind(done); |
aoqi@0 | 289 | |
aoqi@0 | 290 | } else { |
aoqi@0 | 291 | |
aoqi@0 | 292 | // Update standard invocation counters. |
aoqi@0 | 293 | Register Rsum_ivc_bec = R4_ARG2; |
aoqi@0 | 294 | __ get_method_counters(R19_method, R3_counters, done); |
aoqi@0 | 295 | __ increment_invocation_counter(R3_counters, Rsum_ivc_bec, R12_scratch2); |
aoqi@0 | 296 | // Increment interpreter invocation counter. |
aoqi@0 | 297 | if (ProfileInterpreter) { // %%% Merge this into methodDataOop. |
aoqi@0 | 298 | __ lwz(R12_scratch2, in_bytes(MethodCounters::interpreter_invocation_counter_offset()), R3_counters); |
aoqi@0 | 299 | __ addi(R12_scratch2, R12_scratch2, 1); |
aoqi@0 | 300 | __ stw(R12_scratch2, in_bytes(MethodCounters::interpreter_invocation_counter_offset()), R3_counters); |
aoqi@0 | 301 | } |
aoqi@0 | 302 | // Check if we must create a method data obj. |
aoqi@0 | 303 | if (ProfileInterpreter && profile_method != NULL) { |
aoqi@0 | 304 | const Register profile_limit = Rscratch1; |
aoqi@0 | 305 | int pl_offs = __ load_const_optimized(profile_limit, &InvocationCounter::InterpreterProfileLimit, R0, true); |
aoqi@0 | 306 | __ lwz(profile_limit, pl_offs, profile_limit); |
aoqi@0 | 307 | // Test to see if we should create a method data oop. |
aoqi@0 | 308 | __ cmpw(CCR0, Rsum_ivc_bec, profile_limit); |
aoqi@0 | 309 | __ blt(CCR0, *profile_method_continue); |
aoqi@0 | 310 | // If no method data exists, go to profile_method. |
aoqi@0 | 311 | __ test_method_data_pointer(*profile_method); |
aoqi@0 | 312 | } |
aoqi@0 | 313 | // Finally check for counter overflow. |
aoqi@0 | 314 | if (overflow) { |
aoqi@0 | 315 | const Register invocation_limit = Rscratch1; |
aoqi@0 | 316 | int il_offs = __ load_const_optimized(invocation_limit, &InvocationCounter::InterpreterInvocationLimit, R0, true); |
aoqi@0 | 317 | __ lwz(invocation_limit, il_offs, invocation_limit); |
aoqi@0 | 318 | assert(4 == sizeof(InvocationCounter::InterpreterInvocationLimit), "unexpected field size"); |
aoqi@0 | 319 | __ cmpw(CCR0, Rsum_ivc_bec, invocation_limit); |
aoqi@0 | 320 | __ bge(CCR0, *overflow); |
aoqi@0 | 321 | } |
aoqi@0 | 322 | |
aoqi@0 | 323 | __ bind(done); |
aoqi@0 | 324 | } |
aoqi@0 | 325 | } |
aoqi@0 | 326 | |
aoqi@0 | 327 | // Generate code to initiate compilation on invocation counter overflow. |
aoqi@0 | 328 | void TemplateInterpreterGenerator::generate_counter_overflow(Label& continue_entry) { |
aoqi@0 | 329 | // Generate code to initiate compilation on the counter overflow. |
aoqi@0 | 330 | |
aoqi@0 | 331 | // InterpreterRuntime::frequency_counter_overflow takes one arguments, |
aoqi@0 | 332 | // which indicates if the counter overflow occurs at a backwards branch (NULL bcp) |
aoqi@0 | 333 | // We pass zero in. |
aoqi@0 | 334 | // The call returns the address of the verified entry point for the method or NULL |
aoqi@0 | 335 | // if the compilation did not complete (either went background or bailed out). |
aoqi@0 | 336 | // |
aoqi@0 | 337 | // Unlike the C++ interpreter above: Check exceptions! |
aoqi@0 | 338 | // Assumption: Caller must set the flag "do_not_unlock_if_sychronized" if the monitor of a sync'ed |
aoqi@0 | 339 | // method has not yet been created. Thus, no unlocking of a non-existing monitor can occur. |
aoqi@0 | 340 | |
aoqi@0 | 341 | __ li(R4_ARG2, 0); |
aoqi@0 | 342 | __ call_VM(noreg, CAST_FROM_FN_PTR(address, InterpreterRuntime::frequency_counter_overflow), R4_ARG2, true); |
aoqi@0 | 343 | |
aoqi@0 | 344 | // Returns verified_entry_point or NULL. |
aoqi@0 | 345 | // We ignore it in any case. |
aoqi@0 | 346 | __ b(continue_entry); |
aoqi@0 | 347 | } |
aoqi@0 | 348 | |
aoqi@0 | 349 | void TemplateInterpreterGenerator::generate_stack_overflow_check(Register Rmem_frame_size, Register Rscratch1) { |
aoqi@0 | 350 | assert_different_registers(Rmem_frame_size, Rscratch1); |
aoqi@0 | 351 | __ generate_stack_overflow_check_with_compare_and_throw(Rmem_frame_size, Rscratch1); |
aoqi@0 | 352 | } |
aoqi@0 | 353 | |
aoqi@0 | 354 | void TemplateInterpreterGenerator::unlock_method(bool check_exceptions) { |
aoqi@0 | 355 | __ unlock_object(R26_monitor, check_exceptions); |
aoqi@0 | 356 | } |
aoqi@0 | 357 | |
aoqi@0 | 358 | // Lock the current method, interpreter register window must be set up! |
aoqi@0 | 359 | void TemplateInterpreterGenerator::lock_method(Register Rflags, Register Rscratch1, Register Rscratch2, bool flags_preloaded) { |
aoqi@0 | 360 | const Register Robj_to_lock = Rscratch2; |
aoqi@0 | 361 | |
aoqi@0 | 362 | { |
aoqi@0 | 363 | if (!flags_preloaded) { |
aoqi@0 | 364 | __ lwz(Rflags, method_(access_flags)); |
aoqi@0 | 365 | } |
aoqi@0 | 366 | |
aoqi@0 | 367 | #ifdef ASSERT |
aoqi@0 | 368 | // Check if methods needs synchronization. |
aoqi@0 | 369 | { |
aoqi@0 | 370 | Label Lok; |
aoqi@0 | 371 | __ testbitdi(CCR0, R0, Rflags, JVM_ACC_SYNCHRONIZED_BIT); |
aoqi@0 | 372 | __ btrue(CCR0,Lok); |
aoqi@0 | 373 | __ stop("method doesn't need synchronization"); |
aoqi@0 | 374 | __ bind(Lok); |
aoqi@0 | 375 | } |
aoqi@0 | 376 | #endif // ASSERT |
aoqi@0 | 377 | } |
aoqi@0 | 378 | |
aoqi@0 | 379 | // Get synchronization object to Rscratch2. |
aoqi@0 | 380 | { |
aoqi@0 | 381 | const int mirror_offset = in_bytes(Klass::java_mirror_offset()); |
aoqi@0 | 382 | Label Lstatic; |
aoqi@0 | 383 | Label Ldone; |
aoqi@0 | 384 | |
aoqi@0 | 385 | __ testbitdi(CCR0, R0, Rflags, JVM_ACC_STATIC_BIT); |
aoqi@0 | 386 | __ btrue(CCR0, Lstatic); |
aoqi@0 | 387 | |
aoqi@0 | 388 | // Non-static case: load receiver obj from stack and we're done. |
aoqi@0 | 389 | __ ld(Robj_to_lock, R18_locals); |
aoqi@0 | 390 | __ b(Ldone); |
aoqi@0 | 391 | |
aoqi@0 | 392 | __ bind(Lstatic); // Static case: Lock the java mirror |
aoqi@0 | 393 | __ ld(Robj_to_lock, in_bytes(Method::const_offset()), R19_method); |
aoqi@0 | 394 | __ ld(Robj_to_lock, in_bytes(ConstMethod::constants_offset()), Robj_to_lock); |
aoqi@0 | 395 | __ ld(Robj_to_lock, ConstantPool::pool_holder_offset_in_bytes(), Robj_to_lock); |
aoqi@0 | 396 | __ ld(Robj_to_lock, mirror_offset, Robj_to_lock); |
aoqi@0 | 397 | |
aoqi@0 | 398 | __ bind(Ldone); |
aoqi@0 | 399 | __ verify_oop(Robj_to_lock); |
aoqi@0 | 400 | } |
aoqi@0 | 401 | |
aoqi@0 | 402 | // Got the oop to lock => execute! |
aoqi@0 | 403 | __ add_monitor_to_stack(true, Rscratch1, R0); |
aoqi@0 | 404 | |
aoqi@0 | 405 | __ std(Robj_to_lock, BasicObjectLock::obj_offset_in_bytes(), R26_monitor); |
aoqi@0 | 406 | __ lock_object(R26_monitor, Robj_to_lock); |
aoqi@0 | 407 | } |
aoqi@0 | 408 | |
aoqi@0 | 409 | // Generate a fixed interpreter frame for pure interpreter |
aoqi@0 | 410 | // and I2N native transition frames. |
aoqi@0 | 411 | // |
aoqi@0 | 412 | // Before (stack grows downwards): |
aoqi@0 | 413 | // |
aoqi@0 | 414 | // | ... | |
aoqi@0 | 415 | // |------------- | |
aoqi@0 | 416 | // | java arg0 | |
aoqi@0 | 417 | // | ... | |
aoqi@0 | 418 | // | java argn | |
aoqi@0 | 419 | // | | <- R15_esp |
aoqi@0 | 420 | // | | |
aoqi@0 | 421 | // |--------------| |
aoqi@0 | 422 | // | abi_112 | |
aoqi@0 | 423 | // | | <- R1_SP |
aoqi@0 | 424 | // |==============| |
aoqi@0 | 425 | // |
aoqi@0 | 426 | // |
aoqi@0 | 427 | // After: |
aoqi@0 | 428 | // |
aoqi@0 | 429 | // | ... | |
aoqi@0 | 430 | // | java arg0 |<- R18_locals |
aoqi@0 | 431 | // | ... | |
aoqi@0 | 432 | // | java argn | |
aoqi@0 | 433 | // |--------------| |
aoqi@0 | 434 | // | | |
aoqi@0 | 435 | // | java locals | |
aoqi@0 | 436 | // | | |
aoqi@0 | 437 | // |--------------| |
aoqi@0 | 438 | // | abi_48 | |
aoqi@0 | 439 | // |==============| |
aoqi@0 | 440 | // | | |
aoqi@0 | 441 | // | istate | |
aoqi@0 | 442 | // | | |
aoqi@0 | 443 | // |--------------| |
aoqi@0 | 444 | // | monitor |<- R26_monitor |
aoqi@0 | 445 | // |--------------| |
aoqi@0 | 446 | // | |<- R15_esp |
aoqi@0 | 447 | // | expression | |
aoqi@0 | 448 | // | stack | |
aoqi@0 | 449 | // | | |
aoqi@0 | 450 | // |--------------| |
aoqi@0 | 451 | // | | |
aoqi@0 | 452 | // | abi_112 |<- R1_SP |
aoqi@0 | 453 | // |==============| |
aoqi@0 | 454 | // |
aoqi@0 | 455 | // The top most frame needs an abi space of 112 bytes. This space is needed, |
aoqi@0 | 456 | // since we call to c. The c function may spill their arguments to the caller |
aoqi@0 | 457 | // frame. When we call to java, we don't need these spill slots. In order to save |
aoqi@0 | 458 | // space on the stack, we resize the caller. However, java local reside in |
aoqi@0 | 459 | // the caller frame and the frame has to be increased. The frame_size for the |
aoqi@0 | 460 | // current frame was calculated based on max_stack as size for the expression |
aoqi@0 | 461 | // stack. At the call, just a part of the expression stack might be used. |
aoqi@0 | 462 | // We don't want to waste this space and cut the frame back accordingly. |
aoqi@0 | 463 | // The resulting amount for resizing is calculated as follows: |
aoqi@0 | 464 | // resize = (number_of_locals - number_of_arguments) * slot_size |
aoqi@0 | 465 | // + (R1_SP - R15_esp) + 48 |
aoqi@0 | 466 | // |
aoqi@0 | 467 | // The size for the callee frame is calculated: |
aoqi@0 | 468 | // framesize = 112 + max_stack + monitor + state_size |
aoqi@0 | 469 | // |
aoqi@0 | 470 | // maxstack: Max number of slots on the expression stack, loaded from the method. |
aoqi@0 | 471 | // monitor: We statically reserve room for one monitor object. |
aoqi@0 | 472 | // state_size: We save the current state of the interpreter to this area. |
aoqi@0 | 473 | // |
aoqi@0 | 474 | void TemplateInterpreterGenerator::generate_fixed_frame(bool native_call, Register Rsize_of_parameters, Register Rsize_of_locals) { |
aoqi@0 | 475 | Register parent_frame_resize = R6_ARG4, // Frame will grow by this number of bytes. |
aoqi@0 | 476 | top_frame_size = R7_ARG5, |
aoqi@0 | 477 | Rconst_method = R8_ARG6; |
aoqi@0 | 478 | |
aoqi@0 | 479 | assert_different_registers(Rsize_of_parameters, Rsize_of_locals, parent_frame_resize, top_frame_size); |
aoqi@0 | 480 | |
aoqi@0 | 481 | __ ld(Rconst_method, method_(const)); |
aoqi@0 | 482 | __ lhz(Rsize_of_parameters /* number of params */, |
aoqi@0 | 483 | in_bytes(ConstMethod::size_of_parameters_offset()), Rconst_method); |
aoqi@0 | 484 | if (native_call) { |
aoqi@0 | 485 | // If we're calling a native method, we reserve space for the worst-case signature |
aoqi@0 | 486 | // handler varargs vector, which is max(Argument::n_register_parameters, parameter_count+2). |
aoqi@0 | 487 | // We add two slots to the parameter_count, one for the jni |
aoqi@0 | 488 | // environment and one for a possible native mirror. |
aoqi@0 | 489 | Label skip_native_calculate_max_stack; |
aoqi@0 | 490 | __ addi(top_frame_size, Rsize_of_parameters, 2); |
aoqi@0 | 491 | __ cmpwi(CCR0, top_frame_size, Argument::n_register_parameters); |
aoqi@0 | 492 | __ bge(CCR0, skip_native_calculate_max_stack); |
aoqi@0 | 493 | __ li(top_frame_size, Argument::n_register_parameters); |
aoqi@0 | 494 | __ bind(skip_native_calculate_max_stack); |
aoqi@0 | 495 | __ sldi(Rsize_of_parameters, Rsize_of_parameters, Interpreter::logStackElementSize); |
aoqi@0 | 496 | __ sldi(top_frame_size, top_frame_size, Interpreter::logStackElementSize); |
aoqi@0 | 497 | __ sub(parent_frame_resize, R1_SP, R15_esp); // <0, off by Interpreter::stackElementSize! |
aoqi@0 | 498 | assert(Rsize_of_locals == noreg, "Rsize_of_locals not initialized"); // Only relevant value is Rsize_of_parameters. |
aoqi@0 | 499 | } else { |
aoqi@0 | 500 | __ lhz(Rsize_of_locals /* number of params */, in_bytes(ConstMethod::size_of_locals_offset()), Rconst_method); |
aoqi@0 | 501 | __ sldi(Rsize_of_parameters, Rsize_of_parameters, Interpreter::logStackElementSize); |
aoqi@0 | 502 | __ sldi(Rsize_of_locals, Rsize_of_locals, Interpreter::logStackElementSize); |
aoqi@0 | 503 | __ lhz(top_frame_size, in_bytes(ConstMethod::max_stack_offset()), Rconst_method); |
aoqi@0 | 504 | __ sub(R11_scratch1, Rsize_of_locals, Rsize_of_parameters); // >=0 |
aoqi@0 | 505 | __ sub(parent_frame_resize, R1_SP, R15_esp); // <0, off by Interpreter::stackElementSize! |
aoqi@0 | 506 | __ sldi(top_frame_size, top_frame_size, Interpreter::logStackElementSize); |
aoqi@0 | 507 | __ add(parent_frame_resize, parent_frame_resize, R11_scratch1); |
aoqi@0 | 508 | } |
aoqi@0 | 509 | |
aoqi@0 | 510 | // Compute top frame size. |
aoqi@0 | 511 | __ addi(top_frame_size, top_frame_size, frame::abi_reg_args_size + frame::ijava_state_size); |
aoqi@0 | 512 | |
aoqi@0 | 513 | // Cut back area between esp and max_stack. |
aoqi@0 | 514 | __ addi(parent_frame_resize, parent_frame_resize, frame::abi_minframe_size - Interpreter::stackElementSize); |
aoqi@0 | 515 | |
aoqi@0 | 516 | __ round_to(top_frame_size, frame::alignment_in_bytes); |
aoqi@0 | 517 | __ round_to(parent_frame_resize, frame::alignment_in_bytes); |
aoqi@0 | 518 | // parent_frame_resize = (locals-parameters) - (ESP-SP-ABI48) Rounded to frame alignment size. |
aoqi@0 | 519 | // Enlarge by locals-parameters (not in case of native_call), shrink by ESP-SP-ABI48. |
aoqi@0 | 520 | |
aoqi@0 | 521 | { |
aoqi@0 | 522 | // -------------------------------------------------------------------------- |
aoqi@0 | 523 | // Stack overflow check |
aoqi@0 | 524 | |
aoqi@0 | 525 | Label cont; |
aoqi@0 | 526 | __ add(R11_scratch1, parent_frame_resize, top_frame_size); |
aoqi@0 | 527 | generate_stack_overflow_check(R11_scratch1, R12_scratch2); |
aoqi@0 | 528 | } |
aoqi@0 | 529 | |
aoqi@0 | 530 | // Set up interpreter state registers. |
aoqi@0 | 531 | |
aoqi@0 | 532 | __ add(R18_locals, R15_esp, Rsize_of_parameters); |
aoqi@0 | 533 | __ ld(R27_constPoolCache, in_bytes(ConstMethod::constants_offset()), Rconst_method); |
aoqi@0 | 534 | __ ld(R27_constPoolCache, ConstantPool::cache_offset_in_bytes(), R27_constPoolCache); |
aoqi@0 | 535 | |
aoqi@0 | 536 | // Set method data pointer. |
aoqi@0 | 537 | if (ProfileInterpreter) { |
aoqi@0 | 538 | Label zero_continue; |
aoqi@0 | 539 | __ ld(R28_mdx, method_(method_data)); |
aoqi@0 | 540 | __ cmpdi(CCR0, R28_mdx, 0); |
aoqi@0 | 541 | __ beq(CCR0, zero_continue); |
aoqi@0 | 542 | __ addi(R28_mdx, R28_mdx, in_bytes(MethodData::data_offset())); |
aoqi@0 | 543 | __ bind(zero_continue); |
aoqi@0 | 544 | } |
aoqi@0 | 545 | |
aoqi@0 | 546 | if (native_call) { |
aoqi@0 | 547 | __ li(R14_bcp, 0); // Must initialize. |
aoqi@0 | 548 | } else { |
aoqi@0 | 549 | __ add(R14_bcp, in_bytes(ConstMethod::codes_offset()), Rconst_method); |
aoqi@0 | 550 | } |
aoqi@0 | 551 | |
aoqi@0 | 552 | // Resize parent frame. |
aoqi@0 | 553 | __ mflr(R12_scratch2); |
aoqi@0 | 554 | __ neg(parent_frame_resize, parent_frame_resize); |
aoqi@0 | 555 | __ resize_frame(parent_frame_resize, R11_scratch1); |
aoqi@0 | 556 | __ std(R12_scratch2, _abi(lr), R1_SP); |
aoqi@0 | 557 | |
aoqi@0 | 558 | __ addi(R26_monitor, R1_SP, - frame::ijava_state_size); |
aoqi@0 | 559 | __ addi(R15_esp, R26_monitor, - Interpreter::stackElementSize); |
aoqi@0 | 560 | |
aoqi@0 | 561 | // Store values. |
aoqi@0 | 562 | // R15_esp, R14_bcp, R26_monitor, R28_mdx are saved at java calls |
aoqi@0 | 563 | // in InterpreterMacroAssembler::call_from_interpreter. |
aoqi@0 | 564 | __ std(R19_method, _ijava_state_neg(method), R1_SP); |
aoqi@0 | 565 | __ std(R21_sender_SP, _ijava_state_neg(sender_sp), R1_SP); |
aoqi@0 | 566 | __ std(R27_constPoolCache, _ijava_state_neg(cpoolCache), R1_SP); |
aoqi@0 | 567 | __ std(R18_locals, _ijava_state_neg(locals), R1_SP); |
aoqi@0 | 568 | |
aoqi@0 | 569 | // Note: esp, bcp, monitor, mdx live in registers. Hence, the correct version can only |
aoqi@0 | 570 | // be found in the frame after save_interpreter_state is done. This is always true |
aoqi@0 | 571 | // for non-top frames. But when a signal occurs, dumping the top frame can go wrong, |
aoqi@0 | 572 | // because e.g. frame::interpreter_frame_bcp() will not access the correct value |
aoqi@0 | 573 | // (Enhanced Stack Trace). |
aoqi@0 | 574 | // The signal handler does not save the interpreter state into the frame. |
aoqi@0 | 575 | __ li(R0, 0); |
aoqi@0 | 576 | #ifdef ASSERT |
aoqi@0 | 577 | // Fill remaining slots with constants. |
aoqi@0 | 578 | __ load_const_optimized(R11_scratch1, 0x5afe); |
aoqi@0 | 579 | __ load_const_optimized(R12_scratch2, 0xdead); |
aoqi@0 | 580 | #endif |
aoqi@0 | 581 | // We have to initialize some frame slots for native calls (accessed by GC). |
aoqi@0 | 582 | if (native_call) { |
aoqi@0 | 583 | __ std(R26_monitor, _ijava_state_neg(monitors), R1_SP); |
aoqi@0 | 584 | __ std(R14_bcp, _ijava_state_neg(bcp), R1_SP); |
aoqi@0 | 585 | if (ProfileInterpreter) { __ std(R28_mdx, _ijava_state_neg(mdx), R1_SP); } |
aoqi@0 | 586 | } |
aoqi@0 | 587 | #ifdef ASSERT |
aoqi@0 | 588 | else { |
aoqi@0 | 589 | __ std(R12_scratch2, _ijava_state_neg(monitors), R1_SP); |
aoqi@0 | 590 | __ std(R12_scratch2, _ijava_state_neg(bcp), R1_SP); |
aoqi@0 | 591 | __ std(R12_scratch2, _ijava_state_neg(mdx), R1_SP); |
aoqi@0 | 592 | } |
aoqi@0 | 593 | __ std(R11_scratch1, _ijava_state_neg(ijava_reserved), R1_SP); |
aoqi@0 | 594 | __ std(R12_scratch2, _ijava_state_neg(esp), R1_SP); |
aoqi@0 | 595 | __ std(R12_scratch2, _ijava_state_neg(lresult), R1_SP); |
aoqi@0 | 596 | __ std(R12_scratch2, _ijava_state_neg(fresult), R1_SP); |
aoqi@0 | 597 | #endif |
aoqi@0 | 598 | __ subf(R12_scratch2, top_frame_size, R1_SP); |
aoqi@0 | 599 | __ std(R0, _ijava_state_neg(oop_tmp), R1_SP); |
aoqi@0 | 600 | __ std(R12_scratch2, _ijava_state_neg(top_frame_sp), R1_SP); |
aoqi@0 | 601 | |
aoqi@0 | 602 | // Push top frame. |
aoqi@0 | 603 | __ push_frame(top_frame_size, R11_scratch1); |
aoqi@0 | 604 | } |
aoqi@0 | 605 | |
aoqi@0 | 606 | // End of helpers |
aoqi@0 | 607 | |
aoqi@0 | 608 | // ============================================================================ |
aoqi@0 | 609 | // Various method entries |
aoqi@0 | 610 | // |
aoqi@0 | 611 | |
aoqi@0 | 612 | // Empty method, generate a very fast return. We must skip this entry if |
aoqi@0 | 613 | // someone's debugging, indicated by the flag |
aoqi@0 | 614 | // "interp_mode" in the Thread obj. |
aoqi@0 | 615 | // Note: empty methods are generated mostly methods that do assertions, which are |
aoqi@0 | 616 | // disabled in the "java opt build". |
aoqi@0 | 617 | address TemplateInterpreterGenerator::generate_empty_entry(void) { |
aoqi@0 | 618 | if (!UseFastEmptyMethods) { |
aoqi@0 | 619 | NOT_PRODUCT(__ should_not_reach_here();) |
aoqi@0 | 620 | return Interpreter::entry_for_kind(Interpreter::zerolocals); |
aoqi@0 | 621 | } |
aoqi@0 | 622 | |
aoqi@0 | 623 | Label Lslow_path; |
aoqi@0 | 624 | const Register Rjvmti_mode = R11_scratch1; |
aoqi@0 | 625 | address entry = __ pc(); |
aoqi@0 | 626 | |
aoqi@0 | 627 | __ lwz(Rjvmti_mode, thread_(interp_only_mode)); |
aoqi@0 | 628 | __ cmpwi(CCR0, Rjvmti_mode, 0); |
aoqi@0 | 629 | __ bne(CCR0, Lslow_path); // jvmti_mode!=0 |
aoqi@0 | 630 | |
aoqi@0 | 631 | // Noone's debuggin: Simply return. |
aoqi@0 | 632 | // Pop c2i arguments (if any) off when we return. |
aoqi@0 | 633 | #ifdef ASSERT |
aoqi@0 | 634 | __ ld(R9_ARG7, 0, R1_SP); |
aoqi@0 | 635 | __ ld(R10_ARG8, 0, R21_sender_SP); |
aoqi@0 | 636 | __ cmpd(CCR0, R9_ARG7, R10_ARG8); |
aoqi@0 | 637 | __ asm_assert_eq("backlink", 0x545); |
aoqi@0 | 638 | #endif // ASSERT |
aoqi@0 | 639 | __ mr(R1_SP, R21_sender_SP); // Cut the stack back to where the caller started. |
aoqi@0 | 640 | |
aoqi@0 | 641 | // And we're done. |
aoqi@0 | 642 | __ blr(); |
aoqi@0 | 643 | |
aoqi@0 | 644 | __ bind(Lslow_path); |
aoqi@0 | 645 | __ branch_to_entry(Interpreter::entry_for_kind(Interpreter::zerolocals), R11_scratch1); |
aoqi@0 | 646 | __ flush(); |
aoqi@0 | 647 | |
aoqi@0 | 648 | return entry; |
aoqi@0 | 649 | } |
aoqi@0 | 650 | |
aoqi@0 | 651 | // Support abs and sqrt like in compiler. |
aoqi@0 | 652 | // For others we can use a normal (native) entry. |
aoqi@0 | 653 | |
aoqi@0 | 654 | inline bool math_entry_available(AbstractInterpreter::MethodKind kind) { |
aoqi@0 | 655 | // Provide math entry with debugging on demand. |
aoqi@0 | 656 | // Note: Debugging changes which code will get executed: |
aoqi@0 | 657 | // Debugging or disabled InlineIntrinsics: java method will get interpreted and performs a native call. |
aoqi@0 | 658 | // Not debugging and enabled InlineIntrinics: processor instruction will get used. |
aoqi@0 | 659 | // Result might differ slightly due to rounding etc. |
aoqi@0 | 660 | if (!InlineIntrinsics && (!FLAG_IS_ERGO(InlineIntrinsics))) return false; // Generate a vanilla entry. |
aoqi@0 | 661 | |
aoqi@0 | 662 | return ((kind==Interpreter::java_lang_math_sqrt && VM_Version::has_fsqrt()) || |
aoqi@0 | 663 | (kind==Interpreter::java_lang_math_abs)); |
aoqi@0 | 664 | } |
aoqi@0 | 665 | |
aoqi@0 | 666 | address TemplateInterpreterGenerator::generate_math_entry(AbstractInterpreter::MethodKind kind) { |
aoqi@0 | 667 | if (!math_entry_available(kind)) { |
aoqi@0 | 668 | NOT_PRODUCT(__ should_not_reach_here();) |
aoqi@0 | 669 | return Interpreter::entry_for_kind(Interpreter::zerolocals); |
aoqi@0 | 670 | } |
aoqi@0 | 671 | |
aoqi@0 | 672 | Label Lslow_path; |
aoqi@0 | 673 | const Register Rjvmti_mode = R11_scratch1; |
aoqi@0 | 674 | address entry = __ pc(); |
aoqi@0 | 675 | |
aoqi@0 | 676 | // Provide math entry with debugging on demand. |
aoqi@0 | 677 | __ lwz(Rjvmti_mode, thread_(interp_only_mode)); |
aoqi@0 | 678 | __ cmpwi(CCR0, Rjvmti_mode, 0); |
aoqi@0 | 679 | __ bne(CCR0, Lslow_path); // jvmti_mode!=0 |
aoqi@0 | 680 | |
aoqi@0 | 681 | __ lfd(F1_RET, Interpreter::stackElementSize, R15_esp); |
aoqi@0 | 682 | |
aoqi@0 | 683 | // Pop c2i arguments (if any) off when we return. |
aoqi@0 | 684 | #ifdef ASSERT |
aoqi@0 | 685 | __ ld(R9_ARG7, 0, R1_SP); |
aoqi@0 | 686 | __ ld(R10_ARG8, 0, R21_sender_SP); |
aoqi@0 | 687 | __ cmpd(CCR0, R9_ARG7, R10_ARG8); |
aoqi@0 | 688 | __ asm_assert_eq("backlink", 0x545); |
aoqi@0 | 689 | #endif // ASSERT |
aoqi@0 | 690 | __ mr(R1_SP, R21_sender_SP); // Cut the stack back to where the caller started. |
aoqi@0 | 691 | |
aoqi@0 | 692 | if (kind == Interpreter::java_lang_math_sqrt) { |
aoqi@0 | 693 | __ fsqrt(F1_RET, F1_RET); |
aoqi@0 | 694 | } else if (kind == Interpreter::java_lang_math_abs) { |
aoqi@0 | 695 | __ fabs(F1_RET, F1_RET); |
aoqi@0 | 696 | } else { |
aoqi@0 | 697 | ShouldNotReachHere(); |
aoqi@0 | 698 | } |
aoqi@0 | 699 | |
aoqi@0 | 700 | // And we're done. |
aoqi@0 | 701 | __ blr(); |
aoqi@0 | 702 | |
aoqi@0 | 703 | // Provide slow path for JVMTI case. |
aoqi@0 | 704 | __ bind(Lslow_path); |
aoqi@0 | 705 | __ branch_to_entry(Interpreter::entry_for_kind(Interpreter::zerolocals), R12_scratch2); |
aoqi@0 | 706 | __ flush(); |
aoqi@0 | 707 | |
aoqi@0 | 708 | return entry; |
aoqi@0 | 709 | } |
aoqi@0 | 710 | |
aoqi@0 | 711 | // Interpreter stub for calling a native method. (asm interpreter) |
aoqi@0 | 712 | // This sets up a somewhat different looking stack for calling the |
aoqi@0 | 713 | // native method than the typical interpreter frame setup. |
aoqi@0 | 714 | // |
aoqi@0 | 715 | // On entry: |
aoqi@0 | 716 | // R19_method - method |
aoqi@0 | 717 | // R16_thread - JavaThread* |
aoqi@0 | 718 | // R15_esp - intptr_t* sender tos |
aoqi@0 | 719 | // |
aoqi@0 | 720 | // abstract stack (grows up) |
aoqi@0 | 721 | // [ IJava (caller of JNI callee) ] <-- ASP |
aoqi@0 | 722 | // ... |
aoqi@0 | 723 | address TemplateInterpreterGenerator::generate_native_entry(bool synchronized) { |
aoqi@0 | 724 | |
aoqi@0 | 725 | address entry = __ pc(); |
aoqi@0 | 726 | |
aoqi@0 | 727 | const bool inc_counter = UseCompiler || CountCompiledCalls; |
aoqi@0 | 728 | |
aoqi@0 | 729 | // ----------------------------------------------------------------------------- |
aoqi@0 | 730 | // Allocate a new frame that represents the native callee (i2n frame). |
aoqi@0 | 731 | // This is not a full-blown interpreter frame, but in particular, the |
aoqi@0 | 732 | // following registers are valid after this: |
aoqi@0 | 733 | // - R19_method |
aoqi@0 | 734 | // - R18_local (points to start of argumuments to native function) |
aoqi@0 | 735 | // |
aoqi@0 | 736 | // abstract stack (grows up) |
aoqi@0 | 737 | // [ IJava (caller of JNI callee) ] <-- ASP |
aoqi@0 | 738 | // ... |
aoqi@0 | 739 | |
aoqi@0 | 740 | const Register signature_handler_fd = R11_scratch1; |
aoqi@0 | 741 | const Register pending_exception = R0; |
aoqi@0 | 742 | const Register result_handler_addr = R31; |
aoqi@0 | 743 | const Register native_method_fd = R11_scratch1; |
aoqi@0 | 744 | const Register access_flags = R22_tmp2; |
aoqi@0 | 745 | const Register active_handles = R11_scratch1; // R26_monitor saved to state. |
aoqi@0 | 746 | const Register sync_state = R12_scratch2; |
aoqi@0 | 747 | const Register sync_state_addr = sync_state; // Address is dead after use. |
aoqi@0 | 748 | const Register suspend_flags = R11_scratch1; |
aoqi@0 | 749 | |
aoqi@0 | 750 | //============================================================================= |
aoqi@0 | 751 | // Allocate new frame and initialize interpreter state. |
aoqi@0 | 752 | |
aoqi@0 | 753 | Label exception_return; |
aoqi@0 | 754 | Label exception_return_sync_check; |
aoqi@0 | 755 | Label stack_overflow_return; |
aoqi@0 | 756 | |
aoqi@0 | 757 | // Generate new interpreter state and jump to stack_overflow_return in case of |
aoqi@0 | 758 | // a stack overflow. |
aoqi@0 | 759 | //generate_compute_interpreter_state(stack_overflow_return); |
aoqi@0 | 760 | |
aoqi@0 | 761 | Register size_of_parameters = R22_tmp2; |
aoqi@0 | 762 | |
aoqi@0 | 763 | generate_fixed_frame(true, size_of_parameters, noreg /* unused */); |
aoqi@0 | 764 | |
aoqi@0 | 765 | //============================================================================= |
aoqi@0 | 766 | // Increment invocation counter. On overflow, entry to JNI method |
aoqi@0 | 767 | // will be compiled. |
aoqi@0 | 768 | Label invocation_counter_overflow, continue_after_compile; |
aoqi@0 | 769 | if (inc_counter) { |
aoqi@0 | 770 | if (synchronized) { |
aoqi@0 | 771 | // Since at this point in the method invocation the exception handler |
aoqi@0 | 772 | // would try to exit the monitor of synchronized methods which hasn't |
aoqi@0 | 773 | // been entered yet, we set the thread local variable |
aoqi@0 | 774 | // _do_not_unlock_if_synchronized to true. If any exception was thrown by |
aoqi@0 | 775 | // runtime, exception handling i.e. unlock_if_synchronized_method will |
aoqi@0 | 776 | // check this thread local flag. |
aoqi@0 | 777 | // This flag has two effects, one is to force an unwind in the topmost |
aoqi@0 | 778 | // interpreter frame and not perform an unlock while doing so. |
aoqi@0 | 779 | __ li(R0, 1); |
aoqi@0 | 780 | __ stb(R0, in_bytes(JavaThread::do_not_unlock_if_synchronized_offset()), R16_thread); |
aoqi@0 | 781 | } |
aoqi@0 | 782 | generate_counter_incr(&invocation_counter_overflow, NULL, NULL); |
aoqi@0 | 783 | |
aoqi@0 | 784 | __ BIND(continue_after_compile); |
aoqi@0 | 785 | // Reset the _do_not_unlock_if_synchronized flag. |
aoqi@0 | 786 | if (synchronized) { |
aoqi@0 | 787 | __ li(R0, 0); |
aoqi@0 | 788 | __ stb(R0, in_bytes(JavaThread::do_not_unlock_if_synchronized_offset()), R16_thread); |
aoqi@0 | 789 | } |
aoqi@0 | 790 | } |
aoqi@0 | 791 | |
aoqi@0 | 792 | // access_flags = method->access_flags(); |
aoqi@0 | 793 | // Load access flags. |
aoqi@0 | 794 | assert(access_flags->is_nonvolatile(), |
aoqi@0 | 795 | "access_flags must be in a non-volatile register"); |
aoqi@0 | 796 | // Type check. |
aoqi@0 | 797 | assert(4 == sizeof(AccessFlags), "unexpected field size"); |
aoqi@0 | 798 | __ lwz(access_flags, method_(access_flags)); |
aoqi@0 | 799 | |
aoqi@0 | 800 | // We don't want to reload R19_method and access_flags after calls |
aoqi@0 | 801 | // to some helper functions. |
aoqi@0 | 802 | assert(R19_method->is_nonvolatile(), |
aoqi@0 | 803 | "R19_method must be a non-volatile register"); |
aoqi@0 | 804 | |
aoqi@0 | 805 | // Check for synchronized methods. Must happen AFTER invocation counter |
aoqi@0 | 806 | // check, so method is not locked if counter overflows. |
aoqi@0 | 807 | |
aoqi@0 | 808 | if (synchronized) { |
aoqi@0 | 809 | lock_method(access_flags, R11_scratch1, R12_scratch2, true); |
aoqi@0 | 810 | |
aoqi@0 | 811 | // Update monitor in state. |
aoqi@0 | 812 | __ ld(R11_scratch1, 0, R1_SP); |
aoqi@0 | 813 | __ std(R26_monitor, _ijava_state_neg(monitors), R11_scratch1); |
aoqi@0 | 814 | } |
aoqi@0 | 815 | |
aoqi@0 | 816 | // jvmti/jvmpi support |
aoqi@0 | 817 | __ notify_method_entry(); |
aoqi@0 | 818 | |
aoqi@0 | 819 | //============================================================================= |
aoqi@0 | 820 | // Get and call the signature handler. |
aoqi@0 | 821 | |
aoqi@0 | 822 | __ ld(signature_handler_fd, method_(signature_handler)); |
aoqi@0 | 823 | Label call_signature_handler; |
aoqi@0 | 824 | |
aoqi@0 | 825 | __ cmpdi(CCR0, signature_handler_fd, 0); |
aoqi@0 | 826 | __ bne(CCR0, call_signature_handler); |
aoqi@0 | 827 | |
aoqi@0 | 828 | // Method has never been called. Either generate a specialized |
aoqi@0 | 829 | // handler or point to the slow one. |
aoqi@0 | 830 | // |
aoqi@0 | 831 | // Pass parameter 'false' to avoid exception check in call_VM. |
aoqi@0 | 832 | __ call_VM(noreg, CAST_FROM_FN_PTR(address, InterpreterRuntime::prepare_native_call), R19_method, false); |
aoqi@0 | 833 | |
aoqi@0 | 834 | // Check for an exception while looking up the target method. If we |
aoqi@0 | 835 | // incurred one, bail. |
aoqi@0 | 836 | __ ld(pending_exception, thread_(pending_exception)); |
aoqi@0 | 837 | __ cmpdi(CCR0, pending_exception, 0); |
aoqi@0 | 838 | __ bne(CCR0, exception_return_sync_check); // Has pending exception. |
aoqi@0 | 839 | |
aoqi@0 | 840 | // Reload signature handler, it may have been created/assigned in the meanwhile. |
aoqi@0 | 841 | __ ld(signature_handler_fd, method_(signature_handler)); |
aoqi@0 | 842 | __ twi_0(signature_handler_fd); // Order wrt. load of klass mirror and entry point (isync is below). |
aoqi@0 | 843 | |
aoqi@0 | 844 | __ BIND(call_signature_handler); |
aoqi@0 | 845 | |
aoqi@0 | 846 | // Before we call the signature handler we push a new frame to |
aoqi@0 | 847 | // protect the interpreter frame volatile registers when we return |
aoqi@0 | 848 | // from jni but before we can get back to Java. |
aoqi@0 | 849 | |
aoqi@0 | 850 | // First set the frame anchor while the SP/FP registers are |
aoqi@0 | 851 | // convenient and the slow signature handler can use this same frame |
aoqi@0 | 852 | // anchor. |
aoqi@0 | 853 | |
aoqi@0 | 854 | // We have a TOP_IJAVA_FRAME here, which belongs to us. |
aoqi@0 | 855 | __ set_top_ijava_frame_at_SP_as_last_Java_frame(R1_SP, R12_scratch2/*tmp*/); |
aoqi@0 | 856 | |
aoqi@0 | 857 | // Now the interpreter frame (and its call chain) have been |
aoqi@0 | 858 | // invalidated and flushed. We are now protected against eager |
aoqi@0 | 859 | // being enabled in native code. Even if it goes eager the |
aoqi@0 | 860 | // registers will be reloaded as clean and we will invalidate after |
aoqi@0 | 861 | // the call so no spurious flush should be possible. |
aoqi@0 | 862 | |
aoqi@0 | 863 | // Call signature handler and pass locals address. |
aoqi@0 | 864 | // |
aoqi@0 | 865 | // Our signature handlers copy required arguments to the C stack |
aoqi@0 | 866 | // (outgoing C args), R3_ARG1 to R10_ARG8, and FARG1 to FARG13. |
aoqi@0 | 867 | __ mr(R3_ARG1, R18_locals); |
kvn@7132 | 868 | #if !defined(ABI_ELFv2) |
aoqi@0 | 869 | __ ld(signature_handler_fd, 0, signature_handler_fd); |
kvn@7132 | 870 | #endif |
aoqi@0 | 871 | |
aoqi@0 | 872 | __ call_stub(signature_handler_fd); |
aoqi@0 | 873 | |
aoqi@0 | 874 | // Remove the register parameter varargs slots we allocated in |
aoqi@0 | 875 | // compute_interpreter_state. SP+16 ends up pointing to the ABI |
aoqi@0 | 876 | // outgoing argument area. |
aoqi@0 | 877 | // |
aoqi@0 | 878 | // Not needed on PPC64. |
aoqi@0 | 879 | //__ add(SP, SP, Argument::n_register_parameters*BytesPerWord); |
aoqi@0 | 880 | |
aoqi@0 | 881 | assert(result_handler_addr->is_nonvolatile(), "result_handler_addr must be in a non-volatile register"); |
aoqi@0 | 882 | // Save across call to native method. |
aoqi@0 | 883 | __ mr(result_handler_addr, R3_RET); |
aoqi@0 | 884 | |
aoqi@0 | 885 | __ isync(); // Acquire signature handler before trying to fetch the native entry point and klass mirror. |
aoqi@0 | 886 | |
aoqi@0 | 887 | // Set up fixed parameters and call the native method. |
aoqi@0 | 888 | // If the method is static, get mirror into R4_ARG2. |
aoqi@0 | 889 | { |
aoqi@0 | 890 | Label method_is_not_static; |
aoqi@0 | 891 | // Access_flags is non-volatile and still, no need to restore it. |
aoqi@0 | 892 | |
aoqi@0 | 893 | // Restore access flags. |
aoqi@0 | 894 | __ testbitdi(CCR0, R0, access_flags, JVM_ACC_STATIC_BIT); |
aoqi@0 | 895 | __ bfalse(CCR0, method_is_not_static); |
aoqi@0 | 896 | |
aoqi@0 | 897 | // constants = method->constants(); |
aoqi@0 | 898 | __ ld(R11_scratch1, in_bytes(Method::const_offset()), R19_method); |
aoqi@0 | 899 | __ ld(R11_scratch1, in_bytes(ConstMethod::constants_offset()), R11_scratch1); |
aoqi@0 | 900 | // pool_holder = method->constants()->pool_holder(); |
aoqi@0 | 901 | __ ld(R11_scratch1/*pool_holder*/, ConstantPool::pool_holder_offset_in_bytes(), |
aoqi@0 | 902 | R11_scratch1/*constants*/); |
aoqi@0 | 903 | |
aoqi@0 | 904 | const int mirror_offset = in_bytes(Klass::java_mirror_offset()); |
aoqi@0 | 905 | |
aoqi@0 | 906 | // mirror = pool_holder->klass_part()->java_mirror(); |
aoqi@0 | 907 | __ ld(R0/*mirror*/, mirror_offset, R11_scratch1/*pool_holder*/); |
aoqi@0 | 908 | // state->_native_mirror = mirror; |
aoqi@0 | 909 | |
aoqi@0 | 910 | __ ld(R11_scratch1, 0, R1_SP); |
aoqi@0 | 911 | __ std(R0/*mirror*/, _ijava_state_neg(oop_tmp), R11_scratch1); |
aoqi@0 | 912 | // R4_ARG2 = &state->_oop_temp; |
aoqi@0 | 913 | __ addi(R4_ARG2, R11_scratch1, _ijava_state_neg(oop_tmp)); |
aoqi@0 | 914 | __ BIND(method_is_not_static); |
aoqi@0 | 915 | } |
aoqi@0 | 916 | |
aoqi@0 | 917 | // At this point, arguments have been copied off the stack into |
aoqi@0 | 918 | // their JNI positions. Oops are boxed in-place on the stack, with |
aoqi@0 | 919 | // handles copied to arguments. The result handler address is in a |
aoqi@0 | 920 | // register. |
aoqi@0 | 921 | |
aoqi@0 | 922 | // Pass JNIEnv address as first parameter. |
aoqi@0 | 923 | __ addir(R3_ARG1, thread_(jni_environment)); |
aoqi@0 | 924 | |
aoqi@0 | 925 | // Load the native_method entry before we change the thread state. |
aoqi@0 | 926 | __ ld(native_method_fd, method_(native_function)); |
aoqi@0 | 927 | |
aoqi@0 | 928 | //============================================================================= |
aoqi@0 | 929 | // Transition from _thread_in_Java to _thread_in_native. As soon as |
aoqi@0 | 930 | // we make this change the safepoint code needs to be certain that |
aoqi@0 | 931 | // the last Java frame we established is good. The pc in that frame |
aoqi@0 | 932 | // just needs to be near here not an actual return address. |
aoqi@0 | 933 | |
aoqi@0 | 934 | // We use release_store_fence to update values like the thread state, where |
aoqi@0 | 935 | // we don't want the current thread to continue until all our prior memory |
aoqi@0 | 936 | // accesses (including the new thread state) are visible to other threads. |
aoqi@0 | 937 | __ li(R0, _thread_in_native); |
aoqi@0 | 938 | __ release(); |
aoqi@0 | 939 | |
aoqi@0 | 940 | // TODO PPC port assert(4 == JavaThread::sz_thread_state(), "unexpected field size"); |
aoqi@0 | 941 | __ stw(R0, thread_(thread_state)); |
aoqi@0 | 942 | |
aoqi@0 | 943 | if (UseMembar) { |
aoqi@0 | 944 | __ fence(); |
aoqi@0 | 945 | } |
aoqi@0 | 946 | |
aoqi@0 | 947 | //============================================================================= |
aoqi@0 | 948 | // Call the native method. Argument registers must not have been |
aoqi@0 | 949 | // overwritten since "__ call_stub(signature_handler);" (except for |
aoqi@0 | 950 | // ARG1 and ARG2 for static methods). |
aoqi@0 | 951 | __ call_c(native_method_fd); |
aoqi@0 | 952 | |
aoqi@0 | 953 | __ li(R0, 0); |
aoqi@0 | 954 | __ ld(R11_scratch1, 0, R1_SP); |
aoqi@0 | 955 | __ std(R3_RET, _ijava_state_neg(lresult), R11_scratch1); |
aoqi@0 | 956 | __ stfd(F1_RET, _ijava_state_neg(fresult), R11_scratch1); |
aoqi@0 | 957 | __ std(R0/*mirror*/, _ijava_state_neg(oop_tmp), R11_scratch1); // reset |
aoqi@0 | 958 | |
aoqi@0 | 959 | // Note: C++ interpreter needs the following here: |
aoqi@0 | 960 | // The frame_manager_lr field, which we use for setting the last |
aoqi@0 | 961 | // java frame, gets overwritten by the signature handler. Restore |
aoqi@0 | 962 | // it now. |
aoqi@0 | 963 | //__ get_PC_trash_LR(R11_scratch1); |
aoqi@0 | 964 | //__ std(R11_scratch1, _top_ijava_frame_abi(frame_manager_lr), R1_SP); |
aoqi@0 | 965 | |
aoqi@0 | 966 | // Because of GC R19_method may no longer be valid. |
aoqi@0 | 967 | |
aoqi@0 | 968 | // Block, if necessary, before resuming in _thread_in_Java state. |
aoqi@0 | 969 | // In order for GC to work, don't clear the last_Java_sp until after |
aoqi@0 | 970 | // blocking. |
aoqi@0 | 971 | |
aoqi@0 | 972 | //============================================================================= |
aoqi@0 | 973 | // Switch thread to "native transition" state before reading the |
aoqi@0 | 974 | // synchronization state. This additional state is necessary |
aoqi@0 | 975 | // because reading and testing the synchronization state is not |
aoqi@0 | 976 | // atomic w.r.t. GC, as this scenario demonstrates: Java thread A, |
aoqi@0 | 977 | // in _thread_in_native state, loads _not_synchronized and is |
aoqi@0 | 978 | // preempted. VM thread changes sync state to synchronizing and |
aoqi@0 | 979 | // suspends threads for GC. Thread A is resumed to finish this |
aoqi@0 | 980 | // native method, but doesn't block here since it didn't see any |
aoqi@0 | 981 | // synchronization in progress, and escapes. |
aoqi@0 | 982 | |
aoqi@0 | 983 | // We use release_store_fence to update values like the thread state, where |
aoqi@0 | 984 | // we don't want the current thread to continue until all our prior memory |
aoqi@0 | 985 | // accesses (including the new thread state) are visible to other threads. |
aoqi@0 | 986 | __ li(R0/*thread_state*/, _thread_in_native_trans); |
aoqi@0 | 987 | __ release(); |
aoqi@0 | 988 | __ stw(R0/*thread_state*/, thread_(thread_state)); |
aoqi@0 | 989 | if (UseMembar) { |
aoqi@0 | 990 | __ fence(); |
aoqi@0 | 991 | } |
aoqi@0 | 992 | // Write serialization page so that the VM thread can do a pseudo remote |
aoqi@0 | 993 | // membar. We use the current thread pointer to calculate a thread |
aoqi@0 | 994 | // specific offset to write to within the page. This minimizes bus |
aoqi@0 | 995 | // traffic due to cache line collision. |
aoqi@0 | 996 | else { |
aoqi@0 | 997 | __ serialize_memory(R16_thread, R11_scratch1, R12_scratch2); |
aoqi@0 | 998 | } |
aoqi@0 | 999 | |
aoqi@0 | 1000 | // Now before we return to java we must look for a current safepoint |
aoqi@0 | 1001 | // (a new safepoint can not start since we entered native_trans). |
aoqi@0 | 1002 | // We must check here because a current safepoint could be modifying |
aoqi@0 | 1003 | // the callers registers right this moment. |
aoqi@0 | 1004 | |
aoqi@0 | 1005 | // Acquire isn't strictly necessary here because of the fence, but |
aoqi@0 | 1006 | // sync_state is declared to be volatile, so we do it anyway |
aoqi@0 | 1007 | // (cmp-br-isync on one path, release (same as acquire on PPC64) on the other path). |
aoqi@0 | 1008 | int sync_state_offs = __ load_const_optimized(sync_state_addr, SafepointSynchronize::address_of_state(), /*temp*/R0, true); |
aoqi@0 | 1009 | |
aoqi@0 | 1010 | // TODO PPC port assert(4 == SafepointSynchronize::sz_state(), "unexpected field size"); |
aoqi@0 | 1011 | __ lwz(sync_state, sync_state_offs, sync_state_addr); |
aoqi@0 | 1012 | |
aoqi@0 | 1013 | // TODO PPC port assert(4 == Thread::sz_suspend_flags(), "unexpected field size"); |
aoqi@0 | 1014 | __ lwz(suspend_flags, thread_(suspend_flags)); |
aoqi@0 | 1015 | |
aoqi@0 | 1016 | Label sync_check_done; |
aoqi@0 | 1017 | Label do_safepoint; |
aoqi@0 | 1018 | // No synchronization in progress nor yet synchronized. |
aoqi@0 | 1019 | __ cmpwi(CCR0, sync_state, SafepointSynchronize::_not_synchronized); |
aoqi@0 | 1020 | // Not suspended. |
aoqi@0 | 1021 | __ cmpwi(CCR1, suspend_flags, 0); |
aoqi@0 | 1022 | |
aoqi@0 | 1023 | __ bne(CCR0, do_safepoint); |
aoqi@0 | 1024 | __ beq(CCR1, sync_check_done); |
aoqi@0 | 1025 | __ bind(do_safepoint); |
aoqi@0 | 1026 | __ isync(); |
aoqi@0 | 1027 | // Block. We do the call directly and leave the current |
aoqi@0 | 1028 | // last_Java_frame setup undisturbed. We must save any possible |
aoqi@0 | 1029 | // native result across the call. No oop is present. |
aoqi@0 | 1030 | |
aoqi@0 | 1031 | __ mr(R3_ARG1, R16_thread); |
kvn@7132 | 1032 | #if defined(ABI_ELFv2) |
kvn@7132 | 1033 | __ call_c(CAST_FROM_FN_PTR(address, JavaThread::check_special_condition_for_native_trans), |
kvn@7132 | 1034 | relocInfo::none); |
kvn@7132 | 1035 | #else |
aoqi@0 | 1036 | __ call_c(CAST_FROM_FN_PTR(FunctionDescriptor*, JavaThread::check_special_condition_for_native_trans), |
aoqi@0 | 1037 | relocInfo::none); |
kvn@7132 | 1038 | #endif |
aoqi@0 | 1039 | |
aoqi@0 | 1040 | __ bind(sync_check_done); |
aoqi@0 | 1041 | |
aoqi@0 | 1042 | //============================================================================= |
aoqi@0 | 1043 | // <<<<<< Back in Interpreter Frame >>>>> |
aoqi@0 | 1044 | |
aoqi@0 | 1045 | // We are in thread_in_native_trans here and back in the normal |
aoqi@0 | 1046 | // interpreter frame. We don't have to do anything special about |
aoqi@0 | 1047 | // safepoints and we can switch to Java mode anytime we are ready. |
aoqi@0 | 1048 | |
aoqi@0 | 1049 | // Note: frame::interpreter_frame_result has a dependency on how the |
aoqi@0 | 1050 | // method result is saved across the call to post_method_exit. For |
aoqi@0 | 1051 | // native methods it assumes that the non-FPU/non-void result is |
aoqi@0 | 1052 | // saved in _native_lresult and a FPU result in _native_fresult. If |
aoqi@0 | 1053 | // this changes then the interpreter_frame_result implementation |
aoqi@0 | 1054 | // will need to be updated too. |
aoqi@0 | 1055 | |
aoqi@0 | 1056 | // On PPC64, we have stored the result directly after the native call. |
aoqi@0 | 1057 | |
aoqi@0 | 1058 | //============================================================================= |
aoqi@0 | 1059 | // Back in Java |
aoqi@0 | 1060 | |
aoqi@0 | 1061 | // We use release_store_fence to update values like the thread state, where |
aoqi@0 | 1062 | // we don't want the current thread to continue until all our prior memory |
aoqi@0 | 1063 | // accesses (including the new thread state) are visible to other threads. |
aoqi@0 | 1064 | __ li(R0/*thread_state*/, _thread_in_Java); |
aoqi@0 | 1065 | __ release(); |
aoqi@0 | 1066 | __ stw(R0/*thread_state*/, thread_(thread_state)); |
aoqi@0 | 1067 | if (UseMembar) { |
aoqi@0 | 1068 | __ fence(); |
aoqi@0 | 1069 | } |
aoqi@0 | 1070 | |
aoqi@0 | 1071 | __ reset_last_Java_frame(); |
aoqi@0 | 1072 | |
aoqi@0 | 1073 | // Jvmdi/jvmpi support. Whether we've got an exception pending or |
aoqi@0 | 1074 | // not, and whether unlocking throws an exception or not, we notify |
aoqi@0 | 1075 | // on native method exit. If we do have an exception, we'll end up |
aoqi@0 | 1076 | // in the caller's context to handle it, so if we don't do the |
aoqi@0 | 1077 | // notify here, we'll drop it on the floor. |
aoqi@0 | 1078 | __ notify_method_exit(true/*native method*/, |
aoqi@0 | 1079 | ilgl /*illegal state (not used for native methods)*/, |
aoqi@0 | 1080 | InterpreterMacroAssembler::NotifyJVMTI, |
aoqi@0 | 1081 | false /*check_exceptions*/); |
aoqi@0 | 1082 | |
aoqi@0 | 1083 | //============================================================================= |
aoqi@0 | 1084 | // Handle exceptions |
aoqi@0 | 1085 | |
aoqi@0 | 1086 | if (synchronized) { |
aoqi@0 | 1087 | // Don't check for exceptions since we're still in the i2n frame. Do that |
aoqi@0 | 1088 | // manually afterwards. |
aoqi@0 | 1089 | unlock_method(false); |
aoqi@0 | 1090 | } |
aoqi@0 | 1091 | |
aoqi@0 | 1092 | // Reset active handles after returning from native. |
aoqi@0 | 1093 | // thread->active_handles()->clear(); |
aoqi@0 | 1094 | __ ld(active_handles, thread_(active_handles)); |
aoqi@0 | 1095 | // TODO PPC port assert(4 == JNIHandleBlock::top_size_in_bytes(), "unexpected field size"); |
aoqi@0 | 1096 | __ li(R0, 0); |
aoqi@0 | 1097 | __ stw(R0, JNIHandleBlock::top_offset_in_bytes(), active_handles); |
aoqi@0 | 1098 | |
aoqi@0 | 1099 | Label exception_return_sync_check_already_unlocked; |
aoqi@0 | 1100 | __ ld(R0/*pending_exception*/, thread_(pending_exception)); |
aoqi@0 | 1101 | __ cmpdi(CCR0, R0/*pending_exception*/, 0); |
aoqi@0 | 1102 | __ bne(CCR0, exception_return_sync_check_already_unlocked); |
aoqi@0 | 1103 | |
aoqi@0 | 1104 | //----------------------------------------------------------------------------- |
aoqi@0 | 1105 | // No exception pending. |
aoqi@0 | 1106 | |
aoqi@0 | 1107 | // Move native method result back into proper registers and return. |
aoqi@0 | 1108 | // Invoke result handler (may unbox/promote). |
aoqi@0 | 1109 | __ ld(R11_scratch1, 0, R1_SP); |
aoqi@0 | 1110 | __ ld(R3_RET, _ijava_state_neg(lresult), R11_scratch1); |
aoqi@0 | 1111 | __ lfd(F1_RET, _ijava_state_neg(fresult), R11_scratch1); |
aoqi@0 | 1112 | __ call_stub(result_handler_addr); |
aoqi@0 | 1113 | |
aoqi@0 | 1114 | __ merge_frames(/*top_frame_sp*/ R21_sender_SP, /*return_pc*/ R0, R11_scratch1, R12_scratch2); |
aoqi@0 | 1115 | |
aoqi@0 | 1116 | // Must use the return pc which was loaded from the caller's frame |
aoqi@0 | 1117 | // as the VM uses return-pc-patching for deoptimization. |
aoqi@0 | 1118 | __ mtlr(R0); |
aoqi@0 | 1119 | __ blr(); |
aoqi@0 | 1120 | |
aoqi@0 | 1121 | //----------------------------------------------------------------------------- |
aoqi@0 | 1122 | // An exception is pending. We call into the runtime only if the |
aoqi@0 | 1123 | // caller was not interpreted. If it was interpreted the |
aoqi@0 | 1124 | // interpreter will do the correct thing. If it isn't interpreted |
aoqi@0 | 1125 | // (call stub/compiled code) we will change our return and continue. |
aoqi@0 | 1126 | |
aoqi@0 | 1127 | __ BIND(exception_return_sync_check); |
aoqi@0 | 1128 | |
aoqi@0 | 1129 | if (synchronized) { |
aoqi@0 | 1130 | // Don't check for exceptions since we're still in the i2n frame. Do that |
aoqi@0 | 1131 | // manually afterwards. |
aoqi@0 | 1132 | unlock_method(false); |
aoqi@0 | 1133 | } |
aoqi@0 | 1134 | __ BIND(exception_return_sync_check_already_unlocked); |
aoqi@0 | 1135 | |
aoqi@0 | 1136 | const Register return_pc = R31; |
aoqi@0 | 1137 | |
aoqi@0 | 1138 | __ ld(return_pc, 0, R1_SP); |
aoqi@0 | 1139 | __ ld(return_pc, _abi(lr), return_pc); |
aoqi@0 | 1140 | |
aoqi@0 | 1141 | // Get the address of the exception handler. |
aoqi@0 | 1142 | __ call_VM_leaf(CAST_FROM_FN_PTR(address, SharedRuntime::exception_handler_for_return_address), |
aoqi@0 | 1143 | R16_thread, |
aoqi@0 | 1144 | return_pc /* return pc */); |
aoqi@0 | 1145 | __ merge_frames(/*top_frame_sp*/ R21_sender_SP, noreg, R11_scratch1, R12_scratch2); |
aoqi@0 | 1146 | |
aoqi@0 | 1147 | // Load the PC of the the exception handler into LR. |
aoqi@0 | 1148 | __ mtlr(R3_RET); |
aoqi@0 | 1149 | |
aoqi@0 | 1150 | // Load exception into R3_ARG1 and clear pending exception in thread. |
aoqi@0 | 1151 | __ ld(R3_ARG1/*exception*/, thread_(pending_exception)); |
aoqi@0 | 1152 | __ li(R4_ARG2, 0); |
aoqi@0 | 1153 | __ std(R4_ARG2, thread_(pending_exception)); |
aoqi@0 | 1154 | |
aoqi@0 | 1155 | // Load the original return pc into R4_ARG2. |
aoqi@0 | 1156 | __ mr(R4_ARG2/*issuing_pc*/, return_pc); |
aoqi@0 | 1157 | |
aoqi@0 | 1158 | // Return to exception handler. |
aoqi@0 | 1159 | __ blr(); |
aoqi@0 | 1160 | |
aoqi@0 | 1161 | //============================================================================= |
aoqi@0 | 1162 | // Counter overflow. |
aoqi@0 | 1163 | |
aoqi@0 | 1164 | if (inc_counter) { |
aoqi@0 | 1165 | // Handle invocation counter overflow. |
aoqi@0 | 1166 | __ bind(invocation_counter_overflow); |
aoqi@0 | 1167 | |
aoqi@0 | 1168 | generate_counter_overflow(continue_after_compile); |
aoqi@0 | 1169 | } |
aoqi@0 | 1170 | |
aoqi@0 | 1171 | return entry; |
aoqi@0 | 1172 | } |
aoqi@0 | 1173 | |
aoqi@0 | 1174 | // Generic interpreted method entry to (asm) interpreter. |
aoqi@0 | 1175 | // |
aoqi@0 | 1176 | address TemplateInterpreterGenerator::generate_normal_entry(bool synchronized) { |
aoqi@0 | 1177 | bool inc_counter = UseCompiler || CountCompiledCalls; |
aoqi@0 | 1178 | address entry = __ pc(); |
aoqi@0 | 1179 | // Generate the code to allocate the interpreter stack frame. |
aoqi@0 | 1180 | Register Rsize_of_parameters = R4_ARG2, // Written by generate_fixed_frame. |
aoqi@0 | 1181 | Rsize_of_locals = R5_ARG3; // Written by generate_fixed_frame. |
aoqi@0 | 1182 | |
aoqi@0 | 1183 | generate_fixed_frame(false, Rsize_of_parameters, Rsize_of_locals); |
aoqi@0 | 1184 | |
aoqi@0 | 1185 | #ifdef FAST_DISPATCH |
aoqi@0 | 1186 | __ unimplemented("Fast dispatch in generate_normal_entry"); |
aoqi@0 | 1187 | #if 0 |
aoqi@0 | 1188 | __ set((intptr_t)Interpreter::dispatch_table(), IdispatchTables); |
aoqi@0 | 1189 | // Set bytecode dispatch table base. |
aoqi@0 | 1190 | #endif |
aoqi@0 | 1191 | #endif |
aoqi@0 | 1192 | |
aoqi@0 | 1193 | // -------------------------------------------------------------------------- |
aoqi@0 | 1194 | // Zero out non-parameter locals. |
aoqi@0 | 1195 | // Note: *Always* zero out non-parameter locals as Sparc does. It's not |
aoqi@0 | 1196 | // worth to ask the flag, just do it. |
aoqi@0 | 1197 | Register Rslot_addr = R6_ARG4, |
aoqi@0 | 1198 | Rnum = R7_ARG5; |
aoqi@0 | 1199 | Label Lno_locals, Lzero_loop; |
aoqi@0 | 1200 | |
aoqi@0 | 1201 | // Set up the zeroing loop. |
aoqi@0 | 1202 | __ subf(Rnum, Rsize_of_parameters, Rsize_of_locals); |
aoqi@0 | 1203 | __ subf(Rslot_addr, Rsize_of_parameters, R18_locals); |
aoqi@0 | 1204 | __ srdi_(Rnum, Rnum, Interpreter::logStackElementSize); |
aoqi@0 | 1205 | __ beq(CCR0, Lno_locals); |
aoqi@0 | 1206 | __ li(R0, 0); |
aoqi@0 | 1207 | __ mtctr(Rnum); |
aoqi@0 | 1208 | |
aoqi@0 | 1209 | // The zero locals loop. |
aoqi@0 | 1210 | __ bind(Lzero_loop); |
aoqi@0 | 1211 | __ std(R0, 0, Rslot_addr); |
aoqi@0 | 1212 | __ addi(Rslot_addr, Rslot_addr, -Interpreter::stackElementSize); |
aoqi@0 | 1213 | __ bdnz(Lzero_loop); |
aoqi@0 | 1214 | |
aoqi@0 | 1215 | __ bind(Lno_locals); |
aoqi@0 | 1216 | |
aoqi@0 | 1217 | // -------------------------------------------------------------------------- |
aoqi@0 | 1218 | // Counter increment and overflow check. |
aoqi@0 | 1219 | Label invocation_counter_overflow, |
aoqi@0 | 1220 | profile_method, |
aoqi@0 | 1221 | profile_method_continue; |
aoqi@0 | 1222 | if (inc_counter || ProfileInterpreter) { |
aoqi@0 | 1223 | |
aoqi@0 | 1224 | Register Rdo_not_unlock_if_synchronized_addr = R11_scratch1; |
aoqi@0 | 1225 | if (synchronized) { |
aoqi@0 | 1226 | // Since at this point in the method invocation the exception handler |
aoqi@0 | 1227 | // would try to exit the monitor of synchronized methods which hasn't |
aoqi@0 | 1228 | // been entered yet, we set the thread local variable |
aoqi@0 | 1229 | // _do_not_unlock_if_synchronized to true. If any exception was thrown by |
aoqi@0 | 1230 | // runtime, exception handling i.e. unlock_if_synchronized_method will |
aoqi@0 | 1231 | // check this thread local flag. |
aoqi@0 | 1232 | // This flag has two effects, one is to force an unwind in the topmost |
aoqi@0 | 1233 | // interpreter frame and not perform an unlock while doing so. |
aoqi@0 | 1234 | __ li(R0, 1); |
aoqi@0 | 1235 | __ stb(R0, in_bytes(JavaThread::do_not_unlock_if_synchronized_offset()), R16_thread); |
aoqi@0 | 1236 | } |
goetz@7424 | 1237 | |
goetz@7424 | 1238 | // Argument and return type profiling. |
goetz@7424 | 1239 | __ profile_parameters_type(R3_ARG1, R4_ARG2, R5_ARG3, R6_ARG4); |
goetz@7424 | 1240 | |
aoqi@0 | 1241 | // Increment invocation counter and check for overflow. |
aoqi@0 | 1242 | if (inc_counter) { |
aoqi@0 | 1243 | generate_counter_incr(&invocation_counter_overflow, &profile_method, &profile_method_continue); |
aoqi@0 | 1244 | } |
aoqi@0 | 1245 | |
aoqi@0 | 1246 | __ bind(profile_method_continue); |
aoqi@0 | 1247 | |
aoqi@0 | 1248 | // Reset the _do_not_unlock_if_synchronized flag. |
aoqi@0 | 1249 | if (synchronized) { |
aoqi@0 | 1250 | __ li(R0, 0); |
aoqi@0 | 1251 | __ stb(R0, in_bytes(JavaThread::do_not_unlock_if_synchronized_offset()), R16_thread); |
aoqi@0 | 1252 | } |
aoqi@0 | 1253 | } |
aoqi@0 | 1254 | |
aoqi@0 | 1255 | // -------------------------------------------------------------------------- |
aoqi@0 | 1256 | // Locking of synchronized methods. Must happen AFTER invocation_counter |
aoqi@0 | 1257 | // check and stack overflow check, so method is not locked if overflows. |
aoqi@0 | 1258 | if (synchronized) { |
aoqi@0 | 1259 | lock_method(R3_ARG1, R4_ARG2, R5_ARG3); |
aoqi@0 | 1260 | } |
aoqi@0 | 1261 | #ifdef ASSERT |
aoqi@0 | 1262 | else { |
aoqi@0 | 1263 | Label Lok; |
aoqi@0 | 1264 | __ lwz(R0, in_bytes(Method::access_flags_offset()), R19_method); |
aoqi@0 | 1265 | __ andi_(R0, R0, JVM_ACC_SYNCHRONIZED); |
aoqi@0 | 1266 | __ asm_assert_eq("method needs synchronization", 0x8521); |
aoqi@0 | 1267 | __ bind(Lok); |
aoqi@0 | 1268 | } |
aoqi@0 | 1269 | #endif // ASSERT |
aoqi@0 | 1270 | |
aoqi@0 | 1271 | __ verify_thread(); |
aoqi@0 | 1272 | |
aoqi@0 | 1273 | // -------------------------------------------------------------------------- |
aoqi@0 | 1274 | // JVMTI support |
aoqi@0 | 1275 | __ notify_method_entry(); |
aoqi@0 | 1276 | |
aoqi@0 | 1277 | // -------------------------------------------------------------------------- |
aoqi@0 | 1278 | // Start executing instructions. |
aoqi@0 | 1279 | __ dispatch_next(vtos); |
aoqi@0 | 1280 | |
aoqi@0 | 1281 | // -------------------------------------------------------------------------- |
aoqi@0 | 1282 | // Out of line counter overflow and MDO creation code. |
aoqi@0 | 1283 | if (ProfileInterpreter) { |
aoqi@0 | 1284 | // We have decided to profile this method in the interpreter. |
aoqi@0 | 1285 | __ bind(profile_method); |
aoqi@0 | 1286 | __ call_VM(noreg, CAST_FROM_FN_PTR(address, InterpreterRuntime::profile_method)); |
aoqi@0 | 1287 | __ set_method_data_pointer_for_bcp(); |
aoqi@0 | 1288 | __ b(profile_method_continue); |
aoqi@0 | 1289 | } |
aoqi@0 | 1290 | |
aoqi@0 | 1291 | if (inc_counter) { |
aoqi@0 | 1292 | // Handle invocation counter overflow. |
aoqi@0 | 1293 | __ bind(invocation_counter_overflow); |
aoqi@0 | 1294 | generate_counter_overflow(profile_method_continue); |
aoqi@0 | 1295 | } |
aoqi@0 | 1296 | return entry; |
aoqi@0 | 1297 | } |
aoqi@0 | 1298 | |
aoqi@0 | 1299 | // ============================================================================= |
aoqi@0 | 1300 | // Entry points |
aoqi@0 | 1301 | |
aoqi@0 | 1302 | address AbstractInterpreterGenerator::generate_method_entry( |
aoqi@0 | 1303 | AbstractInterpreter::MethodKind kind) { |
aoqi@0 | 1304 | // Determine code generation flags. |
aoqi@0 | 1305 | bool synchronized = false; |
aoqi@0 | 1306 | address entry_point = NULL; |
aoqi@0 | 1307 | |
aoqi@0 | 1308 | switch (kind) { |
aoqi@0 | 1309 | case Interpreter::zerolocals : break; |
aoqi@0 | 1310 | case Interpreter::zerolocals_synchronized: synchronized = true; break; |
aoqi@0 | 1311 | case Interpreter::native : entry_point = ((InterpreterGenerator*) this)->generate_native_entry(false); break; |
aoqi@0 | 1312 | case Interpreter::native_synchronized : entry_point = ((InterpreterGenerator*) this)->generate_native_entry(true); break; |
aoqi@0 | 1313 | case Interpreter::empty : entry_point = ((InterpreterGenerator*) this)->generate_empty_entry(); break; |
aoqi@0 | 1314 | case Interpreter::accessor : entry_point = ((InterpreterGenerator*) this)->generate_accessor_entry(); break; |
aoqi@0 | 1315 | case Interpreter::abstract : entry_point = ((InterpreterGenerator*) this)->generate_abstract_entry(); break; |
aoqi@0 | 1316 | |
aoqi@0 | 1317 | case Interpreter::java_lang_math_sin : // fall thru |
aoqi@0 | 1318 | case Interpreter::java_lang_math_cos : // fall thru |
aoqi@0 | 1319 | case Interpreter::java_lang_math_tan : // fall thru |
aoqi@0 | 1320 | case Interpreter::java_lang_math_abs : // fall thru |
aoqi@0 | 1321 | case Interpreter::java_lang_math_log : // fall thru |
aoqi@0 | 1322 | case Interpreter::java_lang_math_log10 : // fall thru |
aoqi@0 | 1323 | case Interpreter::java_lang_math_sqrt : // fall thru |
aoqi@0 | 1324 | case Interpreter::java_lang_math_pow : // fall thru |
aoqi@0 | 1325 | case Interpreter::java_lang_math_exp : entry_point = ((InterpreterGenerator*) this)->generate_math_entry(kind); break; |
aoqi@0 | 1326 | case Interpreter::java_lang_ref_reference_get |
aoqi@0 | 1327 | : entry_point = ((InterpreterGenerator*)this)->generate_Reference_get_entry(); break; |
aoqi@0 | 1328 | default : ShouldNotReachHere(); break; |
aoqi@0 | 1329 | } |
aoqi@0 | 1330 | |
aoqi@0 | 1331 | if (entry_point) { |
aoqi@0 | 1332 | return entry_point; |
aoqi@0 | 1333 | } |
aoqi@0 | 1334 | |
aoqi@0 | 1335 | return ((InterpreterGenerator*) this)->generate_normal_entry(synchronized); |
aoqi@0 | 1336 | } |
aoqi@0 | 1337 | |
aoqi@0 | 1338 | // These should never be compiled since the interpreter will prefer |
aoqi@0 | 1339 | // the compiled version to the intrinsic version. |
aoqi@0 | 1340 | bool AbstractInterpreter::can_be_compiled(methodHandle m) { |
aoqi@0 | 1341 | return !math_entry_available(method_kind(m)); |
aoqi@0 | 1342 | } |
aoqi@0 | 1343 | |
aoqi@0 | 1344 | // How much stack a method activation needs in stack slots. |
aoqi@0 | 1345 | // We must calc this exactly like in generate_fixed_frame. |
aoqi@0 | 1346 | // Note: This returns the conservative size assuming maximum alignment. |
aoqi@0 | 1347 | int AbstractInterpreter::size_top_interpreter_activation(Method* method) { |
aoqi@0 | 1348 | const int max_alignment_size = 2; |
aoqi@0 | 1349 | const int abi_scratch = frame::abi_reg_args_size; |
aoqi@0 | 1350 | return method->max_locals() + method->max_stack() + |
aoqi@0 | 1351 | frame::interpreter_frame_monitor_size() + max_alignment_size + abi_scratch; |
aoqi@0 | 1352 | } |
aoqi@0 | 1353 | |
aoqi@0 | 1354 | // Returns number of stackElementWords needed for the interpreter frame with the |
aoqi@0 | 1355 | // given sections. |
aoqi@0 | 1356 | // This overestimates the stack by one slot in case of alignments. |
aoqi@0 | 1357 | int AbstractInterpreter::size_activation(int max_stack, |
aoqi@0 | 1358 | int temps, |
aoqi@0 | 1359 | int extra_args, |
aoqi@0 | 1360 | int monitors, |
aoqi@0 | 1361 | int callee_params, |
aoqi@0 | 1362 | int callee_locals, |
aoqi@0 | 1363 | bool is_top_frame) { |
aoqi@0 | 1364 | // Note: This calculation must exactly parallel the frame setup |
aoqi@0 | 1365 | // in AbstractInterpreterGenerator::generate_method_entry. |
aoqi@0 | 1366 | assert(Interpreter::stackElementWords == 1, "sanity"); |
aoqi@0 | 1367 | const int max_alignment_space = StackAlignmentInBytes / Interpreter::stackElementSize; |
aoqi@0 | 1368 | const int abi_scratch = is_top_frame ? (frame::abi_reg_args_size / Interpreter::stackElementSize) : |
aoqi@0 | 1369 | (frame::abi_minframe_size / Interpreter::stackElementSize); |
aoqi@0 | 1370 | const int size = |
aoqi@0 | 1371 | max_stack + |
aoqi@0 | 1372 | (callee_locals - callee_params) + |
aoqi@0 | 1373 | monitors * frame::interpreter_frame_monitor_size() + |
aoqi@0 | 1374 | max_alignment_space + |
aoqi@0 | 1375 | abi_scratch + |
aoqi@0 | 1376 | frame::ijava_state_size / Interpreter::stackElementSize; |
aoqi@0 | 1377 | |
aoqi@0 | 1378 | // Fixed size of an interpreter frame, align to 16-byte. |
aoqi@0 | 1379 | return (size & -2); |
aoqi@0 | 1380 | } |
aoqi@0 | 1381 | |
aoqi@0 | 1382 | // Fills a sceletal interpreter frame generated during deoptimizations. |
aoqi@0 | 1383 | // |
aoqi@0 | 1384 | // Parameters: |
aoqi@0 | 1385 | // |
aoqi@0 | 1386 | // interpreter_frame != NULL: |
aoqi@0 | 1387 | // set up the method, locals, and monitors. |
aoqi@0 | 1388 | // The frame interpreter_frame, if not NULL, is guaranteed to be the |
aoqi@0 | 1389 | // right size, as determined by a previous call to this method. |
aoqi@0 | 1390 | // It is also guaranteed to be walkable even though it is in a skeletal state |
aoqi@0 | 1391 | // |
aoqi@0 | 1392 | // is_top_frame == true: |
aoqi@0 | 1393 | // We're processing the *oldest* interpreter frame! |
aoqi@0 | 1394 | // |
aoqi@0 | 1395 | // pop_frame_extra_args: |
aoqi@0 | 1396 | // If this is != 0 we are returning to a deoptimized frame by popping |
aoqi@0 | 1397 | // off the callee frame. We want to re-execute the call that called the |
aoqi@0 | 1398 | // callee interpreted, but since the return to the interpreter would pop |
aoqi@0 | 1399 | // the arguments off advance the esp by dummy popframe_extra_args slots. |
aoqi@0 | 1400 | // Popping off those will establish the stack layout as it was before the call. |
aoqi@0 | 1401 | // |
aoqi@0 | 1402 | void AbstractInterpreter::layout_activation(Method* method, |
aoqi@0 | 1403 | int tempcount, |
aoqi@0 | 1404 | int popframe_extra_args, |
aoqi@0 | 1405 | int moncount, |
aoqi@0 | 1406 | int caller_actual_parameters, |
aoqi@0 | 1407 | int callee_param_count, |
aoqi@0 | 1408 | int callee_locals_count, |
aoqi@0 | 1409 | frame* caller, |
aoqi@0 | 1410 | frame* interpreter_frame, |
aoqi@0 | 1411 | bool is_top_frame, |
aoqi@0 | 1412 | bool is_bottom_frame) { |
aoqi@0 | 1413 | |
aoqi@0 | 1414 | const int abi_scratch = is_top_frame ? (frame::abi_reg_args_size / Interpreter::stackElementSize) : |
aoqi@0 | 1415 | (frame::abi_minframe_size / Interpreter::stackElementSize); |
aoqi@0 | 1416 | |
aoqi@0 | 1417 | intptr_t* locals_base = (caller->is_interpreted_frame()) ? |
aoqi@0 | 1418 | caller->interpreter_frame_esp() + caller_actual_parameters : |
aoqi@0 | 1419 | caller->sp() + method->max_locals() - 1 + (frame::abi_minframe_size / Interpreter::stackElementSize) ; |
aoqi@0 | 1420 | |
aoqi@0 | 1421 | intptr_t* monitor_base = caller->sp() - frame::ijava_state_size / Interpreter::stackElementSize ; |
aoqi@0 | 1422 | intptr_t* monitor = monitor_base - (moncount * frame::interpreter_frame_monitor_size()); |
aoqi@0 | 1423 | intptr_t* esp_base = monitor - 1; |
aoqi@0 | 1424 | intptr_t* esp = esp_base - tempcount - popframe_extra_args; |
aoqi@0 | 1425 | intptr_t* sp = (intptr_t *) (((intptr_t) (esp_base - callee_locals_count + callee_param_count - method->max_stack()- abi_scratch)) & -StackAlignmentInBytes); |
aoqi@0 | 1426 | intptr_t* sender_sp = caller->sp() + (frame::abi_minframe_size - frame::abi_reg_args_size) / Interpreter::stackElementSize; |
aoqi@0 | 1427 | intptr_t* top_frame_sp = is_top_frame ? sp : sp + (frame::abi_minframe_size - frame::abi_reg_args_size) / Interpreter::stackElementSize; |
aoqi@0 | 1428 | |
aoqi@0 | 1429 | interpreter_frame->interpreter_frame_set_method(method); |
aoqi@0 | 1430 | interpreter_frame->interpreter_frame_set_locals(locals_base); |
aoqi@0 | 1431 | interpreter_frame->interpreter_frame_set_cpcache(method->constants()->cache()); |
aoqi@0 | 1432 | interpreter_frame->interpreter_frame_set_esp(esp); |
aoqi@0 | 1433 | interpreter_frame->interpreter_frame_set_monitor_end((BasicObjectLock *)monitor); |
aoqi@0 | 1434 | interpreter_frame->interpreter_frame_set_top_frame_sp(top_frame_sp); |
aoqi@0 | 1435 | if (!is_bottom_frame) { |
aoqi@0 | 1436 | interpreter_frame->interpreter_frame_set_sender_sp(sender_sp); |
aoqi@0 | 1437 | } |
aoqi@0 | 1438 | } |
aoqi@0 | 1439 | |
aoqi@0 | 1440 | // ============================================================================= |
aoqi@0 | 1441 | // Exceptions |
aoqi@0 | 1442 | |
aoqi@0 | 1443 | void TemplateInterpreterGenerator::generate_throw_exception() { |
aoqi@0 | 1444 | Register Rexception = R17_tos, |
aoqi@0 | 1445 | Rcontinuation = R3_RET; |
aoqi@0 | 1446 | |
aoqi@0 | 1447 | // -------------------------------------------------------------------------- |
aoqi@0 | 1448 | // Entry point if an method returns with a pending exception (rethrow). |
aoqi@0 | 1449 | Interpreter::_rethrow_exception_entry = __ pc(); |
aoqi@0 | 1450 | { |
aoqi@0 | 1451 | __ restore_interpreter_state(R11_scratch1); // Sets R11_scratch1 = fp. |
aoqi@0 | 1452 | __ ld(R12_scratch2, _ijava_state_neg(top_frame_sp), R11_scratch1); |
aoqi@0 | 1453 | __ resize_frame_absolute(R12_scratch2, R11_scratch1, R0); |
aoqi@0 | 1454 | |
aoqi@0 | 1455 | // Compiled code destroys templateTableBase, reload. |
aoqi@0 | 1456 | __ load_const_optimized(R25_templateTableBase, (address)Interpreter::dispatch_table((TosState)0), R11_scratch1); |
aoqi@0 | 1457 | } |
aoqi@0 | 1458 | |
aoqi@0 | 1459 | // Entry point if a interpreted method throws an exception (throw). |
aoqi@0 | 1460 | Interpreter::_throw_exception_entry = __ pc(); |
aoqi@0 | 1461 | { |
aoqi@0 | 1462 | __ mr(Rexception, R3_RET); |
aoqi@0 | 1463 | |
aoqi@0 | 1464 | __ verify_thread(); |
aoqi@0 | 1465 | __ verify_oop(Rexception); |
aoqi@0 | 1466 | |
aoqi@0 | 1467 | // Expression stack must be empty before entering the VM in case of an exception. |
aoqi@0 | 1468 | __ empty_expression_stack(); |
aoqi@0 | 1469 | // Find exception handler address and preserve exception oop. |
aoqi@0 | 1470 | // Call C routine to find handler and jump to it. |
aoqi@0 | 1471 | __ call_VM(Rexception, CAST_FROM_FN_PTR(address, InterpreterRuntime::exception_handler_for_exception), Rexception); |
aoqi@0 | 1472 | __ mtctr(Rcontinuation); |
aoqi@0 | 1473 | // Push exception for exception handler bytecodes. |
aoqi@0 | 1474 | __ push_ptr(Rexception); |
aoqi@0 | 1475 | |
aoqi@0 | 1476 | // Jump to exception handler (may be remove activation entry!). |
aoqi@0 | 1477 | __ bctr(); |
aoqi@0 | 1478 | } |
aoqi@0 | 1479 | |
aoqi@0 | 1480 | // If the exception is not handled in the current frame the frame is |
aoqi@0 | 1481 | // removed and the exception is rethrown (i.e. exception |
aoqi@0 | 1482 | // continuation is _rethrow_exception). |
aoqi@0 | 1483 | // |
aoqi@0 | 1484 | // Note: At this point the bci is still the bxi for the instruction |
aoqi@0 | 1485 | // which caused the exception and the expression stack is |
aoqi@0 | 1486 | // empty. Thus, for any VM calls at this point, GC will find a legal |
aoqi@0 | 1487 | // oop map (with empty expression stack). |
aoqi@0 | 1488 | |
aoqi@0 | 1489 | // In current activation |
aoqi@0 | 1490 | // tos: exception |
aoqi@0 | 1491 | // bcp: exception bcp |
aoqi@0 | 1492 | |
aoqi@0 | 1493 | // -------------------------------------------------------------------------- |
aoqi@0 | 1494 | // JVMTI PopFrame support |
aoqi@0 | 1495 | |
aoqi@0 | 1496 | Interpreter::_remove_activation_preserving_args_entry = __ pc(); |
aoqi@0 | 1497 | { |
aoqi@0 | 1498 | // Set the popframe_processing bit in popframe_condition indicating that we are |
aoqi@0 | 1499 | // currently handling popframe, so that call_VMs that may happen later do not |
aoqi@0 | 1500 | // trigger new popframe handling cycles. |
aoqi@0 | 1501 | __ lwz(R11_scratch1, in_bytes(JavaThread::popframe_condition_offset()), R16_thread); |
aoqi@0 | 1502 | __ ori(R11_scratch1, R11_scratch1, JavaThread::popframe_processing_bit); |
aoqi@0 | 1503 | __ stw(R11_scratch1, in_bytes(JavaThread::popframe_condition_offset()), R16_thread); |
aoqi@0 | 1504 | |
aoqi@0 | 1505 | // Empty the expression stack, as in normal exception handling. |
aoqi@0 | 1506 | __ empty_expression_stack(); |
aoqi@0 | 1507 | __ unlock_if_synchronized_method(vtos, /* throw_monitor_exception */ false, /* install_monitor_exception */ false); |
aoqi@0 | 1508 | |
aoqi@0 | 1509 | // Check to see whether we are returning to a deoptimized frame. |
aoqi@0 | 1510 | // (The PopFrame call ensures that the caller of the popped frame is |
aoqi@0 | 1511 | // either interpreted or compiled and deoptimizes it if compiled.) |
aoqi@0 | 1512 | // Note that we don't compare the return PC against the |
aoqi@0 | 1513 | // deoptimization blob's unpack entry because of the presence of |
aoqi@0 | 1514 | // adapter frames in C2. |
aoqi@0 | 1515 | Label Lcaller_not_deoptimized; |
aoqi@0 | 1516 | Register return_pc = R3_ARG1; |
aoqi@0 | 1517 | __ ld(return_pc, 0, R1_SP); |
aoqi@0 | 1518 | __ ld(return_pc, _abi(lr), return_pc); |
aoqi@0 | 1519 | __ call_VM_leaf(CAST_FROM_FN_PTR(address, InterpreterRuntime::interpreter_contains), return_pc); |
aoqi@0 | 1520 | __ cmpdi(CCR0, R3_RET, 0); |
aoqi@0 | 1521 | __ bne(CCR0, Lcaller_not_deoptimized); |
aoqi@0 | 1522 | |
aoqi@0 | 1523 | // The deoptimized case. |
aoqi@0 | 1524 | // In this case, we can't call dispatch_next() after the frame is |
aoqi@0 | 1525 | // popped, but instead must save the incoming arguments and restore |
aoqi@0 | 1526 | // them after deoptimization has occurred. |
aoqi@0 | 1527 | __ ld(R4_ARG2, in_bytes(Method::const_offset()), R19_method); |
aoqi@0 | 1528 | __ lhz(R4_ARG2 /* number of params */, in_bytes(ConstMethod::size_of_parameters_offset()), R4_ARG2); |
aoqi@0 | 1529 | __ slwi(R4_ARG2, R4_ARG2, Interpreter::logStackElementSize); |
aoqi@0 | 1530 | __ addi(R5_ARG3, R18_locals, Interpreter::stackElementSize); |
aoqi@0 | 1531 | __ subf(R5_ARG3, R4_ARG2, R5_ARG3); |
aoqi@0 | 1532 | // Save these arguments. |
aoqi@0 | 1533 | __ call_VM_leaf(CAST_FROM_FN_PTR(address, Deoptimization::popframe_preserve_args), R16_thread, R4_ARG2, R5_ARG3); |
aoqi@0 | 1534 | |
aoqi@0 | 1535 | // Inform deoptimization that it is responsible for restoring these arguments. |
aoqi@0 | 1536 | __ load_const_optimized(R11_scratch1, JavaThread::popframe_force_deopt_reexecution_bit); |
aoqi@0 | 1537 | __ stw(R11_scratch1, in_bytes(JavaThread::popframe_condition_offset()), R16_thread); |
aoqi@0 | 1538 | |
aoqi@0 | 1539 | // Return from the current method into the deoptimization blob. Will eventually |
aoqi@0 | 1540 | // end up in the deopt interpeter entry, deoptimization prepared everything that |
aoqi@0 | 1541 | // we will reexecute the call that called us. |
aoqi@0 | 1542 | __ merge_frames(/*top_frame_sp*/ R21_sender_SP, /*reload return_pc*/ return_pc, R11_scratch1, R12_scratch2); |
aoqi@0 | 1543 | __ mtlr(return_pc); |
aoqi@0 | 1544 | __ blr(); |
aoqi@0 | 1545 | |
aoqi@0 | 1546 | // The non-deoptimized case. |
aoqi@0 | 1547 | __ bind(Lcaller_not_deoptimized); |
aoqi@0 | 1548 | |
aoqi@0 | 1549 | // Clear the popframe condition flag. |
aoqi@0 | 1550 | __ li(R0, 0); |
aoqi@0 | 1551 | __ stw(R0, in_bytes(JavaThread::popframe_condition_offset()), R16_thread); |
aoqi@0 | 1552 | |
aoqi@0 | 1553 | // Get out of the current method and re-execute the call that called us. |
aoqi@0 | 1554 | __ merge_frames(/*top_frame_sp*/ R21_sender_SP, /*return_pc*/ noreg, R11_scratch1, R12_scratch2); |
aoqi@0 | 1555 | __ restore_interpreter_state(R11_scratch1); |
aoqi@0 | 1556 | __ ld(R12_scratch2, _ijava_state_neg(top_frame_sp), R11_scratch1); |
aoqi@0 | 1557 | __ resize_frame_absolute(R12_scratch2, R11_scratch1, R0); |
aoqi@0 | 1558 | if (ProfileInterpreter) { |
aoqi@0 | 1559 | __ set_method_data_pointer_for_bcp(); |
goetz@7424 | 1560 | __ ld(R11_scratch1, 0, R1_SP); |
goetz@7424 | 1561 | __ std(R28_mdx, _ijava_state_neg(mdx), R11_scratch1); |
aoqi@0 | 1562 | } |
aoqi@0 | 1563 | #if INCLUDE_JVMTI |
aoqi@0 | 1564 | Label L_done; |
aoqi@0 | 1565 | |
aoqi@0 | 1566 | __ lbz(R11_scratch1, 0, R14_bcp); |
aoqi@0 | 1567 | __ cmpwi(CCR0, R11_scratch1, Bytecodes::_invokestatic); |
aoqi@0 | 1568 | __ bne(CCR0, L_done); |
aoqi@0 | 1569 | |
aoqi@0 | 1570 | // The member name argument must be restored if _invokestatic is re-executed after a PopFrame call. |
aoqi@0 | 1571 | // Detect such a case in the InterpreterRuntime function and return the member name argument, or NULL. |
aoqi@0 | 1572 | __ ld(R4_ARG2, 0, R18_locals); |
goetz@7424 | 1573 | __ MacroAssembler::call_VM(R4_ARG2, CAST_FROM_FN_PTR(address, InterpreterRuntime::member_name_arg_or_null), R4_ARG2, R19_method, R14_bcp, false); |
goetz@7424 | 1574 | __ restore_interpreter_state(R11_scratch1, /*bcp_and_mdx_only*/ true); |
goetz@7424 | 1575 | __ cmpdi(CCR0, R4_ARG2, 0); |
aoqi@0 | 1576 | __ beq(CCR0, L_done); |
goetz@7424 | 1577 | __ std(R4_ARG2, wordSize, R15_esp); |
aoqi@0 | 1578 | __ bind(L_done); |
aoqi@0 | 1579 | #endif // INCLUDE_JVMTI |
aoqi@0 | 1580 | __ dispatch_next(vtos); |
aoqi@0 | 1581 | } |
aoqi@0 | 1582 | // end of JVMTI PopFrame support |
aoqi@0 | 1583 | |
aoqi@0 | 1584 | // -------------------------------------------------------------------------- |
aoqi@0 | 1585 | // Remove activation exception entry. |
aoqi@0 | 1586 | // This is jumped to if an interpreted method can't handle an exception itself |
aoqi@0 | 1587 | // (we come from the throw/rethrow exception entry above). We're going to call |
aoqi@0 | 1588 | // into the VM to find the exception handler in the caller, pop the current |
aoqi@0 | 1589 | // frame and return the handler we calculated. |
aoqi@0 | 1590 | Interpreter::_remove_activation_entry = __ pc(); |
aoqi@0 | 1591 | { |
aoqi@0 | 1592 | __ pop_ptr(Rexception); |
aoqi@0 | 1593 | __ verify_thread(); |
aoqi@0 | 1594 | __ verify_oop(Rexception); |
aoqi@0 | 1595 | __ std(Rexception, in_bytes(JavaThread::vm_result_offset()), R16_thread); |
aoqi@0 | 1596 | |
aoqi@0 | 1597 | __ unlock_if_synchronized_method(vtos, /* throw_monitor_exception */ false, true); |
aoqi@0 | 1598 | __ notify_method_exit(false, vtos, InterpreterMacroAssembler::SkipNotifyJVMTI, false); |
aoqi@0 | 1599 | |
aoqi@0 | 1600 | __ get_vm_result(Rexception); |
aoqi@0 | 1601 | |
aoqi@0 | 1602 | // We are done with this activation frame; find out where to go next. |
aoqi@0 | 1603 | // The continuation point will be an exception handler, which expects |
aoqi@0 | 1604 | // the following registers set up: |
aoqi@0 | 1605 | // |
aoqi@0 | 1606 | // RET: exception oop |
aoqi@0 | 1607 | // ARG2: Issuing PC (see generate_exception_blob()), only used if the caller is compiled. |
aoqi@0 | 1608 | |
aoqi@0 | 1609 | Register return_pc = R31; // Needs to survive the runtime call. |
aoqi@0 | 1610 | __ ld(return_pc, 0, R1_SP); |
aoqi@0 | 1611 | __ ld(return_pc, _abi(lr), return_pc); |
aoqi@0 | 1612 | __ call_VM_leaf(CAST_FROM_FN_PTR(address, SharedRuntime::exception_handler_for_return_address), R16_thread, return_pc); |
aoqi@0 | 1613 | |
aoqi@0 | 1614 | // Remove the current activation. |
aoqi@0 | 1615 | __ merge_frames(/*top_frame_sp*/ R21_sender_SP, /*return_pc*/ noreg, R11_scratch1, R12_scratch2); |
aoqi@0 | 1616 | |
aoqi@0 | 1617 | __ mr(R4_ARG2, return_pc); |
aoqi@0 | 1618 | __ mtlr(R3_RET); |
aoqi@0 | 1619 | __ mr(R3_RET, Rexception); |
aoqi@0 | 1620 | __ blr(); |
aoqi@0 | 1621 | } |
aoqi@0 | 1622 | } |
aoqi@0 | 1623 | |
aoqi@0 | 1624 | // JVMTI ForceEarlyReturn support. |
aoqi@0 | 1625 | // Returns "in the middle" of a method with a "fake" return value. |
aoqi@0 | 1626 | address TemplateInterpreterGenerator::generate_earlyret_entry_for(TosState state) { |
aoqi@0 | 1627 | |
aoqi@0 | 1628 | Register Rscratch1 = R11_scratch1, |
aoqi@0 | 1629 | Rscratch2 = R12_scratch2; |
aoqi@0 | 1630 | |
aoqi@0 | 1631 | address entry = __ pc(); |
aoqi@0 | 1632 | __ empty_expression_stack(); |
aoqi@0 | 1633 | |
aoqi@0 | 1634 | __ load_earlyret_value(state, Rscratch1); |
aoqi@0 | 1635 | |
aoqi@0 | 1636 | __ ld(Rscratch1, in_bytes(JavaThread::jvmti_thread_state_offset()), R16_thread); |
aoqi@0 | 1637 | // Clear the earlyret state. |
aoqi@0 | 1638 | __ li(R0, 0); |
aoqi@0 | 1639 | __ stw(R0, in_bytes(JvmtiThreadState::earlyret_state_offset()), Rscratch1); |
aoqi@0 | 1640 | |
aoqi@0 | 1641 | __ remove_activation(state, false, false); |
aoqi@0 | 1642 | // Copied from TemplateTable::_return. |
aoqi@0 | 1643 | // Restoration of lr done by remove_activation. |
aoqi@0 | 1644 | switch (state) { |
aoqi@0 | 1645 | case ltos: |
aoqi@0 | 1646 | case btos: |
aoqi@0 | 1647 | case ctos: |
aoqi@0 | 1648 | case stos: |
aoqi@0 | 1649 | case atos: |
aoqi@0 | 1650 | case itos: __ mr(R3_RET, R17_tos); break; |
aoqi@0 | 1651 | case ftos: |
aoqi@0 | 1652 | case dtos: __ fmr(F1_RET, F15_ftos); break; |
aoqi@0 | 1653 | case vtos: // This might be a constructor. Final fields (and volatile fields on PPC64) need |
aoqi@0 | 1654 | // to get visible before the reference to the object gets stored anywhere. |
aoqi@0 | 1655 | __ membar(Assembler::StoreStore); break; |
aoqi@0 | 1656 | default : ShouldNotReachHere(); |
aoqi@0 | 1657 | } |
aoqi@0 | 1658 | __ blr(); |
aoqi@0 | 1659 | |
aoqi@0 | 1660 | return entry; |
aoqi@0 | 1661 | } // end of ForceEarlyReturn support |
aoqi@0 | 1662 | |
aoqi@0 | 1663 | //----------------------------------------------------------------------------- |
aoqi@0 | 1664 | // Helper for vtos entry point generation |
aoqi@0 | 1665 | |
aoqi@0 | 1666 | void TemplateInterpreterGenerator::set_vtos_entry_points(Template* t, |
aoqi@0 | 1667 | address& bep, |
aoqi@0 | 1668 | address& cep, |
aoqi@0 | 1669 | address& sep, |
aoqi@0 | 1670 | address& aep, |
aoqi@0 | 1671 | address& iep, |
aoqi@0 | 1672 | address& lep, |
aoqi@0 | 1673 | address& fep, |
aoqi@0 | 1674 | address& dep, |
aoqi@0 | 1675 | address& vep) { |
aoqi@0 | 1676 | assert(t->is_valid() && t->tos_in() == vtos, "illegal template"); |
aoqi@0 | 1677 | Label L; |
aoqi@0 | 1678 | |
aoqi@0 | 1679 | aep = __ pc(); __ push_ptr(); __ b(L); |
aoqi@0 | 1680 | fep = __ pc(); __ push_f(); __ b(L); |
aoqi@0 | 1681 | dep = __ pc(); __ push_d(); __ b(L); |
aoqi@0 | 1682 | lep = __ pc(); __ push_l(); __ b(L); |
aoqi@0 | 1683 | __ align(32, 12, 24); // align L |
aoqi@0 | 1684 | bep = cep = sep = |
aoqi@0 | 1685 | iep = __ pc(); __ push_i(); |
aoqi@0 | 1686 | vep = __ pc(); |
aoqi@0 | 1687 | __ bind(L); |
aoqi@0 | 1688 | generate_and_dispatch(t); |
aoqi@0 | 1689 | } |
aoqi@0 | 1690 | |
aoqi@0 | 1691 | //----------------------------------------------------------------------------- |
aoqi@0 | 1692 | // Generation of individual instructions |
aoqi@0 | 1693 | |
aoqi@0 | 1694 | // helpers for generate_and_dispatch |
aoqi@0 | 1695 | |
aoqi@0 | 1696 | InterpreterGenerator::InterpreterGenerator(StubQueue* code) |
aoqi@0 | 1697 | : TemplateInterpreterGenerator(code) { |
aoqi@0 | 1698 | generate_all(); // Down here so it can be "virtual". |
aoqi@0 | 1699 | } |
aoqi@0 | 1700 | |
aoqi@0 | 1701 | //----------------------------------------------------------------------------- |
aoqi@0 | 1702 | |
aoqi@0 | 1703 | // Non-product code |
aoqi@0 | 1704 | #ifndef PRODUCT |
aoqi@0 | 1705 | address TemplateInterpreterGenerator::generate_trace_code(TosState state) { |
aoqi@0 | 1706 | //__ flush_bundle(); |
aoqi@0 | 1707 | address entry = __ pc(); |
aoqi@0 | 1708 | |
aoqi@0 | 1709 | const char *bname = NULL; |
aoqi@0 | 1710 | uint tsize = 0; |
aoqi@0 | 1711 | switch(state) { |
aoqi@0 | 1712 | case ftos: |
aoqi@0 | 1713 | bname = "trace_code_ftos {"; |
aoqi@0 | 1714 | tsize = 2; |
aoqi@0 | 1715 | break; |
aoqi@0 | 1716 | case btos: |
aoqi@0 | 1717 | bname = "trace_code_btos {"; |
aoqi@0 | 1718 | tsize = 2; |
aoqi@0 | 1719 | break; |
aoqi@0 | 1720 | case ctos: |
aoqi@0 | 1721 | bname = "trace_code_ctos {"; |
aoqi@0 | 1722 | tsize = 2; |
aoqi@0 | 1723 | break; |
aoqi@0 | 1724 | case stos: |
aoqi@0 | 1725 | bname = "trace_code_stos {"; |
aoqi@0 | 1726 | tsize = 2; |
aoqi@0 | 1727 | break; |
aoqi@0 | 1728 | case itos: |
aoqi@0 | 1729 | bname = "trace_code_itos {"; |
aoqi@0 | 1730 | tsize = 2; |
aoqi@0 | 1731 | break; |
aoqi@0 | 1732 | case ltos: |
aoqi@0 | 1733 | bname = "trace_code_ltos {"; |
aoqi@0 | 1734 | tsize = 3; |
aoqi@0 | 1735 | break; |
aoqi@0 | 1736 | case atos: |
aoqi@0 | 1737 | bname = "trace_code_atos {"; |
aoqi@0 | 1738 | tsize = 2; |
aoqi@0 | 1739 | break; |
aoqi@0 | 1740 | case vtos: |
aoqi@0 | 1741 | // Note: In case of vtos, the topmost of stack value could be a int or doubl |
aoqi@0 | 1742 | // In case of a double (2 slots) we won't see the 2nd stack value. |
aoqi@0 | 1743 | // Maybe we simply should print the topmost 3 stack slots to cope with the problem. |
aoqi@0 | 1744 | bname = "trace_code_vtos {"; |
aoqi@0 | 1745 | tsize = 2; |
aoqi@0 | 1746 | |
aoqi@0 | 1747 | break; |
aoqi@0 | 1748 | case dtos: |
aoqi@0 | 1749 | bname = "trace_code_dtos {"; |
aoqi@0 | 1750 | tsize = 3; |
aoqi@0 | 1751 | break; |
aoqi@0 | 1752 | default: |
aoqi@0 | 1753 | ShouldNotReachHere(); |
aoqi@0 | 1754 | } |
aoqi@0 | 1755 | BLOCK_COMMENT(bname); |
aoqi@0 | 1756 | |
aoqi@0 | 1757 | // Support short-cut for TraceBytecodesAt. |
aoqi@0 | 1758 | // Don't call into the VM if we don't want to trace to speed up things. |
aoqi@0 | 1759 | Label Lskip_vm_call; |
aoqi@0 | 1760 | if (TraceBytecodesAt > 0 && TraceBytecodesAt < max_intx) { |
aoqi@0 | 1761 | int offs1 = __ load_const_optimized(R11_scratch1, (address) &TraceBytecodesAt, R0, true); |
aoqi@0 | 1762 | int offs2 = __ load_const_optimized(R12_scratch2, (address) &BytecodeCounter::_counter_value, R0, true); |
aoqi@0 | 1763 | __ ld(R11_scratch1, offs1, R11_scratch1); |
aoqi@0 | 1764 | __ lwa(R12_scratch2, offs2, R12_scratch2); |
aoqi@0 | 1765 | __ cmpd(CCR0, R12_scratch2, R11_scratch1); |
aoqi@0 | 1766 | __ blt(CCR0, Lskip_vm_call); |
aoqi@0 | 1767 | } |
aoqi@0 | 1768 | |
aoqi@0 | 1769 | __ push(state); |
aoqi@0 | 1770 | // Load 2 topmost expression stack values. |
aoqi@0 | 1771 | __ ld(R6_ARG4, tsize*Interpreter::stackElementSize, R15_esp); |
aoqi@0 | 1772 | __ ld(R5_ARG3, Interpreter::stackElementSize, R15_esp); |
aoqi@0 | 1773 | __ mflr(R31); |
aoqi@0 | 1774 | __ call_VM(noreg, CAST_FROM_FN_PTR(address, SharedRuntime::trace_bytecode), /* unused */ R4_ARG2, R5_ARG3, R6_ARG4, false); |
aoqi@0 | 1775 | __ mtlr(R31); |
aoqi@0 | 1776 | __ pop(state); |
aoqi@0 | 1777 | |
aoqi@0 | 1778 | if (TraceBytecodesAt > 0 && TraceBytecodesAt < max_intx) { |
aoqi@0 | 1779 | __ bind(Lskip_vm_call); |
aoqi@0 | 1780 | } |
aoqi@0 | 1781 | __ blr(); |
aoqi@0 | 1782 | BLOCK_COMMENT("} trace_code"); |
aoqi@0 | 1783 | return entry; |
aoqi@0 | 1784 | } |
aoqi@0 | 1785 | |
aoqi@0 | 1786 | void TemplateInterpreterGenerator::count_bytecode() { |
aoqi@0 | 1787 | int offs = __ load_const_optimized(R11_scratch1, (address) &BytecodeCounter::_counter_value, R12_scratch2, true); |
aoqi@0 | 1788 | __ lwz(R12_scratch2, offs, R11_scratch1); |
aoqi@0 | 1789 | __ addi(R12_scratch2, R12_scratch2, 1); |
aoqi@0 | 1790 | __ stw(R12_scratch2, offs, R11_scratch1); |
aoqi@0 | 1791 | } |
aoqi@0 | 1792 | |
aoqi@0 | 1793 | void TemplateInterpreterGenerator::histogram_bytecode(Template* t) { |
aoqi@0 | 1794 | int offs = __ load_const_optimized(R11_scratch1, (address) &BytecodeHistogram::_counters[t->bytecode()], R12_scratch2, true); |
aoqi@0 | 1795 | __ lwz(R12_scratch2, offs, R11_scratch1); |
aoqi@0 | 1796 | __ addi(R12_scratch2, R12_scratch2, 1); |
aoqi@0 | 1797 | __ stw(R12_scratch2, offs, R11_scratch1); |
aoqi@0 | 1798 | } |
aoqi@0 | 1799 | |
aoqi@0 | 1800 | void TemplateInterpreterGenerator::histogram_bytecode_pair(Template* t) { |
aoqi@0 | 1801 | const Register addr = R11_scratch1, |
aoqi@0 | 1802 | tmp = R12_scratch2; |
aoqi@0 | 1803 | // Get index, shift out old bytecode, bring in new bytecode, and store it. |
aoqi@0 | 1804 | // _index = (_index >> log2_number_of_codes) | |
aoqi@0 | 1805 | // (bytecode << log2_number_of_codes); |
aoqi@0 | 1806 | int offs1 = __ load_const_optimized(addr, (address)&BytecodePairHistogram::_index, tmp, true); |
aoqi@0 | 1807 | __ lwz(tmp, offs1, addr); |
aoqi@0 | 1808 | __ srwi(tmp, tmp, BytecodePairHistogram::log2_number_of_codes); |
aoqi@0 | 1809 | __ ori(tmp, tmp, ((int) t->bytecode()) << BytecodePairHistogram::log2_number_of_codes); |
aoqi@0 | 1810 | __ stw(tmp, offs1, addr); |
aoqi@0 | 1811 | |
aoqi@0 | 1812 | // Bump bucket contents. |
aoqi@0 | 1813 | // _counters[_index] ++; |
aoqi@0 | 1814 | int offs2 = __ load_const_optimized(addr, (address)&BytecodePairHistogram::_counters, R0, true); |
aoqi@0 | 1815 | __ sldi(tmp, tmp, LogBytesPerInt); |
aoqi@0 | 1816 | __ add(addr, tmp, addr); |
aoqi@0 | 1817 | __ lwz(tmp, offs2, addr); |
aoqi@0 | 1818 | __ addi(tmp, tmp, 1); |
aoqi@0 | 1819 | __ stw(tmp, offs2, addr); |
aoqi@0 | 1820 | } |
aoqi@0 | 1821 | |
aoqi@0 | 1822 | void TemplateInterpreterGenerator::trace_bytecode(Template* t) { |
aoqi@0 | 1823 | // Call a little run-time stub to avoid blow-up for each bytecode. |
aoqi@0 | 1824 | // The run-time runtime saves the right registers, depending on |
aoqi@0 | 1825 | // the tosca in-state for the given template. |
aoqi@0 | 1826 | |
aoqi@0 | 1827 | assert(Interpreter::trace_code(t->tos_in()) != NULL, |
aoqi@0 | 1828 | "entry must have been generated"); |
aoqi@0 | 1829 | |
aoqi@0 | 1830 | // Note: we destroy LR here. |
aoqi@0 | 1831 | __ bl(Interpreter::trace_code(t->tos_in())); |
aoqi@0 | 1832 | } |
aoqi@0 | 1833 | |
aoqi@0 | 1834 | void TemplateInterpreterGenerator::stop_interpreter_at() { |
aoqi@0 | 1835 | Label L; |
aoqi@0 | 1836 | int offs1 = __ load_const_optimized(R11_scratch1, (address) &StopInterpreterAt, R0, true); |
aoqi@0 | 1837 | int offs2 = __ load_const_optimized(R12_scratch2, (address) &BytecodeCounter::_counter_value, R0, true); |
aoqi@0 | 1838 | __ ld(R11_scratch1, offs1, R11_scratch1); |
aoqi@0 | 1839 | __ lwa(R12_scratch2, offs2, R12_scratch2); |
aoqi@0 | 1840 | __ cmpd(CCR0, R12_scratch2, R11_scratch1); |
aoqi@0 | 1841 | __ bne(CCR0, L); |
aoqi@0 | 1842 | __ illtrap(); |
aoqi@0 | 1843 | __ bind(L); |
aoqi@0 | 1844 | } |
aoqi@0 | 1845 | |
aoqi@0 | 1846 | #endif // !PRODUCT |
aoqi@0 | 1847 | #endif // !CC_INTERP |