Wed, 15 Feb 2012 16:29:40 -0800
7142680: default GC affected by jvm path
Summary: Removed old tiered code
Reviewed-by: never, kvn
duke@435 | 1 | /* |
never@2462 | 2 | * Copyright (c) 1997, 2011, Oracle and/or its affiliates. All rights reserved. |
duke@435 | 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
duke@435 | 4 | * |
duke@435 | 5 | * This code is free software; you can redistribute it and/or modify it |
duke@435 | 6 | * under the terms of the GNU General Public License version 2 only, as |
duke@435 | 7 | * published by the Free Software Foundation. |
duke@435 | 8 | * |
duke@435 | 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
duke@435 | 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
duke@435 | 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
duke@435 | 12 | * version 2 for more details (a copy is included in the LICENSE file that |
duke@435 | 13 | * accompanied this code). |
duke@435 | 14 | * |
duke@435 | 15 | * You should have received a copy of the GNU General Public License version |
duke@435 | 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
duke@435 | 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
duke@435 | 18 | * |
trims@1907 | 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
trims@1907 | 20 | * or visit www.oracle.com if you need additional information or have any |
trims@1907 | 21 | * questions. |
duke@435 | 22 | * |
duke@435 | 23 | */ |
duke@435 | 24 | |
stefank@2314 | 25 | #include "precompiled.hpp" |
stefank@2314 | 26 | #include "classfile/vmSymbols.hpp" |
stefank@2314 | 27 | #include "interpreter/interpreter.hpp" |
stefank@2314 | 28 | #include "memory/allocation.inline.hpp" |
stefank@2314 | 29 | #include "memory/resourceArea.hpp" |
stefank@2314 | 30 | #include "memory/universe.inline.hpp" |
stefank@2314 | 31 | #include "oops/methodDataOop.hpp" |
stefank@2314 | 32 | #include "oops/oop.inline.hpp" |
stefank@2314 | 33 | #include "prims/jvmtiThreadState.hpp" |
stefank@2314 | 34 | #include "runtime/handles.inline.hpp" |
stefank@2314 | 35 | #include "runtime/monitorChunk.hpp" |
stefank@2314 | 36 | #include "runtime/sharedRuntime.hpp" |
stefank@2314 | 37 | #include "runtime/vframe.hpp" |
stefank@2314 | 38 | #include "runtime/vframeArray.hpp" |
stefank@2314 | 39 | #include "runtime/vframe_hp.hpp" |
stefank@2314 | 40 | #include "utilities/events.hpp" |
stefank@2314 | 41 | #ifdef COMPILER2 |
stefank@2314 | 42 | #include "opto/runtime.hpp" |
stefank@2314 | 43 | #endif |
duke@435 | 44 | |
duke@435 | 45 | |
duke@435 | 46 | int vframeArrayElement:: bci(void) const { return (_bci == SynchronizationEntryBCI ? 0 : _bci); } |
duke@435 | 47 | |
duke@435 | 48 | void vframeArrayElement::free_monitors(JavaThread* jt) { |
duke@435 | 49 | if (_monitors != NULL) { |
duke@435 | 50 | MonitorChunk* chunk = _monitors; |
duke@435 | 51 | _monitors = NULL; |
duke@435 | 52 | jt->remove_monitor_chunk(chunk); |
duke@435 | 53 | delete chunk; |
duke@435 | 54 | } |
duke@435 | 55 | } |
duke@435 | 56 | |
duke@435 | 57 | void vframeArrayElement::fill_in(compiledVFrame* vf) { |
duke@435 | 58 | |
duke@435 | 59 | // Copy the information from the compiled vframe to the |
duke@435 | 60 | // interpreter frame we will be creating to replace vf |
duke@435 | 61 | |
duke@435 | 62 | _method = vf->method(); |
duke@435 | 63 | _bci = vf->raw_bci(); |
cfang@1335 | 64 | _reexecute = vf->should_reexecute(); |
duke@435 | 65 | |
duke@435 | 66 | int index; |
duke@435 | 67 | |
duke@435 | 68 | // Get the monitors off-stack |
duke@435 | 69 | |
duke@435 | 70 | GrowableArray<MonitorInfo*>* list = vf->monitors(); |
duke@435 | 71 | if (list->is_empty()) { |
duke@435 | 72 | _monitors = NULL; |
duke@435 | 73 | } else { |
duke@435 | 74 | |
duke@435 | 75 | // Allocate monitor chunk |
duke@435 | 76 | _monitors = new MonitorChunk(list->length()); |
duke@435 | 77 | vf->thread()->add_monitor_chunk(_monitors); |
duke@435 | 78 | |
duke@435 | 79 | // Migrate the BasicLocks from the stack to the monitor chunk |
duke@435 | 80 | for (index = 0; index < list->length(); index++) { |
duke@435 | 81 | MonitorInfo* monitor = list->at(index); |
kvn@1253 | 82 | assert(!monitor->owner_is_scalar_replaced(), "object should be reallocated already"); |
duke@435 | 83 | assert(monitor->owner() == NULL || (!monitor->owner()->is_unlocked() && !monitor->owner()->has_bias_pattern()), "object must be null or locked, and unbiased"); |
duke@435 | 84 | BasicObjectLock* dest = _monitors->at(index); |
duke@435 | 85 | dest->set_obj(monitor->owner()); |
duke@435 | 86 | monitor->lock()->move_to(monitor->owner(), dest->lock()); |
duke@435 | 87 | } |
duke@435 | 88 | } |
duke@435 | 89 | |
duke@435 | 90 | // Convert the vframe locals and expressions to off stack |
duke@435 | 91 | // values. Because we will not gc all oops can be converted to |
duke@435 | 92 | // intptr_t (i.e. a stack slot) and we are fine. This is |
duke@435 | 93 | // good since we are inside a HandleMark and the oops in our |
duke@435 | 94 | // collection would go away between packing them here and |
duke@435 | 95 | // unpacking them in unpack_on_stack. |
duke@435 | 96 | |
duke@435 | 97 | // First the locals go off-stack |
duke@435 | 98 | |
duke@435 | 99 | // FIXME this seems silly it creates a StackValueCollection |
duke@435 | 100 | // in order to get the size to then copy them and |
duke@435 | 101 | // convert the types to intptr_t size slots. Seems like it |
duke@435 | 102 | // could do it in place... Still uses less memory than the |
duke@435 | 103 | // old way though |
duke@435 | 104 | |
duke@435 | 105 | StackValueCollection *locs = vf->locals(); |
duke@435 | 106 | _locals = new StackValueCollection(locs->size()); |
duke@435 | 107 | for(index = 0; index < locs->size(); index++) { |
duke@435 | 108 | StackValue* value = locs->at(index); |
duke@435 | 109 | switch(value->type()) { |
duke@435 | 110 | case T_OBJECT: |
kvn@1253 | 111 | assert(!value->obj_is_scalar_replaced(), "object should be reallocated already"); |
duke@435 | 112 | // preserve object type |
duke@435 | 113 | _locals->add( new StackValue((intptr_t) (value->get_obj()()), T_OBJECT )); |
duke@435 | 114 | break; |
duke@435 | 115 | case T_CONFLICT: |
duke@435 | 116 | // A dead local. Will be initialized to null/zero. |
duke@435 | 117 | _locals->add( new StackValue()); |
duke@435 | 118 | break; |
duke@435 | 119 | case T_INT: |
duke@435 | 120 | _locals->add( new StackValue(value->get_int())); |
duke@435 | 121 | break; |
duke@435 | 122 | default: |
duke@435 | 123 | ShouldNotReachHere(); |
duke@435 | 124 | } |
duke@435 | 125 | } |
duke@435 | 126 | |
duke@435 | 127 | // Now the expressions off-stack |
duke@435 | 128 | // Same silliness as above |
duke@435 | 129 | |
duke@435 | 130 | StackValueCollection *exprs = vf->expressions(); |
duke@435 | 131 | _expressions = new StackValueCollection(exprs->size()); |
duke@435 | 132 | for(index = 0; index < exprs->size(); index++) { |
duke@435 | 133 | StackValue* value = exprs->at(index); |
duke@435 | 134 | switch(value->type()) { |
duke@435 | 135 | case T_OBJECT: |
kvn@1253 | 136 | assert(!value->obj_is_scalar_replaced(), "object should be reallocated already"); |
duke@435 | 137 | // preserve object type |
duke@435 | 138 | _expressions->add( new StackValue((intptr_t) (value->get_obj()()), T_OBJECT )); |
duke@435 | 139 | break; |
duke@435 | 140 | case T_CONFLICT: |
duke@435 | 141 | // A dead stack element. Will be initialized to null/zero. |
duke@435 | 142 | // This can occur when the compiler emits a state in which stack |
duke@435 | 143 | // elements are known to be dead (because of an imminent exception). |
duke@435 | 144 | _expressions->add( new StackValue()); |
duke@435 | 145 | break; |
duke@435 | 146 | case T_INT: |
duke@435 | 147 | _expressions->add( new StackValue(value->get_int())); |
duke@435 | 148 | break; |
duke@435 | 149 | default: |
duke@435 | 150 | ShouldNotReachHere(); |
duke@435 | 151 | } |
duke@435 | 152 | } |
duke@435 | 153 | } |
duke@435 | 154 | |
duke@435 | 155 | int unpack_counter = 0; |
duke@435 | 156 | |
never@2901 | 157 | void vframeArrayElement::unpack_on_stack(int caller_actual_parameters, |
never@2901 | 158 | int callee_parameters, |
duke@435 | 159 | int callee_locals, |
duke@435 | 160 | frame* caller, |
duke@435 | 161 | bool is_top_frame, |
duke@435 | 162 | int exec_mode) { |
duke@435 | 163 | JavaThread* thread = (JavaThread*) Thread::current(); |
duke@435 | 164 | |
duke@435 | 165 | // Look at bci and decide on bcp and continuation pc |
duke@435 | 166 | address bcp; |
duke@435 | 167 | // C++ interpreter doesn't need a pc since it will figure out what to do when it |
duke@435 | 168 | // begins execution |
duke@435 | 169 | address pc; |
cfang@1335 | 170 | bool use_next_mdp = false; // true if we should use the mdp associated with the next bci |
cfang@1335 | 171 | // rather than the one associated with bcp |
duke@435 | 172 | if (raw_bci() == SynchronizationEntryBCI) { |
duke@435 | 173 | // We are deoptimizing while hanging in prologue code for synchronized method |
duke@435 | 174 | bcp = method()->bcp_from(0); // first byte code |
duke@435 | 175 | pc = Interpreter::deopt_entry(vtos, 0); // step = 0 since we don't skip current bytecode |
cfang@1335 | 176 | } else if (should_reexecute()) { //reexecute this bytecode |
cfang@1335 | 177 | assert(is_top_frame, "reexecute allowed only for the top frame"); |
cfang@1335 | 178 | bcp = method()->bcp_from(bci()); |
cfang@1335 | 179 | pc = Interpreter::deopt_reexecute_entry(method(), bcp); |
duke@435 | 180 | } else { |
duke@435 | 181 | bcp = method()->bcp_from(bci()); |
cfang@1335 | 182 | pc = Interpreter::deopt_continue_after_entry(method(), bcp, callee_parameters, is_top_frame); |
cfang@1335 | 183 | use_next_mdp = true; |
duke@435 | 184 | } |
duke@435 | 185 | assert(Bytecodes::is_defined(*bcp), "must be a valid bytecode"); |
duke@435 | 186 | |
duke@435 | 187 | // Monitorenter and pending exceptions: |
duke@435 | 188 | // |
duke@435 | 189 | // For Compiler2, there should be no pending exception when deoptimizing at monitorenter |
duke@435 | 190 | // because there is no safepoint at the null pointer check (it is either handled explicitly |
duke@435 | 191 | // or prior to the monitorenter) and asynchronous exceptions are not made "pending" by the |
duke@435 | 192 | // runtime interface for the slow case (see JRT_ENTRY_FOR_MONITORENTER). If an asynchronous |
duke@435 | 193 | // exception was processed, the bytecode pointer would have to be extended one bytecode beyond |
duke@435 | 194 | // the monitorenter to place it in the proper exception range. |
duke@435 | 195 | // |
duke@435 | 196 | // For Compiler1, deoptimization can occur while throwing a NullPointerException at monitorenter, |
duke@435 | 197 | // in which case bcp should point to the monitorenter since it is within the exception's range. |
duke@435 | 198 | |
duke@435 | 199 | assert(*bcp != Bytecodes::_monitorenter || is_top_frame, "a _monitorenter must be a top frame"); |
iveresov@2169 | 200 | assert(thread->deopt_nmethod() != NULL, "nmethod should be known"); |
iveresov@2169 | 201 | guarantee(!(thread->deopt_nmethod()->is_compiled_by_c2() && |
iveresov@2169 | 202 | *bcp == Bytecodes::_monitorenter && |
iveresov@2169 | 203 | exec_mode == Deoptimization::Unpack_exception), |
iveresov@2169 | 204 | "shouldn't get exception during monitorenter"); |
duke@435 | 205 | |
duke@435 | 206 | int popframe_preserved_args_size_in_bytes = 0; |
duke@435 | 207 | int popframe_preserved_args_size_in_words = 0; |
duke@435 | 208 | if (is_top_frame) { |
kvn@1690 | 209 | JvmtiThreadState *state = thread->jvmti_thread_state(); |
duke@435 | 210 | if (JvmtiExport::can_pop_frame() && |
duke@435 | 211 | (thread->has_pending_popframe() || thread->popframe_forcing_deopt_reexecution())) { |
duke@435 | 212 | if (thread->has_pending_popframe()) { |
duke@435 | 213 | // Pop top frame after deoptimization |
duke@435 | 214 | #ifndef CC_INTERP |
duke@435 | 215 | pc = Interpreter::remove_activation_preserving_args_entry(); |
duke@435 | 216 | #else |
duke@435 | 217 | // Do an uncommon trap type entry. c++ interpreter will know |
duke@435 | 218 | // to pop frame and preserve the args |
duke@435 | 219 | pc = Interpreter::deopt_entry(vtos, 0); |
duke@435 | 220 | use_next_mdp = false; |
duke@435 | 221 | #endif |
duke@435 | 222 | } else { |
duke@435 | 223 | // Reexecute invoke in top frame |
duke@435 | 224 | pc = Interpreter::deopt_entry(vtos, 0); |
duke@435 | 225 | use_next_mdp = false; |
duke@435 | 226 | popframe_preserved_args_size_in_bytes = in_bytes(thread->popframe_preserved_args_size()); |
duke@435 | 227 | // Note: the PopFrame-related extension of the expression stack size is done in |
duke@435 | 228 | // Deoptimization::fetch_unroll_info_helper |
duke@435 | 229 | popframe_preserved_args_size_in_words = in_words(thread->popframe_preserved_args_size_in_words()); |
duke@435 | 230 | } |
duke@435 | 231 | } else if (JvmtiExport::can_force_early_return() && state != NULL && state->is_earlyret_pending()) { |
duke@435 | 232 | // Force early return from top frame after deoptimization |
duke@435 | 233 | #ifndef CC_INTERP |
duke@435 | 234 | pc = Interpreter::remove_activation_early_entry(state->earlyret_tos()); |
duke@435 | 235 | #else |
duke@435 | 236 | // TBD: Need to implement ForceEarlyReturn for CC_INTERP (ia64) |
duke@435 | 237 | #endif |
duke@435 | 238 | } else { |
duke@435 | 239 | // Possibly override the previous pc computation of the top (youngest) frame |
duke@435 | 240 | switch (exec_mode) { |
duke@435 | 241 | case Deoptimization::Unpack_deopt: |
duke@435 | 242 | // use what we've got |
duke@435 | 243 | break; |
duke@435 | 244 | case Deoptimization::Unpack_exception: |
duke@435 | 245 | // exception is pending |
twisti@1730 | 246 | pc = SharedRuntime::raw_exception_handler_for_return_address(thread, pc); |
duke@435 | 247 | // [phh] We're going to end up in some handler or other, so it doesn't |
duke@435 | 248 | // matter what mdp we point to. See exception_handler_for_exception() |
duke@435 | 249 | // in interpreterRuntime.cpp. |
duke@435 | 250 | break; |
duke@435 | 251 | case Deoptimization::Unpack_uncommon_trap: |
duke@435 | 252 | case Deoptimization::Unpack_reexecute: |
duke@435 | 253 | // redo last byte code |
duke@435 | 254 | pc = Interpreter::deopt_entry(vtos, 0); |
duke@435 | 255 | use_next_mdp = false; |
duke@435 | 256 | break; |
duke@435 | 257 | default: |
duke@435 | 258 | ShouldNotReachHere(); |
duke@435 | 259 | } |
duke@435 | 260 | } |
duke@435 | 261 | } |
duke@435 | 262 | |
duke@435 | 263 | // Setup the interpreter frame |
duke@435 | 264 | |
duke@435 | 265 | assert(method() != NULL, "method must exist"); |
duke@435 | 266 | int temps = expressions()->size(); |
duke@435 | 267 | |
duke@435 | 268 | int locks = monitors() == NULL ? 0 : monitors()->number_of_monitors(); |
duke@435 | 269 | |
duke@435 | 270 | Interpreter::layout_activation(method(), |
duke@435 | 271 | temps + callee_parameters, |
duke@435 | 272 | popframe_preserved_args_size_in_words, |
duke@435 | 273 | locks, |
never@2901 | 274 | caller_actual_parameters, |
duke@435 | 275 | callee_parameters, |
duke@435 | 276 | callee_locals, |
duke@435 | 277 | caller, |
duke@435 | 278 | iframe(), |
duke@435 | 279 | is_top_frame); |
duke@435 | 280 | |
duke@435 | 281 | // Update the pc in the frame object and overwrite the temporary pc |
duke@435 | 282 | // we placed in the skeletal frame now that we finally know the |
duke@435 | 283 | // exact interpreter address we should use. |
duke@435 | 284 | |
duke@435 | 285 | _frame.patch_pc(thread, pc); |
duke@435 | 286 | |
duke@435 | 287 | assert (!method()->is_synchronized() || locks > 0, "synchronized methods must have monitors"); |
duke@435 | 288 | |
duke@435 | 289 | BasicObjectLock* top = iframe()->interpreter_frame_monitor_begin(); |
duke@435 | 290 | for (int index = 0; index < locks; index++) { |
duke@435 | 291 | top = iframe()->previous_monitor_in_interpreter_frame(top); |
duke@435 | 292 | BasicObjectLock* src = _monitors->at(index); |
duke@435 | 293 | top->set_obj(src->obj()); |
duke@435 | 294 | src->lock()->move_to(src->obj(), top->lock()); |
duke@435 | 295 | } |
duke@435 | 296 | if (ProfileInterpreter) { |
duke@435 | 297 | iframe()->interpreter_frame_set_mdx(0); // clear out the mdp. |
duke@435 | 298 | } |
duke@435 | 299 | iframe()->interpreter_frame_set_bcx((intptr_t)bcp); // cannot use bcp because frame is not initialized yet |
duke@435 | 300 | if (ProfileInterpreter) { |
duke@435 | 301 | methodDataOop mdo = method()->method_data(); |
duke@435 | 302 | if (mdo != NULL) { |
duke@435 | 303 | int bci = iframe()->interpreter_frame_bci(); |
duke@435 | 304 | if (use_next_mdp) ++bci; |
duke@435 | 305 | address mdp = mdo->bci_to_dp(bci); |
duke@435 | 306 | iframe()->interpreter_frame_set_mdp(mdp); |
duke@435 | 307 | } |
duke@435 | 308 | } |
duke@435 | 309 | |
duke@435 | 310 | // Unpack expression stack |
duke@435 | 311 | // If this is an intermediate frame (i.e. not top frame) then this |
duke@435 | 312 | // only unpacks the part of the expression stack not used by callee |
duke@435 | 313 | // as parameters. The callee parameters are unpacked as part of the |
duke@435 | 314 | // callee locals. |
duke@435 | 315 | int i; |
duke@435 | 316 | for(i = 0; i < expressions()->size(); i++) { |
duke@435 | 317 | StackValue *value = expressions()->at(i); |
duke@435 | 318 | intptr_t* addr = iframe()->interpreter_frame_expression_stack_at(i); |
duke@435 | 319 | switch(value->type()) { |
duke@435 | 320 | case T_INT: |
duke@435 | 321 | *addr = value->get_int(); |
duke@435 | 322 | break; |
duke@435 | 323 | case T_OBJECT: |
duke@435 | 324 | *addr = value->get_int(T_OBJECT); |
duke@435 | 325 | break; |
duke@435 | 326 | case T_CONFLICT: |
duke@435 | 327 | // A dead stack slot. Initialize to null in case it is an oop. |
duke@435 | 328 | *addr = NULL_WORD; |
duke@435 | 329 | break; |
duke@435 | 330 | default: |
duke@435 | 331 | ShouldNotReachHere(); |
duke@435 | 332 | } |
duke@435 | 333 | } |
duke@435 | 334 | |
duke@435 | 335 | |
duke@435 | 336 | // Unpack the locals |
duke@435 | 337 | for(i = 0; i < locals()->size(); i++) { |
duke@435 | 338 | StackValue *value = locals()->at(i); |
duke@435 | 339 | intptr_t* addr = iframe()->interpreter_frame_local_at(i); |
duke@435 | 340 | switch(value->type()) { |
duke@435 | 341 | case T_INT: |
duke@435 | 342 | *addr = value->get_int(); |
duke@435 | 343 | break; |
duke@435 | 344 | case T_OBJECT: |
duke@435 | 345 | *addr = value->get_int(T_OBJECT); |
duke@435 | 346 | break; |
duke@435 | 347 | case T_CONFLICT: |
duke@435 | 348 | // A dead location. If it is an oop then we need a NULL to prevent GC from following it |
duke@435 | 349 | *addr = NULL_WORD; |
duke@435 | 350 | break; |
duke@435 | 351 | default: |
duke@435 | 352 | ShouldNotReachHere(); |
duke@435 | 353 | } |
duke@435 | 354 | } |
duke@435 | 355 | |
duke@435 | 356 | if (is_top_frame && JvmtiExport::can_pop_frame() && thread->popframe_forcing_deopt_reexecution()) { |
duke@435 | 357 | // An interpreted frame was popped but it returns to a deoptimized |
duke@435 | 358 | // frame. The incoming arguments to the interpreted activation |
duke@435 | 359 | // were preserved in thread-local storage by the |
duke@435 | 360 | // remove_activation_preserving_args_entry in the interpreter; now |
duke@435 | 361 | // we put them back into the just-unpacked interpreter frame. |
duke@435 | 362 | // Note that this assumes that the locals arena grows toward lower |
duke@435 | 363 | // addresses. |
duke@435 | 364 | if (popframe_preserved_args_size_in_words != 0) { |
duke@435 | 365 | void* saved_args = thread->popframe_preserved_args(); |
duke@435 | 366 | assert(saved_args != NULL, "must have been saved by interpreter"); |
duke@435 | 367 | #ifdef ASSERT |
duke@435 | 368 | assert(popframe_preserved_args_size_in_words <= |
twisti@1861 | 369 | iframe()->interpreter_frame_expression_stack_size()*Interpreter::stackElementWords, |
duke@435 | 370 | "expression stack size should have been extended"); |
duke@435 | 371 | #endif // ASSERT |
duke@435 | 372 | int top_element = iframe()->interpreter_frame_expression_stack_size()-1; |
duke@435 | 373 | intptr_t* base; |
duke@435 | 374 | if (frame::interpreter_frame_expression_stack_direction() < 0) { |
duke@435 | 375 | base = iframe()->interpreter_frame_expression_stack_at(top_element); |
duke@435 | 376 | } else { |
duke@435 | 377 | base = iframe()->interpreter_frame_expression_stack(); |
duke@435 | 378 | } |
kvn@1958 | 379 | Copy::conjoint_jbytes(saved_args, |
kvn@1958 | 380 | base, |
kvn@1958 | 381 | popframe_preserved_args_size_in_bytes); |
duke@435 | 382 | thread->popframe_free_preserved_args(); |
duke@435 | 383 | } |
duke@435 | 384 | } |
duke@435 | 385 | |
duke@435 | 386 | #ifndef PRODUCT |
duke@435 | 387 | if (TraceDeoptimization && Verbose) { |
duke@435 | 388 | ttyLocker ttyl; |
duke@435 | 389 | tty->print_cr("[%d Interpreted Frame]", ++unpack_counter); |
duke@435 | 390 | iframe()->print_on(tty); |
duke@435 | 391 | RegisterMap map(thread); |
duke@435 | 392 | vframe* f = vframe::new_vframe(iframe(), &map, thread); |
duke@435 | 393 | f->print(); |
duke@435 | 394 | |
duke@435 | 395 | tty->print_cr("locals size %d", locals()->size()); |
duke@435 | 396 | tty->print_cr("expression size %d", expressions()->size()); |
duke@435 | 397 | |
duke@435 | 398 | method()->print_value(); |
duke@435 | 399 | tty->cr(); |
duke@435 | 400 | // method()->print_codes(); |
duke@435 | 401 | } else if (TraceDeoptimization) { |
duke@435 | 402 | tty->print(" "); |
duke@435 | 403 | method()->print_value(); |
never@2462 | 404 | Bytecodes::Code code = Bytecodes::java_code_at(method(), bcp); |
duke@435 | 405 | int bci = method()->bci_from(bcp); |
duke@435 | 406 | tty->print(" - %s", Bytecodes::name(code)); |
duke@435 | 407 | tty->print(" @ bci %d ", bci); |
duke@435 | 408 | tty->print_cr("sp = " PTR_FORMAT, iframe()->sp()); |
duke@435 | 409 | } |
duke@435 | 410 | #endif // PRODUCT |
duke@435 | 411 | |
duke@435 | 412 | // The expression stack and locals are in the resource area don't leave |
duke@435 | 413 | // a dangling pointer in the vframeArray we leave around for debug |
duke@435 | 414 | // purposes |
duke@435 | 415 | |
duke@435 | 416 | _locals = _expressions = NULL; |
duke@435 | 417 | |
duke@435 | 418 | } |
duke@435 | 419 | |
never@2901 | 420 | int vframeArrayElement::on_stack_size(int caller_actual_parameters, |
never@2901 | 421 | int callee_parameters, |
duke@435 | 422 | int callee_locals, |
duke@435 | 423 | bool is_top_frame, |
duke@435 | 424 | int popframe_extra_stack_expression_els) const { |
duke@435 | 425 | assert(method()->max_locals() == locals()->size(), "just checking"); |
duke@435 | 426 | int locks = monitors() == NULL ? 0 : monitors()->number_of_monitors(); |
duke@435 | 427 | int temps = expressions()->size(); |
duke@435 | 428 | return Interpreter::size_activation(method(), |
duke@435 | 429 | temps + callee_parameters, |
duke@435 | 430 | popframe_extra_stack_expression_els, |
duke@435 | 431 | locks, |
never@2901 | 432 | caller_actual_parameters, |
duke@435 | 433 | callee_parameters, |
duke@435 | 434 | callee_locals, |
duke@435 | 435 | is_top_frame); |
duke@435 | 436 | } |
duke@435 | 437 | |
duke@435 | 438 | |
duke@435 | 439 | |
duke@435 | 440 | vframeArray* vframeArray::allocate(JavaThread* thread, int frame_size, GrowableArray<compiledVFrame*>* chunk, |
duke@435 | 441 | RegisterMap *reg_map, frame sender, frame caller, frame self) { |
duke@435 | 442 | |
duke@435 | 443 | // Allocate the vframeArray |
duke@435 | 444 | vframeArray * result = (vframeArray*) AllocateHeap(sizeof(vframeArray) + // fixed part |
duke@435 | 445 | sizeof(vframeArrayElement) * (chunk->length() - 1), // variable part |
duke@435 | 446 | "vframeArray::allocate"); |
duke@435 | 447 | result->_frames = chunk->length(); |
duke@435 | 448 | result->_owner_thread = thread; |
duke@435 | 449 | result->_sender = sender; |
duke@435 | 450 | result->_caller = caller; |
duke@435 | 451 | result->_original = self; |
duke@435 | 452 | result->set_unroll_block(NULL); // initialize it |
duke@435 | 453 | result->fill_in(thread, frame_size, chunk, reg_map); |
duke@435 | 454 | return result; |
duke@435 | 455 | } |
duke@435 | 456 | |
duke@435 | 457 | void vframeArray::fill_in(JavaThread* thread, |
duke@435 | 458 | int frame_size, |
duke@435 | 459 | GrowableArray<compiledVFrame*>* chunk, |
duke@435 | 460 | const RegisterMap *reg_map) { |
duke@435 | 461 | // Set owner first, it is used when adding monitor chunks |
duke@435 | 462 | |
duke@435 | 463 | _frame_size = frame_size; |
duke@435 | 464 | for(int i = 0; i < chunk->length(); i++) { |
duke@435 | 465 | element(i)->fill_in(chunk->at(i)); |
duke@435 | 466 | } |
duke@435 | 467 | |
duke@435 | 468 | // Copy registers for callee-saved registers |
duke@435 | 469 | if (reg_map != NULL) { |
duke@435 | 470 | for(int i = 0; i < RegisterMap::reg_count; i++) { |
duke@435 | 471 | #ifdef AMD64 |
duke@435 | 472 | // The register map has one entry for every int (32-bit value), so |
duke@435 | 473 | // 64-bit physical registers have two entries in the map, one for |
duke@435 | 474 | // each half. Ignore the high halves of 64-bit registers, just like |
duke@435 | 475 | // frame::oopmapreg_to_location does. |
duke@435 | 476 | // |
duke@435 | 477 | // [phh] FIXME: this is a temporary hack! This code *should* work |
duke@435 | 478 | // correctly w/o this hack, possibly by changing RegisterMap::pd_location |
duke@435 | 479 | // in frame_amd64.cpp and the values of the phantom high half registers |
duke@435 | 480 | // in amd64.ad. |
duke@435 | 481 | // if (VMReg::Name(i) < SharedInfo::stack0 && is_even(i)) { |
duke@435 | 482 | intptr_t* src = (intptr_t*) reg_map->location(VMRegImpl::as_VMReg(i)); |
duke@435 | 483 | _callee_registers[i] = src != NULL ? *src : NULL_WORD; |
duke@435 | 484 | // } else { |
duke@435 | 485 | // jint* src = (jint*) reg_map->location(VMReg::Name(i)); |
duke@435 | 486 | // _callee_registers[i] = src != NULL ? *src : NULL_WORD; |
duke@435 | 487 | // } |
duke@435 | 488 | #else |
duke@435 | 489 | jint* src = (jint*) reg_map->location(VMRegImpl::as_VMReg(i)); |
duke@435 | 490 | _callee_registers[i] = src != NULL ? *src : NULL_WORD; |
duke@435 | 491 | #endif |
duke@435 | 492 | if (src == NULL) { |
duke@435 | 493 | set_location_valid(i, false); |
duke@435 | 494 | } else { |
duke@435 | 495 | set_location_valid(i, true); |
duke@435 | 496 | jint* dst = (jint*) register_location(i); |
duke@435 | 497 | *dst = *src; |
duke@435 | 498 | } |
duke@435 | 499 | } |
duke@435 | 500 | } |
duke@435 | 501 | } |
duke@435 | 502 | |
never@2901 | 503 | void vframeArray::unpack_to_stack(frame &unpack_frame, int exec_mode, int caller_actual_parameters) { |
duke@435 | 504 | // stack picture |
duke@435 | 505 | // unpack_frame |
duke@435 | 506 | // [new interpreter frames ] (frames are skeletal but walkable) |
duke@435 | 507 | // caller_frame |
duke@435 | 508 | // |
duke@435 | 509 | // This routine fills in the missing data for the skeletal interpreter frames |
duke@435 | 510 | // in the above picture. |
duke@435 | 511 | |
duke@435 | 512 | // Find the skeletal interpreter frames to unpack into |
duke@435 | 513 | RegisterMap map(JavaThread::current(), false); |
duke@435 | 514 | // Get the youngest frame we will unpack (last to be unpacked) |
duke@435 | 515 | frame me = unpack_frame.sender(&map); |
duke@435 | 516 | int index; |
duke@435 | 517 | for (index = 0; index < frames(); index++ ) { |
duke@435 | 518 | *element(index)->iframe() = me; |
duke@435 | 519 | // Get the caller frame (possibly skeletal) |
duke@435 | 520 | me = me.sender(&map); |
duke@435 | 521 | } |
duke@435 | 522 | |
duke@435 | 523 | frame caller_frame = me; |
duke@435 | 524 | |
duke@435 | 525 | // Do the unpacking of interpreter frames; the frame at index 0 represents the top activation, so it has no callee |
duke@435 | 526 | |
duke@435 | 527 | // Unpack the frames from the oldest (frames() -1) to the youngest (0) |
duke@435 | 528 | |
duke@435 | 529 | for (index = frames() - 1; index >= 0 ; index--) { |
duke@435 | 530 | int callee_parameters = index == 0 ? 0 : element(index-1)->method()->size_of_parameters(); |
duke@435 | 531 | int callee_locals = index == 0 ? 0 : element(index-1)->method()->max_locals(); |
never@2901 | 532 | element(index)->unpack_on_stack(caller_actual_parameters, |
never@2901 | 533 | callee_parameters, |
duke@435 | 534 | callee_locals, |
duke@435 | 535 | &caller_frame, |
duke@435 | 536 | index == 0, |
duke@435 | 537 | exec_mode); |
duke@435 | 538 | if (index == frames() - 1) { |
duke@435 | 539 | Deoptimization::unwind_callee_save_values(element(index)->iframe(), this); |
duke@435 | 540 | } |
duke@435 | 541 | caller_frame = *element(index)->iframe(); |
never@2901 | 542 | caller_actual_parameters = callee_parameters; |
duke@435 | 543 | } |
duke@435 | 544 | |
duke@435 | 545 | |
duke@435 | 546 | deallocate_monitor_chunks(); |
duke@435 | 547 | } |
duke@435 | 548 | |
duke@435 | 549 | void vframeArray::deallocate_monitor_chunks() { |
duke@435 | 550 | JavaThread* jt = JavaThread::current(); |
duke@435 | 551 | for (int index = 0; index < frames(); index++ ) { |
duke@435 | 552 | element(index)->free_monitors(jt); |
duke@435 | 553 | } |
duke@435 | 554 | } |
duke@435 | 555 | |
duke@435 | 556 | #ifndef PRODUCT |
duke@435 | 557 | |
duke@435 | 558 | bool vframeArray::structural_compare(JavaThread* thread, GrowableArray<compiledVFrame*>* chunk) { |
duke@435 | 559 | if (owner_thread() != thread) return false; |
duke@435 | 560 | int index = 0; |
duke@435 | 561 | #if 0 // FIXME can't do this comparison |
duke@435 | 562 | |
duke@435 | 563 | // Compare only within vframe array. |
duke@435 | 564 | for (deoptimizedVFrame* vf = deoptimizedVFrame::cast(vframe_at(first_index())); vf; vf = vf->deoptimized_sender_or_null()) { |
duke@435 | 565 | if (index >= chunk->length() || !vf->structural_compare(chunk->at(index))) return false; |
duke@435 | 566 | index++; |
duke@435 | 567 | } |
duke@435 | 568 | if (index != chunk->length()) return false; |
duke@435 | 569 | #endif |
duke@435 | 570 | |
duke@435 | 571 | return true; |
duke@435 | 572 | } |
duke@435 | 573 | |
duke@435 | 574 | #endif |
duke@435 | 575 | |
duke@435 | 576 | address vframeArray::register_location(int i) const { |
duke@435 | 577 | assert(0 <= i && i < RegisterMap::reg_count, "index out of bounds"); |
duke@435 | 578 | return (address) & _callee_registers[i]; |
duke@435 | 579 | } |
duke@435 | 580 | |
duke@435 | 581 | |
duke@435 | 582 | #ifndef PRODUCT |
duke@435 | 583 | |
duke@435 | 584 | // Printing |
duke@435 | 585 | |
duke@435 | 586 | // Note: we cannot have print_on as const, as we allocate inside the method |
duke@435 | 587 | void vframeArray::print_on_2(outputStream* st) { |
duke@435 | 588 | st->print_cr(" - sp: " INTPTR_FORMAT, sp()); |
duke@435 | 589 | st->print(" - thread: "); |
duke@435 | 590 | Thread::current()->print(); |
duke@435 | 591 | st->print_cr(" - frame size: %d", frame_size()); |
duke@435 | 592 | for (int index = 0; index < frames() ; index++ ) { |
duke@435 | 593 | element(index)->print(st); |
duke@435 | 594 | } |
duke@435 | 595 | } |
duke@435 | 596 | |
duke@435 | 597 | void vframeArrayElement::print(outputStream* st) { |
kvn@1690 | 598 | st->print_cr(" - interpreter_frame -> sp: " INTPTR_FORMAT, iframe()->sp()); |
duke@435 | 599 | } |
duke@435 | 600 | |
duke@435 | 601 | void vframeArray::print_value_on(outputStream* st) const { |
duke@435 | 602 | st->print_cr("vframeArray [%d] ", frames()); |
duke@435 | 603 | } |
duke@435 | 604 | |
duke@435 | 605 | |
duke@435 | 606 | #endif |