Mon, 13 Sep 2010 12:10:49 -0700
6919069: client compiler needs to capture more profile information for tiered work
Summary: Added profiling of instanceof and aastore.
Reviewed-by: kvn, jrose, never
duke@435 | 1 | /* |
iveresov@2138 | 2 | * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved. |
duke@435 | 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
duke@435 | 4 | * |
duke@435 | 5 | * This code is free software; you can redistribute it and/or modify it |
duke@435 | 6 | * under the terms of the GNU General Public License version 2 only, as |
duke@435 | 7 | * published by the Free Software Foundation. |
duke@435 | 8 | * |
duke@435 | 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
duke@435 | 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
duke@435 | 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
duke@435 | 12 | * version 2 for more details (a copy is included in the LICENSE file that |
duke@435 | 13 | * accompanied this code). |
duke@435 | 14 | * |
duke@435 | 15 | * You should have received a copy of the GNU General Public License version |
duke@435 | 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
duke@435 | 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
duke@435 | 18 | * |
trims@1907 | 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
trims@1907 | 20 | * or visit www.oracle.com if you need additional information or have any |
trims@1907 | 21 | * questions. |
duke@435 | 22 | * |
duke@435 | 23 | */ |
duke@435 | 24 | |
duke@435 | 25 | # include "incls/_precompiled.incl" |
duke@435 | 26 | # include "incls/_c1_LIRGenerator_sparc.cpp.incl" |
duke@435 | 27 | |
duke@435 | 28 | #ifdef ASSERT |
duke@435 | 29 | #define __ gen()->lir(__FILE__, __LINE__)-> |
duke@435 | 30 | #else |
duke@435 | 31 | #define __ gen()->lir()-> |
duke@435 | 32 | #endif |
duke@435 | 33 | |
duke@435 | 34 | void LIRItem::load_byte_item() { |
duke@435 | 35 | // byte loads use same registers as other loads |
duke@435 | 36 | load_item(); |
duke@435 | 37 | } |
duke@435 | 38 | |
duke@435 | 39 | |
duke@435 | 40 | void LIRItem::load_nonconstant() { |
duke@435 | 41 | LIR_Opr r = value()->operand(); |
duke@435 | 42 | if (_gen->can_inline_as_constant(value())) { |
duke@435 | 43 | if (!r->is_constant()) { |
duke@435 | 44 | r = LIR_OprFact::value_type(value()->type()); |
duke@435 | 45 | } |
duke@435 | 46 | _result = r; |
duke@435 | 47 | } else { |
duke@435 | 48 | load_item(); |
duke@435 | 49 | } |
duke@435 | 50 | } |
duke@435 | 51 | |
duke@435 | 52 | |
duke@435 | 53 | //-------------------------------------------------------------- |
duke@435 | 54 | // LIRGenerator |
duke@435 | 55 | //-------------------------------------------------------------- |
duke@435 | 56 | |
duke@435 | 57 | LIR_Opr LIRGenerator::exceptionOopOpr() { return FrameMap::Oexception_opr; } |
duke@435 | 58 | LIR_Opr LIRGenerator::exceptionPcOpr() { return FrameMap::Oissuing_pc_opr; } |
duke@435 | 59 | LIR_Opr LIRGenerator::syncTempOpr() { return new_register(T_OBJECT); } |
duke@435 | 60 | LIR_Opr LIRGenerator::getThreadTemp() { return rlock_callee_saved(T_INT); } |
duke@435 | 61 | |
duke@435 | 62 | LIR_Opr LIRGenerator::result_register_for(ValueType* type, bool callee) { |
duke@435 | 63 | LIR_Opr opr; |
duke@435 | 64 | switch (type->tag()) { |
duke@435 | 65 | case intTag: opr = callee ? FrameMap::I0_opr : FrameMap::O0_opr; break; |
duke@435 | 66 | case objectTag: opr = callee ? FrameMap::I0_oop_opr : FrameMap::O0_oop_opr; break; |
duke@435 | 67 | case longTag: opr = callee ? FrameMap::in_long_opr : FrameMap::out_long_opr; break; |
duke@435 | 68 | case floatTag: opr = FrameMap::F0_opr; break; |
duke@435 | 69 | case doubleTag: opr = FrameMap::F0_double_opr; break; |
duke@435 | 70 | |
duke@435 | 71 | case addressTag: |
duke@435 | 72 | default: ShouldNotReachHere(); return LIR_OprFact::illegalOpr; |
duke@435 | 73 | } |
duke@435 | 74 | |
duke@435 | 75 | assert(opr->type_field() == as_OprType(as_BasicType(type)), "type mismatch"); |
duke@435 | 76 | return opr; |
duke@435 | 77 | } |
duke@435 | 78 | |
duke@435 | 79 | LIR_Opr LIRGenerator::rlock_callee_saved(BasicType type) { |
duke@435 | 80 | LIR_Opr reg = new_register(type); |
duke@435 | 81 | set_vreg_flag(reg, callee_saved); |
duke@435 | 82 | return reg; |
duke@435 | 83 | } |
duke@435 | 84 | |
duke@435 | 85 | |
duke@435 | 86 | LIR_Opr LIRGenerator::rlock_byte(BasicType type) { |
duke@435 | 87 | return new_register(T_INT); |
duke@435 | 88 | } |
duke@435 | 89 | |
duke@435 | 90 | |
duke@435 | 91 | |
duke@435 | 92 | |
duke@435 | 93 | |
duke@435 | 94 | //--------- loading items into registers -------------------------------- |
duke@435 | 95 | |
duke@435 | 96 | // SPARC cannot inline all constants |
duke@435 | 97 | bool LIRGenerator::can_store_as_constant(Value v, BasicType type) const { |
duke@435 | 98 | if (v->type()->as_IntConstant() != NULL) { |
duke@435 | 99 | return v->type()->as_IntConstant()->value() == 0; |
duke@435 | 100 | } else if (v->type()->as_LongConstant() != NULL) { |
duke@435 | 101 | return v->type()->as_LongConstant()->value() == 0L; |
duke@435 | 102 | } else if (v->type()->as_ObjectConstant() != NULL) { |
duke@435 | 103 | return v->type()->as_ObjectConstant()->value()->is_null_object(); |
duke@435 | 104 | } else { |
duke@435 | 105 | return false; |
duke@435 | 106 | } |
duke@435 | 107 | } |
duke@435 | 108 | |
duke@435 | 109 | |
duke@435 | 110 | // only simm13 constants can be inlined |
duke@435 | 111 | bool LIRGenerator:: can_inline_as_constant(Value i) const { |
duke@435 | 112 | if (i->type()->as_IntConstant() != NULL) { |
duke@435 | 113 | return Assembler::is_simm13(i->type()->as_IntConstant()->value()); |
duke@435 | 114 | } else { |
duke@435 | 115 | return can_store_as_constant(i, as_BasicType(i->type())); |
duke@435 | 116 | } |
duke@435 | 117 | } |
duke@435 | 118 | |
duke@435 | 119 | |
duke@435 | 120 | bool LIRGenerator:: can_inline_as_constant(LIR_Const* c) const { |
duke@435 | 121 | if (c->type() == T_INT) { |
duke@435 | 122 | return Assembler::is_simm13(c->as_jint()); |
duke@435 | 123 | } |
duke@435 | 124 | return false; |
duke@435 | 125 | } |
duke@435 | 126 | |
duke@435 | 127 | |
duke@435 | 128 | LIR_Opr LIRGenerator::safepoint_poll_register() { |
duke@435 | 129 | return new_register(T_INT); |
duke@435 | 130 | } |
duke@435 | 131 | |
duke@435 | 132 | |
duke@435 | 133 | |
duke@435 | 134 | LIR_Address* LIRGenerator::generate_address(LIR_Opr base, LIR_Opr index, |
duke@435 | 135 | int shift, int disp, BasicType type) { |
duke@435 | 136 | assert(base->is_register(), "must be"); |
duke@435 | 137 | |
duke@435 | 138 | // accumulate fixed displacements |
duke@435 | 139 | if (index->is_constant()) { |
duke@435 | 140 | disp += index->as_constant_ptr()->as_jint() << shift; |
duke@435 | 141 | index = LIR_OprFact::illegalOpr; |
duke@435 | 142 | } |
duke@435 | 143 | |
duke@435 | 144 | if (index->is_register()) { |
duke@435 | 145 | // apply the shift and accumulate the displacement |
duke@435 | 146 | if (shift > 0) { |
roland@1495 | 147 | LIR_Opr tmp = new_pointer_register(); |
duke@435 | 148 | __ shift_left(index, shift, tmp); |
duke@435 | 149 | index = tmp; |
duke@435 | 150 | } |
duke@435 | 151 | if (disp != 0) { |
roland@1495 | 152 | LIR_Opr tmp = new_pointer_register(); |
duke@435 | 153 | if (Assembler::is_simm13(disp)) { |
roland@1495 | 154 | __ add(tmp, LIR_OprFact::intptrConst(disp), tmp); |
duke@435 | 155 | index = tmp; |
duke@435 | 156 | } else { |
roland@1495 | 157 | __ move(LIR_OprFact::intptrConst(disp), tmp); |
duke@435 | 158 | __ add(tmp, index, tmp); |
duke@435 | 159 | index = tmp; |
duke@435 | 160 | } |
duke@435 | 161 | disp = 0; |
duke@435 | 162 | } |
duke@435 | 163 | } else if (disp != 0 && !Assembler::is_simm13(disp)) { |
duke@435 | 164 | // index is illegal so replace it with the displacement loaded into a register |
roland@1495 | 165 | index = new_pointer_register(); |
roland@1495 | 166 | __ move(LIR_OprFact::intptrConst(disp), index); |
duke@435 | 167 | disp = 0; |
duke@435 | 168 | } |
duke@435 | 169 | |
duke@435 | 170 | // at this point we either have base + index or base + displacement |
duke@435 | 171 | if (disp == 0) { |
duke@435 | 172 | return new LIR_Address(base, index, type); |
duke@435 | 173 | } else { |
duke@435 | 174 | assert(Assembler::is_simm13(disp), "must be"); |
duke@435 | 175 | return new LIR_Address(base, disp, type); |
duke@435 | 176 | } |
duke@435 | 177 | } |
duke@435 | 178 | |
duke@435 | 179 | |
duke@435 | 180 | LIR_Address* LIRGenerator::emit_array_address(LIR_Opr array_opr, LIR_Opr index_opr, |
duke@435 | 181 | BasicType type, bool needs_card_mark) { |
kvn@464 | 182 | int elem_size = type2aelembytes(type); |
duke@435 | 183 | int shift = exact_log2(elem_size); |
duke@435 | 184 | |
duke@435 | 185 | LIR_Opr base_opr; |
duke@435 | 186 | int offset = arrayOopDesc::base_offset_in_bytes(type); |
duke@435 | 187 | |
duke@435 | 188 | if (index_opr->is_constant()) { |
duke@435 | 189 | int i = index_opr->as_constant_ptr()->as_jint(); |
duke@435 | 190 | int array_offset = i * elem_size; |
duke@435 | 191 | if (Assembler::is_simm13(array_offset + offset)) { |
duke@435 | 192 | base_opr = array_opr; |
duke@435 | 193 | offset = array_offset + offset; |
duke@435 | 194 | } else { |
duke@435 | 195 | base_opr = new_pointer_register(); |
duke@435 | 196 | if (Assembler::is_simm13(array_offset)) { |
duke@435 | 197 | __ add(array_opr, LIR_OprFact::intptrConst(array_offset), base_opr); |
duke@435 | 198 | } else { |
duke@435 | 199 | __ move(LIR_OprFact::intptrConst(array_offset), base_opr); |
duke@435 | 200 | __ add(base_opr, array_opr, base_opr); |
duke@435 | 201 | } |
duke@435 | 202 | } |
duke@435 | 203 | } else { |
duke@435 | 204 | #ifdef _LP64 |
duke@435 | 205 | if (index_opr->type() == T_INT) { |
duke@435 | 206 | LIR_Opr tmp = new_register(T_LONG); |
duke@435 | 207 | __ convert(Bytecodes::_i2l, index_opr, tmp); |
duke@435 | 208 | index_opr = tmp; |
duke@435 | 209 | } |
duke@435 | 210 | #endif |
duke@435 | 211 | |
duke@435 | 212 | base_opr = new_pointer_register(); |
duke@435 | 213 | assert (index_opr->is_register(), "Must be register"); |
duke@435 | 214 | if (shift > 0) { |
duke@435 | 215 | __ shift_left(index_opr, shift, base_opr); |
duke@435 | 216 | __ add(base_opr, array_opr, base_opr); |
duke@435 | 217 | } else { |
duke@435 | 218 | __ add(index_opr, array_opr, base_opr); |
duke@435 | 219 | } |
duke@435 | 220 | } |
duke@435 | 221 | if (needs_card_mark) { |
duke@435 | 222 | LIR_Opr ptr = new_pointer_register(); |
duke@435 | 223 | __ add(base_opr, LIR_OprFact::intptrConst(offset), ptr); |
iveresov@1927 | 224 | return new LIR_Address(ptr, type); |
duke@435 | 225 | } else { |
duke@435 | 226 | return new LIR_Address(base_opr, offset, type); |
duke@435 | 227 | } |
duke@435 | 228 | } |
duke@435 | 229 | |
iveresov@2138 | 230 | LIR_Opr LIRGenerator::load_immediate(int x, BasicType type) { |
iveresov@2138 | 231 | LIR_Opr r; |
iveresov@2138 | 232 | if (type == T_LONG) { |
iveresov@2138 | 233 | r = LIR_OprFact::longConst(x); |
iveresov@2138 | 234 | } else if (type == T_INT) { |
iveresov@2138 | 235 | r = LIR_OprFact::intConst(x); |
iveresov@2138 | 236 | } else { |
iveresov@2138 | 237 | ShouldNotReachHere(); |
iveresov@2138 | 238 | } |
iveresov@2138 | 239 | if (!Assembler::is_simm13(x)) { |
iveresov@2138 | 240 | LIR_Opr tmp = new_register(type); |
iveresov@2138 | 241 | __ move(r, tmp); |
iveresov@2138 | 242 | return tmp; |
iveresov@2138 | 243 | } |
iveresov@2138 | 244 | return r; |
iveresov@2138 | 245 | } |
duke@435 | 246 | |
iveresov@2138 | 247 | void LIRGenerator::increment_counter(address counter, BasicType type, int step) { |
duke@435 | 248 | LIR_Opr pointer = new_pointer_register(); |
duke@435 | 249 | __ move(LIR_OprFact::intptrConst(counter), pointer); |
iveresov@2138 | 250 | LIR_Address* addr = new LIR_Address(pointer, type); |
duke@435 | 251 | increment_counter(addr, step); |
duke@435 | 252 | } |
duke@435 | 253 | |
duke@435 | 254 | void LIRGenerator::increment_counter(LIR_Address* addr, int step) { |
iveresov@2138 | 255 | LIR_Opr temp = new_register(addr->type()); |
duke@435 | 256 | __ move(addr, temp); |
iveresov@2138 | 257 | __ add(temp, load_immediate(step, addr->type()), temp); |
duke@435 | 258 | __ move(temp, addr); |
duke@435 | 259 | } |
duke@435 | 260 | |
duke@435 | 261 | void LIRGenerator::cmp_mem_int(LIR_Condition condition, LIR_Opr base, int disp, int c, CodeEmitInfo* info) { |
duke@435 | 262 | LIR_Opr o7opr = FrameMap::O7_opr; |
duke@435 | 263 | __ load(new LIR_Address(base, disp, T_INT), o7opr, info); |
duke@435 | 264 | __ cmp(condition, o7opr, c); |
duke@435 | 265 | } |
duke@435 | 266 | |
duke@435 | 267 | |
duke@435 | 268 | void LIRGenerator::cmp_reg_mem(LIR_Condition condition, LIR_Opr reg, LIR_Opr base, int disp, BasicType type, CodeEmitInfo* info) { |
duke@435 | 269 | LIR_Opr o7opr = FrameMap::O7_opr; |
duke@435 | 270 | __ load(new LIR_Address(base, disp, type), o7opr, info); |
duke@435 | 271 | __ cmp(condition, reg, o7opr); |
duke@435 | 272 | } |
duke@435 | 273 | |
duke@435 | 274 | |
duke@435 | 275 | void LIRGenerator::cmp_reg_mem(LIR_Condition condition, LIR_Opr reg, LIR_Opr base, LIR_Opr disp, BasicType type, CodeEmitInfo* info) { |
duke@435 | 276 | LIR_Opr o7opr = FrameMap::O7_opr; |
duke@435 | 277 | __ load(new LIR_Address(base, disp, type), o7opr, info); |
duke@435 | 278 | __ cmp(condition, reg, o7opr); |
duke@435 | 279 | } |
duke@435 | 280 | |
duke@435 | 281 | |
duke@435 | 282 | bool LIRGenerator::strength_reduce_multiply(LIR_Opr left, int c, LIR_Opr result, LIR_Opr tmp) { |
duke@435 | 283 | assert(left != result, "should be different registers"); |
duke@435 | 284 | if (is_power_of_2(c + 1)) { |
duke@435 | 285 | __ shift_left(left, log2_intptr(c + 1), result); |
duke@435 | 286 | __ sub(result, left, result); |
duke@435 | 287 | return true; |
duke@435 | 288 | } else if (is_power_of_2(c - 1)) { |
duke@435 | 289 | __ shift_left(left, log2_intptr(c - 1), result); |
duke@435 | 290 | __ add(result, left, result); |
duke@435 | 291 | return true; |
duke@435 | 292 | } |
duke@435 | 293 | return false; |
duke@435 | 294 | } |
duke@435 | 295 | |
duke@435 | 296 | |
duke@435 | 297 | void LIRGenerator::store_stack_parameter (LIR_Opr item, ByteSize offset_from_sp) { |
duke@435 | 298 | BasicType t = item->type(); |
duke@435 | 299 | LIR_Opr sp_opr = FrameMap::SP_opr; |
duke@435 | 300 | if ((t == T_LONG || t == T_DOUBLE) && |
duke@435 | 301 | ((in_bytes(offset_from_sp) - STACK_BIAS) % 8 != 0)) { |
duke@435 | 302 | __ unaligned_move(item, new LIR_Address(sp_opr, in_bytes(offset_from_sp), t)); |
duke@435 | 303 | } else { |
duke@435 | 304 | __ move(item, new LIR_Address(sp_opr, in_bytes(offset_from_sp), t)); |
duke@435 | 305 | } |
duke@435 | 306 | } |
duke@435 | 307 | |
duke@435 | 308 | //---------------------------------------------------------------------- |
duke@435 | 309 | // visitor functions |
duke@435 | 310 | //---------------------------------------------------------------------- |
duke@435 | 311 | |
duke@435 | 312 | |
duke@435 | 313 | void LIRGenerator::do_StoreIndexed(StoreIndexed* x) { |
duke@435 | 314 | assert(x->is_root(),""); |
duke@435 | 315 | bool needs_range_check = true; |
duke@435 | 316 | bool use_length = x->length() != NULL; |
duke@435 | 317 | bool obj_store = x->elt_type() == T_ARRAY || x->elt_type() == T_OBJECT; |
duke@435 | 318 | bool needs_store_check = obj_store && (x->value()->as_Constant() == NULL || |
duke@435 | 319 | !get_jobject_constant(x->value())->is_null_object()); |
duke@435 | 320 | |
duke@435 | 321 | LIRItem array(x->array(), this); |
duke@435 | 322 | LIRItem index(x->index(), this); |
duke@435 | 323 | LIRItem value(x->value(), this); |
duke@435 | 324 | LIRItem length(this); |
duke@435 | 325 | |
duke@435 | 326 | array.load_item(); |
duke@435 | 327 | index.load_nonconstant(); |
duke@435 | 328 | |
duke@435 | 329 | if (use_length) { |
duke@435 | 330 | needs_range_check = x->compute_needs_range_check(); |
duke@435 | 331 | if (needs_range_check) { |
duke@435 | 332 | length.set_instruction(x->length()); |
duke@435 | 333 | length.load_item(); |
duke@435 | 334 | } |
duke@435 | 335 | } |
duke@435 | 336 | if (needs_store_check) { |
duke@435 | 337 | value.load_item(); |
duke@435 | 338 | } else { |
duke@435 | 339 | value.load_for_store(x->elt_type()); |
duke@435 | 340 | } |
duke@435 | 341 | |
duke@435 | 342 | set_no_result(x); |
duke@435 | 343 | |
duke@435 | 344 | // the CodeEmitInfo must be duplicated for each different |
duke@435 | 345 | // LIR-instruction because spilling can occur anywhere between two |
duke@435 | 346 | // instructions and so the debug information must be different |
duke@435 | 347 | CodeEmitInfo* range_check_info = state_for(x); |
duke@435 | 348 | CodeEmitInfo* null_check_info = NULL; |
duke@435 | 349 | if (x->needs_null_check()) { |
duke@435 | 350 | null_check_info = new CodeEmitInfo(range_check_info); |
duke@435 | 351 | } |
duke@435 | 352 | |
duke@435 | 353 | // emit array address setup early so it schedules better |
duke@435 | 354 | LIR_Address* array_addr = emit_array_address(array.result(), index.result(), x->elt_type(), obj_store); |
duke@435 | 355 | |
duke@435 | 356 | if (GenerateRangeChecks && needs_range_check) { |
duke@435 | 357 | if (use_length) { |
duke@435 | 358 | __ cmp(lir_cond_belowEqual, length.result(), index.result()); |
duke@435 | 359 | __ branch(lir_cond_belowEqual, T_INT, new RangeCheckStub(range_check_info, index.result())); |
duke@435 | 360 | } else { |
duke@435 | 361 | array_range_check(array.result(), index.result(), null_check_info, range_check_info); |
duke@435 | 362 | // range_check also does the null check |
duke@435 | 363 | null_check_info = NULL; |
duke@435 | 364 | } |
duke@435 | 365 | } |
duke@435 | 366 | |
duke@435 | 367 | if (GenerateArrayStoreCheck && needs_store_check) { |
duke@435 | 368 | LIR_Opr tmp1 = FrameMap::G1_opr; |
duke@435 | 369 | LIR_Opr tmp2 = FrameMap::G3_opr; |
duke@435 | 370 | LIR_Opr tmp3 = FrameMap::G5_opr; |
duke@435 | 371 | |
duke@435 | 372 | CodeEmitInfo* store_check_info = new CodeEmitInfo(range_check_info); |
duke@435 | 373 | __ store_check(value.result(), array.result(), tmp1, tmp2, tmp3, store_check_info); |
duke@435 | 374 | } |
duke@435 | 375 | |
ysr@777 | 376 | if (obj_store) { |
ysr@777 | 377 | // Needs GC write barriers. |
ysr@777 | 378 | pre_barrier(LIR_OprFact::address(array_addr), false, NULL); |
ysr@777 | 379 | } |
duke@435 | 380 | __ move(value.result(), array_addr, null_check_info); |
duke@435 | 381 | if (obj_store) { |
never@1254 | 382 | // Precise card mark |
duke@435 | 383 | post_barrier(LIR_OprFact::address(array_addr), value.result()); |
duke@435 | 384 | } |
duke@435 | 385 | } |
duke@435 | 386 | |
duke@435 | 387 | |
duke@435 | 388 | void LIRGenerator::do_MonitorEnter(MonitorEnter* x) { |
duke@435 | 389 | assert(x->is_root(),""); |
duke@435 | 390 | LIRItem obj(x->obj(), this); |
duke@435 | 391 | obj.load_item(); |
duke@435 | 392 | |
duke@435 | 393 | set_no_result(x); |
duke@435 | 394 | |
duke@435 | 395 | LIR_Opr lock = FrameMap::G1_opr; |
duke@435 | 396 | LIR_Opr scratch = FrameMap::G3_opr; |
duke@435 | 397 | LIR_Opr hdr = FrameMap::G4_opr; |
duke@435 | 398 | |
duke@435 | 399 | CodeEmitInfo* info_for_exception = NULL; |
duke@435 | 400 | if (x->needs_null_check()) { |
duke@435 | 401 | info_for_exception = state_for(x, x->lock_stack_before()); |
duke@435 | 402 | } |
duke@435 | 403 | |
duke@435 | 404 | // this CodeEmitInfo must not have the xhandlers because here the |
duke@435 | 405 | // object is already locked (xhandlers expects object to be unlocked) |
duke@435 | 406 | CodeEmitInfo* info = state_for(x, x->state(), true); |
duke@435 | 407 | monitor_enter(obj.result(), lock, hdr, scratch, x->monitor_no(), info_for_exception, info); |
duke@435 | 408 | } |
duke@435 | 409 | |
duke@435 | 410 | |
duke@435 | 411 | void LIRGenerator::do_MonitorExit(MonitorExit* x) { |
duke@435 | 412 | assert(x->is_root(),""); |
duke@435 | 413 | LIRItem obj(x->obj(), this); |
duke@435 | 414 | obj.dont_load_item(); |
duke@435 | 415 | |
duke@435 | 416 | set_no_result(x); |
duke@435 | 417 | LIR_Opr lock = FrameMap::G1_opr; |
duke@435 | 418 | LIR_Opr hdr = FrameMap::G3_opr; |
duke@435 | 419 | LIR_Opr obj_temp = FrameMap::G4_opr; |
bobv@2036 | 420 | monitor_exit(obj_temp, lock, hdr, LIR_OprFact::illegalOpr, x->monitor_no()); |
duke@435 | 421 | } |
duke@435 | 422 | |
duke@435 | 423 | |
duke@435 | 424 | // _ineg, _lneg, _fneg, _dneg |
duke@435 | 425 | void LIRGenerator::do_NegateOp(NegateOp* x) { |
duke@435 | 426 | LIRItem value(x->x(), this); |
duke@435 | 427 | value.load_item(); |
duke@435 | 428 | LIR_Opr reg = rlock_result(x); |
duke@435 | 429 | __ negate(value.result(), reg); |
duke@435 | 430 | } |
duke@435 | 431 | |
duke@435 | 432 | |
duke@435 | 433 | |
duke@435 | 434 | // for _fadd, _fmul, _fsub, _fdiv, _frem |
duke@435 | 435 | // _dadd, _dmul, _dsub, _ddiv, _drem |
duke@435 | 436 | void LIRGenerator::do_ArithmeticOp_FPU(ArithmeticOp* x) { |
duke@435 | 437 | switch (x->op()) { |
duke@435 | 438 | case Bytecodes::_fadd: |
duke@435 | 439 | case Bytecodes::_fmul: |
duke@435 | 440 | case Bytecodes::_fsub: |
duke@435 | 441 | case Bytecodes::_fdiv: |
duke@435 | 442 | case Bytecodes::_dadd: |
duke@435 | 443 | case Bytecodes::_dmul: |
duke@435 | 444 | case Bytecodes::_dsub: |
duke@435 | 445 | case Bytecodes::_ddiv: { |
duke@435 | 446 | LIRItem left(x->x(), this); |
duke@435 | 447 | LIRItem right(x->y(), this); |
duke@435 | 448 | left.load_item(); |
duke@435 | 449 | right.load_item(); |
duke@435 | 450 | rlock_result(x); |
duke@435 | 451 | arithmetic_op_fpu(x->op(), x->operand(), left.result(), right.result(), x->is_strictfp()); |
duke@435 | 452 | } |
duke@435 | 453 | break; |
duke@435 | 454 | |
duke@435 | 455 | case Bytecodes::_frem: |
duke@435 | 456 | case Bytecodes::_drem: { |
duke@435 | 457 | address entry; |
duke@435 | 458 | switch (x->op()) { |
duke@435 | 459 | case Bytecodes::_frem: |
duke@435 | 460 | entry = CAST_FROM_FN_PTR(address, SharedRuntime::frem); |
duke@435 | 461 | break; |
duke@435 | 462 | case Bytecodes::_drem: |
duke@435 | 463 | entry = CAST_FROM_FN_PTR(address, SharedRuntime::drem); |
duke@435 | 464 | break; |
duke@435 | 465 | default: |
duke@435 | 466 | ShouldNotReachHere(); |
duke@435 | 467 | } |
duke@435 | 468 | LIR_Opr result = call_runtime(x->x(), x->y(), entry, x->type(), NULL); |
duke@435 | 469 | set_result(x, result); |
duke@435 | 470 | } |
duke@435 | 471 | break; |
duke@435 | 472 | |
duke@435 | 473 | default: ShouldNotReachHere(); |
duke@435 | 474 | } |
duke@435 | 475 | } |
duke@435 | 476 | |
duke@435 | 477 | |
duke@435 | 478 | // for _ladd, _lmul, _lsub, _ldiv, _lrem |
duke@435 | 479 | void LIRGenerator::do_ArithmeticOp_Long(ArithmeticOp* x) { |
duke@435 | 480 | switch (x->op()) { |
duke@435 | 481 | case Bytecodes::_lrem: |
duke@435 | 482 | case Bytecodes::_lmul: |
duke@435 | 483 | case Bytecodes::_ldiv: { |
duke@435 | 484 | |
duke@435 | 485 | if (x->op() == Bytecodes::_ldiv || x->op() == Bytecodes::_lrem) { |
duke@435 | 486 | LIRItem right(x->y(), this); |
duke@435 | 487 | right.load_item(); |
duke@435 | 488 | |
duke@435 | 489 | CodeEmitInfo* info = state_for(x); |
duke@435 | 490 | LIR_Opr item = right.result(); |
duke@435 | 491 | assert(item->is_register(), "must be"); |
duke@435 | 492 | __ cmp(lir_cond_equal, item, LIR_OprFact::longConst(0)); |
duke@435 | 493 | __ branch(lir_cond_equal, T_LONG, new DivByZeroStub(info)); |
duke@435 | 494 | } |
duke@435 | 495 | |
duke@435 | 496 | address entry; |
duke@435 | 497 | switch (x->op()) { |
duke@435 | 498 | case Bytecodes::_lrem: |
duke@435 | 499 | entry = CAST_FROM_FN_PTR(address, SharedRuntime::lrem); |
duke@435 | 500 | break; // check if dividend is 0 is done elsewhere |
duke@435 | 501 | case Bytecodes::_ldiv: |
duke@435 | 502 | entry = CAST_FROM_FN_PTR(address, SharedRuntime::ldiv); |
duke@435 | 503 | break; // check if dividend is 0 is done elsewhere |
duke@435 | 504 | case Bytecodes::_lmul: |
duke@435 | 505 | entry = CAST_FROM_FN_PTR(address, SharedRuntime::lmul); |
duke@435 | 506 | break; |
duke@435 | 507 | default: |
duke@435 | 508 | ShouldNotReachHere(); |
duke@435 | 509 | } |
duke@435 | 510 | |
duke@435 | 511 | // order of arguments to runtime call is reversed. |
duke@435 | 512 | LIR_Opr result = call_runtime(x->y(), x->x(), entry, x->type(), NULL); |
duke@435 | 513 | set_result(x, result); |
duke@435 | 514 | break; |
duke@435 | 515 | } |
duke@435 | 516 | case Bytecodes::_ladd: |
duke@435 | 517 | case Bytecodes::_lsub: { |
duke@435 | 518 | LIRItem left(x->x(), this); |
duke@435 | 519 | LIRItem right(x->y(), this); |
duke@435 | 520 | left.load_item(); |
duke@435 | 521 | right.load_item(); |
duke@435 | 522 | rlock_result(x); |
duke@435 | 523 | |
duke@435 | 524 | arithmetic_op_long(x->op(), x->operand(), left.result(), right.result(), NULL); |
duke@435 | 525 | break; |
duke@435 | 526 | } |
duke@435 | 527 | default: ShouldNotReachHere(); |
duke@435 | 528 | } |
duke@435 | 529 | } |
duke@435 | 530 | |
duke@435 | 531 | |
duke@435 | 532 | // Returns if item is an int constant that can be represented by a simm13 |
duke@435 | 533 | static bool is_simm13(LIR_Opr item) { |
duke@435 | 534 | if (item->is_constant() && item->type() == T_INT) { |
duke@435 | 535 | return Assembler::is_simm13(item->as_constant_ptr()->as_jint()); |
duke@435 | 536 | } else { |
duke@435 | 537 | return false; |
duke@435 | 538 | } |
duke@435 | 539 | } |
duke@435 | 540 | |
duke@435 | 541 | |
duke@435 | 542 | // for: _iadd, _imul, _isub, _idiv, _irem |
duke@435 | 543 | void LIRGenerator::do_ArithmeticOp_Int(ArithmeticOp* x) { |
duke@435 | 544 | bool is_div_rem = x->op() == Bytecodes::_idiv || x->op() == Bytecodes::_irem; |
duke@435 | 545 | LIRItem left(x->x(), this); |
duke@435 | 546 | LIRItem right(x->y(), this); |
duke@435 | 547 | // missing test if instr is commutative and if we should swap |
duke@435 | 548 | right.load_nonconstant(); |
duke@435 | 549 | assert(right.is_constant() || right.is_register(), "wrong state of right"); |
duke@435 | 550 | left.load_item(); |
duke@435 | 551 | rlock_result(x); |
duke@435 | 552 | if (is_div_rem) { |
duke@435 | 553 | CodeEmitInfo* info = state_for(x); |
duke@435 | 554 | LIR_Opr tmp = FrameMap::G1_opr; |
duke@435 | 555 | if (x->op() == Bytecodes::_irem) { |
duke@435 | 556 | __ irem(left.result(), right.result(), x->operand(), tmp, info); |
duke@435 | 557 | } else if (x->op() == Bytecodes::_idiv) { |
duke@435 | 558 | __ idiv(left.result(), right.result(), x->operand(), tmp, info); |
duke@435 | 559 | } |
duke@435 | 560 | } else { |
duke@435 | 561 | arithmetic_op_int(x->op(), x->operand(), left.result(), right.result(), FrameMap::G1_opr); |
duke@435 | 562 | } |
duke@435 | 563 | } |
duke@435 | 564 | |
duke@435 | 565 | |
duke@435 | 566 | void LIRGenerator::do_ArithmeticOp(ArithmeticOp* x) { |
duke@435 | 567 | ValueTag tag = x->type()->tag(); |
duke@435 | 568 | assert(x->x()->type()->tag() == tag && x->y()->type()->tag() == tag, "wrong parameters"); |
duke@435 | 569 | switch (tag) { |
duke@435 | 570 | case floatTag: |
duke@435 | 571 | case doubleTag: do_ArithmeticOp_FPU(x); return; |
duke@435 | 572 | case longTag: do_ArithmeticOp_Long(x); return; |
duke@435 | 573 | case intTag: do_ArithmeticOp_Int(x); return; |
duke@435 | 574 | } |
duke@435 | 575 | ShouldNotReachHere(); |
duke@435 | 576 | } |
duke@435 | 577 | |
duke@435 | 578 | |
duke@435 | 579 | // _ishl, _lshl, _ishr, _lshr, _iushr, _lushr |
duke@435 | 580 | void LIRGenerator::do_ShiftOp(ShiftOp* x) { |
duke@435 | 581 | LIRItem value(x->x(), this); |
duke@435 | 582 | LIRItem count(x->y(), this); |
duke@435 | 583 | // Long shift destroys count register |
duke@435 | 584 | if (value.type()->is_long()) { |
duke@435 | 585 | count.set_destroys_register(); |
duke@435 | 586 | } |
duke@435 | 587 | value.load_item(); |
duke@435 | 588 | // the old backend doesn't support this |
duke@435 | 589 | if (count.is_constant() && count.type()->as_IntConstant() != NULL && value.type()->is_int()) { |
duke@435 | 590 | jint c = count.get_jint_constant() & 0x1f; |
duke@435 | 591 | assert(c >= 0 && c < 32, "should be small"); |
duke@435 | 592 | count.dont_load_item(); |
duke@435 | 593 | } else { |
duke@435 | 594 | count.load_item(); |
duke@435 | 595 | } |
duke@435 | 596 | LIR_Opr reg = rlock_result(x); |
duke@435 | 597 | shift_op(x->op(), reg, value.result(), count.result(), LIR_OprFact::illegalOpr); |
duke@435 | 598 | } |
duke@435 | 599 | |
duke@435 | 600 | |
duke@435 | 601 | // _iand, _land, _ior, _lor, _ixor, _lxor |
duke@435 | 602 | void LIRGenerator::do_LogicOp(LogicOp* x) { |
duke@435 | 603 | LIRItem left(x->x(), this); |
duke@435 | 604 | LIRItem right(x->y(), this); |
duke@435 | 605 | |
duke@435 | 606 | left.load_item(); |
duke@435 | 607 | right.load_nonconstant(); |
duke@435 | 608 | LIR_Opr reg = rlock_result(x); |
duke@435 | 609 | |
duke@435 | 610 | logic_op(x->op(), reg, left.result(), right.result()); |
duke@435 | 611 | } |
duke@435 | 612 | |
duke@435 | 613 | |
duke@435 | 614 | |
duke@435 | 615 | // _lcmp, _fcmpl, _fcmpg, _dcmpl, _dcmpg |
duke@435 | 616 | void LIRGenerator::do_CompareOp(CompareOp* x) { |
duke@435 | 617 | LIRItem left(x->x(), this); |
duke@435 | 618 | LIRItem right(x->y(), this); |
duke@435 | 619 | left.load_item(); |
duke@435 | 620 | right.load_item(); |
duke@435 | 621 | LIR_Opr reg = rlock_result(x); |
duke@435 | 622 | if (x->x()->type()->is_float_kind()) { |
duke@435 | 623 | Bytecodes::Code code = x->op(); |
duke@435 | 624 | __ fcmp2int(left.result(), right.result(), reg, (code == Bytecodes::_fcmpl || code == Bytecodes::_dcmpl)); |
duke@435 | 625 | } else if (x->x()->type()->tag() == longTag) { |
duke@435 | 626 | __ lcmp2int(left.result(), right.result(), reg); |
duke@435 | 627 | } else { |
duke@435 | 628 | Unimplemented(); |
duke@435 | 629 | } |
duke@435 | 630 | } |
duke@435 | 631 | |
duke@435 | 632 | |
duke@435 | 633 | void LIRGenerator::do_AttemptUpdate(Intrinsic* x) { |
duke@435 | 634 | assert(x->number_of_arguments() == 3, "wrong type"); |
duke@435 | 635 | LIRItem obj (x->argument_at(0), this); // AtomicLong object |
duke@435 | 636 | LIRItem cmp_value (x->argument_at(1), this); // value to compare with field |
duke@435 | 637 | LIRItem new_value (x->argument_at(2), this); // replace field with new_value if it matches cmp_value |
duke@435 | 638 | |
duke@435 | 639 | obj.load_item(); |
duke@435 | 640 | cmp_value.load_item(); |
duke@435 | 641 | new_value.load_item(); |
duke@435 | 642 | |
duke@435 | 643 | // generate compare-and-swap and produce zero condition if swap occurs |
duke@435 | 644 | int value_offset = sun_misc_AtomicLongCSImpl::value_offset(); |
duke@435 | 645 | LIR_Opr addr = FrameMap::O7_opr; |
duke@435 | 646 | __ add(obj.result(), LIR_OprFact::intConst(value_offset), addr); |
duke@435 | 647 | LIR_Opr t1 = FrameMap::G1_opr; // temp for 64-bit value |
duke@435 | 648 | LIR_Opr t2 = FrameMap::G3_opr; // temp for 64-bit value |
duke@435 | 649 | __ cas_long(addr, cmp_value.result(), new_value.result(), t1, t2); |
duke@435 | 650 | |
duke@435 | 651 | // generate conditional move of boolean result |
duke@435 | 652 | LIR_Opr result = rlock_result(x); |
duke@435 | 653 | __ cmove(lir_cond_equal, LIR_OprFact::intConst(1), LIR_OprFact::intConst(0), result); |
duke@435 | 654 | } |
duke@435 | 655 | |
duke@435 | 656 | |
duke@435 | 657 | void LIRGenerator::do_CompareAndSwap(Intrinsic* x, ValueType* type) { |
duke@435 | 658 | assert(x->number_of_arguments() == 4, "wrong type"); |
duke@435 | 659 | LIRItem obj (x->argument_at(0), this); // object |
duke@435 | 660 | LIRItem offset(x->argument_at(1), this); // offset of field |
duke@435 | 661 | LIRItem cmp (x->argument_at(2), this); // value to compare with field |
duke@435 | 662 | LIRItem val (x->argument_at(3), this); // replace field with val if matches cmp |
duke@435 | 663 | |
duke@435 | 664 | // Use temps to avoid kills |
duke@435 | 665 | LIR_Opr t1 = FrameMap::G1_opr; |
duke@435 | 666 | LIR_Opr t2 = FrameMap::G3_opr; |
duke@435 | 667 | LIR_Opr addr = new_pointer_register(); |
duke@435 | 668 | |
duke@435 | 669 | // get address of field |
duke@435 | 670 | obj.load_item(); |
duke@435 | 671 | offset.load_item(); |
duke@435 | 672 | cmp.load_item(); |
duke@435 | 673 | val.load_item(); |
duke@435 | 674 | |
duke@435 | 675 | __ add(obj.result(), offset.result(), addr); |
duke@435 | 676 | |
ysr@777 | 677 | if (type == objectType) { // Write-barrier needed for Object fields. |
johnc@1433 | 678 | pre_barrier(addr, false, NULL); |
ysr@777 | 679 | } |
ysr@777 | 680 | |
duke@435 | 681 | if (type == objectType) |
duke@435 | 682 | __ cas_obj(addr, cmp.result(), val.result(), t1, t2); |
duke@435 | 683 | else if (type == intType) |
duke@435 | 684 | __ cas_int(addr, cmp.result(), val.result(), t1, t2); |
duke@435 | 685 | else if (type == longType) |
duke@435 | 686 | __ cas_long(addr, cmp.result(), val.result(), t1, t2); |
duke@435 | 687 | else { |
duke@435 | 688 | ShouldNotReachHere(); |
duke@435 | 689 | } |
duke@435 | 690 | |
duke@435 | 691 | // generate conditional move of boolean result |
duke@435 | 692 | LIR_Opr result = rlock_result(x); |
duke@435 | 693 | __ cmove(lir_cond_equal, LIR_OprFact::intConst(1), LIR_OprFact::intConst(0), result); |
duke@435 | 694 | if (type == objectType) { // Write-barrier needed for Object fields. |
never@1254 | 695 | // Precise card mark since could either be object or array |
ysr@777 | 696 | post_barrier(addr, val.result()); |
duke@435 | 697 | } |
duke@435 | 698 | } |
duke@435 | 699 | |
duke@435 | 700 | |
duke@435 | 701 | void LIRGenerator::do_MathIntrinsic(Intrinsic* x) { |
duke@435 | 702 | switch (x->id()) { |
duke@435 | 703 | case vmIntrinsics::_dabs: |
duke@435 | 704 | case vmIntrinsics::_dsqrt: { |
duke@435 | 705 | assert(x->number_of_arguments() == 1, "wrong type"); |
duke@435 | 706 | LIRItem value(x->argument_at(0), this); |
duke@435 | 707 | value.load_item(); |
duke@435 | 708 | LIR_Opr dst = rlock_result(x); |
duke@435 | 709 | |
duke@435 | 710 | switch (x->id()) { |
duke@435 | 711 | case vmIntrinsics::_dsqrt: { |
duke@435 | 712 | __ sqrt(value.result(), dst, LIR_OprFact::illegalOpr); |
duke@435 | 713 | break; |
duke@435 | 714 | } |
duke@435 | 715 | case vmIntrinsics::_dabs: { |
duke@435 | 716 | __ abs(value.result(), dst, LIR_OprFact::illegalOpr); |
duke@435 | 717 | break; |
duke@435 | 718 | } |
duke@435 | 719 | } |
duke@435 | 720 | break; |
duke@435 | 721 | } |
duke@435 | 722 | case vmIntrinsics::_dlog10: // fall through |
duke@435 | 723 | case vmIntrinsics::_dlog: // fall through |
duke@435 | 724 | case vmIntrinsics::_dsin: // fall through |
duke@435 | 725 | case vmIntrinsics::_dtan: // fall through |
duke@435 | 726 | case vmIntrinsics::_dcos: { |
duke@435 | 727 | assert(x->number_of_arguments() == 1, "wrong type"); |
duke@435 | 728 | |
duke@435 | 729 | address runtime_entry = NULL; |
duke@435 | 730 | switch (x->id()) { |
duke@435 | 731 | case vmIntrinsics::_dsin: |
duke@435 | 732 | runtime_entry = CAST_FROM_FN_PTR(address, SharedRuntime::dsin); |
duke@435 | 733 | break; |
duke@435 | 734 | case vmIntrinsics::_dcos: |
duke@435 | 735 | runtime_entry = CAST_FROM_FN_PTR(address, SharedRuntime::dcos); |
duke@435 | 736 | break; |
duke@435 | 737 | case vmIntrinsics::_dtan: |
duke@435 | 738 | runtime_entry = CAST_FROM_FN_PTR(address, SharedRuntime::dtan); |
duke@435 | 739 | break; |
duke@435 | 740 | case vmIntrinsics::_dlog: |
duke@435 | 741 | runtime_entry = CAST_FROM_FN_PTR(address, SharedRuntime::dlog); |
duke@435 | 742 | break; |
duke@435 | 743 | case vmIntrinsics::_dlog10: |
duke@435 | 744 | runtime_entry = CAST_FROM_FN_PTR(address, SharedRuntime::dlog10); |
duke@435 | 745 | break; |
duke@435 | 746 | default: |
duke@435 | 747 | ShouldNotReachHere(); |
duke@435 | 748 | } |
duke@435 | 749 | |
duke@435 | 750 | LIR_Opr result = call_runtime(x->argument_at(0), runtime_entry, x->type(), NULL); |
duke@435 | 751 | set_result(x, result); |
duke@435 | 752 | } |
duke@435 | 753 | } |
duke@435 | 754 | } |
duke@435 | 755 | |
duke@435 | 756 | |
duke@435 | 757 | void LIRGenerator::do_ArrayCopy(Intrinsic* x) { |
duke@435 | 758 | assert(x->number_of_arguments() == 5, "wrong type"); |
never@1363 | 759 | |
never@1363 | 760 | // Make all state_for calls early since they can emit code |
never@1363 | 761 | CodeEmitInfo* info = state_for(x, x->state()); |
never@1363 | 762 | |
duke@435 | 763 | // Note: spill caller save before setting the item |
duke@435 | 764 | LIRItem src (x->argument_at(0), this); |
duke@435 | 765 | LIRItem src_pos (x->argument_at(1), this); |
duke@435 | 766 | LIRItem dst (x->argument_at(2), this); |
duke@435 | 767 | LIRItem dst_pos (x->argument_at(3), this); |
duke@435 | 768 | LIRItem length (x->argument_at(4), this); |
duke@435 | 769 | // load all values in callee_save_registers, as this makes the |
duke@435 | 770 | // parameter passing to the fast case simpler |
duke@435 | 771 | src.load_item_force (rlock_callee_saved(T_OBJECT)); |
duke@435 | 772 | src_pos.load_item_force (rlock_callee_saved(T_INT)); |
duke@435 | 773 | dst.load_item_force (rlock_callee_saved(T_OBJECT)); |
duke@435 | 774 | dst_pos.load_item_force (rlock_callee_saved(T_INT)); |
duke@435 | 775 | length.load_item_force (rlock_callee_saved(T_INT)); |
duke@435 | 776 | |
duke@435 | 777 | int flags; |
duke@435 | 778 | ciArrayKlass* expected_type; |
duke@435 | 779 | arraycopy_helper(x, &flags, &expected_type); |
duke@435 | 780 | |
duke@435 | 781 | __ arraycopy(src.result(), src_pos.result(), dst.result(), dst_pos.result(), |
duke@435 | 782 | length.result(), rlock_callee_saved(T_INT), |
duke@435 | 783 | expected_type, flags, info); |
duke@435 | 784 | set_no_result(x); |
duke@435 | 785 | } |
duke@435 | 786 | |
duke@435 | 787 | // _i2l, _i2f, _i2d, _l2i, _l2f, _l2d, _f2i, _f2l, _f2d, _d2i, _d2l, _d2f |
duke@435 | 788 | // _i2b, _i2c, _i2s |
duke@435 | 789 | void LIRGenerator::do_Convert(Convert* x) { |
duke@435 | 790 | |
duke@435 | 791 | switch (x->op()) { |
duke@435 | 792 | case Bytecodes::_f2l: |
duke@435 | 793 | case Bytecodes::_d2l: |
duke@435 | 794 | case Bytecodes::_d2i: |
duke@435 | 795 | case Bytecodes::_l2f: |
duke@435 | 796 | case Bytecodes::_l2d: { |
duke@435 | 797 | |
duke@435 | 798 | address entry; |
duke@435 | 799 | switch (x->op()) { |
duke@435 | 800 | case Bytecodes::_l2f: |
duke@435 | 801 | entry = CAST_FROM_FN_PTR(address, SharedRuntime::l2f); |
duke@435 | 802 | break; |
duke@435 | 803 | case Bytecodes::_l2d: |
duke@435 | 804 | entry = CAST_FROM_FN_PTR(address, SharedRuntime::l2d); |
duke@435 | 805 | break; |
duke@435 | 806 | case Bytecodes::_f2l: |
duke@435 | 807 | entry = CAST_FROM_FN_PTR(address, SharedRuntime::f2l); |
duke@435 | 808 | break; |
duke@435 | 809 | case Bytecodes::_d2l: |
duke@435 | 810 | entry = CAST_FROM_FN_PTR(address, SharedRuntime::d2l); |
duke@435 | 811 | break; |
duke@435 | 812 | case Bytecodes::_d2i: |
duke@435 | 813 | entry = CAST_FROM_FN_PTR(address, SharedRuntime::d2i); |
duke@435 | 814 | break; |
duke@435 | 815 | default: |
duke@435 | 816 | ShouldNotReachHere(); |
duke@435 | 817 | } |
duke@435 | 818 | LIR_Opr result = call_runtime(x->value(), entry, x->type(), NULL); |
duke@435 | 819 | set_result(x, result); |
duke@435 | 820 | break; |
duke@435 | 821 | } |
duke@435 | 822 | |
duke@435 | 823 | case Bytecodes::_i2f: |
duke@435 | 824 | case Bytecodes::_i2d: { |
duke@435 | 825 | LIRItem value(x->value(), this); |
duke@435 | 826 | |
duke@435 | 827 | LIR_Opr reg = rlock_result(x); |
duke@435 | 828 | // To convert an int to double, we need to load the 32-bit int |
duke@435 | 829 | // from memory into a single precision floating point register |
duke@435 | 830 | // (even numbered). Then the sparc fitod instruction takes care |
duke@435 | 831 | // of the conversion. This is a bit ugly, but is the best way to |
duke@435 | 832 | // get the int value in a single precision floating point register |
duke@435 | 833 | value.load_item(); |
duke@435 | 834 | LIR_Opr tmp = force_to_spill(value.result(), T_FLOAT); |
duke@435 | 835 | __ convert(x->op(), tmp, reg); |
duke@435 | 836 | break; |
duke@435 | 837 | } |
duke@435 | 838 | break; |
duke@435 | 839 | |
duke@435 | 840 | case Bytecodes::_i2l: |
duke@435 | 841 | case Bytecodes::_i2b: |
duke@435 | 842 | case Bytecodes::_i2c: |
duke@435 | 843 | case Bytecodes::_i2s: |
duke@435 | 844 | case Bytecodes::_l2i: |
duke@435 | 845 | case Bytecodes::_f2d: |
duke@435 | 846 | case Bytecodes::_d2f: { // inline code |
duke@435 | 847 | LIRItem value(x->value(), this); |
duke@435 | 848 | |
duke@435 | 849 | value.load_item(); |
duke@435 | 850 | LIR_Opr reg = rlock_result(x); |
duke@435 | 851 | __ convert(x->op(), value.result(), reg, false); |
duke@435 | 852 | } |
duke@435 | 853 | break; |
duke@435 | 854 | |
duke@435 | 855 | case Bytecodes::_f2i: { |
duke@435 | 856 | LIRItem value (x->value(), this); |
duke@435 | 857 | value.set_destroys_register(); |
duke@435 | 858 | value.load_item(); |
duke@435 | 859 | LIR_Opr reg = rlock_result(x); |
duke@435 | 860 | set_vreg_flag(reg, must_start_in_memory); |
duke@435 | 861 | __ convert(x->op(), value.result(), reg, false); |
duke@435 | 862 | } |
duke@435 | 863 | break; |
duke@435 | 864 | |
duke@435 | 865 | default: ShouldNotReachHere(); |
duke@435 | 866 | } |
duke@435 | 867 | } |
duke@435 | 868 | |
duke@435 | 869 | |
duke@435 | 870 | void LIRGenerator::do_NewInstance(NewInstance* x) { |
duke@435 | 871 | // This instruction can be deoptimized in the slow path : use |
duke@435 | 872 | // O0 as result register. |
duke@435 | 873 | const LIR_Opr reg = result_register_for(x->type()); |
duke@435 | 874 | |
duke@435 | 875 | if (PrintNotLoaded && !x->klass()->is_loaded()) { |
duke@435 | 876 | tty->print_cr(" ###class not loaded at new bci %d", x->bci()); |
duke@435 | 877 | } |
duke@435 | 878 | CodeEmitInfo* info = state_for(x, x->state()); |
duke@435 | 879 | LIR_Opr tmp1 = FrameMap::G1_oop_opr; |
duke@435 | 880 | LIR_Opr tmp2 = FrameMap::G3_oop_opr; |
duke@435 | 881 | LIR_Opr tmp3 = FrameMap::G4_oop_opr; |
duke@435 | 882 | LIR_Opr tmp4 = FrameMap::O1_oop_opr; |
duke@435 | 883 | LIR_Opr klass_reg = FrameMap::G5_oop_opr; |
duke@435 | 884 | new_instance(reg, x->klass(), tmp1, tmp2, tmp3, tmp4, klass_reg, info); |
duke@435 | 885 | LIR_Opr result = rlock_result(x); |
duke@435 | 886 | __ move(reg, result); |
duke@435 | 887 | } |
duke@435 | 888 | |
duke@435 | 889 | |
duke@435 | 890 | void LIRGenerator::do_NewTypeArray(NewTypeArray* x) { |
never@1363 | 891 | // Evaluate state_for early since it may emit code |
never@1363 | 892 | CodeEmitInfo* info = state_for(x, x->state()); |
never@1363 | 893 | |
duke@435 | 894 | LIRItem length(x->length(), this); |
duke@435 | 895 | length.load_item(); |
duke@435 | 896 | |
duke@435 | 897 | LIR_Opr reg = result_register_for(x->type()); |
duke@435 | 898 | LIR_Opr tmp1 = FrameMap::G1_oop_opr; |
duke@435 | 899 | LIR_Opr tmp2 = FrameMap::G3_oop_opr; |
duke@435 | 900 | LIR_Opr tmp3 = FrameMap::G4_oop_opr; |
duke@435 | 901 | LIR_Opr tmp4 = FrameMap::O1_oop_opr; |
duke@435 | 902 | LIR_Opr klass_reg = FrameMap::G5_oop_opr; |
duke@435 | 903 | LIR_Opr len = length.result(); |
duke@435 | 904 | BasicType elem_type = x->elt_type(); |
duke@435 | 905 | |
jrose@1424 | 906 | __ oop2reg(ciTypeArrayKlass::make(elem_type)->constant_encoding(), klass_reg); |
duke@435 | 907 | |
duke@435 | 908 | CodeStub* slow_path = new NewTypeArrayStub(klass_reg, len, reg, info); |
duke@435 | 909 | __ allocate_array(reg, len, tmp1, tmp2, tmp3, tmp4, elem_type, klass_reg, slow_path); |
duke@435 | 910 | |
duke@435 | 911 | LIR_Opr result = rlock_result(x); |
duke@435 | 912 | __ move(reg, result); |
duke@435 | 913 | } |
duke@435 | 914 | |
duke@435 | 915 | |
duke@435 | 916 | void LIRGenerator::do_NewObjectArray(NewObjectArray* x) { |
never@1363 | 917 | // Evaluate state_for early since it may emit code. |
never@1363 | 918 | CodeEmitInfo* info = state_for(x, x->state()); |
duke@435 | 919 | // in case of patching (i.e., object class is not yet loaded), we need to reexecute the instruction |
duke@435 | 920 | // and therefore provide the state before the parameters have been consumed |
duke@435 | 921 | CodeEmitInfo* patching_info = NULL; |
duke@435 | 922 | if (!x->klass()->is_loaded() || PatchALot) { |
duke@435 | 923 | patching_info = state_for(x, x->state_before()); |
duke@435 | 924 | } |
duke@435 | 925 | |
never@1363 | 926 | LIRItem length(x->length(), this); |
duke@435 | 927 | length.load_item(); |
duke@435 | 928 | |
duke@435 | 929 | const LIR_Opr reg = result_register_for(x->type()); |
duke@435 | 930 | LIR_Opr tmp1 = FrameMap::G1_oop_opr; |
duke@435 | 931 | LIR_Opr tmp2 = FrameMap::G3_oop_opr; |
duke@435 | 932 | LIR_Opr tmp3 = FrameMap::G4_oop_opr; |
duke@435 | 933 | LIR_Opr tmp4 = FrameMap::O1_oop_opr; |
duke@435 | 934 | LIR_Opr klass_reg = FrameMap::G5_oop_opr; |
duke@435 | 935 | LIR_Opr len = length.result(); |
duke@435 | 936 | |
duke@435 | 937 | CodeStub* slow_path = new NewObjectArrayStub(klass_reg, len, reg, info); |
duke@435 | 938 | ciObject* obj = (ciObject*) ciObjArrayKlass::make(x->klass()); |
duke@435 | 939 | if (obj == ciEnv::unloaded_ciobjarrayklass()) { |
duke@435 | 940 | BAILOUT("encountered unloaded_ciobjarrayklass due to out of memory error"); |
duke@435 | 941 | } |
duke@435 | 942 | jobject2reg_with_patching(klass_reg, obj, patching_info); |
duke@435 | 943 | __ allocate_array(reg, len, tmp1, tmp2, tmp3, tmp4, T_OBJECT, klass_reg, slow_path); |
duke@435 | 944 | |
duke@435 | 945 | LIR_Opr result = rlock_result(x); |
duke@435 | 946 | __ move(reg, result); |
duke@435 | 947 | } |
duke@435 | 948 | |
duke@435 | 949 | |
duke@435 | 950 | void LIRGenerator::do_NewMultiArray(NewMultiArray* x) { |
duke@435 | 951 | Values* dims = x->dims(); |
duke@435 | 952 | int i = dims->length(); |
duke@435 | 953 | LIRItemList* items = new LIRItemList(dims->length(), NULL); |
duke@435 | 954 | while (i-- > 0) { |
duke@435 | 955 | LIRItem* size = new LIRItem(dims->at(i), this); |
duke@435 | 956 | items->at_put(i, size); |
duke@435 | 957 | } |
duke@435 | 958 | |
never@1363 | 959 | // Evaluate state_for early since it may emit code. |
duke@435 | 960 | CodeEmitInfo* patching_info = NULL; |
duke@435 | 961 | if (!x->klass()->is_loaded() || PatchALot) { |
duke@435 | 962 | patching_info = state_for(x, x->state_before()); |
duke@435 | 963 | |
duke@435 | 964 | // cannot re-use same xhandlers for multiple CodeEmitInfos, so |
never@1368 | 965 | // clone all handlers. This is handled transparently in other |
never@1368 | 966 | // places by the CodeEmitInfo cloning logic but is handled |
never@1368 | 967 | // specially here because a stub isn't being used. |
duke@435 | 968 | x->set_exception_handlers(new XHandlers(x->exception_handlers())); |
duke@435 | 969 | } |
never@1368 | 970 | CodeEmitInfo* info = state_for(x, x->state()); |
duke@435 | 971 | |
duke@435 | 972 | i = dims->length(); |
duke@435 | 973 | while (i-- > 0) { |
duke@435 | 974 | LIRItem* size = items->at(i); |
duke@435 | 975 | size->load_item(); |
duke@435 | 976 | store_stack_parameter (size->result(), |
duke@435 | 977 | in_ByteSize(STACK_BIAS + |
never@739 | 978 | frame::memory_parameter_word_sp_offset * wordSize + |
never@739 | 979 | i * sizeof(jint))); |
duke@435 | 980 | } |
duke@435 | 981 | |
duke@435 | 982 | // This instruction can be deoptimized in the slow path : use |
duke@435 | 983 | // O0 as result register. |
duke@435 | 984 | const LIR_Opr reg = result_register_for(x->type()); |
duke@435 | 985 | jobject2reg_with_patching(reg, x->klass(), patching_info); |
duke@435 | 986 | LIR_Opr rank = FrameMap::O1_opr; |
duke@435 | 987 | __ move(LIR_OprFact::intConst(x->rank()), rank); |
duke@435 | 988 | LIR_Opr varargs = FrameMap::as_pointer_opr(O2); |
duke@435 | 989 | int offset_from_sp = (frame::memory_parameter_word_sp_offset * wordSize) + STACK_BIAS; |
duke@435 | 990 | __ add(FrameMap::SP_opr, |
duke@435 | 991 | LIR_OprFact::intptrConst(offset_from_sp), |
duke@435 | 992 | varargs); |
duke@435 | 993 | LIR_OprList* args = new LIR_OprList(3); |
duke@435 | 994 | args->append(reg); |
duke@435 | 995 | args->append(rank); |
duke@435 | 996 | args->append(varargs); |
duke@435 | 997 | __ call_runtime(Runtime1::entry_for(Runtime1::new_multi_array_id), |
duke@435 | 998 | LIR_OprFact::illegalOpr, |
duke@435 | 999 | reg, args, info); |
duke@435 | 1000 | |
duke@435 | 1001 | LIR_Opr result = rlock_result(x); |
duke@435 | 1002 | __ move(reg, result); |
duke@435 | 1003 | } |
duke@435 | 1004 | |
duke@435 | 1005 | |
duke@435 | 1006 | void LIRGenerator::do_BlockBegin(BlockBegin* x) { |
duke@435 | 1007 | } |
duke@435 | 1008 | |
duke@435 | 1009 | |
duke@435 | 1010 | void LIRGenerator::do_CheckCast(CheckCast* x) { |
duke@435 | 1011 | LIRItem obj(x->obj(), this); |
duke@435 | 1012 | CodeEmitInfo* patching_info = NULL; |
duke@435 | 1013 | if (!x->klass()->is_loaded() || (PatchALot && !x->is_incompatible_class_change_check())) { |
duke@435 | 1014 | // must do this before locking the destination register as an oop register, |
duke@435 | 1015 | // and before the obj is loaded (so x->obj()->item() is valid for creating a debug info location) |
duke@435 | 1016 | patching_info = state_for(x, x->state_before()); |
duke@435 | 1017 | } |
duke@435 | 1018 | obj.load_item(); |
duke@435 | 1019 | LIR_Opr out_reg = rlock_result(x); |
duke@435 | 1020 | CodeStub* stub; |
duke@435 | 1021 | CodeEmitInfo* info_for_exception = state_for(x, x->state()->copy_locks()); |
duke@435 | 1022 | |
duke@435 | 1023 | if (x->is_incompatible_class_change_check()) { |
duke@435 | 1024 | assert(patching_info == NULL, "can't patch this"); |
duke@435 | 1025 | stub = new SimpleExceptionStub(Runtime1::throw_incompatible_class_change_error_id, LIR_OprFact::illegalOpr, info_for_exception); |
duke@435 | 1026 | } else { |
duke@435 | 1027 | stub = new SimpleExceptionStub(Runtime1::throw_class_cast_exception_id, obj.result(), info_for_exception); |
duke@435 | 1028 | } |
duke@435 | 1029 | LIR_Opr tmp1 = FrameMap::G1_oop_opr; |
duke@435 | 1030 | LIR_Opr tmp2 = FrameMap::G3_oop_opr; |
duke@435 | 1031 | LIR_Opr tmp3 = FrameMap::G4_oop_opr; |
duke@435 | 1032 | __ checkcast(out_reg, obj.result(), x->klass(), tmp1, tmp2, tmp3, |
duke@435 | 1033 | x->direct_compare(), info_for_exception, patching_info, stub, |
duke@435 | 1034 | x->profiled_method(), x->profiled_bci()); |
duke@435 | 1035 | } |
duke@435 | 1036 | |
duke@435 | 1037 | |
duke@435 | 1038 | void LIRGenerator::do_InstanceOf(InstanceOf* x) { |
duke@435 | 1039 | LIRItem obj(x->obj(), this); |
duke@435 | 1040 | CodeEmitInfo* patching_info = NULL; |
duke@435 | 1041 | if (!x->klass()->is_loaded() || PatchALot) { |
duke@435 | 1042 | patching_info = state_for(x, x->state_before()); |
duke@435 | 1043 | } |
duke@435 | 1044 | // ensure the result register is not the input register because the result is initialized before the patching safepoint |
duke@435 | 1045 | obj.load_item(); |
duke@435 | 1046 | LIR_Opr out_reg = rlock_result(x); |
duke@435 | 1047 | LIR_Opr tmp1 = FrameMap::G1_oop_opr; |
duke@435 | 1048 | LIR_Opr tmp2 = FrameMap::G3_oop_opr; |
duke@435 | 1049 | LIR_Opr tmp3 = FrameMap::G4_oop_opr; |
iveresov@2146 | 1050 | __ instanceof(out_reg, obj.result(), x->klass(), tmp1, tmp2, tmp3, |
iveresov@2146 | 1051 | x->direct_compare(), patching_info, |
iveresov@2146 | 1052 | x->profiled_method(), x->profiled_bci()); |
duke@435 | 1053 | } |
duke@435 | 1054 | |
duke@435 | 1055 | |
duke@435 | 1056 | void LIRGenerator::do_If(If* x) { |
duke@435 | 1057 | assert(x->number_of_sux() == 2, "inconsistency"); |
duke@435 | 1058 | ValueTag tag = x->x()->type()->tag(); |
duke@435 | 1059 | LIRItem xitem(x->x(), this); |
duke@435 | 1060 | LIRItem yitem(x->y(), this); |
duke@435 | 1061 | LIRItem* xin = &xitem; |
duke@435 | 1062 | LIRItem* yin = &yitem; |
duke@435 | 1063 | If::Condition cond = x->cond(); |
duke@435 | 1064 | |
duke@435 | 1065 | if (tag == longTag) { |
duke@435 | 1066 | // for longs, only conditions "eql", "neq", "lss", "geq" are valid; |
duke@435 | 1067 | // mirror for other conditions |
duke@435 | 1068 | if (cond == If::gtr || cond == If::leq) { |
duke@435 | 1069 | // swap inputs |
duke@435 | 1070 | cond = Instruction::mirror(cond); |
duke@435 | 1071 | xin = &yitem; |
duke@435 | 1072 | yin = &xitem; |
duke@435 | 1073 | } |
duke@435 | 1074 | xin->set_destroys_register(); |
duke@435 | 1075 | } |
duke@435 | 1076 | |
duke@435 | 1077 | LIR_Opr left = LIR_OprFact::illegalOpr; |
duke@435 | 1078 | LIR_Opr right = LIR_OprFact::illegalOpr; |
duke@435 | 1079 | |
duke@435 | 1080 | xin->load_item(); |
duke@435 | 1081 | left = xin->result(); |
duke@435 | 1082 | |
duke@435 | 1083 | if (is_simm13(yin->result())) { |
duke@435 | 1084 | // inline int constants which are small enough to be immediate operands |
duke@435 | 1085 | right = LIR_OprFact::value_type(yin->value()->type()); |
duke@435 | 1086 | } else if (tag == longTag && yin->is_constant() && yin->get_jlong_constant() == 0 && |
duke@435 | 1087 | (cond == If::eql || cond == If::neq)) { |
duke@435 | 1088 | // inline long zero |
duke@435 | 1089 | right = LIR_OprFact::value_type(yin->value()->type()); |
duke@435 | 1090 | } else if (tag == objectTag && yin->is_constant() && (yin->get_jobject_constant()->is_null_object())) { |
duke@435 | 1091 | right = LIR_OprFact::value_type(yin->value()->type()); |
duke@435 | 1092 | } else { |
duke@435 | 1093 | yin->load_item(); |
duke@435 | 1094 | right = yin->result(); |
duke@435 | 1095 | } |
duke@435 | 1096 | set_no_result(x); |
duke@435 | 1097 | |
duke@435 | 1098 | // add safepoint before generating condition code so it can be recomputed |
duke@435 | 1099 | if (x->is_safepoint()) { |
duke@435 | 1100 | // increment backedge counter if needed |
iveresov@2138 | 1101 | increment_backedge_counter(state_for(x, x->state_before()), x->profiled_bci()); |
duke@435 | 1102 | __ safepoint(new_register(T_INT), state_for(x, x->state_before())); |
duke@435 | 1103 | } |
duke@435 | 1104 | |
duke@435 | 1105 | __ cmp(lir_cond(cond), left, right); |
iveresov@2138 | 1106 | // Generate branch profiling. Profiling code doesn't kill flags. |
duke@435 | 1107 | profile_branch(x, cond); |
duke@435 | 1108 | move_to_phi(x->state()); |
duke@435 | 1109 | if (x->x()->type()->is_float_kind()) { |
duke@435 | 1110 | __ branch(lir_cond(cond), right->type(), x->tsux(), x->usux()); |
duke@435 | 1111 | } else { |
duke@435 | 1112 | __ branch(lir_cond(cond), right->type(), x->tsux()); |
duke@435 | 1113 | } |
duke@435 | 1114 | assert(x->default_sux() == x->fsux(), "wrong destination above"); |
duke@435 | 1115 | __ jump(x->default_sux()); |
duke@435 | 1116 | } |
duke@435 | 1117 | |
duke@435 | 1118 | |
duke@435 | 1119 | LIR_Opr LIRGenerator::getThreadPointer() { |
duke@435 | 1120 | return FrameMap::as_pointer_opr(G2); |
duke@435 | 1121 | } |
duke@435 | 1122 | |
duke@435 | 1123 | |
duke@435 | 1124 | void LIRGenerator::trace_block_entry(BlockBegin* block) { |
duke@435 | 1125 | __ move(LIR_OprFact::intConst(block->block_id()), FrameMap::O0_opr); |
duke@435 | 1126 | LIR_OprList* args = new LIR_OprList(1); |
duke@435 | 1127 | args->append(FrameMap::O0_opr); |
duke@435 | 1128 | address func = CAST_FROM_FN_PTR(address, Runtime1::trace_block_entry); |
duke@435 | 1129 | __ call_runtime_leaf(func, rlock_callee_saved(T_INT), LIR_OprFact::illegalOpr, args); |
duke@435 | 1130 | } |
duke@435 | 1131 | |
duke@435 | 1132 | |
duke@435 | 1133 | void LIRGenerator::volatile_field_store(LIR_Opr value, LIR_Address* address, |
duke@435 | 1134 | CodeEmitInfo* info) { |
duke@435 | 1135 | #ifdef _LP64 |
duke@435 | 1136 | __ store(value, address, info); |
duke@435 | 1137 | #else |
duke@435 | 1138 | __ volatile_store_mem_reg(value, address, info); |
duke@435 | 1139 | #endif |
duke@435 | 1140 | } |
duke@435 | 1141 | |
duke@435 | 1142 | void LIRGenerator::volatile_field_load(LIR_Address* address, LIR_Opr result, |
duke@435 | 1143 | CodeEmitInfo* info) { |
duke@435 | 1144 | #ifdef _LP64 |
duke@435 | 1145 | __ load(address, result, info); |
duke@435 | 1146 | #else |
duke@435 | 1147 | __ volatile_load_mem_reg(address, result, info); |
duke@435 | 1148 | #endif |
duke@435 | 1149 | } |
duke@435 | 1150 | |
duke@435 | 1151 | |
duke@435 | 1152 | void LIRGenerator::put_Object_unsafe(LIR_Opr src, LIR_Opr offset, LIR_Opr data, |
duke@435 | 1153 | BasicType type, bool is_volatile) { |
duke@435 | 1154 | LIR_Opr base_op = src; |
duke@435 | 1155 | LIR_Opr index_op = offset; |
duke@435 | 1156 | |
duke@435 | 1157 | bool is_obj = (type == T_ARRAY || type == T_OBJECT); |
duke@435 | 1158 | #ifndef _LP64 |
duke@435 | 1159 | if (is_volatile && type == T_LONG) { |
duke@435 | 1160 | __ volatile_store_unsafe_reg(data, src, offset, type, NULL, lir_patch_none); |
duke@435 | 1161 | } else |
duke@435 | 1162 | #endif |
duke@435 | 1163 | { |
duke@435 | 1164 | if (type == T_BOOLEAN) { |
duke@435 | 1165 | type = T_BYTE; |
duke@435 | 1166 | } |
duke@435 | 1167 | LIR_Address* addr; |
duke@435 | 1168 | if (type == T_ARRAY || type == T_OBJECT) { |
duke@435 | 1169 | LIR_Opr tmp = new_pointer_register(); |
duke@435 | 1170 | __ add(base_op, index_op, tmp); |
iveresov@1927 | 1171 | addr = new LIR_Address(tmp, type); |
duke@435 | 1172 | } else { |
duke@435 | 1173 | addr = new LIR_Address(base_op, index_op, type); |
duke@435 | 1174 | } |
duke@435 | 1175 | |
ysr@777 | 1176 | if (is_obj) { |
ysr@777 | 1177 | pre_barrier(LIR_OprFact::address(addr), false, NULL); |
ysr@777 | 1178 | // _bs->c1_write_barrier_pre(this, LIR_OprFact::address(addr)); |
ysr@777 | 1179 | } |
duke@435 | 1180 | __ move(data, addr); |
duke@435 | 1181 | if (is_obj) { |
duke@435 | 1182 | // This address is precise |
duke@435 | 1183 | post_barrier(LIR_OprFact::address(addr), data); |
duke@435 | 1184 | } |
duke@435 | 1185 | } |
duke@435 | 1186 | } |
duke@435 | 1187 | |
duke@435 | 1188 | |
duke@435 | 1189 | void LIRGenerator::get_Object_unsafe(LIR_Opr dst, LIR_Opr src, LIR_Opr offset, |
duke@435 | 1190 | BasicType type, bool is_volatile) { |
duke@435 | 1191 | #ifndef _LP64 |
duke@435 | 1192 | if (is_volatile && type == T_LONG) { |
duke@435 | 1193 | __ volatile_load_unsafe_reg(src, offset, dst, type, NULL, lir_patch_none); |
duke@435 | 1194 | } else |
duke@435 | 1195 | #endif |
duke@435 | 1196 | { |
duke@435 | 1197 | LIR_Address* addr = new LIR_Address(src, offset, type); |
duke@435 | 1198 | __ load(addr, dst); |
duke@435 | 1199 | } |
duke@435 | 1200 | } |