Thu, 20 Sep 2012 16:49:17 +0200
7023898: Intrinsify AtomicLongFieldUpdater.getAndIncrement()
Summary: use shorter instruction sequences for atomic add and atomic exchange when possible.
Reviewed-by: kvn, jrose
1 /*
2 * Copyright (c) 1999, 2011, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
25 #ifndef SHARE_VM_OPTO_LOCKNODE_HPP
26 #define SHARE_VM_OPTO_LOCKNODE_HPP
28 #include "opto/node.hpp"
29 #include "opto/opcodes.hpp"
30 #include "opto/subnode.hpp"
31 #ifdef TARGET_ARCH_MODEL_x86_32
32 # include "adfiles/ad_x86_32.hpp"
33 #endif
34 #ifdef TARGET_ARCH_MODEL_x86_64
35 # include "adfiles/ad_x86_64.hpp"
36 #endif
37 #ifdef TARGET_ARCH_MODEL_sparc
38 # include "adfiles/ad_sparc.hpp"
39 #endif
40 #ifdef TARGET_ARCH_MODEL_zero
41 # include "adfiles/ad_zero.hpp"
42 #endif
43 #ifdef TARGET_ARCH_MODEL_arm
44 # include "adfiles/ad_arm.hpp"
45 #endif
46 #ifdef TARGET_ARCH_MODEL_ppc
47 # include "adfiles/ad_ppc.hpp"
48 #endif
50 //------------------------------BoxLockNode------------------------------------
51 class BoxLockNode : public Node {
52 const int _slot; // stack slot
53 RegMask _inmask; // OptoReg corresponding to stack slot
54 bool _is_eliminated; // Associated locks were safely eliminated
56 public:
57 BoxLockNode( int lock );
58 virtual int Opcode() const;
59 virtual void emit(CodeBuffer &cbuf, PhaseRegAlloc *ra_) const;
60 virtual uint size(PhaseRegAlloc *ra_) const;
61 virtual const RegMask &in_RegMask(uint) const;
62 virtual const RegMask &out_RegMask() const;
63 virtual uint size_of() const;
64 virtual uint hash() const;
65 virtual uint cmp( const Node &n ) const;
66 virtual const class Type *bottom_type() const { return TypeRawPtr::BOTTOM; }
67 virtual uint ideal_reg() const { return Op_RegP; }
69 static OptoReg::Name reg(Node* box_node);
70 static BoxLockNode* box_node(Node* box_node);
71 static bool same_slot(Node* box1, Node* box2) {
72 return box1->as_BoxLock()->_slot == box2->as_BoxLock()->_slot;
73 }
74 int stack_slot() const { return _slot; }
76 bool is_eliminated() const { return _is_eliminated; }
77 // mark lock as eliminated.
78 void set_eliminated() { _is_eliminated = true; }
80 // Is BoxLock node used for one simple lock region?
81 bool is_simple_lock_region(LockNode** unique_lock, Node* obj);
83 #ifndef PRODUCT
84 virtual void format( PhaseRegAlloc *, outputStream *st ) const;
85 virtual void dump_spec(outputStream *st) const { st->print(" Lock %d",_slot); }
86 #endif
87 };
89 //------------------------------FastLockNode-----------------------------------
90 class FastLockNode: public CmpNode {
91 private:
92 BiasedLockingCounters* _counters;
94 public:
95 FastLockNode(Node *ctrl, Node *oop, Node *box) : CmpNode(oop,box) {
96 init_req(0,ctrl);
97 init_class_id(Class_FastLock);
98 _counters = NULL;
99 }
100 Node* obj_node() const { return in(1); }
101 Node* box_node() const { return in(2); }
102 void set_box_node(Node* box) { set_req(2, box); }
104 // FastLock and FastUnlockNode do not hash, we need one for each correspoding
105 // LockNode/UnLockNode to avoid creating Phi's.
106 virtual uint hash() const ; // { return NO_HASH; }
107 virtual uint cmp( const Node &n ) const ; // Always fail, except on self
108 virtual int Opcode() const;
109 virtual const Type *Value( PhaseTransform *phase ) const { return TypeInt::CC; }
110 const Type *sub(const Type *t1, const Type *t2) const { return TypeInt::CC;}
112 void create_lock_counter(JVMState* s);
113 BiasedLockingCounters* counters() const { return _counters; }
114 };
117 //------------------------------FastUnlockNode---------------------------------
118 class FastUnlockNode: public CmpNode {
119 public:
120 FastUnlockNode(Node *ctrl, Node *oop, Node *box) : CmpNode(oop,box) {
121 init_req(0,ctrl);
122 init_class_id(Class_FastUnlock);
123 }
124 Node* obj_node() const { return in(1); }
125 Node* box_node() const { return in(2); }
128 // FastLock and FastUnlockNode do not hash, we need one for each correspoding
129 // LockNode/UnLockNode to avoid creating Phi's.
130 virtual uint hash() const ; // { return NO_HASH; }
131 virtual uint cmp( const Node &n ) const ; // Always fail, except on self
132 virtual int Opcode() const;
133 virtual const Type *Value( PhaseTransform *phase ) const { return TypeInt::CC; }
134 const Type *sub(const Type *t1, const Type *t2) const { return TypeInt::CC;}
136 };
138 #endif // SHARE_VM_OPTO_LOCKNODE_HPP