src/cpu/x86/vm/x86.ad

Wed, 17 Jun 2015 17:48:25 -0700

author
ascarpino
date
Wed, 17 Jun 2015 17:48:25 -0700
changeset 9788
44ef77ad417c
parent 9333
2fccf735a116
child 9448
73d689add964
permissions
-rw-r--r--

8073108: Use x86 and SPARC CPU instructions for GHASH acceleration
Reviewed-by: kvn, jrose, phh

kvn@3390 1 //
kevinw@9333 2 // Copyright (c) 2011, 2018, Oracle and/or its affiliates. All rights reserved.
kvn@3390 3 // DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
kvn@3390 4 //
kvn@3390 5 // This code is free software; you can redistribute it and/or modify it
kvn@3390 6 // under the terms of the GNU General Public License version 2 only, as
kvn@3390 7 // published by the Free Software Foundation.
kvn@3390 8 //
kvn@3390 9 // This code is distributed in the hope that it will be useful, but WITHOUT
kvn@3390 10 // ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
kvn@3390 11 // FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
kvn@3390 12 // version 2 for more details (a copy is included in the LICENSE file that
kvn@3390 13 // accompanied this code).
kvn@3390 14 //
kvn@3390 15 // You should have received a copy of the GNU General Public License version
kvn@3390 16 // 2 along with this work; if not, write to the Free Software Foundation,
kvn@3390 17 // Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
kvn@3390 18 //
kvn@3390 19 // Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
kvn@3390 20 // or visit www.oracle.com if you need additional information or have any
kvn@3390 21 // questions.
kvn@3390 22 //
kvn@3390 23 //
kvn@3390 24
kvn@3390 25 // X86 Common Architecture Description File
kvn@3390 26
kvn@3882 27 //----------REGISTER DEFINITION BLOCK------------------------------------------
kvn@3882 28 // This information is used by the matcher and the register allocator to
kvn@3882 29 // describe individual registers and classes of registers within the target
kvn@3882 30 // archtecture.
kvn@3882 31
kvn@3882 32 register %{
kvn@3882 33 //----------Architecture Description Register Definitions----------------------
kvn@3882 34 // General Registers
kvn@3882 35 // "reg_def" name ( register save type, C convention save type,
kvn@3882 36 // ideal register type, encoding );
kvn@3882 37 // Register Save Types:
kvn@3882 38 //
kvn@3882 39 // NS = No-Save: The register allocator assumes that these registers
kvn@3882 40 // can be used without saving upon entry to the method, &
kvn@3882 41 // that they do not need to be saved at call sites.
kvn@3882 42 //
kvn@3882 43 // SOC = Save-On-Call: The register allocator assumes that these registers
kvn@3882 44 // can be used without saving upon entry to the method,
kvn@3882 45 // but that they must be saved at call sites.
kvn@3882 46 //
kvn@3882 47 // SOE = Save-On-Entry: The register allocator assumes that these registers
kvn@3882 48 // must be saved before using them upon entry to the
kvn@3882 49 // method, but they do not need to be saved at call
kvn@3882 50 // sites.
kvn@3882 51 //
kvn@3882 52 // AS = Always-Save: The register allocator assumes that these registers
kvn@3882 53 // must be saved before using them upon entry to the
kvn@3882 54 // method, & that they must be saved at call sites.
kvn@3882 55 //
kvn@3882 56 // Ideal Register Type is used to determine how to save & restore a
kvn@3882 57 // register. Op_RegI will get spilled with LoadI/StoreI, Op_RegP will get
kvn@3882 58 // spilled with LoadP/StoreP. If the register supports both, use Op_RegI.
kvn@3882 59 //
kvn@3882 60 // The encoding number is the actual bit-pattern placed into the opcodes.
kvn@3882 61
kvn@3882 62 // XMM registers. 256-bit registers or 8 words each, labeled (a)-h.
kvn@3882 63 // Word a in each register holds a Float, words ab hold a Double.
kvn@3882 64 // The whole registers are used in SSE4.2 version intrinsics,
kvn@3882 65 // array copy stubs and superword operations (see UseSSE42Intrinsics,
kvn@3882 66 // UseXMMForArrayCopy and UseSuperword flags).
kvn@3882 67 // XMM8-XMM15 must be encoded with REX (VEX for UseAVX).
kvn@3882 68 // Linux ABI: No register preserved across function calls
kvn@3882 69 // XMM0-XMM7 might hold parameters
kvn@3882 70 // Windows ABI: XMM6-XMM15 preserved across function calls
kvn@3882 71 // XMM0-XMM3 might hold parameters
kvn@3882 72
kvn@3882 73 reg_def XMM0 ( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg());
kvn@3929 74 reg_def XMM0b( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(1));
kvn@3929 75 reg_def XMM0c( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(2));
kvn@3929 76 reg_def XMM0d( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(3));
kvn@3929 77 reg_def XMM0e( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(4));
kvn@3929 78 reg_def XMM0f( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(5));
kvn@3929 79 reg_def XMM0g( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(6));
kvn@3929 80 reg_def XMM0h( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(7));
kvn@3882 81
kvn@3882 82 reg_def XMM1 ( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg());
kvn@3929 83 reg_def XMM1b( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(1));
kvn@3929 84 reg_def XMM1c( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(2));
kvn@3929 85 reg_def XMM1d( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(3));
kvn@3929 86 reg_def XMM1e( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(4));
kvn@3929 87 reg_def XMM1f( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(5));
kvn@3929 88 reg_def XMM1g( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(6));
kvn@3929 89 reg_def XMM1h( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(7));
kvn@3882 90
kvn@3882 91 reg_def XMM2 ( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg());
kvn@3929 92 reg_def XMM2b( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(1));
kvn@3929 93 reg_def XMM2c( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(2));
kvn@3929 94 reg_def XMM2d( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(3));
kvn@3929 95 reg_def XMM2e( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(4));
kvn@3929 96 reg_def XMM2f( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(5));
kvn@3929 97 reg_def XMM2g( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(6));
kvn@3929 98 reg_def XMM2h( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(7));
kvn@3882 99
kvn@3882 100 reg_def XMM3 ( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg());
kvn@3929 101 reg_def XMM3b( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(1));
kvn@3929 102 reg_def XMM3c( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(2));
kvn@3929 103 reg_def XMM3d( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(3));
kvn@3929 104 reg_def XMM3e( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(4));
kvn@3929 105 reg_def XMM3f( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(5));
kvn@3929 106 reg_def XMM3g( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(6));
kvn@3929 107 reg_def XMM3h( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(7));
kvn@3882 108
kvn@3882 109 reg_def XMM4 ( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg());
kvn@3929 110 reg_def XMM4b( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(1));
kvn@3929 111 reg_def XMM4c( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(2));
kvn@3929 112 reg_def XMM4d( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(3));
kvn@3929 113 reg_def XMM4e( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(4));
kvn@3929 114 reg_def XMM4f( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(5));
kvn@3929 115 reg_def XMM4g( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(6));
kvn@3929 116 reg_def XMM4h( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(7));
kvn@3882 117
kvn@3882 118 reg_def XMM5 ( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg());
kvn@3929 119 reg_def XMM5b( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(1));
kvn@3929 120 reg_def XMM5c( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(2));
kvn@3929 121 reg_def XMM5d( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(3));
kvn@3929 122 reg_def XMM5e( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(4));
kvn@3929 123 reg_def XMM5f( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(5));
kvn@3929 124 reg_def XMM5g( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(6));
kvn@3929 125 reg_def XMM5h( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(7));
kvn@3882 126
kvn@3882 127 #ifdef _WIN64
kvn@3882 128
kvn@3882 129 reg_def XMM6 ( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg());
kvn@3929 130 reg_def XMM6b( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(1));
kvn@3929 131 reg_def XMM6c( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(2));
kvn@3929 132 reg_def XMM6d( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(3));
kvn@3929 133 reg_def XMM6e( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(4));
kvn@3929 134 reg_def XMM6f( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(5));
kvn@3929 135 reg_def XMM6g( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(6));
kvn@3929 136 reg_def XMM6h( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(7));
kvn@3882 137
kvn@3882 138 reg_def XMM7 ( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg());
kvn@3929 139 reg_def XMM7b( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(1));
kvn@3929 140 reg_def XMM7c( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(2));
kvn@3929 141 reg_def XMM7d( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(3));
kvn@3929 142 reg_def XMM7e( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(4));
kvn@3929 143 reg_def XMM7f( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(5));
kvn@3929 144 reg_def XMM7g( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(6));
kvn@3929 145 reg_def XMM7h( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(7));
kvn@3882 146
kvn@3882 147 reg_def XMM8 ( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg());
kvn@3929 148 reg_def XMM8b( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(1));
kvn@3929 149 reg_def XMM8c( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(2));
kvn@3929 150 reg_def XMM8d( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(3));
kvn@3929 151 reg_def XMM8e( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(4));
kvn@3929 152 reg_def XMM8f( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(5));
kvn@3929 153 reg_def XMM8g( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(6));
kvn@3929 154 reg_def XMM8h( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(7));
kvn@3882 155
kvn@3882 156 reg_def XMM9 ( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg());
kvn@3929 157 reg_def XMM9b( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(1));
kvn@3929 158 reg_def XMM9c( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(2));
kvn@3929 159 reg_def XMM9d( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(3));
kvn@3929 160 reg_def XMM9e( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(4));
kvn@3929 161 reg_def XMM9f( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(5));
kvn@3929 162 reg_def XMM9g( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(6));
kvn@3929 163 reg_def XMM9h( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(7));
kvn@3882 164
kvn@3882 165 reg_def XMM10 ( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg());
kvn@3929 166 reg_def XMM10b( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(1));
kvn@3929 167 reg_def XMM10c( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(2));
kvn@3929 168 reg_def XMM10d( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(3));
kvn@3929 169 reg_def XMM10e( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(4));
kvn@3929 170 reg_def XMM10f( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(5));
kvn@3929 171 reg_def XMM10g( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(6));
kvn@3929 172 reg_def XMM10h( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(7));
kvn@3882 173
kvn@3882 174 reg_def XMM11 ( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg());
kvn@3929 175 reg_def XMM11b( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(1));
kvn@3929 176 reg_def XMM11c( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(2));
kvn@3929 177 reg_def XMM11d( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(3));
kvn@3929 178 reg_def XMM11e( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(4));
kvn@3929 179 reg_def XMM11f( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(5));
kvn@3929 180 reg_def XMM11g( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(6));
kvn@3929 181 reg_def XMM11h( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(7));
kvn@3882 182
kvn@3882 183 reg_def XMM12 ( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg());
kvn@3929 184 reg_def XMM12b( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(1));
kvn@3929 185 reg_def XMM12c( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(2));
kvn@3929 186 reg_def XMM12d( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(3));
kvn@3929 187 reg_def XMM12e( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(4));
kvn@3929 188 reg_def XMM12f( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(5));
kvn@3929 189 reg_def XMM12g( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(6));
kvn@3929 190 reg_def XMM12h( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(7));
kvn@3882 191
kvn@3882 192 reg_def XMM13 ( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg());
kvn@3929 193 reg_def XMM13b( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(1));
kvn@3929 194 reg_def XMM13c( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(2));
kvn@3929 195 reg_def XMM13d( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(3));
kvn@3929 196 reg_def XMM13e( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(4));
kvn@3929 197 reg_def XMM13f( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(5));
kvn@3929 198 reg_def XMM13g( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(6));
kvn@3929 199 reg_def XMM13h( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(7));
kvn@3882 200
kvn@3882 201 reg_def XMM14 ( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg());
kvn@3929 202 reg_def XMM14b( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(1));
kvn@3929 203 reg_def XMM14c( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(2));
kvn@3929 204 reg_def XMM14d( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(3));
kvn@3929 205 reg_def XMM14e( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(4));
kvn@3929 206 reg_def XMM14f( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(5));
kvn@3929 207 reg_def XMM14g( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(6));
kvn@3929 208 reg_def XMM14h( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(7));
kvn@3882 209
kvn@3882 210 reg_def XMM15 ( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg());
kvn@3929 211 reg_def XMM15b( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(1));
kvn@3929 212 reg_def XMM15c( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(2));
kvn@3929 213 reg_def XMM15d( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(3));
kvn@3929 214 reg_def XMM15e( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(4));
kvn@3929 215 reg_def XMM15f( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(5));
kvn@3929 216 reg_def XMM15g( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(6));
kvn@3929 217 reg_def XMM15h( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(7));
kvn@3882 218
kvn@3882 219 #else // _WIN64
kvn@3882 220
kvn@3882 221 reg_def XMM6 ( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg());
kvn@3929 222 reg_def XMM6b( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(1));
kvn@3929 223 reg_def XMM6c( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(2));
kvn@3929 224 reg_def XMM6d( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(3));
kvn@3929 225 reg_def XMM6e( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(4));
kvn@3929 226 reg_def XMM6f( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(5));
kvn@3929 227 reg_def XMM6g( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(6));
kvn@3929 228 reg_def XMM6h( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(7));
kvn@3882 229
kvn@3882 230 reg_def XMM7 ( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg());
kvn@3929 231 reg_def XMM7b( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(1));
kvn@3929 232 reg_def XMM7c( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(2));
kvn@3929 233 reg_def XMM7d( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(3));
kvn@3929 234 reg_def XMM7e( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(4));
kvn@3929 235 reg_def XMM7f( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(5));
kvn@3929 236 reg_def XMM7g( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(6));
kvn@3929 237 reg_def XMM7h( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(7));
kvn@3882 238
kvn@3882 239 #ifdef _LP64
kvn@3882 240
kvn@3882 241 reg_def XMM8 ( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg());
kvn@3929 242 reg_def XMM8b( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(1));
kvn@3929 243 reg_def XMM8c( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(2));
kvn@3929 244 reg_def XMM8d( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(3));
kvn@3929 245 reg_def XMM8e( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(4));
kvn@3929 246 reg_def XMM8f( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(5));
kvn@3929 247 reg_def XMM8g( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(6));
kvn@3929 248 reg_def XMM8h( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(7));
kvn@3882 249
kvn@3882 250 reg_def XMM9 ( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg());
kvn@3929 251 reg_def XMM9b( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(1));
kvn@3929 252 reg_def XMM9c( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(2));
kvn@3929 253 reg_def XMM9d( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(3));
kvn@3929 254 reg_def XMM9e( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(4));
kvn@3929 255 reg_def XMM9f( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(5));
kvn@3929 256 reg_def XMM9g( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(6));
kvn@3929 257 reg_def XMM9h( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(7));
kvn@3882 258
kvn@3882 259 reg_def XMM10 ( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg());
kvn@3929 260 reg_def XMM10b( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(1));
kvn@3929 261 reg_def XMM10c( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(2));
kvn@3929 262 reg_def XMM10d( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(3));
kvn@3929 263 reg_def XMM10e( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(4));
kvn@3929 264 reg_def XMM10f( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(5));
kvn@3929 265 reg_def XMM10g( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(6));
kvn@3929 266 reg_def XMM10h( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(7));
kvn@3882 267
kvn@3882 268 reg_def XMM11 ( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg());
kvn@3929 269 reg_def XMM11b( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(1));
kvn@3929 270 reg_def XMM11c( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(2));
kvn@3929 271 reg_def XMM11d( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(3));
kvn@3929 272 reg_def XMM11e( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(4));
kvn@3929 273 reg_def XMM11f( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(5));
kvn@3929 274 reg_def XMM11g( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(6));
kvn@3929 275 reg_def XMM11h( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(7));
kvn@3882 276
kvn@3882 277 reg_def XMM12 ( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg());
kvn@3929 278 reg_def XMM12b( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(1));
kvn@3929 279 reg_def XMM12c( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(2));
kvn@3929 280 reg_def XMM12d( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(3));
kvn@3929 281 reg_def XMM12e( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(4));
kvn@3929 282 reg_def XMM12f( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(5));
kvn@3929 283 reg_def XMM12g( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(6));
kvn@3929 284 reg_def XMM12h( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(7));
kvn@3882 285
kvn@3882 286 reg_def XMM13 ( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg());
kvn@3929 287 reg_def XMM13b( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(1));
kvn@3929 288 reg_def XMM13c( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(2));
kvn@3929 289 reg_def XMM13d( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(3));
kvn@3929 290 reg_def XMM13e( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(4));
kvn@3929 291 reg_def XMM13f( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(5));
kvn@3929 292 reg_def XMM13g( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(6));
kvn@3929 293 reg_def XMM13h( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(7));
kvn@3882 294
kvn@3882 295 reg_def XMM14 ( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg());
kvn@3929 296 reg_def XMM14b( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(1));
kvn@3929 297 reg_def XMM14c( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(2));
kvn@3929 298 reg_def XMM14d( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(3));
kvn@3929 299 reg_def XMM14e( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(4));
kvn@3929 300 reg_def XMM14f( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(5));
kvn@3929 301 reg_def XMM14g( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(6));
kvn@3929 302 reg_def XMM14h( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(7));
kvn@3882 303
kvn@3882 304 reg_def XMM15 ( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg());
kvn@3929 305 reg_def XMM15b( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(1));
kvn@3929 306 reg_def XMM15c( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(2));
kvn@3929 307 reg_def XMM15d( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(3));
kvn@3929 308 reg_def XMM15e( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(4));
kvn@3929 309 reg_def XMM15f( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(5));
kvn@3929 310 reg_def XMM15g( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(6));
kvn@3929 311 reg_def XMM15h( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(7));
kvn@3882 312
kvn@3882 313 #endif // _LP64
kvn@3882 314
kvn@3882 315 #endif // _WIN64
kvn@3882 316
kvn@3882 317 #ifdef _LP64
kvn@3882 318 reg_def RFLAGS(SOC, SOC, 0, 16, VMRegImpl::Bad());
kvn@3882 319 #else
kvn@3882 320 reg_def RFLAGS(SOC, SOC, 0, 8, VMRegImpl::Bad());
kvn@3882 321 #endif // _LP64
kvn@3882 322
kvn@3882 323 alloc_class chunk1(XMM0, XMM0b, XMM0c, XMM0d, XMM0e, XMM0f, XMM0g, XMM0h,
kvn@3882 324 XMM1, XMM1b, XMM1c, XMM1d, XMM1e, XMM1f, XMM1g, XMM1h,
kvn@3882 325 XMM2, XMM2b, XMM2c, XMM2d, XMM2e, XMM2f, XMM2g, XMM2h,
kvn@3882 326 XMM3, XMM3b, XMM3c, XMM3d, XMM3e, XMM3f, XMM3g, XMM3h,
kvn@3882 327 XMM4, XMM4b, XMM4c, XMM4d, XMM4e, XMM4f, XMM4g, XMM4h,
kvn@3882 328 XMM5, XMM5b, XMM5c, XMM5d, XMM5e, XMM5f, XMM5g, XMM5h,
kvn@3882 329 XMM6, XMM6b, XMM6c, XMM6d, XMM6e, XMM6f, XMM6g, XMM6h,
kvn@3882 330 XMM7, XMM7b, XMM7c, XMM7d, XMM7e, XMM7f, XMM7g, XMM7h
kvn@3882 331 #ifdef _LP64
kvn@3882 332 ,XMM8, XMM8b, XMM8c, XMM8d, XMM8e, XMM8f, XMM8g, XMM8h,
kvn@3882 333 XMM9, XMM9b, XMM9c, XMM9d, XMM9e, XMM9f, XMM9g, XMM9h,
kvn@3882 334 XMM10, XMM10b, XMM10c, XMM10d, XMM10e, XMM10f, XMM10g, XMM10h,
kvn@3882 335 XMM11, XMM11b, XMM11c, XMM11d, XMM11e, XMM11f, XMM11g, XMM11h,
kvn@3882 336 XMM12, XMM12b, XMM12c, XMM12d, XMM12e, XMM12f, XMM12g, XMM12h,
kvn@3882 337 XMM13, XMM13b, XMM13c, XMM13d, XMM13e, XMM13f, XMM13g, XMM13h,
kvn@3882 338 XMM14, XMM14b, XMM14c, XMM14d, XMM14e, XMM14f, XMM14g, XMM14h,
kvn@3882 339 XMM15, XMM15b, XMM15c, XMM15d, XMM15e, XMM15f, XMM15g, XMM15h
kvn@3882 340 #endif
kvn@3882 341 );
kvn@3882 342
kvn@3882 343 // flags allocation class should be last.
kvn@3882 344 alloc_class chunk2(RFLAGS);
kvn@3882 345
kvn@3882 346 // Singleton class for condition codes
kvn@3882 347 reg_class int_flags(RFLAGS);
kvn@3882 348
kvn@3882 349 // Class for all float registers
kvn@3882 350 reg_class float_reg(XMM0,
kvn@3882 351 XMM1,
kvn@3882 352 XMM2,
kvn@3882 353 XMM3,
kvn@3882 354 XMM4,
kvn@3882 355 XMM5,
kvn@3882 356 XMM6,
kvn@3882 357 XMM7
kvn@3882 358 #ifdef _LP64
kvn@3882 359 ,XMM8,
kvn@3882 360 XMM9,
kvn@3882 361 XMM10,
kvn@3882 362 XMM11,
kvn@3882 363 XMM12,
kvn@3882 364 XMM13,
kvn@3882 365 XMM14,
kvn@3882 366 XMM15
kvn@3882 367 #endif
kvn@3882 368 );
kvn@3882 369
kvn@3882 370 // Class for all double registers
kvn@3882 371 reg_class double_reg(XMM0, XMM0b,
kvn@3882 372 XMM1, XMM1b,
kvn@3882 373 XMM2, XMM2b,
kvn@3882 374 XMM3, XMM3b,
kvn@3882 375 XMM4, XMM4b,
kvn@3882 376 XMM5, XMM5b,
kvn@3882 377 XMM6, XMM6b,
kvn@3882 378 XMM7, XMM7b
kvn@3882 379 #ifdef _LP64
kvn@3882 380 ,XMM8, XMM8b,
kvn@3882 381 XMM9, XMM9b,
kvn@3882 382 XMM10, XMM10b,
kvn@3882 383 XMM11, XMM11b,
kvn@3882 384 XMM12, XMM12b,
kvn@3882 385 XMM13, XMM13b,
kvn@3882 386 XMM14, XMM14b,
kvn@3882 387 XMM15, XMM15b
kvn@3882 388 #endif
kvn@3882 389 );
kvn@3882 390
kvn@3882 391 // Class for all 32bit vector registers
kvn@3882 392 reg_class vectors_reg(XMM0,
kvn@3882 393 XMM1,
kvn@3882 394 XMM2,
kvn@3882 395 XMM3,
kvn@3882 396 XMM4,
kvn@3882 397 XMM5,
kvn@3882 398 XMM6,
kvn@3882 399 XMM7
kvn@3882 400 #ifdef _LP64
kvn@3882 401 ,XMM8,
kvn@3882 402 XMM9,
kvn@3882 403 XMM10,
kvn@3882 404 XMM11,
kvn@3882 405 XMM12,
kvn@3882 406 XMM13,
kvn@3882 407 XMM14,
kvn@3882 408 XMM15
kvn@3882 409 #endif
kvn@3882 410 );
kvn@3882 411
kvn@3882 412 // Class for all 64bit vector registers
kvn@3882 413 reg_class vectord_reg(XMM0, XMM0b,
kvn@3882 414 XMM1, XMM1b,
kvn@3882 415 XMM2, XMM2b,
kvn@3882 416 XMM3, XMM3b,
kvn@3882 417 XMM4, XMM4b,
kvn@3882 418 XMM5, XMM5b,
kvn@3882 419 XMM6, XMM6b,
kvn@3882 420 XMM7, XMM7b
kvn@3882 421 #ifdef _LP64
kvn@3882 422 ,XMM8, XMM8b,
kvn@3882 423 XMM9, XMM9b,
kvn@3882 424 XMM10, XMM10b,
kvn@3882 425 XMM11, XMM11b,
kvn@3882 426 XMM12, XMM12b,
kvn@3882 427 XMM13, XMM13b,
kvn@3882 428 XMM14, XMM14b,
kvn@3882 429 XMM15, XMM15b
kvn@3882 430 #endif
kvn@3882 431 );
kvn@3882 432
kvn@3882 433 // Class for all 128bit vector registers
kvn@3882 434 reg_class vectorx_reg(XMM0, XMM0b, XMM0c, XMM0d,
kvn@3882 435 XMM1, XMM1b, XMM1c, XMM1d,
kvn@3882 436 XMM2, XMM2b, XMM2c, XMM2d,
kvn@3882 437 XMM3, XMM3b, XMM3c, XMM3d,
kvn@3882 438 XMM4, XMM4b, XMM4c, XMM4d,
kvn@3882 439 XMM5, XMM5b, XMM5c, XMM5d,
kvn@3882 440 XMM6, XMM6b, XMM6c, XMM6d,
kvn@3882 441 XMM7, XMM7b, XMM7c, XMM7d
kvn@3882 442 #ifdef _LP64
kvn@3882 443 ,XMM8, XMM8b, XMM8c, XMM8d,
kvn@3882 444 XMM9, XMM9b, XMM9c, XMM9d,
kvn@3882 445 XMM10, XMM10b, XMM10c, XMM10d,
kvn@3882 446 XMM11, XMM11b, XMM11c, XMM11d,
kvn@3882 447 XMM12, XMM12b, XMM12c, XMM12d,
kvn@3882 448 XMM13, XMM13b, XMM13c, XMM13d,
kvn@3882 449 XMM14, XMM14b, XMM14c, XMM14d,
kvn@3882 450 XMM15, XMM15b, XMM15c, XMM15d
kvn@3882 451 #endif
kvn@3882 452 );
kvn@3882 453
kvn@3882 454 // Class for all 256bit vector registers
kvn@3882 455 reg_class vectory_reg(XMM0, XMM0b, XMM0c, XMM0d, XMM0e, XMM0f, XMM0g, XMM0h,
kvn@3882 456 XMM1, XMM1b, XMM1c, XMM1d, XMM1e, XMM1f, XMM1g, XMM1h,
kvn@3882 457 XMM2, XMM2b, XMM2c, XMM2d, XMM2e, XMM2f, XMM2g, XMM2h,
kvn@3882 458 XMM3, XMM3b, XMM3c, XMM3d, XMM3e, XMM3f, XMM3g, XMM3h,
kvn@3882 459 XMM4, XMM4b, XMM4c, XMM4d, XMM4e, XMM4f, XMM4g, XMM4h,
kvn@3882 460 XMM5, XMM5b, XMM5c, XMM5d, XMM5e, XMM5f, XMM5g, XMM5h,
kvn@3882 461 XMM6, XMM6b, XMM6c, XMM6d, XMM6e, XMM6f, XMM6g, XMM6h,
kvn@3882 462 XMM7, XMM7b, XMM7c, XMM7d, XMM7e, XMM7f, XMM7g, XMM7h
kvn@3882 463 #ifdef _LP64
kvn@3882 464 ,XMM8, XMM8b, XMM8c, XMM8d, XMM8e, XMM8f, XMM8g, XMM8h,
kvn@3882 465 XMM9, XMM9b, XMM9c, XMM9d, XMM9e, XMM9f, XMM9g, XMM9h,
kvn@3882 466 XMM10, XMM10b, XMM10c, XMM10d, XMM10e, XMM10f, XMM10g, XMM10h,
kvn@3882 467 XMM11, XMM11b, XMM11c, XMM11d, XMM11e, XMM11f, XMM11g, XMM11h,
kvn@3882 468 XMM12, XMM12b, XMM12c, XMM12d, XMM12e, XMM12f, XMM12g, XMM12h,
kvn@3882 469 XMM13, XMM13b, XMM13c, XMM13d, XMM13e, XMM13f, XMM13g, XMM13h,
kvn@3882 470 XMM14, XMM14b, XMM14c, XMM14d, XMM14e, XMM14f, XMM14g, XMM14h,
kvn@3882 471 XMM15, XMM15b, XMM15c, XMM15d, XMM15e, XMM15f, XMM15g, XMM15h
kvn@3882 472 #endif
kvn@3882 473 );
kvn@3882 474
kvn@3882 475 %}
kvn@3882 476
goetz@6517 477
goetz@6517 478 //----------SOURCE BLOCK-------------------------------------------------------
goetz@6517 479 // This is a block of C++ code which provides values, functions, and
goetz@6517 480 // definitions necessary in the rest of the architecture description
goetz@6517 481
goetz@6517 482 source_hpp %{
goetz@6517 483 // Header information of the source block.
goetz@6517 484 // Method declarations/definitions which are used outside
goetz@6517 485 // the ad-scope can conveniently be defined here.
goetz@6517 486 //
goetz@6517 487 // To keep related declarations/definitions/uses close together,
goetz@6517 488 // we switch between source %{ }% and source_hpp %{ }% freely as needed.
goetz@6517 489
goetz@6517 490 class CallStubImpl {
goetz@6517 491
goetz@6517 492 //--------------------------------------------------------------
goetz@6517 493 //---< Used for optimization in Compile::shorten_branches >---
goetz@6517 494 //--------------------------------------------------------------
goetz@6517 495
goetz@6517 496 public:
goetz@6517 497 // Size of call trampoline stub.
goetz@6517 498 static uint size_call_trampoline() {
goetz@6517 499 return 0; // no call trampolines on this platform
goetz@6517 500 }
goetz@6517 501
goetz@6517 502 // number of relocations needed by a call trampoline stub
goetz@6517 503 static uint reloc_call_trampoline() {
goetz@6517 504 return 0; // no call trampolines on this platform
goetz@6517 505 }
goetz@6517 506 };
goetz@6517 507
goetz@6517 508 class HandlerImpl {
goetz@6517 509
goetz@6517 510 public:
goetz@6517 511
goetz@6517 512 static int emit_exception_handler(CodeBuffer &cbuf);
goetz@6517 513 static int emit_deopt_handler(CodeBuffer& cbuf);
goetz@6517 514
goetz@6517 515 static uint size_exception_handler() {
goetz@6517 516 // NativeCall instruction size is the same as NativeJump.
goetz@6517 517 // exception handler starts out as jump and can be patched to
goetz@6517 518 // a call be deoptimization. (4932387)
goetz@6517 519 // Note that this value is also credited (in output.cpp) to
goetz@6517 520 // the size of the code section.
goetz@6517 521 return NativeJump::instruction_size;
goetz@6517 522 }
goetz@6517 523
goetz@6517 524 #ifdef _LP64
goetz@6517 525 static uint size_deopt_handler() {
goetz@6517 526 // three 5 byte instructions
goetz@6517 527 return 15;
goetz@6517 528 }
goetz@6517 529 #else
goetz@6517 530 static uint size_deopt_handler() {
goetz@6517 531 // NativeCall instruction size is the same as NativeJump.
goetz@6517 532 // exception handler starts out as jump and can be patched to
goetz@6517 533 // a call be deoptimization. (4932387)
goetz@6517 534 // Note that this value is also credited (in output.cpp) to
goetz@6517 535 // the size of the code section.
goetz@6517 536 return 5 + NativeJump::instruction_size; // pushl(); jmp;
goetz@6517 537 }
goetz@6517 538 #endif
goetz@6517 539 };
goetz@6517 540
goetz@6517 541 %} // end source_hpp
goetz@6517 542
kvn@3390 543 source %{
goetz@6517 544
goetz@6517 545 // Emit exception handler code.
goetz@6517 546 // Stuff framesize into a register and call a VM stub routine.
goetz@6517 547 int HandlerImpl::emit_exception_handler(CodeBuffer& cbuf) {
goetz@6517 548
goetz@6517 549 // Note that the code buffer's insts_mark is always relative to insts.
goetz@6517 550 // That's why we must use the macroassembler to generate a handler.
goetz@6517 551 MacroAssembler _masm(&cbuf);
goetz@6517 552 address base = __ start_a_stub(size_exception_handler());
vkempik@8427 553 if (base == NULL) {
vkempik@8427 554 ciEnv::current()->record_failure("CodeCache is full");
vkempik@8427 555 return 0; // CodeBuffer::expand failed
vkempik@8427 556 }
goetz@6517 557 int offset = __ offset();
goetz@6517 558 __ jump(RuntimeAddress(OptoRuntime::exception_blob()->entry_point()));
goetz@6517 559 assert(__ offset() - offset <= (int) size_exception_handler(), "overflow");
goetz@6517 560 __ end_a_stub();
goetz@6517 561 return offset;
goetz@6517 562 }
goetz@6517 563
goetz@6517 564 // Emit deopt handler code.
goetz@6517 565 int HandlerImpl::emit_deopt_handler(CodeBuffer& cbuf) {
goetz@6517 566
goetz@6517 567 // Note that the code buffer's insts_mark is always relative to insts.
goetz@6517 568 // That's why we must use the macroassembler to generate a handler.
goetz@6517 569 MacroAssembler _masm(&cbuf);
goetz@6517 570 address base = __ start_a_stub(size_deopt_handler());
vkempik@8427 571 if (base == NULL) {
vkempik@8427 572 ciEnv::current()->record_failure("CodeCache is full");
vkempik@8427 573 return 0; // CodeBuffer::expand failed
vkempik@8427 574 }
goetz@6517 575 int offset = __ offset();
goetz@6517 576
goetz@6517 577 #ifdef _LP64
goetz@6517 578 address the_pc = (address) __ pc();
goetz@6517 579 Label next;
goetz@6517 580 // push a "the_pc" on the stack without destroying any registers
goetz@6517 581 // as they all may be live.
goetz@6517 582
goetz@6517 583 // push address of "next"
goetz@6517 584 __ call(next, relocInfo::none); // reloc none is fine since it is a disp32
goetz@6517 585 __ bind(next);
goetz@6517 586 // adjust it so it matches "the_pc"
goetz@6517 587 __ subptr(Address(rsp, 0), __ offset() - offset);
goetz@6517 588 #else
goetz@6517 589 InternalAddress here(__ pc());
goetz@6517 590 __ pushptr(here.addr());
goetz@6517 591 #endif
goetz@6517 592
goetz@6517 593 __ jump(RuntimeAddress(SharedRuntime::deopt_blob()->unpack()));
goetz@6517 594 assert(__ offset() - offset <= (int) size_deopt_handler(), "overflow");
goetz@6517 595 __ end_a_stub();
goetz@6517 596 return offset;
goetz@6517 597 }
goetz@6517 598
goetz@6517 599
goetz@6517 600 //=============================================================================
goetz@6517 601
kvn@3390 602 // Float masks come from different places depending on platform.
kvn@3390 603 #ifdef _LP64
kvn@3390 604 static address float_signmask() { return StubRoutines::x86::float_sign_mask(); }
kvn@3390 605 static address float_signflip() { return StubRoutines::x86::float_sign_flip(); }
kvn@3390 606 static address double_signmask() { return StubRoutines::x86::double_sign_mask(); }
kvn@3390 607 static address double_signflip() { return StubRoutines::x86::double_sign_flip(); }
kvn@3390 608 #else
kvn@3390 609 static address float_signmask() { return (address)float_signmask_pool; }
kvn@3390 610 static address float_signflip() { return (address)float_signflip_pool; }
kvn@3390 611 static address double_signmask() { return (address)double_signmask_pool; }
kvn@3390 612 static address double_signflip() { return (address)double_signflip_pool; }
kvn@3390 613 #endif
kvn@3577 614
kvn@3882 615
kvn@4001 616 const bool Matcher::match_rule_supported(int opcode) {
kvn@4001 617 if (!has_match_rule(opcode))
kvn@4001 618 return false;
kvn@4001 619
kvn@4001 620 switch (opcode) {
kvn@4001 621 case Op_PopCountI:
kvn@4001 622 case Op_PopCountL:
kvn@4001 623 if (!UsePopCountInstruction)
kvn@4001 624 return false;
kvn@4103 625 break;
kvn@4001 626 case Op_MulVI:
kvn@4001 627 if ((UseSSE < 4) && (UseAVX < 1)) // only with SSE4_1 or AVX
kvn@4001 628 return false;
kvn@4001 629 break;
roland@4106 630 case Op_CompareAndSwapL:
roland@4106 631 #ifdef _LP64
roland@4106 632 case Op_CompareAndSwapP:
roland@4106 633 #endif
roland@4106 634 if (!VM_Version::supports_cx8())
roland@4106 635 return false;
roland@4106 636 break;
kvn@4001 637 }
kvn@4001 638
kvn@4001 639 return true; // Per default match rules are supported.
kvn@4001 640 }
kvn@4001 641
kvn@3882 642 // Max vector size in bytes. 0 if not supported.
kvn@3882 643 const int Matcher::vector_width_in_bytes(BasicType bt) {
kvn@3882 644 assert(is_java_primitive(bt), "only primitive type vectors");
kvn@3882 645 if (UseSSE < 2) return 0;
kvn@3882 646 // SSE2 supports 128bit vectors for all types.
kvn@3882 647 // AVX2 supports 256bit vectors for all types.
kvn@3882 648 int size = (UseAVX > 1) ? 32 : 16;
kvn@3882 649 // AVX1 supports 256bit vectors only for FLOAT and DOUBLE.
kvn@3882 650 if (UseAVX > 0 && (bt == T_FLOAT || bt == T_DOUBLE))
kvn@3882 651 size = 32;
kvn@3882 652 // Use flag to limit vector size.
kvn@3882 653 size = MIN2(size,(int)MaxVectorSize);
kvn@3882 654 // Minimum 2 values in vector (or 4 for bytes).
kvn@3882 655 switch (bt) {
kvn@3882 656 case T_DOUBLE:
kvn@3882 657 case T_LONG:
kvn@3882 658 if (size < 16) return 0;
kvn@3882 659 case T_FLOAT:
kvn@3882 660 case T_INT:
kvn@3882 661 if (size < 8) return 0;
kvn@3882 662 case T_BOOLEAN:
kvn@3882 663 case T_BYTE:
kvn@3882 664 case T_CHAR:
kvn@3882 665 case T_SHORT:
kvn@3882 666 if (size < 4) return 0;
kvn@3882 667 break;
kvn@3882 668 default:
kvn@3882 669 ShouldNotReachHere();
kvn@3882 670 }
kvn@3882 671 return size;
kvn@3882 672 }
kvn@3882 673
kvn@3882 674 // Limits on vector size (number of elements) loaded into vector.
kvn@3882 675 const int Matcher::max_vector_size(const BasicType bt) {
kvn@3882 676 return vector_width_in_bytes(bt)/type2aelembytes(bt);
kvn@3882 677 }
kvn@3882 678 const int Matcher::min_vector_size(const BasicType bt) {
kvn@3882 679 int max_size = max_vector_size(bt);
kvn@3882 680 // Min size which can be loaded into vector is 4 bytes.
kvn@3882 681 int size = (type2aelembytes(bt) == 1) ? 4 : 2;
kvn@3882 682 return MIN2(size,max_size);
kvn@3882 683 }
kvn@3882 684
kvn@3882 685 // Vector ideal reg corresponding to specidied size in bytes
kevinw@9333 686 const uint Matcher::vector_ideal_reg(int size) {
kvn@3882 687 assert(MaxVectorSize >= size, "");
kvn@3882 688 switch(size) {
kvn@3882 689 case 4: return Op_VecS;
kvn@3882 690 case 8: return Op_VecD;
kvn@3882 691 case 16: return Op_VecX;
kvn@3882 692 case 32: return Op_VecY;
kvn@3882 693 }
kvn@3882 694 ShouldNotReachHere();
kvn@3882 695 return 0;
kvn@3882 696 }
kvn@3882 697
kvn@4134 698 // Only lowest bits of xmm reg are used for vector shift count.
kevinw@9333 699 const uint Matcher::vector_shift_count_ideal_reg(int size) {
kvn@4134 700 return Op_VecS;
kvn@4134 701 }
kvn@4134 702
kvn@3882 703 // x86 supports misaligned vectors store/load.
kvn@3882 704 const bool Matcher::misaligned_vectors_ok() {
kvn@3882 705 return !AlignVector; // can be changed by flag
kvn@3882 706 }
kvn@3882 707
kvn@6312 708 // x86 AES instructions are compatible with SunJCE expanded
kvn@6312 709 // keys, hence we do not need to pass the original key to stubs
kvn@6312 710 const bool Matcher::pass_original_key_for_aes() {
kvn@6312 711 return false;
kvn@6312 712 }
kvn@6312 713
kvn@3882 714 // Helper methods for MachSpillCopyNode::implementation().
kvn@3882 715 static int vec_mov_helper(CodeBuffer *cbuf, bool do_size, int src_lo, int dst_lo,
kvn@3882 716 int src_hi, int dst_hi, uint ireg, outputStream* st) {
kvn@3882 717 // In 64-bit VM size calculation is very complex. Emitting instructions
kvn@3882 718 // into scratch buffer is used to get size in 64-bit VM.
kvn@3882 719 LP64_ONLY( assert(!do_size, "this method calculates size only for 32-bit VM"); )
kvn@3882 720 assert(ireg == Op_VecS || // 32bit vector
kvn@3882 721 (src_lo & 1) == 0 && (src_lo + 1) == src_hi &&
kvn@3882 722 (dst_lo & 1) == 0 && (dst_lo + 1) == dst_hi,
kvn@3882 723 "no non-adjacent vector moves" );
kvn@3882 724 if (cbuf) {
kvn@3882 725 MacroAssembler _masm(cbuf);
kvn@3882 726 int offset = __ offset();
kvn@3882 727 switch (ireg) {
kvn@3882 728 case Op_VecS: // copy whole register
kvn@3882 729 case Op_VecD:
kvn@3882 730 case Op_VecX:
kvn@3882 731 __ movdqu(as_XMMRegister(Matcher::_regEncode[dst_lo]), as_XMMRegister(Matcher::_regEncode[src_lo]));
kvn@3882 732 break;
kvn@3882 733 case Op_VecY:
kvn@3882 734 __ vmovdqu(as_XMMRegister(Matcher::_regEncode[dst_lo]), as_XMMRegister(Matcher::_regEncode[src_lo]));
kvn@3882 735 break;
kvn@3882 736 default:
kvn@3882 737 ShouldNotReachHere();
kvn@3882 738 }
kvn@3882 739 int size = __ offset() - offset;
kvn@3882 740 #ifdef ASSERT
kvn@3882 741 // VEX_2bytes prefix is used if UseAVX > 0, so it takes the same 2 bytes as SIMD prefix.
kvn@3882 742 assert(!do_size || size == 4, "incorrect size calculattion");
kvn@3882 743 #endif
kvn@3882 744 return size;
kvn@3882 745 #ifndef PRODUCT
kvn@3882 746 } else if (!do_size) {
kvn@3882 747 switch (ireg) {
kvn@3882 748 case Op_VecS:
kvn@3882 749 case Op_VecD:
kvn@3882 750 case Op_VecX:
kvn@3882 751 st->print("movdqu %s,%s\t# spill",Matcher::regName[dst_lo],Matcher::regName[src_lo]);
kvn@3882 752 break;
kvn@3882 753 case Op_VecY:
kvn@3882 754 st->print("vmovdqu %s,%s\t# spill",Matcher::regName[dst_lo],Matcher::regName[src_lo]);
kvn@3882 755 break;
kvn@3882 756 default:
kvn@3882 757 ShouldNotReachHere();
kvn@3882 758 }
kvn@3882 759 #endif
kvn@3882 760 }
kvn@3882 761 // VEX_2bytes prefix is used if UseAVX > 0, and it takes the same 2 bytes as SIMD prefix.
kvn@3882 762 return 4;
kvn@3882 763 }
kvn@3882 764
kvn@3882 765 static int vec_spill_helper(CodeBuffer *cbuf, bool do_size, bool is_load,
kvn@3882 766 int stack_offset, int reg, uint ireg, outputStream* st) {
kvn@3882 767 // In 64-bit VM size calculation is very complex. Emitting instructions
kvn@3882 768 // into scratch buffer is used to get size in 64-bit VM.
kvn@3882 769 LP64_ONLY( assert(!do_size, "this method calculates size only for 32-bit VM"); )
kvn@3882 770 if (cbuf) {
kvn@3882 771 MacroAssembler _masm(cbuf);
kvn@3882 772 int offset = __ offset();
kvn@3882 773 if (is_load) {
kvn@3882 774 switch (ireg) {
kvn@3882 775 case Op_VecS:
kvn@3882 776 __ movdl(as_XMMRegister(Matcher::_regEncode[reg]), Address(rsp, stack_offset));
kvn@3882 777 break;
kvn@3882 778 case Op_VecD:
kvn@3882 779 __ movq(as_XMMRegister(Matcher::_regEncode[reg]), Address(rsp, stack_offset));
kvn@3882 780 break;
kvn@3882 781 case Op_VecX:
kvn@3882 782 __ movdqu(as_XMMRegister(Matcher::_regEncode[reg]), Address(rsp, stack_offset));
kvn@3882 783 break;
kvn@3882 784 case Op_VecY:
kvn@3882 785 __ vmovdqu(as_XMMRegister(Matcher::_regEncode[reg]), Address(rsp, stack_offset));
kvn@3882 786 break;
kvn@3882 787 default:
kvn@3882 788 ShouldNotReachHere();
kvn@3882 789 }
kvn@3882 790 } else { // store
kvn@3882 791 switch (ireg) {
kvn@3882 792 case Op_VecS:
kvn@3882 793 __ movdl(Address(rsp, stack_offset), as_XMMRegister(Matcher::_regEncode[reg]));
kvn@3882 794 break;
kvn@3882 795 case Op_VecD:
kvn@3882 796 __ movq(Address(rsp, stack_offset), as_XMMRegister(Matcher::_regEncode[reg]));
kvn@3882 797 break;
kvn@3882 798 case Op_VecX:
kvn@3882 799 __ movdqu(Address(rsp, stack_offset), as_XMMRegister(Matcher::_regEncode[reg]));
kvn@3882 800 break;
kvn@3882 801 case Op_VecY:
kvn@3882 802 __ vmovdqu(Address(rsp, stack_offset), as_XMMRegister(Matcher::_regEncode[reg]));
kvn@3882 803 break;
kvn@3882 804 default:
kvn@3882 805 ShouldNotReachHere();
kvn@3882 806 }
kvn@3882 807 }
kvn@3882 808 int size = __ offset() - offset;
kvn@3882 809 #ifdef ASSERT
kvn@3882 810 int offset_size = (stack_offset == 0) ? 0 : ((stack_offset < 0x80) ? 1 : 4);
kvn@3882 811 // VEX_2bytes prefix is used if UseAVX > 0, so it takes the same 2 bytes as SIMD prefix.
kvn@3882 812 assert(!do_size || size == (5+offset_size), "incorrect size calculattion");
kvn@3882 813 #endif
kvn@3882 814 return size;
kvn@3882 815 #ifndef PRODUCT
kvn@3882 816 } else if (!do_size) {
kvn@3882 817 if (is_load) {
kvn@3882 818 switch (ireg) {
kvn@3882 819 case Op_VecS:
kvn@3882 820 st->print("movd %s,[rsp + %d]\t# spill", Matcher::regName[reg], stack_offset);
kvn@3882 821 break;
kvn@3882 822 case Op_VecD:
kvn@3882 823 st->print("movq %s,[rsp + %d]\t# spill", Matcher::regName[reg], stack_offset);
kvn@3882 824 break;
kvn@3882 825 case Op_VecX:
kvn@3882 826 st->print("movdqu %s,[rsp + %d]\t# spill", Matcher::regName[reg], stack_offset);
kvn@3882 827 break;
kvn@3882 828 case Op_VecY:
kvn@3882 829 st->print("vmovdqu %s,[rsp + %d]\t# spill", Matcher::regName[reg], stack_offset);
kvn@3882 830 break;
kvn@3882 831 default:
kvn@3882 832 ShouldNotReachHere();
kvn@3882 833 }
kvn@3882 834 } else { // store
kvn@3882 835 switch (ireg) {
kvn@3882 836 case Op_VecS:
kvn@3882 837 st->print("movd [rsp + %d],%s\t# spill", stack_offset, Matcher::regName[reg]);
kvn@3882 838 break;
kvn@3882 839 case Op_VecD:
kvn@3882 840 st->print("movq [rsp + %d],%s\t# spill", stack_offset, Matcher::regName[reg]);
kvn@3882 841 break;
kvn@3882 842 case Op_VecX:
kvn@3882 843 st->print("movdqu [rsp + %d],%s\t# spill", stack_offset, Matcher::regName[reg]);
kvn@3882 844 break;
kvn@3882 845 case Op_VecY:
kvn@3882 846 st->print("vmovdqu [rsp + %d],%s\t# spill", stack_offset, Matcher::regName[reg]);
kvn@3882 847 break;
kvn@3882 848 default:
kvn@3882 849 ShouldNotReachHere();
kvn@3882 850 }
kvn@3882 851 }
kvn@3882 852 #endif
kvn@3882 853 }
kvn@3882 854 int offset_size = (stack_offset == 0) ? 0 : ((stack_offset < 0x80) ? 1 : 4);
kvn@3882 855 // VEX_2bytes prefix is used if UseAVX > 0, so it takes the same 2 bytes as SIMD prefix.
kvn@3882 856 return 5+offset_size;
kvn@3882 857 }
kvn@3882 858
kvn@3882 859 static inline jfloat replicate4_imm(int con, int width) {
kvn@3882 860 // Load a constant of "width" (in bytes) and replicate it to fill 32bit.
kvn@3882 861 assert(width == 1 || width == 2, "only byte or short types here");
kvn@3882 862 int bit_width = width * 8;
kvn@3882 863 jint val = con;
kvn@3882 864 val &= (1 << bit_width) - 1; // mask off sign bits
kvn@3882 865 while(bit_width < 32) {
kvn@3882 866 val |= (val << bit_width);
kvn@3882 867 bit_width <<= 1;
kvn@3882 868 }
kvn@3882 869 jfloat fval = *((jfloat*) &val); // coerce to float type
kvn@3882 870 return fval;
kvn@3882 871 }
kvn@3882 872
kvn@3882 873 static inline jdouble replicate8_imm(int con, int width) {
kvn@3882 874 // Load a constant of "width" (in bytes) and replicate it to fill 64bit.
kvn@3882 875 assert(width == 1 || width == 2 || width == 4, "only byte, short or int types here");
kvn@3882 876 int bit_width = width * 8;
kvn@3882 877 jlong val = con;
kvn@3882 878 val &= (((jlong) 1) << bit_width) - 1; // mask off sign bits
kvn@3882 879 while(bit_width < 64) {
kvn@3882 880 val |= (val << bit_width);
kvn@3882 881 bit_width <<= 1;
kvn@3882 882 }
kvn@3882 883 jdouble dval = *((jdouble*) &val); // coerce to double type
kvn@3882 884 return dval;
kvn@3882 885 }
kvn@3882 886
kvn@3577 887 #ifndef PRODUCT
kvn@3577 888 void MachNopNode::format(PhaseRegAlloc*, outputStream* st) const {
kvn@3577 889 st->print("nop \t# %d bytes pad for loops and calls", _count);
kvn@3577 890 }
kvn@3577 891 #endif
kvn@3577 892
kvn@3577 893 void MachNopNode::emit(CodeBuffer &cbuf, PhaseRegAlloc*) const {
kvn@3577 894 MacroAssembler _masm(&cbuf);
kvn@3577 895 __ nop(_count);
kvn@3577 896 }
kvn@3577 897
kvn@3577 898 uint MachNopNode::size(PhaseRegAlloc*) const {
kvn@3577 899 return _count;
kvn@3577 900 }
kvn@3577 901
kvn@3577 902 #ifndef PRODUCT
kvn@3577 903 void MachBreakpointNode::format(PhaseRegAlloc*, outputStream* st) const {
kvn@3577 904 st->print("# breakpoint");
kvn@3577 905 }
kvn@3577 906 #endif
kvn@3577 907
kvn@3577 908 void MachBreakpointNode::emit(CodeBuffer &cbuf, PhaseRegAlloc* ra_) const {
kvn@3577 909 MacroAssembler _masm(&cbuf);
kvn@3577 910 __ int3();
kvn@3577 911 }
kvn@3577 912
kvn@3577 913 uint MachBreakpointNode::size(PhaseRegAlloc* ra_) const {
kvn@3577 914 return MachNode::size(ra_);
kvn@3577 915 }
kvn@3577 916
kvn@3577 917 %}
kvn@3577 918
kvn@3577 919 encode %{
kvn@3577 920
kvn@3577 921 enc_class call_epilog %{
kvn@3577 922 if (VerifyStackAtCalls) {
kvn@3577 923 // Check that stack depth is unchanged: find majik cookie on stack
kvn@3577 924 int framesize = ra_->reg2offset_unchecked(OptoReg::add(ra_->_matcher._old_SP, -3*VMRegImpl::slots_per_word));
kvn@3577 925 MacroAssembler _masm(&cbuf);
kvn@3577 926 Label L;
kvn@3577 927 __ cmpptr(Address(rsp, framesize), (int32_t)0xbadb100d);
kvn@3577 928 __ jccb(Assembler::equal, L);
kvn@3577 929 // Die if stack mismatch
kvn@3577 930 __ int3();
kvn@3577 931 __ bind(L);
kvn@3577 932 }
kvn@3577 933 %}
kvn@3577 934
kvn@3390 935 %}
kvn@3390 936
kvn@3882 937
kvn@3882 938 //----------OPERANDS-----------------------------------------------------------
kvn@3882 939 // Operand definitions must precede instruction definitions for correct parsing
kvn@3882 940 // in the ADLC because operands constitute user defined types which are used in
kvn@3882 941 // instruction definitions.
kvn@3882 942
kvn@3882 943 // Vectors
kvn@3882 944 operand vecS() %{
kvn@3882 945 constraint(ALLOC_IN_RC(vectors_reg));
kvn@3882 946 match(VecS);
kvn@3882 947
kvn@3882 948 format %{ %}
kvn@3882 949 interface(REG_INTER);
kvn@3882 950 %}
kvn@3882 951
kvn@3882 952 operand vecD() %{
kvn@3882 953 constraint(ALLOC_IN_RC(vectord_reg));
kvn@3882 954 match(VecD);
kvn@3882 955
kvn@3882 956 format %{ %}
kvn@3882 957 interface(REG_INTER);
kvn@3882 958 %}
kvn@3882 959
kvn@3882 960 operand vecX() %{
kvn@3882 961 constraint(ALLOC_IN_RC(vectorx_reg));
kvn@3882 962 match(VecX);
kvn@3882 963
kvn@3882 964 format %{ %}
kvn@3882 965 interface(REG_INTER);
kvn@3882 966 %}
kvn@3882 967
kvn@3882 968 operand vecY() %{
kvn@3882 969 constraint(ALLOC_IN_RC(vectory_reg));
kvn@3882 970 match(VecY);
kvn@3882 971
kvn@3882 972 format %{ %}
kvn@3882 973 interface(REG_INTER);
kvn@3882 974 %}
kvn@3882 975
kvn@3882 976
kvn@3390 977 // INSTRUCTIONS -- Platform independent definitions (same for 32- and 64-bit)
kvn@3390 978
kvn@3577 979 // ============================================================================
kvn@3577 980
kvn@3577 981 instruct ShouldNotReachHere() %{
kvn@3577 982 match(Halt);
kvn@3577 983 format %{ "int3\t# ShouldNotReachHere" %}
kvn@3577 984 ins_encode %{
kvn@3577 985 __ int3();
kvn@3577 986 %}
kvn@3577 987 ins_pipe(pipe_slow);
kvn@3577 988 %}
kvn@3577 989
kvn@3577 990 // ============================================================================
kvn@3577 991
kvn@3390 992 instruct addF_reg(regF dst, regF src) %{
kvn@3390 993 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 994 match(Set dst (AddF dst src));
kvn@3390 995
kvn@3390 996 format %{ "addss $dst, $src" %}
kvn@3390 997 ins_cost(150);
kvn@3390 998 ins_encode %{
kvn@3390 999 __ addss($dst$$XMMRegister, $src$$XMMRegister);
kvn@3390 1000 %}
kvn@3390 1001 ins_pipe(pipe_slow);
kvn@3390 1002 %}
kvn@3390 1003
kvn@3390 1004 instruct addF_mem(regF dst, memory src) %{
kvn@3390 1005 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1006 match(Set dst (AddF dst (LoadF src)));
kvn@3390 1007
kvn@3390 1008 format %{ "addss $dst, $src" %}
kvn@3390 1009 ins_cost(150);
kvn@3390 1010 ins_encode %{
kvn@3390 1011 __ addss($dst$$XMMRegister, $src$$Address);
kvn@3390 1012 %}
kvn@3390 1013 ins_pipe(pipe_slow);
kvn@3390 1014 %}
kvn@3390 1015
kvn@3390 1016 instruct addF_imm(regF dst, immF con) %{
kvn@3390 1017 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1018 match(Set dst (AddF dst con));
kvn@3390 1019 format %{ "addss $dst, [$constantaddress]\t# load from constant table: float=$con" %}
kvn@3390 1020 ins_cost(150);
kvn@3390 1021 ins_encode %{
kvn@3390 1022 __ addss($dst$$XMMRegister, $constantaddress($con));
kvn@3390 1023 %}
kvn@3390 1024 ins_pipe(pipe_slow);
kvn@3390 1025 %}
kvn@3390 1026
kvn@3929 1027 instruct addF_reg_reg(regF dst, regF src1, regF src2) %{
kvn@3390 1028 predicate(UseAVX > 0);
kvn@3390 1029 match(Set dst (AddF src1 src2));
kvn@3390 1030
kvn@3390 1031 format %{ "vaddss $dst, $src1, $src2" %}
kvn@3390 1032 ins_cost(150);
kvn@3390 1033 ins_encode %{
kvn@3390 1034 __ vaddss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister);
kvn@3390 1035 %}
kvn@3390 1036 ins_pipe(pipe_slow);
kvn@3390 1037 %}
kvn@3390 1038
kvn@3929 1039 instruct addF_reg_mem(regF dst, regF src1, memory src2) %{
kvn@3390 1040 predicate(UseAVX > 0);
kvn@3390 1041 match(Set dst (AddF src1 (LoadF src2)));
kvn@3390 1042
kvn@3390 1043 format %{ "vaddss $dst, $src1, $src2" %}
kvn@3390 1044 ins_cost(150);
kvn@3390 1045 ins_encode %{
kvn@3390 1046 __ vaddss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address);
kvn@3390 1047 %}
kvn@3390 1048 ins_pipe(pipe_slow);
kvn@3390 1049 %}
kvn@3390 1050
kvn@3929 1051 instruct addF_reg_imm(regF dst, regF src, immF con) %{
kvn@3390 1052 predicate(UseAVX > 0);
kvn@3390 1053 match(Set dst (AddF src con));
kvn@3390 1054
kvn@3390 1055 format %{ "vaddss $dst, $src, [$constantaddress]\t# load from constant table: float=$con" %}
kvn@3390 1056 ins_cost(150);
kvn@3390 1057 ins_encode %{
kvn@3390 1058 __ vaddss($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con));
kvn@3390 1059 %}
kvn@3390 1060 ins_pipe(pipe_slow);
kvn@3390 1061 %}
kvn@3390 1062
kvn@3390 1063 instruct addD_reg(regD dst, regD src) %{
kvn@3390 1064 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1065 match(Set dst (AddD dst src));
kvn@3390 1066
kvn@3390 1067 format %{ "addsd $dst, $src" %}
kvn@3390 1068 ins_cost(150);
kvn@3390 1069 ins_encode %{
kvn@3390 1070 __ addsd($dst$$XMMRegister, $src$$XMMRegister);
kvn@3390 1071 %}
kvn@3390 1072 ins_pipe(pipe_slow);
kvn@3390 1073 %}
kvn@3390 1074
kvn@3390 1075 instruct addD_mem(regD dst, memory src) %{
kvn@3390 1076 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1077 match(Set dst (AddD dst (LoadD src)));
kvn@3390 1078
kvn@3390 1079 format %{ "addsd $dst, $src" %}
kvn@3390 1080 ins_cost(150);
kvn@3390 1081 ins_encode %{
kvn@3390 1082 __ addsd($dst$$XMMRegister, $src$$Address);
kvn@3390 1083 %}
kvn@3390 1084 ins_pipe(pipe_slow);
kvn@3390 1085 %}
kvn@3390 1086
kvn@3390 1087 instruct addD_imm(regD dst, immD con) %{
kvn@3390 1088 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1089 match(Set dst (AddD dst con));
kvn@3390 1090 format %{ "addsd $dst, [$constantaddress]\t# load from constant table: double=$con" %}
kvn@3390 1091 ins_cost(150);
kvn@3390 1092 ins_encode %{
kvn@3390 1093 __ addsd($dst$$XMMRegister, $constantaddress($con));
kvn@3390 1094 %}
kvn@3390 1095 ins_pipe(pipe_slow);
kvn@3390 1096 %}
kvn@3390 1097
kvn@3929 1098 instruct addD_reg_reg(regD dst, regD src1, regD src2) %{
kvn@3390 1099 predicate(UseAVX > 0);
kvn@3390 1100 match(Set dst (AddD src1 src2));
kvn@3390 1101
kvn@3390 1102 format %{ "vaddsd $dst, $src1, $src2" %}
kvn@3390 1103 ins_cost(150);
kvn@3390 1104 ins_encode %{
kvn@3390 1105 __ vaddsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister);
kvn@3390 1106 %}
kvn@3390 1107 ins_pipe(pipe_slow);
kvn@3390 1108 %}
kvn@3390 1109
kvn@3929 1110 instruct addD_reg_mem(regD dst, regD src1, memory src2) %{
kvn@3390 1111 predicate(UseAVX > 0);
kvn@3390 1112 match(Set dst (AddD src1 (LoadD src2)));
kvn@3390 1113
kvn@3390 1114 format %{ "vaddsd $dst, $src1, $src2" %}
kvn@3390 1115 ins_cost(150);
kvn@3390 1116 ins_encode %{
kvn@3390 1117 __ vaddsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address);
kvn@3390 1118 %}
kvn@3390 1119 ins_pipe(pipe_slow);
kvn@3390 1120 %}
kvn@3390 1121
kvn@3929 1122 instruct addD_reg_imm(regD dst, regD src, immD con) %{
kvn@3390 1123 predicate(UseAVX > 0);
kvn@3390 1124 match(Set dst (AddD src con));
kvn@3390 1125
kvn@3390 1126 format %{ "vaddsd $dst, $src, [$constantaddress]\t# load from constant table: double=$con" %}
kvn@3390 1127 ins_cost(150);
kvn@3390 1128 ins_encode %{
kvn@3390 1129 __ vaddsd($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con));
kvn@3390 1130 %}
kvn@3390 1131 ins_pipe(pipe_slow);
kvn@3390 1132 %}
kvn@3390 1133
kvn@3390 1134 instruct subF_reg(regF dst, regF src) %{
kvn@3390 1135 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1136 match(Set dst (SubF dst src));
kvn@3390 1137
kvn@3390 1138 format %{ "subss $dst, $src" %}
kvn@3390 1139 ins_cost(150);
kvn@3390 1140 ins_encode %{
kvn@3390 1141 __ subss($dst$$XMMRegister, $src$$XMMRegister);
kvn@3390 1142 %}
kvn@3390 1143 ins_pipe(pipe_slow);
kvn@3390 1144 %}
kvn@3390 1145
kvn@3390 1146 instruct subF_mem(regF dst, memory src) %{
kvn@3390 1147 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1148 match(Set dst (SubF dst (LoadF src)));
kvn@3390 1149
kvn@3390 1150 format %{ "subss $dst, $src" %}
kvn@3390 1151 ins_cost(150);
kvn@3390 1152 ins_encode %{
kvn@3390 1153 __ subss($dst$$XMMRegister, $src$$Address);
kvn@3390 1154 %}
kvn@3390 1155 ins_pipe(pipe_slow);
kvn@3390 1156 %}
kvn@3390 1157
kvn@3390 1158 instruct subF_imm(regF dst, immF con) %{
kvn@3390 1159 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1160 match(Set dst (SubF dst con));
kvn@3390 1161 format %{ "subss $dst, [$constantaddress]\t# load from constant table: float=$con" %}
kvn@3390 1162 ins_cost(150);
kvn@3390 1163 ins_encode %{
kvn@3390 1164 __ subss($dst$$XMMRegister, $constantaddress($con));
kvn@3390 1165 %}
kvn@3390 1166 ins_pipe(pipe_slow);
kvn@3390 1167 %}
kvn@3390 1168
kvn@3929 1169 instruct subF_reg_reg(regF dst, regF src1, regF src2) %{
kvn@3390 1170 predicate(UseAVX > 0);
kvn@3390 1171 match(Set dst (SubF src1 src2));
kvn@3390 1172
kvn@3390 1173 format %{ "vsubss $dst, $src1, $src2" %}
kvn@3390 1174 ins_cost(150);
kvn@3390 1175 ins_encode %{
kvn@3390 1176 __ vsubss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister);
kvn@3390 1177 %}
kvn@3390 1178 ins_pipe(pipe_slow);
kvn@3390 1179 %}
kvn@3390 1180
kvn@3929 1181 instruct subF_reg_mem(regF dst, regF src1, memory src2) %{
kvn@3390 1182 predicate(UseAVX > 0);
kvn@3390 1183 match(Set dst (SubF src1 (LoadF src2)));
kvn@3390 1184
kvn@3390 1185 format %{ "vsubss $dst, $src1, $src2" %}
kvn@3390 1186 ins_cost(150);
kvn@3390 1187 ins_encode %{
kvn@3390 1188 __ vsubss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address);
kvn@3390 1189 %}
kvn@3390 1190 ins_pipe(pipe_slow);
kvn@3390 1191 %}
kvn@3390 1192
kvn@3929 1193 instruct subF_reg_imm(regF dst, regF src, immF con) %{
kvn@3390 1194 predicate(UseAVX > 0);
kvn@3390 1195 match(Set dst (SubF src con));
kvn@3390 1196
kvn@3390 1197 format %{ "vsubss $dst, $src, [$constantaddress]\t# load from constant table: float=$con" %}
kvn@3390 1198 ins_cost(150);
kvn@3390 1199 ins_encode %{
kvn@3390 1200 __ vsubss($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con));
kvn@3390 1201 %}
kvn@3390 1202 ins_pipe(pipe_slow);
kvn@3390 1203 %}
kvn@3390 1204
kvn@3390 1205 instruct subD_reg(regD dst, regD src) %{
kvn@3390 1206 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1207 match(Set dst (SubD dst src));
kvn@3390 1208
kvn@3390 1209 format %{ "subsd $dst, $src" %}
kvn@3390 1210 ins_cost(150);
kvn@3390 1211 ins_encode %{
kvn@3390 1212 __ subsd($dst$$XMMRegister, $src$$XMMRegister);
kvn@3390 1213 %}
kvn@3390 1214 ins_pipe(pipe_slow);
kvn@3390 1215 %}
kvn@3390 1216
kvn@3390 1217 instruct subD_mem(regD dst, memory src) %{
kvn@3390 1218 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1219 match(Set dst (SubD dst (LoadD src)));
kvn@3390 1220
kvn@3390 1221 format %{ "subsd $dst, $src" %}
kvn@3390 1222 ins_cost(150);
kvn@3390 1223 ins_encode %{
kvn@3390 1224 __ subsd($dst$$XMMRegister, $src$$Address);
kvn@3390 1225 %}
kvn@3390 1226 ins_pipe(pipe_slow);
kvn@3390 1227 %}
kvn@3390 1228
kvn@3390 1229 instruct subD_imm(regD dst, immD con) %{
kvn@3390 1230 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1231 match(Set dst (SubD dst con));
kvn@3390 1232 format %{ "subsd $dst, [$constantaddress]\t# load from constant table: double=$con" %}
kvn@3390 1233 ins_cost(150);
kvn@3390 1234 ins_encode %{
kvn@3390 1235 __ subsd($dst$$XMMRegister, $constantaddress($con));
kvn@3390 1236 %}
kvn@3390 1237 ins_pipe(pipe_slow);
kvn@3390 1238 %}
kvn@3390 1239
kvn@3929 1240 instruct subD_reg_reg(regD dst, regD src1, regD src2) %{
kvn@3390 1241 predicate(UseAVX > 0);
kvn@3390 1242 match(Set dst (SubD src1 src2));
kvn@3390 1243
kvn@3390 1244 format %{ "vsubsd $dst, $src1, $src2" %}
kvn@3390 1245 ins_cost(150);
kvn@3390 1246 ins_encode %{
kvn@3390 1247 __ vsubsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister);
kvn@3390 1248 %}
kvn@3390 1249 ins_pipe(pipe_slow);
kvn@3390 1250 %}
kvn@3390 1251
kvn@3929 1252 instruct subD_reg_mem(regD dst, regD src1, memory src2) %{
kvn@3390 1253 predicate(UseAVX > 0);
kvn@3390 1254 match(Set dst (SubD src1 (LoadD src2)));
kvn@3390 1255
kvn@3390 1256 format %{ "vsubsd $dst, $src1, $src2" %}
kvn@3390 1257 ins_cost(150);
kvn@3390 1258 ins_encode %{
kvn@3390 1259 __ vsubsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address);
kvn@3390 1260 %}
kvn@3390 1261 ins_pipe(pipe_slow);
kvn@3390 1262 %}
kvn@3390 1263
kvn@3929 1264 instruct subD_reg_imm(regD dst, regD src, immD con) %{
kvn@3390 1265 predicate(UseAVX > 0);
kvn@3390 1266 match(Set dst (SubD src con));
kvn@3390 1267
kvn@3390 1268 format %{ "vsubsd $dst, $src, [$constantaddress]\t# load from constant table: double=$con" %}
kvn@3390 1269 ins_cost(150);
kvn@3390 1270 ins_encode %{
kvn@3390 1271 __ vsubsd($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con));
kvn@3390 1272 %}
kvn@3390 1273 ins_pipe(pipe_slow);
kvn@3390 1274 %}
kvn@3390 1275
kvn@3390 1276 instruct mulF_reg(regF dst, regF src) %{
kvn@3390 1277 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1278 match(Set dst (MulF dst src));
kvn@3390 1279
kvn@3390 1280 format %{ "mulss $dst, $src" %}
kvn@3390 1281 ins_cost(150);
kvn@3390 1282 ins_encode %{
kvn@3390 1283 __ mulss($dst$$XMMRegister, $src$$XMMRegister);
kvn@3390 1284 %}
kvn@3390 1285 ins_pipe(pipe_slow);
kvn@3390 1286 %}
kvn@3390 1287
kvn@3390 1288 instruct mulF_mem(regF dst, memory src) %{
kvn@3390 1289 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1290 match(Set dst (MulF dst (LoadF src)));
kvn@3390 1291
kvn@3390 1292 format %{ "mulss $dst, $src" %}
kvn@3390 1293 ins_cost(150);
kvn@3390 1294 ins_encode %{
kvn@3390 1295 __ mulss($dst$$XMMRegister, $src$$Address);
kvn@3390 1296 %}
kvn@3390 1297 ins_pipe(pipe_slow);
kvn@3390 1298 %}
kvn@3390 1299
kvn@3390 1300 instruct mulF_imm(regF dst, immF con) %{
kvn@3390 1301 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1302 match(Set dst (MulF dst con));
kvn@3390 1303 format %{ "mulss $dst, [$constantaddress]\t# load from constant table: float=$con" %}
kvn@3390 1304 ins_cost(150);
kvn@3390 1305 ins_encode %{
kvn@3390 1306 __ mulss($dst$$XMMRegister, $constantaddress($con));
kvn@3390 1307 %}
kvn@3390 1308 ins_pipe(pipe_slow);
kvn@3390 1309 %}
kvn@3390 1310
kvn@3929 1311 instruct mulF_reg_reg(regF dst, regF src1, regF src2) %{
kvn@3390 1312 predicate(UseAVX > 0);
kvn@3390 1313 match(Set dst (MulF src1 src2));
kvn@3390 1314
kvn@3390 1315 format %{ "vmulss $dst, $src1, $src2" %}
kvn@3390 1316 ins_cost(150);
kvn@3390 1317 ins_encode %{
kvn@3390 1318 __ vmulss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister);
kvn@3390 1319 %}
kvn@3390 1320 ins_pipe(pipe_slow);
kvn@3390 1321 %}
kvn@3390 1322
kvn@3929 1323 instruct mulF_reg_mem(regF dst, regF src1, memory src2) %{
kvn@3390 1324 predicate(UseAVX > 0);
kvn@3390 1325 match(Set dst (MulF src1 (LoadF src2)));
kvn@3390 1326
kvn@3390 1327 format %{ "vmulss $dst, $src1, $src2" %}
kvn@3390 1328 ins_cost(150);
kvn@3390 1329 ins_encode %{
kvn@3390 1330 __ vmulss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address);
kvn@3390 1331 %}
kvn@3390 1332 ins_pipe(pipe_slow);
kvn@3390 1333 %}
kvn@3390 1334
kvn@3929 1335 instruct mulF_reg_imm(regF dst, regF src, immF con) %{
kvn@3390 1336 predicate(UseAVX > 0);
kvn@3390 1337 match(Set dst (MulF src con));
kvn@3390 1338
kvn@3390 1339 format %{ "vmulss $dst, $src, [$constantaddress]\t# load from constant table: float=$con" %}
kvn@3390 1340 ins_cost(150);
kvn@3390 1341 ins_encode %{
kvn@3390 1342 __ vmulss($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con));
kvn@3390 1343 %}
kvn@3390 1344 ins_pipe(pipe_slow);
kvn@3390 1345 %}
kvn@3390 1346
kvn@3390 1347 instruct mulD_reg(regD dst, regD src) %{
kvn@3390 1348 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1349 match(Set dst (MulD dst src));
kvn@3390 1350
kvn@3390 1351 format %{ "mulsd $dst, $src" %}
kvn@3390 1352 ins_cost(150);
kvn@3390 1353 ins_encode %{
kvn@3390 1354 __ mulsd($dst$$XMMRegister, $src$$XMMRegister);
kvn@3390 1355 %}
kvn@3390 1356 ins_pipe(pipe_slow);
kvn@3390 1357 %}
kvn@3390 1358
kvn@3390 1359 instruct mulD_mem(regD dst, memory src) %{
kvn@3390 1360 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1361 match(Set dst (MulD dst (LoadD src)));
kvn@3390 1362
kvn@3390 1363 format %{ "mulsd $dst, $src" %}
kvn@3390 1364 ins_cost(150);
kvn@3390 1365 ins_encode %{
kvn@3390 1366 __ mulsd($dst$$XMMRegister, $src$$Address);
kvn@3390 1367 %}
kvn@3390 1368 ins_pipe(pipe_slow);
kvn@3390 1369 %}
kvn@3390 1370
kvn@3390 1371 instruct mulD_imm(regD dst, immD con) %{
kvn@3390 1372 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1373 match(Set dst (MulD dst con));
kvn@3390 1374 format %{ "mulsd $dst, [$constantaddress]\t# load from constant table: double=$con" %}
kvn@3390 1375 ins_cost(150);
kvn@3390 1376 ins_encode %{
kvn@3390 1377 __ mulsd($dst$$XMMRegister, $constantaddress($con));
kvn@3390 1378 %}
kvn@3390 1379 ins_pipe(pipe_slow);
kvn@3390 1380 %}
kvn@3390 1381
kvn@3929 1382 instruct mulD_reg_reg(regD dst, regD src1, regD src2) %{
kvn@3390 1383 predicate(UseAVX > 0);
kvn@3390 1384 match(Set dst (MulD src1 src2));
kvn@3390 1385
kvn@3390 1386 format %{ "vmulsd $dst, $src1, $src2" %}
kvn@3390 1387 ins_cost(150);
kvn@3390 1388 ins_encode %{
kvn@3390 1389 __ vmulsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister);
kvn@3390 1390 %}
kvn@3390 1391 ins_pipe(pipe_slow);
kvn@3390 1392 %}
kvn@3390 1393
kvn@3929 1394 instruct mulD_reg_mem(regD dst, regD src1, memory src2) %{
kvn@3390 1395 predicate(UseAVX > 0);
kvn@3390 1396 match(Set dst (MulD src1 (LoadD src2)));
kvn@3390 1397
kvn@3390 1398 format %{ "vmulsd $dst, $src1, $src2" %}
kvn@3390 1399 ins_cost(150);
kvn@3390 1400 ins_encode %{
kvn@3390 1401 __ vmulsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address);
kvn@3390 1402 %}
kvn@3390 1403 ins_pipe(pipe_slow);
kvn@3390 1404 %}
kvn@3390 1405
kvn@3929 1406 instruct mulD_reg_imm(regD dst, regD src, immD con) %{
kvn@3390 1407 predicate(UseAVX > 0);
kvn@3390 1408 match(Set dst (MulD src con));
kvn@3390 1409
kvn@3390 1410 format %{ "vmulsd $dst, $src, [$constantaddress]\t# load from constant table: double=$con" %}
kvn@3390 1411 ins_cost(150);
kvn@3390 1412 ins_encode %{
kvn@3390 1413 __ vmulsd($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con));
kvn@3390 1414 %}
kvn@3390 1415 ins_pipe(pipe_slow);
kvn@3390 1416 %}
kvn@3390 1417
kvn@3390 1418 instruct divF_reg(regF dst, regF src) %{
kvn@3390 1419 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1420 match(Set dst (DivF dst src));
kvn@3390 1421
kvn@3390 1422 format %{ "divss $dst, $src" %}
kvn@3390 1423 ins_cost(150);
kvn@3390 1424 ins_encode %{
kvn@3390 1425 __ divss($dst$$XMMRegister, $src$$XMMRegister);
kvn@3390 1426 %}
kvn@3390 1427 ins_pipe(pipe_slow);
kvn@3390 1428 %}
kvn@3390 1429
kvn@3390 1430 instruct divF_mem(regF dst, memory src) %{
kvn@3390 1431 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1432 match(Set dst (DivF dst (LoadF src)));
kvn@3390 1433
kvn@3390 1434 format %{ "divss $dst, $src" %}
kvn@3390 1435 ins_cost(150);
kvn@3390 1436 ins_encode %{
kvn@3390 1437 __ divss($dst$$XMMRegister, $src$$Address);
kvn@3390 1438 %}
kvn@3390 1439 ins_pipe(pipe_slow);
kvn@3390 1440 %}
kvn@3390 1441
kvn@3390 1442 instruct divF_imm(regF dst, immF con) %{
kvn@3390 1443 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1444 match(Set dst (DivF dst con));
kvn@3390 1445 format %{ "divss $dst, [$constantaddress]\t# load from constant table: float=$con" %}
kvn@3390 1446 ins_cost(150);
kvn@3390 1447 ins_encode %{
kvn@3390 1448 __ divss($dst$$XMMRegister, $constantaddress($con));
kvn@3390 1449 %}
kvn@3390 1450 ins_pipe(pipe_slow);
kvn@3390 1451 %}
kvn@3390 1452
kvn@3929 1453 instruct divF_reg_reg(regF dst, regF src1, regF src2) %{
kvn@3390 1454 predicate(UseAVX > 0);
kvn@3390 1455 match(Set dst (DivF src1 src2));
kvn@3390 1456
kvn@3390 1457 format %{ "vdivss $dst, $src1, $src2" %}
kvn@3390 1458 ins_cost(150);
kvn@3390 1459 ins_encode %{
kvn@3390 1460 __ vdivss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister);
kvn@3390 1461 %}
kvn@3390 1462 ins_pipe(pipe_slow);
kvn@3390 1463 %}
kvn@3390 1464
kvn@3929 1465 instruct divF_reg_mem(regF dst, regF src1, memory src2) %{
kvn@3390 1466 predicate(UseAVX > 0);
kvn@3390 1467 match(Set dst (DivF src1 (LoadF src2)));
kvn@3390 1468
kvn@3390 1469 format %{ "vdivss $dst, $src1, $src2" %}
kvn@3390 1470 ins_cost(150);
kvn@3390 1471 ins_encode %{
kvn@3390 1472 __ vdivss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address);
kvn@3390 1473 %}
kvn@3390 1474 ins_pipe(pipe_slow);
kvn@3390 1475 %}
kvn@3390 1476
kvn@3929 1477 instruct divF_reg_imm(regF dst, regF src, immF con) %{
kvn@3390 1478 predicate(UseAVX > 0);
kvn@3390 1479 match(Set dst (DivF src con));
kvn@3390 1480
kvn@3390 1481 format %{ "vdivss $dst, $src, [$constantaddress]\t# load from constant table: float=$con" %}
kvn@3390 1482 ins_cost(150);
kvn@3390 1483 ins_encode %{
kvn@3390 1484 __ vdivss($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con));
kvn@3390 1485 %}
kvn@3390 1486 ins_pipe(pipe_slow);
kvn@3390 1487 %}
kvn@3390 1488
kvn@3390 1489 instruct divD_reg(regD dst, regD src) %{
kvn@3390 1490 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1491 match(Set dst (DivD dst src));
kvn@3390 1492
kvn@3390 1493 format %{ "divsd $dst, $src" %}
kvn@3390 1494 ins_cost(150);
kvn@3390 1495 ins_encode %{
kvn@3390 1496 __ divsd($dst$$XMMRegister, $src$$XMMRegister);
kvn@3390 1497 %}
kvn@3390 1498 ins_pipe(pipe_slow);
kvn@3390 1499 %}
kvn@3390 1500
kvn@3390 1501 instruct divD_mem(regD dst, memory src) %{
kvn@3390 1502 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1503 match(Set dst (DivD dst (LoadD src)));
kvn@3390 1504
kvn@3390 1505 format %{ "divsd $dst, $src" %}
kvn@3390 1506 ins_cost(150);
kvn@3390 1507 ins_encode %{
kvn@3390 1508 __ divsd($dst$$XMMRegister, $src$$Address);
kvn@3390 1509 %}
kvn@3390 1510 ins_pipe(pipe_slow);
kvn@3390 1511 %}
kvn@3390 1512
kvn@3390 1513 instruct divD_imm(regD dst, immD con) %{
kvn@3390 1514 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1515 match(Set dst (DivD dst con));
kvn@3390 1516 format %{ "divsd $dst, [$constantaddress]\t# load from constant table: double=$con" %}
kvn@3390 1517 ins_cost(150);
kvn@3390 1518 ins_encode %{
kvn@3390 1519 __ divsd($dst$$XMMRegister, $constantaddress($con));
kvn@3390 1520 %}
kvn@3390 1521 ins_pipe(pipe_slow);
kvn@3390 1522 %}
kvn@3390 1523
kvn@3929 1524 instruct divD_reg_reg(regD dst, regD src1, regD src2) %{
kvn@3390 1525 predicate(UseAVX > 0);
kvn@3390 1526 match(Set dst (DivD src1 src2));
kvn@3390 1527
kvn@3390 1528 format %{ "vdivsd $dst, $src1, $src2" %}
kvn@3390 1529 ins_cost(150);
kvn@3390 1530 ins_encode %{
kvn@3390 1531 __ vdivsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister);
kvn@3390 1532 %}
kvn@3390 1533 ins_pipe(pipe_slow);
kvn@3390 1534 %}
kvn@3390 1535
kvn@3929 1536 instruct divD_reg_mem(regD dst, regD src1, memory src2) %{
kvn@3390 1537 predicate(UseAVX > 0);
kvn@3390 1538 match(Set dst (DivD src1 (LoadD src2)));
kvn@3390 1539
kvn@3390 1540 format %{ "vdivsd $dst, $src1, $src2" %}
kvn@3390 1541 ins_cost(150);
kvn@3390 1542 ins_encode %{
kvn@3390 1543 __ vdivsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address);
kvn@3390 1544 %}
kvn@3390 1545 ins_pipe(pipe_slow);
kvn@3390 1546 %}
kvn@3390 1547
kvn@3929 1548 instruct divD_reg_imm(regD dst, regD src, immD con) %{
kvn@3390 1549 predicate(UseAVX > 0);
kvn@3390 1550 match(Set dst (DivD src con));
kvn@3390 1551
kvn@3390 1552 format %{ "vdivsd $dst, $src, [$constantaddress]\t# load from constant table: double=$con" %}
kvn@3390 1553 ins_cost(150);
kvn@3390 1554 ins_encode %{
kvn@3390 1555 __ vdivsd($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con));
kvn@3390 1556 %}
kvn@3390 1557 ins_pipe(pipe_slow);
kvn@3390 1558 %}
kvn@3390 1559
kvn@3390 1560 instruct absF_reg(regF dst) %{
kvn@3390 1561 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1562 match(Set dst (AbsF dst));
kvn@3390 1563 ins_cost(150);
kvn@3390 1564 format %{ "andps $dst, [0x7fffffff]\t# abs float by sign masking" %}
kvn@3390 1565 ins_encode %{
kvn@3390 1566 __ andps($dst$$XMMRegister, ExternalAddress(float_signmask()));
kvn@3390 1567 %}
kvn@3390 1568 ins_pipe(pipe_slow);
kvn@3390 1569 %}
kvn@3390 1570
kvn@3929 1571 instruct absF_reg_reg(regF dst, regF src) %{
kvn@3390 1572 predicate(UseAVX > 0);
kvn@3390 1573 match(Set dst (AbsF src));
kvn@3390 1574 ins_cost(150);
kvn@3390 1575 format %{ "vandps $dst, $src, [0x7fffffff]\t# abs float by sign masking" %}
kvn@3390 1576 ins_encode %{
kvn@4001 1577 bool vector256 = false;
kvn@3390 1578 __ vandps($dst$$XMMRegister, $src$$XMMRegister,
kvn@4001 1579 ExternalAddress(float_signmask()), vector256);
kvn@3390 1580 %}
kvn@3390 1581 ins_pipe(pipe_slow);
kvn@3390 1582 %}
kvn@3390 1583
kvn@3390 1584 instruct absD_reg(regD dst) %{
kvn@3390 1585 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1586 match(Set dst (AbsD dst));
kvn@3390 1587 ins_cost(150);
kvn@3390 1588 format %{ "andpd $dst, [0x7fffffffffffffff]\t"
kvn@3390 1589 "# abs double by sign masking" %}
kvn@3390 1590 ins_encode %{
kvn@3390 1591 __ andpd($dst$$XMMRegister, ExternalAddress(double_signmask()));
kvn@3390 1592 %}
kvn@3390 1593 ins_pipe(pipe_slow);
kvn@3390 1594 %}
kvn@3390 1595
kvn@3929 1596 instruct absD_reg_reg(regD dst, regD src) %{
kvn@3390 1597 predicate(UseAVX > 0);
kvn@3390 1598 match(Set dst (AbsD src));
kvn@3390 1599 ins_cost(150);
kvn@3390 1600 format %{ "vandpd $dst, $src, [0x7fffffffffffffff]\t"
kvn@3390 1601 "# abs double by sign masking" %}
kvn@3390 1602 ins_encode %{
kvn@4001 1603 bool vector256 = false;
kvn@3390 1604 __ vandpd($dst$$XMMRegister, $src$$XMMRegister,
kvn@4001 1605 ExternalAddress(double_signmask()), vector256);
kvn@3390 1606 %}
kvn@3390 1607 ins_pipe(pipe_slow);
kvn@3390 1608 %}
kvn@3390 1609
kvn@3390 1610 instruct negF_reg(regF dst) %{
kvn@3390 1611 predicate((UseSSE>=1) && (UseAVX == 0));
kvn@3390 1612 match(Set dst (NegF dst));
kvn@3390 1613 ins_cost(150);
kvn@3390 1614 format %{ "xorps $dst, [0x80000000]\t# neg float by sign flipping" %}
kvn@3390 1615 ins_encode %{
kvn@3390 1616 __ xorps($dst$$XMMRegister, ExternalAddress(float_signflip()));
kvn@3390 1617 %}
kvn@3390 1618 ins_pipe(pipe_slow);
kvn@3390 1619 %}
kvn@3390 1620
kvn@3929 1621 instruct negF_reg_reg(regF dst, regF src) %{
kvn@3390 1622 predicate(UseAVX > 0);
kvn@3390 1623 match(Set dst (NegF src));
kvn@3390 1624 ins_cost(150);
kvn@3390 1625 format %{ "vxorps $dst, $src, [0x80000000]\t# neg float by sign flipping" %}
kvn@3390 1626 ins_encode %{
kvn@4001 1627 bool vector256 = false;
kvn@3390 1628 __ vxorps($dst$$XMMRegister, $src$$XMMRegister,
kvn@4001 1629 ExternalAddress(float_signflip()), vector256);
kvn@3390 1630 %}
kvn@3390 1631 ins_pipe(pipe_slow);
kvn@3390 1632 %}
kvn@3390 1633
kvn@3390 1634 instruct negD_reg(regD dst) %{
kvn@3390 1635 predicate((UseSSE>=2) && (UseAVX == 0));
kvn@3390 1636 match(Set dst (NegD dst));
kvn@3390 1637 ins_cost(150);
kvn@3390 1638 format %{ "xorpd $dst, [0x8000000000000000]\t"
kvn@3390 1639 "# neg double by sign flipping" %}
kvn@3390 1640 ins_encode %{
kvn@3390 1641 __ xorpd($dst$$XMMRegister, ExternalAddress(double_signflip()));
kvn@3390 1642 %}
kvn@3390 1643 ins_pipe(pipe_slow);
kvn@3390 1644 %}
kvn@3390 1645
kvn@3929 1646 instruct negD_reg_reg(regD dst, regD src) %{
kvn@3390 1647 predicate(UseAVX > 0);
kvn@3390 1648 match(Set dst (NegD src));
kvn@3390 1649 ins_cost(150);
kvn@3390 1650 format %{ "vxorpd $dst, $src, [0x8000000000000000]\t"
kvn@3390 1651 "# neg double by sign flipping" %}
kvn@3390 1652 ins_encode %{
kvn@4001 1653 bool vector256 = false;
kvn@3390 1654 __ vxorpd($dst$$XMMRegister, $src$$XMMRegister,
kvn@4001 1655 ExternalAddress(double_signflip()), vector256);
kvn@3390 1656 %}
kvn@3390 1657 ins_pipe(pipe_slow);
kvn@3390 1658 %}
kvn@3390 1659
kvn@3390 1660 instruct sqrtF_reg(regF dst, regF src) %{
kvn@3390 1661 predicate(UseSSE>=1);
kvn@3390 1662 match(Set dst (ConvD2F (SqrtD (ConvF2D src))));
kvn@3390 1663
kvn@3390 1664 format %{ "sqrtss $dst, $src" %}
kvn@3390 1665 ins_cost(150);
kvn@3390 1666 ins_encode %{
kvn@3390 1667 __ sqrtss($dst$$XMMRegister, $src$$XMMRegister);
kvn@3390 1668 %}
kvn@3390 1669 ins_pipe(pipe_slow);
kvn@3390 1670 %}
kvn@3390 1671
kvn@3390 1672 instruct sqrtF_mem(regF dst, memory src) %{
kvn@3390 1673 predicate(UseSSE>=1);
kvn@3390 1674 match(Set dst (ConvD2F (SqrtD (ConvF2D (LoadF src)))));
kvn@3390 1675
kvn@3390 1676 format %{ "sqrtss $dst, $src" %}
kvn@3390 1677 ins_cost(150);
kvn@3390 1678 ins_encode %{
kvn@3390 1679 __ sqrtss($dst$$XMMRegister, $src$$Address);
kvn@3390 1680 %}
kvn@3390 1681 ins_pipe(pipe_slow);
kvn@3390 1682 %}
kvn@3390 1683
kvn@3390 1684 instruct sqrtF_imm(regF dst, immF con) %{
kvn@3390 1685 predicate(UseSSE>=1);
kvn@3390 1686 match(Set dst (ConvD2F (SqrtD (ConvF2D con))));
kvn@3390 1687 format %{ "sqrtss $dst, [$constantaddress]\t# load from constant table: float=$con" %}
kvn@3390 1688 ins_cost(150);
kvn@3390 1689 ins_encode %{
kvn@3390 1690 __ sqrtss($dst$$XMMRegister, $constantaddress($con));
kvn@3390 1691 %}
kvn@3390 1692 ins_pipe(pipe_slow);
kvn@3390 1693 %}
kvn@3390 1694
kvn@3390 1695 instruct sqrtD_reg(regD dst, regD src) %{
kvn@3390 1696 predicate(UseSSE>=2);
kvn@3390 1697 match(Set dst (SqrtD src));
kvn@3390 1698
kvn@3390 1699 format %{ "sqrtsd $dst, $src" %}
kvn@3390 1700 ins_cost(150);
kvn@3390 1701 ins_encode %{
kvn@3390 1702 __ sqrtsd($dst$$XMMRegister, $src$$XMMRegister);
kvn@3390 1703 %}
kvn@3390 1704 ins_pipe(pipe_slow);
kvn@3390 1705 %}
kvn@3390 1706
kvn@3390 1707 instruct sqrtD_mem(regD dst, memory src) %{
kvn@3390 1708 predicate(UseSSE>=2);
kvn@3390 1709 match(Set dst (SqrtD (LoadD src)));
kvn@3390 1710
kvn@3390 1711 format %{ "sqrtsd $dst, $src" %}
kvn@3390 1712 ins_cost(150);
kvn@3390 1713 ins_encode %{
kvn@3390 1714 __ sqrtsd($dst$$XMMRegister, $src$$Address);
kvn@3390 1715 %}
kvn@3390 1716 ins_pipe(pipe_slow);
kvn@3390 1717 %}
kvn@3390 1718
kvn@3390 1719 instruct sqrtD_imm(regD dst, immD con) %{
kvn@3390 1720 predicate(UseSSE>=2);
kvn@3390 1721 match(Set dst (SqrtD con));
kvn@3390 1722 format %{ "sqrtsd $dst, [$constantaddress]\t# load from constant table: double=$con" %}
kvn@3390 1723 ins_cost(150);
kvn@3390 1724 ins_encode %{
kvn@3390 1725 __ sqrtsd($dst$$XMMRegister, $constantaddress($con));
kvn@3390 1726 %}
kvn@3390 1727 ins_pipe(pipe_slow);
kvn@3390 1728 %}
kvn@3390 1729
kvn@3882 1730
kvn@3882 1731 // ====================VECTOR INSTRUCTIONS=====================================
kvn@3882 1732
kvn@3882 1733 // Load vectors (4 bytes long)
kvn@3882 1734 instruct loadV4(vecS dst, memory mem) %{
kvn@3882 1735 predicate(n->as_LoadVector()->memory_size() == 4);
kvn@3882 1736 match(Set dst (LoadVector mem));
kvn@3882 1737 ins_cost(125);
kvn@3882 1738 format %{ "movd $dst,$mem\t! load vector (4 bytes)" %}
kvn@3882 1739 ins_encode %{
kvn@3882 1740 __ movdl($dst$$XMMRegister, $mem$$Address);
kvn@3882 1741 %}
kvn@3882 1742 ins_pipe( pipe_slow );
kvn@3882 1743 %}
kvn@3882 1744
kvn@3882 1745 // Load vectors (8 bytes long)
kvn@3882 1746 instruct loadV8(vecD dst, memory mem) %{
kvn@3882 1747 predicate(n->as_LoadVector()->memory_size() == 8);
kvn@3882 1748 match(Set dst (LoadVector mem));
kvn@3882 1749 ins_cost(125);
kvn@3882 1750 format %{ "movq $dst,$mem\t! load vector (8 bytes)" %}
kvn@3882 1751 ins_encode %{
kvn@3882 1752 __ movq($dst$$XMMRegister, $mem$$Address);
kvn@3882 1753 %}
kvn@3882 1754 ins_pipe( pipe_slow );
kvn@3882 1755 %}
kvn@3882 1756
kvn@3882 1757 // Load vectors (16 bytes long)
kvn@3882 1758 instruct loadV16(vecX dst, memory mem) %{
kvn@3882 1759 predicate(n->as_LoadVector()->memory_size() == 16);
kvn@3882 1760 match(Set dst (LoadVector mem));
kvn@3882 1761 ins_cost(125);
kvn@3882 1762 format %{ "movdqu $dst,$mem\t! load vector (16 bytes)" %}
kvn@3882 1763 ins_encode %{
kvn@3882 1764 __ movdqu($dst$$XMMRegister, $mem$$Address);
kvn@3882 1765 %}
kvn@3882 1766 ins_pipe( pipe_slow );
kvn@3882 1767 %}
kvn@3882 1768
kvn@3882 1769 // Load vectors (32 bytes long)
kvn@3882 1770 instruct loadV32(vecY dst, memory mem) %{
kvn@3882 1771 predicate(n->as_LoadVector()->memory_size() == 32);
kvn@3882 1772 match(Set dst (LoadVector mem));
kvn@3882 1773 ins_cost(125);
kvn@3882 1774 format %{ "vmovdqu $dst,$mem\t! load vector (32 bytes)" %}
kvn@3882 1775 ins_encode %{
kvn@3882 1776 __ vmovdqu($dst$$XMMRegister, $mem$$Address);
kvn@3882 1777 %}
kvn@3882 1778 ins_pipe( pipe_slow );
kvn@3882 1779 %}
kvn@3882 1780
kvn@3882 1781 // Store vectors
kvn@3882 1782 instruct storeV4(memory mem, vecS src) %{
kvn@3882 1783 predicate(n->as_StoreVector()->memory_size() == 4);
kvn@3882 1784 match(Set mem (StoreVector mem src));
kvn@3882 1785 ins_cost(145);
kvn@3882 1786 format %{ "movd $mem,$src\t! store vector (4 bytes)" %}
kvn@3882 1787 ins_encode %{
kvn@3882 1788 __ movdl($mem$$Address, $src$$XMMRegister);
kvn@3882 1789 %}
kvn@3882 1790 ins_pipe( pipe_slow );
kvn@3882 1791 %}
kvn@3882 1792
kvn@3882 1793 instruct storeV8(memory mem, vecD src) %{
kvn@3882 1794 predicate(n->as_StoreVector()->memory_size() == 8);
kvn@3882 1795 match(Set mem (StoreVector mem src));
kvn@3882 1796 ins_cost(145);
kvn@3882 1797 format %{ "movq $mem,$src\t! store vector (8 bytes)" %}
kvn@3882 1798 ins_encode %{
kvn@3882 1799 __ movq($mem$$Address, $src$$XMMRegister);
kvn@3882 1800 %}
kvn@3882 1801 ins_pipe( pipe_slow );
kvn@3882 1802 %}
kvn@3882 1803
kvn@3882 1804 instruct storeV16(memory mem, vecX src) %{
kvn@3882 1805 predicate(n->as_StoreVector()->memory_size() == 16);
kvn@3882 1806 match(Set mem (StoreVector mem src));
kvn@3882 1807 ins_cost(145);
kvn@3882 1808 format %{ "movdqu $mem,$src\t! store vector (16 bytes)" %}
kvn@3882 1809 ins_encode %{
kvn@3882 1810 __ movdqu($mem$$Address, $src$$XMMRegister);
kvn@3882 1811 %}
kvn@3882 1812 ins_pipe( pipe_slow );
kvn@3882 1813 %}
kvn@3882 1814
kvn@3882 1815 instruct storeV32(memory mem, vecY src) %{
kvn@3882 1816 predicate(n->as_StoreVector()->memory_size() == 32);
kvn@3882 1817 match(Set mem (StoreVector mem src));
kvn@3882 1818 ins_cost(145);
kvn@3882 1819 format %{ "vmovdqu $mem,$src\t! store vector (32 bytes)" %}
kvn@3882 1820 ins_encode %{
kvn@3882 1821 __ vmovdqu($mem$$Address, $src$$XMMRegister);
kvn@3882 1822 %}
kvn@3882 1823 ins_pipe( pipe_slow );
kvn@3882 1824 %}
kvn@3882 1825
kvn@3882 1826 // Replicate byte scalar to be vector
kvn@3882 1827 instruct Repl4B(vecS dst, rRegI src) %{
kvn@3882 1828 predicate(n->as_Vector()->length() == 4);
kvn@3882 1829 match(Set dst (ReplicateB src));
kvn@3882 1830 format %{ "movd $dst,$src\n\t"
kvn@3882 1831 "punpcklbw $dst,$dst\n\t"
kvn@3882 1832 "pshuflw $dst,$dst,0x00\t! replicate4B" %}
kvn@3882 1833 ins_encode %{
kvn@3882 1834 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 1835 __ punpcklbw($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 1836 __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3882 1837 %}
kvn@3882 1838 ins_pipe( pipe_slow );
kvn@3882 1839 %}
kvn@3882 1840
kvn@3882 1841 instruct Repl8B(vecD dst, rRegI src) %{
kvn@3882 1842 predicate(n->as_Vector()->length() == 8);
kvn@3882 1843 match(Set dst (ReplicateB src));
kvn@3882 1844 format %{ "movd $dst,$src\n\t"
kvn@3882 1845 "punpcklbw $dst,$dst\n\t"
kvn@3882 1846 "pshuflw $dst,$dst,0x00\t! replicate8B" %}
kvn@3882 1847 ins_encode %{
kvn@3882 1848 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 1849 __ punpcklbw($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 1850 __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3882 1851 %}
kvn@3882 1852 ins_pipe( pipe_slow );
kvn@3882 1853 %}
kvn@3882 1854
kvn@3882 1855 instruct Repl16B(vecX dst, rRegI src) %{
kvn@3882 1856 predicate(n->as_Vector()->length() == 16);
kvn@3882 1857 match(Set dst (ReplicateB src));
kvn@3882 1858 format %{ "movd $dst,$src\n\t"
kvn@3882 1859 "punpcklbw $dst,$dst\n\t"
kvn@3882 1860 "pshuflw $dst,$dst,0x00\n\t"
kvn@3929 1861 "punpcklqdq $dst,$dst\t! replicate16B" %}
kvn@3882 1862 ins_encode %{
kvn@3882 1863 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 1864 __ punpcklbw($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 1865 __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3929 1866 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 1867 %}
kvn@3882 1868 ins_pipe( pipe_slow );
kvn@3882 1869 %}
kvn@3882 1870
kvn@3882 1871 instruct Repl32B(vecY dst, rRegI src) %{
kvn@3882 1872 predicate(n->as_Vector()->length() == 32);
kvn@3882 1873 match(Set dst (ReplicateB src));
kvn@3882 1874 format %{ "movd $dst,$src\n\t"
kvn@3882 1875 "punpcklbw $dst,$dst\n\t"
kvn@3882 1876 "pshuflw $dst,$dst,0x00\n\t"
kvn@3929 1877 "punpcklqdq $dst,$dst\n\t"
kvn@3929 1878 "vinserti128h $dst,$dst,$dst\t! replicate32B" %}
kvn@3882 1879 ins_encode %{
kvn@3882 1880 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 1881 __ punpcklbw($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 1882 __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3929 1883 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3929 1884 __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 1885 %}
kvn@3882 1886 ins_pipe( pipe_slow );
kvn@3882 1887 %}
kvn@3882 1888
kvn@3882 1889 // Replicate byte scalar immediate to be vector by loading from const table.
kvn@3882 1890 instruct Repl4B_imm(vecS dst, immI con) %{
kvn@3882 1891 predicate(n->as_Vector()->length() == 4);
kvn@3882 1892 match(Set dst (ReplicateB con));
kvn@3929 1893 format %{ "movdl $dst,[$constantaddress]\t! replicate4B($con)" %}
kvn@3882 1894 ins_encode %{
kvn@3929 1895 __ movdl($dst$$XMMRegister, $constantaddress(replicate4_imm($con$$constant, 1)));
kvn@3882 1896 %}
kvn@3882 1897 ins_pipe( pipe_slow );
kvn@3882 1898 %}
kvn@3882 1899
kvn@3882 1900 instruct Repl8B_imm(vecD dst, immI con) %{
kvn@3882 1901 predicate(n->as_Vector()->length() == 8);
kvn@3882 1902 match(Set dst (ReplicateB con));
kvn@3929 1903 format %{ "movq $dst,[$constantaddress]\t! replicate8B($con)" %}
kvn@3882 1904 ins_encode %{
kvn@3929 1905 __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 1)));
kvn@3882 1906 %}
kvn@3882 1907 ins_pipe( pipe_slow );
kvn@3882 1908 %}
kvn@3882 1909
kvn@3882 1910 instruct Repl16B_imm(vecX dst, immI con) %{
kvn@3882 1911 predicate(n->as_Vector()->length() == 16);
kvn@3882 1912 match(Set dst (ReplicateB con));
kvn@3929 1913 format %{ "movq $dst,[$constantaddress]\n\t"
kvn@3929 1914 "punpcklqdq $dst,$dst\t! replicate16B($con)" %}
kvn@3882 1915 ins_encode %{
kvn@3929 1916 __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 1)));
kvn@3929 1917 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 1918 %}
kvn@3882 1919 ins_pipe( pipe_slow );
kvn@3882 1920 %}
kvn@3882 1921
kvn@3882 1922 instruct Repl32B_imm(vecY dst, immI con) %{
kvn@3882 1923 predicate(n->as_Vector()->length() == 32);
kvn@3882 1924 match(Set dst (ReplicateB con));
kvn@3929 1925 format %{ "movq $dst,[$constantaddress]\n\t"
kvn@3929 1926 "punpcklqdq $dst,$dst\n\t"
kvn@3929 1927 "vinserti128h $dst,$dst,$dst\t! lreplicate32B($con)" %}
kvn@3882 1928 ins_encode %{
kvn@3929 1929 __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 1)));
kvn@3929 1930 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3929 1931 __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 1932 %}
kvn@3882 1933 ins_pipe( pipe_slow );
kvn@3882 1934 %}
kvn@3882 1935
kvn@3882 1936 // Replicate byte scalar zero to be vector
kvn@3882 1937 instruct Repl4B_zero(vecS dst, immI0 zero) %{
kvn@3882 1938 predicate(n->as_Vector()->length() == 4);
kvn@3882 1939 match(Set dst (ReplicateB zero));
kvn@3882 1940 format %{ "pxor $dst,$dst\t! replicate4B zero" %}
kvn@3882 1941 ins_encode %{
kvn@3882 1942 __ pxor($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 1943 %}
kvn@3882 1944 ins_pipe( fpu_reg_reg );
kvn@3882 1945 %}
kvn@3882 1946
kvn@3882 1947 instruct Repl8B_zero(vecD dst, immI0 zero) %{
kvn@3882 1948 predicate(n->as_Vector()->length() == 8);
kvn@3882 1949 match(Set dst (ReplicateB zero));
kvn@3882 1950 format %{ "pxor $dst,$dst\t! replicate8B zero" %}
kvn@3882 1951 ins_encode %{
kvn@3882 1952 __ pxor($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 1953 %}
kvn@3882 1954 ins_pipe( fpu_reg_reg );
kvn@3882 1955 %}
kvn@3882 1956
kvn@3882 1957 instruct Repl16B_zero(vecX dst, immI0 zero) %{
kvn@3882 1958 predicate(n->as_Vector()->length() == 16);
kvn@3882 1959 match(Set dst (ReplicateB zero));
kvn@3882 1960 format %{ "pxor $dst,$dst\t! replicate16B zero" %}
kvn@3882 1961 ins_encode %{
kvn@3882 1962 __ pxor($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 1963 %}
kvn@3882 1964 ins_pipe( fpu_reg_reg );
kvn@3882 1965 %}
kvn@3882 1966
kvn@3882 1967 instruct Repl32B_zero(vecY dst, immI0 zero) %{
kvn@3882 1968 predicate(n->as_Vector()->length() == 32);
kvn@3882 1969 match(Set dst (ReplicateB zero));
kvn@3929 1970 format %{ "vpxor $dst,$dst,$dst\t! replicate32B zero" %}
kvn@3882 1971 ins_encode %{
kvn@3882 1972 // Use vxorpd since AVX does not have vpxor for 256-bit (AVX2 will have it).
kvn@3882 1973 bool vector256 = true;
kvn@3929 1974 __ vpxor($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister, vector256);
kvn@3882 1975 %}
kvn@3882 1976 ins_pipe( fpu_reg_reg );
kvn@3882 1977 %}
kvn@3882 1978
kvn@3882 1979 // Replicate char/short (2 byte) scalar to be vector
kvn@3882 1980 instruct Repl2S(vecS dst, rRegI src) %{
kvn@3882 1981 predicate(n->as_Vector()->length() == 2);
kvn@3882 1982 match(Set dst (ReplicateS src));
kvn@3882 1983 format %{ "movd $dst,$src\n\t"
kvn@3882 1984 "pshuflw $dst,$dst,0x00\t! replicate2S" %}
kvn@3882 1985 ins_encode %{
kvn@3882 1986 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 1987 __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3882 1988 %}
kvn@3882 1989 ins_pipe( fpu_reg_reg );
kvn@3882 1990 %}
kvn@3882 1991
kvn@3882 1992 instruct Repl4S(vecD dst, rRegI src) %{
kvn@3882 1993 predicate(n->as_Vector()->length() == 4);
kvn@3882 1994 match(Set dst (ReplicateS src));
kvn@3882 1995 format %{ "movd $dst,$src\n\t"
kvn@3882 1996 "pshuflw $dst,$dst,0x00\t! replicate4S" %}
kvn@3882 1997 ins_encode %{
kvn@3882 1998 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 1999 __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3882 2000 %}
kvn@3882 2001 ins_pipe( fpu_reg_reg );
kvn@3882 2002 %}
kvn@3882 2003
kvn@3882 2004 instruct Repl8S(vecX dst, rRegI src) %{
kvn@3882 2005 predicate(n->as_Vector()->length() == 8);
kvn@3882 2006 match(Set dst (ReplicateS src));
kvn@3882 2007 format %{ "movd $dst,$src\n\t"
kvn@3882 2008 "pshuflw $dst,$dst,0x00\n\t"
kvn@3929 2009 "punpcklqdq $dst,$dst\t! replicate8S" %}
kvn@3882 2010 ins_encode %{
kvn@3882 2011 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 2012 __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3929 2013 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2014 %}
kvn@3882 2015 ins_pipe( pipe_slow );
kvn@3882 2016 %}
kvn@3882 2017
kvn@3882 2018 instruct Repl16S(vecY dst, rRegI src) %{
kvn@3882 2019 predicate(n->as_Vector()->length() == 16);
kvn@3882 2020 match(Set dst (ReplicateS src));
kvn@3882 2021 format %{ "movd $dst,$src\n\t"
kvn@3882 2022 "pshuflw $dst,$dst,0x00\n\t"
kvn@3929 2023 "punpcklqdq $dst,$dst\n\t"
kvn@3929 2024 "vinserti128h $dst,$dst,$dst\t! replicate16S" %}
kvn@3882 2025 ins_encode %{
kvn@3882 2026 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 2027 __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3929 2028 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3929 2029 __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2030 %}
kvn@3882 2031 ins_pipe( pipe_slow );
kvn@3882 2032 %}
kvn@3882 2033
kvn@3882 2034 // Replicate char/short (2 byte) scalar immediate to be vector by loading from const table.
kvn@3882 2035 instruct Repl2S_imm(vecS dst, immI con) %{
kvn@3882 2036 predicate(n->as_Vector()->length() == 2);
kvn@3882 2037 match(Set dst (ReplicateS con));
kvn@3929 2038 format %{ "movdl $dst,[$constantaddress]\t! replicate2S($con)" %}
kvn@3882 2039 ins_encode %{
kvn@3929 2040 __ movdl($dst$$XMMRegister, $constantaddress(replicate4_imm($con$$constant, 2)));
kvn@3882 2041 %}
kvn@3882 2042 ins_pipe( fpu_reg_reg );
kvn@3882 2043 %}
kvn@3882 2044
kvn@3882 2045 instruct Repl4S_imm(vecD dst, immI con) %{
kvn@3882 2046 predicate(n->as_Vector()->length() == 4);
kvn@3882 2047 match(Set dst (ReplicateS con));
kvn@3929 2048 format %{ "movq $dst,[$constantaddress]\t! replicate4S($con)" %}
kvn@3882 2049 ins_encode %{
kvn@3929 2050 __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 2)));
kvn@3882 2051 %}
kvn@3882 2052 ins_pipe( fpu_reg_reg );
kvn@3882 2053 %}
kvn@3882 2054
kvn@3882 2055 instruct Repl8S_imm(vecX dst, immI con) %{
kvn@3882 2056 predicate(n->as_Vector()->length() == 8);
kvn@3882 2057 match(Set dst (ReplicateS con));
kvn@3929 2058 format %{ "movq $dst,[$constantaddress]\n\t"
kvn@3929 2059 "punpcklqdq $dst,$dst\t! replicate8S($con)" %}
kvn@3882 2060 ins_encode %{
kvn@3929 2061 __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 2)));
kvn@3929 2062 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2063 %}
kvn@3882 2064 ins_pipe( pipe_slow );
kvn@3882 2065 %}
kvn@3882 2066
kvn@3882 2067 instruct Repl16S_imm(vecY dst, immI con) %{
kvn@3882 2068 predicate(n->as_Vector()->length() == 16);
kvn@3882 2069 match(Set dst (ReplicateS con));
kvn@3929 2070 format %{ "movq $dst,[$constantaddress]\n\t"
kvn@3929 2071 "punpcklqdq $dst,$dst\n\t"
kvn@3929 2072 "vinserti128h $dst,$dst,$dst\t! replicate16S($con)" %}
kvn@3882 2073 ins_encode %{
kvn@3929 2074 __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 2)));
kvn@3929 2075 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3929 2076 __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2077 %}
kvn@3882 2078 ins_pipe( pipe_slow );
kvn@3882 2079 %}
kvn@3882 2080
kvn@3882 2081 // Replicate char/short (2 byte) scalar zero to be vector
kvn@3882 2082 instruct Repl2S_zero(vecS dst, immI0 zero) %{
kvn@3882 2083 predicate(n->as_Vector()->length() == 2);
kvn@3882 2084 match(Set dst (ReplicateS zero));
kvn@3882 2085 format %{ "pxor $dst,$dst\t! replicate2S zero" %}
kvn@3882 2086 ins_encode %{
kvn@3882 2087 __ pxor($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2088 %}
kvn@3882 2089 ins_pipe( fpu_reg_reg );
kvn@3882 2090 %}
kvn@3882 2091
kvn@3882 2092 instruct Repl4S_zero(vecD dst, immI0 zero) %{
kvn@3882 2093 predicate(n->as_Vector()->length() == 4);
kvn@3882 2094 match(Set dst (ReplicateS zero));
kvn@3882 2095 format %{ "pxor $dst,$dst\t! replicate4S zero" %}
kvn@3882 2096 ins_encode %{
kvn@3882 2097 __ pxor($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2098 %}
kvn@3882 2099 ins_pipe( fpu_reg_reg );
kvn@3882 2100 %}
kvn@3882 2101
kvn@3882 2102 instruct Repl8S_zero(vecX dst, immI0 zero) %{
kvn@3882 2103 predicate(n->as_Vector()->length() == 8);
kvn@3882 2104 match(Set dst (ReplicateS zero));
kvn@3882 2105 format %{ "pxor $dst,$dst\t! replicate8S zero" %}
kvn@3882 2106 ins_encode %{
kvn@3882 2107 __ pxor($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2108 %}
kvn@3882 2109 ins_pipe( fpu_reg_reg );
kvn@3882 2110 %}
kvn@3882 2111
kvn@3882 2112 instruct Repl16S_zero(vecY dst, immI0 zero) %{
kvn@3882 2113 predicate(n->as_Vector()->length() == 16);
kvn@3882 2114 match(Set dst (ReplicateS zero));
kvn@3929 2115 format %{ "vpxor $dst,$dst,$dst\t! replicate16S zero" %}
kvn@3882 2116 ins_encode %{
kvn@3882 2117 // Use vxorpd since AVX does not have vpxor for 256-bit (AVX2 will have it).
kvn@3882 2118 bool vector256 = true;
kvn@3929 2119 __ vpxor($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister, vector256);
kvn@3882 2120 %}
kvn@3882 2121 ins_pipe( fpu_reg_reg );
kvn@3882 2122 %}
kvn@3882 2123
kvn@3882 2124 // Replicate integer (4 byte) scalar to be vector
kvn@3882 2125 instruct Repl2I(vecD dst, rRegI src) %{
kvn@3882 2126 predicate(n->as_Vector()->length() == 2);
kvn@3882 2127 match(Set dst (ReplicateI src));
kvn@3882 2128 format %{ "movd $dst,$src\n\t"
kvn@3882 2129 "pshufd $dst,$dst,0x00\t! replicate2I" %}
kvn@3882 2130 ins_encode %{
kvn@3882 2131 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 2132 __ pshufd($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3882 2133 %}
kvn@3882 2134 ins_pipe( fpu_reg_reg );
kvn@3882 2135 %}
kvn@3882 2136
kvn@3882 2137 instruct Repl4I(vecX dst, rRegI src) %{
kvn@3882 2138 predicate(n->as_Vector()->length() == 4);
kvn@3882 2139 match(Set dst (ReplicateI src));
kvn@3882 2140 format %{ "movd $dst,$src\n\t"
kvn@3882 2141 "pshufd $dst,$dst,0x00\t! replicate4I" %}
kvn@3882 2142 ins_encode %{
kvn@3882 2143 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 2144 __ pshufd($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3882 2145 %}
kvn@3882 2146 ins_pipe( pipe_slow );
kvn@3882 2147 %}
kvn@3882 2148
kvn@3882 2149 instruct Repl8I(vecY dst, rRegI src) %{
kvn@3882 2150 predicate(n->as_Vector()->length() == 8);
kvn@3882 2151 match(Set dst (ReplicateI src));
kvn@3882 2152 format %{ "movd $dst,$src\n\t"
kvn@3882 2153 "pshufd $dst,$dst,0x00\n\t"
kvn@3929 2154 "vinserti128h $dst,$dst,$dst\t! replicate8I" %}
kvn@3882 2155 ins_encode %{
kvn@3882 2156 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 2157 __ pshufd($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3929 2158 __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2159 %}
kvn@3882 2160 ins_pipe( pipe_slow );
kvn@3882 2161 %}
kvn@3882 2162
kvn@3882 2163 // Replicate integer (4 byte) scalar immediate to be vector by loading from const table.
kvn@3882 2164 instruct Repl2I_imm(vecD dst, immI con) %{
kvn@3882 2165 predicate(n->as_Vector()->length() == 2);
kvn@3882 2166 match(Set dst (ReplicateI con));
kvn@3929 2167 format %{ "movq $dst,[$constantaddress]\t! replicate2I($con)" %}
kvn@3882 2168 ins_encode %{
kvn@3929 2169 __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 4)));
kvn@3882 2170 %}
kvn@3882 2171 ins_pipe( fpu_reg_reg );
kvn@3882 2172 %}
kvn@3882 2173
kvn@3882 2174 instruct Repl4I_imm(vecX dst, immI con) %{
kvn@3882 2175 predicate(n->as_Vector()->length() == 4);
kvn@3882 2176 match(Set dst (ReplicateI con));
kvn@3929 2177 format %{ "movq $dst,[$constantaddress]\t! replicate4I($con)\n\t"
kvn@3929 2178 "punpcklqdq $dst,$dst" %}
kvn@3882 2179 ins_encode %{
kvn@3929 2180 __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 4)));
kvn@3929 2181 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2182 %}
kvn@3882 2183 ins_pipe( pipe_slow );
kvn@3882 2184 %}
kvn@3882 2185
kvn@3882 2186 instruct Repl8I_imm(vecY dst, immI con) %{
kvn@3882 2187 predicate(n->as_Vector()->length() == 8);
kvn@3882 2188 match(Set dst (ReplicateI con));
kvn@3929 2189 format %{ "movq $dst,[$constantaddress]\t! replicate8I($con)\n\t"
kvn@3929 2190 "punpcklqdq $dst,$dst\n\t"
kvn@3929 2191 "vinserti128h $dst,$dst,$dst" %}
kvn@3882 2192 ins_encode %{
kvn@3929 2193 __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 4)));
kvn@3929 2194 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3929 2195 __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2196 %}
kvn@3882 2197 ins_pipe( pipe_slow );
kvn@3882 2198 %}
kvn@3882 2199
kvn@3882 2200 // Integer could be loaded into xmm register directly from memory.
kvn@3882 2201 instruct Repl2I_mem(vecD dst, memory mem) %{
kvn@3882 2202 predicate(n->as_Vector()->length() == 2);
kvn@3929 2203 match(Set dst (ReplicateI (LoadI mem)));
kvn@3882 2204 format %{ "movd $dst,$mem\n\t"
kvn@3882 2205 "pshufd $dst,$dst,0x00\t! replicate2I" %}
kvn@3882 2206 ins_encode %{
kvn@3882 2207 __ movdl($dst$$XMMRegister, $mem$$Address);
kvn@3882 2208 __ pshufd($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3882 2209 %}
kvn@3882 2210 ins_pipe( fpu_reg_reg );
kvn@3882 2211 %}
kvn@3882 2212
kvn@3882 2213 instruct Repl4I_mem(vecX dst, memory mem) %{
kvn@3882 2214 predicate(n->as_Vector()->length() == 4);
kvn@3929 2215 match(Set dst (ReplicateI (LoadI mem)));
kvn@3882 2216 format %{ "movd $dst,$mem\n\t"
kvn@3882 2217 "pshufd $dst,$dst,0x00\t! replicate4I" %}
kvn@3882 2218 ins_encode %{
kvn@3882 2219 __ movdl($dst$$XMMRegister, $mem$$Address);
kvn@3882 2220 __ pshufd($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3882 2221 %}
kvn@3882 2222 ins_pipe( pipe_slow );
kvn@3882 2223 %}
kvn@3882 2224
kvn@3882 2225 instruct Repl8I_mem(vecY dst, memory mem) %{
kvn@3882 2226 predicate(n->as_Vector()->length() == 8);
kvn@3929 2227 match(Set dst (ReplicateI (LoadI mem)));
kvn@3882 2228 format %{ "movd $dst,$mem\n\t"
kvn@3882 2229 "pshufd $dst,$dst,0x00\n\t"
kvn@3929 2230 "vinserti128h $dst,$dst,$dst\t! replicate8I" %}
kvn@3882 2231 ins_encode %{
kvn@3882 2232 __ movdl($dst$$XMMRegister, $mem$$Address);
kvn@3882 2233 __ pshufd($dst$$XMMRegister, $dst$$XMMRegister, 0x00);
kvn@3929 2234 __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2235 %}
kvn@3882 2236 ins_pipe( pipe_slow );
kvn@3882 2237 %}
kvn@3882 2238
kvn@3882 2239 // Replicate integer (4 byte) scalar zero to be vector
kvn@3882 2240 instruct Repl2I_zero(vecD dst, immI0 zero) %{
kvn@3882 2241 predicate(n->as_Vector()->length() == 2);
kvn@3882 2242 match(Set dst (ReplicateI zero));
kvn@3882 2243 format %{ "pxor $dst,$dst\t! replicate2I" %}
kvn@3882 2244 ins_encode %{
kvn@3882 2245 __ pxor($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2246 %}
kvn@3882 2247 ins_pipe( fpu_reg_reg );
kvn@3882 2248 %}
kvn@3882 2249
kvn@3882 2250 instruct Repl4I_zero(vecX dst, immI0 zero) %{
kvn@3882 2251 predicate(n->as_Vector()->length() == 4);
kvn@3882 2252 match(Set dst (ReplicateI zero));
kvn@3882 2253 format %{ "pxor $dst,$dst\t! replicate4I zero)" %}
kvn@3882 2254 ins_encode %{
kvn@3882 2255 __ pxor($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2256 %}
kvn@3882 2257 ins_pipe( fpu_reg_reg );
kvn@3882 2258 %}
kvn@3882 2259
kvn@3882 2260 instruct Repl8I_zero(vecY dst, immI0 zero) %{
kvn@3882 2261 predicate(n->as_Vector()->length() == 8);
kvn@3882 2262 match(Set dst (ReplicateI zero));
kvn@3929 2263 format %{ "vpxor $dst,$dst,$dst\t! replicate8I zero" %}
kvn@3882 2264 ins_encode %{
kvn@3882 2265 // Use vxorpd since AVX does not have vpxor for 256-bit (AVX2 will have it).
kvn@3882 2266 bool vector256 = true;
kvn@3929 2267 __ vpxor($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister, vector256);
kvn@3882 2268 %}
kvn@3882 2269 ins_pipe( fpu_reg_reg );
kvn@3882 2270 %}
kvn@3882 2271
kvn@3882 2272 // Replicate long (8 byte) scalar to be vector
kvn@3882 2273 #ifdef _LP64
kvn@3882 2274 instruct Repl2L(vecX dst, rRegL src) %{
kvn@3882 2275 predicate(n->as_Vector()->length() == 2);
kvn@3882 2276 match(Set dst (ReplicateL src));
kvn@3882 2277 format %{ "movdq $dst,$src\n\t"
kvn@3929 2278 "punpcklqdq $dst,$dst\t! replicate2L" %}
kvn@3882 2279 ins_encode %{
kvn@3882 2280 __ movdq($dst$$XMMRegister, $src$$Register);
kvn@3929 2281 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2282 %}
kvn@3882 2283 ins_pipe( pipe_slow );
kvn@3882 2284 %}
kvn@3882 2285
kvn@3882 2286 instruct Repl4L(vecY dst, rRegL src) %{
kvn@3882 2287 predicate(n->as_Vector()->length() == 4);
kvn@3882 2288 match(Set dst (ReplicateL src));
kvn@3882 2289 format %{ "movdq $dst,$src\n\t"
kvn@3929 2290 "punpcklqdq $dst,$dst\n\t"
kvn@3929 2291 "vinserti128h $dst,$dst,$dst\t! replicate4L" %}
kvn@3882 2292 ins_encode %{
kvn@3882 2293 __ movdq($dst$$XMMRegister, $src$$Register);
kvn@3929 2294 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3929 2295 __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2296 %}
kvn@3882 2297 ins_pipe( pipe_slow );
kvn@3882 2298 %}
kvn@3882 2299 #else // _LP64
kvn@3882 2300 instruct Repl2L(vecX dst, eRegL src, regD tmp) %{
kvn@3882 2301 predicate(n->as_Vector()->length() == 2);
kvn@3882 2302 match(Set dst (ReplicateL src));
kvn@3882 2303 effect(TEMP dst, USE src, TEMP tmp);
kvn@3882 2304 format %{ "movdl $dst,$src.lo\n\t"
kvn@3882 2305 "movdl $tmp,$src.hi\n\t"
kvn@3882 2306 "punpckldq $dst,$tmp\n\t"
kvn@3929 2307 "punpcklqdq $dst,$dst\t! replicate2L"%}
kvn@3882 2308 ins_encode %{
kvn@3882 2309 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 2310 __ movdl($tmp$$XMMRegister, HIGH_FROM_LOW($src$$Register));
kvn@3882 2311 __ punpckldq($dst$$XMMRegister, $tmp$$XMMRegister);
kvn@3929 2312 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2313 %}
kvn@3882 2314 ins_pipe( pipe_slow );
kvn@3882 2315 %}
kvn@3882 2316
kvn@3882 2317 instruct Repl4L(vecY dst, eRegL src, regD tmp) %{
kvn@3882 2318 predicate(n->as_Vector()->length() == 4);
kvn@3882 2319 match(Set dst (ReplicateL src));
kvn@3882 2320 effect(TEMP dst, USE src, TEMP tmp);
kvn@3882 2321 format %{ "movdl $dst,$src.lo\n\t"
kvn@3882 2322 "movdl $tmp,$src.hi\n\t"
kvn@3882 2323 "punpckldq $dst,$tmp\n\t"
kvn@3929 2324 "punpcklqdq $dst,$dst\n\t"
kvn@3929 2325 "vinserti128h $dst,$dst,$dst\t! replicate4L" %}
kvn@3882 2326 ins_encode %{
kvn@3882 2327 __ movdl($dst$$XMMRegister, $src$$Register);
kvn@3882 2328 __ movdl($tmp$$XMMRegister, HIGH_FROM_LOW($src$$Register));
kvn@3882 2329 __ punpckldq($dst$$XMMRegister, $tmp$$XMMRegister);
kvn@3929 2330 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3929 2331 __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2332 %}
kvn@3882 2333 ins_pipe( pipe_slow );
kvn@3882 2334 %}
kvn@3882 2335 #endif // _LP64
kvn@3882 2336
kvn@3882 2337 // Replicate long (8 byte) scalar immediate to be vector by loading from const table.
kvn@3882 2338 instruct Repl2L_imm(vecX dst, immL con) %{
kvn@3882 2339 predicate(n->as_Vector()->length() == 2);
kvn@3882 2340 match(Set dst (ReplicateL con));
kvn@3929 2341 format %{ "movq $dst,[$constantaddress]\n\t"
kvn@3929 2342 "punpcklqdq $dst,$dst\t! replicate2L($con)" %}
kvn@3882 2343 ins_encode %{
kvn@3929 2344 __ movq($dst$$XMMRegister, $constantaddress($con));
kvn@3929 2345 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2346 %}
kvn@3882 2347 ins_pipe( pipe_slow );
kvn@3882 2348 %}
kvn@3882 2349
kvn@3882 2350 instruct Repl4L_imm(vecY dst, immL con) %{
kvn@3882 2351 predicate(n->as_Vector()->length() == 4);
kvn@3882 2352 match(Set dst (ReplicateL con));
kvn@3929 2353 format %{ "movq $dst,[$constantaddress]\n\t"
kvn@3929 2354 "punpcklqdq $dst,$dst\n\t"
kvn@3929 2355 "vinserti128h $dst,$dst,$dst\t! replicate4L($con)" %}
kvn@3882 2356 ins_encode %{
kvn@3929 2357 __ movq($dst$$XMMRegister, $constantaddress($con));
kvn@3929 2358 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3929 2359 __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2360 %}
kvn@3882 2361 ins_pipe( pipe_slow );
kvn@3882 2362 %}
kvn@3882 2363
kvn@3882 2364 // Long could be loaded into xmm register directly from memory.
kvn@3882 2365 instruct Repl2L_mem(vecX dst, memory mem) %{
kvn@3882 2366 predicate(n->as_Vector()->length() == 2);
kvn@3929 2367 match(Set dst (ReplicateL (LoadL mem)));
kvn@3882 2368 format %{ "movq $dst,$mem\n\t"
kvn@3929 2369 "punpcklqdq $dst,$dst\t! replicate2L" %}
kvn@3882 2370 ins_encode %{
kvn@3882 2371 __ movq($dst$$XMMRegister, $mem$$Address);
kvn@3929 2372 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2373 %}
kvn@3882 2374 ins_pipe( pipe_slow );
kvn@3882 2375 %}
kvn@3882 2376
kvn@3882 2377 instruct Repl4L_mem(vecY dst, memory mem) %{
kvn@3882 2378 predicate(n->as_Vector()->length() == 4);
kvn@3929 2379 match(Set dst (ReplicateL (LoadL mem)));
kvn@3882 2380 format %{ "movq $dst,$mem\n\t"
kvn@3929 2381 "punpcklqdq $dst,$dst\n\t"
kvn@3929 2382 "vinserti128h $dst,$dst,$dst\t! replicate4L" %}
kvn@3882 2383 ins_encode %{
kvn@3882 2384 __ movq($dst$$XMMRegister, $mem$$Address);
kvn@3929 2385 __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3929 2386 __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2387 %}
kvn@3882 2388 ins_pipe( pipe_slow );
kvn@3882 2389 %}
kvn@3882 2390
kvn@3882 2391 // Replicate long (8 byte) scalar zero to be vector
kvn@3882 2392 instruct Repl2L_zero(vecX dst, immL0 zero) %{
kvn@3882 2393 predicate(n->as_Vector()->length() == 2);
kvn@3882 2394 match(Set dst (ReplicateL zero));
kvn@3882 2395 format %{ "pxor $dst,$dst\t! replicate2L zero" %}
kvn@3882 2396 ins_encode %{
kvn@3882 2397 __ pxor($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2398 %}
kvn@3882 2399 ins_pipe( fpu_reg_reg );
kvn@3882 2400 %}
kvn@3882 2401
kvn@3882 2402 instruct Repl4L_zero(vecY dst, immL0 zero) %{
kvn@3882 2403 predicate(n->as_Vector()->length() == 4);
kvn@3882 2404 match(Set dst (ReplicateL zero));
kvn@3929 2405 format %{ "vpxor $dst,$dst,$dst\t! replicate4L zero" %}
kvn@3882 2406 ins_encode %{
kvn@3882 2407 // Use vxorpd since AVX does not have vpxor for 256-bit (AVX2 will have it).
kvn@3882 2408 bool vector256 = true;
kvn@3929 2409 __ vpxor($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister, vector256);
kvn@3882 2410 %}
kvn@3882 2411 ins_pipe( fpu_reg_reg );
kvn@3882 2412 %}
kvn@3882 2413
kvn@3882 2414 // Replicate float (4 byte) scalar to be vector
kvn@3882 2415 instruct Repl2F(vecD dst, regF src) %{
kvn@3882 2416 predicate(n->as_Vector()->length() == 2);
kvn@3882 2417 match(Set dst (ReplicateF src));
kvn@3882 2418 format %{ "pshufd $dst,$dst,0x00\t! replicate2F" %}
kvn@3882 2419 ins_encode %{
kvn@3882 2420 __ pshufd($dst$$XMMRegister, $src$$XMMRegister, 0x00);
kvn@3882 2421 %}
kvn@3882 2422 ins_pipe( fpu_reg_reg );
kvn@3882 2423 %}
kvn@3882 2424
kvn@3882 2425 instruct Repl4F(vecX dst, regF src) %{
kvn@3882 2426 predicate(n->as_Vector()->length() == 4);
kvn@3882 2427 match(Set dst (ReplicateF src));
kvn@3882 2428 format %{ "pshufd $dst,$dst,0x00\t! replicate4F" %}
kvn@3882 2429 ins_encode %{
kvn@3882 2430 __ pshufd($dst$$XMMRegister, $src$$XMMRegister, 0x00);
kvn@3882 2431 %}
kvn@3882 2432 ins_pipe( pipe_slow );
kvn@3882 2433 %}
kvn@3882 2434
kvn@3882 2435 instruct Repl8F(vecY dst, regF src) %{
kvn@3882 2436 predicate(n->as_Vector()->length() == 8);
kvn@3882 2437 match(Set dst (ReplicateF src));
kvn@3882 2438 format %{ "pshufd $dst,$src,0x00\n\t"
kvn@3882 2439 "vinsertf128h $dst,$dst,$dst\t! replicate8F" %}
kvn@3882 2440 ins_encode %{
kvn@3882 2441 __ pshufd($dst$$XMMRegister, $src$$XMMRegister, 0x00);
kvn@3882 2442 __ vinsertf128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2443 %}
kvn@3882 2444 ins_pipe( pipe_slow );
kvn@3882 2445 %}
kvn@3882 2446
kvn@3882 2447 // Replicate float (4 byte) scalar zero to be vector
kvn@3882 2448 instruct Repl2F_zero(vecD dst, immF0 zero) %{
kvn@3882 2449 predicate(n->as_Vector()->length() == 2);
kvn@3882 2450 match(Set dst (ReplicateF zero));
kvn@3882 2451 format %{ "xorps $dst,$dst\t! replicate2F zero" %}
kvn@3882 2452 ins_encode %{
kvn@3882 2453 __ xorps($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2454 %}
kvn@3882 2455 ins_pipe( fpu_reg_reg );
kvn@3882 2456 %}
kvn@3882 2457
kvn@3882 2458 instruct Repl4F_zero(vecX dst, immF0 zero) %{
kvn@3882 2459 predicate(n->as_Vector()->length() == 4);
kvn@3882 2460 match(Set dst (ReplicateF zero));
kvn@3882 2461 format %{ "xorps $dst,$dst\t! replicate4F zero" %}
kvn@3882 2462 ins_encode %{
kvn@3882 2463 __ xorps($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2464 %}
kvn@3882 2465 ins_pipe( fpu_reg_reg );
kvn@3882 2466 %}
kvn@3882 2467
kvn@3882 2468 instruct Repl8F_zero(vecY dst, immF0 zero) %{
kvn@3882 2469 predicate(n->as_Vector()->length() == 8);
kvn@3882 2470 match(Set dst (ReplicateF zero));
kvn@3882 2471 format %{ "vxorps $dst,$dst,$dst\t! replicate8F zero" %}
kvn@3882 2472 ins_encode %{
kvn@3882 2473 bool vector256 = true;
kvn@3882 2474 __ vxorps($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister, vector256);
kvn@3882 2475 %}
kvn@3882 2476 ins_pipe( fpu_reg_reg );
kvn@3882 2477 %}
kvn@3882 2478
kvn@3882 2479 // Replicate double (8 bytes) scalar to be vector
kvn@3882 2480 instruct Repl2D(vecX dst, regD src) %{
kvn@3882 2481 predicate(n->as_Vector()->length() == 2);
kvn@3882 2482 match(Set dst (ReplicateD src));
kvn@3882 2483 format %{ "pshufd $dst,$src,0x44\t! replicate2D" %}
kvn@3882 2484 ins_encode %{
kvn@3882 2485 __ pshufd($dst$$XMMRegister, $src$$XMMRegister, 0x44);
kvn@3882 2486 %}
kvn@3882 2487 ins_pipe( pipe_slow );
kvn@3882 2488 %}
kvn@3882 2489
kvn@3882 2490 instruct Repl4D(vecY dst, regD src) %{
kvn@3882 2491 predicate(n->as_Vector()->length() == 4);
kvn@3882 2492 match(Set dst (ReplicateD src));
kvn@3882 2493 format %{ "pshufd $dst,$src,0x44\n\t"
kvn@3882 2494 "vinsertf128h $dst,$dst,$dst\t! replicate4D" %}
kvn@3882 2495 ins_encode %{
kvn@3882 2496 __ pshufd($dst$$XMMRegister, $src$$XMMRegister, 0x44);
kvn@3882 2497 __ vinsertf128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2498 %}
kvn@3882 2499 ins_pipe( pipe_slow );
kvn@3882 2500 %}
kvn@3882 2501
kvn@3882 2502 // Replicate double (8 byte) scalar zero to be vector
kvn@3882 2503 instruct Repl2D_zero(vecX dst, immD0 zero) %{
kvn@3882 2504 predicate(n->as_Vector()->length() == 2);
kvn@3882 2505 match(Set dst (ReplicateD zero));
kvn@3882 2506 format %{ "xorpd $dst,$dst\t! replicate2D zero" %}
kvn@3882 2507 ins_encode %{
kvn@3882 2508 __ xorpd($dst$$XMMRegister, $dst$$XMMRegister);
kvn@3882 2509 %}
kvn@3882 2510 ins_pipe( fpu_reg_reg );
kvn@3882 2511 %}
kvn@3882 2512
kvn@3882 2513 instruct Repl4D_zero(vecY dst, immD0 zero) %{
kvn@3882 2514 predicate(n->as_Vector()->length() == 4);
kvn@3882 2515 match(Set dst (ReplicateD zero));
kvn@3882 2516 format %{ "vxorpd $dst,$dst,$dst,vect256\t! replicate4D zero" %}
kvn@3882 2517 ins_encode %{
kvn@3882 2518 bool vector256 = true;
kvn@3882 2519 __ vxorpd($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister, vector256);
kvn@3882 2520 %}
kvn@3882 2521 ins_pipe( fpu_reg_reg );
kvn@3882 2522 %}
kvn@3882 2523
kvn@4001 2524 // ====================VECTOR ARITHMETIC=======================================
kvn@4001 2525
kvn@4001 2526 // --------------------------------- ADD --------------------------------------
kvn@4001 2527
kvn@4001 2528 // Bytes vector add
kvn@4001 2529 instruct vadd4B(vecS dst, vecS src) %{
kvn@4001 2530 predicate(n->as_Vector()->length() == 4);
kvn@4001 2531 match(Set dst (AddVB dst src));
kvn@4001 2532 format %{ "paddb $dst,$src\t! add packed4B" %}
kvn@4001 2533 ins_encode %{
kvn@4001 2534 __ paddb($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2535 %}
kvn@4001 2536 ins_pipe( pipe_slow );
kvn@4001 2537 %}
kvn@4001 2538
kvn@4001 2539 instruct vadd4B_reg(vecS dst, vecS src1, vecS src2) %{
kvn@4001 2540 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 2541 match(Set dst (AddVB src1 src2));
kvn@4001 2542 format %{ "vpaddb $dst,$src1,$src2\t! add packed4B" %}
kvn@4001 2543 ins_encode %{
kvn@4001 2544 bool vector256 = false;
kvn@4001 2545 __ vpaddb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2546 %}
kvn@4001 2547 ins_pipe( pipe_slow );
kvn@4001 2548 %}
kvn@4001 2549
kvn@4001 2550 instruct vadd8B(vecD dst, vecD src) %{
kvn@4001 2551 predicate(n->as_Vector()->length() == 8);
kvn@4001 2552 match(Set dst (AddVB dst src));
kvn@4001 2553 format %{ "paddb $dst,$src\t! add packed8B" %}
kvn@4001 2554 ins_encode %{
kvn@4001 2555 __ paddb($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2556 %}
kvn@4001 2557 ins_pipe( pipe_slow );
kvn@4001 2558 %}
kvn@4001 2559
kvn@4001 2560 instruct vadd8B_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 2561 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 2562 match(Set dst (AddVB src1 src2));
kvn@4001 2563 format %{ "vpaddb $dst,$src1,$src2\t! add packed8B" %}
kvn@4001 2564 ins_encode %{
kvn@4001 2565 bool vector256 = false;
kvn@4001 2566 __ vpaddb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2567 %}
kvn@4001 2568 ins_pipe( pipe_slow );
kvn@4001 2569 %}
kvn@4001 2570
kvn@4001 2571 instruct vadd16B(vecX dst, vecX src) %{
kvn@4001 2572 predicate(n->as_Vector()->length() == 16);
kvn@4001 2573 match(Set dst (AddVB dst src));
kvn@4001 2574 format %{ "paddb $dst,$src\t! add packed16B" %}
kvn@4001 2575 ins_encode %{
kvn@4001 2576 __ paddb($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2577 %}
kvn@4001 2578 ins_pipe( pipe_slow );
kvn@4001 2579 %}
kvn@4001 2580
kvn@4001 2581 instruct vadd16B_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 2582 predicate(UseAVX > 0 && n->as_Vector()->length() == 16);
kvn@4001 2583 match(Set dst (AddVB src1 src2));
kvn@4001 2584 format %{ "vpaddb $dst,$src1,$src2\t! add packed16B" %}
kvn@4001 2585 ins_encode %{
kvn@4001 2586 bool vector256 = false;
kvn@4001 2587 __ vpaddb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2588 %}
kvn@4001 2589 ins_pipe( pipe_slow );
kvn@4001 2590 %}
kvn@4001 2591
kvn@4001 2592 instruct vadd16B_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 2593 predicate(UseAVX > 0 && n->as_Vector()->length() == 16);
kvn@4001 2594 match(Set dst (AddVB src (LoadVector mem)));
kvn@4001 2595 format %{ "vpaddb $dst,$src,$mem\t! add packed16B" %}
kvn@4001 2596 ins_encode %{
kvn@4001 2597 bool vector256 = false;
kvn@4001 2598 __ vpaddb($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 2599 %}
kvn@4001 2600 ins_pipe( pipe_slow );
kvn@4001 2601 %}
kvn@4001 2602
kvn@4001 2603 instruct vadd32B_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 2604 predicate(UseAVX > 1 && n->as_Vector()->length() == 32);
kvn@4001 2605 match(Set dst (AddVB src1 src2));
kvn@4001 2606 format %{ "vpaddb $dst,$src1,$src2\t! add packed32B" %}
kvn@4001 2607 ins_encode %{
kvn@4001 2608 bool vector256 = true;
kvn@4001 2609 __ vpaddb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2610 %}
kvn@4001 2611 ins_pipe( pipe_slow );
kvn@4001 2612 %}
kvn@4001 2613
kvn@4001 2614 instruct vadd32B_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 2615 predicate(UseAVX > 1 && n->as_Vector()->length() == 32);
kvn@4001 2616 match(Set dst (AddVB src (LoadVector mem)));
kvn@4001 2617 format %{ "vpaddb $dst,$src,$mem\t! add packed32B" %}
kvn@4001 2618 ins_encode %{
kvn@4001 2619 bool vector256 = true;
kvn@4001 2620 __ vpaddb($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 2621 %}
kvn@4001 2622 ins_pipe( pipe_slow );
kvn@4001 2623 %}
kvn@4001 2624
kvn@4001 2625 // Shorts/Chars vector add
kvn@4001 2626 instruct vadd2S(vecS dst, vecS src) %{
kvn@4001 2627 predicate(n->as_Vector()->length() == 2);
kvn@4001 2628 match(Set dst (AddVS dst src));
kvn@4001 2629 format %{ "paddw $dst,$src\t! add packed2S" %}
kvn@4001 2630 ins_encode %{
kvn@4001 2631 __ paddw($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2632 %}
kvn@4001 2633 ins_pipe( pipe_slow );
kvn@4001 2634 %}
kvn@4001 2635
kvn@4001 2636 instruct vadd2S_reg(vecS dst, vecS src1, vecS src2) %{
kvn@4001 2637 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 2638 match(Set dst (AddVS src1 src2));
kvn@4001 2639 format %{ "vpaddw $dst,$src1,$src2\t! add packed2S" %}
kvn@4001 2640 ins_encode %{
kvn@4001 2641 bool vector256 = false;
kvn@4001 2642 __ vpaddw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2643 %}
kvn@4001 2644 ins_pipe( pipe_slow );
kvn@4001 2645 %}
kvn@4001 2646
kvn@4001 2647 instruct vadd4S(vecD dst, vecD src) %{
kvn@4001 2648 predicate(n->as_Vector()->length() == 4);
kvn@4001 2649 match(Set dst (AddVS dst src));
kvn@4001 2650 format %{ "paddw $dst,$src\t! add packed4S" %}
kvn@4001 2651 ins_encode %{
kvn@4001 2652 __ paddw($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2653 %}
kvn@4001 2654 ins_pipe( pipe_slow );
kvn@4001 2655 %}
kvn@4001 2656
kvn@4001 2657 instruct vadd4S_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 2658 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 2659 match(Set dst (AddVS src1 src2));
kvn@4001 2660 format %{ "vpaddw $dst,$src1,$src2\t! add packed4S" %}
kvn@4001 2661 ins_encode %{
kvn@4001 2662 bool vector256 = false;
kvn@4001 2663 __ vpaddw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2664 %}
kvn@4001 2665 ins_pipe( pipe_slow );
kvn@4001 2666 %}
kvn@4001 2667
kvn@4001 2668 instruct vadd8S(vecX dst, vecX src) %{
kvn@4001 2669 predicate(n->as_Vector()->length() == 8);
kvn@4001 2670 match(Set dst (AddVS dst src));
kvn@4001 2671 format %{ "paddw $dst,$src\t! add packed8S" %}
kvn@4001 2672 ins_encode %{
kvn@4001 2673 __ paddw($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2674 %}
kvn@4001 2675 ins_pipe( pipe_slow );
kvn@4001 2676 %}
kvn@4001 2677
kvn@4001 2678 instruct vadd8S_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 2679 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 2680 match(Set dst (AddVS src1 src2));
kvn@4001 2681 format %{ "vpaddw $dst,$src1,$src2\t! add packed8S" %}
kvn@4001 2682 ins_encode %{
kvn@4001 2683 bool vector256 = false;
kvn@4001 2684 __ vpaddw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2685 %}
kvn@4001 2686 ins_pipe( pipe_slow );
kvn@4001 2687 %}
kvn@4001 2688
kvn@4001 2689 instruct vadd8S_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 2690 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 2691 match(Set dst (AddVS src (LoadVector mem)));
kvn@4001 2692 format %{ "vpaddw $dst,$src,$mem\t! add packed8S" %}
kvn@4001 2693 ins_encode %{
kvn@4001 2694 bool vector256 = false;
kvn@4001 2695 __ vpaddw($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 2696 %}
kvn@4001 2697 ins_pipe( pipe_slow );
kvn@4001 2698 %}
kvn@4001 2699
kvn@4001 2700 instruct vadd16S_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 2701 predicate(UseAVX > 1 && n->as_Vector()->length() == 16);
kvn@4001 2702 match(Set dst (AddVS src1 src2));
kvn@4001 2703 format %{ "vpaddw $dst,$src1,$src2\t! add packed16S" %}
kvn@4001 2704 ins_encode %{
kvn@4001 2705 bool vector256 = true;
kvn@4001 2706 __ vpaddw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2707 %}
kvn@4001 2708 ins_pipe( pipe_slow );
kvn@4001 2709 %}
kvn@4001 2710
kvn@4001 2711 instruct vadd16S_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 2712 predicate(UseAVX > 1 && n->as_Vector()->length() == 16);
kvn@4001 2713 match(Set dst (AddVS src (LoadVector mem)));
kvn@4001 2714 format %{ "vpaddw $dst,$src,$mem\t! add packed16S" %}
kvn@4001 2715 ins_encode %{
kvn@4001 2716 bool vector256 = true;
kvn@4001 2717 __ vpaddw($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 2718 %}
kvn@4001 2719 ins_pipe( pipe_slow );
kvn@4001 2720 %}
kvn@4001 2721
kvn@4001 2722 // Integers vector add
kvn@4001 2723 instruct vadd2I(vecD dst, vecD src) %{
kvn@4001 2724 predicate(n->as_Vector()->length() == 2);
kvn@4001 2725 match(Set dst (AddVI dst src));
kvn@4001 2726 format %{ "paddd $dst,$src\t! add packed2I" %}
kvn@4001 2727 ins_encode %{
kvn@4001 2728 __ paddd($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2729 %}
kvn@4001 2730 ins_pipe( pipe_slow );
kvn@4001 2731 %}
kvn@4001 2732
kvn@4001 2733 instruct vadd2I_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 2734 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 2735 match(Set dst (AddVI src1 src2));
kvn@4001 2736 format %{ "vpaddd $dst,$src1,$src2\t! add packed2I" %}
kvn@4001 2737 ins_encode %{
kvn@4001 2738 bool vector256 = false;
kvn@4001 2739 __ vpaddd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2740 %}
kvn@4001 2741 ins_pipe( pipe_slow );
kvn@4001 2742 %}
kvn@4001 2743
kvn@4001 2744 instruct vadd4I(vecX dst, vecX src) %{
kvn@4001 2745 predicate(n->as_Vector()->length() == 4);
kvn@4001 2746 match(Set dst (AddVI dst src));
kvn@4001 2747 format %{ "paddd $dst,$src\t! add packed4I" %}
kvn@4001 2748 ins_encode %{
kvn@4001 2749 __ paddd($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2750 %}
kvn@4001 2751 ins_pipe( pipe_slow );
kvn@4001 2752 %}
kvn@4001 2753
kvn@4001 2754 instruct vadd4I_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 2755 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 2756 match(Set dst (AddVI src1 src2));
kvn@4001 2757 format %{ "vpaddd $dst,$src1,$src2\t! add packed4I" %}
kvn@4001 2758 ins_encode %{
kvn@4001 2759 bool vector256 = false;
kvn@4001 2760 __ vpaddd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2761 %}
kvn@4001 2762 ins_pipe( pipe_slow );
kvn@4001 2763 %}
kvn@4001 2764
kvn@4001 2765 instruct vadd4I_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 2766 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 2767 match(Set dst (AddVI src (LoadVector mem)));
kvn@4001 2768 format %{ "vpaddd $dst,$src,$mem\t! add packed4I" %}
kvn@4001 2769 ins_encode %{
kvn@4001 2770 bool vector256 = false;
kvn@4001 2771 __ vpaddd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 2772 %}
kvn@4001 2773 ins_pipe( pipe_slow );
kvn@4001 2774 %}
kvn@4001 2775
kvn@4001 2776 instruct vadd8I_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 2777 predicate(UseAVX > 1 && n->as_Vector()->length() == 8);
kvn@4001 2778 match(Set dst (AddVI src1 src2));
kvn@4001 2779 format %{ "vpaddd $dst,$src1,$src2\t! add packed8I" %}
kvn@4001 2780 ins_encode %{
kvn@4001 2781 bool vector256 = true;
kvn@4001 2782 __ vpaddd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2783 %}
kvn@4001 2784 ins_pipe( pipe_slow );
kvn@4001 2785 %}
kvn@4001 2786
kvn@4001 2787 instruct vadd8I_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 2788 predicate(UseAVX > 1 && n->as_Vector()->length() == 8);
kvn@4001 2789 match(Set dst (AddVI src (LoadVector mem)));
kvn@4001 2790 format %{ "vpaddd $dst,$src,$mem\t! add packed8I" %}
kvn@4001 2791 ins_encode %{
kvn@4001 2792 bool vector256 = true;
kvn@4001 2793 __ vpaddd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 2794 %}
kvn@4001 2795 ins_pipe( pipe_slow );
kvn@4001 2796 %}
kvn@4001 2797
kvn@4001 2798 // Longs vector add
kvn@4001 2799 instruct vadd2L(vecX dst, vecX src) %{
kvn@4001 2800 predicate(n->as_Vector()->length() == 2);
kvn@4001 2801 match(Set dst (AddVL dst src));
kvn@4001 2802 format %{ "paddq $dst,$src\t! add packed2L" %}
kvn@4001 2803 ins_encode %{
kvn@4001 2804 __ paddq($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2805 %}
kvn@4001 2806 ins_pipe( pipe_slow );
kvn@4001 2807 %}
kvn@4001 2808
kvn@4001 2809 instruct vadd2L_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 2810 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 2811 match(Set dst (AddVL src1 src2));
kvn@4001 2812 format %{ "vpaddq $dst,$src1,$src2\t! add packed2L" %}
kvn@4001 2813 ins_encode %{
kvn@4001 2814 bool vector256 = false;
kvn@4001 2815 __ vpaddq($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2816 %}
kvn@4001 2817 ins_pipe( pipe_slow );
kvn@4001 2818 %}
kvn@4001 2819
kvn@4001 2820 instruct vadd2L_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 2821 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 2822 match(Set dst (AddVL src (LoadVector mem)));
kvn@4001 2823 format %{ "vpaddq $dst,$src,$mem\t! add packed2L" %}
kvn@4001 2824 ins_encode %{
kvn@4001 2825 bool vector256 = false;
kvn@4001 2826 __ vpaddq($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 2827 %}
kvn@4001 2828 ins_pipe( pipe_slow );
kvn@4001 2829 %}
kvn@4001 2830
kvn@4001 2831 instruct vadd4L_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 2832 predicate(UseAVX > 1 && n->as_Vector()->length() == 4);
kvn@4001 2833 match(Set dst (AddVL src1 src2));
kvn@4001 2834 format %{ "vpaddq $dst,$src1,$src2\t! add packed4L" %}
kvn@4001 2835 ins_encode %{
kvn@4001 2836 bool vector256 = true;
kvn@4001 2837 __ vpaddq($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2838 %}
kvn@4001 2839 ins_pipe( pipe_slow );
kvn@4001 2840 %}
kvn@4001 2841
kvn@4001 2842 instruct vadd4L_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 2843 predicate(UseAVX > 1 && n->as_Vector()->length() == 4);
kvn@4001 2844 match(Set dst (AddVL src (LoadVector mem)));
kvn@4001 2845 format %{ "vpaddq $dst,$src,$mem\t! add packed4L" %}
kvn@4001 2846 ins_encode %{
kvn@4001 2847 bool vector256 = true;
kvn@4001 2848 __ vpaddq($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 2849 %}
kvn@4001 2850 ins_pipe( pipe_slow );
kvn@4001 2851 %}
kvn@4001 2852
kvn@4001 2853 // Floats vector add
kvn@4001 2854 instruct vadd2F(vecD dst, vecD src) %{
kvn@4001 2855 predicate(n->as_Vector()->length() == 2);
kvn@4001 2856 match(Set dst (AddVF dst src));
kvn@4001 2857 format %{ "addps $dst,$src\t! add packed2F" %}
kvn@4001 2858 ins_encode %{
kvn@4001 2859 __ addps($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2860 %}
kvn@4001 2861 ins_pipe( pipe_slow );
kvn@4001 2862 %}
kvn@4001 2863
kvn@4001 2864 instruct vadd2F_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 2865 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 2866 match(Set dst (AddVF src1 src2));
kvn@4001 2867 format %{ "vaddps $dst,$src1,$src2\t! add packed2F" %}
kvn@4001 2868 ins_encode %{
kvn@4001 2869 bool vector256 = false;
kvn@4001 2870 __ vaddps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2871 %}
kvn@4001 2872 ins_pipe( pipe_slow );
kvn@4001 2873 %}
kvn@4001 2874
kvn@4001 2875 instruct vadd4F(vecX dst, vecX src) %{
kvn@4001 2876 predicate(n->as_Vector()->length() == 4);
kvn@4001 2877 match(Set dst (AddVF dst src));
kvn@4001 2878 format %{ "addps $dst,$src\t! add packed4F" %}
kvn@4001 2879 ins_encode %{
kvn@4001 2880 __ addps($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2881 %}
kvn@4001 2882 ins_pipe( pipe_slow );
kvn@4001 2883 %}
kvn@4001 2884
kvn@4001 2885 instruct vadd4F_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 2886 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 2887 match(Set dst (AddVF src1 src2));
kvn@4001 2888 format %{ "vaddps $dst,$src1,$src2\t! add packed4F" %}
kvn@4001 2889 ins_encode %{
kvn@4001 2890 bool vector256 = false;
kvn@4001 2891 __ vaddps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2892 %}
kvn@4001 2893 ins_pipe( pipe_slow );
kvn@4001 2894 %}
kvn@4001 2895
kvn@4001 2896 instruct vadd4F_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 2897 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 2898 match(Set dst (AddVF src (LoadVector mem)));
kvn@4001 2899 format %{ "vaddps $dst,$src,$mem\t! add packed4F" %}
kvn@4001 2900 ins_encode %{
kvn@4001 2901 bool vector256 = false;
kvn@4001 2902 __ vaddps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 2903 %}
kvn@4001 2904 ins_pipe( pipe_slow );
kvn@4001 2905 %}
kvn@4001 2906
kvn@4001 2907 instruct vadd8F_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 2908 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 2909 match(Set dst (AddVF src1 src2));
kvn@4001 2910 format %{ "vaddps $dst,$src1,$src2\t! add packed8F" %}
kvn@4001 2911 ins_encode %{
kvn@4001 2912 bool vector256 = true;
kvn@4001 2913 __ vaddps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2914 %}
kvn@4001 2915 ins_pipe( pipe_slow );
kvn@4001 2916 %}
kvn@4001 2917
kvn@4001 2918 instruct vadd8F_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 2919 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 2920 match(Set dst (AddVF src (LoadVector mem)));
kvn@4001 2921 format %{ "vaddps $dst,$src,$mem\t! add packed8F" %}
kvn@4001 2922 ins_encode %{
kvn@4001 2923 bool vector256 = true;
kvn@4001 2924 __ vaddps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 2925 %}
kvn@4001 2926 ins_pipe( pipe_slow );
kvn@4001 2927 %}
kvn@4001 2928
kvn@4001 2929 // Doubles vector add
kvn@4001 2930 instruct vadd2D(vecX dst, vecX src) %{
kvn@4001 2931 predicate(n->as_Vector()->length() == 2);
kvn@4001 2932 match(Set dst (AddVD dst src));
kvn@4001 2933 format %{ "addpd $dst,$src\t! add packed2D" %}
kvn@4001 2934 ins_encode %{
kvn@4001 2935 __ addpd($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2936 %}
kvn@4001 2937 ins_pipe( pipe_slow );
kvn@4001 2938 %}
kvn@4001 2939
kvn@4001 2940 instruct vadd2D_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 2941 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 2942 match(Set dst (AddVD src1 src2));
kvn@4001 2943 format %{ "vaddpd $dst,$src1,$src2\t! add packed2D" %}
kvn@4001 2944 ins_encode %{
kvn@4001 2945 bool vector256 = false;
kvn@4001 2946 __ vaddpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2947 %}
kvn@4001 2948 ins_pipe( pipe_slow );
kvn@4001 2949 %}
kvn@4001 2950
kvn@4001 2951 instruct vadd2D_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 2952 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 2953 match(Set dst (AddVD src (LoadVector mem)));
kvn@4001 2954 format %{ "vaddpd $dst,$src,$mem\t! add packed2D" %}
kvn@4001 2955 ins_encode %{
kvn@4001 2956 bool vector256 = false;
kvn@4001 2957 __ vaddpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 2958 %}
kvn@4001 2959 ins_pipe( pipe_slow );
kvn@4001 2960 %}
kvn@4001 2961
kvn@4001 2962 instruct vadd4D_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 2963 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 2964 match(Set dst (AddVD src1 src2));
kvn@4001 2965 format %{ "vaddpd $dst,$src1,$src2\t! add packed4D" %}
kvn@4001 2966 ins_encode %{
kvn@4001 2967 bool vector256 = true;
kvn@4001 2968 __ vaddpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 2969 %}
kvn@4001 2970 ins_pipe( pipe_slow );
kvn@4001 2971 %}
kvn@4001 2972
kvn@4001 2973 instruct vadd4D_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 2974 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 2975 match(Set dst (AddVD src (LoadVector mem)));
kvn@4001 2976 format %{ "vaddpd $dst,$src,$mem\t! add packed4D" %}
kvn@4001 2977 ins_encode %{
kvn@4001 2978 bool vector256 = true;
kvn@4001 2979 __ vaddpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 2980 %}
kvn@4001 2981 ins_pipe( pipe_slow );
kvn@4001 2982 %}
kvn@4001 2983
kvn@4001 2984 // --------------------------------- SUB --------------------------------------
kvn@4001 2985
kvn@4001 2986 // Bytes vector sub
kvn@4001 2987 instruct vsub4B(vecS dst, vecS src) %{
kvn@4001 2988 predicate(n->as_Vector()->length() == 4);
kvn@4001 2989 match(Set dst (SubVB dst src));
kvn@4001 2990 format %{ "psubb $dst,$src\t! sub packed4B" %}
kvn@4001 2991 ins_encode %{
kvn@4001 2992 __ psubb($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 2993 %}
kvn@4001 2994 ins_pipe( pipe_slow );
kvn@4001 2995 %}
kvn@4001 2996
kvn@4001 2997 instruct vsub4B_reg(vecS dst, vecS src1, vecS src2) %{
kvn@4001 2998 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 2999 match(Set dst (SubVB src1 src2));
kvn@4001 3000 format %{ "vpsubb $dst,$src1,$src2\t! sub packed4B" %}
kvn@4001 3001 ins_encode %{
kvn@4001 3002 bool vector256 = false;
kvn@4001 3003 __ vpsubb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3004 %}
kvn@4001 3005 ins_pipe( pipe_slow );
kvn@4001 3006 %}
kvn@4001 3007
kvn@4001 3008 instruct vsub8B(vecD dst, vecD src) %{
kvn@4001 3009 predicate(n->as_Vector()->length() == 8);
kvn@4001 3010 match(Set dst (SubVB dst src));
kvn@4001 3011 format %{ "psubb $dst,$src\t! sub packed8B" %}
kvn@4001 3012 ins_encode %{
kvn@4001 3013 __ psubb($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3014 %}
kvn@4001 3015 ins_pipe( pipe_slow );
kvn@4001 3016 %}
kvn@4001 3017
kvn@4001 3018 instruct vsub8B_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 3019 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 3020 match(Set dst (SubVB src1 src2));
kvn@4001 3021 format %{ "vpsubb $dst,$src1,$src2\t! sub packed8B" %}
kvn@4001 3022 ins_encode %{
kvn@4001 3023 bool vector256 = false;
kvn@4001 3024 __ vpsubb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3025 %}
kvn@4001 3026 ins_pipe( pipe_slow );
kvn@4001 3027 %}
kvn@4001 3028
kvn@4001 3029 instruct vsub16B(vecX dst, vecX src) %{
kvn@4001 3030 predicate(n->as_Vector()->length() == 16);
kvn@4001 3031 match(Set dst (SubVB dst src));
kvn@4001 3032 format %{ "psubb $dst,$src\t! sub packed16B" %}
kvn@4001 3033 ins_encode %{
kvn@4001 3034 __ psubb($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3035 %}
kvn@4001 3036 ins_pipe( pipe_slow );
kvn@4001 3037 %}
kvn@4001 3038
kvn@4001 3039 instruct vsub16B_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 3040 predicate(UseAVX > 0 && n->as_Vector()->length() == 16);
kvn@4001 3041 match(Set dst (SubVB src1 src2));
kvn@4001 3042 format %{ "vpsubb $dst,$src1,$src2\t! sub packed16B" %}
kvn@4001 3043 ins_encode %{
kvn@4001 3044 bool vector256 = false;
kvn@4001 3045 __ vpsubb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3046 %}
kvn@4001 3047 ins_pipe( pipe_slow );
kvn@4001 3048 %}
kvn@4001 3049
kvn@4001 3050 instruct vsub16B_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 3051 predicate(UseAVX > 0 && n->as_Vector()->length() == 16);
kvn@4001 3052 match(Set dst (SubVB src (LoadVector mem)));
kvn@4001 3053 format %{ "vpsubb $dst,$src,$mem\t! sub packed16B" %}
kvn@4001 3054 ins_encode %{
kvn@4001 3055 bool vector256 = false;
kvn@4001 3056 __ vpsubb($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3057 %}
kvn@4001 3058 ins_pipe( pipe_slow );
kvn@4001 3059 %}
kvn@4001 3060
kvn@4001 3061 instruct vsub32B_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 3062 predicate(UseAVX > 1 && n->as_Vector()->length() == 32);
kvn@4001 3063 match(Set dst (SubVB src1 src2));
kvn@4001 3064 format %{ "vpsubb $dst,$src1,$src2\t! sub packed32B" %}
kvn@4001 3065 ins_encode %{
kvn@4001 3066 bool vector256 = true;
kvn@4001 3067 __ vpsubb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3068 %}
kvn@4001 3069 ins_pipe( pipe_slow );
kvn@4001 3070 %}
kvn@4001 3071
kvn@4001 3072 instruct vsub32B_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 3073 predicate(UseAVX > 1 && n->as_Vector()->length() == 32);
kvn@4001 3074 match(Set dst (SubVB src (LoadVector mem)));
kvn@4001 3075 format %{ "vpsubb $dst,$src,$mem\t! sub packed32B" %}
kvn@4001 3076 ins_encode %{
kvn@4001 3077 bool vector256 = true;
kvn@4001 3078 __ vpsubb($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3079 %}
kvn@4001 3080 ins_pipe( pipe_slow );
kvn@4001 3081 %}
kvn@4001 3082
kvn@4001 3083 // Shorts/Chars vector sub
kvn@4001 3084 instruct vsub2S(vecS dst, vecS src) %{
kvn@4001 3085 predicate(n->as_Vector()->length() == 2);
kvn@4001 3086 match(Set dst (SubVS dst src));
kvn@4001 3087 format %{ "psubw $dst,$src\t! sub packed2S" %}
kvn@4001 3088 ins_encode %{
kvn@4001 3089 __ psubw($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3090 %}
kvn@4001 3091 ins_pipe( pipe_slow );
kvn@4001 3092 %}
kvn@4001 3093
kvn@4001 3094 instruct vsub2S_reg(vecS dst, vecS src1, vecS src2) %{
kvn@4001 3095 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3096 match(Set dst (SubVS src1 src2));
kvn@4001 3097 format %{ "vpsubw $dst,$src1,$src2\t! sub packed2S" %}
kvn@4001 3098 ins_encode %{
kvn@4001 3099 bool vector256 = false;
kvn@4001 3100 __ vpsubw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3101 %}
kvn@4001 3102 ins_pipe( pipe_slow );
kvn@4001 3103 %}
kvn@4001 3104
kvn@4001 3105 instruct vsub4S(vecD dst, vecD src) %{
kvn@4001 3106 predicate(n->as_Vector()->length() == 4);
kvn@4001 3107 match(Set dst (SubVS dst src));
kvn@4001 3108 format %{ "psubw $dst,$src\t! sub packed4S" %}
kvn@4001 3109 ins_encode %{
kvn@4001 3110 __ psubw($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3111 %}
kvn@4001 3112 ins_pipe( pipe_slow );
kvn@4001 3113 %}
kvn@4001 3114
kvn@4001 3115 instruct vsub4S_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 3116 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3117 match(Set dst (SubVS src1 src2));
kvn@4001 3118 format %{ "vpsubw $dst,$src1,$src2\t! sub packed4S" %}
kvn@4001 3119 ins_encode %{
kvn@4001 3120 bool vector256 = false;
kvn@4001 3121 __ vpsubw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3122 %}
kvn@4001 3123 ins_pipe( pipe_slow );
kvn@4001 3124 %}
kvn@4001 3125
kvn@4001 3126 instruct vsub8S(vecX dst, vecX src) %{
kvn@4001 3127 predicate(n->as_Vector()->length() == 8);
kvn@4001 3128 match(Set dst (SubVS dst src));
kvn@4001 3129 format %{ "psubw $dst,$src\t! sub packed8S" %}
kvn@4001 3130 ins_encode %{
kvn@4001 3131 __ psubw($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3132 %}
kvn@4001 3133 ins_pipe( pipe_slow );
kvn@4001 3134 %}
kvn@4001 3135
kvn@4001 3136 instruct vsub8S_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 3137 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 3138 match(Set dst (SubVS src1 src2));
kvn@4001 3139 format %{ "vpsubw $dst,$src1,$src2\t! sub packed8S" %}
kvn@4001 3140 ins_encode %{
kvn@4001 3141 bool vector256 = false;
kvn@4001 3142 __ vpsubw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3143 %}
kvn@4001 3144 ins_pipe( pipe_slow );
kvn@4001 3145 %}
kvn@4001 3146
kvn@4001 3147 instruct vsub8S_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 3148 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 3149 match(Set dst (SubVS src (LoadVector mem)));
kvn@4001 3150 format %{ "vpsubw $dst,$src,$mem\t! sub packed8S" %}
kvn@4001 3151 ins_encode %{
kvn@4001 3152 bool vector256 = false;
kvn@4001 3153 __ vpsubw($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3154 %}
kvn@4001 3155 ins_pipe( pipe_slow );
kvn@4001 3156 %}
kvn@4001 3157
kvn@4001 3158 instruct vsub16S_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 3159 predicate(UseAVX > 1 && n->as_Vector()->length() == 16);
kvn@4001 3160 match(Set dst (SubVS src1 src2));
kvn@4001 3161 format %{ "vpsubw $dst,$src1,$src2\t! sub packed16S" %}
kvn@4001 3162 ins_encode %{
kvn@4001 3163 bool vector256 = true;
kvn@4001 3164 __ vpsubw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3165 %}
kvn@4001 3166 ins_pipe( pipe_slow );
kvn@4001 3167 %}
kvn@4001 3168
kvn@4001 3169 instruct vsub16S_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 3170 predicate(UseAVX > 1 && n->as_Vector()->length() == 16);
kvn@4001 3171 match(Set dst (SubVS src (LoadVector mem)));
kvn@4001 3172 format %{ "vpsubw $dst,$src,$mem\t! sub packed16S" %}
kvn@4001 3173 ins_encode %{
kvn@4001 3174 bool vector256 = true;
kvn@4001 3175 __ vpsubw($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3176 %}
kvn@4001 3177 ins_pipe( pipe_slow );
kvn@4001 3178 %}
kvn@4001 3179
kvn@4001 3180 // Integers vector sub
kvn@4001 3181 instruct vsub2I(vecD dst, vecD src) %{
kvn@4001 3182 predicate(n->as_Vector()->length() == 2);
kvn@4001 3183 match(Set dst (SubVI dst src));
kvn@4001 3184 format %{ "psubd $dst,$src\t! sub packed2I" %}
kvn@4001 3185 ins_encode %{
kvn@4001 3186 __ psubd($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3187 %}
kvn@4001 3188 ins_pipe( pipe_slow );
kvn@4001 3189 %}
kvn@4001 3190
kvn@4001 3191 instruct vsub2I_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 3192 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3193 match(Set dst (SubVI src1 src2));
kvn@4001 3194 format %{ "vpsubd $dst,$src1,$src2\t! sub packed2I" %}
kvn@4001 3195 ins_encode %{
kvn@4001 3196 bool vector256 = false;
kvn@4001 3197 __ vpsubd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3198 %}
kvn@4001 3199 ins_pipe( pipe_slow );
kvn@4001 3200 %}
kvn@4001 3201
kvn@4001 3202 instruct vsub4I(vecX dst, vecX src) %{
kvn@4001 3203 predicate(n->as_Vector()->length() == 4);
kvn@4001 3204 match(Set dst (SubVI dst src));
kvn@4001 3205 format %{ "psubd $dst,$src\t! sub packed4I" %}
kvn@4001 3206 ins_encode %{
kvn@4001 3207 __ psubd($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3208 %}
kvn@4001 3209 ins_pipe( pipe_slow );
kvn@4001 3210 %}
kvn@4001 3211
kvn@4001 3212 instruct vsub4I_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 3213 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3214 match(Set dst (SubVI src1 src2));
kvn@4001 3215 format %{ "vpsubd $dst,$src1,$src2\t! sub packed4I" %}
kvn@4001 3216 ins_encode %{
kvn@4001 3217 bool vector256 = false;
kvn@4001 3218 __ vpsubd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3219 %}
kvn@4001 3220 ins_pipe( pipe_slow );
kvn@4001 3221 %}
kvn@4001 3222
kvn@4001 3223 instruct vsub4I_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 3224 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3225 match(Set dst (SubVI src (LoadVector mem)));
kvn@4001 3226 format %{ "vpsubd $dst,$src,$mem\t! sub packed4I" %}
kvn@4001 3227 ins_encode %{
kvn@4001 3228 bool vector256 = false;
kvn@4001 3229 __ vpsubd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3230 %}
kvn@4001 3231 ins_pipe( pipe_slow );
kvn@4001 3232 %}
kvn@4001 3233
kvn@4001 3234 instruct vsub8I_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 3235 predicate(UseAVX > 1 && n->as_Vector()->length() == 8);
kvn@4001 3236 match(Set dst (SubVI src1 src2));
kvn@4001 3237 format %{ "vpsubd $dst,$src1,$src2\t! sub packed8I" %}
kvn@4001 3238 ins_encode %{
kvn@4001 3239 bool vector256 = true;
kvn@4001 3240 __ vpsubd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3241 %}
kvn@4001 3242 ins_pipe( pipe_slow );
kvn@4001 3243 %}
kvn@4001 3244
kvn@4001 3245 instruct vsub8I_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 3246 predicate(UseAVX > 1 && n->as_Vector()->length() == 8);
kvn@4001 3247 match(Set dst (SubVI src (LoadVector mem)));
kvn@4001 3248 format %{ "vpsubd $dst,$src,$mem\t! sub packed8I" %}
kvn@4001 3249 ins_encode %{
kvn@4001 3250 bool vector256 = true;
kvn@4001 3251 __ vpsubd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3252 %}
kvn@4001 3253 ins_pipe( pipe_slow );
kvn@4001 3254 %}
kvn@4001 3255
kvn@4001 3256 // Longs vector sub
kvn@4001 3257 instruct vsub2L(vecX dst, vecX src) %{
kvn@4001 3258 predicate(n->as_Vector()->length() == 2);
kvn@4001 3259 match(Set dst (SubVL dst src));
kvn@4001 3260 format %{ "psubq $dst,$src\t! sub packed2L" %}
kvn@4001 3261 ins_encode %{
kvn@4001 3262 __ psubq($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3263 %}
kvn@4001 3264 ins_pipe( pipe_slow );
kvn@4001 3265 %}
kvn@4001 3266
kvn@4001 3267 instruct vsub2L_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 3268 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3269 match(Set dst (SubVL src1 src2));
kvn@4001 3270 format %{ "vpsubq $dst,$src1,$src2\t! sub packed2L" %}
kvn@4001 3271 ins_encode %{
kvn@4001 3272 bool vector256 = false;
kvn@4001 3273 __ vpsubq($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3274 %}
kvn@4001 3275 ins_pipe( pipe_slow );
kvn@4001 3276 %}
kvn@4001 3277
kvn@4001 3278 instruct vsub2L_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 3279 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3280 match(Set dst (SubVL src (LoadVector mem)));
kvn@4001 3281 format %{ "vpsubq $dst,$src,$mem\t! sub packed2L" %}
kvn@4001 3282 ins_encode %{
kvn@4001 3283 bool vector256 = false;
kvn@4001 3284 __ vpsubq($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3285 %}
kvn@4001 3286 ins_pipe( pipe_slow );
kvn@4001 3287 %}
kvn@4001 3288
kvn@4001 3289 instruct vsub4L_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 3290 predicate(UseAVX > 1 && n->as_Vector()->length() == 4);
kvn@4001 3291 match(Set dst (SubVL src1 src2));
kvn@4001 3292 format %{ "vpsubq $dst,$src1,$src2\t! sub packed4L" %}
kvn@4001 3293 ins_encode %{
kvn@4001 3294 bool vector256 = true;
kvn@4001 3295 __ vpsubq($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3296 %}
kvn@4001 3297 ins_pipe( pipe_slow );
kvn@4001 3298 %}
kvn@4001 3299
kvn@4001 3300 instruct vsub4L_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 3301 predicate(UseAVX > 1 && n->as_Vector()->length() == 4);
kvn@4001 3302 match(Set dst (SubVL src (LoadVector mem)));
kvn@4001 3303 format %{ "vpsubq $dst,$src,$mem\t! sub packed4L" %}
kvn@4001 3304 ins_encode %{
kvn@4001 3305 bool vector256 = true;
kvn@4001 3306 __ vpsubq($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3307 %}
kvn@4001 3308 ins_pipe( pipe_slow );
kvn@4001 3309 %}
kvn@4001 3310
kvn@4001 3311 // Floats vector sub
kvn@4001 3312 instruct vsub2F(vecD dst, vecD src) %{
kvn@4001 3313 predicate(n->as_Vector()->length() == 2);
kvn@4001 3314 match(Set dst (SubVF dst src));
kvn@4001 3315 format %{ "subps $dst,$src\t! sub packed2F" %}
kvn@4001 3316 ins_encode %{
kvn@4001 3317 __ subps($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3318 %}
kvn@4001 3319 ins_pipe( pipe_slow );
kvn@4001 3320 %}
kvn@4001 3321
kvn@4001 3322 instruct vsub2F_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 3323 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3324 match(Set dst (SubVF src1 src2));
kvn@4001 3325 format %{ "vsubps $dst,$src1,$src2\t! sub packed2F" %}
kvn@4001 3326 ins_encode %{
kvn@4001 3327 bool vector256 = false;
kvn@4001 3328 __ vsubps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3329 %}
kvn@4001 3330 ins_pipe( pipe_slow );
kvn@4001 3331 %}
kvn@4001 3332
kvn@4001 3333 instruct vsub4F(vecX dst, vecX src) %{
kvn@4001 3334 predicate(n->as_Vector()->length() == 4);
kvn@4001 3335 match(Set dst (SubVF dst src));
kvn@4001 3336 format %{ "subps $dst,$src\t! sub packed4F" %}
kvn@4001 3337 ins_encode %{
kvn@4001 3338 __ subps($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3339 %}
kvn@4001 3340 ins_pipe( pipe_slow );
kvn@4001 3341 %}
kvn@4001 3342
kvn@4001 3343 instruct vsub4F_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 3344 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3345 match(Set dst (SubVF src1 src2));
kvn@4001 3346 format %{ "vsubps $dst,$src1,$src2\t! sub packed4F" %}
kvn@4001 3347 ins_encode %{
kvn@4001 3348 bool vector256 = false;
kvn@4001 3349 __ vsubps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3350 %}
kvn@4001 3351 ins_pipe( pipe_slow );
kvn@4001 3352 %}
kvn@4001 3353
kvn@4001 3354 instruct vsub4F_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 3355 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3356 match(Set dst (SubVF src (LoadVector mem)));
kvn@4001 3357 format %{ "vsubps $dst,$src,$mem\t! sub packed4F" %}
kvn@4001 3358 ins_encode %{
kvn@4001 3359 bool vector256 = false;
kvn@4001 3360 __ vsubps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3361 %}
kvn@4001 3362 ins_pipe( pipe_slow );
kvn@4001 3363 %}
kvn@4001 3364
kvn@4001 3365 instruct vsub8F_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 3366 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 3367 match(Set dst (SubVF src1 src2));
kvn@4001 3368 format %{ "vsubps $dst,$src1,$src2\t! sub packed8F" %}
kvn@4001 3369 ins_encode %{
kvn@4001 3370 bool vector256 = true;
kvn@4001 3371 __ vsubps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3372 %}
kvn@4001 3373 ins_pipe( pipe_slow );
kvn@4001 3374 %}
kvn@4001 3375
kvn@4001 3376 instruct vsub8F_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 3377 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 3378 match(Set dst (SubVF src (LoadVector mem)));
kvn@4001 3379 format %{ "vsubps $dst,$src,$mem\t! sub packed8F" %}
kvn@4001 3380 ins_encode %{
kvn@4001 3381 bool vector256 = true;
kvn@4001 3382 __ vsubps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3383 %}
kvn@4001 3384 ins_pipe( pipe_slow );
kvn@4001 3385 %}
kvn@4001 3386
kvn@4001 3387 // Doubles vector sub
kvn@4001 3388 instruct vsub2D(vecX dst, vecX src) %{
kvn@4001 3389 predicate(n->as_Vector()->length() == 2);
kvn@4001 3390 match(Set dst (SubVD dst src));
kvn@4001 3391 format %{ "subpd $dst,$src\t! sub packed2D" %}
kvn@4001 3392 ins_encode %{
kvn@4001 3393 __ subpd($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3394 %}
kvn@4001 3395 ins_pipe( pipe_slow );
kvn@4001 3396 %}
kvn@4001 3397
kvn@4001 3398 instruct vsub2D_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 3399 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3400 match(Set dst (SubVD src1 src2));
kvn@4001 3401 format %{ "vsubpd $dst,$src1,$src2\t! sub packed2D" %}
kvn@4001 3402 ins_encode %{
kvn@4001 3403 bool vector256 = false;
kvn@4001 3404 __ vsubpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3405 %}
kvn@4001 3406 ins_pipe( pipe_slow );
kvn@4001 3407 %}
kvn@4001 3408
kvn@4001 3409 instruct vsub2D_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 3410 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3411 match(Set dst (SubVD src (LoadVector mem)));
kvn@4001 3412 format %{ "vsubpd $dst,$src,$mem\t! sub packed2D" %}
kvn@4001 3413 ins_encode %{
kvn@4001 3414 bool vector256 = false;
kvn@4001 3415 __ vsubpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3416 %}
kvn@4001 3417 ins_pipe( pipe_slow );
kvn@4001 3418 %}
kvn@4001 3419
kvn@4001 3420 instruct vsub4D_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 3421 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3422 match(Set dst (SubVD src1 src2));
kvn@4001 3423 format %{ "vsubpd $dst,$src1,$src2\t! sub packed4D" %}
kvn@4001 3424 ins_encode %{
kvn@4001 3425 bool vector256 = true;
kvn@4001 3426 __ vsubpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3427 %}
kvn@4001 3428 ins_pipe( pipe_slow );
kvn@4001 3429 %}
kvn@4001 3430
kvn@4001 3431 instruct vsub4D_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 3432 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3433 match(Set dst (SubVD src (LoadVector mem)));
kvn@4001 3434 format %{ "vsubpd $dst,$src,$mem\t! sub packed4D" %}
kvn@4001 3435 ins_encode %{
kvn@4001 3436 bool vector256 = true;
kvn@4001 3437 __ vsubpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3438 %}
kvn@4001 3439 ins_pipe( pipe_slow );
kvn@4001 3440 %}
kvn@4001 3441
kvn@4001 3442 // --------------------------------- MUL --------------------------------------
kvn@4001 3443
kvn@4001 3444 // Shorts/Chars vector mul
kvn@4001 3445 instruct vmul2S(vecS dst, vecS src) %{
kvn@4001 3446 predicate(n->as_Vector()->length() == 2);
kvn@4001 3447 match(Set dst (MulVS dst src));
kvn@4001 3448 format %{ "pmullw $dst,$src\t! mul packed2S" %}
kvn@4001 3449 ins_encode %{
kvn@4001 3450 __ pmullw($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3451 %}
kvn@4001 3452 ins_pipe( pipe_slow );
kvn@4001 3453 %}
kvn@4001 3454
kvn@4001 3455 instruct vmul2S_reg(vecS dst, vecS src1, vecS src2) %{
kvn@4001 3456 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3457 match(Set dst (MulVS src1 src2));
kvn@4001 3458 format %{ "vpmullw $dst,$src1,$src2\t! mul packed2S" %}
kvn@4001 3459 ins_encode %{
kvn@4001 3460 bool vector256 = false;
kvn@4001 3461 __ vpmullw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3462 %}
kvn@4001 3463 ins_pipe( pipe_slow );
kvn@4001 3464 %}
kvn@4001 3465
kvn@4001 3466 instruct vmul4S(vecD dst, vecD src) %{
kvn@4001 3467 predicate(n->as_Vector()->length() == 4);
kvn@4001 3468 match(Set dst (MulVS dst src));
kvn@4001 3469 format %{ "pmullw $dst,$src\t! mul packed4S" %}
kvn@4001 3470 ins_encode %{
kvn@4001 3471 __ pmullw($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3472 %}
kvn@4001 3473 ins_pipe( pipe_slow );
kvn@4001 3474 %}
kvn@4001 3475
kvn@4001 3476 instruct vmul4S_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 3477 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3478 match(Set dst (MulVS src1 src2));
kvn@4001 3479 format %{ "vpmullw $dst,$src1,$src2\t! mul packed4S" %}
kvn@4001 3480 ins_encode %{
kvn@4001 3481 bool vector256 = false;
kvn@4001 3482 __ vpmullw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3483 %}
kvn@4001 3484 ins_pipe( pipe_slow );
kvn@4001 3485 %}
kvn@4001 3486
kvn@4001 3487 instruct vmul8S(vecX dst, vecX src) %{
kvn@4001 3488 predicate(n->as_Vector()->length() == 8);
kvn@4001 3489 match(Set dst (MulVS dst src));
kvn@4001 3490 format %{ "pmullw $dst,$src\t! mul packed8S" %}
kvn@4001 3491 ins_encode %{
kvn@4001 3492 __ pmullw($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3493 %}
kvn@4001 3494 ins_pipe( pipe_slow );
kvn@4001 3495 %}
kvn@4001 3496
kvn@4001 3497 instruct vmul8S_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 3498 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 3499 match(Set dst (MulVS src1 src2));
kvn@4001 3500 format %{ "vpmullw $dst,$src1,$src2\t! mul packed8S" %}
kvn@4001 3501 ins_encode %{
kvn@4001 3502 bool vector256 = false;
kvn@4001 3503 __ vpmullw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3504 %}
kvn@4001 3505 ins_pipe( pipe_slow );
kvn@4001 3506 %}
kvn@4001 3507
kvn@4001 3508 instruct vmul8S_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 3509 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 3510 match(Set dst (MulVS src (LoadVector mem)));
kvn@4001 3511 format %{ "vpmullw $dst,$src,$mem\t! mul packed8S" %}
kvn@4001 3512 ins_encode %{
kvn@4001 3513 bool vector256 = false;
kvn@4001 3514 __ vpmullw($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3515 %}
kvn@4001 3516 ins_pipe( pipe_slow );
kvn@4001 3517 %}
kvn@4001 3518
kvn@4001 3519 instruct vmul16S_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 3520 predicate(UseAVX > 1 && n->as_Vector()->length() == 16);
kvn@4001 3521 match(Set dst (MulVS src1 src2));
kvn@4001 3522 format %{ "vpmullw $dst,$src1,$src2\t! mul packed16S" %}
kvn@4001 3523 ins_encode %{
kvn@4001 3524 bool vector256 = true;
kvn@4001 3525 __ vpmullw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3526 %}
kvn@4001 3527 ins_pipe( pipe_slow );
kvn@4001 3528 %}
kvn@4001 3529
kvn@4001 3530 instruct vmul16S_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 3531 predicate(UseAVX > 1 && n->as_Vector()->length() == 16);
kvn@4001 3532 match(Set dst (MulVS src (LoadVector mem)));
kvn@4001 3533 format %{ "vpmullw $dst,$src,$mem\t! mul packed16S" %}
kvn@4001 3534 ins_encode %{
kvn@4001 3535 bool vector256 = true;
kvn@4001 3536 __ vpmullw($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3537 %}
kvn@4001 3538 ins_pipe( pipe_slow );
kvn@4001 3539 %}
kvn@4001 3540
kvn@4001 3541 // Integers vector mul (sse4_1)
kvn@4001 3542 instruct vmul2I(vecD dst, vecD src) %{
kvn@4001 3543 predicate(UseSSE > 3 && n->as_Vector()->length() == 2);
kvn@4001 3544 match(Set dst (MulVI dst src));
kvn@4001 3545 format %{ "pmulld $dst,$src\t! mul packed2I" %}
kvn@4001 3546 ins_encode %{
kvn@4001 3547 __ pmulld($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3548 %}
kvn@4001 3549 ins_pipe( pipe_slow );
kvn@4001 3550 %}
kvn@4001 3551
kvn@4001 3552 instruct vmul2I_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 3553 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3554 match(Set dst (MulVI src1 src2));
kvn@4001 3555 format %{ "vpmulld $dst,$src1,$src2\t! mul packed2I" %}
kvn@4001 3556 ins_encode %{
kvn@4001 3557 bool vector256 = false;
kvn@4001 3558 __ vpmulld($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3559 %}
kvn@4001 3560 ins_pipe( pipe_slow );
kvn@4001 3561 %}
kvn@4001 3562
kvn@4001 3563 instruct vmul4I(vecX dst, vecX src) %{
kvn@4001 3564 predicate(UseSSE > 3 && n->as_Vector()->length() == 4);
kvn@4001 3565 match(Set dst (MulVI dst src));
kvn@4001 3566 format %{ "pmulld $dst,$src\t! mul packed4I" %}
kvn@4001 3567 ins_encode %{
kvn@4001 3568 __ pmulld($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3569 %}
kvn@4001 3570 ins_pipe( pipe_slow );
kvn@4001 3571 %}
kvn@4001 3572
kvn@4001 3573 instruct vmul4I_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 3574 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3575 match(Set dst (MulVI src1 src2));
kvn@4001 3576 format %{ "vpmulld $dst,$src1,$src2\t! mul packed4I" %}
kvn@4001 3577 ins_encode %{
kvn@4001 3578 bool vector256 = false;
kvn@4001 3579 __ vpmulld($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3580 %}
kvn@4001 3581 ins_pipe( pipe_slow );
kvn@4001 3582 %}
kvn@4001 3583
kvn@4001 3584 instruct vmul4I_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 3585 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3586 match(Set dst (MulVI src (LoadVector mem)));
kvn@4001 3587 format %{ "vpmulld $dst,$src,$mem\t! mul packed4I" %}
kvn@4001 3588 ins_encode %{
kvn@4001 3589 bool vector256 = false;
kvn@4001 3590 __ vpmulld($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3591 %}
kvn@4001 3592 ins_pipe( pipe_slow );
kvn@4001 3593 %}
kvn@4001 3594
kvn@4001 3595 instruct vmul8I_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 3596 predicate(UseAVX > 1 && n->as_Vector()->length() == 8);
kvn@4001 3597 match(Set dst (MulVI src1 src2));
kvn@4001 3598 format %{ "vpmulld $dst,$src1,$src2\t! mul packed8I" %}
kvn@4001 3599 ins_encode %{
kvn@4001 3600 bool vector256 = true;
kvn@4001 3601 __ vpmulld($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3602 %}
kvn@4001 3603 ins_pipe( pipe_slow );
kvn@4001 3604 %}
kvn@4001 3605
kvn@4001 3606 instruct vmul8I_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 3607 predicate(UseAVX > 1 && n->as_Vector()->length() == 8);
kvn@4001 3608 match(Set dst (MulVI src (LoadVector mem)));
kvn@4001 3609 format %{ "vpmulld $dst,$src,$mem\t! mul packed8I" %}
kvn@4001 3610 ins_encode %{
kvn@4001 3611 bool vector256 = true;
kvn@4001 3612 __ vpmulld($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3613 %}
kvn@4001 3614 ins_pipe( pipe_slow );
kvn@4001 3615 %}
kvn@4001 3616
kvn@4001 3617 // Floats vector mul
kvn@4001 3618 instruct vmul2F(vecD dst, vecD src) %{
kvn@4001 3619 predicate(n->as_Vector()->length() == 2);
kvn@4001 3620 match(Set dst (MulVF dst src));
kvn@4001 3621 format %{ "mulps $dst,$src\t! mul packed2F" %}
kvn@4001 3622 ins_encode %{
kvn@4001 3623 __ mulps($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3624 %}
kvn@4001 3625 ins_pipe( pipe_slow );
kvn@4001 3626 %}
kvn@4001 3627
kvn@4001 3628 instruct vmul2F_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 3629 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3630 match(Set dst (MulVF src1 src2));
kvn@4001 3631 format %{ "vmulps $dst,$src1,$src2\t! mul packed2F" %}
kvn@4001 3632 ins_encode %{
kvn@4001 3633 bool vector256 = false;
kvn@4001 3634 __ vmulps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3635 %}
kvn@4001 3636 ins_pipe( pipe_slow );
kvn@4001 3637 %}
kvn@4001 3638
kvn@4001 3639 instruct vmul4F(vecX dst, vecX src) %{
kvn@4001 3640 predicate(n->as_Vector()->length() == 4);
kvn@4001 3641 match(Set dst (MulVF dst src));
kvn@4001 3642 format %{ "mulps $dst,$src\t! mul packed4F" %}
kvn@4001 3643 ins_encode %{
kvn@4001 3644 __ mulps($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3645 %}
kvn@4001 3646 ins_pipe( pipe_slow );
kvn@4001 3647 %}
kvn@4001 3648
kvn@4001 3649 instruct vmul4F_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 3650 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3651 match(Set dst (MulVF src1 src2));
kvn@4001 3652 format %{ "vmulps $dst,$src1,$src2\t! mul packed4F" %}
kvn@4001 3653 ins_encode %{
kvn@4001 3654 bool vector256 = false;
kvn@4001 3655 __ vmulps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3656 %}
kvn@4001 3657 ins_pipe( pipe_slow );
kvn@4001 3658 %}
kvn@4001 3659
kvn@4001 3660 instruct vmul4F_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 3661 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3662 match(Set dst (MulVF src (LoadVector mem)));
kvn@4001 3663 format %{ "vmulps $dst,$src,$mem\t! mul packed4F" %}
kvn@4001 3664 ins_encode %{
kvn@4001 3665 bool vector256 = false;
kvn@4001 3666 __ vmulps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3667 %}
kvn@4001 3668 ins_pipe( pipe_slow );
kvn@4001 3669 %}
kvn@4001 3670
kvn@4001 3671 instruct vmul8F_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 3672 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 3673 match(Set dst (MulVF src1 src2));
kvn@4001 3674 format %{ "vmulps $dst,$src1,$src2\t! mul packed8F" %}
kvn@4001 3675 ins_encode %{
kvn@4001 3676 bool vector256 = true;
kvn@4001 3677 __ vmulps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3678 %}
kvn@4001 3679 ins_pipe( pipe_slow );
kvn@4001 3680 %}
kvn@4001 3681
kvn@4001 3682 instruct vmul8F_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 3683 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 3684 match(Set dst (MulVF src (LoadVector mem)));
kvn@4001 3685 format %{ "vmulps $dst,$src,$mem\t! mul packed8F" %}
kvn@4001 3686 ins_encode %{
kvn@4001 3687 bool vector256 = true;
kvn@4001 3688 __ vmulps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3689 %}
kvn@4001 3690 ins_pipe( pipe_slow );
kvn@4001 3691 %}
kvn@4001 3692
kvn@4001 3693 // Doubles vector mul
kvn@4001 3694 instruct vmul2D(vecX dst, vecX src) %{
kvn@4001 3695 predicate(n->as_Vector()->length() == 2);
kvn@4001 3696 match(Set dst (MulVD dst src));
kvn@4001 3697 format %{ "mulpd $dst,$src\t! mul packed2D" %}
kvn@4001 3698 ins_encode %{
kvn@4001 3699 __ mulpd($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3700 %}
kvn@4001 3701 ins_pipe( pipe_slow );
kvn@4001 3702 %}
kvn@4001 3703
kvn@4001 3704 instruct vmul2D_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 3705 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3706 match(Set dst (MulVD src1 src2));
kvn@4001 3707 format %{ "vmulpd $dst,$src1,$src2\t! mul packed2D" %}
kvn@4001 3708 ins_encode %{
kvn@4001 3709 bool vector256 = false;
kvn@4001 3710 __ vmulpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3711 %}
kvn@4001 3712 ins_pipe( pipe_slow );
kvn@4001 3713 %}
kvn@4001 3714
kvn@4001 3715 instruct vmul2D_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 3716 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3717 match(Set dst (MulVD src (LoadVector mem)));
kvn@4001 3718 format %{ "vmulpd $dst,$src,$mem\t! mul packed2D" %}
kvn@4001 3719 ins_encode %{
kvn@4001 3720 bool vector256 = false;
kvn@4001 3721 __ vmulpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3722 %}
kvn@4001 3723 ins_pipe( pipe_slow );
kvn@4001 3724 %}
kvn@4001 3725
kvn@4001 3726 instruct vmul4D_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 3727 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3728 match(Set dst (MulVD src1 src2));
kvn@4001 3729 format %{ "vmulpd $dst,$src1,$src2\t! mul packed4D" %}
kvn@4001 3730 ins_encode %{
kvn@4001 3731 bool vector256 = true;
kvn@4001 3732 __ vmulpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3733 %}
kvn@4001 3734 ins_pipe( pipe_slow );
kvn@4001 3735 %}
kvn@4001 3736
kvn@4001 3737 instruct vmul4D_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 3738 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3739 match(Set dst (MulVD src (LoadVector mem)));
kvn@4001 3740 format %{ "vmulpd $dst,$src,$mem\t! mul packed4D" %}
kvn@4001 3741 ins_encode %{
kvn@4001 3742 bool vector256 = true;
kvn@4001 3743 __ vmulpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3744 %}
kvn@4001 3745 ins_pipe( pipe_slow );
kvn@4001 3746 %}
kvn@4001 3747
kvn@4001 3748 // --------------------------------- DIV --------------------------------------
kvn@4001 3749
kvn@4001 3750 // Floats vector div
kvn@4001 3751 instruct vdiv2F(vecD dst, vecD src) %{
kvn@4001 3752 predicate(n->as_Vector()->length() == 2);
kvn@4001 3753 match(Set dst (DivVF dst src));
kvn@4001 3754 format %{ "divps $dst,$src\t! div packed2F" %}
kvn@4001 3755 ins_encode %{
kvn@4001 3756 __ divps($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3757 %}
kvn@4001 3758 ins_pipe( pipe_slow );
kvn@4001 3759 %}
kvn@4001 3760
kvn@4001 3761 instruct vdiv2F_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 3762 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3763 match(Set dst (DivVF src1 src2));
kvn@4001 3764 format %{ "vdivps $dst,$src1,$src2\t! div packed2F" %}
kvn@4001 3765 ins_encode %{
kvn@4001 3766 bool vector256 = false;
kvn@4001 3767 __ vdivps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3768 %}
kvn@4001 3769 ins_pipe( pipe_slow );
kvn@4001 3770 %}
kvn@4001 3771
kvn@4001 3772 instruct vdiv4F(vecX dst, vecX src) %{
kvn@4001 3773 predicate(n->as_Vector()->length() == 4);
kvn@4001 3774 match(Set dst (DivVF dst src));
kvn@4001 3775 format %{ "divps $dst,$src\t! div packed4F" %}
kvn@4001 3776 ins_encode %{
kvn@4001 3777 __ divps($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3778 %}
kvn@4001 3779 ins_pipe( pipe_slow );
kvn@4001 3780 %}
kvn@4001 3781
kvn@4001 3782 instruct vdiv4F_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 3783 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3784 match(Set dst (DivVF src1 src2));
kvn@4001 3785 format %{ "vdivps $dst,$src1,$src2\t! div packed4F" %}
kvn@4001 3786 ins_encode %{
kvn@4001 3787 bool vector256 = false;
kvn@4001 3788 __ vdivps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3789 %}
kvn@4001 3790 ins_pipe( pipe_slow );
kvn@4001 3791 %}
kvn@4001 3792
kvn@4001 3793 instruct vdiv4F_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 3794 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3795 match(Set dst (DivVF src (LoadVector mem)));
kvn@4001 3796 format %{ "vdivps $dst,$src,$mem\t! div packed4F" %}
kvn@4001 3797 ins_encode %{
kvn@4001 3798 bool vector256 = false;
kvn@4001 3799 __ vdivps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3800 %}
kvn@4001 3801 ins_pipe( pipe_slow );
kvn@4001 3802 %}
kvn@4001 3803
kvn@4001 3804 instruct vdiv8F_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 3805 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 3806 match(Set dst (DivVF src1 src2));
kvn@4001 3807 format %{ "vdivps $dst,$src1,$src2\t! div packed8F" %}
kvn@4001 3808 ins_encode %{
kvn@4001 3809 bool vector256 = true;
kvn@4001 3810 __ vdivps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3811 %}
kvn@4001 3812 ins_pipe( pipe_slow );
kvn@4001 3813 %}
kvn@4001 3814
kvn@4001 3815 instruct vdiv8F_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 3816 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 3817 match(Set dst (DivVF src (LoadVector mem)));
kvn@4001 3818 format %{ "vdivps $dst,$src,$mem\t! div packed8F" %}
kvn@4001 3819 ins_encode %{
kvn@4001 3820 bool vector256 = true;
kvn@4001 3821 __ vdivps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3822 %}
kvn@4001 3823 ins_pipe( pipe_slow );
kvn@4001 3824 %}
kvn@4001 3825
kvn@4001 3826 // Doubles vector div
kvn@4001 3827 instruct vdiv2D(vecX dst, vecX src) %{
kvn@4001 3828 predicate(n->as_Vector()->length() == 2);
kvn@4001 3829 match(Set dst (DivVD dst src));
kvn@4001 3830 format %{ "divpd $dst,$src\t! div packed2D" %}
kvn@4001 3831 ins_encode %{
kvn@4001 3832 __ divpd($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 3833 %}
kvn@4001 3834 ins_pipe( pipe_slow );
kvn@4001 3835 %}
kvn@4001 3836
kvn@4001 3837 instruct vdiv2D_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 3838 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3839 match(Set dst (DivVD src1 src2));
kvn@4001 3840 format %{ "vdivpd $dst,$src1,$src2\t! div packed2D" %}
kvn@4001 3841 ins_encode %{
kvn@4001 3842 bool vector256 = false;
kvn@4001 3843 __ vdivpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3844 %}
kvn@4001 3845 ins_pipe( pipe_slow );
kvn@4001 3846 %}
kvn@4001 3847
kvn@4001 3848 instruct vdiv2D_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 3849 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3850 match(Set dst (DivVD src (LoadVector mem)));
kvn@4001 3851 format %{ "vdivpd $dst,$src,$mem\t! div packed2D" %}
kvn@4001 3852 ins_encode %{
kvn@4001 3853 bool vector256 = false;
kvn@4001 3854 __ vdivpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3855 %}
kvn@4001 3856 ins_pipe( pipe_slow );
kvn@4001 3857 %}
kvn@4001 3858
kvn@4001 3859 instruct vdiv4D_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 3860 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3861 match(Set dst (DivVD src1 src2));
kvn@4001 3862 format %{ "vdivpd $dst,$src1,$src2\t! div packed4D" %}
kvn@4001 3863 ins_encode %{
kvn@4001 3864 bool vector256 = true;
kvn@4001 3865 __ vdivpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 3866 %}
kvn@4001 3867 ins_pipe( pipe_slow );
kvn@4001 3868 %}
kvn@4001 3869
kvn@4001 3870 instruct vdiv4D_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 3871 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3872 match(Set dst (DivVD src (LoadVector mem)));
kvn@4001 3873 format %{ "vdivpd $dst,$src,$mem\t! div packed4D" %}
kvn@4001 3874 ins_encode %{
kvn@4001 3875 bool vector256 = true;
kvn@4001 3876 __ vdivpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 3877 %}
kvn@4001 3878 ins_pipe( pipe_slow );
kvn@4001 3879 %}
kvn@4001 3880
kvn@4134 3881 // ------------------------------ Shift ---------------------------------------
kvn@4134 3882
kvn@4134 3883 // Left and right shift count vectors are the same on x86
kvn@4134 3884 // (only lowest bits of xmm reg are used for count).
kvn@4134 3885 instruct vshiftcnt(vecS dst, rRegI cnt) %{
kvn@4134 3886 match(Set dst (LShiftCntV cnt));
kvn@4134 3887 match(Set dst (RShiftCntV cnt));
kvn@4134 3888 format %{ "movd $dst,$cnt\t! load shift count" %}
kvn@4134 3889 ins_encode %{
kvn@4134 3890 __ movdl($dst$$XMMRegister, $cnt$$Register);
kvn@4134 3891 %}
kvn@4134 3892 ins_pipe( pipe_slow );
kvn@4134 3893 %}
kvn@4134 3894
kvn@4001 3895 // ------------------------------ LeftShift -----------------------------------
kvn@4001 3896
kvn@4001 3897 // Shorts/Chars vector left shift
kvn@4134 3898 instruct vsll2S(vecS dst, vecS shift) %{
kvn@4001 3899 predicate(n->as_Vector()->length() == 2);
kvn@4001 3900 match(Set dst (LShiftVS dst shift));
kvn@4001 3901 format %{ "psllw $dst,$shift\t! left shift packed2S" %}
kvn@4001 3902 ins_encode %{
kvn@4001 3903 __ psllw($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 3904 %}
kvn@4001 3905 ins_pipe( pipe_slow );
kvn@4001 3906 %}
kvn@4001 3907
kvn@4001 3908 instruct vsll2S_imm(vecS dst, immI8 shift) %{
kvn@4001 3909 predicate(n->as_Vector()->length() == 2);
kvn@4001 3910 match(Set dst (LShiftVS dst shift));
kvn@4001 3911 format %{ "psllw $dst,$shift\t! left shift packed2S" %}
kvn@4001 3912 ins_encode %{
kvn@4001 3913 __ psllw($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 3914 %}
kvn@4001 3915 ins_pipe( pipe_slow );
kvn@4001 3916 %}
kvn@4001 3917
kvn@4134 3918 instruct vsll2S_reg(vecS dst, vecS src, vecS shift) %{
kvn@4001 3919 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3920 match(Set dst (LShiftVS src shift));
kvn@4001 3921 format %{ "vpsllw $dst,$src,$shift\t! left shift packed2S" %}
kvn@4001 3922 ins_encode %{
kvn@4001 3923 bool vector256 = false;
kvn@4001 3924 __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 3925 %}
kvn@4001 3926 ins_pipe( pipe_slow );
kvn@4001 3927 %}
kvn@4001 3928
kvn@4001 3929 instruct vsll2S_reg_imm(vecS dst, vecS src, immI8 shift) %{
kvn@4001 3930 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 3931 match(Set dst (LShiftVS src shift));
kvn@4001 3932 format %{ "vpsllw $dst,$src,$shift\t! left shift packed2S" %}
kvn@4001 3933 ins_encode %{
kvn@4001 3934 bool vector256 = false;
kvn@4001 3935 __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 3936 %}
kvn@4001 3937 ins_pipe( pipe_slow );
kvn@4001 3938 %}
kvn@4001 3939
kvn@4134 3940 instruct vsll4S(vecD dst, vecS shift) %{
kvn@4001 3941 predicate(n->as_Vector()->length() == 4);
kvn@4001 3942 match(Set dst (LShiftVS dst shift));
kvn@4001 3943 format %{ "psllw $dst,$shift\t! left shift packed4S" %}
kvn@4001 3944 ins_encode %{
kvn@4001 3945 __ psllw($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 3946 %}
kvn@4001 3947 ins_pipe( pipe_slow );
kvn@4001 3948 %}
kvn@4001 3949
kvn@4001 3950 instruct vsll4S_imm(vecD dst, immI8 shift) %{
kvn@4001 3951 predicate(n->as_Vector()->length() == 4);
kvn@4001 3952 match(Set dst (LShiftVS dst shift));
kvn@4001 3953 format %{ "psllw $dst,$shift\t! left shift packed4S" %}
kvn@4001 3954 ins_encode %{
kvn@4001 3955 __ psllw($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 3956 %}
kvn@4001 3957 ins_pipe( pipe_slow );
kvn@4001 3958 %}
kvn@4001 3959
kvn@4134 3960 instruct vsll4S_reg(vecD dst, vecD src, vecS shift) %{
kvn@4001 3961 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3962 match(Set dst (LShiftVS src shift));
kvn@4001 3963 format %{ "vpsllw $dst,$src,$shift\t! left shift packed4S" %}
kvn@4001 3964 ins_encode %{
kvn@4001 3965 bool vector256 = false;
kvn@4001 3966 __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 3967 %}
kvn@4001 3968 ins_pipe( pipe_slow );
kvn@4001 3969 %}
kvn@4001 3970
kvn@4001 3971 instruct vsll4S_reg_imm(vecD dst, vecD src, immI8 shift) %{
kvn@4001 3972 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 3973 match(Set dst (LShiftVS src shift));
kvn@4001 3974 format %{ "vpsllw $dst,$src,$shift\t! left shift packed4S" %}
kvn@4001 3975 ins_encode %{
kvn@4001 3976 bool vector256 = false;
kvn@4001 3977 __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 3978 %}
kvn@4001 3979 ins_pipe( pipe_slow );
kvn@4001 3980 %}
kvn@4001 3981
kvn@4134 3982 instruct vsll8S(vecX dst, vecS shift) %{
kvn@4001 3983 predicate(n->as_Vector()->length() == 8);
kvn@4001 3984 match(Set dst (LShiftVS dst shift));
kvn@4001 3985 format %{ "psllw $dst,$shift\t! left shift packed8S" %}
kvn@4001 3986 ins_encode %{
kvn@4001 3987 __ psllw($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 3988 %}
kvn@4001 3989 ins_pipe( pipe_slow );
kvn@4001 3990 %}
kvn@4001 3991
kvn@4001 3992 instruct vsll8S_imm(vecX dst, immI8 shift) %{
kvn@4001 3993 predicate(n->as_Vector()->length() == 8);
kvn@4001 3994 match(Set dst (LShiftVS dst shift));
kvn@4001 3995 format %{ "psllw $dst,$shift\t! left shift packed8S" %}
kvn@4001 3996 ins_encode %{
kvn@4001 3997 __ psllw($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 3998 %}
kvn@4001 3999 ins_pipe( pipe_slow );
kvn@4001 4000 %}
kvn@4001 4001
kvn@4134 4002 instruct vsll8S_reg(vecX dst, vecX src, vecS shift) %{
kvn@4001 4003 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 4004 match(Set dst (LShiftVS src shift));
kvn@4001 4005 format %{ "vpsllw $dst,$src,$shift\t! left shift packed8S" %}
kvn@4001 4006 ins_encode %{
kvn@4001 4007 bool vector256 = false;
kvn@4001 4008 __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4009 %}
kvn@4001 4010 ins_pipe( pipe_slow );
kvn@4001 4011 %}
kvn@4001 4012
kvn@4001 4013 instruct vsll8S_reg_imm(vecX dst, vecX src, immI8 shift) %{
kvn@4001 4014 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 4015 match(Set dst (LShiftVS src shift));
kvn@4001 4016 format %{ "vpsllw $dst,$src,$shift\t! left shift packed8S" %}
kvn@4001 4017 ins_encode %{
kvn@4001 4018 bool vector256 = false;
kvn@4001 4019 __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4020 %}
kvn@4001 4021 ins_pipe( pipe_slow );
kvn@4001 4022 %}
kvn@4001 4023
kvn@4134 4024 instruct vsll16S_reg(vecY dst, vecY src, vecS shift) %{
kvn@4001 4025 predicate(UseAVX > 1 && n->as_Vector()->length() == 16);
kvn@4001 4026 match(Set dst (LShiftVS src shift));
kvn@4001 4027 format %{ "vpsllw $dst,$src,$shift\t! left shift packed16S" %}
kvn@4001 4028 ins_encode %{
kvn@4001 4029 bool vector256 = true;
kvn@4001 4030 __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4031 %}
kvn@4001 4032 ins_pipe( pipe_slow );
kvn@4001 4033 %}
kvn@4001 4034
kvn@4001 4035 instruct vsll16S_reg_imm(vecY dst, vecY src, immI8 shift) %{
kvn@4001 4036 predicate(UseAVX > 1 && n->as_Vector()->length() == 16);
kvn@4001 4037 match(Set dst (LShiftVS src shift));
kvn@4001 4038 format %{ "vpsllw $dst,$src,$shift\t! left shift packed16S" %}
kvn@4001 4039 ins_encode %{
kvn@4001 4040 bool vector256 = true;
kvn@4001 4041 __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4042 %}
kvn@4001 4043 ins_pipe( pipe_slow );
kvn@4001 4044 %}
kvn@4001 4045
kvn@4001 4046 // Integers vector left shift
kvn@4134 4047 instruct vsll2I(vecD dst, vecS shift) %{
kvn@4001 4048 predicate(n->as_Vector()->length() == 2);
kvn@4001 4049 match(Set dst (LShiftVI dst shift));
kvn@4001 4050 format %{ "pslld $dst,$shift\t! left shift packed2I" %}
kvn@4001 4051 ins_encode %{
kvn@4001 4052 __ pslld($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 4053 %}
kvn@4001 4054 ins_pipe( pipe_slow );
kvn@4001 4055 %}
kvn@4001 4056
kvn@4001 4057 instruct vsll2I_imm(vecD dst, immI8 shift) %{
kvn@4001 4058 predicate(n->as_Vector()->length() == 2);
kvn@4001 4059 match(Set dst (LShiftVI dst shift));
kvn@4001 4060 format %{ "pslld $dst,$shift\t! left shift packed2I" %}
kvn@4001 4061 ins_encode %{
kvn@4001 4062 __ pslld($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 4063 %}
kvn@4001 4064 ins_pipe( pipe_slow );
kvn@4001 4065 %}
kvn@4001 4066
kvn@4134 4067 instruct vsll2I_reg(vecD dst, vecD src, vecS shift) %{
kvn@4001 4068 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 4069 match(Set dst (LShiftVI src shift));
kvn@4001 4070 format %{ "vpslld $dst,$src,$shift\t! left shift packed2I" %}
kvn@4001 4071 ins_encode %{
kvn@4001 4072 bool vector256 = false;
kvn@4001 4073 __ vpslld($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4074 %}
kvn@4001 4075 ins_pipe( pipe_slow );
kvn@4001 4076 %}
kvn@4001 4077
kvn@4001 4078 instruct vsll2I_reg_imm(vecD dst, vecD src, immI8 shift) %{
kvn@4001 4079 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 4080 match(Set dst (LShiftVI src shift));
kvn@4001 4081 format %{ "vpslld $dst,$src,$shift\t! left shift packed2I" %}
kvn@4001 4082 ins_encode %{
kvn@4001 4083 bool vector256 = false;
kvn@4001 4084 __ vpslld($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4085 %}
kvn@4001 4086 ins_pipe( pipe_slow );
kvn@4001 4087 %}
kvn@4001 4088
kvn@4134 4089 instruct vsll4I(vecX dst, vecS shift) %{
kvn@4001 4090 predicate(n->as_Vector()->length() == 4);
kvn@4001 4091 match(Set dst (LShiftVI dst shift));
kvn@4001 4092 format %{ "pslld $dst,$shift\t! left shift packed4I" %}
kvn@4001 4093 ins_encode %{
kvn@4001 4094 __ pslld($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 4095 %}
kvn@4001 4096 ins_pipe( pipe_slow );
kvn@4001 4097 %}
kvn@4001 4098
kvn@4001 4099 instruct vsll4I_imm(vecX dst, immI8 shift) %{
kvn@4001 4100 predicate(n->as_Vector()->length() == 4);
kvn@4001 4101 match(Set dst (LShiftVI dst shift));
kvn@4001 4102 format %{ "pslld $dst,$shift\t! left shift packed4I" %}
kvn@4001 4103 ins_encode %{
kvn@4001 4104 __ pslld($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 4105 %}
kvn@4001 4106 ins_pipe( pipe_slow );
kvn@4001 4107 %}
kvn@4001 4108
kvn@4134 4109 instruct vsll4I_reg(vecX dst, vecX src, vecS shift) %{
kvn@4001 4110 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 4111 match(Set dst (LShiftVI src shift));
kvn@4001 4112 format %{ "vpslld $dst,$src,$shift\t! left shift packed4I" %}
kvn@4001 4113 ins_encode %{
kvn@4001 4114 bool vector256 = false;
kvn@4001 4115 __ vpslld($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4116 %}
kvn@4001 4117 ins_pipe( pipe_slow );
kvn@4001 4118 %}
kvn@4001 4119
kvn@4001 4120 instruct vsll4I_reg_imm(vecX dst, vecX src, immI8 shift) %{
kvn@4001 4121 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 4122 match(Set dst (LShiftVI src shift));
kvn@4001 4123 format %{ "vpslld $dst,$src,$shift\t! left shift packed4I" %}
kvn@4001 4124 ins_encode %{
kvn@4001 4125 bool vector256 = false;
kvn@4001 4126 __ vpslld($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4127 %}
kvn@4001 4128 ins_pipe( pipe_slow );
kvn@4001 4129 %}
kvn@4001 4130
kvn@4134 4131 instruct vsll8I_reg(vecY dst, vecY src, vecS shift) %{
kvn@4001 4132 predicate(UseAVX > 1 && n->as_Vector()->length() == 8);
kvn@4001 4133 match(Set dst (LShiftVI src shift));
kvn@4001 4134 format %{ "vpslld $dst,$src,$shift\t! left shift packed8I" %}
kvn@4001 4135 ins_encode %{
kvn@4001 4136 bool vector256 = true;
kvn@4001 4137 __ vpslld($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4138 %}
kvn@4001 4139 ins_pipe( pipe_slow );
kvn@4001 4140 %}
kvn@4001 4141
kvn@4001 4142 instruct vsll8I_reg_imm(vecY dst, vecY src, immI8 shift) %{
kvn@4001 4143 predicate(UseAVX > 1 && n->as_Vector()->length() == 8);
kvn@4001 4144 match(Set dst (LShiftVI src shift));
kvn@4001 4145 format %{ "vpslld $dst,$src,$shift\t! left shift packed8I" %}
kvn@4001 4146 ins_encode %{
kvn@4001 4147 bool vector256 = true;
kvn@4001 4148 __ vpslld($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4149 %}
kvn@4001 4150 ins_pipe( pipe_slow );
kvn@4001 4151 %}
kvn@4001 4152
kvn@4001 4153 // Longs vector left shift
kvn@4134 4154 instruct vsll2L(vecX dst, vecS shift) %{
kvn@4001 4155 predicate(n->as_Vector()->length() == 2);
kvn@4001 4156 match(Set dst (LShiftVL dst shift));
kvn@4001 4157 format %{ "psllq $dst,$shift\t! left shift packed2L" %}
kvn@4001 4158 ins_encode %{
kvn@4001 4159 __ psllq($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 4160 %}
kvn@4001 4161 ins_pipe( pipe_slow );
kvn@4001 4162 %}
kvn@4001 4163
kvn@4001 4164 instruct vsll2L_imm(vecX dst, immI8 shift) %{
kvn@4001 4165 predicate(n->as_Vector()->length() == 2);
kvn@4001 4166 match(Set dst (LShiftVL dst shift));
kvn@4001 4167 format %{ "psllq $dst,$shift\t! left shift packed2L" %}
kvn@4001 4168 ins_encode %{
kvn@4001 4169 __ psllq($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 4170 %}
kvn@4001 4171 ins_pipe( pipe_slow );
kvn@4001 4172 %}
kvn@4001 4173
kvn@4134 4174 instruct vsll2L_reg(vecX dst, vecX src, vecS shift) %{
kvn@4001 4175 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 4176 match(Set dst (LShiftVL src shift));
kvn@4001 4177 format %{ "vpsllq $dst,$src,$shift\t! left shift packed2L" %}
kvn@4001 4178 ins_encode %{
kvn@4001 4179 bool vector256 = false;
kvn@4001 4180 __ vpsllq($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4181 %}
kvn@4001 4182 ins_pipe( pipe_slow );
kvn@4001 4183 %}
kvn@4001 4184
kvn@4001 4185 instruct vsll2L_reg_imm(vecX dst, vecX src, immI8 shift) %{
kvn@4001 4186 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 4187 match(Set dst (LShiftVL src shift));
kvn@4001 4188 format %{ "vpsllq $dst,$src,$shift\t! left shift packed2L" %}
kvn@4001 4189 ins_encode %{
kvn@4001 4190 bool vector256 = false;
kvn@4001 4191 __ vpsllq($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4192 %}
kvn@4001 4193 ins_pipe( pipe_slow );
kvn@4001 4194 %}
kvn@4001 4195
kvn@4134 4196 instruct vsll4L_reg(vecY dst, vecY src, vecS shift) %{
kvn@4001 4197 predicate(UseAVX > 1 && n->as_Vector()->length() == 4);
kvn@4001 4198 match(Set dst (LShiftVL src shift));
kvn@4001 4199 format %{ "vpsllq $dst,$src,$shift\t! left shift packed4L" %}
kvn@4001 4200 ins_encode %{
kvn@4001 4201 bool vector256 = true;
kvn@4001 4202 __ vpsllq($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4203 %}
kvn@4001 4204 ins_pipe( pipe_slow );
kvn@4001 4205 %}
kvn@4001 4206
kvn@4001 4207 instruct vsll4L_reg_imm(vecY dst, vecY src, immI8 shift) %{
kvn@4001 4208 predicate(UseAVX > 1 && n->as_Vector()->length() == 4);
kvn@4001 4209 match(Set dst (LShiftVL src shift));
kvn@4001 4210 format %{ "vpsllq $dst,$src,$shift\t! left shift packed4L" %}
kvn@4001 4211 ins_encode %{
kvn@4001 4212 bool vector256 = true;
kvn@4001 4213 __ vpsllq($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4214 %}
kvn@4001 4215 ins_pipe( pipe_slow );
kvn@4001 4216 %}
kvn@4001 4217
kvn@4001 4218 // ----------------------- LogicalRightShift -----------------------------------
kvn@4001 4219
kvn@4204 4220 // Shorts vector logical right shift produces incorrect Java result
kvn@4001 4221 // for negative data because java code convert short value into int with
kvn@4204 4222 // sign extension before a shift. But char vectors are fine since chars are
kvn@4204 4223 // unsigned values.
kvn@4204 4224
kvn@4204 4225 instruct vsrl2S(vecS dst, vecS shift) %{
kvn@4204 4226 predicate(n->as_Vector()->length() == 2);
kvn@4204 4227 match(Set dst (URShiftVS dst shift));
kvn@4204 4228 format %{ "psrlw $dst,$shift\t! logical right shift packed2S" %}
kvn@4204 4229 ins_encode %{
kvn@4204 4230 __ psrlw($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4204 4231 %}
kvn@4204 4232 ins_pipe( pipe_slow );
kvn@4204 4233 %}
kvn@4204 4234
kvn@4204 4235 instruct vsrl2S_imm(vecS dst, immI8 shift) %{
kvn@4204 4236 predicate(n->as_Vector()->length() == 2);
kvn@4204 4237 match(Set dst (URShiftVS dst shift));
kvn@4204 4238 format %{ "psrlw $dst,$shift\t! logical right shift packed2S" %}
kvn@4204 4239 ins_encode %{
kvn@4204 4240 __ psrlw($dst$$XMMRegister, (int)$shift$$constant);
kvn@4204 4241 %}
kvn@4204 4242 ins_pipe( pipe_slow );
kvn@4204 4243 %}
kvn@4204 4244
kvn@4204 4245 instruct vsrl2S_reg(vecS dst, vecS src, vecS shift) %{
kvn@4204 4246 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4204 4247 match(Set dst (URShiftVS src shift));
kvn@4204 4248 format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed2S" %}
kvn@4204 4249 ins_encode %{
kvn@4204 4250 bool vector256 = false;
kvn@4204 4251 __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4204 4252 %}
kvn@4204 4253 ins_pipe( pipe_slow );
kvn@4204 4254 %}
kvn@4204 4255
kvn@4204 4256 instruct vsrl2S_reg_imm(vecS dst, vecS src, immI8 shift) %{
kvn@4204 4257 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4204 4258 match(Set dst (URShiftVS src shift));
kvn@4204 4259 format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed2S" %}
kvn@4204 4260 ins_encode %{
kvn@4204 4261 bool vector256 = false;
kvn@4204 4262 __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4204 4263 %}
kvn@4204 4264 ins_pipe( pipe_slow );
kvn@4204 4265 %}
kvn@4204 4266
kvn@4204 4267 instruct vsrl4S(vecD dst, vecS shift) %{
kvn@4204 4268 predicate(n->as_Vector()->length() == 4);
kvn@4204 4269 match(Set dst (URShiftVS dst shift));
kvn@4204 4270 format %{ "psrlw $dst,$shift\t! logical right shift packed4S" %}
kvn@4204 4271 ins_encode %{
kvn@4204 4272 __ psrlw($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4204 4273 %}
kvn@4204 4274 ins_pipe( pipe_slow );
kvn@4204 4275 %}
kvn@4204 4276
kvn@4204 4277 instruct vsrl4S_imm(vecD dst, immI8 shift) %{
kvn@4204 4278 predicate(n->as_Vector()->length() == 4);
kvn@4204 4279 match(Set dst (URShiftVS dst shift));
kvn@4204 4280 format %{ "psrlw $dst,$shift\t! logical right shift packed4S" %}
kvn@4204 4281 ins_encode %{
kvn@4204 4282 __ psrlw($dst$$XMMRegister, (int)$shift$$constant);
kvn@4204 4283 %}
kvn@4204 4284 ins_pipe( pipe_slow );
kvn@4204 4285 %}
kvn@4204 4286
kvn@4204 4287 instruct vsrl4S_reg(vecD dst, vecD src, vecS shift) %{
kvn@4204 4288 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4204 4289 match(Set dst (URShiftVS src shift));
kvn@4204 4290 format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed4S" %}
kvn@4204 4291 ins_encode %{
kvn@4204 4292 bool vector256 = false;
kvn@4204 4293 __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4204 4294 %}
kvn@4204 4295 ins_pipe( pipe_slow );
kvn@4204 4296 %}
kvn@4204 4297
kvn@4204 4298 instruct vsrl4S_reg_imm(vecD dst, vecD src, immI8 shift) %{
kvn@4204 4299 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4204 4300 match(Set dst (URShiftVS src shift));
kvn@4204 4301 format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed4S" %}
kvn@4204 4302 ins_encode %{
kvn@4204 4303 bool vector256 = false;
kvn@4204 4304 __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4204 4305 %}
kvn@4204 4306 ins_pipe( pipe_slow );
kvn@4204 4307 %}
kvn@4204 4308
kvn@4204 4309 instruct vsrl8S(vecX dst, vecS shift) %{
kvn@4204 4310 predicate(n->as_Vector()->length() == 8);
kvn@4204 4311 match(Set dst (URShiftVS dst shift));
kvn@4204 4312 format %{ "psrlw $dst,$shift\t! logical right shift packed8S" %}
kvn@4204 4313 ins_encode %{
kvn@4204 4314 __ psrlw($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4204 4315 %}
kvn@4204 4316 ins_pipe( pipe_slow );
kvn@4204 4317 %}
kvn@4204 4318
kvn@4204 4319 instruct vsrl8S_imm(vecX dst, immI8 shift) %{
kvn@4204 4320 predicate(n->as_Vector()->length() == 8);
kvn@4204 4321 match(Set dst (URShiftVS dst shift));
kvn@4204 4322 format %{ "psrlw $dst,$shift\t! logical right shift packed8S" %}
kvn@4204 4323 ins_encode %{
kvn@4204 4324 __ psrlw($dst$$XMMRegister, (int)$shift$$constant);
kvn@4204 4325 %}
kvn@4204 4326 ins_pipe( pipe_slow );
kvn@4204 4327 %}
kvn@4204 4328
kvn@4204 4329 instruct vsrl8S_reg(vecX dst, vecX src, vecS shift) %{
kvn@4204 4330 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4204 4331 match(Set dst (URShiftVS src shift));
kvn@4204 4332 format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed8S" %}
kvn@4204 4333 ins_encode %{
kvn@4204 4334 bool vector256 = false;
kvn@4204 4335 __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4204 4336 %}
kvn@4204 4337 ins_pipe( pipe_slow );
kvn@4204 4338 %}
kvn@4204 4339
kvn@4204 4340 instruct vsrl8S_reg_imm(vecX dst, vecX src, immI8 shift) %{
kvn@4204 4341 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4204 4342 match(Set dst (URShiftVS src shift));
kvn@4204 4343 format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed8S" %}
kvn@4204 4344 ins_encode %{
kvn@4204 4345 bool vector256 = false;
kvn@4204 4346 __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4204 4347 %}
kvn@4204 4348 ins_pipe( pipe_slow );
kvn@4204 4349 %}
kvn@4204 4350
kvn@4204 4351 instruct vsrl16S_reg(vecY dst, vecY src, vecS shift) %{
kvn@4204 4352 predicate(UseAVX > 1 && n->as_Vector()->length() == 16);
kvn@4204 4353 match(Set dst (URShiftVS src shift));
kvn@4204 4354 format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed16S" %}
kvn@4204 4355 ins_encode %{
kvn@4204 4356 bool vector256 = true;
kvn@4204 4357 __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4204 4358 %}
kvn@4204 4359 ins_pipe( pipe_slow );
kvn@4204 4360 %}
kvn@4204 4361
kvn@4204 4362 instruct vsrl16S_reg_imm(vecY dst, vecY src, immI8 shift) %{
kvn@4204 4363 predicate(UseAVX > 1 && n->as_Vector()->length() == 16);
kvn@4204 4364 match(Set dst (URShiftVS src shift));
kvn@4204 4365 format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed16S" %}
kvn@4204 4366 ins_encode %{
kvn@4204 4367 bool vector256 = true;
kvn@4204 4368 __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4204 4369 %}
kvn@4204 4370 ins_pipe( pipe_slow );
kvn@4204 4371 %}
kvn@4001 4372
kvn@4001 4373 // Integers vector logical right shift
kvn@4134 4374 instruct vsrl2I(vecD dst, vecS shift) %{
kvn@4001 4375 predicate(n->as_Vector()->length() == 2);
kvn@4001 4376 match(Set dst (URShiftVI dst shift));
kvn@4001 4377 format %{ "psrld $dst,$shift\t! logical right shift packed2I" %}
kvn@4001 4378 ins_encode %{
kvn@4001 4379 __ psrld($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 4380 %}
kvn@4001 4381 ins_pipe( pipe_slow );
kvn@4001 4382 %}
kvn@4001 4383
kvn@4001 4384 instruct vsrl2I_imm(vecD dst, immI8 shift) %{
kvn@4001 4385 predicate(n->as_Vector()->length() == 2);
kvn@4001 4386 match(Set dst (URShiftVI dst shift));
kvn@4001 4387 format %{ "psrld $dst,$shift\t! logical right shift packed2I" %}
kvn@4001 4388 ins_encode %{
kvn@4001 4389 __ psrld($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 4390 %}
kvn@4001 4391 ins_pipe( pipe_slow );
kvn@4001 4392 %}
kvn@4001 4393
kvn@4134 4394 instruct vsrl2I_reg(vecD dst, vecD src, vecS shift) %{
kvn@4001 4395 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 4396 match(Set dst (URShiftVI src shift));
kvn@4001 4397 format %{ "vpsrld $dst,$src,$shift\t! logical right shift packed2I" %}
kvn@4001 4398 ins_encode %{
kvn@4001 4399 bool vector256 = false;
kvn@4001 4400 __ vpsrld($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4401 %}
kvn@4001 4402 ins_pipe( pipe_slow );
kvn@4001 4403 %}
kvn@4001 4404
kvn@4001 4405 instruct vsrl2I_reg_imm(vecD dst, vecD src, immI8 shift) %{
kvn@4001 4406 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 4407 match(Set dst (URShiftVI src shift));
kvn@4001 4408 format %{ "vpsrld $dst,$src,$shift\t! logical right shift packed2I" %}
kvn@4001 4409 ins_encode %{
kvn@4001 4410 bool vector256 = false;
kvn@4001 4411 __ vpsrld($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4412 %}
kvn@4001 4413 ins_pipe( pipe_slow );
kvn@4001 4414 %}
kvn@4001 4415
kvn@4134 4416 instruct vsrl4I(vecX dst, vecS shift) %{
kvn@4001 4417 predicate(n->as_Vector()->length() == 4);
kvn@4001 4418 match(Set dst (URShiftVI dst shift));
kvn@4001 4419 format %{ "psrld $dst,$shift\t! logical right shift packed4I" %}
kvn@4001 4420 ins_encode %{
kvn@4001 4421 __ psrld($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 4422 %}
kvn@4001 4423 ins_pipe( pipe_slow );
kvn@4001 4424 %}
kvn@4001 4425
kvn@4001 4426 instruct vsrl4I_imm(vecX dst, immI8 shift) %{
kvn@4001 4427 predicate(n->as_Vector()->length() == 4);
kvn@4001 4428 match(Set dst (URShiftVI dst shift));
kvn@4001 4429 format %{ "psrld $dst,$shift\t! logical right shift packed4I" %}
kvn@4001 4430 ins_encode %{
kvn@4001 4431 __ psrld($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 4432 %}
kvn@4001 4433 ins_pipe( pipe_slow );
kvn@4001 4434 %}
kvn@4001 4435
kvn@4134 4436 instruct vsrl4I_reg(vecX dst, vecX src, vecS shift) %{
kvn@4001 4437 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 4438 match(Set dst (URShiftVI src shift));
kvn@4001 4439 format %{ "vpsrld $dst,$src,$shift\t! logical right shift packed4I" %}
kvn@4001 4440 ins_encode %{
kvn@4001 4441 bool vector256 = false;
kvn@4001 4442 __ vpsrld($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4443 %}
kvn@4001 4444 ins_pipe( pipe_slow );
kvn@4001 4445 %}
kvn@4001 4446
kvn@4001 4447 instruct vsrl4I_reg_imm(vecX dst, vecX src, immI8 shift) %{
kvn@4001 4448 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 4449 match(Set dst (URShiftVI src shift));
kvn@4001 4450 format %{ "vpsrld $dst,$src,$shift\t! logical right shift packed4I" %}
kvn@4001 4451 ins_encode %{
kvn@4001 4452 bool vector256 = false;
kvn@4001 4453 __ vpsrld($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4454 %}
kvn@4001 4455 ins_pipe( pipe_slow );
kvn@4001 4456 %}
kvn@4001 4457
kvn@4134 4458 instruct vsrl8I_reg(vecY dst, vecY src, vecS shift) %{
kvn@4001 4459 predicate(UseAVX > 1 && n->as_Vector()->length() == 8);
kvn@4001 4460 match(Set dst (URShiftVI src shift));
kvn@4001 4461 format %{ "vpsrld $dst,$src,$shift\t! logical right shift packed8I" %}
kvn@4001 4462 ins_encode %{
kvn@4001 4463 bool vector256 = true;
kvn@4001 4464 __ vpsrld($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4465 %}
kvn@4001 4466 ins_pipe( pipe_slow );
kvn@4001 4467 %}
kvn@4001 4468
kvn@4001 4469 instruct vsrl8I_reg_imm(vecY dst, vecY src, immI8 shift) %{
kvn@4001 4470 predicate(UseAVX > 1 && n->as_Vector()->length() == 8);
kvn@4001 4471 match(Set dst (URShiftVI src shift));
kvn@4001 4472 format %{ "vpsrld $dst,$src,$shift\t! logical right shift packed8I" %}
kvn@4001 4473 ins_encode %{
kvn@4001 4474 bool vector256 = true;
kvn@4001 4475 __ vpsrld($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4476 %}
kvn@4001 4477 ins_pipe( pipe_slow );
kvn@4001 4478 %}
kvn@4001 4479
kvn@4001 4480 // Longs vector logical right shift
kvn@4134 4481 instruct vsrl2L(vecX dst, vecS shift) %{
kvn@4001 4482 predicate(n->as_Vector()->length() == 2);
kvn@4001 4483 match(Set dst (URShiftVL dst shift));
kvn@4001 4484 format %{ "psrlq $dst,$shift\t! logical right shift packed2L" %}
kvn@4001 4485 ins_encode %{
kvn@4001 4486 __ psrlq($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 4487 %}
kvn@4001 4488 ins_pipe( pipe_slow );
kvn@4001 4489 %}
kvn@4001 4490
kvn@4001 4491 instruct vsrl2L_imm(vecX dst, immI8 shift) %{
kvn@4001 4492 predicate(n->as_Vector()->length() == 2);
kvn@4001 4493 match(Set dst (URShiftVL dst shift));
kvn@4001 4494 format %{ "psrlq $dst,$shift\t! logical right shift packed2L" %}
kvn@4001 4495 ins_encode %{
kvn@4001 4496 __ psrlq($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 4497 %}
kvn@4001 4498 ins_pipe( pipe_slow );
kvn@4001 4499 %}
kvn@4001 4500
kvn@4134 4501 instruct vsrl2L_reg(vecX dst, vecX src, vecS shift) %{
kvn@4001 4502 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 4503 match(Set dst (URShiftVL src shift));
kvn@4001 4504 format %{ "vpsrlq $dst,$src,$shift\t! logical right shift packed2L" %}
kvn@4001 4505 ins_encode %{
kvn@4001 4506 bool vector256 = false;
kvn@4001 4507 __ vpsrlq($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4508 %}
kvn@4001 4509 ins_pipe( pipe_slow );
kvn@4001 4510 %}
kvn@4001 4511
kvn@4001 4512 instruct vsrl2L_reg_imm(vecX dst, vecX src, immI8 shift) %{
kvn@4001 4513 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 4514 match(Set dst (URShiftVL src shift));
kvn@4001 4515 format %{ "vpsrlq $dst,$src,$shift\t! logical right shift packed2L" %}
kvn@4001 4516 ins_encode %{
kvn@4001 4517 bool vector256 = false;
kvn@4001 4518 __ vpsrlq($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4519 %}
kvn@4001 4520 ins_pipe( pipe_slow );
kvn@4001 4521 %}
kvn@4001 4522
kvn@4134 4523 instruct vsrl4L_reg(vecY dst, vecY src, vecS shift) %{
kvn@4001 4524 predicate(UseAVX > 1 && n->as_Vector()->length() == 4);
kvn@4001 4525 match(Set dst (URShiftVL src shift));
kvn@4001 4526 format %{ "vpsrlq $dst,$src,$shift\t! logical right shift packed4L" %}
kvn@4001 4527 ins_encode %{
kvn@4001 4528 bool vector256 = true;
kvn@4001 4529 __ vpsrlq($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4530 %}
kvn@4001 4531 ins_pipe( pipe_slow );
kvn@4001 4532 %}
kvn@4001 4533
kvn@4001 4534 instruct vsrl4L_reg_imm(vecY dst, vecY src, immI8 shift) %{
kvn@4001 4535 predicate(UseAVX > 1 && n->as_Vector()->length() == 4);
kvn@4001 4536 match(Set dst (URShiftVL src shift));
kvn@4001 4537 format %{ "vpsrlq $dst,$src,$shift\t! logical right shift packed4L" %}
kvn@4001 4538 ins_encode %{
kvn@4001 4539 bool vector256 = true;
kvn@4001 4540 __ vpsrlq($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4541 %}
kvn@4001 4542 ins_pipe( pipe_slow );
kvn@4001 4543 %}
kvn@4001 4544
kvn@4001 4545 // ------------------- ArithmeticRightShift -----------------------------------
kvn@4001 4546
kvn@4001 4547 // Shorts/Chars vector arithmetic right shift
kvn@4134 4548 instruct vsra2S(vecS dst, vecS shift) %{
kvn@4001 4549 predicate(n->as_Vector()->length() == 2);
kvn@4001 4550 match(Set dst (RShiftVS dst shift));
kvn@4001 4551 format %{ "psraw $dst,$shift\t! arithmetic right shift packed2S" %}
kvn@4001 4552 ins_encode %{
kvn@4001 4553 __ psraw($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 4554 %}
kvn@4001 4555 ins_pipe( pipe_slow );
kvn@4001 4556 %}
kvn@4001 4557
kvn@4001 4558 instruct vsra2S_imm(vecS dst, immI8 shift) %{
kvn@4001 4559 predicate(n->as_Vector()->length() == 2);
kvn@4001 4560 match(Set dst (RShiftVS dst shift));
kvn@4001 4561 format %{ "psraw $dst,$shift\t! arithmetic right shift packed2S" %}
kvn@4001 4562 ins_encode %{
kvn@4001 4563 __ psraw($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 4564 %}
kvn@4001 4565 ins_pipe( pipe_slow );
kvn@4001 4566 %}
kvn@4001 4567
kvn@4134 4568 instruct vsra2S_reg(vecS dst, vecS src, vecS shift) %{
kvn@4001 4569 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 4570 match(Set dst (RShiftVS src shift));
kvn@4001 4571 format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed2S" %}
kvn@4001 4572 ins_encode %{
kvn@4001 4573 bool vector256 = false;
kvn@4001 4574 __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4575 %}
kvn@4001 4576 ins_pipe( pipe_slow );
kvn@4001 4577 %}
kvn@4001 4578
kvn@4001 4579 instruct vsra2S_reg_imm(vecS dst, vecS src, immI8 shift) %{
kvn@4001 4580 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 4581 match(Set dst (RShiftVS src shift));
kvn@4001 4582 format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed2S" %}
kvn@4001 4583 ins_encode %{
kvn@4001 4584 bool vector256 = false;
kvn@4001 4585 __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4586 %}
kvn@4001 4587 ins_pipe( pipe_slow );
kvn@4001 4588 %}
kvn@4001 4589
kvn@4134 4590 instruct vsra4S(vecD dst, vecS shift) %{
kvn@4001 4591 predicate(n->as_Vector()->length() == 4);
kvn@4001 4592 match(Set dst (RShiftVS dst shift));
kvn@4001 4593 format %{ "psraw $dst,$shift\t! arithmetic right shift packed4S" %}
kvn@4001 4594 ins_encode %{
kvn@4001 4595 __ psraw($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 4596 %}
kvn@4001 4597 ins_pipe( pipe_slow );
kvn@4001 4598 %}
kvn@4001 4599
kvn@4001 4600 instruct vsra4S_imm(vecD dst, immI8 shift) %{
kvn@4001 4601 predicate(n->as_Vector()->length() == 4);
kvn@4001 4602 match(Set dst (RShiftVS dst shift));
kvn@4001 4603 format %{ "psraw $dst,$shift\t! arithmetic right shift packed4S" %}
kvn@4001 4604 ins_encode %{
kvn@4001 4605 __ psraw($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 4606 %}
kvn@4001 4607 ins_pipe( pipe_slow );
kvn@4001 4608 %}
kvn@4001 4609
kvn@4134 4610 instruct vsra4S_reg(vecD dst, vecD src, vecS shift) %{
kvn@4001 4611 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 4612 match(Set dst (RShiftVS src shift));
kvn@4001 4613 format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed4S" %}
kvn@4001 4614 ins_encode %{
kvn@4001 4615 bool vector256 = false;
kvn@4001 4616 __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4617 %}
kvn@4001 4618 ins_pipe( pipe_slow );
kvn@4001 4619 %}
kvn@4001 4620
kvn@4001 4621 instruct vsra4S_reg_imm(vecD dst, vecD src, immI8 shift) %{
kvn@4001 4622 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 4623 match(Set dst (RShiftVS src shift));
kvn@4001 4624 format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed4S" %}
kvn@4001 4625 ins_encode %{
kvn@4001 4626 bool vector256 = false;
kvn@4001 4627 __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4628 %}
kvn@4001 4629 ins_pipe( pipe_slow );
kvn@4001 4630 %}
kvn@4001 4631
kvn@4134 4632 instruct vsra8S(vecX dst, vecS shift) %{
kvn@4001 4633 predicate(n->as_Vector()->length() == 8);
kvn@4001 4634 match(Set dst (RShiftVS dst shift));
kvn@4001 4635 format %{ "psraw $dst,$shift\t! arithmetic right shift packed8S" %}
kvn@4001 4636 ins_encode %{
kvn@4001 4637 __ psraw($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 4638 %}
kvn@4001 4639 ins_pipe( pipe_slow );
kvn@4001 4640 %}
kvn@4001 4641
kvn@4001 4642 instruct vsra8S_imm(vecX dst, immI8 shift) %{
kvn@4001 4643 predicate(n->as_Vector()->length() == 8);
kvn@4001 4644 match(Set dst (RShiftVS dst shift));
kvn@4001 4645 format %{ "psraw $dst,$shift\t! arithmetic right shift packed8S" %}
kvn@4001 4646 ins_encode %{
kvn@4001 4647 __ psraw($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 4648 %}
kvn@4001 4649 ins_pipe( pipe_slow );
kvn@4001 4650 %}
kvn@4001 4651
kvn@4134 4652 instruct vsra8S_reg(vecX dst, vecX src, vecS shift) %{
kvn@4001 4653 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 4654 match(Set dst (RShiftVS src shift));
kvn@4001 4655 format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed8S" %}
kvn@4001 4656 ins_encode %{
kvn@4001 4657 bool vector256 = false;
kvn@4001 4658 __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4659 %}
kvn@4001 4660 ins_pipe( pipe_slow );
kvn@4001 4661 %}
kvn@4001 4662
kvn@4001 4663 instruct vsra8S_reg_imm(vecX dst, vecX src, immI8 shift) %{
kvn@4001 4664 predicate(UseAVX > 0 && n->as_Vector()->length() == 8);
kvn@4001 4665 match(Set dst (RShiftVS src shift));
kvn@4001 4666 format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed8S" %}
kvn@4001 4667 ins_encode %{
kvn@4001 4668 bool vector256 = false;
kvn@4001 4669 __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4670 %}
kvn@4001 4671 ins_pipe( pipe_slow );
kvn@4001 4672 %}
kvn@4001 4673
kvn@4134 4674 instruct vsra16S_reg(vecY dst, vecY src, vecS shift) %{
kvn@4001 4675 predicate(UseAVX > 1 && n->as_Vector()->length() == 16);
kvn@4001 4676 match(Set dst (RShiftVS src shift));
kvn@4001 4677 format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed16S" %}
kvn@4001 4678 ins_encode %{
kvn@4001 4679 bool vector256 = true;
kvn@4001 4680 __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4681 %}
kvn@4001 4682 ins_pipe( pipe_slow );
kvn@4001 4683 %}
kvn@4001 4684
kvn@4001 4685 instruct vsra16S_reg_imm(vecY dst, vecY src, immI8 shift) %{
kvn@4001 4686 predicate(UseAVX > 1 && n->as_Vector()->length() == 16);
kvn@4001 4687 match(Set dst (RShiftVS src shift));
kvn@4001 4688 format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed16S" %}
kvn@4001 4689 ins_encode %{
kvn@4001 4690 bool vector256 = true;
kvn@4001 4691 __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4692 %}
kvn@4001 4693 ins_pipe( pipe_slow );
kvn@4001 4694 %}
kvn@4001 4695
kvn@4001 4696 // Integers vector arithmetic right shift
kvn@4134 4697 instruct vsra2I(vecD dst, vecS shift) %{
kvn@4001 4698 predicate(n->as_Vector()->length() == 2);
kvn@4001 4699 match(Set dst (RShiftVI dst shift));
kvn@4001 4700 format %{ "psrad $dst,$shift\t! arithmetic right shift packed2I" %}
kvn@4001 4701 ins_encode %{
kvn@4001 4702 __ psrad($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 4703 %}
kvn@4001 4704 ins_pipe( pipe_slow );
kvn@4001 4705 %}
kvn@4001 4706
kvn@4001 4707 instruct vsra2I_imm(vecD dst, immI8 shift) %{
kvn@4001 4708 predicate(n->as_Vector()->length() == 2);
kvn@4001 4709 match(Set dst (RShiftVI dst shift));
kvn@4001 4710 format %{ "psrad $dst,$shift\t! arithmetic right shift packed2I" %}
kvn@4001 4711 ins_encode %{
kvn@4001 4712 __ psrad($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 4713 %}
kvn@4001 4714 ins_pipe( pipe_slow );
kvn@4001 4715 %}
kvn@4001 4716
kvn@4134 4717 instruct vsra2I_reg(vecD dst, vecD src, vecS shift) %{
kvn@4001 4718 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 4719 match(Set dst (RShiftVI src shift));
kvn@4001 4720 format %{ "vpsrad $dst,$src,$shift\t! arithmetic right shift packed2I" %}
kvn@4001 4721 ins_encode %{
kvn@4001 4722 bool vector256 = false;
kvn@4001 4723 __ vpsrad($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4724 %}
kvn@4001 4725 ins_pipe( pipe_slow );
kvn@4001 4726 %}
kvn@4001 4727
kvn@4001 4728 instruct vsra2I_reg_imm(vecD dst, vecD src, immI8 shift) %{
kvn@4001 4729 predicate(UseAVX > 0 && n->as_Vector()->length() == 2);
kvn@4001 4730 match(Set dst (RShiftVI src shift));
kvn@4001 4731 format %{ "vpsrad $dst,$src,$shift\t! arithmetic right shift packed2I" %}
kvn@4001 4732 ins_encode %{
kvn@4001 4733 bool vector256 = false;
kvn@4001 4734 __ vpsrad($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4735 %}
kvn@4001 4736 ins_pipe( pipe_slow );
kvn@4001 4737 %}
kvn@4001 4738
kvn@4134 4739 instruct vsra4I(vecX dst, vecS shift) %{
kvn@4001 4740 predicate(n->as_Vector()->length() == 4);
kvn@4001 4741 match(Set dst (RShiftVI dst shift));
kvn@4001 4742 format %{ "psrad $dst,$shift\t! arithmetic right shift packed4I" %}
kvn@4001 4743 ins_encode %{
kvn@4001 4744 __ psrad($dst$$XMMRegister, $shift$$XMMRegister);
kvn@4001 4745 %}
kvn@4001 4746 ins_pipe( pipe_slow );
kvn@4001 4747 %}
kvn@4001 4748
kvn@4001 4749 instruct vsra4I_imm(vecX dst, immI8 shift) %{
kvn@4001 4750 predicate(n->as_Vector()->length() == 4);
kvn@4001 4751 match(Set dst (RShiftVI dst shift));
kvn@4001 4752 format %{ "psrad $dst,$shift\t! arithmetic right shift packed4I" %}
kvn@4001 4753 ins_encode %{
kvn@4001 4754 __ psrad($dst$$XMMRegister, (int)$shift$$constant);
kvn@4001 4755 %}
kvn@4001 4756 ins_pipe( pipe_slow );
kvn@4001 4757 %}
kvn@4001 4758
kvn@4134 4759 instruct vsra4I_reg(vecX dst, vecX src, vecS shift) %{
kvn@4001 4760 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 4761 match(Set dst (RShiftVI src shift));
kvn@4001 4762 format %{ "vpsrad $dst,$src,$shift\t! arithmetic right shift packed4I" %}
kvn@4001 4763 ins_encode %{
kvn@4001 4764 bool vector256 = false;
kvn@4001 4765 __ vpsrad($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4766 %}
kvn@4001 4767 ins_pipe( pipe_slow );
kvn@4001 4768 %}
kvn@4001 4769
kvn@4001 4770 instruct vsra4I_reg_imm(vecX dst, vecX src, immI8 shift) %{
kvn@4001 4771 predicate(UseAVX > 0 && n->as_Vector()->length() == 4);
kvn@4001 4772 match(Set dst (RShiftVI src shift));
kvn@4001 4773 format %{ "vpsrad $dst,$src,$shift\t! arithmetic right shift packed4I" %}
kvn@4001 4774 ins_encode %{
kvn@4001 4775 bool vector256 = false;
kvn@4001 4776 __ vpsrad($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4777 %}
kvn@4001 4778 ins_pipe( pipe_slow );
kvn@4001 4779 %}
kvn@4001 4780
kvn@4134 4781 instruct vsra8I_reg(vecY dst, vecY src, vecS shift) %{
kvn@4001 4782 predicate(UseAVX > 1 && n->as_Vector()->length() == 8);
kvn@4001 4783 match(Set dst (RShiftVI src shift));
kvn@4001 4784 format %{ "vpsrad $dst,$src,$shift\t! arithmetic right shift packed8I" %}
kvn@4001 4785 ins_encode %{
kvn@4001 4786 bool vector256 = true;
kvn@4001 4787 __ vpsrad($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256);
kvn@4001 4788 %}
kvn@4001 4789 ins_pipe( pipe_slow );
kvn@4001 4790 %}
kvn@4001 4791
kvn@4001 4792 instruct vsra8I_reg_imm(vecY dst, vecY src, immI8 shift) %{
kvn@4001 4793 predicate(UseAVX > 1 && n->as_Vector()->length() == 8);
kvn@4001 4794 match(Set dst (RShiftVI src shift));
kvn@4001 4795 format %{ "vpsrad $dst,$src,$shift\t! arithmetic right shift packed8I" %}
kvn@4001 4796 ins_encode %{
kvn@4001 4797 bool vector256 = true;
kvn@4001 4798 __ vpsrad($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256);
kvn@4001 4799 %}
kvn@4001 4800 ins_pipe( pipe_slow );
kvn@4001 4801 %}
kvn@4001 4802
kvn@4001 4803 // There are no longs vector arithmetic right shift instructions.
kvn@4001 4804
kvn@4001 4805
kvn@4001 4806 // --------------------------------- AND --------------------------------------
kvn@4001 4807
kvn@4001 4808 instruct vand4B(vecS dst, vecS src) %{
kvn@4001 4809 predicate(n->as_Vector()->length_in_bytes() == 4);
kvn@4001 4810 match(Set dst (AndV dst src));
kvn@4001 4811 format %{ "pand $dst,$src\t! and vectors (4 bytes)" %}
kvn@4001 4812 ins_encode %{
kvn@4001 4813 __ pand($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 4814 %}
kvn@4001 4815 ins_pipe( pipe_slow );
kvn@4001 4816 %}
kvn@4001 4817
kvn@4001 4818 instruct vand4B_reg(vecS dst, vecS src1, vecS src2) %{
kvn@4001 4819 predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 4);
kvn@4001 4820 match(Set dst (AndV src1 src2));
kvn@4001 4821 format %{ "vpand $dst,$src1,$src2\t! and vectors (4 bytes)" %}
kvn@4001 4822 ins_encode %{
kvn@4001 4823 bool vector256 = false;
kvn@4001 4824 __ vpand($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 4825 %}
kvn@4001 4826 ins_pipe( pipe_slow );
kvn@4001 4827 %}
kvn@4001 4828
kvn@4001 4829 instruct vand8B(vecD dst, vecD src) %{
kvn@4001 4830 predicate(n->as_Vector()->length_in_bytes() == 8);
kvn@4001 4831 match(Set dst (AndV dst src));
kvn@4001 4832 format %{ "pand $dst,$src\t! and vectors (8 bytes)" %}
kvn@4001 4833 ins_encode %{
kvn@4001 4834 __ pand($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 4835 %}
kvn@4001 4836 ins_pipe( pipe_slow );
kvn@4001 4837 %}
kvn@4001 4838
kvn@4001 4839 instruct vand8B_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 4840 predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 8);
kvn@4001 4841 match(Set dst (AndV src1 src2));
kvn@4001 4842 format %{ "vpand $dst,$src1,$src2\t! and vectors (8 bytes)" %}
kvn@4001 4843 ins_encode %{
kvn@4001 4844 bool vector256 = false;
kvn@4001 4845 __ vpand($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 4846 %}
kvn@4001 4847 ins_pipe( pipe_slow );
kvn@4001 4848 %}
kvn@4001 4849
kvn@4001 4850 instruct vand16B(vecX dst, vecX src) %{
kvn@4001 4851 predicate(n->as_Vector()->length_in_bytes() == 16);
kvn@4001 4852 match(Set dst (AndV dst src));
kvn@4001 4853 format %{ "pand $dst,$src\t! and vectors (16 bytes)" %}
kvn@4001 4854 ins_encode %{
kvn@4001 4855 __ pand($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 4856 %}
kvn@4001 4857 ins_pipe( pipe_slow );
kvn@4001 4858 %}
kvn@4001 4859
kvn@4001 4860 instruct vand16B_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 4861 predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 16);
kvn@4001 4862 match(Set dst (AndV src1 src2));
kvn@4001 4863 format %{ "vpand $dst,$src1,$src2\t! and vectors (16 bytes)" %}
kvn@4001 4864 ins_encode %{
kvn@4001 4865 bool vector256 = false;
kvn@4001 4866 __ vpand($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 4867 %}
kvn@4001 4868 ins_pipe( pipe_slow );
kvn@4001 4869 %}
kvn@4001 4870
kvn@4001 4871 instruct vand16B_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 4872 predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 16);
kvn@4001 4873 match(Set dst (AndV src (LoadVector mem)));
kvn@4001 4874 format %{ "vpand $dst,$src,$mem\t! and vectors (16 bytes)" %}
kvn@4001 4875 ins_encode %{
kvn@4001 4876 bool vector256 = false;
kvn@4001 4877 __ vpand($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 4878 %}
kvn@4001 4879 ins_pipe( pipe_slow );
kvn@4001 4880 %}
kvn@4001 4881
kvn@4001 4882 instruct vand32B_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 4883 predicate(UseAVX > 1 && n->as_Vector()->length_in_bytes() == 32);
kvn@4001 4884 match(Set dst (AndV src1 src2));
kvn@4001 4885 format %{ "vpand $dst,$src1,$src2\t! and vectors (32 bytes)" %}
kvn@4001 4886 ins_encode %{
kvn@4001 4887 bool vector256 = true;
kvn@4001 4888 __ vpand($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 4889 %}
kvn@4001 4890 ins_pipe( pipe_slow );
kvn@4001 4891 %}
kvn@4001 4892
kvn@4001 4893 instruct vand32B_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 4894 predicate(UseAVX > 1 && n->as_Vector()->length_in_bytes() == 32);
kvn@4001 4895 match(Set dst (AndV src (LoadVector mem)));
kvn@4001 4896 format %{ "vpand $dst,$src,$mem\t! and vectors (32 bytes)" %}
kvn@4001 4897 ins_encode %{
kvn@4001 4898 bool vector256 = true;
kvn@4001 4899 __ vpand($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 4900 %}
kvn@4001 4901 ins_pipe( pipe_slow );
kvn@4001 4902 %}
kvn@4001 4903
kvn@4001 4904 // --------------------------------- OR ---------------------------------------
kvn@4001 4905
kvn@4001 4906 instruct vor4B(vecS dst, vecS src) %{
kvn@4001 4907 predicate(n->as_Vector()->length_in_bytes() == 4);
kvn@4001 4908 match(Set dst (OrV dst src));
kvn@4001 4909 format %{ "por $dst,$src\t! or vectors (4 bytes)" %}
kvn@4001 4910 ins_encode %{
kvn@4001 4911 __ por($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 4912 %}
kvn@4001 4913 ins_pipe( pipe_slow );
kvn@4001 4914 %}
kvn@4001 4915
kvn@4001 4916 instruct vor4B_reg(vecS dst, vecS src1, vecS src2) %{
kvn@4001 4917 predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 4);
kvn@4001 4918 match(Set dst (OrV src1 src2));
kvn@4001 4919 format %{ "vpor $dst,$src1,$src2\t! or vectors (4 bytes)" %}
kvn@4001 4920 ins_encode %{
kvn@4001 4921 bool vector256 = false;
kvn@4001 4922 __ vpor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 4923 %}
kvn@4001 4924 ins_pipe( pipe_slow );
kvn@4001 4925 %}
kvn@4001 4926
kvn@4001 4927 instruct vor8B(vecD dst, vecD src) %{
kvn@4001 4928 predicate(n->as_Vector()->length_in_bytes() == 8);
kvn@4001 4929 match(Set dst (OrV dst src));
kvn@4001 4930 format %{ "por $dst,$src\t! or vectors (8 bytes)" %}
kvn@4001 4931 ins_encode %{
kvn@4001 4932 __ por($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 4933 %}
kvn@4001 4934 ins_pipe( pipe_slow );
kvn@4001 4935 %}
kvn@4001 4936
kvn@4001 4937 instruct vor8B_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 4938 predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 8);
kvn@4001 4939 match(Set dst (OrV src1 src2));
kvn@4001 4940 format %{ "vpor $dst,$src1,$src2\t! or vectors (8 bytes)" %}
kvn@4001 4941 ins_encode %{
kvn@4001 4942 bool vector256 = false;
kvn@4001 4943 __ vpor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 4944 %}
kvn@4001 4945 ins_pipe( pipe_slow );
kvn@4001 4946 %}
kvn@4001 4947
kvn@4001 4948 instruct vor16B(vecX dst, vecX src) %{
kvn@4001 4949 predicate(n->as_Vector()->length_in_bytes() == 16);
kvn@4001 4950 match(Set dst (OrV dst src));
kvn@4001 4951 format %{ "por $dst,$src\t! or vectors (16 bytes)" %}
kvn@4001 4952 ins_encode %{
kvn@4001 4953 __ por($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 4954 %}
kvn@4001 4955 ins_pipe( pipe_slow );
kvn@4001 4956 %}
kvn@4001 4957
kvn@4001 4958 instruct vor16B_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 4959 predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 16);
kvn@4001 4960 match(Set dst (OrV src1 src2));
kvn@4001 4961 format %{ "vpor $dst,$src1,$src2\t! or vectors (16 bytes)" %}
kvn@4001 4962 ins_encode %{
kvn@4001 4963 bool vector256 = false;
kvn@4001 4964 __ vpor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 4965 %}
kvn@4001 4966 ins_pipe( pipe_slow );
kvn@4001 4967 %}
kvn@4001 4968
kvn@4001 4969 instruct vor16B_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 4970 predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 16);
kvn@4001 4971 match(Set dst (OrV src (LoadVector mem)));
kvn@4001 4972 format %{ "vpor $dst,$src,$mem\t! or vectors (16 bytes)" %}
kvn@4001 4973 ins_encode %{
kvn@4001 4974 bool vector256 = false;
kvn@4001 4975 __ vpor($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 4976 %}
kvn@4001 4977 ins_pipe( pipe_slow );
kvn@4001 4978 %}
kvn@4001 4979
kvn@4001 4980 instruct vor32B_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 4981 predicate(UseAVX > 1 && n->as_Vector()->length_in_bytes() == 32);
kvn@4001 4982 match(Set dst (OrV src1 src2));
kvn@4001 4983 format %{ "vpor $dst,$src1,$src2\t! or vectors (32 bytes)" %}
kvn@4001 4984 ins_encode %{
kvn@4001 4985 bool vector256 = true;
kvn@4001 4986 __ vpor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 4987 %}
kvn@4001 4988 ins_pipe( pipe_slow );
kvn@4001 4989 %}
kvn@4001 4990
kvn@4001 4991 instruct vor32B_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 4992 predicate(UseAVX > 1 && n->as_Vector()->length_in_bytes() == 32);
kvn@4001 4993 match(Set dst (OrV src (LoadVector mem)));
kvn@4001 4994 format %{ "vpor $dst,$src,$mem\t! or vectors (32 bytes)" %}
kvn@4001 4995 ins_encode %{
kvn@4001 4996 bool vector256 = true;
kvn@4001 4997 __ vpor($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 4998 %}
kvn@4001 4999 ins_pipe( pipe_slow );
kvn@4001 5000 %}
kvn@4001 5001
kvn@4001 5002 // --------------------------------- XOR --------------------------------------
kvn@4001 5003
kvn@4001 5004 instruct vxor4B(vecS dst, vecS src) %{
kvn@4001 5005 predicate(n->as_Vector()->length_in_bytes() == 4);
kvn@4001 5006 match(Set dst (XorV dst src));
kvn@4001 5007 format %{ "pxor $dst,$src\t! xor vectors (4 bytes)" %}
kvn@4001 5008 ins_encode %{
kvn@4001 5009 __ pxor($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 5010 %}
kvn@4001 5011 ins_pipe( pipe_slow );
kvn@4001 5012 %}
kvn@4001 5013
kvn@4001 5014 instruct vxor4B_reg(vecS dst, vecS src1, vecS src2) %{
kvn@4001 5015 predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 4);
kvn@4001 5016 match(Set dst (XorV src1 src2));
kvn@4001 5017 format %{ "vpxor $dst,$src1,$src2\t! xor vectors (4 bytes)" %}
kvn@4001 5018 ins_encode %{
kvn@4001 5019 bool vector256 = false;
kvn@4001 5020 __ vpxor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 5021 %}
kvn@4001 5022 ins_pipe( pipe_slow );
kvn@4001 5023 %}
kvn@4001 5024
kvn@4001 5025 instruct vxor8B(vecD dst, vecD src) %{
kvn@4001 5026 predicate(n->as_Vector()->length_in_bytes() == 8);
kvn@4001 5027 match(Set dst (XorV dst src));
kvn@4001 5028 format %{ "pxor $dst,$src\t! xor vectors (8 bytes)" %}
kvn@4001 5029 ins_encode %{
kvn@4001 5030 __ pxor($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 5031 %}
kvn@4001 5032 ins_pipe( pipe_slow );
kvn@4001 5033 %}
kvn@4001 5034
kvn@4001 5035 instruct vxor8B_reg(vecD dst, vecD src1, vecD src2) %{
kvn@4001 5036 predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 8);
kvn@4001 5037 match(Set dst (XorV src1 src2));
kvn@4001 5038 format %{ "vpxor $dst,$src1,$src2\t! xor vectors (8 bytes)" %}
kvn@4001 5039 ins_encode %{
kvn@4001 5040 bool vector256 = false;
kvn@4001 5041 __ vpxor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 5042 %}
kvn@4001 5043 ins_pipe( pipe_slow );
kvn@4001 5044 %}
kvn@4001 5045
kvn@4001 5046 instruct vxor16B(vecX dst, vecX src) %{
kvn@4001 5047 predicate(n->as_Vector()->length_in_bytes() == 16);
kvn@4001 5048 match(Set dst (XorV dst src));
kvn@4001 5049 format %{ "pxor $dst,$src\t! xor vectors (16 bytes)" %}
kvn@4001 5050 ins_encode %{
kvn@4001 5051 __ pxor($dst$$XMMRegister, $src$$XMMRegister);
kvn@4001 5052 %}
kvn@4001 5053 ins_pipe( pipe_slow );
kvn@4001 5054 %}
kvn@4001 5055
kvn@4001 5056 instruct vxor16B_reg(vecX dst, vecX src1, vecX src2) %{
kvn@4001 5057 predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 16);
kvn@4001 5058 match(Set dst (XorV src1 src2));
kvn@4001 5059 format %{ "vpxor $dst,$src1,$src2\t! xor vectors (16 bytes)" %}
kvn@4001 5060 ins_encode %{
kvn@4001 5061 bool vector256 = false;
kvn@4001 5062 __ vpxor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 5063 %}
kvn@4001 5064 ins_pipe( pipe_slow );
kvn@4001 5065 %}
kvn@4001 5066
kvn@4001 5067 instruct vxor16B_mem(vecX dst, vecX src, memory mem) %{
kvn@4001 5068 predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 16);
kvn@4001 5069 match(Set dst (XorV src (LoadVector mem)));
kvn@4001 5070 format %{ "vpxor $dst,$src,$mem\t! xor vectors (16 bytes)" %}
kvn@4001 5071 ins_encode %{
kvn@4001 5072 bool vector256 = false;
kvn@4001 5073 __ vpxor($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 5074 %}
kvn@4001 5075 ins_pipe( pipe_slow );
kvn@4001 5076 %}
kvn@4001 5077
kvn@4001 5078 instruct vxor32B_reg(vecY dst, vecY src1, vecY src2) %{
kvn@4001 5079 predicate(UseAVX > 1 && n->as_Vector()->length_in_bytes() == 32);
kvn@4001 5080 match(Set dst (XorV src1 src2));
kvn@4001 5081 format %{ "vpxor $dst,$src1,$src2\t! xor vectors (32 bytes)" %}
kvn@4001 5082 ins_encode %{
kvn@4001 5083 bool vector256 = true;
kvn@4001 5084 __ vpxor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256);
kvn@4001 5085 %}
kvn@4001 5086 ins_pipe( pipe_slow );
kvn@4001 5087 %}
kvn@4001 5088
kvn@4001 5089 instruct vxor32B_mem(vecY dst, vecY src, memory mem) %{
kvn@4001 5090 predicate(UseAVX > 1 && n->as_Vector()->length_in_bytes() == 32);
kvn@4001 5091 match(Set dst (XorV src (LoadVector mem)));
kvn@4001 5092 format %{ "vpxor $dst,$src,$mem\t! xor vectors (32 bytes)" %}
kvn@4001 5093 ins_encode %{
kvn@4001 5094 bool vector256 = true;
kvn@4001 5095 __ vpxor($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256);
kvn@4001 5096 %}
kvn@4001 5097 ins_pipe( pipe_slow );
kvn@4001 5098 %}
kvn@4001 5099

mercurial