Thu, 21 Apr 2016 16:19:33 +0300
8130309: Need to bailout cleanly if creation of stubs fails when codecache is out of space
Summary: Check for failed expansion of stub section in code buffer and bailout.
Reviewed-by: kvn, thartmann
kvn@3390 | 1 | // |
kvn@3577 | 2 | // Copyright (c) 2011, 2012, Oracle and/or its affiliates. All rights reserved. |
kvn@3390 | 3 | // DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
kvn@3390 | 4 | // |
kvn@3390 | 5 | // This code is free software; you can redistribute it and/or modify it |
kvn@3390 | 6 | // under the terms of the GNU General Public License version 2 only, as |
kvn@3390 | 7 | // published by the Free Software Foundation. |
kvn@3390 | 8 | // |
kvn@3390 | 9 | // This code is distributed in the hope that it will be useful, but WITHOUT |
kvn@3390 | 10 | // ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
kvn@3390 | 11 | // FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
kvn@3390 | 12 | // version 2 for more details (a copy is included in the LICENSE file that |
kvn@3390 | 13 | // accompanied this code). |
kvn@3390 | 14 | // |
kvn@3390 | 15 | // You should have received a copy of the GNU General Public License version |
kvn@3390 | 16 | // 2 along with this work; if not, write to the Free Software Foundation, |
kvn@3390 | 17 | // Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
kvn@3390 | 18 | // |
kvn@3390 | 19 | // Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
kvn@3390 | 20 | // or visit www.oracle.com if you need additional information or have any |
kvn@3390 | 21 | // questions. |
kvn@3390 | 22 | // |
kvn@3390 | 23 | // |
kvn@3390 | 24 | |
kvn@3390 | 25 | // X86 Common Architecture Description File |
kvn@3390 | 26 | |
kvn@3882 | 27 | //----------REGISTER DEFINITION BLOCK------------------------------------------ |
kvn@3882 | 28 | // This information is used by the matcher and the register allocator to |
kvn@3882 | 29 | // describe individual registers and classes of registers within the target |
kvn@3882 | 30 | // archtecture. |
kvn@3882 | 31 | |
kvn@3882 | 32 | register %{ |
kvn@3882 | 33 | //----------Architecture Description Register Definitions---------------------- |
kvn@3882 | 34 | // General Registers |
kvn@3882 | 35 | // "reg_def" name ( register save type, C convention save type, |
kvn@3882 | 36 | // ideal register type, encoding ); |
kvn@3882 | 37 | // Register Save Types: |
kvn@3882 | 38 | // |
kvn@3882 | 39 | // NS = No-Save: The register allocator assumes that these registers |
kvn@3882 | 40 | // can be used without saving upon entry to the method, & |
kvn@3882 | 41 | // that they do not need to be saved at call sites. |
kvn@3882 | 42 | // |
kvn@3882 | 43 | // SOC = Save-On-Call: The register allocator assumes that these registers |
kvn@3882 | 44 | // can be used without saving upon entry to the method, |
kvn@3882 | 45 | // but that they must be saved at call sites. |
kvn@3882 | 46 | // |
kvn@3882 | 47 | // SOE = Save-On-Entry: The register allocator assumes that these registers |
kvn@3882 | 48 | // must be saved before using them upon entry to the |
kvn@3882 | 49 | // method, but they do not need to be saved at call |
kvn@3882 | 50 | // sites. |
kvn@3882 | 51 | // |
kvn@3882 | 52 | // AS = Always-Save: The register allocator assumes that these registers |
kvn@3882 | 53 | // must be saved before using them upon entry to the |
kvn@3882 | 54 | // method, & that they must be saved at call sites. |
kvn@3882 | 55 | // |
kvn@3882 | 56 | // Ideal Register Type is used to determine how to save & restore a |
kvn@3882 | 57 | // register. Op_RegI will get spilled with LoadI/StoreI, Op_RegP will get |
kvn@3882 | 58 | // spilled with LoadP/StoreP. If the register supports both, use Op_RegI. |
kvn@3882 | 59 | // |
kvn@3882 | 60 | // The encoding number is the actual bit-pattern placed into the opcodes. |
kvn@3882 | 61 | |
kvn@3882 | 62 | // XMM registers. 256-bit registers or 8 words each, labeled (a)-h. |
kvn@3882 | 63 | // Word a in each register holds a Float, words ab hold a Double. |
kvn@3882 | 64 | // The whole registers are used in SSE4.2 version intrinsics, |
kvn@3882 | 65 | // array copy stubs and superword operations (see UseSSE42Intrinsics, |
kvn@3882 | 66 | // UseXMMForArrayCopy and UseSuperword flags). |
kvn@3882 | 67 | // XMM8-XMM15 must be encoded with REX (VEX for UseAVX). |
kvn@3882 | 68 | // Linux ABI: No register preserved across function calls |
kvn@3882 | 69 | // XMM0-XMM7 might hold parameters |
kvn@3882 | 70 | // Windows ABI: XMM6-XMM15 preserved across function calls |
kvn@3882 | 71 | // XMM0-XMM3 might hold parameters |
kvn@3882 | 72 | |
kvn@3882 | 73 | reg_def XMM0 ( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()); |
kvn@3929 | 74 | reg_def XMM0b( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(1)); |
kvn@3929 | 75 | reg_def XMM0c( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(2)); |
kvn@3929 | 76 | reg_def XMM0d( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(3)); |
kvn@3929 | 77 | reg_def XMM0e( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(4)); |
kvn@3929 | 78 | reg_def XMM0f( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(5)); |
kvn@3929 | 79 | reg_def XMM0g( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(6)); |
kvn@3929 | 80 | reg_def XMM0h( SOC, SOC, Op_RegF, 0, xmm0->as_VMReg()->next(7)); |
kvn@3882 | 81 | |
kvn@3882 | 82 | reg_def XMM1 ( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()); |
kvn@3929 | 83 | reg_def XMM1b( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(1)); |
kvn@3929 | 84 | reg_def XMM1c( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(2)); |
kvn@3929 | 85 | reg_def XMM1d( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(3)); |
kvn@3929 | 86 | reg_def XMM1e( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(4)); |
kvn@3929 | 87 | reg_def XMM1f( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(5)); |
kvn@3929 | 88 | reg_def XMM1g( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(6)); |
kvn@3929 | 89 | reg_def XMM1h( SOC, SOC, Op_RegF, 1, xmm1->as_VMReg()->next(7)); |
kvn@3882 | 90 | |
kvn@3882 | 91 | reg_def XMM2 ( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()); |
kvn@3929 | 92 | reg_def XMM2b( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(1)); |
kvn@3929 | 93 | reg_def XMM2c( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(2)); |
kvn@3929 | 94 | reg_def XMM2d( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(3)); |
kvn@3929 | 95 | reg_def XMM2e( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(4)); |
kvn@3929 | 96 | reg_def XMM2f( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(5)); |
kvn@3929 | 97 | reg_def XMM2g( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(6)); |
kvn@3929 | 98 | reg_def XMM2h( SOC, SOC, Op_RegF, 2, xmm2->as_VMReg()->next(7)); |
kvn@3882 | 99 | |
kvn@3882 | 100 | reg_def XMM3 ( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()); |
kvn@3929 | 101 | reg_def XMM3b( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(1)); |
kvn@3929 | 102 | reg_def XMM3c( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(2)); |
kvn@3929 | 103 | reg_def XMM3d( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(3)); |
kvn@3929 | 104 | reg_def XMM3e( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(4)); |
kvn@3929 | 105 | reg_def XMM3f( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(5)); |
kvn@3929 | 106 | reg_def XMM3g( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(6)); |
kvn@3929 | 107 | reg_def XMM3h( SOC, SOC, Op_RegF, 3, xmm3->as_VMReg()->next(7)); |
kvn@3882 | 108 | |
kvn@3882 | 109 | reg_def XMM4 ( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()); |
kvn@3929 | 110 | reg_def XMM4b( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(1)); |
kvn@3929 | 111 | reg_def XMM4c( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(2)); |
kvn@3929 | 112 | reg_def XMM4d( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(3)); |
kvn@3929 | 113 | reg_def XMM4e( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(4)); |
kvn@3929 | 114 | reg_def XMM4f( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(5)); |
kvn@3929 | 115 | reg_def XMM4g( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(6)); |
kvn@3929 | 116 | reg_def XMM4h( SOC, SOC, Op_RegF, 4, xmm4->as_VMReg()->next(7)); |
kvn@3882 | 117 | |
kvn@3882 | 118 | reg_def XMM5 ( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()); |
kvn@3929 | 119 | reg_def XMM5b( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(1)); |
kvn@3929 | 120 | reg_def XMM5c( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(2)); |
kvn@3929 | 121 | reg_def XMM5d( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(3)); |
kvn@3929 | 122 | reg_def XMM5e( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(4)); |
kvn@3929 | 123 | reg_def XMM5f( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(5)); |
kvn@3929 | 124 | reg_def XMM5g( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(6)); |
kvn@3929 | 125 | reg_def XMM5h( SOC, SOC, Op_RegF, 5, xmm5->as_VMReg()->next(7)); |
kvn@3882 | 126 | |
kvn@3882 | 127 | #ifdef _WIN64 |
kvn@3882 | 128 | |
kvn@3882 | 129 | reg_def XMM6 ( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()); |
kvn@3929 | 130 | reg_def XMM6b( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(1)); |
kvn@3929 | 131 | reg_def XMM6c( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(2)); |
kvn@3929 | 132 | reg_def XMM6d( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(3)); |
kvn@3929 | 133 | reg_def XMM6e( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(4)); |
kvn@3929 | 134 | reg_def XMM6f( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(5)); |
kvn@3929 | 135 | reg_def XMM6g( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(6)); |
kvn@3929 | 136 | reg_def XMM6h( SOC, SOE, Op_RegF, 6, xmm6->as_VMReg()->next(7)); |
kvn@3882 | 137 | |
kvn@3882 | 138 | reg_def XMM7 ( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()); |
kvn@3929 | 139 | reg_def XMM7b( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(1)); |
kvn@3929 | 140 | reg_def XMM7c( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(2)); |
kvn@3929 | 141 | reg_def XMM7d( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(3)); |
kvn@3929 | 142 | reg_def XMM7e( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(4)); |
kvn@3929 | 143 | reg_def XMM7f( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(5)); |
kvn@3929 | 144 | reg_def XMM7g( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(6)); |
kvn@3929 | 145 | reg_def XMM7h( SOC, SOE, Op_RegF, 7, xmm7->as_VMReg()->next(7)); |
kvn@3882 | 146 | |
kvn@3882 | 147 | reg_def XMM8 ( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()); |
kvn@3929 | 148 | reg_def XMM8b( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(1)); |
kvn@3929 | 149 | reg_def XMM8c( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(2)); |
kvn@3929 | 150 | reg_def XMM8d( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(3)); |
kvn@3929 | 151 | reg_def XMM8e( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(4)); |
kvn@3929 | 152 | reg_def XMM8f( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(5)); |
kvn@3929 | 153 | reg_def XMM8g( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(6)); |
kvn@3929 | 154 | reg_def XMM8h( SOC, SOE, Op_RegF, 8, xmm8->as_VMReg()->next(7)); |
kvn@3882 | 155 | |
kvn@3882 | 156 | reg_def XMM9 ( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()); |
kvn@3929 | 157 | reg_def XMM9b( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(1)); |
kvn@3929 | 158 | reg_def XMM9c( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(2)); |
kvn@3929 | 159 | reg_def XMM9d( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(3)); |
kvn@3929 | 160 | reg_def XMM9e( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(4)); |
kvn@3929 | 161 | reg_def XMM9f( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(5)); |
kvn@3929 | 162 | reg_def XMM9g( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(6)); |
kvn@3929 | 163 | reg_def XMM9h( SOC, SOE, Op_RegF, 9, xmm9->as_VMReg()->next(7)); |
kvn@3882 | 164 | |
kvn@3882 | 165 | reg_def XMM10 ( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()); |
kvn@3929 | 166 | reg_def XMM10b( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(1)); |
kvn@3929 | 167 | reg_def XMM10c( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(2)); |
kvn@3929 | 168 | reg_def XMM10d( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(3)); |
kvn@3929 | 169 | reg_def XMM10e( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(4)); |
kvn@3929 | 170 | reg_def XMM10f( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(5)); |
kvn@3929 | 171 | reg_def XMM10g( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(6)); |
kvn@3929 | 172 | reg_def XMM10h( SOC, SOE, Op_RegF, 10, xmm10->as_VMReg()->next(7)); |
kvn@3882 | 173 | |
kvn@3882 | 174 | reg_def XMM11 ( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()); |
kvn@3929 | 175 | reg_def XMM11b( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(1)); |
kvn@3929 | 176 | reg_def XMM11c( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(2)); |
kvn@3929 | 177 | reg_def XMM11d( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(3)); |
kvn@3929 | 178 | reg_def XMM11e( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(4)); |
kvn@3929 | 179 | reg_def XMM11f( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(5)); |
kvn@3929 | 180 | reg_def XMM11g( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(6)); |
kvn@3929 | 181 | reg_def XMM11h( SOC, SOE, Op_RegF, 11, xmm11->as_VMReg()->next(7)); |
kvn@3882 | 182 | |
kvn@3882 | 183 | reg_def XMM12 ( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()); |
kvn@3929 | 184 | reg_def XMM12b( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(1)); |
kvn@3929 | 185 | reg_def XMM12c( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(2)); |
kvn@3929 | 186 | reg_def XMM12d( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(3)); |
kvn@3929 | 187 | reg_def XMM12e( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(4)); |
kvn@3929 | 188 | reg_def XMM12f( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(5)); |
kvn@3929 | 189 | reg_def XMM12g( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(6)); |
kvn@3929 | 190 | reg_def XMM12h( SOC, SOE, Op_RegF, 12, xmm12->as_VMReg()->next(7)); |
kvn@3882 | 191 | |
kvn@3882 | 192 | reg_def XMM13 ( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()); |
kvn@3929 | 193 | reg_def XMM13b( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(1)); |
kvn@3929 | 194 | reg_def XMM13c( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(2)); |
kvn@3929 | 195 | reg_def XMM13d( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(3)); |
kvn@3929 | 196 | reg_def XMM13e( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(4)); |
kvn@3929 | 197 | reg_def XMM13f( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(5)); |
kvn@3929 | 198 | reg_def XMM13g( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(6)); |
kvn@3929 | 199 | reg_def XMM13h( SOC, SOE, Op_RegF, 13, xmm13->as_VMReg()->next(7)); |
kvn@3882 | 200 | |
kvn@3882 | 201 | reg_def XMM14 ( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()); |
kvn@3929 | 202 | reg_def XMM14b( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(1)); |
kvn@3929 | 203 | reg_def XMM14c( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(2)); |
kvn@3929 | 204 | reg_def XMM14d( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(3)); |
kvn@3929 | 205 | reg_def XMM14e( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(4)); |
kvn@3929 | 206 | reg_def XMM14f( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(5)); |
kvn@3929 | 207 | reg_def XMM14g( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(6)); |
kvn@3929 | 208 | reg_def XMM14h( SOC, SOE, Op_RegF, 14, xmm14->as_VMReg()->next(7)); |
kvn@3882 | 209 | |
kvn@3882 | 210 | reg_def XMM15 ( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()); |
kvn@3929 | 211 | reg_def XMM15b( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(1)); |
kvn@3929 | 212 | reg_def XMM15c( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(2)); |
kvn@3929 | 213 | reg_def XMM15d( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(3)); |
kvn@3929 | 214 | reg_def XMM15e( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(4)); |
kvn@3929 | 215 | reg_def XMM15f( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(5)); |
kvn@3929 | 216 | reg_def XMM15g( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(6)); |
kvn@3929 | 217 | reg_def XMM15h( SOC, SOE, Op_RegF, 15, xmm15->as_VMReg()->next(7)); |
kvn@3882 | 218 | |
kvn@3882 | 219 | #else // _WIN64 |
kvn@3882 | 220 | |
kvn@3882 | 221 | reg_def XMM6 ( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()); |
kvn@3929 | 222 | reg_def XMM6b( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(1)); |
kvn@3929 | 223 | reg_def XMM6c( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(2)); |
kvn@3929 | 224 | reg_def XMM6d( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(3)); |
kvn@3929 | 225 | reg_def XMM6e( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(4)); |
kvn@3929 | 226 | reg_def XMM6f( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(5)); |
kvn@3929 | 227 | reg_def XMM6g( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(6)); |
kvn@3929 | 228 | reg_def XMM6h( SOC, SOC, Op_RegF, 6, xmm6->as_VMReg()->next(7)); |
kvn@3882 | 229 | |
kvn@3882 | 230 | reg_def XMM7 ( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()); |
kvn@3929 | 231 | reg_def XMM7b( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(1)); |
kvn@3929 | 232 | reg_def XMM7c( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(2)); |
kvn@3929 | 233 | reg_def XMM7d( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(3)); |
kvn@3929 | 234 | reg_def XMM7e( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(4)); |
kvn@3929 | 235 | reg_def XMM7f( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(5)); |
kvn@3929 | 236 | reg_def XMM7g( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(6)); |
kvn@3929 | 237 | reg_def XMM7h( SOC, SOC, Op_RegF, 7, xmm7->as_VMReg()->next(7)); |
kvn@3882 | 238 | |
kvn@3882 | 239 | #ifdef _LP64 |
kvn@3882 | 240 | |
kvn@3882 | 241 | reg_def XMM8 ( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()); |
kvn@3929 | 242 | reg_def XMM8b( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(1)); |
kvn@3929 | 243 | reg_def XMM8c( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(2)); |
kvn@3929 | 244 | reg_def XMM8d( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(3)); |
kvn@3929 | 245 | reg_def XMM8e( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(4)); |
kvn@3929 | 246 | reg_def XMM8f( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(5)); |
kvn@3929 | 247 | reg_def XMM8g( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(6)); |
kvn@3929 | 248 | reg_def XMM8h( SOC, SOC, Op_RegF, 8, xmm8->as_VMReg()->next(7)); |
kvn@3882 | 249 | |
kvn@3882 | 250 | reg_def XMM9 ( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()); |
kvn@3929 | 251 | reg_def XMM9b( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(1)); |
kvn@3929 | 252 | reg_def XMM9c( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(2)); |
kvn@3929 | 253 | reg_def XMM9d( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(3)); |
kvn@3929 | 254 | reg_def XMM9e( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(4)); |
kvn@3929 | 255 | reg_def XMM9f( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(5)); |
kvn@3929 | 256 | reg_def XMM9g( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(6)); |
kvn@3929 | 257 | reg_def XMM9h( SOC, SOC, Op_RegF, 9, xmm9->as_VMReg()->next(7)); |
kvn@3882 | 258 | |
kvn@3882 | 259 | reg_def XMM10 ( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()); |
kvn@3929 | 260 | reg_def XMM10b( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(1)); |
kvn@3929 | 261 | reg_def XMM10c( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(2)); |
kvn@3929 | 262 | reg_def XMM10d( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(3)); |
kvn@3929 | 263 | reg_def XMM10e( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(4)); |
kvn@3929 | 264 | reg_def XMM10f( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(5)); |
kvn@3929 | 265 | reg_def XMM10g( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(6)); |
kvn@3929 | 266 | reg_def XMM10h( SOC, SOC, Op_RegF, 10, xmm10->as_VMReg()->next(7)); |
kvn@3882 | 267 | |
kvn@3882 | 268 | reg_def XMM11 ( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()); |
kvn@3929 | 269 | reg_def XMM11b( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(1)); |
kvn@3929 | 270 | reg_def XMM11c( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(2)); |
kvn@3929 | 271 | reg_def XMM11d( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(3)); |
kvn@3929 | 272 | reg_def XMM11e( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(4)); |
kvn@3929 | 273 | reg_def XMM11f( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(5)); |
kvn@3929 | 274 | reg_def XMM11g( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(6)); |
kvn@3929 | 275 | reg_def XMM11h( SOC, SOC, Op_RegF, 11, xmm11->as_VMReg()->next(7)); |
kvn@3882 | 276 | |
kvn@3882 | 277 | reg_def XMM12 ( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()); |
kvn@3929 | 278 | reg_def XMM12b( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(1)); |
kvn@3929 | 279 | reg_def XMM12c( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(2)); |
kvn@3929 | 280 | reg_def XMM12d( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(3)); |
kvn@3929 | 281 | reg_def XMM12e( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(4)); |
kvn@3929 | 282 | reg_def XMM12f( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(5)); |
kvn@3929 | 283 | reg_def XMM12g( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(6)); |
kvn@3929 | 284 | reg_def XMM12h( SOC, SOC, Op_RegF, 12, xmm12->as_VMReg()->next(7)); |
kvn@3882 | 285 | |
kvn@3882 | 286 | reg_def XMM13 ( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()); |
kvn@3929 | 287 | reg_def XMM13b( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(1)); |
kvn@3929 | 288 | reg_def XMM13c( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(2)); |
kvn@3929 | 289 | reg_def XMM13d( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(3)); |
kvn@3929 | 290 | reg_def XMM13e( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(4)); |
kvn@3929 | 291 | reg_def XMM13f( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(5)); |
kvn@3929 | 292 | reg_def XMM13g( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(6)); |
kvn@3929 | 293 | reg_def XMM13h( SOC, SOC, Op_RegF, 13, xmm13->as_VMReg()->next(7)); |
kvn@3882 | 294 | |
kvn@3882 | 295 | reg_def XMM14 ( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()); |
kvn@3929 | 296 | reg_def XMM14b( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(1)); |
kvn@3929 | 297 | reg_def XMM14c( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(2)); |
kvn@3929 | 298 | reg_def XMM14d( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(3)); |
kvn@3929 | 299 | reg_def XMM14e( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(4)); |
kvn@3929 | 300 | reg_def XMM14f( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(5)); |
kvn@3929 | 301 | reg_def XMM14g( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(6)); |
kvn@3929 | 302 | reg_def XMM14h( SOC, SOC, Op_RegF, 14, xmm14->as_VMReg()->next(7)); |
kvn@3882 | 303 | |
kvn@3882 | 304 | reg_def XMM15 ( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()); |
kvn@3929 | 305 | reg_def XMM15b( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(1)); |
kvn@3929 | 306 | reg_def XMM15c( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(2)); |
kvn@3929 | 307 | reg_def XMM15d( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(3)); |
kvn@3929 | 308 | reg_def XMM15e( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(4)); |
kvn@3929 | 309 | reg_def XMM15f( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(5)); |
kvn@3929 | 310 | reg_def XMM15g( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(6)); |
kvn@3929 | 311 | reg_def XMM15h( SOC, SOC, Op_RegF, 15, xmm15->as_VMReg()->next(7)); |
kvn@3882 | 312 | |
kvn@3882 | 313 | #endif // _LP64 |
kvn@3882 | 314 | |
kvn@3882 | 315 | #endif // _WIN64 |
kvn@3882 | 316 | |
kvn@3882 | 317 | #ifdef _LP64 |
kvn@3882 | 318 | reg_def RFLAGS(SOC, SOC, 0, 16, VMRegImpl::Bad()); |
kvn@3882 | 319 | #else |
kvn@3882 | 320 | reg_def RFLAGS(SOC, SOC, 0, 8, VMRegImpl::Bad()); |
kvn@3882 | 321 | #endif // _LP64 |
kvn@3882 | 322 | |
kvn@3882 | 323 | alloc_class chunk1(XMM0, XMM0b, XMM0c, XMM0d, XMM0e, XMM0f, XMM0g, XMM0h, |
kvn@3882 | 324 | XMM1, XMM1b, XMM1c, XMM1d, XMM1e, XMM1f, XMM1g, XMM1h, |
kvn@3882 | 325 | XMM2, XMM2b, XMM2c, XMM2d, XMM2e, XMM2f, XMM2g, XMM2h, |
kvn@3882 | 326 | XMM3, XMM3b, XMM3c, XMM3d, XMM3e, XMM3f, XMM3g, XMM3h, |
kvn@3882 | 327 | XMM4, XMM4b, XMM4c, XMM4d, XMM4e, XMM4f, XMM4g, XMM4h, |
kvn@3882 | 328 | XMM5, XMM5b, XMM5c, XMM5d, XMM5e, XMM5f, XMM5g, XMM5h, |
kvn@3882 | 329 | XMM6, XMM6b, XMM6c, XMM6d, XMM6e, XMM6f, XMM6g, XMM6h, |
kvn@3882 | 330 | XMM7, XMM7b, XMM7c, XMM7d, XMM7e, XMM7f, XMM7g, XMM7h |
kvn@3882 | 331 | #ifdef _LP64 |
kvn@3882 | 332 | ,XMM8, XMM8b, XMM8c, XMM8d, XMM8e, XMM8f, XMM8g, XMM8h, |
kvn@3882 | 333 | XMM9, XMM9b, XMM9c, XMM9d, XMM9e, XMM9f, XMM9g, XMM9h, |
kvn@3882 | 334 | XMM10, XMM10b, XMM10c, XMM10d, XMM10e, XMM10f, XMM10g, XMM10h, |
kvn@3882 | 335 | XMM11, XMM11b, XMM11c, XMM11d, XMM11e, XMM11f, XMM11g, XMM11h, |
kvn@3882 | 336 | XMM12, XMM12b, XMM12c, XMM12d, XMM12e, XMM12f, XMM12g, XMM12h, |
kvn@3882 | 337 | XMM13, XMM13b, XMM13c, XMM13d, XMM13e, XMM13f, XMM13g, XMM13h, |
kvn@3882 | 338 | XMM14, XMM14b, XMM14c, XMM14d, XMM14e, XMM14f, XMM14g, XMM14h, |
kvn@3882 | 339 | XMM15, XMM15b, XMM15c, XMM15d, XMM15e, XMM15f, XMM15g, XMM15h |
kvn@3882 | 340 | #endif |
kvn@3882 | 341 | ); |
kvn@3882 | 342 | |
kvn@3882 | 343 | // flags allocation class should be last. |
kvn@3882 | 344 | alloc_class chunk2(RFLAGS); |
kvn@3882 | 345 | |
kvn@3882 | 346 | // Singleton class for condition codes |
kvn@3882 | 347 | reg_class int_flags(RFLAGS); |
kvn@3882 | 348 | |
kvn@3882 | 349 | // Class for all float registers |
kvn@3882 | 350 | reg_class float_reg(XMM0, |
kvn@3882 | 351 | XMM1, |
kvn@3882 | 352 | XMM2, |
kvn@3882 | 353 | XMM3, |
kvn@3882 | 354 | XMM4, |
kvn@3882 | 355 | XMM5, |
kvn@3882 | 356 | XMM6, |
kvn@3882 | 357 | XMM7 |
kvn@3882 | 358 | #ifdef _LP64 |
kvn@3882 | 359 | ,XMM8, |
kvn@3882 | 360 | XMM9, |
kvn@3882 | 361 | XMM10, |
kvn@3882 | 362 | XMM11, |
kvn@3882 | 363 | XMM12, |
kvn@3882 | 364 | XMM13, |
kvn@3882 | 365 | XMM14, |
kvn@3882 | 366 | XMM15 |
kvn@3882 | 367 | #endif |
kvn@3882 | 368 | ); |
kvn@3882 | 369 | |
kvn@3882 | 370 | // Class for all double registers |
kvn@3882 | 371 | reg_class double_reg(XMM0, XMM0b, |
kvn@3882 | 372 | XMM1, XMM1b, |
kvn@3882 | 373 | XMM2, XMM2b, |
kvn@3882 | 374 | XMM3, XMM3b, |
kvn@3882 | 375 | XMM4, XMM4b, |
kvn@3882 | 376 | XMM5, XMM5b, |
kvn@3882 | 377 | XMM6, XMM6b, |
kvn@3882 | 378 | XMM7, XMM7b |
kvn@3882 | 379 | #ifdef _LP64 |
kvn@3882 | 380 | ,XMM8, XMM8b, |
kvn@3882 | 381 | XMM9, XMM9b, |
kvn@3882 | 382 | XMM10, XMM10b, |
kvn@3882 | 383 | XMM11, XMM11b, |
kvn@3882 | 384 | XMM12, XMM12b, |
kvn@3882 | 385 | XMM13, XMM13b, |
kvn@3882 | 386 | XMM14, XMM14b, |
kvn@3882 | 387 | XMM15, XMM15b |
kvn@3882 | 388 | #endif |
kvn@3882 | 389 | ); |
kvn@3882 | 390 | |
kvn@3882 | 391 | // Class for all 32bit vector registers |
kvn@3882 | 392 | reg_class vectors_reg(XMM0, |
kvn@3882 | 393 | XMM1, |
kvn@3882 | 394 | XMM2, |
kvn@3882 | 395 | XMM3, |
kvn@3882 | 396 | XMM4, |
kvn@3882 | 397 | XMM5, |
kvn@3882 | 398 | XMM6, |
kvn@3882 | 399 | XMM7 |
kvn@3882 | 400 | #ifdef _LP64 |
kvn@3882 | 401 | ,XMM8, |
kvn@3882 | 402 | XMM9, |
kvn@3882 | 403 | XMM10, |
kvn@3882 | 404 | XMM11, |
kvn@3882 | 405 | XMM12, |
kvn@3882 | 406 | XMM13, |
kvn@3882 | 407 | XMM14, |
kvn@3882 | 408 | XMM15 |
kvn@3882 | 409 | #endif |
kvn@3882 | 410 | ); |
kvn@3882 | 411 | |
kvn@3882 | 412 | // Class for all 64bit vector registers |
kvn@3882 | 413 | reg_class vectord_reg(XMM0, XMM0b, |
kvn@3882 | 414 | XMM1, XMM1b, |
kvn@3882 | 415 | XMM2, XMM2b, |
kvn@3882 | 416 | XMM3, XMM3b, |
kvn@3882 | 417 | XMM4, XMM4b, |
kvn@3882 | 418 | XMM5, XMM5b, |
kvn@3882 | 419 | XMM6, XMM6b, |
kvn@3882 | 420 | XMM7, XMM7b |
kvn@3882 | 421 | #ifdef _LP64 |
kvn@3882 | 422 | ,XMM8, XMM8b, |
kvn@3882 | 423 | XMM9, XMM9b, |
kvn@3882 | 424 | XMM10, XMM10b, |
kvn@3882 | 425 | XMM11, XMM11b, |
kvn@3882 | 426 | XMM12, XMM12b, |
kvn@3882 | 427 | XMM13, XMM13b, |
kvn@3882 | 428 | XMM14, XMM14b, |
kvn@3882 | 429 | XMM15, XMM15b |
kvn@3882 | 430 | #endif |
kvn@3882 | 431 | ); |
kvn@3882 | 432 | |
kvn@3882 | 433 | // Class for all 128bit vector registers |
kvn@3882 | 434 | reg_class vectorx_reg(XMM0, XMM0b, XMM0c, XMM0d, |
kvn@3882 | 435 | XMM1, XMM1b, XMM1c, XMM1d, |
kvn@3882 | 436 | XMM2, XMM2b, XMM2c, XMM2d, |
kvn@3882 | 437 | XMM3, XMM3b, XMM3c, XMM3d, |
kvn@3882 | 438 | XMM4, XMM4b, XMM4c, XMM4d, |
kvn@3882 | 439 | XMM5, XMM5b, XMM5c, XMM5d, |
kvn@3882 | 440 | XMM6, XMM6b, XMM6c, XMM6d, |
kvn@3882 | 441 | XMM7, XMM7b, XMM7c, XMM7d |
kvn@3882 | 442 | #ifdef _LP64 |
kvn@3882 | 443 | ,XMM8, XMM8b, XMM8c, XMM8d, |
kvn@3882 | 444 | XMM9, XMM9b, XMM9c, XMM9d, |
kvn@3882 | 445 | XMM10, XMM10b, XMM10c, XMM10d, |
kvn@3882 | 446 | XMM11, XMM11b, XMM11c, XMM11d, |
kvn@3882 | 447 | XMM12, XMM12b, XMM12c, XMM12d, |
kvn@3882 | 448 | XMM13, XMM13b, XMM13c, XMM13d, |
kvn@3882 | 449 | XMM14, XMM14b, XMM14c, XMM14d, |
kvn@3882 | 450 | XMM15, XMM15b, XMM15c, XMM15d |
kvn@3882 | 451 | #endif |
kvn@3882 | 452 | ); |
kvn@3882 | 453 | |
kvn@3882 | 454 | // Class for all 256bit vector registers |
kvn@3882 | 455 | reg_class vectory_reg(XMM0, XMM0b, XMM0c, XMM0d, XMM0e, XMM0f, XMM0g, XMM0h, |
kvn@3882 | 456 | XMM1, XMM1b, XMM1c, XMM1d, XMM1e, XMM1f, XMM1g, XMM1h, |
kvn@3882 | 457 | XMM2, XMM2b, XMM2c, XMM2d, XMM2e, XMM2f, XMM2g, XMM2h, |
kvn@3882 | 458 | XMM3, XMM3b, XMM3c, XMM3d, XMM3e, XMM3f, XMM3g, XMM3h, |
kvn@3882 | 459 | XMM4, XMM4b, XMM4c, XMM4d, XMM4e, XMM4f, XMM4g, XMM4h, |
kvn@3882 | 460 | XMM5, XMM5b, XMM5c, XMM5d, XMM5e, XMM5f, XMM5g, XMM5h, |
kvn@3882 | 461 | XMM6, XMM6b, XMM6c, XMM6d, XMM6e, XMM6f, XMM6g, XMM6h, |
kvn@3882 | 462 | XMM7, XMM7b, XMM7c, XMM7d, XMM7e, XMM7f, XMM7g, XMM7h |
kvn@3882 | 463 | #ifdef _LP64 |
kvn@3882 | 464 | ,XMM8, XMM8b, XMM8c, XMM8d, XMM8e, XMM8f, XMM8g, XMM8h, |
kvn@3882 | 465 | XMM9, XMM9b, XMM9c, XMM9d, XMM9e, XMM9f, XMM9g, XMM9h, |
kvn@3882 | 466 | XMM10, XMM10b, XMM10c, XMM10d, XMM10e, XMM10f, XMM10g, XMM10h, |
kvn@3882 | 467 | XMM11, XMM11b, XMM11c, XMM11d, XMM11e, XMM11f, XMM11g, XMM11h, |
kvn@3882 | 468 | XMM12, XMM12b, XMM12c, XMM12d, XMM12e, XMM12f, XMM12g, XMM12h, |
kvn@3882 | 469 | XMM13, XMM13b, XMM13c, XMM13d, XMM13e, XMM13f, XMM13g, XMM13h, |
kvn@3882 | 470 | XMM14, XMM14b, XMM14c, XMM14d, XMM14e, XMM14f, XMM14g, XMM14h, |
kvn@3882 | 471 | XMM15, XMM15b, XMM15c, XMM15d, XMM15e, XMM15f, XMM15g, XMM15h |
kvn@3882 | 472 | #endif |
kvn@3882 | 473 | ); |
kvn@3882 | 474 | |
kvn@3882 | 475 | %} |
kvn@3882 | 476 | |
goetz@6517 | 477 | |
goetz@6517 | 478 | //----------SOURCE BLOCK------------------------------------------------------- |
goetz@6517 | 479 | // This is a block of C++ code which provides values, functions, and |
goetz@6517 | 480 | // definitions necessary in the rest of the architecture description |
goetz@6517 | 481 | |
goetz@6517 | 482 | source_hpp %{ |
goetz@6517 | 483 | // Header information of the source block. |
goetz@6517 | 484 | // Method declarations/definitions which are used outside |
goetz@6517 | 485 | // the ad-scope can conveniently be defined here. |
goetz@6517 | 486 | // |
goetz@6517 | 487 | // To keep related declarations/definitions/uses close together, |
goetz@6517 | 488 | // we switch between source %{ }% and source_hpp %{ }% freely as needed. |
goetz@6517 | 489 | |
goetz@6517 | 490 | class CallStubImpl { |
goetz@6517 | 491 | |
goetz@6517 | 492 | //-------------------------------------------------------------- |
goetz@6517 | 493 | //---< Used for optimization in Compile::shorten_branches >--- |
goetz@6517 | 494 | //-------------------------------------------------------------- |
goetz@6517 | 495 | |
goetz@6517 | 496 | public: |
goetz@6517 | 497 | // Size of call trampoline stub. |
goetz@6517 | 498 | static uint size_call_trampoline() { |
goetz@6517 | 499 | return 0; // no call trampolines on this platform |
goetz@6517 | 500 | } |
goetz@6517 | 501 | |
goetz@6517 | 502 | // number of relocations needed by a call trampoline stub |
goetz@6517 | 503 | static uint reloc_call_trampoline() { |
goetz@6517 | 504 | return 0; // no call trampolines on this platform |
goetz@6517 | 505 | } |
goetz@6517 | 506 | }; |
goetz@6517 | 507 | |
goetz@6517 | 508 | class HandlerImpl { |
goetz@6517 | 509 | |
goetz@6517 | 510 | public: |
goetz@6517 | 511 | |
goetz@6517 | 512 | static int emit_exception_handler(CodeBuffer &cbuf); |
goetz@6517 | 513 | static int emit_deopt_handler(CodeBuffer& cbuf); |
goetz@6517 | 514 | |
goetz@6517 | 515 | static uint size_exception_handler() { |
goetz@6517 | 516 | // NativeCall instruction size is the same as NativeJump. |
goetz@6517 | 517 | // exception handler starts out as jump and can be patched to |
goetz@6517 | 518 | // a call be deoptimization. (4932387) |
goetz@6517 | 519 | // Note that this value is also credited (in output.cpp) to |
goetz@6517 | 520 | // the size of the code section. |
goetz@6517 | 521 | return NativeJump::instruction_size; |
goetz@6517 | 522 | } |
goetz@6517 | 523 | |
goetz@6517 | 524 | #ifdef _LP64 |
goetz@6517 | 525 | static uint size_deopt_handler() { |
goetz@6517 | 526 | // three 5 byte instructions |
goetz@6517 | 527 | return 15; |
goetz@6517 | 528 | } |
goetz@6517 | 529 | #else |
goetz@6517 | 530 | static uint size_deopt_handler() { |
goetz@6517 | 531 | // NativeCall instruction size is the same as NativeJump. |
goetz@6517 | 532 | // exception handler starts out as jump and can be patched to |
goetz@6517 | 533 | // a call be deoptimization. (4932387) |
goetz@6517 | 534 | // Note that this value is also credited (in output.cpp) to |
goetz@6517 | 535 | // the size of the code section. |
goetz@6517 | 536 | return 5 + NativeJump::instruction_size; // pushl(); jmp; |
goetz@6517 | 537 | } |
goetz@6517 | 538 | #endif |
goetz@6517 | 539 | }; |
goetz@6517 | 540 | |
goetz@6517 | 541 | %} // end source_hpp |
goetz@6517 | 542 | |
kvn@3390 | 543 | source %{ |
goetz@6517 | 544 | |
goetz@6517 | 545 | // Emit exception handler code. |
goetz@6517 | 546 | // Stuff framesize into a register and call a VM stub routine. |
goetz@6517 | 547 | int HandlerImpl::emit_exception_handler(CodeBuffer& cbuf) { |
goetz@6517 | 548 | |
goetz@6517 | 549 | // Note that the code buffer's insts_mark is always relative to insts. |
goetz@6517 | 550 | // That's why we must use the macroassembler to generate a handler. |
goetz@6517 | 551 | MacroAssembler _masm(&cbuf); |
goetz@6517 | 552 | address base = __ start_a_stub(size_exception_handler()); |
vkempik@8427 | 553 | if (base == NULL) { |
vkempik@8427 | 554 | ciEnv::current()->record_failure("CodeCache is full"); |
vkempik@8427 | 555 | return 0; // CodeBuffer::expand failed |
vkempik@8427 | 556 | } |
goetz@6517 | 557 | int offset = __ offset(); |
goetz@6517 | 558 | __ jump(RuntimeAddress(OptoRuntime::exception_blob()->entry_point())); |
goetz@6517 | 559 | assert(__ offset() - offset <= (int) size_exception_handler(), "overflow"); |
goetz@6517 | 560 | __ end_a_stub(); |
goetz@6517 | 561 | return offset; |
goetz@6517 | 562 | } |
goetz@6517 | 563 | |
goetz@6517 | 564 | // Emit deopt handler code. |
goetz@6517 | 565 | int HandlerImpl::emit_deopt_handler(CodeBuffer& cbuf) { |
goetz@6517 | 566 | |
goetz@6517 | 567 | // Note that the code buffer's insts_mark is always relative to insts. |
goetz@6517 | 568 | // That's why we must use the macroassembler to generate a handler. |
goetz@6517 | 569 | MacroAssembler _masm(&cbuf); |
goetz@6517 | 570 | address base = __ start_a_stub(size_deopt_handler()); |
vkempik@8427 | 571 | if (base == NULL) { |
vkempik@8427 | 572 | ciEnv::current()->record_failure("CodeCache is full"); |
vkempik@8427 | 573 | return 0; // CodeBuffer::expand failed |
vkempik@8427 | 574 | } |
goetz@6517 | 575 | int offset = __ offset(); |
goetz@6517 | 576 | |
goetz@6517 | 577 | #ifdef _LP64 |
goetz@6517 | 578 | address the_pc = (address) __ pc(); |
goetz@6517 | 579 | Label next; |
goetz@6517 | 580 | // push a "the_pc" on the stack without destroying any registers |
goetz@6517 | 581 | // as they all may be live. |
goetz@6517 | 582 | |
goetz@6517 | 583 | // push address of "next" |
goetz@6517 | 584 | __ call(next, relocInfo::none); // reloc none is fine since it is a disp32 |
goetz@6517 | 585 | __ bind(next); |
goetz@6517 | 586 | // adjust it so it matches "the_pc" |
goetz@6517 | 587 | __ subptr(Address(rsp, 0), __ offset() - offset); |
goetz@6517 | 588 | #else |
goetz@6517 | 589 | InternalAddress here(__ pc()); |
goetz@6517 | 590 | __ pushptr(here.addr()); |
goetz@6517 | 591 | #endif |
goetz@6517 | 592 | |
goetz@6517 | 593 | __ jump(RuntimeAddress(SharedRuntime::deopt_blob()->unpack())); |
goetz@6517 | 594 | assert(__ offset() - offset <= (int) size_deopt_handler(), "overflow"); |
goetz@6517 | 595 | __ end_a_stub(); |
goetz@6517 | 596 | return offset; |
goetz@6517 | 597 | } |
goetz@6517 | 598 | |
goetz@6517 | 599 | |
goetz@6517 | 600 | //============================================================================= |
goetz@6517 | 601 | |
kvn@3390 | 602 | // Float masks come from different places depending on platform. |
kvn@3390 | 603 | #ifdef _LP64 |
kvn@3390 | 604 | static address float_signmask() { return StubRoutines::x86::float_sign_mask(); } |
kvn@3390 | 605 | static address float_signflip() { return StubRoutines::x86::float_sign_flip(); } |
kvn@3390 | 606 | static address double_signmask() { return StubRoutines::x86::double_sign_mask(); } |
kvn@3390 | 607 | static address double_signflip() { return StubRoutines::x86::double_sign_flip(); } |
kvn@3390 | 608 | #else |
kvn@3390 | 609 | static address float_signmask() { return (address)float_signmask_pool; } |
kvn@3390 | 610 | static address float_signflip() { return (address)float_signflip_pool; } |
kvn@3390 | 611 | static address double_signmask() { return (address)double_signmask_pool; } |
kvn@3390 | 612 | static address double_signflip() { return (address)double_signflip_pool; } |
kvn@3390 | 613 | #endif |
kvn@3577 | 614 | |
kvn@3882 | 615 | |
kvn@4001 | 616 | const bool Matcher::match_rule_supported(int opcode) { |
kvn@4001 | 617 | if (!has_match_rule(opcode)) |
kvn@4001 | 618 | return false; |
kvn@4001 | 619 | |
kvn@4001 | 620 | switch (opcode) { |
kvn@4001 | 621 | case Op_PopCountI: |
kvn@4001 | 622 | case Op_PopCountL: |
kvn@4001 | 623 | if (!UsePopCountInstruction) |
kvn@4001 | 624 | return false; |
kvn@4103 | 625 | break; |
kvn@4001 | 626 | case Op_MulVI: |
kvn@4001 | 627 | if ((UseSSE < 4) && (UseAVX < 1)) // only with SSE4_1 or AVX |
kvn@4001 | 628 | return false; |
kvn@4001 | 629 | break; |
roland@4106 | 630 | case Op_CompareAndSwapL: |
roland@4106 | 631 | #ifdef _LP64 |
roland@4106 | 632 | case Op_CompareAndSwapP: |
roland@4106 | 633 | #endif |
roland@4106 | 634 | if (!VM_Version::supports_cx8()) |
roland@4106 | 635 | return false; |
roland@4106 | 636 | break; |
kvn@4001 | 637 | } |
kvn@4001 | 638 | |
kvn@4001 | 639 | return true; // Per default match rules are supported. |
kvn@4001 | 640 | } |
kvn@4001 | 641 | |
kvn@3882 | 642 | // Max vector size in bytes. 0 if not supported. |
kvn@3882 | 643 | const int Matcher::vector_width_in_bytes(BasicType bt) { |
kvn@3882 | 644 | assert(is_java_primitive(bt), "only primitive type vectors"); |
kvn@3882 | 645 | if (UseSSE < 2) return 0; |
kvn@3882 | 646 | // SSE2 supports 128bit vectors for all types. |
kvn@3882 | 647 | // AVX2 supports 256bit vectors for all types. |
kvn@3882 | 648 | int size = (UseAVX > 1) ? 32 : 16; |
kvn@3882 | 649 | // AVX1 supports 256bit vectors only for FLOAT and DOUBLE. |
kvn@3882 | 650 | if (UseAVX > 0 && (bt == T_FLOAT || bt == T_DOUBLE)) |
kvn@3882 | 651 | size = 32; |
kvn@3882 | 652 | // Use flag to limit vector size. |
kvn@3882 | 653 | size = MIN2(size,(int)MaxVectorSize); |
kvn@3882 | 654 | // Minimum 2 values in vector (or 4 for bytes). |
kvn@3882 | 655 | switch (bt) { |
kvn@3882 | 656 | case T_DOUBLE: |
kvn@3882 | 657 | case T_LONG: |
kvn@3882 | 658 | if (size < 16) return 0; |
kvn@3882 | 659 | case T_FLOAT: |
kvn@3882 | 660 | case T_INT: |
kvn@3882 | 661 | if (size < 8) return 0; |
kvn@3882 | 662 | case T_BOOLEAN: |
kvn@3882 | 663 | case T_BYTE: |
kvn@3882 | 664 | case T_CHAR: |
kvn@3882 | 665 | case T_SHORT: |
kvn@3882 | 666 | if (size < 4) return 0; |
kvn@3882 | 667 | break; |
kvn@3882 | 668 | default: |
kvn@3882 | 669 | ShouldNotReachHere(); |
kvn@3882 | 670 | } |
kvn@3882 | 671 | return size; |
kvn@3882 | 672 | } |
kvn@3882 | 673 | |
kvn@3882 | 674 | // Limits on vector size (number of elements) loaded into vector. |
kvn@3882 | 675 | const int Matcher::max_vector_size(const BasicType bt) { |
kvn@3882 | 676 | return vector_width_in_bytes(bt)/type2aelembytes(bt); |
kvn@3882 | 677 | } |
kvn@3882 | 678 | const int Matcher::min_vector_size(const BasicType bt) { |
kvn@3882 | 679 | int max_size = max_vector_size(bt); |
kvn@3882 | 680 | // Min size which can be loaded into vector is 4 bytes. |
kvn@3882 | 681 | int size = (type2aelembytes(bt) == 1) ? 4 : 2; |
kvn@3882 | 682 | return MIN2(size,max_size); |
kvn@3882 | 683 | } |
kvn@3882 | 684 | |
kvn@3882 | 685 | // Vector ideal reg corresponding to specidied size in bytes |
kvn@3882 | 686 | const int Matcher::vector_ideal_reg(int size) { |
kvn@3882 | 687 | assert(MaxVectorSize >= size, ""); |
kvn@3882 | 688 | switch(size) { |
kvn@3882 | 689 | case 4: return Op_VecS; |
kvn@3882 | 690 | case 8: return Op_VecD; |
kvn@3882 | 691 | case 16: return Op_VecX; |
kvn@3882 | 692 | case 32: return Op_VecY; |
kvn@3882 | 693 | } |
kvn@3882 | 694 | ShouldNotReachHere(); |
kvn@3882 | 695 | return 0; |
kvn@3882 | 696 | } |
kvn@3882 | 697 | |
kvn@4134 | 698 | // Only lowest bits of xmm reg are used for vector shift count. |
kvn@4134 | 699 | const int Matcher::vector_shift_count_ideal_reg(int size) { |
kvn@4134 | 700 | return Op_VecS; |
kvn@4134 | 701 | } |
kvn@4134 | 702 | |
kvn@3882 | 703 | // x86 supports misaligned vectors store/load. |
kvn@3882 | 704 | const bool Matcher::misaligned_vectors_ok() { |
kvn@3882 | 705 | return !AlignVector; // can be changed by flag |
kvn@3882 | 706 | } |
kvn@3882 | 707 | |
kvn@6312 | 708 | // x86 AES instructions are compatible with SunJCE expanded |
kvn@6312 | 709 | // keys, hence we do not need to pass the original key to stubs |
kvn@6312 | 710 | const bool Matcher::pass_original_key_for_aes() { |
kvn@6312 | 711 | return false; |
kvn@6312 | 712 | } |
kvn@6312 | 713 | |
kvn@3882 | 714 | // Helper methods for MachSpillCopyNode::implementation(). |
kvn@3882 | 715 | static int vec_mov_helper(CodeBuffer *cbuf, bool do_size, int src_lo, int dst_lo, |
kvn@3882 | 716 | int src_hi, int dst_hi, uint ireg, outputStream* st) { |
kvn@3882 | 717 | // In 64-bit VM size calculation is very complex. Emitting instructions |
kvn@3882 | 718 | // into scratch buffer is used to get size in 64-bit VM. |
kvn@3882 | 719 | LP64_ONLY( assert(!do_size, "this method calculates size only for 32-bit VM"); ) |
kvn@3882 | 720 | assert(ireg == Op_VecS || // 32bit vector |
kvn@3882 | 721 | (src_lo & 1) == 0 && (src_lo + 1) == src_hi && |
kvn@3882 | 722 | (dst_lo & 1) == 0 && (dst_lo + 1) == dst_hi, |
kvn@3882 | 723 | "no non-adjacent vector moves" ); |
kvn@3882 | 724 | if (cbuf) { |
kvn@3882 | 725 | MacroAssembler _masm(cbuf); |
kvn@3882 | 726 | int offset = __ offset(); |
kvn@3882 | 727 | switch (ireg) { |
kvn@3882 | 728 | case Op_VecS: // copy whole register |
kvn@3882 | 729 | case Op_VecD: |
kvn@3882 | 730 | case Op_VecX: |
kvn@3882 | 731 | __ movdqu(as_XMMRegister(Matcher::_regEncode[dst_lo]), as_XMMRegister(Matcher::_regEncode[src_lo])); |
kvn@3882 | 732 | break; |
kvn@3882 | 733 | case Op_VecY: |
kvn@3882 | 734 | __ vmovdqu(as_XMMRegister(Matcher::_regEncode[dst_lo]), as_XMMRegister(Matcher::_regEncode[src_lo])); |
kvn@3882 | 735 | break; |
kvn@3882 | 736 | default: |
kvn@3882 | 737 | ShouldNotReachHere(); |
kvn@3882 | 738 | } |
kvn@3882 | 739 | int size = __ offset() - offset; |
kvn@3882 | 740 | #ifdef ASSERT |
kvn@3882 | 741 | // VEX_2bytes prefix is used if UseAVX > 0, so it takes the same 2 bytes as SIMD prefix. |
kvn@3882 | 742 | assert(!do_size || size == 4, "incorrect size calculattion"); |
kvn@3882 | 743 | #endif |
kvn@3882 | 744 | return size; |
kvn@3882 | 745 | #ifndef PRODUCT |
kvn@3882 | 746 | } else if (!do_size) { |
kvn@3882 | 747 | switch (ireg) { |
kvn@3882 | 748 | case Op_VecS: |
kvn@3882 | 749 | case Op_VecD: |
kvn@3882 | 750 | case Op_VecX: |
kvn@3882 | 751 | st->print("movdqu %s,%s\t# spill",Matcher::regName[dst_lo],Matcher::regName[src_lo]); |
kvn@3882 | 752 | break; |
kvn@3882 | 753 | case Op_VecY: |
kvn@3882 | 754 | st->print("vmovdqu %s,%s\t# spill",Matcher::regName[dst_lo],Matcher::regName[src_lo]); |
kvn@3882 | 755 | break; |
kvn@3882 | 756 | default: |
kvn@3882 | 757 | ShouldNotReachHere(); |
kvn@3882 | 758 | } |
kvn@3882 | 759 | #endif |
kvn@3882 | 760 | } |
kvn@3882 | 761 | // VEX_2bytes prefix is used if UseAVX > 0, and it takes the same 2 bytes as SIMD prefix. |
kvn@3882 | 762 | return 4; |
kvn@3882 | 763 | } |
kvn@3882 | 764 | |
kvn@3882 | 765 | static int vec_spill_helper(CodeBuffer *cbuf, bool do_size, bool is_load, |
kvn@3882 | 766 | int stack_offset, int reg, uint ireg, outputStream* st) { |
kvn@3882 | 767 | // In 64-bit VM size calculation is very complex. Emitting instructions |
kvn@3882 | 768 | // into scratch buffer is used to get size in 64-bit VM. |
kvn@3882 | 769 | LP64_ONLY( assert(!do_size, "this method calculates size only for 32-bit VM"); ) |
kvn@3882 | 770 | if (cbuf) { |
kvn@3882 | 771 | MacroAssembler _masm(cbuf); |
kvn@3882 | 772 | int offset = __ offset(); |
kvn@3882 | 773 | if (is_load) { |
kvn@3882 | 774 | switch (ireg) { |
kvn@3882 | 775 | case Op_VecS: |
kvn@3882 | 776 | __ movdl(as_XMMRegister(Matcher::_regEncode[reg]), Address(rsp, stack_offset)); |
kvn@3882 | 777 | break; |
kvn@3882 | 778 | case Op_VecD: |
kvn@3882 | 779 | __ movq(as_XMMRegister(Matcher::_regEncode[reg]), Address(rsp, stack_offset)); |
kvn@3882 | 780 | break; |
kvn@3882 | 781 | case Op_VecX: |
kvn@3882 | 782 | __ movdqu(as_XMMRegister(Matcher::_regEncode[reg]), Address(rsp, stack_offset)); |
kvn@3882 | 783 | break; |
kvn@3882 | 784 | case Op_VecY: |
kvn@3882 | 785 | __ vmovdqu(as_XMMRegister(Matcher::_regEncode[reg]), Address(rsp, stack_offset)); |
kvn@3882 | 786 | break; |
kvn@3882 | 787 | default: |
kvn@3882 | 788 | ShouldNotReachHere(); |
kvn@3882 | 789 | } |
kvn@3882 | 790 | } else { // store |
kvn@3882 | 791 | switch (ireg) { |
kvn@3882 | 792 | case Op_VecS: |
kvn@3882 | 793 | __ movdl(Address(rsp, stack_offset), as_XMMRegister(Matcher::_regEncode[reg])); |
kvn@3882 | 794 | break; |
kvn@3882 | 795 | case Op_VecD: |
kvn@3882 | 796 | __ movq(Address(rsp, stack_offset), as_XMMRegister(Matcher::_regEncode[reg])); |
kvn@3882 | 797 | break; |
kvn@3882 | 798 | case Op_VecX: |
kvn@3882 | 799 | __ movdqu(Address(rsp, stack_offset), as_XMMRegister(Matcher::_regEncode[reg])); |
kvn@3882 | 800 | break; |
kvn@3882 | 801 | case Op_VecY: |
kvn@3882 | 802 | __ vmovdqu(Address(rsp, stack_offset), as_XMMRegister(Matcher::_regEncode[reg])); |
kvn@3882 | 803 | break; |
kvn@3882 | 804 | default: |
kvn@3882 | 805 | ShouldNotReachHere(); |
kvn@3882 | 806 | } |
kvn@3882 | 807 | } |
kvn@3882 | 808 | int size = __ offset() - offset; |
kvn@3882 | 809 | #ifdef ASSERT |
kvn@3882 | 810 | int offset_size = (stack_offset == 0) ? 0 : ((stack_offset < 0x80) ? 1 : 4); |
kvn@3882 | 811 | // VEX_2bytes prefix is used if UseAVX > 0, so it takes the same 2 bytes as SIMD prefix. |
kvn@3882 | 812 | assert(!do_size || size == (5+offset_size), "incorrect size calculattion"); |
kvn@3882 | 813 | #endif |
kvn@3882 | 814 | return size; |
kvn@3882 | 815 | #ifndef PRODUCT |
kvn@3882 | 816 | } else if (!do_size) { |
kvn@3882 | 817 | if (is_load) { |
kvn@3882 | 818 | switch (ireg) { |
kvn@3882 | 819 | case Op_VecS: |
kvn@3882 | 820 | st->print("movd %s,[rsp + %d]\t# spill", Matcher::regName[reg], stack_offset); |
kvn@3882 | 821 | break; |
kvn@3882 | 822 | case Op_VecD: |
kvn@3882 | 823 | st->print("movq %s,[rsp + %d]\t# spill", Matcher::regName[reg], stack_offset); |
kvn@3882 | 824 | break; |
kvn@3882 | 825 | case Op_VecX: |
kvn@3882 | 826 | st->print("movdqu %s,[rsp + %d]\t# spill", Matcher::regName[reg], stack_offset); |
kvn@3882 | 827 | break; |
kvn@3882 | 828 | case Op_VecY: |
kvn@3882 | 829 | st->print("vmovdqu %s,[rsp + %d]\t# spill", Matcher::regName[reg], stack_offset); |
kvn@3882 | 830 | break; |
kvn@3882 | 831 | default: |
kvn@3882 | 832 | ShouldNotReachHere(); |
kvn@3882 | 833 | } |
kvn@3882 | 834 | } else { // store |
kvn@3882 | 835 | switch (ireg) { |
kvn@3882 | 836 | case Op_VecS: |
kvn@3882 | 837 | st->print("movd [rsp + %d],%s\t# spill", stack_offset, Matcher::regName[reg]); |
kvn@3882 | 838 | break; |
kvn@3882 | 839 | case Op_VecD: |
kvn@3882 | 840 | st->print("movq [rsp + %d],%s\t# spill", stack_offset, Matcher::regName[reg]); |
kvn@3882 | 841 | break; |
kvn@3882 | 842 | case Op_VecX: |
kvn@3882 | 843 | st->print("movdqu [rsp + %d],%s\t# spill", stack_offset, Matcher::regName[reg]); |
kvn@3882 | 844 | break; |
kvn@3882 | 845 | case Op_VecY: |
kvn@3882 | 846 | st->print("vmovdqu [rsp + %d],%s\t# spill", stack_offset, Matcher::regName[reg]); |
kvn@3882 | 847 | break; |
kvn@3882 | 848 | default: |
kvn@3882 | 849 | ShouldNotReachHere(); |
kvn@3882 | 850 | } |
kvn@3882 | 851 | } |
kvn@3882 | 852 | #endif |
kvn@3882 | 853 | } |
kvn@3882 | 854 | int offset_size = (stack_offset == 0) ? 0 : ((stack_offset < 0x80) ? 1 : 4); |
kvn@3882 | 855 | // VEX_2bytes prefix is used if UseAVX > 0, so it takes the same 2 bytes as SIMD prefix. |
kvn@3882 | 856 | return 5+offset_size; |
kvn@3882 | 857 | } |
kvn@3882 | 858 | |
kvn@3882 | 859 | static inline jfloat replicate4_imm(int con, int width) { |
kvn@3882 | 860 | // Load a constant of "width" (in bytes) and replicate it to fill 32bit. |
kvn@3882 | 861 | assert(width == 1 || width == 2, "only byte or short types here"); |
kvn@3882 | 862 | int bit_width = width * 8; |
kvn@3882 | 863 | jint val = con; |
kvn@3882 | 864 | val &= (1 << bit_width) - 1; // mask off sign bits |
kvn@3882 | 865 | while(bit_width < 32) { |
kvn@3882 | 866 | val |= (val << bit_width); |
kvn@3882 | 867 | bit_width <<= 1; |
kvn@3882 | 868 | } |
kvn@3882 | 869 | jfloat fval = *((jfloat*) &val); // coerce to float type |
kvn@3882 | 870 | return fval; |
kvn@3882 | 871 | } |
kvn@3882 | 872 | |
kvn@3882 | 873 | static inline jdouble replicate8_imm(int con, int width) { |
kvn@3882 | 874 | // Load a constant of "width" (in bytes) and replicate it to fill 64bit. |
kvn@3882 | 875 | assert(width == 1 || width == 2 || width == 4, "only byte, short or int types here"); |
kvn@3882 | 876 | int bit_width = width * 8; |
kvn@3882 | 877 | jlong val = con; |
kvn@3882 | 878 | val &= (((jlong) 1) << bit_width) - 1; // mask off sign bits |
kvn@3882 | 879 | while(bit_width < 64) { |
kvn@3882 | 880 | val |= (val << bit_width); |
kvn@3882 | 881 | bit_width <<= 1; |
kvn@3882 | 882 | } |
kvn@3882 | 883 | jdouble dval = *((jdouble*) &val); // coerce to double type |
kvn@3882 | 884 | return dval; |
kvn@3882 | 885 | } |
kvn@3882 | 886 | |
kvn@3577 | 887 | #ifndef PRODUCT |
kvn@3577 | 888 | void MachNopNode::format(PhaseRegAlloc*, outputStream* st) const { |
kvn@3577 | 889 | st->print("nop \t# %d bytes pad for loops and calls", _count); |
kvn@3577 | 890 | } |
kvn@3577 | 891 | #endif |
kvn@3577 | 892 | |
kvn@3577 | 893 | void MachNopNode::emit(CodeBuffer &cbuf, PhaseRegAlloc*) const { |
kvn@3577 | 894 | MacroAssembler _masm(&cbuf); |
kvn@3577 | 895 | __ nop(_count); |
kvn@3577 | 896 | } |
kvn@3577 | 897 | |
kvn@3577 | 898 | uint MachNopNode::size(PhaseRegAlloc*) const { |
kvn@3577 | 899 | return _count; |
kvn@3577 | 900 | } |
kvn@3577 | 901 | |
kvn@3577 | 902 | #ifndef PRODUCT |
kvn@3577 | 903 | void MachBreakpointNode::format(PhaseRegAlloc*, outputStream* st) const { |
kvn@3577 | 904 | st->print("# breakpoint"); |
kvn@3577 | 905 | } |
kvn@3577 | 906 | #endif |
kvn@3577 | 907 | |
kvn@3577 | 908 | void MachBreakpointNode::emit(CodeBuffer &cbuf, PhaseRegAlloc* ra_) const { |
kvn@3577 | 909 | MacroAssembler _masm(&cbuf); |
kvn@3577 | 910 | __ int3(); |
kvn@3577 | 911 | } |
kvn@3577 | 912 | |
kvn@3577 | 913 | uint MachBreakpointNode::size(PhaseRegAlloc* ra_) const { |
kvn@3577 | 914 | return MachNode::size(ra_); |
kvn@3577 | 915 | } |
kvn@3577 | 916 | |
kvn@3577 | 917 | %} |
kvn@3577 | 918 | |
kvn@3577 | 919 | encode %{ |
kvn@3577 | 920 | |
kvn@3577 | 921 | enc_class call_epilog %{ |
kvn@3577 | 922 | if (VerifyStackAtCalls) { |
kvn@3577 | 923 | // Check that stack depth is unchanged: find majik cookie on stack |
kvn@3577 | 924 | int framesize = ra_->reg2offset_unchecked(OptoReg::add(ra_->_matcher._old_SP, -3*VMRegImpl::slots_per_word)); |
kvn@3577 | 925 | MacroAssembler _masm(&cbuf); |
kvn@3577 | 926 | Label L; |
kvn@3577 | 927 | __ cmpptr(Address(rsp, framesize), (int32_t)0xbadb100d); |
kvn@3577 | 928 | __ jccb(Assembler::equal, L); |
kvn@3577 | 929 | // Die if stack mismatch |
kvn@3577 | 930 | __ int3(); |
kvn@3577 | 931 | __ bind(L); |
kvn@3577 | 932 | } |
kvn@3577 | 933 | %} |
kvn@3577 | 934 | |
kvn@3390 | 935 | %} |
kvn@3390 | 936 | |
kvn@3882 | 937 | |
kvn@3882 | 938 | //----------OPERANDS----------------------------------------------------------- |
kvn@3882 | 939 | // Operand definitions must precede instruction definitions for correct parsing |
kvn@3882 | 940 | // in the ADLC because operands constitute user defined types which are used in |
kvn@3882 | 941 | // instruction definitions. |
kvn@3882 | 942 | |
kvn@3882 | 943 | // Vectors |
kvn@3882 | 944 | operand vecS() %{ |
kvn@3882 | 945 | constraint(ALLOC_IN_RC(vectors_reg)); |
kvn@3882 | 946 | match(VecS); |
kvn@3882 | 947 | |
kvn@3882 | 948 | format %{ %} |
kvn@3882 | 949 | interface(REG_INTER); |
kvn@3882 | 950 | %} |
kvn@3882 | 951 | |
kvn@3882 | 952 | operand vecD() %{ |
kvn@3882 | 953 | constraint(ALLOC_IN_RC(vectord_reg)); |
kvn@3882 | 954 | match(VecD); |
kvn@3882 | 955 | |
kvn@3882 | 956 | format %{ %} |
kvn@3882 | 957 | interface(REG_INTER); |
kvn@3882 | 958 | %} |
kvn@3882 | 959 | |
kvn@3882 | 960 | operand vecX() %{ |
kvn@3882 | 961 | constraint(ALLOC_IN_RC(vectorx_reg)); |
kvn@3882 | 962 | match(VecX); |
kvn@3882 | 963 | |
kvn@3882 | 964 | format %{ %} |
kvn@3882 | 965 | interface(REG_INTER); |
kvn@3882 | 966 | %} |
kvn@3882 | 967 | |
kvn@3882 | 968 | operand vecY() %{ |
kvn@3882 | 969 | constraint(ALLOC_IN_RC(vectory_reg)); |
kvn@3882 | 970 | match(VecY); |
kvn@3882 | 971 | |
kvn@3882 | 972 | format %{ %} |
kvn@3882 | 973 | interface(REG_INTER); |
kvn@3882 | 974 | %} |
kvn@3882 | 975 | |
kvn@3882 | 976 | |
kvn@3390 | 977 | // INSTRUCTIONS -- Platform independent definitions (same for 32- and 64-bit) |
kvn@3390 | 978 | |
kvn@3577 | 979 | // ============================================================================ |
kvn@3577 | 980 | |
kvn@3577 | 981 | instruct ShouldNotReachHere() %{ |
kvn@3577 | 982 | match(Halt); |
kvn@3577 | 983 | format %{ "int3\t# ShouldNotReachHere" %} |
kvn@3577 | 984 | ins_encode %{ |
kvn@3577 | 985 | __ int3(); |
kvn@3577 | 986 | %} |
kvn@3577 | 987 | ins_pipe(pipe_slow); |
kvn@3577 | 988 | %} |
kvn@3577 | 989 | |
kvn@3577 | 990 | // ============================================================================ |
kvn@3577 | 991 | |
kvn@3390 | 992 | instruct addF_reg(regF dst, regF src) %{ |
kvn@3390 | 993 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 994 | match(Set dst (AddF dst src)); |
kvn@3390 | 995 | |
kvn@3390 | 996 | format %{ "addss $dst, $src" %} |
kvn@3390 | 997 | ins_cost(150); |
kvn@3390 | 998 | ins_encode %{ |
kvn@3390 | 999 | __ addss($dst$$XMMRegister, $src$$XMMRegister); |
kvn@3390 | 1000 | %} |
kvn@3390 | 1001 | ins_pipe(pipe_slow); |
kvn@3390 | 1002 | %} |
kvn@3390 | 1003 | |
kvn@3390 | 1004 | instruct addF_mem(regF dst, memory src) %{ |
kvn@3390 | 1005 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1006 | match(Set dst (AddF dst (LoadF src))); |
kvn@3390 | 1007 | |
kvn@3390 | 1008 | format %{ "addss $dst, $src" %} |
kvn@3390 | 1009 | ins_cost(150); |
kvn@3390 | 1010 | ins_encode %{ |
kvn@3390 | 1011 | __ addss($dst$$XMMRegister, $src$$Address); |
kvn@3390 | 1012 | %} |
kvn@3390 | 1013 | ins_pipe(pipe_slow); |
kvn@3390 | 1014 | %} |
kvn@3390 | 1015 | |
kvn@3390 | 1016 | instruct addF_imm(regF dst, immF con) %{ |
kvn@3390 | 1017 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1018 | match(Set dst (AddF dst con)); |
kvn@3390 | 1019 | format %{ "addss $dst, [$constantaddress]\t# load from constant table: float=$con" %} |
kvn@3390 | 1020 | ins_cost(150); |
kvn@3390 | 1021 | ins_encode %{ |
kvn@3390 | 1022 | __ addss($dst$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1023 | %} |
kvn@3390 | 1024 | ins_pipe(pipe_slow); |
kvn@3390 | 1025 | %} |
kvn@3390 | 1026 | |
kvn@3929 | 1027 | instruct addF_reg_reg(regF dst, regF src1, regF src2) %{ |
kvn@3390 | 1028 | predicate(UseAVX > 0); |
kvn@3390 | 1029 | match(Set dst (AddF src1 src2)); |
kvn@3390 | 1030 | |
kvn@3390 | 1031 | format %{ "vaddss $dst, $src1, $src2" %} |
kvn@3390 | 1032 | ins_cost(150); |
kvn@3390 | 1033 | ins_encode %{ |
kvn@3390 | 1034 | __ vaddss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister); |
kvn@3390 | 1035 | %} |
kvn@3390 | 1036 | ins_pipe(pipe_slow); |
kvn@3390 | 1037 | %} |
kvn@3390 | 1038 | |
kvn@3929 | 1039 | instruct addF_reg_mem(regF dst, regF src1, memory src2) %{ |
kvn@3390 | 1040 | predicate(UseAVX > 0); |
kvn@3390 | 1041 | match(Set dst (AddF src1 (LoadF src2))); |
kvn@3390 | 1042 | |
kvn@3390 | 1043 | format %{ "vaddss $dst, $src1, $src2" %} |
kvn@3390 | 1044 | ins_cost(150); |
kvn@3390 | 1045 | ins_encode %{ |
kvn@3390 | 1046 | __ vaddss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address); |
kvn@3390 | 1047 | %} |
kvn@3390 | 1048 | ins_pipe(pipe_slow); |
kvn@3390 | 1049 | %} |
kvn@3390 | 1050 | |
kvn@3929 | 1051 | instruct addF_reg_imm(regF dst, regF src, immF con) %{ |
kvn@3390 | 1052 | predicate(UseAVX > 0); |
kvn@3390 | 1053 | match(Set dst (AddF src con)); |
kvn@3390 | 1054 | |
kvn@3390 | 1055 | format %{ "vaddss $dst, $src, [$constantaddress]\t# load from constant table: float=$con" %} |
kvn@3390 | 1056 | ins_cost(150); |
kvn@3390 | 1057 | ins_encode %{ |
kvn@3390 | 1058 | __ vaddss($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1059 | %} |
kvn@3390 | 1060 | ins_pipe(pipe_slow); |
kvn@3390 | 1061 | %} |
kvn@3390 | 1062 | |
kvn@3390 | 1063 | instruct addD_reg(regD dst, regD src) %{ |
kvn@3390 | 1064 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1065 | match(Set dst (AddD dst src)); |
kvn@3390 | 1066 | |
kvn@3390 | 1067 | format %{ "addsd $dst, $src" %} |
kvn@3390 | 1068 | ins_cost(150); |
kvn@3390 | 1069 | ins_encode %{ |
kvn@3390 | 1070 | __ addsd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@3390 | 1071 | %} |
kvn@3390 | 1072 | ins_pipe(pipe_slow); |
kvn@3390 | 1073 | %} |
kvn@3390 | 1074 | |
kvn@3390 | 1075 | instruct addD_mem(regD dst, memory src) %{ |
kvn@3390 | 1076 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1077 | match(Set dst (AddD dst (LoadD src))); |
kvn@3390 | 1078 | |
kvn@3390 | 1079 | format %{ "addsd $dst, $src" %} |
kvn@3390 | 1080 | ins_cost(150); |
kvn@3390 | 1081 | ins_encode %{ |
kvn@3390 | 1082 | __ addsd($dst$$XMMRegister, $src$$Address); |
kvn@3390 | 1083 | %} |
kvn@3390 | 1084 | ins_pipe(pipe_slow); |
kvn@3390 | 1085 | %} |
kvn@3390 | 1086 | |
kvn@3390 | 1087 | instruct addD_imm(regD dst, immD con) %{ |
kvn@3390 | 1088 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1089 | match(Set dst (AddD dst con)); |
kvn@3390 | 1090 | format %{ "addsd $dst, [$constantaddress]\t# load from constant table: double=$con" %} |
kvn@3390 | 1091 | ins_cost(150); |
kvn@3390 | 1092 | ins_encode %{ |
kvn@3390 | 1093 | __ addsd($dst$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1094 | %} |
kvn@3390 | 1095 | ins_pipe(pipe_slow); |
kvn@3390 | 1096 | %} |
kvn@3390 | 1097 | |
kvn@3929 | 1098 | instruct addD_reg_reg(regD dst, regD src1, regD src2) %{ |
kvn@3390 | 1099 | predicate(UseAVX > 0); |
kvn@3390 | 1100 | match(Set dst (AddD src1 src2)); |
kvn@3390 | 1101 | |
kvn@3390 | 1102 | format %{ "vaddsd $dst, $src1, $src2" %} |
kvn@3390 | 1103 | ins_cost(150); |
kvn@3390 | 1104 | ins_encode %{ |
kvn@3390 | 1105 | __ vaddsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister); |
kvn@3390 | 1106 | %} |
kvn@3390 | 1107 | ins_pipe(pipe_slow); |
kvn@3390 | 1108 | %} |
kvn@3390 | 1109 | |
kvn@3929 | 1110 | instruct addD_reg_mem(regD dst, regD src1, memory src2) %{ |
kvn@3390 | 1111 | predicate(UseAVX > 0); |
kvn@3390 | 1112 | match(Set dst (AddD src1 (LoadD src2))); |
kvn@3390 | 1113 | |
kvn@3390 | 1114 | format %{ "vaddsd $dst, $src1, $src2" %} |
kvn@3390 | 1115 | ins_cost(150); |
kvn@3390 | 1116 | ins_encode %{ |
kvn@3390 | 1117 | __ vaddsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address); |
kvn@3390 | 1118 | %} |
kvn@3390 | 1119 | ins_pipe(pipe_slow); |
kvn@3390 | 1120 | %} |
kvn@3390 | 1121 | |
kvn@3929 | 1122 | instruct addD_reg_imm(regD dst, regD src, immD con) %{ |
kvn@3390 | 1123 | predicate(UseAVX > 0); |
kvn@3390 | 1124 | match(Set dst (AddD src con)); |
kvn@3390 | 1125 | |
kvn@3390 | 1126 | format %{ "vaddsd $dst, $src, [$constantaddress]\t# load from constant table: double=$con" %} |
kvn@3390 | 1127 | ins_cost(150); |
kvn@3390 | 1128 | ins_encode %{ |
kvn@3390 | 1129 | __ vaddsd($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1130 | %} |
kvn@3390 | 1131 | ins_pipe(pipe_slow); |
kvn@3390 | 1132 | %} |
kvn@3390 | 1133 | |
kvn@3390 | 1134 | instruct subF_reg(regF dst, regF src) %{ |
kvn@3390 | 1135 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1136 | match(Set dst (SubF dst src)); |
kvn@3390 | 1137 | |
kvn@3390 | 1138 | format %{ "subss $dst, $src" %} |
kvn@3390 | 1139 | ins_cost(150); |
kvn@3390 | 1140 | ins_encode %{ |
kvn@3390 | 1141 | __ subss($dst$$XMMRegister, $src$$XMMRegister); |
kvn@3390 | 1142 | %} |
kvn@3390 | 1143 | ins_pipe(pipe_slow); |
kvn@3390 | 1144 | %} |
kvn@3390 | 1145 | |
kvn@3390 | 1146 | instruct subF_mem(regF dst, memory src) %{ |
kvn@3390 | 1147 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1148 | match(Set dst (SubF dst (LoadF src))); |
kvn@3390 | 1149 | |
kvn@3390 | 1150 | format %{ "subss $dst, $src" %} |
kvn@3390 | 1151 | ins_cost(150); |
kvn@3390 | 1152 | ins_encode %{ |
kvn@3390 | 1153 | __ subss($dst$$XMMRegister, $src$$Address); |
kvn@3390 | 1154 | %} |
kvn@3390 | 1155 | ins_pipe(pipe_slow); |
kvn@3390 | 1156 | %} |
kvn@3390 | 1157 | |
kvn@3390 | 1158 | instruct subF_imm(regF dst, immF con) %{ |
kvn@3390 | 1159 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1160 | match(Set dst (SubF dst con)); |
kvn@3390 | 1161 | format %{ "subss $dst, [$constantaddress]\t# load from constant table: float=$con" %} |
kvn@3390 | 1162 | ins_cost(150); |
kvn@3390 | 1163 | ins_encode %{ |
kvn@3390 | 1164 | __ subss($dst$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1165 | %} |
kvn@3390 | 1166 | ins_pipe(pipe_slow); |
kvn@3390 | 1167 | %} |
kvn@3390 | 1168 | |
kvn@3929 | 1169 | instruct subF_reg_reg(regF dst, regF src1, regF src2) %{ |
kvn@3390 | 1170 | predicate(UseAVX > 0); |
kvn@3390 | 1171 | match(Set dst (SubF src1 src2)); |
kvn@3390 | 1172 | |
kvn@3390 | 1173 | format %{ "vsubss $dst, $src1, $src2" %} |
kvn@3390 | 1174 | ins_cost(150); |
kvn@3390 | 1175 | ins_encode %{ |
kvn@3390 | 1176 | __ vsubss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister); |
kvn@3390 | 1177 | %} |
kvn@3390 | 1178 | ins_pipe(pipe_slow); |
kvn@3390 | 1179 | %} |
kvn@3390 | 1180 | |
kvn@3929 | 1181 | instruct subF_reg_mem(regF dst, regF src1, memory src2) %{ |
kvn@3390 | 1182 | predicate(UseAVX > 0); |
kvn@3390 | 1183 | match(Set dst (SubF src1 (LoadF src2))); |
kvn@3390 | 1184 | |
kvn@3390 | 1185 | format %{ "vsubss $dst, $src1, $src2" %} |
kvn@3390 | 1186 | ins_cost(150); |
kvn@3390 | 1187 | ins_encode %{ |
kvn@3390 | 1188 | __ vsubss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address); |
kvn@3390 | 1189 | %} |
kvn@3390 | 1190 | ins_pipe(pipe_slow); |
kvn@3390 | 1191 | %} |
kvn@3390 | 1192 | |
kvn@3929 | 1193 | instruct subF_reg_imm(regF dst, regF src, immF con) %{ |
kvn@3390 | 1194 | predicate(UseAVX > 0); |
kvn@3390 | 1195 | match(Set dst (SubF src con)); |
kvn@3390 | 1196 | |
kvn@3390 | 1197 | format %{ "vsubss $dst, $src, [$constantaddress]\t# load from constant table: float=$con" %} |
kvn@3390 | 1198 | ins_cost(150); |
kvn@3390 | 1199 | ins_encode %{ |
kvn@3390 | 1200 | __ vsubss($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1201 | %} |
kvn@3390 | 1202 | ins_pipe(pipe_slow); |
kvn@3390 | 1203 | %} |
kvn@3390 | 1204 | |
kvn@3390 | 1205 | instruct subD_reg(regD dst, regD src) %{ |
kvn@3390 | 1206 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1207 | match(Set dst (SubD dst src)); |
kvn@3390 | 1208 | |
kvn@3390 | 1209 | format %{ "subsd $dst, $src" %} |
kvn@3390 | 1210 | ins_cost(150); |
kvn@3390 | 1211 | ins_encode %{ |
kvn@3390 | 1212 | __ subsd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@3390 | 1213 | %} |
kvn@3390 | 1214 | ins_pipe(pipe_slow); |
kvn@3390 | 1215 | %} |
kvn@3390 | 1216 | |
kvn@3390 | 1217 | instruct subD_mem(regD dst, memory src) %{ |
kvn@3390 | 1218 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1219 | match(Set dst (SubD dst (LoadD src))); |
kvn@3390 | 1220 | |
kvn@3390 | 1221 | format %{ "subsd $dst, $src" %} |
kvn@3390 | 1222 | ins_cost(150); |
kvn@3390 | 1223 | ins_encode %{ |
kvn@3390 | 1224 | __ subsd($dst$$XMMRegister, $src$$Address); |
kvn@3390 | 1225 | %} |
kvn@3390 | 1226 | ins_pipe(pipe_slow); |
kvn@3390 | 1227 | %} |
kvn@3390 | 1228 | |
kvn@3390 | 1229 | instruct subD_imm(regD dst, immD con) %{ |
kvn@3390 | 1230 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1231 | match(Set dst (SubD dst con)); |
kvn@3390 | 1232 | format %{ "subsd $dst, [$constantaddress]\t# load from constant table: double=$con" %} |
kvn@3390 | 1233 | ins_cost(150); |
kvn@3390 | 1234 | ins_encode %{ |
kvn@3390 | 1235 | __ subsd($dst$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1236 | %} |
kvn@3390 | 1237 | ins_pipe(pipe_slow); |
kvn@3390 | 1238 | %} |
kvn@3390 | 1239 | |
kvn@3929 | 1240 | instruct subD_reg_reg(regD dst, regD src1, regD src2) %{ |
kvn@3390 | 1241 | predicate(UseAVX > 0); |
kvn@3390 | 1242 | match(Set dst (SubD src1 src2)); |
kvn@3390 | 1243 | |
kvn@3390 | 1244 | format %{ "vsubsd $dst, $src1, $src2" %} |
kvn@3390 | 1245 | ins_cost(150); |
kvn@3390 | 1246 | ins_encode %{ |
kvn@3390 | 1247 | __ vsubsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister); |
kvn@3390 | 1248 | %} |
kvn@3390 | 1249 | ins_pipe(pipe_slow); |
kvn@3390 | 1250 | %} |
kvn@3390 | 1251 | |
kvn@3929 | 1252 | instruct subD_reg_mem(regD dst, regD src1, memory src2) %{ |
kvn@3390 | 1253 | predicate(UseAVX > 0); |
kvn@3390 | 1254 | match(Set dst (SubD src1 (LoadD src2))); |
kvn@3390 | 1255 | |
kvn@3390 | 1256 | format %{ "vsubsd $dst, $src1, $src2" %} |
kvn@3390 | 1257 | ins_cost(150); |
kvn@3390 | 1258 | ins_encode %{ |
kvn@3390 | 1259 | __ vsubsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address); |
kvn@3390 | 1260 | %} |
kvn@3390 | 1261 | ins_pipe(pipe_slow); |
kvn@3390 | 1262 | %} |
kvn@3390 | 1263 | |
kvn@3929 | 1264 | instruct subD_reg_imm(regD dst, regD src, immD con) %{ |
kvn@3390 | 1265 | predicate(UseAVX > 0); |
kvn@3390 | 1266 | match(Set dst (SubD src con)); |
kvn@3390 | 1267 | |
kvn@3390 | 1268 | format %{ "vsubsd $dst, $src, [$constantaddress]\t# load from constant table: double=$con" %} |
kvn@3390 | 1269 | ins_cost(150); |
kvn@3390 | 1270 | ins_encode %{ |
kvn@3390 | 1271 | __ vsubsd($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1272 | %} |
kvn@3390 | 1273 | ins_pipe(pipe_slow); |
kvn@3390 | 1274 | %} |
kvn@3390 | 1275 | |
kvn@3390 | 1276 | instruct mulF_reg(regF dst, regF src) %{ |
kvn@3390 | 1277 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1278 | match(Set dst (MulF dst src)); |
kvn@3390 | 1279 | |
kvn@3390 | 1280 | format %{ "mulss $dst, $src" %} |
kvn@3390 | 1281 | ins_cost(150); |
kvn@3390 | 1282 | ins_encode %{ |
kvn@3390 | 1283 | __ mulss($dst$$XMMRegister, $src$$XMMRegister); |
kvn@3390 | 1284 | %} |
kvn@3390 | 1285 | ins_pipe(pipe_slow); |
kvn@3390 | 1286 | %} |
kvn@3390 | 1287 | |
kvn@3390 | 1288 | instruct mulF_mem(regF dst, memory src) %{ |
kvn@3390 | 1289 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1290 | match(Set dst (MulF dst (LoadF src))); |
kvn@3390 | 1291 | |
kvn@3390 | 1292 | format %{ "mulss $dst, $src" %} |
kvn@3390 | 1293 | ins_cost(150); |
kvn@3390 | 1294 | ins_encode %{ |
kvn@3390 | 1295 | __ mulss($dst$$XMMRegister, $src$$Address); |
kvn@3390 | 1296 | %} |
kvn@3390 | 1297 | ins_pipe(pipe_slow); |
kvn@3390 | 1298 | %} |
kvn@3390 | 1299 | |
kvn@3390 | 1300 | instruct mulF_imm(regF dst, immF con) %{ |
kvn@3390 | 1301 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1302 | match(Set dst (MulF dst con)); |
kvn@3390 | 1303 | format %{ "mulss $dst, [$constantaddress]\t# load from constant table: float=$con" %} |
kvn@3390 | 1304 | ins_cost(150); |
kvn@3390 | 1305 | ins_encode %{ |
kvn@3390 | 1306 | __ mulss($dst$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1307 | %} |
kvn@3390 | 1308 | ins_pipe(pipe_slow); |
kvn@3390 | 1309 | %} |
kvn@3390 | 1310 | |
kvn@3929 | 1311 | instruct mulF_reg_reg(regF dst, regF src1, regF src2) %{ |
kvn@3390 | 1312 | predicate(UseAVX > 0); |
kvn@3390 | 1313 | match(Set dst (MulF src1 src2)); |
kvn@3390 | 1314 | |
kvn@3390 | 1315 | format %{ "vmulss $dst, $src1, $src2" %} |
kvn@3390 | 1316 | ins_cost(150); |
kvn@3390 | 1317 | ins_encode %{ |
kvn@3390 | 1318 | __ vmulss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister); |
kvn@3390 | 1319 | %} |
kvn@3390 | 1320 | ins_pipe(pipe_slow); |
kvn@3390 | 1321 | %} |
kvn@3390 | 1322 | |
kvn@3929 | 1323 | instruct mulF_reg_mem(regF dst, regF src1, memory src2) %{ |
kvn@3390 | 1324 | predicate(UseAVX > 0); |
kvn@3390 | 1325 | match(Set dst (MulF src1 (LoadF src2))); |
kvn@3390 | 1326 | |
kvn@3390 | 1327 | format %{ "vmulss $dst, $src1, $src2" %} |
kvn@3390 | 1328 | ins_cost(150); |
kvn@3390 | 1329 | ins_encode %{ |
kvn@3390 | 1330 | __ vmulss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address); |
kvn@3390 | 1331 | %} |
kvn@3390 | 1332 | ins_pipe(pipe_slow); |
kvn@3390 | 1333 | %} |
kvn@3390 | 1334 | |
kvn@3929 | 1335 | instruct mulF_reg_imm(regF dst, regF src, immF con) %{ |
kvn@3390 | 1336 | predicate(UseAVX > 0); |
kvn@3390 | 1337 | match(Set dst (MulF src con)); |
kvn@3390 | 1338 | |
kvn@3390 | 1339 | format %{ "vmulss $dst, $src, [$constantaddress]\t# load from constant table: float=$con" %} |
kvn@3390 | 1340 | ins_cost(150); |
kvn@3390 | 1341 | ins_encode %{ |
kvn@3390 | 1342 | __ vmulss($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1343 | %} |
kvn@3390 | 1344 | ins_pipe(pipe_slow); |
kvn@3390 | 1345 | %} |
kvn@3390 | 1346 | |
kvn@3390 | 1347 | instruct mulD_reg(regD dst, regD src) %{ |
kvn@3390 | 1348 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1349 | match(Set dst (MulD dst src)); |
kvn@3390 | 1350 | |
kvn@3390 | 1351 | format %{ "mulsd $dst, $src" %} |
kvn@3390 | 1352 | ins_cost(150); |
kvn@3390 | 1353 | ins_encode %{ |
kvn@3390 | 1354 | __ mulsd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@3390 | 1355 | %} |
kvn@3390 | 1356 | ins_pipe(pipe_slow); |
kvn@3390 | 1357 | %} |
kvn@3390 | 1358 | |
kvn@3390 | 1359 | instruct mulD_mem(regD dst, memory src) %{ |
kvn@3390 | 1360 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1361 | match(Set dst (MulD dst (LoadD src))); |
kvn@3390 | 1362 | |
kvn@3390 | 1363 | format %{ "mulsd $dst, $src" %} |
kvn@3390 | 1364 | ins_cost(150); |
kvn@3390 | 1365 | ins_encode %{ |
kvn@3390 | 1366 | __ mulsd($dst$$XMMRegister, $src$$Address); |
kvn@3390 | 1367 | %} |
kvn@3390 | 1368 | ins_pipe(pipe_slow); |
kvn@3390 | 1369 | %} |
kvn@3390 | 1370 | |
kvn@3390 | 1371 | instruct mulD_imm(regD dst, immD con) %{ |
kvn@3390 | 1372 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1373 | match(Set dst (MulD dst con)); |
kvn@3390 | 1374 | format %{ "mulsd $dst, [$constantaddress]\t# load from constant table: double=$con" %} |
kvn@3390 | 1375 | ins_cost(150); |
kvn@3390 | 1376 | ins_encode %{ |
kvn@3390 | 1377 | __ mulsd($dst$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1378 | %} |
kvn@3390 | 1379 | ins_pipe(pipe_slow); |
kvn@3390 | 1380 | %} |
kvn@3390 | 1381 | |
kvn@3929 | 1382 | instruct mulD_reg_reg(regD dst, regD src1, regD src2) %{ |
kvn@3390 | 1383 | predicate(UseAVX > 0); |
kvn@3390 | 1384 | match(Set dst (MulD src1 src2)); |
kvn@3390 | 1385 | |
kvn@3390 | 1386 | format %{ "vmulsd $dst, $src1, $src2" %} |
kvn@3390 | 1387 | ins_cost(150); |
kvn@3390 | 1388 | ins_encode %{ |
kvn@3390 | 1389 | __ vmulsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister); |
kvn@3390 | 1390 | %} |
kvn@3390 | 1391 | ins_pipe(pipe_slow); |
kvn@3390 | 1392 | %} |
kvn@3390 | 1393 | |
kvn@3929 | 1394 | instruct mulD_reg_mem(regD dst, regD src1, memory src2) %{ |
kvn@3390 | 1395 | predicate(UseAVX > 0); |
kvn@3390 | 1396 | match(Set dst (MulD src1 (LoadD src2))); |
kvn@3390 | 1397 | |
kvn@3390 | 1398 | format %{ "vmulsd $dst, $src1, $src2" %} |
kvn@3390 | 1399 | ins_cost(150); |
kvn@3390 | 1400 | ins_encode %{ |
kvn@3390 | 1401 | __ vmulsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address); |
kvn@3390 | 1402 | %} |
kvn@3390 | 1403 | ins_pipe(pipe_slow); |
kvn@3390 | 1404 | %} |
kvn@3390 | 1405 | |
kvn@3929 | 1406 | instruct mulD_reg_imm(regD dst, regD src, immD con) %{ |
kvn@3390 | 1407 | predicate(UseAVX > 0); |
kvn@3390 | 1408 | match(Set dst (MulD src con)); |
kvn@3390 | 1409 | |
kvn@3390 | 1410 | format %{ "vmulsd $dst, $src, [$constantaddress]\t# load from constant table: double=$con" %} |
kvn@3390 | 1411 | ins_cost(150); |
kvn@3390 | 1412 | ins_encode %{ |
kvn@3390 | 1413 | __ vmulsd($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1414 | %} |
kvn@3390 | 1415 | ins_pipe(pipe_slow); |
kvn@3390 | 1416 | %} |
kvn@3390 | 1417 | |
kvn@3390 | 1418 | instruct divF_reg(regF dst, regF src) %{ |
kvn@3390 | 1419 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1420 | match(Set dst (DivF dst src)); |
kvn@3390 | 1421 | |
kvn@3390 | 1422 | format %{ "divss $dst, $src" %} |
kvn@3390 | 1423 | ins_cost(150); |
kvn@3390 | 1424 | ins_encode %{ |
kvn@3390 | 1425 | __ divss($dst$$XMMRegister, $src$$XMMRegister); |
kvn@3390 | 1426 | %} |
kvn@3390 | 1427 | ins_pipe(pipe_slow); |
kvn@3390 | 1428 | %} |
kvn@3390 | 1429 | |
kvn@3390 | 1430 | instruct divF_mem(regF dst, memory src) %{ |
kvn@3390 | 1431 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1432 | match(Set dst (DivF dst (LoadF src))); |
kvn@3390 | 1433 | |
kvn@3390 | 1434 | format %{ "divss $dst, $src" %} |
kvn@3390 | 1435 | ins_cost(150); |
kvn@3390 | 1436 | ins_encode %{ |
kvn@3390 | 1437 | __ divss($dst$$XMMRegister, $src$$Address); |
kvn@3390 | 1438 | %} |
kvn@3390 | 1439 | ins_pipe(pipe_slow); |
kvn@3390 | 1440 | %} |
kvn@3390 | 1441 | |
kvn@3390 | 1442 | instruct divF_imm(regF dst, immF con) %{ |
kvn@3390 | 1443 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1444 | match(Set dst (DivF dst con)); |
kvn@3390 | 1445 | format %{ "divss $dst, [$constantaddress]\t# load from constant table: float=$con" %} |
kvn@3390 | 1446 | ins_cost(150); |
kvn@3390 | 1447 | ins_encode %{ |
kvn@3390 | 1448 | __ divss($dst$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1449 | %} |
kvn@3390 | 1450 | ins_pipe(pipe_slow); |
kvn@3390 | 1451 | %} |
kvn@3390 | 1452 | |
kvn@3929 | 1453 | instruct divF_reg_reg(regF dst, regF src1, regF src2) %{ |
kvn@3390 | 1454 | predicate(UseAVX > 0); |
kvn@3390 | 1455 | match(Set dst (DivF src1 src2)); |
kvn@3390 | 1456 | |
kvn@3390 | 1457 | format %{ "vdivss $dst, $src1, $src2" %} |
kvn@3390 | 1458 | ins_cost(150); |
kvn@3390 | 1459 | ins_encode %{ |
kvn@3390 | 1460 | __ vdivss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister); |
kvn@3390 | 1461 | %} |
kvn@3390 | 1462 | ins_pipe(pipe_slow); |
kvn@3390 | 1463 | %} |
kvn@3390 | 1464 | |
kvn@3929 | 1465 | instruct divF_reg_mem(regF dst, regF src1, memory src2) %{ |
kvn@3390 | 1466 | predicate(UseAVX > 0); |
kvn@3390 | 1467 | match(Set dst (DivF src1 (LoadF src2))); |
kvn@3390 | 1468 | |
kvn@3390 | 1469 | format %{ "vdivss $dst, $src1, $src2" %} |
kvn@3390 | 1470 | ins_cost(150); |
kvn@3390 | 1471 | ins_encode %{ |
kvn@3390 | 1472 | __ vdivss($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address); |
kvn@3390 | 1473 | %} |
kvn@3390 | 1474 | ins_pipe(pipe_slow); |
kvn@3390 | 1475 | %} |
kvn@3390 | 1476 | |
kvn@3929 | 1477 | instruct divF_reg_imm(regF dst, regF src, immF con) %{ |
kvn@3390 | 1478 | predicate(UseAVX > 0); |
kvn@3390 | 1479 | match(Set dst (DivF src con)); |
kvn@3390 | 1480 | |
kvn@3390 | 1481 | format %{ "vdivss $dst, $src, [$constantaddress]\t# load from constant table: float=$con" %} |
kvn@3390 | 1482 | ins_cost(150); |
kvn@3390 | 1483 | ins_encode %{ |
kvn@3390 | 1484 | __ vdivss($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1485 | %} |
kvn@3390 | 1486 | ins_pipe(pipe_slow); |
kvn@3390 | 1487 | %} |
kvn@3390 | 1488 | |
kvn@3390 | 1489 | instruct divD_reg(regD dst, regD src) %{ |
kvn@3390 | 1490 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1491 | match(Set dst (DivD dst src)); |
kvn@3390 | 1492 | |
kvn@3390 | 1493 | format %{ "divsd $dst, $src" %} |
kvn@3390 | 1494 | ins_cost(150); |
kvn@3390 | 1495 | ins_encode %{ |
kvn@3390 | 1496 | __ divsd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@3390 | 1497 | %} |
kvn@3390 | 1498 | ins_pipe(pipe_slow); |
kvn@3390 | 1499 | %} |
kvn@3390 | 1500 | |
kvn@3390 | 1501 | instruct divD_mem(regD dst, memory src) %{ |
kvn@3390 | 1502 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1503 | match(Set dst (DivD dst (LoadD src))); |
kvn@3390 | 1504 | |
kvn@3390 | 1505 | format %{ "divsd $dst, $src" %} |
kvn@3390 | 1506 | ins_cost(150); |
kvn@3390 | 1507 | ins_encode %{ |
kvn@3390 | 1508 | __ divsd($dst$$XMMRegister, $src$$Address); |
kvn@3390 | 1509 | %} |
kvn@3390 | 1510 | ins_pipe(pipe_slow); |
kvn@3390 | 1511 | %} |
kvn@3390 | 1512 | |
kvn@3390 | 1513 | instruct divD_imm(regD dst, immD con) %{ |
kvn@3390 | 1514 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1515 | match(Set dst (DivD dst con)); |
kvn@3390 | 1516 | format %{ "divsd $dst, [$constantaddress]\t# load from constant table: double=$con" %} |
kvn@3390 | 1517 | ins_cost(150); |
kvn@3390 | 1518 | ins_encode %{ |
kvn@3390 | 1519 | __ divsd($dst$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1520 | %} |
kvn@3390 | 1521 | ins_pipe(pipe_slow); |
kvn@3390 | 1522 | %} |
kvn@3390 | 1523 | |
kvn@3929 | 1524 | instruct divD_reg_reg(regD dst, regD src1, regD src2) %{ |
kvn@3390 | 1525 | predicate(UseAVX > 0); |
kvn@3390 | 1526 | match(Set dst (DivD src1 src2)); |
kvn@3390 | 1527 | |
kvn@3390 | 1528 | format %{ "vdivsd $dst, $src1, $src2" %} |
kvn@3390 | 1529 | ins_cost(150); |
kvn@3390 | 1530 | ins_encode %{ |
kvn@3390 | 1531 | __ vdivsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister); |
kvn@3390 | 1532 | %} |
kvn@3390 | 1533 | ins_pipe(pipe_slow); |
kvn@3390 | 1534 | %} |
kvn@3390 | 1535 | |
kvn@3929 | 1536 | instruct divD_reg_mem(regD dst, regD src1, memory src2) %{ |
kvn@3390 | 1537 | predicate(UseAVX > 0); |
kvn@3390 | 1538 | match(Set dst (DivD src1 (LoadD src2))); |
kvn@3390 | 1539 | |
kvn@3390 | 1540 | format %{ "vdivsd $dst, $src1, $src2" %} |
kvn@3390 | 1541 | ins_cost(150); |
kvn@3390 | 1542 | ins_encode %{ |
kvn@3390 | 1543 | __ vdivsd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$Address); |
kvn@3390 | 1544 | %} |
kvn@3390 | 1545 | ins_pipe(pipe_slow); |
kvn@3390 | 1546 | %} |
kvn@3390 | 1547 | |
kvn@3929 | 1548 | instruct divD_reg_imm(regD dst, regD src, immD con) %{ |
kvn@3390 | 1549 | predicate(UseAVX > 0); |
kvn@3390 | 1550 | match(Set dst (DivD src con)); |
kvn@3390 | 1551 | |
kvn@3390 | 1552 | format %{ "vdivsd $dst, $src, [$constantaddress]\t# load from constant table: double=$con" %} |
kvn@3390 | 1553 | ins_cost(150); |
kvn@3390 | 1554 | ins_encode %{ |
kvn@3390 | 1555 | __ vdivsd($dst$$XMMRegister, $src$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1556 | %} |
kvn@3390 | 1557 | ins_pipe(pipe_slow); |
kvn@3390 | 1558 | %} |
kvn@3390 | 1559 | |
kvn@3390 | 1560 | instruct absF_reg(regF dst) %{ |
kvn@3390 | 1561 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1562 | match(Set dst (AbsF dst)); |
kvn@3390 | 1563 | ins_cost(150); |
kvn@3390 | 1564 | format %{ "andps $dst, [0x7fffffff]\t# abs float by sign masking" %} |
kvn@3390 | 1565 | ins_encode %{ |
kvn@3390 | 1566 | __ andps($dst$$XMMRegister, ExternalAddress(float_signmask())); |
kvn@3390 | 1567 | %} |
kvn@3390 | 1568 | ins_pipe(pipe_slow); |
kvn@3390 | 1569 | %} |
kvn@3390 | 1570 | |
kvn@3929 | 1571 | instruct absF_reg_reg(regF dst, regF src) %{ |
kvn@3390 | 1572 | predicate(UseAVX > 0); |
kvn@3390 | 1573 | match(Set dst (AbsF src)); |
kvn@3390 | 1574 | ins_cost(150); |
kvn@3390 | 1575 | format %{ "vandps $dst, $src, [0x7fffffff]\t# abs float by sign masking" %} |
kvn@3390 | 1576 | ins_encode %{ |
kvn@4001 | 1577 | bool vector256 = false; |
kvn@3390 | 1578 | __ vandps($dst$$XMMRegister, $src$$XMMRegister, |
kvn@4001 | 1579 | ExternalAddress(float_signmask()), vector256); |
kvn@3390 | 1580 | %} |
kvn@3390 | 1581 | ins_pipe(pipe_slow); |
kvn@3390 | 1582 | %} |
kvn@3390 | 1583 | |
kvn@3390 | 1584 | instruct absD_reg(regD dst) %{ |
kvn@3390 | 1585 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1586 | match(Set dst (AbsD dst)); |
kvn@3390 | 1587 | ins_cost(150); |
kvn@3390 | 1588 | format %{ "andpd $dst, [0x7fffffffffffffff]\t" |
kvn@3390 | 1589 | "# abs double by sign masking" %} |
kvn@3390 | 1590 | ins_encode %{ |
kvn@3390 | 1591 | __ andpd($dst$$XMMRegister, ExternalAddress(double_signmask())); |
kvn@3390 | 1592 | %} |
kvn@3390 | 1593 | ins_pipe(pipe_slow); |
kvn@3390 | 1594 | %} |
kvn@3390 | 1595 | |
kvn@3929 | 1596 | instruct absD_reg_reg(regD dst, regD src) %{ |
kvn@3390 | 1597 | predicate(UseAVX > 0); |
kvn@3390 | 1598 | match(Set dst (AbsD src)); |
kvn@3390 | 1599 | ins_cost(150); |
kvn@3390 | 1600 | format %{ "vandpd $dst, $src, [0x7fffffffffffffff]\t" |
kvn@3390 | 1601 | "# abs double by sign masking" %} |
kvn@3390 | 1602 | ins_encode %{ |
kvn@4001 | 1603 | bool vector256 = false; |
kvn@3390 | 1604 | __ vandpd($dst$$XMMRegister, $src$$XMMRegister, |
kvn@4001 | 1605 | ExternalAddress(double_signmask()), vector256); |
kvn@3390 | 1606 | %} |
kvn@3390 | 1607 | ins_pipe(pipe_slow); |
kvn@3390 | 1608 | %} |
kvn@3390 | 1609 | |
kvn@3390 | 1610 | instruct negF_reg(regF dst) %{ |
kvn@3390 | 1611 | predicate((UseSSE>=1) && (UseAVX == 0)); |
kvn@3390 | 1612 | match(Set dst (NegF dst)); |
kvn@3390 | 1613 | ins_cost(150); |
kvn@3390 | 1614 | format %{ "xorps $dst, [0x80000000]\t# neg float by sign flipping" %} |
kvn@3390 | 1615 | ins_encode %{ |
kvn@3390 | 1616 | __ xorps($dst$$XMMRegister, ExternalAddress(float_signflip())); |
kvn@3390 | 1617 | %} |
kvn@3390 | 1618 | ins_pipe(pipe_slow); |
kvn@3390 | 1619 | %} |
kvn@3390 | 1620 | |
kvn@3929 | 1621 | instruct negF_reg_reg(regF dst, regF src) %{ |
kvn@3390 | 1622 | predicate(UseAVX > 0); |
kvn@3390 | 1623 | match(Set dst (NegF src)); |
kvn@3390 | 1624 | ins_cost(150); |
kvn@3390 | 1625 | format %{ "vxorps $dst, $src, [0x80000000]\t# neg float by sign flipping" %} |
kvn@3390 | 1626 | ins_encode %{ |
kvn@4001 | 1627 | bool vector256 = false; |
kvn@3390 | 1628 | __ vxorps($dst$$XMMRegister, $src$$XMMRegister, |
kvn@4001 | 1629 | ExternalAddress(float_signflip()), vector256); |
kvn@3390 | 1630 | %} |
kvn@3390 | 1631 | ins_pipe(pipe_slow); |
kvn@3390 | 1632 | %} |
kvn@3390 | 1633 | |
kvn@3390 | 1634 | instruct negD_reg(regD dst) %{ |
kvn@3390 | 1635 | predicate((UseSSE>=2) && (UseAVX == 0)); |
kvn@3390 | 1636 | match(Set dst (NegD dst)); |
kvn@3390 | 1637 | ins_cost(150); |
kvn@3390 | 1638 | format %{ "xorpd $dst, [0x8000000000000000]\t" |
kvn@3390 | 1639 | "# neg double by sign flipping" %} |
kvn@3390 | 1640 | ins_encode %{ |
kvn@3390 | 1641 | __ xorpd($dst$$XMMRegister, ExternalAddress(double_signflip())); |
kvn@3390 | 1642 | %} |
kvn@3390 | 1643 | ins_pipe(pipe_slow); |
kvn@3390 | 1644 | %} |
kvn@3390 | 1645 | |
kvn@3929 | 1646 | instruct negD_reg_reg(regD dst, regD src) %{ |
kvn@3390 | 1647 | predicate(UseAVX > 0); |
kvn@3390 | 1648 | match(Set dst (NegD src)); |
kvn@3390 | 1649 | ins_cost(150); |
kvn@3390 | 1650 | format %{ "vxorpd $dst, $src, [0x8000000000000000]\t" |
kvn@3390 | 1651 | "# neg double by sign flipping" %} |
kvn@3390 | 1652 | ins_encode %{ |
kvn@4001 | 1653 | bool vector256 = false; |
kvn@3390 | 1654 | __ vxorpd($dst$$XMMRegister, $src$$XMMRegister, |
kvn@4001 | 1655 | ExternalAddress(double_signflip()), vector256); |
kvn@3390 | 1656 | %} |
kvn@3390 | 1657 | ins_pipe(pipe_slow); |
kvn@3390 | 1658 | %} |
kvn@3390 | 1659 | |
kvn@3390 | 1660 | instruct sqrtF_reg(regF dst, regF src) %{ |
kvn@3390 | 1661 | predicate(UseSSE>=1); |
kvn@3390 | 1662 | match(Set dst (ConvD2F (SqrtD (ConvF2D src)))); |
kvn@3390 | 1663 | |
kvn@3390 | 1664 | format %{ "sqrtss $dst, $src" %} |
kvn@3390 | 1665 | ins_cost(150); |
kvn@3390 | 1666 | ins_encode %{ |
kvn@3390 | 1667 | __ sqrtss($dst$$XMMRegister, $src$$XMMRegister); |
kvn@3390 | 1668 | %} |
kvn@3390 | 1669 | ins_pipe(pipe_slow); |
kvn@3390 | 1670 | %} |
kvn@3390 | 1671 | |
kvn@3390 | 1672 | instruct sqrtF_mem(regF dst, memory src) %{ |
kvn@3390 | 1673 | predicate(UseSSE>=1); |
kvn@3390 | 1674 | match(Set dst (ConvD2F (SqrtD (ConvF2D (LoadF src))))); |
kvn@3390 | 1675 | |
kvn@3390 | 1676 | format %{ "sqrtss $dst, $src" %} |
kvn@3390 | 1677 | ins_cost(150); |
kvn@3390 | 1678 | ins_encode %{ |
kvn@3390 | 1679 | __ sqrtss($dst$$XMMRegister, $src$$Address); |
kvn@3390 | 1680 | %} |
kvn@3390 | 1681 | ins_pipe(pipe_slow); |
kvn@3390 | 1682 | %} |
kvn@3390 | 1683 | |
kvn@3390 | 1684 | instruct sqrtF_imm(regF dst, immF con) %{ |
kvn@3390 | 1685 | predicate(UseSSE>=1); |
kvn@3390 | 1686 | match(Set dst (ConvD2F (SqrtD (ConvF2D con)))); |
kvn@3390 | 1687 | format %{ "sqrtss $dst, [$constantaddress]\t# load from constant table: float=$con" %} |
kvn@3390 | 1688 | ins_cost(150); |
kvn@3390 | 1689 | ins_encode %{ |
kvn@3390 | 1690 | __ sqrtss($dst$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1691 | %} |
kvn@3390 | 1692 | ins_pipe(pipe_slow); |
kvn@3390 | 1693 | %} |
kvn@3390 | 1694 | |
kvn@3390 | 1695 | instruct sqrtD_reg(regD dst, regD src) %{ |
kvn@3390 | 1696 | predicate(UseSSE>=2); |
kvn@3390 | 1697 | match(Set dst (SqrtD src)); |
kvn@3390 | 1698 | |
kvn@3390 | 1699 | format %{ "sqrtsd $dst, $src" %} |
kvn@3390 | 1700 | ins_cost(150); |
kvn@3390 | 1701 | ins_encode %{ |
kvn@3390 | 1702 | __ sqrtsd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@3390 | 1703 | %} |
kvn@3390 | 1704 | ins_pipe(pipe_slow); |
kvn@3390 | 1705 | %} |
kvn@3390 | 1706 | |
kvn@3390 | 1707 | instruct sqrtD_mem(regD dst, memory src) %{ |
kvn@3390 | 1708 | predicate(UseSSE>=2); |
kvn@3390 | 1709 | match(Set dst (SqrtD (LoadD src))); |
kvn@3390 | 1710 | |
kvn@3390 | 1711 | format %{ "sqrtsd $dst, $src" %} |
kvn@3390 | 1712 | ins_cost(150); |
kvn@3390 | 1713 | ins_encode %{ |
kvn@3390 | 1714 | __ sqrtsd($dst$$XMMRegister, $src$$Address); |
kvn@3390 | 1715 | %} |
kvn@3390 | 1716 | ins_pipe(pipe_slow); |
kvn@3390 | 1717 | %} |
kvn@3390 | 1718 | |
kvn@3390 | 1719 | instruct sqrtD_imm(regD dst, immD con) %{ |
kvn@3390 | 1720 | predicate(UseSSE>=2); |
kvn@3390 | 1721 | match(Set dst (SqrtD con)); |
kvn@3390 | 1722 | format %{ "sqrtsd $dst, [$constantaddress]\t# load from constant table: double=$con" %} |
kvn@3390 | 1723 | ins_cost(150); |
kvn@3390 | 1724 | ins_encode %{ |
kvn@3390 | 1725 | __ sqrtsd($dst$$XMMRegister, $constantaddress($con)); |
kvn@3390 | 1726 | %} |
kvn@3390 | 1727 | ins_pipe(pipe_slow); |
kvn@3390 | 1728 | %} |
kvn@3390 | 1729 | |
kvn@3882 | 1730 | |
kvn@3882 | 1731 | // ====================VECTOR INSTRUCTIONS===================================== |
kvn@3882 | 1732 | |
kvn@3882 | 1733 | // Load vectors (4 bytes long) |
kvn@3882 | 1734 | instruct loadV4(vecS dst, memory mem) %{ |
kvn@3882 | 1735 | predicate(n->as_LoadVector()->memory_size() == 4); |
kvn@3882 | 1736 | match(Set dst (LoadVector mem)); |
kvn@3882 | 1737 | ins_cost(125); |
kvn@3882 | 1738 | format %{ "movd $dst,$mem\t! load vector (4 bytes)" %} |
kvn@3882 | 1739 | ins_encode %{ |
kvn@3882 | 1740 | __ movdl($dst$$XMMRegister, $mem$$Address); |
kvn@3882 | 1741 | %} |
kvn@3882 | 1742 | ins_pipe( pipe_slow ); |
kvn@3882 | 1743 | %} |
kvn@3882 | 1744 | |
kvn@3882 | 1745 | // Load vectors (8 bytes long) |
kvn@3882 | 1746 | instruct loadV8(vecD dst, memory mem) %{ |
kvn@3882 | 1747 | predicate(n->as_LoadVector()->memory_size() == 8); |
kvn@3882 | 1748 | match(Set dst (LoadVector mem)); |
kvn@3882 | 1749 | ins_cost(125); |
kvn@3882 | 1750 | format %{ "movq $dst,$mem\t! load vector (8 bytes)" %} |
kvn@3882 | 1751 | ins_encode %{ |
kvn@3882 | 1752 | __ movq($dst$$XMMRegister, $mem$$Address); |
kvn@3882 | 1753 | %} |
kvn@3882 | 1754 | ins_pipe( pipe_slow ); |
kvn@3882 | 1755 | %} |
kvn@3882 | 1756 | |
kvn@3882 | 1757 | // Load vectors (16 bytes long) |
kvn@3882 | 1758 | instruct loadV16(vecX dst, memory mem) %{ |
kvn@3882 | 1759 | predicate(n->as_LoadVector()->memory_size() == 16); |
kvn@3882 | 1760 | match(Set dst (LoadVector mem)); |
kvn@3882 | 1761 | ins_cost(125); |
kvn@3882 | 1762 | format %{ "movdqu $dst,$mem\t! load vector (16 bytes)" %} |
kvn@3882 | 1763 | ins_encode %{ |
kvn@3882 | 1764 | __ movdqu($dst$$XMMRegister, $mem$$Address); |
kvn@3882 | 1765 | %} |
kvn@3882 | 1766 | ins_pipe( pipe_slow ); |
kvn@3882 | 1767 | %} |
kvn@3882 | 1768 | |
kvn@3882 | 1769 | // Load vectors (32 bytes long) |
kvn@3882 | 1770 | instruct loadV32(vecY dst, memory mem) %{ |
kvn@3882 | 1771 | predicate(n->as_LoadVector()->memory_size() == 32); |
kvn@3882 | 1772 | match(Set dst (LoadVector mem)); |
kvn@3882 | 1773 | ins_cost(125); |
kvn@3882 | 1774 | format %{ "vmovdqu $dst,$mem\t! load vector (32 bytes)" %} |
kvn@3882 | 1775 | ins_encode %{ |
kvn@3882 | 1776 | __ vmovdqu($dst$$XMMRegister, $mem$$Address); |
kvn@3882 | 1777 | %} |
kvn@3882 | 1778 | ins_pipe( pipe_slow ); |
kvn@3882 | 1779 | %} |
kvn@3882 | 1780 | |
kvn@3882 | 1781 | // Store vectors |
kvn@3882 | 1782 | instruct storeV4(memory mem, vecS src) %{ |
kvn@3882 | 1783 | predicate(n->as_StoreVector()->memory_size() == 4); |
kvn@3882 | 1784 | match(Set mem (StoreVector mem src)); |
kvn@3882 | 1785 | ins_cost(145); |
kvn@3882 | 1786 | format %{ "movd $mem,$src\t! store vector (4 bytes)" %} |
kvn@3882 | 1787 | ins_encode %{ |
kvn@3882 | 1788 | __ movdl($mem$$Address, $src$$XMMRegister); |
kvn@3882 | 1789 | %} |
kvn@3882 | 1790 | ins_pipe( pipe_slow ); |
kvn@3882 | 1791 | %} |
kvn@3882 | 1792 | |
kvn@3882 | 1793 | instruct storeV8(memory mem, vecD src) %{ |
kvn@3882 | 1794 | predicate(n->as_StoreVector()->memory_size() == 8); |
kvn@3882 | 1795 | match(Set mem (StoreVector mem src)); |
kvn@3882 | 1796 | ins_cost(145); |
kvn@3882 | 1797 | format %{ "movq $mem,$src\t! store vector (8 bytes)" %} |
kvn@3882 | 1798 | ins_encode %{ |
kvn@3882 | 1799 | __ movq($mem$$Address, $src$$XMMRegister); |
kvn@3882 | 1800 | %} |
kvn@3882 | 1801 | ins_pipe( pipe_slow ); |
kvn@3882 | 1802 | %} |
kvn@3882 | 1803 | |
kvn@3882 | 1804 | instruct storeV16(memory mem, vecX src) %{ |
kvn@3882 | 1805 | predicate(n->as_StoreVector()->memory_size() == 16); |
kvn@3882 | 1806 | match(Set mem (StoreVector mem src)); |
kvn@3882 | 1807 | ins_cost(145); |
kvn@3882 | 1808 | format %{ "movdqu $mem,$src\t! store vector (16 bytes)" %} |
kvn@3882 | 1809 | ins_encode %{ |
kvn@3882 | 1810 | __ movdqu($mem$$Address, $src$$XMMRegister); |
kvn@3882 | 1811 | %} |
kvn@3882 | 1812 | ins_pipe( pipe_slow ); |
kvn@3882 | 1813 | %} |
kvn@3882 | 1814 | |
kvn@3882 | 1815 | instruct storeV32(memory mem, vecY src) %{ |
kvn@3882 | 1816 | predicate(n->as_StoreVector()->memory_size() == 32); |
kvn@3882 | 1817 | match(Set mem (StoreVector mem src)); |
kvn@3882 | 1818 | ins_cost(145); |
kvn@3882 | 1819 | format %{ "vmovdqu $mem,$src\t! store vector (32 bytes)" %} |
kvn@3882 | 1820 | ins_encode %{ |
kvn@3882 | 1821 | __ vmovdqu($mem$$Address, $src$$XMMRegister); |
kvn@3882 | 1822 | %} |
kvn@3882 | 1823 | ins_pipe( pipe_slow ); |
kvn@3882 | 1824 | %} |
kvn@3882 | 1825 | |
kvn@3882 | 1826 | // Replicate byte scalar to be vector |
kvn@3882 | 1827 | instruct Repl4B(vecS dst, rRegI src) %{ |
kvn@3882 | 1828 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 1829 | match(Set dst (ReplicateB src)); |
kvn@3882 | 1830 | format %{ "movd $dst,$src\n\t" |
kvn@3882 | 1831 | "punpcklbw $dst,$dst\n\t" |
kvn@3882 | 1832 | "pshuflw $dst,$dst,0x00\t! replicate4B" %} |
kvn@3882 | 1833 | ins_encode %{ |
kvn@3882 | 1834 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 1835 | __ punpcklbw($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 1836 | __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3882 | 1837 | %} |
kvn@3882 | 1838 | ins_pipe( pipe_slow ); |
kvn@3882 | 1839 | %} |
kvn@3882 | 1840 | |
kvn@3882 | 1841 | instruct Repl8B(vecD dst, rRegI src) %{ |
kvn@3882 | 1842 | predicate(n->as_Vector()->length() == 8); |
kvn@3882 | 1843 | match(Set dst (ReplicateB src)); |
kvn@3882 | 1844 | format %{ "movd $dst,$src\n\t" |
kvn@3882 | 1845 | "punpcklbw $dst,$dst\n\t" |
kvn@3882 | 1846 | "pshuflw $dst,$dst,0x00\t! replicate8B" %} |
kvn@3882 | 1847 | ins_encode %{ |
kvn@3882 | 1848 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 1849 | __ punpcklbw($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 1850 | __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3882 | 1851 | %} |
kvn@3882 | 1852 | ins_pipe( pipe_slow ); |
kvn@3882 | 1853 | %} |
kvn@3882 | 1854 | |
kvn@3882 | 1855 | instruct Repl16B(vecX dst, rRegI src) %{ |
kvn@3882 | 1856 | predicate(n->as_Vector()->length() == 16); |
kvn@3882 | 1857 | match(Set dst (ReplicateB src)); |
kvn@3882 | 1858 | format %{ "movd $dst,$src\n\t" |
kvn@3882 | 1859 | "punpcklbw $dst,$dst\n\t" |
kvn@3882 | 1860 | "pshuflw $dst,$dst,0x00\n\t" |
kvn@3929 | 1861 | "punpcklqdq $dst,$dst\t! replicate16B" %} |
kvn@3882 | 1862 | ins_encode %{ |
kvn@3882 | 1863 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 1864 | __ punpcklbw($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 1865 | __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3929 | 1866 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 1867 | %} |
kvn@3882 | 1868 | ins_pipe( pipe_slow ); |
kvn@3882 | 1869 | %} |
kvn@3882 | 1870 | |
kvn@3882 | 1871 | instruct Repl32B(vecY dst, rRegI src) %{ |
kvn@3882 | 1872 | predicate(n->as_Vector()->length() == 32); |
kvn@3882 | 1873 | match(Set dst (ReplicateB src)); |
kvn@3882 | 1874 | format %{ "movd $dst,$src\n\t" |
kvn@3882 | 1875 | "punpcklbw $dst,$dst\n\t" |
kvn@3882 | 1876 | "pshuflw $dst,$dst,0x00\n\t" |
kvn@3929 | 1877 | "punpcklqdq $dst,$dst\n\t" |
kvn@3929 | 1878 | "vinserti128h $dst,$dst,$dst\t! replicate32B" %} |
kvn@3882 | 1879 | ins_encode %{ |
kvn@3882 | 1880 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 1881 | __ punpcklbw($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 1882 | __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3929 | 1883 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3929 | 1884 | __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 1885 | %} |
kvn@3882 | 1886 | ins_pipe( pipe_slow ); |
kvn@3882 | 1887 | %} |
kvn@3882 | 1888 | |
kvn@3882 | 1889 | // Replicate byte scalar immediate to be vector by loading from const table. |
kvn@3882 | 1890 | instruct Repl4B_imm(vecS dst, immI con) %{ |
kvn@3882 | 1891 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 1892 | match(Set dst (ReplicateB con)); |
kvn@3929 | 1893 | format %{ "movdl $dst,[$constantaddress]\t! replicate4B($con)" %} |
kvn@3882 | 1894 | ins_encode %{ |
kvn@3929 | 1895 | __ movdl($dst$$XMMRegister, $constantaddress(replicate4_imm($con$$constant, 1))); |
kvn@3882 | 1896 | %} |
kvn@3882 | 1897 | ins_pipe( pipe_slow ); |
kvn@3882 | 1898 | %} |
kvn@3882 | 1899 | |
kvn@3882 | 1900 | instruct Repl8B_imm(vecD dst, immI con) %{ |
kvn@3882 | 1901 | predicate(n->as_Vector()->length() == 8); |
kvn@3882 | 1902 | match(Set dst (ReplicateB con)); |
kvn@3929 | 1903 | format %{ "movq $dst,[$constantaddress]\t! replicate8B($con)" %} |
kvn@3882 | 1904 | ins_encode %{ |
kvn@3929 | 1905 | __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 1))); |
kvn@3882 | 1906 | %} |
kvn@3882 | 1907 | ins_pipe( pipe_slow ); |
kvn@3882 | 1908 | %} |
kvn@3882 | 1909 | |
kvn@3882 | 1910 | instruct Repl16B_imm(vecX dst, immI con) %{ |
kvn@3882 | 1911 | predicate(n->as_Vector()->length() == 16); |
kvn@3882 | 1912 | match(Set dst (ReplicateB con)); |
kvn@3929 | 1913 | format %{ "movq $dst,[$constantaddress]\n\t" |
kvn@3929 | 1914 | "punpcklqdq $dst,$dst\t! replicate16B($con)" %} |
kvn@3882 | 1915 | ins_encode %{ |
kvn@3929 | 1916 | __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 1))); |
kvn@3929 | 1917 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 1918 | %} |
kvn@3882 | 1919 | ins_pipe( pipe_slow ); |
kvn@3882 | 1920 | %} |
kvn@3882 | 1921 | |
kvn@3882 | 1922 | instruct Repl32B_imm(vecY dst, immI con) %{ |
kvn@3882 | 1923 | predicate(n->as_Vector()->length() == 32); |
kvn@3882 | 1924 | match(Set dst (ReplicateB con)); |
kvn@3929 | 1925 | format %{ "movq $dst,[$constantaddress]\n\t" |
kvn@3929 | 1926 | "punpcklqdq $dst,$dst\n\t" |
kvn@3929 | 1927 | "vinserti128h $dst,$dst,$dst\t! lreplicate32B($con)" %} |
kvn@3882 | 1928 | ins_encode %{ |
kvn@3929 | 1929 | __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 1))); |
kvn@3929 | 1930 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3929 | 1931 | __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 1932 | %} |
kvn@3882 | 1933 | ins_pipe( pipe_slow ); |
kvn@3882 | 1934 | %} |
kvn@3882 | 1935 | |
kvn@3882 | 1936 | // Replicate byte scalar zero to be vector |
kvn@3882 | 1937 | instruct Repl4B_zero(vecS dst, immI0 zero) %{ |
kvn@3882 | 1938 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 1939 | match(Set dst (ReplicateB zero)); |
kvn@3882 | 1940 | format %{ "pxor $dst,$dst\t! replicate4B zero" %} |
kvn@3882 | 1941 | ins_encode %{ |
kvn@3882 | 1942 | __ pxor($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 1943 | %} |
kvn@3882 | 1944 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 1945 | %} |
kvn@3882 | 1946 | |
kvn@3882 | 1947 | instruct Repl8B_zero(vecD dst, immI0 zero) %{ |
kvn@3882 | 1948 | predicate(n->as_Vector()->length() == 8); |
kvn@3882 | 1949 | match(Set dst (ReplicateB zero)); |
kvn@3882 | 1950 | format %{ "pxor $dst,$dst\t! replicate8B zero" %} |
kvn@3882 | 1951 | ins_encode %{ |
kvn@3882 | 1952 | __ pxor($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 1953 | %} |
kvn@3882 | 1954 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 1955 | %} |
kvn@3882 | 1956 | |
kvn@3882 | 1957 | instruct Repl16B_zero(vecX dst, immI0 zero) %{ |
kvn@3882 | 1958 | predicate(n->as_Vector()->length() == 16); |
kvn@3882 | 1959 | match(Set dst (ReplicateB zero)); |
kvn@3882 | 1960 | format %{ "pxor $dst,$dst\t! replicate16B zero" %} |
kvn@3882 | 1961 | ins_encode %{ |
kvn@3882 | 1962 | __ pxor($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 1963 | %} |
kvn@3882 | 1964 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 1965 | %} |
kvn@3882 | 1966 | |
kvn@3882 | 1967 | instruct Repl32B_zero(vecY dst, immI0 zero) %{ |
kvn@3882 | 1968 | predicate(n->as_Vector()->length() == 32); |
kvn@3882 | 1969 | match(Set dst (ReplicateB zero)); |
kvn@3929 | 1970 | format %{ "vpxor $dst,$dst,$dst\t! replicate32B zero" %} |
kvn@3882 | 1971 | ins_encode %{ |
kvn@3882 | 1972 | // Use vxorpd since AVX does not have vpxor for 256-bit (AVX2 will have it). |
kvn@3882 | 1973 | bool vector256 = true; |
kvn@3929 | 1974 | __ vpxor($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister, vector256); |
kvn@3882 | 1975 | %} |
kvn@3882 | 1976 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 1977 | %} |
kvn@3882 | 1978 | |
kvn@3882 | 1979 | // Replicate char/short (2 byte) scalar to be vector |
kvn@3882 | 1980 | instruct Repl2S(vecS dst, rRegI src) %{ |
kvn@3882 | 1981 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 1982 | match(Set dst (ReplicateS src)); |
kvn@3882 | 1983 | format %{ "movd $dst,$src\n\t" |
kvn@3882 | 1984 | "pshuflw $dst,$dst,0x00\t! replicate2S" %} |
kvn@3882 | 1985 | ins_encode %{ |
kvn@3882 | 1986 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 1987 | __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3882 | 1988 | %} |
kvn@3882 | 1989 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 1990 | %} |
kvn@3882 | 1991 | |
kvn@3882 | 1992 | instruct Repl4S(vecD dst, rRegI src) %{ |
kvn@3882 | 1993 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 1994 | match(Set dst (ReplicateS src)); |
kvn@3882 | 1995 | format %{ "movd $dst,$src\n\t" |
kvn@3882 | 1996 | "pshuflw $dst,$dst,0x00\t! replicate4S" %} |
kvn@3882 | 1997 | ins_encode %{ |
kvn@3882 | 1998 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 1999 | __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3882 | 2000 | %} |
kvn@3882 | 2001 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2002 | %} |
kvn@3882 | 2003 | |
kvn@3882 | 2004 | instruct Repl8S(vecX dst, rRegI src) %{ |
kvn@3882 | 2005 | predicate(n->as_Vector()->length() == 8); |
kvn@3882 | 2006 | match(Set dst (ReplicateS src)); |
kvn@3882 | 2007 | format %{ "movd $dst,$src\n\t" |
kvn@3882 | 2008 | "pshuflw $dst,$dst,0x00\n\t" |
kvn@3929 | 2009 | "punpcklqdq $dst,$dst\t! replicate8S" %} |
kvn@3882 | 2010 | ins_encode %{ |
kvn@3882 | 2011 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 2012 | __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3929 | 2013 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2014 | %} |
kvn@3882 | 2015 | ins_pipe( pipe_slow ); |
kvn@3882 | 2016 | %} |
kvn@3882 | 2017 | |
kvn@3882 | 2018 | instruct Repl16S(vecY dst, rRegI src) %{ |
kvn@3882 | 2019 | predicate(n->as_Vector()->length() == 16); |
kvn@3882 | 2020 | match(Set dst (ReplicateS src)); |
kvn@3882 | 2021 | format %{ "movd $dst,$src\n\t" |
kvn@3882 | 2022 | "pshuflw $dst,$dst,0x00\n\t" |
kvn@3929 | 2023 | "punpcklqdq $dst,$dst\n\t" |
kvn@3929 | 2024 | "vinserti128h $dst,$dst,$dst\t! replicate16S" %} |
kvn@3882 | 2025 | ins_encode %{ |
kvn@3882 | 2026 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 2027 | __ pshuflw($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3929 | 2028 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3929 | 2029 | __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2030 | %} |
kvn@3882 | 2031 | ins_pipe( pipe_slow ); |
kvn@3882 | 2032 | %} |
kvn@3882 | 2033 | |
kvn@3882 | 2034 | // Replicate char/short (2 byte) scalar immediate to be vector by loading from const table. |
kvn@3882 | 2035 | instruct Repl2S_imm(vecS dst, immI con) %{ |
kvn@3882 | 2036 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2037 | match(Set dst (ReplicateS con)); |
kvn@3929 | 2038 | format %{ "movdl $dst,[$constantaddress]\t! replicate2S($con)" %} |
kvn@3882 | 2039 | ins_encode %{ |
kvn@3929 | 2040 | __ movdl($dst$$XMMRegister, $constantaddress(replicate4_imm($con$$constant, 2))); |
kvn@3882 | 2041 | %} |
kvn@3882 | 2042 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2043 | %} |
kvn@3882 | 2044 | |
kvn@3882 | 2045 | instruct Repl4S_imm(vecD dst, immI con) %{ |
kvn@3882 | 2046 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2047 | match(Set dst (ReplicateS con)); |
kvn@3929 | 2048 | format %{ "movq $dst,[$constantaddress]\t! replicate4S($con)" %} |
kvn@3882 | 2049 | ins_encode %{ |
kvn@3929 | 2050 | __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 2))); |
kvn@3882 | 2051 | %} |
kvn@3882 | 2052 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2053 | %} |
kvn@3882 | 2054 | |
kvn@3882 | 2055 | instruct Repl8S_imm(vecX dst, immI con) %{ |
kvn@3882 | 2056 | predicate(n->as_Vector()->length() == 8); |
kvn@3882 | 2057 | match(Set dst (ReplicateS con)); |
kvn@3929 | 2058 | format %{ "movq $dst,[$constantaddress]\n\t" |
kvn@3929 | 2059 | "punpcklqdq $dst,$dst\t! replicate8S($con)" %} |
kvn@3882 | 2060 | ins_encode %{ |
kvn@3929 | 2061 | __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 2))); |
kvn@3929 | 2062 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2063 | %} |
kvn@3882 | 2064 | ins_pipe( pipe_slow ); |
kvn@3882 | 2065 | %} |
kvn@3882 | 2066 | |
kvn@3882 | 2067 | instruct Repl16S_imm(vecY dst, immI con) %{ |
kvn@3882 | 2068 | predicate(n->as_Vector()->length() == 16); |
kvn@3882 | 2069 | match(Set dst (ReplicateS con)); |
kvn@3929 | 2070 | format %{ "movq $dst,[$constantaddress]\n\t" |
kvn@3929 | 2071 | "punpcklqdq $dst,$dst\n\t" |
kvn@3929 | 2072 | "vinserti128h $dst,$dst,$dst\t! replicate16S($con)" %} |
kvn@3882 | 2073 | ins_encode %{ |
kvn@3929 | 2074 | __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 2))); |
kvn@3929 | 2075 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3929 | 2076 | __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2077 | %} |
kvn@3882 | 2078 | ins_pipe( pipe_slow ); |
kvn@3882 | 2079 | %} |
kvn@3882 | 2080 | |
kvn@3882 | 2081 | // Replicate char/short (2 byte) scalar zero to be vector |
kvn@3882 | 2082 | instruct Repl2S_zero(vecS dst, immI0 zero) %{ |
kvn@3882 | 2083 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2084 | match(Set dst (ReplicateS zero)); |
kvn@3882 | 2085 | format %{ "pxor $dst,$dst\t! replicate2S zero" %} |
kvn@3882 | 2086 | ins_encode %{ |
kvn@3882 | 2087 | __ pxor($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2088 | %} |
kvn@3882 | 2089 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2090 | %} |
kvn@3882 | 2091 | |
kvn@3882 | 2092 | instruct Repl4S_zero(vecD dst, immI0 zero) %{ |
kvn@3882 | 2093 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2094 | match(Set dst (ReplicateS zero)); |
kvn@3882 | 2095 | format %{ "pxor $dst,$dst\t! replicate4S zero" %} |
kvn@3882 | 2096 | ins_encode %{ |
kvn@3882 | 2097 | __ pxor($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2098 | %} |
kvn@3882 | 2099 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2100 | %} |
kvn@3882 | 2101 | |
kvn@3882 | 2102 | instruct Repl8S_zero(vecX dst, immI0 zero) %{ |
kvn@3882 | 2103 | predicate(n->as_Vector()->length() == 8); |
kvn@3882 | 2104 | match(Set dst (ReplicateS zero)); |
kvn@3882 | 2105 | format %{ "pxor $dst,$dst\t! replicate8S zero" %} |
kvn@3882 | 2106 | ins_encode %{ |
kvn@3882 | 2107 | __ pxor($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2108 | %} |
kvn@3882 | 2109 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2110 | %} |
kvn@3882 | 2111 | |
kvn@3882 | 2112 | instruct Repl16S_zero(vecY dst, immI0 zero) %{ |
kvn@3882 | 2113 | predicate(n->as_Vector()->length() == 16); |
kvn@3882 | 2114 | match(Set dst (ReplicateS zero)); |
kvn@3929 | 2115 | format %{ "vpxor $dst,$dst,$dst\t! replicate16S zero" %} |
kvn@3882 | 2116 | ins_encode %{ |
kvn@3882 | 2117 | // Use vxorpd since AVX does not have vpxor for 256-bit (AVX2 will have it). |
kvn@3882 | 2118 | bool vector256 = true; |
kvn@3929 | 2119 | __ vpxor($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister, vector256); |
kvn@3882 | 2120 | %} |
kvn@3882 | 2121 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2122 | %} |
kvn@3882 | 2123 | |
kvn@3882 | 2124 | // Replicate integer (4 byte) scalar to be vector |
kvn@3882 | 2125 | instruct Repl2I(vecD dst, rRegI src) %{ |
kvn@3882 | 2126 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2127 | match(Set dst (ReplicateI src)); |
kvn@3882 | 2128 | format %{ "movd $dst,$src\n\t" |
kvn@3882 | 2129 | "pshufd $dst,$dst,0x00\t! replicate2I" %} |
kvn@3882 | 2130 | ins_encode %{ |
kvn@3882 | 2131 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 2132 | __ pshufd($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3882 | 2133 | %} |
kvn@3882 | 2134 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2135 | %} |
kvn@3882 | 2136 | |
kvn@3882 | 2137 | instruct Repl4I(vecX dst, rRegI src) %{ |
kvn@3882 | 2138 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2139 | match(Set dst (ReplicateI src)); |
kvn@3882 | 2140 | format %{ "movd $dst,$src\n\t" |
kvn@3882 | 2141 | "pshufd $dst,$dst,0x00\t! replicate4I" %} |
kvn@3882 | 2142 | ins_encode %{ |
kvn@3882 | 2143 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 2144 | __ pshufd($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3882 | 2145 | %} |
kvn@3882 | 2146 | ins_pipe( pipe_slow ); |
kvn@3882 | 2147 | %} |
kvn@3882 | 2148 | |
kvn@3882 | 2149 | instruct Repl8I(vecY dst, rRegI src) %{ |
kvn@3882 | 2150 | predicate(n->as_Vector()->length() == 8); |
kvn@3882 | 2151 | match(Set dst (ReplicateI src)); |
kvn@3882 | 2152 | format %{ "movd $dst,$src\n\t" |
kvn@3882 | 2153 | "pshufd $dst,$dst,0x00\n\t" |
kvn@3929 | 2154 | "vinserti128h $dst,$dst,$dst\t! replicate8I" %} |
kvn@3882 | 2155 | ins_encode %{ |
kvn@3882 | 2156 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 2157 | __ pshufd($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3929 | 2158 | __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2159 | %} |
kvn@3882 | 2160 | ins_pipe( pipe_slow ); |
kvn@3882 | 2161 | %} |
kvn@3882 | 2162 | |
kvn@3882 | 2163 | // Replicate integer (4 byte) scalar immediate to be vector by loading from const table. |
kvn@3882 | 2164 | instruct Repl2I_imm(vecD dst, immI con) %{ |
kvn@3882 | 2165 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2166 | match(Set dst (ReplicateI con)); |
kvn@3929 | 2167 | format %{ "movq $dst,[$constantaddress]\t! replicate2I($con)" %} |
kvn@3882 | 2168 | ins_encode %{ |
kvn@3929 | 2169 | __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 4))); |
kvn@3882 | 2170 | %} |
kvn@3882 | 2171 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2172 | %} |
kvn@3882 | 2173 | |
kvn@3882 | 2174 | instruct Repl4I_imm(vecX dst, immI con) %{ |
kvn@3882 | 2175 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2176 | match(Set dst (ReplicateI con)); |
kvn@3929 | 2177 | format %{ "movq $dst,[$constantaddress]\t! replicate4I($con)\n\t" |
kvn@3929 | 2178 | "punpcklqdq $dst,$dst" %} |
kvn@3882 | 2179 | ins_encode %{ |
kvn@3929 | 2180 | __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 4))); |
kvn@3929 | 2181 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2182 | %} |
kvn@3882 | 2183 | ins_pipe( pipe_slow ); |
kvn@3882 | 2184 | %} |
kvn@3882 | 2185 | |
kvn@3882 | 2186 | instruct Repl8I_imm(vecY dst, immI con) %{ |
kvn@3882 | 2187 | predicate(n->as_Vector()->length() == 8); |
kvn@3882 | 2188 | match(Set dst (ReplicateI con)); |
kvn@3929 | 2189 | format %{ "movq $dst,[$constantaddress]\t! replicate8I($con)\n\t" |
kvn@3929 | 2190 | "punpcklqdq $dst,$dst\n\t" |
kvn@3929 | 2191 | "vinserti128h $dst,$dst,$dst" %} |
kvn@3882 | 2192 | ins_encode %{ |
kvn@3929 | 2193 | __ movq($dst$$XMMRegister, $constantaddress(replicate8_imm($con$$constant, 4))); |
kvn@3929 | 2194 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3929 | 2195 | __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2196 | %} |
kvn@3882 | 2197 | ins_pipe( pipe_slow ); |
kvn@3882 | 2198 | %} |
kvn@3882 | 2199 | |
kvn@3882 | 2200 | // Integer could be loaded into xmm register directly from memory. |
kvn@3882 | 2201 | instruct Repl2I_mem(vecD dst, memory mem) %{ |
kvn@3882 | 2202 | predicate(n->as_Vector()->length() == 2); |
kvn@3929 | 2203 | match(Set dst (ReplicateI (LoadI mem))); |
kvn@3882 | 2204 | format %{ "movd $dst,$mem\n\t" |
kvn@3882 | 2205 | "pshufd $dst,$dst,0x00\t! replicate2I" %} |
kvn@3882 | 2206 | ins_encode %{ |
kvn@3882 | 2207 | __ movdl($dst$$XMMRegister, $mem$$Address); |
kvn@3882 | 2208 | __ pshufd($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3882 | 2209 | %} |
kvn@3882 | 2210 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2211 | %} |
kvn@3882 | 2212 | |
kvn@3882 | 2213 | instruct Repl4I_mem(vecX dst, memory mem) %{ |
kvn@3882 | 2214 | predicate(n->as_Vector()->length() == 4); |
kvn@3929 | 2215 | match(Set dst (ReplicateI (LoadI mem))); |
kvn@3882 | 2216 | format %{ "movd $dst,$mem\n\t" |
kvn@3882 | 2217 | "pshufd $dst,$dst,0x00\t! replicate4I" %} |
kvn@3882 | 2218 | ins_encode %{ |
kvn@3882 | 2219 | __ movdl($dst$$XMMRegister, $mem$$Address); |
kvn@3882 | 2220 | __ pshufd($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3882 | 2221 | %} |
kvn@3882 | 2222 | ins_pipe( pipe_slow ); |
kvn@3882 | 2223 | %} |
kvn@3882 | 2224 | |
kvn@3882 | 2225 | instruct Repl8I_mem(vecY dst, memory mem) %{ |
kvn@3882 | 2226 | predicate(n->as_Vector()->length() == 8); |
kvn@3929 | 2227 | match(Set dst (ReplicateI (LoadI mem))); |
kvn@3882 | 2228 | format %{ "movd $dst,$mem\n\t" |
kvn@3882 | 2229 | "pshufd $dst,$dst,0x00\n\t" |
kvn@3929 | 2230 | "vinserti128h $dst,$dst,$dst\t! replicate8I" %} |
kvn@3882 | 2231 | ins_encode %{ |
kvn@3882 | 2232 | __ movdl($dst$$XMMRegister, $mem$$Address); |
kvn@3882 | 2233 | __ pshufd($dst$$XMMRegister, $dst$$XMMRegister, 0x00); |
kvn@3929 | 2234 | __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2235 | %} |
kvn@3882 | 2236 | ins_pipe( pipe_slow ); |
kvn@3882 | 2237 | %} |
kvn@3882 | 2238 | |
kvn@3882 | 2239 | // Replicate integer (4 byte) scalar zero to be vector |
kvn@3882 | 2240 | instruct Repl2I_zero(vecD dst, immI0 zero) %{ |
kvn@3882 | 2241 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2242 | match(Set dst (ReplicateI zero)); |
kvn@3882 | 2243 | format %{ "pxor $dst,$dst\t! replicate2I" %} |
kvn@3882 | 2244 | ins_encode %{ |
kvn@3882 | 2245 | __ pxor($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2246 | %} |
kvn@3882 | 2247 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2248 | %} |
kvn@3882 | 2249 | |
kvn@3882 | 2250 | instruct Repl4I_zero(vecX dst, immI0 zero) %{ |
kvn@3882 | 2251 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2252 | match(Set dst (ReplicateI zero)); |
kvn@3882 | 2253 | format %{ "pxor $dst,$dst\t! replicate4I zero)" %} |
kvn@3882 | 2254 | ins_encode %{ |
kvn@3882 | 2255 | __ pxor($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2256 | %} |
kvn@3882 | 2257 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2258 | %} |
kvn@3882 | 2259 | |
kvn@3882 | 2260 | instruct Repl8I_zero(vecY dst, immI0 zero) %{ |
kvn@3882 | 2261 | predicate(n->as_Vector()->length() == 8); |
kvn@3882 | 2262 | match(Set dst (ReplicateI zero)); |
kvn@3929 | 2263 | format %{ "vpxor $dst,$dst,$dst\t! replicate8I zero" %} |
kvn@3882 | 2264 | ins_encode %{ |
kvn@3882 | 2265 | // Use vxorpd since AVX does not have vpxor for 256-bit (AVX2 will have it). |
kvn@3882 | 2266 | bool vector256 = true; |
kvn@3929 | 2267 | __ vpxor($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister, vector256); |
kvn@3882 | 2268 | %} |
kvn@3882 | 2269 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2270 | %} |
kvn@3882 | 2271 | |
kvn@3882 | 2272 | // Replicate long (8 byte) scalar to be vector |
kvn@3882 | 2273 | #ifdef _LP64 |
kvn@3882 | 2274 | instruct Repl2L(vecX dst, rRegL src) %{ |
kvn@3882 | 2275 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2276 | match(Set dst (ReplicateL src)); |
kvn@3882 | 2277 | format %{ "movdq $dst,$src\n\t" |
kvn@3929 | 2278 | "punpcklqdq $dst,$dst\t! replicate2L" %} |
kvn@3882 | 2279 | ins_encode %{ |
kvn@3882 | 2280 | __ movdq($dst$$XMMRegister, $src$$Register); |
kvn@3929 | 2281 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2282 | %} |
kvn@3882 | 2283 | ins_pipe( pipe_slow ); |
kvn@3882 | 2284 | %} |
kvn@3882 | 2285 | |
kvn@3882 | 2286 | instruct Repl4L(vecY dst, rRegL src) %{ |
kvn@3882 | 2287 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2288 | match(Set dst (ReplicateL src)); |
kvn@3882 | 2289 | format %{ "movdq $dst,$src\n\t" |
kvn@3929 | 2290 | "punpcklqdq $dst,$dst\n\t" |
kvn@3929 | 2291 | "vinserti128h $dst,$dst,$dst\t! replicate4L" %} |
kvn@3882 | 2292 | ins_encode %{ |
kvn@3882 | 2293 | __ movdq($dst$$XMMRegister, $src$$Register); |
kvn@3929 | 2294 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3929 | 2295 | __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2296 | %} |
kvn@3882 | 2297 | ins_pipe( pipe_slow ); |
kvn@3882 | 2298 | %} |
kvn@3882 | 2299 | #else // _LP64 |
kvn@3882 | 2300 | instruct Repl2L(vecX dst, eRegL src, regD tmp) %{ |
kvn@3882 | 2301 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2302 | match(Set dst (ReplicateL src)); |
kvn@3882 | 2303 | effect(TEMP dst, USE src, TEMP tmp); |
kvn@3882 | 2304 | format %{ "movdl $dst,$src.lo\n\t" |
kvn@3882 | 2305 | "movdl $tmp,$src.hi\n\t" |
kvn@3882 | 2306 | "punpckldq $dst,$tmp\n\t" |
kvn@3929 | 2307 | "punpcklqdq $dst,$dst\t! replicate2L"%} |
kvn@3882 | 2308 | ins_encode %{ |
kvn@3882 | 2309 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 2310 | __ movdl($tmp$$XMMRegister, HIGH_FROM_LOW($src$$Register)); |
kvn@3882 | 2311 | __ punpckldq($dst$$XMMRegister, $tmp$$XMMRegister); |
kvn@3929 | 2312 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2313 | %} |
kvn@3882 | 2314 | ins_pipe( pipe_slow ); |
kvn@3882 | 2315 | %} |
kvn@3882 | 2316 | |
kvn@3882 | 2317 | instruct Repl4L(vecY dst, eRegL src, regD tmp) %{ |
kvn@3882 | 2318 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2319 | match(Set dst (ReplicateL src)); |
kvn@3882 | 2320 | effect(TEMP dst, USE src, TEMP tmp); |
kvn@3882 | 2321 | format %{ "movdl $dst,$src.lo\n\t" |
kvn@3882 | 2322 | "movdl $tmp,$src.hi\n\t" |
kvn@3882 | 2323 | "punpckldq $dst,$tmp\n\t" |
kvn@3929 | 2324 | "punpcklqdq $dst,$dst\n\t" |
kvn@3929 | 2325 | "vinserti128h $dst,$dst,$dst\t! replicate4L" %} |
kvn@3882 | 2326 | ins_encode %{ |
kvn@3882 | 2327 | __ movdl($dst$$XMMRegister, $src$$Register); |
kvn@3882 | 2328 | __ movdl($tmp$$XMMRegister, HIGH_FROM_LOW($src$$Register)); |
kvn@3882 | 2329 | __ punpckldq($dst$$XMMRegister, $tmp$$XMMRegister); |
kvn@3929 | 2330 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3929 | 2331 | __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2332 | %} |
kvn@3882 | 2333 | ins_pipe( pipe_slow ); |
kvn@3882 | 2334 | %} |
kvn@3882 | 2335 | #endif // _LP64 |
kvn@3882 | 2336 | |
kvn@3882 | 2337 | // Replicate long (8 byte) scalar immediate to be vector by loading from const table. |
kvn@3882 | 2338 | instruct Repl2L_imm(vecX dst, immL con) %{ |
kvn@3882 | 2339 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2340 | match(Set dst (ReplicateL con)); |
kvn@3929 | 2341 | format %{ "movq $dst,[$constantaddress]\n\t" |
kvn@3929 | 2342 | "punpcklqdq $dst,$dst\t! replicate2L($con)" %} |
kvn@3882 | 2343 | ins_encode %{ |
kvn@3929 | 2344 | __ movq($dst$$XMMRegister, $constantaddress($con)); |
kvn@3929 | 2345 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2346 | %} |
kvn@3882 | 2347 | ins_pipe( pipe_slow ); |
kvn@3882 | 2348 | %} |
kvn@3882 | 2349 | |
kvn@3882 | 2350 | instruct Repl4L_imm(vecY dst, immL con) %{ |
kvn@3882 | 2351 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2352 | match(Set dst (ReplicateL con)); |
kvn@3929 | 2353 | format %{ "movq $dst,[$constantaddress]\n\t" |
kvn@3929 | 2354 | "punpcklqdq $dst,$dst\n\t" |
kvn@3929 | 2355 | "vinserti128h $dst,$dst,$dst\t! replicate4L($con)" %} |
kvn@3882 | 2356 | ins_encode %{ |
kvn@3929 | 2357 | __ movq($dst$$XMMRegister, $constantaddress($con)); |
kvn@3929 | 2358 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3929 | 2359 | __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2360 | %} |
kvn@3882 | 2361 | ins_pipe( pipe_slow ); |
kvn@3882 | 2362 | %} |
kvn@3882 | 2363 | |
kvn@3882 | 2364 | // Long could be loaded into xmm register directly from memory. |
kvn@3882 | 2365 | instruct Repl2L_mem(vecX dst, memory mem) %{ |
kvn@3882 | 2366 | predicate(n->as_Vector()->length() == 2); |
kvn@3929 | 2367 | match(Set dst (ReplicateL (LoadL mem))); |
kvn@3882 | 2368 | format %{ "movq $dst,$mem\n\t" |
kvn@3929 | 2369 | "punpcklqdq $dst,$dst\t! replicate2L" %} |
kvn@3882 | 2370 | ins_encode %{ |
kvn@3882 | 2371 | __ movq($dst$$XMMRegister, $mem$$Address); |
kvn@3929 | 2372 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2373 | %} |
kvn@3882 | 2374 | ins_pipe( pipe_slow ); |
kvn@3882 | 2375 | %} |
kvn@3882 | 2376 | |
kvn@3882 | 2377 | instruct Repl4L_mem(vecY dst, memory mem) %{ |
kvn@3882 | 2378 | predicate(n->as_Vector()->length() == 4); |
kvn@3929 | 2379 | match(Set dst (ReplicateL (LoadL mem))); |
kvn@3882 | 2380 | format %{ "movq $dst,$mem\n\t" |
kvn@3929 | 2381 | "punpcklqdq $dst,$dst\n\t" |
kvn@3929 | 2382 | "vinserti128h $dst,$dst,$dst\t! replicate4L" %} |
kvn@3882 | 2383 | ins_encode %{ |
kvn@3882 | 2384 | __ movq($dst$$XMMRegister, $mem$$Address); |
kvn@3929 | 2385 | __ punpcklqdq($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3929 | 2386 | __ vinserti128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2387 | %} |
kvn@3882 | 2388 | ins_pipe( pipe_slow ); |
kvn@3882 | 2389 | %} |
kvn@3882 | 2390 | |
kvn@3882 | 2391 | // Replicate long (8 byte) scalar zero to be vector |
kvn@3882 | 2392 | instruct Repl2L_zero(vecX dst, immL0 zero) %{ |
kvn@3882 | 2393 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2394 | match(Set dst (ReplicateL zero)); |
kvn@3882 | 2395 | format %{ "pxor $dst,$dst\t! replicate2L zero" %} |
kvn@3882 | 2396 | ins_encode %{ |
kvn@3882 | 2397 | __ pxor($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2398 | %} |
kvn@3882 | 2399 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2400 | %} |
kvn@3882 | 2401 | |
kvn@3882 | 2402 | instruct Repl4L_zero(vecY dst, immL0 zero) %{ |
kvn@3882 | 2403 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2404 | match(Set dst (ReplicateL zero)); |
kvn@3929 | 2405 | format %{ "vpxor $dst,$dst,$dst\t! replicate4L zero" %} |
kvn@3882 | 2406 | ins_encode %{ |
kvn@3882 | 2407 | // Use vxorpd since AVX does not have vpxor for 256-bit (AVX2 will have it). |
kvn@3882 | 2408 | bool vector256 = true; |
kvn@3929 | 2409 | __ vpxor($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister, vector256); |
kvn@3882 | 2410 | %} |
kvn@3882 | 2411 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2412 | %} |
kvn@3882 | 2413 | |
kvn@3882 | 2414 | // Replicate float (4 byte) scalar to be vector |
kvn@3882 | 2415 | instruct Repl2F(vecD dst, regF src) %{ |
kvn@3882 | 2416 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2417 | match(Set dst (ReplicateF src)); |
kvn@3882 | 2418 | format %{ "pshufd $dst,$dst,0x00\t! replicate2F" %} |
kvn@3882 | 2419 | ins_encode %{ |
kvn@3882 | 2420 | __ pshufd($dst$$XMMRegister, $src$$XMMRegister, 0x00); |
kvn@3882 | 2421 | %} |
kvn@3882 | 2422 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2423 | %} |
kvn@3882 | 2424 | |
kvn@3882 | 2425 | instruct Repl4F(vecX dst, regF src) %{ |
kvn@3882 | 2426 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2427 | match(Set dst (ReplicateF src)); |
kvn@3882 | 2428 | format %{ "pshufd $dst,$dst,0x00\t! replicate4F" %} |
kvn@3882 | 2429 | ins_encode %{ |
kvn@3882 | 2430 | __ pshufd($dst$$XMMRegister, $src$$XMMRegister, 0x00); |
kvn@3882 | 2431 | %} |
kvn@3882 | 2432 | ins_pipe( pipe_slow ); |
kvn@3882 | 2433 | %} |
kvn@3882 | 2434 | |
kvn@3882 | 2435 | instruct Repl8F(vecY dst, regF src) %{ |
kvn@3882 | 2436 | predicate(n->as_Vector()->length() == 8); |
kvn@3882 | 2437 | match(Set dst (ReplicateF src)); |
kvn@3882 | 2438 | format %{ "pshufd $dst,$src,0x00\n\t" |
kvn@3882 | 2439 | "vinsertf128h $dst,$dst,$dst\t! replicate8F" %} |
kvn@3882 | 2440 | ins_encode %{ |
kvn@3882 | 2441 | __ pshufd($dst$$XMMRegister, $src$$XMMRegister, 0x00); |
kvn@3882 | 2442 | __ vinsertf128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2443 | %} |
kvn@3882 | 2444 | ins_pipe( pipe_slow ); |
kvn@3882 | 2445 | %} |
kvn@3882 | 2446 | |
kvn@3882 | 2447 | // Replicate float (4 byte) scalar zero to be vector |
kvn@3882 | 2448 | instruct Repl2F_zero(vecD dst, immF0 zero) %{ |
kvn@3882 | 2449 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2450 | match(Set dst (ReplicateF zero)); |
kvn@3882 | 2451 | format %{ "xorps $dst,$dst\t! replicate2F zero" %} |
kvn@3882 | 2452 | ins_encode %{ |
kvn@3882 | 2453 | __ xorps($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2454 | %} |
kvn@3882 | 2455 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2456 | %} |
kvn@3882 | 2457 | |
kvn@3882 | 2458 | instruct Repl4F_zero(vecX dst, immF0 zero) %{ |
kvn@3882 | 2459 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2460 | match(Set dst (ReplicateF zero)); |
kvn@3882 | 2461 | format %{ "xorps $dst,$dst\t! replicate4F zero" %} |
kvn@3882 | 2462 | ins_encode %{ |
kvn@3882 | 2463 | __ xorps($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2464 | %} |
kvn@3882 | 2465 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2466 | %} |
kvn@3882 | 2467 | |
kvn@3882 | 2468 | instruct Repl8F_zero(vecY dst, immF0 zero) %{ |
kvn@3882 | 2469 | predicate(n->as_Vector()->length() == 8); |
kvn@3882 | 2470 | match(Set dst (ReplicateF zero)); |
kvn@3882 | 2471 | format %{ "vxorps $dst,$dst,$dst\t! replicate8F zero" %} |
kvn@3882 | 2472 | ins_encode %{ |
kvn@3882 | 2473 | bool vector256 = true; |
kvn@3882 | 2474 | __ vxorps($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister, vector256); |
kvn@3882 | 2475 | %} |
kvn@3882 | 2476 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2477 | %} |
kvn@3882 | 2478 | |
kvn@3882 | 2479 | // Replicate double (8 bytes) scalar to be vector |
kvn@3882 | 2480 | instruct Repl2D(vecX dst, regD src) %{ |
kvn@3882 | 2481 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2482 | match(Set dst (ReplicateD src)); |
kvn@3882 | 2483 | format %{ "pshufd $dst,$src,0x44\t! replicate2D" %} |
kvn@3882 | 2484 | ins_encode %{ |
kvn@3882 | 2485 | __ pshufd($dst$$XMMRegister, $src$$XMMRegister, 0x44); |
kvn@3882 | 2486 | %} |
kvn@3882 | 2487 | ins_pipe( pipe_slow ); |
kvn@3882 | 2488 | %} |
kvn@3882 | 2489 | |
kvn@3882 | 2490 | instruct Repl4D(vecY dst, regD src) %{ |
kvn@3882 | 2491 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2492 | match(Set dst (ReplicateD src)); |
kvn@3882 | 2493 | format %{ "pshufd $dst,$src,0x44\n\t" |
kvn@3882 | 2494 | "vinsertf128h $dst,$dst,$dst\t! replicate4D" %} |
kvn@3882 | 2495 | ins_encode %{ |
kvn@3882 | 2496 | __ pshufd($dst$$XMMRegister, $src$$XMMRegister, 0x44); |
kvn@3882 | 2497 | __ vinsertf128h($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2498 | %} |
kvn@3882 | 2499 | ins_pipe( pipe_slow ); |
kvn@3882 | 2500 | %} |
kvn@3882 | 2501 | |
kvn@3882 | 2502 | // Replicate double (8 byte) scalar zero to be vector |
kvn@3882 | 2503 | instruct Repl2D_zero(vecX dst, immD0 zero) %{ |
kvn@3882 | 2504 | predicate(n->as_Vector()->length() == 2); |
kvn@3882 | 2505 | match(Set dst (ReplicateD zero)); |
kvn@3882 | 2506 | format %{ "xorpd $dst,$dst\t! replicate2D zero" %} |
kvn@3882 | 2507 | ins_encode %{ |
kvn@3882 | 2508 | __ xorpd($dst$$XMMRegister, $dst$$XMMRegister); |
kvn@3882 | 2509 | %} |
kvn@3882 | 2510 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2511 | %} |
kvn@3882 | 2512 | |
kvn@3882 | 2513 | instruct Repl4D_zero(vecY dst, immD0 zero) %{ |
kvn@3882 | 2514 | predicate(n->as_Vector()->length() == 4); |
kvn@3882 | 2515 | match(Set dst (ReplicateD zero)); |
kvn@3882 | 2516 | format %{ "vxorpd $dst,$dst,$dst,vect256\t! replicate4D zero" %} |
kvn@3882 | 2517 | ins_encode %{ |
kvn@3882 | 2518 | bool vector256 = true; |
kvn@3882 | 2519 | __ vxorpd($dst$$XMMRegister, $dst$$XMMRegister, $dst$$XMMRegister, vector256); |
kvn@3882 | 2520 | %} |
kvn@3882 | 2521 | ins_pipe( fpu_reg_reg ); |
kvn@3882 | 2522 | %} |
kvn@3882 | 2523 | |
kvn@4001 | 2524 | // ====================VECTOR ARITHMETIC======================================= |
kvn@4001 | 2525 | |
kvn@4001 | 2526 | // --------------------------------- ADD -------------------------------------- |
kvn@4001 | 2527 | |
kvn@4001 | 2528 | // Bytes vector add |
kvn@4001 | 2529 | instruct vadd4B(vecS dst, vecS src) %{ |
kvn@4001 | 2530 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 2531 | match(Set dst (AddVB dst src)); |
kvn@4001 | 2532 | format %{ "paddb $dst,$src\t! add packed4B" %} |
kvn@4001 | 2533 | ins_encode %{ |
kvn@4001 | 2534 | __ paddb($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2535 | %} |
kvn@4001 | 2536 | ins_pipe( pipe_slow ); |
kvn@4001 | 2537 | %} |
kvn@4001 | 2538 | |
kvn@4001 | 2539 | instruct vadd4B_reg(vecS dst, vecS src1, vecS src2) %{ |
kvn@4001 | 2540 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 2541 | match(Set dst (AddVB src1 src2)); |
kvn@4001 | 2542 | format %{ "vpaddb $dst,$src1,$src2\t! add packed4B" %} |
kvn@4001 | 2543 | ins_encode %{ |
kvn@4001 | 2544 | bool vector256 = false; |
kvn@4001 | 2545 | __ vpaddb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2546 | %} |
kvn@4001 | 2547 | ins_pipe( pipe_slow ); |
kvn@4001 | 2548 | %} |
kvn@4001 | 2549 | |
kvn@4001 | 2550 | instruct vadd8B(vecD dst, vecD src) %{ |
kvn@4001 | 2551 | predicate(n->as_Vector()->length() == 8); |
kvn@4001 | 2552 | match(Set dst (AddVB dst src)); |
kvn@4001 | 2553 | format %{ "paddb $dst,$src\t! add packed8B" %} |
kvn@4001 | 2554 | ins_encode %{ |
kvn@4001 | 2555 | __ paddb($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2556 | %} |
kvn@4001 | 2557 | ins_pipe( pipe_slow ); |
kvn@4001 | 2558 | %} |
kvn@4001 | 2559 | |
kvn@4001 | 2560 | instruct vadd8B_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 2561 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 2562 | match(Set dst (AddVB src1 src2)); |
kvn@4001 | 2563 | format %{ "vpaddb $dst,$src1,$src2\t! add packed8B" %} |
kvn@4001 | 2564 | ins_encode %{ |
kvn@4001 | 2565 | bool vector256 = false; |
kvn@4001 | 2566 | __ vpaddb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2567 | %} |
kvn@4001 | 2568 | ins_pipe( pipe_slow ); |
kvn@4001 | 2569 | %} |
kvn@4001 | 2570 | |
kvn@4001 | 2571 | instruct vadd16B(vecX dst, vecX src) %{ |
kvn@4001 | 2572 | predicate(n->as_Vector()->length() == 16); |
kvn@4001 | 2573 | match(Set dst (AddVB dst src)); |
kvn@4001 | 2574 | format %{ "paddb $dst,$src\t! add packed16B" %} |
kvn@4001 | 2575 | ins_encode %{ |
kvn@4001 | 2576 | __ paddb($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2577 | %} |
kvn@4001 | 2578 | ins_pipe( pipe_slow ); |
kvn@4001 | 2579 | %} |
kvn@4001 | 2580 | |
kvn@4001 | 2581 | instruct vadd16B_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 2582 | predicate(UseAVX > 0 && n->as_Vector()->length() == 16); |
kvn@4001 | 2583 | match(Set dst (AddVB src1 src2)); |
kvn@4001 | 2584 | format %{ "vpaddb $dst,$src1,$src2\t! add packed16B" %} |
kvn@4001 | 2585 | ins_encode %{ |
kvn@4001 | 2586 | bool vector256 = false; |
kvn@4001 | 2587 | __ vpaddb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2588 | %} |
kvn@4001 | 2589 | ins_pipe( pipe_slow ); |
kvn@4001 | 2590 | %} |
kvn@4001 | 2591 | |
kvn@4001 | 2592 | instruct vadd16B_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 2593 | predicate(UseAVX > 0 && n->as_Vector()->length() == 16); |
kvn@4001 | 2594 | match(Set dst (AddVB src (LoadVector mem))); |
kvn@4001 | 2595 | format %{ "vpaddb $dst,$src,$mem\t! add packed16B" %} |
kvn@4001 | 2596 | ins_encode %{ |
kvn@4001 | 2597 | bool vector256 = false; |
kvn@4001 | 2598 | __ vpaddb($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 2599 | %} |
kvn@4001 | 2600 | ins_pipe( pipe_slow ); |
kvn@4001 | 2601 | %} |
kvn@4001 | 2602 | |
kvn@4001 | 2603 | instruct vadd32B_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 2604 | predicate(UseAVX > 1 && n->as_Vector()->length() == 32); |
kvn@4001 | 2605 | match(Set dst (AddVB src1 src2)); |
kvn@4001 | 2606 | format %{ "vpaddb $dst,$src1,$src2\t! add packed32B" %} |
kvn@4001 | 2607 | ins_encode %{ |
kvn@4001 | 2608 | bool vector256 = true; |
kvn@4001 | 2609 | __ vpaddb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2610 | %} |
kvn@4001 | 2611 | ins_pipe( pipe_slow ); |
kvn@4001 | 2612 | %} |
kvn@4001 | 2613 | |
kvn@4001 | 2614 | instruct vadd32B_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 2615 | predicate(UseAVX > 1 && n->as_Vector()->length() == 32); |
kvn@4001 | 2616 | match(Set dst (AddVB src (LoadVector mem))); |
kvn@4001 | 2617 | format %{ "vpaddb $dst,$src,$mem\t! add packed32B" %} |
kvn@4001 | 2618 | ins_encode %{ |
kvn@4001 | 2619 | bool vector256 = true; |
kvn@4001 | 2620 | __ vpaddb($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 2621 | %} |
kvn@4001 | 2622 | ins_pipe( pipe_slow ); |
kvn@4001 | 2623 | %} |
kvn@4001 | 2624 | |
kvn@4001 | 2625 | // Shorts/Chars vector add |
kvn@4001 | 2626 | instruct vadd2S(vecS dst, vecS src) %{ |
kvn@4001 | 2627 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 2628 | match(Set dst (AddVS dst src)); |
kvn@4001 | 2629 | format %{ "paddw $dst,$src\t! add packed2S" %} |
kvn@4001 | 2630 | ins_encode %{ |
kvn@4001 | 2631 | __ paddw($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2632 | %} |
kvn@4001 | 2633 | ins_pipe( pipe_slow ); |
kvn@4001 | 2634 | %} |
kvn@4001 | 2635 | |
kvn@4001 | 2636 | instruct vadd2S_reg(vecS dst, vecS src1, vecS src2) %{ |
kvn@4001 | 2637 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 2638 | match(Set dst (AddVS src1 src2)); |
kvn@4001 | 2639 | format %{ "vpaddw $dst,$src1,$src2\t! add packed2S" %} |
kvn@4001 | 2640 | ins_encode %{ |
kvn@4001 | 2641 | bool vector256 = false; |
kvn@4001 | 2642 | __ vpaddw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2643 | %} |
kvn@4001 | 2644 | ins_pipe( pipe_slow ); |
kvn@4001 | 2645 | %} |
kvn@4001 | 2646 | |
kvn@4001 | 2647 | instruct vadd4S(vecD dst, vecD src) %{ |
kvn@4001 | 2648 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 2649 | match(Set dst (AddVS dst src)); |
kvn@4001 | 2650 | format %{ "paddw $dst,$src\t! add packed4S" %} |
kvn@4001 | 2651 | ins_encode %{ |
kvn@4001 | 2652 | __ paddw($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2653 | %} |
kvn@4001 | 2654 | ins_pipe( pipe_slow ); |
kvn@4001 | 2655 | %} |
kvn@4001 | 2656 | |
kvn@4001 | 2657 | instruct vadd4S_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 2658 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 2659 | match(Set dst (AddVS src1 src2)); |
kvn@4001 | 2660 | format %{ "vpaddw $dst,$src1,$src2\t! add packed4S" %} |
kvn@4001 | 2661 | ins_encode %{ |
kvn@4001 | 2662 | bool vector256 = false; |
kvn@4001 | 2663 | __ vpaddw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2664 | %} |
kvn@4001 | 2665 | ins_pipe( pipe_slow ); |
kvn@4001 | 2666 | %} |
kvn@4001 | 2667 | |
kvn@4001 | 2668 | instruct vadd8S(vecX dst, vecX src) %{ |
kvn@4001 | 2669 | predicate(n->as_Vector()->length() == 8); |
kvn@4001 | 2670 | match(Set dst (AddVS dst src)); |
kvn@4001 | 2671 | format %{ "paddw $dst,$src\t! add packed8S" %} |
kvn@4001 | 2672 | ins_encode %{ |
kvn@4001 | 2673 | __ paddw($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2674 | %} |
kvn@4001 | 2675 | ins_pipe( pipe_slow ); |
kvn@4001 | 2676 | %} |
kvn@4001 | 2677 | |
kvn@4001 | 2678 | instruct vadd8S_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 2679 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 2680 | match(Set dst (AddVS src1 src2)); |
kvn@4001 | 2681 | format %{ "vpaddw $dst,$src1,$src2\t! add packed8S" %} |
kvn@4001 | 2682 | ins_encode %{ |
kvn@4001 | 2683 | bool vector256 = false; |
kvn@4001 | 2684 | __ vpaddw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2685 | %} |
kvn@4001 | 2686 | ins_pipe( pipe_slow ); |
kvn@4001 | 2687 | %} |
kvn@4001 | 2688 | |
kvn@4001 | 2689 | instruct vadd8S_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 2690 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 2691 | match(Set dst (AddVS src (LoadVector mem))); |
kvn@4001 | 2692 | format %{ "vpaddw $dst,$src,$mem\t! add packed8S" %} |
kvn@4001 | 2693 | ins_encode %{ |
kvn@4001 | 2694 | bool vector256 = false; |
kvn@4001 | 2695 | __ vpaddw($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 2696 | %} |
kvn@4001 | 2697 | ins_pipe( pipe_slow ); |
kvn@4001 | 2698 | %} |
kvn@4001 | 2699 | |
kvn@4001 | 2700 | instruct vadd16S_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 2701 | predicate(UseAVX > 1 && n->as_Vector()->length() == 16); |
kvn@4001 | 2702 | match(Set dst (AddVS src1 src2)); |
kvn@4001 | 2703 | format %{ "vpaddw $dst,$src1,$src2\t! add packed16S" %} |
kvn@4001 | 2704 | ins_encode %{ |
kvn@4001 | 2705 | bool vector256 = true; |
kvn@4001 | 2706 | __ vpaddw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2707 | %} |
kvn@4001 | 2708 | ins_pipe( pipe_slow ); |
kvn@4001 | 2709 | %} |
kvn@4001 | 2710 | |
kvn@4001 | 2711 | instruct vadd16S_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 2712 | predicate(UseAVX > 1 && n->as_Vector()->length() == 16); |
kvn@4001 | 2713 | match(Set dst (AddVS src (LoadVector mem))); |
kvn@4001 | 2714 | format %{ "vpaddw $dst,$src,$mem\t! add packed16S" %} |
kvn@4001 | 2715 | ins_encode %{ |
kvn@4001 | 2716 | bool vector256 = true; |
kvn@4001 | 2717 | __ vpaddw($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 2718 | %} |
kvn@4001 | 2719 | ins_pipe( pipe_slow ); |
kvn@4001 | 2720 | %} |
kvn@4001 | 2721 | |
kvn@4001 | 2722 | // Integers vector add |
kvn@4001 | 2723 | instruct vadd2I(vecD dst, vecD src) %{ |
kvn@4001 | 2724 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 2725 | match(Set dst (AddVI dst src)); |
kvn@4001 | 2726 | format %{ "paddd $dst,$src\t! add packed2I" %} |
kvn@4001 | 2727 | ins_encode %{ |
kvn@4001 | 2728 | __ paddd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2729 | %} |
kvn@4001 | 2730 | ins_pipe( pipe_slow ); |
kvn@4001 | 2731 | %} |
kvn@4001 | 2732 | |
kvn@4001 | 2733 | instruct vadd2I_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 2734 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 2735 | match(Set dst (AddVI src1 src2)); |
kvn@4001 | 2736 | format %{ "vpaddd $dst,$src1,$src2\t! add packed2I" %} |
kvn@4001 | 2737 | ins_encode %{ |
kvn@4001 | 2738 | bool vector256 = false; |
kvn@4001 | 2739 | __ vpaddd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2740 | %} |
kvn@4001 | 2741 | ins_pipe( pipe_slow ); |
kvn@4001 | 2742 | %} |
kvn@4001 | 2743 | |
kvn@4001 | 2744 | instruct vadd4I(vecX dst, vecX src) %{ |
kvn@4001 | 2745 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 2746 | match(Set dst (AddVI dst src)); |
kvn@4001 | 2747 | format %{ "paddd $dst,$src\t! add packed4I" %} |
kvn@4001 | 2748 | ins_encode %{ |
kvn@4001 | 2749 | __ paddd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2750 | %} |
kvn@4001 | 2751 | ins_pipe( pipe_slow ); |
kvn@4001 | 2752 | %} |
kvn@4001 | 2753 | |
kvn@4001 | 2754 | instruct vadd4I_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 2755 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 2756 | match(Set dst (AddVI src1 src2)); |
kvn@4001 | 2757 | format %{ "vpaddd $dst,$src1,$src2\t! add packed4I" %} |
kvn@4001 | 2758 | ins_encode %{ |
kvn@4001 | 2759 | bool vector256 = false; |
kvn@4001 | 2760 | __ vpaddd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2761 | %} |
kvn@4001 | 2762 | ins_pipe( pipe_slow ); |
kvn@4001 | 2763 | %} |
kvn@4001 | 2764 | |
kvn@4001 | 2765 | instruct vadd4I_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 2766 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 2767 | match(Set dst (AddVI src (LoadVector mem))); |
kvn@4001 | 2768 | format %{ "vpaddd $dst,$src,$mem\t! add packed4I" %} |
kvn@4001 | 2769 | ins_encode %{ |
kvn@4001 | 2770 | bool vector256 = false; |
kvn@4001 | 2771 | __ vpaddd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 2772 | %} |
kvn@4001 | 2773 | ins_pipe( pipe_slow ); |
kvn@4001 | 2774 | %} |
kvn@4001 | 2775 | |
kvn@4001 | 2776 | instruct vadd8I_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 2777 | predicate(UseAVX > 1 && n->as_Vector()->length() == 8); |
kvn@4001 | 2778 | match(Set dst (AddVI src1 src2)); |
kvn@4001 | 2779 | format %{ "vpaddd $dst,$src1,$src2\t! add packed8I" %} |
kvn@4001 | 2780 | ins_encode %{ |
kvn@4001 | 2781 | bool vector256 = true; |
kvn@4001 | 2782 | __ vpaddd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2783 | %} |
kvn@4001 | 2784 | ins_pipe( pipe_slow ); |
kvn@4001 | 2785 | %} |
kvn@4001 | 2786 | |
kvn@4001 | 2787 | instruct vadd8I_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 2788 | predicate(UseAVX > 1 && n->as_Vector()->length() == 8); |
kvn@4001 | 2789 | match(Set dst (AddVI src (LoadVector mem))); |
kvn@4001 | 2790 | format %{ "vpaddd $dst,$src,$mem\t! add packed8I" %} |
kvn@4001 | 2791 | ins_encode %{ |
kvn@4001 | 2792 | bool vector256 = true; |
kvn@4001 | 2793 | __ vpaddd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 2794 | %} |
kvn@4001 | 2795 | ins_pipe( pipe_slow ); |
kvn@4001 | 2796 | %} |
kvn@4001 | 2797 | |
kvn@4001 | 2798 | // Longs vector add |
kvn@4001 | 2799 | instruct vadd2L(vecX dst, vecX src) %{ |
kvn@4001 | 2800 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 2801 | match(Set dst (AddVL dst src)); |
kvn@4001 | 2802 | format %{ "paddq $dst,$src\t! add packed2L" %} |
kvn@4001 | 2803 | ins_encode %{ |
kvn@4001 | 2804 | __ paddq($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2805 | %} |
kvn@4001 | 2806 | ins_pipe( pipe_slow ); |
kvn@4001 | 2807 | %} |
kvn@4001 | 2808 | |
kvn@4001 | 2809 | instruct vadd2L_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 2810 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 2811 | match(Set dst (AddVL src1 src2)); |
kvn@4001 | 2812 | format %{ "vpaddq $dst,$src1,$src2\t! add packed2L" %} |
kvn@4001 | 2813 | ins_encode %{ |
kvn@4001 | 2814 | bool vector256 = false; |
kvn@4001 | 2815 | __ vpaddq($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2816 | %} |
kvn@4001 | 2817 | ins_pipe( pipe_slow ); |
kvn@4001 | 2818 | %} |
kvn@4001 | 2819 | |
kvn@4001 | 2820 | instruct vadd2L_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 2821 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 2822 | match(Set dst (AddVL src (LoadVector mem))); |
kvn@4001 | 2823 | format %{ "vpaddq $dst,$src,$mem\t! add packed2L" %} |
kvn@4001 | 2824 | ins_encode %{ |
kvn@4001 | 2825 | bool vector256 = false; |
kvn@4001 | 2826 | __ vpaddq($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 2827 | %} |
kvn@4001 | 2828 | ins_pipe( pipe_slow ); |
kvn@4001 | 2829 | %} |
kvn@4001 | 2830 | |
kvn@4001 | 2831 | instruct vadd4L_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 2832 | predicate(UseAVX > 1 && n->as_Vector()->length() == 4); |
kvn@4001 | 2833 | match(Set dst (AddVL src1 src2)); |
kvn@4001 | 2834 | format %{ "vpaddq $dst,$src1,$src2\t! add packed4L" %} |
kvn@4001 | 2835 | ins_encode %{ |
kvn@4001 | 2836 | bool vector256 = true; |
kvn@4001 | 2837 | __ vpaddq($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2838 | %} |
kvn@4001 | 2839 | ins_pipe( pipe_slow ); |
kvn@4001 | 2840 | %} |
kvn@4001 | 2841 | |
kvn@4001 | 2842 | instruct vadd4L_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 2843 | predicate(UseAVX > 1 && n->as_Vector()->length() == 4); |
kvn@4001 | 2844 | match(Set dst (AddVL src (LoadVector mem))); |
kvn@4001 | 2845 | format %{ "vpaddq $dst,$src,$mem\t! add packed4L" %} |
kvn@4001 | 2846 | ins_encode %{ |
kvn@4001 | 2847 | bool vector256 = true; |
kvn@4001 | 2848 | __ vpaddq($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 2849 | %} |
kvn@4001 | 2850 | ins_pipe( pipe_slow ); |
kvn@4001 | 2851 | %} |
kvn@4001 | 2852 | |
kvn@4001 | 2853 | // Floats vector add |
kvn@4001 | 2854 | instruct vadd2F(vecD dst, vecD src) %{ |
kvn@4001 | 2855 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 2856 | match(Set dst (AddVF dst src)); |
kvn@4001 | 2857 | format %{ "addps $dst,$src\t! add packed2F" %} |
kvn@4001 | 2858 | ins_encode %{ |
kvn@4001 | 2859 | __ addps($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2860 | %} |
kvn@4001 | 2861 | ins_pipe( pipe_slow ); |
kvn@4001 | 2862 | %} |
kvn@4001 | 2863 | |
kvn@4001 | 2864 | instruct vadd2F_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 2865 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 2866 | match(Set dst (AddVF src1 src2)); |
kvn@4001 | 2867 | format %{ "vaddps $dst,$src1,$src2\t! add packed2F" %} |
kvn@4001 | 2868 | ins_encode %{ |
kvn@4001 | 2869 | bool vector256 = false; |
kvn@4001 | 2870 | __ vaddps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2871 | %} |
kvn@4001 | 2872 | ins_pipe( pipe_slow ); |
kvn@4001 | 2873 | %} |
kvn@4001 | 2874 | |
kvn@4001 | 2875 | instruct vadd4F(vecX dst, vecX src) %{ |
kvn@4001 | 2876 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 2877 | match(Set dst (AddVF dst src)); |
kvn@4001 | 2878 | format %{ "addps $dst,$src\t! add packed4F" %} |
kvn@4001 | 2879 | ins_encode %{ |
kvn@4001 | 2880 | __ addps($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2881 | %} |
kvn@4001 | 2882 | ins_pipe( pipe_slow ); |
kvn@4001 | 2883 | %} |
kvn@4001 | 2884 | |
kvn@4001 | 2885 | instruct vadd4F_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 2886 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 2887 | match(Set dst (AddVF src1 src2)); |
kvn@4001 | 2888 | format %{ "vaddps $dst,$src1,$src2\t! add packed4F" %} |
kvn@4001 | 2889 | ins_encode %{ |
kvn@4001 | 2890 | bool vector256 = false; |
kvn@4001 | 2891 | __ vaddps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2892 | %} |
kvn@4001 | 2893 | ins_pipe( pipe_slow ); |
kvn@4001 | 2894 | %} |
kvn@4001 | 2895 | |
kvn@4001 | 2896 | instruct vadd4F_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 2897 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 2898 | match(Set dst (AddVF src (LoadVector mem))); |
kvn@4001 | 2899 | format %{ "vaddps $dst,$src,$mem\t! add packed4F" %} |
kvn@4001 | 2900 | ins_encode %{ |
kvn@4001 | 2901 | bool vector256 = false; |
kvn@4001 | 2902 | __ vaddps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 2903 | %} |
kvn@4001 | 2904 | ins_pipe( pipe_slow ); |
kvn@4001 | 2905 | %} |
kvn@4001 | 2906 | |
kvn@4001 | 2907 | instruct vadd8F_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 2908 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 2909 | match(Set dst (AddVF src1 src2)); |
kvn@4001 | 2910 | format %{ "vaddps $dst,$src1,$src2\t! add packed8F" %} |
kvn@4001 | 2911 | ins_encode %{ |
kvn@4001 | 2912 | bool vector256 = true; |
kvn@4001 | 2913 | __ vaddps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2914 | %} |
kvn@4001 | 2915 | ins_pipe( pipe_slow ); |
kvn@4001 | 2916 | %} |
kvn@4001 | 2917 | |
kvn@4001 | 2918 | instruct vadd8F_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 2919 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 2920 | match(Set dst (AddVF src (LoadVector mem))); |
kvn@4001 | 2921 | format %{ "vaddps $dst,$src,$mem\t! add packed8F" %} |
kvn@4001 | 2922 | ins_encode %{ |
kvn@4001 | 2923 | bool vector256 = true; |
kvn@4001 | 2924 | __ vaddps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 2925 | %} |
kvn@4001 | 2926 | ins_pipe( pipe_slow ); |
kvn@4001 | 2927 | %} |
kvn@4001 | 2928 | |
kvn@4001 | 2929 | // Doubles vector add |
kvn@4001 | 2930 | instruct vadd2D(vecX dst, vecX src) %{ |
kvn@4001 | 2931 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 2932 | match(Set dst (AddVD dst src)); |
kvn@4001 | 2933 | format %{ "addpd $dst,$src\t! add packed2D" %} |
kvn@4001 | 2934 | ins_encode %{ |
kvn@4001 | 2935 | __ addpd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2936 | %} |
kvn@4001 | 2937 | ins_pipe( pipe_slow ); |
kvn@4001 | 2938 | %} |
kvn@4001 | 2939 | |
kvn@4001 | 2940 | instruct vadd2D_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 2941 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 2942 | match(Set dst (AddVD src1 src2)); |
kvn@4001 | 2943 | format %{ "vaddpd $dst,$src1,$src2\t! add packed2D" %} |
kvn@4001 | 2944 | ins_encode %{ |
kvn@4001 | 2945 | bool vector256 = false; |
kvn@4001 | 2946 | __ vaddpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2947 | %} |
kvn@4001 | 2948 | ins_pipe( pipe_slow ); |
kvn@4001 | 2949 | %} |
kvn@4001 | 2950 | |
kvn@4001 | 2951 | instruct vadd2D_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 2952 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 2953 | match(Set dst (AddVD src (LoadVector mem))); |
kvn@4001 | 2954 | format %{ "vaddpd $dst,$src,$mem\t! add packed2D" %} |
kvn@4001 | 2955 | ins_encode %{ |
kvn@4001 | 2956 | bool vector256 = false; |
kvn@4001 | 2957 | __ vaddpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 2958 | %} |
kvn@4001 | 2959 | ins_pipe( pipe_slow ); |
kvn@4001 | 2960 | %} |
kvn@4001 | 2961 | |
kvn@4001 | 2962 | instruct vadd4D_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 2963 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 2964 | match(Set dst (AddVD src1 src2)); |
kvn@4001 | 2965 | format %{ "vaddpd $dst,$src1,$src2\t! add packed4D" %} |
kvn@4001 | 2966 | ins_encode %{ |
kvn@4001 | 2967 | bool vector256 = true; |
kvn@4001 | 2968 | __ vaddpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 2969 | %} |
kvn@4001 | 2970 | ins_pipe( pipe_slow ); |
kvn@4001 | 2971 | %} |
kvn@4001 | 2972 | |
kvn@4001 | 2973 | instruct vadd4D_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 2974 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 2975 | match(Set dst (AddVD src (LoadVector mem))); |
kvn@4001 | 2976 | format %{ "vaddpd $dst,$src,$mem\t! add packed4D" %} |
kvn@4001 | 2977 | ins_encode %{ |
kvn@4001 | 2978 | bool vector256 = true; |
kvn@4001 | 2979 | __ vaddpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 2980 | %} |
kvn@4001 | 2981 | ins_pipe( pipe_slow ); |
kvn@4001 | 2982 | %} |
kvn@4001 | 2983 | |
kvn@4001 | 2984 | // --------------------------------- SUB -------------------------------------- |
kvn@4001 | 2985 | |
kvn@4001 | 2986 | // Bytes vector sub |
kvn@4001 | 2987 | instruct vsub4B(vecS dst, vecS src) %{ |
kvn@4001 | 2988 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 2989 | match(Set dst (SubVB dst src)); |
kvn@4001 | 2990 | format %{ "psubb $dst,$src\t! sub packed4B" %} |
kvn@4001 | 2991 | ins_encode %{ |
kvn@4001 | 2992 | __ psubb($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 2993 | %} |
kvn@4001 | 2994 | ins_pipe( pipe_slow ); |
kvn@4001 | 2995 | %} |
kvn@4001 | 2996 | |
kvn@4001 | 2997 | instruct vsub4B_reg(vecS dst, vecS src1, vecS src2) %{ |
kvn@4001 | 2998 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 2999 | match(Set dst (SubVB src1 src2)); |
kvn@4001 | 3000 | format %{ "vpsubb $dst,$src1,$src2\t! sub packed4B" %} |
kvn@4001 | 3001 | ins_encode %{ |
kvn@4001 | 3002 | bool vector256 = false; |
kvn@4001 | 3003 | __ vpsubb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3004 | %} |
kvn@4001 | 3005 | ins_pipe( pipe_slow ); |
kvn@4001 | 3006 | %} |
kvn@4001 | 3007 | |
kvn@4001 | 3008 | instruct vsub8B(vecD dst, vecD src) %{ |
kvn@4001 | 3009 | predicate(n->as_Vector()->length() == 8); |
kvn@4001 | 3010 | match(Set dst (SubVB dst src)); |
kvn@4001 | 3011 | format %{ "psubb $dst,$src\t! sub packed8B" %} |
kvn@4001 | 3012 | ins_encode %{ |
kvn@4001 | 3013 | __ psubb($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3014 | %} |
kvn@4001 | 3015 | ins_pipe( pipe_slow ); |
kvn@4001 | 3016 | %} |
kvn@4001 | 3017 | |
kvn@4001 | 3018 | instruct vsub8B_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 3019 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 3020 | match(Set dst (SubVB src1 src2)); |
kvn@4001 | 3021 | format %{ "vpsubb $dst,$src1,$src2\t! sub packed8B" %} |
kvn@4001 | 3022 | ins_encode %{ |
kvn@4001 | 3023 | bool vector256 = false; |
kvn@4001 | 3024 | __ vpsubb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3025 | %} |
kvn@4001 | 3026 | ins_pipe( pipe_slow ); |
kvn@4001 | 3027 | %} |
kvn@4001 | 3028 | |
kvn@4001 | 3029 | instruct vsub16B(vecX dst, vecX src) %{ |
kvn@4001 | 3030 | predicate(n->as_Vector()->length() == 16); |
kvn@4001 | 3031 | match(Set dst (SubVB dst src)); |
kvn@4001 | 3032 | format %{ "psubb $dst,$src\t! sub packed16B" %} |
kvn@4001 | 3033 | ins_encode %{ |
kvn@4001 | 3034 | __ psubb($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3035 | %} |
kvn@4001 | 3036 | ins_pipe( pipe_slow ); |
kvn@4001 | 3037 | %} |
kvn@4001 | 3038 | |
kvn@4001 | 3039 | instruct vsub16B_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 3040 | predicate(UseAVX > 0 && n->as_Vector()->length() == 16); |
kvn@4001 | 3041 | match(Set dst (SubVB src1 src2)); |
kvn@4001 | 3042 | format %{ "vpsubb $dst,$src1,$src2\t! sub packed16B" %} |
kvn@4001 | 3043 | ins_encode %{ |
kvn@4001 | 3044 | bool vector256 = false; |
kvn@4001 | 3045 | __ vpsubb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3046 | %} |
kvn@4001 | 3047 | ins_pipe( pipe_slow ); |
kvn@4001 | 3048 | %} |
kvn@4001 | 3049 | |
kvn@4001 | 3050 | instruct vsub16B_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 3051 | predicate(UseAVX > 0 && n->as_Vector()->length() == 16); |
kvn@4001 | 3052 | match(Set dst (SubVB src (LoadVector mem))); |
kvn@4001 | 3053 | format %{ "vpsubb $dst,$src,$mem\t! sub packed16B" %} |
kvn@4001 | 3054 | ins_encode %{ |
kvn@4001 | 3055 | bool vector256 = false; |
kvn@4001 | 3056 | __ vpsubb($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3057 | %} |
kvn@4001 | 3058 | ins_pipe( pipe_slow ); |
kvn@4001 | 3059 | %} |
kvn@4001 | 3060 | |
kvn@4001 | 3061 | instruct vsub32B_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 3062 | predicate(UseAVX > 1 && n->as_Vector()->length() == 32); |
kvn@4001 | 3063 | match(Set dst (SubVB src1 src2)); |
kvn@4001 | 3064 | format %{ "vpsubb $dst,$src1,$src2\t! sub packed32B" %} |
kvn@4001 | 3065 | ins_encode %{ |
kvn@4001 | 3066 | bool vector256 = true; |
kvn@4001 | 3067 | __ vpsubb($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3068 | %} |
kvn@4001 | 3069 | ins_pipe( pipe_slow ); |
kvn@4001 | 3070 | %} |
kvn@4001 | 3071 | |
kvn@4001 | 3072 | instruct vsub32B_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 3073 | predicate(UseAVX > 1 && n->as_Vector()->length() == 32); |
kvn@4001 | 3074 | match(Set dst (SubVB src (LoadVector mem))); |
kvn@4001 | 3075 | format %{ "vpsubb $dst,$src,$mem\t! sub packed32B" %} |
kvn@4001 | 3076 | ins_encode %{ |
kvn@4001 | 3077 | bool vector256 = true; |
kvn@4001 | 3078 | __ vpsubb($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3079 | %} |
kvn@4001 | 3080 | ins_pipe( pipe_slow ); |
kvn@4001 | 3081 | %} |
kvn@4001 | 3082 | |
kvn@4001 | 3083 | // Shorts/Chars vector sub |
kvn@4001 | 3084 | instruct vsub2S(vecS dst, vecS src) %{ |
kvn@4001 | 3085 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 3086 | match(Set dst (SubVS dst src)); |
kvn@4001 | 3087 | format %{ "psubw $dst,$src\t! sub packed2S" %} |
kvn@4001 | 3088 | ins_encode %{ |
kvn@4001 | 3089 | __ psubw($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3090 | %} |
kvn@4001 | 3091 | ins_pipe( pipe_slow ); |
kvn@4001 | 3092 | %} |
kvn@4001 | 3093 | |
kvn@4001 | 3094 | instruct vsub2S_reg(vecS dst, vecS src1, vecS src2) %{ |
kvn@4001 | 3095 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3096 | match(Set dst (SubVS src1 src2)); |
kvn@4001 | 3097 | format %{ "vpsubw $dst,$src1,$src2\t! sub packed2S" %} |
kvn@4001 | 3098 | ins_encode %{ |
kvn@4001 | 3099 | bool vector256 = false; |
kvn@4001 | 3100 | __ vpsubw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3101 | %} |
kvn@4001 | 3102 | ins_pipe( pipe_slow ); |
kvn@4001 | 3103 | %} |
kvn@4001 | 3104 | |
kvn@4001 | 3105 | instruct vsub4S(vecD dst, vecD src) %{ |
kvn@4001 | 3106 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 3107 | match(Set dst (SubVS dst src)); |
kvn@4001 | 3108 | format %{ "psubw $dst,$src\t! sub packed4S" %} |
kvn@4001 | 3109 | ins_encode %{ |
kvn@4001 | 3110 | __ psubw($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3111 | %} |
kvn@4001 | 3112 | ins_pipe( pipe_slow ); |
kvn@4001 | 3113 | %} |
kvn@4001 | 3114 | |
kvn@4001 | 3115 | instruct vsub4S_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 3116 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3117 | match(Set dst (SubVS src1 src2)); |
kvn@4001 | 3118 | format %{ "vpsubw $dst,$src1,$src2\t! sub packed4S" %} |
kvn@4001 | 3119 | ins_encode %{ |
kvn@4001 | 3120 | bool vector256 = false; |
kvn@4001 | 3121 | __ vpsubw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3122 | %} |
kvn@4001 | 3123 | ins_pipe( pipe_slow ); |
kvn@4001 | 3124 | %} |
kvn@4001 | 3125 | |
kvn@4001 | 3126 | instruct vsub8S(vecX dst, vecX src) %{ |
kvn@4001 | 3127 | predicate(n->as_Vector()->length() == 8); |
kvn@4001 | 3128 | match(Set dst (SubVS dst src)); |
kvn@4001 | 3129 | format %{ "psubw $dst,$src\t! sub packed8S" %} |
kvn@4001 | 3130 | ins_encode %{ |
kvn@4001 | 3131 | __ psubw($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3132 | %} |
kvn@4001 | 3133 | ins_pipe( pipe_slow ); |
kvn@4001 | 3134 | %} |
kvn@4001 | 3135 | |
kvn@4001 | 3136 | instruct vsub8S_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 3137 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 3138 | match(Set dst (SubVS src1 src2)); |
kvn@4001 | 3139 | format %{ "vpsubw $dst,$src1,$src2\t! sub packed8S" %} |
kvn@4001 | 3140 | ins_encode %{ |
kvn@4001 | 3141 | bool vector256 = false; |
kvn@4001 | 3142 | __ vpsubw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3143 | %} |
kvn@4001 | 3144 | ins_pipe( pipe_slow ); |
kvn@4001 | 3145 | %} |
kvn@4001 | 3146 | |
kvn@4001 | 3147 | instruct vsub8S_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 3148 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 3149 | match(Set dst (SubVS src (LoadVector mem))); |
kvn@4001 | 3150 | format %{ "vpsubw $dst,$src,$mem\t! sub packed8S" %} |
kvn@4001 | 3151 | ins_encode %{ |
kvn@4001 | 3152 | bool vector256 = false; |
kvn@4001 | 3153 | __ vpsubw($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3154 | %} |
kvn@4001 | 3155 | ins_pipe( pipe_slow ); |
kvn@4001 | 3156 | %} |
kvn@4001 | 3157 | |
kvn@4001 | 3158 | instruct vsub16S_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 3159 | predicate(UseAVX > 1 && n->as_Vector()->length() == 16); |
kvn@4001 | 3160 | match(Set dst (SubVS src1 src2)); |
kvn@4001 | 3161 | format %{ "vpsubw $dst,$src1,$src2\t! sub packed16S" %} |
kvn@4001 | 3162 | ins_encode %{ |
kvn@4001 | 3163 | bool vector256 = true; |
kvn@4001 | 3164 | __ vpsubw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3165 | %} |
kvn@4001 | 3166 | ins_pipe( pipe_slow ); |
kvn@4001 | 3167 | %} |
kvn@4001 | 3168 | |
kvn@4001 | 3169 | instruct vsub16S_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 3170 | predicate(UseAVX > 1 && n->as_Vector()->length() == 16); |
kvn@4001 | 3171 | match(Set dst (SubVS src (LoadVector mem))); |
kvn@4001 | 3172 | format %{ "vpsubw $dst,$src,$mem\t! sub packed16S" %} |
kvn@4001 | 3173 | ins_encode %{ |
kvn@4001 | 3174 | bool vector256 = true; |
kvn@4001 | 3175 | __ vpsubw($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3176 | %} |
kvn@4001 | 3177 | ins_pipe( pipe_slow ); |
kvn@4001 | 3178 | %} |
kvn@4001 | 3179 | |
kvn@4001 | 3180 | // Integers vector sub |
kvn@4001 | 3181 | instruct vsub2I(vecD dst, vecD src) %{ |
kvn@4001 | 3182 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 3183 | match(Set dst (SubVI dst src)); |
kvn@4001 | 3184 | format %{ "psubd $dst,$src\t! sub packed2I" %} |
kvn@4001 | 3185 | ins_encode %{ |
kvn@4001 | 3186 | __ psubd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3187 | %} |
kvn@4001 | 3188 | ins_pipe( pipe_slow ); |
kvn@4001 | 3189 | %} |
kvn@4001 | 3190 | |
kvn@4001 | 3191 | instruct vsub2I_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 3192 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3193 | match(Set dst (SubVI src1 src2)); |
kvn@4001 | 3194 | format %{ "vpsubd $dst,$src1,$src2\t! sub packed2I" %} |
kvn@4001 | 3195 | ins_encode %{ |
kvn@4001 | 3196 | bool vector256 = false; |
kvn@4001 | 3197 | __ vpsubd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3198 | %} |
kvn@4001 | 3199 | ins_pipe( pipe_slow ); |
kvn@4001 | 3200 | %} |
kvn@4001 | 3201 | |
kvn@4001 | 3202 | instruct vsub4I(vecX dst, vecX src) %{ |
kvn@4001 | 3203 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 3204 | match(Set dst (SubVI dst src)); |
kvn@4001 | 3205 | format %{ "psubd $dst,$src\t! sub packed4I" %} |
kvn@4001 | 3206 | ins_encode %{ |
kvn@4001 | 3207 | __ psubd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3208 | %} |
kvn@4001 | 3209 | ins_pipe( pipe_slow ); |
kvn@4001 | 3210 | %} |
kvn@4001 | 3211 | |
kvn@4001 | 3212 | instruct vsub4I_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 3213 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3214 | match(Set dst (SubVI src1 src2)); |
kvn@4001 | 3215 | format %{ "vpsubd $dst,$src1,$src2\t! sub packed4I" %} |
kvn@4001 | 3216 | ins_encode %{ |
kvn@4001 | 3217 | bool vector256 = false; |
kvn@4001 | 3218 | __ vpsubd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3219 | %} |
kvn@4001 | 3220 | ins_pipe( pipe_slow ); |
kvn@4001 | 3221 | %} |
kvn@4001 | 3222 | |
kvn@4001 | 3223 | instruct vsub4I_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 3224 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3225 | match(Set dst (SubVI src (LoadVector mem))); |
kvn@4001 | 3226 | format %{ "vpsubd $dst,$src,$mem\t! sub packed4I" %} |
kvn@4001 | 3227 | ins_encode %{ |
kvn@4001 | 3228 | bool vector256 = false; |
kvn@4001 | 3229 | __ vpsubd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3230 | %} |
kvn@4001 | 3231 | ins_pipe( pipe_slow ); |
kvn@4001 | 3232 | %} |
kvn@4001 | 3233 | |
kvn@4001 | 3234 | instruct vsub8I_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 3235 | predicate(UseAVX > 1 && n->as_Vector()->length() == 8); |
kvn@4001 | 3236 | match(Set dst (SubVI src1 src2)); |
kvn@4001 | 3237 | format %{ "vpsubd $dst,$src1,$src2\t! sub packed8I" %} |
kvn@4001 | 3238 | ins_encode %{ |
kvn@4001 | 3239 | bool vector256 = true; |
kvn@4001 | 3240 | __ vpsubd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3241 | %} |
kvn@4001 | 3242 | ins_pipe( pipe_slow ); |
kvn@4001 | 3243 | %} |
kvn@4001 | 3244 | |
kvn@4001 | 3245 | instruct vsub8I_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 3246 | predicate(UseAVX > 1 && n->as_Vector()->length() == 8); |
kvn@4001 | 3247 | match(Set dst (SubVI src (LoadVector mem))); |
kvn@4001 | 3248 | format %{ "vpsubd $dst,$src,$mem\t! sub packed8I" %} |
kvn@4001 | 3249 | ins_encode %{ |
kvn@4001 | 3250 | bool vector256 = true; |
kvn@4001 | 3251 | __ vpsubd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3252 | %} |
kvn@4001 | 3253 | ins_pipe( pipe_slow ); |
kvn@4001 | 3254 | %} |
kvn@4001 | 3255 | |
kvn@4001 | 3256 | // Longs vector sub |
kvn@4001 | 3257 | instruct vsub2L(vecX dst, vecX src) %{ |
kvn@4001 | 3258 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 3259 | match(Set dst (SubVL dst src)); |
kvn@4001 | 3260 | format %{ "psubq $dst,$src\t! sub packed2L" %} |
kvn@4001 | 3261 | ins_encode %{ |
kvn@4001 | 3262 | __ psubq($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3263 | %} |
kvn@4001 | 3264 | ins_pipe( pipe_slow ); |
kvn@4001 | 3265 | %} |
kvn@4001 | 3266 | |
kvn@4001 | 3267 | instruct vsub2L_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 3268 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3269 | match(Set dst (SubVL src1 src2)); |
kvn@4001 | 3270 | format %{ "vpsubq $dst,$src1,$src2\t! sub packed2L" %} |
kvn@4001 | 3271 | ins_encode %{ |
kvn@4001 | 3272 | bool vector256 = false; |
kvn@4001 | 3273 | __ vpsubq($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3274 | %} |
kvn@4001 | 3275 | ins_pipe( pipe_slow ); |
kvn@4001 | 3276 | %} |
kvn@4001 | 3277 | |
kvn@4001 | 3278 | instruct vsub2L_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 3279 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3280 | match(Set dst (SubVL src (LoadVector mem))); |
kvn@4001 | 3281 | format %{ "vpsubq $dst,$src,$mem\t! sub packed2L" %} |
kvn@4001 | 3282 | ins_encode %{ |
kvn@4001 | 3283 | bool vector256 = false; |
kvn@4001 | 3284 | __ vpsubq($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3285 | %} |
kvn@4001 | 3286 | ins_pipe( pipe_slow ); |
kvn@4001 | 3287 | %} |
kvn@4001 | 3288 | |
kvn@4001 | 3289 | instruct vsub4L_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 3290 | predicate(UseAVX > 1 && n->as_Vector()->length() == 4); |
kvn@4001 | 3291 | match(Set dst (SubVL src1 src2)); |
kvn@4001 | 3292 | format %{ "vpsubq $dst,$src1,$src2\t! sub packed4L" %} |
kvn@4001 | 3293 | ins_encode %{ |
kvn@4001 | 3294 | bool vector256 = true; |
kvn@4001 | 3295 | __ vpsubq($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3296 | %} |
kvn@4001 | 3297 | ins_pipe( pipe_slow ); |
kvn@4001 | 3298 | %} |
kvn@4001 | 3299 | |
kvn@4001 | 3300 | instruct vsub4L_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 3301 | predicate(UseAVX > 1 && n->as_Vector()->length() == 4); |
kvn@4001 | 3302 | match(Set dst (SubVL src (LoadVector mem))); |
kvn@4001 | 3303 | format %{ "vpsubq $dst,$src,$mem\t! sub packed4L" %} |
kvn@4001 | 3304 | ins_encode %{ |
kvn@4001 | 3305 | bool vector256 = true; |
kvn@4001 | 3306 | __ vpsubq($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3307 | %} |
kvn@4001 | 3308 | ins_pipe( pipe_slow ); |
kvn@4001 | 3309 | %} |
kvn@4001 | 3310 | |
kvn@4001 | 3311 | // Floats vector sub |
kvn@4001 | 3312 | instruct vsub2F(vecD dst, vecD src) %{ |
kvn@4001 | 3313 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 3314 | match(Set dst (SubVF dst src)); |
kvn@4001 | 3315 | format %{ "subps $dst,$src\t! sub packed2F" %} |
kvn@4001 | 3316 | ins_encode %{ |
kvn@4001 | 3317 | __ subps($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3318 | %} |
kvn@4001 | 3319 | ins_pipe( pipe_slow ); |
kvn@4001 | 3320 | %} |
kvn@4001 | 3321 | |
kvn@4001 | 3322 | instruct vsub2F_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 3323 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3324 | match(Set dst (SubVF src1 src2)); |
kvn@4001 | 3325 | format %{ "vsubps $dst,$src1,$src2\t! sub packed2F" %} |
kvn@4001 | 3326 | ins_encode %{ |
kvn@4001 | 3327 | bool vector256 = false; |
kvn@4001 | 3328 | __ vsubps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3329 | %} |
kvn@4001 | 3330 | ins_pipe( pipe_slow ); |
kvn@4001 | 3331 | %} |
kvn@4001 | 3332 | |
kvn@4001 | 3333 | instruct vsub4F(vecX dst, vecX src) %{ |
kvn@4001 | 3334 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 3335 | match(Set dst (SubVF dst src)); |
kvn@4001 | 3336 | format %{ "subps $dst,$src\t! sub packed4F" %} |
kvn@4001 | 3337 | ins_encode %{ |
kvn@4001 | 3338 | __ subps($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3339 | %} |
kvn@4001 | 3340 | ins_pipe( pipe_slow ); |
kvn@4001 | 3341 | %} |
kvn@4001 | 3342 | |
kvn@4001 | 3343 | instruct vsub4F_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 3344 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3345 | match(Set dst (SubVF src1 src2)); |
kvn@4001 | 3346 | format %{ "vsubps $dst,$src1,$src2\t! sub packed4F" %} |
kvn@4001 | 3347 | ins_encode %{ |
kvn@4001 | 3348 | bool vector256 = false; |
kvn@4001 | 3349 | __ vsubps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3350 | %} |
kvn@4001 | 3351 | ins_pipe( pipe_slow ); |
kvn@4001 | 3352 | %} |
kvn@4001 | 3353 | |
kvn@4001 | 3354 | instruct vsub4F_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 3355 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3356 | match(Set dst (SubVF src (LoadVector mem))); |
kvn@4001 | 3357 | format %{ "vsubps $dst,$src,$mem\t! sub packed4F" %} |
kvn@4001 | 3358 | ins_encode %{ |
kvn@4001 | 3359 | bool vector256 = false; |
kvn@4001 | 3360 | __ vsubps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3361 | %} |
kvn@4001 | 3362 | ins_pipe( pipe_slow ); |
kvn@4001 | 3363 | %} |
kvn@4001 | 3364 | |
kvn@4001 | 3365 | instruct vsub8F_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 3366 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 3367 | match(Set dst (SubVF src1 src2)); |
kvn@4001 | 3368 | format %{ "vsubps $dst,$src1,$src2\t! sub packed8F" %} |
kvn@4001 | 3369 | ins_encode %{ |
kvn@4001 | 3370 | bool vector256 = true; |
kvn@4001 | 3371 | __ vsubps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3372 | %} |
kvn@4001 | 3373 | ins_pipe( pipe_slow ); |
kvn@4001 | 3374 | %} |
kvn@4001 | 3375 | |
kvn@4001 | 3376 | instruct vsub8F_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 3377 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 3378 | match(Set dst (SubVF src (LoadVector mem))); |
kvn@4001 | 3379 | format %{ "vsubps $dst,$src,$mem\t! sub packed8F" %} |
kvn@4001 | 3380 | ins_encode %{ |
kvn@4001 | 3381 | bool vector256 = true; |
kvn@4001 | 3382 | __ vsubps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3383 | %} |
kvn@4001 | 3384 | ins_pipe( pipe_slow ); |
kvn@4001 | 3385 | %} |
kvn@4001 | 3386 | |
kvn@4001 | 3387 | // Doubles vector sub |
kvn@4001 | 3388 | instruct vsub2D(vecX dst, vecX src) %{ |
kvn@4001 | 3389 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 3390 | match(Set dst (SubVD dst src)); |
kvn@4001 | 3391 | format %{ "subpd $dst,$src\t! sub packed2D" %} |
kvn@4001 | 3392 | ins_encode %{ |
kvn@4001 | 3393 | __ subpd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3394 | %} |
kvn@4001 | 3395 | ins_pipe( pipe_slow ); |
kvn@4001 | 3396 | %} |
kvn@4001 | 3397 | |
kvn@4001 | 3398 | instruct vsub2D_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 3399 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3400 | match(Set dst (SubVD src1 src2)); |
kvn@4001 | 3401 | format %{ "vsubpd $dst,$src1,$src2\t! sub packed2D" %} |
kvn@4001 | 3402 | ins_encode %{ |
kvn@4001 | 3403 | bool vector256 = false; |
kvn@4001 | 3404 | __ vsubpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3405 | %} |
kvn@4001 | 3406 | ins_pipe( pipe_slow ); |
kvn@4001 | 3407 | %} |
kvn@4001 | 3408 | |
kvn@4001 | 3409 | instruct vsub2D_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 3410 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3411 | match(Set dst (SubVD src (LoadVector mem))); |
kvn@4001 | 3412 | format %{ "vsubpd $dst,$src,$mem\t! sub packed2D" %} |
kvn@4001 | 3413 | ins_encode %{ |
kvn@4001 | 3414 | bool vector256 = false; |
kvn@4001 | 3415 | __ vsubpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3416 | %} |
kvn@4001 | 3417 | ins_pipe( pipe_slow ); |
kvn@4001 | 3418 | %} |
kvn@4001 | 3419 | |
kvn@4001 | 3420 | instruct vsub4D_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 3421 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3422 | match(Set dst (SubVD src1 src2)); |
kvn@4001 | 3423 | format %{ "vsubpd $dst,$src1,$src2\t! sub packed4D" %} |
kvn@4001 | 3424 | ins_encode %{ |
kvn@4001 | 3425 | bool vector256 = true; |
kvn@4001 | 3426 | __ vsubpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3427 | %} |
kvn@4001 | 3428 | ins_pipe( pipe_slow ); |
kvn@4001 | 3429 | %} |
kvn@4001 | 3430 | |
kvn@4001 | 3431 | instruct vsub4D_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 3432 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3433 | match(Set dst (SubVD src (LoadVector mem))); |
kvn@4001 | 3434 | format %{ "vsubpd $dst,$src,$mem\t! sub packed4D" %} |
kvn@4001 | 3435 | ins_encode %{ |
kvn@4001 | 3436 | bool vector256 = true; |
kvn@4001 | 3437 | __ vsubpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3438 | %} |
kvn@4001 | 3439 | ins_pipe( pipe_slow ); |
kvn@4001 | 3440 | %} |
kvn@4001 | 3441 | |
kvn@4001 | 3442 | // --------------------------------- MUL -------------------------------------- |
kvn@4001 | 3443 | |
kvn@4001 | 3444 | // Shorts/Chars vector mul |
kvn@4001 | 3445 | instruct vmul2S(vecS dst, vecS src) %{ |
kvn@4001 | 3446 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 3447 | match(Set dst (MulVS dst src)); |
kvn@4001 | 3448 | format %{ "pmullw $dst,$src\t! mul packed2S" %} |
kvn@4001 | 3449 | ins_encode %{ |
kvn@4001 | 3450 | __ pmullw($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3451 | %} |
kvn@4001 | 3452 | ins_pipe( pipe_slow ); |
kvn@4001 | 3453 | %} |
kvn@4001 | 3454 | |
kvn@4001 | 3455 | instruct vmul2S_reg(vecS dst, vecS src1, vecS src2) %{ |
kvn@4001 | 3456 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3457 | match(Set dst (MulVS src1 src2)); |
kvn@4001 | 3458 | format %{ "vpmullw $dst,$src1,$src2\t! mul packed2S" %} |
kvn@4001 | 3459 | ins_encode %{ |
kvn@4001 | 3460 | bool vector256 = false; |
kvn@4001 | 3461 | __ vpmullw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3462 | %} |
kvn@4001 | 3463 | ins_pipe( pipe_slow ); |
kvn@4001 | 3464 | %} |
kvn@4001 | 3465 | |
kvn@4001 | 3466 | instruct vmul4S(vecD dst, vecD src) %{ |
kvn@4001 | 3467 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 3468 | match(Set dst (MulVS dst src)); |
kvn@4001 | 3469 | format %{ "pmullw $dst,$src\t! mul packed4S" %} |
kvn@4001 | 3470 | ins_encode %{ |
kvn@4001 | 3471 | __ pmullw($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3472 | %} |
kvn@4001 | 3473 | ins_pipe( pipe_slow ); |
kvn@4001 | 3474 | %} |
kvn@4001 | 3475 | |
kvn@4001 | 3476 | instruct vmul4S_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 3477 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3478 | match(Set dst (MulVS src1 src2)); |
kvn@4001 | 3479 | format %{ "vpmullw $dst,$src1,$src2\t! mul packed4S" %} |
kvn@4001 | 3480 | ins_encode %{ |
kvn@4001 | 3481 | bool vector256 = false; |
kvn@4001 | 3482 | __ vpmullw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3483 | %} |
kvn@4001 | 3484 | ins_pipe( pipe_slow ); |
kvn@4001 | 3485 | %} |
kvn@4001 | 3486 | |
kvn@4001 | 3487 | instruct vmul8S(vecX dst, vecX src) %{ |
kvn@4001 | 3488 | predicate(n->as_Vector()->length() == 8); |
kvn@4001 | 3489 | match(Set dst (MulVS dst src)); |
kvn@4001 | 3490 | format %{ "pmullw $dst,$src\t! mul packed8S" %} |
kvn@4001 | 3491 | ins_encode %{ |
kvn@4001 | 3492 | __ pmullw($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3493 | %} |
kvn@4001 | 3494 | ins_pipe( pipe_slow ); |
kvn@4001 | 3495 | %} |
kvn@4001 | 3496 | |
kvn@4001 | 3497 | instruct vmul8S_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 3498 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 3499 | match(Set dst (MulVS src1 src2)); |
kvn@4001 | 3500 | format %{ "vpmullw $dst,$src1,$src2\t! mul packed8S" %} |
kvn@4001 | 3501 | ins_encode %{ |
kvn@4001 | 3502 | bool vector256 = false; |
kvn@4001 | 3503 | __ vpmullw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3504 | %} |
kvn@4001 | 3505 | ins_pipe( pipe_slow ); |
kvn@4001 | 3506 | %} |
kvn@4001 | 3507 | |
kvn@4001 | 3508 | instruct vmul8S_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 3509 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 3510 | match(Set dst (MulVS src (LoadVector mem))); |
kvn@4001 | 3511 | format %{ "vpmullw $dst,$src,$mem\t! mul packed8S" %} |
kvn@4001 | 3512 | ins_encode %{ |
kvn@4001 | 3513 | bool vector256 = false; |
kvn@4001 | 3514 | __ vpmullw($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3515 | %} |
kvn@4001 | 3516 | ins_pipe( pipe_slow ); |
kvn@4001 | 3517 | %} |
kvn@4001 | 3518 | |
kvn@4001 | 3519 | instruct vmul16S_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 3520 | predicate(UseAVX > 1 && n->as_Vector()->length() == 16); |
kvn@4001 | 3521 | match(Set dst (MulVS src1 src2)); |
kvn@4001 | 3522 | format %{ "vpmullw $dst,$src1,$src2\t! mul packed16S" %} |
kvn@4001 | 3523 | ins_encode %{ |
kvn@4001 | 3524 | bool vector256 = true; |
kvn@4001 | 3525 | __ vpmullw($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3526 | %} |
kvn@4001 | 3527 | ins_pipe( pipe_slow ); |
kvn@4001 | 3528 | %} |
kvn@4001 | 3529 | |
kvn@4001 | 3530 | instruct vmul16S_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 3531 | predicate(UseAVX > 1 && n->as_Vector()->length() == 16); |
kvn@4001 | 3532 | match(Set dst (MulVS src (LoadVector mem))); |
kvn@4001 | 3533 | format %{ "vpmullw $dst,$src,$mem\t! mul packed16S" %} |
kvn@4001 | 3534 | ins_encode %{ |
kvn@4001 | 3535 | bool vector256 = true; |
kvn@4001 | 3536 | __ vpmullw($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3537 | %} |
kvn@4001 | 3538 | ins_pipe( pipe_slow ); |
kvn@4001 | 3539 | %} |
kvn@4001 | 3540 | |
kvn@4001 | 3541 | // Integers vector mul (sse4_1) |
kvn@4001 | 3542 | instruct vmul2I(vecD dst, vecD src) %{ |
kvn@4001 | 3543 | predicate(UseSSE > 3 && n->as_Vector()->length() == 2); |
kvn@4001 | 3544 | match(Set dst (MulVI dst src)); |
kvn@4001 | 3545 | format %{ "pmulld $dst,$src\t! mul packed2I" %} |
kvn@4001 | 3546 | ins_encode %{ |
kvn@4001 | 3547 | __ pmulld($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3548 | %} |
kvn@4001 | 3549 | ins_pipe( pipe_slow ); |
kvn@4001 | 3550 | %} |
kvn@4001 | 3551 | |
kvn@4001 | 3552 | instruct vmul2I_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 3553 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3554 | match(Set dst (MulVI src1 src2)); |
kvn@4001 | 3555 | format %{ "vpmulld $dst,$src1,$src2\t! mul packed2I" %} |
kvn@4001 | 3556 | ins_encode %{ |
kvn@4001 | 3557 | bool vector256 = false; |
kvn@4001 | 3558 | __ vpmulld($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3559 | %} |
kvn@4001 | 3560 | ins_pipe( pipe_slow ); |
kvn@4001 | 3561 | %} |
kvn@4001 | 3562 | |
kvn@4001 | 3563 | instruct vmul4I(vecX dst, vecX src) %{ |
kvn@4001 | 3564 | predicate(UseSSE > 3 && n->as_Vector()->length() == 4); |
kvn@4001 | 3565 | match(Set dst (MulVI dst src)); |
kvn@4001 | 3566 | format %{ "pmulld $dst,$src\t! mul packed4I" %} |
kvn@4001 | 3567 | ins_encode %{ |
kvn@4001 | 3568 | __ pmulld($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3569 | %} |
kvn@4001 | 3570 | ins_pipe( pipe_slow ); |
kvn@4001 | 3571 | %} |
kvn@4001 | 3572 | |
kvn@4001 | 3573 | instruct vmul4I_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 3574 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3575 | match(Set dst (MulVI src1 src2)); |
kvn@4001 | 3576 | format %{ "vpmulld $dst,$src1,$src2\t! mul packed4I" %} |
kvn@4001 | 3577 | ins_encode %{ |
kvn@4001 | 3578 | bool vector256 = false; |
kvn@4001 | 3579 | __ vpmulld($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3580 | %} |
kvn@4001 | 3581 | ins_pipe( pipe_slow ); |
kvn@4001 | 3582 | %} |
kvn@4001 | 3583 | |
kvn@4001 | 3584 | instruct vmul4I_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 3585 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3586 | match(Set dst (MulVI src (LoadVector mem))); |
kvn@4001 | 3587 | format %{ "vpmulld $dst,$src,$mem\t! mul packed4I" %} |
kvn@4001 | 3588 | ins_encode %{ |
kvn@4001 | 3589 | bool vector256 = false; |
kvn@4001 | 3590 | __ vpmulld($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3591 | %} |
kvn@4001 | 3592 | ins_pipe( pipe_slow ); |
kvn@4001 | 3593 | %} |
kvn@4001 | 3594 | |
kvn@4001 | 3595 | instruct vmul8I_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 3596 | predicate(UseAVX > 1 && n->as_Vector()->length() == 8); |
kvn@4001 | 3597 | match(Set dst (MulVI src1 src2)); |
kvn@4001 | 3598 | format %{ "vpmulld $dst,$src1,$src2\t! mul packed8I" %} |
kvn@4001 | 3599 | ins_encode %{ |
kvn@4001 | 3600 | bool vector256 = true; |
kvn@4001 | 3601 | __ vpmulld($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3602 | %} |
kvn@4001 | 3603 | ins_pipe( pipe_slow ); |
kvn@4001 | 3604 | %} |
kvn@4001 | 3605 | |
kvn@4001 | 3606 | instruct vmul8I_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 3607 | predicate(UseAVX > 1 && n->as_Vector()->length() == 8); |
kvn@4001 | 3608 | match(Set dst (MulVI src (LoadVector mem))); |
kvn@4001 | 3609 | format %{ "vpmulld $dst,$src,$mem\t! mul packed8I" %} |
kvn@4001 | 3610 | ins_encode %{ |
kvn@4001 | 3611 | bool vector256 = true; |
kvn@4001 | 3612 | __ vpmulld($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3613 | %} |
kvn@4001 | 3614 | ins_pipe( pipe_slow ); |
kvn@4001 | 3615 | %} |
kvn@4001 | 3616 | |
kvn@4001 | 3617 | // Floats vector mul |
kvn@4001 | 3618 | instruct vmul2F(vecD dst, vecD src) %{ |
kvn@4001 | 3619 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 3620 | match(Set dst (MulVF dst src)); |
kvn@4001 | 3621 | format %{ "mulps $dst,$src\t! mul packed2F" %} |
kvn@4001 | 3622 | ins_encode %{ |
kvn@4001 | 3623 | __ mulps($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3624 | %} |
kvn@4001 | 3625 | ins_pipe( pipe_slow ); |
kvn@4001 | 3626 | %} |
kvn@4001 | 3627 | |
kvn@4001 | 3628 | instruct vmul2F_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 3629 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3630 | match(Set dst (MulVF src1 src2)); |
kvn@4001 | 3631 | format %{ "vmulps $dst,$src1,$src2\t! mul packed2F" %} |
kvn@4001 | 3632 | ins_encode %{ |
kvn@4001 | 3633 | bool vector256 = false; |
kvn@4001 | 3634 | __ vmulps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3635 | %} |
kvn@4001 | 3636 | ins_pipe( pipe_slow ); |
kvn@4001 | 3637 | %} |
kvn@4001 | 3638 | |
kvn@4001 | 3639 | instruct vmul4F(vecX dst, vecX src) %{ |
kvn@4001 | 3640 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 3641 | match(Set dst (MulVF dst src)); |
kvn@4001 | 3642 | format %{ "mulps $dst,$src\t! mul packed4F" %} |
kvn@4001 | 3643 | ins_encode %{ |
kvn@4001 | 3644 | __ mulps($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3645 | %} |
kvn@4001 | 3646 | ins_pipe( pipe_slow ); |
kvn@4001 | 3647 | %} |
kvn@4001 | 3648 | |
kvn@4001 | 3649 | instruct vmul4F_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 3650 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3651 | match(Set dst (MulVF src1 src2)); |
kvn@4001 | 3652 | format %{ "vmulps $dst,$src1,$src2\t! mul packed4F" %} |
kvn@4001 | 3653 | ins_encode %{ |
kvn@4001 | 3654 | bool vector256 = false; |
kvn@4001 | 3655 | __ vmulps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3656 | %} |
kvn@4001 | 3657 | ins_pipe( pipe_slow ); |
kvn@4001 | 3658 | %} |
kvn@4001 | 3659 | |
kvn@4001 | 3660 | instruct vmul4F_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 3661 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3662 | match(Set dst (MulVF src (LoadVector mem))); |
kvn@4001 | 3663 | format %{ "vmulps $dst,$src,$mem\t! mul packed4F" %} |
kvn@4001 | 3664 | ins_encode %{ |
kvn@4001 | 3665 | bool vector256 = false; |
kvn@4001 | 3666 | __ vmulps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3667 | %} |
kvn@4001 | 3668 | ins_pipe( pipe_slow ); |
kvn@4001 | 3669 | %} |
kvn@4001 | 3670 | |
kvn@4001 | 3671 | instruct vmul8F_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 3672 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 3673 | match(Set dst (MulVF src1 src2)); |
kvn@4001 | 3674 | format %{ "vmulps $dst,$src1,$src2\t! mul packed8F" %} |
kvn@4001 | 3675 | ins_encode %{ |
kvn@4001 | 3676 | bool vector256 = true; |
kvn@4001 | 3677 | __ vmulps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3678 | %} |
kvn@4001 | 3679 | ins_pipe( pipe_slow ); |
kvn@4001 | 3680 | %} |
kvn@4001 | 3681 | |
kvn@4001 | 3682 | instruct vmul8F_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 3683 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 3684 | match(Set dst (MulVF src (LoadVector mem))); |
kvn@4001 | 3685 | format %{ "vmulps $dst,$src,$mem\t! mul packed8F" %} |
kvn@4001 | 3686 | ins_encode %{ |
kvn@4001 | 3687 | bool vector256 = true; |
kvn@4001 | 3688 | __ vmulps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3689 | %} |
kvn@4001 | 3690 | ins_pipe( pipe_slow ); |
kvn@4001 | 3691 | %} |
kvn@4001 | 3692 | |
kvn@4001 | 3693 | // Doubles vector mul |
kvn@4001 | 3694 | instruct vmul2D(vecX dst, vecX src) %{ |
kvn@4001 | 3695 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 3696 | match(Set dst (MulVD dst src)); |
kvn@4001 | 3697 | format %{ "mulpd $dst,$src\t! mul packed2D" %} |
kvn@4001 | 3698 | ins_encode %{ |
kvn@4001 | 3699 | __ mulpd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3700 | %} |
kvn@4001 | 3701 | ins_pipe( pipe_slow ); |
kvn@4001 | 3702 | %} |
kvn@4001 | 3703 | |
kvn@4001 | 3704 | instruct vmul2D_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 3705 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3706 | match(Set dst (MulVD src1 src2)); |
kvn@4001 | 3707 | format %{ "vmulpd $dst,$src1,$src2\t! mul packed2D" %} |
kvn@4001 | 3708 | ins_encode %{ |
kvn@4001 | 3709 | bool vector256 = false; |
kvn@4001 | 3710 | __ vmulpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3711 | %} |
kvn@4001 | 3712 | ins_pipe( pipe_slow ); |
kvn@4001 | 3713 | %} |
kvn@4001 | 3714 | |
kvn@4001 | 3715 | instruct vmul2D_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 3716 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3717 | match(Set dst (MulVD src (LoadVector mem))); |
kvn@4001 | 3718 | format %{ "vmulpd $dst,$src,$mem\t! mul packed2D" %} |
kvn@4001 | 3719 | ins_encode %{ |
kvn@4001 | 3720 | bool vector256 = false; |
kvn@4001 | 3721 | __ vmulpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3722 | %} |
kvn@4001 | 3723 | ins_pipe( pipe_slow ); |
kvn@4001 | 3724 | %} |
kvn@4001 | 3725 | |
kvn@4001 | 3726 | instruct vmul4D_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 3727 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3728 | match(Set dst (MulVD src1 src2)); |
kvn@4001 | 3729 | format %{ "vmulpd $dst,$src1,$src2\t! mul packed4D" %} |
kvn@4001 | 3730 | ins_encode %{ |
kvn@4001 | 3731 | bool vector256 = true; |
kvn@4001 | 3732 | __ vmulpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3733 | %} |
kvn@4001 | 3734 | ins_pipe( pipe_slow ); |
kvn@4001 | 3735 | %} |
kvn@4001 | 3736 | |
kvn@4001 | 3737 | instruct vmul4D_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 3738 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3739 | match(Set dst (MulVD src (LoadVector mem))); |
kvn@4001 | 3740 | format %{ "vmulpd $dst,$src,$mem\t! mul packed4D" %} |
kvn@4001 | 3741 | ins_encode %{ |
kvn@4001 | 3742 | bool vector256 = true; |
kvn@4001 | 3743 | __ vmulpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3744 | %} |
kvn@4001 | 3745 | ins_pipe( pipe_slow ); |
kvn@4001 | 3746 | %} |
kvn@4001 | 3747 | |
kvn@4001 | 3748 | // --------------------------------- DIV -------------------------------------- |
kvn@4001 | 3749 | |
kvn@4001 | 3750 | // Floats vector div |
kvn@4001 | 3751 | instruct vdiv2F(vecD dst, vecD src) %{ |
kvn@4001 | 3752 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 3753 | match(Set dst (DivVF dst src)); |
kvn@4001 | 3754 | format %{ "divps $dst,$src\t! div packed2F" %} |
kvn@4001 | 3755 | ins_encode %{ |
kvn@4001 | 3756 | __ divps($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3757 | %} |
kvn@4001 | 3758 | ins_pipe( pipe_slow ); |
kvn@4001 | 3759 | %} |
kvn@4001 | 3760 | |
kvn@4001 | 3761 | instruct vdiv2F_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 3762 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3763 | match(Set dst (DivVF src1 src2)); |
kvn@4001 | 3764 | format %{ "vdivps $dst,$src1,$src2\t! div packed2F" %} |
kvn@4001 | 3765 | ins_encode %{ |
kvn@4001 | 3766 | bool vector256 = false; |
kvn@4001 | 3767 | __ vdivps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3768 | %} |
kvn@4001 | 3769 | ins_pipe( pipe_slow ); |
kvn@4001 | 3770 | %} |
kvn@4001 | 3771 | |
kvn@4001 | 3772 | instruct vdiv4F(vecX dst, vecX src) %{ |
kvn@4001 | 3773 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 3774 | match(Set dst (DivVF dst src)); |
kvn@4001 | 3775 | format %{ "divps $dst,$src\t! div packed4F" %} |
kvn@4001 | 3776 | ins_encode %{ |
kvn@4001 | 3777 | __ divps($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3778 | %} |
kvn@4001 | 3779 | ins_pipe( pipe_slow ); |
kvn@4001 | 3780 | %} |
kvn@4001 | 3781 | |
kvn@4001 | 3782 | instruct vdiv4F_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 3783 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3784 | match(Set dst (DivVF src1 src2)); |
kvn@4001 | 3785 | format %{ "vdivps $dst,$src1,$src2\t! div packed4F" %} |
kvn@4001 | 3786 | ins_encode %{ |
kvn@4001 | 3787 | bool vector256 = false; |
kvn@4001 | 3788 | __ vdivps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3789 | %} |
kvn@4001 | 3790 | ins_pipe( pipe_slow ); |
kvn@4001 | 3791 | %} |
kvn@4001 | 3792 | |
kvn@4001 | 3793 | instruct vdiv4F_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 3794 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3795 | match(Set dst (DivVF src (LoadVector mem))); |
kvn@4001 | 3796 | format %{ "vdivps $dst,$src,$mem\t! div packed4F" %} |
kvn@4001 | 3797 | ins_encode %{ |
kvn@4001 | 3798 | bool vector256 = false; |
kvn@4001 | 3799 | __ vdivps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3800 | %} |
kvn@4001 | 3801 | ins_pipe( pipe_slow ); |
kvn@4001 | 3802 | %} |
kvn@4001 | 3803 | |
kvn@4001 | 3804 | instruct vdiv8F_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 3805 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 3806 | match(Set dst (DivVF src1 src2)); |
kvn@4001 | 3807 | format %{ "vdivps $dst,$src1,$src2\t! div packed8F" %} |
kvn@4001 | 3808 | ins_encode %{ |
kvn@4001 | 3809 | bool vector256 = true; |
kvn@4001 | 3810 | __ vdivps($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3811 | %} |
kvn@4001 | 3812 | ins_pipe( pipe_slow ); |
kvn@4001 | 3813 | %} |
kvn@4001 | 3814 | |
kvn@4001 | 3815 | instruct vdiv8F_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 3816 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 3817 | match(Set dst (DivVF src (LoadVector mem))); |
kvn@4001 | 3818 | format %{ "vdivps $dst,$src,$mem\t! div packed8F" %} |
kvn@4001 | 3819 | ins_encode %{ |
kvn@4001 | 3820 | bool vector256 = true; |
kvn@4001 | 3821 | __ vdivps($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3822 | %} |
kvn@4001 | 3823 | ins_pipe( pipe_slow ); |
kvn@4001 | 3824 | %} |
kvn@4001 | 3825 | |
kvn@4001 | 3826 | // Doubles vector div |
kvn@4001 | 3827 | instruct vdiv2D(vecX dst, vecX src) %{ |
kvn@4001 | 3828 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 3829 | match(Set dst (DivVD dst src)); |
kvn@4001 | 3830 | format %{ "divpd $dst,$src\t! div packed2D" %} |
kvn@4001 | 3831 | ins_encode %{ |
kvn@4001 | 3832 | __ divpd($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 3833 | %} |
kvn@4001 | 3834 | ins_pipe( pipe_slow ); |
kvn@4001 | 3835 | %} |
kvn@4001 | 3836 | |
kvn@4001 | 3837 | instruct vdiv2D_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 3838 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3839 | match(Set dst (DivVD src1 src2)); |
kvn@4001 | 3840 | format %{ "vdivpd $dst,$src1,$src2\t! div packed2D" %} |
kvn@4001 | 3841 | ins_encode %{ |
kvn@4001 | 3842 | bool vector256 = false; |
kvn@4001 | 3843 | __ vdivpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3844 | %} |
kvn@4001 | 3845 | ins_pipe( pipe_slow ); |
kvn@4001 | 3846 | %} |
kvn@4001 | 3847 | |
kvn@4001 | 3848 | instruct vdiv2D_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 3849 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3850 | match(Set dst (DivVD src (LoadVector mem))); |
kvn@4001 | 3851 | format %{ "vdivpd $dst,$src,$mem\t! div packed2D" %} |
kvn@4001 | 3852 | ins_encode %{ |
kvn@4001 | 3853 | bool vector256 = false; |
kvn@4001 | 3854 | __ vdivpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3855 | %} |
kvn@4001 | 3856 | ins_pipe( pipe_slow ); |
kvn@4001 | 3857 | %} |
kvn@4001 | 3858 | |
kvn@4001 | 3859 | instruct vdiv4D_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 3860 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3861 | match(Set dst (DivVD src1 src2)); |
kvn@4001 | 3862 | format %{ "vdivpd $dst,$src1,$src2\t! div packed4D" %} |
kvn@4001 | 3863 | ins_encode %{ |
kvn@4001 | 3864 | bool vector256 = true; |
kvn@4001 | 3865 | __ vdivpd($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 3866 | %} |
kvn@4001 | 3867 | ins_pipe( pipe_slow ); |
kvn@4001 | 3868 | %} |
kvn@4001 | 3869 | |
kvn@4001 | 3870 | instruct vdiv4D_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 3871 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3872 | match(Set dst (DivVD src (LoadVector mem))); |
kvn@4001 | 3873 | format %{ "vdivpd $dst,$src,$mem\t! div packed4D" %} |
kvn@4001 | 3874 | ins_encode %{ |
kvn@4001 | 3875 | bool vector256 = true; |
kvn@4001 | 3876 | __ vdivpd($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 3877 | %} |
kvn@4001 | 3878 | ins_pipe( pipe_slow ); |
kvn@4001 | 3879 | %} |
kvn@4001 | 3880 | |
kvn@4134 | 3881 | // ------------------------------ Shift --------------------------------------- |
kvn@4134 | 3882 | |
kvn@4134 | 3883 | // Left and right shift count vectors are the same on x86 |
kvn@4134 | 3884 | // (only lowest bits of xmm reg are used for count). |
kvn@4134 | 3885 | instruct vshiftcnt(vecS dst, rRegI cnt) %{ |
kvn@4134 | 3886 | match(Set dst (LShiftCntV cnt)); |
kvn@4134 | 3887 | match(Set dst (RShiftCntV cnt)); |
kvn@4134 | 3888 | format %{ "movd $dst,$cnt\t! load shift count" %} |
kvn@4134 | 3889 | ins_encode %{ |
kvn@4134 | 3890 | __ movdl($dst$$XMMRegister, $cnt$$Register); |
kvn@4134 | 3891 | %} |
kvn@4134 | 3892 | ins_pipe( pipe_slow ); |
kvn@4134 | 3893 | %} |
kvn@4134 | 3894 | |
kvn@4001 | 3895 | // ------------------------------ LeftShift ----------------------------------- |
kvn@4001 | 3896 | |
kvn@4001 | 3897 | // Shorts/Chars vector left shift |
kvn@4134 | 3898 | instruct vsll2S(vecS dst, vecS shift) %{ |
kvn@4001 | 3899 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 3900 | match(Set dst (LShiftVS dst shift)); |
kvn@4001 | 3901 | format %{ "psllw $dst,$shift\t! left shift packed2S" %} |
kvn@4001 | 3902 | ins_encode %{ |
kvn@4001 | 3903 | __ psllw($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 3904 | %} |
kvn@4001 | 3905 | ins_pipe( pipe_slow ); |
kvn@4001 | 3906 | %} |
kvn@4001 | 3907 | |
kvn@4001 | 3908 | instruct vsll2S_imm(vecS dst, immI8 shift) %{ |
kvn@4001 | 3909 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 3910 | match(Set dst (LShiftVS dst shift)); |
kvn@4001 | 3911 | format %{ "psllw $dst,$shift\t! left shift packed2S" %} |
kvn@4001 | 3912 | ins_encode %{ |
kvn@4001 | 3913 | __ psllw($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 3914 | %} |
kvn@4001 | 3915 | ins_pipe( pipe_slow ); |
kvn@4001 | 3916 | %} |
kvn@4001 | 3917 | |
kvn@4134 | 3918 | instruct vsll2S_reg(vecS dst, vecS src, vecS shift) %{ |
kvn@4001 | 3919 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3920 | match(Set dst (LShiftVS src shift)); |
kvn@4001 | 3921 | format %{ "vpsllw $dst,$src,$shift\t! left shift packed2S" %} |
kvn@4001 | 3922 | ins_encode %{ |
kvn@4001 | 3923 | bool vector256 = false; |
kvn@4001 | 3924 | __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 3925 | %} |
kvn@4001 | 3926 | ins_pipe( pipe_slow ); |
kvn@4001 | 3927 | %} |
kvn@4001 | 3928 | |
kvn@4001 | 3929 | instruct vsll2S_reg_imm(vecS dst, vecS src, immI8 shift) %{ |
kvn@4001 | 3930 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 3931 | match(Set dst (LShiftVS src shift)); |
kvn@4001 | 3932 | format %{ "vpsllw $dst,$src,$shift\t! left shift packed2S" %} |
kvn@4001 | 3933 | ins_encode %{ |
kvn@4001 | 3934 | bool vector256 = false; |
kvn@4001 | 3935 | __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 3936 | %} |
kvn@4001 | 3937 | ins_pipe( pipe_slow ); |
kvn@4001 | 3938 | %} |
kvn@4001 | 3939 | |
kvn@4134 | 3940 | instruct vsll4S(vecD dst, vecS shift) %{ |
kvn@4001 | 3941 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 3942 | match(Set dst (LShiftVS dst shift)); |
kvn@4001 | 3943 | format %{ "psllw $dst,$shift\t! left shift packed4S" %} |
kvn@4001 | 3944 | ins_encode %{ |
kvn@4001 | 3945 | __ psllw($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 3946 | %} |
kvn@4001 | 3947 | ins_pipe( pipe_slow ); |
kvn@4001 | 3948 | %} |
kvn@4001 | 3949 | |
kvn@4001 | 3950 | instruct vsll4S_imm(vecD dst, immI8 shift) %{ |
kvn@4001 | 3951 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 3952 | match(Set dst (LShiftVS dst shift)); |
kvn@4001 | 3953 | format %{ "psllw $dst,$shift\t! left shift packed4S" %} |
kvn@4001 | 3954 | ins_encode %{ |
kvn@4001 | 3955 | __ psllw($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 3956 | %} |
kvn@4001 | 3957 | ins_pipe( pipe_slow ); |
kvn@4001 | 3958 | %} |
kvn@4001 | 3959 | |
kvn@4134 | 3960 | instruct vsll4S_reg(vecD dst, vecD src, vecS shift) %{ |
kvn@4001 | 3961 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3962 | match(Set dst (LShiftVS src shift)); |
kvn@4001 | 3963 | format %{ "vpsllw $dst,$src,$shift\t! left shift packed4S" %} |
kvn@4001 | 3964 | ins_encode %{ |
kvn@4001 | 3965 | bool vector256 = false; |
kvn@4001 | 3966 | __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 3967 | %} |
kvn@4001 | 3968 | ins_pipe( pipe_slow ); |
kvn@4001 | 3969 | %} |
kvn@4001 | 3970 | |
kvn@4001 | 3971 | instruct vsll4S_reg_imm(vecD dst, vecD src, immI8 shift) %{ |
kvn@4001 | 3972 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 3973 | match(Set dst (LShiftVS src shift)); |
kvn@4001 | 3974 | format %{ "vpsllw $dst,$src,$shift\t! left shift packed4S" %} |
kvn@4001 | 3975 | ins_encode %{ |
kvn@4001 | 3976 | bool vector256 = false; |
kvn@4001 | 3977 | __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 3978 | %} |
kvn@4001 | 3979 | ins_pipe( pipe_slow ); |
kvn@4001 | 3980 | %} |
kvn@4001 | 3981 | |
kvn@4134 | 3982 | instruct vsll8S(vecX dst, vecS shift) %{ |
kvn@4001 | 3983 | predicate(n->as_Vector()->length() == 8); |
kvn@4001 | 3984 | match(Set dst (LShiftVS dst shift)); |
kvn@4001 | 3985 | format %{ "psllw $dst,$shift\t! left shift packed8S" %} |
kvn@4001 | 3986 | ins_encode %{ |
kvn@4001 | 3987 | __ psllw($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 3988 | %} |
kvn@4001 | 3989 | ins_pipe( pipe_slow ); |
kvn@4001 | 3990 | %} |
kvn@4001 | 3991 | |
kvn@4001 | 3992 | instruct vsll8S_imm(vecX dst, immI8 shift) %{ |
kvn@4001 | 3993 | predicate(n->as_Vector()->length() == 8); |
kvn@4001 | 3994 | match(Set dst (LShiftVS dst shift)); |
kvn@4001 | 3995 | format %{ "psllw $dst,$shift\t! left shift packed8S" %} |
kvn@4001 | 3996 | ins_encode %{ |
kvn@4001 | 3997 | __ psllw($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 3998 | %} |
kvn@4001 | 3999 | ins_pipe( pipe_slow ); |
kvn@4001 | 4000 | %} |
kvn@4001 | 4001 | |
kvn@4134 | 4002 | instruct vsll8S_reg(vecX dst, vecX src, vecS shift) %{ |
kvn@4001 | 4003 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 4004 | match(Set dst (LShiftVS src shift)); |
kvn@4001 | 4005 | format %{ "vpsllw $dst,$src,$shift\t! left shift packed8S" %} |
kvn@4001 | 4006 | ins_encode %{ |
kvn@4001 | 4007 | bool vector256 = false; |
kvn@4001 | 4008 | __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4009 | %} |
kvn@4001 | 4010 | ins_pipe( pipe_slow ); |
kvn@4001 | 4011 | %} |
kvn@4001 | 4012 | |
kvn@4001 | 4013 | instruct vsll8S_reg_imm(vecX dst, vecX src, immI8 shift) %{ |
kvn@4001 | 4014 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 4015 | match(Set dst (LShiftVS src shift)); |
kvn@4001 | 4016 | format %{ "vpsllw $dst,$src,$shift\t! left shift packed8S" %} |
kvn@4001 | 4017 | ins_encode %{ |
kvn@4001 | 4018 | bool vector256 = false; |
kvn@4001 | 4019 | __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4020 | %} |
kvn@4001 | 4021 | ins_pipe( pipe_slow ); |
kvn@4001 | 4022 | %} |
kvn@4001 | 4023 | |
kvn@4134 | 4024 | instruct vsll16S_reg(vecY dst, vecY src, vecS shift) %{ |
kvn@4001 | 4025 | predicate(UseAVX > 1 && n->as_Vector()->length() == 16); |
kvn@4001 | 4026 | match(Set dst (LShiftVS src shift)); |
kvn@4001 | 4027 | format %{ "vpsllw $dst,$src,$shift\t! left shift packed16S" %} |
kvn@4001 | 4028 | ins_encode %{ |
kvn@4001 | 4029 | bool vector256 = true; |
kvn@4001 | 4030 | __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4031 | %} |
kvn@4001 | 4032 | ins_pipe( pipe_slow ); |
kvn@4001 | 4033 | %} |
kvn@4001 | 4034 | |
kvn@4001 | 4035 | instruct vsll16S_reg_imm(vecY dst, vecY src, immI8 shift) %{ |
kvn@4001 | 4036 | predicate(UseAVX > 1 && n->as_Vector()->length() == 16); |
kvn@4001 | 4037 | match(Set dst (LShiftVS src shift)); |
kvn@4001 | 4038 | format %{ "vpsllw $dst,$src,$shift\t! left shift packed16S" %} |
kvn@4001 | 4039 | ins_encode %{ |
kvn@4001 | 4040 | bool vector256 = true; |
kvn@4001 | 4041 | __ vpsllw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4042 | %} |
kvn@4001 | 4043 | ins_pipe( pipe_slow ); |
kvn@4001 | 4044 | %} |
kvn@4001 | 4045 | |
kvn@4001 | 4046 | // Integers vector left shift |
kvn@4134 | 4047 | instruct vsll2I(vecD dst, vecS shift) %{ |
kvn@4001 | 4048 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 4049 | match(Set dst (LShiftVI dst shift)); |
kvn@4001 | 4050 | format %{ "pslld $dst,$shift\t! left shift packed2I" %} |
kvn@4001 | 4051 | ins_encode %{ |
kvn@4001 | 4052 | __ pslld($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 4053 | %} |
kvn@4001 | 4054 | ins_pipe( pipe_slow ); |
kvn@4001 | 4055 | %} |
kvn@4001 | 4056 | |
kvn@4001 | 4057 | instruct vsll2I_imm(vecD dst, immI8 shift) %{ |
kvn@4001 | 4058 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 4059 | match(Set dst (LShiftVI dst shift)); |
kvn@4001 | 4060 | format %{ "pslld $dst,$shift\t! left shift packed2I" %} |
kvn@4001 | 4061 | ins_encode %{ |
kvn@4001 | 4062 | __ pslld($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 4063 | %} |
kvn@4001 | 4064 | ins_pipe( pipe_slow ); |
kvn@4001 | 4065 | %} |
kvn@4001 | 4066 | |
kvn@4134 | 4067 | instruct vsll2I_reg(vecD dst, vecD src, vecS shift) %{ |
kvn@4001 | 4068 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 4069 | match(Set dst (LShiftVI src shift)); |
kvn@4001 | 4070 | format %{ "vpslld $dst,$src,$shift\t! left shift packed2I" %} |
kvn@4001 | 4071 | ins_encode %{ |
kvn@4001 | 4072 | bool vector256 = false; |
kvn@4001 | 4073 | __ vpslld($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4074 | %} |
kvn@4001 | 4075 | ins_pipe( pipe_slow ); |
kvn@4001 | 4076 | %} |
kvn@4001 | 4077 | |
kvn@4001 | 4078 | instruct vsll2I_reg_imm(vecD dst, vecD src, immI8 shift) %{ |
kvn@4001 | 4079 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 4080 | match(Set dst (LShiftVI src shift)); |
kvn@4001 | 4081 | format %{ "vpslld $dst,$src,$shift\t! left shift packed2I" %} |
kvn@4001 | 4082 | ins_encode %{ |
kvn@4001 | 4083 | bool vector256 = false; |
kvn@4001 | 4084 | __ vpslld($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4085 | %} |
kvn@4001 | 4086 | ins_pipe( pipe_slow ); |
kvn@4001 | 4087 | %} |
kvn@4001 | 4088 | |
kvn@4134 | 4089 | instruct vsll4I(vecX dst, vecS shift) %{ |
kvn@4001 | 4090 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 4091 | match(Set dst (LShiftVI dst shift)); |
kvn@4001 | 4092 | format %{ "pslld $dst,$shift\t! left shift packed4I" %} |
kvn@4001 | 4093 | ins_encode %{ |
kvn@4001 | 4094 | __ pslld($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 4095 | %} |
kvn@4001 | 4096 | ins_pipe( pipe_slow ); |
kvn@4001 | 4097 | %} |
kvn@4001 | 4098 | |
kvn@4001 | 4099 | instruct vsll4I_imm(vecX dst, immI8 shift) %{ |
kvn@4001 | 4100 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 4101 | match(Set dst (LShiftVI dst shift)); |
kvn@4001 | 4102 | format %{ "pslld $dst,$shift\t! left shift packed4I" %} |
kvn@4001 | 4103 | ins_encode %{ |
kvn@4001 | 4104 | __ pslld($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 4105 | %} |
kvn@4001 | 4106 | ins_pipe( pipe_slow ); |
kvn@4001 | 4107 | %} |
kvn@4001 | 4108 | |
kvn@4134 | 4109 | instruct vsll4I_reg(vecX dst, vecX src, vecS shift) %{ |
kvn@4001 | 4110 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 4111 | match(Set dst (LShiftVI src shift)); |
kvn@4001 | 4112 | format %{ "vpslld $dst,$src,$shift\t! left shift packed4I" %} |
kvn@4001 | 4113 | ins_encode %{ |
kvn@4001 | 4114 | bool vector256 = false; |
kvn@4001 | 4115 | __ vpslld($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4116 | %} |
kvn@4001 | 4117 | ins_pipe( pipe_slow ); |
kvn@4001 | 4118 | %} |
kvn@4001 | 4119 | |
kvn@4001 | 4120 | instruct vsll4I_reg_imm(vecX dst, vecX src, immI8 shift) %{ |
kvn@4001 | 4121 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 4122 | match(Set dst (LShiftVI src shift)); |
kvn@4001 | 4123 | format %{ "vpslld $dst,$src,$shift\t! left shift packed4I" %} |
kvn@4001 | 4124 | ins_encode %{ |
kvn@4001 | 4125 | bool vector256 = false; |
kvn@4001 | 4126 | __ vpslld($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4127 | %} |
kvn@4001 | 4128 | ins_pipe( pipe_slow ); |
kvn@4001 | 4129 | %} |
kvn@4001 | 4130 | |
kvn@4134 | 4131 | instruct vsll8I_reg(vecY dst, vecY src, vecS shift) %{ |
kvn@4001 | 4132 | predicate(UseAVX > 1 && n->as_Vector()->length() == 8); |
kvn@4001 | 4133 | match(Set dst (LShiftVI src shift)); |
kvn@4001 | 4134 | format %{ "vpslld $dst,$src,$shift\t! left shift packed8I" %} |
kvn@4001 | 4135 | ins_encode %{ |
kvn@4001 | 4136 | bool vector256 = true; |
kvn@4001 | 4137 | __ vpslld($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4138 | %} |
kvn@4001 | 4139 | ins_pipe( pipe_slow ); |
kvn@4001 | 4140 | %} |
kvn@4001 | 4141 | |
kvn@4001 | 4142 | instruct vsll8I_reg_imm(vecY dst, vecY src, immI8 shift) %{ |
kvn@4001 | 4143 | predicate(UseAVX > 1 && n->as_Vector()->length() == 8); |
kvn@4001 | 4144 | match(Set dst (LShiftVI src shift)); |
kvn@4001 | 4145 | format %{ "vpslld $dst,$src,$shift\t! left shift packed8I" %} |
kvn@4001 | 4146 | ins_encode %{ |
kvn@4001 | 4147 | bool vector256 = true; |
kvn@4001 | 4148 | __ vpslld($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4149 | %} |
kvn@4001 | 4150 | ins_pipe( pipe_slow ); |
kvn@4001 | 4151 | %} |
kvn@4001 | 4152 | |
kvn@4001 | 4153 | // Longs vector left shift |
kvn@4134 | 4154 | instruct vsll2L(vecX dst, vecS shift) %{ |
kvn@4001 | 4155 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 4156 | match(Set dst (LShiftVL dst shift)); |
kvn@4001 | 4157 | format %{ "psllq $dst,$shift\t! left shift packed2L" %} |
kvn@4001 | 4158 | ins_encode %{ |
kvn@4001 | 4159 | __ psllq($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 4160 | %} |
kvn@4001 | 4161 | ins_pipe( pipe_slow ); |
kvn@4001 | 4162 | %} |
kvn@4001 | 4163 | |
kvn@4001 | 4164 | instruct vsll2L_imm(vecX dst, immI8 shift) %{ |
kvn@4001 | 4165 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 4166 | match(Set dst (LShiftVL dst shift)); |
kvn@4001 | 4167 | format %{ "psllq $dst,$shift\t! left shift packed2L" %} |
kvn@4001 | 4168 | ins_encode %{ |
kvn@4001 | 4169 | __ psllq($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 4170 | %} |
kvn@4001 | 4171 | ins_pipe( pipe_slow ); |
kvn@4001 | 4172 | %} |
kvn@4001 | 4173 | |
kvn@4134 | 4174 | instruct vsll2L_reg(vecX dst, vecX src, vecS shift) %{ |
kvn@4001 | 4175 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 4176 | match(Set dst (LShiftVL src shift)); |
kvn@4001 | 4177 | format %{ "vpsllq $dst,$src,$shift\t! left shift packed2L" %} |
kvn@4001 | 4178 | ins_encode %{ |
kvn@4001 | 4179 | bool vector256 = false; |
kvn@4001 | 4180 | __ vpsllq($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4181 | %} |
kvn@4001 | 4182 | ins_pipe( pipe_slow ); |
kvn@4001 | 4183 | %} |
kvn@4001 | 4184 | |
kvn@4001 | 4185 | instruct vsll2L_reg_imm(vecX dst, vecX src, immI8 shift) %{ |
kvn@4001 | 4186 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 4187 | match(Set dst (LShiftVL src shift)); |
kvn@4001 | 4188 | format %{ "vpsllq $dst,$src,$shift\t! left shift packed2L" %} |
kvn@4001 | 4189 | ins_encode %{ |
kvn@4001 | 4190 | bool vector256 = false; |
kvn@4001 | 4191 | __ vpsllq($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4192 | %} |
kvn@4001 | 4193 | ins_pipe( pipe_slow ); |
kvn@4001 | 4194 | %} |
kvn@4001 | 4195 | |
kvn@4134 | 4196 | instruct vsll4L_reg(vecY dst, vecY src, vecS shift) %{ |
kvn@4001 | 4197 | predicate(UseAVX > 1 && n->as_Vector()->length() == 4); |
kvn@4001 | 4198 | match(Set dst (LShiftVL src shift)); |
kvn@4001 | 4199 | format %{ "vpsllq $dst,$src,$shift\t! left shift packed4L" %} |
kvn@4001 | 4200 | ins_encode %{ |
kvn@4001 | 4201 | bool vector256 = true; |
kvn@4001 | 4202 | __ vpsllq($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4203 | %} |
kvn@4001 | 4204 | ins_pipe( pipe_slow ); |
kvn@4001 | 4205 | %} |
kvn@4001 | 4206 | |
kvn@4001 | 4207 | instruct vsll4L_reg_imm(vecY dst, vecY src, immI8 shift) %{ |
kvn@4001 | 4208 | predicate(UseAVX > 1 && n->as_Vector()->length() == 4); |
kvn@4001 | 4209 | match(Set dst (LShiftVL src shift)); |
kvn@4001 | 4210 | format %{ "vpsllq $dst,$src,$shift\t! left shift packed4L" %} |
kvn@4001 | 4211 | ins_encode %{ |
kvn@4001 | 4212 | bool vector256 = true; |
kvn@4001 | 4213 | __ vpsllq($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4214 | %} |
kvn@4001 | 4215 | ins_pipe( pipe_slow ); |
kvn@4001 | 4216 | %} |
kvn@4001 | 4217 | |
kvn@4001 | 4218 | // ----------------------- LogicalRightShift ----------------------------------- |
kvn@4001 | 4219 | |
kvn@4204 | 4220 | // Shorts vector logical right shift produces incorrect Java result |
kvn@4001 | 4221 | // for negative data because java code convert short value into int with |
kvn@4204 | 4222 | // sign extension before a shift. But char vectors are fine since chars are |
kvn@4204 | 4223 | // unsigned values. |
kvn@4204 | 4224 | |
kvn@4204 | 4225 | instruct vsrl2S(vecS dst, vecS shift) %{ |
kvn@4204 | 4226 | predicate(n->as_Vector()->length() == 2); |
kvn@4204 | 4227 | match(Set dst (URShiftVS dst shift)); |
kvn@4204 | 4228 | format %{ "psrlw $dst,$shift\t! logical right shift packed2S" %} |
kvn@4204 | 4229 | ins_encode %{ |
kvn@4204 | 4230 | __ psrlw($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4204 | 4231 | %} |
kvn@4204 | 4232 | ins_pipe( pipe_slow ); |
kvn@4204 | 4233 | %} |
kvn@4204 | 4234 | |
kvn@4204 | 4235 | instruct vsrl2S_imm(vecS dst, immI8 shift) %{ |
kvn@4204 | 4236 | predicate(n->as_Vector()->length() == 2); |
kvn@4204 | 4237 | match(Set dst (URShiftVS dst shift)); |
kvn@4204 | 4238 | format %{ "psrlw $dst,$shift\t! logical right shift packed2S" %} |
kvn@4204 | 4239 | ins_encode %{ |
kvn@4204 | 4240 | __ psrlw($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4204 | 4241 | %} |
kvn@4204 | 4242 | ins_pipe( pipe_slow ); |
kvn@4204 | 4243 | %} |
kvn@4204 | 4244 | |
kvn@4204 | 4245 | instruct vsrl2S_reg(vecS dst, vecS src, vecS shift) %{ |
kvn@4204 | 4246 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4204 | 4247 | match(Set dst (URShiftVS src shift)); |
kvn@4204 | 4248 | format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed2S" %} |
kvn@4204 | 4249 | ins_encode %{ |
kvn@4204 | 4250 | bool vector256 = false; |
kvn@4204 | 4251 | __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4204 | 4252 | %} |
kvn@4204 | 4253 | ins_pipe( pipe_slow ); |
kvn@4204 | 4254 | %} |
kvn@4204 | 4255 | |
kvn@4204 | 4256 | instruct vsrl2S_reg_imm(vecS dst, vecS src, immI8 shift) %{ |
kvn@4204 | 4257 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4204 | 4258 | match(Set dst (URShiftVS src shift)); |
kvn@4204 | 4259 | format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed2S" %} |
kvn@4204 | 4260 | ins_encode %{ |
kvn@4204 | 4261 | bool vector256 = false; |
kvn@4204 | 4262 | __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4204 | 4263 | %} |
kvn@4204 | 4264 | ins_pipe( pipe_slow ); |
kvn@4204 | 4265 | %} |
kvn@4204 | 4266 | |
kvn@4204 | 4267 | instruct vsrl4S(vecD dst, vecS shift) %{ |
kvn@4204 | 4268 | predicate(n->as_Vector()->length() == 4); |
kvn@4204 | 4269 | match(Set dst (URShiftVS dst shift)); |
kvn@4204 | 4270 | format %{ "psrlw $dst,$shift\t! logical right shift packed4S" %} |
kvn@4204 | 4271 | ins_encode %{ |
kvn@4204 | 4272 | __ psrlw($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4204 | 4273 | %} |
kvn@4204 | 4274 | ins_pipe( pipe_slow ); |
kvn@4204 | 4275 | %} |
kvn@4204 | 4276 | |
kvn@4204 | 4277 | instruct vsrl4S_imm(vecD dst, immI8 shift) %{ |
kvn@4204 | 4278 | predicate(n->as_Vector()->length() == 4); |
kvn@4204 | 4279 | match(Set dst (URShiftVS dst shift)); |
kvn@4204 | 4280 | format %{ "psrlw $dst,$shift\t! logical right shift packed4S" %} |
kvn@4204 | 4281 | ins_encode %{ |
kvn@4204 | 4282 | __ psrlw($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4204 | 4283 | %} |
kvn@4204 | 4284 | ins_pipe( pipe_slow ); |
kvn@4204 | 4285 | %} |
kvn@4204 | 4286 | |
kvn@4204 | 4287 | instruct vsrl4S_reg(vecD dst, vecD src, vecS shift) %{ |
kvn@4204 | 4288 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4204 | 4289 | match(Set dst (URShiftVS src shift)); |
kvn@4204 | 4290 | format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed4S" %} |
kvn@4204 | 4291 | ins_encode %{ |
kvn@4204 | 4292 | bool vector256 = false; |
kvn@4204 | 4293 | __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4204 | 4294 | %} |
kvn@4204 | 4295 | ins_pipe( pipe_slow ); |
kvn@4204 | 4296 | %} |
kvn@4204 | 4297 | |
kvn@4204 | 4298 | instruct vsrl4S_reg_imm(vecD dst, vecD src, immI8 shift) %{ |
kvn@4204 | 4299 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4204 | 4300 | match(Set dst (URShiftVS src shift)); |
kvn@4204 | 4301 | format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed4S" %} |
kvn@4204 | 4302 | ins_encode %{ |
kvn@4204 | 4303 | bool vector256 = false; |
kvn@4204 | 4304 | __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4204 | 4305 | %} |
kvn@4204 | 4306 | ins_pipe( pipe_slow ); |
kvn@4204 | 4307 | %} |
kvn@4204 | 4308 | |
kvn@4204 | 4309 | instruct vsrl8S(vecX dst, vecS shift) %{ |
kvn@4204 | 4310 | predicate(n->as_Vector()->length() == 8); |
kvn@4204 | 4311 | match(Set dst (URShiftVS dst shift)); |
kvn@4204 | 4312 | format %{ "psrlw $dst,$shift\t! logical right shift packed8S" %} |
kvn@4204 | 4313 | ins_encode %{ |
kvn@4204 | 4314 | __ psrlw($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4204 | 4315 | %} |
kvn@4204 | 4316 | ins_pipe( pipe_slow ); |
kvn@4204 | 4317 | %} |
kvn@4204 | 4318 | |
kvn@4204 | 4319 | instruct vsrl8S_imm(vecX dst, immI8 shift) %{ |
kvn@4204 | 4320 | predicate(n->as_Vector()->length() == 8); |
kvn@4204 | 4321 | match(Set dst (URShiftVS dst shift)); |
kvn@4204 | 4322 | format %{ "psrlw $dst,$shift\t! logical right shift packed8S" %} |
kvn@4204 | 4323 | ins_encode %{ |
kvn@4204 | 4324 | __ psrlw($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4204 | 4325 | %} |
kvn@4204 | 4326 | ins_pipe( pipe_slow ); |
kvn@4204 | 4327 | %} |
kvn@4204 | 4328 | |
kvn@4204 | 4329 | instruct vsrl8S_reg(vecX dst, vecX src, vecS shift) %{ |
kvn@4204 | 4330 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4204 | 4331 | match(Set dst (URShiftVS src shift)); |
kvn@4204 | 4332 | format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed8S" %} |
kvn@4204 | 4333 | ins_encode %{ |
kvn@4204 | 4334 | bool vector256 = false; |
kvn@4204 | 4335 | __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4204 | 4336 | %} |
kvn@4204 | 4337 | ins_pipe( pipe_slow ); |
kvn@4204 | 4338 | %} |
kvn@4204 | 4339 | |
kvn@4204 | 4340 | instruct vsrl8S_reg_imm(vecX dst, vecX src, immI8 shift) %{ |
kvn@4204 | 4341 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4204 | 4342 | match(Set dst (URShiftVS src shift)); |
kvn@4204 | 4343 | format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed8S" %} |
kvn@4204 | 4344 | ins_encode %{ |
kvn@4204 | 4345 | bool vector256 = false; |
kvn@4204 | 4346 | __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4204 | 4347 | %} |
kvn@4204 | 4348 | ins_pipe( pipe_slow ); |
kvn@4204 | 4349 | %} |
kvn@4204 | 4350 | |
kvn@4204 | 4351 | instruct vsrl16S_reg(vecY dst, vecY src, vecS shift) %{ |
kvn@4204 | 4352 | predicate(UseAVX > 1 && n->as_Vector()->length() == 16); |
kvn@4204 | 4353 | match(Set dst (URShiftVS src shift)); |
kvn@4204 | 4354 | format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed16S" %} |
kvn@4204 | 4355 | ins_encode %{ |
kvn@4204 | 4356 | bool vector256 = true; |
kvn@4204 | 4357 | __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4204 | 4358 | %} |
kvn@4204 | 4359 | ins_pipe( pipe_slow ); |
kvn@4204 | 4360 | %} |
kvn@4204 | 4361 | |
kvn@4204 | 4362 | instruct vsrl16S_reg_imm(vecY dst, vecY src, immI8 shift) %{ |
kvn@4204 | 4363 | predicate(UseAVX > 1 && n->as_Vector()->length() == 16); |
kvn@4204 | 4364 | match(Set dst (URShiftVS src shift)); |
kvn@4204 | 4365 | format %{ "vpsrlw $dst,$src,$shift\t! logical right shift packed16S" %} |
kvn@4204 | 4366 | ins_encode %{ |
kvn@4204 | 4367 | bool vector256 = true; |
kvn@4204 | 4368 | __ vpsrlw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4204 | 4369 | %} |
kvn@4204 | 4370 | ins_pipe( pipe_slow ); |
kvn@4204 | 4371 | %} |
kvn@4001 | 4372 | |
kvn@4001 | 4373 | // Integers vector logical right shift |
kvn@4134 | 4374 | instruct vsrl2I(vecD dst, vecS shift) %{ |
kvn@4001 | 4375 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 4376 | match(Set dst (URShiftVI dst shift)); |
kvn@4001 | 4377 | format %{ "psrld $dst,$shift\t! logical right shift packed2I" %} |
kvn@4001 | 4378 | ins_encode %{ |
kvn@4001 | 4379 | __ psrld($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 4380 | %} |
kvn@4001 | 4381 | ins_pipe( pipe_slow ); |
kvn@4001 | 4382 | %} |
kvn@4001 | 4383 | |
kvn@4001 | 4384 | instruct vsrl2I_imm(vecD dst, immI8 shift) %{ |
kvn@4001 | 4385 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 4386 | match(Set dst (URShiftVI dst shift)); |
kvn@4001 | 4387 | format %{ "psrld $dst,$shift\t! logical right shift packed2I" %} |
kvn@4001 | 4388 | ins_encode %{ |
kvn@4001 | 4389 | __ psrld($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 4390 | %} |
kvn@4001 | 4391 | ins_pipe( pipe_slow ); |
kvn@4001 | 4392 | %} |
kvn@4001 | 4393 | |
kvn@4134 | 4394 | instruct vsrl2I_reg(vecD dst, vecD src, vecS shift) %{ |
kvn@4001 | 4395 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 4396 | match(Set dst (URShiftVI src shift)); |
kvn@4001 | 4397 | format %{ "vpsrld $dst,$src,$shift\t! logical right shift packed2I" %} |
kvn@4001 | 4398 | ins_encode %{ |
kvn@4001 | 4399 | bool vector256 = false; |
kvn@4001 | 4400 | __ vpsrld($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4401 | %} |
kvn@4001 | 4402 | ins_pipe( pipe_slow ); |
kvn@4001 | 4403 | %} |
kvn@4001 | 4404 | |
kvn@4001 | 4405 | instruct vsrl2I_reg_imm(vecD dst, vecD src, immI8 shift) %{ |
kvn@4001 | 4406 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 4407 | match(Set dst (URShiftVI src shift)); |
kvn@4001 | 4408 | format %{ "vpsrld $dst,$src,$shift\t! logical right shift packed2I" %} |
kvn@4001 | 4409 | ins_encode %{ |
kvn@4001 | 4410 | bool vector256 = false; |
kvn@4001 | 4411 | __ vpsrld($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4412 | %} |
kvn@4001 | 4413 | ins_pipe( pipe_slow ); |
kvn@4001 | 4414 | %} |
kvn@4001 | 4415 | |
kvn@4134 | 4416 | instruct vsrl4I(vecX dst, vecS shift) %{ |
kvn@4001 | 4417 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 4418 | match(Set dst (URShiftVI dst shift)); |
kvn@4001 | 4419 | format %{ "psrld $dst,$shift\t! logical right shift packed4I" %} |
kvn@4001 | 4420 | ins_encode %{ |
kvn@4001 | 4421 | __ psrld($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 4422 | %} |
kvn@4001 | 4423 | ins_pipe( pipe_slow ); |
kvn@4001 | 4424 | %} |
kvn@4001 | 4425 | |
kvn@4001 | 4426 | instruct vsrl4I_imm(vecX dst, immI8 shift) %{ |
kvn@4001 | 4427 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 4428 | match(Set dst (URShiftVI dst shift)); |
kvn@4001 | 4429 | format %{ "psrld $dst,$shift\t! logical right shift packed4I" %} |
kvn@4001 | 4430 | ins_encode %{ |
kvn@4001 | 4431 | __ psrld($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 4432 | %} |
kvn@4001 | 4433 | ins_pipe( pipe_slow ); |
kvn@4001 | 4434 | %} |
kvn@4001 | 4435 | |
kvn@4134 | 4436 | instruct vsrl4I_reg(vecX dst, vecX src, vecS shift) %{ |
kvn@4001 | 4437 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 4438 | match(Set dst (URShiftVI src shift)); |
kvn@4001 | 4439 | format %{ "vpsrld $dst,$src,$shift\t! logical right shift packed4I" %} |
kvn@4001 | 4440 | ins_encode %{ |
kvn@4001 | 4441 | bool vector256 = false; |
kvn@4001 | 4442 | __ vpsrld($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4443 | %} |
kvn@4001 | 4444 | ins_pipe( pipe_slow ); |
kvn@4001 | 4445 | %} |
kvn@4001 | 4446 | |
kvn@4001 | 4447 | instruct vsrl4I_reg_imm(vecX dst, vecX src, immI8 shift) %{ |
kvn@4001 | 4448 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 4449 | match(Set dst (URShiftVI src shift)); |
kvn@4001 | 4450 | format %{ "vpsrld $dst,$src,$shift\t! logical right shift packed4I" %} |
kvn@4001 | 4451 | ins_encode %{ |
kvn@4001 | 4452 | bool vector256 = false; |
kvn@4001 | 4453 | __ vpsrld($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4454 | %} |
kvn@4001 | 4455 | ins_pipe( pipe_slow ); |
kvn@4001 | 4456 | %} |
kvn@4001 | 4457 | |
kvn@4134 | 4458 | instruct vsrl8I_reg(vecY dst, vecY src, vecS shift) %{ |
kvn@4001 | 4459 | predicate(UseAVX > 1 && n->as_Vector()->length() == 8); |
kvn@4001 | 4460 | match(Set dst (URShiftVI src shift)); |
kvn@4001 | 4461 | format %{ "vpsrld $dst,$src,$shift\t! logical right shift packed8I" %} |
kvn@4001 | 4462 | ins_encode %{ |
kvn@4001 | 4463 | bool vector256 = true; |
kvn@4001 | 4464 | __ vpsrld($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4465 | %} |
kvn@4001 | 4466 | ins_pipe( pipe_slow ); |
kvn@4001 | 4467 | %} |
kvn@4001 | 4468 | |
kvn@4001 | 4469 | instruct vsrl8I_reg_imm(vecY dst, vecY src, immI8 shift) %{ |
kvn@4001 | 4470 | predicate(UseAVX > 1 && n->as_Vector()->length() == 8); |
kvn@4001 | 4471 | match(Set dst (URShiftVI src shift)); |
kvn@4001 | 4472 | format %{ "vpsrld $dst,$src,$shift\t! logical right shift packed8I" %} |
kvn@4001 | 4473 | ins_encode %{ |
kvn@4001 | 4474 | bool vector256 = true; |
kvn@4001 | 4475 | __ vpsrld($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4476 | %} |
kvn@4001 | 4477 | ins_pipe( pipe_slow ); |
kvn@4001 | 4478 | %} |
kvn@4001 | 4479 | |
kvn@4001 | 4480 | // Longs vector logical right shift |
kvn@4134 | 4481 | instruct vsrl2L(vecX dst, vecS shift) %{ |
kvn@4001 | 4482 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 4483 | match(Set dst (URShiftVL dst shift)); |
kvn@4001 | 4484 | format %{ "psrlq $dst,$shift\t! logical right shift packed2L" %} |
kvn@4001 | 4485 | ins_encode %{ |
kvn@4001 | 4486 | __ psrlq($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 4487 | %} |
kvn@4001 | 4488 | ins_pipe( pipe_slow ); |
kvn@4001 | 4489 | %} |
kvn@4001 | 4490 | |
kvn@4001 | 4491 | instruct vsrl2L_imm(vecX dst, immI8 shift) %{ |
kvn@4001 | 4492 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 4493 | match(Set dst (URShiftVL dst shift)); |
kvn@4001 | 4494 | format %{ "psrlq $dst,$shift\t! logical right shift packed2L" %} |
kvn@4001 | 4495 | ins_encode %{ |
kvn@4001 | 4496 | __ psrlq($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 4497 | %} |
kvn@4001 | 4498 | ins_pipe( pipe_slow ); |
kvn@4001 | 4499 | %} |
kvn@4001 | 4500 | |
kvn@4134 | 4501 | instruct vsrl2L_reg(vecX dst, vecX src, vecS shift) %{ |
kvn@4001 | 4502 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 4503 | match(Set dst (URShiftVL src shift)); |
kvn@4001 | 4504 | format %{ "vpsrlq $dst,$src,$shift\t! logical right shift packed2L" %} |
kvn@4001 | 4505 | ins_encode %{ |
kvn@4001 | 4506 | bool vector256 = false; |
kvn@4001 | 4507 | __ vpsrlq($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4508 | %} |
kvn@4001 | 4509 | ins_pipe( pipe_slow ); |
kvn@4001 | 4510 | %} |
kvn@4001 | 4511 | |
kvn@4001 | 4512 | instruct vsrl2L_reg_imm(vecX dst, vecX src, immI8 shift) %{ |
kvn@4001 | 4513 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 4514 | match(Set dst (URShiftVL src shift)); |
kvn@4001 | 4515 | format %{ "vpsrlq $dst,$src,$shift\t! logical right shift packed2L" %} |
kvn@4001 | 4516 | ins_encode %{ |
kvn@4001 | 4517 | bool vector256 = false; |
kvn@4001 | 4518 | __ vpsrlq($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4519 | %} |
kvn@4001 | 4520 | ins_pipe( pipe_slow ); |
kvn@4001 | 4521 | %} |
kvn@4001 | 4522 | |
kvn@4134 | 4523 | instruct vsrl4L_reg(vecY dst, vecY src, vecS shift) %{ |
kvn@4001 | 4524 | predicate(UseAVX > 1 && n->as_Vector()->length() == 4); |
kvn@4001 | 4525 | match(Set dst (URShiftVL src shift)); |
kvn@4001 | 4526 | format %{ "vpsrlq $dst,$src,$shift\t! logical right shift packed4L" %} |
kvn@4001 | 4527 | ins_encode %{ |
kvn@4001 | 4528 | bool vector256 = true; |
kvn@4001 | 4529 | __ vpsrlq($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4530 | %} |
kvn@4001 | 4531 | ins_pipe( pipe_slow ); |
kvn@4001 | 4532 | %} |
kvn@4001 | 4533 | |
kvn@4001 | 4534 | instruct vsrl4L_reg_imm(vecY dst, vecY src, immI8 shift) %{ |
kvn@4001 | 4535 | predicate(UseAVX > 1 && n->as_Vector()->length() == 4); |
kvn@4001 | 4536 | match(Set dst (URShiftVL src shift)); |
kvn@4001 | 4537 | format %{ "vpsrlq $dst,$src,$shift\t! logical right shift packed4L" %} |
kvn@4001 | 4538 | ins_encode %{ |
kvn@4001 | 4539 | bool vector256 = true; |
kvn@4001 | 4540 | __ vpsrlq($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4541 | %} |
kvn@4001 | 4542 | ins_pipe( pipe_slow ); |
kvn@4001 | 4543 | %} |
kvn@4001 | 4544 | |
kvn@4001 | 4545 | // ------------------- ArithmeticRightShift ----------------------------------- |
kvn@4001 | 4546 | |
kvn@4001 | 4547 | // Shorts/Chars vector arithmetic right shift |
kvn@4134 | 4548 | instruct vsra2S(vecS dst, vecS shift) %{ |
kvn@4001 | 4549 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 4550 | match(Set dst (RShiftVS dst shift)); |
kvn@4001 | 4551 | format %{ "psraw $dst,$shift\t! arithmetic right shift packed2S" %} |
kvn@4001 | 4552 | ins_encode %{ |
kvn@4001 | 4553 | __ psraw($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 4554 | %} |
kvn@4001 | 4555 | ins_pipe( pipe_slow ); |
kvn@4001 | 4556 | %} |
kvn@4001 | 4557 | |
kvn@4001 | 4558 | instruct vsra2S_imm(vecS dst, immI8 shift) %{ |
kvn@4001 | 4559 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 4560 | match(Set dst (RShiftVS dst shift)); |
kvn@4001 | 4561 | format %{ "psraw $dst,$shift\t! arithmetic right shift packed2S" %} |
kvn@4001 | 4562 | ins_encode %{ |
kvn@4001 | 4563 | __ psraw($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 4564 | %} |
kvn@4001 | 4565 | ins_pipe( pipe_slow ); |
kvn@4001 | 4566 | %} |
kvn@4001 | 4567 | |
kvn@4134 | 4568 | instruct vsra2S_reg(vecS dst, vecS src, vecS shift) %{ |
kvn@4001 | 4569 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 4570 | match(Set dst (RShiftVS src shift)); |
kvn@4001 | 4571 | format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed2S" %} |
kvn@4001 | 4572 | ins_encode %{ |
kvn@4001 | 4573 | bool vector256 = false; |
kvn@4001 | 4574 | __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4575 | %} |
kvn@4001 | 4576 | ins_pipe( pipe_slow ); |
kvn@4001 | 4577 | %} |
kvn@4001 | 4578 | |
kvn@4001 | 4579 | instruct vsra2S_reg_imm(vecS dst, vecS src, immI8 shift) %{ |
kvn@4001 | 4580 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 4581 | match(Set dst (RShiftVS src shift)); |
kvn@4001 | 4582 | format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed2S" %} |
kvn@4001 | 4583 | ins_encode %{ |
kvn@4001 | 4584 | bool vector256 = false; |
kvn@4001 | 4585 | __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4586 | %} |
kvn@4001 | 4587 | ins_pipe( pipe_slow ); |
kvn@4001 | 4588 | %} |
kvn@4001 | 4589 | |
kvn@4134 | 4590 | instruct vsra4S(vecD dst, vecS shift) %{ |
kvn@4001 | 4591 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 4592 | match(Set dst (RShiftVS dst shift)); |
kvn@4001 | 4593 | format %{ "psraw $dst,$shift\t! arithmetic right shift packed4S" %} |
kvn@4001 | 4594 | ins_encode %{ |
kvn@4001 | 4595 | __ psraw($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 4596 | %} |
kvn@4001 | 4597 | ins_pipe( pipe_slow ); |
kvn@4001 | 4598 | %} |
kvn@4001 | 4599 | |
kvn@4001 | 4600 | instruct vsra4S_imm(vecD dst, immI8 shift) %{ |
kvn@4001 | 4601 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 4602 | match(Set dst (RShiftVS dst shift)); |
kvn@4001 | 4603 | format %{ "psraw $dst,$shift\t! arithmetic right shift packed4S" %} |
kvn@4001 | 4604 | ins_encode %{ |
kvn@4001 | 4605 | __ psraw($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 4606 | %} |
kvn@4001 | 4607 | ins_pipe( pipe_slow ); |
kvn@4001 | 4608 | %} |
kvn@4001 | 4609 | |
kvn@4134 | 4610 | instruct vsra4S_reg(vecD dst, vecD src, vecS shift) %{ |
kvn@4001 | 4611 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 4612 | match(Set dst (RShiftVS src shift)); |
kvn@4001 | 4613 | format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed4S" %} |
kvn@4001 | 4614 | ins_encode %{ |
kvn@4001 | 4615 | bool vector256 = false; |
kvn@4001 | 4616 | __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4617 | %} |
kvn@4001 | 4618 | ins_pipe( pipe_slow ); |
kvn@4001 | 4619 | %} |
kvn@4001 | 4620 | |
kvn@4001 | 4621 | instruct vsra4S_reg_imm(vecD dst, vecD src, immI8 shift) %{ |
kvn@4001 | 4622 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 4623 | match(Set dst (RShiftVS src shift)); |
kvn@4001 | 4624 | format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed4S" %} |
kvn@4001 | 4625 | ins_encode %{ |
kvn@4001 | 4626 | bool vector256 = false; |
kvn@4001 | 4627 | __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4628 | %} |
kvn@4001 | 4629 | ins_pipe( pipe_slow ); |
kvn@4001 | 4630 | %} |
kvn@4001 | 4631 | |
kvn@4134 | 4632 | instruct vsra8S(vecX dst, vecS shift) %{ |
kvn@4001 | 4633 | predicate(n->as_Vector()->length() == 8); |
kvn@4001 | 4634 | match(Set dst (RShiftVS dst shift)); |
kvn@4001 | 4635 | format %{ "psraw $dst,$shift\t! arithmetic right shift packed8S" %} |
kvn@4001 | 4636 | ins_encode %{ |
kvn@4001 | 4637 | __ psraw($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 4638 | %} |
kvn@4001 | 4639 | ins_pipe( pipe_slow ); |
kvn@4001 | 4640 | %} |
kvn@4001 | 4641 | |
kvn@4001 | 4642 | instruct vsra8S_imm(vecX dst, immI8 shift) %{ |
kvn@4001 | 4643 | predicate(n->as_Vector()->length() == 8); |
kvn@4001 | 4644 | match(Set dst (RShiftVS dst shift)); |
kvn@4001 | 4645 | format %{ "psraw $dst,$shift\t! arithmetic right shift packed8S" %} |
kvn@4001 | 4646 | ins_encode %{ |
kvn@4001 | 4647 | __ psraw($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 4648 | %} |
kvn@4001 | 4649 | ins_pipe( pipe_slow ); |
kvn@4001 | 4650 | %} |
kvn@4001 | 4651 | |
kvn@4134 | 4652 | instruct vsra8S_reg(vecX dst, vecX src, vecS shift) %{ |
kvn@4001 | 4653 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 4654 | match(Set dst (RShiftVS src shift)); |
kvn@4001 | 4655 | format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed8S" %} |
kvn@4001 | 4656 | ins_encode %{ |
kvn@4001 | 4657 | bool vector256 = false; |
kvn@4001 | 4658 | __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4659 | %} |
kvn@4001 | 4660 | ins_pipe( pipe_slow ); |
kvn@4001 | 4661 | %} |
kvn@4001 | 4662 | |
kvn@4001 | 4663 | instruct vsra8S_reg_imm(vecX dst, vecX src, immI8 shift) %{ |
kvn@4001 | 4664 | predicate(UseAVX > 0 && n->as_Vector()->length() == 8); |
kvn@4001 | 4665 | match(Set dst (RShiftVS src shift)); |
kvn@4001 | 4666 | format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed8S" %} |
kvn@4001 | 4667 | ins_encode %{ |
kvn@4001 | 4668 | bool vector256 = false; |
kvn@4001 | 4669 | __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4670 | %} |
kvn@4001 | 4671 | ins_pipe( pipe_slow ); |
kvn@4001 | 4672 | %} |
kvn@4001 | 4673 | |
kvn@4134 | 4674 | instruct vsra16S_reg(vecY dst, vecY src, vecS shift) %{ |
kvn@4001 | 4675 | predicate(UseAVX > 1 && n->as_Vector()->length() == 16); |
kvn@4001 | 4676 | match(Set dst (RShiftVS src shift)); |
kvn@4001 | 4677 | format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed16S" %} |
kvn@4001 | 4678 | ins_encode %{ |
kvn@4001 | 4679 | bool vector256 = true; |
kvn@4001 | 4680 | __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4681 | %} |
kvn@4001 | 4682 | ins_pipe( pipe_slow ); |
kvn@4001 | 4683 | %} |
kvn@4001 | 4684 | |
kvn@4001 | 4685 | instruct vsra16S_reg_imm(vecY dst, vecY src, immI8 shift) %{ |
kvn@4001 | 4686 | predicate(UseAVX > 1 && n->as_Vector()->length() == 16); |
kvn@4001 | 4687 | match(Set dst (RShiftVS src shift)); |
kvn@4001 | 4688 | format %{ "vpsraw $dst,$src,$shift\t! arithmetic right shift packed16S" %} |
kvn@4001 | 4689 | ins_encode %{ |
kvn@4001 | 4690 | bool vector256 = true; |
kvn@4001 | 4691 | __ vpsraw($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4692 | %} |
kvn@4001 | 4693 | ins_pipe( pipe_slow ); |
kvn@4001 | 4694 | %} |
kvn@4001 | 4695 | |
kvn@4001 | 4696 | // Integers vector arithmetic right shift |
kvn@4134 | 4697 | instruct vsra2I(vecD dst, vecS shift) %{ |
kvn@4001 | 4698 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 4699 | match(Set dst (RShiftVI dst shift)); |
kvn@4001 | 4700 | format %{ "psrad $dst,$shift\t! arithmetic right shift packed2I" %} |
kvn@4001 | 4701 | ins_encode %{ |
kvn@4001 | 4702 | __ psrad($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 4703 | %} |
kvn@4001 | 4704 | ins_pipe( pipe_slow ); |
kvn@4001 | 4705 | %} |
kvn@4001 | 4706 | |
kvn@4001 | 4707 | instruct vsra2I_imm(vecD dst, immI8 shift) %{ |
kvn@4001 | 4708 | predicate(n->as_Vector()->length() == 2); |
kvn@4001 | 4709 | match(Set dst (RShiftVI dst shift)); |
kvn@4001 | 4710 | format %{ "psrad $dst,$shift\t! arithmetic right shift packed2I" %} |
kvn@4001 | 4711 | ins_encode %{ |
kvn@4001 | 4712 | __ psrad($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 4713 | %} |
kvn@4001 | 4714 | ins_pipe( pipe_slow ); |
kvn@4001 | 4715 | %} |
kvn@4001 | 4716 | |
kvn@4134 | 4717 | instruct vsra2I_reg(vecD dst, vecD src, vecS shift) %{ |
kvn@4001 | 4718 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 4719 | match(Set dst (RShiftVI src shift)); |
kvn@4001 | 4720 | format %{ "vpsrad $dst,$src,$shift\t! arithmetic right shift packed2I" %} |
kvn@4001 | 4721 | ins_encode %{ |
kvn@4001 | 4722 | bool vector256 = false; |
kvn@4001 | 4723 | __ vpsrad($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4724 | %} |
kvn@4001 | 4725 | ins_pipe( pipe_slow ); |
kvn@4001 | 4726 | %} |
kvn@4001 | 4727 | |
kvn@4001 | 4728 | instruct vsra2I_reg_imm(vecD dst, vecD src, immI8 shift) %{ |
kvn@4001 | 4729 | predicate(UseAVX > 0 && n->as_Vector()->length() == 2); |
kvn@4001 | 4730 | match(Set dst (RShiftVI src shift)); |
kvn@4001 | 4731 | format %{ "vpsrad $dst,$src,$shift\t! arithmetic right shift packed2I" %} |
kvn@4001 | 4732 | ins_encode %{ |
kvn@4001 | 4733 | bool vector256 = false; |
kvn@4001 | 4734 | __ vpsrad($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4735 | %} |
kvn@4001 | 4736 | ins_pipe( pipe_slow ); |
kvn@4001 | 4737 | %} |
kvn@4001 | 4738 | |
kvn@4134 | 4739 | instruct vsra4I(vecX dst, vecS shift) %{ |
kvn@4001 | 4740 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 4741 | match(Set dst (RShiftVI dst shift)); |
kvn@4001 | 4742 | format %{ "psrad $dst,$shift\t! arithmetic right shift packed4I" %} |
kvn@4001 | 4743 | ins_encode %{ |
kvn@4001 | 4744 | __ psrad($dst$$XMMRegister, $shift$$XMMRegister); |
kvn@4001 | 4745 | %} |
kvn@4001 | 4746 | ins_pipe( pipe_slow ); |
kvn@4001 | 4747 | %} |
kvn@4001 | 4748 | |
kvn@4001 | 4749 | instruct vsra4I_imm(vecX dst, immI8 shift) %{ |
kvn@4001 | 4750 | predicate(n->as_Vector()->length() == 4); |
kvn@4001 | 4751 | match(Set dst (RShiftVI dst shift)); |
kvn@4001 | 4752 | format %{ "psrad $dst,$shift\t! arithmetic right shift packed4I" %} |
kvn@4001 | 4753 | ins_encode %{ |
kvn@4001 | 4754 | __ psrad($dst$$XMMRegister, (int)$shift$$constant); |
kvn@4001 | 4755 | %} |
kvn@4001 | 4756 | ins_pipe( pipe_slow ); |
kvn@4001 | 4757 | %} |
kvn@4001 | 4758 | |
kvn@4134 | 4759 | instruct vsra4I_reg(vecX dst, vecX src, vecS shift) %{ |
kvn@4001 | 4760 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 4761 | match(Set dst (RShiftVI src shift)); |
kvn@4001 | 4762 | format %{ "vpsrad $dst,$src,$shift\t! arithmetic right shift packed4I" %} |
kvn@4001 | 4763 | ins_encode %{ |
kvn@4001 | 4764 | bool vector256 = false; |
kvn@4001 | 4765 | __ vpsrad($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4766 | %} |
kvn@4001 | 4767 | ins_pipe( pipe_slow ); |
kvn@4001 | 4768 | %} |
kvn@4001 | 4769 | |
kvn@4001 | 4770 | instruct vsra4I_reg_imm(vecX dst, vecX src, immI8 shift) %{ |
kvn@4001 | 4771 | predicate(UseAVX > 0 && n->as_Vector()->length() == 4); |
kvn@4001 | 4772 | match(Set dst (RShiftVI src shift)); |
kvn@4001 | 4773 | format %{ "vpsrad $dst,$src,$shift\t! arithmetic right shift packed4I" %} |
kvn@4001 | 4774 | ins_encode %{ |
kvn@4001 | 4775 | bool vector256 = false; |
kvn@4001 | 4776 | __ vpsrad($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4777 | %} |
kvn@4001 | 4778 | ins_pipe( pipe_slow ); |
kvn@4001 | 4779 | %} |
kvn@4001 | 4780 | |
kvn@4134 | 4781 | instruct vsra8I_reg(vecY dst, vecY src, vecS shift) %{ |
kvn@4001 | 4782 | predicate(UseAVX > 1 && n->as_Vector()->length() == 8); |
kvn@4001 | 4783 | match(Set dst (RShiftVI src shift)); |
kvn@4001 | 4784 | format %{ "vpsrad $dst,$src,$shift\t! arithmetic right shift packed8I" %} |
kvn@4001 | 4785 | ins_encode %{ |
kvn@4001 | 4786 | bool vector256 = true; |
kvn@4001 | 4787 | __ vpsrad($dst$$XMMRegister, $src$$XMMRegister, $shift$$XMMRegister, vector256); |
kvn@4001 | 4788 | %} |
kvn@4001 | 4789 | ins_pipe( pipe_slow ); |
kvn@4001 | 4790 | %} |
kvn@4001 | 4791 | |
kvn@4001 | 4792 | instruct vsra8I_reg_imm(vecY dst, vecY src, immI8 shift) %{ |
kvn@4001 | 4793 | predicate(UseAVX > 1 && n->as_Vector()->length() == 8); |
kvn@4001 | 4794 | match(Set dst (RShiftVI src shift)); |
kvn@4001 | 4795 | format %{ "vpsrad $dst,$src,$shift\t! arithmetic right shift packed8I" %} |
kvn@4001 | 4796 | ins_encode %{ |
kvn@4001 | 4797 | bool vector256 = true; |
kvn@4001 | 4798 | __ vpsrad($dst$$XMMRegister, $src$$XMMRegister, (int)$shift$$constant, vector256); |
kvn@4001 | 4799 | %} |
kvn@4001 | 4800 | ins_pipe( pipe_slow ); |
kvn@4001 | 4801 | %} |
kvn@4001 | 4802 | |
kvn@4001 | 4803 | // There are no longs vector arithmetic right shift instructions. |
kvn@4001 | 4804 | |
kvn@4001 | 4805 | |
kvn@4001 | 4806 | // --------------------------------- AND -------------------------------------- |
kvn@4001 | 4807 | |
kvn@4001 | 4808 | instruct vand4B(vecS dst, vecS src) %{ |
kvn@4001 | 4809 | predicate(n->as_Vector()->length_in_bytes() == 4); |
kvn@4001 | 4810 | match(Set dst (AndV dst src)); |
kvn@4001 | 4811 | format %{ "pand $dst,$src\t! and vectors (4 bytes)" %} |
kvn@4001 | 4812 | ins_encode %{ |
kvn@4001 | 4813 | __ pand($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 4814 | %} |
kvn@4001 | 4815 | ins_pipe( pipe_slow ); |
kvn@4001 | 4816 | %} |
kvn@4001 | 4817 | |
kvn@4001 | 4818 | instruct vand4B_reg(vecS dst, vecS src1, vecS src2) %{ |
kvn@4001 | 4819 | predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 4); |
kvn@4001 | 4820 | match(Set dst (AndV src1 src2)); |
kvn@4001 | 4821 | format %{ "vpand $dst,$src1,$src2\t! and vectors (4 bytes)" %} |
kvn@4001 | 4822 | ins_encode %{ |
kvn@4001 | 4823 | bool vector256 = false; |
kvn@4001 | 4824 | __ vpand($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 4825 | %} |
kvn@4001 | 4826 | ins_pipe( pipe_slow ); |
kvn@4001 | 4827 | %} |
kvn@4001 | 4828 | |
kvn@4001 | 4829 | instruct vand8B(vecD dst, vecD src) %{ |
kvn@4001 | 4830 | predicate(n->as_Vector()->length_in_bytes() == 8); |
kvn@4001 | 4831 | match(Set dst (AndV dst src)); |
kvn@4001 | 4832 | format %{ "pand $dst,$src\t! and vectors (8 bytes)" %} |
kvn@4001 | 4833 | ins_encode %{ |
kvn@4001 | 4834 | __ pand($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 4835 | %} |
kvn@4001 | 4836 | ins_pipe( pipe_slow ); |
kvn@4001 | 4837 | %} |
kvn@4001 | 4838 | |
kvn@4001 | 4839 | instruct vand8B_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 4840 | predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 8); |
kvn@4001 | 4841 | match(Set dst (AndV src1 src2)); |
kvn@4001 | 4842 | format %{ "vpand $dst,$src1,$src2\t! and vectors (8 bytes)" %} |
kvn@4001 | 4843 | ins_encode %{ |
kvn@4001 | 4844 | bool vector256 = false; |
kvn@4001 | 4845 | __ vpand($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 4846 | %} |
kvn@4001 | 4847 | ins_pipe( pipe_slow ); |
kvn@4001 | 4848 | %} |
kvn@4001 | 4849 | |
kvn@4001 | 4850 | instruct vand16B(vecX dst, vecX src) %{ |
kvn@4001 | 4851 | predicate(n->as_Vector()->length_in_bytes() == 16); |
kvn@4001 | 4852 | match(Set dst (AndV dst src)); |
kvn@4001 | 4853 | format %{ "pand $dst,$src\t! and vectors (16 bytes)" %} |
kvn@4001 | 4854 | ins_encode %{ |
kvn@4001 | 4855 | __ pand($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 4856 | %} |
kvn@4001 | 4857 | ins_pipe( pipe_slow ); |
kvn@4001 | 4858 | %} |
kvn@4001 | 4859 | |
kvn@4001 | 4860 | instruct vand16B_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 4861 | predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 16); |
kvn@4001 | 4862 | match(Set dst (AndV src1 src2)); |
kvn@4001 | 4863 | format %{ "vpand $dst,$src1,$src2\t! and vectors (16 bytes)" %} |
kvn@4001 | 4864 | ins_encode %{ |
kvn@4001 | 4865 | bool vector256 = false; |
kvn@4001 | 4866 | __ vpand($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 4867 | %} |
kvn@4001 | 4868 | ins_pipe( pipe_slow ); |
kvn@4001 | 4869 | %} |
kvn@4001 | 4870 | |
kvn@4001 | 4871 | instruct vand16B_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 4872 | predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 16); |
kvn@4001 | 4873 | match(Set dst (AndV src (LoadVector mem))); |
kvn@4001 | 4874 | format %{ "vpand $dst,$src,$mem\t! and vectors (16 bytes)" %} |
kvn@4001 | 4875 | ins_encode %{ |
kvn@4001 | 4876 | bool vector256 = false; |
kvn@4001 | 4877 | __ vpand($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 4878 | %} |
kvn@4001 | 4879 | ins_pipe( pipe_slow ); |
kvn@4001 | 4880 | %} |
kvn@4001 | 4881 | |
kvn@4001 | 4882 | instruct vand32B_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 4883 | predicate(UseAVX > 1 && n->as_Vector()->length_in_bytes() == 32); |
kvn@4001 | 4884 | match(Set dst (AndV src1 src2)); |
kvn@4001 | 4885 | format %{ "vpand $dst,$src1,$src2\t! and vectors (32 bytes)" %} |
kvn@4001 | 4886 | ins_encode %{ |
kvn@4001 | 4887 | bool vector256 = true; |
kvn@4001 | 4888 | __ vpand($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 4889 | %} |
kvn@4001 | 4890 | ins_pipe( pipe_slow ); |
kvn@4001 | 4891 | %} |
kvn@4001 | 4892 | |
kvn@4001 | 4893 | instruct vand32B_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 4894 | predicate(UseAVX > 1 && n->as_Vector()->length_in_bytes() == 32); |
kvn@4001 | 4895 | match(Set dst (AndV src (LoadVector mem))); |
kvn@4001 | 4896 | format %{ "vpand $dst,$src,$mem\t! and vectors (32 bytes)" %} |
kvn@4001 | 4897 | ins_encode %{ |
kvn@4001 | 4898 | bool vector256 = true; |
kvn@4001 | 4899 | __ vpand($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 4900 | %} |
kvn@4001 | 4901 | ins_pipe( pipe_slow ); |
kvn@4001 | 4902 | %} |
kvn@4001 | 4903 | |
kvn@4001 | 4904 | // --------------------------------- OR --------------------------------------- |
kvn@4001 | 4905 | |
kvn@4001 | 4906 | instruct vor4B(vecS dst, vecS src) %{ |
kvn@4001 | 4907 | predicate(n->as_Vector()->length_in_bytes() == 4); |
kvn@4001 | 4908 | match(Set dst (OrV dst src)); |
kvn@4001 | 4909 | format %{ "por $dst,$src\t! or vectors (4 bytes)" %} |
kvn@4001 | 4910 | ins_encode %{ |
kvn@4001 | 4911 | __ por($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 4912 | %} |
kvn@4001 | 4913 | ins_pipe( pipe_slow ); |
kvn@4001 | 4914 | %} |
kvn@4001 | 4915 | |
kvn@4001 | 4916 | instruct vor4B_reg(vecS dst, vecS src1, vecS src2) %{ |
kvn@4001 | 4917 | predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 4); |
kvn@4001 | 4918 | match(Set dst (OrV src1 src2)); |
kvn@4001 | 4919 | format %{ "vpor $dst,$src1,$src2\t! or vectors (4 bytes)" %} |
kvn@4001 | 4920 | ins_encode %{ |
kvn@4001 | 4921 | bool vector256 = false; |
kvn@4001 | 4922 | __ vpor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 4923 | %} |
kvn@4001 | 4924 | ins_pipe( pipe_slow ); |
kvn@4001 | 4925 | %} |
kvn@4001 | 4926 | |
kvn@4001 | 4927 | instruct vor8B(vecD dst, vecD src) %{ |
kvn@4001 | 4928 | predicate(n->as_Vector()->length_in_bytes() == 8); |
kvn@4001 | 4929 | match(Set dst (OrV dst src)); |
kvn@4001 | 4930 | format %{ "por $dst,$src\t! or vectors (8 bytes)" %} |
kvn@4001 | 4931 | ins_encode %{ |
kvn@4001 | 4932 | __ por($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 4933 | %} |
kvn@4001 | 4934 | ins_pipe( pipe_slow ); |
kvn@4001 | 4935 | %} |
kvn@4001 | 4936 | |
kvn@4001 | 4937 | instruct vor8B_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 4938 | predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 8); |
kvn@4001 | 4939 | match(Set dst (OrV src1 src2)); |
kvn@4001 | 4940 | format %{ "vpor $dst,$src1,$src2\t! or vectors (8 bytes)" %} |
kvn@4001 | 4941 | ins_encode %{ |
kvn@4001 | 4942 | bool vector256 = false; |
kvn@4001 | 4943 | __ vpor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 4944 | %} |
kvn@4001 | 4945 | ins_pipe( pipe_slow ); |
kvn@4001 | 4946 | %} |
kvn@4001 | 4947 | |
kvn@4001 | 4948 | instruct vor16B(vecX dst, vecX src) %{ |
kvn@4001 | 4949 | predicate(n->as_Vector()->length_in_bytes() == 16); |
kvn@4001 | 4950 | match(Set dst (OrV dst src)); |
kvn@4001 | 4951 | format %{ "por $dst,$src\t! or vectors (16 bytes)" %} |
kvn@4001 | 4952 | ins_encode %{ |
kvn@4001 | 4953 | __ por($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 4954 | %} |
kvn@4001 | 4955 | ins_pipe( pipe_slow ); |
kvn@4001 | 4956 | %} |
kvn@4001 | 4957 | |
kvn@4001 | 4958 | instruct vor16B_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 4959 | predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 16); |
kvn@4001 | 4960 | match(Set dst (OrV src1 src2)); |
kvn@4001 | 4961 | format %{ "vpor $dst,$src1,$src2\t! or vectors (16 bytes)" %} |
kvn@4001 | 4962 | ins_encode %{ |
kvn@4001 | 4963 | bool vector256 = false; |
kvn@4001 | 4964 | __ vpor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 4965 | %} |
kvn@4001 | 4966 | ins_pipe( pipe_slow ); |
kvn@4001 | 4967 | %} |
kvn@4001 | 4968 | |
kvn@4001 | 4969 | instruct vor16B_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 4970 | predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 16); |
kvn@4001 | 4971 | match(Set dst (OrV src (LoadVector mem))); |
kvn@4001 | 4972 | format %{ "vpor $dst,$src,$mem\t! or vectors (16 bytes)" %} |
kvn@4001 | 4973 | ins_encode %{ |
kvn@4001 | 4974 | bool vector256 = false; |
kvn@4001 | 4975 | __ vpor($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 4976 | %} |
kvn@4001 | 4977 | ins_pipe( pipe_slow ); |
kvn@4001 | 4978 | %} |
kvn@4001 | 4979 | |
kvn@4001 | 4980 | instruct vor32B_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 4981 | predicate(UseAVX > 1 && n->as_Vector()->length_in_bytes() == 32); |
kvn@4001 | 4982 | match(Set dst (OrV src1 src2)); |
kvn@4001 | 4983 | format %{ "vpor $dst,$src1,$src2\t! or vectors (32 bytes)" %} |
kvn@4001 | 4984 | ins_encode %{ |
kvn@4001 | 4985 | bool vector256 = true; |
kvn@4001 | 4986 | __ vpor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 4987 | %} |
kvn@4001 | 4988 | ins_pipe( pipe_slow ); |
kvn@4001 | 4989 | %} |
kvn@4001 | 4990 | |
kvn@4001 | 4991 | instruct vor32B_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 4992 | predicate(UseAVX > 1 && n->as_Vector()->length_in_bytes() == 32); |
kvn@4001 | 4993 | match(Set dst (OrV src (LoadVector mem))); |
kvn@4001 | 4994 | format %{ "vpor $dst,$src,$mem\t! or vectors (32 bytes)" %} |
kvn@4001 | 4995 | ins_encode %{ |
kvn@4001 | 4996 | bool vector256 = true; |
kvn@4001 | 4997 | __ vpor($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 4998 | %} |
kvn@4001 | 4999 | ins_pipe( pipe_slow ); |
kvn@4001 | 5000 | %} |
kvn@4001 | 5001 | |
kvn@4001 | 5002 | // --------------------------------- XOR -------------------------------------- |
kvn@4001 | 5003 | |
kvn@4001 | 5004 | instruct vxor4B(vecS dst, vecS src) %{ |
kvn@4001 | 5005 | predicate(n->as_Vector()->length_in_bytes() == 4); |
kvn@4001 | 5006 | match(Set dst (XorV dst src)); |
kvn@4001 | 5007 | format %{ "pxor $dst,$src\t! xor vectors (4 bytes)" %} |
kvn@4001 | 5008 | ins_encode %{ |
kvn@4001 | 5009 | __ pxor($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 5010 | %} |
kvn@4001 | 5011 | ins_pipe( pipe_slow ); |
kvn@4001 | 5012 | %} |
kvn@4001 | 5013 | |
kvn@4001 | 5014 | instruct vxor4B_reg(vecS dst, vecS src1, vecS src2) %{ |
kvn@4001 | 5015 | predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 4); |
kvn@4001 | 5016 | match(Set dst (XorV src1 src2)); |
kvn@4001 | 5017 | format %{ "vpxor $dst,$src1,$src2\t! xor vectors (4 bytes)" %} |
kvn@4001 | 5018 | ins_encode %{ |
kvn@4001 | 5019 | bool vector256 = false; |
kvn@4001 | 5020 | __ vpxor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 5021 | %} |
kvn@4001 | 5022 | ins_pipe( pipe_slow ); |
kvn@4001 | 5023 | %} |
kvn@4001 | 5024 | |
kvn@4001 | 5025 | instruct vxor8B(vecD dst, vecD src) %{ |
kvn@4001 | 5026 | predicate(n->as_Vector()->length_in_bytes() == 8); |
kvn@4001 | 5027 | match(Set dst (XorV dst src)); |
kvn@4001 | 5028 | format %{ "pxor $dst,$src\t! xor vectors (8 bytes)" %} |
kvn@4001 | 5029 | ins_encode %{ |
kvn@4001 | 5030 | __ pxor($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 5031 | %} |
kvn@4001 | 5032 | ins_pipe( pipe_slow ); |
kvn@4001 | 5033 | %} |
kvn@4001 | 5034 | |
kvn@4001 | 5035 | instruct vxor8B_reg(vecD dst, vecD src1, vecD src2) %{ |
kvn@4001 | 5036 | predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 8); |
kvn@4001 | 5037 | match(Set dst (XorV src1 src2)); |
kvn@4001 | 5038 | format %{ "vpxor $dst,$src1,$src2\t! xor vectors (8 bytes)" %} |
kvn@4001 | 5039 | ins_encode %{ |
kvn@4001 | 5040 | bool vector256 = false; |
kvn@4001 | 5041 | __ vpxor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 5042 | %} |
kvn@4001 | 5043 | ins_pipe( pipe_slow ); |
kvn@4001 | 5044 | %} |
kvn@4001 | 5045 | |
kvn@4001 | 5046 | instruct vxor16B(vecX dst, vecX src) %{ |
kvn@4001 | 5047 | predicate(n->as_Vector()->length_in_bytes() == 16); |
kvn@4001 | 5048 | match(Set dst (XorV dst src)); |
kvn@4001 | 5049 | format %{ "pxor $dst,$src\t! xor vectors (16 bytes)" %} |
kvn@4001 | 5050 | ins_encode %{ |
kvn@4001 | 5051 | __ pxor($dst$$XMMRegister, $src$$XMMRegister); |
kvn@4001 | 5052 | %} |
kvn@4001 | 5053 | ins_pipe( pipe_slow ); |
kvn@4001 | 5054 | %} |
kvn@4001 | 5055 | |
kvn@4001 | 5056 | instruct vxor16B_reg(vecX dst, vecX src1, vecX src2) %{ |
kvn@4001 | 5057 | predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 16); |
kvn@4001 | 5058 | match(Set dst (XorV src1 src2)); |
kvn@4001 | 5059 | format %{ "vpxor $dst,$src1,$src2\t! xor vectors (16 bytes)" %} |
kvn@4001 | 5060 | ins_encode %{ |
kvn@4001 | 5061 | bool vector256 = false; |
kvn@4001 | 5062 | __ vpxor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 5063 | %} |
kvn@4001 | 5064 | ins_pipe( pipe_slow ); |
kvn@4001 | 5065 | %} |
kvn@4001 | 5066 | |
kvn@4001 | 5067 | instruct vxor16B_mem(vecX dst, vecX src, memory mem) %{ |
kvn@4001 | 5068 | predicate(UseAVX > 0 && n->as_Vector()->length_in_bytes() == 16); |
kvn@4001 | 5069 | match(Set dst (XorV src (LoadVector mem))); |
kvn@4001 | 5070 | format %{ "vpxor $dst,$src,$mem\t! xor vectors (16 bytes)" %} |
kvn@4001 | 5071 | ins_encode %{ |
kvn@4001 | 5072 | bool vector256 = false; |
kvn@4001 | 5073 | __ vpxor($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 5074 | %} |
kvn@4001 | 5075 | ins_pipe( pipe_slow ); |
kvn@4001 | 5076 | %} |
kvn@4001 | 5077 | |
kvn@4001 | 5078 | instruct vxor32B_reg(vecY dst, vecY src1, vecY src2) %{ |
kvn@4001 | 5079 | predicate(UseAVX > 1 && n->as_Vector()->length_in_bytes() == 32); |
kvn@4001 | 5080 | match(Set dst (XorV src1 src2)); |
kvn@4001 | 5081 | format %{ "vpxor $dst,$src1,$src2\t! xor vectors (32 bytes)" %} |
kvn@4001 | 5082 | ins_encode %{ |
kvn@4001 | 5083 | bool vector256 = true; |
kvn@4001 | 5084 | __ vpxor($dst$$XMMRegister, $src1$$XMMRegister, $src2$$XMMRegister, vector256); |
kvn@4001 | 5085 | %} |
kvn@4001 | 5086 | ins_pipe( pipe_slow ); |
kvn@4001 | 5087 | %} |
kvn@4001 | 5088 | |
kvn@4001 | 5089 | instruct vxor32B_mem(vecY dst, vecY src, memory mem) %{ |
kvn@4001 | 5090 | predicate(UseAVX > 1 && n->as_Vector()->length_in_bytes() == 32); |
kvn@4001 | 5091 | match(Set dst (XorV src (LoadVector mem))); |
kvn@4001 | 5092 | format %{ "vpxor $dst,$src,$mem\t! xor vectors (32 bytes)" %} |
kvn@4001 | 5093 | ins_encode %{ |
kvn@4001 | 5094 | bool vector256 = true; |
kvn@4001 | 5095 | __ vpxor($dst$$XMMRegister, $src$$XMMRegister, $mem$$Address, vector256); |
kvn@4001 | 5096 | %} |
kvn@4001 | 5097 | ins_pipe( pipe_slow ); |
kvn@4001 | 5098 | %} |
kvn@4001 | 5099 |