summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--llvm/lib/Target/X86/X86InstrSSE.td4
-rw-r--r--llvm/test/CodeGen/X86/avx-intrinsics-x86_64.ll42
2 files changed, 17 insertions, 29 deletions
diff --git a/llvm/lib/Target/X86/X86InstrSSE.td b/llvm/lib/Target/X86/X86InstrSSE.td
index afc9a4f35ef..09e80e3f120 100644
--- a/llvm/lib/Target/X86/X86InstrSSE.td
+++ b/llvm/lib/Target/X86/X86InstrSSE.td
@@ -8057,7 +8057,9 @@ def : Pat<(v16i16 (X86VPerm2x128 VR256:$src1,
// VZERO - Zero YMM registers
//
let Defs = [YMM0, YMM1, YMM2, YMM3, YMM4, YMM5, YMM6, YMM7,
- YMM8, YMM9, YMM10, YMM11, YMM12, YMM13, YMM14, YMM15] in {
+ YMM8, YMM9, YMM10, YMM11, YMM12, YMM13, YMM14, YMM15,
+ YMM16, YMM17, YMM18, YMM19, YMM20, YMM21, YMM22, YMM23,
+ YMM24, YMM25, YMM26, YMM27, YMM28, YMM29, YMM30, YMM31] in {
// Zero All YMM registers
def VZEROALL : I<0x77, RawFrm, (outs), (ins), "vzeroall",
[(int_x86_avx_vzeroall)]>, PS, VEX, VEX_L, Requires<[HasAVX]>;
diff --git a/llvm/test/CodeGen/X86/avx-intrinsics-x86_64.ll b/llvm/test/CodeGen/X86/avx-intrinsics-x86_64.ll
index 9f18c0930fd..24ab870cac8 100644
--- a/llvm/test/CodeGen/X86/avx-intrinsics-x86_64.ll
+++ b/llvm/test/CodeGen/X86/avx-intrinsics-x86_64.ll
@@ -68,20 +68,13 @@ define i64 @test_x86_sse_cvttss2si64(<4 x float> %a0) {
declare i64 @llvm.x86.sse.cvttss2si64(<4 x float>) nounwind readnone
define <4 x double> @test_x86_avx_vzeroall(<4 x double> %a, <4 x double> %b) {
-; AVX-LABEL: test_x86_avx_vzeroall:
-; AVX: ## BB#0:
-; AVX-NEXT: vaddpd %ymm1, %ymm0, %ymm0
-; AVX-NEXT: vmovupd %ymm0, -{{[0-9]+}}(%rsp) ## 32-byte Spill
-; AVX-NEXT: vzeroall
-; AVX-NEXT: vmovups -{{[0-9]+}}(%rsp), %ymm0 ## 32-byte Reload
-; AVX-NEXT: retq
-;
-; AVX512VL-LABEL: test_x86_avx_vzeroall:
-; AVX512VL: ## BB#0:
-; AVX512VL-NEXT: vaddpd %ymm1, %ymm0, %ymm16
-; AVX512VL-NEXT: vzeroall
-; AVX512VL-NEXT: vmovapd %ymm16, %ymm0
-; AVX512VL-NEXT: retq
+; CHECK-LABEL: test_x86_avx_vzeroall:
+; CHECK: ## BB#0:
+; CHECK-NEXT: vaddpd %ymm1, %ymm0, %ymm0
+; CHECK-NEXT: vmovupd %ymm0, -{{[0-9]+}}(%rsp) ## 32-byte Spill
+; CHECK-NEXT: vzeroall
+; CHECK-NEXT: vmovups -{{[0-9]+}}(%rsp), %ymm0 ## 32-byte Reload
+; CHECK-NEXT: retq
%c = fadd <4 x double> %a, %b
call void @llvm.x86.avx.vzeroall()
ret <4 x double> %c
@@ -89,20 +82,13 @@ define <4 x double> @test_x86_avx_vzeroall(<4 x double> %a, <4 x double> %b) {
declare void @llvm.x86.avx.vzeroall() nounwind
define <4 x double> @test_x86_avx_vzeroupper(<4 x double> %a, <4 x double> %b) {
-; AVX-LABEL: test_x86_avx_vzeroupper:
-; AVX: ## BB#0:
-; AVX-NEXT: vaddpd %ymm1, %ymm0, %ymm0
-; AVX-NEXT: vmovupd %ymm0, -{{[0-9]+}}(%rsp) ## 32-byte Spill
-; AVX-NEXT: vzeroupper
-; AVX-NEXT: vmovups -{{[0-9]+}}(%rsp), %ymm0 ## 32-byte Reload
-; AVX-NEXT: retq
-;
-; AVX512VL-LABEL: test_x86_avx_vzeroupper:
-; AVX512VL: ## BB#0:
-; AVX512VL-NEXT: vaddpd %ymm1, %ymm0, %ymm16
-; AVX512VL-NEXT: vzeroupper
-; AVX512VL-NEXT: vmovapd %ymm16, %ymm0
-; AVX512VL-NEXT: retq
+; CHECK-LABEL: test_x86_avx_vzeroupper:
+; CHECK: ## BB#0:
+; CHECK-NEXT: vaddpd %ymm1, %ymm0, %ymm0
+; CHECK-NEXT: vmovupd %ymm0, -{{[0-9]+}}(%rsp) ## 32-byte Spill
+; CHECK-NEXT: vzeroupper
+; CHECK-NEXT: vmovups -{{[0-9]+}}(%rsp), %ymm0 ## 32-byte Reload
+; CHECK-NEXT: retq
%c = fadd <4 x double> %a, %b
call void @llvm.x86.avx.vzeroupper()
ret <4 x double> %c
OpenPOWER on IntegriCloud