diff options
-rw-r--r-- | llvm/lib/Target/X86/X86InstrSSE.td | 4 | ||||
-rw-r--r-- | llvm/test/CodeGen/X86/avx-intrinsics-x86_64.ll | 42 |
2 files changed, 17 insertions, 29 deletions
diff --git a/llvm/lib/Target/X86/X86InstrSSE.td b/llvm/lib/Target/X86/X86InstrSSE.td index afc9a4f35ef..09e80e3f120 100644 --- a/llvm/lib/Target/X86/X86InstrSSE.td +++ b/llvm/lib/Target/X86/X86InstrSSE.td @@ -8057,7 +8057,9 @@ def : Pat<(v16i16 (X86VPerm2x128 VR256:$src1, // VZERO - Zero YMM registers // let Defs = [YMM0, YMM1, YMM2, YMM3, YMM4, YMM5, YMM6, YMM7, - YMM8, YMM9, YMM10, YMM11, YMM12, YMM13, YMM14, YMM15] in { + YMM8, YMM9, YMM10, YMM11, YMM12, YMM13, YMM14, YMM15, + YMM16, YMM17, YMM18, YMM19, YMM20, YMM21, YMM22, YMM23, + YMM24, YMM25, YMM26, YMM27, YMM28, YMM29, YMM30, YMM31] in { // Zero All YMM registers def VZEROALL : I<0x77, RawFrm, (outs), (ins), "vzeroall", [(int_x86_avx_vzeroall)]>, PS, VEX, VEX_L, Requires<[HasAVX]>; diff --git a/llvm/test/CodeGen/X86/avx-intrinsics-x86_64.ll b/llvm/test/CodeGen/X86/avx-intrinsics-x86_64.ll index 9f18c0930fd..24ab870cac8 100644 --- a/llvm/test/CodeGen/X86/avx-intrinsics-x86_64.ll +++ b/llvm/test/CodeGen/X86/avx-intrinsics-x86_64.ll @@ -68,20 +68,13 @@ define i64 @test_x86_sse_cvttss2si64(<4 x float> %a0) { declare i64 @llvm.x86.sse.cvttss2si64(<4 x float>) nounwind readnone define <4 x double> @test_x86_avx_vzeroall(<4 x double> %a, <4 x double> %b) { -; AVX-LABEL: test_x86_avx_vzeroall: -; AVX: ## BB#0: -; AVX-NEXT: vaddpd %ymm1, %ymm0, %ymm0 -; AVX-NEXT: vmovupd %ymm0, -{{[0-9]+}}(%rsp) ## 32-byte Spill -; AVX-NEXT: vzeroall -; AVX-NEXT: vmovups -{{[0-9]+}}(%rsp), %ymm0 ## 32-byte Reload -; AVX-NEXT: retq -; -; AVX512VL-LABEL: test_x86_avx_vzeroall: -; AVX512VL: ## BB#0: -; AVX512VL-NEXT: vaddpd %ymm1, %ymm0, %ymm16 -; AVX512VL-NEXT: vzeroall -; AVX512VL-NEXT: vmovapd %ymm16, %ymm0 -; AVX512VL-NEXT: retq +; CHECK-LABEL: test_x86_avx_vzeroall: +; CHECK: ## BB#0: +; CHECK-NEXT: vaddpd %ymm1, %ymm0, %ymm0 +; CHECK-NEXT: vmovupd %ymm0, -{{[0-9]+}}(%rsp) ## 32-byte Spill +; CHECK-NEXT: vzeroall +; CHECK-NEXT: vmovups -{{[0-9]+}}(%rsp), %ymm0 ## 32-byte Reload +; CHECK-NEXT: retq %c = fadd <4 x double> %a, %b call void @llvm.x86.avx.vzeroall() ret <4 x double> %c @@ -89,20 +82,13 @@ define <4 x double> @test_x86_avx_vzeroall(<4 x double> %a, <4 x double> %b) { declare void @llvm.x86.avx.vzeroall() nounwind define <4 x double> @test_x86_avx_vzeroupper(<4 x double> %a, <4 x double> %b) { -; AVX-LABEL: test_x86_avx_vzeroupper: -; AVX: ## BB#0: -; AVX-NEXT: vaddpd %ymm1, %ymm0, %ymm0 -; AVX-NEXT: vmovupd %ymm0, -{{[0-9]+}}(%rsp) ## 32-byte Spill -; AVX-NEXT: vzeroupper -; AVX-NEXT: vmovups -{{[0-9]+}}(%rsp), %ymm0 ## 32-byte Reload -; AVX-NEXT: retq -; -; AVX512VL-LABEL: test_x86_avx_vzeroupper: -; AVX512VL: ## BB#0: -; AVX512VL-NEXT: vaddpd %ymm1, %ymm0, %ymm16 -; AVX512VL-NEXT: vzeroupper -; AVX512VL-NEXT: vmovapd %ymm16, %ymm0 -; AVX512VL-NEXT: retq +; CHECK-LABEL: test_x86_avx_vzeroupper: +; CHECK: ## BB#0: +; CHECK-NEXT: vaddpd %ymm1, %ymm0, %ymm0 +; CHECK-NEXT: vmovupd %ymm0, -{{[0-9]+}}(%rsp) ## 32-byte Spill +; CHECK-NEXT: vzeroupper +; CHECK-NEXT: vmovups -{{[0-9]+}}(%rsp), %ymm0 ## 32-byte Reload +; CHECK-NEXT: retq %c = fadd <4 x double> %a, %b call void @llvm.x86.avx.vzeroupper() ret <4 x double> %c |