diff options
| author | Craig Topper <craig.topper@intel.com> | 2018-05-12 23:14:39 +0000 |
|---|---|---|
| committer | Craig Topper <craig.topper@intel.com> | 2018-05-12 23:14:39 +0000 |
| commit | 38ad7ddabc95623d9c022adf34d57d5230c726d5 (patch) | |
| tree | 206d2c7565c3254e8caa2fe8729f9cd99d4677a4 /llvm/test/CodeGen | |
| parent | 095d69507e026151c599fd644e9149b6505cc043 (diff) | |
| download | bcm5719-llvm-38ad7ddabc95623d9c022adf34d57d5230c726d5.tar.gz bcm5719-llvm-38ad7ddabc95623d9c022adf34d57d5230c726d5.zip | |
[X86] Remove and autoupgrade cvtsi2ss/cvtsi2sd intrinsics to match what clang has used for a very long time.
llvm-svn: 332186
Diffstat (limited to 'llvm/test/CodeGen')
| -rw-r--r-- | llvm/test/CodeGen/X86/sse-intrinsics-x86-upgrade.ll | 10 | ||||
| -rw-r--r-- | llvm/test/CodeGen/X86/sse-intrinsics-x86.ll | 24 | ||||
| -rw-r--r-- | llvm/test/CodeGen/X86/sse-intrinsics-x86_64-upgrade.ll | 28 | ||||
| -rw-r--r-- | llvm/test/CodeGen/X86/sse-intrinsics-x86_64.ll | 25 | ||||
| -rw-r--r-- | llvm/test/CodeGen/X86/sse2-intrinsics-x86-upgrade.ll | 28 | ||||
| -rw-r--r-- | llvm/test/CodeGen/X86/sse2-intrinsics-x86.ll | 45 | ||||
| -rw-r--r-- | llvm/test/CodeGen/X86/sse2-intrinsics-x86_64-upgrade.ll | 28 | ||||
| -rw-r--r-- | llvm/test/CodeGen/X86/sse2-intrinsics-x86_64.ll | 25 | ||||
| -rw-r--r-- | llvm/test/CodeGen/X86/stack-folding-fp-sse42.ll | 16 | ||||
| -rw-r--r-- | llvm/test/CodeGen/X86/vec_ss_load_fold.ll | 16 |
10 files changed, 118 insertions, 127 deletions
diff --git a/llvm/test/CodeGen/X86/sse-intrinsics-x86-upgrade.ll b/llvm/test/CodeGen/X86/sse-intrinsics-x86-upgrade.ll index f7f9dff9beb..47c3c0b2261 100644 --- a/llvm/test/CodeGen/X86/sse-intrinsics-x86-upgrade.ll +++ b/llvm/test/CodeGen/X86/sse-intrinsics-x86-upgrade.ll @@ -56,3 +56,13 @@ define <4 x float> @test_x86_sse_div_ss(<4 x float> %a0, <4 x float> %a1) { } declare <4 x float> @llvm.x86.sse.div.ss(<4 x float>, <4 x float>) nounwind readnone + +define <4 x float> @test_x86_sse_cvtsi2ss(<4 x float> %a0, i32 %a1) { +; CHECK-LABEL: test_x86_sse_cvtsi2ss: +; CHECK: ## %bb.0: +; CHECK-NEXT: cvtsi2ssl {{[0-9]+}}(%esp), %xmm0 +; CHECK-NEXT: retl + %res = call <4 x float> @llvm.x86.sse.cvtsi2ss(<4 x float> %a0, i32 %a1) ; <<4 x float>> [#uses=1] + ret <4 x float> %res +} +declare <4 x float> @llvm.x86.sse.cvtsi2ss(<4 x float>, i32) nounwind readnone diff --git a/llvm/test/CodeGen/X86/sse-intrinsics-x86.ll b/llvm/test/CodeGen/X86/sse-intrinsics-x86.ll index 04a4352acca..3eb64698905 100644 --- a/llvm/test/CodeGen/X86/sse-intrinsics-x86.ll +++ b/llvm/test/CodeGen/X86/sse-intrinsics-x86.ll @@ -209,30 +209,6 @@ define i32 @test_x86_sse_comineq_ss(<4 x float> %a0, <4 x float> %a1) { declare i32 @llvm.x86.sse.comineq.ss(<4 x float>, <4 x float>) nounwind readnone -define <4 x float> @test_x86_sse_cvtsi2ss(<4 x float> %a0) { -; SSE-LABEL: test_x86_sse_cvtsi2ss: -; SSE: ## %bb.0: -; SSE-NEXT: movl $7, %eax ## encoding: [0xb8,0x07,0x00,0x00,0x00] -; SSE-NEXT: cvtsi2ssl %eax, %xmm0 ## encoding: [0xf3,0x0f,0x2a,0xc0] -; SSE-NEXT: retl ## encoding: [0xc3] -; -; AVX2-LABEL: test_x86_sse_cvtsi2ss: -; AVX2: ## %bb.0: -; AVX2-NEXT: movl $7, %eax ## encoding: [0xb8,0x07,0x00,0x00,0x00] -; AVX2-NEXT: vcvtsi2ssl %eax, %xmm0, %xmm0 ## encoding: [0xc5,0xfa,0x2a,0xc0] -; AVX2-NEXT: retl ## encoding: [0xc3] -; -; SKX-LABEL: test_x86_sse_cvtsi2ss: -; SKX: ## %bb.0: -; SKX-NEXT: movl $7, %eax ## encoding: [0xb8,0x07,0x00,0x00,0x00] -; SKX-NEXT: vcvtsi2ssl %eax, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfa,0x2a,0xc0] -; SKX-NEXT: retl ## encoding: [0xc3] - %res = call <4 x float> @llvm.x86.sse.cvtsi2ss(<4 x float> %a0, i32 7) ; <<4 x float>> [#uses=1] - ret <4 x float> %res -} -declare <4 x float> @llvm.x86.sse.cvtsi2ss(<4 x float>, i32) nounwind readnone - - define i32 @test_x86_sse_cvtss2si(<4 x float> %a0) { ; SSE-LABEL: test_x86_sse_cvtss2si: ; SSE: ## %bb.0: diff --git a/llvm/test/CodeGen/X86/sse-intrinsics-x86_64-upgrade.ll b/llvm/test/CodeGen/X86/sse-intrinsics-x86_64-upgrade.ll new file mode 100644 index 00000000000..25353d442a4 --- /dev/null +++ b/llvm/test/CodeGen/X86/sse-intrinsics-x86_64-upgrade.ll @@ -0,0 +1,28 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=x86_64-apple-darwin -mattr=-avx,+sse -show-mc-encoding | FileCheck %s --check-prefix=SSE +; RUN: llc < %s -mtriple=x86_64-apple-darwin -mattr=+avx2 -show-mc-encoding | FileCheck %s --check-prefix=VCHECK --check-prefix=AVX2 +; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=skx -show-mc-encoding | FileCheck %s --check-prefix=VCHECK --check-prefix=SKX + +define <4 x float> @test_x86_sse_cvtsi642ss(<4 x float> %a0, i64 %a1) { +; CHECK-LABEL: test_x86_sse_cvtsi642ss: +; CHECK: ## %bb.0: +; CHECK-NEXT: vcvtsi2ssq %rdi, %xmm0, %xmm0 +; CHECK-NEXT: retq +; SSE-LABEL: test_x86_sse_cvtsi642ss: +; SSE: ## %bb.0: +; SSE-NEXT: cvtsi2ssq %rdi, %xmm0 ## encoding: [0xf3,0x48,0x0f,0x2a,0xc7] +; SSE-NEXT: retq ## encoding: [0xc3] +; +; AVX2-LABEL: test_x86_sse_cvtsi642ss: +; AVX2: ## %bb.0: +; AVX2-NEXT: vcvtsi2ssq %rdi, %xmm0, %xmm0 ## encoding: [0xc4,0xe1,0xfa,0x2a,0xc7] +; AVX2-NEXT: retq ## encoding: [0xc3] +; +; SKX-LABEL: test_x86_sse_cvtsi642ss: +; SKX: ## %bb.0: +; SKX-NEXT: vcvtsi2ssq %rdi, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe1,0xfa,0x2a,0xc7] +; SKX-NEXT: retq ## encoding: [0xc3] + %res = call <4 x float> @llvm.x86.sse.cvtsi642ss(<4 x float> %a0, i64 %a1) ; <<4 x float>> [#uses=1] + ret <4 x float> %res +} +declare <4 x float> @llvm.x86.sse.cvtsi642ss(<4 x float>, i64) nounwind readnone diff --git a/llvm/test/CodeGen/X86/sse-intrinsics-x86_64.ll b/llvm/test/CodeGen/X86/sse-intrinsics-x86_64.ll index 6f95b8d9ea8..c2939907213 100644 --- a/llvm/test/CodeGen/X86/sse-intrinsics-x86_64.ll +++ b/llvm/test/CodeGen/X86/sse-intrinsics-x86_64.ll @@ -28,31 +28,6 @@ define i64 @test_x86_sse_cvtss2si64(<4 x float> %a0) { declare i64 @llvm.x86.sse.cvtss2si64(<4 x float>) nounwind readnone -define <4 x float> @test_x86_sse_cvtsi642ss(<4 x float> %a0, i64 %a1) { -; CHECK-LABEL: test_x86_sse_cvtsi642ss: -; CHECK: ## %bb.0: -; CHECK-NEXT: vcvtsi2ssq %rdi, %xmm0, %xmm0 -; CHECK-NEXT: retq -; SSE-LABEL: test_x86_sse_cvtsi642ss: -; SSE: ## %bb.0: -; SSE-NEXT: cvtsi2ssq %rdi, %xmm0 ## encoding: [0xf3,0x48,0x0f,0x2a,0xc7] -; SSE-NEXT: retq ## encoding: [0xc3] -; -; AVX2-LABEL: test_x86_sse_cvtsi642ss: -; AVX2: ## %bb.0: -; AVX2-NEXT: vcvtsi2ssq %rdi, %xmm0, %xmm0 ## encoding: [0xc4,0xe1,0xfa,0x2a,0xc7] -; AVX2-NEXT: retq ## encoding: [0xc3] -; -; SKX-LABEL: test_x86_sse_cvtsi642ss: -; SKX: ## %bb.0: -; SKX-NEXT: vcvtsi2ssq %rdi, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe1,0xfa,0x2a,0xc7] -; SKX-NEXT: retq ## encoding: [0xc3] - %res = call <4 x float> @llvm.x86.sse.cvtsi642ss(<4 x float> %a0, i64 %a1) ; <<4 x float>> [#uses=1] - ret <4 x float> %res -} -declare <4 x float> @llvm.x86.sse.cvtsi642ss(<4 x float>, i64) nounwind readnone - - define i64 @test_x86_sse_cvttss2si64(<4 x float> %a0) { ; CHECK-LABEL: test_x86_sse_cvttss2si64: ; CHECK: ## %bb.0: diff --git a/llvm/test/CodeGen/X86/sse2-intrinsics-x86-upgrade.ll b/llvm/test/CodeGen/X86/sse2-intrinsics-x86-upgrade.ll index ccc4b1208e0..2d51c49796f 100644 --- a/llvm/test/CodeGen/X86/sse2-intrinsics-x86-upgrade.ll +++ b/llvm/test/CodeGen/X86/sse2-intrinsics-x86-upgrade.ll @@ -256,3 +256,31 @@ define <2 x i64> @test_x86_sse2_pmulu_dq(<4 x i32> %a0, <4 x i32> %a1) { ret <2 x i64> %res } declare <2 x i64> @llvm.x86.sse2.pmulu.dq(<4 x i32>, <4 x i32>) nounwind readnone + + +define <2 x double> @test_x86_sse2_cvtsi2sd(<2 x double> %a0, i32 %a1) { +; SSE-LABEL: test_x86_sse2_cvtsi2sd: +; SSE: ## %bb.0: +; SSE-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04] +; SSE-NEXT: cvtsi2sdl %eax, %xmm0 ## encoding: [0xf2,0x0f,0x2a,0xc0] +; SSE-NEXT: retl ## encoding: [0xc3] +; +; AVX2-LABEL: test_x86_sse2_cvtsi2sd: +; AVX2: ## %bb.0: +; AVX2-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04] +; AVX2-NEXT: vcvtsi2sdl %eax, %xmm0, %xmm0 ## encoding: [0xc5,0xfb,0x2a,0xc0] +; AVX2-NEXT: retl ## encoding: [0xc3] +; +; SKX-LABEL: test_x86_sse2_cvtsi2sd: +; SKX: ## %bb.0: +; SKX-NEXT: movl {{[0-9]+}}(%esp), %eax ## encoding: [0x8b,0x44,0x24,0x04] +; SKX-NEXT: vcvtsi2sdl %eax, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfb,0x2a,0xc0] +; SKX-NEXT: retl ## encoding: [0xc3] +; CHECK-LABEL: test_x86_sse2_cvtsi2sd: +; CHECK: ## %bb.0: +; CHECK-NEXT: cvtsi2sdl {{[0-9]+}}(%esp), %xmm0 +; CHECK-NEXT: retl + %res = call <2 x double> @llvm.x86.sse2.cvtsi2sd(<2 x double> %a0, i32 %a1) ; <<2 x double>> [#uses=1] + ret <2 x double> %res +} +declare <2 x double> @llvm.x86.sse2.cvtsi2sd(<2 x double>, i32) nounwind readnone diff --git a/llvm/test/CodeGen/X86/sse2-intrinsics-x86.ll b/llvm/test/CodeGen/X86/sse2-intrinsics-x86.ll index ba787cb58b3..6c06997b58d 100644 --- a/llvm/test/CodeGen/X86/sse2-intrinsics-x86.ll +++ b/llvm/test/CodeGen/X86/sse2-intrinsics-x86.ll @@ -457,27 +457,6 @@ define <4 x float> @test_x86_sse2_cvtsd2ss_load_optsize(<4 x float> %a0, <2 x do } -define <2 x double> @test_x86_sse2_cvtsi2sd(<2 x double> %a0, i32 %a1) { -; SSE-LABEL: test_x86_sse2_cvtsi2sd: -; SSE: ## %bb.0: -; SSE-NEXT: cvtsi2sdl {{[0-9]+}}(%esp), %xmm0 ## encoding: [0xf2,0x0f,0x2a,0x44,0x24,0x04] -; SSE-NEXT: retl ## encoding: [0xc3] -; -; AVX2-LABEL: test_x86_sse2_cvtsi2sd: -; AVX2: ## %bb.0: -; AVX2-NEXT: vcvtsi2sdl {{[0-9]+}}(%esp), %xmm0, %xmm0 ## encoding: [0xc5,0xfb,0x2a,0x44,0x24,0x04] -; AVX2-NEXT: retl ## encoding: [0xc3] -; -; SKX-LABEL: test_x86_sse2_cvtsi2sd: -; SKX: ## %bb.0: -; SKX-NEXT: vcvtsi2sdl {{[0-9]+}}(%esp), %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc5,0xfb,0x2a,0x44,0x24,0x04] -; SKX-NEXT: retl ## encoding: [0xc3] - %res = call <2 x double> @llvm.x86.sse2.cvtsi2sd(<2 x double> %a0, i32 %a1) ; <<2 x double>> [#uses=1] - ret <2 x double> %res -} -declare <2 x double> @llvm.x86.sse2.cvtsi2sd(<2 x double>, i32) nounwind readnone - - define <2 x double> @test_x86_sse2_cvtss2sd(<2 x double> %a0, <4 x float> %a1) { ; SSE-LABEL: test_x86_sse2_cvtss2sd: ; SSE: ## %bb.0: @@ -767,21 +746,21 @@ define <8 x i16> @test_x86_sse2_packssdw_128_fold() { ; SSE: ## %bb.0: ; SSE-NEXT: movaps {{.*#+}} xmm0 = [0,0,0,0,32767,32767,65535,32768] ; SSE-NEXT: ## encoding: [0x0f,0x28,0x05,A,A,A,A] -; SSE-NEXT: ## fixup A - offset: 3, value: LCPI35_0, kind: FK_Data_4 +; SSE-NEXT: ## fixup A - offset: 3, value: LCPI34_0, kind: FK_Data_4 ; SSE-NEXT: retl ## encoding: [0xc3] ; ; AVX2-LABEL: test_x86_sse2_packssdw_128_fold: ; AVX2: ## %bb.0: ; AVX2-NEXT: vmovaps {{.*#+}} xmm0 = [0,0,0,0,32767,32767,65535,32768] ; AVX2-NEXT: ## encoding: [0xc5,0xf8,0x28,0x05,A,A,A,A] -; AVX2-NEXT: ## fixup A - offset: 4, value: LCPI35_0, kind: FK_Data_4 +; AVX2-NEXT: ## fixup A - offset: 4, value: LCPI34_0, kind: FK_Data_4 ; AVX2-NEXT: retl ## encoding: [0xc3] ; ; SKX-LABEL: test_x86_sse2_packssdw_128_fold: ; SKX: ## %bb.0: -; SKX-NEXT: vmovaps LCPI35_0, %xmm0 ## EVEX TO VEX Compression xmm0 = [0,0,0,0,32767,32767,65535,32768] +; SKX-NEXT: vmovaps LCPI34_0, %xmm0 ## EVEX TO VEX Compression xmm0 = [0,0,0,0,32767,32767,65535,32768] ; SKX-NEXT: ## encoding: [0xc5,0xf8,0x28,0x05,A,A,A,A] -; SKX-NEXT: ## fixup A - offset: 4, value: LCPI35_0, kind: FK_Data_4 +; SKX-NEXT: ## fixup A - offset: 4, value: LCPI34_0, kind: FK_Data_4 ; SKX-NEXT: retl ## encoding: [0xc3] %res = call <8 x i16> @llvm.x86.sse2.packssdw.128(<4 x i32> zeroinitializer, <4 x i32> <i32 65535, i32 65536, i32 -1, i32 -131072>) ret <8 x i16> %res @@ -814,21 +793,21 @@ define <16 x i8> @test_x86_sse2_packsswb_128_fold() { ; SSE: ## %bb.0: ; SSE-NEXT: movaps {{.*#+}} xmm0 = [0,127,127,255,255,128,128,128,0,0,0,0,0,0,0,0] ; SSE-NEXT: ## encoding: [0x0f,0x28,0x05,A,A,A,A] -; SSE-NEXT: ## fixup A - offset: 3, value: LCPI37_0, kind: FK_Data_4 +; SSE-NEXT: ## fixup A - offset: 3, value: LCPI36_0, kind: FK_Data_4 ; SSE-NEXT: retl ## encoding: [0xc3] ; ; AVX2-LABEL: test_x86_sse2_packsswb_128_fold: ; AVX2: ## %bb.0: ; AVX2-NEXT: vmovaps {{.*#+}} xmm0 = [0,127,127,255,255,128,128,128,0,0,0,0,0,0,0,0] ; AVX2-NEXT: ## encoding: [0xc5,0xf8,0x28,0x05,A,A,A,A] -; AVX2-NEXT: ## fixup A - offset: 4, value: LCPI37_0, kind: FK_Data_4 +; AVX2-NEXT: ## fixup A - offset: 4, value: LCPI36_0, kind: FK_Data_4 ; AVX2-NEXT: retl ## encoding: [0xc3] ; ; SKX-LABEL: test_x86_sse2_packsswb_128_fold: ; SKX: ## %bb.0: -; SKX-NEXT: vmovaps LCPI37_0, %xmm0 ## EVEX TO VEX Compression xmm0 = [0,127,127,255,255,128,128,128,0,0,0,0,0,0,0,0] +; SKX-NEXT: vmovaps LCPI36_0, %xmm0 ## EVEX TO VEX Compression xmm0 = [0,127,127,255,255,128,128,128,0,0,0,0,0,0,0,0] ; SKX-NEXT: ## encoding: [0xc5,0xf8,0x28,0x05,A,A,A,A] -; SKX-NEXT: ## fixup A - offset: 4, value: LCPI37_0, kind: FK_Data_4 +; SKX-NEXT: ## fixup A - offset: 4, value: LCPI36_0, kind: FK_Data_4 ; SKX-NEXT: retl ## encoding: [0xc3] %res = call <16 x i8> @llvm.x86.sse2.packsswb.128(<8 x i16> <i16 0, i16 255, i16 256, i16 65535, i16 -1, i16 -255, i16 -256, i16 -32678>, <8 x i16> zeroinitializer) ret <16 x i8> %res @@ -861,21 +840,21 @@ define <16 x i8> @test_x86_sse2_packuswb_128_fold() { ; SSE: ## %bb.0: ; SSE-NEXT: movaps {{.*#+}} xmm0 = [0,255,255,0,0,0,0,0,0,0,0,0,0,0,0,0] ; SSE-NEXT: ## encoding: [0x0f,0x28,0x05,A,A,A,A] -; SSE-NEXT: ## fixup A - offset: 3, value: LCPI39_0, kind: FK_Data_4 +; SSE-NEXT: ## fixup A - offset: 3, value: LCPI38_0, kind: FK_Data_4 ; SSE-NEXT: retl ## encoding: [0xc3] ; ; AVX2-LABEL: test_x86_sse2_packuswb_128_fold: ; AVX2: ## %bb.0: ; AVX2-NEXT: vmovaps {{.*#+}} xmm0 = [0,255,255,0,0,0,0,0,0,0,0,0,0,0,0,0] ; AVX2-NEXT: ## encoding: [0xc5,0xf8,0x28,0x05,A,A,A,A] -; AVX2-NEXT: ## fixup A - offset: 4, value: LCPI39_0, kind: FK_Data_4 +; AVX2-NEXT: ## fixup A - offset: 4, value: LCPI38_0, kind: FK_Data_4 ; AVX2-NEXT: retl ## encoding: [0xc3] ; ; SKX-LABEL: test_x86_sse2_packuswb_128_fold: ; SKX: ## %bb.0: -; SKX-NEXT: vmovaps LCPI39_0, %xmm0 ## EVEX TO VEX Compression xmm0 = [0,255,255,0,0,0,0,0,0,0,0,0,0,0,0,0] +; SKX-NEXT: vmovaps LCPI38_0, %xmm0 ## EVEX TO VEX Compression xmm0 = [0,255,255,0,0,0,0,0,0,0,0,0,0,0,0,0] ; SKX-NEXT: ## encoding: [0xc5,0xf8,0x28,0x05,A,A,A,A] -; SKX-NEXT: ## fixup A - offset: 4, value: LCPI39_0, kind: FK_Data_4 +; SKX-NEXT: ## fixup A - offset: 4, value: LCPI38_0, kind: FK_Data_4 ; SKX-NEXT: retl ## encoding: [0xc3] %res = call <16 x i8> @llvm.x86.sse2.packuswb.128(<8 x i16> <i16 0, i16 255, i16 256, i16 65535, i16 -1, i16 -255, i16 -256, i16 -32678>, <8 x i16> zeroinitializer) ret <16 x i8> %res diff --git a/llvm/test/CodeGen/X86/sse2-intrinsics-x86_64-upgrade.ll b/llvm/test/CodeGen/X86/sse2-intrinsics-x86_64-upgrade.ll new file mode 100644 index 00000000000..ccc72079b61 --- /dev/null +++ b/llvm/test/CodeGen/X86/sse2-intrinsics-x86_64-upgrade.ll @@ -0,0 +1,28 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -disable-peephole -mtriple=x86_64-apple-darwin -mattr=-avx,+sse2 -show-mc-encoding | FileCheck %s --check-prefix=SSE +; RUN: llc < %s -disable-peephole -mtriple=x86_64-apple-darwin -mattr=+avx2 -show-mc-encoding | FileCheck %s --check-prefix=VCHECK --check-prefix=AVX2 +; RUN: llc < %s -disable-peephole -mtriple=x86_64-apple-darwin -mcpu=skx -show-mc-encoding | FileCheck %s --check-prefix=VCHECK --check-prefix=SKX + +define <2 x double> @test_x86_sse2_cvtsi642sd(<2 x double> %a0, i64 %a1) { +; CHECK-LABEL: test_x86_sse2_cvtsi642sd: +; CHECK: ## %bb.0: +; CHECK-NEXT: vcvtsi2sdq %rdi, %xmm0, %xmm0 +; CHECK-NEXT: retq +; SSE-LABEL: test_x86_sse2_cvtsi642sd: +; SSE: ## %bb.0: +; SSE-NEXT: cvtsi2sdq %rdi, %xmm0 ## encoding: [0xf2,0x48,0x0f,0x2a,0xc7] +; SSE-NEXT: retq ## encoding: [0xc3] +; +; AVX2-LABEL: test_x86_sse2_cvtsi642sd: +; AVX2: ## %bb.0: +; AVX2-NEXT: vcvtsi2sdq %rdi, %xmm0, %xmm0 ## encoding: [0xc4,0xe1,0xfb,0x2a,0xc7] +; AVX2-NEXT: retq ## encoding: [0xc3] +; +; SKX-LABEL: test_x86_sse2_cvtsi642sd: +; SKX: ## %bb.0: +; SKX-NEXT: vcvtsi2sdq %rdi, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe1,0xfb,0x2a,0xc7] +; SKX-NEXT: retq ## encoding: [0xc3] + %res = call <2 x double> @llvm.x86.sse2.cvtsi642sd(<2 x double> %a0, i64 %a1) ; <<2 x double>> [#uses=1] + ret <2 x double> %res +} +declare <2 x double> @llvm.x86.sse2.cvtsi642sd(<2 x double>, i64) nounwind readnone diff --git a/llvm/test/CodeGen/X86/sse2-intrinsics-x86_64.ll b/llvm/test/CodeGen/X86/sse2-intrinsics-x86_64.ll index 41b4b2905dc..6c7b92d2830 100644 --- a/llvm/test/CodeGen/X86/sse2-intrinsics-x86_64.ll +++ b/llvm/test/CodeGen/X86/sse2-intrinsics-x86_64.ll @@ -28,31 +28,6 @@ define i64 @test_x86_sse2_cvtsd2si64(<2 x double> %a0) { declare i64 @llvm.x86.sse2.cvtsd2si64(<2 x double>) nounwind readnone -define <2 x double> @test_x86_sse2_cvtsi642sd(<2 x double> %a0, i64 %a1) { -; CHECK-LABEL: test_x86_sse2_cvtsi642sd: -; CHECK: ## %bb.0: -; CHECK-NEXT: vcvtsi2sdq %rdi, %xmm0, %xmm0 -; CHECK-NEXT: retq -; SSE-LABEL: test_x86_sse2_cvtsi642sd: -; SSE: ## %bb.0: -; SSE-NEXT: cvtsi2sdq %rdi, %xmm0 ## encoding: [0xf2,0x48,0x0f,0x2a,0xc7] -; SSE-NEXT: retq ## encoding: [0xc3] -; -; AVX2-LABEL: test_x86_sse2_cvtsi642sd: -; AVX2: ## %bb.0: -; AVX2-NEXT: vcvtsi2sdq %rdi, %xmm0, %xmm0 ## encoding: [0xc4,0xe1,0xfb,0x2a,0xc7] -; AVX2-NEXT: retq ## encoding: [0xc3] -; -; SKX-LABEL: test_x86_sse2_cvtsi642sd: -; SKX: ## %bb.0: -; SKX-NEXT: vcvtsi2sdq %rdi, %xmm0, %xmm0 ## EVEX TO VEX Compression encoding: [0xc4,0xe1,0xfb,0x2a,0xc7] -; SKX-NEXT: retq ## encoding: [0xc3] - %res = call <2 x double> @llvm.x86.sse2.cvtsi642sd(<2 x double> %a0, i64 %a1) ; <<2 x double>> [#uses=1] - ret <2 x double> %res -} -declare <2 x double> @llvm.x86.sse2.cvtsi642sd(<2 x double>, i64) nounwind readnone - - define i64 @test_x86_sse2_cvttsd2si64(<2 x double> %a0) { ; CHECK-LABEL: test_x86_sse2_cvttsd2si64: ; CHECK: ## %bb.0: diff --git a/llvm/test/CodeGen/X86/stack-folding-fp-sse42.ll b/llvm/test/CodeGen/X86/stack-folding-fp-sse42.ll index a28b3dfdcfa..e99ad3f6d2c 100644 --- a/llvm/test/CodeGen/X86/stack-folding-fp-sse42.ll +++ b/llvm/test/CodeGen/X86/stack-folding-fp-sse42.ll @@ -361,11 +361,11 @@ define double @stack_fold_cvtsi2sd(i32 %a0) minsize { ret double %2 } -define <2 x double> @stack_fold_cvtsi2sd_int(i32 %a0) { +define <2 x double> @stack_fold_cvtsi2sd_int(i32 %a0, <2 x double> %b0) { ;CHECK-LABEL: stack_fold_cvtsi2sd_int ;CHECK: cvtsi2sdl {{-?[0-9]*}}(%rsp), {{%xmm[0-9][0-9]*}} {{.*#+}} 4-byte Folded Reload %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{rax},~{rbx},~{rcx},~{rdx},~{rsi},~{rdi},~{rbp},~{r8},~{r9},~{r10},~{r11},~{r12},~{r13},~{r14},~{r15}"() - %2 = call <2 x double> @llvm.x86.sse2.cvtsi2sd(<2 x double> <double 0x0, double 0x0>, i32 %a0) + %2 = call <2 x double> @llvm.x86.sse2.cvtsi2sd(<2 x double> %b0, i32 %a0) ret <2 x double> %2 } declare <2 x double> @llvm.x86.sse2.cvtsi2sd(<2 x double>, i32) nounwind readnone @@ -378,11 +378,11 @@ define double @stack_fold_cvtsi642sd(i64 %a0) optsize { ret double %2 } -define <2 x double> @stack_fold_cvtsi642sd_int(i64 %a0) { +define <2 x double> @stack_fold_cvtsi642sd_int(i64 %a0, <2 x double> %b0) { ;CHECK-LABEL: stack_fold_cvtsi642sd_int ;CHECK: cvtsi2sdq {{-?[0-9]*}}(%rsp), {{%xmm[0-9][0-9]*}} {{.*#+}} 8-byte Folded Reload %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{rax},~{rbx},~{rcx},~{rdx},~{rsi},~{rdi},~{rbp},~{r8},~{r9},~{r10},~{r11},~{r12},~{r13},~{r14},~{r15}"() - %2 = call <2 x double> @llvm.x86.sse2.cvtsi642sd(<2 x double> <double 0x0, double 0x0>, i64 %a0) + %2 = call <2 x double> @llvm.x86.sse2.cvtsi642sd(<2 x double> %b0, i64 %a0) ret <2 x double> %2 } declare <2 x double> @llvm.x86.sse2.cvtsi642sd(<2 x double>, i64) nounwind readnone @@ -395,11 +395,11 @@ define float @stack_fold_cvtsi2ss(i32 %a0) minsize { ret float %2 } -define <4 x float> @stack_fold_cvtsi2ss_int(i32 %a0) { +define <4 x float> @stack_fold_cvtsi2ss_int(i32 %a0, <4 x float> %b0) { ;CHECK-LABEL: stack_fold_cvtsi2ss_int ;CHECK: cvtsi2ssl {{-?[0-9]*}}(%rsp), {{%xmm[0-9][0-9]*}} {{.*#+}} 4-byte Folded Reload %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{rax},~{rbx},~{rcx},~{rdx},~{rsi},~{rdi},~{rbp},~{r8},~{r9},~{r10},~{r11},~{r12},~{r13},~{r14},~{r15}"() - %2 = call <4 x float> @llvm.x86.sse.cvtsi2ss(<4 x float> <float 0x0, float 0x0, float 0x0, float 0x0>, i32 %a0) + %2 = call <4 x float> @llvm.x86.sse.cvtsi2ss(<4 x float> %b0, i32 %a0) ret <4 x float> %2 } declare <4 x float> @llvm.x86.sse.cvtsi2ss(<4 x float>, i32) nounwind readnone @@ -412,11 +412,11 @@ define float @stack_fold_cvtsi642ss(i64 %a0) optsize { ret float %2 } -define <4 x float> @stack_fold_cvtsi642ss_int(i64 %a0) { +define <4 x float> @stack_fold_cvtsi642ss_int(i64 %a0, <4 x float> %b0) { ;CHECK-LABEL: stack_fold_cvtsi642ss_int ;CHECK: cvtsi2ssq {{-?[0-9]*}}(%rsp), {{%xmm[0-9][0-9]*}} {{.*#+}} 8-byte Folded Reload %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{rax},~{rbx},~{rcx},~{rdx},~{rsi},~{rdi},~{rbp},~{r8},~{r9},~{r10},~{r11},~{r12},~{r13},~{r14},~{r15}"() - %2 = call <4 x float> @llvm.x86.sse.cvtsi642ss(<4 x float> <float 0x0, float 0x0, float 0x0, float 0x0>, i64 %a0) + %2 = call <4 x float> @llvm.x86.sse.cvtsi642ss(<4 x float> %b0, i64 %a0) ret <4 x float> %2 } declare <4 x float> @llvm.x86.sse.cvtsi642ss(<4 x float>, i64) nounwind readnone diff --git a/llvm/test/CodeGen/X86/vec_ss_load_fold.ll b/llvm/test/CodeGen/X86/vec_ss_load_fold.ll index d9ab5a1df7c..b6373cf997e 100644 --- a/llvm/test/CodeGen/X86/vec_ss_load_fold.ll +++ b/llvm/test/CodeGen/X86/vec_ss_load_fold.ll @@ -268,30 +268,22 @@ define <4 x float> @test4(<4 x float> %A, float *%b, i32 %C) nounwind { define <2 x double> @test5() nounwind uwtable readnone noinline { ; X32-LABEL: test5: ; X32: ## %bb.0: ## %entry -; X32-NEXT: movaps {{.*#+}} xmm0 = [4.569870e+02,1.233210e+02] -; X32-NEXT: movl $128, %eax -; X32-NEXT: cvtsi2sdl %eax, %xmm0 +; X32-NEXT: movaps {{.*#+}} xmm0 = [1.280000e+02,1.233210e+02] ; X32-NEXT: retl ; ; X64-LABEL: test5: ; X64: ## %bb.0: ## %entry -; X64-NEXT: movaps {{.*#+}} xmm0 = [4.569870e+02,1.233210e+02] -; X64-NEXT: movl $128, %eax -; X64-NEXT: cvtsi2sdl %eax, %xmm0 +; X64-NEXT: movaps {{.*#+}} xmm0 = [1.280000e+02,1.233210e+02] ; X64-NEXT: retq ; ; X32_AVX-LABEL: test5: ; X32_AVX: ## %bb.0: ## %entry -; X32_AVX-NEXT: vmovaps {{.*#+}} xmm0 = [4.569870e+02,1.233210e+02] -; X32_AVX-NEXT: movl $128, %eax -; X32_AVX-NEXT: vcvtsi2sdl %eax, %xmm0, %xmm0 +; X32_AVX-NEXT: vmovaps {{.*#+}} xmm0 = [1.280000e+02,1.233210e+02] ; X32_AVX-NEXT: retl ; ; X64_AVX-LABEL: test5: ; X64_AVX: ## %bb.0: ## %entry -; X64_AVX-NEXT: vmovaps {{.*#+}} xmm0 = [4.569870e+02,1.233210e+02] -; X64_AVX-NEXT: movl $128, %eax -; X64_AVX-NEXT: vcvtsi2sdl %eax, %xmm0, %xmm0 +; X64_AVX-NEXT: vmovaps {{.*#+}} xmm0 = [1.280000e+02,1.233210e+02] ; X64_AVX-NEXT: retq entry: %0 = tail call <2 x double> @llvm.x86.sse2.cvtsi2sd(<2 x double> <double 4.569870e+02, double 1.233210e+02>, i32 128) nounwind readnone |

