diff options
Diffstat (limited to 'llvm')
-rw-r--r-- | llvm/lib/Target/X86/X86InstrSSE.td | 8 | ||||
-rw-r--r-- | llvm/test/CodeGen/X86/WidenArith.ll | 2 | ||||
-rw-r--r-- | llvm/test/CodeGen/X86/merge-consecutive-loads-256.ll | 26 | ||||
-rw-r--r-- | llvm/test/CodeGen/X86/v8i1-masks.ll | 4 | ||||
-rw-r--r-- | llvm/test/CodeGen/X86/vec_int_to_fp.ll | 30 | ||||
-rw-r--r-- | llvm/test/CodeGen/X86/vec_uint_to_fp-fastmath.ll | 26 |
6 files changed, 42 insertions, 54 deletions
diff --git a/llvm/lib/Target/X86/X86InstrSSE.td b/llvm/lib/Target/X86/X86InstrSSE.td index 15d6d6d440c..7830b48280e 100644 --- a/llvm/lib/Target/X86/X86InstrSSE.td +++ b/llvm/lib/Target/X86/X86InstrSSE.td @@ -2952,13 +2952,13 @@ let isCommutable = 0 in // AVX1 requires type coercions in order to fold loads directly into logical // operations. let Predicates = [HasAVX1Only] in { - def : Pat<(bc_v8f32 (and VR256:$src1, (loadv4i64 addr:$src2))), + def : Pat<(and VR256:$src1, (loadv4i64 addr:$src2)), (VANDPSYrm VR256:$src1, addr:$src2)>; - def : Pat<(bc_v8f32 (or VR256:$src1, (loadv4i64 addr:$src2))), + def : Pat<(or VR256:$src1, (loadv4i64 addr:$src2)), (VORPSYrm VR256:$src1, addr:$src2)>; - def : Pat<(bc_v8f32 (xor VR256:$src1, (loadv4i64 addr:$src2))), + def : Pat<(xor VR256:$src1, (loadv4i64 addr:$src2)), (VXORPSYrm VR256:$src1, addr:$src2)>; - def : Pat<(bc_v8f32 (X86andnp VR256:$src1, (loadv4i64 addr:$src2))), + def : Pat<(X86andnp VR256:$src1, (loadv4i64 addr:$src2)), (VANDNPSYrm VR256:$src1, addr:$src2)>; } diff --git a/llvm/test/CodeGen/X86/WidenArith.ll b/llvm/test/CodeGen/X86/WidenArith.ll index cdd1a2818b2..cc5fcba6670 100644 --- a/llvm/test/CodeGen/X86/WidenArith.ll +++ b/llvm/test/CodeGen/X86/WidenArith.ll @@ -9,8 +9,8 @@ define <8 x i32> @test(<8 x float> %a, <8 x float> %b) { ; CHECK-NEXT: vsubps %ymm2, %ymm1, %ymm3 ; CHECK-NEXT: vcmpltps %ymm1, %ymm0, %ymm0 ; CHECK-NEXT: vcmpltps %ymm3, %ymm2, %ymm1 -; CHECK-NEXT: vandps {{.*}}(%rip), %ymm1, %ymm1 ; CHECK-NEXT: vandps %ymm1, %ymm0, %ymm0 +; CHECK-NEXT: vandps {{.*}}(%rip), %ymm0, %ymm0 ; CHECK-NEXT: retq %c1 = fadd <8 x float> %a, %b %b1 = fmul <8 x float> %b, %a diff --git a/llvm/test/CodeGen/X86/merge-consecutive-loads-256.ll b/llvm/test/CodeGen/X86/merge-consecutive-loads-256.ll index 8c2e9372900..dc268d9bdf8 100644 --- a/llvm/test/CodeGen/X86/merge-consecutive-loads-256.ll +++ b/llvm/test/CodeGen/X86/merge-consecutive-loads-256.ll @@ -547,29 +547,17 @@ define <16 x i16> @merge_16i16_i16_0uu3uuuuuuuuCuEF(i16* %ptr) nounwind uwtable } define <16 x i16> @merge_16i16_i16_0uu3zzuuuuuzCuEF(i16* %ptr) nounwind uwtable noinline ssp { -; AVX1-LABEL: merge_16i16_i16_0uu3zzuuuuuzCuEF: -; AVX1: # BB#0: -; AVX1-NEXT: vmovaps {{.*#+}} ymm0 = [65535,0,0,65535,0,0,0,0,0,0,0,0,65535,0,65535,65535] -; AVX1-NEXT: vandps (%rdi), %ymm0, %ymm0 -; AVX1-NEXT: retq -; -; AVX2-LABEL: merge_16i16_i16_0uu3zzuuuuuzCuEF: -; AVX2: # BB#0: -; AVX2-NEXT: vmovups (%rdi), %ymm0 -; AVX2-NEXT: vandps {{.*}}(%rip), %ymm0, %ymm0 -; AVX2-NEXT: retq -; -; AVX512F-LABEL: merge_16i16_i16_0uu3zzuuuuuzCuEF: -; AVX512F: # BB#0: -; AVX512F-NEXT: vmovups (%rdi), %ymm0 -; AVX512F-NEXT: vandps {{.*}}(%rip), %ymm0, %ymm0 -; AVX512F-NEXT: retq +; AVX-LABEL: merge_16i16_i16_0uu3zzuuuuuzCuEF: +; AVX: # BB#0: +; AVX-NEXT: vmovups (%rdi), %ymm0 +; AVX-NEXT: vandps {{.*}}(%rip), %ymm0, %ymm0 +; AVX-NEXT: retq ; ; X32-AVX-LABEL: merge_16i16_i16_0uu3zzuuuuuzCuEF: ; X32-AVX: # BB#0: ; X32-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax -; X32-AVX-NEXT: vmovaps {{.*#+}} ymm0 = [65535,0,0,65535,0,0,0,0,0,0,0,0,65535,0,65535,65535] -; X32-AVX-NEXT: vandps (%eax), %ymm0, %ymm0 +; X32-AVX-NEXT: vmovups (%eax), %ymm0 +; X32-AVX-NEXT: vandps {{\.LCPI.*}}, %ymm0, %ymm0 ; X32-AVX-NEXT: retl %ptr0 = getelementptr inbounds i16, i16* %ptr, i64 0 %ptr3 = getelementptr inbounds i16, i16* %ptr, i64 3 diff --git a/llvm/test/CodeGen/X86/v8i1-masks.ll b/llvm/test/CodeGen/X86/v8i1-masks.ll index 0135832ad92..d5c31506e98 100644 --- a/llvm/test/CodeGen/X86/v8i1-masks.ll +++ b/llvm/test/CodeGen/X86/v8i1-masks.ll @@ -13,8 +13,8 @@ define void @and_masks(<8 x float>* %a, <8 x float>* %b, <8 x float>* %c) nounwi ; X32-NEXT: vcmpltps %ymm0, %ymm1, %ymm1 ; X32-NEXT: vmovups (%eax), %ymm2 ; X32-NEXT: vcmpltps %ymm0, %ymm2, %ymm0 -; X32-NEXT: vandps LCPI0_0, %ymm1, %ymm1 ; X32-NEXT: vandps %ymm1, %ymm0, %ymm0 +; X32-NEXT: vandps LCPI0_0, %ymm0, %ymm0 ; X32-NEXT: vmovaps %ymm0, (%eax) ; X32-NEXT: vzeroupper ; X32-NEXT: retl @@ -26,8 +26,8 @@ define void @and_masks(<8 x float>* %a, <8 x float>* %b, <8 x float>* %c) nounwi ; X64-NEXT: vcmpltps %ymm0, %ymm1, %ymm1 ; X64-NEXT: vmovups (%rdx), %ymm2 ; X64-NEXT: vcmpltps %ymm0, %ymm2, %ymm0 -; X64-NEXT: vandps {{.*}}(%rip), %ymm1, %ymm1 ; X64-NEXT: vandps %ymm1, %ymm0, %ymm0 +; X64-NEXT: vandps {{.*}}(%rip), %ymm0, %ymm0 ; X64-NEXT: vmovaps %ymm0, (%rax) ; X64-NEXT: vzeroupper ; X64-NEXT: retq diff --git a/llvm/test/CodeGen/X86/vec_int_to_fp.ll b/llvm/test/CodeGen/X86/vec_int_to_fp.ll index 4a3ffde01f1..34daa607172 100644 --- a/llvm/test/CodeGen/X86/vec_int_to_fp.ll +++ b/llvm/test/CodeGen/X86/vec_int_to_fp.ll @@ -1953,15 +1953,15 @@ define <8 x float> @uitofp_8i32_to_8f32(<8 x i32> %a) { ; ; AVX1-LABEL: uitofp_8i32_to_8f32: ; AVX1: # BB#0: -; AVX1-NEXT: vandps {{.*}}(%rip), %ymm0, %ymm1 +; AVX1-NEXT: vpsrld $16, %xmm0, %xmm1 +; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2 +; AVX1-NEXT: vpsrld $16, %xmm2, %xmm2 +; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm1, %ymm1 ; AVX1-NEXT: vcvtdq2ps %ymm1, %ymm1 -; AVX1-NEXT: vpsrld $16, %xmm0, %xmm2 -; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm0 -; AVX1-NEXT: vpsrld $16, %xmm0, %xmm0 -; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm2, %ymm0 +; AVX1-NEXT: vmulps {{.*}}(%rip), %ymm1, %ymm1 +; AVX1-NEXT: vandps {{.*}}(%rip), %ymm0, %ymm0 ; AVX1-NEXT: vcvtdq2ps %ymm0, %ymm0 -; AVX1-NEXT: vmulps {{.*}}(%rip), %ymm0, %ymm0 -; AVX1-NEXT: vaddps %ymm1, %ymm0, %ymm0 +; AVX1-NEXT: vaddps %ymm0, %ymm1, %ymm0 ; AVX1-NEXT: retq ; ; AVX2-LABEL: uitofp_8i32_to_8f32: @@ -3890,16 +3890,16 @@ define <8 x float> @uitofp_load_8i32_to_8f32(<8 x i32> *%a) { ; ; AVX1-LABEL: uitofp_load_8i32_to_8f32: ; AVX1: # BB#0: -; AVX1-NEXT: vmovaps (%rdi), %ymm0 -; AVX1-NEXT: vandps {{.*}}(%rip), %ymm0, %ymm1 +; AVX1-NEXT: vmovdqa (%rdi), %ymm0 +; AVX1-NEXT: vpsrld $16, %xmm0, %xmm1 +; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2 +; AVX1-NEXT: vpsrld $16, %xmm2, %xmm2 +; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm1, %ymm1 ; AVX1-NEXT: vcvtdq2ps %ymm1, %ymm1 -; AVX1-NEXT: vpsrld $16, %xmm0, %xmm2 -; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm0 -; AVX1-NEXT: vpsrld $16, %xmm0, %xmm0 -; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm2, %ymm0 +; AVX1-NEXT: vmulps {{.*}}(%rip), %ymm1, %ymm1 +; AVX1-NEXT: vandps {{.*}}(%rip), %ymm0, %ymm0 ; AVX1-NEXT: vcvtdq2ps %ymm0, %ymm0 -; AVX1-NEXT: vmulps {{.*}}(%rip), %ymm0, %ymm0 -; AVX1-NEXT: vaddps %ymm1, %ymm0, %ymm0 +; AVX1-NEXT: vaddps %ymm0, %ymm1, %ymm0 ; AVX1-NEXT: retq ; ; AVX2-LABEL: uitofp_load_8i32_to_8f32: diff --git a/llvm/test/CodeGen/X86/vec_uint_to_fp-fastmath.ll b/llvm/test/CodeGen/X86/vec_uint_to_fp-fastmath.ll index c0e02bd1599..cb8e2096585 100644 --- a/llvm/test/CodeGen/X86/vec_uint_to_fp-fastmath.ll +++ b/llvm/test/CodeGen/X86/vec_uint_to_fp-fastmath.ll @@ -78,18 +78,18 @@ define <4 x float> @test_uitofp_v4i32_to_v4f32(<4 x i32> %arg) { ret <4 x float> %tmp } -; AVX: [[MASKCSTADDR_v8:.LCPI[0-9_]+]]: -; AVX-NEXT: .long 65535 # 0xffff -; AVX-NEXT: .long 65535 # 0xffff -; AVX-NEXT: .long 65535 # 0xffff -; AVX-NEXT: .long 65535 # 0xffff - ; AVX: [[FPMASKCSTADDR_v8:.LCPI[0-9_]+]]: ; AVX-NEXT: .long 1199570944 # float 65536 ; AVX-NEXT: .long 1199570944 # float 65536 ; AVX-NEXT: .long 1199570944 # float 65536 ; AVX-NEXT: .long 1199570944 # float 65536 +; AVX: [[MASKCSTADDR_v8:.LCPI[0-9_]+]]: +; AVX-NEXT: .long 65535 # 0xffff +; AVX-NEXT: .long 65535 # 0xffff +; AVX-NEXT: .long 65535 # 0xffff +; AVX-NEXT: .long 65535 # 0xffff + ; AVX2: [[FPMASKCSTADDR_v8:.LCPI[0-9_]+]]: ; AVX2-NEXT: .long 1199570944 # float 65536 @@ -119,15 +119,15 @@ define <8 x float> @test_uitofp_v8i32_to_v8f32(<8 x i32> %arg) { ; ; AVX-LABEL: test_uitofp_v8i32_to_v8f32: ; AVX: # BB#0: -; AVX-NEXT: vandps [[MASKCSTADDR_v8]](%rip), %ymm0, %ymm1 +; AVX-NEXT: vpsrld $16, %xmm0, %xmm1 +; AVX-NEXT: vextractf128 $1, %ymm0, %xmm2 +; AVX-NEXT: vpsrld $16, %xmm2, %xmm2 +; AVX-NEXT: vinsertf128 $1, %xmm2, %ymm1, %ymm1 ; AVX-NEXT: vcvtdq2ps %ymm1, %ymm1 -; AVX-NEXT: vpsrld $16, %xmm0, %xmm2 -; AVX-NEXT: vextractf128 $1, %ymm0, %xmm0 -; AVX-NEXT: vpsrld $16, %xmm0, %xmm0 -; AVX-NEXT: vinsertf128 $1, %xmm0, %ymm2, %ymm0 +; AVX-NEXT: vmulps [[FPMASKCSTADDR_v8]](%rip), %ymm1, %ymm1 +; AVX-NEXT: vandps [[MASKCSTADDR_v8]](%rip), %ymm0, %ymm0 ; AVX-NEXT: vcvtdq2ps %ymm0, %ymm0 -; AVX-NEXT: vmulps [[FPMASKCSTADDR_v8]](%rip), %ymm0, %ymm0 -; AVX-NEXT: vaddps %ymm1, %ymm0, %ymm0 +; AVX-NEXT: vaddps %ymm0, %ymm1, %ymm0 ; AVX-NEXT: retq ; ; AVX2-LABEL: test_uitofp_v8i32_to_v8f32: |