diff options
| author | Craig Topper <craig.topper@intel.com> | 2018-05-10 21:49:16 +0000 |
|---|---|---|
| committer | Craig Topper <craig.topper@intel.com> | 2018-05-10 21:49:16 +0000 |
| commit | 1ee19ae126291314d047c0032a04549f222dbc96 (patch) | |
| tree | b52a7999039d480fb5b7a529899d8b8975fdbc32 /llvm/test/CodeGen/X86 | |
| parent | 2903a9bb0201fdb56da3e102bb5de917f16583b0 (diff) | |
| download | bcm5719-llvm-1ee19ae126291314d047c0032a04549f222dbc96.tar.gz bcm5719-llvm-1ee19ae126291314d047c0032a04549f222dbc96.zip | |
[X86] Add new patterns for masked scalar load/store to match clang's codegen from r331958.
Clang's codegen now uses 128-bit masked load/store intrinsics in IR. The backend will widen to 512-bits on AVX512F targets.
So this patch adds patterns to detect codegen's widening and patterns for AVX512VL that don't get widened.
We may be able to drop some of the old patterns, but I leave that for a future patch.
llvm-svn: 332049
Diffstat (limited to 'llvm/test/CodeGen/X86')
| -rw-r--r-- | llvm/test/CodeGen/X86/avx512-load-store.ll | 153 |
1 files changed, 153 insertions, 0 deletions
diff --git a/llvm/test/CodeGen/X86/avx512-load-store.ll b/llvm/test/CodeGen/X86/avx512-load-store.ll index b6000c09da2..c82843d1b8d 100644 --- a/llvm/test/CodeGen/X86/avx512-load-store.ll +++ b/llvm/test/CodeGen/X86/avx512-load-store.ll @@ -1,6 +1,8 @@ ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc < %s -O2 -mattr=avx512f -mtriple=x86_64-unknown | FileCheck %s --check-prefix=CHECK --check-prefix=CHECK64 ; RUN: llc < %s -O2 -mattr=avx512f -mtriple=i386-unknown | FileCheck %s --check-prefix=CHECK --check-prefix=CHECK32 +; RUN: llc < %s -O2 -mattr=avx512vl -mtriple=x86_64-unknown | FileCheck %s --check-prefix=CHECK --check-prefix=CHECK64 +; RUN: llc < %s -O2 -mattr=avx512vl -mtriple=i386-unknown | FileCheck %s --check-prefix=CHECK --check-prefix=CHECK32 define <4 x float> @test_mm_mask_move_ss(<4 x float> %__W, i8 zeroext %__U, <4 x float> %__A, <4 x float> %__B) local_unnamed_addr #0 { ; CHECK64-LABEL: test_mm_mask_move_ss: @@ -237,6 +239,149 @@ entry: ret <2 x double> %shuffle.i } +; The tests below match clang's newer codegen that uses 128-bit masked load/stores. + +define void @test_mm_mask_store_ss_2(float* %__P, i8 zeroext %__U, <4 x float> %__A) { +; CHECK64-LABEL: test_mm_mask_store_ss_2: +; CHECK64: # %bb.0: # %entry +; CHECK64-NEXT: kmovw %esi, %k1 +; CHECK64-NEXT: vmovss %xmm0, (%rdi) {%k1} +; CHECK64-NEXT: retq +; +; CHECK32-LABEL: test_mm_mask_store_ss_2: +; CHECK32: # %bb.0: # %entry +; CHECK32-NEXT: movl {{[0-9]+}}(%esp), %eax +; CHECK32-NEXT: movb {{[0-9]+}}(%esp), %cl +; CHECK32-NEXT: kmovw %ecx, %k1 +; CHECK32-NEXT: vmovss %xmm0, (%eax) {%k1} +; CHECK32-NEXT: retl +entry: + %0 = bitcast float* %__P to <4 x float>* + %1 = and i8 %__U, 1 + %2 = bitcast i8 %1 to <8 x i1> + %extract.i = shufflevector <8 x i1> %2, <8 x i1> undef, <4 x i32> <i32 0, i32 1, i32 2, i32 3> + tail call void @llvm.masked.store.v4f32.p0v4f32(<4 x float> %__A, <4 x float>* %0, i32 1, <4 x i1> %extract.i) + ret void +} + +define void @test_mm_mask_store_sd_2(double* %__P, i8 zeroext %__U, <2 x double> %__A) { +; CHECK64-LABEL: test_mm_mask_store_sd_2: +; CHECK64: # %bb.0: # %entry +; CHECK64-NEXT: kmovw %esi, %k1 +; CHECK64-NEXT: vmovsd %xmm0, (%rdi) {%k1} +; CHECK64-NEXT: retq +; +; CHECK32-LABEL: test_mm_mask_store_sd_2: +; CHECK32: # %bb.0: # %entry +; CHECK32-NEXT: movl {{[0-9]+}}(%esp), %eax +; CHECK32-NEXT: movb {{[0-9]+}}(%esp), %cl +; CHECK32-NEXT: kmovw %ecx, %k1 +; CHECK32-NEXT: vmovsd %xmm0, (%eax) {%k1} +; CHECK32-NEXT: retl +entry: + %0 = bitcast double* %__P to <2 x double>* + %1 = and i8 %__U, 1 + %2 = bitcast i8 %1 to <8 x i1> + %extract.i = shufflevector <8 x i1> %2, <8 x i1> undef, <2 x i32> <i32 0, i32 1> + tail call void @llvm.masked.store.v2f64.p0v2f64(<2 x double> %__A, <2 x double>* %0, i32 1, <2 x i1> %extract.i) + ret void +} + +define <4 x float> @test_mm_mask_load_ss_2(<4 x float> %__A, i8 zeroext %__U, float* readonly %__W) { +; CHECK64-LABEL: test_mm_mask_load_ss_2: +; CHECK64: # %bb.0: # %entry +; CHECK64-NEXT: kmovw %edi, %k1 +; CHECK64-NEXT: vmovss (%rsi), %xmm0 {%k1} +; CHECK64-NEXT: retq +; +; CHECK32-LABEL: test_mm_mask_load_ss_2: +; CHECK32: # %bb.0: # %entry +; CHECK32-NEXT: movl {{[0-9]+}}(%esp), %eax +; CHECK32-NEXT: movb {{[0-9]+}}(%esp), %cl +; CHECK32-NEXT: kmovw %ecx, %k1 +; CHECK32-NEXT: vmovss (%eax), %xmm0 {%k1} +; CHECK32-NEXT: retl +entry: + %shuffle.i = shufflevector <4 x float> %__A, <4 x float> <float 0.000000e+00, float undef, float undef, float undef>, <4 x i32> <i32 0, i32 4, i32 4, i32 4> + %0 = bitcast float* %__W to <4 x float>* + %1 = and i8 %__U, 1 + %2 = bitcast i8 %1 to <8 x i1> + %extract.i = shufflevector <8 x i1> %2, <8 x i1> undef, <4 x i32> <i32 0, i32 1, i32 2, i32 3> + %3 = tail call <4 x float> @llvm.masked.load.v4f32.p0v4f32(<4 x float>* %0, i32 1, <4 x i1> %extract.i, <4 x float> %shuffle.i) + ret <4 x float> %3 +} + +define <4 x float> @test_mm_maskz_load_ss_2(i8 zeroext %__U, float* readonly %__W) { +; CHECK64-LABEL: test_mm_maskz_load_ss_2: +; CHECK64: # %bb.0: # %entry +; CHECK64-NEXT: kmovw %edi, %k1 +; CHECK64-NEXT: vmovss (%rsi), %xmm0 {%k1} {z} +; CHECK64-NEXT: retq +; +; CHECK32-LABEL: test_mm_maskz_load_ss_2: +; CHECK32: # %bb.0: # %entry +; CHECK32-NEXT: movl {{[0-9]+}}(%esp), %eax +; CHECK32-NEXT: movb {{[0-9]+}}(%esp), %cl +; CHECK32-NEXT: kmovw %ecx, %k1 +; CHECK32-NEXT: vmovss (%eax), %xmm0 {%k1} {z} +; CHECK32-NEXT: retl +entry: + %0 = bitcast float* %__W to <4 x float>* + %1 = and i8 %__U, 1 + %2 = bitcast i8 %1 to <8 x i1> + %extract.i = shufflevector <8 x i1> %2, <8 x i1> undef, <4 x i32> <i32 0, i32 1, i32 2, i32 3> + %3 = tail call <4 x float> @llvm.masked.load.v4f32.p0v4f32(<4 x float>* %0, i32 1, <4 x i1> %extract.i, <4 x float> zeroinitializer) + ret <4 x float> %3 +} + +define <2 x double> @test_mm_mask_load_sd_2(<2 x double> %__A, i8 zeroext %__U, double* readonly %__W) { +; CHECK64-LABEL: test_mm_mask_load_sd_2: +; CHECK64: # %bb.0: # %entry +; CHECK64-NEXT: kmovw %edi, %k1 +; CHECK64-NEXT: vmovsd (%rsi), %xmm0 {%k1} +; CHECK64-NEXT: retq +; +; CHECK32-LABEL: test_mm_mask_load_sd_2: +; CHECK32: # %bb.0: # %entry +; CHECK32-NEXT: movl {{[0-9]+}}(%esp), %eax +; CHECK32-NEXT: movb {{[0-9]+}}(%esp), %cl +; CHECK32-NEXT: kmovw %ecx, %k1 +; CHECK32-NEXT: vmovsd (%eax), %xmm0 {%k1} +; CHECK32-NEXT: retl +entry: + %shuffle3.i = insertelement <2 x double> %__A, double 0.000000e+00, i32 1 + %0 = bitcast double* %__W to <2 x double>* + %1 = and i8 %__U, 1 + %2 = bitcast i8 %1 to <8 x i1> + %extract.i = shufflevector <8 x i1> %2, <8 x i1> undef, <2 x i32> <i32 0, i32 1> + %3 = tail call <2 x double> @llvm.masked.load.v2f64.p0v2f64(<2 x double>* %0, i32 1, <2 x i1> %extract.i, <2 x double> %shuffle3.i) + ret <2 x double> %3 +} + +define <2 x double> @test_mm_maskz_load_sd_2(i8 zeroext %__U, double* readonly %__W) { +; CHECK64-LABEL: test_mm_maskz_load_sd_2: +; CHECK64: # %bb.0: # %entry +; CHECK64-NEXT: kmovw %edi, %k1 +; CHECK64-NEXT: vmovsd (%rsi), %xmm0 {%k1} {z} +; CHECK64-NEXT: retq +; +; CHECK32-LABEL: test_mm_maskz_load_sd_2: +; CHECK32: # %bb.0: # %entry +; CHECK32-NEXT: movl {{[0-9]+}}(%esp), %eax +; CHECK32-NEXT: movb {{[0-9]+}}(%esp), %cl +; CHECK32-NEXT: kmovw %ecx, %k1 +; CHECK32-NEXT: vmovsd (%eax), %xmm0 {%k1} {z} +; CHECK32-NEXT: retl +entry: + %0 = bitcast double* %__W to <2 x double>* + %1 = and i8 %__U, 1 + %2 = bitcast i8 %1 to <8 x i1> + %extract.i = shufflevector <8 x i1> %2, <8 x i1> undef, <2 x i32> <i32 0, i32 1> + %3 = tail call <2 x double> @llvm.masked.load.v2f64.p0v2f64(<2 x double>* %0, i32 1, <2 x i1> %extract.i, <2 x double> zeroinitializer) + ret <2 x double> %3 +} + + declare void @llvm.masked.store.v16f32.p0v16f32(<16 x float>, <16 x float>*, i32, <16 x i1>) #3 declare void @llvm.masked.store.v8f64.p0v8f64(<8 x double>, <8 x double>*, i32, <8 x i1>) #3 @@ -244,3 +389,11 @@ declare void @llvm.masked.store.v8f64.p0v8f64(<8 x double>, <8 x double>*, i32, declare <16 x float> @llvm.masked.load.v16f32.p0v16f32(<16 x float>*, i32, <16 x i1>, <16 x float>) #4 declare <8 x double> @llvm.masked.load.v8f64.p0v8f64(<8 x double>*, i32, <8 x i1>, <8 x double>) #4 + +declare void @llvm.masked.store.v4f32.p0v4f32(<4 x float>, <4 x float>*, i32, <4 x i1>) + +declare void @llvm.masked.store.v2f64.p0v2f64(<2 x double>, <2 x double>*, i32, <2 x i1>) + +declare <4 x float> @llvm.masked.load.v4f32.p0v4f32(<4 x float>*, i32, <4 x i1>, <4 x float>) + +declare <2 x double> @llvm.masked.load.v2f64.p0v2f64(<2 x double>*, i32, <2 x i1>, <2 x double>) |

