diff options
| author | Craig Topper <craig.topper@intel.com> | 2018-06-12 21:42:42 +0000 |
|---|---|---|
| committer | Craig Topper <craig.topper@intel.com> | 2018-06-12 21:42:42 +0000 |
| commit | 651ef359ab1ca11a99de4f5e8fd312fdbec9be24 (patch) | |
| tree | 0d224df3e720559474d1ce466544cca996c0eb16 /llvm | |
| parent | 5d89760ea4d2c65d45f88e1386baf6ffea226aa8 (diff) | |
| download | bcm5719-llvm-651ef359ab1ca11a99de4f5e8fd312fdbec9be24.tar.gz bcm5719-llvm-651ef359ab1ca11a99de4f5e8fd312fdbec9be24.zip | |
[X86] add avx512 tests for potentially miscompiling cvttp2si/cvttp2ui (PR37751).
llvm-svn: 334551
Diffstat (limited to 'llvm')
| -rw-r--r-- | llvm/test/CodeGen/X86/avx512-cvttp2i.ll | 358 |
1 files changed, 358 insertions, 0 deletions
diff --git a/llvm/test/CodeGen/X86/avx512-cvttp2i.ll b/llvm/test/CodeGen/X86/avx512-cvttp2i.ll new file mode 100644 index 00000000000..40b1143dcf5 --- /dev/null +++ b/llvm/test/CodeGen/X86/avx512-cvttp2i.ll @@ -0,0 +1,358 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=x86_64-- -mattr=avx512f,avx512vl,avx512dq | FileCheck %s --check-prefixes=CHECK + +; PR37751 - https://bugs.llvm.org/show_bug.cgi?id=37751 +; We can't combine into 'round' instructions because the behavior is different for out-of-range values. + +declare <16 x i32> @llvm.x86.avx512.mask.cvttps2dq.512(<16 x float>, <16 x i32>, i16, i32) +declare <4 x i32> @llvm.x86.avx512.mask.cvttps2udq.128(<4 x float>, <4 x i32>, i8) +declare <8 x i32> @llvm.x86.avx512.mask.cvttps2udq.256(<8 x float>, <8 x i32>, i8) +declare <16 x i32> @llvm.x86.avx512.mask.cvttps2udq.512(<16 x float>, <16 x i32>, i16, i32) +declare <4 x i32> @llvm.x86.avx512.mask.cvttpd2udq.256(<4 x double>, <4 x i32>, i8) +declare <8 x i32> @llvm.x86.avx512.mask.cvttpd2udq.512(<8 x double>, <8 x i32>, i8, i32) +declare <4 x i64> @llvm.x86.avx512.mask.cvttps2qq.256(<4 x float>, <4 x i64>, i8) +declare <8 x i64> @llvm.x86.avx512.mask.cvttps2qq.512(<8 x float>, <8 x i64>, i8, i32) +declare <4 x i64> @llvm.x86.avx512.mask.cvttps2uqq.256(<4 x float>, <4 x i64>, i8) +declare <8 x i64> @llvm.x86.avx512.mask.cvttps2uqq.512(<8 x float>, <8 x i64>, i8, i32) +declare <2 x i64> @llvm.x86.avx512.mask.cvttpd2qq.128(<2 x double>, <2 x i64>, i8) +declare <4 x i64> @llvm.x86.avx512.mask.cvttpd2qq.256(<4 x double>, <4 x i64>, i8) +declare <8 x i64> @llvm.x86.avx512.mask.cvttpd2qq.512(<8 x double>, <8 x i64>, i8, i32) +declare <2 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.128(<2 x double>, <2 x i64>, i8) +declare <4 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.256(<4 x double>, <4 x i64>, i8) +declare <8 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.512(<8 x double>, <8 x i64>, i8, i32) + +define <16 x float> @float_to_sint_to_float_mem_v16f32(<16 x float>* %p) { +; CHECK-LABEL: float_to_sint_to_float_mem_v16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vrndscaleps $11, (%rdi), %zmm0 +; CHECK-NEXT: retq + %x = load <16 x float>, <16 x float>* %p + %fptosi = tail call <16 x i32> @llvm.x86.avx512.mask.cvttps2dq.512(<16 x float> %x, <16 x i32> undef, i16 -1, i32 4) + %sitofp = sitofp <16 x i32> %fptosi to <16 x float> + ret <16 x float> %sitofp +} + +define <16 x float> @float_to_sint_to_float_reg_v16f32(<16 x float> %x) { +; CHECK-LABEL: float_to_sint_to_float_reg_v16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vrndscaleps $11, %zmm0, %zmm0 +; CHECK-NEXT: retq + %fptosi = tail call <16 x i32> @llvm.x86.avx512.mask.cvttps2dq.512(<16 x float> %x, <16 x i32> undef, i16 -1, i32 4) + %sitofp = sitofp <16 x i32> %fptosi to <16 x float> + ret <16 x float> %sitofp +} + +define <16 x float> @float_to_uint_to_float_mem_v16f32(<16 x float>* %p) { +; CHECK-LABEL: float_to_uint_to_float_mem_v16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vrndscaleps $11, (%rdi), %zmm0 +; CHECK-NEXT: retq + %x = load <16 x float>, <16 x float>* %p + %fptoui = tail call <16 x i32> @llvm.x86.avx512.mask.cvttps2udq.512(<16 x float> %x, <16 x i32> undef, i16 -1, i32 4) + %uitofp = uitofp <16 x i32> %fptoui to <16 x float> + ret <16 x float> %uitofp +} + +define <16 x float> @float_to_uint_to_float_reg_v16f32(<16 x float> %x) { +; CHECK-LABEL: float_to_uint_to_float_reg_v16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vrndscaleps $11, %zmm0, %zmm0 +; CHECK-NEXT: retq + %fptoui = tail call <16 x i32> @llvm.x86.avx512.mask.cvttps2udq.512(<16 x float> %x, <16 x i32> undef, i16 -1, i32 4) + %uitofp = uitofp <16 x i32> %fptoui to <16 x float> + ret <16 x float> %uitofp +} + +define <4 x float> @float_to_uint_to_float_mem_v4f32(<4 x float>* %p) { +; CHECK-LABEL: float_to_uint_to_float_mem_v4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundps $11, (%rdi), %xmm0 +; CHECK-NEXT: retq + %x = load <4 x float>, <4 x float>* %p + %fptoui = tail call <4 x i32> @llvm.x86.avx512.mask.cvttps2udq.128(<4 x float> %x, <4 x i32> undef, i8 -1) + %uitofp = uitofp <4 x i32> %fptoui to <4 x float> + ret <4 x float> %uitofp +} + +define <4 x float> @float_to_uint_to_float_reg_v4f32(<4 x float> %x) { +; CHECK-LABEL: float_to_uint_to_float_reg_v4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundps $11, %xmm0, %xmm0 +; CHECK-NEXT: retq + %fptoui = tail call <4 x i32> @llvm.x86.avx512.mask.cvttps2udq.128(<4 x float> %x, <4 x i32> undef, i8 -1) + %uitofp = uitofp <4 x i32> %fptoui to <4 x float> + ret <4 x float> %uitofp +} + +define <8 x float> @float_to_uint_to_float_mem_v8f32(<8 x float>* %p) { +; CHECK-LABEL: float_to_uint_to_float_mem_v8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundps $11, (%rdi), %ymm0 +; CHECK-NEXT: retq + %x = load <8 x float>, <8 x float>* %p + %fptoui = tail call <8 x i32> @llvm.x86.avx512.mask.cvttps2udq.256(<8 x float> %x, <8 x i32> undef, i8 -1) + %uitofp = uitofp <8 x i32> %fptoui to <8 x float> + ret <8 x float> %uitofp +} + +define <8 x float> @float_to_uint_to_float_reg_v8f32(<8 x float> %x) { +; CHECK-LABEL: float_to_uint_to_float_reg_v8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundps $11, %ymm0, %ymm0 +; CHECK-NEXT: retq + %fptoui = tail call <8 x i32> @llvm.x86.avx512.mask.cvttps2udq.256(<8 x float> %x, <8 x i32> undef, i8 -1) + %uitofp = uitofp <8 x i32> %fptoui to <8 x float> + ret <8 x float> %uitofp +} + +define <4 x double> @double_to_uint_to_double_mem_v4f64(<4 x double>* %p) { +; CHECK-LABEL: double_to_uint_to_double_mem_v4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundpd $11, (%rdi), %ymm0 +; CHECK-NEXT: retq + %x = load <4 x double>, <4 x double>* %p + %fptoui = tail call <4 x i32> @llvm.x86.avx512.mask.cvttpd2udq.256(<4 x double> %x, <4 x i32> undef, i8 -1) + %uitofp = uitofp <4 x i32> %fptoui to <4 x double> + ret <4 x double> %uitofp +} + +define <4 x double> @double_to_uint_to_double_reg_v4f64(<4 x double> %x) { +; CHECK-LABEL: double_to_uint_to_double_reg_v4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundpd $11, %ymm0, %ymm0 +; CHECK-NEXT: retq + %fptoui = tail call <4 x i32> @llvm.x86.avx512.mask.cvttpd2udq.256(<4 x double> %x, <4 x i32> undef, i8 -1) + %uitofp = uitofp <4 x i32> %fptoui to <4 x double> + ret <4 x double> %uitofp +} + +define <8 x double> @double_to_uint_to_double_mem_v8f64(<8 x double>* %p) { +; CHECK-LABEL: double_to_uint_to_double_mem_v8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vrndscalepd $11, (%rdi), %zmm0 +; CHECK-NEXT: retq + %x = load <8 x double>, <8 x double>* %p + %fptoui = tail call <8 x i32> @llvm.x86.avx512.mask.cvttpd2udq.512(<8 x double> %x, <8 x i32> undef, i8 -1, i32 4) + %uitofp = uitofp <8 x i32> %fptoui to <8 x double> + ret <8 x double> %uitofp +} + +define <8 x double> @double_to_uint_to_double_reg_v8f64(<8 x double> %x) { +; CHECK-LABEL: double_to_uint_to_double_reg_v8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vrndscalepd $11, %zmm0, %zmm0 +; CHECK-NEXT: retq + %fptoui = tail call <8 x i32> @llvm.x86.avx512.mask.cvttpd2udq.512(<8 x double> %x, <8 x i32> undef, i8 -1, i32 4) + %uitofp = uitofp <8 x i32> %fptoui to <8 x double> + ret <8 x double> %uitofp +} + +define <4 x float> @float_to_sint64_to_float_mem_v4f32(<4 x float>* %p) { +; CHECK-LABEL: float_to_sint64_to_float_mem_v4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundps $11, (%rdi), %xmm0 +; CHECK-NEXT: retq + %x = load <4 x float>, <4 x float>* %p + %fptosi = tail call <4 x i64> @llvm.x86.avx512.mask.cvttps2qq.256(<4 x float> %x, <4 x i64> undef, i8 -1) + %sitofp = sitofp <4 x i64> %fptosi to <4 x float> + ret <4 x float> %sitofp +} + +define <4 x float> @float_to_sint64_to_float_reg_v4f32(<4 x float> %x) { +; CHECK-LABEL: float_to_sint64_to_float_reg_v4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundps $11, %xmm0, %xmm0 +; CHECK-NEXT: retq + %fptosi = tail call <4 x i64> @llvm.x86.avx512.mask.cvttps2qq.256(<4 x float> %x, <4 x i64> undef, i8 -1) + %sitofp = sitofp <4 x i64> %fptosi to <4 x float> + ret <4 x float> %sitofp +} + +define <4 x float> @float_to_uint64_to_float_mem_v4f32(<4 x float>* %p) { +; CHECK-LABEL: float_to_uint64_to_float_mem_v4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundps $11, (%rdi), %xmm0 +; CHECK-NEXT: retq + %x = load <4 x float>, <4 x float>* %p + %fptoui = tail call <4 x i64> @llvm.x86.avx512.mask.cvttps2uqq.256(<4 x float> %x, <4 x i64> undef, i8 -1) + %uitofp = uitofp <4 x i64> %fptoui to <4 x float> + ret <4 x float> %uitofp +} + +define <4 x float> @float_to_uint64_to_float_reg_v4f32(<4 x float> %x) { +; CHECK-LABEL: float_to_uint64_to_float_reg_v4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundps $11, %xmm0, %xmm0 +; CHECK-NEXT: retq + %fptoui = tail call <4 x i64> @llvm.x86.avx512.mask.cvttps2uqq.256(<4 x float> %x, <4 x i64> undef, i8 -1) + %uitofp = uitofp <4 x i64> %fptoui to <4 x float> + ret <4 x float> %uitofp +} + +define <8 x float> @float_to_sint64_to_float_mem_v8f32(<8 x float>* %p) { +; CHECK-LABEL: float_to_sint64_to_float_mem_v8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundps $11, (%rdi), %ymm0 +; CHECK-NEXT: retq + %x = load <8 x float>, <8 x float>* %p + %fptosi = tail call <8 x i64> @llvm.x86.avx512.mask.cvttps2qq.512(<8 x float> %x, <8 x i64> undef, i8 -1, i32 4) + %sitofp = sitofp <8 x i64> %fptosi to <8 x float> + ret <8 x float> %sitofp +} + +define <8 x float> @float_to_sint64_to_float_reg_v8f32(<8 x float> %x) { +; CHECK-LABEL: float_to_sint64_to_float_reg_v8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundps $11, %ymm0, %ymm0 +; CHECK-NEXT: retq + %fptosi = tail call <8 x i64> @llvm.x86.avx512.mask.cvttps2qq.512(<8 x float> %x, <8 x i64> undef, i8 -1, i32 4) + %sitofp = sitofp <8 x i64> %fptosi to <8 x float> + ret <8 x float> %sitofp +} + +define <8 x float> @float_to_uint64_to_float_mem_v8f32(<8 x float>* %p) { +; CHECK-LABEL: float_to_uint64_to_float_mem_v8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundps $11, (%rdi), %ymm0 +; CHECK-NEXT: retq + %x = load <8 x float>, <8 x float>* %p + %fptoui = tail call <8 x i64> @llvm.x86.avx512.mask.cvttps2uqq.512(<8 x float> %x, <8 x i64> undef, i8 -1, i32 4) + %uitofp = uitofp <8 x i64> %fptoui to <8 x float> + ret <8 x float> %uitofp +} + +define <8 x float> @float_to_uint64_to_float_reg_v8f32(<8 x float> %x) { +; CHECK-LABEL: float_to_uint64_to_float_reg_v8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundps $11, %ymm0, %ymm0 +; CHECK-NEXT: retq + %fptoui = tail call <8 x i64> @llvm.x86.avx512.mask.cvttps2uqq.512(<8 x float> %x, <8 x i64> undef, i8 -1, i32 4) + %uitofp = uitofp <8 x i64> %fptoui to <8 x float> + ret <8 x float> %uitofp +} + +define <2 x double> @double_to_sint64_to_double_mem_v2f64(<2 x double>* %p) { +; CHECK-LABEL: double_to_sint64_to_double_mem_v2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundpd $11, (%rdi), %xmm0 +; CHECK-NEXT: retq + %x = load <2 x double>, <2 x double>* %p + %fptosi = tail call <2 x i64> @llvm.x86.avx512.mask.cvttpd2qq.128(<2 x double> %x, <2 x i64> undef, i8 -1) + %sitofp = sitofp <2 x i64> %fptosi to <2 x double> + ret <2 x double> %sitofp +} + +define <2 x double> @double_to_sint64_to_double_reg_v2f64(<2 x double> %x) { +; CHECK-LABEL: double_to_sint64_to_double_reg_v2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundpd $11, %xmm0, %xmm0 +; CHECK-NEXT: retq + %fptosi = tail call <2 x i64> @llvm.x86.avx512.mask.cvttpd2qq.128(<2 x double> %x, <2 x i64> undef, i8 -1) + %sitofp = sitofp <2 x i64> %fptosi to <2 x double> + ret <2 x double> %sitofp +} + +define <2 x double> @double_to_uint64_to_double_mem_v2f64(<2 x double>* %p) { +; CHECK-LABEL: double_to_uint64_to_double_mem_v2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundpd $11, (%rdi), %xmm0 +; CHECK-NEXT: retq + %x = load <2 x double>, <2 x double>* %p + %fptoui = tail call <2 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.128(<2 x double> %x, <2 x i64> undef, i8 -1) + %uitofp = uitofp <2 x i64> %fptoui to <2 x double> + ret <2 x double> %uitofp +} + +define <2 x double> @double_to_uint64_to_double_reg_v2f64(<2 x double> %x) { +; CHECK-LABEL: double_to_uint64_to_double_reg_v2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundpd $11, %xmm0, %xmm0 +; CHECK-NEXT: retq + %fptoui = tail call <2 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.128(<2 x double> %x, <2 x i64> undef, i8 -1) + %uitofp = uitofp <2 x i64> %fptoui to <2 x double> + ret <2 x double> %uitofp +} + +define <4 x double> @double_to_sint64_to_double_mem_v4f64(<4 x double>* %p) { +; CHECK-LABEL: double_to_sint64_to_double_mem_v4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundpd $11, (%rdi), %ymm0 +; CHECK-NEXT: retq + %x = load <4 x double>, <4 x double>* %p + %fptosi = tail call <4 x i64> @llvm.x86.avx512.mask.cvttpd2qq.256(<4 x double> %x, <4 x i64> undef, i8 -1) + %sitofp = sitofp <4 x i64> %fptosi to <4 x double> + ret <4 x double> %sitofp +} + +define <4 x double> @double_to_sint64_to_double_reg_v4f64(<4 x double> %x) { +; CHECK-LABEL: double_to_sint64_to_double_reg_v4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundpd $11, %ymm0, %ymm0 +; CHECK-NEXT: retq + %fptosi = tail call <4 x i64> @llvm.x86.avx512.mask.cvttpd2qq.256(<4 x double> %x, <4 x i64> undef, i8 -1) + %sitofp = sitofp <4 x i64> %fptosi to <4 x double> + ret <4 x double> %sitofp +} + +define <4 x double> @double_to_uint64_to_double_mem_v4f64(<4 x double>* %p) { +; CHECK-LABEL: double_to_uint64_to_double_mem_v4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundpd $11, (%rdi), %ymm0 +; CHECK-NEXT: retq + %x = load <4 x double>, <4 x double>* %p + %fptoui = tail call <4 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.256(<4 x double> %x, <4 x i64> undef, i8 -1) + %uitofp = uitofp <4 x i64> %fptoui to <4 x double> + ret <4 x double> %uitofp +} + +define <4 x double> @double_to_uint64_to_double_reg_v4f64(<4 x double> %x) { +; CHECK-LABEL: double_to_uint64_to_double_reg_v4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vroundpd $11, %ymm0, %ymm0 +; CHECK-NEXT: retq + %fptoui = tail call <4 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.256(<4 x double> %x, <4 x i64> undef, i8 -1) + %uitofp = uitofp <4 x i64> %fptoui to <4 x double> + ret <4 x double> %uitofp +} + +define <8 x double> @double_to_sint64_to_double_mem_v8f64(<8 x double>* %p) { +; CHECK-LABEL: double_to_sint64_to_double_mem_v8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vrndscalepd $11, (%rdi), %zmm0 +; CHECK-NEXT: retq + %x = load <8 x double>, <8 x double>* %p + %fptosi = tail call <8 x i64> @llvm.x86.avx512.mask.cvttpd2qq.512(<8 x double> %x, <8 x i64> undef, i8 -1, i32 4) + %sitofp = sitofp <8 x i64> %fptosi to <8 x double> + ret <8 x double> %sitofp +} + +define <8 x double> @double_to_sint64_to_double_reg_v8f64(<8 x double> %x) { +; CHECK-LABEL: double_to_sint64_to_double_reg_v8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vrndscalepd $11, %zmm0, %zmm0 +; CHECK-NEXT: retq + %fptosi = tail call <8 x i64> @llvm.x86.avx512.mask.cvttpd2qq.512(<8 x double> %x, <8 x i64> undef, i8 -1, i32 4) + %sitofp = sitofp <8 x i64> %fptosi to <8 x double> + ret <8 x double> %sitofp +} + +define <8 x double> @double_to_uint64_to_double_mem_v8f64(<8 x double>* %p) { +; CHECK-LABEL: double_to_uint64_to_double_mem_v8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vrndscalepd $11, (%rdi), %zmm0 +; CHECK-NEXT: retq + %x = load <8 x double>, <8 x double>* %p + %fptoui = tail call <8 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.512(<8 x double> %x, <8 x i64> undef, i8 -1, i32 4) + %uitofp = uitofp <8 x i64> %fptoui to <8 x double> + ret <8 x double> %uitofp +} + +define <8 x double> @double_to_uint64_to_double_reg_v8f64(<8 x double> %x) { +; CHECK-LABEL: double_to_uint64_to_double_reg_v8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vrndscalepd $11, %zmm0, %zmm0 +; CHECK-NEXT: retq + %fptoui = tail call <8 x i64> @llvm.x86.avx512.mask.cvttpd2uqq.512(<8 x double> %x, <8 x i64> undef, i8 -1, i32 4) + %uitofp = uitofp <8 x i64> %fptoui to <8 x double> + ret <8 x double> %uitofp +} |

