summaryrefslogtreecommitdiffstats
path: root/llvm/test/CodeGen/X86/combine-add-ssat.ll
diff options
context:
space:
mode:
authorSimon Pilgrim <llvm-dev@redking.me.uk>2019-01-14 12:12:42 +0000
committerSimon Pilgrim <llvm-dev@redking.me.uk>2019-01-14 12:12:42 +0000
commit67610926fc74d03061da2c1ce4c895d17bee236f (patch)
tree37448332f8936ea2d271542c7c2518e9f8ef5edb /llvm/test/CodeGen/X86/combine-add-ssat.ll
parent8987d00653a15214c5e21691c139ea8304bc99e6 (diff)
downloadbcm5719-llvm-67610926fc74d03061da2c1ce4c895d17bee236f.tar.gz
bcm5719-llvm-67610926fc74d03061da2c1ce4c895d17bee236f.zip
[DAGCombiner] Add add saturation constant folding tests.
Exposes an issue with sadd_sat for computeOverflowKind, so I've disabled it for now. llvm-svn: 351057
Diffstat (limited to 'llvm/test/CodeGen/X86/combine-add-ssat.ll')
-rw-r--r--llvm/test/CodeGen/X86/combine-add-ssat.ll49
1 files changed, 41 insertions, 8 deletions
diff --git a/llvm/test/CodeGen/X86/combine-add-ssat.ll b/llvm/test/CodeGen/X86/combine-add-ssat.ll
index 217e91a62a4..63fde10a325 100644
--- a/llvm/test/CodeGen/X86/combine-add-ssat.ll
+++ b/llvm/test/CodeGen/X86/combine-add-ssat.ll
@@ -11,6 +11,39 @@ declare i32 @llvm.sadd.sat.i32 (i32, i32)
declare i64 @llvm.sadd.sat.i64 (i64, i64)
declare <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16>, <8 x i16>)
+; fold (sadd_sat c1, c2) -> c3
+define i32 @combine_constfold_i32() {
+; CHECK-LABEL: combine_constfold_i32:
+; CHECK: # %bb.0:
+; CHECK-NEXT: movl $2147483647, %eax # imm = 0x7FFFFFFF
+; CHECK-NEXT: xorl %ecx, %ecx
+; CHECK-NEXT: movl $2147483647, %edx # imm = 0x7FFFFFFF
+; CHECK-NEXT: addl $100, %edx
+; CHECK-NEXT: setns %cl
+; CHECK-NEXT: addl $2147483647, %ecx # imm = 0x7FFFFFFF
+; CHECK-NEXT: addl $100, %eax
+; CHECK-NEXT: cmovol %ecx, %eax
+; CHECK-NEXT: retq
+ %res = call i32 @llvm.sadd.sat.i32(i32 2147483647, i32 100)
+ ret i32 %res
+}
+
+define <8 x i16> @combine_constfold_v8i16() {
+; SSE-LABEL: combine_constfold_v8i16:
+; SSE: # %bb.0:
+; SSE-NEXT: movdqa {{.*#+}} xmm0 = [0,1,255,65535,65535,65281,1,1]
+; SSE-NEXT: paddsw {{.*}}(%rip), %xmm0
+; SSE-NEXT: retq
+;
+; AVX-LABEL: combine_constfold_v8i16:
+; AVX: # %bb.0:
+; AVX-NEXT: vmovdqa {{.*#+}} xmm0 = [0,1,255,65535,65535,65281,1,1]
+; AVX-NEXT: vpaddsw {{.*}}(%rip), %xmm0, %xmm0
+; AVX-NEXT: retq
+ %res = call <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16> <i16 0, i16 1, i16 255, i16 65535, i16 -1, i16 -255, i16 -65535, i16 1>, <8 x i16> <i16 1, i16 65535, i16 1, i16 65535, i16 1, i16 65535, i16 1, i16 65535>)
+ ret <8 x i16> %res
+}
+
; fold (sadd_sat c, x) -> (sadd_sat x, c)
define i32 @combine_constant_i32(i32 %a0) {
; CHECK-LABEL: combine_constant_i32:
@@ -23,8 +56,8 @@ define i32 @combine_constant_i32(i32 %a0) {
; CHECK-NEXT: incl %edi
; CHECK-NEXT: cmovnol %edi, %eax
; CHECK-NEXT: retq
- %res = call i32 @llvm.sadd.sat.i32(i32 1, i32 %a0);
- ret i32 %res;
+ %res = call i32 @llvm.sadd.sat.i32(i32 1, i32 %a0)
+ ret i32 %res
}
define <8 x i16> @combine_constant_v8i16(<8 x i16> %a0) {
@@ -37,8 +70,8 @@ define <8 x i16> @combine_constant_v8i16(<8 x i16> %a0) {
; AVX: # %bb.0:
; AVX-NEXT: vpaddsw {{.*}}(%rip), %xmm0, %xmm0
; AVX-NEXT: retq
- %res = call <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16> <i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1>, <8 x i16> %a0);
- ret <8 x i16> %res;
+ %res = call <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16> <i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1>, <8 x i16> %a0)
+ ret <8 x i16> %res
}
; fold (sadd_sat c, 0) -> x
@@ -47,7 +80,7 @@ define i32 @combine_zero_i32(i32 %a0) {
; CHECK: # %bb.0:
; CHECK-NEXT: movl %edi, %eax
; CHECK-NEXT: retq
- %1 = call i32 @llvm.sadd.sat.i32(i32 %a0, i32 0);
+ %1 = call i32 @llvm.sadd.sat.i32(i32 %a0, i32 0)
ret i32 %1
}
@@ -55,7 +88,7 @@ define <8 x i16> @combine_zero_v8i16(<8 x i16> %a0) {
; CHECK-LABEL: combine_zero_v8i16:
; CHECK: # %bb.0:
; CHECK-NEXT: retq
- %1 = call <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16> %a0, <8 x i16> zeroinitializer);
+ %1 = call <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16> %a0, <8 x i16> zeroinitializer)
ret <8 x i16> %1
}
@@ -75,7 +108,7 @@ define i32 @combine_no_overflow_i32(i32 %a0, i32 %a1) {
; CHECK-NEXT: retq
%1 = ashr i32 %a0, 16
%2 = lshr i32 %a1, 16
- %3 = call i32 @llvm.sadd.sat.i32(i32 %1, i32 %2);
+ %3 = call i32 @llvm.sadd.sat.i32(i32 %1, i32 %2)
ret i32 %3
}
@@ -95,6 +128,6 @@ define <8 x i16> @combine_no_overflow_v8i16(<8 x i16> %a0, <8 x i16> %a1) {
; AVX-NEXT: retq
%1 = ashr <8 x i16> %a0, <i16 10, i16 10, i16 10, i16 10, i16 10, i16 10, i16 10, i16 10>
%2 = lshr <8 x i16> %a1, <i16 10, i16 10, i16 10, i16 10, i16 10, i16 10, i16 10, i16 10>
- %3 = call <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16> %1, <8 x i16> %2);
+ %3 = call <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16> %1, <8 x i16> %2)
ret <8 x i16> %3
}
OpenPOWER on IntegriCloud