diff options
author | Louis Gerbarg <lgg@apple.com> | 2014-05-09 17:02:49 +0000 |
---|---|---|
committer | Louis Gerbarg <lgg@apple.com> | 2014-05-09 17:02:49 +0000 |
commit | 3342bf1451f4d0844ca9355abc976cdadecbfac1 (patch) | |
tree | 8f9f6d7e64673e43dd72776de79729869c7449c5 /llvm/test/CodeGen/ARM/intrinsics-overflow.ll | |
parent | 1f54b821643b8a246a7a6161847caa177b81ee64 (diff) | |
download | bcm5719-llvm-3342bf1451f4d0844ca9355abc976cdadecbfac1.tar.gz bcm5719-llvm-3342bf1451f4d0844ca9355abc976cdadecbfac1.zip |
Add custom lowering for add/sub with overflow intrinsics to ARM
This patch adds support to ARM for custom lowering of the
llvm.{u|s}add.with.overflow.i32 intrinsics for i32/i64. This is particularly useful
for handling idiomatic saturating math functions as generated by
InstCombineCompare.
Test cases included.
rdar://14853450
llvm-svn: 208435
Diffstat (limited to 'llvm/test/CodeGen/ARM/intrinsics-overflow.ll')
-rw-r--r-- | llvm/test/CodeGen/ARM/intrinsics-overflow.ll | 57 |
1 files changed, 57 insertions, 0 deletions
diff --git a/llvm/test/CodeGen/ARM/intrinsics-overflow.ll b/llvm/test/CodeGen/ARM/intrinsics-overflow.ll new file mode 100644 index 00000000000..c414add59f9 --- /dev/null +++ b/llvm/test/CodeGen/ARM/intrinsics-overflow.ll @@ -0,0 +1,57 @@ +; RUN: llc < %s -march=arm -mcpu=generic | FileCheck %s + +define i32 @uadd_overflow(i32 %a, i32 %b) #0 { + %sadd = tail call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %a, i32 %b) + %1 = extractvalue { i32, i1 } %sadd, 1 + %2 = zext i1 %1 to i32 + ret i32 %2 + + ; CHECK-LABEL: uadd_overflow: + ; CHECK: add r[[R2:[0-9]+]], r[[R0:[0-9]+]], r[[R1:[0-9]+]] + ; CHECK: mov r[[R1]], #1 + ; CHECK: cmp r[[R2]], r[[R0]] + ; CHECK: movhs r[[R1]], #0 +} + + +define i32 @sadd_overflow(i32 %a, i32 %b) #0 { + %sadd = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) + %1 = extractvalue { i32, i1 } %sadd, 1 + %2 = zext i1 %1 to i32 + ret i32 %2 + + ; CHECK-LABEL: sadd_overflow: + ; CHECK: add r[[R2:[0-9]+]], r[[R0:[0-9]+]], r[[R1:[0-9]+]] + ; CHECK: mov r[[R1]], #1 + ; CHECK: cmp r[[R2]], r[[R0]] + ; CHECK: movvc r[[R1]], #0 +} + +define i32 @usub_overflow(i32 %a, i32 %b) #0 { + %sadd = tail call { i32, i1 } @llvm.usub.with.overflow.i32(i32 %a, i32 %b) + %1 = extractvalue { i32, i1 } %sadd, 1 + %2 = zext i1 %1 to i32 + ret i32 %2 + + ; CHECK-LABEL: usub_overflow: + ; CHECK: mov r[[R2]], #1 + ; CHECK: cmp r[[R0]], r[[R1]] + ; CHECK: movhs r[[R2]], #0 +} + +define i32 @ssub_overflow(i32 %a, i32 %b) #0 { + %sadd = tail call { i32, i1 } @llvm.ssub.with.overflow.i32(i32 %a, i32 %b) + %1 = extractvalue { i32, i1 } %sadd, 1 + %2 = zext i1 %1 to i32 + ret i32 %2 + + ; CHECK-LABEL: ssub_overflow: + ; CHECK: mov r[[R2]], #1 + ; CHECK: cmp r[[R0]], r[[R1]] + ; CHECK: movvc r[[R2]], #0 +} + +declare { i32, i1 } @llvm.uadd.with.overflow.i32(i32, i32) #1 +declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) #2 +declare { i32, i1 } @llvm.usub.with.overflow.i32(i32, i32) #3 +declare { i32, i1 } @llvm.ssub.with.overflow.i32(i32, i32) #4 |