diff options
Diffstat (limited to 'llvm/test/Transforms/CodeGenPrepare/NVPTX/bypass-slow-div.ll')
-rw-r--r-- | llvm/test/Transforms/CodeGenPrepare/NVPTX/bypass-slow-div.ll | 77 |
1 files changed, 77 insertions, 0 deletions
diff --git a/llvm/test/Transforms/CodeGenPrepare/NVPTX/bypass-slow-div.ll b/llvm/test/Transforms/CodeGenPrepare/NVPTX/bypass-slow-div.ll index 4846d52f4d2..4d824e450ff 100644 --- a/llvm/test/Transforms/CodeGenPrepare/NVPTX/bypass-slow-div.ll +++ b/llvm/test/Transforms/CodeGenPrepare/NVPTX/bypass-slow-div.ll @@ -27,3 +27,80 @@ define void @rem_only(i64 %a, i64 %b, i64* %retptr) { store i64 %d, i64* %retptr ret void } + +; CHECK-LABEL: @udiv_by_constant( +define i64 @udiv_by_constant(i32 %a) { +; CHECK-NEXT: [[A_ZEXT:%.*]] = zext i32 [[A:%.*]] to i64 +; CHECK-NEXT: [[TMP1:%.*]] = trunc i64 [[A_ZEXT]] to i32 +; CHECK-NEXT: [[TMP2:%.*]] = udiv i32 [[TMP1]], 50 +; CHECK-NEXT: [[TMP3:%.*]] = zext i32 [[TMP2]] to i64 +; CHECK-NEXT: ret i64 [[TMP3]] + + %a.zext = zext i32 %a to i64 + %wide.div = udiv i64 %a.zext, 50 + ret i64 %wide.div +} + +; CHECK-LABEL: @urem_by_constant( +define i64 @urem_by_constant(i32 %a) { +; CHECK-NEXT: [[A_ZEXT:%.*]] = zext i32 [[A:%.*]] to i64 +; CHECK-NEXT: [[TMP1:%.*]] = trunc i64 [[A_ZEXT]] to i32 +; CHECK-NEXT: [[TMP2:%.*]] = urem i32 [[TMP1]], 50 +; CHECK-NEXT: [[TMP3:%.*]] = zext i32 [[TMP2]] to i64 +; CHECK-NEXT: ret i64 [[TMP3]] + + %a.zext = zext i32 %a to i64 + %wide.div = urem i64 %a.zext, 50 + ret i64 %wide.div +} + +; Negative test: instead of emitting a runtime check on %a, we prefer to let the +; DAGCombiner transform this division by constant into a multiplication (with a +; "magic constant"). +; +; CHECK-LABEL: @udiv_by_constant_negative_0( +define i64 @udiv_by_constant_negative_0(i64 %a) { +; CHECK-NEXT: [[WIDE_DIV:%.*]] = udiv i64 [[A:%.*]], 50 +; CHECK-NEXT: ret i64 [[WIDE_DIV]] + + %wide.div = udiv i64 %a, 50 + ret i64 %wide.div +} + +; Negative test: while we know the dividend is short, the divisor isn't. This +; test is here for completeness, but instcombine will optimize this to return 0. +; +; CHECK-LABEL: @udiv_by_constant_negative_1( +define i64 @udiv_by_constant_negative_1(i32 %a) { +; CHECK-NEXT: [[A_ZEXT:%.*]] = zext i32 [[A:%.*]] to i64 +; CHECK-NEXT: [[WIDE_DIV:%.*]] = udiv i64 [[A_ZEXT]], 8589934592 +; CHECK-NEXT: ret i64 [[WIDE_DIV]] + + %a.zext = zext i32 %a to i64 + %wide.div = udiv i64 %a.zext, 8589934592 ;; == 1 << 33 + ret i64 %wide.div +} + +; URem version of udiv_by_constant_negative_0 +; +; CHECK-LABEL: @urem_by_constant_negative_0( +define i64 @urem_by_constant_negative_0(i64 %a) { +; CHECK-NEXT: [[WIDE_DIV:%.*]] = urem i64 [[A:%.*]], 50 +; CHECK-NEXT: ret i64 [[WIDE_DIV]] + + %wide.div = urem i64 %a, 50 + ret i64 %wide.div +} + +; URem version of udiv_by_constant_negative_1 +; +; CHECK-LABEL: @urem_by_constant_negative_1( +define i64 @urem_by_constant_negative_1(i32 %a) { +; CHECK-NEXT: [[A_ZEXT:%.*]] = zext i32 [[A:%.*]] to i64 +; CHECK-NEXT: [[WIDE_DIV:%.*]] = urem i64 [[A_ZEXT]], 8589934592 +; CHECK-NEXT: ret i64 [[WIDE_DIV]] + + %a.zext = zext i32 %a to i64 + %wide.div = urem i64 %a.zext, 8589934592 ;; == 1 << 33 + ret i64 %wide.div +} |