summaryrefslogtreecommitdiffstats
path: root/llvm/test/CodeGen/AMDGPU/dpp_combine.ll
diff options
context:
space:
mode:
Diffstat (limited to 'llvm/test/CodeGen/AMDGPU/dpp_combine.ll')
-rw-r--r--llvm/test/CodeGen/AMDGPU/dpp_combine.ll185
1 files changed, 185 insertions, 0 deletions
diff --git a/llvm/test/CodeGen/AMDGPU/dpp_combine.ll b/llvm/test/CodeGen/AMDGPU/dpp_combine.ll
new file mode 100644
index 00000000000..36356a72a77
--- /dev/null
+++ b/llvm/test/CodeGen/AMDGPU/dpp_combine.ll
@@ -0,0 +1,185 @@
+; RUN: llc -march=amdgcn -mcpu=tonga -amdgpu-dpp-combine -verify-machineinstrs < %s | FileCheck %s
+
+; VOP2 with literal cannot be combined
+; CHECK-LABEL: {{^}}dpp_combine_i32_literal:
+; CHECK: v_mov_b32_dpp [[OLD:v[0-9]+]], {{v[0-9]+}} quad_perm:[1,0,0,0] row_mask:0x2 bank_mask:0x1 bound_ctrl:0
+; CHECK: v_add_u32_e32 {{v[0-9]+}}, vcc, 42, [[OLD]]
+define amdgpu_kernel void @dpp_combine_i32_literal(i32 addrspace(1)* %out, i32 %in) {
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 1, i32 2, i32 1, i1 1) #0
+ %res = add nsw i32 %dpp, 42
+ store i32 %res, i32 addrspace(1)* %out
+ ret void
+}
+
+; CHECK-LABEL: {{^}}dpp_combine_i32_bz:
+; CHECK: v_add_u32_dpp {{v[0-9]+}}, vcc, {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0
+define amdgpu_kernel void @dpp_combine_i32_bz(i32 addrspace(1)* %out, i32 %in) {
+ %x = tail call i32 @llvm.amdgcn.workitem.id.x()
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 1, i32 1, i32 1, i1 1) #0
+ %res = add nsw i32 %dpp, %x
+ store i32 %res, i32 addrspace(1)* %out
+ ret void
+}
+
+; CHECK-LABEL: {{^}}dpp_combine_i32_boff_undef:
+; CHECK: v_add_u32_dpp {{v[0-9]+}}, vcc, {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1
+define amdgpu_kernel void @dpp_combine_i32_boff_undef(i32 addrspace(1)* %out, i32 %in) {
+ %x = tail call i32 @llvm.amdgcn.workitem.id.x()
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 1, i32 1, i32 1, i1 0) #0
+ %res = add nsw i32 %dpp, %x
+ store i32 %res, i32 addrspace(1)* %out
+ ret void
+}
+
+; CHECK-LABEL: {{^}}dpp_combine_i32_boff_0:
+; CHECK: v_add_u32_dpp {{v[0-9]+}}, vcc, {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0
+define amdgpu_kernel void @dpp_combine_i32_boff_0(i32 addrspace(1)* %out, i32 %in) {
+ %x = tail call i32 @llvm.amdgcn.workitem.id.x()
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 0, i32 %in, i32 1, i32 1, i32 1, i1 0) #0
+ %res = add nsw i32 %dpp, %x
+ store i32 %res, i32 addrspace(1)* %out
+ ret void
+}
+
+; CHECK-LABEL: {{^}}dpp_combine_i32_boff_max:
+; CHECK: v_bfrev_b32_e32 [[OLD:v[0-9]+]], -2
+; CHECK: v_max_i32_dpp [[OLD]], {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1
+define amdgpu_kernel void @dpp_combine_i32_boff_max(i32 addrspace(1)* %out, i32 %in) {
+ %x = tail call i32 @llvm.amdgcn.workitem.id.x()
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 2147483647, i32 %in, i32 1, i32 1, i32 1, i1 0) #0
+ %cmp = icmp sge i32 %dpp, %x
+ %res = select i1 %cmp, i32 %dpp, i32 %x
+ store i32 %res, i32 addrspace(1)* %out
+ ret void
+}
+
+; CHECK-LABEL: {{^}}dpp_combine_i32_boff_min:
+; CHECK: v_bfrev_b32_e32 [[OLD:v[0-9]+]], 1
+; CHECK: v_min_i32_dpp [[OLD]], {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1
+define amdgpu_kernel void @dpp_combine_i32_boff_min(i32 addrspace(1)* %out, i32 %in) {
+ %x = tail call i32 @llvm.amdgcn.workitem.id.x()
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 -2147483648, i32 %in, i32 1, i32 1, i32 1, i1 0) #0
+ %cmp = icmp sle i32 %dpp, %x
+ %res = select i1 %cmp, i32 %dpp, i32 %x
+ store i32 %res, i32 addrspace(1)* %out
+ ret void
+}
+
+; CHECK-LABEL: {{^}}dpp_combine_i32_boff_mul:
+; CHECK: v_mul_i32_i24_dpp v0, v3, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1
+define amdgpu_kernel void @dpp_combine_i32_boff_mul(i32 addrspace(1)* %out, i32 %in) {
+ %x = tail call i32 @llvm.amdgcn.workitem.id.x()
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 1, i32 %in, i32 1, i32 1, i32 1, i1 0) #0
+
+ %dpp.shl = shl i32 %dpp, 8
+ %dpp.24 = ashr i32 %dpp.shl, 8
+ %x.shl = shl i32 %x, 8
+ %x.24 = ashr i32 %x.shl, 8
+ %res = mul i32 %dpp.24, %x.24
+ store i32 %res, i32 addrspace(1)* %out
+ ret void
+}
+
+; CHECK-LABEL: {{^}}dpp_combine_i32_commute:
+; CHECK: v_subrev_u32_dpp {{v[0-9]+}}, vcc, {{v[0-9]+}}, v0 quad_perm:[2,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0
+define amdgpu_kernel void @dpp_combine_i32_commute(i32 addrspace(1)* %out, i32 %in) {
+ %x = tail call i32 @llvm.amdgcn.workitem.id.x()
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 2, i32 1, i32 1, i1 1) #0
+ %res = sub nsw i32 %x, %dpp
+ store i32 %res, i32 addrspace(1)* %out
+ ret void
+}
+
+; CHECK-LABEL: {{^}}dpp_combine_f32:
+; CHECK: v_add_f32_dpp {{v[0-9]+}}, {{v[0-9]+}}, v0 quad_perm:[3,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0
+define amdgpu_kernel void @dpp_combine_f32(i32 addrspace(1)* %out, i32 %in) {
+ %x = tail call i32 @llvm.amdgcn.workitem.id.x()
+
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 3, i32 1, i32 1, i1 1) #0
+ %dpp.f32 = bitcast i32 %dpp to float
+ %x.f32 = bitcast i32 %x to float
+ %res.f32 = fadd float %x.f32, %dpp.f32
+ %res = bitcast float %res.f32 to i32
+ store i32 %res, i32 addrspace(1)* %out
+ ret void
+}
+
+; CHECK-LABEL: {{^}}dpp_combine_test_f32_mods:
+; CHECK: v_mul_f32_dpp {{v[0-9]+}}, |{{v[0-9]+}}|, -v0 quad_perm:[0,1,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0
+define amdgpu_kernel void @dpp_combine_test_f32_mods(i32 addrspace(1)* %out, i32 %in) {
+ %x = tail call i32 @llvm.amdgcn.workitem.id.x()
+
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 4, i32 1, i32 1, i1 1) #0
+
+ %x.f32 = bitcast i32 %x to float
+ %x.f32.neg = fsub float -0.000000e+00, %x.f32
+
+ %dpp.f32 = bitcast i32 %dpp to float
+ %dpp.f32.cmp = fcmp fast olt float %dpp.f32, 0.000000e+00
+ %dpp.f32.sign = select i1 %dpp.f32.cmp, float -1.000000e+00, float 1.000000e+00
+ %dpp.f32.abs = fmul fast float %dpp.f32, %dpp.f32.sign
+
+ %res.f32 = fmul float %x.f32.neg, %dpp.f32.abs
+ %res = bitcast float %res.f32 to i32
+ store i32 %res, i32 addrspace(1)* %out
+ ret void
+}
+
+; CHECK-LABEL: {{^}}dpp_combine_mac:
+; CHECK: v_mac_f32_dpp v0, {{v[0-9]+}}, v1 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0
+define amdgpu_kernel void @dpp_combine_mac(float addrspace(1)* %out, i32 %in) {
+ %x = tail call i32 @llvm.amdgcn.workitem.id.x()
+ %y = tail call i32 @llvm.amdgcn.workitem.id.y()
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 1, i32 1, i32 1, i1 1) #0
+ %dpp.f32 = bitcast i32 %dpp to float
+ %x.f32 = bitcast i32 %x to float
+ %y.f32 = bitcast i32 %y to float
+
+ %mult = fmul float %dpp.f32, %y.f32
+ %res = fadd float %mult, %x.f32
+ store float %res, float addrspace(1)* %out
+ ret void
+}
+
+; CHECK-LABEL: {{^}}dpp_combine_sequence:
+define amdgpu_kernel void @dpp_combine_sequence(i32 addrspace(1)* %out, i32 %in, i1 %cmp) {
+ %x = tail call i32 @llvm.amdgcn.workitem.id.x()
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 1, i32 1, i32 1, i1 1) #0
+ br i1 %cmp, label %bb1, label %bb2
+bb1:
+; CHECK: v_add_u32_dpp {{v[0-9]+}}, vcc, {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0
+ %resadd = add nsw i32 %dpp, %x
+ br label %bb3
+bb2:
+; CHECK: v_subrev_u32_dpp {{v[0-9]+}}, vcc, {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0
+ %ressub = sub nsw i32 %x, %dpp
+ br label %bb3
+bb3:
+ %res = phi i32 [%resadd, %bb1], [%ressub, %bb2]
+ store i32 %res, i32 addrspace(1)* %out
+ ret void
+}
+
+; CHECK-LABEL: {{^}}dpp_combine_sequence_negative:
+; CHECK: v_mov_b32_dpp v1, v1 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0
+define amdgpu_kernel void @dpp_combine_sequence_negative(i32 addrspace(1)* %out, i32 %in, i1 %cmp) {
+ %x = tail call i32 @llvm.amdgcn.workitem.id.x()
+ %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 1, i32 1, i32 1, i1 1) #0
+ br i1 %cmp, label %bb1, label %bb2
+bb1:
+ %resadd = add nsw i32 %dpp, %x
+ br label %bb3
+bb2:
+ %ressub = sub nsw i32 2, %dpp ; break seq
+ br label %bb3
+bb3:
+ %res = phi i32 [%resadd, %bb1], [%ressub, %bb2]
+ store i32 %res, i32 addrspace(1)* %out
+ ret void
+}
+
+declare i32 @llvm.amdgcn.workitem.id.x()
+declare i32 @llvm.amdgcn.workitem.id.y()
+declare i32 @llvm.amdgcn.update.dpp.i32(i32, i32, i32, i32, i32, i1) #0
+
+attributes #0 = { nounwind readnone convergent }
OpenPOWER on IntegriCloud