diff options
| author | Valery Pykhtin <Valery.Pykhtin@amd.com> | 2019-01-09 15:21:53 +0000 |
|---|---|---|
| committer | Valery Pykhtin <Valery.Pykhtin@amd.com> | 2019-01-09 15:21:53 +0000 |
| commit | b7a459547d41ff87e7fed9143901531608e2b321 (patch) | |
| tree | efb7022230b57a7ac25c27674c0dbd3f0408c144 /llvm/test/CodeGen/AMDGPU | |
| parent | c650ff77eb1055d1529b9eeab13214388322ad8f (diff) | |
| download | bcm5719-llvm-b7a459547d41ff87e7fed9143901531608e2b321.tar.gz bcm5719-llvm-b7a459547d41ff87e7fed9143901531608e2b321.zip | |
Revert "[AMDGPU] Fix DPP combiner"
This reverts commit e3e2923a39cbec3b3bc3a7d3f0e9a77a4115080e, svn revision rL350721
llvm-svn: 350730
Diffstat (limited to 'llvm/test/CodeGen/AMDGPU')
| -rw-r--r-- | llvm/test/CodeGen/AMDGPU/dpp_combine.ll | 185 | ||||
| -rw-r--r-- | llvm/test/CodeGen/AMDGPU/dpp_combine.mir | 525 | ||||
| -rw-r--r-- | llvm/test/CodeGen/AMDGPU/dpp_combine_subregs.mir | 143 |
3 files changed, 328 insertions, 525 deletions
diff --git a/llvm/test/CodeGen/AMDGPU/dpp_combine.ll b/llvm/test/CodeGen/AMDGPU/dpp_combine.ll new file mode 100644 index 00000000000..36356a72a77 --- /dev/null +++ b/llvm/test/CodeGen/AMDGPU/dpp_combine.ll @@ -0,0 +1,185 @@ +; RUN: llc -march=amdgcn -mcpu=tonga -amdgpu-dpp-combine -verify-machineinstrs < %s | FileCheck %s + +; VOP2 with literal cannot be combined +; CHECK-LABEL: {{^}}dpp_combine_i32_literal: +; CHECK: v_mov_b32_dpp [[OLD:v[0-9]+]], {{v[0-9]+}} quad_perm:[1,0,0,0] row_mask:0x2 bank_mask:0x1 bound_ctrl:0 +; CHECK: v_add_u32_e32 {{v[0-9]+}}, vcc, 42, [[OLD]] +define amdgpu_kernel void @dpp_combine_i32_literal(i32 addrspace(1)* %out, i32 %in) { + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 1, i32 2, i32 1, i1 1) #0 + %res = add nsw i32 %dpp, 42 + store i32 %res, i32 addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}dpp_combine_i32_bz: +; CHECK: v_add_u32_dpp {{v[0-9]+}}, vcc, {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0 +define amdgpu_kernel void @dpp_combine_i32_bz(i32 addrspace(1)* %out, i32 %in) { + %x = tail call i32 @llvm.amdgcn.workitem.id.x() + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 1, i32 1, i32 1, i1 1) #0 + %res = add nsw i32 %dpp, %x + store i32 %res, i32 addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}dpp_combine_i32_boff_undef: +; CHECK: v_add_u32_dpp {{v[0-9]+}}, vcc, {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 +define amdgpu_kernel void @dpp_combine_i32_boff_undef(i32 addrspace(1)* %out, i32 %in) { + %x = tail call i32 @llvm.amdgcn.workitem.id.x() + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 1, i32 1, i32 1, i1 0) #0 + %res = add nsw i32 %dpp, %x + store i32 %res, i32 addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}dpp_combine_i32_boff_0: +; CHECK: v_add_u32_dpp {{v[0-9]+}}, vcc, {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0 +define amdgpu_kernel void @dpp_combine_i32_boff_0(i32 addrspace(1)* %out, i32 %in) { + %x = tail call i32 @llvm.amdgcn.workitem.id.x() + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 0, i32 %in, i32 1, i32 1, i32 1, i1 0) #0 + %res = add nsw i32 %dpp, %x + store i32 %res, i32 addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}dpp_combine_i32_boff_max: +; CHECK: v_bfrev_b32_e32 [[OLD:v[0-9]+]], -2 +; CHECK: v_max_i32_dpp [[OLD]], {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 +define amdgpu_kernel void @dpp_combine_i32_boff_max(i32 addrspace(1)* %out, i32 %in) { + %x = tail call i32 @llvm.amdgcn.workitem.id.x() + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 2147483647, i32 %in, i32 1, i32 1, i32 1, i1 0) #0 + %cmp = icmp sge i32 %dpp, %x + %res = select i1 %cmp, i32 %dpp, i32 %x + store i32 %res, i32 addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}dpp_combine_i32_boff_min: +; CHECK: v_bfrev_b32_e32 [[OLD:v[0-9]+]], 1 +; CHECK: v_min_i32_dpp [[OLD]], {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 +define amdgpu_kernel void @dpp_combine_i32_boff_min(i32 addrspace(1)* %out, i32 %in) { + %x = tail call i32 @llvm.amdgcn.workitem.id.x() + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 -2147483648, i32 %in, i32 1, i32 1, i32 1, i1 0) #0 + %cmp = icmp sle i32 %dpp, %x + %res = select i1 %cmp, i32 %dpp, i32 %x + store i32 %res, i32 addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}dpp_combine_i32_boff_mul: +; CHECK: v_mul_i32_i24_dpp v0, v3, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 +define amdgpu_kernel void @dpp_combine_i32_boff_mul(i32 addrspace(1)* %out, i32 %in) { + %x = tail call i32 @llvm.amdgcn.workitem.id.x() + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 1, i32 %in, i32 1, i32 1, i32 1, i1 0) #0 + + %dpp.shl = shl i32 %dpp, 8 + %dpp.24 = ashr i32 %dpp.shl, 8 + %x.shl = shl i32 %x, 8 + %x.24 = ashr i32 %x.shl, 8 + %res = mul i32 %dpp.24, %x.24 + store i32 %res, i32 addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}dpp_combine_i32_commute: +; CHECK: v_subrev_u32_dpp {{v[0-9]+}}, vcc, {{v[0-9]+}}, v0 quad_perm:[2,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0 +define amdgpu_kernel void @dpp_combine_i32_commute(i32 addrspace(1)* %out, i32 %in) { + %x = tail call i32 @llvm.amdgcn.workitem.id.x() + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 2, i32 1, i32 1, i1 1) #0 + %res = sub nsw i32 %x, %dpp + store i32 %res, i32 addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}dpp_combine_f32: +; CHECK: v_add_f32_dpp {{v[0-9]+}}, {{v[0-9]+}}, v0 quad_perm:[3,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0 +define amdgpu_kernel void @dpp_combine_f32(i32 addrspace(1)* %out, i32 %in) { + %x = tail call i32 @llvm.amdgcn.workitem.id.x() + + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 3, i32 1, i32 1, i1 1) #0 + %dpp.f32 = bitcast i32 %dpp to float + %x.f32 = bitcast i32 %x to float + %res.f32 = fadd float %x.f32, %dpp.f32 + %res = bitcast float %res.f32 to i32 + store i32 %res, i32 addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}dpp_combine_test_f32_mods: +; CHECK: v_mul_f32_dpp {{v[0-9]+}}, |{{v[0-9]+}}|, -v0 quad_perm:[0,1,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0 +define amdgpu_kernel void @dpp_combine_test_f32_mods(i32 addrspace(1)* %out, i32 %in) { + %x = tail call i32 @llvm.amdgcn.workitem.id.x() + + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 4, i32 1, i32 1, i1 1) #0 + + %x.f32 = bitcast i32 %x to float + %x.f32.neg = fsub float -0.000000e+00, %x.f32 + + %dpp.f32 = bitcast i32 %dpp to float + %dpp.f32.cmp = fcmp fast olt float %dpp.f32, 0.000000e+00 + %dpp.f32.sign = select i1 %dpp.f32.cmp, float -1.000000e+00, float 1.000000e+00 + %dpp.f32.abs = fmul fast float %dpp.f32, %dpp.f32.sign + + %res.f32 = fmul float %x.f32.neg, %dpp.f32.abs + %res = bitcast float %res.f32 to i32 + store i32 %res, i32 addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}dpp_combine_mac: +; CHECK: v_mac_f32_dpp v0, {{v[0-9]+}}, v1 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0 +define amdgpu_kernel void @dpp_combine_mac(float addrspace(1)* %out, i32 %in) { + %x = tail call i32 @llvm.amdgcn.workitem.id.x() + %y = tail call i32 @llvm.amdgcn.workitem.id.y() + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 1, i32 1, i32 1, i1 1) #0 + %dpp.f32 = bitcast i32 %dpp to float + %x.f32 = bitcast i32 %x to float + %y.f32 = bitcast i32 %y to float + + %mult = fmul float %dpp.f32, %y.f32 + %res = fadd float %mult, %x.f32 + store float %res, float addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}dpp_combine_sequence: +define amdgpu_kernel void @dpp_combine_sequence(i32 addrspace(1)* %out, i32 %in, i1 %cmp) { + %x = tail call i32 @llvm.amdgcn.workitem.id.x() + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 1, i32 1, i32 1, i1 1) #0 + br i1 %cmp, label %bb1, label %bb2 +bb1: +; CHECK: v_add_u32_dpp {{v[0-9]+}}, vcc, {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0 + %resadd = add nsw i32 %dpp, %x + br label %bb3 +bb2: +; CHECK: v_subrev_u32_dpp {{v[0-9]+}}, vcc, {{v[0-9]+}}, v0 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0 + %ressub = sub nsw i32 %x, %dpp + br label %bb3 +bb3: + %res = phi i32 [%resadd, %bb1], [%ressub, %bb2] + store i32 %res, i32 addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}dpp_combine_sequence_negative: +; CHECK: v_mov_b32_dpp v1, v1 quad_perm:[1,0,0,0] row_mask:0x1 bank_mask:0x1 bound_ctrl:0 +define amdgpu_kernel void @dpp_combine_sequence_negative(i32 addrspace(1)* %out, i32 %in, i1 %cmp) { + %x = tail call i32 @llvm.amdgcn.workitem.id.x() + %dpp = call i32 @llvm.amdgcn.update.dpp.i32(i32 undef, i32 %in, i32 1, i32 1, i32 1, i1 1) #0 + br i1 %cmp, label %bb1, label %bb2 +bb1: + %resadd = add nsw i32 %dpp, %x + br label %bb3 +bb2: + %ressub = sub nsw i32 2, %dpp ; break seq + br label %bb3 +bb3: + %res = phi i32 [%resadd, %bb1], [%ressub, %bb2] + store i32 %res, i32 addrspace(1)* %out + ret void +} + +declare i32 @llvm.amdgcn.workitem.id.x() +declare i32 @llvm.amdgcn.workitem.id.y() +declare i32 @llvm.amdgcn.update.dpp.i32(i32, i32, i32, i32, i32, i1) #0 + +attributes #0 = { nounwind readnone convergent } diff --git a/llvm/test/CodeGen/AMDGPU/dpp_combine.mir b/llvm/test/CodeGen/AMDGPU/dpp_combine.mir deleted file mode 100644 index def3725d544..00000000000 --- a/llvm/test/CodeGen/AMDGPU/dpp_combine.mir +++ /dev/null @@ -1,525 +0,0 @@ -# RUN: llc -march=amdgcn -mcpu=tonga -run-pass=gcn-dpp-combine -o - %s | FileCheck %s - ---- -# old is undefined: only combine when masks are fully enabled and -# bound_ctrl:0 is set, otherwise the result of DPP VALU op can be undefined. -# CHECK-LABEL: name: old_is_undef -# CHECK: %2:vgpr_32 = IMPLICIT_DEF -# VOP2: -# CHECK: %4:vgpr_32 = V_ADD_U32_dpp %2, %0, %1, 1, 15, 15, 1, implicit $exec -# CHECK: %6:vgpr_32 = V_ADD_U32_e32 %5, %1, implicit $exec -# CHECK: %8:vgpr_32 = V_ADD_U32_e32 %7, %1, implicit $exec -# CHECK: %10:vgpr_32 = V_ADD_U32_e32 %9, %1, implicit $exec -# VOP1: -# CHECK: %12:vgpr_32 = V_NOT_B32_dpp %2, %0, 1, 15, 15, 1, implicit $exec -# CHECK: %14:vgpr_32 = V_NOT_B32_e32 %13, implicit $exec -# CHECK: %16:vgpr_32 = V_NOT_B32_e32 %15, implicit $exec -# CHECK: %18:vgpr_32 = V_NOT_B32_e32 %17, implicit $exec -name: old_is_undef -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - %0:vgpr_32 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - %2:vgpr_32 = IMPLICIT_DEF - - ; VOP2 - %3:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 15, 15, 1, implicit $exec - %4:vgpr_32 = V_ADD_U32_e32 %3, %1, implicit $exec - - %5:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 15, 15, 0, implicit $exec - %6:vgpr_32 = V_ADD_U32_e32 %5, %1, implicit $exec - - %7:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 1, implicit $exec - %8:vgpr_32 = V_ADD_U32_e32 %7, %1, implicit $exec - - %9:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 0, implicit $exec - %10:vgpr_32 = V_ADD_U32_e32 %9, %1, implicit $exec - - ; VOP1 - %11:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 15, 15, 1, implicit $exec - %12:vgpr_32 = V_NOT_B32_e32 %11, implicit $exec - - %13:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 15, 15, 0, implicit $exec - %14:vgpr_32 = V_NOT_B32_e32 %13, implicit $exec - - %15:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 1, implicit $exec - %16:vgpr_32 = V_NOT_B32_e32 %15, implicit $exec - - %17:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 0, implicit $exec - %18:vgpr_32 = V_NOT_B32_e32 %17, implicit $exec -... - -# old is the same as DPP src register: only combine when masks are fully enabled -# othervise the result of DPP VALU can be undefined as its old is undef -# CHECK-LABEL: name: old_is_same_as_src -# VOP2: -# CHECK: [[U1:%[0-9]+]]:vgpr_32 = IMPLICIT_DEF -# CHECK: %3:vgpr_32 = V_ADD_U32_dpp [[U1]], %0, %1, 1, 15, 15, 1, implicit $exec -# CHECK: [[U2:%[0-9]+]]:vgpr_32 = IMPLICIT_DEF -# CHECK: %5:vgpr_32 = V_ADD_U32_dpp [[U2]], %0, %1, 1, 15, 15, 0, implicit $exec -# CHECK: %7:vgpr_32 = V_ADD_U32_e32 %6, %1, implicit $exec -# CHECK: %9:vgpr_32 = V_ADD_U32_e32 %8, %1, implicit $exec -# VOP1: -# CHECK: [[U3:%[0-9]+]]:vgpr_32 = IMPLICIT_DEF -# CHECK: %11:vgpr_32 = V_NOT_B32_dpp [[U3]], %0, 1, 15, 15, 1, implicit $exec -# CHECK: [[U4:%[0-9]+]]:vgpr_32 = IMPLICIT_DEF -# CHECK: %13:vgpr_32 = V_NOT_B32_dpp [[U4]], %0, 1, 15, 15, 0, implicit $exec -# CHECK: %15:vgpr_32 = V_NOT_B32_e32 %14, implicit $exec -# CHECK: %17:vgpr_32 = V_NOT_B32_e32 %16, implicit $exec - -name: old_is_same_as_src -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - %0:vgpr_32 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - - ; VOP2 - %2:vgpr_32 = V_MOV_B32_dpp %0, %0, 1, 15, 15, 1, implicit $exec - %3:vgpr_32 = V_ADD_U32_e32 %2, %1, implicit $exec - - %4:vgpr_32 = V_MOV_B32_dpp %0, %0, 1, 15, 15, 0, implicit $exec - %5:vgpr_32 = V_ADD_U32_e32 %4, %1, implicit $exec - - %6:vgpr_32 = V_MOV_B32_dpp %0, %0, 1, 14, 15, 1, implicit $exec - %7:vgpr_32 = V_ADD_U32_e32 %6, %1, implicit $exec - - %8:vgpr_32 = V_MOV_B32_dpp %0, %0, 1, 14, 15, 0, implicit $exec - %9:vgpr_32 = V_ADD_U32_e32 %8, %1, implicit $exec - - ; VOP1 - %10:vgpr_32 = V_MOV_B32_dpp %0, %0, 1, 15, 15, 1, implicit $exec - %11:vgpr_32 = V_NOT_B32_e32 %10, implicit $exec - - %12:vgpr_32 = V_MOV_B32_dpp %0, %0, 1, 15, 15, 0, implicit $exec - %13:vgpr_32 = V_NOT_B32_e32 %12, implicit $exec - - %14:vgpr_32 = V_MOV_B32_dpp %0, %0, 1, 14, 15, 1, implicit $exec - %15:vgpr_32 = V_NOT_B32_e32 %14, implicit $exec - - %16:vgpr_32 = V_MOV_B32_dpp %0, %0, 1, 14, 15, 0, implicit $exec - %17:vgpr_32 = V_NOT_B32_e32 %16, implicit $exec -... - -# old is zero cases: - -# CHECK-LABEL: name: old_is_0 - -# VOP2: -# case 1: old is zero, masks are fully enabled, bound_ctrl:0 is on: -# the DPP mov result would be either zero ({src lane disabled}|{src lane is -# out of range}) or active src lane result - can combine with old = undef. -# undef is preffered as it makes life easier for the regalloc. -# CHECK: [[U1:%[0-9]+]]:vgpr_32 = IMPLICIT_DEF -# CHECK: %4:vgpr_32 = V_ADD_U32_dpp [[U1]], %0, %1, 1, 15, 15, 1, implicit $exec - -# case 2: old is zero, masks are fully enabled, bound_ctrl:0 is off: -# as the DPP mov old is zero this case is no different from case 1 - combine it -# setting bound_ctrl0 on for the combined DPP VALU op to make old undefined -# CHECK: [[U2:%[0-9]+]]:vgpr_32 = IMPLICIT_DEF -# CHECK: %6:vgpr_32 = V_ADD_U32_dpp [[U2]], %0, %1, 1, 15, 15, 1, implicit $exec - -# case 3: masks are partialy disabled, bound_ctrl:0 is on: -# the DPP mov result would be either zero ({src lane disabled}|{src lane is -# out of range} or {the DPP mov's dest VGPR write is disabled by masks}) or -# active src lane result - can combine with old = src1 of the VALU op. -# The VALU op should have the same masks as DPP mov as they select lanes -# with identity value. -# Special case: the bound_ctrl for the combined DPP VALU op isn't important -# here but let's make it off to keep the combiner's logic simpler. -# CHECK: %8:vgpr_32 = V_ADD_U32_dpp %1, %0, %1, 1, 14, 15, 0, implicit $exec - -# case 4: masks are partialy disabled, bound_ctrl:0 is off: -# the DPP mov result would be either zero ({src lane disabled}|{src lane is -# out of range} or {the DPP mov's dest VGPR write is disabled by masks}) or -# active src lane result - can combine with old = src1 of the VALU op. -# The VALU op should have the same masks as DPP mov as they select -# lanes with identity value -# CHECK: %10:vgpr_32 = V_ADD_U32_dpp %1, %0, %1, 1, 14, 15, 0, implicit $exec - -# VOP1: -# see case 1 -# CHECK: [[U3:%[0-9]+]]:vgpr_32 = IMPLICIT_DEF -# CHECK: %12:vgpr_32 = V_NOT_B32_dpp [[U3]], %0, 1, 15, 15, 1, implicit $exec -# see case 2 -# CHECK: [[U4:%[0-9]+]]:vgpr_32 = IMPLICIT_DEF -# CHECK: %14:vgpr_32 = V_NOT_B32_dpp [[U4]], %0, 1, 15, 15, 1, implicit $exec -# case 3 and 4 not appliable as there is no way to specify unchanged result -# for the unary VALU op -# CHECK: %16:vgpr_32 = V_NOT_B32_e32 %15, implicit $exec -# CHECK: %18:vgpr_32 = V_NOT_B32_e32 %17, implicit $exec - -name: old_is_0 -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - %0:vgpr_32 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - %2:vgpr_32 = V_MOV_B32_e32 0, implicit $exec - - ; VOP2 - %3:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 15, 15, 1, implicit $exec - %4:vgpr_32 = V_ADD_U32_e32 %3, %1, implicit $exec - - %5:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 15, 15, 0, implicit $exec - %6:vgpr_32 = V_ADD_U32_e32 %5, %1, implicit $exec - - %7:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 1, implicit $exec - %8:vgpr_32 = V_ADD_U32_e32 %7, %1, implicit $exec - - %9:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 0, implicit $exec - %10:vgpr_32 = V_ADD_U32_e32 %9, %1, implicit $exec - - ; VOP1 - %11:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 15, 15, 1, implicit $exec - %12:vgpr_32 = V_NOT_B32_e32 %11, implicit $exec - - %13:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 15, 15, 0, implicit $exec - %14:vgpr_32 = V_NOT_B32_e32 %13, implicit $exec - - %15:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 1, implicit $exec - %16:vgpr_32 = V_NOT_B32_e32 %15, implicit $exec - - %17:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 0, implicit $exec - %18:vgpr_32 = V_NOT_B32_e32 %17, implicit $exec -... - -# old is nonzero identity cases: - -# old is nonzero identity, masks are fully enabled, bound_ctrl:0 is off: -# the DPP mov result would be either identity ({src lane disabled}|{out of -# range}) or src lane result - can combine with old = src1 of the VALU op -# The DPP VALU op should have the same masks (and bctrl) as DPP mov as they -# select lanes with identity value - -# CHECK-LABEL: name: nonzero_old_is_identity_masks_enabled_bctl_off -# CHECK: %4:vgpr_32 = V_MUL_U32_U24_dpp %1, %0, %1, 1, 15, 15, 0, implicit $exec -# CHECK: %7:vgpr_32 = V_AND_B32_dpp %1, %0, %1, 1, 15, 15, 0, implicit $exec -# CHECK: %10:vgpr_32 = V_MAX_I32_dpp %1, %0, %1, 1, 15, 15, 0, implicit $exec -# CHECK: %13:vgpr_32 = V_MIN_I32_dpp %1, %0, %1, 1, 15, 15, 0, implicit $exec - -name: nonzero_old_is_identity_masks_enabled_bctl_off -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - %0:vgpr_32 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - - %2:vgpr_32 = V_MOV_B32_e32 1, implicit $exec - %3:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 15, 15, 0, implicit $exec - %4:vgpr_32 = V_MUL_U32_U24_e32 %3, %1, implicit $exec - - %5:vgpr_32 = V_MOV_B32_e32 4294967295, implicit $exec - %6:vgpr_32 = V_MOV_B32_dpp %5, %0, 1, 15, 15, 0, implicit $exec - %7:vgpr_32 = V_AND_B32_e32 %6, %1, implicit $exec - - %8:vgpr_32 = V_MOV_B32_e32 2147483647, implicit $exec - %9:vgpr_32 = V_MOV_B32_dpp %8, %0, 1, 15, 15, 0, implicit $exec - %10:vgpr_32 = V_MAX_I32_e32 %9, %1, implicit $exec - - %11:vgpr_32 = V_MOV_B32_e32 -2147483648, implicit $exec - %12:vgpr_32 = V_MOV_B32_dpp %11, %0, 1, 15, 15, 0, implicit $exec - %13:vgpr_32 = V_MIN_I32_e32 %12, %1, implicit $exec -... - -# old is nonzero identity, masks are partially enabled, bound_ctrl:0 is off: -# the DPP mov result would be either identity ({src lane disabled}|{src lane is -# out of range} or {the DPP mov's dest VGPR write is disabled by masks}) or -# active src lane result - can combine with old = src1 of the VALU op. -# The DPP VALU op should have the same masks (and bctrl) as DPP mov as they -# select lanes with identity value - -# CHECK-LABEL: name: nonzero_old_is_identity_masks_partially_disabled_bctl_off -# CHECK: %4:vgpr_32 = V_MUL_U32_U24_dpp %1, %0, %1, 1, 14, 15, 0, implicit $exec -# CHECK: %7:vgpr_32 = V_AND_B32_dpp %1, %0, %1, 1, 15, 14, 0, implicit $exec -# CHECK: %10:vgpr_32 = V_MAX_I32_dpp %1, %0, %1, 1, 14, 15, 0, implicit $exec -# CHECK: %13:vgpr_32 = V_MIN_I32_dpp %1, %0, %1, 1, 15, 14, 0, implicit $exec - -name: nonzero_old_is_identity_masks_partially_disabled_bctl_off -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - %0:vgpr_32 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - - %2:vgpr_32 = V_MOV_B32_e32 1, implicit $exec - %3:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 0, implicit $exec - %4:vgpr_32 = V_MUL_U32_U24_e32 %3, %1, implicit $exec - - %5:vgpr_32 = V_MOV_B32_e32 4294967295, implicit $exec - %6:vgpr_32 = V_MOV_B32_dpp %5, %0, 1, 15, 14, 0, implicit $exec - %7:vgpr_32 = V_AND_B32_e32 %6, %1, implicit $exec - - %8:vgpr_32 = V_MOV_B32_e32 2147483647, implicit $exec - %9:vgpr_32 = V_MOV_B32_dpp %8, %0, 1, 14, 15, 0, implicit $exec - %10:vgpr_32 = V_MAX_I32_e32 %9, %1, implicit $exec - - %11:vgpr_32 = V_MOV_B32_e32 -2147483648, implicit $exec - %12:vgpr_32 = V_MOV_B32_dpp %11, %0, 1, 15, 14, 0, implicit $exec - %13:vgpr_32 = V_MIN_I32_e32 %12, %1, implicit $exec -... - -# old is nonzero identity, masks are partially enabled, bound_ctrl:0 is on: -# the DPP mov result may have 3 different values: -# 1. the active src lane result -# 2. 0 if the src lane is disabled|out of range -# 3. DPP mov's old value if the mov's dest VGPR write is disabled by masks -# can't combine - -# CHECK-LABEL: name: nonzero_old_is_identity_masks_partially_disabled_bctl0 -# CHECK: %4:vgpr_32 = V_MUL_U32_U24_e32 %3, %1, implicit $exec -# CHECK: %7:vgpr_32 = V_AND_B32_e32 %6, %1, implicit $exec -# CHECK: %10:vgpr_32 = V_MAX_I32_e32 %9, %1, implicit $exec -# CHECK: %13:vgpr_32 = V_MIN_I32_e32 %12, %1, implicit $exec - -name: nonzero_old_is_identity_masks_partially_disabled_bctl0 -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - %0:vgpr_32 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - - %2:vgpr_32 = V_MOV_B32_e32 1, implicit $exec - %3:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 1, implicit $exec - %4:vgpr_32 = V_MUL_U32_U24_e32 %3, %1, implicit $exec - - %5:vgpr_32 = V_MOV_B32_e32 4294967295, implicit $exec - %6:vgpr_32 = V_MOV_B32_dpp %5, %0, 1, 15, 14, 1, implicit $exec - %7:vgpr_32 = V_AND_B32_e32 %6, %1, implicit $exec - - %8:vgpr_32 = V_MOV_B32_e32 2147483647, implicit $exec - %9:vgpr_32 = V_MOV_B32_dpp %8, %0, 1, 14, 15, 1, implicit $exec - %10:vgpr_32 = V_MAX_I32_e32 %9, %1, implicit $exec - - %11:vgpr_32 = V_MOV_B32_e32 -2147483648, implicit $exec - %12:vgpr_32 = V_MOV_B32_dpp %11, %0, 1, 15, 14, 1, implicit $exec - %13:vgpr_32 = V_MIN_I32_e32 %12, %1, implicit $exec -... - -# when the DPP source isn't a src0 operand the operation should be commuted if possible -# CHECK-LABEL: name: dpp_commute -# CHECK: %4:vgpr_32 = V_MUL_U32_U24_dpp %1, %0, %1, 1, 14, 15, 0, implicit $exec -# CHECK: %7:vgpr_32 = V_AND_B32_dpp %1, %0, %1, 1, 15, 14, 0, implicit $exec -# CHECK: %10:vgpr_32 = V_MAX_I32_dpp %1, %0, %1, 1, 14, 15, 0, implicit $exec -# CHECK: %13:vgpr_32 = V_MIN_I32_dpp %1, %0, %1, 1, 15, 14, 0, implicit $exec -# CHECK: %16:vgpr_32 = V_SUBREV_I32_dpp %1, %0, %1, 1, 14, 15, 0, implicit-def $vcc, implicit $exec -# CHECK: %19:vgpr_32 = V_ADD_I32_e32 5, %18, implicit-def $vcc, implicit $exec -name: dpp_commute -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - - %0:vgpr_32 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - - %2:vgpr_32 = V_MOV_B32_e32 1, implicit $exec - %3:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 0, implicit $exec - %4:vgpr_32 = V_MUL_U32_U24_e32 %1, %3, implicit $exec - - %5:vgpr_32 = V_MOV_B32_e32 4294967295, implicit $exec - %6:vgpr_32 = V_MOV_B32_dpp %5, %0, 1, 15, 14, 0, implicit $exec - %7:vgpr_32 = V_AND_B32_e32 %1, %6, implicit $exec - - %8:vgpr_32 = V_MOV_B32_e32 2147483647, implicit $exec - %9:vgpr_32 = V_MOV_B32_dpp %8, %0, 1, 14, 15, 0, implicit $exec - %10:vgpr_32 = V_MAX_I32_e32 %1, %9, implicit $exec - - %11:vgpr_32 = V_MOV_B32_e32 -2147483648, implicit $exec - %12:vgpr_32 = V_MOV_B32_dpp %11, %0, 1, 15, 14, 0, implicit $exec - %13:vgpr_32 = V_MIN_I32_e32 %1, %12, implicit $exec - - %14:vgpr_32 = V_MOV_B32_e32 0, implicit $exec - %15:vgpr_32 = V_MOV_B32_dpp %14, %0, 1, 14, 15, 0, implicit $exec - %16:vgpr_32 = V_SUB_I32_e32 %1, %15, implicit-def $vcc, implicit $exec - - ; this cannot be combined because immediate as src0 isn't commutable - %17:vgpr_32 = V_MOV_B32_e32 0, implicit $exec - %18:vgpr_32 = V_MOV_B32_dpp %17, %0, 1, 14, 15, 0, implicit $exec - %19:vgpr_32 = V_ADD_I32_e32 5, %18, implicit-def $vcc, implicit $exec -... - -# check for floating point modifiers -# CHECK-LABEL: name: add_f32_e64 -# CHECK: %3:vgpr_32 = V_MOV_B32_dpp undef %2, %1, 1, 15, 15, 1, implicit $exec -# CHECK: %4:vgpr_32 = V_ADD_F32_e64 0, %3, 0, %0, 0, 1, implicit $exec -# CHECK: %6:vgpr_32 = V_ADD_F32_dpp %2, 0, %1, 0, %0, 1, 15, 15, 1, implicit $exec -# CHECK: %8:vgpr_32 = V_ADD_F32_dpp %2, 1, %1, 2, %0, 1, 15, 15, 1, implicit $exec -# CHECK: %10:vgpr_32 = V_ADD_F32_e64 4, %9, 8, %0, 0, 0, implicit $exec - -name: add_f32_e64 -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - - %0:vgpr_32 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - %2:vgpr_32 = IMPLICIT_DEF - - ; this shouldn't be combined as omod is set - %3:vgpr_32 = V_MOV_B32_dpp undef %2, %1, 1, 15, 15, 1, implicit $exec - %4:vgpr_32 = V_ADD_F32_e64 0, %3, 0, %0, 0, 1, implicit $exec - - ; this should be combined as all modifiers are default - %5:vgpr_32 = V_MOV_B32_dpp undef %2, %1, 1, 15, 15, 1, implicit $exec - %6:vgpr_32 = V_ADD_F32_e64 0, %5, 0, %0, 0, 0, implicit $exec - - ; this should be combined as modifiers other than abs|neg are default - %7:vgpr_32 = V_MOV_B32_dpp undef %2, %1, 1, 15, 15, 1, implicit $exec - %8:vgpr_32 = V_ADD_F32_e64 1, %7, 2, %0, 0, 0, implicit $exec - - ; this shouldn't be combined as modifiers aren't abs|neg - %9:vgpr_32 = V_MOV_B32_dpp undef %2, %1, 1, 15, 15, 1, implicit $exec - %10:vgpr_32 = V_ADD_F32_e64 4, %9, 8, %0, 0, 0, implicit $exec -... - -# tests on sequences of dpp consumers -# CHECK-LABEL: name: dpp_seq -# CHECK: %4:vgpr_32 = V_ADD_I32_dpp %1, %0, %1, 1, 14, 15, 0, implicit-def $vcc, implicit $exec -# CHECK: %5:vgpr_32 = V_SUBREV_I32_dpp %1, %0, %1, 1, 14, 15, 0, implicit-def $vcc, implicit $exec -# CHECK: %6:vgpr_32 = V_OR_B32_dpp %1, %0, %1, 1, 14, 15, 0, implicit $exec -# broken sequence: -# CHECK: %7:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 0, implicit $exec - -name: dpp_seq -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - %0:vgpr_32 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - %2:vgpr_32 = V_MOV_B32_e32 0, implicit $exec - - %3:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 0, implicit $exec - %4:vgpr_32 = V_ADD_I32_e32 %3, %1, implicit-def $vcc, implicit $exec - %5:vgpr_32 = V_SUB_I32_e32 %1, %3, implicit-def $vcc, implicit $exec - %6:vgpr_32 = V_OR_B32_e32 %3, %1, implicit $exec - - %7:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 14, 15, 0, implicit $exec - %8:vgpr_32 = V_ADD_I32_e32 %7, %1, implicit-def $vcc, implicit $exec - ; this breaks the sequence - %9:vgpr_32 = V_SUB_I32_e32 5, %7, implicit-def $vcc, implicit $exec -... - -# old reg def is in diff BB - cannot combine -# CHECK-LABEL: name: old_in_diff_bb -# CHECK: %3:vgpr_32 = V_MOV_B32_dpp %2, %1, 1, 1, 1, 0, implicit $exec - -name: old_in_diff_bb -tracksRegLiveness: true -body: | - bb.0: - successors: %bb.1 - liveins: $vgpr0, $vgpr1 - - %0:vgpr_32 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - %2:vgpr_32 = V_MOV_B32_e32 0, implicit $exec - S_BRANCH %bb.1 - - bb.1: - %3:vgpr_32 = V_MOV_B32_dpp %2, %1, 1, 1, 1, 0, implicit $exec - %4:vgpr_32 = V_ADD_U32_e32 %3, %0, implicit $exec -... - -# old reg def is in diff BB but bound_ctrl:0 - can combine -# CHECK-LABEL: name: old_in_diff_bb_bctrl_zero -# CHECK: %4:vgpr_32 = V_ADD_U32_dpp {{%[0-9]}}, %0, %1, 1, 15, 15, 1, implicit $exec - -name: old_in_diff_bb_bctrl_zero -tracksRegLiveness: true -body: | - bb.0: - successors: %bb.1 - liveins: $vgpr0, $vgpr1 - - %0:vgpr_32 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - %2:vgpr_32 = V_MOV_B32_e32 0, implicit $exec - S_BRANCH %bb.1 - - bb.1: - %3:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 15, 15, 1, implicit $exec - %4:vgpr_32 = V_ADD_U32_e32 %3, %1, implicit $exec -... - -# EXEC mask changed between def and use - cannot combine -# CHECK-LABEL: name: exec_changed -# CHECK: %3:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 15, 15, 1, implicit $exec - -name: exec_changed -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - - %0:vgpr_32 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - %2:vgpr_32 = V_MOV_B32_e32 0, implicit $exec - %3:vgpr_32 = V_MOV_B32_dpp %2, %0, 1, 15, 15, 1, implicit $exec - %4:vgpr_32 = V_ADD_U32_e32 %3, %1, implicit $exec - %5:sreg_64 = COPY $exec, implicit-def $exec - %6:vgpr_32 = V_ADD_U32_e32 %3, %1, implicit $exec -... - -# test if $old definition is correctly tracked through subreg manipulation pseudos - -# CHECK-LABEL: name: mul_old_subreg -# CHECK: %7:vgpr_32 = V_MUL_I32_I24_dpp %0.sub1, %1, %0.sub1, 1, 1, 1, 0, implicit $exec - -name: mul_old_subreg -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - - %0:vreg_64 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - %2:vgpr_32 = V_MOV_B32_e32 1, implicit $exec - %3:vgpr_32 = V_MOV_B32_e32 42, implicit $exec - %4:vreg_64 = REG_SEQUENCE %2, %subreg.sub0, %3, %subreg.sub1 - %5:vreg_64 = INSERT_SUBREG %4, %1, %subreg.sub1 ; %5.sub0 is taken from %4 - %6:vgpr_32 = V_MOV_B32_dpp %5.sub0, %1, 1, 1, 1, 0, implicit $exec - %7:vgpr_32 = V_MUL_I32_I24_e32 %6, %0.sub1, implicit $exec -... - -# CHECK-LABEL: name: add_old_subreg -# CHECK: %5:vgpr_32 = V_ADD_U32_dpp %0.sub1, %1, %0.sub1, 1, 1, 1, 0, implicit $exec - -name: add_old_subreg -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - - %0:vreg_64 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - %2:vgpr_32 = V_MOV_B32_e32 0, implicit $exec - %3:vreg_64 = INSERT_SUBREG %0, %2, %subreg.sub1 ; %3.sub1 is inserted - %4:vgpr_32 = V_MOV_B32_dpp %3.sub1, %1, 1, 1, 1, 0, implicit $exec - %5:vgpr_32 = V_ADD_U32_e32 %4, %0.sub1, implicit $exec -... - -# CHECK-LABEL: name: add_old_subreg_undef -# CHECK: %5:vgpr_32 = V_ADD_U32_dpp %3.sub1, %1, %0.sub1, 1, 15, 15, 1, implicit $exec - -name: add_old_subreg_undef -tracksRegLiveness: true -body: | - bb.0: - liveins: $vgpr0, $vgpr1 - - %0:vreg_64 = COPY $vgpr0 - %1:vgpr_32 = COPY $vgpr1 - %2:vgpr_32 = V_MOV_B32_e32 0, implicit $exec - %3:vreg_64 = REG_SEQUENCE %2, %subreg.sub0 ; %3.sub1 is undef - %4:vgpr_32 = V_MOV_B32_dpp %3.sub1, %1, 1, 15, 15, 1, implicit $exec - %5:vgpr_32 = V_ADD_U32_e32 %4, %0.sub1, implicit $exec diff --git a/llvm/test/CodeGen/AMDGPU/dpp_combine_subregs.mir b/llvm/test/CodeGen/AMDGPU/dpp_combine_subregs.mir new file mode 100644 index 00000000000..83f992f492a --- /dev/null +++ b/llvm/test/CodeGen/AMDGPU/dpp_combine_subregs.mir @@ -0,0 +1,143 @@ +# RUN: llc -march=amdgcn -mcpu=tonga -run-pass=gcn-dpp-combine -o - %s | FileCheck %s + +# test if $old definition is correctly tracked through subreg manipulation pseudos + +--- +# CHECK-LABEL: name: mul_old_subreg +# CHECK: %7:vgpr_32 = V_MUL_I32_I24_dpp %0.sub1, %1, %0.sub1, 1, 1, 1, 0, implicit $exec + +name: mul_old_subreg +tracksRegLiveness: true +registers: + - { id: 0, class: vreg_64 } + - { id: 1, class: vgpr_32 } + - { id: 2, class: vgpr_32 } + - { id: 3, class: vgpr_32 } + - { id: 4, class: vreg_64 } + - { id: 5, class: vreg_64 } + - { id: 6, class: vgpr_32 } + - { id: 7, class: vgpr_32 } + +liveins: + - { reg: '$vgpr0', virtual-reg: '%0' } + - { reg: '$vgpr1', virtual-reg: '%1' } +body: | + bb.0: + liveins: $vgpr0, $vgpr1 + + %0:vreg_64 = COPY $vgpr0 + %1:vgpr_32 = COPY $vgpr1 + %2:vgpr_32 = V_MOV_B32_e32 1, implicit $exec + %3:vgpr_32 = V_MOV_B32_e32 42, implicit $exec + %4 = REG_SEQUENCE %2, %subreg.sub0, %3, %subreg.sub1 + %5 = INSERT_SUBREG %4, %1, %subreg.sub1 ; %5.sub0 is taken from %4 + %6:vgpr_32 = V_MOV_B32_dpp %5.sub0, %1, 1, 1, 1, 0, implicit $exec + %7:vgpr_32 = V_MUL_I32_I24_e32 %6, %0.sub1, implicit $exec +... + +# CHECK-LABEL: name: add_old_subreg +# CHECK: [[OLD:\%[0-9]+]]:vgpr_32 = IMPLICIT_DEF +# CHECK: %5:vgpr_32 = V_ADD_U32_dpp [[OLD]], %1, %0.sub1, 1, 1, 1, 1, implicit $exec + +name: add_old_subreg +tracksRegLiveness: true +registers: + - { id: 0, class: vreg_64 } + - { id: 1, class: vgpr_32 } + - { id: 2, class: vgpr_32 } + - { id: 3, class: vreg_64 } + - { id: 4, class: vgpr_32 } + - { id: 5, class: vgpr_32 } + +liveins: + - { reg: '$vgpr0', virtual-reg: '%0' } + - { reg: '$vgpr1', virtual-reg: '%1' } +body: | + bb.0: + liveins: $vgpr0, $vgpr1 + + %0:vreg_64 = COPY $vgpr0 + %1:vgpr_32 = COPY $vgpr1 + %2:vgpr_32 = V_MOV_B32_e32 0, implicit $exec + %3:vreg_64 = INSERT_SUBREG %0, %2, %subreg.sub1 ; %3.sub1 is inserted + %4:vgpr_32 = V_MOV_B32_dpp %3.sub1, %1, 1, 1, 1, 0, implicit $exec + %5:vgpr_32 = V_ADD_U32_e32 %4, %0.sub1, implicit $exec +... + +# CHECK-LABEL: name: add_old_subreg_undef +# CHECK: %5:vgpr_32 = V_ADD_U32_dpp %3.sub1, %1, %0.sub1, 1, 1, 1, 0, implicit $exec + +name: add_old_subreg_undef +tracksRegLiveness: true +registers: + - { id: 0, class: vreg_64 } + - { id: 1, class: vgpr_32 } + - { id: 2, class: vgpr_32 } + - { id: 3, class: vreg_64 } + - { id: 4, class: vgpr_32 } + - { id: 5, class: vgpr_32 } + +liveins: + - { reg: '$vgpr0', virtual-reg: '%0' } + - { reg: '$vgpr1', virtual-reg: '%1' } +body: | + bb.0: + liveins: $vgpr0, $vgpr1 + + %0:vreg_64 = COPY $vgpr0 + %1:vgpr_32 = COPY $vgpr1 + %2:vgpr_32 = V_MOV_B32_e32 0, implicit $exec + %3:vreg_64 = REG_SEQUENCE %2, %subreg.sub0 ; %3.sub1 is undef + %4:vgpr_32 = V_MOV_B32_dpp %3.sub1, %1, 1, 1, 1, 0, implicit $exec + %5:vgpr_32 = V_ADD_U32_e32 %4, %0.sub1, implicit $exec +... + +# CHECK-LABEL: name: add_f32_e64 +# CHECK: %3:vgpr_32 = V_MOV_B32_dpp undef %2, %1, 1, 1, 1, 1, implicit $exec +# CHECK: %4:vgpr_32 = V_ADD_F32_e64 0, %3, 0, %0, 0, 1, implicit $exec +# CHECK: %6:vgpr_32 = V_ADD_F32_dpp %2, 0, %1, 0, %0, 1, 1, 1, 1, implicit $exec +# CHECK: %7:vgpr_32 = V_ADD_F32_dpp %2, 1, %1, 2, %0, 1, 1, 1, 1, implicit $exec +# CHECK: %9:vgpr_32 = V_ADD_F32_e64 4, %8, 8, %0, 0, 0, implicit $exec + +name: add_f32_e64 +tracksRegLiveness: true +registers: + - { id: 0, class: vgpr_32 } + - { id: 1, class: vgpr_32 } + - { id: 2, class: vgpr_32 } + - { id: 3, class: vgpr_32 } + - { id: 4, class: vgpr_32 } + - { id: 5, class: vgpr_32 } + - { id: 6, class: vgpr_32 } + - { id: 7, class: vgpr_32 } + - { id: 8, class: vgpr_32 } + - { id: 9, class: vgpr_32 } + +liveins: + - { reg: '$vgpr0', virtual-reg: '%0' } + - { reg: '$vgpr1', virtual-reg: '%1' } +body: | + bb.0: + liveins: $vgpr0, $vgpr1 + + %0:vgpr_32 = COPY $vgpr0 + %1:vgpr_32 = COPY $vgpr1 + %2:vgpr_32 = IMPLICIT_DEF + %3:vgpr_32 = V_MOV_B32_dpp undef %2, %1, 1, 1, 1, 1, implicit $exec + + ; this shouldn't be combined as omod is set + %4:vgpr_32 = V_ADD_F32_e64 0, %3, 0, %0, 0, 1, implicit $exec + + %5:vgpr_32 = V_MOV_B32_dpp undef %2, %1, 1, 1, 1, 1, implicit $exec + + ; this should be combined as all modifiers are default + %6:vgpr_32 = V_ADD_F32_e64 0, %5, 0, %0, 0, 0, implicit $exec + + ; this should be combined as modifiers other than abs|neg are default + %7:vgpr_32 = V_ADD_F32_e64 1, %5, 2, %0, 0, 0, implicit $exec + + %8:vgpr_32 = V_MOV_B32_dpp undef %2, %1, 1, 1, 1, 1, implicit $exec + + ; this shouldn't be combined as modifiers aren't abs|neg + %9:vgpr_32 = V_ADD_F32_e64 4, %8, 8, %0, 0, 0, implicit $exec +... |

