diff options
| author | Stanislav Mekhanoshin <Stanislav.Mekhanoshin@amd.com> | 2018-06-29 16:26:53 +0000 | 
|---|---|---|
| committer | Stanislav Mekhanoshin <Stanislav.Mekhanoshin@amd.com> | 2018-06-29 16:26:53 +0000 | 
| commit | 20d4795d933c3764e34d9a9a7a3e61f797980a6b (patch) | |
| tree | 783b01b028e4793a834f127826db0442eabd9273 /llvm/test/CodeGen/AMDGPU | |
| parent | 3994bafbc7b0bb1d4f21a2506be2a134aa6f886c (diff) | |
| download | bcm5719-llvm-20d4795d933c3764e34d9a9a7a3e61f797980a6b.tar.gz bcm5719-llvm-20d4795d933c3764e34d9a9a7a3e61f797980a6b.zip  | |
[AMDGPU] Enable LICM in the BE pipeline
This allows to hoist code portion to compute reciprocal of loop
invariant denominator in integer division after codegen prepare
expansion.
Differential Revision: https://reviews.llvm.org/D48604
llvm-svn: 335988
Diffstat (limited to 'llvm/test/CodeGen/AMDGPU')
| -rw-r--r-- | llvm/test/CodeGen/AMDGPU/amdpal_scratch_mergedshader.ll | 2 | ||||
| -rw-r--r-- | llvm/test/CodeGen/AMDGPU/collapse-endcf.ll | 2 | ||||
| -rw-r--r-- | llvm/test/CodeGen/AMDGPU/i1-copy-from-loop.ll | 12 | ||||
| -rw-r--r-- | llvm/test/CodeGen/AMDGPU/idiv-licm.ll | 249 | ||||
| -rw-r--r-- | llvm/test/CodeGen/AMDGPU/infinite-loop.ll | 19 | ||||
| -rw-r--r-- | llvm/test/CodeGen/AMDGPU/multilevel-break.ll | 2 | ||||
| -rw-r--r-- | llvm/test/CodeGen/AMDGPU/si-annotate-cf.ll | 34 | ||||
| -rw-r--r-- | llvm/test/CodeGen/AMDGPU/smrd.ll | 1 | 
8 files changed, 289 insertions, 32 deletions
diff --git a/llvm/test/CodeGen/AMDGPU/amdpal_scratch_mergedshader.ll b/llvm/test/CodeGen/AMDGPU/amdpal_scratch_mergedshader.ll index b062d9316c2..691078739a2 100644 --- a/llvm/test/CodeGen/AMDGPU/amdpal_scratch_mergedshader.ll +++ b/llvm/test/CodeGen/AMDGPU/amdpal_scratch_mergedshader.ll @@ -24,7 +24,7 @@ define amdgpu_hs void @_amdgpu_hs_main(i32 inreg %arg, i32 inreg %arg1, i32 inre  .endls:                                           ; preds = %.beginls, %.entry    %.fca.2.gep120.i = getelementptr inbounds [3 x <4 x float>], [3 x <4 x float>] addrspace(5)* %__llpc_global_proxy_7.i, i64 0, i64 2 -  store <4 x float> <float 9.000000e+00, float 1.000000e+01, float 1.100000e+01, float 1.200000e+01>, <4 x float> addrspace(5)* %.fca.2.gep120.i, align 16 +  store volatile <4 x float> <float 9.000000e+00, float 1.000000e+01, float 1.100000e+01, float 1.200000e+01>, <4 x float> addrspace(5)* %.fca.2.gep120.i, align 16    br label %bb  bb:                                               ; preds = %bb, %.endls diff --git a/llvm/test/CodeGen/AMDGPU/collapse-endcf.ll b/llvm/test/CodeGen/AMDGPU/collapse-endcf.ll index 2b369231b79..76b8f2d42da 100644 --- a/llvm/test/CodeGen/AMDGPU/collapse-endcf.ll +++ b/llvm/test/CodeGen/AMDGPU/collapse-endcf.ll @@ -210,7 +210,7 @@ bb.end:                                           ; preds = %bb.then, %bb  ; GCN-NEXT: s_cbranch_execnz [[BB1_LOOP]]  ; GCN: buffer_load_dword v{{[0-9]+}}, v{{[0-9]+}}, s{{\[[0-9]+:[0-9]+\]}}, s{{[0-9]+}} offen -; GCN: s_and_b64 exec, exec, vcc +; GCN: s_and_b64 exec, exec, {{vcc|s\[[0-9:]+\]}}  ; GCN-NOT: s_or_b64 exec, exec diff --git a/llvm/test/CodeGen/AMDGPU/i1-copy-from-loop.ll b/llvm/test/CodeGen/AMDGPU/i1-copy-from-loop.ll index 0a4a99b5bdc..84a2d3d3a7b 100644 --- a/llvm/test/CodeGen/AMDGPU/i1-copy-from-loop.ll +++ b/llvm/test/CodeGen/AMDGPU/i1-copy-from-loop.ll @@ -7,11 +7,13 @@  ; only contain the lanes that were active during the last loop iteration.  ;  ; SI: ; %for.body -; SI: v_cmp_gt_u32_e64 [[SREG:s\[[0-9]+:[0-9]+\]]], 4, -; SI: v_cndmask_b32_e64 [[VREG:v[0-9]+]], 0, -1, [[SREG]] -; SI-NOT: [[VREG]] -; SI: ; %for.end -; SI: v_cmp_ne_u32_e32 vcc, 0, [[VREG]] +; SI:      v_cmp_gt_u32_e64 [[SREG:s\[[0-9]+:[0-9]+\]]], 4, +; SI:      v_cndmask_b32_e64 [[VREG:v[0-9]+]], 0, -1, [[SREG]] +; SI-NEXT: s_cbranch_vccnz [[ENDIF:BB[0-9_]+]] +; SI:      [[ENDIF]]: +; SI-NOT:  [[VREG]] +; SI:      ; %for.end +; SI:      v_cmp_ne_u32_e32 vcc, 0, [[VREG]]  define amdgpu_ps void @i1_copy_from_loop(<4 x i32> inreg %rsrc, i32 %tid) {  entry:    br label %for.body diff --git a/llvm/test/CodeGen/AMDGPU/idiv-licm.ll b/llvm/test/CodeGen/AMDGPU/idiv-licm.ll new file mode 100644 index 00000000000..c2560a070ca --- /dev/null +++ b/llvm/test/CodeGen/AMDGPU/idiv-licm.ll @@ -0,0 +1,249 @@ +; RUN: llc -march=amdgcn -verify-machineinstrs < %s | FileCheck -enable-var-scope -check-prefix=GCN %s + +; GCN-LABEL: {{^}}udiv32_invariant_denom: +; GCN:     v_cvt_f32_u32 +; GCN:     v_rcp_iflag_f32 +; GCN:     v_mul_f32_e32 v{{[0-9]+}}, 0x4f800000, +; GCN:     v_cvt_u32_f32_e32 +; GCN-DAG: v_mul_hi_u32 +; GCN-DAG: v_mul_lo_i32 +; GCN-DAG: v_sub_i32_e32 +; GCN-DAG: v_cmp_eq_u32_e64 +; GCN-DAG: v_cndmask_b32_e64 +; GCN-DAG: v_mul_hi_u32 +; GCN-DAG: v_add_i32_e32 +; GCN-DAG: v_subrev_i32_e32 +; GCN-DAG: v_cndmask_b32_e64 +; GCN:     [[LOOP:BB[0-9_]+]]: +; GCN-NOT: v_rcp +; GCN:     s_cbranch_scc0 [[LOOP]] +; GCN:     s_endpgm +define amdgpu_kernel void @udiv32_invariant_denom(i32 addrspace(1)* nocapture %arg, i32 %arg1) { +bb: +  br label %bb3 + +bb2:                                              ; preds = %bb3 +  ret void + +bb3:                                              ; preds = %bb3, %bb +  %tmp = phi i32 [ 0, %bb ], [ %tmp7, %bb3 ] +  %tmp4 = udiv i32 %tmp, %arg1 +  %tmp5 = zext i32 %tmp to i64 +  %tmp6 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 %tmp5 +  store i32 %tmp4, i32 addrspace(1)* %tmp6, align 4 +  %tmp7 = add nuw nsw i32 %tmp, 1 +  %tmp8 = icmp eq i32 %tmp7, 1024 +  br i1 %tmp8, label %bb2, label %bb3 +} + +; GCN-LABEL: {{^}}urem32_invariant_denom: +; GCN:     v_cvt_f32_u32 +; GCN:     v_rcp_iflag_f32 +; GCN:     v_mul_f32_e32 v{{[0-9]+}}, 0x4f800000, +; GCN:     v_cvt_u32_f32_e32 +; GCN-DAG: v_mul_hi_u32 +; GCN-DAG: v_mul_lo_i32 +; GCN-DAG: v_sub_i32_e32 +; GCN-DAG: v_cmp_eq_u32_e64 +; GCN-DAG: v_cndmask_b32_e64 +; GCN-DAG: v_mul_hi_u32 +; GCN-DAG: v_add_i32_e32 +; GCN-DAG: v_subrev_i32_e32 +; GCN-DAG: v_cndmask_b32_e64 +; GCN:     [[LOOP:BB[0-9_]+]]: +; GCN-NOT: v_rcp +; GCN:     s_cbranch_scc0 [[LOOP]] +; GCN:     s_endpgm +define amdgpu_kernel void @urem32_invariant_denom(i32 addrspace(1)* nocapture %arg, i32 %arg1) { +bb: +  br label %bb3 + +bb2:                                              ; preds = %bb3 +  ret void + +bb3:                                              ; preds = %bb3, %bb +  %tmp = phi i32 [ 0, %bb ], [ %tmp7, %bb3 ] +  %tmp4 = urem i32 %tmp, %arg1 +  %tmp5 = zext i32 %tmp to i64 +  %tmp6 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 %tmp5 +  store i32 %tmp4, i32 addrspace(1)* %tmp6, align 4 +  %tmp7 = add nuw nsw i32 %tmp, 1 +  %tmp8 = icmp eq i32 %tmp7, 1024 +  br i1 %tmp8, label %bb2, label %bb3 +} + +; GCN-LABEL: {{^}}sdiv32_invariant_denom: +; GCN:     v_cvt_f32_u32 +; GCN:     v_rcp_iflag_f32 +; GCN:     v_mul_f32_e32 v{{[0-9]+}}, 0x4f800000, +; GCN:     v_cvt_u32_f32_e32 +; GCN-DAG: v_mul_hi_u32 +; GCN-DAG: v_mul_lo_i32 +; GCN-DAG: v_sub_i32_e32 +; GCN-DAG: v_cmp_eq_u32_e64 +; GCN-DAG: v_cndmask_b32_e64 +; GCN-DAG: v_mul_hi_u32 +; GCN-DAG: v_add_i32_e32 +; GCN-DAG: v_subrev_i32_e32 +; GCN-DAG: v_cndmask_b32_e64 +; GCN:     [[LOOP:BB[0-9_]+]]: +; GCN-NOT: v_rcp +; GCN:     s_cbranch_scc0 [[LOOP]] +; GCN:     s_endpgm +define amdgpu_kernel void @sdiv32_invariant_denom(i32 addrspace(1)* nocapture %arg, i32 %arg1) { +bb: +  br label %bb3 + +bb2:                                              ; preds = %bb3 +  ret void + +bb3:                                              ; preds = %bb3, %bb +  %tmp = phi i32 [ 0, %bb ], [ %tmp7, %bb3 ] +  %tmp4 = sdiv i32 %tmp, %arg1 +  %tmp5 = zext i32 %tmp to i64 +  %tmp6 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 %tmp5 +  store i32 %tmp4, i32 addrspace(1)* %tmp6, align 4 +  %tmp7 = add nuw nsw i32 %tmp, 1 +  %tmp8 = icmp eq i32 %tmp7, 1024 +  br i1 %tmp8, label %bb2, label %bb3 +} + +; GCN-LABEL: {{^}}srem32_invariant_denom: +; GCN:     v_cvt_f32_u32 +; GCN:     v_rcp_iflag_f32 +; GCN:     v_mul_f32_e32 v{{[0-9]+}}, 0x4f800000, +; GCN:     v_cvt_u32_f32_e32 +; GCN-DAG: v_mul_hi_u32 +; GCN-DAG: v_mul_lo_i32 +; GCN-DAG: v_sub_i32_e32 +; GCN-DAG: v_cmp_eq_u32_e64 +; GCN-DAG: v_cndmask_b32_e64 +; GCN-DAG: v_mul_hi_u32 +; GCN-DAG: v_add_i32_e32 +; GCN-DAG: v_subrev_i32_e32 +; GCN-DAG: v_cndmask_b32_e64 +; GCN:     [[LOOP:BB[0-9_]+]]: +; GCN-NOT: v_rcp +; GCN:     s_cbranch_scc0 [[LOOP]] +; GCN:     s_endpgm +define amdgpu_kernel void @srem32_invariant_denom(i32 addrspace(1)* nocapture %arg, i32 %arg1) { +bb: +  br label %bb3 + +bb2:                                              ; preds = %bb3 +  ret void + +bb3:                                              ; preds = %bb3, %bb +  %tmp = phi i32 [ 0, %bb ], [ %tmp7, %bb3 ] +  %tmp4 = srem i32 %tmp, %arg1 +  %tmp5 = zext i32 %tmp to i64 +  %tmp6 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 %tmp5 +  store i32 %tmp4, i32 addrspace(1)* %tmp6, align 4 +  %tmp7 = add nuw nsw i32 %tmp, 1 +  %tmp8 = icmp eq i32 %tmp7, 1024 +  br i1 %tmp8, label %bb2, label %bb3 +} + +; GCN-LABEL: {{^}}udiv16_invariant_denom: +; GCN:     v_cvt_f32_u32 +; GCN:     v_rcp_iflag_f32 +; GCN:     [[LOOP:BB[0-9_]+]]: +; GCN-NOT: v_rcp +; GCN:     s_cbranch_scc0 [[LOOP]] +; GCN:     s_endpgm +define amdgpu_kernel void @udiv16_invariant_denom(i16 addrspace(1)* nocapture %arg, i16 %arg1) { +bb: +  br label %bb3 + +bb2:                                              ; preds = %bb3 +  ret void + +bb3:                                              ; preds = %bb3, %bb +  %tmp = phi i16 [ 0, %bb ], [ %tmp7, %bb3 ] +  %tmp4 = udiv i16 %tmp, %arg1 +  %tmp5 = zext i16 %tmp to i64 +  %tmp6 = getelementptr inbounds i16, i16 addrspace(1)* %arg, i64 %tmp5 +  store i16 %tmp4, i16 addrspace(1)* %tmp6, align 2 +  %tmp7 = add nuw nsw i16 %tmp, 1 +  %tmp8 = icmp eq i16 %tmp7, 1024 +  br i1 %tmp8, label %bb2, label %bb3 +} + +; GCN-LABEL: {{^}}urem16_invariant_denom: +; GCN:     v_cvt_f32_u32 +; GCN:     v_rcp_iflag_f32 +; GCN:     [[LOOP:BB[0-9_]+]]: +; GCN-NOT: v_rcp +; GCN:     s_cbranch_scc0 [[LOOP]] +; GCN:     s_endpgm +define amdgpu_kernel void @urem16_invariant_denom(i16 addrspace(1)* nocapture %arg, i16 %arg1) { +bb: +  br label %bb3 + +bb2:                                              ; preds = %bb3 +  ret void + +bb3:                                              ; preds = %bb3, %bb +  %tmp = phi i16 [ 0, %bb ], [ %tmp7, %bb3 ] +  %tmp4 = urem i16 %tmp, %arg1 +  %tmp5 = zext i16 %tmp to i64 +  %tmp6 = getelementptr inbounds i16, i16 addrspace(1)* %arg, i64 %tmp5 +  store i16 %tmp4, i16 addrspace(1)* %tmp6, align 2 +  %tmp7 = add nuw nsw i16 %tmp, 1 +  %tmp8 = icmp eq i16 %tmp7, 1024 +  br i1 %tmp8, label %bb2, label %bb3 +} + +; GCN-LABEL: {{^}}sdiv16_invariant_denom: +; GCN-DAG: s_sext_i32_i16 +; GCN-DAG: v_and_b32_e32 v{{[0-9]+}}, 0x7fffffff +; GCN-DAG: v_cvt_f32_i32 +; GCN-DAG: v_rcp_iflag_f32 +; GCN:     [[LOOP:BB[0-9_]+]]: +; GCN-NOT: v_rcp +; GCN:     s_cbranch_scc0 [[LOOP]] +; GCN:     s_endpgm +define amdgpu_kernel void @sdiv16_invariant_denom(i16 addrspace(1)* nocapture %arg, i16 %arg1) { +bb: +  br label %bb3 + +bb2:                                              ; preds = %bb3 +  ret void + +bb3:                                              ; preds = %bb3, %bb +  %tmp = phi i16 [ 0, %bb ], [ %tmp7, %bb3 ] +  %tmp4 = sdiv i16 %tmp, %arg1 +  %tmp5 = zext i16 %tmp to i64 +  %tmp6 = getelementptr inbounds i16, i16 addrspace(1)* %arg, i64 %tmp5 +  store i16 %tmp4, i16 addrspace(1)* %tmp6, align 2 +  %tmp7 = add nuw nsw i16 %tmp, 1 +  %tmp8 = icmp eq i16 %tmp7, 1024 +  br i1 %tmp8, label %bb2, label %bb3 +} + +; GCN-LABEL: {{^}}srem16_invariant_denom: +; GCN-DAG: s_sext_i32_i16 +; GCN-DAG: v_and_b32_e32 v{{[0-9]+}}, 0x7fffffff +; GCN-DAG: v_cvt_f32_i32 +; GCN-DAG: v_rcp_iflag_f32 +; GCN:     [[LOOP:BB[0-9_]+]]: +; GCN-NOT: v_rcp +; GCN:     s_cbranch_scc0 [[LOOP]] +; GCN:     s_endpgm +define amdgpu_kernel void @srem16_invariant_denom(i16 addrspace(1)* nocapture %arg, i16 %arg1) { +bb: +  br label %bb3 + +bb2:                                              ; preds = %bb3 +  ret void + +bb3:                                              ; preds = %bb3, %bb +  %tmp = phi i16 [ 0, %bb ], [ %tmp7, %bb3 ] +  %tmp4 = srem i16 %tmp, %arg1 +  %tmp5 = zext i16 %tmp to i64 +  %tmp6 = getelementptr inbounds i16, i16 addrspace(1)* %arg, i64 %tmp5 +  store i16 %tmp4, i16 addrspace(1)* %tmp6, align 2 +  %tmp7 = add nuw nsw i16 %tmp, 1 +  %tmp8 = icmp eq i16 %tmp7, 1024 +  br i1 %tmp8, label %bb2, label %bb3 +} diff --git a/llvm/test/CodeGen/AMDGPU/infinite-loop.ll b/llvm/test/CodeGen/AMDGPU/infinite-loop.ll index 567b5a9f450..5005b781ca3 100644 --- a/llvm/test/CodeGen/AMDGPU/infinite-loop.ll +++ b/llvm/test/CodeGen/AMDGPU/infinite-loop.ll @@ -12,7 +12,7 @@ entry:    br label %loop  loop: -  store i32 999, i32 addrspace(1)* %out, align 4 +  store volatile i32 999, i32 addrspace(1)* %out, align 4    br label %loop  } @@ -21,7 +21,7 @@ loop:  ; IR:  br i1 %cond, label %loop, label %UnifiedReturnBlock  ; IR: loop: -; IR: store i32 999, i32 addrspace(1)* %out, align 4 +; IR: store volatile i32 999, i32 addrspace(1)* %out, align 4  ; IR: br i1 true, label %loop, label %UnifiedReturnBlock  ; IR: UnifiedReturnBlock: @@ -47,7 +47,7 @@ entry:    br i1 %cond, label %loop, label %return  loop: -  store i32 999, i32 addrspace(1)* %out, align 4 +  store volatile i32 999, i32 addrspace(1)* %out, align 4    br label %loop  return: @@ -59,11 +59,11 @@ return:  ; IR: br i1 undef, label %loop1, label %loop2  ; IR: loop1: -; IR: store i32 999, i32 addrspace(1)* %out, align 4 +; IR: store volatile i32 999, i32 addrspace(1)* %out, align 4  ; IR: br i1 true, label %loop1, label %DummyReturnBlock  ; IR: loop2: -; IR: store i32 888, i32 addrspace(1)* %out, align 4 +; IR: store volatile i32 888, i32 addrspace(1)* %out, align 4  ; IR: br i1 true, label %loop2, label %DummyReturnBlock  ; IR: DummyReturnBlock: @@ -96,11 +96,11 @@ entry:    br i1 undef, label %loop1, label %loop2  loop1: -  store i32 999, i32 addrspace(1)* %out, align 4 +  store volatile i32 999, i32 addrspace(1)* %out, align 4    br label %loop1  loop2: -  store i32 888, i32 addrspace(1)* %out, align 4 +  store volatile i32 888, i32 addrspace(1)* %out, align 4    br label %loop2  } @@ -113,7 +113,7 @@ loop2:  ; IR: br label %inner_loop  ; IR: inner_loop: -; IR: store i32 999, i32 addrspace(1)* %out, align 4 +; IR: store volatile i32 999, i32 addrspace(1)* %out, align 4  ; IR: %cond3 = icmp eq i32 %tmp, 3  ; IR: br i1 true, label %TransitionBlock, label %UnifiedReturnBlock @@ -132,7 +132,6 @@ loop2:  ; SI: [[INNER_LOOP:BB[0-9]+_[0-9]+]]:  ; %inner_loop  ; SI: s_waitcnt expcnt(0)  ; SI: v_mov_b32_e32 [[REG:v[0-9]+]], 0x3e7 -; SI: v_cmp_ne_u32_e32  ; SI: s_waitcnt lgkmcnt(0)  ; SI: buffer_store_dword [[REG]] @@ -156,7 +155,7 @@ outer_loop:   br label %inner_loop  inner_loop:                                     ; preds = %LeafBlock, %LeafBlock1 -  store i32 999, i32 addrspace(1)* %out, align 4 +  store volatile i32 999, i32 addrspace(1)* %out, align 4    %cond3 = icmp eq i32 %tmp, 3    br i1 %cond3, label %inner_loop, label %outer_loop diff --git a/llvm/test/CodeGen/AMDGPU/multilevel-break.ll b/llvm/test/CodeGen/AMDGPU/multilevel-break.ll index cfca7b644b2..3f7df7b6005 100644 --- a/llvm/test/CodeGen/AMDGPU/multilevel-break.ll +++ b/llvm/test/CodeGen/AMDGPU/multilevel-break.ll @@ -36,7 +36,7 @@  ; GCN-NEXT: s_andn2_b64 exec, exec, [[OR_BREAK]]  ; GCN-NEXT: s_cbranch_execnz [[INNER_LOOP]] -; GCN: ; %bb.{{[0-9]+}}: ; %Flow1{{$}} +; GCN: ; %bb.{{[0-9]+}}: ; %Flow2{{$}}  ; GCN-NEXT: ; in Loop: Header=[[OUTER_LOOP]] Depth=1  ; Ensure copy is eliminated diff --git a/llvm/test/CodeGen/AMDGPU/si-annotate-cf.ll b/llvm/test/CodeGen/AMDGPU/si-annotate-cf.ll index 7e7f6fc5e6d..73e56593ce8 100644 --- a/llvm/test/CodeGen/AMDGPU/si-annotate-cf.ll +++ b/llvm/test/CodeGen/AMDGPU/si-annotate-cf.ll @@ -1,5 +1,5 @@ -; RUN: llc < %s -march=amdgcn -mcpu=verde -verify-machineinstrs | FileCheck --check-prefix=SI --check-prefix=FUNC %s -; RUN: llc < %s -march=amdgcn -mcpu=tonga -mattr=-flat-for-global -verify-machineinstrs | FileCheck --check-prefix=SI --check-prefix=FUNC %s +; RUN: llc < %s -march=amdgcn -mcpu=verde -asm-verbose=0 -verify-machineinstrs | FileCheck --check-prefix=SI --check-prefix=FUNC %s +; RUN: llc < %s -march=amdgcn -mcpu=tonga -mattr=-flat-for-global -asm-verbose=0 -verify-machineinstrs | FileCheck --check-prefix=SI --check-prefix=FUNC %s  ; FUNC-LABEL: {{^}}break_inserted_outside_of_loop: @@ -89,17 +89,24 @@ declare float @llvm.fabs.f32(float) nounwind readnone  ; This broke the old AMDIL cfg structurizer  ; FUNC-LABEL: {{^}}loop_land_info_assert: -; SI: s_cmp_lt_i32 -; SI-NEXT: s_cbranch_scc1 [[ENDPGM:BB[0-9]+_[0-9]+]] - -; SI: s_cmpk_lt_i32 -; SI-NEXT: s_cbranch_scc0 [[ENDPGM]] - -; SI: [[INFLOOP:BB[0-9]+_[0-9]+]] -; SI: s_cbranch_vccnz [[INFLOOP]] - -; SI: [[ENDPGM]]: -; SI: s_endpgm +; SI:      v_cmp_lt_i32_e64 [[CMP4:s\[[0-9:]+\]]], s{{[0-9]+}}, 4{{$}} +; SI:      s_and_b64 vcc, exec, [[CMP4]] +; SI-NEXT: s_cbranch_vccnz [[BR1:BB[0-9_]+]] +; SI-NEXT: s_branch [[BR2:BB[0-9_]+]] +; SI-NEXT: BB{{[0-9_]+}}: +; SI-NEXT: buffer_store_dword + +; SI:      [[INFLOOP:BB[0-9]+_[0-9]+]]: + +; SI:      [[BR1]]: +; SI-NEXT: s_and_b64 vcc, exec, +; SI-NEXT: s_cbranch_vccnz [[ENDPGM:BB[0-9]+_[0-9]+]] +; SI:      s_branch [[INFLOOP]] +; SI-NEXT: [[BR2]]: +; SI:      s_cbranch_vccz [[ENDPGM]] + +; SI:      [[ENDPGM]]: +; SI-NEXT: s_endpgm  define amdgpu_kernel void @loop_land_info_assert(i32 %c0, i32 %c1, i32 %c2, i32 %c3, i32 %x, i32 %y, i1 %arg) nounwind {  entry:    %cmp = icmp sgt i32 %c0, 0 @@ -144,7 +151,6 @@ return:    ret void  } -  declare i32 @llvm.amdgcn.mbcnt.lo(i32, i32) #0  attributes #0 = { nounwind readnone } diff --git a/llvm/test/CodeGen/AMDGPU/smrd.ll b/llvm/test/CodeGen/AMDGPU/smrd.ll index d16dcd9ccac..aa18da122f5 100644 --- a/llvm/test/CodeGen/AMDGPU/smrd.ll +++ b/llvm/test/CodeGen/AMDGPU/smrd.ll @@ -328,6 +328,7 @@ ret_block:                                       ; preds = %.outer, %.label22, %  .inner_loop_body:    %descriptor = load <4 x i32>, <4 x i32> addrspace(4)* %descptr, align 16, !invariant.load !0    %load1result = call float @llvm.SI.load.const.v4i32(<4 x i32> %descriptor, i32 0) +  store float %load1result, float addrspace(1)* undef    %inner_br2 = icmp uge i32 %1, 10    br i1 %inner_br2, label %.inner_loop_header, label %.outer_loop_body  | 

