summaryrefslogtreecommitdiffstats
path: root/llvm/test/CodeGen/AMDGPU/stack-pointer-offset-relative-frameindex.ll
diff options
context:
space:
mode:
authorcdevadas <cdevadas@amd.com>2020-01-10 22:23:27 +0530
committercdevadas <cdevadas@amd.com>2020-01-15 15:18:16 +0530
commit0dc6c249bffac9f23a605ce4e42a84341da3ddbd (patch)
tree113cc776987199087010ef82f8fd4728b06d0c8b /llvm/test/CodeGen/AMDGPU/stack-pointer-offset-relative-frameindex.ll
parent064859bde79ccd221fd5196fd2d889014c5435c4 (diff)
downloadbcm5719-llvm-0dc6c249bffac9f23a605ce4e42a84341da3ddbd.tar.gz
bcm5719-llvm-0dc6c249bffac9f23a605ce4e42a84341da3ddbd.zip
[AMDGPU] Invert the handling of skip insertion.
The current implementation of skip insertion (SIInsertSkip) makes it a mandatory pass required for correctness. Initially, the idea was to have an optional pass. This patch inserts the s_cbranch_execz upfront during SILowerControlFlow to skip over the sections of code when no lanes are active. Later, SIRemoveShortExecBranches removes the skips for short branches, unless there is a sideeffect and the skip branch is really necessary. This new pass will replace the handling of skip insertion in the existing SIInsertSkip Pass. Differential revision: https://reviews.llvm.org/D68092
Diffstat (limited to 'llvm/test/CodeGen/AMDGPU/stack-pointer-offset-relative-frameindex.ll')
-rw-r--r--llvm/test/CodeGen/AMDGPU/stack-pointer-offset-relative-frameindex.ll3
1 files changed, 1 insertions, 2 deletions
diff --git a/llvm/test/CodeGen/AMDGPU/stack-pointer-offset-relative-frameindex.ll b/llvm/test/CodeGen/AMDGPU/stack-pointer-offset-relative-frameindex.ll
index 00ae166a6ce..be60a34b420 100644
--- a/llvm/test/CodeGen/AMDGPU/stack-pointer-offset-relative-frameindex.ll
+++ b/llvm/test/CodeGen/AMDGPU/stack-pointer-offset-relative-frameindex.ll
@@ -28,9 +28,8 @@ define amdgpu_kernel void @kernel_background_evaluate(float addrspace(5)* %kg, <
; GCN-NEXT: s_swappc_b64 s[30:31], s[4:5]
; GCN-NEXT: v_cmp_ne_u32_e32 vcc_lo, 0, v0
; GCN-NEXT: s_and_saveexec_b32 s0, vcc_lo
-; GCN-NEXT: ; mask branch BB0_2
; GCN-NEXT: s_cbranch_execz BB0_2
-; GCN-NEXT: BB0_1: ; %if.then4.i
+; GCN-NEXT: ; %bb.1: ; %if.then4.i
; GCN-NEXT: buffer_load_dword v0, v32, s[36:39], s32 offen
; GCN-NEXT: buffer_load_dword v1, v32, s[36:39], s32 offen offset:4
; GCN-NEXT: s_waitcnt vmcnt(0)
OpenPOWER on IntegriCloud