diff options
-rw-r--r-- | llvm/lib/Transforms/Scalar/LoopStrengthReduce.cpp | 16 | ||||
-rw-r--r-- | llvm/test/Transforms/LoopStrengthReduce/AMDGPU/preserve-addrspace-assert.ll | 43 |
2 files changed, 5 insertions, 54 deletions
diff --git a/llvm/lib/Transforms/Scalar/LoopStrengthReduce.cpp b/llvm/lib/Transforms/Scalar/LoopStrengthReduce.cpp index e1876ae30f8..2caa526eb23 100644 --- a/llvm/lib/Transforms/Scalar/LoopStrengthReduce.cpp +++ b/llvm/lib/Transforms/Scalar/LoopStrengthReduce.cpp @@ -822,7 +822,7 @@ static bool isAddressUse(const TargetTransformInfo &TTI, /// Return the type of the memory being accessed. static MemAccessTy getAccessType(const TargetTransformInfo &TTI, - Instruction *Inst, Value *OperandVal) { + Instruction *Inst) { MemAccessTy AccessTy(Inst->getType(), MemAccessTy::UnknownAddressSpace); if (const StoreInst *SI = dyn_cast<StoreInst>(Inst)) { AccessTy.MemTy = SI->getOperand(0)->getType(); @@ -836,14 +836,7 @@ static MemAccessTy getAccessType(const TargetTransformInfo &TTI, } else if (IntrinsicInst *II = dyn_cast<IntrinsicInst>(Inst)) { switch (II->getIntrinsicID()) { case Intrinsic::prefetch: - case Intrinsic::memset: AccessTy.AddrSpace = II->getArgOperand(0)->getType()->getPointerAddressSpace(); - AccessTy.MemTy = OperandVal->getType(); - break; - case Intrinsic::memmove: - case Intrinsic::memcpy: - AccessTy.AddrSpace = OperandVal->getType()->getPointerAddressSpace(); - AccessTy.MemTy = OperandVal->getType(); break; default: { MemIntrinsicInfo IntrInfo; @@ -2415,8 +2408,7 @@ LSRInstance::OptimizeLoopTermCond() { C->getValue().isMinSignedValue()) goto decline_post_inc; // Check for possible scaled-address reuse. - MemAccessTy AccessTy = - getAccessType(TTI, UI->getUser(), UI->getOperandValToReplace()); + MemAccessTy AccessTy = getAccessType(TTI, UI->getUser()); int64_t Scale = C->getSExtValue(); if (TTI.isLegalAddressingMode(AccessTy.MemTy, /*BaseGV=*/nullptr, /*BaseOffset=*/0, @@ -3090,7 +3082,7 @@ static bool canFoldIVIncExpr(const SCEV *IncExpr, Instruction *UserInst, if (IncConst->getAPInt().getMinSignedBits() > 64) return false; - MemAccessTy AccessTy = getAccessType(TTI, UserInst, Operand); + MemAccessTy AccessTy = getAccessType(TTI, UserInst); int64_t IncOffset = IncConst->getValue()->getSExtValue(); if (!isAlwaysFoldable(TTI, LSRUse::Address, AccessTy, /*BaseGV=*/nullptr, IncOffset, /*HaseBaseReg=*/false)) @@ -3218,7 +3210,7 @@ void LSRInstance::CollectFixupsAndInitialFormulae() { MemAccessTy AccessTy; if (isAddressUse(TTI, UserInst, U.getOperandValToReplace())) { Kind = LSRUse::Address; - AccessTy = getAccessType(TTI, UserInst, U.getOperandValToReplace()); + AccessTy = getAccessType(TTI, UserInst); } const SCEV *S = IU.getExpr(U); diff --git a/llvm/test/Transforms/LoopStrengthReduce/AMDGPU/preserve-addrspace-assert.ll b/llvm/test/Transforms/LoopStrengthReduce/AMDGPU/preserve-addrspace-assert.ll index 0dd84a9a038..9eba0c3051d 100644 --- a/llvm/test/Transforms/LoopStrengthReduce/AMDGPU/preserve-addrspace-assert.ll +++ b/llvm/test/Transforms/LoopStrengthReduce/AMDGPU/preserve-addrspace-assert.ll @@ -10,7 +10,7 @@ target datalayout = "e-p:32:32-p1:64:64-p2:64:64-p3:32:32-p4:64:64-p5:32:32-p24: ; CHECK-LABEL: @lsr_crash_preserve_addrspace_unknown_type( ; CHECK: %tmp4 = bitcast %0 addrspace(3)* %tmp to double addrspace(3)* ; CHECK: %scevgep5 = getelementptr double, double addrspace(3)* %tmp4, i32 1 -; CHECK: load double, double addrspace(3)* %scevgep5 +; CHEC: load double, double addrspace(3)* %scevgep5 ; CHECK: %scevgep = getelementptr i32, i32 addrspace(3)* %tmp1, i32 4 ; CHECK:%tmp14 = load i32, i32 addrspace(3)* %scevgep @@ -50,46 +50,5 @@ bb17: ; preds = %bb12, %bb8 br label %bb1 } -; CHECK-LABEL: @lsr_crash_preserve_addrspace_unknown_type2( -; CHECK: %scevgep3 = getelementptr i8, i8 addrspace(5)* %array, i32 %j -; CHECK: %scevgep2 = getelementptr i8, i8 addrspace(5)* %array, i32 %j -; CHECK: %n8 = load i8, i8 addrspace(5)* %scevgep2, align 4 -; CHECK: call void @llvm.memcpy.p5i8.p3i8.i64(i8 addrspace(5)* %scevgep3, i8 addrspace(3)* %scevgep4, i64 42, i1 false) -; CHECK: call void @llvm.memmove.p5i8.p3i8.i64(i8 addrspace(5)* %scevgep3, i8 addrspace(3)* %scevgep4, i64 42, i1 false) -; CHECK: call void @llvm.memset.p5i8.i64(i8 addrspace(5)* %scevgep3, i8 42, i64 42, i1 false) -define void @lsr_crash_preserve_addrspace_unknown_type2(i8 addrspace(5)* %array, i8 addrspace(3)* %array2) { -entry: - br label %for.body - -for.body: ; preds = %entry, %for.inc - %j = phi i32 [ %add, %for.inc ], [ 0, %entry ] - %idx = getelementptr inbounds i8, i8 addrspace(5)* %array, i32 %j - %idx1 = getelementptr inbounds i8, i8 addrspace(3)* %array2, i32 %j - %t = getelementptr inbounds i8, i8 addrspace(5)* %array, i32 %j - %n8 = load i8, i8 addrspace(5)* %t, align 4 - %n7 = getelementptr inbounds i8, i8 addrspace(5)* %t, i32 42 - %n9 = load i8, i8 addrspace(5)* %n7, align 4 - %cmp = icmp sgt i32 %j, 42 - %add = add nuw nsw i32 %j, 1 - br i1 %cmp, label %if.then17, label %for.inc - -if.then17: ; preds = %for.body - call void @llvm.memcpy.p5i8.p5i8.i64(i8 addrspace(5)* %idx, i8 addrspace(3)* %idx1, i64 42, i1 false) - call void @llvm.memmove.p5i8.p5i8.i64(i8 addrspace(5)* %idx, i8 addrspace(3)* %idx1, i64 42, i1 false) - call void @llvm.memset.p5i8.i64(i8 addrspace(5)* %idx, i8 42, i64 42, i1 false) - br label %for.inc - -for.inc: ; preds = %for.body, %if.then17 - %exitcond = icmp eq i1 %cmp, 1 - br i1 %exitcond, label %end, label %for.body - -end: ; preds = %for.inc - ret void -} - -declare void @llvm.memcpy.p5i8.p5i8.i64(i8 addrspace(5)*, i8 addrspace(3)*, i64, i1) -declare void @llvm.memmove.p5i8.p5i8.i64(i8 addrspace(5)*, i8 addrspace(3)*, i64, i1) -declare void @llvm.memset.p5i8.i64(i8 addrspace(5)*, i8, i64, i1) - attributes #0 = { nounwind } attributes #1 = { nounwind readnone } |