summaryrefslogtreecommitdiffstats
path: root/llvm/lib
diff options
context:
space:
mode:
authorEvgeniy Stepanov <eugeni.stepanov@gmail.com>2019-05-03 17:31:49 +0000
committerEvgeniy Stepanov <eugeni.stepanov@gmail.com>2019-05-03 17:31:49 +0000
commit46ec57e57605e92121d972438bff04bd694f74d5 (patch)
treea008cf6c119b4c1f6ac2b5ce31bab576f0f14a60 /llvm/lib
parent30649ce09bd5ca5b3430819c3e257aefabf02cb8 (diff)
downloadbcm5719-llvm-46ec57e57605e92121d972438bff04bd694f74d5.tar.gz
bcm5719-llvm-46ec57e57605e92121d972438bff04bd694f74d5.zip
Revert "[CodeGenPrepare] limit overflow intrinsic matching to a single basic block"
This reverts commit r359879, which introduced a compiler crash. llvm-svn: 359908
Diffstat (limited to 'llvm/lib')
-rw-r--r--llvm/lib/CodeGen/CodeGenPrepare.cpp89
1 files changed, 47 insertions, 42 deletions
diff --git a/llvm/lib/CodeGen/CodeGenPrepare.cpp b/llvm/lib/CodeGen/CodeGenPrepare.cpp
index 30b09d35939..4335f10117e 100644
--- a/llvm/lib/CodeGen/CodeGenPrepare.cpp
+++ b/llvm/lib/CodeGen/CodeGenPrepare.cpp
@@ -388,9 +388,9 @@ class TypePromotionTransaction;
bool tryToSinkFreeOperands(Instruction *I);
bool replaceMathCmpWithIntrinsic(BinaryOperator *BO, CmpInst *Cmp,
Intrinsic::ID IID);
- bool optimizeCmp(CmpInst *Cmp);
- bool combineToUSubWithOverflow(CmpInst *Cmp);
- bool combineToUAddWithOverflow(CmpInst *Cmp);
+ bool optimizeCmp(CmpInst *Cmp, bool &ModifiedDT);
+ bool combineToUSubWithOverflow(CmpInst *Cmp, bool &ModifiedDT);
+ bool combineToUAddWithOverflow(CmpInst *Cmp, bool &ModifiedDT);
};
} // end anonymous namespace
@@ -484,13 +484,9 @@ bool CodeGenPrepare::runOnFunction(Function &F) {
if (!LargeOffsetGEPMap.empty())
MadeChange |= splitLargeGEPOffsets();
- // Really free instructions removed during promotion or kept around to
- // improve efficiency (see comments in overflow intrinsic transforms).
- for (Instruction *I : RemovedInsts) {
- if (I->getParent())
- I->removeFromParent();
+ // Really free removed instructions during promotion.
+ for (Instruction *I : RemovedInsts)
I->deleteValue();
- }
EverMadeChange |= MadeChange;
SeenChainsForSExt.clear();
@@ -1181,20 +1177,6 @@ static bool OptimizeNoopCopyExpression(CastInst *CI, const TargetLowering &TLI,
bool CodeGenPrepare::replaceMathCmpWithIntrinsic(BinaryOperator *BO,
CmpInst *Cmp,
Intrinsic::ID IID) {
- if (BO->getParent() != Cmp->getParent()) {
- // We used to use a dominator tree here to allow multi-block optimization.
- // But that was problematic because:
- // 1. It could cause a perf regression by hoisting the math op into the
- // critical path.
- // 2. It could cause a perf regression by creating a value that was live
- // across multiple blocks and increasing register pressure.
- // 3. Use of a dominator tree could cause large compile-time regression.
- // This is because we recompute the DT on every change in the main CGP
- // run-loop. The recomputing is probably unnecessary in many cases, so if
- // that was fixed, using a DT here would be ok.
- return false;
- }
-
// We allow matching the canonical IR (add X, C) back to (usubo X, -C).
Value *Arg0 = BO->getOperand(0);
Value *Arg1 = BO->getOperand(1);
@@ -1204,28 +1186,45 @@ bool CodeGenPrepare::replaceMathCmpWithIntrinsic(BinaryOperator *BO,
Arg1 = ConstantExpr::getNeg(cast<Constant>(Arg1));
}
- // Insert at the first instruction of the pair.
- Instruction *InsertPt = nullptr;
- for (Instruction &Iter : *Cmp->getParent()) {
- if (&Iter == BO || &Iter == Cmp) {
- InsertPt = &Iter;
- break;
+ Instruction *InsertPt;
+ if (BO->hasOneUse() && BO->user_back() == Cmp) {
+ // If the math is only used by the compare, insert at the compare to keep
+ // the condition in the same block as its users. (CGP aggressively sinks
+ // compares to help out SDAG.)
+ InsertPt = Cmp;
+ } else {
+ // The math and compare may be independent instructions. Check dominance to
+ // determine the insertion point for the intrinsic.
+ bool MathDominates = getDT(*BO->getFunction()).dominates(BO, Cmp);
+ if (!MathDominates && !getDT(*BO->getFunction()).dominates(Cmp, BO))
+ return false;
+
+ BasicBlock *MathBB = BO->getParent(), *CmpBB = Cmp->getParent();
+ if (MathBB != CmpBB) {
+ // Avoid hoisting an extra op into a dominating block and creating a
+ // potentially longer critical path.
+ if (!MathDominates)
+ return false;
+ // Check that the insertion doesn't create a value that is live across
+ // more than two blocks, so to minimise the increase in register pressure.
+ BasicBlock *Dominator = MathDominates ? MathBB : CmpBB;
+ BasicBlock *Dominated = MathDominates ? CmpBB : MathBB;
+ auto Successors = successors(Dominator);
+ if (llvm::find(Successors, Dominated) == Successors.end())
+ return false;
}
+
+ InsertPt = MathDominates ? cast<Instruction>(BO) : cast<Instruction>(Cmp);
}
- assert(InsertPt != nullptr && "Parent block did not contain cmp or binop");
IRBuilder<> Builder(InsertPt);
Value *MathOV = Builder.CreateBinaryIntrinsic(IID, Arg0, Arg1);
Value *Math = Builder.CreateExtractValue(MathOV, 0, "math");
Value *OV = Builder.CreateExtractValue(MathOV, 1, "ov");
-
- // Delay the actual removal/deletion of the binop and compare for efficiency.
- // If we delete those now, we would invalidate the instruction iterator and
- // trigger dominator tree updates.
BO->replaceAllUsesWith(Math);
Cmp->replaceAllUsesWith(OV);
- RemovedInsts.insert(BO);
- RemovedInsts.insert(Cmp);
+ BO->eraseFromParent();
+ Cmp->eraseFromParent();
return true;
}
@@ -1261,7 +1260,8 @@ static bool matchUAddWithOverflowConstantEdgeCases(CmpInst *Cmp,
/// Try to combine the compare into a call to the llvm.uadd.with.overflow
/// intrinsic. Return true if any changes were made.
-bool CodeGenPrepare::combineToUAddWithOverflow(CmpInst *Cmp) {
+bool CodeGenPrepare::combineToUAddWithOverflow(CmpInst *Cmp,
+ bool &ModifiedDT) {
Value *A, *B;
BinaryOperator *Add;
if (!match(Cmp, m_UAddWithOverflow(m_Value(A), m_Value(B), m_BinOp(Add))))
@@ -1281,10 +1281,13 @@ bool CodeGenPrepare::combineToUAddWithOverflow(CmpInst *Cmp) {
if (!replaceMathCmpWithIntrinsic(Add, Cmp, Intrinsic::uadd_with_overflow))
return false;
+ // Reset callers - do not crash by iterating over a dead instruction.
+ ModifiedDT = true;
return true;
}
-bool CodeGenPrepare::combineToUSubWithOverflow(CmpInst *Cmp) {
+bool CodeGenPrepare::combineToUSubWithOverflow(CmpInst *Cmp,
+ bool &ModifiedDT) {
// We are not expecting non-canonical/degenerate code. Just bail out.
Value *A = Cmp->getOperand(0), *B = Cmp->getOperand(1);
if (isa<Constant>(A) && isa<Constant>(B))
@@ -1339,6 +1342,8 @@ bool CodeGenPrepare::combineToUSubWithOverflow(CmpInst *Cmp) {
if (!replaceMathCmpWithIntrinsic(Sub, Cmp, Intrinsic::usub_with_overflow))
return false;
+ // Reset callers - do not crash by iterating over a dead instruction.
+ ModifiedDT = true;
return true;
}
@@ -1408,14 +1413,14 @@ static bool sinkCmpExpression(CmpInst *Cmp, const TargetLowering &TLI) {
return MadeChange;
}
-bool CodeGenPrepare::optimizeCmp(CmpInst *Cmp) {
+bool CodeGenPrepare::optimizeCmp(CmpInst *Cmp, bool &ModifiedDT) {
if (sinkCmpExpression(Cmp, *TLI))
return true;
- if (combineToUAddWithOverflow(Cmp))
+ if (combineToUAddWithOverflow(Cmp, ModifiedDT))
return true;
- if (combineToUSubWithOverflow(Cmp))
+ if (combineToUSubWithOverflow(Cmp, ModifiedDT))
return true;
return false;
@@ -6940,7 +6945,7 @@ bool CodeGenPrepare::optimizeInst(Instruction *I, bool &ModifiedDT) {
}
if (auto *Cmp = dyn_cast<CmpInst>(I))
- if (TLI && optimizeCmp(Cmp))
+ if (TLI && optimizeCmp(Cmp, ModifiedDT))
return true;
if (LoadInst *LI = dyn_cast<LoadInst>(I)) {
OpenPOWER on IntegriCloud