@@ -383,12 +383,12 @@ enum OverwriteResult {
383
383
// / write to the same underlying object. In that case, use isPartialOverwrite to
384
384
// / check if \p Later partially overwrites \p Earlier. Returns 'OW_Unknown' if
385
385
// / nothing can be determined.
386
- static OverwriteResult isOverwrite ( const MemoryLocation &Later,
387
- const MemoryLocation &Earlier,
388
- const DataLayout &DL ,
389
- const TargetLibraryInfo &TLI,
390
- int64_t &EarlierOff, int64_t &LaterOff,
391
- AliasAnalysis &AA, const Function *F) {
386
+ template < typename AATy>
387
+ static OverwriteResult
388
+ isOverwrite ( const MemoryLocation &Later, const MemoryLocation &Earlier ,
389
+ const DataLayout &DL, const TargetLibraryInfo &TLI,
390
+ int64_t &EarlierOff, int64_t &LaterOff, AATy &AA ,
391
+ const Function *F) {
392
392
// FIXME: Vet that this works for size upper-bounds. Seems unlikely that we'll
393
393
// get imprecise values here, though (except for unknown sizes).
394
394
if (!Later.Size .isPrecise () || !Earlier.Size .isPrecise ())
@@ -643,11 +643,10 @@ static bool isPossibleSelfRead(Instruction *Inst,
643
643
// / modified between the first and the second instruction.
644
644
// / Precondition: Second instruction must be dominated by the first
645
645
// / instruction.
646
- static bool memoryIsNotModifiedBetween (Instruction *FirstI,
647
- Instruction *SecondI,
648
- AliasAnalysis *AA,
649
- const DataLayout &DL,
650
- DominatorTree *DT) {
646
+ template <typename AATy>
647
+ static bool
648
+ memoryIsNotModifiedBetween (Instruction *FirstI, Instruction *SecondI, AATy &AA,
649
+ const DataLayout &DL, DominatorTree *DT) {
651
650
// Do a backwards scan through the CFG from SecondI to FirstI. Look for
652
651
// instructions which can modify the memory location accessed by SecondI.
653
652
//
@@ -696,7 +695,7 @@ static bool memoryIsNotModifiedBetween(Instruction *FirstI,
696
695
for (; BI != EI; ++BI) {
697
696
Instruction *I = &*BI;
698
697
if (I->mayWriteToMemory () && I != SecondI)
699
- if (isModSet (AA-> getModRefInfo (I, MemLoc.getWithNewPtr (Ptr ))))
698
+ if (isModSet (AA. getModRefInfo (I, MemLoc.getWithNewPtr (Ptr ))))
700
699
return false ;
701
700
}
702
701
if (B != FirstBB) {
@@ -1132,7 +1131,7 @@ static bool eliminateNoopStore(Instruction *Inst, BasicBlock::iterator &BBI,
1132
1131
if (LoadInst *DepLoad = dyn_cast<LoadInst>(SI->getValueOperand ())) {
1133
1132
if (SI->getPointerOperand () == DepLoad->getPointerOperand () &&
1134
1133
isRemovable (SI) &&
1135
- memoryIsNotModifiedBetween (DepLoad, SI, AA, DL, DT)) {
1134
+ memoryIsNotModifiedBetween (DepLoad, SI, * AA, DL, DT)) {
1136
1135
1137
1136
LLVM_DEBUG (
1138
1137
dbgs () << " DSE: Remove Store Of Load from same pointer:\n LOAD: "
@@ -1151,7 +1150,7 @@ static bool eliminateNoopStore(Instruction *Inst, BasicBlock::iterator &BBI,
1151
1150
dyn_cast<Instruction>(getUnderlyingObject (SI->getPointerOperand ()));
1152
1151
1153
1152
if (UnderlyingPointer && isCallocLikeFn (UnderlyingPointer, TLI) &&
1154
- memoryIsNotModifiedBetween (UnderlyingPointer, SI, AA, DL, DT)) {
1153
+ memoryIsNotModifiedBetween (UnderlyingPointer, SI, * AA, DL, DT)) {
1155
1154
LLVM_DEBUG (
1156
1155
dbgs () << " DSE: Remove null store to the calloc'ed object:\n DEAD: "
1157
1156
<< *Inst << " \n OBJECT: " << *UnderlyingPointer << ' \n ' );
@@ -1164,11 +1163,10 @@ static bool eliminateNoopStore(Instruction *Inst, BasicBlock::iterator &BBI,
1164
1163
return false ;
1165
1164
}
1166
1165
1167
- static Constant *
1168
- tryToMergePartialOverlappingStores (StoreInst *Earlier, StoreInst *Later,
1169
- int64_t InstWriteOffset,
1170
- int64_t DepWriteOffset, const DataLayout &DL,
1171
- AliasAnalysis *AA, DominatorTree *DT) {
1166
+ template <typename AATy>
1167
+ static Constant *tryToMergePartialOverlappingStores (
1168
+ StoreInst *Earlier, StoreInst *Later, int64_t InstWriteOffset,
1169
+ int64_t DepWriteOffset, const DataLayout &DL, AATy &AA, DominatorTree *DT) {
1172
1170
1173
1171
if (Earlier && isa<ConstantInt>(Earlier->getValueOperand ()) &&
1174
1172
DL.typeSizeEqualsStoreSize (Earlier->getValueOperand ()->getType ()) &&
@@ -1361,7 +1359,7 @@ static bool eliminateDeadStores(BasicBlock &BB, AliasAnalysis *AA,
1361
1359
auto *Earlier = dyn_cast<StoreInst>(DepWrite);
1362
1360
auto *Later = dyn_cast<StoreInst>(Inst);
1363
1361
if (Constant *C = tryToMergePartialOverlappingStores (
1364
- Earlier, Later, InstWriteOffset, DepWriteOffset, DL, AA,
1362
+ Earlier, Later, InstWriteOffset, DepWriteOffset, DL, * AA,
1365
1363
DT)) {
1366
1364
auto *SI = new StoreInst (
1367
1365
C, Earlier->getPointerOperand (), false , Earlier->getAlign (),
@@ -1507,6 +1505,16 @@ bool canSkipDef(MemoryDef *D, bool DefVisibleToCaller) {
1507
1505
struct DSEState {
1508
1506
Function &F;
1509
1507
AliasAnalysis &AA;
1508
+
1509
+ // / The single BatchAA instance that is used to cache AA queries. It will
1510
+ // / not be invalidated over the whole run. This is safe, because:
1511
+ // / 1. Only memory writes are removed, so the alias cache for memory
1512
+ // / locations remains valid.
1513
+ // / 2. No new instructions are added (only instructions removed), so cached
1514
+ // / information for a deleted value cannot be accessed by a re-used new
1515
+ // / value pointer.
1516
+ BatchAAResults BatchAA;
1517
+
1510
1518
MemorySSA &MSSA;
1511
1519
DominatorTree &DT;
1512
1520
PostDominatorTree &PDT;
@@ -1534,7 +1542,7 @@ struct DSEState {
1534
1542
1535
1543
DSEState (Function &F, AliasAnalysis &AA, MemorySSA &MSSA, DominatorTree &DT,
1536
1544
PostDominatorTree &PDT, const TargetLibraryInfo &TLI)
1537
- : F(F), AA(AA), MSSA(MSSA), DT(DT), PDT(PDT), TLI(TLI) {}
1545
+ : F(F), AA(AA), BatchAA(AA), MSSA(MSSA), DT(DT), PDT(PDT), TLI(TLI) {}
1538
1546
1539
1547
static DSEState get (Function &F, AliasAnalysis &AA, MemorySSA &MSSA,
1540
1548
DominatorTree &DT, PostDominatorTree &PDT,
@@ -1623,7 +1631,7 @@ struct DSEState {
1623
1631
}
1624
1632
1625
1633
// / Returns true if \p Use completely overwrites \p DefLoc.
1626
- bool isCompleteOverwrite (MemoryLocation DefLoc, Instruction *UseInst) const {
1634
+ bool isCompleteOverwrite (MemoryLocation DefLoc, Instruction *UseInst) {
1627
1635
// UseInst has a MemoryDef associated in MemorySSA. It's possible for a
1628
1636
// MemoryDef to not write to memory, e.g. a volatile load is modeled as a
1629
1637
// MemoryDef.
@@ -1638,7 +1646,7 @@ struct DSEState {
1638
1646
auto CC = getLocForWriteEx (UseInst);
1639
1647
const DataLayout &DL = F.getParent ()->getDataLayout ();
1640
1648
return CC && isOverwrite (*CC, DefLoc, DL, TLI, DepWriteOffset,
1641
- InstWriteOffset, AA , &F) == OW_Complete;
1649
+ InstWriteOffset, BatchAA , &F) == OW_Complete;
1642
1650
}
1643
1651
1644
1652
// / Returns true if \p Def is not read before returning from the function.
@@ -1717,7 +1725,7 @@ struct DSEState {
1717
1725
1718
1726
// / Returns true if \p MaybeTerm is a memory terminator for the same
1719
1727
// / underlying object as \p DefLoc.
1720
- bool isMemTerminator (MemoryLocation DefLoc, Instruction *MaybeTerm) const {
1728
+ bool isMemTerminator (MemoryLocation DefLoc, Instruction *MaybeTerm) {
1721
1729
Optional<std::pair<MemoryLocation, bool >> MaybeTermLoc =
1722
1730
getLocForTerminator (MaybeTerm);
1723
1731
@@ -1730,19 +1738,19 @@ struct DSEState {
1730
1738
DataLayout DL = MaybeTerm->getParent ()->getModule ()->getDataLayout ();
1731
1739
DefLoc = MemoryLocation (getUnderlyingObject (DefLoc.Ptr ));
1732
1740
}
1733
- return AA .isMustAlias (MaybeTermLoc->first , DefLoc);
1741
+ return BatchAA .isMustAlias (MaybeTermLoc->first , DefLoc);
1734
1742
}
1735
1743
1736
1744
// Returns true if \p Use may read from \p DefLoc.
1737
- bool isReadClobber (MemoryLocation DefLoc, Instruction *UseInst) const {
1745
+ bool isReadClobber (MemoryLocation DefLoc, Instruction *UseInst) {
1738
1746
if (!UseInst->mayReadFromMemory ())
1739
1747
return false ;
1740
1748
1741
1749
if (auto *CB = dyn_cast<CallBase>(UseInst))
1742
1750
if (CB->onlyAccessesInaccessibleMemory ())
1743
1751
return false ;
1744
1752
1745
- ModRefInfo MR = AA .getModRefInfo (UseInst, DefLoc);
1753
+ ModRefInfo MR = BatchAA .getModRefInfo (UseInst, DefLoc);
1746
1754
// If necessary, perform additional analysis.
1747
1755
if (isRefSet (MR))
1748
1756
MR = AA.callCapturesBefore (UseInst, DefLoc, &DT);
@@ -1758,7 +1766,7 @@ struct DSEState {
1758
1766
Optional<MemoryAccess *>
1759
1767
getDomMemoryDef (MemoryDef *KillingDef, MemoryAccess *Current,
1760
1768
MemoryLocation DefLoc, bool DefVisibleToCallerBeforeRet,
1761
- bool DefVisibleToCallerAfterRet, unsigned &ScanLimit) const {
1769
+ bool DefVisibleToCallerAfterRet, unsigned &ScanLimit) {
1762
1770
if (ScanLimit == 0 ) {
1763
1771
LLVM_DEBUG (dbgs () << " \n ... hit scan limit\n " );
1764
1772
return None;
@@ -2285,7 +2293,7 @@ bool eliminateDeadStoresMemorySSA(Function &F, AliasAnalysis &AA,
2285
2293
// Check if NI overwrites SI.
2286
2294
int64_t InstWriteOffset, DepWriteOffset;
2287
2295
OverwriteResult OR = isOverwrite (SILoc, NILoc, DL, TLI, DepWriteOffset,
2288
- InstWriteOffset, State.AA , &F);
2296
+ InstWriteOffset, State.BatchAA , &F);
2289
2297
if (OR == OW_MaybePartial) {
2290
2298
auto Iter = State.IOLs .insert (
2291
2299
std::make_pair<BasicBlock *, InstOverlapIntervalsTy>(
@@ -2303,8 +2311,8 @@ bool eliminateDeadStoresMemorySSA(Function &F, AliasAnalysis &AA,
2303
2311
// TODO: implement tryToMergeParialOverlappingStores using MemorySSA.
2304
2312
if (Earlier && Later && DT.dominates (Earlier, Later)) {
2305
2313
if (Constant *Merged = tryToMergePartialOverlappingStores (
2306
- Earlier, Later, InstWriteOffset, DepWriteOffset, DL, &AA,
2307
- &DT)) {
2314
+ Earlier, Later, InstWriteOffset, DepWriteOffset, DL,
2315
+ State. BatchAA , &DT)) {
2308
2316
2309
2317
// Update stored value of earlier store to merged constant.
2310
2318
Earlier->setOperand (0 , Merged);
0 commit comments