Newer
Older
//===- DeadStoreElimination.cpp - Fast Dead Store Elimination -------------===//
Owen Anderson
committed
//
// The LLVM Compiler Infrastructure
//
// This file is distributed under the University of Illinois Open Source
// License. See LICENSE.TXT for details.
Owen Anderson
committed
//
//===----------------------------------------------------------------------===//
//
// This file implements a trivial dead store elimination that only considers
// basic-block local redundant stores.
//
// FIXME: This should eventually be extended to be a post-dominator tree
// traversal. Doing so would be pretty trivial.
//
//===----------------------------------------------------------------------===//
Owen Anderson
committed
#include "llvm/Transforms/Scalar.h"
Owen Anderson
committed
#include "llvm/Constants.h"
Owen Anderson
committed
#include "llvm/Function.h"
#include "llvm/Instructions.h"
Owen Anderson
committed
#include "llvm/IntrinsicInst.h"
Owen Anderson
committed
#include "llvm/Pass.h"
#include "llvm/ADT/SmallPtrSet.h"
#include "llvm/ADT/Statistic.h"
Owen Anderson
committed
#include "llvm/Analysis/AliasAnalysis.h"
Owen Anderson
committed
#include "llvm/Analysis/Dominators.h"
#include "llvm/Analysis/MemoryBuiltins.h"
Owen Anderson
committed
#include "llvm/Analysis/MemoryDependenceAnalysis.h"
Owen Anderson
committed
#include "llvm/Target/TargetData.h"
Owen Anderson
committed
#include "llvm/Transforms/Utils/Local.h"
using namespace llvm;
STATISTIC(NumFastStores, "Number of stores deleted");
STATISTIC(NumFastOther , "Number of other instrs removed");
namespace {
struct DSE : public FunctionPass {
TargetData *TD;
Owen Anderson
committed
static char ID; // Pass identification, replacement for typeid
Owen Anderson
committed
DSE() : FunctionPass(ID) {
initializeDSEPass(*PassRegistry::getPassRegistry());
}
Owen Anderson
committed
virtual bool runOnFunction(Function &F) {
bool Changed = false;
DominatorTree &DT = getAnalysis<DominatorTree>();
Owen Anderson
committed
for (Function::iterator I = F.begin(), E = F.end(); I != E; ++I)
// Only check non-dead blocks. Dead blocks may have strange pointer
// cycles that will confuse alias analysis.
if (DT.isReachableFromEntry(I))
Changed |= runOnBasicBlock(*I);
Owen Anderson
committed
return Changed;
}
Owen Anderson
committed
bool runOnBasicBlock(BasicBlock &BB);
bool HandleFree(CallInst *F);
bool handleEndBlock(BasicBlock &BB);
bool RemoveUndeadPointers(Value *Ptr, uint64_t killPointerSize,
BasicBlock::iterator &BBI,
SmallPtrSet<Value*, 64> &deadPointers);
void DeleteDeadInstruction(Instruction *I,
SmallPtrSet<Value*, 64> *deadPointers = 0);
Owen Anderson
committed
// getAnalysisUsage - We require post dominance frontiers (aka Control
// Dependence Graph)
virtual void getAnalysisUsage(AnalysisUsage &AU) const {
AU.setPreservesCFG();
Owen Anderson
committed
AU.addRequired<DominatorTree>();
Owen Anderson
committed
AU.addRequired<AliasAnalysis>();
Owen Anderson
committed
AU.addRequired<MemoryDependenceAnalysis>();
Owen Anderson
committed
AU.addPreserved<DominatorTree>();
Owen Anderson
committed
AU.addPreserved<MemoryDependenceAnalysis>();
}
uint64_t getPointerSize(Value *V) const;
Owen Anderson
committed
};
}
char DSE::ID = 0;
Owen Anderson
committed
INITIALIZE_PASS_BEGIN(DSE, "dse", "Dead Store Elimination", false, false)
INITIALIZE_PASS_DEPENDENCY(DominatorTree)
INITIALIZE_PASS_DEPENDENCY(MemoryDependenceAnalysis)
INITIALIZE_AG_DEPENDENCY(AliasAnalysis)
INITIALIZE_PASS_END(DSE, "dse", "Dead Store Elimination", false, false)
FunctionPass *llvm::createDeadStoreEliminationPass() { return new DSE(); }
Owen Anderson
committed
/// hasMemoryWrite - Does this instruction write some memory? This only returns
/// true for things that we can analyze with other helpers below.
static bool hasMemoryWrite(Instruction *I) {
Nick Lewycky
committed
if (isa<StoreInst>(I))
return true;
if (IntrinsicInst *II = dyn_cast<IntrinsicInst>(I)) {
switch (II->getIntrinsicID()) {
default:
return false;
case Intrinsic::memset:
case Intrinsic::memmove:
case Intrinsic::memcpy:
case Intrinsic::init_trampoline:
case Intrinsic::lifetime_end:
return true;
Nick Lewycky
committed
}
}
return false;
}
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
/// getLocForWrite - Return a Location stored to by the specified instruction.
static AliasAnalysis::Location
getLocForWrite(Instruction *Inst, AliasAnalysis &AA) {
if (StoreInst *SI = dyn_cast<StoreInst>(Inst))
return AA.getLocation(SI);
if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(Inst)) {
// memcpy/memmove/memset.
AliasAnalysis::Location Loc = AA.getLocationForDest(MI);
// If we don't have target data around, an unknown size in Location means
// that we should use the size of the pointee type. This isn't valid for
// memset/memcpy, which writes more than an i8.
if (Loc.Size == AliasAnalysis::UnknownSize && AA.getTargetData() == 0)
return AliasAnalysis::Location();
return Loc;
}
IntrinsicInst *II = dyn_cast<IntrinsicInst>(Inst);
if (II == 0) return AliasAnalysis::Location();
switch (II->getIntrinsicID()) {
default: return AliasAnalysis::Location(); // Unhandled intrinsic.
case Intrinsic::init_trampoline:
// If we don't have target data around, an unknown size in Location means
// that we should use the size of the pointee type. This isn't valid for
// init.trampoline, which writes more than an i8.
if (AA.getTargetData() == 0) return AliasAnalysis::Location();
// FIXME: We don't know the size of the trampoline, so we can't really
// handle it here.
return AliasAnalysis::Location(II->getArgOperand(0));
case Intrinsic::lifetime_end: {
uint64_t Len = cast<ConstantInt>(II->getArgOperand(0))->getZExtValue();
return AliasAnalysis::Location(II->getArgOperand(1), Len);
}
}
}
/// isRemovable - If the value of this instruction and the memory it writes to
/// is unused, may we delete this instruction?
static bool isRemovable(Instruction *I) {
assert(hasMemoryWrite(I));
Nick Lewycky
committed
if (IntrinsicInst *II = dyn_cast<IntrinsicInst>(I))
return II->getIntrinsicID() != Intrinsic::lifetime_end;
if (StoreInst *SI = dyn_cast<StoreInst>(I))
return !SI->isVolatile();
return true;
}
/// getPointerOperand - Return the pointer that is being written to.
Nick Lewycky
committed
static Value *getPointerOperand(Instruction *I) {
assert(hasMemoryWrite(I));
Nick Lewycky
committed
if (StoreInst *SI = dyn_cast<StoreInst>(I))
return SI->getPointerOperand();
if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(I))
Gabor Greif
committed
return MI->getArgOperand(0);
IntrinsicInst *II = cast<IntrinsicInst>(I);
switch (II->getIntrinsicID()) {
default: assert(false && "Unexpected intrinsic!");
case Intrinsic::init_trampoline:
Gabor Greif
committed
return II->getArgOperand(0);
case Intrinsic::lifetime_end:
Gabor Greif
committed
return II->getArgOperand(1);
Nick Lewycky
committed
}
/// isCompleteOverwrite - Return true if a store to the 'Later' location
/// completely overwrites a store to the 'Earlier' location.
static bool isCompleteOverwrite(const AliasAnalysis::Location &Later,
const AliasAnalysis::Location &Earlier,
AliasAnalysis &AA, const TargetData *TD) {
const Value *P1 = Later.Ptr->stripPointerCasts();
const Value *P2 = Earlier.Ptr->stripPointerCasts();
// Make sure that the start pointers are the same.
if (P1 != P2)
return false;
Nick Lewycky
committed
// If we have no TargetData information around, then the size of the store is
// inferrable from the pointee type. If they are the same type, then we know
// that the store is safe.
if (TD == 0)
return Later.Ptr->getType() == Earlier.Ptr->getType();
// Make sure that the Later size is >= the Earlier size.
if (Later.Size < Earlier.Size)
return false;
return true;
MemoryDependenceAnalysis &MD = getAnalysis<MemoryDependenceAnalysis>();
AliasAnalysis &AA = getAnalysis<AliasAnalysis>();
TD = getAnalysisIfAvailable<TargetData>();
Owen Anderson
committed
bool MadeChange = false;
// Do a top-down walk on the BB.
for (BasicBlock::iterator BBI = BB.begin(), BBE = BB.end(); BBI != BBE; ) {
Instruction *Inst = BBI++;
// Handle 'free' calls specially.
if (CallInst *F = isFreeCall(Inst)) {
MadeChange |= HandleFree(F);
continue;
}
// If we find something that writes memory, get its memory dependence.
if (!hasMemoryWrite(Inst))
MemDepResult InstDep = MD.getDependency(Inst);
// Ignore non-local store liveness.
// FIXME: cross-block DSE would be fun. :)
if (InstDep.isNonLocal() ||
// Ignore self dependence, which happens in the entry block of the
// function.
InstDep.getInst() == Inst)
continue;
// If we're storing the same value back to a pointer that we just
// loaded from, then the store can be removed.
if (StoreInst *SI = dyn_cast<StoreInst>(Inst)) {
if (LoadInst *DepLoad = dyn_cast<LoadInst>(InstDep.getInst())) {
if (SI->getPointerOperand() == DepLoad->getPointerOperand() &&
SI->getOperand(0) == DepLoad && !SI->isVolatile()) {
// DeleteDeadInstruction can delete the current instruction. Save BBI
// in case we need it.
WeakVH NextInst(BBI);
DeleteDeadInstruction(SI);
if (NextInst == 0) // Next instruction deleted.
BBI = BB.begin();
else if (BBI != BB.begin()) // Revisit this instruction if possible.
--BBI;
++NumFastStores;
MadeChange = true;
continue;
}
}
}
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
// Figure out what location is being stored to.
AliasAnalysis::Location Loc = getLocForWrite(Inst, AA);
// If we didn't get a useful location, fail.
if (Loc.Ptr == 0)
continue;
while (!InstDep.isNonLocal()) {
// Get the memory clobbered by the instruction we depend on. MemDep will
// skip any instructions that 'Loc' clearly doesn't interact with. If we
// end up depending on a may- or must-aliased load, then we can't optimize
// away the store and we bail out. However, if we depend on on something
// that overwrites the memory location we *can* potentially optimize it.
//
// Find out what memory location the dependant instruction stores.
Instruction *DepWrite = InstDep.getInst();
AliasAnalysis::Location DepLoc = getLocForWrite(DepWrite, AA);
// If we didn't get a useful location, or if it isn't a size, bail out.
if (DepLoc.Ptr == 0)
break;
// If we find a removable write that is completely obliterated by the
// store to 'Loc' then we can remove it.
if (isRemovable(DepWrite) && isCompleteOverwrite(Loc, DepLoc, AA, TD)) {
// Delete the store and now-dead instructions that feed it.
DeleteDeadInstruction(DepWrite);
++NumFastStores;
MadeChange = true;
// DeleteDeadInstruction can delete the current instruction in loop
// cases, reset BBI.
BBI = Inst;
if (BBI != BB.begin())
--BBI;
break;
}
// If this is a may-aliased store that is clobbering the store value, we
// can keep searching past it for another must-aliased pointer that stores
// to the same location. For example, in:
// store -> P
// store -> Q
// store -> P
// we can remove the first store to P even though we don't know if P and Q
// alias.
if (DepWrite == &BB.front()) break;
// Can't look past this instruction if it might read 'Loc'.
if (AA.getModRefInfo(DepWrite, Loc) & AliasAnalysis::Ref)
break;
InstDep = MD.getPointerDependencyFrom(Loc, false, DepWrite, &BB);
Nick Lewycky
committed
}
Owen Anderson
committed
}
// If this block ends in a return, unwind, or unreachable, all allocas are
// dead at its end, which means stores to them are also dead.
Owen Anderson
committed
if (BB.getTerminator()->getNumSuccessors() == 0)
MadeChange |= handleEndBlock(BB);
Owen Anderson
committed
return MadeChange;
}
/// HandleFree - Handle frees of entire structures whose dependency is a store
/// to a field of that structure.
bool DSE::HandleFree(CallInst *F) {
Owen Anderson
committed
AliasAnalysis &AA = getAnalysis<AliasAnalysis>();
MemoryDependenceAnalysis &MD = getAnalysis<MemoryDependenceAnalysis>();
MemDepResult Dep = MD.getDependency(F);
if (Dep.isNonLocal()) return false;
Instruction *Dependency = Dep.getInst();
if (!hasMemoryWrite(Dependency) || !isRemovable(Dependency))
return false;
Owen Anderson
committed
Value *DepPointer = getPointerOperand(Dependency)->getUnderlyingObject();
// Check for aliasing.
if (AA.alias(F->getArgOperand(0), 1, DepPointer, 1) !=
AliasAnalysis::MustAlias)
return false;
Owen Anderson
committed
// DCE instructions only used to calculate that store
DeleteDeadInstruction(Dependency);
++NumFastStores;
// Inst's old Dependency is now deleted. Compute the next dependency,
// which may also be dead, as in
// s[0] = 0;
// s[1] = 0; // This has just been deleted.
// free(s);
Dep = MD.getDependency(F);
} while (!Dep.isNonLocal());
Owen Anderson
committed
}
/// handleEndBlock - Remove dead stores to stack-allocated locations in the
/// function end block. Ex:
/// %A = alloca i32
/// ...
/// store i32 1, i32* %A
/// ret void
bool DSE::handleEndBlock(BasicBlock &BB) {
Owen Anderson
committed
AliasAnalysis &AA = getAnalysis<AliasAnalysis>();
bool MadeChange = false;
// Pointers alloca'd in this function are dead in the end block
Owen Anderson
committed
SmallPtrSet<Value*, 64> deadPointers;
Owen Anderson
committed
// Find all of the alloca'd pointers in the entry block.
Owen Anderson
committed
BasicBlock *Entry = BB.getParent()->begin();
for (BasicBlock::iterator I = Entry->begin(), E = Entry->end(); I != E; ++I)
if (AllocaInst *AI = dyn_cast<AllocaInst>(I))
deadPointers.insert(AI);
// Treat byval arguments the same, stores to them are dead at the end of the
// function.
Owen Anderson
committed
for (Function::arg_iterator AI = BB.getParent()->arg_begin(),
AE = BB.getParent()->arg_end(); AI != AE; ++AI)
if (AI->hasByValAttr())
deadPointers.insert(AI);
Owen Anderson
committed
// Scan the basic block backwards
for (BasicBlock::iterator BBI = BB.end(); BBI != BB.begin(); ){
--BBI;
// If we find a store whose pointer is dead.
if (hasMemoryWrite(BBI)) {
if (isRemovable(BBI)) {
Nick Lewycky
committed
Value *pointerOperand = getPointerOperand(BBI)->getUnderlyingObject();
Owen Anderson
committed
// Alloca'd pointers or byval arguments (which are functionally like
// alloca's) are valid candidates for removal.
Owen Anderson
committed
if (deadPointers.count(pointerOperand)) {
// DCE instructions only used to calculate that store.
Nick Lewycky
committed
Instruction *Dead = BBI;
++BBI;
Nick Lewycky
committed
DeleteDeadInstruction(Dead, &deadPointers);
++NumFastStores;
Nick Lewycky
committed
continue;
Owen Anderson
committed
}
Nick Lewycky
committed
// Because a memcpy or memmove is also a load, we can't skip it if we
// didn't remove it.
if (!isa<MemTransferInst>(BBI))
Owen Anderson
committed
continue;
}
Value *killPointer = 0;
uint64_t killPointerSize = AliasAnalysis::UnknownSize;
Owen Anderson
committed
// If we encounter a use of the pointer, it is no longer considered dead
if (LoadInst *L = dyn_cast<LoadInst>(BBI)) {
// However, if this load is unused and not volatile, we can go ahead and
// remove it, and not have to worry about it making our pointer undead!
if (L->use_empty() && !L->isVolatile()) {
++BBI;
DeleteDeadInstruction(L, &deadPointers);
++NumFastOther;
Owen Anderson
committed
MadeChange = true;
continue;
}
Owen Anderson
committed
killPointer = L->getPointerOperand();
} else if (VAArgInst *V = dyn_cast<VAArgInst>(BBI)) {
Owen Anderson
committed
killPointer = V->getOperand(0);
Nick Lewycky
committed
} else if (isa<MemTransferInst>(BBI) &&
isa<ConstantInt>(cast<MemTransferInst>(BBI)->getLength())) {
killPointer = cast<MemTransferInst>(BBI)->getSource();
killPointerSize = cast<ConstantInt>(
Nick Lewycky
committed
cast<MemTransferInst>(BBI)->getLength())->getZExtValue();
} else if (AllocaInst *A = dyn_cast<AllocaInst>(BBI)) {
Owen Anderson
committed
deadPointers.erase(A);
Owen Anderson
committed
// Dead alloca's can be DCE'd when we reach them
Nick Lewycky
committed
if (A->use_empty()) {
++BBI;
DeleteDeadInstruction(A, &deadPointers);
++NumFastOther;
Owen Anderson
committed
MadeChange = true;
}
Owen Anderson
committed
continue;
Gabor Greif
committed
} else if (CallSite CS = cast<Value>(BBI)) {
Owen Anderson
committed
// If this call does not access memory, it can't
// be undeadifying any of our pointers.
if (AA.doesNotAccessMemory(CS))
Owen Anderson
committed
continue;
unsigned modRef = 0;
unsigned other = 0;
Owen Anderson
committed
// Remove any pointers made undead by the call from the dead set
Owen Anderson
committed
std::vector<Value*> dead;
for (SmallPtrSet<Value*, 64>::iterator I = deadPointers.begin(),
Owen Anderson
committed
E = deadPointers.end(); I != E; ++I) {
// HACK: if we detect that our AA is imprecise, it's not
// worth it to scan the rest of the deadPointers set. Just
// assume that the AA will return ModRef for everything, and
// go ahead and bail.
if (modRef >= 16 && other == 0) {
deadPointers.clear();
return MadeChange;
}
Owen Anderson
committed
// See if the call site touches it
AliasAnalysis::ModRefResult A = AA.getModRefInfo(CS, *I,
getPointerSize(*I));
if (A == AliasAnalysis::ModRef)
++modRef;
else
++other;
Owen Anderson
committed
if (A == AliasAnalysis::ModRef || A == AliasAnalysis::Ref)
Owen Anderson
committed
dead.push_back(*I);
}
Owen Anderson
committed
for (std::vector<Value*>::iterator I = dead.begin(), E = dead.end();
Owen Anderson
committed
I != E; ++I)
Owen Anderson
committed
deadPointers.erase(*I);
Owen Anderson
committed
continue;
} else if (isInstructionTriviallyDead(BBI)) {
Owen Anderson
committed
// For any non-memory-affecting non-terminators, DCE them as we reach them
Instruction *Inst = BBI;
++BBI;
DeleteDeadInstruction(Inst, &deadPointers);
++NumFastOther;
MadeChange = true;
continue;
Owen Anderson
committed
}
if (!killPointer)
continue;
killPointer = killPointer->getUnderlyingObject();
Owen Anderson
committed
// Deal with undead pointers
MadeChange |= RemoveUndeadPointers(killPointer, killPointerSize, BBI,
deadPointers);
Owen Anderson
committed
}
return MadeChange;
}
/// RemoveUndeadPointers - check for uses of a pointer that make it
/// undead when scanning for dead stores to alloca's.
bool DSE::RemoveUndeadPointers(Value *killPointer, uint64_t killPointerSize,
BasicBlock::iterator &BBI,
SmallPtrSet<Value*, 64> &deadPointers) {
Owen Anderson
committed
AliasAnalysis &AA = getAnalysis<AliasAnalysis>();
// If the kill pointer can be easily reduced to an alloca,
// don't bother doing extraneous AA queries.
Owen Anderson
committed
if (deadPointers.count(killPointer)) {
deadPointers.erase(killPointer);
return false;
}
// A global can't be in the dead pointer set.
if (isa<GlobalValue>(killPointer))
return false;
Owen Anderson
committed
bool MadeChange = false;
SmallVector<Value*, 16> undead;
Owen Anderson
committed
for (SmallPtrSet<Value*, 64>::iterator I = deadPointers.begin(),
E = deadPointers.end(); I != E; ++I) {
Owen Anderson
committed
// See if this pointer could alias it
AliasAnalysis::AliasResult A = AA.alias(*I, getPointerSize(*I),
killPointer, killPointerSize);
Owen Anderson
committed
// If it must-alias and a store, we can delete it
if (isa<StoreInst>(BBI) && A == AliasAnalysis::MustAlias) {
StoreInst *S = cast<StoreInst>(BBI);
Owen Anderson
committed
// Remove it!
++BBI;
DeleteDeadInstruction(S, &deadPointers);
++NumFastStores;
Owen Anderson
committed
MadeChange = true;
continue;
// Otherwise, it is undead
} else if (A != AliasAnalysis::NoAlias)
undead.push_back(*I);
Owen Anderson
committed
}
for (SmallVector<Value*, 16>::iterator I = undead.begin(), E = undead.end();
Owen Anderson
committed
I != E; ++I)
Owen Anderson
committed
deadPointers.erase(*I);
Owen Anderson
committed
return MadeChange;
}
/// DeleteDeadInstruction - Delete this instruction. Before we do, go through
/// and zero out all the operands of this instruction. If any of them become
/// dead, delete them and the computation tree that feeds them.
///
/// If ValueSet is non-null, remove any deleted instructions from it as well.
///
void DSE::DeleteDeadInstruction(Instruction *I,
SmallPtrSet<Value*, 64> *ValueSet) {
SmallVector<Instruction*, 32> NowDeadInsts;
NowDeadInsts.push_back(I);
--NumFastOther;
// Before we touch this instruction, remove it from memdep!
MemoryDependenceAnalysis &MDA = getAnalysis<MemoryDependenceAnalysis>();
do {
Instruction *DeadInst = NowDeadInsts.pop_back_val();
++NumFastOther;
// This instruction is dead, zap it, in stages. Start by removing it from
// MemDep, which needs to know the operands and needs it to be in the
// function.
MDA.removeInstruction(DeadInst);
for (unsigned op = 0, e = DeadInst->getNumOperands(); op != e; ++op) {
Value *Op = DeadInst->getOperand(op);
DeadInst->setOperand(op, 0);
// If this operand just became dead, add it to the NowDeadInsts list.
if (!Op->use_empty()) continue;
if (Instruction *OpI = dyn_cast<Instruction>(Op))
if (isInstructionTriviallyDead(OpI))
NowDeadInsts.push_back(OpI);
}
DeadInst->eraseFromParent();
if (ValueSet) ValueSet->erase(DeadInst);
} while (!NowDeadInsts.empty());
Owen Anderson
committed
}
uint64_t DSE::getPointerSize(Value *V) const {
if (TD) {
if (AllocaInst *A = dyn_cast<AllocaInst>(V)) {
// Get size information for the alloca
if (ConstantInt *C = dyn_cast<ConstantInt>(A->getArraySize()))
return C->getZExtValue() * TD->getTypeAllocSize(A->getAllocatedType());
} else {
assert(isa<Argument>(V) && "Expected AllocaInst or Argument!");
const PointerType *PT = cast<PointerType>(V->getType());
return TD->getTypeAllocSize(PT->getElementType());
}
}
return AliasAnalysis::UnknownSize;