Files
llvm/bolt/Passes/AllocCombiner.cpp
Rafael Auler d850ca3622 [BOLT] Add shrink wrapping pass
Summary:
Add an implementation for shrink wrapping, a frame optimization
that moves callee-saved register spills from hot prologues to cold
successors.

(cherry picked from FBD4983706)
2017-05-01 16:52:54 -07:00

117 lines
3.2 KiB
C++

#include "AllocCombiner.h"
#define DEBUG_TYPE "alloccombiner"
using namespace llvm;
namespace opts {
extern bool shouldProcess(const bolt::BinaryFunction &Function);
extern cl::opt<bolt::FrameOptimizationType> FrameOptimization;
} // end namespace opts
namespace llvm {
namespace bolt {
namespace {
bool getStackAdjustmentSize(const BinaryContext &BC, const MCInst &Inst,
int64_t &Adjustment) {
return BC.MIA->evaluateSimple(Inst, Adjustment,
std::make_pair(BC.MIA->getStackPointer(), 0LL),
std::make_pair(0, 0LL));
}
bool isIndifferentToSP(const MCInst &Inst, const BinaryContext &BC) {
if (BC.MIA->isCFI(Inst))
return true;
const auto II = BC.MII->get(Inst.getOpcode());
if (BC.MIA->isTerminator(Inst) ||
II.hasImplicitDefOfPhysReg(BC.MIA->getStackPointer(), BC.MRI.get()) ||
II.hasImplicitUseOfPhysReg(BC.MIA->getStackPointer()))
return false;
for (int I = 0, E = Inst.getNumOperands(); I != E; ++I) {
const auto &Operand = Inst.getOperand(I);
if (Operand.isReg() && Operand.getReg() == BC.MIA->getStackPointer()) {
return false;
}
}
return true;
}
bool shouldProc(BinaryFunction &Function) {
return Function.isSimple() && Function.hasCFG() &&
opts::shouldProcess(Function) && (Function.getSize() > 0);
}
void runForAllWeCare(std::map<uint64_t, BinaryFunction> &BFs,
std::function<void(BinaryFunction &)> Task) {
for (auto &It : BFs) {
auto &Function = It.second;
if (shouldProc(Function))
Task(Function);
}
}
} // end anonymous namespace
void AllocCombinerPass::combineAdjustments(BinaryContext &BC,
BinaryFunction &BF) {
for (auto &BB : BF) {
MCInst *Prev = nullptr;
for (auto I = BB.rbegin(), E = BB.rend(); I != E; ++I) {
auto &Inst = *I;
if (isIndifferentToSP(Inst, BC))
continue; // Skip updating Prev
int64_t Adjustment{0LL};
if (!Prev || !BC.MIA->isStackAdjustment(Inst) ||
!BC.MIA->isStackAdjustment(*Prev) ||
!getStackAdjustmentSize(BC, *Prev, Adjustment)) {
Prev = &Inst;
continue;
}
DEBUG({
dbgs() << "At \"" << BF.getPrintName() << "\", combining: \n";
Inst.dump();
Prev->dump();
dbgs() << "Adjustment: " << Adjustment << "\n";
});
if (BC.MIA->isSUB(Inst))
Adjustment = -Adjustment;
BC.MIA->addToImm(Inst, Adjustment, BC.Ctx.get());
DEBUG({
dbgs() << "After adjustment:\n";
Inst.dump();
});
BB.eraseInstruction(Prev);
++NumCombined;
Prev = &Inst;
}
}
}
void AllocCombinerPass::runOnFunctions(BinaryContext &BC,
std::map<uint64_t, BinaryFunction> &BFs,
std::set<uint64_t> &LargeFunctions) {
if (opts::FrameOptimization == FOP_NONE)
return;
runForAllWeCare(
BFs, [&](BinaryFunction &Function) { combineAdjustments(BC, Function); });
outs() << "BOLT-INFO: Allocation combiner: " << NumCoalesced
<< " empty spaces coalesced.\n";
}
} // end namespace bolt
} // end namespace llvm