You've already forked linux-packaging-mono
Imported Upstream version 5.18.0.225
Former-commit-id: 10196d987d5fc5564b9d3b33b1fdf13190f4d0b5
This commit is contained in:
parent
32d52ae4ca
commit
f32dbaf0b2
208
external/llvm/lib/Target/AMDGPU/AMDGPUInline.cpp
vendored
Normal file
208
external/llvm/lib/Target/AMDGPU/AMDGPUInline.cpp
vendored
Normal file
@ -0,0 +1,208 @@
|
||||
//===- AMDGPUInline.cpp - Code to perform simple function inlining --------===//
|
||||
//
|
||||
// The LLVM Compiler Infrastructure
|
||||
//
|
||||
// This file is distributed under the University of Illinois Open Source
|
||||
// License. See LICENSE.TXT for details.
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
//
|
||||
/// \file
|
||||
/// \brief This is AMDGPU specific replacement of the standard inliner.
|
||||
/// The main purpose is to account for the fact that calls not only expensive
|
||||
/// on the AMDGPU, but much more expensive if a private memory pointer is
|
||||
/// passed to a function as an argument. In this situation, we are unable to
|
||||
/// eliminate private memory in the caller unless inlined and end up with slow
|
||||
/// and expensive scratch access. Thus, we boost the inline threshold for such
|
||||
/// functions here.
|
||||
///
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
|
||||
#include "AMDGPU.h"
|
||||
#include "llvm/Transforms/IPO.h"
|
||||
#include "llvm/Analysis/AssumptionCache.h"
|
||||
#include "llvm/Analysis/CallGraph.h"
|
||||
#include "llvm/Analysis/InlineCost.h"
|
||||
#include "llvm/Analysis/ValueTracking.h"
|
||||
#include "llvm/Analysis/TargetTransformInfo.h"
|
||||
#include "llvm/IR/CallSite.h"
|
||||
#include "llvm/IR/DataLayout.h"
|
||||
#include "llvm/IR/Instructions.h"
|
||||
#include "llvm/IR/Module.h"
|
||||
#include "llvm/IR/Type.h"
|
||||
#include "llvm/Support/CommandLine.h"
|
||||
#include "llvm/Support/Debug.h"
|
||||
#include "llvm/Transforms/IPO/Inliner.h"
|
||||
|
||||
using namespace llvm;
|
||||
|
||||
#define DEBUG_TYPE "inline"
|
||||
|
||||
static cl::opt<int>
|
||||
ArgAllocaCost("amdgpu-inline-arg-alloca-cost", cl::Hidden, cl::init(2200),
|
||||
cl::desc("Cost of alloca argument"));
|
||||
|
||||
// If the amount of scratch memory to eliminate exceeds our ability to allocate
|
||||
// it into registers we gain nothing by agressively inlining functions for that
|
||||
// heuristic.
|
||||
static cl::opt<unsigned>
|
||||
ArgAllocaCutoff("amdgpu-inline-arg-alloca-cutoff", cl::Hidden, cl::init(256),
|
||||
cl::desc("Maximum alloca size to use for inline cost"));
|
||||
|
||||
namespace {
|
||||
|
||||
class AMDGPUInliner : public LegacyInlinerBase {
|
||||
|
||||
public:
|
||||
AMDGPUInliner() : LegacyInlinerBase(ID) {
|
||||
initializeAMDGPUInlinerPass(*PassRegistry::getPassRegistry());
|
||||
Params = getInlineParams();
|
||||
}
|
||||
|
||||
static char ID; // Pass identification, replacement for typeid
|
||||
|
||||
unsigned getInlineThreshold(CallSite CS) const;
|
||||
|
||||
InlineCost getInlineCost(CallSite CS) override;
|
||||
|
||||
bool runOnSCC(CallGraphSCC &SCC) override;
|
||||
|
||||
void getAnalysisUsage(AnalysisUsage &AU) const override;
|
||||
|
||||
private:
|
||||
TargetTransformInfoWrapperPass *TTIWP;
|
||||
|
||||
InlineParams Params;
|
||||
};
|
||||
|
||||
} // end anonymous namespace
|
||||
|
||||
char AMDGPUInliner::ID = 0;
|
||||
INITIALIZE_PASS_BEGIN(AMDGPUInliner, "amdgpu-inline",
|
||||
"AMDGPU Function Integration/Inlining", false, false)
|
||||
INITIALIZE_PASS_DEPENDENCY(AssumptionCacheTracker)
|
||||
INITIALIZE_PASS_DEPENDENCY(CallGraphWrapperPass)
|
||||
INITIALIZE_PASS_DEPENDENCY(ProfileSummaryInfoWrapperPass)
|
||||
INITIALIZE_PASS_DEPENDENCY(TargetTransformInfoWrapperPass)
|
||||
INITIALIZE_PASS_DEPENDENCY(TargetLibraryInfoWrapperPass)
|
||||
INITIALIZE_PASS_END(AMDGPUInliner, "amdgpu-inline",
|
||||
"AMDGPU Function Integration/Inlining", false, false)
|
||||
|
||||
Pass *llvm::createAMDGPUFunctionInliningPass() { return new AMDGPUInliner(); }
|
||||
|
||||
bool AMDGPUInliner::runOnSCC(CallGraphSCC &SCC) {
|
||||
TTIWP = &getAnalysis<TargetTransformInfoWrapperPass>();
|
||||
return LegacyInlinerBase::runOnSCC(SCC);
|
||||
}
|
||||
|
||||
void AMDGPUInliner::getAnalysisUsage(AnalysisUsage &AU) const {
|
||||
AU.addRequired<TargetTransformInfoWrapperPass>();
|
||||
LegacyInlinerBase::getAnalysisUsage(AU);
|
||||
}
|
||||
|
||||
unsigned AMDGPUInliner::getInlineThreshold(CallSite CS) const {
|
||||
int Thres = Params.DefaultThreshold;
|
||||
|
||||
Function *Caller = CS.getCaller();
|
||||
// Listen to the inlinehint attribute when it would increase the threshold
|
||||
// and the caller does not need to minimize its size.
|
||||
Function *Callee = CS.getCalledFunction();
|
||||
bool InlineHint = Callee && !Callee->isDeclaration() &&
|
||||
Callee->hasFnAttribute(Attribute::InlineHint);
|
||||
if (InlineHint && Params.HintThreshold && Params.HintThreshold > Thres
|
||||
&& !Caller->hasFnAttribute(Attribute::MinSize))
|
||||
Thres = Params.HintThreshold.getValue();
|
||||
|
||||
const DataLayout &DL = Caller->getParent()->getDataLayout();
|
||||
if (!Callee)
|
||||
return (unsigned)Thres;
|
||||
|
||||
const AMDGPUAS AS = AMDGPU::getAMDGPUAS(*Caller->getParent());
|
||||
|
||||
// If we have a pointer to private array passed into a function
|
||||
// it will not be optimized out, leaving scratch usage.
|
||||
// Increase the inline threshold to allow inliniting in this case.
|
||||
uint64_t AllocaSize = 0;
|
||||
SmallPtrSet<const AllocaInst *, 8> AIVisited;
|
||||
for (Value *PtrArg : CS.args()) {
|
||||
Type *Ty = PtrArg->getType();
|
||||
if (!Ty->isPointerTy() ||
|
||||
Ty->getPointerAddressSpace() != AS.PRIVATE_ADDRESS)
|
||||
continue;
|
||||
PtrArg = GetUnderlyingObject(PtrArg, DL);
|
||||
if (const AllocaInst *AI = dyn_cast<AllocaInst>(PtrArg)) {
|
||||
if (!AI->isStaticAlloca() || !AIVisited.insert(AI).second)
|
||||
continue;
|
||||
AllocaSize += DL.getTypeAllocSize(AI->getAllocatedType());
|
||||
// If the amount of stack memory is excessive we will not be able
|
||||
// to get rid of the scratch anyway, bail out.
|
||||
if (AllocaSize > ArgAllocaCutoff) {
|
||||
AllocaSize = 0;
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
if (AllocaSize)
|
||||
Thres += ArgAllocaCost;
|
||||
|
||||
return (unsigned)Thres;
|
||||
}
|
||||
|
||||
// Check if call is just a wrapper around another call.
|
||||
// In this case we only have call and ret instructions.
|
||||
static bool isWrapperOnlyCall(CallSite CS) {
|
||||
Function *Callee = CS.getCalledFunction();
|
||||
if (!Callee || Callee->size() != 1)
|
||||
return false;
|
||||
const BasicBlock &BB = Callee->getEntryBlock();
|
||||
if (const Instruction *I = BB.getFirstNonPHI()) {
|
||||
if (!isa<CallInst>(I)) {
|
||||
return false;
|
||||
}
|
||||
if (isa<ReturnInst>(*std::next(I->getIterator()))) {
|
||||
DEBUG(dbgs() << " Wrapper only call detected: "
|
||||
<< Callee->getName() << '\n');
|
||||
return true;
|
||||
}
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
InlineCost AMDGPUInliner::getInlineCost(CallSite CS) {
|
||||
Function *Callee = CS.getCalledFunction();
|
||||
Function *Caller = CS.getCaller();
|
||||
TargetTransformInfo &TTI = TTIWP->getTTI(*Callee);
|
||||
|
||||
if (!Callee || Callee->isDeclaration() || CS.isNoInline() ||
|
||||
!TTI.areInlineCompatible(Caller, Callee))
|
||||
return llvm::InlineCost::getNever();
|
||||
|
||||
if (CS.hasFnAttr(Attribute::AlwaysInline)) {
|
||||
if (isInlineViable(*Callee))
|
||||
return llvm::InlineCost::getAlways();
|
||||
return llvm::InlineCost::getNever();
|
||||
}
|
||||
|
||||
if (isWrapperOnlyCall(CS))
|
||||
return llvm::InlineCost::getAlways();
|
||||
|
||||
InlineParams LocalParams = Params;
|
||||
LocalParams.DefaultThreshold = (int)getInlineThreshold(CS);
|
||||
bool RemarksEnabled = false;
|
||||
const auto &BBs = Caller->getBasicBlockList();
|
||||
if (!BBs.empty()) {
|
||||
auto DI = OptimizationRemark(DEBUG_TYPE, "", DebugLoc(), &BBs.front());
|
||||
if (DI.isEnabled())
|
||||
RemarksEnabled = true;
|
||||
}
|
||||
|
||||
OptimizationRemarkEmitter ORE(Caller);
|
||||
std::function<AssumptionCache &(Function &)> GetAssumptionCache =
|
||||
[this](Function &F) -> AssumptionCache & {
|
||||
return ACT->getAssumptionCache(F);
|
||||
};
|
||||
|
||||
return llvm::getInlineCost(CS, Callee, LocalParams, TTI, GetAssumptionCache,
|
||||
None, PSI, RemarksEnabled ? &ORE : nullptr);
|
||||
}
|
Reference in New Issue
Block a user