You've already forked linux-packaging-mono
Imported Upstream version 5.18.0.207
Former-commit-id: 3b152f462918d427ce18620a2cbe4f8b79650449
This commit is contained in:
parent
8e12397d70
commit
eb85e2fc17
@ -1,206 +0,0 @@
|
||||
//===-- AMDGPUAsmBackend.cpp - AMDGPU Assembler Backend -------------------===//
|
||||
//
|
||||
// The LLVM Compiler Infrastructure
|
||||
//
|
||||
// This file is distributed under the University of Illinois Open Source
|
||||
// License. See LICENSE.TXT for details.
|
||||
//
|
||||
/// \file
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
#include "MCTargetDesc/AMDGPUFixupKinds.h"
|
||||
#include "MCTargetDesc/AMDGPUMCTargetDesc.h"
|
||||
#include "llvm/ADT/StringRef.h"
|
||||
#include "llvm/BinaryFormat/ELF.h"
|
||||
#include "llvm/MC/MCAsmBackend.h"
|
||||
#include "llvm/MC/MCAssembler.h"
|
||||
#include "llvm/MC/MCContext.h"
|
||||
#include "llvm/MC/MCFixupKindInfo.h"
|
||||
#include "llvm/MC/MCObjectWriter.h"
|
||||
#include "llvm/MC/MCValue.h"
|
||||
#include "llvm/Support/TargetRegistry.h"
|
||||
|
||||
using namespace llvm;
|
||||
|
||||
namespace {
|
||||
|
||||
class AMDGPUAsmBackend : public MCAsmBackend {
|
||||
public:
|
||||
AMDGPUAsmBackend(const Target &T)
|
||||
: MCAsmBackend() {}
|
||||
|
||||
unsigned getNumFixupKinds() const override { return AMDGPU::NumTargetFixupKinds; };
|
||||
|
||||
void applyFixup(const MCAssembler &Asm, const MCFixup &Fixup,
|
||||
const MCValue &Target, MutableArrayRef<char> Data,
|
||||
uint64_t Value, bool IsResolved) const override;
|
||||
bool fixupNeedsRelaxation(const MCFixup &Fixup, uint64_t Value,
|
||||
const MCRelaxableFragment *DF,
|
||||
const MCAsmLayout &Layout) const override {
|
||||
return false;
|
||||
}
|
||||
void relaxInstruction(const MCInst &Inst, const MCSubtargetInfo &STI,
|
||||
MCInst &Res) const override {
|
||||
llvm_unreachable("Not implemented");
|
||||
}
|
||||
bool mayNeedRelaxation(const MCInst &Inst) const override { return false; }
|
||||
|
||||
unsigned getMinimumNopSize() const override;
|
||||
bool writeNopData(uint64_t Count, MCObjectWriter *OW) const override;
|
||||
|
||||
const MCFixupKindInfo &getFixupKindInfo(MCFixupKind Kind) const override;
|
||||
};
|
||||
|
||||
} //End anonymous namespace
|
||||
|
||||
static unsigned getFixupKindNumBytes(unsigned Kind) {
|
||||
switch (Kind) {
|
||||
case AMDGPU::fixup_si_sopp_br:
|
||||
return 2;
|
||||
case FK_SecRel_1:
|
||||
case FK_Data_1:
|
||||
return 1;
|
||||
case FK_SecRel_2:
|
||||
case FK_Data_2:
|
||||
return 2;
|
||||
case FK_SecRel_4:
|
||||
case FK_Data_4:
|
||||
case FK_PCRel_4:
|
||||
return 4;
|
||||
case FK_SecRel_8:
|
||||
case FK_Data_8:
|
||||
return 8;
|
||||
default:
|
||||
llvm_unreachable("Unknown fixup kind!");
|
||||
}
|
||||
}
|
||||
|
||||
static uint64_t adjustFixupValue(const MCFixup &Fixup, uint64_t Value,
|
||||
MCContext *Ctx) {
|
||||
int64_t SignedValue = static_cast<int64_t>(Value);
|
||||
|
||||
switch (static_cast<unsigned>(Fixup.getKind())) {
|
||||
case AMDGPU::fixup_si_sopp_br: {
|
||||
int64_t BrImm = (SignedValue - 4) / 4;
|
||||
|
||||
if (Ctx && !isInt<16>(BrImm))
|
||||
Ctx->reportError(Fixup.getLoc(), "branch size exceeds simm16");
|
||||
|
||||
return BrImm;
|
||||
}
|
||||
case FK_Data_1:
|
||||
case FK_Data_2:
|
||||
case FK_Data_4:
|
||||
case FK_Data_8:
|
||||
case FK_PCRel_4:
|
||||
case FK_SecRel_4:
|
||||
return Value;
|
||||
default:
|
||||
llvm_unreachable("unhandled fixup kind");
|
||||
}
|
||||
}
|
||||
|
||||
void AMDGPUAsmBackend::applyFixup(const MCAssembler &Asm, const MCFixup &Fixup,
|
||||
const MCValue &Target,
|
||||
MutableArrayRef<char> Data, uint64_t Value,
|
||||
bool IsResolved) const {
|
||||
Value = adjustFixupValue(Fixup, Value, &Asm.getContext());
|
||||
if (!Value)
|
||||
return; // Doesn't change encoding.
|
||||
|
||||
MCFixupKindInfo Info = getFixupKindInfo(Fixup.getKind());
|
||||
|
||||
// Shift the value into position.
|
||||
Value <<= Info.TargetOffset;
|
||||
|
||||
unsigned NumBytes = getFixupKindNumBytes(Fixup.getKind());
|
||||
uint32_t Offset = Fixup.getOffset();
|
||||
assert(Offset + NumBytes <= Data.size() && "Invalid fixup offset!");
|
||||
|
||||
// For each byte of the fragment that the fixup touches, mask in the bits from
|
||||
// the fixup value.
|
||||
for (unsigned i = 0; i != NumBytes; ++i)
|
||||
Data[Offset + i] |= static_cast<uint8_t>((Value >> (i * 8)) & 0xff);
|
||||
}
|
||||
|
||||
const MCFixupKindInfo &AMDGPUAsmBackend::getFixupKindInfo(
|
||||
MCFixupKind Kind) const {
|
||||
const static MCFixupKindInfo Infos[AMDGPU::NumTargetFixupKinds] = {
|
||||
// name offset bits flags
|
||||
{ "fixup_si_sopp_br", 0, 16, MCFixupKindInfo::FKF_IsPCRel },
|
||||
};
|
||||
|
||||
if (Kind < FirstTargetFixupKind)
|
||||
return MCAsmBackend::getFixupKindInfo(Kind);
|
||||
|
||||
return Infos[Kind - FirstTargetFixupKind];
|
||||
}
|
||||
|
||||
unsigned AMDGPUAsmBackend::getMinimumNopSize() const {
|
||||
return 4;
|
||||
}
|
||||
|
||||
bool AMDGPUAsmBackend::writeNopData(uint64_t Count, MCObjectWriter *OW) const {
|
||||
// If the count is not 4-byte aligned, we must be writing data into the text
|
||||
// section (otherwise we have unaligned instructions, and thus have far
|
||||
// bigger problems), so just write zeros instead.
|
||||
OW->WriteZeros(Count % 4);
|
||||
|
||||
// We are properly aligned, so write NOPs as requested.
|
||||
Count /= 4;
|
||||
|
||||
// FIXME: R600 support.
|
||||
// s_nop 0
|
||||
const uint32_t Encoded_S_NOP_0 = 0xbf800000;
|
||||
|
||||
for (uint64_t I = 0; I != Count; ++I)
|
||||
OW->write32(Encoded_S_NOP_0);
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// ELFAMDGPUAsmBackend class
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
namespace {
|
||||
|
||||
class ELFAMDGPUAsmBackend : public AMDGPUAsmBackend {
|
||||
bool Is64Bit;
|
||||
bool HasRelocationAddend;
|
||||
uint8_t OSABI = ELF::ELFOSABI_NONE;
|
||||
|
||||
public:
|
||||
ELFAMDGPUAsmBackend(const Target &T, const Triple &TT) :
|
||||
AMDGPUAsmBackend(T), Is64Bit(TT.getArch() == Triple::amdgcn),
|
||||
HasRelocationAddend(TT.getOS() == Triple::AMDHSA) {
|
||||
switch (TT.getOS()) {
|
||||
case Triple::AMDHSA:
|
||||
OSABI = ELF::ELFOSABI_AMDGPU_HSA;
|
||||
break;
|
||||
case Triple::AMDPAL:
|
||||
OSABI = ELF::ELFOSABI_AMDGPU_PAL;
|
||||
break;
|
||||
case Triple::Mesa3D:
|
||||
OSABI = ELF::ELFOSABI_AMDGPU_MESA3D;
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
std::unique_ptr<MCObjectWriter>
|
||||
createObjectWriter(raw_pwrite_stream &OS) const override {
|
||||
return createAMDGPUELFObjectWriter(Is64Bit, OSABI, HasRelocationAddend, OS);
|
||||
}
|
||||
};
|
||||
|
||||
} // end anonymous namespace
|
||||
|
||||
MCAsmBackend *llvm::createAMDGPUAsmBackend(const Target &T,
|
||||
const MCSubtargetInfo &STI,
|
||||
const MCRegisterInfo &MRI,
|
||||
const MCTargetOptions &Options) {
|
||||
// Use 64-bit ELF for amdgcn
|
||||
return new ELFAMDGPUAsmBackend(T, STI.getTargetTriple());
|
||||
}
|
Reference in New Issue
Block a user