radeon/llvm: Remove AMDILUtilityFunctions.cpp

This commit is contained in:
Tom Stellard 2012-05-08 15:37:59 -04:00
parent 21ab46eae8
commit 5aaaa6a426
13 changed files with 399 additions and 1041 deletions

View File

@ -176,26 +176,6 @@ typedef union ResourceRec {
} // namespace AMDILAS
// The OpSwizzle encodes a subset of all possible
// swizzle combinations into a number of bits using
// only the combinations utilized by the backend.
// The lower 128 are for source swizzles and the
// upper 128 or for destination swizzles.
// The valid mappings can be found in the
// getSrcSwizzle and getDstSwizzle functions of
// AMDILUtilityFunctions.cpp.
typedef union SwizzleRec {
struct {
#ifdef __BIG_ENDIAN__
unsigned char dst : 1;
unsigned char swizzle : 7;
#else
unsigned char swizzle : 7;
unsigned char dst : 1;
#endif
} bits;
unsigned char u8all;
} OpSwizzle;
// Enums corresponding to AMDIL condition codes for IL. These
// values must be kept in sync with the ones in the .td file.
namespace AMDILCC {

View File

@ -3152,10 +3152,6 @@ struct CFGStructTraits<AMDILCFGStructurizer>
iterEnd = srcBlk->end();
iter != iterEnd; ++iter) {
MachineInstr *instr = func->CloneMachineInstr(iter);
// This is a workaround for LLVM bugzilla 8420 because CloneMachineInstr
// does not clone the AsmPrinterFlags.
instr->setAsmPrinterFlag(
(llvm::MachineInstr::CommentFlag)iter->getAsmPrinterFlags());
newBlk->push_back(instr);
}
return newBlk;

View File

@ -13,9 +13,12 @@
#include "AMDILDevices.h"
#include "AMDILTargetMachine.h"
#include "AMDILUtilityFunctions.h"
#include "llvm/ADT/ValueMap.h"
#include "llvm/CodeGen/PseudoSourceValue.h"
#include "llvm/CodeGen/SelectionDAGISel.h"
#include "llvm/Support/Compiler.h"
#include <list>
#include <queue>
using namespace llvm;
@ -35,13 +38,21 @@ class AMDILDAGToDAGISel : public SelectionDAGISel {
public:
AMDILDAGToDAGISel(AMDILTargetMachine &TM AMDIL_OPT_LEVEL_DECL);
virtual ~AMDILDAGToDAGISel();
inline SDValue getSmallIPtrImm(unsigned Imm);
SDNode *Select(SDNode *N);
virtual const char *getPassName() const;
private:
inline SDValue getSmallIPtrImm(unsigned Imm);
// Complex pattern selectors
bool SelectADDRParam(SDValue Addr, SDValue& R1, SDValue& R2);
bool SelectADDR(SDValue N, SDValue &R1, SDValue &R2);
bool SelectADDR64(SDValue N, SDValue &R1, SDValue &R2);
static bool checkType(const Value *ptr, unsigned int addrspace);
static const Value *getBasePointerValue(const Value *V);
static bool isGlobalStore(const StoreSDNode *N);
static bool isPrivateStore(const StoreSDNode *N);
static bool isLocalStore(const StoreSDNode *N);
@ -54,8 +65,6 @@ public:
static bool isLocalLoad(const LoadSDNode *N);
static bool isRegionLoad(const LoadSDNode *N);
virtual const char *getPassName() const;
private:
SDNode *xformAtomicInst(SDNode *N);
// Include the pieces autogenerated from the target description.
@ -165,26 +174,75 @@ SDNode *AMDILDAGToDAGISel::Select(SDNode *N) {
return SelectCode(N);
}
bool AMDILDAGToDAGISel::checkType(const Value *ptr, unsigned int addrspace) {
if (!ptr) {
return false;
}
Type *ptrType = ptr->getType();
return dyn_cast<PointerType>(ptrType)->getAddressSpace() == addrspace;
}
const Value * AMDILDAGToDAGISel::getBasePointerValue(const Value *V)
{
if (!V) {
return NULL;
}
const Value *ret = NULL;
ValueMap<const Value *, bool> ValueBitMap;
std::queue<const Value *, std::list<const Value *> > ValueQueue;
ValueQueue.push(V);
while (!ValueQueue.empty()) {
V = ValueQueue.front();
if (ValueBitMap.find(V) == ValueBitMap.end()) {
ValueBitMap[V] = true;
if (dyn_cast<Argument>(V) && dyn_cast<PointerType>(V->getType())) {
ret = V;
break;
} else if (dyn_cast<GlobalVariable>(V)) {
ret = V;
break;
} else if (dyn_cast<Constant>(V)) {
const ConstantExpr *CE = dyn_cast<ConstantExpr>(V);
if (CE) {
ValueQueue.push(CE->getOperand(0));
}
} else if (const AllocaInst *AI = dyn_cast<AllocaInst>(V)) {
ret = AI;
break;
} else if (const Instruction *I = dyn_cast<Instruction>(V)) {
uint32_t numOps = I->getNumOperands();
for (uint32_t x = 0; x < numOps; ++x) {
ValueQueue.push(I->getOperand(x));
}
} else {
// assert(0 && "Found a Value that we didn't know how to handle!");
}
}
ValueQueue.pop();
}
return ret;
}
bool AMDILDAGToDAGISel::isGlobalStore(const StoreSDNode *N) {
return check_type(N->getSrcValue(), AMDILAS::GLOBAL_ADDRESS);
return checkType(N->getSrcValue(), AMDILAS::GLOBAL_ADDRESS);
}
bool AMDILDAGToDAGISel::isPrivateStore(const StoreSDNode *N) {
return (!check_type(N->getSrcValue(), AMDILAS::LOCAL_ADDRESS)
&& !check_type(N->getSrcValue(), AMDILAS::GLOBAL_ADDRESS)
&& !check_type(N->getSrcValue(), AMDILAS::REGION_ADDRESS));
return (!checkType(N->getSrcValue(), AMDILAS::LOCAL_ADDRESS)
&& !checkType(N->getSrcValue(), AMDILAS::GLOBAL_ADDRESS)
&& !checkType(N->getSrcValue(), AMDILAS::REGION_ADDRESS));
}
bool AMDILDAGToDAGISel::isLocalStore(const StoreSDNode *N) {
return check_type(N->getSrcValue(), AMDILAS::LOCAL_ADDRESS);
return checkType(N->getSrcValue(), AMDILAS::LOCAL_ADDRESS);
}
bool AMDILDAGToDAGISel::isRegionStore(const StoreSDNode *N) {
return check_type(N->getSrcValue(), AMDILAS::REGION_ADDRESS);
return checkType(N->getSrcValue(), AMDILAS::REGION_ADDRESS);
}
bool AMDILDAGToDAGISel::isConstantLoad(const LoadSDNode *N, int cbID) {
if (check_type(N->getSrcValue(), AMDILAS::CONSTANT_ADDRESS)) {
if (checkType(N->getSrcValue(), AMDILAS::CONSTANT_ADDRESS)) {
return true;
}
MachineMemOperand *MMO = N->getMemOperand();
@ -195,27 +253,27 @@ bool AMDILDAGToDAGISel::isConstantLoad(const LoadSDNode *N, int cbID) {
&& ((V && dyn_cast<GlobalValue>(V))
|| (BV && dyn_cast<GlobalValue>(
getBasePointerValue(MMO->getValue()))))) {
return check_type(N->getSrcValue(), AMDILAS::PRIVATE_ADDRESS);
return checkType(N->getSrcValue(), AMDILAS::PRIVATE_ADDRESS);
} else {
return false;
}
}
bool AMDILDAGToDAGISel::isGlobalLoad(const LoadSDNode *N) {
return check_type(N->getSrcValue(), AMDILAS::GLOBAL_ADDRESS);
return checkType(N->getSrcValue(), AMDILAS::GLOBAL_ADDRESS);
}
bool AMDILDAGToDAGISel::isLocalLoad(const LoadSDNode *N) {
return check_type(N->getSrcValue(), AMDILAS::LOCAL_ADDRESS);
return checkType(N->getSrcValue(), AMDILAS::LOCAL_ADDRESS);
}
bool AMDILDAGToDAGISel::isRegionLoad(const LoadSDNode *N) {
return check_type(N->getSrcValue(), AMDILAS::REGION_ADDRESS);
return checkType(N->getSrcValue(), AMDILAS::REGION_ADDRESS);
}
bool AMDILDAGToDAGISel::isCPLoad(const LoadSDNode *N) {
MachineMemOperand *MMO = N->getMemOperand();
if (check_type(N->getSrcValue(), AMDILAS::PRIVATE_ADDRESS)) {
if (checkType(N->getSrcValue(), AMDILAS::PRIVATE_ADDRESS)) {
if (MMO) {
const Value *V = MMO->getValue();
const PseudoSourceValue *PSV = dyn_cast<PseudoSourceValue>(V);
@ -228,19 +286,19 @@ bool AMDILDAGToDAGISel::isCPLoad(const LoadSDNode *N) {
}
bool AMDILDAGToDAGISel::isPrivateLoad(const LoadSDNode *N) {
if (check_type(N->getSrcValue(), AMDILAS::PRIVATE_ADDRESS)) {
if (checkType(N->getSrcValue(), AMDILAS::PRIVATE_ADDRESS)) {
// Check to make sure we are not a constant pool load or a constant load
// that is marked as a private load
if (isCPLoad(N) || isConstantLoad(N, -1)) {
return false;
}
}
if (!check_type(N->getSrcValue(), AMDILAS::LOCAL_ADDRESS)
&& !check_type(N->getSrcValue(), AMDILAS::GLOBAL_ADDRESS)
&& !check_type(N->getSrcValue(), AMDILAS::REGION_ADDRESS)
&& !check_type(N->getSrcValue(), AMDILAS::CONSTANT_ADDRESS)
&& !check_type(N->getSrcValue(), AMDILAS::PARAM_D_ADDRESS)
&& !check_type(N->getSrcValue(), AMDILAS::PARAM_I_ADDRESS))
if (!checkType(N->getSrcValue(), AMDILAS::LOCAL_ADDRESS)
&& !checkType(N->getSrcValue(), AMDILAS::GLOBAL_ADDRESS)
&& !checkType(N->getSrcValue(), AMDILAS::REGION_ADDRESS)
&& !checkType(N->getSrcValue(), AMDILAS::CONSTANT_ADDRESS)
&& !checkType(N->getSrcValue(), AMDILAS::PARAM_D_ADDRESS)
&& !checkType(N->getSrcValue(), AMDILAS::PARAM_I_ADDRESS))
{
return true;
}

View File

@ -623,6 +623,48 @@ translateToOpcode(uint64_t CCCode, unsigned int regClass)
assert(0 && "Unknown opcode retrieved");
return 0;
}
/// Helper function used by LowerFormalArguments
static const TargetRegisterClass*
getRegClassFromType(unsigned int type) {
switch (type) {
default:
assert(0 && "Passed in type does not match any register classes.");
case MVT::i8:
return &AMDIL::GPRI8RegClass;
case MVT::i16:
return &AMDIL::GPRI16RegClass;
case MVT::i32:
return &AMDIL::GPRI32RegClass;
case MVT::f32:
return &AMDIL::GPRF32RegClass;
case MVT::i64:
return &AMDIL::GPRI64RegClass;
case MVT::f64:
return &AMDIL::GPRF64RegClass;
case MVT::v4f32:
return &AMDIL::GPRV4F32RegClass;
case MVT::v4i8:
return &AMDIL::GPRV4I8RegClass;
case MVT::v4i16:
return &AMDIL::GPRV4I16RegClass;
case MVT::v4i32:
return &AMDIL::GPRV4I32RegClass;
case MVT::v2f32:
return &AMDIL::GPRV2F32RegClass;
case MVT::v2i8:
return &AMDIL::GPRV2I8RegClass;
case MVT::v2i16:
return &AMDIL::GPRV2I16RegClass;
case MVT::v2i32:
return &AMDIL::GPRV2I32RegClass;
case MVT::v2f64:
return &AMDIL::GPRV2F64RegClass;
case MVT::v2i64:
return &AMDIL::GPRV2I64RegClass;
}
}
SDValue
AMDILTargetLowering::LowerMemArgument(
SDValue Chain,
@ -2189,6 +2231,7 @@ AMDILTargetLowering::LowerExternalSymbol(SDValue Op, SelectionDAG &DAG) const
SDValue Result = DAG.getTargetExternalSymbol(Sym, MVT::i32);
return Result;
}
/// LowerFORMAL_ARGUMENTS - transform physical registers into
/// virtual registers and generate load operations for
/// arguments places on the stack.
@ -3843,7 +3886,6 @@ SDValue
AMDILTargetLowering::LowerBUILD_VECTOR( SDValue Op, SelectionDAG &DAG ) const
{
EVT VT = Op.getValueType();
//printSDValue(Op, 1);
SDValue Nodes1;
SDValue second;
SDValue third;
@ -3965,7 +4007,6 @@ AMDILTargetLowering::LowerEXTRACT_VECTOR_ELT(SDValue Op,
SelectionDAG &DAG) const
{
EVT VT = Op.getValueType();
//printSDValue(Op, 1);
const ConstantSDNode *CSDN = dyn_cast<ConstantSDNode>(Op.getOperand(1));
uint64_t swizzleNum = 0;
DebugLoc DL = Op.getDebugLoc();
@ -4782,7 +4823,7 @@ uint32_t
AMDILTargetLowering::genVReg(uint32_t regType) const
{
return mBB->getParent()->getRegInfo().createVirtualRegister(
getRegClassFromID(regType));
getTargetMachine().getRegisterInfo()->getRegClass(regType));
}
MachineInstrBuilder

View File

@ -10,13 +10,10 @@
// This file contains the AMDIL implementation of the TargetInstrInfo class.
//
//===----------------------------------------------------------------------===//
#include "AMDILInstrInfo.h"
#include "AMDILUtilityFunctions.h"
#define GET_INSTRINFO_CTOR
#include "AMDILGenInstrInfo.inc"
#include "AMDILInstrInfo.h"
#include "AMDIL.h"
#include "AMDILISelLowering.h"
#include "AMDILUtilityFunctions.h"
#include "llvm/CodeGen/MachineFrameInfo.h"
#include "llvm/CodeGen/MachineInstrBuilder.h"
@ -24,6 +21,9 @@
#include "llvm/CodeGen/PseudoSourceValue.h"
#include "llvm/Instructions.h"
#define GET_INSTRINFO_CTOR
#include "AMDILGenInstrInfo.inc"
using namespace llvm;
AMDILInstrInfo::AMDILInstrInfo(AMDILTargetMachine &tm)
@ -334,15 +334,11 @@ AMDILInstrInfo::storeRegToStackSlot(MachineBasicBlock &MBB,
if (MI != MBB.end()) {
DL = MI->getDebugLoc();
}
MachineInstr *nMI = BuildMI(MBB, MI, DL, get(Opc))
BuildMI(MBB, MI, DL, get(Opc))
.addReg(SrcReg, getKillRegState(isKill))
.addFrameIndex(FrameIndex)
.addMemOperand(MMO)
.addImm(0);
AMDILAS::InstrResEnc curRes;
curRes.bits.ResourceID
= TM.getSubtargetImpl()->device()->getResourceID(AMDILDevice::SCRATCH_ID);
setAsmPrinterFlags(nMI, curRes);
}
void
@ -418,16 +414,11 @@ AMDILInstrInfo::loadRegFromStackSlot(MachineBasicBlock &MBB,
if (MI != MBB.end()) {
DL = MI->getDebugLoc();
}
MachineInstr* nMI = BuildMI(MBB, MI, DL, get(Opc))
BuildMI(MBB, MI, DL, get(Opc))
.addReg(DestReg, RegState::Define)
.addFrameIndex(FrameIndex)
.addMemOperand(MMO)
.addImm(0);
AMDILAS::InstrResEnc curRes;
curRes.bits.ResourceID
= TM.getSubtargetImpl()->device()->getResourceID(AMDILDevice::SCRATCH_ID);
setAsmPrinterFlags(nMI, curRes);
}
MachineInstr *
AMDILInstrInfo::foldMemoryOperandImpl(MachineFunction &MF,
@ -476,65 +467,6 @@ AMDILInstrInfo::getOpcodeAfterMemoryUnfold(unsigned Opc,
return 0;
}
bool
AMDILInstrInfo::areLoadsFromSameBasePtr(SDNode *Load1, SDNode *Load2,
int64_t &Offset1,
int64_t &Offset2) const {
return false;
if (!Load1->isMachineOpcode() || !Load2->isMachineOpcode()) {
return false;
}
const MachineSDNode *mload1 = dyn_cast<MachineSDNode>(Load1);
const MachineSDNode *mload2 = dyn_cast<MachineSDNode>(Load2);
if (!mload1 || !mload2) {
return false;
}
if (mload1->memoperands_empty() ||
mload2->memoperands_empty()) {
return false;
}
MachineMemOperand *memOp1 = (*mload1->memoperands_begin());
MachineMemOperand *memOp2 = (*mload2->memoperands_begin());
const Value *mv1 = memOp1->getValue();
const Value *mv2 = memOp2->getValue();
if (!memOp1->isLoad() || !memOp2->isLoad()) {
return false;
}
if (getBasePointerValue(mv1) == getBasePointerValue(mv2)) {
if (isa<GetElementPtrInst>(mv1) && isa<GetElementPtrInst>(mv2)) {
const GetElementPtrInst *gep1 = dyn_cast<GetElementPtrInst>(mv1);
const GetElementPtrInst *gep2 = dyn_cast<GetElementPtrInst>(mv2);
if (!gep1 || !gep2) {
return false;
}
if (gep1->getNumOperands() != gep2->getNumOperands()) {
return false;
}
for (unsigned i = 0, e = gep1->getNumOperands() - 1; i < e; ++i) {
const Value *op1 = gep1->getOperand(i);
const Value *op2 = gep2->getOperand(i);
if (op1 != op2) {
// If any value except the last one is different, return false.
return false;
}
}
unsigned size = gep1->getNumOperands()-1;
if (!isa<ConstantInt>(gep1->getOperand(size))
|| !isa<ConstantInt>(gep2->getOperand(size))) {
return false;
}
Offset1 = dyn_cast<ConstantInt>(gep1->getOperand(size))->getSExtValue();
Offset2 = dyn_cast<ConstantInt>(gep2->getOperand(size))->getSExtValue();
return true;
} else if (isa<Argument>(mv1) && isa<Argument>(mv2)) {
return false;
} else if (isa<GlobalValue>(mv1) && isa<GlobalValue>(mv2)) {
return false;
}
}
return false;
}
bool AMDILInstrInfo::shouldScheduleLoadsNear(SDNode *Load1, SDNode *Load2,
int64_t Offset1, int64_t Offset2,
unsigned NumLoads) const {
@ -585,3 +517,113 @@ AMDILInstrInfo::isSafeToMoveRegClassDefs(const TargetRegisterClass *RC) const {
// TODO: Implement this function
return true;
}
bool AMDILInstrInfo::isLoadInst(MachineInstr *MI) const {
if (strstr(getName(MI->getOpcode()), "LOADCONST")) {
return false;
}
return strstr(getName(MI->getOpcode()), "LOAD");
}
bool AMDILInstrInfo::isSWSExtLoadInst(MachineInstr *MI) const
{
switch (MI->getOpcode()) {
default:
break;
ExpandCaseToByteShortTypes(AMDIL::LOCALLOAD);
ExpandCaseToByteShortTypes(AMDIL::GLOBALLOAD);
ExpandCaseToByteShortTypes(AMDIL::REGIONLOAD);
ExpandCaseToByteShortTypes(AMDIL::PRIVATELOAD);
ExpandCaseToByteShortTypes(AMDIL::CPOOLLOAD);
ExpandCaseToByteShortTypes(AMDIL::CONSTANTLOAD);
return true;
};
return false;
}
bool AMDILInstrInfo::isExtLoadInst(MachineInstr *MI) const {
return strstr(getName(MI->getOpcode()), "EXTLOAD");
}
bool AMDILInstrInfo::isSExtLoadInst(MachineInstr *MI) const {
return strstr(getName(MI->getOpcode()), "SEXTLOAD");
}
bool AMDILInstrInfo::isAExtLoadInst(MachineInstr *MI) const {
return strstr(getName(MI->getOpcode()), "AEXTLOAD");
}
bool AMDILInstrInfo::isZExtLoadInst(MachineInstr *MI) const {
return strstr(getName(MI->getOpcode()), "ZEXTLOAD");
}
bool AMDILInstrInfo::isStoreInst(MachineInstr *MI) const {
return strstr(getName(MI->getOpcode()), "STORE");
}
bool AMDILInstrInfo::isTruncStoreInst(MachineInstr *MI) const {
return strstr(getName(MI->getOpcode()), "TRUNCSTORE");
}
bool AMDILInstrInfo::isAtomicInst(MachineInstr *MI) const {
return strstr(getName(MI->getOpcode()), "ATOM");
}
bool AMDILInstrInfo::isVolatileInst(MachineInstr *MI) const {
if (!MI->memoperands_empty()) {
for (MachineInstr::mmo_iterator mob = MI->memoperands_begin(),
moe = MI->memoperands_end(); mob != moe; ++mob) {
// If there is a volatile mem operand, this is a volatile instruction.
if ((*mob)->isVolatile()) {
return true;
}
}
}
return false;
}
bool AMDILInstrInfo::isGlobalInst(llvm::MachineInstr *MI) const
{
return strstr(getName(MI->getOpcode()), "GLOBAL");
}
bool AMDILInstrInfo::isPrivateInst(llvm::MachineInstr *MI) const
{
return strstr(getName(MI->getOpcode()), "PRIVATE");
}
bool AMDILInstrInfo::isConstantInst(llvm::MachineInstr *MI) const
{
return strstr(getName(MI->getOpcode()), "CONSTANT")
|| strstr(getName(MI->getOpcode()), "CPOOL");
}
bool AMDILInstrInfo::isRegionInst(llvm::MachineInstr *MI) const
{
return strstr(getName(MI->getOpcode()), "REGION");
}
bool AMDILInstrInfo::isLocalInst(llvm::MachineInstr *MI) const
{
return strstr(getName(MI->getOpcode()), "LOCAL");
}
bool AMDILInstrInfo::isImageInst(llvm::MachineInstr *MI) const
{
return strstr(getName(MI->getOpcode()), "IMAGE");
}
bool AMDILInstrInfo::isAppendInst(llvm::MachineInstr *MI) const
{
return strstr(getName(MI->getOpcode()), "APPEND");
}
bool AMDILInstrInfo::isRegionAtomic(llvm::MachineInstr *MI) const
{
return strstr(getName(MI->getOpcode()), "ATOM_R");
}
bool AMDILInstrInfo::isLocalAtomic(llvm::MachineInstr *MI) const
{
return strstr(getName(MI->getOpcode()), "ATOM_L");
}
bool AMDILInstrInfo::isGlobalAtomic(llvm::MachineInstr *MI) const
{
return strstr(getName(MI->getOpcode()), "ATOM_G")
|| isArenaAtomic(MI);
}
bool AMDILInstrInfo::isArenaAtomic(llvm::MachineInstr *MI) const
{
return strstr(getName(MI->getOpcode()), "ATOM_A");
}

View File

@ -110,8 +110,6 @@ public:
unsigned getOpcodeAfterMemoryUnfold(unsigned Opc,
bool UnfoldLoad, bool UnfoldStore,
unsigned *LoadRegIndex = 0) const;
bool areLoadsFromSameBasePtr(SDNode *Load1, SDNode *Load2,
int64_t &Offset1, int64_t &Offset2) const;
bool shouldScheduleLoadsNear(SDNode *Load1, SDNode *Load2,
int64_t Offset1, int64_t Offset2,
unsigned NumLoads) const;
@ -126,7 +124,30 @@ public:
std::vector<MachineOperand> &Pred) const;
bool isPredicable(MachineInstr *MI) const;
bool isSafeToMoveRegClassDefs(const TargetRegisterClass *RC) const;
};
// Helper functions that check the opcode for status information
bool isLoadInst(llvm::MachineInstr *MI) const;
bool isExtLoadInst(llvm::MachineInstr *MI) const;
bool isSWSExtLoadInst(llvm::MachineInstr *MI) const;
bool isSExtLoadInst(llvm::MachineInstr *MI) const;
bool isZExtLoadInst(llvm::MachineInstr *MI) const;
bool isAExtLoadInst(llvm::MachineInstr *MI) const;
bool isStoreInst(llvm::MachineInstr *MI) const;
bool isTruncStoreInst(llvm::MachineInstr *MI) const;
bool isAtomicInst(llvm::MachineInstr *MI) const;
bool isVolatileInst(llvm::MachineInstr *MI) const;
bool isGlobalInst(llvm::MachineInstr *MI) const;
bool isPrivateInst(llvm::MachineInstr *MI) const;
bool isConstantInst(llvm::MachineInstr *MI) const;
bool isRegionInst(llvm::MachineInstr *MI) const;
bool isLocalInst(llvm::MachineInstr *MI) const;
bool isImageInst(llvm::MachineInstr *MI) const;
bool isAppendInst(llvm::MachineInstr *MI) const;
bool isRegionAtomic(llvm::MachineInstr *MI) const;
bool isLocalAtomic(llvm::MachineInstr *MI) const;
bool isGlobalAtomic(llvm::MachineInstr *MI) const;
bool isArenaAtomic(llvm::MachineInstr *MI) const;
};
}

View File

@ -9,9 +9,10 @@
#include "AMDIL.h"
#include "AMDILInstrInfo.h"
#include "AMDILSubtarget.h"
#include "AMDILUtilityFunctions.h"
#include "llvm/CodeGen/MachineFunctionPass.h"
#include "llvm/CodeGen/MachineInstrBuilder.h"
#include "llvm/CodeGen/MachineRegisterInfo.h"
#include "llvm/Support/Debug.h"
#include "llvm/Target/TargetMachine.h"
@ -57,6 +58,8 @@ AMDILMachinePeephole::runOnMachineFunction(MachineFunction &MF)
{
bool Changed = false;
const AMDILSubtarget *STM = &TM.getSubtarget<AMDILSubtarget>();
const AMDILInstrInfo * AMDILII =
static_cast<const AMDILInstrInfo *>(TM.getInstrInfo());
for (MachineFunction::iterator MBB = MF.begin(), MBE = MF.end();
MBB != MBE; ++MBB) {
MachineBasicBlock *mb = MBB;
@ -67,7 +70,7 @@ AMDILMachinePeephole::runOnMachineFunction(MachineFunction &MF)
name = TM.getInstrInfo()->getName(mi->getOpcode());
switch (mi->getOpcode()) {
default:
if (isAtomicInst(TM.getInstrInfo(), mi)) {
if (AMDILII->isAtomicInst(mi)) {
// If we don't support the hardware accellerated address spaces,
// then the atomic needs to be transformed to the global atomic.
if (strstr(name, "_L_")
@ -87,7 +90,8 @@ AMDILMachinePeephole::runOnMachineFunction(MachineFunction &MF)
TM.getInstrInfo()->get(
(mi->getOpcode() - AMDIL::ATOM_R_ADD) + AMDIL::ATOM_G_ADD));
}
} else if ((isLoadInst(TM.getInstrInfo(), mi) || isStoreInst(TM.getInstrInfo(), mi)) && isVolatileInst(TM.getInstrInfo(), mi)) {
} else if ((AMDILII->isLoadInst(mi) || AMDILII->isStoreInst(mi))
&& AMDILII->isVolatileInst(mi)) {
insertFence(MIB);
}
continue;

View File

@ -9,11 +9,12 @@
#include "AMDILAlgorithms.tpp"
#include "AMDILDevices.h"
#include "AMDILUtilityFunctions.h"
#include "AMDILInstrInfo.h"
#include "llvm/ADT/Statistic.h"
#include "llvm/ADT/StringExtras.h"
#include "llvm/ADT/StringRef.h"
#include "llvm/ADT/Twine.h"
#include "llvm/Constants.h"
#include "llvm/CodeGen/MachineFunction.h"
#include "llvm/CodeGen/MachineFunctionAnalysis.h"
#include "llvm/Function.h"
@ -34,6 +35,9 @@ using namespace llvm;
// The Peephole optimization pass is used to do simple last minute optimizations
// that are required for correct code or to remove redundant functions
namespace {
class OpaqueType;
class LLVM_LIBRARY_VISIBILITY AMDILPeepholeOpt : public FunctionPass {
public:
TargetMachine &TM;
@ -107,6 +111,19 @@ private:
// samplers at compile time.
bool propagateSamplerInst(CallInst *CI);
// Helper functions
// Group of functions that recursively calculate the size of a structure based
// on it's sub-types.
size_t getTypeSize(Type * const T, bool dereferencePtr = false);
size_t getTypeSize(StructType * const ST, bool dereferencePtr = false);
size_t getTypeSize(IntegerType * const IT, bool dereferencePtr = false);
size_t getTypeSize(FunctionType * const FT,bool dereferencePtr = false);
size_t getTypeSize(ArrayType * const AT, bool dereferencePtr = false);
size_t getTypeSize(VectorType * const VT, bool dereferencePtr = false);
size_t getTypeSize(PointerType * const PT, bool dereferencePtr = false);
size_t getTypeSize(OpaqueType * const OT, bool dereferencePtr = false);
LLVMContext *mCTX;
Function *mF;
const AMDILSubtarget *mSTM;
@ -1129,3 +1146,106 @@ AMDILPeepholeOpt::getAnalysisUsage(AnalysisUsage &AU) const
FunctionPass::getAnalysisUsage(AU);
AU.setPreservesAll();
}
size_t AMDILPeepholeOpt::getTypeSize(Type * const T, bool dereferencePtr) {
size_t size = 0;
if (!T) {
return size;
}
switch (T->getTypeID()) {
case Type::X86_FP80TyID:
case Type::FP128TyID:
case Type::PPC_FP128TyID:
case Type::LabelTyID:
assert(0 && "These types are not supported by this backend");
default:
case Type::FloatTyID:
case Type::DoubleTyID:
size = T->getPrimitiveSizeInBits() >> 3;
break;
case Type::PointerTyID:
size = getTypeSize(dyn_cast<PointerType>(T), dereferencePtr);
break;
case Type::IntegerTyID:
size = getTypeSize(dyn_cast<IntegerType>(T), dereferencePtr);
break;
case Type::StructTyID:
size = getTypeSize(dyn_cast<StructType>(T), dereferencePtr);
break;
case Type::ArrayTyID:
size = getTypeSize(dyn_cast<ArrayType>(T), dereferencePtr);
break;
case Type::FunctionTyID:
size = getTypeSize(dyn_cast<FunctionType>(T), dereferencePtr);
break;
case Type::VectorTyID:
size = getTypeSize(dyn_cast<VectorType>(T), dereferencePtr);
break;
};
return size;
}
size_t AMDILPeepholeOpt::getTypeSize(StructType * const ST,
bool dereferencePtr) {
size_t size = 0;
if (!ST) {
return size;
}
Type *curType;
StructType::element_iterator eib;
StructType::element_iterator eie;
for (eib = ST->element_begin(), eie = ST->element_end(); eib != eie; ++eib) {
curType = *eib;
size += getTypeSize(curType, dereferencePtr);
}
return size;
}
size_t AMDILPeepholeOpt::getTypeSize(IntegerType * const IT,
bool dereferencePtr) {
return IT ? (IT->getBitWidth() >> 3) : 0;
}
size_t AMDILPeepholeOpt::getTypeSize(FunctionType * const FT,
bool dereferencePtr) {
assert(0 && "Should not be able to calculate the size of an function type");
return 0;
}
size_t AMDILPeepholeOpt::getTypeSize(ArrayType * const AT,
bool dereferencePtr) {
return (size_t)(AT ? (getTypeSize(AT->getElementType(),
dereferencePtr) * AT->getNumElements())
: 0);
}
size_t AMDILPeepholeOpt::getTypeSize(VectorType * const VT,
bool dereferencePtr) {
return VT ? (VT->getBitWidth() >> 3) : 0;
}
size_t AMDILPeepholeOpt::getTypeSize(PointerType * const PT,
bool dereferencePtr) {
if (!PT) {
return 0;
}
Type *CT = PT->getElementType();
if (CT->getTypeID() == Type::StructTyID &&
PT->getAddressSpace() == AMDILAS::PRIVATE_ADDRESS) {
return getTypeSize(dyn_cast<StructType>(CT));
} else if (dereferencePtr) {
size_t size = 0;
for (size_t x = 0, y = PT->getNumContainedTypes(); x < y; ++x) {
size += getTypeSize(PT->getContainedType(x), dereferencePtr);
}
return size;
} else {
return 4;
}
}
size_t AMDILPeepholeOpt::getTypeSize(OpaqueType * const OT,
bool dereferencePtr) {
//assert(0 && "Should not be able to calculate the size of an opaque type");
return 4;
}

View File

@ -20,7 +20,8 @@
#include "AMDILRegisterInfo.h"
#include "AMDIL.h"
#include "AMDILUtilityFunctions.h"
#include "AMDILInstrInfo.h"
#include "AMDILTargetMachine.h"
#include "llvm/ADT/BitVector.h"
#include "llvm/CodeGen/MachineFrameInfo.h"
#include "llvm/CodeGen/MachineRegisterInfo.h"
@ -109,7 +110,9 @@ AMDILRegisterInfo::eliminateFrameIndex(MachineBasicBlock::iterator II,
if (!MI.getOperand(x).isFI()) {
continue;
}
bool def = isStoreInst(TM.getInstrInfo(), &MI);
const AMDILInstrInfo * AMDILII =
static_cast<const AMDILInstrInfo *>(TM.getInstrInfo());
bool def = AMDILII->isStoreInst(&MI);
int FrameIndex = MI.getOperand(x).getIndex();
int64_t Offset = MFI->getObjectOffset(FrameIndex);
//int64_t Size = MF.getFrameInfo()->getObjectSize(FrameIndex);

View File

@ -1,683 +0,0 @@
//===-- AMDILUtilityFunctions.cpp - AMDIL Utility Functions ---------===//
//
// The LLVM Compiler Infrastructure
//
// This file is distributed under the University of Illinois Open Source
// License. See LICENSE.TXT for details.
//
//==-----------------------------------------------------------------------===//
//
// This file provides the implementations of functions that are declared in the
// AMDILUtilityFUnctions.h file.
//
//===----------------------------------------------------------------------===//
#include "AMDILUtilityFunctions.h"
#include "AMDILISelLowering.h"
#include "llvm/ADT/ValueMap.h"
#include "llvm/CodeGen/MachineInstr.h"
#include "llvm/Constants.h"
#include "llvm/DerivedTypes.h"
#include "llvm/Instruction.h"
#include "llvm/Instructions.h"
#include "llvm/Support/Debug.h"
#include "llvm/Support/FormattedStream.h"
#include "llvm/Type.h"
#include <cstdio>
#include <list>
#include <queue>
#define GET_OPCODE_NAME(TII, MI) \
TII->getName(MI->getOpcode())
using namespace llvm;
int64_t GET_SCALAR_SIZE(llvm::Type *A) {
return A->getScalarSizeInBits();
}
const TargetRegisterClass * getRegClassFromID(unsigned int ID) {
switch (ID) {
default:
assert(0 && "Passed in ID does not match any register classes.");
return NULL;
case AMDIL::GPRI8RegClassID:
return &AMDIL::GPRI8RegClass;
case AMDIL::GPRI16RegClassID:
return &AMDIL::GPRI16RegClass;
case AMDIL::GPRI32RegClassID:
return &AMDIL::GPRI32RegClass;
case AMDIL::GPRF32RegClassID:
return &AMDIL::GPRF32RegClass;
case AMDIL::GPRI64RegClassID:
return &AMDIL::GPRI64RegClass;
case AMDIL::GPRF64RegClassID:
return &AMDIL::GPRF64RegClass;
case AMDIL::GPRV4F32RegClassID:
return &AMDIL::GPRV4F32RegClass;
case AMDIL::GPRV4I8RegClassID:
return &AMDIL::GPRV4I8RegClass;
case AMDIL::GPRV4I16RegClassID:
return &AMDIL::GPRV4I16RegClass;
case AMDIL::GPRV4I32RegClassID:
return &AMDIL::GPRV4I32RegClass;
case AMDIL::GPRV2F32RegClassID:
return &AMDIL::GPRV2F32RegClass;
case AMDIL::GPRV2I8RegClassID:
return &AMDIL::GPRV2I8RegClass;
case AMDIL::GPRV2I16RegClassID:
return &AMDIL::GPRV2I16RegClass;
case AMDIL::GPRV2I32RegClassID:
return &AMDIL::GPRV2I32RegClass;
case AMDIL::GPRV2F64RegClassID:
return &AMDIL::GPRV2F64RegClass;
case AMDIL::GPRV2I64RegClassID:
return &AMDIL::GPRV2I64RegClass;
};
}
unsigned int getMoveInstFromID(unsigned int ID) {
switch (ID) {
default:
assert(0 && "Passed in ID does not match any move instructions.");
case AMDIL::GPRI8RegClassID:
return AMDIL::MOVE_i8;
case AMDIL::GPRI16RegClassID:
return AMDIL::MOVE_i16;
case AMDIL::GPRI32RegClassID:
return AMDIL::MOVE_i32;
case AMDIL::GPRF32RegClassID:
return AMDIL::MOVE_f32;
case AMDIL::GPRI64RegClassID:
return AMDIL::MOVE_i64;
case AMDIL::GPRF64RegClassID:
return AMDIL::MOVE_f64;
case AMDIL::GPRV4F32RegClassID:
return AMDIL::MOVE_v4f32;
case AMDIL::GPRV4I8RegClassID:
return AMDIL::MOVE_v4i8;
case AMDIL::GPRV4I16RegClassID:
return AMDIL::MOVE_v4i16;
case AMDIL::GPRV4I32RegClassID:
return AMDIL::MOVE_v4i32;
case AMDIL::GPRV2F32RegClassID:
return AMDIL::MOVE_v2f32;
case AMDIL::GPRV2I8RegClassID:
return AMDIL::MOVE_v2i8;
case AMDIL::GPRV2I16RegClassID:
return AMDIL::MOVE_v2i16;
case AMDIL::GPRV2I32RegClassID:
return AMDIL::MOVE_v2i32;
case AMDIL::GPRV2F64RegClassID:
return AMDIL::MOVE_v2f64;
case AMDIL::GPRV2I64RegClassID:
return AMDIL::MOVE_v2i64;
};
return -1;
}
unsigned int getPHIMoveInstFromID(unsigned int ID) {
switch (ID) {
default:
assert(0 && "Passed in ID does not match any move instructions.");
case AMDIL::GPRI8RegClassID:
return AMDIL::PHIMOVE_i8;
case AMDIL::GPRI16RegClassID:
return AMDIL::PHIMOVE_i16;
case AMDIL::GPRI32RegClassID:
return AMDIL::PHIMOVE_i32;
case AMDIL::GPRF32RegClassID:
return AMDIL::PHIMOVE_f32;
case AMDIL::GPRI64RegClassID:
return AMDIL::PHIMOVE_i64;
case AMDIL::GPRF64RegClassID:
return AMDIL::PHIMOVE_f64;
case AMDIL::GPRV4F32RegClassID:
return AMDIL::PHIMOVE_v4f32;
case AMDIL::GPRV4I8RegClassID:
return AMDIL::PHIMOVE_v4i8;
case AMDIL::GPRV4I16RegClassID:
return AMDIL::PHIMOVE_v4i16;
case AMDIL::GPRV4I32RegClassID:
return AMDIL::PHIMOVE_v4i32;
case AMDIL::GPRV2F32RegClassID:
return AMDIL::PHIMOVE_v2f32;
case AMDIL::GPRV2I8RegClassID:
return AMDIL::PHIMOVE_v2i8;
case AMDIL::GPRV2I16RegClassID:
return AMDIL::PHIMOVE_v2i16;
case AMDIL::GPRV2I32RegClassID:
return AMDIL::PHIMOVE_v2i32;
case AMDIL::GPRV2F64RegClassID:
return AMDIL::PHIMOVE_v2f64;
case AMDIL::GPRV2I64RegClassID:
return AMDIL::PHIMOVE_v2i64;
};
return -1;
}
const TargetRegisterClass* getRegClassFromType(unsigned int type) {
switch (type) {
default:
assert(0 && "Passed in type does not match any register classes.");
case MVT::i8:
return &AMDIL::GPRI8RegClass;
case MVT::i16:
return &AMDIL::GPRI16RegClass;
case MVT::i32:
return &AMDIL::GPRI32RegClass;
case MVT::f32:
return &AMDIL::GPRF32RegClass;
case MVT::i64:
return &AMDIL::GPRI64RegClass;
case MVT::f64:
return &AMDIL::GPRF64RegClass;
case MVT::v4f32:
return &AMDIL::GPRV4F32RegClass;
case MVT::v4i8:
return &AMDIL::GPRV4I8RegClass;
case MVT::v4i16:
return &AMDIL::GPRV4I16RegClass;
case MVT::v4i32:
return &AMDIL::GPRV4I32RegClass;
case MVT::v2f32:
return &AMDIL::GPRV2F32RegClass;
case MVT::v2i8:
return &AMDIL::GPRV2I8RegClass;
case MVT::v2i16:
return &AMDIL::GPRV2I16RegClass;
case MVT::v2i32:
return &AMDIL::GPRV2I32RegClass;
case MVT::v2f64:
return &AMDIL::GPRV2F64RegClass;
case MVT::v2i64:
return &AMDIL::GPRV2I64RegClass;
}
}
void printSDNode(const SDNode *N) {
printf("Opcode: %d isTargetOpcode: %d isMachineOpcode: %d\n",
N->getOpcode(), N->isTargetOpcode(), N->isMachineOpcode());
printf("Empty: %d OneUse: %d Size: %d NodeID: %d\n",
N->use_empty(), N->hasOneUse(), (int)N->use_size(), N->getNodeId());
for (unsigned int i = 0; i < N->getNumOperands(); ++i) {
printf("OperandNum: %d ValueCount: %d ValueType: %d\n",
i, N->getNumValues(), N->getValueType(0) .getSimpleVT().SimpleTy);
printSDValue(N->getOperand(i), 0);
}
}
void printSDValue(const SDValue &Op, int level) {
printf("\nOp: %p OpCode: %d NumOperands: %d ", (void*)&Op, Op.getOpcode(),
Op.getNumOperands());
printf("IsTarget: %d IsMachine: %d ", Op.isTargetOpcode(),
Op.isMachineOpcode());
if (Op.isMachineOpcode()) {
printf("MachineOpcode: %d\n", Op.getMachineOpcode());
} else {
printf("\n");
}
EVT vt = Op.getValueType();
printf("ValueType: %d \n", vt.getSimpleVT().SimpleTy);
printf("UseEmpty: %d OneUse: %d\n", Op.use_empty(), Op.hasOneUse());
if (level) {
printf("Children for %d:\n", level);
for (unsigned int i = 0; i < Op.getNumOperands(); ++i) {
printf("Child %d->%d:", level, i);
printSDValue(Op.getOperand(i), level - 1);
}
}
}
bool isPHIMove(unsigned int opcode) {
switch (opcode) {
default:
return false;
ExpandCaseToAllTypes(AMDIL::PHIMOVE);
return true;
}
return false;
}
bool isMove(unsigned int opcode) {
switch (opcode) {
default:
return false;
ExpandCaseToAllTypes(AMDIL::MOVE);
return true;
}
return false;
}
bool isMoveOrEquivalent(unsigned int opcode) {
switch (opcode) {
default:
return isMove(opcode) || isPHIMove(opcode);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASCHAR);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASSHORT);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASINT);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASLONG);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASDOUBLE);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASFLOAT);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASV2CHAR);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASV2SHORT);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASV2INT);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASV2FLOAT);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASV2LONG);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASV2DOUBLE);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASV4CHAR);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASV4SHORT);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASV4INT);
ExpandCaseToAllScalarTypes(AMDIL::IL_ASV4FLOAT);
case AMDIL::INTTOANY_i8:
case AMDIL::INTTOANY_i16:
case AMDIL::INTTOANY_i32:
case AMDIL::INTTOANY_f32:
case AMDIL::DLO:
case AMDIL::LLO:
case AMDIL::LLO_v2i64:
return true;
};
return false;
}
bool check_type(const Value *ptr, unsigned int addrspace) {
if (!ptr) {
return false;
}
Type *ptrType = ptr->getType();
return dyn_cast<PointerType>(ptrType)->getAddressSpace() == addrspace;
}
size_t getTypeSize(Type * const T, bool dereferencePtr) {
size_t size = 0;
if (!T) {
return size;
}
switch (T->getTypeID()) {
case Type::X86_FP80TyID:
case Type::FP128TyID:
case Type::PPC_FP128TyID:
case Type::LabelTyID:
assert(0 && "These types are not supported by this backend");
default:
case Type::FloatTyID:
case Type::DoubleTyID:
size = T->getPrimitiveSizeInBits() >> 3;
break;
case Type::PointerTyID:
size = getTypeSize(dyn_cast<PointerType>(T), dereferencePtr);
break;
case Type::IntegerTyID:
size = getTypeSize(dyn_cast<IntegerType>(T), dereferencePtr);
break;
case Type::StructTyID:
size = getTypeSize(dyn_cast<StructType>(T), dereferencePtr);
break;
case Type::ArrayTyID:
size = getTypeSize(dyn_cast<ArrayType>(T), dereferencePtr);
break;
case Type::FunctionTyID:
size = getTypeSize(dyn_cast<FunctionType>(T), dereferencePtr);
break;
case Type::VectorTyID:
size = getTypeSize(dyn_cast<VectorType>(T), dereferencePtr);
break;
};
return size;
}
size_t getTypeSize(StructType * const ST, bool dereferencePtr) {
size_t size = 0;
if (!ST) {
return size;
}
Type *curType;
StructType::element_iterator eib;
StructType::element_iterator eie;
for (eib = ST->element_begin(), eie = ST->element_end(); eib != eie; ++eib) {
curType = *eib;
size += getTypeSize(curType, dereferencePtr);
}
return size;
}
size_t getTypeSize(IntegerType * const IT, bool dereferencePtr) {
return IT ? (IT->getBitWidth() >> 3) : 0;
}
size_t getTypeSize(FunctionType * const FT, bool dereferencePtr) {
assert(0 && "Should not be able to calculate the size of an function type");
return 0;
}
size_t getTypeSize(ArrayType * const AT, bool dereferencePtr) {
return (size_t)(AT ? (getTypeSize(AT->getElementType(),
dereferencePtr) * AT->getNumElements())
: 0);
}
size_t getTypeSize(VectorType * const VT, bool dereferencePtr) {
return VT ? (VT->getBitWidth() >> 3) : 0;
}
size_t getTypeSize(PointerType * const PT, bool dereferencePtr) {
if (!PT) {
return 0;
}
Type *CT = PT->getElementType();
if (CT->getTypeID() == Type::StructTyID &&
PT->getAddressSpace() == AMDILAS::PRIVATE_ADDRESS) {
return getTypeSize(dyn_cast<StructType>(CT));
} else if (dereferencePtr) {
size_t size = 0;
for (size_t x = 0, y = PT->getNumContainedTypes(); x < y; ++x) {
size += getTypeSize(PT->getContainedType(x), dereferencePtr);
}
return size;
} else {
return 4;
}
}
size_t getTypeSize(OpaqueType * const OT, bool dereferencePtr) {
//assert(0 && "Should not be able to calculate the size of an opaque type");
return 4;
}
size_t getNumElements(Type * const T) {
size_t size = 0;
if (!T) {
return size;
}
switch (T->getTypeID()) {
case Type::X86_FP80TyID:
case Type::FP128TyID:
case Type::PPC_FP128TyID:
case Type::LabelTyID:
assert(0 && "These types are not supported by this backend");
default:
case Type::FloatTyID:
case Type::DoubleTyID:
size = 1;
break;
case Type::PointerTyID:
size = getNumElements(dyn_cast<PointerType>(T));
break;
case Type::IntegerTyID:
size = getNumElements(dyn_cast<IntegerType>(T));
break;
case Type::StructTyID:
size = getNumElements(dyn_cast<StructType>(T));
break;
case Type::ArrayTyID:
size = getNumElements(dyn_cast<ArrayType>(T));
break;
case Type::FunctionTyID:
size = getNumElements(dyn_cast<FunctionType>(T));
break;
case Type::VectorTyID:
size = getNumElements(dyn_cast<VectorType>(T));
break;
};
return size;
}
size_t getNumElements(StructType * const ST) {
size_t size = 0;
if (!ST) {
return size;
}
Type *curType;
StructType::element_iterator eib;
StructType::element_iterator eie;
for (eib = ST->element_begin(), eie = ST->element_end();
eib != eie; ++eib) {
curType = *eib;
size += getNumElements(curType);
}
return size;
}
size_t getNumElements(IntegerType * const IT) {
return (!IT) ? 0 : 1;
}
size_t getNumElements(FunctionType * const FT) {
assert(0 && "Should not be able to calculate the number of "
"elements of a function type");
return 0;
}
size_t getNumElements(ArrayType * const AT) {
return (!AT) ? 0
: (size_t)(getNumElements(AT->getElementType()) *
AT->getNumElements());
}
size_t getNumElements(VectorType * const VT) {
return (!VT) ? 0
: VT->getNumElements() * getNumElements(VT->getElementType());
}
size_t getNumElements(PointerType * const PT) {
size_t size = 0;
if (!PT) {
return size;
}
for (size_t x = 0, y = PT->getNumContainedTypes(); x < y; ++x) {
size += getNumElements(PT->getContainedType(x));
}
return size;
}
const llvm::Value *getBasePointerValue(const llvm::Value *V)
{
if (!V) {
return NULL;
}
const Value *ret = NULL;
ValueMap<const Value *, bool> ValueBitMap;
std::queue<const Value *, std::list<const Value *> > ValueQueue;
ValueQueue.push(V);
while (!ValueQueue.empty()) {
V = ValueQueue.front();
if (ValueBitMap.find(V) == ValueBitMap.end()) {
ValueBitMap[V] = true;
if (dyn_cast<Argument>(V) && dyn_cast<PointerType>(V->getType())) {
ret = V;
break;
} else if (dyn_cast<GlobalVariable>(V)) {
ret = V;
break;
} else if (dyn_cast<Constant>(V)) {
const ConstantExpr *CE = dyn_cast<ConstantExpr>(V);
if (CE) {
ValueQueue.push(CE->getOperand(0));
}
} else if (const AllocaInst *AI = dyn_cast<AllocaInst>(V)) {
ret = AI;
break;
} else if (const Instruction *I = dyn_cast<Instruction>(V)) {
uint32_t numOps = I->getNumOperands();
for (uint32_t x = 0; x < numOps; ++x) {
ValueQueue.push(I->getOperand(x));
}
} else {
// assert(0 && "Found a Value that we didn't know how to handle!");
}
}
ValueQueue.pop();
}
return ret;
}
const llvm::Value *getBasePointerValue(const llvm::MachineInstr *MI) {
const Value *moVal = NULL;
if (!MI->memoperands_empty()) {
const MachineMemOperand *memOp = (*MI->memoperands_begin());
moVal = memOp ? memOp->getValue() : NULL;
moVal = getBasePointerValue(moVal);
}
return moVal;
}
bool commaPrint(int i, llvm::raw_ostream &O) {
O << ":" << i;
return false;
}
bool isLoadInst(const llvm::TargetInstrInfo * TII, MachineInstr *MI) {
if (strstr(GET_OPCODE_NAME(TII, MI), "LOADCONST")) {
return false;
}
return strstr(GET_OPCODE_NAME(TII, MI), "LOAD");
}
bool isSWSExtLoadInst(MachineInstr *MI)
{
switch (MI->getOpcode()) {
default:
break;
ExpandCaseToByteShortTypes(AMDIL::LOCALLOAD);
ExpandCaseToByteShortTypes(AMDIL::GLOBALLOAD);
ExpandCaseToByteShortTypes(AMDIL::REGIONLOAD);
ExpandCaseToByteShortTypes(AMDIL::PRIVATELOAD);
ExpandCaseToByteShortTypes(AMDIL::CPOOLLOAD);
ExpandCaseToByteShortTypes(AMDIL::CONSTANTLOAD);
return true;
};
return false;
}
bool isExtLoadInst(const llvm::TargetInstrInfo * TII, MachineInstr *MI) {
return strstr(GET_OPCODE_NAME(TII, MI), "EXTLOAD");
}
bool isSExtLoadInst(const llvm::TargetInstrInfo * TII, MachineInstr *MI) {
return strstr(GET_OPCODE_NAME(TII, MI), "SEXTLOAD");
}
bool isAExtLoadInst(const llvm::TargetInstrInfo * TII, MachineInstr *MI) {
return strstr(GET_OPCODE_NAME(TII, MI), "AEXTLOAD");
}
bool isZExtLoadInst(const llvm::TargetInstrInfo * TII, MachineInstr *MI) {
return strstr(GET_OPCODE_NAME(TII, MI), "ZEXTLOAD");
}
bool isStoreInst(const llvm::TargetInstrInfo * TII, MachineInstr *MI) {
return strstr(GET_OPCODE_NAME(TII, MI), "STORE");
}
bool isTruncStoreInst(const llvm::TargetInstrInfo * TII, MachineInstr *MI) {
return strstr(GET_OPCODE_NAME(TII, MI), "TRUNCSTORE");
}
bool isAtomicInst(const llvm::TargetInstrInfo * TII, MachineInstr *MI) {
return strstr(GET_OPCODE_NAME(TII, MI), "ATOM");
}
bool isVolatileInst(const llvm::TargetInstrInfo * TII, MachineInstr *MI) {
if (!MI->memoperands_empty()) {
for (MachineInstr::mmo_iterator mob = MI->memoperands_begin(),
moe = MI->memoperands_end(); mob != moe; ++mob) {
// If there is a volatile mem operand, this is a volatile instruction.
if ((*mob)->isVolatile()) {
return true;
}
}
}
return false;
}
bool isGlobalInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI)
{
return strstr(GET_OPCODE_NAME(TII, MI), "GLOBAL");
}
bool isPrivateInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI)
{
return strstr(GET_OPCODE_NAME(TII, MI), "PRIVATE");
}
bool isConstantInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI)
{
return strstr(GET_OPCODE_NAME(TII, MI), "CONSTANT")
|| strstr(GET_OPCODE_NAME(TII, MI), "CPOOL");
}
bool isRegionInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI)
{
return strstr(GET_OPCODE_NAME(TII, MI), "REGION");
}
bool isLocalInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI)
{
return strstr(GET_OPCODE_NAME(TII, MI), "LOCAL");
}
bool isImageInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI)
{
return strstr(GET_OPCODE_NAME(TII, MI), "IMAGE");
}
bool isAppendInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI)
{
return strstr(GET_OPCODE_NAME(TII, MI), "APPEND");
}
bool isRegionAtomic(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI)
{
return strstr(GET_OPCODE_NAME(TII, MI), "ATOM_R");
}
bool isLocalAtomic(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI)
{
return strstr(GET_OPCODE_NAME(TII, MI), "ATOM_L");
}
bool isGlobalAtomic(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI)
{
return strstr(GET_OPCODE_NAME(TII, MI), "ATOM_G")
|| isArenaAtomic(TII, MI);
}
bool isArenaAtomic(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI)
{
return strstr(GET_OPCODE_NAME(TII, MI), "ATOM_A");
}
const char* getSrcSwizzle(unsigned idx) {
const char *srcSwizzles[] = {
"", ".x000", ".0x00", ".00x0", ".000x", ".y000", ".0y00", ".00y0", ".000y",
".z000", ".0z00", ".00z0", ".000z", ".w000", ".0w00", ".00w0", ".000w",
".xy00", ".00xy", ".zw00", ".00zw", ".xyz0", ".0xyz", ".xyzw", ".0000",
".xxxx", ".yyyy", ".zzzz", ".wwww", ".xyxy", ".zwzw", ".xzxz", ".ywyw",
".x0y0", ".0x0y", ".xy_neg(y)", "_neg(yw)", "_neg(x)", ".xy_neg(xy)",
"_neg(xyzw)", ".0yzw", ".x0zw", ".xy0w", ".x", ".y", ".z", ".w", ".xy",
".zw"
};
assert(idx < sizeof(srcSwizzles)/sizeof(srcSwizzles[0])
&& "Idx passed in is invalid!");
return srcSwizzles[idx];
}
const char* getDstSwizzle(unsigned idx) {
const char *dstSwizzles[] = {
"", ".x___", ".xy__", ".xyz_", ".xyzw", "._y__", "._yz_", "._yzw", ".__z_",
".__zw", ".___w", ".x_zw", ".xy_w", ".x_z_", ".x__w", "._y_w",
};
assert(idx < sizeof(dstSwizzles)/sizeof(dstSwizzles[0])
&& "Idx passed in is invalid!");
return dstSwizzles[idx];
}
/// Helper function to get the currently set flags
void getAsmPrinterFlags(MachineInstr *MI, AMDILAS::InstrResEnc &curRes)
{
// We need 16 bits of information, but LLVMr127097 cut the field in half.
// So we have to use two different fields to store all of our information.
uint16_t upper = MI->getFlags() << 8;
uint16_t lower = MI->getAsmPrinterFlags();
curRes.u16all = upper | lower;
}
/// Helper function to clear the currently set flags and add the new flags.
void setAsmPrinterFlags(MachineInstr *MI, AMDILAS::InstrResEnc &curRes)
{
// We need 16 bits of information, but LLVMr127097 cut the field in half.
// So we have to use two different fields to store all of our information.
MI->clearAsmPrinterFlags();
MI->setFlags(0);
uint8_t lower = curRes.u16all & 0xFF;
uint8_t upper = (curRes.u16all >> 8) & 0xFF;
MI->setFlags(upper);
MI->setAsmPrinterFlag((llvm::MachineInstr::CommentFlag)lower);
}

View File

@ -7,191 +7,12 @@
//
//==-----------------------------------------------------------------------===//
//
// This file provides declarations for functions that are used across different
// classes and provide various conversions or utility to shorten the code
// This file provides helper macros for expanding case statements.
//
//===----------------------------------------------------------------------===//
#ifndef AMDILUTILITYFUNCTIONS_H_
#define AMDILUTILITYFUNCTIONS_H_
#include "AMDIL.h"
#include "AMDILTargetMachine.h"
#include "llvm/ADT/SmallVector.h"
// Utility functions from ID
//
namespace llvm {
class TargetRegisterClass;
class SDValue;
class SDNode;
class Value;
class Type;
class StructType;
class IntegerType;
class FunctionType;
class VectorType;
class ArrayType;
class PointerType;
class OpaqueType;
class MachineInstr;
}
enum SrcSwizzles {
AMDIL_SRC_SWIZZLE_DEFAULT = 0,
AMDIL_SRC_SWIZZLE_X000,
AMDIL_SRC_SWIZZLE_0X00,
AMDIL_SRC_SWIZZLE_00X0,
AMDIL_SRC_SWIZZLE_000X,
AMDIL_SRC_SWIZZLE_Y000,
AMDIL_SRC_SWIZZLE_0Y00,
AMDIL_SRC_SWIZZLE_00Y0,
AMDIL_SRC_SWIZZLE_000Y,
AMDIL_SRC_SWIZZLE_Z000,
AMDIL_SRC_SWIZZLE_0Z00,
AMDIL_SRC_SWIZZLE_00Z0,
AMDIL_SRC_SWIZZLE_000Z,
AMDIL_SRC_SWIZZLE_W000,
AMDIL_SRC_SWIZZLE_0W00,
AMDIL_SRC_SWIZZLE_00W0,
AMDIL_SRC_SWIZZLE_000W,
AMDIL_SRC_SWIZZLE_XY00,
AMDIL_SRC_SWIZZLE_00XY,
AMDIL_SRC_SWIZZLE_ZW00,
AMDIL_SRC_SWIZZLE_00ZW,
AMDIL_SRC_SWIZZLE_XYZ0,
AMDIL_SRC_SWIZZLE_0XYZ,
AMDIL_SRC_SWIZZLE_XYZW,
AMDIL_SRC_SWIZZLE_0000,
AMDIL_SRC_SWIZZLE_XXXX,
AMDIL_SRC_SWIZZLE_YYYY,
AMDIL_SRC_SWIZZLE_ZZZZ,
AMDIL_SRC_SWIZZLE_WWWW,
AMDIL_SRC_SWIZZLE_XYXY,
AMDIL_SRC_SWIZZLE_ZWZW,
AMDIL_SRC_SWIZZLE_XZXZ,
AMDIL_SRC_SWIZZLE_YWYW,
AMDIL_SRC_SWIZZLE_X0Y0,
AMDIL_SRC_SWIZZLE_0X0Y,
AMDIL_SRC_SWIZZLE_XY_NEGY,
AMDIL_SRC_SWIZZLE_NEGYW,
AMDIL_SRC_SWIZZLE_NEGX,
AMDIL_SRC_SWIZZLE_XY_NEGXY,
AMDIL_SRC_SWIZZLE_NEG_XYZW,
AMDIL_SRC_SWIZZLE_0YZW,
AMDIL_SRC_SWIZZLE_X0ZW,
AMDIL_SRC_SWIZZLE_XY0W,
AMDIL_SRC_SWIZZLE_X,
AMDIL_SRC_SWIZZLE_Y,
AMDIL_SRC_SWIZZLE_Z,
AMDIL_SRC_SWIZZLE_W,
AMDIL_SRC_SWIZZLE_XY,
AMDIL_SRC_SWIZZLE_ZW,
AMDIL_SRC_SWIZZLE_LAST
};
enum DstSwizzles {
AMDIL_DST_SWIZZLE_DEFAULT = 0,
AMDIL_DST_SWIZZLE_X___,
AMDIL_DST_SWIZZLE_XY__,
AMDIL_DST_SWIZZLE_XYZ_,
AMDIL_DST_SWIZZLE_XYZW,
AMDIL_DST_SWIZZLE__Y__,
AMDIL_DST_SWIZZLE__YZ_,
AMDIL_DST_SWIZZLE__YZW,
AMDIL_DST_SWIZZLE___Z_,
AMDIL_DST_SWIZZLE___ZW,
AMDIL_DST_SWIZZLE____W,
AMDIL_DST_SWIZZLE_X_ZW,
AMDIL_DST_SWIZZLE_XY_W,
AMDIL_DST_SWIZZLE_X_Z_,
AMDIL_DST_SWIZZLE_X__W,
AMDIL_DST_SWIZZLE__Y_W,
AMDIL_DST_SWIZZLE_LAST
};
// Function to get the correct src swizzle string from ID
const char *getSrcSwizzle(unsigned);
// Function to get the correct dst swizzle string from ID
const char *getDstSwizzle(unsigned);
const llvm::TargetRegisterClass *getRegClassFromID(unsigned int ID);
unsigned int getMoveInstFromID(unsigned int ID);
unsigned int getPHIMoveInstFromID(unsigned int ID);
// Utility functions from Type.
const llvm::TargetRegisterClass *getRegClassFromType(unsigned int type);
unsigned int getTargetIndependentMoveFromType(unsigned int type);
// Debug functions for SDNode and SDValue.
void printSDValue(const llvm::SDValue &Op, int level);
void printSDNode(const llvm::SDNode *N);
// Functions to check if an opcode is a specific type.
bool isMove(unsigned int opcode);
bool isPHIMove(unsigned int opcode);
bool isMoveOrEquivalent(unsigned int opcode);
// Function to check address space
bool check_type(const llvm::Value *ptr, unsigned int addrspace);
// Group of functions that recursively calculate the size of a structure based
// on it's sub-types.
size_t getTypeSize(llvm::Type * const T, bool dereferencePtr = false);
size_t
getTypeSize(llvm::StructType * const ST, bool dereferencePtr = false);
size_t
getTypeSize(llvm::IntegerType * const IT, bool dereferencePtr = false);
size_t
getTypeSize(llvm::FunctionType * const FT, bool dereferencePtr = false);
size_t
getTypeSize(llvm::ArrayType * const AT, bool dereferencePtr = false);
size_t
getTypeSize(llvm::VectorType * const VT, bool dereferencePtr = false);
size_t
getTypeSize(llvm::PointerType * const PT, bool dereferencePtr = false);
size_t
getTypeSize(llvm::OpaqueType * const OT, bool dereferencePtr = false);
// Group of functions that recursively calculate the number of elements of a
// structure based on it's sub-types.
size_t getNumElements(llvm::Type * const T);
size_t getNumElements(llvm::StructType * const ST);
size_t getNumElements(llvm::IntegerType * const IT);
size_t getNumElements(llvm::FunctionType * const FT);
size_t getNumElements(llvm::ArrayType * const AT);
size_t getNumElements(llvm::VectorType * const VT);
size_t getNumElements(llvm::PointerType * const PT);
size_t getNumElements(llvm::OpaqueType * const OT);
const llvm::Value *getBasePointerValue(const llvm::Value *V);
const llvm::Value *getBasePointerValue(const llvm::MachineInstr *MI);
int64_t GET_SCALAR_SIZE(llvm::Type* A);
// Helper functions that check the opcode for status information
bool isLoadInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isExtLoadInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isSWSExtLoadInst(llvm::MachineInstr *MI);
bool isSExtLoadInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isZExtLoadInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isAExtLoadInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isStoreInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isTruncStoreInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isAtomicInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isVolatileInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isGlobalInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isPrivateInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isConstantInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isRegionInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isLocalInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isImageInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isAppendInst(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isRegionAtomic(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isLocalAtomic(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isGlobalAtomic(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
bool isArenaAtomic(const llvm::TargetInstrInfo * TII, llvm::MachineInstr *MI);
// Macros that are used to help with switch statements for various data types
// However, these macro's do not return anything unlike the second set below.
#define ExpandCaseTo32bitIntTypes(Instr) \
@ -354,9 +175,4 @@ case Instr##_v4f32: \
case Instr##_v2i64: \
case Instr##_v2f64:
bool commaPrint(int i, llvm::raw_ostream &O);
/// Helper function to get the currently get/set flags.
void getAsmPrinterFlags(llvm::MachineInstr *MI, llvm::AMDILAS::InstrResEnc &curRes);
void setAsmPrinterFlags(llvm::MachineInstr *MI, llvm::AMDILAS::InstrResEnc &curRes);
#endif // AMDILUTILITYFUNCTIONS_H_

View File

@ -37,7 +37,6 @@ CPP_SOURCES := \
AMDILSIDevice.cpp \
AMDILSubtarget.cpp \
AMDILTargetMachine.cpp \
AMDILUtilityFunctions.cpp \
AMDGPUTargetMachine.cpp \
AMDGPUISelLowering.cpp \
AMDGPUConvertToISA.cpp \

View File

@ -724,44 +724,5 @@ RegElement maskBitToElement(unsigned int maskBit)
}
}
unsigned int dstSwizzleToWriteMask(unsigned swizzle)
{
switch(swizzle) {
default:
case AMDIL_DST_SWIZZLE_DEFAULT:
return WRITE_MASK_X | WRITE_MASK_Y | WRITE_MASK_Z | WRITE_MASK_W;
case AMDIL_DST_SWIZZLE_X___:
return WRITE_MASK_X;
case AMDIL_DST_SWIZZLE_XY__:
return WRITE_MASK_X | WRITE_MASK_Y;
case AMDIL_DST_SWIZZLE_XYZ_:
return WRITE_MASK_X | WRITE_MASK_Y | WRITE_MASK_Z;
case AMDIL_DST_SWIZZLE_XYZW:
return WRITE_MASK_X | WRITE_MASK_Y | WRITE_MASK_Z | WRITE_MASK_W;
case AMDIL_DST_SWIZZLE__Y__:
return WRITE_MASK_Y;
case AMDIL_DST_SWIZZLE__YZ_:
return WRITE_MASK_Y | WRITE_MASK_Z;
case AMDIL_DST_SWIZZLE__YZW:
return WRITE_MASK_Y | WRITE_MASK_Z | WRITE_MASK_W;
case AMDIL_DST_SWIZZLE___Z_:
return WRITE_MASK_Z;
case AMDIL_DST_SWIZZLE___ZW:
return WRITE_MASK_Z | WRITE_MASK_W;
case AMDIL_DST_SWIZZLE____W:
return WRITE_MASK_W;
case AMDIL_DST_SWIZZLE_X_ZW:
return WRITE_MASK_X | WRITE_MASK_Z | WRITE_MASK_W;
case AMDIL_DST_SWIZZLE_XY_W:
return WRITE_MASK_X | WRITE_MASK_Y | WRITE_MASK_W;
case AMDIL_DST_SWIZZLE_X_Z_:
return WRITE_MASK_X | WRITE_MASK_Z;
case AMDIL_DST_SWIZZLE_X__W:
return WRITE_MASK_X | WRITE_MASK_W;
case AMDIL_DST_SWIZZLE__Y_W:
return WRITE_MASK_Y | WRITE_MASK_W;
}
}
#include "AMDILGenCodeEmitter.inc"