mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-22 18:54:02 +01:00
[Hexagon] Break up DAG mutations into separate classes, move to subtarget
llvm-svn: 311895
This commit is contained in:
parent
ad38f1f0c7
commit
ab8e9a3099
@ -12,11 +12,9 @@
|
|||||||
//
|
//
|
||||||
//===----------------------------------------------------------------------===//
|
//===----------------------------------------------------------------------===//
|
||||||
|
|
||||||
#include "HexagonInstrInfo.h"
|
|
||||||
#include "HexagonMachineScheduler.h"
|
#include "HexagonMachineScheduler.h"
|
||||||
#include "HexagonSubtarget.h"
|
#include "HexagonSubtarget.h"
|
||||||
#include "llvm/CodeGen/MachineLoopInfo.h"
|
#include "llvm/CodeGen/MachineLoopInfo.h"
|
||||||
#include "llvm/CodeGen/ScheduleDAGMutation.h"
|
|
||||||
#include "llvm/IR/Function.h"
|
#include "llvm/IR/Function.h"
|
||||||
|
|
||||||
#include <iomanip>
|
#include <iomanip>
|
||||||
@ -25,9 +23,6 @@
|
|||||||
static cl::opt<bool> IgnoreBBRegPressure("ignore-bb-reg-pressure",
|
static cl::opt<bool> IgnoreBBRegPressure("ignore-bb-reg-pressure",
|
||||||
cl::Hidden, cl::ZeroOrMore, cl::init(false));
|
cl::Hidden, cl::ZeroOrMore, cl::init(false));
|
||||||
|
|
||||||
static cl::opt<bool> SchedPredsCloser("sched-preds-closer",
|
|
||||||
cl::Hidden, cl::ZeroOrMore, cl::init(true));
|
|
||||||
|
|
||||||
static cl::opt<unsigned> SchedDebugVerboseLevel("misched-verbose-level",
|
static cl::opt<unsigned> SchedDebugVerboseLevel("misched-verbose-level",
|
||||||
cl::Hidden, cl::ZeroOrMore, cl::init(1));
|
cl::Hidden, cl::ZeroOrMore, cl::init(1));
|
||||||
|
|
||||||
@ -40,9 +35,6 @@ static cl::opt<bool> BotUseShorterTie("bot-use-shorter-tie",
|
|||||||
static cl::opt<bool> DisableTCTie("disable-tc-tie",
|
static cl::opt<bool> DisableTCTie("disable-tc-tie",
|
||||||
cl::Hidden, cl::ZeroOrMore, cl::init(false));
|
cl::Hidden, cl::ZeroOrMore, cl::init(false));
|
||||||
|
|
||||||
static cl::opt<bool> SchedRetvalOptimization("sched-retval-optimization",
|
|
||||||
cl::Hidden, cl::ZeroOrMore, cl::init(true));
|
|
||||||
|
|
||||||
// Check if the scheduler should penalize instructions that are available to
|
// Check if the scheduler should penalize instructions that are available to
|
||||||
// early due to a zero-latency dependence.
|
// early due to a zero-latency dependence.
|
||||||
static cl::opt<bool> CheckEarlyAvail("check-early-avail", cl::Hidden,
|
static cl::opt<bool> CheckEarlyAvail("check-early-avail", cl::Hidden,
|
||||||
@ -52,77 +44,6 @@ using namespace llvm;
|
|||||||
|
|
||||||
#define DEBUG_TYPE "machine-scheduler"
|
#define DEBUG_TYPE "machine-scheduler"
|
||||||
|
|
||||||
// Check if a call and subsequent A2_tfrpi instructions should maintain
|
|
||||||
// scheduling affinity. We are looking for the TFRI to be consumed in
|
|
||||||
// the next instruction. This should help reduce the instances of
|
|
||||||
// double register pairs being allocated and scheduled before a call
|
|
||||||
// when not used until after the call. This situation is exacerbated
|
|
||||||
// by the fact that we allocate the pair from the callee saves list,
|
|
||||||
// leading to excess spills and restores.
|
|
||||||
bool HexagonCallMutation::shouldTFRICallBind(const HexagonInstrInfo &HII,
|
|
||||||
const SUnit &Inst1, const SUnit &Inst2) const {
|
|
||||||
if (Inst1.getInstr()->getOpcode() != Hexagon::A2_tfrpi)
|
|
||||||
return false;
|
|
||||||
|
|
||||||
// TypeXTYPE are 64 bit operations.
|
|
||||||
unsigned Type = HII.getType(*Inst2.getInstr());
|
|
||||||
if (Type == HexagonII::TypeS_2op || Type == HexagonII::TypeS_3op ||
|
|
||||||
Type == HexagonII::TypeALU64 || Type == HexagonII::TypeM)
|
|
||||||
return true;
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
void HexagonCallMutation::apply(ScheduleDAGInstrs *DAG) {
|
|
||||||
SUnit* LastSequentialCall = nullptr;
|
|
||||||
unsigned VRegHoldingRet = 0;
|
|
||||||
unsigned RetRegister;
|
|
||||||
SUnit* LastUseOfRet = nullptr;
|
|
||||||
auto &TRI = *DAG->MF.getSubtarget().getRegisterInfo();
|
|
||||||
auto &HII = *DAG->MF.getSubtarget<HexagonSubtarget>().getInstrInfo();
|
|
||||||
|
|
||||||
// Currently we only catch the situation when compare gets scheduled
|
|
||||||
// before preceding call.
|
|
||||||
for (unsigned su = 0, e = DAG->SUnits.size(); su != e; ++su) {
|
|
||||||
// Remember the call.
|
|
||||||
if (DAG->SUnits[su].getInstr()->isCall())
|
|
||||||
LastSequentialCall = &DAG->SUnits[su];
|
|
||||||
// Look for a compare that defines a predicate.
|
|
||||||
else if (DAG->SUnits[su].getInstr()->isCompare() && LastSequentialCall)
|
|
||||||
DAG->SUnits[su].addPred(SDep(LastSequentialCall, SDep::Barrier));
|
|
||||||
// Look for call and tfri* instructions.
|
|
||||||
else if (SchedPredsCloser && LastSequentialCall && su > 1 && su < e-1 &&
|
|
||||||
shouldTFRICallBind(HII, DAG->SUnits[su], DAG->SUnits[su+1]))
|
|
||||||
DAG->SUnits[su].addPred(SDep(&DAG->SUnits[su-1], SDep::Barrier));
|
|
||||||
// Prevent redundant register copies between two calls, which are caused by
|
|
||||||
// both the return value and the argument for the next call being in %R0.
|
|
||||||
// Example:
|
|
||||||
// 1: <call1>
|
|
||||||
// 2: %VregX = COPY %R0
|
|
||||||
// 3: <use of %VregX>
|
|
||||||
// 4: %R0 = ...
|
|
||||||
// 5: <call2>
|
|
||||||
// The scheduler would often swap 3 and 4, so an additional register is
|
|
||||||
// needed. This code inserts a Barrier dependence between 3 & 4 to prevent
|
|
||||||
// this. The same applies for %D0 and %V0/%W0, which are also handled.
|
|
||||||
else if (SchedRetvalOptimization) {
|
|
||||||
const MachineInstr *MI = DAG->SUnits[su].getInstr();
|
|
||||||
if (MI->isCopy() && (MI->readsRegister(Hexagon::R0, &TRI) ||
|
|
||||||
MI->readsRegister(Hexagon::V0, &TRI))) {
|
|
||||||
// %vregX = COPY %R0
|
|
||||||
VRegHoldingRet = MI->getOperand(0).getReg();
|
|
||||||
RetRegister = MI->getOperand(1).getReg();
|
|
||||||
LastUseOfRet = nullptr;
|
|
||||||
} else if (VRegHoldingRet && MI->readsVirtualRegister(VRegHoldingRet))
|
|
||||||
// <use of %vregX>
|
|
||||||
LastUseOfRet = &DAG->SUnits[su];
|
|
||||||
else if (LastUseOfRet && MI->definesRegister(RetRegister, &TRI))
|
|
||||||
// %R0 = ...
|
|
||||||
DAG->SUnits[su].addPred(SDep(LastUseOfRet, SDep::Barrier));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
/// Save the last formed packet
|
/// Save the last formed packet
|
||||||
void VLIWResourceModel::savePacket() {
|
void VLIWResourceModel::savePacket() {
|
||||||
OldPacket = Packet;
|
OldPacket = Packet;
|
||||||
|
@ -249,14 +249,6 @@ protected:
|
|||||||
#endif
|
#endif
|
||||||
};
|
};
|
||||||
|
|
||||||
class HexagonCallMutation : public ScheduleDAGMutation {
|
|
||||||
public:
|
|
||||||
void apply(ScheduleDAGInstrs *DAG) override;
|
|
||||||
private:
|
|
||||||
bool shouldTFRICallBind(const HexagonInstrInfo &HII,
|
|
||||||
const SUnit &Inst1, const SUnit &Inst2) const;
|
|
||||||
};
|
|
||||||
|
|
||||||
} // namespace
|
} // namespace
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
@ -87,6 +87,13 @@ static cl::opt<bool> EnablePredicatedCalls("hexagon-pred-calls",
|
|||||||
cl::Hidden, cl::ZeroOrMore, cl::init(false),
|
cl::Hidden, cl::ZeroOrMore, cl::init(false),
|
||||||
cl::desc("Consider calls to be predicable"));
|
cl::desc("Consider calls to be predicable"));
|
||||||
|
|
||||||
|
static cl::opt<bool> SchedPredsCloser("sched-preds-closer",
|
||||||
|
cl::Hidden, cl::ZeroOrMore, cl::init(true));
|
||||||
|
|
||||||
|
static cl::opt<bool> SchedRetvalOptimization("sched-retval-optimization",
|
||||||
|
cl::Hidden, cl::ZeroOrMore, cl::init(true));
|
||||||
|
|
||||||
|
|
||||||
void HexagonSubtarget::initializeEnvironment() {
|
void HexagonSubtarget::initializeEnvironment() {
|
||||||
UseMemOps = false;
|
UseMemOps = false;
|
||||||
ModeIEEERndNear = false;
|
ModeIEEERndNear = false;
|
||||||
@ -126,6 +133,121 @@ HexagonSubtarget::initializeSubtargetDependencies(StringRef CPU, StringRef FS) {
|
|||||||
return *this;
|
return *this;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void HexagonSubtarget::UsrOverflowMutation::apply(ScheduleDAGInstrs *DAG) {
|
||||||
|
for (SUnit &SU : DAG->SUnits) {
|
||||||
|
if (!SU.isInstr())
|
||||||
|
continue;
|
||||||
|
SmallVector<SDep, 4> Erase;
|
||||||
|
for (auto &D : SU.Preds)
|
||||||
|
if (D.getKind() == SDep::Output && D.getReg() == Hexagon::USR_OVF)
|
||||||
|
Erase.push_back(D);
|
||||||
|
for (auto &E : Erase)
|
||||||
|
SU.removePred(E);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void HexagonSubtarget::HVXMemLatencyMutation::apply(ScheduleDAGInstrs *DAG) {
|
||||||
|
for (SUnit &SU : DAG->SUnits) {
|
||||||
|
// Update the latency of chain edges between v60 vector load or store
|
||||||
|
// instructions to be 1. These instruction cannot be scheduled in the
|
||||||
|
// same packet.
|
||||||
|
MachineInstr &MI1 = *SU.getInstr();
|
||||||
|
auto *QII = static_cast<const HexagonInstrInfo*>(DAG->TII);
|
||||||
|
bool IsStoreMI1 = MI1.mayStore();
|
||||||
|
bool IsLoadMI1 = MI1.mayLoad();
|
||||||
|
if (!QII->isHVXVec(MI1) || !(IsStoreMI1 || IsLoadMI1))
|
||||||
|
continue;
|
||||||
|
for (SDep &SI : SU.Succs) {
|
||||||
|
if (SI.getKind() != SDep::Order || SI.getLatency() != 0)
|
||||||
|
continue;
|
||||||
|
MachineInstr &MI2 = *SI.getSUnit()->getInstr();
|
||||||
|
if (!QII->isHVXVec(MI2))
|
||||||
|
continue;
|
||||||
|
if ((IsStoreMI1 && MI2.mayStore()) || (IsLoadMI1 && MI2.mayLoad())) {
|
||||||
|
SI.setLatency(1);
|
||||||
|
SU.setHeightDirty();
|
||||||
|
// Change the dependence in the opposite direction too.
|
||||||
|
for (SDep &PI : SI.getSUnit()->Preds) {
|
||||||
|
if (PI.getSUnit() != &SU || PI.getKind() != SDep::Order)
|
||||||
|
continue;
|
||||||
|
PI.setLatency(1);
|
||||||
|
SI.getSUnit()->setDepthDirty();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check if a call and subsequent A2_tfrpi instructions should maintain
|
||||||
|
// scheduling affinity. We are looking for the TFRI to be consumed in
|
||||||
|
// the next instruction. This should help reduce the instances of
|
||||||
|
// double register pairs being allocated and scheduled before a call
|
||||||
|
// when not used until after the call. This situation is exacerbated
|
||||||
|
// by the fact that we allocate the pair from the callee saves list,
|
||||||
|
// leading to excess spills and restores.
|
||||||
|
bool HexagonSubtarget::CallMutation::shouldTFRICallBind(
|
||||||
|
const HexagonInstrInfo &HII, const SUnit &Inst1,
|
||||||
|
const SUnit &Inst2) const {
|
||||||
|
if (Inst1.getInstr()->getOpcode() != Hexagon::A2_tfrpi)
|
||||||
|
return false;
|
||||||
|
|
||||||
|
// TypeXTYPE are 64 bit operations.
|
||||||
|
unsigned Type = HII.getType(*Inst2.getInstr());
|
||||||
|
return Type == HexagonII::TypeS_2op || Type == HexagonII::TypeS_3op ||
|
||||||
|
Type == HexagonII::TypeALU64 || Type == HexagonII::TypeM;
|
||||||
|
}
|
||||||
|
|
||||||
|
void HexagonSubtarget::CallMutation::apply(ScheduleDAGInstrs *DAG) {
|
||||||
|
SUnit* LastSequentialCall = nullptr;
|
||||||
|
unsigned VRegHoldingRet = 0;
|
||||||
|
unsigned RetRegister;
|
||||||
|
SUnit* LastUseOfRet = nullptr;
|
||||||
|
auto &TRI = *DAG->MF.getSubtarget().getRegisterInfo();
|
||||||
|
auto &HII = *DAG->MF.getSubtarget<HexagonSubtarget>().getInstrInfo();
|
||||||
|
|
||||||
|
// Currently we only catch the situation when compare gets scheduled
|
||||||
|
// before preceding call.
|
||||||
|
for (unsigned su = 0, e = DAG->SUnits.size(); su != e; ++su) {
|
||||||
|
// Remember the call.
|
||||||
|
if (DAG->SUnits[su].getInstr()->isCall())
|
||||||
|
LastSequentialCall = &DAG->SUnits[su];
|
||||||
|
// Look for a compare that defines a predicate.
|
||||||
|
else if (DAG->SUnits[su].getInstr()->isCompare() && LastSequentialCall)
|
||||||
|
DAG->SUnits[su].addPred(SDep(LastSequentialCall, SDep::Barrier));
|
||||||
|
// Look for call and tfri* instructions.
|
||||||
|
else if (SchedPredsCloser && LastSequentialCall && su > 1 && su < e-1 &&
|
||||||
|
shouldTFRICallBind(HII, DAG->SUnits[su], DAG->SUnits[su+1]))
|
||||||
|
DAG->SUnits[su].addPred(SDep(&DAG->SUnits[su-1], SDep::Barrier));
|
||||||
|
// Prevent redundant register copies between two calls, which are caused by
|
||||||
|
// both the return value and the argument for the next call being in %R0.
|
||||||
|
// Example:
|
||||||
|
// 1: <call1>
|
||||||
|
// 2: %VregX = COPY %R0
|
||||||
|
// 3: <use of %VregX>
|
||||||
|
// 4: %R0 = ...
|
||||||
|
// 5: <call2>
|
||||||
|
// The scheduler would often swap 3 and 4, so an additional register is
|
||||||
|
// needed. This code inserts a Barrier dependence between 3 & 4 to prevent
|
||||||
|
// this. The same applies for %D0 and %V0/%W0, which are also handled.
|
||||||
|
else if (SchedRetvalOptimization) {
|
||||||
|
const MachineInstr *MI = DAG->SUnits[su].getInstr();
|
||||||
|
if (MI->isCopy() && (MI->readsRegister(Hexagon::R0, &TRI) ||
|
||||||
|
MI->readsRegister(Hexagon::V0, &TRI))) {
|
||||||
|
// %vregX = COPY %R0
|
||||||
|
VRegHoldingRet = MI->getOperand(0).getReg();
|
||||||
|
RetRegister = MI->getOperand(1).getReg();
|
||||||
|
LastUseOfRet = nullptr;
|
||||||
|
} else if (VRegHoldingRet && MI->readsVirtualRegister(VRegHoldingRet))
|
||||||
|
// <use of %vregX>
|
||||||
|
LastUseOfRet = &DAG->SUnits[su];
|
||||||
|
else if (LastUseOfRet && MI->definesRegister(RetRegister, &TRI))
|
||||||
|
// %R0 = ...
|
||||||
|
DAG->SUnits[su].addPred(SDep(LastUseOfRet, SDep::Barrier));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
HexagonSubtarget::HexagonSubtarget(const Triple &TT, StringRef CPU,
|
HexagonSubtarget::HexagonSubtarget(const Triple &TT, StringRef CPU,
|
||||||
StringRef FS, const TargetMachine &TM)
|
StringRef FS, const TargetMachine &TM)
|
||||||
: HexagonGenSubtargetInfo(TT, CPU, FS), CPUString(CPU),
|
: HexagonGenSubtargetInfo(TT, CPU, FS), CPUString(CPU),
|
||||||
@ -204,59 +326,16 @@ void HexagonSubtarget::adjustSchedDependency(SUnit *Src, SUnit *Dst,
|
|||||||
updateLatency(*SrcInst, *DstInst, Dep);
|
updateLatency(*SrcInst, *DstInst, Dep);
|
||||||
}
|
}
|
||||||
|
|
||||||
void HexagonSubtarget::HexagonDAGMutation::apply(ScheduleDAGInstrs *DAG) {
|
|
||||||
for (auto &SU : DAG->SUnits) {
|
|
||||||
if (!SU.isInstr())
|
|
||||||
continue;
|
|
||||||
SmallVector<SDep, 4> Erase;
|
|
||||||
for (auto &D : SU.Preds)
|
|
||||||
if (D.getKind() == SDep::Output && D.getReg() == Hexagon::USR_OVF)
|
|
||||||
Erase.push_back(D);
|
|
||||||
for (auto &E : Erase)
|
|
||||||
SU.removePred(E);
|
|
||||||
}
|
|
||||||
|
|
||||||
for (auto &SU : DAG->SUnits) {
|
|
||||||
// Update the latency of chain edges between v60 vector load or store
|
|
||||||
// instructions to be 1. These instruction cannot be scheduled in the
|
|
||||||
// same packet.
|
|
||||||
MachineInstr &MI1 = *SU.getInstr();
|
|
||||||
auto *QII = static_cast<const HexagonInstrInfo*>(DAG->TII);
|
|
||||||
bool IsStoreMI1 = MI1.mayStore();
|
|
||||||
bool IsLoadMI1 = MI1.mayLoad();
|
|
||||||
if (!QII->isHVXVec(MI1) || !(IsStoreMI1 || IsLoadMI1))
|
|
||||||
continue;
|
|
||||||
for (auto &SI : SU.Succs) {
|
|
||||||
if (SI.getKind() != SDep::Order || SI.getLatency() != 0)
|
|
||||||
continue;
|
|
||||||
MachineInstr &MI2 = *SI.getSUnit()->getInstr();
|
|
||||||
if (!QII->isHVXVec(MI2))
|
|
||||||
continue;
|
|
||||||
if ((IsStoreMI1 && MI2.mayStore()) || (IsLoadMI1 && MI2.mayLoad())) {
|
|
||||||
SI.setLatency(1);
|
|
||||||
SU.setHeightDirty();
|
|
||||||
// Change the dependence in the opposite direction too.
|
|
||||||
for (auto &PI : SI.getSUnit()->Preds) {
|
|
||||||
if (PI.getSUnit() != &SU || PI.getKind() != SDep::Order)
|
|
||||||
continue;
|
|
||||||
PI.setLatency(1);
|
|
||||||
SI.getSUnit()->setDepthDirty();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void HexagonSubtarget::getPostRAMutations(
|
void HexagonSubtarget::getPostRAMutations(
|
||||||
std::vector<std::unique_ptr<ScheduleDAGMutation>> &Mutations) const {
|
std::vector<std::unique_ptr<ScheduleDAGMutation>> &Mutations) const {
|
||||||
Mutations.push_back(
|
Mutations.push_back(llvm::make_unique<UsrOverflowMutation>());
|
||||||
llvm::make_unique<HexagonSubtarget::HexagonDAGMutation>());
|
Mutations.push_back(llvm::make_unique<HVXMemLatencyMutation>());
|
||||||
}
|
}
|
||||||
|
|
||||||
void HexagonSubtarget::getSMSMutations(
|
void HexagonSubtarget::getSMSMutations(
|
||||||
std::vector<std::unique_ptr<ScheduleDAGMutation>> &Mutations) const {
|
std::vector<std::unique_ptr<ScheduleDAGMutation>> &Mutations) const {
|
||||||
Mutations.push_back(
|
Mutations.push_back(llvm::make_unique<UsrOverflowMutation>());
|
||||||
llvm::make_unique<HexagonSubtarget::HexagonDAGMutation>());
|
Mutations.push_back(llvm::make_unique<HVXMemLatencyMutation>());
|
||||||
}
|
}
|
||||||
|
|
||||||
// Pin the vtable to this file.
|
// Pin the vtable to this file.
|
||||||
|
@ -56,10 +56,18 @@ public:
|
|||||||
/// default for V60.
|
/// default for V60.
|
||||||
bool UseBSBScheduling;
|
bool UseBSBScheduling;
|
||||||
|
|
||||||
class HexagonDAGMutation : public ScheduleDAGMutation {
|
struct UsrOverflowMutation : public ScheduleDAGMutation {
|
||||||
public:
|
|
||||||
void apply(ScheduleDAGInstrs *DAG) override;
|
void apply(ScheduleDAGInstrs *DAG) override;
|
||||||
};
|
};
|
||||||
|
struct HVXMemLatencyMutation : public ScheduleDAGMutation {
|
||||||
|
void apply(ScheduleDAGInstrs *DAG) override;
|
||||||
|
};
|
||||||
|
struct CallMutation : public ScheduleDAGMutation {
|
||||||
|
void apply(ScheduleDAGInstrs *DAG) override;
|
||||||
|
private:
|
||||||
|
bool shouldTFRICallBind(const HexagonInstrInfo &HII,
|
||||||
|
const SUnit &Inst1, const SUnit &Inst2) const;
|
||||||
|
};
|
||||||
|
|
||||||
private:
|
private:
|
||||||
std::string CPUString;
|
std::string CPUString;
|
||||||
|
@ -102,8 +102,9 @@ int HexagonTargetMachineModule = 0;
|
|||||||
static ScheduleDAGInstrs *createVLIWMachineSched(MachineSchedContext *C) {
|
static ScheduleDAGInstrs *createVLIWMachineSched(MachineSchedContext *C) {
|
||||||
ScheduleDAGMILive *DAG =
|
ScheduleDAGMILive *DAG =
|
||||||
new VLIWMachineScheduler(C, make_unique<ConvergingVLIWScheduler>());
|
new VLIWMachineScheduler(C, make_unique<ConvergingVLIWScheduler>());
|
||||||
DAG->addMutation(make_unique<HexagonSubtarget::HexagonDAGMutation>());
|
DAG->addMutation(make_unique<HexagonSubtarget::UsrOverflowMutation>());
|
||||||
DAG->addMutation(make_unique<HexagonCallMutation>());
|
DAG->addMutation(make_unique<HexagonSubtarget::HVXMemLatencyMutation>());
|
||||||
|
DAG->addMutation(make_unique<HexagonSubtarget::CallMutation>());
|
||||||
DAG->addMutation(createCopyConstrainDAGMutation(DAG->TII, DAG->TRI));
|
DAG->addMutation(createCopyConstrainDAGMutation(DAG->TII, DAG->TRI));
|
||||||
return DAG;
|
return DAG;
|
||||||
}
|
}
|
||||||
|
@ -103,7 +103,8 @@ HexagonPacketizerList::HexagonPacketizerList(MachineFunction &MF,
|
|||||||
HII = MF.getSubtarget<HexagonSubtarget>().getInstrInfo();
|
HII = MF.getSubtarget<HexagonSubtarget>().getInstrInfo();
|
||||||
HRI = MF.getSubtarget<HexagonSubtarget>().getRegisterInfo();
|
HRI = MF.getSubtarget<HexagonSubtarget>().getRegisterInfo();
|
||||||
|
|
||||||
addMutation(make_unique<HexagonSubtarget::HexagonDAGMutation>());
|
addMutation(make_unique<HexagonSubtarget::UsrOverflowMutation>());
|
||||||
|
addMutation(make_unique<HexagonSubtarget::HVXMemLatencyMutation>());
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check if FirstI modifies a register that SecondI reads.
|
// Check if FirstI modifies a register that SecondI reads.
|
||||||
|
Loading…
Reference in New Issue
Block a user