mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-23 03:02:36 +01:00
a315353f43
This patch changes return type of tryCandidate from void to bool: 1. Methods in some targets already follow this convention. 2. This would help if some target wants to re-use generic code. 3. It looks more intuitive if these try-method returns the same type. We may need to change return type of them from bool to some enum further, to make it less confusing. Reviewed By: foad Differential Revision: https://reviews.llvm.org/D103951
252 lines
9.5 KiB
C++
252 lines
9.5 KiB
C++
//===- PPCMachineScheduler.cpp - MI Scheduler for PowerPC -------------===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "PPCMachineScheduler.h"
|
|
#include "MCTargetDesc/PPCMCTargetDesc.h"
|
|
|
|
using namespace llvm;
|
|
|
|
static cl::opt<bool>
|
|
DisableAddiLoadHeuristic("disable-ppc-sched-addi-load",
|
|
cl::desc("Disable scheduling addi instruction before"
|
|
"load for ppc"), cl::Hidden);
|
|
static cl::opt<bool>
|
|
EnableAddiHeuristic("ppc-postra-bias-addi",
|
|
cl::desc("Enable scheduling addi instruction as early"
|
|
"as possible post ra"),
|
|
cl::Hidden, cl::init(true));
|
|
|
|
static bool isADDIInstr(const GenericScheduler::SchedCandidate &Cand) {
|
|
return Cand.SU->getInstr()->getOpcode() == PPC::ADDI ||
|
|
Cand.SU->getInstr()->getOpcode() == PPC::ADDI8;
|
|
}
|
|
|
|
bool PPCPreRASchedStrategy::biasAddiLoadCandidate(SchedCandidate &Cand,
|
|
SchedCandidate &TryCand,
|
|
SchedBoundary &Zone) const {
|
|
if (DisableAddiLoadHeuristic)
|
|
return false;
|
|
|
|
SchedCandidate &FirstCand = Zone.isTop() ? TryCand : Cand;
|
|
SchedCandidate &SecondCand = Zone.isTop() ? Cand : TryCand;
|
|
if (isADDIInstr(FirstCand) && SecondCand.SU->getInstr()->mayLoad()) {
|
|
TryCand.Reason = Stall;
|
|
return true;
|
|
}
|
|
if (FirstCand.SU->getInstr()->mayLoad() && isADDIInstr(SecondCand)) {
|
|
TryCand.Reason = NoCand;
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool PPCPreRASchedStrategy::tryCandidate(SchedCandidate &Cand,
|
|
SchedCandidate &TryCand,
|
|
SchedBoundary *Zone) const {
|
|
// From GenericScheduler::tryCandidate
|
|
|
|
// Initialize the candidate if needed.
|
|
if (!Cand.isValid()) {
|
|
TryCand.Reason = NodeOrder;
|
|
return true;
|
|
}
|
|
|
|
// Bias PhysReg Defs and copies to their uses and defined respectively.
|
|
if (tryGreater(biasPhysReg(TryCand.SU, TryCand.AtTop),
|
|
biasPhysReg(Cand.SU, Cand.AtTop), TryCand, Cand, PhysReg))
|
|
return TryCand.Reason != NoCand;
|
|
|
|
// Avoid exceeding the target's limit.
|
|
if (DAG->isTrackingPressure() &&
|
|
tryPressure(TryCand.RPDelta.Excess, Cand.RPDelta.Excess, TryCand, Cand,
|
|
RegExcess, TRI, DAG->MF))
|
|
return TryCand.Reason != NoCand;
|
|
|
|
// Avoid increasing the max critical pressure in the scheduled region.
|
|
if (DAG->isTrackingPressure() &&
|
|
tryPressure(TryCand.RPDelta.CriticalMax, Cand.RPDelta.CriticalMax,
|
|
TryCand, Cand, RegCritical, TRI, DAG->MF))
|
|
return TryCand.Reason != NoCand;
|
|
|
|
// We only compare a subset of features when comparing nodes between
|
|
// Top and Bottom boundary. Some properties are simply incomparable, in many
|
|
// other instances we should only override the other boundary if something
|
|
// is a clear good pick on one boundary. Skip heuristics that are more
|
|
// "tie-breaking" in nature.
|
|
bool SameBoundary = Zone != nullptr;
|
|
if (SameBoundary) {
|
|
// For loops that are acyclic path limited, aggressively schedule for
|
|
// latency. Within an single cycle, whenever CurrMOps > 0, allow normal
|
|
// heuristics to take precedence.
|
|
if (Rem.IsAcyclicLatencyLimited && !Zone->getCurrMOps() &&
|
|
tryLatency(TryCand, Cand, *Zone))
|
|
return TryCand.Reason != NoCand;
|
|
|
|
// Prioritize instructions that read unbuffered resources by stall cycles.
|
|
if (tryLess(Zone->getLatencyStallCycles(TryCand.SU),
|
|
Zone->getLatencyStallCycles(Cand.SU), TryCand, Cand, Stall))
|
|
return TryCand.Reason != NoCand;
|
|
}
|
|
|
|
// Keep clustered nodes together to encourage downstream peephole
|
|
// optimizations which may reduce resource requirements.
|
|
//
|
|
// This is a best effort to set things up for a post-RA pass. Optimizations
|
|
// like generating loads of multiple registers should ideally be done within
|
|
// the scheduler pass by combining the loads during DAG postprocessing.
|
|
const SUnit *CandNextClusterSU =
|
|
Cand.AtTop ? DAG->getNextClusterSucc() : DAG->getNextClusterPred();
|
|
const SUnit *TryCandNextClusterSU =
|
|
TryCand.AtTop ? DAG->getNextClusterSucc() : DAG->getNextClusterPred();
|
|
if (tryGreater(TryCand.SU == TryCandNextClusterSU,
|
|
Cand.SU == CandNextClusterSU, TryCand, Cand, Cluster))
|
|
return TryCand.Reason != NoCand;
|
|
|
|
if (SameBoundary) {
|
|
// Weak edges are for clustering and other constraints.
|
|
if (tryLess(getWeakLeft(TryCand.SU, TryCand.AtTop),
|
|
getWeakLeft(Cand.SU, Cand.AtTop), TryCand, Cand, Weak))
|
|
return TryCand.Reason != NoCand;
|
|
}
|
|
|
|
// Avoid increasing the max pressure of the entire region.
|
|
if (DAG->isTrackingPressure() &&
|
|
tryPressure(TryCand.RPDelta.CurrentMax, Cand.RPDelta.CurrentMax, TryCand,
|
|
Cand, RegMax, TRI, DAG->MF))
|
|
return TryCand.Reason != NoCand;
|
|
|
|
if (SameBoundary) {
|
|
// Avoid critical resource consumption and balance the schedule.
|
|
TryCand.initResourceDelta(DAG, SchedModel);
|
|
if (tryLess(TryCand.ResDelta.CritResources, Cand.ResDelta.CritResources,
|
|
TryCand, Cand, ResourceReduce))
|
|
return TryCand.Reason != NoCand;
|
|
if (tryGreater(TryCand.ResDelta.DemandedResources,
|
|
Cand.ResDelta.DemandedResources, TryCand, Cand,
|
|
ResourceDemand))
|
|
return TryCand.Reason != NoCand;
|
|
|
|
// Avoid serializing long latency dependence chains.
|
|
// For acyclic path limited loops, latency was already checked above.
|
|
if (!RegionPolicy.DisableLatencyHeuristic && TryCand.Policy.ReduceLatency &&
|
|
!Rem.IsAcyclicLatencyLimited && tryLatency(TryCand, Cand, *Zone))
|
|
return TryCand.Reason != NoCand;
|
|
|
|
// Fall through to original instruction order.
|
|
if ((Zone->isTop() && TryCand.SU->NodeNum < Cand.SU->NodeNum) ||
|
|
(!Zone->isTop() && TryCand.SU->NodeNum > Cand.SU->NodeNum)) {
|
|
TryCand.Reason = NodeOrder;
|
|
}
|
|
}
|
|
|
|
// GenericScheduler::tryCandidate end
|
|
|
|
// Add powerpc specific heuristic only when TryCand isn't selected or
|
|
// selected as node order.
|
|
if (TryCand.Reason != NodeOrder && TryCand.Reason != NoCand)
|
|
return true;
|
|
|
|
// There are some benefits to schedule the ADDI before the load to hide the
|
|
// latency, as RA may create a true dependency between the load and addi.
|
|
if (SameBoundary) {
|
|
if (biasAddiLoadCandidate(Cand, TryCand, *Zone))
|
|
return TryCand.Reason != NoCand;
|
|
}
|
|
|
|
return TryCand.Reason != NoCand;
|
|
}
|
|
|
|
bool PPCPostRASchedStrategy::biasAddiCandidate(SchedCandidate &Cand,
|
|
SchedCandidate &TryCand) const {
|
|
if (!EnableAddiHeuristic)
|
|
return false;
|
|
|
|
if (isADDIInstr(TryCand) && !isADDIInstr(Cand)) {
|
|
TryCand.Reason = Stall;
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
bool PPCPostRASchedStrategy::tryCandidate(SchedCandidate &Cand,
|
|
SchedCandidate &TryCand) {
|
|
// From PostGenericScheduler::tryCandidate
|
|
|
|
// Initialize the candidate if needed.
|
|
if (!Cand.isValid()) {
|
|
TryCand.Reason = NodeOrder;
|
|
return true;
|
|
}
|
|
|
|
// Prioritize instructions that read unbuffered resources by stall cycles.
|
|
if (tryLess(Top.getLatencyStallCycles(TryCand.SU),
|
|
Top.getLatencyStallCycles(Cand.SU), TryCand, Cand, Stall))
|
|
return TryCand.Reason != NoCand;
|
|
|
|
// Keep clustered nodes together.
|
|
if (tryGreater(TryCand.SU == DAG->getNextClusterSucc(),
|
|
Cand.SU == DAG->getNextClusterSucc(), TryCand, Cand, Cluster))
|
|
return TryCand.Reason != NoCand;
|
|
|
|
// Avoid critical resource consumption and balance the schedule.
|
|
if (tryLess(TryCand.ResDelta.CritResources, Cand.ResDelta.CritResources,
|
|
TryCand, Cand, ResourceReduce))
|
|
return TryCand.Reason != NoCand;
|
|
if (tryGreater(TryCand.ResDelta.DemandedResources,
|
|
Cand.ResDelta.DemandedResources, TryCand, Cand,
|
|
ResourceDemand))
|
|
return TryCand.Reason != NoCand;
|
|
|
|
// Avoid serializing long latency dependence chains.
|
|
if (Cand.Policy.ReduceLatency && tryLatency(TryCand, Cand, Top)) {
|
|
return TryCand.Reason != NoCand;
|
|
}
|
|
|
|
// Fall through to original instruction order.
|
|
if (TryCand.SU->NodeNum < Cand.SU->NodeNum)
|
|
TryCand.Reason = NodeOrder;
|
|
|
|
// PostGenericScheduler::tryCandidate end
|
|
|
|
// Add powerpc post ra specific heuristic only when TryCand isn't selected or
|
|
// selected as node order.
|
|
if (TryCand.Reason != NodeOrder && TryCand.Reason != NoCand)
|
|
return true;
|
|
|
|
// There are some benefits to schedule the ADDI as early as possible post ra
|
|
// to avoid stalled by vector instructions which take up all the hw units.
|
|
// And ADDI is usually used to post inc the loop indvar, which matters the
|
|
// performance.
|
|
if (biasAddiCandidate(Cand, TryCand))
|
|
return TryCand.Reason != NoCand;
|
|
|
|
return TryCand.Reason != NoCand;
|
|
}
|
|
|
|
void PPCPostRASchedStrategy::enterMBB(MachineBasicBlock *MBB) {
|
|
// Custom PPC PostRA specific behavior here.
|
|
PostGenericScheduler::enterMBB(MBB);
|
|
}
|
|
|
|
void PPCPostRASchedStrategy::leaveMBB() {
|
|
// Custom PPC PostRA specific behavior here.
|
|
PostGenericScheduler::leaveMBB();
|
|
}
|
|
|
|
void PPCPostRASchedStrategy::initialize(ScheduleDAGMI *Dag) {
|
|
// Custom PPC PostRA specific initialization here.
|
|
PostGenericScheduler::initialize(Dag);
|
|
}
|
|
|
|
SUnit *PPCPostRASchedStrategy::pickNode(bool &IsTopNode) {
|
|
// Custom PPC PostRA specific scheduling here.
|
|
return PostGenericScheduler::pickNode(IsTopNode);
|
|
}
|
|
|