20#define DEBUG_TYPE "machine-scheduler"
31 void schedule()
override {}
47 if (
MI.isDebugInstr())
50 unsigned Opc =
MI.getOpcode();
53 if (
Opc == AMDGPU::ATOMIC_FENCE ||
Opc == AMDGPU::S_WAIT_ASYNCCNT ||
54 Opc == AMDGPU::S_WAIT_TENSORCNT ||
Opc == AMDGPU::S_BARRIER_WAIT ||
55 Opc == AMDGPU::S_BARRIER_SIGNAL_IMM)
83 for (
auto *PrioritySU : PrioritySUs) {
84 if (!PrioritySU->isTopReady())
91 unsigned MinDepth = std::numeric_limits<unsigned int>::max();
92 SUnit *TargetSU =
nullptr;
93 for (
auto *SU : AllSUs) {
100 if (SU->getDepth() < MinDepth) {
101 MinDepth = SU->getDepth();
110 bool Inserted = AllSUs.insert(SU);
116 TotalCycles += BlockingCycles;
118 if (PrioritySUs.empty()) {
119 PrioritySUs.insert(SU);
123 unsigned CurrDepth = (*PrioritySUs.begin())->getDepth();
124 if (SUDepth > CurrDepth)
127 if (SUDepth == CurrDepth) {
128 PrioritySUs.insert(SU);
134 PrioritySUs.insert(SU);
141 if (TotalCycles == 0)
145 PrioritySUs.remove(SU);
147 TotalCycles -= BlockingCycles;
151 if (PrioritySUs.empty()) {
152 for (
auto SU : AllSUs) {
153 if (PrioritySUs.empty()) {
154 PrioritySUs.insert(SU);
158 unsigned CurrDepth = (*PrioritySUs.begin())->getDepth();
159 if (SUDepth > CurrDepth)
162 if (SUDepth == CurrDepth) {
163 PrioritySUs.insert(SU);
169 PrioritySUs.insert(SU);
176 for (
auto &HWUICand :
HWUInfo) {
177 if (HWUICand.getType() == Flavor) {
186 unsigned ReleaseAtCycle = 0;
191 ReleaseAtCycle = std::max(ReleaseAtCycle, (
unsigned)PI->ReleaseAtCycle);
193 return ReleaseAtCycle;
215 for (
unsigned I = 0;
I <
HWUInfo.size();
I++) {
231 for (
auto &SU :
DAG->SUnits) {
242 <<
" (" <<
DAG->SUnits.size() <<
" SUs) ===\n";
244 dbgs() <<
"\nHWUI Resource Pressure:\n";
246 if (HWUI.getTotalCycles() == 0)
250 dbgs() <<
" " << Name <<
": " << HWUI.getTotalCycles() <<
" cycles, "
251 << HWUI.size() <<
" instrs\n";
260 if (
A.producesCoexecWindow() !=
B.producesCoexecWindow())
261 return A.producesCoexecWindow();
264 if (
A.getTotalCycles() !=
B.getTotalCycles())
265 return A.getTotalCycles() >
B.getTotalCycles();
268 if (
A.size() !=
B.size())
269 return A.size() <
B.size();
272 return (
unsigned)
A.getType() < (
unsigned)
B.getType();
280 auto HasPrioritySU = [
this, &Cand, &TryCand](
unsigned ResourceIdx) {
297 auto TryEnablesResource = [&Cand, &TryCand,
this](
unsigned ResourceIdx) {
307 TargetSU != Cand.
SU &&
DAG->IsReachable(TargetSU, Cand.
SU);
308 bool TryCandEnables =
309 TargetSU != TryCand.
SU &&
DAG->IsReachable(TargetSU, TryCand.
SU);
311 if (!CandEnables && !TryCandEnables)
314 if (CandEnables && !TryCandEnables) {
321 if (!CandEnables && TryCandEnables) {
330 if (CandHeight > TryCandHeight) {
337 if (CandHeight < TryCandHeight) {
349 for (
unsigned I = 0;
I <
HWUInfo.size();
I++) {
352 if (!HasPrioritySU(
I))
355 bool Enabled = TryEnablesResource(
I);
366 for (
unsigned I = 0;
I <
HWUInfo.size();
I++) {
370 bool TryCandUsesCrit = HWUI.
contains(TryCand.
SU);
372 if (!CandUsesCrit && !TryCandUsesCrit)
375 if (CandUsesCrit != TryCandUsesCrit) {
396 if (Match == Cand.
SU) {
424 "coexec scheduler only supports top-down scheduling");
440 Heurs.updateForScheduling(SU);
446 "coexec scheduler only supports top-down scheduling");
448 if (
DAG->top() ==
DAG->bottom()) {
450 Bot.Available.empty() &&
Bot.Pending.empty() &&
"ReadyQ garbage");
454 bool PickedPending =
false;
460 PickedPending =
false;
466 PickedPending,
false);
480 unsigned CurrentCycle =
Top.getCurrCycle();
481 if (ReadyCycle > CurrentCycle)
482 Top.bumpCycle(ReadyCycle);
485 while (
Top.checkHazard(SU))
486 Top.bumpCycle(
Top.getCurrCycle() + 1);
488 Top.releasePending();
499 assert(IsTopNode &&
"coexec scheduler must only schedule from top boundary");
506 bool &PickedPending,
bool IsBottomUp) {
507 assert(Zone.
isTop() &&
"coexec scheduler only supports top boundary");
508 assert(!IsBottomUp &&
"coexec scheduler only supports top-down scheduling");
512 unsigned SGPRPressure = 0;
513 unsigned VGPRPressure = 0;
514 PickedPending =
false;
515 if (
DAG->isTrackingPressure()) {
517 SGPRPressure =
Pressure[AMDGPU::RegisterPressureSets::SReg_32];
518 VGPRPressure =
Pressure[AMDGPU::RegisterPressureSets::VGPR_32];
525 auto EvaluateQueue = [&](
ReadyQueue &Q,
bool FromPending) {
526 for (
SUnit *SU : Q) {
529 VGPRPressure, IsBottomUp);
536 PickedPending = FromPending;
548 EvaluateQueue(Zone.
Pending,
true);
555 unsigned Cycle = IsTopNode ?
Top.getCurrCycle() :
Bot.getCurrCycle();
557 dbgs() <<
"=== Pick @ Cycle " <<
Cycle <<
" ===\n";
565 dbgs() <<
" Reason: ";
593 if (
DAG->isTrackingPressure() &&
603 bool SameBoundary = Zone !=
nullptr;
610 Heurs.sortHWUIResources();
611 if (
Heurs.tryCriticalResource(TryCand, Cand, Zone)) {
616 if (
Heurs.tryCriticalResourceDependency(TryCand, Cand, Zone)) {
630 bool CandIsClusterSucc =
632 bool TryCandIsClusterSucc =
635 if (
tryGreater(TryCandIsClusterSucc, CandIsClusterSucc, TryCand, Cand,
647 if (
DAG->isTrackingPressure() &&
656 !
Rem.IsAcyclicLatencyLimited &&
tryLatency(TryCand, Cand, *Zone))
677 unsigned Structural = 0;
679 unsigned Effective = 0;
683 auto GetStallCosts = [&](
SUnit *SU) {
684 unsigned ReadyCycle = Zone.
isTop() ? SU->TopReadyCycle : SU->BotReadyCycle;
686 Costs.Ready = ReadyCycle > CurrCycle ? ReadyCycle - CurrCycle : 0;
689 Costs.Effective = std::max({Costs.Ready, Costs.Structural, Costs.Latency});
693 StallCosts TryCosts = GetStallCosts(TryCand.
SU);
694 StallCosts CandCosts = GetStallCosts(Cand.
SU);
696 LLVM_DEBUG(
if (TryCosts.Effective || CandCosts.Effective) {
697 dbgs() <<
"Effective stalls: try=" << TryCosts.Effective
698 <<
" (ready=" << TryCosts.Ready <<
", struct=" << TryCosts.Structural
699 <<
", lat=" << TryCosts.Latency <<
") cand=" << CandCosts.Effective
700 <<
" (ready=" << CandCosts.Ready
701 <<
", struct=" << CandCosts.Structural
702 <<
", lat=" << CandCosts.Latency <<
")\n";
705 return tryLess(TryCosts.Effective, CandCosts.Effective, TryCand, Cand,
Stall);
710 LLVM_DEBUG(
dbgs() <<
"AMDGPU coexec preRA scheduler selected for "
711 <<
C->MF->getName() <<
'\n');
713 C, std::make_unique<AMDGPUCoExecSchedStrategy>(
C));
719 <<
C->MF->getName() <<
'\n');
720 return new GCNNoopPostScheduleDAG(
C);
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
static SUnit * pickOnlyChoice(SchedBoundary &Zone)
Coexecution-focused scheduling strategy for AMDGPU.
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
Register const TargetRegisterInfo * TRI
bool tryEffectiveStall(SchedCandidate &Cand, SchedCandidate &TryCand, SchedBoundary &Zone) const
AMDGPU::AMDGPUSchedReason LastAMDGPUReason
void initPolicy(MachineBasicBlock::iterator Begin, MachineBasicBlock::iterator End, unsigned NumRegionInstrs) override
Optionally override the per-region scheduling policy.
CandidateHeuristics Heurs
SUnit * pickNode(bool &IsTopNode) override
Pick the next node to schedule, or return NULL.
void pickNodeFromQueue(SchedBoundary &Zone, const CandPolicy &ZonePolicy, const RegPressureTracker &RPTracker, SchedCandidate &Cand, bool &PickedPending, bool IsBottomUp)
void initialize(ScheduleDAGMI *DAG) override
Initialize the strategy after building the DAG for a new region.
void schedNode(SUnit *SU, bool IsTopNode) override
Notify MachineSchedStrategy that ScheduleDAGMI has scheduled an instruction and updated scheduled/rem...
AMDGPUCoExecSchedStrategy(const MachineSchedContext *C)
void dumpPickSummary(SUnit *SU, bool IsTopNode, SchedCandidate &Cand)
bool tryCandidateCoexec(SchedCandidate &Cand, SchedCandidate &TryCand, SchedBoundary *Zone)
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
void updateForScheduling(SUnit *SU)
Update the state to reflect that SU is going to be scheduled.
HardwareUnitInfo * getHWUIFromFlavor(AMDGPU::InstructionFlavor Flavor)
Given a Flavor , find the corresponding HardwareUnit.
void sortHWUIResources()
Sort the HWUInfo vector.
bool tryCriticalResource(GenericSchedulerBase::SchedCandidate &TryCand, GenericSchedulerBase::SchedCandidate &Cand, SchedBoundary *Zone) const
Check for critical resource consumption.
bool tryCriticalResourceDependency(GenericSchedulerBase::SchedCandidate &TryCand, GenericSchedulerBase::SchedCandidate &Cand, SchedBoundary *Zone) const
Check for dependencies of instructions that use prioritized HardwareUnits.
SmallVector< HardwareUnitInfo, 8 > HWUInfo
const SIRegisterInfo * SRI
const TargetSchedModel * SchedModel
void collectHWUIPressure()
Walk over the region and collect total usage per HardwareUnit.
void initialize(ScheduleDAGMI *DAG, const TargetSchedModel *SchedModel, const TargetRegisterInfo *TRI)
unsigned getHWUICyclesForInst(SUnit *SU)
Compute the blocking cycles for the appropriate HardwareUnit given an SU.
GCNDownwardRPTracker DownwardTracker
bool useGCNTrackers() const
GCNSchedStrategy(const MachineSchedContext *C)
SmallVector< GCNSchedStageID, 4 > SchedStages
void schedNode(SUnit *SU, bool IsTopNode) override
Notify MachineSchedStrategy that ScheduleDAGMI has scheduled an instruction and updated scheduled/rem...
std::vector< unsigned > Pressure
void initialize(ScheduleDAGMI *DAG) override
Initialize the strategy after building the DAG for a new region.
void printCandidateDecision(const SchedCandidate &Current, const SchedCandidate &Preferred)
unsigned getStructuralStallCycles(SchedBoundary &Zone, SUnit *SU) const
Estimate how many cycles SU must wait due to structural hazards at the current boundary cycle.
void initCandidate(SchedCandidate &Cand, SUnit *SU, bool AtTop, const RegPressureTracker &RPTracker, const SIRegisterInfo *SRI, unsigned SGPRPressure, unsigned VGPRPressure, bool IsBottomUp)
MachineSchedPolicy RegionPolicy
const TargetSchedModel * SchedModel
static const char * getReasonStr(GenericSchedulerBase::CandReason Reason)
const TargetRegisterInfo * TRI
SchedCandidate TopCand
Candidate last picked from Top boundary.
HardwareUnitInfo is a wrapper class which maps to some real hardware resource.
void markScheduled(SUnit *SU, unsigned BlockingCycles)
Update the state for SU being scheduled by removing it from the AllSus and reducing its BlockingCycle...
bool contains(SUnit *SU) const
SUnit * getNextTargetSU(bool LookDeep=false) const
void insert(SUnit *SU, unsigned BlockingCycles)
Insert the SU into the AllSUs and account its BlockingCycles into the TotalCycles.
AMDGPU::InstructionFlavor getType() const
SUnit * getHigherPriority(SUnit *SU, SUnit *Other) const
int getNumber() const
MachineBasicBlocks are uniquely numbered at the function level, unless they're not in a MachineFuncti...
MachineInstrBundleIterator< MachineInstr > iterator
Representation of each machine instruction.
LLVM_ABI void print(raw_ostream &OS, bool IsStandalone=true, bool SkipOpers=false, bool SkipDebugLoc=false, bool AddNewLine=true, const TargetInstrInfo *TII=nullptr) const
Print this MI to OS.
virtual void initPolicy(MachineBasicBlock::iterator Begin, MachineBasicBlock::iterator End, unsigned NumRegionInstrs)
Optionally override the per-region scheduling policy.
Helpers for implementing custom MachineSchedStrategy classes.
Track the current register pressure at some position in the instruction stream, and remember the high...
const std::vector< unsigned > & getRegSetPressureAtPos() const
Get the register set pressure at the current position, which may be less than the pressure across the...
static bool isDS(const MachineInstr &MI)
static bool isFLATScratch(const MachineInstr &MI)
static bool isSALU(const MachineInstr &MI)
static bool isMFMAorWMMA(const MachineInstr &MI)
static bool isFLATGlobal(const MachineInstr &MI)
static bool isTRANS(const MachineInstr &MI)
static bool isFLAT(const MachineInstr &MI)
static bool isVALU(const MachineInstr &MI)
static bool isLDSDMA(const MachineInstr &MI)
Scheduling unit. This is a node in the scheduling DAG.
unsigned TopReadyCycle
Cycle relative to start when node is ready.
unsigned NodeNum
Entry # of node in the node vector.
unsigned getHeight() const
Returns the height of this node, which is the length of the maximum path down to any node which has n...
unsigned getDepth() const
Returns the depth of this node, which is the length of the maximum path up to any node which has no p...
bool isScheduled
True once scheduled.
unsigned ParentClusterIdx
The parent cluster id.
bool isBottomReady() const
MachineInstr * getInstr() const
Returns the representative MachineInstr for this SUnit.
Each Scheduling boundary is associated with ready queues.
LLVM_ABI unsigned getLatencyStallCycles(SUnit *SU)
Get the difference between the given SUnit's ready time and the current cycle.
LLVM_ABI SUnit * pickOnlyChoice()
Call this before applying any other heuristics to the Available queue.
unsigned getCurrCycle() const
Number of cycles to issue the instructions scheduled in this zone.
A ScheduleDAG for scheduling lists of MachineInstr.
ScheduleDAGMI is an implementation of ScheduleDAGInstrs that simply schedules machine instructions ac...
StringRef - Represent a constant reference to a string, i.e.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
Provide an instruction scheduling machine model to CodeGen passes.
const MCWriteProcResEntry * ProcResIter
InstructionFlavor classifyFlavor(const MachineInstr &MI, const SIInstrInfo &SII)
StringRef getReasonName(AMDGPUSchedReason R)
StringRef getFlavorName(InstructionFlavor F)
@ C
The default llvm calling convention, compatible with C.
This is an optimization pass for GlobalISel generic memory operations.
LLVM_ABI unsigned getWeakLeft(const SUnit *SU, bool isTop)
LLVM_ABI bool tryPressure(const PressureChange &TryP, const PressureChange &CandP, GenericSchedulerBase::SchedCandidate &TryCand, GenericSchedulerBase::SchedCandidate &Cand, GenericSchedulerBase::CandReason Reason, const TargetRegisterInfo *TRI, const MachineFunction &MF)
void sort(IteratorTy Start, IteratorTy End)
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
ScheduleDAGInstrs * createGCNNoopPostMachineScheduler(MachineSchedContext *C)
LLVM_ABI bool tryLatency(GenericSchedulerBase::SchedCandidate &TryCand, GenericSchedulerBase::SchedCandidate &Cand, SchedBoundary &Zone)
ScheduleDAGInstrs * createGCNCoExecMachineScheduler(MachineSchedContext *C)
bool isTheSameCluster(unsigned A, unsigned B)
Return whether the input cluster ID's are the same and valid.
LLVM_ABI bool tryGreater(int TryVal, int CandVal, GenericSchedulerBase::SchedCandidate &TryCand, GenericSchedulerBase::SchedCandidate &Cand, GenericSchedulerBase::CandReason Reason)
LLVM_ABI bool tryLess(int TryVal, int CandVal, GenericSchedulerBase::SchedCandidate &TryCand, GenericSchedulerBase::SchedCandidate &Cand, GenericSchedulerBase::CandReason Reason)
Return true if this heuristic determines order.
@ Enabled
Convert any .debug_str_offsets tables to DWARF64 if needed.
LLVM_ABI cl::opt< MISched::Direction > PreRADirection
LLVM_ABI int biasPhysReg(const SUnit *SU, bool isTop)
Minimize physical register live ranges.
Policy for scheduling the next instruction in the candidate's zone.
Store the state used by GenericScheduler heuristics, required for the lifetime of one invocation of p...
void setBest(SchedCandidate &Best)
LLVM_ABI void initResourceDelta(const ScheduleDAGMI *DAG, const TargetSchedModel *SchedModel)
SchedResourceDelta ResDelta
Status of an instruction's critical resource consumption.
Summarize the scheduling resources required for an instruction of a particular scheduling class.
MachineSchedContext provides enough context from the MachineScheduler pass for the target to instanti...
PressureChange CurrentMax