//===--------------------- Scheduler.cpp ------------------------*- C++ -*-===// // // The LLVM Compiler Infrastructure // // This file is distributed under the University of Illinois Open Source // License. See LICENSE.TXT for details. // //===----------------------------------------------------------------------===// // // A scheduler for processor resource units and processor resource groups. // //===----------------------------------------------------------------------===// #include "Scheduler.h" #include "Backend.h" #include "HWEventListener.h" #include "Support.h" #include "llvm/Support/Debug.h" #include "llvm/Support/raw_ostream.h" #define DEBUG_TYPE "llvm-mca" namespace mca { using namespace llvm; uint64_t ResourceState::selectNextInSequence() { assert(isReady()); uint64_t Next = getNextInSequence(); while (!isSubResourceReady(Next)) { updateNextInSequence(); Next = getNextInSequence(); } return Next; } #ifndef NDEBUG void ResourceState::dump() const { dbgs() << "MASK: " << ResourceMask << ", SIZE_MASK: " << ResourceSizeMask << ", NEXT: " << NextInSequenceMask << ", RDYMASK: " << ReadyMask << ", BufferSize=" << BufferSize << ", AvailableSlots=" << AvailableSlots << ", Reserved=" << Unavailable << '\n'; } #endif void ResourceManager::initialize(const llvm::MCSchedModel &SM) { computeProcResourceMasks(SM, ProcResID2Mask); for (unsigned I = 0, E = SM.getNumProcResourceKinds(); I < E; ++I) addResource(*SM.getProcResource(I), I, ProcResID2Mask[I]); } // Adds a new resource state in Resources, as well as a new descriptor in // ResourceDescriptor. Map 'Resources' allows to quickly obtain ResourceState // objects from resource mask identifiers. void ResourceManager::addResource(const MCProcResourceDesc &Desc, unsigned Index, uint64_t Mask) { assert(Resources.find(Mask) == Resources.end() && "Resource already added!"); Resources[Mask] = llvm::make_unique(Desc, Index, Mask); } // Returns the actual resource consumed by this Use. // First, is the primary resource ID. // Second, is the specific sub-resource ID. std::pair ResourceManager::selectPipe(uint64_t ResourceID) { ResourceState &RS = *Resources[ResourceID]; uint64_t SubResourceID = RS.selectNextInSequence(); if (RS.isAResourceGroup()) return selectPipe(SubResourceID); return std::pair(ResourceID, SubResourceID); } void ResourceState::removeFromNextInSequence(uint64_t ID) { assert(NextInSequenceMask); assert(countPopulation(ID) == 1); if (ID > getNextInSequence()) RemovedFromNextInSequence |= ID; NextInSequenceMask = NextInSequenceMask & (~ID); if (!NextInSequenceMask) { NextInSequenceMask = ResourceSizeMask; assert(NextInSequenceMask != RemovedFromNextInSequence); NextInSequenceMask ^= RemovedFromNextInSequence; RemovedFromNextInSequence = 0; } } void ResourceManager::use(ResourceRef RR) { // Mark the sub-resource referenced by RR as used. ResourceState &RS = *Resources[RR.first]; RS.markSubResourceAsUsed(RR.second); // If there are still available units in RR.first, // then we are done. if (RS.isReady()) return; // Notify to other resources that RR.first is no longer available. for (const std::pair &Res : Resources) { ResourceState &Current = *Res.second.get(); if (!Current.isAResourceGroup() || Current.getResourceMask() == RR.first) continue; if (Current.containsResource(RR.first)) { Current.markSubResourceAsUsed(RR.first); Current.removeFromNextInSequence(RR.first); } } } void ResourceManager::release(ResourceRef RR) { ResourceState &RS = *Resources[RR.first]; bool WasFullyUsed = !RS.isReady(); RS.releaseSubResource(RR.second); if (!WasFullyUsed) return; for (const std::pair &Res : Resources) { ResourceState &Current = *Res.second.get(); if (!Current.isAResourceGroup() || Current.getResourceMask() == RR.first) continue; if (Current.containsResource(RR.first)) Current.releaseSubResource(RR.first); } } ResourceStateEvent ResourceManager::canBeDispatched(ArrayRef Buffers) const { ResourceStateEvent Result = ResourceStateEvent::RS_BUFFER_AVAILABLE; for (uint64_t Buffer : Buffers) { Result = isBufferAvailable(Buffer); if (Result != ResourceStateEvent::RS_BUFFER_AVAILABLE) break; } return Result; } void ResourceManager::reserveBuffers(ArrayRef Buffers) { for (const uint64_t R : Buffers) { reserveBuffer(R); ResourceState &Resource = *Resources[R]; if (Resource.isADispatchHazard()) { assert(!Resource.isReserved()); Resource.setReserved(); } } } void ResourceManager::releaseBuffers(ArrayRef Buffers) { for (const uint64_t R : Buffers) releaseBuffer(R); } bool ResourceManager::canBeIssued(const InstrDesc &Desc) const { return std::all_of(Desc.Resources.begin(), Desc.Resources.end(), [&](const std::pair &E) { unsigned NumUnits = E.second.isReserved() ? 0U : E.second.NumUnits; return isReady(E.first, NumUnits); }); } // Returns true if all resources are in-order, and there is at least one // resource which is a dispatch hazard (BufferSize = 0). bool ResourceManager::mustIssueImmediately(const InstrDesc &Desc) { if (!canBeIssued(Desc)) return false; bool AllInOrderResources = std::all_of( Desc.Buffers.begin(), Desc.Buffers.end(), [&](const unsigned BufferMask) { const ResourceState &Resource = *Resources[BufferMask]; return Resource.isInOrder() || Resource.isADispatchHazard(); }); if (!AllInOrderResources) return false; return std::any_of(Desc.Buffers.begin(), Desc.Buffers.end(), [&](const unsigned BufferMask) { return Resources[BufferMask]->isADispatchHazard(); }); } void ResourceManager::issueInstruction( const InstrDesc &Desc, SmallVectorImpl> &Pipes) { for (const std::pair &R : Desc.Resources) { const CycleSegment &CS = R.second.CS; if (!CS.size()) { releaseResource(R.first); continue; } assert(CS.begin() == 0 && "Invalid {Start, End} cycles!"); if (!R.second.isReserved()) { ResourceRef Pipe = selectPipe(R.first); use(Pipe); BusyResources[Pipe] += CS.size(); // Replace the resource mask with a valid processor resource index. const ResourceState &RS = *Resources[Pipe.first]; Pipe.first = RS.getProcResourceID(); Pipes.emplace_back( std::pair(Pipe, static_cast(CS.size()))); } else { assert((countPopulation(R.first) > 1) && "Expected a group!"); // Mark this group as reserved. assert(R.second.isReserved()); reserveResource(R.first); BusyResources[ResourceRef(R.first, R.first)] += CS.size(); } } } void ResourceManager::cycleEvent(SmallVectorImpl &ResourcesFreed) { for (std::pair &BR : BusyResources) { if (BR.second) BR.second--; if (!BR.second) { // Release this resource. const ResourceRef &RR = BR.first; if (countPopulation(RR.first) == 1) release(RR); releaseResource(RR.first); ResourcesFreed.push_back(RR); } } for (const ResourceRef &RF : ResourcesFreed) BusyResources.erase(RF); } void Scheduler::scheduleInstruction(InstRef &IR) { const unsigned Idx = IR.getSourceIndex(); assert(WaitQueue.find(Idx) == WaitQueue.end()); assert(ReadyQueue.find(Idx) == ReadyQueue.end()); assert(IssuedQueue.find(Idx) == IssuedQueue.end()); // Reserve a slot in each buffered resource. Also, mark units with // BufferSize=0 as reserved. Resources with a buffer size of zero will only // be released after MCIS is issued, and all the ResourceCycles for those // units have been consumed. const InstrDesc &Desc = IR.getInstruction()->getDesc(); reserveBuffers(Desc.Buffers); notifyReservedBuffers(Desc.Buffers); // If necessary, reserve queue entries in the load-store unit (LSU). bool Reserved = LSU->reserve(IR); if (!IR.getInstruction()->isReady() || (Reserved && !LSU->isReady(IR))) { DEBUG(dbgs() << "[SCHEDULER] Adding " << Idx << " to the Wait Queue\n"); WaitQueue[Idx] = IR.getInstruction(); return; } notifyInstructionReady(IR); // Don't add a zero-latency instruction to the Wait or Ready queue. // A zero-latency instruction doesn't consume any scheduler resources. That is // because it doesn't need to be executed, and it is often removed at register // renaming stage. For example, register-register moves are often optimized at // register renaming stage by simply updating register aliases. On some // targets, zero-idiom instructions (for example: a xor that clears the value // of a register) are treated speacially, and are often eliminated at register // renaming stage. bool IsZeroLatency = !Desc.MaxLatency && Desc.Resources.empty(); // Instructions that use an in-order dispatch/issue processor resource must be // issued immediately to the pipeline(s). Any other in-order buffered // resources (i.e. BufferSize=1) is consumed. if (!IsZeroLatency && !Resources->mustIssueImmediately(Desc)) { DEBUG(dbgs() << "[SCHEDULER] Adding " << IR << " to the Ready Queue\n"); ReadyQueue[IR.getSourceIndex()] = IR.getInstruction(); return; } DEBUG(dbgs() << "[SCHEDULER] Instruction " << IR << " issued immediately\n"); // Release buffered resources and issue MCIS to the underlying pipelines. issueInstruction(IR); } void Scheduler::cycleEvent() { SmallVector ResourcesFreed; Resources->cycleEvent(ResourcesFreed); for (const ResourceRef &RR : ResourcesFreed) notifyResourceAvailable(RR); SmallVector InstructionIDs; updateIssuedQueue(InstructionIDs); for (const InstRef &IR : InstructionIDs) notifyInstructionExecuted(IR); InstructionIDs.clear(); updatePendingQueue(InstructionIDs); for (const InstRef &IR : InstructionIDs) notifyInstructionReady(IR); InstructionIDs.clear(); InstRef IR = select(); while (IR.isValid()) { issueInstruction(IR); // Instructions that have been issued during this cycle might have unblocked // other dependent instructions. Dependent instructions may be issued during // this same cycle if operands have ReadAdvance entries. Promote those // instructions to the ReadyQueue and tell to the caller that we need // another round of 'issue()'. promoteToReadyQueue(InstructionIDs); for (const InstRef &I : InstructionIDs) notifyInstructionReady(I); InstructionIDs.clear(); // Select the next instruction to issue. IR = select(); } } #ifndef NDEBUG void Scheduler::dump() const { dbgs() << "[SCHEDULER]: WaitQueue size is: " << WaitQueue.size() << '\n'; dbgs() << "[SCHEDULER]: ReadyQueue size is: " << ReadyQueue.size() << '\n'; dbgs() << "[SCHEDULER]: IssuedQueue size is: " << IssuedQueue.size() << '\n'; Resources->dump(); } #endif bool Scheduler::canBeDispatched(const InstRef &IR) const { HWStallEvent::GenericEventType Type = HWStallEvent::Invalid; const InstrDesc &Desc = IR.getInstruction()->getDesc(); if (Desc.MayLoad && LSU->isLQFull()) Type = HWStallEvent::LoadQueueFull; else if (Desc.MayStore && LSU->isSQFull()) Type = HWStallEvent::StoreQueueFull; else { switch (Resources->canBeDispatched(Desc.Buffers)) { default: return true; case ResourceStateEvent::RS_BUFFER_UNAVAILABLE: Type = HWStallEvent::SchedulerQueueFull; break; case ResourceStateEvent::RS_RESERVED: Type = HWStallEvent::DispatchGroupStall; } } Owner->notifyStallEvent(HWStallEvent(Type, IR)); return false; } void Scheduler::issueInstructionImpl( InstRef &IR, SmallVectorImpl> &UsedResources) { Instruction *IS = IR.getInstruction(); const InstrDesc &D = IS->getDesc(); // Issue the instruction and collect all the consumed resources // into a vector. That vector is then used to notify the listener. Resources->issueInstruction(D, UsedResources); // Notify the instruction that it started executing. // This updates the internal state of each write. IS->execute(); if (IS->isExecuting()) IssuedQueue[IR.getSourceIndex()] = IS; } void Scheduler::issueInstruction(InstRef &IR) { // Release buffered resources. const InstrDesc &Desc = IR.getInstruction()->getDesc(); releaseBuffers(Desc.Buffers); notifyReleasedBuffers(Desc.Buffers); // Issue IS to the underlying pipelines and notify listeners. SmallVector, 4> Pipes; issueInstructionImpl(IR, Pipes); notifyInstructionIssued(IR, Pipes); if (IR.getInstruction()->isExecuted()) notifyInstructionExecuted(IR); } void Scheduler::promoteToReadyQueue(SmallVectorImpl &Ready) { // Scan the set of waiting instructions and promote them to the // ready queue if operands are all ready. for (auto I = WaitQueue.begin(), E = WaitQueue.end(); I != E;) { const unsigned IID = I->first; Instruction *IS = I->second; // Check if this instruction is now ready. In case, force // a transition in state using method 'update()'. IS->update(); const InstrDesc &Desc = IS->getDesc(); bool IsMemOp = Desc.MayLoad || Desc.MayStore; if (!IS->isReady() || (IsMemOp && !LSU->isReady({IID, IS}))) { ++I; continue; } Ready.emplace_back(IID, IS); ReadyQueue[IID] = IS; auto ToRemove = I; ++I; WaitQueue.erase(ToRemove); } } InstRef Scheduler::select() { // Give priority to older instructions in the ReadyQueue. Since the ready // queue is ordered by key, this will always prioritize older instructions. const auto It = std::find_if(ReadyQueue.begin(), ReadyQueue.end(), [&](const QueueEntryTy &Entry) { const InstrDesc &D = Entry.second->getDesc(); return Resources->canBeIssued(D); }); if (It == ReadyQueue.end()) return {0, nullptr}; // We found an instruction to issue. InstRef IR(It->first, It->second); ReadyQueue.erase(It); return IR; } void Scheduler::updatePendingQueue(SmallVectorImpl &Ready) { // Notify to instructions in the pending queue that a new cycle just // started. for (QueueEntryTy Entry : WaitQueue) Entry.second->cycleEvent(); promoteToReadyQueue(Ready); } void Scheduler::updateIssuedQueue(SmallVectorImpl &Executed) { for (auto I = IssuedQueue.begin(), E = IssuedQueue.end(); I != E;) { const QueueEntryTy Entry = *I; Instruction *IS = Entry.second; IS->cycleEvent(); if (IS->isExecuted()) { Executed.push_back({Entry.first, Entry.second}); auto ToRemove = I; ++I; IssuedQueue.erase(ToRemove); } else { DEBUG(dbgs() << "[SCHEDULER]: Instruction " << Entry.first << " is still executing.\n"); ++I; } } } void Scheduler::notifyInstructionIssued( const InstRef &IR, ArrayRef> Used) { DEBUG({ dbgs() << "[E] Instruction Issued: " << IR << '\n'; for (const std::pair &Resource : Used) { dbgs() << "[E] Resource Used: [" << Resource.first.first << '.' << Resource.first.second << "]\n"; dbgs() << " cycles: " << Resource.second << '\n'; } }); Owner->notifyInstructionEvent(HWInstructionIssuedEvent(IR, Used)); } void Scheduler::notifyInstructionExecuted(const InstRef &IR) { LSU->onInstructionExecuted(IR); DEBUG(dbgs() << "[E] Instruction Executed: " << IR << '\n'); Owner->notifyInstructionEvent( HWInstructionEvent(HWInstructionEvent::Executed, IR)); DU->onInstructionExecuted(IR.getInstruction()->getRCUTokenID()); } void Scheduler::notifyInstructionReady(const InstRef &IR) { DEBUG(dbgs() << "[E] Instruction Ready: " << IR << '\n'); Owner->notifyInstructionEvent( HWInstructionEvent(HWInstructionEvent::Ready, IR)); } void Scheduler::notifyResourceAvailable(const ResourceRef &RR) { Owner->notifyResourceAvailable(RR); } void Scheduler::notifyReservedBuffers(ArrayRef Buffers) { if (Buffers.empty()) return; SmallVector BufferIDs(Buffers.begin(), Buffers.end()); std::transform( Buffers.begin(), Buffers.end(), BufferIDs.begin(), [&](uint64_t Op) { return Resources->resolveResourceMask(Op); }); Owner->notifyReservedBuffers(BufferIDs); } void Scheduler::notifyReleasedBuffers(ArrayRef Buffers) { if (Buffers.empty()) return; SmallVector BufferIDs(Buffers.begin(), Buffers.end()); std::transform( Buffers.begin(), Buffers.end(), BufferIDs.begin(), [&](uint64_t Op) { return Resources->resolveResourceMask(Op); }); Owner->notifyReleasedBuffers(BufferIDs); } } // namespace mca