mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-22 18:54:02 +01:00
1f0605d432
The values of registers in inactive lanes needs to be saved during function calls. Save all registers used for whole wave mode, similar to how it is done for VGPRs that are used for SGPR spilling. Differential Revision: https://reviews.llvm.org/D99429 Reapply with fixed tests on window.
259 lines
7.2 KiB
C++
259 lines
7.2 KiB
C++
//===- SIPreAllocateWWMRegs.cpp - WWM Register Pre-allocation -------------===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
/// \file
|
|
/// Pass to pre-allocated WWM registers
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "AMDGPU.h"
|
|
#include "GCNSubtarget.h"
|
|
#include "MCTargetDesc/AMDGPUMCTargetDesc.h"
|
|
#include "SIMachineFunctionInfo.h"
|
|
#include "llvm/ADT/PostOrderIterator.h"
|
|
#include "llvm/CodeGen/LiveIntervals.h"
|
|
#include "llvm/CodeGen/LiveRegMatrix.h"
|
|
#include "llvm/CodeGen/MachineFunctionPass.h"
|
|
#include "llvm/InitializePasses.h"
|
|
|
|
using namespace llvm;
|
|
|
|
#define DEBUG_TYPE "si-pre-allocate-wwm-regs"
|
|
|
|
namespace {
|
|
|
|
class SIPreAllocateWWMRegs : public MachineFunctionPass {
|
|
private:
|
|
const SIInstrInfo *TII;
|
|
const SIRegisterInfo *TRI;
|
|
MachineRegisterInfo *MRI;
|
|
LiveIntervals *LIS;
|
|
LiveRegMatrix *Matrix;
|
|
VirtRegMap *VRM;
|
|
RegisterClassInfo RegClassInfo;
|
|
|
|
std::vector<unsigned> RegsToRewrite;
|
|
#ifndef NDEBUG
|
|
void printWWMInfo(const MachineInstr &MI);
|
|
#endif
|
|
|
|
public:
|
|
static char ID;
|
|
|
|
SIPreAllocateWWMRegs() : MachineFunctionPass(ID) {
|
|
initializeSIPreAllocateWWMRegsPass(*PassRegistry::getPassRegistry());
|
|
}
|
|
|
|
bool runOnMachineFunction(MachineFunction &MF) override;
|
|
|
|
void getAnalysisUsage(AnalysisUsage &AU) const override {
|
|
AU.addRequired<LiveIntervals>();
|
|
AU.addPreserved<LiveIntervals>();
|
|
AU.addRequired<VirtRegMap>();
|
|
AU.addRequired<LiveRegMatrix>();
|
|
AU.addPreserved<SlotIndexes>();
|
|
AU.setPreservesCFG();
|
|
MachineFunctionPass::getAnalysisUsage(AU);
|
|
}
|
|
|
|
private:
|
|
bool processDef(MachineOperand &MO);
|
|
void rewriteRegs(MachineFunction &MF);
|
|
};
|
|
|
|
} // End anonymous namespace.
|
|
|
|
INITIALIZE_PASS_BEGIN(SIPreAllocateWWMRegs, DEBUG_TYPE,
|
|
"SI Pre-allocate WWM Registers", false, false)
|
|
INITIALIZE_PASS_DEPENDENCY(LiveIntervals)
|
|
INITIALIZE_PASS_DEPENDENCY(VirtRegMap)
|
|
INITIALIZE_PASS_DEPENDENCY(LiveRegMatrix)
|
|
INITIALIZE_PASS_END(SIPreAllocateWWMRegs, DEBUG_TYPE,
|
|
"SI Pre-allocate WWM Registers", false, false)
|
|
|
|
char SIPreAllocateWWMRegs::ID = 0;
|
|
|
|
char &llvm::SIPreAllocateWWMRegsID = SIPreAllocateWWMRegs::ID;
|
|
|
|
FunctionPass *llvm::createSIPreAllocateWWMRegsPass() {
|
|
return new SIPreAllocateWWMRegs();
|
|
}
|
|
|
|
bool SIPreAllocateWWMRegs::processDef(MachineOperand &MO) {
|
|
if (!MO.isReg())
|
|
return false;
|
|
|
|
Register Reg = MO.getReg();
|
|
if (Reg.isPhysical())
|
|
return false;
|
|
|
|
if (!TRI->isVGPR(*MRI, Reg))
|
|
return false;
|
|
|
|
if (VRM->hasPhys(Reg))
|
|
return false;
|
|
|
|
LiveInterval &LI = LIS->getInterval(Reg);
|
|
|
|
for (MCRegister PhysReg : RegClassInfo.getOrder(MRI->getRegClass(Reg))) {
|
|
if (!MRI->isPhysRegUsed(PhysReg) &&
|
|
Matrix->checkInterference(LI, PhysReg) == LiveRegMatrix::IK_Free) {
|
|
Matrix->assign(LI, PhysReg);
|
|
assert(PhysReg != 0);
|
|
RegsToRewrite.push_back(Reg);
|
|
return true;
|
|
}
|
|
}
|
|
|
|
llvm_unreachable("physreg not found for WWM expression");
|
|
return false;
|
|
}
|
|
|
|
void SIPreAllocateWWMRegs::rewriteRegs(MachineFunction &MF) {
|
|
for (MachineBasicBlock &MBB : MF) {
|
|
for (MachineInstr &MI : MBB) {
|
|
for (MachineOperand &MO : MI.operands()) {
|
|
if (!MO.isReg())
|
|
continue;
|
|
|
|
const Register VirtReg = MO.getReg();
|
|
if (VirtReg.isPhysical())
|
|
continue;
|
|
|
|
if (!VRM->hasPhys(VirtReg))
|
|
continue;
|
|
|
|
Register PhysReg = VRM->getPhys(VirtReg);
|
|
const unsigned SubReg = MO.getSubReg();
|
|
if (SubReg != 0) {
|
|
PhysReg = TRI->getSubReg(PhysReg, SubReg);
|
|
MO.setSubReg(0);
|
|
}
|
|
|
|
MO.setReg(PhysReg);
|
|
MO.setIsRenamable(false);
|
|
}
|
|
}
|
|
}
|
|
|
|
SIMachineFunctionInfo *MFI = MF.getInfo<SIMachineFunctionInfo>();
|
|
MachineFrameInfo &FrameInfo = MF.getFrameInfo();
|
|
|
|
for (unsigned Reg : RegsToRewrite) {
|
|
LIS->removeInterval(Reg);
|
|
|
|
const Register PhysReg = VRM->getPhys(Reg);
|
|
assert(PhysReg != 0);
|
|
|
|
// Check if PhysReg is already reserved
|
|
if (!MFI->WWMReservedRegs.count(PhysReg)) {
|
|
Optional<int> FI;
|
|
if (!MFI->isEntryFunction()) {
|
|
// Create a stack object for a possible spill in the function prologue.
|
|
// Note: Non-CSR VGPR also need this as we may overwrite inactive lanes.
|
|
const TargetRegisterClass *RC = TRI->getPhysRegClass(PhysReg);
|
|
FI = FrameInfo.CreateSpillStackObject(TRI->getSpillSize(*RC),
|
|
TRI->getSpillAlign(*RC));
|
|
}
|
|
MFI->reserveWWMRegister(PhysReg, FI);
|
|
}
|
|
}
|
|
|
|
RegsToRewrite.clear();
|
|
|
|
// Update the set of reserved registers to include WWM ones.
|
|
MRI->freezeReservedRegs(MF);
|
|
}
|
|
|
|
#ifndef NDEBUG
|
|
LLVM_DUMP_METHOD void
|
|
SIPreAllocateWWMRegs::printWWMInfo(const MachineInstr &MI) {
|
|
|
|
unsigned Opc = MI.getOpcode();
|
|
|
|
if (Opc == AMDGPU::ENTER_STRICT_WWM || Opc == AMDGPU::ENTER_STRICT_WQM) {
|
|
dbgs() << "Entering ";
|
|
} else {
|
|
assert(Opc == AMDGPU::EXIT_STRICT_WWM || Opc == AMDGPU::EXIT_STRICT_WQM);
|
|
dbgs() << "Exiting ";
|
|
}
|
|
|
|
if (Opc == AMDGPU::ENTER_STRICT_WWM || Opc == AMDGPU::EXIT_STRICT_WWM) {
|
|
dbgs() << "Strict WWM ";
|
|
} else {
|
|
assert(Opc == AMDGPU::ENTER_STRICT_WQM || Opc == AMDGPU::EXIT_STRICT_WQM);
|
|
dbgs() << "Strict WQM ";
|
|
}
|
|
|
|
dbgs() << "region: " << MI;
|
|
}
|
|
|
|
#endif
|
|
|
|
bool SIPreAllocateWWMRegs::runOnMachineFunction(MachineFunction &MF) {
|
|
LLVM_DEBUG(dbgs() << "SIPreAllocateWWMRegs: function " << MF.getName() << "\n");
|
|
|
|
const GCNSubtarget &ST = MF.getSubtarget<GCNSubtarget>();
|
|
|
|
TII = ST.getInstrInfo();
|
|
TRI = &TII->getRegisterInfo();
|
|
MRI = &MF.getRegInfo();
|
|
|
|
LIS = &getAnalysis<LiveIntervals>();
|
|
Matrix = &getAnalysis<LiveRegMatrix>();
|
|
VRM = &getAnalysis<VirtRegMap>();
|
|
|
|
RegClassInfo.runOnMachineFunction(MF);
|
|
|
|
bool RegsAssigned = false;
|
|
|
|
// We use a reverse post-order traversal of the control-flow graph to
|
|
// guarantee that we visit definitions in dominance order. Since WWM
|
|
// expressions are guaranteed to never involve phi nodes, and we can only
|
|
// escape WWM through the special WWM instruction, this means that this is a
|
|
// perfect elimination order, so we can never do any better.
|
|
ReversePostOrderTraversal<MachineFunction*> RPOT(&MF);
|
|
|
|
for (MachineBasicBlock *MBB : RPOT) {
|
|
bool InWWM = false;
|
|
for (MachineInstr &MI : *MBB) {
|
|
if (MI.getOpcode() == AMDGPU::V_SET_INACTIVE_B32 ||
|
|
MI.getOpcode() == AMDGPU::V_SET_INACTIVE_B64)
|
|
RegsAssigned |= processDef(MI.getOperand(0));
|
|
|
|
if (MI.getOpcode() == AMDGPU::ENTER_STRICT_WWM ||
|
|
MI.getOpcode() == AMDGPU::ENTER_STRICT_WQM) {
|
|
LLVM_DEBUG(printWWMInfo(MI));
|
|
InWWM = true;
|
|
continue;
|
|
}
|
|
|
|
if (MI.getOpcode() == AMDGPU::EXIT_STRICT_WWM ||
|
|
MI.getOpcode() == AMDGPU::EXIT_STRICT_WQM) {
|
|
LLVM_DEBUG(printWWMInfo(MI));
|
|
InWWM = false;
|
|
}
|
|
|
|
if (!InWWM)
|
|
continue;
|
|
|
|
LLVM_DEBUG(dbgs() << "Processing " << MI);
|
|
|
|
for (MachineOperand &DefOpnd : MI.defs()) {
|
|
RegsAssigned |= processDef(DefOpnd);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (!RegsAssigned)
|
|
return false;
|
|
|
|
rewriteRegs(MF);
|
|
return true;
|
|
}
|