Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
//===- AMDGPUInstructionSelector.cpp ----------------------------*- C++ -*-==//
|
|
|
|
//
|
2019-01-19 09:50:56 +01:00
|
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
/// \file
|
|
|
|
/// This file implements the targeting of the InstructionSelector class for
|
|
|
|
/// AMDGPU.
|
|
|
|
/// \todo This should be generated by TableGen.
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
|
|
#include "AMDGPUInstructionSelector.h"
|
2020-12-25 16:52:14 +01:00
|
|
|
#include "AMDGPU.h"
|
2020-01-03 15:10:00 +01:00
|
|
|
#include "AMDGPUGlobalISelUtils.h"
|
2020-12-25 16:52:14 +01:00
|
|
|
#include "AMDGPUInstrInfo.h"
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
#include "AMDGPURegisterBankInfo.h"
|
AMDGPU/GlobalISel: Enable TableGen'd instruction selector
Reviewers: arsenm, nhaehnle
Reviewed By: arsenm
Subscribers: kzhuravl, wdng, mgorny, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D45994
llvm-svn: 332039
2018-05-10 22:53:06 +02:00
|
|
|
#include "AMDGPUTargetMachine.h"
|
2019-07-16 21:22:21 +02:00
|
|
|
#include "SIMachineFunctionInfo.h"
|
2021-05-07 12:43:29 +02:00
|
|
|
#include "Utils/AMDGPUBaseInfo.h"
|
2019-09-09 17:39:32 +02:00
|
|
|
#include "llvm/CodeGen/GlobalISel/GISelKnownBits.h"
|
AMDGPU/GlobalISel: Enable TableGen'd instruction selector
Reviewers: arsenm, nhaehnle
Reviewed By: arsenm
Subscribers: kzhuravl, wdng, mgorny, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D45994
llvm-svn: 332039
2018-05-10 22:53:06 +02:00
|
|
|
#include "llvm/CodeGen/GlobalISel/InstructionSelectorImpl.h"
|
2019-07-16 21:22:21 +02:00
|
|
|
#include "llvm/CodeGen/GlobalISel/MIPatternMatch.h"
|
2020-12-25 16:52:14 +01:00
|
|
|
#include "llvm/CodeGen/GlobalISel/MachineIRBuilder.h"
|
2020-09-11 00:10:52 +02:00
|
|
|
#include "llvm/IR/DiagnosticInfo.h"
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
|
|
|
|
#define DEBUG_TYPE "amdgpu-isel"
|
|
|
|
|
|
|
|
using namespace llvm;
|
2019-07-16 21:22:21 +02:00
|
|
|
using namespace MIPatternMatch;
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
|
2020-05-20 05:03:39 +02:00
|
|
|
static cl::opt<bool> AllowRiskySelect(
|
|
|
|
"amdgpu-global-isel-risky-select",
|
|
|
|
cl::desc("Allow GlobalISel to select cases that are likely to not work yet"),
|
|
|
|
cl::init(false),
|
|
|
|
cl::ReallyHidden);
|
|
|
|
|
AMDGPU/GlobalISel: Enable TableGen'd instruction selector
Reviewers: arsenm, nhaehnle
Reviewed By: arsenm
Subscribers: kzhuravl, wdng, mgorny, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D45994
llvm-svn: 332039
2018-05-10 22:53:06 +02:00
|
|
|
#define GET_GLOBALISEL_IMPL
|
2018-07-11 22:59:01 +02:00
|
|
|
#define AMDGPUSubtarget GCNSubtarget
|
AMDGPU/GlobalISel: Enable TableGen'd instruction selector
Reviewers: arsenm, nhaehnle
Reviewed By: arsenm
Subscribers: kzhuravl, wdng, mgorny, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D45994
llvm-svn: 332039
2018-05-10 22:53:06 +02:00
|
|
|
#include "AMDGPUGenGlobalISel.inc"
|
|
|
|
#undef GET_GLOBALISEL_IMPL
|
2018-07-11 22:59:01 +02:00
|
|
|
#undef AMDGPUSubtarget
|
AMDGPU/GlobalISel: Enable TableGen'd instruction selector
Reviewers: arsenm, nhaehnle
Reviewed By: arsenm
Subscribers: kzhuravl, wdng, mgorny, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D45994
llvm-svn: 332039
2018-05-10 22:53:06 +02:00
|
|
|
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
AMDGPUInstructionSelector::AMDGPUInstructionSelector(
|
2018-07-11 22:59:01 +02:00
|
|
|
const GCNSubtarget &STI, const AMDGPURegisterBankInfo &RBI,
|
AMDGPU/GlobalISel: Enable TableGen'd instruction selector
Reviewers: arsenm, nhaehnle
Reviewed By: arsenm
Subscribers: kzhuravl, wdng, mgorny, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D45994
llvm-svn: 332039
2018-05-10 22:53:06 +02:00
|
|
|
const AMDGPUTargetMachine &TM)
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
: InstructionSelector(), TII(*STI.getInstrInfo()),
|
AMDGPU/GlobalISel: Enable TableGen'd instruction selector
Reviewers: arsenm, nhaehnle
Reviewed By: arsenm
Subscribers: kzhuravl, wdng, mgorny, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D45994
llvm-svn: 332039
2018-05-10 22:53:06 +02:00
|
|
|
TRI(*STI.getRegisterInfo()), RBI(RBI), TM(TM),
|
|
|
|
STI(STI),
|
|
|
|
EnableLateStructurizeCFG(AMDGPUTargetMachine::EnableLateStructurizeCFG),
|
|
|
|
#define GET_GLOBALISEL_PREDICATES_INIT
|
|
|
|
#include "AMDGPUGenGlobalISel.inc"
|
|
|
|
#undef GET_GLOBALISEL_PREDICATES_INIT
|
|
|
|
#define GET_GLOBALISEL_TEMPORARIES_INIT
|
|
|
|
#include "AMDGPUGenGlobalISel.inc"
|
|
|
|
#undef GET_GLOBALISEL_TEMPORARIES_INIT
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
const char *AMDGPUInstructionSelector::getName() { return DEBUG_TYPE; }
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
|
2021-02-25 07:45:25 +01:00
|
|
|
void AMDGPUInstructionSelector::setupMF(MachineFunction &MF, GISelKnownBits *KB,
|
|
|
|
CodeGenCoverage &CoverageInfo,
|
|
|
|
ProfileSummaryInfo *PSI,
|
|
|
|
BlockFrequencyInfo *BFI) {
|
2019-09-28 05:41:13 +02:00
|
|
|
MRI = &MF.getRegInfo();
|
2020-09-14 10:39:25 +02:00
|
|
|
Subtarget = &MF.getSubtarget<GCNSubtarget>();
|
2021-02-25 07:45:25 +01:00
|
|
|
InstructionSelector::setupMF(MF, KB, CoverageInfo, PSI, BFI);
|
2019-09-28 05:41:13 +02:00
|
|
|
}
|
|
|
|
|
2019-07-01 18:06:02 +02:00
|
|
|
bool AMDGPUInstructionSelector::isVCC(Register Reg,
|
|
|
|
const MachineRegisterInfo &MRI) const {
|
2020-07-22 05:24:02 +02:00
|
|
|
// The verifier is oblivious to s1 being a valid value for wavesize registers.
|
|
|
|
if (Reg.isPhysical())
|
|
|
|
return false;
|
2019-07-01 15:22:07 +02:00
|
|
|
|
|
|
|
auto &RegClassOrBank = MRI.getRegClassOrRegBank(Reg);
|
|
|
|
const TargetRegisterClass *RC =
|
|
|
|
RegClassOrBank.dyn_cast<const TargetRegisterClass*>();
|
|
|
|
if (RC) {
|
2019-07-15 21:44:07 +02:00
|
|
|
const LLT Ty = MRI.getType(Reg);
|
2019-07-01 18:06:02 +02:00
|
|
|
return RC->hasSuperClassEq(TRI.getBoolRC()) &&
|
2019-07-15 21:44:07 +02:00
|
|
|
Ty.isValid() && Ty.getSizeInBits() == 1;
|
2019-07-01 15:22:07 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
const RegisterBank *RB = RegClassOrBank.get<const RegisterBank *>();
|
|
|
|
return RB->getID() == AMDGPU::VCCRegBankID;
|
|
|
|
}
|
|
|
|
|
2020-01-17 06:22:23 +01:00
|
|
|
bool AMDGPUInstructionSelector::constrainCopyLikeIntrin(MachineInstr &MI,
|
|
|
|
unsigned NewOpc) const {
|
|
|
|
MI.setDesc(TII.get(NewOpc));
|
|
|
|
MI.RemoveOperand(1); // Remove intrinsic ID.
|
|
|
|
MI.addOperand(*MF, MachineOperand::CreateReg(AMDGPU::EXEC, false, true));
|
|
|
|
|
|
|
|
MachineOperand &Dst = MI.getOperand(0);
|
|
|
|
MachineOperand &Src = MI.getOperand(1);
|
|
|
|
|
|
|
|
// TODO: This should be legalized to s32 if needed
|
|
|
|
if (MRI->getType(Dst.getReg()) == LLT::scalar(1))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const TargetRegisterClass *DstRC
|
|
|
|
= TRI.getConstrainedRegClassForOperand(Dst, *MRI);
|
|
|
|
const TargetRegisterClass *SrcRC
|
|
|
|
= TRI.getConstrainedRegClassForOperand(Src, *MRI);
|
|
|
|
if (!DstRC || DstRC != SrcRC)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return RBI.constrainGenericRegister(Dst.getReg(), *DstRC, *MRI) &&
|
|
|
|
RBI.constrainGenericRegister(Src.getReg(), *SrcRC, *MRI);
|
|
|
|
}
|
|
|
|
|
2018-05-10 23:20:10 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectCOPY(MachineInstr &I) const {
|
2019-07-15 21:44:07 +02:00
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
2018-05-10 23:20:10 +02:00
|
|
|
MachineBasicBlock *BB = I.getParent();
|
|
|
|
I.setDesc(TII.get(TargetOpcode::COPY));
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
|
|
|
|
const MachineOperand &Src = I.getOperand(1);
|
2019-07-15 21:44:07 +02:00
|
|
|
MachineOperand &Dst = I.getOperand(0);
|
|
|
|
Register DstReg = Dst.getReg();
|
|
|
|
Register SrcReg = Src.getReg();
|
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
if (isVCC(DstReg, *MRI)) {
|
2019-07-15 21:44:07 +02:00
|
|
|
if (SrcReg == AMDGPU::SCC) {
|
|
|
|
const TargetRegisterClass *RC
|
2019-09-28 05:41:13 +02:00
|
|
|
= TRI.getConstrainedRegClassForOperand(Dst, *MRI);
|
2019-07-15 21:44:07 +02:00
|
|
|
if (!RC)
|
|
|
|
return true;
|
2019-09-28 05:41:13 +02:00
|
|
|
return RBI.constrainGenericRegister(DstReg, *RC, *MRI);
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
}
|
2019-07-15 21:44:07 +02:00
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
if (!isVCC(SrcReg, *MRI)) {
|
2019-07-15 21:46:48 +02:00
|
|
|
// TODO: Should probably leave the copy and let copyPhysReg expand it.
|
2019-09-28 05:41:13 +02:00
|
|
|
if (!RBI.constrainGenericRegister(DstReg, *TRI.getBoolRC(), *MRI))
|
2019-07-15 21:46:48 +02:00
|
|
|
return false;
|
|
|
|
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
const TargetRegisterClass *SrcRC
|
|
|
|
= TRI.getConstrainedRegClassForOperand(Src, *MRI);
|
|
|
|
|
2021-01-28 11:12:12 +01:00
|
|
|
Optional<ValueAndVReg> ConstVal =
|
|
|
|
getConstantVRegValWithLookThrough(SrcReg, *MRI, true, true);
|
|
|
|
if (ConstVal) {
|
|
|
|
unsigned MovOpc =
|
|
|
|
STI.isWave64() ? AMDGPU::S_MOV_B64 : AMDGPU::S_MOV_B32;
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(MovOpc), DstReg)
|
|
|
|
.addImm(ConstVal->Value.getBoolValue() ? -1 : 0);
|
|
|
|
} else {
|
|
|
|
Register MaskedReg = MRI->createVirtualRegister(SrcRC);
|
|
|
|
|
|
|
|
// We can't trust the high bits at this point, so clear them.
|
|
|
|
|
|
|
|
// TODO: Skip masking high bits if def is known boolean.
|
|
|
|
|
|
|
|
unsigned AndOpc =
|
|
|
|
TRI.isSGPRClass(SrcRC) ? AMDGPU::S_AND_B32 : AMDGPU::V_AND_B32_e32;
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(AndOpc), MaskedReg)
|
|
|
|
.addImm(1)
|
|
|
|
.addReg(SrcReg);
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::V_CMP_NE_U32_e64), DstReg)
|
|
|
|
.addImm(0)
|
|
|
|
.addReg(MaskedReg);
|
|
|
|
}
|
2019-07-15 21:46:48 +02:00
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
if (!MRI->getRegClassOrNull(SrcReg))
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
MRI->setRegClass(SrcReg, SrcRC);
|
2019-07-15 21:46:48 +02:00
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
2019-07-15 21:48:36 +02:00
|
|
|
|
|
|
|
const TargetRegisterClass *RC =
|
2019-09-28 05:41:13 +02:00
|
|
|
TRI.getConstrainedRegClassForOperand(Dst, *MRI);
|
|
|
|
if (RC && !RBI.constrainGenericRegister(DstReg, *RC, *MRI))
|
2019-07-15 21:48:36 +02:00
|
|
|
return false;
|
|
|
|
|
|
|
|
return true;
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
}
|
|
|
|
|
2018-05-10 23:20:10 +02:00
|
|
|
for (const MachineOperand &MO : I.operands()) {
|
2020-08-20 18:46:16 +02:00
|
|
|
if (MO.getReg().isPhysical())
|
2018-05-10 23:20:10 +02:00
|
|
|
continue;
|
|
|
|
|
|
|
|
const TargetRegisterClass *RC =
|
2019-09-28 05:41:13 +02:00
|
|
|
TRI.getConstrainedRegClassForOperand(MO, *MRI);
|
2018-05-10 23:20:10 +02:00
|
|
|
if (!RC)
|
|
|
|
continue;
|
2019-09-28 05:41:13 +02:00
|
|
|
RBI.constrainGenericRegister(MO.getReg(), *RC, *MRI);
|
2018-05-10 23:20:10 +02:00
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-07-01 18:32:47 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectPHI(MachineInstr &I) const {
|
|
|
|
const Register DefReg = I.getOperand(0).getReg();
|
2019-09-28 05:41:13 +02:00
|
|
|
const LLT DefTy = MRI->getType(DefReg);
|
2020-05-20 05:03:39 +02:00
|
|
|
if (DefTy == LLT::scalar(1)) {
|
|
|
|
if (!AllowRiskySelect) {
|
|
|
|
LLVM_DEBUG(dbgs() << "Skipping risky boolean phi\n");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
LLVM_DEBUG(dbgs() << "Selecting risky boolean phi\n");
|
|
|
|
}
|
2019-07-01 18:32:47 +02:00
|
|
|
|
|
|
|
// TODO: Verify this doesn't have insane operands (i.e. VGPR to SGPR copy)
|
|
|
|
|
|
|
|
const RegClassOrRegBank &RegClassOrBank =
|
2019-09-28 05:41:13 +02:00
|
|
|
MRI->getRegClassOrRegBank(DefReg);
|
2019-07-01 18:32:47 +02:00
|
|
|
|
|
|
|
const TargetRegisterClass *DefRC
|
|
|
|
= RegClassOrBank.dyn_cast<const TargetRegisterClass *>();
|
|
|
|
if (!DefRC) {
|
|
|
|
if (!DefTy.isValid()) {
|
|
|
|
LLVM_DEBUG(dbgs() << "PHI operand has no type, not a gvreg?\n");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
const RegisterBank &RB = *RegClassOrBank.get<const RegisterBank *>();
|
2019-09-28 05:41:13 +02:00
|
|
|
DefRC = TRI.getRegClassForTypeOnBank(DefTy, RB, *MRI);
|
2019-07-01 18:32:47 +02:00
|
|
|
if (!DefRC) {
|
|
|
|
LLVM_DEBUG(dbgs() << "PHI operand has unexpected size/bank\n");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
// TODO: Verify that all registers have the same bank
|
2019-07-01 18:32:47 +02:00
|
|
|
I.setDesc(TII.get(TargetOpcode::PHI));
|
2019-09-28 05:41:13 +02:00
|
|
|
return RBI.constrainGenericRegister(DefReg, *DefRC, *MRI);
|
2019-07-01 18:32:47 +02:00
|
|
|
}
|
|
|
|
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
MachineOperand
|
|
|
|
AMDGPUInstructionSelector::getSubOperand64(MachineOperand &MO,
|
2019-07-01 18:34:48 +02:00
|
|
|
const TargetRegisterClass &SubRC,
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
unsigned SubIdx) const {
|
|
|
|
|
|
|
|
MachineInstr *MI = MO.getParent();
|
|
|
|
MachineBasicBlock *BB = MO.getParent()->getParent();
|
2019-09-28 05:41:13 +02:00
|
|
|
Register DstReg = MRI->createVirtualRegister(&SubRC);
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
|
|
|
|
if (MO.isReg()) {
|
|
|
|
unsigned ComposedSubIdx = TRI.composeSubRegIndices(MO.getSubReg(), SubIdx);
|
Apply llvm-prefer-register-over-unsigned from clang-tidy to LLVM
Summary:
This clang-tidy check is looking for unsigned integer variables whose initializer
starts with an implicit cast from llvm::Register and changes the type of the
variable to llvm::Register (dropping the llvm:: where possible).
Partial reverts in:
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
X86FixupLEAs.cpp - Some functions return unsigned and arguably should be MCRegister
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
HexagonBitSimplify.cpp - Function takes BitTracker::RegisterRef which appears to be unsigned&
MachineVerifier.cpp - Ambiguous operator==() given MCRegister and const Register
PPCFastISel.cpp - No Register::operator-=()
PeepholeOptimizer.cpp - TargetInstrInfo::optimizeLoadInstr() takes an unsigned&
MachineTraceMetrics.cpp - MachineTraceMetrics lacks a suitable constructor
Manual fixups in:
ARMFastISel.cpp - ARMEmitLoad() now takes a Register& instead of unsigned&
HexagonSplitDouble.cpp - Ternary operator was ambiguous between unsigned/Register
HexagonConstExtenders.cpp - Has a local class named Register, used llvm::Register instead of Register.
PPCFastISel.cpp - PPCEmitLoad() now takes a Register& instead of unsigned&
Depends on D65919
Reviewers: arsenm, bogner, craig.topper, RKSimon
Reviewed By: arsenm
Subscribers: RKSimon, craig.topper, lenary, aemerson, wuzish, jholewinski, MatzeB, qcolombet, dschuff, jyknight, dylanmckay, sdardis, nemanjai, jvesely, wdng, nhaehnle, sbc100, jgravelle-google, kristof.beyls, hiraditya, aheejin, kbarton, fedor.sergeev, javed.absar, asb, rbar, johnrusso, simoncook, apazos, sabuasal, niosHD, jrtc27, MaskRay, zzheng, edward-jones, atanasyan, rogfer01, MartinMosbeck, brucehoult, the_o, tpr, PkmX, jocewei, jsji, Petar.Avramovic, asbirlea, Jim, s.egerton, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D65962
llvm-svn: 369041
2019-08-15 21:22:08 +02:00
|
|
|
Register Reg = MO.getReg();
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
BuildMI(*BB, MI, MI->getDebugLoc(), TII.get(AMDGPU::COPY), DstReg)
|
|
|
|
.addReg(Reg, 0, ComposedSubIdx);
|
|
|
|
|
|
|
|
return MachineOperand::CreateReg(DstReg, MO.isDef(), MO.isImplicit(),
|
|
|
|
MO.isKill(), MO.isDead(), MO.isUndef(),
|
|
|
|
MO.isEarlyClobber(), 0, MO.isDebug(),
|
|
|
|
MO.isInternalRead());
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(MO.isImm());
|
|
|
|
|
|
|
|
APInt Imm(64, MO.getImm());
|
|
|
|
|
|
|
|
switch (SubIdx) {
|
|
|
|
default:
|
|
|
|
llvm_unreachable("do not know to split immediate with this sub index.");
|
|
|
|
case AMDGPU::sub0:
|
|
|
|
return MachineOperand::CreateImm(Imm.getLoBits(32).getSExtValue());
|
|
|
|
case AMDGPU::sub1:
|
|
|
|
return MachineOperand::CreateImm(Imm.getHiBits(32).getSExtValue());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-07-15 21:50:07 +02:00
|
|
|
static unsigned getLogicalBitOpcode(unsigned Opc, bool Is64) {
|
|
|
|
switch (Opc) {
|
|
|
|
case AMDGPU::G_AND:
|
|
|
|
return Is64 ? AMDGPU::S_AND_B64 : AMDGPU::S_AND_B32;
|
|
|
|
case AMDGPU::G_OR:
|
|
|
|
return Is64 ? AMDGPU::S_OR_B64 : AMDGPU::S_OR_B32;
|
|
|
|
case AMDGPU::G_XOR:
|
|
|
|
return Is64 ? AMDGPU::S_XOR_B64 : AMDGPU::S_XOR_B32;
|
|
|
|
default:
|
|
|
|
llvm_unreachable("not a bit op");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AMDGPUInstructionSelector::selectG_AND_OR_XOR(MachineInstr &I) const {
|
2020-07-26 16:47:08 +02:00
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
2019-09-28 05:41:13 +02:00
|
|
|
unsigned Size = RBI.getSizeInBits(DstReg, *MRI, TRI);
|
2019-07-15 21:50:07 +02:00
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
const RegisterBank *DstRB = RBI.getRegBank(DstReg, *MRI, TRI);
|
2020-07-26 16:47:08 +02:00
|
|
|
if (DstRB->getID() != AMDGPU::SGPRRegBankID &&
|
|
|
|
DstRB->getID() != AMDGPU::VCCRegBankID)
|
|
|
|
return false;
|
2019-07-15 21:50:07 +02:00
|
|
|
|
2020-07-26 16:47:08 +02:00
|
|
|
bool Is64 = Size > 32 || (DstRB->getID() == AMDGPU::VCCRegBankID &&
|
|
|
|
STI.isWave64());
|
|
|
|
I.setDesc(TII.get(getLogicalBitOpcode(I.getOpcode(), Is64)));
|
|
|
|
|
|
|
|
// Dead implicit-def of scc
|
|
|
|
I.addOperand(MachineOperand::CreateReg(AMDGPU::SCC, true, // isDef
|
|
|
|
true, // isImp
|
|
|
|
false, // isKill
|
|
|
|
true)); // isDead
|
|
|
|
return constrainSelectedInstRegOperands(I, TII, TRI, RBI);
|
2019-07-15 21:50:07 +02:00
|
|
|
}
|
|
|
|
|
2019-07-09 16:05:11 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_ADD_SUB(MachineInstr &I) const {
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
MachineBasicBlock *BB = I.getParent();
|
|
|
|
MachineFunction *MF = BB->getParent();
|
2019-07-01 18:34:48 +02:00
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
2020-02-19 19:23:11 +01:00
|
|
|
LLT Ty = MRI->getType(DstReg);
|
|
|
|
if (Ty.isVector())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
unsigned Size = Ty.getSizeInBits();
|
2019-09-28 05:41:13 +02:00
|
|
|
const RegisterBank *DstRB = RBI.getRegBank(DstReg, *MRI, TRI);
|
2019-07-01 18:34:48 +02:00
|
|
|
const bool IsSALU = DstRB->getID() == AMDGPU::SGPRRegBankID;
|
2019-07-09 16:05:11 +02:00
|
|
|
const bool Sub = I.getOpcode() == TargetOpcode::G_SUB;
|
2019-07-01 18:34:48 +02:00
|
|
|
|
|
|
|
if (Size == 32) {
|
|
|
|
if (IsSALU) {
|
2019-07-09 16:05:11 +02:00
|
|
|
const unsigned Opc = Sub ? AMDGPU::S_SUB_U32 : AMDGPU::S_ADD_U32;
|
2019-07-01 18:34:48 +02:00
|
|
|
MachineInstr *Add =
|
2019-07-09 16:05:11 +02:00
|
|
|
BuildMI(*BB, &I, DL, TII.get(Opc), DstReg)
|
2019-07-01 18:34:48 +02:00
|
|
|
.add(I.getOperand(1))
|
|
|
|
.add(I.getOperand(2));
|
|
|
|
I.eraseFromParent();
|
|
|
|
return constrainSelectedInstRegOperands(*Add, TII, TRI, RBI);
|
|
|
|
}
|
2017-01-31 16:24:11 +01:00
|
|
|
|
2019-07-01 18:34:48 +02:00
|
|
|
if (STI.hasAddNoCarry()) {
|
2019-07-09 16:05:11 +02:00
|
|
|
const unsigned Opc = Sub ? AMDGPU::V_SUB_U32_e64 : AMDGPU::V_ADD_U32_e64;
|
|
|
|
I.setDesc(TII.get(Opc));
|
2019-07-01 18:34:48 +02:00
|
|
|
I.addOperand(*MF, MachineOperand::CreateImm(0));
|
|
|
|
I.addOperand(*MF, MachineOperand::CreateReg(AMDGPU::EXEC, false, true));
|
|
|
|
return constrainSelectedInstRegOperands(I, TII, TRI, RBI);
|
|
|
|
}
|
2017-01-31 16:24:11 +01:00
|
|
|
|
2020-07-14 15:18:36 +02:00
|
|
|
const unsigned Opc = Sub ? AMDGPU::V_SUB_CO_U32_e64 : AMDGPU::V_ADD_CO_U32_e64;
|
2019-07-09 16:05:11 +02:00
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
Register UnusedCarry = MRI->createVirtualRegister(TRI.getWaveMaskRegClass());
|
2019-07-01 18:34:48 +02:00
|
|
|
MachineInstr *Add
|
2019-07-09 16:05:11 +02:00
|
|
|
= BuildMI(*BB, &I, DL, TII.get(Opc), DstReg)
|
2019-07-01 18:34:48 +02:00
|
|
|
.addDef(UnusedCarry, RegState::Dead)
|
|
|
|
.add(I.getOperand(1))
|
|
|
|
.add(I.getOperand(2))
|
|
|
|
.addImm(0);
|
|
|
|
I.eraseFromParent();
|
|
|
|
return constrainSelectedInstRegOperands(*Add, TII, TRI, RBI);
|
|
|
|
}
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
|
2019-07-09 16:05:11 +02:00
|
|
|
assert(!Sub && "illegal sub should not reach here");
|
|
|
|
|
2019-07-01 18:34:48 +02:00
|
|
|
const TargetRegisterClass &RC
|
|
|
|
= IsSALU ? AMDGPU::SReg_64_XEXECRegClass : AMDGPU::VReg_64RegClass;
|
|
|
|
const TargetRegisterClass &HalfRC
|
|
|
|
= IsSALU ? AMDGPU::SReg_32RegClass : AMDGPU::VGPR_32RegClass;
|
|
|
|
|
|
|
|
MachineOperand Lo1(getSubOperand64(I.getOperand(1), HalfRC, AMDGPU::sub0));
|
|
|
|
MachineOperand Lo2(getSubOperand64(I.getOperand(2), HalfRC, AMDGPU::sub0));
|
|
|
|
MachineOperand Hi1(getSubOperand64(I.getOperand(1), HalfRC, AMDGPU::sub1));
|
|
|
|
MachineOperand Hi2(getSubOperand64(I.getOperand(2), HalfRC, AMDGPU::sub1));
|
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
Register DstLo = MRI->createVirtualRegister(&HalfRC);
|
|
|
|
Register DstHi = MRI->createVirtualRegister(&HalfRC);
|
2019-07-01 18:34:48 +02:00
|
|
|
|
|
|
|
if (IsSALU) {
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::S_ADD_U32), DstLo)
|
|
|
|
.add(Lo1)
|
|
|
|
.add(Lo2);
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::S_ADDC_U32), DstHi)
|
|
|
|
.add(Hi1)
|
|
|
|
.add(Hi2);
|
|
|
|
} else {
|
|
|
|
const TargetRegisterClass *CarryRC = TRI.getWaveMaskRegClass();
|
2019-09-28 05:41:13 +02:00
|
|
|
Register CarryReg = MRI->createVirtualRegister(CarryRC);
|
2020-07-14 15:18:36 +02:00
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::V_ADD_CO_U32_e64), DstLo)
|
2019-07-01 18:34:48 +02:00
|
|
|
.addDef(CarryReg)
|
|
|
|
.add(Lo1)
|
|
|
|
.add(Lo2)
|
|
|
|
.addImm(0);
|
2019-07-02 16:40:22 +02:00
|
|
|
MachineInstr *Addc = BuildMI(*BB, &I, DL, TII.get(AMDGPU::V_ADDC_U32_e64), DstHi)
|
2019-09-28 05:41:13 +02:00
|
|
|
.addDef(MRI->createVirtualRegister(CarryRC), RegState::Dead)
|
2019-07-01 18:34:48 +02:00
|
|
|
.add(Hi1)
|
|
|
|
.add(Hi2)
|
|
|
|
.addReg(CarryReg, RegState::Kill)
|
|
|
|
.addImm(0);
|
2019-07-02 16:40:22 +02:00
|
|
|
|
|
|
|
if (!constrainSelectedInstRegOperands(*Addc, TII, TRI, RBI))
|
|
|
|
return false;
|
2019-07-01 18:34:48 +02:00
|
|
|
}
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
|
2019-07-01 18:34:48 +02:00
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::REG_SEQUENCE), DstReg)
|
|
|
|
.addReg(DstLo)
|
|
|
|
.addImm(AMDGPU::sub0)
|
|
|
|
.addReg(DstHi)
|
|
|
|
.addImm(AMDGPU::sub1);
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
|
2019-07-02 16:40:22 +02:00
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
if (!RBI.constrainGenericRegister(DstReg, RC, *MRI))
|
2019-07-01 18:34:48 +02:00
|
|
|
return false;
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
|
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-12-25 15:06:05 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectG_UADDO_USUBO_UADDE_USUBE(
|
|
|
|
MachineInstr &I) const {
|
2019-10-01 03:23:13 +02:00
|
|
|
MachineBasicBlock *BB = I.getParent();
|
|
|
|
MachineFunction *MF = BB->getParent();
|
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
Register Dst0Reg = I.getOperand(0).getReg();
|
|
|
|
Register Dst1Reg = I.getOperand(1).getReg();
|
2019-12-25 15:06:05 +01:00
|
|
|
const bool IsAdd = I.getOpcode() == AMDGPU::G_UADDO ||
|
|
|
|
I.getOpcode() == AMDGPU::G_UADDE;
|
|
|
|
const bool HasCarryIn = I.getOpcode() == AMDGPU::G_UADDE ||
|
|
|
|
I.getOpcode() == AMDGPU::G_USUBE;
|
2019-10-01 03:23:13 +02:00
|
|
|
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
if (isVCC(Dst1Reg, *MRI)) {
|
2020-07-20 17:37:31 +02:00
|
|
|
unsigned NoCarryOpc =
|
|
|
|
IsAdd ? AMDGPU::V_ADD_CO_U32_e64 : AMDGPU::V_SUB_CO_U32_e64;
|
2019-12-25 15:06:05 +01:00
|
|
|
unsigned CarryOpc = IsAdd ? AMDGPU::V_ADDC_U32_e64 : AMDGPU::V_SUBB_U32_e64;
|
|
|
|
I.setDesc(TII.get(HasCarryIn ? CarryOpc : NoCarryOpc));
|
2019-10-01 03:23:13 +02:00
|
|
|
I.addOperand(*MF, MachineOperand::CreateReg(AMDGPU::EXEC, false, true));
|
|
|
|
I.addOperand(*MF, MachineOperand::CreateImm(0));
|
|
|
|
return constrainSelectedInstRegOperands(I, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
|
|
|
Register Src0Reg = I.getOperand(2).getReg();
|
|
|
|
Register Src1Reg = I.getOperand(3).getReg();
|
2019-12-25 15:06:05 +01:00
|
|
|
|
|
|
|
if (HasCarryIn) {
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::COPY), AMDGPU::SCC)
|
|
|
|
.addReg(I.getOperand(4).getReg());
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned NoCarryOpc = IsAdd ? AMDGPU::S_ADD_U32 : AMDGPU::S_SUB_U32;
|
|
|
|
unsigned CarryOpc = IsAdd ? AMDGPU::S_ADDC_U32 : AMDGPU::S_SUBB_U32;
|
|
|
|
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(HasCarryIn ? CarryOpc : NoCarryOpc), Dst0Reg)
|
2019-10-01 03:23:13 +02:00
|
|
|
.add(I.getOperand(2))
|
|
|
|
.add(I.getOperand(3));
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::COPY), Dst1Reg)
|
|
|
|
.addReg(AMDGPU::SCC);
|
|
|
|
|
2019-12-29 15:05:56 +01:00
|
|
|
if (!MRI->getRegClassOrNull(Dst1Reg))
|
|
|
|
MRI->setRegClass(Dst1Reg, &AMDGPU::SReg_32RegClass);
|
2019-10-01 03:23:13 +02:00
|
|
|
|
2019-12-29 15:05:56 +01:00
|
|
|
if (!RBI.constrainGenericRegister(Dst0Reg, AMDGPU::SReg_32RegClass, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(Src0Reg, AMDGPU::SReg_32RegClass, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(Src1Reg, AMDGPU::SReg_32RegClass, *MRI))
|
2019-10-01 03:23:13 +02:00
|
|
|
return false;
|
|
|
|
|
2019-12-25 15:06:05 +01:00
|
|
|
if (HasCarryIn &&
|
|
|
|
!RBI.constrainGenericRegister(I.getOperand(4).getReg(),
|
|
|
|
AMDGPU::SReg_32RegClass, *MRI))
|
|
|
|
return false;
|
|
|
|
|
2019-10-01 03:23:13 +02:00
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-02-07 21:18:58 +01:00
|
|
|
// TODO: We should probably legalize these to only using 32-bit results.
|
2019-03-01 00:37:48 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectG_EXTRACT(MachineInstr &I) const {
|
|
|
|
MachineBasicBlock *BB = I.getParent();
|
2020-01-01 21:23:58 +01:00
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
|
|
|
Register SrcReg = I.getOperand(1).getReg();
|
|
|
|
LLT DstTy = MRI->getType(DstReg);
|
|
|
|
LLT SrcTy = MRI->getType(SrcReg);
|
|
|
|
const unsigned SrcSize = SrcTy.getSizeInBits();
|
2020-05-20 19:26:10 +02:00
|
|
|
unsigned DstSize = DstTy.getSizeInBits();
|
2020-01-01 21:23:58 +01:00
|
|
|
|
|
|
|
// TODO: Should handle any multiple of 32 offset.
|
2019-10-06 03:41:22 +02:00
|
|
|
unsigned Offset = I.getOperand(2).getImm();
|
2020-02-07 21:18:58 +01:00
|
|
|
if (Offset % 32 != 0 || DstSize > 128)
|
|
|
|
return false;
|
|
|
|
|
2020-05-20 19:26:10 +02:00
|
|
|
// 16-bit operations really use 32-bit registers.
|
|
|
|
// FIXME: Probably should not allow 16-bit G_EXTRACT results.
|
|
|
|
if (DstSize == 16)
|
|
|
|
DstSize = 32;
|
|
|
|
|
2020-02-07 21:18:58 +01:00
|
|
|
const TargetRegisterClass *DstRC =
|
|
|
|
TRI.getConstrainedRegClassForOperand(I.getOperand(0), *MRI);
|
|
|
|
if (!DstRC || !RBI.constrainGenericRegister(DstReg, *DstRC, *MRI))
|
2019-10-06 03:41:22 +02:00
|
|
|
return false;
|
|
|
|
|
2020-01-01 21:23:58 +01:00
|
|
|
const RegisterBank *SrcBank = RBI.getRegBank(SrcReg, *MRI, TRI);
|
|
|
|
const TargetRegisterClass *SrcRC =
|
|
|
|
TRI.getRegClassForSizeOnBank(SrcSize, *SrcBank, *MRI);
|
|
|
|
if (!SrcRC)
|
|
|
|
return false;
|
2020-02-07 21:18:58 +01:00
|
|
|
unsigned SubReg = SIRegisterInfo::getSubRegFromChannel(Offset / 32,
|
|
|
|
DstSize / 32);
|
|
|
|
SrcRC = TRI.getSubClassWithSubReg(SrcRC, SubReg);
|
|
|
|
if (!SrcRC)
|
|
|
|
return false;
|
2020-01-01 21:23:58 +01:00
|
|
|
|
2020-02-07 21:18:58 +01:00
|
|
|
SrcReg = constrainOperandRegClass(*MF, TRI, *MRI, TII, RBI, I,
|
|
|
|
*SrcRC, I.getOperand(1));
|
2019-03-01 00:37:48 +01:00
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
2020-02-07 21:18:58 +01:00
|
|
|
BuildMI(*BB, &I, DL, TII.get(TargetOpcode::COPY), DstReg)
|
|
|
|
.addReg(SrcReg, 0, SubReg);
|
2019-03-01 00:37:48 +01:00
|
|
|
|
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-07-09 16:02:20 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_MERGE_VALUES(MachineInstr &MI) const {
|
|
|
|
MachineBasicBlock *BB = MI.getParent();
|
|
|
|
Register DstReg = MI.getOperand(0).getReg();
|
2019-09-28 05:41:13 +02:00
|
|
|
LLT DstTy = MRI->getType(DstReg);
|
|
|
|
LLT SrcTy = MRI->getType(MI.getOperand(1).getReg());
|
2019-07-09 16:02:20 +02:00
|
|
|
|
|
|
|
const unsigned SrcSize = SrcTy.getSizeInBits();
|
2019-07-15 19:26:43 +02:00
|
|
|
if (SrcSize < 32)
|
2019-09-02 09:27:20 +02:00
|
|
|
return selectImpl(MI, *CoverageInfo);
|
2019-07-15 19:26:43 +02:00
|
|
|
|
2019-07-09 16:02:20 +02:00
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
2019-09-28 05:41:13 +02:00
|
|
|
const RegisterBank *DstBank = RBI.getRegBank(DstReg, *MRI, TRI);
|
2019-07-09 16:02:20 +02:00
|
|
|
const unsigned DstSize = DstTy.getSizeInBits();
|
|
|
|
const TargetRegisterClass *DstRC =
|
2019-09-28 05:41:13 +02:00
|
|
|
TRI.getRegClassForSizeOnBank(DstSize, *DstBank, *MRI);
|
2019-07-09 16:02:20 +02:00
|
|
|
if (!DstRC)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
ArrayRef<int16_t> SubRegs = TRI.getRegSplitParts(DstRC, SrcSize / 8);
|
|
|
|
MachineInstrBuilder MIB =
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(TargetOpcode::REG_SEQUENCE), DstReg);
|
|
|
|
for (int I = 0, E = MI.getNumOperands() - 1; I != E; ++I) {
|
|
|
|
MachineOperand &Src = MI.getOperand(I + 1);
|
|
|
|
MIB.addReg(Src.getReg(), getUndefRegState(Src.isUndef()));
|
|
|
|
MIB.addImm(SubRegs[I]);
|
|
|
|
|
|
|
|
const TargetRegisterClass *SrcRC
|
2019-09-28 05:41:13 +02:00
|
|
|
= TRI.getConstrainedRegClassForOperand(Src, *MRI);
|
|
|
|
if (SrcRC && !RBI.constrainGenericRegister(Src.getReg(), *SrcRC, *MRI))
|
2019-07-09 16:02:20 +02:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
if (!RBI.constrainGenericRegister(DstReg, *DstRC, *MRI))
|
2019-07-09 16:02:20 +02:00
|
|
|
return false;
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-07-09 16:02:26 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_UNMERGE_VALUES(MachineInstr &MI) const {
|
|
|
|
MachineBasicBlock *BB = MI.getParent();
|
|
|
|
const int NumDst = MI.getNumOperands() - 1;
|
|
|
|
|
|
|
|
MachineOperand &Src = MI.getOperand(NumDst);
|
|
|
|
|
|
|
|
Register SrcReg = Src.getReg();
|
|
|
|
Register DstReg0 = MI.getOperand(0).getReg();
|
2019-09-28 05:41:13 +02:00
|
|
|
LLT DstTy = MRI->getType(DstReg0);
|
|
|
|
LLT SrcTy = MRI->getType(SrcReg);
|
2019-07-09 16:02:26 +02:00
|
|
|
|
|
|
|
const unsigned DstSize = DstTy.getSizeInBits();
|
|
|
|
const unsigned SrcSize = SrcTy.getSizeInBits();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
2019-09-28 05:41:13 +02:00
|
|
|
const RegisterBank *SrcBank = RBI.getRegBank(SrcReg, *MRI, TRI);
|
2019-07-09 16:02:26 +02:00
|
|
|
|
|
|
|
const TargetRegisterClass *SrcRC =
|
2019-09-28 05:41:13 +02:00
|
|
|
TRI.getRegClassForSizeOnBank(SrcSize, *SrcBank, *MRI);
|
|
|
|
if (!SrcRC || !RBI.constrainGenericRegister(SrcReg, *SrcRC, *MRI))
|
2019-07-09 16:02:26 +02:00
|
|
|
return false;
|
|
|
|
|
|
|
|
// Note we could have mixed SGPR and VGPR destination banks for an SGPR
|
|
|
|
// source, and this relies on the fact that the same subregister indices are
|
|
|
|
// used for both.
|
|
|
|
ArrayRef<int16_t> SubRegs = TRI.getRegSplitParts(SrcRC, DstSize / 8);
|
|
|
|
for (int I = 0, E = NumDst; I != E; ++I) {
|
|
|
|
MachineOperand &Dst = MI.getOperand(I);
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(TargetOpcode::COPY), Dst.getReg())
|
2020-08-09 21:01:21 +02:00
|
|
|
.addReg(SrcReg, 0, SubRegs[I]);
|
2019-07-09 16:02:26 +02:00
|
|
|
|
2020-08-04 15:19:43 +02:00
|
|
|
// Make sure the subregister index is valid for the source register.
|
|
|
|
SrcRC = TRI.getSubClassWithSubReg(SrcRC, SubRegs[I]);
|
|
|
|
if (!SrcRC || !RBI.constrainGenericRegister(SrcReg, *SrcRC, *MRI))
|
|
|
|
return false;
|
|
|
|
|
2019-07-09 16:02:26 +02:00
|
|
|
const TargetRegisterClass *DstRC =
|
2019-09-28 05:41:13 +02:00
|
|
|
TRI.getConstrainedRegClassForOperand(Dst, *MRI);
|
|
|
|
if (DstRC && !RBI.constrainGenericRegister(Dst.getReg(), *DstRC, *MRI))
|
2019-07-09 16:02:26 +02:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-02-15 23:14:17 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectG_BUILD_VECTOR_TRUNC(
|
|
|
|
MachineInstr &MI) const {
|
|
|
|
if (selectImpl(MI, *CoverageInfo))
|
|
|
|
return true;
|
|
|
|
|
|
|
|
const LLT S32 = LLT::scalar(32);
|
2021-06-24 10:58:21 +02:00
|
|
|
const LLT V2S16 = LLT::fixed_vector(2, 16);
|
2020-02-15 23:14:17 +01:00
|
|
|
|
|
|
|
Register Dst = MI.getOperand(0).getReg();
|
|
|
|
if (MRI->getType(Dst) != V2S16)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const RegisterBank *DstBank = RBI.getRegBank(Dst, *MRI, TRI);
|
|
|
|
if (DstBank->getID() != AMDGPU::SGPRRegBankID)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
Register Src0 = MI.getOperand(1).getReg();
|
|
|
|
Register Src1 = MI.getOperand(2).getReg();
|
|
|
|
if (MRI->getType(Src0) != S32)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
MachineBasicBlock *BB = MI.getParent();
|
|
|
|
|
2021-01-20 11:38:05 +01:00
|
|
|
auto ConstSrc1 =
|
|
|
|
getConstantVRegValWithLookThrough(Src1, *MRI, true, true, true);
|
2020-07-18 21:30:59 +02:00
|
|
|
if (ConstSrc1) {
|
2021-01-20 11:38:05 +01:00
|
|
|
auto ConstSrc0 =
|
|
|
|
getConstantVRegValWithLookThrough(Src0, *MRI, true, true, true);
|
2020-07-18 21:30:59 +02:00
|
|
|
if (ConstSrc0) {
|
2020-11-03 15:50:17 +01:00
|
|
|
const int64_t K0 = ConstSrc0->Value.getSExtValue();
|
|
|
|
const int64_t K1 = ConstSrc1->Value.getSExtValue();
|
|
|
|
uint32_t Lo16 = static_cast<uint32_t>(K0) & 0xffff;
|
|
|
|
uint32_t Hi16 = static_cast<uint32_t>(K1) & 0xffff;
|
2020-07-18 21:30:59 +02:00
|
|
|
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::S_MOV_B32), Dst)
|
|
|
|
.addImm(Lo16 | (Hi16 << 16));
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return RBI.constrainGenericRegister(Dst, AMDGPU::SReg_32RegClass, *MRI);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-02-15 23:14:17 +01:00
|
|
|
// TODO: This should probably be a combine somewhere
|
|
|
|
// (build_vector_trunc $src0, undef -> copy $src0
|
|
|
|
MachineInstr *Src1Def = getDefIgnoringCopies(Src1, *MRI);
|
|
|
|
if (Src1Def && Src1Def->getOpcode() == AMDGPU::G_IMPLICIT_DEF) {
|
|
|
|
MI.setDesc(TII.get(AMDGPU::COPY));
|
|
|
|
MI.RemoveOperand(2);
|
|
|
|
return RBI.constrainGenericRegister(Dst, AMDGPU::SReg_32RegClass, *MRI) &&
|
|
|
|
RBI.constrainGenericRegister(Src0, AMDGPU::SReg_32RegClass, *MRI);
|
|
|
|
}
|
|
|
|
|
|
|
|
Register ShiftSrc0;
|
|
|
|
Register ShiftSrc1;
|
|
|
|
|
|
|
|
// With multiple uses of the shift, this will duplicate the shift and
|
|
|
|
// increase register pressure.
|
|
|
|
//
|
|
|
|
// (build_vector_trunc (lshr_oneuse $src0, 16), (lshr_oneuse $src1, 16)
|
|
|
|
// => (S_PACK_HH_B32_B16 $src0, $src1)
|
|
|
|
// (build_vector_trunc $src0, (lshr_oneuse SReg_32:$src1, 16))
|
|
|
|
// => (S_PACK_LH_B32_B16 $src0, $src1)
|
|
|
|
// (build_vector_trunc $src0, $src1)
|
|
|
|
// => (S_PACK_LL_B32_B16 $src0, $src1)
|
|
|
|
|
|
|
|
bool Shift0 = mi_match(
|
2020-11-13 01:42:32 +01:00
|
|
|
Src0, *MRI, m_OneUse(m_GLShr(m_Reg(ShiftSrc0), m_SpecificICst(16))));
|
2020-02-15 23:14:17 +01:00
|
|
|
|
|
|
|
bool Shift1 = mi_match(
|
2020-11-13 01:42:32 +01:00
|
|
|
Src1, *MRI, m_OneUse(m_GLShr(m_Reg(ShiftSrc1), m_SpecificICst(16))));
|
2020-02-15 23:14:17 +01:00
|
|
|
|
|
|
|
unsigned Opc = AMDGPU::S_PACK_LL_B32_B16;
|
|
|
|
if (Shift0 && Shift1) {
|
|
|
|
Opc = AMDGPU::S_PACK_HH_B32_B16;
|
|
|
|
MI.getOperand(1).setReg(ShiftSrc0);
|
|
|
|
MI.getOperand(2).setReg(ShiftSrc1);
|
|
|
|
} else if (Shift1) {
|
|
|
|
Opc = AMDGPU::S_PACK_LH_B32_B16;
|
|
|
|
MI.getOperand(2).setReg(ShiftSrc1);
|
2020-07-18 21:30:59 +02:00
|
|
|
} else if (Shift0 && ConstSrc1 && ConstSrc1->Value == 0) {
|
2020-02-15 23:14:17 +01:00
|
|
|
// build_vector_trunc (lshr $src0, 16), 0 -> s_lshr_b32 $src0, 16
|
|
|
|
auto MIB = BuildMI(*BB, &MI, DL, TII.get(AMDGPU::S_LSHR_B32), Dst)
|
|
|
|
.addReg(ShiftSrc0)
|
|
|
|
.addImm(16);
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return constrainSelectedInstRegOperands(*MIB, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
|
|
|
MI.setDesc(TII.get(Opc));
|
|
|
|
return constrainSelectedInstRegOperands(MI, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
[globalisel] Rename G_GEP to G_PTR_ADD
Summary:
G_GEP is rather poorly named. It's a simple pointer+scalar addition and
doesn't support any of the complexities of getelementptr. I therefore
propose that we rename it. There's a G_PTR_MASK so let's follow that
convention and go with G_PTR_ADD
Reviewers: volkan, aditya_nandakumar, bogner, rovka, arsenm
Subscribers: sdardis, jvesely, wdng, nhaehnle, hiraditya, jrtc27, atanasyan, arphaman, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D69734
2019-11-01 21:18:00 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectG_PTR_ADD(MachineInstr &I) const {
|
2019-07-09 16:05:11 +02:00
|
|
|
return selectG_ADD_SUB(I);
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
}
|
|
|
|
|
2018-06-22 01:38:20 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_IMPLICIT_DEF(MachineInstr &I) const {
|
|
|
|
const MachineOperand &MO = I.getOperand(0);
|
2019-06-24 18:24:03 +02:00
|
|
|
|
|
|
|
// FIXME: Interface for getConstrainedRegClassForOperand needs work. The
|
|
|
|
// regbank check here is to know why getConstrainedRegClassForOperand failed.
|
2019-09-28 05:41:13 +02:00
|
|
|
const TargetRegisterClass *RC = TRI.getConstrainedRegClassForOperand(MO, *MRI);
|
|
|
|
if ((!RC && !MRI->getRegBankOrNull(MO.getReg())) ||
|
|
|
|
(RC && RBI.constrainGenericRegister(MO.getReg(), *RC, *MRI))) {
|
2019-06-24 18:24:03 +02:00
|
|
|
I.setDesc(TII.get(TargetOpcode::IMPLICIT_DEF));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
2018-06-22 01:38:20 +02:00
|
|
|
}
|
|
|
|
|
AMDGPU/GlobalISel: Implement select for G_INSERT
Re-commit r344310.
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D53116
llvm-svn: 355159
2019-03-01 01:50:26 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectG_INSERT(MachineInstr &I) const {
|
|
|
|
MachineBasicBlock *BB = I.getParent();
|
2019-10-07 20:43:31 +02:00
|
|
|
|
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
2019-09-16 16:26:14 +02:00
|
|
|
Register Src0Reg = I.getOperand(1).getReg();
|
|
|
|
Register Src1Reg = I.getOperand(2).getReg();
|
2019-09-28 05:41:13 +02:00
|
|
|
LLT Src1Ty = MRI->getType(Src1Reg);
|
2019-10-07 20:43:31 +02:00
|
|
|
|
|
|
|
unsigned DstSize = MRI->getType(DstReg).getSizeInBits();
|
|
|
|
unsigned InsSize = Src1Ty.getSizeInBits();
|
2019-09-16 16:26:14 +02:00
|
|
|
|
|
|
|
int64_t Offset = I.getOperand(3).getImm();
|
2020-05-20 19:26:10 +02:00
|
|
|
|
|
|
|
// FIXME: These cases should have been illegal and unnecessary to check here.
|
|
|
|
if (Offset % 32 != 0 || InsSize % 32 != 0)
|
2019-09-16 16:26:14 +02:00
|
|
|
return false;
|
|
|
|
|
2020-06-16 02:13:24 +02:00
|
|
|
// Currently not handled by getSubRegFromChannel.
|
|
|
|
if (InsSize > 128)
|
|
|
|
return false;
|
|
|
|
|
2019-10-07 21:16:26 +02:00
|
|
|
unsigned SubReg = TRI.getSubRegFromChannel(Offset / 32, InsSize / 32);
|
2019-10-07 20:43:31 +02:00
|
|
|
if (SubReg == AMDGPU::NoSubRegister)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const RegisterBank *DstBank = RBI.getRegBank(DstReg, *MRI, TRI);
|
|
|
|
const TargetRegisterClass *DstRC =
|
|
|
|
TRI.getRegClassForSizeOnBank(DstSize, *DstBank, *MRI);
|
|
|
|
if (!DstRC)
|
|
|
|
return false;
|
2019-09-16 16:26:14 +02:00
|
|
|
|
2019-10-07 20:43:31 +02:00
|
|
|
const RegisterBank *Src0Bank = RBI.getRegBank(Src0Reg, *MRI, TRI);
|
|
|
|
const RegisterBank *Src1Bank = RBI.getRegBank(Src1Reg, *MRI, TRI);
|
|
|
|
const TargetRegisterClass *Src0RC =
|
|
|
|
TRI.getRegClassForSizeOnBank(DstSize, *Src0Bank, *MRI);
|
|
|
|
const TargetRegisterClass *Src1RC =
|
|
|
|
TRI.getRegClassForSizeOnBank(InsSize, *Src1Bank, *MRI);
|
|
|
|
|
|
|
|
// Deal with weird cases where the class only partially supports the subreg
|
|
|
|
// index.
|
|
|
|
Src0RC = TRI.getSubClassWithSubReg(Src0RC, SubReg);
|
2020-02-06 14:34:18 +01:00
|
|
|
if (!Src0RC || !Src1RC)
|
2019-10-07 20:43:31 +02:00
|
|
|
return false;
|
AMDGPU/GlobalISel: Implement select for G_INSERT
Re-commit r344310.
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D53116
llvm-svn: 355159
2019-03-01 01:50:26 +01:00
|
|
|
|
2019-10-07 20:43:31 +02:00
|
|
|
if (!RBI.constrainGenericRegister(DstReg, *DstRC, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(Src0Reg, *Src0RC, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(Src1Reg, *Src1RC, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(TargetOpcode::INSERT_SUBREG), DstReg)
|
|
|
|
.addReg(Src0Reg)
|
|
|
|
.addReg(Src1Reg)
|
|
|
|
.addImm(SubReg);
|
AMDGPU/GlobalISel: Implement select for G_INSERT
Re-commit r344310.
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D53116
llvm-svn: 355159
2019-03-01 01:50:26 +01:00
|
|
|
|
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2021-04-30 15:57:44 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_SBFX_UBFX(MachineInstr &MI) const {
|
|
|
|
Register DstReg = MI.getOperand(0).getReg();
|
|
|
|
Register SrcReg = MI.getOperand(1).getReg();
|
|
|
|
Register OffsetReg = MI.getOperand(2).getReg();
|
|
|
|
Register WidthReg = MI.getOperand(3).getReg();
|
|
|
|
|
|
|
|
assert(RBI.getRegBank(DstReg, *MRI, TRI)->getID() == AMDGPU::VGPRRegBankID &&
|
|
|
|
"scalar BFX instructions are expanded in regbankselect");
|
|
|
|
assert(MRI->getType(MI.getOperand(0).getReg()).getSizeInBits() == 32 &&
|
|
|
|
"64-bit vector BFX instructions are expanded in regbankselect");
|
|
|
|
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
MachineBasicBlock *MBB = MI.getParent();
|
|
|
|
|
|
|
|
bool IsSigned = MI.getOpcode() == TargetOpcode::G_SBFX;
|
|
|
|
unsigned Opc = IsSigned ? AMDGPU::V_BFE_I32_e64 : AMDGPU::V_BFE_U32_e64;
|
|
|
|
auto MIB = BuildMI(*MBB, &MI, DL, TII.get(Opc), DstReg)
|
|
|
|
.addReg(SrcReg)
|
|
|
|
.addReg(OffsetReg)
|
|
|
|
.addReg(WidthReg);
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return constrainSelectedInstRegOperands(*MIB, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
2020-01-12 18:32:03 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectInterpP1F16(MachineInstr &MI) const {
|
|
|
|
if (STI.getLDSBankCount() != 16)
|
|
|
|
return selectImpl(MI, *CoverageInfo);
|
|
|
|
|
|
|
|
Register Dst = MI.getOperand(0).getReg();
|
|
|
|
Register Src0 = MI.getOperand(2).getReg();
|
|
|
|
Register M0Val = MI.getOperand(6).getReg();
|
|
|
|
if (!RBI.constrainGenericRegister(M0Val, AMDGPU::SReg_32RegClass, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(Dst, AMDGPU::VGPR_32RegClass, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(Src0, AMDGPU::VGPR_32RegClass, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// This requires 2 instructions. It is possible to write a pattern to support
|
|
|
|
// this, but the generated isel emitter doesn't correctly deal with multiple
|
|
|
|
// output instructions using the same physical register input. The copy to m0
|
|
|
|
// is incorrectly placed before the second instruction.
|
|
|
|
//
|
|
|
|
// TODO: Match source modifiers.
|
|
|
|
|
|
|
|
Register InterpMov = MRI->createVirtualRegister(&AMDGPU::VGPR_32RegClass);
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
MachineBasicBlock *MBB = MI.getParent();
|
|
|
|
|
|
|
|
BuildMI(*MBB, &MI, DL, TII.get(AMDGPU::COPY), AMDGPU::M0)
|
|
|
|
.addReg(M0Val);
|
|
|
|
BuildMI(*MBB, &MI, DL, TII.get(AMDGPU::V_INTERP_MOV_F32), InterpMov)
|
|
|
|
.addImm(2)
|
|
|
|
.addImm(MI.getOperand(4).getImm()) // $attr
|
|
|
|
.addImm(MI.getOperand(3).getImm()); // $attrchan
|
|
|
|
|
|
|
|
BuildMI(*MBB, &MI, DL, TII.get(AMDGPU::V_INTERP_P1LV_F16), Dst)
|
|
|
|
.addImm(0) // $src0_modifiers
|
|
|
|
.addReg(Src0) // $src0
|
|
|
|
.addImm(MI.getOperand(4).getImm()) // $attr
|
|
|
|
.addImm(MI.getOperand(3).getImm()) // $attrchan
|
|
|
|
.addImm(0) // $src2_modifiers
|
|
|
|
.addReg(InterpMov) // $src2 - 2 f16 values selected by high
|
|
|
|
.addImm(MI.getOperand(5).getImm()) // $high
|
|
|
|
.addImm(0) // $clamp
|
|
|
|
.addImm(0); // $omod
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-07-23 05:47:35 +02:00
|
|
|
// Writelane is special in that it can use SGPR and M0 (which would normally
|
|
|
|
// count as using the constant bus twice - but in this case it is allowed since
|
|
|
|
// the lane selector doesn't count as a use of the constant bus). However, it is
|
|
|
|
// still required to abide by the 1 SGPR rule. Fix this up if we might have
|
|
|
|
// multiple SGPRs.
|
|
|
|
bool AMDGPUInstructionSelector::selectWritelane(MachineInstr &MI) const {
|
|
|
|
// With a constant bus limit of at least 2, there's no issue.
|
|
|
|
if (STI.getConstantBusLimit(AMDGPU::V_WRITELANE_B32) > 1)
|
|
|
|
return selectImpl(MI, *CoverageInfo);
|
|
|
|
|
|
|
|
MachineBasicBlock *MBB = MI.getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
Register VDst = MI.getOperand(0).getReg();
|
|
|
|
Register Val = MI.getOperand(2).getReg();
|
|
|
|
Register LaneSelect = MI.getOperand(3).getReg();
|
|
|
|
Register VDstIn = MI.getOperand(4).getReg();
|
|
|
|
|
|
|
|
auto MIB = BuildMI(*MBB, &MI, DL, TII.get(AMDGPU::V_WRITELANE_B32), VDst);
|
|
|
|
|
|
|
|
Optional<ValueAndVReg> ConstSelect =
|
|
|
|
getConstantVRegValWithLookThrough(LaneSelect, *MRI, true, true);
|
|
|
|
if (ConstSelect) {
|
|
|
|
// The selector has to be an inline immediate, so we can use whatever for
|
|
|
|
// the other operands.
|
|
|
|
MIB.addReg(Val);
|
2020-11-03 15:50:17 +01:00
|
|
|
MIB.addImm(ConstSelect->Value.getSExtValue() &
|
2020-07-23 05:47:35 +02:00
|
|
|
maskTrailingOnes<uint64_t>(STI.getWavefrontSizeLog2()));
|
|
|
|
} else {
|
|
|
|
Optional<ValueAndVReg> ConstVal =
|
|
|
|
getConstantVRegValWithLookThrough(Val, *MRI, true, true);
|
|
|
|
|
|
|
|
// If the value written is an inline immediate, we can get away without a
|
|
|
|
// copy to m0.
|
2020-11-03 15:50:17 +01:00
|
|
|
if (ConstVal && AMDGPU::isInlinableLiteral32(ConstVal->Value.getSExtValue(),
|
2020-07-23 05:47:35 +02:00
|
|
|
STI.hasInv2PiInlineImm())) {
|
2020-11-03 15:50:17 +01:00
|
|
|
MIB.addImm(ConstVal->Value.getSExtValue());
|
2020-07-23 05:47:35 +02:00
|
|
|
MIB.addReg(LaneSelect);
|
|
|
|
} else {
|
|
|
|
MIB.addReg(Val);
|
|
|
|
|
|
|
|
// If the lane selector was originally in a VGPR and copied with
|
|
|
|
// readfirstlane, there's a hazard to read the same SGPR from the
|
|
|
|
// VALU. Constrain to a different SGPR to help avoid needing a nop later.
|
|
|
|
RBI.constrainGenericRegister(LaneSelect, AMDGPU::SReg_32_XM0RegClass, *MRI);
|
|
|
|
|
|
|
|
BuildMI(*MBB, *MIB, DL, TII.get(AMDGPU::COPY), AMDGPU::M0)
|
|
|
|
.addReg(LaneSelect);
|
|
|
|
MIB.addReg(AMDGPU::M0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
MIB.addReg(VDstIn);
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return constrainSelectedInstRegOperands(*MIB, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
2020-04-06 03:16:00 +02:00
|
|
|
// We need to handle this here because tablegen doesn't support matching
|
|
|
|
// instructions with multiple outputs.
|
|
|
|
bool AMDGPUInstructionSelector::selectDivScale(MachineInstr &MI) const {
|
|
|
|
Register Dst0 = MI.getOperand(0).getReg();
|
|
|
|
Register Dst1 = MI.getOperand(1).getReg();
|
|
|
|
|
|
|
|
LLT Ty = MRI->getType(Dst0);
|
|
|
|
unsigned Opc;
|
|
|
|
if (Ty == LLT::scalar(32))
|
2021-01-07 19:56:02 +01:00
|
|
|
Opc = AMDGPU::V_DIV_SCALE_F32_e64;
|
2020-04-06 03:16:00 +02:00
|
|
|
else if (Ty == LLT::scalar(64))
|
2021-01-07 19:56:02 +01:00
|
|
|
Opc = AMDGPU::V_DIV_SCALE_F64_e64;
|
2020-04-06 03:16:00 +02:00
|
|
|
else
|
|
|
|
return false;
|
|
|
|
|
2020-10-27 13:29:11 +01:00
|
|
|
// TODO: Match source modifiers.
|
|
|
|
|
2020-04-06 03:16:00 +02:00
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
MachineBasicBlock *MBB = MI.getParent();
|
|
|
|
|
|
|
|
Register Numer = MI.getOperand(3).getReg();
|
|
|
|
Register Denom = MI.getOperand(4).getReg();
|
|
|
|
unsigned ChooseDenom = MI.getOperand(5).getImm();
|
|
|
|
|
|
|
|
Register Src0 = ChooseDenom != 0 ? Numer : Denom;
|
|
|
|
|
|
|
|
auto MIB = BuildMI(*MBB, &MI, DL, TII.get(Opc), Dst0)
|
|
|
|
.addDef(Dst1)
|
2020-10-27 13:29:11 +01:00
|
|
|
.addImm(0) // $src0_modifiers
|
|
|
|
.addUse(Src0) // $src0
|
|
|
|
.addImm(0) // $src1_modifiers
|
|
|
|
.addUse(Denom) // $src1
|
|
|
|
.addImm(0) // $src2_modifiers
|
|
|
|
.addUse(Numer) // $src2
|
|
|
|
.addImm(0) // $clamp
|
|
|
|
.addImm(0); // $omod
|
2020-04-06 03:16:00 +02:00
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return constrainSelectedInstRegOperands(*MIB, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
2019-08-13 08:26:59 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_INTRINSIC(MachineInstr &I) const {
|
2019-10-02 03:02:27 +02:00
|
|
|
unsigned IntrinsicID = I.getIntrinsicID();
|
2018-06-14 21:26:37 +02:00
|
|
|
switch (IntrinsicID) {
|
2019-07-15 20:25:24 +02:00
|
|
|
case Intrinsic::amdgcn_if_break: {
|
|
|
|
MachineBasicBlock *BB = I.getParent();
|
|
|
|
|
|
|
|
// FIXME: Manually selecting to avoid dealiing with the SReg_1 trick
|
|
|
|
// SelectionDAG uses for wave32 vs wave64.
|
|
|
|
BuildMI(*BB, &I, I.getDebugLoc(), TII.get(AMDGPU::SI_IF_BREAK))
|
|
|
|
.add(I.getOperand(0))
|
|
|
|
.add(I.getOperand(2))
|
|
|
|
.add(I.getOperand(3));
|
|
|
|
|
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
|
|
|
Register Src0Reg = I.getOperand(2).getReg();
|
|
|
|
Register Src1Reg = I.getOperand(3).getReg();
|
|
|
|
|
|
|
|
I.eraseFromParent();
|
|
|
|
|
2019-10-18 20:26:37 +02:00
|
|
|
for (Register Reg : { DstReg, Src0Reg, Src1Reg })
|
|
|
|
MRI->setRegClass(Reg, TRI.getWaveMaskRegClass());
|
2019-07-15 20:25:24 +02:00
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
2020-01-12 18:32:03 +01:00
|
|
|
case Intrinsic::amdgcn_interp_p1_f16:
|
|
|
|
return selectInterpP1F16(I);
|
2020-01-17 06:22:23 +01:00
|
|
|
case Intrinsic::amdgcn_wqm:
|
|
|
|
return constrainCopyLikeIntrin(I, AMDGPU::WQM);
|
|
|
|
case Intrinsic::amdgcn_softwqm:
|
|
|
|
return constrainCopyLikeIntrin(I, AMDGPU::SOFT_WQM);
|
2021-03-02 09:38:07 +01:00
|
|
|
case Intrinsic::amdgcn_strict_wwm:
|
2020-01-17 06:22:23 +01:00
|
|
|
case Intrinsic::amdgcn_wwm:
|
2021-03-02 09:38:07 +01:00
|
|
|
return constrainCopyLikeIntrin(I, AMDGPU::STRICT_WWM);
|
2021-03-03 10:18:28 +01:00
|
|
|
case Intrinsic::amdgcn_strict_wqm:
|
|
|
|
return constrainCopyLikeIntrin(I, AMDGPU::STRICT_WQM);
|
2020-07-23 05:47:35 +02:00
|
|
|
case Intrinsic::amdgcn_writelane:
|
|
|
|
return selectWritelane(I);
|
2020-04-06 03:16:00 +02:00
|
|
|
case Intrinsic::amdgcn_div_scale:
|
|
|
|
return selectDivScale(I);
|
2020-06-30 10:57:41 +02:00
|
|
|
case Intrinsic::amdgcn_icmp:
|
|
|
|
return selectIntrinsicIcmp(I);
|
2020-07-13 11:44:18 +02:00
|
|
|
case Intrinsic::amdgcn_ballot:
|
|
|
|
return selectBallot(I);
|
2020-07-22 01:41:24 +02:00
|
|
|
case Intrinsic::amdgcn_reloc_constant:
|
|
|
|
return selectRelocConstant(I);
|
2020-07-26 21:43:48 +02:00
|
|
|
case Intrinsic::amdgcn_groupstaticsize:
|
|
|
|
return selectGroupStaticSize(I);
|
2020-07-20 03:26:02 +02:00
|
|
|
case Intrinsic::returnaddress:
|
|
|
|
return selectReturnAddress(I);
|
2019-07-02 16:52:16 +02:00
|
|
|
default:
|
2019-08-13 08:26:59 +02:00
|
|
|
return selectImpl(I, *CoverageInfo);
|
2018-06-14 21:26:37 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-07-01 15:34:26 +02:00
|
|
|
static int getV_CMPOpcode(CmpInst::Predicate P, unsigned Size) {
|
|
|
|
if (Size != 32 && Size != 64)
|
|
|
|
return -1;
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
switch (P) {
|
|
|
|
default:
|
|
|
|
llvm_unreachable("Unknown condition code!");
|
|
|
|
case CmpInst::ICMP_NE:
|
|
|
|
return Size == 32 ? AMDGPU::V_CMP_NE_U32_e64 : AMDGPU::V_CMP_NE_U64_e64;
|
|
|
|
case CmpInst::ICMP_EQ:
|
|
|
|
return Size == 32 ? AMDGPU::V_CMP_EQ_U32_e64 : AMDGPU::V_CMP_EQ_U64_e64;
|
|
|
|
case CmpInst::ICMP_SGT:
|
|
|
|
return Size == 32 ? AMDGPU::V_CMP_GT_I32_e64 : AMDGPU::V_CMP_GT_I64_e64;
|
|
|
|
case CmpInst::ICMP_SGE:
|
|
|
|
return Size == 32 ? AMDGPU::V_CMP_GE_I32_e64 : AMDGPU::V_CMP_GE_I64_e64;
|
|
|
|
case CmpInst::ICMP_SLT:
|
|
|
|
return Size == 32 ? AMDGPU::V_CMP_LT_I32_e64 : AMDGPU::V_CMP_LT_I64_e64;
|
|
|
|
case CmpInst::ICMP_SLE:
|
|
|
|
return Size == 32 ? AMDGPU::V_CMP_LE_I32_e64 : AMDGPU::V_CMP_LE_I64_e64;
|
|
|
|
case CmpInst::ICMP_UGT:
|
|
|
|
return Size == 32 ? AMDGPU::V_CMP_GT_U32_e64 : AMDGPU::V_CMP_GT_U64_e64;
|
|
|
|
case CmpInst::ICMP_UGE:
|
|
|
|
return Size == 32 ? AMDGPU::V_CMP_GE_U32_e64 : AMDGPU::V_CMP_GE_U64_e64;
|
|
|
|
case CmpInst::ICMP_ULT:
|
|
|
|
return Size == 32 ? AMDGPU::V_CMP_LT_U32_e64 : AMDGPU::V_CMP_LT_U64_e64;
|
|
|
|
case CmpInst::ICMP_ULE:
|
|
|
|
return Size == 32 ? AMDGPU::V_CMP_LE_U32_e64 : AMDGPU::V_CMP_LE_U64_e64;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-07-01 15:34:26 +02:00
|
|
|
int AMDGPUInstructionSelector::getS_CMPOpcode(CmpInst::Predicate P,
|
|
|
|
unsigned Size) const {
|
|
|
|
if (Size == 64) {
|
|
|
|
if (!STI.hasScalarCompareEq64())
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
switch (P) {
|
|
|
|
case CmpInst::ICMP_NE:
|
|
|
|
return AMDGPU::S_CMP_LG_U64;
|
|
|
|
case CmpInst::ICMP_EQ:
|
|
|
|
return AMDGPU::S_CMP_EQ_U64;
|
|
|
|
default:
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (Size != 32)
|
|
|
|
return -1;
|
|
|
|
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
switch (P) {
|
|
|
|
case CmpInst::ICMP_NE:
|
|
|
|
return AMDGPU::S_CMP_LG_U32;
|
|
|
|
case CmpInst::ICMP_EQ:
|
|
|
|
return AMDGPU::S_CMP_EQ_U32;
|
|
|
|
case CmpInst::ICMP_SGT:
|
|
|
|
return AMDGPU::S_CMP_GT_I32;
|
|
|
|
case CmpInst::ICMP_SGE:
|
|
|
|
return AMDGPU::S_CMP_GE_I32;
|
|
|
|
case CmpInst::ICMP_SLT:
|
|
|
|
return AMDGPU::S_CMP_LT_I32;
|
|
|
|
case CmpInst::ICMP_SLE:
|
|
|
|
return AMDGPU::S_CMP_LE_I32;
|
|
|
|
case CmpInst::ICMP_UGT:
|
|
|
|
return AMDGPU::S_CMP_GT_U32;
|
|
|
|
case CmpInst::ICMP_UGE:
|
|
|
|
return AMDGPU::S_CMP_GE_U32;
|
|
|
|
case CmpInst::ICMP_ULT:
|
|
|
|
return AMDGPU::S_CMP_LT_U32;
|
|
|
|
case CmpInst::ICMP_ULE:
|
|
|
|
return AMDGPU::S_CMP_LE_U32;
|
2019-07-01 15:34:26 +02:00
|
|
|
default:
|
|
|
|
llvm_unreachable("Unknown condition code!");
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AMDGPUInstructionSelector::selectG_ICMP(MachineInstr &I) const {
|
|
|
|
MachineBasicBlock *BB = I.getParent();
|
2019-07-15 21:39:31 +02:00
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
|
Apply llvm-prefer-register-over-unsigned from clang-tidy to LLVM
Summary:
This clang-tidy check is looking for unsigned integer variables whose initializer
starts with an implicit cast from llvm::Register and changes the type of the
variable to llvm::Register (dropping the llvm:: where possible).
Partial reverts in:
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
X86FixupLEAs.cpp - Some functions return unsigned and arguably should be MCRegister
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
HexagonBitSimplify.cpp - Function takes BitTracker::RegisterRef which appears to be unsigned&
MachineVerifier.cpp - Ambiguous operator==() given MCRegister and const Register
PPCFastISel.cpp - No Register::operator-=()
PeepholeOptimizer.cpp - TargetInstrInfo::optimizeLoadInstr() takes an unsigned&
MachineTraceMetrics.cpp - MachineTraceMetrics lacks a suitable constructor
Manual fixups in:
ARMFastISel.cpp - ARMEmitLoad() now takes a Register& instead of unsigned&
HexagonSplitDouble.cpp - Ternary operator was ambiguous between unsigned/Register
HexagonConstExtenders.cpp - Has a local class named Register, used llvm::Register instead of Register.
PPCFastISel.cpp - PPCEmitLoad() now takes a Register& instead of unsigned&
Depends on D65919
Reviewers: arsenm, bogner, craig.topper, RKSimon
Reviewed By: arsenm
Subscribers: RKSimon, craig.topper, lenary, aemerson, wuzish, jholewinski, MatzeB, qcolombet, dschuff, jyknight, dylanmckay, sdardis, nemanjai, jvesely, wdng, nhaehnle, sbc100, jgravelle-google, kristof.beyls, hiraditya, aheejin, kbarton, fedor.sergeev, javed.absar, asb, rbar, johnrusso, simoncook, apazos, sabuasal, niosHD, jrtc27, MaskRay, zzheng, edward-jones, atanasyan, rogfer01, MartinMosbeck, brucehoult, the_o, tpr, PkmX, jocewei, jsji, Petar.Avramovic, asbirlea, Jim, s.egerton, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D65962
llvm-svn: 369041
2019-08-15 21:22:08 +02:00
|
|
|
Register SrcReg = I.getOperand(2).getReg();
|
2019-09-28 05:41:13 +02:00
|
|
|
unsigned Size = RBI.getSizeInBits(SrcReg, *MRI, TRI);
|
2019-07-01 15:34:26 +02:00
|
|
|
|
|
|
|
auto Pred = (CmpInst::Predicate)I.getOperand(1).getPredicate();
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
|
Apply llvm-prefer-register-over-unsigned from clang-tidy to LLVM
Summary:
This clang-tidy check is looking for unsigned integer variables whose initializer
starts with an implicit cast from llvm::Register and changes the type of the
variable to llvm::Register (dropping the llvm:: where possible).
Partial reverts in:
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
X86FixupLEAs.cpp - Some functions return unsigned and arguably should be MCRegister
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
HexagonBitSimplify.cpp - Function takes BitTracker::RegisterRef which appears to be unsigned&
MachineVerifier.cpp - Ambiguous operator==() given MCRegister and const Register
PPCFastISel.cpp - No Register::operator-=()
PeepholeOptimizer.cpp - TargetInstrInfo::optimizeLoadInstr() takes an unsigned&
MachineTraceMetrics.cpp - MachineTraceMetrics lacks a suitable constructor
Manual fixups in:
ARMFastISel.cpp - ARMEmitLoad() now takes a Register& instead of unsigned&
HexagonSplitDouble.cpp - Ternary operator was ambiguous between unsigned/Register
HexagonConstExtenders.cpp - Has a local class named Register, used llvm::Register instead of Register.
PPCFastISel.cpp - PPCEmitLoad() now takes a Register& instead of unsigned&
Depends on D65919
Reviewers: arsenm, bogner, craig.topper, RKSimon
Reviewed By: arsenm
Subscribers: RKSimon, craig.topper, lenary, aemerson, wuzish, jholewinski, MatzeB, qcolombet, dschuff, jyknight, dylanmckay, sdardis, nemanjai, jvesely, wdng, nhaehnle, sbc100, jgravelle-google, kristof.beyls, hiraditya, aheejin, kbarton, fedor.sergeev, javed.absar, asb, rbar, johnrusso, simoncook, apazos, sabuasal, niosHD, jrtc27, MaskRay, zzheng, edward-jones, atanasyan, rogfer01, MartinMosbeck, brucehoult, the_o, tpr, PkmX, jocewei, jsji, Petar.Avramovic, asbirlea, Jim, s.egerton, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D65962
llvm-svn: 369041
2019-08-15 21:22:08 +02:00
|
|
|
Register CCReg = I.getOperand(0).getReg();
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
if (!isVCC(CCReg, *MRI)) {
|
2019-07-01 15:34:26 +02:00
|
|
|
int Opcode = getS_CMPOpcode(Pred, Size);
|
|
|
|
if (Opcode == -1)
|
|
|
|
return false;
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
MachineInstr *ICmp = BuildMI(*BB, &I, DL, TII.get(Opcode))
|
|
|
|
.add(I.getOperand(2))
|
|
|
|
.add(I.getOperand(3));
|
2019-06-25 15:18:11 +02:00
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::COPY), CCReg)
|
|
|
|
.addReg(AMDGPU::SCC);
|
|
|
|
bool Ret =
|
|
|
|
constrainSelectedInstRegOperands(*ICmp, TII, TRI, RBI) &&
|
2019-09-28 05:41:13 +02:00
|
|
|
RBI.constrainGenericRegister(CCReg, AMDGPU::SReg_32RegClass, *MRI);
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
I.eraseFromParent();
|
|
|
|
return Ret;
|
|
|
|
}
|
|
|
|
|
2019-07-01 15:34:26 +02:00
|
|
|
int Opcode = getV_CMPOpcode(Pred, Size);
|
|
|
|
if (Opcode == -1)
|
|
|
|
return false;
|
|
|
|
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
MachineInstr *ICmp = BuildMI(*BB, &I, DL, TII.get(Opcode),
|
|
|
|
I.getOperand(0).getReg())
|
|
|
|
.add(I.getOperand(2))
|
|
|
|
.add(I.getOperand(3));
|
|
|
|
RBI.constrainGenericRegister(ICmp->getOperand(0).getReg(),
|
2019-09-28 05:41:13 +02:00
|
|
|
*TRI.getBoolRC(), *MRI);
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
bool Ret = constrainSelectedInstRegOperands(*ICmp, TII, TRI, RBI);
|
|
|
|
I.eraseFromParent();
|
|
|
|
return Ret;
|
|
|
|
}
|
|
|
|
|
2020-06-30 10:57:41 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectIntrinsicIcmp(MachineInstr &I) const {
|
|
|
|
Register Dst = I.getOperand(0).getReg();
|
|
|
|
if (isVCC(Dst, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (MRI->getType(Dst).getSizeInBits() != STI.getWavefrontSize())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
MachineBasicBlock *BB = I.getParent();
|
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
Register SrcReg = I.getOperand(2).getReg();
|
|
|
|
unsigned Size = RBI.getSizeInBits(SrcReg, *MRI, TRI);
|
|
|
|
auto Pred = static_cast<CmpInst::Predicate>(I.getOperand(4).getImm());
|
|
|
|
|
|
|
|
int Opcode = getV_CMPOpcode(Pred, Size);
|
|
|
|
if (Opcode == -1)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
MachineInstr *ICmp = BuildMI(*BB, &I, DL, TII.get(Opcode), Dst)
|
|
|
|
.add(I.getOperand(2))
|
|
|
|
.add(I.getOperand(3));
|
|
|
|
RBI.constrainGenericRegister(ICmp->getOperand(0).getReg(), *TRI.getBoolRC(),
|
|
|
|
*MRI);
|
|
|
|
bool Ret = constrainSelectedInstRegOperands(*ICmp, TII, TRI, RBI);
|
|
|
|
I.eraseFromParent();
|
|
|
|
return Ret;
|
|
|
|
}
|
|
|
|
|
2020-07-13 11:44:18 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectBallot(MachineInstr &I) const {
|
|
|
|
MachineBasicBlock *BB = I.getParent();
|
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
|
|
|
const unsigned Size = MRI->getType(DstReg).getSizeInBits();
|
|
|
|
const bool Is64 = Size == 64;
|
|
|
|
|
|
|
|
if (Size != STI.getWavefrontSize())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
Optional<ValueAndVReg> Arg =
|
|
|
|
getConstantVRegValWithLookThrough(I.getOperand(2).getReg(), *MRI, true);
|
|
|
|
|
|
|
|
if (Arg.hasValue()) {
|
2020-11-03 15:50:17 +01:00
|
|
|
const int64_t Value = Arg.getValue().Value.getSExtValue();
|
2020-07-13 11:44:18 +02:00
|
|
|
if (Value == 0) {
|
|
|
|
unsigned Opcode = Is64 ? AMDGPU::S_MOV_B64 : AMDGPU::S_MOV_B32;
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(Opcode), DstReg).addImm(0);
|
|
|
|
} else if (Value == -1) { // all ones
|
|
|
|
Register SrcReg = Is64 ? AMDGPU::EXEC : AMDGPU::EXEC_LO;
|
2020-07-13 13:35:34 +02:00
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::COPY), DstReg).addReg(SrcReg);
|
2020-07-13 11:44:18 +02:00
|
|
|
} else
|
|
|
|
return false;
|
|
|
|
} else {
|
|
|
|
Register SrcReg = I.getOperand(2).getReg();
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::COPY), DstReg).addReg(SrcReg);
|
|
|
|
}
|
|
|
|
|
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-07-22 01:41:24 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectRelocConstant(MachineInstr &I) const {
|
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
|
|
|
const RegisterBank *DstBank = RBI.getRegBank(DstReg, *MRI, TRI);
|
|
|
|
const TargetRegisterClass *DstRC =
|
|
|
|
TRI.getRegClassForSizeOnBank(32, *DstBank, *MRI);
|
|
|
|
if (!DstRC || !RBI.constrainGenericRegister(DstReg, *DstRC, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const bool IsVALU = DstBank->getID() == AMDGPU::VGPRRegBankID;
|
|
|
|
|
|
|
|
Module *M = MF->getFunction().getParent();
|
|
|
|
const MDNode *Metadata = I.getOperand(2).getMetadata();
|
|
|
|
auto SymbolName = cast<MDString>(Metadata->getOperand(0))->getString();
|
|
|
|
auto RelocSymbol = cast<GlobalVariable>(
|
|
|
|
M->getOrInsertGlobal(SymbolName, Type::getInt32Ty(M->getContext())));
|
|
|
|
|
|
|
|
MachineBasicBlock *BB = I.getParent();
|
|
|
|
BuildMI(*BB, &I, I.getDebugLoc(),
|
|
|
|
TII.get(IsVALU ? AMDGPU::V_MOV_B32_e32 : AMDGPU::S_MOV_B32), DstReg)
|
|
|
|
.addGlobalAddress(RelocSymbol, 0, SIInstrInfo::MO_ABS32_LO);
|
|
|
|
|
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-07-26 21:43:48 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectGroupStaticSize(MachineInstr &I) const {
|
|
|
|
Triple::OSType OS = MF->getTarget().getTargetTriple().getOS();
|
|
|
|
|
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
|
|
|
const RegisterBank *DstRB = RBI.getRegBank(DstReg, *MRI, TRI);
|
|
|
|
unsigned Mov = DstRB->getID() == AMDGPU::SGPRRegBankID ?
|
|
|
|
AMDGPU::S_MOV_B32 : AMDGPU::V_MOV_B32_e32;
|
|
|
|
|
|
|
|
MachineBasicBlock *MBB = I.getParent();
|
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
|
|
|
|
auto MIB = BuildMI(*MBB, &I, DL, TII.get(Mov), DstReg);
|
|
|
|
|
|
|
|
if (OS == Triple::AMDHSA || OS == Triple::AMDPAL) {
|
|
|
|
const SIMachineFunctionInfo *MFI = MF->getInfo<SIMachineFunctionInfo>();
|
|
|
|
MIB.addImm(MFI->getLDSSize());
|
|
|
|
} else {
|
|
|
|
Module *M = MF->getFunction().getParent();
|
|
|
|
const GlobalValue *GV
|
|
|
|
= Intrinsic::getDeclaration(M, Intrinsic::amdgcn_groupstaticsize);
|
|
|
|
MIB.addGlobalAddress(GV, 0, SIInstrInfo::MO_ABS32_LO);
|
|
|
|
}
|
|
|
|
|
|
|
|
I.eraseFromParent();
|
|
|
|
return constrainSelectedInstRegOperands(*MIB, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
2020-07-20 03:26:02 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectReturnAddress(MachineInstr &I) const {
|
|
|
|
MachineBasicBlock *MBB = I.getParent();
|
|
|
|
MachineFunction &MF = *MBB->getParent();
|
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
|
|
|
|
MachineOperand &Dst = I.getOperand(0);
|
|
|
|
Register DstReg = Dst.getReg();
|
|
|
|
unsigned Depth = I.getOperand(2).getImm();
|
|
|
|
|
|
|
|
const TargetRegisterClass *RC
|
|
|
|
= TRI.getConstrainedRegClassForOperand(Dst, *MRI);
|
|
|
|
if (!RC->hasSubClassEq(&AMDGPU::SGPR_64RegClass) ||
|
|
|
|
!RBI.constrainGenericRegister(DstReg, *RC, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Check for kernel and shader functions
|
|
|
|
if (Depth != 0 ||
|
|
|
|
MF.getInfo<SIMachineFunctionInfo>()->isEntryFunction()) {
|
|
|
|
BuildMI(*MBB, &I, DL, TII.get(AMDGPU::S_MOV_B64), DstReg)
|
|
|
|
.addImm(0);
|
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
MachineFrameInfo &MFI = MF.getFrameInfo();
|
|
|
|
// There is a call to @llvm.returnaddress in this function
|
|
|
|
MFI.setReturnAddressIsTaken(true);
|
|
|
|
|
|
|
|
// Get the return address reg and mark it as an implicit live-in
|
2020-07-20 23:10:42 +02:00
|
|
|
Register ReturnAddrReg = TRI.getReturnAddressReg(MF);
|
|
|
|
Register LiveIn = getFunctionLiveInPhysReg(MF, TII, ReturnAddrReg,
|
|
|
|
AMDGPU::SReg_64RegClass);
|
2020-07-20 03:26:02 +02:00
|
|
|
BuildMI(*MBB, &I, DL, TII.get(AMDGPU::COPY), DstReg)
|
|
|
|
.addReg(LiveIn);
|
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-01-26 09:01:53 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectEndCfIntrinsic(MachineInstr &MI) const {
|
|
|
|
// FIXME: Manually selecting to avoid dealiing with the SReg_1 trick
|
|
|
|
// SelectionDAG uses for wave32 vs wave64.
|
|
|
|
MachineBasicBlock *BB = MI.getParent();
|
|
|
|
BuildMI(*BB, &MI, MI.getDebugLoc(), TII.get(AMDGPU::SI_END_CF))
|
|
|
|
.add(MI.getOperand(1));
|
|
|
|
|
|
|
|
Register Reg = MI.getOperand(1).getReg();
|
|
|
|
MI.eraseFromParent();
|
|
|
|
|
|
|
|
if (!MRI->getRegClassOrNull(Reg))
|
|
|
|
MRI->setRegClass(Reg, TRI.getWaveMaskRegClass());
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-01-05 20:26:53 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectDSOrderedIntrinsic(
|
|
|
|
MachineInstr &MI, Intrinsic::ID IntrID) const {
|
|
|
|
MachineBasicBlock *MBB = MI.getParent();
|
|
|
|
MachineFunction *MF = MBB->getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
|
|
|
|
unsigned IndexOperand = MI.getOperand(7).getImm();
|
|
|
|
bool WaveRelease = MI.getOperand(8).getImm() != 0;
|
|
|
|
bool WaveDone = MI.getOperand(9).getImm() != 0;
|
|
|
|
|
|
|
|
if (WaveDone && !WaveRelease)
|
|
|
|
report_fatal_error("ds_ordered_count: wave_done requires wave_release");
|
|
|
|
|
|
|
|
unsigned OrderedCountIndex = IndexOperand & 0x3f;
|
|
|
|
IndexOperand &= ~0x3f;
|
|
|
|
unsigned CountDw = 0;
|
|
|
|
|
|
|
|
if (STI.getGeneration() >= AMDGPUSubtarget::GFX10) {
|
|
|
|
CountDw = (IndexOperand >> 24) & 0xf;
|
|
|
|
IndexOperand &= ~(0xf << 24);
|
|
|
|
|
|
|
|
if (CountDw < 1 || CountDw > 4) {
|
|
|
|
report_fatal_error(
|
|
|
|
"ds_ordered_count: dword count must be between 1 and 4");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (IndexOperand)
|
|
|
|
report_fatal_error("ds_ordered_count: bad index operand");
|
|
|
|
|
|
|
|
unsigned Instruction = IntrID == Intrinsic::amdgcn_ds_ordered_add ? 0 : 1;
|
2020-07-26 17:30:44 +02:00
|
|
|
unsigned ShaderType = SIInstrInfo::getDSShaderTypeValue(*MF);
|
2020-01-05 20:26:53 +01:00
|
|
|
|
|
|
|
unsigned Offset0 = OrderedCountIndex << 2;
|
|
|
|
unsigned Offset1 = WaveRelease | (WaveDone << 1) | (ShaderType << 2) |
|
|
|
|
(Instruction << 4);
|
|
|
|
|
|
|
|
if (STI.getGeneration() >= AMDGPUSubtarget::GFX10)
|
|
|
|
Offset1 |= (CountDw - 1) << 6;
|
|
|
|
|
|
|
|
unsigned Offset = Offset0 | (Offset1 << 8);
|
|
|
|
|
|
|
|
Register M0Val = MI.getOperand(2).getReg();
|
|
|
|
BuildMI(*MBB, &MI, DL, TII.get(AMDGPU::COPY), AMDGPU::M0)
|
|
|
|
.addReg(M0Val);
|
|
|
|
|
|
|
|
Register DstReg = MI.getOperand(0).getReg();
|
|
|
|
Register ValReg = MI.getOperand(3).getReg();
|
|
|
|
MachineInstrBuilder DS =
|
|
|
|
BuildMI(*MBB, &MI, DL, TII.get(AMDGPU::DS_ORDERED_COUNT), DstReg)
|
|
|
|
.addReg(ValReg)
|
|
|
|
.addImm(Offset)
|
|
|
|
.cloneMemRefs(MI);
|
|
|
|
|
|
|
|
if (!RBI.constrainGenericRegister(M0Val, AMDGPU::SReg_32RegClass, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
bool Ret = constrainSelectedInstRegOperands(*DS, TII, TRI, RBI);
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return Ret;
|
|
|
|
}
|
|
|
|
|
2020-01-13 01:04:24 +01:00
|
|
|
static unsigned gwsIntrinToOpcode(unsigned IntrID) {
|
|
|
|
switch (IntrID) {
|
|
|
|
case Intrinsic::amdgcn_ds_gws_init:
|
|
|
|
return AMDGPU::DS_GWS_INIT;
|
|
|
|
case Intrinsic::amdgcn_ds_gws_barrier:
|
|
|
|
return AMDGPU::DS_GWS_BARRIER;
|
|
|
|
case Intrinsic::amdgcn_ds_gws_sema_v:
|
|
|
|
return AMDGPU::DS_GWS_SEMA_V;
|
|
|
|
case Intrinsic::amdgcn_ds_gws_sema_br:
|
|
|
|
return AMDGPU::DS_GWS_SEMA_BR;
|
|
|
|
case Intrinsic::amdgcn_ds_gws_sema_p:
|
|
|
|
return AMDGPU::DS_GWS_SEMA_P;
|
|
|
|
case Intrinsic::amdgcn_ds_gws_sema_release_all:
|
|
|
|
return AMDGPU::DS_GWS_SEMA_RELEASE_ALL;
|
|
|
|
default:
|
|
|
|
llvm_unreachable("not a gws intrinsic");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AMDGPUInstructionSelector::selectDSGWSIntrinsic(MachineInstr &MI,
|
|
|
|
Intrinsic::ID IID) const {
|
|
|
|
if (IID == Intrinsic::amdgcn_ds_gws_sema_release_all &&
|
|
|
|
!STI.hasGWSSemaReleaseAll())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// intrinsic ID, vsrc, offset
|
|
|
|
const bool HasVSrc = MI.getNumOperands() == 3;
|
|
|
|
assert(HasVSrc || MI.getNumOperands() == 2);
|
|
|
|
|
|
|
|
Register BaseOffset = MI.getOperand(HasVSrc ? 2 : 1).getReg();
|
|
|
|
const RegisterBank *OffsetRB = RBI.getRegBank(BaseOffset, *MRI, TRI);
|
|
|
|
if (OffsetRB->getID() != AMDGPU::SGPRRegBankID)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
MachineInstr *OffsetDef = getDefIgnoringCopies(BaseOffset, *MRI);
|
|
|
|
assert(OffsetDef);
|
|
|
|
|
|
|
|
unsigned ImmOffset;
|
|
|
|
|
|
|
|
MachineBasicBlock *MBB = MI.getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
|
|
|
|
MachineInstr *Readfirstlane = nullptr;
|
|
|
|
|
|
|
|
// If we legalized the VGPR input, strip out the readfirstlane to analyze the
|
|
|
|
// incoming offset, in case there's an add of a constant. We'll have to put it
|
|
|
|
// back later.
|
|
|
|
if (OffsetDef->getOpcode() == AMDGPU::V_READFIRSTLANE_B32) {
|
|
|
|
Readfirstlane = OffsetDef;
|
|
|
|
BaseOffset = OffsetDef->getOperand(1).getReg();
|
|
|
|
OffsetDef = getDefIgnoringCopies(BaseOffset, *MRI);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (OffsetDef->getOpcode() == AMDGPU::G_CONSTANT) {
|
|
|
|
// If we have a constant offset, try to use the 0 in m0 as the base.
|
|
|
|
// TODO: Look into changing the default m0 initialization value. If the
|
|
|
|
// default -1 only set the low 16-bits, we could leave it as-is and add 1 to
|
|
|
|
// the immediate offset.
|
|
|
|
|
|
|
|
ImmOffset = OffsetDef->getOperand(1).getCImm()->getZExtValue();
|
|
|
|
BuildMI(*MBB, &MI, DL, TII.get(AMDGPU::S_MOV_B32), AMDGPU::M0)
|
|
|
|
.addImm(0);
|
|
|
|
} else {
|
2020-11-09 11:46:38 +01:00
|
|
|
std::tie(BaseOffset, ImmOffset) =
|
|
|
|
AMDGPU::getBaseWithConstantOffset(*MRI, BaseOffset);
|
2020-01-13 01:04:24 +01:00
|
|
|
|
|
|
|
if (Readfirstlane) {
|
|
|
|
// We have the constant offset now, so put the readfirstlane back on the
|
|
|
|
// variable component.
|
|
|
|
if (!RBI.constrainGenericRegister(BaseOffset, AMDGPU::VGPR_32RegClass, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
Readfirstlane->getOperand(1).setReg(BaseOffset);
|
|
|
|
BaseOffset = Readfirstlane->getOperand(0).getReg();
|
|
|
|
} else {
|
|
|
|
if (!RBI.constrainGenericRegister(BaseOffset,
|
|
|
|
AMDGPU::SReg_32RegClass, *MRI))
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
Register M0Base = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
BuildMI(*MBB, &MI, DL, TII.get(AMDGPU::S_LSHL_B32), M0Base)
|
|
|
|
.addReg(BaseOffset)
|
|
|
|
.addImm(16);
|
|
|
|
|
|
|
|
BuildMI(*MBB, &MI, DL, TII.get(AMDGPU::COPY), AMDGPU::M0)
|
|
|
|
.addReg(M0Base);
|
|
|
|
}
|
|
|
|
|
|
|
|
// The resource id offset is computed as (<isa opaque base> + M0[21:16] +
|
|
|
|
// offset field) % 64. Some versions of the programming guide omit the m0
|
|
|
|
// part, or claim it's from offset 0.
|
|
|
|
auto MIB = BuildMI(*MBB, &MI, DL, TII.get(gwsIntrinToOpcode(IID)));
|
|
|
|
|
|
|
|
if (HasVSrc) {
|
|
|
|
Register VSrc = MI.getOperand(1).getReg();
|
2021-05-26 01:37:01 +02:00
|
|
|
|
|
|
|
if (STI.needsAlignedVGPRs()) {
|
|
|
|
// Add implicit aligned super-reg to force alignment on the data operand.
|
|
|
|
Register Undef = MRI->createVirtualRegister(&AMDGPU::VGPR_32RegClass);
|
|
|
|
BuildMI(*MBB, &*MIB, DL, TII.get(AMDGPU::IMPLICIT_DEF), Undef);
|
|
|
|
Register NewVR =
|
|
|
|
MRI->createVirtualRegister(&AMDGPU::VReg_64_Align2RegClass);
|
|
|
|
BuildMI(*MBB, &*MIB, DL, TII.get(AMDGPU::REG_SEQUENCE), NewVR)
|
|
|
|
.addReg(VSrc, 0, MI.getOperand(1).getSubReg())
|
|
|
|
.addImm(AMDGPU::sub0)
|
|
|
|
.addReg(Undef)
|
|
|
|
.addImm(AMDGPU::sub1);
|
|
|
|
MIB.addReg(NewVR, 0, AMDGPU::sub0);
|
|
|
|
MIB.addReg(NewVR, RegState::Implicit);
|
|
|
|
} else {
|
|
|
|
MIB.addReg(VSrc);
|
|
|
|
}
|
|
|
|
|
2020-01-13 01:04:24 +01:00
|
|
|
if (!RBI.constrainGenericRegister(VSrc, AMDGPU::VGPR_32RegClass, *MRI))
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
MIB.addImm(ImmOffset)
|
|
|
|
.cloneMemRefs(MI);
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-01-17 04:27:31 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectDSAppendConsume(MachineInstr &MI,
|
|
|
|
bool IsAppend) const {
|
|
|
|
Register PtrBase = MI.getOperand(2).getReg();
|
|
|
|
LLT PtrTy = MRI->getType(PtrBase);
|
|
|
|
bool IsGDS = PtrTy.getAddressSpace() == AMDGPUAS::REGION_ADDRESS;
|
|
|
|
|
|
|
|
unsigned Offset;
|
|
|
|
std::tie(PtrBase, Offset) = selectDS1Addr1OffsetImpl(MI.getOperand(2));
|
|
|
|
|
|
|
|
// TODO: Should this try to look through readfirstlane like GWS?
|
2020-11-02 14:05:15 +01:00
|
|
|
if (!isDSOffsetLegal(PtrBase, Offset)) {
|
2020-01-17 04:27:31 +01:00
|
|
|
PtrBase = MI.getOperand(2).getReg();
|
|
|
|
Offset = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
MachineBasicBlock *MBB = MI.getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
const unsigned Opc = IsAppend ? AMDGPU::DS_APPEND : AMDGPU::DS_CONSUME;
|
|
|
|
|
|
|
|
BuildMI(*MBB, &MI, DL, TII.get(AMDGPU::COPY), AMDGPU::M0)
|
|
|
|
.addReg(PtrBase);
|
2020-07-25 17:56:33 +02:00
|
|
|
if (!RBI.constrainGenericRegister(PtrBase, AMDGPU::SReg_32RegClass, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
auto MIB = BuildMI(*MBB, &MI, DL, TII.get(Opc), MI.getOperand(0).getReg())
|
2020-01-17 04:27:31 +01:00
|
|
|
.addImm(Offset)
|
|
|
|
.addImm(IsGDS ? -1 : 0)
|
|
|
|
.cloneMemRefs(MI);
|
|
|
|
MI.eraseFromParent();
|
2020-07-25 17:56:33 +02:00
|
|
|
return constrainSelectedInstRegOperands(*MIB, TII, TRI, RBI);
|
2020-01-17 04:27:31 +01:00
|
|
|
}
|
|
|
|
|
2020-08-26 12:41:39 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectSBarrier(MachineInstr &MI) const {
|
|
|
|
if (TM.getOptLevel() > CodeGenOpt::None) {
|
|
|
|
unsigned WGSize = STI.getFlatWorkGroupSizes(MF->getFunction()).second;
|
|
|
|
if (WGSize <= STI.getWavefrontSize()) {
|
|
|
|
MachineBasicBlock *MBB = MI.getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
BuildMI(*MBB, &MI, DL, TII.get(AMDGPU::WAVE_BARRIER));
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return selectImpl(MI, *CoverageInfo);
|
|
|
|
}
|
|
|
|
|
2020-02-08 05:37:54 +01:00
|
|
|
static bool parseTexFail(uint64_t TexFailCtrl, bool &TFE, bool &LWE,
|
|
|
|
bool &IsTexFail) {
|
|
|
|
if (TexFailCtrl)
|
|
|
|
IsTexFail = true;
|
|
|
|
|
|
|
|
TFE = (TexFailCtrl & 0x1) ? 1 : 0;
|
|
|
|
TexFailCtrl &= ~(uint64_t)0x1;
|
|
|
|
LWE = (TexFailCtrl & 0x2) ? 1 : 0;
|
|
|
|
TexFailCtrl &= ~(uint64_t)0x2;
|
|
|
|
|
|
|
|
return TexFailCtrl == 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AMDGPUInstructionSelector::selectImageIntrinsic(
|
|
|
|
MachineInstr &MI, const AMDGPU::ImageDimIntrinsicInfo *Intr) const {
|
|
|
|
MachineBasicBlock *MBB = MI.getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
|
|
|
|
const AMDGPU::MIMGBaseOpcodeInfo *BaseOpcode =
|
|
|
|
AMDGPU::getMIMGBaseOpcodeInfo(Intr->BaseOpcode);
|
|
|
|
|
|
|
|
const AMDGPU::MIMGDimInfo *DimInfo = AMDGPU::getMIMGDimInfo(Intr->Dim);
|
|
|
|
const AMDGPU::MIMGLZMappingInfo *LZMappingInfo =
|
|
|
|
AMDGPU::getMIMGLZMappingInfo(Intr->BaseOpcode);
|
|
|
|
const AMDGPU::MIMGMIPMappingInfo *MIPMappingInfo =
|
|
|
|
AMDGPU::getMIMGMIPMappingInfo(Intr->BaseOpcode);
|
|
|
|
unsigned IntrOpcode = Intr->BaseOpcode;
|
2020-11-25 12:51:23 +01:00
|
|
|
const bool IsGFX10Plus = AMDGPU::isGFX10Plus(STI);
|
2020-02-08 05:37:54 +01:00
|
|
|
|
2020-09-30 14:15:24 +02:00
|
|
|
const unsigned ArgOffset = MI.getNumExplicitDefs() + 1;
|
2020-02-08 05:37:54 +01:00
|
|
|
|
2020-02-09 01:08:34 +01:00
|
|
|
Register VDataIn, VDataOut;
|
2020-02-08 05:37:54 +01:00
|
|
|
LLT VDataTy;
|
|
|
|
int NumVDataDwords = -1;
|
|
|
|
bool IsD16 = false;
|
|
|
|
|
|
|
|
bool Unorm;
|
2020-09-30 14:15:24 +02:00
|
|
|
if (!BaseOpcode->Sampler)
|
2020-02-08 05:37:54 +01:00
|
|
|
Unorm = true;
|
2020-09-30 14:15:24 +02:00
|
|
|
else
|
|
|
|
Unorm = MI.getOperand(ArgOffset + Intr->UnormIndex).getImm() != 0;
|
2020-02-08 05:37:54 +01:00
|
|
|
|
|
|
|
bool TFE;
|
|
|
|
bool LWE;
|
|
|
|
bool IsTexFail = false;
|
2020-09-30 14:15:24 +02:00
|
|
|
if (!parseTexFail(MI.getOperand(ArgOffset + Intr->TexFailCtrlIndex).getImm(),
|
|
|
|
TFE, LWE, IsTexFail))
|
2020-02-08 05:37:54 +01:00
|
|
|
return false;
|
|
|
|
|
2020-09-30 14:15:24 +02:00
|
|
|
const int Flags = MI.getOperand(ArgOffset + Intr->NumArgs).getImm();
|
2020-03-25 08:38:22 +01:00
|
|
|
const bool IsA16 = (Flags & 1) != 0;
|
|
|
|
const bool IsG16 = (Flags & 2) != 0;
|
|
|
|
|
2021-05-07 12:43:29 +02:00
|
|
|
// A16 implies 16 bit gradients if subtarget doesn't support G16
|
|
|
|
if (IsA16 && !STI.hasG16() && !IsG16)
|
2020-03-25 08:38:22 +01:00
|
|
|
return false;
|
|
|
|
|
2020-02-08 05:37:54 +01:00
|
|
|
unsigned DMask = 0;
|
|
|
|
unsigned DMaskLanes = 0;
|
|
|
|
|
|
|
|
if (BaseOpcode->Atomic) {
|
2020-02-09 01:08:34 +01:00
|
|
|
VDataOut = MI.getOperand(0).getReg();
|
|
|
|
VDataIn = MI.getOperand(2).getReg();
|
|
|
|
LLT Ty = MRI->getType(VDataIn);
|
|
|
|
|
|
|
|
// Be careful to allow atomic swap on 16-bit element vectors.
|
|
|
|
const bool Is64Bit = BaseOpcode->AtomicX2 ?
|
|
|
|
Ty.getSizeInBits() == 128 :
|
|
|
|
Ty.getSizeInBits() == 64;
|
|
|
|
|
|
|
|
if (BaseOpcode->AtomicX2) {
|
|
|
|
assert(MI.getOperand(3).getReg() == AMDGPU::NoRegister);
|
|
|
|
|
|
|
|
DMask = Is64Bit ? 0xf : 0x3;
|
|
|
|
NumVDataDwords = Is64Bit ? 4 : 2;
|
|
|
|
} else {
|
|
|
|
DMask = Is64Bit ? 0x3 : 0x1;
|
|
|
|
NumVDataDwords = Is64Bit ? 2 : 1;
|
|
|
|
}
|
2020-02-08 05:37:54 +01:00
|
|
|
} else {
|
2020-09-30 14:15:24 +02:00
|
|
|
DMask = MI.getOperand(ArgOffset + Intr->DMaskIndex).getImm();
|
2020-02-08 05:37:54 +01:00
|
|
|
DMaskLanes = BaseOpcode->Gather4 ? 4 : countPopulation(DMask);
|
|
|
|
|
[AMDGPU] Implement hardware bug workaround for image instructions
Summary:
This implements a workaround for a hardware bug in gfx8 and gfx9,
where register usage is not estimated correctly for image_store and
image_gather4 instructions when D16 is used.
Change-Id: I4e30744da6796acac53a9b5ad37ac1c2035c8899
Subscribers: arsenm, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, dstuttard, tpr, t-tye, hiraditya, kerbowa, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D81172
2020-04-03 23:37:51 +02:00
|
|
|
// One memoperand is mandatory, except for getresinfo.
|
|
|
|
// FIXME: Check this in verifier.
|
|
|
|
if (!MI.memoperands_empty()) {
|
|
|
|
const MachineMemOperand *MMO = *MI.memoperands_begin();
|
|
|
|
|
|
|
|
// Infer d16 from the memory size, as the register type will be mangled by
|
|
|
|
// unpacked subtargets, or by TFE.
|
|
|
|
IsD16 = ((8 * MMO->getSize()) / DMaskLanes) < 32;
|
|
|
|
}
|
|
|
|
|
2020-02-08 05:37:54 +01:00
|
|
|
if (BaseOpcode->Store) {
|
2020-02-09 01:08:34 +01:00
|
|
|
VDataIn = MI.getOperand(1).getReg();
|
|
|
|
VDataTy = MRI->getType(VDataIn);
|
2020-02-08 05:37:54 +01:00
|
|
|
NumVDataDwords = (VDataTy.getSizeInBits() + 31) / 32;
|
|
|
|
} else {
|
2020-02-09 01:08:34 +01:00
|
|
|
VDataOut = MI.getOperand(0).getReg();
|
|
|
|
VDataTy = MRI->getType(VDataOut);
|
2020-02-08 05:37:54 +01:00
|
|
|
NumVDataDwords = DMaskLanes;
|
|
|
|
|
[AMDGPU] Implement hardware bug workaround for image instructions
Summary:
This implements a workaround for a hardware bug in gfx8 and gfx9,
where register usage is not estimated correctly for image_store and
image_gather4 instructions when D16 is used.
Change-Id: I4e30744da6796acac53a9b5ad37ac1c2035c8899
Subscribers: arsenm, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, dstuttard, tpr, t-tye, hiraditya, kerbowa, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D81172
2020-04-03 23:37:51 +02:00
|
|
|
if (IsD16 && !STI.hasUnpackedD16VMem())
|
|
|
|
NumVDataDwords = (DMaskLanes + 1) / 2;
|
2020-02-08 05:37:54 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Optimize _L to _LZ when _L is zero
|
|
|
|
if (LZMappingInfo) {
|
|
|
|
// The legalizer replaced the register with an immediate 0 if we need to
|
|
|
|
// change the opcode.
|
2020-09-30 14:15:24 +02:00
|
|
|
const MachineOperand &Lod = MI.getOperand(ArgOffset + Intr->LodIndex);
|
2020-02-08 05:37:54 +01:00
|
|
|
if (Lod.isImm()) {
|
|
|
|
assert(Lod.getImm() == 0);
|
|
|
|
IntrOpcode = LZMappingInfo->LZ; // set new opcode to _lz variant of _l
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Optimize _mip away, when 'lod' is zero
|
|
|
|
if (MIPMappingInfo) {
|
2020-09-30 14:15:24 +02:00
|
|
|
const MachineOperand &Lod = MI.getOperand(ArgOffset + Intr->MipIndex);
|
2020-02-08 05:37:54 +01:00
|
|
|
if (Lod.isImm()) {
|
|
|
|
assert(Lod.getImm() == 0);
|
|
|
|
IntrOpcode = MIPMappingInfo->NONMIP; // set new opcode to variant without _mip
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-03-25 08:38:22 +01:00
|
|
|
// Set G16 opcode
|
|
|
|
if (IsG16 && !IsA16) {
|
|
|
|
const AMDGPU::MIMGG16MappingInfo *G16MappingInfo =
|
|
|
|
AMDGPU::getMIMGG16MappingInfo(Intr->BaseOpcode);
|
|
|
|
assert(G16MappingInfo);
|
|
|
|
IntrOpcode = G16MappingInfo->G16; // set opcode to variant with _g16
|
|
|
|
}
|
|
|
|
|
2020-02-08 05:37:54 +01:00
|
|
|
// TODO: Check this in verifier.
|
2020-04-01 15:36:33 +02:00
|
|
|
assert((!IsTexFail || DMaskLanes >= 1) && "should have legalized this");
|
2020-02-08 05:37:54 +01:00
|
|
|
|
2021-02-09 01:36:10 +01:00
|
|
|
unsigned CPol = MI.getOperand(ArgOffset + Intr->CachePolicyIndex).getImm();
|
|
|
|
if (BaseOpcode->Atomic)
|
|
|
|
CPol |= AMDGPU::CPol::GLC; // TODO no-return optimization
|
|
|
|
if (CPol & ~AMDGPU::CPol::ALL)
|
|
|
|
return false;
|
2020-02-08 05:37:54 +01:00
|
|
|
|
|
|
|
int NumVAddrRegs = 0;
|
|
|
|
int NumVAddrDwords = 0;
|
2020-09-30 14:15:24 +02:00
|
|
|
for (unsigned I = Intr->VAddrStart; I < Intr->VAddrEnd; I++) {
|
2020-02-08 05:37:54 +01:00
|
|
|
// Skip the $noregs and 0s inserted during legalization.
|
2020-09-30 14:15:24 +02:00
|
|
|
MachineOperand &AddrOp = MI.getOperand(ArgOffset + I);
|
2020-02-08 05:37:54 +01:00
|
|
|
if (!AddrOp.isReg())
|
|
|
|
continue; // XXX - Break?
|
|
|
|
|
|
|
|
Register Addr = AddrOp.getReg();
|
|
|
|
if (!Addr)
|
|
|
|
break;
|
|
|
|
|
|
|
|
++NumVAddrRegs;
|
|
|
|
NumVAddrDwords += (MRI->getType(Addr).getSizeInBits() + 31) / 32;
|
|
|
|
}
|
|
|
|
|
|
|
|
// The legalizer preprocessed the intrinsic arguments. If we aren't using
|
|
|
|
// NSA, these should have beeen packed into a single value in the first
|
|
|
|
// address register
|
|
|
|
const bool UseNSA = NumVAddrRegs != 1 && NumVAddrDwords == NumVAddrRegs;
|
|
|
|
if (UseNSA && !STI.hasFeature(AMDGPU::FeatureNSAEncoding)) {
|
|
|
|
LLVM_DEBUG(dbgs() << "Trying to use NSA on non-NSA target\n");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (IsTexFail)
|
|
|
|
++NumVDataDwords;
|
|
|
|
|
|
|
|
int Opcode = -1;
|
2020-11-25 12:51:23 +01:00
|
|
|
if (IsGFX10Plus) {
|
2020-02-08 05:37:54 +01:00
|
|
|
Opcode = AMDGPU::getMIMGOpcode(IntrOpcode,
|
|
|
|
UseNSA ? AMDGPU::MIMGEncGfx10NSA
|
|
|
|
: AMDGPU::MIMGEncGfx10Default,
|
|
|
|
NumVDataDwords, NumVAddrDwords);
|
|
|
|
} else {
|
|
|
|
if (STI.getGeneration() >= AMDGPUSubtarget::VOLCANIC_ISLANDS)
|
|
|
|
Opcode = AMDGPU::getMIMGOpcode(IntrOpcode, AMDGPU::MIMGEncGfx8,
|
|
|
|
NumVDataDwords, NumVAddrDwords);
|
|
|
|
if (Opcode == -1)
|
|
|
|
Opcode = AMDGPU::getMIMGOpcode(IntrOpcode, AMDGPU::MIMGEncGfx6,
|
|
|
|
NumVDataDwords, NumVAddrDwords);
|
|
|
|
}
|
|
|
|
assert(Opcode != -1);
|
|
|
|
|
|
|
|
auto MIB = BuildMI(*MBB, &MI, DL, TII.get(Opcode))
|
|
|
|
.cloneMemRefs(MI);
|
|
|
|
|
2020-02-09 01:08:34 +01:00
|
|
|
if (VDataOut) {
|
|
|
|
if (BaseOpcode->AtomicX2) {
|
|
|
|
const bool Is64 = MRI->getType(VDataOut).getSizeInBits() == 64;
|
|
|
|
|
|
|
|
Register TmpReg = MRI->createVirtualRegister(
|
|
|
|
Is64 ? &AMDGPU::VReg_128RegClass : &AMDGPU::VReg_64RegClass);
|
|
|
|
unsigned SubReg = Is64 ? AMDGPU::sub0_sub1 : AMDGPU::sub0;
|
|
|
|
|
|
|
|
MIB.addDef(TmpReg);
|
2021-04-28 13:11:44 +02:00
|
|
|
if (!MRI->use_empty(VDataOut)) {
|
|
|
|
BuildMI(*MBB, &MI, DL, TII.get(AMDGPU::COPY), VDataOut)
|
|
|
|
.addReg(TmpReg, RegState::Kill, SubReg);
|
|
|
|
}
|
2020-02-09 01:08:34 +01:00
|
|
|
|
|
|
|
} else {
|
|
|
|
MIB.addDef(VDataOut); // vdata output
|
|
|
|
}
|
|
|
|
}
|
2020-02-08 05:37:54 +01:00
|
|
|
|
2020-02-09 01:08:34 +01:00
|
|
|
if (VDataIn)
|
|
|
|
MIB.addReg(VDataIn); // vdata input
|
2020-02-08 05:37:54 +01:00
|
|
|
|
2020-09-30 14:15:24 +02:00
|
|
|
for (int I = 0; I != NumVAddrRegs; ++I) {
|
|
|
|
MachineOperand &SrcOp = MI.getOperand(ArgOffset + Intr->VAddrStart + I);
|
2020-02-08 05:37:54 +01:00
|
|
|
if (SrcOp.isReg()) {
|
|
|
|
assert(SrcOp.getReg() != 0);
|
|
|
|
MIB.addReg(SrcOp.getReg());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-09-30 14:15:24 +02:00
|
|
|
MIB.addReg(MI.getOperand(ArgOffset + Intr->RsrcIndex).getReg());
|
2020-02-08 05:37:54 +01:00
|
|
|
if (BaseOpcode->Sampler)
|
2020-09-30 14:15:24 +02:00
|
|
|
MIB.addReg(MI.getOperand(ArgOffset + Intr->SampIndex).getReg());
|
2020-02-08 05:37:54 +01:00
|
|
|
|
|
|
|
MIB.addImm(DMask); // dmask
|
|
|
|
|
2020-11-25 12:51:23 +01:00
|
|
|
if (IsGFX10Plus)
|
2020-02-08 05:37:54 +01:00
|
|
|
MIB.addImm(DimInfo->Encoding);
|
|
|
|
MIB.addImm(Unorm);
|
|
|
|
|
2021-02-09 01:36:10 +01:00
|
|
|
MIB.addImm(CPol);
|
2020-02-08 05:37:54 +01:00
|
|
|
MIB.addImm(IsA16 && // a16 or r128
|
|
|
|
STI.hasFeature(AMDGPU::FeatureR128A16) ? -1 : 0);
|
2020-11-25 12:51:23 +01:00
|
|
|
if (IsGFX10Plus)
|
2020-02-08 05:37:54 +01:00
|
|
|
MIB.addImm(IsA16 ? -1 : 0);
|
|
|
|
|
|
|
|
MIB.addImm(TFE); // tfe
|
|
|
|
MIB.addImm(LWE); // lwe
|
2020-11-25 12:51:23 +01:00
|
|
|
if (!IsGFX10Plus)
|
2020-02-08 05:37:54 +01:00
|
|
|
MIB.addImm(DimInfo->DA ? -1 : 0);
|
|
|
|
if (BaseOpcode->HasD16)
|
|
|
|
MIB.addImm(IsD16 ? -1 : 0);
|
|
|
|
|
2021-03-31 17:45:21 +02:00
|
|
|
if (IsTexFail) {
|
|
|
|
// An image load instruction with TFE/LWE only conditionally writes to its
|
|
|
|
// result registers. Initialize them to zero so that we always get well
|
|
|
|
// defined result values.
|
|
|
|
assert(VDataOut && !VDataIn);
|
|
|
|
Register Tied = MRI->cloneVirtualRegister(VDataOut);
|
|
|
|
Register Zero = MRI->createVirtualRegister(&AMDGPU::VGPR_32RegClass);
|
|
|
|
BuildMI(*MBB, *MIB, DL, TII.get(AMDGPU::V_MOV_B32_e32), Zero)
|
|
|
|
.addImm(0);
|
|
|
|
auto Parts = TRI.getRegSplitParts(MRI->getRegClass(Tied), 4);
|
|
|
|
if (STI.usePRTStrictNull()) {
|
|
|
|
// With enable-prt-strict-null enabled, initialize all result registers to
|
|
|
|
// zero.
|
|
|
|
auto RegSeq =
|
|
|
|
BuildMI(*MBB, *MIB, DL, TII.get(AMDGPU::REG_SEQUENCE), Tied);
|
|
|
|
for (auto Sub : Parts)
|
|
|
|
RegSeq.addReg(Zero).addImm(Sub);
|
|
|
|
} else {
|
|
|
|
// With enable-prt-strict-null disabled, only initialize the extra TFE/LWE
|
|
|
|
// result register.
|
|
|
|
Register Undef = MRI->createVirtualRegister(&AMDGPU::VGPR_32RegClass);
|
|
|
|
BuildMI(*MBB, *MIB, DL, TII.get(AMDGPU::IMPLICIT_DEF), Undef);
|
|
|
|
auto RegSeq =
|
|
|
|
BuildMI(*MBB, *MIB, DL, TII.get(AMDGPU::REG_SEQUENCE), Tied);
|
|
|
|
for (auto Sub : Parts.drop_back(1))
|
|
|
|
RegSeq.addReg(Undef).addImm(Sub);
|
|
|
|
RegSeq.addReg(Zero).addImm(Parts.back());
|
|
|
|
}
|
|
|
|
MIB.addReg(Tied, RegState::Implicit);
|
|
|
|
MIB->tieOperands(0, MIB->getNumOperands() - 1);
|
|
|
|
}
|
|
|
|
|
2020-02-08 05:37:54 +01:00
|
|
|
MI.eraseFromParent();
|
|
|
|
return constrainSelectedInstRegOperands(*MIB, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
2018-07-13 23:05:14 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_INTRINSIC_W_SIDE_EFFECTS(
|
2019-08-13 08:26:59 +02:00
|
|
|
MachineInstr &I) const {
|
2019-09-13 06:12:12 +02:00
|
|
|
unsigned IntrinsicID = I.getIntrinsicID();
|
2018-07-13 23:05:14 +02:00
|
|
|
switch (IntrinsicID) {
|
2020-01-26 09:01:53 +01:00
|
|
|
case Intrinsic::amdgcn_end_cf:
|
|
|
|
return selectEndCfIntrinsic(I);
|
2020-01-05 20:26:53 +01:00
|
|
|
case Intrinsic::amdgcn_ds_ordered_add:
|
|
|
|
case Intrinsic::amdgcn_ds_ordered_swap:
|
|
|
|
return selectDSOrderedIntrinsic(I, IntrinsicID);
|
2020-01-13 01:04:24 +01:00
|
|
|
case Intrinsic::amdgcn_ds_gws_init:
|
|
|
|
case Intrinsic::amdgcn_ds_gws_barrier:
|
|
|
|
case Intrinsic::amdgcn_ds_gws_sema_v:
|
|
|
|
case Intrinsic::amdgcn_ds_gws_sema_br:
|
|
|
|
case Intrinsic::amdgcn_ds_gws_sema_p:
|
|
|
|
case Intrinsic::amdgcn_ds_gws_sema_release_all:
|
|
|
|
return selectDSGWSIntrinsic(I, IntrinsicID);
|
2020-01-17 04:27:31 +01:00
|
|
|
case Intrinsic::amdgcn_ds_append:
|
|
|
|
return selectDSAppendConsume(I, true);
|
|
|
|
case Intrinsic::amdgcn_ds_consume:
|
|
|
|
return selectDSAppendConsume(I, false);
|
2020-08-26 12:41:39 +02:00
|
|
|
case Intrinsic::amdgcn_s_barrier:
|
|
|
|
return selectSBarrier(I);
|
2020-09-11 00:10:52 +02:00
|
|
|
case Intrinsic::amdgcn_global_atomic_fadd:
|
2021-03-01 18:24:57 +01:00
|
|
|
return selectGlobalAtomicFadd(I, I.getOperand(2), I.getOperand(3));
|
2020-02-08 05:37:54 +01:00
|
|
|
default: {
|
2019-08-13 08:26:59 +02:00
|
|
|
return selectImpl(I, *CoverageInfo);
|
2018-07-13 23:05:14 +02:00
|
|
|
}
|
2020-02-08 05:37:54 +01:00
|
|
|
}
|
2018-07-13 23:05:14 +02:00
|
|
|
}
|
|
|
|
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_SELECT(MachineInstr &I) const {
|
2020-01-24 19:11:23 +01:00
|
|
|
if (selectImpl(I, *CoverageInfo))
|
|
|
|
return true;
|
|
|
|
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
MachineBasicBlock *BB = I.getParent();
|
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
|
Apply llvm-prefer-register-over-unsigned from clang-tidy to LLVM
Summary:
This clang-tidy check is looking for unsigned integer variables whose initializer
starts with an implicit cast from llvm::Register and changes the type of the
variable to llvm::Register (dropping the llvm:: where possible).
Partial reverts in:
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
X86FixupLEAs.cpp - Some functions return unsigned and arguably should be MCRegister
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
HexagonBitSimplify.cpp - Function takes BitTracker::RegisterRef which appears to be unsigned&
MachineVerifier.cpp - Ambiguous operator==() given MCRegister and const Register
PPCFastISel.cpp - No Register::operator-=()
PeepholeOptimizer.cpp - TargetInstrInfo::optimizeLoadInstr() takes an unsigned&
MachineTraceMetrics.cpp - MachineTraceMetrics lacks a suitable constructor
Manual fixups in:
ARMFastISel.cpp - ARMEmitLoad() now takes a Register& instead of unsigned&
HexagonSplitDouble.cpp - Ternary operator was ambiguous between unsigned/Register
HexagonConstExtenders.cpp - Has a local class named Register, used llvm::Register instead of Register.
PPCFastISel.cpp - PPCEmitLoad() now takes a Register& instead of unsigned&
Depends on D65919
Reviewers: arsenm, bogner, craig.topper, RKSimon
Reviewed By: arsenm
Subscribers: RKSimon, craig.topper, lenary, aemerson, wuzish, jholewinski, MatzeB, qcolombet, dschuff, jyknight, dylanmckay, sdardis, nemanjai, jvesely, wdng, nhaehnle, sbc100, jgravelle-google, kristof.beyls, hiraditya, aheejin, kbarton, fedor.sergeev, javed.absar, asb, rbar, johnrusso, simoncook, apazos, sabuasal, niosHD, jrtc27, MaskRay, zzheng, edward-jones, atanasyan, rogfer01, MartinMosbeck, brucehoult, the_o, tpr, PkmX, jocewei, jsji, Petar.Avramovic, asbirlea, Jim, s.egerton, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D65962
llvm-svn: 369041
2019-08-15 21:22:08 +02:00
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
2019-09-28 05:41:13 +02:00
|
|
|
unsigned Size = RBI.getSizeInBits(DstReg, *MRI, TRI);
|
2019-07-01 17:42:47 +02:00
|
|
|
assert(Size <= 32 || Size == 64);
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
const MachineOperand &CCOp = I.getOperand(1);
|
Apply llvm-prefer-register-over-unsigned from clang-tidy to LLVM
Summary:
This clang-tidy check is looking for unsigned integer variables whose initializer
starts with an implicit cast from llvm::Register and changes the type of the
variable to llvm::Register (dropping the llvm:: where possible).
Partial reverts in:
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
X86FixupLEAs.cpp - Some functions return unsigned and arguably should be MCRegister
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
HexagonBitSimplify.cpp - Function takes BitTracker::RegisterRef which appears to be unsigned&
MachineVerifier.cpp - Ambiguous operator==() given MCRegister and const Register
PPCFastISel.cpp - No Register::operator-=()
PeepholeOptimizer.cpp - TargetInstrInfo::optimizeLoadInstr() takes an unsigned&
MachineTraceMetrics.cpp - MachineTraceMetrics lacks a suitable constructor
Manual fixups in:
ARMFastISel.cpp - ARMEmitLoad() now takes a Register& instead of unsigned&
HexagonSplitDouble.cpp - Ternary operator was ambiguous between unsigned/Register
HexagonConstExtenders.cpp - Has a local class named Register, used llvm::Register instead of Register.
PPCFastISel.cpp - PPCEmitLoad() now takes a Register& instead of unsigned&
Depends on D65919
Reviewers: arsenm, bogner, craig.topper, RKSimon
Reviewed By: arsenm
Subscribers: RKSimon, craig.topper, lenary, aemerson, wuzish, jholewinski, MatzeB, qcolombet, dschuff, jyknight, dylanmckay, sdardis, nemanjai, jvesely, wdng, nhaehnle, sbc100, jgravelle-google, kristof.beyls, hiraditya, aheejin, kbarton, fedor.sergeev, javed.absar, asb, rbar, johnrusso, simoncook, apazos, sabuasal, niosHD, jrtc27, MaskRay, zzheng, edward-jones, atanasyan, rogfer01, MartinMosbeck, brucehoult, the_o, tpr, PkmX, jocewei, jsji, Petar.Avramovic, asbirlea, Jim, s.egerton, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D65962
llvm-svn: 369041
2019-08-15 21:22:08 +02:00
|
|
|
Register CCReg = CCOp.getReg();
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
if (!isVCC(CCReg, *MRI)) {
|
2019-07-01 17:42:47 +02:00
|
|
|
unsigned SelectOpcode = Size == 64 ? AMDGPU::S_CSELECT_B64 :
|
|
|
|
AMDGPU::S_CSELECT_B32;
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
MachineInstr *CopySCC = BuildMI(*BB, &I, DL, TII.get(AMDGPU::COPY), AMDGPU::SCC)
|
|
|
|
.addReg(CCReg);
|
|
|
|
|
|
|
|
// The generic constrainSelectedInstRegOperands doesn't work for the scc register
|
|
|
|
// bank, because it does not cover the register class that we used to represent
|
|
|
|
// for it. So we need to manually set the register class here.
|
2019-09-28 05:41:13 +02:00
|
|
|
if (!MRI->getRegClassOrNull(CCReg))
|
|
|
|
MRI->setRegClass(CCReg, TRI.getConstrainedRegClassForOperand(CCOp, *MRI));
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
MachineInstr *Select = BuildMI(*BB, &I, DL, TII.get(SelectOpcode), DstReg)
|
|
|
|
.add(I.getOperand(2))
|
|
|
|
.add(I.getOperand(3));
|
|
|
|
|
|
|
|
bool Ret = constrainSelectedInstRegOperands(*Select, TII, TRI, RBI) |
|
|
|
|
constrainSelectedInstRegOperands(*CopySCC, TII, TRI, RBI);
|
|
|
|
I.eraseFromParent();
|
|
|
|
return Ret;
|
|
|
|
}
|
|
|
|
|
2019-07-01 17:42:47 +02:00
|
|
|
// Wide VGPR select should have been split in RegBankSelect.
|
|
|
|
if (Size > 32)
|
|
|
|
return false;
|
|
|
|
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
MachineInstr *Select =
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::V_CNDMASK_B32_e64), DstReg)
|
|
|
|
.addImm(0)
|
|
|
|
.add(I.getOperand(3))
|
|
|
|
.addImm(0)
|
|
|
|
.add(I.getOperand(2))
|
|
|
|
.add(I.getOperand(1));
|
|
|
|
|
|
|
|
bool Ret = constrainSelectedInstRegOperands(*Select, TII, TRI, RBI);
|
|
|
|
I.eraseFromParent();
|
|
|
|
return Ret;
|
|
|
|
}
|
|
|
|
|
2019-06-24 20:02:18 +02:00
|
|
|
static int sizeToSubRegIndex(unsigned Size) {
|
|
|
|
switch (Size) {
|
|
|
|
case 32:
|
|
|
|
return AMDGPU::sub0;
|
|
|
|
case 64:
|
|
|
|
return AMDGPU::sub0_sub1;
|
|
|
|
case 96:
|
|
|
|
return AMDGPU::sub0_sub1_sub2;
|
|
|
|
case 128:
|
|
|
|
return AMDGPU::sub0_sub1_sub2_sub3;
|
|
|
|
case 256:
|
|
|
|
return AMDGPU::sub0_sub1_sub2_sub3_sub4_sub5_sub6_sub7;
|
|
|
|
default:
|
|
|
|
if (Size < 32)
|
|
|
|
return AMDGPU::sub0;
|
|
|
|
if (Size > 256)
|
|
|
|
return -1;
|
|
|
|
return sizeToSubRegIndex(PowerOf2Ceil(Size));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AMDGPUInstructionSelector::selectG_TRUNC(MachineInstr &I) const {
|
Apply llvm-prefer-register-over-unsigned from clang-tidy to LLVM
Summary:
This clang-tidy check is looking for unsigned integer variables whose initializer
starts with an implicit cast from llvm::Register and changes the type of the
variable to llvm::Register (dropping the llvm:: where possible).
Partial reverts in:
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
X86FixupLEAs.cpp - Some functions return unsigned and arguably should be MCRegister
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
HexagonBitSimplify.cpp - Function takes BitTracker::RegisterRef which appears to be unsigned&
MachineVerifier.cpp - Ambiguous operator==() given MCRegister and const Register
PPCFastISel.cpp - No Register::operator-=()
PeepholeOptimizer.cpp - TargetInstrInfo::optimizeLoadInstr() takes an unsigned&
MachineTraceMetrics.cpp - MachineTraceMetrics lacks a suitable constructor
Manual fixups in:
ARMFastISel.cpp - ARMEmitLoad() now takes a Register& instead of unsigned&
HexagonSplitDouble.cpp - Ternary operator was ambiguous between unsigned/Register
HexagonConstExtenders.cpp - Has a local class named Register, used llvm::Register instead of Register.
PPCFastISel.cpp - PPCEmitLoad() now takes a Register& instead of unsigned&
Depends on D65919
Reviewers: arsenm, bogner, craig.topper, RKSimon
Reviewed By: arsenm
Subscribers: RKSimon, craig.topper, lenary, aemerson, wuzish, jholewinski, MatzeB, qcolombet, dschuff, jyknight, dylanmckay, sdardis, nemanjai, jvesely, wdng, nhaehnle, sbc100, jgravelle-google, kristof.beyls, hiraditya, aheejin, kbarton, fedor.sergeev, javed.absar, asb, rbar, johnrusso, simoncook, apazos, sabuasal, niosHD, jrtc27, MaskRay, zzheng, edward-jones, atanasyan, rogfer01, MartinMosbeck, brucehoult, the_o, tpr, PkmX, jocewei, jsji, Petar.Avramovic, asbirlea, Jim, s.egerton, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D65962
llvm-svn: 369041
2019-08-15 21:22:08 +02:00
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
|
|
|
Register SrcReg = I.getOperand(1).getReg();
|
2019-09-28 05:41:13 +02:00
|
|
|
const LLT DstTy = MRI->getType(DstReg);
|
|
|
|
const LLT SrcTy = MRI->getType(SrcReg);
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
const LLT S1 = LLT::scalar(1);
|
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
const RegisterBank *SrcRB = RBI.getRegBank(SrcReg, *MRI, TRI);
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
const RegisterBank *DstRB;
|
|
|
|
if (DstTy == S1) {
|
|
|
|
// This is a special case. We don't treat s1 for legalization artifacts as
|
|
|
|
// vcc booleans.
|
|
|
|
DstRB = SrcRB;
|
|
|
|
} else {
|
|
|
|
DstRB = RBI.getRegBank(DstReg, *MRI, TRI);
|
|
|
|
if (SrcRB != DstRB)
|
|
|
|
return false;
|
|
|
|
}
|
2019-06-24 20:02:18 +02:00
|
|
|
|
2020-02-16 01:18:47 +01:00
|
|
|
const bool IsVALU = DstRB->getID() == AMDGPU::VGPRRegBankID;
|
|
|
|
|
2019-06-24 20:02:18 +02:00
|
|
|
unsigned DstSize = DstTy.getSizeInBits();
|
|
|
|
unsigned SrcSize = SrcTy.getSizeInBits();
|
|
|
|
|
|
|
|
const TargetRegisterClass *SrcRC
|
2019-09-28 05:41:13 +02:00
|
|
|
= TRI.getRegClassForSizeOnBank(SrcSize, *SrcRB, *MRI);
|
2019-06-24 20:02:18 +02:00
|
|
|
const TargetRegisterClass *DstRC
|
2019-09-28 05:41:13 +02:00
|
|
|
= TRI.getRegClassForSizeOnBank(DstSize, *DstRB, *MRI);
|
2020-04-15 20:37:31 +02:00
|
|
|
if (!SrcRC || !DstRC)
|
|
|
|
return false;
|
2019-06-24 20:02:18 +02:00
|
|
|
|
2020-02-16 01:18:47 +01:00
|
|
|
if (!RBI.constrainGenericRegister(SrcReg, *SrcRC, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(DstReg, *DstRC, *MRI)) {
|
|
|
|
LLVM_DEBUG(dbgs() << "Failed to constrain G_TRUNC\n");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2021-06-24 10:58:21 +02:00
|
|
|
if (DstTy == LLT::fixed_vector(2, 16) && SrcTy == LLT::fixed_vector(2, 32)) {
|
2020-02-16 01:18:47 +01:00
|
|
|
MachineBasicBlock *MBB = I.getParent();
|
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
|
|
|
|
Register LoReg = MRI->createVirtualRegister(DstRC);
|
|
|
|
Register HiReg = MRI->createVirtualRegister(DstRC);
|
|
|
|
BuildMI(*MBB, I, DL, TII.get(AMDGPU::COPY), LoReg)
|
|
|
|
.addReg(SrcReg, 0, AMDGPU::sub0);
|
|
|
|
BuildMI(*MBB, I, DL, TII.get(AMDGPU::COPY), HiReg)
|
|
|
|
.addReg(SrcReg, 0, AMDGPU::sub1);
|
|
|
|
|
|
|
|
if (IsVALU && STI.hasSDWA()) {
|
|
|
|
// Write the low 16-bits of the high element into the high 16-bits of the
|
|
|
|
// low element.
|
|
|
|
MachineInstr *MovSDWA =
|
|
|
|
BuildMI(*MBB, I, DL, TII.get(AMDGPU::V_MOV_B32_sdwa), DstReg)
|
|
|
|
.addImm(0) // $src0_modifiers
|
|
|
|
.addReg(HiReg) // $src0
|
|
|
|
.addImm(0) // $clamp
|
|
|
|
.addImm(AMDGPU::SDWA::WORD_1) // $dst_sel
|
|
|
|
.addImm(AMDGPU::SDWA::UNUSED_PRESERVE) // $dst_unused
|
|
|
|
.addImm(AMDGPU::SDWA::WORD_0) // $src0_sel
|
|
|
|
.addReg(LoReg, RegState::Implicit);
|
|
|
|
MovSDWA->tieOperands(0, MovSDWA->getNumOperands() - 1);
|
|
|
|
} else {
|
|
|
|
Register TmpReg0 = MRI->createVirtualRegister(DstRC);
|
|
|
|
Register TmpReg1 = MRI->createVirtualRegister(DstRC);
|
|
|
|
Register ImmReg = MRI->createVirtualRegister(DstRC);
|
|
|
|
if (IsVALU) {
|
|
|
|
BuildMI(*MBB, I, DL, TII.get(AMDGPU::V_LSHLREV_B32_e64), TmpReg0)
|
|
|
|
.addImm(16)
|
|
|
|
.addReg(HiReg);
|
|
|
|
} else {
|
|
|
|
BuildMI(*MBB, I, DL, TII.get(AMDGPU::S_LSHL_B32), TmpReg0)
|
|
|
|
.addReg(HiReg)
|
|
|
|
.addImm(16);
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned MovOpc = IsVALU ? AMDGPU::V_MOV_B32_e32 : AMDGPU::S_MOV_B32;
|
|
|
|
unsigned AndOpc = IsVALU ? AMDGPU::V_AND_B32_e64 : AMDGPU::S_AND_B32;
|
|
|
|
unsigned OrOpc = IsVALU ? AMDGPU::V_OR_B32_e64 : AMDGPU::S_OR_B32;
|
|
|
|
|
|
|
|
BuildMI(*MBB, I, DL, TII.get(MovOpc), ImmReg)
|
|
|
|
.addImm(0xffff);
|
|
|
|
BuildMI(*MBB, I, DL, TII.get(AndOpc), TmpReg1)
|
|
|
|
.addReg(LoReg)
|
|
|
|
.addReg(ImmReg);
|
|
|
|
BuildMI(*MBB, I, DL, TII.get(OrOpc), DstReg)
|
|
|
|
.addReg(TmpReg0)
|
|
|
|
.addReg(TmpReg1);
|
|
|
|
}
|
|
|
|
|
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!DstTy.isScalar())
|
|
|
|
return false;
|
|
|
|
|
2019-06-24 20:02:18 +02:00
|
|
|
if (SrcSize > 32) {
|
|
|
|
int SubRegIdx = sizeToSubRegIndex(DstSize);
|
|
|
|
if (SubRegIdx == -1)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Deal with weird cases where the class only partially supports the subreg
|
|
|
|
// index.
|
2020-02-16 01:18:47 +01:00
|
|
|
const TargetRegisterClass *SrcWithSubRC
|
|
|
|
= TRI.getSubClassWithSubReg(SrcRC, SubRegIdx);
|
|
|
|
if (!SrcWithSubRC)
|
2019-06-24 20:02:18 +02:00
|
|
|
return false;
|
|
|
|
|
2020-02-16 01:18:47 +01:00
|
|
|
if (SrcWithSubRC != SrcRC) {
|
|
|
|
if (!RBI.constrainGenericRegister(SrcReg, *SrcWithSubRC, *MRI))
|
|
|
|
return false;
|
|
|
|
}
|
2019-06-24 20:02:18 +02:00
|
|
|
|
2020-02-16 01:18:47 +01:00
|
|
|
I.getOperand(1).setSubReg(SubRegIdx);
|
2019-06-24 20:02:18 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
I.setDesc(TII.get(TargetOpcode::COPY));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-07-01 15:22:06 +02:00
|
|
|
/// \returns true if a bitmask for \p Size bits will be an inline immediate.
|
|
|
|
static bool shouldUseAndMask(unsigned Size, unsigned &Mask) {
|
|
|
|
Mask = maskTrailingOnes<unsigned>(Size);
|
|
|
|
int SignedMask = static_cast<int>(Mask);
|
|
|
|
return SignedMask >= -16 && SignedMask <= 64;
|
|
|
|
}
|
|
|
|
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
// Like RegisterBankInfo::getRegBank, but don't assume vcc for s1.
|
|
|
|
const RegisterBank *AMDGPUInstructionSelector::getArtifactRegBank(
|
|
|
|
Register Reg, const MachineRegisterInfo &MRI,
|
|
|
|
const TargetRegisterInfo &TRI) const {
|
|
|
|
const RegClassOrRegBank &RegClassOrBank = MRI.getRegClassOrRegBank(Reg);
|
|
|
|
if (auto *RB = RegClassOrBank.dyn_cast<const RegisterBank *>())
|
|
|
|
return RB;
|
|
|
|
|
|
|
|
// Ignore the type, since we don't use vcc in artifacts.
|
|
|
|
if (auto *RC = RegClassOrBank.dyn_cast<const TargetRegisterClass *>())
|
|
|
|
return &RBI.getRegBankFromRegClass(*RC, LLT());
|
|
|
|
return nullptr;
|
|
|
|
}
|
|
|
|
|
2019-06-25 15:18:11 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_SZA_EXT(MachineInstr &I) const {
|
2020-01-25 01:38:53 +01:00
|
|
|
bool InReg = I.getOpcode() == AMDGPU::G_SEXT_INREG;
|
|
|
|
bool Signed = I.getOpcode() == AMDGPU::G_SEXT || InReg;
|
2019-06-25 15:18:11 +02:00
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
MachineBasicBlock &MBB = *I.getParent();
|
Apply llvm-prefer-register-over-unsigned from clang-tidy to LLVM
Summary:
This clang-tidy check is looking for unsigned integer variables whose initializer
starts with an implicit cast from llvm::Register and changes the type of the
variable to llvm::Register (dropping the llvm:: where possible).
Partial reverts in:
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
X86FixupLEAs.cpp - Some functions return unsigned and arguably should be MCRegister
X86FrameLowering.cpp - Some functions return unsigned and arguably should be MCRegister
HexagonBitSimplify.cpp - Function takes BitTracker::RegisterRef which appears to be unsigned&
MachineVerifier.cpp - Ambiguous operator==() given MCRegister and const Register
PPCFastISel.cpp - No Register::operator-=()
PeepholeOptimizer.cpp - TargetInstrInfo::optimizeLoadInstr() takes an unsigned&
MachineTraceMetrics.cpp - MachineTraceMetrics lacks a suitable constructor
Manual fixups in:
ARMFastISel.cpp - ARMEmitLoad() now takes a Register& instead of unsigned&
HexagonSplitDouble.cpp - Ternary operator was ambiguous between unsigned/Register
HexagonConstExtenders.cpp - Has a local class named Register, used llvm::Register instead of Register.
PPCFastISel.cpp - PPCEmitLoad() now takes a Register& instead of unsigned&
Depends on D65919
Reviewers: arsenm, bogner, craig.topper, RKSimon
Reviewed By: arsenm
Subscribers: RKSimon, craig.topper, lenary, aemerson, wuzish, jholewinski, MatzeB, qcolombet, dschuff, jyknight, dylanmckay, sdardis, nemanjai, jvesely, wdng, nhaehnle, sbc100, jgravelle-google, kristof.beyls, hiraditya, aheejin, kbarton, fedor.sergeev, javed.absar, asb, rbar, johnrusso, simoncook, apazos, sabuasal, niosHD, jrtc27, MaskRay, zzheng, edward-jones, atanasyan, rogfer01, MartinMosbeck, brucehoult, the_o, tpr, PkmX, jocewei, jsji, Petar.Avramovic, asbirlea, Jim, s.egerton, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D65962
llvm-svn: 369041
2019-08-15 21:22:08 +02:00
|
|
|
const Register DstReg = I.getOperand(0).getReg();
|
|
|
|
const Register SrcReg = I.getOperand(1).getReg();
|
2019-06-25 15:18:11 +02:00
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
const LLT DstTy = MRI->getType(DstReg);
|
|
|
|
const LLT SrcTy = MRI->getType(SrcReg);
|
2020-01-25 01:38:53 +01:00
|
|
|
const unsigned SrcSize = I.getOpcode() == AMDGPU::G_SEXT_INREG ?
|
|
|
|
I.getOperand(2).getImm() : SrcTy.getSizeInBits();
|
2019-06-25 15:18:11 +02:00
|
|
|
const unsigned DstSize = DstTy.getSizeInBits();
|
|
|
|
if (!DstTy.isScalar())
|
|
|
|
return false;
|
|
|
|
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
// Artifact casts should never use vcc.
|
|
|
|
const RegisterBank *SrcBank = getArtifactRegBank(SrcReg, *MRI, TRI);
|
|
|
|
|
2020-07-23 03:07:03 +02:00
|
|
|
// FIXME: This should probably be illegal and split earlier.
|
|
|
|
if (I.getOpcode() == AMDGPU::G_ANYEXT) {
|
|
|
|
if (DstSize <= 32)
|
|
|
|
return selectCOPY(I);
|
|
|
|
|
|
|
|
const TargetRegisterClass *SrcRC =
|
|
|
|
TRI.getRegClassForTypeOnBank(SrcTy, *SrcBank, *MRI);
|
|
|
|
const RegisterBank *DstBank = RBI.getRegBank(DstReg, *MRI, TRI);
|
|
|
|
const TargetRegisterClass *DstRC =
|
|
|
|
TRI.getRegClassForSizeOnBank(DstSize, *DstBank, *MRI);
|
|
|
|
|
|
|
|
Register UndefReg = MRI->createVirtualRegister(SrcRC);
|
|
|
|
BuildMI(MBB, I, DL, TII.get(AMDGPU::IMPLICIT_DEF), UndefReg);
|
|
|
|
BuildMI(MBB, I, DL, TII.get(AMDGPU::REG_SEQUENCE), DstReg)
|
|
|
|
.addReg(SrcReg)
|
|
|
|
.addImm(AMDGPU::sub0)
|
|
|
|
.addReg(UndefReg)
|
|
|
|
.addImm(AMDGPU::sub1);
|
|
|
|
I.eraseFromParent();
|
|
|
|
|
|
|
|
return RBI.constrainGenericRegister(DstReg, *DstRC, *MRI) &&
|
|
|
|
RBI.constrainGenericRegister(SrcReg, *SrcRC, *MRI);
|
|
|
|
}
|
|
|
|
|
2019-06-25 15:18:11 +02:00
|
|
|
if (SrcBank->getID() == AMDGPU::VGPRRegBankID && DstSize <= 32) {
|
|
|
|
// 64-bit should have been split up in RegBankSelect
|
2019-07-01 15:22:06 +02:00
|
|
|
|
|
|
|
// Try to use an and with a mask if it will save code size.
|
|
|
|
unsigned Mask;
|
|
|
|
if (!Signed && shouldUseAndMask(SrcSize, Mask)) {
|
|
|
|
MachineInstr *ExtI =
|
|
|
|
BuildMI(MBB, I, DL, TII.get(AMDGPU::V_AND_B32_e32), DstReg)
|
|
|
|
.addImm(Mask)
|
|
|
|
.addReg(SrcReg);
|
2019-07-24 18:05:53 +02:00
|
|
|
I.eraseFromParent();
|
2019-07-01 15:22:06 +02:00
|
|
|
return constrainSelectedInstRegOperands(*ExtI, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
2021-01-07 19:56:02 +01:00
|
|
|
const unsigned BFE = Signed ? AMDGPU::V_BFE_I32_e64 : AMDGPU::V_BFE_U32_e64;
|
2019-06-25 15:18:11 +02:00
|
|
|
MachineInstr *ExtI =
|
|
|
|
BuildMI(MBB, I, DL, TII.get(BFE), DstReg)
|
|
|
|
.addReg(SrcReg)
|
|
|
|
.addImm(0) // Offset
|
|
|
|
.addImm(SrcSize); // Width
|
2019-07-24 18:05:53 +02:00
|
|
|
I.eraseFromParent();
|
2019-06-25 15:18:11 +02:00
|
|
|
return constrainSelectedInstRegOperands(*ExtI, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (SrcBank->getID() == AMDGPU::SGPRRegBankID && DstSize <= 64) {
|
2020-01-25 01:38:53 +01:00
|
|
|
const TargetRegisterClass &SrcRC = InReg && DstSize > 32 ?
|
|
|
|
AMDGPU::SReg_64RegClass : AMDGPU::SReg_32RegClass;
|
|
|
|
if (!RBI.constrainGenericRegister(SrcReg, SrcRC, *MRI))
|
2019-06-25 15:18:11 +02:00
|
|
|
return false;
|
|
|
|
|
|
|
|
if (Signed && DstSize == 32 && (SrcSize == 8 || SrcSize == 16)) {
|
|
|
|
const unsigned SextOpc = SrcSize == 8 ?
|
|
|
|
AMDGPU::S_SEXT_I32_I8 : AMDGPU::S_SEXT_I32_I16;
|
|
|
|
BuildMI(MBB, I, DL, TII.get(SextOpc), DstReg)
|
|
|
|
.addReg(SrcReg);
|
2019-07-24 18:05:53 +02:00
|
|
|
I.eraseFromParent();
|
2019-09-28 05:41:13 +02:00
|
|
|
return RBI.constrainGenericRegister(DstReg, AMDGPU::SReg_32RegClass, *MRI);
|
2019-06-25 15:18:11 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
const unsigned BFE64 = Signed ? AMDGPU::S_BFE_I64 : AMDGPU::S_BFE_U64;
|
|
|
|
const unsigned BFE32 = Signed ? AMDGPU::S_BFE_I32 : AMDGPU::S_BFE_U32;
|
|
|
|
|
|
|
|
// Scalar BFE is encoded as S1[5:0] = offset, S1[22:16]= width.
|
2020-01-25 01:38:53 +01:00
|
|
|
if (DstSize > 32 && (SrcSize <= 32 || InReg)) {
|
2019-06-25 15:18:11 +02:00
|
|
|
// We need a 64-bit register source, but the high bits don't matter.
|
2019-09-28 05:41:13 +02:00
|
|
|
Register ExtReg = MRI->createVirtualRegister(&AMDGPU::SReg_64RegClass);
|
|
|
|
Register UndefReg = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
2020-01-25 01:38:53 +01:00
|
|
|
unsigned SubReg = InReg ? AMDGPU::sub0 : 0;
|
|
|
|
|
2019-06-25 15:18:11 +02:00
|
|
|
BuildMI(MBB, I, DL, TII.get(AMDGPU::IMPLICIT_DEF), UndefReg);
|
|
|
|
BuildMI(MBB, I, DL, TII.get(AMDGPU::REG_SEQUENCE), ExtReg)
|
2020-01-25 01:38:53 +01:00
|
|
|
.addReg(SrcReg, 0, SubReg)
|
2019-06-25 15:18:11 +02:00
|
|
|
.addImm(AMDGPU::sub0)
|
|
|
|
.addReg(UndefReg)
|
|
|
|
.addImm(AMDGPU::sub1);
|
|
|
|
|
|
|
|
BuildMI(MBB, I, DL, TII.get(BFE64), DstReg)
|
|
|
|
.addReg(ExtReg)
|
|
|
|
.addImm(SrcSize << 16);
|
|
|
|
|
2019-07-24 18:05:53 +02:00
|
|
|
I.eraseFromParent();
|
2019-09-28 05:41:13 +02:00
|
|
|
return RBI.constrainGenericRegister(DstReg, AMDGPU::SReg_64RegClass, *MRI);
|
2019-06-25 15:18:11 +02:00
|
|
|
}
|
|
|
|
|
2019-07-01 15:22:06 +02:00
|
|
|
unsigned Mask;
|
|
|
|
if (!Signed && shouldUseAndMask(SrcSize, Mask)) {
|
|
|
|
BuildMI(MBB, I, DL, TII.get(AMDGPU::S_AND_B32), DstReg)
|
|
|
|
.addReg(SrcReg)
|
|
|
|
.addImm(Mask);
|
|
|
|
} else {
|
|
|
|
BuildMI(MBB, I, DL, TII.get(BFE32), DstReg)
|
|
|
|
.addReg(SrcReg)
|
|
|
|
.addImm(SrcSize << 16);
|
|
|
|
}
|
|
|
|
|
2019-07-24 18:05:53 +02:00
|
|
|
I.eraseFromParent();
|
2019-09-28 05:41:13 +02:00
|
|
|
return RBI.constrainGenericRegister(DstReg, AMDGPU::SReg_32RegClass, *MRI);
|
2019-06-25 15:18:11 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectG_CONSTANT(MachineInstr &I) const {
|
|
|
|
MachineBasicBlock *BB = I.getParent();
|
2018-05-15 19:57:09 +02:00
|
|
|
MachineOperand &ImmOp = I.getOperand(1);
|
2020-07-25 19:21:31 +02:00
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
|
|
|
unsigned Size = MRI->getType(DstReg).getSizeInBits();
|
2018-05-15 19:57:09 +02:00
|
|
|
|
|
|
|
// The AMDGPU backend only supports Imm operands and not CImm or FPImm.
|
|
|
|
if (ImmOp.isFPImm()) {
|
|
|
|
const APInt &Imm = ImmOp.getFPImm()->getValueAPF().bitcastToAPInt();
|
|
|
|
ImmOp.ChangeToImmediate(Imm.getZExtValue());
|
|
|
|
} else if (ImmOp.isCImm()) {
|
2020-07-25 20:37:29 +02:00
|
|
|
ImmOp.ChangeToImmediate(ImmOp.getCImm()->getSExtValue());
|
2020-07-25 19:21:31 +02:00
|
|
|
} else {
|
|
|
|
llvm_unreachable("Not supported by g_constants");
|
2018-05-15 19:57:09 +02:00
|
|
|
}
|
|
|
|
|
2020-07-25 19:21:31 +02:00
|
|
|
const RegisterBank *DstRB = RBI.getRegBank(DstReg, *MRI, TRI);
|
|
|
|
const bool IsSgpr = DstRB->getID() == AMDGPU::SGPRRegBankID;
|
|
|
|
|
|
|
|
unsigned Opcode;
|
|
|
|
if (DstRB->getID() == AMDGPU::VCCRegBankID) {
|
|
|
|
Opcode = STI.isWave32() ? AMDGPU::S_MOV_B32 : AMDGPU::S_MOV_B64;
|
2018-05-15 19:57:09 +02:00
|
|
|
} else {
|
2020-07-25 19:21:31 +02:00
|
|
|
Opcode = IsSgpr ? AMDGPU::S_MOV_B32 : AMDGPU::V_MOV_B32_e32;
|
2018-05-15 19:57:09 +02:00
|
|
|
|
2020-07-25 19:21:31 +02:00
|
|
|
// We should never produce s1 values on banks other than VCC. If the user of
|
|
|
|
// this already constrained the register, we may incorrectly think it's VCC
|
|
|
|
// if it wasn't originally.
|
|
|
|
if (Size == 1)
|
|
|
|
return false;
|
|
|
|
}
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
|
2020-07-25 19:21:31 +02:00
|
|
|
if (Size != 64) {
|
2018-05-15 19:57:09 +02:00
|
|
|
I.setDesc(TII.get(Opcode));
|
|
|
|
I.addImplicitDefUseOperands(*MF);
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
return constrainSelectedInstRegOperands(I, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
2019-10-07 21:07:19 +02:00
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
|
|
|
|
APInt Imm(Size, I.getOperand(1).getImm());
|
|
|
|
|
|
|
|
MachineInstr *ResInst;
|
|
|
|
if (IsSgpr && TII.isInlineConstant(Imm)) {
|
|
|
|
ResInst = BuildMI(*BB, &I, DL, TII.get(AMDGPU::S_MOV_B64), DstReg)
|
|
|
|
.addImm(I.getOperand(1).getImm());
|
|
|
|
} else {
|
|
|
|
const TargetRegisterClass *RC = IsSgpr ?
|
2019-10-18 20:26:37 +02:00
|
|
|
&AMDGPU::SReg_32RegClass : &AMDGPU::VGPR_32RegClass;
|
2019-10-07 21:07:19 +02:00
|
|
|
Register LoReg = MRI->createVirtualRegister(RC);
|
|
|
|
Register HiReg = MRI->createVirtualRegister(RC);
|
|
|
|
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(Opcode), LoReg)
|
|
|
|
.addImm(Imm.trunc(32).getZExtValue());
|
|
|
|
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(Opcode), HiReg)
|
|
|
|
.addImm(Imm.ashr(32).getZExtValue());
|
|
|
|
|
|
|
|
ResInst = BuildMI(*BB, &I, DL, TII.get(AMDGPU::REG_SEQUENCE), DstReg)
|
|
|
|
.addReg(LoReg)
|
|
|
|
.addImm(AMDGPU::sub0)
|
|
|
|
.addReg(HiReg)
|
|
|
|
.addImm(AMDGPU::sub1);
|
|
|
|
}
|
2018-05-15 19:57:09 +02:00
|
|
|
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
// We can't call constrainSelectedInstRegOperands here, because it doesn't
|
|
|
|
// work for target independent opcodes
|
|
|
|
I.eraseFromParent();
|
2018-05-15 19:57:09 +02:00
|
|
|
const TargetRegisterClass *DstRC =
|
2019-10-07 21:07:19 +02:00
|
|
|
TRI.getConstrainedRegClassForOperand(ResInst->getOperand(0), *MRI);
|
2018-05-15 19:57:09 +02:00
|
|
|
if (!DstRC)
|
|
|
|
return true;
|
2019-09-28 05:41:13 +02:00
|
|
|
return RBI.constrainGenericRegister(DstReg, *DstRC, *MRI);
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
}
|
|
|
|
|
2020-01-26 01:34:22 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectG_FNEG(MachineInstr &MI) const {
|
|
|
|
// Only manually handle the f64 SGPR case.
|
|
|
|
//
|
|
|
|
// FIXME: This is a workaround for 2.5 different tablegen problems. Because
|
|
|
|
// the bit ops theoretically have a second result due to the implicit def of
|
|
|
|
// SCC, the GlobalISelEmitter is overly conservative and rejects it. Fixing
|
|
|
|
// that is easy by disabling the check. The result works, but uses a
|
|
|
|
// nonsensical sreg32orlds_and_sreg_1 regclass.
|
|
|
|
//
|
|
|
|
// The DAG emitter is more problematic, and incorrectly adds both S_XOR_B32 to
|
|
|
|
// the variadic REG_SEQUENCE operands.
|
|
|
|
|
|
|
|
Register Dst = MI.getOperand(0).getReg();
|
|
|
|
const RegisterBank *DstRB = RBI.getRegBank(Dst, *MRI, TRI);
|
|
|
|
if (DstRB->getID() != AMDGPU::SGPRRegBankID ||
|
|
|
|
MRI->getType(Dst) != LLT::scalar(64))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
Register Src = MI.getOperand(1).getReg();
|
|
|
|
MachineInstr *Fabs = getOpcodeDef(TargetOpcode::G_FABS, Src, *MRI);
|
|
|
|
if (Fabs)
|
|
|
|
Src = Fabs->getOperand(1).getReg();
|
|
|
|
|
|
|
|
if (!RBI.constrainGenericRegister(Src, AMDGPU::SReg_64RegClass, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(Dst, AMDGPU::SReg_64RegClass, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
MachineBasicBlock *BB = MI.getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
Register LoReg = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
Register HiReg = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
Register ConstReg = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
Register OpReg = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::COPY), LoReg)
|
|
|
|
.addReg(Src, 0, AMDGPU::sub0);
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::COPY), HiReg)
|
|
|
|
.addReg(Src, 0, AMDGPU::sub1);
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::S_MOV_B32), ConstReg)
|
|
|
|
.addImm(0x80000000);
|
|
|
|
|
|
|
|
// Set or toggle sign bit.
|
|
|
|
unsigned Opc = Fabs ? AMDGPU::S_OR_B32 : AMDGPU::S_XOR_B32;
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(Opc), OpReg)
|
|
|
|
.addReg(HiReg)
|
|
|
|
.addReg(ConstReg);
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::REG_SEQUENCE), Dst)
|
|
|
|
.addReg(LoReg)
|
|
|
|
.addImm(AMDGPU::sub0)
|
|
|
|
.addReg(OpReg)
|
|
|
|
.addImm(AMDGPU::sub1);
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-04-07 17:33:26 +02:00
|
|
|
// FIXME: This is a workaround for the same tablegen problems as G_FNEG
|
|
|
|
bool AMDGPUInstructionSelector::selectG_FABS(MachineInstr &MI) const {
|
|
|
|
Register Dst = MI.getOperand(0).getReg();
|
|
|
|
const RegisterBank *DstRB = RBI.getRegBank(Dst, *MRI, TRI);
|
|
|
|
if (DstRB->getID() != AMDGPU::SGPRRegBankID ||
|
|
|
|
MRI->getType(Dst) != LLT::scalar(64))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
Register Src = MI.getOperand(1).getReg();
|
|
|
|
MachineBasicBlock *BB = MI.getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
Register LoReg = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
Register HiReg = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
Register ConstReg = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
Register OpReg = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
|
|
|
|
if (!RBI.constrainGenericRegister(Src, AMDGPU::SReg_64RegClass, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(Dst, AMDGPU::SReg_64RegClass, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::COPY), LoReg)
|
|
|
|
.addReg(Src, 0, AMDGPU::sub0);
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::COPY), HiReg)
|
|
|
|
.addReg(Src, 0, AMDGPU::sub1);
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::S_MOV_B32), ConstReg)
|
|
|
|
.addImm(0x7fffffff);
|
|
|
|
|
|
|
|
// Clear sign bit.
|
|
|
|
// TODO: Should this used S_BITSET0_*?
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::S_AND_B32), OpReg)
|
|
|
|
.addReg(HiReg)
|
|
|
|
.addReg(ConstReg);
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::REG_SEQUENCE), Dst)
|
|
|
|
.addReg(LoReg)
|
|
|
|
.addImm(AMDGPU::sub0)
|
|
|
|
.addReg(OpReg)
|
|
|
|
.addImm(AMDGPU::sub1);
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
static bool isConstant(const MachineInstr &MI) {
|
|
|
|
return MI.getOpcode() == TargetOpcode::G_CONSTANT;
|
|
|
|
}
|
|
|
|
|
|
|
|
void AMDGPUInstructionSelector::getAddrModeInfo(const MachineInstr &Load,
|
|
|
|
const MachineRegisterInfo &MRI, SmallVectorImpl<GEPInfo> &AddrInfo) const {
|
|
|
|
|
|
|
|
const MachineInstr *PtrMI = MRI.getUniqueVRegDef(Load.getOperand(1).getReg());
|
|
|
|
|
|
|
|
assert(PtrMI);
|
|
|
|
|
[globalisel] Rename G_GEP to G_PTR_ADD
Summary:
G_GEP is rather poorly named. It's a simple pointer+scalar addition and
doesn't support any of the complexities of getelementptr. I therefore
propose that we rename it. There's a G_PTR_MASK so let's follow that
convention and go with G_PTR_ADD
Reviewers: volkan, aditya_nandakumar, bogner, rovka, arsenm
Subscribers: sdardis, jvesely, wdng, nhaehnle, hiraditya, jrtc27, atanasyan, arphaman, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D69734
2019-11-01 21:18:00 +01:00
|
|
|
if (PtrMI->getOpcode() != TargetOpcode::G_PTR_ADD)
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
return;
|
|
|
|
|
|
|
|
GEPInfo GEPInfo(*PtrMI);
|
|
|
|
|
2019-09-05 04:20:25 +02:00
|
|
|
for (unsigned i = 1; i != 3; ++i) {
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
const MachineOperand &GEPOp = PtrMI->getOperand(i);
|
|
|
|
const MachineInstr *OpDef = MRI.getUniqueVRegDef(GEPOp.getReg());
|
|
|
|
assert(OpDef);
|
2019-09-05 04:20:25 +02:00
|
|
|
if (i == 2 && isConstant(*OpDef)) {
|
|
|
|
// TODO: Could handle constant base + variable offset, but a combine
|
|
|
|
// probably should have commuted it.
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
assert(GEPInfo.Imm == 0);
|
|
|
|
GEPInfo.Imm = OpDef->getOperand(1).getCImm()->getSExtValue();
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
const RegisterBank *OpBank = RBI.getRegBank(GEPOp.getReg(), MRI, TRI);
|
|
|
|
if (OpBank->getID() == AMDGPU::SGPRRegBankID)
|
|
|
|
GEPInfo.SgprParts.push_back(GEPOp.getReg());
|
|
|
|
else
|
|
|
|
GEPInfo.VgprParts.push_back(GEPOp.getReg());
|
|
|
|
}
|
|
|
|
|
|
|
|
AddrInfo.push_back(GEPInfo);
|
|
|
|
getAddrModeInfo(*PtrMI, MRI, AddrInfo);
|
|
|
|
}
|
|
|
|
|
2020-11-10 00:01:29 +01:00
|
|
|
bool AMDGPUInstructionSelector::isSGPR(Register Reg) const {
|
|
|
|
return RBI.getRegBank(Reg, *MRI, TRI)->getID() == AMDGPU::SGPRRegBankID;
|
|
|
|
}
|
|
|
|
|
AMDGPU/GlobalISel: Move SMRD selection logic to TableGen
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: volkan, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52922
llvm-svn: 354516
2019-02-20 22:02:37 +01:00
|
|
|
bool AMDGPUInstructionSelector::isInstrUniform(const MachineInstr &MI) const {
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
if (!MI.hasOneMemOperand())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const MachineMemOperand *MMO = *MI.memoperands_begin();
|
|
|
|
const Value *Ptr = MMO->getValue();
|
|
|
|
|
|
|
|
// UndefValue means this is a load of a kernel input. These are uniform.
|
|
|
|
// Sometimes LDS instructions have constant pointers.
|
|
|
|
// If Ptr is null, then that means this mem operand contains a
|
|
|
|
// PseudoSourceValue like GOT.
|
|
|
|
if (!Ptr || isa<UndefValue>(Ptr) || isa<Argument>(Ptr) ||
|
|
|
|
isa<Constant>(Ptr) || isa<GlobalValue>(Ptr))
|
|
|
|
return true;
|
|
|
|
|
2018-02-09 17:57:57 +01:00
|
|
|
if (MMO->getAddrSpace() == AMDGPUAS::CONSTANT_ADDRESS_32BIT)
|
|
|
|
return true;
|
|
|
|
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
const Instruction *I = dyn_cast<Instruction>(Ptr);
|
|
|
|
return I && I->getMetadata("amdgpu.uniform");
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AMDGPUInstructionSelector::hasVgprParts(ArrayRef<GEPInfo> AddrInfo) const {
|
|
|
|
for (const GEPInfo &GEPInfo : AddrInfo) {
|
|
|
|
if (!GEPInfo.VgprParts.empty())
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2019-08-01 05:09:15 +02:00
|
|
|
void AMDGPUInstructionSelector::initM0(MachineInstr &I) const {
|
2019-09-28 05:41:13 +02:00
|
|
|
const LLT PtrTy = MRI->getType(I.getOperand(1).getReg());
|
2019-08-01 02:53:38 +02:00
|
|
|
unsigned AS = PtrTy.getAddressSpace();
|
|
|
|
if ((AS == AMDGPUAS::LOCAL_ADDRESS || AS == AMDGPUAS::REGION_ADDRESS) &&
|
|
|
|
STI.ldsRequiresM0Init()) {
|
2020-08-09 21:26:30 +02:00
|
|
|
MachineBasicBlock *BB = I.getParent();
|
|
|
|
|
2019-08-01 02:53:38 +02:00
|
|
|
// If DS instructions require M0 initializtion, insert it before selecting.
|
|
|
|
BuildMI(*BB, &I, I.getDebugLoc(), TII.get(AMDGPU::S_MOV_B32), AMDGPU::M0)
|
|
|
|
.addImm(-1);
|
|
|
|
}
|
2019-08-01 05:09:15 +02:00
|
|
|
}
|
2019-08-01 02:53:38 +02:00
|
|
|
|
2020-08-09 21:26:30 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_LOAD_STORE_ATOMICRMW(
|
|
|
|
MachineInstr &I) const {
|
2021-03-01 18:24:57 +01:00
|
|
|
if (I.getOpcode() == TargetOpcode::G_ATOMICRMW_FADD) {
|
|
|
|
const LLT PtrTy = MRI->getType(I.getOperand(1).getReg());
|
|
|
|
unsigned AS = PtrTy.getAddressSpace();
|
|
|
|
if (AS == AMDGPUAS::GLOBAL_ADDRESS)
|
|
|
|
return selectGlobalAtomicFadd(I, I.getOperand(1), I.getOperand(2));
|
|
|
|
}
|
|
|
|
|
2019-08-01 05:09:15 +02:00
|
|
|
initM0(I);
|
2019-08-13 08:26:59 +02:00
|
|
|
return selectImpl(I, *CoverageInfo);
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
}
|
|
|
|
|
2020-02-10 04:32:44 +01:00
|
|
|
// TODO: No rtn optimization.
|
|
|
|
bool AMDGPUInstructionSelector::selectG_AMDGPU_ATOMIC_CMPXCHG(
|
|
|
|
MachineInstr &MI) const {
|
|
|
|
Register PtrReg = MI.getOperand(1).getReg();
|
|
|
|
const LLT PtrTy = MRI->getType(PtrReg);
|
|
|
|
if (PtrTy.getAddressSpace() == AMDGPUAS::FLAT_ADDRESS ||
|
|
|
|
STI.useFlatForGlobal())
|
|
|
|
return selectImpl(MI, *CoverageInfo);
|
|
|
|
|
|
|
|
Register DstReg = MI.getOperand(0).getReg();
|
|
|
|
const LLT Ty = MRI->getType(DstReg);
|
|
|
|
const bool Is64 = Ty.getSizeInBits() == 64;
|
|
|
|
const unsigned SubReg = Is64 ? AMDGPU::sub0_sub1 : AMDGPU::sub0;
|
|
|
|
Register TmpReg = MRI->createVirtualRegister(
|
|
|
|
Is64 ? &AMDGPU::VReg_128RegClass : &AMDGPU::VReg_64RegClass);
|
|
|
|
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
MachineBasicBlock *BB = MI.getParent();
|
|
|
|
|
|
|
|
Register VAddr, RSrcReg, SOffset;
|
|
|
|
int64_t Offset = 0;
|
|
|
|
|
|
|
|
unsigned Opcode;
|
|
|
|
if (selectMUBUFOffsetImpl(MI.getOperand(1), RSrcReg, SOffset, Offset)) {
|
|
|
|
Opcode = Is64 ? AMDGPU::BUFFER_ATOMIC_CMPSWAP_X2_OFFSET_RTN :
|
|
|
|
AMDGPU::BUFFER_ATOMIC_CMPSWAP_OFFSET_RTN;
|
|
|
|
} else if (selectMUBUFAddr64Impl(MI.getOperand(1), VAddr,
|
|
|
|
RSrcReg, SOffset, Offset)) {
|
|
|
|
Opcode = Is64 ? AMDGPU::BUFFER_ATOMIC_CMPSWAP_X2_ADDR64_RTN :
|
|
|
|
AMDGPU::BUFFER_ATOMIC_CMPSWAP_ADDR64_RTN;
|
|
|
|
} else
|
|
|
|
return selectImpl(MI, *CoverageInfo);
|
|
|
|
|
|
|
|
auto MIB = BuildMI(*BB, &MI, DL, TII.get(Opcode), TmpReg)
|
|
|
|
.addReg(MI.getOperand(2).getReg());
|
|
|
|
|
|
|
|
if (VAddr)
|
|
|
|
MIB.addReg(VAddr);
|
|
|
|
|
|
|
|
MIB.addReg(RSrcReg);
|
|
|
|
if (SOffset)
|
|
|
|
MIB.addReg(SOffset);
|
|
|
|
else
|
|
|
|
MIB.addImm(0);
|
|
|
|
|
|
|
|
MIB.addImm(Offset);
|
2021-02-09 01:36:10 +01:00
|
|
|
MIB.addImm(AMDGPU::CPol::GLC);
|
2020-02-10 04:32:44 +01:00
|
|
|
MIB.cloneMemRefs(MI);
|
|
|
|
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::COPY), DstReg)
|
|
|
|
.addReg(TmpReg, RegState::Kill, SubReg);
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
|
|
|
|
MRI->setRegClass(
|
|
|
|
DstReg, Is64 ? &AMDGPU::VReg_64RegClass : &AMDGPU::VGPR_32RegClass);
|
|
|
|
return constrainSelectedInstRegOperands(*MIB, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
2019-07-01 17:39:27 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_BRCOND(MachineInstr &I) const {
|
|
|
|
MachineBasicBlock *BB = I.getParent();
|
|
|
|
MachineOperand &CondOp = I.getOperand(0);
|
|
|
|
Register CondReg = CondOp.getReg();
|
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
|
2019-07-01 18:06:02 +02:00
|
|
|
unsigned BrOpcode;
|
|
|
|
Register CondPhysReg;
|
|
|
|
const TargetRegisterClass *ConstrainRC;
|
|
|
|
|
|
|
|
// In SelectionDAG, we inspect the IR block for uniformity metadata to decide
|
|
|
|
// whether the branch is uniform when selecting the instruction. In
|
|
|
|
// GlobalISel, we should push that decision into RegBankSelect. Assume for now
|
|
|
|
// RegBankSelect knows what it's doing if the branch condition is scc, even
|
|
|
|
// though it currently does not.
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
if (!isVCC(CondReg, *MRI)) {
|
|
|
|
if (MRI->getType(CondReg) != LLT::scalar(32))
|
|
|
|
return false;
|
|
|
|
|
2019-07-01 18:06:02 +02:00
|
|
|
CondPhysReg = AMDGPU::SCC;
|
|
|
|
BrOpcode = AMDGPU::S_CBRANCH_SCC1;
|
2020-07-26 16:47:08 +02:00
|
|
|
ConstrainRC = &AMDGPU::SReg_32RegClass;
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
} else {
|
2019-07-01 18:06:02 +02:00
|
|
|
// FIXME: Do we have to insert an and with exec here, like in SelectionDAG?
|
|
|
|
// We sort of know that a VCC producer based on the register bank, that ands
|
|
|
|
// inactive lanes with 0. What if there was a logical operation with vcc
|
|
|
|
// producers in different blocks/with different exec masks?
|
|
|
|
// FIXME: Should scc->vcc copies and with exec?
|
|
|
|
CondPhysReg = TRI.getVCC();
|
|
|
|
BrOpcode = AMDGPU::S_CBRANCH_VCCNZ;
|
|
|
|
ConstrainRC = TRI.getBoolRC();
|
AMDGPU/GlobalISel: Replace handling of boolean values
This solves selection failures with generated selection patterns,
which would fail due to inferring the SGPR reg bank for virtual
registers with a set register class instead of VCC bank. Use
instruction selection would constrain the virtual register to a
specific class, so when the def was selected later the bank no longer
was set to VCC.
Remove the SCC reg bank. SCC isn't directly addressable, so it
requires copying from SCC to an allocatable 32-bit register during
selection, so these might as well be treated as 32-bit SGPR values.
Now any scalar boolean value that will produce an outupt in SCC should
be widened during RegBankSelect to s32. Any s1 value should be a
vector boolean during selection. This makes the vcc register bank
unambiguous with a normal SGPR during selection.
Summary of how this should now work:
- G_TRUNC is always a no-op, and never should use a vcc bank result.
- SALU boolean operations should be promoted to s32 in RegBankSelect
apply mapping
- An s1 value means vcc bank at selection. The exception is for
legalization artifacts that use s1, which are never VCC. All other
contexts should infer the VCC register classes for s1 typed
registers. The LLT for the register is now needed to infer the
correct register class. Extensions with vcc sources should be
legalized to a select of constants during RegBankSelect.
- Copy from non-vcc to vcc ensures high bits of the input value are
cleared during selection.
- SALU boolean inputs should ensure the inputs are 0/1. This includes
select, conditional branches, and carry-ins.
There are a few somewhat dirty details. One is that G_TRUNC/G_*EXT
selection ignores the usual register-bank from register class
functions, and can't handle truncates with VCC result banks. I think
this is OK, since the artifacts are specially treated anyway. This
does require some care to avoid producing cases with vcc. There will
also be no 100% reliable way to verify this rule is followed in
selection in case of register classes, and violations manifests
themselves as invalid copy instructions much later.
Standard phi handling also only considers the bank of the result
register, and doesn't insert copies to make the source banks
match. This doesn't work for vcc, so we have to manually correct phi
inputs in this case. We should add a verifier check to make sure there
are no phis with mixed vcc and non-vcc register bank inputs.
There's also some duplication with the LegalizerHelper, and some code
which should live in the helper. I don't see a good way to share
special knowledge about what types to use for intermediate operations
depending on the bank for example. Using the helper to replace
extensions with selects also seems somewhat awkward to me.
Another issue is there are some contexts calling
getRegBankFromRegClass that apparently don't have the LLT type for the
register, but I haven't yet run into a real issue from this.
This also introduces new unnecessary instructions in most cases, since
we don't yet try to optimize out the zext when the source is known to
come from a compare.
2019-11-02 17:30:59 +01:00
|
|
|
}
|
2019-07-01 17:39:27 +02:00
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
if (!MRI->getRegClassOrNull(CondReg))
|
|
|
|
MRI->setRegClass(CondReg, ConstrainRC);
|
2019-07-01 17:39:27 +02:00
|
|
|
|
2019-07-01 18:06:02 +02:00
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::COPY), CondPhysReg)
|
|
|
|
.addReg(CondReg);
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(BrOpcode))
|
|
|
|
.addMBB(I.getOperand(1).getMBB());
|
|
|
|
|
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
2019-07-01 17:39:27 +02:00
|
|
|
}
|
|
|
|
|
2020-07-31 22:01:38 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_GLOBAL_VALUE(
|
2020-01-26 05:20:38 +01:00
|
|
|
MachineInstr &I) const {
|
2019-07-01 17:48:18 +02:00
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
2019-09-28 05:41:13 +02:00
|
|
|
const RegisterBank *DstRB = RBI.getRegBank(DstReg, *MRI, TRI);
|
2019-07-01 17:48:18 +02:00
|
|
|
const bool IsVGPR = DstRB->getID() == AMDGPU::VGPRRegBankID;
|
|
|
|
I.setDesc(TII.get(IsVGPR ? AMDGPU::V_MOV_B32_e32 : AMDGPU::S_MOV_B32));
|
|
|
|
if (IsVGPR)
|
|
|
|
I.addOperand(*MF, MachineOperand::CreateReg(AMDGPU::EXEC, false, true));
|
|
|
|
|
|
|
|
return RBI.constrainGenericRegister(
|
2019-09-28 05:41:13 +02:00
|
|
|
DstReg, IsVGPR ? AMDGPU::VGPR_32RegClass : AMDGPU::SReg_32RegClass, *MRI);
|
2019-07-01 17:48:18 +02:00
|
|
|
}
|
|
|
|
|
2020-05-16 00:33:01 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectG_PTRMASK(MachineInstr &I) const {
|
2019-09-09 17:46:13 +02:00
|
|
|
Register DstReg = I.getOperand(0).getReg();
|
|
|
|
Register SrcReg = I.getOperand(1).getReg();
|
2020-05-23 23:10:58 +02:00
|
|
|
Register MaskReg = I.getOperand(2).getReg();
|
|
|
|
LLT Ty = MRI->getType(DstReg);
|
|
|
|
LLT MaskTy = MRI->getType(MaskReg);
|
2019-09-09 17:46:13 +02:00
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
const RegisterBank *DstRB = RBI.getRegBank(DstReg, *MRI, TRI);
|
|
|
|
const RegisterBank *SrcRB = RBI.getRegBank(SrcReg, *MRI, TRI);
|
2020-05-23 23:10:58 +02:00
|
|
|
const RegisterBank *MaskRB = RBI.getRegBank(MaskReg, *MRI, TRI);
|
2019-09-09 17:46:13 +02:00
|
|
|
const bool IsVGPR = DstRB->getID() == AMDGPU::VGPRRegBankID;
|
2020-05-23 23:10:58 +02:00
|
|
|
if (DstRB != SrcRB) // Should only happen for hand written MIR.
|
|
|
|
return false;
|
|
|
|
|
2019-09-09 17:46:13 +02:00
|
|
|
unsigned NewOpc = IsVGPR ? AMDGPU::V_AND_B32_e64 : AMDGPU::S_AND_B32;
|
|
|
|
const TargetRegisterClass &RegRC
|
|
|
|
= IsVGPR ? AMDGPU::VGPR_32RegClass : AMDGPU::SReg_32RegClass;
|
|
|
|
|
|
|
|
const TargetRegisterClass *DstRC = TRI.getRegClassForTypeOnBank(Ty, *DstRB,
|
2019-09-28 05:41:13 +02:00
|
|
|
*MRI);
|
2019-09-09 17:46:13 +02:00
|
|
|
const TargetRegisterClass *SrcRC = TRI.getRegClassForTypeOnBank(Ty, *SrcRB,
|
2019-09-28 05:41:13 +02:00
|
|
|
*MRI);
|
2020-05-23 23:10:58 +02:00
|
|
|
const TargetRegisterClass *MaskRC =
|
|
|
|
TRI.getRegClassForTypeOnBank(MaskTy, *MaskRB, *MRI);
|
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
if (!RBI.constrainGenericRegister(DstReg, *DstRC, *MRI) ||
|
2020-05-23 23:10:58 +02:00
|
|
|
!RBI.constrainGenericRegister(SrcReg, *SrcRC, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(MaskReg, *MaskRC, *MRI))
|
2019-09-09 17:46:13 +02:00
|
|
|
return false;
|
|
|
|
|
2020-05-23 23:10:58 +02:00
|
|
|
MachineBasicBlock *BB = I.getParent();
|
2019-09-09 17:46:13 +02:00
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
if (Ty.getSizeInBits() == 32) {
|
2020-05-23 23:10:58 +02:00
|
|
|
assert(MaskTy.getSizeInBits() == 32 &&
|
|
|
|
"ptrmask should have been narrowed during legalize");
|
|
|
|
|
2019-09-09 17:46:13 +02:00
|
|
|
BuildMI(*BB, &I, DL, TII.get(NewOpc), DstReg)
|
|
|
|
.addReg(SrcReg)
|
2020-05-23 23:10:58 +02:00
|
|
|
.addReg(MaskReg);
|
2019-09-09 17:46:13 +02:00
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
Register HiReg = MRI->createVirtualRegister(&RegRC);
|
|
|
|
Register LoReg = MRI->createVirtualRegister(&RegRC);
|
2019-09-09 17:46:13 +02:00
|
|
|
|
2020-05-23 23:10:58 +02:00
|
|
|
// Extract the subregisters from the source pointer.
|
2019-09-09 17:46:13 +02:00
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::COPY), LoReg)
|
|
|
|
.addReg(SrcReg, 0, AMDGPU::sub0);
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::COPY), HiReg)
|
|
|
|
.addReg(SrcReg, 0, AMDGPU::sub1);
|
|
|
|
|
2020-05-23 23:10:58 +02:00
|
|
|
Register MaskedLo, MaskedHi;
|
|
|
|
|
|
|
|
// Try to avoid emitting a bit operation when we only need to touch half of
|
|
|
|
// the 64-bit pointer.
|
|
|
|
APInt MaskOnes = KnownBits->getKnownOnes(MaskReg).zextOrSelf(64);
|
|
|
|
|
|
|
|
const APInt MaskHi32 = APInt::getHighBitsSet(64, 32);
|
|
|
|
const APInt MaskLo32 = APInt::getLowBitsSet(64, 32);
|
|
|
|
if ((MaskOnes & MaskLo32) == MaskLo32) {
|
|
|
|
// If all the bits in the low half are 1, we only need a copy for it.
|
|
|
|
MaskedLo = LoReg;
|
|
|
|
} else {
|
|
|
|
// Extract the mask subregister and apply the and.
|
|
|
|
Register MaskLo = MRI->createVirtualRegister(&RegRC);
|
|
|
|
MaskedLo = MRI->createVirtualRegister(&RegRC);
|
|
|
|
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::COPY), MaskLo)
|
|
|
|
.addReg(MaskReg, 0, AMDGPU::sub0);
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(NewOpc), MaskedLo)
|
|
|
|
.addReg(LoReg)
|
|
|
|
.addReg(MaskLo);
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((MaskOnes & MaskHi32) == MaskHi32) {
|
|
|
|
// If all the bits in the high half are 1, we only need a copy for it.
|
|
|
|
MaskedHi = HiReg;
|
|
|
|
} else {
|
|
|
|
Register MaskHi = MRI->createVirtualRegister(&RegRC);
|
|
|
|
MaskedHi = MRI->createVirtualRegister(&RegRC);
|
|
|
|
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::COPY), MaskHi)
|
|
|
|
.addReg(MaskReg, 0, AMDGPU::sub1);
|
|
|
|
BuildMI(*BB, &I, DL, TII.get(NewOpc), MaskedHi)
|
|
|
|
.addReg(HiReg)
|
|
|
|
.addReg(MaskHi);
|
|
|
|
}
|
|
|
|
|
2019-09-09 17:46:13 +02:00
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::REG_SEQUENCE), DstReg)
|
2020-05-23 23:10:58 +02:00
|
|
|
.addReg(MaskedLo)
|
2019-09-09 17:46:13 +02:00
|
|
|
.addImm(AMDGPU::sub0)
|
2020-05-23 23:10:58 +02:00
|
|
|
.addReg(MaskedHi)
|
2019-09-09 17:46:13 +02:00
|
|
|
.addImm(AMDGPU::sub1);
|
|
|
|
I.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-01-03 16:07:51 +01:00
|
|
|
/// Return the register to use for the index value, and the subregister to use
|
|
|
|
/// for the indirectly accessed register.
|
|
|
|
static std::pair<Register, unsigned>
|
|
|
|
computeIndirectRegIndex(MachineRegisterInfo &MRI,
|
|
|
|
const SIRegisterInfo &TRI,
|
|
|
|
const TargetRegisterClass *SuperRC,
|
|
|
|
Register IdxReg,
|
|
|
|
unsigned EltSize) {
|
|
|
|
Register IdxBaseReg;
|
|
|
|
int Offset;
|
|
|
|
|
2020-11-09 11:46:38 +01:00
|
|
|
std::tie(IdxBaseReg, Offset) = AMDGPU::getBaseWithConstantOffset(MRI, IdxReg);
|
2020-02-06 23:18:17 +01:00
|
|
|
if (IdxBaseReg == AMDGPU::NoRegister) {
|
|
|
|
// This will happen if the index is a known constant. This should ordinarily
|
|
|
|
// be legalized out, but handle it as a register just in case.
|
|
|
|
assert(Offset == 0);
|
|
|
|
IdxBaseReg = IdxReg;
|
|
|
|
}
|
2020-01-03 16:07:51 +01:00
|
|
|
|
|
|
|
ArrayRef<int16_t> SubRegs = TRI.getRegSplitParts(SuperRC, EltSize);
|
|
|
|
|
|
|
|
// Skip out of bounds offsets, or else we would end up using an undefined
|
|
|
|
// register.
|
|
|
|
if (static_cast<unsigned>(Offset) >= SubRegs.size())
|
|
|
|
return std::make_pair(IdxReg, SubRegs[0]);
|
|
|
|
return std::make_pair(IdxBaseReg, SubRegs[Offset]);
|
|
|
|
}
|
|
|
|
|
2020-01-02 22:45:33 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectG_EXTRACT_VECTOR_ELT(
|
|
|
|
MachineInstr &MI) const {
|
|
|
|
Register DstReg = MI.getOperand(0).getReg();
|
|
|
|
Register SrcReg = MI.getOperand(1).getReg();
|
|
|
|
Register IdxReg = MI.getOperand(2).getReg();
|
|
|
|
|
|
|
|
LLT DstTy = MRI->getType(DstReg);
|
|
|
|
LLT SrcTy = MRI->getType(SrcReg);
|
|
|
|
|
|
|
|
const RegisterBank *DstRB = RBI.getRegBank(DstReg, *MRI, TRI);
|
|
|
|
const RegisterBank *SrcRB = RBI.getRegBank(SrcReg, *MRI, TRI);
|
|
|
|
const RegisterBank *IdxRB = RBI.getRegBank(IdxReg, *MRI, TRI);
|
|
|
|
|
|
|
|
// The index must be scalar. If it wasn't RegBankSelect should have moved this
|
|
|
|
// into a waterfall loop.
|
|
|
|
if (IdxRB->getID() != AMDGPU::SGPRRegBankID)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const TargetRegisterClass *SrcRC = TRI.getRegClassForTypeOnBank(SrcTy, *SrcRB,
|
|
|
|
*MRI);
|
|
|
|
const TargetRegisterClass *DstRC = TRI.getRegClassForTypeOnBank(DstTy, *DstRB,
|
|
|
|
*MRI);
|
2020-04-17 12:49:38 +02:00
|
|
|
if (!SrcRC || !DstRC)
|
|
|
|
return false;
|
2020-01-02 22:45:33 +01:00
|
|
|
if (!RBI.constrainGenericRegister(SrcReg, *SrcRC, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(DstReg, *DstRC, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(IdxReg, AMDGPU::SReg_32RegClass, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
MachineBasicBlock *BB = MI.getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
const bool Is64 = DstTy.getSizeInBits() == 64;
|
|
|
|
|
2020-01-03 16:07:51 +01:00
|
|
|
unsigned SubReg;
|
|
|
|
std::tie(IdxReg, SubReg) = computeIndirectRegIndex(*MRI, TRI, SrcRC, IdxReg,
|
|
|
|
DstTy.getSizeInBits() / 8);
|
2020-01-02 22:45:33 +01:00
|
|
|
|
|
|
|
if (SrcRB->getID() == AMDGPU::SGPRRegBankID) {
|
|
|
|
if (DstTy.getSizeInBits() != 32 && !Is64)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::COPY), AMDGPU::M0)
|
|
|
|
.addReg(IdxReg);
|
|
|
|
|
|
|
|
unsigned Opc = Is64 ? AMDGPU::S_MOVRELS_B64 : AMDGPU::S_MOVRELS_B32;
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(Opc), DstReg)
|
|
|
|
.addReg(SrcReg, 0, SubReg)
|
|
|
|
.addReg(SrcReg, RegState::Implicit);
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (SrcRB->getID() != AMDGPU::VGPRRegBankID || DstTy.getSizeInBits() != 32)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (!STI.useVGPRIndexMode()) {
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::COPY), AMDGPU::M0)
|
|
|
|
.addReg(IdxReg);
|
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::V_MOVRELS_B32_e32), DstReg)
|
2020-07-30 02:17:45 +02:00
|
|
|
.addReg(SrcReg, 0, SubReg)
|
2020-01-02 22:45:33 +01:00
|
|
|
.addReg(SrcReg, RegState::Implicit);
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-11-06 08:43:58 +01:00
|
|
|
const MCInstrDesc &GPRIDXDesc =
|
|
|
|
TII.getIndirectGPRIDXPseudo(TRI.getRegSizeInBits(*SrcRC), true);
|
|
|
|
BuildMI(*BB, MI, DL, GPRIDXDesc, DstReg)
|
|
|
|
.addReg(SrcReg)
|
|
|
|
.addReg(IdxReg)
|
|
|
|
.addImm(SubReg);
|
2020-01-02 22:45:33 +01:00
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-01-03 18:17:56 +01:00
|
|
|
// TODO: Fold insert_vector_elt (extract_vector_elt) into movrelsd
|
|
|
|
bool AMDGPUInstructionSelector::selectG_INSERT_VECTOR_ELT(
|
|
|
|
MachineInstr &MI) const {
|
|
|
|
Register DstReg = MI.getOperand(0).getReg();
|
|
|
|
Register VecReg = MI.getOperand(1).getReg();
|
|
|
|
Register ValReg = MI.getOperand(2).getReg();
|
|
|
|
Register IdxReg = MI.getOperand(3).getReg();
|
|
|
|
|
|
|
|
LLT VecTy = MRI->getType(DstReg);
|
|
|
|
LLT ValTy = MRI->getType(ValReg);
|
|
|
|
unsigned VecSize = VecTy.getSizeInBits();
|
|
|
|
unsigned ValSize = ValTy.getSizeInBits();
|
|
|
|
|
|
|
|
const RegisterBank *VecRB = RBI.getRegBank(VecReg, *MRI, TRI);
|
|
|
|
const RegisterBank *ValRB = RBI.getRegBank(ValReg, *MRI, TRI);
|
|
|
|
const RegisterBank *IdxRB = RBI.getRegBank(IdxReg, *MRI, TRI);
|
|
|
|
|
|
|
|
assert(VecTy.getElementType() == ValTy);
|
|
|
|
|
|
|
|
// The index must be scalar. If it wasn't RegBankSelect should have moved this
|
|
|
|
// into a waterfall loop.
|
|
|
|
if (IdxRB->getID() != AMDGPU::SGPRRegBankID)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const TargetRegisterClass *VecRC = TRI.getRegClassForTypeOnBank(VecTy, *VecRB,
|
|
|
|
*MRI);
|
|
|
|
const TargetRegisterClass *ValRC = TRI.getRegClassForTypeOnBank(ValTy, *ValRB,
|
|
|
|
*MRI);
|
|
|
|
|
|
|
|
if (!RBI.constrainGenericRegister(VecReg, *VecRC, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(DstReg, *VecRC, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(ValReg, *ValRC, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(IdxReg, AMDGPU::SReg_32RegClass, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (VecRB->getID() == AMDGPU::VGPRRegBankID && ValSize != 32)
|
|
|
|
return false;
|
|
|
|
|
2020-01-04 05:13:15 +01:00
|
|
|
unsigned SubReg;
|
|
|
|
std::tie(IdxReg, SubReg) = computeIndirectRegIndex(*MRI, TRI, VecRC, IdxReg,
|
|
|
|
ValSize / 8);
|
|
|
|
|
2020-01-03 18:17:56 +01:00
|
|
|
const bool IndexMode = VecRB->getID() == AMDGPU::VGPRRegBankID &&
|
|
|
|
STI.useVGPRIndexMode();
|
|
|
|
|
|
|
|
MachineBasicBlock *BB = MI.getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
|
2020-11-06 08:43:58 +01:00
|
|
|
if (!IndexMode) {
|
2020-01-03 18:17:56 +01:00
|
|
|
BuildMI(*BB, &MI, DL, TII.get(AMDGPU::COPY), AMDGPU::M0)
|
|
|
|
.addReg(IdxReg);
|
|
|
|
|
2020-11-06 08:43:58 +01:00
|
|
|
const MCInstrDesc &RegWriteOp = TII.getIndirectRegWriteMovRelPseudo(
|
|
|
|
VecSize, ValSize, VecRB->getID() == AMDGPU::SGPRRegBankID);
|
|
|
|
BuildMI(*BB, MI, DL, RegWriteOp, DstReg)
|
|
|
|
.addReg(VecReg)
|
|
|
|
.addReg(ValReg)
|
|
|
|
.addImm(SubReg);
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
2020-01-03 18:17:56 +01:00
|
|
|
|
2020-11-06 08:43:58 +01:00
|
|
|
const MCInstrDesc &GPRIDXDesc =
|
|
|
|
TII.getIndirectGPRIDXPseudo(TRI.getRegSizeInBits(*VecRC), false);
|
|
|
|
BuildMI(*BB, MI, DL, GPRIDXDesc, DstReg)
|
|
|
|
.addReg(VecReg)
|
|
|
|
.addReg(ValReg)
|
|
|
|
.addReg(IdxReg)
|
|
|
|
.addImm(SubReg);
|
2020-01-03 18:17:56 +01:00
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-02-16 05:56:56 +01:00
|
|
|
static bool isZeroOrUndef(int X) {
|
|
|
|
return X == 0 || X == -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool isOneOrUndef(int X) {
|
|
|
|
return X == 1 || X == -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool isZeroOrOneOrUndef(int X) {
|
|
|
|
return X == 0 || X == 1 || X == -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Normalize a VOP3P shuffle mask to refer to the low/high half of a single
|
|
|
|
// 32-bit register.
|
|
|
|
static Register normalizeVOP3PMask(int NewMask[2], Register Src0, Register Src1,
|
|
|
|
ArrayRef<int> Mask) {
|
|
|
|
NewMask[0] = Mask[0];
|
|
|
|
NewMask[1] = Mask[1];
|
|
|
|
if (isZeroOrOneOrUndef(Mask[0]) && isZeroOrOneOrUndef(Mask[1]))
|
|
|
|
return Src0;
|
|
|
|
|
|
|
|
assert(NewMask[0] == 2 || NewMask[0] == 3 || NewMask[0] == -1);
|
|
|
|
assert(NewMask[1] == 2 || NewMask[1] == 3 || NewMask[1] == -1);
|
|
|
|
|
|
|
|
// Shift the mask inputs to be 0/1;
|
|
|
|
NewMask[0] = NewMask[0] == -1 ? -1 : NewMask[0] - 2;
|
|
|
|
NewMask[1] = NewMask[1] == -1 ? -1 : NewMask[1] - 2;
|
|
|
|
return Src1;
|
|
|
|
}
|
|
|
|
|
|
|
|
// This is only legal with VOP3P instructions as an aid to op_sel matching.
|
|
|
|
bool AMDGPUInstructionSelector::selectG_SHUFFLE_VECTOR(
|
|
|
|
MachineInstr &MI) const {
|
|
|
|
Register DstReg = MI.getOperand(0).getReg();
|
|
|
|
Register Src0Reg = MI.getOperand(1).getReg();
|
|
|
|
Register Src1Reg = MI.getOperand(2).getReg();
|
|
|
|
ArrayRef<int> ShufMask = MI.getOperand(3).getShuffleMask();
|
|
|
|
|
2021-06-24 10:58:21 +02:00
|
|
|
const LLT V2S16 = LLT::fixed_vector(2, 16);
|
2020-02-16 05:56:56 +01:00
|
|
|
if (MRI->getType(DstReg) != V2S16 || MRI->getType(Src0Reg) != V2S16)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (!AMDGPU::isLegalVOP3PShuffleMask(ShufMask))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
assert(ShufMask.size() == 2);
|
|
|
|
assert(STI.hasSDWA() && "no target has VOP3P but not SDWA");
|
|
|
|
|
|
|
|
MachineBasicBlock *MBB = MI.getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
|
|
|
|
const RegisterBank *DstRB = RBI.getRegBank(DstReg, *MRI, TRI);
|
|
|
|
const bool IsVALU = DstRB->getID() == AMDGPU::VGPRRegBankID;
|
|
|
|
const TargetRegisterClass &RC = IsVALU ?
|
|
|
|
AMDGPU::VGPR_32RegClass : AMDGPU::SReg_32RegClass;
|
|
|
|
|
|
|
|
// Handle the degenerate case which should have folded out.
|
|
|
|
if (ShufMask[0] == -1 && ShufMask[1] == -1) {
|
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::IMPLICIT_DEF), DstReg);
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return RBI.constrainGenericRegister(DstReg, RC, *MRI);
|
|
|
|
}
|
|
|
|
|
|
|
|
// A legal VOP3P mask only reads one of the sources.
|
|
|
|
int Mask[2];
|
|
|
|
Register SrcVec = normalizeVOP3PMask(Mask, Src0Reg, Src1Reg, ShufMask);
|
|
|
|
|
|
|
|
if (!RBI.constrainGenericRegister(DstReg, RC, *MRI) ||
|
|
|
|
!RBI.constrainGenericRegister(SrcVec, RC, *MRI))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// TODO: This also should have been folded out
|
|
|
|
if (isZeroOrUndef(Mask[0]) && isOneOrUndef(Mask[1])) {
|
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::COPY), DstReg)
|
|
|
|
.addReg(SrcVec);
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (Mask[0] == 1 && Mask[1] == -1) {
|
|
|
|
if (IsVALU) {
|
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::V_LSHRREV_B32_e64), DstReg)
|
|
|
|
.addImm(16)
|
|
|
|
.addReg(SrcVec);
|
|
|
|
} else {
|
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::S_LSHR_B32), DstReg)
|
|
|
|
.addReg(SrcVec)
|
|
|
|
.addImm(16);
|
|
|
|
}
|
AMDGPU/GlobalISel: Better code for one case of G_SHUFFLE_VECTOR on v2i16
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, dstuttard, tpr, t-tye, hiraditya, kerbowa, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D74987
2020-02-21 21:59:44 +01:00
|
|
|
} else if (Mask[0] == -1 && Mask[1] == 0) {
|
|
|
|
if (IsVALU) {
|
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::V_LSHLREV_B32_e64), DstReg)
|
|
|
|
.addImm(16)
|
|
|
|
.addReg(SrcVec);
|
|
|
|
} else {
|
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::S_LSHL_B32), DstReg)
|
|
|
|
.addReg(SrcVec)
|
|
|
|
.addImm(16);
|
|
|
|
}
|
|
|
|
} else if (Mask[0] == 0 && Mask[1] == 0) {
|
2020-02-16 05:56:56 +01:00
|
|
|
if (IsVALU) {
|
|
|
|
// Write low half of the register into the high half.
|
|
|
|
MachineInstr *MovSDWA =
|
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::V_MOV_B32_sdwa), DstReg)
|
|
|
|
.addImm(0) // $src0_modifiers
|
|
|
|
.addReg(SrcVec) // $src0
|
|
|
|
.addImm(0) // $clamp
|
|
|
|
.addImm(AMDGPU::SDWA::WORD_1) // $dst_sel
|
|
|
|
.addImm(AMDGPU::SDWA::UNUSED_PRESERVE) // $dst_unused
|
|
|
|
.addImm(AMDGPU::SDWA::WORD_0) // $src0_sel
|
|
|
|
.addReg(SrcVec, RegState::Implicit);
|
|
|
|
MovSDWA->tieOperands(0, MovSDWA->getNumOperands() - 1);
|
|
|
|
} else {
|
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::S_PACK_LL_B32_B16), DstReg)
|
|
|
|
.addReg(SrcVec)
|
|
|
|
.addReg(SrcVec);
|
|
|
|
}
|
|
|
|
} else if (Mask[0] == 1 && Mask[1] == 1) {
|
|
|
|
if (IsVALU) {
|
|
|
|
// Write high half of the register into the low half.
|
|
|
|
MachineInstr *MovSDWA =
|
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::V_MOV_B32_sdwa), DstReg)
|
|
|
|
.addImm(0) // $src0_modifiers
|
|
|
|
.addReg(SrcVec) // $src0
|
|
|
|
.addImm(0) // $clamp
|
|
|
|
.addImm(AMDGPU::SDWA::WORD_0) // $dst_sel
|
|
|
|
.addImm(AMDGPU::SDWA::UNUSED_PRESERVE) // $dst_unused
|
|
|
|
.addImm(AMDGPU::SDWA::WORD_1) // $src0_sel
|
|
|
|
.addReg(SrcVec, RegState::Implicit);
|
|
|
|
MovSDWA->tieOperands(0, MovSDWA->getNumOperands() - 1);
|
|
|
|
} else {
|
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::S_PACK_HH_B32_B16), DstReg)
|
|
|
|
.addReg(SrcVec)
|
|
|
|
.addReg(SrcVec);
|
|
|
|
}
|
|
|
|
} else if (Mask[0] == 1 && Mask[1] == 0) {
|
|
|
|
if (IsVALU) {
|
2021-01-07 19:56:02 +01:00
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::V_ALIGNBIT_B32_e64), DstReg)
|
2020-02-16 05:56:56 +01:00
|
|
|
.addReg(SrcVec)
|
|
|
|
.addReg(SrcVec)
|
|
|
|
.addImm(16);
|
|
|
|
} else {
|
|
|
|
Register TmpReg = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::S_LSHR_B32), TmpReg)
|
|
|
|
.addReg(SrcVec)
|
|
|
|
.addImm(16);
|
|
|
|
BuildMI(*MBB, MI, DL, TII.get(AMDGPU::S_PACK_LL_B32_B16), DstReg)
|
|
|
|
.addReg(TmpReg)
|
|
|
|
.addReg(SrcVec);
|
|
|
|
}
|
|
|
|
} else
|
|
|
|
llvm_unreachable("all shuffle masks should be handled");
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-09-11 00:10:52 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectAMDGPU_BUFFER_ATOMIC_FADD(
|
|
|
|
MachineInstr &MI) const {
|
2021-02-17 22:37:46 +01:00
|
|
|
if (STI.hasGFX90AInsts())
|
|
|
|
return selectImpl(MI, *CoverageInfo);
|
2020-09-11 00:10:52 +02:00
|
|
|
|
|
|
|
MachineBasicBlock *MBB = MI.getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
|
|
|
|
if (!MRI->use_nodbg_empty(MI.getOperand(0).getReg())) {
|
|
|
|
Function &F = MBB->getParent()->getFunction();
|
|
|
|
DiagnosticInfoUnsupported
|
|
|
|
NoFpRet(F, "return versions of fp atomics not supported",
|
|
|
|
MI.getDebugLoc(), DS_Error);
|
|
|
|
F.getContext().diagnose(NoFpRet);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// FIXME: This is only needed because tablegen requires number of dst operands
|
|
|
|
// in match and replace pattern to be the same. Otherwise patterns can be
|
|
|
|
// exported from SDag path.
|
|
|
|
MachineOperand &VDataIn = MI.getOperand(1);
|
|
|
|
MachineOperand &VIndex = MI.getOperand(3);
|
|
|
|
MachineOperand &VOffset = MI.getOperand(4);
|
|
|
|
MachineOperand &SOffset = MI.getOperand(5);
|
|
|
|
int16_t Offset = MI.getOperand(6).getImm();
|
|
|
|
|
|
|
|
bool HasVOffset = !isOperandImmEqual(VOffset, 0, *MRI);
|
|
|
|
bool HasVIndex = !isOperandImmEqual(VIndex, 0, *MRI);
|
|
|
|
|
|
|
|
unsigned Opcode;
|
|
|
|
if (HasVOffset) {
|
|
|
|
Opcode = HasVIndex ? AMDGPU::BUFFER_ATOMIC_ADD_F32_BOTHEN
|
|
|
|
: AMDGPU::BUFFER_ATOMIC_ADD_F32_OFFEN;
|
|
|
|
} else {
|
|
|
|
Opcode = HasVIndex ? AMDGPU::BUFFER_ATOMIC_ADD_F32_IDXEN
|
|
|
|
: AMDGPU::BUFFER_ATOMIC_ADD_F32_OFFSET;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (MRI->getType(VDataIn.getReg()).isVector()) {
|
|
|
|
switch (Opcode) {
|
|
|
|
case AMDGPU::BUFFER_ATOMIC_ADD_F32_BOTHEN:
|
|
|
|
Opcode = AMDGPU::BUFFER_ATOMIC_PK_ADD_F16_BOTHEN;
|
|
|
|
break;
|
|
|
|
case AMDGPU::BUFFER_ATOMIC_ADD_F32_OFFEN:
|
|
|
|
Opcode = AMDGPU::BUFFER_ATOMIC_PK_ADD_F16_OFFEN;
|
|
|
|
break;
|
|
|
|
case AMDGPU::BUFFER_ATOMIC_ADD_F32_IDXEN:
|
|
|
|
Opcode = AMDGPU::BUFFER_ATOMIC_PK_ADD_F16_IDXEN;
|
|
|
|
break;
|
|
|
|
case AMDGPU::BUFFER_ATOMIC_ADD_F32_OFFSET:
|
|
|
|
Opcode = AMDGPU::BUFFER_ATOMIC_PK_ADD_F16_OFFSET;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
auto I = BuildMI(*MBB, MI, DL, TII.get(Opcode));
|
|
|
|
I.add(VDataIn);
|
|
|
|
|
|
|
|
if (Opcode == AMDGPU::BUFFER_ATOMIC_ADD_F32_BOTHEN ||
|
|
|
|
Opcode == AMDGPU::BUFFER_ATOMIC_PK_ADD_F16_BOTHEN) {
|
2021-02-19 14:57:14 +01:00
|
|
|
Register IdxReg = MRI->createVirtualRegister(TRI.getVGPR64Class());
|
2020-09-11 00:10:52 +02:00
|
|
|
BuildMI(*MBB, &*I, DL, TII.get(AMDGPU::REG_SEQUENCE), IdxReg)
|
|
|
|
.addReg(VIndex.getReg())
|
|
|
|
.addImm(AMDGPU::sub0)
|
|
|
|
.addReg(VOffset.getReg())
|
|
|
|
.addImm(AMDGPU::sub1);
|
|
|
|
|
|
|
|
I.addReg(IdxReg);
|
|
|
|
} else if (HasVIndex) {
|
|
|
|
I.add(VIndex);
|
|
|
|
} else if (HasVOffset) {
|
|
|
|
I.add(VOffset);
|
|
|
|
}
|
|
|
|
|
|
|
|
I.add(MI.getOperand(2)); // rsrc
|
|
|
|
I.add(SOffset);
|
|
|
|
I.addImm(Offset);
|
2021-02-09 01:36:10 +01:00
|
|
|
I.addImm(MI.getOperand(7).getImm()); // cpol
|
2020-09-11 00:10:52 +02:00
|
|
|
I.cloneMemRefs(MI);
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2021-03-01 18:24:57 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectGlobalAtomicFadd(
|
|
|
|
MachineInstr &MI, MachineOperand &AddrOp, MachineOperand &DataOp) const {
|
2020-09-11 00:10:52 +02:00
|
|
|
|
2021-03-01 18:24:57 +01:00
|
|
|
if (STI.hasGFX90AInsts()) {
|
|
|
|
// gfx90a adds return versions of the global atomic fadd instructions so no
|
|
|
|
// special handling is required.
|
2021-02-17 22:37:46 +01:00
|
|
|
return selectImpl(MI, *CoverageInfo);
|
2021-03-01 18:24:57 +01:00
|
|
|
}
|
2021-02-17 22:37:46 +01:00
|
|
|
|
2020-09-11 00:10:52 +02:00
|
|
|
MachineBasicBlock *MBB = MI.getParent();
|
|
|
|
const DebugLoc &DL = MI.getDebugLoc();
|
|
|
|
|
|
|
|
if (!MRI->use_nodbg_empty(MI.getOperand(0).getReg())) {
|
|
|
|
Function &F = MBB->getParent()->getFunction();
|
|
|
|
DiagnosticInfoUnsupported
|
|
|
|
NoFpRet(F, "return versions of fp atomics not supported",
|
|
|
|
MI.getDebugLoc(), DS_Error);
|
|
|
|
F.getContext().diagnose(NoFpRet);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// FIXME: This is only needed because tablegen requires number of dst operands
|
|
|
|
// in match and replace pattern to be the same. Otherwise patterns can be
|
|
|
|
// exported from SDag path.
|
2021-03-30 14:38:07 +02:00
|
|
|
auto Addr = selectFlatOffsetImpl(AddrOp, SIInstrFlags::FlatGlobal);
|
2020-09-11 00:10:52 +02:00
|
|
|
|
2021-03-01 18:24:57 +01:00
|
|
|
Register Data = DataOp.getReg();
|
2020-09-11 00:10:52 +02:00
|
|
|
const unsigned Opc = MRI->getType(Data).isVector() ?
|
|
|
|
AMDGPU::GLOBAL_ATOMIC_PK_ADD_F16 : AMDGPU::GLOBAL_ATOMIC_ADD_F32;
|
|
|
|
auto MIB = BuildMI(*MBB, &MI, DL, TII.get(Opc))
|
|
|
|
.addReg(Addr.first)
|
|
|
|
.addReg(Data)
|
|
|
|
.addImm(Addr.second)
|
2021-02-09 01:36:10 +01:00
|
|
|
.addImm(0) // cpol
|
2020-09-11 00:10:52 +02:00
|
|
|
.cloneMemRefs(MI);
|
|
|
|
|
|
|
|
MI.eraseFromParent();
|
|
|
|
return constrainSelectedInstRegOperands(*MIB, TII, TRI, RBI);
|
|
|
|
}
|
|
|
|
|
2020-09-17 20:44:52 +02:00
|
|
|
bool AMDGPUInstructionSelector::selectBVHIntrinsic(MachineInstr &MI) const{
|
|
|
|
MI.setDesc(TII.get(MI.getOperand(1).getImm()));
|
|
|
|
MI.RemoveOperand(1);
|
|
|
|
MI.addImplicitDefUseOperands(*MI.getParent()->getParent());
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-08-13 08:26:59 +02:00
|
|
|
bool AMDGPUInstructionSelector::select(MachineInstr &I) {
|
2019-07-01 18:32:47 +02:00
|
|
|
if (I.isPHI())
|
|
|
|
return selectPHI(I);
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
|
2019-10-07 20:43:29 +02:00
|
|
|
if (!I.isPreISelOpcode()) {
|
2018-06-22 02:44:29 +02:00
|
|
|
if (I.isCopy())
|
|
|
|
return selectCOPY(I);
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
return true;
|
2018-06-22 02:44:29 +02:00
|
|
|
}
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
|
|
|
|
switch (I.getOpcode()) {
|
2019-07-15 21:50:07 +02:00
|
|
|
case TargetOpcode::G_AND:
|
|
|
|
case TargetOpcode::G_OR:
|
|
|
|
case TargetOpcode::G_XOR:
|
2020-02-21 03:12:04 +01:00
|
|
|
if (selectImpl(I, *CoverageInfo))
|
2019-07-15 21:50:07 +02:00
|
|
|
return true;
|
2020-02-21 03:12:04 +01:00
|
|
|
return selectG_AND_OR_XOR(I);
|
AMDGPU/GlobalISel: Implement select for 32-bit G_ADD
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: hiraditya, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D58804
llvm-svn: 364797
2019-07-01 18:09:33 +02:00
|
|
|
case TargetOpcode::G_ADD:
|
2019-07-09 16:05:11 +02:00
|
|
|
case TargetOpcode::G_SUB:
|
2019-09-09 17:20:44 +02:00
|
|
|
if (selectImpl(I, *CoverageInfo))
|
AMDGPU/GlobalISel: Implement select for 32-bit G_ADD
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: hiraditya, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D58804
llvm-svn: 364797
2019-07-01 18:09:33 +02:00
|
|
|
return true;
|
2019-09-09 17:20:44 +02:00
|
|
|
return selectG_ADD_SUB(I);
|
2019-10-01 03:23:13 +02:00
|
|
|
case TargetOpcode::G_UADDO:
|
|
|
|
case TargetOpcode::G_USUBO:
|
2019-12-25 15:06:05 +01:00
|
|
|
case TargetOpcode::G_UADDE:
|
|
|
|
case TargetOpcode::G_USUBE:
|
|
|
|
return selectG_UADDO_USUBO_UADDE_USUBE(I);
|
AMDGPU/GlobalISel: Add support for G_INTTOPTR
Summary: This is a no-op.
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52916
llvm-svn: 343839
2018-10-05 06:34:09 +02:00
|
|
|
case TargetOpcode::G_INTTOPTR:
|
2018-05-10 23:20:10 +02:00
|
|
|
case TargetOpcode::G_BITCAST:
|
2019-10-04 10:35:37 +02:00
|
|
|
case TargetOpcode::G_PTRTOINT:
|
2018-05-10 23:20:10 +02:00
|
|
|
return selectCOPY(I);
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
case TargetOpcode::G_CONSTANT:
|
2018-05-15 19:57:09 +02:00
|
|
|
case TargetOpcode::G_FCONSTANT:
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
return selectG_CONSTANT(I);
|
2020-01-26 01:34:22 +01:00
|
|
|
case TargetOpcode::G_FNEG:
|
|
|
|
if (selectImpl(I, *CoverageInfo))
|
|
|
|
return true;
|
|
|
|
return selectG_FNEG(I);
|
2020-04-07 17:33:26 +02:00
|
|
|
case TargetOpcode::G_FABS:
|
|
|
|
if (selectImpl(I, *CoverageInfo))
|
|
|
|
return true;
|
|
|
|
return selectG_FABS(I);
|
2019-03-01 00:37:48 +01:00
|
|
|
case TargetOpcode::G_EXTRACT:
|
|
|
|
return selectG_EXTRACT(I);
|
2019-07-09 16:02:20 +02:00
|
|
|
case TargetOpcode::G_MERGE_VALUES:
|
2019-07-15 19:26:43 +02:00
|
|
|
case TargetOpcode::G_BUILD_VECTOR:
|
2019-07-09 16:02:20 +02:00
|
|
|
case TargetOpcode::G_CONCAT_VECTORS:
|
|
|
|
return selectG_MERGE_VALUES(I);
|
2019-07-09 16:02:26 +02:00
|
|
|
case TargetOpcode::G_UNMERGE_VALUES:
|
|
|
|
return selectG_UNMERGE_VALUES(I);
|
2020-02-15 23:14:17 +01:00
|
|
|
case TargetOpcode::G_BUILD_VECTOR_TRUNC:
|
|
|
|
return selectG_BUILD_VECTOR_TRUNC(I);
|
[globalisel] Rename G_GEP to G_PTR_ADD
Summary:
G_GEP is rather poorly named. It's a simple pointer+scalar addition and
doesn't support any of the complexities of getelementptr. I therefore
propose that we rename it. There's a G_PTR_MASK so let's follow that
convention and go with G_PTR_ADD
Reviewers: volkan, aditya_nandakumar, bogner, rovka, arsenm
Subscribers: sdardis, jvesely, wdng, nhaehnle, hiraditya, jrtc27, atanasyan, arphaman, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D69734
2019-11-01 21:18:00 +01:00
|
|
|
case TargetOpcode::G_PTR_ADD:
|
|
|
|
return selectG_PTR_ADD(I);
|
2018-06-22 01:38:20 +02:00
|
|
|
case TargetOpcode::G_IMPLICIT_DEF:
|
|
|
|
return selectG_IMPLICIT_DEF(I);
|
2020-07-16 11:09:35 +02:00
|
|
|
case TargetOpcode::G_FREEZE:
|
|
|
|
return selectCOPY(I);
|
AMDGPU/GlobalISel: Implement select for G_INSERT
Re-commit r344310.
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D53116
llvm-svn: 355159
2019-03-01 01:50:26 +01:00
|
|
|
case TargetOpcode::G_INSERT:
|
|
|
|
return selectG_INSERT(I);
|
2018-06-14 21:26:37 +02:00
|
|
|
case TargetOpcode::G_INTRINSIC:
|
2019-08-13 08:26:59 +02:00
|
|
|
return selectG_INTRINSIC(I);
|
2018-07-13 23:05:14 +02:00
|
|
|
case TargetOpcode::G_INTRINSIC_W_SIDE_EFFECTS:
|
2019-08-13 08:26:59 +02:00
|
|
|
return selectG_INTRINSIC_W_SIDE_EFFECTS(I);
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
case TargetOpcode::G_ICMP:
|
2019-07-01 15:34:26 +02:00
|
|
|
if (selectG_ICMP(I))
|
|
|
|
return true;
|
2019-08-13 08:26:59 +02:00
|
|
|
return selectImpl(I, *CoverageInfo);
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
case TargetOpcode::G_LOAD:
|
2020-08-09 21:26:30 +02:00
|
|
|
case TargetOpcode::G_STORE:
|
2019-08-01 05:29:01 +02:00
|
|
|
case TargetOpcode::G_ATOMIC_CMPXCHG:
|
|
|
|
case TargetOpcode::G_ATOMICRMW_XCHG:
|
|
|
|
case TargetOpcode::G_ATOMICRMW_ADD:
|
|
|
|
case TargetOpcode::G_ATOMICRMW_SUB:
|
|
|
|
case TargetOpcode::G_ATOMICRMW_AND:
|
|
|
|
case TargetOpcode::G_ATOMICRMW_OR:
|
|
|
|
case TargetOpcode::G_ATOMICRMW_XOR:
|
|
|
|
case TargetOpcode::G_ATOMICRMW_MIN:
|
|
|
|
case TargetOpcode::G_ATOMICRMW_MAX:
|
|
|
|
case TargetOpcode::G_ATOMICRMW_UMIN:
|
|
|
|
case TargetOpcode::G_ATOMICRMW_UMAX:
|
2019-08-01 05:33:15 +02:00
|
|
|
case TargetOpcode::G_ATOMICRMW_FADD:
|
2020-07-26 18:20:29 +02:00
|
|
|
case AMDGPU::G_AMDGPU_ATOMIC_INC:
|
|
|
|
case AMDGPU::G_AMDGPU_ATOMIC_DEC:
|
2020-07-26 17:46:23 +02:00
|
|
|
case AMDGPU::G_AMDGPU_ATOMIC_FMIN:
|
|
|
|
case AMDGPU::G_AMDGPU_ATOMIC_FMAX:
|
2020-08-09 21:26:30 +02:00
|
|
|
return selectG_LOAD_STORE_ATOMICRMW(I);
|
2020-02-10 04:32:44 +01:00
|
|
|
case AMDGPU::G_AMDGPU_ATOMIC_CMPXCHG:
|
|
|
|
return selectG_AMDGPU_ATOMIC_CMPXCHG(I);
|
AMDGPU/GlobalISel: Implement select for G_ICMP and G_SELECT
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D60640
llvm-svn: 363576
2019-06-17 18:27:43 +02:00
|
|
|
case TargetOpcode::G_SELECT:
|
|
|
|
return selectG_SELECT(I);
|
2019-06-24 20:02:18 +02:00
|
|
|
case TargetOpcode::G_TRUNC:
|
|
|
|
return selectG_TRUNC(I);
|
2019-06-25 15:18:11 +02:00
|
|
|
case TargetOpcode::G_SEXT:
|
|
|
|
case TargetOpcode::G_ZEXT:
|
|
|
|
case TargetOpcode::G_ANYEXT:
|
2020-01-25 01:38:53 +01:00
|
|
|
case TargetOpcode::G_SEXT_INREG:
|
2020-01-07 19:10:08 +01:00
|
|
|
if (selectImpl(I, *CoverageInfo))
|
|
|
|
return true;
|
2019-07-24 18:05:53 +02:00
|
|
|
return selectG_SZA_EXT(I);
|
2019-07-01 17:39:27 +02:00
|
|
|
case TargetOpcode::G_BRCOND:
|
|
|
|
return selectG_BRCOND(I);
|
2020-01-26 05:20:38 +01:00
|
|
|
case TargetOpcode::G_GLOBAL_VALUE:
|
2020-07-31 22:01:38 +02:00
|
|
|
return selectG_GLOBAL_VALUE(I);
|
2020-05-16 00:33:01 +02:00
|
|
|
case TargetOpcode::G_PTRMASK:
|
|
|
|
return selectG_PTRMASK(I);
|
2020-01-02 22:45:33 +01:00
|
|
|
case TargetOpcode::G_EXTRACT_VECTOR_ELT:
|
|
|
|
return selectG_EXTRACT_VECTOR_ELT(I);
|
2020-01-03 18:17:56 +01:00
|
|
|
case TargetOpcode::G_INSERT_VECTOR_ELT:
|
|
|
|
return selectG_INSERT_VECTOR_ELT(I);
|
2020-02-16 05:56:56 +01:00
|
|
|
case TargetOpcode::G_SHUFFLE_VECTOR:
|
|
|
|
return selectG_SHUFFLE_VECTOR(I);
|
2020-02-08 05:37:54 +01:00
|
|
|
case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD:
|
|
|
|
case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE: {
|
|
|
|
const AMDGPU::ImageDimIntrinsicInfo *Intr
|
|
|
|
= AMDGPU::getImageDimIntrinsicInfo(I.getIntrinsicID());
|
|
|
|
assert(Intr && "not an image intrinsic with image pseudo");
|
|
|
|
return selectImageIntrinsic(I, Intr);
|
|
|
|
}
|
2020-09-17 20:44:52 +02:00
|
|
|
case AMDGPU::G_AMDGPU_INTRIN_BVH_INTERSECT_RAY:
|
|
|
|
return selectBVHIntrinsic(I);
|
2020-09-11 00:10:52 +02:00
|
|
|
case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FADD:
|
|
|
|
return selectAMDGPU_BUFFER_ATOMIC_FADD(I);
|
2021-04-30 15:57:44 +02:00
|
|
|
case AMDGPU::G_SBFX:
|
|
|
|
case AMDGPU::G_UBFX:
|
|
|
|
return selectG_SBFX_UBFX(I);
|
2019-09-09 17:20:44 +02:00
|
|
|
default:
|
|
|
|
return selectImpl(I, *CoverageInfo);
|
Re-commit AMDGPU/GlobalISel: Add support for simple shaders
Fix build when global-isel is disabled and fix a warning.
Summary: We can select constant/global G_LOAD, global G_STORE, and G_GEP.
Reviewers: qcolombet, MatzeB, t.p.northover, ab, arsenm
Subscribers: mehdi_amini, vkalintiris, kzhuravl, wdng, nhaehnle, mgorny, yaxunl, tony-tye, modocache, llvm-commits, dberris
Differential Revision: https://reviews.llvm.org/D26730
llvm-svn: 293551
2017-01-30 22:56:46 +01:00
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
AMDGPU/GlobalISel: Enable TableGen'd instruction selector
Reviewers: arsenm, nhaehnle
Reviewed By: arsenm
Subscribers: kzhuravl, wdng, mgorny, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D45994
llvm-svn: 332039
2018-05-10 22:53:06 +02:00
|
|
|
|
2018-06-22 04:54:57 +02:00
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectVCSRC(MachineOperand &Root) const {
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.add(Root); }
|
|
|
|
}};
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2019-07-01 17:18:56 +02:00
|
|
|
std::pair<Register, unsigned>
|
2020-10-27 13:29:11 +01:00
|
|
|
AMDGPUInstructionSelector::selectVOP3ModsImpl(MachineOperand &Root,
|
|
|
|
bool AllowAbs) const {
|
2020-02-20 21:34:51 +01:00
|
|
|
Register Src = Root.getReg();
|
|
|
|
Register OrigSrc = Src;
|
2019-07-01 17:18:56 +02:00
|
|
|
unsigned Mods = 0;
|
2020-01-26 02:25:37 +01:00
|
|
|
MachineInstr *MI = getDefIgnoringCopies(Src, *MRI);
|
2019-07-01 17:18:56 +02:00
|
|
|
|
|
|
|
if (MI && MI->getOpcode() == AMDGPU::G_FNEG) {
|
|
|
|
Src = MI->getOperand(1).getReg();
|
|
|
|
Mods |= SISrcMods::NEG;
|
2020-01-26 02:25:37 +01:00
|
|
|
MI = getDefIgnoringCopies(Src, *MRI);
|
2019-07-01 17:18:56 +02:00
|
|
|
}
|
|
|
|
|
2020-10-27 13:29:11 +01:00
|
|
|
if (AllowAbs && MI && MI->getOpcode() == AMDGPU::G_FABS) {
|
2019-07-01 17:18:56 +02:00
|
|
|
Src = MI->getOperand(1).getReg();
|
|
|
|
Mods |= SISrcMods::ABS;
|
|
|
|
}
|
|
|
|
|
2020-02-20 21:34:51 +01:00
|
|
|
if (Mods != 0 &&
|
|
|
|
RBI.getRegBank(Src, *MRI, TRI)->getID() != AMDGPU::VGPRRegBankID) {
|
|
|
|
MachineInstr *UseMI = Root.getParent();
|
|
|
|
|
|
|
|
// If we looked through copies to find source modifiers on an SGPR operand,
|
|
|
|
// we now have an SGPR register source. To avoid potentially violating the
|
|
|
|
// constant bus restriction, we need to insert a copy to a VGPR.
|
|
|
|
Register VGPRSrc = MRI->cloneVirtualRegister(OrigSrc);
|
|
|
|
BuildMI(*UseMI->getParent(), UseMI, UseMI->getDebugLoc(),
|
|
|
|
TII.get(AMDGPU::COPY), VGPRSrc)
|
|
|
|
.addReg(Src);
|
|
|
|
Src = VGPRSrc;
|
|
|
|
}
|
|
|
|
|
2019-07-01 17:18:56 +02:00
|
|
|
return std::make_pair(Src, Mods);
|
|
|
|
}
|
|
|
|
|
AMDGPU/GlobalISel: Enable TableGen'd instruction selector
Reviewers: arsenm, nhaehnle
Reviewed By: arsenm
Subscribers: kzhuravl, wdng, mgorny, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D45994
llvm-svn: 332039
2018-05-10 22:53:06 +02:00
|
|
|
///
|
|
|
|
/// This will select either an SGPR or VGPR operand and will save us from
|
|
|
|
/// having to write an extra tablegen pattern.
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectVSRC0(MachineOperand &Root) const {
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.add(Root); }
|
|
|
|
}};
|
|
|
|
}
|
2018-05-11 07:44:16 +02:00
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectVOP3Mods0(MachineOperand &Root) const {
|
2019-07-01 17:18:56 +02:00
|
|
|
Register Src;
|
|
|
|
unsigned Mods;
|
2020-02-20 21:34:51 +01:00
|
|
|
std::tie(Src, Mods) = selectVOP3ModsImpl(Root);
|
2019-07-01 17:18:56 +02:00
|
|
|
|
2019-09-09 20:29:45 +02:00
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(Src); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(Mods); }, // src0_mods
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(0); }, // clamp
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(0); } // omod
|
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
2020-10-27 13:29:11 +01:00
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectVOP3BMods0(MachineOperand &Root) const {
|
|
|
|
Register Src;
|
|
|
|
unsigned Mods;
|
|
|
|
std::tie(Src, Mods) = selectVOP3ModsImpl(Root, /* AllowAbs */ false);
|
|
|
|
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(Src); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(Mods); }, // src0_mods
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(0); }, // clamp
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(0); } // omod
|
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
2018-06-22 04:34:29 +02:00
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectVOP3OMods(MachineOperand &Root) const {
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.add(Root); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(0); }, // clamp
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(0); } // omod
|
|
|
|
}};
|
|
|
|
}
|
2018-06-14 00:30:47 +02:00
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectVOP3Mods(MachineOperand &Root) const {
|
2019-07-01 17:18:56 +02:00
|
|
|
Register Src;
|
|
|
|
unsigned Mods;
|
2020-02-20 21:34:51 +01:00
|
|
|
std::tie(Src, Mods) = selectVOP3ModsImpl(Root);
|
2019-07-01 17:18:56 +02:00
|
|
|
|
2018-06-14 00:30:47 +02:00
|
|
|
return {{
|
2019-07-01 17:18:56 +02:00
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(Src); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(Mods); } // src_mods
|
2018-06-14 00:30:47 +02:00
|
|
|
}};
|
|
|
|
}
|
AMDGPU/GlobalISel: Move SMRD selection logic to TableGen
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: volkan, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52922
llvm-svn: 354516
2019-02-20 22:02:37 +01:00
|
|
|
|
2020-10-27 13:29:11 +01:00
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectVOP3BMods(MachineOperand &Root) const {
|
|
|
|
Register Src;
|
|
|
|
unsigned Mods;
|
|
|
|
std::tie(Src, Mods) = selectVOP3ModsImpl(Root, /* AllowAbs */ false);
|
|
|
|
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(Src); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(Mods); } // src_mods
|
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
2020-01-26 06:25:35 +01:00
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectVOP3NoMods(MachineOperand &Root) const {
|
|
|
|
Register Reg = Root.getReg();
|
|
|
|
const MachineInstr *Def = getDefIgnoringCopies(Reg, *MRI);
|
|
|
|
if (Def && (Def->getOpcode() == AMDGPU::G_FNEG ||
|
|
|
|
Def->getOpcode() == AMDGPU::G_FABS))
|
|
|
|
return {};
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(Reg); },
|
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
2019-09-01 04:39:00 +02:00
|
|
|
std::pair<Register, unsigned>
|
|
|
|
AMDGPUInstructionSelector::selectVOP3PModsImpl(
|
|
|
|
Register Src, const MachineRegisterInfo &MRI) const {
|
|
|
|
unsigned Mods = 0;
|
|
|
|
MachineInstr *MI = MRI.getVRegDef(Src);
|
|
|
|
|
2020-02-19 18:35:12 +01:00
|
|
|
if (MI && MI->getOpcode() == AMDGPU::G_FNEG &&
|
|
|
|
// It's possible to see an f32 fneg here, but unlikely.
|
|
|
|
// TODO: Treat f32 fneg as only high bit.
|
2021-06-24 10:58:21 +02:00
|
|
|
MRI.getType(Src) == LLT::fixed_vector(2, 16)) {
|
2019-09-01 04:39:00 +02:00
|
|
|
Mods ^= (SISrcMods::NEG | SISrcMods::NEG_HI);
|
|
|
|
Src = MI->getOperand(1).getReg();
|
|
|
|
MI = MRI.getVRegDef(Src);
|
|
|
|
}
|
|
|
|
|
|
|
|
// TODO: Match op_sel through g_build_vector_trunc and g_shuffle_vector.
|
|
|
|
|
|
|
|
// Packed instructions do not have abs modifiers.
|
|
|
|
Mods |= SISrcMods::OP_SEL_1;
|
|
|
|
|
|
|
|
return std::make_pair(Src, Mods);
|
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectVOP3PMods(MachineOperand &Root) const {
|
|
|
|
MachineRegisterInfo &MRI
|
|
|
|
= Root.getParent()->getParent()->getParent()->getRegInfo();
|
|
|
|
|
|
|
|
Register Src;
|
|
|
|
unsigned Mods;
|
|
|
|
std::tie(Src, Mods) = selectVOP3PModsImpl(Root.getReg(), MRI);
|
|
|
|
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(Src); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(Mods); } // src_mods
|
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
2019-09-09 20:29:37 +02:00
|
|
|
InstructionSelector::ComplexRendererFns
|
2020-01-07 18:21:18 +01:00
|
|
|
AMDGPUInstructionSelector::selectVOP3Mods_nnan(MachineOperand &Root) const {
|
|
|
|
Register Src;
|
|
|
|
unsigned Mods;
|
2020-02-20 21:34:51 +01:00
|
|
|
std::tie(Src, Mods) = selectVOP3ModsImpl(Root);
|
2020-09-14 12:03:36 +02:00
|
|
|
if (!isKnownNeverNaN(Src, *MRI))
|
2020-01-07 18:21:18 +01:00
|
|
|
return None;
|
|
|
|
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(Src); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(Mods); } // src_mods
|
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
2019-09-09 20:29:37 +02:00
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectVOP3OpSelMods(MachineOperand &Root) const {
|
|
|
|
// FIXME: Handle op_sel
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(Root.getReg()); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(0); } // src_mods
|
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
AMDGPU/GlobalISel: Move SMRD selection logic to TableGen
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: volkan, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52922
llvm-svn: 354516
2019-02-20 22:02:37 +01:00
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectSmrdImm(MachineOperand &Root) const {
|
|
|
|
SmallVector<GEPInfo, 4> AddrInfo;
|
2019-09-28 05:41:13 +02:00
|
|
|
getAddrModeInfo(*Root.getParent(), *MRI, AddrInfo);
|
AMDGPU/GlobalISel: Move SMRD selection logic to TableGen
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: volkan, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52922
llvm-svn: 354516
2019-02-20 22:02:37 +01:00
|
|
|
|
|
|
|
if (AddrInfo.empty() || AddrInfo[0].SgprParts.size() != 1)
|
|
|
|
return None;
|
|
|
|
|
|
|
|
const GEPInfo &GEPInfo = AddrInfo[0];
|
2020-03-30 08:32:20 +02:00
|
|
|
Optional<int64_t> EncodedImm =
|
|
|
|
AMDGPU::getSMRDEncodedOffset(STI, GEPInfo.Imm, false);
|
2020-01-31 00:41:10 +01:00
|
|
|
if (!EncodedImm)
|
AMDGPU/GlobalISel: Move SMRD selection logic to TableGen
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: volkan, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52922
llvm-svn: 354516
2019-02-20 22:02:37 +01:00
|
|
|
return None;
|
|
|
|
|
|
|
|
unsigned PtrReg = GEPInfo.SgprParts[0];
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(PtrReg); },
|
2020-01-31 00:41:10 +01:00
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(*EncodedImm); }
|
AMDGPU/GlobalISel: Move SMRD selection logic to TableGen
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: volkan, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52922
llvm-svn: 354516
2019-02-20 22:02:37 +01:00
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectSmrdImm32(MachineOperand &Root) const {
|
|
|
|
SmallVector<GEPInfo, 4> AddrInfo;
|
2019-09-28 05:41:13 +02:00
|
|
|
getAddrModeInfo(*Root.getParent(), *MRI, AddrInfo);
|
AMDGPU/GlobalISel: Move SMRD selection logic to TableGen
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: volkan, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52922
llvm-svn: 354516
2019-02-20 22:02:37 +01:00
|
|
|
|
|
|
|
if (AddrInfo.empty() || AddrInfo[0].SgprParts.size() != 1)
|
|
|
|
return None;
|
|
|
|
|
|
|
|
const GEPInfo &GEPInfo = AddrInfo[0];
|
2020-04-21 21:06:53 +02:00
|
|
|
Register PtrReg = GEPInfo.SgprParts[0];
|
2020-01-31 00:41:10 +01:00
|
|
|
Optional<int64_t> EncodedImm =
|
|
|
|
AMDGPU::getSMRDEncodedLiteralOffset32(STI, GEPInfo.Imm);
|
|
|
|
if (!EncodedImm)
|
AMDGPU/GlobalISel: Move SMRD selection logic to TableGen
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: volkan, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52922
llvm-svn: 354516
2019-02-20 22:02:37 +01:00
|
|
|
return None;
|
|
|
|
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(PtrReg); },
|
2020-01-31 00:41:10 +01:00
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(*EncodedImm); }
|
AMDGPU/GlobalISel: Move SMRD selection logic to TableGen
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: volkan, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52922
llvm-svn: 354516
2019-02-20 22:02:37 +01:00
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectSmrdSgpr(MachineOperand &Root) const {
|
|
|
|
MachineInstr *MI = Root.getParent();
|
|
|
|
MachineBasicBlock *MBB = MI->getParent();
|
|
|
|
|
|
|
|
SmallVector<GEPInfo, 4> AddrInfo;
|
2019-09-28 05:41:13 +02:00
|
|
|
getAddrModeInfo(*MI, *MRI, AddrInfo);
|
AMDGPU/GlobalISel: Move SMRD selection logic to TableGen
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: volkan, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52922
llvm-svn: 354516
2019-02-20 22:02:37 +01:00
|
|
|
|
|
|
|
// FIXME: We should shrink the GEP if the offset is known to be <= 32-bits,
|
|
|
|
// then we can select all ptr + 32-bit offsets not just immediate offsets.
|
|
|
|
if (AddrInfo.empty() || AddrInfo[0].SgprParts.size() != 1)
|
|
|
|
return None;
|
|
|
|
|
|
|
|
const GEPInfo &GEPInfo = AddrInfo[0];
|
2020-03-30 08:32:20 +02:00
|
|
|
// SGPR offset is unsigned.
|
|
|
|
if (!GEPInfo.Imm || GEPInfo.Imm < 0 || !isUInt<32>(GEPInfo.Imm))
|
AMDGPU/GlobalISel: Move SMRD selection logic to TableGen
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: volkan, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52922
llvm-svn: 354516
2019-02-20 22:02:37 +01:00
|
|
|
return None;
|
|
|
|
|
|
|
|
// If we make it this far we have a load with an 32-bit immediate offset.
|
|
|
|
// It is OK to select this using a sgpr offset, because we have already
|
|
|
|
// failed trying to select this load into one of the _IMM variants since
|
|
|
|
// the _IMM Patterns are considered before the _SGPR patterns.
|
2020-04-21 21:06:53 +02:00
|
|
|
Register PtrReg = GEPInfo.SgprParts[0];
|
2019-10-18 20:26:37 +02:00
|
|
|
Register OffsetReg = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
AMDGPU/GlobalISel: Move SMRD selection logic to TableGen
Reviewers: arsenm
Reviewed By: arsenm
Subscribers: volkan, kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, kristof.beyls, dstuttard, tpr, t-tye, llvm-commits
Differential Revision: https://reviews.llvm.org/D52922
llvm-svn: 354516
2019-02-20 22:02:37 +01:00
|
|
|
BuildMI(*MBB, MI, MI->getDebugLoc(), TII.get(AMDGPU::S_MOV_B32), OffsetReg)
|
|
|
|
.addImm(GEPInfo.Imm);
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(PtrReg); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(OffsetReg); }
|
|
|
|
}};
|
|
|
|
}
|
2019-07-16 20:05:29 +02:00
|
|
|
|
2020-09-11 00:10:52 +02:00
|
|
|
std::pair<Register, int>
|
2021-03-30 14:38:07 +02:00
|
|
|
AMDGPUInstructionSelector::selectFlatOffsetImpl(MachineOperand &Root,
|
|
|
|
uint64_t FlatVariant) const {
|
2019-07-16 20:05:29 +02:00
|
|
|
MachineInstr *MI = Root.getParent();
|
|
|
|
|
2020-09-11 00:10:52 +02:00
|
|
|
auto Default = std::make_pair(Root.getReg(), 0);
|
2019-07-16 20:05:29 +02:00
|
|
|
|
|
|
|
if (!STI.hasFlatInstOffsets())
|
|
|
|
return Default;
|
|
|
|
|
2020-12-22 13:53:38 +01:00
|
|
|
Register PtrBase;
|
|
|
|
int64_t ConstOffset;
|
|
|
|
std::tie(PtrBase, ConstOffset) =
|
|
|
|
getPtrBaseWithConstantOffset(Root.getReg(), *MRI);
|
|
|
|
if (ConstOffset == 0)
|
2019-07-16 20:05:29 +02:00
|
|
|
return Default;
|
|
|
|
|
|
|
|
unsigned AddrSpace = (*MI->memoperands_begin())->getAddrSpace();
|
2021-03-30 14:38:07 +02:00
|
|
|
if (!TII.isLegalFLATOffset(ConstOffset, AddrSpace, FlatVariant))
|
2019-07-16 20:05:29 +02:00
|
|
|
return Default;
|
|
|
|
|
2020-12-22 13:53:38 +01:00
|
|
|
return std::make_pair(PtrBase, ConstOffset);
|
2019-07-16 20:05:29 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectFlatOffset(MachineOperand &Root) const {
|
2021-03-30 14:38:07 +02:00
|
|
|
auto PtrWithOffset = selectFlatOffsetImpl(Root, SIInstrFlags::FLAT);
|
2020-09-11 00:10:52 +02:00
|
|
|
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(PtrWithOffset.first); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(PtrWithOffset.second); },
|
|
|
|
}};
|
2019-07-16 20:05:29 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
2021-03-30 14:38:07 +02:00
|
|
|
AMDGPUInstructionSelector::selectGlobalOffset(MachineOperand &Root) const {
|
|
|
|
auto PtrWithOffset = selectFlatOffsetImpl(Root, SIInstrFlags::FlatGlobal);
|
|
|
|
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(PtrWithOffset.first); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(PtrWithOffset.second); },
|
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectScratchOffset(MachineOperand &Root) const {
|
|
|
|
auto PtrWithOffset = selectFlatOffsetImpl(Root, SIInstrFlags::FlatScratch);
|
2020-09-11 00:10:52 +02:00
|
|
|
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(PtrWithOffset.first); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(PtrWithOffset.second); },
|
|
|
|
}};
|
2019-07-16 20:05:29 +02:00
|
|
|
}
|
2019-07-16 21:22:21 +02:00
|
|
|
|
2020-08-16 17:18:52 +02:00
|
|
|
/// Match a zero extend from a 32-bit value to 64-bits.
|
|
|
|
static Register matchZeroExtendFromS32(MachineRegisterInfo &MRI, Register Reg) {
|
|
|
|
Register ZExtSrc;
|
|
|
|
if (mi_match(Reg, MRI, m_GZExt(m_Reg(ZExtSrc))))
|
|
|
|
return MRI.getType(ZExtSrc) == LLT::scalar(32) ? ZExtSrc : Register();
|
|
|
|
|
|
|
|
// Match legalized form %zext = G_MERGE_VALUES (s32 %x), (s32 0)
|
|
|
|
const MachineInstr *Def = getDefIgnoringCopies(Reg, MRI);
|
|
|
|
if (Def->getOpcode() != AMDGPU::G_MERGE_VALUES)
|
|
|
|
return false;
|
|
|
|
|
2020-11-13 01:42:32 +01:00
|
|
|
if (mi_match(Def->getOperand(2).getReg(), MRI, m_ZeroInt())) {
|
2020-08-16 17:18:52 +02:00
|
|
|
return Def->getOperand(1).getReg();
|
|
|
|
}
|
|
|
|
|
|
|
|
return Register();
|
|
|
|
}
|
|
|
|
|
|
|
|
// Match (64-bit SGPR base) + (zext vgpr offset) + sext(imm offset)
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectGlobalSAddr(MachineOperand &Root) const {
|
2020-11-10 00:01:29 +01:00
|
|
|
Register Addr = Root.getReg();
|
2020-08-16 17:18:52 +02:00
|
|
|
Register PtrBase;
|
2020-11-10 00:01:29 +01:00
|
|
|
int64_t ConstOffset;
|
|
|
|
int64_t ImmOffset = 0;
|
2020-08-16 17:18:52 +02:00
|
|
|
|
|
|
|
// Match the immediate offset first, which canonically is moved as low as
|
|
|
|
// possible.
|
2020-11-10 00:01:29 +01:00
|
|
|
std::tie(PtrBase, ConstOffset) = getPtrBaseWithConstantOffset(Addr, *MRI);
|
|
|
|
|
|
|
|
if (ConstOffset != 0) {
|
2021-03-30 14:38:07 +02:00
|
|
|
if (TII.isLegalFLATOffset(ConstOffset, AMDGPUAS::GLOBAL_ADDRESS,
|
|
|
|
SIInstrFlags::FlatGlobal)) {
|
2020-11-10 00:01:29 +01:00
|
|
|
Addr = PtrBase;
|
|
|
|
ImmOffset = ConstOffset;
|
2021-05-03 20:01:13 +02:00
|
|
|
} else {
|
2020-11-10 00:01:29 +01:00
|
|
|
auto PtrBaseDef = getDefSrcRegIgnoringCopies(PtrBase, *MRI);
|
|
|
|
if (!PtrBaseDef)
|
|
|
|
return None;
|
|
|
|
|
|
|
|
if (isSGPR(PtrBaseDef->Reg)) {
|
2021-05-03 20:01:13 +02:00
|
|
|
if (ConstOffset > 0) {
|
|
|
|
// Offset is too large.
|
|
|
|
//
|
|
|
|
// saddr + large_offset -> saddr +
|
|
|
|
// (voffset = large_offset & ~MaxOffset) +
|
|
|
|
// (large_offset & MaxOffset);
|
|
|
|
int64_t SplitImmOffset, RemainderOffset;
|
|
|
|
std::tie(SplitImmOffset, RemainderOffset) = TII.splitFlatOffset(
|
|
|
|
ConstOffset, AMDGPUAS::GLOBAL_ADDRESS, SIInstrFlags::FlatGlobal);
|
|
|
|
|
|
|
|
if (isUInt<32>(RemainderOffset)) {
|
|
|
|
MachineInstr *MI = Root.getParent();
|
|
|
|
MachineBasicBlock *MBB = MI->getParent();
|
|
|
|
Register HighBits =
|
|
|
|
MRI->createVirtualRegister(&AMDGPU::VGPR_32RegClass);
|
|
|
|
|
|
|
|
BuildMI(*MBB, MI, MI->getDebugLoc(), TII.get(AMDGPU::V_MOV_B32_e32),
|
|
|
|
HighBits)
|
|
|
|
.addImm(RemainderOffset);
|
|
|
|
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(PtrBase); }, // saddr
|
|
|
|
[=](MachineInstrBuilder &MIB) {
|
|
|
|
MIB.addReg(HighBits);
|
|
|
|
}, // voffset
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(SplitImmOffset); },
|
|
|
|
}};
|
|
|
|
}
|
2020-11-10 00:01:29 +01:00
|
|
|
}
|
2021-05-03 20:01:13 +02:00
|
|
|
|
|
|
|
// We are adding a 64 bit SGPR and a constant. If constant bus limit
|
|
|
|
// is 1 we would need to perform 1 or 2 extra moves for each half of
|
|
|
|
// the constant and it is better to do a scalar add and then issue a
|
|
|
|
// single VALU instruction to materialize zero. Otherwise it is less
|
|
|
|
// instructions to perform VALU adds with immediates or inline literals.
|
|
|
|
unsigned NumLiterals =
|
|
|
|
!TII.isInlineConstant(APInt(32, ConstOffset & 0xffffffff)) +
|
|
|
|
!TII.isInlineConstant(APInt(32, ConstOffset >> 32));
|
|
|
|
if (STI.getConstantBusLimit(AMDGPU::V_ADD_U32_e64) > NumLiterals)
|
|
|
|
return None;
|
2020-11-10 00:01:29 +01:00
|
|
|
}
|
|
|
|
}
|
2020-08-16 17:18:52 +02:00
|
|
|
}
|
|
|
|
|
2020-11-10 00:01:29 +01:00
|
|
|
auto AddrDef = getDefSrcRegIgnoringCopies(Addr, *MRI);
|
|
|
|
if (!AddrDef)
|
|
|
|
return None;
|
|
|
|
|
2020-08-16 17:18:52 +02:00
|
|
|
// Match the variable offset.
|
2021-05-03 20:01:13 +02:00
|
|
|
if (AddrDef->MI->getOpcode() == AMDGPU::G_PTR_ADD) {
|
|
|
|
// Look through the SGPR->VGPR copy.
|
|
|
|
Register SAddr =
|
|
|
|
getSrcRegIgnoringCopies(AddrDef->MI->getOperand(1).getReg(), *MRI);
|
|
|
|
|
|
|
|
if (SAddr && isSGPR(SAddr)) {
|
|
|
|
Register PtrBaseOffset = AddrDef->MI->getOperand(2).getReg();
|
|
|
|
|
|
|
|
// It's possible voffset is an SGPR here, but the copy to VGPR will be
|
|
|
|
// inserted later.
|
|
|
|
if (Register VOffset = matchZeroExtendFromS32(*MRI, PtrBaseOffset)) {
|
|
|
|
return {{[=](MachineInstrBuilder &MIB) { // saddr
|
|
|
|
MIB.addReg(SAddr);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // voffset
|
|
|
|
MIB.addReg(VOffset);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // offset
|
|
|
|
MIB.addImm(ImmOffset);
|
|
|
|
}}};
|
|
|
|
}
|
|
|
|
}
|
2020-11-10 17:06:59 +01:00
|
|
|
}
|
2020-08-16 17:18:52 +02:00
|
|
|
|
2021-05-03 20:01:13 +02:00
|
|
|
// FIXME: We should probably have folded COPY (G_IMPLICIT_DEF) earlier, and
|
|
|
|
// drop this.
|
|
|
|
if (AddrDef->MI->getOpcode() == AMDGPU::G_IMPLICIT_DEF ||
|
|
|
|
AddrDef->MI->getOpcode() == AMDGPU::G_CONSTANT || !isSGPR(AddrDef->Reg))
|
2020-08-16 17:18:52 +02:00
|
|
|
return None;
|
|
|
|
|
2021-05-03 20:01:13 +02:00
|
|
|
// It's cheaper to materialize a single 32-bit zero for vaddr than the two
|
|
|
|
// moves required to copy a 64-bit SGPR to VGPR.
|
|
|
|
MachineInstr *MI = Root.getParent();
|
|
|
|
MachineBasicBlock *MBB = MI->getParent();
|
|
|
|
Register VOffset = MRI->createVirtualRegister(&AMDGPU::VGPR_32RegClass);
|
2020-08-16 17:18:52 +02:00
|
|
|
|
2021-05-03 20:01:13 +02:00
|
|
|
BuildMI(*MBB, MI, MI->getDebugLoc(), TII.get(AMDGPU::V_MOV_B32_e32), VOffset)
|
|
|
|
.addImm(0);
|
2020-08-16 17:18:52 +02:00
|
|
|
|
2021-05-03 20:01:13 +02:00
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(AddrDef->Reg); }, // saddr
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(VOffset); }, // voffset
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(ImmOffset); } // offset
|
|
|
|
}};
|
2020-08-16 17:18:52 +02:00
|
|
|
}
|
|
|
|
|
2020-12-21 21:31:06 +01:00
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectScratchSAddr(MachineOperand &Root) const {
|
|
|
|
Register Addr = Root.getReg();
|
|
|
|
Register PtrBase;
|
|
|
|
int64_t ConstOffset;
|
|
|
|
int64_t ImmOffset = 0;
|
|
|
|
|
|
|
|
// Match the immediate offset first, which canonically is moved as low as
|
|
|
|
// possible.
|
|
|
|
std::tie(PtrBase, ConstOffset) = getPtrBaseWithConstantOffset(Addr, *MRI);
|
|
|
|
|
|
|
|
if (ConstOffset != 0 &&
|
2021-03-30 14:38:07 +02:00
|
|
|
TII.isLegalFLATOffset(ConstOffset, AMDGPUAS::PRIVATE_ADDRESS,
|
|
|
|
SIInstrFlags::FlatScratch)) {
|
2020-12-21 21:31:06 +01:00
|
|
|
Addr = PtrBase;
|
|
|
|
ImmOffset = ConstOffset;
|
|
|
|
}
|
|
|
|
|
|
|
|
auto AddrDef = getDefSrcRegIgnoringCopies(Addr, *MRI);
|
|
|
|
if (!AddrDef)
|
|
|
|
return None;
|
|
|
|
|
|
|
|
if (AddrDef->MI->getOpcode() == AMDGPU::G_FRAME_INDEX) {
|
|
|
|
int FI = AddrDef->MI->getOperand(1).getIndex();
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addFrameIndex(FI); }, // saddr
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(ImmOffset); } // offset
|
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
|
|
|
Register SAddr = AddrDef->Reg;
|
|
|
|
|
|
|
|
if (AddrDef->MI->getOpcode() == AMDGPU::G_PTR_ADD) {
|
|
|
|
Register LHS = AddrDef->MI->getOperand(1).getReg();
|
|
|
|
Register RHS = AddrDef->MI->getOperand(2).getReg();
|
|
|
|
auto LHSDef = getDefSrcRegIgnoringCopies(LHS, *MRI);
|
|
|
|
auto RHSDef = getDefSrcRegIgnoringCopies(RHS, *MRI);
|
|
|
|
|
|
|
|
if (LHSDef && RHSDef &&
|
|
|
|
LHSDef->MI->getOpcode() == AMDGPU::G_FRAME_INDEX &&
|
|
|
|
isSGPR(RHSDef->Reg)) {
|
|
|
|
int FI = LHSDef->MI->getOperand(1).getIndex();
|
|
|
|
MachineInstr &I = *Root.getParent();
|
|
|
|
MachineBasicBlock *BB = I.getParent();
|
|
|
|
const DebugLoc &DL = I.getDebugLoc();
|
|
|
|
SAddr = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
|
2021-06-07 16:09:48 +02:00
|
|
|
BuildMI(*BB, &I, DL, TII.get(AMDGPU::S_ADD_I32), SAddr)
|
|
|
|
.addFrameIndex(FI)
|
|
|
|
.addReg(RHSDef->Reg);
|
2020-12-21 21:31:06 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!isSGPR(SAddr))
|
|
|
|
return None;
|
|
|
|
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(SAddr); }, // saddr
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(ImmOffset); } // offset
|
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
2019-07-16 21:22:21 +02:00
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectMUBUFScratchOffen(MachineOperand &Root) const {
|
|
|
|
MachineInstr *MI = Root.getParent();
|
|
|
|
MachineBasicBlock *MBB = MI->getParent();
|
|
|
|
MachineFunction *MF = MBB->getParent();
|
|
|
|
const SIMachineFunctionInfo *Info = MF->getInfo<SIMachineFunctionInfo>();
|
|
|
|
|
|
|
|
int64_t Offset = 0;
|
2020-05-24 18:28:59 +02:00
|
|
|
if (mi_match(Root.getReg(), *MRI, m_ICst(Offset)) &&
|
|
|
|
Offset != TM.getNullPointerValue(AMDGPUAS::PRIVATE_ADDRESS)) {
|
2019-09-28 05:41:13 +02:00
|
|
|
Register HighBits = MRI->createVirtualRegister(&AMDGPU::VGPR_32RegClass);
|
2019-07-16 21:22:21 +02:00
|
|
|
|
|
|
|
// TODO: Should this be inside the render function? The iterator seems to
|
|
|
|
// move.
|
|
|
|
BuildMI(*MBB, MI, MI->getDebugLoc(), TII.get(AMDGPU::V_MOV_B32_e32),
|
|
|
|
HighBits)
|
|
|
|
.addImm(Offset & ~4095);
|
|
|
|
|
|
|
|
return {{[=](MachineInstrBuilder &MIB) { // rsrc
|
|
|
|
MIB.addReg(Info->getScratchRSrcReg());
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // vaddr
|
|
|
|
MIB.addReg(HighBits);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // soffset
|
2021-01-15 11:32:29 +01:00
|
|
|
// Use constant zero for soffset and rely on eliminateFrameIndex
|
|
|
|
// to choose the appropriate frame register if need be.
|
|
|
|
MIB.addImm(0);
|
2019-07-16 21:22:21 +02:00
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // offset
|
|
|
|
MIB.addImm(Offset & 4095);
|
|
|
|
}}};
|
|
|
|
}
|
|
|
|
|
2020-05-24 18:28:59 +02:00
|
|
|
assert(Offset == 0 || Offset == -1);
|
2019-07-16 21:22:21 +02:00
|
|
|
|
|
|
|
// Try to fold a frame index directly into the MUBUF vaddr field, and any
|
|
|
|
// offsets.
|
|
|
|
Optional<int> FI;
|
|
|
|
Register VAddr = Root.getReg();
|
2019-09-28 05:41:13 +02:00
|
|
|
if (const MachineInstr *RootDef = MRI->getVRegDef(Root.getReg())) {
|
2021-03-04 23:19:49 +01:00
|
|
|
Register PtrBase;
|
|
|
|
int64_t ConstOffset;
|
|
|
|
std::tie(PtrBase, ConstOffset) = getPtrBaseWithConstantOffset(VAddr, *MRI);
|
|
|
|
if (ConstOffset != 0) {
|
|
|
|
if (SIInstrInfo::isLegalMUBUFImmOffset(ConstOffset) &&
|
|
|
|
(!STI.privateMemoryResourceIsRangeChecked() ||
|
|
|
|
KnownBits->signBitIsZero(PtrBase))) {
|
|
|
|
const MachineInstr *PtrBaseDef = MRI->getVRegDef(PtrBase);
|
|
|
|
if (PtrBaseDef->getOpcode() == AMDGPU::G_FRAME_INDEX)
|
|
|
|
FI = PtrBaseDef->getOperand(1).getIndex();
|
|
|
|
else
|
|
|
|
VAddr = PtrBase;
|
|
|
|
Offset = ConstOffset;
|
2019-07-16 21:22:21 +02:00
|
|
|
}
|
|
|
|
} else if (RootDef->getOpcode() == AMDGPU::G_FRAME_INDEX) {
|
|
|
|
FI = RootDef->getOperand(1).getIndex();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return {{[=](MachineInstrBuilder &MIB) { // rsrc
|
|
|
|
MIB.addReg(Info->getScratchRSrcReg());
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // vaddr
|
|
|
|
if (FI.hasValue())
|
|
|
|
MIB.addFrameIndex(FI.getValue());
|
|
|
|
else
|
|
|
|
MIB.addReg(VAddr);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // soffset
|
2021-01-15 11:32:29 +01:00
|
|
|
// Use constant zero for soffset and rely on eliminateFrameIndex
|
|
|
|
// to choose the appropriate frame register if need be.
|
|
|
|
MIB.addImm(0);
|
2019-07-16 21:22:21 +02:00
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // offset
|
|
|
|
MIB.addImm(Offset);
|
|
|
|
}}};
|
|
|
|
}
|
|
|
|
|
2020-01-17 04:27:31 +01:00
|
|
|
bool AMDGPUInstructionSelector::isDSOffsetLegal(Register Base,
|
2020-11-02 14:05:15 +01:00
|
|
|
int64_t Offset) const {
|
|
|
|
if (!isUInt<16>(Offset))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (STI.hasUsableDSOffset() || STI.unsafeDSOffsetFoldingEnabled())
|
|
|
|
return true;
|
|
|
|
|
|
|
|
// On Southern Islands instruction with a negative base value and an offset
|
|
|
|
// don't seem to work.
|
|
|
|
return KnownBits->signBitIsZero(Base);
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AMDGPUInstructionSelector::isDSOffset2Legal(Register Base, int64_t Offset0,
|
|
|
|
int64_t Offset1,
|
|
|
|
unsigned Size) const {
|
|
|
|
if (Offset0 % Size != 0 || Offset1 % Size != 0)
|
|
|
|
return false;
|
|
|
|
if (!isUInt<8>(Offset0 / Size) || !isUInt<8>(Offset1 / Size))
|
2019-08-01 02:53:38 +02:00
|
|
|
return false;
|
|
|
|
|
|
|
|
if (STI.hasUsableDSOffset() || STI.unsafeDSOffsetFoldingEnabled())
|
|
|
|
return true;
|
|
|
|
|
|
|
|
// On Southern Islands instruction with a negative base value and an offset
|
|
|
|
// don't seem to work.
|
2020-01-17 04:27:31 +01:00
|
|
|
return KnownBits->signBitIsZero(Base);
|
2019-08-01 02:53:38 +02:00
|
|
|
}
|
|
|
|
|
2019-07-16 21:22:21 +02:00
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectMUBUFScratchOffset(
|
|
|
|
MachineOperand &Root) const {
|
|
|
|
MachineInstr *MI = Root.getParent();
|
|
|
|
MachineBasicBlock *MBB = MI->getParent();
|
|
|
|
|
|
|
|
int64_t Offset = 0;
|
2019-09-28 05:41:13 +02:00
|
|
|
if (!mi_match(Root.getReg(), *MRI, m_ICst(Offset)) ||
|
2019-07-16 21:22:21 +02:00
|
|
|
!SIInstrInfo::isLegalMUBUFImmOffset(Offset))
|
|
|
|
return {};
|
|
|
|
|
|
|
|
const MachineFunction *MF = MBB->getParent();
|
|
|
|
const SIMachineFunctionInfo *Info = MF->getInfo<SIMachineFunctionInfo>();
|
|
|
|
|
|
|
|
return {{
|
2020-01-21 23:27:57 +01:00
|
|
|
[=](MachineInstrBuilder &MIB) { // rsrc
|
2019-07-16 21:22:21 +02:00
|
|
|
MIB.addReg(Info->getScratchRSrcReg());
|
2020-01-21 23:27:57 +01:00
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // soffset
|
2021-05-27 18:32:42 +02:00
|
|
|
MIB.addImm(0);
|
2020-01-21 23:27:57 +01:00
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(Offset); } // offset
|
2019-07-16 21:22:21 +02:00
|
|
|
}};
|
|
|
|
}
|
2019-08-01 02:53:38 +02:00
|
|
|
|
2020-01-17 04:27:31 +01:00
|
|
|
std::pair<Register, unsigned>
|
|
|
|
AMDGPUInstructionSelector::selectDS1Addr1OffsetImpl(MachineOperand &Root) const {
|
2019-09-28 05:41:13 +02:00
|
|
|
const MachineInstr *RootDef = MRI->getVRegDef(Root.getReg());
|
2020-01-17 04:27:31 +01:00
|
|
|
if (!RootDef)
|
|
|
|
return std::make_pair(Root.getReg(), 0);
|
2019-08-01 02:53:38 +02:00
|
|
|
|
|
|
|
int64_t ConstAddr = 0;
|
2020-02-07 17:29:23 +01:00
|
|
|
|
|
|
|
Register PtrBase;
|
|
|
|
int64_t Offset;
|
|
|
|
std::tie(PtrBase, Offset) =
|
|
|
|
getPtrBaseWithConstantOffset(Root.getReg(), *MRI);
|
|
|
|
|
|
|
|
if (Offset) {
|
2020-11-02 14:05:15 +01:00
|
|
|
if (isDSOffsetLegal(PtrBase, Offset)) {
|
2020-02-07 17:29:23 +01:00
|
|
|
// (add n0, c0)
|
|
|
|
return std::make_pair(PtrBase, Offset);
|
2019-08-01 02:53:38 +02:00
|
|
|
}
|
|
|
|
} else if (RootDef->getOpcode() == AMDGPU::G_SUB) {
|
2020-01-17 04:27:31 +01:00
|
|
|
// TODO
|
2019-08-01 02:53:38 +02:00
|
|
|
|
|
|
|
|
2019-09-28 05:41:13 +02:00
|
|
|
} else if (mi_match(Root.getReg(), *MRI, m_ICst(ConstAddr))) {
|
2020-01-17 04:27:31 +01:00
|
|
|
// TODO
|
2019-08-01 02:53:38 +02:00
|
|
|
|
|
|
|
}
|
|
|
|
|
2020-01-17 04:27:31 +01:00
|
|
|
return std::make_pair(Root.getReg(), 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectDS1Addr1Offset(MachineOperand &Root) const {
|
|
|
|
Register Reg;
|
|
|
|
unsigned Offset;
|
|
|
|
std::tie(Reg, Offset) = selectDS1Addr1OffsetImpl(Root);
|
2019-08-01 02:53:38 +02:00
|
|
|
return {{
|
2020-01-17 04:27:31 +01:00
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(Reg); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(Offset); }
|
2019-08-01 02:53:38 +02:00
|
|
|
}};
|
|
|
|
}
|
2019-10-01 04:07:25 +02:00
|
|
|
|
[AMDGPU][GlobalISel] Select 8-byte LDS Ops with 4-byte alignment
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D73585
2020-01-07 22:13:05 +01:00
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectDS64Bit4ByteAligned(MachineOperand &Root) const {
|
2020-11-02 14:05:15 +01:00
|
|
|
return selectDSReadWrite2(Root, 4);
|
2020-08-21 11:42:49 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectDS128Bit8ByteAligned(MachineOperand &Root) const {
|
2020-11-02 14:05:15 +01:00
|
|
|
return selectDSReadWrite2(Root, 8);
|
2020-08-21 11:42:49 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectDSReadWrite2(MachineOperand &Root,
|
2020-11-02 14:05:15 +01:00
|
|
|
unsigned Size) const {
|
2020-02-07 17:29:23 +01:00
|
|
|
Register Reg;
|
|
|
|
unsigned Offset;
|
2020-11-02 14:05:15 +01:00
|
|
|
std::tie(Reg, Offset) = selectDSReadWrite2Impl(Root, Size);
|
2020-02-07 17:29:23 +01:00
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addReg(Reg); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(Offset); },
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(Offset+1); }
|
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
|
|
|
std::pair<Register, unsigned>
|
2020-08-21 11:42:49 +02:00
|
|
|
AMDGPUInstructionSelector::selectDSReadWrite2Impl(MachineOperand &Root,
|
2020-11-02 14:05:15 +01:00
|
|
|
unsigned Size) const {
|
[AMDGPU][GlobalISel] Select 8-byte LDS Ops with 4-byte alignment
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D73585
2020-01-07 22:13:05 +01:00
|
|
|
const MachineInstr *RootDef = MRI->getVRegDef(Root.getReg());
|
2020-02-07 17:29:23 +01:00
|
|
|
if (!RootDef)
|
|
|
|
return std::make_pair(Root.getReg(), 0);
|
[AMDGPU][GlobalISel] Select 8-byte LDS Ops with 4-byte alignment
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D73585
2020-01-07 22:13:05 +01:00
|
|
|
|
|
|
|
int64_t ConstAddr = 0;
|
2020-02-07 17:29:23 +01:00
|
|
|
|
[AMDGPU][GlobalISel] Select 8-byte LDS Ops with 4-byte alignment
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D73585
2020-01-07 22:13:05 +01:00
|
|
|
Register PtrBase;
|
|
|
|
int64_t Offset;
|
|
|
|
std::tie(PtrBase, Offset) =
|
|
|
|
getPtrBaseWithConstantOffset(Root.getReg(), *MRI);
|
|
|
|
|
|
|
|
if (Offset) {
|
2020-11-02 14:05:15 +01:00
|
|
|
int64_t OffsetValue0 = Offset;
|
|
|
|
int64_t OffsetValue1 = Offset + Size;
|
|
|
|
if (isDSOffset2Legal(PtrBase, OffsetValue0, OffsetValue1, Size)) {
|
[AMDGPU][GlobalISel] Select 8-byte LDS Ops with 4-byte alignment
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D73585
2020-01-07 22:13:05 +01:00
|
|
|
// (add n0, c0)
|
2020-11-02 14:05:15 +01:00
|
|
|
return std::make_pair(PtrBase, OffsetValue0 / Size);
|
[AMDGPU][GlobalISel] Select 8-byte LDS Ops with 4-byte alignment
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D73585
2020-01-07 22:13:05 +01:00
|
|
|
}
|
|
|
|
} else if (RootDef->getOpcode() == AMDGPU::G_SUB) {
|
|
|
|
// TODO
|
|
|
|
|
|
|
|
} else if (mi_match(Root.getReg(), *MRI, m_ICst(ConstAddr))) {
|
|
|
|
// TODO
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2020-02-07 17:29:23 +01:00
|
|
|
return std::make_pair(Root.getReg(), 0);
|
[AMDGPU][GlobalISel] Select 8-byte LDS Ops with 4-byte alignment
Reviewers: arsenm
Subscribers: kzhuravl, jvesely, wdng, nhaehnle, yaxunl, rovka, dstuttard, tpr, t-tye, hiraditya, Petar.Avramovic, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D73585
2020-01-07 22:13:05 +01:00
|
|
|
}
|
|
|
|
|
2020-01-18 03:28:58 +01:00
|
|
|
/// If \p Root is a G_PTR_ADD with a G_CONSTANT on the right hand side, return
|
|
|
|
/// the base value with the constant offset. There may be intervening copies
|
|
|
|
/// between \p Root and the identified constant. Returns \p Root, 0 if this does
|
|
|
|
/// not match the pattern.
|
|
|
|
std::pair<Register, int64_t>
|
|
|
|
AMDGPUInstructionSelector::getPtrBaseWithConstantOffset(
|
|
|
|
Register Root, const MachineRegisterInfo &MRI) const {
|
2020-08-15 17:41:08 +02:00
|
|
|
MachineInstr *RootI = getDefIgnoringCopies(Root, MRI);
|
2020-01-18 03:28:58 +01:00
|
|
|
if (RootI->getOpcode() != TargetOpcode::G_PTR_ADD)
|
|
|
|
return {Root, 0};
|
|
|
|
|
|
|
|
MachineOperand &RHS = RootI->getOperand(2);
|
|
|
|
Optional<ValueAndVReg> MaybeOffset
|
|
|
|
= getConstantVRegValWithLookThrough(RHS.getReg(), MRI, true);
|
|
|
|
if (!MaybeOffset)
|
|
|
|
return {Root, 0};
|
2020-11-03 15:50:17 +01:00
|
|
|
return {RootI->getOperand(1).getReg(), MaybeOffset->Value.getSExtValue()};
|
2020-01-18 03:28:58 +01:00
|
|
|
}
|
|
|
|
|
2020-01-22 22:56:48 +01:00
|
|
|
static void addZeroImm(MachineInstrBuilder &MIB) {
|
|
|
|
MIB.addImm(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Return a resource descriptor for use with an arbitrary 64-bit pointer. If \p
|
2020-01-18 03:28:58 +01:00
|
|
|
/// BasePtr is not valid, a null base pointer will be used.
|
|
|
|
static Register buildRSRC(MachineIRBuilder &B, MachineRegisterInfo &MRI,
|
|
|
|
uint32_t FormatLo, uint32_t FormatHi,
|
|
|
|
Register BasePtr) {
|
2020-01-22 22:56:48 +01:00
|
|
|
Register RSrc2 = MRI.createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
Register RSrc3 = MRI.createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
Register RSrcHi = MRI.createVirtualRegister(&AMDGPU::SReg_64RegClass);
|
|
|
|
Register RSrc = MRI.createVirtualRegister(&AMDGPU::SGPR_128RegClass);
|
|
|
|
|
2020-01-18 03:28:58 +01:00
|
|
|
B.buildInstr(AMDGPU::S_MOV_B32)
|
|
|
|
.addDef(RSrc2)
|
|
|
|
.addImm(FormatLo);
|
|
|
|
B.buildInstr(AMDGPU::S_MOV_B32)
|
|
|
|
.addDef(RSrc3)
|
|
|
|
.addImm(FormatHi);
|
2020-01-22 22:56:48 +01:00
|
|
|
|
|
|
|
// Build the half of the subregister with the constants before building the
|
|
|
|
// full 128-bit register. If we are building multiple resource descriptors,
|
|
|
|
// this will allow CSEing of the 2-component register.
|
2020-01-18 03:28:58 +01:00
|
|
|
B.buildInstr(AMDGPU::REG_SEQUENCE)
|
|
|
|
.addDef(RSrcHi)
|
2020-01-22 22:56:48 +01:00
|
|
|
.addReg(RSrc2)
|
|
|
|
.addImm(AMDGPU::sub0)
|
|
|
|
.addReg(RSrc3)
|
|
|
|
.addImm(AMDGPU::sub1);
|
|
|
|
|
|
|
|
Register RSrcLo = BasePtr;
|
|
|
|
if (!BasePtr) {
|
|
|
|
RSrcLo = MRI.createVirtualRegister(&AMDGPU::SReg_64RegClass);
|
2020-01-18 03:28:58 +01:00
|
|
|
B.buildInstr(AMDGPU::S_MOV_B64)
|
|
|
|
.addDef(RSrcLo)
|
2020-01-22 22:56:48 +01:00
|
|
|
.addImm(0);
|
|
|
|
}
|
|
|
|
|
2020-01-18 03:28:58 +01:00
|
|
|
B.buildInstr(AMDGPU::REG_SEQUENCE)
|
|
|
|
.addDef(RSrc)
|
2020-01-22 22:56:48 +01:00
|
|
|
.addReg(RSrcLo)
|
|
|
|
.addImm(AMDGPU::sub0_sub1)
|
|
|
|
.addReg(RSrcHi)
|
|
|
|
.addImm(AMDGPU::sub2_sub3);
|
|
|
|
|
|
|
|
return RSrc;
|
|
|
|
}
|
|
|
|
|
2020-01-18 03:28:58 +01:00
|
|
|
static Register buildAddr64RSrc(MachineIRBuilder &B, MachineRegisterInfo &MRI,
|
|
|
|
const SIInstrInfo &TII, Register BasePtr) {
|
|
|
|
uint64_t DefaultFormat = TII.getDefaultRsrcDataFormat();
|
|
|
|
|
|
|
|
// FIXME: Why are half the "default" bits ignored based on the addressing
|
|
|
|
// mode?
|
|
|
|
return buildRSRC(B, MRI, 0, Hi_32(DefaultFormat), BasePtr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static Register buildOffsetSrc(MachineIRBuilder &B, MachineRegisterInfo &MRI,
|
|
|
|
const SIInstrInfo &TII, Register BasePtr) {
|
|
|
|
uint64_t DefaultFormat = TII.getDefaultRsrcDataFormat();
|
|
|
|
|
|
|
|
// FIXME: Why are half the "default" bits ignored based on the addressing
|
|
|
|
// mode?
|
|
|
|
return buildRSRC(B, MRI, -1, Hi_32(DefaultFormat), BasePtr);
|
|
|
|
}
|
|
|
|
|
|
|
|
AMDGPUInstructionSelector::MUBUFAddressData
|
|
|
|
AMDGPUInstructionSelector::parseMUBUFAddress(Register Src) const {
|
|
|
|
MUBUFAddressData Data;
|
|
|
|
Data.N0 = Src;
|
|
|
|
|
|
|
|
Register PtrBase;
|
|
|
|
int64_t Offset;
|
|
|
|
|
|
|
|
std::tie(PtrBase, Offset) = getPtrBaseWithConstantOffset(Src, *MRI);
|
|
|
|
if (isUInt<32>(Offset)) {
|
|
|
|
Data.N0 = PtrBase;
|
|
|
|
Data.Offset = Offset;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (MachineInstr *InputAdd
|
|
|
|
= getOpcodeDef(TargetOpcode::G_PTR_ADD, Data.N0, *MRI)) {
|
|
|
|
Data.N2 = InputAdd->getOperand(1).getReg();
|
|
|
|
Data.N3 = InputAdd->getOperand(2).getReg();
|
|
|
|
|
|
|
|
// FIXME: Need to fix extra SGPR->VGPRcopies inserted
|
|
|
|
// FIXME: Don't know this was defined by operand 0
|
|
|
|
//
|
|
|
|
// TODO: Remove this when we have copy folding optimizations after
|
|
|
|
// RegBankSelect.
|
|
|
|
Data.N2 = getDefIgnoringCopies(Data.N2, *MRI)->getOperand(0).getReg();
|
|
|
|
Data.N3 = getDefIgnoringCopies(Data.N3, *MRI)->getOperand(0).getReg();
|
|
|
|
}
|
|
|
|
|
|
|
|
return Data;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Return if the addr64 mubuf mode should be used for the given address.
|
|
|
|
bool AMDGPUInstructionSelector::shouldUseAddr64(MUBUFAddressData Addr) const {
|
|
|
|
// (ptr_add N2, N3) -> addr64, or
|
|
|
|
// (ptr_add (ptr_add N2, N3), C1) -> addr64
|
|
|
|
if (Addr.N2)
|
|
|
|
return true;
|
|
|
|
|
|
|
|
const RegisterBank *N0Bank = RBI.getRegBank(Addr.N0, *MRI, TRI);
|
|
|
|
return N0Bank->getID() == AMDGPU::VGPRRegBankID;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Split an immediate offset \p ImmOffset depending on whether it fits in the
|
|
|
|
/// immediate field. Modifies \p ImmOffset and sets \p SOffset to the variable
|
|
|
|
/// component.
|
|
|
|
void AMDGPUInstructionSelector::splitIllegalMUBUFOffset(
|
|
|
|
MachineIRBuilder &B, Register &SOffset, int64_t &ImmOffset) const {
|
|
|
|
if (SIInstrInfo::isLegalMUBUFImmOffset(ImmOffset))
|
|
|
|
return;
|
|
|
|
|
|
|
|
// Illegal offset, store it in soffset.
|
|
|
|
SOffset = MRI->createVirtualRegister(&AMDGPU::SReg_32RegClass);
|
|
|
|
B.buildInstr(AMDGPU::S_MOV_B32)
|
|
|
|
.addDef(SOffset)
|
|
|
|
.addImm(ImmOffset);
|
|
|
|
ImmOffset = 0;
|
|
|
|
}
|
|
|
|
|
2020-01-24 17:45:33 +01:00
|
|
|
bool AMDGPUInstructionSelector::selectMUBUFAddr64Impl(
|
|
|
|
MachineOperand &Root, Register &VAddr, Register &RSrcReg,
|
|
|
|
Register &SOffset, int64_t &Offset) const {
|
2020-01-22 22:56:48 +01:00
|
|
|
// FIXME: Predicates should stop this from reaching here.
|
|
|
|
// addr64 bit was removed for volcanic islands.
|
|
|
|
if (!STI.hasAddr64() || STI.useFlatForGlobal())
|
2020-01-24 17:45:33 +01:00
|
|
|
return false;
|
2020-01-22 22:56:48 +01:00
|
|
|
|
2020-01-18 03:28:58 +01:00
|
|
|
MUBUFAddressData AddrData = parseMUBUFAddress(Root.getReg());
|
|
|
|
if (!shouldUseAddr64(AddrData))
|
2020-01-24 17:45:33 +01:00
|
|
|
return false;
|
2020-01-18 03:28:58 +01:00
|
|
|
|
|
|
|
Register N0 = AddrData.N0;
|
|
|
|
Register N2 = AddrData.N2;
|
|
|
|
Register N3 = AddrData.N3;
|
2020-01-24 17:45:33 +01:00
|
|
|
Offset = AddrData.Offset;
|
2020-01-18 03:28:58 +01:00
|
|
|
|
|
|
|
// Base pointer for the SRD.
|
|
|
|
Register SRDPtr;
|
|
|
|
|
|
|
|
if (N2) {
|
|
|
|
if (RBI.getRegBank(N2, *MRI, TRI)->getID() == AMDGPU::VGPRRegBankID) {
|
|
|
|
assert(N3);
|
|
|
|
if (RBI.getRegBank(N3, *MRI, TRI)->getID() == AMDGPU::VGPRRegBankID) {
|
|
|
|
// Both N2 and N3 are divergent. Use N0 (the result of the add) as the
|
|
|
|
// addr64, and construct the default resource from a 0 address.
|
|
|
|
VAddr = N0;
|
|
|
|
} else {
|
|
|
|
SRDPtr = N3;
|
|
|
|
VAddr = N2;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// N2 is not divergent.
|
|
|
|
SRDPtr = N2;
|
|
|
|
VAddr = N3;
|
|
|
|
}
|
|
|
|
} else if (RBI.getRegBank(N0, *MRI, TRI)->getID() == AMDGPU::VGPRRegBankID) {
|
|
|
|
// Use the default null pointer in the resource
|
|
|
|
VAddr = N0;
|
|
|
|
} else {
|
|
|
|
// N0 -> offset, or
|
|
|
|
// (N0 + C1) -> offset
|
|
|
|
SRDPtr = N0;
|
|
|
|
}
|
2020-01-22 22:56:48 +01:00
|
|
|
|
2020-01-18 03:28:58 +01:00
|
|
|
MachineIRBuilder B(*Root.getParent());
|
2020-01-24 17:45:33 +01:00
|
|
|
RSrcReg = buildAddr64RSrc(B, *MRI, TII, SRDPtr);
|
|
|
|
splitIllegalMUBUFOffset(B, SOffset, Offset);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AMDGPUInstructionSelector::selectMUBUFOffsetImpl(
|
|
|
|
MachineOperand &Root, Register &RSrcReg, Register &SOffset,
|
|
|
|
int64_t &Offset) const {
|
2020-08-15 17:41:08 +02:00
|
|
|
|
|
|
|
// FIXME: Pattern should not reach here.
|
|
|
|
if (STI.useFlatForGlobal())
|
|
|
|
return false;
|
|
|
|
|
2020-01-24 17:45:33 +01:00
|
|
|
MUBUFAddressData AddrData = parseMUBUFAddress(Root.getReg());
|
|
|
|
if (shouldUseAddr64(AddrData))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// N0 -> offset, or
|
|
|
|
// (N0 + C1) -> offset
|
|
|
|
Register SRDPtr = AddrData.N0;
|
|
|
|
Offset = AddrData.Offset;
|
|
|
|
|
|
|
|
// TODO: Look through extensions for 32-bit soffset.
|
|
|
|
MachineIRBuilder B(*Root.getParent());
|
|
|
|
|
|
|
|
RSrcReg = buildOffsetSrc(B, *MRI, TII, SRDPtr);
|
2020-01-18 03:28:58 +01:00
|
|
|
splitIllegalMUBUFOffset(B, SOffset, Offset);
|
2020-01-24 17:45:33 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectMUBUFAddr64(MachineOperand &Root) const {
|
|
|
|
Register VAddr;
|
|
|
|
Register RSrcReg;
|
|
|
|
Register SOffset;
|
|
|
|
int64_t Offset = 0;
|
|
|
|
|
|
|
|
if (!selectMUBUFAddr64Impl(Root, VAddr, RSrcReg, SOffset, Offset))
|
|
|
|
return {};
|
2020-01-22 22:56:48 +01:00
|
|
|
|
|
|
|
// FIXME: Use defaulted operands for trailing 0s and remove from the complex
|
|
|
|
// pattern.
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { // rsrc
|
2020-01-18 03:28:58 +01:00
|
|
|
MIB.addReg(RSrcReg);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // vaddr
|
|
|
|
MIB.addReg(VAddr);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // soffset
|
|
|
|
if (SOffset)
|
|
|
|
MIB.addReg(SOffset);
|
|
|
|
else
|
|
|
|
MIB.addImm(0);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // offset
|
|
|
|
MIB.addImm(Offset);
|
|
|
|
},
|
2021-02-09 01:36:10 +01:00
|
|
|
addZeroImm, // cpol
|
2020-01-18 03:28:58 +01:00
|
|
|
addZeroImm, // tfe
|
2021-02-09 01:36:10 +01:00
|
|
|
addZeroImm // swz
|
2020-01-18 03:28:58 +01:00
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectMUBUFOffset(MachineOperand &Root) const {
|
2020-01-24 17:45:33 +01:00
|
|
|
Register RSrcReg;
|
2020-01-18 03:28:58 +01:00
|
|
|
Register SOffset;
|
2020-01-24 17:45:33 +01:00
|
|
|
int64_t Offset = 0;
|
2020-01-18 03:28:58 +01:00
|
|
|
|
2020-01-24 17:45:33 +01:00
|
|
|
if (!selectMUBUFOffsetImpl(Root, RSrcReg, SOffset, Offset))
|
|
|
|
return {};
|
2020-01-18 03:28:58 +01:00
|
|
|
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { // rsrc
|
|
|
|
MIB.addReg(RSrcReg);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // soffset
|
|
|
|
if (SOffset)
|
|
|
|
MIB.addReg(SOffset);
|
|
|
|
else
|
|
|
|
MIB.addImm(0);
|
2020-01-22 22:56:48 +01:00
|
|
|
},
|
2020-01-18 03:28:58 +01:00
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(Offset); }, // offset
|
2021-02-09 01:36:10 +01:00
|
|
|
addZeroImm, // cpol
|
2020-01-22 22:56:48 +01:00
|
|
|
addZeroImm, // tfe
|
2021-02-17 22:37:46 +01:00
|
|
|
addZeroImm, // swz
|
2020-01-22 22:56:48 +01:00
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
2020-01-24 17:55:46 +01:00
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectMUBUFAddr64Atomic(MachineOperand &Root) const {
|
|
|
|
Register VAddr;
|
|
|
|
Register RSrcReg;
|
|
|
|
Register SOffset;
|
|
|
|
int64_t Offset = 0;
|
|
|
|
|
|
|
|
if (!selectMUBUFAddr64Impl(Root, VAddr, RSrcReg, SOffset, Offset))
|
|
|
|
return {};
|
|
|
|
|
|
|
|
// FIXME: Use defaulted operands for trailing 0s and remove from the complex
|
|
|
|
// pattern.
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { // rsrc
|
|
|
|
MIB.addReg(RSrcReg);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // vaddr
|
|
|
|
MIB.addReg(VAddr);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // soffset
|
|
|
|
if (SOffset)
|
|
|
|
MIB.addReg(SOffset);
|
|
|
|
else
|
|
|
|
MIB.addImm(0);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // offset
|
|
|
|
MIB.addImm(Offset);
|
|
|
|
},
|
2021-02-09 01:36:10 +01:00
|
|
|
[=](MachineInstrBuilder &MIB) {
|
|
|
|
MIB.addImm(AMDGPU::CPol::GLC); // cpol
|
|
|
|
}
|
2020-01-24 17:55:46 +01:00
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectMUBUFOffsetAtomic(MachineOperand &Root) const {
|
|
|
|
Register RSrcReg;
|
|
|
|
Register SOffset;
|
|
|
|
int64_t Offset = 0;
|
|
|
|
|
|
|
|
if (!selectMUBUFOffsetImpl(Root, RSrcReg, SOffset, Offset))
|
|
|
|
return {};
|
|
|
|
|
|
|
|
return {{
|
|
|
|
[=](MachineInstrBuilder &MIB) { // rsrc
|
|
|
|
MIB.addReg(RSrcReg);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { // soffset
|
|
|
|
if (SOffset)
|
|
|
|
MIB.addReg(SOffset);
|
|
|
|
else
|
|
|
|
MIB.addImm(0);
|
|
|
|
},
|
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(Offset); }, // offset
|
2021-02-09 01:36:10 +01:00
|
|
|
[=](MachineInstrBuilder &MIB) { MIB.addImm(AMDGPU::CPol::GLC); } // cpol
|
2020-01-24 17:55:46 +01:00
|
|
|
}};
|
|
|
|
}
|
|
|
|
|
2020-01-30 02:34:32 +01:00
|
|
|
/// Get an immediate that must be 32-bits, and treated as zero extended.
|
|
|
|
static Optional<uint64_t> getConstantZext32Val(Register Reg,
|
|
|
|
const MachineRegisterInfo &MRI) {
|
|
|
|
// getConstantVRegVal sexts any values, so see if that matters.
|
2020-11-03 15:50:17 +01:00
|
|
|
Optional<int64_t> OffsetVal = getConstantVRegSExtVal(Reg, MRI);
|
2020-01-30 02:34:32 +01:00
|
|
|
if (!OffsetVal || !isInt<32>(*OffsetVal))
|
|
|
|
return None;
|
|
|
|
return Lo_32(*OffsetVal);
|
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectSMRDBufferImm(MachineOperand &Root) const {
|
|
|
|
Optional<uint64_t> OffsetVal = getConstantZext32Val(Root.getReg(), *MRI);
|
|
|
|
if (!OffsetVal)
|
|
|
|
return {};
|
|
|
|
|
2020-03-30 08:32:20 +02:00
|
|
|
Optional<int64_t> EncodedImm =
|
|
|
|
AMDGPU::getSMRDEncodedOffset(STI, *OffsetVal, true);
|
2020-01-30 02:34:32 +01:00
|
|
|
if (!EncodedImm)
|
|
|
|
return {};
|
|
|
|
|
|
|
|
return {{ [=](MachineInstrBuilder &MIB) { MIB.addImm(*EncodedImm); } }};
|
|
|
|
}
|
|
|
|
|
|
|
|
InstructionSelector::ComplexRendererFns
|
|
|
|
AMDGPUInstructionSelector::selectSMRDBufferImm32(MachineOperand &Root) const {
|
|
|
|
assert(STI.getGeneration() == AMDGPUSubtarget::SEA_ISLANDS);
|
|
|
|
|
|
|
|
Optional<uint64_t> OffsetVal = getConstantZext32Val(Root.getReg(), *MRI);
|
|
|
|
if (!OffsetVal)
|
|
|
|
return {};
|
|
|
|
|
|
|
|
Optional<int64_t> EncodedImm
|
|
|
|
= AMDGPU::getSMRDEncodedLiteralOffset32(STI, *OffsetVal);
|
|
|
|
if (!EncodedImm)
|
|
|
|
return {};
|
|
|
|
|
|
|
|
return {{ [=](MachineInstrBuilder &MIB) { MIB.addImm(*EncodedImm); } }};
|
|
|
|
}
|
|
|
|
|
2019-10-01 04:07:25 +02:00
|
|
|
void AMDGPUInstructionSelector::renderTruncImm32(MachineInstrBuilder &MIB,
|
TableGen/GlobalISel: Add way for SDNodeXForm to work on timm
The current implementation assumes there is an instruction associated
with the transform, but this is not the case for
timm/TargetConstant/immarg values. These transforms should directly
operate on a specific MachineOperand in the source
instruction. TableGen would assert if you attempted to define an
equivalent GISDNodeXFormEquiv using timm when it failed to find the
instruction matcher.
Specially recognize SDNodeXForms on timm, and pass the operand index
to the render function.
Ideally this would be a separate render function type that looks like
void renderFoo(MachineInstrBuilder, const MachineOperand&), but this
proved to be somewhat mechanically painful. Add an optional operand
index which will only be passed if the transform should only look at
the one source operand.
Theoretically it would also be possible to only ever pass the
MachineOperand, and the existing renderers would check the parent. I
think that would be somewhat ugly for the standard usage which may
want to inspect other operands, and I also think MachineOperand should
eventually not carry a pointer to the parent instruction.
Use it in one sample pattern. This isn't a great example, since the
transform exists to satisfy DAG type constraints. This could also be
avoided by just changing the MachineInstr's arbitrary choice of
operand type from i16 to i32. Other patterns have nontrivial uses, but
this serves as the simplest example.
One flaw this still has is if you try to use an SDNodeXForm defined
for imm, but the source pattern uses timm, you still see the "Failed
to lookup instruction" assert. However, there is now a way to avoid
it.
2020-01-08 18:53:15 +01:00
|
|
|
const MachineInstr &MI,
|
|
|
|
int OpIdx) const {
|
|
|
|
assert(MI.getOpcode() == TargetOpcode::G_CONSTANT && OpIdx == -1 &&
|
|
|
|
"Expected G_CONSTANT");
|
2020-01-30 15:25:47 +01:00
|
|
|
MIB.addImm(MI.getOperand(1).getCImm()->getSExtValue());
|
2019-10-01 04:07:25 +02:00
|
|
|
}
|
2020-01-06 20:39:13 +01:00
|
|
|
|
2020-01-07 17:29:05 +01:00
|
|
|
void AMDGPUInstructionSelector::renderNegateImm(MachineInstrBuilder &MIB,
|
TableGen/GlobalISel: Add way for SDNodeXForm to work on timm
The current implementation assumes there is an instruction associated
with the transform, but this is not the case for
timm/TargetConstant/immarg values. These transforms should directly
operate on a specific MachineOperand in the source
instruction. TableGen would assert if you attempted to define an
equivalent GISDNodeXFormEquiv using timm when it failed to find the
instruction matcher.
Specially recognize SDNodeXForms on timm, and pass the operand index
to the render function.
Ideally this would be a separate render function type that looks like
void renderFoo(MachineInstrBuilder, const MachineOperand&), but this
proved to be somewhat mechanically painful. Add an optional operand
index which will only be passed if the transform should only look at
the one source operand.
Theoretically it would also be possible to only ever pass the
MachineOperand, and the existing renderers would check the parent. I
think that would be somewhat ugly for the standard usage which may
want to inspect other operands, and I also think MachineOperand should
eventually not carry a pointer to the parent instruction.
Use it in one sample pattern. This isn't a great example, since the
transform exists to satisfy DAG type constraints. This could also be
avoided by just changing the MachineInstr's arbitrary choice of
operand type from i16 to i32. Other patterns have nontrivial uses, but
this serves as the simplest example.
One flaw this still has is if you try to use an SDNodeXForm defined
for imm, but the source pattern uses timm, you still see the "Failed
to lookup instruction" assert. However, there is now a way to avoid
it.
2020-01-08 18:53:15 +01:00
|
|
|
const MachineInstr &MI,
|
|
|
|
int OpIdx) const {
|
|
|
|
assert(MI.getOpcode() == TargetOpcode::G_CONSTANT && OpIdx == -1 &&
|
|
|
|
"Expected G_CONSTANT");
|
2020-01-07 17:29:05 +01:00
|
|
|
MIB.addImm(-MI.getOperand(1).getCImm()->getSExtValue());
|
|
|
|
}
|
|
|
|
|
2020-01-07 17:45:10 +01:00
|
|
|
void AMDGPUInstructionSelector::renderBitcastImm(MachineInstrBuilder &MIB,
|
TableGen/GlobalISel: Add way for SDNodeXForm to work on timm
The current implementation assumes there is an instruction associated
with the transform, but this is not the case for
timm/TargetConstant/immarg values. These transforms should directly
operate on a specific MachineOperand in the source
instruction. TableGen would assert if you attempted to define an
equivalent GISDNodeXFormEquiv using timm when it failed to find the
instruction matcher.
Specially recognize SDNodeXForms on timm, and pass the operand index
to the render function.
Ideally this would be a separate render function type that looks like
void renderFoo(MachineInstrBuilder, const MachineOperand&), but this
proved to be somewhat mechanically painful. Add an optional operand
index which will only be passed if the transform should only look at
the one source operand.
Theoretically it would also be possible to only ever pass the
MachineOperand, and the existing renderers would check the parent. I
think that would be somewhat ugly for the standard usage which may
want to inspect other operands, and I also think MachineOperand should
eventually not carry a pointer to the parent instruction.
Use it in one sample pattern. This isn't a great example, since the
transform exists to satisfy DAG type constraints. This could also be
avoided by just changing the MachineInstr's arbitrary choice of
operand type from i16 to i32. Other patterns have nontrivial uses, but
this serves as the simplest example.
One flaw this still has is if you try to use an SDNodeXForm defined
for imm, but the source pattern uses timm, you still see the "Failed
to lookup instruction" assert. However, there is now a way to avoid
it.
2020-01-08 18:53:15 +01:00
|
|
|
const MachineInstr &MI,
|
|
|
|
int OpIdx) const {
|
|
|
|
assert(OpIdx == -1);
|
|
|
|
|
2020-01-07 17:45:10 +01:00
|
|
|
const MachineOperand &Op = MI.getOperand(1);
|
|
|
|
if (MI.getOpcode() == TargetOpcode::G_FCONSTANT)
|
|
|
|
MIB.addImm(Op.getFPImm()->getValueAPF().bitcastToAPInt().getZExtValue());
|
|
|
|
else {
|
|
|
|
assert(MI.getOpcode() == TargetOpcode::G_CONSTANT && "Expected G_CONSTANT");
|
|
|
|
MIB.addImm(Op.getCImm()->getSExtValue());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-01-07 18:38:35 +01:00
|
|
|
void AMDGPUInstructionSelector::renderPopcntImm(MachineInstrBuilder &MIB,
|
TableGen/GlobalISel: Add way for SDNodeXForm to work on timm
The current implementation assumes there is an instruction associated
with the transform, but this is not the case for
timm/TargetConstant/immarg values. These transforms should directly
operate on a specific MachineOperand in the source
instruction. TableGen would assert if you attempted to define an
equivalent GISDNodeXFormEquiv using timm when it failed to find the
instruction matcher.
Specially recognize SDNodeXForms on timm, and pass the operand index
to the render function.
Ideally this would be a separate render function type that looks like
void renderFoo(MachineInstrBuilder, const MachineOperand&), but this
proved to be somewhat mechanically painful. Add an optional operand
index which will only be passed if the transform should only look at
the one source operand.
Theoretically it would also be possible to only ever pass the
MachineOperand, and the existing renderers would check the parent. I
think that would be somewhat ugly for the standard usage which may
want to inspect other operands, and I also think MachineOperand should
eventually not carry a pointer to the parent instruction.
Use it in one sample pattern. This isn't a great example, since the
transform exists to satisfy DAG type constraints. This could also be
avoided by just changing the MachineInstr's arbitrary choice of
operand type from i16 to i32. Other patterns have nontrivial uses, but
this serves as the simplest example.
One flaw this still has is if you try to use an SDNodeXForm defined
for imm, but the source pattern uses timm, you still see the "Failed
to lookup instruction" assert. However, there is now a way to avoid
it.
2020-01-08 18:53:15 +01:00
|
|
|
const MachineInstr &MI,
|
|
|
|
int OpIdx) const {
|
|
|
|
assert(MI.getOpcode() == TargetOpcode::G_CONSTANT && OpIdx == -1 &&
|
|
|
|
"Expected G_CONSTANT");
|
2020-01-07 18:38:35 +01:00
|
|
|
MIB.addImm(MI.getOperand(1).getCImm()->getValue().countPopulation());
|
|
|
|
}
|
|
|
|
|
TableGen/GlobalISel: Add way for SDNodeXForm to work on timm
The current implementation assumes there is an instruction associated
with the transform, but this is not the case for
timm/TargetConstant/immarg values. These transforms should directly
operate on a specific MachineOperand in the source
instruction. TableGen would assert if you attempted to define an
equivalent GISDNodeXFormEquiv using timm when it failed to find the
instruction matcher.
Specially recognize SDNodeXForms on timm, and pass the operand index
to the render function.
Ideally this would be a separate render function type that looks like
void renderFoo(MachineInstrBuilder, const MachineOperand&), but this
proved to be somewhat mechanically painful. Add an optional operand
index which will only be passed if the transform should only look at
the one source operand.
Theoretically it would also be possible to only ever pass the
MachineOperand, and the existing renderers would check the parent. I
think that would be somewhat ugly for the standard usage which may
want to inspect other operands, and I also think MachineOperand should
eventually not carry a pointer to the parent instruction.
Use it in one sample pattern. This isn't a great example, since the
transform exists to satisfy DAG type constraints. This could also be
avoided by just changing the MachineInstr's arbitrary choice of
operand type from i16 to i32. Other patterns have nontrivial uses, but
this serves as the simplest example.
One flaw this still has is if you try to use an SDNodeXForm defined
for imm, but the source pattern uses timm, you still see the "Failed
to lookup instruction" assert. However, there is now a way to avoid
it.
2020-01-08 18:53:15 +01:00
|
|
|
/// This only really exists to satisfy DAG type checking machinery, so is a
|
|
|
|
/// no-op here.
|
2020-01-14 00:18:56 +01:00
|
|
|
void AMDGPUInstructionSelector::renderTruncTImm(MachineInstrBuilder &MIB,
|
|
|
|
const MachineInstr &MI,
|
|
|
|
int OpIdx) const {
|
TableGen/GlobalISel: Add way for SDNodeXForm to work on timm
The current implementation assumes there is an instruction associated
with the transform, but this is not the case for
timm/TargetConstant/immarg values. These transforms should directly
operate on a specific MachineOperand in the source
instruction. TableGen would assert if you attempted to define an
equivalent GISDNodeXFormEquiv using timm when it failed to find the
instruction matcher.
Specially recognize SDNodeXForms on timm, and pass the operand index
to the render function.
Ideally this would be a separate render function type that looks like
void renderFoo(MachineInstrBuilder, const MachineOperand&), but this
proved to be somewhat mechanically painful. Add an optional operand
index which will only be passed if the transform should only look at
the one source operand.
Theoretically it would also be possible to only ever pass the
MachineOperand, and the existing renderers would check the parent. I
think that would be somewhat ugly for the standard usage which may
want to inspect other operands, and I also think MachineOperand should
eventually not carry a pointer to the parent instruction.
Use it in one sample pattern. This isn't a great example, since the
transform exists to satisfy DAG type constraints. This could also be
avoided by just changing the MachineInstr's arbitrary choice of
operand type from i16 to i32. Other patterns have nontrivial uses, but
this serves as the simplest example.
One flaw this still has is if you try to use an SDNodeXForm defined
for imm, but the source pattern uses timm, you still see the "Failed
to lookup instruction" assert. However, there is now a way to avoid
it.
2020-01-08 18:53:15 +01:00
|
|
|
MIB.addImm(MI.getOperand(OpIdx).getImm());
|
|
|
|
}
|
|
|
|
|
2021-02-09 01:36:10 +01:00
|
|
|
void AMDGPUInstructionSelector::renderExtractCPol(MachineInstrBuilder &MIB,
|
|
|
|
const MachineInstr &MI,
|
|
|
|
int OpIdx) const {
|
2019-09-19 06:29:20 +02:00
|
|
|
assert(OpIdx >= 0 && "expected to match an immediate operand");
|
2021-02-09 01:36:10 +01:00
|
|
|
MIB.addImm(MI.getOperand(OpIdx).getImm() & AMDGPU::CPol::ALL);
|
2019-09-19 06:29:20 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
void AMDGPUInstructionSelector::renderExtractSWZ(MachineInstrBuilder &MIB,
|
|
|
|
const MachineInstr &MI,
|
|
|
|
int OpIdx) const {
|
|
|
|
assert(OpIdx >= 0 && "expected to match an immediate operand");
|
|
|
|
MIB.addImm((MI.getOperand(OpIdx).getImm() >> 3) & 1);
|
|
|
|
}
|
|
|
|
|
2021-02-09 01:36:10 +01:00
|
|
|
void AMDGPUInstructionSelector::renderSetGLC(MachineInstrBuilder &MIB,
|
|
|
|
const MachineInstr &MI,
|
|
|
|
int OpIdx) const {
|
2021-02-17 22:37:46 +01:00
|
|
|
assert(OpIdx >= 0 && "expected to match an immediate operand");
|
2021-02-09 01:36:10 +01:00
|
|
|
MIB.addImm(MI.getOperand(OpIdx).getImm() | AMDGPU::CPol::GLC);
|
2021-02-17 22:37:46 +01:00
|
|
|
}
|
|
|
|
|
2020-07-31 22:01:38 +02:00
|
|
|
void AMDGPUInstructionSelector::renderFrameIndex(MachineInstrBuilder &MIB,
|
|
|
|
const MachineInstr &MI,
|
|
|
|
int OpIdx) const {
|
|
|
|
MIB.addFrameIndex((MI.getOperand(1).getIndex()));
|
|
|
|
}
|
|
|
|
|
2020-01-06 20:39:13 +01:00
|
|
|
bool AMDGPUInstructionSelector::isInlineImmediate16(int64_t Imm) const {
|
|
|
|
return AMDGPU::isInlinableLiteral16(Imm, STI.hasInv2PiInlineImm());
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AMDGPUInstructionSelector::isInlineImmediate32(int64_t Imm) const {
|
|
|
|
return AMDGPU::isInlinableLiteral32(Imm, STI.hasInv2PiInlineImm());
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AMDGPUInstructionSelector::isInlineImmediate64(int64_t Imm) const {
|
|
|
|
return AMDGPU::isInlinableLiteral64(Imm, STI.hasInv2PiInlineImm());
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AMDGPUInstructionSelector::isInlineImmediate(const APFloat &Imm) const {
|
|
|
|
return TII.isInlineConstant(Imm);
|
|
|
|
}
|