mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-24 19:52:54 +01:00
187863ee97
When passing SVE types as arguments to function calls we can run out of hardware SVE registers. This is normally fine, since we switch to an indirect mode where we pass a pointer to a SVE stack object in a GPR. However, if we switch over part-way through processing a SVE tuple then part of it will be in registers and the other part will be on the stack. I've fixed this by ensuring that: 1. When we don't have enough registers to allocate the whole block we mark any remaining SVE registers temporarily as allocated. 2. We temporarily remove the InConsecutiveRegs flags from the last tuple part argument and reinvoke the autogenerated calling convention handler. Doing this prevents the code from entering an infinite recursion and, in combination with 1), ensures we switch over to the Indirect mode. 3. After allocating a GPR register for the pointer to the tuple we then deallocate any SVE registers we marked as allocated in 1). We also set the InConsecutiveRegs flags back how they were before. 4. I've changed the AArch64ISelLowering LowerCALL and LowerFormalArguments functions to detect the start of a tuple, which involves allocating a single stack object and doing the correct numbers of legal loads and stores. Differential Revision: https://reviews.llvm.org/D90219
208 lines
8.4 KiB
C++
208 lines
8.4 KiB
C++
//=== AArch64CallingConvention.cpp - AArch64 CC impl ------------*- C++ -*-===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// This file contains the table-generated and custom routines for the AArch64
|
|
// Calling Convention.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "AArch64CallingConvention.h"
|
|
#include "AArch64.h"
|
|
#include "AArch64InstrInfo.h"
|
|
#include "AArch64Subtarget.h"
|
|
#include "llvm/CodeGen/CallingConvLower.h"
|
|
#include "llvm/CodeGen/TargetInstrInfo.h"
|
|
#include "llvm/IR/CallingConv.h"
|
|
using namespace llvm;
|
|
|
|
static const MCPhysReg XRegList[] = {AArch64::X0, AArch64::X1, AArch64::X2,
|
|
AArch64::X3, AArch64::X4, AArch64::X5,
|
|
AArch64::X6, AArch64::X7};
|
|
static const MCPhysReg HRegList[] = {AArch64::H0, AArch64::H1, AArch64::H2,
|
|
AArch64::H3, AArch64::H4, AArch64::H5,
|
|
AArch64::H6, AArch64::H7};
|
|
static const MCPhysReg SRegList[] = {AArch64::S0, AArch64::S1, AArch64::S2,
|
|
AArch64::S3, AArch64::S4, AArch64::S5,
|
|
AArch64::S6, AArch64::S7};
|
|
static const MCPhysReg DRegList[] = {AArch64::D0, AArch64::D1, AArch64::D2,
|
|
AArch64::D3, AArch64::D4, AArch64::D5,
|
|
AArch64::D6, AArch64::D7};
|
|
static const MCPhysReg QRegList[] = {AArch64::Q0, AArch64::Q1, AArch64::Q2,
|
|
AArch64::Q3, AArch64::Q4, AArch64::Q5,
|
|
AArch64::Q6, AArch64::Q7};
|
|
static const MCPhysReg ZRegList[] = {AArch64::Z0, AArch64::Z1, AArch64::Z2,
|
|
AArch64::Z3, AArch64::Z4, AArch64::Z5,
|
|
AArch64::Z6, AArch64::Z7};
|
|
|
|
static bool finishStackBlock(SmallVectorImpl<CCValAssign> &PendingMembers,
|
|
MVT LocVT, ISD::ArgFlagsTy &ArgFlags,
|
|
CCState &State, Align SlotAlign) {
|
|
if (LocVT.isScalableVector()) {
|
|
const AArch64Subtarget &Subtarget = static_cast<const AArch64Subtarget &>(
|
|
State.getMachineFunction().getSubtarget());
|
|
const AArch64TargetLowering *TLI = Subtarget.getTargetLowering();
|
|
|
|
// We are about to reinvoke the CCAssignFn auto-generated handler. If we
|
|
// don't unset these flags we will get stuck in an infinite loop forever
|
|
// invoking the custom handler.
|
|
ArgFlags.setInConsecutiveRegs(false);
|
|
ArgFlags.setInConsecutiveRegsLast(false);
|
|
|
|
// The calling convention for passing SVE tuples states that in the event
|
|
// we cannot allocate enough registers for the tuple we should still leave
|
|
// any remaining registers unallocated. However, when we call the
|
|
// CCAssignFn again we want it to behave as if all remaining registers are
|
|
// allocated. This will force the code to pass the tuple indirectly in
|
|
// accordance with the PCS.
|
|
bool RegsAllocated[8];
|
|
for (int I = 0; I < 8; I++) {
|
|
RegsAllocated[I] = State.isAllocated(ZRegList[I]);
|
|
State.AllocateReg(ZRegList[I]);
|
|
}
|
|
|
|
auto &It = PendingMembers[0];
|
|
CCAssignFn *AssignFn =
|
|
TLI->CCAssignFnForCall(State.getCallingConv(), /*IsVarArg=*/false);
|
|
if (AssignFn(It.getValNo(), It.getValVT(), It.getValVT(), CCValAssign::Full,
|
|
ArgFlags, State))
|
|
llvm_unreachable("Call operand has unhandled type");
|
|
|
|
// Return the flags to how they were before.
|
|
ArgFlags.setInConsecutiveRegs(true);
|
|
ArgFlags.setInConsecutiveRegsLast(true);
|
|
|
|
// Return the register state back to how it was before, leaving any
|
|
// unallocated registers available for other smaller types.
|
|
for (int I = 0; I < 8; I++)
|
|
if (!RegsAllocated[I])
|
|
State.DeallocateReg(ZRegList[I]);
|
|
|
|
// All pending members have now been allocated
|
|
PendingMembers.clear();
|
|
return true;
|
|
}
|
|
|
|
unsigned Size = LocVT.getSizeInBits() / 8;
|
|
const Align StackAlign =
|
|
State.getMachineFunction().getDataLayout().getStackAlignment();
|
|
const Align OrigAlign = ArgFlags.getNonZeroOrigAlign();
|
|
const Align Alignment = std::min(OrigAlign, StackAlign);
|
|
|
|
for (auto &It : PendingMembers) {
|
|
It.convertToMem(State.AllocateStack(Size, std::max(Alignment, SlotAlign)));
|
|
State.addLoc(It);
|
|
SlotAlign = Align(1);
|
|
}
|
|
|
|
// All pending members have now been allocated
|
|
PendingMembers.clear();
|
|
return true;
|
|
}
|
|
|
|
/// The Darwin variadic PCS places anonymous arguments in 8-byte stack slots. An
|
|
/// [N x Ty] type must still be contiguous in memory though.
|
|
static bool CC_AArch64_Custom_Stack_Block(
|
|
unsigned &ValNo, MVT &ValVT, MVT &LocVT, CCValAssign::LocInfo &LocInfo,
|
|
ISD::ArgFlagsTy &ArgFlags, CCState &State) {
|
|
SmallVectorImpl<CCValAssign> &PendingMembers = State.getPendingLocs();
|
|
|
|
// Add the argument to the list to be allocated once we know the size of the
|
|
// block.
|
|
PendingMembers.push_back(
|
|
CCValAssign::getPending(ValNo, ValVT, LocVT, LocInfo));
|
|
|
|
if (!ArgFlags.isInConsecutiveRegsLast())
|
|
return true;
|
|
|
|
return finishStackBlock(PendingMembers, LocVT, ArgFlags, State, Align(8));
|
|
}
|
|
|
|
/// Given an [N x Ty] block, it should be passed in a consecutive sequence of
|
|
/// registers. If no such sequence is available, mark the rest of the registers
|
|
/// of that type as used and place the argument on the stack.
|
|
static bool CC_AArch64_Custom_Block(unsigned &ValNo, MVT &ValVT, MVT &LocVT,
|
|
CCValAssign::LocInfo &LocInfo,
|
|
ISD::ArgFlagsTy &ArgFlags, CCState &State) {
|
|
const AArch64Subtarget &Subtarget = static_cast<const AArch64Subtarget &>(
|
|
State.getMachineFunction().getSubtarget());
|
|
bool IsDarwinILP32 = Subtarget.isTargetILP32() && Subtarget.isTargetMachO();
|
|
|
|
// Try to allocate a contiguous block of registers, each of the correct
|
|
// size to hold one member.
|
|
ArrayRef<MCPhysReg> RegList;
|
|
if (LocVT.SimpleTy == MVT::i64 || (IsDarwinILP32 && LocVT.SimpleTy == MVT::i32))
|
|
RegList = XRegList;
|
|
else if (LocVT.SimpleTy == MVT::f16)
|
|
RegList = HRegList;
|
|
else if (LocVT.SimpleTy == MVT::f32 || LocVT.is32BitVector())
|
|
RegList = SRegList;
|
|
else if (LocVT.SimpleTy == MVT::f64 || LocVT.is64BitVector())
|
|
RegList = DRegList;
|
|
else if (LocVT.SimpleTy == MVT::f128 || LocVT.is128BitVector())
|
|
RegList = QRegList;
|
|
else if (LocVT.isScalableVector())
|
|
RegList = ZRegList;
|
|
else {
|
|
// Not an array we want to split up after all.
|
|
return false;
|
|
}
|
|
|
|
SmallVectorImpl<CCValAssign> &PendingMembers = State.getPendingLocs();
|
|
|
|
// Add the argument to the list to be allocated once we know the size of the
|
|
// block.
|
|
PendingMembers.push_back(
|
|
CCValAssign::getPending(ValNo, ValVT, LocVT, LocInfo));
|
|
|
|
if (!ArgFlags.isInConsecutiveRegsLast())
|
|
return true;
|
|
|
|
// [N x i32] arguments get packed into x-registers on Darwin's arm64_32
|
|
// because that's how the armv7k Clang front-end emits small structs.
|
|
unsigned EltsPerReg = (IsDarwinILP32 && LocVT.SimpleTy == MVT::i32) ? 2 : 1;
|
|
unsigned RegResult = State.AllocateRegBlock(
|
|
RegList, alignTo(PendingMembers.size(), EltsPerReg) / EltsPerReg);
|
|
if (RegResult && EltsPerReg == 1) {
|
|
for (auto &It : PendingMembers) {
|
|
It.convertToReg(RegResult);
|
|
State.addLoc(It);
|
|
++RegResult;
|
|
}
|
|
PendingMembers.clear();
|
|
return true;
|
|
} else if (RegResult) {
|
|
assert(EltsPerReg == 2 && "unexpected ABI");
|
|
bool UseHigh = false;
|
|
CCValAssign::LocInfo Info;
|
|
for (auto &It : PendingMembers) {
|
|
Info = UseHigh ? CCValAssign::AExtUpper : CCValAssign::ZExt;
|
|
State.addLoc(CCValAssign::getReg(It.getValNo(), MVT::i32, RegResult,
|
|
MVT::i64, Info));
|
|
UseHigh = !UseHigh;
|
|
if (!UseHigh)
|
|
++RegResult;
|
|
}
|
|
PendingMembers.clear();
|
|
return true;
|
|
}
|
|
|
|
if (!LocVT.isScalableVector()) {
|
|
// Mark all regs in the class as unavailable
|
|
for (auto Reg : RegList)
|
|
State.AllocateReg(Reg);
|
|
}
|
|
|
|
const Align SlotAlign = Subtarget.isTargetDarwin() ? Align(1) : Align(8);
|
|
|
|
return finishStackBlock(PendingMembers, LocVT, ArgFlags, State, SlotAlign);
|
|
}
|
|
|
|
// TableGen provides definitions of the calling convention analysis entry
|
|
// points.
|
|
#include "AArch64GenCallingConv.inc"
|