mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-22 02:33:06 +01:00
00169a8661
Linker scripts might not handle COMDAT sections. SLSHardeing adds new section for each __llvm_slsblr_thunk_xN. This new option allows the generation of the thunks into the normal text section to handle these exceptional cases. ,comdat or ,noncomdat can be added to harden-sls to control the codegen. -mharden-sls=[all|retbr|blr],nocomdat. Reviewed By: kristof.beyls Differential Revision: https://reviews.llvm.org/D100546
448 lines
16 KiB
C++
448 lines
16 KiB
C++
//===- AArch64SLSHardening.cpp - Harden Straight Line Missspeculation -----===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// This file contains a pass to insert code to mitigate against side channel
|
|
// vulnerabilities that may happen under straight line miss-speculation.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "AArch64InstrInfo.h"
|
|
#include "AArch64Subtarget.h"
|
|
#include "Utils/AArch64BaseInfo.h"
|
|
#include "llvm/ADT/BitVector.h"
|
|
#include "llvm/ADT/SmallVector.h"
|
|
#include "llvm/CodeGen/IndirectThunks.h"
|
|
#include "llvm/CodeGen/MachineBasicBlock.h"
|
|
#include "llvm/CodeGen/MachineFunction.h"
|
|
#include "llvm/CodeGen/MachineFunctionPass.h"
|
|
#include "llvm/CodeGen/MachineInstr.h"
|
|
#include "llvm/CodeGen/MachineInstrBuilder.h"
|
|
#include "llvm/CodeGen/MachineOperand.h"
|
|
#include "llvm/CodeGen/MachineRegisterInfo.h"
|
|
#include "llvm/CodeGen/RegisterScavenging.h"
|
|
#include "llvm/IR/DebugLoc.h"
|
|
#include "llvm/Pass.h"
|
|
#include "llvm/Support/CodeGen.h"
|
|
#include "llvm/Support/Debug.h"
|
|
#include "llvm/Target/TargetMachine.h"
|
|
#include <cassert>
|
|
|
|
using namespace llvm;
|
|
|
|
#define DEBUG_TYPE "aarch64-sls-hardening"
|
|
|
|
#define AARCH64_SLS_HARDENING_NAME "AArch64 sls hardening pass"
|
|
|
|
namespace {
|
|
|
|
class AArch64SLSHardening : public MachineFunctionPass {
|
|
public:
|
|
const TargetInstrInfo *TII;
|
|
const TargetRegisterInfo *TRI;
|
|
const AArch64Subtarget *ST;
|
|
|
|
static char ID;
|
|
|
|
AArch64SLSHardening() : MachineFunctionPass(ID) {
|
|
initializeAArch64SLSHardeningPass(*PassRegistry::getPassRegistry());
|
|
}
|
|
|
|
bool runOnMachineFunction(MachineFunction &Fn) override;
|
|
|
|
StringRef getPassName() const override { return AARCH64_SLS_HARDENING_NAME; }
|
|
|
|
private:
|
|
bool hardenReturnsAndBRs(MachineBasicBlock &MBB) const;
|
|
bool hardenBLRs(MachineBasicBlock &MBB) const;
|
|
MachineBasicBlock &ConvertBLRToBL(MachineBasicBlock &MBB,
|
|
MachineBasicBlock::iterator) const;
|
|
};
|
|
|
|
} // end anonymous namespace
|
|
|
|
char AArch64SLSHardening::ID = 0;
|
|
|
|
INITIALIZE_PASS(AArch64SLSHardening, "aarch64-sls-hardening",
|
|
AARCH64_SLS_HARDENING_NAME, false, false)
|
|
|
|
static void insertSpeculationBarrier(const AArch64Subtarget *ST,
|
|
MachineBasicBlock &MBB,
|
|
MachineBasicBlock::iterator MBBI,
|
|
DebugLoc DL,
|
|
bool AlwaysUseISBDSB = false) {
|
|
assert(MBBI != MBB.begin() &&
|
|
"Must not insert SpeculationBarrierEndBB as only instruction in MBB.");
|
|
assert(std::prev(MBBI)->isBarrier() &&
|
|
"SpeculationBarrierEndBB must only follow unconditional control flow "
|
|
"instructions.");
|
|
assert(std::prev(MBBI)->isTerminator() &&
|
|
"SpeculationBarrierEndBB must only follow terminators.");
|
|
const TargetInstrInfo *TII = ST->getInstrInfo();
|
|
unsigned BarrierOpc = ST->hasSB() && !AlwaysUseISBDSB
|
|
? AArch64::SpeculationBarrierSBEndBB
|
|
: AArch64::SpeculationBarrierISBDSBEndBB;
|
|
if (MBBI == MBB.end() ||
|
|
(MBBI->getOpcode() != AArch64::SpeculationBarrierSBEndBB &&
|
|
MBBI->getOpcode() != AArch64::SpeculationBarrierISBDSBEndBB))
|
|
BuildMI(MBB, MBBI, DL, TII->get(BarrierOpc));
|
|
}
|
|
|
|
bool AArch64SLSHardening::runOnMachineFunction(MachineFunction &MF) {
|
|
ST = &MF.getSubtarget<AArch64Subtarget>();
|
|
TII = MF.getSubtarget().getInstrInfo();
|
|
TRI = MF.getSubtarget().getRegisterInfo();
|
|
|
|
bool Modified = false;
|
|
for (auto &MBB : MF) {
|
|
Modified |= hardenReturnsAndBRs(MBB);
|
|
Modified |= hardenBLRs(MBB);
|
|
}
|
|
|
|
return Modified;
|
|
}
|
|
|
|
static bool isBLR(const MachineInstr &MI) {
|
|
switch (MI.getOpcode()) {
|
|
case AArch64::BLR:
|
|
case AArch64::BLRNoIP:
|
|
return true;
|
|
case AArch64::BLRAA:
|
|
case AArch64::BLRAB:
|
|
case AArch64::BLRAAZ:
|
|
case AArch64::BLRABZ:
|
|
llvm_unreachable("Currently, LLVM's code generator does not support "
|
|
"producing BLRA* instructions. Therefore, there's no "
|
|
"support in this pass for those instructions.");
|
|
}
|
|
return false;
|
|
}
|
|
|
|
bool AArch64SLSHardening::hardenReturnsAndBRs(MachineBasicBlock &MBB) const {
|
|
if (!ST->hardenSlsRetBr())
|
|
return false;
|
|
bool Modified = false;
|
|
MachineBasicBlock::iterator MBBI = MBB.getFirstTerminator(), E = MBB.end();
|
|
MachineBasicBlock::iterator NextMBBI;
|
|
for (; MBBI != E; MBBI = NextMBBI) {
|
|
MachineInstr &MI = *MBBI;
|
|
NextMBBI = std::next(MBBI);
|
|
if (MI.isReturn() || isIndirectBranchOpcode(MI.getOpcode())) {
|
|
assert(MI.isTerminator());
|
|
insertSpeculationBarrier(ST, MBB, std::next(MBBI), MI.getDebugLoc());
|
|
Modified = true;
|
|
}
|
|
}
|
|
return Modified;
|
|
}
|
|
|
|
static const char SLSBLRNamePrefix[] = "__llvm_slsblr_thunk_";
|
|
|
|
static const struct ThunkNameAndReg {
|
|
const char* Name;
|
|
Register Reg;
|
|
} SLSBLRThunks[] = {
|
|
{ "__llvm_slsblr_thunk_x0", AArch64::X0},
|
|
{ "__llvm_slsblr_thunk_x1", AArch64::X1},
|
|
{ "__llvm_slsblr_thunk_x2", AArch64::X2},
|
|
{ "__llvm_slsblr_thunk_x3", AArch64::X3},
|
|
{ "__llvm_slsblr_thunk_x4", AArch64::X4},
|
|
{ "__llvm_slsblr_thunk_x5", AArch64::X5},
|
|
{ "__llvm_slsblr_thunk_x6", AArch64::X6},
|
|
{ "__llvm_slsblr_thunk_x7", AArch64::X7},
|
|
{ "__llvm_slsblr_thunk_x8", AArch64::X8},
|
|
{ "__llvm_slsblr_thunk_x9", AArch64::X9},
|
|
{ "__llvm_slsblr_thunk_x10", AArch64::X10},
|
|
{ "__llvm_slsblr_thunk_x11", AArch64::X11},
|
|
{ "__llvm_slsblr_thunk_x12", AArch64::X12},
|
|
{ "__llvm_slsblr_thunk_x13", AArch64::X13},
|
|
{ "__llvm_slsblr_thunk_x14", AArch64::X14},
|
|
{ "__llvm_slsblr_thunk_x15", AArch64::X15},
|
|
// X16 and X17 are deliberately missing, as the mitigation requires those
|
|
// register to not be used in BLR. See comment in ConvertBLRToBL for more
|
|
// details.
|
|
{ "__llvm_slsblr_thunk_x18", AArch64::X18},
|
|
{ "__llvm_slsblr_thunk_x19", AArch64::X19},
|
|
{ "__llvm_slsblr_thunk_x20", AArch64::X20},
|
|
{ "__llvm_slsblr_thunk_x21", AArch64::X21},
|
|
{ "__llvm_slsblr_thunk_x22", AArch64::X22},
|
|
{ "__llvm_slsblr_thunk_x23", AArch64::X23},
|
|
{ "__llvm_slsblr_thunk_x24", AArch64::X24},
|
|
{ "__llvm_slsblr_thunk_x25", AArch64::X25},
|
|
{ "__llvm_slsblr_thunk_x26", AArch64::X26},
|
|
{ "__llvm_slsblr_thunk_x27", AArch64::X27},
|
|
{ "__llvm_slsblr_thunk_x28", AArch64::X28},
|
|
{ "__llvm_slsblr_thunk_x29", AArch64::FP},
|
|
// X30 is deliberately missing, for similar reasons as X16 and X17 are
|
|
// missing.
|
|
{ "__llvm_slsblr_thunk_x31", AArch64::XZR},
|
|
};
|
|
|
|
namespace {
|
|
struct SLSBLRThunkInserter : ThunkInserter<SLSBLRThunkInserter> {
|
|
const char *getThunkPrefix() { return SLSBLRNamePrefix; }
|
|
bool mayUseThunk(const MachineFunction &MF) {
|
|
ComdatThunks &= !MF.getSubtarget<AArch64Subtarget>().hardenSlsNoComdat();
|
|
// FIXME: This could also check if there are any BLRs in the function
|
|
// to more accurately reflect if a thunk will be needed.
|
|
return MF.getSubtarget<AArch64Subtarget>().hardenSlsBlr();
|
|
}
|
|
void insertThunks(MachineModuleInfo &MMI);
|
|
void populateThunk(MachineFunction &MF);
|
|
|
|
private:
|
|
bool ComdatThunks = true;
|
|
};
|
|
} // namespace
|
|
|
|
void SLSBLRThunkInserter::insertThunks(MachineModuleInfo &MMI) {
|
|
// FIXME: It probably would be possible to filter which thunks to produce
|
|
// based on which registers are actually used in BLR instructions in this
|
|
// function. But would that be a worthwhile optimization?
|
|
for (auto T : SLSBLRThunks)
|
|
createThunkFunction(MMI, T.Name, ComdatThunks);
|
|
}
|
|
|
|
void SLSBLRThunkInserter::populateThunk(MachineFunction &MF) {
|
|
// FIXME: How to better communicate Register number, rather than through
|
|
// name and lookup table?
|
|
assert(MF.getName().startswith(getThunkPrefix()));
|
|
auto ThunkIt = llvm::find_if(
|
|
SLSBLRThunks, [&MF](auto T) { return T.Name == MF.getName(); });
|
|
assert(ThunkIt != std::end(SLSBLRThunks));
|
|
Register ThunkReg = ThunkIt->Reg;
|
|
|
|
const TargetInstrInfo *TII =
|
|
MF.getSubtarget<AArch64Subtarget>().getInstrInfo();
|
|
assert (MF.size() == 1);
|
|
MachineBasicBlock *Entry = &MF.front();
|
|
Entry->clear();
|
|
|
|
// These thunks need to consist of the following instructions:
|
|
// __llvm_slsblr_thunk_xN:
|
|
// BR xN
|
|
// barrierInsts
|
|
Entry->addLiveIn(ThunkReg);
|
|
// MOV X16, ThunkReg == ORR X16, XZR, ThunkReg, LSL #0
|
|
BuildMI(Entry, DebugLoc(), TII->get(AArch64::ORRXrs), AArch64::X16)
|
|
.addReg(AArch64::XZR)
|
|
.addReg(ThunkReg)
|
|
.addImm(0);
|
|
BuildMI(Entry, DebugLoc(), TII->get(AArch64::BR)).addReg(AArch64::X16);
|
|
// Make sure the thunks do not make use of the SB extension in case there is
|
|
// a function somewhere that will call to it that for some reason disabled
|
|
// the SB extension locally on that function, even though it's enabled for
|
|
// the module otherwise. Therefore set AlwaysUseISBSDB to true.
|
|
insertSpeculationBarrier(&MF.getSubtarget<AArch64Subtarget>(), *Entry,
|
|
Entry->end(), DebugLoc(), true /*AlwaysUseISBDSB*/);
|
|
}
|
|
|
|
MachineBasicBlock &
|
|
AArch64SLSHardening::ConvertBLRToBL(MachineBasicBlock &MBB,
|
|
MachineBasicBlock::iterator MBBI) const {
|
|
// Transform a BLR to a BL as follows:
|
|
// Before:
|
|
// |-----------------------------|
|
|
// | ... |
|
|
// | instI |
|
|
// | BLR xN |
|
|
// | instJ |
|
|
// | ... |
|
|
// |-----------------------------|
|
|
//
|
|
// After:
|
|
// |-----------------------------|
|
|
// | ... |
|
|
// | instI |
|
|
// | BL __llvm_slsblr_thunk_xN |
|
|
// | instJ |
|
|
// | ... |
|
|
// |-----------------------------|
|
|
//
|
|
// __llvm_slsblr_thunk_xN:
|
|
// |-----------------------------|
|
|
// | BR xN |
|
|
// | barrierInsts |
|
|
// |-----------------------------|
|
|
//
|
|
// The __llvm_slsblr_thunk_xN thunks are created by the SLSBLRThunkInserter.
|
|
// This function merely needs to transform BLR xN into BL
|
|
// __llvm_slsblr_thunk_xN.
|
|
//
|
|
// Since linkers are allowed to clobber X16 and X17 on function calls, the
|
|
// above mitigation only works if the original BLR instruction was not
|
|
// BLR X16 nor BLR X17. Code generation before must make sure that no BLR
|
|
// X16|X17 was produced if the mitigation is enabled.
|
|
|
|
MachineInstr &BLR = *MBBI;
|
|
assert(isBLR(BLR));
|
|
unsigned BLOpcode;
|
|
Register Reg;
|
|
bool RegIsKilled;
|
|
switch (BLR.getOpcode()) {
|
|
case AArch64::BLR:
|
|
case AArch64::BLRNoIP:
|
|
BLOpcode = AArch64::BL;
|
|
Reg = BLR.getOperand(0).getReg();
|
|
assert(Reg != AArch64::X16 && Reg != AArch64::X17 && Reg != AArch64::LR);
|
|
RegIsKilled = BLR.getOperand(0).isKill();
|
|
break;
|
|
case AArch64::BLRAA:
|
|
case AArch64::BLRAB:
|
|
case AArch64::BLRAAZ:
|
|
case AArch64::BLRABZ:
|
|
llvm_unreachable("BLRA instructions cannot yet be produced by LLVM, "
|
|
"therefore there is no need to support them for now.");
|
|
default:
|
|
llvm_unreachable("unhandled BLR");
|
|
}
|
|
DebugLoc DL = BLR.getDebugLoc();
|
|
|
|
// If we'd like to support also BLRAA and BLRAB instructions, we'd need
|
|
// a lot more different kind of thunks.
|
|
// For example, a
|
|
//
|
|
// BLRAA xN, xM
|
|
//
|
|
// instruction probably would need to be transformed to something like:
|
|
//
|
|
// BL __llvm_slsblraa_thunk_x<N>_x<M>
|
|
//
|
|
// __llvm_slsblraa_thunk_x<N>_x<M>:
|
|
// BRAA x<N>, x<M>
|
|
// barrierInsts
|
|
//
|
|
// Given that about 30 different values of N are possible and about 30
|
|
// different values of M are possible in the above, with the current way
|
|
// of producing indirect thunks, we'd be producing about 30 times 30, i.e.
|
|
// about 900 thunks (where most might not be actually called). This would
|
|
// multiply further by two to support both BLRAA and BLRAB variants of those
|
|
// instructions.
|
|
// If we'd want to support this, we'd probably need to look into a different
|
|
// way to produce thunk functions, based on which variants are actually
|
|
// needed, rather than producing all possible variants.
|
|
// So far, LLVM does never produce BLRA* instructions, so let's leave this
|
|
// for the future when LLVM can start producing BLRA* instructions.
|
|
MachineFunction &MF = *MBBI->getMF();
|
|
MCContext &Context = MBB.getParent()->getContext();
|
|
auto ThunkIt =
|
|
llvm::find_if(SLSBLRThunks, [Reg](auto T) { return T.Reg == Reg; });
|
|
assert (ThunkIt != std::end(SLSBLRThunks));
|
|
MCSymbol *Sym = Context.getOrCreateSymbol(ThunkIt->Name);
|
|
|
|
MachineInstr *BL = BuildMI(MBB, MBBI, DL, TII->get(BLOpcode)).addSym(Sym);
|
|
|
|
// Now copy the implicit operands from BLR to BL and copy other necessary
|
|
// info.
|
|
// However, both BLR and BL instructions implictly use SP and implicitly
|
|
// define LR. Blindly copying implicit operands would result in SP and LR
|
|
// operands to be present multiple times. While this may not be too much of
|
|
// an issue, let's avoid that for cleanliness, by removing those implicit
|
|
// operands from the BL created above before we copy over all implicit
|
|
// operands from the BLR.
|
|
int ImpLROpIdx = -1;
|
|
int ImpSPOpIdx = -1;
|
|
for (unsigned OpIdx = BL->getNumExplicitOperands();
|
|
OpIdx < BL->getNumOperands(); OpIdx++) {
|
|
MachineOperand Op = BL->getOperand(OpIdx);
|
|
if (!Op.isReg())
|
|
continue;
|
|
if (Op.getReg() == AArch64::LR && Op.isDef())
|
|
ImpLROpIdx = OpIdx;
|
|
if (Op.getReg() == AArch64::SP && !Op.isDef())
|
|
ImpSPOpIdx = OpIdx;
|
|
}
|
|
assert(ImpLROpIdx != -1);
|
|
assert(ImpSPOpIdx != -1);
|
|
int FirstOpIdxToRemove = std::max(ImpLROpIdx, ImpSPOpIdx);
|
|
int SecondOpIdxToRemove = std::min(ImpLROpIdx, ImpSPOpIdx);
|
|
BL->RemoveOperand(FirstOpIdxToRemove);
|
|
BL->RemoveOperand(SecondOpIdxToRemove);
|
|
// Now copy over the implicit operands from the original BLR
|
|
BL->copyImplicitOps(MF, BLR);
|
|
MF.moveCallSiteInfo(&BLR, BL);
|
|
// Also add the register called in the BLR as being used in the called thunk.
|
|
BL->addOperand(MachineOperand::CreateReg(Reg, false /*isDef*/, true /*isImp*/,
|
|
RegIsKilled /*isKill*/));
|
|
// Remove BLR instruction
|
|
MBB.erase(MBBI);
|
|
|
|
return MBB;
|
|
}
|
|
|
|
bool AArch64SLSHardening::hardenBLRs(MachineBasicBlock &MBB) const {
|
|
if (!ST->hardenSlsBlr())
|
|
return false;
|
|
bool Modified = false;
|
|
MachineBasicBlock::iterator MBBI = MBB.begin(), E = MBB.end();
|
|
MachineBasicBlock::iterator NextMBBI;
|
|
for (; MBBI != E; MBBI = NextMBBI) {
|
|
MachineInstr &MI = *MBBI;
|
|
NextMBBI = std::next(MBBI);
|
|
if (isBLR(MI)) {
|
|
ConvertBLRToBL(MBB, MBBI);
|
|
Modified = true;
|
|
}
|
|
}
|
|
return Modified;
|
|
}
|
|
|
|
FunctionPass *llvm::createAArch64SLSHardeningPass() {
|
|
return new AArch64SLSHardening();
|
|
}
|
|
|
|
namespace {
|
|
class AArch64IndirectThunks : public MachineFunctionPass {
|
|
public:
|
|
static char ID;
|
|
|
|
AArch64IndirectThunks() : MachineFunctionPass(ID) {}
|
|
|
|
StringRef getPassName() const override { return "AArch64 Indirect Thunks"; }
|
|
|
|
bool doInitialization(Module &M) override;
|
|
bool runOnMachineFunction(MachineFunction &MF) override;
|
|
|
|
private:
|
|
std::tuple<SLSBLRThunkInserter> TIs;
|
|
|
|
// FIXME: When LLVM moves to C++17, these can become folds
|
|
template <typename... ThunkInserterT>
|
|
static void initTIs(Module &M,
|
|
std::tuple<ThunkInserterT...> &ThunkInserters) {
|
|
(void)std::initializer_list<int>{
|
|
(std::get<ThunkInserterT>(ThunkInserters).init(M), 0)...};
|
|
}
|
|
template <typename... ThunkInserterT>
|
|
static bool runTIs(MachineModuleInfo &MMI, MachineFunction &MF,
|
|
std::tuple<ThunkInserterT...> &ThunkInserters) {
|
|
bool Modified = false;
|
|
(void)std::initializer_list<int>{
|
|
Modified |= std::get<ThunkInserterT>(ThunkInserters).run(MMI, MF)...};
|
|
return Modified;
|
|
}
|
|
};
|
|
|
|
} // end anonymous namespace
|
|
|
|
char AArch64IndirectThunks::ID = 0;
|
|
|
|
FunctionPass *llvm::createAArch64IndirectThunks() {
|
|
return new AArch64IndirectThunks();
|
|
}
|
|
|
|
bool AArch64IndirectThunks::doInitialization(Module &M) {
|
|
initTIs(M, TIs);
|
|
return false;
|
|
}
|
|
|
|
bool AArch64IndirectThunks::runOnMachineFunction(MachineFunction &MF) {
|
|
LLVM_DEBUG(dbgs() << getPassName() << '\n');
|
|
auto &MMI = getAnalysis<MachineModuleInfoWrapperPass>().getMMI();
|
|
return runTIs(MMI, MF, TIs);
|
|
}
|