mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-25 12:12:47 +01:00
58e3420193
Add explicit type i64 to RV64 only patterns to stop emitting unneeded i32 patterns. It can reduce the isel table size. Reviewed By: craig.topper Differential Revision: https://reviews.llvm.org/D100089
391 lines
17 KiB
TableGen
391 lines
17 KiB
TableGen
//===-- RISCVInstrInfoA.td - RISC-V 'A' instructions -------*- tablegen -*-===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// This file describes the RISC-V instructions from the standard 'A', Atomic
|
|
// Instructions extension.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// Operand and SDNode transformation definitions.
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// A parse method for (${gpr}) or 0(${gpr}), where the 0 is be silently ignored.
|
|
// Used for GNU as Compatibility.
|
|
def AtomicMemOpOperand : AsmOperandClass {
|
|
let Name = "AtomicMemOpOperand";
|
|
let RenderMethod = "addRegOperands";
|
|
let PredicateMethod = "isGPR";
|
|
let ParserMethod = "parseAtomicMemOp";
|
|
}
|
|
|
|
def GPRMemAtomic : RegisterOperand<GPR> {
|
|
let ParserMatchClass = AtomicMemOpOperand;
|
|
let PrintMethod = "printAtomicMemOp";
|
|
}
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// Instruction class templates
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
let hasSideEffects = 0, mayLoad = 1, mayStore = 0 in
|
|
class LR_r<bit aq, bit rl, bits<3> funct3, string opcodestr>
|
|
: RVInstRAtomic<0b00010, aq, rl, funct3, OPC_AMO,
|
|
(outs GPR:$rd), (ins GPRMemAtomic:$rs1),
|
|
opcodestr, "$rd, $rs1"> {
|
|
let rs2 = 0;
|
|
}
|
|
|
|
multiclass LR_r_aq_rl<bits<3> funct3, string opcodestr> {
|
|
def "" : LR_r<0, 0, funct3, opcodestr>;
|
|
def _AQ : LR_r<1, 0, funct3, opcodestr # ".aq">;
|
|
def _RL : LR_r<0, 1, funct3, opcodestr # ".rl">;
|
|
def _AQ_RL : LR_r<1, 1, funct3, opcodestr # ".aqrl">;
|
|
}
|
|
|
|
let hasSideEffects = 0, mayLoad = 1, mayStore = 1 in
|
|
class AMO_rr<bits<5> funct5, bit aq, bit rl, bits<3> funct3, string opcodestr>
|
|
: RVInstRAtomic<funct5, aq, rl, funct3, OPC_AMO,
|
|
(outs GPR:$rd), (ins GPRMemAtomic:$rs1, GPR:$rs2),
|
|
opcodestr, "$rd, $rs2, $rs1">;
|
|
|
|
multiclass AMO_rr_aq_rl<bits<5> funct5, bits<3> funct3, string opcodestr> {
|
|
def "" : AMO_rr<funct5, 0, 0, funct3, opcodestr>;
|
|
def _AQ : AMO_rr<funct5, 1, 0, funct3, opcodestr # ".aq">;
|
|
def _RL : AMO_rr<funct5, 0, 1, funct3, opcodestr # ".rl">;
|
|
def _AQ_RL : AMO_rr<funct5, 1, 1, funct3, opcodestr # ".aqrl">;
|
|
}
|
|
|
|
multiclass AtomicStPat<PatFrag StoreOp, RVInst Inst, RegisterClass StTy,
|
|
ValueType vt = XLenVT> {
|
|
def : Pat<(StoreOp BaseAddr:$rs1, (vt StTy:$rs2)),
|
|
(Inst StTy:$rs2, BaseAddr:$rs1, 0)>;
|
|
def : Pat<(StoreOp (add BaseAddr:$rs1, simm12:$imm12), (vt StTy:$rs2)),
|
|
(Inst StTy:$rs2, BaseAddr:$rs1, simm12:$imm12)>;
|
|
def : Pat<(StoreOp (IsOrAdd AddrFI:$rs1, simm12:$imm12), (vt StTy:$rs2)),
|
|
(Inst StTy:$rs2, AddrFI:$rs1, simm12:$imm12)>;
|
|
}
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// Instructions
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
let Predicates = [HasStdExtA] in {
|
|
defm LR_W : LR_r_aq_rl<0b010, "lr.w">, Sched<[WriteAtomicLDW, ReadAtomicLDW]>;
|
|
defm SC_W : AMO_rr_aq_rl<0b00011, 0b010, "sc.w">,
|
|
Sched<[WriteAtomicSTW, ReadAtomicSTW, ReadAtomicSTW]>;
|
|
defm AMOSWAP_W : AMO_rr_aq_rl<0b00001, 0b010, "amoswap.w">,
|
|
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
|
defm AMOADD_W : AMO_rr_aq_rl<0b00000, 0b010, "amoadd.w">,
|
|
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
|
defm AMOXOR_W : AMO_rr_aq_rl<0b00100, 0b010, "amoxor.w">,
|
|
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
|
defm AMOAND_W : AMO_rr_aq_rl<0b01100, 0b010, "amoand.w">,
|
|
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
|
defm AMOOR_W : AMO_rr_aq_rl<0b01000, 0b010, "amoor.w">,
|
|
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
|
defm AMOMIN_W : AMO_rr_aq_rl<0b10000, 0b010, "amomin.w">,
|
|
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
|
defm AMOMAX_W : AMO_rr_aq_rl<0b10100, 0b010, "amomax.w">,
|
|
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
|
defm AMOMINU_W : AMO_rr_aq_rl<0b11000, 0b010, "amominu.w">,
|
|
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
|
defm AMOMAXU_W : AMO_rr_aq_rl<0b11100, 0b010, "amomaxu.w">,
|
|
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
|
} // Predicates = [HasStdExtA]
|
|
|
|
let Predicates = [HasStdExtA, IsRV64] in {
|
|
defm LR_D : LR_r_aq_rl<0b011, "lr.d">, Sched<[WriteAtomicLDD, ReadAtomicLDD]>;
|
|
defm SC_D : AMO_rr_aq_rl<0b00011, 0b011, "sc.d">,
|
|
Sched<[WriteAtomicSTD, ReadAtomicSTD, ReadAtomicSTD]>;
|
|
defm AMOSWAP_D : AMO_rr_aq_rl<0b00001, 0b011, "amoswap.d">,
|
|
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
|
defm AMOADD_D : AMO_rr_aq_rl<0b00000, 0b011, "amoadd.d">,
|
|
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
|
defm AMOXOR_D : AMO_rr_aq_rl<0b00100, 0b011, "amoxor.d">,
|
|
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
|
defm AMOAND_D : AMO_rr_aq_rl<0b01100, 0b011, "amoand.d">,
|
|
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
|
defm AMOOR_D : AMO_rr_aq_rl<0b01000, 0b011, "amoor.d">,
|
|
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
|
defm AMOMIN_D : AMO_rr_aq_rl<0b10000, 0b011, "amomin.d">,
|
|
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
|
defm AMOMAX_D : AMO_rr_aq_rl<0b10100, 0b011, "amomax.d">,
|
|
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
|
defm AMOMINU_D : AMO_rr_aq_rl<0b11000, 0b011, "amominu.d">,
|
|
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
|
defm AMOMAXU_D : AMO_rr_aq_rl<0b11100, 0b011, "amomaxu.d">,
|
|
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
|
} // Predicates = [HasStdExtA, IsRV64]
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// Pseudo-instructions and codegen patterns
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
let Predicates = [HasStdExtA] in {
|
|
|
|
/// Atomic loads and stores
|
|
|
|
// Fences will be inserted for atomic load/stores according to the logic in
|
|
// RISCVTargetLowering::{emitLeadingFence,emitTrailingFence}.
|
|
|
|
defm : LdPat<atomic_load_8, LB>;
|
|
defm : LdPat<atomic_load_16, LH>;
|
|
defm : LdPat<atomic_load_32, LW>;
|
|
|
|
defm : AtomicStPat<atomic_store_8, SB, GPR>;
|
|
defm : AtomicStPat<atomic_store_16, SH, GPR>;
|
|
defm : AtomicStPat<atomic_store_32, SW, GPR>;
|
|
|
|
/// AMOs
|
|
|
|
multiclass AMOPat<string AtomicOp, string BaseInst> {
|
|
def : PatGprGpr<!cast<PatFrag>(AtomicOp#"_monotonic"),
|
|
!cast<RVInst>(BaseInst)>;
|
|
def : PatGprGpr<!cast<PatFrag>(AtomicOp#"_acquire"),
|
|
!cast<RVInst>(BaseInst#"_AQ")>;
|
|
def : PatGprGpr<!cast<PatFrag>(AtomicOp#"_release"),
|
|
!cast<RVInst>(BaseInst#"_RL")>;
|
|
def : PatGprGpr<!cast<PatFrag>(AtomicOp#"_acq_rel"),
|
|
!cast<RVInst>(BaseInst#"_AQ_RL")>;
|
|
def : PatGprGpr<!cast<PatFrag>(AtomicOp#"_seq_cst"),
|
|
!cast<RVInst>(BaseInst#"_AQ_RL")>;
|
|
}
|
|
|
|
defm : AMOPat<"atomic_swap_32", "AMOSWAP_W">;
|
|
defm : AMOPat<"atomic_load_add_32", "AMOADD_W">;
|
|
defm : AMOPat<"atomic_load_and_32", "AMOAND_W">;
|
|
defm : AMOPat<"atomic_load_or_32", "AMOOR_W">;
|
|
defm : AMOPat<"atomic_load_xor_32", "AMOXOR_W">;
|
|
defm : AMOPat<"atomic_load_max_32", "AMOMAX_W">;
|
|
defm : AMOPat<"atomic_load_min_32", "AMOMIN_W">;
|
|
defm : AMOPat<"atomic_load_umax_32", "AMOMAXU_W">;
|
|
defm : AMOPat<"atomic_load_umin_32", "AMOMINU_W">;
|
|
|
|
def : Pat<(atomic_load_sub_32_monotonic GPR:$addr, GPR:$incr),
|
|
(AMOADD_W GPR:$addr, (SUB X0, GPR:$incr))>;
|
|
def : Pat<(atomic_load_sub_32_acquire GPR:$addr, GPR:$incr),
|
|
(AMOADD_W_AQ GPR:$addr, (SUB X0, GPR:$incr))>;
|
|
def : Pat<(atomic_load_sub_32_release GPR:$addr, GPR:$incr),
|
|
(AMOADD_W_RL GPR:$addr, (SUB X0, GPR:$incr))>;
|
|
def : Pat<(atomic_load_sub_32_acq_rel GPR:$addr, GPR:$incr),
|
|
(AMOADD_W_AQ_RL GPR:$addr, (SUB X0, GPR:$incr))>;
|
|
def : Pat<(atomic_load_sub_32_seq_cst GPR:$addr, GPR:$incr),
|
|
(AMOADD_W_AQ_RL GPR:$addr, (SUB X0, GPR:$incr))>;
|
|
|
|
/// Pseudo AMOs
|
|
|
|
class PseudoAMO : Pseudo<(outs GPR:$res, GPR:$scratch),
|
|
(ins GPR:$addr, GPR:$incr, ixlenimm:$ordering), []> {
|
|
let Constraints = "@earlyclobber $res,@earlyclobber $scratch";
|
|
let mayLoad = 1;
|
|
let mayStore = 1;
|
|
let hasSideEffects = 0;
|
|
}
|
|
|
|
def PseudoAtomicLoadNand32 : PseudoAMO;
|
|
// Ordering constants must be kept in sync with the AtomicOrdering enum in
|
|
// AtomicOrdering.h.
|
|
def : Pat<(atomic_load_nand_32_monotonic GPR:$addr, GPR:$incr),
|
|
(PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 2)>;
|
|
def : Pat<(atomic_load_nand_32_acquire GPR:$addr, GPR:$incr),
|
|
(PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 4)>;
|
|
def : Pat<(atomic_load_nand_32_release GPR:$addr, GPR:$incr),
|
|
(PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 5)>;
|
|
def : Pat<(atomic_load_nand_32_acq_rel GPR:$addr, GPR:$incr),
|
|
(PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 6)>;
|
|
def : Pat<(atomic_load_nand_32_seq_cst GPR:$addr, GPR:$incr),
|
|
(PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 7)>;
|
|
|
|
class PseudoMaskedAMO
|
|
: Pseudo<(outs GPR:$res, GPR:$scratch),
|
|
(ins GPR:$addr, GPR:$incr, GPR:$mask, ixlenimm:$ordering), []> {
|
|
let Constraints = "@earlyclobber $res,@earlyclobber $scratch";
|
|
let mayLoad = 1;
|
|
let mayStore = 1;
|
|
let hasSideEffects = 0;
|
|
}
|
|
|
|
class PseudoMaskedAMOMinMax
|
|
: Pseudo<(outs GPR:$res, GPR:$scratch1, GPR:$scratch2),
|
|
(ins GPR:$addr, GPR:$incr, GPR:$mask, ixlenimm:$sextshamt,
|
|
ixlenimm:$ordering), []> {
|
|
let Constraints = "@earlyclobber $res,@earlyclobber $scratch1,"
|
|
"@earlyclobber $scratch2";
|
|
let mayLoad = 1;
|
|
let mayStore = 1;
|
|
let hasSideEffects = 0;
|
|
}
|
|
|
|
class PseudoMaskedAMOUMinUMax
|
|
: Pseudo<(outs GPR:$res, GPR:$scratch1, GPR:$scratch2),
|
|
(ins GPR:$addr, GPR:$incr, GPR:$mask, ixlenimm:$ordering), []> {
|
|
let Constraints = "@earlyclobber $res,@earlyclobber $scratch1,"
|
|
"@earlyclobber $scratch2";
|
|
let mayLoad = 1;
|
|
let mayStore = 1;
|
|
let hasSideEffects = 0;
|
|
}
|
|
|
|
class PseudoMaskedAMOPat<Intrinsic intrin, Pseudo AMOInst>
|
|
: Pat<(intrin GPR:$addr, GPR:$incr, GPR:$mask, timm:$ordering),
|
|
(AMOInst GPR:$addr, GPR:$incr, GPR:$mask, timm:$ordering)>;
|
|
|
|
class PseudoMaskedAMOMinMaxPat<Intrinsic intrin, Pseudo AMOInst>
|
|
: Pat<(intrin GPR:$addr, GPR:$incr, GPR:$mask, GPR:$shiftamt,
|
|
timm:$ordering),
|
|
(AMOInst GPR:$addr, GPR:$incr, GPR:$mask, GPR:$shiftamt,
|
|
timm:$ordering)>;
|
|
|
|
def PseudoMaskedAtomicSwap32 : PseudoMaskedAMO;
|
|
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_xchg_i32,
|
|
PseudoMaskedAtomicSwap32>;
|
|
def PseudoMaskedAtomicLoadAdd32 : PseudoMaskedAMO;
|
|
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_add_i32,
|
|
PseudoMaskedAtomicLoadAdd32>;
|
|
def PseudoMaskedAtomicLoadSub32 : PseudoMaskedAMO;
|
|
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_sub_i32,
|
|
PseudoMaskedAtomicLoadSub32>;
|
|
def PseudoMaskedAtomicLoadNand32 : PseudoMaskedAMO;
|
|
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_nand_i32,
|
|
PseudoMaskedAtomicLoadNand32>;
|
|
def PseudoMaskedAtomicLoadMax32 : PseudoMaskedAMOMinMax;
|
|
def : PseudoMaskedAMOMinMaxPat<int_riscv_masked_atomicrmw_max_i32,
|
|
PseudoMaskedAtomicLoadMax32>;
|
|
def PseudoMaskedAtomicLoadMin32 : PseudoMaskedAMOMinMax;
|
|
def : PseudoMaskedAMOMinMaxPat<int_riscv_masked_atomicrmw_min_i32,
|
|
PseudoMaskedAtomicLoadMin32>;
|
|
def PseudoMaskedAtomicLoadUMax32 : PseudoMaskedAMOUMinUMax;
|
|
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_umax_i32,
|
|
PseudoMaskedAtomicLoadUMax32>;
|
|
def PseudoMaskedAtomicLoadUMin32 : PseudoMaskedAMOUMinUMax;
|
|
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_umin_i32,
|
|
PseudoMaskedAtomicLoadUMin32>;
|
|
|
|
/// Compare and exchange
|
|
|
|
class PseudoCmpXchg
|
|
: Pseudo<(outs GPR:$res, GPR:$scratch),
|
|
(ins GPR:$addr, GPR:$cmpval, GPR:$newval, ixlenimm:$ordering), []> {
|
|
let Constraints = "@earlyclobber $res,@earlyclobber $scratch";
|
|
let mayLoad = 1;
|
|
let mayStore = 1;
|
|
let hasSideEffects = 0;
|
|
}
|
|
|
|
// Ordering constants must be kept in sync with the AtomicOrdering enum in
|
|
// AtomicOrdering.h.
|
|
multiclass PseudoCmpXchgPat<string Op, Pseudo CmpXchgInst> {
|
|
def : Pat<(!cast<PatFrag>(Op#"_monotonic") GPR:$addr, GPR:$cmp, GPR:$new),
|
|
(CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 2)>;
|
|
def : Pat<(!cast<PatFrag>(Op#"_acquire") GPR:$addr, GPR:$cmp, GPR:$new),
|
|
(CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 4)>;
|
|
def : Pat<(!cast<PatFrag>(Op#"_release") GPR:$addr, GPR:$cmp, GPR:$new),
|
|
(CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 5)>;
|
|
def : Pat<(!cast<PatFrag>(Op#"_acq_rel") GPR:$addr, GPR:$cmp, GPR:$new),
|
|
(CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 6)>;
|
|
def : Pat<(!cast<PatFrag>(Op#"_seq_cst") GPR:$addr, GPR:$cmp, GPR:$new),
|
|
(CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 7)>;
|
|
}
|
|
|
|
def PseudoCmpXchg32 : PseudoCmpXchg;
|
|
defm : PseudoCmpXchgPat<"atomic_cmp_swap_32", PseudoCmpXchg32>;
|
|
|
|
def PseudoMaskedCmpXchg32
|
|
: Pseudo<(outs GPR:$res, GPR:$scratch),
|
|
(ins GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask,
|
|
ixlenimm:$ordering), []> {
|
|
let Constraints = "@earlyclobber $res,@earlyclobber $scratch";
|
|
let mayLoad = 1;
|
|
let mayStore = 1;
|
|
let hasSideEffects = 0;
|
|
}
|
|
|
|
def : Pat<(int_riscv_masked_cmpxchg_i32
|
|
GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, timm:$ordering),
|
|
(PseudoMaskedCmpXchg32
|
|
GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, timm:$ordering)>;
|
|
|
|
} // Predicates = [HasStdExtA]
|
|
|
|
let Predicates = [HasStdExtA, IsRV64] in {
|
|
|
|
/// 64-bit atomic loads and stores
|
|
|
|
// Fences will be inserted for atomic load/stores according to the logic in
|
|
// RISCVTargetLowering::{emitLeadingFence,emitTrailingFence}.
|
|
defm : LdPat<atomic_load_64, LD, i64>;
|
|
defm : AtomicStPat<atomic_store_64, SD, GPR, i64>;
|
|
|
|
defm : AMOPat<"atomic_swap_64", "AMOSWAP_D">;
|
|
defm : AMOPat<"atomic_load_add_64", "AMOADD_D">;
|
|
defm : AMOPat<"atomic_load_and_64", "AMOAND_D">;
|
|
defm : AMOPat<"atomic_load_or_64", "AMOOR_D">;
|
|
defm : AMOPat<"atomic_load_xor_64", "AMOXOR_D">;
|
|
defm : AMOPat<"atomic_load_max_64", "AMOMAX_D">;
|
|
defm : AMOPat<"atomic_load_min_64", "AMOMIN_D">;
|
|
defm : AMOPat<"atomic_load_umax_64", "AMOMAXU_D">;
|
|
defm : AMOPat<"atomic_load_umin_64", "AMOMINU_D">;
|
|
|
|
/// 64-bit AMOs
|
|
|
|
def : Pat<(i64 (atomic_load_sub_64_monotonic GPR:$addr, GPR:$incr)),
|
|
(AMOADD_D GPR:$addr, (SUB X0, GPR:$incr))>;
|
|
def : Pat<(i64 (atomic_load_sub_64_acquire GPR:$addr, GPR:$incr)),
|
|
(AMOADD_D_AQ GPR:$addr, (SUB X0, GPR:$incr))>;
|
|
def : Pat<(i64 (atomic_load_sub_64_release GPR:$addr, GPR:$incr)),
|
|
(AMOADD_D_RL GPR:$addr, (SUB X0, GPR:$incr))>;
|
|
def : Pat<(i64 (atomic_load_sub_64_acq_rel GPR:$addr, GPR:$incr)),
|
|
(AMOADD_D_AQ_RL GPR:$addr, (SUB X0, GPR:$incr))>;
|
|
def : Pat<(i64 (atomic_load_sub_64_seq_cst GPR:$addr, GPR:$incr)),
|
|
(AMOADD_D_AQ_RL GPR:$addr, (SUB X0, GPR:$incr))>;
|
|
|
|
/// 64-bit pseudo AMOs
|
|
|
|
def PseudoAtomicLoadNand64 : PseudoAMO;
|
|
// Ordering constants must be kept in sync with the AtomicOrdering enum in
|
|
// AtomicOrdering.h.
|
|
def : Pat<(i64 (atomic_load_nand_64_monotonic GPR:$addr, GPR:$incr)),
|
|
(PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 2)>;
|
|
def : Pat<(i64 (atomic_load_nand_64_acquire GPR:$addr, GPR:$incr)),
|
|
(PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 4)>;
|
|
def : Pat<(i64 (atomic_load_nand_64_release GPR:$addr, GPR:$incr)),
|
|
(PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 5)>;
|
|
def : Pat<(i64 (atomic_load_nand_64_acq_rel GPR:$addr, GPR:$incr)),
|
|
(PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 6)>;
|
|
def : Pat<(i64 (atomic_load_nand_64_seq_cst GPR:$addr, GPR:$incr)),
|
|
(PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 7)>;
|
|
|
|
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_xchg_i64,
|
|
PseudoMaskedAtomicSwap32>;
|
|
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_add_i64,
|
|
PseudoMaskedAtomicLoadAdd32>;
|
|
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_sub_i64,
|
|
PseudoMaskedAtomicLoadSub32>;
|
|
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_nand_i64,
|
|
PseudoMaskedAtomicLoadNand32>;
|
|
def : PseudoMaskedAMOMinMaxPat<int_riscv_masked_atomicrmw_max_i64,
|
|
PseudoMaskedAtomicLoadMax32>;
|
|
def : PseudoMaskedAMOMinMaxPat<int_riscv_masked_atomicrmw_min_i64,
|
|
PseudoMaskedAtomicLoadMin32>;
|
|
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_umax_i64,
|
|
PseudoMaskedAtomicLoadUMax32>;
|
|
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_umin_i64,
|
|
PseudoMaskedAtomicLoadUMin32>;
|
|
|
|
/// 64-bit compare and exchange
|
|
|
|
def PseudoCmpXchg64 : PseudoCmpXchg;
|
|
defm : PseudoCmpXchgPat<"atomic_cmp_swap_64", PseudoCmpXchg64>;
|
|
|
|
def : Pat<(int_riscv_masked_cmpxchg_i64
|
|
GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, timm:$ordering),
|
|
(PseudoMaskedCmpXchg32
|
|
GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, timm:$ordering)>;
|
|
} // Predicates = [HasStdExtA, IsRV64]
|