2020-12-07 16:44:40 +01:00
|
|
|
//===-- MVETailPredUtils.h - Tail predication utility functions -*- C++-*-===//
|
|
|
|
//
|
|
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
|
|
|
// This file contains utility functions for low overhead and tail predicated
|
|
|
|
// loops, shared between the ARMLowOverheadLoops pass and anywhere else that
|
|
|
|
// needs them.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
|
|
#ifndef LLVM_LIB_TARGET_ARM_MVETAILPREDUTILS_H
|
|
|
|
#define LLVM_LIB_TARGET_ARM_MVETAILPREDUTILS_H
|
|
|
|
|
|
|
|
#include "llvm/CodeGen/MachineInstr.h"
|
|
|
|
#include "llvm/CodeGen/MachineInstrBuilder.h"
|
|
|
|
#include "llvm/CodeGen/MachineOperand.h"
|
|
|
|
#include "llvm/CodeGen/TargetInstrInfo.h"
|
|
|
|
|
|
|
|
namespace llvm {
|
|
|
|
|
|
|
|
static inline unsigned VCTPOpcodeToLSTP(unsigned Opcode, bool IsDoLoop) {
|
|
|
|
switch (Opcode) {
|
|
|
|
default:
|
|
|
|
llvm_unreachable("unhandled vctp opcode");
|
|
|
|
break;
|
|
|
|
case ARM::MVE_VCTP8:
|
|
|
|
return IsDoLoop ? ARM::MVE_DLSTP_8 : ARM::MVE_WLSTP_8;
|
|
|
|
case ARM::MVE_VCTP16:
|
|
|
|
return IsDoLoop ? ARM::MVE_DLSTP_16 : ARM::MVE_WLSTP_16;
|
|
|
|
case ARM::MVE_VCTP32:
|
|
|
|
return IsDoLoop ? ARM::MVE_DLSTP_32 : ARM::MVE_WLSTP_32;
|
|
|
|
case ARM::MVE_VCTP64:
|
|
|
|
return IsDoLoop ? ARM::MVE_DLSTP_64 : ARM::MVE_WLSTP_64;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline unsigned getTailPredVectorWidth(unsigned Opcode) {
|
|
|
|
switch (Opcode) {
|
|
|
|
default:
|
|
|
|
llvm_unreachable("unhandled vctp opcode");
|
|
|
|
case ARM::MVE_VCTP8:
|
|
|
|
return 16;
|
|
|
|
case ARM::MVE_VCTP16:
|
|
|
|
return 8;
|
|
|
|
case ARM::MVE_VCTP32:
|
|
|
|
return 4;
|
|
|
|
case ARM::MVE_VCTP64:
|
|
|
|
return 2;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline bool isVCTP(const MachineInstr *MI) {
|
|
|
|
switch (MI->getOpcode()) {
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
case ARM::MVE_VCTP8:
|
|
|
|
case ARM::MVE_VCTP16:
|
|
|
|
case ARM::MVE_VCTP32:
|
|
|
|
case ARM::MVE_VCTP64:
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2021-06-13 14:55:34 +02:00
|
|
|
static inline bool isDoLoopStart(const MachineInstr &MI) {
|
2020-12-07 16:44:40 +01:00
|
|
|
return MI.getOpcode() == ARM::t2DoLoopStart ||
|
2021-06-13 14:55:34 +02:00
|
|
|
MI.getOpcode() == ARM::t2DoLoopStartTP;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline bool isWhileLoopStart(const MachineInstr &MI) {
|
|
|
|
return MI.getOpcode() == ARM::t2WhileLoopStart ||
|
|
|
|
MI.getOpcode() == ARM::t2WhileLoopStartLR ||
|
|
|
|
MI.getOpcode() == ARM::t2WhileLoopStartTP;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline bool isLoopStart(const MachineInstr &MI) {
|
|
|
|
return isDoLoopStart(MI) || isWhileLoopStart(MI);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return the TargetBB stored in a t2WhileLoopStartLR/t2WhileLoopStartTP.
|
|
|
|
inline MachineBasicBlock *getWhileLoopStartTargetBB(const MachineInstr &MI) {
|
|
|
|
assert(isWhileLoopStart(MI) && "Expected WhileLoopStart!");
|
|
|
|
unsigned Op = MI.getOpcode() == ARM::t2WhileLoopStartTP ? 3 : 2;
|
|
|
|
return MI.getOperand(Op).getMBB();
|
2020-12-07 16:44:40 +01:00
|
|
|
}
|
|
|
|
|
[ARM] Improve WLS lowering
Recently we improved the lowering of low overhead loops and tail
predicated loops, but concentrated first on the DLS do style loops. This
extends those improvements over to the WLS while loops, improving the
chance of lowering them successfully. To do this the lowering has to
change a little as the instructions are terminators that produce a value
- something that needs to be treated carefully.
Lowering starts at the Hardware Loop pass, inserting a new
llvm.test.start.loop.iterations that produces both an i1 to control the
loop entry and an i32 similar to the llvm.start.loop.iterations
intrinsic added for do loops. This feeds into the loop phi, properly
gluing the values together:
%wls = call { i32, i1 } @llvm.test.start.loop.iterations.i32(i32 %div)
%wls0 = extractvalue { i32, i1 } %wls, 0
%wls1 = extractvalue { i32, i1 } %wls, 1
br i1 %wls1, label %loop.ph, label %loop.exit
...
loop:
%lsr.iv = phi i32 [ %wls0, %loop.ph ], [ %iv.next, %loop ]
..
%iv.next = call i32 @llvm.loop.decrement.reg.i32(i32 %lsr.iv, i32 1)
%cmp = icmp ne i32 %iv.next, 0
br i1 %cmp, label %loop, label %loop.exit
The llvm.test.start.loop.iterations need to be lowered through ISel
lowering as a pair of WLS and WLSSETUP nodes, which each get converted
to t2WhileLoopSetup and t2WhileLoopStart Pseudos. This helps prevent
t2WhileLoopStart from being a terminator that produces a value,
something difficult to control at that stage in the pipeline. Instead
the t2WhileLoopSetup produces the value of LR (essentially acting as a
lr = subs rn, 0), t2WhileLoopStart consumes that lr value (the Bcc).
These are then converted into a single t2WhileLoopStartLR at the same
point as t2DoLoopStartTP and t2LoopEndDec. Otherwise we revert the loop
to prevent them from progressing further in the pipeline. The
t2WhileLoopStartLR is a single instruction that takes a GPR and produces
LR, similar to the WLS instruction.
%1:gprlr = t2WhileLoopStartLR %0:rgpr, %bb.3
t2B %bb.1
...
bb.2.loop:
%2:gprlr = PHI %1:gprlr, %bb.1, %3:gprlr, %bb.2
...
%3:gprlr = t2LoopEndDec %2:gprlr, %bb.2
t2B %bb.3
The t2WhileLoopStartLR can then be treated similar to the other low
overhead loop pseudos, eventually being lowered to a WLS providing the
branches are within range.
Differential Revision: https://reviews.llvm.org/D97729
2021-03-11 15:06:04 +01:00
|
|
|
// WhileLoopStart holds the exit block, so produce a subs Op0, Op1, 0 and then a
|
2020-12-07 16:44:40 +01:00
|
|
|
// beq that branches to the exit branch.
|
2021-03-25 17:44:15 +01:00
|
|
|
// If UseCmp is true, this will create a t2CMP instead of a t2SUBri, meaning the
|
|
|
|
// value of LR into the loop will not be setup. This is used if the LR setup is
|
|
|
|
// done via another means (via a t2DoLoopStart, for example).
|
[ARM] Improve WLS lowering
Recently we improved the lowering of low overhead loops and tail
predicated loops, but concentrated first on the DLS do style loops. This
extends those improvements over to the WLS while loops, improving the
chance of lowering them successfully. To do this the lowering has to
change a little as the instructions are terminators that produce a value
- something that needs to be treated carefully.
Lowering starts at the Hardware Loop pass, inserting a new
llvm.test.start.loop.iterations that produces both an i1 to control the
loop entry and an i32 similar to the llvm.start.loop.iterations
intrinsic added for do loops. This feeds into the loop phi, properly
gluing the values together:
%wls = call { i32, i1 } @llvm.test.start.loop.iterations.i32(i32 %div)
%wls0 = extractvalue { i32, i1 } %wls, 0
%wls1 = extractvalue { i32, i1 } %wls, 1
br i1 %wls1, label %loop.ph, label %loop.exit
...
loop:
%lsr.iv = phi i32 [ %wls0, %loop.ph ], [ %iv.next, %loop ]
..
%iv.next = call i32 @llvm.loop.decrement.reg.i32(i32 %lsr.iv, i32 1)
%cmp = icmp ne i32 %iv.next, 0
br i1 %cmp, label %loop, label %loop.exit
The llvm.test.start.loop.iterations need to be lowered through ISel
lowering as a pair of WLS and WLSSETUP nodes, which each get converted
to t2WhileLoopSetup and t2WhileLoopStart Pseudos. This helps prevent
t2WhileLoopStart from being a terminator that produces a value,
something difficult to control at that stage in the pipeline. Instead
the t2WhileLoopSetup produces the value of LR (essentially acting as a
lr = subs rn, 0), t2WhileLoopStart consumes that lr value (the Bcc).
These are then converted into a single t2WhileLoopStartLR at the same
point as t2DoLoopStartTP and t2LoopEndDec. Otherwise we revert the loop
to prevent them from progressing further in the pipeline. The
t2WhileLoopStartLR is a single instruction that takes a GPR and produces
LR, similar to the WLS instruction.
%1:gprlr = t2WhileLoopStartLR %0:rgpr, %bb.3
t2B %bb.1
...
bb.2.loop:
%2:gprlr = PHI %1:gprlr, %bb.1, %3:gprlr, %bb.2
...
%3:gprlr = t2LoopEndDec %2:gprlr, %bb.2
t2B %bb.3
The t2WhileLoopStartLR can then be treated similar to the other low
overhead loop pseudos, eventually being lowered to a WLS providing the
branches are within range.
Differential Revision: https://reviews.llvm.org/D97729
2021-03-11 15:06:04 +01:00
|
|
|
inline void RevertWhileLoopStartLR(MachineInstr *MI, const TargetInstrInfo *TII,
|
2021-03-25 17:44:15 +01:00
|
|
|
unsigned BrOpc = ARM::t2Bcc,
|
|
|
|
bool UseCmp = false) {
|
2020-12-07 16:44:40 +01:00
|
|
|
MachineBasicBlock *MBB = MI->getParent();
|
2021-06-13 14:55:34 +02:00
|
|
|
assert((MI->getOpcode() == ARM::t2WhileLoopStartLR ||
|
|
|
|
MI->getOpcode() == ARM::t2WhileLoopStartTP) &&
|
|
|
|
"Only expected a t2WhileLoopStartLR/TP in RevertWhileLoopStartLR!");
|
2020-12-07 16:44:40 +01:00
|
|
|
|
2021-03-25 17:44:15 +01:00
|
|
|
// Subs/Cmp
|
|
|
|
if (UseCmp) {
|
|
|
|
MachineInstrBuilder MIB =
|
|
|
|
BuildMI(*MBB, MI, MI->getDebugLoc(), TII->get(ARM::t2CMPri));
|
|
|
|
MIB.add(MI->getOperand(1));
|
|
|
|
MIB.addImm(0);
|
|
|
|
MIB.addImm(ARMCC::AL);
|
|
|
|
MIB.addReg(ARM::NoRegister);
|
|
|
|
} else {
|
|
|
|
MachineInstrBuilder MIB =
|
|
|
|
BuildMI(*MBB, MI, MI->getDebugLoc(), TII->get(ARM::t2SUBri));
|
|
|
|
MIB.add(MI->getOperand(0));
|
|
|
|
MIB.add(MI->getOperand(1));
|
|
|
|
MIB.addImm(0);
|
|
|
|
MIB.addImm(ARMCC::AL);
|
|
|
|
MIB.addReg(ARM::NoRegister);
|
|
|
|
MIB.addReg(ARM::CPSR, RegState::Define);
|
|
|
|
}
|
2020-12-07 16:44:40 +01:00
|
|
|
|
|
|
|
// Branch
|
2021-03-25 17:44:15 +01:00
|
|
|
MachineInstrBuilder MIB =
|
|
|
|
BuildMI(*MBB, MI, MI->getDebugLoc(), TII->get(BrOpc));
|
2021-06-13 14:55:34 +02:00
|
|
|
MIB.addMBB(getWhileLoopStartTargetBB(*MI)); // branch target
|
|
|
|
MIB.addImm(ARMCC::EQ); // condition code
|
2020-12-07 16:44:40 +01:00
|
|
|
MIB.addReg(ARM::CPSR);
|
|
|
|
|
|
|
|
MI->eraseFromParent();
|
|
|
|
}
|
|
|
|
|
|
|
|
inline void RevertDoLoopStart(MachineInstr *MI, const TargetInstrInfo *TII) {
|
|
|
|
MachineBasicBlock *MBB = MI->getParent();
|
|
|
|
BuildMI(*MBB, MI, MI->getDebugLoc(), TII->get(ARM::tMOVr))
|
|
|
|
.add(MI->getOperand(0))
|
|
|
|
.add(MI->getOperand(1))
|
|
|
|
.add(predOps(ARMCC::AL));
|
|
|
|
|
|
|
|
MI->eraseFromParent();
|
|
|
|
}
|
|
|
|
|
|
|
|
inline void RevertLoopDec(MachineInstr *MI, const TargetInstrInfo *TII,
|
|
|
|
bool SetFlags = false) {
|
|
|
|
MachineBasicBlock *MBB = MI->getParent();
|
|
|
|
|
|
|
|
MachineInstrBuilder MIB =
|
|
|
|
BuildMI(*MBB, MI, MI->getDebugLoc(), TII->get(ARM::t2SUBri));
|
|
|
|
MIB.add(MI->getOperand(0));
|
|
|
|
MIB.add(MI->getOperand(1));
|
|
|
|
MIB.add(MI->getOperand(2));
|
|
|
|
MIB.addImm(ARMCC::AL);
|
|
|
|
MIB.addReg(0);
|
|
|
|
|
|
|
|
if (SetFlags) {
|
|
|
|
MIB.addReg(ARM::CPSR);
|
|
|
|
MIB->getOperand(5).setIsDef(true);
|
|
|
|
} else
|
|
|
|
MIB.addReg(0);
|
|
|
|
|
|
|
|
MI->eraseFromParent();
|
|
|
|
}
|
|
|
|
|
|
|
|
// Generate a subs, or sub and cmp, and a branch instead of an LE.
|
|
|
|
inline void RevertLoopEnd(MachineInstr *MI, const TargetInstrInfo *TII,
|
|
|
|
unsigned BrOpc = ARM::t2Bcc, bool SkipCmp = false) {
|
|
|
|
MachineBasicBlock *MBB = MI->getParent();
|
|
|
|
|
|
|
|
// Create cmp
|
|
|
|
if (!SkipCmp) {
|
|
|
|
MachineInstrBuilder MIB =
|
|
|
|
BuildMI(*MBB, MI, MI->getDebugLoc(), TII->get(ARM::t2CMPri));
|
|
|
|
MIB.add(MI->getOperand(0));
|
|
|
|
MIB.addImm(0);
|
|
|
|
MIB.addImm(ARMCC::AL);
|
|
|
|
MIB.addReg(ARM::NoRegister);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create bne
|
|
|
|
MachineInstrBuilder MIB =
|
|
|
|
BuildMI(*MBB, MI, MI->getDebugLoc(), TII->get(BrOpc));
|
|
|
|
MIB.add(MI->getOperand(1)); // branch target
|
|
|
|
MIB.addImm(ARMCC::NE); // condition code
|
|
|
|
MIB.addReg(ARM::CPSR);
|
|
|
|
MI->eraseFromParent();
|
|
|
|
}
|
|
|
|
|
|
|
|
} // end namespace llvm
|
|
|
|
|
|
|
|
#endif // LLVM_LIB_TARGET_ARM_MVETAILPREDUTILS_H
|