mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-22 10:42:39 +01:00
[ARM][MVE][Intrinsics] Add MVE VMUL intrinsics. Remove annoying "t1" from VMUL* instructions. Add unit tests.
Summary: Add MVE VMUL intrinsics. Remove annoying "t1" from VMUL* instructions. Add unit tests. Reviewers: simon_tatham, ostannard, dmgreen Subscribers: kristof.beyls, hiraditya, cfe-commits, llvm-commits Tags: #clang, #llvm Differential Revision: https://reviews.llvm.org/D70546
This commit is contained in:
parent
99e7f8127a
commit
8d87cd0246
@ -805,6 +805,9 @@ def int_arm_mve_add_predicated: Intrinsic<[llvm_anyvector_ty],
|
|||||||
def int_arm_mve_sub_predicated: Intrinsic<[llvm_anyvector_ty],
|
def int_arm_mve_sub_predicated: Intrinsic<[llvm_anyvector_ty],
|
||||||
[LLVMMatchType<0>, LLVMMatchType<0>, llvm_anyvector_ty, LLVMMatchType<0>],
|
[LLVMMatchType<0>, LLVMMatchType<0>, llvm_anyvector_ty, LLVMMatchType<0>],
|
||||||
[IntrNoMem]>;
|
[IntrNoMem]>;
|
||||||
|
def int_arm_mve_mul_predicated: Intrinsic<[llvm_anyvector_ty],
|
||||||
|
[LLVMMatchType<0>, LLVMMatchType<0>, llvm_anyvector_ty, LLVMMatchType<0>],
|
||||||
|
[IntrNoMem]>;
|
||||||
|
|
||||||
defm int_arm_mve_minv: IntrinsicSignSuffix<[llvm_i32_ty],
|
defm int_arm_mve_minv: IntrinsicSignSuffix<[llvm_i32_ty],
|
||||||
[llvm_i32_ty, llvm_anyvector_ty], [IntrNoMem]>;
|
[llvm_i32_ty, llvm_anyvector_ty], [IntrNoMem]>;
|
||||||
|
@ -1512,8 +1512,9 @@ class MVE_int<string iname, string suffix, bits<2> size, list<dag> pattern=[]>
|
|||||||
let Inst{3-1} = Qm{2-0};
|
let Inst{3-1} = Qm{2-0};
|
||||||
}
|
}
|
||||||
|
|
||||||
class MVE_VMULt1<string suffix, bits<2> size, list<dag> pattern=[]>
|
class MVE_VMULt1<string iname, string suffix, bits<2> size,
|
||||||
: MVE_int<"vmul", suffix, size, pattern> {
|
list<dag> pattern=[]>
|
||||||
|
: MVE_int<iname, suffix, size, pattern> {
|
||||||
|
|
||||||
let Inst{28} = 0b0;
|
let Inst{28} = 0b0;
|
||||||
let Inst{25-23} = 0b110;
|
let Inst{25-23} = 0b110;
|
||||||
@ -1524,19 +1525,33 @@ class MVE_VMULt1<string suffix, bits<2> size, list<dag> pattern=[]>
|
|||||||
let validForTailPredication = 1;
|
let validForTailPredication = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
def MVE_VMULt1i8 : MVE_VMULt1<"i8", 0b00>;
|
multiclass MVE_VMUL_m<string iname, MVEVectorVTInfo VTI,
|
||||||
def MVE_VMULt1i16 : MVE_VMULt1<"i16", 0b01>;
|
SDNode unpred_op, Intrinsic pred_int> {
|
||||||
def MVE_VMULt1i32 : MVE_VMULt1<"i32", 0b10>;
|
def "" : MVE_VMULt1<iname, VTI.Suffix, VTI.Size>;
|
||||||
|
|
||||||
let Predicates = [HasMVEInt] in {
|
let Predicates = [HasMVEInt] in {
|
||||||
def : Pat<(v16i8 (mul (v16i8 MQPR:$val1), (v16i8 MQPR:$val2))),
|
// Unpredicated multiply
|
||||||
(v16i8 (MVE_VMULt1i8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
|
def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn))),
|
||||||
def : Pat<(v8i16 (mul (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))),
|
(VTI.Vec (!cast<Instruction>(NAME)
|
||||||
(v8i16 (MVE_VMULt1i16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
|
(VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
|
||||||
def : Pat<(v4i32 (mul (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))),
|
|
||||||
(v4i32 (MVE_VMULt1i32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)))>;
|
// Predicated multiply
|
||||||
|
def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
|
||||||
|
(VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive))),
|
||||||
|
(VTI.Vec (!cast<Instruction>(NAME)
|
||||||
|
(VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
|
||||||
|
(i32 1), (VTI.Pred VCCR:$mask),
|
||||||
|
(VTI.Vec MQPR:$inactive)))>;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
multiclass MVE_VMUL<MVEVectorVTInfo VTI>
|
||||||
|
: MVE_VMUL_m<"vmul", VTI, mul, int_arm_mve_mul_predicated>;
|
||||||
|
|
||||||
|
defm MVE_VMULi8 : MVE_VMUL<MVE_v16i8>;
|
||||||
|
defm MVE_VMULi16 : MVE_VMUL<MVE_v8i16>;
|
||||||
|
defm MVE_VMULi32 : MVE_VMUL<MVE_v4i32>;
|
||||||
|
|
||||||
class MVE_VQxDMULH<string iname, string suffix, bits<2> size, bit rounding,
|
class MVE_VQxDMULH<string iname, string suffix, bits<2> size, bit rounding,
|
||||||
list<dag> pattern=[]>
|
list<dag> pattern=[]>
|
||||||
: MVE_int<iname, suffix, size, pattern> {
|
: MVE_int<iname, suffix, size, pattern> {
|
||||||
@ -2805,8 +2820,8 @@ class MVEFloatArithNeon<string iname, string suffix, bit size,
|
|||||||
let Inst{16} = 0b0;
|
let Inst{16} = 0b0;
|
||||||
}
|
}
|
||||||
|
|
||||||
class MVE_VMUL_fp<string suffix, bit size, list<dag> pattern=[]>
|
class MVE_VMUL_fp<string iname, string suffix, bit size, list<dag> pattern=[]>
|
||||||
: MVEFloatArithNeon<"vmul", suffix, size, (outs MQPR:$Qd),
|
: MVEFloatArithNeon<iname, suffix, size, (outs MQPR:$Qd),
|
||||||
(ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm", vpred_r, "",
|
(ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm", vpred_r, "",
|
||||||
pattern> {
|
pattern> {
|
||||||
bits<4> Qd;
|
bits<4> Qd;
|
||||||
@ -2824,16 +2839,29 @@ class MVE_VMUL_fp<string suffix, bit size, list<dag> pattern=[]>
|
|||||||
let validForTailPredication = 1;
|
let validForTailPredication = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
def MVE_VMULf32 : MVE_VMUL_fp<"f32", 0b0>;
|
multiclass MVE_VMULT_fp_m<string iname, bit bit_21, MVEVectorVTInfo VTI,
|
||||||
def MVE_VMULf16 : MVE_VMUL_fp<"f16", 0b1>;
|
SDNode unpred_op, Intrinsic pred_int> {
|
||||||
|
def "" : MVE_VMUL_fp<iname, VTI.Suffix, VTI.Size{0}>;
|
||||||
|
|
||||||
let Predicates = [HasMVEFloat] in {
|
let Predicates = [HasMVEFloat] in {
|
||||||
def : Pat<(v4f32 (fmul (v4f32 MQPR:$val1), (v4f32 MQPR:$val2))),
|
def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn))),
|
||||||
(v4f32 (MVE_VMULf32 (v4f32 MQPR:$val1), (v4f32 MQPR:$val2)))>;
|
(VTI.Vec (!cast<Instruction>(NAME)
|
||||||
def : Pat<(v8f16 (fmul (v8f16 MQPR:$val1), (v8f16 MQPR:$val2))),
|
(VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
|
||||||
(v8f16 (MVE_VMULf16 (v8f16 MQPR:$val1), (v8f16 MQPR:$val2)))>;
|
def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
|
||||||
|
(VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive))),
|
||||||
|
(VTI.Vec (!cast<Instruction>(NAME)
|
||||||
|
(VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
|
||||||
|
(i32 1), (VTI.Pred VCCR:$mask),
|
||||||
|
(VTI.Vec MQPR:$inactive)))>;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
multiclass MVE_VMUL_fp_m<MVEVectorVTInfo VTI>
|
||||||
|
: MVE_VMULT_fp_m<"vmul", 0, VTI, fmul, int_arm_mve_mul_predicated>;
|
||||||
|
|
||||||
|
defm MVE_VMULf32 : MVE_VMUL_fp_m<MVE_v4f32>;
|
||||||
|
defm MVE_VMULf16 : MVE_VMUL_fp_m<MVE_v8f16>;
|
||||||
|
|
||||||
class MVE_VCMLA<string suffix, bit size, list<dag> pattern=[]>
|
class MVE_VCMLA<string suffix, bit size, list<dag> pattern=[]>
|
||||||
: MVEFloatArithNeon<"vcmla", suffix, size, (outs MQPR:$Qd),
|
: MVEFloatArithNeon<"vcmla", suffix, size, (outs MQPR:$Qd),
|
||||||
(ins MQPR:$Qd_src, MQPR:$Qn, MQPR:$Qm, complexrotateop:$rot),
|
(ins MQPR:$Qd_src, MQPR:$Qn, MQPR:$Qm, complexrotateop:$rot),
|
||||||
|
@ -211,7 +211,7 @@ body: |
|
|||||||
; CHECK: renamable $r4 = t2ADDrr renamable $r0, renamable $r12, 14, $noreg, $noreg
|
; CHECK: renamable $r4 = t2ADDrr renamable $r0, renamable $r12, 14, $noreg, $noreg
|
||||||
; CHECK: renamable $r12 = t2ADDri killed renamable $r12, 16, 14, $noreg, $noreg
|
; CHECK: renamable $r12 = t2ADDri killed renamable $r12, 16, 14, $noreg, $noreg
|
||||||
; CHECK: renamable $r3, dead $cpsr = tSUBi8 killed renamable $r3, 16, 14, $noreg
|
; CHECK: renamable $r3, dead $cpsr = tSUBi8 killed renamable $r3, 16, 14, $noreg
|
||||||
; CHECK: renamable $q0 = MVE_VMULt1i8 killed renamable $q1, killed renamable $q0, 0, $noreg, undef renamable $q0
|
; CHECK: renamable $q0 = MVE_VMULi8 killed renamable $q1, killed renamable $q0, 0, $noreg, undef renamable $q0
|
||||||
; CHECK: MVE_VSTRBU8 killed renamable $q0, killed renamable $r4, 0, 0, killed $noreg :: (store 16 into %ir.scevgep1, align 1)
|
; CHECK: MVE_VSTRBU8 killed renamable $q0, killed renamable $r4, 0, 0, killed $noreg :: (store 16 into %ir.scevgep1, align 1)
|
||||||
; CHECK: $lr = MVE_LETP renamable $lr, %bb.2
|
; CHECK: $lr = MVE_LETP renamable $lr, %bb.2
|
||||||
; CHECK: bb.3.for.cond.cleanup:
|
; CHECK: bb.3.for.cond.cleanup:
|
||||||
@ -252,7 +252,7 @@ body: |
|
|||||||
renamable $r4 = t2ADDrr renamable $r0, renamable $r12, 14, $noreg, $noreg
|
renamable $r4 = t2ADDrr renamable $r0, renamable $r12, 14, $noreg, $noreg
|
||||||
renamable $r12 = t2ADDri killed renamable $r12, 16, 14, $noreg, $noreg
|
renamable $r12 = t2ADDri killed renamable $r12, 16, 14, $noreg, $noreg
|
||||||
renamable $r3, dead $cpsr = tSUBi8 killed renamable $r3, 16, 14, $noreg
|
renamable $r3, dead $cpsr = tSUBi8 killed renamable $r3, 16, 14, $noreg
|
||||||
renamable $q0 = MVE_VMULt1i8 killed renamable $q1, killed renamable $q0, 0, $noreg, undef renamable $q0
|
renamable $q0 = MVE_VMULi8 killed renamable $q1, killed renamable $q0, 0, $noreg, undef renamable $q0
|
||||||
MVE_VPST 8, implicit $vpr
|
MVE_VPST 8, implicit $vpr
|
||||||
MVE_VSTRBU8 killed renamable $q0, killed renamable $r4, 0, 1, killed renamable $vpr :: (store 16 into %ir.scevgep1, align 1)
|
MVE_VSTRBU8 killed renamable $q0, killed renamable $r4, 0, 1, killed renamable $vpr :: (store 16 into %ir.scevgep1, align 1)
|
||||||
renamable $lr = t2LoopDec killed renamable $lr, 1
|
renamable $lr = t2LoopDec killed renamable $lr, 1
|
||||||
@ -325,7 +325,7 @@ body: |
|
|||||||
; CHECK: liveins: $lr, $r0, $r1, $r2, $r3
|
; CHECK: liveins: $lr, $r0, $r1, $r2, $r3
|
||||||
; CHECK: renamable $q0 = MVE_VLDRHU16 renamable $r1, 0, 0, $noreg :: (load 16 from %ir.lsr.iv57, align 2)
|
; CHECK: renamable $q0 = MVE_VLDRHU16 renamable $r1, 0, 0, $noreg :: (load 16 from %ir.lsr.iv57, align 2)
|
||||||
; CHECK: renamable $q1 = MVE_VLDRHU16 renamable $r2, 0, 0, $noreg :: (load 16 from %ir.lsr.iv24, align 2)
|
; CHECK: renamable $q1 = MVE_VLDRHU16 renamable $r2, 0, 0, $noreg :: (load 16 from %ir.lsr.iv24, align 2)
|
||||||
; CHECK: renamable $q0 = MVE_VMULt1i16 killed renamable $q1, killed renamable $q0, 0, $noreg, undef renamable $q0
|
; CHECK: renamable $q0 = MVE_VMULi16 killed renamable $q1, killed renamable $q0, 0, $noreg, undef renamable $q0
|
||||||
; CHECK: MVE_VSTRHU16 killed renamable $q0, renamable $r0, 0, 0, killed $noreg :: (store 16 into %ir.lsr.iv1, align 2)
|
; CHECK: MVE_VSTRHU16 killed renamable $q0, renamable $r0, 0, 0, killed $noreg :: (store 16 into %ir.lsr.iv1, align 2)
|
||||||
; CHECK: renamable $r1, dead $cpsr = tADDi8 killed renamable $r1, 16, 14, $noreg
|
; CHECK: renamable $r1, dead $cpsr = tADDi8 killed renamable $r1, 16, 14, $noreg
|
||||||
; CHECK: renamable $r2, dead $cpsr = tADDi8 killed renamable $r2, 16, 14, $noreg
|
; CHECK: renamable $r2, dead $cpsr = tADDi8 killed renamable $r2, 16, 14, $noreg
|
||||||
@ -358,7 +358,7 @@ body: |
|
|||||||
MVE_VPST 4, implicit $vpr
|
MVE_VPST 4, implicit $vpr
|
||||||
renamable $q0 = MVE_VLDRHU16 renamable $r1, 0, 1, renamable $vpr :: (load 16 from %ir.lsr.iv57, align 2)
|
renamable $q0 = MVE_VLDRHU16 renamable $r1, 0, 1, renamable $vpr :: (load 16 from %ir.lsr.iv57, align 2)
|
||||||
renamable $q1 = MVE_VLDRHU16 renamable $r2, 0, 1, renamable $vpr :: (load 16 from %ir.lsr.iv24, align 2)
|
renamable $q1 = MVE_VLDRHU16 renamable $r2, 0, 1, renamable $vpr :: (load 16 from %ir.lsr.iv24, align 2)
|
||||||
renamable $q0 = MVE_VMULt1i16 killed renamable $q1, killed renamable $q0, 0, $noreg, undef renamable $q0
|
renamable $q0 = MVE_VMULi16 killed renamable $q1, killed renamable $q0, 0, $noreg, undef renamable $q0
|
||||||
MVE_VPST 8, implicit $vpr
|
MVE_VPST 8, implicit $vpr
|
||||||
MVE_VSTRHU16 killed renamable $q0, renamable $r0, 0, 1, killed renamable $vpr :: (store 16 into %ir.lsr.iv1, align 2)
|
MVE_VSTRHU16 killed renamable $q0, renamable $r0, 0, 1, killed renamable $vpr :: (store 16 into %ir.lsr.iv1, align 2)
|
||||||
renamable $r1, dead $cpsr = tADDi8 killed renamable $r1, 16, 14, $noreg
|
renamable $r1, dead $cpsr = tADDi8 killed renamable $r1, 16, 14, $noreg
|
||||||
@ -441,7 +441,7 @@ body: |
|
|||||||
; CHECK: renamable $q1 = MVE_VLDRWU32 renamable $r0, 0, 0, $noreg :: (load 16 from %ir.lsr.iv24, align 4)
|
; CHECK: renamable $q1 = MVE_VLDRWU32 renamable $r0, 0, 0, $noreg :: (load 16 from %ir.lsr.iv24, align 4)
|
||||||
; CHECK: renamable $q2 = MVE_VLDRWU32 renamable $r1, 0, 0, killed $noreg :: (load 16 from %ir.lsr.iv1, align 4)
|
; CHECK: renamable $q2 = MVE_VLDRWU32 renamable $r1, 0, 0, killed $noreg :: (load 16 from %ir.lsr.iv1, align 4)
|
||||||
; CHECK: $r3 = tMOVr $r2, 14, $noreg
|
; CHECK: $r3 = tMOVr $r2, 14, $noreg
|
||||||
; CHECK: renamable $q1 = nsw MVE_VMULt1i32 killed renamable $q2, killed renamable $q1, 0, $noreg, undef renamable $q1
|
; CHECK: renamable $q1 = nsw MVE_VMULi32 killed renamable $q2, killed renamable $q1, 0, $noreg, undef renamable $q1
|
||||||
; CHECK: renamable $r0, dead $cpsr = tADDi8 killed renamable $r0, 16, 14, $noreg
|
; CHECK: renamable $r0, dead $cpsr = tADDi8 killed renamable $r0, 16, 14, $noreg
|
||||||
; CHECK: renamable $r1, dead $cpsr = tADDi8 killed renamable $r1, 16, 14, $noreg
|
; CHECK: renamable $r1, dead $cpsr = tADDi8 killed renamable $r1, 16, 14, $noreg
|
||||||
; CHECK: renamable $r2, dead $cpsr = tSUBi8 killed $r2, 4, 14, $noreg
|
; CHECK: renamable $r2, dead $cpsr = tSUBi8 killed $r2, 4, 14, $noreg
|
||||||
@ -490,7 +490,7 @@ body: |
|
|||||||
renamable $q1 = MVE_VLDRWU32 renamable $r0, 0, 1, renamable $vpr :: (load 16 from %ir.lsr.iv24, align 4)
|
renamable $q1 = MVE_VLDRWU32 renamable $r0, 0, 1, renamable $vpr :: (load 16 from %ir.lsr.iv24, align 4)
|
||||||
renamable $q2 = MVE_VLDRWU32 renamable $r1, 0, 1, killed renamable $vpr :: (load 16 from %ir.lsr.iv1, align 4)
|
renamable $q2 = MVE_VLDRWU32 renamable $r1, 0, 1, killed renamable $vpr :: (load 16 from %ir.lsr.iv1, align 4)
|
||||||
$r3 = tMOVr $r2, 14, $noreg
|
$r3 = tMOVr $r2, 14, $noreg
|
||||||
renamable $q1 = nsw MVE_VMULt1i32 killed renamable $q2, killed renamable $q1, 0, $noreg, undef renamable $q1
|
renamable $q1 = nsw MVE_VMULi32 killed renamable $q2, killed renamable $q1, 0, $noreg, undef renamable $q1
|
||||||
renamable $r0, dead $cpsr = tADDi8 killed renamable $r0, 16, 14, $noreg
|
renamable $r0, dead $cpsr = tADDi8 killed renamable $r0, 16, 14, $noreg
|
||||||
renamable $r1, dead $cpsr = tADDi8 killed renamable $r1, 16, 14, $noreg
|
renamable $r1, dead $cpsr = tADDi8 killed renamable $r1, 16, 14, $noreg
|
||||||
renamable $r2, dead $cpsr = tSUBi8 killed $r2, 4, 14, $noreg
|
renamable $r2, dead $cpsr = tSUBi8 killed $r2, 4, 14, $noreg
|
||||||
|
58
test/CodeGen/Thumb2/mve-intrinsics/vmulq.ll
Normal file
58
test/CodeGen/Thumb2/mve-intrinsics/vmulq.ll
Normal file
@ -0,0 +1,58 @@
|
|||||||
|
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
||||||
|
; RUN: llc -mtriple=thumbv8.1m.main -mattr=+mve.fp -verify-machineinstrs -o - %s | FileCheck %s
|
||||||
|
|
||||||
|
define arm_aapcs_vfpcc <4 x i32> @test_vmulq_u32(<4 x i32> %a, <4 x i32> %b) {
|
||||||
|
; CHECK-LABEL: test_vmulq_u32:
|
||||||
|
; CHECK: @ %bb.0: @ %entry
|
||||||
|
; CHECK-NEXT: vmul.i32 q0, q1, q0
|
||||||
|
; CHECK-NEXT: bx lr
|
||||||
|
entry:
|
||||||
|
%0 = mul <4 x i32> %b, %a
|
||||||
|
ret <4 x i32> %0
|
||||||
|
}
|
||||||
|
|
||||||
|
define arm_aapcs_vfpcc <4 x float> @test_vmulq_f32(<4 x float> %a, <4 x float> %b) {
|
||||||
|
; CHECK-LABEL: test_vmulq_f32:
|
||||||
|
; CHECK: @ %bb.0: @ %entry
|
||||||
|
; CHECK-NEXT: vmul.f32 q0, q1, q0
|
||||||
|
; CHECK-NEXT: bx lr
|
||||||
|
entry:
|
||||||
|
%0 = fmul <4 x float> %b, %a
|
||||||
|
ret <4 x float> %0
|
||||||
|
}
|
||||||
|
|
||||||
|
define arm_aapcs_vfpcc <16 x i8> @test_vmulq_m_s8(<16 x i8> %inactive, <16 x i8> %a, <16 x i8> %b, i16 zeroext %p) {
|
||||||
|
; CHECK-LABEL: test_vmulq_m_s8:
|
||||||
|
; CHECK: @ %bb.0: @ %entry
|
||||||
|
; CHECK-NEXT: vmsr p0, r0
|
||||||
|
; CHECK-NEXT: vpst
|
||||||
|
; CHECK-NEXT: vmult.i8 q0, q1, q2
|
||||||
|
; CHECK-NEXT: bx lr
|
||||||
|
entry:
|
||||||
|
%0 = zext i16 %p to i32
|
||||||
|
%1 = tail call <16 x i1> @llvm.arm.mve.pred.i2v.v16i1(i32 %0)
|
||||||
|
%2 = tail call <16 x i8> @llvm.arm.mve.mul.predicated.v16i8.v16i1(<16 x i8> %a, <16 x i8> %b, <16 x i1> %1, <16 x i8> %inactive)
|
||||||
|
ret <16 x i8> %2
|
||||||
|
}
|
||||||
|
|
||||||
|
declare <16 x i1> @llvm.arm.mve.pred.i2v.v16i1(i32)
|
||||||
|
|
||||||
|
declare <16 x i8> @llvm.arm.mve.mul.predicated.v16i8.v16i1(<16 x i8>, <16 x i8>, <16 x i1>, <16 x i8>)
|
||||||
|
|
||||||
|
define arm_aapcs_vfpcc <8 x half> @test_vmulq_m_f16(<8 x half> %inactive, <8 x half> %a, <8 x half> %b, i16 zeroext %p) {
|
||||||
|
; CHECK-LABEL: test_vmulq_m_f16:
|
||||||
|
; CHECK: @ %bb.0: @ %entry
|
||||||
|
; CHECK-NEXT: vmsr p0, r0
|
||||||
|
; CHECK-NEXT: vpst
|
||||||
|
; CHECK-NEXT: vmult.f16 q0, q1, q2
|
||||||
|
; CHECK-NEXT: bx lr
|
||||||
|
entry:
|
||||||
|
%0 = zext i16 %p to i32
|
||||||
|
%1 = tail call <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32 %0)
|
||||||
|
%2 = tail call <8 x half> @llvm.arm.mve.mul.predicated.v8f16.v8i1(<8 x half> %a, <8 x half> %b, <8 x i1> %1, <8 x half> %inactive)
|
||||||
|
ret <8 x half> %2
|
||||||
|
}
|
||||||
|
|
||||||
|
declare <8 x i1> @llvm.arm.mve.pred.i2v.v8i1(i32)
|
||||||
|
|
||||||
|
declare <8 x half> @llvm.arm.mve.mul.predicated.v8f16.v8i1(<8 x half>, <8 x half>, <8 x i1>, <8 x half>)
|
@ -250,9 +250,9 @@ TEST(MachineInstrValidTailPredication, IsCorrect) {
|
|||||||
case MVE_VMUL_qr_i8:
|
case MVE_VMUL_qr_i8:
|
||||||
case MVE_VMULf16:
|
case MVE_VMULf16:
|
||||||
case MVE_VMULf32:
|
case MVE_VMULf32:
|
||||||
case MVE_VMULt1i16:
|
case MVE_VMULi16:
|
||||||
case MVE_VMULt1i8:
|
case MVE_VMULi8:
|
||||||
case MVE_VMULt1i32:
|
case MVE_VMULi32:
|
||||||
case MVE_VMVN:
|
case MVE_VMVN:
|
||||||
case MVE_VMVNimmi16:
|
case MVE_VMVNimmi16:
|
||||||
case MVE_VMVNimmi32:
|
case MVE_VMVNimmi32:
|
||||||
|
Loading…
Reference in New Issue
Block a user