mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-24 03:33:20 +01:00
7e501cf4c3
This update was done with the following bash script: find test/CodeGen -name "*.ll" | \ while read NAME; do echo "$NAME" if ! grep -q "^; *RUN: *llc.*debug" $NAME; then TEMP=`mktemp -t temp` cp $NAME $TEMP sed -n "s/^define [^@]*@\([A-Za-z0-9_]*\)(.*$/\1/p" < $NAME | \ while read FUNC; do sed -i '' "s/;\(.*\)\([A-Za-z0-9_-]*\):\( *\)$FUNC: *\$/;\1\2-LABEL:\3$FUNC:/g" $TEMP done sed -i '' "s/;\(.*\)-LABEL-LABEL:/;\1-LABEL:/" $TEMP sed -i '' "s/;\(.*\)-NEXT-LABEL:/;\1-NEXT:/" $TEMP sed -i '' "s/;\(.*\)-NOT-LABEL:/;\1-NOT:/" $TEMP sed -i '' "s/;\(.*\)-DAG-LABEL:/;\1-DAG:/" $TEMP mv $TEMP $NAME fi done llvm-svn: 186280
154 lines
4.1 KiB
LLVM
154 lines
4.1 KiB
LLVM
; We currently estimate the cost of sext/zext/trunc v8(v16)i32 <-> v8(v16)i8
|
|
; instructions as expensive. If lowering is improved the cost model needs to
|
|
; change.
|
|
; RUN: opt < %s -cost-model -analyze -mtriple=thumbv7-apple-ios6.0.0 -march=arm -mcpu=cortex-a8 | FileCheck %s --check-prefix=COST
|
|
%T0_5 = type <8 x i8>
|
|
%T1_5 = type <8 x i32>
|
|
; CHECK-LABEL: func_cvt5:
|
|
define void @func_cvt5(%T0_5* %loadaddr, %T1_5* %storeaddr) {
|
|
; CHECK: vmovl.s8
|
|
; CHECK: vmovl.s16
|
|
; CHECK: vmovl.s16
|
|
%v0 = load %T0_5* %loadaddr
|
|
; COST: func_cvt5
|
|
; COST: cost of 3 {{.*}} sext
|
|
%r = sext %T0_5 %v0 to %T1_5
|
|
store %T1_5 %r, %T1_5* %storeaddr
|
|
ret void
|
|
}
|
|
;; We currently estimate the cost of this instruction as expensive. If lowering
|
|
;; is improved the cost needs to change.
|
|
%TA0_5 = type <8 x i8>
|
|
%TA1_5 = type <8 x i32>
|
|
; CHECK-LABEL: func_cvt1:
|
|
define void @func_cvt1(%TA0_5* %loadaddr, %TA1_5* %storeaddr) {
|
|
; CHECK: vmovl.u8
|
|
; CHECK: vmovl.u16
|
|
; CHECK: vmovl.u16
|
|
%v0 = load %TA0_5* %loadaddr
|
|
; COST: func_cvt1
|
|
; COST: cost of 3 {{.*}} zext
|
|
%r = zext %TA0_5 %v0 to %TA1_5
|
|
store %TA1_5 %r, %TA1_5* %storeaddr
|
|
ret void
|
|
}
|
|
|
|
%T0_51 = type <8 x i32>
|
|
%T1_51 = type <8 x i8>
|
|
; CHECK-LABEL: func_cvt51:
|
|
define void @func_cvt51(%T0_51* %loadaddr, %T1_51* %storeaddr) {
|
|
; CHECK: vmovn.i32
|
|
; CHECK: vmovn.i32
|
|
; CHECK: vmovn.i16
|
|
%v0 = load %T0_51* %loadaddr
|
|
; COST: func_cvt51
|
|
; COST: cost of 3 {{.*}} trunc
|
|
%r = trunc %T0_51 %v0 to %T1_51
|
|
store %T1_51 %r, %T1_51* %storeaddr
|
|
ret void
|
|
}
|
|
|
|
%TT0_5 = type <16 x i8>
|
|
%TT1_5 = type <16 x i32>
|
|
; CHECK-LABEL: func_cvt52:
|
|
define void @func_cvt52(%TT0_5* %loadaddr, %TT1_5* %storeaddr) {
|
|
; CHECK: vmovl.s16
|
|
; CHECK: vmovl.s16
|
|
; CHECK: vmovl.s16
|
|
; CHECK: vmovl.s16
|
|
%v0 = load %TT0_5* %loadaddr
|
|
; COST: func_cvt52
|
|
; COST: cost of 6 {{.*}} sext
|
|
%r = sext %TT0_5 %v0 to %TT1_5
|
|
store %TT1_5 %r, %TT1_5* %storeaddr
|
|
ret void
|
|
}
|
|
;; We currently estimate the cost of this instruction as expensive. If lowering
|
|
;; is improved the cost needs to change.
|
|
%TTA0_5 = type <16 x i8>
|
|
%TTA1_5 = type <16 x i32>
|
|
; CHECK-LABEL: func_cvt12:
|
|
define void @func_cvt12(%TTA0_5* %loadaddr, %TTA1_5* %storeaddr) {
|
|
; CHECK: vmovl.u16
|
|
; CHECK: vmovl.u16
|
|
; CHECK: vmovl.u16
|
|
; CHECK: vmovl.u16
|
|
%v0 = load %TTA0_5* %loadaddr
|
|
; COST: func_cvt12
|
|
; COST: cost of 6 {{.*}} zext
|
|
%r = zext %TTA0_5 %v0 to %TTA1_5
|
|
store %TTA1_5 %r, %TTA1_5* %storeaddr
|
|
ret void
|
|
}
|
|
|
|
%TT0_51 = type <16 x i32>
|
|
%TT1_51 = type <16 x i8>
|
|
; CHECK-LABEL: func_cvt512:
|
|
define void @func_cvt512(%TT0_51* %loadaddr, %TT1_51* %storeaddr) {
|
|
; CHECK: vmovn.i32
|
|
; CHECK: vmovn.i32
|
|
; CHECK: vmovn.i32
|
|
; CHECK: vmovn.i32
|
|
; CHECK: vmovn.i16
|
|
; CHECK: vmovn.i16
|
|
%v0 = load %TT0_51* %loadaddr
|
|
; COST: func_cvt512
|
|
; COST: cost of 6 {{.*}} trunc
|
|
%r = trunc %TT0_51 %v0 to %TT1_51
|
|
store %TT1_51 %r, %TT1_51* %storeaddr
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: sext_v4i16_v4i64:
|
|
define void @sext_v4i16_v4i64(<4 x i16>* %loadaddr, <4 x i64>* %storeaddr) {
|
|
; CHECK: vmovl.s32
|
|
; CHECK: vmovl.s32
|
|
%v0 = load <4 x i16>* %loadaddr
|
|
; COST: sext_v4i16_v4i64
|
|
; COST: cost of 3 {{.*}} sext
|
|
%r = sext <4 x i16> %v0 to <4 x i64>
|
|
store <4 x i64> %r, <4 x i64>* %storeaddr
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: zext_v4i16_v4i64:
|
|
define void @zext_v4i16_v4i64(<4 x i16>* %loadaddr, <4 x i64>* %storeaddr) {
|
|
; CHECK: vmovl.u32
|
|
; CHECK: vmovl.u32
|
|
%v0 = load <4 x i16>* %loadaddr
|
|
; COST: zext_v4i16_v4i64
|
|
; COST: cost of 3 {{.*}} zext
|
|
%r = zext <4 x i16> %v0 to <4 x i64>
|
|
store <4 x i64> %r, <4 x i64>* %storeaddr
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: sext_v8i16_v8i64:
|
|
define void @sext_v8i16_v8i64(<8 x i16>* %loadaddr, <8 x i64>* %storeaddr) {
|
|
; CHECK: vmovl.s32
|
|
; CHECK: vmovl.s32
|
|
; CHECK: vmovl.s32
|
|
; CHECK: vmovl.s32
|
|
%v0 = load <8 x i16>* %loadaddr
|
|
; COST: sext_v8i16_v8i64
|
|
; COST: cost of 6 {{.*}} sext
|
|
%r = sext <8 x i16> %v0 to <8 x i64>
|
|
store <8 x i64> %r, <8 x i64>* %storeaddr
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: zext_v8i16_v8i64:
|
|
define void @zext_v8i16_v8i64(<8 x i16>* %loadaddr, <8 x i64>* %storeaddr) {
|
|
; CHECK: vmovl.u32
|
|
; CHECK: vmovl.u32
|
|
; CHECK: vmovl.u32
|
|
; CHECK: vmovl.u32
|
|
%v0 = load <8 x i16>* %loadaddr
|
|
; COST: zext_v8i16_v8i64
|
|
; COST: cost of 6 {{.*}} zext
|
|
%r = zext <8 x i16> %v0 to <8 x i64>
|
|
store <8 x i64> %r, <8 x i64>* %storeaddr
|
|
ret void
|
|
}
|
|
|