mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-26 04:32:44 +01:00
bbe3f8278e
Summary: When promoting the over flow vector for these ops we should use the target's desired setcc result type. This way a v8i32 result type will use a v8i32 overflow vector instead of a v8i16 overflow vector. A v8i16 overflow vector will cause LegalizeDAG/LegalizeVectorOps to have to use v8i32 and truncate to v8i16 in its expansion. By doing this in type legalization instead, we get the truncate into the DAG earlier and give DAG combine more of a chance to optimize it. We also have to fix unrolling to use the scalar setcc result type for the scalarized operation, and convert it to the required vector element type after the scalar operation. We have to observe the vector boolean contents when doing this conversion. The previous code was just taking the scalar result and putting it in the vector. But for X86 and AArch64 that would have only put a the boolean value in bit 0 of the element and left all other bits in the element 0. We need to ensure all bits in the element are the same. I'm using a select with constants here because that's what setcc unrolling in LegalizeVectorOps used. Reviewers: spatel, RKSimon, nikic Reviewed By: nikic Subscribers: javed.absar, kristof.beyls, dmgreen, llvm-commits Tags: #llvm Differential Revision: https://reviews.llvm.org/D58567 llvm-svn: 354753
308 lines
12 KiB
LLVM
308 lines
12 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc < %s -mtriple=aarch64-none-linux-gnu -mattr=+neon | FileCheck %s --check-prefix=CHECK
|
|
|
|
declare {<1 x i32>, <1 x i1>} @llvm.uadd.with.overflow.v1i32(<1 x i32>, <1 x i32>)
|
|
declare {<2 x i32>, <2 x i1>} @llvm.uadd.with.overflow.v2i32(<2 x i32>, <2 x i32>)
|
|
declare {<3 x i32>, <3 x i1>} @llvm.uadd.with.overflow.v3i32(<3 x i32>, <3 x i32>)
|
|
declare {<4 x i32>, <4 x i1>} @llvm.uadd.with.overflow.v4i32(<4 x i32>, <4 x i32>)
|
|
declare {<6 x i32>, <6 x i1>} @llvm.uadd.with.overflow.v6i32(<6 x i32>, <6 x i32>)
|
|
declare {<8 x i32>, <8 x i1>} @llvm.uadd.with.overflow.v8i32(<8 x i32>, <8 x i32>)
|
|
|
|
declare {<16 x i8>, <16 x i1>} @llvm.uadd.with.overflow.v16i8(<16 x i8>, <16 x i8>)
|
|
declare {<8 x i16>, <8 x i1>} @llvm.uadd.with.overflow.v8i16(<8 x i16>, <8 x i16>)
|
|
declare {<2 x i64>, <2 x i1>} @llvm.uadd.with.overflow.v2i64(<2 x i64>, <2 x i64>)
|
|
|
|
declare {<4 x i24>, <4 x i1>} @llvm.uadd.with.overflow.v4i24(<4 x i24>, <4 x i24>)
|
|
declare {<4 x i1>, <4 x i1>} @llvm.uadd.with.overflow.v4i1(<4 x i1>, <4 x i1>)
|
|
declare {<2 x i128>, <2 x i1>} @llvm.uadd.with.overflow.v2i128(<2 x i128>, <2 x i128>)
|
|
|
|
define <1 x i32> @uaddo_v1i32(<1 x i32> %a0, <1 x i32> %a1, <1 x i32>* %p2) nounwind {
|
|
; CHECK-LABEL: uaddo_v1i32:
|
|
; CHECK: // %bb.0:
|
|
; CHECK-NEXT: add v1.2s, v0.2s, v1.2s
|
|
; CHECK-NEXT: cmhi v0.2s, v0.2s, v1.2s
|
|
; CHECK-NEXT: str s1, [x0]
|
|
; CHECK-NEXT: ret
|
|
%t = call {<1 x i32>, <1 x i1>} @llvm.uadd.with.overflow.v1i32(<1 x i32> %a0, <1 x i32> %a1)
|
|
%val = extractvalue {<1 x i32>, <1 x i1>} %t, 0
|
|
%obit = extractvalue {<1 x i32>, <1 x i1>} %t, 1
|
|
%res = sext <1 x i1> %obit to <1 x i32>
|
|
store <1 x i32> %val, <1 x i32>* %p2
|
|
ret <1 x i32> %res
|
|
}
|
|
|
|
define <2 x i32> @uaddo_v2i32(<2 x i32> %a0, <2 x i32> %a1, <2 x i32>* %p2) nounwind {
|
|
; CHECK-LABEL: uaddo_v2i32:
|
|
; CHECK: // %bb.0:
|
|
; CHECK-NEXT: add v1.2s, v0.2s, v1.2s
|
|
; CHECK-NEXT: cmhi v0.2s, v0.2s, v1.2s
|
|
; CHECK-NEXT: str d1, [x0]
|
|
; CHECK-NEXT: ret
|
|
%t = call {<2 x i32>, <2 x i1>} @llvm.uadd.with.overflow.v2i32(<2 x i32> %a0, <2 x i32> %a1)
|
|
%val = extractvalue {<2 x i32>, <2 x i1>} %t, 0
|
|
%obit = extractvalue {<2 x i32>, <2 x i1>} %t, 1
|
|
%res = sext <2 x i1> %obit to <2 x i32>
|
|
store <2 x i32> %val, <2 x i32>* %p2
|
|
ret <2 x i32> %res
|
|
}
|
|
|
|
define <3 x i32> @uaddo_v3i32(<3 x i32> %a0, <3 x i32> %a1, <3 x i32>* %p2) nounwind {
|
|
; CHECK-LABEL: uaddo_v3i32:
|
|
; CHECK: // %bb.0:
|
|
; CHECK-NEXT: add v1.4s, v0.4s, v1.4s
|
|
; CHECK-NEXT: add x8, x0, #8 // =8
|
|
; CHECK-NEXT: cmhi v0.4s, v0.4s, v1.4s
|
|
; CHECK-NEXT: st1 { v1.s }[2], [x8]
|
|
; CHECK-NEXT: str d1, [x0]
|
|
; CHECK-NEXT: ret
|
|
%t = call {<3 x i32>, <3 x i1>} @llvm.uadd.with.overflow.v3i32(<3 x i32> %a0, <3 x i32> %a1)
|
|
%val = extractvalue {<3 x i32>, <3 x i1>} %t, 0
|
|
%obit = extractvalue {<3 x i32>, <3 x i1>} %t, 1
|
|
%res = sext <3 x i1> %obit to <3 x i32>
|
|
store <3 x i32> %val, <3 x i32>* %p2
|
|
ret <3 x i32> %res
|
|
}
|
|
|
|
define <4 x i32> @uaddo_v4i32(<4 x i32> %a0, <4 x i32> %a1, <4 x i32>* %p2) nounwind {
|
|
; CHECK-LABEL: uaddo_v4i32:
|
|
; CHECK: // %bb.0:
|
|
; CHECK-NEXT: add v1.4s, v0.4s, v1.4s
|
|
; CHECK-NEXT: cmhi v0.4s, v0.4s, v1.4s
|
|
; CHECK-NEXT: str q1, [x0]
|
|
; CHECK-NEXT: ret
|
|
%t = call {<4 x i32>, <4 x i1>} @llvm.uadd.with.overflow.v4i32(<4 x i32> %a0, <4 x i32> %a1)
|
|
%val = extractvalue {<4 x i32>, <4 x i1>} %t, 0
|
|
%obit = extractvalue {<4 x i32>, <4 x i1>} %t, 1
|
|
%res = sext <4 x i1> %obit to <4 x i32>
|
|
store <4 x i32> %val, <4 x i32>* %p2
|
|
ret <4 x i32> %res
|
|
}
|
|
|
|
define <6 x i32> @uaddo_v6i32(<6 x i32> %a0, <6 x i32> %a1, <6 x i32>* %p2) nounwind {
|
|
; CHECK-LABEL: uaddo_v6i32:
|
|
; CHECK: // %bb.0:
|
|
; CHECK-NEXT: fmov s2, w6
|
|
; CHECK-NEXT: ldr s0, [sp, #16]
|
|
; CHECK-NEXT: mov x9, sp
|
|
; CHECK-NEXT: mov v2.s[1], w7
|
|
; CHECK-NEXT: ld1 { v2.s }[2], [x9]
|
|
; CHECK-NEXT: add x8, sp, #24 // =24
|
|
; CHECK-NEXT: add x10, sp, #8 // =8
|
|
; CHECK-NEXT: ld1 { v0.s }[1], [x8]
|
|
; CHECK-NEXT: fmov s3, w0
|
|
; CHECK-NEXT: ldr x11, [sp, #32]
|
|
; CHECK-NEXT: ld1 { v2.s }[3], [x10]
|
|
; CHECK-NEXT: fmov s1, w4
|
|
; CHECK-NEXT: mov v3.s[1], w1
|
|
; CHECK-NEXT: mov v1.s[1], w5
|
|
; CHECK-NEXT: mov v3.s[2], w2
|
|
; CHECK-NEXT: mov v3.s[3], w3
|
|
; CHECK-NEXT: add v0.4s, v1.4s, v0.4s
|
|
; CHECK-NEXT: cmhi v1.4s, v1.4s, v0.4s
|
|
; CHECK-NEXT: str d0, [x11, #16]
|
|
; CHECK-NEXT: add v0.4s, v3.4s, v2.4s
|
|
; CHECK-NEXT: cmhi v2.4s, v3.4s, v0.4s
|
|
; CHECK-NEXT: mov w5, v1.s[1]
|
|
; CHECK-NEXT: mov w1, v2.s[1]
|
|
; CHECK-NEXT: mov w2, v2.s[2]
|
|
; CHECK-NEXT: mov w3, v2.s[3]
|
|
; CHECK-NEXT: fmov w4, s1
|
|
; CHECK-NEXT: fmov w0, s2
|
|
; CHECK-NEXT: str q0, [x11]
|
|
; CHECK-NEXT: ret
|
|
%t = call {<6 x i32>, <6 x i1>} @llvm.uadd.with.overflow.v6i32(<6 x i32> %a0, <6 x i32> %a1)
|
|
%val = extractvalue {<6 x i32>, <6 x i1>} %t, 0
|
|
%obit = extractvalue {<6 x i32>, <6 x i1>} %t, 1
|
|
%res = sext <6 x i1> %obit to <6 x i32>
|
|
store <6 x i32> %val, <6 x i32>* %p2
|
|
ret <6 x i32> %res
|
|
}
|
|
|
|
define <8 x i32> @uaddo_v8i32(<8 x i32> %a0, <8 x i32> %a1, <8 x i32>* %p2) nounwind {
|
|
; CHECK-LABEL: uaddo_v8i32:
|
|
; CHECK: // %bb.0:
|
|
; CHECK-NEXT: add v2.4s, v0.4s, v2.4s
|
|
; CHECK-NEXT: add v3.4s, v1.4s, v3.4s
|
|
; CHECK-NEXT: cmhi v0.4s, v0.4s, v2.4s
|
|
; CHECK-NEXT: cmhi v1.4s, v1.4s, v3.4s
|
|
; CHECK-NEXT: stp q2, q3, [x0]
|
|
; CHECK-NEXT: ret
|
|
%t = call {<8 x i32>, <8 x i1>} @llvm.uadd.with.overflow.v8i32(<8 x i32> %a0, <8 x i32> %a1)
|
|
%val = extractvalue {<8 x i32>, <8 x i1>} %t, 0
|
|
%obit = extractvalue {<8 x i32>, <8 x i1>} %t, 1
|
|
%res = sext <8 x i1> %obit to <8 x i32>
|
|
store <8 x i32> %val, <8 x i32>* %p2
|
|
ret <8 x i32> %res
|
|
}
|
|
|
|
define <16 x i32> @uaddo_v16i8(<16 x i8> %a0, <16 x i8> %a1, <16 x i8>* %p2) nounwind {
|
|
; CHECK-LABEL: uaddo_v16i8:
|
|
; CHECK: // %bb.0:
|
|
; CHECK-NEXT: add v4.16b, v0.16b, v1.16b
|
|
; CHECK-NEXT: cmhi v0.16b, v0.16b, v4.16b
|
|
; CHECK-NEXT: zip1 v1.8b, v0.8b, v0.8b
|
|
; CHECK-NEXT: zip2 v2.8b, v0.8b, v0.8b
|
|
; CHECK-NEXT: ext v0.16b, v0.16b, v0.16b, #8
|
|
; CHECK-NEXT: ushll v1.4s, v1.4h, #0
|
|
; CHECK-NEXT: ushll v2.4s, v2.4h, #0
|
|
; CHECK-NEXT: zip1 v3.8b, v0.8b, v0.8b
|
|
; CHECK-NEXT: zip2 v0.8b, v0.8b, v0.8b
|
|
; CHECK-NEXT: shl v1.4s, v1.4s, #31
|
|
; CHECK-NEXT: shl v2.4s, v2.4s, #31
|
|
; CHECK-NEXT: ushll v3.4s, v3.4h, #0
|
|
; CHECK-NEXT: ushll v5.4s, v0.4h, #0
|
|
; CHECK-NEXT: sshr v0.4s, v1.4s, #31
|
|
; CHECK-NEXT: sshr v1.4s, v2.4s, #31
|
|
; CHECK-NEXT: shl v2.4s, v3.4s, #31
|
|
; CHECK-NEXT: shl v3.4s, v5.4s, #31
|
|
; CHECK-NEXT: sshr v2.4s, v2.4s, #31
|
|
; CHECK-NEXT: sshr v3.4s, v3.4s, #31
|
|
; CHECK-NEXT: str q4, [x0]
|
|
; CHECK-NEXT: ret
|
|
%t = call {<16 x i8>, <16 x i1>} @llvm.uadd.with.overflow.v16i8(<16 x i8> %a0, <16 x i8> %a1)
|
|
%val = extractvalue {<16 x i8>, <16 x i1>} %t, 0
|
|
%obit = extractvalue {<16 x i8>, <16 x i1>} %t, 1
|
|
%res = sext <16 x i1> %obit to <16 x i32>
|
|
store <16 x i8> %val, <16 x i8>* %p2
|
|
ret <16 x i32> %res
|
|
}
|
|
|
|
define <8 x i32> @uaddo_v8i16(<8 x i16> %a0, <8 x i16> %a1, <8 x i16>* %p2) nounwind {
|
|
; CHECK-LABEL: uaddo_v8i16:
|
|
; CHECK: // %bb.0:
|
|
; CHECK-NEXT: add v2.8h, v0.8h, v1.8h
|
|
; CHECK-NEXT: cmhi v0.8h, v0.8h, v2.8h
|
|
; CHECK-NEXT: xtn v0.8b, v0.8h
|
|
; CHECK-NEXT: zip1 v1.8b, v0.8b, v0.8b
|
|
; CHECK-NEXT: zip2 v0.8b, v0.8b, v0.8b
|
|
; CHECK-NEXT: ushll v1.4s, v1.4h, #0
|
|
; CHECK-NEXT: ushll v0.4s, v0.4h, #0
|
|
; CHECK-NEXT: shl v1.4s, v1.4s, #31
|
|
; CHECK-NEXT: shl v3.4s, v0.4s, #31
|
|
; CHECK-NEXT: sshr v0.4s, v1.4s, #31
|
|
; CHECK-NEXT: sshr v1.4s, v3.4s, #31
|
|
; CHECK-NEXT: str q2, [x0]
|
|
; CHECK-NEXT: ret
|
|
%t = call {<8 x i16>, <8 x i1>} @llvm.uadd.with.overflow.v8i16(<8 x i16> %a0, <8 x i16> %a1)
|
|
%val = extractvalue {<8 x i16>, <8 x i1>} %t, 0
|
|
%obit = extractvalue {<8 x i16>, <8 x i1>} %t, 1
|
|
%res = sext <8 x i1> %obit to <8 x i32>
|
|
store <8 x i16> %val, <8 x i16>* %p2
|
|
ret <8 x i32> %res
|
|
}
|
|
|
|
define <2 x i32> @uaddo_v2i64(<2 x i64> %a0, <2 x i64> %a1, <2 x i64>* %p2) nounwind {
|
|
; CHECK-LABEL: uaddo_v2i64:
|
|
; CHECK: // %bb.0:
|
|
; CHECK-NEXT: add v1.2d, v0.2d, v1.2d
|
|
; CHECK-NEXT: cmhi v0.2d, v0.2d, v1.2d
|
|
; CHECK-NEXT: xtn v0.2s, v0.2d
|
|
; CHECK-NEXT: str q1, [x0]
|
|
; CHECK-NEXT: ret
|
|
%t = call {<2 x i64>, <2 x i1>} @llvm.uadd.with.overflow.v2i64(<2 x i64> %a0, <2 x i64> %a1)
|
|
%val = extractvalue {<2 x i64>, <2 x i1>} %t, 0
|
|
%obit = extractvalue {<2 x i64>, <2 x i1>} %t, 1
|
|
%res = sext <2 x i1> %obit to <2 x i32>
|
|
store <2 x i64> %val, <2 x i64>* %p2
|
|
ret <2 x i32> %res
|
|
}
|
|
|
|
define <4 x i32> @uaddo_v4i24(<4 x i24> %a0, <4 x i24> %a1, <4 x i24>* %p2) nounwind {
|
|
; CHECK-LABEL: uaddo_v4i24:
|
|
; CHECK: // %bb.0:
|
|
; CHECK-NEXT: bic v1.4s, #255, lsl #24
|
|
; CHECK-NEXT: bic v0.4s, #255, lsl #24
|
|
; CHECK-NEXT: add v0.4s, v0.4s, v1.4s
|
|
; CHECK-NEXT: mov v1.16b, v0.16b
|
|
; CHECK-NEXT: mov w8, v0.s[3]
|
|
; CHECK-NEXT: bic v1.4s, #255, lsl #24
|
|
; CHECK-NEXT: mov w9, v0.s[2]
|
|
; CHECK-NEXT: mov w10, v0.s[1]
|
|
; CHECK-NEXT: sturh w8, [x0, #9]
|
|
; CHECK-NEXT: lsr w8, w8, #16
|
|
; CHECK-NEXT: cmeq v1.4s, v1.4s, v0.4s
|
|
; CHECK-NEXT: fmov w11, s0
|
|
; CHECK-NEXT: strh w9, [x0, #6]
|
|
; CHECK-NEXT: sturh w10, [x0, #3]
|
|
; CHECK-NEXT: lsr w9, w9, #16
|
|
; CHECK-NEXT: lsr w10, w10, #16
|
|
; CHECK-NEXT: strb w8, [x0, #11]
|
|
; CHECK-NEXT: mvn v0.16b, v1.16b
|
|
; CHECK-NEXT: lsr w8, w11, #16
|
|
; CHECK-NEXT: strh w11, [x0]
|
|
; CHECK-NEXT: strb w9, [x0, #8]
|
|
; CHECK-NEXT: strb w10, [x0, #5]
|
|
; CHECK-NEXT: strb w8, [x0, #2]
|
|
; CHECK-NEXT: ret
|
|
%t = call {<4 x i24>, <4 x i1>} @llvm.uadd.with.overflow.v4i24(<4 x i24> %a0, <4 x i24> %a1)
|
|
%val = extractvalue {<4 x i24>, <4 x i1>} %t, 0
|
|
%obit = extractvalue {<4 x i24>, <4 x i1>} %t, 1
|
|
%res = sext <4 x i1> %obit to <4 x i32>
|
|
store <4 x i24> %val, <4 x i24>* %p2
|
|
ret <4 x i32> %res
|
|
}
|
|
|
|
define <4 x i32> @uaddo_v4i1(<4 x i1> %a0, <4 x i1> %a1, <4 x i1>* %p2) nounwind {
|
|
; CHECK-LABEL: uaddo_v4i1:
|
|
; CHECK: // %bb.0:
|
|
; CHECK-NEXT: movi v2.4h, #1
|
|
; CHECK-NEXT: and v1.8b, v1.8b, v2.8b
|
|
; CHECK-NEXT: and v0.8b, v0.8b, v2.8b
|
|
; CHECK-NEXT: add v1.4h, v0.4h, v1.4h
|
|
; CHECK-NEXT: umov w9, v1.h[1]
|
|
; CHECK-NEXT: umov w8, v1.h[0]
|
|
; CHECK-NEXT: and w9, w9, #0x1
|
|
; CHECK-NEXT: bfi w8, w9, #1, #1
|
|
; CHECK-NEXT: umov w9, v1.h[2]
|
|
; CHECK-NEXT: and v0.8b, v1.8b, v2.8b
|
|
; CHECK-NEXT: and w9, w9, #0x1
|
|
; CHECK-NEXT: cmeq v0.4h, v0.4h, v1.4h
|
|
; CHECK-NEXT: bfi w8, w9, #2, #1
|
|
; CHECK-NEXT: umov w9, v1.h[3]
|
|
; CHECK-NEXT: mvn v0.8b, v0.8b
|
|
; CHECK-NEXT: bfi w8, w9, #3, #29
|
|
; CHECK-NEXT: sshll v0.4s, v0.4h, #0
|
|
; CHECK-NEXT: and w8, w8, #0xf
|
|
; CHECK-NEXT: strb w8, [x0]
|
|
; CHECK-NEXT: ret
|
|
%t = call {<4 x i1>, <4 x i1>} @llvm.uadd.with.overflow.v4i1(<4 x i1> %a0, <4 x i1> %a1)
|
|
%val = extractvalue {<4 x i1>, <4 x i1>} %t, 0
|
|
%obit = extractvalue {<4 x i1>, <4 x i1>} %t, 1
|
|
%res = sext <4 x i1> %obit to <4 x i32>
|
|
store <4 x i1> %val, <4 x i1>* %p2
|
|
ret <4 x i32> %res
|
|
}
|
|
|
|
define <2 x i32> @uaddo_v2i128(<2 x i128> %a0, <2 x i128> %a1, <2 x i128>* %p2) nounwind {
|
|
; CHECK-LABEL: uaddo_v2i128:
|
|
; CHECK: // %bb.0:
|
|
; CHECK-NEXT: adds x9, x2, x6
|
|
; CHECK-NEXT: adcs x10, x3, x7
|
|
; CHECK-NEXT: cmp x9, x2
|
|
; CHECK-NEXT: cset w11, lo
|
|
; CHECK-NEXT: cmp x10, x3
|
|
; CHECK-NEXT: cset w12, lo
|
|
; CHECK-NEXT: csel w11, w11, w12, eq
|
|
; CHECK-NEXT: adds x12, x0, x4
|
|
; CHECK-NEXT: adcs x13, x1, x5
|
|
; CHECK-NEXT: cmp x12, x0
|
|
; CHECK-NEXT: cset w14, lo
|
|
; CHECK-NEXT: cmp x13, x1
|
|
; CHECK-NEXT: cset w15, lo
|
|
; CHECK-NEXT: csel w14, w14, w15, eq
|
|
; CHECK-NEXT: ldr x8, [sp]
|
|
; CHECK-NEXT: fmov s0, w14
|
|
; CHECK-NEXT: mov v0.s[1], w11
|
|
; CHECK-NEXT: shl v0.2s, v0.2s, #31
|
|
; CHECK-NEXT: sshr v0.2s, v0.2s, #31
|
|
; CHECK-NEXT: stp x9, x10, [x8, #16]
|
|
; CHECK-NEXT: stp x12, x13, [x8]
|
|
; CHECK-NEXT: ret
|
|
%t = call {<2 x i128>, <2 x i1>} @llvm.uadd.with.overflow.v2i128(<2 x i128> %a0, <2 x i128> %a1)
|
|
%val = extractvalue {<2 x i128>, <2 x i1>} %t, 0
|
|
%obit = extractvalue {<2 x i128>, <2 x i1>} %t, 1
|
|
%res = sext <2 x i1> %obit to <2 x i32>
|
|
store <2 x i128> %val, <2 x i128>* %p2
|
|
ret <2 x i32> %res
|
|
}
|