mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-26 04:32:44 +01:00
8a77617721
Summary: This patch fixes an issue so that the right alias is printed when the instruction has tied operands. It checks the number of operands in the resulting instruction as opposed to the alias, and then skips over tied operands that should not be printed in the alias. This allows to generate the preferred assembly syntax for the AArch64 'ins' instruction, which should always be displayed as 'mov' according to the ARM Architecture Reference Manual. Several unit tests have changed as a result, but only to reflect the preferred disassembly. Some other InstAlias patterns (movk/bic/orr) needed a slight adjustment to stop them becoming the default and breaking other unit tests. Please note that the patch is mostly the same as https://reviews.llvm.org/D29219 which was reverted because of an issue found when running TableGen with the Address Sanitizer. That issue has been addressed in this iteration of the patch. Reviewers: rengolin, stoklund, huntergr, SjoerdMeijer, rovka Reviewed By: rengolin, SjoerdMeijer Subscribers: fhahn, aemerson, javed.absar, kristof.beyls, llvm-commits Differential Revision: https://reviews.llvm.org/D40030 llvm-svn: 318650
179 lines
6.0 KiB
LLVM
179 lines
6.0 KiB
LLVM
; RUN: llc < %s -mtriple aarch64-apple-darwin -asm-verbose=false -disable-post-ra | FileCheck %s
|
|
|
|
target datalayout = "e-m:o-i64:64-i128:128-n32:64-S128"
|
|
|
|
;============ v1f32
|
|
|
|
; WidenVecRes same
|
|
define <1 x float> @test_copysign_v1f32_v1f32(<1 x float> %a, <1 x float> %b) #0 {
|
|
; CHECK-LABEL: test_copysign_v1f32_v1f32:
|
|
; CHECK-NEXT: movi.2s v2, #128, lsl #24
|
|
; CHECK-NEXT: bit.8b v0, v1, v2
|
|
; CHECK-NEXT: ret
|
|
%r = call <1 x float> @llvm.copysign.v1f32(<1 x float> %a, <1 x float> %b)
|
|
ret <1 x float> %r
|
|
}
|
|
|
|
; WidenVecRes mismatched
|
|
define <1 x float> @test_copysign_v1f32_v1f64(<1 x float> %a, <1 x double> %b) #0 {
|
|
; CHECK-LABEL: test_copysign_v1f32_v1f64:
|
|
; CHECK-NEXT: fcvt s1, d1
|
|
; CHECK-NEXT: movi.4s v2, #128, lsl #24
|
|
; CHECK-NEXT: bit.16b v0, v1, v2
|
|
; CHECK-NEXT: ret
|
|
%tmp0 = fptrunc <1 x double> %b to <1 x float>
|
|
%r = call <1 x float> @llvm.copysign.v1f32(<1 x float> %a, <1 x float> %tmp0)
|
|
ret <1 x float> %r
|
|
}
|
|
|
|
declare <1 x float> @llvm.copysign.v1f32(<1 x float> %a, <1 x float> %b) #0
|
|
|
|
;============ v1f64
|
|
|
|
; WidenVecOp #1
|
|
define <1 x double> @test_copysign_v1f64_v1f32(<1 x double> %a, <1 x float> %b) #0 {
|
|
; CHECK-LABEL: test_copysign_v1f64_v1f32:
|
|
; CHECK-NEXT: fcvt d1, s1
|
|
; CHECK-NEXT: movi.2d v2, #0000000000000000
|
|
; CHECK-NEXT: fneg.2d v2, v2
|
|
; CHECK-NEXT: bit.16b v0, v1, v2
|
|
; CHECK-NEXT: ret
|
|
%tmp0 = fpext <1 x float> %b to <1 x double>
|
|
%r = call <1 x double> @llvm.copysign.v1f64(<1 x double> %a, <1 x double> %tmp0)
|
|
ret <1 x double> %r
|
|
}
|
|
|
|
define <1 x double> @test_copysign_v1f64_v1f64(<1 x double> %a, <1 x double> %b) #0 {
|
|
; CHECK-LABEL: test_copysign_v1f64_v1f64:
|
|
; CHECK-NEXT: movi.2d v2, #0000000000000000
|
|
; CHECK-NEXT: fneg.2d v2, v2
|
|
; CHECK-NEXT: bit.16b v0, v1, v2
|
|
; CHECK-NEXT: ret
|
|
%r = call <1 x double> @llvm.copysign.v1f64(<1 x double> %a, <1 x double> %b)
|
|
ret <1 x double> %r
|
|
}
|
|
|
|
declare <1 x double> @llvm.copysign.v1f64(<1 x double> %a, <1 x double> %b) #0
|
|
|
|
;============ v2f32
|
|
|
|
define <2 x float> @test_copysign_v2f32_v2f32(<2 x float> %a, <2 x float> %b) #0 {
|
|
; CHECK-LABEL: test_copysign_v2f32_v2f32:
|
|
; CHECK-NEXT: movi.2s v2, #128, lsl #24
|
|
; CHECK-NEXT: bit.8b v0, v1, v2
|
|
; CHECK-NEXT: ret
|
|
%r = call <2 x float> @llvm.copysign.v2f32(<2 x float> %a, <2 x float> %b)
|
|
ret <2 x float> %r
|
|
}
|
|
|
|
define <2 x float> @test_copysign_v2f32_v2f64(<2 x float> %a, <2 x double> %b) #0 {
|
|
; CHECK-LABEL: test_copysign_v2f32_v2f64:
|
|
; CHECK-NEXT: fcvtn v1.2s, v1.2d
|
|
; CHECK-NEXT: movi.2s v2, #128, lsl #24
|
|
; CHECK-NEXT: bit.8b v0, v1, v2
|
|
; CHECK-NEXT: ret
|
|
%tmp0 = fptrunc <2 x double> %b to <2 x float>
|
|
%r = call <2 x float> @llvm.copysign.v2f32(<2 x float> %a, <2 x float> %tmp0)
|
|
ret <2 x float> %r
|
|
}
|
|
|
|
declare <2 x float> @llvm.copysign.v2f32(<2 x float> %a, <2 x float> %b) #0
|
|
|
|
;============ v4f32
|
|
|
|
define <4 x float> @test_copysign_v4f32_v4f32(<4 x float> %a, <4 x float> %b) #0 {
|
|
; CHECK-LABEL: test_copysign_v4f32_v4f32:
|
|
; CHECK-NEXT: movi.4s v2, #128, lsl #24
|
|
; CHECK-NEXT: bit.16b v0, v1, v2
|
|
; CHECK-NEXT: ret
|
|
%r = call <4 x float> @llvm.copysign.v4f32(<4 x float> %a, <4 x float> %b)
|
|
ret <4 x float> %r
|
|
}
|
|
|
|
; SplitVecOp #1
|
|
define <4 x float> @test_copysign_v4f32_v4f64(<4 x float> %a, <4 x double> %b) #0 {
|
|
; CHECK-LABEL: test_copysign_v4f32_v4f64:
|
|
; CHECK-NEXT: mov s3, v0[1]
|
|
; CHECK-NEXT: movi.4s v4, #128, lsl #24
|
|
; CHECK-NEXT: fcvt s5, d1
|
|
; CHECK-NEXT: mov s6, v0[2]
|
|
; CHECK-NEXT: mov s7, v0[3]
|
|
; CHECK-NEXT: bit.16b v0, v5, v4
|
|
; CHECK-NEXT: fcvt s5, d2
|
|
; CHECK-NEXT: bit.16b v6, v5, v4
|
|
; CHECK-NEXT: mov d1, v1[1]
|
|
; CHECK-NEXT: fcvt s1, d1
|
|
; CHECK-NEXT: bit.16b v3, v1, v4
|
|
; CHECK-NEXT: mov d1, v2[1]
|
|
; CHECK-NEXT: fcvt s1, d1
|
|
; CHECK-NEXT: mov.s v0[1], v3[0]
|
|
; CHECK-NEXT: mov.s v0[2], v6[0]
|
|
; CHECK-NEXT: bit.16b v7, v1, v4
|
|
; CHECK-NEXT: mov.s v0[3], v7[0]
|
|
; CHECK-NEXT: ret
|
|
%tmp0 = fptrunc <4 x double> %b to <4 x float>
|
|
%r = call <4 x float> @llvm.copysign.v4f32(<4 x float> %a, <4 x float> %tmp0)
|
|
ret <4 x float> %r
|
|
}
|
|
|
|
declare <4 x float> @llvm.copysign.v4f32(<4 x float> %a, <4 x float> %b) #0
|
|
|
|
;============ v2f64
|
|
|
|
define <2 x double> @test_copysign_v2f64_v232(<2 x double> %a, <2 x float> %b) #0 {
|
|
; CHECK-LABEL: test_copysign_v2f64_v232:
|
|
; CHECK-NEXT: movi.2d v2, #0000000000000000
|
|
; CHECK-NEXT: fneg.2d v2, v2
|
|
; CHECK-NEXT: fcvtl v1.2d, v1.2s
|
|
; CHECK-NEXT: bit.16b v0, v1, v2
|
|
; CHECK-NEXT: ret
|
|
%tmp0 = fpext <2 x float> %b to <2 x double>
|
|
%r = call <2 x double> @llvm.copysign.v2f64(<2 x double> %a, <2 x double> %tmp0)
|
|
ret <2 x double> %r
|
|
}
|
|
|
|
define <2 x double> @test_copysign_v2f64_v2f64(<2 x double> %a, <2 x double> %b) #0 {
|
|
; CHECK-LABEL: test_copysign_v2f64_v2f64:
|
|
; CHECK-NEXT: movi.2d v2, #0000000000000000
|
|
; CHECK-NEXT: fneg.2d v2, v2
|
|
; CHECK-NEXT: bit.16b v0, v1, v2
|
|
; CHECK-NEXT: ret
|
|
%r = call <2 x double> @llvm.copysign.v2f64(<2 x double> %a, <2 x double> %b)
|
|
ret <2 x double> %r
|
|
}
|
|
|
|
declare <2 x double> @llvm.copysign.v2f64(<2 x double> %a, <2 x double> %b) #0
|
|
|
|
;============ v4f64
|
|
|
|
; SplitVecRes mismatched
|
|
define <4 x double> @test_copysign_v4f64_v4f32(<4 x double> %a, <4 x float> %b) #0 {
|
|
; CHECK-LABEL: test_copysign_v4f64_v4f32:
|
|
; CHECK-NEXT: movi.2d v3, #0000000000000000
|
|
; CHECK-NEXT: fcvtl2 v4.2d, v2.4s
|
|
; CHECK-NEXT: fcvtl v2.2d, v2.2s
|
|
; CHECK-NEXT: fneg.2d v3, v3
|
|
; CHECK-NEXT: bit.16b v1, v4, v3
|
|
; CHECK-NEXT: bit.16b v0, v2, v3
|
|
; CHECK-NEXT: ret
|
|
%tmp0 = fpext <4 x float> %b to <4 x double>
|
|
%r = call <4 x double> @llvm.copysign.v4f64(<4 x double> %a, <4 x double> %tmp0)
|
|
ret <4 x double> %r
|
|
}
|
|
|
|
; SplitVecRes same
|
|
define <4 x double> @test_copysign_v4f64_v4f64(<4 x double> %a, <4 x double> %b) #0 {
|
|
; CHECK-LABEL: test_copysign_v4f64_v4f64:
|
|
; CHECK-NEXT: movi.2d v4, #0000000000000000
|
|
; CHECK-NEXT: fneg.2d v4, v4
|
|
; CHECK-NEXT: bit.16b v0, v2, v4
|
|
; CHECK-NEXT: bit.16b v1, v3, v4
|
|
; CHECK-NEXT: ret
|
|
%r = call <4 x double> @llvm.copysign.v4f64(<4 x double> %a, <4 x double> %b)
|
|
ret <4 x double> %r
|
|
}
|
|
|
|
declare <4 x double> @llvm.copysign.v4f64(<4 x double> %a, <4 x double> %b) #0
|
|
|
|
attributes #0 = { nounwind }
|