mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-22 18:54:02 +01:00
95cb757dac
This patch adds support for S_ANDN2, S_ORN2 32-bit and 64-bit instructions and adds splits to move them to the vector unit (for which there is no equivalent instruction). It modifies the way that the more complex scalar instructions are lowered to vector instructions by first breaking them down to sequences of simpler scalar instructions which are then lowered through the existing code paths. The pattern for S_XNOR has also been updated to apply inversion to one input rather than the output of the XOR as the result is equivalent and may allow leaving the NOT instruction on the scalar unit. A new tests for NAND, NOR, ANDN2 and ORN2 have been added, and existing tests now hit the new instructions (and have been modified accordingly). Differential: https://reviews.llvm.org/D54714 llvm-svn: 347877
111 lines
3.9 KiB
LLVM
111 lines
3.9 KiB
LLVM
; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=tahiti < %s | FileCheck -check-prefix=SI -check-prefix=FUNC %s
|
|
; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=bonaire < %s | FileCheck -check-prefix=CI -check-prefix=FUNC %s
|
|
; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=tonga < %s | FileCheck -check-prefix=CI -check-prefix=FUNC %s
|
|
|
|
declare double @llvm.trunc.f64(double) nounwind readnone
|
|
declare <2 x double> @llvm.trunc.v2f64(<2 x double>) nounwind readnone
|
|
declare <3 x double> @llvm.trunc.v3f64(<3 x double>) nounwind readnone
|
|
declare <4 x double> @llvm.trunc.v4f64(<4 x double>) nounwind readnone
|
|
declare <8 x double> @llvm.trunc.v8f64(<8 x double>) nounwind readnone
|
|
declare <16 x double> @llvm.trunc.v16f64(<16 x double>) nounwind readnone
|
|
|
|
; FUNC-LABEL: {{^}}v_ftrunc_f64:
|
|
; CI: v_trunc_f64
|
|
; SI: v_bfe_u32 {{v[0-9]+}}, {{v[0-9]+}}, 20, 11
|
|
; SI: s_endpgm
|
|
define amdgpu_kernel void @v_ftrunc_f64(double addrspace(1)* %out, double addrspace(1)* %in) {
|
|
%x = load double, double addrspace(1)* %in, align 8
|
|
%y = call double @llvm.trunc.f64(double %x) nounwind readnone
|
|
store double %y, double addrspace(1)* %out, align 8
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: {{^}}ftrunc_f64:
|
|
; CI: v_trunc_f64_e32
|
|
|
|
; SI: s_bfe_u32 [[SEXP:s[0-9]+]], {{s[0-9]+}}, 0xb0014
|
|
; SI-DAG: s_and_b32 s{{[0-9]+}}, s{{[0-9]+}}, 0x80000000
|
|
; SI-DAG: s_add_i32 [[SEXP1:s[0-9]+]], [[SEXP]], 0xfffffc01
|
|
; SI-DAG: s_lshr_b64 s[{{[0-9]+:[0-9]+}}], s[{{[0-9]+:[0-9]+}}], [[SEXP1]]
|
|
; SI-DAG: s_andn2_b64
|
|
; SI-DAG: cmp_gt_i32
|
|
; SI-DAG: cndmask_b32
|
|
; SI-DAG: cndmask_b32
|
|
; SI-DAG: cmp_lt_i32
|
|
; SI-DAG: cndmask_b32
|
|
; SI-DAG: cndmask_b32
|
|
; SI: s_endpgm
|
|
define amdgpu_kernel void @ftrunc_f64(double addrspace(1)* %out, double %x) {
|
|
%y = call double @llvm.trunc.f64(double %x) nounwind readnone
|
|
store double %y, double addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: {{^}}ftrunc_v2f64:
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
define amdgpu_kernel void @ftrunc_v2f64(<2 x double> addrspace(1)* %out, <2 x double> %x) {
|
|
%y = call <2 x double> @llvm.trunc.v2f64(<2 x double> %x) nounwind readnone
|
|
store <2 x double> %y, <2 x double> addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FIXME-FUNC-LABEL: {{^}}ftrunc_v3f64:
|
|
; FIXME-CI: v_trunc_f64_e32
|
|
; FIXME-CI: v_trunc_f64_e32
|
|
; FIXME-CI: v_trunc_f64_e32
|
|
; define amdgpu_kernel void @ftrunc_v3f64(<3 x double> addrspace(1)* %out, <3 x double> %x) {
|
|
; %y = call <3 x double> @llvm.trunc.v3f64(<3 x double> %x) nounwind readnone
|
|
; store <3 x double> %y, <3 x double> addrspace(1)* %out
|
|
; ret void
|
|
; }
|
|
|
|
; FUNC-LABEL: {{^}}ftrunc_v4f64:
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
define amdgpu_kernel void @ftrunc_v4f64(<4 x double> addrspace(1)* %out, <4 x double> %x) {
|
|
%y = call <4 x double> @llvm.trunc.v4f64(<4 x double> %x) nounwind readnone
|
|
store <4 x double> %y, <4 x double> addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: {{^}}ftrunc_v8f64:
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
define amdgpu_kernel void @ftrunc_v8f64(<8 x double> addrspace(1)* %out, <8 x double> %x) {
|
|
%y = call <8 x double> @llvm.trunc.v8f64(<8 x double> %x) nounwind readnone
|
|
store <8 x double> %y, <8 x double> addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: {{^}}ftrunc_v16f64:
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
; CI: v_trunc_f64_e32
|
|
define amdgpu_kernel void @ftrunc_v16f64(<16 x double> addrspace(1)* %out, <16 x double> %x) {
|
|
%y = call <16 x double> @llvm.trunc.v16f64(<16 x double> %x) nounwind readnone
|
|
store <16 x double> %y, <16 x double> addrspace(1)* %out
|
|
ret void
|
|
}
|