mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-23 03:02:36 +01:00
dd9ab77318
Currently the default C calling convention functions are treated the same as compute kernels. Make this explicit so the default calling convention can be changed to a non-kernel. Converted with perl -pi -e 's/define void/define amdgpu_kernel void/' on the relevant test directories (and undoing in one place that actually wanted a non-kernel). llvm-svn: 298444
129 lines
4.4 KiB
LLVM
129 lines
4.4 KiB
LLVM
; RUN: llc -march=amdgcn -verify-machineinstrs -enable-unsafe-fp-math < %s | FileCheck -check-prefix=SI -check-prefix=FUNC %s
|
|
; RUN: llc -march=amdgcn -mcpu=bonaire -verify-machineinstrs -enable-unsafe-fp-math < %s | FileCheck -check-prefix=CI -check-prefix=FUNC %s
|
|
; RUN: llc -march=amdgcn -mcpu=tonga -mattr=-flat-for-global -verify-machineinstrs -enable-unsafe-fp-math < %s | FileCheck -check-prefix=CI -check-prefix=FUNC %s
|
|
|
|
declare double @llvm.fabs.f64(double %Val)
|
|
declare double @llvm.floor.f64(double) nounwind readnone
|
|
declare <2 x double> @llvm.floor.v2f64(<2 x double>) nounwind readnone
|
|
declare <3 x double> @llvm.floor.v3f64(<3 x double>) nounwind readnone
|
|
declare <4 x double> @llvm.floor.v4f64(<4 x double>) nounwind readnone
|
|
declare <8 x double> @llvm.floor.v8f64(<8 x double>) nounwind readnone
|
|
declare <16 x double> @llvm.floor.v16f64(<16 x double>) nounwind readnone
|
|
|
|
; FUNC-LABEL: {{^}}ffloor_f64:
|
|
; CI: v_floor_f64_e32
|
|
; SI: v_fract_f64_e32
|
|
; SI-DAG: v_min_f64
|
|
; SI-DAG: v_cmp_class_f64_e64 vcc
|
|
; SI: v_cndmask_b32_e32
|
|
; SI: v_cndmask_b32_e32
|
|
; SI: v_add_f64
|
|
; SI: s_endpgm
|
|
define amdgpu_kernel void @ffloor_f64(double addrspace(1)* %out, double %x) {
|
|
%y = call double @llvm.floor.f64(double %x) nounwind readnone
|
|
store double %y, double addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: {{^}}ffloor_f64_neg:
|
|
; CI: v_floor_f64_e64
|
|
; SI: v_fract_f64_e64 {{v[[0-9]+:[0-9]+]}}, -[[INPUT:s[[0-9]+:[0-9]+]]]
|
|
; SI-DAG: v_min_f64
|
|
; SI-DAG: v_cmp_class_f64_e64 vcc
|
|
; SI: v_cndmask_b32_e32
|
|
; SI: v_cndmask_b32_e32
|
|
; SI: v_add_f64 {{v[[0-9]+:[0-9]+]}}, -[[INPUT]]
|
|
; SI: s_endpgm
|
|
define amdgpu_kernel void @ffloor_f64_neg(double addrspace(1)* %out, double %x) {
|
|
%neg = fsub double 0.0, %x
|
|
%y = call double @llvm.floor.f64(double %neg) nounwind readnone
|
|
store double %y, double addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: {{^}}ffloor_f64_neg_abs:
|
|
; CI: v_floor_f64_e64
|
|
; SI: v_fract_f64_e64 {{v[[0-9]+:[0-9]+]}}, -|[[INPUT:s[[0-9]+:[0-9]+]]]|
|
|
; SI-DAG: v_min_f64
|
|
; SI-DAG: v_cmp_class_f64_e64 vcc
|
|
; SI: v_cndmask_b32_e32
|
|
; SI: v_cndmask_b32_e32
|
|
; SI: v_add_f64 {{v[[0-9]+:[0-9]+]}}, -|[[INPUT]]|
|
|
; SI: s_endpgm
|
|
define amdgpu_kernel void @ffloor_f64_neg_abs(double addrspace(1)* %out, double %x) {
|
|
%abs = call double @llvm.fabs.f64(double %x)
|
|
%neg = fsub double 0.0, %abs
|
|
%y = call double @llvm.floor.f64(double %neg) nounwind readnone
|
|
store double %y, double addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: {{^}}ffloor_v2f64:
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
define amdgpu_kernel void @ffloor_v2f64(<2 x double> addrspace(1)* %out, <2 x double> %x) {
|
|
%y = call <2 x double> @llvm.floor.v2f64(<2 x double> %x) nounwind readnone
|
|
store <2 x double> %y, <2 x double> addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: {{^}}ffloor_v3f64:
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI-NOT: v_floor_f64_e32
|
|
define amdgpu_kernel void @ffloor_v3f64(<3 x double> addrspace(1)* %out, <3 x double> %x) {
|
|
%y = call <3 x double> @llvm.floor.v3f64(<3 x double> %x) nounwind readnone
|
|
store <3 x double> %y, <3 x double> addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: {{^}}ffloor_v4f64:
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
define amdgpu_kernel void @ffloor_v4f64(<4 x double> addrspace(1)* %out, <4 x double> %x) {
|
|
%y = call <4 x double> @llvm.floor.v4f64(<4 x double> %x) nounwind readnone
|
|
store <4 x double> %y, <4 x double> addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: {{^}}ffloor_v8f64:
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
define amdgpu_kernel void @ffloor_v8f64(<8 x double> addrspace(1)* %out, <8 x double> %x) {
|
|
%y = call <8 x double> @llvm.floor.v8f64(<8 x double> %x) nounwind readnone
|
|
store <8 x double> %y, <8 x double> addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: {{^}}ffloor_v16f64:
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
; CI: v_floor_f64_e32
|
|
define amdgpu_kernel void @ffloor_v16f64(<16 x double> addrspace(1)* %out, <16 x double> %x) {
|
|
%y = call <16 x double> @llvm.floor.v16f64(<16 x double> %x) nounwind readnone
|
|
store <16 x double> %y, <16 x double> addrspace(1)* %out
|
|
ret void
|
|
}
|