mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-10-23 21:13:02 +02:00
a52916b4b8
Summary: G_LOAD/G_STORE, add alternative RegisterBank mapping. For G_LOAD, Fast and Greedy mode choose the same RegisterBank mapping (GprRegBank ) for the G_GLOAD + G_FADD , can't get rid of cross register bank copy GprRegBank->VecRegBank. Reviewers: zvi, rovka, qcolombet, ab Reviewed By: zvi Subscribers: llvm-commits, dberris, kristof.beyls, eladcohen, guyblank Differential Revision: https://reviews.llvm.org/D30979 llvm-svn: 298907
187 lines
5.1 KiB
LLVM
187 lines
5.1 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc -mtriple=x86_64-linux-gnu -global-isel < %s -o - | FileCheck %s --check-prefix=ALL --check-prefix=SSE
|
|
; RUN: llc -mtriple=x86_64-linux-gnu -mattr=+avx -global-isel < %s -o - | FileCheck %s --check-prefix=ALL --check-prefix=ALL_AVX --check-prefix=AVX
|
|
; RUN: llc -mtriple=x86_64-linux-gnu -mattr=+avx512f -global-isel < %s -o - | FileCheck %s --check-prefix=ALL --check-prefix=ALL_AVX --check-prefix=AVX512F
|
|
; RUN: llc -mtriple=x86_64-linux-gnu -mattr=+avx512f -mattr=+avx512vl -global-isel < %s -o - | FileCheck %s --check-prefix=ALL --check-prefix=ALL_AVX --check-prefix=AVX512VL
|
|
|
|
define i64 @test_add_i64(i64 %arg1, i64 %arg2) {
|
|
; ALL-LABEL: test_add_i64:
|
|
; ALL: # BB#0:
|
|
; ALL-NEXT: leaq (%rsi,%rdi), %rax
|
|
; ALL-NEXT: retq
|
|
%ret = add i64 %arg1, %arg2
|
|
ret i64 %ret
|
|
}
|
|
|
|
define i32 @test_add_i32(i32 %arg1, i32 %arg2) {
|
|
; ALL-LABEL: test_add_i32:
|
|
; ALL: # BB#0:
|
|
; ALL-NEXT: # kill: %EDI<def> %EDI<kill> %RDI<def>
|
|
; ALL-NEXT: # kill: %ESI<def> %ESI<kill> %RSI<def>
|
|
; ALL-NEXT: leal (%rsi,%rdi), %eax
|
|
; ALL-NEXT: retq
|
|
%ret = add i32 %arg1, %arg2
|
|
ret i32 %ret
|
|
}
|
|
|
|
define i64 @test_sub_i64(i64 %arg1, i64 %arg2) {
|
|
; ALL-LABEL: test_sub_i64:
|
|
; ALL: # BB#0:
|
|
; ALL-NEXT: subq %rsi, %rdi
|
|
; ALL-NEXT: movq %rdi, %rax
|
|
; ALL-NEXT: retq
|
|
%ret = sub i64 %arg1, %arg2
|
|
ret i64 %ret
|
|
}
|
|
|
|
define i32 @test_sub_i32(i32 %arg1, i32 %arg2) {
|
|
; ALL-LABEL: test_sub_i32:
|
|
; ALL: # BB#0:
|
|
; ALL-NEXT: subl %esi, %edi
|
|
; ALL-NEXT: movl %edi, %eax
|
|
; ALL-NEXT: retq
|
|
%ret = sub i32 %arg1, %arg2
|
|
ret i32 %ret
|
|
}
|
|
|
|
define float @test_add_float(float %arg1, float %arg2) {
|
|
; SSE-LABEL: test_add_float:
|
|
; SSE: # BB#0:
|
|
; SSE-NEXT: addss %xmm1, %xmm0
|
|
; SSE-NEXT: retq
|
|
;
|
|
; ALL_AVX-LABEL: test_add_float:
|
|
; ALL_AVX: # BB#0:
|
|
; ALL_AVX-NEXT: vaddss %xmm1, %xmm0, %xmm0
|
|
; ALL_AVX-NEXT: retq
|
|
%ret = fadd float %arg1, %arg2
|
|
ret float %ret
|
|
}
|
|
|
|
define double @test_add_double(double %arg1, double %arg2) {
|
|
; SSE-LABEL: test_add_double:
|
|
; SSE: # BB#0:
|
|
; SSE-NEXT: addsd %xmm1, %xmm0
|
|
; SSE-NEXT: retq
|
|
;
|
|
; ALL_AVX-LABEL: test_add_double:
|
|
; ALL_AVX: # BB#0:
|
|
; ALL_AVX-NEXT: vaddsd %xmm1, %xmm0, %xmm0
|
|
; ALL_AVX-NEXT: retq
|
|
%ret = fadd double %arg1, %arg2
|
|
ret double %ret
|
|
}
|
|
|
|
define float @test_sub_float(float %arg1, float %arg2) {
|
|
; SSE-LABEL: test_sub_float:
|
|
; SSE: # BB#0:
|
|
; SSE-NEXT: subss %xmm1, %xmm0
|
|
; SSE-NEXT: retq
|
|
;
|
|
; ALL_AVX-LABEL: test_sub_float:
|
|
; ALL_AVX: # BB#0:
|
|
; ALL_AVX-NEXT: vsubss %xmm1, %xmm0, %xmm0
|
|
; ALL_AVX-NEXT: retq
|
|
%ret = fsub float %arg1, %arg2
|
|
ret float %ret
|
|
}
|
|
|
|
define double @test_sub_double(double %arg1, double %arg2) {
|
|
; SSE-LABEL: test_sub_double:
|
|
; SSE: # BB#0:
|
|
; SSE-NEXT: subsd %xmm1, %xmm0
|
|
; SSE-NEXT: retq
|
|
;
|
|
; ALL_AVX-LABEL: test_sub_double:
|
|
; ALL_AVX: # BB#0:
|
|
; ALL_AVX-NEXT: vsubsd %xmm1, %xmm0, %xmm0
|
|
; ALL_AVX-NEXT: retq
|
|
%ret = fsub double %arg1, %arg2
|
|
ret double %ret
|
|
}
|
|
|
|
define <4 x i32> @test_add_v4i32(<4 x i32> %arg1, <4 x i32> %arg2) {
|
|
; SSE-LABEL: test_add_v4i32:
|
|
; SSE: # BB#0:
|
|
; SSE-NEXT: paddd %xmm1, %xmm0
|
|
; SSE-NEXT: retq
|
|
;
|
|
; ALL_AVX-LABEL: test_add_v4i32:
|
|
; ALL_AVX: # BB#0:
|
|
; ALL_AVX-NEXT: vpaddd %xmm1, %xmm0, %xmm0
|
|
; ALL_AVX-NEXT: retq
|
|
%ret = add <4 x i32> %arg1, %arg2
|
|
ret <4 x i32> %ret
|
|
}
|
|
|
|
define <4 x i32> @test_sub_v4i32(<4 x i32> %arg1, <4 x i32> %arg2) {
|
|
; SSE-LABEL: test_sub_v4i32:
|
|
; SSE: # BB#0:
|
|
; SSE-NEXT: psubd %xmm1, %xmm0
|
|
; SSE-NEXT: retq
|
|
;
|
|
; ALL_AVX-LABEL: test_sub_v4i32:
|
|
; ALL_AVX: # BB#0:
|
|
; ALL_AVX-NEXT: vpsubd %xmm1, %xmm0, %xmm0
|
|
; ALL_AVX-NEXT: retq
|
|
%ret = sub <4 x i32> %arg1, %arg2
|
|
ret <4 x i32> %ret
|
|
}
|
|
|
|
define <4 x float> @test_add_v4f32(<4 x float> %arg1, <4 x float> %arg2) {
|
|
; SSE-LABEL: test_add_v4f32:
|
|
; SSE: # BB#0:
|
|
; SSE-NEXT: addps %xmm1, %xmm0
|
|
; SSE-NEXT: retq
|
|
;
|
|
; ALL_AVX-LABEL: test_add_v4f32:
|
|
; ALL_AVX: # BB#0:
|
|
; ALL_AVX-NEXT: vaddps %xmm1, %xmm0, %xmm0
|
|
; ALL_AVX-NEXT: retq
|
|
%ret = fadd <4 x float> %arg1, %arg2
|
|
ret <4 x float> %ret
|
|
}
|
|
|
|
define <4 x float> @test_sub_v4f32(<4 x float> %arg1, <4 x float> %arg2) {
|
|
; SSE-LABEL: test_sub_v4f32:
|
|
; SSE: # BB#0:
|
|
; SSE-NEXT: subps %xmm1, %xmm0
|
|
; SSE-NEXT: retq
|
|
;
|
|
; ALL_AVX-LABEL: test_sub_v4f32:
|
|
; ALL_AVX: # BB#0:
|
|
; ALL_AVX-NEXT: vsubps %xmm1, %xmm0, %xmm0
|
|
; ALL_AVX-NEXT: retq
|
|
%ret = fsub <4 x float> %arg1, %arg2
|
|
ret <4 x float> %ret
|
|
}
|
|
|
|
define i32 @test_copy_float(float %val) {
|
|
; SSE-LABEL: test_copy_float:
|
|
; SSE: # BB#0:
|
|
; SSE-NEXT: movd %xmm0, %eax
|
|
; SSE-NEXT: retq
|
|
;
|
|
; ALL_AVX-LABEL: test_copy_float:
|
|
; ALL_AVX: # BB#0:
|
|
; ALL_AVX-NEXT: vmovd %xmm0, %eax
|
|
; ALL_AVX-NEXT: retq
|
|
%r = bitcast float %val to i32
|
|
ret i32 %r
|
|
}
|
|
|
|
define float @test_copy_i32(i32 %val) {
|
|
; SSE-LABEL: test_copy_i32:
|
|
; SSE: # BB#0:
|
|
; SSE-NEXT: movd %edi, %xmm0
|
|
; SSE-NEXT: retq
|
|
;
|
|
; ALL_AVX-LABEL: test_copy_i32:
|
|
; ALL_AVX: # BB#0:
|
|
; ALL_AVX-NEXT: vmovd %edi, %xmm0
|
|
; ALL_AVX-NEXT: retq
|
|
%r = bitcast i32 %val to float
|
|
ret float %r
|
|
}
|
|
|