mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-23 11:13:28 +01:00
4f480d4b83
This patch adds two new target-independent calling conventions for runtime calls - PreserveMost and PreserveAll. The target-specific implementation for X86-64 is defined as following: - Arguments are passed as for the default C calling convention - The same applies for the return value(s) - PreserveMost preserves all GPRs - except R11 - PreserveAll preserves all GPRs and all XMMs/YMMs - except R11 Reviewed by Lang and Philip llvm-svn: 199508
105 lines
4.2 KiB
LLVM
105 lines
4.2 KiB
LLVM
; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=corei7 | FileCheck --check-prefix=SSE %s
|
|
; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=corei7-avx | FileCheck --check-prefix=AVX %s
|
|
|
|
define preserve_allcc void @preserve_allcc1() nounwind {
|
|
entry:
|
|
;SSE-LABEL: preserve_allcc1
|
|
;SSE: pushq %r10
|
|
;SSE-NEXT: pushq %r9
|
|
;SSE-NEXT: pushq %r8
|
|
;SSE-NEXT: pushq %rdi
|
|
;SSE-NEXT: pushq %rsi
|
|
;SSE-NEXT: pushq %rdx
|
|
;SSE-NEXT: pushq %rcx
|
|
;SSE-NEXT: pushq %rax
|
|
;SSE-NEXT: pushq %rbp
|
|
;SSE-NEXT: pushq %r15
|
|
;SSE-NEXT: pushq %r14
|
|
;SSE-NEXT: pushq %r13
|
|
;SSE-NEXT: pushq %r12
|
|
;SSE-NEXT: pushq %rbx
|
|
;SSE: movaps %xmm15
|
|
;SSE-NEXT: movaps %xmm14
|
|
;SSE-NEXT: movaps %xmm13
|
|
;SSE-NEXT: movaps %xmm12
|
|
;SSE-NEXT: movaps %xmm11
|
|
;SSE-NEXT: movaps %xmm10
|
|
;SSE-NEXT: movaps %xmm9
|
|
;SSE-NEXT: movaps %xmm8
|
|
;SSE-NEXT: movaps %xmm7
|
|
;SSE-NEXT: movaps %xmm6
|
|
;SSE-NEXT: movaps %xmm5
|
|
;SSE-NEXT: movaps %xmm4
|
|
;SSE-NEXT: movaps %xmm3
|
|
;SSE-NEXT: movaps %xmm2
|
|
;SSE-NEXT: movaps %xmm1
|
|
;SSE-NEXT: movaps %xmm0
|
|
;AVX-LABEL: preserve_allcc1
|
|
;AVX: pushq %r10
|
|
;AVX-NEXT: pushq %r9
|
|
;AVX-NEXT: pushq %r8
|
|
;AVX-NEXT: pushq %rdi
|
|
;AVX-NEXT: pushq %rsi
|
|
;AVX-NEXT: pushq %rdx
|
|
;AVX-NEXT: pushq %rcx
|
|
;AVX-NEXT: pushq %rax
|
|
;AVX-NEXT: pushq %rbp
|
|
;AVX-NEXT: pushq %r15
|
|
;AVX-NEXT: pushq %r14
|
|
;AVX-NEXT: pushq %r13
|
|
;AVX-NEXT: pushq %r12
|
|
;AVX-NEXT: pushq %rbx
|
|
;AVX: vmovups %ymm15
|
|
;AVX-NEXT: vmovups %ymm14
|
|
;AVX-NEXT: vmovups %ymm13
|
|
;AVX-NEXT: vmovups %ymm12
|
|
;AVX-NEXT: vmovups %ymm11
|
|
;AVX-NEXT: vmovups %ymm10
|
|
;AVX-NEXT: vmovups %ymm9
|
|
;AVX-NEXT: vmovups %ymm8
|
|
;AVX-NEXT: vmovups %ymm7
|
|
;AVX-NEXT: vmovups %ymm6
|
|
;AVX-NEXT: vmovups %ymm5
|
|
;AVX-NEXT: vmovups %ymm4
|
|
;AVX-NEXT: vmovups %ymm3
|
|
;AVX-NEXT: vmovups %ymm2
|
|
;AVX-NEXT: vmovups %ymm1
|
|
;AVX-NEXT: vmovups %ymm0
|
|
call void asm sideeffect "", "~{rax},~{rbx},~{rcx},~{rdx},~{rsi},~{rdi},~{r8},~{r9},~{r10},~{r11},~{r12},~{r13},~{r14},~{r15},~{rbp},~{xmm0},~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15}"()
|
|
ret void
|
|
}
|
|
|
|
; Make sure only R11 is saved before the call
|
|
declare preserve_allcc void @bar(i64, i64, double, double)
|
|
define void @preserve_allcc2() nounwind {
|
|
entry:
|
|
;SSE-LABEL: preserve_allcc2
|
|
;SSE: movq %r11, [[REG:%[a-z0-9]+]]
|
|
;SSE-NOT: movaps %xmm
|
|
;SSE: movq [[REG]], %r11
|
|
%a0 = call i64 asm sideeffect "", "={rax}"() nounwind
|
|
%a1 = call i64 asm sideeffect "", "={rcx}"() nounwind
|
|
%a2 = call i64 asm sideeffect "", "={rdx}"() nounwind
|
|
%a3 = call i64 asm sideeffect "", "={r8}"() nounwind
|
|
%a4 = call i64 asm sideeffect "", "={r9}"() nounwind
|
|
%a5 = call i64 asm sideeffect "", "={r10}"() nounwind
|
|
%a6 = call i64 asm sideeffect "", "={r11}"() nounwind
|
|
%a10 = call <2 x double> asm sideeffect "", "={xmm2}"() nounwind
|
|
%a11 = call <2 x double> asm sideeffect "", "={xmm3}"() nounwind
|
|
%a12 = call <2 x double> asm sideeffect "", "={xmm4}"() nounwind
|
|
%a13 = call <2 x double> asm sideeffect "", "={xmm5}"() nounwind
|
|
%a14 = call <2 x double> asm sideeffect "", "={xmm6}"() nounwind
|
|
%a15 = call <2 x double> asm sideeffect "", "={xmm7}"() nounwind
|
|
%a16 = call <2 x double> asm sideeffect "", "={xmm8}"() nounwind
|
|
%a17 = call <2 x double> asm sideeffect "", "={xmm9}"() nounwind
|
|
%a18 = call <2 x double> asm sideeffect "", "={xmm10}"() nounwind
|
|
%a19 = call <2 x double> asm sideeffect "", "={xmm11}"() nounwind
|
|
%a20 = call <2 x double> asm sideeffect "", "={xmm12}"() nounwind
|
|
%a21 = call <2 x double> asm sideeffect "", "={xmm13}"() nounwind
|
|
%a22 = call <2 x double> asm sideeffect "", "={xmm14}"() nounwind
|
|
%a23 = call <2 x double> asm sideeffect "", "={xmm15}"() nounwind
|
|
call preserve_allcc void @bar(i64 1, i64 2, double 3.0, double 4.0)
|
|
call void asm sideeffect "", "{rax},{rcx},{rdx},{r8},{r9},{r10},{r11},{xmm2},{xmm3},{xmm4},{xmm5},{xmm6},{xmm7},{xmm8},{xmm9},{xmm10},{xmm11},{xmm12},{xmm13},{xmm14},{xmm15}"(i64 %a0, i64 %a1, i64 %a2, i64 %a3, i64 %a4, i64 %a5, i64 %a6, <2 x double> %a10, <2 x double> %a11, <2 x double> %a12, <2 x double> %a13, <2 x double> %a14, <2 x double> %a15, <2 x double> %a16, <2 x double> %a17, <2 x double> %a18, <2 x double> %a19, <2 x double> %a20, <2 x double> %a21, <2 x double> %a22, <2 x double> %a23)
|
|
ret void
|
|
}
|