1
0
mirror of https://github.com/RPCS3/llvm-mirror.git synced 2024-11-22 18:54:02 +01:00
llvm-mirror/test/CodeGen/AArch64/sponentry.ll
Martin Storsjö ba79f9fef9 [AArch64] Match the windows canonical callee saved register order
On windows, the callee saved registers in a canonical prologue are
ordered starting from a lower register number at a lower stack
address (with the possible gap for aligning the stack at the top);
this is the opposite order that llvm normally produces.

To achieve this, reverse the order of the registers in the
assignCalleeSavedSpillSlots callback, to get the stack objects
laid out by PrologEpilogInserter in the right order, and adjust
computeCalleeSaveRegisterPairs to lay them out from the bottom up.

This allows generated prologs more often to match the format that
allows the unwind info to be written as packed info.

Differential Revision: https://reviews.llvm.org/D88677
2020-10-03 21:37:22 +03:00

105 lines
3.1 KiB
LLVM

; RUN: llc -mtriple=aarch64-windows-msvc -frame-pointer=all %s -o - | FileCheck %s
; RUN: llc -mtriple=aarch64-windows-msvc -fast-isel -frame-pointer=all %s -o - | FileCheck %s
; RUN: llc -mtriple=aarch64-windows-msvc %s -o - | FileCheck %s --check-prefix=NOFP
; RUN: llc -mtriple=aarch64-windows-msvc -fast-isel %s -o - | FileCheck %s --check-prefix=NOFP
@env2 = common dso_local global [24 x i64]* null, align 8
define dso_local void @bar() {
%1 = call i8* @llvm.sponentry()
%2 = load [24 x i64]*, [24 x i64]** @env2, align 8
%3 = getelementptr inbounds [24 x i64], [24 x i64]* %2, i32 0, i32 0
%4 = bitcast i64* %3 to i8*
%5 = call i32 @_setjmpex(i8* %4, i8* %1) #2
ret void
}
; CHECK: bar:
; CHECK: mov x29, sp
; CHECK: add x1, x29, #16
; CHECK: bl _setjmpex
; NOFP: str x30, [sp, #-16]!
; NOFP: add x1, sp, #16
define dso_local void @foo([24 x i64]*) {
%2 = alloca [24 x i64]*, align 8
%3 = alloca i32, align 4
%4 = alloca [100 x i32], align 4
store [24 x i64]* %0, [24 x i64]** %2, align 8
%5 = call i8* @llvm.sponentry()
%6 = load [24 x i64]*, [24 x i64]** %2, align 8
%7 = getelementptr inbounds [24 x i64], [24 x i64]* %6, i32 0, i32 0
%8 = bitcast i64* %7 to i8*
%9 = call i32 @_setjmpex(i8* %8, i8* %5)
store i32 %9, i32* %3, align 4
ret void
}
; CHECK: foo:
; CHECK: sub sp, sp, #448
; CHECK: add x29, sp, #424
; CHECK: add x1, x29, #24
; CHECK: bl _setjmpex
; NOFP: sub sp, sp, #432
; NOFP: add x1, sp, #432
define dso_local void @var_args(i8*, ...) {
%2 = alloca i8*, align 8
%3 = alloca i8*, align 8
store i8* %0, i8** %2, align 8
%4 = bitcast i8** %3 to i8*
call void @llvm.va_start(i8* %4)
%5 = load i8*, i8** %3, align 8
%6 = getelementptr inbounds i8, i8* %5, i64 8
store i8* %6, i8** %3, align 8
%7 = bitcast i8* %5 to i32*
%8 = load i32, i32* %7, align 8
%9 = bitcast i8** %3 to i8*
call void @llvm.va_end(i8* %9)
%10 = call i8* @llvm.sponentry()
%11 = load [24 x i64]*, [24 x i64]** @env2, align 8
%12 = getelementptr inbounds [24 x i64], [24 x i64]* %11, i32 0, i32 0
%13 = bitcast i64* %12 to i8*
%14 = call i32 @_setjmpex(i8* %13, i8* %10) #3
ret void
}
; CHECK: var_args:
; CHECK: sub sp, sp, #96
; CHECK: add x29, sp, #16
; CHECK: add x1, x29, #80
; CHECK: bl _setjmpex
; NOFP: sub sp, sp, #96
; NOFP: add x1, sp, #96
define dso_local void @manyargs(i64 %x1, i64 %x2, i64 %x3, i64 %x4, i64 %x5, i64 %x6, i64 %x7, i64 %x8, i64 %x9, i64 %x10) {
%1 = call i8* @llvm.sponentry()
%2 = load [24 x i64]*, [24 x i64]** @env2, align 8
%3 = getelementptr inbounds [24 x i64], [24 x i64]* %2, i32 0, i32 0
%4 = bitcast i64* %3 to i8*
%5 = call i32 @_setjmpex(i8* %4, i8* %1) #2
ret void
}
; CHECK: manyargs:
; CHECK: stp x29, x30, [sp, #-16]!
; CHECK: add x1, x29, #16
; NOFP: str x30, [sp, #-16]!
; NOFP: add x1, sp, #16
; Function Attrs: nounwind readnone
declare i8* @llvm.sponentry()
; Function Attrs: returns_twice
declare dso_local i32 @_setjmpex(i8*, i8*)
; Function Attrs: nounwind
declare void @llvm.va_start(i8*) #1
; Function Attrs: nounwind
declare void @llvm.va_end(i8*) #1