mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-23 11:13:28 +01:00
68d81cc406
Any llvm function with the "packed-stack" attribute will be compiled to use the packed stack layout which reuses unused parts of the incoming register save area. This is needed for building the Linux kernel. Review: Ulrich Weigand https://reviews.llvm.org/D70821
446 lines
16 KiB
LLVM
446 lines
16 KiB
LLVM
; Like frame-03.ll, but for z13. In this case we have 16 more registers
|
|
; available.
|
|
;
|
|
; RUN: llc < %s -mtriple=s390x-linux-gnu -mcpu=z13 | FileCheck %s
|
|
|
|
; This function should require all FPRs, but no other spill slots.
|
|
; We need to save and restore 8 of the 16 FPRs, so the frame size
|
|
; should be exactly 8 * 8 = 64. The CFA offset is 160
|
|
; (the caller-allocated part of the frame) + 64.
|
|
define void @f1(double *%ptr) {
|
|
; CHECK-LABEL: f1:
|
|
; CHECK: aghi %r15, -64
|
|
; CHECK: .cfi_def_cfa_offset 224
|
|
; CHECK: std %f8, 56(%r15)
|
|
; CHECK: std %f9, 48(%r15)
|
|
; CHECK: std %f10, 40(%r15)
|
|
; CHECK: std %f11, 32(%r15)
|
|
; CHECK: std %f12, 24(%r15)
|
|
; CHECK: std %f13, 16(%r15)
|
|
; CHECK: std %f14, 8(%r15)
|
|
; CHECK: std %f15, 0(%r15)
|
|
; CHECK: .cfi_offset %f8, -168
|
|
; CHECK: .cfi_offset %f9, -176
|
|
; CHECK: .cfi_offset %f10, -184
|
|
; CHECK: .cfi_offset %f11, -192
|
|
; CHECK: .cfi_offset %f12, -200
|
|
; CHECK: .cfi_offset %f13, -208
|
|
; CHECK: .cfi_offset %f14, -216
|
|
; CHECK: .cfi_offset %f15, -224
|
|
; CHECK-DAG: ld %f0, 0(%r2)
|
|
; CHECK-DAG: ld %f7, 0(%r2)
|
|
; CHECK-DAG: ld %f8, 0(%r2)
|
|
; CHECK-DAG: ld %f15, 0(%r2)
|
|
; CHECK-DAG: vlrepg %v16, 0(%r2)
|
|
; CHECK-DAG: vlrepg %v23, 0(%r2)
|
|
; CHECK-DAG: vlrepg %v24, 0(%r2)
|
|
; CHECK-DAG: vlrepg %v31, 0(%r2)
|
|
; CHECK: ld %f8, 56(%r15)
|
|
; CHECK: ld %f9, 48(%r15)
|
|
; CHECK: ld %f10, 40(%r15)
|
|
; CHECK: ld %f11, 32(%r15)
|
|
; CHECK: ld %f12, 24(%r15)
|
|
; CHECK: ld %f13, 16(%r15)
|
|
; CHECK: ld %f14, 8(%r15)
|
|
; CHECK: ld %f15, 0(%r15)
|
|
; CHECK: aghi %r15, 64
|
|
; CHECK: br %r14
|
|
%l0 = load volatile double, double *%ptr
|
|
%l1 = load volatile double, double *%ptr
|
|
%l2 = load volatile double, double *%ptr
|
|
%l3 = load volatile double, double *%ptr
|
|
%l4 = load volatile double, double *%ptr
|
|
%l5 = load volatile double, double *%ptr
|
|
%l6 = load volatile double, double *%ptr
|
|
%l7 = load volatile double, double *%ptr
|
|
%l8 = load volatile double, double *%ptr
|
|
%l9 = load volatile double, double *%ptr
|
|
%l10 = load volatile double, double *%ptr
|
|
%l11 = load volatile double, double *%ptr
|
|
%l12 = load volatile double, double *%ptr
|
|
%l13 = load volatile double, double *%ptr
|
|
%l14 = load volatile double, double *%ptr
|
|
%l15 = load volatile double, double *%ptr
|
|
%l16 = load volatile double, double *%ptr
|
|
%l17 = load volatile double, double *%ptr
|
|
%l18 = load volatile double, double *%ptr
|
|
%l19 = load volatile double, double *%ptr
|
|
%l20 = load volatile double, double *%ptr
|
|
%l21 = load volatile double, double *%ptr
|
|
%l22 = load volatile double, double *%ptr
|
|
%l23 = load volatile double, double *%ptr
|
|
%l24 = load volatile double, double *%ptr
|
|
%l25 = load volatile double, double *%ptr
|
|
%l26 = load volatile double, double *%ptr
|
|
%l27 = load volatile double, double *%ptr
|
|
%l28 = load volatile double, double *%ptr
|
|
%l29 = load volatile double, double *%ptr
|
|
%l30 = load volatile double, double *%ptr
|
|
%l31 = load volatile double, double *%ptr
|
|
%acc0 = fsub double %l0, %l0
|
|
%acc1 = fsub double %l1, %acc0
|
|
%acc2 = fsub double %l2, %acc1
|
|
%acc3 = fsub double %l3, %acc2
|
|
%acc4 = fsub double %l4, %acc3
|
|
%acc5 = fsub double %l5, %acc4
|
|
%acc6 = fsub double %l6, %acc5
|
|
%acc7 = fsub double %l7, %acc6
|
|
%acc8 = fsub double %l8, %acc7
|
|
%acc9 = fsub double %l9, %acc8
|
|
%acc10 = fsub double %l10, %acc9
|
|
%acc11 = fsub double %l11, %acc10
|
|
%acc12 = fsub double %l12, %acc11
|
|
%acc13 = fsub double %l13, %acc12
|
|
%acc14 = fsub double %l14, %acc13
|
|
%acc15 = fsub double %l15, %acc14
|
|
%acc16 = fsub double %l16, %acc15
|
|
%acc17 = fsub double %l17, %acc16
|
|
%acc18 = fsub double %l18, %acc17
|
|
%acc19 = fsub double %l19, %acc18
|
|
%acc20 = fsub double %l20, %acc19
|
|
%acc21 = fsub double %l21, %acc20
|
|
%acc22 = fsub double %l22, %acc21
|
|
%acc23 = fsub double %l23, %acc22
|
|
%acc24 = fsub double %l24, %acc23
|
|
%acc25 = fsub double %l25, %acc24
|
|
%acc26 = fsub double %l26, %acc25
|
|
%acc27 = fsub double %l27, %acc26
|
|
%acc28 = fsub double %l28, %acc27
|
|
%acc29 = fsub double %l29, %acc28
|
|
%acc30 = fsub double %l30, %acc29
|
|
%acc31 = fsub double %l31, %acc30
|
|
store volatile double %acc0, double *%ptr
|
|
store volatile double %acc1, double *%ptr
|
|
store volatile double %acc2, double *%ptr
|
|
store volatile double %acc3, double *%ptr
|
|
store volatile double %acc4, double *%ptr
|
|
store volatile double %acc5, double *%ptr
|
|
store volatile double %acc6, double *%ptr
|
|
store volatile double %acc7, double *%ptr
|
|
store volatile double %acc8, double *%ptr
|
|
store volatile double %acc9, double *%ptr
|
|
store volatile double %acc10, double *%ptr
|
|
store volatile double %acc11, double *%ptr
|
|
store volatile double %acc12, double *%ptr
|
|
store volatile double %acc13, double *%ptr
|
|
store volatile double %acc14, double *%ptr
|
|
store volatile double %acc15, double *%ptr
|
|
store volatile double %acc16, double *%ptr
|
|
store volatile double %acc17, double *%ptr
|
|
store volatile double %acc18, double *%ptr
|
|
store volatile double %acc19, double *%ptr
|
|
store volatile double %acc20, double *%ptr
|
|
store volatile double %acc21, double *%ptr
|
|
store volatile double %acc22, double *%ptr
|
|
store volatile double %acc23, double *%ptr
|
|
store volatile double %acc24, double *%ptr
|
|
store volatile double %acc25, double *%ptr
|
|
store volatile double %acc26, double *%ptr
|
|
store volatile double %acc27, double *%ptr
|
|
store volatile double %acc28, double *%ptr
|
|
store volatile double %acc29, double *%ptr
|
|
store volatile double %acc30, double *%ptr
|
|
store volatile double %acc31, double *%ptr
|
|
ret void
|
|
}
|
|
|
|
; Like f1, but requires one fewer FPR. We allocate in numerical order,
|
|
; so %f15 is the one that gets dropped.
|
|
define void @f2(double *%ptr) {
|
|
; CHECK-LABEL: f2:
|
|
; CHECK: aghi %r15, -56
|
|
; CHECK: .cfi_def_cfa_offset 216
|
|
; CHECK: std %f8, 48(%r15)
|
|
; CHECK: std %f9, 40(%r15)
|
|
; CHECK: std %f10, 32(%r15)
|
|
; CHECK: std %f11, 24(%r15)
|
|
; CHECK: std %f12, 16(%r15)
|
|
; CHECK: std %f13, 8(%r15)
|
|
; CHECK: std %f14, 0(%r15)
|
|
; CHECK: .cfi_offset %f8, -168
|
|
; CHECK: .cfi_offset %f9, -176
|
|
; CHECK: .cfi_offset %f10, -184
|
|
; CHECK: .cfi_offset %f11, -192
|
|
; CHECK: .cfi_offset %f12, -200
|
|
; CHECK: .cfi_offset %f13, -208
|
|
; CHECK: .cfi_offset %f14, -216
|
|
; CHECK-NOT: %v15
|
|
; CHECK-NOT: %f15
|
|
; CHECK: ld %f8, 48(%r15)
|
|
; CHECK: ld %f9, 40(%r15)
|
|
; CHECK: ld %f10, 32(%r15)
|
|
; CHECK: ld %f11, 24(%r15)
|
|
; CHECK: ld %f12, 16(%r15)
|
|
; CHECK: ld %f13, 8(%r15)
|
|
; CHECK: ld %f14, 0(%r15)
|
|
; CHECK: aghi %r15, 56
|
|
; CHECK: br %r14
|
|
%l0 = load volatile double, double *%ptr
|
|
%l1 = load volatile double, double *%ptr
|
|
%l2 = load volatile double, double *%ptr
|
|
%l3 = load volatile double, double *%ptr
|
|
%l4 = load volatile double, double *%ptr
|
|
%l5 = load volatile double, double *%ptr
|
|
%l6 = load volatile double, double *%ptr
|
|
%l7 = load volatile double, double *%ptr
|
|
%l8 = load volatile double, double *%ptr
|
|
%l9 = load volatile double, double *%ptr
|
|
%l10 = load volatile double, double *%ptr
|
|
%l11 = load volatile double, double *%ptr
|
|
%l12 = load volatile double, double *%ptr
|
|
%l13 = load volatile double, double *%ptr
|
|
%l14 = load volatile double, double *%ptr
|
|
%l16 = load volatile double, double *%ptr
|
|
%l17 = load volatile double, double *%ptr
|
|
%l18 = load volatile double, double *%ptr
|
|
%l19 = load volatile double, double *%ptr
|
|
%l20 = load volatile double, double *%ptr
|
|
%l21 = load volatile double, double *%ptr
|
|
%l22 = load volatile double, double *%ptr
|
|
%l23 = load volatile double, double *%ptr
|
|
%l24 = load volatile double, double *%ptr
|
|
%l25 = load volatile double, double *%ptr
|
|
%l26 = load volatile double, double *%ptr
|
|
%l27 = load volatile double, double *%ptr
|
|
%l28 = load volatile double, double *%ptr
|
|
%l29 = load volatile double, double *%ptr
|
|
%l30 = load volatile double, double *%ptr
|
|
%l31 = load volatile double, double *%ptr
|
|
%acc0 = fsub double %l0, %l0
|
|
%acc1 = fsub double %l1, %acc0
|
|
%acc2 = fsub double %l2, %acc1
|
|
%acc3 = fsub double %l3, %acc2
|
|
%acc4 = fsub double %l4, %acc3
|
|
%acc5 = fsub double %l5, %acc4
|
|
%acc6 = fsub double %l6, %acc5
|
|
%acc7 = fsub double %l7, %acc6
|
|
%acc8 = fsub double %l8, %acc7
|
|
%acc9 = fsub double %l9, %acc8
|
|
%acc10 = fsub double %l10, %acc9
|
|
%acc11 = fsub double %l11, %acc10
|
|
%acc12 = fsub double %l12, %acc11
|
|
%acc13 = fsub double %l13, %acc12
|
|
%acc14 = fsub double %l14, %acc13
|
|
%acc16 = fsub double %l16, %acc14
|
|
%acc17 = fsub double %l17, %acc16
|
|
%acc18 = fsub double %l18, %acc17
|
|
%acc19 = fsub double %l19, %acc18
|
|
%acc20 = fsub double %l20, %acc19
|
|
%acc21 = fsub double %l21, %acc20
|
|
%acc22 = fsub double %l22, %acc21
|
|
%acc23 = fsub double %l23, %acc22
|
|
%acc24 = fsub double %l24, %acc23
|
|
%acc25 = fsub double %l25, %acc24
|
|
%acc26 = fsub double %l26, %acc25
|
|
%acc27 = fsub double %l27, %acc26
|
|
%acc28 = fsub double %l28, %acc27
|
|
%acc29 = fsub double %l29, %acc28
|
|
%acc30 = fsub double %l30, %acc29
|
|
%acc31 = fsub double %l31, %acc30
|
|
store volatile double %acc0, double *%ptr
|
|
store volatile double %acc1, double *%ptr
|
|
store volatile double %acc2, double *%ptr
|
|
store volatile double %acc3, double *%ptr
|
|
store volatile double %acc4, double *%ptr
|
|
store volatile double %acc5, double *%ptr
|
|
store volatile double %acc6, double *%ptr
|
|
store volatile double %acc7, double *%ptr
|
|
store volatile double %acc8, double *%ptr
|
|
store volatile double %acc9, double *%ptr
|
|
store volatile double %acc10, double *%ptr
|
|
store volatile double %acc11, double *%ptr
|
|
store volatile double %acc12, double *%ptr
|
|
store volatile double %acc13, double *%ptr
|
|
store volatile double %acc14, double *%ptr
|
|
store volatile double %acc16, double *%ptr
|
|
store volatile double %acc17, double *%ptr
|
|
store volatile double %acc18, double *%ptr
|
|
store volatile double %acc19, double *%ptr
|
|
store volatile double %acc20, double *%ptr
|
|
store volatile double %acc21, double *%ptr
|
|
store volatile double %acc22, double *%ptr
|
|
store volatile double %acc23, double *%ptr
|
|
store volatile double %acc24, double *%ptr
|
|
store volatile double %acc25, double *%ptr
|
|
store volatile double %acc26, double *%ptr
|
|
store volatile double %acc27, double *%ptr
|
|
store volatile double %acc28, double *%ptr
|
|
store volatile double %acc29, double *%ptr
|
|
store volatile double %acc30, double *%ptr
|
|
store volatile double %acc31, double *%ptr
|
|
ret void
|
|
}
|
|
|
|
; Like f1, but should require only one call-saved FPR.
|
|
define void @f3(double *%ptr) {
|
|
; CHECK-LABEL: f3:
|
|
; CHECK: aghi %r15, -8
|
|
; CHECK: .cfi_def_cfa_offset 168
|
|
; CHECK: std %f8, 0(%r15)
|
|
; CHECK: .cfi_offset %f8, -168
|
|
; CHECK-NOT: {{%[fv]9}}
|
|
; CHECK-NOT: {{%[fv]1[0-5]}}
|
|
; CHECK: ld %f8, 0(%r15)
|
|
; CHECK: aghi %r15, 8
|
|
; CHECK: br %r14
|
|
%l0 = load volatile double, double *%ptr
|
|
%l1 = load volatile double, double *%ptr
|
|
%l2 = load volatile double, double *%ptr
|
|
%l3 = load volatile double, double *%ptr
|
|
%l4 = load volatile double, double *%ptr
|
|
%l5 = load volatile double, double *%ptr
|
|
%l6 = load volatile double, double *%ptr
|
|
%l7 = load volatile double, double *%ptr
|
|
%l8 = load volatile double, double *%ptr
|
|
%l16 = load volatile double, double *%ptr
|
|
%l17 = load volatile double, double *%ptr
|
|
%l18 = load volatile double, double *%ptr
|
|
%l19 = load volatile double, double *%ptr
|
|
%l20 = load volatile double, double *%ptr
|
|
%l21 = load volatile double, double *%ptr
|
|
%l22 = load volatile double, double *%ptr
|
|
%l23 = load volatile double, double *%ptr
|
|
%l24 = load volatile double, double *%ptr
|
|
%l25 = load volatile double, double *%ptr
|
|
%l26 = load volatile double, double *%ptr
|
|
%l27 = load volatile double, double *%ptr
|
|
%l28 = load volatile double, double *%ptr
|
|
%l29 = load volatile double, double *%ptr
|
|
%l30 = load volatile double, double *%ptr
|
|
%l31 = load volatile double, double *%ptr
|
|
%acc0 = fsub double %l0, %l0
|
|
%acc1 = fsub double %l1, %acc0
|
|
%acc2 = fsub double %l2, %acc1
|
|
%acc3 = fsub double %l3, %acc2
|
|
%acc4 = fsub double %l4, %acc3
|
|
%acc5 = fsub double %l5, %acc4
|
|
%acc6 = fsub double %l6, %acc5
|
|
%acc7 = fsub double %l7, %acc6
|
|
%acc8 = fsub double %l8, %acc7
|
|
%acc16 = fsub double %l16, %acc8
|
|
%acc17 = fsub double %l17, %acc16
|
|
%acc18 = fsub double %l18, %acc17
|
|
%acc19 = fsub double %l19, %acc18
|
|
%acc20 = fsub double %l20, %acc19
|
|
%acc21 = fsub double %l21, %acc20
|
|
%acc22 = fsub double %l22, %acc21
|
|
%acc23 = fsub double %l23, %acc22
|
|
%acc24 = fsub double %l24, %acc23
|
|
%acc25 = fsub double %l25, %acc24
|
|
%acc26 = fsub double %l26, %acc25
|
|
%acc27 = fsub double %l27, %acc26
|
|
%acc28 = fsub double %l28, %acc27
|
|
%acc29 = fsub double %l29, %acc28
|
|
%acc30 = fsub double %l30, %acc29
|
|
%acc31 = fsub double %l31, %acc30
|
|
store volatile double %acc0, double *%ptr
|
|
store volatile double %acc1, double *%ptr
|
|
store volatile double %acc2, double *%ptr
|
|
store volatile double %acc3, double *%ptr
|
|
store volatile double %acc4, double *%ptr
|
|
store volatile double %acc5, double *%ptr
|
|
store volatile double %acc6, double *%ptr
|
|
store volatile double %acc7, double *%ptr
|
|
store volatile double %acc8, double *%ptr
|
|
store volatile double %acc16, double *%ptr
|
|
store volatile double %acc17, double *%ptr
|
|
store volatile double %acc18, double *%ptr
|
|
store volatile double %acc19, double *%ptr
|
|
store volatile double %acc20, double *%ptr
|
|
store volatile double %acc21, double *%ptr
|
|
store volatile double %acc22, double *%ptr
|
|
store volatile double %acc23, double *%ptr
|
|
store volatile double %acc24, double *%ptr
|
|
store volatile double %acc25, double *%ptr
|
|
store volatile double %acc26, double *%ptr
|
|
store volatile double %acc27, double *%ptr
|
|
store volatile double %acc28, double *%ptr
|
|
store volatile double %acc29, double *%ptr
|
|
store volatile double %acc30, double *%ptr
|
|
store volatile double %acc31, double *%ptr
|
|
ret void
|
|
}
|
|
|
|
; This function should use all call-clobbered FPRs and vector registers
|
|
; but no call-saved ones. It shouldn't need to create a frame.
|
|
define void @f4(double *%ptr) {
|
|
; CHECK-LABEL: f4:
|
|
; CHECK-NOT: %r15
|
|
; CHECK-NOT: {{%[fv][89]}}
|
|
; CHECK-NOT: {{%[fv]1[0-5]}}
|
|
; CHECK: br %r14
|
|
%l0 = load volatile double, double *%ptr
|
|
%l1 = load volatile double, double *%ptr
|
|
%l2 = load volatile double, double *%ptr
|
|
%l3 = load volatile double, double *%ptr
|
|
%l4 = load volatile double, double *%ptr
|
|
%l5 = load volatile double, double *%ptr
|
|
%l6 = load volatile double, double *%ptr
|
|
%l7 = load volatile double, double *%ptr
|
|
%l16 = load volatile double, double *%ptr
|
|
%l17 = load volatile double, double *%ptr
|
|
%l18 = load volatile double, double *%ptr
|
|
%l19 = load volatile double, double *%ptr
|
|
%l20 = load volatile double, double *%ptr
|
|
%l21 = load volatile double, double *%ptr
|
|
%l22 = load volatile double, double *%ptr
|
|
%l23 = load volatile double, double *%ptr
|
|
%l24 = load volatile double, double *%ptr
|
|
%l25 = load volatile double, double *%ptr
|
|
%l26 = load volatile double, double *%ptr
|
|
%l27 = load volatile double, double *%ptr
|
|
%l28 = load volatile double, double *%ptr
|
|
%l29 = load volatile double, double *%ptr
|
|
%l30 = load volatile double, double *%ptr
|
|
%l31 = load volatile double, double *%ptr
|
|
%acc0 = fsub double %l0, %l0
|
|
%acc1 = fsub double %l1, %acc0
|
|
%acc2 = fsub double %l2, %acc1
|
|
%acc3 = fsub double %l3, %acc2
|
|
%acc4 = fsub double %l4, %acc3
|
|
%acc5 = fsub double %l5, %acc4
|
|
%acc6 = fsub double %l6, %acc5
|
|
%acc7 = fsub double %l7, %acc6
|
|
%acc16 = fsub double %l16, %acc7
|
|
%acc17 = fsub double %l17, %acc16
|
|
%acc18 = fsub double %l18, %acc17
|
|
%acc19 = fsub double %l19, %acc18
|
|
%acc20 = fsub double %l20, %acc19
|
|
%acc21 = fsub double %l21, %acc20
|
|
%acc22 = fsub double %l22, %acc21
|
|
%acc23 = fsub double %l23, %acc22
|
|
%acc24 = fsub double %l24, %acc23
|
|
%acc25 = fsub double %l25, %acc24
|
|
%acc26 = fsub double %l26, %acc25
|
|
%acc27 = fsub double %l27, %acc26
|
|
%acc28 = fsub double %l28, %acc27
|
|
%acc29 = fsub double %l29, %acc28
|
|
%acc30 = fsub double %l30, %acc29
|
|
%acc31 = fsub double %l31, %acc30
|
|
store volatile double %acc0, double *%ptr
|
|
store volatile double %acc1, double *%ptr
|
|
store volatile double %acc2, double *%ptr
|
|
store volatile double %acc3, double *%ptr
|
|
store volatile double %acc4, double *%ptr
|
|
store volatile double %acc5, double *%ptr
|
|
store volatile double %acc6, double *%ptr
|
|
store volatile double %acc7, double *%ptr
|
|
store volatile double %acc16, double *%ptr
|
|
store volatile double %acc17, double *%ptr
|
|
store volatile double %acc18, double *%ptr
|
|
store volatile double %acc19, double *%ptr
|
|
store volatile double %acc20, double *%ptr
|
|
store volatile double %acc21, double *%ptr
|
|
store volatile double %acc22, double *%ptr
|
|
store volatile double %acc23, double *%ptr
|
|
store volatile double %acc24, double *%ptr
|
|
store volatile double %acc25, double *%ptr
|
|
store volatile double %acc26, double *%ptr
|
|
store volatile double %acc27, double *%ptr
|
|
store volatile double %acc28, double *%ptr
|
|
store volatile double %acc29, double *%ptr
|
|
store volatile double %acc30, double *%ptr
|
|
store volatile double %acc31, double *%ptr
|
|
ret void
|
|
}
|