1
0
mirror of https://github.com/RPCS3/llvm-mirror.git synced 2024-11-25 04:02:41 +01:00
llvm-mirror/lib/Target/VE/VECallingConv.td
Kazushi (Jam) Marukawa 920c28fbd1 [VE] Correct VMP allocation in calling conv
VE used to allocate VM1, VM2, VMP2 (VM4+VM5), and VM3.  This patch
corrects to allocate VM1, VM2, VMP2 (VM4+VM5), and VM6.  Also add
a regression test.

Reviewed By: simoll

Differential Revision: https://reviews.llvm.org/D93570
2020-12-21 22:42:24 +09:00

154 lines
5.3 KiB
TableGen

//===-- VECallingConv.td - Calling Conventions VE ----------*- tablegen -*-===//
//
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
// See https://llvm.org/LICENSE.txt for license information.
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
//
//===----------------------------------------------------------------------===//
//
// This describes the calling conventions for the VE architectures.
//
//===----------------------------------------------------------------------===//
//===----------------------------------------------------------------------===//
// Aurora VE
//===----------------------------------------------------------------------===//
def CC_VE_C_Stack: CallingConv<[
// F128 are assigned to the stack in 16-byte aligned units
CCIfType<[f128], CCAssignToStackWithShadow<16, 16, [SX7]>>,
// All of the rest are assigned to the stack in 8-byte aligned units.
CCAssignToStack<0, 8>
]>;
///// C Calling Convention (VE ABI v2.1) /////
//
// Reference: https://www.nec.com/en/global/prod/hpc/aurora/document/VE-ABI_v2.1.pdf
//
def CC_VE_C : CallingConv<[
// All arguments get passed in generic registers if there is space.
// Promote i1/i8/i16/i32 arguments to i64.
CCIfType<[i1, i8, i16, i32], CCPromoteToType<i64>>,
// Convert float arguments to i64 with padding.
// 63 31 0
// +------+------+
// | float| 0 |
// +------+------+
CCIfType<[f32], CCBitConvertToType<i64>>,
// bool, char, int, enum, long, long long, float, double
// --> generic 64 bit registers
CCIfType<[i64, f64],
CCAssignToReg<[SX0, SX1, SX2, SX3, SX4, SX5, SX6, SX7]>>,
// long double --> pair of generic 64 bit registers
//
// NOTE: If Q1 is allocated while SX1 is free, llvm tries to allocate SX1 for
// following operands, this masks SX1 to avoid such behavior.
CCIfType<[f128],
CCAssignToRegWithShadow<[Q0, Q1, Q2, Q3],
[SX0, SX1, SX3, SX5]>>,
// Alternatively, they are assigned to the stack in 8-byte aligned units.
CCDelegateTo<CC_VE_C_Stack>
]>;
///// Standard vararg C Calling Convention (VE ABI v2.1) /////
// All arguments get passed in stack for varargs function or non-prototyped
// function.
def CC_VE2 : CallingConv<[
// Promote i1/i8/i16/i32 arguments to i64.
CCIfType<[i1, i8, i16, i32], CCPromoteToType<i64>>,
// Convert float arguments to i64 with padding.
// 63 31 0
// +------+------+
// | float| 0 |
// +------+------+
CCIfType<[f32], CCBitConvertToType<i64>>,
// F128 are assigned to the stack in 16-byte aligned units
CCIfType<[f128], CCAssignToStack<16, 16>>,
CCAssignToStack<0, 8>
]>;
def RetCC_VE_C : CallingConv<[
// Promote i1/i8/i16/i32 return values to i64.
CCIfType<[i1, i8, i16, i32], CCPromoteToType<i64>>,
// Convert float return values to i64 with padding.
// 63 31 0
// +------+------+
// | float| 0 |
// +------+------+
CCIfType<[f32], CCBitConvertToType<i64>>,
// bool, char, int, enum, long, long long, float, double
// --> generic 64 bit registers
CCIfType<[i64, f64],
CCAssignToReg<[SX0, SX1, SX2, SX3, SX4, SX5, SX6, SX7]>>,
// long double --> pair of generic 64 bit registers
CCIfType<[f128],
CCAssignToRegWithShadow<[Q0, Q1, Q2, Q3],
[SX0, SX1, SX3, SX5]>>,
]>;
///// Custom fastcc /////
//
// This passes vector params and return values in registers. Scalar values are
// handled conforming to the standard cc.
def CC_VE_Fast : CallingConv<[
// vector --> generic vector registers
CCIfType<[v256i32, v256f32, v256i64, v256f64],
CCAssignToReg<[V0, V1, V2, V3, V4, V5, V6, V7]>>,
// TODO: make this conditional on packed mode
CCIfType<[v512i32, v512f32],
CCAssignToReg<[V0, V1, V2, V3, V4, V5, V6, V7]>>,
// vector mask --> generic vector mask registers
CCIfType<[v256i1],
CCAssignToReg<[VM1, VM2, VM3, VM4, VM5, VM6, VM7]>>,
// pair of vector mask --> generic vector mask registers
CCIfType<[v512i1],
CCAssignToRegWithShadow<[VMP1, VMP2, VMP3],
[VM1, VM3, VM5]>>,
// Follow the standard C CC for scalars.
CCDelegateTo<CC_VE_C>
]>;
def RetCC_VE_Fast : CallingConv<[
// vector --> generic vector registers
CCIfType<[v256i32, v256f32, v256i64, v256f64],
CCAssignToReg<[V0, V1, V2, V3, V4, V5, V6, V7]>>,
// TODO: make this conditional on packed mode
CCIfType<[v512i32, v512f32],
CCAssignToReg<[V0, V1, V2, V3, V4, V5, V6, V7]>>,
// vector mask --> generic vector mask registers
CCIfType<[v256i1],
CCAssignToReg<[VM1, VM2, VM3, VM4, VM5, VM6, VM7]>>,
// pair of vector mask --> generic vector mask registers
CCIfType<[v512i1],
CCAssignToRegWithShadow<[VMP1, VMP2, VMP3],
[VM1, VM3, VM5]>>,
// Follow the standard C CC for scalars.
CCDelegateTo<RetCC_VE_C>
]>;
// Callee-saved registers
def CSR : CalleeSavedRegs<(add (sequence "SX%u", 18, 33))>;
def CSR_NoRegs : CalleeSavedRegs<(add)>;
// PreserveAll (clobbers s62,s63) - used for ve_grow_stack
def CSR_preserve_all : CalleeSavedRegs<(add (sequence "SX%u", 0, 61),
(sequence "V%u", 0, 63),
(sequence "VM%u", 1, 15))>;