mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-23 19:23:23 +01:00
3a9a1ac61b
For some reason there are both of these available, except for scalar 64-bit compares which only has u64. I'm not sure why there are both (I'm guessing it's for the one bit inputs we don't use), but for consistency always using the unsigned one. llvm-svn: 282832
186 lines
4.8 KiB
LLVM
186 lines
4.8 KiB
LLVM
; RUN: llc -march=amdgcn -verify-machineinstrs < %s | FileCheck %s
|
|
; RUN: llc -march=amdgcn -mcpu=tonga -verify-machineinstrs < %s | FileCheck %s
|
|
|
|
; CHECK-LABEL: {{^}}inline_asm:
|
|
; CHECK: s_endpgm
|
|
; CHECK: s_endpgm
|
|
define void @inline_asm(i32 addrspace(1)* %out) {
|
|
entry:
|
|
store i32 5, i32 addrspace(1)* %out
|
|
call void asm sideeffect "s_endpgm", ""()
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}inline_asm_shader:
|
|
; CHECK: s_endpgm
|
|
; CHECK: s_endpgm
|
|
define amdgpu_ps void @inline_asm_shader() {
|
|
entry:
|
|
call void asm sideeffect "s_endpgm", ""()
|
|
ret void
|
|
}
|
|
|
|
|
|
; CHECK: {{^}}branch_on_asm:
|
|
; Make sure inline assembly is treted as divergent.
|
|
; CHECK: s_mov_b32 s{{[0-9]+}}, 0
|
|
; CHECK: s_and_saveexec_b64
|
|
define void @branch_on_asm(i32 addrspace(1)* %out) {
|
|
%zero = call i32 asm "s_mov_b32 $0, 0", "=s"()
|
|
%cmp = icmp eq i32 %zero, 0
|
|
br i1 %cmp, label %if, label %endif
|
|
|
|
if:
|
|
store i32 0, i32 addrspace(1)* %out
|
|
br label %endif
|
|
|
|
endif:
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}v_cmp_asm:
|
|
; CHECK: v_mov_b32_e32 [[SRC:v[0-9]+]], s{{[0-9]+}}
|
|
; CHECK: v_cmp_ne_u32_e64 s{{\[}}[[MASK_LO:[0-9]+]]:[[MASK_HI:[0-9]+]]{{\]}}, 0, [[SRC]]
|
|
; CHECK-DAG: v_mov_b32_e32 v[[V_LO:[0-9]+]], s[[MASK_LO]]
|
|
; CHECK-DAG: v_mov_b32_e32 v[[V_HI:[0-9]+]], s[[MASK_HI]]
|
|
; CHECK: buffer_store_dwordx2 v{{\[}}[[V_LO]]:[[V_HI]]{{\]}}
|
|
define void @v_cmp_asm(i64 addrspace(1)* %out, i32 %in) {
|
|
%sgpr = tail call i64 asm "v_cmp_ne_u32_e64 $0, 0, $1", "=s,v"(i32 %in)
|
|
store i64 %sgpr, i64 addrspace(1)* %out
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm:
|
|
; CHECK: codeLenInByte = 12
|
|
define void @code_size_inline_asm(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "v_nop_e64", ""()
|
|
ret void
|
|
}
|
|
|
|
; All inlineasm instructions are assumed to be the maximum size
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_small_inst:
|
|
; CHECK: codeLenInByte = 12
|
|
define void @code_size_inline_asm_small_inst(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "v_nop_e32", ""()
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_2_inst:
|
|
; CHECK: codeLenInByte = 20
|
|
define void @code_size_inline_asm_2_inst(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "
|
|
v_nop_e64
|
|
v_nop_e64
|
|
", ""()
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_2_inst_extra_newline:
|
|
; CHECK: codeLenInByte = 20
|
|
define void @code_size_inline_asm_2_inst_extra_newline(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "
|
|
v_nop_e64
|
|
|
|
v_nop_e64
|
|
", ""()
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_0_inst:
|
|
; CHECK: codeLenInByte = 4
|
|
define void @code_size_inline_asm_0_inst(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "", ""()
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_1_comment:
|
|
; CHECK: codeLenInByte = 4
|
|
define void @code_size_inline_asm_1_comment(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "; comment", ""()
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_newline_1_comment:
|
|
; CHECK: codeLenInByte = 4
|
|
define void @code_size_inline_asm_newline_1_comment(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "
|
|
; comment", ""()
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_1_comment_newline:
|
|
; CHECK: codeLenInByte = 4
|
|
define void @code_size_inline_asm_1_comment_newline(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "; comment
|
|
", ""()
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_2_comments_line:
|
|
; CHECK: codeLenInByte = 4
|
|
define void @code_size_inline_asm_2_comments_line(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "; first comment ; second comment", ""()
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_2_comments_line_nospace:
|
|
; CHECK: codeLenInByte = 4
|
|
define void @code_size_inline_asm_2_comments_line_nospace(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "; first comment;second comment", ""()
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_mixed_comments0:
|
|
; CHECK: codeLenInByte = 20
|
|
define void @code_size_inline_asm_mixed_comments0(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "; comment
|
|
v_nop_e64 ; inline comment
|
|
; separate comment
|
|
v_nop_e64
|
|
|
|
; trailing comment
|
|
; extra comment
|
|
", ""()
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_mixed_comments1:
|
|
; CHECK: codeLenInByte = 20
|
|
define void @code_size_inline_asm_mixed_comments1(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "v_nop_e64 ; inline comment
|
|
; separate comment
|
|
v_nop_e64
|
|
|
|
; trailing comment
|
|
; extra comment
|
|
", ""()
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: {{^}}code_size_inline_asm_mixed_comments_operands:
|
|
; CHECK: codeLenInByte = 20
|
|
define void @code_size_inline_asm_mixed_comments_operands(i32 addrspace(1)* %out) {
|
|
entry:
|
|
call void asm sideeffect "; comment
|
|
v_add_i32_e32 v0, vcc, v1, v2 ; inline comment
|
|
; separate comment
|
|
v_bfrev_b32_e32 v0, 1
|
|
|
|
; trailing comment
|
|
; extra comment
|
|
", ""()
|
|
ret void
|
|
}
|