mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-26 12:43:36 +01:00
b6b7acf7ed
This is the conservatively correct default. It is always safe to assume xnack is enabled, but not the converse. Introduce a feature to blacklist targets where xnack can never be meaningfully enabled. I'm not sure the targets this is applied to is 100% correct. llvm-svn: 360903
79 lines
3.3 KiB
LLVM
79 lines
3.3 KiB
LLVM
; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=tahiti -verify-machineinstrs -stress-regalloc=6 < %s | FileCheck %s
|
|
|
|
; Inline spiller can decide to move a spill as early as possible in the basic block.
|
|
; It will skip phis and label, but we also need to make sure it skips instructions
|
|
; in the basic block prologue which restore exec mask.
|
|
; Make sure instruction to restore exec mask immediately follows label
|
|
|
|
; CHECK-LABEL: {{^}}spill_cfg_position:
|
|
; CHECK: s_cbranch_execz [[LABEL1:BB[0-9_]+]]
|
|
; CHECK: {{^}}[[LABEL1]]:
|
|
; CHECK: s_cbranch_execz [[LABEL2:BB[0-9_]+]]
|
|
; CHECK: {{^}}[[LABEL2]]:
|
|
; CHECK-NEXT: s_or_b64 exec
|
|
; CHECK: buffer_
|
|
|
|
define amdgpu_kernel void @spill_cfg_position(i32 addrspace(1)* nocapture %arg) {
|
|
bb:
|
|
%tmp1 = tail call i32 @llvm.amdgcn.workitem.id.x() #0
|
|
%tmp14 = load i32, i32 addrspace(1)* %arg, align 4
|
|
%tmp15 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 1
|
|
%tmp16 = load i32, i32 addrspace(1)* %tmp15, align 4
|
|
%tmp17 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 2
|
|
%tmp18 = load i32, i32 addrspace(1)* %tmp17, align 4
|
|
%tmp19 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 3
|
|
%tmp20 = load i32, i32 addrspace(1)* %tmp19, align 4
|
|
%tmp21 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 4
|
|
%tmp22 = load i32, i32 addrspace(1)* %tmp21, align 4
|
|
%tmp23 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 5
|
|
%tmp24 = load i32, i32 addrspace(1)* %tmp23, align 4
|
|
%tmp25 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 6
|
|
%tmp26 = load i32, i32 addrspace(1)* %tmp25, align 4
|
|
%tmp27 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 7
|
|
%tmp28 = load i32, i32 addrspace(1)* %tmp27, align 4
|
|
%tmp29 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 8
|
|
%tmp30 = load i32, i32 addrspace(1)* %tmp29, align 4
|
|
%tmp33 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i32 %tmp1
|
|
%tmp34 = load i32, i32 addrspace(1)* %tmp33, align 4
|
|
%tmp35 = icmp eq i32 %tmp34, 0
|
|
br i1 %tmp35, label %bb44, label %bb36
|
|
|
|
bb36: ; preds = %bb
|
|
%tmp37 = mul nsw i32 %tmp20, %tmp18
|
|
%tmp38 = add nsw i32 %tmp37, %tmp16
|
|
%tmp39 = mul nsw i32 %tmp24, %tmp22
|
|
%tmp40 = add nsw i32 %tmp38, %tmp39
|
|
%tmp41 = mul nsw i32 %tmp28, %tmp26
|
|
%tmp42 = add nsw i32 %tmp40, %tmp41
|
|
%tmp43 = add nsw i32 %tmp42, %tmp30
|
|
br label %bb52
|
|
|
|
bb44: ; preds = %bb
|
|
%tmp45 = mul nsw i32 %tmp18, %tmp16
|
|
%tmp46 = mul nsw i32 %tmp22, %tmp20
|
|
%tmp47 = add nsw i32 %tmp46, %tmp45
|
|
%tmp48 = mul nsw i32 %tmp26, %tmp24
|
|
%tmp49 = add nsw i32 %tmp47, %tmp48
|
|
%tmp50 = mul nsw i32 %tmp30, %tmp28
|
|
%tmp51 = add nsw i32 %tmp49, %tmp50
|
|
br label %bb52
|
|
|
|
bb52: ; preds = %bb44, %bb36
|
|
%tmp53 = phi i32 [ %tmp43, %bb36 ], [ %tmp51, %bb44 ]
|
|
%tmp54 = mul nsw i32 %tmp16, %tmp14
|
|
%tmp55 = mul nsw i32 %tmp22, %tmp18
|
|
%tmp56 = mul nsw i32 %tmp24, %tmp20
|
|
%tmp57 = mul nsw i32 %tmp30, %tmp26
|
|
%tmp58 = add i32 %tmp55, %tmp54
|
|
%tmp59 = add i32 %tmp58, %tmp56
|
|
%tmp60 = add i32 %tmp59, %tmp28
|
|
%tmp61 = add i32 %tmp60, %tmp57
|
|
%tmp62 = add i32 %tmp61, %tmp53
|
|
store i32 %tmp62, i32 addrspace(1)* %tmp33, align 4
|
|
ret void
|
|
}
|
|
|
|
declare i32 @llvm.amdgcn.workitem.id.x() #0
|
|
|
|
attributes #0 = { nounwind readnone }
|