mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-23 11:13:28 +01:00
d654e7d40c
Enable enableMultipleCopyHints() on X86. Original Patch by @jonpa: While enabling the mischeduler for SystemZ, it was discovered that for some reason a test needed one extra seemingly needless COPY (test/CodeGen/SystemZ/call-03.ll). The handling for that is resulted in this patch, which improves the register coalescing by providing not just one copy hint, but a sorted list of copy hints. On SystemZ, this gives ~12500 less register moves on SPEC, as well as marginally less spilling. Instead of improving just the SystemZ backend, the improvement has been implemented in common-code (calculateSpillWeightAndHint(). This gives a lot of test failures, but since this should be a general improvement I hope that the involved targets will help and review the test updates. Differential Revision: https://reviews.llvm.org/D38128 llvm-svn: 342578
340 lines
10 KiB
LLVM
340 lines
10 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc < %s -mtriple=x86_64-- | FileCheck %s --check-prefix=CHECK --check-prefix=FASTINCDEC
|
|
; RUN: llc < %s -mtriple=x86_64-- -mattr=slow-incdec | FileCheck %s --check-prefix=CHECK --check-prefix=SLOWINCDEC
|
|
|
|
define i32 @test_add_1_cmov_slt(i64* %p, i32 %a0, i32 %a1) #0 {
|
|
; FASTINCDEC-LABEL: test_add_1_cmov_slt:
|
|
; FASTINCDEC: # %bb.0: # %entry
|
|
; FASTINCDEC-NEXT: movl %esi, %eax
|
|
; FASTINCDEC-NEXT: lock incq (%rdi)
|
|
; FASTINCDEC-NEXT: cmovgl %edx, %eax
|
|
; FASTINCDEC-NEXT: retq
|
|
;
|
|
; SLOWINCDEC-LABEL: test_add_1_cmov_slt:
|
|
; SLOWINCDEC: # %bb.0: # %entry
|
|
; SLOWINCDEC-NEXT: movl %esi, %eax
|
|
; SLOWINCDEC-NEXT: lock addq $1, (%rdi)
|
|
; SLOWINCDEC-NEXT: cmovgl %edx, %eax
|
|
; SLOWINCDEC-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw add i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp slt i64 %tmp0, 0
|
|
%tmp2 = select i1 %tmp1, i32 %a0, i32 %a1
|
|
ret i32 %tmp2
|
|
}
|
|
|
|
define i32 @test_add_1_cmov_sge(i64* %p, i32 %a0, i32 %a1) #0 {
|
|
; FASTINCDEC-LABEL: test_add_1_cmov_sge:
|
|
; FASTINCDEC: # %bb.0: # %entry
|
|
; FASTINCDEC-NEXT: movl %esi, %eax
|
|
; FASTINCDEC-NEXT: lock incq (%rdi)
|
|
; FASTINCDEC-NEXT: cmovlel %edx, %eax
|
|
; FASTINCDEC-NEXT: retq
|
|
;
|
|
; SLOWINCDEC-LABEL: test_add_1_cmov_sge:
|
|
; SLOWINCDEC: # %bb.0: # %entry
|
|
; SLOWINCDEC-NEXT: movl %esi, %eax
|
|
; SLOWINCDEC-NEXT: lock addq $1, (%rdi)
|
|
; SLOWINCDEC-NEXT: cmovlel %edx, %eax
|
|
; SLOWINCDEC-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw add i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp sge i64 %tmp0, 0
|
|
%tmp2 = select i1 %tmp1, i32 %a0, i32 %a1
|
|
ret i32 %tmp2
|
|
}
|
|
|
|
define i32 @test_sub_1_cmov_sle(i64* %p, i32 %a0, i32 %a1) #0 {
|
|
; FASTINCDEC-LABEL: test_sub_1_cmov_sle:
|
|
; FASTINCDEC: # %bb.0: # %entry
|
|
; FASTINCDEC-NEXT: movl %esi, %eax
|
|
; FASTINCDEC-NEXT: lock decq (%rdi)
|
|
; FASTINCDEC-NEXT: cmovgel %edx, %eax
|
|
; FASTINCDEC-NEXT: retq
|
|
;
|
|
; SLOWINCDEC-LABEL: test_sub_1_cmov_sle:
|
|
; SLOWINCDEC: # %bb.0: # %entry
|
|
; SLOWINCDEC-NEXT: movl %esi, %eax
|
|
; SLOWINCDEC-NEXT: lock addq $-1, (%rdi)
|
|
; SLOWINCDEC-NEXT: cmovgel %edx, %eax
|
|
; SLOWINCDEC-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw sub i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp sle i64 %tmp0, 0
|
|
%tmp2 = select i1 %tmp1, i32 %a0, i32 %a1
|
|
ret i32 %tmp2
|
|
}
|
|
|
|
define i32 @test_sub_1_cmov_sgt(i64* %p, i32 %a0, i32 %a1) #0 {
|
|
; FASTINCDEC-LABEL: test_sub_1_cmov_sgt:
|
|
; FASTINCDEC: # %bb.0: # %entry
|
|
; FASTINCDEC-NEXT: movl %esi, %eax
|
|
; FASTINCDEC-NEXT: lock decq (%rdi)
|
|
; FASTINCDEC-NEXT: cmovll %edx, %eax
|
|
; FASTINCDEC-NEXT: retq
|
|
;
|
|
; SLOWINCDEC-LABEL: test_sub_1_cmov_sgt:
|
|
; SLOWINCDEC: # %bb.0: # %entry
|
|
; SLOWINCDEC-NEXT: movl %esi, %eax
|
|
; SLOWINCDEC-NEXT: lock addq $-1, (%rdi)
|
|
; SLOWINCDEC-NEXT: cmovll %edx, %eax
|
|
; SLOWINCDEC-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw sub i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp sgt i64 %tmp0, 0
|
|
%tmp2 = select i1 %tmp1, i32 %a0, i32 %a1
|
|
ret i32 %tmp2
|
|
}
|
|
|
|
; FIXME: (setcc slt x, 0) gets combined into shr early.
|
|
define i8 @test_add_1_setcc_slt(i64* %p) #0 {
|
|
; CHECK-LABEL: test_add_1_setcc_slt:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movl $1, %eax
|
|
; CHECK-NEXT: lock xaddq %rax, (%rdi)
|
|
; CHECK-NEXT: shrq $63, %rax
|
|
; CHECK-NEXT: # kill: def $al killed $al killed $rax
|
|
; CHECK-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw add i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp slt i64 %tmp0, 0
|
|
%tmp2 = zext i1 %tmp1 to i8
|
|
ret i8 %tmp2
|
|
}
|
|
|
|
define i8 @test_sub_1_setcc_sgt(i64* %p) #0 {
|
|
; FASTINCDEC-LABEL: test_sub_1_setcc_sgt:
|
|
; FASTINCDEC: # %bb.0: # %entry
|
|
; FASTINCDEC-NEXT: lock decq (%rdi)
|
|
; FASTINCDEC-NEXT: setge %al
|
|
; FASTINCDEC-NEXT: retq
|
|
;
|
|
; SLOWINCDEC-LABEL: test_sub_1_setcc_sgt:
|
|
; SLOWINCDEC: # %bb.0: # %entry
|
|
; SLOWINCDEC-NEXT: lock addq $-1, (%rdi)
|
|
; SLOWINCDEC-NEXT: setge %al
|
|
; SLOWINCDEC-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw sub i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp sgt i64 %tmp0, 0
|
|
%tmp2 = zext i1 %tmp1 to i8
|
|
ret i8 %tmp2
|
|
}
|
|
|
|
define i32 @test_add_1_brcond_sge(i64* %p, i32 %a0, i32 %a1) #0 {
|
|
; FASTINCDEC-LABEL: test_add_1_brcond_sge:
|
|
; FASTINCDEC: # %bb.0: # %entry
|
|
; FASTINCDEC-NEXT: lock incq (%rdi)
|
|
; FASTINCDEC-NEXT: jle .LBB6_2
|
|
; FASTINCDEC-NEXT: # %bb.1: # %t
|
|
; FASTINCDEC-NEXT: movl %esi, %eax
|
|
; FASTINCDEC-NEXT: retq
|
|
; FASTINCDEC-NEXT: .LBB6_2: # %f
|
|
; FASTINCDEC-NEXT: movl %edx, %eax
|
|
; FASTINCDEC-NEXT: retq
|
|
;
|
|
; SLOWINCDEC-LABEL: test_add_1_brcond_sge:
|
|
; SLOWINCDEC: # %bb.0: # %entry
|
|
; SLOWINCDEC-NEXT: lock addq $1, (%rdi)
|
|
; SLOWINCDEC-NEXT: jle .LBB6_2
|
|
; SLOWINCDEC-NEXT: # %bb.1: # %t
|
|
; SLOWINCDEC-NEXT: movl %esi, %eax
|
|
; SLOWINCDEC-NEXT: retq
|
|
; SLOWINCDEC-NEXT: .LBB6_2: # %f
|
|
; SLOWINCDEC-NEXT: movl %edx, %eax
|
|
; SLOWINCDEC-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw add i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp sge i64 %tmp0, 0
|
|
br i1 %tmp1, label %t, label %f
|
|
t:
|
|
ret i32 %a0
|
|
f:
|
|
ret i32 %a1
|
|
}
|
|
|
|
; Also make sure we don't muck with condition codes that we should ignore.
|
|
; No need to test unsigned comparisons, as they should all be simplified.
|
|
|
|
define i32 @test_add_1_cmov_sle(i64* %p, i32 %a0, i32 %a1) #0 {
|
|
; CHECK-LABEL: test_add_1_cmov_sle:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movl %esi, %eax
|
|
; CHECK-NEXT: movl $1, %ecx
|
|
; CHECK-NEXT: lock xaddq %rcx, (%rdi)
|
|
; CHECK-NEXT: testq %rcx, %rcx
|
|
; CHECK-NEXT: cmovgl %edx, %eax
|
|
; CHECK-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw add i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp sle i64 %tmp0, 0
|
|
%tmp2 = select i1 %tmp1, i32 %a0, i32 %a1
|
|
ret i32 %tmp2
|
|
}
|
|
|
|
define i32 @test_add_1_cmov_sgt(i64* %p, i32 %a0, i32 %a1) #0 {
|
|
; CHECK-LABEL: test_add_1_cmov_sgt:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movl %esi, %eax
|
|
; CHECK-NEXT: movl $1, %ecx
|
|
; CHECK-NEXT: lock xaddq %rcx, (%rdi)
|
|
; CHECK-NEXT: testq %rcx, %rcx
|
|
; CHECK-NEXT: cmovlel %edx, %eax
|
|
; CHECK-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw add i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp sgt i64 %tmp0, 0
|
|
%tmp2 = select i1 %tmp1, i32 %a0, i32 %a1
|
|
ret i32 %tmp2
|
|
}
|
|
|
|
; Test a result being used by more than just the comparison.
|
|
|
|
define i8 @test_add_1_setcc_sgt_reuse(i64* %p, i64* %p2) #0 {
|
|
; CHECK-LABEL: test_add_1_setcc_sgt_reuse:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movl $1, %ecx
|
|
; CHECK-NEXT: lock xaddq %rcx, (%rdi)
|
|
; CHECK-NEXT: testq %rcx, %rcx
|
|
; CHECK-NEXT: setg %al
|
|
; CHECK-NEXT: movq %rcx, (%rsi)
|
|
; CHECK-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw add i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp sgt i64 %tmp0, 0
|
|
%tmp2 = zext i1 %tmp1 to i8
|
|
store i64 %tmp0, i64* %p2
|
|
ret i8 %tmp2
|
|
}
|
|
|
|
define i8 @test_sub_2_setcc_sgt(i64* %p) #0 {
|
|
; CHECK-LABEL: test_sub_2_setcc_sgt:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movq $-2, %rax
|
|
; CHECK-NEXT: lock xaddq %rax, (%rdi)
|
|
; CHECK-NEXT: testq %rax, %rax
|
|
; CHECK-NEXT: setg %al
|
|
; CHECK-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw sub i64* %p, i64 2 seq_cst
|
|
%tmp1 = icmp sgt i64 %tmp0, 0
|
|
%tmp2 = zext i1 %tmp1 to i8
|
|
ret i8 %tmp2
|
|
}
|
|
|
|
define i8 @test_add_1_cmov_cmov(i64* %p, i8* %q) #0 {
|
|
; TODO: It's possible to use "lock inc" here, but both cmovs need to be updated.
|
|
; CHECK-LABEL: test_add_1_cmov_cmov:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movl $1, %eax
|
|
; CHECK-NEXT: lock xaddq %rax, (%rdi)
|
|
; CHECK-NEXT: testq %rax, %rax
|
|
entry:
|
|
%add = atomicrmw add i64* %p, i64 1 seq_cst
|
|
%cmp = icmp slt i64 %add, 0
|
|
%s1 = select i1 %cmp, i8 12, i8 34
|
|
store i8 %s1, i8* %q
|
|
%s2 = select i1 %cmp, i8 56, i8 78
|
|
ret i8 %s2
|
|
}
|
|
|
|
define i8 @test_sub_1_cmp_1_setcc_eq(i64* %p) #0 {
|
|
; FASTINCDEC-LABEL: test_sub_1_cmp_1_setcc_eq:
|
|
; FASTINCDEC: # %bb.0: # %entry
|
|
; FASTINCDEC-NEXT: lock decq (%rdi)
|
|
; FASTINCDEC-NEXT: sete %al
|
|
; FASTINCDEC-NEXT: retq
|
|
;
|
|
; SLOWINCDEC-LABEL: test_sub_1_cmp_1_setcc_eq:
|
|
; SLOWINCDEC: # %bb.0: # %entry
|
|
; SLOWINCDEC-NEXT: lock subq $1, (%rdi)
|
|
; SLOWINCDEC-NEXT: sete %al
|
|
; SLOWINCDEC-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw sub i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp eq i64 %tmp0, 1
|
|
%tmp2 = zext i1 %tmp1 to i8
|
|
ret i8 %tmp2
|
|
}
|
|
|
|
define i8 @test_sub_1_cmp_1_setcc_ne(i64* %p) #0 {
|
|
; FASTINCDEC-LABEL: test_sub_1_cmp_1_setcc_ne:
|
|
; FASTINCDEC: # %bb.0: # %entry
|
|
; FASTINCDEC-NEXT: lock decq (%rdi)
|
|
; FASTINCDEC-NEXT: setne %al
|
|
; FASTINCDEC-NEXT: retq
|
|
;
|
|
; SLOWINCDEC-LABEL: test_sub_1_cmp_1_setcc_ne:
|
|
; SLOWINCDEC: # %bb.0: # %entry
|
|
; SLOWINCDEC-NEXT: lock subq $1, (%rdi)
|
|
; SLOWINCDEC-NEXT: setne %al
|
|
; SLOWINCDEC-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw sub i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp ne i64 %tmp0, 1
|
|
%tmp2 = zext i1 %tmp1 to i8
|
|
ret i8 %tmp2
|
|
}
|
|
|
|
define i8 @test_sub_1_cmp_1_setcc_ugt(i64* %p) #0 {
|
|
; CHECK-LABEL: test_sub_1_cmp_1_setcc_ugt:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: lock subq $1, (%rdi)
|
|
; CHECK-NEXT: seta %al
|
|
; CHECK-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw sub i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp ugt i64 %tmp0, 1
|
|
%tmp2 = zext i1 %tmp1 to i8
|
|
ret i8 %tmp2
|
|
}
|
|
|
|
; FIXME: This test canonicalizes in a way that hides the fact that the
|
|
; comparison can be folded into the atomic subtract.
|
|
define i8 @test_sub_1_cmp_1_setcc_sle(i64* %p) #0 {
|
|
; CHECK-LABEL: test_sub_1_cmp_1_setcc_sle:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movq $-1, %rax
|
|
; CHECK-NEXT: lock xaddq %rax, (%rdi)
|
|
; CHECK-NEXT: cmpq $2, %rax
|
|
; CHECK-NEXT: setl %al
|
|
; CHECK-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw sub i64* %p, i64 1 seq_cst
|
|
%tmp1 = icmp sle i64 %tmp0, 1
|
|
%tmp2 = zext i1 %tmp1 to i8
|
|
ret i8 %tmp2
|
|
}
|
|
|
|
define i8 @test_sub_3_cmp_3_setcc_eq(i64* %p) #0 {
|
|
; CHECK-LABEL: test_sub_3_cmp_3_setcc_eq:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: lock subq $3, (%rdi)
|
|
; CHECK-NEXT: sete %al
|
|
; CHECK-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw sub i64* %p, i64 3 seq_cst
|
|
%tmp1 = icmp eq i64 %tmp0, 3
|
|
%tmp2 = zext i1 %tmp1 to i8
|
|
ret i8 %tmp2
|
|
}
|
|
|
|
; FIXME: This test canonicalizes in a way that hides the fact that the
|
|
; comparison can be folded into the atomic subtract.
|
|
define i8 @test_sub_3_cmp_3_setcc_uge(i64* %p) #0 {
|
|
; CHECK-LABEL: test_sub_3_cmp_3_setcc_uge:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movq $-3, %rax
|
|
; CHECK-NEXT: lock xaddq %rax, (%rdi)
|
|
; CHECK-NEXT: cmpq $2, %rax
|
|
; CHECK-NEXT: seta %al
|
|
; CHECK-NEXT: retq
|
|
entry:
|
|
%tmp0 = atomicrmw sub i64* %p, i64 3 seq_cst
|
|
%tmp1 = icmp uge i64 %tmp0, 3
|
|
%tmp2 = zext i1 %tmp1 to i8
|
|
ret i8 %tmp2
|
|
}
|
|
|
|
attributes #0 = { nounwind }
|