mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-26 12:43:36 +01:00
2be00d9e54
Current tail duplication integrated in bb layout is designed to increase the fallthrough from a BB's predecessor to its successor, but we have observed cases that duplication doesn't increase fallthrough, or it brings too much size overhead. To overcome these two issues in function canTailDuplicateUnplacedPreds I add two checks: make sure there is at least one duplication in current work set. the number of duplication should not exceed the number of successors. The modification in hasBetterLayoutPredecessor fixes a bug that potential predecessor must be at the bottom of a chain. Differential Revision: https://reviews.llvm.org/D64376
97 lines
3.5 KiB
LLVM
97 lines
3.5 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+mmx,+sse2 | FileCheck %s
|
|
|
|
%SA = type <{ %union.anon, i32, [4 x i8], i8*, i8*, i8*, i32, [4 x i8] }>
|
|
%union.anon = type { <1 x i64> }
|
|
|
|
; Check that extra movd (copy) instructions aren't generated.
|
|
|
|
define i32 @test(%SA* %pSA, i16* %A, i32 %B, i32 %C, i32 %D, i8* %E) {
|
|
; CHECK-LABEL: test:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: pshufw $238, (%rdi), %mm0 # mm0 = mem[2,3,2,3]
|
|
; CHECK-NEXT: movd %mm0, %eax
|
|
; CHECK-NEXT: testl %eax, %eax
|
|
; CHECK-NEXT: je .LBB0_1
|
|
; CHECK-NEXT: # %bb.2: # %if.B
|
|
; CHECK-NEXT: pshufw $238, %mm0, %mm0 # mm0 = mm0[2,3,2,3]
|
|
; CHECK-NEXT: movq %mm0, %rax
|
|
; CHECK-NEXT: jmp .LBB0_3
|
|
; CHECK-NEXT: .LBB0_1: # %if.A
|
|
; CHECK-NEXT: movd %edx, %mm1
|
|
; CHECK-NEXT: psllq %mm1, %mm0
|
|
; CHECK-NEXT: movq %mm0, %rax
|
|
; CHECK-NEXT: testq %rax, %rax
|
|
; CHECK-NEXT: jne .LBB0_4
|
|
; CHECK-NEXT: .LBB0_3: # %if.C
|
|
; CHECK-NEXT: testl %eax, %eax
|
|
; CHECK-NEXT: je .LBB0_1
|
|
; CHECK-NEXT: .LBB0_4: # %merge
|
|
; CHECK-NEXT: pshufw $238, %mm0, %mm0 # mm0 = mm0[2,3,2,3]
|
|
; CHECK-NEXT: movd %mm0, %eax
|
|
; CHECK-NEXT: retq
|
|
entry:
|
|
%shl = shl i32 1, %B
|
|
%shl1 = shl i32 %C, %B
|
|
%shl2 = shl i32 1, %D
|
|
%v = getelementptr inbounds %SA, %SA* %pSA, i64 0, i32 0, i32 0
|
|
%v0 = load <1 x i64>, <1 x i64>* %v, align 8
|
|
%SA0 = getelementptr inbounds %SA, %SA* %pSA, i64 0, i32 1
|
|
%v1 = load i32, i32* %SA0, align 4
|
|
%SA1 = getelementptr inbounds %SA, %SA* %pSA, i64 0, i32 3
|
|
%v2 = load i8*, i8** %SA1, align 8
|
|
%SA2 = getelementptr inbounds %SA, %SA* %pSA, i64 0, i32 4
|
|
%v3 = load i8*, i8** %SA2, align 8
|
|
%v4 = bitcast <1 x i64> %v0 to <4 x i16>
|
|
%v5 = bitcast <4 x i16> %v4 to x86_mmx
|
|
%v6 = tail call x86_mmx @llvm.x86.sse.pshuf.w(x86_mmx %v5, i8 -18)
|
|
%v7 = bitcast x86_mmx %v6 to <4 x i16>
|
|
%v8 = bitcast <4 x i16> %v7 to <1 x i64>
|
|
%v9 = extractelement <1 x i64> %v8, i32 0
|
|
%v10 = bitcast i64 %v9 to <2 x i32>
|
|
%v11 = extractelement <2 x i32> %v10, i32 0
|
|
%cmp = icmp eq i32 %v11, 0
|
|
br i1 %cmp, label %if.A, label %if.B
|
|
|
|
if.A:
|
|
%pa = phi <1 x i64> [ %v8, %entry ], [ %vx, %if.C ]
|
|
%v17 = extractelement <1 x i64> %pa, i32 0
|
|
%v18 = bitcast i64 %v17 to x86_mmx
|
|
%v19 = tail call x86_mmx @llvm.x86.mmx.pslli.q(x86_mmx %v18, i32 %B) #2
|
|
%v20 = bitcast x86_mmx %v19 to i64
|
|
%v21 = insertelement <1 x i64> undef, i64 %v20, i32 0
|
|
%cmp3 = icmp eq i64 %v20, 0
|
|
br i1 %cmp3, label %if.C, label %merge
|
|
|
|
if.B:
|
|
%v34 = bitcast <1 x i64> %v8 to <4 x i16>
|
|
%v35 = bitcast <4 x i16> %v34 to x86_mmx
|
|
%v36 = tail call x86_mmx @llvm.x86.sse.pshuf.w(x86_mmx %v35, i8 -18)
|
|
%v37 = bitcast x86_mmx %v36 to <4 x i16>
|
|
%v38 = bitcast <4 x i16> %v37 to <1 x i64>
|
|
br label %if.C
|
|
|
|
if.C:
|
|
%vx = phi <1 x i64> [ %v21, %if.A ], [ %v38, %if.B ]
|
|
%cvt = bitcast <1 x i64> %vx to <2 x i32>
|
|
%ex = extractelement <2 x i32> %cvt, i32 0
|
|
%cmp2 = icmp eq i32 %ex, 0
|
|
br i1 %cmp2, label %if.A, label %merge
|
|
|
|
merge:
|
|
%vy = phi <1 x i64> [ %v21, %if.A ], [ %vx, %if.C ]
|
|
%v130 = bitcast <1 x i64> %vy to <4 x i16>
|
|
%v131 = bitcast <4 x i16> %v130 to x86_mmx
|
|
%v132 = tail call x86_mmx @llvm.x86.sse.pshuf.w(x86_mmx %v131, i8 -18)
|
|
%v133 = bitcast x86_mmx %v132 to <4 x i16>
|
|
%v134 = bitcast <4 x i16> %v133 to <1 x i64>
|
|
%v135 = extractelement <1 x i64> %v134, i32 0
|
|
%v136 = bitcast i64 %v135 to <2 x i32>
|
|
%v137 = extractelement <2 x i32> %v136, i32 0
|
|
ret i32 %v137
|
|
}
|
|
|
|
|
|
declare x86_mmx @llvm.x86.sse.pshuf.w(x86_mmx, i8)
|
|
declare x86_mmx @llvm.x86.mmx.pslli.q(x86_mmx, i32)
|