1
0
mirror of https://github.com/RPCS3/llvm-mirror.git synced 2025-01-31 20:51:52 +01:00
llvm-mirror/test/CodeGen/X86/2012-08-07-CmpISelBug.ll
Simon Pilgrim 5ef0d1424c [TargetLowering] Add SimplifyMultipleUseDemandedBits
This patch introduces the DAG version of SimplifyMultipleUseDemandedBits, which attempts to peek through ops (mainly and/or/xor so far) that don't contribute to the demandedbits/elts of a node - which means we can do this even in cases where we have multiple uses of an op, which normally requires us to demanded all bits/elts. The intention is to remove a similar instruction - SelectionDAG::GetDemandedBits - once SimplifyMultipleUseDemandedBits has matured.

The InstCombine version of SimplifyMultipleUseDemandedBits can constant fold which I haven't added here yet, and so far I've only wired this up to some basic binops (and/or/xor/add/sub/mul) to demonstrate its use.

We do see a couple of regressions that need to be addressed:

    AMDGPU unsigned dot product codegen retains an AND mask (for ZERO_EXTEND) that it previously removed (but otherwise the dotproduct codegen is a lot better).
	
    X86/AVX2 has poor handling of vector ANY_EXTEND/ANY_EXTEND_VECTOR_INREG - it prematurely gets converted to ZERO_EXTEND_VECTOR_INREG.

The code owners have confirmed its ok for these cases to fixed up in future patches.

Differential Revision: https://reviews.llvm.org/D63281

llvm-svn: 366799
2019-07-23 12:39:08 +00:00

51 lines
1.6 KiB
LLVM

; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc < %s -mtriple=x86_64-apple-macosx | FileCheck %s
; Cmp lowering should not look past the truncate unless the high bits are known
; zero.
; rdar://12027825
define void @foo(i8 %arg4, i32 %arg5, i32* %arg14) nounwind {
; CHECK-LABEL: foo:
; CHECK: ## %bb.0: ## %bb
; CHECK-NEXT: ## kill: def $edi killed $edi def $rdi
; CHECK-NEXT: andl $32, %edi
; CHECK-NEXT: leal 13(%rdi), %eax
; CHECK-NEXT: xorb $-14, %al
; CHECK-NEXT: addb $82, %al
; CHECK-NEXT: movzbl %al, %eax
; CHECK-NEXT: testl %esi, %edi
; CHECK-NEXT: movl $1, %ecx
; CHECK-NEXT: cmovnel %eax, %ecx
; CHECK-NEXT: xorb $81, %cl
; CHECK-NEXT: movzbl %cl, %eax
; CHECK-NEXT: addl $2062143348, %eax ## imm = 0x7AE9CF74
; CHECK-NEXT: movl %eax, (%rdx)
; CHECK-NEXT: retq
bb:
%tmp48 = zext i8 %arg4 to i32
%tmp49 = and i32 %tmp48, 32
%tmp50 = add i32 %tmp49, 1593371643
%tmp55 = sub i32 %tmp50, 0
%tmp56 = add i32 %tmp55, 7787538
%tmp57 = xor i32 %tmp56, 1601159181
%tmp58 = xor i32 %arg5, 1601159181
%tmp59 = and i32 %tmp57, %tmp58
%tmp60 = add i32 %tmp59, -1263900958
%tmp67 = sub i32 %tmp60, 0
%tmp103 = xor i32 %tmp56, 13
%tmp104 = trunc i32 %tmp103 to i8
%tmp105 = sub i8 0, %tmp104
%tmp106 = add i8 %tmp105, -103
%tmp113 = sub i8 %tmp106, 0
%tmp114 = add i8 %tmp113, -72
%tmp141 = icmp ne i32 %tmp67, -1263900958
%tmp142 = select i1 %tmp141, i8 %tmp114, i8 1
%tmp143 = xor i8 %tmp142, 81
%tmp144 = zext i8 %tmp143 to i32
%tmp145 = add i32 %tmp144, 2062143348
%tmp152 = sub i32 %tmp145, 0
store i32 %tmp152, i32* %arg14
ret void
}