1
0
mirror of https://github.com/RPCS3/llvm-mirror.git synced 2024-11-23 11:13:28 +01:00
llvm-mirror/test/CodeGen/X86/pr46455.ll
Simon Pilgrim 21418c85a1 [X86][SSE] getV4X86ShuffleImm8 - canonicalize broadcast masks
If the mask input to getV4X86ShuffleImm8 only refers to a single source element (+ undefs) then canonicalize to a full broadcast.

getV4X86ShuffleImm8 defaults to inline values for undefs, which can be useful for shuffle widening/narrowing but does leave SimplifyDemanded* calls thinking the shuffle depends on unnecessary elements.

I'm still investigating what we should do more generally to avoid these undemanded elements, but broadcast cases was a simpler win.
2020-07-29 11:32:44 +01:00

40 lines
1.9 KiB
LLVM

; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc < %s -mtriple=x86_64-unknown-linux-gnu -mattr=avx512vl,avx512bw,avx512dq | FileCheck %s
define void @EntryModule(i8** %buffer_table) {
; CHECK-LABEL: EntryModule:
; CHECK: # %bb.0: # %entry
; CHECK-NEXT: vxorps %xmm0, %xmm0, %xmm0
; CHECK-NEXT: movq (%rdi), %rax
; CHECK-NEXT: movq 24(%rdi), %rcx
; CHECK-NEXT: vcmpneqps (%rax), %ymm0, %ymm0
; CHECK-NEXT: vpsrld $31, %xmm0, %xmm1
; CHECK-NEXT: vpshufd {{.*#+}} xmm2 = xmm1[1,1,1,1]
; CHECK-NEXT: vpshufd {{.*#+}} xmm3 = xmm1[2,3,2,3]
; CHECK-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[3,3,3,3]
; CHECK-NEXT: vpaddd %xmm1, %xmm3, %xmm1
; CHECK-NEXT: vpsubd %xmm0, %xmm2, %xmm0
; CHECK-NEXT: vpaddd %xmm1, %xmm0, %xmm0
; CHECK-NEXT: vmovd %xmm0, (%rcx)
; CHECK-NEXT: vzeroupper
; CHECK-NEXT: retq
entry:
%i = bitcast i8** %buffer_table to <8 x float>**
%i1 = load <8 x float>*, <8 x float>** %i, align 8
%i6 = load <8 x float>, <8 x float>* %i1, align 16
%i7 = fcmp une <8 x float> %i6, zeroinitializer
%i8 = zext <8 x i1> %i7 to <8 x i32>
%i18 = getelementptr inbounds i8*, i8** %buffer_table, i64 3
%i19 = load i8*, i8** %i18, align 8
%shift = shufflevector <8 x i32> %i8, <8 x i32> undef, <8 x i32> <i32 1, i32 undef, i32 undef, i32 undef, i32 undef, i32 undef, i32 undef, i32 undef>
%i20 = add nuw nsw <8 x i32> %shift, %i8
%shift13 = shufflevector <8 x i32> %i8, <8 x i32> undef, <8 x i32> <i32 2, i32 undef, i32 undef, i32 undef, i32 undef, i32 undef, i32 undef, i32 undef>
%i21 = add nuw nsw <8 x i32> %i20, %shift13
%shift14 = shufflevector <8 x i32> %i8, <8 x i32> undef, <8 x i32> <i32 3, i32 undef, i32 undef, i32 undef, i32 undef, i32 undef, i32 undef, i32 undef>
%i22 = add nuw nsw <8 x i32> %i21, %shift14
%i23 = extractelement <8 x i32> %i22, i32 0
%i24 = bitcast i8* %i19 to i32*
store i32 %i23, i32* %i24, align 8
ret void
}