mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2025-02-01 13:11:39 +01:00
96817ba3db
Narrowing vector binops came up in the demanded bits discussion in D52912. I don't think we're going to be able to do this transform in IR as a canonicalization because of the risk of creating unsupported widths for vector ops, but we already have a DAG TLI hook to allow what I was hoping for: isExtractSubvectorCheap(). This is currently enabled for x86, ARM, and AArch64 (although only x86 has existing regression test diffs). This is artificially limited to not look through bitcasts because there are so many test diffs already, but that's marked with a TODO and is a small follow-up. Differential Revision: https://reviews.llvm.org/D53784 llvm-svn: 345602
57 lines
3.4 KiB
LLVM
57 lines
3.4 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=core-avx2 | FileCheck %s
|
|
|
|
; rdar://11314175: SD Scheduler, BuildSchedUnits assert:
|
|
; N->getNodeId() == -1 && "Node already inserted!
|
|
|
|
define void @func(<4 x float> %a, <16 x i8> %b, <16 x i8> %c, <8 x float> %d, <8 x float> %e, <8 x float>* %f) nounwind ssp {
|
|
; CHECK-LABEL: func:
|
|
; CHECK: ## %bb.0:
|
|
; CHECK-NEXT: vmovdqu 0, %xmm0
|
|
; CHECK-NEXT: vpalignr {{.*#+}} xmm1 = xmm0[4,5,6,7,8,9,10,11,12,13,14,15],xmm1[0,1,2,3]
|
|
; CHECK-NEXT: vxorps %xmm2, %xmm2, %xmm2
|
|
; CHECK-NEXT: vmulps %xmm1, %xmm1, %xmm1
|
|
; CHECK-NEXT: vmulps %xmm0, %xmm0, %xmm0
|
|
; CHECK-NEXT: vaddps %xmm1, %xmm0, %xmm0
|
|
; CHECK-NEXT: vaddps %xmm0, %xmm0, %xmm0
|
|
; CHECK-NEXT: vmulps %xmm0, %xmm0, %xmm0
|
|
; CHECK-NEXT: vperm2f128 {{.*#+}} ymm0 = zero,zero,ymm0[0,1]
|
|
; CHECK-NEXT: vaddps %ymm0, %ymm0, %ymm0
|
|
; CHECK-NEXT: vhaddps %ymm4, %ymm0, %ymm0
|
|
; CHECK-NEXT: vsubps %ymm0, %ymm0, %ymm0
|
|
; CHECK-NEXT: vhaddps %ymm0, %ymm2, %ymm0
|
|
; CHECK-NEXT: vmovaps %ymm0, (%rdi)
|
|
; CHECK-NEXT: vzeroupper
|
|
; CHECK-NEXT: retq
|
|
%tmp = load <4 x float>, <4 x float>* null, align 1
|
|
%tmp14 = getelementptr <4 x float>, <4 x float>* null, i32 2
|
|
%tmp15 = load <4 x float>, <4 x float>* %tmp14, align 1
|
|
%tmp16 = shufflevector <4 x float> %tmp, <4 x float> <float 0.000000e+00, float undef, float undef, float undef>, <8 x i32> <i32 0, i32 1, i32 2, i32 3, i32 4, i32 4, i32 4, i32 4>
|
|
%tmp17 = call <8 x float> @llvm.x86.avx.vinsertf128.ps.256(<8 x float> %tmp16, <4 x float> %a, i8 1)
|
|
%tmp18 = bitcast <4 x float> %tmp to <16 x i8>
|
|
%tmp19 = shufflevector <16 x i8> %tmp18, <16 x i8> %b, <16 x i32> <i32 4, i32 5, i32 6, i32 7, i32 8, i32 9, i32 10, i32 11, i32 12, i32 13, i32 14, i32 15, i32 16, i32 17, i32 18, i32 19>
|
|
%tmp20 = bitcast <16 x i8> %tmp19 to <4 x float>
|
|
%tmp21 = bitcast <4 x float> %tmp15 to <16 x i8>
|
|
%tmp22 = shufflevector <16 x i8> %c, <16 x i8> %tmp21, <16 x i32> <i32 4, i32 5, i32 6, i32 7, i32 8, i32 9, i32 10, i32 11, i32 12, i32 13, i32 14, i32 15, i32 16, i32 17, i32 18, i32 19>
|
|
%tmp23 = bitcast <16 x i8> %tmp22 to <4 x float>
|
|
%tmp24 = shufflevector <4 x float> %tmp20, <4 x float> <float 0.000000e+00, float undef, float undef, float undef>, <8 x i32> <i32 0, i32 1, i32 2, i32 3, i32 4, i32 4, i32 4, i32 4>
|
|
%tmp25 = call <8 x float> @llvm.x86.avx.vinsertf128.ps.256(<8 x float> %tmp24, <4 x float> %tmp23, i8 1)
|
|
%tmp26 = fmul <8 x float> %tmp17, %tmp17
|
|
%tmp27 = fmul <8 x float> %tmp25, %tmp25
|
|
%tmp28 = fadd <8 x float> %tmp26, %tmp27
|
|
%tmp29 = fadd <8 x float> %tmp28, %tmp28
|
|
%tmp30 = shufflevector <8 x float> %tmp29, <8 x float> %d, <4 x i32> <i32 0, i32 1, i32 2, i32 3>
|
|
%tmp31 = fmul <4 x float> %tmp30, %tmp30
|
|
%tmp32 = call <8 x float> @llvm.x86.avx.vinsertf128.ps.256(<8 x float> zeroinitializer, <4 x float> %tmp31, i8 1)
|
|
%tmp33 = fadd <8 x float> %tmp32, %tmp32
|
|
%tmp34 = call <8 x float> @llvm.x86.avx.hadd.ps.256(<8 x float> %tmp33, <8 x float> %e) nounwind
|
|
%tmp35 = fsub <8 x float> %tmp34, %tmp34
|
|
%tmp36 = call <8 x float> @llvm.x86.avx.hadd.ps.256(<8 x float> zeroinitializer, <8 x float> %tmp35) nounwind
|
|
store <8 x float> %tmp36, <8 x float>* %f, align 32
|
|
ret void
|
|
}
|
|
|
|
declare <8 x float> @llvm.x86.avx.vinsertf128.ps.256(<8 x float>, <4 x float>, i8) nounwind readnone
|
|
|
|
declare <8 x float> @llvm.x86.avx.hadd.ps.256(<8 x float>, <8 x float>) nounwind readnone
|