1
0
mirror of https://github.com/RPCS3/llvm-mirror.git synced 2024-10-23 21:13:02 +02:00
llvm-mirror/test/CodeGen/X86/chain_order.ll
Chandler Carruth 7b066daf55 [SDAG] Revert r296476 (and r296486, r296668, r296690).
This patch causes compile times for some patterns to explode. I have
a (large, unreduced) test case that slows down by more than 20x and
several test cases slow down by 2x. I'm sending some of the test cases
directly to Nirav and following up with more details in the review log,
but this should unblock anyone else hitting this.

llvm-svn: 296862
2017-03-03 10:02:25 +00:00

42 lines
1.9 KiB
LLVM

; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc < %s -mcpu=corei7-avx -mtriple=x86_64-linux | FileCheck %s
; A test from pifft (after SLP-vectorization) that fails when we drop the chain on newly merged loads.
define void @cftx020(double* nocapture %a) {
; CHECK-LABEL: cftx020:
; CHECK: # BB#0: # %entry
; CHECK-NEXT: vmovsd {{.*#+}} xmm0 = mem[0],zero
; CHECK-NEXT: vmovsd {{.*#+}} xmm1 = mem[0],zero
; CHECK-NEXT: vmovhpd {{.*#+}} xmm0 = xmm0[0],mem[0]
; CHECK-NEXT: vmovhpd {{.*#+}} xmm1 = xmm1[0],mem[0]
; CHECK-NEXT: vaddpd %xmm1, %xmm0, %xmm0
; CHECK-NEXT: vmovupd (%rdi), %xmm1
; CHECK-NEXT: vsubpd 16(%rdi), %xmm1, %xmm1
; CHECK-NEXT: vmovupd %xmm0, (%rdi)
; CHECK-NEXT: vmovupd %xmm1, 16(%rdi)
; CHECK-NEXT: retq
entry:
%0 = load double, double* %a, align 8
%arrayidx1 = getelementptr inbounds double, double* %a, i64 2
%1 = load double, double* %arrayidx1, align 8
%arrayidx2 = getelementptr inbounds double, double* %a, i64 1
%2 = load double, double* %arrayidx2, align 8
%arrayidx3 = getelementptr inbounds double, double* %a, i64 3
%3 = load double, double* %arrayidx3, align 8
%4 = insertelement <2 x double> undef, double %0, i32 0
%5 = insertelement <2 x double> %4, double %3, i32 1
%6 = insertelement <2 x double> undef, double %1, i32 0
%7 = insertelement <2 x double> %6, double %2, i32 1
%8 = fadd <2 x double> %5, %7
%9 = bitcast double* %a to <2 x double>*
store <2 x double> %8, <2 x double>* %9, align 8
%10 = insertelement <2 x double> undef, double %0, i32 0
%11 = insertelement <2 x double> %10, double %2, i32 1
%12 = insertelement <2 x double> undef, double %1, i32 0
%13 = insertelement <2 x double> %12, double %3, i32 1
%14 = fsub <2 x double> %11, %13
%15 = bitcast double* %arrayidx1 to <2 x double>*
store <2 x double> %14, <2 x double>* %15, align 8
ret void
}