From 90953f232d38e1d6ad2d90701696e7f17190537d Mon Sep 17 00:00:00 2001 From: Simon Pilgrim Date: Tue, 3 Nov 2015 21:58:35 +0000 Subject: [PATCH] [X86][AVX] Tweaked shuffle stack folding tests To avoid alternative lowerings. llvm-svn: 251986 --- test/CodeGen/X86/stack-folding-fp-avx1.ll | 2 +- test/CodeGen/X86/stack-folding-int-avx2.ll | 8 ++++---- 2 files changed, 5 insertions(+), 5 deletions(-) diff --git a/test/CodeGen/X86/stack-folding-fp-avx1.ll b/test/CodeGen/X86/stack-folding-fp-avx1.ll index e74ccb03c1b..c7a2143b5b2 100644 --- a/test/CodeGen/X86/stack-folding-fp-avx1.ll +++ b/test/CodeGen/X86/stack-folding-fp-avx1.ll @@ -1494,7 +1494,7 @@ define <8 x float> @stack_fold_shufps_ymm(<8 x float> %a0, <8 x float> %a1) { ;CHECK-LABEL: stack_fold_shufps_ymm ;CHECK: vshufps $148, {{-?[0-9]*}}(%rsp), {{%ymm[0-9][0-9]*}}, {{%ymm[0-9][0-9]*}} {{.*#+}} 32-byte Folded Reload %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() - %2 = shufflevector <8 x float> %a0, <8 x float> %a1, <8 x i32> + %2 = shufflevector <8 x float> %a0, <8 x float> %a1, <8 x i32> ret <8 x float> %2 } diff --git a/test/CodeGen/X86/stack-folding-int-avx2.ll b/test/CodeGen/X86/stack-folding-int-avx2.ll index d0e51911091..7febee242cf 100644 --- a/test/CodeGen/X86/stack-folding-int-avx2.ll +++ b/test/CodeGen/X86/stack-folding-int-avx2.ll @@ -447,9 +447,9 @@ declare <8 x i32> @llvm.x86.avx2.permd(<8 x i32>, <8 x i32>) nounwind readonly define <4 x double> @stack_fold_permpd(<4 x double> %a0) { ;CHECK-LABEL: stack_fold_permpd - ;CHECK: vpermpd $255, {{-?[0-9]*}}(%rsp), {{%ymm[0-9][0-9]*}} {{.*#+}} 32-byte Folded Reload + ;CHECK: vpermpd $235, {{-?[0-9]*}}(%rsp), {{%ymm[0-9][0-9]*}} {{.*#+}} 32-byte Folded Reload %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() - %2 = shufflevector <4 x double> %a0, <4 x double> undef, <4 x i32> + %2 = shufflevector <4 x double> %a0, <4 x double> undef, <4 x i32> ; fadd forces execution domain %3 = fadd <4 x double> %2, ret <4 x double> %3 @@ -466,9 +466,9 @@ declare <8 x float> @llvm.x86.avx2.permps(<8 x float>, <8 x float>) nounwind rea define <4 x i64> @stack_fold_permq(<4 x i64> %a0) { ;CHECK-LABEL: stack_fold_permq - ;CHECK: vpermq $255, {{-?[0-9]*}}(%rsp), {{%ymm[0-9][0-9]*}} {{.*#+}} 32-byte Folded Reload + ;CHECK: vpermq $235, {{-?[0-9]*}}(%rsp), {{%ymm[0-9][0-9]*}} {{.*#+}} 32-byte Folded Reload %1 = tail call <2 x i64> asm sideeffect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{flags}"() - %2 = shufflevector <4 x i64> %a0, <4 x i64> undef, <4 x i32> + %2 = shufflevector <4 x i64> %a0, <4 x i64> undef, <4 x i32> ; add forces execution domain %3 = add <4 x i64> %2, ret <4 x i64> %3