1
0
mirror of https://github.com/RPCS3/llvm-mirror.git synced 2024-11-26 04:32:44 +01:00

[NFC][SLP] Add test case exposing SLP cost model bug.

The bug is related to aggregate build cost model adjustment
that adds a bias to cost triggering vectorization of actually
unprofitable to vectorize tree.

Differential Revision: https://reviews.llvm.org/D80682
This commit is contained in:
Valery N Dmitriev 2020-05-28 08:54:04 -07:00
parent 312b2495ca
commit ad3765c441

View File

@ -0,0 +1,118 @@
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
; RUN: opt < %s -slp-vectorizer -mtriple=x86_64-unknown-linux-gnu -mcpu=skylake -S | FileCheck %s
; Check no vectorization triggered with any portion of
; insertelement <8 x i32> instructions that build entire vector.
; Vectorization triggered by cost bias caused by subtracting
; the cost of entire "aggregate build" sequence while
; building vectorizable tree from only a portion of it.
; FIXME: this is unprofitable to vectorize.
define void @test(i32* nocapture %t2) {
; CHECK-LABEL: @test(
; CHECK-NEXT: [[T3:%.*]] = load i32, i32* [[T2:%.*]], align 4
; CHECK-NEXT: [[T4:%.*]] = getelementptr inbounds i32, i32* [[T2]], i64 7
; CHECK-NEXT: [[T5:%.*]] = load i32, i32* [[T4]], align 4
; CHECK-NEXT: [[T8:%.*]] = getelementptr inbounds i32, i32* [[T2]], i64 1
; CHECK-NEXT: [[T9:%.*]] = load i32, i32* [[T8]], align 4
; CHECK-NEXT: [[T10:%.*]] = getelementptr inbounds i32, i32* [[T2]], i64 6
; CHECK-NEXT: [[T11:%.*]] = load i32, i32* [[T10]], align 4
; CHECK-NEXT: [[T14:%.*]] = getelementptr inbounds i32, i32* [[T2]], i64 2
; CHECK-NEXT: [[T15:%.*]] = load i32, i32* [[T14]], align 4
; CHECK-NEXT: [[T16:%.*]] = getelementptr inbounds i32, i32* [[T2]], i64 5
; CHECK-NEXT: [[T17:%.*]] = load i32, i32* [[T16]], align 4
; CHECK-NEXT: [[T20:%.*]] = getelementptr inbounds i32, i32* [[T2]], i64 3
; CHECK-NEXT: [[T21:%.*]] = load i32, i32* [[T20]], align 4
; CHECK-NEXT: [[T22:%.*]] = getelementptr inbounds i32, i32* [[T2]], i64 4
; CHECK-NEXT: [[T23:%.*]] = load i32, i32* [[T22]], align 4
; CHECK-NEXT: [[T24:%.*]] = add nsw i32 [[T23]], [[T21]]
; CHECK-NEXT: [[T25:%.*]] = sub nsw i32 [[T21]], [[T23]]
; CHECK-NEXT: [[T27:%.*]] = sub nsw i32 [[T3]], [[T24]]
; CHECK-NEXT: [[T29:%.*]] = sub nsw i32 [[T9]], [[T15]]
; CHECK-NEXT: [[T30:%.*]] = add nsw i32 [[T27]], [[T29]]
; CHECK-NEXT: [[T31:%.*]] = mul nsw i32 [[T30]], 4433
; CHECK-NEXT: [[T34:%.*]] = mul nsw i32 [[T29]], -15137
; CHECK-NEXT: [[T37:%.*]] = add nsw i32 [[T25]], [[T11]]
; CHECK-NEXT: [[T38:%.*]] = add nsw i32 [[T17]], [[T5]]
; CHECK-NEXT: [[T39:%.*]] = add nsw i32 [[T37]], [[T38]]
; CHECK-NEXT: [[T40:%.*]] = mul nsw i32 [[T39]], 9633
; CHECK-NEXT: [[T41:%.*]] = mul nsw i32 [[T25]], 2446
; CHECK-NEXT: [[T42:%.*]] = mul nsw i32 [[T17]], 16819
; CHECK-NEXT: [[T47:%.*]] = mul nsw i32 [[T37]], -16069
; CHECK-NEXT: [[T48:%.*]] = mul nsw i32 [[T38]], -3196
; CHECK-NEXT: [[TMP1:%.*]] = insertelement <4 x i32> undef, i32 [[T15]], i32 0
; CHECK-NEXT: [[TMP2:%.*]] = insertelement <4 x i32> [[TMP1]], i32 [[T40]], i32 1
; CHECK-NEXT: [[TMP3:%.*]] = insertelement <4 x i32> [[TMP2]], i32 [[T27]], i32 2
; CHECK-NEXT: [[TMP4:%.*]] = insertelement <4 x i32> [[TMP3]], i32 [[T40]], i32 3
; CHECK-NEXT: [[TMP5:%.*]] = insertelement <4 x i32> undef, i32 [[T9]], i32 0
; CHECK-NEXT: [[TMP6:%.*]] = insertelement <4 x i32> [[TMP5]], i32 [[T48]], i32 1
; CHECK-NEXT: [[TMP7:%.*]] = insertelement <4 x i32> [[TMP6]], i32 6270, i32 2
; CHECK-NEXT: [[TMP8:%.*]] = insertelement <4 x i32> [[TMP7]], i32 [[T47]], i32 3
; CHECK-NEXT: [[TMP9:%.*]] = add nsw <4 x i32> [[TMP4]], [[TMP8]]
; CHECK-NEXT: [[TMP10:%.*]] = mul nsw <4 x i32> [[TMP4]], [[TMP8]]
; CHECK-NEXT: [[TMP11:%.*]] = shufflevector <4 x i32> [[TMP9]], <4 x i32> [[TMP10]], <4 x i32> <i32 0, i32 1, i32 6, i32 3>
; CHECK-NEXT: [[TMP12:%.*]] = extractelement <4 x i32> [[TMP11]], i32 0
; CHECK-NEXT: [[T65:%.*]] = insertelement <8 x i32> undef, i32 [[TMP12]], i32 0
; CHECK-NEXT: [[TMP13:%.*]] = extractelement <4 x i32> [[TMP11]], i32 1
; CHECK-NEXT: [[T66:%.*]] = insertelement <8 x i32> [[T65]], i32 [[TMP13]], i32 1
; CHECK-NEXT: [[TMP14:%.*]] = extractelement <4 x i32> [[TMP11]], i32 2
; CHECK-NEXT: [[T67:%.*]] = insertelement <8 x i32> [[T66]], i32 [[TMP14]], i32 2
; CHECK-NEXT: [[TMP15:%.*]] = extractelement <4 x i32> [[TMP11]], i32 3
; CHECK-NEXT: [[T68:%.*]] = insertelement <8 x i32> [[T67]], i32 [[TMP15]], i32 3
; CHECK-NEXT: [[T69:%.*]] = insertelement <8 x i32> [[T68]], i32 [[TMP12]], i32 4
; CHECK-NEXT: [[T70:%.*]] = insertelement <8 x i32> [[T69]], i32 [[TMP13]], i32 5
; CHECK-NEXT: [[T71:%.*]] = insertelement <8 x i32> [[T70]], i32 [[T34]], i32 6
; CHECK-NEXT: [[T72:%.*]] = insertelement <8 x i32> [[T71]], i32 [[TMP15]], i32 7
; CHECK-NEXT: [[T76:%.*]] = shl <8 x i32> [[T72]], <i32 3, i32 3, i32 3, i32 3, i32 3, i32 3, i32 3, i32 3>
; CHECK-NEXT: [[T79:%.*]] = bitcast i32* [[T2]] to <8 x i32>*
; CHECK-NEXT: store <8 x i32> [[T76]], <8 x i32>* [[T79]], align 4
; CHECK-NEXT: ret void
;
%t3 = load i32, i32* %t2, align 4
%t4 = getelementptr inbounds i32, i32* %t2, i64 7
%t5 = load i32, i32* %t4, align 4
%t8 = getelementptr inbounds i32, i32* %t2, i64 1
%t9 = load i32, i32* %t8, align 4
%t10 = getelementptr inbounds i32, i32* %t2, i64 6
%t11 = load i32, i32* %t10, align 4
%t14 = getelementptr inbounds i32, i32* %t2, i64 2
%t15 = load i32, i32* %t14, align 4
%t16 = getelementptr inbounds i32, i32* %t2, i64 5
%t17 = load i32, i32* %t16, align 4
%t20 = getelementptr inbounds i32, i32* %t2, i64 3
%t21 = load i32, i32* %t20, align 4
%t22 = getelementptr inbounds i32, i32* %t2, i64 4
%t23 = load i32, i32* %t22, align 4
%t24 = add nsw i32 %t23, %t21
%t25 = sub nsw i32 %t21, %t23
%t27 = sub nsw i32 %t3, %t24
%t28 = add nsw i32 %t15, %t9
%t29 = sub nsw i32 %t9, %t15
%t30 = add nsw i32 %t27, %t29
%t31 = mul nsw i32 %t30, 4433
%t32 = mul nsw i32 %t27, 6270
%t34 = mul nsw i32 %t29, -15137
%t37 = add nsw i32 %t25, %t11
%t38 = add nsw i32 %t17, %t5
%t39 = add nsw i32 %t37, %t38
%t40 = mul nsw i32 %t39, 9633
%t41 = mul nsw i32 %t25, 2446
%t42 = mul nsw i32 %t17, 16819
%t47 = mul nsw i32 %t37, -16069
%t48 = mul nsw i32 %t38, -3196
%t49 = add nsw i32 %t40, %t47
%t50 = add nsw i32 %t40, %t48
%t65 = insertelement <8 x i32> undef, i32 %t28, i32 0
%t66 = insertelement <8 x i32> %t65, i32 %t50, i32 1
%t67 = insertelement <8 x i32> %t66, i32 %t32, i32 2
%t68 = insertelement <8 x i32> %t67, i32 %t49, i32 3
%t69 = insertelement <8 x i32> %t68, i32 %t28, i32 4
%t70 = insertelement <8 x i32> %t69, i32 %t50, i32 5
%t71 = insertelement <8 x i32> %t70, i32 %t34, i32 6
%t72 = insertelement <8 x i32> %t71, i32 %t49, i32 7
%t76 = shl <8 x i32> %t72, <i32 3, i32 3, i32 3, i32 3, i32 3, i32 3, i32 3, i32 3>
%t79 = bitcast i32* %t2 to <8 x i32>*
store <8 x i32> %t76, <8 x i32>* %t79, align 4
ret void
}