mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-23 19:23:23 +01:00
0773b05cfa
This changes the definition of t2DoLoopStart from t2DoLoopStart rGPR to GPRlr = t2DoLoopStart rGPR This will hopefully mean that low overhead loops are more tied together, and we can more reliably generate loops without reverting or being at the whims of the register allocator. This is a fairly simple change in itself, but leads to a number of other required alterations. - The hardware loop pass, if UsePhi is set, now generates loops of the form: %start = llvm.start.loop.iterations(%N) loop: %p = phi [%start], [%dec] %dec = llvm.loop.decrement.reg(%p, 1) %c = icmp ne %dec, 0 br %c, loop, exit - For this a new llvm.start.loop.iterations intrinsic was added, identical to llvm.set.loop.iterations but produces a value as seen above, gluing the loop together more through def-use chains. - This new instrinsic conceptually produces the same output as input, which is taught to SCEV so that the checks in MVETailPredication are not affected. - Some minor changes are needed to the ARMLowOverheadLoop pass, but it has been left mostly as before. We should now more reliably be able to tell that the t2DoLoopStart is correct without having to prove it, but t2WhileLoopStart and tail-predicated loops will remain the same. - And all the tests have been updated. There are a lot of them! This patch on it's own might cause more trouble that it helps, with more tail-predicated loops being reverted, but some additional patches can hopefully improve upon that to get to something that is better overall. Differential Revision: https://reviews.llvm.org/D89881
551 lines
31 KiB
LLVM
551 lines
31 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
|
|
; RUN: opt -mtriple=thumbv8.1m.main -mve-tail-predication -tail-predication=enabled -mattr=+mve %s -S -o - | FileCheck %s
|
|
|
|
define dso_local void @foo(i32* noalias nocapture %A, i32* noalias nocapture readonly %B, i32* noalias nocapture readonly %C, i32* noalias nocapture readnone %D, i32 %N) local_unnamed_addr #0 {
|
|
; CHECK-LABEL: @foo(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: [[START:%.*]] = call i32 @llvm.start.loop.iterations.i32(i32 8001)
|
|
; CHECK-NEXT: br label [[VECTOR_BODY:%.*]]
|
|
; CHECK: vector.body:
|
|
; CHECK-NEXT: [[LSR_IV14:%.*]] = phi i32* [ [[SCEVGEP15:%.*]], [[VECTOR_BODY]] ], [ [[A:%.*]], [[ENTRY:%.*]] ]
|
|
; CHECK-NEXT: [[LSR_IV11:%.*]] = phi i32* [ [[SCEVGEP12:%.*]], [[VECTOR_BODY]] ], [ [[C:%.*]], [[ENTRY]] ]
|
|
; CHECK-NEXT: [[LSR_IV:%.*]] = phi i32* [ [[SCEVGEP:%.*]], [[VECTOR_BODY]] ], [ [[B:%.*]], [[ENTRY]] ]
|
|
; CHECK-NEXT: [[INDEX:%.*]] = phi i32 [ 0, [[ENTRY]] ], [ [[INDEX_NEXT:%.*]], [[VECTOR_BODY]] ]
|
|
; CHECK-NEXT: [[TMP0:%.*]] = phi i32 [ [[START]], [[ENTRY]] ], [ [[TMP5:%.*]], [[VECTOR_BODY]] ]
|
|
; CHECK-NEXT: [[TMP1:%.*]] = phi i32 [ 32003, [[ENTRY]] ], [ [[TMP3:%.*]], [[VECTOR_BODY]] ]
|
|
; CHECK-NEXT: [[LSR_IV1416:%.*]] = bitcast i32* [[LSR_IV14]] to <4 x i32>*
|
|
; CHECK-NEXT: [[LSR_IV1113:%.*]] = bitcast i32* [[LSR_IV11]] to <4 x i32>*
|
|
; CHECK-NEXT: [[LSR_IV10:%.*]] = bitcast i32* [[LSR_IV]] to <4 x i32>*
|
|
; CHECK-NEXT: [[BROADCAST_SPLATINSERT:%.*]] = insertelement <4 x i32> undef, i32 [[INDEX]], i32 0
|
|
; CHECK-NEXT: [[BROADCAST_SPLAT:%.*]] = shufflevector <4 x i32> [[BROADCAST_SPLATINSERT]], <4 x i32> undef, <4 x i32> zeroinitializer
|
|
; CHECK-NEXT: [[INDUCTION:%.*]] = add <4 x i32> [[BROADCAST_SPLAT]], <i32 0, i32 1, i32 2, i32 3>
|
|
; CHECK-NEXT: [[TMP2:%.*]] = call <4 x i1> @llvm.arm.mve.vctp32(i32 [[TMP1]])
|
|
; CHECK-NEXT: [[TMP3]] = sub i32 [[TMP1]], 4
|
|
; CHECK-NEXT: [[WIDE_MASKED_LOAD:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* [[LSR_IV10]], i32 4, <4 x i1> [[TMP2]], <4 x i32> undef)
|
|
; CHECK-NEXT: [[WIDE_MASKED_LOAD9:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* [[LSR_IV1113]], i32 4, <4 x i1> [[TMP2]], <4 x i32> undef)
|
|
; CHECK-NEXT: [[TMP4:%.*]] = add nsw <4 x i32> [[WIDE_MASKED_LOAD9]], [[WIDE_MASKED_LOAD]]
|
|
; CHECK-NEXT: call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> [[TMP4]], <4 x i32>* [[LSR_IV1416]], i32 4, <4 x i1> [[TMP2]])
|
|
; CHECK-NEXT: [[INDEX_NEXT]] = add i32 [[INDEX]], 4
|
|
; CHECK-NEXT: [[SCEVGEP]] = getelementptr i32, i32* [[LSR_IV]], i32 4
|
|
; CHECK-NEXT: [[SCEVGEP12]] = getelementptr i32, i32* [[LSR_IV11]], i32 4
|
|
; CHECK-NEXT: [[SCEVGEP15]] = getelementptr i32, i32* [[LSR_IV14]], i32 4
|
|
; CHECK-NEXT: [[TMP5]] = call i32 @llvm.loop.decrement.reg.i32(i32 [[TMP0]], i32 1)
|
|
; CHECK-NEXT: [[TMP6:%.*]] = icmp ne i32 [[TMP5]], 0
|
|
; CHECK-NEXT: br i1 [[TMP6]], label [[VECTOR_BODY]], label [[FOR_COND_CLEANUP:%.*]]
|
|
; CHECK: for.cond.cleanup:
|
|
; CHECK-NEXT: ret void
|
|
;
|
|
entry:
|
|
%start = call i32 @llvm.start.loop.iterations.i32(i32 8001)
|
|
br label %vector.body
|
|
|
|
vector.body:
|
|
%lsr.iv14 = phi i32* [ %scevgep15, %vector.body ], [ %A, %entry ]
|
|
%lsr.iv11 = phi i32* [ %scevgep12, %vector.body ], [ %C, %entry ]
|
|
%lsr.iv = phi i32* [ %scevgep, %vector.body ], [ %B, %entry ]
|
|
%index = phi i32 [ 0, %entry ], [ %index.next, %vector.body ]
|
|
%0 = phi i32 [ %start, %entry ], [ %3, %vector.body ]
|
|
%lsr.iv1416 = bitcast i32* %lsr.iv14 to <4 x i32>*
|
|
%lsr.iv1113 = bitcast i32* %lsr.iv11 to <4 x i32>*
|
|
%lsr.iv10 = bitcast i32* %lsr.iv to <4 x i32>*
|
|
%broadcast.splatinsert = insertelement <4 x i32> undef, i32 %index, i32 0
|
|
%broadcast.splat = shufflevector <4 x i32> %broadcast.splatinsert, <4 x i32> undef, <4 x i32> zeroinitializer
|
|
%induction = add <4 x i32> %broadcast.splat, <i32 0, i32 1, i32 2, i32 3>
|
|
|
|
; %1 = icmp ult <4 x i32> %induction, <i32 32002, i32 32002, i32 32002, i32 32002>
|
|
%1 = call <4 x i1> @llvm.get.active.lane.mask.v4i1.i32(i32 %index, i32 32003)
|
|
|
|
%wide.masked.load = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv10, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%wide.masked.load9 = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv1113, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%2 = add nsw <4 x i32> %wide.masked.load9, %wide.masked.load
|
|
call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> %2, <4 x i32>* %lsr.iv1416, i32 4, <4 x i1> %1)
|
|
%index.next = add i32 %index, 4
|
|
%scevgep = getelementptr i32, i32* %lsr.iv, i32 4
|
|
%scevgep12 = getelementptr i32, i32* %lsr.iv11, i32 4
|
|
%scevgep15 = getelementptr i32, i32* %lsr.iv14, i32 4
|
|
%3 = call i32 @llvm.loop.decrement.reg.i32(i32 %0, i32 1)
|
|
%4 = icmp ne i32 %3, 0
|
|
br i1 %4, label %vector.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
ret void
|
|
}
|
|
|
|
; Silly test case: the loop count is constant and a multiple of the vectorisation
|
|
; factor. So, the vectoriser should not produce masked loads/stores and there's
|
|
; nothing to tail-predicate here, just checking.
|
|
define dso_local void @foo2(i32* noalias nocapture %A, i32* noalias nocapture readonly %B, i32* noalias nocapture readonly %C, i32* noalias nocapture readnone %D, i32 %N) local_unnamed_addr #0 {
|
|
; CHECK-LABEL: @foo2(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: [[START:%.*]] = call i32 @llvm.start.loop.iterations.i32(i32 2000)
|
|
; CHECK-NEXT: br label [[VECTOR_BODY:%.*]]
|
|
; CHECK: vector.body:
|
|
; CHECK-NEXT: [[LSR_IV14:%.*]] = phi i32* [ [[SCEVGEP15:%.*]], [[VECTOR_BODY]] ], [ [[A:%.*]], [[ENTRY:%.*]] ]
|
|
; CHECK-NEXT: [[LSR_IV11:%.*]] = phi i32* [ [[SCEVGEP12:%.*]], [[VECTOR_BODY]] ], [ [[C:%.*]], [[ENTRY]] ]
|
|
; CHECK-NEXT: [[LSR_IV:%.*]] = phi i32* [ [[SCEVGEP:%.*]], [[VECTOR_BODY]] ], [ [[B:%.*]], [[ENTRY]] ]
|
|
; CHECK-NEXT: [[TMP0:%.*]] = phi i32 [ [[START]], [[ENTRY]] ], [ [[TMP2:%.*]], [[VECTOR_BODY]] ]
|
|
; CHECK-NEXT: [[LSR_IV1416:%.*]] = bitcast i32* [[LSR_IV14]] to <4 x i32>*
|
|
; CHECK-NEXT: [[LSR_IV1113:%.*]] = bitcast i32* [[LSR_IV11]] to <4 x i32>*
|
|
; CHECK-NEXT: [[LSR_IV10:%.*]] = bitcast i32* [[LSR_IV]] to <4 x i32>*
|
|
; CHECK-NEXT: [[WIDE_LOAD:%.*]] = load <4 x i32>, <4 x i32>* [[LSR_IV10]], align 4
|
|
; CHECK-NEXT: [[WIDE_LOAD9:%.*]] = load <4 x i32>, <4 x i32>* [[LSR_IV1113]], align 4
|
|
; CHECK-NEXT: [[TMP1:%.*]] = add nsw <4 x i32> [[WIDE_LOAD9]], [[WIDE_LOAD]]
|
|
; CHECK-NEXT: store <4 x i32> [[TMP1]], <4 x i32>* [[LSR_IV1416]], align 4
|
|
; CHECK-NEXT: [[SCEVGEP]] = getelementptr i32, i32* [[LSR_IV]], i32 4
|
|
; CHECK-NEXT: [[SCEVGEP12]] = getelementptr i32, i32* [[LSR_IV11]], i32 4
|
|
; CHECK-NEXT: [[SCEVGEP15]] = getelementptr i32, i32* [[LSR_IV14]], i32 4
|
|
; CHECK-NEXT: [[TMP2]] = call i32 @llvm.loop.decrement.reg.i32(i32 [[TMP0]], i32 1)
|
|
; CHECK-NEXT: [[TMP3:%.*]] = icmp ne i32 [[TMP2]], 0
|
|
; CHECK-NEXT: br i1 [[TMP3]], label [[VECTOR_BODY]], label [[FOR_COND_CLEANUP:%.*]]
|
|
; CHECK: for.cond.cleanup:
|
|
; CHECK-NEXT: ret void
|
|
;
|
|
entry:
|
|
%start = call i32 @llvm.start.loop.iterations.i32(i32 2000)
|
|
br label %vector.body
|
|
|
|
vector.body:
|
|
%lsr.iv14 = phi i32* [ %scevgep15, %vector.body ], [ %A, %entry ]
|
|
%lsr.iv11 = phi i32* [ %scevgep12, %vector.body ], [ %C, %entry ]
|
|
%lsr.iv = phi i32* [ %scevgep, %vector.body ], [ %B, %entry ]
|
|
%0 = phi i32 [ %start, %entry ], [ %2, %vector.body ]
|
|
%lsr.iv1416 = bitcast i32* %lsr.iv14 to <4 x i32>*
|
|
%lsr.iv1113 = bitcast i32* %lsr.iv11 to <4 x i32>*
|
|
%lsr.iv10 = bitcast i32* %lsr.iv to <4 x i32>*
|
|
%wide.load = load <4 x i32>, <4 x i32>* %lsr.iv10, align 4
|
|
%wide.load9 = load <4 x i32>, <4 x i32>* %lsr.iv1113, align 4
|
|
%1 = add nsw <4 x i32> %wide.load9, %wide.load
|
|
store <4 x i32> %1, <4 x i32>* %lsr.iv1416, align 4
|
|
%scevgep = getelementptr i32, i32* %lsr.iv, i32 4
|
|
%scevgep12 = getelementptr i32, i32* %lsr.iv11, i32 4
|
|
%scevgep15 = getelementptr i32, i32* %lsr.iv14, i32 4
|
|
%2 = call i32 @llvm.loop.decrement.reg.i32(i32 %0, i32 1)
|
|
%3 = icmp ne i32 %2, 0
|
|
br i1 %3, label %vector.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
ret void
|
|
}
|
|
|
|
; Check that the icmp is a ult
|
|
define dso_local void @foo3(i32* noalias nocapture %A, i32* noalias nocapture readonly %B, i32* noalias nocapture readonly %C, i32* noalias nocapture readnone %D, i32 %N) local_unnamed_addr #0 {
|
|
; CHECK-LABEL: @foo3(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: [[START:%.*]] = call i32 @llvm.start.loop.iterations.i32(i32 8001)
|
|
; CHECK-NEXT: br label [[VECTOR_BODY:%.*]]
|
|
; CHECK: vector.body:
|
|
; CHECK-NEXT: [[LSR_IV14:%.*]] = phi i32* [ [[SCEVGEP15:%.*]], [[VECTOR_BODY]] ], [ [[A:%.*]], [[ENTRY:%.*]] ]
|
|
; CHECK-NEXT: [[LSR_IV11:%.*]] = phi i32* [ [[SCEVGEP12:%.*]], [[VECTOR_BODY]] ], [ [[C:%.*]], [[ENTRY]] ]
|
|
; CHECK-NEXT: [[LSR_IV:%.*]] = phi i32* [ [[SCEVGEP:%.*]], [[VECTOR_BODY]] ], [ [[B:%.*]], [[ENTRY]] ]
|
|
; CHECK-NEXT: [[INDEX:%.*]] = phi i32 [ 0, [[ENTRY]] ], [ [[INDEX_NEXT:%.*]], [[VECTOR_BODY]] ]
|
|
; CHECK-NEXT: [[TMP0:%.*]] = phi i32 [ [[START]], [[ENTRY]] ], [ [[TMP3:%.*]], [[VECTOR_BODY]] ]
|
|
; CHECK-NEXT: [[LSR_IV1416:%.*]] = bitcast i32* [[LSR_IV14]] to <4 x i32>*
|
|
; CHECK-NEXT: [[LSR_IV1113:%.*]] = bitcast i32* [[LSR_IV11]] to <4 x i32>*
|
|
; CHECK-NEXT: [[LSR_IV10:%.*]] = bitcast i32* [[LSR_IV]] to <4 x i32>*
|
|
; CHECK-NEXT: [[BROADCAST_SPLATINSERT:%.*]] = insertelement <4 x i32> undef, i32 [[INDEX]], i32 0
|
|
; CHECK-NEXT: [[BROADCAST_SPLAT:%.*]] = shufflevector <4 x i32> [[BROADCAST_SPLATINSERT]], <4 x i32> undef, <4 x i32> zeroinitializer
|
|
; CHECK-NEXT: [[INDUCTION:%.*]] = add <4 x i32> [[BROADCAST_SPLAT]], <i32 0, i32 1, i32 2, i32 3>
|
|
; CHECK-NEXT: [[TMP1:%.*]] = icmp ugt <4 x i32> [[INDUCTION]], <i32 32002, i32 32002, i32 32002, i32 32002>
|
|
; CHECK-NEXT: [[WIDE_MASKED_LOAD:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* [[LSR_IV10]], i32 4, <4 x i1> [[TMP1]], <4 x i32> undef)
|
|
; CHECK-NEXT: [[WIDE_MASKED_LOAD9:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* [[LSR_IV1113]], i32 4, <4 x i1> [[TMP1]], <4 x i32> undef)
|
|
; CHECK-NEXT: [[TMP2:%.*]] = add nsw <4 x i32> [[WIDE_MASKED_LOAD9]], [[WIDE_MASKED_LOAD]]
|
|
; CHECK-NEXT: call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> [[TMP2]], <4 x i32>* [[LSR_IV1416]], i32 4, <4 x i1> [[TMP1]])
|
|
; CHECK-NEXT: [[INDEX_NEXT]] = add i32 [[INDEX]], 4
|
|
; CHECK-NEXT: [[SCEVGEP]] = getelementptr i32, i32* [[LSR_IV]], i32 4
|
|
; CHECK-NEXT: [[SCEVGEP12]] = getelementptr i32, i32* [[LSR_IV11]], i32 4
|
|
; CHECK-NEXT: [[SCEVGEP15]] = getelementptr i32, i32* [[LSR_IV14]], i32 4
|
|
; CHECK-NEXT: [[TMP3]] = call i32 @llvm.loop.decrement.reg.i32(i32 [[TMP0]], i32 1)
|
|
; CHECK-NEXT: [[TMP4:%.*]] = icmp ne i32 [[TMP3]], 0
|
|
; CHECK-NEXT: br i1 [[TMP4]], label [[VECTOR_BODY]], label [[FOR_COND_CLEANUP:%.*]]
|
|
; CHECK: for.cond.cleanup:
|
|
; CHECK-NEXT: ret void
|
|
;
|
|
entry:
|
|
%start = call i32 @llvm.start.loop.iterations.i32(i32 8001)
|
|
br label %vector.body
|
|
|
|
vector.body:
|
|
%lsr.iv14 = phi i32* [ %scevgep15, %vector.body ], [ %A, %entry ]
|
|
%lsr.iv11 = phi i32* [ %scevgep12, %vector.body ], [ %C, %entry ]
|
|
%lsr.iv = phi i32* [ %scevgep, %vector.body ], [ %B, %entry ]
|
|
%index = phi i32 [ 0, %entry ], [ %index.next, %vector.body ]
|
|
%0 = phi i32 [ %start, %entry ], [ %3, %vector.body ]
|
|
%lsr.iv1416 = bitcast i32* %lsr.iv14 to <4 x i32>*
|
|
%lsr.iv1113 = bitcast i32* %lsr.iv11 to <4 x i32>*
|
|
%lsr.iv10 = bitcast i32* %lsr.iv to <4 x i32>*
|
|
%broadcast.splatinsert = insertelement <4 x i32> undef, i32 %index, i32 0
|
|
%broadcast.splat = shufflevector <4 x i32> %broadcast.splatinsert, <4 x i32> undef, <4 x i32> zeroinitializer
|
|
%induction = add <4 x i32> %broadcast.splat, <i32 0, i32 1, i32 2, i32 3>
|
|
|
|
; UGT here:
|
|
%1 = icmp ugt <4 x i32> %induction, <i32 32002, i32 32002, i32 32002, i32 32002>
|
|
|
|
%wide.masked.load = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv10, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%wide.masked.load9 = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv1113, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%2 = add nsw <4 x i32> %wide.masked.load9, %wide.masked.load
|
|
call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> %2, <4 x i32>* %lsr.iv1416, i32 4, <4 x i1> %1)
|
|
%index.next = add i32 %index, 4
|
|
%scevgep = getelementptr i32, i32* %lsr.iv, i32 4
|
|
%scevgep12 = getelementptr i32, i32* %lsr.iv11, i32 4
|
|
%scevgep15 = getelementptr i32, i32* %lsr.iv14, i32 4
|
|
%3 = call i32 @llvm.loop.decrement.reg.i32(i32 %0, i32 1)
|
|
%4 = icmp ne i32 %3, 0
|
|
br i1 %4, label %vector.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
ret void
|
|
}
|
|
|
|
define dso_local void @foo5(i32* noalias nocapture %A, i32* noalias nocapture readonly %B, i32* noalias nocapture readonly %C, i32* noalias nocapture readnone %D, i32 %N) local_unnamed_addr #0 {
|
|
; CHECK-LABEL: @foo5(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: [[START:%.*]] = call i32 @llvm.start.loop.iterations.i32(i32 8001)
|
|
; CHECK-NEXT: br label [[VECTOR_BODY:%.*]]
|
|
; CHECK: vector.body:
|
|
; CHECK-NEXT: [[LSR_IV14:%.*]] = phi i32* [ [[SCEVGEP15:%.*]], [[VECTOR_BODY]] ], [ [[A:%.*]], [[ENTRY:%.*]] ]
|
|
; CHECK-NEXT: [[LSR_IV11:%.*]] = phi i32* [ [[SCEVGEP12:%.*]], [[VECTOR_BODY]] ], [ [[C:%.*]], [[ENTRY]] ]
|
|
; CHECK-NEXT: [[LSR_IV:%.*]] = phi i32* [ [[SCEVGEP:%.*]], [[VECTOR_BODY]] ], [ [[B:%.*]], [[ENTRY]] ]
|
|
; CHECK-NEXT: [[INDEX:%.*]] = phi i32 [ 0, [[ENTRY]] ], [ [[INDEX_NEXT:%.*]], [[VECTOR_BODY]] ]
|
|
; CHECK-NEXT: [[TMP0:%.*]] = phi i32 [ [[START]], [[ENTRY]] ], [ [[TMP3:%.*]], [[VECTOR_BODY]] ]
|
|
; CHECK-NEXT: [[LSR_IV1416:%.*]] = bitcast i32* [[LSR_IV14]] to <4 x i32>*
|
|
; CHECK-NEXT: [[LSR_IV1113:%.*]] = bitcast i32* [[LSR_IV11]] to <4 x i32>*
|
|
; CHECK-NEXT: [[LSR_IV10:%.*]] = bitcast i32* [[LSR_IV]] to <4 x i32>*
|
|
; CHECK-NEXT: [[BROADCAST_SPLATINSERT:%.*]] = insertelement <4 x i32> undef, i32 [[INDEX]], i32 0
|
|
; CHECK-NEXT: [[BROADCAST_SPLAT:%.*]] = shufflevector <4 x i32> [[BROADCAST_SPLATINSERT]], <4 x i32> undef, <4 x i32> zeroinitializer
|
|
; CHECK-NEXT: [[INDUCTION:%.*]] = add <4 x i32> [[BROADCAST_SPLAT]], <i32 0, i32 1, i32 2, i32 3>
|
|
; CHECK-NEXT: [[TMP1:%.*]] = icmp ult <4 x i32> [[INDUCTION]], <i32 0, i32 3200, i32 32002, i32 32002>
|
|
; CHECK-NEXT: [[WIDE_MASKED_LOAD:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* [[LSR_IV10]], i32 4, <4 x i1> [[TMP1]], <4 x i32> undef)
|
|
; CHECK-NEXT: [[WIDE_MASKED_LOAD9:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* [[LSR_IV1113]], i32 4, <4 x i1> [[TMP1]], <4 x i32> undef)
|
|
; CHECK-NEXT: [[TMP2:%.*]] = add nsw <4 x i32> [[WIDE_MASKED_LOAD9]], [[WIDE_MASKED_LOAD]]
|
|
; CHECK-NEXT: call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> [[TMP2]], <4 x i32>* [[LSR_IV1416]], i32 4, <4 x i1> [[TMP1]])
|
|
; CHECK-NEXT: [[INDEX_NEXT]] = add i32 [[INDEX]], 4
|
|
; CHECK-NEXT: [[SCEVGEP]] = getelementptr i32, i32* [[LSR_IV]], i32 4
|
|
; CHECK-NEXT: [[SCEVGEP12]] = getelementptr i32, i32* [[LSR_IV11]], i32 4
|
|
; CHECK-NEXT: [[SCEVGEP15]] = getelementptr i32, i32* [[LSR_IV14]], i32 4
|
|
; CHECK-NEXT: [[TMP3]] = call i32 @llvm.loop.decrement.reg.i32(i32 [[TMP0]], i32 1)
|
|
; CHECK-NEXT: [[TMP4:%.*]] = icmp ne i32 [[TMP3]], 0
|
|
; CHECK-NEXT: br i1 [[TMP4]], label [[VECTOR_BODY]], label [[FOR_COND_CLEANUP:%.*]]
|
|
; CHECK: for.cond.cleanup:
|
|
; CHECK-NEXT: ret void
|
|
;
|
|
entry:
|
|
%start = call i32 @llvm.start.loop.iterations.i32(i32 8001)
|
|
br label %vector.body
|
|
|
|
vector.body:
|
|
%lsr.iv14 = phi i32* [ %scevgep15, %vector.body ], [ %A, %entry ]
|
|
%lsr.iv11 = phi i32* [ %scevgep12, %vector.body ], [ %C, %entry ]
|
|
%lsr.iv = phi i32* [ %scevgep, %vector.body ], [ %B, %entry ]
|
|
%index = phi i32 [ 0, %entry ], [ %index.next, %vector.body ]
|
|
%0 = phi i32 [ %start, %entry ], [ %3, %vector.body ]
|
|
%lsr.iv1416 = bitcast i32* %lsr.iv14 to <4 x i32>*
|
|
%lsr.iv1113 = bitcast i32* %lsr.iv11 to <4 x i32>*
|
|
%lsr.iv10 = bitcast i32* %lsr.iv to <4 x i32>*
|
|
%broadcast.splatinsert = insertelement <4 x i32> undef, i32 %index, i32 0
|
|
%broadcast.splat = shufflevector <4 x i32> %broadcast.splatinsert, <4 x i32> undef, <4 x i32> zeroinitializer
|
|
%induction = add <4 x i32> %broadcast.splat, <i32 0, i32 1, i32 2, i32 3>
|
|
|
|
; Non-uniform constant vector here. This can't be represented with
|
|
; @llvm.get.active.lane.mask, but let's keep this test as a sanity check:
|
|
%1 = icmp ult <4 x i32> %induction, <i32 0, i32 3200, i32 32002, i32 32002>
|
|
|
|
%wide.masked.load = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv10, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%wide.masked.load9 = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv1113, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%2 = add nsw <4 x i32> %wide.masked.load9, %wide.masked.load
|
|
call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> %2, <4 x i32>* %lsr.iv1416, i32 4, <4 x i1> %1)
|
|
%index.next = add i32 %index, 4
|
|
%scevgep = getelementptr i32, i32* %lsr.iv, i32 4
|
|
%scevgep12 = getelementptr i32, i32* %lsr.iv11, i32 4
|
|
%scevgep15 = getelementptr i32, i32* %lsr.iv14, i32 4
|
|
%3 = call i32 @llvm.loop.decrement.reg.i32(i32 %0, i32 1)
|
|
%4 = icmp ne i32 %3, 0
|
|
br i1 %4, label %vector.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @inconsistent_tripcounts(
|
|
; CHECK: vector.body:
|
|
; CHECK-NOT: @llvm.arm.mve.vctp32
|
|
; CHECK: @llvm.get.active.lane.mask
|
|
; CHECK: ret void
|
|
;
|
|
define dso_local void @inconsistent_tripcounts(i32* noalias nocapture %A, i32* noalias nocapture readonly %B, i32* noalias nocapture readonly %C, i32* noalias nocapture readnone %D, i32 %N) local_unnamed_addr #0 {
|
|
entry:
|
|
%start = call i32 @llvm.start.loop.iterations.i32(i32 8001)
|
|
br label %vector.body
|
|
|
|
vector.body:
|
|
%lsr.iv14 = phi i32* [ %scevgep15, %vector.body ], [ %A, %entry ]
|
|
%lsr.iv11 = phi i32* [ %scevgep12, %vector.body ], [ %C, %entry ]
|
|
%lsr.iv = phi i32* [ %scevgep, %vector.body ], [ %B, %entry ]
|
|
%index = phi i32 [ 0, %entry ], [ %index.next, %vector.body ]
|
|
%0 = phi i32 [ %start, %entry ], [ %3, %vector.body ]
|
|
%lsr.iv1416 = bitcast i32* %lsr.iv14 to <4 x i32>*
|
|
%lsr.iv1113 = bitcast i32* %lsr.iv11 to <4 x i32>*
|
|
%lsr.iv10 = bitcast i32* %lsr.iv to <4 x i32>*
|
|
%broadcast.splatinsert = insertelement <4 x i32> undef, i32 %index, i32 0
|
|
%broadcast.splat = shufflevector <4 x i32> %broadcast.splatinsert, <4 x i32> undef, <4 x i32> zeroinitializer
|
|
%induction = add <4 x i32> %broadcast.splat, <i32 0, i32 1, i32 2, i32 3>
|
|
|
|
; BTC = UINT_MAX, and scalar trip count BTC + 1 would overflow:
|
|
%1 = call <4 x i1> @llvm.get.active.lane.mask.v4i1.i32(i32 %index, i32 4294967295)
|
|
|
|
%wide.masked.load = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv10, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%wide.masked.load9 = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv1113, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%2 = add nsw <4 x i32> %wide.masked.load9, %wide.masked.load
|
|
call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> %2, <4 x i32>* %lsr.iv1416, i32 4, <4 x i1> %1)
|
|
%index.next = add i32 %index, 4
|
|
%scevgep = getelementptr i32, i32* %lsr.iv, i32 4
|
|
%scevgep12 = getelementptr i32, i32* %lsr.iv11, i32 4
|
|
%scevgep15 = getelementptr i32, i32* %lsr.iv14, i32 4
|
|
%3 = call i32 @llvm.loop.decrement.reg.i32(i32 %0, i32 1)
|
|
%4 = icmp ne i32 %3, 0
|
|
br i1 %4, label %vector.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @overflow_in_sub(
|
|
; CHECK: vector.body:
|
|
; CHECK-NOT: @llvm.arm.mve.vctp32
|
|
; CHECK: @llvm.get.active.lane.mask
|
|
; CHECK: ret void
|
|
;
|
|
define dso_local void @overflow_in_sub(i32* noalias nocapture %A, i32* noalias nocapture readonly %B, i32* noalias nocapture readonly %C, i32* noalias nocapture readnone %D, i32 %N) local_unnamed_addr #0 {
|
|
entry:
|
|
%start = call i32 @llvm.start.loop.iterations.i32(i32 1073741824)
|
|
br label %vector.body
|
|
|
|
vector.body:
|
|
%lsr.iv14 = phi i32* [ %scevgep15, %vector.body ], [ %A, %entry ]
|
|
%lsr.iv11 = phi i32* [ %scevgep12, %vector.body ], [ %C, %entry ]
|
|
%lsr.iv = phi i32* [ %scevgep, %vector.body ], [ %B, %entry ]
|
|
%index = phi i32 [ 0, %entry ], [ %index.next, %vector.body ]
|
|
%0 = phi i32 [ %start, %entry ], [ %3, %vector.body ]
|
|
%lsr.iv1416 = bitcast i32* %lsr.iv14 to <4 x i32>*
|
|
%lsr.iv1113 = bitcast i32* %lsr.iv11 to <4 x i32>*
|
|
%lsr.iv10 = bitcast i32* %lsr.iv to <4 x i32>*
|
|
%broadcast.splatinsert = insertelement <4 x i32> undef, i32 %index, i32 0
|
|
%broadcast.splat = shufflevector <4 x i32> %broadcast.splatinsert, <4 x i32> undef, <4 x i32> zeroinitializer
|
|
%induction = add <4 x i32> %broadcast.splat, <i32 0, i32 1, i32 2, i32 3>
|
|
|
|
%1 = call <4 x i1> @llvm.get.active.lane.mask.v4i1.i32(i32 %index, i32 32003)
|
|
|
|
%wide.masked.load = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv10, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%wide.masked.load9 = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv1113, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%2 = add nsw <4 x i32> %wide.masked.load9, %wide.masked.load
|
|
call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> %2, <4 x i32>* %lsr.iv1416, i32 4, <4 x i1> %1)
|
|
%index.next = add i32 %index, 4
|
|
%scevgep = getelementptr i32, i32* %lsr.iv, i32 4
|
|
%scevgep12 = getelementptr i32, i32* %lsr.iv11, i32 4
|
|
%scevgep15 = getelementptr i32, i32* %lsr.iv14, i32 4
|
|
%3 = call i32 @llvm.loop.decrement.reg.i32(i32 %0, i32 1)
|
|
%4 = icmp ne i32 %3, 0
|
|
br i1 %4, label %vector.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
ret void
|
|
}
|
|
|
|
|
|
; CHECK-LABEL: @IV_not_an_induction(
|
|
; CHECK: vector.body:
|
|
; CHECK-NOT: @llvm.arm.mve.vctp32
|
|
; CHECK: @llvm.get.active.lane.mask
|
|
; CHECK: ret void
|
|
;
|
|
define dso_local void @IV_not_an_induction(i32* noalias nocapture %A, i32* noalias nocapture readonly %B, i32* noalias nocapture readonly %C, i32* noalias nocapture readnone %D, i32 %N) local_unnamed_addr #0 {
|
|
entry:
|
|
%start = call i32 @llvm.start.loop.iterations.i32(i32 8001)
|
|
br label %vector.body
|
|
|
|
vector.body:
|
|
%lsr.iv14 = phi i32* [ %scevgep15, %vector.body ], [ %A, %entry ]
|
|
%lsr.iv11 = phi i32* [ %scevgep12, %vector.body ], [ %C, %entry ]
|
|
%lsr.iv = phi i32* [ %scevgep, %vector.body ], [ %B, %entry ]
|
|
%index = phi i32 [ 0, %entry ], [ %index.next, %vector.body ]
|
|
%0 = phi i32 [ %start, %entry ], [ %3, %vector.body ]
|
|
%lsr.iv1416 = bitcast i32* %lsr.iv14 to <4 x i32>*
|
|
%lsr.iv1113 = bitcast i32* %lsr.iv11 to <4 x i32>*
|
|
%lsr.iv10 = bitcast i32* %lsr.iv to <4 x i32>*
|
|
%broadcast.splatinsert = insertelement <4 x i32> undef, i32 %index, i32 0
|
|
%broadcast.splat = shufflevector <4 x i32> %broadcast.splatinsert, <4 x i32> undef, <4 x i32> zeroinitializer
|
|
%induction = add <4 x i32> %broadcast.splat, <i32 0, i32 1, i32 2, i32 3>
|
|
|
|
; The induction variable %N is not an IV:
|
|
%1 = call <4 x i1> @llvm.get.active.lane.mask.v4i1.i32(i32 %N, i32 32003)
|
|
|
|
%wide.masked.load = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv10, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%wide.masked.load9 = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv1113, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%2 = add nsw <4 x i32> %wide.masked.load9, %wide.masked.load
|
|
call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> %2, <4 x i32>* %lsr.iv1416, i32 4, <4 x i1> %1)
|
|
%index.next = add i32 %index, 4
|
|
%scevgep = getelementptr i32, i32* %lsr.iv, i32 4
|
|
%scevgep12 = getelementptr i32, i32* %lsr.iv11, i32 4
|
|
%scevgep15 = getelementptr i32, i32* %lsr.iv14, i32 4
|
|
%3 = call i32 @llvm.loop.decrement.reg.i32(i32 %0, i32 1)
|
|
%4 = icmp ne i32 %3, 0
|
|
br i1 %4, label %vector.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @IV_wrong_step(
|
|
; CHECK: vector.body:
|
|
; CHECK-NOT: @llvm.arm.mve.vctp32
|
|
; CHECK: @llvm.get.active.lane.mask
|
|
; CHECK: ret void
|
|
;
|
|
define dso_local void @IV_wrong_step(i32* noalias nocapture %A, i32* noalias nocapture readonly %B, i32* noalias nocapture readonly %C, i32* noalias nocapture readnone %D, i32 %N) local_unnamed_addr #0 {
|
|
entry:
|
|
%start = call i32 @llvm.start.loop.iterations.i32(i32 8001)
|
|
br label %vector.body
|
|
|
|
vector.body:
|
|
%lsr.iv14 = phi i32* [ %scevgep15, %vector.body ], [ %A, %entry ]
|
|
%lsr.iv11 = phi i32* [ %scevgep12, %vector.body ], [ %C, %entry ]
|
|
%lsr.iv = phi i32* [ %scevgep, %vector.body ], [ %B, %entry ]
|
|
%index = phi i32 [ 0, %entry ], [ %index.next, %vector.body ]
|
|
%0 = phi i32 [ %start, %entry ], [ %3, %vector.body ]
|
|
%lsr.iv1416 = bitcast i32* %lsr.iv14 to <4 x i32>*
|
|
%lsr.iv1113 = bitcast i32* %lsr.iv11 to <4 x i32>*
|
|
%lsr.iv10 = bitcast i32* %lsr.iv to <4 x i32>*
|
|
%broadcast.splatinsert = insertelement <4 x i32> undef, i32 %index, i32 0
|
|
%broadcast.splat = shufflevector <4 x i32> %broadcast.splatinsert, <4 x i32> undef, <4 x i32> zeroinitializer
|
|
%induction = add <4 x i32> %broadcast.splat, <i32 0, i32 1, i32 2, i32 3>
|
|
|
|
%1 = call <4 x i1> @llvm.get.active.lane.mask.v4i1.i32(i32 %index, i32 32003)
|
|
|
|
%wide.masked.load = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv10, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%wide.masked.load9 = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv1113, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%2 = add nsw <4 x i32> %wide.masked.load9, %wide.masked.load
|
|
call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> %2, <4 x i32>* %lsr.iv1416, i32 4, <4 x i1> %1)
|
|
|
|
; %index is incremented with 3 and not 4, which is the vectorisation factor
|
|
; that we expect here:
|
|
%index.next = add i32 %index, 3
|
|
|
|
%scevgep = getelementptr i32, i32* %lsr.iv, i32 4
|
|
%scevgep12 = getelementptr i32, i32* %lsr.iv11, i32 4
|
|
%scevgep15 = getelementptr i32, i32* %lsr.iv14, i32 4
|
|
%3 = call i32 @llvm.loop.decrement.reg.i32(i32 %0, i32 1)
|
|
%4 = icmp ne i32 %3, 0
|
|
br i1 %4, label %vector.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @IV_step_not_constant(
|
|
; CHECK: vector.body:
|
|
; CHECK-NOT: @llvm.arm.mve.vctp32
|
|
; CHECK: @llvm.get.active.lane.mask
|
|
; CHECK: ret void
|
|
;
|
|
define dso_local void @IV_step_not_constant(i32* noalias nocapture %A, i32* noalias nocapture readonly %B, i32* noalias nocapture readonly %C, i32* noalias nocapture readnone %D, i32 %N) local_unnamed_addr #0 {
|
|
entry:
|
|
%start = call i32 @llvm.start.loop.iterations.i32(i32 8001)
|
|
br label %vector.body
|
|
|
|
vector.body:
|
|
%lsr.iv14 = phi i32* [ %scevgep15, %vector.body ], [ %A, %entry ]
|
|
%lsr.iv11 = phi i32* [ %scevgep12, %vector.body ], [ %C, %entry ]
|
|
%lsr.iv = phi i32* [ %scevgep, %vector.body ], [ %B, %entry ]
|
|
%index = phi i32 [ 0, %entry ], [ %index.next, %vector.body ]
|
|
%0 = phi i32 [ %start, %entry ], [ %3, %vector.body ]
|
|
%lsr.iv1416 = bitcast i32* %lsr.iv14 to <4 x i32>*
|
|
%lsr.iv1113 = bitcast i32* %lsr.iv11 to <4 x i32>*
|
|
%lsr.iv10 = bitcast i32* %lsr.iv to <4 x i32>*
|
|
%broadcast.splatinsert = insertelement <4 x i32> undef, i32 %index, i32 0
|
|
%broadcast.splat = shufflevector <4 x i32> %broadcast.splatinsert, <4 x i32> undef, <4 x i32> zeroinitializer
|
|
%induction = add <4 x i32> %broadcast.splat, <i32 0, i32 1, i32 2, i32 3>
|
|
%1 = call <4 x i1> @llvm.get.active.lane.mask.v4i1.i32(i32 %index, i32 32003)
|
|
%wide.masked.load = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv10, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%wide.masked.load9 = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv1113, i32 4, <4 x i1> %1, <4 x i32> undef)
|
|
%2 = add nsw <4 x i32> %wide.masked.load9, %wide.masked.load
|
|
call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> %2, <4 x i32>* %lsr.iv1416, i32 4, <4 x i1> %1)
|
|
|
|
; %index is incremented with some runtime value, i.e. not a constant:
|
|
%index.next = add i32 %index, %N
|
|
|
|
%scevgep = getelementptr i32, i32* %lsr.iv, i32 4
|
|
%scevgep12 = getelementptr i32, i32* %lsr.iv11, i32 4
|
|
%scevgep15 = getelementptr i32, i32* %lsr.iv14, i32 4
|
|
%3 = call i32 @llvm.loop.decrement.reg.i32(i32 %0, i32 1)
|
|
%4 = icmp ne i32 %3, 0
|
|
br i1 %4, label %vector.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
ret void
|
|
}
|
|
|
|
; CHECK-LABEL: @outerloop_phi(
|
|
; CHECK: vector.body:
|
|
; CHECK-NOT: @llvm.arm.mve.vctp32
|
|
; CHECK: @llvm.get.active.lane.mask
|
|
; CHECK: ret void
|
|
;
|
|
define dso_local void @outerloop_phi(i32* noalias nocapture %A, i32* noalias nocapture readonly %B, i32* noalias nocapture readonly %C, i32 %N) local_unnamed_addr #0 {
|
|
entry:
|
|
%cmp24 = icmp eq i32 %N, 0
|
|
br i1 %cmp24, label %for.cond.cleanup, label %vector.ph.preheader
|
|
|
|
vector.ph.preheader: ; preds = %entry
|
|
br label %vector.ph
|
|
|
|
vector.ph: ; preds = %vector.ph.preheader, %for.cond.cleanup3
|
|
%lsr.iv36 = phi i32* [ %B, %vector.ph.preheader ], [ %scevgep37, %for.cond.cleanup3 ]
|
|
%lsr.iv31 = phi i32* [ %C, %vector.ph.preheader ], [ %scevgep32, %for.cond.cleanup3 ]
|
|
%lsr.iv = phi i32* [ %A, %vector.ph.preheader ], [ %scevgep, %for.cond.cleanup3 ]
|
|
%j.025 = phi i32 [ %inc11, %for.cond.cleanup3 ], [ 0, %vector.ph.preheader ]
|
|
%start = call i32 @llvm.start.loop.iterations.i32(i32 1025)
|
|
br label %vector.body
|
|
|
|
vector.body: ; preds = %vector.body, %vector.ph
|
|
%lsr.iv38 = phi i32* [ %scevgep39, %vector.body ], [ %lsr.iv36, %vector.ph ]
|
|
%lsr.iv33 = phi i32* [ %scevgep34, %vector.body ], [ %lsr.iv31, %vector.ph ]
|
|
%lsr.iv28 = phi i32* [ %scevgep29, %vector.body ], [ %lsr.iv, %vector.ph ]
|
|
%index = phi i32 [ 0, %vector.ph ], [ %index.next, %vector.body ]
|
|
%0 = phi i32 [ %start, %vector.ph ], [ %2, %vector.body ]
|
|
%lsr.iv3840 = bitcast i32* %lsr.iv38 to <4 x i32>*
|
|
%lsr.iv3335 = bitcast i32* %lsr.iv33 to <4 x i32>*
|
|
%lsr.iv2830 = bitcast i32* %lsr.iv28 to <4 x i32>*
|
|
|
|
; It's using %j.025, the induction variable from its outer loop:
|
|
%active.lane.mask = call <4 x i1> @llvm.get.active.lane.mask.v4i1.i32(i32 %j.025, i32 4096)
|
|
|
|
%wide.masked.load = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv3840, i32 4, <4 x i1> %active.lane.mask, <4 x i32> undef)
|
|
%wide.masked.load27 = call <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>* %lsr.iv3335, i32 4, <4 x i1> %active.lane.mask, <4 x i32> undef)
|
|
%1 = add nsw <4 x i32> %wide.masked.load27, %wide.masked.load
|
|
call void @llvm.masked.store.v4i32.p0v4i32(<4 x i32> %1, <4 x i32>* %lsr.iv2830, i32 4, <4 x i1> %active.lane.mask)
|
|
%index.next = add i32 %index, 4
|
|
%scevgep29 = getelementptr i32, i32* %lsr.iv28, i32 4
|
|
%scevgep34 = getelementptr i32, i32* %lsr.iv33, i32 4
|
|
%scevgep39 = getelementptr i32, i32* %lsr.iv38, i32 4
|
|
%2 = call i32 @llvm.loop.decrement.reg.i32(i32 %0, i32 1)
|
|
%3 = icmp ne i32 %2, 0
|
|
br i1 %3, label %vector.body, label %for.cond.cleanup3
|
|
|
|
for.cond.cleanup: ; preds = %for.cond.cleanup3, %entry
|
|
ret void
|
|
|
|
for.cond.cleanup3: ; preds = %vector.body
|
|
%inc11 = add nuw i32 %j.025, 1
|
|
%scevgep = getelementptr i32, i32* %lsr.iv, i32 1
|
|
%scevgep32 = getelementptr i32, i32* %lsr.iv31, i32 1
|
|
%scevgep37 = getelementptr i32, i32* %lsr.iv36, i32 1
|
|
%exitcond26 = icmp eq i32 %inc11, %N
|
|
br i1 %exitcond26, label %for.cond.cleanup, label %vector.ph
|
|
}
|
|
|
|
|
|
declare <4 x i32> @llvm.masked.load.v4i32.p0v4i32(<4 x i32>*, i32 immarg, <4 x i1>, <4 x i32>) #1
|
|
declare void @llvm.masked.store.v4i32.p0v4i32(<4 x i32>, <4 x i32>*, i32 immarg, <4 x i1>) #2
|
|
declare i32 @llvm.loop.decrement.reg.i32(i32 , i32 )
|
|
declare i32 @llvm.start.loop.iterations.i32(i32)
|
|
declare <4 x i1> @llvm.get.active.lane.mask.v4i1.i32(i32, i32)
|