mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-25 12:12:47 +01:00
[Hexagon] Unmasked and masked load pair to dame bae -? one load and selects
This commit is contained in:
parent
543196c49f
commit
5e52d75c0d
@ -521,11 +521,6 @@ auto AlignVectors::createAddressGroups() -> bool {
|
||||
return !llvm::any_of(
|
||||
G.second, [&](auto &I) { return HVC.HST.isTypeForHVX(I.ValTy); });
|
||||
});
|
||||
// Remove groups where everything is properly aligned.
|
||||
erase_if(AddrGroups, [&](auto &G) {
|
||||
return llvm::all_of(G.second,
|
||||
[&](auto &I) { return I.HaveAlign >= I.NeedAlign; });
|
||||
});
|
||||
|
||||
return !AddrGroups.empty();
|
||||
}
|
||||
@ -1382,6 +1377,11 @@ auto HexagonVectorCombine::isSafeToMoveBeforeInBB(const Instruction &In,
|
||||
const Instruction &I = *It;
|
||||
if (llvm::is_contained(Ignore, &I))
|
||||
continue;
|
||||
// assume intrinsic can be ignored
|
||||
if (auto *II = dyn_cast<IntrinsicInst>(&I)) {
|
||||
if (II->getIntrinsicID() == Intrinsic::assume)
|
||||
continue;
|
||||
}
|
||||
// Parts based on isSafeToMoveBefore from CoveMoverUtils.cpp.
|
||||
if (I.mayThrow())
|
||||
return false;
|
||||
|
38
test/CodeGen/Hexagon/vcombine_zero_diff_ptrs.ll
Normal file
38
test/CodeGen/Hexagon/vcombine_zero_diff_ptrs.ll
Normal file
@ -0,0 +1,38 @@
|
||||
; RUN: opt -march=hexagon -hexagon-vc -S < %s | FileCheck %s
|
||||
|
||||
; Test that the HexagonVectorCombine pass identifies instruction
|
||||
; pairs whose difference in pointers is zero. This creates a vector
|
||||
; load to handle masked and unmasked loads that have no base
|
||||
; pointer difference and replaces the masked and unmasked loads
|
||||
; with selects
|
||||
|
||||
; CHECK: select
|
||||
|
||||
target datalayout = "e-m:e-p:32:32:32-a:0-n16:32-i64:64:64-i32:32:32-i16:16:16-i1:8:8-f32:32:32-f64:64:64-v32:32:32-v64:64:64-v512:512:512-v1024:1024:1024-v2048:2048:2048"
|
||||
target triple = "hexagon"
|
||||
|
||||
define dllexport void @f0(i8** %a0) local_unnamed_addr #0 {
|
||||
b0:
|
||||
%v0 = load i8*, i8** %a0, align 4
|
||||
%v1 = getelementptr i8, i8* %v0, i32 1794
|
||||
%v2 = bitcast i8* %v1 to <64 x i16>*
|
||||
call void @llvm.assume(i1 true) [ "align"(i8* %v0, i32 128) ]
|
||||
%v3 = load <64 x i16>, <64 x i16>* %v2, align 128
|
||||
%v4 = add <64 x i16> %v3, %v3
|
||||
call void @llvm.assume(i1 true) [ "align"(i8* %v0, i32 128) ]
|
||||
%v5 = tail call <64 x i16> @llvm.masked.load.v64i16.p0v64i16(<64 x i16>* %v2, i32 128, <64 x i1> <i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true, i1 false, i1 true>, <64 x i16> undef)
|
||||
call void @llvm.assume(i1 true) [ "align"(i8* %v0, i32 128) ]
|
||||
%v6 = add <64 x i16> %v4, %v5
|
||||
store <64 x i16> %v6, <64 x i16>* %v2, align 128
|
||||
ret void
|
||||
}
|
||||
|
||||
; Function Attrs: nofree nosync nounwind willreturn
|
||||
declare void @llvm.assume(i1 noundef) #1
|
||||
|
||||
; Function Attrs: argmemonly nofree nosync nounwind readonly willreturn
|
||||
declare <64 x i16> @llvm.masked.load.v64i16.p0v64i16(<64 x i16>*, i32 immarg, <64 x i1>, <64 x i16>) #2
|
||||
|
||||
attributes #0 = { "target-features"="+hvxv68,+hvx-length128b,+hvxv68,+hvx-length128b" }
|
||||
attributes #1 = { nofree nosync nounwind willreturn }
|
||||
attributes #2 = { argmemonly nofree nosync nounwind readonly willreturn }
|
Loading…
Reference in New Issue
Block a user