mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-24 19:52:54 +01:00
c8018d2237
Make the kind of cost explicit throughout the cost model which, apart from making the cost clear, will allow the generic parts to calculate better costs. It will also allow some backends to approximate and correlate the different costs if they wish. Another benefit is that it will also help simplify the cost model around immediate and intrinsic costs, where we currently have multiple APIs. RFC thread: http://lists.llvm.org/pipermail/llvm-dev/2020-April/141263.html Differential Revision: https://reviews.llvm.org/D79002
834 lines
36 KiB
C++
834 lines
36 KiB
C++
//===- SpeculateAroundPHIs.cpp --------------------------------------------===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "llvm/Transforms/Scalar/SpeculateAroundPHIs.h"
|
|
#include "llvm/ADT/PostOrderIterator.h"
|
|
#include "llvm/ADT/Sequence.h"
|
|
#include "llvm/ADT/SetVector.h"
|
|
#include "llvm/ADT/Statistic.h"
|
|
#include "llvm/Analysis/TargetTransformInfo.h"
|
|
#include "llvm/Analysis/ValueTracking.h"
|
|
#include "llvm/IR/BasicBlock.h"
|
|
#include "llvm/IR/IRBuilder.h"
|
|
#include "llvm/IR/Instructions.h"
|
|
#include "llvm/IR/IntrinsicInst.h"
|
|
#include "llvm/Support/Debug.h"
|
|
#include "llvm/Transforms/Utils/BasicBlockUtils.h"
|
|
|
|
using namespace llvm;
|
|
|
|
#define DEBUG_TYPE "spec-phis"
|
|
|
|
STATISTIC(NumPHIsSpeculated, "Number of PHI nodes we speculated around");
|
|
STATISTIC(NumEdgesSplit,
|
|
"Number of critical edges which were split for speculation");
|
|
STATISTIC(NumSpeculatedInstructions,
|
|
"Number of instructions we speculated around the PHI nodes");
|
|
STATISTIC(NumNewRedundantInstructions,
|
|
"Number of new, redundant instructions inserted");
|
|
|
|
/// Check whether speculating the users of a PHI node around the PHI
|
|
/// will be safe.
|
|
///
|
|
/// This checks both that all of the users are safe and also that all of their
|
|
/// operands are either recursively safe or already available along an incoming
|
|
/// edge to the PHI.
|
|
///
|
|
/// This routine caches both all the safe nodes explored in `PotentialSpecSet`
|
|
/// and the chain of nodes that definitively reach any unsafe node in
|
|
/// `UnsafeSet`. By preserving these between repeated calls to this routine for
|
|
/// PHIs in the same basic block, the exploration here can be reused. However,
|
|
/// these caches must no be reused for PHIs in a different basic block as they
|
|
/// reflect what is available along incoming edges.
|
|
static bool
|
|
isSafeToSpeculatePHIUsers(PHINode &PN, DominatorTree &DT,
|
|
SmallPtrSetImpl<Instruction *> &PotentialSpecSet,
|
|
SmallPtrSetImpl<Instruction *> &UnsafeSet) {
|
|
auto *PhiBB = PN.getParent();
|
|
SmallPtrSet<Instruction *, 4> Visited;
|
|
SmallVector<std::pair<Instruction *, User::value_op_iterator>, 16> DFSStack;
|
|
|
|
// Walk each user of the PHI node.
|
|
for (Use &U : PN.uses()) {
|
|
auto *UI = cast<Instruction>(U.getUser());
|
|
|
|
// Ensure the use post-dominates the PHI node. This ensures that, in the
|
|
// absence of unwinding, the use will actually be reached.
|
|
// FIXME: We use a blunt hammer of requiring them to be in the same basic
|
|
// block. We should consider using actual post-dominance here in the
|
|
// future.
|
|
if (UI->getParent() != PhiBB) {
|
|
LLVM_DEBUG(dbgs() << " Unsafe: use in a different BB: " << *UI << "\n");
|
|
return false;
|
|
}
|
|
|
|
if (const auto *CS = dyn_cast<CallBase>(UI)) {
|
|
if (CS->isConvergent() || CS->cannotDuplicate()) {
|
|
LLVM_DEBUG(dbgs() << " Unsafe: convergent "
|
|
"callsite cannot de duplicated: " << *UI << '\n');
|
|
return false;
|
|
}
|
|
}
|
|
|
|
// FIXME: This check is much too conservative. We're not going to move these
|
|
// instructions onto new dynamic paths through the program unless there is
|
|
// a call instruction between the use and the PHI node. And memory isn't
|
|
// changing unless there is a store in that same sequence. We should
|
|
// probably change this to do at least a limited scan of the intervening
|
|
// instructions and allow handling stores in easily proven safe cases.
|
|
if (mayBeMemoryDependent(*UI)) {
|
|
LLVM_DEBUG(dbgs() << " Unsafe: can't speculate use: " << *UI << "\n");
|
|
return false;
|
|
}
|
|
|
|
// Now do a depth-first search of everything these users depend on to make
|
|
// sure they are transitively safe. This is a depth-first search, but we
|
|
// check nodes in preorder to minimize the amount of checking.
|
|
Visited.insert(UI);
|
|
DFSStack.push_back({UI, UI->value_op_begin()});
|
|
do {
|
|
User::value_op_iterator OpIt;
|
|
std::tie(UI, OpIt) = DFSStack.pop_back_val();
|
|
|
|
while (OpIt != UI->value_op_end()) {
|
|
auto *OpI = dyn_cast<Instruction>(*OpIt);
|
|
// Increment to the next operand for whenever we continue.
|
|
++OpIt;
|
|
// No need to visit non-instructions, which can't form dependencies.
|
|
if (!OpI)
|
|
continue;
|
|
|
|
// Now do the main pre-order checks that this operand is a viable
|
|
// dependency of something we want to speculate.
|
|
|
|
// First do a few checks for instructions that won't require
|
|
// speculation at all because they are trivially available on the
|
|
// incoming edge (either through dominance or through an incoming value
|
|
// to a PHI).
|
|
//
|
|
// The cases in the current block will be trivially dominated by the
|
|
// edge.
|
|
auto *ParentBB = OpI->getParent();
|
|
if (ParentBB == PhiBB) {
|
|
if (isa<PHINode>(OpI)) {
|
|
// We can trivially map through phi nodes in the same block.
|
|
continue;
|
|
}
|
|
} else if (DT.dominates(ParentBB, PhiBB)) {
|
|
// Instructions from dominating blocks are already available.
|
|
continue;
|
|
}
|
|
|
|
// Once we know that we're considering speculating the operand, check
|
|
// if we've already explored this subgraph and found it to be safe.
|
|
if (PotentialSpecSet.count(OpI))
|
|
continue;
|
|
|
|
// If we've already explored this subgraph and found it unsafe, bail.
|
|
// If when we directly test whether this is safe it fails, bail.
|
|
if (UnsafeSet.count(OpI) || ParentBB != PhiBB ||
|
|
mayBeMemoryDependent(*OpI)) {
|
|
LLVM_DEBUG(dbgs() << " Unsafe: can't speculate transitive use: "
|
|
<< *OpI << "\n");
|
|
// Record the stack of instructions which reach this node as unsafe
|
|
// so we prune subsequent searches.
|
|
UnsafeSet.insert(OpI);
|
|
for (auto &StackPair : DFSStack) {
|
|
Instruction *I = StackPair.first;
|
|
UnsafeSet.insert(I);
|
|
}
|
|
return false;
|
|
}
|
|
|
|
// Skip any operands we're already recursively checking.
|
|
if (!Visited.insert(OpI).second)
|
|
continue;
|
|
|
|
// Push onto the stack and descend. We can directly continue this
|
|
// loop when ascending.
|
|
DFSStack.push_back({UI, OpIt});
|
|
UI = OpI;
|
|
OpIt = OpI->value_op_begin();
|
|
}
|
|
|
|
// This node and all its operands are safe. Go ahead and cache that for
|
|
// reuse later.
|
|
PotentialSpecSet.insert(UI);
|
|
|
|
// Continue with the next node on the stack.
|
|
} while (!DFSStack.empty());
|
|
}
|
|
|
|
#ifndef NDEBUG
|
|
// Every visited operand should have been marked as safe for speculation at
|
|
// this point. Verify this and return success.
|
|
for (auto *I : Visited)
|
|
assert(PotentialSpecSet.count(I) &&
|
|
"Failed to mark a visited instruction as safe!");
|
|
#endif
|
|
return true;
|
|
}
|
|
|
|
/// Check whether, in isolation, a given PHI node is both safe and profitable
|
|
/// to speculate users around.
|
|
///
|
|
/// This handles checking whether there are any constant operands to a PHI
|
|
/// which could represent a useful speculation candidate, whether the users of
|
|
/// the PHI are safe to speculate including all their transitive dependencies,
|
|
/// and whether after speculation there will be some cost savings (profit) to
|
|
/// folding the operands into the users of the PHI node. Returns true if both
|
|
/// safe and profitable with relevant cost savings updated in the map and with
|
|
/// an update to the `PotentialSpecSet`. Returns false if either safety or
|
|
/// profitability are absent. Some new entries may be made to the
|
|
/// `PotentialSpecSet` even when this routine returns false, but they remain
|
|
/// conservatively correct.
|
|
///
|
|
/// The profitability check here is a local one, but it checks this in an
|
|
/// interesting way. Beyond checking that the total cost of materializing the
|
|
/// constants will be less than the cost of folding them into their users, it
|
|
/// also checks that no one incoming constant will have a higher cost when
|
|
/// folded into its users rather than materialized. This higher cost could
|
|
/// result in a dynamic *path* that is more expensive even when the total cost
|
|
/// is lower. Currently, all of the interesting cases where this optimization
|
|
/// should fire are ones where it is a no-loss operation in this sense. If we
|
|
/// ever want to be more aggressive here, we would need to balance the
|
|
/// different incoming edges' cost by looking at their respective
|
|
/// probabilities.
|
|
static bool isSafeAndProfitableToSpeculateAroundPHI(
|
|
PHINode &PN, SmallDenseMap<PHINode *, int, 16> &CostSavingsMap,
|
|
SmallPtrSetImpl<Instruction *> &PotentialSpecSet,
|
|
SmallPtrSetImpl<Instruction *> &UnsafeSet, DominatorTree &DT,
|
|
TargetTransformInfo &TTI) {
|
|
// First see whether there is any cost savings to speculating around this
|
|
// PHI, and build up a map of the constant inputs to how many times they
|
|
// occur.
|
|
bool NonFreeMat = false;
|
|
struct CostsAndCount {
|
|
int MatCost = TargetTransformInfo::TCC_Free;
|
|
int FoldedCost = TargetTransformInfo::TCC_Free;
|
|
int Count = 0;
|
|
};
|
|
SmallDenseMap<ConstantInt *, CostsAndCount, 16> CostsAndCounts;
|
|
SmallPtrSet<BasicBlock *, 16> IncomingConstantBlocks;
|
|
for (int i : llvm::seq<int>(0, PN.getNumIncomingValues())) {
|
|
auto *IncomingC = dyn_cast<ConstantInt>(PN.getIncomingValue(i));
|
|
if (!IncomingC)
|
|
continue;
|
|
|
|
// Only visit each incoming edge with a constant input once.
|
|
if (!IncomingConstantBlocks.insert(PN.getIncomingBlock(i)).second)
|
|
continue;
|
|
|
|
auto InsertResult = CostsAndCounts.insert({IncomingC, {}});
|
|
// Count how many edges share a given incoming costant.
|
|
++InsertResult.first->second.Count;
|
|
// Only compute the cost the first time we see a particular constant.
|
|
if (!InsertResult.second)
|
|
continue;
|
|
|
|
int &MatCost = InsertResult.first->second.MatCost;
|
|
MatCost = TTI.getIntImmCost(IncomingC->getValue(), IncomingC->getType(),
|
|
TargetTransformInfo::TCK_SizeAndLatency);
|
|
NonFreeMat |= MatCost != TTI.TCC_Free;
|
|
}
|
|
if (!NonFreeMat) {
|
|
LLVM_DEBUG(dbgs() << " Free: " << PN << "\n");
|
|
// No profit in free materialization.
|
|
return false;
|
|
}
|
|
|
|
// Now check that the uses of this PHI can actually be speculated,
|
|
// otherwise we'll still have to materialize the PHI value.
|
|
if (!isSafeToSpeculatePHIUsers(PN, DT, PotentialSpecSet, UnsafeSet)) {
|
|
LLVM_DEBUG(dbgs() << " Unsafe PHI: " << PN << "\n");
|
|
return false;
|
|
}
|
|
|
|
// Compute how much (if any) savings are available by speculating around this
|
|
// PHI.
|
|
for (Use &U : PN.uses()) {
|
|
auto *UserI = cast<Instruction>(U.getUser());
|
|
// Now check whether there is any savings to folding the incoming constants
|
|
// into this use.
|
|
unsigned Idx = U.getOperandNo();
|
|
|
|
// If we have a binary operator that is commutative, an actual constant
|
|
// operand would end up on the RHS, so pretend the use of the PHI is on the
|
|
// RHS.
|
|
//
|
|
// Technically, this is a bit weird if *both* operands are PHIs we're
|
|
// speculating. But if that is the case, giving an "optimistic" cost isn't
|
|
// a bad thing because after speculation it will constant fold. And
|
|
// moreover, such cases should likely have been constant folded already by
|
|
// some other pass, so we shouldn't worry about "modeling" them terribly
|
|
// accurately here. Similarly, if the other operand is a constant, it still
|
|
// seems fine to be "optimistic" in our cost modeling, because when the
|
|
// incoming operand from the PHI node is also a constant, we will end up
|
|
// constant folding.
|
|
if (UserI->isBinaryOp() && UserI->isCommutative() && Idx != 1)
|
|
// Assume we will commute the constant to the RHS to be canonical.
|
|
Idx = 1;
|
|
|
|
// Get the intrinsic ID if this user is an intrinsic.
|
|
Intrinsic::ID IID = Intrinsic::not_intrinsic;
|
|
if (auto *UserII = dyn_cast<IntrinsicInst>(UserI))
|
|
IID = UserII->getIntrinsicID();
|
|
|
|
for (auto &IncomingConstantAndCostsAndCount : CostsAndCounts) {
|
|
ConstantInt *IncomingC = IncomingConstantAndCostsAndCount.first;
|
|
int MatCost = IncomingConstantAndCostsAndCount.second.MatCost;
|
|
int &FoldedCost = IncomingConstantAndCostsAndCount.second.FoldedCost;
|
|
if (IID)
|
|
FoldedCost +=
|
|
TTI.getIntImmCostIntrin(IID, Idx, IncomingC->getValue(),
|
|
IncomingC->getType(),
|
|
TargetTransformInfo::TCK_SizeAndLatency);
|
|
else
|
|
FoldedCost +=
|
|
TTI.getIntImmCostInst(UserI->getOpcode(), Idx,
|
|
IncomingC->getValue(), IncomingC->getType(),
|
|
TargetTransformInfo::TCK_SizeAndLatency);
|
|
|
|
// If we accumulate more folded cost for this incoming constant than
|
|
// materialized cost, then we'll regress any edge with this constant so
|
|
// just bail. We're only interested in cases where folding the incoming
|
|
// constants is at least break-even on all paths.
|
|
if (FoldedCost > MatCost) {
|
|
LLVM_DEBUG(dbgs() << " Not profitable to fold imm: " << *IncomingC
|
|
<< "\n"
|
|
" Materializing cost: "
|
|
<< MatCost
|
|
<< "\n"
|
|
" Accumulated folded cost: "
|
|
<< FoldedCost << "\n");
|
|
return false;
|
|
}
|
|
}
|
|
}
|
|
|
|
// Compute the total cost savings afforded by this PHI node.
|
|
int TotalMatCost = TTI.TCC_Free, TotalFoldedCost = TTI.TCC_Free;
|
|
for (auto IncomingConstantAndCostsAndCount : CostsAndCounts) {
|
|
int MatCost = IncomingConstantAndCostsAndCount.second.MatCost;
|
|
int FoldedCost = IncomingConstantAndCostsAndCount.second.FoldedCost;
|
|
int Count = IncomingConstantAndCostsAndCount.second.Count;
|
|
|
|
TotalMatCost += MatCost * Count;
|
|
TotalFoldedCost += FoldedCost * Count;
|
|
}
|
|
assert(TotalFoldedCost <= TotalMatCost && "If each constant's folded cost is "
|
|
"less that its materialized cost, "
|
|
"the sum must be as well.");
|
|
|
|
LLVM_DEBUG(dbgs() << " Cost savings " << (TotalMatCost - TotalFoldedCost)
|
|
<< ": " << PN << "\n");
|
|
CostSavingsMap[&PN] = TotalMatCost - TotalFoldedCost;
|
|
return true;
|
|
}
|
|
|
|
/// Simple helper to walk all the users of a list of phis depth first, and call
|
|
/// a visit function on each one in post-order.
|
|
///
|
|
/// All of the PHIs should be in the same basic block, and this is primarily
|
|
/// used to make a single depth-first walk across their collective users
|
|
/// without revisiting any subgraphs. Callers should provide a fast, idempotent
|
|
/// callable to test whether a node has been visited and the more important
|
|
/// callable to actually visit a particular node.
|
|
///
|
|
/// Depth-first and postorder here refer to the *operand* graph -- we start
|
|
/// from a collection of users of PHI nodes and walk "up" the operands
|
|
/// depth-first.
|
|
template <typename IsVisitedT, typename VisitT>
|
|
static void visitPHIUsersAndDepsInPostOrder(ArrayRef<PHINode *> PNs,
|
|
IsVisitedT IsVisited,
|
|
VisitT Visit) {
|
|
SmallVector<std::pair<Instruction *, User::value_op_iterator>, 16> DFSStack;
|
|
for (auto *PN : PNs)
|
|
for (Use &U : PN->uses()) {
|
|
auto *UI = cast<Instruction>(U.getUser());
|
|
if (IsVisited(UI))
|
|
// Already visited this user, continue across the roots.
|
|
continue;
|
|
|
|
// Otherwise, walk the operand graph depth-first and visit each
|
|
// dependency in postorder.
|
|
DFSStack.push_back({UI, UI->value_op_begin()});
|
|
do {
|
|
User::value_op_iterator OpIt;
|
|
std::tie(UI, OpIt) = DFSStack.pop_back_val();
|
|
while (OpIt != UI->value_op_end()) {
|
|
auto *OpI = dyn_cast<Instruction>(*OpIt);
|
|
// Increment to the next operand for whenever we continue.
|
|
++OpIt;
|
|
// No need to visit non-instructions, which can't form dependencies,
|
|
// or instructions outside of our potential dependency set that we
|
|
// were given. Finally, if we've already visited the node, continue
|
|
// to the next.
|
|
if (!OpI || IsVisited(OpI))
|
|
continue;
|
|
|
|
// Push onto the stack and descend. We can directly continue this
|
|
// loop when ascending.
|
|
DFSStack.push_back({UI, OpIt});
|
|
UI = OpI;
|
|
OpIt = OpI->value_op_begin();
|
|
}
|
|
|
|
// Finished visiting children, visit this node.
|
|
assert(!IsVisited(UI) && "Should not have already visited a node!");
|
|
Visit(UI);
|
|
} while (!DFSStack.empty());
|
|
}
|
|
}
|
|
|
|
/// Find profitable PHIs to speculate.
|
|
///
|
|
/// For a PHI node to be profitable, we need the cost of speculating its users
|
|
/// (and their dependencies) to not exceed the savings of folding the PHI's
|
|
/// constant operands into the speculated users.
|
|
///
|
|
/// Computing this is surprisingly challenging. Because users of two different
|
|
/// PHI nodes can depend on each other or on common other instructions, it may
|
|
/// be profitable to speculate two PHI nodes together even though neither one
|
|
/// in isolation is profitable. The straightforward way to find all the
|
|
/// profitable PHIs would be to check each combination of PHIs' cost, but this
|
|
/// is exponential in complexity.
|
|
///
|
|
/// Even if we assume that we only care about cases where we can consider each
|
|
/// PHI node in isolation (rather than considering cases where none are
|
|
/// profitable in isolation but some subset are profitable as a set), we still
|
|
/// have a challenge. The obvious way to find all individually profitable PHIs
|
|
/// is to iterate until reaching a fixed point, but this will be quadratic in
|
|
/// complexity. =/
|
|
///
|
|
/// This code currently uses a linear-to-compute order for a greedy approach.
|
|
/// It won't find cases where a set of PHIs must be considered together, but it
|
|
/// handles most cases of order dependence without quadratic iteration. The
|
|
/// specific order used is the post-order across the operand DAG. When the last
|
|
/// user of a PHI is visited in this postorder walk, we check it for
|
|
/// profitability.
|
|
///
|
|
/// There is an orthogonal extra complexity to all of this: computing the cost
|
|
/// itself can easily become a linear computation making everything again (at
|
|
/// best) quadratic. Using a postorder over the operand graph makes it
|
|
/// particularly easy to avoid this through dynamic programming. As we do the
|
|
/// postorder walk, we build the transitive cost of that subgraph. It is also
|
|
/// straightforward to then update these costs when we mark a PHI for
|
|
/// speculation so that subsequent PHIs don't re-pay the cost of already
|
|
/// speculated instructions.
|
|
static SmallVector<PHINode *, 16>
|
|
findProfitablePHIs(ArrayRef<PHINode *> PNs,
|
|
const SmallDenseMap<PHINode *, int, 16> &CostSavingsMap,
|
|
const SmallPtrSetImpl<Instruction *> &PotentialSpecSet,
|
|
int NumPreds, DominatorTree &DT, TargetTransformInfo &TTI) {
|
|
SmallVector<PHINode *, 16> SpecPNs;
|
|
|
|
// First, establish a reverse mapping from immediate users of the PHI nodes
|
|
// to the nodes themselves, and count how many users each PHI node has in
|
|
// a way we can update while processing them.
|
|
SmallDenseMap<Instruction *, TinyPtrVector<PHINode *>, 16> UserToPNMap;
|
|
SmallDenseMap<PHINode *, int, 16> PNUserCountMap;
|
|
SmallPtrSet<Instruction *, 16> UserSet;
|
|
for (auto *PN : PNs) {
|
|
assert(UserSet.empty() && "Must start with an empty user set!");
|
|
for (Use &U : PN->uses())
|
|
UserSet.insert(cast<Instruction>(U.getUser()));
|
|
PNUserCountMap[PN] = UserSet.size();
|
|
for (auto *UI : UserSet)
|
|
UserToPNMap.insert({UI, {}}).first->second.push_back(PN);
|
|
UserSet.clear();
|
|
}
|
|
|
|
// Now do a DFS across the operand graph of the users, computing cost as we
|
|
// go and when all costs for a given PHI are known, checking that PHI for
|
|
// profitability.
|
|
SmallDenseMap<Instruction *, int, 16> SpecCostMap;
|
|
visitPHIUsersAndDepsInPostOrder(
|
|
PNs,
|
|
/*IsVisited*/
|
|
[&](Instruction *I) {
|
|
// We consider anything that isn't potentially speculated to be
|
|
// "visited" as it is already handled. Similarly, anything that *is*
|
|
// potentially speculated but for which we have an entry in our cost
|
|
// map, we're done.
|
|
return !PotentialSpecSet.count(I) || SpecCostMap.count(I);
|
|
},
|
|
/*Visit*/
|
|
[&](Instruction *I) {
|
|
// We've fully visited the operands, so sum their cost with this node
|
|
// and update the cost map.
|
|
int Cost = TTI.TCC_Free;
|
|
for (Value *OpV : I->operand_values())
|
|
if (auto *OpI = dyn_cast<Instruction>(OpV)) {
|
|
auto CostMapIt = SpecCostMap.find(OpI);
|
|
if (CostMapIt != SpecCostMap.end())
|
|
Cost += CostMapIt->second;
|
|
}
|
|
Cost += TTI.getUserCost(I, TargetTransformInfo::TCK_SizeAndLatency);
|
|
bool Inserted = SpecCostMap.insert({I, Cost}).second;
|
|
(void)Inserted;
|
|
assert(Inserted && "Must not re-insert a cost during the DFS!");
|
|
|
|
// Now check if this node had a corresponding PHI node using it. If so,
|
|
// we need to decrement the outstanding user count for it.
|
|
auto UserPNsIt = UserToPNMap.find(I);
|
|
if (UserPNsIt == UserToPNMap.end())
|
|
return;
|
|
auto &UserPNs = UserPNsIt->second;
|
|
auto UserPNsSplitIt = std::stable_partition(
|
|
UserPNs.begin(), UserPNs.end(), [&](PHINode *UserPN) {
|
|
int &PNUserCount = PNUserCountMap.find(UserPN)->second;
|
|
assert(
|
|
PNUserCount > 0 &&
|
|
"Should never re-visit a PN after its user count hits zero!");
|
|
--PNUserCount;
|
|
return PNUserCount != 0;
|
|
});
|
|
|
|
// FIXME: Rather than one at a time, we should sum the savings as the
|
|
// cost will be completely shared.
|
|
SmallVector<Instruction *, 16> SpecWorklist;
|
|
for (auto *PN : llvm::make_range(UserPNsSplitIt, UserPNs.end())) {
|
|
int SpecCost = TTI.TCC_Free;
|
|
for (Use &U : PN->uses())
|
|
SpecCost +=
|
|
SpecCostMap.find(cast<Instruction>(U.getUser()))->second;
|
|
SpecCost *= (NumPreds - 1);
|
|
// When the user count of a PHI node hits zero, we should check its
|
|
// profitability. If profitable, we should mark it for speculation
|
|
// and zero out the cost of everything it depends on.
|
|
int CostSavings = CostSavingsMap.find(PN)->second;
|
|
if (SpecCost > CostSavings) {
|
|
LLVM_DEBUG(dbgs() << " Not profitable, speculation cost: " << *PN
|
|
<< "\n"
|
|
" Cost savings: "
|
|
<< CostSavings
|
|
<< "\n"
|
|
" Speculation cost: "
|
|
<< SpecCost << "\n");
|
|
continue;
|
|
}
|
|
|
|
// We're going to speculate this user-associated PHI. Copy it out and
|
|
// add its users to the worklist to update their cost.
|
|
SpecPNs.push_back(PN);
|
|
for (Use &U : PN->uses()) {
|
|
auto *UI = cast<Instruction>(U.getUser());
|
|
auto CostMapIt = SpecCostMap.find(UI);
|
|
if (CostMapIt->second == 0)
|
|
continue;
|
|
// Zero out this cost entry to avoid duplicates.
|
|
CostMapIt->second = 0;
|
|
SpecWorklist.push_back(UI);
|
|
}
|
|
}
|
|
|
|
// Now walk all the operands of the users in the worklist transitively
|
|
// to zero out all the memoized costs.
|
|
while (!SpecWorklist.empty()) {
|
|
Instruction *SpecI = SpecWorklist.pop_back_val();
|
|
assert(SpecCostMap.find(SpecI)->second == 0 &&
|
|
"Didn't zero out a cost!");
|
|
|
|
// Walk the operands recursively to zero out their cost as well.
|
|
for (auto *OpV : SpecI->operand_values()) {
|
|
auto *OpI = dyn_cast<Instruction>(OpV);
|
|
if (!OpI)
|
|
continue;
|
|
auto CostMapIt = SpecCostMap.find(OpI);
|
|
if (CostMapIt == SpecCostMap.end() || CostMapIt->second == 0)
|
|
continue;
|
|
CostMapIt->second = 0;
|
|
SpecWorklist.push_back(OpI);
|
|
}
|
|
}
|
|
});
|
|
|
|
return SpecPNs;
|
|
}
|
|
|
|
/// Speculate users around a set of PHI nodes.
|
|
///
|
|
/// This routine does the actual speculation around a set of PHI nodes where we
|
|
/// have determined this to be both safe and profitable.
|
|
///
|
|
/// This routine handles any spliting of critical edges necessary to create
|
|
/// a safe block to speculate into as well as cloning the instructions and
|
|
/// rewriting all uses.
|
|
static void speculatePHIs(ArrayRef<PHINode *> SpecPNs,
|
|
SmallPtrSetImpl<Instruction *> &PotentialSpecSet,
|
|
SmallSetVector<BasicBlock *, 16> &PredSet,
|
|
DominatorTree &DT) {
|
|
LLVM_DEBUG(dbgs() << " Speculating around " << SpecPNs.size() << " PHIs!\n");
|
|
NumPHIsSpeculated += SpecPNs.size();
|
|
|
|
// Split any critical edges so that we have a block to hoist into.
|
|
auto *ParentBB = SpecPNs[0]->getParent();
|
|
SmallVector<BasicBlock *, 16> SpecPreds;
|
|
SpecPreds.reserve(PredSet.size());
|
|
for (auto *PredBB : PredSet) {
|
|
auto *NewPredBB = SplitCriticalEdge(
|
|
PredBB, ParentBB,
|
|
CriticalEdgeSplittingOptions(&DT).setMergeIdenticalEdges());
|
|
if (NewPredBB) {
|
|
++NumEdgesSplit;
|
|
LLVM_DEBUG(dbgs() << " Split critical edge from: " << PredBB->getName()
|
|
<< "\n");
|
|
SpecPreds.push_back(NewPredBB);
|
|
} else {
|
|
assert(PredBB->getSingleSuccessor() == ParentBB &&
|
|
"We need a non-critical predecessor to speculate into.");
|
|
assert(!isa<InvokeInst>(PredBB->getTerminator()) &&
|
|
"Cannot have a non-critical invoke!");
|
|
|
|
// Already non-critical, use existing pred.
|
|
SpecPreds.push_back(PredBB);
|
|
}
|
|
}
|
|
|
|
SmallPtrSet<Instruction *, 16> SpecSet;
|
|
SmallVector<Instruction *, 16> SpecList;
|
|
visitPHIUsersAndDepsInPostOrder(SpecPNs,
|
|
/*IsVisited*/
|
|
[&](Instruction *I) {
|
|
// This is visited if we don't need to
|
|
// speculate it or we already have
|
|
// speculated it.
|
|
return !PotentialSpecSet.count(I) ||
|
|
SpecSet.count(I);
|
|
},
|
|
/*Visit*/
|
|
[&](Instruction *I) {
|
|
// All operands scheduled, schedule this
|
|
// node.
|
|
SpecSet.insert(I);
|
|
SpecList.push_back(I);
|
|
});
|
|
|
|
int NumSpecInsts = SpecList.size() * SpecPreds.size();
|
|
int NumRedundantInsts = NumSpecInsts - SpecList.size();
|
|
LLVM_DEBUG(dbgs() << " Inserting " << NumSpecInsts
|
|
<< " speculated instructions, " << NumRedundantInsts
|
|
<< " redundancies\n");
|
|
NumSpeculatedInstructions += NumSpecInsts;
|
|
NumNewRedundantInstructions += NumRedundantInsts;
|
|
|
|
// Each predecessor is numbered by its index in `SpecPreds`, so for each
|
|
// instruction we speculate, the speculated instruction is stored in that
|
|
// index of the vector associated with the original instruction. We also
|
|
// store the incoming values for each predecessor from any PHIs used.
|
|
SmallDenseMap<Instruction *, SmallVector<Value *, 2>, 16> SpeculatedValueMap;
|
|
|
|
// Inject the synthetic mappings to rewrite PHIs to the appropriate incoming
|
|
// value. This handles both the PHIs we are speculating around and any other
|
|
// PHIs that happen to be used.
|
|
for (auto *OrigI : SpecList)
|
|
for (auto *OpV : OrigI->operand_values()) {
|
|
auto *OpPN = dyn_cast<PHINode>(OpV);
|
|
if (!OpPN || OpPN->getParent() != ParentBB)
|
|
continue;
|
|
|
|
auto InsertResult = SpeculatedValueMap.insert({OpPN, {}});
|
|
if (!InsertResult.second)
|
|
continue;
|
|
|
|
auto &SpeculatedVals = InsertResult.first->second;
|
|
|
|
// Populating our structure for mapping is particularly annoying because
|
|
// finding an incoming value for a particular predecessor block in a PHI
|
|
// node is a linear time operation! To avoid quadratic behavior, we build
|
|
// a map for this PHI node's incoming values and then translate it into
|
|
// the more compact representation used below.
|
|
SmallDenseMap<BasicBlock *, Value *, 16> IncomingValueMap;
|
|
for (int i : llvm::seq<int>(0, OpPN->getNumIncomingValues()))
|
|
IncomingValueMap[OpPN->getIncomingBlock(i)] = OpPN->getIncomingValue(i);
|
|
|
|
for (auto *PredBB : SpecPreds)
|
|
SpeculatedVals.push_back(IncomingValueMap.find(PredBB)->second);
|
|
}
|
|
|
|
// Speculate into each predecessor.
|
|
for (int PredIdx : llvm::seq<int>(0, SpecPreds.size())) {
|
|
auto *PredBB = SpecPreds[PredIdx];
|
|
assert(PredBB->getSingleSuccessor() == ParentBB &&
|
|
"We need a non-critical predecessor to speculate into.");
|
|
|
|
for (auto *OrigI : SpecList) {
|
|
auto *NewI = OrigI->clone();
|
|
NewI->setName(Twine(OrigI->getName()) + "." + Twine(PredIdx));
|
|
NewI->insertBefore(PredBB->getTerminator());
|
|
|
|
// Rewrite all the operands to the previously speculated instructions.
|
|
// Because we're walking in-order, the defs must precede the uses and we
|
|
// should already have these mappings.
|
|
for (Use &U : NewI->operands()) {
|
|
auto *OpI = dyn_cast<Instruction>(U.get());
|
|
if (!OpI)
|
|
continue;
|
|
auto MapIt = SpeculatedValueMap.find(OpI);
|
|
if (MapIt == SpeculatedValueMap.end())
|
|
continue;
|
|
const auto &SpeculatedVals = MapIt->second;
|
|
assert(SpeculatedVals[PredIdx] &&
|
|
"Must have a speculated value for this predecessor!");
|
|
assert(SpeculatedVals[PredIdx]->getType() == OpI->getType() &&
|
|
"Speculated value has the wrong type!");
|
|
|
|
// Rewrite the use to this predecessor's speculated instruction.
|
|
U.set(SpeculatedVals[PredIdx]);
|
|
}
|
|
|
|
// Commute instructions which now have a constant in the LHS but not the
|
|
// RHS.
|
|
if (NewI->isBinaryOp() && NewI->isCommutative() &&
|
|
isa<Constant>(NewI->getOperand(0)) &&
|
|
!isa<Constant>(NewI->getOperand(1)))
|
|
NewI->getOperandUse(0).swap(NewI->getOperandUse(1));
|
|
|
|
SpeculatedValueMap[OrigI].push_back(NewI);
|
|
assert(SpeculatedValueMap[OrigI][PredIdx] == NewI &&
|
|
"Mismatched speculated instruction index!");
|
|
}
|
|
}
|
|
|
|
// Walk the speculated instruction list and if they have uses, insert a PHI
|
|
// for them from the speculated versions, and replace the uses with the PHI.
|
|
// Then erase the instructions as they have been fully speculated. The walk
|
|
// needs to be in reverse so that we don't think there are users when we'll
|
|
// actually eventually remove them later.
|
|
IRBuilder<> IRB(SpecPNs[0]);
|
|
for (auto *OrigI : llvm::reverse(SpecList)) {
|
|
// Check if we need a PHI for any remaining users and if so, insert it.
|
|
if (!OrigI->use_empty()) {
|
|
auto *SpecIPN = IRB.CreatePHI(OrigI->getType(), SpecPreds.size(),
|
|
Twine(OrigI->getName()) + ".phi");
|
|
// Add the incoming values we speculated.
|
|
auto &SpeculatedVals = SpeculatedValueMap.find(OrigI)->second;
|
|
for (int PredIdx : llvm::seq<int>(0, SpecPreds.size()))
|
|
SpecIPN->addIncoming(SpeculatedVals[PredIdx], SpecPreds[PredIdx]);
|
|
|
|
// And replace the uses with the PHI node.
|
|
OrigI->replaceAllUsesWith(SpecIPN);
|
|
}
|
|
|
|
// It is important to immediately erase this so that it stops using other
|
|
// instructions. This avoids inserting needless PHIs of them.
|
|
OrigI->eraseFromParent();
|
|
}
|
|
|
|
// All of the uses of the speculated phi nodes should be removed at this
|
|
// point, so erase them.
|
|
for (auto *SpecPN : SpecPNs) {
|
|
assert(SpecPN->use_empty() && "All users should have been speculated!");
|
|
SpecPN->eraseFromParent();
|
|
}
|
|
}
|
|
|
|
/// Try to speculate around a series of PHIs from a single basic block.
|
|
///
|
|
/// This routine checks whether any of these PHIs are profitable to speculate
|
|
/// users around. If safe and profitable, it does the speculation. It returns
|
|
/// true when at least some speculation occurs.
|
|
static bool tryToSpeculatePHIs(SmallVectorImpl<PHINode *> &PNs,
|
|
DominatorTree &DT, TargetTransformInfo &TTI) {
|
|
LLVM_DEBUG(dbgs() << "Evaluating phi nodes for speculation:\n");
|
|
|
|
// Savings in cost from speculating around a PHI node.
|
|
SmallDenseMap<PHINode *, int, 16> CostSavingsMap;
|
|
|
|
// Remember the set of instructions that are candidates for speculation so
|
|
// that we can quickly walk things within that space. This prunes out
|
|
// instructions already available along edges, etc.
|
|
SmallPtrSet<Instruction *, 16> PotentialSpecSet;
|
|
|
|
// Remember the set of instructions that are (transitively) unsafe to
|
|
// speculate into the incoming edges of this basic block. This avoids
|
|
// recomputing them for each PHI node we check. This set is specific to this
|
|
// block though as things are pruned out of it based on what is available
|
|
// along incoming edges.
|
|
SmallPtrSet<Instruction *, 16> UnsafeSet;
|
|
|
|
// For each PHI node in this block, check whether there are immediate folding
|
|
// opportunities from speculation, and whether that speculation will be
|
|
// valid. This determise the set of safe PHIs to speculate.
|
|
PNs.erase(llvm::remove_if(PNs,
|
|
[&](PHINode *PN) {
|
|
return !isSafeAndProfitableToSpeculateAroundPHI(
|
|
*PN, CostSavingsMap, PotentialSpecSet,
|
|
UnsafeSet, DT, TTI);
|
|
}),
|
|
PNs.end());
|
|
// If no PHIs were profitable, skip.
|
|
if (PNs.empty()) {
|
|
LLVM_DEBUG(dbgs() << " No safe and profitable PHIs found!\n");
|
|
return false;
|
|
}
|
|
|
|
// We need to know how much speculation will cost which is determined by how
|
|
// many incoming edges will need a copy of each speculated instruction.
|
|
SmallSetVector<BasicBlock *, 16> PredSet;
|
|
for (auto *PredBB : PNs[0]->blocks()) {
|
|
if (!PredSet.insert(PredBB))
|
|
continue;
|
|
|
|
// We cannot speculate when a predecessor is an indirect branch.
|
|
// FIXME: We also can't reliably create a non-critical edge block for
|
|
// speculation if the predecessor is an invoke. This doesn't seem
|
|
// fundamental and we should probably be splitting critical edges
|
|
// differently.
|
|
const auto *TermInst = PredBB->getTerminator();
|
|
if (isa<IndirectBrInst>(TermInst) ||
|
|
isa<InvokeInst>(TermInst) ||
|
|
isa<CallBrInst>(TermInst)) {
|
|
LLVM_DEBUG(dbgs() << " Invalid: predecessor terminator: "
|
|
<< PredBB->getName() << "\n");
|
|
return false;
|
|
}
|
|
}
|
|
if (PredSet.size() < 2) {
|
|
LLVM_DEBUG(dbgs() << " Unimportant: phi with only one predecessor\n");
|
|
return false;
|
|
}
|
|
|
|
SmallVector<PHINode *, 16> SpecPNs = findProfitablePHIs(
|
|
PNs, CostSavingsMap, PotentialSpecSet, PredSet.size(), DT, TTI);
|
|
if (SpecPNs.empty())
|
|
// Nothing to do.
|
|
return false;
|
|
|
|
speculatePHIs(SpecPNs, PotentialSpecSet, PredSet, DT);
|
|
return true;
|
|
}
|
|
|
|
PreservedAnalyses SpeculateAroundPHIsPass::run(Function &F,
|
|
FunctionAnalysisManager &AM) {
|
|
auto &DT = AM.getResult<DominatorTreeAnalysis>(F);
|
|
auto &TTI = AM.getResult<TargetIRAnalysis>(F);
|
|
|
|
bool Changed = false;
|
|
for (auto *BB : ReversePostOrderTraversal<Function *>(&F)) {
|
|
SmallVector<PHINode *, 16> PNs;
|
|
auto BBI = BB->begin();
|
|
while (auto *PN = dyn_cast<PHINode>(&*BBI)) {
|
|
PNs.push_back(PN);
|
|
++BBI;
|
|
}
|
|
|
|
if (PNs.empty())
|
|
continue;
|
|
|
|
Changed |= tryToSpeculatePHIs(PNs, DT, TTI);
|
|
}
|
|
|
|
if (!Changed)
|
|
return PreservedAnalyses::all();
|
|
|
|
PreservedAnalyses PA;
|
|
return PA;
|
|
}
|