mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-25 20:23:11 +01:00
d9118fe9e0
This reverts commit 848a68a032d1c59274526abb3220714202d4757e.
4200 lines
168 KiB
C++
4200 lines
168 KiB
C++
//===- NewGVN.cpp - Global Value Numbering Pass ---------------------------===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
/// \file
|
|
/// This file implements the new LLVM's Global Value Numbering pass.
|
|
/// GVN partitions values computed by a function into congruence classes.
|
|
/// Values ending up in the same congruence class are guaranteed to be the same
|
|
/// for every execution of the program. In that respect, congruency is a
|
|
/// compile-time approximation of equivalence of values at runtime.
|
|
/// The algorithm implemented here uses a sparse formulation and it's based
|
|
/// on the ideas described in the paper:
|
|
/// "A Sparse Algorithm for Predicated Global Value Numbering" from
|
|
/// Karthik Gargi.
|
|
///
|
|
/// A brief overview of the algorithm: The algorithm is essentially the same as
|
|
/// the standard RPO value numbering algorithm (a good reference is the paper
|
|
/// "SCC based value numbering" by L. Taylor Simpson) with one major difference:
|
|
/// The RPO algorithm proceeds, on every iteration, to process every reachable
|
|
/// block and every instruction in that block. This is because the standard RPO
|
|
/// algorithm does not track what things have the same value number, it only
|
|
/// tracks what the value number of a given operation is (the mapping is
|
|
/// operation -> value number). Thus, when a value number of an operation
|
|
/// changes, it must reprocess everything to ensure all uses of a value number
|
|
/// get updated properly. In constrast, the sparse algorithm we use *also*
|
|
/// tracks what operations have a given value number (IE it also tracks the
|
|
/// reverse mapping from value number -> operations with that value number), so
|
|
/// that it only needs to reprocess the instructions that are affected when
|
|
/// something's value number changes. The vast majority of complexity and code
|
|
/// in this file is devoted to tracking what value numbers could change for what
|
|
/// instructions when various things happen. The rest of the algorithm is
|
|
/// devoted to performing symbolic evaluation, forward propagation, and
|
|
/// simplification of operations based on the value numbers deduced so far
|
|
///
|
|
/// In order to make the GVN mostly-complete, we use a technique derived from
|
|
/// "Detection of Redundant Expressions: A Complete and Polynomial-time
|
|
/// Algorithm in SSA" by R.R. Pai. The source of incompleteness in most SSA
|
|
/// based GVN algorithms is related to their inability to detect equivalence
|
|
/// between phi of ops (IE phi(a+b, c+d)) and op of phis (phi(a,c) + phi(b, d)).
|
|
/// We resolve this issue by generating the equivalent "phi of ops" form for
|
|
/// each op of phis we see, in a way that only takes polynomial time to resolve.
|
|
///
|
|
/// We also do not perform elimination by using any published algorithm. All
|
|
/// published algorithms are O(Instructions). Instead, we use a technique that
|
|
/// is O(number of operations with the same value number), enabling us to skip
|
|
/// trying to eliminate things that have unique value numbers.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "llvm/Transforms/Scalar/NewGVN.h"
|
|
#include "llvm/ADT/ArrayRef.h"
|
|
#include "llvm/ADT/BitVector.h"
|
|
#include "llvm/ADT/DenseMap.h"
|
|
#include "llvm/ADT/DenseMapInfo.h"
|
|
#include "llvm/ADT/DenseSet.h"
|
|
#include "llvm/ADT/DepthFirstIterator.h"
|
|
#include "llvm/ADT/GraphTraits.h"
|
|
#include "llvm/ADT/Hashing.h"
|
|
#include "llvm/ADT/PointerIntPair.h"
|
|
#include "llvm/ADT/PostOrderIterator.h"
|
|
#include "llvm/ADT/SmallPtrSet.h"
|
|
#include "llvm/ADT/SmallVector.h"
|
|
#include "llvm/ADT/SparseBitVector.h"
|
|
#include "llvm/ADT/Statistic.h"
|
|
#include "llvm/ADT/iterator_range.h"
|
|
#include "llvm/Analysis/AliasAnalysis.h"
|
|
#include "llvm/Analysis/AssumptionCache.h"
|
|
#include "llvm/Analysis/CFGPrinter.h"
|
|
#include "llvm/Analysis/ConstantFolding.h"
|
|
#include "llvm/Analysis/GlobalsModRef.h"
|
|
#include "llvm/Analysis/InstructionSimplify.h"
|
|
#include "llvm/Analysis/MemoryBuiltins.h"
|
|
#include "llvm/Analysis/MemorySSA.h"
|
|
#include "llvm/Analysis/TargetLibraryInfo.h"
|
|
#include "llvm/IR/Argument.h"
|
|
#include "llvm/IR/BasicBlock.h"
|
|
#include "llvm/IR/Constant.h"
|
|
#include "llvm/IR/Constants.h"
|
|
#include "llvm/IR/Dominators.h"
|
|
#include "llvm/IR/Function.h"
|
|
#include "llvm/IR/InstrTypes.h"
|
|
#include "llvm/IR/Instruction.h"
|
|
#include "llvm/IR/Instructions.h"
|
|
#include "llvm/IR/IntrinsicInst.h"
|
|
#include "llvm/IR/Intrinsics.h"
|
|
#include "llvm/IR/LLVMContext.h"
|
|
#include "llvm/IR/PatternMatch.h"
|
|
#include "llvm/IR/Type.h"
|
|
#include "llvm/IR/Use.h"
|
|
#include "llvm/IR/User.h"
|
|
#include "llvm/IR/Value.h"
|
|
#include "llvm/InitializePasses.h"
|
|
#include "llvm/Pass.h"
|
|
#include "llvm/Support/Allocator.h"
|
|
#include "llvm/Support/ArrayRecycler.h"
|
|
#include "llvm/Support/Casting.h"
|
|
#include "llvm/Support/CommandLine.h"
|
|
#include "llvm/Support/Debug.h"
|
|
#include "llvm/Support/DebugCounter.h"
|
|
#include "llvm/Support/ErrorHandling.h"
|
|
#include "llvm/Support/PointerLikeTypeTraits.h"
|
|
#include "llvm/Support/raw_ostream.h"
|
|
#include "llvm/Transforms/Scalar.h"
|
|
#include "llvm/Transforms/Scalar/GVNExpression.h"
|
|
#include "llvm/Transforms/Utils/AssumeBundleBuilder.h"
|
|
#include "llvm/Transforms/Utils/Local.h"
|
|
#include "llvm/Transforms/Utils/PredicateInfo.h"
|
|
#include "llvm/Transforms/Utils/VNCoercion.h"
|
|
#include <algorithm>
|
|
#include <cassert>
|
|
#include <cstdint>
|
|
#include <iterator>
|
|
#include <map>
|
|
#include <memory>
|
|
#include <set>
|
|
#include <string>
|
|
#include <tuple>
|
|
#include <utility>
|
|
#include <vector>
|
|
|
|
using namespace llvm;
|
|
using namespace llvm::GVNExpression;
|
|
using namespace llvm::VNCoercion;
|
|
using namespace llvm::PatternMatch;
|
|
|
|
#define DEBUG_TYPE "newgvn"
|
|
|
|
STATISTIC(NumGVNInstrDeleted, "Number of instructions deleted");
|
|
STATISTIC(NumGVNBlocksDeleted, "Number of blocks deleted");
|
|
STATISTIC(NumGVNOpsSimplified, "Number of Expressions simplified");
|
|
STATISTIC(NumGVNPhisAllSame, "Number of PHIs whos arguments are all the same");
|
|
STATISTIC(NumGVNMaxIterations,
|
|
"Maximum Number of iterations it took to converge GVN");
|
|
STATISTIC(NumGVNLeaderChanges, "Number of leader changes");
|
|
STATISTIC(NumGVNSortedLeaderChanges, "Number of sorted leader changes");
|
|
STATISTIC(NumGVNAvoidedSortedLeaderChanges,
|
|
"Number of avoided sorted leader changes");
|
|
STATISTIC(NumGVNDeadStores, "Number of redundant/dead stores eliminated");
|
|
STATISTIC(NumGVNPHIOfOpsCreated, "Number of PHI of ops created");
|
|
STATISTIC(NumGVNPHIOfOpsEliminations,
|
|
"Number of things eliminated using PHI of ops");
|
|
DEBUG_COUNTER(VNCounter, "newgvn-vn",
|
|
"Controls which instructions are value numbered");
|
|
DEBUG_COUNTER(PHIOfOpsCounter, "newgvn-phi",
|
|
"Controls which instructions we create phi of ops for");
|
|
// Currently store defining access refinement is too slow due to basicaa being
|
|
// egregiously slow. This flag lets us keep it working while we work on this
|
|
// issue.
|
|
static cl::opt<bool> EnableStoreRefinement("enable-store-refinement",
|
|
cl::init(false), cl::Hidden);
|
|
|
|
/// Currently, the generation "phi of ops" can result in correctness issues.
|
|
static cl::opt<bool> EnablePhiOfOps("enable-phi-of-ops", cl::init(true),
|
|
cl::Hidden);
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
// GVN Pass
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// Anchor methods.
|
|
namespace llvm {
|
|
namespace GVNExpression {
|
|
|
|
Expression::~Expression() = default;
|
|
BasicExpression::~BasicExpression() = default;
|
|
CallExpression::~CallExpression() = default;
|
|
LoadExpression::~LoadExpression() = default;
|
|
StoreExpression::~StoreExpression() = default;
|
|
AggregateValueExpression::~AggregateValueExpression() = default;
|
|
PHIExpression::~PHIExpression() = default;
|
|
|
|
} // end namespace GVNExpression
|
|
} // end namespace llvm
|
|
|
|
namespace {
|
|
|
|
// Tarjan's SCC finding algorithm with Nuutila's improvements
|
|
// SCCIterator is actually fairly complex for the simple thing we want.
|
|
// It also wants to hand us SCC's that are unrelated to the phi node we ask
|
|
// about, and have us process them there or risk redoing work.
|
|
// Graph traits over a filter iterator also doesn't work that well here.
|
|
// This SCC finder is specialized to walk use-def chains, and only follows
|
|
// instructions,
|
|
// not generic values (arguments, etc).
|
|
struct TarjanSCC {
|
|
TarjanSCC() : Components(1) {}
|
|
|
|
void Start(const Instruction *Start) {
|
|
if (Root.lookup(Start) == 0)
|
|
FindSCC(Start);
|
|
}
|
|
|
|
const SmallPtrSetImpl<const Value *> &getComponentFor(const Value *V) const {
|
|
unsigned ComponentID = ValueToComponent.lookup(V);
|
|
|
|
assert(ComponentID > 0 &&
|
|
"Asking for a component for a value we never processed");
|
|
return Components[ComponentID];
|
|
}
|
|
|
|
private:
|
|
void FindSCC(const Instruction *I) {
|
|
Root[I] = ++DFSNum;
|
|
// Store the DFS Number we had before it possibly gets incremented.
|
|
unsigned int OurDFS = DFSNum;
|
|
for (auto &Op : I->operands()) {
|
|
if (auto *InstOp = dyn_cast<Instruction>(Op)) {
|
|
if (Root.lookup(Op) == 0)
|
|
FindSCC(InstOp);
|
|
if (!InComponent.count(Op))
|
|
Root[I] = std::min(Root.lookup(I), Root.lookup(Op));
|
|
}
|
|
}
|
|
// See if we really were the root of a component, by seeing if we still have
|
|
// our DFSNumber. If we do, we are the root of the component, and we have
|
|
// completed a component. If we do not, we are not the root of a component,
|
|
// and belong on the component stack.
|
|
if (Root.lookup(I) == OurDFS) {
|
|
unsigned ComponentID = Components.size();
|
|
Components.resize(Components.size() + 1);
|
|
auto &Component = Components.back();
|
|
Component.insert(I);
|
|
LLVM_DEBUG(dbgs() << "Component root is " << *I << "\n");
|
|
InComponent.insert(I);
|
|
ValueToComponent[I] = ComponentID;
|
|
// Pop a component off the stack and label it.
|
|
while (!Stack.empty() && Root.lookup(Stack.back()) >= OurDFS) {
|
|
auto *Member = Stack.back();
|
|
LLVM_DEBUG(dbgs() << "Component member is " << *Member << "\n");
|
|
Component.insert(Member);
|
|
InComponent.insert(Member);
|
|
ValueToComponent[Member] = ComponentID;
|
|
Stack.pop_back();
|
|
}
|
|
} else {
|
|
// Part of a component, push to stack
|
|
Stack.push_back(I);
|
|
}
|
|
}
|
|
|
|
unsigned int DFSNum = 1;
|
|
SmallPtrSet<const Value *, 8> InComponent;
|
|
DenseMap<const Value *, unsigned int> Root;
|
|
SmallVector<const Value *, 8> Stack;
|
|
|
|
// Store the components as vector of ptr sets, because we need the topo order
|
|
// of SCC's, but not individual member order
|
|
SmallVector<SmallPtrSet<const Value *, 8>, 8> Components;
|
|
|
|
DenseMap<const Value *, unsigned> ValueToComponent;
|
|
};
|
|
|
|
// Congruence classes represent the set of expressions/instructions
|
|
// that are all the same *during some scope in the function*.
|
|
// That is, because of the way we perform equality propagation, and
|
|
// because of memory value numbering, it is not correct to assume
|
|
// you can willy-nilly replace any member with any other at any
|
|
// point in the function.
|
|
//
|
|
// For any Value in the Member set, it is valid to replace any dominated member
|
|
// with that Value.
|
|
//
|
|
// Every congruence class has a leader, and the leader is used to symbolize
|
|
// instructions in a canonical way (IE every operand of an instruction that is a
|
|
// member of the same congruence class will always be replaced with leader
|
|
// during symbolization). To simplify symbolization, we keep the leader as a
|
|
// constant if class can be proved to be a constant value. Otherwise, the
|
|
// leader is the member of the value set with the smallest DFS number. Each
|
|
// congruence class also has a defining expression, though the expression may be
|
|
// null. If it exists, it can be used for forward propagation and reassociation
|
|
// of values.
|
|
|
|
// For memory, we also track a representative MemoryAccess, and a set of memory
|
|
// members for MemoryPhis (which have no real instructions). Note that for
|
|
// memory, it seems tempting to try to split the memory members into a
|
|
// MemoryCongruenceClass or something. Unfortunately, this does not work
|
|
// easily. The value numbering of a given memory expression depends on the
|
|
// leader of the memory congruence class, and the leader of memory congruence
|
|
// class depends on the value numbering of a given memory expression. This
|
|
// leads to wasted propagation, and in some cases, missed optimization. For
|
|
// example: If we had value numbered two stores together before, but now do not,
|
|
// we move them to a new value congruence class. This in turn will move at one
|
|
// of the memorydefs to a new memory congruence class. Which in turn, affects
|
|
// the value numbering of the stores we just value numbered (because the memory
|
|
// congruence class is part of the value number). So while theoretically
|
|
// possible to split them up, it turns out to be *incredibly* complicated to get
|
|
// it to work right, because of the interdependency. While structurally
|
|
// slightly messier, it is algorithmically much simpler and faster to do what we
|
|
// do here, and track them both at once in the same class.
|
|
// Note: The default iterators for this class iterate over values
|
|
class CongruenceClass {
|
|
public:
|
|
using MemberType = Value;
|
|
using MemberSet = SmallPtrSet<MemberType *, 4>;
|
|
using MemoryMemberType = MemoryPhi;
|
|
using MemoryMemberSet = SmallPtrSet<const MemoryMemberType *, 2>;
|
|
|
|
explicit CongruenceClass(unsigned ID) : ID(ID) {}
|
|
CongruenceClass(unsigned ID, Value *Leader, const Expression *E)
|
|
: ID(ID), RepLeader(Leader), DefiningExpr(E) {}
|
|
|
|
unsigned getID() const { return ID; }
|
|
|
|
// True if this class has no members left. This is mainly used for assertion
|
|
// purposes, and for skipping empty classes.
|
|
bool isDead() const {
|
|
// If it's both dead from a value perspective, and dead from a memory
|
|
// perspective, it's really dead.
|
|
return empty() && memory_empty();
|
|
}
|
|
|
|
// Leader functions
|
|
Value *getLeader() const { return RepLeader; }
|
|
void setLeader(Value *Leader) { RepLeader = Leader; }
|
|
const std::pair<Value *, unsigned int> &getNextLeader() const {
|
|
return NextLeader;
|
|
}
|
|
void resetNextLeader() { NextLeader = {nullptr, ~0}; }
|
|
void addPossibleNextLeader(std::pair<Value *, unsigned int> LeaderPair) {
|
|
if (LeaderPair.second < NextLeader.second)
|
|
NextLeader = LeaderPair;
|
|
}
|
|
|
|
Value *getStoredValue() const { return RepStoredValue; }
|
|
void setStoredValue(Value *Leader) { RepStoredValue = Leader; }
|
|
const MemoryAccess *getMemoryLeader() const { return RepMemoryAccess; }
|
|
void setMemoryLeader(const MemoryAccess *Leader) { RepMemoryAccess = Leader; }
|
|
|
|
// Forward propagation info
|
|
const Expression *getDefiningExpr() const { return DefiningExpr; }
|
|
|
|
// Value member set
|
|
bool empty() const { return Members.empty(); }
|
|
unsigned size() const { return Members.size(); }
|
|
MemberSet::const_iterator begin() const { return Members.begin(); }
|
|
MemberSet::const_iterator end() const { return Members.end(); }
|
|
void insert(MemberType *M) { Members.insert(M); }
|
|
void erase(MemberType *M) { Members.erase(M); }
|
|
void swap(MemberSet &Other) { Members.swap(Other); }
|
|
|
|
// Memory member set
|
|
bool memory_empty() const { return MemoryMembers.empty(); }
|
|
unsigned memory_size() const { return MemoryMembers.size(); }
|
|
MemoryMemberSet::const_iterator memory_begin() const {
|
|
return MemoryMembers.begin();
|
|
}
|
|
MemoryMemberSet::const_iterator memory_end() const {
|
|
return MemoryMembers.end();
|
|
}
|
|
iterator_range<MemoryMemberSet::const_iterator> memory() const {
|
|
return make_range(memory_begin(), memory_end());
|
|
}
|
|
|
|
void memory_insert(const MemoryMemberType *M) { MemoryMembers.insert(M); }
|
|
void memory_erase(const MemoryMemberType *M) { MemoryMembers.erase(M); }
|
|
|
|
// Store count
|
|
unsigned getStoreCount() const { return StoreCount; }
|
|
void incStoreCount() { ++StoreCount; }
|
|
void decStoreCount() {
|
|
assert(StoreCount != 0 && "Store count went negative");
|
|
--StoreCount;
|
|
}
|
|
|
|
// True if this class has no memory members.
|
|
bool definesNoMemory() const { return StoreCount == 0 && memory_empty(); }
|
|
|
|
// Return true if two congruence classes are equivalent to each other. This
|
|
// means that every field but the ID number and the dead field are equivalent.
|
|
bool isEquivalentTo(const CongruenceClass *Other) const {
|
|
if (!Other)
|
|
return false;
|
|
if (this == Other)
|
|
return true;
|
|
|
|
if (std::tie(StoreCount, RepLeader, RepStoredValue, RepMemoryAccess) !=
|
|
std::tie(Other->StoreCount, Other->RepLeader, Other->RepStoredValue,
|
|
Other->RepMemoryAccess))
|
|
return false;
|
|
if (DefiningExpr != Other->DefiningExpr)
|
|
if (!DefiningExpr || !Other->DefiningExpr ||
|
|
*DefiningExpr != *Other->DefiningExpr)
|
|
return false;
|
|
|
|
if (Members.size() != Other->Members.size())
|
|
return false;
|
|
|
|
return all_of(Members,
|
|
[&](const Value *V) { return Other->Members.count(V); });
|
|
}
|
|
|
|
private:
|
|
unsigned ID;
|
|
|
|
// Representative leader.
|
|
Value *RepLeader = nullptr;
|
|
|
|
// The most dominating leader after our current leader, because the member set
|
|
// is not sorted and is expensive to keep sorted all the time.
|
|
std::pair<Value *, unsigned int> NextLeader = {nullptr, ~0U};
|
|
|
|
// If this is represented by a store, the value of the store.
|
|
Value *RepStoredValue = nullptr;
|
|
|
|
// If this class contains MemoryDefs or MemoryPhis, this is the leading memory
|
|
// access.
|
|
const MemoryAccess *RepMemoryAccess = nullptr;
|
|
|
|
// Defining Expression.
|
|
const Expression *DefiningExpr = nullptr;
|
|
|
|
// Actual members of this class.
|
|
MemberSet Members;
|
|
|
|
// This is the set of MemoryPhis that exist in the class. MemoryDefs and
|
|
// MemoryUses have real instructions representing them, so we only need to
|
|
// track MemoryPhis here.
|
|
MemoryMemberSet MemoryMembers;
|
|
|
|
// Number of stores in this congruence class.
|
|
// This is used so we can detect store equivalence changes properly.
|
|
int StoreCount = 0;
|
|
};
|
|
|
|
} // end anonymous namespace
|
|
|
|
namespace llvm {
|
|
|
|
struct ExactEqualsExpression {
|
|
const Expression &E;
|
|
|
|
explicit ExactEqualsExpression(const Expression &E) : E(E) {}
|
|
|
|
hash_code getComputedHash() const { return E.getComputedHash(); }
|
|
|
|
bool operator==(const Expression &Other) const {
|
|
return E.exactlyEquals(Other);
|
|
}
|
|
};
|
|
|
|
template <> struct DenseMapInfo<const Expression *> {
|
|
static const Expression *getEmptyKey() {
|
|
auto Val = static_cast<uintptr_t>(-1);
|
|
Val <<= PointerLikeTypeTraits<const Expression *>::NumLowBitsAvailable;
|
|
return reinterpret_cast<const Expression *>(Val);
|
|
}
|
|
|
|
static const Expression *getTombstoneKey() {
|
|
auto Val = static_cast<uintptr_t>(~1U);
|
|
Val <<= PointerLikeTypeTraits<const Expression *>::NumLowBitsAvailable;
|
|
return reinterpret_cast<const Expression *>(Val);
|
|
}
|
|
|
|
static unsigned getHashValue(const Expression *E) {
|
|
return E->getComputedHash();
|
|
}
|
|
|
|
static unsigned getHashValue(const ExactEqualsExpression &E) {
|
|
return E.getComputedHash();
|
|
}
|
|
|
|
static bool isEqual(const ExactEqualsExpression &LHS, const Expression *RHS) {
|
|
if (RHS == getTombstoneKey() || RHS == getEmptyKey())
|
|
return false;
|
|
return LHS == *RHS;
|
|
}
|
|
|
|
static bool isEqual(const Expression *LHS, const Expression *RHS) {
|
|
if (LHS == RHS)
|
|
return true;
|
|
if (LHS == getTombstoneKey() || RHS == getTombstoneKey() ||
|
|
LHS == getEmptyKey() || RHS == getEmptyKey())
|
|
return false;
|
|
// Compare hashes before equality. This is *not* what the hashtable does,
|
|
// since it is computing it modulo the number of buckets, whereas we are
|
|
// using the full hash keyspace. Since the hashes are precomputed, this
|
|
// check is *much* faster than equality.
|
|
if (LHS->getComputedHash() != RHS->getComputedHash())
|
|
return false;
|
|
return *LHS == *RHS;
|
|
}
|
|
};
|
|
|
|
} // end namespace llvm
|
|
|
|
namespace {
|
|
|
|
class NewGVN {
|
|
Function &F;
|
|
DominatorTree *DT = nullptr;
|
|
const TargetLibraryInfo *TLI = nullptr;
|
|
AliasAnalysis *AA = nullptr;
|
|
MemorySSA *MSSA = nullptr;
|
|
MemorySSAWalker *MSSAWalker = nullptr;
|
|
AssumptionCache *AC = nullptr;
|
|
const DataLayout &DL;
|
|
std::unique_ptr<PredicateInfo> PredInfo;
|
|
|
|
// These are the only two things the create* functions should have
|
|
// side-effects on due to allocating memory.
|
|
mutable BumpPtrAllocator ExpressionAllocator;
|
|
mutable ArrayRecycler<Value *> ArgRecycler;
|
|
mutable TarjanSCC SCCFinder;
|
|
const SimplifyQuery SQ;
|
|
|
|
// Number of function arguments, used by ranking
|
|
unsigned int NumFuncArgs = 0;
|
|
|
|
// RPOOrdering of basic blocks
|
|
DenseMap<const DomTreeNode *, unsigned> RPOOrdering;
|
|
|
|
// Congruence class info.
|
|
|
|
// This class is called INITIAL in the paper. It is the class everything
|
|
// startsout in, and represents any value. Being an optimistic analysis,
|
|
// anything in the TOP class has the value TOP, which is indeterminate and
|
|
// equivalent to everything.
|
|
CongruenceClass *TOPClass = nullptr;
|
|
std::vector<CongruenceClass *> CongruenceClasses;
|
|
unsigned NextCongruenceNum = 0;
|
|
|
|
// Value Mappings.
|
|
DenseMap<Value *, CongruenceClass *> ValueToClass;
|
|
DenseMap<Value *, const Expression *> ValueToExpression;
|
|
|
|
// Value PHI handling, used to make equivalence between phi(op, op) and
|
|
// op(phi, phi).
|
|
// These mappings just store various data that would normally be part of the
|
|
// IR.
|
|
SmallPtrSet<const Instruction *, 8> PHINodeUses;
|
|
|
|
DenseMap<const Value *, bool> OpSafeForPHIOfOps;
|
|
|
|
// Map a temporary instruction we created to a parent block.
|
|
DenseMap<const Value *, BasicBlock *> TempToBlock;
|
|
|
|
// Map between the already in-program instructions and the temporary phis we
|
|
// created that they are known equivalent to.
|
|
DenseMap<const Value *, PHINode *> RealToTemp;
|
|
|
|
// In order to know when we should re-process instructions that have
|
|
// phi-of-ops, we track the set of expressions that they needed as
|
|
// leaders. When we discover new leaders for those expressions, we process the
|
|
// associated phi-of-op instructions again in case they have changed. The
|
|
// other way they may change is if they had leaders, and those leaders
|
|
// disappear. However, at the point they have leaders, there are uses of the
|
|
// relevant operands in the created phi node, and so they will get reprocessed
|
|
// through the normal user marking we perform.
|
|
mutable DenseMap<const Value *, SmallPtrSet<Value *, 2>> AdditionalUsers;
|
|
DenseMap<const Expression *, SmallPtrSet<Instruction *, 2>>
|
|
ExpressionToPhiOfOps;
|
|
|
|
// Map from temporary operation to MemoryAccess.
|
|
DenseMap<const Instruction *, MemoryUseOrDef *> TempToMemory;
|
|
|
|
// Set of all temporary instructions we created.
|
|
// Note: This will include instructions that were just created during value
|
|
// numbering. The way to test if something is using them is to check
|
|
// RealToTemp.
|
|
DenseSet<Instruction *> AllTempInstructions;
|
|
|
|
// This is the set of instructions to revisit on a reachability change. At
|
|
// the end of the main iteration loop it will contain at least all the phi of
|
|
// ops instructions that will be changed to phis, as well as regular phis.
|
|
// During the iteration loop, it may contain other things, such as phi of ops
|
|
// instructions that used edge reachability to reach a result, and so need to
|
|
// be revisited when the edge changes, independent of whether the phi they
|
|
// depended on changes.
|
|
DenseMap<BasicBlock *, SparseBitVector<>> RevisitOnReachabilityChange;
|
|
|
|
// Mapping from predicate info we used to the instructions we used it with.
|
|
// In order to correctly ensure propagation, we must keep track of what
|
|
// comparisons we used, so that when the values of the comparisons change, we
|
|
// propagate the information to the places we used the comparison.
|
|
mutable DenseMap<const Value *, SmallPtrSet<Instruction *, 2>>
|
|
PredicateToUsers;
|
|
|
|
// the same reasoning as PredicateToUsers. When we skip MemoryAccesses for
|
|
// stores, we no longer can rely solely on the def-use chains of MemorySSA.
|
|
mutable DenseMap<const MemoryAccess *, SmallPtrSet<MemoryAccess *, 2>>
|
|
MemoryToUsers;
|
|
|
|
// A table storing which memorydefs/phis represent a memory state provably
|
|
// equivalent to another memory state.
|
|
// We could use the congruence class machinery, but the MemoryAccess's are
|
|
// abstract memory states, so they can only ever be equivalent to each other,
|
|
// and not to constants, etc.
|
|
DenseMap<const MemoryAccess *, CongruenceClass *> MemoryAccessToClass;
|
|
|
|
// We could, if we wanted, build MemoryPhiExpressions and
|
|
// MemoryVariableExpressions, etc, and value number them the same way we value
|
|
// number phi expressions. For the moment, this seems like overkill. They
|
|
// can only exist in one of three states: they can be TOP (equal to
|
|
// everything), Equivalent to something else, or unique. Because we do not
|
|
// create expressions for them, we need to simulate leader change not just
|
|
// when they change class, but when they change state. Note: We can do the
|
|
// same thing for phis, and avoid having phi expressions if we wanted, We
|
|
// should eventually unify in one direction or the other, so this is a little
|
|
// bit of an experiment in which turns out easier to maintain.
|
|
enum MemoryPhiState { MPS_Invalid, MPS_TOP, MPS_Equivalent, MPS_Unique };
|
|
DenseMap<const MemoryPhi *, MemoryPhiState> MemoryPhiState;
|
|
|
|
enum InstCycleState { ICS_Unknown, ICS_CycleFree, ICS_Cycle };
|
|
mutable DenseMap<const Instruction *, InstCycleState> InstCycleState;
|
|
|
|
// Expression to class mapping.
|
|
using ExpressionClassMap = DenseMap<const Expression *, CongruenceClass *>;
|
|
ExpressionClassMap ExpressionToClass;
|
|
|
|
// We have a single expression that represents currently DeadExpressions.
|
|
// For dead expressions we can prove will stay dead, we mark them with
|
|
// DFS number zero. However, it's possible in the case of phi nodes
|
|
// for us to assume/prove all arguments are dead during fixpointing.
|
|
// We use DeadExpression for that case.
|
|
DeadExpression *SingletonDeadExpression = nullptr;
|
|
|
|
// Which values have changed as a result of leader changes.
|
|
SmallPtrSet<Value *, 8> LeaderChanges;
|
|
|
|
// Reachability info.
|
|
using BlockEdge = BasicBlockEdge;
|
|
DenseSet<BlockEdge> ReachableEdges;
|
|
SmallPtrSet<const BasicBlock *, 8> ReachableBlocks;
|
|
|
|
// This is a bitvector because, on larger functions, we may have
|
|
// thousands of touched instructions at once (entire blocks,
|
|
// instructions with hundreds of uses, etc). Even with optimization
|
|
// for when we mark whole blocks as touched, when this was a
|
|
// SmallPtrSet or DenseSet, for some functions, we spent >20% of all
|
|
// the time in GVN just managing this list. The bitvector, on the
|
|
// other hand, efficiently supports test/set/clear of both
|
|
// individual and ranges, as well as "find next element" This
|
|
// enables us to use it as a worklist with essentially 0 cost.
|
|
BitVector TouchedInstructions;
|
|
|
|
DenseMap<const BasicBlock *, std::pair<unsigned, unsigned>> BlockInstRange;
|
|
|
|
#ifndef NDEBUG
|
|
// Debugging for how many times each block and instruction got processed.
|
|
DenseMap<const Value *, unsigned> ProcessedCount;
|
|
#endif
|
|
|
|
// DFS info.
|
|
// This contains a mapping from Instructions to DFS numbers.
|
|
// The numbering starts at 1. An instruction with DFS number zero
|
|
// means that the instruction is dead.
|
|
DenseMap<const Value *, unsigned> InstrDFS;
|
|
|
|
// This contains the mapping DFS numbers to instructions.
|
|
SmallVector<Value *, 32> DFSToInstr;
|
|
|
|
// Deletion info.
|
|
SmallPtrSet<Instruction *, 8> InstructionsToErase;
|
|
|
|
public:
|
|
NewGVN(Function &F, DominatorTree *DT, AssumptionCache *AC,
|
|
TargetLibraryInfo *TLI, AliasAnalysis *AA, MemorySSA *MSSA,
|
|
const DataLayout &DL)
|
|
: F(F), DT(DT), TLI(TLI), AA(AA), MSSA(MSSA), AC(AC), DL(DL),
|
|
PredInfo(std::make_unique<PredicateInfo>(F, *DT, *AC)),
|
|
SQ(DL, TLI, DT, AC, /*CtxI=*/nullptr, /*UseInstrInfo=*/false,
|
|
/*CanUseUndef=*/false) {}
|
|
|
|
bool runGVN();
|
|
|
|
private:
|
|
// Expression handling.
|
|
const Expression *createExpression(Instruction *) const;
|
|
const Expression *createBinaryExpression(unsigned, Type *, Value *, Value *,
|
|
Instruction *) const;
|
|
|
|
// Our canonical form for phi arguments is a pair of incoming value, incoming
|
|
// basic block.
|
|
using ValPair = std::pair<Value *, BasicBlock *>;
|
|
|
|
PHIExpression *createPHIExpression(ArrayRef<ValPair>, const Instruction *,
|
|
BasicBlock *, bool &HasBackEdge,
|
|
bool &OriginalOpsConstant) const;
|
|
const DeadExpression *createDeadExpression() const;
|
|
const VariableExpression *createVariableExpression(Value *) const;
|
|
const ConstantExpression *createConstantExpression(Constant *) const;
|
|
const Expression *createVariableOrConstant(Value *V) const;
|
|
const UnknownExpression *createUnknownExpression(Instruction *) const;
|
|
const StoreExpression *createStoreExpression(StoreInst *,
|
|
const MemoryAccess *) const;
|
|
LoadExpression *createLoadExpression(Type *, Value *, LoadInst *,
|
|
const MemoryAccess *) const;
|
|
const CallExpression *createCallExpression(CallInst *,
|
|
const MemoryAccess *) const;
|
|
const AggregateValueExpression *
|
|
createAggregateValueExpression(Instruction *) const;
|
|
bool setBasicExpressionInfo(Instruction *, BasicExpression *) const;
|
|
|
|
// Congruence class handling.
|
|
CongruenceClass *createCongruenceClass(Value *Leader, const Expression *E) {
|
|
auto *result = new CongruenceClass(NextCongruenceNum++, Leader, E);
|
|
CongruenceClasses.emplace_back(result);
|
|
return result;
|
|
}
|
|
|
|
CongruenceClass *createMemoryClass(MemoryAccess *MA) {
|
|
auto *CC = createCongruenceClass(nullptr, nullptr);
|
|
CC->setMemoryLeader(MA);
|
|
return CC;
|
|
}
|
|
|
|
CongruenceClass *ensureLeaderOfMemoryClass(MemoryAccess *MA) {
|
|
auto *CC = getMemoryClass(MA);
|
|
if (CC->getMemoryLeader() != MA)
|
|
CC = createMemoryClass(MA);
|
|
return CC;
|
|
}
|
|
|
|
CongruenceClass *createSingletonCongruenceClass(Value *Member) {
|
|
CongruenceClass *CClass = createCongruenceClass(Member, nullptr);
|
|
CClass->insert(Member);
|
|
ValueToClass[Member] = CClass;
|
|
return CClass;
|
|
}
|
|
|
|
void initializeCongruenceClasses(Function &F);
|
|
const Expression *makePossiblePHIOfOps(Instruction *,
|
|
SmallPtrSetImpl<Value *> &);
|
|
Value *findLeaderForInst(Instruction *ValueOp,
|
|
SmallPtrSetImpl<Value *> &Visited,
|
|
MemoryAccess *MemAccess, Instruction *OrigInst,
|
|
BasicBlock *PredBB);
|
|
bool OpIsSafeForPHIOfOpsHelper(Value *V, const BasicBlock *PHIBlock,
|
|
SmallPtrSetImpl<const Value *> &Visited,
|
|
SmallVectorImpl<Instruction *> &Worklist);
|
|
bool OpIsSafeForPHIOfOps(Value *Op, const BasicBlock *PHIBlock,
|
|
SmallPtrSetImpl<const Value *> &);
|
|
void addPhiOfOps(PHINode *Op, BasicBlock *BB, Instruction *ExistingValue);
|
|
void removePhiOfOps(Instruction *I, PHINode *PHITemp);
|
|
|
|
// Value number an Instruction or MemoryPhi.
|
|
void valueNumberMemoryPhi(MemoryPhi *);
|
|
void valueNumberInstruction(Instruction *);
|
|
|
|
// Symbolic evaluation.
|
|
const Expression *checkSimplificationResults(Expression *, Instruction *,
|
|
Value *) const;
|
|
const Expression *performSymbolicEvaluation(Value *,
|
|
SmallPtrSetImpl<Value *> &) const;
|
|
const Expression *performSymbolicLoadCoercion(Type *, Value *, LoadInst *,
|
|
Instruction *,
|
|
MemoryAccess *) const;
|
|
const Expression *performSymbolicLoadEvaluation(Instruction *) const;
|
|
const Expression *performSymbolicStoreEvaluation(Instruction *) const;
|
|
const Expression *performSymbolicCallEvaluation(Instruction *) const;
|
|
void sortPHIOps(MutableArrayRef<ValPair> Ops) const;
|
|
const Expression *performSymbolicPHIEvaluation(ArrayRef<ValPair>,
|
|
Instruction *I,
|
|
BasicBlock *PHIBlock) const;
|
|
const Expression *performSymbolicAggrValueEvaluation(Instruction *) const;
|
|
const Expression *performSymbolicCmpEvaluation(Instruction *) const;
|
|
const Expression *performSymbolicPredicateInfoEvaluation(Instruction *) const;
|
|
|
|
// Congruence finding.
|
|
bool someEquivalentDominates(const Instruction *, const Instruction *) const;
|
|
Value *lookupOperandLeader(Value *) const;
|
|
CongruenceClass *getClassForExpression(const Expression *E) const;
|
|
void performCongruenceFinding(Instruction *, const Expression *);
|
|
void moveValueToNewCongruenceClass(Instruction *, const Expression *,
|
|
CongruenceClass *, CongruenceClass *);
|
|
void moveMemoryToNewCongruenceClass(Instruction *, MemoryAccess *,
|
|
CongruenceClass *, CongruenceClass *);
|
|
Value *getNextValueLeader(CongruenceClass *) const;
|
|
const MemoryAccess *getNextMemoryLeader(CongruenceClass *) const;
|
|
bool setMemoryClass(const MemoryAccess *From, CongruenceClass *To);
|
|
CongruenceClass *getMemoryClass(const MemoryAccess *MA) const;
|
|
const MemoryAccess *lookupMemoryLeader(const MemoryAccess *) const;
|
|
bool isMemoryAccessTOP(const MemoryAccess *) const;
|
|
|
|
// Ranking
|
|
unsigned int getRank(const Value *) const;
|
|
bool shouldSwapOperands(const Value *, const Value *) const;
|
|
|
|
// Reachability handling.
|
|
void updateReachableEdge(BasicBlock *, BasicBlock *);
|
|
void processOutgoingEdges(Instruction *, BasicBlock *);
|
|
Value *findConditionEquivalence(Value *) const;
|
|
|
|
// Elimination.
|
|
struct ValueDFS;
|
|
void convertClassToDFSOrdered(const CongruenceClass &,
|
|
SmallVectorImpl<ValueDFS> &,
|
|
DenseMap<const Value *, unsigned int> &,
|
|
SmallPtrSetImpl<Instruction *> &) const;
|
|
void convertClassToLoadsAndStores(const CongruenceClass &,
|
|
SmallVectorImpl<ValueDFS> &) const;
|
|
|
|
bool eliminateInstructions(Function &);
|
|
void replaceInstruction(Instruction *, Value *);
|
|
void markInstructionForDeletion(Instruction *);
|
|
void deleteInstructionsInBlock(BasicBlock *);
|
|
Value *findPHIOfOpsLeader(const Expression *, const Instruction *,
|
|
const BasicBlock *) const;
|
|
|
|
// New instruction creation.
|
|
void handleNewInstruction(Instruction *) {}
|
|
|
|
// Various instruction touch utilities
|
|
template <typename Map, typename KeyType, typename Func>
|
|
void for_each_found(Map &, const KeyType &, Func);
|
|
template <typename Map, typename KeyType>
|
|
void touchAndErase(Map &, const KeyType &);
|
|
void markUsersTouched(Value *);
|
|
void markMemoryUsersTouched(const MemoryAccess *);
|
|
void markMemoryDefTouched(const MemoryAccess *);
|
|
void markPredicateUsersTouched(Instruction *);
|
|
void markValueLeaderChangeTouched(CongruenceClass *CC);
|
|
void markMemoryLeaderChangeTouched(CongruenceClass *CC);
|
|
void markPhiOfOpsChanged(const Expression *E);
|
|
void addPredicateUsers(const PredicateBase *, Instruction *) const;
|
|
void addMemoryUsers(const MemoryAccess *To, MemoryAccess *U) const;
|
|
void addAdditionalUsers(Value *To, Value *User) const;
|
|
|
|
// Main loop of value numbering
|
|
void iterateTouchedInstructions();
|
|
|
|
// Utilities.
|
|
void cleanupTables();
|
|
std::pair<unsigned, unsigned> assignDFSNumbers(BasicBlock *, unsigned);
|
|
void updateProcessedCount(const Value *V);
|
|
void verifyMemoryCongruency() const;
|
|
void verifyIterationSettled(Function &F);
|
|
void verifyStoreExpressions() const;
|
|
bool singleReachablePHIPath(SmallPtrSet<const MemoryAccess *, 8> &,
|
|
const MemoryAccess *, const MemoryAccess *) const;
|
|
BasicBlock *getBlockForValue(Value *V) const;
|
|
void deleteExpression(const Expression *E) const;
|
|
MemoryUseOrDef *getMemoryAccess(const Instruction *) const;
|
|
MemoryAccess *getDefiningAccess(const MemoryAccess *) const;
|
|
MemoryPhi *getMemoryAccess(const BasicBlock *) const;
|
|
template <class T, class Range> T *getMinDFSOfRange(const Range &) const;
|
|
|
|
unsigned InstrToDFSNum(const Value *V) const {
|
|
assert(isa<Instruction>(V) && "This should not be used for MemoryAccesses");
|
|
return InstrDFS.lookup(V);
|
|
}
|
|
|
|
unsigned InstrToDFSNum(const MemoryAccess *MA) const {
|
|
return MemoryToDFSNum(MA);
|
|
}
|
|
|
|
Value *InstrFromDFSNum(unsigned DFSNum) { return DFSToInstr[DFSNum]; }
|
|
|
|
// Given a MemoryAccess, return the relevant instruction DFS number. Note:
|
|
// This deliberately takes a value so it can be used with Use's, which will
|
|
// auto-convert to Value's but not to MemoryAccess's.
|
|
unsigned MemoryToDFSNum(const Value *MA) const {
|
|
assert(isa<MemoryAccess>(MA) &&
|
|
"This should not be used with instructions");
|
|
return isa<MemoryUseOrDef>(MA)
|
|
? InstrToDFSNum(cast<MemoryUseOrDef>(MA)->getMemoryInst())
|
|
: InstrDFS.lookup(MA);
|
|
}
|
|
|
|
bool isCycleFree(const Instruction *) const;
|
|
bool isBackedge(BasicBlock *From, BasicBlock *To) const;
|
|
|
|
// Debug counter info. When verifying, we have to reset the value numbering
|
|
// debug counter to the same state it started in to get the same results.
|
|
int64_t StartingVNCounter = 0;
|
|
};
|
|
|
|
} // end anonymous namespace
|
|
|
|
template <typename T>
|
|
static bool equalsLoadStoreHelper(const T &LHS, const Expression &RHS) {
|
|
if (!isa<LoadExpression>(RHS) && !isa<StoreExpression>(RHS))
|
|
return false;
|
|
return LHS.MemoryExpression::equals(RHS);
|
|
}
|
|
|
|
bool LoadExpression::equals(const Expression &Other) const {
|
|
return equalsLoadStoreHelper(*this, Other);
|
|
}
|
|
|
|
bool StoreExpression::equals(const Expression &Other) const {
|
|
if (!equalsLoadStoreHelper(*this, Other))
|
|
return false;
|
|
// Make sure that store vs store includes the value operand.
|
|
if (const auto *S = dyn_cast<StoreExpression>(&Other))
|
|
if (getStoredValue() != S->getStoredValue())
|
|
return false;
|
|
return true;
|
|
}
|
|
|
|
// Determine if the edge From->To is a backedge
|
|
bool NewGVN::isBackedge(BasicBlock *From, BasicBlock *To) const {
|
|
return From == To ||
|
|
RPOOrdering.lookup(DT->getNode(From)) >=
|
|
RPOOrdering.lookup(DT->getNode(To));
|
|
}
|
|
|
|
#ifndef NDEBUG
|
|
static std::string getBlockName(const BasicBlock *B) {
|
|
return DOTGraphTraits<DOTFuncInfo *>::getSimpleNodeLabel(B, nullptr);
|
|
}
|
|
#endif
|
|
|
|
// Get a MemoryAccess for an instruction, fake or real.
|
|
MemoryUseOrDef *NewGVN::getMemoryAccess(const Instruction *I) const {
|
|
auto *Result = MSSA->getMemoryAccess(I);
|
|
return Result ? Result : TempToMemory.lookup(I);
|
|
}
|
|
|
|
// Get a MemoryPhi for a basic block. These are all real.
|
|
MemoryPhi *NewGVN::getMemoryAccess(const BasicBlock *BB) const {
|
|
return MSSA->getMemoryAccess(BB);
|
|
}
|
|
|
|
// Get the basic block from an instruction/memory value.
|
|
BasicBlock *NewGVN::getBlockForValue(Value *V) const {
|
|
if (auto *I = dyn_cast<Instruction>(V)) {
|
|
auto *Parent = I->getParent();
|
|
if (Parent)
|
|
return Parent;
|
|
Parent = TempToBlock.lookup(V);
|
|
assert(Parent && "Every fake instruction should have a block");
|
|
return Parent;
|
|
}
|
|
|
|
auto *MP = dyn_cast<MemoryPhi>(V);
|
|
assert(MP && "Should have been an instruction or a MemoryPhi");
|
|
return MP->getBlock();
|
|
}
|
|
|
|
// Delete a definitely dead expression, so it can be reused by the expression
|
|
// allocator. Some of these are not in creation functions, so we have to accept
|
|
// const versions.
|
|
void NewGVN::deleteExpression(const Expression *E) const {
|
|
assert(isa<BasicExpression>(E));
|
|
auto *BE = cast<BasicExpression>(E);
|
|
const_cast<BasicExpression *>(BE)->deallocateOperands(ArgRecycler);
|
|
ExpressionAllocator.Deallocate(E);
|
|
}
|
|
|
|
// If V is a predicateinfo copy, get the thing it is a copy of.
|
|
static Value *getCopyOf(const Value *V) {
|
|
if (auto *II = dyn_cast<IntrinsicInst>(V))
|
|
if (II->getIntrinsicID() == Intrinsic::ssa_copy)
|
|
return II->getOperand(0);
|
|
return nullptr;
|
|
}
|
|
|
|
// Return true if V is really PN, even accounting for predicateinfo copies.
|
|
static bool isCopyOfPHI(const Value *V, const PHINode *PN) {
|
|
return V == PN || getCopyOf(V) == PN;
|
|
}
|
|
|
|
static bool isCopyOfAPHI(const Value *V) {
|
|
auto *CO = getCopyOf(V);
|
|
return CO && isa<PHINode>(CO);
|
|
}
|
|
|
|
// Sort PHI Operands into a canonical order. What we use here is an RPO
|
|
// order. The BlockInstRange numbers are generated in an RPO walk of the basic
|
|
// blocks.
|
|
void NewGVN::sortPHIOps(MutableArrayRef<ValPair> Ops) const {
|
|
llvm::sort(Ops, [&](const ValPair &P1, const ValPair &P2) {
|
|
return BlockInstRange.lookup(P1.second).first <
|
|
BlockInstRange.lookup(P2.second).first;
|
|
});
|
|
}
|
|
|
|
// Return true if V is a value that will always be available (IE can
|
|
// be placed anywhere) in the function. We don't do globals here
|
|
// because they are often worse to put in place.
|
|
static bool alwaysAvailable(Value *V) {
|
|
return isa<Constant>(V) || isa<Argument>(V);
|
|
}
|
|
|
|
// Create a PHIExpression from an array of {incoming edge, value} pairs. I is
|
|
// the original instruction we are creating a PHIExpression for (but may not be
|
|
// a phi node). We require, as an invariant, that all the PHIOperands in the
|
|
// same block are sorted the same way. sortPHIOps will sort them into a
|
|
// canonical order.
|
|
PHIExpression *NewGVN::createPHIExpression(ArrayRef<ValPair> PHIOperands,
|
|
const Instruction *I,
|
|
BasicBlock *PHIBlock,
|
|
bool &HasBackedge,
|
|
bool &OriginalOpsConstant) const {
|
|
unsigned NumOps = PHIOperands.size();
|
|
auto *E = new (ExpressionAllocator) PHIExpression(NumOps, PHIBlock);
|
|
|
|
E->allocateOperands(ArgRecycler, ExpressionAllocator);
|
|
E->setType(PHIOperands.begin()->first->getType());
|
|
E->setOpcode(Instruction::PHI);
|
|
|
|
// Filter out unreachable phi operands.
|
|
auto Filtered = make_filter_range(PHIOperands, [&](const ValPair &P) {
|
|
auto *BB = P.second;
|
|
if (auto *PHIOp = dyn_cast<PHINode>(I))
|
|
if (isCopyOfPHI(P.first, PHIOp))
|
|
return false;
|
|
if (!ReachableEdges.count({BB, PHIBlock}))
|
|
return false;
|
|
// Things in TOPClass are equivalent to everything.
|
|
if (ValueToClass.lookup(P.first) == TOPClass)
|
|
return false;
|
|
OriginalOpsConstant = OriginalOpsConstant && isa<Constant>(P.first);
|
|
HasBackedge = HasBackedge || isBackedge(BB, PHIBlock);
|
|
return lookupOperandLeader(P.first) != I;
|
|
});
|
|
std::transform(Filtered.begin(), Filtered.end(), op_inserter(E),
|
|
[&](const ValPair &P) -> Value * {
|
|
return lookupOperandLeader(P.first);
|
|
});
|
|
return E;
|
|
}
|
|
|
|
// Set basic expression info (Arguments, type, opcode) for Expression
|
|
// E from Instruction I in block B.
|
|
bool NewGVN::setBasicExpressionInfo(Instruction *I, BasicExpression *E) const {
|
|
bool AllConstant = true;
|
|
if (auto *GEP = dyn_cast<GetElementPtrInst>(I))
|
|
E->setType(GEP->getSourceElementType());
|
|
else
|
|
E->setType(I->getType());
|
|
E->setOpcode(I->getOpcode());
|
|
E->allocateOperands(ArgRecycler, ExpressionAllocator);
|
|
|
|
// Transform the operand array into an operand leader array, and keep track of
|
|
// whether all members are constant.
|
|
std::transform(I->op_begin(), I->op_end(), op_inserter(E), [&](Value *O) {
|
|
auto Operand = lookupOperandLeader(O);
|
|
AllConstant = AllConstant && isa<Constant>(Operand);
|
|
return Operand;
|
|
});
|
|
|
|
return AllConstant;
|
|
}
|
|
|
|
const Expression *NewGVN::createBinaryExpression(unsigned Opcode, Type *T,
|
|
Value *Arg1, Value *Arg2,
|
|
Instruction *I) const {
|
|
auto *E = new (ExpressionAllocator) BasicExpression(2);
|
|
|
|
E->setType(T);
|
|
E->setOpcode(Opcode);
|
|
E->allocateOperands(ArgRecycler, ExpressionAllocator);
|
|
if (Instruction::isCommutative(Opcode)) {
|
|
// Ensure that commutative instructions that only differ by a permutation
|
|
// of their operands get the same value number by sorting the operand value
|
|
// numbers. Since all commutative instructions have two operands it is more
|
|
// efficient to sort by hand rather than using, say, std::sort.
|
|
if (shouldSwapOperands(Arg1, Arg2))
|
|
std::swap(Arg1, Arg2);
|
|
}
|
|
E->op_push_back(lookupOperandLeader(Arg1));
|
|
E->op_push_back(lookupOperandLeader(Arg2));
|
|
|
|
Value *V = SimplifyBinOp(Opcode, E->getOperand(0), E->getOperand(1), SQ);
|
|
if (const Expression *SimplifiedE = checkSimplificationResults(E, I, V))
|
|
return SimplifiedE;
|
|
return E;
|
|
}
|
|
|
|
// Take a Value returned by simplification of Expression E/Instruction
|
|
// I, and see if it resulted in a simpler expression. If so, return
|
|
// that expression.
|
|
const Expression *NewGVN::checkSimplificationResults(Expression *E,
|
|
Instruction *I,
|
|
Value *V) const {
|
|
if (!V)
|
|
return nullptr;
|
|
if (auto *C = dyn_cast<Constant>(V)) {
|
|
if (I)
|
|
LLVM_DEBUG(dbgs() << "Simplified " << *I << " to "
|
|
<< " constant " << *C << "\n");
|
|
NumGVNOpsSimplified++;
|
|
assert(isa<BasicExpression>(E) &&
|
|
"We should always have had a basic expression here");
|
|
deleteExpression(E);
|
|
return createConstantExpression(C);
|
|
} else if (isa<Argument>(V) || isa<GlobalVariable>(V)) {
|
|
if (I)
|
|
LLVM_DEBUG(dbgs() << "Simplified " << *I << " to "
|
|
<< " variable " << *V << "\n");
|
|
deleteExpression(E);
|
|
return createVariableExpression(V);
|
|
}
|
|
|
|
CongruenceClass *CC = ValueToClass.lookup(V);
|
|
if (CC) {
|
|
if (CC->getLeader() && CC->getLeader() != I) {
|
|
// If we simplified to something else, we need to communicate
|
|
// that we're users of the value we simplified to.
|
|
if (I != V) {
|
|
// Don't add temporary instructions to the user lists.
|
|
if (!AllTempInstructions.count(I))
|
|
addAdditionalUsers(V, I);
|
|
}
|
|
return createVariableOrConstant(CC->getLeader());
|
|
}
|
|
if (CC->getDefiningExpr()) {
|
|
// If we simplified to something else, we need to communicate
|
|
// that we're users of the value we simplified to.
|
|
if (I != V) {
|
|
// Don't add temporary instructions to the user lists.
|
|
if (!AllTempInstructions.count(I))
|
|
addAdditionalUsers(V, I);
|
|
}
|
|
|
|
if (I)
|
|
LLVM_DEBUG(dbgs() << "Simplified " << *I << " to "
|
|
<< " expression " << *CC->getDefiningExpr() << "\n");
|
|
NumGVNOpsSimplified++;
|
|
deleteExpression(E);
|
|
return CC->getDefiningExpr();
|
|
}
|
|
}
|
|
|
|
return nullptr;
|
|
}
|
|
|
|
// Create a value expression from the instruction I, replacing operands with
|
|
// their leaders.
|
|
|
|
const Expression *NewGVN::createExpression(Instruction *I) const {
|
|
auto *E = new (ExpressionAllocator) BasicExpression(I->getNumOperands());
|
|
|
|
bool AllConstant = setBasicExpressionInfo(I, E);
|
|
|
|
if (I->isCommutative()) {
|
|
// Ensure that commutative instructions that only differ by a permutation
|
|
// of their operands get the same value number by sorting the operand value
|
|
// numbers. Since all commutative instructions have two operands it is more
|
|
// efficient to sort by hand rather than using, say, std::sort.
|
|
assert(I->getNumOperands() == 2 && "Unsupported commutative instruction!");
|
|
if (shouldSwapOperands(E->getOperand(0), E->getOperand(1)))
|
|
E->swapOperands(0, 1);
|
|
}
|
|
// Perform simplification.
|
|
if (auto *CI = dyn_cast<CmpInst>(I)) {
|
|
// Sort the operand value numbers so x<y and y>x get the same value
|
|
// number.
|
|
CmpInst::Predicate Predicate = CI->getPredicate();
|
|
if (shouldSwapOperands(E->getOperand(0), E->getOperand(1))) {
|
|
E->swapOperands(0, 1);
|
|
Predicate = CmpInst::getSwappedPredicate(Predicate);
|
|
}
|
|
E->setOpcode((CI->getOpcode() << 8) | Predicate);
|
|
// TODO: 25% of our time is spent in SimplifyCmpInst with pointer operands
|
|
assert(I->getOperand(0)->getType() == I->getOperand(1)->getType() &&
|
|
"Wrong types on cmp instruction");
|
|
assert((E->getOperand(0)->getType() == I->getOperand(0)->getType() &&
|
|
E->getOperand(1)->getType() == I->getOperand(1)->getType()));
|
|
Value *V =
|
|
SimplifyCmpInst(Predicate, E->getOperand(0), E->getOperand(1), SQ);
|
|
if (const Expression *SimplifiedE = checkSimplificationResults(E, I, V))
|
|
return SimplifiedE;
|
|
} else if (isa<SelectInst>(I)) {
|
|
if (isa<Constant>(E->getOperand(0)) ||
|
|
E->getOperand(1) == E->getOperand(2)) {
|
|
assert(E->getOperand(1)->getType() == I->getOperand(1)->getType() &&
|
|
E->getOperand(2)->getType() == I->getOperand(2)->getType());
|
|
Value *V = SimplifySelectInst(E->getOperand(0), E->getOperand(1),
|
|
E->getOperand(2), SQ);
|
|
if (const Expression *SimplifiedE = checkSimplificationResults(E, I, V))
|
|
return SimplifiedE;
|
|
}
|
|
} else if (I->isBinaryOp()) {
|
|
Value *V =
|
|
SimplifyBinOp(E->getOpcode(), E->getOperand(0), E->getOperand(1), SQ);
|
|
if (const Expression *SimplifiedE = checkSimplificationResults(E, I, V))
|
|
return SimplifiedE;
|
|
} else if (auto *CI = dyn_cast<CastInst>(I)) {
|
|
Value *V =
|
|
SimplifyCastInst(CI->getOpcode(), E->getOperand(0), CI->getType(), SQ);
|
|
if (const Expression *SimplifiedE = checkSimplificationResults(E, I, V))
|
|
return SimplifiedE;
|
|
} else if (isa<GetElementPtrInst>(I)) {
|
|
Value *V = SimplifyGEPInst(
|
|
E->getType(), ArrayRef<Value *>(E->op_begin(), E->op_end()), SQ);
|
|
if (const Expression *SimplifiedE = checkSimplificationResults(E, I, V))
|
|
return SimplifiedE;
|
|
} else if (AllConstant) {
|
|
// We don't bother trying to simplify unless all of the operands
|
|
// were constant.
|
|
// TODO: There are a lot of Simplify*'s we could call here, if we
|
|
// wanted to. The original motivating case for this code was a
|
|
// zext i1 false to i8, which we don't have an interface to
|
|
// simplify (IE there is no SimplifyZExt).
|
|
|
|
SmallVector<Constant *, 8> C;
|
|
for (Value *Arg : E->operands())
|
|
C.emplace_back(cast<Constant>(Arg));
|
|
|
|
if (Value *V = ConstantFoldInstOperands(I, C, DL, TLI))
|
|
if (const Expression *SimplifiedE = checkSimplificationResults(E, I, V))
|
|
return SimplifiedE;
|
|
}
|
|
return E;
|
|
}
|
|
|
|
const AggregateValueExpression *
|
|
NewGVN::createAggregateValueExpression(Instruction *I) const {
|
|
if (auto *II = dyn_cast<InsertValueInst>(I)) {
|
|
auto *E = new (ExpressionAllocator)
|
|
AggregateValueExpression(I->getNumOperands(), II->getNumIndices());
|
|
setBasicExpressionInfo(I, E);
|
|
E->allocateIntOperands(ExpressionAllocator);
|
|
std::copy(II->idx_begin(), II->idx_end(), int_op_inserter(E));
|
|
return E;
|
|
} else if (auto *EI = dyn_cast<ExtractValueInst>(I)) {
|
|
auto *E = new (ExpressionAllocator)
|
|
AggregateValueExpression(I->getNumOperands(), EI->getNumIndices());
|
|
setBasicExpressionInfo(EI, E);
|
|
E->allocateIntOperands(ExpressionAllocator);
|
|
std::copy(EI->idx_begin(), EI->idx_end(), int_op_inserter(E));
|
|
return E;
|
|
}
|
|
llvm_unreachable("Unhandled type of aggregate value operation");
|
|
}
|
|
|
|
const DeadExpression *NewGVN::createDeadExpression() const {
|
|
// DeadExpression has no arguments and all DeadExpression's are the same,
|
|
// so we only need one of them.
|
|
return SingletonDeadExpression;
|
|
}
|
|
|
|
const VariableExpression *NewGVN::createVariableExpression(Value *V) const {
|
|
auto *E = new (ExpressionAllocator) VariableExpression(V);
|
|
E->setOpcode(V->getValueID());
|
|
return E;
|
|
}
|
|
|
|
const Expression *NewGVN::createVariableOrConstant(Value *V) const {
|
|
if (auto *C = dyn_cast<Constant>(V))
|
|
return createConstantExpression(C);
|
|
return createVariableExpression(V);
|
|
}
|
|
|
|
const ConstantExpression *NewGVN::createConstantExpression(Constant *C) const {
|
|
auto *E = new (ExpressionAllocator) ConstantExpression(C);
|
|
E->setOpcode(C->getValueID());
|
|
return E;
|
|
}
|
|
|
|
const UnknownExpression *NewGVN::createUnknownExpression(Instruction *I) const {
|
|
auto *E = new (ExpressionAllocator) UnknownExpression(I);
|
|
E->setOpcode(I->getOpcode());
|
|
return E;
|
|
}
|
|
|
|
const CallExpression *
|
|
NewGVN::createCallExpression(CallInst *CI, const MemoryAccess *MA) const {
|
|
// FIXME: Add operand bundles for calls.
|
|
// FIXME: Allow commutative matching for intrinsics.
|
|
auto *E =
|
|
new (ExpressionAllocator) CallExpression(CI->getNumOperands(), CI, MA);
|
|
setBasicExpressionInfo(CI, E);
|
|
return E;
|
|
}
|
|
|
|
// Return true if some equivalent of instruction Inst dominates instruction U.
|
|
bool NewGVN::someEquivalentDominates(const Instruction *Inst,
|
|
const Instruction *U) const {
|
|
auto *CC = ValueToClass.lookup(Inst);
|
|
// This must be an instruction because we are only called from phi nodes
|
|
// in the case that the value it needs to check against is an instruction.
|
|
|
|
// The most likely candidates for dominance are the leader and the next leader.
|
|
// The leader or nextleader will dominate in all cases where there is an
|
|
// equivalent that is higher up in the dom tree.
|
|
// We can't *only* check them, however, because the
|
|
// dominator tree could have an infinite number of non-dominating siblings
|
|
// with instructions that are in the right congruence class.
|
|
// A
|
|
// B C D E F G
|
|
// |
|
|
// H
|
|
// Instruction U could be in H, with equivalents in every other sibling.
|
|
// Depending on the rpo order picked, the leader could be the equivalent in
|
|
// any of these siblings.
|
|
if (!CC)
|
|
return false;
|
|
if (alwaysAvailable(CC->getLeader()))
|
|
return true;
|
|
if (DT->dominates(cast<Instruction>(CC->getLeader()), U))
|
|
return true;
|
|
if (CC->getNextLeader().first &&
|
|
DT->dominates(cast<Instruction>(CC->getNextLeader().first), U))
|
|
return true;
|
|
return llvm::any_of(*CC, [&](const Value *Member) {
|
|
return Member != CC->getLeader() &&
|
|
DT->dominates(cast<Instruction>(Member), U);
|
|
});
|
|
}
|
|
|
|
// See if we have a congruence class and leader for this operand, and if so,
|
|
// return it. Otherwise, return the operand itself.
|
|
Value *NewGVN::lookupOperandLeader(Value *V) const {
|
|
CongruenceClass *CC = ValueToClass.lookup(V);
|
|
if (CC) {
|
|
// Everything in TOP is represented by undef, as it can be any value.
|
|
// We do have to make sure we get the type right though, so we can't set the
|
|
// RepLeader to undef.
|
|
if (CC == TOPClass)
|
|
return UndefValue::get(V->getType());
|
|
return CC->getStoredValue() ? CC->getStoredValue() : CC->getLeader();
|
|
}
|
|
|
|
return V;
|
|
}
|
|
|
|
const MemoryAccess *NewGVN::lookupMemoryLeader(const MemoryAccess *MA) const {
|
|
auto *CC = getMemoryClass(MA);
|
|
assert(CC->getMemoryLeader() &&
|
|
"Every MemoryAccess should be mapped to a congruence class with a "
|
|
"representative memory access");
|
|
return CC->getMemoryLeader();
|
|
}
|
|
|
|
// Return true if the MemoryAccess is really equivalent to everything. This is
|
|
// equivalent to the lattice value "TOP" in most lattices. This is the initial
|
|
// state of all MemoryAccesses.
|
|
bool NewGVN::isMemoryAccessTOP(const MemoryAccess *MA) const {
|
|
return getMemoryClass(MA) == TOPClass;
|
|
}
|
|
|
|
LoadExpression *NewGVN::createLoadExpression(Type *LoadType, Value *PointerOp,
|
|
LoadInst *LI,
|
|
const MemoryAccess *MA) const {
|
|
auto *E =
|
|
new (ExpressionAllocator) LoadExpression(1, LI, lookupMemoryLeader(MA));
|
|
E->allocateOperands(ArgRecycler, ExpressionAllocator);
|
|
E->setType(LoadType);
|
|
|
|
// Give store and loads same opcode so they value number together.
|
|
E->setOpcode(0);
|
|
E->op_push_back(PointerOp);
|
|
|
|
// TODO: Value number heap versions. We may be able to discover
|
|
// things alias analysis can't on it's own (IE that a store and a
|
|
// load have the same value, and thus, it isn't clobbering the load).
|
|
return E;
|
|
}
|
|
|
|
const StoreExpression *
|
|
NewGVN::createStoreExpression(StoreInst *SI, const MemoryAccess *MA) const {
|
|
auto *StoredValueLeader = lookupOperandLeader(SI->getValueOperand());
|
|
auto *E = new (ExpressionAllocator)
|
|
StoreExpression(SI->getNumOperands(), SI, StoredValueLeader, MA);
|
|
E->allocateOperands(ArgRecycler, ExpressionAllocator);
|
|
E->setType(SI->getValueOperand()->getType());
|
|
|
|
// Give store and loads same opcode so they value number together.
|
|
E->setOpcode(0);
|
|
E->op_push_back(lookupOperandLeader(SI->getPointerOperand()));
|
|
|
|
// TODO: Value number heap versions. We may be able to discover
|
|
// things alias analysis can't on it's own (IE that a store and a
|
|
// load have the same value, and thus, it isn't clobbering the load).
|
|
return E;
|
|
}
|
|
|
|
const Expression *NewGVN::performSymbolicStoreEvaluation(Instruction *I) const {
|
|
// Unlike loads, we never try to eliminate stores, so we do not check if they
|
|
// are simple and avoid value numbering them.
|
|
auto *SI = cast<StoreInst>(I);
|
|
auto *StoreAccess = getMemoryAccess(SI);
|
|
// Get the expression, if any, for the RHS of the MemoryDef.
|
|
const MemoryAccess *StoreRHS = StoreAccess->getDefiningAccess();
|
|
if (EnableStoreRefinement)
|
|
StoreRHS = MSSAWalker->getClobberingMemoryAccess(StoreAccess);
|
|
// If we bypassed the use-def chains, make sure we add a use.
|
|
StoreRHS = lookupMemoryLeader(StoreRHS);
|
|
if (StoreRHS != StoreAccess->getDefiningAccess())
|
|
addMemoryUsers(StoreRHS, StoreAccess);
|
|
// If we are defined by ourselves, use the live on entry def.
|
|
if (StoreRHS == StoreAccess)
|
|
StoreRHS = MSSA->getLiveOnEntryDef();
|
|
|
|
if (SI->isSimple()) {
|
|
// See if we are defined by a previous store expression, it already has a
|
|
// value, and it's the same value as our current store. FIXME: Right now, we
|
|
// only do this for simple stores, we should expand to cover memcpys, etc.
|
|
const auto *LastStore = createStoreExpression(SI, StoreRHS);
|
|
const auto *LastCC = ExpressionToClass.lookup(LastStore);
|
|
// We really want to check whether the expression we matched was a store. No
|
|
// easy way to do that. However, we can check that the class we found has a
|
|
// store, which, assuming the value numbering state is not corrupt, is
|
|
// sufficient, because we must also be equivalent to that store's expression
|
|
// for it to be in the same class as the load.
|
|
if (LastCC && LastCC->getStoredValue() == LastStore->getStoredValue())
|
|
return LastStore;
|
|
// Also check if our value operand is defined by a load of the same memory
|
|
// location, and the memory state is the same as it was then (otherwise, it
|
|
// could have been overwritten later. See test32 in
|
|
// transforms/DeadStoreElimination/simple.ll).
|
|
if (auto *LI = dyn_cast<LoadInst>(LastStore->getStoredValue()))
|
|
if ((lookupOperandLeader(LI->getPointerOperand()) ==
|
|
LastStore->getOperand(0)) &&
|
|
(lookupMemoryLeader(getMemoryAccess(LI)->getDefiningAccess()) ==
|
|
StoreRHS))
|
|
return LastStore;
|
|
deleteExpression(LastStore);
|
|
}
|
|
|
|
// If the store is not equivalent to anything, value number it as a store that
|
|
// produces a unique memory state (instead of using it's MemoryUse, we use
|
|
// it's MemoryDef).
|
|
return createStoreExpression(SI, StoreAccess);
|
|
}
|
|
|
|
// See if we can extract the value of a loaded pointer from a load, a store, or
|
|
// a memory instruction.
|
|
const Expression *
|
|
NewGVN::performSymbolicLoadCoercion(Type *LoadType, Value *LoadPtr,
|
|
LoadInst *LI, Instruction *DepInst,
|
|
MemoryAccess *DefiningAccess) const {
|
|
assert((!LI || LI->isSimple()) && "Not a simple load");
|
|
if (auto *DepSI = dyn_cast<StoreInst>(DepInst)) {
|
|
// Can't forward from non-atomic to atomic without violating memory model.
|
|
// Also don't need to coerce if they are the same type, we will just
|
|
// propagate.
|
|
if (LI->isAtomic() > DepSI->isAtomic() ||
|
|
LoadType == DepSI->getValueOperand()->getType())
|
|
return nullptr;
|
|
int Offset = analyzeLoadFromClobberingStore(LoadType, LoadPtr, DepSI, DL);
|
|
if (Offset >= 0) {
|
|
if (auto *C = dyn_cast<Constant>(
|
|
lookupOperandLeader(DepSI->getValueOperand()))) {
|
|
LLVM_DEBUG(dbgs() << "Coercing load from store " << *DepSI
|
|
<< " to constant " << *C << "\n");
|
|
return createConstantExpression(
|
|
getConstantStoreValueForLoad(C, Offset, LoadType, DL));
|
|
}
|
|
}
|
|
} else if (auto *DepLI = dyn_cast<LoadInst>(DepInst)) {
|
|
// Can't forward from non-atomic to atomic without violating memory model.
|
|
if (LI->isAtomic() > DepLI->isAtomic())
|
|
return nullptr;
|
|
int Offset = analyzeLoadFromClobberingLoad(LoadType, LoadPtr, DepLI, DL);
|
|
if (Offset >= 0) {
|
|
// We can coerce a constant load into a load.
|
|
if (auto *C = dyn_cast<Constant>(lookupOperandLeader(DepLI)))
|
|
if (auto *PossibleConstant =
|
|
getConstantLoadValueForLoad(C, Offset, LoadType, DL)) {
|
|
LLVM_DEBUG(dbgs() << "Coercing load from load " << *LI
|
|
<< " to constant " << *PossibleConstant << "\n");
|
|
return createConstantExpression(PossibleConstant);
|
|
}
|
|
}
|
|
} else if (auto *DepMI = dyn_cast<MemIntrinsic>(DepInst)) {
|
|
int Offset = analyzeLoadFromClobberingMemInst(LoadType, LoadPtr, DepMI, DL);
|
|
if (Offset >= 0) {
|
|
if (auto *PossibleConstant =
|
|
getConstantMemInstValueForLoad(DepMI, Offset, LoadType, DL)) {
|
|
LLVM_DEBUG(dbgs() << "Coercing load from meminst " << *DepMI
|
|
<< " to constant " << *PossibleConstant << "\n");
|
|
return createConstantExpression(PossibleConstant);
|
|
}
|
|
}
|
|
}
|
|
|
|
// All of the below are only true if the loaded pointer is produced
|
|
// by the dependent instruction.
|
|
if (LoadPtr != lookupOperandLeader(DepInst) &&
|
|
!AA->isMustAlias(LoadPtr, DepInst))
|
|
return nullptr;
|
|
// If this load really doesn't depend on anything, then we must be loading an
|
|
// undef value. This can happen when loading for a fresh allocation with no
|
|
// intervening stores, for example. Note that this is only true in the case
|
|
// that the result of the allocation is pointer equal to the load ptr.
|
|
if (isa<AllocaInst>(DepInst) || isMallocLikeFn(DepInst, TLI) ||
|
|
isAlignedAllocLikeFn(DepInst, TLI)) {
|
|
return createConstantExpression(UndefValue::get(LoadType));
|
|
}
|
|
// If this load occurs either right after a lifetime begin,
|
|
// then the loaded value is undefined.
|
|
else if (auto *II = dyn_cast<IntrinsicInst>(DepInst)) {
|
|
if (II->getIntrinsicID() == Intrinsic::lifetime_start)
|
|
return createConstantExpression(UndefValue::get(LoadType));
|
|
}
|
|
// If this load follows a calloc (which zero initializes memory),
|
|
// then the loaded value is zero
|
|
else if (isCallocLikeFn(DepInst, TLI)) {
|
|
return createConstantExpression(Constant::getNullValue(LoadType));
|
|
}
|
|
|
|
return nullptr;
|
|
}
|
|
|
|
const Expression *NewGVN::performSymbolicLoadEvaluation(Instruction *I) const {
|
|
auto *LI = cast<LoadInst>(I);
|
|
|
|
// We can eliminate in favor of non-simple loads, but we won't be able to
|
|
// eliminate the loads themselves.
|
|
if (!LI->isSimple())
|
|
return nullptr;
|
|
|
|
Value *LoadAddressLeader = lookupOperandLeader(LI->getPointerOperand());
|
|
// Load of undef is undef.
|
|
if (isa<UndefValue>(LoadAddressLeader))
|
|
return createConstantExpression(UndefValue::get(LI->getType()));
|
|
MemoryAccess *OriginalAccess = getMemoryAccess(I);
|
|
MemoryAccess *DefiningAccess =
|
|
MSSAWalker->getClobberingMemoryAccess(OriginalAccess);
|
|
|
|
if (!MSSA->isLiveOnEntryDef(DefiningAccess)) {
|
|
if (auto *MD = dyn_cast<MemoryDef>(DefiningAccess)) {
|
|
Instruction *DefiningInst = MD->getMemoryInst();
|
|
// If the defining instruction is not reachable, replace with undef.
|
|
if (!ReachableBlocks.count(DefiningInst->getParent()))
|
|
return createConstantExpression(UndefValue::get(LI->getType()));
|
|
// This will handle stores and memory insts. We only do if it the
|
|
// defining access has a different type, or it is a pointer produced by
|
|
// certain memory operations that cause the memory to have a fixed value
|
|
// (IE things like calloc).
|
|
if (const auto *CoercionResult =
|
|
performSymbolicLoadCoercion(LI->getType(), LoadAddressLeader, LI,
|
|
DefiningInst, DefiningAccess))
|
|
return CoercionResult;
|
|
}
|
|
}
|
|
|
|
const auto *LE = createLoadExpression(LI->getType(), LoadAddressLeader, LI,
|
|
DefiningAccess);
|
|
// If our MemoryLeader is not our defining access, add a use to the
|
|
// MemoryLeader, so that we get reprocessed when it changes.
|
|
if (LE->getMemoryLeader() != DefiningAccess)
|
|
addMemoryUsers(LE->getMemoryLeader(), OriginalAccess);
|
|
return LE;
|
|
}
|
|
|
|
const Expression *
|
|
NewGVN::performSymbolicPredicateInfoEvaluation(Instruction *I) const {
|
|
auto *PI = PredInfo->getPredicateInfoFor(I);
|
|
if (!PI)
|
|
return nullptr;
|
|
|
|
LLVM_DEBUG(dbgs() << "Found predicate info from instruction !\n");
|
|
|
|
const Optional<PredicateConstraint> &Constraint = PI->getConstraint();
|
|
if (!Constraint)
|
|
return nullptr;
|
|
|
|
CmpInst::Predicate Predicate = Constraint->Predicate;
|
|
Value *CmpOp0 = I->getOperand(0);
|
|
Value *CmpOp1 = Constraint->OtherOp;
|
|
|
|
Value *FirstOp = lookupOperandLeader(CmpOp0);
|
|
Value *SecondOp = lookupOperandLeader(CmpOp1);
|
|
Value *AdditionallyUsedValue = CmpOp0;
|
|
|
|
// Sort the ops.
|
|
if (shouldSwapOperands(FirstOp, SecondOp)) {
|
|
std::swap(FirstOp, SecondOp);
|
|
Predicate = CmpInst::getSwappedPredicate(Predicate);
|
|
AdditionallyUsedValue = CmpOp1;
|
|
}
|
|
|
|
if (Predicate == CmpInst::ICMP_EQ) {
|
|
addPredicateUsers(PI, I);
|
|
addAdditionalUsers(AdditionallyUsedValue, I);
|
|
return createVariableOrConstant(FirstOp);
|
|
}
|
|
|
|
// Handle the special case of floating point.
|
|
if (Predicate == CmpInst::FCMP_OEQ && isa<ConstantFP>(FirstOp) &&
|
|
!cast<ConstantFP>(FirstOp)->isZero()) {
|
|
addPredicateUsers(PI, I);
|
|
addAdditionalUsers(AdditionallyUsedValue, I);
|
|
return createConstantExpression(cast<Constant>(FirstOp));
|
|
}
|
|
|
|
return nullptr;
|
|
}
|
|
|
|
// Evaluate read only and pure calls, and create an expression result.
|
|
const Expression *NewGVN::performSymbolicCallEvaluation(Instruction *I) const {
|
|
auto *CI = cast<CallInst>(I);
|
|
if (auto *II = dyn_cast<IntrinsicInst>(I)) {
|
|
// Intrinsics with the returned attribute are copies of arguments.
|
|
if (auto *ReturnedValue = II->getReturnedArgOperand()) {
|
|
if (II->getIntrinsicID() == Intrinsic::ssa_copy)
|
|
if (const auto *Result = performSymbolicPredicateInfoEvaluation(I))
|
|
return Result;
|
|
return createVariableOrConstant(ReturnedValue);
|
|
}
|
|
}
|
|
if (AA->doesNotAccessMemory(CI)) {
|
|
return createCallExpression(CI, TOPClass->getMemoryLeader());
|
|
} else if (AA->onlyReadsMemory(CI)) {
|
|
if (auto *MA = MSSA->getMemoryAccess(CI)) {
|
|
auto *DefiningAccess = MSSAWalker->getClobberingMemoryAccess(MA);
|
|
return createCallExpression(CI, DefiningAccess);
|
|
} else // MSSA determined that CI does not access memory.
|
|
return createCallExpression(CI, TOPClass->getMemoryLeader());
|
|
}
|
|
return nullptr;
|
|
}
|
|
|
|
// Retrieve the memory class for a given MemoryAccess.
|
|
CongruenceClass *NewGVN::getMemoryClass(const MemoryAccess *MA) const {
|
|
auto *Result = MemoryAccessToClass.lookup(MA);
|
|
assert(Result && "Should have found memory class");
|
|
return Result;
|
|
}
|
|
|
|
// Update the MemoryAccess equivalence table to say that From is equal to To,
|
|
// and return true if this is different from what already existed in the table.
|
|
bool NewGVN::setMemoryClass(const MemoryAccess *From,
|
|
CongruenceClass *NewClass) {
|
|
assert(NewClass &&
|
|
"Every MemoryAccess should be getting mapped to a non-null class");
|
|
LLVM_DEBUG(dbgs() << "Setting " << *From);
|
|
LLVM_DEBUG(dbgs() << " equivalent to congruence class ");
|
|
LLVM_DEBUG(dbgs() << NewClass->getID()
|
|
<< " with current MemoryAccess leader ");
|
|
LLVM_DEBUG(dbgs() << *NewClass->getMemoryLeader() << "\n");
|
|
|
|
auto LookupResult = MemoryAccessToClass.find(From);
|
|
bool Changed = false;
|
|
// If it's already in the table, see if the value changed.
|
|
if (LookupResult != MemoryAccessToClass.end()) {
|
|
auto *OldClass = LookupResult->second;
|
|
if (OldClass != NewClass) {
|
|
// If this is a phi, we have to handle memory member updates.
|
|
if (auto *MP = dyn_cast<MemoryPhi>(From)) {
|
|
OldClass->memory_erase(MP);
|
|
NewClass->memory_insert(MP);
|
|
// This may have killed the class if it had no non-memory members
|
|
if (OldClass->getMemoryLeader() == From) {
|
|
if (OldClass->definesNoMemory()) {
|
|
OldClass->setMemoryLeader(nullptr);
|
|
} else {
|
|
OldClass->setMemoryLeader(getNextMemoryLeader(OldClass));
|
|
LLVM_DEBUG(dbgs() << "Memory class leader change for class "
|
|
<< OldClass->getID() << " to "
|
|
<< *OldClass->getMemoryLeader()
|
|
<< " due to removal of a memory member " << *From
|
|
<< "\n");
|
|
markMemoryLeaderChangeTouched(OldClass);
|
|
}
|
|
}
|
|
}
|
|
// It wasn't equivalent before, and now it is.
|
|
LookupResult->second = NewClass;
|
|
Changed = true;
|
|
}
|
|
}
|
|
|
|
return Changed;
|
|
}
|
|
|
|
// Determine if a instruction is cycle-free. That means the values in the
|
|
// instruction don't depend on any expressions that can change value as a result
|
|
// of the instruction. For example, a non-cycle free instruction would be v =
|
|
// phi(0, v+1).
|
|
bool NewGVN::isCycleFree(const Instruction *I) const {
|
|
// In order to compute cycle-freeness, we do SCC finding on the instruction,
|
|
// and see what kind of SCC it ends up in. If it is a singleton, it is
|
|
// cycle-free. If it is not in a singleton, it is only cycle free if the
|
|
// other members are all phi nodes (as they do not compute anything, they are
|
|
// copies).
|
|
auto ICS = InstCycleState.lookup(I);
|
|
if (ICS == ICS_Unknown) {
|
|
SCCFinder.Start(I);
|
|
auto &SCC = SCCFinder.getComponentFor(I);
|
|
// It's cycle free if it's size 1 or the SCC is *only* phi nodes.
|
|
if (SCC.size() == 1)
|
|
InstCycleState.insert({I, ICS_CycleFree});
|
|
else {
|
|
bool AllPhis = llvm::all_of(SCC, [](const Value *V) {
|
|
return isa<PHINode>(V) || isCopyOfAPHI(V);
|
|
});
|
|
ICS = AllPhis ? ICS_CycleFree : ICS_Cycle;
|
|
for (auto *Member : SCC)
|
|
if (auto *MemberPhi = dyn_cast<PHINode>(Member))
|
|
InstCycleState.insert({MemberPhi, ICS});
|
|
}
|
|
}
|
|
if (ICS == ICS_Cycle)
|
|
return false;
|
|
return true;
|
|
}
|
|
|
|
// Evaluate PHI nodes symbolically and create an expression result.
|
|
const Expression *
|
|
NewGVN::performSymbolicPHIEvaluation(ArrayRef<ValPair> PHIOps,
|
|
Instruction *I,
|
|
BasicBlock *PHIBlock) const {
|
|
// True if one of the incoming phi edges is a backedge.
|
|
bool HasBackedge = false;
|
|
// All constant tracks the state of whether all the *original* phi operands
|
|
// This is really shorthand for "this phi cannot cycle due to forward
|
|
// change in value of the phi is guaranteed not to later change the value of
|
|
// the phi. IE it can't be v = phi(undef, v+1)
|
|
bool OriginalOpsConstant = true;
|
|
auto *E = cast<PHIExpression>(createPHIExpression(
|
|
PHIOps, I, PHIBlock, HasBackedge, OriginalOpsConstant));
|
|
// We match the semantics of SimplifyPhiNode from InstructionSimplify here.
|
|
// See if all arguments are the same.
|
|
// We track if any were undef because they need special handling.
|
|
bool HasUndef = false;
|
|
auto Filtered = make_filter_range(E->operands(), [&](Value *Arg) {
|
|
if (isa<UndefValue>(Arg)) {
|
|
HasUndef = true;
|
|
return false;
|
|
}
|
|
return true;
|
|
});
|
|
// If we are left with no operands, it's dead.
|
|
if (Filtered.empty()) {
|
|
// If it has undef at this point, it means there are no-non-undef arguments,
|
|
// and thus, the value of the phi node must be undef.
|
|
if (HasUndef) {
|
|
LLVM_DEBUG(
|
|
dbgs() << "PHI Node " << *I
|
|
<< " has no non-undef arguments, valuing it as undef\n");
|
|
return createConstantExpression(UndefValue::get(I->getType()));
|
|
}
|
|
|
|
LLVM_DEBUG(dbgs() << "No arguments of PHI node " << *I << " are live\n");
|
|
deleteExpression(E);
|
|
return createDeadExpression();
|
|
}
|
|
Value *AllSameValue = *(Filtered.begin());
|
|
++Filtered.begin();
|
|
// Can't use std::equal here, sadly, because filter.begin moves.
|
|
if (llvm::all_of(Filtered, [&](Value *Arg) { return Arg == AllSameValue; })) {
|
|
// In LLVM's non-standard representation of phi nodes, it's possible to have
|
|
// phi nodes with cycles (IE dependent on other phis that are .... dependent
|
|
// on the original phi node), especially in weird CFG's where some arguments
|
|
// are unreachable, or uninitialized along certain paths. This can cause
|
|
// infinite loops during evaluation. We work around this by not trying to
|
|
// really evaluate them independently, but instead using a variable
|
|
// expression to say if one is equivalent to the other.
|
|
// We also special case undef, so that if we have an undef, we can't use the
|
|
// common value unless it dominates the phi block.
|
|
if (HasUndef) {
|
|
// If we have undef and at least one other value, this is really a
|
|
// multivalued phi, and we need to know if it's cycle free in order to
|
|
// evaluate whether we can ignore the undef. The other parts of this are
|
|
// just shortcuts. If there is no backedge, or all operands are
|
|
// constants, it also must be cycle free.
|
|
if (HasBackedge && !OriginalOpsConstant &&
|
|
!isa<UndefValue>(AllSameValue) && !isCycleFree(I))
|
|
return E;
|
|
|
|
// Only have to check for instructions
|
|
if (auto *AllSameInst = dyn_cast<Instruction>(AllSameValue))
|
|
if (!someEquivalentDominates(AllSameInst, I))
|
|
return E;
|
|
}
|
|
// Can't simplify to something that comes later in the iteration.
|
|
// Otherwise, when and if it changes congruence class, we will never catch
|
|
// up. We will always be a class behind it.
|
|
if (isa<Instruction>(AllSameValue) &&
|
|
InstrToDFSNum(AllSameValue) > InstrToDFSNum(I))
|
|
return E;
|
|
NumGVNPhisAllSame++;
|
|
LLVM_DEBUG(dbgs() << "Simplified PHI node " << *I << " to " << *AllSameValue
|
|
<< "\n");
|
|
deleteExpression(E);
|
|
return createVariableOrConstant(AllSameValue);
|
|
}
|
|
return E;
|
|
}
|
|
|
|
const Expression *
|
|
NewGVN::performSymbolicAggrValueEvaluation(Instruction *I) const {
|
|
if (auto *EI = dyn_cast<ExtractValueInst>(I)) {
|
|
auto *WO = dyn_cast<WithOverflowInst>(EI->getAggregateOperand());
|
|
if (WO && EI->getNumIndices() == 1 && *EI->idx_begin() == 0)
|
|
// EI is an extract from one of our with.overflow intrinsics. Synthesize
|
|
// a semantically equivalent expression instead of an extract value
|
|
// expression.
|
|
return createBinaryExpression(WO->getBinaryOp(), EI->getType(),
|
|
WO->getLHS(), WO->getRHS(), I);
|
|
}
|
|
|
|
return createAggregateValueExpression(I);
|
|
}
|
|
|
|
const Expression *NewGVN::performSymbolicCmpEvaluation(Instruction *I) const {
|
|
assert(isa<CmpInst>(I) && "Expected a cmp instruction.");
|
|
|
|
auto *CI = cast<CmpInst>(I);
|
|
// See if our operands are equal to those of a previous predicate, and if so,
|
|
// if it implies true or false.
|
|
auto Op0 = lookupOperandLeader(CI->getOperand(0));
|
|
auto Op1 = lookupOperandLeader(CI->getOperand(1));
|
|
auto OurPredicate = CI->getPredicate();
|
|
if (shouldSwapOperands(Op0, Op1)) {
|
|
std::swap(Op0, Op1);
|
|
OurPredicate = CI->getSwappedPredicate();
|
|
}
|
|
|
|
// Avoid processing the same info twice.
|
|
const PredicateBase *LastPredInfo = nullptr;
|
|
// See if we know something about the comparison itself, like it is the target
|
|
// of an assume.
|
|
auto *CmpPI = PredInfo->getPredicateInfoFor(I);
|
|
if (dyn_cast_or_null<PredicateAssume>(CmpPI))
|
|
return createConstantExpression(ConstantInt::getTrue(CI->getType()));
|
|
|
|
if (Op0 == Op1) {
|
|
// This condition does not depend on predicates, no need to add users
|
|
if (CI->isTrueWhenEqual())
|
|
return createConstantExpression(ConstantInt::getTrue(CI->getType()));
|
|
else if (CI->isFalseWhenEqual())
|
|
return createConstantExpression(ConstantInt::getFalse(CI->getType()));
|
|
}
|
|
|
|
// NOTE: Because we are comparing both operands here and below, and using
|
|
// previous comparisons, we rely on fact that predicateinfo knows to mark
|
|
// comparisons that use renamed operands as users of the earlier comparisons.
|
|
// It is *not* enough to just mark predicateinfo renamed operands as users of
|
|
// the earlier comparisons, because the *other* operand may have changed in a
|
|
// previous iteration.
|
|
// Example:
|
|
// icmp slt %a, %b
|
|
// %b.0 = ssa.copy(%b)
|
|
// false branch:
|
|
// icmp slt %c, %b.0
|
|
|
|
// %c and %a may start out equal, and thus, the code below will say the second
|
|
// %icmp is false. c may become equal to something else, and in that case the
|
|
// %second icmp *must* be reexamined, but would not if only the renamed
|
|
// %operands are considered users of the icmp.
|
|
|
|
// *Currently* we only check one level of comparisons back, and only mark one
|
|
// level back as touched when changes happen. If you modify this code to look
|
|
// back farther through comparisons, you *must* mark the appropriate
|
|
// comparisons as users in PredicateInfo.cpp, or you will cause bugs. See if
|
|
// we know something just from the operands themselves
|
|
|
|
// See if our operands have predicate info, so that we may be able to derive
|
|
// something from a previous comparison.
|
|
for (const auto &Op : CI->operands()) {
|
|
auto *PI = PredInfo->getPredicateInfoFor(Op);
|
|
if (const auto *PBranch = dyn_cast_or_null<PredicateBranch>(PI)) {
|
|
if (PI == LastPredInfo)
|
|
continue;
|
|
LastPredInfo = PI;
|
|
// In phi of ops cases, we may have predicate info that we are evaluating
|
|
// in a different context.
|
|
if (!DT->dominates(PBranch->To, getBlockForValue(I)))
|
|
continue;
|
|
// TODO: Along the false edge, we may know more things too, like
|
|
// icmp of
|
|
// same operands is false.
|
|
// TODO: We only handle actual comparison conditions below, not
|
|
// and/or.
|
|
auto *BranchCond = dyn_cast<CmpInst>(PBranch->Condition);
|
|
if (!BranchCond)
|
|
continue;
|
|
auto *BranchOp0 = lookupOperandLeader(BranchCond->getOperand(0));
|
|
auto *BranchOp1 = lookupOperandLeader(BranchCond->getOperand(1));
|
|
auto BranchPredicate = BranchCond->getPredicate();
|
|
if (shouldSwapOperands(BranchOp0, BranchOp1)) {
|
|
std::swap(BranchOp0, BranchOp1);
|
|
BranchPredicate = BranchCond->getSwappedPredicate();
|
|
}
|
|
if (BranchOp0 == Op0 && BranchOp1 == Op1) {
|
|
if (PBranch->TrueEdge) {
|
|
// If we know the previous predicate is true and we are in the true
|
|
// edge then we may be implied true or false.
|
|
if (CmpInst::isImpliedTrueByMatchingCmp(BranchPredicate,
|
|
OurPredicate)) {
|
|
addPredicateUsers(PI, I);
|
|
return createConstantExpression(
|
|
ConstantInt::getTrue(CI->getType()));
|
|
}
|
|
|
|
if (CmpInst::isImpliedFalseByMatchingCmp(BranchPredicate,
|
|
OurPredicate)) {
|
|
addPredicateUsers(PI, I);
|
|
return createConstantExpression(
|
|
ConstantInt::getFalse(CI->getType()));
|
|
}
|
|
} else {
|
|
// Just handle the ne and eq cases, where if we have the same
|
|
// operands, we may know something.
|
|
if (BranchPredicate == OurPredicate) {
|
|
addPredicateUsers(PI, I);
|
|
// Same predicate, same ops,we know it was false, so this is false.
|
|
return createConstantExpression(
|
|
ConstantInt::getFalse(CI->getType()));
|
|
} else if (BranchPredicate ==
|
|
CmpInst::getInversePredicate(OurPredicate)) {
|
|
addPredicateUsers(PI, I);
|
|
// Inverse predicate, we know the other was false, so this is true.
|
|
return createConstantExpression(
|
|
ConstantInt::getTrue(CI->getType()));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
// Create expression will take care of simplifyCmpInst
|
|
return createExpression(I);
|
|
}
|
|
|
|
// Substitute and symbolize the value before value numbering.
|
|
const Expression *
|
|
NewGVN::performSymbolicEvaluation(Value *V,
|
|
SmallPtrSetImpl<Value *> &Visited) const {
|
|
const Expression *E = nullptr;
|
|
if (auto *C = dyn_cast<Constant>(V))
|
|
E = createConstantExpression(C);
|
|
else if (isa<Argument>(V) || isa<GlobalVariable>(V)) {
|
|
E = createVariableExpression(V);
|
|
} else {
|
|
// TODO: memory intrinsics.
|
|
// TODO: Some day, we should do the forward propagation and reassociation
|
|
// parts of the algorithm.
|
|
auto *I = cast<Instruction>(V);
|
|
switch (I->getOpcode()) {
|
|
case Instruction::ExtractValue:
|
|
case Instruction::InsertValue:
|
|
E = performSymbolicAggrValueEvaluation(I);
|
|
break;
|
|
case Instruction::PHI: {
|
|
SmallVector<ValPair, 3> Ops;
|
|
auto *PN = cast<PHINode>(I);
|
|
for (unsigned i = 0; i < PN->getNumOperands(); ++i)
|
|
Ops.push_back({PN->getIncomingValue(i), PN->getIncomingBlock(i)});
|
|
// Sort to ensure the invariant createPHIExpression requires is met.
|
|
sortPHIOps(Ops);
|
|
E = performSymbolicPHIEvaluation(Ops, I, getBlockForValue(I));
|
|
} break;
|
|
case Instruction::Call:
|
|
E = performSymbolicCallEvaluation(I);
|
|
break;
|
|
case Instruction::Store:
|
|
E = performSymbolicStoreEvaluation(I);
|
|
break;
|
|
case Instruction::Load:
|
|
E = performSymbolicLoadEvaluation(I);
|
|
break;
|
|
case Instruction::BitCast:
|
|
case Instruction::AddrSpaceCast:
|
|
E = createExpression(I);
|
|
break;
|
|
case Instruction::ICmp:
|
|
case Instruction::FCmp:
|
|
E = performSymbolicCmpEvaluation(I);
|
|
break;
|
|
case Instruction::FNeg:
|
|
case Instruction::Add:
|
|
case Instruction::FAdd:
|
|
case Instruction::Sub:
|
|
case Instruction::FSub:
|
|
case Instruction::Mul:
|
|
case Instruction::FMul:
|
|
case Instruction::UDiv:
|
|
case Instruction::SDiv:
|
|
case Instruction::FDiv:
|
|
case Instruction::URem:
|
|
case Instruction::SRem:
|
|
case Instruction::FRem:
|
|
case Instruction::Shl:
|
|
case Instruction::LShr:
|
|
case Instruction::AShr:
|
|
case Instruction::And:
|
|
case Instruction::Or:
|
|
case Instruction::Xor:
|
|
case Instruction::Trunc:
|
|
case Instruction::ZExt:
|
|
case Instruction::SExt:
|
|
case Instruction::FPToUI:
|
|
case Instruction::FPToSI:
|
|
case Instruction::UIToFP:
|
|
case Instruction::SIToFP:
|
|
case Instruction::FPTrunc:
|
|
case Instruction::FPExt:
|
|
case Instruction::PtrToInt:
|
|
case Instruction::IntToPtr:
|
|
case Instruction::Select:
|
|
case Instruction::ExtractElement:
|
|
case Instruction::InsertElement:
|
|
case Instruction::GetElementPtr:
|
|
E = createExpression(I);
|
|
break;
|
|
case Instruction::ShuffleVector:
|
|
// FIXME: Add support for shufflevector to createExpression.
|
|
return nullptr;
|
|
default:
|
|
return nullptr;
|
|
}
|
|
}
|
|
return E;
|
|
}
|
|
|
|
// Look up a container in a map, and then call a function for each thing in the
|
|
// found container.
|
|
template <typename Map, typename KeyType, typename Func>
|
|
void NewGVN::for_each_found(Map &M, const KeyType &Key, Func F) {
|
|
const auto Result = M.find_as(Key);
|
|
if (Result != M.end())
|
|
for (typename Map::mapped_type::value_type Mapped : Result->second)
|
|
F(Mapped);
|
|
}
|
|
|
|
// Look up a container of values/instructions in a map, and touch all the
|
|
// instructions in the container. Then erase value from the map.
|
|
template <typename Map, typename KeyType>
|
|
void NewGVN::touchAndErase(Map &M, const KeyType &Key) {
|
|
const auto Result = M.find_as(Key);
|
|
if (Result != M.end()) {
|
|
for (const typename Map::mapped_type::value_type Mapped : Result->second)
|
|
TouchedInstructions.set(InstrToDFSNum(Mapped));
|
|
M.erase(Result);
|
|
}
|
|
}
|
|
|
|
void NewGVN::addAdditionalUsers(Value *To, Value *User) const {
|
|
assert(User && To != User);
|
|
if (isa<Instruction>(To))
|
|
AdditionalUsers[To].insert(User);
|
|
}
|
|
|
|
void NewGVN::markUsersTouched(Value *V) {
|
|
// Now mark the users as touched.
|
|
for (auto *User : V->users()) {
|
|
assert(isa<Instruction>(User) && "Use of value not within an instruction?");
|
|
TouchedInstructions.set(InstrToDFSNum(User));
|
|
}
|
|
touchAndErase(AdditionalUsers, V);
|
|
}
|
|
|
|
void NewGVN::addMemoryUsers(const MemoryAccess *To, MemoryAccess *U) const {
|
|
LLVM_DEBUG(dbgs() << "Adding memory user " << *U << " to " << *To << "\n");
|
|
MemoryToUsers[To].insert(U);
|
|
}
|
|
|
|
void NewGVN::markMemoryDefTouched(const MemoryAccess *MA) {
|
|
TouchedInstructions.set(MemoryToDFSNum(MA));
|
|
}
|
|
|
|
void NewGVN::markMemoryUsersTouched(const MemoryAccess *MA) {
|
|
if (isa<MemoryUse>(MA))
|
|
return;
|
|
for (auto U : MA->users())
|
|
TouchedInstructions.set(MemoryToDFSNum(U));
|
|
touchAndErase(MemoryToUsers, MA);
|
|
}
|
|
|
|
// Add I to the set of users of a given predicate.
|
|
void NewGVN::addPredicateUsers(const PredicateBase *PB, Instruction *I) const {
|
|
// Don't add temporary instructions to the user lists.
|
|
if (AllTempInstructions.count(I))
|
|
return;
|
|
|
|
if (auto *PBranch = dyn_cast<PredicateBranch>(PB))
|
|
PredicateToUsers[PBranch->Condition].insert(I);
|
|
else if (auto *PAssume = dyn_cast<PredicateAssume>(PB))
|
|
PredicateToUsers[PAssume->Condition].insert(I);
|
|
}
|
|
|
|
// Touch all the predicates that depend on this instruction.
|
|
void NewGVN::markPredicateUsersTouched(Instruction *I) {
|
|
touchAndErase(PredicateToUsers, I);
|
|
}
|
|
|
|
// Mark users affected by a memory leader change.
|
|
void NewGVN::markMemoryLeaderChangeTouched(CongruenceClass *CC) {
|
|
for (auto M : CC->memory())
|
|
markMemoryDefTouched(M);
|
|
}
|
|
|
|
// Touch the instructions that need to be updated after a congruence class has a
|
|
// leader change, and mark changed values.
|
|
void NewGVN::markValueLeaderChangeTouched(CongruenceClass *CC) {
|
|
for (auto M : *CC) {
|
|
if (auto *I = dyn_cast<Instruction>(M))
|
|
TouchedInstructions.set(InstrToDFSNum(I));
|
|
LeaderChanges.insert(M);
|
|
}
|
|
}
|
|
|
|
// Give a range of things that have instruction DFS numbers, this will return
|
|
// the member of the range with the smallest dfs number.
|
|
template <class T, class Range>
|
|
T *NewGVN::getMinDFSOfRange(const Range &R) const {
|
|
std::pair<T *, unsigned> MinDFS = {nullptr, ~0U};
|
|
for (const auto X : R) {
|
|
auto DFSNum = InstrToDFSNum(X);
|
|
if (DFSNum < MinDFS.second)
|
|
MinDFS = {X, DFSNum};
|
|
}
|
|
return MinDFS.first;
|
|
}
|
|
|
|
// This function returns the MemoryAccess that should be the next leader of
|
|
// congruence class CC, under the assumption that the current leader is going to
|
|
// disappear.
|
|
const MemoryAccess *NewGVN::getNextMemoryLeader(CongruenceClass *CC) const {
|
|
// TODO: If this ends up to slow, we can maintain a next memory leader like we
|
|
// do for regular leaders.
|
|
// Make sure there will be a leader to find.
|
|
assert(!CC->definesNoMemory() && "Can't get next leader if there is none");
|
|
if (CC->getStoreCount() > 0) {
|
|
if (auto *NL = dyn_cast_or_null<StoreInst>(CC->getNextLeader().first))
|
|
return getMemoryAccess(NL);
|
|
// Find the store with the minimum DFS number.
|
|
auto *V = getMinDFSOfRange<Value>(make_filter_range(
|
|
*CC, [&](const Value *V) { return isa<StoreInst>(V); }));
|
|
return getMemoryAccess(cast<StoreInst>(V));
|
|
}
|
|
assert(CC->getStoreCount() == 0);
|
|
|
|
// Given our assertion, hitting this part must mean
|
|
// !OldClass->memory_empty()
|
|
if (CC->memory_size() == 1)
|
|
return *CC->memory_begin();
|
|
return getMinDFSOfRange<const MemoryPhi>(CC->memory());
|
|
}
|
|
|
|
// This function returns the next value leader of a congruence class, under the
|
|
// assumption that the current leader is going away. This should end up being
|
|
// the next most dominating member.
|
|
Value *NewGVN::getNextValueLeader(CongruenceClass *CC) const {
|
|
// We don't need to sort members if there is only 1, and we don't care about
|
|
// sorting the TOP class because everything either gets out of it or is
|
|
// unreachable.
|
|
|
|
if (CC->size() == 1 || CC == TOPClass) {
|
|
return *(CC->begin());
|
|
} else if (CC->getNextLeader().first) {
|
|
++NumGVNAvoidedSortedLeaderChanges;
|
|
return CC->getNextLeader().first;
|
|
} else {
|
|
++NumGVNSortedLeaderChanges;
|
|
// NOTE: If this ends up to slow, we can maintain a dual structure for
|
|
// member testing/insertion, or keep things mostly sorted, and sort only
|
|
// here, or use SparseBitVector or ....
|
|
return getMinDFSOfRange<Value>(*CC);
|
|
}
|
|
}
|
|
|
|
// Move a MemoryAccess, currently in OldClass, to NewClass, including updates to
|
|
// the memory members, etc for the move.
|
|
//
|
|
// The invariants of this function are:
|
|
//
|
|
// - I must be moving to NewClass from OldClass
|
|
// - The StoreCount of OldClass and NewClass is expected to have been updated
|
|
// for I already if it is a store.
|
|
// - The OldClass memory leader has not been updated yet if I was the leader.
|
|
void NewGVN::moveMemoryToNewCongruenceClass(Instruction *I,
|
|
MemoryAccess *InstMA,
|
|
CongruenceClass *OldClass,
|
|
CongruenceClass *NewClass) {
|
|
// If the leader is I, and we had a representative MemoryAccess, it should
|
|
// be the MemoryAccess of OldClass.
|
|
assert((!InstMA || !OldClass->getMemoryLeader() ||
|
|
OldClass->getLeader() != I ||
|
|
MemoryAccessToClass.lookup(OldClass->getMemoryLeader()) ==
|
|
MemoryAccessToClass.lookup(InstMA)) &&
|
|
"Representative MemoryAccess mismatch");
|
|
// First, see what happens to the new class
|
|
if (!NewClass->getMemoryLeader()) {
|
|
// Should be a new class, or a store becoming a leader of a new class.
|
|
assert(NewClass->size() == 1 ||
|
|
(isa<StoreInst>(I) && NewClass->getStoreCount() == 1));
|
|
NewClass->setMemoryLeader(InstMA);
|
|
// Mark it touched if we didn't just create a singleton
|
|
LLVM_DEBUG(dbgs() << "Memory class leader change for class "
|
|
<< NewClass->getID()
|
|
<< " due to new memory instruction becoming leader\n");
|
|
markMemoryLeaderChangeTouched(NewClass);
|
|
}
|
|
setMemoryClass(InstMA, NewClass);
|
|
// Now, fixup the old class if necessary
|
|
if (OldClass->getMemoryLeader() == InstMA) {
|
|
if (!OldClass->definesNoMemory()) {
|
|
OldClass->setMemoryLeader(getNextMemoryLeader(OldClass));
|
|
LLVM_DEBUG(dbgs() << "Memory class leader change for class "
|
|
<< OldClass->getID() << " to "
|
|
<< *OldClass->getMemoryLeader()
|
|
<< " due to removal of old leader " << *InstMA << "\n");
|
|
markMemoryLeaderChangeTouched(OldClass);
|
|
} else
|
|
OldClass->setMemoryLeader(nullptr);
|
|
}
|
|
}
|
|
|
|
// Move a value, currently in OldClass, to be part of NewClass
|
|
// Update OldClass and NewClass for the move (including changing leaders, etc).
|
|
void NewGVN::moveValueToNewCongruenceClass(Instruction *I, const Expression *E,
|
|
CongruenceClass *OldClass,
|
|
CongruenceClass *NewClass) {
|
|
if (I == OldClass->getNextLeader().first)
|
|
OldClass->resetNextLeader();
|
|
|
|
OldClass->erase(I);
|
|
NewClass->insert(I);
|
|
|
|
if (NewClass->getLeader() != I)
|
|
NewClass->addPossibleNextLeader({I, InstrToDFSNum(I)});
|
|
// Handle our special casing of stores.
|
|
if (auto *SI = dyn_cast<StoreInst>(I)) {
|
|
OldClass->decStoreCount();
|
|
// Okay, so when do we want to make a store a leader of a class?
|
|
// If we have a store defined by an earlier load, we want the earlier load
|
|
// to lead the class.
|
|
// If we have a store defined by something else, we want the store to lead
|
|
// the class so everything else gets the "something else" as a value.
|
|
// If we have a store as the single member of the class, we want the store
|
|
// as the leader
|
|
if (NewClass->getStoreCount() == 0 && !NewClass->getStoredValue()) {
|
|
// If it's a store expression we are using, it means we are not equivalent
|
|
// to something earlier.
|
|
if (auto *SE = dyn_cast<StoreExpression>(E)) {
|
|
NewClass->setStoredValue(SE->getStoredValue());
|
|
markValueLeaderChangeTouched(NewClass);
|
|
// Shift the new class leader to be the store
|
|
LLVM_DEBUG(dbgs() << "Changing leader of congruence class "
|
|
<< NewClass->getID() << " from "
|
|
<< *NewClass->getLeader() << " to " << *SI
|
|
<< " because store joined class\n");
|
|
// If we changed the leader, we have to mark it changed because we don't
|
|
// know what it will do to symbolic evaluation.
|
|
NewClass->setLeader(SI);
|
|
}
|
|
// We rely on the code below handling the MemoryAccess change.
|
|
}
|
|
NewClass->incStoreCount();
|
|
}
|
|
// True if there is no memory instructions left in a class that had memory
|
|
// instructions before.
|
|
|
|
// If it's not a memory use, set the MemoryAccess equivalence
|
|
auto *InstMA = dyn_cast_or_null<MemoryDef>(getMemoryAccess(I));
|
|
if (InstMA)
|
|
moveMemoryToNewCongruenceClass(I, InstMA, OldClass, NewClass);
|
|
ValueToClass[I] = NewClass;
|
|
// See if we destroyed the class or need to swap leaders.
|
|
if (OldClass->empty() && OldClass != TOPClass) {
|
|
if (OldClass->getDefiningExpr()) {
|
|
LLVM_DEBUG(dbgs() << "Erasing expression " << *OldClass->getDefiningExpr()
|
|
<< " from table\n");
|
|
// We erase it as an exact expression to make sure we don't just erase an
|
|
// equivalent one.
|
|
auto Iter = ExpressionToClass.find_as(
|
|
ExactEqualsExpression(*OldClass->getDefiningExpr()));
|
|
if (Iter != ExpressionToClass.end())
|
|
ExpressionToClass.erase(Iter);
|
|
#ifdef EXPENSIVE_CHECKS
|
|
assert(
|
|
(*OldClass->getDefiningExpr() != *E || ExpressionToClass.lookup(E)) &&
|
|
"We erased the expression we just inserted, which should not happen");
|
|
#endif
|
|
}
|
|
} else if (OldClass->getLeader() == I) {
|
|
// When the leader changes, the value numbering of
|
|
// everything may change due to symbolization changes, so we need to
|
|
// reprocess.
|
|
LLVM_DEBUG(dbgs() << "Value class leader change for class "
|
|
<< OldClass->getID() << "\n");
|
|
++NumGVNLeaderChanges;
|
|
// Destroy the stored value if there are no more stores to represent it.
|
|
// Note that this is basically clean up for the expression removal that
|
|
// happens below. If we remove stores from a class, we may leave it as a
|
|
// class of equivalent memory phis.
|
|
if (OldClass->getStoreCount() == 0) {
|
|
if (OldClass->getStoredValue())
|
|
OldClass->setStoredValue(nullptr);
|
|
}
|
|
OldClass->setLeader(getNextValueLeader(OldClass));
|
|
OldClass->resetNextLeader();
|
|
markValueLeaderChangeTouched(OldClass);
|
|
}
|
|
}
|
|
|
|
// For a given expression, mark the phi of ops instructions that could have
|
|
// changed as a result.
|
|
void NewGVN::markPhiOfOpsChanged(const Expression *E) {
|
|
touchAndErase(ExpressionToPhiOfOps, E);
|
|
}
|
|
|
|
// Perform congruence finding on a given value numbering expression.
|
|
void NewGVN::performCongruenceFinding(Instruction *I, const Expression *E) {
|
|
// This is guaranteed to return something, since it will at least find
|
|
// TOP.
|
|
|
|
CongruenceClass *IClass = ValueToClass.lookup(I);
|
|
assert(IClass && "Should have found a IClass");
|
|
// Dead classes should have been eliminated from the mapping.
|
|
assert(!IClass->isDead() && "Found a dead class");
|
|
|
|
CongruenceClass *EClass = nullptr;
|
|
if (const auto *VE = dyn_cast<VariableExpression>(E)) {
|
|
EClass = ValueToClass.lookup(VE->getVariableValue());
|
|
} else if (isa<DeadExpression>(E)) {
|
|
EClass = TOPClass;
|
|
}
|
|
if (!EClass) {
|
|
auto lookupResult = ExpressionToClass.insert({E, nullptr});
|
|
|
|
// If it's not in the value table, create a new congruence class.
|
|
if (lookupResult.second) {
|
|
CongruenceClass *NewClass = createCongruenceClass(nullptr, E);
|
|
auto place = lookupResult.first;
|
|
place->second = NewClass;
|
|
|
|
// Constants and variables should always be made the leader.
|
|
if (const auto *CE = dyn_cast<ConstantExpression>(E)) {
|
|
NewClass->setLeader(CE->getConstantValue());
|
|
} else if (const auto *SE = dyn_cast<StoreExpression>(E)) {
|
|
StoreInst *SI = SE->getStoreInst();
|
|
NewClass->setLeader(SI);
|
|
NewClass->setStoredValue(SE->getStoredValue());
|
|
// The RepMemoryAccess field will be filled in properly by the
|
|
// moveValueToNewCongruenceClass call.
|
|
} else {
|
|
NewClass->setLeader(I);
|
|
}
|
|
assert(!isa<VariableExpression>(E) &&
|
|
"VariableExpression should have been handled already");
|
|
|
|
EClass = NewClass;
|
|
LLVM_DEBUG(dbgs() << "Created new congruence class for " << *I
|
|
<< " using expression " << *E << " at "
|
|
<< NewClass->getID() << " and leader "
|
|
<< *(NewClass->getLeader()));
|
|
if (NewClass->getStoredValue())
|
|
LLVM_DEBUG(dbgs() << " and stored value "
|
|
<< *(NewClass->getStoredValue()));
|
|
LLVM_DEBUG(dbgs() << "\n");
|
|
} else {
|
|
EClass = lookupResult.first->second;
|
|
if (isa<ConstantExpression>(E))
|
|
assert((isa<Constant>(EClass->getLeader()) ||
|
|
(EClass->getStoredValue() &&
|
|
isa<Constant>(EClass->getStoredValue()))) &&
|
|
"Any class with a constant expression should have a "
|
|
"constant leader");
|
|
|
|
assert(EClass && "Somehow don't have an eclass");
|
|
|
|
assert(!EClass->isDead() && "We accidentally looked up a dead class");
|
|
}
|
|
}
|
|
bool ClassChanged = IClass != EClass;
|
|
bool LeaderChanged = LeaderChanges.erase(I);
|
|
if (ClassChanged || LeaderChanged) {
|
|
LLVM_DEBUG(dbgs() << "New class " << EClass->getID() << " for expression "
|
|
<< *E << "\n");
|
|
if (ClassChanged) {
|
|
moveValueToNewCongruenceClass(I, E, IClass, EClass);
|
|
markPhiOfOpsChanged(E);
|
|
}
|
|
|
|
markUsersTouched(I);
|
|
if (MemoryAccess *MA = getMemoryAccess(I))
|
|
markMemoryUsersTouched(MA);
|
|
if (auto *CI = dyn_cast<CmpInst>(I))
|
|
markPredicateUsersTouched(CI);
|
|
}
|
|
// If we changed the class of the store, we want to ensure nothing finds the
|
|
// old store expression. In particular, loads do not compare against stored
|
|
// value, so they will find old store expressions (and associated class
|
|
// mappings) if we leave them in the table.
|
|
if (ClassChanged && isa<StoreInst>(I)) {
|
|
auto *OldE = ValueToExpression.lookup(I);
|
|
// It could just be that the old class died. We don't want to erase it if we
|
|
// just moved classes.
|
|
if (OldE && isa<StoreExpression>(OldE) && *E != *OldE) {
|
|
// Erase this as an exact expression to ensure we don't erase expressions
|
|
// equivalent to it.
|
|
auto Iter = ExpressionToClass.find_as(ExactEqualsExpression(*OldE));
|
|
if (Iter != ExpressionToClass.end())
|
|
ExpressionToClass.erase(Iter);
|
|
}
|
|
}
|
|
ValueToExpression[I] = E;
|
|
}
|
|
|
|
// Process the fact that Edge (from, to) is reachable, including marking
|
|
// any newly reachable blocks and instructions for processing.
|
|
void NewGVN::updateReachableEdge(BasicBlock *From, BasicBlock *To) {
|
|
// Check if the Edge was reachable before.
|
|
if (ReachableEdges.insert({From, To}).second) {
|
|
// If this block wasn't reachable before, all instructions are touched.
|
|
if (ReachableBlocks.insert(To).second) {
|
|
LLVM_DEBUG(dbgs() << "Block " << getBlockName(To)
|
|
<< " marked reachable\n");
|
|
const auto &InstRange = BlockInstRange.lookup(To);
|
|
TouchedInstructions.set(InstRange.first, InstRange.second);
|
|
} else {
|
|
LLVM_DEBUG(dbgs() << "Block " << getBlockName(To)
|
|
<< " was reachable, but new edge {"
|
|
<< getBlockName(From) << "," << getBlockName(To)
|
|
<< "} to it found\n");
|
|
|
|
// We've made an edge reachable to an existing block, which may
|
|
// impact predicates. Otherwise, only mark the phi nodes as touched, as
|
|
// they are the only thing that depend on new edges. Anything using their
|
|
// values will get propagated to if necessary.
|
|
if (MemoryAccess *MemPhi = getMemoryAccess(To))
|
|
TouchedInstructions.set(InstrToDFSNum(MemPhi));
|
|
|
|
// FIXME: We should just add a union op on a Bitvector and
|
|
// SparseBitVector. We can do it word by word faster than we are doing it
|
|
// here.
|
|
for (auto InstNum : RevisitOnReachabilityChange[To])
|
|
TouchedInstructions.set(InstNum);
|
|
}
|
|
}
|
|
}
|
|
|
|
// Given a predicate condition (from a switch, cmp, or whatever) and a block,
|
|
// see if we know some constant value for it already.
|
|
Value *NewGVN::findConditionEquivalence(Value *Cond) const {
|
|
auto Result = lookupOperandLeader(Cond);
|
|
return isa<Constant>(Result) ? Result : nullptr;
|
|
}
|
|
|
|
// Process the outgoing edges of a block for reachability.
|
|
void NewGVN::processOutgoingEdges(Instruction *TI, BasicBlock *B) {
|
|
// Evaluate reachability of terminator instruction.
|
|
Value *Cond;
|
|
BasicBlock *TrueSucc, *FalseSucc;
|
|
if (match(TI, m_Br(m_Value(Cond), TrueSucc, FalseSucc))) {
|
|
Value *CondEvaluated = findConditionEquivalence(Cond);
|
|
if (!CondEvaluated) {
|
|
if (auto *I = dyn_cast<Instruction>(Cond)) {
|
|
const Expression *E = createExpression(I);
|
|
if (const auto *CE = dyn_cast<ConstantExpression>(E)) {
|
|
CondEvaluated = CE->getConstantValue();
|
|
}
|
|
} else if (isa<ConstantInt>(Cond)) {
|
|
CondEvaluated = Cond;
|
|
}
|
|
}
|
|
ConstantInt *CI;
|
|
if (CondEvaluated && (CI = dyn_cast<ConstantInt>(CondEvaluated))) {
|
|
if (CI->isOne()) {
|
|
LLVM_DEBUG(dbgs() << "Condition for Terminator " << *TI
|
|
<< " evaluated to true\n");
|
|
updateReachableEdge(B, TrueSucc);
|
|
} else if (CI->isZero()) {
|
|
LLVM_DEBUG(dbgs() << "Condition for Terminator " << *TI
|
|
<< " evaluated to false\n");
|
|
updateReachableEdge(B, FalseSucc);
|
|
}
|
|
} else {
|
|
updateReachableEdge(B, TrueSucc);
|
|
updateReachableEdge(B, FalseSucc);
|
|
}
|
|
} else if (auto *SI = dyn_cast<SwitchInst>(TI)) {
|
|
// For switches, propagate the case values into the case
|
|
// destinations.
|
|
|
|
Value *SwitchCond = SI->getCondition();
|
|
Value *CondEvaluated = findConditionEquivalence(SwitchCond);
|
|
// See if we were able to turn this switch statement into a constant.
|
|
if (CondEvaluated && isa<ConstantInt>(CondEvaluated)) {
|
|
auto *CondVal = cast<ConstantInt>(CondEvaluated);
|
|
// We should be able to get case value for this.
|
|
auto Case = *SI->findCaseValue(CondVal);
|
|
if (Case.getCaseSuccessor() == SI->getDefaultDest()) {
|
|
// We proved the value is outside of the range of the case.
|
|
// We can't do anything other than mark the default dest as reachable,
|
|
// and go home.
|
|
updateReachableEdge(B, SI->getDefaultDest());
|
|
return;
|
|
}
|
|
// Now get where it goes and mark it reachable.
|
|
BasicBlock *TargetBlock = Case.getCaseSuccessor();
|
|
updateReachableEdge(B, TargetBlock);
|
|
} else {
|
|
for (unsigned i = 0, e = SI->getNumSuccessors(); i != e; ++i) {
|
|
BasicBlock *TargetBlock = SI->getSuccessor(i);
|
|
updateReachableEdge(B, TargetBlock);
|
|
}
|
|
}
|
|
} else {
|
|
// Otherwise this is either unconditional, or a type we have no
|
|
// idea about. Just mark successors as reachable.
|
|
for (unsigned i = 0, e = TI->getNumSuccessors(); i != e; ++i) {
|
|
BasicBlock *TargetBlock = TI->getSuccessor(i);
|
|
updateReachableEdge(B, TargetBlock);
|
|
}
|
|
|
|
// This also may be a memory defining terminator, in which case, set it
|
|
// equivalent only to itself.
|
|
//
|
|
auto *MA = getMemoryAccess(TI);
|
|
if (MA && !isa<MemoryUse>(MA)) {
|
|
auto *CC = ensureLeaderOfMemoryClass(MA);
|
|
if (setMemoryClass(MA, CC))
|
|
markMemoryUsersTouched(MA);
|
|
}
|
|
}
|
|
}
|
|
|
|
// Remove the PHI of Ops PHI for I
|
|
void NewGVN::removePhiOfOps(Instruction *I, PHINode *PHITemp) {
|
|
InstrDFS.erase(PHITemp);
|
|
// It's still a temp instruction. We keep it in the array so it gets erased.
|
|
// However, it's no longer used by I, or in the block
|
|
TempToBlock.erase(PHITemp);
|
|
RealToTemp.erase(I);
|
|
// We don't remove the users from the phi node uses. This wastes a little
|
|
// time, but such is life. We could use two sets to track which were there
|
|
// are the start of NewGVN, and which were added, but right nowt he cost of
|
|
// tracking is more than the cost of checking for more phi of ops.
|
|
}
|
|
|
|
// Add PHI Op in BB as a PHI of operations version of ExistingValue.
|
|
void NewGVN::addPhiOfOps(PHINode *Op, BasicBlock *BB,
|
|
Instruction *ExistingValue) {
|
|
InstrDFS[Op] = InstrToDFSNum(ExistingValue);
|
|
AllTempInstructions.insert(Op);
|
|
TempToBlock[Op] = BB;
|
|
RealToTemp[ExistingValue] = Op;
|
|
// Add all users to phi node use, as they are now uses of the phi of ops phis
|
|
// and may themselves be phi of ops.
|
|
for (auto *U : ExistingValue->users())
|
|
if (auto *UI = dyn_cast<Instruction>(U))
|
|
PHINodeUses.insert(UI);
|
|
}
|
|
|
|
static bool okayForPHIOfOps(const Instruction *I) {
|
|
if (!EnablePhiOfOps)
|
|
return false;
|
|
return isa<BinaryOperator>(I) || isa<SelectInst>(I) || isa<CmpInst>(I) ||
|
|
isa<LoadInst>(I);
|
|
}
|
|
|
|
bool NewGVN::OpIsSafeForPHIOfOpsHelper(
|
|
Value *V, const BasicBlock *PHIBlock,
|
|
SmallPtrSetImpl<const Value *> &Visited,
|
|
SmallVectorImpl<Instruction *> &Worklist) {
|
|
|
|
if (!isa<Instruction>(V))
|
|
return true;
|
|
auto OISIt = OpSafeForPHIOfOps.find(V);
|
|
if (OISIt != OpSafeForPHIOfOps.end())
|
|
return OISIt->second;
|
|
|
|
// Keep walking until we either dominate the phi block, or hit a phi, or run
|
|
// out of things to check.
|
|
if (DT->properlyDominates(getBlockForValue(V), PHIBlock)) {
|
|
OpSafeForPHIOfOps.insert({V, true});
|
|
return true;
|
|
}
|
|
// PHI in the same block.
|
|
if (isa<PHINode>(V) && getBlockForValue(V) == PHIBlock) {
|
|
OpSafeForPHIOfOps.insert({V, false});
|
|
return false;
|
|
}
|
|
|
|
auto *OrigI = cast<Instruction>(V);
|
|
for (auto *Op : OrigI->operand_values()) {
|
|
if (!isa<Instruction>(Op))
|
|
continue;
|
|
// Stop now if we find an unsafe operand.
|
|
auto OISIt = OpSafeForPHIOfOps.find(OrigI);
|
|
if (OISIt != OpSafeForPHIOfOps.end()) {
|
|
if (!OISIt->second) {
|
|
OpSafeForPHIOfOps.insert({V, false});
|
|
return false;
|
|
}
|
|
continue;
|
|
}
|
|
if (!Visited.insert(Op).second)
|
|
continue;
|
|
Worklist.push_back(cast<Instruction>(Op));
|
|
}
|
|
return true;
|
|
}
|
|
|
|
// Return true if this operand will be safe to use for phi of ops.
|
|
//
|
|
// The reason some operands are unsafe is that we are not trying to recursively
|
|
// translate everything back through phi nodes. We actually expect some lookups
|
|
// of expressions to fail. In particular, a lookup where the expression cannot
|
|
// exist in the predecessor. This is true even if the expression, as shown, can
|
|
// be determined to be constant.
|
|
bool NewGVN::OpIsSafeForPHIOfOps(Value *V, const BasicBlock *PHIBlock,
|
|
SmallPtrSetImpl<const Value *> &Visited) {
|
|
SmallVector<Instruction *, 4> Worklist;
|
|
if (!OpIsSafeForPHIOfOpsHelper(V, PHIBlock, Visited, Worklist))
|
|
return false;
|
|
while (!Worklist.empty()) {
|
|
auto *I = Worklist.pop_back_val();
|
|
if (!OpIsSafeForPHIOfOpsHelper(I, PHIBlock, Visited, Worklist))
|
|
return false;
|
|
}
|
|
OpSafeForPHIOfOps.insert({V, true});
|
|
return true;
|
|
}
|
|
|
|
// Try to find a leader for instruction TransInst, which is a phi translated
|
|
// version of something in our original program. Visited is used to ensure we
|
|
// don't infinite loop during translations of cycles. OrigInst is the
|
|
// instruction in the original program, and PredBB is the predecessor we
|
|
// translated it through.
|
|
Value *NewGVN::findLeaderForInst(Instruction *TransInst,
|
|
SmallPtrSetImpl<Value *> &Visited,
|
|
MemoryAccess *MemAccess, Instruction *OrigInst,
|
|
BasicBlock *PredBB) {
|
|
unsigned IDFSNum = InstrToDFSNum(OrigInst);
|
|
// Make sure it's marked as a temporary instruction.
|
|
AllTempInstructions.insert(TransInst);
|
|
// and make sure anything that tries to add it's DFS number is
|
|
// redirected to the instruction we are making a phi of ops
|
|
// for.
|
|
TempToBlock.insert({TransInst, PredBB});
|
|
InstrDFS.insert({TransInst, IDFSNum});
|
|
|
|
const Expression *E = performSymbolicEvaluation(TransInst, Visited);
|
|
InstrDFS.erase(TransInst);
|
|
AllTempInstructions.erase(TransInst);
|
|
TempToBlock.erase(TransInst);
|
|
if (MemAccess)
|
|
TempToMemory.erase(TransInst);
|
|
if (!E)
|
|
return nullptr;
|
|
auto *FoundVal = findPHIOfOpsLeader(E, OrigInst, PredBB);
|
|
if (!FoundVal) {
|
|
ExpressionToPhiOfOps[E].insert(OrigInst);
|
|
LLVM_DEBUG(dbgs() << "Cannot find phi of ops operand for " << *TransInst
|
|
<< " in block " << getBlockName(PredBB) << "\n");
|
|
return nullptr;
|
|
}
|
|
if (auto *SI = dyn_cast<StoreInst>(FoundVal))
|
|
FoundVal = SI->getValueOperand();
|
|
return FoundVal;
|
|
}
|
|
|
|
// When we see an instruction that is an op of phis, generate the equivalent phi
|
|
// of ops form.
|
|
const Expression *
|
|
NewGVN::makePossiblePHIOfOps(Instruction *I,
|
|
SmallPtrSetImpl<Value *> &Visited) {
|
|
if (!okayForPHIOfOps(I))
|
|
return nullptr;
|
|
|
|
if (!Visited.insert(I).second)
|
|
return nullptr;
|
|
// For now, we require the instruction be cycle free because we don't
|
|
// *always* create a phi of ops for instructions that could be done as phi
|
|
// of ops, we only do it if we think it is useful. If we did do it all the
|
|
// time, we could remove the cycle free check.
|
|
if (!isCycleFree(I))
|
|
return nullptr;
|
|
|
|
SmallPtrSet<const Value *, 8> ProcessedPHIs;
|
|
// TODO: We don't do phi translation on memory accesses because it's
|
|
// complicated. For a load, we'd need to be able to simulate a new memoryuse,
|
|
// which we don't have a good way of doing ATM.
|
|
auto *MemAccess = getMemoryAccess(I);
|
|
// If the memory operation is defined by a memory operation this block that
|
|
// isn't a MemoryPhi, transforming the pointer backwards through a scalar phi
|
|
// can't help, as it would still be killed by that memory operation.
|
|
if (MemAccess && !isa<MemoryPhi>(MemAccess->getDefiningAccess()) &&
|
|
MemAccess->getDefiningAccess()->getBlock() == I->getParent())
|
|
return nullptr;
|
|
|
|
// Convert op of phis to phi of ops
|
|
SmallPtrSet<const Value *, 10> VisitedOps;
|
|
SmallVector<Value *, 4> Ops(I->operand_values());
|
|
BasicBlock *SamePHIBlock = nullptr;
|
|
PHINode *OpPHI = nullptr;
|
|
if (!DebugCounter::shouldExecute(PHIOfOpsCounter))
|
|
return nullptr;
|
|
for (auto *Op : Ops) {
|
|
if (!isa<PHINode>(Op)) {
|
|
auto *ValuePHI = RealToTemp.lookup(Op);
|
|
if (!ValuePHI)
|
|
continue;
|
|
LLVM_DEBUG(dbgs() << "Found possible dependent phi of ops\n");
|
|
Op = ValuePHI;
|
|
}
|
|
OpPHI = cast<PHINode>(Op);
|
|
if (!SamePHIBlock) {
|
|
SamePHIBlock = getBlockForValue(OpPHI);
|
|
} else if (SamePHIBlock != getBlockForValue(OpPHI)) {
|
|
LLVM_DEBUG(
|
|
dbgs()
|
|
<< "PHIs for operands are not all in the same block, aborting\n");
|
|
return nullptr;
|
|
}
|
|
// No point in doing this for one-operand phis.
|
|
if (OpPHI->getNumOperands() == 1) {
|
|
OpPHI = nullptr;
|
|
continue;
|
|
}
|
|
}
|
|
|
|
if (!OpPHI)
|
|
return nullptr;
|
|
|
|
SmallVector<ValPair, 4> PHIOps;
|
|
SmallPtrSet<Value *, 4> Deps;
|
|
auto *PHIBlock = getBlockForValue(OpPHI);
|
|
RevisitOnReachabilityChange[PHIBlock].reset(InstrToDFSNum(I));
|
|
for (unsigned PredNum = 0; PredNum < OpPHI->getNumOperands(); ++PredNum) {
|
|
auto *PredBB = OpPHI->getIncomingBlock(PredNum);
|
|
Value *FoundVal = nullptr;
|
|
SmallPtrSet<Value *, 4> CurrentDeps;
|
|
// We could just skip unreachable edges entirely but it's tricky to do
|
|
// with rewriting existing phi nodes.
|
|
if (ReachableEdges.count({PredBB, PHIBlock})) {
|
|
// Clone the instruction, create an expression from it that is
|
|
// translated back into the predecessor, and see if we have a leader.
|
|
Instruction *ValueOp = I->clone();
|
|
if (MemAccess)
|
|
TempToMemory.insert({ValueOp, MemAccess});
|
|
bool SafeForPHIOfOps = true;
|
|
VisitedOps.clear();
|
|
for (auto &Op : ValueOp->operands()) {
|
|
auto *OrigOp = &*Op;
|
|
// When these operand changes, it could change whether there is a
|
|
// leader for us or not, so we have to add additional users.
|
|
if (isa<PHINode>(Op)) {
|
|
Op = Op->DoPHITranslation(PHIBlock, PredBB);
|
|
if (Op != OrigOp && Op != I)
|
|
CurrentDeps.insert(Op);
|
|
} else if (auto *ValuePHI = RealToTemp.lookup(Op)) {
|
|
if (getBlockForValue(ValuePHI) == PHIBlock)
|
|
Op = ValuePHI->getIncomingValueForBlock(PredBB);
|
|
}
|
|
// If we phi-translated the op, it must be safe.
|
|
SafeForPHIOfOps =
|
|
SafeForPHIOfOps &&
|
|
(Op != OrigOp || OpIsSafeForPHIOfOps(Op, PHIBlock, VisitedOps));
|
|
}
|
|
// FIXME: For those things that are not safe we could generate
|
|
// expressions all the way down, and see if this comes out to a
|
|
// constant. For anything where that is true, and unsafe, we should
|
|
// have made a phi-of-ops (or value numbered it equivalent to something)
|
|
// for the pieces already.
|
|
FoundVal = !SafeForPHIOfOps ? nullptr
|
|
: findLeaderForInst(ValueOp, Visited,
|
|
MemAccess, I, PredBB);
|
|
ValueOp->deleteValue();
|
|
if (!FoundVal) {
|
|
// We failed to find a leader for the current ValueOp, but this might
|
|
// change in case of the translated operands change.
|
|
if (SafeForPHIOfOps)
|
|
for (auto Dep : CurrentDeps)
|
|
addAdditionalUsers(Dep, I);
|
|
|
|
return nullptr;
|
|
}
|
|
Deps.insert(CurrentDeps.begin(), CurrentDeps.end());
|
|
} else {
|
|
LLVM_DEBUG(dbgs() << "Skipping phi of ops operand for incoming block "
|
|
<< getBlockName(PredBB)
|
|
<< " because the block is unreachable\n");
|
|
FoundVal = UndefValue::get(I->getType());
|
|
RevisitOnReachabilityChange[PHIBlock].set(InstrToDFSNum(I));
|
|
}
|
|
|
|
PHIOps.push_back({FoundVal, PredBB});
|
|
LLVM_DEBUG(dbgs() << "Found phi of ops operand " << *FoundVal << " in "
|
|
<< getBlockName(PredBB) << "\n");
|
|
}
|
|
for (auto Dep : Deps)
|
|
addAdditionalUsers(Dep, I);
|
|
sortPHIOps(PHIOps);
|
|
auto *E = performSymbolicPHIEvaluation(PHIOps, I, PHIBlock);
|
|
if (isa<ConstantExpression>(E) || isa<VariableExpression>(E)) {
|
|
LLVM_DEBUG(
|
|
dbgs()
|
|
<< "Not creating real PHI of ops because it simplified to existing "
|
|
"value or constant\n");
|
|
return E;
|
|
}
|
|
auto *ValuePHI = RealToTemp.lookup(I);
|
|
bool NewPHI = false;
|
|
if (!ValuePHI) {
|
|
ValuePHI =
|
|
PHINode::Create(I->getType(), OpPHI->getNumOperands(), "phiofops");
|
|
addPhiOfOps(ValuePHI, PHIBlock, I);
|
|
NewPHI = true;
|
|
NumGVNPHIOfOpsCreated++;
|
|
}
|
|
if (NewPHI) {
|
|
for (auto PHIOp : PHIOps)
|
|
ValuePHI->addIncoming(PHIOp.first, PHIOp.second);
|
|
} else {
|
|
TempToBlock[ValuePHI] = PHIBlock;
|
|
unsigned int i = 0;
|
|
for (auto PHIOp : PHIOps) {
|
|
ValuePHI->setIncomingValue(i, PHIOp.first);
|
|
ValuePHI->setIncomingBlock(i, PHIOp.second);
|
|
++i;
|
|
}
|
|
}
|
|
RevisitOnReachabilityChange[PHIBlock].set(InstrToDFSNum(I));
|
|
LLVM_DEBUG(dbgs() << "Created phi of ops " << *ValuePHI << " for " << *I
|
|
<< "\n");
|
|
|
|
return E;
|
|
}
|
|
|
|
// The algorithm initially places the values of the routine in the TOP
|
|
// congruence class. The leader of TOP is the undetermined value `undef`.
|
|
// When the algorithm has finished, values still in TOP are unreachable.
|
|
void NewGVN::initializeCongruenceClasses(Function &F) {
|
|
NextCongruenceNum = 0;
|
|
|
|
// Note that even though we use the live on entry def as a representative
|
|
// MemoryAccess, it is *not* the same as the actual live on entry def. We
|
|
// have no real equivalemnt to undef for MemoryAccesses, and so we really
|
|
// should be checking whether the MemoryAccess is top if we want to know if it
|
|
// is equivalent to everything. Otherwise, what this really signifies is that
|
|
// the access "it reaches all the way back to the beginning of the function"
|
|
|
|
// Initialize all other instructions to be in TOP class.
|
|
TOPClass = createCongruenceClass(nullptr, nullptr);
|
|
TOPClass->setMemoryLeader(MSSA->getLiveOnEntryDef());
|
|
// The live on entry def gets put into it's own class
|
|
MemoryAccessToClass[MSSA->getLiveOnEntryDef()] =
|
|
createMemoryClass(MSSA->getLiveOnEntryDef());
|
|
|
|
for (auto DTN : nodes(DT)) {
|
|
BasicBlock *BB = DTN->getBlock();
|
|
// All MemoryAccesses are equivalent to live on entry to start. They must
|
|
// be initialized to something so that initial changes are noticed. For
|
|
// the maximal answer, we initialize them all to be the same as
|
|
// liveOnEntry.
|
|
auto *MemoryBlockDefs = MSSA->getBlockDefs(BB);
|
|
if (MemoryBlockDefs)
|
|
for (const auto &Def : *MemoryBlockDefs) {
|
|
MemoryAccessToClass[&Def] = TOPClass;
|
|
auto *MD = dyn_cast<MemoryDef>(&Def);
|
|
// Insert the memory phis into the member list.
|
|
if (!MD) {
|
|
const MemoryPhi *MP = cast<MemoryPhi>(&Def);
|
|
TOPClass->memory_insert(MP);
|
|
MemoryPhiState.insert({MP, MPS_TOP});
|
|
}
|
|
|
|
if (MD && isa<StoreInst>(MD->getMemoryInst()))
|
|
TOPClass->incStoreCount();
|
|
}
|
|
|
|
// FIXME: This is trying to discover which instructions are uses of phi
|
|
// nodes. We should move this into one of the myriad of places that walk
|
|
// all the operands already.
|
|
for (auto &I : *BB) {
|
|
if (isa<PHINode>(&I))
|
|
for (auto *U : I.users())
|
|
if (auto *UInst = dyn_cast<Instruction>(U))
|
|
if (InstrToDFSNum(UInst) != 0 && okayForPHIOfOps(UInst))
|
|
PHINodeUses.insert(UInst);
|
|
// Don't insert void terminators into the class. We don't value number
|
|
// them, and they just end up sitting in TOP.
|
|
if (I.isTerminator() && I.getType()->isVoidTy())
|
|
continue;
|
|
TOPClass->insert(&I);
|
|
ValueToClass[&I] = TOPClass;
|
|
}
|
|
}
|
|
|
|
// Initialize arguments to be in their own unique congruence classes
|
|
for (auto &FA : F.args())
|
|
createSingletonCongruenceClass(&FA);
|
|
}
|
|
|
|
void NewGVN::cleanupTables() {
|
|
for (unsigned i = 0, e = CongruenceClasses.size(); i != e; ++i) {
|
|
LLVM_DEBUG(dbgs() << "Congruence class " << CongruenceClasses[i]->getID()
|
|
<< " has " << CongruenceClasses[i]->size()
|
|
<< " members\n");
|
|
// Make sure we delete the congruence class (probably worth switching to
|
|
// a unique_ptr at some point.
|
|
delete CongruenceClasses[i];
|
|
CongruenceClasses[i] = nullptr;
|
|
}
|
|
|
|
// Destroy the value expressions
|
|
SmallVector<Instruction *, 8> TempInst(AllTempInstructions.begin(),
|
|
AllTempInstructions.end());
|
|
AllTempInstructions.clear();
|
|
|
|
// We have to drop all references for everything first, so there are no uses
|
|
// left as we delete them.
|
|
for (auto *I : TempInst) {
|
|
I->dropAllReferences();
|
|
}
|
|
|
|
while (!TempInst.empty()) {
|
|
auto *I = TempInst.back();
|
|
TempInst.pop_back();
|
|
I->deleteValue();
|
|
}
|
|
|
|
ValueToClass.clear();
|
|
ArgRecycler.clear(ExpressionAllocator);
|
|
ExpressionAllocator.Reset();
|
|
CongruenceClasses.clear();
|
|
ExpressionToClass.clear();
|
|
ValueToExpression.clear();
|
|
RealToTemp.clear();
|
|
AdditionalUsers.clear();
|
|
ExpressionToPhiOfOps.clear();
|
|
TempToBlock.clear();
|
|
TempToMemory.clear();
|
|
PHINodeUses.clear();
|
|
OpSafeForPHIOfOps.clear();
|
|
ReachableBlocks.clear();
|
|
ReachableEdges.clear();
|
|
#ifndef NDEBUG
|
|
ProcessedCount.clear();
|
|
#endif
|
|
InstrDFS.clear();
|
|
InstructionsToErase.clear();
|
|
DFSToInstr.clear();
|
|
BlockInstRange.clear();
|
|
TouchedInstructions.clear();
|
|
MemoryAccessToClass.clear();
|
|
PredicateToUsers.clear();
|
|
MemoryToUsers.clear();
|
|
RevisitOnReachabilityChange.clear();
|
|
}
|
|
|
|
// Assign local DFS number mapping to instructions, and leave space for Value
|
|
// PHI's.
|
|
std::pair<unsigned, unsigned> NewGVN::assignDFSNumbers(BasicBlock *B,
|
|
unsigned Start) {
|
|
unsigned End = Start;
|
|
if (MemoryAccess *MemPhi = getMemoryAccess(B)) {
|
|
InstrDFS[MemPhi] = End++;
|
|
DFSToInstr.emplace_back(MemPhi);
|
|
}
|
|
|
|
// Then the real block goes next.
|
|
for (auto &I : *B) {
|
|
// There's no need to call isInstructionTriviallyDead more than once on
|
|
// an instruction. Therefore, once we know that an instruction is dead
|
|
// we change its DFS number so that it doesn't get value numbered.
|
|
if (isInstructionTriviallyDead(&I, TLI)) {
|
|
InstrDFS[&I] = 0;
|
|
LLVM_DEBUG(dbgs() << "Skipping trivially dead instruction " << I << "\n");
|
|
markInstructionForDeletion(&I);
|
|
continue;
|
|
}
|
|
if (isa<PHINode>(&I))
|
|
RevisitOnReachabilityChange[B].set(End);
|
|
InstrDFS[&I] = End++;
|
|
DFSToInstr.emplace_back(&I);
|
|
}
|
|
|
|
// All of the range functions taken half-open ranges (open on the end side).
|
|
// So we do not subtract one from count, because at this point it is one
|
|
// greater than the last instruction.
|
|
return std::make_pair(Start, End);
|
|
}
|
|
|
|
void NewGVN::updateProcessedCount(const Value *V) {
|
|
#ifndef NDEBUG
|
|
if (ProcessedCount.count(V) == 0) {
|
|
ProcessedCount.insert({V, 1});
|
|
} else {
|
|
++ProcessedCount[V];
|
|
assert(ProcessedCount[V] < 100 &&
|
|
"Seem to have processed the same Value a lot");
|
|
}
|
|
#endif
|
|
}
|
|
|
|
// Evaluate MemoryPhi nodes symbolically, just like PHI nodes
|
|
void NewGVN::valueNumberMemoryPhi(MemoryPhi *MP) {
|
|
// If all the arguments are the same, the MemoryPhi has the same value as the
|
|
// argument. Filter out unreachable blocks and self phis from our operands.
|
|
// TODO: We could do cycle-checking on the memory phis to allow valueizing for
|
|
// self-phi checking.
|
|
const BasicBlock *PHIBlock = MP->getBlock();
|
|
auto Filtered = make_filter_range(MP->operands(), [&](const Use &U) {
|
|
return cast<MemoryAccess>(U) != MP &&
|
|
!isMemoryAccessTOP(cast<MemoryAccess>(U)) &&
|
|
ReachableEdges.count({MP->getIncomingBlock(U), PHIBlock});
|
|
});
|
|
// If all that is left is nothing, our memoryphi is undef. We keep it as
|
|
// InitialClass. Note: The only case this should happen is if we have at
|
|
// least one self-argument.
|
|
if (Filtered.begin() == Filtered.end()) {
|
|
if (setMemoryClass(MP, TOPClass))
|
|
markMemoryUsersTouched(MP);
|
|
return;
|
|
}
|
|
|
|
// Transform the remaining operands into operand leaders.
|
|
// FIXME: mapped_iterator should have a range version.
|
|
auto LookupFunc = [&](const Use &U) {
|
|
return lookupMemoryLeader(cast<MemoryAccess>(U));
|
|
};
|
|
auto MappedBegin = map_iterator(Filtered.begin(), LookupFunc);
|
|
auto MappedEnd = map_iterator(Filtered.end(), LookupFunc);
|
|
|
|
// and now check if all the elements are equal.
|
|
// Sadly, we can't use std::equals since these are random access iterators.
|
|
const auto *AllSameValue = *MappedBegin;
|
|
++MappedBegin;
|
|
bool AllEqual = std::all_of(
|
|
MappedBegin, MappedEnd,
|
|
[&AllSameValue](const MemoryAccess *V) { return V == AllSameValue; });
|
|
|
|
if (AllEqual)
|
|
LLVM_DEBUG(dbgs() << "Memory Phi value numbered to " << *AllSameValue
|
|
<< "\n");
|
|
else
|
|
LLVM_DEBUG(dbgs() << "Memory Phi value numbered to itself\n");
|
|
// If it's equal to something, it's in that class. Otherwise, it has to be in
|
|
// a class where it is the leader (other things may be equivalent to it, but
|
|
// it needs to start off in its own class, which means it must have been the
|
|
// leader, and it can't have stopped being the leader because it was never
|
|
// removed).
|
|
CongruenceClass *CC =
|
|
AllEqual ? getMemoryClass(AllSameValue) : ensureLeaderOfMemoryClass(MP);
|
|
auto OldState = MemoryPhiState.lookup(MP);
|
|
assert(OldState != MPS_Invalid && "Invalid memory phi state");
|
|
auto NewState = AllEqual ? MPS_Equivalent : MPS_Unique;
|
|
MemoryPhiState[MP] = NewState;
|
|
if (setMemoryClass(MP, CC) || OldState != NewState)
|
|
markMemoryUsersTouched(MP);
|
|
}
|
|
|
|
// Value number a single instruction, symbolically evaluating, performing
|
|
// congruence finding, and updating mappings.
|
|
void NewGVN::valueNumberInstruction(Instruction *I) {
|
|
LLVM_DEBUG(dbgs() << "Processing instruction " << *I << "\n");
|
|
if (!I->isTerminator()) {
|
|
const Expression *Symbolized = nullptr;
|
|
SmallPtrSet<Value *, 2> Visited;
|
|
if (DebugCounter::shouldExecute(VNCounter)) {
|
|
Symbolized = performSymbolicEvaluation(I, Visited);
|
|
// Make a phi of ops if necessary
|
|
if (Symbolized && !isa<ConstantExpression>(Symbolized) &&
|
|
!isa<VariableExpression>(Symbolized) && PHINodeUses.count(I)) {
|
|
auto *PHIE = makePossiblePHIOfOps(I, Visited);
|
|
// If we created a phi of ops, use it.
|
|
// If we couldn't create one, make sure we don't leave one lying around
|
|
if (PHIE) {
|
|
Symbolized = PHIE;
|
|
} else if (auto *Op = RealToTemp.lookup(I)) {
|
|
removePhiOfOps(I, Op);
|
|
}
|
|
}
|
|
} else {
|
|
// Mark the instruction as unused so we don't value number it again.
|
|
InstrDFS[I] = 0;
|
|
}
|
|
// If we couldn't come up with a symbolic expression, use the unknown
|
|
// expression
|
|
if (Symbolized == nullptr)
|
|
Symbolized = createUnknownExpression(I);
|
|
performCongruenceFinding(I, Symbolized);
|
|
} else {
|
|
// Handle terminators that return values. All of them produce values we
|
|
// don't currently understand. We don't place non-value producing
|
|
// terminators in a class.
|
|
if (!I->getType()->isVoidTy()) {
|
|
auto *Symbolized = createUnknownExpression(I);
|
|
performCongruenceFinding(I, Symbolized);
|
|
}
|
|
processOutgoingEdges(I, I->getParent());
|
|
}
|
|
}
|
|
|
|
// Check if there is a path, using single or equal argument phi nodes, from
|
|
// First to Second.
|
|
bool NewGVN::singleReachablePHIPath(
|
|
SmallPtrSet<const MemoryAccess *, 8> &Visited, const MemoryAccess *First,
|
|
const MemoryAccess *Second) const {
|
|
if (First == Second)
|
|
return true;
|
|
if (MSSA->isLiveOnEntryDef(First))
|
|
return false;
|
|
|
|
// This is not perfect, but as we're just verifying here, we can live with
|
|
// the loss of precision. The real solution would be that of doing strongly
|
|
// connected component finding in this routine, and it's probably not worth
|
|
// the complexity for the time being. So, we just keep a set of visited
|
|
// MemoryAccess and return true when we hit a cycle.
|
|
if (Visited.count(First))
|
|
return true;
|
|
Visited.insert(First);
|
|
|
|
const auto *EndDef = First;
|
|
for (auto *ChainDef : optimized_def_chain(First)) {
|
|
if (ChainDef == Second)
|
|
return true;
|
|
if (MSSA->isLiveOnEntryDef(ChainDef))
|
|
return false;
|
|
EndDef = ChainDef;
|
|
}
|
|
auto *MP = cast<MemoryPhi>(EndDef);
|
|
auto ReachableOperandPred = [&](const Use &U) {
|
|
return ReachableEdges.count({MP->getIncomingBlock(U), MP->getBlock()});
|
|
};
|
|
auto FilteredPhiArgs =
|
|
make_filter_range(MP->operands(), ReachableOperandPred);
|
|
SmallVector<const Value *, 32> OperandList;
|
|
llvm::copy(FilteredPhiArgs, std::back_inserter(OperandList));
|
|
bool Okay = is_splat(OperandList);
|
|
if (Okay)
|
|
return singleReachablePHIPath(Visited, cast<MemoryAccess>(OperandList[0]),
|
|
Second);
|
|
return false;
|
|
}
|
|
|
|
// Verify the that the memory equivalence table makes sense relative to the
|
|
// congruence classes. Note that this checking is not perfect, and is currently
|
|
// subject to very rare false negatives. It is only useful for
|
|
// testing/debugging.
|
|
void NewGVN::verifyMemoryCongruency() const {
|
|
#ifndef NDEBUG
|
|
// Verify that the memory table equivalence and memory member set match
|
|
for (const auto *CC : CongruenceClasses) {
|
|
if (CC == TOPClass || CC->isDead())
|
|
continue;
|
|
if (CC->getStoreCount() != 0) {
|
|
assert((CC->getStoredValue() || !isa<StoreInst>(CC->getLeader())) &&
|
|
"Any class with a store as a leader should have a "
|
|
"representative stored value");
|
|
assert(CC->getMemoryLeader() &&
|
|
"Any congruence class with a store should have a "
|
|
"representative access");
|
|
}
|
|
|
|
if (CC->getMemoryLeader())
|
|
assert(MemoryAccessToClass.lookup(CC->getMemoryLeader()) == CC &&
|
|
"Representative MemoryAccess does not appear to be reverse "
|
|
"mapped properly");
|
|
for (auto M : CC->memory())
|
|
assert(MemoryAccessToClass.lookup(M) == CC &&
|
|
"Memory member does not appear to be reverse mapped properly");
|
|
}
|
|
|
|
// Anything equivalent in the MemoryAccess table should be in the same
|
|
// congruence class.
|
|
|
|
// Filter out the unreachable and trivially dead entries, because they may
|
|
// never have been updated if the instructions were not processed.
|
|
auto ReachableAccessPred =
|
|
[&](const std::pair<const MemoryAccess *, CongruenceClass *> Pair) {
|
|
bool Result = ReachableBlocks.count(Pair.first->getBlock());
|
|
if (!Result || MSSA->isLiveOnEntryDef(Pair.first) ||
|
|
MemoryToDFSNum(Pair.first) == 0)
|
|
return false;
|
|
if (auto *MemDef = dyn_cast<MemoryDef>(Pair.first))
|
|
return !isInstructionTriviallyDead(MemDef->getMemoryInst());
|
|
|
|
// We could have phi nodes which operands are all trivially dead,
|
|
// so we don't process them.
|
|
if (auto *MemPHI = dyn_cast<MemoryPhi>(Pair.first)) {
|
|
for (auto &U : MemPHI->incoming_values()) {
|
|
if (auto *I = dyn_cast<Instruction>(&*U)) {
|
|
if (!isInstructionTriviallyDead(I))
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
};
|
|
|
|
auto Filtered = make_filter_range(MemoryAccessToClass, ReachableAccessPred);
|
|
for (auto KV : Filtered) {
|
|
if (auto *FirstMUD = dyn_cast<MemoryUseOrDef>(KV.first)) {
|
|
auto *SecondMUD = dyn_cast<MemoryUseOrDef>(KV.second->getMemoryLeader());
|
|
if (FirstMUD && SecondMUD) {
|
|
SmallPtrSet<const MemoryAccess *, 8> VisitedMAS;
|
|
assert((singleReachablePHIPath(VisitedMAS, FirstMUD, SecondMUD) ||
|
|
ValueToClass.lookup(FirstMUD->getMemoryInst()) ==
|
|
ValueToClass.lookup(SecondMUD->getMemoryInst())) &&
|
|
"The instructions for these memory operations should have "
|
|
"been in the same congruence class or reachable through"
|
|
"a single argument phi");
|
|
}
|
|
} else if (auto *FirstMP = dyn_cast<MemoryPhi>(KV.first)) {
|
|
// We can only sanely verify that MemoryDefs in the operand list all have
|
|
// the same class.
|
|
auto ReachableOperandPred = [&](const Use &U) {
|
|
return ReachableEdges.count(
|
|
{FirstMP->getIncomingBlock(U), FirstMP->getBlock()}) &&
|
|
isa<MemoryDef>(U);
|
|
|
|
};
|
|
// All arguments should in the same class, ignoring unreachable arguments
|
|
auto FilteredPhiArgs =
|
|
make_filter_range(FirstMP->operands(), ReachableOperandPred);
|
|
SmallVector<const CongruenceClass *, 16> PhiOpClasses;
|
|
std::transform(FilteredPhiArgs.begin(), FilteredPhiArgs.end(),
|
|
std::back_inserter(PhiOpClasses), [&](const Use &U) {
|
|
const MemoryDef *MD = cast<MemoryDef>(U);
|
|
return ValueToClass.lookup(MD->getMemoryInst());
|
|
});
|
|
assert(is_splat(PhiOpClasses) &&
|
|
"All MemoryPhi arguments should be in the same class");
|
|
}
|
|
}
|
|
#endif
|
|
}
|
|
|
|
// Verify that the sparse propagation we did actually found the maximal fixpoint
|
|
// We do this by storing the value to class mapping, touching all instructions,
|
|
// and redoing the iteration to see if anything changed.
|
|
void NewGVN::verifyIterationSettled(Function &F) {
|
|
#ifndef NDEBUG
|
|
LLVM_DEBUG(dbgs() << "Beginning iteration verification\n");
|
|
if (DebugCounter::isCounterSet(VNCounter))
|
|
DebugCounter::setCounterValue(VNCounter, StartingVNCounter);
|
|
|
|
// Note that we have to store the actual classes, as we may change existing
|
|
// classes during iteration. This is because our memory iteration propagation
|
|
// is not perfect, and so may waste a little work. But it should generate
|
|
// exactly the same congruence classes we have now, with different IDs.
|
|
std::map<const Value *, CongruenceClass> BeforeIteration;
|
|
|
|
for (auto &KV : ValueToClass) {
|
|
if (auto *I = dyn_cast<Instruction>(KV.first))
|
|
// Skip unused/dead instructions.
|
|
if (InstrToDFSNum(I) == 0)
|
|
continue;
|
|
BeforeIteration.insert({KV.first, *KV.second});
|
|
}
|
|
|
|
TouchedInstructions.set();
|
|
TouchedInstructions.reset(0);
|
|
iterateTouchedInstructions();
|
|
DenseSet<std::pair<const CongruenceClass *, const CongruenceClass *>>
|
|
EqualClasses;
|
|
for (const auto &KV : ValueToClass) {
|
|
if (auto *I = dyn_cast<Instruction>(KV.first))
|
|
// Skip unused/dead instructions.
|
|
if (InstrToDFSNum(I) == 0)
|
|
continue;
|
|
// We could sink these uses, but i think this adds a bit of clarity here as
|
|
// to what we are comparing.
|
|
auto *BeforeCC = &BeforeIteration.find(KV.first)->second;
|
|
auto *AfterCC = KV.second;
|
|
// Note that the classes can't change at this point, so we memoize the set
|
|
// that are equal.
|
|
if (!EqualClasses.count({BeforeCC, AfterCC})) {
|
|
assert(BeforeCC->isEquivalentTo(AfterCC) &&
|
|
"Value number changed after main loop completed!");
|
|
EqualClasses.insert({BeforeCC, AfterCC});
|
|
}
|
|
}
|
|
#endif
|
|
}
|
|
|
|
// Verify that for each store expression in the expression to class mapping,
|
|
// only the latest appears, and multiple ones do not appear.
|
|
// Because loads do not use the stored value when doing equality with stores,
|
|
// if we don't erase the old store expressions from the table, a load can find
|
|
// a no-longer valid StoreExpression.
|
|
void NewGVN::verifyStoreExpressions() const {
|
|
#ifndef NDEBUG
|
|
// This is the only use of this, and it's not worth defining a complicated
|
|
// densemapinfo hash/equality function for it.
|
|
std::set<
|
|
std::pair<const Value *,
|
|
std::tuple<const Value *, const CongruenceClass *, Value *>>>
|
|
StoreExpressionSet;
|
|
for (const auto &KV : ExpressionToClass) {
|
|
if (auto *SE = dyn_cast<StoreExpression>(KV.first)) {
|
|
// Make sure a version that will conflict with loads is not already there
|
|
auto Res = StoreExpressionSet.insert(
|
|
{SE->getOperand(0), std::make_tuple(SE->getMemoryLeader(), KV.second,
|
|
SE->getStoredValue())});
|
|
bool Okay = Res.second;
|
|
// It's okay to have the same expression already in there if it is
|
|
// identical in nature.
|
|
// This can happen when the leader of the stored value changes over time.
|
|
if (!Okay)
|
|
Okay = (std::get<1>(Res.first->second) == KV.second) &&
|
|
(lookupOperandLeader(std::get<2>(Res.first->second)) ==
|
|
lookupOperandLeader(SE->getStoredValue()));
|
|
assert(Okay && "Stored expression conflict exists in expression table");
|
|
auto *ValueExpr = ValueToExpression.lookup(SE->getStoreInst());
|
|
assert(ValueExpr && ValueExpr->equals(*SE) &&
|
|
"StoreExpression in ExpressionToClass is not latest "
|
|
"StoreExpression for value");
|
|
}
|
|
}
|
|
#endif
|
|
}
|
|
|
|
// This is the main value numbering loop, it iterates over the initial touched
|
|
// instruction set, propagating value numbers, marking things touched, etc,
|
|
// until the set of touched instructions is completely empty.
|
|
void NewGVN::iterateTouchedInstructions() {
|
|
unsigned int Iterations = 0;
|
|
// Figure out where touchedinstructions starts
|
|
int FirstInstr = TouchedInstructions.find_first();
|
|
// Nothing set, nothing to iterate, just return.
|
|
if (FirstInstr == -1)
|
|
return;
|
|
const BasicBlock *LastBlock = getBlockForValue(InstrFromDFSNum(FirstInstr));
|
|
while (TouchedInstructions.any()) {
|
|
++Iterations;
|
|
// Walk through all the instructions in all the blocks in RPO.
|
|
// TODO: As we hit a new block, we should push and pop equalities into a
|
|
// table lookupOperandLeader can use, to catch things PredicateInfo
|
|
// might miss, like edge-only equivalences.
|
|
for (unsigned InstrNum : TouchedInstructions.set_bits()) {
|
|
|
|
// This instruction was found to be dead. We don't bother looking
|
|
// at it again.
|
|
if (InstrNum == 0) {
|
|
TouchedInstructions.reset(InstrNum);
|
|
continue;
|
|
}
|
|
|
|
Value *V = InstrFromDFSNum(InstrNum);
|
|
const BasicBlock *CurrBlock = getBlockForValue(V);
|
|
|
|
// If we hit a new block, do reachability processing.
|
|
if (CurrBlock != LastBlock) {
|
|
LastBlock = CurrBlock;
|
|
bool BlockReachable = ReachableBlocks.count(CurrBlock);
|
|
const auto &CurrInstRange = BlockInstRange.lookup(CurrBlock);
|
|
|
|
// If it's not reachable, erase any touched instructions and move on.
|
|
if (!BlockReachable) {
|
|
TouchedInstructions.reset(CurrInstRange.first, CurrInstRange.second);
|
|
LLVM_DEBUG(dbgs() << "Skipping instructions in block "
|
|
<< getBlockName(CurrBlock)
|
|
<< " because it is unreachable\n");
|
|
continue;
|
|
}
|
|
updateProcessedCount(CurrBlock);
|
|
}
|
|
// Reset after processing (because we may mark ourselves as touched when
|
|
// we propagate equalities).
|
|
TouchedInstructions.reset(InstrNum);
|
|
|
|
if (auto *MP = dyn_cast<MemoryPhi>(V)) {
|
|
LLVM_DEBUG(dbgs() << "Processing MemoryPhi " << *MP << "\n");
|
|
valueNumberMemoryPhi(MP);
|
|
} else if (auto *I = dyn_cast<Instruction>(V)) {
|
|
valueNumberInstruction(I);
|
|
} else {
|
|
llvm_unreachable("Should have been a MemoryPhi or Instruction");
|
|
}
|
|
updateProcessedCount(V);
|
|
}
|
|
}
|
|
NumGVNMaxIterations = std::max(NumGVNMaxIterations.getValue(), Iterations);
|
|
}
|
|
|
|
// This is the main transformation entry point.
|
|
bool NewGVN::runGVN() {
|
|
if (DebugCounter::isCounterSet(VNCounter))
|
|
StartingVNCounter = DebugCounter::getCounterValue(VNCounter);
|
|
bool Changed = false;
|
|
NumFuncArgs = F.arg_size();
|
|
MSSAWalker = MSSA->getWalker();
|
|
SingletonDeadExpression = new (ExpressionAllocator) DeadExpression();
|
|
|
|
// Count number of instructions for sizing of hash tables, and come
|
|
// up with a global dfs numbering for instructions.
|
|
unsigned ICount = 1;
|
|
// Add an empty instruction to account for the fact that we start at 1
|
|
DFSToInstr.emplace_back(nullptr);
|
|
// Note: We want ideal RPO traversal of the blocks, which is not quite the
|
|
// same as dominator tree order, particularly with regard whether backedges
|
|
// get visited first or second, given a block with multiple successors.
|
|
// If we visit in the wrong order, we will end up performing N times as many
|
|
// iterations.
|
|
// The dominator tree does guarantee that, for a given dom tree node, it's
|
|
// parent must occur before it in the RPO ordering. Thus, we only need to sort
|
|
// the siblings.
|
|
ReversePostOrderTraversal<Function *> RPOT(&F);
|
|
unsigned Counter = 0;
|
|
for (auto &B : RPOT) {
|
|
auto *Node = DT->getNode(B);
|
|
assert(Node && "RPO and Dominator tree should have same reachability");
|
|
RPOOrdering[Node] = ++Counter;
|
|
}
|
|
// Sort dominator tree children arrays into RPO.
|
|
for (auto &B : RPOT) {
|
|
auto *Node = DT->getNode(B);
|
|
if (Node->getNumChildren() > 1)
|
|
llvm::sort(Node->begin(), Node->end(),
|
|
[&](const DomTreeNode *A, const DomTreeNode *B) {
|
|
return RPOOrdering[A] < RPOOrdering[B];
|
|
});
|
|
}
|
|
|
|
// Now a standard depth first ordering of the domtree is equivalent to RPO.
|
|
for (auto DTN : depth_first(DT->getRootNode())) {
|
|
BasicBlock *B = DTN->getBlock();
|
|
const auto &BlockRange = assignDFSNumbers(B, ICount);
|
|
BlockInstRange.insert({B, BlockRange});
|
|
ICount += BlockRange.second - BlockRange.first;
|
|
}
|
|
initializeCongruenceClasses(F);
|
|
|
|
TouchedInstructions.resize(ICount);
|
|
// Ensure we don't end up resizing the expressionToClass map, as
|
|
// that can be quite expensive. At most, we have one expression per
|
|
// instruction.
|
|
ExpressionToClass.reserve(ICount);
|
|
|
|
// Initialize the touched instructions to include the entry block.
|
|
const auto &InstRange = BlockInstRange.lookup(&F.getEntryBlock());
|
|
TouchedInstructions.set(InstRange.first, InstRange.second);
|
|
LLVM_DEBUG(dbgs() << "Block " << getBlockName(&F.getEntryBlock())
|
|
<< " marked reachable\n");
|
|
ReachableBlocks.insert(&F.getEntryBlock());
|
|
|
|
iterateTouchedInstructions();
|
|
verifyMemoryCongruency();
|
|
verifyIterationSettled(F);
|
|
verifyStoreExpressions();
|
|
|
|
Changed |= eliminateInstructions(F);
|
|
|
|
// Delete all instructions marked for deletion.
|
|
for (Instruction *ToErase : InstructionsToErase) {
|
|
if (!ToErase->use_empty())
|
|
ToErase->replaceAllUsesWith(UndefValue::get(ToErase->getType()));
|
|
|
|
assert(ToErase->getParent() &&
|
|
"BB containing ToErase deleted unexpectedly!");
|
|
ToErase->eraseFromParent();
|
|
}
|
|
Changed |= !InstructionsToErase.empty();
|
|
|
|
// Delete all unreachable blocks.
|
|
auto UnreachableBlockPred = [&](const BasicBlock &BB) {
|
|
return !ReachableBlocks.count(&BB);
|
|
};
|
|
|
|
for (auto &BB : make_filter_range(F, UnreachableBlockPred)) {
|
|
LLVM_DEBUG(dbgs() << "We believe block " << getBlockName(&BB)
|
|
<< " is unreachable\n");
|
|
deleteInstructionsInBlock(&BB);
|
|
Changed = true;
|
|
}
|
|
|
|
cleanupTables();
|
|
return Changed;
|
|
}
|
|
|
|
struct NewGVN::ValueDFS {
|
|
int DFSIn = 0;
|
|
int DFSOut = 0;
|
|
int LocalNum = 0;
|
|
|
|
// Only one of Def and U will be set.
|
|
// The bool in the Def tells us whether the Def is the stored value of a
|
|
// store.
|
|
PointerIntPair<Value *, 1, bool> Def;
|
|
Use *U = nullptr;
|
|
|
|
bool operator<(const ValueDFS &Other) const {
|
|
// It's not enough that any given field be less than - we have sets
|
|
// of fields that need to be evaluated together to give a proper ordering.
|
|
// For example, if you have;
|
|
// DFS (1, 3)
|
|
// Val 0
|
|
// DFS (1, 2)
|
|
// Val 50
|
|
// We want the second to be less than the first, but if we just go field
|
|
// by field, we will get to Val 0 < Val 50 and say the first is less than
|
|
// the second. We only want it to be less than if the DFS orders are equal.
|
|
//
|
|
// Each LLVM instruction only produces one value, and thus the lowest-level
|
|
// differentiator that really matters for the stack (and what we use as as a
|
|
// replacement) is the local dfs number.
|
|
// Everything else in the structure is instruction level, and only affects
|
|
// the order in which we will replace operands of a given instruction.
|
|
//
|
|
// For a given instruction (IE things with equal dfsin, dfsout, localnum),
|
|
// the order of replacement of uses does not matter.
|
|
// IE given,
|
|
// a = 5
|
|
// b = a + a
|
|
// When you hit b, you will have two valuedfs with the same dfsin, out, and
|
|
// localnum.
|
|
// The .val will be the same as well.
|
|
// The .u's will be different.
|
|
// You will replace both, and it does not matter what order you replace them
|
|
// in (IE whether you replace operand 2, then operand 1, or operand 1, then
|
|
// operand 2).
|
|
// Similarly for the case of same dfsin, dfsout, localnum, but different
|
|
// .val's
|
|
// a = 5
|
|
// b = 6
|
|
// c = a + b
|
|
// in c, we will a valuedfs for a, and one for b,with everything the same
|
|
// but .val and .u.
|
|
// It does not matter what order we replace these operands in.
|
|
// You will always end up with the same IR, and this is guaranteed.
|
|
return std::tie(DFSIn, DFSOut, LocalNum, Def, U) <
|
|
std::tie(Other.DFSIn, Other.DFSOut, Other.LocalNum, Other.Def,
|
|
Other.U);
|
|
}
|
|
};
|
|
|
|
// This function converts the set of members for a congruence class from values,
|
|
// to sets of defs and uses with associated DFS info. The total number of
|
|
// reachable uses for each value is stored in UseCount, and instructions that
|
|
// seem
|
|
// dead (have no non-dead uses) are stored in ProbablyDead.
|
|
void NewGVN::convertClassToDFSOrdered(
|
|
const CongruenceClass &Dense, SmallVectorImpl<ValueDFS> &DFSOrderedSet,
|
|
DenseMap<const Value *, unsigned int> &UseCounts,
|
|
SmallPtrSetImpl<Instruction *> &ProbablyDead) const {
|
|
for (auto D : Dense) {
|
|
// First add the value.
|
|
BasicBlock *BB = getBlockForValue(D);
|
|
// Constants are handled prior to ever calling this function, so
|
|
// we should only be left with instructions as members.
|
|
assert(BB && "Should have figured out a basic block for value");
|
|
ValueDFS VDDef;
|
|
DomTreeNode *DomNode = DT->getNode(BB);
|
|
VDDef.DFSIn = DomNode->getDFSNumIn();
|
|
VDDef.DFSOut = DomNode->getDFSNumOut();
|
|
// If it's a store, use the leader of the value operand, if it's always
|
|
// available, or the value operand. TODO: We could do dominance checks to
|
|
// find a dominating leader, but not worth it ATM.
|
|
if (auto *SI = dyn_cast<StoreInst>(D)) {
|
|
auto Leader = lookupOperandLeader(SI->getValueOperand());
|
|
if (alwaysAvailable(Leader)) {
|
|
VDDef.Def.setPointer(Leader);
|
|
} else {
|
|
VDDef.Def.setPointer(SI->getValueOperand());
|
|
VDDef.Def.setInt(true);
|
|
}
|
|
} else {
|
|
VDDef.Def.setPointer(D);
|
|
}
|
|
assert(isa<Instruction>(D) &&
|
|
"The dense set member should always be an instruction");
|
|
Instruction *Def = cast<Instruction>(D);
|
|
VDDef.LocalNum = InstrToDFSNum(D);
|
|
DFSOrderedSet.push_back(VDDef);
|
|
// If there is a phi node equivalent, add it
|
|
if (auto *PN = RealToTemp.lookup(Def)) {
|
|
auto *PHIE =
|
|
dyn_cast_or_null<PHIExpression>(ValueToExpression.lookup(Def));
|
|
if (PHIE) {
|
|
VDDef.Def.setInt(false);
|
|
VDDef.Def.setPointer(PN);
|
|
VDDef.LocalNum = 0;
|
|
DFSOrderedSet.push_back(VDDef);
|
|
}
|
|
}
|
|
|
|
unsigned int UseCount = 0;
|
|
// Now add the uses.
|
|
for (auto &U : Def->uses()) {
|
|
if (auto *I = dyn_cast<Instruction>(U.getUser())) {
|
|
// Don't try to replace into dead uses
|
|
if (InstructionsToErase.count(I))
|
|
continue;
|
|
ValueDFS VDUse;
|
|
// Put the phi node uses in the incoming block.
|
|
BasicBlock *IBlock;
|
|
if (auto *P = dyn_cast<PHINode>(I)) {
|
|
IBlock = P->getIncomingBlock(U);
|
|
// Make phi node users appear last in the incoming block
|
|
// they are from.
|
|
VDUse.LocalNum = InstrDFS.size() + 1;
|
|
} else {
|
|
IBlock = getBlockForValue(I);
|
|
VDUse.LocalNum = InstrToDFSNum(I);
|
|
}
|
|
|
|
// Skip uses in unreachable blocks, as we're going
|
|
// to delete them.
|
|
if (ReachableBlocks.count(IBlock) == 0)
|
|
continue;
|
|
|
|
DomTreeNode *DomNode = DT->getNode(IBlock);
|
|
VDUse.DFSIn = DomNode->getDFSNumIn();
|
|
VDUse.DFSOut = DomNode->getDFSNumOut();
|
|
VDUse.U = &U;
|
|
++UseCount;
|
|
DFSOrderedSet.emplace_back(VDUse);
|
|
}
|
|
}
|
|
|
|
// If there are no uses, it's probably dead (but it may have side-effects,
|
|
// so not definitely dead. Otherwise, store the number of uses so we can
|
|
// track if it becomes dead later).
|
|
if (UseCount == 0)
|
|
ProbablyDead.insert(Def);
|
|
else
|
|
UseCounts[Def] = UseCount;
|
|
}
|
|
}
|
|
|
|
// This function converts the set of members for a congruence class from values,
|
|
// to the set of defs for loads and stores, with associated DFS info.
|
|
void NewGVN::convertClassToLoadsAndStores(
|
|
const CongruenceClass &Dense,
|
|
SmallVectorImpl<ValueDFS> &LoadsAndStores) const {
|
|
for (auto D : Dense) {
|
|
if (!isa<LoadInst>(D) && !isa<StoreInst>(D))
|
|
continue;
|
|
|
|
BasicBlock *BB = getBlockForValue(D);
|
|
ValueDFS VD;
|
|
DomTreeNode *DomNode = DT->getNode(BB);
|
|
VD.DFSIn = DomNode->getDFSNumIn();
|
|
VD.DFSOut = DomNode->getDFSNumOut();
|
|
VD.Def.setPointer(D);
|
|
|
|
// If it's an instruction, use the real local dfs number.
|
|
if (auto *I = dyn_cast<Instruction>(D))
|
|
VD.LocalNum = InstrToDFSNum(I);
|
|
else
|
|
llvm_unreachable("Should have been an instruction");
|
|
|
|
LoadsAndStores.emplace_back(VD);
|
|
}
|
|
}
|
|
|
|
static void patchAndReplaceAllUsesWith(Instruction *I, Value *Repl) {
|
|
patchReplacementInstruction(I, Repl);
|
|
I->replaceAllUsesWith(Repl);
|
|
}
|
|
|
|
void NewGVN::deleteInstructionsInBlock(BasicBlock *BB) {
|
|
LLVM_DEBUG(dbgs() << " BasicBlock Dead:" << *BB);
|
|
++NumGVNBlocksDeleted;
|
|
|
|
// Delete the instructions backwards, as it has a reduced likelihood of having
|
|
// to update as many def-use and use-def chains. Start after the terminator.
|
|
auto StartPoint = BB->rbegin();
|
|
++StartPoint;
|
|
// Note that we explicitly recalculate BB->rend() on each iteration,
|
|
// as it may change when we remove the first instruction.
|
|
for (BasicBlock::reverse_iterator I(StartPoint); I != BB->rend();) {
|
|
Instruction &Inst = *I++;
|
|
if (!Inst.use_empty())
|
|
Inst.replaceAllUsesWith(UndefValue::get(Inst.getType()));
|
|
if (isa<LandingPadInst>(Inst))
|
|
continue;
|
|
salvageKnowledge(&Inst, AC);
|
|
|
|
Inst.eraseFromParent();
|
|
++NumGVNInstrDeleted;
|
|
}
|
|
// Now insert something that simplifycfg will turn into an unreachable.
|
|
Type *Int8Ty = Type::getInt8Ty(BB->getContext());
|
|
new StoreInst(UndefValue::get(Int8Ty),
|
|
Constant::getNullValue(Int8Ty->getPointerTo()),
|
|
BB->getTerminator());
|
|
}
|
|
|
|
void NewGVN::markInstructionForDeletion(Instruction *I) {
|
|
LLVM_DEBUG(dbgs() << "Marking " << *I << " for deletion\n");
|
|
InstructionsToErase.insert(I);
|
|
}
|
|
|
|
void NewGVN::replaceInstruction(Instruction *I, Value *V) {
|
|
LLVM_DEBUG(dbgs() << "Replacing " << *I << " with " << *V << "\n");
|
|
patchAndReplaceAllUsesWith(I, V);
|
|
// We save the actual erasing to avoid invalidating memory
|
|
// dependencies until we are done with everything.
|
|
markInstructionForDeletion(I);
|
|
}
|
|
|
|
namespace {
|
|
|
|
// This is a stack that contains both the value and dfs info of where
|
|
// that value is valid.
|
|
class ValueDFSStack {
|
|
public:
|
|
Value *back() const { return ValueStack.back(); }
|
|
std::pair<int, int> dfs_back() const { return DFSStack.back(); }
|
|
|
|
void push_back(Value *V, int DFSIn, int DFSOut) {
|
|
ValueStack.emplace_back(V);
|
|
DFSStack.emplace_back(DFSIn, DFSOut);
|
|
}
|
|
|
|
bool empty() const { return DFSStack.empty(); }
|
|
|
|
bool isInScope(int DFSIn, int DFSOut) const {
|
|
if (empty())
|
|
return false;
|
|
return DFSIn >= DFSStack.back().first && DFSOut <= DFSStack.back().second;
|
|
}
|
|
|
|
void popUntilDFSScope(int DFSIn, int DFSOut) {
|
|
|
|
// These two should always be in sync at this point.
|
|
assert(ValueStack.size() == DFSStack.size() &&
|
|
"Mismatch between ValueStack and DFSStack");
|
|
while (
|
|
!DFSStack.empty() &&
|
|
!(DFSIn >= DFSStack.back().first && DFSOut <= DFSStack.back().second)) {
|
|
DFSStack.pop_back();
|
|
ValueStack.pop_back();
|
|
}
|
|
}
|
|
|
|
private:
|
|
SmallVector<Value *, 8> ValueStack;
|
|
SmallVector<std::pair<int, int>, 8> DFSStack;
|
|
};
|
|
|
|
} // end anonymous namespace
|
|
|
|
// Given an expression, get the congruence class for it.
|
|
CongruenceClass *NewGVN::getClassForExpression(const Expression *E) const {
|
|
if (auto *VE = dyn_cast<VariableExpression>(E))
|
|
return ValueToClass.lookup(VE->getVariableValue());
|
|
else if (isa<DeadExpression>(E))
|
|
return TOPClass;
|
|
return ExpressionToClass.lookup(E);
|
|
}
|
|
|
|
// Given a value and a basic block we are trying to see if it is available in,
|
|
// see if the value has a leader available in that block.
|
|
Value *NewGVN::findPHIOfOpsLeader(const Expression *E,
|
|
const Instruction *OrigInst,
|
|
const BasicBlock *BB) const {
|
|
// It would already be constant if we could make it constant
|
|
if (auto *CE = dyn_cast<ConstantExpression>(E))
|
|
return CE->getConstantValue();
|
|
if (auto *VE = dyn_cast<VariableExpression>(E)) {
|
|
auto *V = VE->getVariableValue();
|
|
if (alwaysAvailable(V) || DT->dominates(getBlockForValue(V), BB))
|
|
return VE->getVariableValue();
|
|
}
|
|
|
|
auto *CC = getClassForExpression(E);
|
|
if (!CC)
|
|
return nullptr;
|
|
if (alwaysAvailable(CC->getLeader()))
|
|
return CC->getLeader();
|
|
|
|
for (auto Member : *CC) {
|
|
auto *MemberInst = dyn_cast<Instruction>(Member);
|
|
if (MemberInst == OrigInst)
|
|
continue;
|
|
// Anything that isn't an instruction is always available.
|
|
if (!MemberInst)
|
|
return Member;
|
|
if (DT->dominates(getBlockForValue(MemberInst), BB))
|
|
return Member;
|
|
}
|
|
return nullptr;
|
|
}
|
|
|
|
bool NewGVN::eliminateInstructions(Function &F) {
|
|
// This is a non-standard eliminator. The normal way to eliminate is
|
|
// to walk the dominator tree in order, keeping track of available
|
|
// values, and eliminating them. However, this is mildly
|
|
// pointless. It requires doing lookups on every instruction,
|
|
// regardless of whether we will ever eliminate it. For
|
|
// instructions part of most singleton congruence classes, we know we
|
|
// will never eliminate them.
|
|
|
|
// Instead, this eliminator looks at the congruence classes directly, sorts
|
|
// them into a DFS ordering of the dominator tree, and then we just
|
|
// perform elimination straight on the sets by walking the congruence
|
|
// class member uses in order, and eliminate the ones dominated by the
|
|
// last member. This is worst case O(E log E) where E = number of
|
|
// instructions in a single congruence class. In theory, this is all
|
|
// instructions. In practice, it is much faster, as most instructions are
|
|
// either in singleton congruence classes or can't possibly be eliminated
|
|
// anyway (if there are no overlapping DFS ranges in class).
|
|
// When we find something not dominated, it becomes the new leader
|
|
// for elimination purposes.
|
|
// TODO: If we wanted to be faster, We could remove any members with no
|
|
// overlapping ranges while sorting, as we will never eliminate anything
|
|
// with those members, as they don't dominate anything else in our set.
|
|
|
|
bool AnythingReplaced = false;
|
|
|
|
// Since we are going to walk the domtree anyway, and we can't guarantee the
|
|
// DFS numbers are updated, we compute some ourselves.
|
|
DT->updateDFSNumbers();
|
|
|
|
// Go through all of our phi nodes, and kill the arguments associated with
|
|
// unreachable edges.
|
|
auto ReplaceUnreachablePHIArgs = [&](PHINode *PHI, BasicBlock *BB) {
|
|
for (auto &Operand : PHI->incoming_values())
|
|
if (!ReachableEdges.count({PHI->getIncomingBlock(Operand), BB})) {
|
|
LLVM_DEBUG(dbgs() << "Replacing incoming value of " << PHI
|
|
<< " for block "
|
|
<< getBlockName(PHI->getIncomingBlock(Operand))
|
|
<< " with undef due to it being unreachable\n");
|
|
Operand.set(UndefValue::get(PHI->getType()));
|
|
}
|
|
};
|
|
// Replace unreachable phi arguments.
|
|
// At this point, RevisitOnReachabilityChange only contains:
|
|
//
|
|
// 1. PHIs
|
|
// 2. Temporaries that will convert to PHIs
|
|
// 3. Operations that are affected by an unreachable edge but do not fit into
|
|
// 1 or 2 (rare).
|
|
// So it is a slight overshoot of what we want. We could make it exact by
|
|
// using two SparseBitVectors per block.
|
|
DenseMap<const BasicBlock *, unsigned> ReachablePredCount;
|
|
for (auto &KV : ReachableEdges)
|
|
ReachablePredCount[KV.getEnd()]++;
|
|
for (auto &BBPair : RevisitOnReachabilityChange) {
|
|
for (auto InstNum : BBPair.second) {
|
|
auto *Inst = InstrFromDFSNum(InstNum);
|
|
auto *PHI = dyn_cast<PHINode>(Inst);
|
|
PHI = PHI ? PHI : dyn_cast_or_null<PHINode>(RealToTemp.lookup(Inst));
|
|
if (!PHI)
|
|
continue;
|
|
auto *BB = BBPair.first;
|
|
if (ReachablePredCount.lookup(BB) != PHI->getNumIncomingValues())
|
|
ReplaceUnreachablePHIArgs(PHI, BB);
|
|
}
|
|
}
|
|
|
|
// Map to store the use counts
|
|
DenseMap<const Value *, unsigned int> UseCounts;
|
|
for (auto *CC : reverse(CongruenceClasses)) {
|
|
LLVM_DEBUG(dbgs() << "Eliminating in congruence class " << CC->getID()
|
|
<< "\n");
|
|
// Track the equivalent store info so we can decide whether to try
|
|
// dead store elimination.
|
|
SmallVector<ValueDFS, 8> PossibleDeadStores;
|
|
SmallPtrSet<Instruction *, 8> ProbablyDead;
|
|
if (CC->isDead() || CC->empty())
|
|
continue;
|
|
// Everything still in the TOP class is unreachable or dead.
|
|
if (CC == TOPClass) {
|
|
for (auto M : *CC) {
|
|
auto *VTE = ValueToExpression.lookup(M);
|
|
if (VTE && isa<DeadExpression>(VTE))
|
|
markInstructionForDeletion(cast<Instruction>(M));
|
|
assert((!ReachableBlocks.count(cast<Instruction>(M)->getParent()) ||
|
|
InstructionsToErase.count(cast<Instruction>(M))) &&
|
|
"Everything in TOP should be unreachable or dead at this "
|
|
"point");
|
|
}
|
|
continue;
|
|
}
|
|
|
|
assert(CC->getLeader() && "We should have had a leader");
|
|
// If this is a leader that is always available, and it's a
|
|
// constant or has no equivalences, just replace everything with
|
|
// it. We then update the congruence class with whatever members
|
|
// are left.
|
|
Value *Leader =
|
|
CC->getStoredValue() ? CC->getStoredValue() : CC->getLeader();
|
|
if (alwaysAvailable(Leader)) {
|
|
CongruenceClass::MemberSet MembersLeft;
|
|
for (auto M : *CC) {
|
|
Value *Member = M;
|
|
// Void things have no uses we can replace.
|
|
if (Member == Leader || !isa<Instruction>(Member) ||
|
|
Member->getType()->isVoidTy()) {
|
|
MembersLeft.insert(Member);
|
|
continue;
|
|
}
|
|
LLVM_DEBUG(dbgs() << "Found replacement " << *(Leader) << " for "
|
|
<< *Member << "\n");
|
|
auto *I = cast<Instruction>(Member);
|
|
assert(Leader != I && "About to accidentally remove our leader");
|
|
replaceInstruction(I, Leader);
|
|
AnythingReplaced = true;
|
|
}
|
|
CC->swap(MembersLeft);
|
|
} else {
|
|
// If this is a singleton, we can skip it.
|
|
if (CC->size() != 1 || RealToTemp.count(Leader)) {
|
|
// This is a stack because equality replacement/etc may place
|
|
// constants in the middle of the member list, and we want to use
|
|
// those constant values in preference to the current leader, over
|
|
// the scope of those constants.
|
|
ValueDFSStack EliminationStack;
|
|
|
|
// Convert the members to DFS ordered sets and then merge them.
|
|
SmallVector<ValueDFS, 8> DFSOrderedSet;
|
|
convertClassToDFSOrdered(*CC, DFSOrderedSet, UseCounts, ProbablyDead);
|
|
|
|
// Sort the whole thing.
|
|
llvm::sort(DFSOrderedSet);
|
|
for (auto &VD : DFSOrderedSet) {
|
|
int MemberDFSIn = VD.DFSIn;
|
|
int MemberDFSOut = VD.DFSOut;
|
|
Value *Def = VD.Def.getPointer();
|
|
bool FromStore = VD.Def.getInt();
|
|
Use *U = VD.U;
|
|
// We ignore void things because we can't get a value from them.
|
|
if (Def && Def->getType()->isVoidTy())
|
|
continue;
|
|
auto *DefInst = dyn_cast_or_null<Instruction>(Def);
|
|
if (DefInst && AllTempInstructions.count(DefInst)) {
|
|
auto *PN = cast<PHINode>(DefInst);
|
|
|
|
// If this is a value phi and that's the expression we used, insert
|
|
// it into the program
|
|
// remove from temp instruction list.
|
|
AllTempInstructions.erase(PN);
|
|
auto *DefBlock = getBlockForValue(Def);
|
|
LLVM_DEBUG(dbgs() << "Inserting fully real phi of ops" << *Def
|
|
<< " into block "
|
|
<< getBlockName(getBlockForValue(Def)) << "\n");
|
|
PN->insertBefore(&DefBlock->front());
|
|
Def = PN;
|
|
NumGVNPHIOfOpsEliminations++;
|
|
}
|
|
|
|
if (EliminationStack.empty()) {
|
|
LLVM_DEBUG(dbgs() << "Elimination Stack is empty\n");
|
|
} else {
|
|
LLVM_DEBUG(dbgs() << "Elimination Stack Top DFS numbers are ("
|
|
<< EliminationStack.dfs_back().first << ","
|
|
<< EliminationStack.dfs_back().second << ")\n");
|
|
}
|
|
|
|
LLVM_DEBUG(dbgs() << "Current DFS numbers are (" << MemberDFSIn << ","
|
|
<< MemberDFSOut << ")\n");
|
|
// First, we see if we are out of scope or empty. If so,
|
|
// and there equivalences, we try to replace the top of
|
|
// stack with equivalences (if it's on the stack, it must
|
|
// not have been eliminated yet).
|
|
// Then we synchronize to our current scope, by
|
|
// popping until we are back within a DFS scope that
|
|
// dominates the current member.
|
|
// Then, what happens depends on a few factors
|
|
// If the stack is now empty, we need to push
|
|
// If we have a constant or a local equivalence we want to
|
|
// start using, we also push.
|
|
// Otherwise, we walk along, processing members who are
|
|
// dominated by this scope, and eliminate them.
|
|
bool ShouldPush = Def && EliminationStack.empty();
|
|
bool OutOfScope =
|
|
!EliminationStack.isInScope(MemberDFSIn, MemberDFSOut);
|
|
|
|
if (OutOfScope || ShouldPush) {
|
|
// Sync to our current scope.
|
|
EliminationStack.popUntilDFSScope(MemberDFSIn, MemberDFSOut);
|
|
bool ShouldPush = Def && EliminationStack.empty();
|
|
if (ShouldPush) {
|
|
EliminationStack.push_back(Def, MemberDFSIn, MemberDFSOut);
|
|
}
|
|
}
|
|
|
|
// Skip the Def's, we only want to eliminate on their uses. But mark
|
|
// dominated defs as dead.
|
|
if (Def) {
|
|
// For anything in this case, what and how we value number
|
|
// guarantees that any side-effets that would have occurred (ie
|
|
// throwing, etc) can be proven to either still occur (because it's
|
|
// dominated by something that has the same side-effects), or never
|
|
// occur. Otherwise, we would not have been able to prove it value
|
|
// equivalent to something else. For these things, we can just mark
|
|
// it all dead. Note that this is different from the "ProbablyDead"
|
|
// set, which may not be dominated by anything, and thus, are only
|
|
// easy to prove dead if they are also side-effect free. Note that
|
|
// because stores are put in terms of the stored value, we skip
|
|
// stored values here. If the stored value is really dead, it will
|
|
// still be marked for deletion when we process it in its own class.
|
|
if (!EliminationStack.empty() && Def != EliminationStack.back() &&
|
|
isa<Instruction>(Def) && !FromStore)
|
|
markInstructionForDeletion(cast<Instruction>(Def));
|
|
continue;
|
|
}
|
|
// At this point, we know it is a Use we are trying to possibly
|
|
// replace.
|
|
|
|
assert(isa<Instruction>(U->get()) &&
|
|
"Current def should have been an instruction");
|
|
assert(isa<Instruction>(U->getUser()) &&
|
|
"Current user should have been an instruction");
|
|
|
|
// If the thing we are replacing into is already marked to be dead,
|
|
// this use is dead. Note that this is true regardless of whether
|
|
// we have anything dominating the use or not. We do this here
|
|
// because we are already walking all the uses anyway.
|
|
Instruction *InstUse = cast<Instruction>(U->getUser());
|
|
if (InstructionsToErase.count(InstUse)) {
|
|
auto &UseCount = UseCounts[U->get()];
|
|
if (--UseCount == 0) {
|
|
ProbablyDead.insert(cast<Instruction>(U->get()));
|
|
}
|
|
}
|
|
|
|
// If we get to this point, and the stack is empty we must have a use
|
|
// with nothing we can use to eliminate this use, so just skip it.
|
|
if (EliminationStack.empty())
|
|
continue;
|
|
|
|
Value *DominatingLeader = EliminationStack.back();
|
|
|
|
auto *II = dyn_cast<IntrinsicInst>(DominatingLeader);
|
|
bool isSSACopy = II && II->getIntrinsicID() == Intrinsic::ssa_copy;
|
|
if (isSSACopy)
|
|
DominatingLeader = II->getOperand(0);
|
|
|
|
// Don't replace our existing users with ourselves.
|
|
if (U->get() == DominatingLeader)
|
|
continue;
|
|
LLVM_DEBUG(dbgs()
|
|
<< "Found replacement " << *DominatingLeader << " for "
|
|
<< *U->get() << " in " << *(U->getUser()) << "\n");
|
|
|
|
// If we replaced something in an instruction, handle the patching of
|
|
// metadata. Skip this if we are replacing predicateinfo with its
|
|
// original operand, as we already know we can just drop it.
|
|
auto *ReplacedInst = cast<Instruction>(U->get());
|
|
auto *PI = PredInfo->getPredicateInfoFor(ReplacedInst);
|
|
if (!PI || DominatingLeader != PI->OriginalOp)
|
|
patchReplacementInstruction(ReplacedInst, DominatingLeader);
|
|
U->set(DominatingLeader);
|
|
// This is now a use of the dominating leader, which means if the
|
|
// dominating leader was dead, it's now live!
|
|
auto &LeaderUseCount = UseCounts[DominatingLeader];
|
|
// It's about to be alive again.
|
|
if (LeaderUseCount == 0 && isa<Instruction>(DominatingLeader))
|
|
ProbablyDead.erase(cast<Instruction>(DominatingLeader));
|
|
// For copy instructions, we use their operand as a leader,
|
|
// which means we remove a user of the copy and it may become dead.
|
|
if (isSSACopy) {
|
|
unsigned &IIUseCount = UseCounts[II];
|
|
if (--IIUseCount == 0)
|
|
ProbablyDead.insert(II);
|
|
}
|
|
++LeaderUseCount;
|
|
AnythingReplaced = true;
|
|
}
|
|
}
|
|
}
|
|
|
|
// At this point, anything still in the ProbablyDead set is actually dead if
|
|
// would be trivially dead.
|
|
for (auto *I : ProbablyDead)
|
|
if (wouldInstructionBeTriviallyDead(I))
|
|
markInstructionForDeletion(I);
|
|
|
|
// Cleanup the congruence class.
|
|
CongruenceClass::MemberSet MembersLeft;
|
|
for (auto *Member : *CC)
|
|
if (!isa<Instruction>(Member) ||
|
|
!InstructionsToErase.count(cast<Instruction>(Member)))
|
|
MembersLeft.insert(Member);
|
|
CC->swap(MembersLeft);
|
|
|
|
// If we have possible dead stores to look at, try to eliminate them.
|
|
if (CC->getStoreCount() > 0) {
|
|
convertClassToLoadsAndStores(*CC, PossibleDeadStores);
|
|
llvm::sort(PossibleDeadStores);
|
|
ValueDFSStack EliminationStack;
|
|
for (auto &VD : PossibleDeadStores) {
|
|
int MemberDFSIn = VD.DFSIn;
|
|
int MemberDFSOut = VD.DFSOut;
|
|
Instruction *Member = cast<Instruction>(VD.Def.getPointer());
|
|
if (EliminationStack.empty() ||
|
|
!EliminationStack.isInScope(MemberDFSIn, MemberDFSOut)) {
|
|
// Sync to our current scope.
|
|
EliminationStack.popUntilDFSScope(MemberDFSIn, MemberDFSOut);
|
|
if (EliminationStack.empty()) {
|
|
EliminationStack.push_back(Member, MemberDFSIn, MemberDFSOut);
|
|
continue;
|
|
}
|
|
}
|
|
// We already did load elimination, so nothing to do here.
|
|
if (isa<LoadInst>(Member))
|
|
continue;
|
|
assert(!EliminationStack.empty());
|
|
Instruction *Leader = cast<Instruction>(EliminationStack.back());
|
|
(void)Leader;
|
|
assert(DT->dominates(Leader->getParent(), Member->getParent()));
|
|
// Member is dominater by Leader, and thus dead
|
|
LLVM_DEBUG(dbgs() << "Marking dead store " << *Member
|
|
<< " that is dominated by " << *Leader << "\n");
|
|
markInstructionForDeletion(Member);
|
|
CC->erase(Member);
|
|
++NumGVNDeadStores;
|
|
}
|
|
}
|
|
}
|
|
return AnythingReplaced;
|
|
}
|
|
|
|
// This function provides global ranking of operations so that we can place them
|
|
// in a canonical order. Note that rank alone is not necessarily enough for a
|
|
// complete ordering, as constants all have the same rank. However, generally,
|
|
// we will simplify an operation with all constants so that it doesn't matter
|
|
// what order they appear in.
|
|
unsigned int NewGVN::getRank(const Value *V) const {
|
|
// Prefer constants to undef to anything else
|
|
// Undef is a constant, have to check it first.
|
|
// Prefer smaller constants to constantexprs
|
|
if (isa<ConstantExpr>(V))
|
|
return 2;
|
|
if (isa<UndefValue>(V))
|
|
return 1;
|
|
if (isa<Constant>(V))
|
|
return 0;
|
|
else if (auto *A = dyn_cast<Argument>(V))
|
|
return 3 + A->getArgNo();
|
|
|
|
// Need to shift the instruction DFS by number of arguments + 3 to account for
|
|
// the constant and argument ranking above.
|
|
unsigned Result = InstrToDFSNum(V);
|
|
if (Result > 0)
|
|
return 4 + NumFuncArgs + Result;
|
|
// Unreachable or something else, just return a really large number.
|
|
return ~0;
|
|
}
|
|
|
|
// This is a function that says whether two commutative operations should
|
|
// have their order swapped when canonicalizing.
|
|
bool NewGVN::shouldSwapOperands(const Value *A, const Value *B) const {
|
|
// Because we only care about a total ordering, and don't rewrite expressions
|
|
// in this order, we order by rank, which will give a strict weak ordering to
|
|
// everything but constants, and then we order by pointer address.
|
|
return std::make_pair(getRank(A), A) > std::make_pair(getRank(B), B);
|
|
}
|
|
|
|
namespace {
|
|
|
|
class NewGVNLegacyPass : public FunctionPass {
|
|
public:
|
|
// Pass identification, replacement for typeid.
|
|
static char ID;
|
|
|
|
NewGVNLegacyPass() : FunctionPass(ID) {
|
|
initializeNewGVNLegacyPassPass(*PassRegistry::getPassRegistry());
|
|
}
|
|
|
|
bool runOnFunction(Function &F) override;
|
|
|
|
private:
|
|
void getAnalysisUsage(AnalysisUsage &AU) const override {
|
|
AU.addRequired<AssumptionCacheTracker>();
|
|
AU.addRequired<DominatorTreeWrapperPass>();
|
|
AU.addRequired<TargetLibraryInfoWrapperPass>();
|
|
AU.addRequired<MemorySSAWrapperPass>();
|
|
AU.addRequired<AAResultsWrapperPass>();
|
|
AU.addPreserved<DominatorTreeWrapperPass>();
|
|
AU.addPreserved<GlobalsAAWrapperPass>();
|
|
}
|
|
};
|
|
|
|
} // end anonymous namespace
|
|
|
|
bool NewGVNLegacyPass::runOnFunction(Function &F) {
|
|
if (skipFunction(F))
|
|
return false;
|
|
return NewGVN(F, &getAnalysis<DominatorTreeWrapperPass>().getDomTree(),
|
|
&getAnalysis<AssumptionCacheTracker>().getAssumptionCache(F),
|
|
&getAnalysis<TargetLibraryInfoWrapperPass>().getTLI(F),
|
|
&getAnalysis<AAResultsWrapperPass>().getAAResults(),
|
|
&getAnalysis<MemorySSAWrapperPass>().getMSSA(),
|
|
F.getParent()->getDataLayout())
|
|
.runGVN();
|
|
}
|
|
|
|
char NewGVNLegacyPass::ID = 0;
|
|
|
|
INITIALIZE_PASS_BEGIN(NewGVNLegacyPass, "newgvn", "Global Value Numbering",
|
|
false, false)
|
|
INITIALIZE_PASS_DEPENDENCY(AssumptionCacheTracker)
|
|
INITIALIZE_PASS_DEPENDENCY(MemorySSAWrapperPass)
|
|
INITIALIZE_PASS_DEPENDENCY(DominatorTreeWrapperPass)
|
|
INITIALIZE_PASS_DEPENDENCY(TargetLibraryInfoWrapperPass)
|
|
INITIALIZE_PASS_DEPENDENCY(AAResultsWrapperPass)
|
|
INITIALIZE_PASS_DEPENDENCY(GlobalsAAWrapperPass)
|
|
INITIALIZE_PASS_END(NewGVNLegacyPass, "newgvn", "Global Value Numbering", false,
|
|
false)
|
|
|
|
// createGVNPass - The public interface to this file.
|
|
FunctionPass *llvm::createNewGVNPass() { return new NewGVNLegacyPass(); }
|
|
|
|
PreservedAnalyses NewGVNPass::run(Function &F, AnalysisManager<Function> &AM) {
|
|
// Apparently the order in which we get these results matter for
|
|
// the old GVN (see Chandler's comment in GVN.cpp). I'll keep
|
|
// the same order here, just in case.
|
|
auto &AC = AM.getResult<AssumptionAnalysis>(F);
|
|
auto &DT = AM.getResult<DominatorTreeAnalysis>(F);
|
|
auto &TLI = AM.getResult<TargetLibraryAnalysis>(F);
|
|
auto &AA = AM.getResult<AAManager>(F);
|
|
auto &MSSA = AM.getResult<MemorySSAAnalysis>(F).getMSSA();
|
|
bool Changed =
|
|
NewGVN(F, &DT, &AC, &TLI, &AA, &MSSA, F.getParent()->getDataLayout())
|
|
.runGVN();
|
|
if (!Changed)
|
|
return PreservedAnalyses::all();
|
|
PreservedAnalyses PA;
|
|
PA.preserve<DominatorTreeAnalysis>();
|
|
PA.preserve<GlobalsAA>();
|
|
return PA;
|
|
}
|