1
0
mirror of https://github.com/RPCS3/llvm-mirror.git synced 2024-11-25 12:12:47 +01:00
llvm-mirror/lib/Analysis/IPA/Andersens.cpp
Jeffrey Yasskin 23ac706aab Fix DenseMap iterator constness.
This patch forbids implicit conversion of DenseMap::const_iterator to
DenseMap::iterator which was possible because DenseMapIterator inherited
(publicly) from DenseMapConstIterator. Conversion the other way around is now
allowed as one may expect.

The template DenseMapConstIterator is removed and the template parameter
IsConst which specifies whether the iterator is constant is added to
DenseMapIterator.

Actually IsConst parameter is not necessary since the constness can be
determined from KeyT but this is not relevant to the fix and can be addressed
later.

Patch by Victor Zverovich!

llvm-svn: 86636
2009-11-10 01:02:17 +00:00

2888 lines
103 KiB
C++

//===- Andersens.cpp - Andersen's Interprocedural Alias Analysis ----------===//
//
// The LLVM Compiler Infrastructure
//
// This file is distributed under the University of Illinois Open Source
// License. See LICENSE.TXT for details.
//
//===----------------------------------------------------------------------===//
//
// This file defines an implementation of Andersen's interprocedural alias
// analysis
//
// In pointer analysis terms, this is a subset-based, flow-insensitive,
// field-sensitive, and context-insensitive algorithm pointer algorithm.
//
// This algorithm is implemented as three stages:
// 1. Object identification.
// 2. Inclusion constraint identification.
// 3. Offline constraint graph optimization
// 4. Inclusion constraint solving.
//
// The object identification stage identifies all of the memory objects in the
// program, which includes globals, heap allocated objects, and stack allocated
// objects.
//
// The inclusion constraint identification stage finds all inclusion constraints
// in the program by scanning the program, looking for pointer assignments and
// other statements that effect the points-to graph. For a statement like "A =
// B", this statement is processed to indicate that A can point to anything that
// B can point to. Constraints can handle copies, loads, and stores, and
// address taking.
//
// The offline constraint graph optimization portion includes offline variable
// substitution algorithms intended to compute pointer and location
// equivalences. Pointer equivalences are those pointers that will have the
// same points-to sets, and location equivalences are those variables that
// always appear together in points-to sets. It also includes an offline
// cycle detection algorithm that allows cycles to be collapsed sooner
// during solving.
//
// The inclusion constraint solving phase iteratively propagates the inclusion
// constraints until a fixed point is reached. This is an O(N^3) algorithm.
//
// Function constraints are handled as if they were structs with X fields.
// Thus, an access to argument X of function Y is an access to node index
// getNode(Y) + X. This representation allows handling of indirect calls
// without any issues. To wit, an indirect call Y(a,b) is equivalent to
// *(Y + 1) = a, *(Y + 2) = b.
// The return node for a function is always located at getNode(F) +
// CallReturnPos. The arguments start at getNode(F) + CallArgPos.
//
// Future Improvements:
// Use of BDD's.
//===----------------------------------------------------------------------===//
#define DEBUG_TYPE "anders-aa"
#include "llvm/Constants.h"
#include "llvm/DerivedTypes.h"
#include "llvm/Instructions.h"
#include "llvm/Module.h"
#include "llvm/Pass.h"
#include "llvm/Support/ErrorHandling.h"
#include "llvm/Support/InstIterator.h"
#include "llvm/Support/InstVisitor.h"
#include "llvm/Analysis/AliasAnalysis.h"
#include "llvm/Analysis/MemoryBuiltins.h"
#include "llvm/Analysis/Passes.h"
#include "llvm/Support/Debug.h"
#include "llvm/System/Atomic.h"
#include "llvm/ADT/Statistic.h"
#include "llvm/ADT/SparseBitVector.h"
#include "llvm/ADT/DenseSet.h"
#include <algorithm>
#include <set>
#include <list>
#include <map>
#include <stack>
#include <vector>
#include <queue>
// Determining the actual set of nodes the universal set can consist of is very
// expensive because it means propagating around very large sets. We rely on
// other analysis being able to determine which nodes can never be pointed to in
// order to disambiguate further than "points-to anything".
#define FULL_UNIVERSAL 0
using namespace llvm;
#ifndef NDEBUG
STATISTIC(NumIters , "Number of iterations to reach convergence");
#endif
STATISTIC(NumConstraints, "Number of constraints");
STATISTIC(NumNodes , "Number of nodes");
STATISTIC(NumUnified , "Number of variables unified");
STATISTIC(NumErased , "Number of redundant constraints erased");
static const unsigned SelfRep = (unsigned)-1;
static const unsigned Unvisited = (unsigned)-1;
// Position of the function return node relative to the function node.
static const unsigned CallReturnPos = 1;
// Position of the function call node relative to the function node.
static const unsigned CallFirstArgPos = 2;
namespace {
struct BitmapKeyInfo {
static inline SparseBitVector<> *getEmptyKey() {
return reinterpret_cast<SparseBitVector<> *>(-1);
}
static inline SparseBitVector<> *getTombstoneKey() {
return reinterpret_cast<SparseBitVector<> *>(-2);
}
static unsigned getHashValue(const SparseBitVector<> *bitmap) {
return bitmap->getHashValue();
}
static bool isEqual(const SparseBitVector<> *LHS,
const SparseBitVector<> *RHS) {
if (LHS == RHS)
return true;
else if (LHS == getEmptyKey() || RHS == getEmptyKey()
|| LHS == getTombstoneKey() || RHS == getTombstoneKey())
return false;
return *LHS == *RHS;
}
static bool isPod() { return true; }
};
class Andersens : public ModulePass, public AliasAnalysis,
private InstVisitor<Andersens> {
struct Node;
/// Constraint - Objects of this structure are used to represent the various
/// constraints identified by the algorithm. The constraints are 'copy',
/// for statements like "A = B", 'load' for statements like "A = *B",
/// 'store' for statements like "*A = B", and AddressOf for statements like
/// A = alloca; The Offset is applied as *(A + K) = B for stores,
/// A = *(B + K) for loads, and A = B + K for copies. It is
/// illegal on addressof constraints (because it is statically
/// resolvable to A = &C where C = B + K)
struct Constraint {
enum ConstraintType { Copy, Load, Store, AddressOf } Type;
unsigned Dest;
unsigned Src;
unsigned Offset;
Constraint(ConstraintType Ty, unsigned D, unsigned S, unsigned O = 0)
: Type(Ty), Dest(D), Src(S), Offset(O) {
assert((Offset == 0 || Ty != AddressOf) &&
"Offset is illegal on addressof constraints");
}
bool operator==(const Constraint &RHS) const {
return RHS.Type == Type
&& RHS.Dest == Dest
&& RHS.Src == Src
&& RHS.Offset == Offset;
}
bool operator!=(const Constraint &RHS) const {
return !(*this == RHS);
}
bool operator<(const Constraint &RHS) const {
if (RHS.Type != Type)
return RHS.Type < Type;
else if (RHS.Dest != Dest)
return RHS.Dest < Dest;
else if (RHS.Src != Src)
return RHS.Src < Src;
return RHS.Offset < Offset;
}
};
// Information DenseSet requires implemented in order to be able to do
// it's thing
struct PairKeyInfo {
static inline std::pair<unsigned, unsigned> getEmptyKey() {
return std::make_pair(~0U, ~0U);
}
static inline std::pair<unsigned, unsigned> getTombstoneKey() {
return std::make_pair(~0U - 1, ~0U - 1);
}
static unsigned getHashValue(const std::pair<unsigned, unsigned> &P) {
return P.first ^ P.second;
}
static unsigned isEqual(const std::pair<unsigned, unsigned> &LHS,
const std::pair<unsigned, unsigned> &RHS) {
return LHS == RHS;
}
};
struct ConstraintKeyInfo {
static inline Constraint getEmptyKey() {
return Constraint(Constraint::Copy, ~0U, ~0U, ~0U);
}
static inline Constraint getTombstoneKey() {
return Constraint(Constraint::Copy, ~0U - 1, ~0U - 1, ~0U - 1);
}
static unsigned getHashValue(const Constraint &C) {
return C.Src ^ C.Dest ^ C.Type ^ C.Offset;
}
static bool isEqual(const Constraint &LHS,
const Constraint &RHS) {
return LHS.Type == RHS.Type && LHS.Dest == RHS.Dest
&& LHS.Src == RHS.Src && LHS.Offset == RHS.Offset;
}
};
// Node class - This class is used to represent a node in the constraint
// graph. Due to various optimizations, it is not always the case that
// there is a mapping from a Node to a Value. In particular, we add
// artificial Node's that represent the set of pointed-to variables shared
// for each location equivalent Node.
struct Node {
private:
static volatile sys::cas_flag Counter;
public:
Value *Val;
SparseBitVector<> *Edges;
SparseBitVector<> *PointsTo;
SparseBitVector<> *OldPointsTo;
std::list<Constraint> Constraints;
// Pointer and location equivalence labels
unsigned PointerEquivLabel;
unsigned LocationEquivLabel;
// Predecessor edges, both real and implicit
SparseBitVector<> *PredEdges;
SparseBitVector<> *ImplicitPredEdges;
// Set of nodes that point to us, only use for location equivalence.
SparseBitVector<> *PointedToBy;
// Number of incoming edges, used during variable substitution to early
// free the points-to sets
unsigned NumInEdges;
// True if our points-to set is in the Set2PEClass map
bool StoredInHash;
// True if our node has no indirect constraints (complex or otherwise)
bool Direct;
// True if the node is address taken, *or* it is part of a group of nodes
// that must be kept together. This is set to true for functions and
// their arg nodes, which must be kept at the same position relative to
// their base function node.
bool AddressTaken;
// Nodes in cycles (or in equivalence classes) are united together using a
// standard union-find representation with path compression. NodeRep
// gives the index into GraphNodes for the representative Node.
unsigned NodeRep;
// Modification timestamp. Assigned from Counter.
// Used for work list prioritization.
unsigned Timestamp;
explicit Node(bool direct = true) :
Val(0), Edges(0), PointsTo(0), OldPointsTo(0),
PointerEquivLabel(0), LocationEquivLabel(0), PredEdges(0),
ImplicitPredEdges(0), PointedToBy(0), NumInEdges(0),
StoredInHash(false), Direct(direct), AddressTaken(false),
NodeRep(SelfRep), Timestamp(0) { }
Node *setValue(Value *V) {
assert(Val == 0 && "Value already set for this node!");
Val = V;
return this;
}
/// getValue - Return the LLVM value corresponding to this node.
///
Value *getValue() const { return Val; }
/// addPointerTo - Add a pointer to the list of pointees of this node,
/// returning true if this caused a new pointer to be added, or false if
/// we already knew about the points-to relation.
bool addPointerTo(unsigned Node) {
return PointsTo->test_and_set(Node);
}
/// intersects - Return true if the points-to set of this node intersects
/// with the points-to set of the specified node.
bool intersects(Node *N) const;
/// intersectsIgnoring - Return true if the points-to set of this node
/// intersects with the points-to set of the specified node on any nodes
/// except for the specified node to ignore.
bool intersectsIgnoring(Node *N, unsigned) const;
// Timestamp a node (used for work list prioritization)
void Stamp() {
Timestamp = sys::AtomicIncrement(&Counter);
--Timestamp;
}
bool isRep() const {
return( (int) NodeRep < 0 );
}
};
struct WorkListElement {
Node* node;
unsigned Timestamp;
WorkListElement(Node* n, unsigned t) : node(n), Timestamp(t) {}
// Note that we reverse the sense of the comparison because we
// actually want to give low timestamps the priority over high,
// whereas priority is typically interpreted as a greater value is
// given high priority.
bool operator<(const WorkListElement& that) const {
return( this->Timestamp > that.Timestamp );
}
};
// Priority-queue based work list specialized for Nodes.
class WorkList {
std::priority_queue<WorkListElement> Q;
public:
void insert(Node* n) {
Q.push( WorkListElement(n, n->Timestamp) );
}
// We automatically discard non-representative nodes and nodes
// that were in the work list twice (we keep a copy of the
// timestamp in the work list so we can detect this situation by
// comparing against the node's current timestamp).
Node* pop() {
while( !Q.empty() ) {
WorkListElement x = Q.top(); Q.pop();
Node* INode = x.node;
if( INode->isRep() &&
INode->Timestamp == x.Timestamp ) {
return(x.node);
}
}
return(0);
}
bool empty() {
return Q.empty();
}
};
/// GraphNodes - This vector is populated as part of the object
/// identification stage of the analysis, which populates this vector with a
/// node for each memory object and fills in the ValueNodes map.
std::vector<Node> GraphNodes;
/// ValueNodes - This map indicates the Node that a particular Value* is
/// represented by. This contains entries for all pointers.
DenseMap<Value*, unsigned> ValueNodes;
/// ObjectNodes - This map contains entries for each memory object in the
/// program: globals, alloca's and mallocs.
DenseMap<Value*, unsigned> ObjectNodes;
/// ReturnNodes - This map contains an entry for each function in the
/// program that returns a value.
DenseMap<Function*, unsigned> ReturnNodes;
/// VarargNodes - This map contains the entry used to represent all pointers
/// passed through the varargs portion of a function call for a particular
/// function. An entry is not present in this map for functions that do not
/// take variable arguments.
DenseMap<Function*, unsigned> VarargNodes;
/// Constraints - This vector contains a list of all of the constraints
/// identified by the program.
std::vector<Constraint> Constraints;
// Map from graph node to maximum K value that is allowed (for functions,
// this is equivalent to the number of arguments + CallFirstArgPos)
std::map<unsigned, unsigned> MaxK;
/// This enum defines the GraphNodes indices that correspond to important
/// fixed sets.
enum {
UniversalSet = 0,
NullPtr = 1,
NullObject = 2,
NumberSpecialNodes
};
// Stack for Tarjan's
std::stack<unsigned> SCCStack;
// Map from Graph Node to DFS number
std::vector<unsigned> Node2DFS;
// Map from Graph Node to Deleted from graph.
std::vector<bool> Node2Deleted;
// Same as Node Maps, but implemented as std::map because it is faster to
// clear
std::map<unsigned, unsigned> Tarjan2DFS;
std::map<unsigned, bool> Tarjan2Deleted;
// Current DFS number
unsigned DFSNumber;
// Work lists.
WorkList w1, w2;
WorkList *CurrWL, *NextWL; // "current" and "next" work lists
// Offline variable substitution related things
// Temporary rep storage, used because we can't collapse SCC's in the
// predecessor graph by uniting the variables permanently, we can only do so
// for the successor graph.
std::vector<unsigned> VSSCCRep;
// Mapping from node to whether we have visited it during SCC finding yet.
std::vector<bool> Node2Visited;
// During variable substitution, we create unknowns to represent the unknown
// value that is a dereference of a variable. These nodes are known as
// "ref" nodes (since they represent the value of dereferences).
unsigned FirstRefNode;
// During HVN, we create represent address taken nodes as if they were
// unknown (since HVN, unlike HU, does not evaluate unions).
unsigned FirstAdrNode;
// Current pointer equivalence class number
unsigned PEClass;
// Mapping from points-to sets to equivalence classes
typedef DenseMap<SparseBitVector<> *, unsigned, BitmapKeyInfo> BitVectorMap;
BitVectorMap Set2PEClass;
// Mapping from pointer equivalences to the representative node. -1 if we
// have no representative node for this pointer equivalence class yet.
std::vector<int> PEClass2Node;
// Mapping from pointer equivalences to representative node. This includes
// pointer equivalent but not location equivalent variables. -1 if we have
// no representative node for this pointer equivalence class yet.
std::vector<int> PENLEClass2Node;
// Union/Find for HCD
std::vector<unsigned> HCDSCCRep;
// HCD's offline-detected cycles; "Statically DeTected"
// -1 if not part of such a cycle, otherwise a representative node.
std::vector<int> SDT;
// Whether to use SDT (UniteNodes can use it during solving, but not before)
bool SDTActive;
public:
static char ID;
Andersens() : ModulePass(&ID) {}
bool runOnModule(Module &M) {
InitializeAliasAnalysis(this);
IdentifyObjects(M);
CollectConstraints(M);
#undef DEBUG_TYPE
#define DEBUG_TYPE "anders-aa-constraints"
DEBUG(PrintConstraints());
#undef DEBUG_TYPE
#define DEBUG_TYPE "anders-aa"
SolveConstraints();
DEBUG(PrintPointsToGraph());
// Free the constraints list, as we don't need it to respond to alias
// requests.
std::vector<Constraint>().swap(Constraints);
//These are needed for Print() (-analyze in opt)
//ObjectNodes.clear();
//ReturnNodes.clear();
//VarargNodes.clear();
return false;
}
void releaseMemory() {
// FIXME: Until we have transitively required passes working correctly,
// this cannot be enabled! Otherwise, using -count-aa with the pass
// causes memory to be freed too early. :(
#if 0
// The memory objects and ValueNodes data structures at the only ones that
// are still live after construction.
std::vector<Node>().swap(GraphNodes);
ValueNodes.clear();
#endif
}
virtual void getAnalysisUsage(AnalysisUsage &AU) const {
AliasAnalysis::getAnalysisUsage(AU);
AU.setPreservesAll(); // Does not transform code
}
//------------------------------------------------
// Implement the AliasAnalysis API
//
AliasResult alias(const Value *V1, unsigned V1Size,
const Value *V2, unsigned V2Size);
virtual ModRefResult getModRefInfo(CallSite CS, Value *P, unsigned Size);
virtual ModRefResult getModRefInfo(CallSite CS1, CallSite CS2);
void getMustAliases(Value *P, std::vector<Value*> &RetVals);
bool pointsToConstantMemory(const Value *P);
virtual void deleteValue(Value *V) {
ValueNodes.erase(V);
getAnalysis<AliasAnalysis>().deleteValue(V);
}
virtual void copyValue(Value *From, Value *To) {
ValueNodes[To] = ValueNodes[From];
getAnalysis<AliasAnalysis>().copyValue(From, To);
}
private:
/// getNode - Return the node corresponding to the specified pointer scalar.
///
unsigned getNode(Value *V) {
if (Constant *C = dyn_cast<Constant>(V))
if (!isa<GlobalValue>(C))
return getNodeForConstantPointer(C);
DenseMap<Value*, unsigned>::iterator I = ValueNodes.find(V);
if (I == ValueNodes.end()) {
#ifndef NDEBUG
V->dump();
#endif
llvm_unreachable("Value does not have a node in the points-to graph!");
}
return I->second;
}
/// getObject - Return the node corresponding to the memory object for the
/// specified global or allocation instruction.
unsigned getObject(Value *V) const {
DenseMap<Value*, unsigned>::const_iterator I = ObjectNodes.find(V);
assert(I != ObjectNodes.end() &&
"Value does not have an object in the points-to graph!");
return I->second;
}
/// getReturnNode - Return the node representing the return value for the
/// specified function.
unsigned getReturnNode(Function *F) const {
DenseMap<Function*, unsigned>::const_iterator I = ReturnNodes.find(F);
assert(I != ReturnNodes.end() && "Function does not return a value!");
return I->second;
}
/// getVarargNode - Return the node representing the variable arguments
/// formal for the specified function.
unsigned getVarargNode(Function *F) const {
DenseMap<Function*, unsigned>::const_iterator I = VarargNodes.find(F);
assert(I != VarargNodes.end() && "Function does not take var args!");
return I->second;
}
/// getNodeValue - Get the node for the specified LLVM value and set the
/// value for it to be the specified value.
unsigned getNodeValue(Value &V) {
unsigned Index = getNode(&V);
GraphNodes[Index].setValue(&V);
return Index;
}
unsigned UniteNodes(unsigned First, unsigned Second,
bool UnionByRank = true);
unsigned FindNode(unsigned Node);
unsigned FindNode(unsigned Node) const;
void IdentifyObjects(Module &M);
void CollectConstraints(Module &M);
bool AnalyzeUsesOfFunction(Value *);
void CreateConstraintGraph();
void OptimizeConstraints();
unsigned FindEquivalentNode(unsigned, unsigned);
void ClumpAddressTaken();
void RewriteConstraints();
void HU();
void HVN();
void HCD();
void Search(unsigned Node);
void UnitePointerEquivalences();
void SolveConstraints();
bool QueryNode(unsigned Node);
void Condense(unsigned Node);
void HUValNum(unsigned Node);
void HVNValNum(unsigned Node);
unsigned getNodeForConstantPointer(Constant *C);
unsigned getNodeForConstantPointerTarget(Constant *C);
void AddGlobalInitializerConstraints(unsigned, Constant *C);
void AddConstraintsForNonInternalLinkage(Function *F);
void AddConstraintsForCall(CallSite CS, Function *F);
bool AddConstraintsForExternalCall(CallSite CS, Function *F);
void PrintNode(const Node *N) const;
void PrintConstraints() const ;
void PrintConstraint(const Constraint &) const;
void PrintLabels() const;
void PrintPointsToGraph() const;
//===------------------------------------------------------------------===//
// Instruction visitation methods for adding constraints
//
friend class InstVisitor<Andersens>;
void visitReturnInst(ReturnInst &RI);
void visitInvokeInst(InvokeInst &II) { visitCallSite(CallSite(&II)); }
void visitCallInst(CallInst &CI) {
if (isMalloc(&CI)) visitAlloc(CI);
else visitCallSite(CallSite(&CI));
}
void visitCallSite(CallSite CS);
void visitAllocaInst(AllocaInst &I);
void visitAlloc(Instruction &I);
void visitLoadInst(LoadInst &LI);
void visitStoreInst(StoreInst &SI);
void visitGetElementPtrInst(GetElementPtrInst &GEP);
void visitPHINode(PHINode &PN);
void visitCastInst(CastInst &CI);
void visitICmpInst(ICmpInst &ICI) {} // NOOP!
void visitFCmpInst(FCmpInst &ICI) {} // NOOP!
void visitSelectInst(SelectInst &SI);
void visitVAArg(VAArgInst &I);
void visitInstruction(Instruction &I);
//===------------------------------------------------------------------===//
// Implement Analyize interface
//
void print(raw_ostream &O, const Module*) const {
PrintPointsToGraph();
}
};
}
char Andersens::ID = 0;
static RegisterPass<Andersens>
X("anders-aa", "Andersen's Interprocedural Alias Analysis (experimental)",
false, true);
static RegisterAnalysisGroup<AliasAnalysis> Y(X);
// Initialize Timestamp Counter (static).
volatile llvm::sys::cas_flag Andersens::Node::Counter = 0;
ModulePass *llvm::createAndersensPass() { return new Andersens(); }
//===----------------------------------------------------------------------===//
// AliasAnalysis Interface Implementation
//===----------------------------------------------------------------------===//
AliasAnalysis::AliasResult Andersens::alias(const Value *V1, unsigned V1Size,
const Value *V2, unsigned V2Size) {
Node *N1 = &GraphNodes[FindNode(getNode(const_cast<Value*>(V1)))];
Node *N2 = &GraphNodes[FindNode(getNode(const_cast<Value*>(V2)))];
// Check to see if the two pointers are known to not alias. They don't alias
// if their points-to sets do not intersect.
if (!N1->intersectsIgnoring(N2, NullObject))
return NoAlias;
return AliasAnalysis::alias(V1, V1Size, V2, V2Size);
}
AliasAnalysis::ModRefResult
Andersens::getModRefInfo(CallSite CS, Value *P, unsigned Size) {
// The only thing useful that we can contribute for mod/ref information is
// when calling external function calls: if we know that memory never escapes
// from the program, it cannot be modified by an external call.
//
// NOTE: This is not really safe, at least not when the entire program is not
// available. The deal is that the external function could call back into the
// program and modify stuff. We ignore this technical niggle for now. This
// is, after all, a "research quality" implementation of Andersen's analysis.
if (Function *F = CS.getCalledFunction())
if (F->isDeclaration()) {
Node *N1 = &GraphNodes[FindNode(getNode(P))];
if (N1->PointsTo->empty())
return NoModRef;
#if FULL_UNIVERSAL
if (!UniversalSet->PointsTo->test(FindNode(getNode(P))))
return NoModRef; // Universal set does not contain P
#else
if (!N1->PointsTo->test(UniversalSet))
return NoModRef; // P doesn't point to the universal set.
#endif
}
return AliasAnalysis::getModRefInfo(CS, P, Size);
}
AliasAnalysis::ModRefResult
Andersens::getModRefInfo(CallSite CS1, CallSite CS2) {
return AliasAnalysis::getModRefInfo(CS1,CS2);
}
/// getMustAlias - We can provide must alias information if we know that a
/// pointer can only point to a specific function or the null pointer.
/// Unfortunately we cannot determine must-alias information for global
/// variables or any other memory memory objects because we do not track whether
/// a pointer points to the beginning of an object or a field of it.
void Andersens::getMustAliases(Value *P, std::vector<Value*> &RetVals) {
Node *N = &GraphNodes[FindNode(getNode(P))];
if (N->PointsTo->count() == 1) {
Node *Pointee = &GraphNodes[N->PointsTo->find_first()];
// If a function is the only object in the points-to set, then it must be
// the destination. Note that we can't handle global variables here,
// because we don't know if the pointer is actually pointing to a field of
// the global or to the beginning of it.
if (Value *V = Pointee->getValue()) {
if (Function *F = dyn_cast<Function>(V))
RetVals.push_back(F);
} else {
// If the object in the points-to set is the null object, then the null
// pointer is a must alias.
if (Pointee == &GraphNodes[NullObject])
RetVals.push_back(Constant::getNullValue(P->getType()));
}
}
AliasAnalysis::getMustAliases(P, RetVals);
}
/// pointsToConstantMemory - If we can determine that this pointer only points
/// to constant memory, return true. In practice, this means that if the
/// pointer can only point to constant globals, functions, or the null pointer,
/// return true.
///
bool Andersens::pointsToConstantMemory(const Value *P) {
Node *N = &GraphNodes[FindNode(getNode(const_cast<Value*>(P)))];
unsigned i;
for (SparseBitVector<>::iterator bi = N->PointsTo->begin();
bi != N->PointsTo->end();
++bi) {
i = *bi;
Node *Pointee = &GraphNodes[i];
if (Value *V = Pointee->getValue()) {
if (!isa<GlobalValue>(V) || (isa<GlobalVariable>(V) &&
!cast<GlobalVariable>(V)->isConstant()))
return AliasAnalysis::pointsToConstantMemory(P);
} else {
if (i != NullObject)
return AliasAnalysis::pointsToConstantMemory(P);
}
}
return true;
}
//===----------------------------------------------------------------------===//
// Object Identification Phase
//===----------------------------------------------------------------------===//
/// IdentifyObjects - This stage scans the program, adding an entry to the
/// GraphNodes list for each memory object in the program (global stack or
/// heap), and populates the ValueNodes and ObjectNodes maps for these objects.
///
void Andersens::IdentifyObjects(Module &M) {
unsigned NumObjects = 0;
// Object #0 is always the universal set: the object that we don't know
// anything about.
assert(NumObjects == UniversalSet && "Something changed!");
++NumObjects;
// Object #1 always represents the null pointer.
assert(NumObjects == NullPtr && "Something changed!");
++NumObjects;
// Object #2 always represents the null object (the object pointed to by null)
assert(NumObjects == NullObject && "Something changed!");
++NumObjects;
// Add all the globals first.
for (Module::global_iterator I = M.global_begin(), E = M.global_end();
I != E; ++I) {
ObjectNodes[I] = NumObjects++;
ValueNodes[I] = NumObjects++;
}
// Add nodes for all of the functions and the instructions inside of them.
for (Module::iterator F = M.begin(), E = M.end(); F != E; ++F) {
// The function itself is a memory object.
unsigned First = NumObjects;
ValueNodes[F] = NumObjects++;
if (isa<PointerType>(F->getFunctionType()->getReturnType()))
ReturnNodes[F] = NumObjects++;
if (F->getFunctionType()->isVarArg())
VarargNodes[F] = NumObjects++;
// Add nodes for all of the incoming pointer arguments.
for (Function::arg_iterator I = F->arg_begin(), E = F->arg_end();
I != E; ++I)
{
if (isa<PointerType>(I->getType()))
ValueNodes[I] = NumObjects++;
}
MaxK[First] = NumObjects - First;
// Scan the function body, creating a memory object for each heap/stack
// allocation in the body of the function and a node to represent all
// pointer values defined by instructions and used as operands.
for (inst_iterator II = inst_begin(F), E = inst_end(F); II != E; ++II) {
// If this is an heap or stack allocation, create a node for the memory
// object.
if (isa<PointerType>(II->getType())) {
ValueNodes[&*II] = NumObjects++;
if (AllocaInst *AI = dyn_cast<AllocaInst>(&*II))
ObjectNodes[AI] = NumObjects++;
else if (isMalloc(&*II))
ObjectNodes[&*II] = NumObjects++;
}
// Calls to inline asm need to be added as well because the callee isn't
// referenced anywhere else.
if (CallInst *CI = dyn_cast<CallInst>(&*II)) {
Value *Callee = CI->getCalledValue();
if (isa<InlineAsm>(Callee))
ValueNodes[Callee] = NumObjects++;
}
}
}
// Now that we know how many objects to create, make them all now!
GraphNodes.resize(NumObjects);
NumNodes += NumObjects;
}
//===----------------------------------------------------------------------===//
// Constraint Identification Phase
//===----------------------------------------------------------------------===//
/// getNodeForConstantPointer - Return the node corresponding to the constant
/// pointer itself.
unsigned Andersens::getNodeForConstantPointer(Constant *C) {
assert(isa<PointerType>(C->getType()) && "Not a constant pointer!");
if (isa<ConstantPointerNull>(C) || isa<UndefValue>(C))
return NullPtr;
else if (GlobalValue *GV = dyn_cast<GlobalValue>(C))
return getNode(GV);
else if (ConstantExpr *CE = dyn_cast<ConstantExpr>(C)) {
switch (CE->getOpcode()) {
case Instruction::GetElementPtr:
return getNodeForConstantPointer(CE->getOperand(0));
case Instruction::IntToPtr:
return UniversalSet;
case Instruction::BitCast:
return getNodeForConstantPointer(CE->getOperand(0));
default:
errs() << "Constant Expr not yet handled: " << *CE << "\n";
llvm_unreachable(0);
}
} else {
llvm_unreachable("Unknown constant pointer!");
}
return 0;
}
/// getNodeForConstantPointerTarget - Return the node POINTED TO by the
/// specified constant pointer.
unsigned Andersens::getNodeForConstantPointerTarget(Constant *C) {
assert(isa<PointerType>(C->getType()) && "Not a constant pointer!");
if (isa<ConstantPointerNull>(C))
return NullObject;
else if (GlobalValue *GV = dyn_cast<GlobalValue>(C))
return getObject(GV);
else if (ConstantExpr *CE = dyn_cast<ConstantExpr>(C)) {
switch (CE->getOpcode()) {
case Instruction::GetElementPtr:
return getNodeForConstantPointerTarget(CE->getOperand(0));
case Instruction::IntToPtr:
return UniversalSet;
case Instruction::BitCast:
return getNodeForConstantPointerTarget(CE->getOperand(0));
default:
errs() << "Constant Expr not yet handled: " << *CE << "\n";
llvm_unreachable(0);
}
} else {
llvm_unreachable("Unknown constant pointer!");
}
return 0;
}
/// AddGlobalInitializerConstraints - Add inclusion constraints for the memory
/// object N, which contains values indicated by C.
void Andersens::AddGlobalInitializerConstraints(unsigned NodeIndex,
Constant *C) {
if (C->getType()->isSingleValueType()) {
if (isa<PointerType>(C->getType()))
Constraints.push_back(Constraint(Constraint::Copy, NodeIndex,
getNodeForConstantPointer(C)));
} else if (C->isNullValue()) {
Constraints.push_back(Constraint(Constraint::Copy, NodeIndex,
NullObject));
return;
} else if (!isa<UndefValue>(C)) {
// If this is an array or struct, include constraints for each element.
assert(isa<ConstantArray>(C) || isa<ConstantStruct>(C));
for (unsigned i = 0, e = C->getNumOperands(); i != e; ++i)
AddGlobalInitializerConstraints(NodeIndex,
cast<Constant>(C->getOperand(i)));
}
}
/// AddConstraintsForNonInternalLinkage - If this function does not have
/// internal linkage, realize that we can't trust anything passed into or
/// returned by this function.
void Andersens::AddConstraintsForNonInternalLinkage(Function *F) {
for (Function::arg_iterator I = F->arg_begin(), E = F->arg_end(); I != E; ++I)
if (isa<PointerType>(I->getType()))
// If this is an argument of an externally accessible function, the
// incoming pointer might point to anything.
Constraints.push_back(Constraint(Constraint::Copy, getNode(I),
UniversalSet));
}
/// AddConstraintsForCall - If this is a call to a "known" function, add the
/// constraints and return true. If this is a call to an unknown function,
/// return false.
bool Andersens::AddConstraintsForExternalCall(CallSite CS, Function *F) {
assert(F->isDeclaration() && "Not an external function!");
// These functions don't induce any points-to constraints.
if (F->getName() == "atoi" || F->getName() == "atof" ||
F->getName() == "atol" || F->getName() == "atoll" ||
F->getName() == "remove" || F->getName() == "unlink" ||
F->getName() == "rename" || F->getName() == "memcmp" ||
F->getName() == "llvm.memset" ||
F->getName() == "strcmp" || F->getName() == "strncmp" ||
F->getName() == "execl" || F->getName() == "execlp" ||
F->getName() == "execle" || F->getName() == "execv" ||
F->getName() == "execvp" || F->getName() == "chmod" ||
F->getName() == "puts" || F->getName() == "write" ||
F->getName() == "open" || F->getName() == "create" ||
F->getName() == "truncate" || F->getName() == "chdir" ||
F->getName() == "mkdir" || F->getName() == "rmdir" ||
F->getName() == "read" || F->getName() == "pipe" ||
F->getName() == "wait" || F->getName() == "time" ||
F->getName() == "stat" || F->getName() == "fstat" ||
F->getName() == "lstat" || F->getName() == "strtod" ||
F->getName() == "strtof" || F->getName() == "strtold" ||
F->getName() == "fopen" || F->getName() == "fdopen" ||
F->getName() == "freopen" ||
F->getName() == "fflush" || F->getName() == "feof" ||
F->getName() == "fileno" || F->getName() == "clearerr" ||
F->getName() == "rewind" || F->getName() == "ftell" ||
F->getName() == "ferror" || F->getName() == "fgetc" ||
F->getName() == "fgetc" || F->getName() == "_IO_getc" ||
F->getName() == "fwrite" || F->getName() == "fread" ||
F->getName() == "fgets" || F->getName() == "ungetc" ||
F->getName() == "fputc" ||
F->getName() == "fputs" || F->getName() == "putc" ||
F->getName() == "ftell" || F->getName() == "rewind" ||
F->getName() == "_IO_putc" || F->getName() == "fseek" ||
F->getName() == "fgetpos" || F->getName() == "fsetpos" ||
F->getName() == "printf" || F->getName() == "fprintf" ||
F->getName() == "sprintf" || F->getName() == "vprintf" ||
F->getName() == "vfprintf" || F->getName() == "vsprintf" ||
F->getName() == "scanf" || F->getName() == "fscanf" ||
F->getName() == "sscanf" || F->getName() == "__assert_fail" ||
F->getName() == "modf")
return true;
// These functions do induce points-to edges.
if (F->getName() == "llvm.memcpy" ||
F->getName() == "llvm.memmove" ||
F->getName() == "memmove") {
const FunctionType *FTy = F->getFunctionType();
if (FTy->getNumParams() > 1 &&
isa<PointerType>(FTy->getParamType(0)) &&
isa<PointerType>(FTy->getParamType(1))) {
// *Dest = *Src, which requires an artificial graph node to represent the
// constraint. It is broken up into *Dest = temp, temp = *Src
unsigned FirstArg = getNode(CS.getArgument(0));
unsigned SecondArg = getNode(CS.getArgument(1));
unsigned TempArg = GraphNodes.size();
GraphNodes.push_back(Node());
Constraints.push_back(Constraint(Constraint::Store,
FirstArg, TempArg));
Constraints.push_back(Constraint(Constraint::Load,
TempArg, SecondArg));
// In addition, Dest = Src
Constraints.push_back(Constraint(Constraint::Copy,
FirstArg, SecondArg));
return true;
}
}
// Result = Arg0
if (F->getName() == "realloc" || F->getName() == "strchr" ||
F->getName() == "strrchr" || F->getName() == "strstr" ||
F->getName() == "strtok") {
const FunctionType *FTy = F->getFunctionType();
if (FTy->getNumParams() > 0 &&
isa<PointerType>(FTy->getParamType(0))) {
Constraints.push_back(Constraint(Constraint::Copy,
getNode(CS.getInstruction()),
getNode(CS.getArgument(0))));
return true;
}
}
return false;
}
/// AnalyzeUsesOfFunction - Look at all of the users of the specified function.
/// If this is used by anything complex (i.e., the address escapes), return
/// true.
bool Andersens::AnalyzeUsesOfFunction(Value *V) {
if (!isa<PointerType>(V->getType())) return true;
for (Value::use_iterator UI = V->use_begin(), E = V->use_end(); UI != E; ++UI)
if (isa<LoadInst>(*UI)) {
return false;
} else if (StoreInst *SI = dyn_cast<StoreInst>(*UI)) {
if (V == SI->getOperand(1)) {
return false;
} else if (SI->getOperand(1)) {
return true; // Storing the pointer
}
} else if (GetElementPtrInst *GEP = dyn_cast<GetElementPtrInst>(*UI)) {
if (AnalyzeUsesOfFunction(GEP)) return true;
} else if (isFreeCall(*UI)) {
return false;
} else if (CallInst *CI = dyn_cast<CallInst>(*UI)) {
// Make sure that this is just the function being called, not that it is
// passing into the function.
for (unsigned i = 1, e = CI->getNumOperands(); i != e; ++i)
if (CI->getOperand(i) == V) return true;
} else if (InvokeInst *II = dyn_cast<InvokeInst>(*UI)) {
// Make sure that this is just the function being called, not that it is
// passing into the function.
for (unsigned i = 3, e = II->getNumOperands(); i != e; ++i)
if (II->getOperand(i) == V) return true;
} else if (ConstantExpr *CE = dyn_cast<ConstantExpr>(*UI)) {
if (CE->getOpcode() == Instruction::GetElementPtr ||
CE->getOpcode() == Instruction::BitCast) {
if (AnalyzeUsesOfFunction(CE))
return true;
} else {
return true;
}
} else if (ICmpInst *ICI = dyn_cast<ICmpInst>(*UI)) {
if (!isa<ConstantPointerNull>(ICI->getOperand(1)))
return true; // Allow comparison against null.
} else {
return true;
}
return false;
}
/// CollectConstraints - This stage scans the program, adding a constraint to
/// the Constraints list for each instruction in the program that induces a
/// constraint, and setting up the initial points-to graph.
///
void Andersens::CollectConstraints(Module &M) {
// First, the universal set points to itself.
Constraints.push_back(Constraint(Constraint::AddressOf, UniversalSet,
UniversalSet));
Constraints.push_back(Constraint(Constraint::Store, UniversalSet,
UniversalSet));
// Next, the null pointer points to the null object.
Constraints.push_back(Constraint(Constraint::AddressOf, NullPtr, NullObject));
// Next, add any constraints on global variables and their initializers.
for (Module::global_iterator I = M.global_begin(), E = M.global_end();
I != E; ++I) {
// Associate the address of the global object as pointing to the memory for
// the global: &G = <G memory>
unsigned ObjectIndex = getObject(I);
Node *Object = &GraphNodes[ObjectIndex];
Object->setValue(I);
Constraints.push_back(Constraint(Constraint::AddressOf, getNodeValue(*I),
ObjectIndex));
if (I->hasDefinitiveInitializer()) {
AddGlobalInitializerConstraints(ObjectIndex, I->getInitializer());
} else {
// If it doesn't have an initializer (i.e. it's defined in another
// translation unit), it points to the universal set.
Constraints.push_back(Constraint(Constraint::Copy, ObjectIndex,
UniversalSet));
}
}
for (Module::iterator F = M.begin(), E = M.end(); F != E; ++F) {
// Set up the return value node.
if (isa<PointerType>(F->getFunctionType()->getReturnType()))
GraphNodes[getReturnNode(F)].setValue(F);
if (F->getFunctionType()->isVarArg())
GraphNodes[getVarargNode(F)].setValue(F);
// Set up incoming argument nodes.
for (Function::arg_iterator I = F->arg_begin(), E = F->arg_end();
I != E; ++I)
if (isa<PointerType>(I->getType()))
getNodeValue(*I);
// At some point we should just add constraints for the escaping functions
// at solve time, but this slows down solving. For now, we simply mark
// address taken functions as escaping and treat them as external.
if (!F->hasLocalLinkage() || AnalyzeUsesOfFunction(F))
AddConstraintsForNonInternalLinkage(F);
if (!F->isDeclaration()) {
// Scan the function body, creating a memory object for each heap/stack
// allocation in the body of the function and a node to represent all
// pointer values defined by instructions and used as operands.
visit(F);
} else {
// External functions that return pointers return the universal set.
if (isa<PointerType>(F->getFunctionType()->getReturnType()))
Constraints.push_back(Constraint(Constraint::Copy,
getReturnNode(F),
UniversalSet));
// Any pointers that are passed into the function have the universal set
// stored into them.
for (Function::arg_iterator I = F->arg_begin(), E = F->arg_end();
I != E; ++I)
if (isa<PointerType>(I->getType())) {
// Pointers passed into external functions could have anything stored
// through them.
Constraints.push_back(Constraint(Constraint::Store, getNode(I),
UniversalSet));
// Memory objects passed into external function calls can have the
// universal set point to them.
#if FULL_UNIVERSAL
Constraints.push_back(Constraint(Constraint::Copy,
UniversalSet,
getNode(I)));
#else
Constraints.push_back(Constraint(Constraint::Copy,
getNode(I),
UniversalSet));
#endif
}
// If this is an external varargs function, it can also store pointers
// into any pointers passed through the varargs section.
if (F->getFunctionType()->isVarArg())
Constraints.push_back(Constraint(Constraint::Store, getVarargNode(F),
UniversalSet));
}
}
NumConstraints += Constraints.size();
}
void Andersens::visitInstruction(Instruction &I) {
#ifdef NDEBUG
return; // This function is just a big assert.
#endif
if (isa<BinaryOperator>(I))
return;
// Most instructions don't have any effect on pointer values.
switch (I.getOpcode()) {
case Instruction::Br:
case Instruction::Switch:
case Instruction::Unwind:
case Instruction::Unreachable:
case Instruction::ICmp:
case Instruction::FCmp:
return;
default:
// Is this something we aren't handling yet?
errs() << "Unknown instruction: " << I;
llvm_unreachable(0);
}
}
void Andersens::visitAllocaInst(AllocaInst &I) {
visitAlloc(I);
}
void Andersens::visitAlloc(Instruction &I) {
unsigned ObjectIndex = getObject(&I);
GraphNodes[ObjectIndex].setValue(&I);
Constraints.push_back(Constraint(Constraint::AddressOf, getNodeValue(I),
ObjectIndex));
}
void Andersens::visitReturnInst(ReturnInst &RI) {
if (RI.getNumOperands() && isa<PointerType>(RI.getOperand(0)->getType()))
// return V --> <Copy/retval{F}/v>
Constraints.push_back(Constraint(Constraint::Copy,
getReturnNode(RI.getParent()->getParent()),
getNode(RI.getOperand(0))));
}
void Andersens::visitLoadInst(LoadInst &LI) {
if (isa<PointerType>(LI.getType()))
// P1 = load P2 --> <Load/P1/P2>
Constraints.push_back(Constraint(Constraint::Load, getNodeValue(LI),
getNode(LI.getOperand(0))));
}
void Andersens::visitStoreInst(StoreInst &SI) {
if (isa<PointerType>(SI.getOperand(0)->getType()))
// store P1, P2 --> <Store/P2/P1>
Constraints.push_back(Constraint(Constraint::Store,
getNode(SI.getOperand(1)),
getNode(SI.getOperand(0))));
}
void Andersens::visitGetElementPtrInst(GetElementPtrInst &GEP) {
// P1 = getelementptr P2, ... --> <Copy/P1/P2>
Constraints.push_back(Constraint(Constraint::Copy, getNodeValue(GEP),
getNode(GEP.getOperand(0))));
}
void Andersens::visitPHINode(PHINode &PN) {
if (isa<PointerType>(PN.getType())) {
unsigned PNN = getNodeValue(PN);
for (unsigned i = 0, e = PN.getNumIncomingValues(); i != e; ++i)
// P1 = phi P2, P3 --> <Copy/P1/P2>, <Copy/P1/P3>, ...
Constraints.push_back(Constraint(Constraint::Copy, PNN,
getNode(PN.getIncomingValue(i))));
}
}
void Andersens::visitCastInst(CastInst &CI) {
Value *Op = CI.getOperand(0);
if (isa<PointerType>(CI.getType())) {
if (isa<PointerType>(Op->getType())) {
// P1 = cast P2 --> <Copy/P1/P2>
Constraints.push_back(Constraint(Constraint::Copy, getNodeValue(CI),
getNode(CI.getOperand(0))));
} else {
// P1 = cast int --> <Copy/P1/Univ>
#if 0
Constraints.push_back(Constraint(Constraint::Copy, getNodeValue(CI),
UniversalSet));
#else
getNodeValue(CI);
#endif
}
} else if (isa<PointerType>(Op->getType())) {
// int = cast P1 --> <Copy/Univ/P1>
#if 0
Constraints.push_back(Constraint(Constraint::Copy,
UniversalSet,
getNode(CI.getOperand(0))));
#else
getNode(CI.getOperand(0));
#endif
}
}
void Andersens::visitSelectInst(SelectInst &SI) {
if (isa<PointerType>(SI.getType())) {
unsigned SIN = getNodeValue(SI);
// P1 = select C, P2, P3 ---> <Copy/P1/P2>, <Copy/P1/P3>
Constraints.push_back(Constraint(Constraint::Copy, SIN,
getNode(SI.getOperand(1))));
Constraints.push_back(Constraint(Constraint::Copy, SIN,
getNode(SI.getOperand(2))));
}
}
void Andersens::visitVAArg(VAArgInst &I) {
llvm_unreachable("vaarg not handled yet!");
}
/// AddConstraintsForCall - Add constraints for a call with actual arguments
/// specified by CS to the function specified by F. Note that the types of
/// arguments might not match up in the case where this is an indirect call and
/// the function pointer has been casted. If this is the case, do something
/// reasonable.
void Andersens::AddConstraintsForCall(CallSite CS, Function *F) {
Value *CallValue = CS.getCalledValue();
bool IsDeref = F == NULL;
// If this is a call to an external function, try to handle it directly to get
// some taste of context sensitivity.
if (F && F->isDeclaration() && AddConstraintsForExternalCall(CS, F))
return;
if (isa<PointerType>(CS.getType())) {
unsigned CSN = getNode(CS.getInstruction());
if (!F || isa<PointerType>(F->getFunctionType()->getReturnType())) {
if (IsDeref)
Constraints.push_back(Constraint(Constraint::Load, CSN,
getNode(CallValue), CallReturnPos));
else
Constraints.push_back(Constraint(Constraint::Copy, CSN,
getNode(CallValue) + CallReturnPos));
} else {
// If the function returns a non-pointer value, handle this just like we
// treat a nonpointer cast to pointer.
Constraints.push_back(Constraint(Constraint::Copy, CSN,
UniversalSet));
}
} else if (F && isa<PointerType>(F->getFunctionType()->getReturnType())) {
#if FULL_UNIVERSAL
Constraints.push_back(Constraint(Constraint::Copy,
UniversalSet,
getNode(CallValue) + CallReturnPos));
#else
Constraints.push_back(Constraint(Constraint::Copy,
getNode(CallValue) + CallReturnPos,
UniversalSet));
#endif
}
CallSite::arg_iterator ArgI = CS.arg_begin(), ArgE = CS.arg_end();
bool external = !F || F->isDeclaration();
if (F) {
// Direct Call
Function::arg_iterator AI = F->arg_begin(), AE = F->arg_end();
for (; AI != AE && ArgI != ArgE; ++AI, ++ArgI)
{
#if !FULL_UNIVERSAL
if (external && isa<PointerType>((*ArgI)->getType()))
{
// Add constraint that ArgI can now point to anything due to
// escaping, as can everything it points to. The second portion of
// this should be taken care of by universal = *universal
Constraints.push_back(Constraint(Constraint::Copy,
getNode(*ArgI),
UniversalSet));
}
#endif
if (isa<PointerType>(AI->getType())) {
if (isa<PointerType>((*ArgI)->getType())) {
// Copy the actual argument into the formal argument.
Constraints.push_back(Constraint(Constraint::Copy, getNode(AI),
getNode(*ArgI)));
} else {
Constraints.push_back(Constraint(Constraint::Copy, getNode(AI),
UniversalSet));
}
} else if (isa<PointerType>((*ArgI)->getType())) {
#if FULL_UNIVERSAL
Constraints.push_back(Constraint(Constraint::Copy,
UniversalSet,
getNode(*ArgI)));
#else
Constraints.push_back(Constraint(Constraint::Copy,
getNode(*ArgI),
UniversalSet));
#endif
}
}
} else {
//Indirect Call
unsigned ArgPos = CallFirstArgPos;
for (; ArgI != ArgE; ++ArgI) {
if (isa<PointerType>((*ArgI)->getType())) {
// Copy the actual argument into the formal argument.
Constraints.push_back(Constraint(Constraint::Store,
getNode(CallValue),
getNode(*ArgI), ArgPos++));
} else {
Constraints.push_back(Constraint(Constraint::Store,
getNode (CallValue),
UniversalSet, ArgPos++));
}
}
}
// Copy all pointers passed through the varargs section to the varargs node.
if (F && F->getFunctionType()->isVarArg())
for (; ArgI != ArgE; ++ArgI)
if (isa<PointerType>((*ArgI)->getType()))
Constraints.push_back(Constraint(Constraint::Copy, getVarargNode(F),
getNode(*ArgI)));
// If more arguments are passed in than we track, just drop them on the floor.
}
void Andersens::visitCallSite(CallSite CS) {
if (isa<PointerType>(CS.getType()))
getNodeValue(*CS.getInstruction());
if (Function *F = CS.getCalledFunction()) {
AddConstraintsForCall(CS, F);
} else {
AddConstraintsForCall(CS, NULL);
}
}
//===----------------------------------------------------------------------===//
// Constraint Solving Phase
//===----------------------------------------------------------------------===//
/// intersects - Return true if the points-to set of this node intersects
/// with the points-to set of the specified node.
bool Andersens::Node::intersects(Node *N) const {
return PointsTo->intersects(N->PointsTo);
}
/// intersectsIgnoring - Return true if the points-to set of this node
/// intersects with the points-to set of the specified node on any nodes
/// except for the specified node to ignore.
bool Andersens::Node::intersectsIgnoring(Node *N, unsigned Ignoring) const {
// TODO: If we are only going to call this with the same value for Ignoring,
// we should move the special values out of the points-to bitmap.
bool WeHadIt = PointsTo->test(Ignoring);
bool NHadIt = N->PointsTo->test(Ignoring);
bool Result = false;
if (WeHadIt)
PointsTo->reset(Ignoring);
if (NHadIt)
N->PointsTo->reset(Ignoring);
Result = PointsTo->intersects(N->PointsTo);
if (WeHadIt)
PointsTo->set(Ignoring);
if (NHadIt)
N->PointsTo->set(Ignoring);
return Result;
}
/// Clump together address taken variables so that the points-to sets use up
/// less space and can be operated on faster.
void Andersens::ClumpAddressTaken() {
#undef DEBUG_TYPE
#define DEBUG_TYPE "anders-aa-renumber"
std::vector<unsigned> Translate;
std::vector<Node> NewGraphNodes;
Translate.resize(GraphNodes.size());
unsigned NewPos = 0;
for (unsigned i = 0; i < Constraints.size(); ++i) {
Constraint &C = Constraints[i];
if (C.Type == Constraint::AddressOf) {
GraphNodes[C.Src].AddressTaken = true;
}
}
for (unsigned i = 0; i < NumberSpecialNodes; ++i) {
unsigned Pos = NewPos++;
Translate[i] = Pos;
NewGraphNodes.push_back(GraphNodes[i]);
DEBUG(errs() << "Renumbering node " << i << " to node " << Pos << "\n");
}
// I believe this ends up being faster than making two vectors and splicing
// them.
for (unsigned i = NumberSpecialNodes; i < GraphNodes.size(); ++i) {
if (GraphNodes[i].AddressTaken) {
unsigned Pos = NewPos++;
Translate[i] = Pos;
NewGraphNodes.push_back(GraphNodes[i]);
DEBUG(errs() << "Renumbering node " << i << " to node " << Pos << "\n");
}
}
for (unsigned i = NumberSpecialNodes; i < GraphNodes.size(); ++i) {
if (!GraphNodes[i].AddressTaken) {
unsigned Pos = NewPos++;
Translate[i] = Pos;
NewGraphNodes.push_back(GraphNodes[i]);
DEBUG(errs() << "Renumbering node " << i << " to node " << Pos << "\n");
}
}
for (DenseMap<Value*, unsigned>::iterator Iter = ValueNodes.begin();
Iter != ValueNodes.end();
++Iter)
Iter->second = Translate[Iter->second];
for (DenseMap<Value*, unsigned>::iterator Iter = ObjectNodes.begin();
Iter != ObjectNodes.end();
++Iter)
Iter->second = Translate[Iter->second];
for (DenseMap<Function*, unsigned>::iterator Iter = ReturnNodes.begin();
Iter != ReturnNodes.end();
++Iter)
Iter->second = Translate[Iter->second];
for (DenseMap<Function*, unsigned>::iterator Iter = VarargNodes.begin();
Iter != VarargNodes.end();
++Iter)
Iter->second = Translate[Iter->second];
for (unsigned i = 0; i < Constraints.size(); ++i) {
Constraint &C = Constraints[i];
C.Src = Translate[C.Src];
C.Dest = Translate[C.Dest];
}
GraphNodes.swap(NewGraphNodes);
#undef DEBUG_TYPE
#define DEBUG_TYPE "anders-aa"
}
/// The technique used here is described in "Exploiting Pointer and Location
/// Equivalence to Optimize Pointer Analysis. In the 14th International Static
/// Analysis Symposium (SAS), August 2007." It is known as the "HVN" algorithm,
/// and is equivalent to value numbering the collapsed constraint graph without
/// evaluating unions. This is used as a pre-pass to HU in order to resolve
/// first order pointer dereferences and speed up/reduce memory usage of HU.
/// Running both is equivalent to HRU without the iteration
/// HVN in more detail:
/// Imagine the set of constraints was simply straight line code with no loops
/// (we eliminate cycles, so there are no loops), such as:
/// E = &D
/// E = &C
/// E = F
/// F = G
/// G = F
/// Applying value numbering to this code tells us:
/// G == F == E
///
/// For HVN, this is as far as it goes. We assign new value numbers to every
/// "address node", and every "reference node".
/// To get the optimal result for this, we use a DFS + SCC (since all nodes in a
/// cycle must have the same value number since the = operation is really
/// inclusion, not overwrite), and value number nodes we receive points-to sets
/// before we value our own node.
/// The advantage of HU over HVN is that HU considers the inclusion property, so
/// that if you have
/// E = &D
/// E = &C
/// E = F
/// F = G
/// F = &D
/// G = F
/// HU will determine that G == F == E. HVN will not, because it cannot prove
/// that the points to information ends up being the same because they all
/// receive &D from E anyway.
void Andersens::HVN() {
DEBUG(errs() << "Beginning HVN\n");
// Build a predecessor graph. This is like our constraint graph with the
// edges going in the opposite direction, and there are edges for all the
// constraints, instead of just copy constraints. We also build implicit
// edges for constraints are implied but not explicit. I.E for the constraint
// a = &b, we add implicit edges *a = b. This helps us capture more cycles
for (unsigned i = 0, e = Constraints.size(); i != e; ++i) {
Constraint &C = Constraints[i];
if (C.Type == Constraint::AddressOf) {
GraphNodes[C.Src].AddressTaken = true;
GraphNodes[C.Src].Direct = false;
// Dest = &src edge
unsigned AdrNode = C.Src + FirstAdrNode;
if (!GraphNodes[C.Dest].PredEdges)
GraphNodes[C.Dest].PredEdges = new SparseBitVector<>;
GraphNodes[C.Dest].PredEdges->set(AdrNode);
// *Dest = src edge
unsigned RefNode = C.Dest + FirstRefNode;
if (!GraphNodes[RefNode].ImplicitPredEdges)
GraphNodes[RefNode].ImplicitPredEdges = new SparseBitVector<>;
GraphNodes[RefNode].ImplicitPredEdges->set(C.Src);
} else if (C.Type == Constraint::Load) {
if (C.Offset == 0) {
// dest = *src edge
if (!GraphNodes[C.Dest].PredEdges)
GraphNodes[C.Dest].PredEdges = new SparseBitVector<>;
GraphNodes[C.Dest].PredEdges->set(C.Src + FirstRefNode);
} else {
GraphNodes[C.Dest].Direct = false;
}
} else if (C.Type == Constraint::Store) {
if (C.Offset == 0) {
// *dest = src edge
unsigned RefNode = C.Dest + FirstRefNode;
if (!GraphNodes[RefNode].PredEdges)
GraphNodes[RefNode].PredEdges = new SparseBitVector<>;
GraphNodes[RefNode].PredEdges->set(C.Src);
}
} else {
// Dest = Src edge and *Dest = *Src edge
if (!GraphNodes[C.Dest].PredEdges)
GraphNodes[C.Dest].PredEdges = new SparseBitVector<>;
GraphNodes[C.Dest].PredEdges->set(C.Src);
unsigned RefNode = C.Dest + FirstRefNode;
if (!GraphNodes[RefNode].ImplicitPredEdges)
GraphNodes[RefNode].ImplicitPredEdges = new SparseBitVector<>;
GraphNodes[RefNode].ImplicitPredEdges->set(C.Src + FirstRefNode);
}
}
PEClass = 1;
// Do SCC finding first to condense our predecessor graph
DFSNumber = 0;
Node2DFS.insert(Node2DFS.begin(), GraphNodes.size(), 0);
Node2Deleted.insert(Node2Deleted.begin(), GraphNodes.size(), false);
Node2Visited.insert(Node2Visited.begin(), GraphNodes.size(), false);
for (unsigned i = 0; i < FirstRefNode; ++i) {
unsigned Node = VSSCCRep[i];
if (!Node2Visited[Node])
HVNValNum(Node);
}
for (BitVectorMap::iterator Iter = Set2PEClass.begin();
Iter != Set2PEClass.end();
++Iter)
delete Iter->first;
Set2PEClass.clear();
Node2DFS.clear();
Node2Deleted.clear();
Node2Visited.clear();
DEBUG(errs() << "Finished HVN\n");
}
/// This is the workhorse of HVN value numbering. We combine SCC finding at the
/// same time because it's easy.
void Andersens::HVNValNum(unsigned NodeIndex) {
unsigned MyDFS = DFSNumber++;
Node *N = &GraphNodes[NodeIndex];
Node2Visited[NodeIndex] = true;
Node2DFS[NodeIndex] = MyDFS;
// First process all our explicit edges
if (N->PredEdges)
for (SparseBitVector<>::iterator Iter = N->PredEdges->begin();
Iter != N->PredEdges->end();
++Iter) {
unsigned j = VSSCCRep[*Iter];
if (!Node2Deleted[j]) {
if (!Node2Visited[j])
HVNValNum(j);
if (Node2DFS[NodeIndex] > Node2DFS[j])
Node2DFS[NodeIndex] = Node2DFS[j];
}
}
// Now process all the implicit edges
if (N->ImplicitPredEdges)
for (SparseBitVector<>::iterator Iter = N->ImplicitPredEdges->begin();
Iter != N->ImplicitPredEdges->end();
++Iter) {
unsigned j = VSSCCRep[*Iter];
if (!Node2Deleted[j]) {
if (!Node2Visited[j])
HVNValNum(j);
if (Node2DFS[NodeIndex] > Node2DFS[j])
Node2DFS[NodeIndex] = Node2DFS[j];
}
}
// See if we found any cycles
if (MyDFS == Node2DFS[NodeIndex]) {
while (!SCCStack.empty() && Node2DFS[SCCStack.top()] >= MyDFS) {
unsigned CycleNodeIndex = SCCStack.top();
Node *CycleNode = &GraphNodes[CycleNodeIndex];
VSSCCRep[CycleNodeIndex] = NodeIndex;
// Unify the nodes
N->Direct &= CycleNode->Direct;
if (CycleNode->PredEdges) {
if (!N->PredEdges)
N->PredEdges = new SparseBitVector<>;
*(N->PredEdges) |= CycleNode->PredEdges;
delete CycleNode->PredEdges;
CycleNode->PredEdges = NULL;
}
if (CycleNode->ImplicitPredEdges) {
if (!N->ImplicitPredEdges)
N->ImplicitPredEdges = new SparseBitVector<>;
*(N->ImplicitPredEdges) |= CycleNode->ImplicitPredEdges;
delete CycleNode->ImplicitPredEdges;
CycleNode->ImplicitPredEdges = NULL;
}
SCCStack.pop();
}
Node2Deleted[NodeIndex] = true;
if (!N->Direct) {
GraphNodes[NodeIndex].PointerEquivLabel = PEClass++;
return;
}
// Collect labels of successor nodes
bool AllSame = true;
unsigned First = ~0;
SparseBitVector<> *Labels = new SparseBitVector<>;
bool Used = false;
if (N->PredEdges)
for (SparseBitVector<>::iterator Iter = N->PredEdges->begin();
Iter != N->PredEdges->end();
++Iter) {
unsigned j = VSSCCRep[*Iter];
unsigned Label = GraphNodes[j].PointerEquivLabel;
// Ignore labels that are equal to us or non-pointers
if (j == NodeIndex || Label == 0)
continue;
if (First == (unsigned)~0)
First = Label;
else if (First != Label)
AllSame = false;
Labels->set(Label);
}
// We either have a non-pointer, a copy of an existing node, or a new node.
// Assign the appropriate pointer equivalence label.
if (Labels->empty()) {
GraphNodes[NodeIndex].PointerEquivLabel = 0;
} else if (AllSame) {
GraphNodes[NodeIndex].PointerEquivLabel = First;
} else {
GraphNodes[NodeIndex].PointerEquivLabel = Set2PEClass[Labels];
if (GraphNodes[NodeIndex].PointerEquivLabel == 0) {
unsigned EquivClass = PEClass++;
Set2PEClass[Labels] = EquivClass;
GraphNodes[NodeIndex].PointerEquivLabel = EquivClass;
Used = true;
}
}
if (!Used)
delete Labels;
} else {
SCCStack.push(NodeIndex);
}
}
/// The technique used here is described in "Exploiting Pointer and Location
/// Equivalence to Optimize Pointer Analysis. In the 14th International Static
/// Analysis Symposium (SAS), August 2007." It is known as the "HU" algorithm,
/// and is equivalent to value numbering the collapsed constraint graph
/// including evaluating unions.
void Andersens::HU() {
DEBUG(errs() << "Beginning HU\n");
// Build a predecessor graph. This is like our constraint graph with the
// edges going in the opposite direction, and there are edges for all the
// constraints, instead of just copy constraints. We also build implicit
// edges for constraints are implied but not explicit. I.E for the constraint
// a = &b, we add implicit edges *a = b. This helps us capture more cycles
for (unsigned i = 0, e = Constraints.size(); i != e; ++i) {
Constraint &C = Constraints[i];
if (C.Type == Constraint::AddressOf) {
GraphNodes[C.Src].AddressTaken = true;
GraphNodes[C.Src].Direct = false;
GraphNodes[C.Dest].PointsTo->set(C.Src);
// *Dest = src edge
unsigned RefNode = C.Dest + FirstRefNode;
if (!GraphNodes[RefNode].ImplicitPredEdges)
GraphNodes[RefNode].ImplicitPredEdges = new SparseBitVector<>;
GraphNodes[RefNode].ImplicitPredEdges->set(C.Src);
GraphNodes[C.Src].PointedToBy->set(C.Dest);
} else if (C.Type == Constraint::Load) {
if (C.Offset == 0) {
// dest = *src edge
if (!GraphNodes[C.Dest].PredEdges)
GraphNodes[C.Dest].PredEdges = new SparseBitVector<>;
GraphNodes[C.Dest].PredEdges->set(C.Src + FirstRefNode);
} else {
GraphNodes[C.Dest].Direct = false;
}
} else if (C.Type == Constraint::Store) {
if (C.Offset == 0) {
// *dest = src edge
unsigned RefNode = C.Dest + FirstRefNode;
if (!GraphNodes[RefNode].PredEdges)
GraphNodes[RefNode].PredEdges = new SparseBitVector<>;
GraphNodes[RefNode].PredEdges->set(C.Src);
}
} else {
// Dest = Src edge and *Dest = *Src edg
if (!GraphNodes[C.Dest].PredEdges)
GraphNodes[C.Dest].PredEdges = new SparseBitVector<>;
GraphNodes[C.Dest].PredEdges->set(C.Src);
unsigned RefNode = C.Dest + FirstRefNode;
if (!GraphNodes[RefNode].ImplicitPredEdges)
GraphNodes[RefNode].ImplicitPredEdges = new SparseBitVector<>;
GraphNodes[RefNode].ImplicitPredEdges->set(C.Src + FirstRefNode);
}
}
PEClass = 1;
// Do SCC finding first to condense our predecessor graph
DFSNumber = 0;
Node2DFS.insert(Node2DFS.begin(), GraphNodes.size(), 0);
Node2Deleted.insert(Node2Deleted.begin(), GraphNodes.size(), false);
Node2Visited.insert(Node2Visited.begin(), GraphNodes.size(), false);
for (unsigned i = 0; i < FirstRefNode; ++i) {
if (FindNode(i) == i) {
unsigned Node = VSSCCRep[i];
if (!Node2Visited[Node])
Condense(Node);
}
}
// Reset tables for actual labeling
Node2DFS.clear();
Node2Visited.clear();
Node2Deleted.clear();
// Pre-grow our densemap so that we don't get really bad behavior
Set2PEClass.resize(GraphNodes.size());
// Visit the condensed graph and generate pointer equivalence labels.
Node2Visited.insert(Node2Visited.begin(), GraphNodes.size(), false);
for (unsigned i = 0; i < FirstRefNode; ++i) {
if (FindNode(i) == i) {
unsigned Node = VSSCCRep[i];
if (!Node2Visited[Node])
HUValNum(Node);
}
}
// PEClass nodes will be deleted by the deleting of N->PointsTo in our caller.
Set2PEClass.clear();
DEBUG(errs() << "Finished HU\n");
}
/// Implementation of standard Tarjan SCC algorithm as modified by Nuutilla.
void Andersens::Condense(unsigned NodeIndex) {
unsigned MyDFS = DFSNumber++;
Node *N = &GraphNodes[NodeIndex];
Node2Visited[NodeIndex] = true;
Node2DFS[NodeIndex] = MyDFS;
// First process all our explicit edges
if (N->PredEdges)
for (SparseBitVector<>::iterator Iter = N->PredEdges->begin();
Iter != N->PredEdges->end();
++Iter) {
unsigned j = VSSCCRep[*Iter];
if (!Node2Deleted[j]) {
if (!Node2Visited[j])
Condense(j);
if (Node2DFS[NodeIndex] > Node2DFS[j])
Node2DFS[NodeIndex] = Node2DFS[j];
}
}
// Now process all the implicit edges
if (N->ImplicitPredEdges)
for (SparseBitVector<>::iterator Iter = N->ImplicitPredEdges->begin();
Iter != N->ImplicitPredEdges->end();
++Iter) {
unsigned j = VSSCCRep[*Iter];
if (!Node2Deleted[j]) {
if (!Node2Visited[j])
Condense(j);
if (Node2DFS[NodeIndex] > Node2DFS[j])
Node2DFS[NodeIndex] = Node2DFS[j];
}
}
// See if we found any cycles
if (MyDFS == Node2DFS[NodeIndex]) {
while (!SCCStack.empty() && Node2DFS[SCCStack.top()] >= MyDFS) {
unsigned CycleNodeIndex = SCCStack.top();
Node *CycleNode = &GraphNodes[CycleNodeIndex];
VSSCCRep[CycleNodeIndex] = NodeIndex;
// Unify the nodes
N->Direct &= CycleNode->Direct;
*(N->PointsTo) |= CycleNode->PointsTo;
delete CycleNode->PointsTo;
CycleNode->PointsTo = NULL;
if (CycleNode->PredEdges) {
if (!N->PredEdges)
N->PredEdges = new SparseBitVector<>;
*(N->PredEdges) |= CycleNode->PredEdges;
delete CycleNode->PredEdges;
CycleNode->PredEdges = NULL;
}
if (CycleNode->ImplicitPredEdges) {
if (!N->ImplicitPredEdges)
N->ImplicitPredEdges = new SparseBitVector<>;
*(N->ImplicitPredEdges) |= CycleNode->ImplicitPredEdges;
delete CycleNode->ImplicitPredEdges;
CycleNode->ImplicitPredEdges = NULL;
}
SCCStack.pop();
}
Node2Deleted[NodeIndex] = true;
// Set up number of incoming edges for other nodes
if (N->PredEdges)
for (SparseBitVector<>::iterator Iter = N->PredEdges->begin();
Iter != N->PredEdges->end();
++Iter)
++GraphNodes[VSSCCRep[*Iter]].NumInEdges;
} else {
SCCStack.push(NodeIndex);
}
}
void Andersens::HUValNum(unsigned NodeIndex) {
Node *N = &GraphNodes[NodeIndex];
Node2Visited[NodeIndex] = true;
// Eliminate dereferences of non-pointers for those non-pointers we have
// already identified. These are ref nodes whose non-ref node:
// 1. Has already been visited determined to point to nothing (and thus, a
// dereference of it must point to nothing)
// 2. Any direct node with no predecessor edges in our graph and with no
// points-to set (since it can't point to anything either, being that it
// receives no points-to sets and has none).
if (NodeIndex >= FirstRefNode) {
unsigned j = VSSCCRep[FindNode(NodeIndex - FirstRefNode)];
if ((Node2Visited[j] && !GraphNodes[j].PointerEquivLabel)
|| (GraphNodes[j].Direct && !GraphNodes[j].PredEdges
&& GraphNodes[j].PointsTo->empty())){
return;
}
}
// Process all our explicit edges
if (N->PredEdges)
for (SparseBitVector<>::iterator Iter = N->PredEdges->begin();
Iter != N->PredEdges->end();
++Iter) {
unsigned j = VSSCCRep[*Iter];
if (!Node2Visited[j])
HUValNum(j);
// If this edge turned out to be the same as us, or got no pointer
// equivalence label (and thus points to nothing) , just decrement our
// incoming edges and continue.
if (j == NodeIndex || GraphNodes[j].PointerEquivLabel == 0) {
--GraphNodes[j].NumInEdges;
continue;
}
*(N->PointsTo) |= GraphNodes[j].PointsTo;
// If we didn't end up storing this in the hash, and we're done with all
// the edges, we don't need the points-to set anymore.
--GraphNodes[j].NumInEdges;
if (!GraphNodes[j].NumInEdges && !GraphNodes[j].StoredInHash) {
delete GraphNodes[j].PointsTo;
GraphNodes[j].PointsTo = NULL;
}
}
// If this isn't a direct node, generate a fresh variable.
if (!N->Direct) {
N->PointsTo->set(FirstRefNode + NodeIndex);
}
// See If we have something equivalent to us, if not, generate a new
// equivalence class.
if (N->PointsTo->empty()) {
delete N->PointsTo;
N->PointsTo = NULL;
} else {
if (N->Direct) {
N->PointerEquivLabel = Set2PEClass[N->PointsTo];
if (N->PointerEquivLabel == 0) {
unsigned EquivClass = PEClass++;
N->StoredInHash = true;
Set2PEClass[N->PointsTo] = EquivClass;
N->PointerEquivLabel = EquivClass;
}
} else {
N->PointerEquivLabel = PEClass++;
}
}
}
/// Rewrite our list of constraints so that pointer equivalent nodes are
/// replaced by their the pointer equivalence class representative.
void Andersens::RewriteConstraints() {
std::vector<Constraint> NewConstraints;
DenseSet<Constraint, ConstraintKeyInfo> Seen;
PEClass2Node.clear();
PENLEClass2Node.clear();
// We may have from 1 to Graphnodes + 1 equivalence classes.
PEClass2Node.insert(PEClass2Node.begin(), GraphNodes.size() + 1, -1);
PENLEClass2Node.insert(PENLEClass2Node.begin(), GraphNodes.size() + 1, -1);
// Rewrite constraints, ignoring non-pointer constraints, uniting equivalent
// nodes, and rewriting constraints to use the representative nodes.
for (unsigned i = 0, e = Constraints.size(); i != e; ++i) {
Constraint &C = Constraints[i];
unsigned RHSNode = FindNode(C.Src);
unsigned LHSNode = FindNode(C.Dest);
unsigned RHSLabel = GraphNodes[VSSCCRep[RHSNode]].PointerEquivLabel;
unsigned LHSLabel = GraphNodes[VSSCCRep[LHSNode]].PointerEquivLabel;
// First we try to eliminate constraints for things we can prove don't point
// to anything.
if (LHSLabel == 0) {
DEBUG(PrintNode(&GraphNodes[LHSNode]));
DEBUG(errs() << " is a non-pointer, ignoring constraint.\n");
continue;
}
if (RHSLabel == 0) {
DEBUG(PrintNode(&GraphNodes[RHSNode]));
DEBUG(errs() << " is a non-pointer, ignoring constraint.\n");
continue;
}
// This constraint may be useless, and it may become useless as we translate
// it.
if (C.Src == C.Dest && C.Type == Constraint::Copy)
continue;
C.Src = FindEquivalentNode(RHSNode, RHSLabel);
C.Dest = FindEquivalentNode(FindNode(LHSNode), LHSLabel);
if ((C.Src == C.Dest && C.Type == Constraint::Copy)
|| Seen.count(C))
continue;
Seen.insert(C);
NewConstraints.push_back(C);
}
Constraints.swap(NewConstraints);
PEClass2Node.clear();
}
/// See if we have a node that is pointer equivalent to the one being asked
/// about, and if so, unite them and return the equivalent node. Otherwise,
/// return the original node.
unsigned Andersens::FindEquivalentNode(unsigned NodeIndex,
unsigned NodeLabel) {
if (!GraphNodes[NodeIndex].AddressTaken) {
if (PEClass2Node[NodeLabel] != -1) {
// We found an existing node with the same pointer label, so unify them.
// We specifically request that Union-By-Rank not be used so that
// PEClass2Node[NodeLabel] U= NodeIndex and not the other way around.
return UniteNodes(PEClass2Node[NodeLabel], NodeIndex, false);
} else {
PEClass2Node[NodeLabel] = NodeIndex;
PENLEClass2Node[NodeLabel] = NodeIndex;
}
} else if (PENLEClass2Node[NodeLabel] == -1) {
PENLEClass2Node[NodeLabel] = NodeIndex;
}
return NodeIndex;
}
void Andersens::PrintLabels() const {
for (unsigned i = 0; i < GraphNodes.size(); ++i) {
if (i < FirstRefNode) {
PrintNode(&GraphNodes[i]);
} else if (i < FirstAdrNode) {
DEBUG(errs() << "REF(");
PrintNode(&GraphNodes[i-FirstRefNode]);
DEBUG(errs() <<")");
} else {
DEBUG(errs() << "ADR(");
PrintNode(&GraphNodes[i-FirstAdrNode]);
DEBUG(errs() <<")");
}
DEBUG(errs() << " has pointer label " << GraphNodes[i].PointerEquivLabel
<< " and SCC rep " << VSSCCRep[i]
<< " and is " << (GraphNodes[i].Direct ? "Direct" : "Not direct")
<< "\n");
}
}
/// The technique used here is described in "The Ant and the
/// Grasshopper: Fast and Accurate Pointer Analysis for Millions of
/// Lines of Code. In Programming Language Design and Implementation
/// (PLDI), June 2007." It is known as the "HCD" (Hybrid Cycle
/// Detection) algorithm. It is called a hybrid because it performs an
/// offline analysis and uses its results during the solving (online)
/// phase. This is just the offline portion; the results of this
/// operation are stored in SDT and are later used in SolveContraints()
/// and UniteNodes().
void Andersens::HCD() {
DEBUG(errs() << "Starting HCD.\n");
HCDSCCRep.resize(GraphNodes.size());
for (unsigned i = 0; i < GraphNodes.size(); ++i) {
GraphNodes[i].Edges = new SparseBitVector<>;
HCDSCCRep[i] = i;
}
for (unsigned i = 0, e = Constraints.size(); i != e; ++i) {
Constraint &C = Constraints[i];
assert (C.Src < GraphNodes.size() && C.Dest < GraphNodes.size());
if (C.Type == Constraint::AddressOf) {
continue;
} else if (C.Type == Constraint::Load) {
if( C.Offset == 0 )
GraphNodes[C.Dest].Edges->set(C.Src + FirstRefNode);
} else if (C.Type == Constraint::Store) {
if( C.Offset == 0 )
GraphNodes[C.Dest + FirstRefNode].Edges->set(C.Src);
} else {
GraphNodes[C.Dest].Edges->set(C.Src);
}
}
Node2DFS.insert(Node2DFS.begin(), GraphNodes.size(), 0);
Node2Deleted.insert(Node2Deleted.begin(), GraphNodes.size(), false);
Node2Visited.insert(Node2Visited.begin(), GraphNodes.size(), false);
SDT.insert(SDT.begin(), GraphNodes.size() / 2, -1);
DFSNumber = 0;
for (unsigned i = 0; i < GraphNodes.size(); ++i) {
unsigned Node = HCDSCCRep[i];
if (!Node2Deleted[Node])
Search(Node);
}
for (unsigned i = 0; i < GraphNodes.size(); ++i)
if (GraphNodes[i].Edges != NULL) {
delete GraphNodes[i].Edges;
GraphNodes[i].Edges = NULL;
}
while( !SCCStack.empty() )
SCCStack.pop();
Node2DFS.clear();
Node2Visited.clear();
Node2Deleted.clear();
HCDSCCRep.clear();
DEBUG(errs() << "HCD complete.\n");
}
// Component of HCD:
// Use Nuutila's variant of Tarjan's algorithm to detect
// Strongly-Connected Components (SCCs). For non-trivial SCCs
// containing ref nodes, insert the appropriate information in SDT.
void Andersens::Search(unsigned Node) {
unsigned MyDFS = DFSNumber++;
Node2Visited[Node] = true;
Node2DFS[Node] = MyDFS;
for (SparseBitVector<>::iterator Iter = GraphNodes[Node].Edges->begin(),
End = GraphNodes[Node].Edges->end();
Iter != End;
++Iter) {
unsigned J = HCDSCCRep[*Iter];
assert(GraphNodes[J].isRep() && "Debug check; must be representative");
if (!Node2Deleted[J]) {
if (!Node2Visited[J])
Search(J);
if (Node2DFS[Node] > Node2DFS[J])
Node2DFS[Node] = Node2DFS[J];
}
}
if( MyDFS != Node2DFS[Node] ) {
SCCStack.push(Node);
return;
}
// This node is the root of a SCC, so process it.
//
// If the SCC is "non-trivial" (not a singleton) and contains a reference
// node, we place this SCC into SDT. We unite the nodes in any case.
if (!SCCStack.empty() && Node2DFS[SCCStack.top()] >= MyDFS) {
SparseBitVector<> SCC;
SCC.set(Node);
bool Ref = (Node >= FirstRefNode);
Node2Deleted[Node] = true;
do {
unsigned P = SCCStack.top(); SCCStack.pop();
Ref |= (P >= FirstRefNode);
SCC.set(P);
HCDSCCRep[P] = Node;
} while (!SCCStack.empty() && Node2DFS[SCCStack.top()] >= MyDFS);
if (Ref) {
unsigned Rep = SCC.find_first();
assert(Rep < FirstRefNode && "The SCC didn't have a non-Ref node!");
SparseBitVector<>::iterator i = SCC.begin();
// Skip over the non-ref nodes
while( *i < FirstRefNode )
++i;
while( i != SCC.end() )
SDT[ (*i++) - FirstRefNode ] = Rep;
}
}
}
/// Optimize the constraints by performing offline variable substitution and
/// other optimizations.
void Andersens::OptimizeConstraints() {
DEBUG(errs() << "Beginning constraint optimization\n");
SDTActive = false;
// Function related nodes need to stay in the same relative position and can't
// be location equivalent.
for (std::map<unsigned, unsigned>::iterator Iter = MaxK.begin();
Iter != MaxK.end();
++Iter) {
for (unsigned i = Iter->first;
i != Iter->first + Iter->second;
++i) {
GraphNodes[i].AddressTaken = true;
GraphNodes[i].Direct = false;
}
}
ClumpAddressTaken();
FirstRefNode = GraphNodes.size();
FirstAdrNode = FirstRefNode + GraphNodes.size();
GraphNodes.insert(GraphNodes.end(), 2 * GraphNodes.size(),
Node(false));
VSSCCRep.resize(GraphNodes.size());
for (unsigned i = 0; i < GraphNodes.size(); ++i) {
VSSCCRep[i] = i;
}
HVN();
for (unsigned i = 0; i < GraphNodes.size(); ++i) {
Node *N = &GraphNodes[i];
delete N->PredEdges;
N->PredEdges = NULL;
delete N->ImplicitPredEdges;
N->ImplicitPredEdges = NULL;
}
#undef DEBUG_TYPE
#define DEBUG_TYPE "anders-aa-labels"
DEBUG(PrintLabels());
#undef DEBUG_TYPE
#define DEBUG_TYPE "anders-aa"
RewriteConstraints();
// Delete the adr nodes.
GraphNodes.resize(FirstRefNode * 2);
// Now perform HU
for (unsigned i = 0; i < GraphNodes.size(); ++i) {
Node *N = &GraphNodes[i];
if (FindNode(i) == i) {
N->PointsTo = new SparseBitVector<>;
N->PointedToBy = new SparseBitVector<>;
// Reset our labels
}
VSSCCRep[i] = i;
N->PointerEquivLabel = 0;
}
HU();
#undef DEBUG_TYPE
#define DEBUG_TYPE "anders-aa-labels"
DEBUG(PrintLabels());
#undef DEBUG_TYPE
#define DEBUG_TYPE "anders-aa"
RewriteConstraints();
for (unsigned i = 0; i < GraphNodes.size(); ++i) {
if (FindNode(i) == i) {
Node *N = &GraphNodes[i];
delete N->PointsTo;
N->PointsTo = NULL;
delete N->PredEdges;
N->PredEdges = NULL;
delete N->ImplicitPredEdges;
N->ImplicitPredEdges = NULL;
delete N->PointedToBy;
N->PointedToBy = NULL;
}
}
// perform Hybrid Cycle Detection (HCD)
HCD();
SDTActive = true;
// No longer any need for the upper half of GraphNodes (for ref nodes).
GraphNodes.erase(GraphNodes.begin() + FirstRefNode, GraphNodes.end());
// HCD complete.
DEBUG(errs() << "Finished constraint optimization\n");
FirstRefNode = 0;
FirstAdrNode = 0;
}
/// Unite pointer but not location equivalent variables, now that the constraint
/// graph is built.
void Andersens::UnitePointerEquivalences() {
DEBUG(errs() << "Uniting remaining pointer equivalences\n");
for (unsigned i = 0; i < GraphNodes.size(); ++i) {
if (GraphNodes[i].AddressTaken && GraphNodes[i].isRep()) {
unsigned Label = GraphNodes[i].PointerEquivLabel;
if (Label && PENLEClass2Node[Label] != -1)
UniteNodes(i, PENLEClass2Node[Label]);
}
}
DEBUG(errs() << "Finished remaining pointer equivalences\n");
PENLEClass2Node.clear();
}
/// Create the constraint graph used for solving points-to analysis.
///
void Andersens::CreateConstraintGraph() {
for (unsigned i = 0, e = Constraints.size(); i != e; ++i) {
Constraint &C = Constraints[i];
assert (C.Src < GraphNodes.size() && C.Dest < GraphNodes.size());
if (C.Type == Constraint::AddressOf)
GraphNodes[C.Dest].PointsTo->set(C.Src);
else if (C.Type == Constraint::Load)
GraphNodes[C.Src].Constraints.push_back(C);
else if (C.Type == Constraint::Store)
GraphNodes[C.Dest].Constraints.push_back(C);
else if (C.Offset != 0)
GraphNodes[C.Src].Constraints.push_back(C);
else
GraphNodes[C.Src].Edges->set(C.Dest);
}
}
// Perform DFS and cycle detection.
bool Andersens::QueryNode(unsigned Node) {
assert(GraphNodes[Node].isRep() && "Querying a non-rep node");
unsigned OurDFS = ++DFSNumber;
SparseBitVector<> ToErase;
SparseBitVector<> NewEdges;
Tarjan2DFS[Node] = OurDFS;
// Changed denotes a change from a recursive call that we will bubble up.
// Merged is set if we actually merge a node ourselves.
bool Changed = false, Merged = false;
for (SparseBitVector<>::iterator bi = GraphNodes[Node].Edges->begin();
bi != GraphNodes[Node].Edges->end();
++bi) {
unsigned RepNode = FindNode(*bi);
// If this edge points to a non-representative node but we are
// already planning to add an edge to its representative, we have no
// need for this edge anymore.
if (RepNode != *bi && NewEdges.test(RepNode)){
ToErase.set(*bi);
continue;
}
// Continue about our DFS.
if (!Tarjan2Deleted[RepNode]){
if (Tarjan2DFS[RepNode] == 0) {
Changed |= QueryNode(RepNode);
// May have been changed by QueryNode
RepNode = FindNode(RepNode);
}
if (Tarjan2DFS[RepNode] < Tarjan2DFS[Node])
Tarjan2DFS[Node] = Tarjan2DFS[RepNode];
}
// We may have just discovered that this node is part of a cycle, in
// which case we can also erase it.
if (RepNode != *bi) {
ToErase.set(*bi);
NewEdges.set(RepNode);
}
}
GraphNodes[Node].Edges->intersectWithComplement(ToErase);
GraphNodes[Node].Edges |= NewEdges;
// If this node is a root of a non-trivial SCC, place it on our
// worklist to be processed.
if (OurDFS == Tarjan2DFS[Node]) {
while (!SCCStack.empty() && Tarjan2DFS[SCCStack.top()] >= OurDFS) {
Node = UniteNodes(Node, SCCStack.top());
SCCStack.pop();
Merged = true;
}
Tarjan2Deleted[Node] = true;
if (Merged)
NextWL->insert(&GraphNodes[Node]);
} else {
SCCStack.push(Node);
}
return(Changed | Merged);
}
/// SolveConstraints - This stage iteratively processes the constraints list
/// propagating constraints (adding edges to the Nodes in the points-to graph)
/// until a fixed point is reached.
///
/// We use a variant of the technique called "Lazy Cycle Detection", which is
/// described in "The Ant and the Grasshopper: Fast and Accurate Pointer
/// Analysis for Millions of Lines of Code. In Programming Language Design and
/// Implementation (PLDI), June 2007."
/// The paper describes performing cycle detection one node at a time, which can
/// be expensive if there are no cycles, but there are long chains of nodes that
/// it heuristically believes are cycles (because it will DFS from each node
/// without state from previous nodes).
/// Instead, we use the heuristic to build a worklist of nodes to check, then
/// cycle detect them all at the same time to do this more cheaply. This
/// catches cycles slightly later than the original technique did, but does it
/// make significantly cheaper.
void Andersens::SolveConstraints() {
CurrWL = &w1;
NextWL = &w2;
OptimizeConstraints();
#undef DEBUG_TYPE
#define DEBUG_TYPE "anders-aa-constraints"
DEBUG(PrintConstraints());
#undef DEBUG_TYPE
#define DEBUG_TYPE "anders-aa"
for (unsigned i = 0; i < GraphNodes.size(); ++i) {
Node *N = &GraphNodes[i];
N->PointsTo = new SparseBitVector<>;
N->OldPointsTo = new SparseBitVector<>;
N->Edges = new SparseBitVector<>;
}
CreateConstraintGraph();
UnitePointerEquivalences();
assert(SCCStack.empty() && "SCC Stack should be empty by now!");
Node2DFS.clear();
Node2Deleted.clear();
Node2DFS.insert(Node2DFS.begin(), GraphNodes.size(), 0);
Node2Deleted.insert(Node2Deleted.begin(), GraphNodes.size(), false);
DFSNumber = 0;
DenseSet<Constraint, ConstraintKeyInfo> Seen;
DenseSet<std::pair<unsigned,unsigned>, PairKeyInfo> EdgesChecked;
// Order graph and add initial nodes to work list.
for (unsigned i = 0; i < GraphNodes.size(); ++i) {
Node *INode = &GraphNodes[i];
// Add to work list if it's a representative and can contribute to the
// calculation right now.
if (INode->isRep() && !INode->PointsTo->empty()
&& (!INode->Edges->empty() || !INode->Constraints.empty())) {
INode->Stamp();
CurrWL->insert(INode);
}
}
std::queue<unsigned int> TarjanWL;
#if !FULL_UNIVERSAL
// "Rep and special variables" - in order for HCD to maintain conservative
// results when !FULL_UNIVERSAL, we need to treat the special variables in
// the same way that the !FULL_UNIVERSAL tweak does throughout the rest of
// the analysis - it's ok to add edges from the special nodes, but never
// *to* the special nodes.
std::vector<unsigned int> RSV;
#endif
while( !CurrWL->empty() ) {
DEBUG(errs() << "Starting iteration #" << ++NumIters << "\n");
Node* CurrNode;
unsigned CurrNodeIndex;
// Actual cycle checking code. We cycle check all of the lazy cycle
// candidates from the last iteration in one go.
if (!TarjanWL.empty()) {
DFSNumber = 0;
Tarjan2DFS.clear();
Tarjan2Deleted.clear();
while (!TarjanWL.empty()) {
unsigned int ToTarjan = TarjanWL.front();
TarjanWL.pop();
if (!Tarjan2Deleted[ToTarjan]
&& GraphNodes[ToTarjan].isRep()
&& Tarjan2DFS[ToTarjan] == 0)
QueryNode(ToTarjan);
}
}
// Add to work list if it's a representative and can contribute to the
// calculation right now.
while( (CurrNode = CurrWL->pop()) != NULL ) {
CurrNodeIndex = CurrNode - &GraphNodes[0];
CurrNode->Stamp();
// Figure out the changed points to bits
SparseBitVector<> CurrPointsTo;
CurrPointsTo.intersectWithComplement(CurrNode->PointsTo,
CurrNode->OldPointsTo);
if (CurrPointsTo.empty())
continue;
*(CurrNode->OldPointsTo) |= CurrPointsTo;
// Check the offline-computed equivalencies from HCD.
bool SCC = false;
unsigned Rep;
if (SDT[CurrNodeIndex] >= 0) {
SCC = true;
Rep = FindNode(SDT[CurrNodeIndex]);
#if !FULL_UNIVERSAL
RSV.clear();
#endif
for (SparseBitVector<>::iterator bi = CurrPointsTo.begin();
bi != CurrPointsTo.end(); ++bi) {
unsigned Node = FindNode(*bi);
#if !FULL_UNIVERSAL
if (Node < NumberSpecialNodes) {
RSV.push_back(Node);
continue;
}
#endif
Rep = UniteNodes(Rep,Node);
}
#if !FULL_UNIVERSAL
RSV.push_back(Rep);
#endif
NextWL->insert(&GraphNodes[Rep]);
if ( ! CurrNode->isRep() )
continue;
}
Seen.clear();
/* Now process the constraints for this node. */
for (std::list<Constraint>::iterator li = CurrNode->Constraints.begin();
li != CurrNode->Constraints.end(); ) {
li->Src = FindNode(li->Src);
li->Dest = FindNode(li->Dest);
// Delete redundant constraints
if( Seen.count(*li) ) {
std::list<Constraint>::iterator lk = li; li++;
CurrNode->Constraints.erase(lk);
++NumErased;
continue;
}
Seen.insert(*li);
// Src and Dest will be the vars we are going to process.
// This may look a bit ugly, but what it does is allow us to process
// both store and load constraints with the same code.
// Load constraints say that every member of our RHS solution has K
// added to it, and that variable gets an edge to LHS. We also union
// RHS+K's solution into the LHS solution.
// Store constraints say that every member of our LHS solution has K
// added to it, and that variable gets an edge from RHS. We also union
// RHS's solution into the LHS+K solution.
unsigned *Src;
unsigned *Dest;
unsigned K = li->Offset;
unsigned CurrMember;
if (li->Type == Constraint::Load) {
Src = &CurrMember;
Dest = &li->Dest;
} else if (li->Type == Constraint::Store) {
Src = &li->Src;
Dest = &CurrMember;
} else {
// TODO Handle offseted copy constraint
li++;
continue;
}
// See if we can use Hybrid Cycle Detection (that is, check
// if it was a statically detected offline equivalence that
// involves pointers; if so, remove the redundant constraints).
if( SCC && K == 0 ) {
#if FULL_UNIVERSAL
CurrMember = Rep;
if (GraphNodes[*Src].Edges->test_and_set(*Dest))
if (GraphNodes[*Dest].PointsTo |= *(GraphNodes[*Src].PointsTo))
NextWL->insert(&GraphNodes[*Dest]);
#else
for (unsigned i=0; i < RSV.size(); ++i) {
CurrMember = RSV[i];
if (*Dest < NumberSpecialNodes)
continue;
if (GraphNodes[*Src].Edges->test_and_set(*Dest))
if (GraphNodes[*Dest].PointsTo |= *(GraphNodes[*Src].PointsTo))
NextWL->insert(&GraphNodes[*Dest]);
}
#endif
// since all future elements of the points-to set will be
// equivalent to the current ones, the complex constraints
// become redundant.
//
std::list<Constraint>::iterator lk = li; li++;
#if !FULL_UNIVERSAL
// In this case, we can still erase the constraints when the
// elements of the points-to sets are referenced by *Dest,
// but not when they are referenced by *Src (i.e. for a Load
// constraint). This is because if another special variable is
// put into the points-to set later, we still need to add the
// new edge from that special variable.
if( lk->Type != Constraint::Load)
#endif
GraphNodes[CurrNodeIndex].Constraints.erase(lk);
} else {
const SparseBitVector<> &Solution = CurrPointsTo;
for (SparseBitVector<>::iterator bi = Solution.begin();
bi != Solution.end();
++bi) {
CurrMember = *bi;
// Need to increment the member by K since that is where we are
// supposed to copy to/from. Note that in positive weight cycles,
// which occur in address taking of fields, K can go past
// MaxK[CurrMember] elements, even though that is all it could point
// to.
if (K > 0 && K > MaxK[CurrMember])
continue;
else
CurrMember = FindNode(CurrMember + K);
// Add an edge to the graph, so we can just do regular
// bitmap ior next time. It may also let us notice a cycle.
#if !FULL_UNIVERSAL
if (*Dest < NumberSpecialNodes)
continue;
#endif
if (GraphNodes[*Src].Edges->test_and_set(*Dest))
if (GraphNodes[*Dest].PointsTo |= *(GraphNodes[*Src].PointsTo))
NextWL->insert(&GraphNodes[*Dest]);
}
li++;
}
}
SparseBitVector<> NewEdges;
SparseBitVector<> ToErase;
// Now all we have left to do is propagate points-to info along the
// edges, erasing the redundant edges.
for (SparseBitVector<>::iterator bi = CurrNode->Edges->begin();
bi != CurrNode->Edges->end();
++bi) {
unsigned DestVar = *bi;
unsigned Rep = FindNode(DestVar);
// If we ended up with this node as our destination, or we've already
// got an edge for the representative, delete the current edge.
if (Rep == CurrNodeIndex ||
(Rep != DestVar && NewEdges.test(Rep))) {
ToErase.set(DestVar);
continue;
}
std::pair<unsigned,unsigned> edge(CurrNodeIndex,Rep);
// This is where we do lazy cycle detection.
// If this is a cycle candidate (equal points-to sets and this
// particular edge has not been cycle-checked previously), add to the
// list to check for cycles on the next iteration.
if (!EdgesChecked.count(edge) &&
*(GraphNodes[Rep].PointsTo) == *(CurrNode->PointsTo)) {
EdgesChecked.insert(edge);
TarjanWL.push(Rep);
}
// Union the points-to sets into the dest
#if !FULL_UNIVERSAL
if (Rep >= NumberSpecialNodes)
#endif
if (GraphNodes[Rep].PointsTo |= CurrPointsTo) {
NextWL->insert(&GraphNodes[Rep]);
}
// If this edge's destination was collapsed, rewrite the edge.
if (Rep != DestVar) {
ToErase.set(DestVar);
NewEdges.set(Rep);
}
}
CurrNode->Edges->intersectWithComplement(ToErase);
CurrNode->Edges |= NewEdges;
}
// Switch to other work list.
WorkList* t = CurrWL; CurrWL = NextWL; NextWL = t;
}
Node2DFS.clear();
Node2Deleted.clear();
for (unsigned i = 0; i < GraphNodes.size(); ++i) {
Node *N = &GraphNodes[i];
delete N->OldPointsTo;
delete N->Edges;
}
SDTActive = false;
SDT.clear();
}
//===----------------------------------------------------------------------===//
// Union-Find
//===----------------------------------------------------------------------===//
// Unite nodes First and Second, returning the one which is now the
// representative node. First and Second are indexes into GraphNodes
unsigned Andersens::UniteNodes(unsigned First, unsigned Second,
bool UnionByRank) {
assert (First < GraphNodes.size() && Second < GraphNodes.size() &&
"Attempting to merge nodes that don't exist");
Node *FirstNode = &GraphNodes[First];
Node *SecondNode = &GraphNodes[Second];
assert (SecondNode->isRep() && FirstNode->isRep() &&
"Trying to unite two non-representative nodes!");
if (First == Second)
return First;
if (UnionByRank) {
int RankFirst = (int) FirstNode ->NodeRep;
int RankSecond = (int) SecondNode->NodeRep;
// Rank starts at -1 and gets decremented as it increases.
// Translation: higher rank, lower NodeRep value, which is always negative.
if (RankFirst > RankSecond) {
unsigned t = First; First = Second; Second = t;
Node* tp = FirstNode; FirstNode = SecondNode; SecondNode = tp;
} else if (RankFirst == RankSecond) {
FirstNode->NodeRep = (unsigned) (RankFirst - 1);
}
}
SecondNode->NodeRep = First;
#if !FULL_UNIVERSAL
if (First >= NumberSpecialNodes)
#endif
if (FirstNode->PointsTo && SecondNode->PointsTo)
FirstNode->PointsTo |= *(SecondNode->PointsTo);
if (FirstNode->Edges && SecondNode->Edges)
FirstNode->Edges |= *(SecondNode->Edges);
if (!SecondNode->Constraints.empty())
FirstNode->Constraints.splice(FirstNode->Constraints.begin(),
SecondNode->Constraints);
if (FirstNode->OldPointsTo) {
delete FirstNode->OldPointsTo;
FirstNode->OldPointsTo = new SparseBitVector<>;
}
// Destroy interesting parts of the merged-from node.
delete SecondNode->OldPointsTo;
delete SecondNode->Edges;
delete SecondNode->PointsTo;
SecondNode->Edges = NULL;
SecondNode->PointsTo = NULL;
SecondNode->OldPointsTo = NULL;
NumUnified++;
DEBUG(errs() << "Unified Node ");
DEBUG(PrintNode(FirstNode));
DEBUG(errs() << " and Node ");
DEBUG(PrintNode(SecondNode));
DEBUG(errs() << "\n");
if (SDTActive)
if (SDT[Second] >= 0) {
if (SDT[First] < 0)
SDT[First] = SDT[Second];
else {
UniteNodes( FindNode(SDT[First]), FindNode(SDT[Second]) );
First = FindNode(First);
}
}
return First;
}
// Find the index into GraphNodes of the node representing Node, performing
// path compression along the way
unsigned Andersens::FindNode(unsigned NodeIndex) {
assert (NodeIndex < GraphNodes.size()
&& "Attempting to find a node that can't exist");
Node *N = &GraphNodes[NodeIndex];
if (N->isRep())
return NodeIndex;
else
return (N->NodeRep = FindNode(N->NodeRep));
}
// Find the index into GraphNodes of the node representing Node,
// don't perform path compression along the way (for Print)
unsigned Andersens::FindNode(unsigned NodeIndex) const {
assert (NodeIndex < GraphNodes.size()
&& "Attempting to find a node that can't exist");
const Node *N = &GraphNodes[NodeIndex];
if (N->isRep())
return NodeIndex;
else
return FindNode(N->NodeRep);
}
//===----------------------------------------------------------------------===//
// Debugging Output
//===----------------------------------------------------------------------===//
void Andersens::PrintNode(const Node *N) const {
if (N == &GraphNodes[UniversalSet]) {
errs() << "<universal>";
return;
} else if (N == &GraphNodes[NullPtr]) {
errs() << "<nullptr>";
return;
} else if (N == &GraphNodes[NullObject]) {
errs() << "<null>";
return;
}
if (!N->getValue()) {
errs() << "artificial" << (intptr_t) N;
return;
}
assert(N->getValue() != 0 && "Never set node label!");
Value *V = N->getValue();
if (Function *F = dyn_cast<Function>(V)) {
if (isa<PointerType>(F->getFunctionType()->getReturnType()) &&
N == &GraphNodes[getReturnNode(F)]) {
errs() << F->getName() << ":retval";
return;
} else if (F->getFunctionType()->isVarArg() &&
N == &GraphNodes[getVarargNode(F)]) {
errs() << F->getName() << ":vararg";
return;
}
}
if (Instruction *I = dyn_cast<Instruction>(V))
errs() << I->getParent()->getParent()->getName() << ":";
else if (Argument *Arg = dyn_cast<Argument>(V))
errs() << Arg->getParent()->getName() << ":";
if (V->hasName())
errs() << V->getName();
else
errs() << "(unnamed)";
if (isa<GlobalValue>(V) || isa<AllocaInst>(V) || isMalloc(V))
if (N == &GraphNodes[getObject(V)])
errs() << "<mem>";
}
void Andersens::PrintConstraint(const Constraint &C) const {
if (C.Type == Constraint::Store) {
errs() << "*";
if (C.Offset != 0)
errs() << "(";
}
PrintNode(&GraphNodes[C.Dest]);
if (C.Type == Constraint::Store && C.Offset != 0)
errs() << " + " << C.Offset << ")";
errs() << " = ";
if (C.Type == Constraint::Load) {
errs() << "*";
if (C.Offset != 0)
errs() << "(";
}
else if (C.Type == Constraint::AddressOf)
errs() << "&";
PrintNode(&GraphNodes[C.Src]);
if (C.Offset != 0 && C.Type != Constraint::Store)
errs() << " + " << C.Offset;
if (C.Type == Constraint::Load && C.Offset != 0)
errs() << ")";
errs() << "\n";
}
void Andersens::PrintConstraints() const {
errs() << "Constraints:\n";
for (unsigned i = 0, e = Constraints.size(); i != e; ++i)
PrintConstraint(Constraints[i]);
}
void Andersens::PrintPointsToGraph() const {
errs() << "Points-to graph:\n";
for (unsigned i = 0, e = GraphNodes.size(); i != e; ++i) {
const Node *N = &GraphNodes[i];
if (FindNode(i) != i) {
PrintNode(N);
errs() << "\t--> same as ";
PrintNode(&GraphNodes[FindNode(i)]);
errs() << "\n";
} else {
errs() << "[" << (N->PointsTo->count()) << "] ";
PrintNode(N);
errs() << "\t--> ";
bool first = true;
for (SparseBitVector<>::iterator bi = N->PointsTo->begin();
bi != N->PointsTo->end();
++bi) {
if (!first)
errs() << ", ";
PrintNode(&GraphNodes[*bi]);
first = false;
}
errs() << "\n";
}
}
}