1
0
mirror of https://github.com/RPCS3/llvm-mirror.git synced 2024-11-22 02:33:06 +01:00
llvm-mirror/utils/check_ninja_deps.py
Simon Tatham 38ef31c22c [utils] New script check_ninja_deps.py
Summary:
This can be run after a ninja-based build, and analyzes the ninja
build files and dependency database to spot any missing dependencies
in the build scripts.

I wrote it in the course of investigating D82659, and it seems likely
to be useful again.

Reviewers: thakis, chandlerc, theraven

Reviewed By: thakis

Subscribers: riccibruno, llvm-commits

Tags: #llvm

Differential Revision: https://reviews.llvm.org/D83032
2020-07-14 14:54:08 +01:00

192 lines
7.7 KiB
Python
Executable File

#!/usr/bin/env python3
#
# ======- check-ninja-deps - build debugging script ----*- python -*--========#
#
# Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
# See https://llvm.org/LICENSE.txt for license information.
# SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
#
# ==------------------------------------------------------------------------==#
"""Script to find missing formal dependencies in a build.ninja file.
Suppose you have a header file that's autogenerated by (for example) Tablegen.
If a C++ compilation step needs to include that header, then it must be
executed after the Tablegen build step that generates the header. So the
dependency graph in build.ninja should have the Tablegen build step as an
ancestor of the C++ one. If it does not, then there's a latent build-failure
bug, because depending on the order that ninja chooses to schedule its build
steps, the C++ build step could run first, and fail because the header it needs
does not exist yet.
But because that kind of bug can easily be latent or intermittent, you might
not notice, if your local test build happens to succeed. What you'd like is a
way to detect problems of this kind reliably, even if they _didn't_ cause a
failure on your first test.
This script tries to do that. It's specific to the 'ninja' build tool, because
ninja has useful auxiliary output modes that produce the necessary data:
- 'ninja -t graph' emits the full DAG of formal dependencies derived from
build.ninja (in Graphviz format)
- 'ninja -t deps' dumps the database of dependencies discovered at build time
by finding out which headers each source file actually included
By cross-checking these two sources of data against each other, you can find
true dependencies shown by 'deps' that are not reflected as formal dependencies
in 'graph', i.e. a generated header that is required by a given source file but
not forced to be built first.
To run it:
- set up a build directory using ninja as the build tool (cmake -G Ninja)
- in that build directory, run ninja to perform an actual build (populating
the dependency database)
- then, in the same build directory, run this script. No arguments are needed
(but -C and -f are accepted, and propagated to ninja for convenience).
Requirements outside core Python: the 'pygraphviz' module, available via pip or
as the 'python3-pygraphviz' package in Debian and Ubuntu.
"""
import sys
import argparse
import subprocess
import pygraphviz
def toposort(g):
"""Topologically sort a graph.
The input g is a pygraphviz graph object representing a DAG. The function
yields the vertices of g in an arbitrary order consistent with the edges,
so that for any edge v->w, v is output before w."""
# Count the number of immediate predecessors *not yet output* for each
# vertex. Initially this is simply their in-degrees.
ideg = {v: g.in_degree(v) for v in g.nodes_iter()}
# Set of vertices which can be output next, which is true if they have no
# immediate predecessor that has not already been output.
ready = {v for v, d in ideg.items() if d == 0}
# Keep outputting vertices while we have any to output.
while len(ready) > 0:
v = next(iter(ready))
yield v
ready.remove(v)
# Having output v, find each immediate successor w, and decrement its
# 'ideg' value by 1, to indicate that one more of its predecessors has
# now been output.
for w in g.out_neighbors(v):
ideg[w] -= 1
if ideg[w] == 0:
# If that counter reaches zero, w is ready to output.
ready.add(w)
def ancestors(g, translate = lambda x: x):
"""Form the set of ancestors for each vertex of a graph.
The input g is a pygraphviz graph object representing a DAG. The function
yields a sequence of pairs (vertex, set of proper ancestors).
The vertex names are all mapped through 'translate' before output. This
allows us to produce output referring to the label rather than the
identifier of every vertex.
"""
# Store the set of (translated) ancestors for each vertex so far. a[v]
# includes (the translation of) v itself.
a = {}
for v in toposort(g):
vm = translate(v)
# Make up a[v], based on a[predecessors of v].
a[v] = {vm} # include v itself
for w in g.in_neighbors(v):
a[v].update(a[w])
# Remove v itself from the set before yielding it, so that the caller
# doesn't get the trivial dependency of v on itself.
yield vm, a[v].difference({vm})
def main():
parser = argparse.ArgumentParser(
description='Find missing formal dependencies on generated include '
'files in a build.ninja file.')
parser.add_argument("-C", "--build-dir",
help="Build directory (default cwd)")
parser.add_argument("-f", "--build-file",
help="Build directory (default build.ninja)")
args = parser.parse_args()
errs = 0
ninja_prefix = ["ninja"]
if args.build_dir is not None:
ninja_prefix.extend(["-C", args.build_dir])
if args.build_file is not None:
ninja_prefix.extend(["-f", args.build_file])
# Get the formal dependency graph and decode it using pygraphviz.
g = pygraphviz.AGraph(subprocess.check_output(
ninja_prefix + ["-t", "graph"]).decode("UTF-8"))
# Helper function to ask for the label of a vertex, which is where ninja's
# Graphviz output keeps the actual file name of the target.
label = lambda v: g.get_node(v).attr["label"]
# Start by making a list of build targets, i.e. generated files. These are
# just any graph vertex with at least one predecessor.
targets = set(label(v) for v in g.nodes_iter() if g.in_degree(v) > 0)
# Find the set of ancestors of each graph vertex. We pass in 'label' as a
# translation function, so that this gives us the set of ancestor _files_
# for a given _file_ rather than arbitrary numeric vertex ids.
deps = dict(ancestors(g, label))
# Fetch the cached dependency data and check it against our formal ancestry
# data.
currtarget = None
for line in (subprocess.check_output(ninja_prefix + ["-t", "deps"])
.decode("UTF-8").splitlines()):
# ninja -t deps output consists of stanzas of the following form,
# separated by a blank line:
#
# target: [other information we don't need]
# some_file.cpp
# some_header.h
# other_header.h
#
# We parse this ad-hoc by detecting the four leading spaces in a
# source-file line, and the colon in a target line. 'currtarget' stores
# the last target name we saw.
if line.startswith(" "):
dep = line[4:]
assert currtarget is not None, "Source file appeared before target"
# We're only interested in this dependency if it's a *generated*
# file, i.e. it is in our set of targets. Also, we must check that
# currtarget is actually a target we know about: the dependency
# cache is not cleared when build.ninja changes, so it can contain
# stale data from targets that existed only in past builds in the
# same directory.
if (dep in targets and currtarget in deps and
dep not in deps[currtarget]):
print("error:", currtarget, "requires", dep,
"but has no dependency on it", file=sys.stderr)
errs += 1
elif ":" in line:
currtarget = line.split(":", 1)[0]
if errs:
sys.exit("{:d} errors found".format(errs))
if __name__ == '__main__':
main()