mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-24 03:33:20 +01:00
ae65e281f3
to reflect the new license. We understand that people may be surprised that we're moving the header entirely to discuss the new license. We checked this carefully with the Foundation's lawyer and we believe this is the correct approach. Essentially, all code in the project is now made available by the LLVM project under our new license, so you will see that the license headers include that license only. Some of our contributors have contributed code under our old license, and accordingly, we have retained a copy of our old license notice in the top-level files in each project and repository. llvm-svn: 351636
621 lines
24 KiB
Python
Executable File
621 lines
24 KiB
Python
Executable File
#!/usr/bin/python3
|
|
##===- utils/llvmbuild - Build the LLVM project ----------------*-python-*-===##
|
|
#
|
|
# Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
# See https://llvm.org/LICENSE.txt for license information.
|
|
# SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
#
|
|
##===----------------------------------------------------------------------===##
|
|
#
|
|
# This script builds many different flavors of the LLVM ecosystem. It
|
|
# will build LLVM, Clang and dragonegg as well as run tests on them.
|
|
# This script is convenient to use to check builds and tests before
|
|
# committing changes to the upstream repository
|
|
#
|
|
# A typical source setup uses three trees and looks like this:
|
|
#
|
|
# official
|
|
# dragonegg
|
|
# llvm
|
|
# tools
|
|
# clang
|
|
# staging
|
|
# dragonegg
|
|
# llvm
|
|
# tools
|
|
# clang
|
|
# commit
|
|
# dragonegg
|
|
# llvm
|
|
# tools
|
|
# clang
|
|
#
|
|
# In a typical workflow, the "official" tree always contains unchanged
|
|
# sources from the main LLVM project repositories. The "staging" tree
|
|
# is where local work is done. A set of changes resides there waiting
|
|
# to be moved upstream. The "commit" tree is where changes from
|
|
# "staging" make their way upstream. Individual incremental changes
|
|
# from "staging" are applied to "commit" and committed upstream after
|
|
# a successful build and test run. A successful build is one in which
|
|
# testing results in no more failures than seen in the testing of the
|
|
# "official" tree.
|
|
#
|
|
# A build may be invoked as such:
|
|
#
|
|
# llvmbuild --src=~/llvm/commit --src=~/llvm/staging --src=~/llvm/official
|
|
# --build=debug --build=release --build=paranoid
|
|
# --prefix=/home/greened/install --builddir=/home/greened/build
|
|
#
|
|
# This will build the LLVM ecosystem, including LLVM, Clangand
|
|
# dragonegg, putting build results in ~/build and installing tools in
|
|
# ~/install. llvm-compilers-check creates separate build and install
|
|
# directories for each source/build flavor. In the above example,
|
|
# llvmbuild will build debug, release and paranoid (debug+checks)
|
|
# flavors from each source tree (official, staging and commit) for a
|
|
# total of nine builds. All builds will be run in parallel.
|
|
#
|
|
# The user may control parallelism via the --jobs and --threads
|
|
# switches. --jobs tells llvm-compilers-checl the maximum total
|
|
# number of builds to activate in parallel. The user may think of it
|
|
# as equivalent to the GNU make -j switch. --threads tells
|
|
# llvm-compilers-check how many worker threads to use to accomplish
|
|
# those builds. If --threads is less than --jobs, --threads workers
|
|
# will be launched and each one will pick a source/flavor combination
|
|
# to build. Then llvm-compilers-check will invoke GNU make with -j
|
|
# (--jobs / --threads) to use up the remaining job capacity. Once a
|
|
# worker is finished with a build, it will pick another combination
|
|
# off the list and start building it.
|
|
#
|
|
##===----------------------------------------------------------------------===##
|
|
|
|
import optparse
|
|
import os
|
|
import sys
|
|
import threading
|
|
import queue
|
|
import logging
|
|
import traceback
|
|
import subprocess
|
|
import re
|
|
|
|
# TODO: Use shutil.which when it is available (3.2 or later)
|
|
def find_executable(executable, path=None):
|
|
"""Try to find 'executable' in the directories listed in 'path' (a
|
|
string listing directories separated by 'os.pathsep'; defaults to
|
|
os.environ['PATH']). Returns the complete filename or None if not
|
|
found
|
|
"""
|
|
if path is None:
|
|
path = os.environ['PATH']
|
|
paths = path.split(os.pathsep)
|
|
extlist = ['']
|
|
if os.name == 'os2':
|
|
(base, ext) = os.path.splitext(executable)
|
|
# executable files on OS/2 can have an arbitrary extension, but
|
|
# .exe is automatically appended if no dot is present in the name
|
|
if not ext:
|
|
executable = executable + ".exe"
|
|
elif sys.platform == 'win32':
|
|
pathext = os.environ['PATHEXT'].lower().split(os.pathsep)
|
|
(base, ext) = os.path.splitext(executable)
|
|
if ext.lower() not in pathext:
|
|
extlist = pathext
|
|
for ext in extlist:
|
|
execname = executable + ext
|
|
if os.path.isfile(execname):
|
|
return execname
|
|
else:
|
|
for p in paths:
|
|
f = os.path.join(p, execname)
|
|
if os.path.isfile(f):
|
|
return f
|
|
else:
|
|
return None
|
|
|
|
def is_executable(fpath):
|
|
return os.path.exists(fpath) and os.access(fpath, os.X_OK)
|
|
|
|
def add_options(parser):
|
|
parser.add_option("-v", "--verbose", action="store_true",
|
|
default=False,
|
|
help=("Output informational messages"
|
|
" [default: %default]"))
|
|
parser.add_option("--src", action="append",
|
|
help=("Top-level source directory [default: %default]"))
|
|
parser.add_option("--build", action="append",
|
|
help=("Build types to run [default: %default]"))
|
|
parser.add_option("--cc", default=find_executable("cc"),
|
|
help=("The C compiler to use [default: %default]"))
|
|
parser.add_option("--cxx", default=find_executable("c++"),
|
|
help=("The C++ compiler to use [default: %default]"))
|
|
parser.add_option("--threads", default=4, type="int",
|
|
help=("The number of worker threads to use "
|
|
"[default: %default]"))
|
|
parser.add_option("--jobs", "-j", default=8, type="int",
|
|
help=("The number of simultaneous build jobs "
|
|
"[default: %default]"))
|
|
parser.add_option("--prefix",
|
|
help=("Root install directory [default: %default]"))
|
|
parser.add_option("--builddir",
|
|
help=("Root build directory [default: %default]"))
|
|
parser.add_option("--extra-llvm-config-flags", default="",
|
|
help=("Extra flags to pass to llvm configure [default: %default]"))
|
|
parser.add_option("--force-configure", default=False, action="store_true",
|
|
help=("Force reconfigure of all components"))
|
|
parser.add_option("--no-dragonegg", default=False, action="store_true",
|
|
help=("Do not build dragonegg"))
|
|
parser.add_option("--no-install", default=False, action="store_true",
|
|
help=("Do not do installs"))
|
|
parser.add_option("--keep-going", default=False, action="store_true",
|
|
help=("Keep going after failures"))
|
|
parser.add_option("--no-flavor-prefix", default=False, action="store_true",
|
|
help=("Do not append the build flavor to the install path"))
|
|
parser.add_option("--enable-werror", default=False, action="store_true",
|
|
help=("Build with -Werror"))
|
|
return
|
|
|
|
def check_options(parser, options, valid_builds):
|
|
# See if we're building valid flavors.
|
|
for build in options.build:
|
|
if (build not in valid_builds):
|
|
parser.error("'" + build + "' is not a valid build flavor "
|
|
+ str(valid_builds))
|
|
|
|
# See if we can find source directories.
|
|
for src in options.src:
|
|
for component in components:
|
|
component = component.rstrip("2")
|
|
compsrc = src + "/" + component
|
|
if (not os.path.isdir(compsrc)):
|
|
parser.error("'" + compsrc + "' does not exist")
|
|
|
|
# See if we can find the compilers
|
|
options.cc = find_executable(options.cc)
|
|
options.cxx = find_executable(options.cxx)
|
|
|
|
return
|
|
|
|
# Find a unique short name for the given set of paths. This searches
|
|
# back through path components until it finds unique component names
|
|
# among all given paths.
|
|
def get_path_abbrevs(paths):
|
|
# Find the number of common starting characters in the last component
|
|
# of the paths.
|
|
unique_paths = list(paths)
|
|
|
|
class NotFoundException(Exception): pass
|
|
|
|
# Find a unique component of each path.
|
|
unique_bases = unique_paths[:]
|
|
found = 0
|
|
while len(unique_paths) > 0:
|
|
bases = [os.path.basename(src) for src in unique_paths]
|
|
components = { c for c in bases }
|
|
# Account for single entry in paths.
|
|
if len(components) > 1 or len(components) == len(bases):
|
|
# We found something unique.
|
|
for c in components:
|
|
if bases.count(c) == 1:
|
|
index = bases.index(c)
|
|
unique_bases[index] = c
|
|
# Remove the corresponding path from the set under
|
|
# consideration.
|
|
unique_paths[index] = None
|
|
unique_paths = [ p for p in unique_paths if p is not None ]
|
|
unique_paths = [os.path.dirname(src) for src in unique_paths]
|
|
|
|
if len(unique_paths) > 0:
|
|
raise NotFoundException()
|
|
|
|
abbrevs = dict(zip(paths, [base for base in unique_bases]))
|
|
|
|
return abbrevs
|
|
|
|
# Given a set of unique names, find a short character sequence that
|
|
# uniquely identifies them.
|
|
def get_short_abbrevs(unique_bases):
|
|
# Find a unique start character for each path base.
|
|
my_unique_bases = unique_bases[:]
|
|
unique_char_starts = unique_bases[:]
|
|
while len(my_unique_bases) > 0:
|
|
for start, char_tuple in enumerate(zip(*[base
|
|
for base in my_unique_bases])):
|
|
chars = { c for c in char_tuple }
|
|
# Account for single path.
|
|
if len(chars) > 1 or len(chars) == len(char_tuple):
|
|
# We found something unique.
|
|
for c in chars:
|
|
if char_tuple.count(c) == 1:
|
|
index = char_tuple.index(c)
|
|
unique_char_starts[index] = start
|
|
# Remove the corresponding path from the set under
|
|
# consideration.
|
|
my_unique_bases[index] = None
|
|
my_unique_bases = [ b for b in my_unique_bases
|
|
if b is not None ]
|
|
break
|
|
|
|
if len(my_unique_bases) > 0:
|
|
raise NotFoundException()
|
|
|
|
abbrevs = [abbrev[start_index:start_index+3]
|
|
for abbrev, start_index
|
|
in zip([base for base in unique_bases],
|
|
[index for index in unique_char_starts])]
|
|
|
|
abbrevs = dict(zip(unique_bases, abbrevs))
|
|
|
|
return abbrevs
|
|
|
|
class Builder(threading.Thread):
|
|
class ExecutableNotFound(Exception): pass
|
|
class FileNotExecutable(Exception): pass
|
|
|
|
def __init__(self, work_queue, jobs,
|
|
build_abbrev, source_abbrev,
|
|
options):
|
|
super().__init__()
|
|
self.work_queue = work_queue
|
|
self.jobs = jobs
|
|
self.cc = options.cc
|
|
self.cxx = options.cxx
|
|
self.build_abbrev = build_abbrev
|
|
self.source_abbrev = source_abbrev
|
|
self.build_prefix = options.builddir
|
|
self.install_prefix = options.prefix
|
|
self.options = options
|
|
self.component_abbrev = dict(
|
|
llvm="llvm",
|
|
dragonegg="degg")
|
|
def run(self):
|
|
while True:
|
|
try:
|
|
source, build = self.work_queue.get()
|
|
self.dobuild(source, build)
|
|
except:
|
|
traceback.print_exc()
|
|
finally:
|
|
self.work_queue.task_done()
|
|
|
|
def execute(self, command, execdir, env, component):
|
|
prefix = self.component_abbrev[component.replace("-", "_")]
|
|
pwd = os.getcwd()
|
|
if not os.path.exists(execdir):
|
|
os.makedirs(execdir)
|
|
|
|
execenv = os.environ.copy()
|
|
|
|
for key, value in env.items():
|
|
execenv[key] = value
|
|
|
|
self.logger.debug("[" + prefix + "] " + "env " + str(env) + " "
|
|
+ " ".join(command));
|
|
|
|
try:
|
|
proc = subprocess.Popen(command,
|
|
cwd=execdir,
|
|
env=execenv,
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.STDOUT)
|
|
|
|
line = proc.stdout.readline()
|
|
while line:
|
|
self.logger.info("[" + prefix + "] "
|
|
+ str(line, "utf-8").rstrip())
|
|
line = proc.stdout.readline()
|
|
|
|
(stdoutdata, stderrdata) = proc.communicate()
|
|
retcode = proc.wait()
|
|
|
|
return retcode
|
|
|
|
except:
|
|
traceback.print_exc()
|
|
|
|
# Get a list of C++ include directories to pass to clang.
|
|
def get_includes(self):
|
|
# Assume we're building with g++ for now.
|
|
command = [self.cxx]
|
|
command += ["-v", "-x", "c++", "/dev/null", "-fsyntax-only"]
|
|
includes = []
|
|
self.logger.debug(command)
|
|
try:
|
|
proc = subprocess.Popen(command,
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.STDOUT)
|
|
|
|
gather = False
|
|
line = proc.stdout.readline()
|
|
while line:
|
|
self.logger.debug(line)
|
|
if re.search("End of search list", str(line)) is not None:
|
|
self.logger.debug("Stop Gather")
|
|
gather = False
|
|
if gather:
|
|
includes.append(str(line, "utf-8").strip())
|
|
if re.search("#include <...> search starts", str(line)) is not None:
|
|
self.logger.debug("Start Gather")
|
|
gather = True
|
|
line = proc.stdout.readline()
|
|
|
|
except:
|
|
traceback.print_exc()
|
|
self.logger.debug(includes)
|
|
return includes
|
|
|
|
def dobuild(self, source, build):
|
|
build_suffix = ""
|
|
|
|
ssabbrev = get_short_abbrevs([ab for ab in self.source_abbrev.values()])
|
|
|
|
prefix = "[" + ssabbrev[self.source_abbrev[source]] + "-" + self.build_abbrev[build] + "]"
|
|
if (not self.options.no_flavor_prefix):
|
|
self.install_prefix += "/" + self.source_abbrev[source] + "/" + build
|
|
|
|
build_suffix += "/" + self.source_abbrev[source] + "/" + build
|
|
|
|
self.logger = logging.getLogger(prefix)
|
|
|
|
self.logger.debug(self.install_prefix)
|
|
|
|
# Assume we're building with gcc for now.
|
|
cxxincludes = self.get_includes()
|
|
cxxroot = os.path.dirname(cxxincludes[0]) # Remove the version
|
|
cxxroot = os.path.dirname(cxxroot) # Remove the c++
|
|
cxxroot = os.path.dirname(cxxroot) # Remove the include
|
|
|
|
configure_flags = dict(
|
|
llvm=dict(debug=["--prefix=" + self.install_prefix,
|
|
"--enable-assertions",
|
|
"--disable-optimized",
|
|
"--with-gcc-toolchain=" + cxxroot],
|
|
release=["--prefix=" + self.install_prefix,
|
|
"--enable-optimized",
|
|
"--with-gcc-toolchain=" + cxxroot],
|
|
paranoid=["--prefix=" + self.install_prefix,
|
|
"--enable-assertions",
|
|
"--enable-expensive-checks",
|
|
"--disable-optimized",
|
|
"--with-gcc-toolchain=" + cxxroot]),
|
|
dragonegg=dict(debug=[],
|
|
release=[],
|
|
paranoid=[]))
|
|
|
|
if (self.options.enable_werror):
|
|
configure_flags["llvm"]["debug"].append("--enable-werror")
|
|
configure_flags["llvm"]["release"].append("--enable-werror")
|
|
configure_flags["llvm"]["paranoid"].append("--enable-werror")
|
|
|
|
configure_env = dict(
|
|
llvm=dict(debug=dict(CC=self.cc,
|
|
CXX=self.cxx),
|
|
release=dict(CC=self.cc,
|
|
CXX=self.cxx),
|
|
paranoid=dict(CC=self.cc,
|
|
CXX=self.cxx)),
|
|
dragonegg=dict(debug=dict(CC=self.cc,
|
|
CXX=self.cxx),
|
|
release=dict(CC=self.cc,
|
|
CXX=self.cxx),
|
|
paranoid=dict(CC=self.cc,
|
|
CXX=self.cxx)))
|
|
|
|
make_flags = dict(
|
|
llvm=dict(debug=["-j" + str(self.jobs)],
|
|
release=["-j" + str(self.jobs)],
|
|
paranoid=["-j" + str(self.jobs)]),
|
|
dragonegg=dict(debug=["-j" + str(self.jobs)],
|
|
release=["-j" + str(self.jobs)],
|
|
paranoid=["-j" + str(self.jobs)]))
|
|
|
|
make_env = dict(
|
|
llvm=dict(debug=dict(),
|
|
release=dict(),
|
|
paranoid=dict()),
|
|
dragonegg=dict(debug=dict(GCC=self.cc,
|
|
LLVM_CONFIG=self.install_prefix + "/bin/llvm-config"),
|
|
release=dict(GCC=self.cc,
|
|
LLVM_CONFIG=self.install_prefix + "/bin/llvm-config"),
|
|
paranoid=dict(GCC=self.cc,
|
|
LLVM_CONFIG=self.install_prefix + "/bin/llvm-config")))
|
|
|
|
make_install_flags = dict(
|
|
llvm=dict(debug=["install"],
|
|
release=["install"],
|
|
paranoid=["install"]),
|
|
dragonegg=dict(debug=["install"],
|
|
release=["install"],
|
|
paranoid=["install"]))
|
|
|
|
make_install_env = dict(
|
|
llvm=dict(debug=dict(),
|
|
release=dict(),
|
|
paranoid=dict()),
|
|
dragonegg=dict(debug=dict(),
|
|
release=dict(),
|
|
paranoid=dict()))
|
|
|
|
make_check_flags = dict(
|
|
llvm=dict(debug=["check"],
|
|
release=["check"],
|
|
paranoid=["check"]),
|
|
dragonegg=dict(debug=["check"],
|
|
release=["check"],
|
|
paranoid=["check"]))
|
|
|
|
make_check_env = dict(
|
|
llvm=dict(debug=dict(),
|
|
release=dict(),
|
|
paranoid=dict()),
|
|
dragonegg=dict(debug=dict(),
|
|
release=dict(),
|
|
paranoid=dict()))
|
|
|
|
for component in components:
|
|
comp = component[:]
|
|
|
|
if (self.options.no_dragonegg):
|
|
if (comp == 'dragonegg'):
|
|
self.logger.info("Skipping " + component + " in "
|
|
+ builddir)
|
|
continue
|
|
|
|
srcdir = source + "/" + comp.rstrip("2")
|
|
builddir = self.build_prefix + "/" + comp + "/" + build_suffix
|
|
installdir = self.install_prefix
|
|
|
|
comp_key = comp.replace("-", "_")
|
|
|
|
config_args = configure_flags[comp_key][build][:]
|
|
config_args.extend(getattr(self.options,
|
|
"extra_" + comp_key.rstrip("2")
|
|
+ "_config_flags",
|
|
"").split())
|
|
|
|
self.logger.info("Configuring " + component + " in " + builddir)
|
|
configrc = self.configure(component, srcdir, builddir,
|
|
config_args,
|
|
configure_env[comp_key][build])
|
|
|
|
if (configrc == None) :
|
|
self.logger.info("[None] Failed to configure " + component + " in " + installdir)
|
|
|
|
if (configrc == 0 or self.options.keep_going) :
|
|
self.logger.info("Building " + component + " in " + builddir)
|
|
self.logger.info("Build: make " + str(make_flags[comp_key][build]))
|
|
buildrc = self.make(component, srcdir, builddir,
|
|
make_flags[comp_key][build],
|
|
make_env[comp_key][build])
|
|
|
|
if (buildrc == None) :
|
|
self.logger.info("[None] Failed to build " + component + " in " + installdir)
|
|
|
|
if (buildrc == 0 or self.options.keep_going) :
|
|
self.logger.info("Testing " + component + " in " + builddir)
|
|
self.logger.info("Test: make "
|
|
+ str(make_check_flags[comp_key][build]))
|
|
testrc = self.make(component, srcdir, builddir,
|
|
make_check_flags[comp_key][build],
|
|
make_check_env[comp_key][build])
|
|
|
|
if (testrc == None) :
|
|
self.logger.info("[None] Failed to test " + component + " in " + installdir)
|
|
|
|
if ((testrc == 0 or self.options.keep_going)
|
|
and not self.options.no_install):
|
|
self.logger.info("Installing " + component + " in " + installdir)
|
|
self.make(component, srcdir, builddir,
|
|
make_install_flags[comp_key][build],
|
|
make_install_env[comp_key][build])
|
|
else :
|
|
self.logger.info("Failed testing " + component + " in " + installdir)
|
|
|
|
else :
|
|
self.logger.info("Failed to build " + component + " in " + installdir)
|
|
|
|
else :
|
|
self.logger.info("Failed to configure " + component + " in " + installdir)
|
|
|
|
def configure(self, component, srcdir, builddir, flags, env):
|
|
prefix = self.component_abbrev[component.replace("-", "_")]
|
|
|
|
self.logger.debug("Configure " + str(flags) + " " + str(srcdir) + " -> "
|
|
+ str(builddir))
|
|
|
|
configure_files = dict(
|
|
llvm=[(srcdir + "/configure", builddir + "/Makefile")],
|
|
dragonegg=[(None,None)])
|
|
|
|
|
|
doconfig = False
|
|
for conf, mf in configure_files[component.replace("-", "_")]:
|
|
if conf is None:
|
|
# No configure necessary
|
|
return 0
|
|
|
|
if not os.path.exists(conf):
|
|
self.logger.info("[" + prefix + "] Configure failed, no configure script " + conf)
|
|
return -1
|
|
|
|
if os.path.exists(conf) and os.path.exists(mf):
|
|
confstat = os.stat(conf)
|
|
makestat = os.stat(mf)
|
|
if confstat.st_mtime > makestat.st_mtime:
|
|
doconfig = True
|
|
break
|
|
else:
|
|
doconfig = True
|
|
break
|
|
|
|
if not doconfig and not self.options.force_configure:
|
|
return 0
|
|
|
|
program = srcdir + "/configure"
|
|
if not is_executable(program):
|
|
self.logger.info("[" + prefix + "] Configure failed, cannot execute " + program)
|
|
return -1
|
|
|
|
args = [program]
|
|
args += ["--verbose"]
|
|
args += flags
|
|
return self.execute(args, builddir, env, component)
|
|
|
|
def make(self, component, srcdir, builddir, flags, env):
|
|
program = find_executable("make")
|
|
if program is None:
|
|
raise ExecutableNotFound
|
|
|
|
if not is_executable(program):
|
|
raise FileNotExecutable
|
|
|
|
args = [program]
|
|
args += flags
|
|
return self.execute(args, builddir, env, component)
|
|
|
|
# Global constants
|
|
build_abbrev = dict(debug="dbg", release="opt", paranoid="par")
|
|
components = ["llvm", "dragonegg"]
|
|
|
|
# Parse options
|
|
parser = optparse.OptionParser(version="%prog 1.0")
|
|
add_options(parser)
|
|
(options, args) = parser.parse_args()
|
|
check_options(parser, options, build_abbrev.keys());
|
|
|
|
if options.verbose:
|
|
logging.basicConfig(level=logging.DEBUG,
|
|
format='%(name)-13s: %(message)s')
|
|
else:
|
|
logging.basicConfig(level=logging.INFO,
|
|
format='%(name)-13s: %(message)s')
|
|
|
|
source_abbrev = get_path_abbrevs(set(options.src))
|
|
|
|
work_queue = queue.Queue()
|
|
|
|
jobs = options.jobs // options.threads
|
|
if jobs == 0:
|
|
jobs = 1
|
|
|
|
numthreads = options.threads
|
|
|
|
logging.getLogger().info("Building with " + str(options.jobs) + " jobs and "
|
|
+ str(numthreads) + " threads using " + str(jobs)
|
|
+ " make jobs")
|
|
|
|
logging.getLogger().info("CC = " + str(options.cc))
|
|
logging.getLogger().info("CXX = " + str(options.cxx))
|
|
|
|
for t in range(numthreads):
|
|
builder = Builder(work_queue, jobs,
|
|
build_abbrev, source_abbrev,
|
|
options)
|
|
builder.daemon = True
|
|
builder.start()
|
|
|
|
for build in set(options.build):
|
|
for source in set(options.src):
|
|
work_queue.put((source, build))
|
|
|
|
work_queue.join()
|