mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-23 19:23:23 +01:00
94946e5da5
Wanted to check if inablility to measure latency of CMOV32rm is a regression from D60041 / D60138, but unable to do that because the llvm-exegesis-{8,9} from debian sid fails with that cryptic, unhelpful error. I suspect this will be a better error. llvm-svn: 357900
518 lines
20 KiB
C++
518 lines
20 KiB
C++
//===-- llvm-exegesis.cpp ---------------------------------------*- C++ -*-===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
///
|
|
/// \file
|
|
/// Measures execution properties (latencies/uops) of an instruction.
|
|
///
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "lib/Analysis.h"
|
|
#include "lib/BenchmarkResult.h"
|
|
#include "lib/BenchmarkRunner.h"
|
|
#include "lib/Clustering.h"
|
|
#include "lib/LlvmState.h"
|
|
#include "lib/PerfHelper.h"
|
|
#include "lib/Target.h"
|
|
#include "llvm/ADT/StringExtras.h"
|
|
#include "llvm/ADT/Twine.h"
|
|
#include "llvm/MC/MCInstBuilder.h"
|
|
#include "llvm/MC/MCObjectFileInfo.h"
|
|
#include "llvm/MC/MCParser/MCAsmParser.h"
|
|
#include "llvm/MC/MCParser/MCTargetAsmParser.h"
|
|
#include "llvm/MC/MCRegisterInfo.h"
|
|
#include "llvm/MC/MCStreamer.h"
|
|
#include "llvm/MC/MCSubtargetInfo.h"
|
|
#include "llvm/Object/ObjectFile.h"
|
|
#include "llvm/Support/CommandLine.h"
|
|
#include "llvm/Support/Format.h"
|
|
#include "llvm/Support/Path.h"
|
|
#include "llvm/Support/SourceMgr.h"
|
|
#include "llvm/Support/TargetRegistry.h"
|
|
#include "llvm/Support/TargetSelect.h"
|
|
#include <algorithm>
|
|
#include <string>
|
|
|
|
namespace llvm {
|
|
namespace exegesis {
|
|
|
|
static cl::OptionCategory Options("llvm-exegesis options");
|
|
static cl::OptionCategory BenchmarkOptions("llvm-exegesis benchmark options");
|
|
static cl::OptionCategory AnalysisOptions("llvm-exegesis analysis options");
|
|
|
|
static cl::opt<int> OpcodeIndex("opcode-index",
|
|
cl::desc("opcode to measure, by index"),
|
|
cl::cat(BenchmarkOptions), cl::init(0));
|
|
|
|
static cl::opt<std::string>
|
|
OpcodeNames("opcode-name",
|
|
cl::desc("comma-separated list of opcodes to measure, by name"),
|
|
cl::cat(BenchmarkOptions), cl::init(""));
|
|
|
|
static cl::opt<std::string> SnippetsFile("snippets-file",
|
|
cl::desc("code snippets to measure"),
|
|
cl::cat(BenchmarkOptions),
|
|
cl::init(""));
|
|
|
|
static cl::opt<std::string>
|
|
BenchmarkFile("benchmarks-file",
|
|
cl::desc("File to read (analysis mode) or write "
|
|
"(latency/uops/inverse_throughput modes) benchmark "
|
|
"results. “-” uses stdin/stdout."),
|
|
cl::cat(Options), cl::init(""));
|
|
|
|
static cl::opt<exegesis::InstructionBenchmark::ModeE> BenchmarkMode(
|
|
"mode", cl::desc("the mode to run"), cl::cat(Options),
|
|
cl::values(clEnumValN(exegesis::InstructionBenchmark::Latency, "latency",
|
|
"Instruction Latency"),
|
|
clEnumValN(exegesis::InstructionBenchmark::InverseThroughput,
|
|
"inverse_throughput",
|
|
"Instruction Inverse Throughput"),
|
|
clEnumValN(exegesis::InstructionBenchmark::Uops, "uops",
|
|
"Uop Decomposition"),
|
|
// When not asking for a specific benchmark mode,
|
|
// we'll analyse the results.
|
|
clEnumValN(exegesis::InstructionBenchmark::Unknown, "analysis",
|
|
"Analysis")));
|
|
|
|
static cl::opt<unsigned>
|
|
NumRepetitions("num-repetitions",
|
|
cl::desc("number of time to repeat the asm snippet"),
|
|
cl::cat(BenchmarkOptions), cl::init(10000));
|
|
|
|
static cl::opt<bool> IgnoreInvalidSchedClass(
|
|
"ignore-invalid-sched-class",
|
|
cl::desc("ignore instructions that do not define a sched class"),
|
|
cl::cat(BenchmarkOptions), cl::init(false));
|
|
|
|
static cl::opt<exegesis::InstructionBenchmarkClustering::ModeE>
|
|
AnalysisClusteringAlgorithm(
|
|
"analysis-clustering", cl::desc("the clustering algorithm to use"),
|
|
cl::cat(AnalysisOptions),
|
|
cl::values(clEnumValN(exegesis::InstructionBenchmarkClustering::Dbscan,
|
|
"dbscan", "use DBSCAN/OPTICS algorithm"),
|
|
clEnumValN(exegesis::InstructionBenchmarkClustering::Naive,
|
|
"naive", "one cluster per opcode")),
|
|
cl::init(exegesis::InstructionBenchmarkClustering::Dbscan));
|
|
|
|
static cl::opt<unsigned> AnalysisDbscanNumPoints(
|
|
"analysis-numpoints",
|
|
cl::desc("minimum number of points in an analysis cluster (dbscan only)"),
|
|
cl::cat(AnalysisOptions), cl::init(3));
|
|
|
|
static cl::opt<float> AnalysisClusteringEpsilon(
|
|
"analysis-clustering-epsilon",
|
|
cl::desc("epsilon for benchmark point clustering"),
|
|
cl::cat(AnalysisOptions), cl::init(0.1));
|
|
|
|
static cl::opt<float> AnalysisInconsistencyEpsilon(
|
|
"analysis-inconsistency-epsilon",
|
|
cl::desc("epsilon for detection of when the cluster is different from the "
|
|
"LLVM schedule profile values"),
|
|
cl::cat(AnalysisOptions), cl::init(0.1));
|
|
|
|
static cl::opt<std::string>
|
|
AnalysisClustersOutputFile("analysis-clusters-output-file", cl::desc(""),
|
|
cl::cat(AnalysisOptions), cl::init(""));
|
|
static cl::opt<std::string>
|
|
AnalysisInconsistenciesOutputFile("analysis-inconsistencies-output-file",
|
|
cl::desc(""), cl::cat(AnalysisOptions),
|
|
cl::init(""));
|
|
|
|
static cl::opt<bool> AnalysisDisplayUnstableOpcodes(
|
|
"analysis-display-unstable-clusters",
|
|
cl::desc("if there is more than one benchmark for an opcode, said "
|
|
"benchmarks may end up not being clustered into the same cluster "
|
|
"if the measured performance characteristics are different. by "
|
|
"default all such opcodes are filtered out. this flag will "
|
|
"instead show only such unstable opcodes"),
|
|
cl::cat(AnalysisOptions), cl::init(false));
|
|
|
|
static cl::opt<std::string> CpuName(
|
|
"mcpu",
|
|
cl::desc("cpu name to use for pfm counters, leave empty to autodetect"),
|
|
cl::cat(Options), cl::init(""));
|
|
|
|
static cl::opt<bool>
|
|
DumpObjectToDisk("dump-object-to-disk",
|
|
cl::desc("dumps the generated benchmark object to disk "
|
|
"and prints a message to access it"),
|
|
cl::cat(BenchmarkOptions), cl::init(true));
|
|
|
|
static ExitOnError ExitOnErr;
|
|
|
|
#ifdef LLVM_EXEGESIS_INITIALIZE_NATIVE_TARGET
|
|
void LLVM_EXEGESIS_INITIALIZE_NATIVE_TARGET();
|
|
#endif
|
|
|
|
// Checks that only one of OpcodeNames, OpcodeIndex or SnippetsFile is provided,
|
|
// and returns the opcode indices or {} if snippets should be read from
|
|
// `SnippetsFile`.
|
|
static std::vector<unsigned>
|
|
getOpcodesOrDie(const llvm::MCInstrInfo &MCInstrInfo) {
|
|
const size_t NumSetFlags = (OpcodeNames.empty() ? 0 : 1) +
|
|
(OpcodeIndex == 0 ? 0 : 1) +
|
|
(SnippetsFile.empty() ? 0 : 1);
|
|
if (NumSetFlags != 1)
|
|
llvm::report_fatal_error(
|
|
"please provide one and only one of 'opcode-index', 'opcode-name' or "
|
|
"'snippets-file'");
|
|
if (!SnippetsFile.empty())
|
|
return {};
|
|
if (OpcodeIndex > 0)
|
|
return {static_cast<unsigned>(OpcodeIndex)};
|
|
if (OpcodeIndex < 0) {
|
|
std::vector<unsigned> Result;
|
|
for (unsigned I = 1, E = MCInstrInfo.getNumOpcodes(); I < E; ++I)
|
|
Result.push_back(I);
|
|
return Result;
|
|
}
|
|
// Resolve opcode name -> opcode.
|
|
const auto ResolveName =
|
|
[&MCInstrInfo](llvm::StringRef OpcodeName) -> unsigned {
|
|
for (unsigned I = 1, E = MCInstrInfo.getNumOpcodes(); I < E; ++I)
|
|
if (MCInstrInfo.getName(I) == OpcodeName)
|
|
return I;
|
|
return 0u;
|
|
};
|
|
llvm::SmallVector<llvm::StringRef, 2> Pieces;
|
|
llvm::StringRef(OpcodeNames.getValue())
|
|
.split(Pieces, ",", /* MaxSplit */ -1, /* KeepEmpty */ false);
|
|
std::vector<unsigned> Result;
|
|
for (const llvm::StringRef OpcodeName : Pieces) {
|
|
if (unsigned Opcode = ResolveName(OpcodeName))
|
|
Result.push_back(Opcode);
|
|
else
|
|
llvm::report_fatal_error(
|
|
llvm::Twine("unknown opcode ").concat(OpcodeName));
|
|
}
|
|
return Result;
|
|
}
|
|
|
|
// Generates code snippets for opcode `Opcode`.
|
|
static llvm::Expected<std::vector<BenchmarkCode>>
|
|
generateSnippets(const LLVMState &State, unsigned Opcode) {
|
|
const Instruction &Instr = State.getIC().getInstr(Opcode);
|
|
const llvm::MCInstrDesc &InstrDesc = *Instr.Description;
|
|
// Ignore instructions that we cannot run.
|
|
if (InstrDesc.isPseudo())
|
|
return llvm::make_error<BenchmarkFailure>("Unsupported opcode: isPseudo");
|
|
if (InstrDesc.isBranch() || InstrDesc.isIndirectBranch())
|
|
return llvm::make_error<BenchmarkFailure>(
|
|
"Unsupported opcode: isBranch/isIndirectBranch");
|
|
if (InstrDesc.isCall() || InstrDesc.isReturn())
|
|
return llvm::make_error<BenchmarkFailure>(
|
|
"Unsupported opcode: isCall/isReturn");
|
|
|
|
const std::unique_ptr<SnippetGenerator> Generator =
|
|
State.getExegesisTarget().createSnippetGenerator(BenchmarkMode, State);
|
|
if (!Generator)
|
|
llvm::report_fatal_error("cannot create snippet generator");
|
|
return Generator->generateConfigurations(Instr);
|
|
}
|
|
|
|
namespace {
|
|
|
|
// An MCStreamer that reads a BenchmarkCode definition from a file.
|
|
// The BenchmarkCode definition is just an asm file, with additional comments to
|
|
// specify which registers should be defined or are live on entry.
|
|
class BenchmarkCodeStreamer : public llvm::MCStreamer,
|
|
public llvm::AsmCommentConsumer {
|
|
public:
|
|
explicit BenchmarkCodeStreamer(llvm::MCContext *Context,
|
|
const llvm::MCRegisterInfo *TheRegInfo,
|
|
BenchmarkCode *Result)
|
|
: llvm::MCStreamer(*Context), RegInfo(TheRegInfo), Result(Result) {}
|
|
|
|
// Implementation of the llvm::MCStreamer interface. We only care about
|
|
// instructions.
|
|
void EmitInstruction(const llvm::MCInst &Instruction,
|
|
const llvm::MCSubtargetInfo &STI) override {
|
|
Result->Instructions.push_back(Instruction);
|
|
}
|
|
|
|
// Implementation of the llvm::AsmCommentConsumer.
|
|
void HandleComment(llvm::SMLoc Loc, llvm::StringRef CommentText) override {
|
|
CommentText = CommentText.trim();
|
|
if (!CommentText.consume_front("LLVM-EXEGESIS-"))
|
|
return;
|
|
if (CommentText.consume_front("DEFREG")) {
|
|
// LLVM-EXEGESIS-DEFREF <reg> <hex_value>
|
|
RegisterValue RegVal;
|
|
llvm::SmallVector<llvm::StringRef, 2> Parts;
|
|
CommentText.split(Parts, ' ', /*unlimited splits*/ -1,
|
|
/*do not keep empty strings*/ false);
|
|
if (Parts.size() != 2) {
|
|
llvm::errs() << "invalid comment 'LLVM-EXEGESIS-DEFREG " << CommentText
|
|
<< "\n";
|
|
++InvalidComments;
|
|
}
|
|
if (!(RegVal.Register = findRegisterByName(Parts[0].trim()))) {
|
|
llvm::errs() << "unknown register in 'LLVM-EXEGESIS-DEFREG "
|
|
<< CommentText << "\n";
|
|
++InvalidComments;
|
|
return;
|
|
}
|
|
const llvm::StringRef HexValue = Parts[1].trim();
|
|
RegVal.Value = llvm::APInt(
|
|
/* each hex digit is 4 bits */ HexValue.size() * 4, HexValue, 16);
|
|
Result->RegisterInitialValues.push_back(std::move(RegVal));
|
|
return;
|
|
}
|
|
if (CommentText.consume_front("LIVEIN")) {
|
|
// LLVM-EXEGESIS-LIVEIN <reg>
|
|
if (unsigned Reg = findRegisterByName(CommentText.ltrim()))
|
|
Result->LiveIns.push_back(Reg);
|
|
else {
|
|
llvm::errs() << "unknown register in 'LLVM-EXEGESIS-LIVEIN "
|
|
<< CommentText << "\n";
|
|
++InvalidComments;
|
|
}
|
|
return;
|
|
}
|
|
}
|
|
|
|
unsigned numInvalidComments() const { return InvalidComments; }
|
|
|
|
private:
|
|
// We only care about instructions, we don't implement this part of the API.
|
|
void EmitCommonSymbol(llvm::MCSymbol *Symbol, uint64_t Size,
|
|
unsigned ByteAlignment) override {}
|
|
bool EmitSymbolAttribute(llvm::MCSymbol *Symbol,
|
|
llvm::MCSymbolAttr Attribute) override {
|
|
return false;
|
|
}
|
|
void EmitValueToAlignment(unsigned ByteAlignment, int64_t Value,
|
|
unsigned ValueSize,
|
|
unsigned MaxBytesToEmit) override {}
|
|
void EmitZerofill(llvm::MCSection *Section, llvm::MCSymbol *Symbol,
|
|
uint64_t Size, unsigned ByteAlignment,
|
|
llvm::SMLoc Loc) override {}
|
|
|
|
unsigned findRegisterByName(const llvm::StringRef RegName) const {
|
|
// FIXME: Can we do better than this ?
|
|
for (unsigned I = 0, E = RegInfo->getNumRegs(); I < E; ++I) {
|
|
if (RegName == RegInfo->getName(I))
|
|
return I;
|
|
}
|
|
llvm::errs() << "'" << RegName
|
|
<< "' is not a valid register name for the target\n";
|
|
return 0;
|
|
}
|
|
|
|
const llvm::MCRegisterInfo *const RegInfo;
|
|
BenchmarkCode *const Result;
|
|
unsigned InvalidComments = 0;
|
|
};
|
|
|
|
} // namespace
|
|
|
|
// Reads code snippets from file `Filename`.
|
|
static llvm::Expected<std::vector<BenchmarkCode>>
|
|
readSnippets(const LLVMState &State, llvm::StringRef Filename) {
|
|
llvm::ErrorOr<std::unique_ptr<llvm::MemoryBuffer>> BufferPtr =
|
|
llvm::MemoryBuffer::getFileOrSTDIN(Filename);
|
|
if (std::error_code EC = BufferPtr.getError()) {
|
|
return llvm::make_error<BenchmarkFailure>(
|
|
"cannot read snippet: " + Filename + ": " + EC.message());
|
|
}
|
|
llvm::SourceMgr SM;
|
|
SM.AddNewSourceBuffer(std::move(BufferPtr.get()), llvm::SMLoc());
|
|
|
|
BenchmarkCode Result;
|
|
|
|
llvm::MCObjectFileInfo ObjectFileInfo;
|
|
const llvm::TargetMachine &TM = State.getTargetMachine();
|
|
llvm::MCContext Context(TM.getMCAsmInfo(), TM.getMCRegisterInfo(),
|
|
&ObjectFileInfo);
|
|
ObjectFileInfo.InitMCObjectFileInfo(TM.getTargetTriple(), /*PIC*/ false,
|
|
Context);
|
|
BenchmarkCodeStreamer Streamer(&Context, TM.getMCRegisterInfo(), &Result);
|
|
const std::unique_ptr<llvm::MCAsmParser> AsmParser(
|
|
llvm::createMCAsmParser(SM, Context, Streamer, *TM.getMCAsmInfo()));
|
|
if (!AsmParser)
|
|
return llvm::make_error<BenchmarkFailure>("cannot create asm parser");
|
|
AsmParser->getLexer().setCommentConsumer(&Streamer);
|
|
|
|
const std::unique_ptr<llvm::MCTargetAsmParser> TargetAsmParser(
|
|
TM.getTarget().createMCAsmParser(*TM.getMCSubtargetInfo(), *AsmParser,
|
|
*TM.getMCInstrInfo(),
|
|
llvm::MCTargetOptions()));
|
|
|
|
if (!TargetAsmParser)
|
|
return llvm::make_error<BenchmarkFailure>(
|
|
"cannot create target asm parser");
|
|
AsmParser->setTargetParser(*TargetAsmParser);
|
|
|
|
if (AsmParser->Run(false))
|
|
return llvm::make_error<BenchmarkFailure>("cannot parse asm file");
|
|
if (Streamer.numInvalidComments())
|
|
return llvm::make_error<BenchmarkFailure>(
|
|
llvm::Twine("found ")
|
|
.concat(llvm::Twine(Streamer.numInvalidComments()))
|
|
.concat(" invalid LLVM-EXEGESIS comments"));
|
|
return std::vector<BenchmarkCode>{std::move(Result)};
|
|
}
|
|
|
|
void benchmarkMain() {
|
|
#ifndef HAVE_LIBPFM
|
|
llvm::report_fatal_error(
|
|
"benchmarking unavaliable, LLVM was built without libpfm.");
|
|
#endif
|
|
|
|
if (exegesis::pfm::pfmInitialize())
|
|
llvm::report_fatal_error("cannot initialize libpfm");
|
|
|
|
llvm::InitializeNativeTarget();
|
|
llvm::InitializeNativeTargetAsmPrinter();
|
|
llvm::InitializeNativeTargetAsmParser();
|
|
#ifdef LLVM_EXEGESIS_INITIALIZE_NATIVE_TARGET
|
|
LLVM_EXEGESIS_INITIALIZE_NATIVE_TARGET();
|
|
#endif
|
|
|
|
const LLVMState State(CpuName);
|
|
const auto Opcodes = getOpcodesOrDie(State.getInstrInfo());
|
|
|
|
std::vector<BenchmarkCode> Configurations;
|
|
if (!Opcodes.empty()) {
|
|
for (const unsigned Opcode : Opcodes) {
|
|
// Ignore instructions without a sched class if
|
|
// -ignore-invalid-sched-class is passed.
|
|
if (IgnoreInvalidSchedClass &&
|
|
State.getInstrInfo().get(Opcode).getSchedClass() == 0) {
|
|
llvm::errs() << State.getInstrInfo().getName(Opcode)
|
|
<< ": ignoring instruction without sched class\n";
|
|
continue;
|
|
}
|
|
auto ConfigsForInstr = generateSnippets(State, Opcode);
|
|
if (!ConfigsForInstr) {
|
|
llvm::logAllUnhandledErrors(
|
|
ConfigsForInstr.takeError(), llvm::errs(),
|
|
llvm::Twine(State.getInstrInfo().getName(Opcode)).concat(": "));
|
|
continue;
|
|
}
|
|
std::move(ConfigsForInstr->begin(), ConfigsForInstr->end(),
|
|
std::back_inserter(Configurations));
|
|
}
|
|
} else {
|
|
Configurations = ExitOnErr(readSnippets(State, SnippetsFile));
|
|
}
|
|
|
|
const std::unique_ptr<BenchmarkRunner> Runner =
|
|
State.getExegesisTarget().createBenchmarkRunner(BenchmarkMode, State);
|
|
if (!Runner) {
|
|
llvm::report_fatal_error("cannot create benchmark runner");
|
|
}
|
|
|
|
if (NumRepetitions == 0)
|
|
llvm::report_fatal_error("--num-repetitions must be greater than zero");
|
|
|
|
// Write to standard output if file is not set.
|
|
if (BenchmarkFile.empty())
|
|
BenchmarkFile = "-";
|
|
|
|
for (const BenchmarkCode &Conf : Configurations) {
|
|
InstructionBenchmark Result =
|
|
Runner->runConfiguration(Conf, NumRepetitions, DumpObjectToDisk);
|
|
ExitOnErr(Result.writeYaml(State, BenchmarkFile));
|
|
}
|
|
exegesis::pfm::pfmTerminate();
|
|
}
|
|
|
|
// Prints the results of running analysis pass `Pass` to file `OutputFilename`
|
|
// if OutputFilename is non-empty.
|
|
template <typename Pass>
|
|
static void maybeRunAnalysis(const Analysis &Analyzer, const std::string &Name,
|
|
const std::string &OutputFilename) {
|
|
if (OutputFilename.empty())
|
|
return;
|
|
if (OutputFilename != "-") {
|
|
llvm::errs() << "Printing " << Name << " results to file '"
|
|
<< OutputFilename << "'\n";
|
|
}
|
|
std::error_code ErrorCode;
|
|
llvm::raw_fd_ostream ClustersOS(OutputFilename, ErrorCode,
|
|
llvm::sys::fs::FA_Read |
|
|
llvm::sys::fs::FA_Write);
|
|
if (ErrorCode)
|
|
llvm::report_fatal_error("cannot open out file: " + OutputFilename);
|
|
if (auto Err = Analyzer.run<Pass>(ClustersOS))
|
|
llvm::report_fatal_error(std::move(Err));
|
|
}
|
|
|
|
static void analysisMain() {
|
|
if (BenchmarkFile.empty())
|
|
llvm::report_fatal_error("--benchmarks-file must be set.");
|
|
|
|
if (AnalysisClustersOutputFile.empty() &&
|
|
AnalysisInconsistenciesOutputFile.empty()) {
|
|
llvm::report_fatal_error(
|
|
"At least one of --analysis-clusters-output-file and "
|
|
"--analysis-inconsistencies-output-file must be specified.");
|
|
}
|
|
|
|
llvm::InitializeNativeTarget();
|
|
llvm::InitializeNativeTargetAsmPrinter();
|
|
llvm::InitializeNativeTargetDisassembler();
|
|
// Read benchmarks.
|
|
const LLVMState State("");
|
|
const std::vector<InstructionBenchmark> Points =
|
|
ExitOnErr(InstructionBenchmark::readYamls(State, BenchmarkFile));
|
|
llvm::outs() << "Parsed " << Points.size() << " benchmark points\n";
|
|
if (Points.empty()) {
|
|
llvm::errs() << "no benchmarks to analyze\n";
|
|
return;
|
|
}
|
|
// FIXME: Check that all points have the same triple/cpu.
|
|
// FIXME: Merge points from several runs (latency and uops).
|
|
|
|
std::string Error;
|
|
const auto *TheTarget =
|
|
llvm::TargetRegistry::lookupTarget(Points[0].LLVMTriple, Error);
|
|
if (!TheTarget) {
|
|
llvm::errs() << "unknown target '" << Points[0].LLVMTriple << "'\n";
|
|
return;
|
|
}
|
|
|
|
std::unique_ptr<llvm::MCInstrInfo> InstrInfo(TheTarget->createMCInstrInfo());
|
|
|
|
const auto Clustering = ExitOnErr(InstructionBenchmarkClustering::create(
|
|
Points, AnalysisClusteringAlgorithm, AnalysisDbscanNumPoints,
|
|
AnalysisClusteringEpsilon, InstrInfo->getNumOpcodes()));
|
|
|
|
const Analysis Analyzer(*TheTarget, std::move(InstrInfo), Clustering,
|
|
AnalysisInconsistencyEpsilon,
|
|
AnalysisDisplayUnstableOpcodes);
|
|
|
|
maybeRunAnalysis<Analysis::PrintClusters>(Analyzer, "analysis clusters",
|
|
AnalysisClustersOutputFile);
|
|
maybeRunAnalysis<Analysis::PrintSchedClassInconsistencies>(
|
|
Analyzer, "sched class consistency analysis",
|
|
AnalysisInconsistenciesOutputFile);
|
|
}
|
|
|
|
} // namespace exegesis
|
|
} // namespace llvm
|
|
|
|
int main(int Argc, char **Argv) {
|
|
using namespace llvm;
|
|
cl::ParseCommandLineOptions(Argc, Argv, "");
|
|
|
|
exegesis::ExitOnErr.setExitCodeMapper([](const llvm::Error &Err) {
|
|
if (Err.isA<llvm::StringError>())
|
|
return EXIT_SUCCESS;
|
|
return EXIT_FAILURE;
|
|
});
|
|
|
|
if (exegesis::BenchmarkMode == exegesis::InstructionBenchmark::Unknown) {
|
|
exegesis::analysisMain();
|
|
} else {
|
|
exegesis::benchmarkMain();
|
|
}
|
|
return EXIT_SUCCESS;
|
|
}
|