mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-23 11:13:28 +01:00
df37c4f17f
Summary: As noted in documentation, different repetition modes have different trade-offs: > .. option:: -repetition-mode=[duplicate|loop] > > Specify the repetition mode. `duplicate` will create a large, straight line > basic block with `num-repetitions` copies of the snippet. `loop` will wrap > the snippet in a loop which will be run `num-repetitions` times. The `loop` > mode tends to better hide the effects of the CPU frontend on architectures > that cache decoded instructions, but consumes a register for counting > iterations. Indeed. Example: >>! In D74156#1873657, @lebedev.ri wrote: > At least for `CMOV`, i'm seeing wildly different results > | | Latency | RThroughput | > | duplicate | 1 | 0.8 | > | loop | 2 | 0.6 | > where latency=1 seems correct, and i'd expect the througput to be close to 1/2 (since there are two execution units). This isn't great for analysis, at least for schedule model development. As discussed in excruciating detail in >>! In D74156#1924514, @gchatelet wrote: >>>! In D74156#1920632, @lebedev.ri wrote: >> ... did that explanation of the question i'm having made any sense? > > Thx for digging in the conversation ! > Ok it makes more sense now. > > I discussed it a bit with @courbet: > - We want the analysis tool to stay simple so we'd rather not make it knowledgeable of the repetition mode. > - We'd like to still be able to select either repetition mode to dig into special cases > > So we could add a third `min` repetition mode that would run both and take the minimum. It could be the default option. > Would you have some time to look what it would take to add this third mode? there appears to be an agreement that it is indeed sub-par, and that we should provide an optional, measurement (not analysis!) -time way to rectify the situation. However, the solutions isn't entirely straight-forward. We can just add an actual 'multiplexer' `MinSnippetRepetitor`, because if we just concatenate snippets produced by `DuplicateSnippetRepetitor` and `LoopSnippetRepetitor` and run+measure that, the measurement will naturally be different from what we'd get by running+measuring them separately and taking the min. ([[ https://www.wolframalpha.com/input/?i=%28x%2By%29%2F2+%21%3D+min%28x%2C+y%29 | `time(D+L)/2 != min(time(D), time(L))` ]]) Also, it seems best to me to have a single snippet instead of generating a snippet per repetition mode, since the only difference here is that the loop repetition mode reserves one register for loop counter. As far as i can tell, we can either teach `BenchmarkRunner::runConfiguration()` to produce a single report given multiple repetitors (as in the patch), or do that one layer higher - don't modify `BenchmarkRunner::runConfiguration()`, produce multiple reports, don't actually print each one, but aggregate them somehow and only print the final one. Initially i've gone ahead with the latter approach, but it didn't look like a natural fit; the former (as in the diff) does seem like a better fit to me. There's also a question of the test coverage. It sure currently does work here: ``` $ ./bin/llvm-exegesis --opcode-name=CMOV64rr --mode=inverse_throughput --repetition-mode=duplicate Check generated assembly with: /usr/bin/objdump -d /tmp/snippet-8fb949.o --- mode: inverse_throughput key: instructions: - 'CMOV64rr RAX RAX R11 i_0x0' - 'CMOV64rr RBP RBP R15 i_0x0' - 'CMOV64rr RBX RBX RBX i_0x0' - 'CMOV64rr RCX RCX RBX i_0x0' - 'CMOV64rr RDI RDI R10 i_0x0' - 'CMOV64rr RDX RDX RAX i_0x0' - 'CMOV64rr RSI RSI RAX i_0x0' - 'CMOV64rr R8 R8 R8 i_0x0' - 'CMOV64rr R9 R9 RDX i_0x0' - 'CMOV64rr R10 R10 RBX i_0x0' - 'CMOV64rr R11 R11 R14 i_0x0' - 'CMOV64rr R12 R12 R9 i_0x0' - 'CMOV64rr R13 R13 R12 i_0x0' - 'CMOV64rr R14 R14 R15 i_0x0' - 'CMOV64rr R15 R15 R13 i_0x0' config: '' register_initial_values: - 'RAX=0x0' - 'R11=0x0' - 'EFLAGS=0x0' - 'RBP=0x0' - 'R15=0x0' - 'RBX=0x0' - 'RCX=0x0' - 'RDI=0x0' - 'R10=0x0' - 'RDX=0x0' - 'RSI=0x0' - 'R8=0x0' - 'R9=0x0' - 'R14=0x0' - 'R12=0x0' - 'R13=0x0' cpu_name: bdver2 llvm_triple: x86_64-unknown-linux-gnu num_repetitions: 10000 measurements: - { key: inverse_throughput, value: 0.819, per_snippet_value: 12.285 } error: '' info: instruction has tied variables, using static renaming. assembled_snippet: 5541574156415541545348B8000000000000000049BB00000000000000004883EC08C7042400000000C7442404000000009D48BD000000000000000049BF000000000000000048BB000000000000000048B9000000000000000048BF000000000000000049BA000000000000000048BA000000000000000048BE000000000000000049B8000000000000000049B9000000000000000049BE000000000000000049BC000000000000000049BD0000000000000000490F40C3490F40EF480F40DB480F40CB490F40FA480F40D0480F40F04D0F40C04C0F40CA4C0F40D34D0F40DE4D0F40E14D0F40EC4D0F40F74D0F40FD490F40C35B415C415D415E415F5DC3 ... $ ./bin/llvm-exegesis --opcode-name=CMOV64rr --mode=inverse_throughput --repetition-mode=loop Check generated assembly with: /usr/bin/objdump -d /tmp/snippet-051eb3.o --- mode: inverse_throughput key: instructions: - 'CMOV64rr RAX RAX R11 i_0x0' - 'CMOV64rr RBP RBP RSI i_0x0' - 'CMOV64rr RBX RBX R9 i_0x0' - 'CMOV64rr RCX RCX RSI i_0x0' - 'CMOV64rr RDI RDI RBP i_0x0' - 'CMOV64rr RDX RDX R9 i_0x0' - 'CMOV64rr RSI RSI RDI i_0x0' - 'CMOV64rr R9 R9 R12 i_0x0' - 'CMOV64rr R10 R10 R11 i_0x0' - 'CMOV64rr R11 R11 R9 i_0x0' - 'CMOV64rr R12 R12 RBP i_0x0' - 'CMOV64rr R13 R13 RSI i_0x0' - 'CMOV64rr R14 R14 R14 i_0x0' - 'CMOV64rr R15 R15 R10 i_0x0' config: '' register_initial_values: - 'RAX=0x0' - 'R11=0x0' - 'EFLAGS=0x0' - 'RBP=0x0' - 'RSI=0x0' - 'RBX=0x0' - 'R9=0x0' - 'RCX=0x0' - 'RDI=0x0' - 'RDX=0x0' - 'R12=0x0' - 'R10=0x0' - 'R13=0x0' - 'R14=0x0' - 'R15=0x0' cpu_name: bdver2 llvm_triple: x86_64-unknown-linux-gnu num_repetitions: 10000 measurements: - { key: inverse_throughput, value: 0.6083, per_snippet_value: 8.5162 } error: '' info: instruction has tied variables, using static renaming. assembled_snippet: 5541574156415541545348B8000000000000000049BB00000000000000004883EC08C7042400000000C7442404000000009D48BD000000000000000048BE000000000000000048BB000000000000000049B9000000000000000048B9000000000000000048BF000000000000000048BA000000000000000049BC000000000000000049BA000000000000000049BD000000000000000049BE000000000000000049BF000000000000000049B80200000000000000490F40C3480F40EE490F40D9480F40CE480F40FD490F40D1480F40F74D0F40CC4D0F40D34D0F40D94C0F40E54C0F40EE4D0F40F64D0F40FA4983C0FF75C25B415C415D415E415F5DC3 ... $ ./bin/llvm-exegesis --opcode-name=CMOV64rr --mode=inverse_throughput --repetition-mode=min Check generated assembly with: /usr/bin/objdump -d /tmp/snippet-c7a47d.o Check generated assembly with: /usr/bin/objdump -d /tmp/snippet-2581f1.o --- mode: inverse_throughput key: instructions: - 'CMOV64rr RAX RAX R11 i_0x0' - 'CMOV64rr RBP RBP R10 i_0x0' - 'CMOV64rr RBX RBX R10 i_0x0' - 'CMOV64rr RCX RCX RDX i_0x0' - 'CMOV64rr RDI RDI RAX i_0x0' - 'CMOV64rr RDX RDX R9 i_0x0' - 'CMOV64rr RSI RSI RAX i_0x0' - 'CMOV64rr R9 R9 RBX i_0x0' - 'CMOV64rr R10 R10 R12 i_0x0' - 'CMOV64rr R11 R11 RDI i_0x0' - 'CMOV64rr R12 R12 RDI i_0x0' - 'CMOV64rr R13 R13 RDI i_0x0' - 'CMOV64rr R14 R14 R9 i_0x0' - 'CMOV64rr R15 R15 RBP i_0x0' config: '' register_initial_values: - 'RAX=0x0' - 'R11=0x0' - 'EFLAGS=0x0' - 'RBP=0x0' - 'R10=0x0' - 'RBX=0x0' - 'RCX=0x0' - 'RDX=0x0' - 'RDI=0x0' - 'R9=0x0' - 'RSI=0x0' - 'R12=0x0' - 'R13=0x0' - 'R14=0x0' - 'R15=0x0' cpu_name: bdver2 llvm_triple: x86_64-unknown-linux-gnu num_repetitions: 10000 measurements: - { key: inverse_throughput, value: 0.6073, per_snippet_value: 8.5022 } error: '' info: instruction has tied variables, using static renaming. assembled_snippet: 5541574156415541545348B8000000000000000049BB00000000000000004883EC08C7042400000000C7442404000000009D48BD000000000000000049BA000000000000000048BB000000000000000048B9000000000000000048BA000000000000000048BF000000000000000049B9000000000000000048BE000000000000000049BC000000000000000049BD000000000000000049BE000000000000000049BF0000000000000000490F40C3490F40EA490F40DA480F40CA480F40F8490F40D1480F40F04C0F40CB4D0F40D44C0F40DF4C0F40E74C0F40EF4D0F40F14C0F40FD490F40C3490F40EA5B415C415D415E415F5DC35541574156415541545348B8000000000000000049BB00000000000000004883EC08C7042400000000C7442404000000009D48BD000000000000000049BA000000000000000048BB000000000000000048B9000000000000000048BA000000000000000048BF000000000000000049B9000000000000000048BE000000000000000049BC000000000000000049BD000000000000000049BE000000000000000049BF000000000000000049B80200000000000000490F40C3490F40EA490F40DA480F40CA480F40F8490F40D1480F40F04C0F40CB4D0F40D44C0F40DF4C0F40E74C0F40EF4D0F40F14C0F40FD4983C0FF75C25B415C415D415E415F5DC3 ... ``` but i open to suggestions as to how test that. I also have gone with the suggestion to default to this new mode. This was irking me for some time, so i'm happy to finally see progress here. Looking forward to feedback. Reviewers: courbet, gchatelet Reviewed By: courbet, gchatelet Subscribers: mstojanovic, RKSimon, llvm-commits, courbet, gchatelet Tags: #llvm Differential Revision: https://reviews.llvm.org/D76921
446 lines
17 KiB
C++
446 lines
17 KiB
C++
//===-- llvm-exegesis.cpp ---------------------------------------*- C++ -*-===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
///
|
|
/// \file
|
|
/// Measures execution properties (latencies/uops) of an instruction.
|
|
///
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "lib/Analysis.h"
|
|
#include "lib/BenchmarkResult.h"
|
|
#include "lib/BenchmarkRunner.h"
|
|
#include "lib/Clustering.h"
|
|
#include "lib/Error.h"
|
|
#include "lib/LlvmState.h"
|
|
#include "lib/PerfHelper.h"
|
|
#include "lib/SnippetFile.h"
|
|
#include "lib/SnippetRepetitor.h"
|
|
#include "lib/Target.h"
|
|
#include "lib/TargetSelect.h"
|
|
#include "llvm/ADT/StringExtras.h"
|
|
#include "llvm/ADT/Twine.h"
|
|
#include "llvm/MC/MCInstBuilder.h"
|
|
#include "llvm/MC/MCObjectFileInfo.h"
|
|
#include "llvm/MC/MCParser/MCAsmParser.h"
|
|
#include "llvm/MC/MCParser/MCTargetAsmParser.h"
|
|
#include "llvm/MC/MCRegisterInfo.h"
|
|
#include "llvm/MC/MCSubtargetInfo.h"
|
|
#include "llvm/Object/ObjectFile.h"
|
|
#include "llvm/Support/CommandLine.h"
|
|
#include "llvm/Support/Format.h"
|
|
#include "llvm/Support/Path.h"
|
|
#include "llvm/Support/SourceMgr.h"
|
|
#include "llvm/Support/TargetRegistry.h"
|
|
#include "llvm/Support/TargetSelect.h"
|
|
#include <algorithm>
|
|
#include <string>
|
|
|
|
namespace llvm {
|
|
namespace exegesis {
|
|
|
|
static cl::OptionCategory Options("llvm-exegesis options");
|
|
static cl::OptionCategory BenchmarkOptions("llvm-exegesis benchmark options");
|
|
static cl::OptionCategory AnalysisOptions("llvm-exegesis analysis options");
|
|
|
|
static cl::opt<int> OpcodeIndex(
|
|
"opcode-index",
|
|
cl::desc("opcode to measure, by index, or -1 to measure all opcodes"),
|
|
cl::cat(BenchmarkOptions), cl::init(0));
|
|
|
|
static cl::opt<std::string>
|
|
OpcodeNames("opcode-name",
|
|
cl::desc("comma-separated list of opcodes to measure, by name"),
|
|
cl::cat(BenchmarkOptions), cl::init(""));
|
|
|
|
static cl::opt<std::string> SnippetsFile("snippets-file",
|
|
cl::desc("code snippets to measure"),
|
|
cl::cat(BenchmarkOptions),
|
|
cl::init(""));
|
|
|
|
static cl::opt<std::string>
|
|
BenchmarkFile("benchmarks-file",
|
|
cl::desc("File to read (analysis mode) or write "
|
|
"(latency/uops/inverse_throughput modes) benchmark "
|
|
"results. “-” uses stdin/stdout."),
|
|
cl::cat(Options), cl::init(""));
|
|
|
|
static cl::opt<exegesis::InstructionBenchmark::ModeE> BenchmarkMode(
|
|
"mode", cl::desc("the mode to run"), cl::cat(Options),
|
|
cl::values(clEnumValN(exegesis::InstructionBenchmark::Latency, "latency",
|
|
"Instruction Latency"),
|
|
clEnumValN(exegesis::InstructionBenchmark::InverseThroughput,
|
|
"inverse_throughput",
|
|
"Instruction Inverse Throughput"),
|
|
clEnumValN(exegesis::InstructionBenchmark::Uops, "uops",
|
|
"Uop Decomposition"),
|
|
// When not asking for a specific benchmark mode,
|
|
// we'll analyse the results.
|
|
clEnumValN(exegesis::InstructionBenchmark::Unknown, "analysis",
|
|
"Analysis")));
|
|
|
|
static cl::opt<exegesis::InstructionBenchmark::RepetitionModeE> RepetitionMode(
|
|
"repetition-mode", cl::desc("how to repeat the instruction snippet"),
|
|
cl::cat(BenchmarkOptions),
|
|
cl::values(
|
|
clEnumValN(exegesis::InstructionBenchmark::Duplicate, "duplicate",
|
|
"Duplicate the snippet"),
|
|
clEnumValN(exegesis::InstructionBenchmark::Loop, "loop",
|
|
"Loop over the snippet"),
|
|
clEnumValN(exegesis::InstructionBenchmark::AggregateMin, "min",
|
|
"All of the above and take the minimum of measurements")),
|
|
cl::init(exegesis::InstructionBenchmark::Duplicate));
|
|
|
|
static cl::opt<unsigned>
|
|
NumRepetitions("num-repetitions",
|
|
cl::desc("number of time to repeat the asm snippet"),
|
|
cl::cat(BenchmarkOptions), cl::init(10000));
|
|
|
|
static cl::opt<unsigned> MaxConfigsPerOpcode(
|
|
"max-configs-per-opcode",
|
|
cl::desc(
|
|
"allow to snippet generator to generate at most that many configs"),
|
|
cl::cat(BenchmarkOptions), cl::init(1));
|
|
|
|
static cl::opt<bool> IgnoreInvalidSchedClass(
|
|
"ignore-invalid-sched-class",
|
|
cl::desc("ignore instructions that do not define a sched class"),
|
|
cl::cat(BenchmarkOptions), cl::init(false));
|
|
|
|
static cl::opt<exegesis::InstructionBenchmarkClustering::ModeE>
|
|
AnalysisClusteringAlgorithm(
|
|
"analysis-clustering", cl::desc("the clustering algorithm to use"),
|
|
cl::cat(AnalysisOptions),
|
|
cl::values(clEnumValN(exegesis::InstructionBenchmarkClustering::Dbscan,
|
|
"dbscan", "use DBSCAN/OPTICS algorithm"),
|
|
clEnumValN(exegesis::InstructionBenchmarkClustering::Naive,
|
|
"naive", "one cluster per opcode")),
|
|
cl::init(exegesis::InstructionBenchmarkClustering::Dbscan));
|
|
|
|
static cl::opt<unsigned> AnalysisDbscanNumPoints(
|
|
"analysis-numpoints",
|
|
cl::desc("minimum number of points in an analysis cluster (dbscan only)"),
|
|
cl::cat(AnalysisOptions), cl::init(3));
|
|
|
|
static cl::opt<float> AnalysisClusteringEpsilon(
|
|
"analysis-clustering-epsilon",
|
|
cl::desc("epsilon for benchmark point clustering"),
|
|
cl::cat(AnalysisOptions), cl::init(0.1));
|
|
|
|
static cl::opt<float> AnalysisInconsistencyEpsilon(
|
|
"analysis-inconsistency-epsilon",
|
|
cl::desc("epsilon for detection of when the cluster is different from the "
|
|
"LLVM schedule profile values"),
|
|
cl::cat(AnalysisOptions), cl::init(0.1));
|
|
|
|
static cl::opt<std::string>
|
|
AnalysisClustersOutputFile("analysis-clusters-output-file", cl::desc(""),
|
|
cl::cat(AnalysisOptions), cl::init(""));
|
|
static cl::opt<std::string>
|
|
AnalysisInconsistenciesOutputFile("analysis-inconsistencies-output-file",
|
|
cl::desc(""), cl::cat(AnalysisOptions),
|
|
cl::init(""));
|
|
|
|
static cl::opt<bool> AnalysisDisplayUnstableOpcodes(
|
|
"analysis-display-unstable-clusters",
|
|
cl::desc("if there is more than one benchmark for an opcode, said "
|
|
"benchmarks may end up not being clustered into the same cluster "
|
|
"if the measured performance characteristics are different. by "
|
|
"default all such opcodes are filtered out. this flag will "
|
|
"instead show only such unstable opcodes"),
|
|
cl::cat(AnalysisOptions), cl::init(false));
|
|
|
|
static cl::opt<std::string> CpuName(
|
|
"mcpu",
|
|
cl::desc("cpu name to use for pfm counters, leave empty to autodetect"),
|
|
cl::cat(Options), cl::init(""));
|
|
|
|
static cl::opt<bool>
|
|
DumpObjectToDisk("dump-object-to-disk",
|
|
cl::desc("dumps the generated benchmark object to disk "
|
|
"and prints a message to access it"),
|
|
cl::cat(BenchmarkOptions), cl::init(true));
|
|
|
|
static ExitOnError ExitOnErr("llvm-exegesis error: ");
|
|
|
|
// Helper function that logs the error(s) and exits.
|
|
template <typename... ArgTs> static void ExitWithError(ArgTs &&... Args) {
|
|
ExitOnErr(make_error<Failure>(std::forward<ArgTs>(Args)...));
|
|
}
|
|
|
|
// Check Err. If it's in a failure state log the file error(s) and exit.
|
|
static void ExitOnFileError(const Twine &FileName, Error Err) {
|
|
if (Err) {
|
|
ExitOnErr(createFileError(FileName, std::move(Err)));
|
|
}
|
|
}
|
|
|
|
// Check E. If it's in a success state then return the contained value.
|
|
// If it's in a failure state log the file error(s) and exit.
|
|
template <typename T>
|
|
T ExitOnFileError(const Twine &FileName, Expected<T> &&E) {
|
|
ExitOnFileError(FileName, E.takeError());
|
|
return std::move(*E);
|
|
}
|
|
|
|
// Checks that only one of OpcodeNames, OpcodeIndex or SnippetsFile is provided,
|
|
// and returns the opcode indices or {} if snippets should be read from
|
|
// `SnippetsFile`.
|
|
static std::vector<unsigned> getOpcodesOrDie(const MCInstrInfo &MCInstrInfo) {
|
|
const size_t NumSetFlags = (OpcodeNames.empty() ? 0 : 1) +
|
|
(OpcodeIndex == 0 ? 0 : 1) +
|
|
(SnippetsFile.empty() ? 0 : 1);
|
|
if (NumSetFlags != 1) {
|
|
ExitOnErr.setBanner("llvm-exegesis: ");
|
|
ExitWithError("please provide one and only one of 'opcode-index', "
|
|
"'opcode-name' or 'snippets-file'");
|
|
}
|
|
if (!SnippetsFile.empty())
|
|
return {};
|
|
if (OpcodeIndex > 0)
|
|
return {static_cast<unsigned>(OpcodeIndex)};
|
|
if (OpcodeIndex < 0) {
|
|
std::vector<unsigned> Result;
|
|
for (unsigned I = 1, E = MCInstrInfo.getNumOpcodes(); I < E; ++I)
|
|
Result.push_back(I);
|
|
return Result;
|
|
}
|
|
// Resolve opcode name -> opcode.
|
|
const auto ResolveName = [&MCInstrInfo](StringRef OpcodeName) -> unsigned {
|
|
for (unsigned I = 1, E = MCInstrInfo.getNumOpcodes(); I < E; ++I)
|
|
if (MCInstrInfo.getName(I) == OpcodeName)
|
|
return I;
|
|
return 0u;
|
|
};
|
|
SmallVector<StringRef, 2> Pieces;
|
|
StringRef(OpcodeNames.getValue())
|
|
.split(Pieces, ",", /* MaxSplit */ -1, /* KeepEmpty */ false);
|
|
std::vector<unsigned> Result;
|
|
for (const StringRef &OpcodeName : Pieces) {
|
|
if (unsigned Opcode = ResolveName(OpcodeName))
|
|
Result.push_back(Opcode);
|
|
else
|
|
ExitWithError(Twine("unknown opcode ").concat(OpcodeName));
|
|
}
|
|
return Result;
|
|
}
|
|
|
|
// Generates code snippets for opcode `Opcode`.
|
|
static Expected<std::vector<BenchmarkCode>>
|
|
generateSnippets(const LLVMState &State, unsigned Opcode,
|
|
const BitVector &ForbiddenRegs) {
|
|
const Instruction &Instr = State.getIC().getInstr(Opcode);
|
|
const MCInstrDesc &InstrDesc = Instr.Description;
|
|
// Ignore instructions that we cannot run.
|
|
if (InstrDesc.isPseudo())
|
|
return make_error<Failure>("Unsupported opcode: isPseudo");
|
|
if (InstrDesc.isBranch() || InstrDesc.isIndirectBranch())
|
|
return make_error<Failure>("Unsupported opcode: isBranch/isIndirectBranch");
|
|
if (InstrDesc.isCall() || InstrDesc.isReturn())
|
|
return make_error<Failure>("Unsupported opcode: isCall/isReturn");
|
|
|
|
const std::vector<InstructionTemplate> InstructionVariants =
|
|
State.getExegesisTarget().generateInstructionVariants(
|
|
Instr, MaxConfigsPerOpcode);
|
|
|
|
SnippetGenerator::Options SnippetOptions;
|
|
SnippetOptions.MaxConfigsPerOpcode = MaxConfigsPerOpcode;
|
|
const std::unique_ptr<SnippetGenerator> Generator =
|
|
State.getExegesisTarget().createSnippetGenerator(BenchmarkMode, State,
|
|
SnippetOptions);
|
|
if (!Generator)
|
|
ExitWithError("cannot create snippet generator");
|
|
|
|
std::vector<BenchmarkCode> Benchmarks;
|
|
for (const InstructionTemplate &Variant : InstructionVariants) {
|
|
if (Benchmarks.size() >= MaxConfigsPerOpcode)
|
|
break;
|
|
if (auto Err = Generator->generateConfigurations(Variant, Benchmarks,
|
|
ForbiddenRegs))
|
|
return std::move(Err);
|
|
}
|
|
return Benchmarks;
|
|
}
|
|
|
|
void benchmarkMain() {
|
|
#ifndef HAVE_LIBPFM
|
|
ExitWithError("benchmarking unavailable, LLVM was built without libpfm.");
|
|
#endif
|
|
|
|
if (exegesis::pfm::pfmInitialize())
|
|
ExitWithError("cannot initialize libpfm");
|
|
|
|
InitializeNativeTarget();
|
|
InitializeNativeTargetAsmPrinter();
|
|
InitializeNativeTargetAsmParser();
|
|
InitializeNativeExegesisTarget();
|
|
|
|
const LLVMState State(CpuName);
|
|
|
|
const std::unique_ptr<BenchmarkRunner> Runner = ExitOnErr(
|
|
State.getExegesisTarget().createBenchmarkRunner(BenchmarkMode, State));
|
|
if (!Runner) {
|
|
ExitWithError("cannot create benchmark runner");
|
|
}
|
|
|
|
const auto Opcodes = getOpcodesOrDie(State.getInstrInfo());
|
|
|
|
SmallVector<std::unique_ptr<const SnippetRepetitor>, 2> Repetitors;
|
|
if (RepetitionMode != InstructionBenchmark::RepetitionModeE::AggregateMin)
|
|
Repetitors.emplace_back(SnippetRepetitor::Create(RepetitionMode, State));
|
|
else {
|
|
for (InstructionBenchmark::RepetitionModeE RepMode :
|
|
{InstructionBenchmark::RepetitionModeE::Duplicate,
|
|
InstructionBenchmark::RepetitionModeE::Loop})
|
|
Repetitors.emplace_back(SnippetRepetitor::Create(RepMode, State));
|
|
}
|
|
|
|
BitVector AllReservedRegs;
|
|
llvm::for_each(Repetitors,
|
|
[&AllReservedRegs](
|
|
const std::unique_ptr<const SnippetRepetitor> &Repetitor) {
|
|
AllReservedRegs |= Repetitor->getReservedRegs();
|
|
});
|
|
|
|
std::vector<BenchmarkCode> Configurations;
|
|
if (!Opcodes.empty()) {
|
|
for (const unsigned Opcode : Opcodes) {
|
|
// Ignore instructions without a sched class if
|
|
// -ignore-invalid-sched-class is passed.
|
|
if (IgnoreInvalidSchedClass &&
|
|
State.getInstrInfo().get(Opcode).getSchedClass() == 0) {
|
|
errs() << State.getInstrInfo().getName(Opcode)
|
|
<< ": ignoring instruction without sched class\n";
|
|
continue;
|
|
}
|
|
|
|
auto ConfigsForInstr = generateSnippets(State, Opcode, AllReservedRegs);
|
|
if (!ConfigsForInstr) {
|
|
logAllUnhandledErrors(
|
|
ConfigsForInstr.takeError(), errs(),
|
|
Twine(State.getInstrInfo().getName(Opcode)).concat(": "));
|
|
continue;
|
|
}
|
|
std::move(ConfigsForInstr->begin(), ConfigsForInstr->end(),
|
|
std::back_inserter(Configurations));
|
|
}
|
|
} else {
|
|
Configurations = ExitOnErr(readSnippets(State, SnippetsFile));
|
|
}
|
|
|
|
if (NumRepetitions == 0) {
|
|
ExitOnErr.setBanner("llvm-exegesis: ");
|
|
ExitWithError("--num-repetitions must be greater than zero");
|
|
}
|
|
|
|
// Write to standard output if file is not set.
|
|
if (BenchmarkFile.empty())
|
|
BenchmarkFile = "-";
|
|
|
|
for (const BenchmarkCode &Conf : Configurations) {
|
|
InstructionBenchmark Result = ExitOnErr(Runner->runConfiguration(
|
|
Conf, NumRepetitions, Repetitors, DumpObjectToDisk));
|
|
ExitOnFileError(BenchmarkFile, Result.writeYaml(State, BenchmarkFile));
|
|
}
|
|
exegesis::pfm::pfmTerminate();
|
|
}
|
|
|
|
// Prints the results of running analysis pass `Pass` to file `OutputFilename`
|
|
// if OutputFilename is non-empty.
|
|
template <typename Pass>
|
|
static void maybeRunAnalysis(const Analysis &Analyzer, const std::string &Name,
|
|
const std::string &OutputFilename) {
|
|
if (OutputFilename.empty())
|
|
return;
|
|
if (OutputFilename != "-") {
|
|
errs() << "Printing " << Name << " results to file '" << OutputFilename
|
|
<< "'\n";
|
|
}
|
|
std::error_code ErrorCode;
|
|
raw_fd_ostream ClustersOS(OutputFilename, ErrorCode,
|
|
sys::fs::FA_Read | sys::fs::FA_Write);
|
|
if (ErrorCode)
|
|
ExitOnFileError(OutputFilename, errorCodeToError(ErrorCode));
|
|
if (auto Err = Analyzer.run<Pass>(ClustersOS))
|
|
ExitOnFileError(OutputFilename, std::move(Err));
|
|
}
|
|
|
|
static void analysisMain() {
|
|
ExitOnErr.setBanner("llvm-exegesis: ");
|
|
if (BenchmarkFile.empty())
|
|
ExitWithError("--benchmarks-file must be set");
|
|
|
|
if (AnalysisClustersOutputFile.empty() &&
|
|
AnalysisInconsistenciesOutputFile.empty()) {
|
|
ExitWithError(
|
|
"for --mode=analysis: At least one of --analysis-clusters-output-file"
|
|
"and --analysis-inconsistencies-output-file must be specified");
|
|
}
|
|
|
|
InitializeNativeTarget();
|
|
InitializeNativeTargetAsmPrinter();
|
|
InitializeNativeTargetDisassembler();
|
|
|
|
// Read benchmarks.
|
|
const LLVMState State("");
|
|
const std::vector<InstructionBenchmark> Points = ExitOnFileError(
|
|
BenchmarkFile, InstructionBenchmark::readYamls(State, BenchmarkFile));
|
|
|
|
outs() << "Parsed " << Points.size() << " benchmark points\n";
|
|
if (Points.empty()) {
|
|
errs() << "no benchmarks to analyze\n";
|
|
return;
|
|
}
|
|
// FIXME: Check that all points have the same triple/cpu.
|
|
// FIXME: Merge points from several runs (latency and uops).
|
|
|
|
std::string Error;
|
|
const auto *TheTarget =
|
|
TargetRegistry::lookupTarget(Points[0].LLVMTriple, Error);
|
|
if (!TheTarget) {
|
|
errs() << "unknown target '" << Points[0].LLVMTriple << "'\n";
|
|
return;
|
|
}
|
|
|
|
std::unique_ptr<MCInstrInfo> InstrInfo(TheTarget->createMCInstrInfo());
|
|
|
|
const auto Clustering = ExitOnErr(InstructionBenchmarkClustering::create(
|
|
Points, AnalysisClusteringAlgorithm, AnalysisDbscanNumPoints,
|
|
AnalysisClusteringEpsilon, InstrInfo->getNumOpcodes()));
|
|
|
|
const Analysis Analyzer(*TheTarget, std::move(InstrInfo), Clustering,
|
|
AnalysisInconsistencyEpsilon,
|
|
AnalysisDisplayUnstableOpcodes);
|
|
|
|
maybeRunAnalysis<Analysis::PrintClusters>(Analyzer, "analysis clusters",
|
|
AnalysisClustersOutputFile);
|
|
maybeRunAnalysis<Analysis::PrintSchedClassInconsistencies>(
|
|
Analyzer, "sched class consistency analysis",
|
|
AnalysisInconsistenciesOutputFile);
|
|
}
|
|
|
|
} // namespace exegesis
|
|
} // namespace llvm
|
|
|
|
int main(int Argc, char **Argv) {
|
|
using namespace llvm;
|
|
cl::ParseCommandLineOptions(Argc, Argv, "");
|
|
|
|
exegesis::ExitOnErr.setExitCodeMapper([](const Error &Err) {
|
|
if (Err.isA<exegesis::ClusteringError>())
|
|
return EXIT_SUCCESS;
|
|
return EXIT_FAILURE;
|
|
});
|
|
|
|
if (exegesis::BenchmarkMode == exegesis::InstructionBenchmark::Unknown) {
|
|
exegesis::analysisMain();
|
|
} else {
|
|
exegesis::benchmarkMain();
|
|
}
|
|
return EXIT_SUCCESS;
|
|
}
|