mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-23 19:23:23 +01:00
10839866a1
This patch implements initial backend support for a -mtune CPU controlled by a "tune-cpu" function attribute. If the attribute is not present X86 will use the resolved CPU from target-cpu attribute or command line. This patch adds MC layer support a tune CPU. Each CPU now has two sets of features stored in their GenSubtargetInfo.inc tables . These features lists are passed separately to the Processor and ProcessorModel classes in tablegen. The tune list defaults to an empty list to avoid changes to non-X86. This annoyingly increases the size of static tables on all target as we now store 24 more bytes per CPU. I haven't quantified the overall impact, but I can if we're concerned. One new test is added to X86 to show a few tuning features with mismatched tune-cpu and target-cpu/target-feature attributes to demonstrate independent control. Another new test is added to demonstrate that the scheduler model follows the tune CPU. I have not added a -mtune to llc/opt or MC layer command line yet. With no attributes we'll just use the -mcpu for both. MC layer tools will always follow the normal CPU for tuning. Differential Revision: https://reviews.llvm.org/D85165 |
||
---|---|---|
.. | ||
AsmParser | ||
Disassembler | ||
MCTargetDesc | ||
TargetInfo | ||
CMakeLists.txt | ||
LLVMBuild.txt | ||
README.txt | ||
SystemZ.h | ||
SystemZ.td | ||
SystemZAsmPrinter.cpp | ||
SystemZAsmPrinter.h | ||
SystemZCallingConv.cpp | ||
SystemZCallingConv.h | ||
SystemZCallingConv.td | ||
SystemZConstantPoolValue.cpp | ||
SystemZConstantPoolValue.h | ||
SystemZCopyPhysRegs.cpp | ||
SystemZElimCompare.cpp | ||
SystemZFeatures.td | ||
SystemZFrameLowering.cpp | ||
SystemZFrameLowering.h | ||
SystemZHazardRecognizer.cpp | ||
SystemZHazardRecognizer.h | ||
SystemZInstrBuilder.h | ||
SystemZInstrDFP.td | ||
SystemZInstrFormats.td | ||
SystemZInstrFP.td | ||
SystemZInstrHFP.td | ||
SystemZInstrInfo.cpp | ||
SystemZInstrInfo.h | ||
SystemZInstrInfo.td | ||
SystemZInstrSystem.td | ||
SystemZInstrVector.td | ||
SystemZISelDAGToDAG.cpp | ||
SystemZISelLowering.cpp | ||
SystemZISelLowering.h | ||
SystemZLDCleanup.cpp | ||
SystemZLongBranch.cpp | ||
SystemZMachineFunctionInfo.cpp | ||
SystemZMachineFunctionInfo.h | ||
SystemZMachineScheduler.cpp | ||
SystemZMachineScheduler.h | ||
SystemZMCInstLower.cpp | ||
SystemZMCInstLower.h | ||
SystemZOperands.td | ||
SystemZOperators.td | ||
SystemZPatterns.td | ||
SystemZPostRewrite.cpp | ||
SystemZProcessors.td | ||
SystemZRegisterInfo.cpp | ||
SystemZRegisterInfo.h | ||
SystemZRegisterInfo.td | ||
SystemZSchedule.td | ||
SystemZScheduleZ13.td | ||
SystemZScheduleZ14.td | ||
SystemZScheduleZ15.td | ||
SystemZScheduleZ196.td | ||
SystemZScheduleZEC12.td | ||
SystemZSelectionDAGInfo.cpp | ||
SystemZSelectionDAGInfo.h | ||
SystemZShortenInst.cpp | ||
SystemZSubtarget.cpp | ||
SystemZSubtarget.h | ||
SystemZTargetMachine.cpp | ||
SystemZTargetMachine.h | ||
SystemZTargetTransformInfo.cpp | ||
SystemZTargetTransformInfo.h | ||
SystemZTDC.cpp |
//===---------------------------------------------------------------------===// // Random notes about and ideas for the SystemZ backend. //===---------------------------------------------------------------------===// The initial backend is deliberately restricted to z10. We should add support for later architectures at some point. -- If an inline asm ties an i32 "r" result to an i64 input, the input will be treated as an i32, leaving the upper bits uninitialised. For example: define void @f4(i32 *%dst) { %val = call i32 asm "blah $0", "=r,0" (i64 103) store i32 %val, i32 *%dst ret void } from CodeGen/SystemZ/asm-09.ll will use LHI rather than LGHI. to load 103. This seems to be a general target-independent problem. -- The tuning of the choice between LOAD ADDRESS (LA) and addition in SystemZISelDAGToDAG.cpp is suspect. It should be tweaked based on performance measurements. -- There is no scheduling support. -- We don't use the BRANCH ON INDEX instructions. -- We only use MVC, XC and CLC for constant-length block operations. We could extend them to variable-length operations too, using EXECUTE RELATIVE LONG. MVCIN, MVCLE and CLCLE may be worthwhile too. -- We don't use CUSE or the TRANSLATE family of instructions for string operations. The TRANSLATE ones are probably more difficult to exploit. -- We don't take full advantage of builtins like fabsl because the calling conventions require f128s to be returned by invisible reference. -- ADD LOGICAL WITH SIGNED IMMEDIATE could be useful when we need to produce a carry. SUBTRACT LOGICAL IMMEDIATE could be useful when we need to produce a borrow. (Note that there are no memory forms of ADD LOGICAL WITH CARRY and SUBTRACT LOGICAL WITH BORROW, so the high part of 128-bit memory operations would probably need to be done via a register.) -- We don't use ICM, STCM, or CLM. -- We don't use ADD (LOGICAL) HIGH, SUBTRACT (LOGICAL) HIGH, or COMPARE (LOGICAL) HIGH yet. -- DAGCombiner doesn't yet fold truncations of extended loads. Functions like: unsigned long f (unsigned long x, unsigned short *y) { return (x << 32) | *y; } therefore end up as: sllg %r2, %r2, 32 llgh %r0, 0(%r3) lr %r2, %r0 br %r14 but truncating the load would give: sllg %r2, %r2, 32 lh %r2, 0(%r3) br %r14 -- Functions like: define i64 @f1(i64 %a) { %and = and i64 %a, 1 ret i64 %and } ought to be implemented as: lhi %r0, 1 ngr %r2, %r0 br %r14 but two-address optimizations reverse the order of the AND and force: lhi %r0, 1 ngr %r0, %r2 lgr %r2, %r0 br %r14 CodeGen/SystemZ/and-04.ll has several examples of this. -- Out-of-range displacements are usually handled by loading the full address into a register. In many cases it would be better to create an anchor point instead. E.g. for: define void @f4a(i128 *%aptr, i64 %base) { %addr = add i64 %base, 524288 %bptr = inttoptr i64 %addr to i128 * %a = load volatile i128 *%aptr %b = load i128 *%bptr %add = add i128 %a, %b store i128 %add, i128 *%aptr ret void } (from CodeGen/SystemZ/int-add-08.ll) we load %base+524288 and %base+524296 into separate registers, rather than using %base+524288 as a base for both. -- Dynamic stack allocations round the size to 8 bytes and then allocate that rounded amount. It would be simpler to subtract the unrounded size from the copy of the stack pointer and then align the result. See CodeGen/SystemZ/alloca-01.ll for an example. -- If needed, we can support 16-byte atomics using LPQ, STPQ and CSDG. -- We might want to model all access registers and use them to spill 32-bit values. -- We might want to use the 'overflow' condition of eg. AR to support llvm.sadd.with.overflow.i32 and related instructions - the generated code for signed overflow check is currently quite bad. This would improve the results of using -ftrapv.