mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-24 03:33:20 +01:00
b5b6e8e953
This adds support for constrained floating-point comparison intrinsics. Specifically, we add: declare <ty2> @llvm.experimental.constrained.fcmp(<type> <op1>, <type> <op2>, metadata <condition code>, metadata <exception behavior>) declare <ty2> @llvm.experimental.constrained.fcmps(<type> <op1>, <type> <op2>, metadata <condition code>, metadata <exception behavior>) The first variant implements an IEEE "quiet" comparison (i.e. we only get an invalid FP exception if either argument is a SNaN), while the second variant implements an IEEE "signaling" comparison (i.e. we get an invalid FP exception if either argument is any NaN). The condition code is implemented as a metadata string. The same set of predicates as for the fcmp instruction is supported (except for the "true" and "false" predicates). These new intrinsics are mapped by SelectionDAG codegen onto two new ISD opcodes, ISD::STRICT_FSETCC and ISD::STRICT_FSETCCS, again representing quiet vs. signaling comparison operations. Otherwise those nodes look like SETCC nodes, with an additional chain argument and result as usual for strict FP nodes. The patch includes support for the common legalization operations for those nodes. The patch also includes full SystemZ back-end support for the new ISD nodes, mapping them to all available SystemZ instruction to fully implement strict semantics (scalar and vector). Differential Revision: https://reviews.llvm.org/D69281
437 lines
17 KiB
LLVM
437 lines
17 KiB
LLVM
; Test 32-bit floating-point signaling comparison. The tests assume a z10
|
|
; implementation of select, using conditional branches rather than LOCGR.
|
|
;
|
|
; RUN: llc < %s -mtriple=s390x-linux-gnu -mcpu=z10 \
|
|
; RUN: | FileCheck -check-prefix=CHECK -check-prefix=CHECK-SCALAR %s
|
|
; RUN: llc < %s -mtriple=s390x-linux-gnu -mcpu=z14 \
|
|
; RUN: | FileCheck -check-prefix=CHECK -check-prefix=CHECK-VECTOR %s
|
|
|
|
declare float @foo()
|
|
|
|
; Check comparison with registers.
|
|
define i64 @f1(i64 %a, i64 %b, float %f1, float %f2) #0 {
|
|
; CHECK-LABEL: f1:
|
|
; CHECK: kebr %f0, %f2
|
|
; CHECK-SCALAR-NEXT: ber %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrne %r2, %r3
|
|
; CHECK: br %r14
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"oeq",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; Check the low end of the KEB range.
|
|
define i64 @f2(i64 %a, i64 %b, float %f1, float *%ptr) #0 {
|
|
; CHECK-LABEL: f2:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: ber %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrne %r2, %r3
|
|
; CHECK: br %r14
|
|
%f2 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"oeq",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; Check the high end of the aligned KEB range.
|
|
define i64 @f3(i64 %a, i64 %b, float %f1, float *%base) #0 {
|
|
; CHECK-LABEL: f3:
|
|
; CHECK: keb %f0, 4092(%r4)
|
|
; CHECK-SCALAR-NEXT: ber %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrne %r2, %r3
|
|
; CHECK: br %r14
|
|
%ptr = getelementptr float, float *%base, i64 1023
|
|
%f2 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"oeq",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; Check the next word up, which needs separate address logic.
|
|
; Other sequences besides this one would be OK.
|
|
define i64 @f4(i64 %a, i64 %b, float %f1, float *%base) #0 {
|
|
; CHECK-LABEL: f4:
|
|
; CHECK: aghi %r4, 4096
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: ber %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrne %r2, %r3
|
|
; CHECK: br %r14
|
|
%ptr = getelementptr float, float *%base, i64 1024
|
|
%f2 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"oeq",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; Check negative displacements, which also need separate address logic.
|
|
define i64 @f5(i64 %a, i64 %b, float %f1, float *%base) #0 {
|
|
; CHECK-LABEL: f5:
|
|
; CHECK: aghi %r4, -4
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: ber %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrne %r2, %r3
|
|
; CHECK: br %r14
|
|
%ptr = getelementptr float, float *%base, i64 -1
|
|
%f2 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"oeq",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; Check that KEB allows indices.
|
|
define i64 @f6(i64 %a, i64 %b, float %f1, float *%base, i64 %index) #0 {
|
|
; CHECK-LABEL: f6:
|
|
; CHECK: sllg %r1, %r5, 2
|
|
; CHECK: keb %f0, 400(%r1,%r4)
|
|
; CHECK-SCALAR-NEXT: ber %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrne %r2, %r3
|
|
; CHECK: br %r14
|
|
%ptr1 = getelementptr float, float *%base, i64 %index
|
|
%ptr2 = getelementptr float, float *%ptr1, i64 100
|
|
%f2 = load float, float *%ptr2
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"oeq",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; Check that comparisons of spilled values can use KEB rather than KEBR.
|
|
define float @f7(float *%ptr0) #0 {
|
|
; CHECK-LABEL: f7:
|
|
; CHECK: brasl %r14, foo@PLT
|
|
; CHECK-SCALAR: keb {{%f[0-9]+}}, 16{{[04]}}(%r15)
|
|
; CHECK: br %r14
|
|
%ptr1 = getelementptr float, float *%ptr0, i64 2
|
|
%ptr2 = getelementptr float, float *%ptr0, i64 4
|
|
%ptr3 = getelementptr float, float *%ptr0, i64 6
|
|
%ptr4 = getelementptr float, float *%ptr0, i64 8
|
|
%ptr5 = getelementptr float, float *%ptr0, i64 10
|
|
%ptr6 = getelementptr float, float *%ptr0, i64 12
|
|
%ptr7 = getelementptr float, float *%ptr0, i64 14
|
|
%ptr8 = getelementptr float, float *%ptr0, i64 16
|
|
%ptr9 = getelementptr float, float *%ptr0, i64 18
|
|
%ptr10 = getelementptr float, float *%ptr0, i64 20
|
|
|
|
%val0 = load float, float *%ptr0
|
|
%val1 = load float, float *%ptr1
|
|
%val2 = load float, float *%ptr2
|
|
%val3 = load float, float *%ptr3
|
|
%val4 = load float, float *%ptr4
|
|
%val5 = load float, float *%ptr5
|
|
%val6 = load float, float *%ptr6
|
|
%val7 = load float, float *%ptr7
|
|
%val8 = load float, float *%ptr8
|
|
%val9 = load float, float *%ptr9
|
|
%val10 = load float, float *%ptr10
|
|
|
|
%ret = call float @foo() #0
|
|
|
|
%cmp0 = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %ret, float %val0,
|
|
metadata !"olt",
|
|
metadata !"fpexcept.strict") #0
|
|
%cmp1 = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %ret, float %val1,
|
|
metadata !"olt",
|
|
metadata !"fpexcept.strict") #0
|
|
%cmp2 = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %ret, float %val2,
|
|
metadata !"olt",
|
|
metadata !"fpexcept.strict") #0
|
|
%cmp3 = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %ret, float %val3,
|
|
metadata !"olt",
|
|
metadata !"fpexcept.strict") #0
|
|
%cmp4 = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %ret, float %val4,
|
|
metadata !"olt",
|
|
metadata !"fpexcept.strict") #0
|
|
%cmp5 = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %ret, float %val5,
|
|
metadata !"olt",
|
|
metadata !"fpexcept.strict") #0
|
|
%cmp6 = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %ret, float %val6,
|
|
metadata !"olt",
|
|
metadata !"fpexcept.strict") #0
|
|
%cmp7 = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %ret, float %val7,
|
|
metadata !"olt",
|
|
metadata !"fpexcept.strict") #0
|
|
%cmp8 = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %ret, float %val8,
|
|
metadata !"olt",
|
|
metadata !"fpexcept.strict") #0
|
|
%cmp9 = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %ret, float %val9,
|
|
metadata !"olt",
|
|
metadata !"fpexcept.strict") #0
|
|
%cmp10 = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %ret, float %val10,
|
|
metadata !"olt",
|
|
metadata !"fpexcept.strict") #0
|
|
|
|
%sel0 = select i1 %cmp0, float %ret, float 0.0
|
|
%sel1 = select i1 %cmp1, float %sel0, float 1.0
|
|
%sel2 = select i1 %cmp2, float %sel1, float 2.0
|
|
%sel3 = select i1 %cmp3, float %sel2, float 3.0
|
|
%sel4 = select i1 %cmp4, float %sel3, float 4.0
|
|
%sel5 = select i1 %cmp5, float %sel4, float 5.0
|
|
%sel6 = select i1 %cmp6, float %sel5, float 6.0
|
|
%sel7 = select i1 %cmp7, float %sel6, float 7.0
|
|
%sel8 = select i1 %cmp8, float %sel7, float 8.0
|
|
%sel9 = select i1 %cmp9, float %sel8, float 9.0
|
|
%sel10 = select i1 %cmp10, float %sel9, float 10.0
|
|
|
|
ret float %sel10
|
|
}
|
|
|
|
; Check comparison with zero - cannot use LOAD AND TEST.
|
|
define i64 @f8(i64 %a, i64 %b, float %f) #0 {
|
|
; CHECK-LABEL: f8:
|
|
; CHECK: lzer [[REG:%f[0-9]+]]
|
|
; CHECK-NEXT: kebr %f0, [[REG]]
|
|
; CHECK-SCALAR-NEXT: ber %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrne %r2, %r3
|
|
; CHECK: br %r14
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f, float 0.0,
|
|
metadata !"oeq",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; Check the comparison can be reversed if that allows KEB to be used,
|
|
; first with oeq.
|
|
define i64 @f9(i64 %a, i64 %b, float %f2, float *%ptr) #0 {
|
|
; CHECK-LABEL: f9:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: ber %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrne %r2, %r3
|
|
; CHECK: br %r14
|
|
%f1 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"oeq",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; ...then one.
|
|
define i64 @f10(i64 %a, i64 %b, float %f2, float *%ptr) #0 {
|
|
; CHECK-LABEL: f10:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: blhr %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrnlh %r2, %r3
|
|
; CHECK: br %r14
|
|
%f1 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"one",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; ...then olt.
|
|
define i64 @f11(i64 %a, i64 %b, float %f2, float *%ptr) #0 {
|
|
; CHECK-LABEL: f11:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: bhr %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrnh %r2, %r3
|
|
; CHECK: br %r14
|
|
%f1 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"olt",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; ...then ole.
|
|
define i64 @f12(i64 %a, i64 %b, float %f2, float *%ptr) #0 {
|
|
; CHECK-LABEL: f12:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: bher %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrnhe %r2, %r3
|
|
; CHECK: br %r14
|
|
%f1 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"ole",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; ...then oge.
|
|
define i64 @f13(i64 %a, i64 %b, float %f2, float *%ptr) #0 {
|
|
; CHECK-LABEL: f13:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: bler %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrnle %r2, %r3
|
|
; CHECK: br %r14
|
|
%f1 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"oge",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; ...then ogt.
|
|
define i64 @f14(i64 %a, i64 %b, float %f2, float *%ptr) #0 {
|
|
; CHECK-LABEL: f14:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: blr %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrnl %r2, %r3
|
|
; CHECK: br %r14
|
|
%f1 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"ogt",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; ...then ueq.
|
|
define i64 @f15(i64 %a, i64 %b, float %f2, float *%ptr) #0 {
|
|
; CHECK-LABEL: f15:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: bnlhr %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrlh %r2, %r3
|
|
; CHECK: br %r14
|
|
%f1 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"ueq",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; ...then une.
|
|
define i64 @f16(i64 %a, i64 %b, float %f2, float *%ptr) #0 {
|
|
; CHECK-LABEL: f16:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: bner %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgre %r2, %r3
|
|
; CHECK: br %r14
|
|
%f1 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"une",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; ...then ult.
|
|
define i64 @f17(i64 %a, i64 %b, float %f2, float *%ptr) #0 {
|
|
; CHECK-LABEL: f17:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: bnler %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrle %r2, %r3
|
|
; CHECK: br %r14
|
|
%f1 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"ult",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; ...then ule.
|
|
define i64 @f18(i64 %a, i64 %b, float %f2, float *%ptr) #0 {
|
|
; CHECK-LABEL: f18:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: bnlr %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrl %r2, %r3
|
|
; CHECK: br %r14
|
|
%f1 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"ule",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; ...then uge.
|
|
define i64 @f19(i64 %a, i64 %b, float %f2, float *%ptr) #0 {
|
|
; CHECK-LABEL: f19:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: bnhr %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrh %r2, %r3
|
|
; CHECK: br %r14
|
|
%f1 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"uge",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
; ...then ugt.
|
|
define i64 @f20(i64 %a, i64 %b, float %f2, float *%ptr) #0 {
|
|
; CHECK-LABEL: f20:
|
|
; CHECK: keb %f0, 0(%r4)
|
|
; CHECK-SCALAR-NEXT: bnher %r14
|
|
; CHECK-SCALAR: lgr %r2, %r3
|
|
; CHECK-VECTOR-NEXT: locgrhe %r2, %r3
|
|
; CHECK: br %r14
|
|
%f1 = load float, float *%ptr
|
|
%cond = call i1 @llvm.experimental.constrained.fcmps.f32(
|
|
float %f1, float %f2,
|
|
metadata !"ugt",
|
|
metadata !"fpexcept.strict") #0
|
|
%res = select i1 %cond, i64 %a, i64 %b
|
|
ret i64 %res
|
|
}
|
|
|
|
attributes #0 = { strictfp }
|
|
|
|
declare i1 @llvm.experimental.constrained.fcmps.f32(float, float, metadata, metadata)
|