mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-26 04:32:44 +01:00
1a3ef94576
This patch implements the following changes: 1) SelectionDAGBuilder::visitConstrainedFPIntrinsic currently treats each constrained intrinsic like a global barrier (e.g. a function call) and fully serializes all pending chains. This is actually not required; it is allowed for constrained intrinsics to be reordered w.r.t one another or (nonvolatile) memory accesses. The MI-level scheduler already allows for that flexibility, so it makes sense to allow it at the DAG level as well. This patch therefore changes the way chains for constrained intrisincs are created, and handles them basically like load operations are handled. This has the effect that constrained intrinsics are no longer serialized against one another or (nonvolatile) loads. They are still serialized against stores, but that seems hard to change with the current DAG chain setup, and it also doesn't seem to be a big problem preventing DAG 2) The OPC_CheckFoldableChainNode check requires that each of the intermediate nodes in a multi-node pattern match only has a single use. This check tends to fail if those intermediate nodes are strict operations as those have a chain output that typically indeed has another use. However, we don't really need to consider chains here at all, since they will all be rewritten anyway by UpdateChains later. Other parts of the matcher therefore already ignore chains, but this hasOneUse check doesn't. This patch replaces hasOneUse by a custom test that verifies there is no more than one use of any non-chain output value. In theory, this change could affect code unrelated to strict FP nodes, but at least on SystemZ I could not find any single instance of that happening 3) The SystemZ back-end currently does not allow matching multiply-and- extend operations (32x32 -> 64bit or 64x64 -> 128bit FP multiply) for strict FP operations. This was not possible in the past due to the problems described under 1) and 2) above. With those issues fixed, it is now possible to fully support those instructions in strict mode as well, and this patch does so. Differential Revision: https://reviews.llvm.org/D70913
436 lines
20 KiB
LLVM
436 lines
20 KiB
LLVM
; Test strict multiplication of two f64s, producing an f128 result.
|
|
;
|
|
; RUN: llc < %s -mtriple=s390x-linux-gnu | FileCheck %s
|
|
|
|
declare fp128 @llvm.experimental.constrained.fmul.f128(fp128, fp128, metadata, metadata)
|
|
declare double @llvm.experimental.constrained.fadd.f64(double, double, metadata, metadata)
|
|
declare double @llvm.experimental.constrained.fptrunc.f64.f128(fp128, metadata, metadata)
|
|
declare fp128 @llvm.experimental.constrained.fpext.f128.f64(double, metadata)
|
|
|
|
declare double @foo()
|
|
|
|
; Check register multiplication. "mxdbr %f0, %f2" is not valid from LLVM's
|
|
; point of view, because %f2 is the low register of the FP128 %f0. Pass the
|
|
; multiplier in %f4 instead.
|
|
define void @f1(double %f1, double %dummy, double %f2, fp128 *%dst) #0 {
|
|
; CHECK-LABEL: f1:
|
|
; CHECK: mxdbr %f0, %f4
|
|
; CHECK: std %f0, 0(%r2)
|
|
; CHECK: std %f2, 8(%r2)
|
|
; CHECK: br %r14
|
|
%f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %f1,
|
|
metadata !"fpexcept.strict") #0
|
|
%f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %f2,
|
|
metadata !"fpexcept.strict") #0
|
|
%res = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %f1x, fp128 %f2x,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
store fp128 %res, fp128 *%dst
|
|
ret void
|
|
}
|
|
|
|
; Check the low end of the MXDB range.
|
|
define void @f2(double %f1, double *%ptr, fp128 *%dst) #0 {
|
|
; CHECK-LABEL: f2:
|
|
; CHECK: mxdb %f0, 0(%r2)
|
|
; CHECK: std %f0, 0(%r3)
|
|
; CHECK: std %f2, 8(%r3)
|
|
; CHECK: br %r14
|
|
%f2 = load double, double *%ptr
|
|
%f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %f1,
|
|
metadata !"fpexcept.strict") #0
|
|
%f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %f2,
|
|
metadata !"fpexcept.strict") #0
|
|
%res = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %f1x, fp128 %f2x,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
store fp128 %res, fp128 *%dst
|
|
ret void
|
|
}
|
|
|
|
; Check the high end of the aligned MXDB range.
|
|
define void @f3(double %f1, double *%base, fp128 *%dst) #0 {
|
|
; CHECK-LABEL: f3:
|
|
; CHECK: mxdb %f0, 4088(%r2)
|
|
; CHECK: std %f0, 0(%r3)
|
|
; CHECK: std %f2, 8(%r3)
|
|
; CHECK: br %r14
|
|
%ptr = getelementptr double, double *%base, i64 511
|
|
%f2 = load double, double *%ptr
|
|
%f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %f1,
|
|
metadata !"fpexcept.strict") #0
|
|
%f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %f2,
|
|
metadata !"fpexcept.strict") #0
|
|
%res = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %f1x, fp128 %f2x,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
store fp128 %res, fp128 *%dst
|
|
ret void
|
|
}
|
|
|
|
; Check the next doubleword up, which needs separate address logic.
|
|
; Other sequences besides this one would be OK.
|
|
define void @f4(double %f1, double *%base, fp128 *%dst) #0 {
|
|
; CHECK-LABEL: f4:
|
|
; CHECK: aghi %r2, 4096
|
|
; CHECK: mxdb %f0, 0(%r2)
|
|
; CHECK: std %f0, 0(%r3)
|
|
; CHECK: std %f2, 8(%r3)
|
|
; CHECK: br %r14
|
|
%ptr = getelementptr double, double *%base, i64 512
|
|
%f2 = load double, double *%ptr
|
|
%f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %f1,
|
|
metadata !"fpexcept.strict") #0
|
|
%f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %f2,
|
|
metadata !"fpexcept.strict") #0
|
|
%res = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %f1x, fp128 %f2x,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
store fp128 %res, fp128 *%dst
|
|
ret void
|
|
}
|
|
|
|
; Check negative displacements, which also need separate address logic.
|
|
define void @f5(double %f1, double *%base, fp128 *%dst) #0 {
|
|
; CHECK-LABEL: f5:
|
|
; CHECK: aghi %r2, -8
|
|
; CHECK: mxdb %f0, 0(%r2)
|
|
; CHECK: std %f0, 0(%r3)
|
|
; CHECK: std %f2, 8(%r3)
|
|
; CHECK: br %r14
|
|
%ptr = getelementptr double, double *%base, i64 -1
|
|
%f2 = load double, double *%ptr
|
|
%f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %f1,
|
|
metadata !"fpexcept.strict") #0
|
|
%f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %f2,
|
|
metadata !"fpexcept.strict") #0
|
|
%res = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %f1x, fp128 %f2x,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
store fp128 %res, fp128 *%dst
|
|
ret void
|
|
}
|
|
|
|
; Check that MXDB allows indices.
|
|
define void @f6(double %f1, double *%base, i64 %index, fp128 *%dst) #0 {
|
|
; CHECK-LABEL: f6:
|
|
; CHECK: sllg %r1, %r3, 3
|
|
; CHECK: mxdb %f0, 800(%r1,%r2)
|
|
; CHECK: std %f0, 0(%r4)
|
|
; CHECK: std %f2, 8(%r4)
|
|
; CHECK: br %r14
|
|
%ptr1 = getelementptr double, double *%base, i64 %index
|
|
%ptr2 = getelementptr double, double *%ptr1, i64 100
|
|
%f2 = load double, double *%ptr2
|
|
%f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %f1,
|
|
metadata !"fpexcept.strict") #0
|
|
%f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %f2,
|
|
metadata !"fpexcept.strict") #0
|
|
%res = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %f1x, fp128 %f2x,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
store fp128 %res, fp128 *%dst
|
|
ret void
|
|
}
|
|
|
|
; Check that multiplications of spilled values can use MXDB rather than MXDBR.
|
|
define double @f7(double *%ptr0) #0 {
|
|
; CHECK-LABEL: f7:
|
|
; CHECK: brasl %r14, foo@PLT
|
|
; CHECK: mxdb %f0, 160(%r15)
|
|
; CHECK: br %r14
|
|
%ptr1 = getelementptr double, double *%ptr0, i64 2
|
|
%ptr2 = getelementptr double, double *%ptr0, i64 4
|
|
%ptr3 = getelementptr double, double *%ptr0, i64 6
|
|
%ptr4 = getelementptr double, double *%ptr0, i64 8
|
|
%ptr5 = getelementptr double, double *%ptr0, i64 10
|
|
%ptr6 = getelementptr double, double *%ptr0, i64 12
|
|
%ptr7 = getelementptr double, double *%ptr0, i64 14
|
|
%ptr8 = getelementptr double, double *%ptr0, i64 16
|
|
%ptr9 = getelementptr double, double *%ptr0, i64 18
|
|
%ptr10 = getelementptr double, double *%ptr0, i64 20
|
|
|
|
%val0 = load double, double *%ptr0
|
|
%val1 = load double, double *%ptr1
|
|
%val2 = load double, double *%ptr2
|
|
%val3 = load double, double *%ptr3
|
|
%val4 = load double, double *%ptr4
|
|
%val5 = load double, double *%ptr5
|
|
%val6 = load double, double *%ptr6
|
|
%val7 = load double, double *%ptr7
|
|
%val8 = load double, double *%ptr8
|
|
%val9 = load double, double *%ptr9
|
|
%val10 = load double, double *%ptr10
|
|
|
|
%frob0 = call double @llvm.experimental.constrained.fadd.f64(
|
|
double %val0, double %val0,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%frob1 = call double @llvm.experimental.constrained.fadd.f64(
|
|
double %val1, double %val1,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%frob2 = call double @llvm.experimental.constrained.fadd.f64(
|
|
double %val2, double %val2,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%frob3 = call double @llvm.experimental.constrained.fadd.f64(
|
|
double %val3, double %val3,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%frob4 = call double @llvm.experimental.constrained.fadd.f64(
|
|
double %val4, double %val4,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%frob5 = call double @llvm.experimental.constrained.fadd.f64(
|
|
double %val5, double %val5,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%frob6 = call double @llvm.experimental.constrained.fadd.f64(
|
|
double %val6, double %val6,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%frob7 = call double @llvm.experimental.constrained.fadd.f64(
|
|
double %val7, double %val7,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%frob8 = call double @llvm.experimental.constrained.fadd.f64(
|
|
double %val8, double %val8,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%frob9 = call double @llvm.experimental.constrained.fadd.f64(
|
|
double %val9, double %val9,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%frob10 = call double @llvm.experimental.constrained.fadd.f64(
|
|
double %val10, double %val10,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
|
|
store double %frob0, double *%ptr0
|
|
store double %frob1, double *%ptr1
|
|
store double %frob2, double *%ptr2
|
|
store double %frob3, double *%ptr3
|
|
store double %frob4, double *%ptr4
|
|
store double %frob5, double *%ptr5
|
|
store double %frob6, double *%ptr6
|
|
store double %frob7, double *%ptr7
|
|
store double %frob8, double *%ptr8
|
|
store double %frob9, double *%ptr9
|
|
store double %frob10, double *%ptr10
|
|
|
|
%ret = call double @foo() #0
|
|
|
|
%accext0 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %ret,
|
|
metadata !"fpexcept.strict") #0
|
|
%ext0 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %frob0,
|
|
metadata !"fpexcept.strict") #0
|
|
%mul0 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %accext0, fp128 %ext0,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%extra0 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %mul0, fp128 0xL00000000000000003fff000001000000,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%trunc0 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
|
|
fp128 %extra0,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
|
|
%accext1 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %trunc0,
|
|
metadata !"fpexcept.strict") #0
|
|
%ext1 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %frob1,
|
|
metadata !"fpexcept.strict") #0
|
|
%mul1 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %accext1, fp128 %ext1,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%extra1 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %mul1, fp128 0xL00000000000000003fff000002000000,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%trunc1 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
|
|
fp128 %extra1,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
|
|
%accext2 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %trunc1,
|
|
metadata !"fpexcept.strict") #0
|
|
%ext2 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %frob2,
|
|
metadata !"fpexcept.strict") #0
|
|
%mul2 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %accext2, fp128 %ext2,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%extra2 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %mul2, fp128 0xL00000000000000003fff000003000000,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%trunc2 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
|
|
fp128 %extra2,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
|
|
%accext3 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %trunc2,
|
|
metadata !"fpexcept.strict") #0
|
|
%ext3 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %frob3,
|
|
metadata !"fpexcept.strict") #0
|
|
%mul3 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %accext3, fp128 %ext3,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%extra3 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %mul3, fp128 0xL00000000000000003fff000004000000,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%trunc3 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
|
|
fp128 %extra3,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
|
|
%accext4 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %trunc3,
|
|
metadata !"fpexcept.strict") #0
|
|
%ext4 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %frob4,
|
|
metadata !"fpexcept.strict") #0
|
|
%mul4 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %accext4, fp128 %ext4,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%extra4 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %mul4, fp128 0xL00000000000000003fff000005000000,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%trunc4 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
|
|
fp128 %extra4,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
|
|
%accext5 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %trunc4,
|
|
metadata !"fpexcept.strict") #0
|
|
%ext5 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %frob5,
|
|
metadata !"fpexcept.strict") #0
|
|
%mul5 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %accext5, fp128 %ext5,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%extra5 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %mul5, fp128 0xL00000000000000003fff000006000000,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%trunc5 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
|
|
fp128 %extra5,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
|
|
%accext6 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %trunc5,
|
|
metadata !"fpexcept.strict") #0
|
|
%ext6 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %frob6,
|
|
metadata !"fpexcept.strict") #0
|
|
%mul6 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %accext6, fp128 %ext6,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%extra6 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %mul6, fp128 0xL00000000000000003fff000007000000,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%trunc6 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
|
|
fp128 %extra6,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
|
|
%accext7 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %trunc6,
|
|
metadata !"fpexcept.strict") #0
|
|
%ext7 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %frob7,
|
|
metadata !"fpexcept.strict") #0
|
|
%mul7 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %accext7, fp128 %ext7,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%extra7 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %mul7, fp128 0xL00000000000000003fff000008000000,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%trunc7 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
|
|
fp128 %extra7,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
|
|
%accext8 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %trunc7,
|
|
metadata !"fpexcept.strict") #0
|
|
%ext8 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %frob8,
|
|
metadata !"fpexcept.strict") #0
|
|
%mul8 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %accext8, fp128 %ext8,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%extra8 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %mul8, fp128 0xL00000000000000003fff000009000000,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%trunc8 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
|
|
fp128 %extra8,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
|
|
%accext9 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %trunc8,
|
|
metadata !"fpexcept.strict") #0
|
|
%ext9 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
|
|
double %frob9,
|
|
metadata !"fpexcept.strict") #0
|
|
%mul9 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %accext9, fp128 %ext9,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%extra9 = call fp128 @llvm.experimental.constrained.fmul.f128(
|
|
fp128 %mul9, fp128 0xL00000000000000003fff00000a000000,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
%trunc9 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
|
|
fp128 %extra9,
|
|
metadata !"round.dynamic",
|
|
metadata !"fpexcept.strict") #0
|
|
|
|
ret double %trunc9
|
|
}
|
|
|
|
attributes #0 = { strictfp }
|