mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-25 20:23:11 +01:00
0e3bafc4e2
Regenerated using: ./llvm/utils/update_llc_test_checks.py -u llvm/test/CodeGen/RISCV/*.ll This has added comments to spill-related instructions and added @plt to some symbols. Differential Revision: https://reviews.llvm.org/D92841
54 lines
2.4 KiB
LLVM
54 lines
2.4 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc -mtriple=riscv64 -verify-machineinstrs < %s | FileCheck %s
|
|
|
|
; The complex floating value will be returned by a single register for LP64 ABI.
|
|
; The test case check that the real part returned by __addsf3 will be
|
|
; cleared upper bits by shifts to avoid corrupting the imaginary part.
|
|
|
|
define i64 @complex_float_add(i64 %a.coerce, i64 %b.coerce) nounwind {
|
|
; CHECK-LABEL: complex_float_add:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: addi sp, sp, -32
|
|
; CHECK-NEXT: sd ra, 24(sp) # 8-byte Folded Spill
|
|
; CHECK-NEXT: sd s0, 16(sp) # 8-byte Folded Spill
|
|
; CHECK-NEXT: sd s1, 8(sp) # 8-byte Folded Spill
|
|
; CHECK-NEXT: sd s2, 0(sp) # 8-byte Folded Spill
|
|
; CHECK-NEXT: srli s2, a0, 32
|
|
; CHECK-NEXT: srli s1, a1, 32
|
|
; CHECK-NEXT: call __addsf3@plt
|
|
; CHECK-NEXT: mv s0, a0
|
|
; CHECK-NEXT: mv a0, s2
|
|
; CHECK-NEXT: mv a1, s1
|
|
; CHECK-NEXT: call __addsf3@plt
|
|
; CHECK-NEXT: slli a0, a0, 32
|
|
; CHECK-NEXT: slli a1, s0, 32
|
|
; CHECK-NEXT: srli a1, a1, 32
|
|
; CHECK-NEXT: or a0, a0, a1
|
|
; CHECK-NEXT: ld s2, 0(sp) # 8-byte Folded Reload
|
|
; CHECK-NEXT: ld s1, 8(sp) # 8-byte Folded Reload
|
|
; CHECK-NEXT: ld s0, 16(sp) # 8-byte Folded Reload
|
|
; CHECK-NEXT: ld ra, 24(sp) # 8-byte Folded Reload
|
|
; CHECK-NEXT: addi sp, sp, 32
|
|
; CHECK-NEXT: ret
|
|
entry:
|
|
%a.sroa.0.0.extract.trunc = trunc i64 %a.coerce to i32
|
|
%0 = bitcast i32 %a.sroa.0.0.extract.trunc to float
|
|
%a.sroa.2.0.extract.shift = lshr i64 %a.coerce, 32
|
|
%a.sroa.2.0.extract.trunc = trunc i64 %a.sroa.2.0.extract.shift to i32
|
|
%1 = bitcast i32 %a.sroa.2.0.extract.trunc to float
|
|
%b.sroa.0.0.extract.trunc = trunc i64 %b.coerce to i32
|
|
%2 = bitcast i32 %b.sroa.0.0.extract.trunc to float
|
|
%b.sroa.2.0.extract.shift = lshr i64 %b.coerce, 32
|
|
%b.sroa.2.0.extract.trunc = trunc i64 %b.sroa.2.0.extract.shift to i32
|
|
%3 = bitcast i32 %b.sroa.2.0.extract.trunc to float
|
|
%add.r = fadd float %0, %2
|
|
%add.i = fadd float %1, %3
|
|
%4 = bitcast float %add.r to i32
|
|
%5 = bitcast float %add.i to i32
|
|
%retval.sroa.2.0.insert.ext = zext i32 %5 to i64
|
|
%retval.sroa.2.0.insert.shift = shl nuw i64 %retval.sroa.2.0.insert.ext, 32
|
|
%retval.sroa.0.0.insert.ext = zext i32 %4 to i64
|
|
%retval.sroa.0.0.insert.insert = or i64 %retval.sroa.2.0.insert.shift, %retval.sroa.0.0.insert.ext
|
|
ret i64 %retval.sroa.0.0.insert.insert
|
|
}
|