mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-23 11:13:28 +01:00
a41ca2b24d
The combineSelect code was casting to i64 without any check that i64 was legal. This can break after type legalization. It also required splitting the mmx register on 32-bit targets. It's not clear that this makes sense. Instead switch to using a cmov pseudo like we do for XMM/YMM/ZMM.
126 lines
3.5 KiB
LLVM
126 lines
3.5 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc -mtriple=x86_64-unknown-unknown -mattr=+mmx < %s | FileCheck %s --check-prefix=X64
|
|
; RUN: llc -mtriple=i686-unknown-unknown -mattr=+mmx < %s | FileCheck %s --check-prefix=I32
|
|
|
|
|
|
; From source: clang -02
|
|
;__m64 test47(int a)
|
|
;{
|
|
; __m64 x = (a)? (__m64)(7): (__m64)(0);
|
|
; return __builtin_ia32_psllw(x, x);
|
|
;}
|
|
|
|
define i64 @test47(i64 %arg) {
|
|
;
|
|
; X64-LABEL: test47:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: testq %rdi, %rdi
|
|
; X64-NEXT: je .LBB0_1
|
|
; X64-NEXT: # %bb.2:
|
|
; X64-NEXT: pxor %mm0, %mm0
|
|
; X64-NEXT: jmp .LBB0_3
|
|
; X64-NEXT: .LBB0_1:
|
|
; X64-NEXT: movl $7, %eax
|
|
; X64-NEXT: movd %eax, %mm0
|
|
; X64-NEXT: .LBB0_3:
|
|
; X64-NEXT: psllw %mm0, %mm0
|
|
; X64-NEXT: movq %mm0, %rax
|
|
; X64-NEXT: retq
|
|
;
|
|
; I32-LABEL: test47:
|
|
; I32: # %bb.0:
|
|
; I32-NEXT: pushl %ebp
|
|
; I32-NEXT: .cfi_def_cfa_offset 8
|
|
; I32-NEXT: .cfi_offset %ebp, -8
|
|
; I32-NEXT: movl %esp, %ebp
|
|
; I32-NEXT: .cfi_def_cfa_register %ebp
|
|
; I32-NEXT: andl $-8, %esp
|
|
; I32-NEXT: subl $8, %esp
|
|
; I32-NEXT: movl 8(%ebp), %eax
|
|
; I32-NEXT: orl 12(%ebp), %eax
|
|
; I32-NEXT: je .LBB0_1
|
|
; I32-NEXT: # %bb.2:
|
|
; I32-NEXT: pxor %mm0, %mm0
|
|
; I32-NEXT: jmp .LBB0_3
|
|
; I32-NEXT: .LBB0_1:
|
|
; I32-NEXT: movl $7, %eax
|
|
; I32-NEXT: movd %eax, %mm0
|
|
; I32-NEXT: .LBB0_3:
|
|
; I32-NEXT: psllw %mm0, %mm0
|
|
; I32-NEXT: movq %mm0, (%esp)
|
|
; I32-NEXT: movl (%esp), %eax
|
|
; I32-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; I32-NEXT: movl %ebp, %esp
|
|
; I32-NEXT: popl %ebp
|
|
; I32-NEXT: .cfi_def_cfa %esp, 4
|
|
; I32-NEXT: retl
|
|
%cond = icmp eq i64 %arg, 0
|
|
%slct = select i1 %cond, x86_mmx bitcast (i64 7 to x86_mmx), x86_mmx bitcast (i64 0 to x86_mmx)
|
|
%psll = tail call x86_mmx @llvm.x86.mmx.psll.w(x86_mmx %slct, x86_mmx %slct)
|
|
%retc = bitcast x86_mmx %psll to i64
|
|
ret i64 %retc
|
|
}
|
|
|
|
|
|
; From source: clang -O2
|
|
;__m64 test49(int a, long long n, long long m)
|
|
;{
|
|
; __m64 x = (a)? (__m64)(n): (__m64)(m);
|
|
; return __builtin_ia32_psllw(x, x);
|
|
;}
|
|
|
|
define i64 @test49(i64 %arg, i64 %x, i64 %y) {
|
|
;
|
|
; X64-LABEL: test49:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: testq %rdi, %rdi
|
|
; X64-NEXT: je .LBB1_1
|
|
; X64-NEXT: # %bb.2:
|
|
; X64-NEXT: movq %rdx, %mm0
|
|
; X64-NEXT: jmp .LBB1_3
|
|
; X64-NEXT: .LBB1_1:
|
|
; X64-NEXT: movq %rsi, %mm0
|
|
; X64-NEXT: .LBB1_3:
|
|
; X64-NEXT: psllw %mm0, %mm0
|
|
; X64-NEXT: movq %mm0, %rax
|
|
; X64-NEXT: retq
|
|
;
|
|
; I32-LABEL: test49:
|
|
; I32: # %bb.0:
|
|
; I32-NEXT: pushl %ebp
|
|
; I32-NEXT: .cfi_def_cfa_offset 8
|
|
; I32-NEXT: .cfi_offset %ebp, -8
|
|
; I32-NEXT: movl %esp, %ebp
|
|
; I32-NEXT: .cfi_def_cfa_register %ebp
|
|
; I32-NEXT: andl $-8, %esp
|
|
; I32-NEXT: subl $8, %esp
|
|
; I32-NEXT: movl 8(%ebp), %eax
|
|
; I32-NEXT: orl 12(%ebp), %eax
|
|
; I32-NEXT: je .LBB1_1
|
|
; I32-NEXT: # %bb.2:
|
|
; I32-NEXT: leal 24(%ebp), %eax
|
|
; I32-NEXT: jmp .LBB1_3
|
|
; I32-NEXT: .LBB1_1:
|
|
; I32-NEXT: leal 16(%ebp), %eax
|
|
; I32-NEXT: .LBB1_3:
|
|
; I32-NEXT: movq (%eax), %mm0
|
|
; I32-NEXT: psllw %mm0, %mm0
|
|
; I32-NEXT: movq %mm0, (%esp)
|
|
; I32-NEXT: movl (%esp), %eax
|
|
; I32-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; I32-NEXT: movl %ebp, %esp
|
|
; I32-NEXT: popl %ebp
|
|
; I32-NEXT: .cfi_def_cfa %esp, 4
|
|
; I32-NEXT: retl
|
|
%cond = icmp eq i64 %arg, 0
|
|
%xmmx = bitcast i64 %x to x86_mmx
|
|
%ymmx = bitcast i64 %y to x86_mmx
|
|
%slct = select i1 %cond, x86_mmx %xmmx, x86_mmx %ymmx
|
|
%psll = tail call x86_mmx @llvm.x86.mmx.psll.w(x86_mmx %slct, x86_mmx %slct)
|
|
%retc = bitcast x86_mmx %psll to i64
|
|
ret i64 %retc
|
|
}
|
|
|
|
declare x86_mmx @llvm.x86.mmx.psll.w(x86_mmx, x86_mmx)
|
|
|