mirror of
https://github.com/RPCS3/llvm-mirror.git
synced 2024-11-23 11:13:28 +01:00
c2a599b0fa
TargetMachine::shouldAssumeDSOLocal currently implies dso_local for such definitions. Adding explicit dso_local makes these tests align with the clang -fno-pic behavior and allow the removal of the TargetMachine::shouldAssumeDSOLocal special case.
341 lines
8.1 KiB
LLVM
341 lines
8.1 KiB
LLVM
; RUN: llc < %s -emulated-tls -mtriple=i386-linux-gnu | FileCheck -check-prefix=X86 %s
|
|
; RUN: llc < %s -emulated-tls -mtriple=x86_64-linux-gnu | FileCheck -check-prefix=X64 %s
|
|
; RUN: llc < %s -emulated-tls -mtriple=i386-linux-android | FileCheck -check-prefix=X86 %s
|
|
; RUN: llc < %s -emulated-tls -mtriple=x86_64-linux-android | FileCheck -check-prefix=X64 %s
|
|
|
|
; RUN: llc < %s -mtriple=i386-linux-gnu | FileCheck -check-prefix=NoEMU %s
|
|
; RUN: llc < %s -mtriple=x86_64-linux-gnu | FileCheck -check-prefix=NoEMU %s
|
|
; RUN: llc < %s -mtriple=i386-linux-android | FileCheck -check-prefix=X86 %s
|
|
; RUN: llc < %s -mtriple=x86_64-linux-android | FileCheck -check-prefix=X64 %s
|
|
|
|
; Copied from tls.ll; emulated TLS model is not implemented
|
|
; for *-pc-win32 and *-pc-windows targets yet.
|
|
|
|
; NoEMU-NOT: __emutls
|
|
|
|
; Use my_emutls_get_address like __emutls_get_address.
|
|
@my_emutls_v_xyz = external global i8*, align 4
|
|
declare i8* @my_emutls_get_address(i8*)
|
|
|
|
define dso_local i32 @my_get_xyz() {
|
|
; X86-LABEL: my_get_xyz:
|
|
; X86: movl $my_emutls_v_xyz, (%esp)
|
|
; X86-NEXT: calll my_emutls_get_address
|
|
; X86-NEXT: movl (%eax), %eax
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: my_get_xyz:
|
|
; X64: movq my_emutls_v_xyz@GOTPCREL(%rip), %rdi
|
|
; X64-NEXT: callq my_emutls_get_address
|
|
; X64-NEXT: movl (%rax), %eax
|
|
; X64-NEXT: popq %rcx
|
|
; X64-NEXT: .cfi_def_cfa_offset 8
|
|
; X64-NEXT: retq
|
|
entry:
|
|
%call = call i8* @my_emutls_get_address(i8* bitcast (i8** @my_emutls_v_xyz to i8*))
|
|
%0 = bitcast i8* %call to i32*
|
|
%1 = load i32, i32* %0, align 4
|
|
ret i32 %1
|
|
}
|
|
|
|
@i1 = dso_local thread_local global i32 15
|
|
@i2 = external thread_local global i32
|
|
@i3 = internal thread_local global i32 15
|
|
@i4 = hidden thread_local global i32 15
|
|
@i5 = external hidden thread_local global i32
|
|
@s1 = dso_local thread_local global i16 15
|
|
@b1 = dso_local thread_local global i8 0
|
|
|
|
define dso_local i32 @f1() {
|
|
; X86-LABEL: f1:
|
|
; X86: movl $__emutls_v.i1, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: movl (%eax), %eax
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: f1:
|
|
; X64: movl $__emutls_v.i1, %edi
|
|
; X64-NEXT: callq __emutls_get_address
|
|
; X64-NEXT: movl (%rax), %eax
|
|
; X64-NEXT: popq %rcx
|
|
; X64-NEXT: .cfi_def_cfa_offset 8
|
|
; X64-NEXT: retq
|
|
entry:
|
|
%tmp1 = load i32, i32* @i1
|
|
ret i32 %tmp1
|
|
}
|
|
|
|
define dso_local i32* @f2() {
|
|
; X86-LABEL: f2:
|
|
; X86: movl $__emutls_v.i1, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: f2:
|
|
; X64: movl $__emutls_v.i1, %edi
|
|
; X64-NEXT: callq __emutls_get_address
|
|
; X64-NEXT: popq %rcx
|
|
; X64-NEXT: .cfi_def_cfa_offset 8
|
|
; X64-NEXT: retq
|
|
entry:
|
|
ret i32* @i1
|
|
}
|
|
|
|
define dso_local i32 @f3() nounwind {
|
|
; X86-LABEL: f3:
|
|
; X86: movl $__emutls_v.i2, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: movl (%eax), %eax
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: retl
|
|
entry:
|
|
%tmp1 = load i32, i32* @i2
|
|
ret i32 %tmp1
|
|
}
|
|
|
|
define dso_local i32* @f4() {
|
|
; X86-LABEL: f4:
|
|
; X86: movl $__emutls_v.i2, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
entry:
|
|
ret i32* @i2
|
|
}
|
|
|
|
define dso_local i32 @f5() nounwind {
|
|
; X86-LABEL: f5:
|
|
; X86: movl $__emutls_v.i3, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: movl (%eax), %eax
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: retl
|
|
entry:
|
|
%tmp1 = load i32, i32* @i3
|
|
ret i32 %tmp1
|
|
}
|
|
|
|
define dso_local i32* @f6() {
|
|
; X86-LABEL: f6:
|
|
; X86: movl $__emutls_v.i3, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
entry:
|
|
ret i32* @i3
|
|
}
|
|
|
|
define dso_local i32 @f7() {
|
|
; X86-LABEL: f7:
|
|
; X86: movl $__emutls_v.i4, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: movl (%eax), %eax
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
entry:
|
|
%tmp1 = load i32, i32* @i4
|
|
ret i32 %tmp1
|
|
}
|
|
|
|
define dso_local i32* @f8() {
|
|
; X86-LABEL: f8:
|
|
; X86: movl $__emutls_v.i4, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
entry:
|
|
ret i32* @i4
|
|
}
|
|
|
|
define dso_local i32 @f9() {
|
|
; X86-LABEL: f9:
|
|
; X86: movl $__emutls_v.i5, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: movl (%eax), %eax
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
entry:
|
|
%tmp1 = load i32, i32* @i5
|
|
ret i32 %tmp1
|
|
}
|
|
|
|
define dso_local i32* @f10() {
|
|
; X86-LABEL: f10:
|
|
; X86: movl $__emutls_v.i5, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
entry:
|
|
ret i32* @i5
|
|
}
|
|
|
|
define dso_local i16 @f11() {
|
|
; X86-LABEL: f11:
|
|
; X86: movl $__emutls_v.s1, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: movzwl (%eax), %eax
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
entry:
|
|
%tmp1 = load i16, i16* @s1
|
|
ret i16 %tmp1
|
|
}
|
|
|
|
define dso_local i32 @f12() {
|
|
; X86-LABEL: f12:
|
|
; X86: movl $__emutls_v.s1, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: movswl (%eax), %eax
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
entry:
|
|
%tmp1 = load i16, i16* @s1
|
|
%tmp2 = sext i16 %tmp1 to i32
|
|
ret i32 %tmp2
|
|
}
|
|
|
|
define dso_local i8 @f13() {
|
|
; X86-LABEL: f13:
|
|
; X86: movl $__emutls_v.b1, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: movb (%eax), %al
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
entry:
|
|
%tmp1 = load i8, i8* @b1
|
|
ret i8 %tmp1
|
|
}
|
|
|
|
define dso_local i32 @f14() {
|
|
; X86-LABEL: f14:
|
|
; X86: movl $__emutls_v.b1, (%esp)
|
|
; X86-NEXT: calll __emutls_get_address
|
|
; X86-NEXT: movsbl (%eax), %eax
|
|
; X86-NEXT: addl $12, %esp
|
|
; X86-NEXT: .cfi_def_cfa_offset 4
|
|
; X86-NEXT: retl
|
|
entry:
|
|
%tmp1 = load i8, i8* @b1
|
|
%tmp2 = sext i8 %tmp1 to i32
|
|
ret i32 %tmp2
|
|
}
|
|
|
|
;;;;;;;;;;;;;; 32-bit __emutls_v. and __emutls_t.
|
|
|
|
; X86-LABEL: __emutls_v.i1:
|
|
; X86-NEXT: .long 4
|
|
; X86-NEXT: .long 4
|
|
; X86-NEXT: .long 0
|
|
; X86-NEXT: .long __emutls_t.i1
|
|
|
|
; X86-LABEL: __emutls_t.i1:
|
|
; X86-NEXT: .long 15
|
|
|
|
; X86-NOT: __emutls_v.i2
|
|
|
|
; X86-LABEL: __emutls_v.i3:
|
|
; X86-NEXT: .long 4
|
|
; X86-NEXT: .long 4
|
|
; X86-NEXT: .long 0
|
|
; X86-NEXT: .long __emutls_t.i3
|
|
|
|
; X86-LABEL: __emutls_t.i3:
|
|
; X86-NEXT: .long 15
|
|
|
|
; X86-LABEL: __emutls_v.i4:
|
|
; X86-NEXT: .long 4
|
|
; X86-NEXT: .long 4
|
|
; X86-NEXT: .long 0
|
|
; X86-NEXT: .long __emutls_t.i4
|
|
|
|
; X86-LABEL: __emutls_t.i4:
|
|
; X86-NEXT: .long 15
|
|
|
|
; X86-NOT: __emutls_v.i5:
|
|
; X86: .hidden __emutls_v.i5
|
|
; X86-NOT: __emutls_v.i5:
|
|
|
|
; X86-LABEL: __emutls_v.s1:
|
|
; X86-NEXT: .long 2
|
|
; X86-NEXT: .long 2
|
|
; X86-NEXT: .long 0
|
|
; X86-NEXT: .long __emutls_t.s1
|
|
|
|
; X86-LABEL: __emutls_t.s1:
|
|
; X86-NEXT: .short 15
|
|
|
|
; X86-LABEL: __emutls_v.b1:
|
|
; X86-NEXT: .long 1
|
|
; X86-NEXT: .long 1
|
|
; X86-NEXT: .long 0
|
|
; X86-NEXT: .long 0
|
|
|
|
; X86-NOT: __emutls_t.b1
|
|
|
|
;;;;;;;;;;;;;; 64-bit __emutls_v. and __emutls_t.
|
|
|
|
; X64-LABEL: __emutls_v.i1:
|
|
; X64-NEXT: .quad 4
|
|
; X64-NEXT: .quad 4
|
|
; X64-NEXT: .quad 0
|
|
; X64-NEXT: .quad __emutls_t.i1
|
|
|
|
; X64-LABEL: __emutls_t.i1:
|
|
; X64-NEXT: .long 15
|
|
|
|
; X64-NOT: __emutls_v.i2
|
|
|
|
; X64-LABEL: __emutls_v.i3:
|
|
; X64-NEXT: .quad 4
|
|
; X64-NEXT: .quad 4
|
|
; X64-NEXT: .quad 0
|
|
; X64-NEXT: .quad __emutls_t.i3
|
|
|
|
; X64-LABEL: __emutls_t.i3:
|
|
; X64-NEXT: .long 15
|
|
|
|
; X64-LABEL: __emutls_v.i4:
|
|
; X64-NEXT: .quad 4
|
|
; X64-NEXT: .quad 4
|
|
; X64-NEXT: .quad 0
|
|
; X64-NEXT: .quad __emutls_t.i4
|
|
|
|
; X64-LABEL: __emutls_t.i4:
|
|
; X64-NEXT: .long 15
|
|
|
|
; X64-NOT: __emutls_v.i5:
|
|
; X64: .hidden __emutls_v.i5
|
|
; X64-NOT: __emutls_v.i5:
|
|
|
|
; X64-LABEL: __emutls_v.s1:
|
|
; X64-NEXT: .quad 2
|
|
; X64-NEXT: .quad 2
|
|
; X64-NEXT: .quad 0
|
|
; X64-NEXT: .quad __emutls_t.s1
|
|
|
|
; X64-LABEL: __emutls_t.s1:
|
|
; X64-NEXT: .short 15
|
|
|
|
; X64-LABEL: __emutls_v.b1:
|
|
; X64-NEXT: .quad 1
|
|
; X64-NEXT: .quad 1
|
|
; X64-NEXT: .quad 0
|
|
; X64-NEXT: .quad 0
|
|
|
|
; X64-NOT: __emutls_t.b1
|