1
0
mirror of https://github.com/RPCS3/llvm-mirror.git synced 2024-10-23 04:52:54 +02:00
llvm-mirror/test/Bitcode/memInstructions.3.2.ll
Tim Northover b9ec29d7c5 IR: add "cmpxchg weak" variant to support permitted failure.
This commit adds a weak variant of the cmpxchg operation, as described
in C++11. A cmpxchg instruction with this modifier is permitted to
fail to store, even if the comparison indicated it should.

As a result, cmpxchg instructions must return a flag indicating
success in addition to their original iN value loaded. Thus, for
uniformity *all* cmpxchg instructions now return "{ iN, i1 }". The
second flag is 1 when the store succeeded.

At the DAG level, a new ATOMIC_CMP_SWAP_WITH_SUCCESS node has been
added as the natural representation for the new cmpxchg instructions.
It is a strong cmpxchg.

By default this gets Expanded to the existing ATOMIC_CMP_SWAP during
Legalization, so existing backends should see no change in behaviour.
If they wish to deal with the enhanced node instead, they can call
setOperationAction on it. Beware: as a node with 2 results, it cannot
be selected from TableGen.

Currently, no use is made of the extra information provided in this
patch. Test updates are almost entirely adapting the input IR to the
new scheme.

Summary for out of tree users:
------------------------------

+ Legacy Bitcode files are upgraded during read.
+ Legacy assembly IR files will be invalid.
+ Front-ends must adapt to different type for "cmpxchg".
+ Backends should be unaffected by default.

llvm-svn: 210903
2014-06-13 14:24:07 +00:00

328 lines
14 KiB
LLVM

; RUN: llvm-dis < %s.bc| FileCheck %s
; memOperations.3.2.ll.bc was generated by passing this file to llvm-as-3.2.
; The test checks that LLVM does not misread memory related instructions of
; older bitcode files.
define void @alloca(){
entry:
; CHECK: %res1 = alloca i8
%res1 = alloca i8
; CHECK-NEXT: %res2 = alloca i8, i32 2
%res2 = alloca i8, i32 2
; CHECK-NEXT: %res3 = alloca i8, i32 2, align 4
%res3 = alloca i8, i32 2, align 4
; CHECK-NEXT: %res4 = alloca i8, align 4
%res4 = alloca i8, align 4
ret void
}
define void @load(){
entry:
%ptr1 = alloca i8
store i8 2, i8* %ptr1
; CHECK: %res1 = load i8* %ptr1
%res1 = load i8* %ptr1
; CHECK-NEXT: %res2 = load volatile i8* %ptr1
%res2 = load volatile i8* %ptr1
; CHECK-NEXT: %res3 = load i8* %ptr1, align 1
%res3 = load i8* %ptr1, align 1
; CHECK-NEXT: %res4 = load volatile i8* %ptr1, align 1
%res4 = load volatile i8* %ptr1, align 1
; CHECK-NEXT: %res5 = load i8* %ptr1, !nontemporal !0
%res5 = load i8* %ptr1, !nontemporal !0
; CHECK-NEXT: %res6 = load volatile i8* %ptr1, !nontemporal !0
%res6 = load volatile i8* %ptr1, !nontemporal !0
; CHECK-NEXT: %res7 = load i8* %ptr1, align 1, !nontemporal !0
%res7 = load i8* %ptr1, align 1, !nontemporal !0
; CHECK-NEXT: %res8 = load volatile i8* %ptr1, align 1, !nontemporal !0
%res8 = load volatile i8* %ptr1, align 1, !nontemporal !0
; CHECK-NEXT: %res9 = load i8* %ptr1, !invariant.load !1
%res9 = load i8* %ptr1, !invariant.load !1
; CHECK-NEXT: %res10 = load volatile i8* %ptr1, !invariant.load !1
%res10 = load volatile i8* %ptr1, !invariant.load !1
; CHECK-NEXT: %res11 = load i8* %ptr1, align 1, !invariant.load !1
%res11 = load i8* %ptr1, align 1, !invariant.load !1
; CHECK-NEXT: %res12 = load volatile i8* %ptr1, align 1, !invariant.load !1
%res12 = load volatile i8* %ptr1, align 1, !invariant.load !1
; CHECK-NEXT: %res13 = load i8* %ptr1, {{[(!nontemporal !0, !invariant.load !1) | (!invariant.load !1, !nontemporal !0)]}}
%res13 = load i8* %ptr1, !nontemporal !0, !invariant.load !1
; CHECK-NEXT: %res14 = load volatile i8* %ptr1, {{[(!nontemporal !0, !invariant.load !1) | (!invariant.load !1, !nontemporal !0)]}}
%res14 = load volatile i8* %ptr1, !nontemporal !0, !invariant.load !1
; CHECK-NEXT: %res15 = load i8* %ptr1, align 1, {{[(!nontemporal !0, !invariant.load !1) | (!invariant.load !1, !nontemporal !0)]}}
%res15 = load i8* %ptr1, align 1, !nontemporal !0, !invariant.load !1
; CHECK-NEXT: %res16 = load volatile i8* %ptr1, align 1, {{[(!nontemporal !0, !invariant.load !1) | (!invariant.load !1, !nontemporal !0)]}}
%res16 = load volatile i8* %ptr1, align 1, !nontemporal !0, !invariant.load !1
ret void
}
define void @loadAtomic(){
entry:
%ptr1 = alloca i8
store i8 2, i8* %ptr1
; CHECK: %res1 = load atomic i8* %ptr1 unordered, align 1
%res1 = load atomic i8* %ptr1 unordered, align 1
; CHECK-NEXT: %res2 = load atomic i8* %ptr1 monotonic, align 1
%res2 = load atomic i8* %ptr1 monotonic, align 1
; CHECK-NEXT: %res3 = load atomic i8* %ptr1 acquire, align 1
%res3 = load atomic i8* %ptr1 acquire, align 1
; CHECK-NEXT: %res4 = load atomic i8* %ptr1 seq_cst, align 1
%res4 = load atomic i8* %ptr1 seq_cst, align 1
; CHECK-NEXT: %res5 = load atomic volatile i8* %ptr1 unordered, align 1
%res5 = load atomic volatile i8* %ptr1 unordered, align 1
; CHECK-NEXT: %res6 = load atomic volatile i8* %ptr1 monotonic, align 1
%res6 = load atomic volatile i8* %ptr1 monotonic, align 1
; CHECK-NEXT: %res7 = load atomic volatile i8* %ptr1 acquire, align 1
%res7 = load atomic volatile i8* %ptr1 acquire, align 1
; CHECK-NEXT: %res8 = load atomic volatile i8* %ptr1 seq_cst, align 1
%res8 = load atomic volatile i8* %ptr1 seq_cst, align 1
; CHECK-NEXT: %res9 = load atomic i8* %ptr1 singlethread unordered, align 1
%res9 = load atomic i8* %ptr1 singlethread unordered, align 1
; CHECK-NEXT: %res10 = load atomic i8* %ptr1 singlethread monotonic, align 1
%res10 = load atomic i8* %ptr1 singlethread monotonic, align 1
; CHECK-NEXT: %res11 = load atomic i8* %ptr1 singlethread acquire, align 1
%res11 = load atomic i8* %ptr1 singlethread acquire, align 1
; CHECK-NEXT: %res12 = load atomic i8* %ptr1 singlethread seq_cst, align 1
%res12 = load atomic i8* %ptr1 singlethread seq_cst, align 1
; CHECK-NEXT: %res13 = load atomic volatile i8* %ptr1 singlethread unordered, align 1
%res13 = load atomic volatile i8* %ptr1 singlethread unordered, align 1
; CHECK-NEXT: %res14 = load atomic volatile i8* %ptr1 singlethread monotonic, align 1
%res14 = load atomic volatile i8* %ptr1 singlethread monotonic, align 1
; CHECK-NEXT: %res15 = load atomic volatile i8* %ptr1 singlethread acquire, align 1
%res15 = load atomic volatile i8* %ptr1 singlethread acquire, align 1
; CHECK-NEXT: %res16 = load atomic volatile i8* %ptr1 singlethread seq_cst, align 1
%res16 = load atomic volatile i8* %ptr1 singlethread seq_cst, align 1
ret void
}
define void @store(){
entry:
%ptr1 = alloca i8
; CHECK: store i8 2, i8* %ptr1
store i8 2, i8* %ptr1
; CHECK-NEXT: store volatile i8 2, i8* %ptr1
store volatile i8 2, i8* %ptr1
; CHECK-NEXT: store i8 2, i8* %ptr1, align 1
store i8 2, i8* %ptr1, align 1
; CHECK-NEXT: store volatile i8 2, i8* %ptr1, align 1
store volatile i8 2, i8* %ptr1, align 1
; CHECK-NEXT: store i8 2, i8* %ptr1, !nontemporal !0
store i8 2, i8* %ptr1, !nontemporal !0
; CHECK-NEXT: store volatile i8 2, i8* %ptr1, !nontemporal !0
store volatile i8 2, i8* %ptr1, !nontemporal !0
; CHECK-NEXT: store i8 2, i8* %ptr1, align 1, !nontemporal !0
store i8 2, i8* %ptr1, align 1, !nontemporal !0
; CHECK-NEXT: store volatile i8 2, i8* %ptr1, align 1, !nontemporal !0
store volatile i8 2, i8* %ptr1, align 1, !nontemporal !0
ret void
}
define void @storeAtomic(){
entry:
%ptr1 = alloca i8
; CHECK: store atomic i8 2, i8* %ptr1 unordered, align 1
store atomic i8 2, i8* %ptr1 unordered, align 1
; CHECK-NEXT: store atomic i8 2, i8* %ptr1 monotonic, align 1
store atomic i8 2, i8* %ptr1 monotonic, align 1
; CHECK-NEXT: store atomic i8 2, i8* %ptr1 release, align 1
store atomic i8 2, i8* %ptr1 release, align 1
; CHECK-NEXT: store atomic i8 2, i8* %ptr1 seq_cst, align 1
store atomic i8 2, i8* %ptr1 seq_cst, align 1
; CHECK-NEXT: store atomic volatile i8 2, i8* %ptr1 unordered, align 1
store atomic volatile i8 2, i8* %ptr1 unordered, align 1
; CHECK-NEXT: store atomic volatile i8 2, i8* %ptr1 monotonic, align 1
store atomic volatile i8 2, i8* %ptr1 monotonic, align 1
; CHECK-NEXT: store atomic volatile i8 2, i8* %ptr1 release, align 1
store atomic volatile i8 2, i8* %ptr1 release, align 1
; CHECK-NEXT: store atomic volatile i8 2, i8* %ptr1 seq_cst, align 1
store atomic volatile i8 2, i8* %ptr1 seq_cst, align 1
; CHECK-NEXT: store atomic i8 2, i8* %ptr1 singlethread unordered, align 1
store atomic i8 2, i8* %ptr1 singlethread unordered, align 1
; CHECK-NEXT: store atomic i8 2, i8* %ptr1 singlethread monotonic, align 1
store atomic i8 2, i8* %ptr1 singlethread monotonic, align 1
; CHECK-NEXT: store atomic i8 2, i8* %ptr1 singlethread release, align 1
store atomic i8 2, i8* %ptr1 singlethread release, align 1
; CHECK-NEXT: store atomic i8 2, i8* %ptr1 singlethread seq_cst, align 1
store atomic i8 2, i8* %ptr1 singlethread seq_cst, align 1
; CHECK-NEXT: store atomic volatile i8 2, i8* %ptr1 singlethread unordered, align 1
store atomic volatile i8 2, i8* %ptr1 singlethread unordered, align 1
; CHECK-NEXT: store atomic volatile i8 2, i8* %ptr1 singlethread monotonic, align 1
store atomic volatile i8 2, i8* %ptr1 singlethread monotonic, align 1
; CHECK-NEXT: store atomic volatile i8 2, i8* %ptr1 singlethread release, align 1
store atomic volatile i8 2, i8* %ptr1 singlethread release, align 1
; CHECK-NEXT: store atomic volatile i8 2, i8* %ptr1 singlethread seq_cst, align 1
store atomic volatile i8 2, i8* %ptr1 singlethread seq_cst, align 1
ret void
}
define void @cmpxchg(i32* %ptr,i32 %cmp,i32 %new){
entry:
;cmpxchg [volatile] <ty>* <pointer>, <ty> <cmp>, <ty> <new> [singlethread] <ordering>
; CHECK: [[TMP:%[a-z0-9]+]] = cmpxchg i32* %ptr, i32 %cmp, i32 %new monotonic monotonic
; CHECK-NEXT: %res1 = extractvalue { i32, i1 } [[TMP]], 0
%res1 = cmpxchg i32* %ptr, i32 %cmp, i32 %new monotonic monotonic
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new monotonic monotonic
; CHECK-NEXT: %res2 = extractvalue { i32, i1 } [[TMP]], 0
%res2 = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new monotonic monotonic
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg i32* %ptr, i32 %cmp, i32 %new singlethread monotonic monotonic
; CHECK-NEXT: %res3 = extractvalue { i32, i1 } [[TMP]], 0
%res3 = cmpxchg i32* %ptr, i32 %cmp, i32 %new singlethread monotonic monotonic
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new singlethread monotonic monotonic
; CHECK-NEXT: %res4 = extractvalue { i32, i1 } [[TMP]], 0
%res4 = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new singlethread monotonic monotonic
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg i32* %ptr, i32 %cmp, i32 %new acquire acquire
; CHECK-NEXT: %res5 = extractvalue { i32, i1 } [[TMP]], 0
%res5 = cmpxchg i32* %ptr, i32 %cmp, i32 %new acquire acquire
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new acquire acquire
; CHECK-NEXT: %res6 = extractvalue { i32, i1 } [[TMP]], 0
%res6 = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new acquire acquire
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg i32* %ptr, i32 %cmp, i32 %new singlethread acquire acquire
; CHECK-NEXT: %res7 = extractvalue { i32, i1 } [[TMP]], 0
%res7 = cmpxchg i32* %ptr, i32 %cmp, i32 %new singlethread acquire acquire
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new singlethread acquire acquire
; CHECK-NEXT: %res8 = extractvalue { i32, i1 } [[TMP]], 0
%res8 = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new singlethread acquire acquire
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg i32* %ptr, i32 %cmp, i32 %new release monotonic
; CHECK-NEXT: %res9 = extractvalue { i32, i1 } [[TMP]], 0
%res9 = cmpxchg i32* %ptr, i32 %cmp, i32 %new release monotonic
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new release monotonic
; CHECK-NEXT: %res10 = extractvalue { i32, i1 } [[TMP]], 0
%res10 = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new release monotonic
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg i32* %ptr, i32 %cmp, i32 %new singlethread release monotonic
; CHECK-NEXT: %res11 = extractvalue { i32, i1 } [[TMP]], 0
%res11 = cmpxchg i32* %ptr, i32 %cmp, i32 %new singlethread release monotonic
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new singlethread release monotonic
; CHECK-NEXT: %res12 = extractvalue { i32, i1 } [[TMP]], 0
%res12 = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new singlethread release monotonic
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg i32* %ptr, i32 %cmp, i32 %new acq_rel acquire
; CHECK-NEXT: %res13 = extractvalue { i32, i1 } [[TMP]], 0
%res13 = cmpxchg i32* %ptr, i32 %cmp, i32 %new acq_rel acquire
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new acq_rel acquire
; CHECK-NEXT: %res14 = extractvalue { i32, i1 } [[TMP]], 0
%res14 = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new acq_rel acquire
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg i32* %ptr, i32 %cmp, i32 %new singlethread acq_rel acquire
; CHECK-NEXT: %res15 = extractvalue { i32, i1 } [[TMP]], 0
%res15 = cmpxchg i32* %ptr, i32 %cmp, i32 %new singlethread acq_rel acquire
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new singlethread acq_rel acquire
; CHECK-NEXT: %res16 = extractvalue { i32, i1 } [[TMP]], 0
%res16 = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new singlethread acq_rel acquire
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg i32* %ptr, i32 %cmp, i32 %new seq_cst seq_cst
; CHECK-NEXT: %res17 = extractvalue { i32, i1 } [[TMP]], 0
%res17 = cmpxchg i32* %ptr, i32 %cmp, i32 %new seq_cst seq_cst
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new seq_cst seq_cst
; CHECK-NEXT: %res18 = extractvalue { i32, i1 } [[TMP]], 0
%res18 = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new seq_cst seq_cst
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg i32* %ptr, i32 %cmp, i32 %new singlethread seq_cst seq_cst
; CHECK-NEXT: %res19 = extractvalue { i32, i1 } [[TMP]], 0
%res19 = cmpxchg i32* %ptr, i32 %cmp, i32 %new singlethread seq_cst seq_cst
; CHECK-NEXT: [[TMP:%[a-z0-9]+]] = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new singlethread seq_cst seq_cst
; CHECK-NEXT: %res20 = extractvalue { i32, i1 } [[TMP]], 0
%res20 = cmpxchg volatile i32* %ptr, i32 %cmp, i32 %new singlethread seq_cst seq_cst
ret void
}
define void @getelementptr({i8, i8}* %s, <4 x i8*> %ptrs, <4 x i64> %offsets ){
entry:
; CHECK: %res1 = getelementptr { i8, i8 }* %s, i32 1, i32 1
%res1 = getelementptr {i8, i8}* %s, i32 1, i32 1
; CHECK-NEXT: %res2 = getelementptr inbounds { i8, i8 }* %s, i32 1, i32 1
%res2 = getelementptr inbounds {i8, i8}* %s, i32 1, i32 1
; CHECK-NEXT: %res3 = getelementptr <4 x i8*> %ptrs, <4 x i64> %offsets
%res3 = getelementptr <4 x i8*> %ptrs, <4 x i64> %offsets
ret void
}
!0 = metadata !{ i32 1 }
!1 = metadata !{}