1
0
mirror of https://github.com/RPCS3/llvm-mirror.git synced 2024-10-25 22:12:57 +02:00
llvm-mirror/test/Transforms/GVN/big-endian.ll
Ulrich Weigand 4c20bac964 [GVN] Fix handling of sub-byte types in big-endian mode
When GVN wants to re-interpret an already available value in a smaller
type, it needs to right-shift the value on big-endian systems to ensure
the correct bytes are accessed.  The shift value is the difference of
the sizes of the two types.

This is correct as long as both types occupy multiples of full bytes.
However, when one of them is a sub-byte type like i1, this no longer
holds true: we still need to shift, but only to access the correct
*byte*.  Accessing bits within the byte requires no shift in either
endianness; e.g. an i1 resides in the least-significant bit of its
containing byte on both big- and little-endian systems.

Therefore, the appropriate shift value to be used is the difference of
the *storage* sizes of the two types.  This is already handled correctly
in one place where such a shift takes place (GetStoreValueForLoad), but
is incorrect in two other places: GetLoadValueForLoad and
CoerceAvailableValueToLoadType.

This patch changes both places to use the storage size as well.

Differential Revision: http://reviews.llvm.org/D18662

llvm-svn: 265684
2016-04-07 15:45:02 +00:00

41 lines
1.8 KiB
LLVM

; RUN: opt -gvn -S < %s | FileCheck %s
target datalayout = "E-m:e-i64:64-n32:64"
target triple = "powerpc64-unknown-linux-gnu"
;; Make sure we use correct bit shift based on storage size for
;; loads reusing a load value.
define i64 @test1({ i1, i8 }* %predA, { i1, i8 }* %predB) {
; CHECK-LABEL: @test1
; CHECK: [[V1:%.*]] = load i16, i16* %{{.*}}
; CHECK: [[V2:%.*]] = lshr i16 [[V1]], 8
; CHECK: trunc i16 [[V2]] to i1
%valueLoadA.fca.0.gep = getelementptr inbounds { i1, i8 }, { i1, i8 }* %predA, i64 0, i32 0
%valueLoadA.fca.0.load = load i1, i1* %valueLoadA.fca.0.gep, align 8
%valueLoadB.fca.0.gep = getelementptr inbounds { i1, i8 }, { i1, i8 }* %predB, i64 0, i32 0
%valueLoadB.fca.0.load = load i1, i1* %valueLoadB.fca.0.gep, align 8
%isTrue = and i1 %valueLoadA.fca.0.load, %valueLoadB.fca.0.load
%valueLoadA.fca.1.gep = getelementptr inbounds { i1, i8 }, { i1, i8 }* %predA, i64 0, i32 1
%valueLoadA.fca.1.load = load i8, i8* %valueLoadA.fca.1.gep, align 1
%isNotNullA = icmp ne i8 %valueLoadA.fca.1.load, 0
%valueLoadB.fca.1.gep = getelementptr inbounds { i1, i8 }, { i1, i8 }* %predB, i64 0, i32 1
%valueLoadB.fca.1.load = load i8, i8* %valueLoadB.fca.1.gep, align 1
%isNotNullB = icmp ne i8 %valueLoadB.fca.1.load, 0
%isNotNull = and i1 %isNotNullA, %isNotNullB
%isTrueAndNotNull = and i1 %isTrue, %isNotNull
%ret = zext i1 %isTrueAndNotNull to i64
ret i64 %ret
}
;; And likewise for loads reusing a store value.
define i1 @test2(i8 %V, i8* %P) {
; CHECK-LABEL: @test2
; CHECK-NOT: lshr
store i8 %V, i8* %P
%P2 = bitcast i8* %P to i1*
%A = load i1, i1* %P2
ret i1 %A
}