diff --git a/lib/Target/AArch64/AArch64InstrAtomics.td b/lib/Target/AArch64/AArch64InstrAtomics.td index 24f75895be0..027381f4036 100644 --- a/lib/Target/AArch64/AArch64InstrAtomics.td +++ b/lib/Target/AArch64/AArch64InstrAtomics.td @@ -204,19 +204,27 @@ def : Pat<(relaxed_store def ldxr_1 : PatFrag<(ops node:$ptr), (int_aarch64_ldxr node:$ptr), [{ return cast(N)->getMemoryVT() == MVT::i8; -}]>; +}]> { + let GISelPredicateCode = [{ return isLoadStoreOfNumBytes(MI, 1); }]; +} def ldxr_2 : PatFrag<(ops node:$ptr), (int_aarch64_ldxr node:$ptr), [{ return cast(N)->getMemoryVT() == MVT::i16; -}]>; +}]> { + let GISelPredicateCode = [{ return isLoadStoreOfNumBytes(MI, 2); }]; +} def ldxr_4 : PatFrag<(ops node:$ptr), (int_aarch64_ldxr node:$ptr), [{ return cast(N)->getMemoryVT() == MVT::i32; -}]>; +}]> { + let GISelPredicateCode = [{ return isLoadStoreOfNumBytes(MI, 4); }]; +} def ldxr_8 : PatFrag<(ops node:$ptr), (int_aarch64_ldxr node:$ptr), [{ return cast(N)->getMemoryVT() == MVT::i64; -}]>; +}]> { + let GISelPredicateCode = [{ return isLoadStoreOfNumBytes(MI, 8); }]; +} def : Pat<(ldxr_1 GPR64sp:$addr), (SUBREG_TO_REG (i64 0), (LDXRB GPR64sp:$addr), sub_32)>; diff --git a/test/CodeGen/AArch64/GlobalISel/select-ldxr-intrin.mir b/test/CodeGen/AArch64/GlobalISel/select-ldxr-intrin.mir new file mode 100644 index 00000000000..167cfc84e2a --- /dev/null +++ b/test/CodeGen/AArch64/GlobalISel/select-ldxr-intrin.mir @@ -0,0 +1,95 @@ +# NOTE: Assertions have been autogenerated by utils/update_mir_test_checks.py +# RUN: llc -mtriple=aarch64-unknown-unknown -run-pass=instruction-select -verify-machineinstrs %s -o - | FileCheck %s + +--- | + define void @test_load_i8(i8* %addr) { ret void } + define void @test_load_i16(i16* %addr) { ret void } + define void @test_load_i32(i32* %addr) { ret void } + define void @test_load_i64(i64* %addr) { ret void } +... +--- +name: test_load_i8 +alignment: 2 +legalized: true +regBankSelected: true +tracksRegLiveness: true +body: | + bb.0: + liveins: $x0 + ; CHECK-LABEL: name: test_load_i8 + ; CHECK: liveins: $x0 + ; CHECK: [[COPY:%[0-9]+]]:gpr64sp = COPY $x0 + ; CHECK: [[LDXRB:%[0-9]+]]:gpr32 = LDXRB [[COPY]] :: (volatile load 1 from %ir.addr) + ; CHECK: [[SUBREG_TO_REG:%[0-9]+]]:gpr64all = SUBREG_TO_REG 0, [[LDXRB]], %subreg.sub_32 + ; CHECK: $x1 = COPY [[SUBREG_TO_REG]] + ; CHECK: RET_ReallyLR implicit $x1 + %0:gpr(p0) = COPY $x0 + %1:gpr(s64) = G_INTRINSIC_W_SIDE_EFFECTS intrinsic(@llvm.aarch64.ldxr), %0(p0) :: (volatile load 1 from %ir.addr) + $x1 = COPY %1(s64) + RET_ReallyLR implicit $x1 + +... +--- +name: test_load_i16 +alignment: 2 +legalized: true +regBankSelected: true +tracksRegLiveness: true +body: | + bb.0: + liveins: $x0 + ; CHECK-LABEL: name: test_load_i16 + ; CHECK: liveins: $x0 + ; CHECK: [[COPY:%[0-9]+]]:gpr64sp = COPY $x0 + ; CHECK: [[LDXRH:%[0-9]+]]:gpr32 = LDXRH [[COPY]] :: (volatile load 2 from %ir.addr) + ; CHECK: [[SUBREG_TO_REG:%[0-9]+]]:gpr64all = SUBREG_TO_REG 0, [[LDXRH]], %subreg.sub_32 + ; CHECK: $x1 = COPY [[SUBREG_TO_REG]] + ; CHECK: RET_ReallyLR implicit $x1 + %0:gpr(p0) = COPY $x0 + %1:gpr(s64) = G_INTRINSIC_W_SIDE_EFFECTS intrinsic(@llvm.aarch64.ldxr), %0(p0) :: (volatile load 2 from %ir.addr) + $x1 = COPY %1(s64) + RET_ReallyLR implicit $x1 + +... +--- +name: test_load_i32 +alignment: 2 +legalized: true +regBankSelected: true +tracksRegLiveness: true +body: | + bb.0: + liveins: $x0 + ; CHECK-LABEL: name: test_load_i32 + ; CHECK: liveins: $x0 + ; CHECK: [[COPY:%[0-9]+]]:gpr64sp = COPY $x0 + ; CHECK: [[LDXRW:%[0-9]+]]:gpr32 = LDXRW [[COPY]] :: (volatile load 4 from %ir.addr) + ; CHECK: [[SUBREG_TO_REG:%[0-9]+]]:gpr64all = SUBREG_TO_REG 0, [[LDXRW]], %subreg.sub_32 + ; CHECK: $x1 = COPY [[SUBREG_TO_REG]] + ; CHECK: RET_ReallyLR implicit $x1 + %0:gpr(p0) = COPY $x0 + %1:gpr(s64) = G_INTRINSIC_W_SIDE_EFFECTS intrinsic(@llvm.aarch64.ldxr), %0(p0) :: (volatile load 4 from %ir.addr) + $x1 = COPY %1(s64) + RET_ReallyLR implicit $x1 + + +... +--- +name: test_load_i64 +alignment: 2 +legalized: true +regBankSelected: true +tracksRegLiveness: true +body: | + bb.0: + liveins: $x0 + ; CHECK-LABEL: name: test_load_i64 + ; CHECK: liveins: $x0 + ; CHECK: [[COPY:%[0-9]+]]:gpr64sp = COPY $x0 + ; CHECK: [[LDXRX:%[0-9]+]]:gpr64 = LDXRX [[COPY]] :: (volatile load 8 from %ir.addr) + ; CHECK: $x1 = COPY [[LDXRX]] + ; CHECK: RET_ReallyLR implicit $x1 + %0:gpr(p0) = COPY $x0 + %1:gpr(s64) = G_INTRINSIC_W_SIDE_EFFECTS intrinsic(@llvm.aarch64.ldxr), %0(p0) :: (volatile load 8 from %ir.addr) + $x1 = COPY %1(s64) + RET_ReallyLR implicit $x1 diff --git a/test/CodeGen/AArch64/arm64-ldxr-stxr.ll b/test/CodeGen/AArch64/arm64-ldxr-stxr.ll index 680b4389dac..e688a29112a 100644 --- a/test/CodeGen/AArch64/arm64-ldxr-stxr.ll +++ b/test/CodeGen/AArch64/arm64-ldxr-stxr.ll @@ -33,6 +33,7 @@ declare i32 @llvm.aarch64.stxp(i64, i64, i8*) nounwind @var = global i64 0, align 8 +; FALLBACK-NOT: remark:{{.*}}test_load_i8 define void @test_load_i8(i8* %addr) { ; CHECK-LABEL: test_load_i8: ; CHECK: ldxrb w[[LOADVAL:[0-9]+]], [x0] @@ -40,6 +41,12 @@ define void @test_load_i8(i8* %addr) { ; CHECK-NOT: and ; CHECK: str x[[LOADVAL]], [{{x[0-9]+}}, :lo12:var] +; FIXME: GlobalISel doesn't fold ands/adds into load/store addressing modes +; right now/ So, we won't get the :lo12:var. +; GISEL-LABEL: test_load_i8: +; GISEL: ldxrb w[[LOADVAL:[0-9]+]], [x0] +; GISEL-NOT: uxtb +; GISEL: str x[[LOADVAL]], [{{x[0-9]+}}] %val = call i64 @llvm.aarch64.ldxr.p0i8(i8* %addr) %shortval = trunc i64 %val to i8 %extval = zext i8 %shortval to i64 @@ -47,6 +54,7 @@ define void @test_load_i8(i8* %addr) { ret void } +; FALLBACK-NOT: remark:{{.*}}test_load_i16 define void @test_load_i16(i16* %addr) { ; CHECK-LABEL: test_load_i16: ; CHECK: ldxrh w[[LOADVAL:[0-9]+]], [x0] @@ -54,6 +62,10 @@ define void @test_load_i16(i16* %addr) { ; CHECK-NOT: and ; CHECK: str x[[LOADVAL]], [{{x[0-9]+}}, :lo12:var] +; GISEL-LABEL: test_load_i16: +; GISEL: ldxrh w[[LOADVAL:[0-9]+]], [x0] +; GISEL-NOT: uxtb +; GISEL: str x[[LOADVAL]], [{{x[0-9]+}}] %val = call i64 @llvm.aarch64.ldxr.p0i16(i16* %addr) %shortval = trunc i64 %val to i16 %extval = zext i16 %shortval to i64 @@ -61,6 +73,7 @@ define void @test_load_i16(i16* %addr) { ret void } +; FALLBACK-NOT: remark:{{.*}}test_load_i32 define void @test_load_i32(i32* %addr) { ; CHECK-LABEL: test_load_i32: ; CHECK: ldxr w[[LOADVAL:[0-9]+]], [x0] @@ -68,6 +81,10 @@ define void @test_load_i32(i32* %addr) { ; CHECK-NOT: and ; CHECK: str x[[LOADVAL]], [{{x[0-9]+}}, :lo12:var] +; GISEL-LABEL: test_load_i32: +; GISEL: ldxr w[[LOADVAL:[0-9]+]], [x0] +; GISEL-NOT: uxtb +; GISEL: str x[[LOADVAL]], [{{x[0-9]+}}] %val = call i64 @llvm.aarch64.ldxr.p0i32(i32* %addr) %shortval = trunc i64 %val to i32 %extval = zext i32 %shortval to i64 @@ -75,11 +92,16 @@ define void @test_load_i32(i32* %addr) { ret void } +; FALLBACK-NOT: remark:{{.*}}test_load_i64 define void @test_load_i64(i64* %addr) { ; CHECK-LABEL: test_load_i64: ; CHECK: ldxr x[[LOADVAL:[0-9]+]], [x0] ; CHECK: str x[[LOADVAL]], [{{x[0-9]+}}, :lo12:var] +; GISEL-LABEL: test_load_i64: +; GISEL: ldxr x[[LOADVAL:[0-9]+]], [x0] +; GISEL-NOT: uxtb +; GISEL: str x[[LOADVAL]], [{{x[0-9]+}}] %val = call i64 @llvm.aarch64.ldxr.p0i64(i64* %addr) store i64 %val, i64* @var, align 8 ret void