; RUN: llc -mtriple=aarch64--linux-gnu -mattr=+sve -asm-verbose=0 < %s 2>%t | FileCheck %s ; RUN: FileCheck --check-prefix=WARN --allow-empty %s <%t ; WARN-NOT: warning ; ; Masked Loads ; define @masked_load_nxv2i64( *%a, %mask) nounwind { ; CHECK-LABEL: masked_load_nxv2i64: ; CHECK-NEXT: ld1d { z0.d }, p0/z, [x0] ; CHECK-NEXT: ret %load = call @llvm.masked.load.nxv2i64( *%a, i32 8, %mask, undef) ret %load } define @masked_load_nxv4i32( *%a, %mask) nounwind { ; CHECK-LABEL: masked_load_nxv4i32: ; CHECK-NEXT: ld1w { z0.s }, p0/z, [x0] ; CHECK-NEXT: ret %load = call @llvm.masked.load.nxv4i32( *%a, i32 4, %mask, undef) ret %load } define @masked_load_nxv8i16( *%a, %mask) nounwind { ; CHECK-LABEL: masked_load_nxv8i16: ; CHECK-NEXT: ld1h { z0.h }, p0/z, [x0] ; CHECK-NEXT: ret %load = call @llvm.masked.load.nxv8i16( *%a, i32 2, %mask, undef) ret %load } define @masked_load_nxv16i8( *%a, %mask) nounwind { ; CHECK-LABEL: masked_load_nxv16i8: ; CHECK-NEXT: ld1b { z0.b }, p0/z, [x0] ; CHECK-NEXT: ret %load = call @llvm.masked.load.nxv16i8( *%a, i32 1, %mask, undef) ret %load } define @masked_load_nxv2f64( *%a, %mask) nounwind { ; CHECK-LABEL: masked_load_nxv2f64: ; CHECK-NEXT: ld1d { z0.d }, p0/z, [x0] ; CHECK-NEXT: ret %load = call @llvm.masked.load.nxv2f64( *%a, i32 8, %mask, undef) ret %load } define @masked_load_nxv2f32( *%a, %mask) nounwind { ; CHECK-LABEL: masked_load_nxv2f32: ; CHECK-NEXT: ld1w { z0.d }, p0/z, [x0] ; CHECK-NEXT: ret %load = call @llvm.masked.load.nxv2f32( *%a, i32 4, %mask, undef) ret %load } define @masked_load_nxv2f16( *%a, %mask) nounwind { ; CHECK-LABEL: masked_load_nxv2f16: ; CHECK-NEXT: ld1h { z0.d }, p0/z, [x0] ; CHECK-NEXT: ret %load = call @llvm.masked.load.nxv2f16( *%a, i32 2, %mask, undef) ret %load } define @masked_load_nxv4f32( *%a, %mask) nounwind { ; CHECK-LABEL: masked_load_nxv4f32: ; CHECK-NEXT: ld1w { z0.s }, p0/z, [x0] ; CHECK-NEXT: ret %load = call @llvm.masked.load.nxv4f32( *%a, i32 4, %mask, undef) ret %load } define @masked_load_nxv4f16( *%a, %mask) nounwind { ; CHECK-LABEL: masked_load_nxv4f16: ; CHECK-NEXT: ld1h { z0.s }, p0/z, [x0] ; CHECK-NEXT: ret %load = call @llvm.masked.load.nxv4f16( *%a, i32 2, %mask, undef) ret %load } define @masked_load_nxv8f16( *%a, %mask) nounwind { ; CHECK-LABEL: masked_load_nxv8f16: ; CHECK-NEXT: ld1h { z0.h }, p0/z, [x0] ; CHECK-NEXT: ret %load = call @llvm.masked.load.nxv8f16( *%a, i32 2, %mask, undef) ret %load } define @masked_load_nxv8bf16( *%a, %mask) nounwind #0 { ; CHECK-LABEL: masked_load_nxv8bf16: ; CHECK-NEXT: ld1h { z0.h }, p0/z, [x0] ; CHECK-NEXT: ret %load = call @llvm.masked.load.nxv8bf16( *%a, i32 2, %mask, undef) ret %load } ; ; Masked Stores ; define void @masked_store_nxv2i64( *%a, %val, %mask) nounwind { ; CHECK-LABEL: masked_store_nxv2i64: ; CHECK-NEXT: st1d { z0.d }, p0, [x0] ; CHECK-NEXT: ret call void @llvm.masked.store.nxv2i64( %val, *%a, i32 8, %mask) ret void } define void @masked_store_nxv4i32( *%a, %val, %mask) nounwind { ; CHECK-LABEL: masked_store_nxv4i32: ; CHECK-NEXT: st1w { z0.s }, p0, [x0] ; CHECK-NEXT: ret call void @llvm.masked.store.nxv4i32( %val, *%a, i32 4, %mask) ret void } define void @masked_store_nxv8i16( *%a, %val, %mask) nounwind { ; CHECK-LABEL: masked_store_nxv8i16: ; CHECK-NEXT: st1h { z0.h }, p0, [x0] ; CHECK-NEXT: ret call void @llvm.masked.store.nxv8i16( %val, *%a, i32 2, %mask) ret void } define void @masked_store_nxv16i8( *%a, %val, %mask) nounwind { ; CHECK-LABEL: masked_store_nxv16i8: ; CHECK-NEXT: st1b { z0.b }, p0, [x0] ; CHECK-NEXT: ret call void @llvm.masked.store.nxv16i8( %val, *%a, i32 1, %mask) ret void } define void @masked_store_nxv2f64( *%a, %val, %mask) nounwind { ; CHECK-LABEL: masked_store_nxv2f64: ; CHECK-NEXT: st1d { z0.d }, p0, [x0] ; CHECK-NEXT: ret call void @llvm.masked.store.nxv2f64( %val, *%a, i32 8, %mask) ret void } define void @masked_store_nxv2f32( *%a, %val, %mask) nounwind { ; CHECK-LABEL: masked_store_nxv2f32: ; CHECK-NEXT: st1w { z0.d }, p0, [x0] ; CHECK-NEXT: ret call void @llvm.masked.store.nxv2f32( %val, *%a, i32 4, %mask) ret void } define void @masked_store_nxv2f16( *%a, %val, %mask) nounwind { ; CHECK-LABEL: masked_store_nxv2f16: ; CHECK-NEXT: st1h { z0.d }, p0, [x0] ; CHECK-NEXT: ret call void @llvm.masked.store.nxv2f16( %val, *%a, i32 4, %mask) ret void } define void @masked_store_nxv4f32( *%a, %val, %mask) nounwind { ; CHECK-LABEL: masked_store_nxv4f32: ; CHECK-NEXT: st1w { z0.s }, p0, [x0] ; CHECK-NEXT: ret call void @llvm.masked.store.nxv4f32( %val, *%a, i32 4, %mask) ret void } define void @masked_store_nxv4f16( *%a, %val, %mask) nounwind { ; CHECK-LABEL: masked_store_nxv4f16: ; CHECK-NEXT: st1h { z0.s }, p0, [x0] ; CHECK-NEXT: ret call void @llvm.masked.store.nxv4f16( %val, *%a, i32 2, %mask) ret void } define void @masked_store_nxv8f16( *%a, %val, %mask) nounwind { ; CHECK-LABEL: masked_store_nxv8f16: ; CHECK-NEXT: st1h { z0.h }, p0, [x0] ; CHECK-NEXT: ret call void @llvm.masked.store.nxv8f16( %val, *%a, i32 2, %mask) ret void } define void @masked_store_nxv8bf16( *%a, %val, %mask) nounwind #0 { ; CHECK-LABEL: masked_store_nxv8bf16: ; CHECK-NEXT: st1h { z0.h }, p0, [x0] ; CHECK-NEXT: ret call void @llvm.masked.store.nxv8bf16( %val, *%a, i32 2, %mask) ret void } declare @llvm.masked.load.nxv2i64(*, i32, , ) declare @llvm.masked.load.nxv4i32(*, i32, , ) declare @llvm.masked.load.nxv8i16(*, i32, , ) declare @llvm.masked.load.nxv16i8(*, i32, , ) declare @llvm.masked.load.nxv2f64(*, i32, , ) declare @llvm.masked.load.nxv2f32(*, i32, , ) declare @llvm.masked.load.nxv2f16(*, i32, , ) declare @llvm.masked.load.nxv4f32(*, i32, , ) declare @llvm.masked.load.nxv4f16(*, i32, , ) declare @llvm.masked.load.nxv8f16(*, i32, , ) declare @llvm.masked.load.nxv8bf16(*, i32, , ) declare void @llvm.masked.store.nxv2i64(, *, i32, ) declare void @llvm.masked.store.nxv4i32(, *, i32, ) declare void @llvm.masked.store.nxv8i16(, *, i32, ) declare void @llvm.masked.store.nxv16i8(, *, i32, ) declare void @llvm.masked.store.nxv2f64(, *, i32, ) declare void @llvm.masked.store.nxv2f32(, *, i32, ) declare void @llvm.masked.store.nxv2f16(, *, i32, ) declare void @llvm.masked.store.nxv4f32(, *, i32, ) declare void @llvm.masked.store.nxv4f16(, *, i32, ) declare void @llvm.masked.store.nxv8f16(, *, i32, ) declare void @llvm.masked.store.nxv8bf16(, *, i32, ) ; +bf16 is required for the bfloat version. attributes #0 = { "target-features"="+sve,+bf16" }