diff options
| author | sayantn <sayantn05@gmail.com> | 2025-06-02 21:38:25 +0530 |
|---|---|---|
| committer | Amanieu d'Antras <amanieu@gmail.com> | 2025-06-02 19:45:08 +0000 |
| commit | 068cc378a6ba1633ae497dcaeee290799107f176 (patch) | |
| tree | 2066c9bcadc9612ecce9f959a5ec80e57614f45b /library/stdarch/crates | |
| parent | 5d4794185490fe1cbec3fba2a5925af286122a7c (diff) | |
| download | rust-068cc378a6ba1633ae497dcaeee290799107f176.tar.gz rust-068cc378a6ba1633ae497dcaeee290799107f176.zip | |
Remove uses of deprecated type-specific pointers from ARM
Diffstat (limited to 'library/stdarch/crates')
4 files changed, 380 insertions, 380 deletions
diff --git a/library/stdarch/crates/core_arch/src/aarch64/neon/generated.rs b/library/stdarch/crates/core_arch/src/aarch64/neon/generated.rs index c7621d95bc9..96ed82021b4 100644 --- a/library/stdarch/crates/core_arch/src/aarch64/neon/generated.rs +++ b/library/stdarch/crates/core_arch/src/aarch64/neon/generated.rs @@ -11421,7 +11421,7 @@ pub unsafe fn vld1_f64_x2(a: *const f64) -> float64x1x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld1x2.v1f64.p0f64" + link_name = "llvm.aarch64.neon.ld1x2.v1f64.p0" )] fn _vld1_f64_x2(a: *const f64) -> float64x1x2_t; } @@ -11439,7 +11439,7 @@ pub unsafe fn vld1_f64_x3(a: *const f64) -> float64x1x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld1x3.v1f64.p0f64" + link_name = "llvm.aarch64.neon.ld1x3.v1f64.p0" )] fn _vld1_f64_x3(a: *const f64) -> float64x1x3_t; } @@ -11457,7 +11457,7 @@ pub unsafe fn vld1_f64_x4(a: *const f64) -> float64x1x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld1x4.v1f64.p0f64" + link_name = "llvm.aarch64.neon.ld1x4.v1f64.p0" )] fn _vld1_f64_x4(a: *const f64) -> float64x1x4_t; } @@ -11475,7 +11475,7 @@ pub unsafe fn vld1q_f64_x2(a: *const f64) -> float64x2x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld1x2.v2f64.p0f64" + link_name = "llvm.aarch64.neon.ld1x2.v2f64.p0" )] fn _vld1q_f64_x2(a: *const f64) -> float64x2x2_t; } @@ -11493,7 +11493,7 @@ pub unsafe fn vld1q_f64_x3(a: *const f64) -> float64x2x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld1x3.v2f64.p0f64" + link_name = "llvm.aarch64.neon.ld1x3.v2f64.p0" )] fn _vld1q_f64_x3(a: *const f64) -> float64x2x3_t; } @@ -11511,7 +11511,7 @@ pub unsafe fn vld1q_f64_x4(a: *const f64) -> float64x2x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld1x4.v2f64.p0f64" + link_name = "llvm.aarch64.neon.ld1x4.v2f64.p0" )] fn _vld1q_f64_x4(a: *const f64) -> float64x2x4_t; } @@ -11529,7 +11529,7 @@ pub unsafe fn vld2_dup_f64(a: *const f64) -> float64x1x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2r.v1f64.p0f64" + link_name = "llvm.aarch64.neon.ld2r.v1f64.p0" )] fn _vld2_dup_f64(ptr: *const f64) -> float64x1x2_t; } @@ -11547,7 +11547,7 @@ pub unsafe fn vld2q_dup_f64(a: *const f64) -> float64x2x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2r.v2f64.p0f64" + link_name = "llvm.aarch64.neon.ld2r.v2f64.p0" )] fn _vld2q_dup_f64(ptr: *const f64) -> float64x2x2_t; } @@ -11565,7 +11565,7 @@ pub unsafe fn vld2q_dup_s64(a: *const i64) -> int64x2x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2r.v2i64.p0i64" + link_name = "llvm.aarch64.neon.ld2r.v2i64.p0" )] fn _vld2q_dup_s64(ptr: *const i64) -> int64x2x2_t; } @@ -11583,7 +11583,7 @@ pub unsafe fn vld2_f64(a: *const f64) -> float64x1x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2.v1f64.p0v1f64" + link_name = "llvm.aarch64.neon.ld2.v1f64.p0" )] fn _vld2_f64(ptr: *const float64x1_t) -> float64x1x2_t; } @@ -11603,7 +11603,7 @@ pub unsafe fn vld2_lane_f64<const LANE: i32>(a: *const f64, b: float64x1x2_t) -> unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2lane.v1f64.p0i8" + link_name = "llvm.aarch64.neon.ld2lane.v1f64.p0" )] fn _vld2_lane_f64(a: float64x1_t, b: float64x1_t, n: i64, ptr: *const i8) -> float64x1x2_t; } @@ -11623,7 +11623,7 @@ pub unsafe fn vld2_lane_s64<const LANE: i32>(a: *const i64, b: int64x1x2_t) -> i unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2lane.v1i64.p0i8" + link_name = "llvm.aarch64.neon.ld2lane.v1i64.p0" )] fn _vld2_lane_s64(a: int64x1_t, b: int64x1_t, n: i64, ptr: *const i8) -> int64x1x2_t; } @@ -11721,7 +11721,7 @@ pub unsafe fn vld2q_f64(a: *const f64) -> float64x2x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2.v2f64.p0v2f64" + link_name = "llvm.aarch64.neon.ld2.v2f64.p0" )] fn _vld2q_f64(ptr: *const float64x2_t) -> float64x2x2_t; } @@ -11739,7 +11739,7 @@ pub unsafe fn vld2q_s64(a: *const i64) -> int64x2x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2.v2i64.p0v2i64" + link_name = "llvm.aarch64.neon.ld2.v2i64.p0" )] fn _vld2q_s64(ptr: *const int64x2_t) -> int64x2x2_t; } @@ -11759,7 +11759,7 @@ pub unsafe fn vld2q_lane_f64<const LANE: i32>(a: *const f64, b: float64x2x2_t) - unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2lane.v2f64.p0i8" + link_name = "llvm.aarch64.neon.ld2lane.v2f64.p0" )] fn _vld2q_lane_f64(a: float64x2_t, b: float64x2_t, n: i64, ptr: *const i8) -> float64x2x2_t; @@ -11780,7 +11780,7 @@ pub unsafe fn vld2q_lane_s8<const LANE: i32>(a: *const i8, b: int8x16x2_t) -> in unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2lane.v16i8.p0i8" + link_name = "llvm.aarch64.neon.ld2lane.v16i8.p0" )] fn _vld2q_lane_s8(a: int8x16_t, b: int8x16_t, n: i64, ptr: *const i8) -> int8x16x2_t; } @@ -11800,7 +11800,7 @@ pub unsafe fn vld2q_lane_s64<const LANE: i32>(a: *const i64, b: int64x2x2_t) -> unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2lane.v2i64.p0i8" + link_name = "llvm.aarch64.neon.ld2lane.v2i64.p0" )] fn _vld2q_lane_s64(a: int64x2_t, b: int64x2_t, n: i64, ptr: *const i8) -> int64x2x2_t; } @@ -11924,7 +11924,7 @@ pub unsafe fn vld3_dup_f64(a: *const f64) -> float64x1x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3r.v1f64.p0f64" + link_name = "llvm.aarch64.neon.ld3r.v1f64.p0" )] fn _vld3_dup_f64(ptr: *const f64) -> float64x1x3_t; } @@ -11942,7 +11942,7 @@ pub unsafe fn vld3q_dup_f64(a: *const f64) -> float64x2x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3r.v2f64.p0f64" + link_name = "llvm.aarch64.neon.ld3r.v2f64.p0" )] fn _vld3q_dup_f64(ptr: *const f64) -> float64x2x3_t; } @@ -11960,7 +11960,7 @@ pub unsafe fn vld3q_dup_s64(a: *const i64) -> int64x2x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3r.v2i64.p0i64" + link_name = "llvm.aarch64.neon.ld3r.v2i64.p0" )] fn _vld3q_dup_s64(ptr: *const i64) -> int64x2x3_t; } @@ -11978,7 +11978,7 @@ pub unsafe fn vld3_f64(a: *const f64) -> float64x1x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3.v1f64.p0v1f64" + link_name = "llvm.aarch64.neon.ld3.v1f64.p0" )] fn _vld3_f64(ptr: *const float64x1_t) -> float64x1x3_t; } @@ -11998,7 +11998,7 @@ pub unsafe fn vld3_lane_f64<const LANE: i32>(a: *const f64, b: float64x1x3_t) -> unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3lane.v1f64.p0i8" + link_name = "llvm.aarch64.neon.ld3lane.v1f64.p0" )] fn _vld3_lane_f64( a: float64x1_t, @@ -12037,7 +12037,7 @@ pub unsafe fn vld3_lane_s64<const LANE: i32>(a: *const i64, b: int64x1x3_t) -> i unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3lane.v1i64.p0i8" + link_name = "llvm.aarch64.neon.ld3lane.v1i64.p0" )] fn _vld3_lane_s64( a: int64x1_t, @@ -12130,7 +12130,7 @@ pub unsafe fn vld3q_f64(a: *const f64) -> float64x2x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3.v2f64.p0v2f64" + link_name = "llvm.aarch64.neon.ld3.v2f64.p0" )] fn _vld3q_f64(ptr: *const float64x2_t) -> float64x2x3_t; } @@ -12148,7 +12148,7 @@ pub unsafe fn vld3q_s64(a: *const i64) -> int64x2x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3.v2i64.p0v2i64" + link_name = "llvm.aarch64.neon.ld3.v2i64.p0" )] fn _vld3q_s64(ptr: *const int64x2_t) -> int64x2x3_t; } @@ -12168,7 +12168,7 @@ pub unsafe fn vld3q_lane_f64<const LANE: i32>(a: *const f64, b: float64x2x3_t) - unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3lane.v2f64.p0i8" + link_name = "llvm.aarch64.neon.ld3lane.v2f64.p0" )] fn _vld3q_lane_f64( a: float64x2_t, @@ -12207,7 +12207,7 @@ pub unsafe fn vld3q_lane_s8<const LANE: i32>(a: *const i8, b: int8x16x3_t) -> in unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3lane.v16i8.p0i8" + link_name = "llvm.aarch64.neon.ld3lane.v16i8.p0" )] fn _vld3q_lane_s8( a: int8x16_t, @@ -12233,7 +12233,7 @@ pub unsafe fn vld3q_lane_s64<const LANE: i32>(a: *const i64, b: int64x2x3_t) -> unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3lane.v2i64.p0i8" + link_name = "llvm.aarch64.neon.ld3lane.v2i64.p0" )] fn _vld3q_lane_s64( a: int64x2_t, @@ -12352,7 +12352,7 @@ pub unsafe fn vld4_dup_f64(a: *const f64) -> float64x1x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v1f64.p0f64" + link_name = "llvm.aarch64.neon.ld4r.v1f64.p0" )] fn _vld4_dup_f64(ptr: *const f64) -> float64x1x4_t; } @@ -12370,7 +12370,7 @@ pub unsafe fn vld4q_dup_f64(a: *const f64) -> float64x2x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v2f64.p0f64" + link_name = "llvm.aarch64.neon.ld4r.v2f64.p0" )] fn _vld4q_dup_f64(ptr: *const f64) -> float64x2x4_t; } @@ -12388,7 +12388,7 @@ pub unsafe fn vld4q_dup_s64(a: *const i64) -> int64x2x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v2i64.p0i64" + link_name = "llvm.aarch64.neon.ld4r.v2i64.p0" )] fn _vld4q_dup_s64(ptr: *const i64) -> int64x2x4_t; } @@ -12406,7 +12406,7 @@ pub unsafe fn vld4_f64(a: *const f64) -> float64x1x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4.v1f64.p0v1f64" + link_name = "llvm.aarch64.neon.ld4.v1f64.p0" )] fn _vld4_f64(ptr: *const float64x1_t) -> float64x1x4_t; } @@ -12426,7 +12426,7 @@ pub unsafe fn vld4_lane_f64<const LANE: i32>(a: *const f64, b: float64x1x4_t) -> unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4lane.v1f64.p0i8" + link_name = "llvm.aarch64.neon.ld4lane.v1f64.p0" )] fn _vld4_lane_f64( a: float64x1_t, @@ -12453,7 +12453,7 @@ pub unsafe fn vld4_lane_s64<const LANE: i32>(a: *const i64, b: int64x1x4_t) -> i unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4lane.v1i64.p0i8" + link_name = "llvm.aarch64.neon.ld4lane.v1i64.p0" )] fn _vld4_lane_s64( a: int64x1_t, @@ -12562,7 +12562,7 @@ pub unsafe fn vld4q_f64(a: *const f64) -> float64x2x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4.v2f64.p0v2f64" + link_name = "llvm.aarch64.neon.ld4.v2f64.p0" )] fn _vld4q_f64(ptr: *const float64x2_t) -> float64x2x4_t; } @@ -12580,7 +12580,7 @@ pub unsafe fn vld4q_s64(a: *const i64) -> int64x2x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4.v2i64.p0v2i64" + link_name = "llvm.aarch64.neon.ld4.v2i64.p0" )] fn _vld4q_s64(ptr: *const int64x2_t) -> int64x2x4_t; } @@ -12600,7 +12600,7 @@ pub unsafe fn vld4q_lane_f64<const LANE: i32>(a: *const f64, b: float64x2x4_t) - unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4lane.v2f64.p0i8" + link_name = "llvm.aarch64.neon.ld4lane.v2f64.p0" )] fn _vld4q_lane_f64( a: float64x2_t, @@ -12627,7 +12627,7 @@ pub unsafe fn vld4q_lane_s8<const LANE: i32>(a: *const i8, b: int8x16x4_t) -> in unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4lane.v16i8.p0i8" + link_name = "llvm.aarch64.neon.ld4lane.v16i8.p0" )] fn _vld4q_lane_s8( a: int8x16_t, @@ -12654,7 +12654,7 @@ pub unsafe fn vld4q_lane_s64<const LANE: i32>(a: *const i64, b: int64x2x4_t) -> unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4lane.v2i64.p0i8" + link_name = "llvm.aarch64.neon.ld4lane.v2i64.p0" )] fn _vld4q_lane_s64( a: int64x2_t, @@ -25979,7 +25979,7 @@ pub unsafe fn vst1_f64_x2(a: *mut f64, b: float64x1x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v1f64.p0f64" + link_name = "llvm.aarch64.neon.st1x2.v1f64.p0" )] fn _vst1_f64_x2(a: float64x1_t, b: float64x1_t, ptr: *mut f64); } @@ -25997,7 +25997,7 @@ pub unsafe fn vst1q_f64_x2(a: *mut f64, b: float64x2x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v2f64.p0f64" + link_name = "llvm.aarch64.neon.st1x2.v2f64.p0" )] fn _vst1q_f64_x2(a: float64x2_t, b: float64x2_t, ptr: *mut f64); } @@ -26015,7 +26015,7 @@ pub unsafe fn vst1_f64_x3(a: *mut f64, b: float64x1x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v1f64.p0f64" + link_name = "llvm.aarch64.neon.st1x3.v1f64.p0" )] fn _vst1_f64_x3(a: float64x1_t, b: float64x1_t, c: float64x1_t, ptr: *mut f64); } @@ -26033,7 +26033,7 @@ pub unsafe fn vst1q_f64_x3(a: *mut f64, b: float64x2x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v2f64.p0f64" + link_name = "llvm.aarch64.neon.st1x3.v2f64.p0" )] fn _vst1q_f64_x3(a: float64x2_t, b: float64x2_t, c: float64x2_t, ptr: *mut f64); } @@ -26051,7 +26051,7 @@ pub unsafe fn vst1_f64_x4(a: *mut f64, b: float64x1x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v1f64.p0f64" + link_name = "llvm.aarch64.neon.st1x4.v1f64.p0" )] fn _vst1_f64_x4( a: float64x1_t, @@ -26075,7 +26075,7 @@ pub unsafe fn vst1q_f64_x4(a: *mut f64, b: float64x2x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v2f64.p0f64" + link_name = "llvm.aarch64.neon.st1x4.v2f64.p0" )] fn _vst1q_f64_x4( a: float64x2_t, @@ -26125,7 +26125,7 @@ pub unsafe fn vst2_f64(a: *mut f64, b: float64x1x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v1f64.p0i8" + link_name = "llvm.aarch64.neon.st2.v1f64.p0" )] fn _vst2_f64(a: float64x1_t, b: float64x1_t, ptr: *mut i8); } @@ -26145,7 +26145,7 @@ pub unsafe fn vst2_lane_f64<const LANE: i32>(a: *mut f64, b: float64x1x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v1f64.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v1f64.p0" )] fn _vst2_lane_f64(a: float64x1_t, b: float64x1_t, n: i64, ptr: *mut i8); } @@ -26165,7 +26165,7 @@ pub unsafe fn vst2_lane_s64<const LANE: i32>(a: *mut i64, b: int64x1x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v1i64.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v1i64.p0" )] fn _vst2_lane_s64(a: int64x1_t, b: int64x1_t, n: i64, ptr: *mut i8); } @@ -26209,7 +26209,7 @@ pub unsafe fn vst2q_f64(a: *mut f64, b: float64x2x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v2f64.p0i8" + link_name = "llvm.aarch64.neon.st2.v2f64.p0" )] fn _vst2q_f64(a: float64x2_t, b: float64x2_t, ptr: *mut i8); } @@ -26227,7 +26227,7 @@ pub unsafe fn vst2q_s64(a: *mut i64, b: int64x2x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v2i64.p0i8" + link_name = "llvm.aarch64.neon.st2.v2i64.p0" )] fn _vst2q_s64(a: int64x2_t, b: int64x2_t, ptr: *mut i8); } @@ -26247,7 +26247,7 @@ pub unsafe fn vst2q_lane_f64<const LANE: i32>(a: *mut f64, b: float64x2x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v2f64.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v2f64.p0" )] fn _vst2q_lane_f64(a: float64x2_t, b: float64x2_t, n: i64, ptr: *mut i8); } @@ -26267,7 +26267,7 @@ pub unsafe fn vst2q_lane_s8<const LANE: i32>(a: *mut i8, b: int8x16x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v16i8.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v16i8.p0" )] fn _vst2q_lane_s8(a: int8x16_t, b: int8x16_t, n: i64, ptr: *mut i8); } @@ -26287,7 +26287,7 @@ pub unsafe fn vst2q_lane_s64<const LANE: i32>(a: *mut i64, b: int64x2x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v2i64.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v2i64.p0" )] fn _vst2q_lane_s64(a: int64x2_t, b: int64x2_t, n: i64, ptr: *mut i8); } @@ -26379,7 +26379,7 @@ pub unsafe fn vst3_f64(a: *mut f64, b: float64x1x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v1f64.p0i8" + link_name = "llvm.aarch64.neon.st3.v1f64.p0" )] fn _vst3_f64(a: float64x1_t, b: float64x1_t, c: float64x1_t, ptr: *mut i8); } @@ -26399,7 +26399,7 @@ pub unsafe fn vst3_lane_f64<const LANE: i32>(a: *mut f64, b: float64x1x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v1f64.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v1f64.p0" )] fn _vst3_lane_f64(a: float64x1_t, b: float64x1_t, c: float64x1_t, n: i64, ptr: *mut i8); } @@ -26419,7 +26419,7 @@ pub unsafe fn vst3_lane_s64<const LANE: i32>(a: *mut i64, b: int64x1x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v1i64.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v1i64.p0" )] fn _vst3_lane_s64(a: int64x1_t, b: int64x1_t, c: int64x1_t, n: i64, ptr: *mut i8); } @@ -26463,7 +26463,7 @@ pub unsafe fn vst3q_f64(a: *mut f64, b: float64x2x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v2f64.p0i8" + link_name = "llvm.aarch64.neon.st3.v2f64.p0" )] fn _vst3q_f64(a: float64x2_t, b: float64x2_t, c: float64x2_t, ptr: *mut i8); } @@ -26481,7 +26481,7 @@ pub unsafe fn vst3q_s64(a: *mut i64, b: int64x2x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v2i64.p0i8" + link_name = "llvm.aarch64.neon.st3.v2i64.p0" )] fn _vst3q_s64(a: int64x2_t, b: int64x2_t, c: int64x2_t, ptr: *mut i8); } @@ -26501,7 +26501,7 @@ pub unsafe fn vst3q_lane_f64<const LANE: i32>(a: *mut f64, b: float64x2x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v2f64.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v2f64.p0" )] fn _vst3q_lane_f64(a: float64x2_t, b: float64x2_t, c: float64x2_t, n: i64, ptr: *mut i8); } @@ -26521,7 +26521,7 @@ pub unsafe fn vst3q_lane_s8<const LANE: i32>(a: *mut i8, b: int8x16x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v16i8.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v16i8.p0" )] fn _vst3q_lane_s8(a: int8x16_t, b: int8x16_t, c: int8x16_t, n: i64, ptr: *mut i8); } @@ -26541,7 +26541,7 @@ pub unsafe fn vst3q_lane_s64<const LANE: i32>(a: *mut i64, b: int64x2x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v2i64.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v2i64.p0" )] fn _vst3q_lane_s64(a: int64x2_t, b: int64x2_t, c: int64x2_t, n: i64, ptr: *mut i8); } @@ -26633,7 +26633,7 @@ pub unsafe fn vst4_f64(a: *mut f64, b: float64x1x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v1f64.p0i8" + link_name = "llvm.aarch64.neon.st4.v1f64.p0" )] fn _vst4_f64(a: float64x1_t, b: float64x1_t, c: float64x1_t, d: float64x1_t, ptr: *mut i8); } @@ -26653,7 +26653,7 @@ pub unsafe fn vst4_lane_f64<const LANE: i32>(a: *mut f64, b: float64x1x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v1f64.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v1f64.p0" )] fn _vst4_lane_f64( a: float64x1_t, @@ -26680,7 +26680,7 @@ pub unsafe fn vst4_lane_s64<const LANE: i32>(a: *mut i64, b: int64x1x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v1i64.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v1i64.p0" )] fn _vst4_lane_s64( a: int64x1_t, @@ -26731,7 +26731,7 @@ pub unsafe fn vst4q_f64(a: *mut f64, b: float64x2x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v2f64.p0i8" + link_name = "llvm.aarch64.neon.st4.v2f64.p0" )] fn _vst4q_f64(a: float64x2_t, b: float64x2_t, c: float64x2_t, d: float64x2_t, ptr: *mut i8); } @@ -26749,7 +26749,7 @@ pub unsafe fn vst4q_s64(a: *mut i64, b: int64x2x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v2i64.p0i8" + link_name = "llvm.aarch64.neon.st4.v2i64.p0" )] fn _vst4q_s64(a: int64x2_t, b: int64x2_t, c: int64x2_t, d: int64x2_t, ptr: *mut i8); } @@ -26769,7 +26769,7 @@ pub unsafe fn vst4q_lane_f64<const LANE: i32>(a: *mut f64, b: float64x2x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v2f64.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v2f64.p0" )] fn _vst4q_lane_f64( a: float64x2_t, @@ -26796,7 +26796,7 @@ pub unsafe fn vst4q_lane_s8<const LANE: i32>(a: *mut i8, b: int8x16x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v16i8.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v16i8.p0" )] fn _vst4q_lane_s8( a: int8x16_t, @@ -26823,7 +26823,7 @@ pub unsafe fn vst4q_lane_s64<const LANE: i32>(a: *mut i64, b: int64x2x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v2i64.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v2i64.p0" )] fn _vst4q_lane_s64( a: int64x2_t, diff --git a/library/stdarch/crates/core_arch/src/arm_shared/neon/generated.rs b/library/stdarch/crates/core_arch/src/arm_shared/neon/generated.rs index 6e2f1ccf80e..286f1868852 100644 --- a/library/stdarch/crates/core_arch/src/arm_shared/neon/generated.rs +++ b/library/stdarch/crates/core_arch/src/arm_shared/neon/generated.rs @@ -14964,9 +14964,9 @@ pub unsafe fn vld1_f16_x2(a: *const f16) -> float16x4x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld1x2.v4f16.p0f16" + link_name = "llvm.aarch64.neon.ld1x2.v4f16.p0" )] - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld1x2.v4f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld1x2.v4f16.p0")] fn _vld1_f16_x2(a: *const f16) -> float16x4x2_t; } _vld1_f16_x2(a) @@ -14988,9 +14988,9 @@ pub unsafe fn vld1_f16_x3(a: *const f16) -> float16x4x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld1x3.v4f16.p0f16" + link_name = "llvm.aarch64.neon.ld1x3.v4f16.p0" )] - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld1x3.v4f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld1x3.v4f16.p0")] fn _vld1_f16_x3(a: *const f16) -> float16x4x3_t; } _vld1_f16_x3(a) @@ -15012,9 +15012,9 @@ pub unsafe fn vld1_f16_x4(a: *const f16) -> float16x4x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld1x4.v4f16.p0f16" + link_name = "llvm.aarch64.neon.ld1x4.v4f16.p0" )] - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld1x4.v4f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld1x4.v4f16.p0")] fn _vld1_f16_x4(a: *const f16) -> float16x4x4_t; } _vld1_f16_x4(a) @@ -15036,9 +15036,9 @@ pub unsafe fn vld1q_f16_x2(a: *const f16) -> float16x8x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld1x2.v8f16.p0f16" + link_name = "llvm.aarch64.neon.ld1x2.v8f16.p0" )] - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld1x2.v8f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld1x2.v8f16.p0")] fn _vld1q_f16_x2(a: *const f16) -> float16x8x2_t; } _vld1q_f16_x2(a) @@ -15060,9 +15060,9 @@ pub unsafe fn vld1q_f16_x3(a: *const f16) -> float16x8x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld1x3.v8f16.p0f16" + link_name = "llvm.aarch64.neon.ld1x3.v8f16.p0" )] - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld1x3.v8f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld1x3.v8f16.p0")] fn _vld1q_f16_x3(a: *const f16) -> float16x8x3_t; } _vld1q_f16_x3(a) @@ -15084,9 +15084,9 @@ pub unsafe fn vld1q_f16_x4(a: *const f16) -> float16x8x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld1x4.v8f16.p0f16" + link_name = "llvm.aarch64.neon.ld1x4.v8f16.p0" )] - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld1x4.v8f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld1x4.v8f16.p0")] fn _vld1q_f16_x4(a: *const f16) -> float16x8x4_t; } _vld1q_f16_x4(a) @@ -19594,7 +19594,7 @@ pub unsafe fn vld1q_dup_p64(ptr: *const p64) -> poly64x2_t { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vld2_dup_f16(a: *const f16) -> float16x4x2_t { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld2dup.v4f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld2dup.v4f16.p0")] fn _vld2_dup_f16(ptr: *const f16, size: i32) -> float16x4x2_t; } _vld2_dup_f16(a as _, 2) @@ -19611,7 +19611,7 @@ pub unsafe fn vld2_dup_f16(a: *const f16) -> float16x4x2_t { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vld2q_dup_f16(a: *const f16) -> float16x8x2_t { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld2dup.v8f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld2dup.v8f16.p0")] fn _vld2q_dup_f16(ptr: *const f16, size: i32) -> float16x8x2_t; } _vld2q_dup_f16(a as _, 2) @@ -19632,7 +19632,7 @@ pub unsafe fn vld2_dup_f16(a: *const f16) -> float16x4x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2r.v4f16.p0f16" + link_name = "llvm.aarch64.neon.ld2r.v4f16.p0" )] fn _vld2_dup_f16(ptr: *const f16) -> float16x4x2_t; } @@ -19654,7 +19654,7 @@ pub unsafe fn vld2q_dup_f16(a: *const f16) -> float16x8x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2r.v8f16.p0f16" + link_name = "llvm.aarch64.neon.ld2r.v8f16.p0" )] fn _vld2q_dup_f16(ptr: *const f16) -> float16x8x2_t; } @@ -20567,7 +20567,7 @@ pub unsafe fn vld2q_dup_p16(a: *const p16) -> poly16x8x2_t { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vld2_f16(a: *const f16) -> float16x4x2_t { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld2.v4f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld2.v4f16.p0")] fn _vld2_f16(ptr: *const f16, size: i32) -> float16x4x2_t; } _vld2_f16(a as _, 2) @@ -20584,7 +20584,7 @@ pub unsafe fn vld2_f16(a: *const f16) -> float16x4x2_t { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vld2q_f16(a: *const f16) -> float16x8x2_t { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld2.v8f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld2.v8f16.p0")] fn _vld2q_f16(ptr: *const f16, size: i32) -> float16x8x2_t; } _vld2q_f16(a as _, 2) @@ -20605,7 +20605,7 @@ pub unsafe fn vld2_f16(a: *const f16) -> float16x4x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2.v4f16.p0f16" + link_name = "llvm.aarch64.neon.ld2.v4f16.p0" )] fn _vld2_f16(ptr: *const f16) -> float16x4x2_t; } @@ -20627,7 +20627,7 @@ pub unsafe fn vld2q_f16(a: *const f16) -> float16x8x2_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld2.v8f16.p0f16" + link_name = "llvm.aarch64.neon.ld2.v8f16.p0" )] fn _vld2q_f16(ptr: *const f16) -> float16x8x2_t; } @@ -22155,7 +22155,7 @@ pub unsafe fn vld2q_p16(a: *const p16) -> poly16x8x2_t { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vld3_dup_f16(a: *const f16) -> float16x4x3_t { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld3dup.v4f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld3dup.v4f16.p0")] fn _vld3_dup_f16(ptr: *const f16, size: i32) -> float16x4x3_t; } _vld3_dup_f16(a as _, 2) @@ -22172,7 +22172,7 @@ pub unsafe fn vld3_dup_f16(a: *const f16) -> float16x4x3_t { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vld3q_dup_f16(a: *const f16) -> float16x8x3_t { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld3dup.v8f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld3dup.v8f16.p0")] fn _vld3q_dup_f16(ptr: *const f16, size: i32) -> float16x8x3_t; } _vld3q_dup_f16(a as _, 2) @@ -22193,7 +22193,7 @@ pub unsafe fn vld3_dup_f16(a: *const f16) -> float16x4x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3r.v4f16.p0f16" + link_name = "llvm.aarch64.neon.ld3r.v4f16.p0" )] fn _vld3_dup_f16(ptr: *const f16) -> float16x4x3_t; } @@ -22215,7 +22215,7 @@ pub unsafe fn vld3q_dup_f16(a: *const f16) -> float16x8x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3r.v8f16.p0f16" + link_name = "llvm.aarch64.neon.ld3r.v8f16.p0" )] fn _vld3q_dup_f16(ptr: *const f16) -> float16x8x3_t; } @@ -23150,7 +23150,7 @@ pub unsafe fn vld3q_dup_p16(a: *const p16) -> poly16x8x3_t { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vld3_f16(a: *const f16) -> float16x4x3_t { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld3.v4f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld3.v4f16.p0")] fn _vld3_f16(ptr: *const f16, size: i32) -> float16x4x3_t; } _vld3_f16(a as _, 2) @@ -23167,7 +23167,7 @@ pub unsafe fn vld3_f16(a: *const f16) -> float16x4x3_t { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vld3q_f16(a: *const f16) -> float16x8x3_t { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld3.v8f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld3.v8f16.p0")] fn _vld3q_f16(ptr: *const f16, size: i32) -> float16x8x3_t; } _vld3q_f16(a as _, 2) @@ -23188,7 +23188,7 @@ pub unsafe fn vld3_f16(a: *const f16) -> float16x4x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3.v4f16.p0f16" + link_name = "llvm.aarch64.neon.ld3.v4f16.p0" )] fn _vld3_f16(ptr: *const f16) -> float16x4x3_t; } @@ -23210,7 +23210,7 @@ pub unsafe fn vld3q_f16(a: *const f16) -> float16x8x3_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld3.v8f16.p0f16" + link_name = "llvm.aarch64.neon.ld3.v8f16.p0" )] fn _vld3q_f16(ptr: *const f16) -> float16x8x3_t; } @@ -24821,7 +24821,7 @@ pub unsafe fn vld3q_lane_f32<const LANE: i32>(a: *const f32, b: float32x4x3_t) - #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vld4_dup_f16(a: *const f16) -> float16x4x4_t { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld4dup.v4f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld4dup.v4f16.p0")] fn _vld4_dup_f16(ptr: *const f16, size: i32) -> float16x4x4_t; } _vld4_dup_f16(a as _, 2) @@ -24838,7 +24838,7 @@ pub unsafe fn vld4_dup_f16(a: *const f16) -> float16x4x4_t { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vld4q_dup_f16(a: *const f16) -> float16x8x4_t { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld4dup.v8f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld4dup.v8f16.p0")] fn _vld4q_dup_f16(ptr: *const f16, size: i32) -> float16x8x4_t; } _vld4q_dup_f16(a as _, 2) @@ -24859,7 +24859,7 @@ pub unsafe fn vld4_dup_f16(a: *const f16) -> float16x4x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v4f16.p0f16" + link_name = "llvm.aarch64.neon.ld4r.v4f16.p0" )] fn _vld4_dup_f16(ptr: *const f16) -> float16x4x4_t; } @@ -24881,7 +24881,7 @@ pub unsafe fn vld4q_dup_f16(a: *const f16) -> float16x8x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v8f16.p0f16" + link_name = "llvm.aarch64.neon.ld4r.v8f16.p0" )] fn _vld4q_dup_f16(ptr: *const f16) -> float16x8x4_t; } @@ -25028,7 +25028,7 @@ pub unsafe fn vld4_dup_f32(a: *const f32) -> float32x2x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v2f32.p0f32.p0" + link_name = "llvm.aarch64.neon.ld4r.v2f32.p0.p0" )] fn _vld4_dup_f32(ptr: *const f32) -> float32x2x4_t; } @@ -25047,7 +25047,7 @@ pub unsafe fn vld4q_dup_f32(a: *const f32) -> float32x4x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v4f32.p0f32.p0" + link_name = "llvm.aarch64.neon.ld4r.v4f32.p0.p0" )] fn _vld4q_dup_f32(ptr: *const f32) -> float32x4x4_t; } @@ -25066,7 +25066,7 @@ pub unsafe fn vld4_dup_s8(a: *const i8) -> int8x8x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v8i8.p0i8.p0" + link_name = "llvm.aarch64.neon.ld4r.v8i8.p0.p0" )] fn _vld4_dup_s8(ptr: *const i8) -> int8x8x4_t; } @@ -25085,7 +25085,7 @@ pub unsafe fn vld4q_dup_s8(a: *const i8) -> int8x16x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v16i8.p0i8.p0" + link_name = "llvm.aarch64.neon.ld4r.v16i8.p0.p0" )] fn _vld4q_dup_s8(ptr: *const i8) -> int8x16x4_t; } @@ -25104,7 +25104,7 @@ pub unsafe fn vld4_dup_s16(a: *const i16) -> int16x4x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v4i16.p0i16.p0" + link_name = "llvm.aarch64.neon.ld4r.v4i16.p0.p0" )] fn _vld4_dup_s16(ptr: *const i16) -> int16x4x4_t; } @@ -25123,7 +25123,7 @@ pub unsafe fn vld4q_dup_s16(a: *const i16) -> int16x8x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v8i16.p0i16.p0" + link_name = "llvm.aarch64.neon.ld4r.v8i16.p0.p0" )] fn _vld4q_dup_s16(ptr: *const i16) -> int16x8x4_t; } @@ -25142,7 +25142,7 @@ pub unsafe fn vld4_dup_s32(a: *const i32) -> int32x2x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v2i32.p0i32.p0" + link_name = "llvm.aarch64.neon.ld4r.v2i32.p0.p0" )] fn _vld4_dup_s32(ptr: *const i32) -> int32x2x4_t; } @@ -25161,7 +25161,7 @@ pub unsafe fn vld4q_dup_s32(a: *const i32) -> int32x4x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v4i32.p0i32.p0" + link_name = "llvm.aarch64.neon.ld4r.v4i32.p0.p0" )] fn _vld4q_dup_s32(ptr: *const i32) -> int32x4x4_t; } @@ -25180,7 +25180,7 @@ pub unsafe fn vld4_dup_s64(a: *const i64) -> int64x1x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4r.v1i64.p0i64.p0" + link_name = "llvm.aarch64.neon.ld4r.v1i64.p0.p0" )] fn _vld4_dup_s64(ptr: *const i64) -> int64x1x4_t; } @@ -25838,7 +25838,7 @@ pub unsafe fn vld4q_dup_p16(a: *const p16) -> poly16x8x4_t { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vld4_f16(a: *const f16) -> float16x4x4_t { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld4.v4f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld4.v4f16.p0")] fn _vld4_f16(ptr: *const f16, size: i32) -> float16x4x4_t; } _vld4_f16(a as _, 2) @@ -25855,7 +25855,7 @@ pub unsafe fn vld4_f16(a: *const f16) -> float16x4x4_t { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vld4q_f16(a: *const f16) -> float16x8x4_t { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld4.v8f16.p0f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vld4.v8f16.p0")] fn _vld4q_f16(ptr: *const f16, size: i32) -> float16x8x4_t; } _vld4q_f16(a as _, 2) @@ -25876,7 +25876,7 @@ pub unsafe fn vld4_f16(a: *const f16) -> float16x4x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4.v4f16.p0f16" + link_name = "llvm.aarch64.neon.ld4.v4f16.p0" )] fn _vld4_f16(ptr: *const f16) -> float16x4x4_t; } @@ -25898,7 +25898,7 @@ pub unsafe fn vld4q_f16(a: *const f16) -> float16x8x4_t { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.ld4.v8f16.p0f16" + link_name = "llvm.aarch64.neon.ld4.v8f16.p0" )] fn _vld4q_f16(ptr: *const f16) -> float16x8x4_t; } @@ -64115,7 +64115,7 @@ pub unsafe fn vst1q_f16(ptr: *mut f16, a: float16x8_t) { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vst1_f16_x2(a: *mut f16, b: float16x4x2_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x2.p0f16.v4f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x2.p0.v4f16")] fn _vst1_f16_x2(ptr: *mut f16, a: float16x4_t, b: float16x4_t); } _vst1_f16_x2(a, b.0, b.1) @@ -64132,7 +64132,7 @@ pub unsafe fn vst1_f16_x2(a: *mut f16, b: float16x4x2_t) { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vst1q_f16_x2(a: *mut f16, b: float16x8x2_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x2.p0f16.v8f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x2.p0.v8f16")] fn _vst1q_f16_x2(ptr: *mut f16, a: float16x8_t, b: float16x8_t); } _vst1q_f16_x2(a, b.0, b.1) @@ -64150,7 +64150,7 @@ pub unsafe fn vst1_f16_x2(a: *mut f16, b: float16x4x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v4f16.p0f16" + link_name = "llvm.aarch64.neon.st1x2.v4f16.p0" )] fn _vst1_f16_x2(a: float16x4_t, b: float16x4_t, ptr: *mut f16); } @@ -64169,7 +64169,7 @@ pub unsafe fn vst1q_f16_x2(a: *mut f16, b: float16x8x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v8f16.p0f16" + link_name = "llvm.aarch64.neon.st1x2.v8f16.p0" )] fn _vst1q_f16_x2(a: float16x8_t, b: float16x8_t, ptr: *mut f16); } @@ -64187,7 +64187,7 @@ pub unsafe fn vst1q_f16_x2(a: *mut f16, b: float16x8x2_t) { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vst1_f16_x3(a: *mut f16, b: float16x4x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0f16.v4f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0.v4f16")] fn _vst1_f16_x3(ptr: *mut f16, a: float16x4_t, b: float16x4_t, c: float16x4_t); } _vst1_f16_x3(a, b.0, b.1, b.2) @@ -64204,7 +64204,7 @@ pub unsafe fn vst1_f16_x3(a: *mut f16, b: float16x4x3_t) { #[unstable(feature = "stdarch_neon_f16", issue = "136306")] pub unsafe fn vst1q_f16_x3(a: *mut f16, b: float16x8x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0f16.v8f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0.v8f16")] fn _vst1q_f16_x3(ptr: *mut f16, a: float16x8_t, b: float16x8_t, c: float16x8_t); } _vst1q_f16_x3(a, b.0, b.1, b.2) @@ -64222,7 +64222,7 @@ pub unsafe fn vst1_f16_x3(a: *mut f16, b: float16x4x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v4f16.p0f16" + link_name = "llvm.aarch64.neon.st1x3.v4f16.p0" )] fn _vst1_f16_x3(a: float16x4_t, b: float16x4_t, c: float16x4_t, ptr: *mut f16); } @@ -64241,7 +64241,7 @@ pub unsafe fn vst1q_f16_x3(a: *mut f16, b: float16x8x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v8f16.p0f16" + link_name = "llvm.aarch64.neon.st1x3.v8f16.p0" )] fn _vst1q_f16_x3(a: float16x8_t, b: float16x8_t, c: float16x8_t, ptr: *mut f16); } @@ -64259,7 +64259,7 @@ pub unsafe fn vst1q_f16_x3(a: *mut f16, b: float16x8x3_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1_f16_x4(a: *mut f16, b: float16x4x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0f16.v4f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0.v4f16")] fn _vst1_f16_x4( ptr: *mut f16, a: float16x4_t, @@ -64282,7 +64282,7 @@ pub unsafe fn vst1_f16_x4(a: *mut f16, b: float16x4x4_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1q_f16_x4(a: *mut f16, b: float16x8x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0f16.v8f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0.v8f16")] fn _vst1q_f16_x4( ptr: *mut f16, a: float16x8_t, @@ -64306,7 +64306,7 @@ pub unsafe fn vst1_f16_x4(a: *mut f16, b: float16x4x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v4f16.p0f16" + link_name = "llvm.aarch64.neon.st1x4.v4f16.p0" )] fn _vst1_f16_x4( a: float16x4_t, @@ -64331,7 +64331,7 @@ pub unsafe fn vst1q_f16_x4(a: *mut f16, b: float16x8x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v8f16.p0f16" + link_name = "llvm.aarch64.neon.st1x4.v8f16.p0" )] fn _vst1q_f16_x4( a: float16x8_t, @@ -64764,7 +64764,7 @@ pub unsafe fn vst1_f32_x2(a: *mut f32, b: float32x2x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v2f32.p0f32" + link_name = "llvm.aarch64.neon.st1x2.v2f32.p0" )] fn _vst1_f32_x2(a: float32x2_t, b: float32x2_t, ptr: *mut f32); } @@ -64783,7 +64783,7 @@ pub unsafe fn vst1q_f32_x2(a: *mut f32, b: float32x4x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v4f32.p0f32" + link_name = "llvm.aarch64.neon.st1x2.v4f32.p0" )] fn _vst1q_f32_x2(a: float32x4_t, b: float32x4_t, ptr: *mut f32); } @@ -64802,7 +64802,7 @@ pub unsafe fn vst1_f32_x3(a: *mut f32, b: float32x2x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v2f32.p0f32" + link_name = "llvm.aarch64.neon.st1x3.v2f32.p0" )] fn _vst1_f32_x3(a: float32x2_t, b: float32x2_t, c: float32x2_t, ptr: *mut f32); } @@ -64821,7 +64821,7 @@ pub unsafe fn vst1q_f32_x3(a: *mut f32, b: float32x4x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v4f32.p0f32" + link_name = "llvm.aarch64.neon.st1x3.v4f32.p0" )] fn _vst1q_f32_x3(a: float32x4_t, b: float32x4_t, c: float32x4_t, ptr: *mut f32); } @@ -64838,7 +64838,7 @@ pub unsafe fn vst1q_f32_x3(a: *mut f32, b: float32x4x3_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1_f32_x4(a: *mut f32, b: float32x2x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0f32.v2f32.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0.v2f32.p0")] fn _vst1_f32_x4( ptr: *mut f32, a: float32x2_t, @@ -64860,7 +64860,7 @@ pub unsafe fn vst1_f32_x4(a: *mut f32, b: float32x2x4_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1q_f32_x4(a: *mut f32, b: float32x4x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0f32.v4f32.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0.v4f32.p0")] fn _vst1q_f32_x4( ptr: *mut f32, a: float32x4_t, @@ -64884,7 +64884,7 @@ pub unsafe fn vst1_f32_x4(a: *mut f32, b: float32x2x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v2f32.p0f32" + link_name = "llvm.aarch64.neon.st1x4.v2f32.p0" )] fn _vst1_f32_x4( a: float32x2_t, @@ -64909,7 +64909,7 @@ pub unsafe fn vst1q_f32_x4(a: *mut f32, b: float32x4x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v4f32.p0f32" + link_name = "llvm.aarch64.neon.st1x4.v4f32.p0" )] fn _vst1q_f32_x4( a: float32x4_t, @@ -65683,7 +65683,7 @@ pub unsafe fn vst1_s8_x2(a: *mut i8, b: int8x8x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v8i8.p0i8" + link_name = "llvm.aarch64.neon.st1x2.v8i8.p0" )] fn _vst1_s8_x2(a: int8x8_t, b: int8x8_t, ptr: *mut i8); } @@ -65702,7 +65702,7 @@ pub unsafe fn vst1q_s8_x2(a: *mut i8, b: int8x16x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v16i8.p0i8" + link_name = "llvm.aarch64.neon.st1x2.v16i8.p0" )] fn _vst1q_s8_x2(a: int8x16_t, b: int8x16_t, ptr: *mut i8); } @@ -65721,7 +65721,7 @@ pub unsafe fn vst1_s16_x2(a: *mut i16, b: int16x4x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v4i16.p0i16" + link_name = "llvm.aarch64.neon.st1x2.v4i16.p0" )] fn _vst1_s16_x2(a: int16x4_t, b: int16x4_t, ptr: *mut i16); } @@ -65740,7 +65740,7 @@ pub unsafe fn vst1q_s16_x2(a: *mut i16, b: int16x8x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v8i16.p0i16" + link_name = "llvm.aarch64.neon.st1x2.v8i16.p0" )] fn _vst1q_s16_x2(a: int16x8_t, b: int16x8_t, ptr: *mut i16); } @@ -65759,7 +65759,7 @@ pub unsafe fn vst1_s32_x2(a: *mut i32, b: int32x2x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v2i32.p0i32" + link_name = "llvm.aarch64.neon.st1x2.v2i32.p0" )] fn _vst1_s32_x2(a: int32x2_t, b: int32x2_t, ptr: *mut i32); } @@ -65778,7 +65778,7 @@ pub unsafe fn vst1q_s32_x2(a: *mut i32, b: int32x4x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v4i32.p0i32" + link_name = "llvm.aarch64.neon.st1x2.v4i32.p0" )] fn _vst1q_s32_x2(a: int32x4_t, b: int32x4_t, ptr: *mut i32); } @@ -65797,7 +65797,7 @@ pub unsafe fn vst1_s64_x2(a: *mut i64, b: int64x1x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v1i64.p0i64" + link_name = "llvm.aarch64.neon.st1x2.v1i64.p0" )] fn _vst1_s64_x2(a: int64x1_t, b: int64x1_t, ptr: *mut i64); } @@ -65816,7 +65816,7 @@ pub unsafe fn vst1q_s64_x2(a: *mut i64, b: int64x2x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x2.v2i64.p0i64" + link_name = "llvm.aarch64.neon.st1x2.v2i64.p0" )] fn _vst1q_s64_x2(a: int64x2_t, b: int64x2_t, ptr: *mut i64); } @@ -65963,7 +65963,7 @@ pub unsafe fn vst1_s8_x3(a: *mut i8, b: int8x8x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v8i8.p0i8" + link_name = "llvm.aarch64.neon.st1x3.v8i8.p0" )] fn _vst1_s8_x3(a: int8x8_t, b: int8x8_t, c: int8x8_t, ptr: *mut i8); } @@ -65982,7 +65982,7 @@ pub unsafe fn vst1q_s8_x3(a: *mut i8, b: int8x16x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v16i8.p0i8" + link_name = "llvm.aarch64.neon.st1x3.v16i8.p0" )] fn _vst1q_s8_x3(a: int8x16_t, b: int8x16_t, c: int8x16_t, ptr: *mut i8); } @@ -66001,7 +66001,7 @@ pub unsafe fn vst1_s16_x3(a: *mut i16, b: int16x4x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v4i16.p0i16" + link_name = "llvm.aarch64.neon.st1x3.v4i16.p0" )] fn _vst1_s16_x3(a: int16x4_t, b: int16x4_t, c: int16x4_t, ptr: *mut i16); } @@ -66020,7 +66020,7 @@ pub unsafe fn vst1q_s16_x3(a: *mut i16, b: int16x8x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v8i16.p0i16" + link_name = "llvm.aarch64.neon.st1x3.v8i16.p0" )] fn _vst1q_s16_x3(a: int16x8_t, b: int16x8_t, c: int16x8_t, ptr: *mut i16); } @@ -66039,7 +66039,7 @@ pub unsafe fn vst1_s32_x3(a: *mut i32, b: int32x2x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v2i32.p0i32" + link_name = "llvm.aarch64.neon.st1x3.v2i32.p0" )] fn _vst1_s32_x3(a: int32x2_t, b: int32x2_t, c: int32x2_t, ptr: *mut i32); } @@ -66058,7 +66058,7 @@ pub unsafe fn vst1q_s32_x3(a: *mut i32, b: int32x4x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v4i32.p0i32" + link_name = "llvm.aarch64.neon.st1x3.v4i32.p0" )] fn _vst1q_s32_x3(a: int32x4_t, b: int32x4_t, c: int32x4_t, ptr: *mut i32); } @@ -66077,7 +66077,7 @@ pub unsafe fn vst1_s64_x3(a: *mut i64, b: int64x1x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v1i64.p0i64" + link_name = "llvm.aarch64.neon.st1x3.v1i64.p0" )] fn _vst1_s64_x3(a: int64x1_t, b: int64x1_t, c: int64x1_t, ptr: *mut i64); } @@ -66096,7 +66096,7 @@ pub unsafe fn vst1q_s64_x3(a: *mut i64, b: int64x2x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x3.v2i64.p0i64" + link_name = "llvm.aarch64.neon.st1x3.v2i64.p0" )] fn _vst1q_s64_x3(a: int64x2_t, b: int64x2_t, c: int64x2_t, ptr: *mut i64); } @@ -66113,7 +66113,7 @@ pub unsafe fn vst1q_s64_x3(a: *mut i64, b: int64x2x3_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1_s8_x3(a: *mut i8, b: int8x8x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0i8.v8i8.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0.v8i8.p0")] fn _vst1_s8_x3(ptr: *mut i8, a: int8x8_t, b: int8x8_t, c: int8x8_t); } _vst1_s8_x3(a, b.0, b.1, b.2) @@ -66129,7 +66129,7 @@ pub unsafe fn vst1_s8_x3(a: *mut i8, b: int8x8x3_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1q_s8_x3(a: *mut i8, b: int8x16x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0i8.v16i8.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0.v16i8.p0")] fn _vst1q_s8_x3(ptr: *mut i8, a: int8x16_t, b: int8x16_t, c: int8x16_t); } _vst1q_s8_x3(a, b.0, b.1, b.2) @@ -66145,7 +66145,7 @@ pub unsafe fn vst1q_s8_x3(a: *mut i8, b: int8x16x3_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1_s16_x3(a: *mut i16, b: int16x4x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0i16.v4i16.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0.v4i16.p0")] fn _vst1_s16_x3(ptr: *mut i16, a: int16x4_t, b: int16x4_t, c: int16x4_t); } _vst1_s16_x3(a, b.0, b.1, b.2) @@ -66161,7 +66161,7 @@ pub unsafe fn vst1_s16_x3(a: *mut i16, b: int16x4x3_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1q_s16_x3(a: *mut i16, b: int16x8x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0i16.v8i16.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0.v8i16.p0")] fn _vst1q_s16_x3(ptr: *mut i16, a: int16x8_t, b: int16x8_t, c: int16x8_t); } _vst1q_s16_x3(a, b.0, b.1, b.2) @@ -66177,7 +66177,7 @@ pub unsafe fn vst1q_s16_x3(a: *mut i16, b: int16x8x3_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1_s32_x3(a: *mut i32, b: int32x2x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0i32.v2i32.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0.v2i32.p0")] fn _vst1_s32_x3(ptr: *mut i32, a: int32x2_t, b: int32x2_t, c: int32x2_t); } _vst1_s32_x3(a, b.0, b.1, b.2) @@ -66193,7 +66193,7 @@ pub unsafe fn vst1_s32_x3(a: *mut i32, b: int32x2x3_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1q_s32_x3(a: *mut i32, b: int32x4x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0i32.v4i32.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0.v4i32.p0")] fn _vst1q_s32_x3(ptr: *mut i32, a: int32x4_t, b: int32x4_t, c: int32x4_t); } _vst1q_s32_x3(a, b.0, b.1, b.2) @@ -66209,7 +66209,7 @@ pub unsafe fn vst1q_s32_x3(a: *mut i32, b: int32x4x3_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1_s64_x3(a: *mut i64, b: int64x1x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0i64.v1i64.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0.v1i64.p0")] fn _vst1_s64_x3(ptr: *mut i64, a: int64x1_t, b: int64x1_t, c: int64x1_t); } _vst1_s64_x3(a, b.0, b.1, b.2) @@ -66225,7 +66225,7 @@ pub unsafe fn vst1_s64_x3(a: *mut i64, b: int64x1x3_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1q_s64_x3(a: *mut i64, b: int64x2x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0i64.v2i64.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x3.p0.v2i64.p0")] fn _vst1q_s64_x3(ptr: *mut i64, a: int64x2_t, b: int64x2_t, c: int64x2_t); } _vst1q_s64_x3(a, b.0, b.1, b.2) @@ -66243,7 +66243,7 @@ pub unsafe fn vst1_s8_x4(a: *mut i8, b: int8x8x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v8i8.p0i8" + link_name = "llvm.aarch64.neon.st1x4.v8i8.p0" )] fn _vst1_s8_x4(a: int8x8_t, b: int8x8_t, c: int8x8_t, d: int8x8_t, ptr: *mut i8); } @@ -66262,7 +66262,7 @@ pub unsafe fn vst1q_s8_x4(a: *mut i8, b: int8x16x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v16i8.p0i8" + link_name = "llvm.aarch64.neon.st1x4.v16i8.p0" )] fn _vst1q_s8_x4(a: int8x16_t, b: int8x16_t, c: int8x16_t, d: int8x16_t, ptr: *mut i8); } @@ -66281,7 +66281,7 @@ pub unsafe fn vst1_s16_x4(a: *mut i16, b: int16x4x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v4i16.p0i16" + link_name = "llvm.aarch64.neon.st1x4.v4i16.p0" )] fn _vst1_s16_x4(a: int16x4_t, b: int16x4_t, c: int16x4_t, d: int16x4_t, ptr: *mut i16); } @@ -66300,7 +66300,7 @@ pub unsafe fn vst1q_s16_x4(a: *mut i16, b: int16x8x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v8i16.p0i16" + link_name = "llvm.aarch64.neon.st1x4.v8i16.p0" )] fn _vst1q_s16_x4(a: int16x8_t, b: int16x8_t, c: int16x8_t, d: int16x8_t, ptr: *mut i16); } @@ -66319,7 +66319,7 @@ pub unsafe fn vst1_s32_x4(a: *mut i32, b: int32x2x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v2i32.p0i32" + link_name = "llvm.aarch64.neon.st1x4.v2i32.p0" )] fn _vst1_s32_x4(a: int32x2_t, b: int32x2_t, c: int32x2_t, d: int32x2_t, ptr: *mut i32); } @@ -66338,7 +66338,7 @@ pub unsafe fn vst1q_s32_x4(a: *mut i32, b: int32x4x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v4i32.p0i32" + link_name = "llvm.aarch64.neon.st1x4.v4i32.p0" )] fn _vst1q_s32_x4(a: int32x4_t, b: int32x4_t, c: int32x4_t, d: int32x4_t, ptr: *mut i32); } @@ -66357,7 +66357,7 @@ pub unsafe fn vst1_s64_x4(a: *mut i64, b: int64x1x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v1i64.p0i64" + link_name = "llvm.aarch64.neon.st1x4.v1i64.p0" )] fn _vst1_s64_x4(a: int64x1_t, b: int64x1_t, c: int64x1_t, d: int64x1_t, ptr: *mut i64); } @@ -66376,7 +66376,7 @@ pub unsafe fn vst1q_s64_x4(a: *mut i64, b: int64x2x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st1x4.v2i64.p0i64" + link_name = "llvm.aarch64.neon.st1x4.v2i64.p0" )] fn _vst1q_s64_x4(a: int64x2_t, b: int64x2_t, c: int64x2_t, d: int64x2_t, ptr: *mut i64); } @@ -66393,7 +66393,7 @@ pub unsafe fn vst1q_s64_x4(a: *mut i64, b: int64x2x4_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1_s8_x4(a: *mut i8, b: int8x8x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0i8.v8i8.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0.v8i8.p0")] fn _vst1_s8_x4(ptr: *mut i8, a: int8x8_t, b: int8x8_t, c: int8x8_t, d: int8x8_t); } _vst1_s8_x4(a, b.0, b.1, b.2, b.3) @@ -66409,7 +66409,7 @@ pub unsafe fn vst1_s8_x4(a: *mut i8, b: int8x8x4_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1q_s8_x4(a: *mut i8, b: int8x16x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0i8.v16i8.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0.v16i8.p0")] fn _vst1q_s8_x4(ptr: *mut i8, a: int8x16_t, b: int8x16_t, c: int8x16_t, d: int8x16_t); } _vst1q_s8_x4(a, b.0, b.1, b.2, b.3) @@ -66425,7 +66425,7 @@ pub unsafe fn vst1q_s8_x4(a: *mut i8, b: int8x16x4_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1_s16_x4(a: *mut i16, b: int16x4x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0i16.v4i16.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0.v4i16.p0")] fn _vst1_s16_x4(ptr: *mut i16, a: int16x4_t, b: int16x4_t, c: int16x4_t, d: int16x4_t); } _vst1_s16_x4(a, b.0, b.1, b.2, b.3) @@ -66441,7 +66441,7 @@ pub unsafe fn vst1_s16_x4(a: *mut i16, b: int16x4x4_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1q_s16_x4(a: *mut i16, b: int16x8x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0i16.v8i16.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0.v8i16.p0")] fn _vst1q_s16_x4(ptr: *mut i16, a: int16x8_t, b: int16x8_t, c: int16x8_t, d: int16x8_t); } _vst1q_s16_x4(a, b.0, b.1, b.2, b.3) @@ -66457,7 +66457,7 @@ pub unsafe fn vst1q_s16_x4(a: *mut i16, b: int16x8x4_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1_s32_x4(a: *mut i32, b: int32x2x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0i32.v2i32.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0.v2i32.p0")] fn _vst1_s32_x4(ptr: *mut i32, a: int32x2_t, b: int32x2_t, c: int32x2_t, d: int32x2_t); } _vst1_s32_x4(a, b.0, b.1, b.2, b.3) @@ -66473,7 +66473,7 @@ pub unsafe fn vst1_s32_x4(a: *mut i32, b: int32x2x4_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1q_s32_x4(a: *mut i32, b: int32x4x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0i32.v4i32.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0.v4i32.p0")] fn _vst1q_s32_x4(ptr: *mut i32, a: int32x4_t, b: int32x4_t, c: int32x4_t, d: int32x4_t); } _vst1q_s32_x4(a, b.0, b.1, b.2, b.3) @@ -66489,7 +66489,7 @@ pub unsafe fn vst1q_s32_x4(a: *mut i32, b: int32x4x4_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1_s64_x4(a: *mut i64, b: int64x1x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0i64.v1i64.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0.v1i64.p0")] fn _vst1_s64_x4(ptr: *mut i64, a: int64x1_t, b: int64x1_t, c: int64x1_t, d: int64x1_t); } _vst1_s64_x4(a, b.0, b.1, b.2, b.3) @@ -66505,7 +66505,7 @@ pub unsafe fn vst1_s64_x4(a: *mut i64, b: int64x1x4_t) { #[cfg_attr(test, assert_instr(vst1))] pub unsafe fn vst1q_s64_x4(a: *mut i64, b: int64x2x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0i64.v2i64.p0")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst1x4.p0.v2i64.p0")] fn _vst1q_s64_x4(ptr: *mut i64, a: int64x2_t, b: int64x2_t, c: int64x2_t, d: int64x2_t); } _vst1q_s64_x4(a, b.0, b.1, b.2, b.3) @@ -67580,7 +67580,7 @@ pub unsafe fn vst2_f16(a: *mut f16, b: float16x4x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v4f16.p0i8" + link_name = "llvm.aarch64.neon.st2.v4f16.p0" )] fn _vst2_f16(a: float16x4_t, b: float16x4_t, ptr: *mut i8); } @@ -67599,7 +67599,7 @@ pub unsafe fn vst2q_f16(a: *mut f16, b: float16x8x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v8f16.p0i8" + link_name = "llvm.aarch64.neon.st2.v8f16.p0" )] fn _vst2q_f16(a: float16x8_t, b: float16x8_t, ptr: *mut i8); } @@ -67617,7 +67617,7 @@ pub unsafe fn vst2q_f16(a: *mut f16, b: float16x8x2_t) { #[cfg_attr(test, assert_instr(vst2))] pub unsafe fn vst2_f16(a: *mut f16, b: float16x4x2_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst2.p0i8.v4f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst2.p0.v4f16")] fn _vst2_f16(ptr: *mut i8, a: float16x4_t, b: float16x4_t, size: i32); } _vst2_f16(a as _, b.0, b.1, 2) @@ -67634,7 +67634,7 @@ pub unsafe fn vst2_f16(a: *mut f16, b: float16x4x2_t) { #[cfg_attr(test, assert_instr(vst2))] pub unsafe fn vst2q_f16(a: *mut f16, b: float16x8x2_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst2.p0i8.v8f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst2.p0.v8f16")] fn _vst2q_f16(ptr: *mut i8, a: float16x8_t, b: float16x8_t, size: i32); } _vst2q_f16(a as _, b.0, b.1, 2) @@ -67652,7 +67652,7 @@ pub unsafe fn vst2_f32(a: *mut f32, b: float32x2x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v2f32.p0i8" + link_name = "llvm.aarch64.neon.st2.v2f32.p0" )] fn _vst2_f32(a: float32x2_t, b: float32x2_t, ptr: *mut i8); } @@ -67671,7 +67671,7 @@ pub unsafe fn vst2q_f32(a: *mut f32, b: float32x4x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v4f32.p0i8" + link_name = "llvm.aarch64.neon.st2.v4f32.p0" )] fn _vst2q_f32(a: float32x4_t, b: float32x4_t, ptr: *mut i8); } @@ -67690,7 +67690,7 @@ pub unsafe fn vst2_s8(a: *mut i8, b: int8x8x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v8i8.p0i8" + link_name = "llvm.aarch64.neon.st2.v8i8.p0" )] fn _vst2_s8(a: int8x8_t, b: int8x8_t, ptr: *mut i8); } @@ -67709,7 +67709,7 @@ pub unsafe fn vst2q_s8(a: *mut i8, b: int8x16x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v16i8.p0i8" + link_name = "llvm.aarch64.neon.st2.v16i8.p0" )] fn _vst2q_s8(a: int8x16_t, b: int8x16_t, ptr: *mut i8); } @@ -67728,7 +67728,7 @@ pub unsafe fn vst2_s16(a: *mut i16, b: int16x4x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v4i16.p0i8" + link_name = "llvm.aarch64.neon.st2.v4i16.p0" )] fn _vst2_s16(a: int16x4_t, b: int16x4_t, ptr: *mut i8); } @@ -67747,7 +67747,7 @@ pub unsafe fn vst2q_s16(a: *mut i16, b: int16x8x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v8i16.p0i8" + link_name = "llvm.aarch64.neon.st2.v8i16.p0" )] fn _vst2q_s16(a: int16x8_t, b: int16x8_t, ptr: *mut i8); } @@ -67766,7 +67766,7 @@ pub unsafe fn vst2_s32(a: *mut i32, b: int32x2x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v2i32.p0i8" + link_name = "llvm.aarch64.neon.st2.v2i32.p0" )] fn _vst2_s32(a: int32x2_t, b: int32x2_t, ptr: *mut i8); } @@ -67785,7 +67785,7 @@ pub unsafe fn vst2q_s32(a: *mut i32, b: int32x4x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v4i32.p0i8" + link_name = "llvm.aarch64.neon.st2.v4i32.p0" )] fn _vst2q_s32(a: int32x4_t, b: int32x4_t, ptr: *mut i8); } @@ -67934,7 +67934,7 @@ pub unsafe fn vst2_lane_f16<const LANE: i32>(a: *mut f16, b: float16x4x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v4f16.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v4f16.p0" )] fn _vst2_lane_f16(a: float16x4_t, b: float16x4_t, n: i64, ptr: *mut i8); } @@ -67955,7 +67955,7 @@ pub unsafe fn vst2q_lane_f16<const LANE: i32>(a: *mut f16, b: float16x8x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v8f16.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v8f16.p0" )] fn _vst2q_lane_f16(a: float16x8_t, b: float16x8_t, n: i64, ptr: *mut i8); } @@ -67975,7 +67975,7 @@ pub unsafe fn vst2q_lane_f16<const LANE: i32>(a: *mut f16, b: float16x8x2_t) { pub unsafe fn vst2_lane_f16<const LANE: i32>(a: *mut f16, b: float16x4x2_t) { static_assert_uimm_bits!(LANE, 2); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst2lane.p0i8.v4f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst2lane.p0.v4f16")] fn _vst2_lane_f16(ptr: *mut i8, a: float16x4_t, b: float16x4_t, n: i32, size: i32); } _vst2_lane_f16(a as _, b.0, b.1, LANE, 2) @@ -67994,7 +67994,7 @@ pub unsafe fn vst2_lane_f16<const LANE: i32>(a: *mut f16, b: float16x4x2_t) { pub unsafe fn vst2q_lane_f16<const LANE: i32>(a: *mut f16, b: float16x8x2_t) { static_assert_uimm_bits!(LANE, 1); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst2lane.p0i8.v8f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst2lane.p0.v8f16")] fn _vst2q_lane_f16(ptr: *mut i8, a: float16x8_t, b: float16x8_t, n: i32, size: i32); } _vst2q_lane_f16(a as _, b.0, b.1, LANE, 2) @@ -68014,7 +68014,7 @@ pub unsafe fn vst2_lane_f32<const LANE: i32>(a: *mut f32, b: float32x2x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v2f32.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v2f32.p0" )] fn _vst2_lane_f32(a: float32x2_t, b: float32x2_t, n: i64, ptr: *mut i8); } @@ -68035,7 +68035,7 @@ pub unsafe fn vst2q_lane_f32<const LANE: i32>(a: *mut f32, b: float32x4x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v4f32.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v4f32.p0" )] fn _vst2q_lane_f32(a: float32x4_t, b: float32x4_t, n: i64, ptr: *mut i8); } @@ -68056,7 +68056,7 @@ pub unsafe fn vst2_lane_s8<const LANE: i32>(a: *mut i8, b: int8x8x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v8i8.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v8i8.p0" )] fn _vst2_lane_s8(a: int8x8_t, b: int8x8_t, n: i64, ptr: *mut i8); } @@ -68077,7 +68077,7 @@ pub unsafe fn vst2_lane_s16<const LANE: i32>(a: *mut i16, b: int16x4x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v4i16.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v4i16.p0" )] fn _vst2_lane_s16(a: int16x4_t, b: int16x4_t, n: i64, ptr: *mut i8); } @@ -68098,7 +68098,7 @@ pub unsafe fn vst2q_lane_s16<const LANE: i32>(a: *mut i16, b: int16x8x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v8i16.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v8i16.p0" )] fn _vst2q_lane_s16(a: int16x8_t, b: int16x8_t, n: i64, ptr: *mut i8); } @@ -68119,7 +68119,7 @@ pub unsafe fn vst2_lane_s32<const LANE: i32>(a: *mut i32, b: int32x2x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v2i32.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v2i32.p0" )] fn _vst2_lane_s32(a: int32x2_t, b: int32x2_t, n: i64, ptr: *mut i8); } @@ -68140,7 +68140,7 @@ pub unsafe fn vst2q_lane_s32<const LANE: i32>(a: *mut i32, b: int32x4x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2lane.v4i32.p0i8" + link_name = "llvm.aarch64.neon.st2lane.v4i32.p0" )] fn _vst2q_lane_s32(a: int32x4_t, b: int32x4_t, n: i64, ptr: *mut i8); } @@ -68524,7 +68524,7 @@ pub unsafe fn vst2_s64(a: *mut i64, b: int64x1x2_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st2.v1i64.p0i8" + link_name = "llvm.aarch64.neon.st2.v1i64.p0" )] fn _vst2_s64(a: int64x1_t, b: int64x1_t, ptr: *mut i8); } @@ -68795,7 +68795,7 @@ pub unsafe fn vst2q_p16(a: *mut p16, b: poly16x8x2_t) { #[cfg_attr(test, assert_instr(vst3))] pub unsafe fn vst3_f16(a: *mut f16, b: float16x4x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0i8.v4f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0.v4f16")] fn _vst3_f16(ptr: *mut i8, a: float16x4_t, b: float16x4_t, c: float16x4_t, size: i32); } _vst3_f16(a as _, b.0, b.1, b.2, 2) @@ -68812,7 +68812,7 @@ pub unsafe fn vst3_f16(a: *mut f16, b: float16x4x3_t) { #[cfg_attr(test, assert_instr(vst3))] pub unsafe fn vst3q_f16(a: *mut f16, b: float16x8x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0i8.v8f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0.v8f16")] fn _vst3q_f16(ptr: *mut i8, a: float16x8_t, b: float16x8_t, c: float16x8_t, size: i32); } _vst3q_f16(a as _, b.0, b.1, b.2, 2) @@ -68830,7 +68830,7 @@ pub unsafe fn vst3_f16(a: *mut f16, b: float16x4x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v4f16.p0i8" + link_name = "llvm.aarch64.neon.st3.v4f16.p0" )] fn _vst3_f16(a: float16x4_t, b: float16x4_t, c: float16x4_t, ptr: *mut i8); } @@ -68849,7 +68849,7 @@ pub unsafe fn vst3q_f16(a: *mut f16, b: float16x8x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v8f16.p0i8" + link_name = "llvm.aarch64.neon.st3.v8f16.p0" )] fn _vst3q_f16(a: float16x8_t, b: float16x8_t, c: float16x8_t, ptr: *mut i8); } @@ -68866,7 +68866,7 @@ pub unsafe fn vst3q_f16(a: *mut f16, b: float16x8x3_t) { #[cfg_attr(test, assert_instr(vst3))] pub unsafe fn vst3_f32(a: *mut f32, b: float32x2x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0i8.v2f32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0.v2f32")] fn _vst3_f32(ptr: *mut i8, a: float32x2_t, b: float32x2_t, c: float32x2_t, size: i32); } _vst3_f32(a as _, b.0, b.1, b.2, 4) @@ -68882,7 +68882,7 @@ pub unsafe fn vst3_f32(a: *mut f32, b: float32x2x3_t) { #[cfg_attr(test, assert_instr(vst3))] pub unsafe fn vst3q_f32(a: *mut f32, b: float32x4x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0i8.v4f32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0.v4f32")] fn _vst3q_f32(ptr: *mut i8, a: float32x4_t, b: float32x4_t, c: float32x4_t, size: i32); } _vst3q_f32(a as _, b.0, b.1, b.2, 4) @@ -68898,7 +68898,7 @@ pub unsafe fn vst3q_f32(a: *mut f32, b: float32x4x3_t) { #[cfg_attr(test, assert_instr(vst3))] pub unsafe fn vst3_s8(a: *mut i8, b: int8x8x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0i8.v8i8")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0.v8i8")] fn _vst3_s8(ptr: *mut i8, a: int8x8_t, b: int8x8_t, c: int8x8_t, size: i32); } _vst3_s8(a as _, b.0, b.1, b.2, 1) @@ -68914,7 +68914,7 @@ pub unsafe fn vst3_s8(a: *mut i8, b: int8x8x3_t) { #[cfg_attr(test, assert_instr(vst3))] pub unsafe fn vst3q_s8(a: *mut i8, b: int8x16x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0i8.v16i8")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0.v16i8")] fn _vst3q_s8(ptr: *mut i8, a: int8x16_t, b: int8x16_t, c: int8x16_t, size: i32); } _vst3q_s8(a as _, b.0, b.1, b.2, 1) @@ -68930,7 +68930,7 @@ pub unsafe fn vst3q_s8(a: *mut i8, b: int8x16x3_t) { #[cfg_attr(test, assert_instr(vst3))] pub unsafe fn vst3_s16(a: *mut i16, b: int16x4x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0i8.v4i16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0.v4i16")] fn _vst3_s16(ptr: *mut i8, a: int16x4_t, b: int16x4_t, c: int16x4_t, size: i32); } _vst3_s16(a as _, b.0, b.1, b.2, 2) @@ -68946,7 +68946,7 @@ pub unsafe fn vst3_s16(a: *mut i16, b: int16x4x3_t) { #[cfg_attr(test, assert_instr(vst3))] pub unsafe fn vst3q_s16(a: *mut i16, b: int16x8x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0i8.v8i16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0.v8i16")] fn _vst3q_s16(ptr: *mut i8, a: int16x8_t, b: int16x8_t, c: int16x8_t, size: i32); } _vst3q_s16(a as _, b.0, b.1, b.2, 2) @@ -68962,7 +68962,7 @@ pub unsafe fn vst3q_s16(a: *mut i16, b: int16x8x3_t) { #[cfg_attr(test, assert_instr(vst3))] pub unsafe fn vst3_s32(a: *mut i32, b: int32x2x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0i8.v2i32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0.v2i32")] fn _vst3_s32(ptr: *mut i8, a: int32x2_t, b: int32x2_t, c: int32x2_t, size: i32); } _vst3_s32(a as _, b.0, b.1, b.2, 4) @@ -68978,7 +68978,7 @@ pub unsafe fn vst3_s32(a: *mut i32, b: int32x2x3_t) { #[cfg_attr(test, assert_instr(vst3))] pub unsafe fn vst3q_s32(a: *mut i32, b: int32x4x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0i8.v4i32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0.v4i32")] fn _vst3q_s32(ptr: *mut i8, a: int32x4_t, b: int32x4_t, c: int32x4_t, size: i32); } _vst3q_s32(a as _, b.0, b.1, b.2, 4) @@ -68996,7 +68996,7 @@ pub unsafe fn vst3_f32(a: *mut f32, b: float32x2x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v2f32.p0i8" + link_name = "llvm.aarch64.neon.st3.v2f32.p0" )] fn _vst3_f32(a: float32x2_t, b: float32x2_t, c: float32x2_t, ptr: *mut i8); } @@ -69015,7 +69015,7 @@ pub unsafe fn vst3q_f32(a: *mut f32, b: float32x4x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v4f32.p0i8" + link_name = "llvm.aarch64.neon.st3.v4f32.p0" )] fn _vst3q_f32(a: float32x4_t, b: float32x4_t, c: float32x4_t, ptr: *mut i8); } @@ -69034,7 +69034,7 @@ pub unsafe fn vst3_s8(a: *mut i8, b: int8x8x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v8i8.p0i8" + link_name = "llvm.aarch64.neon.st3.v8i8.p0" )] fn _vst3_s8(a: int8x8_t, b: int8x8_t, c: int8x8_t, ptr: *mut i8); } @@ -69053,7 +69053,7 @@ pub unsafe fn vst3q_s8(a: *mut i8, b: int8x16x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v16i8.p0i8" + link_name = "llvm.aarch64.neon.st3.v16i8.p0" )] fn _vst3q_s8(a: int8x16_t, b: int8x16_t, c: int8x16_t, ptr: *mut i8); } @@ -69072,7 +69072,7 @@ pub unsafe fn vst3_s16(a: *mut i16, b: int16x4x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v4i16.p0i8" + link_name = "llvm.aarch64.neon.st3.v4i16.p0" )] fn _vst3_s16(a: int16x4_t, b: int16x4_t, c: int16x4_t, ptr: *mut i8); } @@ -69091,7 +69091,7 @@ pub unsafe fn vst3q_s16(a: *mut i16, b: int16x8x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v8i16.p0i8" + link_name = "llvm.aarch64.neon.st3.v8i16.p0" )] fn _vst3q_s16(a: int16x8_t, b: int16x8_t, c: int16x8_t, ptr: *mut i8); } @@ -69110,7 +69110,7 @@ pub unsafe fn vst3_s32(a: *mut i32, b: int32x2x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v2i32.p0i8" + link_name = "llvm.aarch64.neon.st3.v2i32.p0" )] fn _vst3_s32(a: int32x2_t, b: int32x2_t, c: int32x2_t, ptr: *mut i8); } @@ -69129,7 +69129,7 @@ pub unsafe fn vst3q_s32(a: *mut i32, b: int32x4x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v4i32.p0i8" + link_name = "llvm.aarch64.neon.st3.v4i32.p0" )] fn _vst3q_s32(a: int32x4_t, b: int32x4_t, c: int32x4_t, ptr: *mut i8); } @@ -69149,7 +69149,7 @@ pub unsafe fn vst3q_s32(a: *mut i32, b: int32x4x3_t) { pub unsafe fn vst3_lane_f16<const LANE: i32>(a: *mut f16, b: float16x4x3_t) { static_assert_uimm_bits!(LANE, 2); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0i8.v4f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0.v4f16")] fn _vst3_lane_f16( ptr: *mut i8, a: float16x4_t, @@ -69175,7 +69175,7 @@ pub unsafe fn vst3_lane_f16<const LANE: i32>(a: *mut f16, b: float16x4x3_t) { pub unsafe fn vst3q_lane_f16<const LANE: i32>(a: *mut f16, b: float16x8x3_t) { static_assert_uimm_bits!(LANE, 3); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0i8.v8f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0.v8f16")] fn _vst3q_lane_f16( ptr: *mut i8, a: float16x8_t, @@ -69202,7 +69202,7 @@ pub unsafe fn vst3_lane_f16<const LANE: i32>(a: *mut f16, b: float16x4x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v4f16.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v4f16.p0" )] fn _vst3_lane_f16(a: float16x4_t, b: float16x4_t, c: float16x4_t, n: i64, ptr: *mut i8); } @@ -69223,7 +69223,7 @@ pub unsafe fn vst3q_lane_f16<const LANE: i32>(a: *mut f16, b: float16x8x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v8f16.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v8f16.p0" )] fn _vst3q_lane_f16(a: float16x8_t, b: float16x8_t, c: float16x8_t, n: i64, ptr: *mut i8); } @@ -69242,7 +69242,7 @@ pub unsafe fn vst3q_lane_f16<const LANE: i32>(a: *mut f16, b: float16x8x3_t) { pub unsafe fn vst3_lane_f32<const LANE: i32>(a: *mut f32, b: float32x2x3_t) { static_assert_uimm_bits!(LANE, 1); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0i8.v2f32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0.v2f32")] fn _vst3_lane_f32( ptr: *mut i8, a: float32x2_t, @@ -69267,7 +69267,7 @@ pub unsafe fn vst3_lane_f32<const LANE: i32>(a: *mut f32, b: float32x2x3_t) { pub unsafe fn vst3q_lane_f32<const LANE: i32>(a: *mut f32, b: float32x4x3_t) { static_assert_uimm_bits!(LANE, 2); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0i8.v4f32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0.v4f32")] fn _vst3q_lane_f32( ptr: *mut i8, a: float32x4_t, @@ -69292,7 +69292,7 @@ pub unsafe fn vst3q_lane_f32<const LANE: i32>(a: *mut f32, b: float32x4x3_t) { pub unsafe fn vst3_lane_s8<const LANE: i32>(a: *mut i8, b: int8x8x3_t) { static_assert_uimm_bits!(LANE, 3); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0i8.v8i8")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0.v8i8")] fn _vst3_lane_s8(ptr: *mut i8, a: int8x8_t, b: int8x8_t, c: int8x8_t, n: i32, size: i32); } _vst3_lane_s8(a as _, b.0, b.1, b.2, LANE, 1) @@ -69310,7 +69310,7 @@ pub unsafe fn vst3_lane_s8<const LANE: i32>(a: *mut i8, b: int8x8x3_t) { pub unsafe fn vst3_lane_s16<const LANE: i32>(a: *mut i16, b: int16x4x3_t) { static_assert_uimm_bits!(LANE, 2); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0i8.v4i16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0.v4i16")] fn _vst3_lane_s16( ptr: *mut i8, a: int16x4_t, @@ -69335,7 +69335,7 @@ pub unsafe fn vst3_lane_s16<const LANE: i32>(a: *mut i16, b: int16x4x3_t) { pub unsafe fn vst3q_lane_s16<const LANE: i32>(a: *mut i16, b: int16x8x3_t) { static_assert_uimm_bits!(LANE, 3); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0i8.v8i16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0.v8i16")] fn _vst3q_lane_s16( ptr: *mut i8, a: int16x8_t, @@ -69360,7 +69360,7 @@ pub unsafe fn vst3q_lane_s16<const LANE: i32>(a: *mut i16, b: int16x8x3_t) { pub unsafe fn vst3_lane_s32<const LANE: i32>(a: *mut i32, b: int32x2x3_t) { static_assert_uimm_bits!(LANE, 1); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0i8.v2i32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0.v2i32")] fn _vst3_lane_s32( ptr: *mut i8, a: int32x2_t, @@ -69385,7 +69385,7 @@ pub unsafe fn vst3_lane_s32<const LANE: i32>(a: *mut i32, b: int32x2x3_t) { pub unsafe fn vst3q_lane_s32<const LANE: i32>(a: *mut i32, b: int32x4x3_t) { static_assert_uimm_bits!(LANE, 2); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0i8.v4i32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3lane.p0.v4i32")] fn _vst3q_lane_s32( ptr: *mut i8, a: int32x4_t, @@ -69412,7 +69412,7 @@ pub unsafe fn vst3_lane_f32<const LANE: i32>(a: *mut f32, b: float32x2x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v2f32.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v2f32.p0" )] fn _vst3_lane_f32(a: float32x2_t, b: float32x2_t, c: float32x2_t, n: i64, ptr: *mut i8); } @@ -69433,7 +69433,7 @@ pub unsafe fn vst3q_lane_f32<const LANE: i32>(a: *mut f32, b: float32x4x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v4f32.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v4f32.p0" )] fn _vst3q_lane_f32(a: float32x4_t, b: float32x4_t, c: float32x4_t, n: i64, ptr: *mut i8); } @@ -69454,7 +69454,7 @@ pub unsafe fn vst3_lane_s8<const LANE: i32>(a: *mut i8, b: int8x8x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v8i8.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v8i8.p0" )] fn _vst3_lane_s8(a: int8x8_t, b: int8x8_t, c: int8x8_t, n: i64, ptr: *mut i8); } @@ -69475,7 +69475,7 @@ pub unsafe fn vst3_lane_s16<const LANE: i32>(a: *mut i16, b: int16x4x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v4i16.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v4i16.p0" )] fn _vst3_lane_s16(a: int16x4_t, b: int16x4_t, c: int16x4_t, n: i64, ptr: *mut i8); } @@ -69496,7 +69496,7 @@ pub unsafe fn vst3q_lane_s16<const LANE: i32>(a: *mut i16, b: int16x8x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v8i16.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v8i16.p0" )] fn _vst3q_lane_s16(a: int16x8_t, b: int16x8_t, c: int16x8_t, n: i64, ptr: *mut i8); } @@ -69517,7 +69517,7 @@ pub unsafe fn vst3_lane_s32<const LANE: i32>(a: *mut i32, b: int32x2x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v2i32.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v2i32.p0" )] fn _vst3_lane_s32(a: int32x2_t, b: int32x2_t, c: int32x2_t, n: i64, ptr: *mut i8); } @@ -69538,7 +69538,7 @@ pub unsafe fn vst3q_lane_s32<const LANE: i32>(a: *mut i32, b: int32x4x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3lane.v4i32.p0i8" + link_name = "llvm.aarch64.neon.st3lane.v4i32.p0" )] fn _vst3q_lane_s32(a: int32x4_t, b: int32x4_t, c: int32x4_t, n: i64, ptr: *mut i8); } @@ -69780,7 +69780,7 @@ pub unsafe fn vst3_s64(a: *mut i64, b: int64x1x3_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st3.v1i64.p0i8" + link_name = "llvm.aarch64.neon.st3.v1i64.p0" )] fn _vst3_s64(a: int64x1_t, b: int64x1_t, c: int64x1_t, ptr: *mut i8); } @@ -69797,7 +69797,7 @@ pub unsafe fn vst3_s64(a: *mut i64, b: int64x1x3_t) { #[cfg_attr(test, assert_instr(nop))] pub unsafe fn vst3_s64(a: *mut i64, b: int64x1x3_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0i8.v1i64")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst3.p0.v1i64")] fn _vst3_s64(ptr: *mut i8, a: int64x1_t, b: int64x1_t, c: int64x1_t, size: i32); } _vst3_s64(a as _, b.0, b.1, b.2, 8) @@ -70067,7 +70067,7 @@ pub unsafe fn vst3q_p16(a: *mut p16, b: poly16x8x3_t) { #[cfg_attr(test, assert_instr(vst4))] pub unsafe fn vst4_f16(a: *mut f16, b: float16x4x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0i8.v4f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0.v4f16")] fn _vst4_f16( ptr: *mut i8, a: float16x4_t, @@ -70091,7 +70091,7 @@ pub unsafe fn vst4_f16(a: *mut f16, b: float16x4x4_t) { #[cfg_attr(test, assert_instr(vst4))] pub unsafe fn vst4q_f16(a: *mut f16, b: float16x8x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0i8.v8f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0.v8f16")] fn _vst4q_f16( ptr: *mut i8, a: float16x8_t, @@ -70116,7 +70116,7 @@ pub unsafe fn vst4_f16(a: *mut f16, b: float16x4x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v4f16.p0i8" + link_name = "llvm.aarch64.neon.st4.v4f16.p0" )] fn _vst4_f16(a: float16x4_t, b: float16x4_t, c: float16x4_t, d: float16x4_t, ptr: *mut i8); } @@ -70135,7 +70135,7 @@ pub unsafe fn vst4q_f16(a: *mut f16, b: float16x8x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v8f16.p0i8" + link_name = "llvm.aarch64.neon.st4.v8f16.p0" )] fn _vst4q_f16(a: float16x8_t, b: float16x8_t, c: float16x8_t, d: float16x8_t, ptr: *mut i8); } @@ -70152,7 +70152,7 @@ pub unsafe fn vst4q_f16(a: *mut f16, b: float16x8x4_t) { #[cfg_attr(test, assert_instr(vst4))] pub unsafe fn vst4_f32(a: *mut f32, b: float32x2x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0i8.v2f32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0.v2f32")] fn _vst4_f32( ptr: *mut i8, a: float32x2_t, @@ -70175,7 +70175,7 @@ pub unsafe fn vst4_f32(a: *mut f32, b: float32x2x4_t) { #[cfg_attr(test, assert_instr(vst4))] pub unsafe fn vst4q_f32(a: *mut f32, b: float32x4x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0i8.v4f32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0.v4f32")] fn _vst4q_f32( ptr: *mut i8, a: float32x4_t, @@ -70198,7 +70198,7 @@ pub unsafe fn vst4q_f32(a: *mut f32, b: float32x4x4_t) { #[cfg_attr(test, assert_instr(vst4))] pub unsafe fn vst4_s8(a: *mut i8, b: int8x8x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0i8.v8i8")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0.v8i8")] fn _vst4_s8(ptr: *mut i8, a: int8x8_t, b: int8x8_t, c: int8x8_t, d: int8x8_t, size: i32); } _vst4_s8(a as _, b.0, b.1, b.2, b.3, 1) @@ -70214,7 +70214,7 @@ pub unsafe fn vst4_s8(a: *mut i8, b: int8x8x4_t) { #[cfg_attr(test, assert_instr(vst4))] pub unsafe fn vst4q_s8(a: *mut i8, b: int8x16x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0i8.v16i8")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0.v16i8")] fn _vst4q_s8( ptr: *mut i8, a: int8x16_t, @@ -70237,7 +70237,7 @@ pub unsafe fn vst4q_s8(a: *mut i8, b: int8x16x4_t) { #[cfg_attr(test, assert_instr(vst4))] pub unsafe fn vst4_s16(a: *mut i16, b: int16x4x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0i8.v4i16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0.v4i16")] fn _vst4_s16( ptr: *mut i8, a: int16x4_t, @@ -70260,7 +70260,7 @@ pub unsafe fn vst4_s16(a: *mut i16, b: int16x4x4_t) { #[cfg_attr(test, assert_instr(vst4))] pub unsafe fn vst4q_s16(a: *mut i16, b: int16x8x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0i8.v8i16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0.v8i16")] fn _vst4q_s16( ptr: *mut i8, a: int16x8_t, @@ -70283,7 +70283,7 @@ pub unsafe fn vst4q_s16(a: *mut i16, b: int16x8x4_t) { #[cfg_attr(test, assert_instr(vst4))] pub unsafe fn vst4_s32(a: *mut i32, b: int32x2x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0i8.v2i32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0.v2i32")] fn _vst4_s32( ptr: *mut i8, a: int32x2_t, @@ -70306,7 +70306,7 @@ pub unsafe fn vst4_s32(a: *mut i32, b: int32x2x4_t) { #[cfg_attr(test, assert_instr(vst4))] pub unsafe fn vst4q_s32(a: *mut i32, b: int32x4x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0i8.v4i32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0.v4i32")] fn _vst4q_s32( ptr: *mut i8, a: int32x4_t, @@ -70331,7 +70331,7 @@ pub unsafe fn vst4_f32(a: *mut f32, b: float32x2x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v2f32.p0i8" + link_name = "llvm.aarch64.neon.st4.v2f32.p0" )] fn _vst4_f32(a: float32x2_t, b: float32x2_t, c: float32x2_t, d: float32x2_t, ptr: *mut i8); } @@ -70350,7 +70350,7 @@ pub unsafe fn vst4q_f32(a: *mut f32, b: float32x4x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v4f32.p0i8" + link_name = "llvm.aarch64.neon.st4.v4f32.p0" )] fn _vst4q_f32(a: float32x4_t, b: float32x4_t, c: float32x4_t, d: float32x4_t, ptr: *mut i8); } @@ -70369,7 +70369,7 @@ pub unsafe fn vst4_s8(a: *mut i8, b: int8x8x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v8i8.p0i8" + link_name = "llvm.aarch64.neon.st4.v8i8.p0" )] fn _vst4_s8(a: int8x8_t, b: int8x8_t, c: int8x8_t, d: int8x8_t, ptr: *mut i8); } @@ -70388,7 +70388,7 @@ pub unsafe fn vst4q_s8(a: *mut i8, b: int8x16x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v16i8.p0i8" + link_name = "llvm.aarch64.neon.st4.v16i8.p0" )] fn _vst4q_s8(a: int8x16_t, b: int8x16_t, c: int8x16_t, d: int8x16_t, ptr: *mut i8); } @@ -70407,7 +70407,7 @@ pub unsafe fn vst4_s16(a: *mut i16, b: int16x4x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v4i16.p0i8" + link_name = "llvm.aarch64.neon.st4.v4i16.p0" )] fn _vst4_s16(a: int16x4_t, b: int16x4_t, c: int16x4_t, d: int16x4_t, ptr: *mut i8); } @@ -70426,7 +70426,7 @@ pub unsafe fn vst4q_s16(a: *mut i16, b: int16x8x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v8i16.p0i8" + link_name = "llvm.aarch64.neon.st4.v8i16.p0" )] fn _vst4q_s16(a: int16x8_t, b: int16x8_t, c: int16x8_t, d: int16x8_t, ptr: *mut i8); } @@ -70445,7 +70445,7 @@ pub unsafe fn vst4_s32(a: *mut i32, b: int32x2x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v2i32.p0i8" + link_name = "llvm.aarch64.neon.st4.v2i32.p0" )] fn _vst4_s32(a: int32x2_t, b: int32x2_t, c: int32x2_t, d: int32x2_t, ptr: *mut i8); } @@ -70464,7 +70464,7 @@ pub unsafe fn vst4q_s32(a: *mut i32, b: int32x4x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v4i32.p0i8" + link_name = "llvm.aarch64.neon.st4.v4i32.p0" )] fn _vst4q_s32(a: int32x4_t, b: int32x4_t, c: int32x4_t, d: int32x4_t, ptr: *mut i8); } @@ -70484,7 +70484,7 @@ pub unsafe fn vst4q_s32(a: *mut i32, b: int32x4x4_t) { pub unsafe fn vst4_lane_f16<const LANE: i32>(a: *mut f16, b: float16x4x4_t) { static_assert_uimm_bits!(LANE, 2); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0i8.v4f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0.v4f16")] fn _vst4_lane_f16( ptr: *mut i8, a: float16x4_t, @@ -70511,7 +70511,7 @@ pub unsafe fn vst4_lane_f16<const LANE: i32>(a: *mut f16, b: float16x4x4_t) { pub unsafe fn vst4q_lane_f16<const LANE: i32>(a: *mut f16, b: float16x8x4_t) { static_assert_uimm_bits!(LANE, 3); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0i8.v8f16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0.v8f16")] fn _vst4q_lane_f16( ptr: *mut i8, a: float16x8_t, @@ -70539,7 +70539,7 @@ pub unsafe fn vst4_lane_f16<const LANE: i32>(a: *mut f16, b: float16x4x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v4f16.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v4f16.p0" )] fn _vst4_lane_f16( a: float16x4_t, @@ -70567,7 +70567,7 @@ pub unsafe fn vst4q_lane_f16<const LANE: i32>(a: *mut f16, b: float16x8x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v8f16.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v8f16.p0" )] fn _vst4q_lane_f16( a: float16x8_t, @@ -70593,7 +70593,7 @@ pub unsafe fn vst4q_lane_f16<const LANE: i32>(a: *mut f16, b: float16x8x4_t) { pub unsafe fn vst4_lane_f32<const LANE: i32>(a: *mut f32, b: float32x2x4_t) { static_assert_uimm_bits!(LANE, 1); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0i8.v2f32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0.v2f32")] fn _vst4_lane_f32( ptr: *mut i8, a: float32x2_t, @@ -70619,7 +70619,7 @@ pub unsafe fn vst4_lane_f32<const LANE: i32>(a: *mut f32, b: float32x2x4_t) { pub unsafe fn vst4q_lane_f32<const LANE: i32>(a: *mut f32, b: float32x4x4_t) { static_assert_uimm_bits!(LANE, 2); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0i8.v4f32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0.v4f32")] fn _vst4q_lane_f32( ptr: *mut i8, a: float32x4_t, @@ -70645,7 +70645,7 @@ pub unsafe fn vst4q_lane_f32<const LANE: i32>(a: *mut f32, b: float32x4x4_t) { pub unsafe fn vst4_lane_s8<const LANE: i32>(a: *mut i8, b: int8x8x4_t) { static_assert_uimm_bits!(LANE, 3); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0i8.v8i8")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0.v8i8")] fn _vst4_lane_s8( ptr: *mut i8, a: int8x8_t, @@ -70671,7 +70671,7 @@ pub unsafe fn vst4_lane_s8<const LANE: i32>(a: *mut i8, b: int8x8x4_t) { pub unsafe fn vst4_lane_s16<const LANE: i32>(a: *mut i16, b: int16x4x4_t) { static_assert_uimm_bits!(LANE, 2); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0i8.v4i16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0.v4i16")] fn _vst4_lane_s16( ptr: *mut i8, a: int16x4_t, @@ -70697,7 +70697,7 @@ pub unsafe fn vst4_lane_s16<const LANE: i32>(a: *mut i16, b: int16x4x4_t) { pub unsafe fn vst4q_lane_s16<const LANE: i32>(a: *mut i16, b: int16x8x4_t) { static_assert_uimm_bits!(LANE, 3); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0i8.v8i16")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0.v8i16")] fn _vst4q_lane_s16( ptr: *mut i8, a: int16x8_t, @@ -70723,7 +70723,7 @@ pub unsafe fn vst4q_lane_s16<const LANE: i32>(a: *mut i16, b: int16x8x4_t) { pub unsafe fn vst4_lane_s32<const LANE: i32>(a: *mut i32, b: int32x2x4_t) { static_assert_uimm_bits!(LANE, 1); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0i8.v2i32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0.v2i32")] fn _vst4_lane_s32( ptr: *mut i8, a: int32x2_t, @@ -70749,7 +70749,7 @@ pub unsafe fn vst4_lane_s32<const LANE: i32>(a: *mut i32, b: int32x2x4_t) { pub unsafe fn vst4q_lane_s32<const LANE: i32>(a: *mut i32, b: int32x4x4_t) { static_assert_uimm_bits!(LANE, 2); unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0i8.v4i32")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4lane.p0.v4i32")] fn _vst4q_lane_s32( ptr: *mut i8, a: int32x4_t, @@ -70777,7 +70777,7 @@ pub unsafe fn vst4_lane_f32<const LANE: i32>(a: *mut f32, b: float32x2x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v2f32.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v2f32.p0" )] fn _vst4_lane_f32( a: float32x2_t, @@ -70805,7 +70805,7 @@ pub unsafe fn vst4q_lane_f32<const LANE: i32>(a: *mut f32, b: float32x4x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v4f32.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v4f32.p0" )] fn _vst4q_lane_f32( a: float32x4_t, @@ -70833,7 +70833,7 @@ pub unsafe fn vst4_lane_s8<const LANE: i32>(a: *mut i8, b: int8x8x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v8i8.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v8i8.p0" )] fn _vst4_lane_s8(a: int8x8_t, b: int8x8_t, c: int8x8_t, d: int8x8_t, n: i64, ptr: *mut i8); } @@ -70854,7 +70854,7 @@ pub unsafe fn vst4_lane_s16<const LANE: i32>(a: *mut i16, b: int16x4x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v4i16.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v4i16.p0" )] fn _vst4_lane_s16( a: int16x4_t, @@ -70882,7 +70882,7 @@ pub unsafe fn vst4q_lane_s16<const LANE: i32>(a: *mut i16, b: int16x8x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v8i16.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v8i16.p0" )] fn _vst4q_lane_s16( a: int16x8_t, @@ -70910,7 +70910,7 @@ pub unsafe fn vst4_lane_s32<const LANE: i32>(a: *mut i32, b: int32x2x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v2i32.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v2i32.p0" )] fn _vst4_lane_s32( a: int32x2_t, @@ -70938,7 +70938,7 @@ pub unsafe fn vst4q_lane_s32<const LANE: i32>(a: *mut i32, b: int32x4x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4lane.v4i32.p0i8" + link_name = "llvm.aarch64.neon.st4lane.v4i32.p0" )] fn _vst4q_lane_s32( a: int32x4_t, @@ -71185,7 +71185,7 @@ pub unsafe fn vst4_p64(a: *mut p64, b: poly64x1x4_t) { #[cfg_attr(test, assert_instr(nop))] pub unsafe fn vst4_s64(a: *mut i64, b: int64x1x4_t) { unsafe extern "unadjusted" { - #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0i8.v1i64")] + #[cfg_attr(target_arch = "arm", link_name = "llvm.arm.neon.vst4.p0.v1i64")] fn _vst4_s64( ptr: *mut i8, a: int64x1_t, @@ -71210,7 +71210,7 @@ pub unsafe fn vst4_s64(a: *mut i64, b: int64x1x4_t) { unsafe extern "unadjusted" { #[cfg_attr( any(target_arch = "aarch64", target_arch = "arm64ec"), - link_name = "llvm.aarch64.neon.st4.v1i64.p0i8" + link_name = "llvm.aarch64.neon.st4.v1i64.p0" )] fn _vst4_s64(a: int64x1_t, b: int64x1_t, c: int64x1_t, d: int64x1_t, ptr: *mut i8); } diff --git a/library/stdarch/crates/stdarch-gen-arm/spec/neon/aarch64.spec.yml b/library/stdarch/crates/stdarch-gen-arm/spec/neon/aarch64.spec.yml index 5f482b2527d..f658267b9a1 100644 --- a/library/stdarch/crates/stdarch-gen-arm/spec/neon/aarch64.spec.yml +++ b/library/stdarch/crates/stdarch-gen-arm/spec/neon/aarch64.spec.yml @@ -3428,7 +3428,7 @@ intrinsics: - LLVMLink: name: "vld1{neon_type[1].no}" links: - - link: "llvm.aarch64.neon.ld1x{neon_type[1].tuple}.v{neon_type[1].lane}f{neon_type[1].base}.p0f64" + - link: "llvm.aarch64.neon.ld1x{neon_type[1].tuple}.v{neon_type[1].lane}f{neon_type[1].base}.p0" arch: aarch64,arm64ec - name: "vld2{neon_type[1].lane_nox}" @@ -3460,7 +3460,7 @@ intrinsics: - "n: i64" - "ptr: *const i8" links: - - link: "llvm.aarch64.neon.ld2lane.v{neon_type[1].lane}{type[2]}.p0i8" + - link: "llvm.aarch64.neon.ld2lane.v{neon_type[1].lane}{type[2]}.p0" arch: aarch64,arm64ec - FnCall: ["_vld2{neon_type[1].lane_nox}", ["b.0", "b.1", "LANE as i64", "a as _"]] @@ -3489,7 +3489,7 @@ intrinsics: - "n: i64" - "ptr: *const i8" links: - - link: "llvm.aarch64.neon.ld2lane.v{neon_type[1].lane}{type[2]}.p0i8" + - link: "llvm.aarch64.neon.ld2lane.v{neon_type[1].lane}{type[2]}.p0" arch: aarch64,arm64ec - FnCall: ["_vld2{neon_type[1].lane_nox}", ["b.0", "b.1", "LANE as i64", "a as _"]] @@ -3613,7 +3613,7 @@ intrinsics: arguments: - "ptr: *const {neon_type[3]}" links: - - link: "llvm.aarch64.neon.ld2.v{neon_type[1].lane}{type[2]}.p0v{neon_type[1].lane}{type[2]}" + - link: "llvm.aarch64.neon.ld2.v{neon_type[1].lane}{type[2]}.p0" arch: aarch64,arm64ec - FnCall: - "_vld2{neon_type[1].nox}" @@ -3635,7 +3635,7 @@ intrinsics: arguments: - "ptr: *const {neon_type[3]}" links: - - link: "llvm.aarch64.neon.ld2.v{neon_type[1].lane}{type[2]}.p0v{neon_type[1].lane}{type[2]}" + - link: "llvm.aarch64.neon.ld2.v{neon_type[1].lane}{type[2]}.p0" arch: aarch64,arm64ec - FnCall: - "_vld2{neon_type[1].nox}" @@ -3696,7 +3696,7 @@ intrinsics: arguments: - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.ld2r.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.aarch64.neon.ld2r.v{neon_type[1].lane}{type[2]}.p0" arch: aarch64,arm64ec - FnCall: - "_vld2{neon_type[1].dup_nox}" @@ -3769,7 +3769,7 @@ intrinsics: - 'n: i64' - 'ptr: *const i8' links: - - link: 'llvm.aarch64.neon.ld3lane.v{neon_type[1].lane}{type[3]}.p0i8' + - link: 'llvm.aarch64.neon.ld3lane.v{neon_type[1].lane}{type[3]}.p0' arch: aarch64,arm64ec - FnCall: ['_vld3{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'LANE as i64', 'a as _']] @@ -3798,7 +3798,7 @@ intrinsics: - 'n: i64' - 'ptr: *const i8' links: - - link: 'llvm.aarch64.neon.ld3lane.v{neon_type[1].lane}{type[3]}.p0i8' + - link: 'llvm.aarch64.neon.ld3lane.v{neon_type[1].lane}{type[3]}.p0' arch: aarch64,arm64ec - FnCall: ['_vld3{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'LANE as i64', 'a as _']] @@ -3827,7 +3827,7 @@ intrinsics: - 'n: i64' - 'ptr: *const i8' links: - - link: 'llvm.aarch64.neon.ld3lane.v{neon_type[1].lane}{type[3]}.p0i8' + - link: 'llvm.aarch64.neon.ld3lane.v{neon_type[1].lane}{type[3]}.p0' arch: aarch64,arm64ec - FnCall: ['_vld3{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'LANE as i64', 'a as _']] @@ -3856,7 +3856,7 @@ intrinsics: - 'n: i64' - 'ptr: *const i8' links: - - link: 'llvm.aarch64.neon.ld3lane.v{neon_type[1].lane}{type[3]}.p0i8' + - link: 'llvm.aarch64.neon.ld3lane.v{neon_type[1].lane}{type[3]}.p0' arch: aarch64,arm64ec - FnCall: ['_vld3{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'LANE as i64', 'a as _']] @@ -3973,7 +3973,7 @@ intrinsics: arguments: - 'ptr: {type[2]}' links: - - link: 'llvm.aarch64.neon.ld3.v{neon_type[1].lane}{type[3]}.p0v{neon_type[1].lane}{type[3]}' + - link: 'llvm.aarch64.neon.ld3.v{neon_type[1].lane}{type[3]}.p0' arch: aarch64,arm64ec - FnCall: ['_vld3{neon_type[1].nox}', ['a as _']] @@ -3993,7 +3993,7 @@ intrinsics: arguments: - 'ptr: {type[2]}' links: - - link: 'llvm.aarch64.neon.ld3.v{neon_type[1].lane}{type[3]}.p0v{neon_type[1].lane}{type[3]}' + - link: 'llvm.aarch64.neon.ld3.v{neon_type[1].lane}{type[3]}.p0' arch: aarch64,arm64ec - FnCall: ['_vld3{neon_type[1].nox}', ['a as _']] @@ -4055,7 +4055,7 @@ intrinsics: arguments: - 'ptr: {type[0]}' links: - - link: 'llvm.aarch64.neon.ld3r.v{neon_type[1].lane}{type[2]}.p0{type[2]}' + - link: 'llvm.aarch64.neon.ld3r.v{neon_type[1].lane}{type[2]}.p0' arch: aarch64,arm64ec - FnCall: ['_vld3{neon_type[1].dup_nox}', ['a as _']] @@ -4117,7 +4117,7 @@ intrinsics: arguments: - 'ptr: {type[3]}' links: - - link: 'llvm.aarch64.neon.ld4.v{neon_type[1].lane}{type[2]}.p0v{neon_type[1].lane}{type[2]}' + - link: 'llvm.aarch64.neon.ld4.v{neon_type[1].lane}{type[2]}.p0' arch: aarch64,arm64ec - FnCall: ['_vld4{neon_type[1].nox}', ['a as _']] @@ -4138,7 +4138,7 @@ intrinsics: arguments: - 'ptr: {type[3]}' links: - - link: 'llvm.aarch64.neon.ld4.v{neon_type[1].lane}{type[2]}.p0v{neon_type[1].lane}{type[2]}' + - link: 'llvm.aarch64.neon.ld4.v{neon_type[1].lane}{type[2]}.p0' arch: aarch64,arm64ec - FnCall: ['_vld4{neon_type[1].nox}', ['a as _']] @@ -4210,7 +4210,7 @@ intrinsics: - 'n: i64' - 'ptr: *const i8' links: - - link: 'llvm.aarch64.neon.ld4lane.v{neon_type[1].lane}{type[3]}.p0i8' + - link: 'llvm.aarch64.neon.ld4lane.v{neon_type[1].lane}{type[3]}.p0' arch: aarch64,arm64ec - FnCall: ['_vld4{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'b.3', 'LANE as i64', 'a as _']] @@ -4241,7 +4241,7 @@ intrinsics: - 'n: i64' - 'ptr: *const i8' links: - - link: 'llvm.aarch64.neon.ld4lane.v{neon_type[1].lane}{type[3]}.p0i8' + - link: 'llvm.aarch64.neon.ld4lane.v{neon_type[1].lane}{type[3]}.p0' arch: aarch64,arm64ec - FnCall: ['_vld4{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'b.3', 'LANE as i64', 'a as _']] @@ -4398,7 +4398,7 @@ intrinsics: - 'b: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st2.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st2.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst2{neon_type[1].nox}', ['b.0', 'b.1', 'a as _']] @@ -4421,7 +4421,7 @@ intrinsics: - 'b: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st2.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st2.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst2{neon_type[1].nox}', ['b.0', 'b.1', 'a as _']] @@ -4448,7 +4448,7 @@ intrinsics: - 'n: i64' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st2lane.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st2lane.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst2{neon_type[1].lane_nox}', ['b.0', 'b.1', 'LANE as i64', 'a as _']] @@ -4476,7 +4476,7 @@ intrinsics: - 'n: i64' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st2lane.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st2lane.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst2{neon_type[1].lane_nox}', ['b.0', 'b.1', 'LANE as i64', 'a as _']] @@ -4611,7 +4611,7 @@ intrinsics: - 'c: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st3.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st3.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst3{neon_type[1].nox}', ['b.0', 'b.1', 'b.2', 'a as _']] @@ -4691,7 +4691,7 @@ intrinsics: - 'c: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st3.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st3.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst3{neon_type[1].nox}', ['b.0', 'b.1', 'b.2', 'a as _']] @@ -4775,7 +4775,7 @@ intrinsics: - 'n: i64' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st3lane.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st3lane.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst3{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'LANE as i64', 'a as _']] @@ -4803,7 +4803,7 @@ intrinsics: - 'n: i64' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st3lane.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st3lane.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst3{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'LANE as i64', 'a as _']] @@ -4826,7 +4826,7 @@ intrinsics: - 'd: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st4.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st4.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst4{neon_type[1].nox}', ['b.0', 'b.1', 'b.2', 'b.3', 'a as _']] @@ -4907,7 +4907,7 @@ intrinsics: - 'd: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st4.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st4.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst4{neon_type[1].nox}', ['b.0', 'b.1', 'b.2', 'b.3', 'a as _']] @@ -4992,7 +4992,7 @@ intrinsics: - 'n: i64' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st4lane.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st4lane.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst4{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'b.3', 'LANE as i64', 'a as _']] @@ -5021,7 +5021,7 @@ intrinsics: - 'n: i64' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st4lane.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st4lane.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst4{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'b.3', 'LANE as i64', 'a as _']] @@ -10142,7 +10142,7 @@ intrinsics: - "b: {neon_type[2]}" - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.st1x{neon_type[1].tuple}.{neon_type[2]}.p0f64" + - link: "llvm.aarch64.neon.st1x{neon_type[1].tuple}.{neon_type[2]}.p0" arch: aarch64,arm64ec - FnCall: ["_vst1{neon_type[1].no}", ['b.0', 'b.1', 'a']] @@ -10166,7 +10166,7 @@ intrinsics: - "c: {neon_type[2]}" - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.st1x{neon_type[1].tuple}.{neon_type[2]}.p0f64" + - link: "llvm.aarch64.neon.st1x{neon_type[1].tuple}.{neon_type[2]}.p0" arch: aarch64,arm64ec - FnCall: ["_vst1{neon_type[1].no}", ['b.0', 'b.1', 'b.2', 'a']] @@ -10191,7 +10191,7 @@ intrinsics: - "d: {neon_type[2]}" - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.st1x{neon_type[1].tuple}.{neon_type[2]}.p0f64" + - link: "llvm.aarch64.neon.st1x{neon_type[1].tuple}.{neon_type[2]}.p0" arch: aarch64,arm64ec - FnCall: ["_vst1{neon_type[1].no}", ['b.0', 'b.1', 'b.2', 'b.3', 'a']] @@ -11998,16 +11998,16 @@ intrinsics: safety: unsafe: [neon] types: - - ["*const i64", int64x2x4_t, "v2i64", "p0i64"] - - ["*const f64", float64x1x4_t, "v1f64", "p0f64"] - - ["*const f64", float64x2x4_t, "v2f64", "p0f64"] + - ["*const i64", int64x2x4_t, "v2i64"] + - ["*const f64", float64x1x4_t, "v1f64"] + - ["*const f64", float64x2x4_t, "v2f64"] compose: - LLVMLink: name: "vld4{neon_type[1].dup_nox}" arguments: - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.ld4r.{type[2]}.{type[3]}" + - link: "llvm.aarch64.neon.ld4r.{type[2]}.p0" arch: aarch64,arm64ec - FnCall: ["_vld4{neon_type[1].dup_nox}", ['a as _']] diff --git a/library/stdarch/crates/stdarch-gen-arm/spec/neon/arm_shared.spec.yml b/library/stdarch/crates/stdarch-gen-arm/spec/neon/arm_shared.spec.yml index 1f8806ed1ee..118f5808f75 100644 --- a/library/stdarch/crates/stdarch-gen-arm/spec/neon/arm_shared.spec.yml +++ b/library/stdarch/crates/stdarch-gen-arm/spec/neon/arm_shared.spec.yml @@ -2756,9 +2756,9 @@ intrinsics: - LLVMLink: name: "vld1x{neon_type[1].tuple}.{neon_type[1]}" links: - - link: "llvm.aarch64.neon.ld1x{neon_type[1].tuple}.v{neon_type[1].lane}f{neon_type[1].base}.p0f{neon_type[1].base}" + - link: "llvm.aarch64.neon.ld1x{neon_type[1].tuple}.v{neon_type[1].lane}f{neon_type[1].base}.p0" arch: aarch64,arm64ec - - link: "llvm.arm.neon.vld1x{neon_type[1].tuple}.v{neon_type[1].lane}f{neon_type[1].base}.p0f{neon_type[1].base}" + - link: "llvm.arm.neon.vld1x{neon_type[1].tuple}.v{neon_type[1].lane}f{neon_type[1].base}.p0" arch: arm - name: "vld1{type[2]}_{neon_type[1]}" @@ -3397,7 +3397,7 @@ intrinsics: - "ptr: {type[0]}" - "size: i32" links: - - link: "llvm.arm.neon.vld2.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.arm.neon.vld2.v{neon_type[1].lane}{type[2]}.p0" arch: arm - FnCall: - "_vld2{neon_type[1].nox}" @@ -3424,7 +3424,7 @@ intrinsics: arguments: - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.ld2.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.aarch64.neon.ld2.v{neon_type[1].lane}{type[2]}.p0" arch: aarch64,arm64ec - FnCall: - "_vld2{neon_type[1].nox}" @@ -3452,7 +3452,7 @@ intrinsics: - "ptr: {type[0]}" - "size: i32" links: - - link: "llvm.arm.neon.vld2dup.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.arm.neon.vld2dup.v{neon_type[1].lane}{type[2]}.p0" arch: arm - FnCall: - "_vld2{neon_type[1].dup_nox}" @@ -3480,7 +3480,7 @@ intrinsics: arguments: - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.ld2r.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.aarch64.neon.ld2r.v{neon_type[1].lane}{type[2]}.p0" arch: aarch64,arm64ec - FnCall: - "_vld2{neon_type[1].dup_nox}" @@ -3592,7 +3592,7 @@ intrinsics: - "ptr: {type[0]}" - "size: i32" links: - - link: "llvm.arm.neon.vld3.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.arm.neon.vld3.v{neon_type[1].lane}{type[2]}.p0" arch: arm - FnCall: - "_vld3{neon_type[1].nox}" @@ -3619,7 +3619,7 @@ intrinsics: arguments: - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.ld3.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.aarch64.neon.ld3.v{neon_type[1].lane}{type[2]}.p0" arch: aarch64,arm64ec - FnCall: - "_vld3{neon_type[1].nox}" @@ -3647,7 +3647,7 @@ intrinsics: - "ptr: {type[0]}" - "size: i32" links: - - link: "llvm.arm.neon.vld3dup.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.arm.neon.vld3dup.v{neon_type[1].lane}{type[2]}.p0" arch: arm - FnCall: - "_vld3{neon_type[1].dup_nox}" @@ -3675,7 +3675,7 @@ intrinsics: arguments: - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.ld3r.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.aarch64.neon.ld3r.v{neon_type[1].lane}{type[2]}.p0" arch: aarch64,arm64ec - FnCall: - "_vld3{neon_type[1].dup_nox}" @@ -4755,7 +4755,7 @@ intrinsics: - 'b: {type[2]}' - 'ptr: *mut {type[0]}' links: - - link: 'llvm.aarch64.neon.st1x2.v{neon_type[1].lane}{type[0]}.p0{type[0]}' + - link: 'llvm.aarch64.neon.st1x2.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst1{neon_type[1].no}', ['b.0', 'b.1', 'a']] @@ -4786,7 +4786,7 @@ intrinsics: - 'c: {type[2]}' - 'ptr: *mut {type[0]}' links: - - link: 'llvm.aarch64.neon.st1x3.v{neon_type[1].lane}{type[0]}.p0{type[0]}' + - link: 'llvm.aarch64.neon.st1x3.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst1{neon_type[1].no}', ['b.0', 'b.1', 'b.2', 'a']] @@ -4818,7 +4818,7 @@ intrinsics: - 'd: {type[2]}' - 'ptr: *mut {type[0]}' links: - - link: 'llvm.aarch64.neon.st1x4.v{neon_type[1].lane}{type[0]}.p0{type[0]}' + - link: 'llvm.aarch64.neon.st1x4.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst1{neon_type[1].no}', ['b.0', 'b.1', 'b.2', 'b.3', 'a']] @@ -4881,7 +4881,7 @@ intrinsics: - 'b: {type[2]}' - 'c: {type[2]}' links: - - link: 'llvm.arm.neon.vst1x3.p0{type[0]}.v{neon_type[1].lane}{type[0]}.p0' + - link: 'llvm.arm.neon.vst1x3.p0.v{neon_type[1].lane}{type[0]}.p0' arch: arm - FnCall: ['_vst1{neon_type[1].no}', ['a', 'b.0', 'b.1', 'b.2']] @@ -4914,7 +4914,7 @@ intrinsics: - 'c: {type[2]}' - 'd: {type[2]}' links: - - link: 'llvm.arm.neon.vst1x4.p0{type[0]}.v{neon_type[1].lane}{type[0]}.p0' + - link: 'llvm.arm.neon.vst1x4.p0.v{neon_type[1].lane}{type[0]}.p0' arch: arm - FnCall: ['_vst1{neon_type[1].no}', ['a', 'b.0', 'b.1', 'b.2', 'b.3']] @@ -4941,7 +4941,7 @@ intrinsics: - 'c: {type[2]}' - 'd: {type[2]}' links: - - link: 'llvm.arm.neon.vst1x4.p0{type[0]}.v{neon_type[1].lane}{type[0]}.p0' + - link: 'llvm.arm.neon.vst1x4.p0.v{neon_type[1].lane}{type[0]}.p0' arch: arm - FnCall: ['_vst1{neon_type[1].no}', ['a', 'b.0', 'b.1', 'b.2', 'b.3']] @@ -4969,7 +4969,7 @@ intrinsics: - 'c: {type[2]}' - 'd: {type[2]}' links: - - link: 'llvm.arm.neon.vst1x4.p0{type[0]}.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst1x4.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst1{neon_type[1].no}', ['a', 'b.0', 'b.1', 'b.2', 'b.3']] @@ -5056,7 +5056,7 @@ intrinsics: - 'b: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st2.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st2.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst2{neon_type[1].nox}', ['b.0', 'b.1', 'a as _']] @@ -5086,7 +5086,7 @@ intrinsics: - 'b: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st2.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st2.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst2{neon_type[1].nox}', ['b.0', 'b.1', 'a as _']] @@ -5112,7 +5112,7 @@ intrinsics: - 'b: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st2.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st2.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst2{neon_type[1].nox}', ['b.0', 'b.1', 'a as _']] @@ -5174,7 +5174,7 @@ intrinsics: - 'n: i64' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st2lane.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st2lane.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst2{neon_type[1].lane_nox}', ['b.0', 'b.1', 'LANE as i64', 'a as _']] @@ -5204,7 +5204,7 @@ intrinsics: - 'n: i64' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st2lane.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st2lane.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst2{neon_type[1].lane_nox}', ['b.0', 'b.1', 'LANE as i64', 'a as _']] @@ -5294,7 +5294,7 @@ intrinsics: - 'b: {type[2]}' - 'size: i32' links: - - link: 'llvm.arm.neon.vst2.p0i8.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst2.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst2{neon_type[1].nox}', ['a as _', 'b.0', 'b.1', "{type[3]}"]] @@ -5362,7 +5362,7 @@ intrinsics: - 'n: i32' - 'size: i32' links: - - link: 'llvm.arm.neon.vst2lane.p0i8.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst2lane.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst2{neon_type[1].lane_nox}', ['a as _', 'b.0', 'b.1', 'LANE', "{type[4]}"]] @@ -5387,7 +5387,7 @@ intrinsics: - 'c: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st3.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st3.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst3{neon_type[1].nox}', ['b.0', 'b.1', 'b.2', 'a as _']] @@ -5433,7 +5433,7 @@ intrinsics: - 'c: {type[2]}' - 'size: i32' links: - - link: 'llvm.arm.neon.vst3.p0i8.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst3.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst3{neon_type[1].nox}', ['a as _', 'b.0', 'b.1', 'b.2', '8']] @@ -5542,7 +5542,7 @@ intrinsics: - 'c: {type[2]}' - 'size: i32' links: - - link: 'llvm.arm.neon.vst3.p0i8.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst3.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst3{neon_type[1].nox}', ['a as _', 'b.0', 'b.1', 'b.2', "{type[3]}"]] @@ -5571,7 +5571,7 @@ intrinsics: - 'c: {type[2]}' - 'size: i32' links: - - link: 'llvm.arm.neon.vst3.p0i8.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst3.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst3{neon_type[1].nox}', ['a as _', 'b.0', 'b.1', 'b.2', "{type[3]}"]] @@ -5608,7 +5608,7 @@ intrinsics: - 'n: i32' - 'size: i32' links: - - link: 'llvm.arm.neon.vst3lane.p0i8.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst3lane.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst3{neon_type[1].lane_nox}', ['a as _', 'b.0', 'b.1', 'b.2', 'LANE', "{type[4]}"]] @@ -5641,7 +5641,7 @@ intrinsics: - 'n: i32' - 'size: i32' links: - - link: 'llvm.arm.neon.vst3lane.p0i8.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst3lane.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst3{neon_type[1].lane_nox}', ['a as _', 'b.0', 'b.1', 'b.2', 'LANE', "{type[4]}"]] @@ -5671,7 +5671,7 @@ intrinsics: - 'c: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st3.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st3.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst3{neon_type[1].nox}', ['b.0', 'b.1', 'b.2', 'a as _']] @@ -5698,7 +5698,7 @@ intrinsics: - 'c: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st3.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st3.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst3{neon_type[1].nox}', ['b.0', 'b.1', 'b.2', 'a as _']] @@ -5733,7 +5733,7 @@ intrinsics: - 'n: i64' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st3lane.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st3lane.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst3{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'LANE as i64', 'a as _']] @@ -5764,7 +5764,7 @@ intrinsics: - 'n: i64' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st3lane.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st3lane.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst3{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'LANE as i64', 'a as _']] @@ -5812,7 +5812,7 @@ intrinsics: - 'd: {type[2]}' - 'size: i32' links: - - link: 'llvm.arm.neon.vst4.p0i8.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst4.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst4{neon_type[1].nox}', ['a as _', 'b.0', 'b.1', 'b.2', 'b.3', '8']] @@ -5837,7 +5837,7 @@ intrinsics: - 'd: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st4.{neon_type[2]}.p0i8' + - link: 'llvm.aarch64.neon.st4.{neon_type[2]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst4{neon_type[1].nox}', ['b.0', 'b.1', 'b.2', 'b.3', 'a as _']] @@ -5947,7 +5947,7 @@ intrinsics: - 'd: {type[2]}' - 'size: i32' links: - - link: 'llvm.arm.neon.vst4.p0i8.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst4.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst4{neon_type[1].nox}', ['a as _', 'b.0', 'b.1', 'b.2', 'b.3', "{type[3]}"]] @@ -5977,7 +5977,7 @@ intrinsics: - 'd: {type[2]}' - 'size: i32' links: - - link: 'llvm.arm.neon.vst4.p0i8.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst4.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst4{neon_type[1].nox}', ['a as _', 'b.0', 'b.1', 'b.2', 'b.3', "{type[3]}"]] @@ -6015,7 +6015,7 @@ intrinsics: - 'n: i32' - 'size: i32' links: - - link: 'llvm.arm.neon.vst4lane.p0i8.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst4lane.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst4{neon_type[1].lane_nox}', ['a as _', 'b.0', 'b.1', 'b.2', 'b.3', 'LANE', "{type[4]}"]] @@ -6048,7 +6048,7 @@ intrinsics: - 'n: i32' - 'size: i32' links: - - link: 'llvm.arm.neon.vst4lane.p0i8.v{neon_type[1].lane}{type[0]}' + - link: 'llvm.arm.neon.vst4lane.p0.v{neon_type[1].lane}{type[0]}' arch: arm - FnCall: ['_vst4{neon_type[1].lane_nox}', ['a as _', 'b.0', 'b.1', 'b.2', 'b.3', 'LANE', "{type[4]}"]] @@ -6079,7 +6079,7 @@ intrinsics: - 'd: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st4.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st4.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst4{neon_type[1].nox}', ['b.0', 'b.1', 'b.2', 'b.3', 'a as _']] @@ -6107,7 +6107,7 @@ intrinsics: - 'd: {type[2]}' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st4.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st4.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst4{neon_type[1].nox}', ['b.0', 'b.1', 'b.2', 'b.3', 'a as _']] @@ -6143,7 +6143,7 @@ intrinsics: - 'n: i64' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st4lane.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st4lane.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst4{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'b.3', 'LANE as i64', 'a as _']] @@ -6175,7 +6175,7 @@ intrinsics: - 'n: i64' - 'ptr: *mut i8' links: - - link: 'llvm.aarch64.neon.st4lane.v{neon_type[1].lane}{type[0]}.p0i8' + - link: 'llvm.aarch64.neon.st4lane.v{neon_type[1].lane}{type[0]}.p0' arch: aarch64,arm64ec - FnCall: ['_vst4{neon_type[1].lane_nox}', ['b.0', 'b.1', 'b.2', 'b.3', 'LANE as i64', 'a as _']] @@ -10097,7 +10097,7 @@ intrinsics: - "a: {neon_type[2]}" - "b: {neon_type[2]}" links: - - link: "llvm.arm.neon.vst1x{neon_type[1].tuple}.p0f16.{neon_type[2]}" + - link: "llvm.arm.neon.vst1x{neon_type[1].tuple}.p0.{neon_type[2]}" arch: arm - FnCall: ["_vst1{neon_type[1].no}", ['a', 'b.0', 'b.1']] @@ -10124,7 +10124,7 @@ intrinsics: - "b: {neon_type[2]}" - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.st1x2.{neon_type[2]}.p0f16" + - link: "llvm.aarch64.neon.st1x2.{neon_type[2]}.p0" arch: aarch64,arm64ec - FnCall: ["_vst1{neon_type[1].no}", ['b.0', 'b.1', a]] @@ -10151,7 +10151,7 @@ intrinsics: - "b: {neon_type[2]}" - "c: {neon_type[2]}" links: - - link: "llvm.arm.neon.vst1x{neon_type[1].tuple}.p0f16.{neon_type[2]}" + - link: "llvm.arm.neon.vst1x{neon_type[1].tuple}.p0.{neon_type[2]}" arch: arm - FnCall: ["_vst1{neon_type[1].no}", ['a', 'b.0', 'b.1', 'b.2']] @@ -10175,7 +10175,7 @@ intrinsics: - "b: {neon_type[2]}" - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.st1x2.{neon_type[2]}.p0f32" + - link: "llvm.aarch64.neon.st1x2.{neon_type[2]}.p0" arch: aarch64,arm64ec - FnCall: ["_vst1{neon_type[1].no}", ['b.0', 'b.1', a]] @@ -10200,7 +10200,7 @@ intrinsics: - "c: {neon_type[2]}" - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.st1x3.{neon_type[2]}.p0f32" + - link: "llvm.aarch64.neon.st1x3.{neon_type[2]}.p0" arch: aarch64,arm64ec - FnCall: ["_vst1{neon_type[1].no}", ['b.0', 'b.1', 'b.2', a]] @@ -10227,7 +10227,7 @@ intrinsics: - "c: {neon_type[2]}" - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.st1x3.{neon_type[2]}.p0f16" + - link: "llvm.aarch64.neon.st1x3.{neon_type[2]}.p0" arch: aarch64,arm64ec - FnCall: ["_vst1{neon_type[1].no}", ['b.0', 'b.1', 'b.2', a]] @@ -10254,7 +10254,7 @@ intrinsics: - "d: {neon_type[2]}" - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.st1x4.{neon_type[2]}.p0f32" + - link: "llvm.aarch64.neon.st1x4.{neon_type[2]}.p0" arch: aarch64,arm64ec - FnCall: ["_vst1{neon_type[1].no}", ['b.0', 'b.1', 'b.2', 'b.3', a]] @@ -10282,7 +10282,7 @@ intrinsics: - "d: {neon_type[2]}" - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.st1x4.{neon_type[2]}.p0f16" + - link: "llvm.aarch64.neon.st1x4.{neon_type[2]}.p0" arch: aarch64,arm64ec - FnCall: ["_vst1{neon_type[1].no}", ['b.0', 'b.1', 'b.2', 'b.3', a]] @@ -11629,22 +11629,22 @@ intrinsics: safety: unsafe: [neon] types: - - ["*const i8", int8x8x4_t, int8x8_t, 'p0i8'] - - ["*const i16", int16x4x4_t, int16x4_t, 'p0i16'] - - ["*const i32", int32x2x4_t, int32x2_t, 'p0i32'] - - ["*const i8", int8x16x4_t, int8x16_t, 'p0i8'] - - ["*const i16", int16x8x4_t, int16x8_t, 'p0i16'] - - ["*const i32", int32x4x4_t, int32x4_t, 'p0i32'] - - ["*const i64", int64x1x4_t, int64x1_t, 'p0i64'] - - ["*const f32", float32x2x4_t, float32x2_t, 'p0f32'] - - ["*const f32", float32x4x4_t, float32x4_t, 'p0f32'] + - ["*const i8", int8x8x4_t, int8x8_t] + - ["*const i16", int16x4x4_t, int16x4_t] + - ["*const i32", int32x2x4_t, int32x2_t] + - ["*const i8", int8x16x4_t, int8x16_t] + - ["*const i16", int16x8x4_t, int16x8_t] + - ["*const i32", int32x4x4_t, int32x4_t] + - ["*const i64", int64x1x4_t, int64x1_t] + - ["*const f32", float32x2x4_t, float32x2_t] + - ["*const f32", float32x4x4_t, float32x4_t] compose: - LLVMLink: name: "vld4{neon_type[1].dup_nox}" arguments: - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.ld4r.{neon_type[2]}.{type[3]}.p0" + - link: "llvm.aarch64.neon.ld4r.{neon_type[2]}.p0.p0" arch: aarch64,arm64ec - FnCall: ["_vld4{neon_type[1].dup_nox}", ['a as _']] @@ -12163,7 +12163,7 @@ intrinsics: - "ptr: {type[0]}" - "size: i32" links: - - link: "llvm.arm.neon.vld4.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.arm.neon.vld4.v{neon_type[1].lane}{type[2]}.p0" arch: arm - FnCall: - "_vld4{neon_type[1].nox}" @@ -12190,7 +12190,7 @@ intrinsics: arguments: - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.ld4.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.aarch64.neon.ld4.v{neon_type[1].lane}{type[2]}.p0" arch: aarch64,arm64ec - FnCall: - "_vld4{neon_type[1].nox}" @@ -12218,7 +12218,7 @@ intrinsics: - "ptr: {type[0]}" - "size: i32" links: - - link: "llvm.arm.neon.vld4dup.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.arm.neon.vld4dup.v{neon_type[1].lane}{type[2]}.p0" arch: arm - FnCall: - "_vld4{neon_type[1].dup_nox}" @@ -12246,7 +12246,7 @@ intrinsics: arguments: - "ptr: {type[0]}" links: - - link: "llvm.aarch64.neon.ld4r.v{neon_type[1].lane}{type[2]}.p0{type[2]}" + - link: "llvm.aarch64.neon.ld4r.v{neon_type[1].lane}{type[2]}.p0" arch: aarch64,arm64ec - FnCall: - "_vld4{neon_type[1].dup_nox}" |
