@@ -10281,7 +10281,10 @@ pub fn svmul_lane_u64<const IMM_INDEX: i32>(op1: svuint64_t, op2: svuint64_t) ->
1028110281#[inline(always)]
1028210282#[target_feature(enable = "sve,sve2")]
1028310283#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10284- #[cfg_attr(test, assert_instr(smullb, IMM_INDEX = 0))]
10284+ #[cfg_attr(
10285+ all(test, not(target_env = "msvc")),
10286+ assert_instr(smullb, IMM_INDEX = 0)
10287+ )]
1028510288pub fn svmullb_lane_s32<const IMM_INDEX: i32>(op1: svint16_t, op2: svint16_t) -> svint32_t {
1028610289 static_assert_range!(IMM_INDEX, 0..=7);
1028710290 unsafe extern "unadjusted" {
@@ -10298,7 +10301,10 @@ pub fn svmullb_lane_s32<const IMM_INDEX: i32>(op1: svint16_t, op2: svint16_t) ->
1029810301#[inline(always)]
1029910302#[target_feature(enable = "sve,sve2")]
1030010303#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10301- #[cfg_attr(test, assert_instr(smullb, IMM_INDEX = 0))]
10304+ #[cfg_attr(
10305+ all(test, not(target_env = "msvc")),
10306+ assert_instr(smullb, IMM_INDEX = 0)
10307+ )]
1030210308pub fn svmullb_lane_s64<const IMM_INDEX: i32>(op1: svint32_t, op2: svint32_t) -> svint64_t {
1030310309 static_assert_range!(IMM_INDEX, 0..=3);
1030410310 unsafe extern "unadjusted" {
@@ -10315,7 +10321,10 @@ pub fn svmullb_lane_s64<const IMM_INDEX: i32>(op1: svint32_t, op2: svint32_t) ->
1031510321#[inline(always)]
1031610322#[target_feature(enable = "sve,sve2")]
1031710323#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10318- #[cfg_attr(test, assert_instr(umullb, IMM_INDEX = 0))]
10324+ #[cfg_attr(
10325+ all(test, not(target_env = "msvc")),
10326+ assert_instr(umullb, IMM_INDEX = 0)
10327+ )]
1031910328pub fn svmullb_lane_u32<const IMM_INDEX: i32>(op1: svuint16_t, op2: svuint16_t) -> svuint32_t {
1032010329 static_assert_range!(IMM_INDEX, 0..=7);
1032110330 unsafe extern "unadjusted" {
@@ -10332,7 +10341,10 @@ pub fn svmullb_lane_u32<const IMM_INDEX: i32>(op1: svuint16_t, op2: svuint16_t)
1033210341#[inline(always)]
1033310342#[target_feature(enable = "sve,sve2")]
1033410343#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10335- #[cfg_attr(test, assert_instr(umullb, IMM_INDEX = 0))]
10344+ #[cfg_attr(
10345+ all(test, not(target_env = "msvc")),
10346+ assert_instr(umullb, IMM_INDEX = 0)
10347+ )]
1033610348pub fn svmullb_lane_u64<const IMM_INDEX: i32>(op1: svuint32_t, op2: svuint32_t) -> svuint64_t {
1033710349 static_assert_range!(IMM_INDEX, 0..=3);
1033810350 unsafe extern "unadjusted" {
@@ -10481,7 +10493,10 @@ pub fn svmullb_n_u64(op1: svuint32_t, op2: u32) -> svuint64_t {
1048110493#[inline(always)]
1048210494#[target_feature(enable = "sve,sve2")]
1048310495#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10484- #[cfg_attr(test, assert_instr(smullt, IMM_INDEX = 0))]
10496+ #[cfg_attr(
10497+ all(test, not(target_env = "msvc")),
10498+ assert_instr(smullt, IMM_INDEX = 0)
10499+ )]
1048510500pub fn svmullt_lane_s32<const IMM_INDEX: i32>(op1: svint16_t, op2: svint16_t) -> svint32_t {
1048610501 static_assert_range!(IMM_INDEX, 0..=7);
1048710502 unsafe extern "unadjusted" {
@@ -10498,7 +10513,10 @@ pub fn svmullt_lane_s32<const IMM_INDEX: i32>(op1: svint16_t, op2: svint16_t) ->
1049810513#[inline(always)]
1049910514#[target_feature(enable = "sve,sve2")]
1050010515#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10501- #[cfg_attr(test, assert_instr(smullt, IMM_INDEX = 0))]
10516+ #[cfg_attr(
10517+ all(test, not(target_env = "msvc")),
10518+ assert_instr(smullt, IMM_INDEX = 0)
10519+ )]
1050210520pub fn svmullt_lane_s64<const IMM_INDEX: i32>(op1: svint32_t, op2: svint32_t) -> svint64_t {
1050310521 static_assert_range!(IMM_INDEX, 0..=3);
1050410522 unsafe extern "unadjusted" {
@@ -10515,7 +10533,10 @@ pub fn svmullt_lane_s64<const IMM_INDEX: i32>(op1: svint32_t, op2: svint32_t) ->
1051510533#[inline(always)]
1051610534#[target_feature(enable = "sve,sve2")]
1051710535#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10518- #[cfg_attr(test, assert_instr(umullt, IMM_INDEX = 0))]
10536+ #[cfg_attr(
10537+ all(test, not(target_env = "msvc")),
10538+ assert_instr(umullt, IMM_INDEX = 0)
10539+ )]
1051910540pub fn svmullt_lane_u32<const IMM_INDEX: i32>(op1: svuint16_t, op2: svuint16_t) -> svuint32_t {
1052010541 static_assert_range!(IMM_INDEX, 0..=7);
1052110542 unsafe extern "unadjusted" {
@@ -10532,7 +10553,10 @@ pub fn svmullt_lane_u32<const IMM_INDEX: i32>(op1: svuint16_t, op2: svuint16_t)
1053210553#[inline(always)]
1053310554#[target_feature(enable = "sve,sve2")]
1053410555#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10535- #[cfg_attr(test, assert_instr(umullt, IMM_INDEX = 0))]
10556+ #[cfg_attr(
10557+ all(test, not(target_env = "msvc")),
10558+ assert_instr(umullt, IMM_INDEX = 0)
10559+ )]
1053610560pub fn svmullt_lane_u64<const IMM_INDEX: i32>(op1: svuint32_t, op2: svuint32_t) -> svuint64_t {
1053710561 static_assert_range!(IMM_INDEX, 0..=3);
1053810562 unsafe extern "unadjusted" {
@@ -10549,7 +10573,7 @@ pub fn svmullt_lane_u64<const IMM_INDEX: i32>(op1: svuint32_t, op2: svuint32_t)
1054910573#[inline(always)]
1055010574#[target_feature(enable = "sve,sve2")]
1055110575#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10552- #[cfg_attr(test, assert_instr(smullt))]
10576+ #[cfg_attr(all( test, not(target_env = "msvc")) , assert_instr(smullt))]
1055310577pub fn svmullt_s16(op1: svint8_t, op2: svint8_t) -> svint16_t {
1055410578 unsafe extern "unadjusted" {
1055510579 #[cfg_attr(target_arch = "aarch64", link_name = "llvm.aarch64.sve.smullt.nxv8i16")]
@@ -10562,7 +10586,7 @@ pub fn svmullt_s16(op1: svint8_t, op2: svint8_t) -> svint16_t {
1056210586#[inline(always)]
1056310587#[target_feature(enable = "sve,sve2")]
1056410588#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10565- #[cfg_attr(test, assert_instr(smullt))]
10589+ #[cfg_attr(all( test, not(target_env = "msvc")) , assert_instr(smullt))]
1056610590pub fn svmullt_n_s16(op1: svint8_t, op2: i8) -> svint16_t {
1056710591 svmullt_s16(op1, svdup_n_s8(op2))
1056810592}
@@ -10571,7 +10595,7 @@ pub fn svmullt_n_s16(op1: svint8_t, op2: i8) -> svint16_t {
1057110595#[inline(always)]
1057210596#[target_feature(enable = "sve,sve2")]
1057310597#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10574- #[cfg_attr(test, assert_instr(smullt))]
10598+ #[cfg_attr(all( test, not(target_env = "msvc")) , assert_instr(smullt))]
1057510599pub fn svmullt_s32(op1: svint16_t, op2: svint16_t) -> svint32_t {
1057610600 unsafe extern "unadjusted" {
1057710601 #[cfg_attr(target_arch = "aarch64", link_name = "llvm.aarch64.sve.smullt.nxv4i32")]
@@ -10584,7 +10608,7 @@ pub fn svmullt_s32(op1: svint16_t, op2: svint16_t) -> svint32_t {
1058410608#[inline(always)]
1058510609#[target_feature(enable = "sve,sve2")]
1058610610#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10587- #[cfg_attr(test, assert_instr(smullt))]
10611+ #[cfg_attr(all( test, not(target_env = "msvc")) , assert_instr(smullt))]
1058810612pub fn svmullt_n_s32(op1: svint16_t, op2: i16) -> svint32_t {
1058910613 svmullt_s32(op1, svdup_n_s16(op2))
1059010614}
@@ -10593,7 +10617,7 @@ pub fn svmullt_n_s32(op1: svint16_t, op2: i16) -> svint32_t {
1059310617#[inline(always)]
1059410618#[target_feature(enable = "sve,sve2")]
1059510619#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10596- #[cfg_attr(test, assert_instr(smullt))]
10620+ #[cfg_attr(all( test, not(target_env = "msvc")) , assert_instr(smullt))]
1059710621pub fn svmullt_s64(op1: svint32_t, op2: svint32_t) -> svint64_t {
1059810622 unsafe extern "unadjusted" {
1059910623 #[cfg_attr(target_arch = "aarch64", link_name = "llvm.aarch64.sve.smullt.nxv2i64")]
@@ -10606,7 +10630,7 @@ pub fn svmullt_s64(op1: svint32_t, op2: svint32_t) -> svint64_t {
1060610630#[inline(always)]
1060710631#[target_feature(enable = "sve,sve2")]
1060810632#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10609- #[cfg_attr(test, assert_instr(smullt))]
10633+ #[cfg_attr(all( test, not(target_env = "msvc")) , assert_instr(smullt))]
1061010634pub fn svmullt_n_s64(op1: svint32_t, op2: i32) -> svint64_t {
1061110635 svmullt_s64(op1, svdup_n_s32(op2))
1061210636}
@@ -10615,7 +10639,7 @@ pub fn svmullt_n_s64(op1: svint32_t, op2: i32) -> svint64_t {
1061510639#[inline(always)]
1061610640#[target_feature(enable = "sve,sve2")]
1061710641#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10618- #[cfg_attr(test, assert_instr(umullt))]
10642+ #[cfg_attr(all( test, not(target_env = "msvc")) , assert_instr(umullt))]
1061910643pub fn svmullt_u16(op1: svuint8_t, op2: svuint8_t) -> svuint16_t {
1062010644 unsafe extern "unadjusted" {
1062110645 #[cfg_attr(target_arch = "aarch64", link_name = "llvm.aarch64.sve.umullt.nxv8i16")]
@@ -10628,7 +10652,7 @@ pub fn svmullt_u16(op1: svuint8_t, op2: svuint8_t) -> svuint16_t {
1062810652#[inline(always)]
1062910653#[target_feature(enable = "sve,sve2")]
1063010654#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10631- #[cfg_attr(test, assert_instr(umullt))]
10655+ #[cfg_attr(all( test, not(target_env = "msvc")) , assert_instr(umullt))]
1063210656pub fn svmullt_n_u16(op1: svuint8_t, op2: u8) -> svuint16_t {
1063310657 svmullt_u16(op1, svdup_n_u8(op2))
1063410658}
@@ -10637,7 +10661,7 @@ pub fn svmullt_n_u16(op1: svuint8_t, op2: u8) -> svuint16_t {
1063710661#[inline(always)]
1063810662#[target_feature(enable = "sve,sve2")]
1063910663#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10640- #[cfg_attr(test, assert_instr(umullt))]
10664+ #[cfg_attr(all( test, not(target_env = "msvc")) , assert_instr(umullt))]
1064110665pub fn svmullt_u32(op1: svuint16_t, op2: svuint16_t) -> svuint32_t {
1064210666 unsafe extern "unadjusted" {
1064310667 #[cfg_attr(target_arch = "aarch64", link_name = "llvm.aarch64.sve.umullt.nxv4i32")]
@@ -10650,7 +10674,7 @@ pub fn svmullt_u32(op1: svuint16_t, op2: svuint16_t) -> svuint32_t {
1065010674#[inline(always)]
1065110675#[target_feature(enable = "sve,sve2")]
1065210676#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10653- #[cfg_attr(test, assert_instr(umullt))]
10677+ #[cfg_attr(all( test, not(target_env = "msvc")) , assert_instr(umullt))]
1065410678pub fn svmullt_n_u32(op1: svuint16_t, op2: u16) -> svuint32_t {
1065510679 svmullt_u32(op1, svdup_n_u16(op2))
1065610680}
@@ -10659,7 +10683,7 @@ pub fn svmullt_n_u32(op1: svuint16_t, op2: u16) -> svuint32_t {
1065910683#[inline(always)]
1066010684#[target_feature(enable = "sve,sve2")]
1066110685#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10662- #[cfg_attr(test, assert_instr(umullt))]
10686+ #[cfg_attr(all( test, not(target_env = "msvc")) , assert_instr(umullt))]
1066310687pub fn svmullt_u64(op1: svuint32_t, op2: svuint32_t) -> svuint64_t {
1066410688 unsafe extern "unadjusted" {
1066510689 #[cfg_attr(target_arch = "aarch64", link_name = "llvm.aarch64.sve.umullt.nxv2i64")]
@@ -10672,7 +10696,7 @@ pub fn svmullt_u64(op1: svuint32_t, op2: svuint32_t) -> svuint64_t {
1067210696#[inline(always)]
1067310697#[target_feature(enable = "sve,sve2")]
1067410698#[unstable(feature = "stdarch_aarch64_sve", issue = "145052")]
10675- #[cfg_attr(test, assert_instr(umullt))]
10699+ #[cfg_attr(all( test, not(target_env = "msvc")) , assert_instr(umullt))]
1067610700pub fn svmullt_n_u64(op1: svuint32_t, op2: u32) -> svuint64_t {
1067710701 svmullt_u64(op1, svdup_n_u32(op2))
1067810702}
0 commit comments