@@ -14609,7 +14609,7 @@ pub unsafe fn vld1q_dup_s32(ptr: *const i32) -> int32x4_t {
14609
14609
#[cfg_attr(all(test, target_arch = "arm"), assert_instr("vldr"))]
14610
14610
#[cfg_attr(
14611
14611
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
14612
- assert_instr(ld1 )
14612
+ assert_instr(ld1r )
14613
14613
)]
14614
14614
#[cfg_attr(
14615
14615
not(target_arch = "arm"),
@@ -14701,7 +14701,7 @@ pub unsafe fn vld1q_dup_u32(ptr: *const u32) -> uint32x4_t {
14701
14701
#[cfg_attr(all(test, target_arch = "arm"), assert_instr("vldr"))]
14702
14702
#[cfg_attr(
14703
14703
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
14704
- assert_instr(ld1 )
14704
+ assert_instr(ld1r )
14705
14705
)]
14706
14706
#[cfg_attr(
14707
14707
not(target_arch = "arm"),
@@ -73031,7 +73031,11 @@ pub fn vtrnq_f16(a: float16x8_t, b: float16x8_t) -> float16x8x2_t {
73031
73031
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vtrn))]
73032
73032
#[cfg_attr(
73033
73033
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
73034
- assert_instr(zip)
73034
+ assert_instr(zip1)
73035
+ )]
73036
+ #[cfg_attr(
73037
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
73038
+ assert_instr(zip2)
73035
73039
)]
73036
73040
#[cfg_attr(
73037
73041
not(target_arch = "arm"),
@@ -73056,7 +73060,11 @@ pub fn vtrn_f32(a: float32x2_t, b: float32x2_t) -> float32x2x2_t {
73056
73060
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vtrn))]
73057
73061
#[cfg_attr(
73058
73062
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
73059
- assert_instr(zip)
73063
+ assert_instr(zip1)
73064
+ )]
73065
+ #[cfg_attr(
73066
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
73067
+ assert_instr(zip2)
73060
73068
)]
73061
73069
#[cfg_attr(
73062
73070
not(target_arch = "arm"),
@@ -73081,7 +73089,11 @@ pub fn vtrn_s32(a: int32x2_t, b: int32x2_t) -> int32x2x2_t {
73081
73089
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vtrn))]
73082
73090
#[cfg_attr(
73083
73091
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
73084
- assert_instr(zip)
73092
+ assert_instr(zip1)
73093
+ )]
73094
+ #[cfg_attr(
73095
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
73096
+ assert_instr(zip2)
73085
73097
)]
73086
73098
#[cfg_attr(
73087
73099
not(target_arch = "arm"),
@@ -74083,7 +74095,11 @@ pub fn vuzpq_f16(a: float16x8_t, b: float16x8_t) -> float16x8x2_t {
74083
74095
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vtrn))]
74084
74096
#[cfg_attr(
74085
74097
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74086
- assert_instr(zip)
74098
+ assert_instr(zip1)
74099
+ )]
74100
+ #[cfg_attr(
74101
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74102
+ assert_instr(zip2)
74087
74103
)]
74088
74104
#[cfg_attr(
74089
74105
not(target_arch = "arm"),
@@ -74108,7 +74124,11 @@ pub fn vuzp_f32(a: float32x2_t, b: float32x2_t) -> float32x2x2_t {
74108
74124
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vtrn))]
74109
74125
#[cfg_attr(
74110
74126
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74111
- assert_instr(zip)
74127
+ assert_instr(zip1)
74128
+ )]
74129
+ #[cfg_attr(
74130
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74131
+ assert_instr(zip2)
74112
74132
)]
74113
74133
#[cfg_attr(
74114
74134
not(target_arch = "arm"),
@@ -74133,7 +74153,11 @@ pub fn vuzp_s32(a: int32x2_t, b: int32x2_t) -> int32x2x2_t {
74133
74153
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vtrn))]
74134
74154
#[cfg_attr(
74135
74155
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74136
- assert_instr(zip)
74156
+ assert_instr(zip1)
74157
+ )]
74158
+ #[cfg_attr(
74159
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74160
+ assert_instr(zip2)
74137
74161
)]
74138
74162
#[cfg_attr(
74139
74163
not(target_arch = "arm"),
@@ -74556,7 +74580,11 @@ pub fn vuzpq_p16(a: poly16x8_t, b: poly16x8_t) -> poly16x8x2_t {
74556
74580
#[cfg_attr(all(test, target_arch = "arm"), assert_instr("vzip.16"))]
74557
74581
#[cfg_attr(
74558
74582
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74559
- assert_instr(zip)
74583
+ assert_instr(zip1)
74584
+ )]
74585
+ #[cfg_attr(
74586
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74587
+ assert_instr(zip2)
74560
74588
)]
74561
74589
#[target_feature(enable = "neon,fp16")]
74562
74590
#[unstable(feature = "stdarch_neon_f16", issue = "136306")]
@@ -74574,7 +74602,11 @@ pub fn vzip_f16(a: float16x4_t, b: float16x4_t) -> float16x4x2_t {
74574
74602
#[cfg_attr(all(test, target_arch = "arm"), assert_instr("vzip.16"))]
74575
74603
#[cfg_attr(
74576
74604
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74577
- assert_instr(zip)
74605
+ assert_instr(zip1)
74606
+ )]
74607
+ #[cfg_attr(
74608
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74609
+ assert_instr(zip2)
74578
74610
)]
74579
74611
#[target_feature(enable = "neon,fp16")]
74580
74612
#[unstable(feature = "stdarch_neon_f16", issue = "136306")]
@@ -74593,7 +74625,11 @@ pub fn vzipq_f16(a: float16x8_t, b: float16x8_t) -> float16x8x2_t {
74593
74625
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vtrn))]
74594
74626
#[cfg_attr(
74595
74627
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74596
- assert_instr(zip)
74628
+ assert_instr(zip1)
74629
+ )]
74630
+ #[cfg_attr(
74631
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74632
+ assert_instr(zip2)
74597
74633
)]
74598
74634
#[cfg_attr(
74599
74635
not(target_arch = "arm"),
@@ -74618,7 +74654,11 @@ pub fn vzip_f32(a: float32x2_t, b: float32x2_t) -> float32x2x2_t {
74618
74654
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vtrn))]
74619
74655
#[cfg_attr(
74620
74656
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74621
- assert_instr(zip)
74657
+ assert_instr(zip1)
74658
+ )]
74659
+ #[cfg_attr(
74660
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74661
+ assert_instr(zip2)
74622
74662
)]
74623
74663
#[cfg_attr(
74624
74664
not(target_arch = "arm"),
@@ -74643,7 +74683,11 @@ pub fn vzip_s32(a: int32x2_t, b: int32x2_t) -> int32x2x2_t {
74643
74683
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vtrn))]
74644
74684
#[cfg_attr(
74645
74685
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74646
- assert_instr(zip)
74686
+ assert_instr(zip1)
74687
+ )]
74688
+ #[cfg_attr(
74689
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74690
+ assert_instr(zip2)
74647
74691
)]
74648
74692
#[cfg_attr(
74649
74693
not(target_arch = "arm"),
@@ -74668,7 +74712,11 @@ pub fn vzip_u32(a: uint32x2_t, b: uint32x2_t) -> uint32x2x2_t {
74668
74712
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vzip))]
74669
74713
#[cfg_attr(
74670
74714
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74671
- assert_instr(zip)
74715
+ assert_instr(zip1)
74716
+ )]
74717
+ #[cfg_attr(
74718
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74719
+ assert_instr(zip2)
74672
74720
)]
74673
74721
#[cfg_attr(
74674
74722
not(target_arch = "arm"),
@@ -74693,7 +74741,11 @@ pub fn vzip_s8(a: int8x8_t, b: int8x8_t) -> int8x8x2_t {
74693
74741
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vzip))]
74694
74742
#[cfg_attr(
74695
74743
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74696
- assert_instr(zip)
74744
+ assert_instr(zip1)
74745
+ )]
74746
+ #[cfg_attr(
74747
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74748
+ assert_instr(zip2)
74697
74749
)]
74698
74750
#[cfg_attr(
74699
74751
not(target_arch = "arm"),
@@ -74718,7 +74770,11 @@ pub fn vzip_s16(a: int16x4_t, b: int16x4_t) -> int16x4x2_t {
74718
74770
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vzip))]
74719
74771
#[cfg_attr(
74720
74772
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74721
- assert_instr(zip)
74773
+ assert_instr(zip1)
74774
+ )]
74775
+ #[cfg_attr(
74776
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74777
+ assert_instr(zip2)
74722
74778
)]
74723
74779
#[cfg_attr(
74724
74780
not(target_arch = "arm"),
@@ -74743,7 +74799,11 @@ pub fn vzip_u8(a: uint8x8_t, b: uint8x8_t) -> uint8x8x2_t {
74743
74799
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vzip))]
74744
74800
#[cfg_attr(
74745
74801
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74746
- assert_instr(zip)
74802
+ assert_instr(zip1)
74803
+ )]
74804
+ #[cfg_attr(
74805
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74806
+ assert_instr(zip2)
74747
74807
)]
74748
74808
#[cfg_attr(
74749
74809
not(target_arch = "arm"),
@@ -74768,7 +74828,11 @@ pub fn vzip_u16(a: uint16x4_t, b: uint16x4_t) -> uint16x4x2_t {
74768
74828
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vzip))]
74769
74829
#[cfg_attr(
74770
74830
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74771
- assert_instr(zip)
74831
+ assert_instr(zip1)
74832
+ )]
74833
+ #[cfg_attr(
74834
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74835
+ assert_instr(zip2)
74772
74836
)]
74773
74837
#[cfg_attr(
74774
74838
not(target_arch = "arm"),
@@ -74793,7 +74857,11 @@ pub fn vzip_p8(a: poly8x8_t, b: poly8x8_t) -> poly8x8x2_t {
74793
74857
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vzip))]
74794
74858
#[cfg_attr(
74795
74859
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74796
- assert_instr(zip)
74860
+ assert_instr(zip1)
74861
+ )]
74862
+ #[cfg_attr(
74863
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74864
+ assert_instr(zip2)
74797
74865
)]
74798
74866
#[cfg_attr(
74799
74867
not(target_arch = "arm"),
@@ -74818,7 +74886,11 @@ pub fn vzip_p16(a: poly16x4_t, b: poly16x4_t) -> poly16x4x2_t {
74818
74886
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vorr))]
74819
74887
#[cfg_attr(
74820
74888
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74821
- assert_instr(zip)
74889
+ assert_instr(zip1)
74890
+ )]
74891
+ #[cfg_attr(
74892
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74893
+ assert_instr(zip2)
74822
74894
)]
74823
74895
#[cfg_attr(
74824
74896
not(target_arch = "arm"),
@@ -74843,7 +74915,11 @@ pub fn vzipq_f32(a: float32x4_t, b: float32x4_t) -> float32x4x2_t {
74843
74915
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vorr))]
74844
74916
#[cfg_attr(
74845
74917
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74846
- assert_instr(zip)
74918
+ assert_instr(zip1)
74919
+ )]
74920
+ #[cfg_attr(
74921
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74922
+ assert_instr(zip2)
74847
74923
)]
74848
74924
#[cfg_attr(
74849
74925
not(target_arch = "arm"),
@@ -74876,7 +74952,11 @@ pub fn vzipq_s8(a: int8x16_t, b: int8x16_t) -> int8x16x2_t {
74876
74952
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vorr))]
74877
74953
#[cfg_attr(
74878
74954
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74879
- assert_instr(zip)
74955
+ assert_instr(zip1)
74956
+ )]
74957
+ #[cfg_attr(
74958
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74959
+ assert_instr(zip2)
74880
74960
)]
74881
74961
#[cfg_attr(
74882
74962
not(target_arch = "arm"),
@@ -74901,7 +74981,11 @@ pub fn vzipq_s16(a: int16x8_t, b: int16x8_t) -> int16x8x2_t {
74901
74981
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vorr))]
74902
74982
#[cfg_attr(
74903
74983
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74904
- assert_instr(zip)
74984
+ assert_instr(zip1)
74985
+ )]
74986
+ #[cfg_attr(
74987
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74988
+ assert_instr(zip2)
74905
74989
)]
74906
74990
#[cfg_attr(
74907
74991
not(target_arch = "arm"),
@@ -74926,7 +75010,11 @@ pub fn vzipq_s32(a: int32x4_t, b: int32x4_t) -> int32x4x2_t {
74926
75010
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vorr))]
74927
75011
#[cfg_attr(
74928
75012
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74929
- assert_instr(zip)
75013
+ assert_instr(zip1)
75014
+ )]
75015
+ #[cfg_attr(
75016
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
75017
+ assert_instr(zip2)
74930
75018
)]
74931
75019
#[cfg_attr(
74932
75020
not(target_arch = "arm"),
@@ -74959,7 +75047,11 @@ pub fn vzipq_u8(a: uint8x16_t, b: uint8x16_t) -> uint8x16x2_t {
74959
75047
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vorr))]
74960
75048
#[cfg_attr(
74961
75049
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74962
- assert_instr(zip)
75050
+ assert_instr(zip1)
75051
+ )]
75052
+ #[cfg_attr(
75053
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
75054
+ assert_instr(zip2)
74963
75055
)]
74964
75056
#[cfg_attr(
74965
75057
not(target_arch = "arm"),
@@ -74984,7 +75076,11 @@ pub fn vzipq_u16(a: uint16x8_t, b: uint16x8_t) -> uint16x8x2_t {
74984
75076
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vorr))]
74985
75077
#[cfg_attr(
74986
75078
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
74987
- assert_instr(zip)
75079
+ assert_instr(zip1)
75080
+ )]
75081
+ #[cfg_attr(
75082
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
75083
+ assert_instr(zip2)
74988
75084
)]
74989
75085
#[cfg_attr(
74990
75086
not(target_arch = "arm"),
@@ -75009,7 +75105,11 @@ pub fn vzipq_u32(a: uint32x4_t, b: uint32x4_t) -> uint32x4x2_t {
75009
75105
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vorr))]
75010
75106
#[cfg_attr(
75011
75107
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
75012
- assert_instr(zip)
75108
+ assert_instr(zip1)
75109
+ )]
75110
+ #[cfg_attr(
75111
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
75112
+ assert_instr(zip2)
75013
75113
)]
75014
75114
#[cfg_attr(
75015
75115
not(target_arch = "arm"),
@@ -75042,7 +75142,11 @@ pub fn vzipq_p8(a: poly8x16_t, b: poly8x16_t) -> poly8x16x2_t {
75042
75142
#[cfg_attr(all(test, target_arch = "arm"), assert_instr(vorr))]
75043
75143
#[cfg_attr(
75044
75144
all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
75045
- assert_instr(zip)
75145
+ assert_instr(zip1)
75146
+ )]
75147
+ #[cfg_attr(
75148
+ all(test, any(target_arch = "aarch64", target_arch = "arm64ec")),
75149
+ assert_instr(zip2)
75046
75150
)]
75047
75151
#[cfg_attr(
75048
75152
not(target_arch = "arm"),
0 commit comments