@@ -2923,46 +2923,46 @@ pub unsafe fn _mm256_sll_epi64(a: __m256i, count: __m128i) -> __m256i {
2923
2923
transmute ( psllq ( a. as_i64x4 ( ) , count. as_i64x2 ( ) ) )
2924
2924
}
2925
2925
2926
- /// Shifts packed 16-bit integers in `a` left by `imm8 ` while
2926
+ /// Shifts packed 16-bit integers in `a` left by `IMM8 ` while
2927
2927
/// shifting in zeros, return the results;
2928
2928
///
2929
2929
/// [Intel's documentation](https://software.intel.com/sites/landingpage/IntrinsicsGuide/#text=_mm256_slli_epi16)
2930
2930
#[ inline]
2931
2931
#[ target_feature( enable = "avx2" ) ]
2932
- #[ cfg_attr( test, assert_instr( vpsllw, imm8 = 7 ) ) ]
2932
+ #[ cfg_attr( test, assert_instr( vpsllw, IMM8 = 7 ) ) ]
2933
2933
#[ rustc_legacy_const_generics( 1 ) ]
2934
2934
#[ stable( feature = "simd_x86" , since = "1.27.0" ) ]
2935
- pub unsafe fn _mm256_slli_epi16 < const imm8 : i32 > ( a : __m256i ) -> __m256i {
2936
- static_assert_imm8 ! ( imm8 ) ;
2937
- transmute ( pslliw ( a. as_i16x16 ( ) , imm8 ) )
2935
+ pub unsafe fn _mm256_slli_epi16 < const IMM8 : i32 > ( a : __m256i ) -> __m256i {
2936
+ static_assert_imm8 ! ( IMM8 ) ;
2937
+ transmute ( pslliw ( a. as_i16x16 ( ) , IMM8 ) )
2938
2938
}
2939
2939
2940
- /// Shifts packed 32-bit integers in `a` left by `imm8 ` while
2940
+ /// Shifts packed 32-bit integers in `a` left by `IMM8 ` while
2941
2941
/// shifting in zeros, return the results;
2942
2942
///
2943
2943
/// [Intel's documentation](https://software.intel.com/sites/landingpage/IntrinsicsGuide/#text=_mm256_slli_epi32)
2944
2944
#[ inline]
2945
2945
#[ target_feature( enable = "avx2" ) ]
2946
- #[ cfg_attr( test, assert_instr( vpslld, imm8 = 7 ) ) ]
2946
+ #[ cfg_attr( test, assert_instr( vpslld, IMM8 = 7 ) ) ]
2947
2947
#[ rustc_legacy_const_generics( 1 ) ]
2948
2948
#[ stable( feature = "simd_x86" , since = "1.27.0" ) ]
2949
- pub unsafe fn _mm256_slli_epi32 < const imm8 : i32 > ( a : __m256i ) -> __m256i {
2950
- static_assert_imm8 ! ( imm8 ) ;
2951
- transmute ( psllid ( a. as_i32x8 ( ) , imm8 ) )
2949
+ pub unsafe fn _mm256_slli_epi32 < const IMM8 : i32 > ( a : __m256i ) -> __m256i {
2950
+ static_assert_imm8 ! ( IMM8 ) ;
2951
+ transmute ( psllid ( a. as_i32x8 ( ) , IMM8 ) )
2952
2952
}
2953
2953
2954
- /// Shifts packed 64-bit integers in `a` left by `imm8 ` while
2954
+ /// Shifts packed 64-bit integers in `a` left by `IMM8 ` while
2955
2955
/// shifting in zeros, return the results;
2956
2956
///
2957
2957
/// [Intel's documentation](https://software.intel.com/sites/landingpage/IntrinsicsGuide/#text=_mm256_slli_epi64)
2958
2958
#[ inline]
2959
2959
#[ target_feature( enable = "avx2" ) ]
2960
- #[ cfg_attr( test, assert_instr( vpsllq, imm8 = 7 ) ) ]
2960
+ #[ cfg_attr( test, assert_instr( vpsllq, IMM8 = 7 ) ) ]
2961
2961
#[ rustc_legacy_const_generics( 1 ) ]
2962
2962
#[ stable( feature = "simd_x86" , since = "1.27.0" ) ]
2963
- pub unsafe fn _mm256_slli_epi64 < const imm8 : i32 > ( a : __m256i ) -> __m256i {
2964
- static_assert_imm8 ! ( imm8 ) ;
2965
- transmute ( pslliq ( a. as_i64x4 ( ) , imm8 ) )
2963
+ pub unsafe fn _mm256_slli_epi64 < const IMM8 : i32 > ( a : __m256i ) -> __m256i {
2964
+ static_assert_imm8 ! ( IMM8 ) ;
2965
+ transmute ( pslliq ( a. as_i64x4 ( ) , IMM8 ) )
2966
2966
}
2967
2967
2968
2968
/// Shifts 128-bit lanes in `a` left by `imm8` bytes while shifting in zeros.
@@ -3077,32 +3077,32 @@ pub unsafe fn _mm256_sra_epi32(a: __m256i, count: __m128i) -> __m256i {
3077
3077
transmute ( psrad ( a. as_i32x8 ( ) , count. as_i32x4 ( ) ) )
3078
3078
}
3079
3079
3080
- /// Shifts packed 16-bit integers in `a` right by `imm8 ` while
3080
+ /// Shifts packed 16-bit integers in `a` right by `IMM8 ` while
3081
3081
/// shifting in sign bits.
3082
3082
///
3083
3083
/// [Intel's documentation](https://software.intel.com/sites/landingpage/IntrinsicsGuide/#text=_mm256_srai_epi16)
3084
3084
#[ inline]
3085
3085
#[ target_feature( enable = "avx2" ) ]
3086
- #[ cfg_attr( test, assert_instr( vpsraw, imm8 = 7 ) ) ]
3086
+ #[ cfg_attr( test, assert_instr( vpsraw, IMM8 = 7 ) ) ]
3087
3087
#[ rustc_legacy_const_generics( 1 ) ]
3088
3088
#[ stable( feature = "simd_x86" , since = "1.27.0" ) ]
3089
- pub unsafe fn _mm256_srai_epi16 < const imm8 : i32 > ( a : __m256i ) -> __m256i {
3090
- static_assert_imm8 ! ( imm8 ) ;
3091
- transmute ( psraiw ( a. as_i16x16 ( ) , imm8 ) )
3089
+ pub unsafe fn _mm256_srai_epi16 < const IMM8 : i32 > ( a : __m256i ) -> __m256i {
3090
+ static_assert_imm8 ! ( IMM8 ) ;
3091
+ transmute ( psraiw ( a. as_i16x16 ( ) , IMM8 ) )
3092
3092
}
3093
3093
3094
- /// Shifts packed 32-bit integers in `a` right by `imm8 ` while
3094
+ /// Shifts packed 32-bit integers in `a` right by `IMM8 ` while
3095
3095
/// shifting in sign bits.
3096
3096
///
3097
3097
/// [Intel's documentation](https://software.intel.com/sites/landingpage/IntrinsicsGuide/#text=_mm256_srai_epi32)
3098
3098
#[ inline]
3099
3099
#[ target_feature( enable = "avx2" ) ]
3100
- #[ cfg_attr( test, assert_instr( vpsrad, imm8 = 7 ) ) ]
3100
+ #[ cfg_attr( test, assert_instr( vpsrad, IMM8 = 7 ) ) ]
3101
3101
#[ rustc_legacy_const_generics( 1 ) ]
3102
3102
#[ stable( feature = "simd_x86" , since = "1.27.0" ) ]
3103
- pub unsafe fn _mm256_srai_epi32 < const imm8 : i32 > ( a : __m256i ) -> __m256i {
3104
- static_assert_imm8 ! ( imm8 ) ;
3105
- transmute ( psraid ( a. as_i32x8 ( ) , imm8 ) )
3103
+ pub unsafe fn _mm256_srai_epi32 < const IMM8 : i32 > ( a : __m256i ) -> __m256i {
3104
+ static_assert_imm8 ! ( IMM8 ) ;
3105
+ transmute ( psraid ( a. as_i32x8 ( ) , IMM8 ) )
3106
3106
}
3107
3107
3108
3108
/// Shifts packed 32-bit integers in `a` right by the amount specified by the
@@ -3201,46 +3201,46 @@ pub unsafe fn _mm256_srl_epi64(a: __m256i, count: __m128i) -> __m256i {
3201
3201
transmute ( psrlq ( a. as_i64x4 ( ) , count. as_i64x2 ( ) ) )
3202
3202
}
3203
3203
3204
- /// Shifts packed 16-bit integers in `a` right by `imm8 ` while shifting in
3204
+ /// Shifts packed 16-bit integers in `a` right by `IMM8 ` while shifting in
3205
3205
/// zeros
3206
3206
///
3207
3207
/// [Intel's documentation](https://software.intel.com/sites/landingpage/IntrinsicsGuide/#text=_mm256_srli_epi16)
3208
3208
#[ inline]
3209
3209
#[ target_feature( enable = "avx2" ) ]
3210
- #[ cfg_attr( test, assert_instr( vpsrlw, imm8 = 7 ) ) ]
3210
+ #[ cfg_attr( test, assert_instr( vpsrlw, IMM8 = 7 ) ) ]
3211
3211
#[ rustc_legacy_const_generics( 1 ) ]
3212
3212
#[ stable( feature = "simd_x86" , since = "1.27.0" ) ]
3213
- pub unsafe fn _mm256_srli_epi16 < const imm8 : i32 > ( a : __m256i ) -> __m256i {
3214
- static_assert_imm8 ! ( imm8 ) ;
3215
- transmute ( psrliw ( a. as_i16x16 ( ) , imm8 ) )
3213
+ pub unsafe fn _mm256_srli_epi16 < const IMM8 : i32 > ( a : __m256i ) -> __m256i {
3214
+ static_assert_imm8 ! ( IMM8 ) ;
3215
+ transmute ( psrliw ( a. as_i16x16 ( ) , IMM8 ) )
3216
3216
}
3217
3217
3218
- /// Shifts packed 32-bit integers in `a` right by `imm8 ` while shifting in
3218
+ /// Shifts packed 32-bit integers in `a` right by `IMM8 ` while shifting in
3219
3219
/// zeros
3220
3220
///
3221
3221
/// [Intel's documentation](https://software.intel.com/sites/landingpage/IntrinsicsGuide/#text=_mm256_srli_epi32)
3222
3222
#[ inline]
3223
3223
#[ target_feature( enable = "avx2" ) ]
3224
- #[ cfg_attr( test, assert_instr( vpsrld, imm8 = 7 ) ) ]
3224
+ #[ cfg_attr( test, assert_instr( vpsrld, IMM8 = 7 ) ) ]
3225
3225
#[ rustc_legacy_const_generics( 1 ) ]
3226
3226
#[ stable( feature = "simd_x86" , since = "1.27.0" ) ]
3227
- pub unsafe fn _mm256_srli_epi32 < const imm8 : i32 > ( a : __m256i ) -> __m256i {
3228
- static_assert_imm8 ! ( imm8 ) ;
3229
- transmute ( psrlid ( a. as_i32x8 ( ) , imm8 ) )
3227
+ pub unsafe fn _mm256_srli_epi32 < const IMM8 : i32 > ( a : __m256i ) -> __m256i {
3228
+ static_assert_imm8 ! ( IMM8 ) ;
3229
+ transmute ( psrlid ( a. as_i32x8 ( ) , IMM8 ) )
3230
3230
}
3231
3231
3232
- /// Shifts packed 64-bit integers in `a` right by `imm8 ` while shifting in
3232
+ /// Shifts packed 64-bit integers in `a` right by `IMM8 ` while shifting in
3233
3233
/// zeros
3234
3234
///
3235
3235
/// [Intel's documentation](https://software.intel.com/sites/landingpage/IntrinsicsGuide/#text=_mm256_srli_epi64)
3236
3236
#[ inline]
3237
3237
#[ target_feature( enable = "avx2" ) ]
3238
- #[ cfg_attr( test, assert_instr( vpsrlq, imm8 = 7 ) ) ]
3238
+ #[ cfg_attr( test, assert_instr( vpsrlq, IMM8 = 7 ) ) ]
3239
3239
#[ rustc_legacy_const_generics( 1 ) ]
3240
3240
#[ stable( feature = "simd_x86" , since = "1.27.0" ) ]
3241
- pub unsafe fn _mm256_srli_epi64 < const imm8 : i32 > ( a : __m256i ) -> __m256i {
3242
- static_assert_imm8 ! ( imm8 ) ;
3243
- transmute ( psrliq ( a. as_i64x4 ( ) , imm8 ) )
3241
+ pub unsafe fn _mm256_srli_epi64 < const IMM8 : i32 > ( a : __m256i ) -> __m256i {
3242
+ static_assert_imm8 ! ( IMM8 ) ;
3243
+ transmute ( psrliq ( a. as_i64x4 ( ) , IMM8 ) )
3244
3244
}
3245
3245
3246
3246
/// Shifts packed 32-bit integers in `a` right by the amount specified by
0 commit comments