Message ID | 20210915090807.17351-1-hongyu.wang@intel.com |
---|---|
State | New |
Headers | show |
Series | AVX512FP16: Adjust builtin name for FP16 builtins to match AVX512F style | expand |
On Wed, Sep 15, 2021 at 5:08 PM Hongyu Wang via Gcc-patches <gcc-patches@gcc.gnu.org> wrote: > > Hi, > > For AVX512FP16 builtins, they all contain format like vaddph_v8hf, > while AVX512F builtins use addps128 which succeeded SSE/AVX style. > Adjust AVX512FP16 builtins to match such format. > > Bootstraped/regtested on x86-64-*-linux, ok for master? > > gcc/ChangeLog: > > * config/i386/avx512fp16intrin.h: Adjust all builtin calls. > * config/i386/avx512fp16vlintrin.h: Likewise. > * config/i386/i386-builtin.def: Adjust builtin name to match > AVX512F style. > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask, "__builtin_ia32_maxsh_mask", IX86_BUILTIN_VMAXSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask, "__builtin_ia32_minsh_mask", IX86_BUILTIN_VMINSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_cmpv8hf3_mask, "__builtin_ia32_cmpph128_mask", IX86_BUILTIN_VCMPPH_V8HF_MASK, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_cmpv16hf3_mask, "__builtin_ia32_cmpph256_mask", IX86_BUILTIN_VCMPPH_V16HF_MASK, UNKNOWN, (int) UHI_FTYPE_V16HF_V16HF_INT_UHI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask, "__builtin_ia32_cmpph512_mask", IX86_BUILTIN_VCMPPH_V32HF_MASK, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv8hf2_mask, "__builtin_ia32_sqrtph128_mask", IX86_BUILTIN_VSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv16hf2_mask, "__builtin_ia32_sqrtph256_mask", IX86_BUILTIN_VSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv8hf2_mask, "__builtin_ia32_rsqrtph128_mask", IX86_BUILTIN_VRSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv16hf2_mask, "__builtin_ia32_rsqrtph256_mask", IX86_BUILTIN_VRSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv32hf2_mask, "__builtin_ia32_rsqrtph512_mask", IX86_BUILTIN_VRSQRTPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrsqrtv8hf2_mask, "__builtin_ia32_rsqrtsh_mask", IX86_BUILTIN_VRSQRTSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv8hf2_mask, "__builtin_ia32_rcpph128_mask", IX86_BUILTIN_VRCPPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv16hf2_mask, "__builtin_ia32_rcpph256_mask", IX86_BUILTIN_VRCPPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv32hf2_mask, "__builtin_ia32_rcpph512_mask", IX86_BUILTIN_VRCPPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrcpv8hf2_mask, "__builtin_ia32_rcpsh_mask", IX86_BUILTIN_VRCPSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_scalefv8hf_mask, "__builtin_ia32_scalefph128_mask", IX86_BUILTIN_VSCALEFPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_scalefv16hf_mask, "__builtin_ia32_scalefph256_mask", IX86_BUILTIN_VSCALEFPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) Could you also change IX86_BUILTIN_VSCALEFPH_V16HF_MASK IX86_BUILTIN_VSCALEFPH256_MASK, similar for others. > gcc/testsuite/ChangeLog: > > * gcc.target/i386/avx-1.c: Adjust builtin macros. > * gcc.target/i386/sse-13.c: Likewise. > * gcc.target/i386/sse-23.c: Likewise. > --- > gcc/config/i386/avx512fp16intrin.h | 1236 ++++++++++++------------ > gcc/config/i386/avx512fp16vlintrin.h | 454 ++++----- > gcc/config/i386/i386-builtin.def | 130 +-- > gcc/testsuite/gcc.target/i386/avx-1.c | 58 +- > gcc/testsuite/gcc.target/i386/sse-13.c | 58 +- > gcc/testsuite/gcc.target/i386/sse-23.c | 58 +- > 6 files changed, 997 insertions(+), 997 deletions(-) > > diff --git a/gcc/config/i386/avx512fp16intrin.h b/gcc/config/i386/avx512fp16intrin.h > index 2fbfc140c44..5d66ca5c820 100644 > --- a/gcc/config/i386/avx512fp16intrin.h > +++ b/gcc/config/i386/avx512fp16intrin.h > @@ -229,15 +229,15 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_add_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > { > - return __builtin_ia32_vaddph_v32hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_addph512_mask (__C, __D, __A, __B); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_add_ph (__mmask32 __A, __m512h __B, __m512h __C) > { > - return __builtin_ia32_vaddph_v32hf_mask (__B, __C, > - _mm512_setzero_ph (), __A); > + return __builtin_ia32_addph512_mask (__B, __C, > + _mm512_setzero_ph (), __A); > } > > extern __inline __m512h > @@ -251,15 +251,15 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_sub_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > { > - return __builtin_ia32_vsubph_v32hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_subph512_mask (__C, __D, __A, __B); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_sub_ph (__mmask32 __A, __m512h __B, __m512h __C) > { > - return __builtin_ia32_vsubph_v32hf_mask (__B, __C, > - _mm512_setzero_ph (), __A); > + return __builtin_ia32_subph512_mask (__B, __C, > + _mm512_setzero_ph (), __A); > } > > extern __inline __m512h > @@ -273,15 +273,15 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_mul_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > { > - return __builtin_ia32_vmulph_v32hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_mulph512_mask (__C, __D, __A, __B); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_mul_ph (__mmask32 __A, __m512h __B, __m512h __C) > { > - return __builtin_ia32_vmulph_v32hf_mask (__B, __C, > - _mm512_setzero_ph (), __A); > + return __builtin_ia32_mulph512_mask (__B, __C, > + _mm512_setzero_ph (), __A); > } > > extern __inline __m512h > @@ -295,15 +295,15 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_div_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > { > - return __builtin_ia32_vdivph_v32hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_divph512_mask (__C, __D, __A, __B); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_div_ph (__mmask32 __A, __m512h __B, __m512h __C) > { > - return __builtin_ia32_vdivph_v32hf_mask (__B, __C, > - _mm512_setzero_ph (), __A); > + return __builtin_ia32_divph512_mask (__B, __C, > + _mm512_setzero_ph (), __A); > } > > #ifdef __OPTIMIZE__ > @@ -311,9 +311,9 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_add_round_ph (__m512h __A, __m512h __B, const int __C) > { > - return __builtin_ia32_vaddph_v32hf_mask_round (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1, __C); > + return __builtin_ia32_addph512_mask_round (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1, __C); > } > > extern __inline __m512h > @@ -321,7 +321,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_add_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > __m512h __D, const int __E) > { > - return __builtin_ia32_vaddph_v32hf_mask_round (__C, __D, __A, __B, __E); > + return __builtin_ia32_addph512_mask_round (__C, __D, __A, __B, __E); > } > > extern __inline __m512h > @@ -329,18 +329,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_add_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > const int __D) > { > - return __builtin_ia32_vaddph_v32hf_mask_round (__B, __C, > - _mm512_setzero_ph (), > - __A, __D); > + return __builtin_ia32_addph512_mask_round (__B, __C, > + _mm512_setzero_ph (), > + __A, __D); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_sub_round_ph (__m512h __A, __m512h __B, const int __C) > { > - return __builtin_ia32_vsubph_v32hf_mask_round (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1, __C); > + return __builtin_ia32_subph512_mask_round (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1, __C); > } > > extern __inline __m512h > @@ -348,7 +348,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_sub_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > __m512h __D, const int __E) > { > - return __builtin_ia32_vsubph_v32hf_mask_round (__C, __D, __A, __B, __E); > + return __builtin_ia32_subph512_mask_round (__C, __D, __A, __B, __E); > } > > extern __inline __m512h > @@ -356,18 +356,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_sub_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > const int __D) > { > - return __builtin_ia32_vsubph_v32hf_mask_round (__B, __C, > - _mm512_setzero_ph (), > - __A, __D); > + return __builtin_ia32_subph512_mask_round (__B, __C, > + _mm512_setzero_ph (), > + __A, __D); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mul_round_ph (__m512h __A, __m512h __B, const int __C) > { > - return __builtin_ia32_vmulph_v32hf_mask_round (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1, __C); > + return __builtin_ia32_mulph512_mask_round (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1, __C); > } > > extern __inline __m512h > @@ -375,7 +375,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_mul_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > __m512h __D, const int __E) > { > - return __builtin_ia32_vmulph_v32hf_mask_round (__C, __D, __A, __B, __E); > + return __builtin_ia32_mulph512_mask_round (__C, __D, __A, __B, __E); > } > > extern __inline __m512h > @@ -383,18 +383,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_mul_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > const int __D) > { > - return __builtin_ia32_vmulph_v32hf_mask_round (__B, __C, > - _mm512_setzero_ph (), > - __A, __D); > + return __builtin_ia32_mulph512_mask_round (__B, __C, > + _mm512_setzero_ph (), > + __A, __D); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_div_round_ph (__m512h __A, __m512h __B, const int __C) > { > - return __builtin_ia32_vdivph_v32hf_mask_round (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1, __C); > + return __builtin_ia32_divph512_mask_round (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1, __C); > } > > extern __inline __m512h > @@ -402,7 +402,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_div_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > __m512h __D, const int __E) > { > - return __builtin_ia32_vdivph_v32hf_mask_round (__C, __D, __A, __B, __E); > + return __builtin_ia32_divph512_mask_round (__C, __D, __A, __B, __E); > } > > extern __inline __m512h > @@ -410,67 +410,67 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_div_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > const int __D) > { > - return __builtin_ia32_vdivph_v32hf_mask_round (__B, __C, > - _mm512_setzero_ph (), > - __A, __D); > + return __builtin_ia32_divph512_mask_round (__B, __C, > + _mm512_setzero_ph (), > + __A, __D); > } > #else > #define _mm512_add_round_ph(A, B, C) \ > - ((__m512h)__builtin_ia32_vaddph_v32hf_mask_round((A), (B), \ > - _mm512_setzero_ph (),\ > - (__mmask32)-1, (C))) > + ((__m512h)__builtin_ia32_addph512_mask_round((A), (B), \ > + _mm512_setzero_ph (), \ > + (__mmask32)-1, (C))) > > -#define _mm512_mask_add_round_ph(A, B, C, D, E) \ > - ((__m512h)__builtin_ia32_vaddph_v32hf_mask_round((C), (D), (A), (B), (E))) > +#define _mm512_mask_add_round_ph(A, B, C, D, E) \ > + ((__m512h)__builtin_ia32_addph512_mask_round((C), (D), (A), (B), (E))) > > #define _mm512_maskz_add_round_ph(A, B, C, D) \ > - ((__m512h)__builtin_ia32_vaddph_v32hf_mask_round((B), (C), \ > - _mm512_setzero_ph (),\ > - (A), (D))) > + ((__m512h)__builtin_ia32_addph512_mask_round((B), (C), \ > + _mm512_setzero_ph (), \ > + (A), (D))) > > #define _mm512_sub_round_ph(A, B, C) \ > - ((__m512h)__builtin_ia32_vsubph_v32hf_mask_round((A), (B), \ > - _mm512_setzero_ph (),\ > - (__mmask32)-1, (C))) > + ((__m512h)__builtin_ia32_subph512_mask_round((A), (B), \ > + _mm512_setzero_ph (), \ > + (__mmask32)-1, (C))) > > -#define _mm512_mask_sub_round_ph(A, B, C, D, E) \ > - ((__m512h)__builtin_ia32_vsubph_v32hf_mask_round((C), (D), (A), (B), (E))) > +#define _mm512_mask_sub_round_ph(A, B, C, D, E) \ > + ((__m512h)__builtin_ia32_subph512_mask_round((C), (D), (A), (B), (E))) > > #define _mm512_maskz_sub_round_ph(A, B, C, D) \ > - ((__m512h)__builtin_ia32_vsubph_v32hf_mask_round((B), (C), \ > - _mm512_setzero_ph (),\ > - (A), (D))) > + ((__m512h)__builtin_ia32_subph512_mask_round((B), (C), \ > + _mm512_setzero_ph (), \ > + (A), (D))) > > #define _mm512_mul_round_ph(A, B, C) \ > - ((__m512h)__builtin_ia32_vmulph_v32hf_mask_round((A), (B), \ > - _mm512_setzero_ph (),\ > - (__mmask32)-1, (C))) > + ((__m512h)__builtin_ia32_mulph512_mask_round((A), (B), \ > + _mm512_setzero_ph (), \ > + (__mmask32)-1, (C))) > > -#define _mm512_mask_mul_round_ph(A, B, C, D, E) \ > - ((__m512h)__builtin_ia32_vmulph_v32hf_mask_round((C), (D), (A), (B), (E))) > +#define _mm512_mask_mul_round_ph(A, B, C, D, E) \ > + ((__m512h)__builtin_ia32_mulph512_mask_round((C), (D), (A), (B), (E))) > > #define _mm512_maskz_mul_round_ph(A, B, C, D) \ > - ((__m512h)__builtin_ia32_vmulph_v32hf_mask_round((B), (C), \ > - _mm512_setzero_ph (),\ > - (A), (D))) > + ((__m512h)__builtin_ia32_mulph512_mask_round((B), (C), \ > + _mm512_setzero_ph (), \ > + (A), (D))) > > #define _mm512_div_round_ph(A, B, C) \ > - ((__m512h)__builtin_ia32_vdivph_v32hf_mask_round((A), (B), \ > - _mm512_setzero_ph (),\ > - (__mmask32)-1, (C))) > + ((__m512h)__builtin_ia32_divph512_mask_round((A), (B), \ > + _mm512_setzero_ph (), \ > + (__mmask32)-1, (C))) > > -#define _mm512_mask_div_round_ph(A, B, C, D, E) \ > - ((__m512h)__builtin_ia32_vdivph_v32hf_mask_round((C), (D), (A), (B), (E))) > +#define _mm512_mask_div_round_ph(A, B, C, D, E) \ > + ((__m512h)__builtin_ia32_divph512_mask_round((C), (D), (A), (B), (E))) > > #define _mm512_maskz_div_round_ph(A, B, C, D) \ > - ((__m512h)__builtin_ia32_vdivph_v32hf_mask_round((B), (C), \ > - _mm512_setzero_ph (),\ > - (A), (D))) > + ((__m512h)__builtin_ia32_divph512_mask_round((B), (C), \ > + _mm512_setzero_ph (), \ > + (A), (D))) > #endif /* __OPTIMIZE__ */ > > /* Intrinsics of v[add,sub,mul,div]sh. */ > extern __inline __m128h > -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_add_sh (__m128h __A, __m128h __B) > { > __A[0] += __B[0]; > @@ -481,15 +481,15 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_add_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vaddsh_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_addsh_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_add_sh (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vaddsh_v8hf_mask (__B, __C, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_addsh_mask (__B, __C, _mm_setzero_ph (), > + __A); > } > > extern __inline __m128h > @@ -504,15 +504,15 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_sub_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vsubsh_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_subsh_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_sub_sh (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vsubsh_v8hf_mask (__B, __C, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_subsh_mask (__B, __C, _mm_setzero_ph (), > + __A); > } > > extern __inline __m128h > @@ -527,14 +527,14 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_mul_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vmulsh_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_mulsh_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_mul_sh (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vmulsh_v8hf_mask (__B, __C, _mm_setzero_ph (), __A); > + return __builtin_ia32_mulsh_mask (__B, __C, _mm_setzero_ph (), __A); > } > > extern __inline __m128h > @@ -549,15 +549,15 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_div_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vdivsh_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_divsh_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_div_sh (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vdivsh_v8hf_mask (__B, __C, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_divsh_mask (__B, __C, _mm_setzero_ph (), > + __A); > } > > #ifdef __OPTIMIZE__ > @@ -565,9 +565,9 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_add_round_sh (__m128h __A, __m128h __B, const int __C) > { > - return __builtin_ia32_vaddsh_v8hf_mask_round (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1, __C); > + return __builtin_ia32_addsh_mask_round (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1, __C); > } > > extern __inline __m128h > @@ -575,7 +575,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_add_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > __m128h __D, const int __E) > { > - return __builtin_ia32_vaddsh_v8hf_mask_round (__C, __D, __A, __B, __E); > + return __builtin_ia32_addsh_mask_round (__C, __D, __A, __B, __E); > } > > extern __inline __m128h > @@ -583,18 +583,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_add_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > const int __D) > { > - return __builtin_ia32_vaddsh_v8hf_mask_round (__B, __C, > - _mm_setzero_ph (), > - __A, __D); > + return __builtin_ia32_addsh_mask_round (__B, __C, > + _mm_setzero_ph (), > + __A, __D); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_sub_round_sh (__m128h __A, __m128h __B, const int __C) > { > - return __builtin_ia32_vsubsh_v8hf_mask_round (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1, __C); > + return __builtin_ia32_subsh_mask_round (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1, __C); > } > > extern __inline __m128h > @@ -602,7 +602,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_sub_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > __m128h __D, const int __E) > { > - return __builtin_ia32_vsubsh_v8hf_mask_round (__C, __D, __A, __B, __E); > + return __builtin_ia32_subsh_mask_round (__C, __D, __A, __B, __E); > } > > extern __inline __m128h > @@ -610,18 +610,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_sub_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > const int __D) > { > - return __builtin_ia32_vsubsh_v8hf_mask_round (__B, __C, > - _mm_setzero_ph (), > - __A, __D); > + return __builtin_ia32_subsh_mask_round (__B, __C, > + _mm_setzero_ph (), > + __A, __D); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mul_round_sh (__m128h __A, __m128h __B, const int __C) > { > - return __builtin_ia32_vmulsh_v8hf_mask_round (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1, __C); > + return __builtin_ia32_mulsh_mask_round (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1, __C); > } > > extern __inline __m128h > @@ -629,7 +629,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_mul_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > __m128h __D, const int __E) > { > - return __builtin_ia32_vmulsh_v8hf_mask_round (__C, __D, __A, __B, __E); > + return __builtin_ia32_mulsh_mask_round (__C, __D, __A, __B, __E); > } > > extern __inline __m128h > @@ -637,18 +637,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_mul_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > const int __D) > { > - return __builtin_ia32_vmulsh_v8hf_mask_round (__B, __C, > - _mm_setzero_ph (), > - __A, __D); > + return __builtin_ia32_mulsh_mask_round (__B, __C, > + _mm_setzero_ph (), > + __A, __D); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_div_round_sh (__m128h __A, __m128h __B, const int __C) > { > - return __builtin_ia32_vdivsh_v8hf_mask_round (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1, __C); > + return __builtin_ia32_divsh_mask_round (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1, __C); > } > > extern __inline __m128h > @@ -656,7 +656,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_div_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > __m128h __D, const int __E) > { > - return __builtin_ia32_vdivsh_v8hf_mask_round (__C, __D, __A, __B, __E); > + return __builtin_ia32_divsh_mask_round (__C, __D, __A, __B, __E); > } > > extern __inline __m128h > @@ -664,62 +664,62 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_div_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > const int __D) > { > - return __builtin_ia32_vdivsh_v8hf_mask_round (__B, __C, > - _mm_setzero_ph (), > - __A, __D); > + return __builtin_ia32_divsh_mask_round (__B, __C, > + _mm_setzero_ph (), > + __A, __D); > } > #else > #define _mm_add_round_sh(A, B, C) \ > - ((__m128h)__builtin_ia32_vaddsh_v8hf_mask_round ((A), (B), \ > - _mm_setzero_ph (), \ > - (__mmask8)-1, (C))) > + ((__m128h)__builtin_ia32_addsh_mask_round ((A), (B), \ > + _mm_setzero_ph (), \ > + (__mmask8)-1, (C))) > > #define _mm_mask_add_round_sh(A, B, C, D, E) \ > - ((__m128h)__builtin_ia32_vaddsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > + ((__m128h)__builtin_ia32_addsh_mask_round ((C), (D), (A), (B), (E))) > > -#define _mm_maskz_add_round_sh(A, B, C, D) \ > - ((__m128h)__builtin_ia32_vaddsh_v8hf_mask_round ((B), (C), \ > - _mm_setzero_ph (), \ > - (A), (D))) > +#define _mm_maskz_add_round_sh(A, B, C, D) \ > + ((__m128h)__builtin_ia32_addsh_mask_round ((B), (C), \ > + _mm_setzero_ph (), \ > + (A), (D))) > > #define _mm_sub_round_sh(A, B, C) \ > - ((__m128h)__builtin_ia32_vsubsh_v8hf_mask_round ((A), (B), \ > - _mm_setzero_ph (), \ > - (__mmask8)-1, (C))) > + ((__m128h)__builtin_ia32_subsh_mask_round ((A), (B), \ > + _mm_setzero_ph (), \ > + (__mmask8)-1, (C))) > > #define _mm_mask_sub_round_sh(A, B, C, D, E) \ > - ((__m128h)__builtin_ia32_vsubsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > + ((__m128h)__builtin_ia32_subsh_mask_round ((C), (D), (A), (B), (E))) > > -#define _mm_maskz_sub_round_sh(A, B, C, D) \ > - ((__m128h)__builtin_ia32_vsubsh_v8hf_mask_round ((B), (C), \ > - _mm_setzero_ph (), \ > - (A), (D))) > +#define _mm_maskz_sub_round_sh(A, B, C, D) \ > + ((__m128h)__builtin_ia32_subsh_mask_round ((B), (C), \ > + _mm_setzero_ph (), \ > + (A), (D))) > > #define _mm_mul_round_sh(A, B, C) \ > - ((__m128h)__builtin_ia32_vmulsh_v8hf_mask_round ((A), (B), \ > - _mm_setzero_ph (), \ > - (__mmask8)-1, (C))) > + ((__m128h)__builtin_ia32_mulsh_mask_round ((A), (B), \ > + _mm_setzero_ph (), \ > + (__mmask8)-1, (C))) > > #define _mm_mask_mul_round_sh(A, B, C, D, E) \ > - ((__m128h)__builtin_ia32_vmulsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > + ((__m128h)__builtin_ia32_mulsh_mask_round ((C), (D), (A), (B), (E))) > > -#define _mm_maskz_mul_round_sh(A, B, C, D) \ > - ((__m128h)__builtin_ia32_vmulsh_v8hf_mask_round ((B), (C), \ > - _mm_setzero_ph (), \ > - (A), (D))) > +#define _mm_maskz_mul_round_sh(A, B, C, D) \ > + ((__m128h)__builtin_ia32_mulsh_mask_round ((B), (C), \ > + _mm_setzero_ph (), \ > + (A), (D))) > > #define _mm_div_round_sh(A, B, C) \ > - ((__m128h)__builtin_ia32_vdivsh_v8hf_mask_round ((A), (B), \ > - _mm_setzero_ph (), \ > - (__mmask8)-1, (C))) > + ((__m128h)__builtin_ia32_divsh_mask_round ((A), (B), \ > + _mm_setzero_ph (), \ > + (__mmask8)-1, (C))) > > #define _mm_mask_div_round_sh(A, B, C, D, E) \ > - ((__m128h)__builtin_ia32_vdivsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > + ((__m128h)__builtin_ia32_divsh_mask_round ((C), (D), (A), (B), (E))) > > -#define _mm_maskz_div_round_sh(A, B, C, D) \ > - ((__m128h)__builtin_ia32_vdivsh_v8hf_mask_round ((B), (C), \ > - _mm_setzero_ph (), \ > - (A), (D))) > +#define _mm_maskz_div_round_sh(A, B, C, D) \ > + ((__m128h)__builtin_ia32_divsh_mask_round ((B), (C), \ > + _mm_setzero_ph (), \ > + (A), (D))) > #endif /* __OPTIMIZE__ */ > > /* Intrinsic vmaxph vminph. */ > @@ -727,48 +727,48 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_max_ph (__m512h __A, __m512h __B) > { > - return __builtin_ia32_vmaxph_v32hf_mask (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1); > + return __builtin_ia32_maxph512_mask (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_max_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > { > - return __builtin_ia32_vmaxph_v32hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_maxph512_mask (__C, __D, __A, __B); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_max_ph (__mmask32 __A, __m512h __B, __m512h __C) > { > - return __builtin_ia32_vmaxph_v32hf_mask (__B, __C, > - _mm512_setzero_ph (), __A); > + return __builtin_ia32_maxph512_mask (__B, __C, > + _mm512_setzero_ph (), __A); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_min_ph (__m512h __A, __m512h __B) > { > - return __builtin_ia32_vminph_v32hf_mask (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1); > + return __builtin_ia32_minph512_mask (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_min_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > { > - return __builtin_ia32_vminph_v32hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_minph512_mask (__C, __D, __A, __B); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_min_ph (__mmask32 __A, __m512h __B, __m512h __C) > { > - return __builtin_ia32_vminph_v32hf_mask (__B, __C, > - _mm512_setzero_ph (), __A); > + return __builtin_ia32_minph512_mask (__B, __C, > + _mm512_setzero_ph (), __A); > } > > #ifdef __OPTIMIZE__ > @@ -776,9 +776,9 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_max_round_ph (__m512h __A, __m512h __B, const int __C) > { > - return __builtin_ia32_vmaxph_v32hf_mask_round (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1, __C); > + return __builtin_ia32_maxph512_mask_round (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1, __C); > } > > extern __inline __m512h > @@ -786,7 +786,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_max_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > __m512h __D, const int __E) > { > - return __builtin_ia32_vmaxph_v32hf_mask_round (__C, __D, __A, __B, __E); > + return __builtin_ia32_maxph512_mask_round (__C, __D, __A, __B, __E); > } > > extern __inline __m512h > @@ -794,18 +794,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_max_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > const int __D) > { > - return __builtin_ia32_vmaxph_v32hf_mask_round (__B, __C, > - _mm512_setzero_ph (), > - __A, __D); > + return __builtin_ia32_maxph512_mask_round (__B, __C, > + _mm512_setzero_ph (), > + __A, __D); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_min_round_ph (__m512h __A, __m512h __B, const int __C) > { > - return __builtin_ia32_vminph_v32hf_mask_round (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1, __C); > + return __builtin_ia32_minph512_mask_round (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1, __C); > } > > extern __inline __m512h > @@ -813,7 +813,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_min_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > __m512h __D, const int __E) > { > - return __builtin_ia32_vminph_v32hf_mask_round (__C, __D, __A, __B, __E); > + return __builtin_ia32_minph512_mask_round (__C, __D, __A, __B, __E); > } > > extern __inline __m512h > @@ -821,37 +821,37 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_min_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > const int __D) > { > - return __builtin_ia32_vminph_v32hf_mask_round (__B, __C, > - _mm512_setzero_ph (), > - __A, __D); > + return __builtin_ia32_minph512_mask_round (__B, __C, > + _mm512_setzero_ph (), > + __A, __D); > } > > #else > -#define _mm512_max_round_ph(A, B, C) \ > - (__builtin_ia32_vmaxph_v32hf_mask_round ((A), (B), \ > - _mm512_setzero_ph (), \ > - (__mmask32)-1, (C))) > +#define _mm512_max_round_ph(A, B, C) \ > + (__builtin_ia32_maxph512_mask_round ((A), (B), \ > + _mm512_setzero_ph (), \ > + (__mmask32)-1, (C))) > > #define _mm512_mask_max_round_ph(A, B, C, D, E) \ > - (__builtin_ia32_vmaxph_v32hf_mask_round ((C), (D), (A), (B), (E))) > + (__builtin_ia32_maxph512_mask_round ((C), (D), (A), (B), (E))) > > -#define _mm512_maskz_max_round_ph(A, B, C, D) \ > - (__builtin_ia32_vmaxph_v32hf_mask_round ((B), (C), \ > - _mm512_setzero_ph (), \ > - (A), (D))) > +#define _mm512_maskz_max_round_ph(A, B, C, D) \ > + (__builtin_ia32_maxph512_mask_round ((B), (C), \ > + _mm512_setzero_ph (), \ > + (A), (D))) > > -#define _mm512_min_round_ph(A, B, C) \ > - (__builtin_ia32_vminph_v32hf_mask_round ((A), (B), \ > - _mm512_setzero_ph (), \ > - (__mmask32)-1, (C))) > +#define _mm512_min_round_ph(A, B, C) \ > + (__builtin_ia32_minph512_mask_round ((A), (B), \ > + _mm512_setzero_ph (), \ > + (__mmask32)-1, (C))) > > #define _mm512_mask_min_round_ph(A, B, C, D, E) \ > - (__builtin_ia32_vminph_v32hf_mask_round ((C), (D), (A), (B), (E))) > + (__builtin_ia32_minph512_mask_round ((C), (D), (A), (B), (E))) > > -#define _mm512_maskz_min_round_ph(A, B, C, D) \ > - (__builtin_ia32_vminph_v32hf_mask_round ((B), (C), \ > - _mm512_setzero_ph (), \ > - (A), (D))) > +#define _mm512_maskz_min_round_ph(A, B, C, D) \ > + (__builtin_ia32_minph512_mask_round ((B), (C), \ > + _mm512_setzero_ph (), \ > + (A), (D))) > #endif /* __OPTIMIZE__ */ > > /* Intrinsic vmaxsh vminsh. */ > @@ -867,15 +867,15 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_max_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vmaxsh_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_maxsh_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_max_sh (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vmaxsh_v8hf_mask (__B, __C, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_maxsh_mask (__B, __C, _mm_setzero_ph (), > + __A); > } > > extern __inline __m128h > @@ -890,15 +890,15 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_min_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vminsh_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_minsh_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_min_sh (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vminsh_v8hf_mask (__B, __C, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_minsh_mask (__B, __C, _mm_setzero_ph (), > + __A); > } > > #ifdef __OPTIMIZE__ > @@ -906,9 +906,9 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_max_round_sh (__m128h __A, __m128h __B, const int __C) > { > - return __builtin_ia32_vmaxsh_v8hf_mask_round (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1, __C); > + return __builtin_ia32_maxsh_mask_round (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1, __C); > } > > extern __inline __m128h > @@ -916,7 +916,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_max_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > __m128h __D, const int __E) > { > - return __builtin_ia32_vmaxsh_v8hf_mask_round (__C, __D, __A, __B, __E); > + return __builtin_ia32_maxsh_mask_round (__C, __D, __A, __B, __E); > } > > extern __inline __m128h > @@ -924,18 +924,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_max_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > const int __D) > { > - return __builtin_ia32_vmaxsh_v8hf_mask_round (__B, __C, > - _mm_setzero_ph (), > - __A, __D); > + return __builtin_ia32_maxsh_mask_round (__B, __C, > + _mm_setzero_ph (), > + __A, __D); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_min_round_sh (__m128h __A, __m128h __B, const int __C) > { > - return __builtin_ia32_vminsh_v8hf_mask_round (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1, __C); > + return __builtin_ia32_minsh_mask_round (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1, __C); > } > > extern __inline __m128h > @@ -943,7 +943,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_min_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > __m128h __D, const int __E) > { > - return __builtin_ia32_vminsh_v8hf_mask_round (__C, __D, __A, __B, __E); > + return __builtin_ia32_minsh_mask_round (__C, __D, __A, __B, __E); > } > > extern __inline __m128h > @@ -951,37 +951,37 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_min_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > const int __D) > { > - return __builtin_ia32_vminsh_v8hf_mask_round (__B, __C, > - _mm_setzero_ph (), > - __A, __D); > + return __builtin_ia32_minsh_mask_round (__B, __C, > + _mm_setzero_ph (), > + __A, __D); > } > > #else > -#define _mm_max_round_sh(A, B, C) \ > - (__builtin_ia32_vmaxsh_v8hf_mask_round ((A), (B), \ > - _mm_setzero_ph (), \ > - (__mmask8)-1, (C))) > +#define _mm_max_round_sh(A, B, C) \ > + (__builtin_ia32_maxsh_mask_round ((A), (B), \ > + _mm_setzero_ph (), \ > + (__mmask8)-1, (C))) > > -#define _mm_mask_max_round_sh(A, B, C, D, E) \ > - (__builtin_ia32_vmaxsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > +#define _mm_mask_max_round_sh(A, B, C, D, E) \ > + (__builtin_ia32_maxsh_mask_round ((C), (D), (A), (B), (E))) > > -#define _mm_maskz_max_round_sh(A, B, C, D) \ > - (__builtin_ia32_vmaxsh_v8hf_mask_round ((B), (C), \ > - _mm_setzero_ph (), \ > - (A), (D))) > +#define _mm_maskz_max_round_sh(A, B, C, D) \ > + (__builtin_ia32_maxsh_mask_round ((B), (C), \ > + _mm_setzero_ph (), \ > + (A), (D))) > > -#define _mm_min_round_sh(A, B, C) \ > - (__builtin_ia32_vminsh_v8hf_mask_round ((A), (B), \ > - _mm_setzero_ph (), \ > - (__mmask8)-1, (C))) > +#define _mm_min_round_sh(A, B, C) \ > + (__builtin_ia32_minsh_mask_round ((A), (B), \ > + _mm_setzero_ph (), \ > + (__mmask8)-1, (C))) > > -#define _mm_mask_min_round_sh(A, B, C, D, E) \ > - (__builtin_ia32_vminsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > +#define _mm_mask_min_round_sh(A, B, C, D, E) \ > + (__builtin_ia32_minsh_mask_round ((C), (D), (A), (B), (E))) > > -#define _mm_maskz_min_round_sh(A, B, C, D) \ > - (__builtin_ia32_vminsh_v8hf_mask_round ((B), (C), \ > - _mm_setzero_ph (), \ > - (A), (D))) > +#define _mm_maskz_min_round_sh(A, B, C, D) \ > + (__builtin_ia32_minsh_mask_round ((B), (C), \ > + _mm_setzero_ph (), \ > + (A), (D))) > > #endif /* __OPTIMIZE__ */ > > @@ -991,8 +991,8 @@ extern __inline __mmask32 > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_cmp_ph_mask (__m512h __A, __m512h __B, const int __C) > { > - return (__mmask32) __builtin_ia32_vcmpph_v32hf_mask (__A, __B, __C, > - (__mmask32) -1); > + return (__mmask32) __builtin_ia32_cmpph512_mask (__A, __B, __C, > + (__mmask32) -1); > } > > extern __inline __mmask32 > @@ -1000,8 +1000,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_cmp_ph_mask (__mmask32 __A, __m512h __B, __m512h __C, > const int __D) > { > - return (__mmask32) __builtin_ia32_vcmpph_v32hf_mask (__B, __C, __D, > - __A); > + return (__mmask32) __builtin_ia32_cmpph512_mask (__B, __C, __D, > + __A); > } > > extern __inline __mmask32 > @@ -1009,9 +1009,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_cmp_round_ph_mask (__m512h __A, __m512h __B, const int __C, > const int __D) > { > - return (__mmask32) __builtin_ia32_vcmpph_v32hf_mask_round (__A, __B, > - __C, (__mmask32) -1, > - __D); > + return (__mmask32) __builtin_ia32_cmpph512_mask_round (__A, __B, > + __C, (__mmask32) -1, > + __D); > } > > extern __inline __mmask32 > @@ -1019,23 +1019,23 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_cmp_round_ph_mask (__mmask32 __A, __m512h __B, __m512h __C, > const int __D, const int __E) > { > - return (__mmask32) __builtin_ia32_vcmpph_v32hf_mask_round (__B, __C, > - __D, __A, > - __E); > + return (__mmask32) __builtin_ia32_cmpph512_mask_round (__B, __C, > + __D, __A, > + __E); > } > > #else > #define _mm512_cmp_ph_mask(A, B, C) \ > - (__builtin_ia32_vcmpph_v32hf_mask ((A), (B), (C), (-1))) > + (__builtin_ia32_cmpph512_mask ((A), (B), (C), (-1))) > > #define _mm512_mask_cmp_ph_mask(A, B, C, D) \ > - (__builtin_ia32_vcmpph_v32hf_mask ((B), (C), (D), (A))) > + (__builtin_ia32_cmpph512_mask ((B), (C), (D), (A))) > > -#define _mm512_cmp_round_ph_mask(A, B, C, D) \ > - (__builtin_ia32_vcmpph_v32hf_mask_round ((A), (B), (C), (-1), (D))) > +#define _mm512_cmp_round_ph_mask(A, B, C, D) \ > + (__builtin_ia32_cmpph512_mask_round ((A), (B), (C), (-1), (D))) > > -#define _mm512_mask_cmp_round_ph_mask(A, B, C, D, E) \ > - (__builtin_ia32_vcmpph_v32hf_mask_round ((B), (C), (D), (A), (E))) > +#define _mm512_mask_cmp_round_ph_mask(A, B, C, D, E) \ > + (__builtin_ia32_cmpph512_mask_round ((B), (C), (D), (A), (E))) > > #endif /* __OPTIMIZE__ */ > > @@ -1046,9 +1046,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_cmp_sh_mask (__m128h __A, __m128h __B, const int __C) > { > return (__mmask8) > - __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, > - __C, (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + __builtin_ia32_cmpsh_mask_round (__A, __B, > + __C, (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __mmask8 > @@ -1057,9 +1057,9 @@ _mm_mask_cmp_sh_mask (__mmask8 __A, __m128h __B, __m128h __C, > const int __D) > { > return (__mmask8) > - __builtin_ia32_vcmpsh_v8hf_mask_round (__B, __C, > - __D, __A, > - _MM_FROUND_CUR_DIRECTION); > + __builtin_ia32_cmpsh_mask_round (__B, __C, > + __D, __A, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __mmask8 > @@ -1067,9 +1067,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_cmp_round_sh_mask (__m128h __A, __m128h __B, const int __C, > const int __D) > { > - return (__mmask8) __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, > - __C, (__mmask8) -1, > - __D); > + return (__mmask8) __builtin_ia32_cmpsh_mask_round (__A, __B, > + __C, (__mmask8) -1, > + __D); > } > > extern __inline __mmask8 > @@ -1077,25 +1077,25 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_cmp_round_sh_mask (__mmask8 __A, __m128h __B, __m128h __C, > const int __D, const int __E) > { > - return (__mmask8) __builtin_ia32_vcmpsh_v8hf_mask_round (__B, __C, > - __D, __A, > - __E); > + return (__mmask8) __builtin_ia32_cmpsh_mask_round (__B, __C, > + __D, __A, > + __E); > } > > #else > -#define _mm_cmp_sh_mask(A, B, C) \ > - (__builtin_ia32_vcmpsh_v8hf_mask_round ((A), (B), (C), (-1), \ > - (_MM_FROUND_CUR_DIRECTION))) > +#define _mm_cmp_sh_mask(A, B, C) \ > + (__builtin_ia32_cmpsh_mask_round ((A), (B), (C), (-1), \ > + (_MM_FROUND_CUR_DIRECTION))) > > -#define _mm_mask_cmp_sh_mask(A, B, C, D) \ > - (__builtin_ia32_vcmpsh_v8hf_mask_round ((B), (C), (D), (A), \ > - (_MM_FROUND_CUR_DIRECTION))) > +#define _mm_mask_cmp_sh_mask(A, B, C, D) \ > + (__builtin_ia32_cmpsh_mask_round ((B), (C), (D), (A), \ > + (_MM_FROUND_CUR_DIRECTION))) > > -#define _mm_cmp_round_sh_mask(A, B, C, D) \ > - (__builtin_ia32_vcmpsh_v8hf_mask_round ((A), (B), (C), (-1), (D))) > +#define _mm_cmp_round_sh_mask(A, B, C, D) \ > + (__builtin_ia32_cmpsh_mask_round ((A), (B), (C), (-1), (D))) > > -#define _mm_mask_cmp_round_sh_mask(A, B, C, D, E) \ > - (__builtin_ia32_vcmpsh_v8hf_mask_round ((B), (C), (D), (A), (E))) > +#define _mm_mask_cmp_round_sh_mask(A, B, C, D, E) \ > + (__builtin_ia32_cmpsh_mask_round ((B), (C), (D), (A), (E))) > > #endif /* __OPTIMIZE__ */ > > @@ -1104,134 +1104,134 @@ extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_comieq_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_EQ_OS, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_EQ_OS, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_comilt_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_LT_OS, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_LT_OS, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_comile_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_LE_OS, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_LE_OS, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_comigt_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_GT_OS, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_GT_OS, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_comige_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_GE_OS, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_GE_OS, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_comineq_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_NEQ_US, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_NEQ_US, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_ucomieq_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_EQ_OQ, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_EQ_OQ, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_ucomilt_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_LT_OQ, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_LT_OQ, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_ucomile_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_LE_OQ, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_LE_OQ, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_ucomigt_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_GT_OQ, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_GT_OQ, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_ucomige_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_GE_OQ, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_GE_OQ, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_ucomineq_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_NEQ_UQ, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_NEQ_UQ, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > #ifdef __OPTIMIZE__ > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > - _mm_comi_sh (__m128h __A, __m128h __B, const int __P) > +_mm_comi_sh (__m128h __A, __m128h __B, const int __P) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, __P, > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, __P, > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline int > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_comi_round_sh (__m128h __A, __m128h __B, const int __P, const int __R) > { > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, __P, > - (__mmask8) -1,__R); > + return __builtin_ia32_cmpsh_mask_round (__A, __B, __P, > + (__mmask8) -1,__R); > } > > #else > -#define _mm_comi_round_sh(A, B, P, R) \ > - (__builtin_ia32_vcmpsh_v8hf_mask_round ((A), (B), (P), (__mmask8) (-1), (R))) > -#define _mm_comi_sh(A, B, P) \ > - (__builtin_ia32_vcmpsh_v8hf_mask_round ((A), (B), (P), (__mmask8) (-1), \ > - _MM_FROUND_CUR_DIRECTION)) > +#define _mm_comi_round_sh(A, B, P, R) \ > + (__builtin_ia32_cmpsh_mask_round ((A), (B), (P), (__mmask8) (-1), (R))) > +#define _mm_comi_sh(A, B, P) \ > + (__builtin_ia32_cmpsh_mask_round ((A), (B), (P), (__mmask8) (-1), \ > + _MM_FROUND_CUR_DIRECTION)) > > #endif /* __OPTIMIZE__ */ > > @@ -1240,28 +1240,28 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_sqrt_ph (__m512h __A) > { > - return __builtin_ia32_vsqrtph_v32hf_mask_round (__A, > - _mm512_setzero_ph(), > - (__mmask32) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_sqrtph512_mask_round (__A, > + _mm512_setzero_ph(), > + (__mmask32) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_sqrt_ph (__m512h __A, __mmask32 __B, __m512h __C) > { > - return __builtin_ia32_vsqrtph_v32hf_mask_round (__C, __A, __B, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_sqrtph512_mask_round (__C, __A, __B, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_sqrt_ph (__mmask32 __A, __m512h __B) > { > - return __builtin_ia32_vsqrtph_v32hf_mask_round (__B, > - _mm512_setzero_ph (), > - __A, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_sqrtph512_mask_round (__B, > + _mm512_setzero_ph (), > + __A, > + _MM_FROUND_CUR_DIRECTION); > } > > #ifdef __OPTIMIZE__ > @@ -1269,9 +1269,9 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_sqrt_round_ph (__m512h __A, const int __B) > { > - return __builtin_ia32_vsqrtph_v32hf_mask_round (__A, > - _mm512_setzero_ph(), > - (__mmask32) -1, __B); > + return __builtin_ia32_sqrtph512_mask_round (__A, > + _mm512_setzero_ph(), > + (__mmask32) -1, __B); > } > > extern __inline __m512h > @@ -1279,31 +1279,31 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_sqrt_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > const int __D) > { > - return __builtin_ia32_vsqrtph_v32hf_mask_round (__C, __A, __B, __D); > + return __builtin_ia32_sqrtph512_mask_round (__C, __A, __B, __D); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_sqrt_round_ph (__mmask32 __A, __m512h __B, const int __C) > { > - return __builtin_ia32_vsqrtph_v32hf_mask_round (__B, > - _mm512_setzero_ph (), > - __A, __C); > + return __builtin_ia32_sqrtph512_mask_round (__B, > + _mm512_setzero_ph (), > + __A, __C); > } > > #else > -#define _mm512_sqrt_round_ph(A, B) \ > - (__builtin_ia32_vsqrtph_v32hf_mask_round ((A), \ > - _mm512_setzero_ph (), \ > - (__mmask32)-1, (B))) > +#define _mm512_sqrt_round_ph(A, B) \ > + (__builtin_ia32_sqrtph512_mask_round ((A), \ > + _mm512_setzero_ph (), \ > + (__mmask32)-1, (B))) > > -#define _mm512_mask_sqrt_round_ph(A, B, C, D) \ > - (__builtin_ia32_vsqrtph_v32hf_mask_round ((C), (A), (B), (D))) > +#define _mm512_mask_sqrt_round_ph(A, B, C, D) \ > + (__builtin_ia32_sqrtph512_mask_round ((C), (A), (B), (D))) > > -#define _mm512_maskz_sqrt_round_ph(A, B, C) \ > - (__builtin_ia32_vsqrtph_v32hf_mask_round ((B), \ > - _mm512_setzero_ph (), \ > - (A), (C))) > +#define _mm512_maskz_sqrt_round_ph(A, B, C) \ > + (__builtin_ia32_sqrtph512_mask_round ((B), \ > + _mm512_setzero_ph (), \ > + (A), (C))) > > #endif /* __OPTIMIZE__ */ > > @@ -1312,23 +1312,23 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_rsqrt_ph (__m512h __A) > { > - return __builtin_ia32_vrsqrtph_v32hf_mask (__A, _mm512_setzero_ph (), > - (__mmask32) -1); > + return __builtin_ia32_rsqrtph512_mask (__A, _mm512_setzero_ph (), > + (__mmask32) -1); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_rsqrt_ph (__m512h __A, __mmask32 __B, __m512h __C) > { > - return __builtin_ia32_vrsqrtph_v32hf_mask (__C, __A, __B); > + return __builtin_ia32_rsqrtph512_mask (__C, __A, __B); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_rsqrt_ph (__mmask32 __A, __m512h __B) > { > - return __builtin_ia32_vrsqrtph_v32hf_mask (__B, _mm512_setzero_ph (), > - __A); > + return __builtin_ia32_rsqrtph512_mask (__B, _mm512_setzero_ph (), > + __A); > } > > /* Intrinsics vrsqrtsh. */ > @@ -1336,23 +1336,23 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_rsqrt_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vrsqrtsh_v8hf_mask (__B, __A, _mm_setzero_ph (), > - (__mmask8) -1); > + return __builtin_ia32_rsqrtsh_mask (__B, __A, _mm_setzero_ph (), > + (__mmask8) -1); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_rsqrt_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vrsqrtsh_v8hf_mask (__D, __C, __A, __B); > + return __builtin_ia32_rsqrtsh_mask (__D, __C, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_rsqrt_sh (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vrsqrtsh_v8hf_mask (__C, __B, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_rsqrtsh_mask (__C, __B, _mm_setzero_ph (), > + __A); > } > > /* Intrinsics vsqrtsh. */ > @@ -1360,27 +1360,27 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_sqrt_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__B, __A, > - _mm_setzero_ph (), > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_sqrtsh_mask_round (__B, __A, > + _mm_setzero_ph (), > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_sqrt_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__D, __C, __A, __B, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_sqrtsh_mask_round (__D, __C, __A, __B, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_sqrt_sh (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__C, __B, > - _mm_setzero_ph (), > - __A, _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_sqrtsh_mask_round (__C, __B, > + _mm_setzero_ph (), > + __A, _MM_FROUND_CUR_DIRECTION); > } > > #ifdef __OPTIMIZE__ > @@ -1388,9 +1388,9 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_sqrt_round_sh (__m128h __A, __m128h __B, const int __C) > { > - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__B, __A, > - _mm_setzero_ph (), > - (__mmask8) -1, __C); > + return __builtin_ia32_sqrtsh_mask_round (__B, __A, > + _mm_setzero_ph (), > + (__mmask8) -1, __C); > } > > extern __inline __m128h > @@ -1398,8 +1398,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_sqrt_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > __m128h __D, const int __E) > { > - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__D, __C, __A, __B, > - __E); > + return __builtin_ia32_sqrtsh_mask_round (__D, __C, __A, __B, > + __E); > } > > extern __inline __m128h > @@ -1407,24 +1407,24 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_sqrt_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > const int __D) > { > - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__C, __B, > - _mm_setzero_ph (), > - __A, __D); > + return __builtin_ia32_sqrtsh_mask_round (__C, __B, > + _mm_setzero_ph (), > + __A, __D); > } > > #else > #define _mm_sqrt_round_sh(A, B, C) \ > - (__builtin_ia32_vsqrtsh_v8hf_mask_round ((B), (A), \ > - _mm_setzero_ph (), \ > - (__mmask8)-1, (C))) > + (__builtin_ia32_sqrtsh_mask_round ((B), (A), \ > + _mm_setzero_ph (), \ > + (__mmask8)-1, (C))) > > #define _mm_mask_sqrt_round_sh(A, B, C, D, E) \ > - (__builtin_ia32_vsqrtsh_v8hf_mask_round ((D), (C), (A), (B), (E))) > + (__builtin_ia32_sqrtsh_mask_round ((D), (C), (A), (B), (E))) > > -#define _mm_maskz_sqrt_round_sh(A, B, C, D) \ > - (__builtin_ia32_vsqrtsh_v8hf_mask_round ((C), (B), \ > - _mm_setzero_ph (), \ > - (A), (D))) > +#define _mm_maskz_sqrt_round_sh(A, B, C, D) \ > + (__builtin_ia32_sqrtsh_mask_round ((C), (B), \ > + _mm_setzero_ph (), \ > + (A), (D))) > > #endif /* __OPTIMIZE__ */ > > @@ -1433,23 +1433,23 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_rcp_ph (__m512h __A) > { > - return __builtin_ia32_vrcpph_v32hf_mask (__A, _mm512_setzero_ph (), > - (__mmask32) -1); > + return __builtin_ia32_rcpph512_mask (__A, _mm512_setzero_ph (), > + (__mmask32) -1); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_rcp_ph (__m512h __A, __mmask32 __B, __m512h __C) > { > - return __builtin_ia32_vrcpph_v32hf_mask (__C, __A, __B); > + return __builtin_ia32_rcpph512_mask (__C, __A, __B); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_rcp_ph (__mmask32 __A, __m512h __B) > { > - return __builtin_ia32_vrcpph_v32hf_mask (__B, _mm512_setzero_ph (), > - __A); > + return __builtin_ia32_rcpph512_mask (__B, _mm512_setzero_ph (), > + __A); > } > > /* Intrinsics vrcpsh. */ > @@ -1457,23 +1457,23 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_rcp_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vrcpsh_v8hf_mask (__B, __A, _mm_setzero_ph (), > - (__mmask8) -1); > + return __builtin_ia32_rcpsh_mask (__B, __A, _mm_setzero_ph (), > + (__mmask8) -1); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_rcp_sh (__m128h __A, __mmask32 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vrcpsh_v8hf_mask (__D, __C, __A, __B); > + return __builtin_ia32_rcpsh_mask (__D, __C, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_rcp_sh (__mmask32 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vrcpsh_v8hf_mask (__C, __B, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_rcpsh_mask (__C, __B, _mm_setzero_ph (), > + __A); > } > > /* Intrinsics vscalefph. */ > @@ -1481,28 +1481,28 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_scalef_ph (__m512h __A, __m512h __B) > { > - return __builtin_ia32_vscalefph_v32hf_mask_round (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_scalefph512_mask_round (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_scalef_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > { > - return __builtin_ia32_vscalefph_v32hf_mask_round (__C, __D, __A, __B, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_scalefph512_mask_round (__C, __D, __A, __B, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_scalef_ph (__mmask32 __A, __m512h __B, __m512h __C) > { > - return __builtin_ia32_vscalefph_v32hf_mask_round (__B, __C, > - _mm512_setzero_ph (), > - __A, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_scalefph512_mask_round (__B, __C, > + _mm512_setzero_ph (), > + __A, > + _MM_FROUND_CUR_DIRECTION); > } > > #ifdef __OPTIMIZE__ > @@ -1510,9 +1510,9 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_scalef_round_ph (__m512h __A, __m512h __B, const int __C) > { > - return __builtin_ia32_vscalefph_v32hf_mask_round (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1, __C); > + return __builtin_ia32_scalefph512_mask_round (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1, __C); > } > > extern __inline __m512h > @@ -1520,8 +1520,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_scalef_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > __m512h __D, const int __E) > { > - return __builtin_ia32_vscalefph_v32hf_mask_round (__C, __D, __A, __B, > - __E); > + return __builtin_ia32_scalefph512_mask_round (__C, __D, __A, __B, > + __E); > } > > extern __inline __m512h > @@ -1529,24 +1529,24 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_scalef_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > const int __D) > { > - return __builtin_ia32_vscalefph_v32hf_mask_round (__B, __C, > - _mm512_setzero_ph (), > - __A, __D); > + return __builtin_ia32_scalefph512_mask_round (__B, __C, > + _mm512_setzero_ph (), > + __A, __D); > } > > #else > -#define _mm512_scalef_round_ph(A, B, C) \ > - (__builtin_ia32_vscalefph_v32hf_mask_round ((A), (B), \ > - _mm512_setzero_ph (), \ > - (__mmask32)-1, (C))) > +#define _mm512_scalef_round_ph(A, B, C) \ > + (__builtin_ia32_scalefph512_mask_round ((A), (B), \ > + _mm512_setzero_ph (), \ > + (__mmask32)-1, (C))) > > #define _mm512_mask_scalef_round_ph(A, B, C, D, E) \ > - (__builtin_ia32_vscalefph_v32hf_mask_round ((C), (D), (A), (B), (E))) > + (__builtin_ia32_scalefph512_mask_round ((C), (D), (A), (B), (E))) > > -#define _mm512_maskz_scalef_round_ph(A, B, C, D) \ > - (__builtin_ia32_vscalefph_v32hf_mask_round ((B), (C), \ > - _mm512_setzero_ph (), \ > - (A), (D))) > +#define _mm512_maskz_scalef_round_ph(A, B, C, D) \ > + (__builtin_ia32_scalefph512_mask_round ((B), (C), \ > + _mm512_setzero_ph (), \ > + (A), (D))) > > #endif /* __OPTIMIZE__ */ > > @@ -1555,28 +1555,28 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_scalef_sh (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vscalefsh_v8hf_mask_round (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_scalefsh_mask_round (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_scalef_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vscalefsh_v8hf_mask_round (__C, __D, __A, __B, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_scalefsh_mask_round (__C, __D, __A, __B, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_scalef_sh (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vscalefsh_v8hf_mask_round (__B, __C, > - _mm_setzero_ph (), > - __A, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_scalefsh_mask_round (__B, __C, > + _mm_setzero_ph (), > + __A, > + _MM_FROUND_CUR_DIRECTION); > } > > #ifdef __OPTIMIZE__ > @@ -1584,9 +1584,9 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_scalef_round_sh (__m128h __A, __m128h __B, const int __C) > { > - return __builtin_ia32_vscalefsh_v8hf_mask_round (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1, __C); > + return __builtin_ia32_scalefsh_mask_round (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1, __C); > } > > extern __inline __m128h > @@ -1594,8 +1594,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_scalef_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > __m128h __D, const int __E) > { > - return __builtin_ia32_vscalefsh_v8hf_mask_round (__C, __D, __A, __B, > - __E); > + return __builtin_ia32_scalefsh_mask_round (__C, __D, __A, __B, > + __E); > } > > extern __inline __m128h > @@ -1603,23 +1603,23 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_scalef_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > const int __D) > { > - return __builtin_ia32_vscalefsh_v8hf_mask_round (__B, __C, > - _mm_setzero_ph (), > - __A, __D); > + return __builtin_ia32_scalefsh_mask_round (__B, __C, > + _mm_setzero_ph (), > + __A, __D); > } > > #else > -#define _mm_scalef_round_sh(A, B, C) \ > - (__builtin_ia32_vscalefsh_v8hf_mask_round ((A), (B), \ > - _mm_setzero_ph (), \ > - (__mmask8)-1, (C))) > +#define _mm_scalef_round_sh(A, B, C) \ > + (__builtin_ia32_scalefsh_mask_round ((A), (B), \ > + _mm_setzero_ph (), \ > + (__mmask8)-1, (C))) > > -#define _mm_mask_scalef_round_sh(A, B, C, D, E) \ > - (__builtin_ia32_vscalefsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > +#define _mm_mask_scalef_round_sh(A, B, C, D, E) \ > + (__builtin_ia32_scalefsh_mask_round ((C), (D), (A), (B), (E))) > > -#define _mm_maskz_scalef_round_sh(A, B, C, D) \ > - (__builtin_ia32_vscalefsh_v8hf_mask_round ((B), (C), _mm_setzero_ph (), \ > - (A), (D))) > +#define _mm_maskz_scalef_round_sh(A, B, C, D) \ > + (__builtin_ia32_scalefsh_mask_round ((B), (C), _mm_setzero_ph (), \ > + (A), (D))) > > #endif /* __OPTIMIZE__ */ > > @@ -1629,37 +1629,37 @@ extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_reduce_ph (__m512h __A, int __B) > { > - return __builtin_ia32_vreduceph_v32hf_mask_round (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_reduceph512_mask_round (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_reduce_ph (__m512h __A, __mmask32 __B, __m512h __C, int __D) > { > - return __builtin_ia32_vreduceph_v32hf_mask_round (__C, __D, __A, __B, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_reduceph512_mask_round (__C, __D, __A, __B, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_reduce_ph (__mmask32 __A, __m512h __B, int __C) > { > - return __builtin_ia32_vreduceph_v32hf_mask_round (__B, __C, > - _mm512_setzero_ph (), > - __A, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_reduceph512_mask_round (__B, __C, > + _mm512_setzero_ph (), > + __A, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_reduce_round_ph (__m512h __A, int __B, const int __C) > { > - return __builtin_ia32_vreduceph_v32hf_mask_round (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1, __C); > + return __builtin_ia32_reduceph512_mask_round (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1, __C); > } > > extern __inline __m512h > @@ -1667,8 +1667,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_reduce_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > int __D, const int __E) > { > - return __builtin_ia32_vreduceph_v32hf_mask_round (__C, __D, __A, __B, > - __E); > + return __builtin_ia32_reduceph512_mask_round (__C, __D, __A, __B, > + __E); > } > > extern __inline __m512h > @@ -1676,39 +1676,39 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_reduce_round_ph (__mmask32 __A, __m512h __B, int __C, > const int __D) > { > - return __builtin_ia32_vreduceph_v32hf_mask_round (__B, __C, > - _mm512_setzero_ph (), > - __A, __D); > + return __builtin_ia32_reduceph512_mask_round (__B, __C, > + _mm512_setzero_ph (), > + __A, __D); > } > > #else > #define _mm512_reduce_ph(A, B) \ > - (__builtin_ia32_vreduceph_v32hf_mask_round ((A), (B), \ > - _mm512_setzero_ph (), \ > - (__mmask32)-1, \ > - _MM_FROUND_CUR_DIRECTION)) > + (__builtin_ia32_reduceph512_mask_round ((A), (B), \ > + _mm512_setzero_ph (), \ > + (__mmask32)-1, \ > + _MM_FROUND_CUR_DIRECTION)) > > #define _mm512_mask_reduce_ph(A, B, C, D) \ > - (__builtin_ia32_vreduceph_v32hf_mask_round ((C), (D), (A), (B), \ > - _MM_FROUND_CUR_DIRECTION)) > + (__builtin_ia32_reduceph512_mask_round ((C), (D), (A), (B), \ > + _MM_FROUND_CUR_DIRECTION)) > > #define _mm512_maskz_reduce_ph(A, B, C) \ > - (__builtin_ia32_vreduceph_v32hf_mask_round ((B), (C), \ > - _mm512_setzero_ph (), \ > - (A), _MM_FROUND_CUR_DIRECTION)) > + (__builtin_ia32_reduceph512_mask_round ((B), (C), \ > + _mm512_setzero_ph (), \ > + (A), _MM_FROUND_CUR_DIRECTION)) > > -#define _mm512_reduce_round_ph(A, B, C) \ > - (__builtin_ia32_vreduceph_v32hf_mask_round ((A), (B), \ > - _mm512_setzero_ph (), \ > - (__mmask32)-1, (C))) > +#define _mm512_reduce_round_ph(A, B, C) \ > + (__builtin_ia32_reduceph512_mask_round ((A), (B), \ > + _mm512_setzero_ph (), \ > + (__mmask32)-1, (C))) > > #define _mm512_mask_reduce_round_ph(A, B, C, D, E) \ > - (__builtin_ia32_vreduceph_v32hf_mask_round ((C), (D), (A), (B), (E))) > + (__builtin_ia32_reduceph512_mask_round ((C), (D), (A), (B), (E))) > > -#define _mm512_maskz_reduce_round_ph(A, B, C, D) \ > - (__builtin_ia32_vreduceph_v32hf_mask_round ((B), (C), \ > - _mm512_setzero_ph (), \ > - (A), (D))) > +#define _mm512_maskz_reduce_round_ph(A, B, C, D) \ > + (__builtin_ia32_reduceph512_mask_round ((B), (C), \ > + _mm512_setzero_ph (), \ > + (A), (D))) > > #endif /* __OPTIMIZE__ */ > > @@ -1718,10 +1718,10 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_reduce_sh (__m128h __A, __m128h __B, int __C) > { > - return __builtin_ia32_vreducesh_v8hf_mask_round (__A, __B, __C, > - _mm_setzero_ph (), > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_reducesh_mask_round (__A, __B, __C, > + _mm_setzero_ph (), > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m128h > @@ -1729,26 +1729,26 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_reduce_sh (__m128h __A, __mmask8 __B, __m128h __C, > __m128h __D, int __E) > { > - return __builtin_ia32_vreducesh_v8hf_mask_round (__C, __D, __E, __A, __B, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_reducesh_mask_round (__C, __D, __E, __A, __B, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_reduce_sh (__mmask8 __A, __m128h __B, __m128h __C, int __D) > { > - return __builtin_ia32_vreducesh_v8hf_mask_round (__B, __C, __D, > - _mm_setzero_ph (), __A, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_reducesh_mask_round (__B, __C, __D, > + _mm_setzero_ph (), __A, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_reduce_round_sh (__m128h __A, __m128h __B, int __C, const int __D) > { > - return __builtin_ia32_vreducesh_v8hf_mask_round (__A, __B, __C, > - _mm_setzero_ph (), > - (__mmask8) -1, __D); > + return __builtin_ia32_reducesh_mask_round (__A, __B, __C, > + _mm_setzero_ph (), > + (__mmask8) -1, __D); > } > > extern __inline __m128h > @@ -1756,8 +1756,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_reduce_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > __m128h __D, int __E, const int __F) > { > - return __builtin_ia32_vreducesh_v8hf_mask_round (__C, __D, __E, __A, > - __B, __F); > + return __builtin_ia32_reducesh_mask_round (__C, __D, __E, __A, > + __B, __F); > } > > extern __inline __m128h > @@ -1765,81 +1765,81 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_reduce_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > int __D, const int __E) > { > - return __builtin_ia32_vreducesh_v8hf_mask_round (__B, __C, __D, > - _mm_setzero_ph (), > - __A, __E); > + return __builtin_ia32_reducesh_mask_round (__B, __C, __D, > + _mm_setzero_ph (), > + __A, __E); > } > > #else > #define _mm_reduce_sh(A, B, C) \ > - (__builtin_ia32_vreducesh_v8hf_mask_round ((A), (B), (C), \ > - _mm_setzero_ph (), \ > - (__mmask8)-1, \ > - _MM_FROUND_CUR_DIRECTION)) > + (__builtin_ia32_reducesh_mask_round ((A), (B), (C), \ > + _mm_setzero_ph (), \ > + (__mmask8)-1, \ > + _MM_FROUND_CUR_DIRECTION)) > > #define _mm_mask_reduce_sh(A, B, C, D, E) \ > - (__builtin_ia32_vreducesh_v8hf_mask_round ((C), (D), (E), (A), (B), \ > - _MM_FROUND_CUR_DIRECTION)) > + (__builtin_ia32_reducesh_mask_round ((C), (D), (E), (A), (B), \ > + _MM_FROUND_CUR_DIRECTION)) > > #define _mm_maskz_reduce_sh(A, B, C, D) \ > - (__builtin_ia32_vreducesh_v8hf_mask_round ((B), (C), (D), \ > - _mm_setzero_ph (), \ > - (A), _MM_FROUND_CUR_DIRECTION)) > + (__builtin_ia32_reducesh_mask_round ((B), (C), (D), \ > + _mm_setzero_ph (), \ > + (A), _MM_FROUND_CUR_DIRECTION)) > > #define _mm_reduce_round_sh(A, B, C, D) \ > - (__builtin_ia32_vreducesh_v8hf_mask_round ((A), (B), (C), \ > - _mm_setzero_ph (), \ > - (__mmask8)-1, (D))) > + (__builtin_ia32_reducesh_mask_round ((A), (B), (C), \ > + _mm_setzero_ph (), \ > + (__mmask8)-1, (D))) > > #define _mm_mask_reduce_round_sh(A, B, C, D, E, F) \ > - (__builtin_ia32_vreducesh_v8hf_mask_round ((C), (D), (E), (A), (B), (F))) > + (__builtin_ia32_reducesh_mask_round ((C), (D), (E), (A), (B), (F))) > > #define _mm_maskz_reduce_round_sh(A, B, C, D, E) \ > - (__builtin_ia32_vreducesh_v8hf_mask_round ((B), (C), (D), \ > - _mm_setzero_ph (), \ > - (A), (E))) > + (__builtin_ia32_reducesh_mask_round ((B), (C), (D), \ > + _mm_setzero_ph (), \ > + (A), (E))) > > #endif /* __OPTIMIZE__ */ > > /* Intrinsics vrndscaleph. */ > #ifdef __OPTIMIZE__ > extern __inline __m512h > -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_roundscale_ph (__m512h __A, int __B) > { > - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_rndscaleph512_mask_round (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_roundscale_ph (__m512h __A, __mmask32 __B, > - __m512h __C, int __D) > + __m512h __C, int __D) > { > - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__C, __D, __A, __B, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_rndscaleph512_mask_round (__C, __D, __A, __B, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_roundscale_ph (__mmask32 __A, __m512h __B, int __C) > { > - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__B, __C, > - _mm512_setzero_ph (), > - __A, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_rndscaleph512_mask_round (__B, __C, > + _mm512_setzero_ph (), > + __A, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m512h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_roundscale_round_ph (__m512h __A, int __B, const int __C) > { > - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__A, __B, > - _mm512_setzero_ph (), > - (__mmask32) -1, > - __C); > + return __builtin_ia32_rndscaleph512_mask_round (__A, __B, > + _mm512_setzero_ph (), > + (__mmask32) -1, > + __C); > } > > extern __inline __m512h > @@ -1847,8 +1847,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_mask_roundscale_round_ph (__m512h __A, __mmask32 __B, > __m512h __C, int __D, const int __E) > { > - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__C, __D, __A, > - __B, __E); > + return __builtin_ia32_rndscaleph512_mask_round (__C, __D, __A, > + __B, __E); > } > > extern __inline __m512h > @@ -1856,52 +1856,52 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_roundscale_round_ph (__mmask32 __A, __m512h __B, int __C, > const int __D) > { > - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__B, __C, > - _mm512_setzero_ph (), > - __A, __D); > + return __builtin_ia32_rndscaleph512_mask_round (__B, __C, > + _mm512_setzero_ph (), > + __A, __D); > } > > #else > -#define _mm512_roundscale_ph(A, B) \ > - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((A), (B), \ > - _mm512_setzero_ph (), \ > - (__mmask32)-1, \ > - _MM_FROUND_CUR_DIRECTION)) > - > -#define _mm512_mask_roundscale_ph(A, B, C, D) \ > - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((C), (D), (A), (B), \ > - _MM_FROUND_CUR_DIRECTION)) > - > -#define _mm512_maskz_roundscale_ph(A, B, C) \ > - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((B), (C), \ > - _mm512_setzero_ph (), \ > - (A), \ > - _MM_FROUND_CUR_DIRECTION)) > -#define _mm512_roundscale_round_ph(A, B, C) \ > - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((A), (B), \ > - _mm512_setzero_ph (), \ > - (__mmask32)-1, (C))) > +#define _mm512_roundscale_ph(A, B) \ > + (__builtin_ia32_rndscaleph512_mask_round ((A), (B), \ > + _mm512_setzero_ph (), \ > + (__mmask32)-1, \ > + _MM_FROUND_CUR_DIRECTION)) > + > +#define _mm512_mask_roundscale_ph(A, B, C, D) \ > + (__builtin_ia32_rndscaleph512_mask_round ((C), (D), (A), (B), \ > + _MM_FROUND_CUR_DIRECTION)) > + > +#define _mm512_maskz_roundscale_ph(A, B, C) \ > + (__builtin_ia32_rndscaleph512_mask_round ((B), (C), \ > + _mm512_setzero_ph (), \ > + (A), \ > + _MM_FROUND_CUR_DIRECTION)) > +#define _mm512_roundscale_round_ph(A, B, C) \ > + (__builtin_ia32_rndscaleph512_mask_round ((A), (B), \ > + _mm512_setzero_ph (), \ > + (__mmask32)-1, (C))) > > #define _mm512_mask_roundscale_round_ph(A, B, C, D, E) \ > - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((C), (D), (A), (B), (E))) > + (__builtin_ia32_rndscaleph512_mask_round ((C), (D), (A), (B), (E))) > > -#define _mm512_maskz_roundscale_round_ph(A, B, C, D) \ > - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((B), (C), \ > - _mm512_setzero_ph (), \ > - (A), (D))) > +#define _mm512_maskz_roundscale_round_ph(A, B, C, D) \ > + (__builtin_ia32_rndscaleph512_mask_round ((B), (C), \ > + _mm512_setzero_ph (), \ > + (A), (D))) > > #endif /* __OPTIMIZE__ */ > > /* Intrinsics vrndscalesh. */ > #ifdef __OPTIMIZE__ > extern __inline __m128h > -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_roundscale_sh (__m128h __A, __m128h __B, int __C) > { > - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__A, __B, __C, > - _mm_setzero_ph (), > - (__mmask8) -1, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_rndscalesh_mask_round (__A, __B, __C, > + _mm_setzero_ph (), > + (__mmask8) -1, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m128h > @@ -1909,27 +1909,27 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_roundscale_sh (__m128h __A, __mmask8 __B, __m128h __C, > __m128h __D, int __E) > { > - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__C, __D, __E, __A, __B, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_rndscalesh_mask_round (__C, __D, __E, __A, __B, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_roundscale_sh (__mmask8 __A, __m128h __B, __m128h __C, int __D) > { > - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__B, __C, __D, > - _mm_setzero_ph (), __A, > - _MM_FROUND_CUR_DIRECTION); > + return __builtin_ia32_rndscalesh_mask_round (__B, __C, __D, > + _mm_setzero_ph (), __A, > + _MM_FROUND_CUR_DIRECTION); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_roundscale_round_sh (__m128h __A, __m128h __B, int __C, const int __D) > { > - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__A, __B, __C, > - _mm_setzero_ph (), > - (__mmask8) -1, > - __D); > + return __builtin_ia32_rndscalesh_mask_round (__A, __B, __C, > + _mm_setzero_ph (), > + (__mmask8) -1, > + __D); > } > > extern __inline __m128h > @@ -1937,8 +1937,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_roundscale_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > __m128h __D, int __E, const int __F) > { > - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__C, __D, __E, > - __A, __B, __F); > + return __builtin_ia32_rndscalesh_mask_round (__C, __D, __E, > + __A, __B, __F); > } > > extern __inline __m128h > @@ -1946,46 +1946,46 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_roundscale_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > int __D, const int __E) > { > - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__B, __C, __D, > - _mm_setzero_ph (), > - __A, __E); > + return __builtin_ia32_rndscalesh_mask_round (__B, __C, __D, > + _mm_setzero_ph (), > + __A, __E); > } > > #else > #define _mm_roundscale_sh(A, B, C) \ > - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((A), (B), (C), \ > - _mm_setzero_ph (), \ > - (__mmask8)-1, \ > - _MM_FROUND_CUR_DIRECTION)) > + (__builtin_ia32_rndscalesh_mask_round ((A), (B), (C), \ > + _mm_setzero_ph (), \ > + (__mmask8)-1, \ > + _MM_FROUND_CUR_DIRECTION)) > > #define _mm_mask_roundscale_sh(A, B, C, D, E) \ > - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((C), (D), (E), (A), (B), \ > - _MM_FROUND_CUR_DIRECTION)) > + (__builtin_ia32_rndscalesh_mask_round ((C), (D), (E), (A), (B), \ > + _MM_FROUND_CUR_DIRECTION)) > > #define _mm_maskz_roundscale_sh(A, B, C, D) \ > - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((B), (C), (D), \ > - _mm_setzero_ph (), \ > - (A), _MM_FROUND_CUR_DIRECTION)) > + (__builtin_ia32_rndscalesh_mask_round ((B), (C), (D), \ > + _mm_setzero_ph (), \ > + (A), _MM_FROUND_CUR_DIRECTION)) > > -#define _mm_roundscale_round_sh(A, B, C, D) \ > - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((A), (B), (C), \ > - _mm_setzero_ph (), \ > - (__mmask8)-1, (D))) > +#define _mm_roundscale_round_sh(A, B, C, D) \ > + (__builtin_ia32_rndscalesh_mask_round ((A), (B), (C), \ > + _mm_setzero_ph (), \ > + (__mmask8)-1, (D))) > > #define _mm_mask_roundscale_round_sh(A, B, C, D, E, F) \ > - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((C), (D), (E), (A), (B), (F))) > + (__builtin_ia32_rndscalesh_mask_round ((C), (D), (E), (A), (B), (F))) > > -#define _mm_maskz_roundscale_round_sh(A, B, C, D, E) \ > - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((B), (C), (D), \ > - _mm_setzero_ph (), \ > - (A), (E))) > +#define _mm_maskz_roundscale_round_sh(A, B, C, D, E) \ > + (__builtin_ia32_rndscalesh_mask_round ((B), (C), (D), \ > + _mm_setzero_ph (), \ > + (A), (E))) > > #endif /* __OPTIMIZE__ */ > > /* Intrinsics vfpclasssh. */ > #ifdef __OPTIMIZE__ > extern __inline __mmask8 > -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_fpclass_sh_mask (__m128h __A, const int __imm) > { > return (__mmask8) __builtin_ia32_fpclasssh_mask ((__v8hf) __A, __imm, > @@ -2031,11 +2031,11 @@ _mm512_fpclass_ph_mask (__m512h __A, const int __imm) > > #else > #define _mm512_mask_fpclass_ph_mask(u, x, c) \ > - ((__mmask32) __builtin_ia32_fpclassph512_mask ((__v32hf) (__m512h) (x),\ > + ((__mmask32) __builtin_ia32_fpclassph512_mask ((__v32hf) (__m512h) (x), \ > (int) (c),(__mmask8)(u))) > > #define _mm512_fpclass_ph_mask(x, c) \ > - ((__mmask32) __builtin_ia32_fpclassph512_mask ((__v32hf) (__m512h) (x),\ > + ((__mmask32) __builtin_ia32_fpclassph512_mask ((__v32hf) (__m512h) (x), \ > (int) (c),(__mmask8)-1)) > #endif /* __OPIMTIZE__ */ > > @@ -2141,9 +2141,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_getexp_round_ph (__m512h __A, const int __R) > { > return (__m512h) __builtin_ia32_getexpph512_mask ((__v32hf) __A, > - (__v32hf) > - _mm512_setzero_ph (), > - (__mmask32) -1, __R); > + (__v32hf) > + _mm512_setzero_ph (), > + (__mmask32) -1, __R); > } > > extern __inline __m512h > @@ -2152,8 +2152,8 @@ _mm512_mask_getexp_round_ph (__m512h __W, __mmask32 __U, __m512h __A, > const int __R) > { > return (__m512h) __builtin_ia32_getexpph512_mask ((__v32hf) __A, > - (__v32hf) __W, > - (__mmask32) __U, __R); > + (__v32hf) __W, > + (__mmask32) __U, __R); > } > > extern __inline __m512h > @@ -2161,37 +2161,37 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm512_maskz_getexp_round_ph (__mmask32 __U, __m512h __A, const int __R) > { > return (__m512h) __builtin_ia32_getexpph512_mask ((__v32hf) __A, > - (__v32hf) > - _mm512_setzero_ph (), > - (__mmask32) __U, __R); > + (__v32hf) > + _mm512_setzero_ph (), > + (__mmask32) __U, __R); > } > > #else > -#define _mm_getexp_round_sh(A, B, R) \ > - ((__m128h)__builtin_ia32_getexpsh_mask_round((__v8hf)(__m128h)(A), \ > - (__v8hf)(__m128h)(B), \ > - (__v8hf)_mm_setzero_ph(), \ > +#define _mm_getexp_round_sh(A, B, R) \ > + ((__m128h)__builtin_ia32_getexpsh_mask_round((__v8hf)(__m128h)(A), \ > + (__v8hf)(__m128h)(B), \ > + (__v8hf)_mm_setzero_ph(), \ > (__mmask8)-1, R)) > > -#define _mm_mask_getexp_round_sh(W, U, A, B, C) \ > +#define _mm_mask_getexp_round_sh(W, U, A, B, C) \ > (__m128h)__builtin_ia32_getexpsh_mask_round(A, B, W, U, C) > > -#define _mm_maskz_getexp_round_sh(U, A, B, C) \ > - (__m128h)__builtin_ia32_getexpsh_mask_round(A, B, \ > - (__v8hf)_mm_setzero_ph(), \ > +#define _mm_maskz_getexp_round_sh(U, A, B, C) \ > + (__m128h)__builtin_ia32_getexpsh_mask_round(A, B, \ > + (__v8hf)_mm_setzero_ph(), \ > U, C) > > -#define _mm512_getexp_round_ph(A, R) \ > - ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ > - (__v32hf)_mm512_setzero_ph(), (__mmask32)-1, R)) > +#define _mm512_getexp_round_ph(A, R) \ > + ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ > + (__v32hf)_mm512_setzero_ph(), (__mmask32)-1, R)) > > -#define _mm512_mask_getexp_round_ph(W, U, A, R) \ > - ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ > - (__v32hf)(__m512h)(W), (__mmask32)(U), R)) > +#define _mm512_mask_getexp_round_ph(W, U, A, R) \ > + ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ > + (__v32hf)(__m512h)(W), (__mmask32)(U), R)) > > -#define _mm512_maskz_getexp_round_ph(U, A, R) \ > - ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ > - (__v32hf)_mm512_setzero_ph(), (__mmask32)(U), R)) > +#define _mm512_maskz_getexp_round_ph(U, A, R) \ > + ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ > + (__v32hf)_mm512_setzero_ph(), (__mmask32)(U), R)) > > #endif /* __OPTIMIZE__ */ > > diff --git a/gcc/config/i386/avx512fp16vlintrin.h b/gcc/config/i386/avx512fp16vlintrin.h > index 206d60407fc..e9478792a03 100644 > --- a/gcc/config/i386/avx512fp16vlintrin.h > +++ b/gcc/config/i386/avx512fp16vlintrin.h > @@ -53,30 +53,30 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_add_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vaddph_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_addph128_mask (__C, __D, __A, __B); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_add_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) > { > - return __builtin_ia32_vaddph_v16hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_addph256_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_add_ph (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vaddph_v8hf_mask (__B, __C, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_addph128_mask (__B, __C, _mm_setzero_ph (), > + __A); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_add_ph (__mmask16 __A, __m256h __B, __m256h __C) > { > - return __builtin_ia32_vaddph_v16hf_mask (__B, __C, > - _mm256_setzero_ph (), __A); > + return __builtin_ia32_addph256_mask (__B, __C, > + _mm256_setzero_ph (), __A); > } > > extern __inline __m128h > @@ -97,30 +97,30 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_sub_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vsubph_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_subph128_mask (__C, __D, __A, __B); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_sub_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) > { > - return __builtin_ia32_vsubph_v16hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_subph256_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_sub_ph (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vsubph_v8hf_mask (__B, __C, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_subph128_mask (__B, __C, _mm_setzero_ph (), > + __A); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_sub_ph (__mmask16 __A, __m256h __B, __m256h __C) > { > - return __builtin_ia32_vsubph_v16hf_mask (__B, __C, > - _mm256_setzero_ph (), __A); > + return __builtin_ia32_subph256_mask (__B, __C, > + _mm256_setzero_ph (), __A); > } > > extern __inline __m128h > @@ -141,30 +141,30 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_mul_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vmulph_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_mulph128_mask (__C, __D, __A, __B); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_mul_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) > { > - return __builtin_ia32_vmulph_v16hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_mulph256_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_mul_ph (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vmulph_v8hf_mask (__B, __C, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_mulph128_mask (__B, __C, _mm_setzero_ph (), > + __A); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_mul_ph (__mmask16 __A, __m256h __B, __m256h __C) > { > - return __builtin_ia32_vmulph_v16hf_mask (__B, __C, > - _mm256_setzero_ph (), __A); > + return __builtin_ia32_mulph256_mask (__B, __C, > + _mm256_setzero_ph (), __A); > } > > extern __inline __m128h > @@ -185,30 +185,30 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_div_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vdivph_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_divph128_mask (__C, __D, __A, __B); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_div_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) > { > - return __builtin_ia32_vdivph_v16hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_divph256_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_div_ph (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vdivph_v8hf_mask (__B, __C, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_divph128_mask (__B, __C, _mm_setzero_ph (), > + __A); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_div_ph (__mmask16 __A, __m256h __B, __m256h __C) > { > - return __builtin_ia32_vdivph_v16hf_mask (__B, __C, > - _mm256_setzero_ph (), __A); > + return __builtin_ia32_divph256_mask (__B, __C, > + _mm256_setzero_ph (), __A); > } > > /* Intrinsics v[max,min]ph. */ > @@ -216,96 +216,96 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_max_ph (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vmaxph_v8hf_mask (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1); > + return __builtin_ia32_maxph128_mask (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_max_ph (__m256h __A, __m256h __B) > { > - return __builtin_ia32_vmaxph_v16hf_mask (__A, __B, > - _mm256_setzero_ph (), > - (__mmask16) -1); > + return __builtin_ia32_maxph256_mask (__A, __B, > + _mm256_setzero_ph (), > + (__mmask16) -1); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_max_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vmaxph_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_maxph128_mask (__C, __D, __A, __B); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_max_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) > { > - return __builtin_ia32_vmaxph_v16hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_maxph256_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_max_ph (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vmaxph_v8hf_mask (__B, __C, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_maxph128_mask (__B, __C, _mm_setzero_ph (), > + __A); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_max_ph (__mmask16 __A, __m256h __B, __m256h __C) > { > - return __builtin_ia32_vmaxph_v16hf_mask (__B, __C, > - _mm256_setzero_ph (), __A); > + return __builtin_ia32_maxph256_mask (__B, __C, > + _mm256_setzero_ph (), __A); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_min_ph (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vminph_v8hf_mask (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1); > + return __builtin_ia32_minph128_mask (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_min_ph (__m256h __A, __m256h __B) > { > - return __builtin_ia32_vminph_v16hf_mask (__A, __B, > - _mm256_setzero_ph (), > - (__mmask16) -1); > + return __builtin_ia32_minph256_mask (__A, __B, > + _mm256_setzero_ph (), > + (__mmask16) -1); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_min_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vminph_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_minph128_mask (__C, __D, __A, __B); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_min_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) > { > - return __builtin_ia32_vminph_v16hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_minph256_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_min_ph (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vminph_v8hf_mask (__B, __C, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_minph128_mask (__B, __C, _mm_setzero_ph (), > + __A); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_min_ph (__mmask16 __A, __m256h __B, __m256h __C) > { > - return __builtin_ia32_vminph_v16hf_mask (__B, __C, > - _mm256_setzero_ph (), __A); > + return __builtin_ia32_minph256_mask (__B, __C, > + _mm256_setzero_ph (), __A); > } > > /* vcmpph */ > @@ -314,8 +314,8 @@ extern __inline __mmask8 > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_cmp_ph_mask (__m128h __A, __m128h __B, const int __C) > { > - return (__mmask8) __builtin_ia32_vcmpph_v8hf_mask (__A, __B, __C, > - (__mmask8) -1); > + return (__mmask8) __builtin_ia32_cmpph128_mask (__A, __B, __C, > + (__mmask8) -1); > } > > extern __inline __mmask8 > @@ -323,15 +323,15 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_cmp_ph_mask (__mmask8 __A, __m128h __B, __m128h __C, > const int __D) > { > - return (__mmask8) __builtin_ia32_vcmpph_v8hf_mask (__B, __C, __D, __A); > + return (__mmask8) __builtin_ia32_cmpph128_mask (__B, __C, __D, __A); > } > > extern __inline __mmask16 > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_cmp_ph_mask (__m256h __A, __m256h __B, const int __C) > { > - return (__mmask16) __builtin_ia32_vcmpph_v16hf_mask (__A, __B, __C, > - (__mmask16) -1); > + return (__mmask16) __builtin_ia32_cmpph256_mask (__A, __B, __C, > + (__mmask16) -1); > } > > extern __inline __mmask16 > @@ -339,22 +339,22 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_cmp_ph_mask (__mmask16 __A, __m256h __B, __m256h __C, > const int __D) > { > - return (__mmask16) __builtin_ia32_vcmpph_v16hf_mask (__B, __C, __D, > - __A); > + return (__mmask16) __builtin_ia32_cmpph256_mask (__B, __C, __D, > + __A); > } > > #else > -#define _mm_cmp_ph_mask(A, B, C) \ > - (__builtin_ia32_vcmpph_v8hf_mask ((A), (B), (C), (-1))) > +#define _mm_cmp_ph_mask(A, B, C) \ > + (__builtin_ia32_cmpph128_mask ((A), (B), (C), (-1))) > > -#define _mm_mask_cmp_ph_mask(A, B, C, D) \ > - (__builtin_ia32_vcmpph_v8hf_mask ((B), (C), (D), (A))) > +#define _mm_mask_cmp_ph_mask(A, B, C, D) \ > + (__builtin_ia32_cmpph128_mask ((B), (C), (D), (A))) > > -#define _mm256_cmp_ph_mask(A, B, C) \ > - (__builtin_ia32_vcmpph_v16hf_mask ((A), (B), (C), (-1))) > +#define _mm256_cmp_ph_mask(A, B, C) \ > + (__builtin_ia32_cmpph256_mask ((A), (B), (C), (-1))) > > -#define _mm256_mask_cmp_ph_mask(A, B, C, D) \ > - (__builtin_ia32_vcmpph_v16hf_mask ((B), (C), (D), (A))) > +#define _mm256_mask_cmp_ph_mask(A, B, C, D) \ > + (__builtin_ia32_cmpph256_mask ((B), (C), (D), (A))) > > #endif /* __OPTIMIZE__ */ > > @@ -363,46 +363,46 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_sqrt_ph (__m128h __A) > { > - return __builtin_ia32_vsqrtph_v8hf_mask (__A, _mm_setzero_ph (), > - (__mmask8) -1); > + return __builtin_ia32_sqrtph128_mask (__A, _mm_setzero_ph (), > + (__mmask8) -1); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_sqrt_ph (__m256h __A) > { > - return __builtin_ia32_vsqrtph_v16hf_mask (__A, _mm256_setzero_ph (), > - (__mmask16) -1); > + return __builtin_ia32_sqrtph256_mask (__A, _mm256_setzero_ph (), > + (__mmask16) -1); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_sqrt_ph (__m128h __A, __mmask8 __B, __m128h __C) > { > - return __builtin_ia32_vsqrtph_v8hf_mask (__C, __A, __B); > + return __builtin_ia32_sqrtph128_mask (__C, __A, __B); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_sqrt_ph (__m256h __A, __mmask16 __B, __m256h __C) > { > - return __builtin_ia32_vsqrtph_v16hf_mask (__C, __A, __B); > + return __builtin_ia32_sqrtph256_mask (__C, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_sqrt_ph (__mmask8 __A, __m128h __B) > { > - return __builtin_ia32_vsqrtph_v8hf_mask (__B, _mm_setzero_ph (), > - __A); > + return __builtin_ia32_sqrtph128_mask (__B, _mm_setzero_ph (), > + __A); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_sqrt_ph (__mmask16 __A, __m256h __B) > { > - return __builtin_ia32_vsqrtph_v16hf_mask (__B, _mm256_setzero_ph (), > - __A); > + return __builtin_ia32_sqrtph256_mask (__B, _mm256_setzero_ph (), > + __A); > } > > /* Intrinsics vrsqrtph. */ > @@ -410,45 +410,45 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_rsqrt_ph (__m128h __A) > { > - return __builtin_ia32_vrsqrtph_v8hf_mask (__A, _mm_setzero_ph (), > - (__mmask8) -1); > + return __builtin_ia32_rsqrtph128_mask (__A, _mm_setzero_ph (), > + (__mmask8) -1); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_rsqrt_ph (__m256h __A) > { > - return __builtin_ia32_vrsqrtph_v16hf_mask (__A, _mm256_setzero_ph (), > - (__mmask16) -1); > + return __builtin_ia32_rsqrtph256_mask (__A, _mm256_setzero_ph (), > + (__mmask16) -1); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_rsqrt_ph (__m128h __A, __mmask8 __B, __m128h __C) > { > - return __builtin_ia32_vrsqrtph_v8hf_mask (__C, __A, __B); > + return __builtin_ia32_rsqrtph128_mask (__C, __A, __B); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_rsqrt_ph (__m256h __A, __mmask16 __B, __m256h __C) > { > - return __builtin_ia32_vrsqrtph_v16hf_mask (__C, __A, __B); > + return __builtin_ia32_rsqrtph256_mask (__C, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_rsqrt_ph (__mmask8 __A, __m128h __B) > { > - return __builtin_ia32_vrsqrtph_v8hf_mask (__B, _mm_setzero_ph (), __A); > + return __builtin_ia32_rsqrtph128_mask (__B, _mm_setzero_ph (), __A); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_rsqrt_ph (__mmask16 __A, __m256h __B) > { > - return __builtin_ia32_vrsqrtph_v16hf_mask (__B, _mm256_setzero_ph (), > - __A); > + return __builtin_ia32_rsqrtph256_mask (__B, _mm256_setzero_ph (), > + __A); > } > > /* Intrinsics vrcpph. */ > @@ -456,45 +456,45 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_rcp_ph (__m128h __A) > { > - return __builtin_ia32_vrcpph_v8hf_mask (__A, _mm_setzero_ph (), > - (__mmask8) -1); > + return __builtin_ia32_rcpph128_mask (__A, _mm_setzero_ph (), > + (__mmask8) -1); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_rcp_ph (__m256h __A) > { > - return __builtin_ia32_vrcpph_v16hf_mask (__A, _mm256_setzero_ph (), > - (__mmask16) -1); > + return __builtin_ia32_rcpph256_mask (__A, _mm256_setzero_ph (), > + (__mmask16) -1); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_rcp_ph (__m128h __A, __mmask8 __B, __m128h __C) > { > - return __builtin_ia32_vrcpph_v8hf_mask (__C, __A, __B); > + return __builtin_ia32_rcpph128_mask (__C, __A, __B); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_rcp_ph (__m256h __A, __mmask16 __B, __m256h __C) > { > - return __builtin_ia32_vrcpph_v16hf_mask (__C, __A, __B); > + return __builtin_ia32_rcpph256_mask (__C, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_rcp_ph (__mmask8 __A, __m128h __B) > { > - return __builtin_ia32_vrcpph_v8hf_mask (__B, _mm_setzero_ph (), __A); > + return __builtin_ia32_rcpph128_mask (__B, _mm_setzero_ph (), __A); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_rcp_ph (__mmask16 __A, __m256h __B) > { > - return __builtin_ia32_vrcpph_v16hf_mask (__B, _mm256_setzero_ph (), > - __A); > + return __builtin_ia32_rcpph256_mask (__B, _mm256_setzero_ph (), > + __A); > } > > /* Intrinsics vscalefph. */ > @@ -502,25 +502,25 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_scalef_ph (__m128h __A, __m128h __B) > { > - return __builtin_ia32_vscalefph_v8hf_mask (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1); > + return __builtin_ia32_scalefph128_mask (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_scalef_ph (__m256h __A, __m256h __B) > { > - return __builtin_ia32_vscalefph_v16hf_mask (__A, __B, > - _mm256_setzero_ph (), > - (__mmask16) -1); > + return __builtin_ia32_scalefph256_mask (__A, __B, > + _mm256_setzero_ph (), > + (__mmask16) -1); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_scalef_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > { > - return __builtin_ia32_vscalefph_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_scalefph128_mask (__C, __D, __A, __B); > } > > extern __inline __m256h > @@ -528,24 +528,24 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_scalef_ph (__m256h __A, __mmask16 __B, __m256h __C, > __m256h __D) > { > - return __builtin_ia32_vscalefph_v16hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_scalefph256_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_scalef_ph (__mmask8 __A, __m128h __B, __m128h __C) > { > - return __builtin_ia32_vscalefph_v8hf_mask (__B, __C, > - _mm_setzero_ph (), __A); > + return __builtin_ia32_scalefph128_mask (__B, __C, > + _mm_setzero_ph (), __A); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_scalef_ph (__mmask16 __A, __m256h __B, __m256h __C) > { > - return __builtin_ia32_vscalefph_v16hf_mask (__B, __C, > - _mm256_setzero_ph (), > - __A); > + return __builtin_ia32_scalefph256_mask (__B, __C, > + _mm256_setzero_ph (), > + __A); > } > > /* Intrinsics vreduceph. */ > @@ -554,109 +554,109 @@ extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_reduce_ph (__m128h __A, int __B) > { > - return __builtin_ia32_vreduceph_v8hf_mask (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1); > + return __builtin_ia32_reduceph128_mask (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_reduce_ph (__m128h __A, __mmask8 __B, __m128h __C, int __D) > { > - return __builtin_ia32_vreduceph_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_reduceph128_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_reduce_ph (__mmask8 __A, __m128h __B, int __C) > { > - return __builtin_ia32_vreduceph_v8hf_mask (__B, __C, > - _mm_setzero_ph (), __A); > + return __builtin_ia32_reduceph128_mask (__B, __C, > + _mm_setzero_ph (), __A); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_reduce_ph (__m256h __A, int __B) > { > - return __builtin_ia32_vreduceph_v16hf_mask (__A, __B, > - _mm256_setzero_ph (), > - (__mmask16) -1); > + return __builtin_ia32_reduceph256_mask (__A, __B, > + _mm256_setzero_ph (), > + (__mmask16) -1); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_reduce_ph (__m256h __A, __mmask16 __B, __m256h __C, int __D) > { > - return __builtin_ia32_vreduceph_v16hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_reduceph256_mask (__C, __D, __A, __B); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_reduce_ph (__mmask16 __A, __m256h __B, int __C) > { > - return __builtin_ia32_vreduceph_v16hf_mask (__B, __C, > - _mm256_setzero_ph (), > - __A); > + return __builtin_ia32_reduceph256_mask (__B, __C, > + _mm256_setzero_ph (), > + __A); > } > > #else > -#define _mm_reduce_ph(A, B) \ > - (__builtin_ia32_vreduceph_v8hf_mask ((A), (B),\ > - _mm_setzero_ph (), \ > - ((__mmask8)-1))) > +#define _mm_reduce_ph(A, B) \ > + (__builtin_ia32_reduceph128_mask ((A), (B), \ > + _mm_setzero_ph (), \ > + ((__mmask8)-1))) > > -#define _mm_mask_reduce_ph(A, B, C, D) \ > - (__builtin_ia32_vreduceph_v8hf_mask ((C), (D), (A), (B))) > +#define _mm_mask_reduce_ph(A, B, C, D) \ > + (__builtin_ia32_reduceph128_mask ((C), (D), (A), (B))) > > -#define _mm_maskz_reduce_ph(A, B, C) \ > - (__builtin_ia32_vreduceph_v8hf_mask ((B), (C), _mm_setzero_ph (), (A))) > +#define _mm_maskz_reduce_ph(A, B, C) \ > + (__builtin_ia32_reduceph128_mask ((B), (C), _mm_setzero_ph (), (A))) > > #define _mm256_reduce_ph(A, B) \ > - (__builtin_ia32_vreduceph_v16hf_mask ((A), (B),\ > - _mm256_setzero_ph (), \ > - ((__mmask16)-1))) > + (__builtin_ia32_reduceph256_mask ((A), (B), \ > + _mm256_setzero_ph (), \ > + ((__mmask16)-1))) > > -#define _mm256_mask_reduce_ph(A, B, C, D) \ > - (__builtin_ia32_vreduceph_v16hf_mask ((C), (D), (A), (B))) > +#define _mm256_mask_reduce_ph(A, B, C, D) \ > + (__builtin_ia32_reduceph256_mask ((C), (D), (A), (B))) > > -#define _mm256_maskz_reduce_ph(A, B, C) \ > - (__builtin_ia32_vreduceph_v16hf_mask ((B), (C), _mm256_setzero_ph (), (A))) > +#define _mm256_maskz_reduce_ph(A, B, C) \ > + (__builtin_ia32_reduceph256_mask ((B), (C), _mm256_setzero_ph (), (A))) > > #endif /* __OPTIMIZE__ */ > > /* Intrinsics vrndscaleph. */ > #ifdef __OPTIMIZE__ > -extern __inline __m128h > -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > -_mm_roundscale_ph (__m128h __A, int __B) > -{ > - return __builtin_ia32_vrndscaleph_v8hf_mask (__A, __B, > - _mm_setzero_ph (), > - (__mmask8) -1); > -} > + extern __inline __m128h > + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > + _mm_roundscale_ph (__m128h __A, int __B) > + { > + return __builtin_ia32_rndscaleph128_mask (__A, __B, > + _mm_setzero_ph (), > + (__mmask8) -1); > + } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_roundscale_ph (__m128h __A, __mmask8 __B, __m128h __C, int __D) > { > - return __builtin_ia32_vrndscaleph_v8hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_rndscaleph128_mask (__C, __D, __A, __B); > } > > extern __inline __m128h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_roundscale_ph (__mmask8 __A, __m128h __B, int __C) > { > - return __builtin_ia32_vrndscaleph_v8hf_mask (__B, __C, > - _mm_setzero_ph (), __A); > + return __builtin_ia32_rndscaleph128_mask (__B, __C, > + _mm_setzero_ph (), __A); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_roundscale_ph (__m256h __A, int __B) > { > - return __builtin_ia32_vrndscaleph_v16hf_mask (__A, __B, > - _mm256_setzero_ph (), > - (__mmask16) -1); > + return __builtin_ia32_rndscaleph256_mask (__A, __B, > + _mm256_setzero_ph (), > + (__mmask16) -1); > } > > extern __inline __m256h > @@ -664,40 +664,40 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_roundscale_ph (__m256h __A, __mmask16 __B, __m256h __C, > int __D) > { > - return __builtin_ia32_vrndscaleph_v16hf_mask (__C, __D, __A, __B); > + return __builtin_ia32_rndscaleph256_mask (__C, __D, __A, __B); > } > > extern __inline __m256h > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_roundscale_ph (__mmask16 __A, __m256h __B, int __C) > { > - return __builtin_ia32_vrndscaleph_v16hf_mask (__B, __C, > - _mm256_setzero_ph (), > - __A); > + return __builtin_ia32_rndscaleph256_mask (__B, __C, > + _mm256_setzero_ph (), > + __A); > } > > #else > -#define _mm_roundscale_ph(A, B) \ > - (__builtin_ia32_vrndscaleph_v8hf_mask ((A), (B), _mm_setzero_ph (), \ > - ((__mmask8)-1))) > +#define _mm_roundscale_ph(A, B) \ > + (__builtin_ia32_rndscaleph128_mask ((A), (B), _mm_setzero_ph (), \ > + ((__mmask8)-1))) > > -#define _mm_mask_roundscale_ph(A, B, C, D) \ > - (__builtin_ia32_vrndscaleph_v8hf_mask ((C), (D), (A), (B))) > +#define _mm_mask_roundscale_ph(A, B, C, D) \ > + (__builtin_ia32_rndscaleph128_mask ((C), (D), (A), (B))) > > -#define _mm_maskz_roundscale_ph(A, B, C) \ > - (__builtin_ia32_vrndscaleph_v8hf_mask ((B), (C), _mm_setzero_ph (), (A))) > +#define _mm_maskz_roundscale_ph(A, B, C) \ > + (__builtin_ia32_rndscaleph128_mask ((B), (C), _mm_setzero_ph (), (A))) > > -#define _mm256_roundscale_ph(A, B) \ > - (__builtin_ia32_vrndscaleph_v16hf_mask ((A), (B), \ > - _mm256_setzero_ph(), \ > - ((__mmask16)-1))) > +#define _mm256_roundscale_ph(A, B) \ > + (__builtin_ia32_rndscaleph256_mask ((A), (B), \ > + _mm256_setzero_ph(), \ > + ((__mmask16)-1))) > > -#define _mm256_mask_roundscale_ph(A, B, C, D) \ > - (__builtin_ia32_vrndscaleph_v16hf_mask ((C), (D), (A), (B))) > +#define _mm256_mask_roundscale_ph(A, B, C, D) \ > + (__builtin_ia32_rndscaleph256_mask ((C), (D), (A), (B))) > > -#define _mm256_maskz_roundscale_ph(A, B, C) \ > - (__builtin_ia32_vrndscaleph_v16hf_mask ((B), (C), \ > - _mm256_setzero_ph (), (A))) > +#define _mm256_maskz_roundscale_ph(A, B, C) \ > + (__builtin_ia32_rndscaleph256_mask ((B), (C), \ > + _mm256_setzero_ph (), (A))) > > #endif /* __OPTIMIZE__ */ > > @@ -705,7 +705,7 @@ _mm256_maskz_roundscale_ph (__mmask16 __A, __m256h __B, int __C) > #ifdef __OPTIMIZE__ > extern __inline __mmask8 > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > -_mm_mask_fpclass_ph_mask (__mmask8 __U, __m128h __A, const int __imm) > + _mm_mask_fpclass_ph_mask (__mmask8 __U, __m128h __A, const int __imm) > { > return (__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) __A, > __imm, __U); > @@ -725,7 +725,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_fpclass_ph_mask (__mmask16 __U, __m256h __A, const int __imm) > { > return (__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) __A, > - __imm, __U); > + __imm, __U); > } > > extern __inline __mmask16 > @@ -733,26 +733,26 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_fpclass_ph_mask (__m256h __A, const int __imm) > { > return (__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) __A, > - __imm, > - (__mmask16) -1); > + __imm, > + (__mmask16) -1); > } > > #else > #define _mm_fpclass_ph_mask(X, C) \ > - ((__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) (__m128h) (X), \ > + ((__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) (__m128h) (X), \ > (int) (C),(__mmask8)-1)) > > #define _mm_mask_fpclass_ph_mask(u, X, C) \ > - ((__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) (__m128h) (X), \ > + ((__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) (__m128h) (X), \ > (int) (C),(__mmask8)(u))) > > #define _mm256_fpclass_ph_mask(X, C) \ > - ((__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) (__m256h) (X), \ > - (int) (C),(__mmask16)-1)) > + ((__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) (__m256h) (X), \ > + (int) (C),(__mmask16)-1)) > > #define _mm256_mask_fpclass_ph_mask(u, X, C) \ > - ((__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) (__m256h) (X), \ > - (int) (C),(__mmask16)(u))) > + ((__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) (__m256h) (X), \ > + (int) (C),(__mmask16)(u))) > #endif /* __OPTIMIZE__ */ > > /* Intrinsics vgetexpph, vgetexpsh. */ > @@ -761,9 +761,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_getexp_ph (__m256h __A) > { > return (__m256h) __builtin_ia32_getexpph256_mask ((__v16hf) __A, > - (__v16hf) > - _mm256_setzero_ph (), > - (__mmask16) -1); > + (__v16hf) > + _mm256_setzero_ph (), > + (__mmask16) -1); > } > > extern __inline __m256h > @@ -771,8 +771,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_mask_getexp_ph (__m256h __W, __mmask16 __U, __m256h __A) > { > return (__m256h) __builtin_ia32_getexpph256_mask ((__v16hf) __A, > - (__v16hf) __W, > - (__mmask16) __U); > + (__v16hf) __W, > + (__mmask16) __U); > } > > extern __inline __m256h > @@ -780,9 +780,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm256_maskz_getexp_ph (__mmask16 __U, __m256h __A) > { > return (__m256h) __builtin_ia32_getexpph256_mask ((__v16hf) __A, > - (__v16hf) > - _mm256_setzero_ph (), > - (__mmask16) __U); > + (__v16hf) > + _mm256_setzero_ph (), > + (__mmask16) __U); > } > > extern __inline __m128h > @@ -790,9 +790,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_getexp_ph (__m128h __A) > { > return (__m128h) __builtin_ia32_getexpph128_mask ((__v8hf) __A, > - (__v8hf) > - _mm_setzero_ph (), > - (__mmask8) -1); > + (__v8hf) > + _mm_setzero_ph (), > + (__mmask8) -1); > } > > extern __inline __m128h > @@ -800,8 +800,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_mask_getexp_ph (__m128h __W, __mmask8 __U, __m128h __A) > { > return (__m128h) __builtin_ia32_getexpph128_mask ((__v8hf) __A, > - (__v8hf) __W, > - (__mmask8) __U); > + (__v8hf) __W, > + (__mmask8) __U); > } > > extern __inline __m128h > @@ -809,9 +809,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > _mm_maskz_getexp_ph (__mmask8 __U, __m128h __A) > { > return (__m128h) __builtin_ia32_getexpph128_mask ((__v8hf) __A, > - (__v8hf) > - _mm_setzero_ph (), > - (__mmask8) __U); > + (__v8hf) > + _mm_setzero_ph (), > + (__mmask8) __U); > } > > > @@ -892,41 +892,41 @@ _mm_maskz_getmant_ph (__mmask8 __U, __m128h __A, > } > > #else > -#define _mm256_getmant_ph(X, B, C) \ > - ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ > - (int)(((C)<<2) | (B)), \ > - (__v16hf)(__m256h)_mm256_setzero_ph (),\ > - (__mmask16)-1)) > - > -#define _mm256_mask_getmant_ph(W, U, X, B, C) \ > - ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ > - (int)(((C)<<2) | (B)), \ > - (__v16hf)(__m256h)(W), \ > - (__mmask16)(U))) > - > -#define _mm256_maskz_getmant_ph(U, X, B, C) \ > - ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ > - (int)(((C)<<2) | (B)), \ > - (__v16hf)(__m256h)_mm256_setzero_ph (),\ > - (__mmask16)(U))) > - > -#define _mm_getmant_ph(X, B, C) \ > - ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ > - (int)(((C)<<2) | (B)), \ > - (__v8hf)(__m128h)_mm_setzero_ph (), \ > - (__mmask8)-1)) > - > -#define _mm_mask_getmant_ph(W, U, X, B, C) \ > - ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ > - (int)(((C)<<2) | (B)), \ > - (__v8hf)(__m128h)(W), \ > - (__mmask8)(U))) > - > -#define _mm_maskz_getmant_ph(U, X, B, C) \ > - ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ > - (int)(((C)<<2) | (B)), \ > - (__v8hf)(__m128h)_mm_setzero_ph (), \ > - (__mmask8)(U))) > +#define _mm256_getmant_ph(X, B, C) \ > + ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ > + (int)(((C)<<2) | (B)), \ > + (__v16hf)(__m256h)_mm256_setzero_ph (), \ > + (__mmask16)-1)) > + > +#define _mm256_mask_getmant_ph(W, U, X, B, C) \ > + ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ > + (int)(((C)<<2) | (B)), \ > + (__v16hf)(__m256h)(W), \ > + (__mmask16)(U))) > + > +#define _mm256_maskz_getmant_ph(U, X, B, C) \ > + ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ > + (int)(((C)<<2) | (B)), \ > + (__v16hf)(__m256h)_mm256_setzero_ph (), \ > + (__mmask16)(U))) > + > +#define _mm_getmant_ph(X, B, C) \ > + ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ > + (int)(((C)<<2) | (B)), \ > + (__v8hf)(__m128h)_mm_setzero_ph (), \ > + (__mmask8)-1)) > + > +#define _mm_mask_getmant_ph(W, U, X, B, C) \ > + ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ > + (int)(((C)<<2) | (B)), \ > + (__v8hf)(__m128h)(W), \ > + (__mmask8)(U))) > + > +#define _mm_maskz_getmant_ph(U, X, B, C) \ > + ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ > + (int)(((C)<<2) | (B)), \ > + (__v8hf)(__m128h)_mm_setzero_ph (), \ > + (__mmask8)(U))) > > #endif /* __OPTIMIZE__ */ > > diff --git a/gcc/config/i386/i386-builtin.def b/gcc/config/i386/i386-builtin.def > index bfa1d56a7b8..10f6fd87cbb 100644 > --- a/gcc/config/i386/i386-builtin.def > +++ b/gcc/config/i386/i386-builtin.def > @@ -2775,49 +2775,49 @@ BDESC (0, OPTION_MASK_ISA2_AVX512BF16, CODE_FOR_avx512f_dpbf16ps_v4sf_mask, "__b > BDESC (0, OPTION_MASK_ISA2_AVX512BF16, CODE_FOR_avx512f_dpbf16ps_v4sf_maskz, "__builtin_ia32_dpbf16ps_v4sf_maskz", IX86_BUILTIN_DPHI16PS_V4SF_MASKZ, UNKNOWN, (int) V4SF_FTYPE_V4SF_V8HI_V8HI_UQI) > > /* AVX512FP16. */ > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv8hf3_mask, "__builtin_ia32_vaddph_v8hf_mask", IX86_BUILTIN_VADDPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv16hf3_mask, "__builtin_ia32_vaddph_v16hf_mask", IX86_BUILTIN_VADDPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv32hf3_mask, "__builtin_ia32_vaddph_v32hf_mask", IX86_BUILTIN_VADDPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv8hf3_mask, "__builtin_ia32_vsubph_v8hf_mask", IX86_BUILTIN_VSUBPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv16hf3_mask, "__builtin_ia32_vsubph_v16hf_mask", IX86_BUILTIN_VSUBPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv32hf3_mask, "__builtin_ia32_vsubph_v32hf_mask", IX86_BUILTIN_VSUBPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv8hf3_mask, "__builtin_ia32_vmulph_v8hf_mask", IX86_BUILTIN_VMULPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv16hf3_mask, "__builtin_ia32_vmulph_v16hf_mask", IX86_BUILTIN_VMULPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv32hf3_mask, "__builtin_ia32_vmulph_v32hf_mask", IX86_BUILTIN_VMULPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv8hf3_mask, "__builtin_ia32_vdivph_v8hf_mask", IX86_BUILTIN_VDIVPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv16hf3_mask, "__builtin_ia32_vdivph_v16hf_mask", IX86_BUILTIN_VDIVPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv32hf3_mask, "__builtin_ia32_vdivph_v32hf_mask", IX86_BUILTIN_VDIVPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmaddv8hf3_mask, "__builtin_ia32_vaddsh_v8hf_mask", IX86_BUILTIN_VADDSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsubv8hf3_mask, "__builtin_ia32_vsubsh_v8hf_mask", IX86_BUILTIN_VSUBSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmmulv8hf3_mask, "__builtin_ia32_vmulsh_v8hf_mask", IX86_BUILTIN_VMULSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmdivv8hf3_mask, "__builtin_ia32_vdivsh_v8hf_mask", IX86_BUILTIN_VDIVSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv8hf3_mask, "__builtin_ia32_vmaxph_v8hf_mask", IX86_BUILTIN_VMAXPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv16hf3_mask, "__builtin_ia32_vmaxph_v16hf_mask", IX86_BUILTIN_VMAXPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv32hf3_mask, "__builtin_ia32_vmaxph_v32hf_mask", IX86_BUILTIN_VMAXPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv8hf3_mask, "__builtin_ia32_vminph_v8hf_mask", IX86_BUILTIN_VMINPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv16hf3_mask, "__builtin_ia32_vminph_v16hf_mask", IX86_BUILTIN_VMINPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv32hf3_mask, "__builtin_ia32_vminph_v32hf_mask", IX86_BUILTIN_VMINPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask, "__builtin_ia32_vmaxsh_v8hf_mask", IX86_BUILTIN_VMAXSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask, "__builtin_ia32_vminsh_v8hf_mask", IX86_BUILTIN_VMINSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_cmpv8hf3_mask, "__builtin_ia32_vcmpph_v8hf_mask", IX86_BUILTIN_VCMPPH_V8HF_MASK, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_cmpv16hf3_mask, "__builtin_ia32_vcmpph_v16hf_mask", IX86_BUILTIN_VCMPPH_V16HF_MASK, UNKNOWN, (int) UHI_FTYPE_V16HF_V16HF_INT_UHI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask, "__builtin_ia32_vcmpph_v32hf_mask", IX86_BUILTIN_VCMPPH_V32HF_MASK, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv8hf2_mask, "__builtin_ia32_vsqrtph_v8hf_mask", IX86_BUILTIN_VSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv16hf2_mask, "__builtin_ia32_vsqrtph_v16hf_mask", IX86_BUILTIN_VSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv8hf2_mask, "__builtin_ia32_vrsqrtph_v8hf_mask", IX86_BUILTIN_VRSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv16hf2_mask, "__builtin_ia32_vrsqrtph_v16hf_mask", IX86_BUILTIN_VRSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv32hf2_mask, "__builtin_ia32_vrsqrtph_v32hf_mask", IX86_BUILTIN_VRSQRTPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrsqrtv8hf2_mask, "__builtin_ia32_vrsqrtsh_v8hf_mask", IX86_BUILTIN_VRSQRTSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv8hf2_mask, "__builtin_ia32_vrcpph_v8hf_mask", IX86_BUILTIN_VRCPPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv16hf2_mask, "__builtin_ia32_vrcpph_v16hf_mask", IX86_BUILTIN_VRCPPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv32hf2_mask, "__builtin_ia32_vrcpph_v32hf_mask", IX86_BUILTIN_VRCPPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrcpv8hf2_mask, "__builtin_ia32_vrcpsh_v8hf_mask", IX86_BUILTIN_VRCPSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_scalefv8hf_mask, "__builtin_ia32_vscalefph_v8hf_mask", IX86_BUILTIN_VSCALEFPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_scalefv16hf_mask, "__builtin_ia32_vscalefph_v16hf_mask", IX86_BUILTIN_VSCALEFPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv8hf_mask, "__builtin_ia32_vreduceph_v8hf_mask", IX86_BUILTIN_VREDUCEPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_INT_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv16hf_mask, "__builtin_ia32_vreduceph_v16hf_mask", IX86_BUILTIN_VREDUCEPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_INT_V16HF_UHI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rndscalev8hf_mask, "__builtin_ia32_vrndscaleph_v8hf_mask", IX86_BUILTIN_VRNDSCALEPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_INT_V8HF_UQI) > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_rndscalev16hf_mask, "__builtin_ia32_vrndscaleph_v16hf_mask", IX86_BUILTIN_VRNDSCALEPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_INT_V16HF_UHI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv8hf3_mask, "__builtin_ia32_addph128_mask", IX86_BUILTIN_VADDPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv16hf3_mask, "__builtin_ia32_addph256_mask", IX86_BUILTIN_VADDPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv32hf3_mask, "__builtin_ia32_addph512_mask", IX86_BUILTIN_VADDPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv8hf3_mask, "__builtin_ia32_subph128_mask", IX86_BUILTIN_VSUBPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv16hf3_mask, "__builtin_ia32_subph256_mask", IX86_BUILTIN_VSUBPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv32hf3_mask, "__builtin_ia32_subph512_mask", IX86_BUILTIN_VSUBPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv8hf3_mask, "__builtin_ia32_mulph128_mask", IX86_BUILTIN_VMULPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv16hf3_mask, "__builtin_ia32_mulph256_mask", IX86_BUILTIN_VMULPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv32hf3_mask, "__builtin_ia32_mulph512_mask", IX86_BUILTIN_VMULPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv8hf3_mask, "__builtin_ia32_divph128_mask", IX86_BUILTIN_VDIVPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv16hf3_mask, "__builtin_ia32_divph256_mask", IX86_BUILTIN_VDIVPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv32hf3_mask, "__builtin_ia32_divph512_mask", IX86_BUILTIN_VDIVPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmaddv8hf3_mask, "__builtin_ia32_addsh_mask", IX86_BUILTIN_VADDSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsubv8hf3_mask, "__builtin_ia32_subsh_mask", IX86_BUILTIN_VSUBSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmmulv8hf3_mask, "__builtin_ia32_mulsh_mask", IX86_BUILTIN_VMULSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmdivv8hf3_mask, "__builtin_ia32_divsh_mask", IX86_BUILTIN_VDIVSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv8hf3_mask, "__builtin_ia32_maxph128_mask", IX86_BUILTIN_VMAXPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv16hf3_mask, "__builtin_ia32_maxph256_mask", IX86_BUILTIN_VMAXPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv32hf3_mask, "__builtin_ia32_maxph512_mask", IX86_BUILTIN_VMAXPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv8hf3_mask, "__builtin_ia32_minph128_mask", IX86_BUILTIN_VMINPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv16hf3_mask, "__builtin_ia32_minph256_mask", IX86_BUILTIN_VMINPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv32hf3_mask, "__builtin_ia32_minph512_mask", IX86_BUILTIN_VMINPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask, "__builtin_ia32_maxsh_mask", IX86_BUILTIN_VMAXSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask, "__builtin_ia32_minsh_mask", IX86_BUILTIN_VMINSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_cmpv8hf3_mask, "__builtin_ia32_cmpph128_mask", IX86_BUILTIN_VCMPPH_V8HF_MASK, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_cmpv16hf3_mask, "__builtin_ia32_cmpph256_mask", IX86_BUILTIN_VCMPPH_V16HF_MASK, UNKNOWN, (int) UHI_FTYPE_V16HF_V16HF_INT_UHI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask, "__builtin_ia32_cmpph512_mask", IX86_BUILTIN_VCMPPH_V32HF_MASK, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv8hf2_mask, "__builtin_ia32_sqrtph128_mask", IX86_BUILTIN_VSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv16hf2_mask, "__builtin_ia32_sqrtph256_mask", IX86_BUILTIN_VSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv8hf2_mask, "__builtin_ia32_rsqrtph128_mask", IX86_BUILTIN_VRSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv16hf2_mask, "__builtin_ia32_rsqrtph256_mask", IX86_BUILTIN_VRSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv32hf2_mask, "__builtin_ia32_rsqrtph512_mask", IX86_BUILTIN_VRSQRTPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrsqrtv8hf2_mask, "__builtin_ia32_rsqrtsh_mask", IX86_BUILTIN_VRSQRTSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv8hf2_mask, "__builtin_ia32_rcpph128_mask", IX86_BUILTIN_VRCPPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv16hf2_mask, "__builtin_ia32_rcpph256_mask", IX86_BUILTIN_VRCPPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv32hf2_mask, "__builtin_ia32_rcpph512_mask", IX86_BUILTIN_VRCPPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrcpv8hf2_mask, "__builtin_ia32_rcpsh_mask", IX86_BUILTIN_VRCPSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_scalefv8hf_mask, "__builtin_ia32_scalefph128_mask", IX86_BUILTIN_VSCALEFPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_scalefv16hf_mask, "__builtin_ia32_scalefph256_mask", IX86_BUILTIN_VSCALEFPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv8hf_mask, "__builtin_ia32_reduceph128_mask", IX86_BUILTIN_VREDUCEPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_INT_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv16hf_mask, "__builtin_ia32_reduceph256_mask", IX86_BUILTIN_VREDUCEPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_INT_V16HF_UHI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rndscalev8hf_mask, "__builtin_ia32_rndscaleph128_mask", IX86_BUILTIN_VRNDSCALEPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_INT_V8HF_UQI) > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_rndscalev16hf_mask, "__builtin_ia32_rndscaleph256_mask", IX86_BUILTIN_VRNDSCALEPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_INT_V16HF_UHI) > BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512dq_fpclassv16hf_mask, "__builtin_ia32_fpclassph256_mask", IX86_BUILTIN_FPCLASSPH256, UNKNOWN, (int) HI_FTYPE_V16HF_INT_UHI) > BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512dq_fpclassv8hf_mask, "__builtin_ia32_fpclassph128_mask", IX86_BUILTIN_FPCLASSPH128, UNKNOWN, (int) QI_FTYPE_V8HF_INT_UQI) > BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512dq_fpclassv32hf_mask, "__builtin_ia32_fpclassph512_mask", IX86_BUILTIN_FPCLASSPH512, UNKNOWN, (int) SI_FTYPE_V32HF_INT_USI) > @@ -3027,28 +3027,28 @@ BDESC (OPTION_MASK_ISA_AVX512DQ, 0, CODE_FOR_avx512dq_rangepv16sf_mask_round, "_ > BDESC (OPTION_MASK_ISA_AVX512DQ, 0, CODE_FOR_avx512dq_rangepv8df_mask_round, "__builtin_ia32_rangepd512_mask", IX86_BUILTIN_RANGEPD512, UNKNOWN, (int) V8DF_FTYPE_V8DF_V8DF_INT_V8DF_QI_INT) > > /* AVX512FP16. */ > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv32hf3_mask_round, "__builtin_ia32_vaddph_v32hf_mask_round", IX86_BUILTIN_VADDPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv32hf3_mask_round, "__builtin_ia32_vsubph_v32hf_mask_round", IX86_BUILTIN_VSUBPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv32hf3_mask_round, "__builtin_ia32_vmulph_v32hf_mask_round", IX86_BUILTIN_VMULPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv32hf3_mask_round, "__builtin_ia32_vdivph_v32hf_mask_round", IX86_BUILTIN_VDIVPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmaddv8hf3_mask_round, "__builtin_ia32_vaddsh_v8hf_mask_round", IX86_BUILTIN_VADDSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsubv8hf3_mask_round, "__builtin_ia32_vsubsh_v8hf_mask_round", IX86_BUILTIN_VSUBSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmmulv8hf3_mask_round, "__builtin_ia32_vmulsh_v8hf_mask_round", IX86_BUILTIN_VMULSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmdivv8hf3_mask_round, "__builtin_ia32_vdivsh_v8hf_mask_round", IX86_BUILTIN_VDIVSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv32hf3_mask_round, "__builtin_ia32_vmaxph_v32hf_mask_round", IX86_BUILTIN_VMAXPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv32hf3_mask_round, "__builtin_ia32_vminph_v32hf_mask_round", IX86_BUILTIN_VMINPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask_round, "__builtin_ia32_vmaxsh_v8hf_mask_round", IX86_BUILTIN_VMAXSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask_round, "__builtin_ia32_vminsh_v8hf_mask_round", IX86_BUILTIN_VMINSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask_round, "__builtin_ia32_vcmpph_v32hf_mask_round", IX86_BUILTIN_VCMPPH_V32HF_MASK_ROUND, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_vmcmpv8hf3_mask_round, "__builtin_ia32_vcmpsh_v8hf_mask_round", IX86_BUILTIN_VCMPSH_V8HF_MASK_ROUND, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv32hf2_mask_round, "__builtin_ia32_vsqrtph_v32hf_mask_round", IX86_BUILTIN_VSQRTPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsqrtv8hf2_mask_round, "__builtin_ia32_vsqrtsh_v8hf_mask_round", IX86_BUILTIN_VSQRTSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_scalefv32hf_mask_round, "__builtin_ia32_vscalefph_v32hf_mask_round", IX86_BUILTIN_VSCALEFPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_vmscalefv8hf_mask_round, "__builtin_ia32_vscalefsh_v8hf_mask_round", IX86_BUILTIN_VSCALEFSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv32hf_mask_round, "__builtin_ia32_vreduceph_v32hf_mask_round", IX86_BUILTIN_VREDUCEPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducesv8hf_mask_round, "__builtin_ia32_vreducesh_v8hf_mask_round", IX86_BUILTIN_VREDUCESH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_INT_V8HF_UQI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_rndscalev32hf_mask_round, "__builtin_ia32_vrndscaleph_v32hf_mask_round", IX86_BUILTIN_VRNDSCALEPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_rndscalev8hf_mask_round, "__builtin_ia32_vrndscalesh_v8hf_mask_round", IX86_BUILTIN_VRNDSCALESH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_INT_V8HF_UQI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv32hf3_mask_round, "__builtin_ia32_addph512_mask_round", IX86_BUILTIN_VADDPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv32hf3_mask_round, "__builtin_ia32_subph512_mask_round", IX86_BUILTIN_VSUBPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv32hf3_mask_round, "__builtin_ia32_mulph512_mask_round", IX86_BUILTIN_VMULPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv32hf3_mask_round, "__builtin_ia32_divph512_mask_round", IX86_BUILTIN_VDIVPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmaddv8hf3_mask_round, "__builtin_ia32_addsh_mask_round", IX86_BUILTIN_VADDSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsubv8hf3_mask_round, "__builtin_ia32_subsh_mask_round", IX86_BUILTIN_VSUBSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmmulv8hf3_mask_round, "__builtin_ia32_mulsh_mask_round", IX86_BUILTIN_VMULSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmdivv8hf3_mask_round, "__builtin_ia32_divsh_mask_round", IX86_BUILTIN_VDIVSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv32hf3_mask_round, "__builtin_ia32_maxph512_mask_round", IX86_BUILTIN_VMAXPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv32hf3_mask_round, "__builtin_ia32_minph512_mask_round", IX86_BUILTIN_VMINPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask_round, "__builtin_ia32_maxsh_mask_round", IX86_BUILTIN_VMAXSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask_round, "__builtin_ia32_minsh_mask_round", IX86_BUILTIN_VMINSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask_round, "__builtin_ia32_cmpph512_mask_round", IX86_BUILTIN_VCMPPH_V32HF_MASK_ROUND, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_vmcmpv8hf3_mask_round, "__builtin_ia32_cmpsh_mask_round", IX86_BUILTIN_VCMPSH_V8HF_MASK_ROUND, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv32hf2_mask_round, "__builtin_ia32_sqrtph512_mask_round", IX86_BUILTIN_VSQRTPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsqrtv8hf2_mask_round, "__builtin_ia32_sqrtsh_mask_round", IX86_BUILTIN_VSQRTSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_scalefv32hf_mask_round, "__builtin_ia32_scalefph512_mask_round", IX86_BUILTIN_VSCALEFPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_vmscalefv8hf_mask_round, "__builtin_ia32_scalefsh_mask_round", IX86_BUILTIN_VSCALEFSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv32hf_mask_round, "__builtin_ia32_reduceph512_mask_round", IX86_BUILTIN_VREDUCEPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducesv8hf_mask_round, "__builtin_ia32_reducesh_mask_round", IX86_BUILTIN_VREDUCESH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_INT_V8HF_UQI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_rndscalev32hf_mask_round, "__builtin_ia32_rndscaleph512_mask_round", IX86_BUILTIN_VRNDSCALEPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_rndscalev8hf_mask_round, "__builtin_ia32_rndscalesh_mask_round", IX86_BUILTIN_VRNDSCALESH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_INT_V8HF_UQI_INT) > BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_getexpv32hf_mask_round, "__builtin_ia32_getexpph512_mask", IX86_BUILTIN_GETEXPPH512, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI_INT) > BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_sgetexpv8hf_mask_round, "__builtin_ia32_getexpsh_mask_round", IX86_BUILTIN_GETEXPSH_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_getmantv32hf_mask_round, "__builtin_ia32_getmantph512_mask", IX86_BUILTIN_GETMANTPH512, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) > diff --git a/gcc/testsuite/gcc.target/i386/avx-1.c b/gcc/testsuite/gcc.target/i386/avx-1.c > index b3cffa0644f..3a96e586418 100644 > --- a/gcc/testsuite/gcc.target/i386/avx-1.c > +++ b/gcc/testsuite/gcc.target/i386/avx-1.c > @@ -686,33 +686,33 @@ > #define __builtin_ia32_vpshld_v2di_mask(A, B, C, D, E) __builtin_ia32_vpshld_v2di_mask(A, B, 1, D, E) > > /* avx512fp16intrin.h */ > -#define __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vcmpph_v32hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v32hf_mask(A, B, 1, D) > -#define __builtin_ia32_vcmpph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpph_v32hf_mask_round(A, B, 1, D, 8) > -#define __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, 1, D, 8) > -#define __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, D) __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, 8) > -#define __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, E) __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, 8) > -#define __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vreduceph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vreduceph_v32hf_mask_round(A, 123, C, D, 8) > -#define __builtin_ia32_vreduceph_v8hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v8hf_mask(A, 123, C, D) > -#define __builtin_ia32_vreduceph_v16hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v16hf_mask(A, 123, C, D) > -#define __builtin_ia32_vreducesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vreducesh_v8hf_mask_round(A, B, 123, D, E, 8) > -#define __builtin_ia32_vrndscaleph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vrndscaleph_v32hf_mask_round(A, 123, C, D, 8) > -#define __builtin_ia32_vrndscaleph_v8hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v8hf_mask(A, 123, C, D) > -#define __builtin_ia32_vrndscaleph_v16hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v16hf_mask(A, 123, C, D) > -#define __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, 123, D, E, 8) > +#define __builtin_ia32_addph512_mask_round(A, B, C, D, E) __builtin_ia32_addph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_subph512_mask_round(A, B, C, D, E) __builtin_ia32_subph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_mulph512_mask_round(A, B, C, D, E) __builtin_ia32_mulph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_divph512_mask_round(A, B, C, D, E) __builtin_ia32_divph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_addsh_mask_round(A, B, C, D, E) __builtin_ia32_addsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_subsh_mask_round(A, B, C, D, E) __builtin_ia32_subsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_mulsh_mask_round(A, B, C, D, E) __builtin_ia32_mulsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_divsh_mask_round(A, B, C, D, E) __builtin_ia32_divsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_maxph512_mask_round(A, B, C, D, E) __builtin_ia32_maxph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_minph512_mask_round(A, B, C, D, E) __builtin_ia32_minph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_maxsh_mask_round(A, B, C, D, E) __builtin_ia32_maxsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_minsh_mask_round(A, B, C, D, E) __builtin_ia32_minsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_cmpph512_mask(A, B, C, D) __builtin_ia32_cmpph512_mask(A, B, 1, D) > +#define __builtin_ia32_cmpph512_mask_round(A, B, C, D, E) __builtin_ia32_cmpph512_mask_round(A, B, 1, D, 8) > +#define __builtin_ia32_cmpsh_mask_round(A, B, C, D, E) __builtin_ia32_cmpsh_mask_round(A, B, 1, D, 8) > +#define __builtin_ia32_sqrtph512_mask_round(C, A, B, D) __builtin_ia32_sqrtph512_mask_round(C, A, B, 8) > +#define __builtin_ia32_sqrtsh_mask_round(D, C, A, B, E) __builtin_ia32_sqrtsh_mask_round(D, C, A, B, 8) > +#define __builtin_ia32_scalefph512_mask_round(A, B, C, D, E) __builtin_ia32_scalefph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_scalefsh_mask_round(A, B, C, D, E) __builtin_ia32_scalefsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_reduceph512_mask_round(A, B, C, D, E) __builtin_ia32_reduceph512_mask_round(A, 123, C, D, 8) > +#define __builtin_ia32_reduceph128_mask(A, B, C, D) __builtin_ia32_reduceph128_mask(A, 123, C, D) > +#define __builtin_ia32_reduceph256_mask(A, B, C, D) __builtin_ia32_reduceph256_mask(A, 123, C, D) > +#define __builtin_ia32_reducesh_mask_round(A, B, C, D, E, F) __builtin_ia32_reducesh_mask_round(A, B, 123, D, E, 8) > +#define __builtin_ia32_rndscaleph512_mask_round(A, B, C, D, E) __builtin_ia32_rndscaleph512_mask_round(A, 123, C, D, 8) > +#define __builtin_ia32_rndscaleph128_mask(A, B, C, D) __builtin_ia32_rndscaleph128_mask(A, 123, C, D) > +#define __builtin_ia32_rndscaleph256_mask(A, B, C, D) __builtin_ia32_rndscaleph256_mask(A, 123, C, D) > +#define __builtin_ia32_rndscalesh_mask_round(A, B, C, D, E, F) __builtin_ia32_rndscalesh_mask_round(A, B, 123, D, E, 8) > #define __builtin_ia32_fpclassph512_mask(A, D, C) __builtin_ia32_fpclassph512_mask(A, 1, C) > #define __builtin_ia32_fpclasssh_mask(A, D, U) __builtin_ia32_fpclasssh_mask(A, 1, U) > #define __builtin_ia32_getexpph512_mask(A, B, C, D) __builtin_ia32_getexpph512_mask(A, B, C, 8) > @@ -721,8 +721,8 @@ > #define __builtin_ia32_getmantsh_mask_round(A, B, C, W, U, D) __builtin_ia32_getmantsh_mask_round(A, B, 1, W, U, 4) > > /* avx512fp16vlintrin.h */ > -#define __builtin_ia32_vcmpph_v8hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v8hf_mask(A, B, 1, D) > -#define __builtin_ia32_vcmpph_v16hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v16hf_mask(A, B, 1, D) > +#define __builtin_ia32_cmpph128_mask(A, B, C, D) __builtin_ia32_cmpph128_mask(A, B, 1, D) > +#define __builtin_ia32_cmpph256_mask(A, B, C, D) __builtin_ia32_cmpph256_mask(A, B, 1, D) > #define __builtin_ia32_fpclassph256_mask(A, D, C) __builtin_ia32_fpclassph256_mask(A, 1, C) > #define __builtin_ia32_fpclassph128_mask(A, D, C) __builtin_ia32_fpclassph128_mask(A, 1, C) > #define __builtin_ia32_getmantph256_mask(A, E, C, D) __builtin_ia32_getmantph256_mask(A, 1, C, D) > diff --git a/gcc/testsuite/gcc.target/i386/sse-13.c b/gcc/testsuite/gcc.target/i386/sse-13.c > index 67ef567e437..aafcd414530 100644 > --- a/gcc/testsuite/gcc.target/i386/sse-13.c > +++ b/gcc/testsuite/gcc.target/i386/sse-13.c > @@ -703,33 +703,33 @@ > #define __builtin_ia32_vpshld_v2di_mask(A, B, C, D, E) __builtin_ia32_vpshld_v2di_mask(A, B, 1, D, E) > > /* avx512fp16intrin.h */ > -#define __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vcmpph_v32hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v32hf_mask(A, B, 1, D) > -#define __builtin_ia32_vcmpph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpph_v32hf_mask_round(A, B, 1, D, 8) > -#define __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, 1, D, 8) > -#define __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, D) __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, 8) > -#define __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, E) __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, 8) > -#define __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vreduceph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vreduceph_v32hf_mask_round(A, 123, C, D, 8) > -#define __builtin_ia32_vreduceph_v8hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v8hf_mask(A, 123, C, D) > -#define __builtin_ia32_vreduceph_v16hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v16hf_mask(A, 123, C, D) > -#define __builtin_ia32_vreducesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vreducesh_v8hf_mask_round(A, B, 123, D, E, 8) > -#define __builtin_ia32_vrndscaleph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vrndscaleph_v32hf_mask_round(A, 123, C, D, 8) > -#define __builtin_ia32_vrndscaleph_v8hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v8hf_mask(A, 123, C, D) > -#define __builtin_ia32_vrndscaleph_v16hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v16hf_mask(A, 123, C, D) > -#define __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, 123, D, E, 8) > +#define __builtin_ia32_addph512_mask_round(A, B, C, D, E) __builtin_ia32_addph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_subph512_mask_round(A, B, C, D, E) __builtin_ia32_subph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_mulph512_mask_round(A, B, C, D, E) __builtin_ia32_mulph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_divph512_mask_round(A, B, C, D, E) __builtin_ia32_divph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_addsh_mask_round(A, B, C, D, E) __builtin_ia32_addsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_subsh_mask_round(A, B, C, D, E) __builtin_ia32_subsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_mulsh_mask_round(A, B, C, D, E) __builtin_ia32_mulsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_divsh_mask_round(A, B, C, D, E) __builtin_ia32_divsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_maxph512_mask_round(A, B, C, D, E) __builtin_ia32_maxph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_minph512_mask_round(A, B, C, D, E) __builtin_ia32_minph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_maxsh_mask_round(A, B, C, D, E) __builtin_ia32_maxsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_minsh_mask_round(A, B, C, D, E) __builtin_ia32_minsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_cmpph512_mask(A, B, C, D) __builtin_ia32_cmpph512_mask(A, B, 1, D) > +#define __builtin_ia32_cmpph512_mask_round(A, B, C, D, E) __builtin_ia32_cmpph512_mask_round(A, B, 1, D, 8) > +#define __builtin_ia32_cmpsh_mask_round(A, B, C, D, E) __builtin_ia32_cmpsh_mask_round(A, B, 1, D, 8) > +#define __builtin_ia32_sqrtph512_mask_round(C, A, B, D) __builtin_ia32_sqrtph512_mask_round(C, A, B, 8) > +#define __builtin_ia32_sqrtsh_mask_round(D, C, A, B, E) __builtin_ia32_sqrtsh_mask_round(D, C, A, B, 8) > +#define __builtin_ia32_scalefph512_mask_round(A, B, C, D, E) __builtin_ia32_scalefph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_scalefsh_mask_round(A, B, C, D, E) __builtin_ia32_scalefsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_reduceph512_mask_round(A, B, C, D, E) __builtin_ia32_reduceph512_mask_round(A, 123, C, D, 8) > +#define __builtin_ia32_reduceph128_mask(A, B, C, D) __builtin_ia32_reduceph128_mask(A, 123, C, D) > +#define __builtin_ia32_reduceph256_mask(A, B, C, D) __builtin_ia32_reduceph256_mask(A, 123, C, D) > +#define __builtin_ia32_reducesh_mask_round(A, B, C, D, E, F) __builtin_ia32_reducesh_mask_round(A, B, 123, D, E, 8) > +#define __builtin_ia32_rndscaleph512_mask_round(A, B, C, D, E) __builtin_ia32_rndscaleph512_mask_round(A, 123, C, D, 8) > +#define __builtin_ia32_rndscaleph128_mask(A, B, C, D) __builtin_ia32_rndscaleph128_mask(A, 123, C, D) > +#define __builtin_ia32_rndscaleph256_mask(A, B, C, D) __builtin_ia32_rndscaleph256_mask(A, 123, C, D) > +#define __builtin_ia32_rndscalesh_mask_round(A, B, C, D, E, F) __builtin_ia32_rndscalesh_mask_round(A, B, 123, D, E, 8) > #define __builtin_ia32_fpclassph512_mask(A, D, C) __builtin_ia32_fpclassph512_mask(A, 1, C) > #define __builtin_ia32_fpclasssh_mask(A, D, U) __builtin_ia32_fpclasssh_mask(A, 1, U) > #define __builtin_ia32_getexpph512_mask(A, B, C, D) __builtin_ia32_getexpph512_mask(A, B, C, 8) > @@ -738,8 +738,8 @@ > #define __builtin_ia32_getmantsh_mask_round(A, B, C, W, U, D) __builtin_ia32_getmantsh_mask_round(A, B, 1, W, U, 4) > > /* avx512fp16vlintrin.h */ > -#define __builtin_ia32_vcmpph_v8hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v8hf_mask(A, B, 1, D) > -#define __builtin_ia32_vcmpph_v16hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v16hf_mask(A, B, 1, D) > +#define __builtin_ia32_cmpph128_mask(A, B, C, D) __builtin_ia32_cmpph128_mask(A, B, 1, D) > +#define __builtin_ia32_cmpph256_mask(A, B, C, D) __builtin_ia32_cmpph256_mask(A, B, 1, D) > #define __builtin_ia32_fpclassph256_mask(A, D, C) __builtin_ia32_fpclassph256_mask(A, 1, C) > #define __builtin_ia32_fpclassph128_mask(A, D, C) __builtin_ia32_fpclassph128_mask(A, 1, C) > #define __builtin_ia32_getmantph256_mask(A, E, C, D) __builtin_ia32_getmantph256_mask(A, 1, C, D) > diff --git a/gcc/testsuite/gcc.target/i386/sse-23.c b/gcc/testsuite/gcc.target/i386/sse-23.c > index b3f07587acb..8b600282c67 100644 > --- a/gcc/testsuite/gcc.target/i386/sse-23.c > +++ b/gcc/testsuite/gcc.target/i386/sse-23.c > @@ -704,33 +704,33 @@ > #define __builtin_ia32_vpshld_v2di_mask(A, B, C, D, E) __builtin_ia32_vpshld_v2di_mask(A, B, 1, D, E) > > /* avx512fp16intrin.h */ > -#define __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vcmpph_v32hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v32hf_mask(A, B, 1, D) > -#define __builtin_ia32_vcmpph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpph_v32hf_mask_round(A, B, 1, D, 8) > -#define __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, 1, D, 8) > -#define __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, D) __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, 8) > -#define __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, E) __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, 8) > -#define __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, 8) > -#define __builtin_ia32_vreduceph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vreduceph_v32hf_mask_round(A, 123, C, D, 8) > -#define __builtin_ia32_vreduceph_v8hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v8hf_mask(A, 123, C, D) > -#define __builtin_ia32_vreduceph_v16hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v16hf_mask(A, 123, C, D) > -#define __builtin_ia32_vreducesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vreducesh_v8hf_mask_round(A, B, 123, D, E, 8) > -#define __builtin_ia32_vrndscaleph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vrndscaleph_v32hf_mask_round(A, 123, C, D, 8) > -#define __builtin_ia32_vrndscaleph_v8hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v8hf_mask(A, 123, C, D) > -#define __builtin_ia32_vrndscaleph_v16hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v16hf_mask(A, 123, C, D) > -#define __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, 123, D, E, 8) > +#define __builtin_ia32_addph512_mask_round(A, B, C, D, E) __builtin_ia32_addph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_subph512_mask_round(A, B, C, D, E) __builtin_ia32_subph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_mulph512_mask_round(A, B, C, D, E) __builtin_ia32_mulph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_divph512_mask_round(A, B, C, D, E) __builtin_ia32_divph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_addsh_mask_round(A, B, C, D, E) __builtin_ia32_addsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_subsh_mask_round(A, B, C, D, E) __builtin_ia32_subsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_mulsh_mask_round(A, B, C, D, E) __builtin_ia32_mulsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_divsh_mask_round(A, B, C, D, E) __builtin_ia32_divsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_maxph512_mask_round(A, B, C, D, E) __builtin_ia32_maxph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_minph512_mask_round(A, B, C, D, E) __builtin_ia32_minph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_maxsh_mask_round(A, B, C, D, E) __builtin_ia32_maxsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_minsh_mask_round(A, B, C, D, E) __builtin_ia32_minsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_cmpph512_mask(A, B, C, D) __builtin_ia32_cmpph512_mask(A, B, 1, D) > +#define __builtin_ia32_cmpph512_mask_round(A, B, C, D, E) __builtin_ia32_cmpph512_mask_round(A, B, 1, D, 8) > +#define __builtin_ia32_cmpsh_mask_round(A, B, C, D, E) __builtin_ia32_cmpsh_mask_round(A, B, 1, D, 8) > +#define __builtin_ia32_sqrtph512_mask_round(C, A, B, D) __builtin_ia32_sqrtph512_mask_round(C, A, B, 8) > +#define __builtin_ia32_sqrtsh_mask_round(D, C, A, B, E) __builtin_ia32_sqrtsh_mask_round(D, C, A, B, 8) > +#define __builtin_ia32_scalefph512_mask_round(A, B, C, D, E) __builtin_ia32_scalefph512_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_scalefsh_mask_round(A, B, C, D, E) __builtin_ia32_scalefsh_mask_round(A, B, C, D, 8) > +#define __builtin_ia32_reduceph512_mask_round(A, B, C, D, E) __builtin_ia32_reduceph512_mask_round(A, 123, C, D, 8) > +#define __builtin_ia32_reduceph128_mask(A, B, C, D) __builtin_ia32_reduceph128_mask(A, 123, C, D) > +#define __builtin_ia32_reduceph256_mask(A, B, C, D) __builtin_ia32_reduceph256_mask(A, 123, C, D) > +#define __builtin_ia32_reducesh_mask_round(A, B, C, D, E, F) __builtin_ia32_reducesh_mask_round(A, B, 123, D, E, 8) > +#define __builtin_ia32_rndscaleph512_mask_round(A, B, C, D, E) __builtin_ia32_rndscaleph512_mask_round(A, 123, C, D, 8) > +#define __builtin_ia32_rndscaleph128_mask(A, B, C, D) __builtin_ia32_rndscaleph128_mask(A, 123, C, D) > +#define __builtin_ia32_rndscaleph256_mask(A, B, C, D) __builtin_ia32_rndscaleph256_mask(A, 123, C, D) > +#define __builtin_ia32_rndscalesh_mask_round(A, B, C, D, E, F) __builtin_ia32_rndscalesh_mask_round(A, B, 123, D, E, 8) > #define __builtin_ia32_fpclassph512_mask(A, D, C) __builtin_ia32_fpclassph512_mask(A, 1, C) > #define __builtin_ia32_fpclasssh_mask(A, D, U) __builtin_ia32_fpclasssh_mask(A, 1, U) > #define __builtin_ia32_getexpph512_mask(A, B, C, D) __builtin_ia32_getexpph512_mask(A, B, C, 8) > @@ -739,8 +739,8 @@ > #define __builtin_ia32_getmantsh_mask_round(A, B, C, W, U, D) __builtin_ia32_getmantsh_mask_round(A, B, 1, W, U, 4) > > /* avx512fp16vlintrin.h */ > -#define __builtin_ia32_vcmpph_v8hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v8hf_mask(A, B, 1, D) > -#define __builtin_ia32_vcmpph_v16hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v16hf_mask(A, B, 1, D) > +#define __builtin_ia32_cmpph128_mask(A, B, C, D) __builtin_ia32_cmpph128_mask(A, B, 1, D) > +#define __builtin_ia32_cmpph256_mask(A, B, C, D) __builtin_ia32_cmpph256_mask(A, B, 1, D) > #define __builtin_ia32_fpclassph256_mask(A, D, C) __builtin_ia32_fpclassph256_mask(A, 1, C) > #define __builtin_ia32_fpclassph128_mask(A, D, C) __builtin_ia32_fpclassph128_mask(A, 1, C) > #define __builtin_ia32_getmantph256_mask(A, E, C, D) __builtin_ia32_getmantph256_mask(A, 1, C, D) > -- > 2.18.1 >
On Wed, Sep 15, 2021 at 5:52 PM Hongyu Wang <wwwhhhyyy333@gmail.com> wrote: > > Sorry for missing that part. > > Updated patch. Ok, thanks. > > Hongtao Liu via Gcc-patches <gcc-patches@gcc.gnu.org> 于2021年9月15日周三 下午5:16写道: > > > > On Wed, Sep 15, 2021 at 5:08 PM Hongyu Wang via Gcc-patches > > <gcc-patches@gcc.gnu.org> wrote: > > > > > > Hi, > > > > > > For AVX512FP16 builtins, they all contain format like vaddph_v8hf, > > > while AVX512F builtins use addps128 which succeeded SSE/AVX style. > > > Adjust AVX512FP16 builtins to match such format. > > > > > > Bootstraped/regtested on x86-64-*-linux, ok for master? > > > > > > gcc/ChangeLog: > > > > > > * config/i386/avx512fp16intrin.h: Adjust all builtin calls. > > > * config/i386/avx512fp16vlintrin.h: Likewise. > > > * config/i386/i386-builtin.def: Adjust builtin name to match > > > AVX512F style. > > > > > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_vmsmaxv8hf3_mask, "__builtin_ia32_maxsh_mask", > > IX86_BUILTIN_VMAXSH_V8HF_MASK, UNKNOWN, (int) > > V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_vmsminv8hf3_mask, "__builtin_ia32_minsh_mask", > > IX86_BUILTIN_VMINSH_V8HF_MASK, UNKNOWN, (int) > > V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_cmpv8hf3_mask, "__builtin_ia32_cmpph128_mask", > > IX86_BUILTIN_VCMPPH_V8HF_MASK, UNKNOWN, (int) > > UQI_FTYPE_V8HF_V8HF_INT_UQI) > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512vl_cmpv16hf3_mask, "__builtin_ia32_cmpph256_mask", > > IX86_BUILTIN_VCMPPH_V16HF_MASK, UNKNOWN, (int) > > UHI_FTYPE_V16HF_V16HF_INT_UHI) > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512bw_cmpv32hf3_mask, "__builtin_ia32_cmpph512_mask", > > IX86_BUILTIN_VCMPPH_V32HF_MASK, UNKNOWN, (int) > > USI_FTYPE_V32HF_V32HF_INT_USI) > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_sqrtv8hf2_mask, "__builtin_ia32_sqrtph128_mask", > > IX86_BUILTIN_VSQRTPH_V8HF_MASK, UNKNOWN, (int) > > V8HF_FTYPE_V8HF_V8HF_UQI) > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_sqrtv16hf2_mask, "__builtin_ia32_sqrtph256_mask", > > IX86_BUILTIN_VSQRTPH_V16HF_MASK, UNKNOWN, (int) > > V16HF_FTYPE_V16HF_V16HF_UHI) > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_rsqrtv8hf2_mask, "__builtin_ia32_rsqrtph128_mask", > > IX86_BUILTIN_VRSQRTPH_V8HF_MASK, UNKNOWN, (int) > > V8HF_FTYPE_V8HF_V8HF_UQI) > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_rsqrtv16hf2_mask, > > "__builtin_ia32_rsqrtph256_mask", IX86_BUILTIN_VRSQRTPH_V16HF_MASK, > > UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_rsqrtv32hf2_mask, > > "__builtin_ia32_rsqrtph512_mask", IX86_BUILTIN_VRSQRTPH_V32HF_MASK, > > UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_vmrsqrtv8hf2_mask, "__builtin_ia32_rsqrtsh_mask", > > IX86_BUILTIN_VRSQRTSH_V8HF_MASK, UNKNOWN, (int) > > V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_rcpv8hf2_mask, "__builtin_ia32_rcpph128_mask", > > IX86_BUILTIN_VRCPPH_V8HF_MASK, UNKNOWN, (int) > > V8HF_FTYPE_V8HF_V8HF_UQI) > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_rcpv16hf2_mask, "__builtin_ia32_rcpph256_mask", > > IX86_BUILTIN_VRCPPH_V16HF_MASK, UNKNOWN, (int) > > V16HF_FTYPE_V16HF_V16HF_UHI) > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_rcpv32hf2_mask, "__builtin_ia32_rcpph512_mask", > > IX86_BUILTIN_VRCPPH_V32HF_MASK, UNKNOWN, (int) > > V32HF_FTYPE_V32HF_V32HF_USI) > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_vmrcpv8hf2_mask, "__builtin_ia32_rcpsh_mask", > > IX86_BUILTIN_VRCPSH_V8HF_MASK, UNKNOWN, (int) > > V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512fp16_scalefv8hf_mask, > > "__builtin_ia32_scalefph128_mask", IX86_BUILTIN_VSCALEFPH_V8HF_MASK, > > UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, > > CODE_FOR_avx512vl_scalefv16hf_mask, "__builtin_ia32_scalefph256_mask", > > IX86_BUILTIN_VSCALEFPH_V16HF_MASK, UNKNOWN, (int) > > V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > Could you also change IX86_BUILTIN_VSCALEFPH_V16HF_MASK > > IX86_BUILTIN_VSCALEFPH256_MASK, similar for others. > > > > > gcc/testsuite/ChangeLog: > > > > > > * gcc.target/i386/avx-1.c: Adjust builtin macros. > > > * gcc.target/i386/sse-13.c: Likewise. > > > * gcc.target/i386/sse-23.c: Likewise. > > > --- > > > gcc/config/i386/avx512fp16intrin.h | 1236 ++++++++++++------------ > > > gcc/config/i386/avx512fp16vlintrin.h | 454 ++++----- > > > gcc/config/i386/i386-builtin.def | 130 +-- > > > gcc/testsuite/gcc.target/i386/avx-1.c | 58 +- > > > gcc/testsuite/gcc.target/i386/sse-13.c | 58 +- > > > gcc/testsuite/gcc.target/i386/sse-23.c | 58 +- > > > 6 files changed, 997 insertions(+), 997 deletions(-) > > > > > > diff --git a/gcc/config/i386/avx512fp16intrin.h b/gcc/config/i386/avx512fp16intrin.h > > > index 2fbfc140c44..5d66ca5c820 100644 > > > --- a/gcc/config/i386/avx512fp16intrin.h > > > +++ b/gcc/config/i386/avx512fp16intrin.h > > > @@ -229,15 +229,15 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_add_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > > > { > > > - return __builtin_ia32_vaddph_v32hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_addph512_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_add_ph (__mmask32 __A, __m512h __B, __m512h __C) > > > { > > > - return __builtin_ia32_vaddph_v32hf_mask (__B, __C, > > > - _mm512_setzero_ph (), __A); > > > + return __builtin_ia32_addph512_mask (__B, __C, > > > + _mm512_setzero_ph (), __A); > > > } > > > > > > extern __inline __m512h > > > @@ -251,15 +251,15 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_sub_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > > > { > > > - return __builtin_ia32_vsubph_v32hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_subph512_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_sub_ph (__mmask32 __A, __m512h __B, __m512h __C) > > > { > > > - return __builtin_ia32_vsubph_v32hf_mask (__B, __C, > > > - _mm512_setzero_ph (), __A); > > > + return __builtin_ia32_subph512_mask (__B, __C, > > > + _mm512_setzero_ph (), __A); > > > } > > > > > > extern __inline __m512h > > > @@ -273,15 +273,15 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_mul_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > > > { > > > - return __builtin_ia32_vmulph_v32hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_mulph512_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_mul_ph (__mmask32 __A, __m512h __B, __m512h __C) > > > { > > > - return __builtin_ia32_vmulph_v32hf_mask (__B, __C, > > > - _mm512_setzero_ph (), __A); > > > + return __builtin_ia32_mulph512_mask (__B, __C, > > > + _mm512_setzero_ph (), __A); > > > } > > > > > > extern __inline __m512h > > > @@ -295,15 +295,15 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_div_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > > > { > > > - return __builtin_ia32_vdivph_v32hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_divph512_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_div_ph (__mmask32 __A, __m512h __B, __m512h __C) > > > { > > > - return __builtin_ia32_vdivph_v32hf_mask (__B, __C, > > > - _mm512_setzero_ph (), __A); > > > + return __builtin_ia32_divph512_mask (__B, __C, > > > + _mm512_setzero_ph (), __A); > > > } > > > > > > #ifdef __OPTIMIZE__ > > > @@ -311,9 +311,9 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_add_round_ph (__m512h __A, __m512h __B, const int __C) > > > { > > > - return __builtin_ia32_vaddph_v32hf_mask_round (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, __C); > > > + return __builtin_ia32_addph512_mask_round (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, __C); > > > } > > > > > > extern __inline __m512h > > > @@ -321,7 +321,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_add_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > > > __m512h __D, const int __E) > > > { > > > - return __builtin_ia32_vaddph_v32hf_mask_round (__C, __D, __A, __B, __E); > > > + return __builtin_ia32_addph512_mask_round (__C, __D, __A, __B, __E); > > > } > > > > > > extern __inline __m512h > > > @@ -329,18 +329,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_add_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vaddph_v32hf_mask_round (__B, __C, > > > - _mm512_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_addph512_mask_round (__B, __C, > > > + _mm512_setzero_ph (), > > > + __A, __D); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_sub_round_ph (__m512h __A, __m512h __B, const int __C) > > > { > > > - return __builtin_ia32_vsubph_v32hf_mask_round (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, __C); > > > + return __builtin_ia32_subph512_mask_round (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, __C); > > > } > > > > > > extern __inline __m512h > > > @@ -348,7 +348,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_sub_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > > > __m512h __D, const int __E) > > > { > > > - return __builtin_ia32_vsubph_v32hf_mask_round (__C, __D, __A, __B, __E); > > > + return __builtin_ia32_subph512_mask_round (__C, __D, __A, __B, __E); > > > } > > > > > > extern __inline __m512h > > > @@ -356,18 +356,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_sub_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vsubph_v32hf_mask_round (__B, __C, > > > - _mm512_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_subph512_mask_round (__B, __C, > > > + _mm512_setzero_ph (), > > > + __A, __D); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mul_round_ph (__m512h __A, __m512h __B, const int __C) > > > { > > > - return __builtin_ia32_vmulph_v32hf_mask_round (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, __C); > > > + return __builtin_ia32_mulph512_mask_round (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, __C); > > > } > > > > > > extern __inline __m512h > > > @@ -375,7 +375,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_mul_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > > > __m512h __D, const int __E) > > > { > > > - return __builtin_ia32_vmulph_v32hf_mask_round (__C, __D, __A, __B, __E); > > > + return __builtin_ia32_mulph512_mask_round (__C, __D, __A, __B, __E); > > > } > > > > > > extern __inline __m512h > > > @@ -383,18 +383,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_mul_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vmulph_v32hf_mask_round (__B, __C, > > > - _mm512_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_mulph512_mask_round (__B, __C, > > > + _mm512_setzero_ph (), > > > + __A, __D); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_div_round_ph (__m512h __A, __m512h __B, const int __C) > > > { > > > - return __builtin_ia32_vdivph_v32hf_mask_round (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, __C); > > > + return __builtin_ia32_divph512_mask_round (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, __C); > > > } > > > > > > extern __inline __m512h > > > @@ -402,7 +402,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_div_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > > > __m512h __D, const int __E) > > > { > > > - return __builtin_ia32_vdivph_v32hf_mask_round (__C, __D, __A, __B, __E); > > > + return __builtin_ia32_divph512_mask_round (__C, __D, __A, __B, __E); > > > } > > > > > > extern __inline __m512h > > > @@ -410,67 +410,67 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_div_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vdivph_v32hf_mask_round (__B, __C, > > > - _mm512_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_divph512_mask_round (__B, __C, > > > + _mm512_setzero_ph (), > > > + __A, __D); > > > } > > > #else > > > #define _mm512_add_round_ph(A, B, C) \ > > > - ((__m512h)__builtin_ia32_vaddph_v32hf_mask_round((A), (B), \ > > > - _mm512_setzero_ph (),\ > > > - (__mmask32)-1, (C))) > > > + ((__m512h)__builtin_ia32_addph512_mask_round((A), (B), \ > > > + _mm512_setzero_ph (), \ > > > + (__mmask32)-1, (C))) > > > > > > -#define _mm512_mask_add_round_ph(A, B, C, D, E) \ > > > - ((__m512h)__builtin_ia32_vaddph_v32hf_mask_round((C), (D), (A), (B), (E))) > > > +#define _mm512_mask_add_round_ph(A, B, C, D, E) \ > > > + ((__m512h)__builtin_ia32_addph512_mask_round((C), (D), (A), (B), (E))) > > > > > > #define _mm512_maskz_add_round_ph(A, B, C, D) \ > > > - ((__m512h)__builtin_ia32_vaddph_v32hf_mask_round((B), (C), \ > > > - _mm512_setzero_ph (),\ > > > - (A), (D))) > > > + ((__m512h)__builtin_ia32_addph512_mask_round((B), (C), \ > > > + _mm512_setzero_ph (), \ > > > + (A), (D))) > > > > > > #define _mm512_sub_round_ph(A, B, C) \ > > > - ((__m512h)__builtin_ia32_vsubph_v32hf_mask_round((A), (B), \ > > > - _mm512_setzero_ph (),\ > > > - (__mmask32)-1, (C))) > > > + ((__m512h)__builtin_ia32_subph512_mask_round((A), (B), \ > > > + _mm512_setzero_ph (), \ > > > + (__mmask32)-1, (C))) > > > > > > -#define _mm512_mask_sub_round_ph(A, B, C, D, E) \ > > > - ((__m512h)__builtin_ia32_vsubph_v32hf_mask_round((C), (D), (A), (B), (E))) > > > +#define _mm512_mask_sub_round_ph(A, B, C, D, E) \ > > > + ((__m512h)__builtin_ia32_subph512_mask_round((C), (D), (A), (B), (E))) > > > > > > #define _mm512_maskz_sub_round_ph(A, B, C, D) \ > > > - ((__m512h)__builtin_ia32_vsubph_v32hf_mask_round((B), (C), \ > > > - _mm512_setzero_ph (),\ > > > - (A), (D))) > > > + ((__m512h)__builtin_ia32_subph512_mask_round((B), (C), \ > > > + _mm512_setzero_ph (), \ > > > + (A), (D))) > > > > > > #define _mm512_mul_round_ph(A, B, C) \ > > > - ((__m512h)__builtin_ia32_vmulph_v32hf_mask_round((A), (B), \ > > > - _mm512_setzero_ph (),\ > > > - (__mmask32)-1, (C))) > > > + ((__m512h)__builtin_ia32_mulph512_mask_round((A), (B), \ > > > + _mm512_setzero_ph (), \ > > > + (__mmask32)-1, (C))) > > > > > > -#define _mm512_mask_mul_round_ph(A, B, C, D, E) \ > > > - ((__m512h)__builtin_ia32_vmulph_v32hf_mask_round((C), (D), (A), (B), (E))) > > > +#define _mm512_mask_mul_round_ph(A, B, C, D, E) \ > > > + ((__m512h)__builtin_ia32_mulph512_mask_round((C), (D), (A), (B), (E))) > > > > > > #define _mm512_maskz_mul_round_ph(A, B, C, D) \ > > > - ((__m512h)__builtin_ia32_vmulph_v32hf_mask_round((B), (C), \ > > > - _mm512_setzero_ph (),\ > > > - (A), (D))) > > > + ((__m512h)__builtin_ia32_mulph512_mask_round((B), (C), \ > > > + _mm512_setzero_ph (), \ > > > + (A), (D))) > > > > > > #define _mm512_div_round_ph(A, B, C) \ > > > - ((__m512h)__builtin_ia32_vdivph_v32hf_mask_round((A), (B), \ > > > - _mm512_setzero_ph (),\ > > > - (__mmask32)-1, (C))) > > > + ((__m512h)__builtin_ia32_divph512_mask_round((A), (B), \ > > > + _mm512_setzero_ph (), \ > > > + (__mmask32)-1, (C))) > > > > > > -#define _mm512_mask_div_round_ph(A, B, C, D, E) \ > > > - ((__m512h)__builtin_ia32_vdivph_v32hf_mask_round((C), (D), (A), (B), (E))) > > > +#define _mm512_mask_div_round_ph(A, B, C, D, E) \ > > > + ((__m512h)__builtin_ia32_divph512_mask_round((C), (D), (A), (B), (E))) > > > > > > #define _mm512_maskz_div_round_ph(A, B, C, D) \ > > > - ((__m512h)__builtin_ia32_vdivph_v32hf_mask_round((B), (C), \ > > > - _mm512_setzero_ph (),\ > > > - (A), (D))) > > > + ((__m512h)__builtin_ia32_divph512_mask_round((B), (C), \ > > > + _mm512_setzero_ph (), \ > > > + (A), (D))) > > > #endif /* __OPTIMIZE__ */ > > > > > > /* Intrinsics of v[add,sub,mul,div]sh. */ > > > extern __inline __m128h > > > -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_add_sh (__m128h __A, __m128h __B) > > > { > > > __A[0] += __B[0]; > > > @@ -481,15 +481,15 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_add_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vaddsh_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_addsh_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_add_sh (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vaddsh_v8hf_mask (__B, __C, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_addsh_mask (__B, __C, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > extern __inline __m128h > > > @@ -504,15 +504,15 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_sub_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vsubsh_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_subsh_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_sub_sh (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vsubsh_v8hf_mask (__B, __C, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_subsh_mask (__B, __C, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > extern __inline __m128h > > > @@ -527,14 +527,14 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_mul_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vmulsh_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_mulsh_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_mul_sh (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vmulsh_v8hf_mask (__B, __C, _mm_setzero_ph (), __A); > > > + return __builtin_ia32_mulsh_mask (__B, __C, _mm_setzero_ph (), __A); > > > } > > > > > > extern __inline __m128h > > > @@ -549,15 +549,15 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_div_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vdivsh_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_divsh_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_div_sh (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vdivsh_v8hf_mask (__B, __C, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_divsh_mask (__B, __C, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > #ifdef __OPTIMIZE__ > > > @@ -565,9 +565,9 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_add_round_sh (__m128h __A, __m128h __B, const int __C) > > > { > > > - return __builtin_ia32_vaddsh_v8hf_mask_round (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, __C); > > > + return __builtin_ia32_addsh_mask_round (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, __C); > > > } > > > > > > extern __inline __m128h > > > @@ -575,7 +575,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_add_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > > > __m128h __D, const int __E) > > > { > > > - return __builtin_ia32_vaddsh_v8hf_mask_round (__C, __D, __A, __B, __E); > > > + return __builtin_ia32_addsh_mask_round (__C, __D, __A, __B, __E); > > > } > > > > > > extern __inline __m128h > > > @@ -583,18 +583,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_add_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vaddsh_v8hf_mask_round (__B, __C, > > > - _mm_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_addsh_mask_round (__B, __C, > > > + _mm_setzero_ph (), > > > + __A, __D); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_sub_round_sh (__m128h __A, __m128h __B, const int __C) > > > { > > > - return __builtin_ia32_vsubsh_v8hf_mask_round (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, __C); > > > + return __builtin_ia32_subsh_mask_round (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, __C); > > > } > > > > > > extern __inline __m128h > > > @@ -602,7 +602,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_sub_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > > > __m128h __D, const int __E) > > > { > > > - return __builtin_ia32_vsubsh_v8hf_mask_round (__C, __D, __A, __B, __E); > > > + return __builtin_ia32_subsh_mask_round (__C, __D, __A, __B, __E); > > > } > > > > > > extern __inline __m128h > > > @@ -610,18 +610,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_sub_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vsubsh_v8hf_mask_round (__B, __C, > > > - _mm_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_subsh_mask_round (__B, __C, > > > + _mm_setzero_ph (), > > > + __A, __D); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mul_round_sh (__m128h __A, __m128h __B, const int __C) > > > { > > > - return __builtin_ia32_vmulsh_v8hf_mask_round (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, __C); > > > + return __builtin_ia32_mulsh_mask_round (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, __C); > > > } > > > > > > extern __inline __m128h > > > @@ -629,7 +629,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_mul_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > > > __m128h __D, const int __E) > > > { > > > - return __builtin_ia32_vmulsh_v8hf_mask_round (__C, __D, __A, __B, __E); > > > + return __builtin_ia32_mulsh_mask_round (__C, __D, __A, __B, __E); > > > } > > > > > > extern __inline __m128h > > > @@ -637,18 +637,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_mul_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vmulsh_v8hf_mask_round (__B, __C, > > > - _mm_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_mulsh_mask_round (__B, __C, > > > + _mm_setzero_ph (), > > > + __A, __D); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_div_round_sh (__m128h __A, __m128h __B, const int __C) > > > { > > > - return __builtin_ia32_vdivsh_v8hf_mask_round (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, __C); > > > + return __builtin_ia32_divsh_mask_round (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, __C); > > > } > > > > > > extern __inline __m128h > > > @@ -656,7 +656,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_div_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > > > __m128h __D, const int __E) > > > { > > > - return __builtin_ia32_vdivsh_v8hf_mask_round (__C, __D, __A, __B, __E); > > > + return __builtin_ia32_divsh_mask_round (__C, __D, __A, __B, __E); > > > } > > > > > > extern __inline __m128h > > > @@ -664,62 +664,62 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_div_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vdivsh_v8hf_mask_round (__B, __C, > > > - _mm_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_divsh_mask_round (__B, __C, > > > + _mm_setzero_ph (), > > > + __A, __D); > > > } > > > #else > > > #define _mm_add_round_sh(A, B, C) \ > > > - ((__m128h)__builtin_ia32_vaddsh_v8hf_mask_round ((A), (B), \ > > > - _mm_setzero_ph (), \ > > > - (__mmask8)-1, (C))) > > > + ((__m128h)__builtin_ia32_addsh_mask_round ((A), (B), \ > > > + _mm_setzero_ph (), \ > > > + (__mmask8)-1, (C))) > > > > > > #define _mm_mask_add_round_sh(A, B, C, D, E) \ > > > - ((__m128h)__builtin_ia32_vaddsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > > > + ((__m128h)__builtin_ia32_addsh_mask_round ((C), (D), (A), (B), (E))) > > > > > > -#define _mm_maskz_add_round_sh(A, B, C, D) \ > > > - ((__m128h)__builtin_ia32_vaddsh_v8hf_mask_round ((B), (C), \ > > > - _mm_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm_maskz_add_round_sh(A, B, C, D) \ > > > + ((__m128h)__builtin_ia32_addsh_mask_round ((B), (C), \ > > > + _mm_setzero_ph (), \ > > > + (A), (D))) > > > > > > #define _mm_sub_round_sh(A, B, C) \ > > > - ((__m128h)__builtin_ia32_vsubsh_v8hf_mask_round ((A), (B), \ > > > - _mm_setzero_ph (), \ > > > - (__mmask8)-1, (C))) > > > + ((__m128h)__builtin_ia32_subsh_mask_round ((A), (B), \ > > > + _mm_setzero_ph (), \ > > > + (__mmask8)-1, (C))) > > > > > > #define _mm_mask_sub_round_sh(A, B, C, D, E) \ > > > - ((__m128h)__builtin_ia32_vsubsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > > > + ((__m128h)__builtin_ia32_subsh_mask_round ((C), (D), (A), (B), (E))) > > > > > > -#define _mm_maskz_sub_round_sh(A, B, C, D) \ > > > - ((__m128h)__builtin_ia32_vsubsh_v8hf_mask_round ((B), (C), \ > > > - _mm_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm_maskz_sub_round_sh(A, B, C, D) \ > > > + ((__m128h)__builtin_ia32_subsh_mask_round ((B), (C), \ > > > + _mm_setzero_ph (), \ > > > + (A), (D))) > > > > > > #define _mm_mul_round_sh(A, B, C) \ > > > - ((__m128h)__builtin_ia32_vmulsh_v8hf_mask_round ((A), (B), \ > > > - _mm_setzero_ph (), \ > > > - (__mmask8)-1, (C))) > > > + ((__m128h)__builtin_ia32_mulsh_mask_round ((A), (B), \ > > > + _mm_setzero_ph (), \ > > > + (__mmask8)-1, (C))) > > > > > > #define _mm_mask_mul_round_sh(A, B, C, D, E) \ > > > - ((__m128h)__builtin_ia32_vmulsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > > > + ((__m128h)__builtin_ia32_mulsh_mask_round ((C), (D), (A), (B), (E))) > > > > > > -#define _mm_maskz_mul_round_sh(A, B, C, D) \ > > > - ((__m128h)__builtin_ia32_vmulsh_v8hf_mask_round ((B), (C), \ > > > - _mm_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm_maskz_mul_round_sh(A, B, C, D) \ > > > + ((__m128h)__builtin_ia32_mulsh_mask_round ((B), (C), \ > > > + _mm_setzero_ph (), \ > > > + (A), (D))) > > > > > > #define _mm_div_round_sh(A, B, C) \ > > > - ((__m128h)__builtin_ia32_vdivsh_v8hf_mask_round ((A), (B), \ > > > - _mm_setzero_ph (), \ > > > - (__mmask8)-1, (C))) > > > + ((__m128h)__builtin_ia32_divsh_mask_round ((A), (B), \ > > > + _mm_setzero_ph (), \ > > > + (__mmask8)-1, (C))) > > > > > > #define _mm_mask_div_round_sh(A, B, C, D, E) \ > > > - ((__m128h)__builtin_ia32_vdivsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > > > + ((__m128h)__builtin_ia32_divsh_mask_round ((C), (D), (A), (B), (E))) > > > > > > -#define _mm_maskz_div_round_sh(A, B, C, D) \ > > > - ((__m128h)__builtin_ia32_vdivsh_v8hf_mask_round ((B), (C), \ > > > - _mm_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm_maskz_div_round_sh(A, B, C, D) \ > > > + ((__m128h)__builtin_ia32_divsh_mask_round ((B), (C), \ > > > + _mm_setzero_ph (), \ > > > + (A), (D))) > > > #endif /* __OPTIMIZE__ */ > > > > > > /* Intrinsic vmaxph vminph. */ > > > @@ -727,48 +727,48 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_max_ph (__m512h __A, __m512h __B) > > > { > > > - return __builtin_ia32_vmaxph_v32hf_mask (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1); > > > + return __builtin_ia32_maxph512_mask (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_max_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > > > { > > > - return __builtin_ia32_vmaxph_v32hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_maxph512_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_max_ph (__mmask32 __A, __m512h __B, __m512h __C) > > > { > > > - return __builtin_ia32_vmaxph_v32hf_mask (__B, __C, > > > - _mm512_setzero_ph (), __A); > > > + return __builtin_ia32_maxph512_mask (__B, __C, > > > + _mm512_setzero_ph (), __A); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_min_ph (__m512h __A, __m512h __B) > > > { > > > - return __builtin_ia32_vminph_v32hf_mask (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1); > > > + return __builtin_ia32_minph512_mask (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_min_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > > > { > > > - return __builtin_ia32_vminph_v32hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_minph512_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_min_ph (__mmask32 __A, __m512h __B, __m512h __C) > > > { > > > - return __builtin_ia32_vminph_v32hf_mask (__B, __C, > > > - _mm512_setzero_ph (), __A); > > > + return __builtin_ia32_minph512_mask (__B, __C, > > > + _mm512_setzero_ph (), __A); > > > } > > > > > > #ifdef __OPTIMIZE__ > > > @@ -776,9 +776,9 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_max_round_ph (__m512h __A, __m512h __B, const int __C) > > > { > > > - return __builtin_ia32_vmaxph_v32hf_mask_round (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, __C); > > > + return __builtin_ia32_maxph512_mask_round (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, __C); > > > } > > > > > > extern __inline __m512h > > > @@ -786,7 +786,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_max_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > > > __m512h __D, const int __E) > > > { > > > - return __builtin_ia32_vmaxph_v32hf_mask_round (__C, __D, __A, __B, __E); > > > + return __builtin_ia32_maxph512_mask_round (__C, __D, __A, __B, __E); > > > } > > > > > > extern __inline __m512h > > > @@ -794,18 +794,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_max_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vmaxph_v32hf_mask_round (__B, __C, > > > - _mm512_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_maxph512_mask_round (__B, __C, > > > + _mm512_setzero_ph (), > > > + __A, __D); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_min_round_ph (__m512h __A, __m512h __B, const int __C) > > > { > > > - return __builtin_ia32_vminph_v32hf_mask_round (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, __C); > > > + return __builtin_ia32_minph512_mask_round (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, __C); > > > } > > > > > > extern __inline __m512h > > > @@ -813,7 +813,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_min_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > > > __m512h __D, const int __E) > > > { > > > - return __builtin_ia32_vminph_v32hf_mask_round (__C, __D, __A, __B, __E); > > > + return __builtin_ia32_minph512_mask_round (__C, __D, __A, __B, __E); > > > } > > > > > > extern __inline __m512h > > > @@ -821,37 +821,37 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_min_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vminph_v32hf_mask_round (__B, __C, > > > - _mm512_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_minph512_mask_round (__B, __C, > > > + _mm512_setzero_ph (), > > > + __A, __D); > > > } > > > > > > #else > > > -#define _mm512_max_round_ph(A, B, C) \ > > > - (__builtin_ia32_vmaxph_v32hf_mask_round ((A), (B), \ > > > - _mm512_setzero_ph (), \ > > > - (__mmask32)-1, (C))) > > > +#define _mm512_max_round_ph(A, B, C) \ > > > + (__builtin_ia32_maxph512_mask_round ((A), (B), \ > > > + _mm512_setzero_ph (), \ > > > + (__mmask32)-1, (C))) > > > > > > #define _mm512_mask_max_round_ph(A, B, C, D, E) \ > > > - (__builtin_ia32_vmaxph_v32hf_mask_round ((C), (D), (A), (B), (E))) > > > + (__builtin_ia32_maxph512_mask_round ((C), (D), (A), (B), (E))) > > > > > > -#define _mm512_maskz_max_round_ph(A, B, C, D) \ > > > - (__builtin_ia32_vmaxph_v32hf_mask_round ((B), (C), \ > > > - _mm512_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm512_maskz_max_round_ph(A, B, C, D) \ > > > + (__builtin_ia32_maxph512_mask_round ((B), (C), \ > > > + _mm512_setzero_ph (), \ > > > + (A), (D))) > > > > > > -#define _mm512_min_round_ph(A, B, C) \ > > > - (__builtin_ia32_vminph_v32hf_mask_round ((A), (B), \ > > > - _mm512_setzero_ph (), \ > > > - (__mmask32)-1, (C))) > > > +#define _mm512_min_round_ph(A, B, C) \ > > > + (__builtin_ia32_minph512_mask_round ((A), (B), \ > > > + _mm512_setzero_ph (), \ > > > + (__mmask32)-1, (C))) > > > > > > #define _mm512_mask_min_round_ph(A, B, C, D, E) \ > > > - (__builtin_ia32_vminph_v32hf_mask_round ((C), (D), (A), (B), (E))) > > > + (__builtin_ia32_minph512_mask_round ((C), (D), (A), (B), (E))) > > > > > > -#define _mm512_maskz_min_round_ph(A, B, C, D) \ > > > - (__builtin_ia32_vminph_v32hf_mask_round ((B), (C), \ > > > - _mm512_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm512_maskz_min_round_ph(A, B, C, D) \ > > > + (__builtin_ia32_minph512_mask_round ((B), (C), \ > > > + _mm512_setzero_ph (), \ > > > + (A), (D))) > > > #endif /* __OPTIMIZE__ */ > > > > > > /* Intrinsic vmaxsh vminsh. */ > > > @@ -867,15 +867,15 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_max_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vmaxsh_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_maxsh_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_max_sh (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vmaxsh_v8hf_mask (__B, __C, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_maxsh_mask (__B, __C, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > extern __inline __m128h > > > @@ -890,15 +890,15 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_min_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vminsh_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_minsh_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_min_sh (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vminsh_v8hf_mask (__B, __C, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_minsh_mask (__B, __C, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > #ifdef __OPTIMIZE__ > > > @@ -906,9 +906,9 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_max_round_sh (__m128h __A, __m128h __B, const int __C) > > > { > > > - return __builtin_ia32_vmaxsh_v8hf_mask_round (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, __C); > > > + return __builtin_ia32_maxsh_mask_round (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, __C); > > > } > > > > > > extern __inline __m128h > > > @@ -916,7 +916,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_max_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > > > __m128h __D, const int __E) > > > { > > > - return __builtin_ia32_vmaxsh_v8hf_mask_round (__C, __D, __A, __B, __E); > > > + return __builtin_ia32_maxsh_mask_round (__C, __D, __A, __B, __E); > > > } > > > > > > extern __inline __m128h > > > @@ -924,18 +924,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_max_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vmaxsh_v8hf_mask_round (__B, __C, > > > - _mm_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_maxsh_mask_round (__B, __C, > > > + _mm_setzero_ph (), > > > + __A, __D); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_min_round_sh (__m128h __A, __m128h __B, const int __C) > > > { > > > - return __builtin_ia32_vminsh_v8hf_mask_round (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, __C); > > > + return __builtin_ia32_minsh_mask_round (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, __C); > > > } > > > > > > extern __inline __m128h > > > @@ -943,7 +943,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_min_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > > > __m128h __D, const int __E) > > > { > > > - return __builtin_ia32_vminsh_v8hf_mask_round (__C, __D, __A, __B, __E); > > > + return __builtin_ia32_minsh_mask_round (__C, __D, __A, __B, __E); > > > } > > > > > > extern __inline __m128h > > > @@ -951,37 +951,37 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_min_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vminsh_v8hf_mask_round (__B, __C, > > > - _mm_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_minsh_mask_round (__B, __C, > > > + _mm_setzero_ph (), > > > + __A, __D); > > > } > > > > > > #else > > > -#define _mm_max_round_sh(A, B, C) \ > > > - (__builtin_ia32_vmaxsh_v8hf_mask_round ((A), (B), \ > > > - _mm_setzero_ph (), \ > > > - (__mmask8)-1, (C))) > > > +#define _mm_max_round_sh(A, B, C) \ > > > + (__builtin_ia32_maxsh_mask_round ((A), (B), \ > > > + _mm_setzero_ph (), \ > > > + (__mmask8)-1, (C))) > > > > > > -#define _mm_mask_max_round_sh(A, B, C, D, E) \ > > > - (__builtin_ia32_vmaxsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > > > +#define _mm_mask_max_round_sh(A, B, C, D, E) \ > > > + (__builtin_ia32_maxsh_mask_round ((C), (D), (A), (B), (E))) > > > > > > -#define _mm_maskz_max_round_sh(A, B, C, D) \ > > > - (__builtin_ia32_vmaxsh_v8hf_mask_round ((B), (C), \ > > > - _mm_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm_maskz_max_round_sh(A, B, C, D) \ > > > + (__builtin_ia32_maxsh_mask_round ((B), (C), \ > > > + _mm_setzero_ph (), \ > > > + (A), (D))) > > > > > > -#define _mm_min_round_sh(A, B, C) \ > > > - (__builtin_ia32_vminsh_v8hf_mask_round ((A), (B), \ > > > - _mm_setzero_ph (), \ > > > - (__mmask8)-1, (C))) > > > +#define _mm_min_round_sh(A, B, C) \ > > > + (__builtin_ia32_minsh_mask_round ((A), (B), \ > > > + _mm_setzero_ph (), \ > > > + (__mmask8)-1, (C))) > > > > > > -#define _mm_mask_min_round_sh(A, B, C, D, E) \ > > > - (__builtin_ia32_vminsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > > > +#define _mm_mask_min_round_sh(A, B, C, D, E) \ > > > + (__builtin_ia32_minsh_mask_round ((C), (D), (A), (B), (E))) > > > > > > -#define _mm_maskz_min_round_sh(A, B, C, D) \ > > > - (__builtin_ia32_vminsh_v8hf_mask_round ((B), (C), \ > > > - _mm_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm_maskz_min_round_sh(A, B, C, D) \ > > > + (__builtin_ia32_minsh_mask_round ((B), (C), \ > > > + _mm_setzero_ph (), \ > > > + (A), (D))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > @@ -991,8 +991,8 @@ extern __inline __mmask32 > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_cmp_ph_mask (__m512h __A, __m512h __B, const int __C) > > > { > > > - return (__mmask32) __builtin_ia32_vcmpph_v32hf_mask (__A, __B, __C, > > > - (__mmask32) -1); > > > + return (__mmask32) __builtin_ia32_cmpph512_mask (__A, __B, __C, > > > + (__mmask32) -1); > > > } > > > > > > extern __inline __mmask32 > > > @@ -1000,8 +1000,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_cmp_ph_mask (__mmask32 __A, __m512h __B, __m512h __C, > > > const int __D) > > > { > > > - return (__mmask32) __builtin_ia32_vcmpph_v32hf_mask (__B, __C, __D, > > > - __A); > > > + return (__mmask32) __builtin_ia32_cmpph512_mask (__B, __C, __D, > > > + __A); > > > } > > > > > > extern __inline __mmask32 > > > @@ -1009,9 +1009,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_cmp_round_ph_mask (__m512h __A, __m512h __B, const int __C, > > > const int __D) > > > { > > > - return (__mmask32) __builtin_ia32_vcmpph_v32hf_mask_round (__A, __B, > > > - __C, (__mmask32) -1, > > > - __D); > > > + return (__mmask32) __builtin_ia32_cmpph512_mask_round (__A, __B, > > > + __C, (__mmask32) -1, > > > + __D); > > > } > > > > > > extern __inline __mmask32 > > > @@ -1019,23 +1019,23 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_cmp_round_ph_mask (__mmask32 __A, __m512h __B, __m512h __C, > > > const int __D, const int __E) > > > { > > > - return (__mmask32) __builtin_ia32_vcmpph_v32hf_mask_round (__B, __C, > > > - __D, __A, > > > - __E); > > > + return (__mmask32) __builtin_ia32_cmpph512_mask_round (__B, __C, > > > + __D, __A, > > > + __E); > > > } > > > > > > #else > > > #define _mm512_cmp_ph_mask(A, B, C) \ > > > - (__builtin_ia32_vcmpph_v32hf_mask ((A), (B), (C), (-1))) > > > + (__builtin_ia32_cmpph512_mask ((A), (B), (C), (-1))) > > > > > > #define _mm512_mask_cmp_ph_mask(A, B, C, D) \ > > > - (__builtin_ia32_vcmpph_v32hf_mask ((B), (C), (D), (A))) > > > + (__builtin_ia32_cmpph512_mask ((B), (C), (D), (A))) > > > > > > -#define _mm512_cmp_round_ph_mask(A, B, C, D) \ > > > - (__builtin_ia32_vcmpph_v32hf_mask_round ((A), (B), (C), (-1), (D))) > > > +#define _mm512_cmp_round_ph_mask(A, B, C, D) \ > > > + (__builtin_ia32_cmpph512_mask_round ((A), (B), (C), (-1), (D))) > > > > > > -#define _mm512_mask_cmp_round_ph_mask(A, B, C, D, E) \ > > > - (__builtin_ia32_vcmpph_v32hf_mask_round ((B), (C), (D), (A), (E))) > > > +#define _mm512_mask_cmp_round_ph_mask(A, B, C, D, E) \ > > > + (__builtin_ia32_cmpph512_mask_round ((B), (C), (D), (A), (E))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > @@ -1046,9 +1046,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_cmp_sh_mask (__m128h __A, __m128h __B, const int __C) > > > { > > > return (__mmask8) > > > - __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, > > > - __C, (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + __builtin_ia32_cmpsh_mask_round (__A, __B, > > > + __C, (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __mmask8 > > > @@ -1057,9 +1057,9 @@ _mm_mask_cmp_sh_mask (__mmask8 __A, __m128h __B, __m128h __C, > > > const int __D) > > > { > > > return (__mmask8) > > > - __builtin_ia32_vcmpsh_v8hf_mask_round (__B, __C, > > > - __D, __A, > > > - _MM_FROUND_CUR_DIRECTION); > > > + __builtin_ia32_cmpsh_mask_round (__B, __C, > > > + __D, __A, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __mmask8 > > > @@ -1067,9 +1067,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_cmp_round_sh_mask (__m128h __A, __m128h __B, const int __C, > > > const int __D) > > > { > > > - return (__mmask8) __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, > > > - __C, (__mmask8) -1, > > > - __D); > > > + return (__mmask8) __builtin_ia32_cmpsh_mask_round (__A, __B, > > > + __C, (__mmask8) -1, > > > + __D); > > > } > > > > > > extern __inline __mmask8 > > > @@ -1077,25 +1077,25 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_cmp_round_sh_mask (__mmask8 __A, __m128h __B, __m128h __C, > > > const int __D, const int __E) > > > { > > > - return (__mmask8) __builtin_ia32_vcmpsh_v8hf_mask_round (__B, __C, > > > - __D, __A, > > > - __E); > > > + return (__mmask8) __builtin_ia32_cmpsh_mask_round (__B, __C, > > > + __D, __A, > > > + __E); > > > } > > > > > > #else > > > -#define _mm_cmp_sh_mask(A, B, C) \ > > > - (__builtin_ia32_vcmpsh_v8hf_mask_round ((A), (B), (C), (-1), \ > > > - (_MM_FROUND_CUR_DIRECTION))) > > > +#define _mm_cmp_sh_mask(A, B, C) \ > > > + (__builtin_ia32_cmpsh_mask_round ((A), (B), (C), (-1), \ > > > + (_MM_FROUND_CUR_DIRECTION))) > > > > > > -#define _mm_mask_cmp_sh_mask(A, B, C, D) \ > > > - (__builtin_ia32_vcmpsh_v8hf_mask_round ((B), (C), (D), (A), \ > > > - (_MM_FROUND_CUR_DIRECTION))) > > > +#define _mm_mask_cmp_sh_mask(A, B, C, D) \ > > > + (__builtin_ia32_cmpsh_mask_round ((B), (C), (D), (A), \ > > > + (_MM_FROUND_CUR_DIRECTION))) > > > > > > -#define _mm_cmp_round_sh_mask(A, B, C, D) \ > > > - (__builtin_ia32_vcmpsh_v8hf_mask_round ((A), (B), (C), (-1), (D))) > > > +#define _mm_cmp_round_sh_mask(A, B, C, D) \ > > > + (__builtin_ia32_cmpsh_mask_round ((A), (B), (C), (-1), (D))) > > > > > > -#define _mm_mask_cmp_round_sh_mask(A, B, C, D, E) \ > > > - (__builtin_ia32_vcmpsh_v8hf_mask_round ((B), (C), (D), (A), (E))) > > > +#define _mm_mask_cmp_round_sh_mask(A, B, C, D, E) \ > > > + (__builtin_ia32_cmpsh_mask_round ((B), (C), (D), (A), (E))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > @@ -1104,134 +1104,134 @@ extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_comieq_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_EQ_OS, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_EQ_OS, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_comilt_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_LT_OS, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_LT_OS, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_comile_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_LE_OS, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_LE_OS, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_comigt_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_GT_OS, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_GT_OS, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_comige_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_GE_OS, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_GE_OS, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_comineq_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_NEQ_US, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_NEQ_US, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_ucomieq_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_EQ_OQ, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_EQ_OQ, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_ucomilt_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_LT_OQ, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_LT_OQ, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_ucomile_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_LE_OQ, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_LE_OQ, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_ucomigt_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_GT_OQ, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_GT_OQ, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_ucomige_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_GE_OQ, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_GE_OQ, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_ucomineq_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_NEQ_UQ, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_NEQ_UQ, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > #ifdef __OPTIMIZE__ > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > - _mm_comi_sh (__m128h __A, __m128h __B, const int __P) > > > +_mm_comi_sh (__m128h __A, __m128h __B, const int __P) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, __P, > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, __P, > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline int > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_comi_round_sh (__m128h __A, __m128h __B, const int __P, const int __R) > > > { > > > - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, __P, > > > - (__mmask8) -1,__R); > > > + return __builtin_ia32_cmpsh_mask_round (__A, __B, __P, > > > + (__mmask8) -1,__R); > > > } > > > > > > #else > > > -#define _mm_comi_round_sh(A, B, P, R) \ > > > - (__builtin_ia32_vcmpsh_v8hf_mask_round ((A), (B), (P), (__mmask8) (-1), (R))) > > > -#define _mm_comi_sh(A, B, P) \ > > > - (__builtin_ia32_vcmpsh_v8hf_mask_round ((A), (B), (P), (__mmask8) (-1), \ > > > - _MM_FROUND_CUR_DIRECTION)) > > > +#define _mm_comi_round_sh(A, B, P, R) \ > > > + (__builtin_ia32_cmpsh_mask_round ((A), (B), (P), (__mmask8) (-1), (R))) > > > +#define _mm_comi_sh(A, B, P) \ > > > + (__builtin_ia32_cmpsh_mask_round ((A), (B), (P), (__mmask8) (-1), \ > > > + _MM_FROUND_CUR_DIRECTION)) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > @@ -1240,28 +1240,28 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_sqrt_ph (__m512h __A) > > > { > > > - return __builtin_ia32_vsqrtph_v32hf_mask_round (__A, > > > - _mm512_setzero_ph(), > > > - (__mmask32) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_sqrtph512_mask_round (__A, > > > + _mm512_setzero_ph(), > > > + (__mmask32) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_sqrt_ph (__m512h __A, __mmask32 __B, __m512h __C) > > > { > > > - return __builtin_ia32_vsqrtph_v32hf_mask_round (__C, __A, __B, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_sqrtph512_mask_round (__C, __A, __B, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_sqrt_ph (__mmask32 __A, __m512h __B) > > > { > > > - return __builtin_ia32_vsqrtph_v32hf_mask_round (__B, > > > - _mm512_setzero_ph (), > > > - __A, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_sqrtph512_mask_round (__B, > > > + _mm512_setzero_ph (), > > > + __A, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > #ifdef __OPTIMIZE__ > > > @@ -1269,9 +1269,9 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_sqrt_round_ph (__m512h __A, const int __B) > > > { > > > - return __builtin_ia32_vsqrtph_v32hf_mask_round (__A, > > > - _mm512_setzero_ph(), > > > - (__mmask32) -1, __B); > > > + return __builtin_ia32_sqrtph512_mask_round (__A, > > > + _mm512_setzero_ph(), > > > + (__mmask32) -1, __B); > > > } > > > > > > extern __inline __m512h > > > @@ -1279,31 +1279,31 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_sqrt_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vsqrtph_v32hf_mask_round (__C, __A, __B, __D); > > > + return __builtin_ia32_sqrtph512_mask_round (__C, __A, __B, __D); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_sqrt_round_ph (__mmask32 __A, __m512h __B, const int __C) > > > { > > > - return __builtin_ia32_vsqrtph_v32hf_mask_round (__B, > > > - _mm512_setzero_ph (), > > > - __A, __C); > > > + return __builtin_ia32_sqrtph512_mask_round (__B, > > > + _mm512_setzero_ph (), > > > + __A, __C); > > > } > > > > > > #else > > > -#define _mm512_sqrt_round_ph(A, B) \ > > > - (__builtin_ia32_vsqrtph_v32hf_mask_round ((A), \ > > > - _mm512_setzero_ph (), \ > > > - (__mmask32)-1, (B))) > > > +#define _mm512_sqrt_round_ph(A, B) \ > > > + (__builtin_ia32_sqrtph512_mask_round ((A), \ > > > + _mm512_setzero_ph (), \ > > > + (__mmask32)-1, (B))) > > > > > > -#define _mm512_mask_sqrt_round_ph(A, B, C, D) \ > > > - (__builtin_ia32_vsqrtph_v32hf_mask_round ((C), (A), (B), (D))) > > > +#define _mm512_mask_sqrt_round_ph(A, B, C, D) \ > > > + (__builtin_ia32_sqrtph512_mask_round ((C), (A), (B), (D))) > > > > > > -#define _mm512_maskz_sqrt_round_ph(A, B, C) \ > > > - (__builtin_ia32_vsqrtph_v32hf_mask_round ((B), \ > > > - _mm512_setzero_ph (), \ > > > - (A), (C))) > > > +#define _mm512_maskz_sqrt_round_ph(A, B, C) \ > > > + (__builtin_ia32_sqrtph512_mask_round ((B), \ > > > + _mm512_setzero_ph (), \ > > > + (A), (C))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > @@ -1312,23 +1312,23 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_rsqrt_ph (__m512h __A) > > > { > > > - return __builtin_ia32_vrsqrtph_v32hf_mask (__A, _mm512_setzero_ph (), > > > - (__mmask32) -1); > > > + return __builtin_ia32_rsqrtph512_mask (__A, _mm512_setzero_ph (), > > > + (__mmask32) -1); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_rsqrt_ph (__m512h __A, __mmask32 __B, __m512h __C) > > > { > > > - return __builtin_ia32_vrsqrtph_v32hf_mask (__C, __A, __B); > > > + return __builtin_ia32_rsqrtph512_mask (__C, __A, __B); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_rsqrt_ph (__mmask32 __A, __m512h __B) > > > { > > > - return __builtin_ia32_vrsqrtph_v32hf_mask (__B, _mm512_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_rsqrtph512_mask (__B, _mm512_setzero_ph (), > > > + __A); > > > } > > > > > > /* Intrinsics vrsqrtsh. */ > > > @@ -1336,23 +1336,23 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_rsqrt_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vrsqrtsh_v8hf_mask (__B, __A, _mm_setzero_ph (), > > > - (__mmask8) -1); > > > + return __builtin_ia32_rsqrtsh_mask (__B, __A, _mm_setzero_ph (), > > > + (__mmask8) -1); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_rsqrt_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vrsqrtsh_v8hf_mask (__D, __C, __A, __B); > > > + return __builtin_ia32_rsqrtsh_mask (__D, __C, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_rsqrt_sh (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vrsqrtsh_v8hf_mask (__C, __B, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_rsqrtsh_mask (__C, __B, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > /* Intrinsics vsqrtsh. */ > > > @@ -1360,27 +1360,27 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_sqrt_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__B, __A, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_sqrtsh_mask_round (__B, __A, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_sqrt_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__D, __C, __A, __B, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_sqrtsh_mask_round (__D, __C, __A, __B, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_sqrt_sh (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__C, __B, > > > - _mm_setzero_ph (), > > > - __A, _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_sqrtsh_mask_round (__C, __B, > > > + _mm_setzero_ph (), > > > + __A, _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > #ifdef __OPTIMIZE__ > > > @@ -1388,9 +1388,9 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_sqrt_round_sh (__m128h __A, __m128h __B, const int __C) > > > { > > > - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__B, __A, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, __C); > > > + return __builtin_ia32_sqrtsh_mask_round (__B, __A, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, __C); > > > } > > > > > > extern __inline __m128h > > > @@ -1398,8 +1398,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_sqrt_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > > > __m128h __D, const int __E) > > > { > > > - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__D, __C, __A, __B, > > > - __E); > > > + return __builtin_ia32_sqrtsh_mask_round (__D, __C, __A, __B, > > > + __E); > > > } > > > > > > extern __inline __m128h > > > @@ -1407,24 +1407,24 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_sqrt_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__C, __B, > > > - _mm_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_sqrtsh_mask_round (__C, __B, > > > + _mm_setzero_ph (), > > > + __A, __D); > > > } > > > > > > #else > > > #define _mm_sqrt_round_sh(A, B, C) \ > > > - (__builtin_ia32_vsqrtsh_v8hf_mask_round ((B), (A), \ > > > - _mm_setzero_ph (), \ > > > - (__mmask8)-1, (C))) > > > + (__builtin_ia32_sqrtsh_mask_round ((B), (A), \ > > > + _mm_setzero_ph (), \ > > > + (__mmask8)-1, (C))) > > > > > > #define _mm_mask_sqrt_round_sh(A, B, C, D, E) \ > > > - (__builtin_ia32_vsqrtsh_v8hf_mask_round ((D), (C), (A), (B), (E))) > > > + (__builtin_ia32_sqrtsh_mask_round ((D), (C), (A), (B), (E))) > > > > > > -#define _mm_maskz_sqrt_round_sh(A, B, C, D) \ > > > - (__builtin_ia32_vsqrtsh_v8hf_mask_round ((C), (B), \ > > > - _mm_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm_maskz_sqrt_round_sh(A, B, C, D) \ > > > + (__builtin_ia32_sqrtsh_mask_round ((C), (B), \ > > > + _mm_setzero_ph (), \ > > > + (A), (D))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > @@ -1433,23 +1433,23 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_rcp_ph (__m512h __A) > > > { > > > - return __builtin_ia32_vrcpph_v32hf_mask (__A, _mm512_setzero_ph (), > > > - (__mmask32) -1); > > > + return __builtin_ia32_rcpph512_mask (__A, _mm512_setzero_ph (), > > > + (__mmask32) -1); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_rcp_ph (__m512h __A, __mmask32 __B, __m512h __C) > > > { > > > - return __builtin_ia32_vrcpph_v32hf_mask (__C, __A, __B); > > > + return __builtin_ia32_rcpph512_mask (__C, __A, __B); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_rcp_ph (__mmask32 __A, __m512h __B) > > > { > > > - return __builtin_ia32_vrcpph_v32hf_mask (__B, _mm512_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_rcpph512_mask (__B, _mm512_setzero_ph (), > > > + __A); > > > } > > > > > > /* Intrinsics vrcpsh. */ > > > @@ -1457,23 +1457,23 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_rcp_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vrcpsh_v8hf_mask (__B, __A, _mm_setzero_ph (), > > > - (__mmask8) -1); > > > + return __builtin_ia32_rcpsh_mask (__B, __A, _mm_setzero_ph (), > > > + (__mmask8) -1); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_rcp_sh (__m128h __A, __mmask32 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vrcpsh_v8hf_mask (__D, __C, __A, __B); > > > + return __builtin_ia32_rcpsh_mask (__D, __C, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_rcp_sh (__mmask32 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vrcpsh_v8hf_mask (__C, __B, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_rcpsh_mask (__C, __B, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > /* Intrinsics vscalefph. */ > > > @@ -1481,28 +1481,28 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_scalef_ph (__m512h __A, __m512h __B) > > > { > > > - return __builtin_ia32_vscalefph_v32hf_mask_round (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_scalefph512_mask_round (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_scalef_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) > > > { > > > - return __builtin_ia32_vscalefph_v32hf_mask_round (__C, __D, __A, __B, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_scalefph512_mask_round (__C, __D, __A, __B, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_scalef_ph (__mmask32 __A, __m512h __B, __m512h __C) > > > { > > > - return __builtin_ia32_vscalefph_v32hf_mask_round (__B, __C, > > > - _mm512_setzero_ph (), > > > - __A, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_scalefph512_mask_round (__B, __C, > > > + _mm512_setzero_ph (), > > > + __A, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > #ifdef __OPTIMIZE__ > > > @@ -1510,9 +1510,9 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_scalef_round_ph (__m512h __A, __m512h __B, const int __C) > > > { > > > - return __builtin_ia32_vscalefph_v32hf_mask_round (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, __C); > > > + return __builtin_ia32_scalefph512_mask_round (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, __C); > > > } > > > > > > extern __inline __m512h > > > @@ -1520,8 +1520,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_scalef_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > > > __m512h __D, const int __E) > > > { > > > - return __builtin_ia32_vscalefph_v32hf_mask_round (__C, __D, __A, __B, > > > - __E); > > > + return __builtin_ia32_scalefph512_mask_round (__C, __D, __A, __B, > > > + __E); > > > } > > > > > > extern __inline __m512h > > > @@ -1529,24 +1529,24 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_scalef_round_ph (__mmask32 __A, __m512h __B, __m512h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vscalefph_v32hf_mask_round (__B, __C, > > > - _mm512_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_scalefph512_mask_round (__B, __C, > > > + _mm512_setzero_ph (), > > > + __A, __D); > > > } > > > > > > #else > > > -#define _mm512_scalef_round_ph(A, B, C) \ > > > - (__builtin_ia32_vscalefph_v32hf_mask_round ((A), (B), \ > > > - _mm512_setzero_ph (), \ > > > - (__mmask32)-1, (C))) > > > +#define _mm512_scalef_round_ph(A, B, C) \ > > > + (__builtin_ia32_scalefph512_mask_round ((A), (B), \ > > > + _mm512_setzero_ph (), \ > > > + (__mmask32)-1, (C))) > > > > > > #define _mm512_mask_scalef_round_ph(A, B, C, D, E) \ > > > - (__builtin_ia32_vscalefph_v32hf_mask_round ((C), (D), (A), (B), (E))) > > > + (__builtin_ia32_scalefph512_mask_round ((C), (D), (A), (B), (E))) > > > > > > -#define _mm512_maskz_scalef_round_ph(A, B, C, D) \ > > > - (__builtin_ia32_vscalefph_v32hf_mask_round ((B), (C), \ > > > - _mm512_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm512_maskz_scalef_round_ph(A, B, C, D) \ > > > + (__builtin_ia32_scalefph512_mask_round ((B), (C), \ > > > + _mm512_setzero_ph (), \ > > > + (A), (D))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > @@ -1555,28 +1555,28 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_scalef_sh (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vscalefsh_v8hf_mask_round (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_scalefsh_mask_round (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_scalef_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vscalefsh_v8hf_mask_round (__C, __D, __A, __B, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_scalefsh_mask_round (__C, __D, __A, __B, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_scalef_sh (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vscalefsh_v8hf_mask_round (__B, __C, > > > - _mm_setzero_ph (), > > > - __A, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_scalefsh_mask_round (__B, __C, > > > + _mm_setzero_ph (), > > > + __A, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > #ifdef __OPTIMIZE__ > > > @@ -1584,9 +1584,9 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_scalef_round_sh (__m128h __A, __m128h __B, const int __C) > > > { > > > - return __builtin_ia32_vscalefsh_v8hf_mask_round (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, __C); > > > + return __builtin_ia32_scalefsh_mask_round (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, __C); > > > } > > > > > > extern __inline __m128h > > > @@ -1594,8 +1594,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_scalef_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > > > __m128h __D, const int __E) > > > { > > > - return __builtin_ia32_vscalefsh_v8hf_mask_round (__C, __D, __A, __B, > > > - __E); > > > + return __builtin_ia32_scalefsh_mask_round (__C, __D, __A, __B, > > > + __E); > > > } > > > > > > extern __inline __m128h > > > @@ -1603,23 +1603,23 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_scalef_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vscalefsh_v8hf_mask_round (__B, __C, > > > - _mm_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_scalefsh_mask_round (__B, __C, > > > + _mm_setzero_ph (), > > > + __A, __D); > > > } > > > > > > #else > > > -#define _mm_scalef_round_sh(A, B, C) \ > > > - (__builtin_ia32_vscalefsh_v8hf_mask_round ((A), (B), \ > > > - _mm_setzero_ph (), \ > > > - (__mmask8)-1, (C))) > > > +#define _mm_scalef_round_sh(A, B, C) \ > > > + (__builtin_ia32_scalefsh_mask_round ((A), (B), \ > > > + _mm_setzero_ph (), \ > > > + (__mmask8)-1, (C))) > > > > > > -#define _mm_mask_scalef_round_sh(A, B, C, D, E) \ > > > - (__builtin_ia32_vscalefsh_v8hf_mask_round ((C), (D), (A), (B), (E))) > > > +#define _mm_mask_scalef_round_sh(A, B, C, D, E) \ > > > + (__builtin_ia32_scalefsh_mask_round ((C), (D), (A), (B), (E))) > > > > > > -#define _mm_maskz_scalef_round_sh(A, B, C, D) \ > > > - (__builtin_ia32_vscalefsh_v8hf_mask_round ((B), (C), _mm_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm_maskz_scalef_round_sh(A, B, C, D) \ > > > + (__builtin_ia32_scalefsh_mask_round ((B), (C), _mm_setzero_ph (), \ > > > + (A), (D))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > @@ -1629,37 +1629,37 @@ extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_reduce_ph (__m512h __A, int __B) > > > { > > > - return __builtin_ia32_vreduceph_v32hf_mask_round (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_reduceph512_mask_round (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_reduce_ph (__m512h __A, __mmask32 __B, __m512h __C, int __D) > > > { > > > - return __builtin_ia32_vreduceph_v32hf_mask_round (__C, __D, __A, __B, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_reduceph512_mask_round (__C, __D, __A, __B, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_reduce_ph (__mmask32 __A, __m512h __B, int __C) > > > { > > > - return __builtin_ia32_vreduceph_v32hf_mask_round (__B, __C, > > > - _mm512_setzero_ph (), > > > - __A, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_reduceph512_mask_round (__B, __C, > > > + _mm512_setzero_ph (), > > > + __A, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_reduce_round_ph (__m512h __A, int __B, const int __C) > > > { > > > - return __builtin_ia32_vreduceph_v32hf_mask_round (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, __C); > > > + return __builtin_ia32_reduceph512_mask_round (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, __C); > > > } > > > > > > extern __inline __m512h > > > @@ -1667,8 +1667,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_reduce_round_ph (__m512h __A, __mmask32 __B, __m512h __C, > > > int __D, const int __E) > > > { > > > - return __builtin_ia32_vreduceph_v32hf_mask_round (__C, __D, __A, __B, > > > - __E); > > > + return __builtin_ia32_reduceph512_mask_round (__C, __D, __A, __B, > > > + __E); > > > } > > > > > > extern __inline __m512h > > > @@ -1676,39 +1676,39 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_reduce_round_ph (__mmask32 __A, __m512h __B, int __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vreduceph_v32hf_mask_round (__B, __C, > > > - _mm512_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_reduceph512_mask_round (__B, __C, > > > + _mm512_setzero_ph (), > > > + __A, __D); > > > } > > > > > > #else > > > #define _mm512_reduce_ph(A, B) \ > > > - (__builtin_ia32_vreduceph_v32hf_mask_round ((A), (B), \ > > > - _mm512_setzero_ph (), \ > > > - (__mmask32)-1, \ > > > - _MM_FROUND_CUR_DIRECTION)) > > > + (__builtin_ia32_reduceph512_mask_round ((A), (B), \ > > > + _mm512_setzero_ph (), \ > > > + (__mmask32)-1, \ > > > + _MM_FROUND_CUR_DIRECTION)) > > > > > > #define _mm512_mask_reduce_ph(A, B, C, D) \ > > > - (__builtin_ia32_vreduceph_v32hf_mask_round ((C), (D), (A), (B), \ > > > - _MM_FROUND_CUR_DIRECTION)) > > > + (__builtin_ia32_reduceph512_mask_round ((C), (D), (A), (B), \ > > > + _MM_FROUND_CUR_DIRECTION)) > > > > > > #define _mm512_maskz_reduce_ph(A, B, C) \ > > > - (__builtin_ia32_vreduceph_v32hf_mask_round ((B), (C), \ > > > - _mm512_setzero_ph (), \ > > > - (A), _MM_FROUND_CUR_DIRECTION)) > > > + (__builtin_ia32_reduceph512_mask_round ((B), (C), \ > > > + _mm512_setzero_ph (), \ > > > + (A), _MM_FROUND_CUR_DIRECTION)) > > > > > > -#define _mm512_reduce_round_ph(A, B, C) \ > > > - (__builtin_ia32_vreduceph_v32hf_mask_round ((A), (B), \ > > > - _mm512_setzero_ph (), \ > > > - (__mmask32)-1, (C))) > > > +#define _mm512_reduce_round_ph(A, B, C) \ > > > + (__builtin_ia32_reduceph512_mask_round ((A), (B), \ > > > + _mm512_setzero_ph (), \ > > > + (__mmask32)-1, (C))) > > > > > > #define _mm512_mask_reduce_round_ph(A, B, C, D, E) \ > > > - (__builtin_ia32_vreduceph_v32hf_mask_round ((C), (D), (A), (B), (E))) > > > + (__builtin_ia32_reduceph512_mask_round ((C), (D), (A), (B), (E))) > > > > > > -#define _mm512_maskz_reduce_round_ph(A, B, C, D) \ > > > - (__builtin_ia32_vreduceph_v32hf_mask_round ((B), (C), \ > > > - _mm512_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm512_maskz_reduce_round_ph(A, B, C, D) \ > > > + (__builtin_ia32_reduceph512_mask_round ((B), (C), \ > > > + _mm512_setzero_ph (), \ > > > + (A), (D))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > @@ -1718,10 +1718,10 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_reduce_sh (__m128h __A, __m128h __B, int __C) > > > { > > > - return __builtin_ia32_vreducesh_v8hf_mask_round (__A, __B, __C, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_reducesh_mask_round (__A, __B, __C, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m128h > > > @@ -1729,26 +1729,26 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_reduce_sh (__m128h __A, __mmask8 __B, __m128h __C, > > > __m128h __D, int __E) > > > { > > > - return __builtin_ia32_vreducesh_v8hf_mask_round (__C, __D, __E, __A, __B, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_reducesh_mask_round (__C, __D, __E, __A, __B, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_reduce_sh (__mmask8 __A, __m128h __B, __m128h __C, int __D) > > > { > > > - return __builtin_ia32_vreducesh_v8hf_mask_round (__B, __C, __D, > > > - _mm_setzero_ph (), __A, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_reducesh_mask_round (__B, __C, __D, > > > + _mm_setzero_ph (), __A, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_reduce_round_sh (__m128h __A, __m128h __B, int __C, const int __D) > > > { > > > - return __builtin_ia32_vreducesh_v8hf_mask_round (__A, __B, __C, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, __D); > > > + return __builtin_ia32_reducesh_mask_round (__A, __B, __C, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, __D); > > > } > > > > > > extern __inline __m128h > > > @@ -1756,8 +1756,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_reduce_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > > > __m128h __D, int __E, const int __F) > > > { > > > - return __builtin_ia32_vreducesh_v8hf_mask_round (__C, __D, __E, __A, > > > - __B, __F); > > > + return __builtin_ia32_reducesh_mask_round (__C, __D, __E, __A, > > > + __B, __F); > > > } > > > > > > extern __inline __m128h > > > @@ -1765,81 +1765,81 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_reduce_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > > > int __D, const int __E) > > > { > > > - return __builtin_ia32_vreducesh_v8hf_mask_round (__B, __C, __D, > > > - _mm_setzero_ph (), > > > - __A, __E); > > > + return __builtin_ia32_reducesh_mask_round (__B, __C, __D, > > > + _mm_setzero_ph (), > > > + __A, __E); > > > } > > > > > > #else > > > #define _mm_reduce_sh(A, B, C) \ > > > - (__builtin_ia32_vreducesh_v8hf_mask_round ((A), (B), (C), \ > > > - _mm_setzero_ph (), \ > > > - (__mmask8)-1, \ > > > - _MM_FROUND_CUR_DIRECTION)) > > > + (__builtin_ia32_reducesh_mask_round ((A), (B), (C), \ > > > + _mm_setzero_ph (), \ > > > + (__mmask8)-1, \ > > > + _MM_FROUND_CUR_DIRECTION)) > > > > > > #define _mm_mask_reduce_sh(A, B, C, D, E) \ > > > - (__builtin_ia32_vreducesh_v8hf_mask_round ((C), (D), (E), (A), (B), \ > > > - _MM_FROUND_CUR_DIRECTION)) > > > + (__builtin_ia32_reducesh_mask_round ((C), (D), (E), (A), (B), \ > > > + _MM_FROUND_CUR_DIRECTION)) > > > > > > #define _mm_maskz_reduce_sh(A, B, C, D) \ > > > - (__builtin_ia32_vreducesh_v8hf_mask_round ((B), (C), (D), \ > > > - _mm_setzero_ph (), \ > > > - (A), _MM_FROUND_CUR_DIRECTION)) > > > + (__builtin_ia32_reducesh_mask_round ((B), (C), (D), \ > > > + _mm_setzero_ph (), \ > > > + (A), _MM_FROUND_CUR_DIRECTION)) > > > > > > #define _mm_reduce_round_sh(A, B, C, D) \ > > > - (__builtin_ia32_vreducesh_v8hf_mask_round ((A), (B), (C), \ > > > - _mm_setzero_ph (), \ > > > - (__mmask8)-1, (D))) > > > + (__builtin_ia32_reducesh_mask_round ((A), (B), (C), \ > > > + _mm_setzero_ph (), \ > > > + (__mmask8)-1, (D))) > > > > > > #define _mm_mask_reduce_round_sh(A, B, C, D, E, F) \ > > > - (__builtin_ia32_vreducesh_v8hf_mask_round ((C), (D), (E), (A), (B), (F))) > > > + (__builtin_ia32_reducesh_mask_round ((C), (D), (E), (A), (B), (F))) > > > > > > #define _mm_maskz_reduce_round_sh(A, B, C, D, E) \ > > > - (__builtin_ia32_vreducesh_v8hf_mask_round ((B), (C), (D), \ > > > - _mm_setzero_ph (), \ > > > - (A), (E))) > > > + (__builtin_ia32_reducesh_mask_round ((B), (C), (D), \ > > > + _mm_setzero_ph (), \ > > > + (A), (E))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > /* Intrinsics vrndscaleph. */ > > > #ifdef __OPTIMIZE__ > > > extern __inline __m512h > > > -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_roundscale_ph (__m512h __A, int __B) > > > { > > > - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_rndscaleph512_mask_round (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_roundscale_ph (__m512h __A, __mmask32 __B, > > > - __m512h __C, int __D) > > > + __m512h __C, int __D) > > > { > > > - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__C, __D, __A, __B, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_rndscaleph512_mask_round (__C, __D, __A, __B, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_roundscale_ph (__mmask32 __A, __m512h __B, int __C) > > > { > > > - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__B, __C, > > > - _mm512_setzero_ph (), > > > - __A, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_rndscaleph512_mask_round (__B, __C, > > > + _mm512_setzero_ph (), > > > + __A, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m512h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_roundscale_round_ph (__m512h __A, int __B, const int __C) > > > { > > > - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__A, __B, > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, > > > - __C); > > > + return __builtin_ia32_rndscaleph512_mask_round (__A, __B, > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, > > > + __C); > > > } > > > > > > extern __inline __m512h > > > @@ -1847,8 +1847,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_mask_roundscale_round_ph (__m512h __A, __mmask32 __B, > > > __m512h __C, int __D, const int __E) > > > { > > > - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__C, __D, __A, > > > - __B, __E); > > > + return __builtin_ia32_rndscaleph512_mask_round (__C, __D, __A, > > > + __B, __E); > > > } > > > > > > extern __inline __m512h > > > @@ -1856,52 +1856,52 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_roundscale_round_ph (__mmask32 __A, __m512h __B, int __C, > > > const int __D) > > > { > > > - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__B, __C, > > > - _mm512_setzero_ph (), > > > - __A, __D); > > > + return __builtin_ia32_rndscaleph512_mask_round (__B, __C, > > > + _mm512_setzero_ph (), > > > + __A, __D); > > > } > > > > > > #else > > > -#define _mm512_roundscale_ph(A, B) \ > > > - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((A), (B), \ > > > - _mm512_setzero_ph (), \ > > > - (__mmask32)-1, \ > > > - _MM_FROUND_CUR_DIRECTION)) > > > - > > > -#define _mm512_mask_roundscale_ph(A, B, C, D) \ > > > - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((C), (D), (A), (B), \ > > > - _MM_FROUND_CUR_DIRECTION)) > > > - > > > -#define _mm512_maskz_roundscale_ph(A, B, C) \ > > > - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((B), (C), \ > > > - _mm512_setzero_ph (), \ > > > - (A), \ > > > - _MM_FROUND_CUR_DIRECTION)) > > > -#define _mm512_roundscale_round_ph(A, B, C) \ > > > - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((A), (B), \ > > > - _mm512_setzero_ph (), \ > > > - (__mmask32)-1, (C))) > > > +#define _mm512_roundscale_ph(A, B) \ > > > + (__builtin_ia32_rndscaleph512_mask_round ((A), (B), \ > > > + _mm512_setzero_ph (), \ > > > + (__mmask32)-1, \ > > > + _MM_FROUND_CUR_DIRECTION)) > > > + > > > +#define _mm512_mask_roundscale_ph(A, B, C, D) \ > > > + (__builtin_ia32_rndscaleph512_mask_round ((C), (D), (A), (B), \ > > > + _MM_FROUND_CUR_DIRECTION)) > > > + > > > +#define _mm512_maskz_roundscale_ph(A, B, C) \ > > > + (__builtin_ia32_rndscaleph512_mask_round ((B), (C), \ > > > + _mm512_setzero_ph (), \ > > > + (A), \ > > > + _MM_FROUND_CUR_DIRECTION)) > > > +#define _mm512_roundscale_round_ph(A, B, C) \ > > > + (__builtin_ia32_rndscaleph512_mask_round ((A), (B), \ > > > + _mm512_setzero_ph (), \ > > > + (__mmask32)-1, (C))) > > > > > > #define _mm512_mask_roundscale_round_ph(A, B, C, D, E) \ > > > - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((C), (D), (A), (B), (E))) > > > + (__builtin_ia32_rndscaleph512_mask_round ((C), (D), (A), (B), (E))) > > > > > > -#define _mm512_maskz_roundscale_round_ph(A, B, C, D) \ > > > - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((B), (C), \ > > > - _mm512_setzero_ph (), \ > > > - (A), (D))) > > > +#define _mm512_maskz_roundscale_round_ph(A, B, C, D) \ > > > + (__builtin_ia32_rndscaleph512_mask_round ((B), (C), \ > > > + _mm512_setzero_ph (), \ > > > + (A), (D))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > /* Intrinsics vrndscalesh. */ > > > #ifdef __OPTIMIZE__ > > > extern __inline __m128h > > > -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_roundscale_sh (__m128h __A, __m128h __B, int __C) > > > { > > > - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__A, __B, __C, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_rndscalesh_mask_round (__A, __B, __C, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m128h > > > @@ -1909,27 +1909,27 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_roundscale_sh (__m128h __A, __mmask8 __B, __m128h __C, > > > __m128h __D, int __E) > > > { > > > - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__C, __D, __E, __A, __B, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_rndscalesh_mask_round (__C, __D, __E, __A, __B, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_roundscale_sh (__mmask8 __A, __m128h __B, __m128h __C, int __D) > > > { > > > - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__B, __C, __D, > > > - _mm_setzero_ph (), __A, > > > - _MM_FROUND_CUR_DIRECTION); > > > + return __builtin_ia32_rndscalesh_mask_round (__B, __C, __D, > > > + _mm_setzero_ph (), __A, > > > + _MM_FROUND_CUR_DIRECTION); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_roundscale_round_sh (__m128h __A, __m128h __B, int __C, const int __D) > > > { > > > - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__A, __B, __C, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1, > > > - __D); > > > + return __builtin_ia32_rndscalesh_mask_round (__A, __B, __C, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1, > > > + __D); > > > } > > > > > > extern __inline __m128h > > > @@ -1937,8 +1937,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_roundscale_round_sh (__m128h __A, __mmask8 __B, __m128h __C, > > > __m128h __D, int __E, const int __F) > > > { > > > - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__C, __D, __E, > > > - __A, __B, __F); > > > + return __builtin_ia32_rndscalesh_mask_round (__C, __D, __E, > > > + __A, __B, __F); > > > } > > > > > > extern __inline __m128h > > > @@ -1946,46 +1946,46 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_roundscale_round_sh (__mmask8 __A, __m128h __B, __m128h __C, > > > int __D, const int __E) > > > { > > > - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__B, __C, __D, > > > - _mm_setzero_ph (), > > > - __A, __E); > > > + return __builtin_ia32_rndscalesh_mask_round (__B, __C, __D, > > > + _mm_setzero_ph (), > > > + __A, __E); > > > } > > > > > > #else > > > #define _mm_roundscale_sh(A, B, C) \ > > > - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((A), (B), (C), \ > > > - _mm_setzero_ph (), \ > > > - (__mmask8)-1, \ > > > - _MM_FROUND_CUR_DIRECTION)) > > > + (__builtin_ia32_rndscalesh_mask_round ((A), (B), (C), \ > > > + _mm_setzero_ph (), \ > > > + (__mmask8)-1, \ > > > + _MM_FROUND_CUR_DIRECTION)) > > > > > > #define _mm_mask_roundscale_sh(A, B, C, D, E) \ > > > - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((C), (D), (E), (A), (B), \ > > > - _MM_FROUND_CUR_DIRECTION)) > > > + (__builtin_ia32_rndscalesh_mask_round ((C), (D), (E), (A), (B), \ > > > + _MM_FROUND_CUR_DIRECTION)) > > > > > > #define _mm_maskz_roundscale_sh(A, B, C, D) \ > > > - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((B), (C), (D), \ > > > - _mm_setzero_ph (), \ > > > - (A), _MM_FROUND_CUR_DIRECTION)) > > > + (__builtin_ia32_rndscalesh_mask_round ((B), (C), (D), \ > > > + _mm_setzero_ph (), \ > > > + (A), _MM_FROUND_CUR_DIRECTION)) > > > > > > -#define _mm_roundscale_round_sh(A, B, C, D) \ > > > - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((A), (B), (C), \ > > > - _mm_setzero_ph (), \ > > > - (__mmask8)-1, (D))) > > > +#define _mm_roundscale_round_sh(A, B, C, D) \ > > > + (__builtin_ia32_rndscalesh_mask_round ((A), (B), (C), \ > > > + _mm_setzero_ph (), \ > > > + (__mmask8)-1, (D))) > > > > > > #define _mm_mask_roundscale_round_sh(A, B, C, D, E, F) \ > > > - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((C), (D), (E), (A), (B), (F))) > > > + (__builtin_ia32_rndscalesh_mask_round ((C), (D), (E), (A), (B), (F))) > > > > > > -#define _mm_maskz_roundscale_round_sh(A, B, C, D, E) \ > > > - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((B), (C), (D), \ > > > - _mm_setzero_ph (), \ > > > - (A), (E))) > > > +#define _mm_maskz_roundscale_round_sh(A, B, C, D, E) \ > > > + (__builtin_ia32_rndscalesh_mask_round ((B), (C), (D), \ > > > + _mm_setzero_ph (), \ > > > + (A), (E))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > /* Intrinsics vfpclasssh. */ > > > #ifdef __OPTIMIZE__ > > > extern __inline __mmask8 > > > -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_fpclass_sh_mask (__m128h __A, const int __imm) > > > { > > > return (__mmask8) __builtin_ia32_fpclasssh_mask ((__v8hf) __A, __imm, > > > @@ -2031,11 +2031,11 @@ _mm512_fpclass_ph_mask (__m512h __A, const int __imm) > > > > > > #else > > > #define _mm512_mask_fpclass_ph_mask(u, x, c) \ > > > - ((__mmask32) __builtin_ia32_fpclassph512_mask ((__v32hf) (__m512h) (x),\ > > > + ((__mmask32) __builtin_ia32_fpclassph512_mask ((__v32hf) (__m512h) (x), \ > > > (int) (c),(__mmask8)(u))) > > > > > > #define _mm512_fpclass_ph_mask(x, c) \ > > > - ((__mmask32) __builtin_ia32_fpclassph512_mask ((__v32hf) (__m512h) (x),\ > > > + ((__mmask32) __builtin_ia32_fpclassph512_mask ((__v32hf) (__m512h) (x), \ > > > (int) (c),(__mmask8)-1)) > > > #endif /* __OPIMTIZE__ */ > > > > > > @@ -2141,9 +2141,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_getexp_round_ph (__m512h __A, const int __R) > > > { > > > return (__m512h) __builtin_ia32_getexpph512_mask ((__v32hf) __A, > > > - (__v32hf) > > > - _mm512_setzero_ph (), > > > - (__mmask32) -1, __R); > > > + (__v32hf) > > > + _mm512_setzero_ph (), > > > + (__mmask32) -1, __R); > > > } > > > > > > extern __inline __m512h > > > @@ -2152,8 +2152,8 @@ _mm512_mask_getexp_round_ph (__m512h __W, __mmask32 __U, __m512h __A, > > > const int __R) > > > { > > > return (__m512h) __builtin_ia32_getexpph512_mask ((__v32hf) __A, > > > - (__v32hf) __W, > > > - (__mmask32) __U, __R); > > > + (__v32hf) __W, > > > + (__mmask32) __U, __R); > > > } > > > > > > extern __inline __m512h > > > @@ -2161,37 +2161,37 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm512_maskz_getexp_round_ph (__mmask32 __U, __m512h __A, const int __R) > > > { > > > return (__m512h) __builtin_ia32_getexpph512_mask ((__v32hf) __A, > > > - (__v32hf) > > > - _mm512_setzero_ph (), > > > - (__mmask32) __U, __R); > > > + (__v32hf) > > > + _mm512_setzero_ph (), > > > + (__mmask32) __U, __R); > > > } > > > > > > #else > > > -#define _mm_getexp_round_sh(A, B, R) \ > > > - ((__m128h)__builtin_ia32_getexpsh_mask_round((__v8hf)(__m128h)(A), \ > > > - (__v8hf)(__m128h)(B), \ > > > - (__v8hf)_mm_setzero_ph(), \ > > > +#define _mm_getexp_round_sh(A, B, R) \ > > > + ((__m128h)__builtin_ia32_getexpsh_mask_round((__v8hf)(__m128h)(A), \ > > > + (__v8hf)(__m128h)(B), \ > > > + (__v8hf)_mm_setzero_ph(), \ > > > (__mmask8)-1, R)) > > > > > > -#define _mm_mask_getexp_round_sh(W, U, A, B, C) \ > > > +#define _mm_mask_getexp_round_sh(W, U, A, B, C) \ > > > (__m128h)__builtin_ia32_getexpsh_mask_round(A, B, W, U, C) > > > > > > -#define _mm_maskz_getexp_round_sh(U, A, B, C) \ > > > - (__m128h)__builtin_ia32_getexpsh_mask_round(A, B, \ > > > - (__v8hf)_mm_setzero_ph(), \ > > > +#define _mm_maskz_getexp_round_sh(U, A, B, C) \ > > > + (__m128h)__builtin_ia32_getexpsh_mask_round(A, B, \ > > > + (__v8hf)_mm_setzero_ph(), \ > > > U, C) > > > > > > -#define _mm512_getexp_round_ph(A, R) \ > > > - ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ > > > - (__v32hf)_mm512_setzero_ph(), (__mmask32)-1, R)) > > > +#define _mm512_getexp_round_ph(A, R) \ > > > + ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ > > > + (__v32hf)_mm512_setzero_ph(), (__mmask32)-1, R)) > > > > > > -#define _mm512_mask_getexp_round_ph(W, U, A, R) \ > > > - ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ > > > - (__v32hf)(__m512h)(W), (__mmask32)(U), R)) > > > +#define _mm512_mask_getexp_round_ph(W, U, A, R) \ > > > + ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ > > > + (__v32hf)(__m512h)(W), (__mmask32)(U), R)) > > > > > > -#define _mm512_maskz_getexp_round_ph(U, A, R) \ > > > - ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ > > > - (__v32hf)_mm512_setzero_ph(), (__mmask32)(U), R)) > > > +#define _mm512_maskz_getexp_round_ph(U, A, R) \ > > > + ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ > > > + (__v32hf)_mm512_setzero_ph(), (__mmask32)(U), R)) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > diff --git a/gcc/config/i386/avx512fp16vlintrin.h b/gcc/config/i386/avx512fp16vlintrin.h > > > index 206d60407fc..e9478792a03 100644 > > > --- a/gcc/config/i386/avx512fp16vlintrin.h > > > +++ b/gcc/config/i386/avx512fp16vlintrin.h > > > @@ -53,30 +53,30 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_add_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vaddph_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_addph128_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_add_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) > > > { > > > - return __builtin_ia32_vaddph_v16hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_addph256_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_add_ph (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vaddph_v8hf_mask (__B, __C, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_addph128_mask (__B, __C, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_add_ph (__mmask16 __A, __m256h __B, __m256h __C) > > > { > > > - return __builtin_ia32_vaddph_v16hf_mask (__B, __C, > > > - _mm256_setzero_ph (), __A); > > > + return __builtin_ia32_addph256_mask (__B, __C, > > > + _mm256_setzero_ph (), __A); > > > } > > > > > > extern __inline __m128h > > > @@ -97,30 +97,30 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_sub_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vsubph_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_subph128_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_sub_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) > > > { > > > - return __builtin_ia32_vsubph_v16hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_subph256_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_sub_ph (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vsubph_v8hf_mask (__B, __C, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_subph128_mask (__B, __C, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_sub_ph (__mmask16 __A, __m256h __B, __m256h __C) > > > { > > > - return __builtin_ia32_vsubph_v16hf_mask (__B, __C, > > > - _mm256_setzero_ph (), __A); > > > + return __builtin_ia32_subph256_mask (__B, __C, > > > + _mm256_setzero_ph (), __A); > > > } > > > > > > extern __inline __m128h > > > @@ -141,30 +141,30 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_mul_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vmulph_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_mulph128_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_mul_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) > > > { > > > - return __builtin_ia32_vmulph_v16hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_mulph256_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_mul_ph (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vmulph_v8hf_mask (__B, __C, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_mulph128_mask (__B, __C, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_mul_ph (__mmask16 __A, __m256h __B, __m256h __C) > > > { > > > - return __builtin_ia32_vmulph_v16hf_mask (__B, __C, > > > - _mm256_setzero_ph (), __A); > > > + return __builtin_ia32_mulph256_mask (__B, __C, > > > + _mm256_setzero_ph (), __A); > > > } > > > > > > extern __inline __m128h > > > @@ -185,30 +185,30 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_div_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vdivph_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_divph128_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_div_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) > > > { > > > - return __builtin_ia32_vdivph_v16hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_divph256_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_div_ph (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vdivph_v8hf_mask (__B, __C, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_divph128_mask (__B, __C, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_div_ph (__mmask16 __A, __m256h __B, __m256h __C) > > > { > > > - return __builtin_ia32_vdivph_v16hf_mask (__B, __C, > > > - _mm256_setzero_ph (), __A); > > > + return __builtin_ia32_divph256_mask (__B, __C, > > > + _mm256_setzero_ph (), __A); > > > } > > > > > > /* Intrinsics v[max,min]ph. */ > > > @@ -216,96 +216,96 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_max_ph (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vmaxph_v8hf_mask (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1); > > > + return __builtin_ia32_maxph128_mask (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_max_ph (__m256h __A, __m256h __B) > > > { > > > - return __builtin_ia32_vmaxph_v16hf_mask (__A, __B, > > > - _mm256_setzero_ph (), > > > - (__mmask16) -1); > > > + return __builtin_ia32_maxph256_mask (__A, __B, > > > + _mm256_setzero_ph (), > > > + (__mmask16) -1); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_max_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vmaxph_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_maxph128_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_max_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) > > > { > > > - return __builtin_ia32_vmaxph_v16hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_maxph256_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_max_ph (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vmaxph_v8hf_mask (__B, __C, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_maxph128_mask (__B, __C, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_max_ph (__mmask16 __A, __m256h __B, __m256h __C) > > > { > > > - return __builtin_ia32_vmaxph_v16hf_mask (__B, __C, > > > - _mm256_setzero_ph (), __A); > > > + return __builtin_ia32_maxph256_mask (__B, __C, > > > + _mm256_setzero_ph (), __A); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_min_ph (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vminph_v8hf_mask (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1); > > > + return __builtin_ia32_minph128_mask (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_min_ph (__m256h __A, __m256h __B) > > > { > > > - return __builtin_ia32_vminph_v16hf_mask (__A, __B, > > > - _mm256_setzero_ph (), > > > - (__mmask16) -1); > > > + return __builtin_ia32_minph256_mask (__A, __B, > > > + _mm256_setzero_ph (), > > > + (__mmask16) -1); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_min_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vminph_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_minph128_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_min_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) > > > { > > > - return __builtin_ia32_vminph_v16hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_minph256_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_min_ph (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vminph_v8hf_mask (__B, __C, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_minph128_mask (__B, __C, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_min_ph (__mmask16 __A, __m256h __B, __m256h __C) > > > { > > > - return __builtin_ia32_vminph_v16hf_mask (__B, __C, > > > - _mm256_setzero_ph (), __A); > > > + return __builtin_ia32_minph256_mask (__B, __C, > > > + _mm256_setzero_ph (), __A); > > > } > > > > > > /* vcmpph */ > > > @@ -314,8 +314,8 @@ extern __inline __mmask8 > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_cmp_ph_mask (__m128h __A, __m128h __B, const int __C) > > > { > > > - return (__mmask8) __builtin_ia32_vcmpph_v8hf_mask (__A, __B, __C, > > > - (__mmask8) -1); > > > + return (__mmask8) __builtin_ia32_cmpph128_mask (__A, __B, __C, > > > + (__mmask8) -1); > > > } > > > > > > extern __inline __mmask8 > > > @@ -323,15 +323,15 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_cmp_ph_mask (__mmask8 __A, __m128h __B, __m128h __C, > > > const int __D) > > > { > > > - return (__mmask8) __builtin_ia32_vcmpph_v8hf_mask (__B, __C, __D, __A); > > > + return (__mmask8) __builtin_ia32_cmpph128_mask (__B, __C, __D, __A); > > > } > > > > > > extern __inline __mmask16 > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_cmp_ph_mask (__m256h __A, __m256h __B, const int __C) > > > { > > > - return (__mmask16) __builtin_ia32_vcmpph_v16hf_mask (__A, __B, __C, > > > - (__mmask16) -1); > > > + return (__mmask16) __builtin_ia32_cmpph256_mask (__A, __B, __C, > > > + (__mmask16) -1); > > > } > > > > > > extern __inline __mmask16 > > > @@ -339,22 +339,22 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_cmp_ph_mask (__mmask16 __A, __m256h __B, __m256h __C, > > > const int __D) > > > { > > > - return (__mmask16) __builtin_ia32_vcmpph_v16hf_mask (__B, __C, __D, > > > - __A); > > > + return (__mmask16) __builtin_ia32_cmpph256_mask (__B, __C, __D, > > > + __A); > > > } > > > > > > #else > > > -#define _mm_cmp_ph_mask(A, B, C) \ > > > - (__builtin_ia32_vcmpph_v8hf_mask ((A), (B), (C), (-1))) > > > +#define _mm_cmp_ph_mask(A, B, C) \ > > > + (__builtin_ia32_cmpph128_mask ((A), (B), (C), (-1))) > > > > > > -#define _mm_mask_cmp_ph_mask(A, B, C, D) \ > > > - (__builtin_ia32_vcmpph_v8hf_mask ((B), (C), (D), (A))) > > > +#define _mm_mask_cmp_ph_mask(A, B, C, D) \ > > > + (__builtin_ia32_cmpph128_mask ((B), (C), (D), (A))) > > > > > > -#define _mm256_cmp_ph_mask(A, B, C) \ > > > - (__builtin_ia32_vcmpph_v16hf_mask ((A), (B), (C), (-1))) > > > +#define _mm256_cmp_ph_mask(A, B, C) \ > > > + (__builtin_ia32_cmpph256_mask ((A), (B), (C), (-1))) > > > > > > -#define _mm256_mask_cmp_ph_mask(A, B, C, D) \ > > > - (__builtin_ia32_vcmpph_v16hf_mask ((B), (C), (D), (A))) > > > +#define _mm256_mask_cmp_ph_mask(A, B, C, D) \ > > > + (__builtin_ia32_cmpph256_mask ((B), (C), (D), (A))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > @@ -363,46 +363,46 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_sqrt_ph (__m128h __A) > > > { > > > - return __builtin_ia32_vsqrtph_v8hf_mask (__A, _mm_setzero_ph (), > > > - (__mmask8) -1); > > > + return __builtin_ia32_sqrtph128_mask (__A, _mm_setzero_ph (), > > > + (__mmask8) -1); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_sqrt_ph (__m256h __A) > > > { > > > - return __builtin_ia32_vsqrtph_v16hf_mask (__A, _mm256_setzero_ph (), > > > - (__mmask16) -1); > > > + return __builtin_ia32_sqrtph256_mask (__A, _mm256_setzero_ph (), > > > + (__mmask16) -1); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_sqrt_ph (__m128h __A, __mmask8 __B, __m128h __C) > > > { > > > - return __builtin_ia32_vsqrtph_v8hf_mask (__C, __A, __B); > > > + return __builtin_ia32_sqrtph128_mask (__C, __A, __B); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_sqrt_ph (__m256h __A, __mmask16 __B, __m256h __C) > > > { > > > - return __builtin_ia32_vsqrtph_v16hf_mask (__C, __A, __B); > > > + return __builtin_ia32_sqrtph256_mask (__C, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_sqrt_ph (__mmask8 __A, __m128h __B) > > > { > > > - return __builtin_ia32_vsqrtph_v8hf_mask (__B, _mm_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_sqrtph128_mask (__B, _mm_setzero_ph (), > > > + __A); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_sqrt_ph (__mmask16 __A, __m256h __B) > > > { > > > - return __builtin_ia32_vsqrtph_v16hf_mask (__B, _mm256_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_sqrtph256_mask (__B, _mm256_setzero_ph (), > > > + __A); > > > } > > > > > > /* Intrinsics vrsqrtph. */ > > > @@ -410,45 +410,45 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_rsqrt_ph (__m128h __A) > > > { > > > - return __builtin_ia32_vrsqrtph_v8hf_mask (__A, _mm_setzero_ph (), > > > - (__mmask8) -1); > > > + return __builtin_ia32_rsqrtph128_mask (__A, _mm_setzero_ph (), > > > + (__mmask8) -1); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_rsqrt_ph (__m256h __A) > > > { > > > - return __builtin_ia32_vrsqrtph_v16hf_mask (__A, _mm256_setzero_ph (), > > > - (__mmask16) -1); > > > + return __builtin_ia32_rsqrtph256_mask (__A, _mm256_setzero_ph (), > > > + (__mmask16) -1); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_rsqrt_ph (__m128h __A, __mmask8 __B, __m128h __C) > > > { > > > - return __builtin_ia32_vrsqrtph_v8hf_mask (__C, __A, __B); > > > + return __builtin_ia32_rsqrtph128_mask (__C, __A, __B); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_rsqrt_ph (__m256h __A, __mmask16 __B, __m256h __C) > > > { > > > - return __builtin_ia32_vrsqrtph_v16hf_mask (__C, __A, __B); > > > + return __builtin_ia32_rsqrtph256_mask (__C, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_rsqrt_ph (__mmask8 __A, __m128h __B) > > > { > > > - return __builtin_ia32_vrsqrtph_v8hf_mask (__B, _mm_setzero_ph (), __A); > > > + return __builtin_ia32_rsqrtph128_mask (__B, _mm_setzero_ph (), __A); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_rsqrt_ph (__mmask16 __A, __m256h __B) > > > { > > > - return __builtin_ia32_vrsqrtph_v16hf_mask (__B, _mm256_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_rsqrtph256_mask (__B, _mm256_setzero_ph (), > > > + __A); > > > } > > > > > > /* Intrinsics vrcpph. */ > > > @@ -456,45 +456,45 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_rcp_ph (__m128h __A) > > > { > > > - return __builtin_ia32_vrcpph_v8hf_mask (__A, _mm_setzero_ph (), > > > - (__mmask8) -1); > > > + return __builtin_ia32_rcpph128_mask (__A, _mm_setzero_ph (), > > > + (__mmask8) -1); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_rcp_ph (__m256h __A) > > > { > > > - return __builtin_ia32_vrcpph_v16hf_mask (__A, _mm256_setzero_ph (), > > > - (__mmask16) -1); > > > + return __builtin_ia32_rcpph256_mask (__A, _mm256_setzero_ph (), > > > + (__mmask16) -1); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_rcp_ph (__m128h __A, __mmask8 __B, __m128h __C) > > > { > > > - return __builtin_ia32_vrcpph_v8hf_mask (__C, __A, __B); > > > + return __builtin_ia32_rcpph128_mask (__C, __A, __B); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_rcp_ph (__m256h __A, __mmask16 __B, __m256h __C) > > > { > > > - return __builtin_ia32_vrcpph_v16hf_mask (__C, __A, __B); > > > + return __builtin_ia32_rcpph256_mask (__C, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_rcp_ph (__mmask8 __A, __m128h __B) > > > { > > > - return __builtin_ia32_vrcpph_v8hf_mask (__B, _mm_setzero_ph (), __A); > > > + return __builtin_ia32_rcpph128_mask (__B, _mm_setzero_ph (), __A); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_rcp_ph (__mmask16 __A, __m256h __B) > > > { > > > - return __builtin_ia32_vrcpph_v16hf_mask (__B, _mm256_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_rcpph256_mask (__B, _mm256_setzero_ph (), > > > + __A); > > > } > > > > > > /* Intrinsics vscalefph. */ > > > @@ -502,25 +502,25 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_scalef_ph (__m128h __A, __m128h __B) > > > { > > > - return __builtin_ia32_vscalefph_v8hf_mask (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1); > > > + return __builtin_ia32_scalefph128_mask (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_scalef_ph (__m256h __A, __m256h __B) > > > { > > > - return __builtin_ia32_vscalefph_v16hf_mask (__A, __B, > > > - _mm256_setzero_ph (), > > > - (__mmask16) -1); > > > + return __builtin_ia32_scalefph256_mask (__A, __B, > > > + _mm256_setzero_ph (), > > > + (__mmask16) -1); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_scalef_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) > > > { > > > - return __builtin_ia32_vscalefph_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_scalefph128_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m256h > > > @@ -528,24 +528,24 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_scalef_ph (__m256h __A, __mmask16 __B, __m256h __C, > > > __m256h __D) > > > { > > > - return __builtin_ia32_vscalefph_v16hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_scalefph256_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_scalef_ph (__mmask8 __A, __m128h __B, __m128h __C) > > > { > > > - return __builtin_ia32_vscalefph_v8hf_mask (__B, __C, > > > - _mm_setzero_ph (), __A); > > > + return __builtin_ia32_scalefph128_mask (__B, __C, > > > + _mm_setzero_ph (), __A); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_scalef_ph (__mmask16 __A, __m256h __B, __m256h __C) > > > { > > > - return __builtin_ia32_vscalefph_v16hf_mask (__B, __C, > > > - _mm256_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_scalefph256_mask (__B, __C, > > > + _mm256_setzero_ph (), > > > + __A); > > > } > > > > > > /* Intrinsics vreduceph. */ > > > @@ -554,109 +554,109 @@ extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_reduce_ph (__m128h __A, int __B) > > > { > > > - return __builtin_ia32_vreduceph_v8hf_mask (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1); > > > + return __builtin_ia32_reduceph128_mask (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_reduce_ph (__m128h __A, __mmask8 __B, __m128h __C, int __D) > > > { > > > - return __builtin_ia32_vreduceph_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_reduceph128_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_reduce_ph (__mmask8 __A, __m128h __B, int __C) > > > { > > > - return __builtin_ia32_vreduceph_v8hf_mask (__B, __C, > > > - _mm_setzero_ph (), __A); > > > + return __builtin_ia32_reduceph128_mask (__B, __C, > > > + _mm_setzero_ph (), __A); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_reduce_ph (__m256h __A, int __B) > > > { > > > - return __builtin_ia32_vreduceph_v16hf_mask (__A, __B, > > > - _mm256_setzero_ph (), > > > - (__mmask16) -1); > > > + return __builtin_ia32_reduceph256_mask (__A, __B, > > > + _mm256_setzero_ph (), > > > + (__mmask16) -1); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_reduce_ph (__m256h __A, __mmask16 __B, __m256h __C, int __D) > > > { > > > - return __builtin_ia32_vreduceph_v16hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_reduceph256_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_reduce_ph (__mmask16 __A, __m256h __B, int __C) > > > { > > > - return __builtin_ia32_vreduceph_v16hf_mask (__B, __C, > > > - _mm256_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_reduceph256_mask (__B, __C, > > > + _mm256_setzero_ph (), > > > + __A); > > > } > > > > > > #else > > > -#define _mm_reduce_ph(A, B) \ > > > - (__builtin_ia32_vreduceph_v8hf_mask ((A), (B),\ > > > - _mm_setzero_ph (), \ > > > - ((__mmask8)-1))) > > > +#define _mm_reduce_ph(A, B) \ > > > + (__builtin_ia32_reduceph128_mask ((A), (B), \ > > > + _mm_setzero_ph (), \ > > > + ((__mmask8)-1))) > > > > > > -#define _mm_mask_reduce_ph(A, B, C, D) \ > > > - (__builtin_ia32_vreduceph_v8hf_mask ((C), (D), (A), (B))) > > > +#define _mm_mask_reduce_ph(A, B, C, D) \ > > > + (__builtin_ia32_reduceph128_mask ((C), (D), (A), (B))) > > > > > > -#define _mm_maskz_reduce_ph(A, B, C) \ > > > - (__builtin_ia32_vreduceph_v8hf_mask ((B), (C), _mm_setzero_ph (), (A))) > > > +#define _mm_maskz_reduce_ph(A, B, C) \ > > > + (__builtin_ia32_reduceph128_mask ((B), (C), _mm_setzero_ph (), (A))) > > > > > > #define _mm256_reduce_ph(A, B) \ > > > - (__builtin_ia32_vreduceph_v16hf_mask ((A), (B),\ > > > - _mm256_setzero_ph (), \ > > > - ((__mmask16)-1))) > > > + (__builtin_ia32_reduceph256_mask ((A), (B), \ > > > + _mm256_setzero_ph (), \ > > > + ((__mmask16)-1))) > > > > > > -#define _mm256_mask_reduce_ph(A, B, C, D) \ > > > - (__builtin_ia32_vreduceph_v16hf_mask ((C), (D), (A), (B))) > > > +#define _mm256_mask_reduce_ph(A, B, C, D) \ > > > + (__builtin_ia32_reduceph256_mask ((C), (D), (A), (B))) > > > > > > -#define _mm256_maskz_reduce_ph(A, B, C) \ > > > - (__builtin_ia32_vreduceph_v16hf_mask ((B), (C), _mm256_setzero_ph (), (A))) > > > +#define _mm256_maskz_reduce_ph(A, B, C) \ > > > + (__builtin_ia32_reduceph256_mask ((B), (C), _mm256_setzero_ph (), (A))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > /* Intrinsics vrndscaleph. */ > > > #ifdef __OPTIMIZE__ > > > -extern __inline __m128h > > > -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > -_mm_roundscale_ph (__m128h __A, int __B) > > > -{ > > > - return __builtin_ia32_vrndscaleph_v8hf_mask (__A, __B, > > > - _mm_setzero_ph (), > > > - (__mmask8) -1); > > > -} > > > + extern __inline __m128h > > > + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > + _mm_roundscale_ph (__m128h __A, int __B) > > > + { > > > + return __builtin_ia32_rndscaleph128_mask (__A, __B, > > > + _mm_setzero_ph (), > > > + (__mmask8) -1); > > > + } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_roundscale_ph (__m128h __A, __mmask8 __B, __m128h __C, int __D) > > > { > > > - return __builtin_ia32_vrndscaleph_v8hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_rndscaleph128_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m128h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_roundscale_ph (__mmask8 __A, __m128h __B, int __C) > > > { > > > - return __builtin_ia32_vrndscaleph_v8hf_mask (__B, __C, > > > - _mm_setzero_ph (), __A); > > > + return __builtin_ia32_rndscaleph128_mask (__B, __C, > > > + _mm_setzero_ph (), __A); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_roundscale_ph (__m256h __A, int __B) > > > { > > > - return __builtin_ia32_vrndscaleph_v16hf_mask (__A, __B, > > > - _mm256_setzero_ph (), > > > - (__mmask16) -1); > > > + return __builtin_ia32_rndscaleph256_mask (__A, __B, > > > + _mm256_setzero_ph (), > > > + (__mmask16) -1); > > > } > > > > > > extern __inline __m256h > > > @@ -664,40 +664,40 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_roundscale_ph (__m256h __A, __mmask16 __B, __m256h __C, > > > int __D) > > > { > > > - return __builtin_ia32_vrndscaleph_v16hf_mask (__C, __D, __A, __B); > > > + return __builtin_ia32_rndscaleph256_mask (__C, __D, __A, __B); > > > } > > > > > > extern __inline __m256h > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_roundscale_ph (__mmask16 __A, __m256h __B, int __C) > > > { > > > - return __builtin_ia32_vrndscaleph_v16hf_mask (__B, __C, > > > - _mm256_setzero_ph (), > > > - __A); > > > + return __builtin_ia32_rndscaleph256_mask (__B, __C, > > > + _mm256_setzero_ph (), > > > + __A); > > > } > > > > > > #else > > > -#define _mm_roundscale_ph(A, B) \ > > > - (__builtin_ia32_vrndscaleph_v8hf_mask ((A), (B), _mm_setzero_ph (), \ > > > - ((__mmask8)-1))) > > > +#define _mm_roundscale_ph(A, B) \ > > > + (__builtin_ia32_rndscaleph128_mask ((A), (B), _mm_setzero_ph (), \ > > > + ((__mmask8)-1))) > > > > > > -#define _mm_mask_roundscale_ph(A, B, C, D) \ > > > - (__builtin_ia32_vrndscaleph_v8hf_mask ((C), (D), (A), (B))) > > > +#define _mm_mask_roundscale_ph(A, B, C, D) \ > > > + (__builtin_ia32_rndscaleph128_mask ((C), (D), (A), (B))) > > > > > > -#define _mm_maskz_roundscale_ph(A, B, C) \ > > > - (__builtin_ia32_vrndscaleph_v8hf_mask ((B), (C), _mm_setzero_ph (), (A))) > > > +#define _mm_maskz_roundscale_ph(A, B, C) \ > > > + (__builtin_ia32_rndscaleph128_mask ((B), (C), _mm_setzero_ph (), (A))) > > > > > > -#define _mm256_roundscale_ph(A, B) \ > > > - (__builtin_ia32_vrndscaleph_v16hf_mask ((A), (B), \ > > > - _mm256_setzero_ph(), \ > > > - ((__mmask16)-1))) > > > +#define _mm256_roundscale_ph(A, B) \ > > > + (__builtin_ia32_rndscaleph256_mask ((A), (B), \ > > > + _mm256_setzero_ph(), \ > > > + ((__mmask16)-1))) > > > > > > -#define _mm256_mask_roundscale_ph(A, B, C, D) \ > > > - (__builtin_ia32_vrndscaleph_v16hf_mask ((C), (D), (A), (B))) > > > +#define _mm256_mask_roundscale_ph(A, B, C, D) \ > > > + (__builtin_ia32_rndscaleph256_mask ((C), (D), (A), (B))) > > > > > > -#define _mm256_maskz_roundscale_ph(A, B, C) \ > > > - (__builtin_ia32_vrndscaleph_v16hf_mask ((B), (C), \ > > > - _mm256_setzero_ph (), (A))) > > > +#define _mm256_maskz_roundscale_ph(A, B, C) \ > > > + (__builtin_ia32_rndscaleph256_mask ((B), (C), \ > > > + _mm256_setzero_ph (), (A))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > @@ -705,7 +705,7 @@ _mm256_maskz_roundscale_ph (__mmask16 __A, __m256h __B, int __C) > > > #ifdef __OPTIMIZE__ > > > extern __inline __mmask8 > > > __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > -_mm_mask_fpclass_ph_mask (__mmask8 __U, __m128h __A, const int __imm) > > > + _mm_mask_fpclass_ph_mask (__mmask8 __U, __m128h __A, const int __imm) > > > { > > > return (__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) __A, > > > __imm, __U); > > > @@ -725,7 +725,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_fpclass_ph_mask (__mmask16 __U, __m256h __A, const int __imm) > > > { > > > return (__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) __A, > > > - __imm, __U); > > > + __imm, __U); > > > } > > > > > > extern __inline __mmask16 > > > @@ -733,26 +733,26 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_fpclass_ph_mask (__m256h __A, const int __imm) > > > { > > > return (__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) __A, > > > - __imm, > > > - (__mmask16) -1); > > > + __imm, > > > + (__mmask16) -1); > > > } > > > > > > #else > > > #define _mm_fpclass_ph_mask(X, C) \ > > > - ((__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) (__m128h) (X), \ > > > + ((__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) (__m128h) (X), \ > > > (int) (C),(__mmask8)-1)) > > > > > > #define _mm_mask_fpclass_ph_mask(u, X, C) \ > > > - ((__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) (__m128h) (X), \ > > > + ((__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) (__m128h) (X), \ > > > (int) (C),(__mmask8)(u))) > > > > > > #define _mm256_fpclass_ph_mask(X, C) \ > > > - ((__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) (__m256h) (X), \ > > > - (int) (C),(__mmask16)-1)) > > > + ((__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) (__m256h) (X), \ > > > + (int) (C),(__mmask16)-1)) > > > > > > #define _mm256_mask_fpclass_ph_mask(u, X, C) \ > > > - ((__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) (__m256h) (X), \ > > > - (int) (C),(__mmask16)(u))) > > > + ((__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) (__m256h) (X), \ > > > + (int) (C),(__mmask16)(u))) > > > #endif /* __OPTIMIZE__ */ > > > > > > /* Intrinsics vgetexpph, vgetexpsh. */ > > > @@ -761,9 +761,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_getexp_ph (__m256h __A) > > > { > > > return (__m256h) __builtin_ia32_getexpph256_mask ((__v16hf) __A, > > > - (__v16hf) > > > - _mm256_setzero_ph (), > > > - (__mmask16) -1); > > > + (__v16hf) > > > + _mm256_setzero_ph (), > > > + (__mmask16) -1); > > > } > > > > > > extern __inline __m256h > > > @@ -771,8 +771,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_mask_getexp_ph (__m256h __W, __mmask16 __U, __m256h __A) > > > { > > > return (__m256h) __builtin_ia32_getexpph256_mask ((__v16hf) __A, > > > - (__v16hf) __W, > > > - (__mmask16) __U); > > > + (__v16hf) __W, > > > + (__mmask16) __U); > > > } > > > > > > extern __inline __m256h > > > @@ -780,9 +780,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm256_maskz_getexp_ph (__mmask16 __U, __m256h __A) > > > { > > > return (__m256h) __builtin_ia32_getexpph256_mask ((__v16hf) __A, > > > - (__v16hf) > > > - _mm256_setzero_ph (), > > > - (__mmask16) __U); > > > + (__v16hf) > > > + _mm256_setzero_ph (), > > > + (__mmask16) __U); > > > } > > > > > > extern __inline __m128h > > > @@ -790,9 +790,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_getexp_ph (__m128h __A) > > > { > > > return (__m128h) __builtin_ia32_getexpph128_mask ((__v8hf) __A, > > > - (__v8hf) > > > - _mm_setzero_ph (), > > > - (__mmask8) -1); > > > + (__v8hf) > > > + _mm_setzero_ph (), > > > + (__mmask8) -1); > > > } > > > > > > extern __inline __m128h > > > @@ -800,8 +800,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_mask_getexp_ph (__m128h __W, __mmask8 __U, __m128h __A) > > > { > > > return (__m128h) __builtin_ia32_getexpph128_mask ((__v8hf) __A, > > > - (__v8hf) __W, > > > - (__mmask8) __U); > > > + (__v8hf) __W, > > > + (__mmask8) __U); > > > } > > > > > > extern __inline __m128h > > > @@ -809,9 +809,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) > > > _mm_maskz_getexp_ph (__mmask8 __U, __m128h __A) > > > { > > > return (__m128h) __builtin_ia32_getexpph128_mask ((__v8hf) __A, > > > - (__v8hf) > > > - _mm_setzero_ph (), > > > - (__mmask8) __U); > > > + (__v8hf) > > > + _mm_setzero_ph (), > > > + (__mmask8) __U); > > > } > > > > > > > > > @@ -892,41 +892,41 @@ _mm_maskz_getmant_ph (__mmask8 __U, __m128h __A, > > > } > > > > > > #else > > > -#define _mm256_getmant_ph(X, B, C) \ > > > - ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ > > > - (int)(((C)<<2) | (B)), \ > > > - (__v16hf)(__m256h)_mm256_setzero_ph (),\ > > > - (__mmask16)-1)) > > > - > > > -#define _mm256_mask_getmant_ph(W, U, X, B, C) \ > > > - ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ > > > - (int)(((C)<<2) | (B)), \ > > > - (__v16hf)(__m256h)(W), \ > > > - (__mmask16)(U))) > > > - > > > -#define _mm256_maskz_getmant_ph(U, X, B, C) \ > > > - ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ > > > - (int)(((C)<<2) | (B)), \ > > > - (__v16hf)(__m256h)_mm256_setzero_ph (),\ > > > - (__mmask16)(U))) > > > - > > > -#define _mm_getmant_ph(X, B, C) \ > > > - ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ > > > - (int)(((C)<<2) | (B)), \ > > > - (__v8hf)(__m128h)_mm_setzero_ph (), \ > > > - (__mmask8)-1)) > > > - > > > -#define _mm_mask_getmant_ph(W, U, X, B, C) \ > > > - ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ > > > - (int)(((C)<<2) | (B)), \ > > > - (__v8hf)(__m128h)(W), \ > > > - (__mmask8)(U))) > > > - > > > -#define _mm_maskz_getmant_ph(U, X, B, C) \ > > > - ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ > > > - (int)(((C)<<2) | (B)), \ > > > - (__v8hf)(__m128h)_mm_setzero_ph (), \ > > > - (__mmask8)(U))) > > > +#define _mm256_getmant_ph(X, B, C) \ > > > + ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ > > > + (int)(((C)<<2) | (B)), \ > > > + (__v16hf)(__m256h)_mm256_setzero_ph (), \ > > > + (__mmask16)-1)) > > > + > > > +#define _mm256_mask_getmant_ph(W, U, X, B, C) \ > > > + ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ > > > + (int)(((C)<<2) | (B)), \ > > > + (__v16hf)(__m256h)(W), \ > > > + (__mmask16)(U))) > > > + > > > +#define _mm256_maskz_getmant_ph(U, X, B, C) \ > > > + ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ > > > + (int)(((C)<<2) | (B)), \ > > > + (__v16hf)(__m256h)_mm256_setzero_ph (), \ > > > + (__mmask16)(U))) > > > + > > > +#define _mm_getmant_ph(X, B, C) \ > > > + ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ > > > + (int)(((C)<<2) | (B)), \ > > > + (__v8hf)(__m128h)_mm_setzero_ph (), \ > > > + (__mmask8)-1)) > > > + > > > +#define _mm_mask_getmant_ph(W, U, X, B, C) \ > > > + ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ > > > + (int)(((C)<<2) | (B)), \ > > > + (__v8hf)(__m128h)(W), \ > > > + (__mmask8)(U))) > > > + > > > +#define _mm_maskz_getmant_ph(U, X, B, C) \ > > > + ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ > > > + (int)(((C)<<2) | (B)), \ > > > + (__v8hf)(__m128h)_mm_setzero_ph (), \ > > > + (__mmask8)(U))) > > > > > > #endif /* __OPTIMIZE__ */ > > > > > > diff --git a/gcc/config/i386/i386-builtin.def b/gcc/config/i386/i386-builtin.def > > > index bfa1d56a7b8..10f6fd87cbb 100644 > > > --- a/gcc/config/i386/i386-builtin.def > > > +++ b/gcc/config/i386/i386-builtin.def > > > @@ -2775,49 +2775,49 @@ BDESC (0, OPTION_MASK_ISA2_AVX512BF16, CODE_FOR_avx512f_dpbf16ps_v4sf_mask, "__b > > > BDESC (0, OPTION_MASK_ISA2_AVX512BF16, CODE_FOR_avx512f_dpbf16ps_v4sf_maskz, "__builtin_ia32_dpbf16ps_v4sf_maskz", IX86_BUILTIN_DPHI16PS_V4SF_MASKZ, UNKNOWN, (int) V4SF_FTYPE_V4SF_V8HI_V8HI_UQI) > > > > > > /* AVX512FP16. */ > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv8hf3_mask, "__builtin_ia32_vaddph_v8hf_mask", IX86_BUILTIN_VADDPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv16hf3_mask, "__builtin_ia32_vaddph_v16hf_mask", IX86_BUILTIN_VADDPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv32hf3_mask, "__builtin_ia32_vaddph_v32hf_mask", IX86_BUILTIN_VADDPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv8hf3_mask, "__builtin_ia32_vsubph_v8hf_mask", IX86_BUILTIN_VSUBPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv16hf3_mask, "__builtin_ia32_vsubph_v16hf_mask", IX86_BUILTIN_VSUBPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv32hf3_mask, "__builtin_ia32_vsubph_v32hf_mask", IX86_BUILTIN_VSUBPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv8hf3_mask, "__builtin_ia32_vmulph_v8hf_mask", IX86_BUILTIN_VMULPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv16hf3_mask, "__builtin_ia32_vmulph_v16hf_mask", IX86_BUILTIN_VMULPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv32hf3_mask, "__builtin_ia32_vmulph_v32hf_mask", IX86_BUILTIN_VMULPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv8hf3_mask, "__builtin_ia32_vdivph_v8hf_mask", IX86_BUILTIN_VDIVPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv16hf3_mask, "__builtin_ia32_vdivph_v16hf_mask", IX86_BUILTIN_VDIVPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv32hf3_mask, "__builtin_ia32_vdivph_v32hf_mask", IX86_BUILTIN_VDIVPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmaddv8hf3_mask, "__builtin_ia32_vaddsh_v8hf_mask", IX86_BUILTIN_VADDSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsubv8hf3_mask, "__builtin_ia32_vsubsh_v8hf_mask", IX86_BUILTIN_VSUBSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmmulv8hf3_mask, "__builtin_ia32_vmulsh_v8hf_mask", IX86_BUILTIN_VMULSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmdivv8hf3_mask, "__builtin_ia32_vdivsh_v8hf_mask", IX86_BUILTIN_VDIVSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv8hf3_mask, "__builtin_ia32_vmaxph_v8hf_mask", IX86_BUILTIN_VMAXPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv16hf3_mask, "__builtin_ia32_vmaxph_v16hf_mask", IX86_BUILTIN_VMAXPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv32hf3_mask, "__builtin_ia32_vmaxph_v32hf_mask", IX86_BUILTIN_VMAXPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv8hf3_mask, "__builtin_ia32_vminph_v8hf_mask", IX86_BUILTIN_VMINPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv16hf3_mask, "__builtin_ia32_vminph_v16hf_mask", IX86_BUILTIN_VMINPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv32hf3_mask, "__builtin_ia32_vminph_v32hf_mask", IX86_BUILTIN_VMINPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask, "__builtin_ia32_vmaxsh_v8hf_mask", IX86_BUILTIN_VMAXSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask, "__builtin_ia32_vminsh_v8hf_mask", IX86_BUILTIN_VMINSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_cmpv8hf3_mask, "__builtin_ia32_vcmpph_v8hf_mask", IX86_BUILTIN_VCMPPH_V8HF_MASK, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_cmpv16hf3_mask, "__builtin_ia32_vcmpph_v16hf_mask", IX86_BUILTIN_VCMPPH_V16HF_MASK, UNKNOWN, (int) UHI_FTYPE_V16HF_V16HF_INT_UHI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask, "__builtin_ia32_vcmpph_v32hf_mask", IX86_BUILTIN_VCMPPH_V32HF_MASK, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv8hf2_mask, "__builtin_ia32_vsqrtph_v8hf_mask", IX86_BUILTIN_VSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv16hf2_mask, "__builtin_ia32_vsqrtph_v16hf_mask", IX86_BUILTIN_VSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv8hf2_mask, "__builtin_ia32_vrsqrtph_v8hf_mask", IX86_BUILTIN_VRSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv16hf2_mask, "__builtin_ia32_vrsqrtph_v16hf_mask", IX86_BUILTIN_VRSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv32hf2_mask, "__builtin_ia32_vrsqrtph_v32hf_mask", IX86_BUILTIN_VRSQRTPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrsqrtv8hf2_mask, "__builtin_ia32_vrsqrtsh_v8hf_mask", IX86_BUILTIN_VRSQRTSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv8hf2_mask, "__builtin_ia32_vrcpph_v8hf_mask", IX86_BUILTIN_VRCPPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv16hf2_mask, "__builtin_ia32_vrcpph_v16hf_mask", IX86_BUILTIN_VRCPPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv32hf2_mask, "__builtin_ia32_vrcpph_v32hf_mask", IX86_BUILTIN_VRCPPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrcpv8hf2_mask, "__builtin_ia32_vrcpsh_v8hf_mask", IX86_BUILTIN_VRCPSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_scalefv8hf_mask, "__builtin_ia32_vscalefph_v8hf_mask", IX86_BUILTIN_VSCALEFPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_scalefv16hf_mask, "__builtin_ia32_vscalefph_v16hf_mask", IX86_BUILTIN_VSCALEFPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv8hf_mask, "__builtin_ia32_vreduceph_v8hf_mask", IX86_BUILTIN_VREDUCEPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_INT_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv16hf_mask, "__builtin_ia32_vreduceph_v16hf_mask", IX86_BUILTIN_VREDUCEPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_INT_V16HF_UHI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rndscalev8hf_mask, "__builtin_ia32_vrndscaleph_v8hf_mask", IX86_BUILTIN_VRNDSCALEPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_INT_V8HF_UQI) > > > -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_rndscalev16hf_mask, "__builtin_ia32_vrndscaleph_v16hf_mask", IX86_BUILTIN_VRNDSCALEPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_INT_V16HF_UHI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv8hf3_mask, "__builtin_ia32_addph128_mask", IX86_BUILTIN_VADDPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv16hf3_mask, "__builtin_ia32_addph256_mask", IX86_BUILTIN_VADDPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv32hf3_mask, "__builtin_ia32_addph512_mask", IX86_BUILTIN_VADDPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv8hf3_mask, "__builtin_ia32_subph128_mask", IX86_BUILTIN_VSUBPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv16hf3_mask, "__builtin_ia32_subph256_mask", IX86_BUILTIN_VSUBPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv32hf3_mask, "__builtin_ia32_subph512_mask", IX86_BUILTIN_VSUBPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv8hf3_mask, "__builtin_ia32_mulph128_mask", IX86_BUILTIN_VMULPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv16hf3_mask, "__builtin_ia32_mulph256_mask", IX86_BUILTIN_VMULPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv32hf3_mask, "__builtin_ia32_mulph512_mask", IX86_BUILTIN_VMULPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv8hf3_mask, "__builtin_ia32_divph128_mask", IX86_BUILTIN_VDIVPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv16hf3_mask, "__builtin_ia32_divph256_mask", IX86_BUILTIN_VDIVPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv32hf3_mask, "__builtin_ia32_divph512_mask", IX86_BUILTIN_VDIVPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmaddv8hf3_mask, "__builtin_ia32_addsh_mask", IX86_BUILTIN_VADDSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsubv8hf3_mask, "__builtin_ia32_subsh_mask", IX86_BUILTIN_VSUBSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmmulv8hf3_mask, "__builtin_ia32_mulsh_mask", IX86_BUILTIN_VMULSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmdivv8hf3_mask, "__builtin_ia32_divsh_mask", IX86_BUILTIN_VDIVSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv8hf3_mask, "__builtin_ia32_maxph128_mask", IX86_BUILTIN_VMAXPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv16hf3_mask, "__builtin_ia32_maxph256_mask", IX86_BUILTIN_VMAXPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv32hf3_mask, "__builtin_ia32_maxph512_mask", IX86_BUILTIN_VMAXPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv8hf3_mask, "__builtin_ia32_minph128_mask", IX86_BUILTIN_VMINPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv16hf3_mask, "__builtin_ia32_minph256_mask", IX86_BUILTIN_VMINPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv32hf3_mask, "__builtin_ia32_minph512_mask", IX86_BUILTIN_VMINPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask, "__builtin_ia32_maxsh_mask", IX86_BUILTIN_VMAXSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask, "__builtin_ia32_minsh_mask", IX86_BUILTIN_VMINSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_cmpv8hf3_mask, "__builtin_ia32_cmpph128_mask", IX86_BUILTIN_VCMPPH_V8HF_MASK, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_cmpv16hf3_mask, "__builtin_ia32_cmpph256_mask", IX86_BUILTIN_VCMPPH_V16HF_MASK, UNKNOWN, (int) UHI_FTYPE_V16HF_V16HF_INT_UHI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask, "__builtin_ia32_cmpph512_mask", IX86_BUILTIN_VCMPPH_V32HF_MASK, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv8hf2_mask, "__builtin_ia32_sqrtph128_mask", IX86_BUILTIN_VSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv16hf2_mask, "__builtin_ia32_sqrtph256_mask", IX86_BUILTIN_VSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv8hf2_mask, "__builtin_ia32_rsqrtph128_mask", IX86_BUILTIN_VRSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv16hf2_mask, "__builtin_ia32_rsqrtph256_mask", IX86_BUILTIN_VRSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv32hf2_mask, "__builtin_ia32_rsqrtph512_mask", IX86_BUILTIN_VRSQRTPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrsqrtv8hf2_mask, "__builtin_ia32_rsqrtsh_mask", IX86_BUILTIN_VRSQRTSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv8hf2_mask, "__builtin_ia32_rcpph128_mask", IX86_BUILTIN_VRCPPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv16hf2_mask, "__builtin_ia32_rcpph256_mask", IX86_BUILTIN_VRCPPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv32hf2_mask, "__builtin_ia32_rcpph512_mask", IX86_BUILTIN_VRCPPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrcpv8hf2_mask, "__builtin_ia32_rcpsh_mask", IX86_BUILTIN_VRCPSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_scalefv8hf_mask, "__builtin_ia32_scalefph128_mask", IX86_BUILTIN_VSCALEFPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_scalefv16hf_mask, "__builtin_ia32_scalefph256_mask", IX86_BUILTIN_VSCALEFPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv8hf_mask, "__builtin_ia32_reduceph128_mask", IX86_BUILTIN_VREDUCEPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_INT_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv16hf_mask, "__builtin_ia32_reduceph256_mask", IX86_BUILTIN_VREDUCEPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_INT_V16HF_UHI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rndscalev8hf_mask, "__builtin_ia32_rndscaleph128_mask", IX86_BUILTIN_VRNDSCALEPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_INT_V8HF_UQI) > > > +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_rndscalev16hf_mask, "__builtin_ia32_rndscaleph256_mask", IX86_BUILTIN_VRNDSCALEPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_INT_V16HF_UHI) > > > BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512dq_fpclassv16hf_mask, "__builtin_ia32_fpclassph256_mask", IX86_BUILTIN_FPCLASSPH256, UNKNOWN, (int) HI_FTYPE_V16HF_INT_UHI) > > > BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512dq_fpclassv8hf_mask, "__builtin_ia32_fpclassph128_mask", IX86_BUILTIN_FPCLASSPH128, UNKNOWN, (int) QI_FTYPE_V8HF_INT_UQI) > > > BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512dq_fpclassv32hf_mask, "__builtin_ia32_fpclassph512_mask", IX86_BUILTIN_FPCLASSPH512, UNKNOWN, (int) SI_FTYPE_V32HF_INT_USI) > > > @@ -3027,28 +3027,28 @@ BDESC (OPTION_MASK_ISA_AVX512DQ, 0, CODE_FOR_avx512dq_rangepv16sf_mask_round, "_ > > > BDESC (OPTION_MASK_ISA_AVX512DQ, 0, CODE_FOR_avx512dq_rangepv8df_mask_round, "__builtin_ia32_rangepd512_mask", IX86_BUILTIN_RANGEPD512, UNKNOWN, (int) V8DF_FTYPE_V8DF_V8DF_INT_V8DF_QI_INT) > > > > > > /* AVX512FP16. */ > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv32hf3_mask_round, "__builtin_ia32_vaddph_v32hf_mask_round", IX86_BUILTIN_VADDPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv32hf3_mask_round, "__builtin_ia32_vsubph_v32hf_mask_round", IX86_BUILTIN_VSUBPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv32hf3_mask_round, "__builtin_ia32_vmulph_v32hf_mask_round", IX86_BUILTIN_VMULPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv32hf3_mask_round, "__builtin_ia32_vdivph_v32hf_mask_round", IX86_BUILTIN_VDIVPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmaddv8hf3_mask_round, "__builtin_ia32_vaddsh_v8hf_mask_round", IX86_BUILTIN_VADDSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsubv8hf3_mask_round, "__builtin_ia32_vsubsh_v8hf_mask_round", IX86_BUILTIN_VSUBSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmmulv8hf3_mask_round, "__builtin_ia32_vmulsh_v8hf_mask_round", IX86_BUILTIN_VMULSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmdivv8hf3_mask_round, "__builtin_ia32_vdivsh_v8hf_mask_round", IX86_BUILTIN_VDIVSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv32hf3_mask_round, "__builtin_ia32_vmaxph_v32hf_mask_round", IX86_BUILTIN_VMAXPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv32hf3_mask_round, "__builtin_ia32_vminph_v32hf_mask_round", IX86_BUILTIN_VMINPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask_round, "__builtin_ia32_vmaxsh_v8hf_mask_round", IX86_BUILTIN_VMAXSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask_round, "__builtin_ia32_vminsh_v8hf_mask_round", IX86_BUILTIN_VMINSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask_round, "__builtin_ia32_vcmpph_v32hf_mask_round", IX86_BUILTIN_VCMPPH_V32HF_MASK_ROUND, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_vmcmpv8hf3_mask_round, "__builtin_ia32_vcmpsh_v8hf_mask_round", IX86_BUILTIN_VCMPSH_V8HF_MASK_ROUND, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv32hf2_mask_round, "__builtin_ia32_vsqrtph_v32hf_mask_round", IX86_BUILTIN_VSQRTPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsqrtv8hf2_mask_round, "__builtin_ia32_vsqrtsh_v8hf_mask_round", IX86_BUILTIN_VSQRTSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_scalefv32hf_mask_round, "__builtin_ia32_vscalefph_v32hf_mask_round", IX86_BUILTIN_VSCALEFPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_vmscalefv8hf_mask_round, "__builtin_ia32_vscalefsh_v8hf_mask_round", IX86_BUILTIN_VSCALEFSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv32hf_mask_round, "__builtin_ia32_vreduceph_v32hf_mask_round", IX86_BUILTIN_VREDUCEPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducesv8hf_mask_round, "__builtin_ia32_vreducesh_v8hf_mask_round", IX86_BUILTIN_VREDUCESH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_INT_V8HF_UQI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_rndscalev32hf_mask_round, "__builtin_ia32_vrndscaleph_v32hf_mask_round", IX86_BUILTIN_VRNDSCALEPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) > > > -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_rndscalev8hf_mask_round, "__builtin_ia32_vrndscalesh_v8hf_mask_round", IX86_BUILTIN_VRNDSCALESH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_INT_V8HF_UQI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv32hf3_mask_round, "__builtin_ia32_addph512_mask_round", IX86_BUILTIN_VADDPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv32hf3_mask_round, "__builtin_ia32_subph512_mask_round", IX86_BUILTIN_VSUBPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv32hf3_mask_round, "__builtin_ia32_mulph512_mask_round", IX86_BUILTIN_VMULPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv32hf3_mask_round, "__builtin_ia32_divph512_mask_round", IX86_BUILTIN_VDIVPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmaddv8hf3_mask_round, "__builtin_ia32_addsh_mask_round", IX86_BUILTIN_VADDSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsubv8hf3_mask_round, "__builtin_ia32_subsh_mask_round", IX86_BUILTIN_VSUBSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmmulv8hf3_mask_round, "__builtin_ia32_mulsh_mask_round", IX86_BUILTIN_VMULSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmdivv8hf3_mask_round, "__builtin_ia32_divsh_mask_round", IX86_BUILTIN_VDIVSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv32hf3_mask_round, "__builtin_ia32_maxph512_mask_round", IX86_BUILTIN_VMAXPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv32hf3_mask_round, "__builtin_ia32_minph512_mask_round", IX86_BUILTIN_VMINPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask_round, "__builtin_ia32_maxsh_mask_round", IX86_BUILTIN_VMAXSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask_round, "__builtin_ia32_minsh_mask_round", IX86_BUILTIN_VMINSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask_round, "__builtin_ia32_cmpph512_mask_round", IX86_BUILTIN_VCMPPH_V32HF_MASK_ROUND, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_vmcmpv8hf3_mask_round, "__builtin_ia32_cmpsh_mask_round", IX86_BUILTIN_VCMPSH_V8HF_MASK_ROUND, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv32hf2_mask_round, "__builtin_ia32_sqrtph512_mask_round", IX86_BUILTIN_VSQRTPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsqrtv8hf2_mask_round, "__builtin_ia32_sqrtsh_mask_round", IX86_BUILTIN_VSQRTSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_scalefv32hf_mask_round, "__builtin_ia32_scalefph512_mask_round", IX86_BUILTIN_VSCALEFPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_vmscalefv8hf_mask_round, "__builtin_ia32_scalefsh_mask_round", IX86_BUILTIN_VSCALEFSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv32hf_mask_round, "__builtin_ia32_reduceph512_mask_round", IX86_BUILTIN_VREDUCEPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducesv8hf_mask_round, "__builtin_ia32_reducesh_mask_round", IX86_BUILTIN_VREDUCESH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_INT_V8HF_UQI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_rndscalev32hf_mask_round, "__builtin_ia32_rndscaleph512_mask_round", IX86_BUILTIN_VRNDSCALEPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) > > > +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_rndscalev8hf_mask_round, "__builtin_ia32_rndscalesh_mask_round", IX86_BUILTIN_VRNDSCALESH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_INT_V8HF_UQI_INT) > > > BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_getexpv32hf_mask_round, "__builtin_ia32_getexpph512_mask", IX86_BUILTIN_GETEXPPH512, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI_INT) > > > BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_sgetexpv8hf_mask_round, "__builtin_ia32_getexpsh_mask_round", IX86_BUILTIN_GETEXPSH_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) > > > BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_getmantv32hf_mask_round, "__builtin_ia32_getmantph512_mask", IX86_BUILTIN_GETMANTPH512, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) > > > diff --git a/gcc/testsuite/gcc.target/i386/avx-1.c b/gcc/testsuite/gcc.target/i386/avx-1.c > > > index b3cffa0644f..3a96e586418 100644 > > > --- a/gcc/testsuite/gcc.target/i386/avx-1.c > > > +++ b/gcc/testsuite/gcc.target/i386/avx-1.c > > > @@ -686,33 +686,33 @@ > > > #define __builtin_ia32_vpshld_v2di_mask(A, B, C, D, E) __builtin_ia32_vpshld_v2di_mask(A, B, 1, D, E) > > > > > > /* avx512fp16intrin.h */ > > > -#define __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vcmpph_v32hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v32hf_mask(A, B, 1, D) > > > -#define __builtin_ia32_vcmpph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpph_v32hf_mask_round(A, B, 1, D, 8) > > > -#define __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, 1, D, 8) > > > -#define __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, D) __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, 8) > > > -#define __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, E) __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, 8) > > > -#define __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vreduceph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vreduceph_v32hf_mask_round(A, 123, C, D, 8) > > > -#define __builtin_ia32_vreduceph_v8hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v8hf_mask(A, 123, C, D) > > > -#define __builtin_ia32_vreduceph_v16hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v16hf_mask(A, 123, C, D) > > > -#define __builtin_ia32_vreducesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vreducesh_v8hf_mask_round(A, B, 123, D, E, 8) > > > -#define __builtin_ia32_vrndscaleph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vrndscaleph_v32hf_mask_round(A, 123, C, D, 8) > > > -#define __builtin_ia32_vrndscaleph_v8hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v8hf_mask(A, 123, C, D) > > > -#define __builtin_ia32_vrndscaleph_v16hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v16hf_mask(A, 123, C, D) > > > -#define __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, 123, D, E, 8) > > > +#define __builtin_ia32_addph512_mask_round(A, B, C, D, E) __builtin_ia32_addph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_subph512_mask_round(A, B, C, D, E) __builtin_ia32_subph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_mulph512_mask_round(A, B, C, D, E) __builtin_ia32_mulph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_divph512_mask_round(A, B, C, D, E) __builtin_ia32_divph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_addsh_mask_round(A, B, C, D, E) __builtin_ia32_addsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_subsh_mask_round(A, B, C, D, E) __builtin_ia32_subsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_mulsh_mask_round(A, B, C, D, E) __builtin_ia32_mulsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_divsh_mask_round(A, B, C, D, E) __builtin_ia32_divsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_maxph512_mask_round(A, B, C, D, E) __builtin_ia32_maxph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_minph512_mask_round(A, B, C, D, E) __builtin_ia32_minph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_maxsh_mask_round(A, B, C, D, E) __builtin_ia32_maxsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_minsh_mask_round(A, B, C, D, E) __builtin_ia32_minsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_cmpph512_mask(A, B, C, D) __builtin_ia32_cmpph512_mask(A, B, 1, D) > > > +#define __builtin_ia32_cmpph512_mask_round(A, B, C, D, E) __builtin_ia32_cmpph512_mask_round(A, B, 1, D, 8) > > > +#define __builtin_ia32_cmpsh_mask_round(A, B, C, D, E) __builtin_ia32_cmpsh_mask_round(A, B, 1, D, 8) > > > +#define __builtin_ia32_sqrtph512_mask_round(C, A, B, D) __builtin_ia32_sqrtph512_mask_round(C, A, B, 8) > > > +#define __builtin_ia32_sqrtsh_mask_round(D, C, A, B, E) __builtin_ia32_sqrtsh_mask_round(D, C, A, B, 8) > > > +#define __builtin_ia32_scalefph512_mask_round(A, B, C, D, E) __builtin_ia32_scalefph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_scalefsh_mask_round(A, B, C, D, E) __builtin_ia32_scalefsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_reduceph512_mask_round(A, B, C, D, E) __builtin_ia32_reduceph512_mask_round(A, 123, C, D, 8) > > > +#define __builtin_ia32_reduceph128_mask(A, B, C, D) __builtin_ia32_reduceph128_mask(A, 123, C, D) > > > +#define __builtin_ia32_reduceph256_mask(A, B, C, D) __builtin_ia32_reduceph256_mask(A, 123, C, D) > > > +#define __builtin_ia32_reducesh_mask_round(A, B, C, D, E, F) __builtin_ia32_reducesh_mask_round(A, B, 123, D, E, 8) > > > +#define __builtin_ia32_rndscaleph512_mask_round(A, B, C, D, E) __builtin_ia32_rndscaleph512_mask_round(A, 123, C, D, 8) > > > +#define __builtin_ia32_rndscaleph128_mask(A, B, C, D) __builtin_ia32_rndscaleph128_mask(A, 123, C, D) > > > +#define __builtin_ia32_rndscaleph256_mask(A, B, C, D) __builtin_ia32_rndscaleph256_mask(A, 123, C, D) > > > +#define __builtin_ia32_rndscalesh_mask_round(A, B, C, D, E, F) __builtin_ia32_rndscalesh_mask_round(A, B, 123, D, E, 8) > > > #define __builtin_ia32_fpclassph512_mask(A, D, C) __builtin_ia32_fpclassph512_mask(A, 1, C) > > > #define __builtin_ia32_fpclasssh_mask(A, D, U) __builtin_ia32_fpclasssh_mask(A, 1, U) > > > #define __builtin_ia32_getexpph512_mask(A, B, C, D) __builtin_ia32_getexpph512_mask(A, B, C, 8) > > > @@ -721,8 +721,8 @@ > > > #define __builtin_ia32_getmantsh_mask_round(A, B, C, W, U, D) __builtin_ia32_getmantsh_mask_round(A, B, 1, W, U, 4) > > > > > > /* avx512fp16vlintrin.h */ > > > -#define __builtin_ia32_vcmpph_v8hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v8hf_mask(A, B, 1, D) > > > -#define __builtin_ia32_vcmpph_v16hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v16hf_mask(A, B, 1, D) > > > +#define __builtin_ia32_cmpph128_mask(A, B, C, D) __builtin_ia32_cmpph128_mask(A, B, 1, D) > > > +#define __builtin_ia32_cmpph256_mask(A, B, C, D) __builtin_ia32_cmpph256_mask(A, B, 1, D) > > > #define __builtin_ia32_fpclassph256_mask(A, D, C) __builtin_ia32_fpclassph256_mask(A, 1, C) > > > #define __builtin_ia32_fpclassph128_mask(A, D, C) __builtin_ia32_fpclassph128_mask(A, 1, C) > > > #define __builtin_ia32_getmantph256_mask(A, E, C, D) __builtin_ia32_getmantph256_mask(A, 1, C, D) > > > diff --git a/gcc/testsuite/gcc.target/i386/sse-13.c b/gcc/testsuite/gcc.target/i386/sse-13.c > > > index 67ef567e437..aafcd414530 100644 > > > --- a/gcc/testsuite/gcc.target/i386/sse-13.c > > > +++ b/gcc/testsuite/gcc.target/i386/sse-13.c > > > @@ -703,33 +703,33 @@ > > > #define __builtin_ia32_vpshld_v2di_mask(A, B, C, D, E) __builtin_ia32_vpshld_v2di_mask(A, B, 1, D, E) > > > > > > /* avx512fp16intrin.h */ > > > -#define __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vcmpph_v32hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v32hf_mask(A, B, 1, D) > > > -#define __builtin_ia32_vcmpph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpph_v32hf_mask_round(A, B, 1, D, 8) > > > -#define __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, 1, D, 8) > > > -#define __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, D) __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, 8) > > > -#define __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, E) __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, 8) > > > -#define __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vreduceph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vreduceph_v32hf_mask_round(A, 123, C, D, 8) > > > -#define __builtin_ia32_vreduceph_v8hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v8hf_mask(A, 123, C, D) > > > -#define __builtin_ia32_vreduceph_v16hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v16hf_mask(A, 123, C, D) > > > -#define __builtin_ia32_vreducesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vreducesh_v8hf_mask_round(A, B, 123, D, E, 8) > > > -#define __builtin_ia32_vrndscaleph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vrndscaleph_v32hf_mask_round(A, 123, C, D, 8) > > > -#define __builtin_ia32_vrndscaleph_v8hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v8hf_mask(A, 123, C, D) > > > -#define __builtin_ia32_vrndscaleph_v16hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v16hf_mask(A, 123, C, D) > > > -#define __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, 123, D, E, 8) > > > +#define __builtin_ia32_addph512_mask_round(A, B, C, D, E) __builtin_ia32_addph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_subph512_mask_round(A, B, C, D, E) __builtin_ia32_subph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_mulph512_mask_round(A, B, C, D, E) __builtin_ia32_mulph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_divph512_mask_round(A, B, C, D, E) __builtin_ia32_divph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_addsh_mask_round(A, B, C, D, E) __builtin_ia32_addsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_subsh_mask_round(A, B, C, D, E) __builtin_ia32_subsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_mulsh_mask_round(A, B, C, D, E) __builtin_ia32_mulsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_divsh_mask_round(A, B, C, D, E) __builtin_ia32_divsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_maxph512_mask_round(A, B, C, D, E) __builtin_ia32_maxph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_minph512_mask_round(A, B, C, D, E) __builtin_ia32_minph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_maxsh_mask_round(A, B, C, D, E) __builtin_ia32_maxsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_minsh_mask_round(A, B, C, D, E) __builtin_ia32_minsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_cmpph512_mask(A, B, C, D) __builtin_ia32_cmpph512_mask(A, B, 1, D) > > > +#define __builtin_ia32_cmpph512_mask_round(A, B, C, D, E) __builtin_ia32_cmpph512_mask_round(A, B, 1, D, 8) > > > +#define __builtin_ia32_cmpsh_mask_round(A, B, C, D, E) __builtin_ia32_cmpsh_mask_round(A, B, 1, D, 8) > > > +#define __builtin_ia32_sqrtph512_mask_round(C, A, B, D) __builtin_ia32_sqrtph512_mask_round(C, A, B, 8) > > > +#define __builtin_ia32_sqrtsh_mask_round(D, C, A, B, E) __builtin_ia32_sqrtsh_mask_round(D, C, A, B, 8) > > > +#define __builtin_ia32_scalefph512_mask_round(A, B, C, D, E) __builtin_ia32_scalefph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_scalefsh_mask_round(A, B, C, D, E) __builtin_ia32_scalefsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_reduceph512_mask_round(A, B, C, D, E) __builtin_ia32_reduceph512_mask_round(A, 123, C, D, 8) > > > +#define __builtin_ia32_reduceph128_mask(A, B, C, D) __builtin_ia32_reduceph128_mask(A, 123, C, D) > > > +#define __builtin_ia32_reduceph256_mask(A, B, C, D) __builtin_ia32_reduceph256_mask(A, 123, C, D) > > > +#define __builtin_ia32_reducesh_mask_round(A, B, C, D, E, F) __builtin_ia32_reducesh_mask_round(A, B, 123, D, E, 8) > > > +#define __builtin_ia32_rndscaleph512_mask_round(A, B, C, D, E) __builtin_ia32_rndscaleph512_mask_round(A, 123, C, D, 8) > > > +#define __builtin_ia32_rndscaleph128_mask(A, B, C, D) __builtin_ia32_rndscaleph128_mask(A, 123, C, D) > > > +#define __builtin_ia32_rndscaleph256_mask(A, B, C, D) __builtin_ia32_rndscaleph256_mask(A, 123, C, D) > > > +#define __builtin_ia32_rndscalesh_mask_round(A, B, C, D, E, F) __builtin_ia32_rndscalesh_mask_round(A, B, 123, D, E, 8) > > > #define __builtin_ia32_fpclassph512_mask(A, D, C) __builtin_ia32_fpclassph512_mask(A, 1, C) > > > #define __builtin_ia32_fpclasssh_mask(A, D, U) __builtin_ia32_fpclasssh_mask(A, 1, U) > > > #define __builtin_ia32_getexpph512_mask(A, B, C, D) __builtin_ia32_getexpph512_mask(A, B, C, 8) > > > @@ -738,8 +738,8 @@ > > > #define __builtin_ia32_getmantsh_mask_round(A, B, C, W, U, D) __builtin_ia32_getmantsh_mask_round(A, B, 1, W, U, 4) > > > > > > /* avx512fp16vlintrin.h */ > > > -#define __builtin_ia32_vcmpph_v8hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v8hf_mask(A, B, 1, D) > > > -#define __builtin_ia32_vcmpph_v16hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v16hf_mask(A, B, 1, D) > > > +#define __builtin_ia32_cmpph128_mask(A, B, C, D) __builtin_ia32_cmpph128_mask(A, B, 1, D) > > > +#define __builtin_ia32_cmpph256_mask(A, B, C, D) __builtin_ia32_cmpph256_mask(A, B, 1, D) > > > #define __builtin_ia32_fpclassph256_mask(A, D, C) __builtin_ia32_fpclassph256_mask(A, 1, C) > > > #define __builtin_ia32_fpclassph128_mask(A, D, C) __builtin_ia32_fpclassph128_mask(A, 1, C) > > > #define __builtin_ia32_getmantph256_mask(A, E, C, D) __builtin_ia32_getmantph256_mask(A, 1, C, D) > > > diff --git a/gcc/testsuite/gcc.target/i386/sse-23.c b/gcc/testsuite/gcc.target/i386/sse-23.c > > > index b3f07587acb..8b600282c67 100644 > > > --- a/gcc/testsuite/gcc.target/i386/sse-23.c > > > +++ b/gcc/testsuite/gcc.target/i386/sse-23.c > > > @@ -704,33 +704,33 @@ > > > #define __builtin_ia32_vpshld_v2di_mask(A, B, C, D, E) __builtin_ia32_vpshld_v2di_mask(A, B, 1, D, E) > > > > > > /* avx512fp16intrin.h */ > > > -#define __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vcmpph_v32hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v32hf_mask(A, B, 1, D) > > > -#define __builtin_ia32_vcmpph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpph_v32hf_mask_round(A, B, 1, D, 8) > > > -#define __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, 1, D, 8) > > > -#define __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, D) __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, 8) > > > -#define __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, E) __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, 8) > > > -#define __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, 8) > > > -#define __builtin_ia32_vreduceph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vreduceph_v32hf_mask_round(A, 123, C, D, 8) > > > -#define __builtin_ia32_vreduceph_v8hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v8hf_mask(A, 123, C, D) > > > -#define __builtin_ia32_vreduceph_v16hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v16hf_mask(A, 123, C, D) > > > -#define __builtin_ia32_vreducesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vreducesh_v8hf_mask_round(A, B, 123, D, E, 8) > > > -#define __builtin_ia32_vrndscaleph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vrndscaleph_v32hf_mask_round(A, 123, C, D, 8) > > > -#define __builtin_ia32_vrndscaleph_v8hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v8hf_mask(A, 123, C, D) > > > -#define __builtin_ia32_vrndscaleph_v16hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v16hf_mask(A, 123, C, D) > > > -#define __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, 123, D, E, 8) > > > +#define __builtin_ia32_addph512_mask_round(A, B, C, D, E) __builtin_ia32_addph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_subph512_mask_round(A, B, C, D, E) __builtin_ia32_subph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_mulph512_mask_round(A, B, C, D, E) __builtin_ia32_mulph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_divph512_mask_round(A, B, C, D, E) __builtin_ia32_divph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_addsh_mask_round(A, B, C, D, E) __builtin_ia32_addsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_subsh_mask_round(A, B, C, D, E) __builtin_ia32_subsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_mulsh_mask_round(A, B, C, D, E) __builtin_ia32_mulsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_divsh_mask_round(A, B, C, D, E) __builtin_ia32_divsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_maxph512_mask_round(A, B, C, D, E) __builtin_ia32_maxph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_minph512_mask_round(A, B, C, D, E) __builtin_ia32_minph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_maxsh_mask_round(A, B, C, D, E) __builtin_ia32_maxsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_minsh_mask_round(A, B, C, D, E) __builtin_ia32_minsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_cmpph512_mask(A, B, C, D) __builtin_ia32_cmpph512_mask(A, B, 1, D) > > > +#define __builtin_ia32_cmpph512_mask_round(A, B, C, D, E) __builtin_ia32_cmpph512_mask_round(A, B, 1, D, 8) > > > +#define __builtin_ia32_cmpsh_mask_round(A, B, C, D, E) __builtin_ia32_cmpsh_mask_round(A, B, 1, D, 8) > > > +#define __builtin_ia32_sqrtph512_mask_round(C, A, B, D) __builtin_ia32_sqrtph512_mask_round(C, A, B, 8) > > > +#define __builtin_ia32_sqrtsh_mask_round(D, C, A, B, E) __builtin_ia32_sqrtsh_mask_round(D, C, A, B, 8) > > > +#define __builtin_ia32_scalefph512_mask_round(A, B, C, D, E) __builtin_ia32_scalefph512_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_scalefsh_mask_round(A, B, C, D, E) __builtin_ia32_scalefsh_mask_round(A, B, C, D, 8) > > > +#define __builtin_ia32_reduceph512_mask_round(A, B, C, D, E) __builtin_ia32_reduceph512_mask_round(A, 123, C, D, 8) > > > +#define __builtin_ia32_reduceph128_mask(A, B, C, D) __builtin_ia32_reduceph128_mask(A, 123, C, D) > > > +#define __builtin_ia32_reduceph256_mask(A, B, C, D) __builtin_ia32_reduceph256_mask(A, 123, C, D) > > > +#define __builtin_ia32_reducesh_mask_round(A, B, C, D, E, F) __builtin_ia32_reducesh_mask_round(A, B, 123, D, E, 8) > > > +#define __builtin_ia32_rndscaleph512_mask_round(A, B, C, D, E) __builtin_ia32_rndscaleph512_mask_round(A, 123, C, D, 8) > > > +#define __builtin_ia32_rndscaleph128_mask(A, B, C, D) __builtin_ia32_rndscaleph128_mask(A, 123, C, D) > > > +#define __builtin_ia32_rndscaleph256_mask(A, B, C, D) __builtin_ia32_rndscaleph256_mask(A, 123, C, D) > > > +#define __builtin_ia32_rndscalesh_mask_round(A, B, C, D, E, F) __builtin_ia32_rndscalesh_mask_round(A, B, 123, D, E, 8) > > > #define __builtin_ia32_fpclassph512_mask(A, D, C) __builtin_ia32_fpclassph512_mask(A, 1, C) > > > #define __builtin_ia32_fpclasssh_mask(A, D, U) __builtin_ia32_fpclasssh_mask(A, 1, U) > > > #define __builtin_ia32_getexpph512_mask(A, B, C, D) __builtin_ia32_getexpph512_mask(A, B, C, 8) > > > @@ -739,8 +739,8 @@ > > > #define __builtin_ia32_getmantsh_mask_round(A, B, C, W, U, D) __builtin_ia32_getmantsh_mask_round(A, B, 1, W, U, 4) > > > > > > /* avx512fp16vlintrin.h */ > > > -#define __builtin_ia32_vcmpph_v8hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v8hf_mask(A, B, 1, D) > > > -#define __builtin_ia32_vcmpph_v16hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v16hf_mask(A, B, 1, D) > > > +#define __builtin_ia32_cmpph128_mask(A, B, C, D) __builtin_ia32_cmpph128_mask(A, B, 1, D) > > > +#define __builtin_ia32_cmpph256_mask(A, B, C, D) __builtin_ia32_cmpph256_mask(A, B, 1, D) > > > #define __builtin_ia32_fpclassph256_mask(A, D, C) __builtin_ia32_fpclassph256_mask(A, 1, C) > > > #define __builtin_ia32_fpclassph128_mask(A, D, C) __builtin_ia32_fpclassph128_mask(A, 1, C) > > > #define __builtin_ia32_getmantph256_mask(A, E, C, D) __builtin_ia32_getmantph256_mask(A, 1, C, D) > > > -- > > > 2.18.1 > > > > > > > > > -- > > BR, > > Hongtao
diff --git a/gcc/config/i386/avx512fp16intrin.h b/gcc/config/i386/avx512fp16intrin.h index 2fbfc140c44..5d66ca5c820 100644 --- a/gcc/config/i386/avx512fp16intrin.h +++ b/gcc/config/i386/avx512fp16intrin.h @@ -229,15 +229,15 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_add_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) { - return __builtin_ia32_vaddph_v32hf_mask (__C, __D, __A, __B); + return __builtin_ia32_addph512_mask (__C, __D, __A, __B); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_add_ph (__mmask32 __A, __m512h __B, __m512h __C) { - return __builtin_ia32_vaddph_v32hf_mask (__B, __C, - _mm512_setzero_ph (), __A); + return __builtin_ia32_addph512_mask (__B, __C, + _mm512_setzero_ph (), __A); } extern __inline __m512h @@ -251,15 +251,15 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_sub_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) { - return __builtin_ia32_vsubph_v32hf_mask (__C, __D, __A, __B); + return __builtin_ia32_subph512_mask (__C, __D, __A, __B); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_sub_ph (__mmask32 __A, __m512h __B, __m512h __C) { - return __builtin_ia32_vsubph_v32hf_mask (__B, __C, - _mm512_setzero_ph (), __A); + return __builtin_ia32_subph512_mask (__B, __C, + _mm512_setzero_ph (), __A); } extern __inline __m512h @@ -273,15 +273,15 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_mul_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) { - return __builtin_ia32_vmulph_v32hf_mask (__C, __D, __A, __B); + return __builtin_ia32_mulph512_mask (__C, __D, __A, __B); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_mul_ph (__mmask32 __A, __m512h __B, __m512h __C) { - return __builtin_ia32_vmulph_v32hf_mask (__B, __C, - _mm512_setzero_ph (), __A); + return __builtin_ia32_mulph512_mask (__B, __C, + _mm512_setzero_ph (), __A); } extern __inline __m512h @@ -295,15 +295,15 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_div_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) { - return __builtin_ia32_vdivph_v32hf_mask (__C, __D, __A, __B); + return __builtin_ia32_divph512_mask (__C, __D, __A, __B); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_div_ph (__mmask32 __A, __m512h __B, __m512h __C) { - return __builtin_ia32_vdivph_v32hf_mask (__B, __C, - _mm512_setzero_ph (), __A); + return __builtin_ia32_divph512_mask (__B, __C, + _mm512_setzero_ph (), __A); } #ifdef __OPTIMIZE__ @@ -311,9 +311,9 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_add_round_ph (__m512h __A, __m512h __B, const int __C) { - return __builtin_ia32_vaddph_v32hf_mask_round (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1, __C); + return __builtin_ia32_addph512_mask_round (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1, __C); } extern __inline __m512h @@ -321,7 +321,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_add_round_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D, const int __E) { - return __builtin_ia32_vaddph_v32hf_mask_round (__C, __D, __A, __B, __E); + return __builtin_ia32_addph512_mask_round (__C, __D, __A, __B, __E); } extern __inline __m512h @@ -329,18 +329,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_add_round_ph (__mmask32 __A, __m512h __B, __m512h __C, const int __D) { - return __builtin_ia32_vaddph_v32hf_mask_round (__B, __C, - _mm512_setzero_ph (), - __A, __D); + return __builtin_ia32_addph512_mask_round (__B, __C, + _mm512_setzero_ph (), + __A, __D); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_sub_round_ph (__m512h __A, __m512h __B, const int __C) { - return __builtin_ia32_vsubph_v32hf_mask_round (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1, __C); + return __builtin_ia32_subph512_mask_round (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1, __C); } extern __inline __m512h @@ -348,7 +348,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_sub_round_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D, const int __E) { - return __builtin_ia32_vsubph_v32hf_mask_round (__C, __D, __A, __B, __E); + return __builtin_ia32_subph512_mask_round (__C, __D, __A, __B, __E); } extern __inline __m512h @@ -356,18 +356,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_sub_round_ph (__mmask32 __A, __m512h __B, __m512h __C, const int __D) { - return __builtin_ia32_vsubph_v32hf_mask_round (__B, __C, - _mm512_setzero_ph (), - __A, __D); + return __builtin_ia32_subph512_mask_round (__B, __C, + _mm512_setzero_ph (), + __A, __D); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mul_round_ph (__m512h __A, __m512h __B, const int __C) { - return __builtin_ia32_vmulph_v32hf_mask_round (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1, __C); + return __builtin_ia32_mulph512_mask_round (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1, __C); } extern __inline __m512h @@ -375,7 +375,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_mul_round_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D, const int __E) { - return __builtin_ia32_vmulph_v32hf_mask_round (__C, __D, __A, __B, __E); + return __builtin_ia32_mulph512_mask_round (__C, __D, __A, __B, __E); } extern __inline __m512h @@ -383,18 +383,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_mul_round_ph (__mmask32 __A, __m512h __B, __m512h __C, const int __D) { - return __builtin_ia32_vmulph_v32hf_mask_round (__B, __C, - _mm512_setzero_ph (), - __A, __D); + return __builtin_ia32_mulph512_mask_round (__B, __C, + _mm512_setzero_ph (), + __A, __D); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_div_round_ph (__m512h __A, __m512h __B, const int __C) { - return __builtin_ia32_vdivph_v32hf_mask_round (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1, __C); + return __builtin_ia32_divph512_mask_round (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1, __C); } extern __inline __m512h @@ -402,7 +402,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_div_round_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D, const int __E) { - return __builtin_ia32_vdivph_v32hf_mask_round (__C, __D, __A, __B, __E); + return __builtin_ia32_divph512_mask_round (__C, __D, __A, __B, __E); } extern __inline __m512h @@ -410,67 +410,67 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_div_round_ph (__mmask32 __A, __m512h __B, __m512h __C, const int __D) { - return __builtin_ia32_vdivph_v32hf_mask_round (__B, __C, - _mm512_setzero_ph (), - __A, __D); + return __builtin_ia32_divph512_mask_round (__B, __C, + _mm512_setzero_ph (), + __A, __D); } #else #define _mm512_add_round_ph(A, B, C) \ - ((__m512h)__builtin_ia32_vaddph_v32hf_mask_round((A), (B), \ - _mm512_setzero_ph (),\ - (__mmask32)-1, (C))) + ((__m512h)__builtin_ia32_addph512_mask_round((A), (B), \ + _mm512_setzero_ph (), \ + (__mmask32)-1, (C))) -#define _mm512_mask_add_round_ph(A, B, C, D, E) \ - ((__m512h)__builtin_ia32_vaddph_v32hf_mask_round((C), (D), (A), (B), (E))) +#define _mm512_mask_add_round_ph(A, B, C, D, E) \ + ((__m512h)__builtin_ia32_addph512_mask_round((C), (D), (A), (B), (E))) #define _mm512_maskz_add_round_ph(A, B, C, D) \ - ((__m512h)__builtin_ia32_vaddph_v32hf_mask_round((B), (C), \ - _mm512_setzero_ph (),\ - (A), (D))) + ((__m512h)__builtin_ia32_addph512_mask_round((B), (C), \ + _mm512_setzero_ph (), \ + (A), (D))) #define _mm512_sub_round_ph(A, B, C) \ - ((__m512h)__builtin_ia32_vsubph_v32hf_mask_round((A), (B), \ - _mm512_setzero_ph (),\ - (__mmask32)-1, (C))) + ((__m512h)__builtin_ia32_subph512_mask_round((A), (B), \ + _mm512_setzero_ph (), \ + (__mmask32)-1, (C))) -#define _mm512_mask_sub_round_ph(A, B, C, D, E) \ - ((__m512h)__builtin_ia32_vsubph_v32hf_mask_round((C), (D), (A), (B), (E))) +#define _mm512_mask_sub_round_ph(A, B, C, D, E) \ + ((__m512h)__builtin_ia32_subph512_mask_round((C), (D), (A), (B), (E))) #define _mm512_maskz_sub_round_ph(A, B, C, D) \ - ((__m512h)__builtin_ia32_vsubph_v32hf_mask_round((B), (C), \ - _mm512_setzero_ph (),\ - (A), (D))) + ((__m512h)__builtin_ia32_subph512_mask_round((B), (C), \ + _mm512_setzero_ph (), \ + (A), (D))) #define _mm512_mul_round_ph(A, B, C) \ - ((__m512h)__builtin_ia32_vmulph_v32hf_mask_round((A), (B), \ - _mm512_setzero_ph (),\ - (__mmask32)-1, (C))) + ((__m512h)__builtin_ia32_mulph512_mask_round((A), (B), \ + _mm512_setzero_ph (), \ + (__mmask32)-1, (C))) -#define _mm512_mask_mul_round_ph(A, B, C, D, E) \ - ((__m512h)__builtin_ia32_vmulph_v32hf_mask_round((C), (D), (A), (B), (E))) +#define _mm512_mask_mul_round_ph(A, B, C, D, E) \ + ((__m512h)__builtin_ia32_mulph512_mask_round((C), (D), (A), (B), (E))) #define _mm512_maskz_mul_round_ph(A, B, C, D) \ - ((__m512h)__builtin_ia32_vmulph_v32hf_mask_round((B), (C), \ - _mm512_setzero_ph (),\ - (A), (D))) + ((__m512h)__builtin_ia32_mulph512_mask_round((B), (C), \ + _mm512_setzero_ph (), \ + (A), (D))) #define _mm512_div_round_ph(A, B, C) \ - ((__m512h)__builtin_ia32_vdivph_v32hf_mask_round((A), (B), \ - _mm512_setzero_ph (),\ - (__mmask32)-1, (C))) + ((__m512h)__builtin_ia32_divph512_mask_round((A), (B), \ + _mm512_setzero_ph (), \ + (__mmask32)-1, (C))) -#define _mm512_mask_div_round_ph(A, B, C, D, E) \ - ((__m512h)__builtin_ia32_vdivph_v32hf_mask_round((C), (D), (A), (B), (E))) +#define _mm512_mask_div_round_ph(A, B, C, D, E) \ + ((__m512h)__builtin_ia32_divph512_mask_round((C), (D), (A), (B), (E))) #define _mm512_maskz_div_round_ph(A, B, C, D) \ - ((__m512h)__builtin_ia32_vdivph_v32hf_mask_round((B), (C), \ - _mm512_setzero_ph (),\ - (A), (D))) + ((__m512h)__builtin_ia32_divph512_mask_round((B), (C), \ + _mm512_setzero_ph (), \ + (A), (D))) #endif /* __OPTIMIZE__ */ /* Intrinsics of v[add,sub,mul,div]sh. */ extern __inline __m128h -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_add_sh (__m128h __A, __m128h __B) { __A[0] += __B[0]; @@ -481,15 +481,15 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_add_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vaddsh_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_addsh_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_add_sh (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vaddsh_v8hf_mask (__B, __C, _mm_setzero_ph (), - __A); + return __builtin_ia32_addsh_mask (__B, __C, _mm_setzero_ph (), + __A); } extern __inline __m128h @@ -504,15 +504,15 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_sub_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vsubsh_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_subsh_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_sub_sh (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vsubsh_v8hf_mask (__B, __C, _mm_setzero_ph (), - __A); + return __builtin_ia32_subsh_mask (__B, __C, _mm_setzero_ph (), + __A); } extern __inline __m128h @@ -527,14 +527,14 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_mul_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vmulsh_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_mulsh_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_mul_sh (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vmulsh_v8hf_mask (__B, __C, _mm_setzero_ph (), __A); + return __builtin_ia32_mulsh_mask (__B, __C, _mm_setzero_ph (), __A); } extern __inline __m128h @@ -549,15 +549,15 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_div_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vdivsh_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_divsh_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_div_sh (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vdivsh_v8hf_mask (__B, __C, _mm_setzero_ph (), - __A); + return __builtin_ia32_divsh_mask (__B, __C, _mm_setzero_ph (), + __A); } #ifdef __OPTIMIZE__ @@ -565,9 +565,9 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_add_round_sh (__m128h __A, __m128h __B, const int __C) { - return __builtin_ia32_vaddsh_v8hf_mask_round (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1, __C); + return __builtin_ia32_addsh_mask_round (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1, __C); } extern __inline __m128h @@ -575,7 +575,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_add_round_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D, const int __E) { - return __builtin_ia32_vaddsh_v8hf_mask_round (__C, __D, __A, __B, __E); + return __builtin_ia32_addsh_mask_round (__C, __D, __A, __B, __E); } extern __inline __m128h @@ -583,18 +583,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_add_round_sh (__mmask8 __A, __m128h __B, __m128h __C, const int __D) { - return __builtin_ia32_vaddsh_v8hf_mask_round (__B, __C, - _mm_setzero_ph (), - __A, __D); + return __builtin_ia32_addsh_mask_round (__B, __C, + _mm_setzero_ph (), + __A, __D); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_sub_round_sh (__m128h __A, __m128h __B, const int __C) { - return __builtin_ia32_vsubsh_v8hf_mask_round (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1, __C); + return __builtin_ia32_subsh_mask_round (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1, __C); } extern __inline __m128h @@ -602,7 +602,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_sub_round_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D, const int __E) { - return __builtin_ia32_vsubsh_v8hf_mask_round (__C, __D, __A, __B, __E); + return __builtin_ia32_subsh_mask_round (__C, __D, __A, __B, __E); } extern __inline __m128h @@ -610,18 +610,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_sub_round_sh (__mmask8 __A, __m128h __B, __m128h __C, const int __D) { - return __builtin_ia32_vsubsh_v8hf_mask_round (__B, __C, - _mm_setzero_ph (), - __A, __D); + return __builtin_ia32_subsh_mask_round (__B, __C, + _mm_setzero_ph (), + __A, __D); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mul_round_sh (__m128h __A, __m128h __B, const int __C) { - return __builtin_ia32_vmulsh_v8hf_mask_round (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1, __C); + return __builtin_ia32_mulsh_mask_round (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1, __C); } extern __inline __m128h @@ -629,7 +629,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_mul_round_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D, const int __E) { - return __builtin_ia32_vmulsh_v8hf_mask_round (__C, __D, __A, __B, __E); + return __builtin_ia32_mulsh_mask_round (__C, __D, __A, __B, __E); } extern __inline __m128h @@ -637,18 +637,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_mul_round_sh (__mmask8 __A, __m128h __B, __m128h __C, const int __D) { - return __builtin_ia32_vmulsh_v8hf_mask_round (__B, __C, - _mm_setzero_ph (), - __A, __D); + return __builtin_ia32_mulsh_mask_round (__B, __C, + _mm_setzero_ph (), + __A, __D); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_div_round_sh (__m128h __A, __m128h __B, const int __C) { - return __builtin_ia32_vdivsh_v8hf_mask_round (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1, __C); + return __builtin_ia32_divsh_mask_round (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1, __C); } extern __inline __m128h @@ -656,7 +656,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_div_round_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D, const int __E) { - return __builtin_ia32_vdivsh_v8hf_mask_round (__C, __D, __A, __B, __E); + return __builtin_ia32_divsh_mask_round (__C, __D, __A, __B, __E); } extern __inline __m128h @@ -664,62 +664,62 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_div_round_sh (__mmask8 __A, __m128h __B, __m128h __C, const int __D) { - return __builtin_ia32_vdivsh_v8hf_mask_round (__B, __C, - _mm_setzero_ph (), - __A, __D); + return __builtin_ia32_divsh_mask_round (__B, __C, + _mm_setzero_ph (), + __A, __D); } #else #define _mm_add_round_sh(A, B, C) \ - ((__m128h)__builtin_ia32_vaddsh_v8hf_mask_round ((A), (B), \ - _mm_setzero_ph (), \ - (__mmask8)-1, (C))) + ((__m128h)__builtin_ia32_addsh_mask_round ((A), (B), \ + _mm_setzero_ph (), \ + (__mmask8)-1, (C))) #define _mm_mask_add_round_sh(A, B, C, D, E) \ - ((__m128h)__builtin_ia32_vaddsh_v8hf_mask_round ((C), (D), (A), (B), (E))) + ((__m128h)__builtin_ia32_addsh_mask_round ((C), (D), (A), (B), (E))) -#define _mm_maskz_add_round_sh(A, B, C, D) \ - ((__m128h)__builtin_ia32_vaddsh_v8hf_mask_round ((B), (C), \ - _mm_setzero_ph (), \ - (A), (D))) +#define _mm_maskz_add_round_sh(A, B, C, D) \ + ((__m128h)__builtin_ia32_addsh_mask_round ((B), (C), \ + _mm_setzero_ph (), \ + (A), (D))) #define _mm_sub_round_sh(A, B, C) \ - ((__m128h)__builtin_ia32_vsubsh_v8hf_mask_round ((A), (B), \ - _mm_setzero_ph (), \ - (__mmask8)-1, (C))) + ((__m128h)__builtin_ia32_subsh_mask_round ((A), (B), \ + _mm_setzero_ph (), \ + (__mmask8)-1, (C))) #define _mm_mask_sub_round_sh(A, B, C, D, E) \ - ((__m128h)__builtin_ia32_vsubsh_v8hf_mask_round ((C), (D), (A), (B), (E))) + ((__m128h)__builtin_ia32_subsh_mask_round ((C), (D), (A), (B), (E))) -#define _mm_maskz_sub_round_sh(A, B, C, D) \ - ((__m128h)__builtin_ia32_vsubsh_v8hf_mask_round ((B), (C), \ - _mm_setzero_ph (), \ - (A), (D))) +#define _mm_maskz_sub_round_sh(A, B, C, D) \ + ((__m128h)__builtin_ia32_subsh_mask_round ((B), (C), \ + _mm_setzero_ph (), \ + (A), (D))) #define _mm_mul_round_sh(A, B, C) \ - ((__m128h)__builtin_ia32_vmulsh_v8hf_mask_round ((A), (B), \ - _mm_setzero_ph (), \ - (__mmask8)-1, (C))) + ((__m128h)__builtin_ia32_mulsh_mask_round ((A), (B), \ + _mm_setzero_ph (), \ + (__mmask8)-1, (C))) #define _mm_mask_mul_round_sh(A, B, C, D, E) \ - ((__m128h)__builtin_ia32_vmulsh_v8hf_mask_round ((C), (D), (A), (B), (E))) + ((__m128h)__builtin_ia32_mulsh_mask_round ((C), (D), (A), (B), (E))) -#define _mm_maskz_mul_round_sh(A, B, C, D) \ - ((__m128h)__builtin_ia32_vmulsh_v8hf_mask_round ((B), (C), \ - _mm_setzero_ph (), \ - (A), (D))) +#define _mm_maskz_mul_round_sh(A, B, C, D) \ + ((__m128h)__builtin_ia32_mulsh_mask_round ((B), (C), \ + _mm_setzero_ph (), \ + (A), (D))) #define _mm_div_round_sh(A, B, C) \ - ((__m128h)__builtin_ia32_vdivsh_v8hf_mask_round ((A), (B), \ - _mm_setzero_ph (), \ - (__mmask8)-1, (C))) + ((__m128h)__builtin_ia32_divsh_mask_round ((A), (B), \ + _mm_setzero_ph (), \ + (__mmask8)-1, (C))) #define _mm_mask_div_round_sh(A, B, C, D, E) \ - ((__m128h)__builtin_ia32_vdivsh_v8hf_mask_round ((C), (D), (A), (B), (E))) + ((__m128h)__builtin_ia32_divsh_mask_round ((C), (D), (A), (B), (E))) -#define _mm_maskz_div_round_sh(A, B, C, D) \ - ((__m128h)__builtin_ia32_vdivsh_v8hf_mask_round ((B), (C), \ - _mm_setzero_ph (), \ - (A), (D))) +#define _mm_maskz_div_round_sh(A, B, C, D) \ + ((__m128h)__builtin_ia32_divsh_mask_round ((B), (C), \ + _mm_setzero_ph (), \ + (A), (D))) #endif /* __OPTIMIZE__ */ /* Intrinsic vmaxph vminph. */ @@ -727,48 +727,48 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_max_ph (__m512h __A, __m512h __B) { - return __builtin_ia32_vmaxph_v32hf_mask (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1); + return __builtin_ia32_maxph512_mask (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_max_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) { - return __builtin_ia32_vmaxph_v32hf_mask (__C, __D, __A, __B); + return __builtin_ia32_maxph512_mask (__C, __D, __A, __B); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_max_ph (__mmask32 __A, __m512h __B, __m512h __C) { - return __builtin_ia32_vmaxph_v32hf_mask (__B, __C, - _mm512_setzero_ph (), __A); + return __builtin_ia32_maxph512_mask (__B, __C, + _mm512_setzero_ph (), __A); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_min_ph (__m512h __A, __m512h __B) { - return __builtin_ia32_vminph_v32hf_mask (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1); + return __builtin_ia32_minph512_mask (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_min_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) { - return __builtin_ia32_vminph_v32hf_mask (__C, __D, __A, __B); + return __builtin_ia32_minph512_mask (__C, __D, __A, __B); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_min_ph (__mmask32 __A, __m512h __B, __m512h __C) { - return __builtin_ia32_vminph_v32hf_mask (__B, __C, - _mm512_setzero_ph (), __A); + return __builtin_ia32_minph512_mask (__B, __C, + _mm512_setzero_ph (), __A); } #ifdef __OPTIMIZE__ @@ -776,9 +776,9 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_max_round_ph (__m512h __A, __m512h __B, const int __C) { - return __builtin_ia32_vmaxph_v32hf_mask_round (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1, __C); + return __builtin_ia32_maxph512_mask_round (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1, __C); } extern __inline __m512h @@ -786,7 +786,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_max_round_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D, const int __E) { - return __builtin_ia32_vmaxph_v32hf_mask_round (__C, __D, __A, __B, __E); + return __builtin_ia32_maxph512_mask_round (__C, __D, __A, __B, __E); } extern __inline __m512h @@ -794,18 +794,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_max_round_ph (__mmask32 __A, __m512h __B, __m512h __C, const int __D) { - return __builtin_ia32_vmaxph_v32hf_mask_round (__B, __C, - _mm512_setzero_ph (), - __A, __D); + return __builtin_ia32_maxph512_mask_round (__B, __C, + _mm512_setzero_ph (), + __A, __D); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_min_round_ph (__m512h __A, __m512h __B, const int __C) { - return __builtin_ia32_vminph_v32hf_mask_round (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1, __C); + return __builtin_ia32_minph512_mask_round (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1, __C); } extern __inline __m512h @@ -813,7 +813,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_min_round_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D, const int __E) { - return __builtin_ia32_vminph_v32hf_mask_round (__C, __D, __A, __B, __E); + return __builtin_ia32_minph512_mask_round (__C, __D, __A, __B, __E); } extern __inline __m512h @@ -821,37 +821,37 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_min_round_ph (__mmask32 __A, __m512h __B, __m512h __C, const int __D) { - return __builtin_ia32_vminph_v32hf_mask_round (__B, __C, - _mm512_setzero_ph (), - __A, __D); + return __builtin_ia32_minph512_mask_round (__B, __C, + _mm512_setzero_ph (), + __A, __D); } #else -#define _mm512_max_round_ph(A, B, C) \ - (__builtin_ia32_vmaxph_v32hf_mask_round ((A), (B), \ - _mm512_setzero_ph (), \ - (__mmask32)-1, (C))) +#define _mm512_max_round_ph(A, B, C) \ + (__builtin_ia32_maxph512_mask_round ((A), (B), \ + _mm512_setzero_ph (), \ + (__mmask32)-1, (C))) #define _mm512_mask_max_round_ph(A, B, C, D, E) \ - (__builtin_ia32_vmaxph_v32hf_mask_round ((C), (D), (A), (B), (E))) + (__builtin_ia32_maxph512_mask_round ((C), (D), (A), (B), (E))) -#define _mm512_maskz_max_round_ph(A, B, C, D) \ - (__builtin_ia32_vmaxph_v32hf_mask_round ((B), (C), \ - _mm512_setzero_ph (), \ - (A), (D))) +#define _mm512_maskz_max_round_ph(A, B, C, D) \ + (__builtin_ia32_maxph512_mask_round ((B), (C), \ + _mm512_setzero_ph (), \ + (A), (D))) -#define _mm512_min_round_ph(A, B, C) \ - (__builtin_ia32_vminph_v32hf_mask_round ((A), (B), \ - _mm512_setzero_ph (), \ - (__mmask32)-1, (C))) +#define _mm512_min_round_ph(A, B, C) \ + (__builtin_ia32_minph512_mask_round ((A), (B), \ + _mm512_setzero_ph (), \ + (__mmask32)-1, (C))) #define _mm512_mask_min_round_ph(A, B, C, D, E) \ - (__builtin_ia32_vminph_v32hf_mask_round ((C), (D), (A), (B), (E))) + (__builtin_ia32_minph512_mask_round ((C), (D), (A), (B), (E))) -#define _mm512_maskz_min_round_ph(A, B, C, D) \ - (__builtin_ia32_vminph_v32hf_mask_round ((B), (C), \ - _mm512_setzero_ph (), \ - (A), (D))) +#define _mm512_maskz_min_round_ph(A, B, C, D) \ + (__builtin_ia32_minph512_mask_round ((B), (C), \ + _mm512_setzero_ph (), \ + (A), (D))) #endif /* __OPTIMIZE__ */ /* Intrinsic vmaxsh vminsh. */ @@ -867,15 +867,15 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_max_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vmaxsh_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_maxsh_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_max_sh (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vmaxsh_v8hf_mask (__B, __C, _mm_setzero_ph (), - __A); + return __builtin_ia32_maxsh_mask (__B, __C, _mm_setzero_ph (), + __A); } extern __inline __m128h @@ -890,15 +890,15 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_min_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vminsh_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_minsh_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_min_sh (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vminsh_v8hf_mask (__B, __C, _mm_setzero_ph (), - __A); + return __builtin_ia32_minsh_mask (__B, __C, _mm_setzero_ph (), + __A); } #ifdef __OPTIMIZE__ @@ -906,9 +906,9 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_max_round_sh (__m128h __A, __m128h __B, const int __C) { - return __builtin_ia32_vmaxsh_v8hf_mask_round (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1, __C); + return __builtin_ia32_maxsh_mask_round (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1, __C); } extern __inline __m128h @@ -916,7 +916,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_max_round_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D, const int __E) { - return __builtin_ia32_vmaxsh_v8hf_mask_round (__C, __D, __A, __B, __E); + return __builtin_ia32_maxsh_mask_round (__C, __D, __A, __B, __E); } extern __inline __m128h @@ -924,18 +924,18 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_max_round_sh (__mmask8 __A, __m128h __B, __m128h __C, const int __D) { - return __builtin_ia32_vmaxsh_v8hf_mask_round (__B, __C, - _mm_setzero_ph (), - __A, __D); + return __builtin_ia32_maxsh_mask_round (__B, __C, + _mm_setzero_ph (), + __A, __D); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_min_round_sh (__m128h __A, __m128h __B, const int __C) { - return __builtin_ia32_vminsh_v8hf_mask_round (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1, __C); + return __builtin_ia32_minsh_mask_round (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1, __C); } extern __inline __m128h @@ -943,7 +943,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_min_round_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D, const int __E) { - return __builtin_ia32_vminsh_v8hf_mask_round (__C, __D, __A, __B, __E); + return __builtin_ia32_minsh_mask_round (__C, __D, __A, __B, __E); } extern __inline __m128h @@ -951,37 +951,37 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_min_round_sh (__mmask8 __A, __m128h __B, __m128h __C, const int __D) { - return __builtin_ia32_vminsh_v8hf_mask_round (__B, __C, - _mm_setzero_ph (), - __A, __D); + return __builtin_ia32_minsh_mask_round (__B, __C, + _mm_setzero_ph (), + __A, __D); } #else -#define _mm_max_round_sh(A, B, C) \ - (__builtin_ia32_vmaxsh_v8hf_mask_round ((A), (B), \ - _mm_setzero_ph (), \ - (__mmask8)-1, (C))) +#define _mm_max_round_sh(A, B, C) \ + (__builtin_ia32_maxsh_mask_round ((A), (B), \ + _mm_setzero_ph (), \ + (__mmask8)-1, (C))) -#define _mm_mask_max_round_sh(A, B, C, D, E) \ - (__builtin_ia32_vmaxsh_v8hf_mask_round ((C), (D), (A), (B), (E))) +#define _mm_mask_max_round_sh(A, B, C, D, E) \ + (__builtin_ia32_maxsh_mask_round ((C), (D), (A), (B), (E))) -#define _mm_maskz_max_round_sh(A, B, C, D) \ - (__builtin_ia32_vmaxsh_v8hf_mask_round ((B), (C), \ - _mm_setzero_ph (), \ - (A), (D))) +#define _mm_maskz_max_round_sh(A, B, C, D) \ + (__builtin_ia32_maxsh_mask_round ((B), (C), \ + _mm_setzero_ph (), \ + (A), (D))) -#define _mm_min_round_sh(A, B, C) \ - (__builtin_ia32_vminsh_v8hf_mask_round ((A), (B), \ - _mm_setzero_ph (), \ - (__mmask8)-1, (C))) +#define _mm_min_round_sh(A, B, C) \ + (__builtin_ia32_minsh_mask_round ((A), (B), \ + _mm_setzero_ph (), \ + (__mmask8)-1, (C))) -#define _mm_mask_min_round_sh(A, B, C, D, E) \ - (__builtin_ia32_vminsh_v8hf_mask_round ((C), (D), (A), (B), (E))) +#define _mm_mask_min_round_sh(A, B, C, D, E) \ + (__builtin_ia32_minsh_mask_round ((C), (D), (A), (B), (E))) -#define _mm_maskz_min_round_sh(A, B, C, D) \ - (__builtin_ia32_vminsh_v8hf_mask_round ((B), (C), \ - _mm_setzero_ph (), \ - (A), (D))) +#define _mm_maskz_min_round_sh(A, B, C, D) \ + (__builtin_ia32_minsh_mask_round ((B), (C), \ + _mm_setzero_ph (), \ + (A), (D))) #endif /* __OPTIMIZE__ */ @@ -991,8 +991,8 @@ extern __inline __mmask32 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_cmp_ph_mask (__m512h __A, __m512h __B, const int __C) { - return (__mmask32) __builtin_ia32_vcmpph_v32hf_mask (__A, __B, __C, - (__mmask32) -1); + return (__mmask32) __builtin_ia32_cmpph512_mask (__A, __B, __C, + (__mmask32) -1); } extern __inline __mmask32 @@ -1000,8 +1000,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_cmp_ph_mask (__mmask32 __A, __m512h __B, __m512h __C, const int __D) { - return (__mmask32) __builtin_ia32_vcmpph_v32hf_mask (__B, __C, __D, - __A); + return (__mmask32) __builtin_ia32_cmpph512_mask (__B, __C, __D, + __A); } extern __inline __mmask32 @@ -1009,9 +1009,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_cmp_round_ph_mask (__m512h __A, __m512h __B, const int __C, const int __D) { - return (__mmask32) __builtin_ia32_vcmpph_v32hf_mask_round (__A, __B, - __C, (__mmask32) -1, - __D); + return (__mmask32) __builtin_ia32_cmpph512_mask_round (__A, __B, + __C, (__mmask32) -1, + __D); } extern __inline __mmask32 @@ -1019,23 +1019,23 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_cmp_round_ph_mask (__mmask32 __A, __m512h __B, __m512h __C, const int __D, const int __E) { - return (__mmask32) __builtin_ia32_vcmpph_v32hf_mask_round (__B, __C, - __D, __A, - __E); + return (__mmask32) __builtin_ia32_cmpph512_mask_round (__B, __C, + __D, __A, + __E); } #else #define _mm512_cmp_ph_mask(A, B, C) \ - (__builtin_ia32_vcmpph_v32hf_mask ((A), (B), (C), (-1))) + (__builtin_ia32_cmpph512_mask ((A), (B), (C), (-1))) #define _mm512_mask_cmp_ph_mask(A, B, C, D) \ - (__builtin_ia32_vcmpph_v32hf_mask ((B), (C), (D), (A))) + (__builtin_ia32_cmpph512_mask ((B), (C), (D), (A))) -#define _mm512_cmp_round_ph_mask(A, B, C, D) \ - (__builtin_ia32_vcmpph_v32hf_mask_round ((A), (B), (C), (-1), (D))) +#define _mm512_cmp_round_ph_mask(A, B, C, D) \ + (__builtin_ia32_cmpph512_mask_round ((A), (B), (C), (-1), (D))) -#define _mm512_mask_cmp_round_ph_mask(A, B, C, D, E) \ - (__builtin_ia32_vcmpph_v32hf_mask_round ((B), (C), (D), (A), (E))) +#define _mm512_mask_cmp_round_ph_mask(A, B, C, D, E) \ + (__builtin_ia32_cmpph512_mask_round ((B), (C), (D), (A), (E))) #endif /* __OPTIMIZE__ */ @@ -1046,9 +1046,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_cmp_sh_mask (__m128h __A, __m128h __B, const int __C) { return (__mmask8) - __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, - __C, (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + __builtin_ia32_cmpsh_mask_round (__A, __B, + __C, (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline __mmask8 @@ -1057,9 +1057,9 @@ _mm_mask_cmp_sh_mask (__mmask8 __A, __m128h __B, __m128h __C, const int __D) { return (__mmask8) - __builtin_ia32_vcmpsh_v8hf_mask_round (__B, __C, - __D, __A, - _MM_FROUND_CUR_DIRECTION); + __builtin_ia32_cmpsh_mask_round (__B, __C, + __D, __A, + _MM_FROUND_CUR_DIRECTION); } extern __inline __mmask8 @@ -1067,9 +1067,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_cmp_round_sh_mask (__m128h __A, __m128h __B, const int __C, const int __D) { - return (__mmask8) __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, - __C, (__mmask8) -1, - __D); + return (__mmask8) __builtin_ia32_cmpsh_mask_round (__A, __B, + __C, (__mmask8) -1, + __D); } extern __inline __mmask8 @@ -1077,25 +1077,25 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_cmp_round_sh_mask (__mmask8 __A, __m128h __B, __m128h __C, const int __D, const int __E) { - return (__mmask8) __builtin_ia32_vcmpsh_v8hf_mask_round (__B, __C, - __D, __A, - __E); + return (__mmask8) __builtin_ia32_cmpsh_mask_round (__B, __C, + __D, __A, + __E); } #else -#define _mm_cmp_sh_mask(A, B, C) \ - (__builtin_ia32_vcmpsh_v8hf_mask_round ((A), (B), (C), (-1), \ - (_MM_FROUND_CUR_DIRECTION))) +#define _mm_cmp_sh_mask(A, B, C) \ + (__builtin_ia32_cmpsh_mask_round ((A), (B), (C), (-1), \ + (_MM_FROUND_CUR_DIRECTION))) -#define _mm_mask_cmp_sh_mask(A, B, C, D) \ - (__builtin_ia32_vcmpsh_v8hf_mask_round ((B), (C), (D), (A), \ - (_MM_FROUND_CUR_DIRECTION))) +#define _mm_mask_cmp_sh_mask(A, B, C, D) \ + (__builtin_ia32_cmpsh_mask_round ((B), (C), (D), (A), \ + (_MM_FROUND_CUR_DIRECTION))) -#define _mm_cmp_round_sh_mask(A, B, C, D) \ - (__builtin_ia32_vcmpsh_v8hf_mask_round ((A), (B), (C), (-1), (D))) +#define _mm_cmp_round_sh_mask(A, B, C, D) \ + (__builtin_ia32_cmpsh_mask_round ((A), (B), (C), (-1), (D))) -#define _mm_mask_cmp_round_sh_mask(A, B, C, D, E) \ - (__builtin_ia32_vcmpsh_v8hf_mask_round ((B), (C), (D), (A), (E))) +#define _mm_mask_cmp_round_sh_mask(A, B, C, D, E) \ + (__builtin_ia32_cmpsh_mask_round ((B), (C), (D), (A), (E))) #endif /* __OPTIMIZE__ */ @@ -1104,134 +1104,134 @@ extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_comieq_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_EQ_OS, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_EQ_OS, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_comilt_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_LT_OS, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_LT_OS, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_comile_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_LE_OS, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_LE_OS, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_comigt_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_GT_OS, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_GT_OS, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_comige_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_GE_OS, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_GE_OS, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_comineq_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_NEQ_US, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_NEQ_US, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_ucomieq_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_EQ_OQ, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_EQ_OQ, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_ucomilt_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_LT_OQ, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_LT_OQ, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_ucomile_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_LE_OQ, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_LE_OQ, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_ucomigt_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_GT_OQ, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_GT_OQ, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_ucomige_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_GE_OQ, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_GE_OQ, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_ucomineq_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, _CMP_NEQ_UQ, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, _CMP_NEQ_UQ, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } #ifdef __OPTIMIZE__ extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) - _mm_comi_sh (__m128h __A, __m128h __B, const int __P) +_mm_comi_sh (__m128h __A, __m128h __B, const int __P) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, __P, - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_cmpsh_mask_round (__A, __B, __P, + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline int __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_comi_round_sh (__m128h __A, __m128h __B, const int __P, const int __R) { - return __builtin_ia32_vcmpsh_v8hf_mask_round (__A, __B, __P, - (__mmask8) -1,__R); + return __builtin_ia32_cmpsh_mask_round (__A, __B, __P, + (__mmask8) -1,__R); } #else -#define _mm_comi_round_sh(A, B, P, R) \ - (__builtin_ia32_vcmpsh_v8hf_mask_round ((A), (B), (P), (__mmask8) (-1), (R))) -#define _mm_comi_sh(A, B, P) \ - (__builtin_ia32_vcmpsh_v8hf_mask_round ((A), (B), (P), (__mmask8) (-1), \ - _MM_FROUND_CUR_DIRECTION)) +#define _mm_comi_round_sh(A, B, P, R) \ + (__builtin_ia32_cmpsh_mask_round ((A), (B), (P), (__mmask8) (-1), (R))) +#define _mm_comi_sh(A, B, P) \ + (__builtin_ia32_cmpsh_mask_round ((A), (B), (P), (__mmask8) (-1), \ + _MM_FROUND_CUR_DIRECTION)) #endif /* __OPTIMIZE__ */ @@ -1240,28 +1240,28 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_sqrt_ph (__m512h __A) { - return __builtin_ia32_vsqrtph_v32hf_mask_round (__A, - _mm512_setzero_ph(), - (__mmask32) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_sqrtph512_mask_round (__A, + _mm512_setzero_ph(), + (__mmask32) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_sqrt_ph (__m512h __A, __mmask32 __B, __m512h __C) { - return __builtin_ia32_vsqrtph_v32hf_mask_round (__C, __A, __B, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_sqrtph512_mask_round (__C, __A, __B, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_sqrt_ph (__mmask32 __A, __m512h __B) { - return __builtin_ia32_vsqrtph_v32hf_mask_round (__B, - _mm512_setzero_ph (), - __A, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_sqrtph512_mask_round (__B, + _mm512_setzero_ph (), + __A, + _MM_FROUND_CUR_DIRECTION); } #ifdef __OPTIMIZE__ @@ -1269,9 +1269,9 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_sqrt_round_ph (__m512h __A, const int __B) { - return __builtin_ia32_vsqrtph_v32hf_mask_round (__A, - _mm512_setzero_ph(), - (__mmask32) -1, __B); + return __builtin_ia32_sqrtph512_mask_round (__A, + _mm512_setzero_ph(), + (__mmask32) -1, __B); } extern __inline __m512h @@ -1279,31 +1279,31 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_sqrt_round_ph (__m512h __A, __mmask32 __B, __m512h __C, const int __D) { - return __builtin_ia32_vsqrtph_v32hf_mask_round (__C, __A, __B, __D); + return __builtin_ia32_sqrtph512_mask_round (__C, __A, __B, __D); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_sqrt_round_ph (__mmask32 __A, __m512h __B, const int __C) { - return __builtin_ia32_vsqrtph_v32hf_mask_round (__B, - _mm512_setzero_ph (), - __A, __C); + return __builtin_ia32_sqrtph512_mask_round (__B, + _mm512_setzero_ph (), + __A, __C); } #else -#define _mm512_sqrt_round_ph(A, B) \ - (__builtin_ia32_vsqrtph_v32hf_mask_round ((A), \ - _mm512_setzero_ph (), \ - (__mmask32)-1, (B))) +#define _mm512_sqrt_round_ph(A, B) \ + (__builtin_ia32_sqrtph512_mask_round ((A), \ + _mm512_setzero_ph (), \ + (__mmask32)-1, (B))) -#define _mm512_mask_sqrt_round_ph(A, B, C, D) \ - (__builtin_ia32_vsqrtph_v32hf_mask_round ((C), (A), (B), (D))) +#define _mm512_mask_sqrt_round_ph(A, B, C, D) \ + (__builtin_ia32_sqrtph512_mask_round ((C), (A), (B), (D))) -#define _mm512_maskz_sqrt_round_ph(A, B, C) \ - (__builtin_ia32_vsqrtph_v32hf_mask_round ((B), \ - _mm512_setzero_ph (), \ - (A), (C))) +#define _mm512_maskz_sqrt_round_ph(A, B, C) \ + (__builtin_ia32_sqrtph512_mask_round ((B), \ + _mm512_setzero_ph (), \ + (A), (C))) #endif /* __OPTIMIZE__ */ @@ -1312,23 +1312,23 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_rsqrt_ph (__m512h __A) { - return __builtin_ia32_vrsqrtph_v32hf_mask (__A, _mm512_setzero_ph (), - (__mmask32) -1); + return __builtin_ia32_rsqrtph512_mask (__A, _mm512_setzero_ph (), + (__mmask32) -1); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_rsqrt_ph (__m512h __A, __mmask32 __B, __m512h __C) { - return __builtin_ia32_vrsqrtph_v32hf_mask (__C, __A, __B); + return __builtin_ia32_rsqrtph512_mask (__C, __A, __B); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_rsqrt_ph (__mmask32 __A, __m512h __B) { - return __builtin_ia32_vrsqrtph_v32hf_mask (__B, _mm512_setzero_ph (), - __A); + return __builtin_ia32_rsqrtph512_mask (__B, _mm512_setzero_ph (), + __A); } /* Intrinsics vrsqrtsh. */ @@ -1336,23 +1336,23 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_rsqrt_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vrsqrtsh_v8hf_mask (__B, __A, _mm_setzero_ph (), - (__mmask8) -1); + return __builtin_ia32_rsqrtsh_mask (__B, __A, _mm_setzero_ph (), + (__mmask8) -1); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_rsqrt_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vrsqrtsh_v8hf_mask (__D, __C, __A, __B); + return __builtin_ia32_rsqrtsh_mask (__D, __C, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_rsqrt_sh (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vrsqrtsh_v8hf_mask (__C, __B, _mm_setzero_ph (), - __A); + return __builtin_ia32_rsqrtsh_mask (__C, __B, _mm_setzero_ph (), + __A); } /* Intrinsics vsqrtsh. */ @@ -1360,27 +1360,27 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_sqrt_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__B, __A, - _mm_setzero_ph (), - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_sqrtsh_mask_round (__B, __A, + _mm_setzero_ph (), + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_sqrt_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__D, __C, __A, __B, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_sqrtsh_mask_round (__D, __C, __A, __B, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_sqrt_sh (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__C, __B, - _mm_setzero_ph (), - __A, _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_sqrtsh_mask_round (__C, __B, + _mm_setzero_ph (), + __A, _MM_FROUND_CUR_DIRECTION); } #ifdef __OPTIMIZE__ @@ -1388,9 +1388,9 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_sqrt_round_sh (__m128h __A, __m128h __B, const int __C) { - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__B, __A, - _mm_setzero_ph (), - (__mmask8) -1, __C); + return __builtin_ia32_sqrtsh_mask_round (__B, __A, + _mm_setzero_ph (), + (__mmask8) -1, __C); } extern __inline __m128h @@ -1398,8 +1398,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_sqrt_round_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D, const int __E) { - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__D, __C, __A, __B, - __E); + return __builtin_ia32_sqrtsh_mask_round (__D, __C, __A, __B, + __E); } extern __inline __m128h @@ -1407,24 +1407,24 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_sqrt_round_sh (__mmask8 __A, __m128h __B, __m128h __C, const int __D) { - return __builtin_ia32_vsqrtsh_v8hf_mask_round (__C, __B, - _mm_setzero_ph (), - __A, __D); + return __builtin_ia32_sqrtsh_mask_round (__C, __B, + _mm_setzero_ph (), + __A, __D); } #else #define _mm_sqrt_round_sh(A, B, C) \ - (__builtin_ia32_vsqrtsh_v8hf_mask_round ((B), (A), \ - _mm_setzero_ph (), \ - (__mmask8)-1, (C))) + (__builtin_ia32_sqrtsh_mask_round ((B), (A), \ + _mm_setzero_ph (), \ + (__mmask8)-1, (C))) #define _mm_mask_sqrt_round_sh(A, B, C, D, E) \ - (__builtin_ia32_vsqrtsh_v8hf_mask_round ((D), (C), (A), (B), (E))) + (__builtin_ia32_sqrtsh_mask_round ((D), (C), (A), (B), (E))) -#define _mm_maskz_sqrt_round_sh(A, B, C, D) \ - (__builtin_ia32_vsqrtsh_v8hf_mask_round ((C), (B), \ - _mm_setzero_ph (), \ - (A), (D))) +#define _mm_maskz_sqrt_round_sh(A, B, C, D) \ + (__builtin_ia32_sqrtsh_mask_round ((C), (B), \ + _mm_setzero_ph (), \ + (A), (D))) #endif /* __OPTIMIZE__ */ @@ -1433,23 +1433,23 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_rcp_ph (__m512h __A) { - return __builtin_ia32_vrcpph_v32hf_mask (__A, _mm512_setzero_ph (), - (__mmask32) -1); + return __builtin_ia32_rcpph512_mask (__A, _mm512_setzero_ph (), + (__mmask32) -1); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_rcp_ph (__m512h __A, __mmask32 __B, __m512h __C) { - return __builtin_ia32_vrcpph_v32hf_mask (__C, __A, __B); + return __builtin_ia32_rcpph512_mask (__C, __A, __B); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_rcp_ph (__mmask32 __A, __m512h __B) { - return __builtin_ia32_vrcpph_v32hf_mask (__B, _mm512_setzero_ph (), - __A); + return __builtin_ia32_rcpph512_mask (__B, _mm512_setzero_ph (), + __A); } /* Intrinsics vrcpsh. */ @@ -1457,23 +1457,23 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_rcp_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vrcpsh_v8hf_mask (__B, __A, _mm_setzero_ph (), - (__mmask8) -1); + return __builtin_ia32_rcpsh_mask (__B, __A, _mm_setzero_ph (), + (__mmask8) -1); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_rcp_sh (__m128h __A, __mmask32 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vrcpsh_v8hf_mask (__D, __C, __A, __B); + return __builtin_ia32_rcpsh_mask (__D, __C, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_rcp_sh (__mmask32 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vrcpsh_v8hf_mask (__C, __B, _mm_setzero_ph (), - __A); + return __builtin_ia32_rcpsh_mask (__C, __B, _mm_setzero_ph (), + __A); } /* Intrinsics vscalefph. */ @@ -1481,28 +1481,28 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_scalef_ph (__m512h __A, __m512h __B) { - return __builtin_ia32_vscalefph_v32hf_mask_round (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_scalefph512_mask_round (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_scalef_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D) { - return __builtin_ia32_vscalefph_v32hf_mask_round (__C, __D, __A, __B, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_scalefph512_mask_round (__C, __D, __A, __B, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_scalef_ph (__mmask32 __A, __m512h __B, __m512h __C) { - return __builtin_ia32_vscalefph_v32hf_mask_round (__B, __C, - _mm512_setzero_ph (), - __A, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_scalefph512_mask_round (__B, __C, + _mm512_setzero_ph (), + __A, + _MM_FROUND_CUR_DIRECTION); } #ifdef __OPTIMIZE__ @@ -1510,9 +1510,9 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_scalef_round_ph (__m512h __A, __m512h __B, const int __C) { - return __builtin_ia32_vscalefph_v32hf_mask_round (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1, __C); + return __builtin_ia32_scalefph512_mask_round (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1, __C); } extern __inline __m512h @@ -1520,8 +1520,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_scalef_round_ph (__m512h __A, __mmask32 __B, __m512h __C, __m512h __D, const int __E) { - return __builtin_ia32_vscalefph_v32hf_mask_round (__C, __D, __A, __B, - __E); + return __builtin_ia32_scalefph512_mask_round (__C, __D, __A, __B, + __E); } extern __inline __m512h @@ -1529,24 +1529,24 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_scalef_round_ph (__mmask32 __A, __m512h __B, __m512h __C, const int __D) { - return __builtin_ia32_vscalefph_v32hf_mask_round (__B, __C, - _mm512_setzero_ph (), - __A, __D); + return __builtin_ia32_scalefph512_mask_round (__B, __C, + _mm512_setzero_ph (), + __A, __D); } #else -#define _mm512_scalef_round_ph(A, B, C) \ - (__builtin_ia32_vscalefph_v32hf_mask_round ((A), (B), \ - _mm512_setzero_ph (), \ - (__mmask32)-1, (C))) +#define _mm512_scalef_round_ph(A, B, C) \ + (__builtin_ia32_scalefph512_mask_round ((A), (B), \ + _mm512_setzero_ph (), \ + (__mmask32)-1, (C))) #define _mm512_mask_scalef_round_ph(A, B, C, D, E) \ - (__builtin_ia32_vscalefph_v32hf_mask_round ((C), (D), (A), (B), (E))) + (__builtin_ia32_scalefph512_mask_round ((C), (D), (A), (B), (E))) -#define _mm512_maskz_scalef_round_ph(A, B, C, D) \ - (__builtin_ia32_vscalefph_v32hf_mask_round ((B), (C), \ - _mm512_setzero_ph (), \ - (A), (D))) +#define _mm512_maskz_scalef_round_ph(A, B, C, D) \ + (__builtin_ia32_scalefph512_mask_round ((B), (C), \ + _mm512_setzero_ph (), \ + (A), (D))) #endif /* __OPTIMIZE__ */ @@ -1555,28 +1555,28 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_scalef_sh (__m128h __A, __m128h __B) { - return __builtin_ia32_vscalefsh_v8hf_mask_round (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_scalefsh_mask_round (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_scalef_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vscalefsh_v8hf_mask_round (__C, __D, __A, __B, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_scalefsh_mask_round (__C, __D, __A, __B, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_scalef_sh (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vscalefsh_v8hf_mask_round (__B, __C, - _mm_setzero_ph (), - __A, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_scalefsh_mask_round (__B, __C, + _mm_setzero_ph (), + __A, + _MM_FROUND_CUR_DIRECTION); } #ifdef __OPTIMIZE__ @@ -1584,9 +1584,9 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_scalef_round_sh (__m128h __A, __m128h __B, const int __C) { - return __builtin_ia32_vscalefsh_v8hf_mask_round (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1, __C); + return __builtin_ia32_scalefsh_mask_round (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1, __C); } extern __inline __m128h @@ -1594,8 +1594,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_scalef_round_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D, const int __E) { - return __builtin_ia32_vscalefsh_v8hf_mask_round (__C, __D, __A, __B, - __E); + return __builtin_ia32_scalefsh_mask_round (__C, __D, __A, __B, + __E); } extern __inline __m128h @@ -1603,23 +1603,23 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_scalef_round_sh (__mmask8 __A, __m128h __B, __m128h __C, const int __D) { - return __builtin_ia32_vscalefsh_v8hf_mask_round (__B, __C, - _mm_setzero_ph (), - __A, __D); + return __builtin_ia32_scalefsh_mask_round (__B, __C, + _mm_setzero_ph (), + __A, __D); } #else -#define _mm_scalef_round_sh(A, B, C) \ - (__builtin_ia32_vscalefsh_v8hf_mask_round ((A), (B), \ - _mm_setzero_ph (), \ - (__mmask8)-1, (C))) +#define _mm_scalef_round_sh(A, B, C) \ + (__builtin_ia32_scalefsh_mask_round ((A), (B), \ + _mm_setzero_ph (), \ + (__mmask8)-1, (C))) -#define _mm_mask_scalef_round_sh(A, B, C, D, E) \ - (__builtin_ia32_vscalefsh_v8hf_mask_round ((C), (D), (A), (B), (E))) +#define _mm_mask_scalef_round_sh(A, B, C, D, E) \ + (__builtin_ia32_scalefsh_mask_round ((C), (D), (A), (B), (E))) -#define _mm_maskz_scalef_round_sh(A, B, C, D) \ - (__builtin_ia32_vscalefsh_v8hf_mask_round ((B), (C), _mm_setzero_ph (), \ - (A), (D))) +#define _mm_maskz_scalef_round_sh(A, B, C, D) \ + (__builtin_ia32_scalefsh_mask_round ((B), (C), _mm_setzero_ph (), \ + (A), (D))) #endif /* __OPTIMIZE__ */ @@ -1629,37 +1629,37 @@ extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_reduce_ph (__m512h __A, int __B) { - return __builtin_ia32_vreduceph_v32hf_mask_round (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_reduceph512_mask_round (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_reduce_ph (__m512h __A, __mmask32 __B, __m512h __C, int __D) { - return __builtin_ia32_vreduceph_v32hf_mask_round (__C, __D, __A, __B, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_reduceph512_mask_round (__C, __D, __A, __B, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_reduce_ph (__mmask32 __A, __m512h __B, int __C) { - return __builtin_ia32_vreduceph_v32hf_mask_round (__B, __C, - _mm512_setzero_ph (), - __A, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_reduceph512_mask_round (__B, __C, + _mm512_setzero_ph (), + __A, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_reduce_round_ph (__m512h __A, int __B, const int __C) { - return __builtin_ia32_vreduceph_v32hf_mask_round (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1, __C); + return __builtin_ia32_reduceph512_mask_round (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1, __C); } extern __inline __m512h @@ -1667,8 +1667,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_reduce_round_ph (__m512h __A, __mmask32 __B, __m512h __C, int __D, const int __E) { - return __builtin_ia32_vreduceph_v32hf_mask_round (__C, __D, __A, __B, - __E); + return __builtin_ia32_reduceph512_mask_round (__C, __D, __A, __B, + __E); } extern __inline __m512h @@ -1676,39 +1676,39 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_reduce_round_ph (__mmask32 __A, __m512h __B, int __C, const int __D) { - return __builtin_ia32_vreduceph_v32hf_mask_round (__B, __C, - _mm512_setzero_ph (), - __A, __D); + return __builtin_ia32_reduceph512_mask_round (__B, __C, + _mm512_setzero_ph (), + __A, __D); } #else #define _mm512_reduce_ph(A, B) \ - (__builtin_ia32_vreduceph_v32hf_mask_round ((A), (B), \ - _mm512_setzero_ph (), \ - (__mmask32)-1, \ - _MM_FROUND_CUR_DIRECTION)) + (__builtin_ia32_reduceph512_mask_round ((A), (B), \ + _mm512_setzero_ph (), \ + (__mmask32)-1, \ + _MM_FROUND_CUR_DIRECTION)) #define _mm512_mask_reduce_ph(A, B, C, D) \ - (__builtin_ia32_vreduceph_v32hf_mask_round ((C), (D), (A), (B), \ - _MM_FROUND_CUR_DIRECTION)) + (__builtin_ia32_reduceph512_mask_round ((C), (D), (A), (B), \ + _MM_FROUND_CUR_DIRECTION)) #define _mm512_maskz_reduce_ph(A, B, C) \ - (__builtin_ia32_vreduceph_v32hf_mask_round ((B), (C), \ - _mm512_setzero_ph (), \ - (A), _MM_FROUND_CUR_DIRECTION)) + (__builtin_ia32_reduceph512_mask_round ((B), (C), \ + _mm512_setzero_ph (), \ + (A), _MM_FROUND_CUR_DIRECTION)) -#define _mm512_reduce_round_ph(A, B, C) \ - (__builtin_ia32_vreduceph_v32hf_mask_round ((A), (B), \ - _mm512_setzero_ph (), \ - (__mmask32)-1, (C))) +#define _mm512_reduce_round_ph(A, B, C) \ + (__builtin_ia32_reduceph512_mask_round ((A), (B), \ + _mm512_setzero_ph (), \ + (__mmask32)-1, (C))) #define _mm512_mask_reduce_round_ph(A, B, C, D, E) \ - (__builtin_ia32_vreduceph_v32hf_mask_round ((C), (D), (A), (B), (E))) + (__builtin_ia32_reduceph512_mask_round ((C), (D), (A), (B), (E))) -#define _mm512_maskz_reduce_round_ph(A, B, C, D) \ - (__builtin_ia32_vreduceph_v32hf_mask_round ((B), (C), \ - _mm512_setzero_ph (), \ - (A), (D))) +#define _mm512_maskz_reduce_round_ph(A, B, C, D) \ + (__builtin_ia32_reduceph512_mask_round ((B), (C), \ + _mm512_setzero_ph (), \ + (A), (D))) #endif /* __OPTIMIZE__ */ @@ -1718,10 +1718,10 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_reduce_sh (__m128h __A, __m128h __B, int __C) { - return __builtin_ia32_vreducesh_v8hf_mask_round (__A, __B, __C, - _mm_setzero_ph (), - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_reducesh_mask_round (__A, __B, __C, + _mm_setzero_ph (), + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m128h @@ -1729,26 +1729,26 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_reduce_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D, int __E) { - return __builtin_ia32_vreducesh_v8hf_mask_round (__C, __D, __E, __A, __B, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_reducesh_mask_round (__C, __D, __E, __A, __B, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_reduce_sh (__mmask8 __A, __m128h __B, __m128h __C, int __D) { - return __builtin_ia32_vreducesh_v8hf_mask_round (__B, __C, __D, - _mm_setzero_ph (), __A, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_reducesh_mask_round (__B, __C, __D, + _mm_setzero_ph (), __A, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_reduce_round_sh (__m128h __A, __m128h __B, int __C, const int __D) { - return __builtin_ia32_vreducesh_v8hf_mask_round (__A, __B, __C, - _mm_setzero_ph (), - (__mmask8) -1, __D); + return __builtin_ia32_reducesh_mask_round (__A, __B, __C, + _mm_setzero_ph (), + (__mmask8) -1, __D); } extern __inline __m128h @@ -1756,8 +1756,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_reduce_round_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D, int __E, const int __F) { - return __builtin_ia32_vreducesh_v8hf_mask_round (__C, __D, __E, __A, - __B, __F); + return __builtin_ia32_reducesh_mask_round (__C, __D, __E, __A, + __B, __F); } extern __inline __m128h @@ -1765,81 +1765,81 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_reduce_round_sh (__mmask8 __A, __m128h __B, __m128h __C, int __D, const int __E) { - return __builtin_ia32_vreducesh_v8hf_mask_round (__B, __C, __D, - _mm_setzero_ph (), - __A, __E); + return __builtin_ia32_reducesh_mask_round (__B, __C, __D, + _mm_setzero_ph (), + __A, __E); } #else #define _mm_reduce_sh(A, B, C) \ - (__builtin_ia32_vreducesh_v8hf_mask_round ((A), (B), (C), \ - _mm_setzero_ph (), \ - (__mmask8)-1, \ - _MM_FROUND_CUR_DIRECTION)) + (__builtin_ia32_reducesh_mask_round ((A), (B), (C), \ + _mm_setzero_ph (), \ + (__mmask8)-1, \ + _MM_FROUND_CUR_DIRECTION)) #define _mm_mask_reduce_sh(A, B, C, D, E) \ - (__builtin_ia32_vreducesh_v8hf_mask_round ((C), (D), (E), (A), (B), \ - _MM_FROUND_CUR_DIRECTION)) + (__builtin_ia32_reducesh_mask_round ((C), (D), (E), (A), (B), \ + _MM_FROUND_CUR_DIRECTION)) #define _mm_maskz_reduce_sh(A, B, C, D) \ - (__builtin_ia32_vreducesh_v8hf_mask_round ((B), (C), (D), \ - _mm_setzero_ph (), \ - (A), _MM_FROUND_CUR_DIRECTION)) + (__builtin_ia32_reducesh_mask_round ((B), (C), (D), \ + _mm_setzero_ph (), \ + (A), _MM_FROUND_CUR_DIRECTION)) #define _mm_reduce_round_sh(A, B, C, D) \ - (__builtin_ia32_vreducesh_v8hf_mask_round ((A), (B), (C), \ - _mm_setzero_ph (), \ - (__mmask8)-1, (D))) + (__builtin_ia32_reducesh_mask_round ((A), (B), (C), \ + _mm_setzero_ph (), \ + (__mmask8)-1, (D))) #define _mm_mask_reduce_round_sh(A, B, C, D, E, F) \ - (__builtin_ia32_vreducesh_v8hf_mask_round ((C), (D), (E), (A), (B), (F))) + (__builtin_ia32_reducesh_mask_round ((C), (D), (E), (A), (B), (F))) #define _mm_maskz_reduce_round_sh(A, B, C, D, E) \ - (__builtin_ia32_vreducesh_v8hf_mask_round ((B), (C), (D), \ - _mm_setzero_ph (), \ - (A), (E))) + (__builtin_ia32_reducesh_mask_round ((B), (C), (D), \ + _mm_setzero_ph (), \ + (A), (E))) #endif /* __OPTIMIZE__ */ /* Intrinsics vrndscaleph. */ #ifdef __OPTIMIZE__ extern __inline __m512h -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_roundscale_ph (__m512h __A, int __B) { - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_rndscaleph512_mask_round (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_roundscale_ph (__m512h __A, __mmask32 __B, - __m512h __C, int __D) + __m512h __C, int __D) { - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__C, __D, __A, __B, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_rndscaleph512_mask_round (__C, __D, __A, __B, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_roundscale_ph (__mmask32 __A, __m512h __B, int __C) { - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__B, __C, - _mm512_setzero_ph (), - __A, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_rndscaleph512_mask_round (__B, __C, + _mm512_setzero_ph (), + __A, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m512h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_roundscale_round_ph (__m512h __A, int __B, const int __C) { - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__A, __B, - _mm512_setzero_ph (), - (__mmask32) -1, - __C); + return __builtin_ia32_rndscaleph512_mask_round (__A, __B, + _mm512_setzero_ph (), + (__mmask32) -1, + __C); } extern __inline __m512h @@ -1847,8 +1847,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_mask_roundscale_round_ph (__m512h __A, __mmask32 __B, __m512h __C, int __D, const int __E) { - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__C, __D, __A, - __B, __E); + return __builtin_ia32_rndscaleph512_mask_round (__C, __D, __A, + __B, __E); } extern __inline __m512h @@ -1856,52 +1856,52 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_roundscale_round_ph (__mmask32 __A, __m512h __B, int __C, const int __D) { - return __builtin_ia32_vrndscaleph_v32hf_mask_round (__B, __C, - _mm512_setzero_ph (), - __A, __D); + return __builtin_ia32_rndscaleph512_mask_round (__B, __C, + _mm512_setzero_ph (), + __A, __D); } #else -#define _mm512_roundscale_ph(A, B) \ - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((A), (B), \ - _mm512_setzero_ph (), \ - (__mmask32)-1, \ - _MM_FROUND_CUR_DIRECTION)) - -#define _mm512_mask_roundscale_ph(A, B, C, D) \ - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((C), (D), (A), (B), \ - _MM_FROUND_CUR_DIRECTION)) - -#define _mm512_maskz_roundscale_ph(A, B, C) \ - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((B), (C), \ - _mm512_setzero_ph (), \ - (A), \ - _MM_FROUND_CUR_DIRECTION)) -#define _mm512_roundscale_round_ph(A, B, C) \ - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((A), (B), \ - _mm512_setzero_ph (), \ - (__mmask32)-1, (C))) +#define _mm512_roundscale_ph(A, B) \ + (__builtin_ia32_rndscaleph512_mask_round ((A), (B), \ + _mm512_setzero_ph (), \ + (__mmask32)-1, \ + _MM_FROUND_CUR_DIRECTION)) + +#define _mm512_mask_roundscale_ph(A, B, C, D) \ + (__builtin_ia32_rndscaleph512_mask_round ((C), (D), (A), (B), \ + _MM_FROUND_CUR_DIRECTION)) + +#define _mm512_maskz_roundscale_ph(A, B, C) \ + (__builtin_ia32_rndscaleph512_mask_round ((B), (C), \ + _mm512_setzero_ph (), \ + (A), \ + _MM_FROUND_CUR_DIRECTION)) +#define _mm512_roundscale_round_ph(A, B, C) \ + (__builtin_ia32_rndscaleph512_mask_round ((A), (B), \ + _mm512_setzero_ph (), \ + (__mmask32)-1, (C))) #define _mm512_mask_roundscale_round_ph(A, B, C, D, E) \ - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((C), (D), (A), (B), (E))) + (__builtin_ia32_rndscaleph512_mask_round ((C), (D), (A), (B), (E))) -#define _mm512_maskz_roundscale_round_ph(A, B, C, D) \ - (__builtin_ia32_vrndscaleph_v32hf_mask_round ((B), (C), \ - _mm512_setzero_ph (), \ - (A), (D))) +#define _mm512_maskz_roundscale_round_ph(A, B, C, D) \ + (__builtin_ia32_rndscaleph512_mask_round ((B), (C), \ + _mm512_setzero_ph (), \ + (A), (D))) #endif /* __OPTIMIZE__ */ /* Intrinsics vrndscalesh. */ #ifdef __OPTIMIZE__ extern __inline __m128h -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_roundscale_sh (__m128h __A, __m128h __B, int __C) { - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__A, __B, __C, - _mm_setzero_ph (), - (__mmask8) -1, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_rndscalesh_mask_round (__A, __B, __C, + _mm_setzero_ph (), + (__mmask8) -1, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m128h @@ -1909,27 +1909,27 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_roundscale_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D, int __E) { - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__C, __D, __E, __A, __B, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_rndscalesh_mask_round (__C, __D, __E, __A, __B, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_roundscale_sh (__mmask8 __A, __m128h __B, __m128h __C, int __D) { - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__B, __C, __D, - _mm_setzero_ph (), __A, - _MM_FROUND_CUR_DIRECTION); + return __builtin_ia32_rndscalesh_mask_round (__B, __C, __D, + _mm_setzero_ph (), __A, + _MM_FROUND_CUR_DIRECTION); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_roundscale_round_sh (__m128h __A, __m128h __B, int __C, const int __D) { - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__A, __B, __C, - _mm_setzero_ph (), - (__mmask8) -1, - __D); + return __builtin_ia32_rndscalesh_mask_round (__A, __B, __C, + _mm_setzero_ph (), + (__mmask8) -1, + __D); } extern __inline __m128h @@ -1937,8 +1937,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_roundscale_round_sh (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D, int __E, const int __F) { - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__C, __D, __E, - __A, __B, __F); + return __builtin_ia32_rndscalesh_mask_round (__C, __D, __E, + __A, __B, __F); } extern __inline __m128h @@ -1946,46 +1946,46 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_roundscale_round_sh (__mmask8 __A, __m128h __B, __m128h __C, int __D, const int __E) { - return __builtin_ia32_vrndscalesh_v8hf_mask_round (__B, __C, __D, - _mm_setzero_ph (), - __A, __E); + return __builtin_ia32_rndscalesh_mask_round (__B, __C, __D, + _mm_setzero_ph (), + __A, __E); } #else #define _mm_roundscale_sh(A, B, C) \ - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((A), (B), (C), \ - _mm_setzero_ph (), \ - (__mmask8)-1, \ - _MM_FROUND_CUR_DIRECTION)) + (__builtin_ia32_rndscalesh_mask_round ((A), (B), (C), \ + _mm_setzero_ph (), \ + (__mmask8)-1, \ + _MM_FROUND_CUR_DIRECTION)) #define _mm_mask_roundscale_sh(A, B, C, D, E) \ - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((C), (D), (E), (A), (B), \ - _MM_FROUND_CUR_DIRECTION)) + (__builtin_ia32_rndscalesh_mask_round ((C), (D), (E), (A), (B), \ + _MM_FROUND_CUR_DIRECTION)) #define _mm_maskz_roundscale_sh(A, B, C, D) \ - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((B), (C), (D), \ - _mm_setzero_ph (), \ - (A), _MM_FROUND_CUR_DIRECTION)) + (__builtin_ia32_rndscalesh_mask_round ((B), (C), (D), \ + _mm_setzero_ph (), \ + (A), _MM_FROUND_CUR_DIRECTION)) -#define _mm_roundscale_round_sh(A, B, C, D) \ - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((A), (B), (C), \ - _mm_setzero_ph (), \ - (__mmask8)-1, (D))) +#define _mm_roundscale_round_sh(A, B, C, D) \ + (__builtin_ia32_rndscalesh_mask_round ((A), (B), (C), \ + _mm_setzero_ph (), \ + (__mmask8)-1, (D))) #define _mm_mask_roundscale_round_sh(A, B, C, D, E, F) \ - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((C), (D), (E), (A), (B), (F))) + (__builtin_ia32_rndscalesh_mask_round ((C), (D), (E), (A), (B), (F))) -#define _mm_maskz_roundscale_round_sh(A, B, C, D, E) \ - (__builtin_ia32_vrndscalesh_v8hf_mask_round ((B), (C), (D), \ - _mm_setzero_ph (), \ - (A), (E))) +#define _mm_maskz_roundscale_round_sh(A, B, C, D, E) \ + (__builtin_ia32_rndscalesh_mask_round ((B), (C), (D), \ + _mm_setzero_ph (), \ + (A), (E))) #endif /* __OPTIMIZE__ */ /* Intrinsics vfpclasssh. */ #ifdef __OPTIMIZE__ extern __inline __mmask8 -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_fpclass_sh_mask (__m128h __A, const int __imm) { return (__mmask8) __builtin_ia32_fpclasssh_mask ((__v8hf) __A, __imm, @@ -2031,11 +2031,11 @@ _mm512_fpclass_ph_mask (__m512h __A, const int __imm) #else #define _mm512_mask_fpclass_ph_mask(u, x, c) \ - ((__mmask32) __builtin_ia32_fpclassph512_mask ((__v32hf) (__m512h) (x),\ + ((__mmask32) __builtin_ia32_fpclassph512_mask ((__v32hf) (__m512h) (x), \ (int) (c),(__mmask8)(u))) #define _mm512_fpclass_ph_mask(x, c) \ - ((__mmask32) __builtin_ia32_fpclassph512_mask ((__v32hf) (__m512h) (x),\ + ((__mmask32) __builtin_ia32_fpclassph512_mask ((__v32hf) (__m512h) (x), \ (int) (c),(__mmask8)-1)) #endif /* __OPIMTIZE__ */ @@ -2141,9 +2141,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_getexp_round_ph (__m512h __A, const int __R) { return (__m512h) __builtin_ia32_getexpph512_mask ((__v32hf) __A, - (__v32hf) - _mm512_setzero_ph (), - (__mmask32) -1, __R); + (__v32hf) + _mm512_setzero_ph (), + (__mmask32) -1, __R); } extern __inline __m512h @@ -2152,8 +2152,8 @@ _mm512_mask_getexp_round_ph (__m512h __W, __mmask32 __U, __m512h __A, const int __R) { return (__m512h) __builtin_ia32_getexpph512_mask ((__v32hf) __A, - (__v32hf) __W, - (__mmask32) __U, __R); + (__v32hf) __W, + (__mmask32) __U, __R); } extern __inline __m512h @@ -2161,37 +2161,37 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm512_maskz_getexp_round_ph (__mmask32 __U, __m512h __A, const int __R) { return (__m512h) __builtin_ia32_getexpph512_mask ((__v32hf) __A, - (__v32hf) - _mm512_setzero_ph (), - (__mmask32) __U, __R); + (__v32hf) + _mm512_setzero_ph (), + (__mmask32) __U, __R); } #else -#define _mm_getexp_round_sh(A, B, R) \ - ((__m128h)__builtin_ia32_getexpsh_mask_round((__v8hf)(__m128h)(A), \ - (__v8hf)(__m128h)(B), \ - (__v8hf)_mm_setzero_ph(), \ +#define _mm_getexp_round_sh(A, B, R) \ + ((__m128h)__builtin_ia32_getexpsh_mask_round((__v8hf)(__m128h)(A), \ + (__v8hf)(__m128h)(B), \ + (__v8hf)_mm_setzero_ph(), \ (__mmask8)-1, R)) -#define _mm_mask_getexp_round_sh(W, U, A, B, C) \ +#define _mm_mask_getexp_round_sh(W, U, A, B, C) \ (__m128h)__builtin_ia32_getexpsh_mask_round(A, B, W, U, C) -#define _mm_maskz_getexp_round_sh(U, A, B, C) \ - (__m128h)__builtin_ia32_getexpsh_mask_round(A, B, \ - (__v8hf)_mm_setzero_ph(), \ +#define _mm_maskz_getexp_round_sh(U, A, B, C) \ + (__m128h)__builtin_ia32_getexpsh_mask_round(A, B, \ + (__v8hf)_mm_setzero_ph(), \ U, C) -#define _mm512_getexp_round_ph(A, R) \ - ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ - (__v32hf)_mm512_setzero_ph(), (__mmask32)-1, R)) +#define _mm512_getexp_round_ph(A, R) \ + ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ + (__v32hf)_mm512_setzero_ph(), (__mmask32)-1, R)) -#define _mm512_mask_getexp_round_ph(W, U, A, R) \ - ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ - (__v32hf)(__m512h)(W), (__mmask32)(U), R)) +#define _mm512_mask_getexp_round_ph(W, U, A, R) \ + ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ + (__v32hf)(__m512h)(W), (__mmask32)(U), R)) -#define _mm512_maskz_getexp_round_ph(U, A, R) \ - ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ - (__v32hf)_mm512_setzero_ph(), (__mmask32)(U), R)) +#define _mm512_maskz_getexp_round_ph(U, A, R) \ + ((__m512h)__builtin_ia32_getexpph512_mask((__v32hf)(__m512h)(A), \ + (__v32hf)_mm512_setzero_ph(), (__mmask32)(U), R)) #endif /* __OPTIMIZE__ */ diff --git a/gcc/config/i386/avx512fp16vlintrin.h b/gcc/config/i386/avx512fp16vlintrin.h index 206d60407fc..e9478792a03 100644 --- a/gcc/config/i386/avx512fp16vlintrin.h +++ b/gcc/config/i386/avx512fp16vlintrin.h @@ -53,30 +53,30 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_add_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vaddph_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_addph128_mask (__C, __D, __A, __B); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_add_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) { - return __builtin_ia32_vaddph_v16hf_mask (__C, __D, __A, __B); + return __builtin_ia32_addph256_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_add_ph (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vaddph_v8hf_mask (__B, __C, _mm_setzero_ph (), - __A); + return __builtin_ia32_addph128_mask (__B, __C, _mm_setzero_ph (), + __A); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_add_ph (__mmask16 __A, __m256h __B, __m256h __C) { - return __builtin_ia32_vaddph_v16hf_mask (__B, __C, - _mm256_setzero_ph (), __A); + return __builtin_ia32_addph256_mask (__B, __C, + _mm256_setzero_ph (), __A); } extern __inline __m128h @@ -97,30 +97,30 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_sub_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vsubph_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_subph128_mask (__C, __D, __A, __B); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_sub_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) { - return __builtin_ia32_vsubph_v16hf_mask (__C, __D, __A, __B); + return __builtin_ia32_subph256_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_sub_ph (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vsubph_v8hf_mask (__B, __C, _mm_setzero_ph (), - __A); + return __builtin_ia32_subph128_mask (__B, __C, _mm_setzero_ph (), + __A); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_sub_ph (__mmask16 __A, __m256h __B, __m256h __C) { - return __builtin_ia32_vsubph_v16hf_mask (__B, __C, - _mm256_setzero_ph (), __A); + return __builtin_ia32_subph256_mask (__B, __C, + _mm256_setzero_ph (), __A); } extern __inline __m128h @@ -141,30 +141,30 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_mul_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vmulph_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_mulph128_mask (__C, __D, __A, __B); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_mul_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) { - return __builtin_ia32_vmulph_v16hf_mask (__C, __D, __A, __B); + return __builtin_ia32_mulph256_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_mul_ph (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vmulph_v8hf_mask (__B, __C, _mm_setzero_ph (), - __A); + return __builtin_ia32_mulph128_mask (__B, __C, _mm_setzero_ph (), + __A); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_mul_ph (__mmask16 __A, __m256h __B, __m256h __C) { - return __builtin_ia32_vmulph_v16hf_mask (__B, __C, - _mm256_setzero_ph (), __A); + return __builtin_ia32_mulph256_mask (__B, __C, + _mm256_setzero_ph (), __A); } extern __inline __m128h @@ -185,30 +185,30 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_div_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vdivph_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_divph128_mask (__C, __D, __A, __B); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_div_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) { - return __builtin_ia32_vdivph_v16hf_mask (__C, __D, __A, __B); + return __builtin_ia32_divph256_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_div_ph (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vdivph_v8hf_mask (__B, __C, _mm_setzero_ph (), - __A); + return __builtin_ia32_divph128_mask (__B, __C, _mm_setzero_ph (), + __A); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_div_ph (__mmask16 __A, __m256h __B, __m256h __C) { - return __builtin_ia32_vdivph_v16hf_mask (__B, __C, - _mm256_setzero_ph (), __A); + return __builtin_ia32_divph256_mask (__B, __C, + _mm256_setzero_ph (), __A); } /* Intrinsics v[max,min]ph. */ @@ -216,96 +216,96 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_max_ph (__m128h __A, __m128h __B) { - return __builtin_ia32_vmaxph_v8hf_mask (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1); + return __builtin_ia32_maxph128_mask (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_max_ph (__m256h __A, __m256h __B) { - return __builtin_ia32_vmaxph_v16hf_mask (__A, __B, - _mm256_setzero_ph (), - (__mmask16) -1); + return __builtin_ia32_maxph256_mask (__A, __B, + _mm256_setzero_ph (), + (__mmask16) -1); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_max_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vmaxph_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_maxph128_mask (__C, __D, __A, __B); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_max_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) { - return __builtin_ia32_vmaxph_v16hf_mask (__C, __D, __A, __B); + return __builtin_ia32_maxph256_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_max_ph (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vmaxph_v8hf_mask (__B, __C, _mm_setzero_ph (), - __A); + return __builtin_ia32_maxph128_mask (__B, __C, _mm_setzero_ph (), + __A); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_max_ph (__mmask16 __A, __m256h __B, __m256h __C) { - return __builtin_ia32_vmaxph_v16hf_mask (__B, __C, - _mm256_setzero_ph (), __A); + return __builtin_ia32_maxph256_mask (__B, __C, + _mm256_setzero_ph (), __A); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_min_ph (__m128h __A, __m128h __B) { - return __builtin_ia32_vminph_v8hf_mask (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1); + return __builtin_ia32_minph128_mask (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_min_ph (__m256h __A, __m256h __B) { - return __builtin_ia32_vminph_v16hf_mask (__A, __B, - _mm256_setzero_ph (), - (__mmask16) -1); + return __builtin_ia32_minph256_mask (__A, __B, + _mm256_setzero_ph (), + (__mmask16) -1); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_min_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vminph_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_minph128_mask (__C, __D, __A, __B); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_min_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) { - return __builtin_ia32_vminph_v16hf_mask (__C, __D, __A, __B); + return __builtin_ia32_minph256_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_min_ph (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vminph_v8hf_mask (__B, __C, _mm_setzero_ph (), - __A); + return __builtin_ia32_minph128_mask (__B, __C, _mm_setzero_ph (), + __A); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_min_ph (__mmask16 __A, __m256h __B, __m256h __C) { - return __builtin_ia32_vminph_v16hf_mask (__B, __C, - _mm256_setzero_ph (), __A); + return __builtin_ia32_minph256_mask (__B, __C, + _mm256_setzero_ph (), __A); } /* vcmpph */ @@ -314,8 +314,8 @@ extern __inline __mmask8 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_cmp_ph_mask (__m128h __A, __m128h __B, const int __C) { - return (__mmask8) __builtin_ia32_vcmpph_v8hf_mask (__A, __B, __C, - (__mmask8) -1); + return (__mmask8) __builtin_ia32_cmpph128_mask (__A, __B, __C, + (__mmask8) -1); } extern __inline __mmask8 @@ -323,15 +323,15 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_cmp_ph_mask (__mmask8 __A, __m128h __B, __m128h __C, const int __D) { - return (__mmask8) __builtin_ia32_vcmpph_v8hf_mask (__B, __C, __D, __A); + return (__mmask8) __builtin_ia32_cmpph128_mask (__B, __C, __D, __A); } extern __inline __mmask16 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_cmp_ph_mask (__m256h __A, __m256h __B, const int __C) { - return (__mmask16) __builtin_ia32_vcmpph_v16hf_mask (__A, __B, __C, - (__mmask16) -1); + return (__mmask16) __builtin_ia32_cmpph256_mask (__A, __B, __C, + (__mmask16) -1); } extern __inline __mmask16 @@ -339,22 +339,22 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_cmp_ph_mask (__mmask16 __A, __m256h __B, __m256h __C, const int __D) { - return (__mmask16) __builtin_ia32_vcmpph_v16hf_mask (__B, __C, __D, - __A); + return (__mmask16) __builtin_ia32_cmpph256_mask (__B, __C, __D, + __A); } #else -#define _mm_cmp_ph_mask(A, B, C) \ - (__builtin_ia32_vcmpph_v8hf_mask ((A), (B), (C), (-1))) +#define _mm_cmp_ph_mask(A, B, C) \ + (__builtin_ia32_cmpph128_mask ((A), (B), (C), (-1))) -#define _mm_mask_cmp_ph_mask(A, B, C, D) \ - (__builtin_ia32_vcmpph_v8hf_mask ((B), (C), (D), (A))) +#define _mm_mask_cmp_ph_mask(A, B, C, D) \ + (__builtin_ia32_cmpph128_mask ((B), (C), (D), (A))) -#define _mm256_cmp_ph_mask(A, B, C) \ - (__builtin_ia32_vcmpph_v16hf_mask ((A), (B), (C), (-1))) +#define _mm256_cmp_ph_mask(A, B, C) \ + (__builtin_ia32_cmpph256_mask ((A), (B), (C), (-1))) -#define _mm256_mask_cmp_ph_mask(A, B, C, D) \ - (__builtin_ia32_vcmpph_v16hf_mask ((B), (C), (D), (A))) +#define _mm256_mask_cmp_ph_mask(A, B, C, D) \ + (__builtin_ia32_cmpph256_mask ((B), (C), (D), (A))) #endif /* __OPTIMIZE__ */ @@ -363,46 +363,46 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_sqrt_ph (__m128h __A) { - return __builtin_ia32_vsqrtph_v8hf_mask (__A, _mm_setzero_ph (), - (__mmask8) -1); + return __builtin_ia32_sqrtph128_mask (__A, _mm_setzero_ph (), + (__mmask8) -1); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_sqrt_ph (__m256h __A) { - return __builtin_ia32_vsqrtph_v16hf_mask (__A, _mm256_setzero_ph (), - (__mmask16) -1); + return __builtin_ia32_sqrtph256_mask (__A, _mm256_setzero_ph (), + (__mmask16) -1); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_sqrt_ph (__m128h __A, __mmask8 __B, __m128h __C) { - return __builtin_ia32_vsqrtph_v8hf_mask (__C, __A, __B); + return __builtin_ia32_sqrtph128_mask (__C, __A, __B); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_sqrt_ph (__m256h __A, __mmask16 __B, __m256h __C) { - return __builtin_ia32_vsqrtph_v16hf_mask (__C, __A, __B); + return __builtin_ia32_sqrtph256_mask (__C, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_sqrt_ph (__mmask8 __A, __m128h __B) { - return __builtin_ia32_vsqrtph_v8hf_mask (__B, _mm_setzero_ph (), - __A); + return __builtin_ia32_sqrtph128_mask (__B, _mm_setzero_ph (), + __A); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_sqrt_ph (__mmask16 __A, __m256h __B) { - return __builtin_ia32_vsqrtph_v16hf_mask (__B, _mm256_setzero_ph (), - __A); + return __builtin_ia32_sqrtph256_mask (__B, _mm256_setzero_ph (), + __A); } /* Intrinsics vrsqrtph. */ @@ -410,45 +410,45 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_rsqrt_ph (__m128h __A) { - return __builtin_ia32_vrsqrtph_v8hf_mask (__A, _mm_setzero_ph (), - (__mmask8) -1); + return __builtin_ia32_rsqrtph128_mask (__A, _mm_setzero_ph (), + (__mmask8) -1); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_rsqrt_ph (__m256h __A) { - return __builtin_ia32_vrsqrtph_v16hf_mask (__A, _mm256_setzero_ph (), - (__mmask16) -1); + return __builtin_ia32_rsqrtph256_mask (__A, _mm256_setzero_ph (), + (__mmask16) -1); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_rsqrt_ph (__m128h __A, __mmask8 __B, __m128h __C) { - return __builtin_ia32_vrsqrtph_v8hf_mask (__C, __A, __B); + return __builtin_ia32_rsqrtph128_mask (__C, __A, __B); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_rsqrt_ph (__m256h __A, __mmask16 __B, __m256h __C) { - return __builtin_ia32_vrsqrtph_v16hf_mask (__C, __A, __B); + return __builtin_ia32_rsqrtph256_mask (__C, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_rsqrt_ph (__mmask8 __A, __m128h __B) { - return __builtin_ia32_vrsqrtph_v8hf_mask (__B, _mm_setzero_ph (), __A); + return __builtin_ia32_rsqrtph128_mask (__B, _mm_setzero_ph (), __A); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_rsqrt_ph (__mmask16 __A, __m256h __B) { - return __builtin_ia32_vrsqrtph_v16hf_mask (__B, _mm256_setzero_ph (), - __A); + return __builtin_ia32_rsqrtph256_mask (__B, _mm256_setzero_ph (), + __A); } /* Intrinsics vrcpph. */ @@ -456,45 +456,45 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_rcp_ph (__m128h __A) { - return __builtin_ia32_vrcpph_v8hf_mask (__A, _mm_setzero_ph (), - (__mmask8) -1); + return __builtin_ia32_rcpph128_mask (__A, _mm_setzero_ph (), + (__mmask8) -1); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_rcp_ph (__m256h __A) { - return __builtin_ia32_vrcpph_v16hf_mask (__A, _mm256_setzero_ph (), - (__mmask16) -1); + return __builtin_ia32_rcpph256_mask (__A, _mm256_setzero_ph (), + (__mmask16) -1); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_rcp_ph (__m128h __A, __mmask8 __B, __m128h __C) { - return __builtin_ia32_vrcpph_v8hf_mask (__C, __A, __B); + return __builtin_ia32_rcpph128_mask (__C, __A, __B); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_rcp_ph (__m256h __A, __mmask16 __B, __m256h __C) { - return __builtin_ia32_vrcpph_v16hf_mask (__C, __A, __B); + return __builtin_ia32_rcpph256_mask (__C, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_rcp_ph (__mmask8 __A, __m128h __B) { - return __builtin_ia32_vrcpph_v8hf_mask (__B, _mm_setzero_ph (), __A); + return __builtin_ia32_rcpph128_mask (__B, _mm_setzero_ph (), __A); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_rcp_ph (__mmask16 __A, __m256h __B) { - return __builtin_ia32_vrcpph_v16hf_mask (__B, _mm256_setzero_ph (), - __A); + return __builtin_ia32_rcpph256_mask (__B, _mm256_setzero_ph (), + __A); } /* Intrinsics vscalefph. */ @@ -502,25 +502,25 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_scalef_ph (__m128h __A, __m128h __B) { - return __builtin_ia32_vscalefph_v8hf_mask (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1); + return __builtin_ia32_scalefph128_mask (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_scalef_ph (__m256h __A, __m256h __B) { - return __builtin_ia32_vscalefph_v16hf_mask (__A, __B, - _mm256_setzero_ph (), - (__mmask16) -1); + return __builtin_ia32_scalefph256_mask (__A, __B, + _mm256_setzero_ph (), + (__mmask16) -1); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_scalef_ph (__m128h __A, __mmask8 __B, __m128h __C, __m128h __D) { - return __builtin_ia32_vscalefph_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_scalefph128_mask (__C, __D, __A, __B); } extern __inline __m256h @@ -528,24 +528,24 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_scalef_ph (__m256h __A, __mmask16 __B, __m256h __C, __m256h __D) { - return __builtin_ia32_vscalefph_v16hf_mask (__C, __D, __A, __B); + return __builtin_ia32_scalefph256_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_scalef_ph (__mmask8 __A, __m128h __B, __m128h __C) { - return __builtin_ia32_vscalefph_v8hf_mask (__B, __C, - _mm_setzero_ph (), __A); + return __builtin_ia32_scalefph128_mask (__B, __C, + _mm_setzero_ph (), __A); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_scalef_ph (__mmask16 __A, __m256h __B, __m256h __C) { - return __builtin_ia32_vscalefph_v16hf_mask (__B, __C, - _mm256_setzero_ph (), - __A); + return __builtin_ia32_scalefph256_mask (__B, __C, + _mm256_setzero_ph (), + __A); } /* Intrinsics vreduceph. */ @@ -554,109 +554,109 @@ extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_reduce_ph (__m128h __A, int __B) { - return __builtin_ia32_vreduceph_v8hf_mask (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1); + return __builtin_ia32_reduceph128_mask (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_reduce_ph (__m128h __A, __mmask8 __B, __m128h __C, int __D) { - return __builtin_ia32_vreduceph_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_reduceph128_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_reduce_ph (__mmask8 __A, __m128h __B, int __C) { - return __builtin_ia32_vreduceph_v8hf_mask (__B, __C, - _mm_setzero_ph (), __A); + return __builtin_ia32_reduceph128_mask (__B, __C, + _mm_setzero_ph (), __A); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_reduce_ph (__m256h __A, int __B) { - return __builtin_ia32_vreduceph_v16hf_mask (__A, __B, - _mm256_setzero_ph (), - (__mmask16) -1); + return __builtin_ia32_reduceph256_mask (__A, __B, + _mm256_setzero_ph (), + (__mmask16) -1); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_reduce_ph (__m256h __A, __mmask16 __B, __m256h __C, int __D) { - return __builtin_ia32_vreduceph_v16hf_mask (__C, __D, __A, __B); + return __builtin_ia32_reduceph256_mask (__C, __D, __A, __B); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_reduce_ph (__mmask16 __A, __m256h __B, int __C) { - return __builtin_ia32_vreduceph_v16hf_mask (__B, __C, - _mm256_setzero_ph (), - __A); + return __builtin_ia32_reduceph256_mask (__B, __C, + _mm256_setzero_ph (), + __A); } #else -#define _mm_reduce_ph(A, B) \ - (__builtin_ia32_vreduceph_v8hf_mask ((A), (B),\ - _mm_setzero_ph (), \ - ((__mmask8)-1))) +#define _mm_reduce_ph(A, B) \ + (__builtin_ia32_reduceph128_mask ((A), (B), \ + _mm_setzero_ph (), \ + ((__mmask8)-1))) -#define _mm_mask_reduce_ph(A, B, C, D) \ - (__builtin_ia32_vreduceph_v8hf_mask ((C), (D), (A), (B))) +#define _mm_mask_reduce_ph(A, B, C, D) \ + (__builtin_ia32_reduceph128_mask ((C), (D), (A), (B))) -#define _mm_maskz_reduce_ph(A, B, C) \ - (__builtin_ia32_vreduceph_v8hf_mask ((B), (C), _mm_setzero_ph (), (A))) +#define _mm_maskz_reduce_ph(A, B, C) \ + (__builtin_ia32_reduceph128_mask ((B), (C), _mm_setzero_ph (), (A))) #define _mm256_reduce_ph(A, B) \ - (__builtin_ia32_vreduceph_v16hf_mask ((A), (B),\ - _mm256_setzero_ph (), \ - ((__mmask16)-1))) + (__builtin_ia32_reduceph256_mask ((A), (B), \ + _mm256_setzero_ph (), \ + ((__mmask16)-1))) -#define _mm256_mask_reduce_ph(A, B, C, D) \ - (__builtin_ia32_vreduceph_v16hf_mask ((C), (D), (A), (B))) +#define _mm256_mask_reduce_ph(A, B, C, D) \ + (__builtin_ia32_reduceph256_mask ((C), (D), (A), (B))) -#define _mm256_maskz_reduce_ph(A, B, C) \ - (__builtin_ia32_vreduceph_v16hf_mask ((B), (C), _mm256_setzero_ph (), (A))) +#define _mm256_maskz_reduce_ph(A, B, C) \ + (__builtin_ia32_reduceph256_mask ((B), (C), _mm256_setzero_ph (), (A))) #endif /* __OPTIMIZE__ */ /* Intrinsics vrndscaleph. */ #ifdef __OPTIMIZE__ -extern __inline __m128h -__attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) -_mm_roundscale_ph (__m128h __A, int __B) -{ - return __builtin_ia32_vrndscaleph_v8hf_mask (__A, __B, - _mm_setzero_ph (), - (__mmask8) -1); -} + extern __inline __m128h + __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) + _mm_roundscale_ph (__m128h __A, int __B) + { + return __builtin_ia32_rndscaleph128_mask (__A, __B, + _mm_setzero_ph (), + (__mmask8) -1); + } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_roundscale_ph (__m128h __A, __mmask8 __B, __m128h __C, int __D) { - return __builtin_ia32_vrndscaleph_v8hf_mask (__C, __D, __A, __B); + return __builtin_ia32_rndscaleph128_mask (__C, __D, __A, __B); } extern __inline __m128h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_roundscale_ph (__mmask8 __A, __m128h __B, int __C) { - return __builtin_ia32_vrndscaleph_v8hf_mask (__B, __C, - _mm_setzero_ph (), __A); + return __builtin_ia32_rndscaleph128_mask (__B, __C, + _mm_setzero_ph (), __A); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_roundscale_ph (__m256h __A, int __B) { - return __builtin_ia32_vrndscaleph_v16hf_mask (__A, __B, - _mm256_setzero_ph (), - (__mmask16) -1); + return __builtin_ia32_rndscaleph256_mask (__A, __B, + _mm256_setzero_ph (), + (__mmask16) -1); } extern __inline __m256h @@ -664,40 +664,40 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_roundscale_ph (__m256h __A, __mmask16 __B, __m256h __C, int __D) { - return __builtin_ia32_vrndscaleph_v16hf_mask (__C, __D, __A, __B); + return __builtin_ia32_rndscaleph256_mask (__C, __D, __A, __B); } extern __inline __m256h __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_roundscale_ph (__mmask16 __A, __m256h __B, int __C) { - return __builtin_ia32_vrndscaleph_v16hf_mask (__B, __C, - _mm256_setzero_ph (), - __A); + return __builtin_ia32_rndscaleph256_mask (__B, __C, + _mm256_setzero_ph (), + __A); } #else -#define _mm_roundscale_ph(A, B) \ - (__builtin_ia32_vrndscaleph_v8hf_mask ((A), (B), _mm_setzero_ph (), \ - ((__mmask8)-1))) +#define _mm_roundscale_ph(A, B) \ + (__builtin_ia32_rndscaleph128_mask ((A), (B), _mm_setzero_ph (), \ + ((__mmask8)-1))) -#define _mm_mask_roundscale_ph(A, B, C, D) \ - (__builtin_ia32_vrndscaleph_v8hf_mask ((C), (D), (A), (B))) +#define _mm_mask_roundscale_ph(A, B, C, D) \ + (__builtin_ia32_rndscaleph128_mask ((C), (D), (A), (B))) -#define _mm_maskz_roundscale_ph(A, B, C) \ - (__builtin_ia32_vrndscaleph_v8hf_mask ((B), (C), _mm_setzero_ph (), (A))) +#define _mm_maskz_roundscale_ph(A, B, C) \ + (__builtin_ia32_rndscaleph128_mask ((B), (C), _mm_setzero_ph (), (A))) -#define _mm256_roundscale_ph(A, B) \ - (__builtin_ia32_vrndscaleph_v16hf_mask ((A), (B), \ - _mm256_setzero_ph(), \ - ((__mmask16)-1))) +#define _mm256_roundscale_ph(A, B) \ + (__builtin_ia32_rndscaleph256_mask ((A), (B), \ + _mm256_setzero_ph(), \ + ((__mmask16)-1))) -#define _mm256_mask_roundscale_ph(A, B, C, D) \ - (__builtin_ia32_vrndscaleph_v16hf_mask ((C), (D), (A), (B))) +#define _mm256_mask_roundscale_ph(A, B, C, D) \ + (__builtin_ia32_rndscaleph256_mask ((C), (D), (A), (B))) -#define _mm256_maskz_roundscale_ph(A, B, C) \ - (__builtin_ia32_vrndscaleph_v16hf_mask ((B), (C), \ - _mm256_setzero_ph (), (A))) +#define _mm256_maskz_roundscale_ph(A, B, C) \ + (__builtin_ia32_rndscaleph256_mask ((B), (C), \ + _mm256_setzero_ph (), (A))) #endif /* __OPTIMIZE__ */ @@ -705,7 +705,7 @@ _mm256_maskz_roundscale_ph (__mmask16 __A, __m256h __B, int __C) #ifdef __OPTIMIZE__ extern __inline __mmask8 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) -_mm_mask_fpclass_ph_mask (__mmask8 __U, __m128h __A, const int __imm) + _mm_mask_fpclass_ph_mask (__mmask8 __U, __m128h __A, const int __imm) { return (__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) __A, __imm, __U); @@ -725,7 +725,7 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_fpclass_ph_mask (__mmask16 __U, __m256h __A, const int __imm) { return (__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) __A, - __imm, __U); + __imm, __U); } extern __inline __mmask16 @@ -733,26 +733,26 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_fpclass_ph_mask (__m256h __A, const int __imm) { return (__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) __A, - __imm, - (__mmask16) -1); + __imm, + (__mmask16) -1); } #else #define _mm_fpclass_ph_mask(X, C) \ - ((__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) (__m128h) (X), \ + ((__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) (__m128h) (X), \ (int) (C),(__mmask8)-1)) #define _mm_mask_fpclass_ph_mask(u, X, C) \ - ((__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) (__m128h) (X), \ + ((__mmask8) __builtin_ia32_fpclassph128_mask ((__v8hf) (__m128h) (X), \ (int) (C),(__mmask8)(u))) #define _mm256_fpclass_ph_mask(X, C) \ - ((__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) (__m256h) (X), \ - (int) (C),(__mmask16)-1)) + ((__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) (__m256h) (X), \ + (int) (C),(__mmask16)-1)) #define _mm256_mask_fpclass_ph_mask(u, X, C) \ - ((__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) (__m256h) (X), \ - (int) (C),(__mmask16)(u))) + ((__mmask16) __builtin_ia32_fpclassph256_mask ((__v16hf) (__m256h) (X), \ + (int) (C),(__mmask16)(u))) #endif /* __OPTIMIZE__ */ /* Intrinsics vgetexpph, vgetexpsh. */ @@ -761,9 +761,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_getexp_ph (__m256h __A) { return (__m256h) __builtin_ia32_getexpph256_mask ((__v16hf) __A, - (__v16hf) - _mm256_setzero_ph (), - (__mmask16) -1); + (__v16hf) + _mm256_setzero_ph (), + (__mmask16) -1); } extern __inline __m256h @@ -771,8 +771,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_mask_getexp_ph (__m256h __W, __mmask16 __U, __m256h __A) { return (__m256h) __builtin_ia32_getexpph256_mask ((__v16hf) __A, - (__v16hf) __W, - (__mmask16) __U); + (__v16hf) __W, + (__mmask16) __U); } extern __inline __m256h @@ -780,9 +780,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm256_maskz_getexp_ph (__mmask16 __U, __m256h __A) { return (__m256h) __builtin_ia32_getexpph256_mask ((__v16hf) __A, - (__v16hf) - _mm256_setzero_ph (), - (__mmask16) __U); + (__v16hf) + _mm256_setzero_ph (), + (__mmask16) __U); } extern __inline __m128h @@ -790,9 +790,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_getexp_ph (__m128h __A) { return (__m128h) __builtin_ia32_getexpph128_mask ((__v8hf) __A, - (__v8hf) - _mm_setzero_ph (), - (__mmask8) -1); + (__v8hf) + _mm_setzero_ph (), + (__mmask8) -1); } extern __inline __m128h @@ -800,8 +800,8 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_mask_getexp_ph (__m128h __W, __mmask8 __U, __m128h __A) { return (__m128h) __builtin_ia32_getexpph128_mask ((__v8hf) __A, - (__v8hf) __W, - (__mmask8) __U); + (__v8hf) __W, + (__mmask8) __U); } extern __inline __m128h @@ -809,9 +809,9 @@ __attribute__ ((__gnu_inline__, __always_inline__, __artificial__)) _mm_maskz_getexp_ph (__mmask8 __U, __m128h __A) { return (__m128h) __builtin_ia32_getexpph128_mask ((__v8hf) __A, - (__v8hf) - _mm_setzero_ph (), - (__mmask8) __U); + (__v8hf) + _mm_setzero_ph (), + (__mmask8) __U); } @@ -892,41 +892,41 @@ _mm_maskz_getmant_ph (__mmask8 __U, __m128h __A, } #else -#define _mm256_getmant_ph(X, B, C) \ - ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ - (int)(((C)<<2) | (B)), \ - (__v16hf)(__m256h)_mm256_setzero_ph (),\ - (__mmask16)-1)) - -#define _mm256_mask_getmant_ph(W, U, X, B, C) \ - ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ - (int)(((C)<<2) | (B)), \ - (__v16hf)(__m256h)(W), \ - (__mmask16)(U))) - -#define _mm256_maskz_getmant_ph(U, X, B, C) \ - ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ - (int)(((C)<<2) | (B)), \ - (__v16hf)(__m256h)_mm256_setzero_ph (),\ - (__mmask16)(U))) - -#define _mm_getmant_ph(X, B, C) \ - ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ - (int)(((C)<<2) | (B)), \ - (__v8hf)(__m128h)_mm_setzero_ph (), \ - (__mmask8)-1)) - -#define _mm_mask_getmant_ph(W, U, X, B, C) \ - ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ - (int)(((C)<<2) | (B)), \ - (__v8hf)(__m128h)(W), \ - (__mmask8)(U))) - -#define _mm_maskz_getmant_ph(U, X, B, C) \ - ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ - (int)(((C)<<2) | (B)), \ - (__v8hf)(__m128h)_mm_setzero_ph (), \ - (__mmask8)(U))) +#define _mm256_getmant_ph(X, B, C) \ + ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ + (int)(((C)<<2) | (B)), \ + (__v16hf)(__m256h)_mm256_setzero_ph (), \ + (__mmask16)-1)) + +#define _mm256_mask_getmant_ph(W, U, X, B, C) \ + ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ + (int)(((C)<<2) | (B)), \ + (__v16hf)(__m256h)(W), \ + (__mmask16)(U))) + +#define _mm256_maskz_getmant_ph(U, X, B, C) \ + ((__m256h) __builtin_ia32_getmantph256_mask ((__v16hf)(__m256h) (X), \ + (int)(((C)<<2) | (B)), \ + (__v16hf)(__m256h)_mm256_setzero_ph (), \ + (__mmask16)(U))) + +#define _mm_getmant_ph(X, B, C) \ + ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ + (int)(((C)<<2) | (B)), \ + (__v8hf)(__m128h)_mm_setzero_ph (), \ + (__mmask8)-1)) + +#define _mm_mask_getmant_ph(W, U, X, B, C) \ + ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ + (int)(((C)<<2) | (B)), \ + (__v8hf)(__m128h)(W), \ + (__mmask8)(U))) + +#define _mm_maskz_getmant_ph(U, X, B, C) \ + ((__m128h) __builtin_ia32_getmantph128_mask ((__v8hf)(__m128h) (X), \ + (int)(((C)<<2) | (B)), \ + (__v8hf)(__m128h)_mm_setzero_ph (), \ + (__mmask8)(U))) #endif /* __OPTIMIZE__ */ diff --git a/gcc/config/i386/i386-builtin.def b/gcc/config/i386/i386-builtin.def index bfa1d56a7b8..10f6fd87cbb 100644 --- a/gcc/config/i386/i386-builtin.def +++ b/gcc/config/i386/i386-builtin.def @@ -2775,49 +2775,49 @@ BDESC (0, OPTION_MASK_ISA2_AVX512BF16, CODE_FOR_avx512f_dpbf16ps_v4sf_mask, "__b BDESC (0, OPTION_MASK_ISA2_AVX512BF16, CODE_FOR_avx512f_dpbf16ps_v4sf_maskz, "__builtin_ia32_dpbf16ps_v4sf_maskz", IX86_BUILTIN_DPHI16PS_V4SF_MASKZ, UNKNOWN, (int) V4SF_FTYPE_V4SF_V8HI_V8HI_UQI) /* AVX512FP16. */ -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv8hf3_mask, "__builtin_ia32_vaddph_v8hf_mask", IX86_BUILTIN_VADDPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv16hf3_mask, "__builtin_ia32_vaddph_v16hf_mask", IX86_BUILTIN_VADDPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv32hf3_mask, "__builtin_ia32_vaddph_v32hf_mask", IX86_BUILTIN_VADDPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv8hf3_mask, "__builtin_ia32_vsubph_v8hf_mask", IX86_BUILTIN_VSUBPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv16hf3_mask, "__builtin_ia32_vsubph_v16hf_mask", IX86_BUILTIN_VSUBPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv32hf3_mask, "__builtin_ia32_vsubph_v32hf_mask", IX86_BUILTIN_VSUBPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv8hf3_mask, "__builtin_ia32_vmulph_v8hf_mask", IX86_BUILTIN_VMULPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv16hf3_mask, "__builtin_ia32_vmulph_v16hf_mask", IX86_BUILTIN_VMULPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv32hf3_mask, "__builtin_ia32_vmulph_v32hf_mask", IX86_BUILTIN_VMULPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv8hf3_mask, "__builtin_ia32_vdivph_v8hf_mask", IX86_BUILTIN_VDIVPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv16hf3_mask, "__builtin_ia32_vdivph_v16hf_mask", IX86_BUILTIN_VDIVPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv32hf3_mask, "__builtin_ia32_vdivph_v32hf_mask", IX86_BUILTIN_VDIVPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmaddv8hf3_mask, "__builtin_ia32_vaddsh_v8hf_mask", IX86_BUILTIN_VADDSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsubv8hf3_mask, "__builtin_ia32_vsubsh_v8hf_mask", IX86_BUILTIN_VSUBSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmmulv8hf3_mask, "__builtin_ia32_vmulsh_v8hf_mask", IX86_BUILTIN_VMULSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmdivv8hf3_mask, "__builtin_ia32_vdivsh_v8hf_mask", IX86_BUILTIN_VDIVSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv8hf3_mask, "__builtin_ia32_vmaxph_v8hf_mask", IX86_BUILTIN_VMAXPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv16hf3_mask, "__builtin_ia32_vmaxph_v16hf_mask", IX86_BUILTIN_VMAXPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv32hf3_mask, "__builtin_ia32_vmaxph_v32hf_mask", IX86_BUILTIN_VMAXPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv8hf3_mask, "__builtin_ia32_vminph_v8hf_mask", IX86_BUILTIN_VMINPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv16hf3_mask, "__builtin_ia32_vminph_v16hf_mask", IX86_BUILTIN_VMINPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv32hf3_mask, "__builtin_ia32_vminph_v32hf_mask", IX86_BUILTIN_VMINPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask, "__builtin_ia32_vmaxsh_v8hf_mask", IX86_BUILTIN_VMAXSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask, "__builtin_ia32_vminsh_v8hf_mask", IX86_BUILTIN_VMINSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_cmpv8hf3_mask, "__builtin_ia32_vcmpph_v8hf_mask", IX86_BUILTIN_VCMPPH_V8HF_MASK, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_cmpv16hf3_mask, "__builtin_ia32_vcmpph_v16hf_mask", IX86_BUILTIN_VCMPPH_V16HF_MASK, UNKNOWN, (int) UHI_FTYPE_V16HF_V16HF_INT_UHI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask, "__builtin_ia32_vcmpph_v32hf_mask", IX86_BUILTIN_VCMPPH_V32HF_MASK, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv8hf2_mask, "__builtin_ia32_vsqrtph_v8hf_mask", IX86_BUILTIN_VSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv16hf2_mask, "__builtin_ia32_vsqrtph_v16hf_mask", IX86_BUILTIN_VSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv8hf2_mask, "__builtin_ia32_vrsqrtph_v8hf_mask", IX86_BUILTIN_VRSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv16hf2_mask, "__builtin_ia32_vrsqrtph_v16hf_mask", IX86_BUILTIN_VRSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv32hf2_mask, "__builtin_ia32_vrsqrtph_v32hf_mask", IX86_BUILTIN_VRSQRTPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrsqrtv8hf2_mask, "__builtin_ia32_vrsqrtsh_v8hf_mask", IX86_BUILTIN_VRSQRTSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv8hf2_mask, "__builtin_ia32_vrcpph_v8hf_mask", IX86_BUILTIN_VRCPPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv16hf2_mask, "__builtin_ia32_vrcpph_v16hf_mask", IX86_BUILTIN_VRCPPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv32hf2_mask, "__builtin_ia32_vrcpph_v32hf_mask", IX86_BUILTIN_VRCPPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrcpv8hf2_mask, "__builtin_ia32_vrcpsh_v8hf_mask", IX86_BUILTIN_VRCPSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_scalefv8hf_mask, "__builtin_ia32_vscalefph_v8hf_mask", IX86_BUILTIN_VSCALEFPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_scalefv16hf_mask, "__builtin_ia32_vscalefph_v16hf_mask", IX86_BUILTIN_VSCALEFPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv8hf_mask, "__builtin_ia32_vreduceph_v8hf_mask", IX86_BUILTIN_VREDUCEPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_INT_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv16hf_mask, "__builtin_ia32_vreduceph_v16hf_mask", IX86_BUILTIN_VREDUCEPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_INT_V16HF_UHI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rndscalev8hf_mask, "__builtin_ia32_vrndscaleph_v8hf_mask", IX86_BUILTIN_VRNDSCALEPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_INT_V8HF_UQI) -BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_rndscalev16hf_mask, "__builtin_ia32_vrndscaleph_v16hf_mask", IX86_BUILTIN_VRNDSCALEPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_INT_V16HF_UHI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv8hf3_mask, "__builtin_ia32_addph128_mask", IX86_BUILTIN_VADDPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv16hf3_mask, "__builtin_ia32_addph256_mask", IX86_BUILTIN_VADDPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv32hf3_mask, "__builtin_ia32_addph512_mask", IX86_BUILTIN_VADDPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv8hf3_mask, "__builtin_ia32_subph128_mask", IX86_BUILTIN_VSUBPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv16hf3_mask, "__builtin_ia32_subph256_mask", IX86_BUILTIN_VSUBPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv32hf3_mask, "__builtin_ia32_subph512_mask", IX86_BUILTIN_VSUBPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv8hf3_mask, "__builtin_ia32_mulph128_mask", IX86_BUILTIN_VMULPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv16hf3_mask, "__builtin_ia32_mulph256_mask", IX86_BUILTIN_VMULPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv32hf3_mask, "__builtin_ia32_mulph512_mask", IX86_BUILTIN_VMULPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv8hf3_mask, "__builtin_ia32_divph128_mask", IX86_BUILTIN_VDIVPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv16hf3_mask, "__builtin_ia32_divph256_mask", IX86_BUILTIN_VDIVPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv32hf3_mask, "__builtin_ia32_divph512_mask", IX86_BUILTIN_VDIVPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmaddv8hf3_mask, "__builtin_ia32_addsh_mask", IX86_BUILTIN_VADDSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsubv8hf3_mask, "__builtin_ia32_subsh_mask", IX86_BUILTIN_VSUBSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmmulv8hf3_mask, "__builtin_ia32_mulsh_mask", IX86_BUILTIN_VMULSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmdivv8hf3_mask, "__builtin_ia32_divsh_mask", IX86_BUILTIN_VDIVSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv8hf3_mask, "__builtin_ia32_maxph128_mask", IX86_BUILTIN_VMAXPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv16hf3_mask, "__builtin_ia32_maxph256_mask", IX86_BUILTIN_VMAXPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv32hf3_mask, "__builtin_ia32_maxph512_mask", IX86_BUILTIN_VMAXPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv8hf3_mask, "__builtin_ia32_minph128_mask", IX86_BUILTIN_VMINPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv16hf3_mask, "__builtin_ia32_minph256_mask", IX86_BUILTIN_VMINPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv32hf3_mask, "__builtin_ia32_minph512_mask", IX86_BUILTIN_VMINPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask, "__builtin_ia32_maxsh_mask", IX86_BUILTIN_VMAXSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask, "__builtin_ia32_minsh_mask", IX86_BUILTIN_VMINSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_cmpv8hf3_mask, "__builtin_ia32_cmpph128_mask", IX86_BUILTIN_VCMPPH_V8HF_MASK, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_cmpv16hf3_mask, "__builtin_ia32_cmpph256_mask", IX86_BUILTIN_VCMPPH_V16HF_MASK, UNKNOWN, (int) UHI_FTYPE_V16HF_V16HF_INT_UHI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask, "__builtin_ia32_cmpph512_mask", IX86_BUILTIN_VCMPPH_V32HF_MASK, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv8hf2_mask, "__builtin_ia32_sqrtph128_mask", IX86_BUILTIN_VSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv16hf2_mask, "__builtin_ia32_sqrtph256_mask", IX86_BUILTIN_VSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv8hf2_mask, "__builtin_ia32_rsqrtph128_mask", IX86_BUILTIN_VRSQRTPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv16hf2_mask, "__builtin_ia32_rsqrtph256_mask", IX86_BUILTIN_VRSQRTPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rsqrtv32hf2_mask, "__builtin_ia32_rsqrtph512_mask", IX86_BUILTIN_VRSQRTPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrsqrtv8hf2_mask, "__builtin_ia32_rsqrtsh_mask", IX86_BUILTIN_VRSQRTSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv8hf2_mask, "__builtin_ia32_rcpph128_mask", IX86_BUILTIN_VRCPPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv16hf2_mask, "__builtin_ia32_rcpph256_mask", IX86_BUILTIN_VRCPPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_UHI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rcpv32hf2_mask, "__builtin_ia32_rcpph512_mask", IX86_BUILTIN_VRCPPH_V32HF_MASK, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmrcpv8hf2_mask, "__builtin_ia32_rcpsh_mask", IX86_BUILTIN_VRCPSH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_scalefv8hf_mask, "__builtin_ia32_scalefph128_mask", IX86_BUILTIN_VSCALEFPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_scalefv16hf_mask, "__builtin_ia32_scalefph256_mask", IX86_BUILTIN_VSCALEFPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_V16HF_V16HF_UHI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv8hf_mask, "__builtin_ia32_reduceph128_mask", IX86_BUILTIN_VREDUCEPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_INT_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv16hf_mask, "__builtin_ia32_reduceph256_mask", IX86_BUILTIN_VREDUCEPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_INT_V16HF_UHI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_rndscalev8hf_mask, "__builtin_ia32_rndscaleph128_mask", IX86_BUILTIN_VRNDSCALEPH_V8HF_MASK, UNKNOWN, (int) V8HF_FTYPE_V8HF_INT_V8HF_UQI) +BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512vl_rndscalev16hf_mask, "__builtin_ia32_rndscaleph256_mask", IX86_BUILTIN_VRNDSCALEPH_V16HF_MASK, UNKNOWN, (int) V16HF_FTYPE_V16HF_INT_V16HF_UHI) BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512dq_fpclassv16hf_mask, "__builtin_ia32_fpclassph256_mask", IX86_BUILTIN_FPCLASSPH256, UNKNOWN, (int) HI_FTYPE_V16HF_INT_UHI) BDESC (OPTION_MASK_ISA_AVX512VL, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512dq_fpclassv8hf_mask, "__builtin_ia32_fpclassph128_mask", IX86_BUILTIN_FPCLASSPH128, UNKNOWN, (int) QI_FTYPE_V8HF_INT_UQI) BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512dq_fpclassv32hf_mask, "__builtin_ia32_fpclassph512_mask", IX86_BUILTIN_FPCLASSPH512, UNKNOWN, (int) SI_FTYPE_V32HF_INT_USI) @@ -3027,28 +3027,28 @@ BDESC (OPTION_MASK_ISA_AVX512DQ, 0, CODE_FOR_avx512dq_rangepv16sf_mask_round, "_ BDESC (OPTION_MASK_ISA_AVX512DQ, 0, CODE_FOR_avx512dq_rangepv8df_mask_round, "__builtin_ia32_rangepd512_mask", IX86_BUILTIN_RANGEPD512, UNKNOWN, (int) V8DF_FTYPE_V8DF_V8DF_INT_V8DF_QI_INT) /* AVX512FP16. */ -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv32hf3_mask_round, "__builtin_ia32_vaddph_v32hf_mask_round", IX86_BUILTIN_VADDPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv32hf3_mask_round, "__builtin_ia32_vsubph_v32hf_mask_round", IX86_BUILTIN_VSUBPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv32hf3_mask_round, "__builtin_ia32_vmulph_v32hf_mask_round", IX86_BUILTIN_VMULPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv32hf3_mask_round, "__builtin_ia32_vdivph_v32hf_mask_round", IX86_BUILTIN_VDIVPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmaddv8hf3_mask_round, "__builtin_ia32_vaddsh_v8hf_mask_round", IX86_BUILTIN_VADDSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsubv8hf3_mask_round, "__builtin_ia32_vsubsh_v8hf_mask_round", IX86_BUILTIN_VSUBSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmmulv8hf3_mask_round, "__builtin_ia32_vmulsh_v8hf_mask_round", IX86_BUILTIN_VMULSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmdivv8hf3_mask_round, "__builtin_ia32_vdivsh_v8hf_mask_round", IX86_BUILTIN_VDIVSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv32hf3_mask_round, "__builtin_ia32_vmaxph_v32hf_mask_round", IX86_BUILTIN_VMAXPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv32hf3_mask_round, "__builtin_ia32_vminph_v32hf_mask_round", IX86_BUILTIN_VMINPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask_round, "__builtin_ia32_vmaxsh_v8hf_mask_round", IX86_BUILTIN_VMAXSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask_round, "__builtin_ia32_vminsh_v8hf_mask_round", IX86_BUILTIN_VMINSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask_round, "__builtin_ia32_vcmpph_v32hf_mask_round", IX86_BUILTIN_VCMPPH_V32HF_MASK_ROUND, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_vmcmpv8hf3_mask_round, "__builtin_ia32_vcmpsh_v8hf_mask_round", IX86_BUILTIN_VCMPSH_V8HF_MASK_ROUND, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv32hf2_mask_round, "__builtin_ia32_vsqrtph_v32hf_mask_round", IX86_BUILTIN_VSQRTPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsqrtv8hf2_mask_round, "__builtin_ia32_vsqrtsh_v8hf_mask_round", IX86_BUILTIN_VSQRTSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_scalefv32hf_mask_round, "__builtin_ia32_vscalefph_v32hf_mask_round", IX86_BUILTIN_VSCALEFPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_vmscalefv8hf_mask_round, "__builtin_ia32_vscalefsh_v8hf_mask_round", IX86_BUILTIN_VSCALEFSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv32hf_mask_round, "__builtin_ia32_vreduceph_v32hf_mask_round", IX86_BUILTIN_VREDUCEPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducesv8hf_mask_round, "__builtin_ia32_vreducesh_v8hf_mask_round", IX86_BUILTIN_VREDUCESH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_INT_V8HF_UQI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_rndscalev32hf_mask_round, "__builtin_ia32_vrndscaleph_v32hf_mask_round", IX86_BUILTIN_VRNDSCALEPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) -BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_rndscalev8hf_mask_round, "__builtin_ia32_vrndscalesh_v8hf_mask_round", IX86_BUILTIN_VRNDSCALESH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_INT_V8HF_UQI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_addv32hf3_mask_round, "__builtin_ia32_addph512_mask_round", IX86_BUILTIN_VADDPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_subv32hf3_mask_round, "__builtin_ia32_subph512_mask_round", IX86_BUILTIN_VSUBPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_mulv32hf3_mask_round, "__builtin_ia32_mulph512_mask_round", IX86_BUILTIN_VMULPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_divv32hf3_mask_round, "__builtin_ia32_divph512_mask_round", IX86_BUILTIN_VDIVPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmaddv8hf3_mask_round, "__builtin_ia32_addsh_mask_round", IX86_BUILTIN_VADDSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsubv8hf3_mask_round, "__builtin_ia32_subsh_mask_round", IX86_BUILTIN_VSUBSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmmulv8hf3_mask_round, "__builtin_ia32_mulsh_mask_round", IX86_BUILTIN_VMULSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmdivv8hf3_mask_round, "__builtin_ia32_divsh_mask_round", IX86_BUILTIN_VDIVSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_smaxv32hf3_mask_round, "__builtin_ia32_maxph512_mask_round", IX86_BUILTIN_VMAXPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_sminv32hf3_mask_round, "__builtin_ia32_minph512_mask_round", IX86_BUILTIN_VMINPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsmaxv8hf3_mask_round, "__builtin_ia32_maxsh_mask_round", IX86_BUILTIN_VMAXSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsminv8hf3_mask_round, "__builtin_ia32_minsh_mask_round", IX86_BUILTIN_VMINSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_cmpv32hf3_mask_round, "__builtin_ia32_cmpph512_mask_round", IX86_BUILTIN_VCMPPH_V32HF_MASK_ROUND, UNKNOWN, (int) USI_FTYPE_V32HF_V32HF_INT_USI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_vmcmpv8hf3_mask_round, "__builtin_ia32_cmpsh_mask_round", IX86_BUILTIN_VCMPSH_V8HF_MASK_ROUND, UNKNOWN, (int) UQI_FTYPE_V8HF_V8HF_INT_UQI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_sqrtv32hf2_mask_round, "__builtin_ia32_sqrtph512_mask_round", IX86_BUILTIN_VSQRTPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512fp16_vmsqrtv8hf2_mask_round, "__builtin_ia32_sqrtsh_mask_round", IX86_BUILTIN_VSQRTSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_scalefv32hf_mask_round, "__builtin_ia32_scalefph512_mask_round", IX86_BUILTIN_VSCALEFPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_V32HF_USI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_vmscalefv8hf_mask_round, "__builtin_ia32_scalefsh_mask_round", IX86_BUILTIN_VSCALEFSH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducepv32hf_mask_round, "__builtin_ia32_reduceph512_mask_round", IX86_BUILTIN_VREDUCEPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_reducesv8hf_mask_round, "__builtin_ia32_reducesh_mask_round", IX86_BUILTIN_VREDUCESH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_INT_V8HF_UQI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_rndscalev32hf_mask_round, "__builtin_ia32_rndscaleph512_mask_round", IX86_BUILTIN_VRNDSCALEPH_V32HF_MASK_ROUND, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) +BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_rndscalev8hf_mask_round, "__builtin_ia32_rndscalesh_mask_round", IX86_BUILTIN_VRNDSCALESH_V8HF_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_INT_V8HF_UQI_INT) BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_getexpv32hf_mask_round, "__builtin_ia32_getexpph512_mask", IX86_BUILTIN_GETEXPPH512, UNKNOWN, (int) V32HF_FTYPE_V32HF_V32HF_USI_INT) BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512f_sgetexpv8hf_mask_round, "__builtin_ia32_getexpsh_mask_round", IX86_BUILTIN_GETEXPSH_MASK_ROUND, UNKNOWN, (int) V8HF_FTYPE_V8HF_V8HF_V8HF_UQI_INT) BDESC (0, OPTION_MASK_ISA2_AVX512FP16, CODE_FOR_avx512bw_getmantv32hf_mask_round, "__builtin_ia32_getmantph512_mask", IX86_BUILTIN_GETMANTPH512, UNKNOWN, (int) V32HF_FTYPE_V32HF_INT_V32HF_USI_INT) diff --git a/gcc/testsuite/gcc.target/i386/avx-1.c b/gcc/testsuite/gcc.target/i386/avx-1.c index b3cffa0644f..3a96e586418 100644 --- a/gcc/testsuite/gcc.target/i386/avx-1.c +++ b/gcc/testsuite/gcc.target/i386/avx-1.c @@ -686,33 +686,33 @@ #define __builtin_ia32_vpshld_v2di_mask(A, B, C, D, E) __builtin_ia32_vpshld_v2di_mask(A, B, 1, D, E) /* avx512fp16intrin.h */ -#define __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vcmpph_v32hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v32hf_mask(A, B, 1, D) -#define __builtin_ia32_vcmpph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpph_v32hf_mask_round(A, B, 1, D, 8) -#define __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, 1, D, 8) -#define __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, D) __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, 8) -#define __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, E) __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, 8) -#define __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vreduceph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vreduceph_v32hf_mask_round(A, 123, C, D, 8) -#define __builtin_ia32_vreduceph_v8hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v8hf_mask(A, 123, C, D) -#define __builtin_ia32_vreduceph_v16hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v16hf_mask(A, 123, C, D) -#define __builtin_ia32_vreducesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vreducesh_v8hf_mask_round(A, B, 123, D, E, 8) -#define __builtin_ia32_vrndscaleph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vrndscaleph_v32hf_mask_round(A, 123, C, D, 8) -#define __builtin_ia32_vrndscaleph_v8hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v8hf_mask(A, 123, C, D) -#define __builtin_ia32_vrndscaleph_v16hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v16hf_mask(A, 123, C, D) -#define __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, 123, D, E, 8) +#define __builtin_ia32_addph512_mask_round(A, B, C, D, E) __builtin_ia32_addph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_subph512_mask_round(A, B, C, D, E) __builtin_ia32_subph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_mulph512_mask_round(A, B, C, D, E) __builtin_ia32_mulph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_divph512_mask_round(A, B, C, D, E) __builtin_ia32_divph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_addsh_mask_round(A, B, C, D, E) __builtin_ia32_addsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_subsh_mask_round(A, B, C, D, E) __builtin_ia32_subsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_mulsh_mask_round(A, B, C, D, E) __builtin_ia32_mulsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_divsh_mask_round(A, B, C, D, E) __builtin_ia32_divsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_maxph512_mask_round(A, B, C, D, E) __builtin_ia32_maxph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_minph512_mask_round(A, B, C, D, E) __builtin_ia32_minph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_maxsh_mask_round(A, B, C, D, E) __builtin_ia32_maxsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_minsh_mask_round(A, B, C, D, E) __builtin_ia32_minsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_cmpph512_mask(A, B, C, D) __builtin_ia32_cmpph512_mask(A, B, 1, D) +#define __builtin_ia32_cmpph512_mask_round(A, B, C, D, E) __builtin_ia32_cmpph512_mask_round(A, B, 1, D, 8) +#define __builtin_ia32_cmpsh_mask_round(A, B, C, D, E) __builtin_ia32_cmpsh_mask_round(A, B, 1, D, 8) +#define __builtin_ia32_sqrtph512_mask_round(C, A, B, D) __builtin_ia32_sqrtph512_mask_round(C, A, B, 8) +#define __builtin_ia32_sqrtsh_mask_round(D, C, A, B, E) __builtin_ia32_sqrtsh_mask_round(D, C, A, B, 8) +#define __builtin_ia32_scalefph512_mask_round(A, B, C, D, E) __builtin_ia32_scalefph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_scalefsh_mask_round(A, B, C, D, E) __builtin_ia32_scalefsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_reduceph512_mask_round(A, B, C, D, E) __builtin_ia32_reduceph512_mask_round(A, 123, C, D, 8) +#define __builtin_ia32_reduceph128_mask(A, B, C, D) __builtin_ia32_reduceph128_mask(A, 123, C, D) +#define __builtin_ia32_reduceph256_mask(A, B, C, D) __builtin_ia32_reduceph256_mask(A, 123, C, D) +#define __builtin_ia32_reducesh_mask_round(A, B, C, D, E, F) __builtin_ia32_reducesh_mask_round(A, B, 123, D, E, 8) +#define __builtin_ia32_rndscaleph512_mask_round(A, B, C, D, E) __builtin_ia32_rndscaleph512_mask_round(A, 123, C, D, 8) +#define __builtin_ia32_rndscaleph128_mask(A, B, C, D) __builtin_ia32_rndscaleph128_mask(A, 123, C, D) +#define __builtin_ia32_rndscaleph256_mask(A, B, C, D) __builtin_ia32_rndscaleph256_mask(A, 123, C, D) +#define __builtin_ia32_rndscalesh_mask_round(A, B, C, D, E, F) __builtin_ia32_rndscalesh_mask_round(A, B, 123, D, E, 8) #define __builtin_ia32_fpclassph512_mask(A, D, C) __builtin_ia32_fpclassph512_mask(A, 1, C) #define __builtin_ia32_fpclasssh_mask(A, D, U) __builtin_ia32_fpclasssh_mask(A, 1, U) #define __builtin_ia32_getexpph512_mask(A, B, C, D) __builtin_ia32_getexpph512_mask(A, B, C, 8) @@ -721,8 +721,8 @@ #define __builtin_ia32_getmantsh_mask_round(A, B, C, W, U, D) __builtin_ia32_getmantsh_mask_round(A, B, 1, W, U, 4) /* avx512fp16vlintrin.h */ -#define __builtin_ia32_vcmpph_v8hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v8hf_mask(A, B, 1, D) -#define __builtin_ia32_vcmpph_v16hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v16hf_mask(A, B, 1, D) +#define __builtin_ia32_cmpph128_mask(A, B, C, D) __builtin_ia32_cmpph128_mask(A, B, 1, D) +#define __builtin_ia32_cmpph256_mask(A, B, C, D) __builtin_ia32_cmpph256_mask(A, B, 1, D) #define __builtin_ia32_fpclassph256_mask(A, D, C) __builtin_ia32_fpclassph256_mask(A, 1, C) #define __builtin_ia32_fpclassph128_mask(A, D, C) __builtin_ia32_fpclassph128_mask(A, 1, C) #define __builtin_ia32_getmantph256_mask(A, E, C, D) __builtin_ia32_getmantph256_mask(A, 1, C, D) diff --git a/gcc/testsuite/gcc.target/i386/sse-13.c b/gcc/testsuite/gcc.target/i386/sse-13.c index 67ef567e437..aafcd414530 100644 --- a/gcc/testsuite/gcc.target/i386/sse-13.c +++ b/gcc/testsuite/gcc.target/i386/sse-13.c @@ -703,33 +703,33 @@ #define __builtin_ia32_vpshld_v2di_mask(A, B, C, D, E) __builtin_ia32_vpshld_v2di_mask(A, B, 1, D, E) /* avx512fp16intrin.h */ -#define __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vcmpph_v32hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v32hf_mask(A, B, 1, D) -#define __builtin_ia32_vcmpph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpph_v32hf_mask_round(A, B, 1, D, 8) -#define __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, 1, D, 8) -#define __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, D) __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, 8) -#define __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, E) __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, 8) -#define __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vreduceph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vreduceph_v32hf_mask_round(A, 123, C, D, 8) -#define __builtin_ia32_vreduceph_v8hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v8hf_mask(A, 123, C, D) -#define __builtin_ia32_vreduceph_v16hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v16hf_mask(A, 123, C, D) -#define __builtin_ia32_vreducesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vreducesh_v8hf_mask_round(A, B, 123, D, E, 8) -#define __builtin_ia32_vrndscaleph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vrndscaleph_v32hf_mask_round(A, 123, C, D, 8) -#define __builtin_ia32_vrndscaleph_v8hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v8hf_mask(A, 123, C, D) -#define __builtin_ia32_vrndscaleph_v16hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v16hf_mask(A, 123, C, D) -#define __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, 123, D, E, 8) +#define __builtin_ia32_addph512_mask_round(A, B, C, D, E) __builtin_ia32_addph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_subph512_mask_round(A, B, C, D, E) __builtin_ia32_subph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_mulph512_mask_round(A, B, C, D, E) __builtin_ia32_mulph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_divph512_mask_round(A, B, C, D, E) __builtin_ia32_divph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_addsh_mask_round(A, B, C, D, E) __builtin_ia32_addsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_subsh_mask_round(A, B, C, D, E) __builtin_ia32_subsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_mulsh_mask_round(A, B, C, D, E) __builtin_ia32_mulsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_divsh_mask_round(A, B, C, D, E) __builtin_ia32_divsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_maxph512_mask_round(A, B, C, D, E) __builtin_ia32_maxph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_minph512_mask_round(A, B, C, D, E) __builtin_ia32_minph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_maxsh_mask_round(A, B, C, D, E) __builtin_ia32_maxsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_minsh_mask_round(A, B, C, D, E) __builtin_ia32_minsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_cmpph512_mask(A, B, C, D) __builtin_ia32_cmpph512_mask(A, B, 1, D) +#define __builtin_ia32_cmpph512_mask_round(A, B, C, D, E) __builtin_ia32_cmpph512_mask_round(A, B, 1, D, 8) +#define __builtin_ia32_cmpsh_mask_round(A, B, C, D, E) __builtin_ia32_cmpsh_mask_round(A, B, 1, D, 8) +#define __builtin_ia32_sqrtph512_mask_round(C, A, B, D) __builtin_ia32_sqrtph512_mask_round(C, A, B, 8) +#define __builtin_ia32_sqrtsh_mask_round(D, C, A, B, E) __builtin_ia32_sqrtsh_mask_round(D, C, A, B, 8) +#define __builtin_ia32_scalefph512_mask_round(A, B, C, D, E) __builtin_ia32_scalefph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_scalefsh_mask_round(A, B, C, D, E) __builtin_ia32_scalefsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_reduceph512_mask_round(A, B, C, D, E) __builtin_ia32_reduceph512_mask_round(A, 123, C, D, 8) +#define __builtin_ia32_reduceph128_mask(A, B, C, D) __builtin_ia32_reduceph128_mask(A, 123, C, D) +#define __builtin_ia32_reduceph256_mask(A, B, C, D) __builtin_ia32_reduceph256_mask(A, 123, C, D) +#define __builtin_ia32_reducesh_mask_round(A, B, C, D, E, F) __builtin_ia32_reducesh_mask_round(A, B, 123, D, E, 8) +#define __builtin_ia32_rndscaleph512_mask_round(A, B, C, D, E) __builtin_ia32_rndscaleph512_mask_round(A, 123, C, D, 8) +#define __builtin_ia32_rndscaleph128_mask(A, B, C, D) __builtin_ia32_rndscaleph128_mask(A, 123, C, D) +#define __builtin_ia32_rndscaleph256_mask(A, B, C, D) __builtin_ia32_rndscaleph256_mask(A, 123, C, D) +#define __builtin_ia32_rndscalesh_mask_round(A, B, C, D, E, F) __builtin_ia32_rndscalesh_mask_round(A, B, 123, D, E, 8) #define __builtin_ia32_fpclassph512_mask(A, D, C) __builtin_ia32_fpclassph512_mask(A, 1, C) #define __builtin_ia32_fpclasssh_mask(A, D, U) __builtin_ia32_fpclasssh_mask(A, 1, U) #define __builtin_ia32_getexpph512_mask(A, B, C, D) __builtin_ia32_getexpph512_mask(A, B, C, 8) @@ -738,8 +738,8 @@ #define __builtin_ia32_getmantsh_mask_round(A, B, C, W, U, D) __builtin_ia32_getmantsh_mask_round(A, B, 1, W, U, 4) /* avx512fp16vlintrin.h */ -#define __builtin_ia32_vcmpph_v8hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v8hf_mask(A, B, 1, D) -#define __builtin_ia32_vcmpph_v16hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v16hf_mask(A, B, 1, D) +#define __builtin_ia32_cmpph128_mask(A, B, C, D) __builtin_ia32_cmpph128_mask(A, B, 1, D) +#define __builtin_ia32_cmpph256_mask(A, B, C, D) __builtin_ia32_cmpph256_mask(A, B, 1, D) #define __builtin_ia32_fpclassph256_mask(A, D, C) __builtin_ia32_fpclassph256_mask(A, 1, C) #define __builtin_ia32_fpclassph128_mask(A, D, C) __builtin_ia32_fpclassph128_mask(A, 1, C) #define __builtin_ia32_getmantph256_mask(A, E, C, D) __builtin_ia32_getmantph256_mask(A, 1, C, D) diff --git a/gcc/testsuite/gcc.target/i386/sse-23.c b/gcc/testsuite/gcc.target/i386/sse-23.c index b3f07587acb..8b600282c67 100644 --- a/gcc/testsuite/gcc.target/i386/sse-23.c +++ b/gcc/testsuite/gcc.target/i386/sse-23.c @@ -704,33 +704,33 @@ #define __builtin_ia32_vpshld_v2di_mask(A, B, C, D, E) __builtin_ia32_vpshld_v2di_mask(A, B, 1, D, E) /* avx512fp16intrin.h */ -#define __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vaddsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vsubsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmulsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vdivsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vminph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vmaxsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vminsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vcmpph_v32hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v32hf_mask(A, B, 1, D) -#define __builtin_ia32_vcmpph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpph_v32hf_mask_round(A, B, 1, D, 8) -#define __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vcmpsh_v8hf_mask_round(A, B, 1, D, 8) -#define __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, D) __builtin_ia32_vsqrtph_v32hf_mask_round(C, A, B, 8) -#define __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, E) __builtin_ia32_vsqrtsh_v8hf_mask_round(D, C, A, B, 8) -#define __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefph_v32hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, E) __builtin_ia32_vscalefsh_v8hf_mask_round(A, B, C, D, 8) -#define __builtin_ia32_vreduceph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vreduceph_v32hf_mask_round(A, 123, C, D, 8) -#define __builtin_ia32_vreduceph_v8hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v8hf_mask(A, 123, C, D) -#define __builtin_ia32_vreduceph_v16hf_mask(A, B, C, D) __builtin_ia32_vreduceph_v16hf_mask(A, 123, C, D) -#define __builtin_ia32_vreducesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vreducesh_v8hf_mask_round(A, B, 123, D, E, 8) -#define __builtin_ia32_vrndscaleph_v32hf_mask_round(A, B, C, D, E) __builtin_ia32_vrndscaleph_v32hf_mask_round(A, 123, C, D, 8) -#define __builtin_ia32_vrndscaleph_v8hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v8hf_mask(A, 123, C, D) -#define __builtin_ia32_vrndscaleph_v16hf_mask(A, B, C, D) __builtin_ia32_vrndscaleph_v16hf_mask(A, 123, C, D) -#define __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, C, D, E, F) __builtin_ia32_vrndscalesh_v8hf_mask_round(A, B, 123, D, E, 8) +#define __builtin_ia32_addph512_mask_round(A, B, C, D, E) __builtin_ia32_addph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_subph512_mask_round(A, B, C, D, E) __builtin_ia32_subph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_mulph512_mask_round(A, B, C, D, E) __builtin_ia32_mulph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_divph512_mask_round(A, B, C, D, E) __builtin_ia32_divph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_addsh_mask_round(A, B, C, D, E) __builtin_ia32_addsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_subsh_mask_round(A, B, C, D, E) __builtin_ia32_subsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_mulsh_mask_round(A, B, C, D, E) __builtin_ia32_mulsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_divsh_mask_round(A, B, C, D, E) __builtin_ia32_divsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_maxph512_mask_round(A, B, C, D, E) __builtin_ia32_maxph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_minph512_mask_round(A, B, C, D, E) __builtin_ia32_minph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_maxsh_mask_round(A, B, C, D, E) __builtin_ia32_maxsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_minsh_mask_round(A, B, C, D, E) __builtin_ia32_minsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_cmpph512_mask(A, B, C, D) __builtin_ia32_cmpph512_mask(A, B, 1, D) +#define __builtin_ia32_cmpph512_mask_round(A, B, C, D, E) __builtin_ia32_cmpph512_mask_round(A, B, 1, D, 8) +#define __builtin_ia32_cmpsh_mask_round(A, B, C, D, E) __builtin_ia32_cmpsh_mask_round(A, B, 1, D, 8) +#define __builtin_ia32_sqrtph512_mask_round(C, A, B, D) __builtin_ia32_sqrtph512_mask_round(C, A, B, 8) +#define __builtin_ia32_sqrtsh_mask_round(D, C, A, B, E) __builtin_ia32_sqrtsh_mask_round(D, C, A, B, 8) +#define __builtin_ia32_scalefph512_mask_round(A, B, C, D, E) __builtin_ia32_scalefph512_mask_round(A, B, C, D, 8) +#define __builtin_ia32_scalefsh_mask_round(A, B, C, D, E) __builtin_ia32_scalefsh_mask_round(A, B, C, D, 8) +#define __builtin_ia32_reduceph512_mask_round(A, B, C, D, E) __builtin_ia32_reduceph512_mask_round(A, 123, C, D, 8) +#define __builtin_ia32_reduceph128_mask(A, B, C, D) __builtin_ia32_reduceph128_mask(A, 123, C, D) +#define __builtin_ia32_reduceph256_mask(A, B, C, D) __builtin_ia32_reduceph256_mask(A, 123, C, D) +#define __builtin_ia32_reducesh_mask_round(A, B, C, D, E, F) __builtin_ia32_reducesh_mask_round(A, B, 123, D, E, 8) +#define __builtin_ia32_rndscaleph512_mask_round(A, B, C, D, E) __builtin_ia32_rndscaleph512_mask_round(A, 123, C, D, 8) +#define __builtin_ia32_rndscaleph128_mask(A, B, C, D) __builtin_ia32_rndscaleph128_mask(A, 123, C, D) +#define __builtin_ia32_rndscaleph256_mask(A, B, C, D) __builtin_ia32_rndscaleph256_mask(A, 123, C, D) +#define __builtin_ia32_rndscalesh_mask_round(A, B, C, D, E, F) __builtin_ia32_rndscalesh_mask_round(A, B, 123, D, E, 8) #define __builtin_ia32_fpclassph512_mask(A, D, C) __builtin_ia32_fpclassph512_mask(A, 1, C) #define __builtin_ia32_fpclasssh_mask(A, D, U) __builtin_ia32_fpclasssh_mask(A, 1, U) #define __builtin_ia32_getexpph512_mask(A, B, C, D) __builtin_ia32_getexpph512_mask(A, B, C, 8) @@ -739,8 +739,8 @@ #define __builtin_ia32_getmantsh_mask_round(A, B, C, W, U, D) __builtin_ia32_getmantsh_mask_round(A, B, 1, W, U, 4) /* avx512fp16vlintrin.h */ -#define __builtin_ia32_vcmpph_v8hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v8hf_mask(A, B, 1, D) -#define __builtin_ia32_vcmpph_v16hf_mask(A, B, C, D) __builtin_ia32_vcmpph_v16hf_mask(A, B, 1, D) +#define __builtin_ia32_cmpph128_mask(A, B, C, D) __builtin_ia32_cmpph128_mask(A, B, 1, D) +#define __builtin_ia32_cmpph256_mask(A, B, C, D) __builtin_ia32_cmpph256_mask(A, B, 1, D) #define __builtin_ia32_fpclassph256_mask(A, D, C) __builtin_ia32_fpclassph256_mask(A, 1, C) #define __builtin_ia32_fpclassph128_mask(A, D, C) __builtin_ia32_fpclassph128_mask(A, 1, C) #define __builtin_ia32_getmantph256_mask(A, E, C, D) __builtin_ia32_getmantph256_mask(A, 1, C, D)