diff mbox series

[06/13] i386: Change mnemonics from V[GETMANT, REDUCENE, RNDSCALENE]PBF16 to V[GETMANT, REDUCE, RNDSCALE]BF16

Message ID 20250121083451.1836888-7-haochen.jiang@intel.com
State New
Headers show
Series Realign x86 GCC after Binutils change [PR118270] | expand

Commit Message

Jiang, Haochen Jan. 21, 2025, 8:34 a.m. UTC
gcc/ChangeLog:

	PR target/118270
	* config/i386/avx10_2-512bf16intrin.h: Change intrin and builtin
	name according to new mnemonics.
	* config/i386/avx10_2bf16intrin.h: Ditto.
	* config/i386/i386-builtin.def (BDESC): Ditto.
	* config/i386/sse.md
	(UNSPEC_VRNDSCALEBF16): Rename from UNSPEC_VRNDSCALENEPBF16.
	(UNSPEC_VREDUCEBF16): Rename from UNSPEC_VREDUCENEPBF16.
	(UNSPEC_VGETMANTBF16): Rename from UNSPEC_VGETMANTPBF16.
	(BF16IMMOP): Adjust iterator due to UNSPEC name change.
	(bf16immop): Ditto.
	(avx10_2_<bf16immop>pbf16_<mode><mask_name>): Rename to...
	(avx10_2_<bf16immop>bf16_<mode><mask_name>): ...this. Change
	instruction name output.

gcc/testsuite/ChangeLog:

	PR target/118270
	* gcc.target/i386/avx10_2-512-bf16-1.c: Adjust output and
	intrin call.
	* gcc.target/i386/avx10_2-512-vgetmantpbf16-2.c: Move to...
	* gcc.target/i386/avx10_2-512-vgetmantbf16-2.c: ...here.
	Adjust intrin call.
	* gcc.target/i386/avx10_2-512-vreducenepbf16-2.c: Move to...
	* gcc.target/i386/avx10_2-512-vreducebf16-2.c: ...here.
	Adjust intrin call.
	* gcc.target/i386/avx10_2-512-vrndscalenepbf16-2.c: Move to...
	* gcc.target/i386/avx10_2-512-vrndscalebf16-2.c: ...here.
	Adjust intrin call.
	* gcc.target/i386/avx10_2-bf16-1.c: Adjust output and intrin
	call.
	* gcc.target/i386/avx10_2-vgetmantpbf16-2.c: Move to...
	* gcc.target/i386/avx10_2-vgetmantbf16-2.c: ...here.
	Adjust intrin call.
	* gcc.target/i386/avx10_2-vreducenepbf16-2.c: Move to...
	* gcc.target/i386/avx10_2-vreducebf16-2.c: ...here.
	Adjust intrin call.
	* gcc.target/i386/avx10_2-vrndscalenepbf16-2.c: Move to...
	* gcc.target/i386/avx10_2-vrndscalebf16-2.c: ...here.
	Adjust intrin call.
	* gcc.target/i386/avx-1.c: Adjust builtin call.
	* gcc.target/i386/sse-13.c: Ditto.
	* gcc.target/i386/sse-23.c: Ditto.
	* gcc.target/i386/sse-14.c: Adjust intrin call.
	* gcc.target/i386/sse-22.c: Ditto.
---
 gcc/config/i386/avx10_2-512bf16intrin.h       | 112 ++++-----
 gcc/config/i386/avx10_2bf16intrin.h           | 232 +++++++++---------
 gcc/config/i386/i386-builtin.def              |  18 +-
 gcc/config/i386/sse.md                        |  22 +-
 gcc/testsuite/gcc.target/i386/avx-1.c         |  18 +-
 .../gcc.target/i386/avx10_2-512-bf16-1.c      |  30 +--
 ...pbf16-2.c => avx10_2-512-vgetmantbf16-2.c} |   0
 ...epbf16-2.c => avx10_2-512-vreducebf16-2.c} |   6 +-
 ...bf16-2.c => avx10_2-512-vrndscalebf16-2.c} |   6 +-
 .../gcc.target/i386/avx10_2-bf16-1.c          |  60 ++---
 ...mantpbf16-2.c => avx10_2-vgetmantbf16-2.c} |   4 +-
 ...ucenepbf16-2.c => avx10_2-vreducebf16-2.c} |   4 +-
 ...enepbf16-2.c => avx10_2-vrndscalebf16-2.c} |   4 +-
 gcc/testsuite/gcc.target/i386/sse-13.c        |  18 +-
 gcc/testsuite/gcc.target/i386/sse-14.c        |  36 +--
 gcc/testsuite/gcc.target/i386/sse-22.c        |  36 +--
 gcc/testsuite/gcc.target/i386/sse-23.c        |  18 +-
 17 files changed, 312 insertions(+), 312 deletions(-)
 rename gcc/testsuite/gcc.target/i386/{avx10_2-512-vgetmantpbf16-2.c => avx10_2-512-vgetmantbf16-2.c} (100%)
 rename gcc/testsuite/gcc.target/i386/{avx10_2-512-vreducenepbf16-2.c => avx10_2-512-vreducebf16-2.c} (87%)
 rename gcc/testsuite/gcc.target/i386/{avx10_2-512-vrndscalenepbf16-2.c => avx10_2-512-vrndscalebf16-2.c} (84%)
 rename gcc/testsuite/gcc.target/i386/{avx10_2-vgetmantpbf16-2.c => avx10_2-vgetmantbf16-2.c} (78%)
 rename gcc/testsuite/gcc.target/i386/{avx10_2-vreducenepbf16-2.c => avx10_2-vreducebf16-2.c} (78%)
 rename gcc/testsuite/gcc.target/i386/{avx10_2-vrndscalenepbf16-2.c => avx10_2-vrndscalebf16-2.c} (77%)
diff mbox series

Patch

diff --git a/gcc/config/i386/avx10_2-512bf16intrin.h b/gcc/config/i386/avx10_2-512bf16intrin.h
index fcd28534ddc..276a43890bd 100644
--- a/gcc/config/i386/avx10_2-512bf16intrin.h
+++ b/gcc/config/i386/avx10_2-512bf16intrin.h
@@ -468,100 +468,100 @@  _mm512_maskz_getexp_pbh (__mmask32 __U, __m512bh __A)
 					__U);
 }
 
-/* Intrinsics vrndscalepbf16.  */
+/* Intrinsics vrndscalebf16.  */
 #ifdef __OPTIMIZE__
 extern __inline__ __m512bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_roundscalene_pbh (__m512bh __A, int B)
+_mm512_roundscale_pbh (__m512bh __A, int B)
 {
   return (__m512bh)
-    __builtin_ia32_rndscalenepbf16512_mask (__A, B,
-					    (__v32bf) _mm512_setzero_si512 (),
-					    (__mmask32) -1);
+    __builtin_ia32_rndscalebf16512_mask (__A, B,
+					 (__v32bf) _mm512_setzero_si512 (),
+					 (__mmask32) -1);
 }
 
 extern __inline__ __m512bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_roundscalene_pbh (__m512bh __W, __mmask32 __U, __m512bh __A, int B)
+_mm512_mask_roundscale_pbh (__m512bh __W, __mmask32 __U, __m512bh __A, int B)
 {
   return (__m512bh)
-    __builtin_ia32_rndscalenepbf16512_mask (__A, B, __W,  __U);
+    __builtin_ia32_rndscalebf16512_mask (__A, B, __W,  __U);
 }
 
 extern __inline__ __m512bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_roundscalene_pbh (__mmask32 __U, __m512bh __A, int B)
+_mm512_maskz_roundscale_pbh (__mmask32 __U, __m512bh __A, int B)
 {
   return (__m512bh)
-    __builtin_ia32_rndscalenepbf16512_mask (__A, B,
-					    (__v32bf) _mm512_setzero_si512 (),
-					    __U);
+    __builtin_ia32_rndscalebf16512_mask (__A, B,
+					 (__v32bf) _mm512_setzero_si512 (),
+					 __U);
 }
 
 #else
-#define _mm512_roundscalene_pbh(A, B)					      \
-  (__builtin_ia32_rndscalenepbf16512_mask ((A), (B),			      \
-					   (__v32bf) _mm512_setzero_si512 (), \
-					   (__mmask32) -1))
+#define _mm512_roundscale_pbh(A, B)					      \
+  (__builtin_ia32_rndscalebf16512_mask ((A), (B),			      \
+					(__v32bf) _mm512_setzero_si512 (),    \
+					(__mmask32) -1))
 
-#define _mm512_mask_roundscalene_pbh(A, B, C, D)	    		      \
-  (__builtin_ia32_rndscalenepbf16512_mask ((C), (D), (A), (B)))
+#define _mm512_mask_roundscale_pbh(A, B, C, D)	    		      \
+  (__builtin_ia32_rndscalebf16512_mask ((C), (D), (A), (B)))
 
-#define _mm512_maskz_roundscalene_pbh(A, B, C)				      \
-  (__builtin_ia32_rndscalenepbf16512_mask ((B), (C),			      \
-					   (__v32bf) _mm512_setzero_si512 (), \
-					   (A)))
+#define _mm512_maskz_roundscale_pbh(A, B, C)				      \
+  (__builtin_ia32_rndscalebf16512_mask ((B), (C),			      \
+					(__v32bf) _mm512_setzero_si512 (),    \
+					(A)))
 
 #endif /* __OPTIMIZE__ */
 
-/* Intrinsics vreducepbf16.  */
+/* Intrinsics vreducebf16.  */
 #ifdef __OPTIMIZE__
 extern __inline__ __m512bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_reducene_pbh (__m512bh __A, int B)
+_mm512_reduce_pbh (__m512bh __A, int B)
 {
   return (__m512bh)
-    __builtin_ia32_reducenepbf16512_mask (__A, B,
-					  (__v32bf) _mm512_setzero_si512 (),
-					  (__mmask32) -1);
+    __builtin_ia32_reducebf16512_mask (__A, B,
+				       (__v32bf) _mm512_setzero_si512 (),
+				       (__mmask32) -1);
 }
 
 extern __inline__ __m512bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_reducene_pbh (__m512bh __W, __mmask32 __U,
-			  __m512bh __A, int B)
+_mm512_mask_reduce_pbh (__m512bh __W, __mmask32 __U,
+			__m512bh __A, int B)
 {
   return (__m512bh)
-    __builtin_ia32_reducenepbf16512_mask (__A, B, __W,  __U);
+    __builtin_ia32_reducebf16512_mask (__A, B, __W,  __U);
 }
 
 extern __inline__ __m512bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_reducene_pbh (__mmask32 __U, __m512bh __A, int B)
+_mm512_maskz_reduce_pbh (__mmask32 __U, __m512bh __A, int B)
 {
   return (__m512bh)
-    __builtin_ia32_reducenepbf16512_mask (__A, B,
+    __builtin_ia32_reducebf16512_mask (__A, B,
 					  (__v32bf) _mm512_setzero_si512 (),
 					  __U);
 }
 
 #else
-#define _mm512_reducene_pbh(A, B)					      \
-  (__builtin_ia32_reducenepbf16512_mask ((A), (B),			      \
-					 (__v32bf) _mm512_setzero_si512 (),   \
-					 (__mmask32) -1))
+#define _mm512_reduce_pbh(A, B)					      \
+  (__builtin_ia32_reducebf16512_mask ((A), (B),			      \
+				      (__v32bf) _mm512_setzero_si512 (),   \
+				      (__mmask32) -1))
 
-#define _mm512_mask_reducene_pbh(A, B, C, D)				      \
-  (__builtin_ia32_reducenepbf16512_mask ((C), (D), (A), (B)))
+#define _mm512_mask_reduce_pbh(A, B, C, D)				      \
+  (__builtin_ia32_reducebf16512_mask ((C), (D), (A), (B)))
 
-#define _mm512_maskz_reducene_pbh(A, B, C)				      \
-  (__builtin_ia32_reducenepbf16512_mask ((B), (C),			      \
-					 (__v32bf) _mm512_setzero_si512 (),   \
-					 (A)))
+#define _mm512_maskz_reduce_pbh(A, B, C)				      \
+  (__builtin_ia32_reducebf16512_mask ((B), (C),			      \
+				      (__v32bf) _mm512_setzero_si512 (),      \
+				      (A)))
 
 #endif /* __OPTIMIZE__ */
 
-/* Intrinsics vgetmantpbf16.  */
+/* Intrinsics vgetmantbf16.  */
 #ifdef __OPTIMIZE__
 extern __inline__ __m512bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
@@ -569,9 +569,9 @@  _mm512_getmant_pbh (__m512bh __A, _MM_MANTISSA_NORM_ENUM __B,
 		    _MM_MANTISSA_SIGN_ENUM __C)
 {
   return (__m512bh)
-    __builtin_ia32_getmantpbf16512_mask (__A, (int) (__C << 2) | __B,
-					 (__v32bf) _mm512_setzero_si512 (),
-					 (__mmask32) -1);
+    __builtin_ia32_getmantbf16512_mask (__A, (int) (__C << 2) | __B,
+					(__v32bf) _mm512_setzero_si512 (),
+					(__mmask32) -1);
 }
 
 extern __inline__ __m512bh
@@ -581,8 +581,8 @@  _mm512_mask_getmant_pbh (__m512bh __W, __mmask32 __U, __m512bh __A,
 			 _MM_MANTISSA_SIGN_ENUM __C)
 {
   return (__m512bh)
-    __builtin_ia32_getmantpbf16512_mask (__A, (int) (__C << 2) | __B,
-					 __W, __U);
+    __builtin_ia32_getmantbf16512_mask (__A, (int) (__C << 2) | __B,
+					__W, __U);
 }
 
 extern __inline__ __m512bh
@@ -592,23 +592,23 @@  _mm512_maskz_getmant_pbh (__mmask32 __U, __m512bh __A,
 			  _MM_MANTISSA_SIGN_ENUM __C)
 {
   return (__m512bh)
-    __builtin_ia32_getmantpbf16512_mask (__A, (int) (__C << 2) | __B,
-					 (__v32bf) _mm512_setzero_si512 (),
-					 __U);
+    __builtin_ia32_getmantbf16512_mask (__A, (int) (__C << 2) | __B,
+					(__v32bf) _mm512_setzero_si512 (),
+					__U);
 }
 
 #else
 #define _mm512_getmant_pbh(A, B, C)					      \
-  (__builtin_ia32_getmantpbf16512_mask ((A), (int)(((C)<<2) | (B)),	      \
-					  (__v32bf) _mm512_setzero_si512 (),  \
-					  (__mmask32) -1))
+  (__builtin_ia32_getmantbf16512_mask ((A), (int)(((C)<<2) | (B)),	      \
+				       (__v32bf) _mm512_setzero_si512 (),     \
+				       (__mmask32) -1))
 
 #define _mm512_mask_getmant_pbh(A, B, C, D, E)				      \
-  (__builtin_ia32_getmantpbf16512_mask ((C), (int)(((D)<<2) | (E)), (A), (B)))
+  (__builtin_ia32_getmantbf16512_mask ((C), (int)(((D)<<2) | (E)), (A), (B)))
 
 #define _mm512_maskz_getmant_pbh(A, B, C, D)				      \
-  (__builtin_ia32_getmantpbf16512_mask ((B), (int)(((C)<<2) | (D)),	      \
-					  (__v32bf) _mm512_setzero_si512 (),  \
+  (__builtin_ia32_getmantbf16512_mask ((B), (int)(((C)<<2) | (D)),	      \
+				       (__v32bf) _mm512_setzero_si512 (),     \
 					  (A)))
 
 #endif /* __OPTIMIZE__ */
diff --git a/gcc/config/i386/avx10_2bf16intrin.h b/gcc/config/i386/avx10_2bf16intrin.h
index 945556da636..891df893c00 100644
--- a/gcc/config/i386/avx10_2bf16intrin.h
+++ b/gcc/config/i386/avx10_2bf16intrin.h
@@ -901,186 +901,186 @@  _mm_maskz_getexp_pbh (__mmask8 __U, __m128bh __A)
 					__U);
 }
 
-/* Intrinsics vrndscalepbf16.  */
+/* Intrinsics vrndscalebf16.  */
 #ifdef __OPTIMIZE__
 extern __inline__ __m256bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_roundscalene_pbh (__m256bh __A, int B)
+_mm256_roundscale_pbh (__m256bh __A, int B)
 {
   return (__m256bh)
-    __builtin_ia32_rndscalenepbf16256_mask (__A, B,
-					    (__v16bf) _mm256_setzero_si256 (),
-					    (__mmask16) -1);
+    __builtin_ia32_rndscalebf16256_mask (__A, B,
+					 (__v16bf) _mm256_setzero_si256 (),
+					 (__mmask16) -1);
 }
 
 extern __inline__ __m256bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_roundscalene_pbh (__m256bh __W, __mmask16 __U,
-			      __m256bh __A, int B)
+_mm256_mask_roundscale_pbh (__m256bh __W, __mmask16 __U,
+			    __m256bh __A, int B)
 {
   return (__m256bh)
-    __builtin_ia32_rndscalenepbf16256_mask (__A, B, __W, __U);
+    __builtin_ia32_rndscalebf16256_mask (__A, B, __W, __U);
 }
 
 extern __inline__ __m256bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_roundscalene_pbh (__mmask16 __U, __m256bh __A, int B)
+_mm256_maskz_roundscale_pbh (__mmask16 __U, __m256bh __A, int B)
 {
   return (__m256bh)
-    __builtin_ia32_rndscalenepbf16256_mask (__A, B,
-					    (__v16bf) _mm256_setzero_si256 (),
-					    __U);
+    __builtin_ia32_rndscalebf16256_mask (__A, B,
+					 (__v16bf) _mm256_setzero_si256 (),
+					 __U);
 }
 
 extern __inline__ __m128bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_roundscalene_pbh (__m128bh __A, int B)
+_mm_roundscale_pbh (__m128bh __A, int B)
 {
   return (__m128bh)
-    __builtin_ia32_rndscalenepbf16128_mask (__A, B,
-					    (__v8bf) _mm_setzero_si128 (),
-					    (__mmask8) -1);
+    __builtin_ia32_rndscalebf16128_mask (__A, B,
+					 (__v8bf) _mm_setzero_si128 (),
+					 (__mmask8) -1);
 }
 
 extern __inline__ __m128bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_roundscalene_pbh (__m128bh __W, __mmask8 __U,
-			   __m128bh __A, int B)
+_mm_mask_roundscale_pbh (__m128bh __W, __mmask8 __U,
+			 __m128bh __A, int B)
 {
   return (__m128bh)
-    __builtin_ia32_rndscalenepbf16128_mask (__A, B, __W, __U);
+    __builtin_ia32_rndscalebf16128_mask (__A, B, __W, __U);
 }
 
 extern __inline__ __m128bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_roundscalene_pbh (__mmask8 __U, __m128bh __A, int B)
+_mm_maskz_roundscale_pbh (__mmask8 __U, __m128bh __A, int B)
 {
   return (__m128bh)
-    __builtin_ia32_rndscalenepbf16128_mask (__A, B,
-					    (__v8bf) _mm_setzero_si128 (),
-					    __U);
+    __builtin_ia32_rndscalebf16128_mask (__A, B,
+					 (__v8bf) _mm_setzero_si128 (),
+					 __U);
 }
 
 #else
-#define _mm256_roundscalene_pbh(A, B)					      \
-  (__builtin_ia32_rndscalenepbf16256_mask ((A), (B),			      \
-					   (__v16bf) _mm256_setzero_si256 (), \
-					   (__mmask16) -1))
+#define _mm256_roundscale_pbh(A, B)					      \
+  (__builtin_ia32_rndscalebf16256_mask ((A), (B),			      \
+					(__v16bf) _mm256_setzero_si256 (),    \
+					(__mmask16) -1))
 
-#define _mm256_mask_roundscalene_pbh(A, B, C, D)	    		      \
-  (__builtin_ia32_rndscalenepbf16256_mask ((C), (D), (A), (B)))
+#define _mm256_mask_roundscale_pbh(A, B, C, D)	    		      \
+  (__builtin_ia32_rndscalebf16256_mask ((C), (D), (A), (B)))
 
-#define _mm256_maskz_roundscalene_pbh(A, B, C)				      \
-  (__builtin_ia32_rndscalenepbf16256_mask ((B), (C),			      \
-					   (__v16bf) _mm256_setzero_si256 (), \
-					   (A)))
+#define _mm256_maskz_roundscale_pbh(A, B, C)				      \
+  (__builtin_ia32_rndscalebf16256_mask ((B), (C),			      \
+					(__v16bf) _mm256_setzero_si256 (),    \
+					(A)))
 
-#define _mm_roundscalene_pbh(A, B)					      \
-  (__builtin_ia32_rndscalenepbf16128_mask ((A), (B),			      \
-					   (__v8bf) _mm_setzero_si128 (),     \
-					   (__mmask8) -1))
+#define _mm_roundscale_pbh(A, B)					      \
+  (__builtin_ia32_rndscalebf16128_mask ((A), (B),			      \
+					(__v8bf) _mm_setzero_si128 (),	      \
+					(__mmask8) -1))
 
-#define _mm_mask_roundscalene_pbh(A, B, C, D)				      \
-  (__builtin_ia32_rndscalenepbf16128_mask ((C), (D), (A), (B)))
+#define _mm_mask_roundscale_pbh(A, B, C, D)				      \
+  (__builtin_ia32_rndscalebf16128_mask ((C), (D), (A), (B)))
 
-#define _mm_maskz_roundscalene_pbh(A, B, C)				      \
-  (__builtin_ia32_rndscalenepbf16128_mask ((B), (C),			      \
-					   (__v8bf) _mm_setzero_si128 (),     \
-					   (A)))
+#define _mm_maskz_roundscale_pbh(A, B, C)				      \
+  (__builtin_ia32_rndscalebf16128_mask ((B), (C),			      \
+					(__v8bf) _mm_setzero_si128 (),	      \
+					(A)))
 
 #endif /* __OPTIMIZE__ */
 
-/* Intrinsics vreducepbf16.  */
+/* Intrinsics vreducebf16.  */
 #ifdef __OPTIMIZE__
 extern __inline__ __m256bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_reducene_pbh (__m256bh __A, int B)
+_mm256_reduce_pbh (__m256bh __A, int B)
 {
   return (__m256bh)
-    __builtin_ia32_reducenepbf16256_mask (__A, B,
-					  (__v16bf) _mm256_setzero_si256 (),
-					  (__mmask16) -1);
+    __builtin_ia32_reducebf16256_mask (__A, B,
+				       (__v16bf) _mm256_setzero_si256 (),
+				       (__mmask16) -1);
 }
 
 extern __inline__ __m256bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_reducene_pbh (__m256bh __W, __mmask16 __U,
-			  __m256bh __A, int B)
+_mm256_mask_reduce_pbh (__m256bh __W, __mmask16 __U,
+			__m256bh __A, int B)
 {
   return (__m256bh)
-    __builtin_ia32_reducenepbf16256_mask (__A, B, __W, __U);
+    __builtin_ia32_reducebf16256_mask (__A, B, __W, __U);
 }
 
 extern __inline__ __m256bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_reducene_pbh (__mmask16 __U, __m256bh __A, int B)
+_mm256_maskz_reduce_pbh (__mmask16 __U, __m256bh __A, int B)
 {
   return (__m256bh)
-    __builtin_ia32_reducenepbf16256_mask (__A, B,
-					  (__v16bf) _mm256_setzero_si256 (),
-					  __U);
+    __builtin_ia32_reducebf16256_mask (__A, B,
+				       (__v16bf) _mm256_setzero_si256 (),
+				       __U);
 }
 
 extern __inline__ __m128bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_reducene_pbh (__m128bh __A, int B)
+_mm_reduce_pbh (__m128bh __A, int B)
 {
   return (__m128bh)
-    __builtin_ia32_reducenepbf16128_mask (__A, B,
-					  (__v8bf) _mm_setzero_si128 (),
-					  (__mmask8) -1);
+    __builtin_ia32_reducebf16128_mask (__A, B,
+				       (__v8bf) _mm_setzero_si128 (),
+				       (__mmask8) -1);
 }
 
 extern __inline__ __m128bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_reducene_pbh (__m128bh __W, __mmask8 __U,
-		       __m128bh __A, int B)
+_mm_mask_reduce_pbh (__m128bh __W, __mmask8 __U,
+		     __m128bh __A, int B)
 {
   return (__m128bh)
-    __builtin_ia32_reducenepbf16128_mask (__A, B, __W, __U);
+    __builtin_ia32_reducebf16128_mask (__A, B, __W, __U);
 }
 
 extern __inline__ __m128bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_reducene_pbh (__mmask8 __U, __m128bh __A, int B)
+_mm_maskz_reduce_pbh (__mmask8 __U, __m128bh __A, int B)
 {
   return (__m128bh)
-    __builtin_ia32_reducenepbf16128_mask (__A, B,
-					  (__v8bf) _mm_setzero_si128 (),
-					  __U);
+    __builtin_ia32_reducebf16128_mask (__A, B,
+				       (__v8bf) _mm_setzero_si128 (),
+				       __U);
 }
 
 #else
-#define _mm256_reducene_pbh(A, B)					      \
-  (__builtin_ia32_reducenepbf16256_mask ((A), (B),			      \
-					 (__v16bf) _mm256_setzero_si256 (),   \
-					 (__mmask16) -1))
+#define _mm256_reduce_pbh(A, B)					      \
+  (__builtin_ia32_reducebf16256_mask ((A), (B),			      \
+				      (__v16bf) _mm256_setzero_si256 (),      \
+				      (__mmask16) -1))
 
-#define _mm256_mask_reducene_pbh(A, B, C, D)				      \
-  (__builtin_ia32_reducenepbf16256_mask ((C), (D), (A), (B)))
+#define _mm256_mask_reduce_pbh(A, B, C, D)				      \
+  (__builtin_ia32_reducebf16256_mask ((C), (D), (A), (B)))
 
-#define _mm256_maskz_reducene_pbh(A, B, C)				      \
-  (__builtin_ia32_reducenepbf16256_mask ((B), (C),			      \
-					 (__v16bf) _mm256_setzero_si256 (),   \
-					 (A)))
+#define _mm256_maskz_reduce_pbh(A, B, C)				      \
+  (__builtin_ia32_reducebf16256_mask ((B), (C),			      \
+				      (__v16bf) _mm256_setzero_si256 (),      \
+				      (A)))
 
-#define _mm_reducene_pbh(A, B)						      \
-  (__builtin_ia32_reducenepbf16128_mask ((A), (B),			      \
-					 (__v8bf) _mm_setzero_si128 (),       \
-					 (__mmask8) -1))
+#define _mm_reduce_pbh(A, B)						      \
+  (__builtin_ia32_reducebf16128_mask ((A), (B),			      \
+				      (__v8bf) _mm_setzero_si128 (),	      \
+				      (__mmask8) -1))
 
-#define _mm_mask_reducene_pbh(A, B, C, D)				      \
-  (__builtin_ia32_reducenepbf16128_mask ((C), (D), (A), (B)))
+#define _mm_mask_reduce_pbh(A, B, C, D)				      \
+  (__builtin_ia32_reducebf16128_mask ((C), (D), (A), (B)))
 
-#define _mm_maskz_reducene_pbh(A, B, C)					      \
-  (__builtin_ia32_reducenepbf16128_mask ((B), (C),			      \
-					 (__v8bf) _mm_setzero_si128 (),       \
-					 (A)))
+#define _mm_maskz_reduce_pbh(A, B, C)					      \
+  (__builtin_ia32_reducebf16128_mask ((B), (C),			      \
+				      (__v8bf) _mm_setzero_si128 (),	      \
+				      (A)))
 
 #endif /* __OPTIMIZE__ */
 
 
-/* Intrinsics vgetmantpbf16.  */
+/* Intrinsics vgetmantbf16.  */
 #ifdef __OPTIMIZE__
 extern __inline__ __m256bh
 __attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
@@ -1088,9 +1088,9 @@  _mm256_getmant_pbh (__m256bh __A, _MM_MANTISSA_NORM_ENUM __B,
 		    _MM_MANTISSA_SIGN_ENUM __C)
 {
   return (__m256bh)
-    __builtin_ia32_getmantpbf16256_mask (__A, (int) (__C << 2) | __B,
-					 (__v16bf) _mm256_setzero_si256 (),
-					 (__mmask16) -1);
+    __builtin_ia32_getmantbf16256_mask (__A, (int) (__C << 2) | __B,
+					(__v16bf) _mm256_setzero_si256 (),
+					(__mmask16) -1);
 }
 
 extern __inline__ __m256bh
@@ -1100,8 +1100,8 @@  _mm256_mask_getmant_pbh (__m256bh __W, __mmask16 __U, __m256bh __A,
 			 _MM_MANTISSA_SIGN_ENUM __C)
 {
   return (__m256bh)
-    __builtin_ia32_getmantpbf16256_mask (__A, (int) (__C << 2) | __B,
-					 __W, __U);
+    __builtin_ia32_getmantbf16256_mask (__A, (int) (__C << 2) | __B,
+					__W, __U);
 }
 
 extern __inline__ __m256bh
@@ -1111,9 +1111,9 @@  _mm256_maskz_getmant_pbh (__mmask16 __U, __m256bh __A,
 			  _MM_MANTISSA_SIGN_ENUM __C)
 {
   return (__m256bh)
-    __builtin_ia32_getmantpbf16256_mask (__A, (int) (__C << 2) | __B,
-					 (__v16bf) _mm256_setzero_si256 (),
-					 __U);
+    __builtin_ia32_getmantbf16256_mask (__A, (int) (__C << 2) | __B,
+					(__v16bf) _mm256_setzero_si256 (),
+					__U);
 }
 
 extern __inline__ __m128bh
@@ -1122,9 +1122,9 @@  _mm_getmant_pbh (__m128bh __A, _MM_MANTISSA_NORM_ENUM __B,
 		 _MM_MANTISSA_SIGN_ENUM __C)
 {
   return (__m128bh)
-    __builtin_ia32_getmantpbf16128_mask (__A, (int) (__C << 2) | __B,
-					 (__v8bf) _mm_setzero_si128 (),
-					 (__mmask8) -1);
+    __builtin_ia32_getmantbf16128_mask (__A, (int) (__C << 2) | __B,
+					(__v8bf) _mm_setzero_si128 (),
+					(__mmask8) -1);
 }
 
 extern __inline__ __m128bh
@@ -1134,8 +1134,8 @@  _mm_mask_getmant_pbh (__m128bh __W, __mmask8 __U, __m128bh __A,
 		      _MM_MANTISSA_SIGN_ENUM __C)
 {
   return (__m128bh)
-    __builtin_ia32_getmantpbf16128_mask (__A, (int) (__C << 2) | __B,
-					 __W, __U);
+    __builtin_ia32_getmantbf16128_mask (__A, (int) (__C << 2) | __B,
+					__W, __U);
 }
 
 extern __inline__ __m128bh
@@ -1145,36 +1145,36 @@  _mm_maskz_getmant_pbh (__mmask8 __U, __m128bh __A,
 		       _MM_MANTISSA_SIGN_ENUM __C)
 {
   return (__m128bh)
-    __builtin_ia32_getmantpbf16128_mask (__A, (int) (__C << 2) | __B,
-					 (__v8bf) _mm_setzero_si128 (),
-					 __U);
+    __builtin_ia32_getmantbf16128_mask (__A, (int) (__C << 2) | __B,
+					(__v8bf) _mm_setzero_si128 (),
+					__U);
 }
 
 #else
 #define _mm256_getmant_pbh(A, B, C)					      \
-  (__builtin_ia32_getmantpbf16256_mask ((A), (int)(((C)<<2) | (B)),	      \
-					   (__v16bf) _mm256_setzero_si256 (), \
-					   (__mmask16) (-1)))
+  (__builtin_ia32_getmantbf16256_mask ((A), (int)(((C)<<2) | (B)),	      \
+				       (__v16bf) _mm256_setzero_si256 (),     \
+				       (__mmask16) (-1)))
 
 #define _mm256_mask_getmant_pbh(A, B, C, D, E)				      \
-  (__builtin_ia32_getmantpbf16256_mask ((C), (int)(((D)<<2) | (E)), (A), (B)))
+  (__builtin_ia32_getmantbf16256_mask ((C), (int)(((D)<<2) | (E)), (A), (B)))
 
 #define _mm256_maskz_getmant_pbh(A, B, C, D)				      \
-  (__builtin_ia32_getmantpbf16256_mask ((B), (int)(((C)<<2) | (D)),	      \
-					   (__v16bf) _mm256_setzero_si256 (), \
-					   (A)))
+  (__builtin_ia32_getmantbf16256_mask ((B), (int)(((C)<<2) | (D)),	      \
+				       (__v16bf) _mm256_setzero_si256 (),     \
+				       (A)))
 
 #define _mm_getmant_pbh(A, B, C)					      \
-  (__builtin_ia32_getmantpbf16128_mask ((A), (int)(((C)<<2) | (B)),	      \
-					(__v8bf) _mm_setzero_si128 (),	      \
-					(__mmask8) (-1)))
+  (__builtin_ia32_getmantbf16128_mask ((A), (int)(((C)<<2) | (B)),	      \
+				       (__v8bf) _mm_setzero_si128 (),	      \
+				       (__mmask8) (-1)))
 
 #define _mm_mask_getmant_pbh(A, B, C, D, E)				      \
-  (__builtin_ia32_getmantpbf16128_mask ((C), (int)(((D)<<2) | (E)), (A), (B)))
+  (__builtin_ia32_getmantbf16128_mask ((C), (int)(((D)<<2) | (E)), (A), (B)))
 
 #define _mm_maskz_getmant_pbh(A, B, C, D)				      \
-  (__builtin_ia32_getmantpbf16128_mask ((B), (int)(((C)<<2) | (D)),	      \
-					(__v8bf) _mm_setzero_si128 (), (A)))
+  (__builtin_ia32_getmantbf16128_mask ((B), (int)(((C)<<2) | (D)),	      \
+				       (__v8bf) _mm_setzero_si128 (), (A)))
 
 #endif /* __OPTIMIZE__ */
 
diff --git a/gcc/config/i386/i386-builtin.def b/gcc/config/i386/i386-builtin.def
index 3d51cb62e49..17f1c173629 100644
--- a/gcc/config/i386/i386-builtin.def
+++ b/gcc/config/i386/i386-builtin.def
@@ -3269,15 +3269,15 @@  BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_rcpbf16_v8bf_mask, "__b
 BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_avx10_2_getexppbf16_v32bf_mask, "__builtin_ia32_getexppbf16512_mask", IX86_BUILTIN_GETEXPPBF16512_MASK, UNKNOWN, (int) V32BF_FTYPE_V32BF_V32BF_USI)
 BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_getexppbf16_v16bf_mask, "__builtin_ia32_getexppbf16256_mask", IX86_BUILTIN_GETEXPPBF16256_MASK, UNKNOWN, (int) V16BF_FTYPE_V16BF_V16BF_UHI)
 BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_getexppbf16_v8bf_mask, "__builtin_ia32_getexppbf16128_mask", IX86_BUILTIN_GETEXPPBF16128_MASK, UNKNOWN, (int) V8BF_FTYPE_V8BF_V8BF_UQI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_avx10_2_rndscalenepbf16_v32bf_mask, "__builtin_ia32_rndscalenepbf16512_mask", IX86_BUILTIN_RNDSCALENEPBF16512_MASK, UNKNOWN, (int) V32BF_FTYPE_V32BF_INT_V32BF_USI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_rndscalenepbf16_v16bf_mask, "__builtin_ia32_rndscalenepbf16256_mask", IX86_BUILTIN_RNDSCALENEPBF16256_MASK, UNKNOWN, (int) V16BF_FTYPE_V16BF_INT_V16BF_UHI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_rndscalenepbf16_v8bf_mask, "__builtin_ia32_rndscalenepbf16128_mask", IX86_BUILTIN_RNDSCALENEPBF16128_MASK, UNKNOWN, (int) V8BF_FTYPE_V8BF_INT_V8BF_UQI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_avx10_2_reducenepbf16_v32bf_mask, "__builtin_ia32_reducenepbf16512_mask", IX86_BUILTIN_REDUCENEPBF16512_MASK, UNKNOWN, (int) V32BF_FTYPE_V32BF_INT_V32BF_USI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_reducenepbf16_v16bf_mask, "__builtin_ia32_reducenepbf16256_mask", IX86_BUILTIN_REDUCENEPBF16256_MASK, UNKNOWN, (int) V16BF_FTYPE_V16BF_INT_V16BF_UHI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_reducenepbf16_v8bf_mask, "__builtin_ia32_reducenepbf16128_mask", IX86_BUILTIN_REDUCENEPBF16128_MASK, UNKNOWN, (int) V8BF_FTYPE_V8BF_INT_V8BF_UQI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_avx10_2_getmantpbf16_v32bf_mask, "__builtin_ia32_getmantpbf16512_mask", IX86_BUILTIN_GETMANTPBF16512_MASK, UNKNOWN, (int) V32BF_FTYPE_V32BF_INT_V32BF_USI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_getmantpbf16_v16bf_mask, "__builtin_ia32_getmantpbf16256_mask", IX86_BUILTIN_GETMANTPBF16256_MASK, UNKNOWN, (int) V16BF_FTYPE_V16BF_INT_V16BF_UHI)
-BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_getmantpbf16_v8bf_mask, "__builtin_ia32_getmantpbf16128_mask", IX86_BUILTIN_GETMANTPBF16128_MASK, UNKNOWN, (int) V8BF_FTYPE_V8BF_INT_V8BF_UQI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_avx10_2_rndscalebf16_v32bf_mask, "__builtin_ia32_rndscalebf16512_mask", IX86_BUILTIN_RNDSCALEBF16512_MASK, UNKNOWN, (int) V32BF_FTYPE_V32BF_INT_V32BF_USI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_rndscalebf16_v16bf_mask, "__builtin_ia32_rndscalebf16256_mask", IX86_BUILTIN_RNDSCALEBF16256_MASK, UNKNOWN, (int) V16BF_FTYPE_V16BF_INT_V16BF_UHI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_rndscalebf16_v8bf_mask, "__builtin_ia32_rndscalebf16128_mask", IX86_BUILTIN_RNDSCALEBF16128_MASK, UNKNOWN, (int) V8BF_FTYPE_V8BF_INT_V8BF_UQI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_avx10_2_reducebf16_v32bf_mask, "__builtin_ia32_reducebf16512_mask", IX86_BUILTIN_REDUCEBF16512_MASK, UNKNOWN, (int) V32BF_FTYPE_V32BF_INT_V32BF_USI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_reducebf16_v16bf_mask, "__builtin_ia32_reducebf16256_mask", IX86_BUILTIN_REDUCEBF16256_MASK, UNKNOWN, (int) V16BF_FTYPE_V16BF_INT_V16BF_UHI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_reducebf16_v8bf_mask, "__builtin_ia32_reducebf16128_mask", IX86_BUILTIN_REDUCEBF16128_MASK, UNKNOWN, (int) V8BF_FTYPE_V8BF_INT_V8BF_UQI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_avx10_2_getmantbf16_v32bf_mask, "__builtin_ia32_getmantbf16512_mask", IX86_BUILTIN_GETMANTBF16512_MASK, UNKNOWN, (int) V32BF_FTYPE_V32BF_INT_V32BF_USI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_getmantbf16_v16bf_mask, "__builtin_ia32_getmantbf16256_mask", IX86_BUILTIN_GETMANTBF16256_MASK, UNKNOWN, (int) V16BF_FTYPE_V16BF_INT_V16BF_UHI)
+BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_getmantbf16_v8bf_mask, "__builtin_ia32_getmantbf16128_mask", IX86_BUILTIN_GETMANTBF16128_MASK, UNKNOWN, (int) V8BF_FTYPE_V8BF_INT_V8BF_UQI)
 BDESC (0, OPTION_MASK_ISA2_AVX10_2_512, CODE_FOR_avx10_2_fpclasspbf16_v32bf_mask, "__builtin_ia32_fpclasspbf16512_mask", IX86_BUILTIN_FPCLASSPBF16512_MASK, UNKNOWN, (int) SI_FTYPE_V32BF_INT_USI)
 BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_fpclasspbf16_v16bf_mask, "__builtin_ia32_fpclasspbf16256_mask", IX86_BUILTIN_FPCLASSPBF16256_MASK, UNKNOWN, (int) HI_FTYPE_V16BF_INT_UHI)
 BDESC (0, OPTION_MASK_ISA2_AVX10_2_256, CODE_FOR_avx10_2_fpclasspbf16_v8bf_mask, "__builtin_ia32_fpclasspbf16128_mask", IX86_BUILTIN_FPCLASSPBF16128_MASK, UNKNOWN, (int) QI_FTYPE_V8BF_INT_UQI)
diff --git a/gcc/config/i386/sse.md b/gcc/config/i386/sse.md
index 44f4e18e7f8..7f844987557 100644
--- a/gcc/config/i386/sse.md
+++ b/gcc/config/i386/sse.md
@@ -230,9 +230,9 @@ 
   UNSPEC_VCVTNEPH2HF8S
   UNSPEC_VCVTHF82PH
   UNSPEC_VSCALEFPBF16
-  UNSPEC_VRNDSCALENEPBF16
-  UNSPEC_VREDUCENEPBF16
-  UNSPEC_VGETMANTPBF16
+  UNSPEC_VRNDSCALEBF16
+  UNSPEC_VREDUCEBF16
+  UNSPEC_VGETMANTBF16
   UNSPEC_VFPCLASSPBF16
   UNSPEC_VCOMSBF16
   UNSPEC_VCVTNEBF162IBS
@@ -32407,23 +32407,23 @@ 
    [(set_attr "prefix" "evex")])
 
 (define_int_iterator BF16IMMOP
-   [UNSPEC_VRNDSCALENEPBF16
-    UNSPEC_VREDUCENEPBF16
-    UNSPEC_VGETMANTPBF16])
+   [UNSPEC_VRNDSCALEBF16
+    UNSPEC_VREDUCEBF16
+    UNSPEC_VGETMANTBF16])
 
 (define_int_attr bf16immop
-   [(UNSPEC_VRNDSCALENEPBF16 "rndscalene")
-    (UNSPEC_VREDUCENEPBF16 "reducene")
-    (UNSPEC_VGETMANTPBF16 "getmant")])
+   [(UNSPEC_VRNDSCALEBF16 "rndscale")
+    (UNSPEC_VREDUCEBF16 "reduce")
+    (UNSPEC_VGETMANTBF16 "getmant")])
 
-(define_insn "avx10_2_<bf16immop>pbf16_<mode><mask_name>"
+(define_insn "avx10_2_<bf16immop>bf16_<mode><mask_name>"
    [(set (match_operand:VBF_AVX10_2 0 "register_operand" "=v")
 	 (unspec:VBF_AVX10_2
 	   [(match_operand:VBF_AVX10_2 1 "nonimmediate_operand" "vm")
 	    (match_operand:SI 2 "const_0_to_255_operand")]
 	  BF16IMMOP))]
    "TARGET_AVX10_2_256"
-   "v<bf16immop>pbf16\t{%2, %1, %0<mask_operand3>|%0<mask_operand3>, %1, %2}"
+   "v<bf16immop>bf16\t{%2, %1, %0<mask_operand3>|%0<mask_operand3>, %1, %2}"
    [(set_attr "prefix" "evex")])
 
 (define_insn "avx10_2_fpclasspbf16_<mode><mask_scalar_merge_name>"
diff --git a/gcc/testsuite/gcc.target/i386/avx-1.c b/gcc/testsuite/gcc.target/i386/avx-1.c
index 36e95a7800a..ba2a2bb1dad 100644
--- a/gcc/testsuite/gcc.target/i386/avx-1.c
+++ b/gcc/testsuite/gcc.target/i386/avx-1.c
@@ -1017,19 +1017,19 @@ 
 #define __builtin_ia32_vcvt2ps2phx512_mask_round(A, B, C, D, E) __builtin_ia32_vcvt2ps2phx512_mask_round(A, B, C, D, 8)
 
 /* avx10_2-512bf16intrin.h */
-#define __builtin_ia32_rndscalenepbf16512_mask(A, B, C, D) __builtin_ia32_rndscalenepbf16512_mask(A, 123, C, D)
-#define __builtin_ia32_reducenepbf16512_mask(A, B, C, D) __builtin_ia32_reducenepbf16512_mask(A, 123, C, D)
-#define __builtin_ia32_getmantpbf16512_mask(A, B, C, D) __builtin_ia32_getmantpbf16512_mask(A, 1, C, D)
+#define __builtin_ia32_rndscalebf16512_mask(A, B, C, D) __builtin_ia32_rndscalebf16512_mask(A, 123, C, D)
+#define __builtin_ia32_reducebf16512_mask(A, B, C, D) __builtin_ia32_reducebf16512_mask(A, 123, C, D)
+#define __builtin_ia32_getmantbf16512_mask(A, B, C, D) __builtin_ia32_getmantbf16512_mask(A, 1, C, D)
 #define __builtin_ia32_fpclasspbf16512_mask(A, B, C) __builtin_ia32_fpclasspbf16512_mask(A, 1, C)
 #define __builtin_ia32_cmpbf16512_mask(A, B, C, D) __builtin_ia32_cmpbf16512_mask(A, B, 1, D)
 
 /* avx10_2bf16intrin.h */
-#define __builtin_ia32_rndscalenepbf16256_mask(A, B, C, D) __builtin_ia32_rndscalenepbf16256_mask(A, 123, C, D)
-#define __builtin_ia32_rndscalenepbf16128_mask(A, B, C, D) __builtin_ia32_rndscalenepbf16128_mask(A, 123, C, D)
-#define __builtin_ia32_reducenepbf16256_mask(A, B, C, D) __builtin_ia32_reducenepbf16256_mask(A, 123, C, D)
-#define __builtin_ia32_reducenepbf16128_mask(A, B, C, D) __builtin_ia32_reducenepbf16128_mask(A, 123, C, D)
-#define __builtin_ia32_getmantpbf16256_mask(A, B, C, D) __builtin_ia32_getmantpbf16256_mask(A, 1, C, D)
-#define __builtin_ia32_getmantpbf16128_mask(A, B, C, D) __builtin_ia32_getmantpbf16128_mask(A, 1, C, D)
+#define __builtin_ia32_rndscalebf16256_mask(A, B, C, D) __builtin_ia32_rndscalebf16256_mask(A, 123, C, D)
+#define __builtin_ia32_rndscalebf16128_mask(A, B, C, D) __builtin_ia32_rndscalebf16128_mask(A, 123, C, D)
+#define __builtin_ia32_reducebf16256_mask(A, B, C, D) __builtin_ia32_reducebf16256_mask(A, 123, C, D)
+#define __builtin_ia32_reducebf16128_mask(A, B, C, D) __builtin_ia32_reducebf16128_mask(A, 123, C, D)
+#define __builtin_ia32_getmantbf16256_mask(A, B, C, D) __builtin_ia32_getmantbf16256_mask(A, 1, C, D)
+#define __builtin_ia32_getmantbf16128_mask(A, B, C, D) __builtin_ia32_getmantbf16128_mask(A, 1, C, D)
 #define __builtin_ia32_fpclasspbf16256_mask(A, B, C) __builtin_ia32_fpclasspbf16256_mask(A, 1, C)
 #define __builtin_ia32_fpclasspbf16128_mask(A, B, C) __builtin_ia32_fpclasspbf16128_mask(A, 1, C)
 #define __builtin_ia32_cmpbf16256_mask(A, B, C, D) __builtin_ia32_cmpbf16256_mask(A, B, 1, D)
diff --git a/gcc/testsuite/gcc.target/i386/avx10_2-512-bf16-1.c b/gcc/testsuite/gcc.target/i386/avx10_2-512-bf16-1.c
index dd4d81e4dfc..df194132f39 100644
--- a/gcc/testsuite/gcc.target/i386/avx10_2-512-bf16-1.c
+++ b/gcc/testsuite/gcc.target/i386/avx10_2-512-bf16-1.c
@@ -49,15 +49,15 @@ 
 /* { dg-final { scan-assembler-times "vgetexppbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\[^\n\r]*%zmm\[0-9\]+(?:\n|\[ \\t\]+#)" 1 } } */
 /* { dg-final { scan-assembler-times "vgetexppbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\[^\n\r]*%zmm\[0-9\]+\{%k\[0-9\]\}\[^\{\n\r]*(?:\n|\[ \\t\]+#)" 1 } } */
 /* { dg-final { scan-assembler-times "vgetexppbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\[^\n\r]*%zmm\[0-9\]+\{%k\[0-9\]\}\{z\}\[^\n\r]*(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vrndscalenepbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
-/* { dg-final { scan-assembler-times "vrndscalenepbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vrndscalenepbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vreducenepbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
-/* { dg-final { scan-assembler-times "vreducenepbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vreducenepbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vgetmantpbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
-/* { dg-final { scan-assembler-times "vgetmantpbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vgetmantpbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vrndscalebf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
+/* { dg-final { scan-assembler-times "vrndscalebf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vrndscalebf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vreducebf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
+/* { dg-final { scan-assembler-times "vreducebf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vreducebf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vgetmantbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
+/* { dg-final { scan-assembler-times "vgetmantbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vgetmantbf16\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
 /* { dg-final { scan-assembler-times "vfpclasspbf16z\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\[^\n^k\]*%k\[0-7\](?:\n|\[ \\t\]+#)" 1 } } */
 /* { dg-final { scan-assembler-times "vfpclasspbf16z\[ \\t\]+\[^\{\n\]*%zmm\[0-9\]+\[^\n^k\]*%k\[0-7\]\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
 /* { dg-final { scan-assembler-times "vcmpbf16\[ \\t\]+\\\$1\[^\n\r]*%zmm\[0-9\]+\[^\n\r]*%zmm\[0-9\]+\[^\n\r]*%k\[0-9\](?:\n|\[ \\t\]+#)" 1 } } */
@@ -125,12 +125,12 @@  avx10_2_512_test (void)
   res = _mm512_mask_getexp_pbh (res, m32, x1);
   res = _mm512_maskz_getexp_pbh (m32, x1);
   
-  res = _mm512_roundscalene_pbh (x1, IMM);
-  res = _mm512_mask_roundscalene_pbh (res, m32, x1, IMM);
-  res = _mm512_maskz_roundscalene_pbh (m32, x1, IMM);
-  res = _mm512_reducene_pbh (x1, IMM);
-  res = _mm512_mask_reducene_pbh (res, m32, x1, IMM);
-  res = _mm512_maskz_reducene_pbh (m32, x1, IMM);
+  res = _mm512_roundscale_pbh (x1, IMM);
+  res = _mm512_mask_roundscale_pbh (res, m32, x1, IMM);
+  res = _mm512_maskz_roundscale_pbh (m32, x1, IMM);
+  res = _mm512_reduce_pbh (x1, IMM);
+  res = _mm512_mask_reduce_pbh (res, m32, x1, IMM);
+  res = _mm512_maskz_reduce_pbh (m32, x1, IMM);
   res = _mm512_getmant_pbh (x1, _MM_MANT_NORM_p75_1p5, _MM_MANT_SIGN_src);
   res = _mm512_mask_getmant_pbh (res, m32, x1, _MM_MANT_NORM_p75_1p5,
 				 _MM_MANT_SIGN_src);
diff --git a/gcc/testsuite/gcc.target/i386/avx10_2-512-vgetmantpbf16-2.c b/gcc/testsuite/gcc.target/i386/avx10_2-512-vgetmantbf16-2.c
similarity index 100%
rename from gcc/testsuite/gcc.target/i386/avx10_2-512-vgetmantpbf16-2.c
rename to gcc/testsuite/gcc.target/i386/avx10_2-512-vgetmantbf16-2.c
diff --git a/gcc/testsuite/gcc.target/i386/avx10_2-512-vreducenepbf16-2.c b/gcc/testsuite/gcc.target/i386/avx10_2-512-vreducebf16-2.c
similarity index 87%
rename from gcc/testsuite/gcc.target/i386/avx10_2-512-vreducenepbf16-2.c
rename to gcc/testsuite/gcc.target/i386/avx10_2-512-vreducebf16-2.c
index 3c19dd2ace4..1bfca414025 100644
--- a/gcc/testsuite/gcc.target/i386/avx10_2-512-vreducenepbf16-2.c
+++ b/gcc/testsuite/gcc.target/i386/avx10_2-512-vreducebf16-2.c
@@ -34,9 +34,9 @@  TEST (void)
       res_ref[i] = res_ref2[i] =  convert_fp32_to_bf16_ne (res);
     }
 
-  res1.x = INTRINSIC (_reducene_pbh) (src1.x, 0x10);
-  res2.x = INTRINSIC (_mask_reducene_pbh) (res2.x, mask, src1.x, 0x10);
-  res3.x = INTRINSIC (_maskz_reducene_pbh) (mask, src1.x, 0x10);
+  res1.x = INTRINSIC (_reduce_pbh) (src1.x, 0x10);
+  res2.x = INTRINSIC (_mask_reduce_pbh) (res2.x, mask, src1.x, 0x10);
+  res3.x = INTRINSIC (_maskz_reduce_pbh) (mask, src1.x, 0x10);
 
   if (UNION_CHECK (AVX512F_LEN, bf16_uw) (res1, res_ref))
     abort ();
diff --git a/gcc/testsuite/gcc.target/i386/avx10_2-512-vrndscalenepbf16-2.c b/gcc/testsuite/gcc.target/i386/avx10_2-512-vrndscalebf16-2.c
similarity index 84%
rename from gcc/testsuite/gcc.target/i386/avx10_2-512-vrndscalenepbf16-2.c
rename to gcc/testsuite/gcc.target/i386/avx10_2-512-vrndscalebf16-2.c
index 9be6bca5824..6f671d89f13 100644
--- a/gcc/testsuite/gcc.target/i386/avx10_2-512-vrndscalenepbf16-2.c
+++ b/gcc/testsuite/gcc.target/i386/avx10_2-512-vrndscalebf16-2.c
@@ -30,9 +30,9 @@  TEST (void)
       res_ref[i] = res_ref2[i] =  convert_fp32_to_bf16_ne (rndscale (x));
     }
 
-  res1.x = INTRINSIC (_roundscalene_pbh) (src1.x, 0x10);
-  res2.x = INTRINSIC (_mask_roundscalene_pbh) (res2.x, mask, src1.x, 0x10);
-  res3.x = INTRINSIC (_maskz_roundscalene_pbh) (mask, src1.x, 0x10);
+  res1.x = INTRINSIC (_roundscale_pbh) (src1.x, 0x10);
+  res2.x = INTRINSIC (_mask_roundscale_pbh) (res2.x, mask, src1.x, 0x10);
+  res3.x = INTRINSIC (_maskz_roundscale_pbh) (mask, src1.x, 0x10);
 
   if (UNION_CHECK (AVX512F_LEN, bf16_uw) (res1, res_ref))
     abort ();
diff --git a/gcc/testsuite/gcc.target/i386/avx10_2-bf16-1.c b/gcc/testsuite/gcc.target/i386/avx10_2-bf16-1.c
index a4841e53be4..74addd9fb19 100644
--- a/gcc/testsuite/gcc.target/i386/avx10_2-bf16-1.c
+++ b/gcc/testsuite/gcc.target/i386/avx10_2-bf16-1.c
@@ -98,24 +98,24 @@ 
 /* { dg-final { scan-assembler-times "vgetexppbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\[^\n\r]*%xmm\[0-9\]+(?:\n|\[ \\t\]+#)" 1 } } */
 /* { dg-final { scan-assembler-times "vgetexppbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\[^\n\r]*%xmm\[0-9\]+\{%k\[0-9\]\}\[^\{\n\r]*(?:\n|\[ \\t\]+#)" 1 } } */
 /* { dg-final { scan-assembler-times "vgetexppbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\[^\n\r]*%xmm\[0-9\]+\{%k\[0-9\]\}\{z\}\[^\n\r]*(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vrndscalenepbf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
-/* { dg-final { scan-assembler-times "vrndscalenepbf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vrndscalenepbf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vrndscalenepbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
-/* { dg-final { scan-assembler-times "vrndscalenepbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vrndscalenepbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vreducenepbf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
-/* { dg-final { scan-assembler-times "vreducenepbf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vreducenepbf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vreducenepbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
-/* { dg-final { scan-assembler-times "vreducenepbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vreducenepbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vgetmantpbf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
-/* { dg-final { scan-assembler-times "vgetmantpbf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vgetmantpbf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vgetmantpbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
-/* { dg-final { scan-assembler-times "vgetmantpbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
-/* { dg-final { scan-assembler-times "vgetmantpbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vrndscalebf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
+/* { dg-final { scan-assembler-times "vrndscalebf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vrndscalebf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vrndscalebf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
+/* { dg-final { scan-assembler-times "vrndscalebf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vrndscalebf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vreducebf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
+/* { dg-final { scan-assembler-times "vreducebf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vreducebf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vreducebf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
+/* { dg-final { scan-assembler-times "vreducebf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vreducebf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vgetmantbf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
+/* { dg-final { scan-assembler-times "vgetmantbf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vgetmantbf16\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vgetmantbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+(?:\n|\[ \\t\]+#)"  1 } } */
+/* { dg-final { scan-assembler-times "vgetmantbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
+/* { dg-final { scan-assembler-times "vgetmantbf16\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\{%k\[1-7\]\}\{z\}(?:\n|\[ \\t\]+#)" 1 } } */
 /* { dg-final { scan-assembler-times "vfpclasspbf16y\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\[^\n^k\]*%k\[0-7\](?:\n|\[ \\t\]+#)" 1 } } */
 /* { dg-final { scan-assembler-times "vfpclasspbf16y\[ \\t\]+\[^\{\n\]*%ymm\[0-9\]+\[^\n^k\]*%k\[0-7\]\{%k\[1-7\]\}(?:\n|\[ \\t\]+#)" 1 } } */
 /* { dg-final { scan-assembler-times "vfpclasspbf16x\[ \\t\]+\[^\{\n\]*%xmm\[0-9\]+\[^\n^k\]*%k\[0-7\](?:\n|\[ \\t\]+#)" 1 } } */
@@ -249,19 +249,19 @@  avx10_2_test (void)
   res1 = _mm_mask_getexp_pbh (res1, m8, x3);
   res1 = _mm_maskz_getexp_pbh (m8, x3);
    
-  res = _mm256_roundscalene_pbh (x1, IMM);
-  res = _mm256_mask_roundscalene_pbh (res, m16, x1, IMM);
-  res = _mm256_maskz_roundscalene_pbh (m16, x1, IMM);
-  res1 = _mm_roundscalene_pbh (x3, IMM);
-  res1 = _mm_mask_roundscalene_pbh (res1, m8, x3, IMM);
-  res1 = _mm_maskz_roundscalene_pbh (m8, x3, IMM);
+  res = _mm256_roundscale_pbh (x1, IMM);
+  res = _mm256_mask_roundscale_pbh (res, m16, x1, IMM);
+  res = _mm256_maskz_roundscale_pbh (m16, x1, IMM);
+  res1 = _mm_roundscale_pbh (x3, IMM);
+  res1 = _mm_mask_roundscale_pbh (res1, m8, x3, IMM);
+  res1 = _mm_maskz_roundscale_pbh (m8, x3, IMM);
   
-  res = _mm256_reducene_pbh (x1, IMM);
-  res = _mm256_mask_reducene_pbh (res, m16, x1, IMM);
-  res = _mm256_maskz_reducene_pbh (m16, x1, IMM);
-  res1 = _mm_reducene_pbh (x3, IMM);
-  res1 = _mm_mask_reducene_pbh (res1, m8, x3, IMM);
-  res1 = _mm_maskz_reducene_pbh (m8, x3, IMM);
+  res = _mm256_reduce_pbh (x1, IMM);
+  res = _mm256_mask_reduce_pbh (res, m16, x1, IMM);
+  res = _mm256_maskz_reduce_pbh (m16, x1, IMM);
+  res1 = _mm_reduce_pbh (x3, IMM);
+  res1 = _mm_mask_reduce_pbh (res1, m8, x3, IMM);
+  res1 = _mm_maskz_reduce_pbh (m8, x3, IMM);
   
   res = _mm256_getmant_pbh (x1, _MM_MANT_NORM_p75_1p5, _MM_MANT_SIGN_src);
   res = _mm256_mask_getmant_pbh (res, m16, x1, _MM_MANT_NORM_p75_1p5,
diff --git a/gcc/testsuite/gcc.target/i386/avx10_2-vgetmantpbf16-2.c b/gcc/testsuite/gcc.target/i386/avx10_2-vgetmantbf16-2.c
similarity index 78%
rename from gcc/testsuite/gcc.target/i386/avx10_2-vgetmantpbf16-2.c
rename to gcc/testsuite/gcc.target/i386/avx10_2-vgetmantbf16-2.c
index 8fbb6e74027..9cdec144a2c 100644
--- a/gcc/testsuite/gcc.target/i386/avx10_2-vgetmantpbf16-2.c
+++ b/gcc/testsuite/gcc.target/i386/avx10_2-vgetmantbf16-2.c
@@ -6,11 +6,11 @@ 
 #define AVX512VL
 #define AVX512F_LEN 256
 #define AVX512F_LEN_HALF 128
-#include "avx10_2-512-vgetmantpbf16-2.c"
+#include "avx10_2-512-vgetmantbf16-2.c"
 
 #undef AVX512F_LEN
 #undef AVX512F_LEN_HALF
 
 #define AVX512F_LEN 128
 #define AVX512F_LEN_HALF 128
-#include "avx10_2-512-vgetmantpbf16-2.c" 
+#include "avx10_2-512-vgetmantbf16-2.c" 
diff --git a/gcc/testsuite/gcc.target/i386/avx10_2-vreducenepbf16-2.c b/gcc/testsuite/gcc.target/i386/avx10_2-vreducebf16-2.c
similarity index 78%
rename from gcc/testsuite/gcc.target/i386/avx10_2-vreducenepbf16-2.c
rename to gcc/testsuite/gcc.target/i386/avx10_2-vreducebf16-2.c
index 9522c83b5b8..318e430d792 100644
--- a/gcc/testsuite/gcc.target/i386/avx10_2-vreducenepbf16-2.c
+++ b/gcc/testsuite/gcc.target/i386/avx10_2-vreducebf16-2.c
@@ -6,11 +6,11 @@ 
 #define AVX512VL
 #define AVX512F_LEN 256
 #define AVX512F_LEN_HALF 128
-#include "avx10_2-512-vreducenepbf16-2.c"
+#include "avx10_2-512-vreducebf16-2.c"
 
 #undef AVX512F_LEN
 #undef AVX512F_LEN_HALF
 
 #define AVX512F_LEN 128
 #define AVX512F_LEN_HALF 128
-#include "avx10_2-512-vreducenepbf16-2.c" 
+#include "avx10_2-512-vreducebf16-2.c" 
diff --git a/gcc/testsuite/gcc.target/i386/avx10_2-vrndscalenepbf16-2.c b/gcc/testsuite/gcc.target/i386/avx10_2-vrndscalebf16-2.c
similarity index 77%
rename from gcc/testsuite/gcc.target/i386/avx10_2-vrndscalenepbf16-2.c
rename to gcc/testsuite/gcc.target/i386/avx10_2-vrndscalebf16-2.c
index e27bf994c1f..57204388e41 100644
--- a/gcc/testsuite/gcc.target/i386/avx10_2-vrndscalenepbf16-2.c
+++ b/gcc/testsuite/gcc.target/i386/avx10_2-vrndscalebf16-2.c
@@ -6,11 +6,11 @@ 
 #define AVX512VL
 #define AVX512F_LEN 256
 #define AVX512F_LEN_HALF 128
-#include "avx10_2-512-vrndscalenepbf16-2.c"
+#include "avx10_2-512-vrndscalebf16-2.c"
 
 #undef AVX512F_LEN
 #undef AVX512F_LEN_HALF
 
 #define AVX512F_LEN 128
 #define AVX512F_LEN_HALF 128
-#include "avx10_2-512-vrndscalenepbf16-2.c" 
+#include "avx10_2-512-vrndscalebf16-2.c" 
diff --git a/gcc/testsuite/gcc.target/i386/sse-13.c b/gcc/testsuite/gcc.target/i386/sse-13.c
index 830935eda27..73ed7458deb 100644
--- a/gcc/testsuite/gcc.target/i386/sse-13.c
+++ b/gcc/testsuite/gcc.target/i386/sse-13.c
@@ -1025,19 +1025,19 @@ 
 #define __builtin_ia32_vcvt2ps2phx512_mask_round(A, B, C, D, E) __builtin_ia32_vcvt2ps2phx512_mask_round(A, B, C, D, 8)
 
 /* avx10_2-512bf16intrin.h */
-#define __builtin_ia32_rndscalenepbf16512_mask(A, B, C, D) __builtin_ia32_rndscalenepbf16512_mask(A, 123, C, D)
-#define __builtin_ia32_reducenepbf16512_mask(A, B, C, D) __builtin_ia32_reducenepbf16512_mask(A, 123, C, D)
-#define __builtin_ia32_getmantpbf16512_mask(A, B, C, D) __builtin_ia32_getmantpbf16512_mask(A, 1, C, D)
+#define __builtin_ia32_rndscalebf16512_mask(A, B, C, D) __builtin_ia32_rndscalebf16512_mask(A, 123, C, D)
+#define __builtin_ia32_reducebf16512_mask(A, B, C, D) __builtin_ia32_reducebf16512_mask(A, 123, C, D)
+#define __builtin_ia32_getmantbf16512_mask(A, B, C, D) __builtin_ia32_getmantbf16512_mask(A, 1, C, D)
 #define __builtin_ia32_fpclasspbf16512_mask(A, B, C) __builtin_ia32_fpclasspbf16512_mask(A, 1, C)
 #define __builtin_ia32_cmpbf16512_mask(A, B, C, D) __builtin_ia32_cmpbf16512_mask(A, B, 1, D)
 
 /* avx10_2bf16intrin.h */
-#define __builtin_ia32_rndscalenepbf16256_mask(A, B, C, D) __builtin_ia32_rndscalenepbf16256_mask(A, 123, C, D)
-#define __builtin_ia32_rndscalenepbf16128_mask(A, B, C, D) __builtin_ia32_rndscalenepbf16128_mask(A, 123, C, D)
-#define __builtin_ia32_reducenepbf16256_mask(A, B, C, D) __builtin_ia32_reducenepbf16256_mask(A, 123, C, D)
-#define __builtin_ia32_reducenepbf16128_mask(A, B, C, D) __builtin_ia32_reducenepbf16128_mask(A, 123, C, D)
-#define __builtin_ia32_getmantpbf16256_mask(A, B, C, D) __builtin_ia32_getmantpbf16256_mask(A, 1, C, D)
-#define __builtin_ia32_getmantpbf16128_mask(A, B, C, D) __builtin_ia32_getmantpbf16128_mask(A, 1, C, D)
+#define __builtin_ia32_rndscalebf16256_mask(A, B, C, D) __builtin_ia32_rndscalebf16256_mask(A, 123, C, D)
+#define __builtin_ia32_rndscalebf16128_mask(A, B, C, D) __builtin_ia32_rndscalebf16128_mask(A, 123, C, D)
+#define __builtin_ia32_reducebf16256_mask(A, B, C, D) __builtin_ia32_reducebf16256_mask(A, 123, C, D)
+#define __builtin_ia32_reducebf16128_mask(A, B, C, D) __builtin_ia32_reducebf16128_mask(A, 123, C, D)
+#define __builtin_ia32_getmantbf16256_mask(A, B, C, D) __builtin_ia32_getmantbf16256_mask(A, 1, C, D)
+#define __builtin_ia32_getmantbf16128_mask(A, B, C, D) __builtin_ia32_getmantbf16128_mask(A, 1, C, D)
 #define __builtin_ia32_fpclasspbf16256_mask(A, B, C) __builtin_ia32_fpclasspbf16256_mask(A, 1, C)
 #define __builtin_ia32_fpclasspbf16128_mask(A, B, C) __builtin_ia32_fpclasspbf16128_mask(A, 1, C)
 #define __builtin_ia32_cmpbf16256_mask(A, B, C, D) __builtin_ia32_cmpbf16256_mask(A, B, 1, D)
diff --git a/gcc/testsuite/gcc.target/i386/sse-14.c b/gcc/testsuite/gcc.target/i386/sse-14.c
index f285ae2481f..16f03e260aa 100644
--- a/gcc/testsuite/gcc.target/i386/sse-14.c
+++ b/gcc/testsuite/gcc.target/i386/sse-14.c
@@ -1390,12 +1390,12 @@  test_2 (_mm256_cvtx_round2ps_ph, __m256h, __m256, __m256, 4)
 test_2 (_mm512_cvtx_round2ps_ph, __m512h, __m512, __m512, 4)
 
 /* avx10_2-512bf16intrin.h */
-test_1 (_mm512_roundscalene_pbh, __m512bh, __m512bh, 123)
-test_2 (_mm512_maskz_roundscalene_pbh, __m512bh, __mmask32, __m512bh, 123)
-test_3 (_mm512_mask_roundscalene_pbh, __m512bh, __m512bh, __mmask32, __m512bh, 123)
-test_1 (_mm512_reducene_pbh, __m512bh, __m512bh, 123)
-test_2 (_mm512_maskz_reducene_pbh, __m512bh, __mmask32, __m512bh, 123)
-test_3 (_mm512_mask_reducene_pbh, __m512bh, __m512bh, __mmask32, __m512bh, 123)
+test_1 (_mm512_roundscale_pbh, __m512bh, __m512bh, 123)
+test_2 (_mm512_maskz_roundscale_pbh, __m512bh, __mmask32, __m512bh, 123)
+test_3 (_mm512_mask_roundscale_pbh, __m512bh, __m512bh, __mmask32, __m512bh, 123)
+test_1 (_mm512_reduce_pbh, __m512bh, __m512bh, 123)
+test_2 (_mm512_maskz_reduce_pbh, __m512bh, __mmask32, __m512bh, 123)
+test_3 (_mm512_mask_reduce_pbh, __m512bh, __m512bh, __mmask32, __m512bh, 123)
 test_1x (_mm512_getmant_pbh, __m512bh, __m512bh, 1, 1)
 test_2x (_mm512_maskz_getmant_pbh, __m512bh, __mmask32,__m512bh, 1, 1)
 test_3x (_mm512_mask_getmant_pbh, __m512bh, __m512bh, __mmask32,__m512bh, 1, 1)
@@ -1405,18 +1405,18 @@  test_2 (_mm512_cmp_pbh_mask, __mmask32, __m512bh, __m512bh, 1)
 test_3 (_mm512_mask_cmp_pbh_mask, __mmask32, __mmask32,__m512bh, __m512bh, 1)
 
 /* avx10_2bf16intrin.h */
-test_1 (_mm256_roundscalene_pbh, __m256bh, __m256bh, 123)
-test_1 (_mm_roundscalene_pbh, __m128bh, __m128bh, 123)
-test_2 (_mm256_maskz_roundscalene_pbh, __m256bh, __mmask16, __m256bh, 123)
-test_2 (_mm_maskz_roundscalene_pbh, __m128bh, __mmask8, __m128bh, 123)
-test_3 (_mm256_mask_roundscalene_pbh, __m256bh, __m256bh, __mmask16, __m256bh, 123)
-test_3 (_mm_mask_roundscalene_pbh, __m128bh, __m128bh, __mmask8, __m128bh, 123)
-test_1 (_mm256_reducene_pbh, __m256bh, __m256bh, 123)
-test_1 (_mm_reducene_pbh, __m128bh, __m128bh, 123)
-test_2 (_mm256_maskz_reducene_pbh, __m256bh, __mmask16, __m256bh, 123)
-test_2 (_mm_maskz_reducene_pbh, __m128bh, __mmask8, __m128bh, 123)
-test_3 (_mm256_mask_reducene_pbh, __m256bh, __m256bh, __mmask16, __m256bh, 123)
-test_3 (_mm_mask_reducene_pbh, __m128bh, __m128bh, __mmask8, __m128bh, 123)
+test_1 (_mm256_roundscale_pbh, __m256bh, __m256bh, 123)
+test_1 (_mm_roundscale_pbh, __m128bh, __m128bh, 123)
+test_2 (_mm256_maskz_roundscale_pbh, __m256bh, __mmask16, __m256bh, 123)
+test_2 (_mm_maskz_roundscale_pbh, __m128bh, __mmask8, __m128bh, 123)
+test_3 (_mm256_mask_roundscale_pbh, __m256bh, __m256bh, __mmask16, __m256bh, 123)
+test_3 (_mm_mask_roundscale_pbh, __m128bh, __m128bh, __mmask8, __m128bh, 123)
+test_1 (_mm256_reduce_pbh, __m256bh, __m256bh, 123)
+test_1 (_mm_reduce_pbh, __m128bh, __m128bh, 123)
+test_2 (_mm256_maskz_reduce_pbh, __m256bh, __mmask16, __m256bh, 123)
+test_2 (_mm_maskz_reduce_pbh, __m128bh, __mmask8, __m128bh, 123)
+test_3 (_mm256_mask_reduce_pbh, __m256bh, __m256bh, __mmask16, __m256bh, 123)
+test_3 (_mm_mask_reduce_pbh, __m128bh, __m128bh, __mmask8, __m128bh, 123)
 test_1x (_mm256_getmant_pbh, __m256bh, __m256bh, 1, 1)
 test_1x (_mm_getmant_pbh, __m128bh, __m128bh, 1, 1)
 test_2x (_mm256_maskz_getmant_pbh, __m256bh, __mmask16,__m256bh, 1, 1)
diff --git a/gcc/testsuite/gcc.target/i386/sse-22.c b/gcc/testsuite/gcc.target/i386/sse-22.c
index 3f0b1c00809..4f22fee1858 100644
--- a/gcc/testsuite/gcc.target/i386/sse-22.c
+++ b/gcc/testsuite/gcc.target/i386/sse-22.c
@@ -1429,12 +1429,12 @@  test_2 (_mm256_cvtx_round2ps_ph, __m256h, __m256, __m256, 4)
 test_2 (_mm512_cvtx_round2ps_ph, __m512h, __m512, __m512, 4)
 
 /* avx10_2-512bf16intrin.h */
-test_1 (_mm512_roundscalene_pbh, __m512bh, __m512bh, 123)
-test_2 (_mm512_maskz_roundscalene_pbh, __m512bh, __mmask32, __m512bh, 123)
-test_3 (_mm512_mask_roundscalene_pbh, __m512bh, __m512bh, __mmask32, __m512bh, 123)
-test_1 (_mm512_reducene_pbh, __m512bh, __m512bh, 123)
-test_2 (_mm512_maskz_reducene_pbh, __m512bh, __mmask32, __m512bh, 123)
-test_3 (_mm512_mask_reducene_pbh, __m512bh, __m512bh, __mmask32, __m512bh, 123)
+test_1 (_mm512_roundscale_pbh, __m512bh, __m512bh, 123)
+test_2 (_mm512_maskz_roundscale_pbh, __m512bh, __mmask32, __m512bh, 123)
+test_3 (_mm512_mask_roundscale_pbh, __m512bh, __m512bh, __mmask32, __m512bh, 123)
+test_1 (_mm512_reduce_pbh, __m512bh, __m512bh, 123)
+test_2 (_mm512_maskz_reduce_pbh, __m512bh, __mmask32, __m512bh, 123)
+test_3 (_mm512_mask_reduce_pbh, __m512bh, __m512bh, __mmask32, __m512bh, 123)
 test_1x (_mm512_getmant_pbh, __m512bh, __m512bh, 1, 1)
 test_2x (_mm512_maskz_getmant_pbh, __m512bh, __mmask32,__m512bh, 1, 1)
 test_3x (_mm512_mask_getmant_pbh, __m512bh, __m512bh, __mmask32,__m512bh, 1, 1)
@@ -1444,18 +1444,18 @@  test_2 (_mm512_cmp_pbh_mask, __mmask32, __m512bh, __m512bh, 1)
 test_3 (_mm512_mask_cmp_pbh_mask, __mmask32, __mmask32,__m512bh, __m512bh, 1)
 
 /* avx10_2bf16intrin.h */
-test_1 (_mm256_roundscalene_pbh, __m256bh, __m256bh, 123)
-test_1 (_mm_roundscalene_pbh, __m128bh, __m128bh, 123)
-test_2 (_mm256_maskz_roundscalene_pbh, __m256bh, __mmask16, __m256bh, 123)
-test_2 (_mm_maskz_roundscalene_pbh, __m128bh, __mmask8, __m128bh, 123)
-test_3 (_mm256_mask_roundscalene_pbh, __m256bh, __m256bh, __mmask16, __m256bh, 123)
-test_3 (_mm_mask_roundscalene_pbh, __m128bh, __m128bh, __mmask8, __m128bh, 123)
-test_1 (_mm256_reducene_pbh, __m256bh, __m256bh, 123)
-test_1 (_mm_reducene_pbh, __m128bh, __m128bh, 123)
-test_2 (_mm256_maskz_reducene_pbh, __m256bh, __mmask16, __m256bh, 123)
-test_2 (_mm_maskz_reducene_pbh, __m128bh, __mmask8, __m128bh, 123)
-test_3 (_mm256_mask_reducene_pbh, __m256bh, __m256bh, __mmask16, __m256bh, 123)
-test_3 (_mm_mask_reducene_pbh, __m128bh, __m128bh, __mmask8, __m128bh, 123)
+test_1 (_mm256_roundscale_pbh, __m256bh, __m256bh, 123)
+test_1 (_mm_roundscale_pbh, __m128bh, __m128bh, 123)
+test_2 (_mm256_maskz_roundscale_pbh, __m256bh, __mmask16, __m256bh, 123)
+test_2 (_mm_maskz_roundscale_pbh, __m128bh, __mmask8, __m128bh, 123)
+test_3 (_mm256_mask_roundscale_pbh, __m256bh, __m256bh, __mmask16, __m256bh, 123)
+test_3 (_mm_mask_roundscale_pbh, __m128bh, __m128bh, __mmask8, __m128bh, 123)
+test_1 (_mm256_reduce_pbh, __m256bh, __m256bh, 123)
+test_1 (_mm_reduce_pbh, __m128bh, __m128bh, 123)
+test_2 (_mm256_maskz_reduce_pbh, __m256bh, __mmask16, __m256bh, 123)
+test_2 (_mm_maskz_reduce_pbh, __m128bh, __mmask8, __m128bh, 123)
+test_3 (_mm256_mask_reduce_pbh, __m256bh, __m256bh, __mmask16, __m256bh, 123)
+test_3 (_mm_mask_reduce_pbh, __m128bh, __m128bh, __mmask8, __m128bh, 123)
 test_1x (_mm256_getmant_pbh, __m256bh, __m256bh, 1, 1)
 test_1x (_mm_getmant_pbh, __m128bh, __m128bh, 1, 1)
 test_2x (_mm256_maskz_getmant_pbh, __m256bh, __mmask16,__m256bh, 1, 1)
diff --git a/gcc/testsuite/gcc.target/i386/sse-23.c b/gcc/testsuite/gcc.target/i386/sse-23.c
index 7ebc4d0c88d..428e4f5fc5a 100644
--- a/gcc/testsuite/gcc.target/i386/sse-23.c
+++ b/gcc/testsuite/gcc.target/i386/sse-23.c
@@ -999,19 +999,19 @@ 
 #define __builtin_ia32_vcvt2ps2phx512_mask_round(A, B, C, D, E) __builtin_ia32_vcvt2ps2phx512_mask_round(A, B, C, D, 8)
 
 /* avx10_2-512bf16intrin.h */
-#define __builtin_ia32_rndscalenepbf16512_mask(A, B, C, D) __builtin_ia32_rndscalenepbf16512_mask(A, 123, C, D)
-#define __builtin_ia32_reducenepbf16512_mask(A, B, C, D) __builtin_ia32_reducenepbf16512_mask(A, 123, C, D)
-#define __builtin_ia32_getmantpbf16512_mask(A, B, C, D) __builtin_ia32_getmantpbf16512_mask(A, 1, C, D)
+#define __builtin_ia32_rndscalebf16512_mask(A, B, C, D) __builtin_ia32_rndscalebf16512_mask(A, 123, C, D)
+#define __builtin_ia32_reducebf16512_mask(A, B, C, D) __builtin_ia32_reducebf16512_mask(A, 123, C, D)
+#define __builtin_ia32_getmantbf16512_mask(A, B, C, D) __builtin_ia32_getmantbf16512_mask(A, 1, C, D)
 #define __builtin_ia32_fpclasspbf16512_mask(A, B, C) __builtin_ia32_fpclasspbf16512_mask(A, 1, C)
 #define __builtin_ia32_cmpbf16512_mask(A, B, C, D) __builtin_ia32_cmpbf16512_mask(A, B, 1, D)
 
 /* avx10_2bf16intrin.h */
-#define __builtin_ia32_rndscalenepbf16256_mask(A, B, C, D) __builtin_ia32_rndscalenepbf16256_mask(A, 123, C, D)
-#define __builtin_ia32_rndscalenepbf16128_mask(A, B, C, D) __builtin_ia32_rndscalenepbf16128_mask(A, 123, C, D)
-#define __builtin_ia32_reducenepbf16256_mask(A, B, C, D) __builtin_ia32_reducenepbf16256_mask(A, 123, C, D)
-#define __builtin_ia32_reducenepbf16128_mask(A, B, C, D) __builtin_ia32_reducenepbf16128_mask(A, 123, C, D)
-#define __builtin_ia32_getmantpbf16256_mask(A, B, C, D) __builtin_ia32_getmantpbf16256_mask(A, 1, C, D)
-#define __builtin_ia32_getmantpbf16128_mask(A, B, C, D) __builtin_ia32_getmantpbf16128_mask(A, 1, C, D)
+#define __builtin_ia32_rndscalebf16256_mask(A, B, C, D) __builtin_ia32_rndscalebf16256_mask(A, 123, C, D)
+#define __builtin_ia32_rndscalebf16128_mask(A, B, C, D) __builtin_ia32_rndscalebf16128_mask(A, 123, C, D)
+#define __builtin_ia32_reducebf16256_mask(A, B, C, D) __builtin_ia32_reducebf16256_mask(A, 123, C, D)
+#define __builtin_ia32_reducebf16128_mask(A, B, C, D) __builtin_ia32_reducebf16128_mask(A, 123, C, D)
+#define __builtin_ia32_getmantbf16256_mask(A, B, C, D) __builtin_ia32_getmantbf16256_mask(A, 1, C, D)
+#define __builtin_ia32_getmantbf16128_mask(A, B, C, D) __builtin_ia32_getmantbf16128_mask(A, 1, C, D)
 #define __builtin_ia32_fpclasspbf16256_mask(A, B, C) __builtin_ia32_fpclasspbf16256_mask(A, 1, C)
 #define __builtin_ia32_fpclasspbf16128_mask(A, B, C) __builtin_ia32_fpclasspbf16128_mask(A, 1, C)
 #define __builtin_ia32_cmpbf16256_mask(A, B, C, D) __builtin_ia32_cmpbf16256_mask(A, B, 1, D)