extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_ipcvtbf16_epi16 (__m512bh __A)
+_mm512_ipcvtbf16_epi8 (__m512bh __A)
{
return
(__m512i) __builtin_ia32_cvtbf162ibs512_mask ((__v32bf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_ipcvtbf16_epi16 (__m512i __W, __mmask32 __U, __m512bh __A)
+_mm512_mask_ipcvtbf16_epi8 (__m512i __W, __mmask32 __U, __m512bh __A)
{
return (__m512i) __builtin_ia32_cvtbf162ibs512_mask ((__v32bf) __A,
(__v32hi) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_ipcvtbf16_epi16 (__mmask32 __U, __m512bh __A)
+_mm512_maskz_ipcvtbf16_epi8 (__mmask32 __U, __m512bh __A)
{
return
(__m512i) __builtin_ia32_cvtbf162ibs512_mask ((__v32bf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_ipcvtbf16_epu16 (__m512bh __A)
+_mm512_ipcvtbf16_epu8 (__m512bh __A)
{
return
(__m512i) __builtin_ia32_cvtbf162iubs512_mask ((__v32bf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_ipcvtbf16_epu16 (__m512i __W, __mmask32 __U, __m512bh __A)
+_mm512_mask_ipcvtbf16_epu8 (__m512i __W, __mmask32 __U, __m512bh __A)
{
return (__m512i) __builtin_ia32_cvtbf162iubs512_mask ((__v32bf) __A,
(__v32hi) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_ipcvtbf16_epu16 (__mmask32 __U, __m512bh __A)
+_mm512_maskz_ipcvtbf16_epu8 (__mmask32 __U, __m512bh __A)
{
return
(__m512i) __builtin_ia32_cvtbf162iubs512_mask ((__v32bf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_ipcvttbf16_epi16 (__m512bh __A)
+_mm512_ipcvttbf16_epi8 (__m512bh __A)
{
return
(__m512i) __builtin_ia32_cvttbf162ibs512_mask ((__v32bf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_ipcvttbf16_epi16 (__m512i __W, __mmask32 __U, __m512bh __A)
+_mm512_mask_ipcvttbf16_epi8 (__m512i __W, __mmask32 __U, __m512bh __A)
{
return (__m512i) __builtin_ia32_cvttbf162ibs512_mask ((__v32bf) __A,
(__v32hi) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_ipcvttbf16_epi16 (__mmask32 __U, __m512bh __A)
+_mm512_maskz_ipcvttbf16_epi8 (__mmask32 __U, __m512bh __A)
{
return
(__m512i) __builtin_ia32_cvttbf162ibs512_mask ((__v32bf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_ipcvttbf16_epu16 (__m512bh __A)
+_mm512_ipcvttbf16_epu8 (__m512bh __A)
{
return (__m512i)
__builtin_ia32_cvttbf162iubs512_mask ((__v32bf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_ipcvttbf16_epu16 (__m512i __W, __mmask32 __U, __m512bh __A)
+_mm512_mask_ipcvttbf16_epu8 (__m512i __W, __mmask32 __U, __m512bh __A)
{
return (__m512i) __builtin_ia32_cvttbf162iubs512_mask ((__v32bf) __A,
(__v32hi) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_ipcvttbf16_epu16 (__mmask32 __U, __m512bh __A)
+_mm512_maskz_ipcvttbf16_epu8 (__mmask32 __U, __m512bh __A)
{
return (__m512i)
__builtin_ia32_cvttbf162iubs512_mask ((__v32bf) __A,
#ifdef __OPTIMIZE__
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_ipcvt_roundph_epi16 (__m512h __A, const int __R)
+_mm512_ipcvt_roundph_epi8 (__m512h __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtph2ibs512_mask_round ((__v32hf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_ipcvt_roundph_epi16 (__m512i __W, __mmask32 __U, __m512h __A,
- const int __R)
+_mm512_mask_ipcvt_roundph_epi8 (__m512i __W, __mmask32 __U, __m512h __A,
+ const int __R)
{
return (__m512i) __builtin_ia32_cvtph2ibs512_mask_round ((__v32hf) __A,
(__v32hi) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_ipcvt_roundph_epi16 (__mmask32 __U, __m512h __A, const int __R)
+_mm512_maskz_ipcvt_roundph_epi8 (__mmask32 __U, __m512h __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtph2ibs512_mask_round ((__v32hf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_ipcvt_roundph_epu16 (__m512h __A, const int __R)
+_mm512_ipcvt_roundph_epu8 (__m512h __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtph2iubs512_mask_round ((__v32hf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_ipcvt_roundph_epu16 (__m512i __W, __mmask32 __U, __m512h __A,
- const int __R)
+_mm512_mask_ipcvt_roundph_epu8 (__m512i __W, __mmask32 __U, __m512h __A,
+ const int __R)
{
return (__m512i) __builtin_ia32_cvtph2iubs512_mask_round ((__v32hf) __A,
(__v32hi) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_ipcvt_roundph_epu16 (__mmask32 __U, __m512h __A, const int __R)
+_mm512_maskz_ipcvt_roundph_epu8 (__mmask32 __U, __m512h __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtph2iubs512_mask_round ((__v32hf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_ipcvt_roundps_epi32 (__m512 __A, const int __R)
+_mm512_ipcvt_roundps_epi8 (__m512 __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtps2ibs512_mask_round ((__v16sf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_ipcvt_roundps_epi32 (__m512i __W, __mmask16 __U, __m512 __A,
- const int __R)
+_mm512_mask_ipcvt_roundps_epi8 (__m512i __W, __mmask16 __U, __m512 __A,
+ const int __R)
{
return (__m512i) __builtin_ia32_cvtps2ibs512_mask_round ((__v16sf) __A,
(__v16si) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_ipcvt_roundps_epi32 (__mmask16 __U, __m512 __A, const int __R)
+_mm512_maskz_ipcvt_roundps_epi8 (__mmask16 __U, __m512 __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtps2ibs512_mask_round ((__v16sf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_ipcvt_roundps_epu32 (__m512 __A, const int __R)
+_mm512_ipcvt_roundps_epu8 (__m512 __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtps2iubs512_mask_round ((__v16sf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_ipcvt_roundps_epu32 (__m512i __W, __mmask16 __U, __m512 __A,
- const int __R)
+_mm512_mask_ipcvt_roundps_epu8 (__m512i __W, __mmask16 __U, __m512 __A,
+ const int __R)
{
return (__m512i) __builtin_ia32_cvtps2iubs512_mask_round ((__v16sf) __A,
(__v16si) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_ipcvt_roundps_epu32 (__mmask16 __U, __m512 __A, const int __R)
+_mm512_maskz_ipcvt_roundps_epu8 (__mmask16 __U, __m512 __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtps2iubs512_mask_round ((__v16sf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_ipcvtt_roundph_epi16 (__m512h __A, const int __R)
+_mm512_ipcvtt_roundph_epi8 (__m512h __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttph2ibs512_mask_round ((__v32hf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_ipcvtt_roundph_epi16 (__m512i __W, __mmask32 __U, __m512h __A,
- const int __R)
+_mm512_mask_ipcvtt_roundph_epi8 (__m512i __W, __mmask32 __U, __m512h __A,
+ const int __R)
{
return (__m512i) __builtin_ia32_cvttph2ibs512_mask_round ((__v32hf) __A,
(__v32hi) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_ipcvtt_roundph_epi16 (__mmask32 __U, __m512h __A, const int __R)
+_mm512_maskz_ipcvtt_roundph_epi8 (__mmask32 __U, __m512h __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvttph2ibs512_mask_round ((__v32hf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_ipcvtt_roundph_epu16 (__m512h __A, const int __R)
+_mm512_ipcvtt_roundph_epu8 (__m512h __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttph2iubs512_mask_round ((__v32hf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_ipcvtt_roundph_epu16 (__m512i __W, __mmask32 __U, __m512h __A,
- const int __R)
+_mm512_mask_ipcvtt_roundph_epu8 (__m512i __W, __mmask32 __U, __m512h __A,
+ const int __R)
{
return (__m512i) __builtin_ia32_cvttph2iubs512_mask_round ((__v32hf) __A,
(__v32hi) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_ipcvtt_roundph_epu16 (__mmask32 __U, __m512h __A, const int __R)
+_mm512_maskz_ipcvtt_roundph_epu8 (__mmask32 __U, __m512h __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttph2iubs512_mask_round ((__v32hf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_ipcvtt_roundps_epi32 (__m512 __A, const int __R)
+_mm512_ipcvtt_roundps_epi8 (__m512 __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttps2ibs512_mask_round ((__v16sf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_ipcvtt_roundps_epi32 (__m512i __W, __mmask16 __U, __m512 __A,
- const int __R)
+_mm512_mask_ipcvtt_roundps_epi8 (__m512i __W, __mmask16 __U, __m512 __A,
+ const int __R)
{
return (__m512i) __builtin_ia32_cvttps2ibs512_mask_round ((__v16sf) __A,
(__v16si) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_ipcvtt_roundps_epi32 (__mmask16 __U, __m512 __A, const int __R)
+_mm512_maskz_ipcvtt_roundps_epi8 (__mmask16 __U, __m512 __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttps2ibs512_mask_round ((__v16sf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_ipcvtt_roundps_epu32 (__m512 __A, const int __R)
+_mm512_ipcvtt_roundps_epu8 (__m512 __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttps2iubs512_mask_round ((__v16sf) __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_ipcvtt_roundps_epu32 (__m512i __W, __mmask16 __U, __m512 __A,
- const int __R)
+_mm512_mask_ipcvtt_roundps_epu8 (__m512i __W, __mmask16 __U, __m512 __A,
+ const int __R)
{
return (__m512i) __builtin_ia32_cvttps2iubs512_mask_round ((__v16sf) __A,
(__v16si) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_ipcvtt_roundps_epu32 (__mmask16 __U, __m512 __A, const int __R)
+_mm512_maskz_ipcvtt_roundps_epu8 (__mmask16 __U, __m512 __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttps2iubs512_mask_round ((__v16sf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvtts_roundpd_epi32 (__m256i __W, __mmask8 __U, __m512d __A,
- const int __R)
+ const int __R)
{
return (__m256i) __builtin_ia32_cvttpd2dqs512_mask_round ((__v8df) __A,
(__v8si) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvtts_roundpd_epi64 (__m512i __W, __mmask8 __U, __m512d __A,
- const int __R)
+ const int __R)
{
return (__m512i) __builtin_ia32_cvttpd2qqs512_mask_round ((__v8df) __A,
(__v8di) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvtts_roundpd_epu32 (__m256i __W, __mmask8 __U, __m512d __A,
- const int __R)
+ const int __R)
{
return (__m256i) __builtin_ia32_cvttpd2udqs512_mask_round ((__v8df) __A,
(__v8si) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvtts_roundpd_epu64 (__m512i __W, __mmask8 __U, __m512d __A,
- const int __R)
+ const int __R)
{
return (__m512i) __builtin_ia32_cvttpd2uqqs512_mask_round ((__v8df) __A,
(__v8di) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvtts_roundps_epi32 (__m512i __W, __mmask16 __U, __m512 __A,
- const int __R)
+ const int __R)
{
return (__m512i) __builtin_ia32_cvttps2dqs512_mask_round ((__v16sf) __A,
(__v16si) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvtts_roundps_epi64 (__m512i __W, __mmask8 __U, __m256 __A,
- const int __R)
+ const int __R)
{
return (__m512i) __builtin_ia32_cvttps2qqs512_mask_round ((__v8sf) __A,
(__v8di) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvtts_roundps_epu32 (__m512i __W, __mmask16 __U, __m512 __A,
- const int __R)
+ const int __R)
{
return (__m512i) __builtin_ia32_cvttps2udqs512_mask_round ((__v16sf) __A,
(__v16si) __W,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvtts_roundps_epu64 (__m512i __W, __mmask8 __U, __m256 __A,
- const int __R)
+ const int __R)
{
return (__m512i) __builtin_ia32_cvttps2uqqs512_mask_round ((__v8sf) __A,
(__v8di) __W,
__R);
}
#else
-#define _mm512_ipcvt_roundph_epi16(A, R) \
+#define _mm512_ipcvt_roundph_epi8(A, R) \
((__m512i) \
__builtin_ia32_cvtph2ibs512_mask_round ((__v32hf) (A), \
(__v32hi) \
(__mmask32) (-1), \
(R)))
-#define _mm512_mask_ipcvt_roundph_epi16(W, U, A, R) \
+#define _mm512_mask_ipcvt_roundph_epi8(W, U, A, R) \
((__m512i) __builtin_ia32_cvtph2ibs512_mask_round ((__v32hf) (A), \
(__v32hi) (W), \
(__mmask32) (U), \
(R)))
-#define _mm512_maskz_ipcvt_roundph_epi16(U, A, R) \
+#define _mm512_maskz_ipcvt_roundph_epi8(U, A, R) \
((__m512i) \
__builtin_ia32_cvtph2ibs512_mask_round ((__v32hf) (A), \
(__v32hi) \
(__mmask32) (U), \
(R)))
-#define _mm512_ipcvt_roundph_epu16(A, R) \
+#define _mm512_ipcvt_roundph_epu8(A, R) \
((__m512i) \
__builtin_ia32_cvtph2iubs512_mask_round ((__v32hf) (A), \
(__v32hi) \
(__mmask32) (-1), \
(R)))
-#define _mm512_mask_ipcvt_roundph_epu16(W, U, A, R) \
+#define _mm512_mask_ipcvt_roundph_epu8(W, U, A, R) \
((__m512i) __builtin_ia32_cvtph2iubs512_mask_round ((__v32hf) (A), \
(__v32hi) (W), \
(__mmask32) (U), \
(R)))
-#define _mm512_maskz_ipcvt_roundph_epu16(U, A, R) \
+#define _mm512_maskz_ipcvt_roundph_epu8(U, A, R) \
((__m512i) \
__builtin_ia32_cvtph2iubs512_mask_round ((__v32hf) (A), \
(__v32hi) \
(__mmask32) (U), \
(R)))
-#define _mm512_ipcvt_roundps_epi32(A, R) \
+#define _mm512_ipcvt_roundps_epi8(A, R) \
((__m512i) \
__builtin_ia32_cvtps2ibs512_mask_round ((__v16sf) (A), \
(__v16si) \
(__mmask16) (-1), \
(R)))
-#define _mm512_mask_ipcvt_roundps_epi32(W, U, A, R) \
+#define _mm512_mask_ipcvt_roundps_epi8(W, U, A, R) \
((__m512i) __builtin_ia32_cvtps2ibs512_mask_round ((__v16sf) (A), \
(__v16si) (W), \
(__mmask16) (U), \
(R)))
-#define _mm512_maskz_ipcvt_roundps_epi32(U, A, R) \
+#define _mm512_maskz_ipcvt_roundps_epi8(U, A, R) \
((__m512i) \
__builtin_ia32_cvtps2ibs512_mask_round ((__v16sf) (A), \
(__v16si) \
(__mmask16) (U), \
(R)))
-#define _mm512_ipcvt_roundps_epu32(A, R) \
+#define _mm512_ipcvt_roundps_epu8(A, R) \
((__m512i) \
__builtin_ia32_cvtps2iubs512_mask_round ((__v16sf) (A), \
(__v16si) \
(__mmask16) (-1), \
(R)))
-#define _mm512_mask_ipcvt_roundps_epu32(W, U, A, R) \
+#define _mm512_mask_ipcvt_roundps_epu8(W, U, A, R) \
((__m512i) __builtin_ia32_cvtps2iubs512_mask_round ((__v16sf) (A), \
(__v16si) (W), \
(__mmask16) (U), \
(R)))
-#define _mm512_maskz_ipcvt_roundps_epu32(U, A, R) \
+#define _mm512_maskz_ipcvt_roundps_epu8(U, A, R) \
((__m512i) \
__builtin_ia32_cvtps2iubs512_mask_round ((__v16sf) (A), \
(__v16si) \
(__mmask16) (U), \
(R)))
-#define _mm512_ipcvtt_roundph_epi16(A, R) \
+#define _mm512_ipcvtt_roundph_epi8(A, R) \
((__m512i) \
__builtin_ia32_cvttph2ibs512_mask_round ((__v32hf) (A), \
(__v32hi) \
(__mmask32) (-1), \
(R)))
-#define _mm512_mask_ipcvtt_roundph_epi16(W, U, A, R) \
+#define _mm512_mask_ipcvtt_roundph_epi8(W, U, A, R) \
((__m512i) __builtin_ia32_cvttph2ibs512_mask_round ((__v32hf) (A), \
(__v32hi) (W), \
(__mmask32) (U), \
(R)))
-#define _mm512_maskz_ipcvtt_roundph_epi16(U, A, R) \
+#define _mm512_maskz_ipcvtt_roundph_epi8(U, A, R) \
((__m512i) \
__builtin_ia32_cvttph2ibs512_mask_round ((__v32hf) (A), \
(__v32hi) \
(__mmask32) (U), \
(R)))
-#define _mm512_ipcvtt_roundph_epu16(A, R) \
+#define _mm512_ipcvtt_roundph_epu8(A, R) \
((__m512i) \
__builtin_ia32_cvttph2iubs512_mask_round ((__v32hf) (A), \
(__v32hi) \
(__mmask32) (-1), \
(R)))
-#define _mm512_mask_ipcvtt_roundph_epu16(W, U, A, R) \
+#define _mm512_mask_ipcvtt_roundph_epu8(W, U, A, R) \
((__m512i) __builtin_ia32_cvttph2iubs512_mask_round ((__v32hf) (A), \
(__v32hi) (W), \
(__mmask32) (U), \
(R)))
-#define _mm512_maskz_ipcvtt_roundph_epu16(U, A, R) \
+#define _mm512_maskz_ipcvtt_roundph_epu8(U, A, R) \
((__m512i) \
__builtin_ia32_cvttph2iubs512_mask_round ((__v32hf) (A), \
(__v32hi) \
(__mmask32) (U), \
(R)))
-#define _mm512_ipcvtt_roundps_epi32(A, R) \
+#define _mm512_ipcvtt_roundps_epi8(A, R) \
((__m512i) \
__builtin_ia32_cvttps2ibs512_mask_round ((__v16sf) (A), \
(__v16si) \
(__mmask16) (-1), \
(R)))
-#define _mm512_mask_ipcvtt_roundps_epi32(W, U, A, R) \
+#define _mm512_mask_ipcvtt_roundps_epi8(W, U, A, R) \
((__m512i) __builtin_ia32_cvttps2ibs512_mask_round ((__v16sf) (A), \
(__v16si) (W), \
(__mmask16) (U), \
(R)))
-#define _mm512_maskz_ipcvtt_roundps_epi32(U, A, R) \
+#define _mm512_maskz_ipcvtt_roundps_epi8(U, A, R) \
((__m512i) \
__builtin_ia32_cvttps2ibs512_mask_round ((__v16sf) (A), \
(__v16si) \
(__mmask16) (U), \
(R)))
-#define _mm512_ipcvtt_roundps_epu32(A, R) \
+#define _mm512_ipcvtt_roundps_epu8(A, R) \
((__m512i) \
__builtin_ia32_cvttps2iubs512_mask_round ((__v16sf) (A), \
(__v16si) \
(__mmask16) (-1), \
(R)))
-#define _mm512_mask_ipcvtt_roundps_epu32(W, U, A, R) \
+#define _mm512_mask_ipcvtt_roundps_epu8(W, U, A, R) \
((__m512i) __builtin_ia32_cvttps2iubs512_mask_round ((__v16sf) (A), \
(__v16si) (W), \
(__mmask16) (U), \
(R)))
-#define _mm512_maskz_ipcvtt_roundps_epu32(U, A, R) \
+#define _mm512_maskz_ipcvtt_roundps_epu8(U, A, R) \
((__m512i) \
__builtin_ia32_cvttps2iubs512_mask_round ((__v16sf) (A), \
(__v16si) \
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_ipcvtbf16_epi16 (__m128bh __A)
+_mm_ipcvtbf16_epi8 (__m128bh __A)
{
return (__m128i) __builtin_ia32_cvtbf162ibs128_mask ((__v8bf) __A,
(__v8hi)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_ipcvtbf16_epi16 (__m128i __W, __mmask8 __U, __m128bh __A)
+_mm_mask_ipcvtbf16_epi8 (__m128i __W, __mmask8 __U, __m128bh __A)
{
return (__m128i) __builtin_ia32_cvtbf162ibs128_mask ((__v8bf) __A,
(__v8hi) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_ipcvtbf16_epi16 (__mmask8 __U, __m128bh __A)
+_mm_maskz_ipcvtbf16_epi8 (__mmask8 __U, __m128bh __A)
{
return (__m128i) __builtin_ia32_cvtbf162ibs128_mask ((__v8bf) __A,
(__v8hi)
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_ipcvtbf16_epi16 (__m256bh __A)
+_mm256_ipcvtbf16_epi8 (__m256bh __A)
{
return
(__m256i) __builtin_ia32_cvtbf162ibs256_mask ((__v16bf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_ipcvtbf16_epi16 (__m256i __W, __mmask16 __U, __m256bh __A)
+_mm256_mask_ipcvtbf16_epi8 (__m256i __W, __mmask16 __U, __m256bh __A)
{
return (__m256i) __builtin_ia32_cvtbf162ibs256_mask ((__v16bf) __A,
(__v16hi) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_ipcvtbf16_epi16 (__mmask16 __U, __m256bh __A)
+_mm256_maskz_ipcvtbf16_epi8 (__mmask16 __U, __m256bh __A)
{
return
(__m256i) __builtin_ia32_cvtbf162ibs256_mask ((__v16bf) __A,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_ipcvtbf16_epu16 (__m128bh __A)
+_mm_ipcvtbf16_epu8 (__m128bh __A)
{
return
(__m128i) __builtin_ia32_cvtbf162iubs128_mask ((__v8bf) __A,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_ipcvtbf16_epu16 (__m128i __W, __mmask8 __U, __m128bh __A)
+_mm_mask_ipcvtbf16_epu8 (__m128i __W, __mmask8 __U, __m128bh __A)
{
return (__m128i) __builtin_ia32_cvtbf162iubs128_mask ((__v8bf) __A,
(__v8hi) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_ipcvtbf16_epu16 (__mmask8 __U, __m128bh __A)
+_mm_maskz_ipcvtbf16_epu8 (__mmask8 __U, __m128bh __A)
{
return
(__m128i) __builtin_ia32_cvtbf162iubs128_mask ((__v8bf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_ipcvtbf16_epu16 (__m256bh __A)
+_mm256_ipcvtbf16_epu8 (__m256bh __A)
{
return
(__m256i) __builtin_ia32_cvtbf162iubs256_mask ((__v16bf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_ipcvtbf16_epu16 (__m256i __W, __mmask16 __U, __m256bh __A)
+_mm256_mask_ipcvtbf16_epu8 (__m256i __W, __mmask16 __U, __m256bh __A)
{
return (__m256i) __builtin_ia32_cvtbf162iubs256_mask ((__v16bf) __A,
(__v16hi) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_ipcvtbf16_epu16 (__mmask16 __U, __m256bh __A)
+_mm256_maskz_ipcvtbf16_epu8 (__mmask16 __U, __m256bh __A)
{
return
(__m256i) __builtin_ia32_cvtbf162iubs256_mask ((__v16bf) __A,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_ipcvtph_epi16 (__m128h __A)
+_mm_ipcvtph_epi8 (__m128h __A)
{
return (__m128i) __builtin_ia32_cvtph2ibs128_mask ((__v8hf) __A,
(__v8hi)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_ipcvtph_epi16 (__m128i __W, __mmask8 __U, __m128h __A)
+_mm_mask_ipcvtph_epi8 (__m128i __W, __mmask8 __U, __m128h __A)
{
return (__m128i) __builtin_ia32_cvtph2ibs128_mask ((__v8hf) __A,
(__v8hi) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_ipcvtph_epi16 (__mmask8 __U, __m128h __A)
+_mm_maskz_ipcvtph_epi8 (__mmask8 __U, __m128h __A)
{
return (__m128i) __builtin_ia32_cvtph2ibs128_mask ((__v8hf) __A,
(__v8hi)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_ipcvtph_epu16 (__m128h __A)
+_mm_ipcvtph_epu8 (__m128h __A)
{
return (__m128i) __builtin_ia32_cvtph2iubs128_mask ((__v8hf) __A,
(__v8hi)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_ipcvtph_epu16 (__m128i __W, __mmask8 __U, __m128h __A)
+_mm_mask_ipcvtph_epu8 (__m128i __W, __mmask8 __U, __m128h __A)
{
return (__m128i) __builtin_ia32_cvtph2iubs128_mask ((__v8hf) __A,
(__v8hi) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_ipcvtph_epu16 (__mmask8 __U, __m128h __A)
+_mm_maskz_ipcvtph_epu8 (__mmask8 __U, __m128h __A)
{
return (__m128i) __builtin_ia32_cvtph2iubs128_mask ((__v8hf) __A,
(__v8hi)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_ipcvtps_epi32 (__m128 __A)
+_mm_ipcvtps_epi8 (__m128 __A)
{
return (__m128i) __builtin_ia32_cvtps2ibs128_mask ((__v4sf) __A,
(__v4si)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_ipcvtps_epi32 (__m128i __W, __mmask8 __U, __m128 __A)
+_mm_mask_ipcvtps_epi8 (__m128i __W, __mmask8 __U, __m128 __A)
{
return (__m128i) __builtin_ia32_cvtps2ibs128_mask ((__v4sf) __A,
(__v4si) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_ipcvtps_epi32 (__mmask8 __U, __m128 __A)
+_mm_maskz_ipcvtps_epi8 (__mmask8 __U, __m128 __A)
{
return (__m128i) __builtin_ia32_cvtps2ibs128_mask ((__v4sf) __A,
(__v4si)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_ipcvtps_epu32 (__m128 __A)
+_mm_ipcvtps_epu8 (__m128 __A)
{
return (__m128i) __builtin_ia32_cvtps2iubs128_mask ((__v4sf) __A,
(__v4si)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_ipcvtps_epu32 (__m128i __W, __mmask8 __U, __m128 __A)
+_mm_mask_ipcvtps_epu8 (__m128i __W, __mmask8 __U, __m128 __A)
{
return (__m128i) __builtin_ia32_cvtps2iubs128_mask ((__v4sf) __A,
(__v4si) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_ipcvtps_epu32 (__mmask8 __U, __m128 __A)
+_mm_maskz_ipcvtps_epu8 (__mmask8 __U, __m128 __A)
{
return (__m128i) __builtin_ia32_cvtps2iubs128_mask ((__v4sf) __A,
(__v4si)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_ipcvttbf16_epi16 (__m128bh __A)
+_mm_ipcvttbf16_epi8 (__m128bh __A)
{
return
(__m128i) __builtin_ia32_cvttbf162ibs128_mask ((__v8bf) __A,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_ipcvttbf16_epi16 (__m128i __W, __mmask8 __U, __m128bh __A)
+_mm_mask_ipcvttbf16_epi8 (__m128i __W, __mmask8 __U, __m128bh __A)
{
return (__m128i) __builtin_ia32_cvttbf162ibs128_mask ((__v8bf) __A,
(__v8hi) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_ipcvttbf16_epi16 (__mmask8 __U, __m128bh __A)
+_mm_maskz_ipcvttbf16_epi8 (__mmask8 __U, __m128bh __A)
{
return (__m128i) __builtin_ia32_cvttbf162ibs128_mask ((__v8bf) __A,
(__v8hi)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_ipcvttbf16_epu16 (__m128bh __A)
+_mm_ipcvttbf16_epu8 (__m128bh __A)
{
return
(__m128i) __builtin_ia32_cvttbf162iubs128_mask ((__v8bf) __A,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_ipcvttbf16_epu16 (__m128i __W, __mmask8 __U, __m128bh __A)
+_mm_mask_ipcvttbf16_epu8 (__m128i __W, __mmask8 __U, __m128bh __A)
{
return (__m128i) __builtin_ia32_cvttbf162iubs128_mask ((__v8bf) __A,
(__v8hi) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_ipcvttbf16_epu16 (__mmask8 __U, __m128bh __A)
+_mm_maskz_ipcvttbf16_epu8 (__mmask8 __U, __m128bh __A)
{
return (__m128i) __builtin_ia32_cvttbf162iubs128_mask ((__v8bf) __A,
(__v8hi)
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_ipcvttbf16_epi16 (__m256bh __A)
+_mm256_ipcvttbf16_epi8 (__m256bh __A)
{
return (__m256i)
__builtin_ia32_cvttbf162ibs256_mask ((__v16bf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_ipcvttbf16_epi16 (__m256i __W, __mmask16 __U, __m256bh __A)
+_mm256_mask_ipcvttbf16_epi8 (__m256i __W, __mmask16 __U, __m256bh __A)
{
return (__m256i) __builtin_ia32_cvttbf162ibs256_mask ((__v16bf) __A,
(__v16hi) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_ipcvttbf16_epi16 (__mmask16 __U, __m256bh __A)
+_mm256_maskz_ipcvttbf16_epi8 (__mmask16 __U, __m256bh __A)
{
return (__m256i)
__builtin_ia32_cvttbf162ibs256_mask ((__v16bf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_ipcvttbf16_epu16 (__m256bh __A)
+_mm256_ipcvttbf16_epu8 (__m256bh __A)
{
return (__m256i)
__builtin_ia32_cvttbf162iubs256_mask ((__v16bf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_ipcvttbf16_epu16 (__m256i __W, __mmask16 __U, __m256bh __A)
+_mm256_mask_ipcvttbf16_epu8 (__m256i __W, __mmask16 __U, __m256bh __A)
{
return (__m256i) __builtin_ia32_cvttbf162iubs256_mask ((__v16bf) __A,
(__v16hi) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_ipcvttbf16_epu16 (__mmask16 __U, __m256bh __A)
+_mm256_maskz_ipcvttbf16_epu8 (__mmask16 __U, __m256bh __A)
{
return (__m256i)
__builtin_ia32_cvttbf162iubs256_mask ((__v16bf) __A,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_ipcvttph_epi16 (__m128h __A)
+_mm_ipcvttph_epi8 (__m128h __A)
{
return (__m128i) __builtin_ia32_cvttph2ibs128_mask ((__v8hf) __A,
(__v8hi)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_ipcvttph_epi16 (__m128i __W, __mmask8 __U, __m128h __A)
+_mm_mask_ipcvttph_epi8 (__m128i __W, __mmask8 __U, __m128h __A)
{
return (__m128i) __builtin_ia32_cvttph2ibs128_mask ((__v8hf) __A,
(__v8hi) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_ipcvttph_epi16 (__mmask8 __U, __m128h __A)
+_mm_maskz_ipcvttph_epi8 (__mmask8 __U, __m128h __A)
{
return (__m128i) __builtin_ia32_cvttph2ibs128_mask ((__v8hf) __A,
(__v8hi)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_ipcvttph_epu16 (__m128h __A)
+_mm_ipcvttph_epu8 (__m128h __A)
{
return (__m128i) __builtin_ia32_cvttph2iubs128_mask ((__v8hf) __A,
(__v8hi)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_ipcvttph_epu16 (__m128i __W, __mmask8 __U, __m128h __A)
+_mm_mask_ipcvttph_epu8 (__m128i __W, __mmask8 __U, __m128h __A)
{
return (__m128i) __builtin_ia32_cvttph2iubs128_mask ((__v8hf) __A,
(__v8hi) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_ipcvttph_epu16 (__mmask8 __U, __m128h __A)
+_mm_maskz_ipcvttph_epu8 (__mmask8 __U, __m128h __A)
{
return (__m128i) __builtin_ia32_cvttph2iubs128_mask ((__v8hf) __A,
(__v8hi)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_ipcvttps_epi32 (__m128 __A)
+_mm_ipcvttps_epi8 (__m128 __A)
{
return (__m128i) __builtin_ia32_cvttps2ibs128_mask ((__v4sf) __A,
(__v4si)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_ipcvttps_epi32 (__m128i __W, __mmask8 __U, __m128 __A)
+_mm_mask_ipcvttps_epi8 (__m128i __W, __mmask8 __U, __m128 __A)
{
return (__m128i) __builtin_ia32_cvttps2ibs128_mask ((__v4sf) __A,
(__v4si) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_ipcvttps_epi32 (__mmask8 __U, __m128 __A)
+_mm_maskz_ipcvttps_epi8 (__mmask8 __U, __m128 __A)
{
return (__m128i) __builtin_ia32_cvttps2ibs128_mask ((__v4sf) __A,
(__v4si)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_ipcvttps_epu32 (__m128 __A)
+_mm_ipcvttps_epu8 (__m128 __A)
{
return (__m128i) __builtin_ia32_cvttps2iubs128_mask ((__v4sf) __A,
(__v4si)
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_ipcvttps_epu32 (__m128i __W, __mmask8 __U, __m128 __A)
+_mm_mask_ipcvttps_epu8 (__m128i __W, __mmask8 __U, __m128 __A)
{
return (__m128i) __builtin_ia32_cvttps2iubs128_mask ((__v4sf) __A,
(__v4si) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_ipcvttps_epu32 (__mmask8 __U, __m128 __A)
+_mm_maskz_ipcvttps_epu8 (__mmask8 __U, __m128 __A)
{
return (__m128i) __builtin_ia32_cvttps2iubs128_mask ((__v4sf) __A,
(__v4si)
#ifdef __OPTIMIZE__
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_ipcvt_roundph_epi16 (__m256h __A, const int __R)
+_mm256_ipcvt_roundph_epi8 (__m256h __A, const int __R)
{
return
(__m256i) __builtin_ia32_cvtph2ibs256_mask_round ((__v16hf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_ipcvt_roundph_epi16 (__m256i __W, __mmask16 __U, __m256h __A,
- const int __R)
+_mm256_mask_ipcvt_roundph_epi8 (__m256i __W, __mmask16 __U, __m256h __A,
+ const int __R)
{
return (__m256i) __builtin_ia32_cvtph2ibs256_mask_round ((__v16hf) __A,
(__v16hi) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_ipcvt_roundph_epi16 (__mmask16 __U, __m256h __A, const int __R)
+_mm256_maskz_ipcvt_roundph_epi8 (__mmask16 __U, __m256h __A, const int __R)
{
return
(__m256i) __builtin_ia32_cvtph2ibs256_mask_round ((__v16hf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_ipcvt_roundph_epu16 (__m256h __A, const int __R)
+_mm256_ipcvt_roundph_epu8 (__m256h __A, const int __R)
{
return (__m256i)
__builtin_ia32_cvtph2iubs256_mask_round ((__v16hf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_ipcvt_roundph_epu16 (__m256i __W, __mmask16 __U, __m256h __A,
- const int __R)
+_mm256_mask_ipcvt_roundph_epu8 (__m256i __W, __mmask16 __U, __m256h __A,
+ const int __R)
{
return (__m256i) __builtin_ia32_cvtph2iubs256_mask_round ((__v16hf) __A,
(__v16hi) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_ipcvt_roundph_epu16 (__mmask16 __U, __m256h __A, const int __R)
+_mm256_maskz_ipcvt_roundph_epu8 (__mmask16 __U, __m256h __A, const int __R)
{
return
(__m256i) __builtin_ia32_cvtph2iubs256_mask_round ((__v16hf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_ipcvt_roundps_epi32 (__m256 __A, const int __R)
+_mm256_ipcvt_roundps_epi8 (__m256 __A, const int __R)
{
return
(__m256i) __builtin_ia32_cvtps2ibs256_mask_round ((__v8sf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_ipcvt_roundps_epi32 (__m256i __W, __mmask8 __U, __m256 __A,
- const int __R)
+_mm256_mask_ipcvt_roundps_epi8 (__m256i __W, __mmask8 __U, __m256 __A,
+ const int __R)
{
return (__m256i) __builtin_ia32_cvtps2ibs256_mask_round ((__v8sf) __A,
(__v8si) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_ipcvt_roundps_epi32 (__mmask8 __U, __m256 __A, const int __R)
+_mm256_maskz_ipcvt_roundps_epi8 (__mmask8 __U, __m256 __A, const int __R)
{
return
(__m256i) __builtin_ia32_cvtps2ibs256_mask_round ((__v8sf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_ipcvt_roundps_epu32 (__m256 __A, const int __R)
+_mm256_ipcvt_roundps_epu8 (__m256 __A, const int __R)
{
return (__m256i)
__builtin_ia32_cvtps2iubs256_mask_round ((__v8sf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_ipcvt_roundps_epu32 (__m256i __W, __mmask8 __U, __m256 __A,
- const int __R)
+_mm256_mask_ipcvt_roundps_epu8 (__m256i __W, __mmask8 __U, __m256 __A,
+ const int __R)
{
return (__m256i) __builtin_ia32_cvtps2iubs256_mask_round ((__v8sf) __A,
(__v8si) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_ipcvt_roundps_epu32 (__mmask8 __U, __m256 __A, const int __R)
+_mm256_maskz_ipcvt_roundps_epu8 (__mmask8 __U, __m256 __A, const int __R)
{
return
(__m256i) __builtin_ia32_cvtps2iubs256_mask_round ((__v8sf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_ipcvtt_roundph_epi16 (__m256h __A, const int __R)
+_mm256_ipcvtt_roundph_epi8 (__m256h __A, const int __R)
{
return (__m256i)
__builtin_ia32_cvttph2ibs256_mask_round ((__v16hf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_ipcvtt_roundph_epi16 (__m256i __W, __mmask16 __U, __m256h __A,
- const int __R)
+_mm256_mask_ipcvtt_roundph_epi8 (__m256i __W, __mmask16 __U, __m256h __A,
+ const int __R)
{
return (__m256i) __builtin_ia32_cvttph2ibs256_mask_round ((__v16hf) __A,
(__v16hi) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_ipcvtt_roundph_epi16 (__mmask16 __U, __m256h __A, const int __R)
+_mm256_maskz_ipcvtt_roundph_epi8 (__mmask16 __U, __m256h __A, const int __R)
{
return
(__m256i) __builtin_ia32_cvttph2ibs256_mask_round ((__v16hf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_ipcvtt_roundph_epu16 (__m256h __A, const int __R)
+_mm256_ipcvtt_roundph_epu8 (__m256h __A, const int __R)
{
return (__m256i)
__builtin_ia32_cvttph2iubs256_mask_round ((__v16hf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_ipcvtt_roundph_epu16 (__m256i __W, __mmask16 __U, __m256h __A,
- const int __R)
+_mm256_mask_ipcvtt_roundph_epu8 (__m256i __W, __mmask16 __U, __m256h __A,
+ const int __R)
{
return (__m256i) __builtin_ia32_cvttph2iubs256_mask_round ((__v16hf) __A,
(__v16hi) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_ipcvtt_roundph_epu16 (__mmask16 __U, __m256h __A, const int __R)
+_mm256_maskz_ipcvtt_roundph_epu8 (__mmask16 __U, __m256h __A, const int __R)
{
return
(__m256i) __builtin_ia32_cvttph2iubs256_mask_round ((__v16hf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_ipcvtt_roundps_epi32 (__m256 __A, const int __R)
+_mm256_ipcvtt_roundps_epi8 (__m256 __A, const int __R)
{
return (__m256i)
__builtin_ia32_cvttps2ibs256_mask_round ((__v8sf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_ipcvtt_roundps_epi32 (__m256i __W, __mmask8 __U, __m256 __A,
- const int __R)
+_mm256_mask_ipcvtt_roundps_epi8 (__m256i __W, __mmask8 __U, __m256 __A,
+ const int __R)
{
return (__m256i) __builtin_ia32_cvttps2ibs256_mask_round ((__v8sf) __A,
(__v8si) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_ipcvtt_roundps_epi32 (__mmask8 __U, __m256 __A, const int __R)
+_mm256_maskz_ipcvtt_roundps_epi8 (__mmask8 __U, __m256 __A, const int __R)
{
return
(__m256i) __builtin_ia32_cvttps2ibs256_mask_round ((__v8sf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_ipcvtt_roundps_epu32 (__m256 __A, const int __R)
+_mm256_ipcvtt_roundps_epu8 (__m256 __A, const int __R)
{
return (__m256i)
__builtin_ia32_cvttps2iubs256_mask_round ((__v8sf) __A,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_ipcvtt_roundps_epu32 (__m256i __W, __mmask8 __U, __m256 __A,
- const int __R)
+_mm256_mask_ipcvtt_roundps_epu8 (__m256i __W, __mmask8 __U, __m256 __A,
+ const int __R)
{
return (__m256i) __builtin_ia32_cvttps2iubs256_mask_round ((__v8sf) __A,
(__v8si) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_ipcvtt_roundps_epu32 (__mmask8 __U, __m256 __A, const int __R)
+_mm256_maskz_ipcvtt_roundps_epu8 (__mmask8 __U, __m256 __A, const int __R)
{
return
(__m256i) __builtin_ia32_cvttps2iubs256_mask_round ((__v8sf) __A,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm256_mask_cvtts_roundpd_epi32 (__m128i __W, __mmask8 __U, __m256d __A,
- const int __R)
+ const int __R)
{
return (__m128i) __builtin_ia32_cvttpd2dqs256_mask_round ((__v4df) __A,
(__v4si) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm256_mask_cvtts_roundpd_epi64 (__m256i __W, __mmask8 __U, __m256d __A,
- const int __R)
+ const int __R)
{
return (__m256i) __builtin_ia32_cvttpd2qqs256_mask_round ((__v4df) __A,
(__v4di) __W,
extern __inline __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm256_mask_cvtts_roundpd_epu32 (__m128i __W, __mmask8 __U, __m256d __A,
- const int __R)
+ const int __R)
{
return (__m128i) __builtin_ia32_cvttpd2udqs256_mask_round ((__v4df) __A,
(__v4si) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm256_mask_cvtts_roundpd_epu64 (__m256i __W, __mmask8 __U, __m256d __A,
- const int __R)
+ const int __R)
{
return (__m256i) __builtin_ia32_cvttpd2uqqs256_mask_round ((__v4df) __A,
(__v4di) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm256_mask_cvtts_roundps_epi32 (__m256i __W, __mmask8 __U, __m256 __A,
- const int __R)
+ const int __R)
{
return (__m256i) __builtin_ia32_cvttps2dqs256_mask_round ((__v8sf) __A,
(__v8si) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm256_mask_cvtts_roundps_epi64 (__m256i __W, __mmask8 __U, __m128 __A,
- const int __R)
+ const int __R)
{
return (__m256i) __builtin_ia32_cvttps2qqs256_mask_round ((__v4sf) __A,
(__v4di) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm256_mask_cvtts_roundps_epu32 (__m256i __W, __mmask8 __U, __m256 __A,
- const int __R)
+ const int __R)
{
return (__m256i) __builtin_ia32_cvttps2udqs256_mask_round ((__v8sf) __A,
(__v8si) __W,
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm256_mask_cvtts_roundps_epu64 (__m256i __W, __mmask8 __U, __m128 __A,
- const int __R)
+ const int __R)
{
return (__m256i) __builtin_ia32_cvttps2uqqs256_mask_round ((__v4sf) __A,
(__v4di) __W,
}
#else
-#define _mm256_ipcvt_roundph_epi16(A, R) \
+#define _mm256_ipcvt_roundph_epi8(A, R) \
((__m256i) \
__builtin_ia32_cvtph2ibs256_mask_round ((__v16hf) (A), \
(__v16hi) \
(__mmask16) (-1), \
(R)))
-#define _mm256_mask_ipcvt_roundph_epi16(W, U, A, R) \
+#define _mm256_mask_ipcvt_roundph_epi8(W, U, A, R) \
((__m256i) __builtin_ia32_cvtph2ibs256_mask_round ((__v16hf) (A), \
(__v16hi) (W), \
(__mmask16) (U), \
(R)))
-#define _mm256_maskz_ipcvt_roundph_epi16(U, A, R) \
+#define _mm256_maskz_ipcvt_roundph_epi8(U, A, R) \
((__m256i) \
__builtin_ia32_cvtph2ibs256_mask_round ((__v16hf) (A), \
(__v16hi) \
(__mmask16) (U), \
(R)))
-#define _mm256_ipcvt_roundph_epu16(A, R) \
+#define _mm256_ipcvt_roundph_epu8(A, R) \
((__m256i) \
__builtin_ia32_cvtph2iubs256_mask_round ((__v16hf) (A), \
(__v16hi) \
(__mmask16) (-1), \
(R)))
-#define _mm256_mask_ipcvt_roundph_epu16(W, U, A, R) \
+#define _mm256_mask_ipcvt_roundph_epu8(W, U, A, R) \
((__m256i) __builtin_ia32_cvtph2iubs256_mask_round ((__v16hf) (A), \
(__v16hi) (W), \
(__mmask16) (U), \
(R)))
-#define _mm256_maskz_ipcvt_roundph_epu16(U, A, R) \
+#define _mm256_maskz_ipcvt_roundph_epu8(U, A, R) \
((__m256i) \
__builtin_ia32_cvtph2iubs256_mask_round ((__v16hf) (A), \
(__v16hi) \
(__mmask16) (U), \
(R)))
-#define _mm256_ipcvt_roundps_epi32(A, R) \
+#define _mm256_ipcvt_roundps_epi8(A, R) \
((__m256i) \
__builtin_ia32_cvtps2ibs256_mask_round ((__v8sf) (A), \
(__v8si) \
(__mmask8) (-1), \
(R)))
-#define _mm256_mask_ipcvt_roundps_epi32(W, U, A, R) \
+#define _mm256_mask_ipcvt_roundps_epi8(W, U, A, R) \
((__m256i) __builtin_ia32_cvtps2ibs256_mask_round ((__v8sf) (A), \
(__v8si) (W), \
(__mmask8) (U), \
(R)))
-#define _mm256_maskz_ipcvt_roundps_epi32(U, A, R) \
+#define _mm256_maskz_ipcvt_roundps_epi8(U, A, R) \
((__m256i) \
__builtin_ia32_cvtps2ibs256_mask_round ((__v8sf) (A), \
(__v8si) \
(__mmask8) (U), \
(R)))
-#define _mm256_ipcvt_roundps_epu32(A, R) \
+#define _mm256_ipcvt_roundps_epu8(A, R) \
((__m256i) \
__builtin_ia32_cvtps2iubs256_mask_round ((__v8sf) (A), \
(__v8si) \
(__mmask8) (-1), \
(R)))
-#define _mm256_mask_ipcvt_roundps_epu32(W, U, A, R) \
+#define _mm256_mask_ipcvt_roundps_epu8(W, U, A, R) \
((__m256i) __builtin_ia32_cvtps2iubs256_mask_round ((__v8sf) (A), \
(__v8si) (W), \
(__mmask8) (U), \
(R)))
-#define _mm256_maskz_ipcvt_roundps_epu32(U, A, R) \
+#define _mm256_maskz_ipcvt_roundps_epu8(U, A, R) \
((__m256i) \
__builtin_ia32_cvtps2iubs256_mask_round ((__v8sf) (A), \
(__v8si) \
(__mmask8) (U), \
(R)))
-#define _mm256_ipcvtt_roundph_epi16(A, R) \
+#define _mm256_ipcvtt_roundph_epi8(A, R) \
((__m256i) \
__builtin_ia32_cvttph2ibs256_mask_round ((__v16hf) (A), \
(__v16hi) \
(__mmask16) (-1), \
(R)))
-#define _mm256_mask_ipcvtt_roundph_epi16(W, U, A, R) \
+#define _mm256_mask_ipcvtt_roundph_epi8(W, U, A, R) \
((__m256i) __builtin_ia32_cvttph2ibs256_mask_round ((__v16hf) (A), \
(__v16hi) (W), \
(__mmask16) (U), \
(R)))
-#define _mm256_maskz_ipcvtt_roundph_epi16(U, A, R) \
+#define _mm256_maskz_ipcvtt_roundph_epi8(U, A, R) \
((__m256i) \
__builtin_ia32_cvttph2ibs256_mask_round ((__v16hf) (A), \
(__v16hi) \
(__mmask16) (U), \
(R)))
-#define _mm256_ipcvtt_roundph_epu16(A, R) \
+#define _mm256_ipcvtt_roundph_epu8(A, R) \
((__m256i) \
__builtin_ia32_cvttph2iubs256_mask_round ((__v16hf) (A), \
(__v16hi) \
(__mmask16) (-1), \
(R)))
-#define _mm256_mask_ipcvtt_roundph_epu16(W, U, A, R) \
+#define _mm256_mask_ipcvtt_roundph_epu8(W, U, A, R) \
((__m256i) __builtin_ia32_cvttph2iubs256_mask_round ((__v16hf) (A), \
(__v16hi) (W), \
(__mmask16) (U), \
(R)))
-#define _mm256_maskz_ipcvtt_roundph_epu16(U, A, R) \
+#define _mm256_maskz_ipcvtt_roundph_epu8(U, A, R) \
((__m256i) \
__builtin_ia32_cvttph2iubs256_mask_round ((__v16hf) (A), \
(__v16hi) \
(__mmask16) (U), \
(R)))
-#define _mm256_ipcvtt_roundps_epi32(A, R) \
+#define _mm256_ipcvtt_roundps_epi8(A, R) \
((__m256i) \
__builtin_ia32_cvttps2ibs256_mask_round ((__v8sf) (A), \
(__v8si) \
(__mmask8) (-1), \
(R)))
-#define _mm256_mask_ipcvtt_roundps_epi32(W, U, A, R) \
+#define _mm256_mask_ipcvtt_roundps_epi8(W, U, A, R) \
((__m256i) __builtin_ia32_cvttps2ibs256_mask_round ((__v8sf) (A), \
(__v8si) (W), \
(__mmask8) (U), \
(R)))
-#define _mm256_maskz_ipcvtt_roundps_epi32(U, A, R) \
+#define _mm256_maskz_ipcvtt_roundps_epi8(U, A, R) \
((__m256i) \
__builtin_ia32_cvttps2ibs256_mask_round ((__v8sf) (A), \
(__v8si) \
(__mmask8) (U), \
(R)))
-#define _mm256_ipcvtt_roundps_epu32(A, R) \
+#define _mm256_ipcvtt_roundps_epu8(A, R) \
((__m256i) \
__builtin_ia32_cvttps2iubs256_mask_round ((__v8sf) (A), \
(__v8si) \
(__mmask8) (-1), \
(R)))
-#define _mm256_mask_ipcvtt_roundps_epu32(W, U, A, R) \
+#define _mm256_mask_ipcvtt_roundps_epu8(W, U, A, R) \
((__m256i) __builtin_ia32_cvttps2iubs256_mask_round ((__v8sf) (A), \
(__v8si) (W), \
(__mmask8) (U), \
(R)))
-#define _mm256_maskz_ipcvtt_roundps_epu32(U, A, R) \
-((__m256i) \
- __builtin_ia32_cvttps2iubs256_mask_round ((__v8sf) (A), \
- (__v8si) \
- (_mm256_setzero_si256 ()), \
- (__mmask8) (U), \
- (R)))
+#define _mm256_maskz_ipcvtt_roundps_epu8(U, A, R) \
+ ((__m256i) \
+ __builtin_ia32_cvttps2iubs256_mask_round ((__v8sf) (A), \
+ (__v8si) \
+ (_mm256_setzero_si256 ()), \
+ (__mmask8) (U), \
+ (R)))
#define _mm256_cvtts_roundpd_epi32(A, R) \
((__m128i) \
void extern
avx10_2_test (void)
{
- xi = _mm512_ipcvt_roundph_epi16 (xh, 4);
- xi = _mm512_mask_ipcvt_roundph_epi16 (xi, m32, xh, 8);
- xi = _mm512_maskz_ipcvt_roundph_epi16 (m32, xh, 11);
+ xi = _mm512_ipcvt_roundph_epi8 (xh, 4);
+ xi = _mm512_mask_ipcvt_roundph_epi8 (xi, m32, xh, 8);
+ xi = _mm512_maskz_ipcvt_roundph_epi8 (m32, xh, 11);
- xi = _mm512_ipcvt_roundph_epu16 (xh, 4);
- xi = _mm512_mask_ipcvt_roundph_epu16 (xi, m32, xh, 8);
- xi = _mm512_maskz_ipcvt_roundph_epu16 (m32, xh, 11);
+ xi = _mm512_ipcvt_roundph_epu8 (xh, 4);
+ xi = _mm512_mask_ipcvt_roundph_epu8 (xi, m32, xh, 8);
+ xi = _mm512_maskz_ipcvt_roundph_epu8 (m32, xh, 11);
- xi = _mm512_ipcvtt_roundph_epi16 (xh, 4);
- xi = _mm512_mask_ipcvtt_roundph_epi16 (xi, m32, xh, 8);
- xi = _mm512_maskz_ipcvtt_roundph_epi16 (m32, xh, 8);
+ xi = _mm512_ipcvtt_roundph_epi8 (xh, 4);
+ xi = _mm512_mask_ipcvtt_roundph_epi8 (xi, m32, xh, 8);
+ xi = _mm512_maskz_ipcvtt_roundph_epi8 (m32, xh, 8);
- xi = _mm512_ipcvtt_roundph_epu16 (xh, 4);
- xi = _mm512_mask_ipcvtt_roundph_epu16 (xi, m32, xh, 8);
- xi = _mm512_maskz_ipcvtt_roundph_epu16 (m32, xh, 8);
+ xi = _mm512_ipcvtt_roundph_epu8 (xh, 4);
+ xi = _mm512_mask_ipcvtt_roundph_epu8 (xi, m32, xh, 8);
+ xi = _mm512_maskz_ipcvtt_roundph_epu8 (m32, xh, 8);
- xi = _mm512_ipcvt_roundps_epi32 (x, 4);
- xi = _mm512_mask_ipcvt_roundps_epi32 (xi, m16, x, 8);
- xi = _mm512_maskz_ipcvt_roundps_epi32 (m16, x, 11);
+ xi = _mm512_ipcvt_roundps_epi8 (x, 4);
+ xi = _mm512_mask_ipcvt_roundps_epi8 (xi, m16, x, 8);
+ xi = _mm512_maskz_ipcvt_roundps_epi8 (m16, x, 11);
- xi = _mm512_ipcvt_roundps_epu32 (x, 4);
- xi = _mm512_mask_ipcvt_roundps_epu32 (xi, m16, x, 8);
- xi = _mm512_maskz_ipcvt_roundps_epu32 (m16, x, 11);
+ xi = _mm512_ipcvt_roundps_epu8 (x, 4);
+ xi = _mm512_mask_ipcvt_roundps_epu8 (xi, m16, x, 8);
+ xi = _mm512_maskz_ipcvt_roundps_epu8 (m16, x, 11);
- xi = _mm512_ipcvtt_roundps_epi32 (x, 4);
- xi = _mm512_mask_ipcvtt_roundps_epi32 (xi, m16, x, 8);
- xi = _mm512_maskz_ipcvtt_roundps_epi32 (m16, x, 8);
+ xi = _mm512_ipcvtt_roundps_epi8 (x, 4);
+ xi = _mm512_mask_ipcvtt_roundps_epi8 (xi, m16, x, 8);
+ xi = _mm512_maskz_ipcvtt_roundps_epi8 (m16, x, 8);
- xi = _mm512_ipcvtt_roundps_epu32 (x, 4);
- xi = _mm512_mask_ipcvtt_roundps_epu32 (xi, m16, x, 8);
- xi = _mm512_maskz_ipcvtt_roundps_epu32 (m16, x, 8);
+ xi = _mm512_ipcvtt_roundps_epu8 (x, 4);
+ xi = _mm512_mask_ipcvtt_roundps_epu8 (xi, m16, x, 8);
+ xi = _mm512_maskz_ipcvtt_roundps_epu8 (m16, x, 8);
- xi = _mm512_ipcvtbf16_epi16 (xbh);
- xi = _mm512_mask_ipcvtbf16_epi16 (xi, m32, xbh);
- xi = _mm512_maskz_ipcvtbf16_epi16 (m32, xbh);
+ xi = _mm512_ipcvtbf16_epi8 (xbh);
+ xi = _mm512_mask_ipcvtbf16_epi8 (xi, m32, xbh);
+ xi = _mm512_maskz_ipcvtbf16_epi8 (m32, xbh);
- xi = _mm512_ipcvtbf16_epu16 (xbh);
- xi = _mm512_mask_ipcvtbf16_epu16 (xi, m32, xbh);
- xi = _mm512_maskz_ipcvtbf16_epu16 (m32, xbh);
+ xi = _mm512_ipcvtbf16_epu8 (xbh);
+ xi = _mm512_mask_ipcvtbf16_epu8 (xi, m32, xbh);
+ xi = _mm512_maskz_ipcvtbf16_epu8 (m32, xbh);
- xi = _mm512_ipcvttbf16_epi16 (xbh);
- xi = _mm512_mask_ipcvttbf16_epi16 (xi, m32, xbh);
- xi = _mm512_maskz_ipcvttbf16_epi16 (m32, xbh);
+ xi = _mm512_ipcvttbf16_epi8 (xbh);
+ xi = _mm512_mask_ipcvttbf16_epi8 (xi, m32, xbh);
+ xi = _mm512_maskz_ipcvttbf16_epi8 (m32, xbh);
- xi = _mm512_ipcvttbf16_epu16 (xbh);
- xi = _mm512_mask_ipcvttbf16_epu16 (xi, m32, xbh);
- xi = _mm512_maskz_ipcvttbf16_epu16 (m32, xbh);
+ xi = _mm512_ipcvttbf16_epu8 (xbh);
+ xi = _mm512_mask_ipcvttbf16_epu8 (xi, m32, xbh);
+ xi = _mm512_maskz_ipcvttbf16_epu8 (m32, xbh);
hxi = _mm512_cvtts_roundpd_epi32 (xd, 8);
hxi = _mm512_mask_cvtts_roundpd_epi32 (hxi, m8, xd, 8);
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
- res1.x = INTRINSIC (_ipcvtbf16_epi16) (s.x);
- res2.x = INTRINSIC (_mask_ipcvtbf16_epi16) (res2.x, mask, s.x);
- res3.x = INTRINSIC (_maskz_ipcvtbf16_epi16) (mask, s.x);
+ res1.x = INTRINSIC (_ipcvtbf16_epi8) (s.x);
+ res2.x = INTRINSIC (_mask_ipcvtbf16_epi8) (res2.x, mask, s.x);
+ res3.x = INTRINSIC (_maskz_ipcvtbf16_epi8) (mask, s.x);
CALC (s.a, res_ref);
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
- res1.x = INTRINSIC (_ipcvtbf16_epu16) (s.x);
- res2.x = INTRINSIC (_mask_ipcvtbf16_epu16) (res2.x, mask, s.x);
- res3.x = INTRINSIC (_maskz_ipcvtbf16_epu16) (mask, s.x);
+ res1.x = INTRINSIC (_ipcvtbf16_epu8) (s.x);
+ res2.x = INTRINSIC (_mask_ipcvtbf16_epu8) (res2.x, mask, s.x);
+ res3.x = INTRINSIC (_maskz_ipcvtbf16_epu8) (mask, s.x);
CALC (s.a, res_ref);
res2.a[i] = DEFAULT_VALUE;
#if AVX512F_LEN == 128
- res1.x = INTRINSIC (_ipcvtph_epi16) (s.x);
- res2.x = INTRINSIC (_mask_ipcvtph_epi16) (res2.x, mask, s.x);
- res3.x = INTRINSIC (_maskz_ipcvtph_epi16) (mask, s.x);
+ res1.x = INTRINSIC (_ipcvtph_epi8) (s.x);
+ res2.x = INTRINSIC (_mask_ipcvtph_epi8) (res2.x, mask, s.x);
+ res3.x = INTRINSIC (_maskz_ipcvtph_epi8) (mask, s.x);
#else
- res1.x = INTRINSIC (_ipcvt_roundph_epi16) (s.x, 8);
- res2.x = INTRINSIC (_mask_ipcvt_roundph_epi16) (res2.x, mask, s.x, 8);
- res3.x = INTRINSIC (_maskz_ipcvt_roundph_epi16) (mask, s.x, 8);
+ res1.x = INTRINSIC (_ipcvt_roundph_epi8) (s.x, 8);
+ res2.x = INTRINSIC (_mask_ipcvt_roundph_epi8) (res2.x, mask, s.x, 8);
+ res3.x = INTRINSIC (_maskz_ipcvt_roundph_epi8) (mask, s.x, 8);
#endif
CALC (s.a, res_ref);
res2.a[i] = DEFAULT_VALUE;
#if AVX512F_LEN == 128
- res1.x = INTRINSIC (_ipcvtph_epu16) (s.x);
- res2.x = INTRINSIC (_mask_ipcvtph_epu16) (res2.x, mask, s.x);
- res3.x = INTRINSIC (_maskz_ipcvtph_epu16) (mask, s.x);
+ res1.x = INTRINSIC (_ipcvtph_epu8) (s.x);
+ res2.x = INTRINSIC (_mask_ipcvtph_epu8) (res2.x, mask, s.x);
+ res3.x = INTRINSIC (_maskz_ipcvtph_epu8) (mask, s.x);
#else
- res1.x = INTRINSIC (_ipcvt_roundph_epu16) (s.x, 8);
- res2.x = INTRINSIC (_mask_ipcvt_roundph_epu16) (res2.x, mask, s.x, 8);
- res3.x = INTRINSIC (_maskz_ipcvt_roundph_epu16) (mask, s.x, 8);
+ res1.x = INTRINSIC (_ipcvt_roundph_epu8) (s.x, 8);
+ res2.x = INTRINSIC (_mask_ipcvt_roundph_epu8) (res2.x, mask, s.x, 8);
+ res3.x = INTRINSIC (_maskz_ipcvt_roundph_epu8) (mask, s.x, 8);
#endif
CALC (s.a, res_ref);
res2.a[i] = DEFAULT_VALUE;
#if AVX512F_LEN == 128
- res1.x = INTRINSIC (_ipcvtps_epi32) (s.x);
- res2.x = INTRINSIC (_mask_ipcvtps_epi32) (res2.x, mask, s.x);
- res3.x = INTRINSIC (_maskz_ipcvtps_epi32) (mask, s.x);
+ res1.x = INTRINSIC (_ipcvtps_epi8) (s.x);
+ res2.x = INTRINSIC (_mask_ipcvtps_epi8) (res2.x, mask, s.x);
+ res3.x = INTRINSIC (_maskz_ipcvtps_epi8) (mask, s.x);
#else
- res1.x = INTRINSIC (_ipcvt_roundps_epi32) (s.x, 8);
- res2.x = INTRINSIC (_mask_ipcvt_roundps_epi32) (res2.x, mask, s.x, 8);
- res3.x = INTRINSIC (_maskz_ipcvt_roundps_epi32) (mask, s.x, 8);
+ res1.x = INTRINSIC (_ipcvt_roundps_epi8) (s.x, 8);
+ res2.x = INTRINSIC (_mask_ipcvt_roundps_epi8) (res2.x, mask, s.x, 8);
+ res3.x = INTRINSIC (_maskz_ipcvt_roundps_epi8) (mask, s.x, 8);
#endif
CALC (s.a, res_ref);
res2.a[i] = DEFAULT_VALUE;
#if AVX512F_LEN == 128
- res1.x = INTRINSIC (_ipcvtps_epu32) (s.x);
- res2.x = INTRINSIC (_mask_ipcvtps_epu32) (res2.x, mask, s.x);
- res3.x = INTRINSIC (_maskz_ipcvtps_epu32) (mask, s.x);
+ res1.x = INTRINSIC (_ipcvtps_epu8) (s.x);
+ res2.x = INTRINSIC (_mask_ipcvtps_epu8) (res2.x, mask, s.x);
+ res3.x = INTRINSIC (_maskz_ipcvtps_epu8) (mask, s.x);
#else
- res1.x = INTRINSIC (_ipcvt_roundps_epu32) (s.x, 8);
- res2.x = INTRINSIC (_mask_ipcvt_roundps_epu32) (res2.x, mask, s.x, 8);
- res3.x = INTRINSIC (_maskz_ipcvt_roundps_epu32) (mask, s.x, 8);
+ res1.x = INTRINSIC (_ipcvt_roundps_epu8) (s.x, 8);
+ res2.x = INTRINSIC (_mask_ipcvt_roundps_epu8) (res2.x, mask, s.x, 8);
+ res3.x = INTRINSIC (_maskz_ipcvt_roundps_epu8) (mask, s.x, 8);
#endif
CALC (s.a, res_ref);
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
- res1.x = INTRINSIC (_ipcvttbf16_epi16) (s.x);
- res2.x = INTRINSIC (_mask_ipcvttbf16_epi16) (res2.x, mask, s.x);
- res3.x = INTRINSIC (_maskz_ipcvttbf16_epi16) (mask, s.x);
+ res1.x = INTRINSIC (_ipcvttbf16_epi8) (s.x);
+ res2.x = INTRINSIC (_mask_ipcvttbf16_epi8) (res2.x, mask, s.x);
+ res3.x = INTRINSIC (_maskz_ipcvttbf16_epi8) (mask, s.x);
CALC (s.a, res_ref);
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
- res1.x = INTRINSIC (_ipcvttbf16_epu16) (s.x);
- res2.x = INTRINSIC (_mask_ipcvttbf16_epu16) (res2.x, mask, s.x);
- res3.x = INTRINSIC (_maskz_ipcvttbf16_epu16) (mask, s.x);
+ res1.x = INTRINSIC (_ipcvttbf16_epu8) (s.x);
+ res2.x = INTRINSIC (_mask_ipcvttbf16_epu8) (res2.x, mask, s.x);
+ res3.x = INTRINSIC (_maskz_ipcvttbf16_epu8) (mask, s.x);
CALC (s.a, res_ref);
res2.a[i] = DEFAULT_VALUE;
#if AVX512F_LEN == 128
- res1.x = INTRINSIC (_ipcvttph_epi16) (s.x);
- res2.x = INTRINSIC (_mask_ipcvttph_epi16) (res2.x, mask, s.x);
- res3.x = INTRINSIC (_maskz_ipcvttph_epi16) (mask, s.x);
+ res1.x = INTRINSIC (_ipcvttph_epi8) (s.x);
+ res2.x = INTRINSIC (_mask_ipcvttph_epi8) (res2.x, mask, s.x);
+ res3.x = INTRINSIC (_maskz_ipcvttph_epi8) (mask, s.x);
#else
- res1.x = INTRINSIC (_ipcvtt_roundph_epi16) (s.x, 8);
- res2.x = INTRINSIC (_mask_ipcvtt_roundph_epi16) (res2.x, mask, s.x, 8);
- res3.x = INTRINSIC (_maskz_ipcvtt_roundph_epi16) (mask, s.x, 8);
+ res1.x = INTRINSIC (_ipcvtt_roundph_epi8) (s.x, 8);
+ res2.x = INTRINSIC (_mask_ipcvtt_roundph_epi8) (res2.x, mask, s.x, 8);
+ res3.x = INTRINSIC (_maskz_ipcvtt_roundph_epi8) (mask, s.x, 8);
#endif
CALC (s.a, res_ref);
res2.a[i] = DEFAULT_VALUE;
#if AVX512F_LEN == 128
- res1.x = INTRINSIC (_ipcvttph_epu16) (s.x);
- res2.x = INTRINSIC (_mask_ipcvttph_epu16) (res2.x, mask, s.x);
- res3.x = INTRINSIC (_maskz_ipcvttph_epu16) (mask, s.x);
+ res1.x = INTRINSIC (_ipcvttph_epu8) (s.x);
+ res2.x = INTRINSIC (_mask_ipcvttph_epu8) (res2.x, mask, s.x);
+ res3.x = INTRINSIC (_maskz_ipcvttph_epu8) (mask, s.x);
#else
- res1.x = INTRINSIC (_ipcvtt_roundph_epu16) (s.x, 8);
- res2.x = INTRINSIC (_mask_ipcvtt_roundph_epu16) (res2.x, mask, s.x, 8);
- res3.x = INTRINSIC (_maskz_ipcvtt_roundph_epu16) (mask, s.x, 8);
+ res1.x = INTRINSIC (_ipcvtt_roundph_epu8) (s.x, 8);
+ res2.x = INTRINSIC (_mask_ipcvtt_roundph_epu8) (res2.x, mask, s.x, 8);
+ res3.x = INTRINSIC (_maskz_ipcvtt_roundph_epu8) (mask, s.x, 8);
#endif
CALC (s.a, res_ref);
res2.a[i] = DEFAULT_VALUE;
#if AVX512F_LEN == 128
- res1.x = INTRINSIC (_ipcvttps_epi32) (s.x);
- res2.x = INTRINSIC (_mask_ipcvttps_epi32) (res2.x, mask, s.x);
- res3.x = INTRINSIC (_maskz_ipcvttps_epi32) (mask, s.x);
+ res1.x = INTRINSIC (_ipcvttps_epi8) (s.x);
+ res2.x = INTRINSIC (_mask_ipcvttps_epi8) (res2.x, mask, s.x);
+ res3.x = INTRINSIC (_maskz_ipcvttps_epi8) (mask, s.x);
#else
- res1.x = INTRINSIC (_ipcvtt_roundps_epi32) (s.x, 8);
- res2.x = INTRINSIC (_mask_ipcvtt_roundps_epi32) (res2.x, mask, s.x, 8);
- res3.x = INTRINSIC (_maskz_ipcvtt_roundps_epi32) (mask, s.x, 8);
+ res1.x = INTRINSIC (_ipcvtt_roundps_epi8) (s.x, 8);
+ res2.x = INTRINSIC (_mask_ipcvtt_roundps_epi8) (res2.x, mask, s.x, 8);
+ res3.x = INTRINSIC (_maskz_ipcvtt_roundps_epi8) (mask, s.x, 8);
#endif
CALC (s.a, res_ref);
res2.a[i] = DEFAULT_VALUE;
#if AVX512F_LEN == 128
- res1.x = INTRINSIC (_ipcvttps_epu32) (s.x);
- res2.x = INTRINSIC (_mask_ipcvttps_epu32) (res2.x, mask, s.x);
- res3.x = INTRINSIC (_maskz_ipcvttps_epu32) (mask, s.x);
+ res1.x = INTRINSIC (_ipcvttps_epu8) (s.x);
+ res2.x = INTRINSIC (_mask_ipcvttps_epu8) (res2.x, mask, s.x);
+ res3.x = INTRINSIC (_maskz_ipcvttps_epu8) (mask, s.x);
#else
- res1.x = INTRINSIC (_ipcvtt_roundps_epu32) (s.x, 8);
- res2.x = INTRINSIC (_mask_ipcvtt_roundps_epu32) (res2.x, mask, s.x, 8);
- res3.x = INTRINSIC (_maskz_ipcvtt_roundps_epu32) (mask, s.x, 8);
+ res1.x = INTRINSIC (_ipcvtt_roundps_epu8) (s.x, 8);
+ res2.x = INTRINSIC (_mask_ipcvtt_roundps_epu8) (res2.x, mask, s.x, 8);
+ res3.x = INTRINSIC (_maskz_ipcvtt_roundps_epu8) (mask, s.x, 8);
#endif
CALC (s.a, res_ref);
void extern
avx10_2_test (void)
{
- xi = _mm256_ipcvt_roundph_epi16 (xh, 4);
- xi = _mm256_mask_ipcvt_roundph_epi16 (xi, m16, xh, 8);
- xi = _mm256_maskz_ipcvt_roundph_epi16 (m16, xh, 11);
+ xi = _mm256_ipcvt_roundph_epi8 (xh, 4);
+ xi = _mm256_mask_ipcvt_roundph_epi8 (xi, m16, xh, 8);
+ xi = _mm256_maskz_ipcvt_roundph_epi8 (m16, xh, 11);
- xi = _mm256_ipcvt_roundph_epu16 (xh, 4);
- xi = _mm256_mask_ipcvt_roundph_epu16 (xi, m16, xh, 8);
- xi = _mm256_maskz_ipcvt_roundph_epu16 (m16, xh, 11);
+ xi = _mm256_ipcvt_roundph_epu8 (xh, 4);
+ xi = _mm256_mask_ipcvt_roundph_epu8 (xi, m16, xh, 8);
+ xi = _mm256_maskz_ipcvt_roundph_epu8 (m16, xh, 11);
- xi = _mm256_ipcvtt_roundph_epi16 (xh, 4);
- xi = _mm256_mask_ipcvtt_roundph_epi16 (xi, m16, xh, 8);
- xi = _mm256_maskz_ipcvtt_roundph_epi16 (m16, xh, 8);
+ xi = _mm256_ipcvtt_roundph_epi8 (xh, 4);
+ xi = _mm256_mask_ipcvtt_roundph_epi8 (xi, m16, xh, 8);
+ xi = _mm256_maskz_ipcvtt_roundph_epi8 (m16, xh, 8);
- xi = _mm256_ipcvtt_roundph_epu16 (xh, 4);
- xi = _mm256_mask_ipcvtt_roundph_epu16 (xi, m16, xh, 8);
- xi = _mm256_maskz_ipcvtt_roundph_epu16 (m16, xh, 8);
+ xi = _mm256_ipcvtt_roundph_epu8 (xh, 4);
+ xi = _mm256_mask_ipcvtt_roundph_epu8 (xi, m16, xh, 8);
+ xi = _mm256_maskz_ipcvtt_roundph_epu8 (m16, xh, 8);
- xi = _mm256_ipcvt_roundps_epi32 (x, 4);
- xi = _mm256_mask_ipcvt_roundps_epi32 (xi, m8, x, 8);
- xi = _mm256_maskz_ipcvt_roundps_epi32 (m8, x, 11);
+ xi = _mm256_ipcvt_roundps_epi8 (x, 4);
+ xi = _mm256_mask_ipcvt_roundps_epi8 (xi, m8, x, 8);
+ xi = _mm256_maskz_ipcvt_roundps_epi8 (m8, x, 11);
- xi = _mm256_ipcvt_roundps_epu32 (x, 4);
- xi = _mm256_mask_ipcvt_roundps_epu32 (xi, m8, x, 8);
- xi = _mm256_maskz_ipcvt_roundps_epu32 (m8, x, 11);
+ xi = _mm256_ipcvt_roundps_epu8 (x, 4);
+ xi = _mm256_mask_ipcvt_roundps_epu8 (xi, m8, x, 8);
+ xi = _mm256_maskz_ipcvt_roundps_epu8 (m8, x, 11);
- xi = _mm256_ipcvtt_roundps_epi32 (x, 4);
- xi = _mm256_mask_ipcvtt_roundps_epi32 (xi, m8, x, 8);
- xi = _mm256_maskz_ipcvtt_roundps_epi32 (m8, x, 8);
+ xi = _mm256_ipcvtt_roundps_epi8 (x, 4);
+ xi = _mm256_mask_ipcvtt_roundps_epi8 (xi, m8, x, 8);
+ xi = _mm256_maskz_ipcvtt_roundps_epi8 (m8, x, 8);
- xi = _mm256_ipcvtt_roundps_epu32 (x, 4);
- xi = _mm256_mask_ipcvtt_roundps_epu32 (xi, m8, x, 8);
- xi = _mm256_maskz_ipcvtt_roundps_epu32 (m8, x, 8);
+ xi = _mm256_ipcvtt_roundps_epu8 (x, 4);
+ xi = _mm256_mask_ipcvtt_roundps_epu8 (xi, m8, x, 8);
+ xi = _mm256_maskz_ipcvtt_roundps_epu8 (m8, x, 8);
- xi = _mm256_ipcvtbf16_epi16 (xbh);
- xi = _mm256_mask_ipcvtbf16_epi16 (xi, m16, xbh);
- xi = _mm256_maskz_ipcvtbf16_epi16 (m16, xbh);
+ xi = _mm256_ipcvtbf16_epi8 (xbh);
+ xi = _mm256_mask_ipcvtbf16_epi8 (xi, m16, xbh);
+ xi = _mm256_maskz_ipcvtbf16_epi8 (m16, xbh);
- xi = _mm256_ipcvtbf16_epu16 (xbh);
- xi = _mm256_mask_ipcvtbf16_epu16 (xi, m16, xbh);
- xi = _mm256_maskz_ipcvtbf16_epu16 (m16, xbh);
+ xi = _mm256_ipcvtbf16_epu8 (xbh);
+ xi = _mm256_mask_ipcvtbf16_epu8 (xi, m16, xbh);
+ xi = _mm256_maskz_ipcvtbf16_epu8 (m16, xbh);
- xi = _mm256_ipcvttbf16_epi16 (xbh);
- xi = _mm256_mask_ipcvttbf16_epi16 (xi, m16, xbh);
- xi = _mm256_maskz_ipcvttbf16_epi16 (m16, xbh);
+ xi = _mm256_ipcvttbf16_epi8 (xbh);
+ xi = _mm256_mask_ipcvttbf16_epi8 (xi, m16, xbh);
+ xi = _mm256_maskz_ipcvttbf16_epi8 (m16, xbh);
- xi = _mm256_ipcvttbf16_epu16 (xbh);
- xi = _mm256_mask_ipcvttbf16_epu16 (xi, m16, xbh);
- xi = _mm256_maskz_ipcvttbf16_epu16 (m16, xbh);
+ xi = _mm256_ipcvttbf16_epu8 (xbh);
+ xi = _mm256_mask_ipcvttbf16_epu8 (xi, m16, xbh);
+ xi = _mm256_maskz_ipcvttbf16_epu8 (m16, xbh);
- hxi = _mm_ipcvtph_epi16 (hxh);
- hxi = _mm_mask_ipcvtph_epi16 (hxi, m8, hxh);
- hxi = _mm_maskz_ipcvtph_epi16 (m8, hxh);
+ hxi = _mm_ipcvtph_epi8 (hxh);
+ hxi = _mm_mask_ipcvtph_epi8 (hxi, m8, hxh);
+ hxi = _mm_maskz_ipcvtph_epi8 (m8, hxh);
- hxi = _mm_ipcvtph_epu16 (hxh);
- hxi = _mm_mask_ipcvtph_epu16 (hxi, m8, hxh);
- hxi = _mm_maskz_ipcvtph_epu16 (m8, hxh);
+ hxi = _mm_ipcvtph_epu8 (hxh);
+ hxi = _mm_mask_ipcvtph_epu8 (hxi, m8, hxh);
+ hxi = _mm_maskz_ipcvtph_epu8 (m8, hxh);
- hxi = _mm_ipcvttph_epi16 (hxh);
- hxi = _mm_mask_ipcvttph_epi16 (hxi, m8, hxh);
- hxi = _mm_maskz_ipcvttph_epi16 (m8, hxh);
+ hxi = _mm_ipcvttph_epi8 (hxh);
+ hxi = _mm_mask_ipcvttph_epi8 (hxi, m8, hxh);
+ hxi = _mm_maskz_ipcvttph_epi8 (m8, hxh);
- hxi = _mm_ipcvttph_epu16 (hxh);
- hxi = _mm_mask_ipcvttph_epu16 (hxi, m8, hxh);
- hxi = _mm_maskz_ipcvttph_epu16 (m8, hxh);
+ hxi = _mm_ipcvttph_epu8 (hxh);
+ hxi = _mm_mask_ipcvttph_epu8 (hxi, m8, hxh);
+ hxi = _mm_maskz_ipcvttph_epu8 (m8, hxh);
- hxi = _mm_ipcvtps_epi32 (hx);
- hxi = _mm_mask_ipcvtps_epi32 (hxi, m8, hx);
- hxi = _mm_maskz_ipcvtps_epi32 (m8, hx);
+ hxi = _mm_ipcvtps_epi8 (hx);
+ hxi = _mm_mask_ipcvtps_epi8 (hxi, m8, hx);
+ hxi = _mm_maskz_ipcvtps_epi8 (m8, hx);
- hxi = _mm_ipcvtps_epu32 (hx);
- hxi = _mm_mask_ipcvtps_epu32 (hxi, m8, hx);
- hxi = _mm_maskz_ipcvtps_epu32 (m8, hx);
+ hxi = _mm_ipcvtps_epu8 (hx);
+ hxi = _mm_mask_ipcvtps_epu8 (hxi, m8, hx);
+ hxi = _mm_maskz_ipcvtps_epu8 (m8, hx);
- hxi = _mm_ipcvttps_epi32 (hx);
- hxi = _mm_mask_ipcvttps_epi32 (hxi, m8, hx);
- hxi = _mm_maskz_ipcvttps_epi32 (m8, hx);
+ hxi = _mm_ipcvttps_epi8 (hx);
+ hxi = _mm_mask_ipcvttps_epi8 (hxi, m8, hx);
+ hxi = _mm_maskz_ipcvttps_epi8 (m8, hx);
- hxi = _mm_ipcvttps_epu32 (hx);
- hxi = _mm_mask_ipcvttps_epu32 (hxi, m8, hx);
- hxi = _mm_maskz_ipcvttps_epu32 (m8, hx);
+ hxi = _mm_ipcvttps_epu8 (hx);
+ hxi = _mm_mask_ipcvttps_epu8 (hxi, m8, hx);
+ hxi = _mm_maskz_ipcvttps_epu8 (m8, hx);
- hxi = _mm_ipcvtbf16_epi16 (hxbh);
- hxi = _mm_mask_ipcvtbf16_epi16 (hxi, m8, hxbh);
- hxi = _mm_maskz_ipcvtbf16_epi16 (m8, hxbh);
+ hxi = _mm_ipcvtbf16_epi8 (hxbh);
+ hxi = _mm_mask_ipcvtbf16_epi8 (hxi, m8, hxbh);
+ hxi = _mm_maskz_ipcvtbf16_epi8 (m8, hxbh);
- hxi = _mm_ipcvtbf16_epu16 (hxbh);
- hxi = _mm_mask_ipcvtbf16_epu16 (hxi, m8, hxbh);
- hxi = _mm_maskz_ipcvtbf16_epu16 (m8, hxbh);
+ hxi = _mm_ipcvtbf16_epu8 (hxbh);
+ hxi = _mm_mask_ipcvtbf16_epu8 (hxi, m8, hxbh);
+ hxi = _mm_maskz_ipcvtbf16_epu8 (m8, hxbh);
- hxi = _mm_ipcvttbf16_epi16 (hxbh);
- hxi = _mm_mask_ipcvttbf16_epi16 (hxi, m8, hxbh);
- hxi = _mm_maskz_ipcvttbf16_epi16 (m8, hxbh);
+ hxi = _mm_ipcvttbf16_epi8 (hxbh);
+ hxi = _mm_mask_ipcvttbf16_epi8 (hxi, m8, hxbh);
+ hxi = _mm_maskz_ipcvttbf16_epi8 (m8, hxbh);
- hxi = _mm_ipcvttbf16_epu16 (hxbh);
- hxi = _mm_mask_ipcvttbf16_epu16 (hxi, m8, hxbh);
- hxi = _mm_maskz_ipcvttbf16_epu16 (m8, hxbh);
+ hxi = _mm_ipcvttbf16_epu8 (hxbh);
+ hxi = _mm_mask_ipcvttbf16_epu8 (hxi, m8, hxbh);
+ hxi = _mm_maskz_ipcvttbf16_epu8 (m8, hxbh);
hxi = _mm256_cvtts_roundpd_epi32 (xd, 8);
hxi = _mm256_mask_cvtts_roundpd_epi32 (hxi, m8, xd, 8);
test_3 (_mm_mask_cmp_pbh_mask, __mmask8, __mmask8, __m128bh, __m128bh, 1)
/* avx10_2-512satcvtintrin.h */
-test_1 (_mm512_ipcvt_roundph_epi16, __m512i, __m512h, 8)
-test_1 (_mm512_ipcvt_roundph_epu16, __m512i, __m512h, 8)
-test_1 (_mm512_ipcvt_roundps_epi32, __m512i, __m512, 8)
-test_1 (_mm512_ipcvt_roundps_epu32, __m512i, __m512, 8)
-test_1 (_mm512_ipcvtt_roundph_epi16, __m512i, __m512h, 8)
-test_1 (_mm512_ipcvtt_roundph_epu16, __m512i, __m512h, 8)
-test_1 (_mm512_ipcvtt_roundps_epi32, __m512i, __m512, 8)
-test_1 (_mm512_ipcvtt_roundps_epu32, __m512i, __m512, 8)
-test_2 (_mm512_maskz_ipcvt_roundph_epi16, __m512i, __mmask32, __m512h, 8)
-test_2 (_mm512_maskz_ipcvt_roundph_epu16, __m512i, __mmask32, __m512h, 8)
-test_2 (_mm512_maskz_ipcvt_roundps_epi32, __m512i, __mmask16, __m512, 8)
-test_2 (_mm512_maskz_ipcvt_roundps_epu32, __m512i, __mmask16, __m512, 8)
-test_2 (_mm512_maskz_ipcvtt_roundph_epi16, __m512i, __mmask32, __m512h, 8)
-test_2 (_mm512_maskz_ipcvtt_roundph_epu16, __m512i, __mmask32, __m512h, 8)
-test_2 (_mm512_maskz_ipcvtt_roundps_epi32, __m512i, __mmask16, __m512, 8)
-test_2 (_mm512_maskz_ipcvtt_roundps_epu32, __m512i, __mmask16, __m512, 8)
-test_3 (_mm512_mask_ipcvt_roundph_epi16, __m512i, __m512i, __mmask32, __m512h, 8)
-test_3 (_mm512_mask_ipcvt_roundph_epu16, __m512i, __m512i, __mmask32, __m512h, 8)
-test_3 (_mm512_mask_ipcvt_roundps_epi32, __m512i, __m512i, __mmask16, __m512, 8)
-test_3 (_mm512_mask_ipcvt_roundps_epu32, __m512i, __m512i, __mmask16, __m512, 8)
-test_3 (_mm512_mask_ipcvtt_roundph_epi16, __m512i, __m512i, __mmask32, __m512h, 8)
-test_3 (_mm512_mask_ipcvtt_roundph_epu16, __m512i, __m512i, __mmask32, __m512h, 8)
-test_3 (_mm512_mask_ipcvtt_roundps_epi32, __m512i, __m512i, __mmask16, __m512, 8)
-test_3 (_mm512_mask_ipcvtt_roundps_epu32, __m512i, __m512i, __mmask16, __m512, 8)
+test_1 (_mm512_ipcvt_roundph_epi8, __m512i, __m512h, 8)
+test_1 (_mm512_ipcvt_roundph_epu8, __m512i, __m512h, 8)
+test_1 (_mm512_ipcvt_roundps_epi8, __m512i, __m512, 8)
+test_1 (_mm512_ipcvt_roundps_epu8, __m512i, __m512, 8)
+test_1 (_mm512_ipcvtt_roundph_epi8, __m512i, __m512h, 8)
+test_1 (_mm512_ipcvtt_roundph_epu8, __m512i, __m512h, 8)
+test_1 (_mm512_ipcvtt_roundps_epi8, __m512i, __m512, 8)
+test_1 (_mm512_ipcvtt_roundps_epu8, __m512i, __m512, 8)
+test_2 (_mm512_maskz_ipcvt_roundph_epi8, __m512i, __mmask32, __m512h, 8)
+test_2 (_mm512_maskz_ipcvt_roundph_epu8, __m512i, __mmask32, __m512h, 8)
+test_2 (_mm512_maskz_ipcvt_roundps_epi8, __m512i, __mmask16, __m512, 8)
+test_2 (_mm512_maskz_ipcvt_roundps_epu8, __m512i, __mmask16, __m512, 8)
+test_2 (_mm512_maskz_ipcvtt_roundph_epi8, __m512i, __mmask32, __m512h, 8)
+test_2 (_mm512_maskz_ipcvtt_roundph_epu8, __m512i, __mmask32, __m512h, 8)
+test_2 (_mm512_maskz_ipcvtt_roundps_epi8, __m512i, __mmask16, __m512, 8)
+test_2 (_mm512_maskz_ipcvtt_roundps_epu8, __m512i, __mmask16, __m512, 8)
+test_3 (_mm512_mask_ipcvt_roundph_epi8, __m512i, __m512i, __mmask32, __m512h, 8)
+test_3 (_mm512_mask_ipcvt_roundph_epu8, __m512i, __m512i, __mmask32, __m512h, 8)
+test_3 (_mm512_mask_ipcvt_roundps_epi8, __m512i, __m512i, __mmask16, __m512, 8)
+test_3 (_mm512_mask_ipcvt_roundps_epu8, __m512i, __m512i, __mmask16, __m512, 8)
+test_3 (_mm512_mask_ipcvtt_roundph_epi8, __m512i, __m512i, __mmask32, __m512h, 8)
+test_3 (_mm512_mask_ipcvtt_roundph_epu8, __m512i, __m512i, __mmask32, __m512h, 8)
+test_3 (_mm512_mask_ipcvtt_roundps_epi8, __m512i, __m512i, __mmask16, __m512, 8)
+test_3 (_mm512_mask_ipcvtt_roundps_epu8, __m512i, __m512i, __mmask16, __m512, 8)
test_1 (_mm512_cvtts_roundpd_epi32, __m256i, __m512d, 8)
test_2 (_mm512_maskz_cvtts_roundpd_epi32, __m256i, __mmask8, __m512d, 8)
test_3 (_mm512_mask_cvtts_roundpd_epi32, __m256i, __m256i, __mmask8, __m512d, 8)
test_3 (_mm512_mask_cvtts_roundps_epu64, __m512i, __m512i, __mmask8, __m256, 8)
/* avx10_2satcvtintrin.h */
-test_1 (_mm256_ipcvt_roundph_epi16, __m256i, __m256h, 8)
-test_1 (_mm256_ipcvt_roundph_epu16, __m256i, __m256h, 8)
-test_1 (_mm256_ipcvt_roundps_epi32, __m256i, __m256, 8)
-test_1 (_mm256_ipcvt_roundps_epu32, __m256i, __m256, 8)
-test_1 (_mm256_ipcvtt_roundph_epi16, __m256i, __m256h, 8)
-test_1 (_mm256_ipcvtt_roundph_epu16, __m256i, __m256h, 8)
-test_1 (_mm256_ipcvtt_roundps_epi32, __m256i, __m256, 8)
-test_1 (_mm256_ipcvtt_roundps_epu32, __m256i, __m256, 8)
-test_2 (_mm256_maskz_ipcvt_roundph_epi16, __m256i, __mmask16, __m256h, 8)
-test_2 (_mm256_maskz_ipcvt_roundph_epu16, __m256i, __mmask16, __m256h, 8)
-test_2 (_mm256_maskz_ipcvt_roundps_epi32, __m256i, __mmask8, __m256, 8)
-test_2 (_mm256_maskz_ipcvt_roundps_epu32, __m256i, __mmask8, __m256, 8)
-test_2 (_mm256_maskz_ipcvtt_roundph_epi16, __m256i, __mmask16, __m256h, 8)
-test_2 (_mm256_maskz_ipcvtt_roundph_epu16, __m256i, __mmask16, __m256h, 8)
-test_2 (_mm256_maskz_ipcvtt_roundps_epi32, __m256i, __mmask8, __m256, 8)
-test_2 (_mm256_maskz_ipcvtt_roundps_epu32, __m256i, __mmask8, __m256, 8)
-test_3 (_mm256_mask_ipcvt_roundph_epi16, __m256i, __m256i, __mmask16, __m256h, 8)
-test_3 (_mm256_mask_ipcvt_roundph_epu16, __m256i, __m256i, __mmask16, __m256h, 8)
-test_3 (_mm256_mask_ipcvt_roundps_epi32, __m256i, __m256i, __mmask8, __m256, 8)
-test_3 (_mm256_mask_ipcvt_roundps_epu32, __m256i, __m256i, __mmask8, __m256, 8)
-test_3 (_mm256_mask_ipcvtt_roundph_epi16, __m256i, __m256i, __mmask16, __m256h, 8)
-test_3 (_mm256_mask_ipcvtt_roundph_epu16, __m256i, __m256i, __mmask16, __m256h, 8)
-test_3 (_mm256_mask_ipcvtt_roundps_epi32, __m256i, __m256i, __mmask8, __m256, 8)
-test_3 (_mm256_mask_ipcvtt_roundps_epu32, __m256i, __m256i, __mmask8, __m256, 8)
+test_1 (_mm256_ipcvt_roundph_epi8, __m256i, __m256h, 8)
+test_1 (_mm256_ipcvt_roundph_epu8, __m256i, __m256h, 8)
+test_1 (_mm256_ipcvt_roundps_epi8, __m256i, __m256, 8)
+test_1 (_mm256_ipcvt_roundps_epu8, __m256i, __m256, 8)
+test_1 (_mm256_ipcvtt_roundph_epi8, __m256i, __m256h, 8)
+test_1 (_mm256_ipcvtt_roundph_epu8, __m256i, __m256h, 8)
+test_1 (_mm256_ipcvtt_roundps_epi8, __m256i, __m256, 8)
+test_1 (_mm256_ipcvtt_roundps_epu8, __m256i, __m256, 8)
+test_2 (_mm256_maskz_ipcvt_roundph_epi8, __m256i, __mmask16, __m256h, 8)
+test_2 (_mm256_maskz_ipcvt_roundph_epu8, __m256i, __mmask16, __m256h, 8)
+test_2 (_mm256_maskz_ipcvt_roundps_epi8, __m256i, __mmask8, __m256, 8)
+test_2 (_mm256_maskz_ipcvt_roundps_epu8, __m256i, __mmask8, __m256, 8)
+test_2 (_mm256_maskz_ipcvtt_roundph_epi8, __m256i, __mmask16, __m256h, 8)
+test_2 (_mm256_maskz_ipcvtt_roundph_epu8, __m256i, __mmask16, __m256h, 8)
+test_2 (_mm256_maskz_ipcvtt_roundps_epi8, __m256i, __mmask8, __m256, 8)
+test_2 (_mm256_maskz_ipcvtt_roundps_epu8, __m256i, __mmask8, __m256, 8)
+test_3 (_mm256_mask_ipcvt_roundph_epi8, __m256i, __m256i, __mmask16, __m256h, 8)
+test_3 (_mm256_mask_ipcvt_roundph_epu8, __m256i, __m256i, __mmask16, __m256h, 8)
+test_3 (_mm256_mask_ipcvt_roundps_epi8, __m256i, __m256i, __mmask8, __m256, 8)
+test_3 (_mm256_mask_ipcvt_roundps_epu8, __m256i, __m256i, __mmask8, __m256, 8)
+test_3 (_mm256_mask_ipcvtt_roundph_epi8, __m256i, __m256i, __mmask16, __m256h, 8)
+test_3 (_mm256_mask_ipcvtt_roundph_epu8, __m256i, __m256i, __mmask16, __m256h, 8)
+test_3 (_mm256_mask_ipcvtt_roundps_epi8, __m256i, __m256i, __mmask8, __m256, 8)
+test_3 (_mm256_mask_ipcvtt_roundps_epu8, __m256i, __m256i, __mmask8, __m256, 8)
test_1 (_mm256_cvtts_roundpd_epi32, __m128i, __m256d, 8)
test_2 (_mm256_maskz_cvtts_roundpd_epi32, __m128i, __mmask8, __m256d, 8)
test_3 (_mm256_mask_cvtts_roundpd_epi32, __m128i, __m128i, __mmask8, __m256d, 8)
test_3 (_mm_mask_cmp_pbh_mask, __mmask8, __mmask8, __m128bh, __m128bh, 1)
/* avx10_2-512satcvtintrin.h */
-test_1 (_mm512_ipcvt_roundph_epi16, __m512i, __m512h, 8)
-test_1 (_mm512_ipcvt_roundph_epu16, __m512i, __m512h, 8)
-test_1 (_mm512_ipcvt_roundps_epi32, __m512i, __m512, 8)
-test_1 (_mm512_ipcvt_roundps_epu32, __m512i, __m512, 8)
-test_1 (_mm512_ipcvtt_roundph_epi16, __m512i, __m512h, 8)
-test_1 (_mm512_ipcvtt_roundph_epu16, __m512i, __m512h, 8)
-test_1 (_mm512_ipcvtt_roundps_epi32, __m512i, __m512, 8)
-test_1 (_mm512_ipcvtt_roundps_epu32, __m512i, __m512, 8)
-test_2 (_mm512_maskz_ipcvt_roundph_epi16, __m512i, __mmask32, __m512h, 8)
-test_2 (_mm512_maskz_ipcvt_roundph_epu16, __m512i, __mmask32, __m512h, 8)
-test_2 (_mm512_maskz_ipcvt_roundps_epi32, __m512i, __mmask16, __m512, 8)
-test_2 (_mm512_maskz_ipcvt_roundps_epu32, __m512i, __mmask16, __m512, 8)
-test_2 (_mm512_maskz_ipcvtt_roundph_epi16, __m512i, __mmask32, __m512h, 8)
-test_2 (_mm512_maskz_ipcvtt_roundph_epu16, __m512i, __mmask32, __m512h, 8)
-test_2 (_mm512_maskz_ipcvtt_roundps_epi32, __m512i, __mmask16, __m512, 8)
-test_2 (_mm512_maskz_ipcvtt_roundps_epu32, __m512i, __mmask16, __m512, 8)
-test_3 (_mm512_mask_ipcvt_roundph_epi16, __m512i, __m512i, __mmask32, __m512h, 8)
-test_3 (_mm512_mask_ipcvt_roundph_epu16, __m512i, __m512i, __mmask32, __m512h, 8)
-test_3 (_mm512_mask_ipcvt_roundps_epi32, __m512i, __m512i, __mmask16, __m512, 8)
-test_3 (_mm512_mask_ipcvt_roundps_epu32, __m512i, __m512i, __mmask16, __m512, 8)
-test_3 (_mm512_mask_ipcvtt_roundph_epi16, __m512i, __m512i, __mmask32, __m512h, 8)
-test_3 (_mm512_mask_ipcvtt_roundph_epu16, __m512i, __m512i, __mmask32, __m512h, 8)
-test_3 (_mm512_mask_ipcvtt_roundps_epi32, __m512i, __m512i, __mmask16, __m512, 8)
-test_3 (_mm512_mask_ipcvtt_roundps_epu32, __m512i, __m512i, __mmask16, __m512, 8)
+test_1 (_mm512_ipcvt_roundph_epi8, __m512i, __m512h, 8)
+test_1 (_mm512_ipcvt_roundph_epu8, __m512i, __m512h, 8)
+test_1 (_mm512_ipcvt_roundps_epi8, __m512i, __m512, 8)
+test_1 (_mm512_ipcvt_roundps_epu8, __m512i, __m512, 8)
+test_1 (_mm512_ipcvtt_roundph_epi8, __m512i, __m512h, 8)
+test_1 (_mm512_ipcvtt_roundph_epu8, __m512i, __m512h, 8)
+test_1 (_mm512_ipcvtt_roundps_epi8, __m512i, __m512, 8)
+test_1 (_mm512_ipcvtt_roundps_epu8, __m512i, __m512, 8)
+test_2 (_mm512_maskz_ipcvt_roundph_epi8, __m512i, __mmask32, __m512h, 8)
+test_2 (_mm512_maskz_ipcvt_roundph_epu8, __m512i, __mmask32, __m512h, 8)
+test_2 (_mm512_maskz_ipcvt_roundps_epi8, __m512i, __mmask16, __m512, 8)
+test_2 (_mm512_maskz_ipcvt_roundps_epu8, __m512i, __mmask16, __m512, 8)
+test_2 (_mm512_maskz_ipcvtt_roundph_epi8, __m512i, __mmask32, __m512h, 8)
+test_2 (_mm512_maskz_ipcvtt_roundph_epu8, __m512i, __mmask32, __m512h, 8)
+test_2 (_mm512_maskz_ipcvtt_roundps_epi8, __m512i, __mmask16, __m512, 8)
+test_2 (_mm512_maskz_ipcvtt_roundps_epu8, __m512i, __mmask16, __m512, 8)
+test_3 (_mm512_mask_ipcvt_roundph_epi8, __m512i, __m512i, __mmask32, __m512h, 8)
+test_3 (_mm512_mask_ipcvt_roundph_epu8, __m512i, __m512i, __mmask32, __m512h, 8)
+test_3 (_mm512_mask_ipcvt_roundps_epi8, __m512i, __m512i, __mmask16, __m512, 8)
+test_3 (_mm512_mask_ipcvt_roundps_epu8, __m512i, __m512i, __mmask16, __m512, 8)
+test_3 (_mm512_mask_ipcvtt_roundph_epi8, __m512i, __m512i, __mmask32, __m512h, 8)
+test_3 (_mm512_mask_ipcvtt_roundph_epu8, __m512i, __m512i, __mmask32, __m512h, 8)
+test_3 (_mm512_mask_ipcvtt_roundps_epi8, __m512i, __m512i, __mmask16, __m512, 8)
+test_3 (_mm512_mask_ipcvtt_roundps_epu8, __m512i, __m512i, __mmask16, __m512, 8)
test_1 (_mm512_cvtts_roundpd_epi32, __m256i, __m512d, 8)
test_2 (_mm512_maskz_cvtts_roundpd_epi32, __m256i, __mmask8, __m512d, 8)
test_3 (_mm512_mask_cvtts_roundpd_epi32, __m256i, __m256i, __mmask8, __m512d, 8)
test_3 (_mm512_mask_cvtts_roundps_epu64, __m512i, __m512i, __mmask8, __m256, 8)
/* avx10_2satcvtintrin.h */
-test_1 (_mm256_ipcvt_roundph_epi16, __m256i, __m256h, 8)
-test_1 (_mm256_ipcvt_roundph_epu16, __m256i, __m256h, 8)
-test_1 (_mm256_ipcvt_roundps_epi32, __m256i, __m256, 8)
-test_1 (_mm256_ipcvt_roundps_epu32, __m256i, __m256, 8)
-test_1 (_mm256_ipcvtt_roundph_epi16, __m256i, __m256h, 8)
-test_1 (_mm256_ipcvtt_roundph_epu16, __m256i, __m256h, 8)
-test_1 (_mm256_ipcvtt_roundps_epi32, __m256i, __m256, 8)
-test_1 (_mm256_ipcvtt_roundps_epu32, __m256i, __m256, 8)
-test_2 (_mm256_maskz_ipcvt_roundph_epi16, __m256i, __mmask16, __m256h, 8)
-test_2 (_mm256_maskz_ipcvt_roundph_epu16, __m256i, __mmask16, __m256h, 8)
-test_2 (_mm256_maskz_ipcvt_roundps_epi32, __m256i, __mmask8, __m256, 8)
-test_2 (_mm256_maskz_ipcvt_roundps_epu32, __m256i, __mmask8, __m256, 8)
-test_2 (_mm256_maskz_ipcvtt_roundph_epi16, __m256i, __mmask16, __m256h, 8)
-test_2 (_mm256_maskz_ipcvtt_roundph_epu16, __m256i, __mmask16, __m256h, 8)
-test_2 (_mm256_maskz_ipcvtt_roundps_epi32, __m256i, __mmask8, __m256, 8)
-test_2 (_mm256_maskz_ipcvtt_roundps_epu32, __m256i, __mmask8, __m256, 8)
-test_3 (_mm256_mask_ipcvt_roundph_epi16, __m256i, __m256i, __mmask16, __m256h, 8)
-test_3 (_mm256_mask_ipcvt_roundph_epu16, __m256i, __m256i, __mmask16, __m256h, 8)
-test_3 (_mm256_mask_ipcvt_roundps_epi32, __m256i, __m256i, __mmask8, __m256, 8)
-test_3 (_mm256_mask_ipcvt_roundps_epu32, __m256i, __m256i, __mmask8, __m256, 8)
-test_3 (_mm256_mask_ipcvtt_roundph_epi16, __m256i, __m256i, __mmask16, __m256h, 8)
-test_3 (_mm256_mask_ipcvtt_roundph_epu16, __m256i, __m256i, __mmask16, __m256h, 8)
-test_3 (_mm256_mask_ipcvtt_roundps_epi32, __m256i, __m256i, __mmask8, __m256, 8)
-test_3 (_mm256_mask_ipcvtt_roundps_epu32, __m256i, __m256i, __mmask8, __m256, 8)
+test_1 (_mm256_ipcvt_roundph_epi8, __m256i, __m256h, 8)
+test_1 (_mm256_ipcvt_roundph_epu8, __m256i, __m256h, 8)
+test_1 (_mm256_ipcvt_roundps_epi8, __m256i, __m256, 8)
+test_1 (_mm256_ipcvt_roundps_epu8, __m256i, __m256, 8)
+test_1 (_mm256_ipcvtt_roundph_epi8, __m256i, __m256h, 8)
+test_1 (_mm256_ipcvtt_roundph_epu8, __m256i, __m256h, 8)
+test_1 (_mm256_ipcvtt_roundps_epi8, __m256i, __m256, 8)
+test_1 (_mm256_ipcvtt_roundps_epu8, __m256i, __m256, 8)
+test_2 (_mm256_maskz_ipcvt_roundph_epi8, __m256i, __mmask16, __m256h, 8)
+test_2 (_mm256_maskz_ipcvt_roundph_epu8, __m256i, __mmask16, __m256h, 8)
+test_2 (_mm256_maskz_ipcvt_roundps_epi8, __m256i, __mmask8, __m256, 8)
+test_2 (_mm256_maskz_ipcvt_roundps_epu8, __m256i, __mmask8, __m256, 8)
+test_2 (_mm256_maskz_ipcvtt_roundph_epi8, __m256i, __mmask16, __m256h, 8)
+test_2 (_mm256_maskz_ipcvtt_roundph_epu8, __m256i, __mmask16, __m256h, 8)
+test_2 (_mm256_maskz_ipcvtt_roundps_epi8, __m256i, __mmask8, __m256, 8)
+test_2 (_mm256_maskz_ipcvtt_roundps_epu8, __m256i, __mmask8, __m256, 8)
+test_3 (_mm256_mask_ipcvt_roundph_epi8, __m256i, __m256i, __mmask16, __m256h, 8)
+test_3 (_mm256_mask_ipcvt_roundph_epu8, __m256i, __m256i, __mmask16, __m256h, 8)
+test_3 (_mm256_mask_ipcvt_roundps_epi8, __m256i, __m256i, __mmask8, __m256, 8)
+test_3 (_mm256_mask_ipcvt_roundps_epu8, __m256i, __m256i, __mmask8, __m256, 8)
+test_3 (_mm256_mask_ipcvtt_roundph_epi8, __m256i, __m256i, __mmask16, __m256h, 8)
+test_3 (_mm256_mask_ipcvtt_roundph_epu8, __m256i, __m256i, __mmask16, __m256h, 8)
+test_3 (_mm256_mask_ipcvtt_roundps_epi8, __m256i, __m256i, __mmask8, __m256, 8)
+test_3 (_mm256_mask_ipcvtt_roundps_epu8, __m256i, __m256i, __mmask8, __m256, 8)
test_1 (_mm256_cvtts_roundpd_epi32, __m128i, __m256d, 8)
test_2 (_mm256_maskz_cvtts_roundpd_epi32, __m128i, __mmask8, __m256d, 8)
test_3 (_mm256_mask_cvtts_roundpd_epi32, __m128i, __m128i, __mmask8, __m256d, 8)