extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_cvtbiasph_pbf8 (__m512i __A, __m512h __B)
+_mm512_cvtbiasph_bf8 (__m512i __A, __m512h __B)
{
return (__m256i) __builtin_ia32_vcvtbiasph2bf8512_mask ((__v64qi) __A,
(__v32hf) __B,
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_cvtbiasph_pbf8 (__m256i __W, __mmask32 __U,
- __m512i __A, __m512h __B)
+_mm512_mask_cvtbiasph_bf8 (__m256i __W, __mmask32 __U,
+ __m512i __A, __m512h __B)
{
return (__m256i) __builtin_ia32_vcvtbiasph2bf8512_mask ((__v64qi) __A,
(__v32hf) __B,
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_cvtbiasph_pbf8 (__mmask32 __U, __m512i __A, __m512h __B)
+_mm512_maskz_cvtbiasph_bf8 (__mmask32 __U, __m512i __A, __m512h __B)
{
return (__m256i) __builtin_ia32_vcvtbiasph2bf8512_mask ((__v64qi) __A,
(__v32hf) __B,
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_cvtbiassph_pbf8 (__m512i __A, __m512h __B)
+_mm512_cvtbiassph_bf8 (__m512i __A, __m512h __B)
{
return (__m256i) __builtin_ia32_vcvtbiasph2bf8s512_mask ((__v64qi) __A,
(__v32hf) __B,
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_cvtbiassph_pbf8 (__m256i __W, __mmask32 __U,
- __m512i __A, __m512h __B)
+_mm512_mask_cvtbiassph_bf8 (__m256i __W, __mmask32 __U,
+ __m512i __A, __m512h __B)
{
return (__m256i) __builtin_ia32_vcvtbiasph2bf8s512_mask ((__v64qi) __A,
(__v32hf) __B,
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_cvtbiassph_pbf8 (__mmask32 __U, __m512i __A, __m512h __B)
+_mm512_maskz_cvtbiassph_bf8 (__mmask32 __U, __m512i __A, __m512h __B)
{
return (__m256i) __builtin_ia32_vcvtbiasph2bf8s512_mask ((__v64qi) __A,
(__v32hf) __B,
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_cvtbiasph_phf8 (__m512i __A, __m512h __B)
+_mm512_cvtbiasph_hf8 (__m512i __A, __m512h __B)
{
return (__m256i) __builtin_ia32_vcvtbiasph2hf8512_mask ((__v64qi) __A,
(__v32hf) __B,
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_cvtbiasph_phf8 (__m256i __W, __mmask32 __U, __m512i __A,
- __m512h __B)
+_mm512_mask_cvtbiasph_hf8 (__m256i __W, __mmask32 __U, __m512i __A,
+ __m512h __B)
{
return (__m256i) __builtin_ia32_vcvtbiasph2hf8512_mask ((__v64qi) __A,
(__v32hf) __B,
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_cvtbiasph_phf8 (__mmask32 __U, __m512i __A, __m512h __B)
+_mm512_maskz_cvtbiasph_hf8 (__mmask32 __U, __m512i __A, __m512h __B)
{
return (__m256i) __builtin_ia32_vcvtbiasph2hf8512_mask ((__v64qi) __A,
(__v32hf) __B,
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_cvtbiassph_phf8 (__m512i __A, __m512h __B)
+_mm512_cvtbiassph_hf8 (__m512i __A, __m512h __B)
{
return (__m256i) __builtin_ia32_vcvtbiasph2hf8s512_mask ((__v64qi) __A,
(__v32hf) __B,
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_cvtbiassph_phf8 (__m256i __W, __mmask32 __U,
- __m512i __A, __m512h __B)
+_mm512_mask_cvtbiassph_hf8 (__m256i __W, __mmask32 __U,
+ __m512i __A, __m512h __B)
{
return (__m256i) __builtin_ia32_vcvtbiasph2hf8s512_mask ((__v64qi) __A,
(__v32hf) __B,
extern __inline__ __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_cvtbiassph_phf8 (__mmask32 __U, __m512i __A, __m512h __B)
+_mm512_maskz_cvtbiassph_hf8 (__mmask32 __U, __m512i __A, __m512h __B)
{
return (__m256i) __builtin_ia32_vcvtbiasph2hf8s512_mask ((__v64qi) __A,
(__v32hf) __B,
extern __inline __m512h
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_cvtpbf8_ph (__m256i __A)
+_mm512_cvtbf8_ph (__m256i __A)
{
return (__m512h) _mm512_castsi512_ph ((__m512i) _mm512_slli_epi16 (
(__m512i) _mm512_cvtepi8_epi16 (__A), 8));
extern __inline __m512h
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_mask_cvtpbf8_ph (__m512h __S, __mmask16 __U, __m256i __A)
+_mm512_mask_cvtbf8_ph (__m512h __S, __mmask16 __U, __m256i __A)
{
return (__m512h) _mm512_castsi512_ph ((__m512i) _mm512_mask_slli_epi16 (
(__m512i) __S, __U, (__m512i) _mm512_cvtepi8_epi16 (__A), 8));
extern __inline __m512h
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm512_maskz_cvtpbf8_ph (__mmask16 __U, __m256i __A)
+_mm512_maskz_cvtbf8_ph (__mmask16 __U, __m256i __A)
{
return (__m512h) _mm512_castsi512_ph ((__m512i) _mm512_slli_epi16 (
(__m512i) _mm512_maskz_cvtepi8_epi16 (__U, __A), 8));
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_cvtbiasph_pbf8 (__m128i __A, __m128h __B)
+_mm_cvtbiasph_bf8 (__m128i __A, __m128h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2bf8128 ((__v16qi) __A,
(__v8hf) __B);
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_cvtbiasph_pbf8 (__m128i __W, __mmask8 __U, __m128i __A,
- __m128h __B)
+_mm_mask_cvtbiasph_bf8 (__m128i __W, __mmask8 __U, __m128i __A,
+ __m128h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2bf8128_mask ((__v16qi) __A,
(__v8hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_cvtbiasph_pbf8 (__mmask8 __U, __m128i __A, __m128h __B)
+_mm_maskz_cvtbiasph_bf8 (__mmask8 __U, __m128i __A, __m128h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2bf8128_mask ((__v16qi) __A,
(__v8hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_cvtbiasph_pbf8 (__m256i __A, __m256h __B)
+_mm256_cvtbiasph_bf8 (__m256i __A, __m256h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2bf8256_mask ((__v32qi) __A,
(__v16hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_cvtbiasph_pbf8 (__m128i __W, __mmask16 __U, __m256i __A,
- __m256h __B)
+_mm256_mask_cvtbiasph_bf8 (__m128i __W, __mmask16 __U, __m256i __A,
+ __m256h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2bf8256_mask ((__v32qi) __A,
(__v16hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_cvtbiasph_pbf8 (__mmask16 __U, __m256i __A, __m256h __B)
+_mm256_maskz_cvtbiasph_bf8 (__mmask16 __U, __m256i __A, __m256h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2bf8256_mask ((__v32qi) __A,
(__v16hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_cvtbiassph_pbf8 (__m128i __A, __m128h __B)
+_mm_cvtbiassph_bf8 (__m128i __A, __m128h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2bf8s128 ((__v16qi) __A,
(__v8hf) __B);
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_cvtbiassph_pbf8 (__m128i __W, __mmask8 __U,
- __m128i __A, __m128h __B)
+_mm_mask_cvtbiassph_bf8 (__m128i __W, __mmask8 __U,
+ __m128i __A, __m128h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2bf8s128_mask ((__v16qi) __A,
(__v8hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_cvtbiassph_pbf8 (__mmask8 __U, __m128i __A, __m128h __B)
+_mm_maskz_cvtbiassph_bf8 (__mmask8 __U, __m128i __A, __m128h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2bf8s128_mask ((__v16qi) __A,
(__v8hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_cvtbiassph_pbf8 (__m256i __A, __m256h __B)
+_mm256_cvtbiassph_bf8 (__m256i __A, __m256h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2bf8s256_mask ((__v32qi) __A,
(__v16hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_cvtbiassph_pbf8 (__m128i __W, __mmask16 __U,
- __m256i __A, __m256h __B)
+_mm256_mask_cvtbiassph_bf8 (__m128i __W, __mmask16 __U,
+ __m256i __A, __m256h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2bf8s256_mask ((__v32qi) __A,
(__v16hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_cvtbiassph_pbf8 (__mmask16 __U, __m256i __A, __m256h __B)
+_mm256_maskz_cvtbiassph_bf8 (__mmask16 __U, __m256i __A, __m256h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2bf8s256_mask ((__v32qi) __A,
(__v16hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_cvtbiasph_phf8 (__m128i __A, __m128h __B)
+_mm_cvtbiasph_hf8 (__m128i __A, __m128h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2hf8128 ((__v16qi) __A,
(__v8hf) __B);
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_cvtbiasph_phf8 (__m128i __W, __mmask8 __U, __m128i __A,
- __m128h __B)
+_mm_mask_cvtbiasph_hf8 (__m128i __W, __mmask8 __U, __m128i __A,
+ __m128h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2hf8128_mask ((__v16qi) __A,
(__v8hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_cvtbiasph_phf8 (__mmask8 __U, __m128i __A, __m128h __B)
+_mm_maskz_cvtbiasph_hf8 (__mmask8 __U, __m128i __A, __m128h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2hf8128_mask ((__v16qi) __A,
(__v8hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_cvtbiasph_phf8 (__m256i __A, __m256h __B)
+_mm256_cvtbiasph_hf8 (__m256i __A, __m256h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2hf8256_mask ((__v32qi) __A,
(__v16hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_cvtbiasph_phf8 (__m128i __W, __mmask16 __U,
- __m256i __A, __m256h __B)
+_mm256_mask_cvtbiasph_hf8 (__m128i __W, __mmask16 __U,
+ __m256i __A, __m256h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2hf8256_mask ((__v32qi) __A,
(__v16hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_cvtbiasph_phf8 (__mmask16 __U, __m256i __A, __m256h __B)
+_mm256_maskz_cvtbiasph_hf8 (__mmask16 __U, __m256i __A, __m256h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2hf8256_mask ((__v32qi) __A,
(__v16hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_cvtbiassph_phf8 (__m128i __A, __m128h __B)
+_mm_cvtbiassph_hf8 (__m128i __A, __m128h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2hf8s128 ((__v16qi) __A,
(__v8hf) __B);
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_cvtbiassph_phf8 (__m128i __W, __mmask8 __U,
- __m128i __A, __m128h __B)
+_mm_mask_cvtbiassph_hf8 (__m128i __W, __mmask8 __U,
+ __m128i __A, __m128h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2hf8s128_mask ((__v16qi) __A,
(__v8hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_cvtbiassph_phf8 (__mmask8 __U, __m128i __A, __m128h __B)
+_mm_maskz_cvtbiassph_hf8 (__mmask8 __U, __m128i __A, __m128h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2hf8s128_mask ((__v16qi) __A,
(__v8hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_cvtbiassph_phf8 (__m256i __A, __m256h __B)
+_mm256_cvtbiassph_hf8 (__m256i __A, __m256h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2hf8s256_mask ((__v32qi) __A,
(__v16hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_cvtbiassph_phf8 (__m128i __W, __mmask16 __U,
- __m256i __A, __m256h __B)
+_mm256_mask_cvtbiassph_hf8 (__m128i __W, __mmask16 __U,
+ __m256i __A, __m256h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2hf8s256_mask ((__v32qi) __A,
(__v16hf) __B,
extern __inline__ __m128i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_cvtbiassph_phf8 (__mmask16 __U, __m256i __A, __m256h __B)
+_mm256_maskz_cvtbiassph_hf8 (__mmask16 __U, __m256i __A, __m256h __B)
{
return (__m128i) __builtin_ia32_vcvtbiasph2hf8s256_mask ((__v32qi) __A,
(__v16hf) __B,
extern __inline __m128h
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_cvtpbf8_ph (__m128i __A)
+_mm_cvtbf8_ph (__m128i __A)
{
return (__m128h) _mm_castsi128_ph ((__m128i) _mm_slli_epi16 (
(__m128i) _mm_cvtepi8_epi16 (__A), 8));
extern __inline __m128h
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_mask_cvtpbf8_ph (__m128h __S, __mmask8 __U, __m128i __A)
+_mm_mask_cvtbf8_ph (__m128h __S, __mmask8 __U, __m128i __A)
{
return (__m128h) _mm_castsi128_ph ((__m128i) _mm_mask_slli_epi16 (
(__m128i) __S, __U, (__m128i) _mm_cvtepi8_epi16 (__A), 8));
extern __inline __m128h
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm_maskz_cvtpbf8_ph (__mmask8 __U, __m128i __A)
+_mm_maskz_cvtbf8_ph (__mmask8 __U, __m128i __A)
{
return (__m128h) _mm_castsi128_ph ((__m128i) _mm_slli_epi16 (
(__m128i) _mm_maskz_cvtepi8_epi16 (__U, __A), 8));
extern __inline __m256h
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_cvtpbf8_ph (__m128i __A)
+_mm256_cvtbf8_ph (__m128i __A)
{
return (__m256h) _mm256_castsi256_ph ((__m256i) _mm256_slli_epi16 (
(__m256i) _mm256_cvtepi8_epi16 (__A), 8));
extern __inline __m256h
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_mask_cvtpbf8_ph (__m256h __S, __mmask8 __U, __m128i __A)
+_mm256_mask_cvtbf8_ph (__m256h __S, __mmask8 __U, __m128i __A)
{
return (__m256h) _mm256_castsi256_ph ((__m256i) _mm256_mask_slli_epi16 (
(__m256i) __S, __U, (__m256i) _mm256_cvtepi8_epi16 (__A), 8));
extern __inline __m256h
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
-_mm256_maskz_cvtpbf8_ph (__mmask8 __U, __m128i __A)
+_mm256_maskz_cvtbf8_ph (__mmask8 __U, __m128i __A)
{
return (__m256h) _mm256_castsi256_ph ((__m256i) _mm256_slli_epi16 (
(__m256i) _mm256_maskz_cvtepi8_epi16 (__U, __A), 8));
void extern
avx10_2_512_vcvtbiasph2bf8_test (void)
{
- x256i = _mm512_cvtbiasph_pbf8 (x512i, x512h);
- x256i = _mm512_mask_cvtbiasph_pbf8 (x256i, m32, x512i, x512h);
- x256i = _mm512_maskz_cvtbiasph_pbf8 (m32, x512i, x512h);
+ x256i = _mm512_cvtbiasph_bf8 (x512i, x512h);
+ x256i = _mm512_mask_cvtbiasph_bf8 (x256i, m32, x512i, x512h);
+ x256i = _mm512_maskz_cvtbiasph_bf8 (m32, x512i, x512h);
}
void extern
avx10_2_512_vcvtbiasph2bf8s_test (void)
{
- x256i = _mm512_cvtbiassph_pbf8 (x512i, x512h);
- x256i = _mm512_mask_cvtbiassph_pbf8 (x256i, m32, x512i, x512h);
- x256i = _mm512_maskz_cvtbiassph_pbf8 (m32, x512i, x512h);
+ x256i = _mm512_cvtbiassph_bf8 (x512i, x512h);
+ x256i = _mm512_mask_cvtbiassph_bf8 (x256i, m32, x512i, x512h);
+ x256i = _mm512_maskz_cvtbiassph_bf8 (m32, x512i, x512h);
}
void extern
avx10_2_512_vcvtbiasph2hf8_test (void)
{
- x256i = _mm512_cvtbiasph_phf8 (x512i, x512h);
- x256i = _mm512_mask_cvtbiasph_phf8 (x256i, m32, x512i, x512h);
- x256i = _mm512_maskz_cvtbiasph_phf8 (m32, x512i, x512h);
+ x256i = _mm512_cvtbiasph_hf8 (x512i, x512h);
+ x256i = _mm512_mask_cvtbiasph_hf8 (x256i, m32, x512i, x512h);
+ x256i = _mm512_maskz_cvtbiasph_hf8 (m32, x512i, x512h);
}
void extern
avx10_2_512_vcvtbiasph2hf8s_test (void)
{
- x256i = _mm512_cvtbiassph_phf8 (x512i, x512h);
- x256i = _mm512_mask_cvtbiassph_phf8 (x256i, m32, x512i, x512h);
- x256i = _mm512_maskz_cvtbiassph_phf8 (m32, x512i, x512h);
+ x256i = _mm512_cvtbiassph_hf8 (x512i, x512h);
+ x256i = _mm512_mask_cvtbiassph_hf8 (x256i, m32, x512i, x512h);
+ x256i = _mm512_maskz_cvtbiassph_hf8 (m32, x512i, x512h);
}
void extern
void extern
avx10_2_512_cvtbf8_fp16_test (void)
{
- y = _mm512_cvtpbf8_ph (z1);
- y = _mm512_mask_cvtpbf8_ph (z, m16, z1);
- y = _mm512_maskz_cvtpbf8_ph (m16, z1);
+ y = _mm512_cvtbf8_ph (z1);
+ y = _mm512_mask_cvtbf8_ph (z, m16, z1);
+ y = _mm512_maskz_cvtbf8_ph (m16, z1);
}
CALC (res_ref, src1.a, src2.a);
- res1.x = INTRINSIC (_cvtbiasph_pbf8) (src1.x, src2.x);
+ res1.x = INTRINSIC (_cvtbiasph_bf8) (src1.x, src2.x);
if (UNION_CHECK (AVX512F_LEN_HALF, i_b) (res1, res_ref))
abort ();
- res2.x = INTRINSIC (_mask_cvtbiasph_pbf8) (res2.x, mask, src1.x, src2.x);
+ res2.x = INTRINSIC (_mask_cvtbiasph_bf8) (res2.x, mask, src1.x, src2.x);
MASK_MERGE (i_b) (res_ref, mask, SIZE);
if (UNION_CHECK (AVX512F_LEN_HALF, i_b) (res2, res_ref))
abort ();
- res3.x = INTRINSIC (_maskz_cvtbiasph_pbf8) (mask, src1.x, src2.x);
+ res3.x = INTRINSIC (_maskz_cvtbiasph_bf8) (mask, src1.x, src2.x);
MASK_ZERO (i_b) (res_ref, mask, SIZE);
if (UNION_CHECK (AVX512F_LEN_HALF, i_b) (res3, res_ref))
abort ();
CALC (res_ref, src1.a, src2.a);
- res1.x = INTRINSIC (_cvtbiassph_pbf8) (src1.x, src2.x);
+ res1.x = INTRINSIC (_cvtbiassph_bf8) (src1.x, src2.x);
if (UNION_CHECK (AVX512F_LEN_HALF, i_b) (res1, res_ref))
abort ();
- res2.x = INTRINSIC (_mask_cvtbiassph_pbf8) (res2.x, mask, src1.x, src2.x);
+ res2.x = INTRINSIC (_mask_cvtbiassph_bf8) (res2.x, mask, src1.x, src2.x);
MASK_MERGE (i_b) (res_ref, mask, SIZE);
if (UNION_CHECK (AVX512F_LEN_HALF, i_b) (res2, res_ref))
abort ();
- res3.x = INTRINSIC (_maskz_cvtbiassph_pbf8) (mask, src1.x, src2.x);
+ res3.x = INTRINSIC (_maskz_cvtbiassph_bf8) (mask, src1.x, src2.x);
MASK_ZERO (i_b) (res_ref, mask, SIZE);
if (UNION_CHECK (AVX512F_LEN_HALF, i_b) (res3, res_ref))
abort ();
CALC (res_ref, src1.a, src2.a);
- res1.x = INTRINSIC (_cvtbiasph_phf8) (src1.x, src2.x);
+ res1.x = INTRINSIC (_cvtbiasph_hf8) (src1.x, src2.x);
if (UNION_CHECK (AVX512F_LEN_HALF, i_b) (res1, res_ref))
abort ();
- res2.x = INTRINSIC (_mask_cvtbiasph_phf8) (res2.x, mask, src1.x, src2.x);
+ res2.x = INTRINSIC (_mask_cvtbiasph_hf8) (res2.x, mask, src1.x, src2.x);
MASK_MERGE (i_b) (res_ref, mask, SIZE);
if (UNION_CHECK (AVX512F_LEN_HALF, i_b) (res2, res_ref))
abort ();
- res3.x = INTRINSIC (_maskz_cvtbiasph_phf8) (mask, src1.x, src2.x);
+ res3.x = INTRINSIC (_maskz_cvtbiasph_hf8) (mask, src1.x, src2.x);
MASK_ZERO (i_b) (res_ref, mask, SIZE);
if (UNION_CHECK (AVX512F_LEN_HALF, i_b) (res3, res_ref))
abort ();
CALC (res_ref, src1.a, src2.a);
- res1.x = INTRINSIC (_cvtbiassph_phf8) (src1.x, src2.x);
+ res1.x = INTRINSIC (_cvtbiassph_hf8) (src1.x, src2.x);
if (UNION_CHECK (AVX512F_LEN_HALF, i_b) (res1, res_ref))
abort ();
- res2.x = INTRINSIC (_mask_cvtbiassph_phf8) (res2.x, mask, src1.x, src2.x);
+ res2.x = INTRINSIC (_mask_cvtbiassph_hf8) (res2.x, mask, src1.x, src2.x);
MASK_MERGE (i_b) (res_ref, mask, SIZE);
if (UNION_CHECK (AVX512F_LEN_HALF, i_b) (res2, res_ref))
abort ();
- res3.x = INTRINSIC (_maskz_cvtbiassph_phf8) (mask, src1.x, src2.x);
+ res3.x = INTRINSIC (_maskz_cvtbiassph_hf8) (mask, src1.x, src2.x);
MASK_ZERO (i_b) (res_ref, mask, SIZE);
if (UNION_CHECK (AVX512F_LEN_HALF, i_b) (res3, res_ref))
abort ();
void extern
avx10_2_vcvtbiasph2bf8_test (void)
{
- x128i = _mm_cvtbiasph_pbf8 (x128i, x128h);
- x128i = _mm_mask_cvtbiasph_pbf8 (x128i, m8, x128i, x128h);
- x128i = _mm_maskz_cvtbiasph_pbf8 (m8, x128i, x128h);
+ x128i = _mm_cvtbiasph_bf8 (x128i, x128h);
+ x128i = _mm_mask_cvtbiasph_bf8 (x128i, m8, x128i, x128h);
+ x128i = _mm_maskz_cvtbiasph_bf8 (m8, x128i, x128h);
- x128i = _mm256_cvtbiasph_pbf8 (x256i, x256h);
- x128i = _mm256_mask_cvtbiasph_pbf8 (x128i, m16, x256i, x256h);
- x128i = _mm256_maskz_cvtbiasph_pbf8 (m16, x256i, x256h);
+ x128i = _mm256_cvtbiasph_bf8 (x256i, x256h);
+ x128i = _mm256_mask_cvtbiasph_bf8 (x128i, m16, x256i, x256h);
+ x128i = _mm256_maskz_cvtbiasph_bf8 (m16, x256i, x256h);
}
void extern
avx10_2_vcvtbiasph2bf8s_test (void)
{
- x128i = _mm_cvtbiassph_pbf8 (x128i, x128h);
- x128i = _mm_mask_cvtbiassph_pbf8 (x128i, m8, x128i, x128h);
- x128i = _mm_maskz_cvtbiassph_pbf8 (m8, x128i, x128h);
+ x128i = _mm_cvtbiassph_bf8 (x128i, x128h);
+ x128i = _mm_mask_cvtbiassph_bf8 (x128i, m8, x128i, x128h);
+ x128i = _mm_maskz_cvtbiassph_bf8 (m8, x128i, x128h);
- x128i = _mm256_cvtbiassph_pbf8 (x256i, x256h);
- x128i = _mm256_mask_cvtbiassph_pbf8 (x128i, m16, x256i, x256h);
- x128i = _mm256_maskz_cvtbiassph_pbf8 (m16, x256i, x256h);
+ x128i = _mm256_cvtbiassph_bf8 (x256i, x256h);
+ x128i = _mm256_mask_cvtbiassph_bf8 (x128i, m16, x256i, x256h);
+ x128i = _mm256_maskz_cvtbiassph_bf8 (m16, x256i, x256h);
}
void extern
avx10_2_vcvtbiasph2hf8_test (void)
{
- x128i = _mm_cvtbiasph_phf8 (x128i, x128h);
- x128i = _mm_mask_cvtbiasph_phf8 (x128i, m8, x128i, x128h);
- x128i = _mm_maskz_cvtbiasph_phf8 (m8, x128i, x128h);
+ x128i = _mm_cvtbiasph_hf8 (x128i, x128h);
+ x128i = _mm_mask_cvtbiasph_hf8 (x128i, m8, x128i, x128h);
+ x128i = _mm_maskz_cvtbiasph_hf8 (m8, x128i, x128h);
- x128i = _mm256_cvtbiasph_phf8 (x256i, x256h);
- x128i = _mm256_mask_cvtbiasph_phf8 (x128i, m16, x256i, x256h);
- x128i = _mm256_maskz_cvtbiasph_phf8 (m16, x256i, x256h);
+ x128i = _mm256_cvtbiasph_hf8 (x256i, x256h);
+ x128i = _mm256_mask_cvtbiasph_hf8 (x128i, m16, x256i, x256h);
+ x128i = _mm256_maskz_cvtbiasph_hf8 (m16, x256i, x256h);
}
void extern
avx10_2_vcvtbiasph2hf8s_test (void)
{
- x128i = _mm_cvtbiassph_phf8 (x128i, x128h);
- x128i = _mm_mask_cvtbiassph_phf8 (x128i, m8, x128i, x128h);
- x128i = _mm_maskz_cvtbiassph_phf8 (m8, x128i, x128h);
+ x128i = _mm_cvtbiassph_hf8 (x128i, x128h);
+ x128i = _mm_mask_cvtbiassph_hf8 (x128i, m8, x128i, x128h);
+ x128i = _mm_maskz_cvtbiassph_hf8 (m8, x128i, x128h);
- x128i = _mm256_cvtbiassph_phf8 (x256i, x256h);
- x128i = _mm256_mask_cvtbiassph_phf8 (x128i, m16, x256i, x256h);
- x128i = _mm256_maskz_cvtbiassph_phf8 (m16, x256i, x256h);
+ x128i = _mm256_cvtbiassph_hf8 (x256i, x256h);
+ x128i = _mm256_mask_cvtbiassph_hf8 (x128i, m16, x256i, x256h);
+ x128i = _mm256_maskz_cvtbiassph_hf8 (m16, x256i, x256h);
}
void extern
void extern
avx10_2_cvtbf8_fp16_test (void)
{
- y = _mm_cvtpbf8_ph (z3);
- y = _mm_mask_cvtpbf8_ph (z, m8, z3);
- y = _mm_maskz_cvtpbf8_ph (m8, z3);
+ y = _mm_cvtbf8_ph (z3);
+ y = _mm_mask_cvtbf8_ph (z, m8, z3);
+ y = _mm_maskz_cvtbf8_ph (m8, z3);
- y2 = _mm256_cvtpbf8_ph (z3);
- y2 = _mm256_mask_cvtpbf8_ph (z2, m8, z3);
- y2 = _mm256_maskz_cvtpbf8_ph (m8, z3);
+ y2 = _mm256_cvtbf8_ph (z3);
+ y2 = _mm256_mask_cvtbf8_ph (z2, m8, z3);
+ y2 = _mm256_maskz_cvtbf8_ph (m8, z3);
}