i386: Add "s_" as Saturation for AVX10.2 SAT CVT Intrinsics.

This patch aims to add "s_" before intrinsic core name represent
saturation.

gcc/ChangeLog:

	* config/i386/avx10_2-512satcvtintrin.h: Add "s_" before
	intrinsics' core name.
	* config/i386/avx10_2satcvtintrin.h: Ditto.

gcc/testsuite/ChangeLog:

	* gcc.target/i386/avx10_2-512-satcvt-1.c: Modify intrinsic name.
	* gcc.target/i386/avx10_2-512-vcvtbf162ibs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvtbf162iubs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvtph2ibs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvtph2iubs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvtps2ibs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvtps2iubs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttbf162ibs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttbf162iubs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttpd2dqs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttpd2qqs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttpd2udqs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttpd2uqqs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttph2ibs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttph2iubs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttps2dqs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttps2ibs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttps2iubs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttps2qqs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttps2udqs-2.c: Ditto.
	* gcc.target/i386/avx10_2-512-vcvttps2uqqs-2.c: Ditto.
	* gcc.target/i386/avx10_2-satcvt-1.c: Ditto.
	* gcc.target/i386/avx10_2-vcvttsd2sis-2.c: Ditto.
	* gcc.target/i386/avx10_2-vcvttsd2usis-2.c: Ditto.
	* gcc.target/i386/avx10_2-vcvttss2sis-2.c: Ditto.
	* gcc.target/i386/avx10_2-vcvttss2usis-2.c: Ditto.
	* gcc.target/i386/sse-14.c: Ditto.
	* gcc.target/i386/sse-22.c: Ditto.
This commit is contained in:
Hu, Lin1 2025-03-18 16:50:24 +08:00
parent 2e7a92a68a
commit c3a211c932
30 changed files with 724 additions and 724 deletions

View file

@ -36,7 +36,7 @@
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvtbf16_epi8 (__m512bh __A)
_mm512_ipcvts_bf16_epi8 (__m512bh __A)
{
return
(__m512i) __builtin_ia32_cvtbf162ibs512_mask ((__v32bf) __A,
@ -47,7 +47,7 @@ _mm512_ipcvtbf16_epi8 (__m512bh __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvtbf16_epi8 (__m512i __W, __mmask32 __U, __m512bh __A)
_mm512_mask_ipcvts_bf16_epi8 (__m512i __W, __mmask32 __U, __m512bh __A)
{
return (__m512i) __builtin_ia32_cvtbf162ibs512_mask ((__v32bf) __A,
(__v32hi) __W,
@ -56,7 +56,7 @@ _mm512_mask_ipcvtbf16_epi8 (__m512i __W, __mmask32 __U, __m512bh __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvtbf16_epi8 (__mmask32 __U, __m512bh __A)
_mm512_maskz_ipcvts_bf16_epi8 (__mmask32 __U, __m512bh __A)
{
return
(__m512i) __builtin_ia32_cvtbf162ibs512_mask ((__v32bf) __A,
@ -67,7 +67,7 @@ _mm512_maskz_ipcvtbf16_epi8 (__mmask32 __U, __m512bh __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvtbf16_epu8 (__m512bh __A)
_mm512_ipcvts_bf16_epu8 (__m512bh __A)
{
return
(__m512i) __builtin_ia32_cvtbf162iubs512_mask ((__v32bf) __A,
@ -78,7 +78,7 @@ _mm512_ipcvtbf16_epu8 (__m512bh __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvtbf16_epu8 (__m512i __W, __mmask32 __U, __m512bh __A)
_mm512_mask_ipcvts_bf16_epu8 (__m512i __W, __mmask32 __U, __m512bh __A)
{
return (__m512i) __builtin_ia32_cvtbf162iubs512_mask ((__v32bf) __A,
(__v32hi) __W,
@ -87,7 +87,7 @@ _mm512_mask_ipcvtbf16_epu8 (__m512i __W, __mmask32 __U, __m512bh __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvtbf16_epu8 (__mmask32 __U, __m512bh __A)
_mm512_maskz_ipcvts_bf16_epu8 (__mmask32 __U, __m512bh __A)
{
return
(__m512i) __builtin_ia32_cvtbf162iubs512_mask ((__v32bf) __A,
@ -98,7 +98,7 @@ _mm512_maskz_ipcvtbf16_epu8 (__mmask32 __U, __m512bh __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvttbf16_epi8 (__m512bh __A)
_mm512_ipcvtts_bf16_epi8 (__m512bh __A)
{
return
(__m512i) __builtin_ia32_cvttbf162ibs512_mask ((__v32bf) __A,
@ -109,7 +109,7 @@ _mm512_ipcvttbf16_epi8 (__m512bh __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvttbf16_epi8 (__m512i __W, __mmask32 __U, __m512bh __A)
_mm512_mask_ipcvtts_bf16_epi8 (__m512i __W, __mmask32 __U, __m512bh __A)
{
return (__m512i) __builtin_ia32_cvttbf162ibs512_mask ((__v32bf) __A,
(__v32hi) __W,
@ -118,7 +118,7 @@ _mm512_mask_ipcvttbf16_epi8 (__m512i __W, __mmask32 __U, __m512bh __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvttbf16_epi8 (__mmask32 __U, __m512bh __A)
_mm512_maskz_ipcvtts_bf16_epi8 (__mmask32 __U, __m512bh __A)
{
return
(__m512i) __builtin_ia32_cvttbf162ibs512_mask ((__v32bf) __A,
@ -129,7 +129,7 @@ _mm512_maskz_ipcvttbf16_epi8 (__mmask32 __U, __m512bh __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvttbf16_epu8 (__m512bh __A)
_mm512_ipcvtts_bf16_epu8 (__m512bh __A)
{
return (__m512i)
__builtin_ia32_cvttbf162iubs512_mask ((__v32bf) __A,
@ -139,7 +139,7 @@ _mm512_ipcvttbf16_epu8 (__m512bh __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvttbf16_epu8 (__m512i __W, __mmask32 __U, __m512bh __A)
_mm512_mask_ipcvtts_bf16_epu8 (__m512i __W, __mmask32 __U, __m512bh __A)
{
return (__m512i) __builtin_ia32_cvttbf162iubs512_mask ((__v32bf) __A,
(__v32hi) __W,
@ -148,7 +148,7 @@ _mm512_mask_ipcvttbf16_epu8 (__m512i __W, __mmask32 __U, __m512bh __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvttbf16_epu8 (__mmask32 __U, __m512bh __A)
_mm512_maskz_ipcvtts_bf16_epu8 (__mmask32 __U, __m512bh __A)
{
return (__m512i)
__builtin_ia32_cvttbf162iubs512_mask ((__v32bf) __A,
@ -159,7 +159,7 @@ _mm512_maskz_ipcvttbf16_epu8 (__mmask32 __U, __m512bh __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvtph_epi8 (__m512h __A)
_mm512_ipcvts_ph_epi8 (__m512h __A)
{
return
(__m512i) __builtin_ia32_cvtph2ibs512_mask ((__v32hf) __A,
@ -170,7 +170,7 @@ _mm512_ipcvtph_epi8 (__m512h __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvtph_epi8 (__m512i __W, __mmask32 __U, __m512h __A)
_mm512_mask_ipcvts_ph_epi8 (__m512i __W, __mmask32 __U, __m512h __A)
{
return (__m512i) __builtin_ia32_cvtph2ibs512_mask ((__v32hf) __A,
(__v32hi) __W,
@ -179,7 +179,7 @@ _mm512_mask_ipcvtph_epi8 (__m512i __W, __mmask32 __U, __m512h __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvtph_epi8 (__mmask32 __U, __m512h __A)
_mm512_maskz_ipcvts_ph_epi8 (__mmask32 __U, __m512h __A)
{
return
(__m512i) __builtin_ia32_cvtph2ibs512_mask ((__v32hf) __A,
@ -190,7 +190,7 @@ _mm512_maskz_ipcvtph_epi8 (__mmask32 __U, __m512h __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvtph_epu8 (__m512h __A)
_mm512_ipcvts_ph_epu8 (__m512h __A)
{
return
(__m512i) __builtin_ia32_cvtph2iubs512_mask ((__v32hf) __A,
@ -201,7 +201,7 @@ _mm512_ipcvtph_epu8 (__m512h __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvtph_epu8 (__m512i __W, __mmask32 __U, __m512h __A)
_mm512_mask_ipcvts_ph_epu8 (__m512i __W, __mmask32 __U, __m512h __A)
{
return (__m512i) __builtin_ia32_cvtph2iubs512_mask ((__v32hf) __A,
(__v32hi) __W,
@ -210,7 +210,7 @@ _mm512_mask_ipcvtph_epu8 (__m512i __W, __mmask32 __U, __m512h __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvtph_epu8 (__mmask32 __U, __m512h __A)
_mm512_maskz_ipcvts_ph_epu8 (__mmask32 __U, __m512h __A)
{
return
(__m512i) __builtin_ia32_cvtph2iubs512_mask ((__v32hf) __A,
@ -221,7 +221,7 @@ _mm512_maskz_ipcvtph_epu8 (__mmask32 __U, __m512h __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvtps_epi8 (__m512 __A)
_mm512_ipcvts_ps_epi8 (__m512 __A)
{
return
(__m512i) __builtin_ia32_cvtps2ibs512_mask ((__v16sf) __A,
@ -232,7 +232,7 @@ _mm512_ipcvtps_epi8 (__m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvtps_epi8 (__m512i __W, __mmask16 __U, __m512 __A)
_mm512_mask_ipcvts_ps_epi8 (__m512i __W, __mmask16 __U, __m512 __A)
{
return (__m512i) __builtin_ia32_cvtps2ibs512_mask ((__v16sf) __A,
(__v16si) __W,
@ -241,7 +241,7 @@ _mm512_mask_ipcvtps_epi8 (__m512i __W, __mmask16 __U, __m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvtps_epi8 (__mmask16 __U, __m512 __A)
_mm512_maskz_ipcvts_ps_epi8 (__mmask16 __U, __m512 __A)
{
return
(__m512i) __builtin_ia32_cvtps2ibs512_mask ((__v16sf) __A,
@ -252,7 +252,7 @@ _mm512_maskz_ipcvtps_epi8 (__mmask16 __U, __m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvtps_epu8 (__m512 __A)
_mm512_ipcvts_ps_epu8 (__m512 __A)
{
return
(__m512i) __builtin_ia32_cvtps2iubs512_mask ((__v16sf) __A,
@ -263,7 +263,7 @@ _mm512_ipcvtps_epu8 (__m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvtps_epu8 (__m512i __W, __mmask16 __U, __m512 __A)
_mm512_mask_ipcvts_ps_epu8 (__m512i __W, __mmask16 __U, __m512 __A)
{
return (__m512i) __builtin_ia32_cvtps2iubs512_mask ((__v16sf) __A,
(__v16si) __W,
@ -272,7 +272,7 @@ _mm512_mask_ipcvtps_epu8 (__m512i __W, __mmask16 __U, __m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvtps_epu8 (__mmask16 __U, __m512 __A)
_mm512_maskz_ipcvts_ps_epu8 (__mmask16 __U, __m512 __A)
{
return
(__m512i) __builtin_ia32_cvtps2iubs512_mask ((__v16sf) __A,
@ -283,7 +283,7 @@ _mm512_maskz_ipcvtps_epu8 (__mmask16 __U, __m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvttph_epi8 (__m512h __A)
_mm512_ipcvtts_ph_epi8 (__m512h __A)
{
return (__m512i)
__builtin_ia32_cvttph2ibs512_mask ((__v32hf) __A,
@ -294,7 +294,7 @@ _mm512_ipcvttph_epi8 (__m512h __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvttph_epi8 (__m512i __W, __mmask32 __U, __m512h __A)
_mm512_mask_ipcvtts_ph_epi8 (__m512i __W, __mmask32 __U, __m512h __A)
{
return (__m512i) __builtin_ia32_cvttph2ibs512_mask ((__v32hf) __A,
(__v32hi) __W,
@ -303,7 +303,7 @@ _mm512_mask_ipcvttph_epi8 (__m512i __W, __mmask32 __U, __m512h __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvttph_epi8 (__mmask32 __U, __m512h __A)
_mm512_maskz_ipcvtts_ph_epi8 (__mmask32 __U, __m512h __A)
{
return
(__m512i) __builtin_ia32_cvttph2ibs512_mask ((__v32hf) __A,
@ -314,7 +314,7 @@ _mm512_maskz_ipcvttph_epi8 (__mmask32 __U, __m512h __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvttph_epu8 (__m512h __A)
_mm512_ipcvtts_ph_epu8 (__m512h __A)
{
return (__m512i)
__builtin_ia32_cvttph2iubs512_mask ((__v32hf) __A,
@ -325,7 +325,7 @@ _mm512_ipcvttph_epu8 (__m512h __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvttph_epu8 (__m512i __W, __mmask32 __U, __m512h __A)
_mm512_mask_ipcvtts_ph_epu8 (__m512i __W, __mmask32 __U, __m512h __A)
{
return (__m512i) __builtin_ia32_cvttph2iubs512_mask ((__v32hf) __A,
(__v32hi) __W,
@ -334,7 +334,7 @@ _mm512_mask_ipcvttph_epu8 (__m512i __W, __mmask32 __U, __m512h __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvttph_epu8 (__mmask32 __U, __m512h __A)
_mm512_maskz_ipcvtts_ph_epu8 (__mmask32 __U, __m512h __A)
{
return (__m512i)
__builtin_ia32_cvttph2iubs512_mask ((__v32hf) __A,
@ -345,7 +345,7 @@ _mm512_maskz_ipcvttph_epu8 (__mmask32 __U, __m512h __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvttps_epi8 (__m512 __A)
_mm512_ipcvtts_ps_epi8 (__m512 __A)
{
return (__m512i)
__builtin_ia32_cvttps2ibs512_mask ((__v16sf) __A,
@ -356,7 +356,7 @@ _mm512_ipcvttps_epi8 (__m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvttps_epi8 (__m512i __W, __mmask16 __U, __m512 __A)
_mm512_mask_ipcvtts_ps_epi8 (__m512i __W, __mmask16 __U, __m512 __A)
{
return (__m512i) __builtin_ia32_cvttps2ibs512_mask ((__v16sf) __A,
(__v16si) __W,
@ -365,7 +365,7 @@ _mm512_mask_ipcvttps_epi8 (__m512i __W, __mmask16 __U, __m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvttps_epi8 (__mmask16 __U, __m512 __A)
_mm512_maskz_ipcvtts_ps_epi8 (__mmask16 __U, __m512 __A)
{
return (__m512i)
__builtin_ia32_cvttps2ibs512_mask ((__v16sf) __A,
@ -376,7 +376,7 @@ _mm512_maskz_ipcvttps_epi8 (__mmask16 __U, __m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvttps_epu8 (__m512 __A)
_mm512_ipcvtts_ps_epu8 (__m512 __A)
{
return (__m512i)
__builtin_ia32_cvttps2iubs512_mask ((__v16sf) __A,
@ -387,7 +387,7 @@ _mm512_ipcvttps_epu8 (__m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvttps_epu8 (__m512i __W, __mmask16 __U, __m512 __A)
_mm512_mask_ipcvtts_ps_epu8 (__m512i __W, __mmask16 __U, __m512 __A)
{
return (__m512i) __builtin_ia32_cvttps2iubs512_mask ((__v16sf) __A,
(__v16si) __W,
@ -396,7 +396,7 @@ _mm512_mask_ipcvttps_epu8 (__m512i __W, __mmask16 __U, __m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvttps_epu8 (__mmask16 __U, __m512 __A)
_mm512_maskz_ipcvtts_ps_epu8 (__mmask16 __U, __m512 __A)
{
return (__m512i)
__builtin_ia32_cvttps2iubs512_mask ((__v16sf) __A,
@ -407,7 +407,7 @@ _mm512_maskz_ipcvttps_epu8 (__mmask16 __U, __m512 __A)
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_cvttspd_epi32 (__m512d __A)
_mm512_cvtts_pd_epi32 (__m512d __A)
{
return (__m256i)
__builtin_ia32_cvttpd2dqs512_mask ((__v8df) __A,
@ -418,7 +418,7 @@ _mm512_cvttspd_epi32 (__m512d __A)
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvttspd_epi32 (__m256i __W, __mmask8 __U, __m512d __A)
_mm512_mask_cvtts_pd_epi32 (__m256i __W, __mmask8 __U, __m512d __A)
{
return (__m256i) __builtin_ia32_cvttpd2dqs512_mask ((__v8df) __A,
(__v8si) __W,
@ -427,7 +427,7 @@ _mm512_mask_cvttspd_epi32 (__m256i __W, __mmask8 __U, __m512d __A)
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_cvttspd_epi32 (__mmask8 __U, __m512d __A)
_mm512_maskz_cvtts_pd_epi32 (__mmask8 __U, __m512d __A)
{
return
(__m256i) __builtin_ia32_cvttpd2dqs512_mask ((__v8df) __A,
@ -438,7 +438,7 @@ _mm512_maskz_cvttspd_epi32 (__mmask8 __U, __m512d __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_cvttspd_epi64 (__m512d __A)
_mm512_cvtts_pd_epi64 (__m512d __A)
{
return (__m512i)
__builtin_ia32_cvttpd2qqs512_mask ((__v8df) __A,
@ -449,7 +449,7 @@ _mm512_cvttspd_epi64 (__m512d __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvttspd_epi64 (__m512i __W, __mmask8 __U, __m512d __A)
_mm512_mask_cvtts_pd_epi64 (__m512i __W, __mmask8 __U, __m512d __A)
{
return (__m512i) __builtin_ia32_cvttpd2qqs512_mask ((__v8df) __A,
(__v8di) __W,
@ -458,7 +458,7 @@ _mm512_mask_cvttspd_epi64 (__m512i __W, __mmask8 __U, __m512d __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_cvttspd_epi64 (__mmask8 __U, __m512d __A)
_mm512_maskz_cvtts_pd_epi64 (__mmask8 __U, __m512d __A)
{
return
(__m512i) __builtin_ia32_cvttpd2qqs512_mask ((__v8df) __A,
@ -469,7 +469,7 @@ _mm512_maskz_cvttspd_epi64 (__mmask8 __U, __m512d __A)
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_cvttspd_epu32 (__m512d __A)
_mm512_cvtts_pd_epu32 (__m512d __A)
{
return (__m256i)
__builtin_ia32_cvttpd2udqs512_mask ((__v8df) __A,
@ -480,7 +480,7 @@ _mm512_cvttspd_epu32 (__m512d __A)
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvttspd_epu32 (__m256i __W, __mmask8 __U, __m512d __A)
_mm512_mask_cvtts_pd_epu32 (__m256i __W, __mmask8 __U, __m512d __A)
{
return (__m256i) __builtin_ia32_cvttpd2udqs512_mask ((__v8df) __A,
(__v8si) __W,
@ -489,7 +489,7 @@ _mm512_mask_cvttspd_epu32 (__m256i __W, __mmask8 __U, __m512d __A)
extern __inline __m256i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_cvttspd_epu32 (__mmask8 __U, __m512d __A)
_mm512_maskz_cvtts_pd_epu32 (__mmask8 __U, __m512d __A)
{
return
(__m256i) __builtin_ia32_cvttpd2udqs512_mask ((__v8df) __A,
@ -500,7 +500,7 @@ _mm512_maskz_cvttspd_epu32 (__mmask8 __U, __m512d __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_cvttspd_epu64 (__m512d __A)
_mm512_cvtts_pd_epu64 (__m512d __A)
{
return (__m512i)
__builtin_ia32_cvttpd2uqqs512_mask ((__v8df) __A,
@ -511,7 +511,7 @@ _mm512_cvttspd_epu64 (__m512d __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvttspd_epu64 (__m512i __W, __mmask8 __U, __m512d __A)
_mm512_mask_cvtts_pd_epu64 (__m512i __W, __mmask8 __U, __m512d __A)
{
return (__m512i) __builtin_ia32_cvttpd2uqqs512_mask ((__v8df) __A,
(__v8di) __W,
@ -520,7 +520,7 @@ _mm512_mask_cvttspd_epu64 (__m512i __W, __mmask8 __U, __m512d __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_cvttspd_epu64 (__mmask8 __U, __m512d __A)
_mm512_maskz_cvtts_pd_epu64 (__mmask8 __U, __m512d __A)
{
return (__m512i)
__builtin_ia32_cvttpd2uqqs512_mask ((__v8df) __A,
@ -531,7 +531,7 @@ _mm512_maskz_cvttspd_epu64 (__mmask8 __U, __m512d __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_cvttsps_epi32 (__m512 __A)
_mm512_cvtts_ps_epi32 (__m512 __A)
{
return (__m512i)
__builtin_ia32_cvttps2dqs512_mask ((__v16sf) __A,
@ -542,7 +542,7 @@ _mm512_cvttsps_epi32 (__m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvttsps_epi32 (__m512i __W, __mmask16 __U, __m512 __A)
_mm512_mask_cvtts_ps_epi32 (__m512i __W, __mmask16 __U, __m512 __A)
{
return (__m512i) __builtin_ia32_cvttps2dqs512_mask ((__v16sf) __A,
(__v16si) __W,
@ -551,7 +551,7 @@ _mm512_mask_cvttsps_epi32 (__m512i __W, __mmask16 __U, __m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_cvttsps_epi32 (__mmask16 __U, __m512 __A)
_mm512_maskz_cvtts_ps_epi32 (__mmask16 __U, __m512 __A)
{
return
(__m512i) __builtin_ia32_cvttps2dqs512_mask ((__v16sf) __A,
@ -562,7 +562,7 @@ _mm512_maskz_cvttsps_epi32 (__mmask16 __U, __m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_cvttsps_epi64 (__m256 __A)
_mm512_cvtts_ps_epi64 (__m256 __A)
{
return (__m512i)
__builtin_ia32_cvttps2qqs512_mask ((__v8sf) __A,
@ -573,7 +573,7 @@ _mm512_cvttsps_epi64 (__m256 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvttsps_epi64 (__m512i __W, __mmask8 __U, __m256 __A)
_mm512_mask_cvtts_ps_epi64 (__m512i __W, __mmask8 __U, __m256 __A)
{
return (__m512i) __builtin_ia32_cvttps2qqs512_mask ((__v8sf) __A,
(__v8di) __W,
@ -582,7 +582,7 @@ _mm512_mask_cvttsps_epi64 (__m512i __W, __mmask8 __U, __m256 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_cvttsps_epi64 (__mmask8 __U, __m256 __A)
_mm512_maskz_cvtts_ps_epi64 (__mmask8 __U, __m256 __A)
{
return
(__m512i) __builtin_ia32_cvttps2qqs512_mask ((__v8sf) __A,
@ -593,7 +593,7 @@ _mm512_maskz_cvttsps_epi64 (__mmask8 __U, __m256 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_cvttsps_epu32 (__m512 __A)
_mm512_cvtts_ps_epu32 (__m512 __A)
{
return (__m512i)
__builtin_ia32_cvttps2udqs512_mask ((__v16sf) __A,
@ -604,7 +604,7 @@ _mm512_cvttsps_epu32 (__m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvttsps_epu32 (__m512i __W, __mmask16 __U, __m512 __A)
_mm512_mask_cvtts_ps_epu32 (__m512i __W, __mmask16 __U, __m512 __A)
{
return (__m512i) __builtin_ia32_cvttps2udqs512_mask ((__v16sf) __A,
(__v16si) __W,
@ -613,7 +613,7 @@ _mm512_mask_cvttsps_epu32 (__m512i __W, __mmask16 __U, __m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_cvttsps_epu32 (__mmask16 __U, __m512 __A)
_mm512_maskz_cvtts_ps_epu32 (__mmask16 __U, __m512 __A)
{
return (__m512i)
__builtin_ia32_cvttps2udqs512_mask ((__v16sf) __A,
@ -624,7 +624,7 @@ _mm512_maskz_cvttsps_epu32 (__mmask16 __U, __m512 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_cvttsps_epu64 (__m256 __A)
_mm512_cvtts_ps_epu64 (__m256 __A)
{
return (__m512i)
__builtin_ia32_cvttps2uqqs512_mask ((__v8sf) __A,
@ -635,7 +635,7 @@ _mm512_cvttsps_epu64 (__m256 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_cvttsps_epu64 (__m512i __W, __mmask8 __U, __m256 __A)
_mm512_mask_cvtts_ps_epu64 (__m512i __W, __mmask8 __U, __m256 __A)
{
return (__m512i) __builtin_ia32_cvttps2uqqs512_mask ((__v8sf) __A,
(__v8di) __W,
@ -644,7 +644,7 @@ _mm512_mask_cvttsps_epu64 (__m512i __W, __mmask8 __U, __m256 __A)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_cvttsps_epu64 (__mmask8 __U, __m256 __A)
_mm512_maskz_cvtts_ps_epu64 (__mmask8 __U, __m256 __A)
{
return
(__m512i) __builtin_ia32_cvttps2uqqs512_mask ((__v8sf) __A,
@ -656,7 +656,7 @@ _mm512_maskz_cvttsps_epu64 (__mmask8 __U, __m256 __A)
#ifdef __OPTIMIZE__
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvt_roundph_epi8 (__m512h __A, const int __R)
_mm512_ipcvts_roundph_epi8 (__m512h __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtph2ibs512_mask_round ((__v32hf) __A,
@ -668,8 +668,8 @@ _mm512_ipcvt_roundph_epi8 (__m512h __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvt_roundph_epi8 (__m512i __W, __mmask32 __U, __m512h __A,
const int __R)
_mm512_mask_ipcvts_roundph_epi8 (__m512i __W, __mmask32 __U, __m512h __A,
const int __R)
{
return (__m512i) __builtin_ia32_cvtph2ibs512_mask_round ((__v32hf) __A,
(__v32hi) __W,
@ -679,7 +679,7 @@ _mm512_mask_ipcvt_roundph_epi8 (__m512i __W, __mmask32 __U, __m512h __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvt_roundph_epi8 (__mmask32 __U, __m512h __A, const int __R)
_mm512_maskz_ipcvts_roundph_epi8 (__mmask32 __U, __m512h __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtph2ibs512_mask_round ((__v32hf) __A,
@ -691,7 +691,7 @@ _mm512_maskz_ipcvt_roundph_epi8 (__mmask32 __U, __m512h __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvt_roundph_epu8 (__m512h __A, const int __R)
_mm512_ipcvts_roundph_epu8 (__m512h __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtph2iubs512_mask_round ((__v32hf) __A,
@ -703,8 +703,8 @@ _mm512_ipcvt_roundph_epu8 (__m512h __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvt_roundph_epu8 (__m512i __W, __mmask32 __U, __m512h __A,
const int __R)
_mm512_mask_ipcvts_roundph_epu8 (__m512i __W, __mmask32 __U, __m512h __A,
const int __R)
{
return (__m512i) __builtin_ia32_cvtph2iubs512_mask_round ((__v32hf) __A,
(__v32hi) __W,
@ -714,7 +714,7 @@ _mm512_mask_ipcvt_roundph_epu8 (__m512i __W, __mmask32 __U, __m512h __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvt_roundph_epu8 (__mmask32 __U, __m512h __A, const int __R)
_mm512_maskz_ipcvts_roundph_epu8 (__mmask32 __U, __m512h __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtph2iubs512_mask_round ((__v32hf) __A,
@ -726,7 +726,7 @@ _mm512_maskz_ipcvt_roundph_epu8 (__mmask32 __U, __m512h __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvt_roundps_epi8 (__m512 __A, const int __R)
_mm512_ipcvts_roundps_epi8 (__m512 __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtps2ibs512_mask_round ((__v16sf) __A,
@ -738,8 +738,8 @@ _mm512_ipcvt_roundps_epi8 (__m512 __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvt_roundps_epi8 (__m512i __W, __mmask16 __U, __m512 __A,
const int __R)
_mm512_mask_ipcvts_roundps_epi8 (__m512i __W, __mmask16 __U, __m512 __A,
const int __R)
{
return (__m512i) __builtin_ia32_cvtps2ibs512_mask_round ((__v16sf) __A,
(__v16si) __W,
@ -749,7 +749,7 @@ _mm512_mask_ipcvt_roundps_epi8 (__m512i __W, __mmask16 __U, __m512 __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvt_roundps_epi8 (__mmask16 __U, __m512 __A, const int __R)
_mm512_maskz_ipcvts_roundps_epi8 (__mmask16 __U, __m512 __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtps2ibs512_mask_round ((__v16sf) __A,
@ -761,7 +761,7 @@ _mm512_maskz_ipcvt_roundps_epi8 (__mmask16 __U, __m512 __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvt_roundps_epu8 (__m512 __A, const int __R)
_mm512_ipcvts_roundps_epu8 (__m512 __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtps2iubs512_mask_round ((__v16sf) __A,
@ -773,8 +773,8 @@ _mm512_ipcvt_roundps_epu8 (__m512 __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvt_roundps_epu8 (__m512i __W, __mmask16 __U, __m512 __A,
const int __R)
_mm512_mask_ipcvts_roundps_epu8 (__m512i __W, __mmask16 __U, __m512 __A,
const int __R)
{
return (__m512i) __builtin_ia32_cvtps2iubs512_mask_round ((__v16sf) __A,
(__v16si) __W,
@ -784,7 +784,7 @@ _mm512_mask_ipcvt_roundps_epu8 (__m512i __W, __mmask16 __U, __m512 __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvt_roundps_epu8 (__mmask16 __U, __m512 __A, const int __R)
_mm512_maskz_ipcvts_roundps_epu8 (__mmask16 __U, __m512 __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvtps2iubs512_mask_round ((__v16sf) __A,
@ -796,7 +796,7 @@ _mm512_maskz_ipcvt_roundps_epu8 (__mmask16 __U, __m512 __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvtt_roundph_epi8 (__m512h __A, const int __R)
_mm512_ipcvtts_roundph_epi8 (__m512h __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttph2ibs512_mask_round ((__v32hf) __A,
@ -808,8 +808,8 @@ _mm512_ipcvtt_roundph_epi8 (__m512h __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvtt_roundph_epi8 (__m512i __W, __mmask32 __U, __m512h __A,
const int __R)
_mm512_mask_ipcvtts_roundph_epi8 (__m512i __W, __mmask32 __U, __m512h __A,
const int __R)
{
return (__m512i) __builtin_ia32_cvttph2ibs512_mask_round ((__v32hf) __A,
(__v32hi) __W,
@ -819,7 +819,7 @@ _mm512_mask_ipcvtt_roundph_epi8 (__m512i __W, __mmask32 __U, __m512h __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvtt_roundph_epi8 (__mmask32 __U, __m512h __A, const int __R)
_mm512_maskz_ipcvtts_roundph_epi8 (__mmask32 __U, __m512h __A, const int __R)
{
return
(__m512i) __builtin_ia32_cvttph2ibs512_mask_round ((__v32hf) __A,
@ -831,7 +831,7 @@ _mm512_maskz_ipcvtt_roundph_epi8 (__mmask32 __U, __m512h __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvtt_roundph_epu8 (__m512h __A, const int __R)
_mm512_ipcvtts_roundph_epu8 (__m512h __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttph2iubs512_mask_round ((__v32hf) __A,
@ -843,8 +843,8 @@ _mm512_ipcvtt_roundph_epu8 (__m512h __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvtt_roundph_epu8 (__m512i __W, __mmask32 __U, __m512h __A,
const int __R)
_mm512_mask_ipcvtts_roundph_epu8 (__m512i __W, __mmask32 __U, __m512h __A,
const int __R)
{
return (__m512i) __builtin_ia32_cvttph2iubs512_mask_round ((__v32hf) __A,
(__v32hi) __W,
@ -854,7 +854,7 @@ _mm512_mask_ipcvtt_roundph_epu8 (__m512i __W, __mmask32 __U, __m512h __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvtt_roundph_epu8 (__mmask32 __U, __m512h __A, const int __R)
_mm512_maskz_ipcvtts_roundph_epu8 (__mmask32 __U, __m512h __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttph2iubs512_mask_round ((__v32hf) __A,
@ -866,7 +866,7 @@ _mm512_maskz_ipcvtt_roundph_epu8 (__mmask32 __U, __m512h __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvtt_roundps_epi8 (__m512 __A, const int __R)
_mm512_ipcvtts_roundps_epi8 (__m512 __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttps2ibs512_mask_round ((__v16sf) __A,
@ -878,8 +878,8 @@ _mm512_ipcvtt_roundps_epi8 (__m512 __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvtt_roundps_epi8 (__m512i __W, __mmask16 __U, __m512 __A,
const int __R)
_mm512_mask_ipcvtts_roundps_epi8 (__m512i __W, __mmask16 __U, __m512 __A,
const int __R)
{
return (__m512i) __builtin_ia32_cvttps2ibs512_mask_round ((__v16sf) __A,
(__v16si) __W,
@ -889,7 +889,7 @@ _mm512_mask_ipcvtt_roundps_epi8 (__m512i __W, __mmask16 __U, __m512 __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvtt_roundps_epi8 (__mmask16 __U, __m512 __A, const int __R)
_mm512_maskz_ipcvtts_roundps_epi8 (__mmask16 __U, __m512 __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttps2ibs512_mask_round ((__v16sf) __A,
@ -901,7 +901,7 @@ _mm512_maskz_ipcvtt_roundps_epi8 (__mmask16 __U, __m512 __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_ipcvtt_roundps_epu8 (__m512 __A, const int __R)
_mm512_ipcvtts_roundps_epu8 (__m512 __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttps2iubs512_mask_round ((__v16sf) __A,
@ -913,8 +913,8 @@ _mm512_ipcvtt_roundps_epu8 (__m512 __A, const int __R)
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_mask_ipcvtt_roundps_epu8 (__m512i __W, __mmask16 __U, __m512 __A,
const int __R)
_mm512_mask_ipcvtts_roundps_epu8 (__m512i __W, __mmask16 __U, __m512 __A,
const int __R)
{
return (__m512i) __builtin_ia32_cvttps2iubs512_mask_round ((__v16sf) __A,
(__v16si) __W,
@ -924,7 +924,7 @@ _mm512_mask_ipcvtt_roundps_epu8 (__m512i __W, __mmask16 __U, __m512 __A,
extern __inline __m512i
__attribute__ ((__gnu_inline__, __always_inline__, __artificial__))
_mm512_maskz_ipcvtt_roundps_epu8 (__mmask16 __U, __m512 __A, const int __R)
_mm512_maskz_ipcvtts_roundps_epu8 (__mmask16 __U, __m512 __A, const int __R)
{
return (__m512i)
__builtin_ia32_cvttps2iubs512_mask_round ((__v16sf) __A,
@ -1214,7 +1214,7 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
__R);
}
#else
#define _mm512_ipcvt_roundph_epi8(A, R) \
#define _mm512_ipcvts_roundph_epi8(A, R) \
((__m512i) \
__builtin_ia32_cvtph2ibs512_mask_round ((__v32hf) (A), \
(__v32hi) \
@ -1222,13 +1222,13 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask32) (-1), \
(R)))
#define _mm512_mask_ipcvt_roundph_epi8(W, U, A, R) \
#define _mm512_mask_ipcvts_roundph_epi8(W, U, A, R) \
((__m512i) __builtin_ia32_cvtph2ibs512_mask_round ((__v32hf) (A), \
(__v32hi) (W), \
(__mmask32) (U), \
(R)))
#define _mm512_maskz_ipcvt_roundph_epi8(U, A, R) \
#define _mm512_maskz_ipcvts_roundph_epi8(U, A, R) \
((__m512i) \
__builtin_ia32_cvtph2ibs512_mask_round ((__v32hf) (A), \
(__v32hi) \
@ -1236,7 +1236,7 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask32) (U), \
(R)))
#define _mm512_ipcvt_roundph_epu8(A, R) \
#define _mm512_ipcvts_roundph_epu8(A, R) \
((__m512i) \
__builtin_ia32_cvtph2iubs512_mask_round ((__v32hf) (A), \
(__v32hi) \
@ -1244,13 +1244,13 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask32) (-1), \
(R)))
#define _mm512_mask_ipcvt_roundph_epu8(W, U, A, R) \
#define _mm512_mask_ipcvts_roundph_epu8(W, U, A, R) \
((__m512i) __builtin_ia32_cvtph2iubs512_mask_round ((__v32hf) (A), \
(__v32hi) (W), \
(__mmask32) (U), \
(R)))
#define _mm512_maskz_ipcvt_roundph_epu8(U, A, R) \
#define _mm512_maskz_ipcvts_roundph_epu8(U, A, R) \
((__m512i) \
__builtin_ia32_cvtph2iubs512_mask_round ((__v32hf) (A), \
(__v32hi) \
@ -1258,7 +1258,7 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask32) (U), \
(R)))
#define _mm512_ipcvt_roundps_epi8(A, R) \
#define _mm512_ipcvts_roundps_epi8(A, R) \
((__m512i) \
__builtin_ia32_cvtps2ibs512_mask_round ((__v16sf) (A), \
(__v16si) \
@ -1266,13 +1266,13 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask16) (-1), \
(R)))
#define _mm512_mask_ipcvt_roundps_epi8(W, U, A, R) \
#define _mm512_mask_ipcvts_roundps_epi8(W, U, A, R) \
((__m512i) __builtin_ia32_cvtps2ibs512_mask_round ((__v16sf) (A), \
(__v16si) (W), \
(__mmask16) (U), \
(R)))
#define _mm512_maskz_ipcvt_roundps_epi8(U, A, R) \
#define _mm512_maskz_ipcvts_roundps_epi8(U, A, R) \
((__m512i) \
__builtin_ia32_cvtps2ibs512_mask_round ((__v16sf) (A), \
(__v16si) \
@ -1280,7 +1280,7 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask16) (U), \
(R)))
#define _mm512_ipcvt_roundps_epu8(A, R) \
#define _mm512_ipcvts_roundps_epu8(A, R) \
((__m512i) \
__builtin_ia32_cvtps2iubs512_mask_round ((__v16sf) (A), \
(__v16si) \
@ -1288,13 +1288,13 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask16) (-1), \
(R)))
#define _mm512_mask_ipcvt_roundps_epu8(W, U, A, R) \
#define _mm512_mask_ipcvts_roundps_epu8(W, U, A, R) \
((__m512i) __builtin_ia32_cvtps2iubs512_mask_round ((__v16sf) (A), \
(__v16si) (W), \
(__mmask16) (U), \
(R)))
#define _mm512_maskz_ipcvt_roundps_epu8(U, A, R) \
#define _mm512_maskz_ipcvts_roundps_epu8(U, A, R) \
((__m512i) \
__builtin_ia32_cvtps2iubs512_mask_round ((__v16sf) (A), \
(__v16si) \
@ -1302,7 +1302,7 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask16) (U), \
(R)))
#define _mm512_ipcvtt_roundph_epi8(A, R) \
#define _mm512_ipcvtts_roundph_epi8(A, R) \
((__m512i) \
__builtin_ia32_cvttph2ibs512_mask_round ((__v32hf) (A), \
(__v32hi) \
@ -1310,13 +1310,13 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask32) (-1), \
(R)))
#define _mm512_mask_ipcvtt_roundph_epi8(W, U, A, R) \
#define _mm512_mask_ipcvtts_roundph_epi8(W, U, A, R) \
((__m512i) __builtin_ia32_cvttph2ibs512_mask_round ((__v32hf) (A), \
(__v32hi) (W), \
(__mmask32) (U), \
(R)))
#define _mm512_maskz_ipcvtt_roundph_epi8(U, A, R) \
#define _mm512_maskz_ipcvtts_roundph_epi8(U, A, R) \
((__m512i) \
__builtin_ia32_cvttph2ibs512_mask_round ((__v32hf) (A), \
(__v32hi) \
@ -1324,7 +1324,7 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask32) (U), \
(R)))
#define _mm512_ipcvtt_roundph_epu8(A, R) \
#define _mm512_ipcvtts_roundph_epu8(A, R) \
((__m512i) \
__builtin_ia32_cvttph2iubs512_mask_round ((__v32hf) (A), \
(__v32hi) \
@ -1332,13 +1332,13 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask32) (-1), \
(R)))
#define _mm512_mask_ipcvtt_roundph_epu8(W, U, A, R) \
#define _mm512_mask_ipcvtts_roundph_epu8(W, U, A, R) \
((__m512i) __builtin_ia32_cvttph2iubs512_mask_round ((__v32hf) (A), \
(__v32hi) (W), \
(__mmask32) (U), \
(R)))
#define _mm512_maskz_ipcvtt_roundph_epu8(U, A, R) \
#define _mm512_maskz_ipcvtts_roundph_epu8(U, A, R) \
((__m512i) \
__builtin_ia32_cvttph2iubs512_mask_round ((__v32hf) (A), \
(__v32hi) \
@ -1346,7 +1346,7 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask32) (U), \
(R)))
#define _mm512_ipcvtt_roundps_epi8(A, R) \
#define _mm512_ipcvtts_roundps_epi8(A, R) \
((__m512i) \
__builtin_ia32_cvttps2ibs512_mask_round ((__v16sf) (A), \
(__v16si) \
@ -1354,13 +1354,13 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask16) (-1), \
(R)))
#define _mm512_mask_ipcvtt_roundps_epi8(W, U, A, R) \
#define _mm512_mask_ipcvtts_roundps_epi8(W, U, A, R) \
((__m512i) __builtin_ia32_cvttps2ibs512_mask_round ((__v16sf) (A), \
(__v16si) (W), \
(__mmask16) (U), \
(R)))
#define _mm512_maskz_ipcvtt_roundps_epi8(U, A, R) \
#define _mm512_maskz_ipcvtts_roundps_epi8(U, A, R) \
((__m512i) \
__builtin_ia32_cvttps2ibs512_mask_round ((__v16sf) (A), \
(__v16si) \
@ -1368,7 +1368,7 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask16) (U), \
(R)))
#define _mm512_ipcvtt_roundps_epu8(A, R) \
#define _mm512_ipcvtts_roundps_epu8(A, R) \
((__m512i) \
__builtin_ia32_cvttps2iubs512_mask_round ((__v16sf) (A), \
(__v16si) \
@ -1376,13 +1376,13 @@ _mm512_maskz_cvtts_roundps_epu64 (__mmask8 __U, __m256 __A, const int __R)
(__mmask16) (-1), \
(R)))
#define _mm512_mask_ipcvtt_roundps_epu8(W, U, A, R) \
#define _mm512_mask_ipcvtts_roundps_epu8(W, U, A, R) \
((__m512i) __builtin_ia32_cvttps2iubs512_mask_round ((__v16sf) (A), \
(__v16si) (W), \
(__mmask16) (U), \
(R)))
#define _mm512_maskz_ipcvtt_roundps_epu8(U, A, R) \
#define _mm512_maskz_ipcvtts_roundps_epu8(U, A, R) \
((__m512i) \
__builtin_ia32_cvttps2iubs512_mask_round ((__v16sf) (A), \
(__v16si) \

File diff suppressed because it is too large Load diff

View file

@ -117,130 +117,130 @@ volatile __mmask32 m32;
void extern
avx10_2_test (void)
{
xi = _mm512_ipcvtph_epi8 (xh);
xi = _mm512_mask_ipcvtph_epi8 (xi, m32, xh);
xi = _mm512_maskz_ipcvtph_epi8 (m32, xh);
xi = _mm512_ipcvt_roundph_epi8 (xh, 4);
xi = _mm512_mask_ipcvt_roundph_epi8 (xi, m32, xh, 8);
xi = _mm512_maskz_ipcvt_roundph_epi8 (m32, xh, 11);
xi = _mm512_ipcvts_ph_epi8 (xh);
xi = _mm512_mask_ipcvts_ph_epi8 (xi, m32, xh);
xi = _mm512_maskz_ipcvts_ph_epi8 (m32, xh);
xi = _mm512_ipcvts_roundph_epi8 (xh, 4);
xi = _mm512_mask_ipcvts_roundph_epi8 (xi, m32, xh, 8);
xi = _mm512_maskz_ipcvts_roundph_epi8 (m32, xh, 11);
xi = _mm512_ipcvtph_epu8 (xh);
xi = _mm512_mask_ipcvtph_epu8 (xi, m32, xh);
xi = _mm512_maskz_ipcvtph_epu8 (m32, xh);
xi = _mm512_ipcvt_roundph_epu8 (xh, 4);
xi = _mm512_mask_ipcvt_roundph_epu8 (xi, m32, xh, 8);
xi = _mm512_maskz_ipcvt_roundph_epu8 (m32, xh, 11);
xi = _mm512_ipcvts_ph_epu8 (xh);
xi = _mm512_mask_ipcvts_ph_epu8 (xi, m32, xh);
xi = _mm512_maskz_ipcvts_ph_epu8 (m32, xh);
xi = _mm512_ipcvts_roundph_epu8 (xh, 4);
xi = _mm512_mask_ipcvts_roundph_epu8 (xi, m32, xh, 8);
xi = _mm512_maskz_ipcvts_roundph_epu8 (m32, xh, 11);
xi = _mm512_ipcvttph_epi8 (xh);
xi = _mm512_mask_ipcvttph_epi8 (xi, m32, xh);
xi = _mm512_maskz_ipcvttph_epi8 (m32, xh);
xi = _mm512_ipcvtt_roundph_epi8 (xh, 4);
xi = _mm512_mask_ipcvtt_roundph_epi8 (xi, m32, xh, 8);
xi = _mm512_maskz_ipcvtt_roundph_epi8 (m32, xh, 8);
xi = _mm512_ipcvtts_ph_epi8 (xh);
xi = _mm512_mask_ipcvtts_ph_epi8 (xi, m32, xh);
xi = _mm512_maskz_ipcvtts_ph_epi8 (m32, xh);
xi = _mm512_ipcvtts_roundph_epi8 (xh, 4);
xi = _mm512_mask_ipcvtts_roundph_epi8 (xi, m32, xh, 8);
xi = _mm512_maskz_ipcvtts_roundph_epi8 (m32, xh, 8);
xi = _mm512_ipcvttph_epu8 (xh);
xi = _mm512_mask_ipcvttph_epu8 (xi, m32, xh);
xi = _mm512_maskz_ipcvttph_epu8 (m32, xh);
xi = _mm512_ipcvtt_roundph_epu8 (xh, 4);
xi = _mm512_mask_ipcvtt_roundph_epu8 (xi, m32, xh, 8);
xi = _mm512_maskz_ipcvtt_roundph_epu8 (m32, xh, 8);
xi = _mm512_ipcvtts_ph_epu8 (xh);
xi = _mm512_mask_ipcvtts_ph_epu8 (xi, m32, xh);
xi = _mm512_maskz_ipcvtts_ph_epu8 (m32, xh);
xi = _mm512_ipcvtts_roundph_epu8 (xh, 4);
xi = _mm512_mask_ipcvtts_roundph_epu8 (xi, m32, xh, 8);
xi = _mm512_maskz_ipcvtts_roundph_epu8 (m32, xh, 8);
xi = _mm512_ipcvtps_epi8 (x);
xi = _mm512_mask_ipcvtps_epi8 (xi, m16, x);
xi = _mm512_maskz_ipcvtps_epi8 (m16, x);
xi = _mm512_ipcvt_roundps_epi8 (x, 4);
xi = _mm512_mask_ipcvt_roundps_epi8 (xi, m16, x, 8);
xi = _mm512_maskz_ipcvt_roundps_epi8 (m16, x, 11);
xi = _mm512_ipcvts_ps_epi8 (x);
xi = _mm512_mask_ipcvts_ps_epi8 (xi, m16, x);
xi = _mm512_maskz_ipcvts_ps_epi8 (m16, x);
xi = _mm512_ipcvts_roundps_epi8 (x, 4);
xi = _mm512_mask_ipcvts_roundps_epi8 (xi, m16, x, 8);
xi = _mm512_maskz_ipcvts_roundps_epi8 (m16, x, 11);
xi = _mm512_ipcvtps_epu8 (x);
xi = _mm512_mask_ipcvtps_epu8 (xi, m16, x);
xi = _mm512_maskz_ipcvtps_epu8 (m16, x);
xi = _mm512_ipcvt_roundps_epu8 (x, 4);
xi = _mm512_mask_ipcvt_roundps_epu8 (xi, m16, x, 8);
xi = _mm512_maskz_ipcvt_roundps_epu8 (m16, x, 11);
xi = _mm512_ipcvts_ps_epu8 (x);
xi = _mm512_mask_ipcvts_ps_epu8 (xi, m16, x);
xi = _mm512_maskz_ipcvts_ps_epu8 (m16, x);
xi = _mm512_ipcvts_roundps_epu8 (x, 4);
xi = _mm512_mask_ipcvts_roundps_epu8 (xi, m16, x, 8);
xi = _mm512_maskz_ipcvts_roundps_epu8 (m16, x, 11);
xi = _mm512_ipcvttps_epi8 (x);
xi = _mm512_mask_ipcvttps_epi8 (xi, m16, x);
xi = _mm512_maskz_ipcvttps_epi8 (m16, x);
xi = _mm512_ipcvtt_roundps_epi8 (x, 4);
xi = _mm512_mask_ipcvtt_roundps_epi8 (xi, m16, x, 8);
xi = _mm512_maskz_ipcvtt_roundps_epi8 (m16, x, 8);
xi = _mm512_ipcvtts_ps_epi8 (x);
xi = _mm512_mask_ipcvtts_ps_epi8 (xi, m16, x);
xi = _mm512_maskz_ipcvtts_ps_epi8 (m16, x);
xi = _mm512_ipcvtts_roundps_epi8 (x, 4);
xi = _mm512_mask_ipcvtts_roundps_epi8 (xi, m16, x, 8);
xi = _mm512_maskz_ipcvtts_roundps_epi8 (m16, x, 8);
xi = _mm512_ipcvttps_epu8 (x);
xi = _mm512_mask_ipcvttps_epu8 (xi, m16, x);
xi = _mm512_maskz_ipcvttps_epu8 (m16, x);
xi = _mm512_ipcvtt_roundps_epu8 (x, 4);
xi = _mm512_mask_ipcvtt_roundps_epu8 (xi, m16, x, 8);
xi = _mm512_maskz_ipcvtt_roundps_epu8 (m16, x, 8);
xi = _mm512_ipcvtts_ps_epu8 (x);
xi = _mm512_mask_ipcvtts_ps_epu8 (xi, m16, x);
xi = _mm512_maskz_ipcvtts_ps_epu8 (m16, x);
xi = _mm512_ipcvtts_roundps_epu8 (x, 4);
xi = _mm512_mask_ipcvtts_roundps_epu8 (xi, m16, x, 8);
xi = _mm512_maskz_ipcvtts_roundps_epu8 (m16, x, 8);
xi = _mm512_ipcvtbf16_epi8 (xbh);
xi = _mm512_mask_ipcvtbf16_epi8 (xi, m32, xbh);
xi = _mm512_maskz_ipcvtbf16_epi8 (m32, xbh);
xi = _mm512_ipcvts_bf16_epi8 (xbh);
xi = _mm512_mask_ipcvts_bf16_epi8 (xi, m32, xbh);
xi = _mm512_maskz_ipcvts_bf16_epi8 (m32, xbh);
xi = _mm512_ipcvtbf16_epu8 (xbh);
xi = _mm512_mask_ipcvtbf16_epu8 (xi, m32, xbh);
xi = _mm512_maskz_ipcvtbf16_epu8 (m32, xbh);
xi = _mm512_ipcvts_bf16_epu8 (xbh);
xi = _mm512_mask_ipcvts_bf16_epu8 (xi, m32, xbh);
xi = _mm512_maskz_ipcvts_bf16_epu8 (m32, xbh);
xi = _mm512_ipcvttbf16_epi8 (xbh);
xi = _mm512_mask_ipcvttbf16_epi8 (xi, m32, xbh);
xi = _mm512_maskz_ipcvttbf16_epi8 (m32, xbh);
xi = _mm512_ipcvtts_bf16_epi8 (xbh);
xi = _mm512_mask_ipcvtts_bf16_epi8 (xi, m32, xbh);
xi = _mm512_maskz_ipcvtts_bf16_epi8 (m32, xbh);
xi = _mm512_ipcvttbf16_epu8 (xbh);
xi = _mm512_mask_ipcvttbf16_epu8 (xi, m32, xbh);
xi = _mm512_maskz_ipcvttbf16_epu8 (m32, xbh);
xi = _mm512_ipcvtts_bf16_epu8 (xbh);
xi = _mm512_mask_ipcvtts_bf16_epu8 (xi, m32, xbh);
xi = _mm512_maskz_ipcvtts_bf16_epu8 (m32, xbh);
hxi = _mm512_cvttspd_epi32 (xd);
hxi = _mm512_mask_cvttspd_epi32 (hxi, m8, xd);
hxi = _mm512_maskz_cvttspd_epi32 (m8, xd);
hxi = _mm512_cvtts_pd_epi32 (xd);
hxi = _mm512_mask_cvtts_pd_epi32 (hxi, m8, xd);
hxi = _mm512_maskz_cvtts_pd_epi32 (m8, xd);
hxi = _mm512_cvtts_roundpd_epi32 (xd, 8);
hxi = _mm512_mask_cvtts_roundpd_epi32 (hxi, m8, xd, 8);
hxi = _mm512_maskz_cvtts_roundpd_epi32 (m8, xd, 8);
xi = _mm512_cvttspd_epi64 (xd);
xi = _mm512_mask_cvttspd_epi64 (xi, m8, xd);
xi = _mm512_maskz_cvttspd_epi64 (m8, xd);
xi = _mm512_cvtts_pd_epi64 (xd);
xi = _mm512_mask_cvtts_pd_epi64 (xi, m8, xd);
xi = _mm512_maskz_cvtts_pd_epi64 (m8, xd);
xi = _mm512_cvtts_roundpd_epi64 (xd, 8);
xi = _mm512_mask_cvtts_roundpd_epi64 (xi, m8, xd, 8);
xi = _mm512_maskz_cvtts_roundpd_epi64 (m8, xd, 8);
hxi = _mm512_cvttspd_epu32 (xd);
hxi = _mm512_mask_cvttspd_epu32 (hxi, m8, xd);
hxi = _mm512_maskz_cvttspd_epu32 (m8, xd);
hxi = _mm512_cvtts_pd_epu32 (xd);
hxi = _mm512_mask_cvtts_pd_epu32 (hxi, m8, xd);
hxi = _mm512_maskz_cvtts_pd_epu32 (m8, xd);
hxi = _mm512_cvtts_roundpd_epu32 (xd, 8);
hxi = _mm512_mask_cvtts_roundpd_epu32 (hxi, m8, xd, 8);
hxi = _mm512_maskz_cvtts_roundpd_epu32 (m8, xd, 8);
xi = _mm512_cvttspd_epu64 (xd);
xi = _mm512_mask_cvttspd_epu64 (xi, m8, xd);
xi = _mm512_maskz_cvttspd_epu64 (m8, xd);
xi = _mm512_cvtts_pd_epu64 (xd);
xi = _mm512_mask_cvtts_pd_epu64 (xi, m8, xd);
xi = _mm512_maskz_cvtts_pd_epu64 (m8, xd);
xi = _mm512_cvtts_roundpd_epu64 (xd, 8);
xi = _mm512_mask_cvtts_roundpd_epu64 (xi, m8, xd, 8);
xi = _mm512_maskz_cvtts_roundpd_epu64 (m8, xd, 8);
xi = _mm512_cvttsps_epi32 (x);
xi = _mm512_mask_cvttsps_epi32 (xi, m16, x);
xi = _mm512_maskz_cvttsps_epi32 (m16, x);
xi = _mm512_cvtts_ps_epi32 (x);
xi = _mm512_mask_cvtts_ps_epi32 (xi, m16, x);
xi = _mm512_maskz_cvtts_ps_epi32 (m16, x);
xi = _mm512_cvtts_roundps_epi32 (x, 8);
xi = _mm512_mask_cvtts_roundps_epi32 (xi, m16, x, 8);
xi = _mm512_maskz_cvtts_roundps_epi32 (m16, x, 8);
xi = _mm512_cvttsps_epi64 (hx);
xi = _mm512_mask_cvttsps_epi64 (xi, m8, hx);
xi = _mm512_maskz_cvttsps_epi64 (m8, hx);
xi = _mm512_cvtts_ps_epi64 (hx);
xi = _mm512_mask_cvtts_ps_epi64 (xi, m8, hx);
xi = _mm512_maskz_cvtts_ps_epi64 (m8, hx);
xi = _mm512_cvtts_roundps_epi64 (hx, 8);
xi = _mm512_mask_cvtts_roundps_epi64 (xi, m8, hx, 8);
xi = _mm512_maskz_cvtts_roundps_epi64 (m8, hx, 8);
xi = _mm512_cvttsps_epu32 (x);
xi = _mm512_mask_cvttsps_epu32 (xi, m16, x);
xi = _mm512_maskz_cvttsps_epu32 (m16, x);
xi = _mm512_cvtts_ps_epu32 (x);
xi = _mm512_mask_cvtts_ps_epu32 (xi, m16, x);
xi = _mm512_maskz_cvtts_ps_epu32 (m16, x);
xi = _mm512_cvtts_roundps_epu32 (x, 8);
xi = _mm512_mask_cvtts_roundps_epu32 (xi, m16, x, 8);
xi = _mm512_maskz_cvtts_roundps_epu32 (m16, x, 8);
xi = _mm512_cvttsps_epu64 (hx);
xi = _mm512_mask_cvttsps_epu64 (xi, m8, hx);
xi = _mm512_maskz_cvttsps_epu64 (m8, hx);
xi = _mm512_cvtts_ps_epu64 (hx);
xi = _mm512_mask_cvtts_ps_epu64 (xi, m8, hx);
xi = _mm512_maskz_cvtts_ps_epu64 (m8, hx);
xi = _mm512_cvtts_roundps_epu64 (hx, 8);
xi = _mm512_mask_cvtts_roundps_epu64 (xi, m8, hx, 8);
xi = _mm512_maskz_cvtts_roundps_epu64 (m8, hx, 8);

View file

@ -50,9 +50,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_ipcvtbf16_epi8) (s.x);
res2.x = INTRINSIC (_mask_ipcvtbf16_epi8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvtbf16_epi8) (mask, s.x);
res1.x = INTRINSIC (_ipcvts_bf16_epi8) (s.x);
res2.x = INTRINSIC (_mask_ipcvts_bf16_epi8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvts_bf16_epi8) (mask, s.x);
CALC (s.a, res_ref);

View file

@ -50,9 +50,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_ipcvtbf16_epu8) (s.x);
res2.x = INTRINSIC (_mask_ipcvtbf16_epu8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvtbf16_epu8) (mask, s.x);
res1.x = INTRINSIC (_ipcvts_bf16_epu8) (s.x);
res2.x = INTRINSIC (_mask_ipcvts_bf16_epu8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvts_bf16_epu8) (mask, s.x);
CALC (s.a, res_ref);

View file

@ -49,9 +49,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_ipcvtph_epi8) (s.x);
res2.x = INTRINSIC (_mask_ipcvtph_epi8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvtph_epi8) (mask, s.x);
res1.x = INTRINSIC (_ipcvts_ph_epi8) (s.x);
res2.x = INTRINSIC (_mask_ipcvts_ph_epi8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvts_ph_epi8) (mask, s.x);
CALC (s.a, res_ref);
@ -67,9 +67,9 @@ TEST (void)
abort ();
#if AVX512F_LEN != 128
res1.x = INTRINSIC (_ipcvt_roundph_epi8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvt_roundph_epi8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvt_roundph_epi8) (mask, s.x, 8);
res1.x = INTRINSIC (_ipcvts_roundph_epi8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvts_roundph_epi8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvts_roundph_epi8) (mask, s.x, 8);
if (UNION_CHECK (AVX512F_LEN, i_w) (res1, res_ref))
abort ();

View file

@ -49,9 +49,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_ipcvtph_epu8) (s.x);
res2.x = INTRINSIC (_mask_ipcvtph_epu8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvtph_epu8) (mask, s.x);
res1.x = INTRINSIC (_ipcvts_ph_epu8) (s.x);
res2.x = INTRINSIC (_mask_ipcvts_ph_epu8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvts_ph_epu8) (mask, s.x);
CALC (s.a, res_ref);
@ -67,9 +67,9 @@ TEST (void)
abort ();
#if AVX512F_LEN != 128
res1.x = INTRINSIC (_ipcvt_roundph_epu8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvt_roundph_epu8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvt_roundph_epu8) (mask, s.x, 8);
res1.x = INTRINSIC (_ipcvts_roundph_epu8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvts_roundph_epu8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvts_roundph_epu8) (mask, s.x, 8);
if (UNION_CHECK (AVX512F_LEN, i_w) (res1, res_ref))
abort ();

View file

@ -50,9 +50,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_ipcvtps_epi8) (s.x);
res2.x = INTRINSIC (_mask_ipcvtps_epi8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvtps_epi8) (mask, s.x);
res1.x = INTRINSIC (_ipcvts_ps_epi8) (s.x);
res2.x = INTRINSIC (_mask_ipcvts_ps_epi8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvts_ps_epi8) (mask, s.x);
CALC (s.a, res_ref);
@ -68,9 +68,9 @@ TEST (void)
abort ();
#if AVX512F_LEN != 128
res1.x = INTRINSIC (_ipcvt_roundps_epi8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvt_roundps_epi8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvt_roundps_epi8) (mask, s.x, 8);
res1.x = INTRINSIC (_ipcvts_roundps_epi8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvts_roundps_epi8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvts_roundps_epi8) (mask, s.x, 8);
if (UNION_CHECK (AVX512F_LEN, i_d) (res1, res_ref))
abort ();

View file

@ -48,9 +48,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_ipcvtps_epu8) (s.x);
res2.x = INTRINSIC (_mask_ipcvtps_epu8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvtps_epu8) (mask, s.x);
res1.x = INTRINSIC (_ipcvts_ps_epu8) (s.x);
res2.x = INTRINSIC (_mask_ipcvts_ps_epu8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvts_ps_epu8) (mask, s.x);
CALC (s.a, res_ref);
@ -66,9 +66,9 @@ TEST (void)
abort ();
#if AVX512F_LEN != 128
res1.x = INTRINSIC (_ipcvt_roundps_epu8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvt_roundps_epu8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvt_roundps_epu8) (mask, s.x, 8);
res1.x = INTRINSIC (_ipcvts_roundps_epu8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvts_roundps_epu8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvts_roundps_epu8) (mask, s.x, 8);
if (UNION_CHECK (AVX512F_LEN, i_d) (res1, res_ref))
abort ();

View file

@ -50,9 +50,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_ipcvttbf16_epi8) (s.x);
res2.x = INTRINSIC (_mask_ipcvttbf16_epi8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvttbf16_epi8) (mask, s.x);
res1.x = INTRINSIC (_ipcvtts_bf16_epi8) (s.x);
res2.x = INTRINSIC (_mask_ipcvtts_bf16_epi8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvtts_bf16_epi8) (mask, s.x);
CALC (s.a, res_ref);

View file

@ -50,9 +50,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_ipcvttbf16_epu8) (s.x);
res2.x = INTRINSIC (_mask_ipcvttbf16_epu8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvttbf16_epu8) (mask, s.x);
res1.x = INTRINSIC (_ipcvtts_bf16_epu8) (s.x);
res2.x = INTRINSIC (_mask_ipcvtts_bf16_epu8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvtts_bf16_epu8) (mask, s.x);
CALC (s.a, res_ref);

View file

@ -48,9 +48,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_cvttspd_epi32) (s.x);
res2.x = INTRINSIC (_mask_cvttspd_epi32) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvttspd_epi32) (mask, s.x);
res1.x = INTRINSIC (_cvtts_pd_epi32) (s.x);
res2.x = INTRINSIC (_mask_cvtts_pd_epi32) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvtts_pd_epi32) (mask, s.x);
CALC (s.a, res_ref);

View file

@ -47,9 +47,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_cvttspd_epi64) (s.x);
res2.x = INTRINSIC (_mask_cvttspd_epi64) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvttspd_epi64) (mask, s.x);
res1.x = INTRINSIC (_cvtts_pd_epi64) (s.x);
res2.x = INTRINSIC (_mask_cvtts_pd_epi64) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvtts_pd_epi64) (mask, s.x);
CALC (s.a, res_ref);

View file

@ -48,9 +48,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_cvttspd_epu32) (s.x);
res2.x = INTRINSIC (_mask_cvttspd_epu32) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvttspd_epu32) (mask, s.x);
res1.x = INTRINSIC (_cvtts_pd_epu32) (s.x);
res2.x = INTRINSIC (_mask_cvtts_pd_epu32) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvtts_pd_epu32) (mask, s.x);
CALC (s.a, res_ref);

View file

@ -47,9 +47,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_cvttspd_epu64) (s.x);
res2.x = INTRINSIC (_mask_cvttspd_epu64) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvttspd_epu64) (mask, s.x);
res1.x = INTRINSIC (_cvtts_pd_epu64) (s.x);
res2.x = INTRINSIC (_mask_cvtts_pd_epu64) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvtts_pd_epu64) (mask, s.x);
CALC (s.a, res_ref);

View file

@ -49,9 +49,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_ipcvttph_epi8) (s.x);
res2.x = INTRINSIC (_mask_ipcvttph_epi8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvttph_epi8) (mask, s.x);
res1.x = INTRINSIC (_ipcvtts_ph_epi8) (s.x);
res2.x = INTRINSIC (_mask_ipcvtts_ph_epi8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvtts_ph_epi8) (mask, s.x);
CALC (s.a, res_ref);
@ -67,9 +67,9 @@ TEST (void)
abort ();
#if AVX512F_LEN != 128
res1.x = INTRINSIC (_ipcvtt_roundph_epi8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvtt_roundph_epi8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvtt_roundph_epi8) (mask, s.x, 8);
res1.x = INTRINSIC (_ipcvtts_roundph_epi8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvtts_roundph_epi8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvtts_roundph_epi8) (mask, s.x, 8);
if (UNION_CHECK (AVX512F_LEN, i_w) (res1, res_ref))
abort ();

View file

@ -49,16 +49,16 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_ipcvttph_epu8) (s.x);
res2.x = INTRINSIC (_mask_ipcvttph_epu8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvttph_epu8) (mask, s.x);
res1.x = INTRINSIC (_ipcvtts_ph_epu8) (s.x);
res2.x = INTRINSIC (_mask_ipcvtts_ph_epu8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvtts_ph_epu8) (mask, s.x);
CALC (s.a, res_ref);
#if AVX512F_LEN != 128
res1.x = INTRINSIC (_ipcvtt_roundph_epu8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvtt_roundph_epu8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvtt_roundph_epu8) (mask, s.x, 8);
res1.x = INTRINSIC (_ipcvtts_roundph_epu8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvtts_roundph_epu8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvtts_roundph_epu8) (mask, s.x, 8);
if (UNION_CHECK (AVX512F_LEN, i_w) (res1, res_ref))
abort ();

View file

@ -47,9 +47,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_cvttsps_epi32) (s.x);
res2.x = INTRINSIC (_mask_cvttsps_epi32) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvttsps_epi32) (mask, s.x);
res1.x = INTRINSIC (_cvtts_ps_epi32) (s.x);
res2.x = INTRINSIC (_mask_cvtts_ps_epi32) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvtts_ps_epi32) (mask, s.x);
CALC (s.a, res_ref);

View file

@ -50,9 +50,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_ipcvttps_epi8) (s.x);
res2.x = INTRINSIC (_mask_ipcvttps_epi8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvttps_epi8) (mask, s.x);
res1.x = INTRINSIC (_ipcvtts_ps_epi8) (s.x);
res2.x = INTRINSIC (_mask_ipcvtts_ps_epi8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvtts_ps_epi8) (mask, s.x);
CALC (s.a, res_ref);
@ -68,9 +68,9 @@ TEST (void)
abort ();
#if AVX512F_LEN != 128
res1.x = INTRINSIC (_ipcvtt_roundps_epi8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvtt_roundps_epi8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvtt_roundps_epi8) (mask, s.x, 8);
res1.x = INTRINSIC (_ipcvtts_roundps_epi8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvtts_roundps_epi8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvtts_roundps_epi8) (mask, s.x, 8);
if (UNION_CHECK (AVX512F_LEN, i_d) (res1, res_ref))
abort ();

View file

@ -48,9 +48,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_ipcvttps_epu8) (s.x);
res2.x = INTRINSIC (_mask_ipcvttps_epu8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvttps_epu8) (mask, s.x);
res1.x = INTRINSIC (_ipcvtts_ps_epu8) (s.x);
res2.x = INTRINSIC (_mask_ipcvtts_ps_epu8) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_ipcvtts_ps_epu8) (mask, s.x);
CALC (s.a, res_ref);
@ -66,9 +66,9 @@ TEST (void)
abort ();
#if AVX512F_LEN != 128
res1.x = INTRINSIC (_ipcvtt_roundps_epu8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvtt_roundps_epu8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvtt_roundps_epu8) (mask, s.x, 8);
res1.x = INTRINSIC (_ipcvtts_roundps_epu8) (s.x, 8);
res2.x = INTRINSIC (_mask_ipcvtts_roundps_epu8) (res2.x, mask, s.x, 8);
res3.x = INTRINSIC (_maskz_ipcvtts_roundps_epu8) (mask, s.x, 8);
if (UNION_CHECK (AVX512F_LEN, i_d) (res1, res_ref))
abort ();

View file

@ -48,9 +48,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_cvttsps_epi64) (s.x);
res2.x = INTRINSIC (_mask_cvttsps_epi64) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvttsps_epi64) (mask, s.x);
res1.x = INTRINSIC (_cvtts_ps_epi64) (s.x);
res2.x = INTRINSIC (_mask_cvtts_ps_epi64) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvtts_ps_epi64) (mask, s.x);
CALC (s.a, res_ref);

View file

@ -47,9 +47,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_cvttsps_epu32) (s.x);
res2.x = INTRINSIC (_mask_cvttsps_epu32) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvttsps_epu32) (mask, s.x);
res1.x = INTRINSIC (_cvtts_ps_epu32) (s.x);
res2.x = INTRINSIC (_mask_cvtts_ps_epu32) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvtts_ps_epu32) (mask, s.x);
CALC (s.a, res_ref);

View file

@ -48,9 +48,9 @@ TEST (void)
for (i = 0; i < SIZE; i++)
res2.a[i] = DEFAULT_VALUE;
res1.x = INTRINSIC (_cvttsps_epu64) (s.x);
res2.x = INTRINSIC (_mask_cvttsps_epu64) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvttsps_epu64) (mask, s.x);
res1.x = INTRINSIC (_cvtts_ps_epu64) (s.x);
res2.x = INTRINSIC (_mask_cvtts_ps_epu64) (res2.x, mask, s.x);
res3.x = INTRINSIC (_maskz_cvtts_ps_epu64) (mask, s.x);
CALC (s.a, res_ref);

View file

@ -199,228 +199,228 @@ volatile unsigned long long ull;
void extern
avx10_2_test (void)
{
xi = _mm256_ipcvtph_epi8 (xh);
xi = _mm256_mask_ipcvtph_epi8 (xi, m16, xh);
xi = _mm256_maskz_ipcvtph_epi8 (m16, xh);
xi = _mm256_ipcvt_roundph_epi8 (xh, 4);
xi = _mm256_mask_ipcvt_roundph_epi8 (xi, m16, xh, 8);
xi = _mm256_maskz_ipcvt_roundph_epi8 (m16, xh, 11);
xi = _mm256_ipcvts_ph_epi8 (xh);
xi = _mm256_mask_ipcvts_ph_epi8 (xi, m16, xh);
xi = _mm256_maskz_ipcvts_ph_epi8 (m16, xh);
xi = _mm256_ipcvts_roundph_epi8 (xh, 4);
xi = _mm256_mask_ipcvts_roundph_epi8 (xi, m16, xh, 8);
xi = _mm256_maskz_ipcvts_roundph_epi8 (m16, xh, 11);
xi = _mm256_ipcvtph_epu8 (xh);
xi = _mm256_mask_ipcvtph_epu8 (xi, m16, xh);
xi = _mm256_maskz_ipcvtph_epu8 (m16, xh);
xi = _mm256_ipcvt_roundph_epu8 (xh, 4);
xi = _mm256_mask_ipcvt_roundph_epu8 (xi, m16, xh, 8);
xi = _mm256_maskz_ipcvt_roundph_epu8 (m16, xh, 11);
xi = _mm256_ipcvts_ph_epu8 (xh);
xi = _mm256_mask_ipcvts_ph_epu8 (xi, m16, xh);
xi = _mm256_maskz_ipcvts_ph_epu8 (m16, xh);
xi = _mm256_ipcvts_roundph_epu8 (xh, 4);
xi = _mm256_mask_ipcvts_roundph_epu8 (xi, m16, xh, 8);
xi = _mm256_maskz_ipcvts_roundph_epu8 (m16, xh, 11);
xi = _mm256_ipcvttph_epi8 (xh);
xi = _mm256_mask_ipcvttph_epi8 (xi, m16, xh);
xi = _mm256_maskz_ipcvttph_epi8 (m16, xh);
xi = _mm256_ipcvtt_roundph_epi8 (xh, 4);
xi = _mm256_mask_ipcvtt_roundph_epi8 (xi, m16, xh, 8);
xi = _mm256_maskz_ipcvtt_roundph_epi8 (m16, xh, 8);
xi = _mm256_ipcvtts_ph_epi8 (xh);
xi = _mm256_mask_ipcvtts_ph_epi8 (xi, m16, xh);
xi = _mm256_maskz_ipcvtts_ph_epi8 (m16, xh);
xi = _mm256_ipcvtts_roundph_epi8 (xh, 4);
xi = _mm256_mask_ipcvtts_roundph_epi8 (xi, m16, xh, 8);
xi = _mm256_maskz_ipcvtts_roundph_epi8 (m16, xh, 8);
xi = _mm256_ipcvttph_epu8 (xh);
xi = _mm256_mask_ipcvttph_epu8 (xi, m16, xh);
xi = _mm256_maskz_ipcvttph_epu8 (m16, xh);
xi = _mm256_ipcvtt_roundph_epu8 (xh, 4);
xi = _mm256_mask_ipcvtt_roundph_epu8 (xi, m16, xh, 8);
xi = _mm256_maskz_ipcvtt_roundph_epu8 (m16, xh, 8);
xi = _mm256_ipcvtts_ph_epu8 (xh);
xi = _mm256_mask_ipcvtts_ph_epu8 (xi, m16, xh);
xi = _mm256_maskz_ipcvtts_ph_epu8 (m16, xh);
xi = _mm256_ipcvtts_roundph_epu8 (xh, 4);
xi = _mm256_mask_ipcvtts_roundph_epu8 (xi, m16, xh, 8);
xi = _mm256_maskz_ipcvtts_roundph_epu8 (m16, xh, 8);
xi = _mm256_ipcvtps_epi8 (x);
xi = _mm256_mask_ipcvtps_epi8 (xi, m8, x);
xi = _mm256_maskz_ipcvtps_epi8 (m8, x);
xi = _mm256_ipcvt_roundps_epi8 (x, 4);
xi = _mm256_mask_ipcvt_roundps_epi8 (xi, m8, x, 8);
xi = _mm256_maskz_ipcvt_roundps_epi8 (m8, x, 11);
xi = _mm256_ipcvts_ps_epi8 (x);
xi = _mm256_mask_ipcvts_ps_epi8 (xi, m8, x);
xi = _mm256_maskz_ipcvts_ps_epi8 (m8, x);
xi = _mm256_ipcvts_roundps_epi8 (x, 4);
xi = _mm256_mask_ipcvts_roundps_epi8 (xi, m8, x, 8);
xi = _mm256_maskz_ipcvts_roundps_epi8 (m8, x, 11);
xi = _mm256_ipcvtps_epu8 (x);
xi = _mm256_mask_ipcvtps_epu8 (xi, m8, x);
xi = _mm256_maskz_ipcvtps_epu8 (m8, x);
xi = _mm256_ipcvt_roundps_epu8 (x, 4);
xi = _mm256_mask_ipcvt_roundps_epu8 (xi, m8, x, 8);
xi = _mm256_maskz_ipcvt_roundps_epu8 (m8, x, 11);
xi = _mm256_ipcvts_ps_epu8 (x);
xi = _mm256_mask_ipcvts_ps_epu8 (xi, m8, x);
xi = _mm256_maskz_ipcvts_ps_epu8 (m8, x);
xi = _mm256_ipcvts_roundps_epu8 (x, 4);
xi = _mm256_mask_ipcvts_roundps_epu8 (xi, m8, x, 8);
xi = _mm256_maskz_ipcvts_roundps_epu8 (m8, x, 11);
xi = _mm256_ipcvttps_epi8 (x);
xi = _mm256_mask_ipcvttps_epi8 (xi, m8, x);
xi = _mm256_maskz_ipcvttps_epi8 (m8, x);
xi = _mm256_ipcvtt_roundps_epi8 (x, 4);
xi = _mm256_mask_ipcvtt_roundps_epi8 (xi, m8, x, 8);
xi = _mm256_maskz_ipcvtt_roundps_epi8 (m8, x, 8);
xi = _mm256_ipcvtts_ps_epi8 (x);
xi = _mm256_mask_ipcvtts_ps_epi8 (xi, m8, x);
xi = _mm256_maskz_ipcvtts_ps_epi8 (m8, x);
xi = _mm256_ipcvtts_roundps_epi8 (x, 4);
xi = _mm256_mask_ipcvtts_roundps_epi8 (xi, m8, x, 8);
xi = _mm256_maskz_ipcvtts_roundps_epi8 (m8, x, 8);
xi = _mm256_ipcvttps_epu8 (x);
xi = _mm256_mask_ipcvttps_epu8 (xi, m8, x);
xi = _mm256_maskz_ipcvttps_epu8 (m8, x);
xi = _mm256_ipcvtt_roundps_epu8 (x, 4);
xi = _mm256_mask_ipcvtt_roundps_epu8 (xi, m8, x, 8);
xi = _mm256_maskz_ipcvtt_roundps_epu8 (m8, x, 8);
xi = _mm256_ipcvtts_ps_epu8 (x);
xi = _mm256_mask_ipcvtts_ps_epu8 (xi, m8, x);
xi = _mm256_maskz_ipcvtts_ps_epu8 (m8, x);
xi = _mm256_ipcvtts_roundps_epu8 (x, 4);
xi = _mm256_mask_ipcvtts_roundps_epu8 (xi, m8, x, 8);
xi = _mm256_maskz_ipcvtts_roundps_epu8 (m8, x, 8);
xi = _mm256_ipcvtbf16_epi8 (xbh);
xi = _mm256_mask_ipcvtbf16_epi8 (xi, m16, xbh);
xi = _mm256_maskz_ipcvtbf16_epi8 (m16, xbh);
xi = _mm256_ipcvts_bf16_epi8 (xbh);
xi = _mm256_mask_ipcvts_bf16_epi8 (xi, m16, xbh);
xi = _mm256_maskz_ipcvts_bf16_epi8 (m16, xbh);
xi = _mm256_ipcvtbf16_epu8 (xbh);
xi = _mm256_mask_ipcvtbf16_epu8 (xi, m16, xbh);
xi = _mm256_maskz_ipcvtbf16_epu8 (m16, xbh);
xi = _mm256_ipcvts_bf16_epu8 (xbh);
xi = _mm256_mask_ipcvts_bf16_epu8 (xi, m16, xbh);
xi = _mm256_maskz_ipcvts_bf16_epu8 (m16, xbh);
xi = _mm256_ipcvttbf16_epi8 (xbh);
xi = _mm256_mask_ipcvttbf16_epi8 (xi, m16, xbh);
xi = _mm256_maskz_ipcvttbf16_epi8 (m16, xbh);
xi = _mm256_ipcvtts_bf16_epi8 (xbh);
xi = _mm256_mask_ipcvtts_bf16_epi8 (xi, m16, xbh);
xi = _mm256_maskz_ipcvtts_bf16_epi8 (m16, xbh);
xi = _mm256_ipcvttbf16_epu8 (xbh);
xi = _mm256_mask_ipcvttbf16_epu8 (xi, m16, xbh);
xi = _mm256_maskz_ipcvttbf16_epu8 (m16, xbh);
xi = _mm256_ipcvtts_bf16_epu8 (xbh);
xi = _mm256_mask_ipcvtts_bf16_epu8 (xi, m16, xbh);
xi = _mm256_maskz_ipcvtts_bf16_epu8 (m16, xbh);
hxi = _mm_ipcvtph_epi8 (hxh);
hxi = _mm_mask_ipcvtph_epi8 (hxi, m8, hxh);
hxi = _mm_maskz_ipcvtph_epi8 (m8, hxh);
hxi = _mm_ipcvts_ph_epi8 (hxh);
hxi = _mm_mask_ipcvts_ph_epi8 (hxi, m8, hxh);
hxi = _mm_maskz_ipcvts_ph_epi8 (m8, hxh);
hxi = _mm_ipcvtph_epu8 (hxh);
hxi = _mm_mask_ipcvtph_epu8 (hxi, m8, hxh);
hxi = _mm_maskz_ipcvtph_epu8 (m8, hxh);
hxi = _mm_ipcvts_ph_epu8 (hxh);
hxi = _mm_mask_ipcvts_ph_epu8 (hxi, m8, hxh);
hxi = _mm_maskz_ipcvts_ph_epu8 (m8, hxh);
hxi = _mm_ipcvttph_epi8 (hxh);
hxi = _mm_mask_ipcvttph_epi8 (hxi, m8, hxh);
hxi = _mm_maskz_ipcvttph_epi8 (m8, hxh);
hxi = _mm_ipcvtts_ph_epi8 (hxh);
hxi = _mm_mask_ipcvtts_ph_epi8 (hxi, m8, hxh);
hxi = _mm_maskz_ipcvtts_ph_epi8 (m8, hxh);
hxi = _mm_ipcvttph_epu8 (hxh);
hxi = _mm_mask_ipcvttph_epu8 (hxi, m8, hxh);
hxi = _mm_maskz_ipcvttph_epu8 (m8, hxh);
hxi = _mm_ipcvtts_ph_epu8 (hxh);
hxi = _mm_mask_ipcvtts_ph_epu8 (hxi, m8, hxh);
hxi = _mm_maskz_ipcvtts_ph_epu8 (m8, hxh);
hxi = _mm_ipcvtps_epi8 (hx);
hxi = _mm_mask_ipcvtps_epi8 (hxi, m8, hx);
hxi = _mm_maskz_ipcvtps_epi8 (m8, hx);
hxi = _mm_ipcvts_ps_epi8 (hx);
hxi = _mm_mask_ipcvts_ps_epi8 (hxi, m8, hx);
hxi = _mm_maskz_ipcvts_ps_epi8 (m8, hx);
hxi = _mm_ipcvtps_epu8 (hx);
hxi = _mm_mask_ipcvtps_epu8 (hxi, m8, hx);
hxi = _mm_maskz_ipcvtps_epu8 (m8, hx);
hxi = _mm_ipcvts_ps_epu8 (hx);
hxi = _mm_mask_ipcvts_ps_epu8 (hxi, m8, hx);
hxi = _mm_maskz_ipcvts_ps_epu8 (m8, hx);
hxi = _mm_ipcvttps_epi8 (hx);
hxi = _mm_mask_ipcvttps_epi8 (hxi, m8, hx);
hxi = _mm_maskz_ipcvttps_epi8 (m8, hx);
hxi = _mm_ipcvtts_ps_epi8 (hx);
hxi = _mm_mask_ipcvtts_ps_epi8 (hxi, m8, hx);
hxi = _mm_maskz_ipcvtts_ps_epi8 (m8, hx);
hxi = _mm_ipcvttps_epu8 (hx);
hxi = _mm_mask_ipcvttps_epu8 (hxi, m8, hx);
hxi = _mm_maskz_ipcvttps_epu8 (m8, hx);
hxi = _mm_ipcvtts_ps_epu8 (hx);
hxi = _mm_mask_ipcvtts_ps_epu8 (hxi, m8, hx);
hxi = _mm_maskz_ipcvtts_ps_epu8 (m8, hx);
hxi = _mm_ipcvtbf16_epi8 (hxbh);
hxi = _mm_mask_ipcvtbf16_epi8 (hxi, m8, hxbh);
hxi = _mm_maskz_ipcvtbf16_epi8 (m8, hxbh);
hxi = _mm_ipcvts_bf16_epi8 (hxbh);
hxi = _mm_mask_ipcvts_bf16_epi8 (hxi, m8, hxbh);
hxi = _mm_maskz_ipcvts_bf16_epi8 (m8, hxbh);
hxi = _mm_ipcvtbf16_epu8 (hxbh);
hxi = _mm_mask_ipcvtbf16_epu8 (hxi, m8, hxbh);
hxi = _mm_maskz_ipcvtbf16_epu8 (m8, hxbh);
hxi = _mm_ipcvts_bf16_epu8 (hxbh);
hxi = _mm_mask_ipcvts_bf16_epu8 (hxi, m8, hxbh);
hxi = _mm_maskz_ipcvts_bf16_epu8 (m8, hxbh);
hxi = _mm_ipcvttbf16_epi8 (hxbh);
hxi = _mm_mask_ipcvttbf16_epi8 (hxi, m8, hxbh);
hxi = _mm_maskz_ipcvttbf16_epi8 (m8, hxbh);
hxi = _mm_ipcvtts_bf16_epi8 (hxbh);
hxi = _mm_mask_ipcvtts_bf16_epi8 (hxi, m8, hxbh);
hxi = _mm_maskz_ipcvtts_bf16_epi8 (m8, hxbh);
hxi = _mm_ipcvttbf16_epu8 (hxbh);
hxi = _mm_mask_ipcvttbf16_epu8 (hxi, m8, hxbh);
hxi = _mm_maskz_ipcvttbf16_epu8 (m8, hxbh);
hxi = _mm_ipcvtts_bf16_epu8 (hxbh);
hxi = _mm_mask_ipcvtts_bf16_epu8 (hxi, m8, hxbh);
hxi = _mm_maskz_ipcvtts_bf16_epu8 (m8, hxbh);
hxi = _mm256_cvttspd_epi32 (xd);
hxi = _mm256_mask_cvttspd_epi32 (hxi, m8, xd);
hxi = _mm256_maskz_cvttspd_epi32 (m8, xd);
hxi = _mm256_cvtts_pd_epi32 (xd);
hxi = _mm256_mask_cvtts_pd_epi32 (hxi, m8, xd);
hxi = _mm256_maskz_cvtts_pd_epi32 (m8, xd);
hxi = _mm256_cvtts_roundpd_epi32 (xd, 8);
hxi = _mm256_mask_cvtts_roundpd_epi32 (hxi, m8, xd, 8);
hxi = _mm256_maskz_cvtts_roundpd_epi32 (m8, xd, 8);
xi = _mm256_cvttspd_epi64 (xd);
xi = _mm256_mask_cvttspd_epi64 (xi, m8, xd);
xi = _mm256_maskz_cvttspd_epi64 (m8, xd);
xi = _mm256_cvtts_pd_epi64 (xd);
xi = _mm256_mask_cvtts_pd_epi64 (xi, m8, xd);
xi = _mm256_maskz_cvtts_pd_epi64 (m8, xd);
xi = _mm256_cvtts_roundpd_epi64 (xd, 8);
xi = _mm256_mask_cvtts_roundpd_epi64 (xi, m8, xd, 8);
xi = _mm256_maskz_cvtts_roundpd_epi64 (m8, xd, 8);
hxi = _mm256_cvttspd_epu32 (xd);
hxi = _mm256_mask_cvttspd_epu32 (hxi, m8, xd);
hxi = _mm256_maskz_cvttspd_epu32 (m8, xd);
hxi = _mm256_cvtts_pd_epu32 (xd);
hxi = _mm256_mask_cvtts_pd_epu32 (hxi, m8, xd);
hxi = _mm256_maskz_cvtts_pd_epu32 (m8, xd);
hxi = _mm256_cvtts_roundpd_epu32 (xd, 8);
hxi = _mm256_mask_cvtts_roundpd_epu32 (hxi, m8, xd, 8);
hxi = _mm256_maskz_cvtts_roundpd_epu32 (m8, xd, 8);
xi = _mm256_cvttspd_epu64 (xd);
xi = _mm256_mask_cvttspd_epu64 (xi, m8, xd);
xi = _mm256_maskz_cvttspd_epu64 (m8, xd);
xi = _mm256_cvtts_pd_epu64 (xd);
xi = _mm256_mask_cvtts_pd_epu64 (xi, m8, xd);
xi = _mm256_maskz_cvtts_pd_epu64 (m8, xd);
xi = _mm256_cvtts_roundpd_epu64 (xd, 8);
xi = _mm256_mask_cvtts_roundpd_epu64 (xi, m8, xd, 8);
xi = _mm256_maskz_cvtts_roundpd_epu64 (m8, xd, 8);
xi = _mm256_cvttsps_epi32 (x);
xi = _mm256_mask_cvttsps_epi32 (xi, m16, x);
xi = _mm256_maskz_cvttsps_epi32 (m16, x);
xi = _mm256_cvtts_ps_epi32 (x);
xi = _mm256_mask_cvtts_ps_epi32 (xi, m16, x);
xi = _mm256_maskz_cvtts_ps_epi32 (m16, x);
xi = _mm256_cvtts_roundps_epi32 (x, 8);
xi = _mm256_mask_cvtts_roundps_epi32 (xi, m16, x, 8);
xi = _mm256_maskz_cvtts_roundps_epi32 (m16, x, 8);
xi = _mm256_cvttsps_epi64 (hx);
xi = _mm256_mask_cvttsps_epi64 (xi, m8, hx);
xi = _mm256_maskz_cvttsps_epi64 (m8, hx);
xi = _mm256_cvtts_ps_epi64 (hx);
xi = _mm256_mask_cvtts_ps_epi64 (xi, m8, hx);
xi = _mm256_maskz_cvtts_ps_epi64 (m8, hx);
xi = _mm256_cvtts_roundps_epi64 (hx, 8);
xi = _mm256_mask_cvtts_roundps_epi64 (xi, m8, hx, 8);
xi = _mm256_maskz_cvtts_roundps_epi64 (m8, hx, 8);
xi = _mm256_cvttsps_epu32 (x);
xi = _mm256_mask_cvttsps_epu32 (xi, m16, x);
xi = _mm256_maskz_cvttsps_epu32 (m16, x);
xi = _mm256_cvtts_ps_epu32 (x);
xi = _mm256_mask_cvtts_ps_epu32 (xi, m16, x);
xi = _mm256_maskz_cvtts_ps_epu32 (m16, x);
xi = _mm256_cvtts_roundps_epu32 (x, 8);
xi = _mm256_mask_cvtts_roundps_epu32 (xi, m16, x, 8);
xi = _mm256_maskz_cvtts_roundps_epu32 (m16, x, 8);
xi = _mm256_cvttsps_epu64 (hx);
xi = _mm256_mask_cvttsps_epu64 (xi, m8, hx);
xi = _mm256_maskz_cvttsps_epu64 (m8, hx);
xi = _mm256_cvtts_ps_epu64 (hx);
xi = _mm256_mask_cvtts_ps_epu64 (xi, m8, hx);
xi = _mm256_maskz_cvtts_ps_epu64 (m8, hx);
xi = _mm256_cvtts_roundps_epu64 (hx, 8);
xi = _mm256_mask_cvtts_roundps_epu64 (xi, m8, hx, 8);
xi = _mm256_maskz_cvtts_roundps_epu64 (m8, hx, 8);
hxi = _mm_cvttspd_epi32 (hxd);
hxi = _mm_mask_cvttspd_epi32 (hxi, m8, hxd);
hxi = _mm_maskz_cvttspd_epi32 (m8, hxd);
hxi = _mm_cvtts_pd_epi32 (hxd);
hxi = _mm_mask_cvtts_pd_epi32 (hxi, m8, hxd);
hxi = _mm_maskz_cvtts_pd_epi32 (m8, hxd);
hxi = _mm_cvttspd_epi64 (hxd);
hxi = _mm_mask_cvttspd_epi64 (hxi, m8, hxd);
hxi = _mm_maskz_cvttspd_epi64 (m8, hxd);
hxi = _mm_cvtts_pd_epi64 (hxd);
hxi = _mm_mask_cvtts_pd_epi64 (hxi, m8, hxd);
hxi = _mm_maskz_cvtts_pd_epi64 (m8, hxd);
hxi = _mm_cvttspd_epu32 (hxd);
hxi = _mm_mask_cvttspd_epu32 (hxi, m8, hxd);
hxi = _mm_maskz_cvttspd_epu32 (m8, hxd);
hxi = _mm_cvtts_pd_epu32 (hxd);
hxi = _mm_mask_cvtts_pd_epu32 (hxi, m8, hxd);
hxi = _mm_maskz_cvtts_pd_epu32 (m8, hxd);
hxi = _mm_cvttspd_epu64 (hxd);
hxi = _mm_mask_cvttspd_epu64 (hxi, m8, hxd);
hxi = _mm_maskz_cvttspd_epu64 (m8, hxd);
hxi = _mm_cvtts_pd_epu64 (hxd);
hxi = _mm_mask_cvtts_pd_epu64 (hxi, m8, hxd);
hxi = _mm_maskz_cvtts_pd_epu64 (m8, hxd);
hxi = _mm_cvttsps_epi32 (hx);
hxi = _mm_mask_cvttsps_epi32 (hxi, m8, hx);
hxi = _mm_maskz_cvttsps_epi32 (m8, hx);
hxi = _mm_cvtts_ps_epi32 (hx);
hxi = _mm_mask_cvtts_ps_epi32 (hxi, m8, hx);
hxi = _mm_maskz_cvtts_ps_epi32 (m8, hx);
hxi = _mm_cvttsps_epi64 (hx);
hxi = _mm_mask_cvttsps_epi64 (hxi, m8, hx);
hxi = _mm_maskz_cvttsps_epi64 (m8, hx);
hxi = _mm_cvtts_ps_epi64 (hx);
hxi = _mm_mask_cvtts_ps_epi64 (hxi, m8, hx);
hxi = _mm_maskz_cvtts_ps_epi64 (m8, hx);
hxi = _mm_cvttsps_epu32 (hx);
hxi = _mm_mask_cvttsps_epu32 (hxi, m8, hx);
hxi = _mm_maskz_cvttsps_epu32 (m8, hx);
hxi = _mm_cvtts_ps_epu32 (hx);
hxi = _mm_mask_cvtts_ps_epu32 (hxi, m8, hx);
hxi = _mm_maskz_cvtts_ps_epu32 (m8, hx);
hxi = _mm_cvttsps_epu64 (hx);
hxi = _mm_mask_cvttsps_epu64 (hxi, m8, hx);
hxi = _mm_maskz_cvttsps_epu64 (m8, hx);
hxi = _mm_cvtts_ps_epu64 (hx);
hxi = _mm_mask_cvtts_ps_epu64 (hxi, m8, hx);
hxi = _mm_maskz_cvtts_ps_epu64 (m8, hx);
i = _mm_cvttssd_epi32 (hxd);
ui = _mm_cvttssd_epu32 (hxd);
i = _mm_cvttsss_epi32 (hx);
ui = _mm_cvttsss_epu32 (hx);
i = _mm_cvtts_sd_epi32 (hxd);
ui = _mm_cvtts_sd_epu32 (hxd);
i = _mm_cvtts_ss_epi32 (hx);
ui = _mm_cvtts_ss_epu32 (hx);
i = _mm_cvtts_roundsd_epi32 (hxd, 8);
ui = _mm_cvtts_roundsd_epu32 (hxd, 8);
i = _mm_cvtts_roundss_epi32 (hx, 8);
ui = _mm_cvtts_roundss_epu32 (hx, 8);
#ifdef __x86_64__
ll = _mm_cvttssd_epi64 (hxd);
ull = _mm_cvttssd_epu64 (hxd);
ll = _mm_cvttsss_epi64 (hx);
ull = _mm_cvttsss_epu64 (hx);
ll = _mm_cvtts_sd_epi64 (hxd);
ull = _mm_cvtts_sd_epu64 (hxd);
ll = _mm_cvtts_ss_epi64 (hx);
ull = _mm_cvtts_ss_epu64 (hx);
ll = _mm_cvtts_roundsd_epi64 (hxd, 8);
ull = _mm_cvtts_roundsd_epu64 (hxd, 8);
ll = _mm_cvtts_roundss_epi64 (hx, 8);

View file

@ -31,7 +31,7 @@ TEST (void)
if (res1 != res1_ref)
abort();
res1 = _mm_cvttssd_epi32 (s.x);
res1 = _mm_cvtts_sd_epi32 (s.x);
if (s.a[0] > INT_MAX)
res1_ref = INT_MAX;
@ -56,7 +56,7 @@ TEST (void)
if (res2 != res2_ref)
abort();
res2 = _mm_cvttssd_epi64 (s.x);
res2 = _mm_cvtts_sd_epi64 (s.x);
if (s.a[0] > LLONG_MAX)
res2_ref = LLONG_MAX;

View file

@ -31,7 +31,7 @@ TEST (void)
if (res1 != res1_ref)
abort();
res1 = _mm_cvttssd_epu32 (s.x);
res1 = _mm_cvtts_sd_epu32 (s.x);
if (s.a[0] > UINT_MAX)
res1_ref = UINT_MAX;
@ -56,7 +56,7 @@ TEST (void)
if (res2 != res2_ref)
abort();
res2 = _mm_cvttssd_epu64 (s.x);
res2 = _mm_cvtts_sd_epu64 (s.x);
if (s.a[0] > ULONG_MAX)
res2_ref = ULONG_MAX;

View file

@ -31,7 +31,7 @@ TEST (void)
if (res1 != res1_ref)
abort();
res1 = _mm_cvttsss_epi32 (s.x);
res1 = _mm_cvtts_ss_epi32 (s.x);
if (s.a[0] > INT_MAX)
res1_ref = INT_MAX;
@ -56,7 +56,7 @@ TEST (void)
if (res2 != res2_ref)
abort();
res2 = _mm_cvttsss_epi64 (s.x);
res2 = _mm_cvtts_ss_epi64 (s.x);
if (s.a[0] > LLONG_MAX)
res2_ref = LLONG_MAX;

View file

@ -30,7 +30,7 @@ TEST (void)
if (res1 != res1_ref)
abort();
res1 = _mm_cvttsss_epu32 (s.x);
res1 = _mm_cvtts_ss_epu32 (s.x);
if (s.a[0] > UINT_MAX)
res1_ref = UINT_MAX;
@ -55,7 +55,7 @@ TEST (void)
if (res2 != res2_ref)
abort();
res2 = _mm_cvttsss_epu64 (s.x);
res2 = _mm_cvtts_ss_epu64 (s.x);
if (s.a[0] > ULONG_MAX)
res2_ref = ULONG_MAX;

View file

@ -1433,30 +1433,30 @@ test_3 (_mm256_mask_cmp_pbh_mask, __mmask16, __mmask16, __m256bh, __m256bh, 1)
test_3 (_mm_mask_cmp_pbh_mask, __mmask8, __mmask8, __m128bh, __m128bh, 1)
/* avx10_2-512satcvtintrin.h */
test_1 (_mm512_ipcvt_roundph_epi8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvt_roundph_epu8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvt_roundps_epi8, __m512i, __m512, 8)
test_1 (_mm512_ipcvt_roundps_epu8, __m512i, __m512, 8)
test_1 (_mm512_ipcvtt_roundph_epi8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvtt_roundph_epu8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvtt_roundps_epi8, __m512i, __m512, 8)
test_1 (_mm512_ipcvtt_roundps_epu8, __m512i, __m512, 8)
test_2 (_mm512_maskz_ipcvt_roundph_epi8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvt_roundph_epu8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvt_roundps_epi8, __m512i, __mmask16, __m512, 8)
test_2 (_mm512_maskz_ipcvt_roundps_epu8, __m512i, __mmask16, __m512, 8)
test_2 (_mm512_maskz_ipcvtt_roundph_epi8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvtt_roundph_epu8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvtt_roundps_epi8, __m512i, __mmask16, __m512, 8)
test_2 (_mm512_maskz_ipcvtt_roundps_epu8, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvt_roundph_epi8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvt_roundph_epu8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvt_roundps_epi8, __m512i, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvt_roundps_epu8, __m512i, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvtt_roundph_epi8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvtt_roundph_epu8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvtt_roundps_epi8, __m512i, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvtt_roundps_epu8, __m512i, __m512i, __mmask16, __m512, 8)
test_1 (_mm512_ipcvts_roundph_epi8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvts_roundph_epu8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvts_roundps_epi8, __m512i, __m512, 8)
test_1 (_mm512_ipcvts_roundps_epu8, __m512i, __m512, 8)
test_1 (_mm512_ipcvtts_roundph_epi8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvtts_roundph_epu8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvtts_roundps_epi8, __m512i, __m512, 8)
test_1 (_mm512_ipcvtts_roundps_epu8, __m512i, __m512, 8)
test_2 (_mm512_maskz_ipcvts_roundph_epi8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvts_roundph_epu8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvts_roundps_epi8, __m512i, __mmask16, __m512, 8)
test_2 (_mm512_maskz_ipcvts_roundps_epu8, __m512i, __mmask16, __m512, 8)
test_2 (_mm512_maskz_ipcvtts_roundph_epi8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvtts_roundph_epu8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvtts_roundps_epi8, __m512i, __mmask16, __m512, 8)
test_2 (_mm512_maskz_ipcvtts_roundps_epu8, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvts_roundph_epi8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvts_roundph_epu8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvts_roundps_epi8, __m512i, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvts_roundps_epu8, __m512i, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvtts_roundph_epi8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvtts_roundph_epu8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvtts_roundps_epi8, __m512i, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvtts_roundps_epu8, __m512i, __m512i, __mmask16, __m512, 8)
test_1 (_mm512_cvtts_roundpd_epi32, __m256i, __m512d, 8)
test_2 (_mm512_maskz_cvtts_roundpd_epi32, __m256i, __mmask8, __m512d, 8)
test_3 (_mm512_mask_cvtts_roundpd_epi32, __m256i, __m256i, __mmask8, __m512d, 8)
@ -1483,30 +1483,30 @@ test_2 (_mm512_maskz_cvtts_roundps_epu64, __m512i, __mmask8, __m256, 8)
test_3 (_mm512_mask_cvtts_roundps_epu64, __m512i, __m512i, __mmask8, __m256, 8)
/* avx10_2satcvtintrin.h */
test_1 (_mm256_ipcvt_roundph_epi8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvt_roundph_epu8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvt_roundps_epi8, __m256i, __m256, 8)
test_1 (_mm256_ipcvt_roundps_epu8, __m256i, __m256, 8)
test_1 (_mm256_ipcvtt_roundph_epi8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvtt_roundph_epu8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvtt_roundps_epi8, __m256i, __m256, 8)
test_1 (_mm256_ipcvtt_roundps_epu8, __m256i, __m256, 8)
test_2 (_mm256_maskz_ipcvt_roundph_epi8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvt_roundph_epu8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvt_roundps_epi8, __m256i, __mmask8, __m256, 8)
test_2 (_mm256_maskz_ipcvt_roundps_epu8, __m256i, __mmask8, __m256, 8)
test_2 (_mm256_maskz_ipcvtt_roundph_epi8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvtt_roundph_epu8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvtt_roundps_epi8, __m256i, __mmask8, __m256, 8)
test_2 (_mm256_maskz_ipcvtt_roundps_epu8, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvt_roundph_epi8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvt_roundph_epu8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvt_roundps_epi8, __m256i, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvt_roundps_epu8, __m256i, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvtt_roundph_epi8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvtt_roundph_epu8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvtt_roundps_epi8, __m256i, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvtt_roundps_epu8, __m256i, __m256i, __mmask8, __m256, 8)
test_1 (_mm256_ipcvts_roundph_epi8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvts_roundph_epu8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvts_roundps_epi8, __m256i, __m256, 8)
test_1 (_mm256_ipcvts_roundps_epu8, __m256i, __m256, 8)
test_1 (_mm256_ipcvtts_roundph_epi8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvtts_roundph_epu8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvtts_roundps_epi8, __m256i, __m256, 8)
test_1 (_mm256_ipcvtts_roundps_epu8, __m256i, __m256, 8)
test_2 (_mm256_maskz_ipcvts_roundph_epi8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvts_roundph_epu8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvts_roundps_epi8, __m256i, __mmask8, __m256, 8)
test_2 (_mm256_maskz_ipcvts_roundps_epu8, __m256i, __mmask8, __m256, 8)
test_2 (_mm256_maskz_ipcvtts_roundph_epi8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvtts_roundph_epu8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvtts_roundps_epi8, __m256i, __mmask8, __m256, 8)
test_2 (_mm256_maskz_ipcvtts_roundps_epu8, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvts_roundph_epi8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvts_roundph_epu8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvts_roundps_epi8, __m256i, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvts_roundps_epu8, __m256i, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvtts_roundph_epi8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvtts_roundph_epu8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvtts_roundps_epi8, __m256i, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvtts_roundps_epu8, __m256i, __m256i, __mmask8, __m256, 8)
test_1 (_mm256_cvtts_roundpd_epi32, __m128i, __m256d, 8)
test_2 (_mm256_maskz_cvtts_roundpd_epi32, __m128i, __mmask8, __m256d, 8)
test_3 (_mm256_mask_cvtts_roundpd_epi32, __m128i, __m128i, __mmask8, __m256d, 8)

View file

@ -1472,30 +1472,30 @@ test_3 (_mm256_mask_cmp_pbh_mask, __mmask16, __mmask16, __m256bh, __m256bh, 1)
test_3 (_mm_mask_cmp_pbh_mask, __mmask8, __mmask8, __m128bh, __m128bh, 1)
/* avx10_2-512satcvtintrin.h */
test_1 (_mm512_ipcvt_roundph_epi8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvt_roundph_epu8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvt_roundps_epi8, __m512i, __m512, 8)
test_1 (_mm512_ipcvt_roundps_epu8, __m512i, __m512, 8)
test_1 (_mm512_ipcvtt_roundph_epi8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvtt_roundph_epu8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvtt_roundps_epi8, __m512i, __m512, 8)
test_1 (_mm512_ipcvtt_roundps_epu8, __m512i, __m512, 8)
test_2 (_mm512_maskz_ipcvt_roundph_epi8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvt_roundph_epu8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvt_roundps_epi8, __m512i, __mmask16, __m512, 8)
test_2 (_mm512_maskz_ipcvt_roundps_epu8, __m512i, __mmask16, __m512, 8)
test_2 (_mm512_maskz_ipcvtt_roundph_epi8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvtt_roundph_epu8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvtt_roundps_epi8, __m512i, __mmask16, __m512, 8)
test_2 (_mm512_maskz_ipcvtt_roundps_epu8, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvt_roundph_epi8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvt_roundph_epu8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvt_roundps_epi8, __m512i, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvt_roundps_epu8, __m512i, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvtt_roundph_epi8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvtt_roundph_epu8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvtt_roundps_epi8, __m512i, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvtt_roundps_epu8, __m512i, __m512i, __mmask16, __m512, 8)
test_1 (_mm512_ipcvts_roundph_epi8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvts_roundph_epu8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvts_roundps_epi8, __m512i, __m512, 8)
test_1 (_mm512_ipcvts_roundps_epu8, __m512i, __m512, 8)
test_1 (_mm512_ipcvtts_roundph_epi8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvtts_roundph_epu8, __m512i, __m512h, 8)
test_1 (_mm512_ipcvtts_roundps_epi8, __m512i, __m512, 8)
test_1 (_mm512_ipcvtts_roundps_epu8, __m512i, __m512, 8)
test_2 (_mm512_maskz_ipcvts_roundph_epi8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvts_roundph_epu8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvts_roundps_epi8, __m512i, __mmask16, __m512, 8)
test_2 (_mm512_maskz_ipcvts_roundps_epu8, __m512i, __mmask16, __m512, 8)
test_2 (_mm512_maskz_ipcvtts_roundph_epi8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvtts_roundph_epu8, __m512i, __mmask32, __m512h, 8)
test_2 (_mm512_maskz_ipcvtts_roundps_epi8, __m512i, __mmask16, __m512, 8)
test_2 (_mm512_maskz_ipcvtts_roundps_epu8, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvts_roundph_epi8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvts_roundph_epu8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvts_roundps_epi8, __m512i, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvts_roundps_epu8, __m512i, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvtts_roundph_epi8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvtts_roundph_epu8, __m512i, __m512i, __mmask32, __m512h, 8)
test_3 (_mm512_mask_ipcvtts_roundps_epi8, __m512i, __m512i, __mmask16, __m512, 8)
test_3 (_mm512_mask_ipcvtts_roundps_epu8, __m512i, __m512i, __mmask16, __m512, 8)
test_1 (_mm512_cvtts_roundpd_epi32, __m256i, __m512d, 8)
test_2 (_mm512_maskz_cvtts_roundpd_epi32, __m256i, __mmask8, __m512d, 8)
test_3 (_mm512_mask_cvtts_roundpd_epi32, __m256i, __m256i, __mmask8, __m512d, 8)
@ -1522,30 +1522,30 @@ test_2 (_mm512_maskz_cvtts_roundps_epu64, __m512i, __mmask8, __m256, 8)
test_3 (_mm512_mask_cvtts_roundps_epu64, __m512i, __m512i, __mmask8, __m256, 8)
/* avx10_2satcvtintrin.h */
test_1 (_mm256_ipcvt_roundph_epi8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvt_roundph_epu8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvt_roundps_epi8, __m256i, __m256, 8)
test_1 (_mm256_ipcvt_roundps_epu8, __m256i, __m256, 8)
test_1 (_mm256_ipcvtt_roundph_epi8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvtt_roundph_epu8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvtt_roundps_epi8, __m256i, __m256, 8)
test_1 (_mm256_ipcvtt_roundps_epu8, __m256i, __m256, 8)
test_2 (_mm256_maskz_ipcvt_roundph_epi8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvt_roundph_epu8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvt_roundps_epi8, __m256i, __mmask8, __m256, 8)
test_2 (_mm256_maskz_ipcvt_roundps_epu8, __m256i, __mmask8, __m256, 8)
test_2 (_mm256_maskz_ipcvtt_roundph_epi8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvtt_roundph_epu8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvtt_roundps_epi8, __m256i, __mmask8, __m256, 8)
test_2 (_mm256_maskz_ipcvtt_roundps_epu8, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvt_roundph_epi8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvt_roundph_epu8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvt_roundps_epi8, __m256i, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvt_roundps_epu8, __m256i, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvtt_roundph_epi8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvtt_roundph_epu8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvtt_roundps_epi8, __m256i, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvtt_roundps_epu8, __m256i, __m256i, __mmask8, __m256, 8)
test_1 (_mm256_ipcvts_roundph_epi8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvts_roundph_epu8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvts_roundps_epi8, __m256i, __m256, 8)
test_1 (_mm256_ipcvts_roundps_epu8, __m256i, __m256, 8)
test_1 (_mm256_ipcvtts_roundph_epi8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvtts_roundph_epu8, __m256i, __m256h, 8)
test_1 (_mm256_ipcvtts_roundps_epi8, __m256i, __m256, 8)
test_1 (_mm256_ipcvtts_roundps_epu8, __m256i, __m256, 8)
test_2 (_mm256_maskz_ipcvts_roundph_epi8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvts_roundph_epu8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvts_roundps_epi8, __m256i, __mmask8, __m256, 8)
test_2 (_mm256_maskz_ipcvts_roundps_epu8, __m256i, __mmask8, __m256, 8)
test_2 (_mm256_maskz_ipcvtts_roundph_epi8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvtts_roundph_epu8, __m256i, __mmask16, __m256h, 8)
test_2 (_mm256_maskz_ipcvtts_roundps_epi8, __m256i, __mmask8, __m256, 8)
test_2 (_mm256_maskz_ipcvtts_roundps_epu8, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvts_roundph_epi8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvts_roundph_epu8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvts_roundps_epi8, __m256i, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvts_roundps_epu8, __m256i, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvtts_roundph_epi8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvtts_roundph_epu8, __m256i, __m256i, __mmask16, __m256h, 8)
test_3 (_mm256_mask_ipcvtts_roundps_epi8, __m256i, __m256i, __mmask8, __m256, 8)
test_3 (_mm256_mask_ipcvtts_roundps_epu8, __m256i, __m256i, __mmask8, __m256, 8)
test_1 (_mm256_cvtts_roundpd_epi32, __m128i, __m256d, 8)
test_2 (_mm256_maskz_cvtts_roundpd_epi32, __m128i, __mmask8, __m256d, 8)
test_3 (_mm256_mask_cvtts_roundpd_epi32, __m128i, __m128i, __mmask8, __m256d, 8)