i386: Auto vectorize usdot_prod, udot_prod with AVXVNNIINT16 instruction.
gcc/ChangeLog: * config/i386/sse.md (VI2_AVX2): Delete V32HI since we actually have the same iterator. Also renaming all the occurence to VI2_AVX2_AVX512BW. (usdot_prod<mode>): New define_expand. (udot_prod<mode>): Ditto. gcc/testsuite/ChangeLog: * gcc.target/i386/vnniint16-auto-vectorize-1.c: New test. * gcc.target/i386/vnniint16-auto-vectorize-2.c: Ditto.
This commit is contained in:
parent
37bdeb8f76
commit
85c50b4a50
3 changed files with 170 additions and 30 deletions
|
@ -545,6 +545,9 @@
|
|||
V32HI (V16HI "TARGET_AVX512VL")])
|
||||
|
||||
(define_mode_iterator VI2_AVX2
|
||||
[(V16HI "TARGET_AVX2") V8HI])
|
||||
|
||||
(define_mode_iterator VI2_AVX2_AVX512BW
|
||||
[(V32HI "TARGET_AVX512BW") (V16HI "TARGET_AVX2") V8HI])
|
||||
|
||||
(define_mode_iterator VI2_AVX512F
|
||||
|
@ -637,9 +640,6 @@
|
|||
(V16HI "TARGET_AVX2") V8HI
|
||||
(V8SI "TARGET_AVX2") V4SI])
|
||||
|
||||
(define_mode_iterator VI2_AVX2_AVX512BW
|
||||
[(V32HI "TARGET_AVX512BW") (V16HI "TARGET_AVX2") V8HI])
|
||||
|
||||
(define_mode_iterator VI248_AVX512VL
|
||||
[V32HI V16SI V8DI
|
||||
(V16HI "TARGET_AVX512VL") (V8SI "TARGET_AVX512VL")
|
||||
|
@ -15298,16 +15298,16 @@
|
|||
})
|
||||
|
||||
(define_expand "mul<mode>3<mask_name>"
|
||||
[(set (match_operand:VI2_AVX2 0 "register_operand")
|
||||
(mult:VI2_AVX2 (match_operand:VI2_AVX2 1 "vector_operand")
|
||||
(match_operand:VI2_AVX2 2 "vector_operand")))]
|
||||
[(set (match_operand:VI2_AVX2_AVX512BW 0 "register_operand")
|
||||
(mult:VI2_AVX2_AVX512BW (match_operand:VI2_AVX2_AVX512BW 1 "vector_operand")
|
||||
(match_operand:VI2_AVX2_AVX512BW 2 "vector_operand")))]
|
||||
"TARGET_SSE2 && <mask_mode512bit_condition> && <mask_avx512bw_condition>"
|
||||
"ix86_fixup_binary_operands_no_copy (MULT, <MODE>mode, operands);")
|
||||
|
||||
(define_insn "*mul<mode>3<mask_name>"
|
||||
[(set (match_operand:VI2_AVX2 0 "register_operand" "=x,<v_Yw>")
|
||||
(mult:VI2_AVX2 (match_operand:VI2_AVX2 1 "vector_operand" "%0,<v_Yw>")
|
||||
(match_operand:VI2_AVX2 2 "vector_operand" "xBm,<v_Yw>m")))]
|
||||
[(set (match_operand:VI2_AVX2_AVX512BW 0 "register_operand" "=x,<v_Yw>")
|
||||
(mult:VI2_AVX2_AVX512BW (match_operand:VI2_AVX2_AVX512BW 1 "vector_operand" "%0,<v_Yw>")
|
||||
(match_operand:VI2_AVX2_AVX512BW 2 "vector_operand" "xBm,<v_Yw>m")))]
|
||||
"TARGET_SSE2 && !(MEM_P (operands[1]) && MEM_P (operands[2]))
|
||||
&& <mask_mode512bit_condition> && <mask_avx512bw_condition>"
|
||||
"@
|
||||
|
@ -15320,28 +15320,28 @@
|
|||
(set_attr "mode" "<sseinsnmode>")])
|
||||
|
||||
(define_expand "<s>mul<mode>3_highpart<mask_name>"
|
||||
[(set (match_operand:VI2_AVX2 0 "register_operand")
|
||||
(truncate:VI2_AVX2
|
||||
[(set (match_operand:VI2_AVX2_AVX512BW 0 "register_operand")
|
||||
(truncate:VI2_AVX2_AVX512BW
|
||||
(lshiftrt:<ssedoublemode>
|
||||
(mult:<ssedoublemode>
|
||||
(any_extend:<ssedoublemode>
|
||||
(match_operand:VI2_AVX2 1 "vector_operand"))
|
||||
(match_operand:VI2_AVX2_AVX512BW 1 "vector_operand"))
|
||||
(any_extend:<ssedoublemode>
|
||||
(match_operand:VI2_AVX2 2 "vector_operand")))
|
||||
(match_operand:VI2_AVX2_AVX512BW 2 "vector_operand")))
|
||||
(const_int 16))))]
|
||||
"TARGET_SSE2
|
||||
&& <mask_mode512bit_condition> && <mask_avx512bw_condition>"
|
||||
"ix86_fixup_binary_operands_no_copy (MULT, <MODE>mode, operands);")
|
||||
|
||||
(define_insn "*<s>mul<mode>3_highpart<mask_name>"
|
||||
[(set (match_operand:VI2_AVX2 0 "register_operand" "=x,<v_Yw>")
|
||||
(truncate:VI2_AVX2
|
||||
[(set (match_operand:VI2_AVX2_AVX512BW 0 "register_operand" "=x,<v_Yw>")
|
||||
(truncate:VI2_AVX2_AVX512BW
|
||||
(lshiftrt:<ssedoublemode>
|
||||
(mult:<ssedoublemode>
|
||||
(any_extend:<ssedoublemode>
|
||||
(match_operand:VI2_AVX2 1 "vector_operand" "%0,<v_Yw>"))
|
||||
(match_operand:VI2_AVX2_AVX512BW 1 "vector_operand" "%0,<v_Yw>"))
|
||||
(any_extend:<ssedoublemode>
|
||||
(match_operand:VI2_AVX2 2 "vector_operand" "xBm,<v_Yw>m")))
|
||||
(match_operand:VI2_AVX2_AVX512BW 2 "vector_operand" "xBm,<v_Yw>m")))
|
||||
(const_int 16))))]
|
||||
"TARGET_SSE2 && !(MEM_P (operands[1]) && MEM_P (operands[2]))
|
||||
&& <mask_mode512bit_condition> && <mask_avx512bw_condition>"
|
||||
|
@ -15591,8 +15591,8 @@
|
|||
(define_insn "avx512bw_pmaddwd512<mode><mask_name>"
|
||||
[(set (match_operand:<sseunpackmode> 0 "register_operand" "=v")
|
||||
(unspec:<sseunpackmode>
|
||||
[(match_operand:VI2_AVX2 1 "register_operand" "v")
|
||||
(match_operand:VI2_AVX2 2 "nonimmediate_operand" "vm")]
|
||||
[(match_operand:VI2_AVX2_AVX512BW 1 "register_operand" "v")
|
||||
(match_operand:VI2_AVX2_AVX512BW 2 "nonimmediate_operand" "vm")]
|
||||
UNSPEC_PMADDWD512))]
|
||||
"TARGET_AVX512BW && <mask_mode512bit_condition>"
|
||||
"vpmaddwd\t{%2, %1, %0<mask_operand3>|%0<mask_operand3>, %1, %2}";
|
||||
|
@ -21569,16 +21569,16 @@
|
|||
})
|
||||
|
||||
(define_expand "smulhrs<mode>3"
|
||||
[(set (match_operand:VI2_AVX2 0 "register_operand")
|
||||
(truncate:VI2_AVX2
|
||||
[(set (match_operand:VI2_AVX2_AVX512BW 0 "register_operand")
|
||||
(truncate:VI2_AVX2_AVX512BW
|
||||
(lshiftrt:<ssedoublemode>
|
||||
(plus:<ssedoublemode>
|
||||
(lshiftrt:<ssedoublemode>
|
||||
(mult:<ssedoublemode>
|
||||
(sign_extend:<ssedoublemode>
|
||||
(match_operand:VI2_AVX2 1 "nonimmediate_operand"))
|
||||
(match_operand:VI2_AVX2_AVX512BW 1 "nonimmediate_operand"))
|
||||
(sign_extend:<ssedoublemode>
|
||||
(match_operand:VI2_AVX2 2 "nonimmediate_operand")))
|
||||
(match_operand:VI2_AVX2_AVX512BW 2 "nonimmediate_operand")))
|
||||
(const_int 14))
|
||||
(match_dup 3))
|
||||
(const_int 1))))]
|
||||
|
@ -21589,18 +21589,18 @@
|
|||
})
|
||||
|
||||
(define_insn "*<ssse3_avx2>_pmulhrsw<mode>3<mask_name>"
|
||||
[(set (match_operand:VI2_AVX2 0 "register_operand" "=x,<v_Yw>")
|
||||
(truncate:VI2_AVX2
|
||||
[(set (match_operand:VI2_AVX2_AVX512BW 0 "register_operand" "=x,<v_Yw>")
|
||||
(truncate:VI2_AVX2_AVX512BW
|
||||
(lshiftrt:<ssedoublemode>
|
||||
(plus:<ssedoublemode>
|
||||
(lshiftrt:<ssedoublemode>
|
||||
(mult:<ssedoublemode>
|
||||
(sign_extend:<ssedoublemode>
|
||||
(match_operand:VI2_AVX2 1 "vector_operand" "%0,<v_Yw>"))
|
||||
(match_operand:VI2_AVX2_AVX512BW 1 "vector_operand" "%0,<v_Yw>"))
|
||||
(sign_extend:<ssedoublemode>
|
||||
(match_operand:VI2_AVX2 2 "vector_operand" "xBm,<v_Yw>m")))
|
||||
(match_operand:VI2_AVX2_AVX512BW 2 "vector_operand" "xBm,<v_Yw>m")))
|
||||
(const_int 14))
|
||||
(match_operand:VI2_AVX2 3 "const1_operand"))
|
||||
(match_operand:VI2_AVX2_AVX512BW 3 "const1_operand"))
|
||||
(const_int 1))))]
|
||||
"TARGET_SSSE3 && <mask_mode512bit_condition> && <mask_avx512bw_condition>
|
||||
&& !(MEM_P (operands[1]) && MEM_P (operands[2]))"
|
||||
|
@ -22327,8 +22327,8 @@
|
|||
(set_attr "mode" "<sseinsnmode>")])
|
||||
|
||||
(define_insn "<sse4_1_avx2>_packusdw<mask_name>"
|
||||
[(set (match_operand:VI2_AVX2 0 "register_operand" "=Yr,*x,<v_Yw>")
|
||||
(unspec:VI2_AVX2
|
||||
[(set (match_operand:VI2_AVX2_AVX512BW 0 "register_operand" "=Yr,*x,<v_Yw>")
|
||||
(unspec:VI2_AVX2_AVX512BW
|
||||
[(match_operand:<sseunpackmode> 1 "register_operand" "0,0,<v_Yw>")
|
||||
(match_operand:<sseunpackmode> 2 "vector_operand" "YrBm,*xBm,<v_Yw>m")]
|
||||
UNSPEC_US_TRUNCATE))]
|
||||
|
@ -30340,6 +30340,42 @@
|
|||
(UNSPEC_VPDPWSUD "wsud") (UNSPEC_VPDPWSUDS "wsuds")
|
||||
(UNSPEC_VPDPWUUD "wuud") (UNSPEC_VPDPWUUDS "wuuds")])
|
||||
|
||||
(define_expand "usdot_prod<mode>"
|
||||
[(match_operand:<sseunpackmode> 0 "register_operand")
|
||||
(match_operand:VI2_AVX2 1 "register_operand")
|
||||
(match_operand:VI2_AVX2 2 "register_operand")
|
||||
(match_operand:<sseunpackmode> 3 "register_operand")]
|
||||
"TARGET_AVXVNNIINT16"
|
||||
{
|
||||
operands[1] = lowpart_subreg (<sseunpackmode>mode,
|
||||
force_reg (<MODE>mode, operands[1]),
|
||||
<MODE>mode);
|
||||
operands[2] = lowpart_subreg (<sseunpackmode>mode,
|
||||
force_reg (<MODE>mode, operands[2]),
|
||||
<MODE>mode);
|
||||
emit_insn (gen_vpdpwusd_<SDOT_VPDP_SUF> (operands[0], operands[3],
|
||||
operands[1], operands[2]));
|
||||
DONE;
|
||||
})
|
||||
|
||||
(define_expand "udot_prod<mode>"
|
||||
[(match_operand:<sseunpackmode> 0 "register_operand")
|
||||
(match_operand:VI2_AVX2 1 "register_operand")
|
||||
(match_operand:VI2_AVX2 2 "register_operand")
|
||||
(match_operand:<sseunpackmode> 3 "register_operand")]
|
||||
"TARGET_AVXVNNIINT16"
|
||||
{
|
||||
operands[1] = lowpart_subreg (<sseunpackmode>mode,
|
||||
force_reg (<MODE>mode, operands[1]),
|
||||
<MODE>mode);
|
||||
operands[2] = lowpart_subreg (<sseunpackmode>mode,
|
||||
force_reg (<MODE>mode, operands[2]),
|
||||
<MODE>mode);
|
||||
emit_insn (gen_vpdpwuud_<SDOT_VPDP_SUF> (operands[0], operands[3],
|
||||
operands[1], operands[2]));
|
||||
DONE;
|
||||
})
|
||||
|
||||
(define_insn "vpdp<vpdpwprodtype>_<mode>"
|
||||
[(set (match_operand:VI4_AVX 0 "register_operand" "=x")
|
||||
(unspec:VI4_AVX
|
||||
|
|
28
gcc/testsuite/gcc.target/i386/vnniint16-auto-vectorize-1.c
Normal file
28
gcc/testsuite/gcc.target/i386/vnniint16-auto-vectorize-1.c
Normal file
|
@ -0,0 +1,28 @@
|
|||
/* { dg-do compile } */
|
||||
/* { dg-options "-mavxvnniint16 -O2" } */
|
||||
/* { dg-final { scan-assembler "vpdpwusd\t" } } */
|
||||
/* { dg-final { scan-assembler "vpdpwuud\t" } } */
|
||||
|
||||
int __attribute__((noinline, noclone, optimize("tree-vectorize")))
|
||||
usdot_prod_hi (unsigned short * restrict a, short * restrict b,
|
||||
int c, int n)
|
||||
{
|
||||
int i;
|
||||
for (i = 0; i < n; i++)
|
||||
{
|
||||
c += ((int) a[i] * (int) b[i]);
|
||||
}
|
||||
return c;
|
||||
}
|
||||
|
||||
int __attribute__((noinline, noclone, optimize("tree-vectorize")))
|
||||
udot_prod_hi (unsigned short * restrict a, unsigned short *restrict b,
|
||||
int c, int n)
|
||||
{
|
||||
int i;
|
||||
for (i = 0; i < n; i++)
|
||||
{
|
||||
c += ((int) a[i] * (int) b[i]);
|
||||
}
|
||||
return c;
|
||||
}
|
76
gcc/testsuite/gcc.target/i386/vnniint16-auto-vectorize-2.c
Normal file
76
gcc/testsuite/gcc.target/i386/vnniint16-auto-vectorize-2.c
Normal file
|
@ -0,0 +1,76 @@
|
|||
/* { dg-do run } */
|
||||
/* { dg-options "-O2 -mavxvnniint16" } */
|
||||
/* { dg-require-effective-target avxvnniint16 } */
|
||||
|
||||
#define AVXVNNIINT16
|
||||
#ifndef CHECK
|
||||
#define CHECK "avx-check.h"
|
||||
#endif
|
||||
|
||||
#ifndef TEST
|
||||
#define TEST avx_test
|
||||
#endif
|
||||
|
||||
#include CHECK
|
||||
#include "vnniint16-auto-vectorize-1.c"
|
||||
|
||||
#define N 256
|
||||
|
||||
short a_i16[N];
|
||||
unsigned short b_u16[N], c_u16[N], d_u16[N];
|
||||
int i16_exp, i16_ref;
|
||||
|
||||
int __attribute__((noinline, noclone, optimize("no-tree-vectorize")))
|
||||
udot_prod_hi_scalar (unsigned short * restrict a, unsigned short * restrict b,
|
||||
int c, int n)
|
||||
{
|
||||
int i;
|
||||
for (i = 0; i < n; i++)
|
||||
{
|
||||
c += ((int) a[i] * (int) b[i]);
|
||||
}
|
||||
return c;
|
||||
}
|
||||
|
||||
int __attribute__((noinline, noclone, optimize("no-tree-vectorize")))
|
||||
usdot_prod_hi_scalar (unsigned short * restrict a, short *restrict b,
|
||||
int c, int n)
|
||||
{
|
||||
int i;
|
||||
for (i = 0; i < n; i++)
|
||||
{
|
||||
c += ((int) a[i] * (int) b[i]);
|
||||
}
|
||||
return c;
|
||||
}
|
||||
|
||||
void init ()
|
||||
{
|
||||
int i;
|
||||
|
||||
i16_exp = i16_ref = 65535;
|
||||
|
||||
for (i = 0; i < N; i++)
|
||||
{
|
||||
a_i16[i] = -i + 2;
|
||||
b_u16[i] = i * 2;
|
||||
c_u16[i] = i * 3;
|
||||
d_u16[i] = i * 4;
|
||||
}
|
||||
}
|
||||
|
||||
void
|
||||
TEST (void)
|
||||
{
|
||||
init ();
|
||||
i16_exp = usdot_prod_hi (a_i16, b_u16, i16_exp, N);
|
||||
i16_ref = usdot_prod_hi_scalar (a_i16, b_u16, i16_ref, N);
|
||||
if (i16_exp != i16_ref)
|
||||
abort ();
|
||||
|
||||
init ();
|
||||
i16_exp = udot_prod_hi (c_u16, d_u16, i16_exp, N);
|
||||
i16_ref = udot_prod_hi_scalar (c_u16, d_u16, i16_ref, N);
|
||||
if (i16_exp != i16_ref)
|
||||
abort ();
|
||||
}
|
Loading…
Add table
Reference in a new issue