SVE2 supports an unpredicated vector integer MUL form that we can emit from our SVE expanders
without using up a predicate registers. This patch does so.
As the SVE MUL expansion currently is templated away through a code iterator I did not split it
off just for this case but instead special-cased it in the define_expand. It seemed somewhat less
invasive than the alternatives but I could split it off more explicitly if others want to.
The div-by-bitmask_1.c testcase is adjusted to expect this new MUL form.
Bootstrapped and tested on aarch64-none-linux-gnu.
gcc/ChangeLog:
PR target/109406
* config/aarch64/aarch64-sve.md (<optab><mode>3): Handle TARGET_SVE2 MUL
case.
* config/aarch64/aarch64-sve2.md (*aarch64_mul_unpredicated_<mode>): New
pattern.
gcc/testsuite/ChangeLog:
PR target/109406
* gcc.target/aarch64/sve2/div-by-bitmask_1.c: Adjust for unpredicated SVE2
MUL.
* gcc.target/aarch64/sve2/unpred_mul_1.c: New test.
UNSPEC_PRED_X))]
"TARGET_SVE"
{
+ /* SVE2 supports the MUL (vectors, unpredicated) form. Emit the simple
+ pattern for it here rather than splitting off the MULT expander
+ separately. */
+ if (TARGET_SVE2 && <CODE> == MULT)
+ {
+ emit_move_insn (operands[0], gen_rtx_MULT (<MODE>mode,
+ operands[1], operands[2]));
+ DONE;
+ }
operands[3] = aarch64_ptrue_reg (<VPRED>mode);
}
)
;; -------------------------------------------------------------------------
;; ---- [INT] Multiplication
;; -------------------------------------------------------------------------
-;; Includes the lane forms of:
+;; Includes the lane and unpredicated forms of:
;; - MUL
;; -------------------------------------------------------------------------
"mul\t%0.<Vetype>, %1.<Vetype>, %2.<Vetype>[%3]"
)
+;; The 2nd and 3rd alternatives are valid for just TARGET_SVE as well but
+;; we include them here to allow matching simpler, unpredicated RTL.
+(define_insn "*aarch64_mul_unpredicated_<mode>"
+ [(set (match_operand:SVE_I 0 "register_operand" "=w,w,?&w")
+ (mult:SVE_I
+ (match_operand:SVE_I 1 "register_operand" "w,0,w")
+ (match_operand:SVE_I 2 "aarch64_sve_vsm_operand" "w,vsm,vsm")))]
+ "TARGET_SVE2"
+ "@
+ mul\t%0.<Vetype>, %1.<Vetype>, %2.<Vetype>
+ mul\t%0.<Vetype>, %0.<Vetype>, #%2
+ movprfx\t%0, %1\;mul\t%0.<Vetype>, %0.<Vetype>, #%2"
+ [(set_attr "movprfx" "*,*,yes")]
+)
+
;; -------------------------------------------------------------------------
;; ---- [INT] Scaled high-part multiplication
;; -------------------------------------------------------------------------
/*
** draw_bitmap1:
** ...
-** mul z[0-9]+.h, p[0-9]+/m, z[0-9]+.h, z[0-9]+.h
+** mul z[0-9]+.h, z[0-9]+.h, z[0-9]+.h
** addhnb z[0-9]+.b, z[0-9]+.h, z[0-9]+.h
** addhnb z[0-9]+.b, z[0-9]+.h, z[0-9]+.h
** ...
/*
** draw_bitmap3:
** ...
-** mul z[0-9]+.s, p[0-9]+/m, z[0-9]+.s, z[0-9]+.s
+** mul z[0-9]+.s, z[0-9]+.s, z[0-9]+.s
** addhnb z[0-9]+.h, z[0-9]+.s, z[0-9]+.s
** addhnb z[0-9]+.h, z[0-9]+.s, z[0-9]+.s
** ...
/*
** draw_bitmap4:
** ...
-** mul z[0-9]+.d, p[0-9]+/m, z[0-9]+.d, z[0-9]+.d
+** mul z[0-9]+.d, z[0-9]+.d, z[0-9]+.d
** addhnb z[0-9]+.s, z[0-9]+.d, z[0-9]+.d
** addhnb z[0-9]+.s, z[0-9]+.d, z[0-9]+.d
** ...
--- /dev/null
+/* { dg-do compile } */
+/* { dg-options "-O2 -ftree-vectorize" } */
+
+#include <stdint.h>
+
+#define N 1024
+
+#define TYPE(N) int##N##_t
+
+#define TEMPLATE(SIZE) \
+void __attribute__ ((noinline, noclone)) \
+f_##SIZE##_##OP \
+ (TYPE(SIZE) *restrict a, TYPE(SIZE) *restrict b, \
+ TYPE(SIZE) *restrict c) \
+{ \
+ for (int i = 0; i < N; i++) \
+ a[i] = b[i] * c[i]; \
+}
+
+TEMPLATE (8);
+TEMPLATE (16);
+TEMPLATE (32);
+TEMPLATE (64);
+
+/* { dg-final { scan-assembler-times {\tmul\tz[0-9]+\.d, z[0-9]+\.d, z[0-9]+\.d} 1 } } */
+/* { dg-final { scan-assembler-times {\tmul\tz[0-9]+\.s, z[0-9]+\.s, z[0-9]+\.s} 1 } } */
+/* { dg-final { scan-assembler-times {\tmul\tz[0-9]+\.h, z[0-9]+\.h, z[0-9]+\.h} 1 } } */
+/* { dg-final { scan-assembler-times {\tmul\tz[0-9]+\.b, z[0-9]+\.b, z[0-9]+\.b} 1 } } */
+