diff mbox series

[v2,24/36] arm: [MVE intrinsics] add vidwdup shape

Message ID 20240904132650.2720446-25-christophe.lyon@linaro.org
State New
Headers show
Series arm: [MVE intrinsics] Re-implement more intrinsics | expand

Commit Message

Christophe Lyon Sept. 4, 2024, 1:26 p.m. UTC
This patch adds the vidwdup shape description for vdwdup and viwdup.

It is very similar to viddup, but accounts for the additional 'wrap'
scalar parameter.

2024-08-21  Christophe Lyon  <christophe.lyon@linaro.org>

	gcc/
	* config/arm/arm-mve-builtins-shapes.cc (vidwdup): New.
	* config/arm/arm-mve-builtins-shapes.h (vidwdup): New.
---
 gcc/config/arm/arm-mve-builtins-shapes.cc | 88 +++++++++++++++++++++++
 gcc/config/arm/arm-mve-builtins-shapes.h  |  1 +
 2 files changed, 89 insertions(+)

Comments

Richard Earnshaw (lists) Oct. 14, 2024, 5:46 p.m. UTC | #1
On 04/09/2024 14:26, Christophe Lyon wrote:
> This patch adds the vidwdup shape description for vdwdup and viwdup.
> 
> It is very similar to viddup, but accounts for the additional 'wrap'
> scalar parameter.
> 
> 2024-08-21  Christophe Lyon  <christophe.lyon@linaro.org>
> 
> 	gcc/
> 	* config/arm/arm-mve-builtins-shapes.cc (vidwdup): New.
> 	* config/arm/arm-mve-builtins-shapes.h (vidwdup): New.

OK.

R.

> ---
>  gcc/config/arm/arm-mve-builtins-shapes.cc | 88 +++++++++++++++++++++++
>  gcc/config/arm/arm-mve-builtins-shapes.h  |  1 +
>  2 files changed, 89 insertions(+)
> 
> diff --git a/gcc/config/arm/arm-mve-builtins-shapes.cc b/gcc/config/arm/arm-mve-builtins-shapes.cc
> index a1d2e243128..510f15ae73a 100644
> --- a/gcc/config/arm/arm-mve-builtins-shapes.cc
> +++ b/gcc/config/arm/arm-mve-builtins-shapes.cc
> @@ -2291,6 +2291,94 @@ struct viddup_def : public overloaded_base<0>
>  };
>  SHAPE (viddup)
>  
> +/* <T0>_t vfoo[_n]_t0(uint32_t, uint32_t, const int)
> +   <T0>_t vfoo[_wb]_t0(uint32_t *, uint32_t, const int)
> +
> +   Shape for vector increment or decrement with wrap and duplicate operations
> +   that take an integer or pointer to integer first argument, an integer second
> +   argument and an immediate, and produce a vector.
> +
> +   Check that 'imm' is one of 1, 2, 4 or 8.
> +
> +   Example: vdwdupq.
> +   uint8x16_t [__arm_]vdwdupq[_n]_u8(uint32_t a, uint32_t b, const int imm)
> +   uint8x16_t [__arm_]vdwdupq[_wb]_u8(uint32_t *a, uint32_t b, const int imm)
> +   uint8x16_t [__arm_]vdwdupq_m[_n_u8](uint8x16_t inactive, uint32_t a, uint32_t b, const int imm, mve_pred16_t p)
> +   uint8x16_t [__arm_]vdwdupq_m[_wb_u8](uint8x16_t inactive, uint32_t *a, uint32_t b, const int imm, mve_pred16_t p)
> +   uint8x16_t [__arm_]vdwdupq_x[_n]_u8(uint32_t a, uint32_t b, const int imm, mve_pred16_t p)
> +   uint8x16_t [__arm_]vdwdupq_x[_wb]_u8(uint32_t *a, uint32_t b, const int imm, mve_pred16_t p)  */
> +struct vidwdup_def : public overloaded_base<0>
> +{
> +  bool
> +  explicit_type_suffix_p (unsigned int i, enum predication_index pred,
> +			  enum mode_suffix_index,
> +			  type_suffix_info) const override
> +  {
> +    return ((i == 0) && (pred != PRED_m));
> +  }
> +
> +  bool
> +  skip_overload_p (enum predication_index, enum mode_suffix_index mode) const override
> +  {
> +    /* For MODE_wb, share the overloaded instance with MODE_n.  */
> +    if (mode == MODE_wb)
> +      return true;
> +
> +    return false;
> +  }
> +
> +  void
> +  build (function_builder &b, const function_group_info &group,
> +	 bool preserve_user_namespace) const override
> +  {
> +    b.add_overloaded_functions (group, MODE_none, preserve_user_namespace);
> +    build_all (b, "v0,su32,su32,su64", group, MODE_n, preserve_user_namespace);
> +    build_all (b, "v0,as,su32,su64", group, MODE_wb, preserve_user_namespace);
> +  }
> +
> +  tree
> +  resolve (function_resolver &r) const override
> +  {
> +    unsigned int i, nargs;
> +    type_suffix_index type_suffix = NUM_TYPE_SUFFIXES;
> +    if (!r.check_gp_argument (3, i, nargs))
> +      return error_mark_node;
> +
> +    type_suffix = r.type_suffix_ids[0];
> +    /* With PRED_m, ther is no type suffix, so infer it from the first (inactive)
> +       argument.  */
> +    if (type_suffix == NUM_TYPE_SUFFIXES)
> +      type_suffix = r.infer_vector_type (0);
> +
> +    unsigned int last_arg = i - 2;
> +    /* Check that last_arg is either scalar or pointer.  */
> +    if (!r.scalar_argument_p (last_arg))
> +      return error_mark_node;
> +
> +    if (!r.scalar_argument_p (last_arg + 1))
> +      return error_mark_node;
> +
> +    if (!r.require_integer_immediate (last_arg + 2))
> +      return error_mark_node;
> +
> +    /* With MODE_n we expect a scalar, with MODE_wb we expect a pointer.  */
> +    mode_suffix_index mode_suffix;
> +    if (POINTER_TYPE_P (r.get_argument_type (last_arg)))
> +      mode_suffix = MODE_wb;
> +    else
> +      mode_suffix = MODE_n;
> +
> +    return r.resolve_to (mode_suffix, type_suffix);
> +  }
> +
> +  bool
> +  check (function_checker &c) const override
> +  {
> +    return c.require_immediate_one_of (2, 1, 2, 4, 8);
> +  }
> +};
> +SHAPE (vidwdup)
> +
>  /* <T0>_t vfoo[_t0](<T0>_t, <T0>_t, mve_pred16_t)
>  
>     i.e. a version of the standard ternary shape in which
> diff --git a/gcc/config/arm/arm-mve-builtins-shapes.h b/gcc/config/arm/arm-mve-builtins-shapes.h
> index 186287c1620..b3d08ab3866 100644
> --- a/gcc/config/arm/arm-mve-builtins-shapes.h
> +++ b/gcc/config/arm/arm-mve-builtins-shapes.h
> @@ -83,6 +83,7 @@ namespace arm_mve
>      extern const function_shape *const vcvt_f32_f16;
>      extern const function_shape *const vcvtx;
>      extern const function_shape *const viddup;
> +    extern const function_shape *const vidwdup;
>      extern const function_shape *const vpsel;
>  
>    } /* end namespace arm_mve::shapes */
diff mbox series

Patch

diff --git a/gcc/config/arm/arm-mve-builtins-shapes.cc b/gcc/config/arm/arm-mve-builtins-shapes.cc
index a1d2e243128..510f15ae73a 100644
--- a/gcc/config/arm/arm-mve-builtins-shapes.cc
+++ b/gcc/config/arm/arm-mve-builtins-shapes.cc
@@ -2291,6 +2291,94 @@  struct viddup_def : public overloaded_base<0>
 };
 SHAPE (viddup)
 
+/* <T0>_t vfoo[_n]_t0(uint32_t, uint32_t, const int)
+   <T0>_t vfoo[_wb]_t0(uint32_t *, uint32_t, const int)
+
+   Shape for vector increment or decrement with wrap and duplicate operations
+   that take an integer or pointer to integer first argument, an integer second
+   argument and an immediate, and produce a vector.
+
+   Check that 'imm' is one of 1, 2, 4 or 8.
+
+   Example: vdwdupq.
+   uint8x16_t [__arm_]vdwdupq[_n]_u8(uint32_t a, uint32_t b, const int imm)
+   uint8x16_t [__arm_]vdwdupq[_wb]_u8(uint32_t *a, uint32_t b, const int imm)
+   uint8x16_t [__arm_]vdwdupq_m[_n_u8](uint8x16_t inactive, uint32_t a, uint32_t b, const int imm, mve_pred16_t p)
+   uint8x16_t [__arm_]vdwdupq_m[_wb_u8](uint8x16_t inactive, uint32_t *a, uint32_t b, const int imm, mve_pred16_t p)
+   uint8x16_t [__arm_]vdwdupq_x[_n]_u8(uint32_t a, uint32_t b, const int imm, mve_pred16_t p)
+   uint8x16_t [__arm_]vdwdupq_x[_wb]_u8(uint32_t *a, uint32_t b, const int imm, mve_pred16_t p)  */
+struct vidwdup_def : public overloaded_base<0>
+{
+  bool
+  explicit_type_suffix_p (unsigned int i, enum predication_index pred,
+			  enum mode_suffix_index,
+			  type_suffix_info) const override
+  {
+    return ((i == 0) && (pred != PRED_m));
+  }
+
+  bool
+  skip_overload_p (enum predication_index, enum mode_suffix_index mode) const override
+  {
+    /* For MODE_wb, share the overloaded instance with MODE_n.  */
+    if (mode == MODE_wb)
+      return true;
+
+    return false;
+  }
+
+  void
+  build (function_builder &b, const function_group_info &group,
+	 bool preserve_user_namespace) const override
+  {
+    b.add_overloaded_functions (group, MODE_none, preserve_user_namespace);
+    build_all (b, "v0,su32,su32,su64", group, MODE_n, preserve_user_namespace);
+    build_all (b, "v0,as,su32,su64", group, MODE_wb, preserve_user_namespace);
+  }
+
+  tree
+  resolve (function_resolver &r) const override
+  {
+    unsigned int i, nargs;
+    type_suffix_index type_suffix = NUM_TYPE_SUFFIXES;
+    if (!r.check_gp_argument (3, i, nargs))
+      return error_mark_node;
+
+    type_suffix = r.type_suffix_ids[0];
+    /* With PRED_m, ther is no type suffix, so infer it from the first (inactive)
+       argument.  */
+    if (type_suffix == NUM_TYPE_SUFFIXES)
+      type_suffix = r.infer_vector_type (0);
+
+    unsigned int last_arg = i - 2;
+    /* Check that last_arg is either scalar or pointer.  */
+    if (!r.scalar_argument_p (last_arg))
+      return error_mark_node;
+
+    if (!r.scalar_argument_p (last_arg + 1))
+      return error_mark_node;
+
+    if (!r.require_integer_immediate (last_arg + 2))
+      return error_mark_node;
+
+    /* With MODE_n we expect a scalar, with MODE_wb we expect a pointer.  */
+    mode_suffix_index mode_suffix;
+    if (POINTER_TYPE_P (r.get_argument_type (last_arg)))
+      mode_suffix = MODE_wb;
+    else
+      mode_suffix = MODE_n;
+
+    return r.resolve_to (mode_suffix, type_suffix);
+  }
+
+  bool
+  check (function_checker &c) const override
+  {
+    return c.require_immediate_one_of (2, 1, 2, 4, 8);
+  }
+};
+SHAPE (vidwdup)
+
 /* <T0>_t vfoo[_t0](<T0>_t, <T0>_t, mve_pred16_t)
 
    i.e. a version of the standard ternary shape in which
diff --git a/gcc/config/arm/arm-mve-builtins-shapes.h b/gcc/config/arm/arm-mve-builtins-shapes.h
index 186287c1620..b3d08ab3866 100644
--- a/gcc/config/arm/arm-mve-builtins-shapes.h
+++ b/gcc/config/arm/arm-mve-builtins-shapes.h
@@ -83,6 +83,7 @@  namespace arm_mve
     extern const function_shape *const vcvt_f32_f16;
     extern const function_shape *const vcvtx;
     extern const function_shape *const viddup;
+    extern const function_shape *const vidwdup;
     extern const function_shape *const vpsel;
 
   } /* end namespace arm_mve::shapes */