Message ID | 20240904132650.2720446-25-christophe.lyon@linaro.org |
---|---|
State | New |
Headers | show |
Series | arm: [MVE intrinsics] Re-implement more intrinsics | expand |
On 04/09/2024 14:26, Christophe Lyon wrote: > This patch adds the vidwdup shape description for vdwdup and viwdup. > > It is very similar to viddup, but accounts for the additional 'wrap' > scalar parameter. > > 2024-08-21 Christophe Lyon <christophe.lyon@linaro.org> > > gcc/ > * config/arm/arm-mve-builtins-shapes.cc (vidwdup): New. > * config/arm/arm-mve-builtins-shapes.h (vidwdup): New. OK. R. > --- > gcc/config/arm/arm-mve-builtins-shapes.cc | 88 +++++++++++++++++++++++ > gcc/config/arm/arm-mve-builtins-shapes.h | 1 + > 2 files changed, 89 insertions(+) > > diff --git a/gcc/config/arm/arm-mve-builtins-shapes.cc b/gcc/config/arm/arm-mve-builtins-shapes.cc > index a1d2e243128..510f15ae73a 100644 > --- a/gcc/config/arm/arm-mve-builtins-shapes.cc > +++ b/gcc/config/arm/arm-mve-builtins-shapes.cc > @@ -2291,6 +2291,94 @@ struct viddup_def : public overloaded_base<0> > }; > SHAPE (viddup) > > +/* <T0>_t vfoo[_n]_t0(uint32_t, uint32_t, const int) > + <T0>_t vfoo[_wb]_t0(uint32_t *, uint32_t, const int) > + > + Shape for vector increment or decrement with wrap and duplicate operations > + that take an integer or pointer to integer first argument, an integer second > + argument and an immediate, and produce a vector. > + > + Check that 'imm' is one of 1, 2, 4 or 8. > + > + Example: vdwdupq. > + uint8x16_t [__arm_]vdwdupq[_n]_u8(uint32_t a, uint32_t b, const int imm) > + uint8x16_t [__arm_]vdwdupq[_wb]_u8(uint32_t *a, uint32_t b, const int imm) > + uint8x16_t [__arm_]vdwdupq_m[_n_u8](uint8x16_t inactive, uint32_t a, uint32_t b, const int imm, mve_pred16_t p) > + uint8x16_t [__arm_]vdwdupq_m[_wb_u8](uint8x16_t inactive, uint32_t *a, uint32_t b, const int imm, mve_pred16_t p) > + uint8x16_t [__arm_]vdwdupq_x[_n]_u8(uint32_t a, uint32_t b, const int imm, mve_pred16_t p) > + uint8x16_t [__arm_]vdwdupq_x[_wb]_u8(uint32_t *a, uint32_t b, const int imm, mve_pred16_t p) */ > +struct vidwdup_def : public overloaded_base<0> > +{ > + bool > + explicit_type_suffix_p (unsigned int i, enum predication_index pred, > + enum mode_suffix_index, > + type_suffix_info) const override > + { > + return ((i == 0) && (pred != PRED_m)); > + } > + > + bool > + skip_overload_p (enum predication_index, enum mode_suffix_index mode) const override > + { > + /* For MODE_wb, share the overloaded instance with MODE_n. */ > + if (mode == MODE_wb) > + return true; > + > + return false; > + } > + > + void > + build (function_builder &b, const function_group_info &group, > + bool preserve_user_namespace) const override > + { > + b.add_overloaded_functions (group, MODE_none, preserve_user_namespace); > + build_all (b, "v0,su32,su32,su64", group, MODE_n, preserve_user_namespace); > + build_all (b, "v0,as,su32,su64", group, MODE_wb, preserve_user_namespace); > + } > + > + tree > + resolve (function_resolver &r) const override > + { > + unsigned int i, nargs; > + type_suffix_index type_suffix = NUM_TYPE_SUFFIXES; > + if (!r.check_gp_argument (3, i, nargs)) > + return error_mark_node; > + > + type_suffix = r.type_suffix_ids[0]; > + /* With PRED_m, ther is no type suffix, so infer it from the first (inactive) > + argument. */ > + if (type_suffix == NUM_TYPE_SUFFIXES) > + type_suffix = r.infer_vector_type (0); > + > + unsigned int last_arg = i - 2; > + /* Check that last_arg is either scalar or pointer. */ > + if (!r.scalar_argument_p (last_arg)) > + return error_mark_node; > + > + if (!r.scalar_argument_p (last_arg + 1)) > + return error_mark_node; > + > + if (!r.require_integer_immediate (last_arg + 2)) > + return error_mark_node; > + > + /* With MODE_n we expect a scalar, with MODE_wb we expect a pointer. */ > + mode_suffix_index mode_suffix; > + if (POINTER_TYPE_P (r.get_argument_type (last_arg))) > + mode_suffix = MODE_wb; > + else > + mode_suffix = MODE_n; > + > + return r.resolve_to (mode_suffix, type_suffix); > + } > + > + bool > + check (function_checker &c) const override > + { > + return c.require_immediate_one_of (2, 1, 2, 4, 8); > + } > +}; > +SHAPE (vidwdup) > + > /* <T0>_t vfoo[_t0](<T0>_t, <T0>_t, mve_pred16_t) > > i.e. a version of the standard ternary shape in which > diff --git a/gcc/config/arm/arm-mve-builtins-shapes.h b/gcc/config/arm/arm-mve-builtins-shapes.h > index 186287c1620..b3d08ab3866 100644 > --- a/gcc/config/arm/arm-mve-builtins-shapes.h > +++ b/gcc/config/arm/arm-mve-builtins-shapes.h > @@ -83,6 +83,7 @@ namespace arm_mve > extern const function_shape *const vcvt_f32_f16; > extern const function_shape *const vcvtx; > extern const function_shape *const viddup; > + extern const function_shape *const vidwdup; > extern const function_shape *const vpsel; > > } /* end namespace arm_mve::shapes */
diff --git a/gcc/config/arm/arm-mve-builtins-shapes.cc b/gcc/config/arm/arm-mve-builtins-shapes.cc index a1d2e243128..510f15ae73a 100644 --- a/gcc/config/arm/arm-mve-builtins-shapes.cc +++ b/gcc/config/arm/arm-mve-builtins-shapes.cc @@ -2291,6 +2291,94 @@ struct viddup_def : public overloaded_base<0> }; SHAPE (viddup) +/* <T0>_t vfoo[_n]_t0(uint32_t, uint32_t, const int) + <T0>_t vfoo[_wb]_t0(uint32_t *, uint32_t, const int) + + Shape for vector increment or decrement with wrap and duplicate operations + that take an integer or pointer to integer first argument, an integer second + argument and an immediate, and produce a vector. + + Check that 'imm' is one of 1, 2, 4 or 8. + + Example: vdwdupq. + uint8x16_t [__arm_]vdwdupq[_n]_u8(uint32_t a, uint32_t b, const int imm) + uint8x16_t [__arm_]vdwdupq[_wb]_u8(uint32_t *a, uint32_t b, const int imm) + uint8x16_t [__arm_]vdwdupq_m[_n_u8](uint8x16_t inactive, uint32_t a, uint32_t b, const int imm, mve_pred16_t p) + uint8x16_t [__arm_]vdwdupq_m[_wb_u8](uint8x16_t inactive, uint32_t *a, uint32_t b, const int imm, mve_pred16_t p) + uint8x16_t [__arm_]vdwdupq_x[_n]_u8(uint32_t a, uint32_t b, const int imm, mve_pred16_t p) + uint8x16_t [__arm_]vdwdupq_x[_wb]_u8(uint32_t *a, uint32_t b, const int imm, mve_pred16_t p) */ +struct vidwdup_def : public overloaded_base<0> +{ + bool + explicit_type_suffix_p (unsigned int i, enum predication_index pred, + enum mode_suffix_index, + type_suffix_info) const override + { + return ((i == 0) && (pred != PRED_m)); + } + + bool + skip_overload_p (enum predication_index, enum mode_suffix_index mode) const override + { + /* For MODE_wb, share the overloaded instance with MODE_n. */ + if (mode == MODE_wb) + return true; + + return false; + } + + void + build (function_builder &b, const function_group_info &group, + bool preserve_user_namespace) const override + { + b.add_overloaded_functions (group, MODE_none, preserve_user_namespace); + build_all (b, "v0,su32,su32,su64", group, MODE_n, preserve_user_namespace); + build_all (b, "v0,as,su32,su64", group, MODE_wb, preserve_user_namespace); + } + + tree + resolve (function_resolver &r) const override + { + unsigned int i, nargs; + type_suffix_index type_suffix = NUM_TYPE_SUFFIXES; + if (!r.check_gp_argument (3, i, nargs)) + return error_mark_node; + + type_suffix = r.type_suffix_ids[0]; + /* With PRED_m, ther is no type suffix, so infer it from the first (inactive) + argument. */ + if (type_suffix == NUM_TYPE_SUFFIXES) + type_suffix = r.infer_vector_type (0); + + unsigned int last_arg = i - 2; + /* Check that last_arg is either scalar or pointer. */ + if (!r.scalar_argument_p (last_arg)) + return error_mark_node; + + if (!r.scalar_argument_p (last_arg + 1)) + return error_mark_node; + + if (!r.require_integer_immediate (last_arg + 2)) + return error_mark_node; + + /* With MODE_n we expect a scalar, with MODE_wb we expect a pointer. */ + mode_suffix_index mode_suffix; + if (POINTER_TYPE_P (r.get_argument_type (last_arg))) + mode_suffix = MODE_wb; + else + mode_suffix = MODE_n; + + return r.resolve_to (mode_suffix, type_suffix); + } + + bool + check (function_checker &c) const override + { + return c.require_immediate_one_of (2, 1, 2, 4, 8); + } +}; +SHAPE (vidwdup) + /* <T0>_t vfoo[_t0](<T0>_t, <T0>_t, mve_pred16_t) i.e. a version of the standard ternary shape in which diff --git a/gcc/config/arm/arm-mve-builtins-shapes.h b/gcc/config/arm/arm-mve-builtins-shapes.h index 186287c1620..b3d08ab3866 100644 --- a/gcc/config/arm/arm-mve-builtins-shapes.h +++ b/gcc/config/arm/arm-mve-builtins-shapes.h @@ -83,6 +83,7 @@ namespace arm_mve extern const function_shape *const vcvt_f32_f16; extern const function_shape *const vcvtx; extern const function_shape *const viddup; + extern const function_shape *const vidwdup; extern const function_shape *const vpsel; } /* end namespace arm_mve::shapes */