diff mbox series

[v2] x86: Add new cpu-flag `Prefer_Non_Temporal`

Message ID 20240811055619.2863839-1-goldstein.w.n@gmail.com
State New
Headers show
Series [v2] x86: Add new cpu-flag `Prefer_Non_Temporal` | expand

Commit Message

Noah Goldstein Aug. 11, 2024, 5:56 a.m. UTC
The goal of this flag is to allow targets which don't prefer/have ERMS
to still access the non-temporal memset implementation.
---
 sysdeps/x86/cpu-tunables.c                    |  2 ++
 sysdeps/x86/dl-cacheinfo.h                    | 32 ++++++++++++++++---
 ...cpu-features-preferred_feature_index_1.def |  1 +
 sysdeps/x86/tst-hwcap-tunables.c              |  6 ++--
 sysdeps/x86_64/multiarch/ifunc-memset.h       | 15 ++++++---
 5 files changed, 45 insertions(+), 11 deletions(-)

Comments

H.J. Lu Aug. 12, 2024, 11:37 a.m. UTC | #1
On Sat, Aug 10, 2024 at 10:56 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
>
> The goal of this flag is to allow targets which don't prefer/have ERMS
> to still access the non-temporal memset implementation.
> ---
>  sysdeps/x86/cpu-tunables.c                    |  2 ++
>  sysdeps/x86/dl-cacheinfo.h                    | 32 ++++++++++++++++---
>  ...cpu-features-preferred_feature_index_1.def |  1 +
>  sysdeps/x86/tst-hwcap-tunables.c              |  6 ++--
>  sysdeps/x86_64/multiarch/ifunc-memset.h       | 15 ++++++---
>  5 files changed, 45 insertions(+), 11 deletions(-)
>
> diff --git a/sysdeps/x86/cpu-tunables.c b/sysdeps/x86/cpu-tunables.c
> index ccc6b64dc2..789646ba26 100644
> --- a/sysdeps/x86/cpu-tunables.c
> +++ b/sysdeps/x86/cpu-tunables.c
> @@ -223,6 +223,8 @@ TUNABLE_CALLBACK (set_hwcaps) (tunable_val_t *valp)
>                                                 Fast_Unaligned_Load, 19);
>               CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
>                                                 Fast_Unaligned_Copy, 19);
> +             CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> +                                               Prefer_Non_Temporal, 19);
>             }
>           break;
>         case 20:
> diff --git a/sysdeps/x86/dl-cacheinfo.h b/sysdeps/x86/dl-cacheinfo.h
> index a1c03b8903..b8ba0c098d 100644
> --- a/sysdeps/x86/dl-cacheinfo.h
> +++ b/sysdeps/x86/dl-cacheinfo.h
> @@ -992,7 +992,8 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
>       non_temporal_threshold. Enable this for both Intel and AMD hardware. */
>    unsigned long int memset_non_temporal_threshold = SIZE_MAX;
>    if (!CPU_FEATURES_ARCH_P (cpu_features, Avoid_Non_Temporal_Memset)
> -      && (cpu_features->basic.kind == arch_kind_intel
> +      && (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
> +         || cpu_features->basic.kind == arch_kind_intel

Please set Prefer_Non_Temporal for Intel and AMD.  There is no
need to check arch_kind_intel nor arch_kind_amd.

>           || cpu_features->basic.kind == arch_kind_amd))
>      memset_non_temporal_threshold = non_temporal_threshold;
>
> @@ -1042,14 +1043,37 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
>         slightly better than ERMS.  */
>      rep_stosb_threshold = SIZE_MAX;
>
> +  /*
> +     For memset, the non-temporal implementation is only accessed through the
> +     stosb code. ie:
> +     ```
> +     if (size >= rep_stosb_thresh)
> +     {
> +       if (size >= non_temporal_thresh)
> +     {
> +     do_non_temporal ();
> +     }
> +       do_stosb ();
> +     }
> +     do_normal_vec_loop ();
> +     ```
> +     So if we prefer non-temporal, set `rep_stosb_thresh = non_temporal_thresh`
> +     to enable the implementation. If `rep_stosb_thresh = non_temporal_thresh`,
> +    `rep stosb` will never be used.
> +   */
> +  TUNABLE_SET_WITH_BOUNDS (x86_memset_non_temporal_threshold,
> +                          memset_non_temporal_threshold,
> +                          minimum_non_temporal_threshold, SIZE_MAX);
> +  if (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> +    rep_stosb_threshold
> +       = TUNABLE_GET (x86_memset_non_temporal_threshold, long int, NULL);
> +
> +
>    TUNABLE_SET_WITH_BOUNDS (x86_data_cache_size, data, 0, SIZE_MAX);
>    TUNABLE_SET_WITH_BOUNDS (x86_shared_cache_size, shared, 0, SIZE_MAX);
>    TUNABLE_SET_WITH_BOUNDS (x86_non_temporal_threshold, non_temporal_threshold,
>                            minimum_non_temporal_threshold,
>                            maximum_non_temporal_threshold);
> -  TUNABLE_SET_WITH_BOUNDS (x86_memset_non_temporal_threshold,
> -                          memset_non_temporal_threshold,
> -                          minimum_non_temporal_threshold, SIZE_MAX);
>    TUNABLE_SET_WITH_BOUNDS (x86_rep_movsb_threshold, rep_movsb_threshold,
>                            minimum_rep_movsb_threshold, SIZE_MAX);
>    TUNABLE_SET_WITH_BOUNDS (x86_rep_stosb_threshold, rep_stosb_threshold, 1,
> diff --git a/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def b/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def
> index 61bbbc2e89..f15344a8be 100644
> --- a/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def
> +++ b/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def
> @@ -34,3 +34,4 @@ BIT (MathVec_Prefer_No_AVX512)
>  BIT (Prefer_FSRM)
>  BIT (Avoid_Short_Distance_REP_MOVSB)
>  BIT (Avoid_Non_Temporal_Memset)
> +BIT (Prefer_Non_Temporal)
> \ No newline at end of file
> diff --git a/sysdeps/x86/tst-hwcap-tunables.c b/sysdeps/x86/tst-hwcap-tunables.c
> index 94307283d7..b33e0e92f4 100644
> --- a/sysdeps/x86/tst-hwcap-tunables.c
> +++ b/sysdeps/x86/tst-hwcap-tunables.c
> @@ -60,7 +60,8 @@ static const struct test_t
>      /* Disable everything.  */
>      "-Prefer_ERMS,-Prefer_FSRM,-AVX,-AVX2,-AVX512F,-AVX512VL,"
>      "-SSE4_1,-SSE4_2,-SSSE3,-Fast_Unaligned_Load,-ERMS,"
> -    "-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset",
> +    "-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset,"
> +    "-Prefer_Non_Temporal",
>      test_1,
>      array_length (test_1)
>    },
> @@ -68,7 +69,8 @@ static const struct test_t
>      /* Same as before, but with some empty suboptions.  */
>      ",-,-Prefer_ERMS,-Prefer_FSRM,-AVX,-AVX2,-AVX512F,-AVX512VL,"
>      "-SSE4_1,-SSE4_2,-SSSE3,-Fast_Unaligned_Load,,-,"
> -    "-ERMS,-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset,-,",
> +    "-ERMS,-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset,"
> +    "-Prefer_Non_Temporal,-,",
>      test_1,
>      array_length (test_1)
>    }
> diff --git a/sysdeps/x86_64/multiarch/ifunc-memset.h b/sysdeps/x86_64/multiarch/ifunc-memset.h
> index 7a637ef7ca..27f04e9dce 100644
> --- a/sysdeps/x86_64/multiarch/ifunc-memset.h
> +++ b/sysdeps/x86_64/multiarch/ifunc-memset.h
> @@ -61,7 +61,8 @@ IFUNC_SELECTOR (void)
>           && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, AVX512BW)
>           && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, BMI2))
>         {
> -         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> +         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> +             || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
>             return OPTIMIZE (avx512_unaligned_erms);
>
>           return OPTIMIZE (avx512_unaligned);
> @@ -76,7 +77,8 @@ IFUNC_SELECTOR (void)
>           && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, AVX512BW)
>           && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, BMI2))
>         {
> -         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> +         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> +             || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
>             return OPTIMIZE (evex_unaligned_erms);
>
>           return OPTIMIZE (evex_unaligned);
> @@ -84,7 +86,8 @@ IFUNC_SELECTOR (void)
>
>        if (CPU_FEATURE_USABLE_P (cpu_features, RTM))
>         {
> -         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> +         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> +             || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
>             return OPTIMIZE (avx2_unaligned_erms_rtm);
>
>           return OPTIMIZE (avx2_unaligned_rtm);
> @@ -93,14 +96,16 @@ IFUNC_SELECTOR (void)
>        if (X86_ISA_CPU_FEATURES_ARCH_P (cpu_features,
>                                        Prefer_No_VZEROUPPER, !))
>         {
> -         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> +         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> +             || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
>             return OPTIMIZE (avx2_unaligned_erms);
>
>           return OPTIMIZE (avx2_unaligned);
>         }
>      }
>
> -  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> +  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> +      || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
>      return OPTIMIZE (sse2_unaligned_erms);
>
>    return OPTIMIZE (sse2_unaligned);
> --
> 2.34.1
>
Noah Goldstein Aug. 12, 2024, 7:06 p.m. UTC | #2
On Mon, Aug 12, 2024 at 7:38 PM H.J. Lu <hjl.tools@gmail.com> wrote:
>
> On Sat, Aug 10, 2024 at 10:56 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> >
> > The goal of this flag is to allow targets which don't prefer/have ERMS
> > to still access the non-temporal memset implementation.
> > ---
> >  sysdeps/x86/cpu-tunables.c                    |  2 ++
> >  sysdeps/x86/dl-cacheinfo.h                    | 32 ++++++++++++++++---
> >  ...cpu-features-preferred_feature_index_1.def |  1 +
> >  sysdeps/x86/tst-hwcap-tunables.c              |  6 ++--
> >  sysdeps/x86_64/multiarch/ifunc-memset.h       | 15 ++++++---
> >  5 files changed, 45 insertions(+), 11 deletions(-)
> >
> > diff --git a/sysdeps/x86/cpu-tunables.c b/sysdeps/x86/cpu-tunables.c
> > index ccc6b64dc2..789646ba26 100644
> > --- a/sysdeps/x86/cpu-tunables.c
> > +++ b/sysdeps/x86/cpu-tunables.c
> > @@ -223,6 +223,8 @@ TUNABLE_CALLBACK (set_hwcaps) (tunable_val_t *valp)
> >                                                 Fast_Unaligned_Load, 19);
> >               CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> >                                                 Fast_Unaligned_Copy, 19);
> > +             CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > +                                               Prefer_Non_Temporal, 19);
> >             }
> >           break;
> >         case 20:
> > diff --git a/sysdeps/x86/dl-cacheinfo.h b/sysdeps/x86/dl-cacheinfo.h
> > index a1c03b8903..b8ba0c098d 100644
> > --- a/sysdeps/x86/dl-cacheinfo.h
> > +++ b/sysdeps/x86/dl-cacheinfo.h
> > @@ -992,7 +992,8 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> >       non_temporal_threshold. Enable this for both Intel and AMD hardware. */
> >    unsigned long int memset_non_temporal_threshold = SIZE_MAX;
> >    if (!CPU_FEATURES_ARCH_P (cpu_features, Avoid_Non_Temporal_Memset)
> > -      && (cpu_features->basic.kind == arch_kind_intel
> > +      && (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
> > +         || cpu_features->basic.kind == arch_kind_intel
>
> Please set Prefer_Non_Temporal for Intel and AMD.  There is no
> need to check arch_kind_intel nor arch_kind_amd.

We shouldn't do that. Think you a bit of a misunderstanding.
Prefer_Non_Temporal will use non-temporal but *not erms*
which is not what we want on intel/amd.

Maybe I should rename the variable as I can see nothing about its
name implies that.

How about `Prefer_Non_Temporal_But_Not_ERMS` or is that too
cumbersome?
>
> >           || cpu_features->basic.kind == arch_kind_amd))
> >      memset_non_temporal_threshold = non_temporal_threshold;
> >
> > @@ -1042,14 +1043,37 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> >         slightly better than ERMS.  */
> >      rep_stosb_threshold = SIZE_MAX;
> >
> > +  /*
> > +     For memset, the non-temporal implementation is only accessed through the
> > +     stosb code. ie:
> > +     ```
> > +     if (size >= rep_stosb_thresh)
> > +     {
> > +       if (size >= non_temporal_thresh)
> > +     {
> > +     do_non_temporal ();
> > +     }
> > +       do_stosb ();
> > +     }
> > +     do_normal_vec_loop ();
> > +     ```
> > +     So if we prefer non-temporal, set `rep_stosb_thresh = non_temporal_thresh`
> > +     to enable the implementation. If `rep_stosb_thresh = non_temporal_thresh`,
> > +    `rep stosb` will never be used.
> > +   */
> > +  TUNABLE_SET_WITH_BOUNDS (x86_memset_non_temporal_threshold,
> > +                          memset_non_temporal_threshold,
> > +                          minimum_non_temporal_threshold, SIZE_MAX);
> > +  if (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> > +    rep_stosb_threshold
> > +       = TUNABLE_GET (x86_memset_non_temporal_threshold, long int, NULL);
> > +
> > +
> >    TUNABLE_SET_WITH_BOUNDS (x86_data_cache_size, data, 0, SIZE_MAX);
> >    TUNABLE_SET_WITH_BOUNDS (x86_shared_cache_size, shared, 0, SIZE_MAX);
> >    TUNABLE_SET_WITH_BOUNDS (x86_non_temporal_threshold, non_temporal_threshold,
> >                            minimum_non_temporal_threshold,
> >                            maximum_non_temporal_threshold);
> > -  TUNABLE_SET_WITH_BOUNDS (x86_memset_non_temporal_threshold,
> > -                          memset_non_temporal_threshold,
> > -                          minimum_non_temporal_threshold, SIZE_MAX);
> >    TUNABLE_SET_WITH_BOUNDS (x86_rep_movsb_threshold, rep_movsb_threshold,
> >                            minimum_rep_movsb_threshold, SIZE_MAX);
> >    TUNABLE_SET_WITH_BOUNDS (x86_rep_stosb_threshold, rep_stosb_threshold, 1,
> > diff --git a/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def b/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def
> > index 61bbbc2e89..f15344a8be 100644
> > --- a/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def
> > +++ b/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def
> > @@ -34,3 +34,4 @@ BIT (MathVec_Prefer_No_AVX512)
> >  BIT (Prefer_FSRM)
> >  BIT (Avoid_Short_Distance_REP_MOVSB)
> >  BIT (Avoid_Non_Temporal_Memset)
> > +BIT (Prefer_Non_Temporal)
> > \ No newline at end of file
> > diff --git a/sysdeps/x86/tst-hwcap-tunables.c b/sysdeps/x86/tst-hwcap-tunables.c
> > index 94307283d7..b33e0e92f4 100644
> > --- a/sysdeps/x86/tst-hwcap-tunables.c
> > +++ b/sysdeps/x86/tst-hwcap-tunables.c
> > @@ -60,7 +60,8 @@ static const struct test_t
> >      /* Disable everything.  */
> >      "-Prefer_ERMS,-Prefer_FSRM,-AVX,-AVX2,-AVX512F,-AVX512VL,"
> >      "-SSE4_1,-SSE4_2,-SSSE3,-Fast_Unaligned_Load,-ERMS,"
> > -    "-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset",
> > +    "-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset,"
> > +    "-Prefer_Non_Temporal",
> >      test_1,
> >      array_length (test_1)
> >    },
> > @@ -68,7 +69,8 @@ static const struct test_t
> >      /* Same as before, but with some empty suboptions.  */
> >      ",-,-Prefer_ERMS,-Prefer_FSRM,-AVX,-AVX2,-AVX512F,-AVX512VL,"
> >      "-SSE4_1,-SSE4_2,-SSSE3,-Fast_Unaligned_Load,,-,"
> > -    "-ERMS,-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset,-,",
> > +    "-ERMS,-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset,"
> > +    "-Prefer_Non_Temporal,-,",
> >      test_1,
> >      array_length (test_1)
> >    }
> > diff --git a/sysdeps/x86_64/multiarch/ifunc-memset.h b/sysdeps/x86_64/multiarch/ifunc-memset.h
> > index 7a637ef7ca..27f04e9dce 100644
> > --- a/sysdeps/x86_64/multiarch/ifunc-memset.h
> > +++ b/sysdeps/x86_64/multiarch/ifunc-memset.h
> > @@ -61,7 +61,8 @@ IFUNC_SELECTOR (void)
> >           && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, AVX512BW)
> >           && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, BMI2))
> >         {
> > -         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> > +         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> > +             || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> >             return OPTIMIZE (avx512_unaligned_erms);
> >
> >           return OPTIMIZE (avx512_unaligned);
> > @@ -76,7 +77,8 @@ IFUNC_SELECTOR (void)
> >           && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, AVX512BW)
> >           && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, BMI2))
> >         {
> > -         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> > +         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> > +             || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> >             return OPTIMIZE (evex_unaligned_erms);
> >
> >           return OPTIMIZE (evex_unaligned);
> > @@ -84,7 +86,8 @@ IFUNC_SELECTOR (void)
> >
> >        if (CPU_FEATURE_USABLE_P (cpu_features, RTM))
> >         {
> > -         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> > +         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> > +             || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> >             return OPTIMIZE (avx2_unaligned_erms_rtm);
> >
> >           return OPTIMIZE (avx2_unaligned_rtm);
> > @@ -93,14 +96,16 @@ IFUNC_SELECTOR (void)
> >        if (X86_ISA_CPU_FEATURES_ARCH_P (cpu_features,
> >                                        Prefer_No_VZEROUPPER, !))
> >         {
> > -         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> > +         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> > +             || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> >             return OPTIMIZE (avx2_unaligned_erms);
> >
> >           return OPTIMIZE (avx2_unaligned);
> >         }
> >      }
> >
> > -  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> > +  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> > +      || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> >      return OPTIMIZE (sse2_unaligned_erms);
> >
> >    return OPTIMIZE (sse2_unaligned);
> > --
> > 2.34.1
> >
>
>
> --
> H.J.
H.J. Lu Aug. 12, 2024, 7:45 p.m. UTC | #3
On Mon, Aug 12, 2024 at 12:06 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
>
> On Mon, Aug 12, 2024 at 7:38 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> >
> > On Sat, Aug 10, 2024 at 10:56 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > >
> > > The goal of this flag is to allow targets which don't prefer/have ERMS
> > > to still access the non-temporal memset implementation.
> > > ---
> > >  sysdeps/x86/cpu-tunables.c                    |  2 ++
> > >  sysdeps/x86/dl-cacheinfo.h                    | 32 ++++++++++++++++---
> > >  ...cpu-features-preferred_feature_index_1.def |  1 +
> > >  sysdeps/x86/tst-hwcap-tunables.c              |  6 ++--
> > >  sysdeps/x86_64/multiarch/ifunc-memset.h       | 15 ++++++---
> > >  5 files changed, 45 insertions(+), 11 deletions(-)
> > >
> > > diff --git a/sysdeps/x86/cpu-tunables.c b/sysdeps/x86/cpu-tunables.c
> > > index ccc6b64dc2..789646ba26 100644
> > > --- a/sysdeps/x86/cpu-tunables.c
> > > +++ b/sysdeps/x86/cpu-tunables.c
> > > @@ -223,6 +223,8 @@ TUNABLE_CALLBACK (set_hwcaps) (tunable_val_t *valp)
> > >                                                 Fast_Unaligned_Load, 19);
> > >               CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > >                                                 Fast_Unaligned_Copy, 19);
> > > +             CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > +                                               Prefer_Non_Temporal, 19);
> > >             }
> > >           break;
> > >         case 20:
> > > diff --git a/sysdeps/x86/dl-cacheinfo.h b/sysdeps/x86/dl-cacheinfo.h
> > > index a1c03b8903..b8ba0c098d 100644
> > > --- a/sysdeps/x86/dl-cacheinfo.h
> > > +++ b/sysdeps/x86/dl-cacheinfo.h
> > > @@ -992,7 +992,8 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> > >       non_temporal_threshold. Enable this for both Intel and AMD hardware. */
> > >    unsigned long int memset_non_temporal_threshold = SIZE_MAX;
> > >    if (!CPU_FEATURES_ARCH_P (cpu_features, Avoid_Non_Temporal_Memset)
> > > -      && (cpu_features->basic.kind == arch_kind_intel
> > > +      && (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
> > > +         || cpu_features->basic.kind == arch_kind_intel
> >
> > Please set Prefer_Non_Temporal for Intel and AMD.  There is no
> > need to check arch_kind_intel nor arch_kind_amd.
>
> We shouldn't do that. Think you a bit of a misunderstanding.
> Prefer_Non_Temporal will use non-temporal but *not erms*
> which is not what we want on intel/amd.
>
> Maybe I should rename the variable as I can see nothing about its
> name implies that.
>
> How about `Prefer_Non_Temporal_But_Not_ERMS` or is that too
> cumbersome?

Prefer_Non_Temporal_Without_ERMS?  Please move
rep_stosb_threshold setting close to

  tunable_size = TUNABLE_GET (x86_memset_non_temporal_threshold, long
int, NULL);
  if (tunable_size > minimum_non_temporal_threshold
      && tunable_size <= maximum_non_temporal_threshold)
    memset_non_temporal_threshold = tunable_size;

> >
> > >           || cpu_features->basic.kind == arch_kind_amd))
> > >      memset_non_temporal_threshold = non_temporal_threshold;
> > >
> > > @@ -1042,14 +1043,37 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> > >         slightly better than ERMS.  */
> > >      rep_stosb_threshold = SIZE_MAX;
> > >
> > > +  /*
> > > +     For memset, the non-temporal implementation is only accessed through the
> > > +     stosb code. ie:
> > > +     ```
> > > +     if (size >= rep_stosb_thresh)
> > > +     {
> > > +       if (size >= non_temporal_thresh)
> > > +     {
> > > +     do_non_temporal ();
> > > +     }
> > > +       do_stosb ();
> > > +     }
> > > +     do_normal_vec_loop ();
> > > +     ```
> > > +     So if we prefer non-temporal, set `rep_stosb_thresh = non_temporal_thresh`
> > > +     to enable the implementation. If `rep_stosb_thresh = non_temporal_thresh`,
> > > +    `rep stosb` will never be used.
> > > +   */
> > > +  TUNABLE_SET_WITH_BOUNDS (x86_memset_non_temporal_threshold,
> > > +                          memset_non_temporal_threshold,
> > > +                          minimum_non_temporal_threshold, SIZE_MAX);
> > > +  if (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> > > +    rep_stosb_threshold
> > > +       = TUNABLE_GET (x86_memset_non_temporal_threshold, long int, NULL);
> > > +
> > > +
> > >    TUNABLE_SET_WITH_BOUNDS (x86_data_cache_size, data, 0, SIZE_MAX);
> > >    TUNABLE_SET_WITH_BOUNDS (x86_shared_cache_size, shared, 0, SIZE_MAX);
> > >    TUNABLE_SET_WITH_BOUNDS (x86_non_temporal_threshold, non_temporal_threshold,
> > >                            minimum_non_temporal_threshold,
> > >                            maximum_non_temporal_threshold);
> > > -  TUNABLE_SET_WITH_BOUNDS (x86_memset_non_temporal_threshold,
> > > -                          memset_non_temporal_threshold,
> > > -                          minimum_non_temporal_threshold, SIZE_MAX);
> > >    TUNABLE_SET_WITH_BOUNDS (x86_rep_movsb_threshold, rep_movsb_threshold,
> > >                            minimum_rep_movsb_threshold, SIZE_MAX);
> > >    TUNABLE_SET_WITH_BOUNDS (x86_rep_stosb_threshold, rep_stosb_threshold, 1,
> > > diff --git a/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def b/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def
> > > index 61bbbc2e89..f15344a8be 100644
> > > --- a/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def
> > > +++ b/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def
> > > @@ -34,3 +34,4 @@ BIT (MathVec_Prefer_No_AVX512)
> > >  BIT (Prefer_FSRM)
> > >  BIT (Avoid_Short_Distance_REP_MOVSB)
> > >  BIT (Avoid_Non_Temporal_Memset)
> > > +BIT (Prefer_Non_Temporal)
> > > \ No newline at end of file
> > > diff --git a/sysdeps/x86/tst-hwcap-tunables.c b/sysdeps/x86/tst-hwcap-tunables.c
> > > index 94307283d7..b33e0e92f4 100644
> > > --- a/sysdeps/x86/tst-hwcap-tunables.c
> > > +++ b/sysdeps/x86/tst-hwcap-tunables.c
> > > @@ -60,7 +60,8 @@ static const struct test_t
> > >      /* Disable everything.  */
> > >      "-Prefer_ERMS,-Prefer_FSRM,-AVX,-AVX2,-AVX512F,-AVX512VL,"
> > >      "-SSE4_1,-SSE4_2,-SSSE3,-Fast_Unaligned_Load,-ERMS,"
> > > -    "-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset",
> > > +    "-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset,"
> > > +    "-Prefer_Non_Temporal",
> > >      test_1,
> > >      array_length (test_1)
> > >    },
> > > @@ -68,7 +69,8 @@ static const struct test_t
> > >      /* Same as before, but with some empty suboptions.  */
> > >      ",-,-Prefer_ERMS,-Prefer_FSRM,-AVX,-AVX2,-AVX512F,-AVX512VL,"
> > >      "-SSE4_1,-SSE4_2,-SSSE3,-Fast_Unaligned_Load,,-,"
> > > -    "-ERMS,-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset,-,",
> > > +    "-ERMS,-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset,"
> > > +    "-Prefer_Non_Temporal,-,",
> > >      test_1,
> > >      array_length (test_1)
> > >    }
> > > diff --git a/sysdeps/x86_64/multiarch/ifunc-memset.h b/sysdeps/x86_64/multiarch/ifunc-memset.h
> > > index 7a637ef7ca..27f04e9dce 100644
> > > --- a/sysdeps/x86_64/multiarch/ifunc-memset.h
> > > +++ b/sysdeps/x86_64/multiarch/ifunc-memset.h
> > > @@ -61,7 +61,8 @@ IFUNC_SELECTOR (void)
> > >           && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, AVX512BW)
> > >           && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, BMI2))
> > >         {
> > > -         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> > > +         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> > > +             || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> > >             return OPTIMIZE (avx512_unaligned_erms);
> > >
> > >           return OPTIMIZE (avx512_unaligned);
> > > @@ -76,7 +77,8 @@ IFUNC_SELECTOR (void)
> > >           && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, AVX512BW)
> > >           && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, BMI2))
> > >         {
> > > -         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> > > +         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> > > +             || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> > >             return OPTIMIZE (evex_unaligned_erms);
> > >
> > >           return OPTIMIZE (evex_unaligned);
> > > @@ -84,7 +86,8 @@ IFUNC_SELECTOR (void)
> > >
> > >        if (CPU_FEATURE_USABLE_P (cpu_features, RTM))
> > >         {
> > > -         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> > > +         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> > > +             || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> > >             return OPTIMIZE (avx2_unaligned_erms_rtm);
> > >
> > >           return OPTIMIZE (avx2_unaligned_rtm);
> > > @@ -93,14 +96,16 @@ IFUNC_SELECTOR (void)
> > >        if (X86_ISA_CPU_FEATURES_ARCH_P (cpu_features,
> > >                                        Prefer_No_VZEROUPPER, !))
> > >         {
> > > -         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> > > +         if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> > > +             || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> > >             return OPTIMIZE (avx2_unaligned_erms);
> > >
> > >           return OPTIMIZE (avx2_unaligned);
> > >         }
> > >      }
> > >
> > > -  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> > > +  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
> > > +      || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
> > >      return OPTIMIZE (sse2_unaligned_erms);
> > >
> > >    return OPTIMIZE (sse2_unaligned);
> > > --
> > > 2.34.1
> > >
> >
> >
> > --
> > H.J.
H.J. Lu Aug. 13, 2024, 1:08 p.m. UTC | #4
On Mon, Aug 12, 2024 at 12:45 PM H.J. Lu <hjl.tools@gmail.com> wrote:
>
> On Mon, Aug 12, 2024 at 12:06 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> >
> > On Mon, Aug 12, 2024 at 7:38 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > >
> > > On Sat, Aug 10, 2024 at 10:56 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > > >
> > > > The goal of this flag is to allow targets which don't prefer/have ERMS
> > > > to still access the non-temporal memset implementation.
> > > > ---
> > > >  sysdeps/x86/cpu-tunables.c                    |  2 ++
> > > >  sysdeps/x86/dl-cacheinfo.h                    | 32 ++++++++++++++++---
> > > >  ...cpu-features-preferred_feature_index_1.def |  1 +
> > > >  sysdeps/x86/tst-hwcap-tunables.c              |  6 ++--
> > > >  sysdeps/x86_64/multiarch/ifunc-memset.h       | 15 ++++++---
> > > >  5 files changed, 45 insertions(+), 11 deletions(-)
> > > >
> > > > diff --git a/sysdeps/x86/cpu-tunables.c b/sysdeps/x86/cpu-tunables.c
> > > > index ccc6b64dc2..789646ba26 100644
> > > > --- a/sysdeps/x86/cpu-tunables.c
> > > > +++ b/sysdeps/x86/cpu-tunables.c
> > > > @@ -223,6 +223,8 @@ TUNABLE_CALLBACK (set_hwcaps) (tunable_val_t *valp)
> > > >                                                 Fast_Unaligned_Load, 19);
> > > >               CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > >                                                 Fast_Unaligned_Copy, 19);
> > > > +             CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > > +                                               Prefer_Non_Temporal, 19);
> > > >             }
> > > >           break;
> > > >         case 20:
> > > > diff --git a/sysdeps/x86/dl-cacheinfo.h b/sysdeps/x86/dl-cacheinfo.h
> > > > index a1c03b8903..b8ba0c098d 100644
> > > > --- a/sysdeps/x86/dl-cacheinfo.h
> > > > +++ b/sysdeps/x86/dl-cacheinfo.h
> > > > @@ -992,7 +992,8 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> > > >       non_temporal_threshold. Enable this for both Intel and AMD hardware. */
> > > >    unsigned long int memset_non_temporal_threshold = SIZE_MAX;
> > > >    if (!CPU_FEATURES_ARCH_P (cpu_features, Avoid_Non_Temporal_Memset)
> > > > -      && (cpu_features->basic.kind == arch_kind_intel
> > > > +      && (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
> > > > +         || cpu_features->basic.kind == arch_kind_intel
> > >
> > > Please set Prefer_Non_Temporal for Intel and AMD.  There is no
> > > need to check arch_kind_intel nor arch_kind_amd.
> >
> > We shouldn't do that. Think you a bit of a misunderstanding.
> > Prefer_Non_Temporal will use non-temporal but *not erms*
> > which is not what we want on intel/amd.
> >
> > Maybe I should rename the variable as I can see nothing about its
> > name implies that.
> >
> > How about `Prefer_Non_Temporal_But_Not_ERMS` or is that too
> > cumbersome?
>
> Prefer_Non_Temporal_Without_ERMS?  Please move
> rep_stosb_threshold setting close to
>
>   tunable_size = TUNABLE_GET (x86_memset_non_temporal_threshold, long
> int, NULL);
>   if (tunable_size > minimum_non_temporal_threshold
>       && tunable_size <= maximum_non_temporal_threshold)
>     memset_non_temporal_threshold = tunable_size;
>

How about something like this?
Noah Goldstein Aug. 13, 2024, 1:25 p.m. UTC | #5
On Tue, Aug 13, 2024 at 9:08 PM H.J. Lu <hjl.tools@gmail.com> wrote:
>
> On Mon, Aug 12, 2024 at 12:45 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> >
> > On Mon, Aug 12, 2024 at 12:06 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > >
> > > On Mon, Aug 12, 2024 at 7:38 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > > >
> > > > On Sat, Aug 10, 2024 at 10:56 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > > > >
> > > > > The goal of this flag is to allow targets which don't prefer/have ERMS
> > > > > to still access the non-temporal memset implementation.
> > > > > ---
> > > > >  sysdeps/x86/cpu-tunables.c                    |  2 ++
> > > > >  sysdeps/x86/dl-cacheinfo.h                    | 32 ++++++++++++++++---
> > > > >  ...cpu-features-preferred_feature_index_1.def |  1 +
> > > > >  sysdeps/x86/tst-hwcap-tunables.c              |  6 ++--
> > > > >  sysdeps/x86_64/multiarch/ifunc-memset.h       | 15 ++++++---
> > > > >  5 files changed, 45 insertions(+), 11 deletions(-)
> > > > >
> > > > > diff --git a/sysdeps/x86/cpu-tunables.c b/sysdeps/x86/cpu-tunables.c
> > > > > index ccc6b64dc2..789646ba26 100644
> > > > > --- a/sysdeps/x86/cpu-tunables.c
> > > > > +++ b/sysdeps/x86/cpu-tunables.c
> > > > > @@ -223,6 +223,8 @@ TUNABLE_CALLBACK (set_hwcaps) (tunable_val_t *valp)
> > > > >                                                 Fast_Unaligned_Load, 19);
> > > > >               CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > > >                                                 Fast_Unaligned_Copy, 19);
> > > > > +             CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > > > +                                               Prefer_Non_Temporal, 19);
> > > > >             }
> > > > >           break;
> > > > >         case 20:
> > > > > diff --git a/sysdeps/x86/dl-cacheinfo.h b/sysdeps/x86/dl-cacheinfo.h
> > > > > index a1c03b8903..b8ba0c098d 100644
> > > > > --- a/sysdeps/x86/dl-cacheinfo.h
> > > > > +++ b/sysdeps/x86/dl-cacheinfo.h
> > > > > @@ -992,7 +992,8 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> > > > >       non_temporal_threshold. Enable this for both Intel and AMD hardware. */
> > > > >    unsigned long int memset_non_temporal_threshold = SIZE_MAX;
> > > > >    if (!CPU_FEATURES_ARCH_P (cpu_features, Avoid_Non_Temporal_Memset)
> > > > > -      && (cpu_features->basic.kind == arch_kind_intel
> > > > > +      && (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
> > > > > +         || cpu_features->basic.kind == arch_kind_intel
> > > >
> > > > Please set Prefer_Non_Temporal for Intel and AMD.  There is no
> > > > need to check arch_kind_intel nor arch_kind_amd.
> > >
> > > We shouldn't do that. Think you a bit of a misunderstanding.
> > > Prefer_Non_Temporal will use non-temporal but *not erms*
> > > which is not what we want on intel/amd.
> > >
> > > Maybe I should rename the variable as I can see nothing about its
> > > name implies that.
> > >
> > > How about `Prefer_Non_Temporal_But_Not_ERMS` or is that too
> > > cumbersome?
> >
> > Prefer_Non_Temporal_Without_ERMS?  Please move
> > rep_stosb_threshold setting close to

Sure, thought about that but was worried is implies ERMS not being
available, but think its clearer than what we have now.
> >
> >   tunable_size = TUNABLE_GET (x86_memset_non_temporal_threshold, long
> > int, NULL);
> >   if (tunable_size > minimum_non_temporal_threshold
> >       && tunable_size <= maximum_non_temporal_threshold)
> >     memset_non_temporal_threshold = tunable_size;
> >
>
> How about something like this?

IIUC that will only skip `rep stosb` if we have ERMS. Personally not a fan
of that restrictions. We may find cases where we don't want `rep stosb` at
all despite having ERMS (as may be the case for SKX).
>
> --
> H.J.
H.J. Lu Aug. 13, 2024, 1:28 p.m. UTC | #6
On Tue, Aug 13, 2024 at 6:25 AM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
>
> On Tue, Aug 13, 2024 at 9:08 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> >
> > On Mon, Aug 12, 2024 at 12:45 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > >
> > > On Mon, Aug 12, 2024 at 12:06 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > > >
> > > > On Mon, Aug 12, 2024 at 7:38 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > > > >
> > > > > On Sat, Aug 10, 2024 at 10:56 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > > > > >
> > > > > > The goal of this flag is to allow targets which don't prefer/have ERMS
> > > > > > to still access the non-temporal memset implementation.
> > > > > > ---
> > > > > >  sysdeps/x86/cpu-tunables.c                    |  2 ++
> > > > > >  sysdeps/x86/dl-cacheinfo.h                    | 32 ++++++++++++++++---
> > > > > >  ...cpu-features-preferred_feature_index_1.def |  1 +
> > > > > >  sysdeps/x86/tst-hwcap-tunables.c              |  6 ++--
> > > > > >  sysdeps/x86_64/multiarch/ifunc-memset.h       | 15 ++++++---
> > > > > >  5 files changed, 45 insertions(+), 11 deletions(-)
> > > > > >
> > > > > > diff --git a/sysdeps/x86/cpu-tunables.c b/sysdeps/x86/cpu-tunables.c
> > > > > > index ccc6b64dc2..789646ba26 100644
> > > > > > --- a/sysdeps/x86/cpu-tunables.c
> > > > > > +++ b/sysdeps/x86/cpu-tunables.c
> > > > > > @@ -223,6 +223,8 @@ TUNABLE_CALLBACK (set_hwcaps) (tunable_val_t *valp)
> > > > > >                                                 Fast_Unaligned_Load, 19);
> > > > > >               CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > > > >                                                 Fast_Unaligned_Copy, 19);
> > > > > > +             CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > > > > +                                               Prefer_Non_Temporal, 19);
> > > > > >             }
> > > > > >           break;
> > > > > >         case 20:
> > > > > > diff --git a/sysdeps/x86/dl-cacheinfo.h b/sysdeps/x86/dl-cacheinfo.h
> > > > > > index a1c03b8903..b8ba0c098d 100644
> > > > > > --- a/sysdeps/x86/dl-cacheinfo.h
> > > > > > +++ b/sysdeps/x86/dl-cacheinfo.h
> > > > > > @@ -992,7 +992,8 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> > > > > >       non_temporal_threshold. Enable this for both Intel and AMD hardware. */
> > > > > >    unsigned long int memset_non_temporal_threshold = SIZE_MAX;
> > > > > >    if (!CPU_FEATURES_ARCH_P (cpu_features, Avoid_Non_Temporal_Memset)
> > > > > > -      && (cpu_features->basic.kind == arch_kind_intel
> > > > > > +      && (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
> > > > > > +         || cpu_features->basic.kind == arch_kind_intel
> > > > >
> > > > > Please set Prefer_Non_Temporal for Intel and AMD.  There is no
> > > > > need to check arch_kind_intel nor arch_kind_amd.
> > > >
> > > > We shouldn't do that. Think you a bit of a misunderstanding.
> > > > Prefer_Non_Temporal will use non-temporal but *not erms*
> > > > which is not what we want on intel/amd.
> > > >
> > > > Maybe I should rename the variable as I can see nothing about its
> > > > name implies that.
> > > >
> > > > How about `Prefer_Non_Temporal_But_Not_ERMS` or is that too
> > > > cumbersome?
> > >
> > > Prefer_Non_Temporal_Without_ERMS?  Please move
> > > rep_stosb_threshold setting close to
>
> Sure, thought about that but was worried is implies ERMS not being
> available, but think its clearer than what we have now.
> > >
> > >   tunable_size = TUNABLE_GET (x86_memset_non_temporal_threshold, long
> > > int, NULL);
> > >   if (tunable_size > minimum_non_temporal_threshold
> > >       && tunable_size <= maximum_non_temporal_threshold)
> > >     memset_non_temporal_threshold = tunable_size;
> > >
> >
> > How about something like this?
>
> IIUC that will only skip `rep stosb` if we have ERMS. Personally not a fan

The code is

+  else if (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
+    && !CPU_FEATURE_USABLE_P (cpu_features, ERMS))
+    rep_stosb_threshold = tunable_size;

It will skip `rep stosb` if we DON'T have ERMS.

> of that restrictions. We may find cases where we don't want `rep stosb` at
> all despite having ERMS (as may be the case for SKX).

We can make ERMS unsable in this case.

> >
> > --
> > H.J.
Noah Goldstein Aug. 13, 2024, 1:35 p.m. UTC | #7
On Tue, Aug 13, 2024 at 9:29 PM H.J. Lu <hjl.tools@gmail.com> wrote:
>
> On Tue, Aug 13, 2024 at 6:25 AM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> >
> > On Tue, Aug 13, 2024 at 9:08 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > >
> > > On Mon, Aug 12, 2024 at 12:45 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > > >
> > > > On Mon, Aug 12, 2024 at 12:06 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > > > >
> > > > > On Mon, Aug 12, 2024 at 7:38 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > > > > >
> > > > > > On Sat, Aug 10, 2024 at 10:56 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > > > > > >
> > > > > > > The goal of this flag is to allow targets which don't prefer/have ERMS
> > > > > > > to still access the non-temporal memset implementation.
> > > > > > > ---
> > > > > > >  sysdeps/x86/cpu-tunables.c                    |  2 ++
> > > > > > >  sysdeps/x86/dl-cacheinfo.h                    | 32 ++++++++++++++++---
> > > > > > >  ...cpu-features-preferred_feature_index_1.def |  1 +
> > > > > > >  sysdeps/x86/tst-hwcap-tunables.c              |  6 ++--
> > > > > > >  sysdeps/x86_64/multiarch/ifunc-memset.h       | 15 ++++++---
> > > > > > >  5 files changed, 45 insertions(+), 11 deletions(-)
> > > > > > >
> > > > > > > diff --git a/sysdeps/x86/cpu-tunables.c b/sysdeps/x86/cpu-tunables.c
> > > > > > > index ccc6b64dc2..789646ba26 100644
> > > > > > > --- a/sysdeps/x86/cpu-tunables.c
> > > > > > > +++ b/sysdeps/x86/cpu-tunables.c
> > > > > > > @@ -223,6 +223,8 @@ TUNABLE_CALLBACK (set_hwcaps) (tunable_val_t *valp)
> > > > > > >                                                 Fast_Unaligned_Load, 19);
> > > > > > >               CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > > > > >                                                 Fast_Unaligned_Copy, 19);
> > > > > > > +             CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > > > > > +                                               Prefer_Non_Temporal, 19);
> > > > > > >             }
> > > > > > >           break;
> > > > > > >         case 20:
> > > > > > > diff --git a/sysdeps/x86/dl-cacheinfo.h b/sysdeps/x86/dl-cacheinfo.h
> > > > > > > index a1c03b8903..b8ba0c098d 100644
> > > > > > > --- a/sysdeps/x86/dl-cacheinfo.h
> > > > > > > +++ b/sysdeps/x86/dl-cacheinfo.h
> > > > > > > @@ -992,7 +992,8 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> > > > > > >       non_temporal_threshold. Enable this for both Intel and AMD hardware. */
> > > > > > >    unsigned long int memset_non_temporal_threshold = SIZE_MAX;
> > > > > > >    if (!CPU_FEATURES_ARCH_P (cpu_features, Avoid_Non_Temporal_Memset)
> > > > > > > -      && (cpu_features->basic.kind == arch_kind_intel
> > > > > > > +      && (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
> > > > > > > +         || cpu_features->basic.kind == arch_kind_intel
> > > > > >
> > > > > > Please set Prefer_Non_Temporal for Intel and AMD.  There is no
> > > > > > need to check arch_kind_intel nor arch_kind_amd.
> > > > >
> > > > > We shouldn't do that. Think you a bit of a misunderstanding.
> > > > > Prefer_Non_Temporal will use non-temporal but *not erms*
> > > > > which is not what we want on intel/amd.
> > > > >
> > > > > Maybe I should rename the variable as I can see nothing about its
> > > > > name implies that.
> > > > >
> > > > > How about `Prefer_Non_Temporal_But_Not_ERMS` or is that too
> > > > > cumbersome?
> > > >
> > > > Prefer_Non_Temporal_Without_ERMS?  Please move
> > > > rep_stosb_threshold setting close to
> >
> > Sure, thought about that but was worried is implies ERMS not being
> > available, but think its clearer than what we have now.
> > > >
> > > >   tunable_size = TUNABLE_GET (x86_memset_non_temporal_threshold, long
> > > > int, NULL);
> > > >   if (tunable_size > minimum_non_temporal_threshold
> > > >       && tunable_size <= maximum_non_temporal_threshold)
> > > >     memset_non_temporal_threshold = tunable_size;
> > > >
> > >
> > > How about something like this?
> >
> > IIUC that will only skip `rep stosb` if we have ERMS. Personally not a fan
>
> The code is
>
> +  else if (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
> +    && !CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> +    rep_stosb_threshold = tunable_size;
>
> It will skip `rep stosb` if we DON'T have ERMS.

Yes but it will only set the `rep stosb` threshold if we don't have ERMS
so req non-erms to use the `rep stosb` codes as staging point for non-temporal
which is the point of this patch.

>
> > of that restrictions. We may find cases where we don't want `rep stosb` at
> > all despite having ERMS (as may be the case for SKX).
>
> We can make ERMS unsable in this case.

That would work, although I don't think its easier to start juggling
capabilities.
I don't think we should start disabling the usability of features to
meet preferences.
>
> > >
> > > --
> > > H.J.
>
>
>
> --
> H.J.
H.J. Lu Aug. 13, 2024, 1:52 p.m. UTC | #8
On Tue, Aug 13, 2024 at 6:35 AM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
>
> On Tue, Aug 13, 2024 at 9:29 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> >
> > On Tue, Aug 13, 2024 at 6:25 AM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > >
> > > On Tue, Aug 13, 2024 at 9:08 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > > >
> > > > On Mon, Aug 12, 2024 at 12:45 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > > > >
> > > > > On Mon, Aug 12, 2024 at 12:06 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > > > > >
> > > > > > On Mon, Aug 12, 2024 at 7:38 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > > > > > >
> > > > > > > On Sat, Aug 10, 2024 at 10:56 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > > > > > > >
> > > > > > > > The goal of this flag is to allow targets which don't prefer/have ERMS
> > > > > > > > to still access the non-temporal memset implementation.
> > > > > > > > ---
> > > > > > > >  sysdeps/x86/cpu-tunables.c                    |  2 ++
> > > > > > > >  sysdeps/x86/dl-cacheinfo.h                    | 32 ++++++++++++++++---
> > > > > > > >  ...cpu-features-preferred_feature_index_1.def |  1 +
> > > > > > > >  sysdeps/x86/tst-hwcap-tunables.c              |  6 ++--
> > > > > > > >  sysdeps/x86_64/multiarch/ifunc-memset.h       | 15 ++++++---
> > > > > > > >  5 files changed, 45 insertions(+), 11 deletions(-)
> > > > > > > >
> > > > > > > > diff --git a/sysdeps/x86/cpu-tunables.c b/sysdeps/x86/cpu-tunables.c
> > > > > > > > index ccc6b64dc2..789646ba26 100644
> > > > > > > > --- a/sysdeps/x86/cpu-tunables.c
> > > > > > > > +++ b/sysdeps/x86/cpu-tunables.c
> > > > > > > > @@ -223,6 +223,8 @@ TUNABLE_CALLBACK (set_hwcaps) (tunable_val_t *valp)
> > > > > > > >                                                 Fast_Unaligned_Load, 19);
> > > > > > > >               CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > > > > > >                                                 Fast_Unaligned_Copy, 19);
> > > > > > > > +             CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > > > > > > +                                               Prefer_Non_Temporal, 19);
> > > > > > > >             }
> > > > > > > >           break;
> > > > > > > >         case 20:
> > > > > > > > diff --git a/sysdeps/x86/dl-cacheinfo.h b/sysdeps/x86/dl-cacheinfo.h
> > > > > > > > index a1c03b8903..b8ba0c098d 100644
> > > > > > > > --- a/sysdeps/x86/dl-cacheinfo.h
> > > > > > > > +++ b/sysdeps/x86/dl-cacheinfo.h
> > > > > > > > @@ -992,7 +992,8 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> > > > > > > >       non_temporal_threshold. Enable this for both Intel and AMD hardware. */
> > > > > > > >    unsigned long int memset_non_temporal_threshold = SIZE_MAX;
> > > > > > > >    if (!CPU_FEATURES_ARCH_P (cpu_features, Avoid_Non_Temporal_Memset)
> > > > > > > > -      && (cpu_features->basic.kind == arch_kind_intel
> > > > > > > > +      && (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
> > > > > > > > +         || cpu_features->basic.kind == arch_kind_intel
> > > > > > >
> > > > > > > Please set Prefer_Non_Temporal for Intel and AMD.  There is no
> > > > > > > need to check arch_kind_intel nor arch_kind_amd.
> > > > > >
> > > > > > We shouldn't do that. Think you a bit of a misunderstanding.
> > > > > > Prefer_Non_Temporal will use non-temporal but *not erms*
> > > > > > which is not what we want on intel/amd.
> > > > > >
> > > > > > Maybe I should rename the variable as I can see nothing about its
> > > > > > name implies that.
> > > > > >
> > > > > > How about `Prefer_Non_Temporal_But_Not_ERMS` or is that too
> > > > > > cumbersome?
> > > > >
> > > > > Prefer_Non_Temporal_Without_ERMS?  Please move
> > > > > rep_stosb_threshold setting close to
> > >
> > > Sure, thought about that but was worried is implies ERMS not being
> > > available, but think its clearer than what we have now.
> > > > >
> > > > >   tunable_size = TUNABLE_GET (x86_memset_non_temporal_threshold, long
> > > > > int, NULL);
> > > > >   if (tunable_size > minimum_non_temporal_threshold
> > > > >       && tunable_size <= maximum_non_temporal_threshold)
> > > > >     memset_non_temporal_threshold = tunable_size;
> > > > >
> > > >
> > > > How about something like this?
> > >
> > > IIUC that will only skip `rep stosb` if we have ERMS. Personally not a fan
> >
> > The code is
> >
> > +  else if (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
> > +    && !CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> > +    rep_stosb_threshold = tunable_size;
> >
> > It will skip `rep stosb` if we DON'T have ERMS.
>
> Yes but it will only set the `rep stosb` threshold if we don't have ERMS
> so req non-erms to use the `rep stosb` codes as staging point for non-temporal
> which is the point of this patch.
>
> >
> > > of that restrictions. We may find cases where we don't want `rep stosb` at
> > > all despite having ERMS (as may be the case for SKX).
> >
> > We can make ERMS unsable in this case.
>
> That would work, although I don't think its easier to start juggling
> capabilities.
> I don't think we should start disabling the usability of features to
> meet preferences.

Each feature has a bit to indicate it is available, another bit to
indicate it is usable.  If a feature should be avoided, we turn
off the usable bit.  This is how it works.

> >
> > > >
> > > > --
> > > > H.J.
> >
> >
> >
> > --
> > H.J.
Noah Goldstein Aug. 13, 2024, 1:54 p.m. UTC | #9
On Tue, Aug 13, 2024 at 9:53 PM H.J. Lu <hjl.tools@gmail.com> wrote:
>
> On Tue, Aug 13, 2024 at 6:35 AM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> >
> > On Tue, Aug 13, 2024 at 9:29 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > >
> > > On Tue, Aug 13, 2024 at 6:25 AM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > > >
> > > > On Tue, Aug 13, 2024 at 9:08 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > > > >
> > > > > On Mon, Aug 12, 2024 at 12:45 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > > > > >
> > > > > > On Mon, Aug 12, 2024 at 12:06 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > > > > > >
> > > > > > > On Mon, Aug 12, 2024 at 7:38 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> > > > > > > >
> > > > > > > > On Sat, Aug 10, 2024 at 10:56 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> > > > > > > > >
> > > > > > > > > The goal of this flag is to allow targets which don't prefer/have ERMS
> > > > > > > > > to still access the non-temporal memset implementation.
> > > > > > > > > ---
> > > > > > > > >  sysdeps/x86/cpu-tunables.c                    |  2 ++
> > > > > > > > >  sysdeps/x86/dl-cacheinfo.h                    | 32 ++++++++++++++++---
> > > > > > > > >  ...cpu-features-preferred_feature_index_1.def |  1 +
> > > > > > > > >  sysdeps/x86/tst-hwcap-tunables.c              |  6 ++--
> > > > > > > > >  sysdeps/x86_64/multiarch/ifunc-memset.h       | 15 ++++++---
> > > > > > > > >  5 files changed, 45 insertions(+), 11 deletions(-)
> > > > > > > > >
> > > > > > > > > diff --git a/sysdeps/x86/cpu-tunables.c b/sysdeps/x86/cpu-tunables.c
> > > > > > > > > index ccc6b64dc2..789646ba26 100644
> > > > > > > > > --- a/sysdeps/x86/cpu-tunables.c
> > > > > > > > > +++ b/sysdeps/x86/cpu-tunables.c
> > > > > > > > > @@ -223,6 +223,8 @@ TUNABLE_CALLBACK (set_hwcaps) (tunable_val_t *valp)
> > > > > > > > >                                                 Fast_Unaligned_Load, 19);
> > > > > > > > >               CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > > > > > > >                                                 Fast_Unaligned_Copy, 19);
> > > > > > > > > +             CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
> > > > > > > > > +                                               Prefer_Non_Temporal, 19);
> > > > > > > > >             }
> > > > > > > > >           break;
> > > > > > > > >         case 20:
> > > > > > > > > diff --git a/sysdeps/x86/dl-cacheinfo.h b/sysdeps/x86/dl-cacheinfo.h
> > > > > > > > > index a1c03b8903..b8ba0c098d 100644
> > > > > > > > > --- a/sysdeps/x86/dl-cacheinfo.h
> > > > > > > > > +++ b/sysdeps/x86/dl-cacheinfo.h
> > > > > > > > > @@ -992,7 +992,8 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> > > > > > > > >       non_temporal_threshold. Enable this for both Intel and AMD hardware. */
> > > > > > > > >    unsigned long int memset_non_temporal_threshold = SIZE_MAX;
> > > > > > > > >    if (!CPU_FEATURES_ARCH_P (cpu_features, Avoid_Non_Temporal_Memset)
> > > > > > > > > -      && (cpu_features->basic.kind == arch_kind_intel
> > > > > > > > > +      && (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
> > > > > > > > > +         || cpu_features->basic.kind == arch_kind_intel
> > > > > > > >
> > > > > > > > Please set Prefer_Non_Temporal for Intel and AMD.  There is no
> > > > > > > > need to check arch_kind_intel nor arch_kind_amd.
> > > > > > >
> > > > > > > We shouldn't do that. Think you a bit of a misunderstanding.
> > > > > > > Prefer_Non_Temporal will use non-temporal but *not erms*
> > > > > > > which is not what we want on intel/amd.
> > > > > > >
> > > > > > > Maybe I should rename the variable as I can see nothing about its
> > > > > > > name implies that.
> > > > > > >
> > > > > > > How about `Prefer_Non_Temporal_But_Not_ERMS` or is that too
> > > > > > > cumbersome?
> > > > > >
> > > > > > Prefer_Non_Temporal_Without_ERMS?  Please move
> > > > > > rep_stosb_threshold setting close to
> > > >
> > > > Sure, thought about that but was worried is implies ERMS not being
> > > > available, but think its clearer than what we have now.
> > > > > >
> > > > > >   tunable_size = TUNABLE_GET (x86_memset_non_temporal_threshold, long
> > > > > > int, NULL);
> > > > > >   if (tunable_size > minimum_non_temporal_threshold
> > > > > >       && tunable_size <= maximum_non_temporal_threshold)
> > > > > >     memset_non_temporal_threshold = tunable_size;
> > > > > >
> > > > >
> > > > > How about something like this?
> > > >
> > > > IIUC that will only skip `rep stosb` if we have ERMS. Personally not a fan
> > >
> > > The code is
> > >
> > > +  else if (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
> > > +    && !CPU_FEATURE_USABLE_P (cpu_features, ERMS))
> > > +    rep_stosb_threshold = tunable_size;
> > >
> > > It will skip `rep stosb` if we DON'T have ERMS.
> >
> > Yes but it will only set the `rep stosb` threshold if we don't have ERMS
> > so req non-erms to use the `rep stosb` codes as staging point for non-temporal
> > which is the point of this patch.
> >
> > >
> > > > of that restrictions. We may find cases where we don't want `rep stosb` at
> > > > all despite having ERMS (as may be the case for SKX).
> > >
> > > We can make ERMS unsable in this case.
> >
> > That would work, although I don't think its easier to start juggling
> > capabilities.
> > I don't think we should start disabling the usability of features to
> > meet preferences.
>
> Each feature has a bit to indicate it is available, another bit to
> indicate it is usable.  If a feature should be avoided, we turn
> off the usable bit.  This is how it works.

Ah good point. Although we would need a bit to distiguish between
stosb vs movsb then.

Let me post with rename + refactor so the code is closer. If its still
unclear we can re-evaluate.
>
> > >
> > > > >
> > > > > --
> > > > > H.J.
> > >
> > >
> > >
> > > --
> > > H.J.
>
>
>
> --
> H.J.
diff mbox series

Patch

diff --git a/sysdeps/x86/cpu-tunables.c b/sysdeps/x86/cpu-tunables.c
index ccc6b64dc2..789646ba26 100644
--- a/sysdeps/x86/cpu-tunables.c
+++ b/sysdeps/x86/cpu-tunables.c
@@ -223,6 +223,8 @@  TUNABLE_CALLBACK (set_hwcaps) (tunable_val_t *valp)
 						Fast_Unaligned_Load, 19);
 	      CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
 						Fast_Unaligned_Copy, 19);
+	      CHECK_GLIBC_IFUNC_PREFERRED_BOTH (n, cpu_features,
+						Prefer_Non_Temporal, 19);
 	    }
 	  break;
 	case 20:
diff --git a/sysdeps/x86/dl-cacheinfo.h b/sysdeps/x86/dl-cacheinfo.h
index a1c03b8903..b8ba0c098d 100644
--- a/sysdeps/x86/dl-cacheinfo.h
+++ b/sysdeps/x86/dl-cacheinfo.h
@@ -992,7 +992,8 @@  dl_init_cacheinfo (struct cpu_features *cpu_features)
      non_temporal_threshold. Enable this for both Intel and AMD hardware. */
   unsigned long int memset_non_temporal_threshold = SIZE_MAX;
   if (!CPU_FEATURES_ARCH_P (cpu_features, Avoid_Non_Temporal_Memset)
-      && (cpu_features->basic.kind == arch_kind_intel
+      && (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal)
+	  || cpu_features->basic.kind == arch_kind_intel
 	  || cpu_features->basic.kind == arch_kind_amd))
     memset_non_temporal_threshold = non_temporal_threshold;
 
@@ -1042,14 +1043,37 @@  dl_init_cacheinfo (struct cpu_features *cpu_features)
        slightly better than ERMS.  */
     rep_stosb_threshold = SIZE_MAX;
 
+  /*
+     For memset, the non-temporal implementation is only accessed through the
+     stosb code. ie:
+     ```
+     if (size >= rep_stosb_thresh)
+     {
+    	if (size >= non_temporal_thresh)
+     {
+     do_non_temporal ();
+     }
+    	do_stosb ();
+     }
+     do_normal_vec_loop ();
+     ```
+     So if we prefer non-temporal, set `rep_stosb_thresh = non_temporal_thresh`
+     to enable the implementation. If `rep_stosb_thresh = non_temporal_thresh`,
+    `rep stosb` will never be used.
+   */
+  TUNABLE_SET_WITH_BOUNDS (x86_memset_non_temporal_threshold,
+			   memset_non_temporal_threshold,
+			   minimum_non_temporal_threshold, SIZE_MAX);
+  if (CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
+    rep_stosb_threshold
+	= TUNABLE_GET (x86_memset_non_temporal_threshold, long int, NULL);
+
+
   TUNABLE_SET_WITH_BOUNDS (x86_data_cache_size, data, 0, SIZE_MAX);
   TUNABLE_SET_WITH_BOUNDS (x86_shared_cache_size, shared, 0, SIZE_MAX);
   TUNABLE_SET_WITH_BOUNDS (x86_non_temporal_threshold, non_temporal_threshold,
 			   minimum_non_temporal_threshold,
 			   maximum_non_temporal_threshold);
-  TUNABLE_SET_WITH_BOUNDS (x86_memset_non_temporal_threshold,
-			   memset_non_temporal_threshold,
-			   minimum_non_temporal_threshold, SIZE_MAX);
   TUNABLE_SET_WITH_BOUNDS (x86_rep_movsb_threshold, rep_movsb_threshold,
 			   minimum_rep_movsb_threshold, SIZE_MAX);
   TUNABLE_SET_WITH_BOUNDS (x86_rep_stosb_threshold, rep_stosb_threshold, 1,
diff --git a/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def b/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def
index 61bbbc2e89..f15344a8be 100644
--- a/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def
+++ b/sysdeps/x86/include/cpu-features-preferred_feature_index_1.def
@@ -34,3 +34,4 @@  BIT (MathVec_Prefer_No_AVX512)
 BIT (Prefer_FSRM)
 BIT (Avoid_Short_Distance_REP_MOVSB)
 BIT (Avoid_Non_Temporal_Memset)
+BIT (Prefer_Non_Temporal)
\ No newline at end of file
diff --git a/sysdeps/x86/tst-hwcap-tunables.c b/sysdeps/x86/tst-hwcap-tunables.c
index 94307283d7..b33e0e92f4 100644
--- a/sysdeps/x86/tst-hwcap-tunables.c
+++ b/sysdeps/x86/tst-hwcap-tunables.c
@@ -60,7 +60,8 @@  static const struct test_t
     /* Disable everything.  */
     "-Prefer_ERMS,-Prefer_FSRM,-AVX,-AVX2,-AVX512F,-AVX512VL,"
     "-SSE4_1,-SSE4_2,-SSSE3,-Fast_Unaligned_Load,-ERMS,"
-    "-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset",
+    "-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset,"
+    "-Prefer_Non_Temporal",
     test_1,
     array_length (test_1)
   },
@@ -68,7 +69,8 @@  static const struct test_t
     /* Same as before, but with some empty suboptions.  */
     ",-,-Prefer_ERMS,-Prefer_FSRM,-AVX,-AVX2,-AVX512F,-AVX512VL,"
     "-SSE4_1,-SSE4_2,-SSSE3,-Fast_Unaligned_Load,,-,"
-    "-ERMS,-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset,-,",
+    "-ERMS,-AVX_Fast_Unaligned_Load,-Avoid_Non_Temporal_Memset,"
+    "-Prefer_Non_Temporal,-,",
     test_1,
     array_length (test_1)
   }
diff --git a/sysdeps/x86_64/multiarch/ifunc-memset.h b/sysdeps/x86_64/multiarch/ifunc-memset.h
index 7a637ef7ca..27f04e9dce 100644
--- a/sysdeps/x86_64/multiarch/ifunc-memset.h
+++ b/sysdeps/x86_64/multiarch/ifunc-memset.h
@@ -61,7 +61,8 @@  IFUNC_SELECTOR (void)
 	  && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, AVX512BW)
 	  && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, BMI2))
 	{
-	  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
+	  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
+	      || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
 	    return OPTIMIZE (avx512_unaligned_erms);
 
 	  return OPTIMIZE (avx512_unaligned);
@@ -76,7 +77,8 @@  IFUNC_SELECTOR (void)
 	  && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, AVX512BW)
 	  && X86_ISA_CPU_FEATURE_USABLE_P (cpu_features, BMI2))
 	{
-	  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
+	  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
+	      || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
 	    return OPTIMIZE (evex_unaligned_erms);
 
 	  return OPTIMIZE (evex_unaligned);
@@ -84,7 +86,8 @@  IFUNC_SELECTOR (void)
 
       if (CPU_FEATURE_USABLE_P (cpu_features, RTM))
 	{
-	  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
+	  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
+	      || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
 	    return OPTIMIZE (avx2_unaligned_erms_rtm);
 
 	  return OPTIMIZE (avx2_unaligned_rtm);
@@ -93,14 +96,16 @@  IFUNC_SELECTOR (void)
       if (X86_ISA_CPU_FEATURES_ARCH_P (cpu_features,
 				       Prefer_No_VZEROUPPER, !))
 	{
-	  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
+	  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
+	      || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
 	    return OPTIMIZE (avx2_unaligned_erms);
 
 	  return OPTIMIZE (avx2_unaligned);
 	}
     }
 
-  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
+  if (CPU_FEATURE_USABLE_P (cpu_features, ERMS)
+      || CPU_FEATURES_ARCH_P (cpu_features, Prefer_Non_Temporal))
     return OPTIMIZE (sse2_unaligned_erms);
 
   return OPTIMIZE (sse2_unaligned);