public inbox for libc-alpha@sourceware.org
 help / color / mirror / Atom feed
From: "H.J. Lu" <hjl.tools@gmail.com>
To: Noah Goldstein <goldstein.w.n@gmail.com>
Cc: GNU C Library <libc-alpha@sourceware.org>
Subject: Re: [PATCH v3] Reversing calculation of __x86_shared_non_temporal_threshold
Date: Wed, 19 Apr 2023 15:43:16 -0700	[thread overview]
Message-ID: <CAMe9rOo5Zd5023k-oMa31w4Q-7NMYSDc4ZYqJty3ecd=b9t+SA@mail.gmail.com> (raw)
In-Reply-To: <CAFUsyf+Kvyfar5md7uhaFXXGj2PUW2VOVM=vMco_z1+KyNBbyg@mail.gmail.com>

On Wed, Apr 19, 2023 at 3:30 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
>
> On Wed, Apr 19, 2023 at 5:26 PM H.J. Lu <hjl.tools@gmail.com> wrote:
> >
> > ---------- Forwarded message ---------
> > From: Patrick McGehearty via Libc-alpha <libc-alpha@sourceware.org>
> > Date: Fri, Sep 25, 2020 at 3:21 PM
> > Subject: [PATCH v3] Reversing calculation of __x86_shared_non_temporal_threshold
> > To: <libc-alpha@sourceware.org>
> >
> >
> > The __x86_shared_non_temporal_threshold determines when memcpy on x86
> > uses non_temporal stores to avoid pushing other data out of the last
> > level cache.
> >
> > This patch proposes to revert the calculation change made by H.J. Lu's
> > patch of June 2, 2017.
> >
> > H.J. Lu's patch selected a threshold suitable for a single thread
> > getting maximum performance. It was tuned using the single threaded
> > large memcpy micro benchmark on an 8 core processor. The last change
> > changes the threshold from using 3/4 of one thread's share of the
> > cache to using 3/4 of the entire cache of a multi-threaded system
> > before switching to non-temporal stores. Multi-threaded systems with
> > more than a few threads are server-class and typically have many
> > active threads. If one thread consumes 3/4 of the available cache for
> > all threads, it will cause other active threads to have data removed
> > from the cache. Two examples show the range of the effect. John
> > McCalpin's widely parallel Stream benchmark, which runs in parallel
> > and fetches data sequentially, saw a 20% slowdown with this patch on
> > an internal system test of 128 threads. This regression was discovered
> > when comparing OL8 performance to OL7.  An example that compares
> > normal stores to non-temporal stores may be found at
> > https://vgatherps.github.io/2018-09-02-nontemporal/.  A simple test
> > shows performance loss of 400 to 500% due to a failure to use
> > nontemporal stores. These performance losses are most likely to occur
> > when the system load is heaviest and good performance is critical.
> >
> > The tunable x86_non_temporal_threshold can be used to override the
> > default for the knowledgable user who really wants maximum cache
> > allocation to a single thread in a multi-threaded system.
> > The manual entry for the tunable has been expanded to provide
> > more information about its purpose.
> >
> >         modified: sysdeps/x86/cacheinfo.c
> >         modified: manual/tunables.texi
> > ---
> >  manual/tunables.texi    |  6 +++++-
> >  sysdeps/x86/cacheinfo.c | 16 +++++++++++-----
> >  2 files changed, 16 insertions(+), 6 deletions(-)
> >
> > diff --git a/manual/tunables.texi b/manual/tunables.texi
> > index b6bb54d..94d4fbd 100644
> > --- a/manual/tunables.texi
> > +++ b/manual/tunables.texi
> > @@ -364,7 +364,11 @@ set shared cache size in bytes for use in memory
> > and string routines.
> >
> >  @deftp Tunable glibc.tune.x86_non_temporal_threshold
> >  The @code{glibc.tune.x86_non_temporal_threshold} tunable allows the user
> > -to set threshold in bytes for non temporal store.
> > +to set threshold in bytes for non temporal store. Non temporal stores
> > +give a hint to the hardware to move data directly to memory without
> > +displacing other data from the cache. This tunable is used by some
> > +platforms to determine when to use non temporal stores in operations
> > +like memmove and memcpy.
> >
> >  This tunable is specific to i386 and x86-64.
> >  @end deftp
> > diff --git a/sysdeps/x86/cacheinfo.c b/sysdeps/x86/cacheinfo.c
> > index b9444dd..42b468d 100644
> > --- a/sysdeps/x86/cacheinfo.c
> > +++ b/sysdeps/x86/cacheinfo.c
> > @@ -778,14 +778,20 @@ intel_bug_no_cache_info:
> >        __x86_shared_cache_size = shared;
> >      }
> >
> > -  /* The large memcpy micro benchmark in glibc shows that 6 times of
> > -     shared cache size is the approximate value above which non-temporal
> > -     store becomes faster on a 8-core processor.  This is the 3/4 of the
> > -     total shared cache size.  */
> > +  /* The default setting for the non_temporal threshold is 3/4 of one
> > +     thread's share of the chip's cache. For most Intel and AMD processors
> > +     with an initial release date between 2017 and 2020, a thread's typical
> > +     share of the cache is from 500 KBytes to 2 MBytes. Using the 3/4
> > +     threshold leaves 125 KBytes to 500 KBytes of the thread's data
> > +     in cache after a maximum temporal copy, which will maintain
> > +     in cache a reasonable portion of the thread's stack and other
> > +     active data. If the threshold is set higher than one thread's
> > +     share of the cache, it has a substantial risk of negatively
> > +     impacting the performance of other threads running on the chip. */
> >    __x86_shared_non_temporal_threshold
> >      = (cpu_features->non_temporal_threshold != 0
> >         ? cpu_features->non_temporal_threshold
> > -       : __x86_shared_cache_size * threads * 3 / 4);
> > +       : __x86_shared_cache_size * 3 / 4);
> >  }
> >
> >  #endif
> > --
> > 1.8.3.1
> >
> >
> >
> > --
> > H.J.
>
>
> I am looking into re-tuning the NT store threshold which appears to be
> too low in many cases.
>
> I've played around with some micro-benchmarks:
> https://github.com/goldsteinn/memcpy-nt-benchmarks
>
> I am finding that for the most part, ERMS stays competitive with
> NT-Stores even as core count increases with heavy read workloads going
> on on other threads.
> See: https://github.com/goldsteinn/memcpy-nt-benchmarks/blob/master/results-skx-pdf/skx-memcpy-4--read.pdf
>
> I saw: https://vgatherps.github.io/2018-09-02-nontemporal/ although
> it's not clear how to reproduce the results in the blog. I also see it
> was only comparing vs standard temporal stores, not ERMS.
>
> Does anyone know of benchmarks or an application that can highlight
> the L3 clobbering issues brought up in this patch?

You can try this:

https://github.com/jeffhammond/STREAM

-- 
H.J.

  reply	other threads:[~2023-04-19 22:43 UTC|newest]

Thread overview: 14+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-09-25 22:21 Patrick McGehearty
2020-09-25 22:26 ` H.J. Lu
2020-09-28 12:55   ` Florian Weimer
2020-09-27 13:54 ` Carlos O'Donell
2020-10-01 16:04   ` Patrick McGehearty
2020-10-01 21:02     ` Carlos O'Donell
     [not found] ` <CAMe9rOr3QUQKGgAnk+UBBq6hLXkU6i8XcNUMKkNRo1iAK=7ceA@mail.gmail.com>
2023-04-19 22:30   ` Noah Goldstein
2023-04-19 22:43     ` H.J. Lu [this message]
2023-04-19 23:24       ` Noah Goldstein
2023-04-20  0:12         ` H.J. Lu
2023-04-20  0:27           ` Noah Goldstein
2023-04-20 16:17             ` H.J. Lu
2023-04-20 20:23               ` Noah Goldstein
2023-04-20 23:50                 ` H.J. Lu

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to='CAMe9rOo5Zd5023k-oMa31w4Q-7NMYSDc4ZYqJty3ecd=b9t+SA@mail.gmail.com' \
    --to=hjl.tools@gmail.com \
    --cc=goldstein.w.n@gmail.com \
    --cc=libc-alpha@sourceware.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).