From: "H.J. Lu" <hjl.tools@gmail.com>
To: Noah Goldstein <goldstein.w.n@gmail.com>
Cc: libc-alpha@sourceware.org, carlos@systemhalted.org
Subject: Re: [PATCH v2] x86: Increase `non_temporal_threshold` to roughly `sizeof_L3 / 2`
Date: Mon, 24 Apr 2023 15:48:28 -0700 [thread overview]
Message-ID: <CAMe9rOokMDgEZwvMp1F8P3CO7E-O0YTPM+jA-PhKQK93Stk8HQ@mail.gmail.com> (raw)
In-Reply-To: <20230424223045.2066606-1-goldstein.w.n@gmail.com>
On Mon, Apr 24, 2023 at 3:30 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
>
> Current `non_temporal_threshold` set to roughly '3/4 * sizeof_L3 /
> ncores_per_socket'. This patch updates that value to roughly
> 'sizeof_L3 / 2`
>
> The original value (specifically dividing the `ncores_per_socket`) was
> done to limit the amount of other threads' data a `memcpy`/`memset`
> could evict.
>
> Dividing by 'ncores_per_socket', however leads to exceedingly low
> non-temporal threshholds and leads to using non-temporal stores in
> cases where `rep movsb` is multiple times faster.
>
> Furthermore, non-temporal stores are written directly to disk so using
> it at a size much smaller than L3 can place soon to be accessed data
> much further away than it otherwise could be. As well, modern machines
> are able to detect streaming patterns (especially if `rep movsb` is
> used) and provide LRU hints to the memory subsystem. This in affect
> caps the total amount of eviction at 1/cache_assosiativity, far below
> meaningfully thrashing the entire cache.
>
> As best I can tell, the benchmarks that lead this small threshold
> where done comparing non-temporal stores versus standard cacheable
> stores. A better comparison (linked below) is to be `rep movsb` which,
> on the measure systems, is nearly 2x faster than non-temporal stores
> at the low-end of the previous threshold, and within 10% for over
> 100MB copies (well past even the current threshold). In cases with a
> low number of threads competing for bandwidth, `rep movsb` is ~2x
> faster up to `sizeof_L3`.
>
> Because there are still valid concerns about performance of large
> memcpy's using cacheable stores (both direct performance and on the
> system), if `rep movsb` is not available this patch also introduces a
> new tunable: `__x86_shared_non_temporal_threshold_no_erms` that will
> continue to use the old calculation and be used if no ERMS memcpy is
> supported by the target.
>
> Benchmarks comparing non-temporal stores, rep movsb, and cacheable
> stores where done using:
> https://github.com/goldsteinn/memcpy-nt-benchmarks
>
> Sheets results (also available in pdf on the github):
> https://docs.google.com/spreadsheets/d/e/2PACX-1vS183r0rW_jRX6tG_E90m9qVuFiMbRIJvi5VAE8yYOvEOIEEc3aSNuEsrFbuXw5c3nGboxMmrupZD7K/pubhtml
> ---
> manual/tunables.texi | 16 +++-
> sysdeps/x86/cacheinfo.h | 8 +-
> sysdeps/x86/dl-cacheinfo.h | 85 +++++++++++++------
> sysdeps/x86/dl-diagnostics-cpu.c | 2 +
> sysdeps/x86/dl-tunables.list | 3 +
> sysdeps/x86/include/cpu-features.h | 4 +-
> .../multiarch/memmove-vec-unaligned-erms.S | 12 ++-
> 7 files changed, 98 insertions(+), 32 deletions(-)
>
> diff --git a/manual/tunables.texi b/manual/tunables.texi
> index 130f94b2bc..8320e724f0 100644
> --- a/manual/tunables.texi
> +++ b/manual/tunables.texi
> @@ -52,6 +52,7 @@ glibc.elision.skip_lock_busy: 3 (min: 0, max: 2147483647)
> glibc.malloc.top_pad: 0x20000 (min: 0x0, max: 0xffffffffffffffff)
> glibc.cpu.x86_rep_stosb_threshold: 0x800 (min: 0x1, max: 0xffffffffffffffff)
> glibc.cpu.x86_non_temporal_threshold: 0xc0000 (min: 0x4040, max: 0xfffffffffffffff)
> +glibc.cpu.x86_non_temporal_threshold_no_erms: 0xc0000 (min: 0x4040, max: 0xfffffffffffffff)
We don't need this. We can use
if (CPU_FEATURE_USABLE_P (cpu_features, ERMS))
to check for ERMS processors.
> glibc.cpu.x86_shstk:
> glibc.pthread.stack_cache_size: 0x2800000 (min: 0x0, max: 0xffffffffffffffff)
> glibc.cpu.hwcap_mask: 0x6 (min: 0x0, max: 0xffffffffffffffff)
> @@ -486,7 +487,8 @@ thread stack originally backup by Huge Pages to default pages.
> @cindex shared_cache_size tunables
> @cindex tunables, shared_cache_size
> @cindex non_temporal_threshold tunables
> -@cindex tunables, non_temporal_threshold
> +@cindex non_temporal_threshold tunables_no_erms
> +@cindex tunables, non_temporal_threshold, non_temporal_threshold_no_erms
>
> @deftp {Tunable namespace} glibc.cpu
> Behavior of @theglibc{} can be tuned to assume specific hardware capabilities
> @@ -559,6 +561,18 @@ like memmove and memcpy.
> This tunable is specific to i386 and x86-64.
> @end deftp
>
> +@deftp Tunable glibc.cpu.x86_non_temporal_threshold_no_erms
> +The @code{glibc.cpu.x86_non_temporal_threshold_no_erms} is similiar to
> +the above, but is used specifically when the ERMS feature is not
> +available. ERMS function are often implemented with optimizations for
> +large streaming workloads. This often makes it a better choice than
> +non-temporal stores for a wider-range of values. When ERMS is not
> +available, however, non-temporal stores become preferable at a much
> +lower threshold.
> +
> +This tunable is specific to i386 and x86-64.
> +@end deftp
> +
> @deftp Tunable glibc.cpu.x86_rep_movsb_threshold
> The @code{glibc.cpu.x86_rep_movsb_threshold} tunable allows the user to
> set threshold in bytes to start using "rep movsb". The value must be
> diff --git a/sysdeps/x86/cacheinfo.h b/sysdeps/x86/cacheinfo.h
> index ec1bc142c4..1083bd6018 100644
> --- a/sysdeps/x86/cacheinfo.h
> +++ b/sysdeps/x86/cacheinfo.h
> @@ -35,9 +35,12 @@ long int __x86_data_cache_size attribute_hidden = 32 * 1024;
> long int __x86_shared_cache_size_half attribute_hidden = 1024 * 1024 / 2;
> long int __x86_shared_cache_size attribute_hidden = 1024 * 1024;
>
> -/* Threshold to use non temporal store. */
> +/* Threshold to use non temporal store if ERMS is available. */
> long int __x86_shared_non_temporal_threshold attribute_hidden;
>
> +/* Threshold to use non temporal store if ERMS is not available. */
> +long int __x86_shared_non_temporal_threshold_no_erms attribute_hidden;
> +
> /* Threshold to use Enhanced REP MOVSB. */
> long int __x86_rep_movsb_threshold attribute_hidden = 2048;
>
> @@ -77,6 +80,9 @@ init_cacheinfo (void)
> __x86_shared_non_temporal_threshold
> = cpu_features->non_temporal_threshold;
>
> + __x86_shared_non_temporal_threshold_no_erms
> + = cpu_features->non_temporal_threshold_no_erms;
> +
> __x86_rep_movsb_threshold = cpu_features->rep_movsb_threshold;
> __x86_rep_stosb_threshold = cpu_features->rep_stosb_threshold;
> __x86_rep_movsb_stop_threshold = cpu_features->rep_movsb_stop_threshold;
> diff --git a/sysdeps/x86/dl-cacheinfo.h b/sysdeps/x86/dl-cacheinfo.h
> index ec88945b39..94d5c6183a 100644
> --- a/sysdeps/x86/dl-cacheinfo.h
> +++ b/sysdeps/x86/dl-cacheinfo.h
> @@ -407,7 +407,7 @@ handle_zhaoxin (int name)
> }
>
> static void
> -get_common_cache_info (long int *shared_ptr, unsigned int *threads_ptr,
> +get_common_cache_info (long int *shared_ptr, long int * shared_per_thread_ptr, unsigned int *threads_ptr,
> long int core)
> {
> unsigned int eax;
> @@ -426,6 +426,7 @@ get_common_cache_info (long int *shared_ptr, unsigned int *threads_ptr,
> unsigned int family = cpu_features->basic.family;
> unsigned int model = cpu_features->basic.model;
> long int shared = *shared_ptr;
> + long int shared_per_thread = *shared_per_thread_ptr;
> unsigned int threads = *threads_ptr;
> bool inclusive_cache = true;
> bool support_count_mask = true;
> @@ -441,6 +442,7 @@ get_common_cache_info (long int *shared_ptr, unsigned int *threads_ptr,
> /* Try L2 otherwise. */
> level = 2;
> shared = core;
> + shared_per_thread = core;
> threads_l2 = 0;
> threads_l3 = -1;
> }
> @@ -597,29 +599,28 @@ get_common_cache_info (long int *shared_ptr, unsigned int *threads_ptr,
> }
> else
> {
> -intel_bug_no_cache_info:
> - /* Assume that all logical threads share the highest cache
> - level. */
> - threads
> - = ((cpu_features->features[CPUID_INDEX_1].cpuid.ebx >> 16)
> - & 0xff);
> - }
> -
> - /* Cap usage of highest cache level to the number of supported
> - threads. */
> - if (shared > 0 && threads > 0)
> - shared /= threads;
> + intel_bug_no_cache_info:
> + /* Assume that all logical threads share the highest cache
> + level. */
> + threads = ((cpu_features->features[CPUID_INDEX_1].cpuid.ebx >> 16)
> + & 0xff);
> +
> + /* Get per-thread size of highest level cache. */
> + if (shared_per_thread > 0 && threads > 0)
> + shared_per_thread /= threads;
> + }
> }
>
> /* Account for non-inclusive L2 and L3 caches. */
> if (!inclusive_cache)
> {
> if (threads_l2 > 0)
> - core /= threads_l2;
> + shared_per_thread += core / threads_l2;
> shared += core;
> }
>
> *shared_ptr = shared;
> + *shared_per_thread_ptr = shared_per_thread;
> *threads_ptr = threads;
> }
>
> @@ -629,6 +630,7 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> /* Find out what brand of processor. */
> long int data = -1;
> long int shared = -1;
> + long int shared_per_thread = -1;
> long int core = -1;
> unsigned int threads = 0;
> unsigned long int level1_icache_size = -1;
> @@ -649,6 +651,7 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> data = handle_intel (_SC_LEVEL1_DCACHE_SIZE, cpu_features);
> core = handle_intel (_SC_LEVEL2_CACHE_SIZE, cpu_features);
> shared = handle_intel (_SC_LEVEL3_CACHE_SIZE, cpu_features);
> + shared_per_thread = shared;
>
> level1_icache_size
> = handle_intel (_SC_LEVEL1_ICACHE_SIZE, cpu_features);
> @@ -672,13 +675,14 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> level4_cache_size
> = handle_intel (_SC_LEVEL4_CACHE_SIZE, cpu_features);
>
> - get_common_cache_info (&shared, &threads, core);
> + get_common_cache_info (&shared, &shared_per_thread, &threads, core);
> }
> else if (cpu_features->basic.kind == arch_kind_zhaoxin)
> {
> data = handle_zhaoxin (_SC_LEVEL1_DCACHE_SIZE);
> core = handle_zhaoxin (_SC_LEVEL2_CACHE_SIZE);
> shared = handle_zhaoxin (_SC_LEVEL3_CACHE_SIZE);
> + shared_per_thread = shared;
>
> level1_icache_size = handle_zhaoxin (_SC_LEVEL1_ICACHE_SIZE);
> level1_icache_linesize = handle_zhaoxin (_SC_LEVEL1_ICACHE_LINESIZE);
> @@ -692,13 +696,14 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> level3_cache_assoc = handle_zhaoxin (_SC_LEVEL3_CACHE_ASSOC);
> level3_cache_linesize = handle_zhaoxin (_SC_LEVEL3_CACHE_LINESIZE);
>
> - get_common_cache_info (&shared, &threads, core);
> + get_common_cache_info (&shared, &shared_per_thread, &threads, core);
> }
> else if (cpu_features->basic.kind == arch_kind_amd)
> {
> data = handle_amd (_SC_LEVEL1_DCACHE_SIZE);
> core = handle_amd (_SC_LEVEL2_CACHE_SIZE);
> shared = handle_amd (_SC_LEVEL3_CACHE_SIZE);
> + shared_per_thread = shared;
>
> level1_icache_size = handle_amd (_SC_LEVEL1_ICACHE_SIZE);
> level1_icache_linesize = handle_amd (_SC_LEVEL1_ICACHE_LINESIZE);
> @@ -715,6 +720,9 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> if (shared <= 0)
> /* No shared L3 cache. All we have is the L2 cache. */
> shared = core;
> +
> + if (shared_per_thread <= 0)
> + shared_per_thread = shared;
> }
>
> cpu_features->level1_icache_size = level1_icache_size;
> @@ -730,17 +738,24 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> cpu_features->level3_cache_linesize = level3_cache_linesize;
> cpu_features->level4_cache_size = level4_cache_size;
>
> - /* The default setting for the non_temporal threshold is 3/4 of one
> - thread's share of the chip's cache. For most Intel and AMD processors
> - with an initial release date between 2017 and 2020, a thread's typical
> - share of the cache is from 500 KBytes to 2 MBytes. Using the 3/4
> - threshold leaves 125 KBytes to 500 KBytes of the thread's data
> - in cache after a maximum temporal copy, which will maintain
> - in cache a reasonable portion of the thread's stack and other
> - active data. If the threshold is set higher than one thread's
> - share of the cache, it has a substantial risk of negatively
> - impacting the performance of other threads running on the chip. */
> - unsigned long int non_temporal_threshold = shared * 3 / 4;
> + /* The default setting for the non_temporal threshold is 1/2 of size
> + of chip's cache. For most Intel and AMD processors with an
> + initial release date between 2017 and 2023, a thread's typical
> + share of the cache is from 18-64MB. Using the 1/2 L3 is meant to
> + estimate the point where non-temporal stores begin outcompeting
> + other methods. As well the point where the fact that non-temporal
> + stores are forced back to disk would already occured to the
> + majority of the lines in the copy. Note, concerns about the
> + entire L3 cache being evicted by the copy are mostly alleviated
> + by the fact that modern HW detects streaming patterns and
> + provides proper LRU hints so that the the maximum thrashing
> + capped at 1/assosiativity. */
> + unsigned long int non_temporal_threshold = shared / 2;
> + /* If no ERMS, we use the per-thread L3 chunking. Normal cacheable stores run
> + a higher risk of actually thrashing the cache as they don't have a HW LRU
> + hint. As well, there performance in highly parallel situations is
> + noticeably worse. */
> + unsigned long int non_temporal_threshold_no_erms = shared_per_thread * 3 / 4;
> /* SIZE_MAX >> 4 because memmove-vec-unaligned-erms right-shifts the value of
> 'x86_non_temporal_threshold' by `LOG_4X_MEMCPY_THRESH` (4) and it is best
> if that operation cannot overflow. Minimum of 0x4040 (16448) because the
> @@ -754,6 +769,11 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> else if (non_temporal_threshold > maximum_non_temporal_threshold)
> non_temporal_threshold = maximum_non_temporal_threshold;
>
> + if (non_temporal_threshold_no_erms < minimum_non_temporal_threshold)
> + non_temporal_threshold_no_erms = minimum_non_temporal_threshold;
> + else if (non_temporal_threshold_no_erms > maximum_non_temporal_threshold)
> + non_temporal_threshold_no_erms = maximum_non_temporal_threshold;
> +
> /* NB: The REP MOVSB threshold must be greater than VEC_SIZE * 8. */
> unsigned int minimum_rep_movsb_threshold;
> /* NB: The default REP MOVSB threshold is 4096 * (VEC_SIZE / 16) for
> @@ -802,6 +822,12 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> && tunable_size <= maximum_non_temporal_threshold)
> non_temporal_threshold = tunable_size;
>
> + tunable_size
> + = TUNABLE_GET (x86_non_temporal_threshold_no_erms, long int, NULL);
> + if (tunable_size > minimum_non_temporal_threshold
> + && tunable_size <= maximum_non_temporal_threshold)
> + non_temporal_threshold_no_erms = tunable_size;
> +
> tunable_size = TUNABLE_GET (x86_rep_movsb_threshold, long int, NULL);
> if (tunable_size > minimum_rep_movsb_threshold)
> rep_movsb_threshold = tunable_size;
> @@ -817,6 +843,9 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> TUNABLE_SET_WITH_BOUNDS (x86_non_temporal_threshold, non_temporal_threshold,
> minimum_non_temporal_threshold,
> maximum_non_temporal_threshold);
> + TUNABLE_SET_WITH_BOUNDS (
> + x86_non_temporal_threshold_no_erms, non_temporal_threshold_no_erms,
> + minimum_non_temporal_threshold, maximum_non_temporal_threshold);
> TUNABLE_SET_WITH_BOUNDS (x86_rep_movsb_threshold, rep_movsb_threshold,
> minimum_rep_movsb_threshold, SIZE_MAX);
> TUNABLE_SET_WITH_BOUNDS (x86_rep_stosb_threshold, rep_stosb_threshold, 1,
> @@ -837,6 +866,8 @@ dl_init_cacheinfo (struct cpu_features *cpu_features)
> cpu_features->data_cache_size = data;
> cpu_features->shared_cache_size = shared;
> cpu_features->non_temporal_threshold = non_temporal_threshold;
> + cpu_features->non_temporal_threshold_no_erms
> + = non_temporal_threshold_no_erms;
> cpu_features->rep_movsb_threshold = rep_movsb_threshold;
> cpu_features->rep_stosb_threshold = rep_stosb_threshold;
> cpu_features->rep_movsb_stop_threshold = rep_movsb_stop_threshold;
> diff --git a/sysdeps/x86/dl-diagnostics-cpu.c b/sysdeps/x86/dl-diagnostics-cpu.c
> index a1578e4665..5c09472a10 100644
> --- a/sysdeps/x86/dl-diagnostics-cpu.c
> +++ b/sysdeps/x86/dl-diagnostics-cpu.c
> @@ -83,6 +83,8 @@ _dl_diagnostics_cpu (void)
> cpu_features->shared_cache_size);
> print_cpu_features_value ("non_temporal_threshold",
> cpu_features->non_temporal_threshold);
> + print_cpu_features_value ("non_temporal_threshold_no_erms",
> + cpu_features->non_temporal_threshold_no_erms);
> print_cpu_features_value ("rep_movsb_threshold",
> cpu_features->rep_movsb_threshold);
> print_cpu_features_value ("rep_movsb_stop_threshold",
> diff --git a/sysdeps/x86/dl-tunables.list b/sysdeps/x86/dl-tunables.list
> index feb7004036..aac6341716 100644
> --- a/sysdeps/x86/dl-tunables.list
> +++ b/sysdeps/x86/dl-tunables.list
> @@ -30,6 +30,9 @@ glibc {
> x86_non_temporal_threshold {
> type: SIZE_T
> }
> + x86_non_temporal_threshold_no_erms {
> + type: SIZE_T
> + }
> x86_rep_movsb_threshold {
> type: SIZE_T
> # Since there is overhead to set up REP MOVSB operation, REP
> diff --git a/sysdeps/x86/include/cpu-features.h b/sysdeps/x86/include/cpu-features.h
> index 40b8129d6a..df6c561eac 100644
> --- a/sysdeps/x86/include/cpu-features.h
> +++ b/sysdeps/x86/include/cpu-features.h
> @@ -913,8 +913,10 @@ struct cpu_features
> /* Shared cache size for use in memory and string routines, typically
> L2 or L3 size. */
> unsigned long int shared_cache_size;
> - /* Threshold to use non temporal store. */
> + /* Threshold to use non temporal store if ERMS is available. */
> unsigned long int non_temporal_threshold;
> + /* Threshold to use non temporal store if ERMS is not available. */
> + unsigned long int non_temporal_threshold_no_erms;
> /* Threshold to use "rep movsb". */
> unsigned long int rep_movsb_threshold;
> /* Threshold to stop using "rep movsb". */
> diff --git a/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S b/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
> index d1b92785b0..856c3daf3b 100644
> --- a/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
> +++ b/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
> @@ -424,8 +424,16 @@ L(more_8x_vec):
> jb L(more_8x_vec_backward_check_nop)
> /* Check if non-temporal move candidate. */
> #if (defined USE_MULTIARCH || VEC_SIZE == 16) && IS_IN (libc)
> - /* Check non-temporal store threshold. */
> - cmp __x86_shared_non_temporal_threshold(%rip), %RDX_LP
> + /* Check non-temporal store threshold if ERMS is not available.
> + NB: This path is only hit if we jumped here from L(more_2x_vec).
> + If we went to L(movsb), then we enter at either the forward loop
> + directly or go to the backward loop.
> +
> + WARNING: `__x86_shared_non_temporal_threshold_no_erms` should
> + NEVER be used in a control flow that could come from
> + L(movsb_more_2x_vec) without checking checkout
> + `__x86_rep_movsb_threshold` first. */
> + cmp __x86_shared_non_temporal_threshold_no_erms(%rip), %RDX_LP
> ja L(large_memcpy_2x)
> #endif
> /* To reach this point there cannot be overlap and dst > src. So
> --
> 2.34.1
>
--
H.J.
next prev parent reply other threads:[~2023-04-24 22:49 UTC|newest]
Thread overview: 76+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-04-24 5:03 [PATCH v1] " Noah Goldstein
2023-04-24 18:09 ` H.J. Lu
2023-04-24 18:34 ` Noah Goldstein
2023-04-24 20:44 ` H.J. Lu
2023-04-24 22:30 ` Noah Goldstein
2023-04-24 22:30 ` [PATCH v2] " Noah Goldstein
2023-04-24 22:48 ` H.J. Lu [this message]
2023-04-25 2:05 ` Noah Goldstein
2023-04-25 2:55 ` H.J. Lu
2023-04-25 3:43 ` Noah Goldstein
2023-04-25 3:43 ` [PATCH v3] " Noah Goldstein
2023-04-25 17:42 ` H.J. Lu
2023-04-25 21:45 ` Noah Goldstein
2023-04-25 21:45 ` [PATCH v4] " Noah Goldstein
2023-04-26 15:59 ` H.J. Lu
2023-04-26 17:15 ` Noah Goldstein
2023-05-04 3:28 ` Noah Goldstein
2023-05-05 18:06 ` H.J. Lu
2023-05-09 3:14 ` Noah Goldstein
2023-05-09 3:13 ` [PATCH v5 1/3] x86: Increase `non_temporal_threshold` to roughly `sizeof_L3 / 4` Noah Goldstein
2023-05-09 3:13 ` [PATCH v5 2/3] x86: Refactor Intel `init_cpu_features` Noah Goldstein
2023-05-09 21:58 ` H.J. Lu
2023-05-10 0:33 ` Noah Goldstein
2023-05-09 3:13 ` [PATCH v5 3/3] x86: Make the divisor in setting `non_temporal_threshold` cpu specific Noah Goldstein
2023-05-10 0:33 ` [PATCH v6 1/4] x86: Increase `non_temporal_threshold` to roughly `sizeof_L3 / 4` Noah Goldstein
2023-05-10 0:33 ` [PATCH v6 2/4] x86: Refactor Intel `init_cpu_features` Noah Goldstein
2023-05-10 22:13 ` H.J. Lu
2023-05-10 23:17 ` Noah Goldstein
2023-05-11 21:36 ` H.J. Lu
2023-05-12 5:11 ` Noah Goldstein
2023-05-10 0:33 ` [PATCH v6 3/4] x86: Make the divisor in setting `non_temporal_threshold` cpu specific Noah Goldstein
2023-05-10 0:33 ` [PATCH v6 4/4] x86: Tune 'Saltwell' microarch the same was a 'Bonnell' Noah Goldstein
2023-05-10 22:04 ` H.J. Lu
2023-05-10 22:12 ` Noah Goldstein
2023-05-10 15:55 ` [PATCH v6 1/4] x86: Increase `non_temporal_threshold` to roughly `sizeof_L3 / 4` H.J. Lu
2023-05-10 16:07 ` Noah Goldstein
2023-05-10 22:12 ` [PATCH v7 2/4] x86: Refactor Intel `init_cpu_features` Noah Goldstein
2023-05-10 22:12 ` [PATCH v7 3/4] x86: Make the divisor in setting `non_temporal_threshold` cpu specific Noah Goldstein
2023-05-10 22:12 ` [PATCH v7 4/4] x86: Tune 'Saltwell' microarch the same was a 'Bonnell' Noah Goldstein
2023-05-12 5:12 ` Noah Goldstein
2023-05-12 5:10 ` [PATCH v8 1/3] x86: Increase `non_temporal_threshold` to roughly `sizeof_L3 / 4` Noah Goldstein
2023-05-12 5:10 ` [PATCH v8 2/3] x86: Refactor Intel `init_cpu_features` Noah Goldstein
2023-05-12 22:17 ` H.J. Lu
2023-05-13 5:18 ` Noah Goldstein
2023-05-12 22:03 ` [PATCH v8 3/3] x86: Make the divisor in setting `non_temporal_threshold` cpu specific Noah Goldstein
2023-05-13 5:19 ` [PATCH v9 1/3] x86: Increase `non_temporal_threshold` to roughly `sizeof_L3 / 4` Noah Goldstein
2023-05-13 5:19 ` [PATCH v9 2/3] x86: Refactor Intel `init_cpu_features` Noah Goldstein
2023-05-15 20:57 ` H.J. Lu
2023-05-26 3:34 ` DJ Delorie
2023-05-27 18:46 ` Noah Goldstein
2023-05-13 5:19 ` [PATCH v9 3/3] x86: Make the divisor in setting `non_temporal_threshold` cpu specific Noah Goldstein
2023-05-26 3:34 ` DJ Delorie
2023-05-27 18:46 ` Noah Goldstein
2023-05-15 18:29 ` [PATCH v9 1/3] x86: Increase `non_temporal_threshold` to roughly `sizeof_L3 / 4` Noah Goldstein
2023-05-17 12:00 ` Carlos O'Donell
2023-05-26 3:34 ` DJ Delorie
2023-05-27 18:46 ` [PATCH v10 " Noah Goldstein
2023-05-27 18:46 ` [PATCH v10 2/3] x86: Refactor Intel `init_cpu_features` Noah Goldstein
2023-05-27 18:46 ` [PATCH v10 3/3] x86: Make the divisor in setting `non_temporal_threshold` cpu specific Noah Goldstein
2023-05-31 2:33 ` DJ Delorie
2023-07-10 5:23 ` Sajan Karumanchi
2023-07-10 15:58 ` Noah Goldstein
2023-07-14 2:21 ` Re: Noah Goldstein
2023-07-14 7:39 ` Re: sajan karumanchi
2023-06-07 0:15 ` [PATCH v10 1/3] x86: Increase `non_temporal_threshold` to roughly `sizeof_L3 / 4` Carlos O'Donell
2023-06-07 18:18 ` Noah Goldstein
2023-06-07 18:18 ` [PATCH v11 " Noah Goldstein
2023-06-07 18:18 ` [PATCH v11 2/3] x86: Refactor Intel `init_cpu_features` Noah Goldstein
2023-06-07 18:18 ` [PATCH v11 3/3] x86: Make the divisor in setting `non_temporal_threshold` cpu specific Noah Goldstein
2023-06-07 18:19 ` [PATCH v11 1/3] x86: Increase `non_temporal_threshold` to roughly `sizeof_L3 / 4` Noah Goldstein
2023-08-14 23:00 ` Noah Goldstein
2023-08-22 15:11 ` Noah Goldstein
2023-08-24 17:06 ` Noah Goldstein
2023-08-28 20:02 ` Noah Goldstein
2023-09-05 15:37 ` Noah Goldstein
2023-09-12 3:50 ` Noah Goldstein
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=CAMe9rOokMDgEZwvMp1F8P3CO7E-O0YTPM+jA-PhKQK93Stk8HQ@mail.gmail.com \
--to=hjl.tools@gmail.com \
--cc=carlos@systemhalted.org \
--cc=goldstein.w.n@gmail.com \
--cc=libc-alpha@sourceware.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).