public inbox for glibc-cvs@sourceware.org
help / color / mirror / Atom feed
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2023-02-09 19:45 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2023-02-09 19:45 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=f0c20d36de584ea846a08cf6b658885f1801dfc2
commit f0c20d36de584ea846a08cf6b658885f1801dfc2
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index 6d71047154..4c0c74091f 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-10-28 17:37 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-10-28 17:37 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=d6e010db175cec9664539308c0d3ffe597feca25
commit d6e010db175cec9664539308c0d3ffe597feca25
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-10-04 12:55 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-10-04 12:55 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=c92ee6e05a05234e94a8eb2e35dd04f56d6db350
commit c92ee6e05a05234e94a8eb2e35dd04f56d6db350
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-06-09 21:15 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-06-09 21:15 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=2cb7de4737f15cd52f80269a67bcaee580324d04
commit 2cb7de4737f15cd52f80269a67bcaee580324d04
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-06-09 13:12 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-06-09 13:12 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=2cb7de4737f15cd52f80269a67bcaee580324d04
commit 2cb7de4737f15cd52f80269a67bcaee580324d04
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-06-03 14:01 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-06-03 14:01 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=18672aea5721e1b7e5fa28be826a4a0aa6098f79
commit 18672aea5721e1b7e5fa28be826a4a0aa6098f79
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-05-13 14:15 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-05-13 14:15 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=01b0212b8d38f0ef87ac39038dc6f57b376c9fc6
commit 01b0212b8d38f0ef87ac39038dc6f57b376c9fc6
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-05-12 19:29 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-05-12 19:29 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=8ac4a0715e41752a2da5c3a295af3de43a69a803
commit 8ac4a0715e41752a2da5c3a295af3de43a69a803
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-05-10 18:19 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-05-10 18:19 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=bb2be0769550e1c5a1dbf70ab0795c47b75e87df
commit bb2be0769550e1c5a1dbf70ab0795c47b75e87df
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-04-29 13:59 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-04-29 13:59 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=9cfa2113b48d3808c53a6ddbe32fc3e08dad908e
commit 9cfa2113b48d3808c53a6ddbe32fc3e08dad908e
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-04-04 12:49 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-04-04 12:49 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=67ddd9f46ffeb3e2ac5cacd14f1afce72bfea13c
commit 67ddd9f46ffeb3e2ac5cacd14f1afce72bfea13c
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-03-31 19:02 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-03-31 19:02 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=d104cb2cc3fa2d6403da92e14241df9079302595
commit d104cb2cc3fa2d6403da92e14241df9079302595
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-03-29 20:25 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-03-29 20:25 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=11f58bb3edb0a3093a400d9aa9599b99e5346657
commit 11f58bb3edb0a3093a400d9aa9599b99e5346657
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-03-16 18:01 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-03-16 18:01 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=0de6f580c244aa0bcfa8dc726d0a0fa5e40efe3a
commit 0de6f580c244aa0bcfa8dc726d0a0fa5e40efe3a
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-03-15 18:39 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-03-15 18:39 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=0de6f580c244aa0bcfa8dc726d0a0fa5e40efe3a
commit 0de6f580c244aa0bcfa8dc726d0a0fa5e40efe3a
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
* [glibc/azanella/clang] malloc: Use C11 atomics on memusage
@ 2022-03-11 17:23 Adhemerval Zanella
0 siblings, 0 replies; 16+ messages in thread
From: Adhemerval Zanella @ 2022-03-11 17:23 UTC (permalink / raw)
To: glibc-cvs
https://sourceware.org/git/gitweb.cgi?p=glibc.git;h=9b4bf3d04fc3d53818956e520f6500af095ca3f2
commit 9b4bf3d04fc3d53818956e520f6500af095ca3f2
Author: Adhemerval Zanella <adhemerval.zanella@linaro.org>
Date: Fri Mar 11 13:53:11 2022 -0300
malloc: Use C11 atomics on memusage
Checked on x86_64-linux-gnu.
Diff:
---
malloc/memusage.c | 132 ++++++++++++++++++++++++++++++------------------------
1 file changed, 73 insertions(+), 59 deletions(-)
diff --git a/malloc/memusage.c b/malloc/memusage.c
index f30906dffb..ddc487422c 100644
--- a/malloc/memusage.c
+++ b/malloc/memusage.c
@@ -134,6 +134,19 @@ gettime (struct entry *e)
#endif
}
+static inline void
+peak_atomic_max (size_t *peak, size_t val)
+{
+ size_t v;
+ do
+ {
+ v = atomic_load_relaxed (peak);
+ if (v >= val)
+ break;
+ }
+ while (! atomic_compare_exchange_weak_acquire (peak, &v, val));
+}
+
/* Update the global data after a successful function call. */
static void
update_data (struct header *result, size_t len, size_t old_len)
@@ -148,8 +161,8 @@ update_data (struct header *result, size_t len, size_t old_len)
/* Compute current heap usage and compare it with the maximum value. */
size_t heap
- = catomic_exchange_and_add (¤t_heap, len - old_len) + len - old_len;
- catomic_max (&peak_heap, heap);
+ = atomic_fetch_add_acquire (¤t_heap, len - old_len) + len - old_len;
+ peak_atomic_max (&peak_heap, heap);
/* Compute current stack usage and compare it with the maximum
value. The base stack pointer might not be set if this is not
@@ -172,15 +185,15 @@ update_data (struct header *result, size_t len, size_t old_len)
start_sp = sp;
size_t current_stack = start_sp - sp;
#endif
- catomic_max (&peak_stack, current_stack);
+ peak_atomic_max (&peak_stack, current_stack);
/* Add up heap and stack usage and compare it with the maximum value. */
- catomic_max (&peak_total, heap + current_stack);
+ peak_atomic_max (&peak_total, heap + current_stack);
/* Store the value only if we are writing to a file. */
if (fd != -1)
{
- uint32_t idx = catomic_exchange_and_add (&buffer_cnt, 1);
+ uint32_t idx = atomic_fetch_add_acquire (&buffer_cnt, 1);
if (idx + 1 >= 2 * buffer_size)
{
/* We try to reset the counter to the correct range. If
@@ -188,7 +201,8 @@ update_data (struct header *result, size_t len, size_t old_len)
counter it does not matter since that thread will take
care of the correction. */
uint32_t reset = (idx + 1) % (2 * buffer_size);
- catomic_compare_and_exchange_val_acq (&buffer_cnt, reset, idx + 1);
+ uint32_t expected = idx + 1;
+ atomic_compare_exchange_weak_acquire (&buffer_cnt, &expected, reset);
if (idx >= 2 * buffer_size)
idx = reset - 1;
}
@@ -362,24 +376,24 @@ malloc (size_t len)
return (*mallocp)(len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_malloc]);
+ atomic_fetch_add_acquire (&calls[idx_malloc], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_malloc], len);
+ atomic_fetch_add_acquire (&total[idx_malloc], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*mallocp)(len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_malloc]);
+ atomic_fetch_add_acquire (&failed[idx_malloc], 1);
return NULL;
}
@@ -430,21 +444,21 @@ realloc (void *old, size_t len)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_realloc]);
+ atomic_fetch_add_acquire (&calls[idx_realloc], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `realloc'. */
- catomic_add (&total[idx_realloc], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_realloc], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
if (len == 0 && old != NULL)
{
/* Special case. */
- catomic_increment (&realloc_free);
+ atomic_fetch_add_acquire (&realloc_free, 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, old_len);
@@ -457,26 +471,26 @@ realloc (void *old, size_t len)
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Do the real work. */
result = (struct header *) (*reallocp)(real, len + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_realloc]);
+ atomic_fetch_add_acquire (&failed[idx_realloc], 1);
return NULL;
}
/* Record whether the reduction/increase happened in place. */
if (real == result)
- catomic_increment (&inplace);
+ atomic_fetch_add_acquire (&inplace, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing);
+ atomic_fetch_add_acquire (&decreasing, 1);
/* Update the allocation data and write out the records if necessary. */
update_data (result, len, old_len);
@@ -508,16 +522,16 @@ calloc (size_t n, size_t len)
return (*callocp)(n, len);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_calloc]);
+ atomic_fetch_add_acquire (&calls[idx_calloc], 1);
/* Keep track of total memory consumption for `calloc'. */
- catomic_add (&total[idx_calloc], size);
+ atomic_fetch_add_acquire (&total[idx_calloc], size);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, size);
+ atomic_fetch_add_acquire (&grand_total, size);
/* Remember the size of the request. */
if (size < 65536)
- catomic_increment (&histogram[size / 16]);
+ atomic_fetch_add_acquire (&histogram[size / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
++calls_total;
@@ -525,7 +539,7 @@ calloc (size_t n, size_t len)
result = (struct header *) (*mallocp)(size + sizeof (struct header));
if (result == NULL)
{
- catomic_increment (&failed[idx_calloc]);
+ atomic_fetch_add_acquire (&failed[idx_calloc], 1);
return NULL;
}
@@ -563,7 +577,7 @@ free (void *ptr)
/* `free (NULL)' has no effect. */
if (ptr == NULL)
{
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
return;
}
@@ -577,9 +591,9 @@ free (void *ptr)
}
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_free]);
+ atomic_fetch_add_acquire (&calls[idx_free], 1);
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_free], real->length);
+ atomic_fetch_add_acquire (&total[idx_free], real->length);
/* Update the allocation data and write out the records if necessary. */
update_data (NULL, 0, real->length);
@@ -614,22 +628,22 @@ mmap (void *start, size_t len, int prot, int flags, int fd, off_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -667,22 +681,22 @@ mmap64 (void *start, size_t len, int prot, int flags, int fd, off64_t offset)
? idx_mmap_a : prot & PROT_WRITE ? idx_mmap_w : idx_mmap_r);
/* Keep track of number of calls. */
- catomic_increment (&calls[idx]);
+ atomic_fetch_add_acquire (&calls[idx], 1);
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx], len);
+ atomic_fetch_add_acquire (&total[idx], len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len);
+ atomic_fetch_add_acquire (&grand_total, len);
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx]);
+ atomic_fetch_add_acquire (&failed[idx], 1);
else if (idx == idx_mmap_w)
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -722,33 +736,33 @@ mremap (void *start, size_t old_len, size_t len, int flags, ...)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_mremap]);
+ atomic_fetch_add_acquire (&calls[idx_mremap], 1);
if (len > old_len)
{
/* Keep track of total memory consumption for `malloc'. */
- catomic_add (&total[idx_mremap], len - old_len);
+ atomic_fetch_add_acquire (&total[idx_mremap], len - old_len);
/* Keep track of total memory requirement. */
- catomic_add (&grand_total, len - old_len);
+ atomic_fetch_add_acquire (&grand_total, len - old_len);
}
/* Remember the size of the request. */
if (len < 65536)
- catomic_increment (&histogram[len / 16]);
+ atomic_fetch_add_acquire (&histogram[len / 16], 1);
else
- catomic_increment (&large);
+ atomic_fetch_add_acquire (&large, 1);
/* Total number of calls of any of the functions. */
- catomic_increment (&calls_total);
+ atomic_fetch_add_acquire (&calls_total, 1);
/* Check for failures. */
if (result == NULL)
- catomic_increment (&failed[idx_mremap]);
+ atomic_fetch_add_acquire (&failed[idx_mremap], 1);
else
{
/* Record whether the reduction/increase happened in place. */
if (start == result)
- catomic_increment (&inplace_mremap);
+ atomic_fetch_add_acquire (&inplace_mremap, 1);
/* Was the buffer increased? */
if (old_len > len)
- catomic_increment (&decreasing_mremap);
+ atomic_fetch_add_acquire (&decreasing_mremap, 1);
/* Update the allocation data and write out the records if
necessary. Note the first parameter is NULL which means
@@ -783,19 +797,19 @@ munmap (void *start, size_t len)
if (!not_me && trace_mmap)
{
/* Keep track of number of calls. */
- catomic_increment (&calls[idx_munmap]);
+ atomic_fetch_add_acquire (&calls[idx_munmap], 1);
if (__glibc_likely (result == 0))
{
/* Keep track of total memory freed using `free'. */
- catomic_add (&total[idx_munmap], len);
+ atomic_fetch_add_acquire (&total[idx_munmap], len);
/* Update the allocation data and write out the records if
necessary. */
update_data (NULL, 0, len);
}
else
- catomic_increment (&failed[idx_munmap]);
+ atomic_fetch_add_acquire (&failed[idx_munmap], 1);
}
return result;
^ permalink raw reply [flat|nested] 16+ messages in thread
end of thread, other threads:[~2023-02-09 19:45 UTC | newest]
Thread overview: 16+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2023-02-09 19:45 [glibc/azanella/clang] malloc: Use C11 atomics on memusage Adhemerval Zanella
-- strict thread matches above, loose matches on Subject: below --
2022-10-28 17:37 Adhemerval Zanella
2022-10-04 12:55 Adhemerval Zanella
2022-06-09 21:15 Adhemerval Zanella
2022-06-09 13:12 Adhemerval Zanella
2022-06-03 14:01 Adhemerval Zanella
2022-05-13 14:15 Adhemerval Zanella
2022-05-12 19:29 Adhemerval Zanella
2022-05-10 18:19 Adhemerval Zanella
2022-04-29 13:59 Adhemerval Zanella
2022-04-04 12:49 Adhemerval Zanella
2022-03-31 19:02 Adhemerval Zanella
2022-03-29 20:25 Adhemerval Zanella
2022-03-16 18:01 Adhemerval Zanella
2022-03-15 18:39 Adhemerval Zanella
2022-03-11 17:23 Adhemerval Zanella
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).