* [RFC, PR 80689] Copy small aggregates element-wise
@ 2017-10-13 16:31 Martin Jambor
2017-10-17 11:44 ` Richard Biener
0 siblings, 1 reply; 26+ messages in thread
From: Martin Jambor @ 2017-10-13 16:31 UTC (permalink / raw)
To: GCC Patches; +Cc: Jan Hubicka
Hi,
I'd like to request comments to the patch below which aims to fix PR
80689, which is an instance of a store-to-load forwarding stall on
x86_64 CPUs in the Image Magick benchmark, which is responsible for a
slow down of up to 9% compared to gcc 6, depending on options and HW
used. (Actually, I have just seen 24% in one specific combination but
for various reasons can no longer verify it today.)
The revision causing the regression is 237074, which increased the
size of the mode for copying aggregates "by pieces" to 128 bits,
incurring big stalls when the values being copied are also still being
stored in a smaller data type or if the copied values are loaded in a
smaller types shortly afterwards. Such situations happen in Image
Magick even across calls, which means that any non-IPA flow-sensitive
approach would not detect them. Therefore, the patch simply changes
the way we copy small BLKmode data that are simple combinations of
records and arrays (meaning no unions, bit-fields but also character
arrays are disallowed) and simply copies them one field and/or element
at a time.
"Small" in this RFC patch means up to 35 bytes on x86_64 and i386 CPUs
(the structure in the benchmark has 32 bytes) but is subject to change
after more benchmarking and is actually zero - meaning element copying
never happens - on other architectures. I believe that any
architecture with a store buffer can benefit but it's probably better
to leave it to their maintainers to find a different default value. I
am not sure this is how such HW-dependant decisions should be done and
is the primary reason, why I am sending this RFC first.
I have decided to implement this change at the expansion level because
at that point the type information is still readily available and at
the same time we can also handle various implicit copies, for example
those passing a parameter. I found I could re-use some bits and
pieces of tree-SRA and so I did, creating tree-sra.h header file in
the process.
I am fully aware that in the final patch the new parameter, or indeed
any new parameters, need to be documented. I have skipped that
intentionally now and will write the documentation if feedback here is
generally good.
I have bootstrapped and tested this patch on x86_64-linux, with
different values of the parameter and only found problems with
unreasonably high values leading to OOM. I have done the same with a
previous version of the patch which was equivalent to the limit being
64 bytes on aarch64-linux, ppc64le-linux and ia64-linux and only ran
into failures of tests which assumed that structure padding was copied
in aggregate copies (mostly gcc.target/aarch64/aapcs64/ stuff but also
for example gcc.dg/vmx/varargs-4.c).
The patch decreases the SPEC 2017 "rate" run-time of imagick by 9% and
8% at -O2 and -Ofast compilation levels respectively on one particular
new AMD CPU and by 6% and 3% on one particular old Intel machine.
Thanks in advance for any comments,
Martin
2017-10-12 Martin Jambor <mjambor@suse.cz>
PR target/80689
* tree-sra.h: New file.
* ipa-prop.h: Moved declaration of build_ref_for_offset to
tree-sra.h.
* expr.c: Include params.h and tree-sra.h.
(emit_move_elementwise): New function.
(store_expr_with_bounds): Optionally use it.
* ipa-cp.c: Include tree-sra.h.
* params.def (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY): New.
* config/i386/i386.c (ix86_option_override_internal): Set
PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY to 35.
* tree-sra.c: Include tree-sra.h.
(scalarizable_type_p): Renamed to
simple_mix_of_records_and_arrays_p, made public, renamed the
second parameter to allow_char_arrays.
(extract_min_max_idx_from_array): New function.
(completely_scalarize): Moved bits of the function to
extract_min_max_idx_from_array.
testsuite/
* gcc.target/i386/pr80689-1.c: New test.
---
gcc/config/i386/i386.c | 4 ++
gcc/expr.c | 103 ++++++++++++++++++++++++++++--
gcc/ipa-cp.c | 1 +
gcc/ipa-prop.h | 4 --
gcc/params.def | 6 ++
gcc/testsuite/gcc.target/i386/pr80689-1.c | 38 +++++++++++
gcc/tree-sra.c | 86 +++++++++++++++----------
gcc/tree-sra.h | 33 ++++++++++
8 files changed, 233 insertions(+), 42 deletions(-)
create mode 100644 gcc/testsuite/gcc.target/i386/pr80689-1.c
create mode 100644 gcc/tree-sra.h
diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c
index 1ee8351c21f..87f602e7ead 100644
--- a/gcc/config/i386/i386.c
+++ b/gcc/config/i386/i386.c
@@ -6511,6 +6511,10 @@ ix86_option_override_internal (bool main_args_p,
ix86_tune_cost->l2_cache_size,
opts->x_param_values,
opts_set->x_param_values);
+ maybe_set_param_value (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
+ 35,
+ opts->x_param_values,
+ opts_set->x_param_values);
/* Enable sw prefetching at -O3 for CPUS that prefetching is helpful. */
if (opts->x_flag_prefetch_loop_arrays < 0
diff --git a/gcc/expr.c b/gcc/expr.c
index 134ee731c29..dff24e7f166 100644
--- a/gcc/expr.c
+++ b/gcc/expr.c
@@ -61,7 +61,8 @@ along with GCC; see the file COPYING3. If not see
#include "tree-chkp.h"
#include "rtl-chkp.h"
#include "ccmp.h"
-
+#include "params.h"
+#include "tree-sra.h"
/* If this is nonzero, we do not bother generating VOLATILE
around volatile memory references, and we are willing to
@@ -5340,6 +5341,80 @@ emit_storent_insn (rtx to, rtx from)
return maybe_expand_insn (code, 2, ops);
}
+/* Generate code for copying data of type TYPE at SOURCE plus OFFSET to TARGET
+ plus OFFSET, but do so element-wise and/or field-wise for each record and
+ array within TYPE. TYPE must either be a register type or an aggregate
+ complying with scalarizable_type_p.
+
+ If CALL_PARAM_P is nonzero, this is a store into a call param on the
+ stack, and block moves may need to be treated specially. */
+
+static void
+emit_move_elementwise (tree type, rtx target, rtx source, HOST_WIDE_INT offset,
+ int call_param_p)
+{
+ switch (TREE_CODE (type))
+ {
+ case RECORD_TYPE:
+ for (tree fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
+ if (TREE_CODE (fld) == FIELD_DECL)
+ {
+ HOST_WIDE_INT fld_offset = offset + int_bit_position (fld);
+ tree ft = TREE_TYPE (fld);
+ emit_move_elementwise (ft, target, source, fld_offset,
+ call_param_p);
+ }
+ break;
+
+ case ARRAY_TYPE:
+ {
+ tree elem_type = TREE_TYPE (type);
+ HOST_WIDE_INT el_size = tree_to_shwi (TYPE_SIZE (elem_type));
+ gcc_assert (el_size > 0);
+
+ offset_int idx, max;
+ /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
+ if (extract_min_max_idx_from_array (type, &idx, &max))
+ {
+ HOST_WIDE_INT el_offset = offset;
+ for (; idx <= max; ++idx)
+ {
+ emit_move_elementwise (elem_type, target, source, el_offset,
+ call_param_p);
+ el_offset += el_size;
+ }
+ }
+ }
+ break;
+ default:
+ machine_mode mode = TYPE_MODE (type);
+
+ rtx ntgt = adjust_address (target, mode, offset / BITS_PER_UNIT);
+ rtx nsrc = adjust_address (source, mode, offset / BITS_PER_UNIT);
+
+ /* TODO: Figure out whether the following is actually necessary. */
+ if (target == ntgt)
+ ntgt = copy_rtx (target);
+ if (source == nsrc)
+ nsrc = copy_rtx (source);
+
+ gcc_assert (mode != VOIDmode);
+ if (mode != BLKmode)
+ emit_move_insn (ntgt, nsrc);
+ else
+ {
+ /* For example vector gimple registers can end up here. */
+ rtx size = expand_expr (TYPE_SIZE_UNIT (type), NULL_RTX,
+ TYPE_MODE (sizetype), EXPAND_NORMAL);
+ emit_block_move (ntgt, nsrc, size,
+ (call_param_p
+ ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
+ }
+ break;
+ }
+ return;
+}
+
/* Generate code for computing expression EXP,
and storing the value into TARGET.
@@ -5713,9 +5788,29 @@ store_expr_with_bounds (tree exp, rtx target, int call_param_p,
emit_group_store (target, temp, TREE_TYPE (exp),
int_size_in_bytes (TREE_TYPE (exp)));
else if (GET_MODE (temp) == BLKmode)
- emit_block_move (target, temp, expr_size (exp),
- (call_param_p
- ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
+ {
+ /* Copying smallish BLKmode structures with emit_block_move and thus
+ by-pieces can result in store-to-load stalls. So copy some simple
+ small aggregates element or field-wise. */
+ if (GET_MODE (target) == BLKmode
+ && AGGREGATE_TYPE_P (TREE_TYPE (exp))
+ && !TREE_ADDRESSABLE (TREE_TYPE (exp))
+ && tree_fits_shwi_p (TYPE_SIZE (TREE_TYPE (exp)))
+ && (tree_to_shwi (TYPE_SIZE (TREE_TYPE (exp)))
+ <= (PARAM_VALUE (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY)
+ * BITS_PER_UNIT))
+ && simple_mix_of_records_and_arrays_p (TREE_TYPE (exp), false))
+ {
+ /* FIXME: Can this happen? What would it mean? */
+ gcc_assert (!reverse);
+ emit_move_elementwise (TREE_TYPE (exp), target, temp, 0,
+ call_param_p);
+ }
+ else
+ emit_block_move (target, temp, expr_size (exp),
+ (call_param_p
+ ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
+ }
/* If we emit a nontemporal store, there is nothing else to do. */
else if (nontemporal && emit_storent_insn (target, temp))
;
diff --git a/gcc/ipa-cp.c b/gcc/ipa-cp.c
index 6b3d8d7364c..7d6019bbd30 100644
--- a/gcc/ipa-cp.c
+++ b/gcc/ipa-cp.c
@@ -124,6 +124,7 @@ along with GCC; see the file COPYING3. If not see
#include "tree-ssa-ccp.h"
#include "stringpool.h"
#include "attribs.h"
+#include "tree-sra.h"
template <typename valtype> class ipcp_value;
diff --git a/gcc/ipa-prop.h b/gcc/ipa-prop.h
index fa5bed49ee0..2313cc884ed 100644
--- a/gcc/ipa-prop.h
+++ b/gcc/ipa-prop.h
@@ -877,10 +877,6 @@ ipa_parm_adjustment *ipa_get_adjustment_candidate (tree **, bool *,
void ipa_release_body_info (struct ipa_func_body_info *);
tree ipa_get_callee_param_type (struct cgraph_edge *e, int i);
-/* From tree-sra.c: */
-tree build_ref_for_offset (location_t, tree, HOST_WIDE_INT, bool, tree,
- gimple_stmt_iterator *, bool);
-
/* In ipa-cp.c */
void ipa_cp_c_finalize (void);
diff --git a/gcc/params.def b/gcc/params.def
index e55afc28053..5e19f1414a0 100644
--- a/gcc/params.def
+++ b/gcc/params.def
@@ -1294,6 +1294,12 @@ DEFPARAM (PARAM_VECT_EPILOGUES_NOMASK,
"Enable loop epilogue vectorization using smaller vector size.",
0, 0, 1)
+DEFPARAM (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
+ "max-size-for-elementwise-copy",
+ "Maximum size in bytes of a structure or array to by considered for "
+ "copying by its individual fields or elements",
+ 0, 0, 512)
+
/*
Local variables:
diff --git a/gcc/testsuite/gcc.target/i386/pr80689-1.c b/gcc/testsuite/gcc.target/i386/pr80689-1.c
new file mode 100644
index 00000000000..4156d4fba45
--- /dev/null
+++ b/gcc/testsuite/gcc.target/i386/pr80689-1.c
@@ -0,0 +1,38 @@
+/* { dg-do compile } */
+/* { dg-options "-O2" } */
+
+typedef struct st1
+{
+ long unsigned int a,b;
+ long int c,d;
+}R;
+
+typedef struct st2
+{
+ int t;
+ R reg;
+}N;
+
+void Set (const R *region, N *n_info );
+
+void test(N *n_obj ,const long unsigned int a, const long unsigned int b, const long int c,const long int d)
+{
+ R reg;
+
+ reg.a=a;
+ reg.b=b;
+ reg.c=c;
+ reg.d=d;
+ Set (®, n_obj);
+
+}
+
+void Set (const R *reg, N *n_obj )
+{
+ n_obj->reg=(*reg);
+}
+
+
+/* { dg-final { scan-assembler-not "%(x|y|z)mm\[0-9\]+" } } */
+/* { dg-final { scan-assembler-not "movdqu" } } */
+/* { dg-final { scan-assembler-not "movups" } } */
diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
index bac593951e7..ade97964205 100644
--- a/gcc/tree-sra.c
+++ b/gcc/tree-sra.c
@@ -104,6 +104,7 @@ along with GCC; see the file COPYING3. If not see
#include "ipa-fnsummary.h"
#include "ipa-utils.h"
#include "builtins.h"
+#include "tree-sra.h"
/* Enumeration of all aggregate reductions we can do. */
enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
@@ -952,14 +953,14 @@ create_access (tree expr, gimple *stmt, bool write)
}
-/* Return true iff TYPE is scalarizable - i.e. a RECORD_TYPE or fixed-length
- ARRAY_TYPE with fields that are either of gimple register types (excluding
- bit-fields) or (recursively) scalarizable types. CONST_DECL must be true if
- we are considering a decl from constant pool. If it is false, char arrays
- will be refused. */
+/* Return true if TYPE consists of RECORD_TYPE or fixed-length ARRAY_TYPE with
+ fields/elements that are not bit-fields and are either register types or
+ recursively comply with simple_mix_of_records_and_arrays_p. Furthermore, if
+ ALLOW_CHAR_ARRAYS is false, the function will return false also if TYPE
+ contains an array of elements that only have one byte. */
-static bool
-scalarizable_type_p (tree type, bool const_decl)
+bool
+simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays)
{
gcc_assert (!is_gimple_reg_type (type));
if (type_contains_placeholder_p (type))
@@ -977,7 +978,7 @@ scalarizable_type_p (tree type, bool const_decl)
return false;
if (!is_gimple_reg_type (ft)
- && !scalarizable_type_p (ft, const_decl))
+ && !simple_mix_of_records_and_arrays_p (ft, allow_char_arrays))
return false;
}
@@ -986,7 +987,7 @@ scalarizable_type_p (tree type, bool const_decl)
case ARRAY_TYPE:
{
HOST_WIDE_INT min_elem_size;
- if (const_decl)
+ if (allow_char_arrays)
min_elem_size = 0;
else
min_elem_size = BITS_PER_UNIT;
@@ -1008,7 +1009,7 @@ scalarizable_type_p (tree type, bool const_decl)
tree elem = TREE_TYPE (type);
if (!is_gimple_reg_type (elem)
- && !scalarizable_type_p (elem, const_decl))
+ && !simple_mix_of_records_and_arrays_p (elem, allow_char_arrays))
return false;
return true;
}
@@ -1017,10 +1018,38 @@ scalarizable_type_p (tree type, bool const_decl)
}
}
-static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree, tree);
+static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree,
+ tree);
+
+/* For a given array TYPE, return false if its domain does not have any maximum
+ value. Otherwise calculate MIN and MAX indices of the first and the last
+ element. */
+
+bool
+extract_min_max_idx_from_array (tree type, offset_int *min, offset_int *max)
+{
+ tree domain = TYPE_DOMAIN (type);
+ tree minidx = TYPE_MIN_VALUE (domain);
+ gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
+ tree maxidx = TYPE_MAX_VALUE (domain);
+ if (!maxidx)
+ return false;
+ gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
+
+ /* MINIDX and MAXIDX are inclusive, and must be interpreted in
+ DOMAIN (e.g. signed int, whereas min/max may be size_int). */
+ *min = wi::to_offset (minidx);
+ *max = wi::to_offset (maxidx);
+ if (!TYPE_UNSIGNED (domain))
+ {
+ *min = wi::sext (*min, TYPE_PRECISION (domain));
+ *max = wi::sext (*max, TYPE_PRECISION (domain));
+ }
+ return true;
+}
/* Create total_scalarization accesses for all scalar fields of a member
- of type DECL_TYPE conforming to scalarizable_type_p. BASE
+ of type DECL_TYPE conforming to simple_mix_of_records_and_arrays_p. BASE
must be the top-most VAR_DECL representing the variable; within that,
OFFSET locates the member and REF must be the memory reference expression for
the member. */
@@ -1047,27 +1076,14 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
{
tree elemtype = TREE_TYPE (decl_type);
tree elem_size = TYPE_SIZE (elemtype);
- gcc_assert (elem_size && tree_fits_shwi_p (elem_size));
HOST_WIDE_INT el_size = tree_to_shwi (elem_size);
gcc_assert (el_size > 0);
- tree minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (decl_type));
- gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
- tree maxidx = TYPE_MAX_VALUE (TYPE_DOMAIN (decl_type));
+ offset_int idx, max;
/* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
- if (maxidx)
+ if (extract_min_max_idx_from_array (decl_type, &idx, &max))
{
- gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
tree domain = TYPE_DOMAIN (decl_type);
- /* MINIDX and MAXIDX are inclusive, and must be interpreted in
- DOMAIN (e.g. signed int, whereas min/max may be size_int). */
- offset_int idx = wi::to_offset (minidx);
- offset_int max = wi::to_offset (maxidx);
- if (!TYPE_UNSIGNED (domain))
- {
- idx = wi::sext (idx, TYPE_PRECISION (domain));
- max = wi::sext (max, TYPE_PRECISION (domain));
- }
for (int el_off = offset; idx <= max; ++idx)
{
tree nref = build4 (ARRAY_REF, elemtype,
@@ -1088,10 +1104,10 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
}
/* Create total_scalarization accesses for a member of type TYPE, which must
- satisfy either is_gimple_reg_type or scalarizable_type_p. BASE must be the
- top-most VAR_DECL representing the variable; within that, POS and SIZE locate
- the member, REVERSE gives its torage order. and REF must be the reference
- expression for it. */
+ satisfy either is_gimple_reg_type or simple_mix_of_records_and_arrays_p.
+ BASE must be the top-most VAR_DECL representing the variable; within that,
+ POS and SIZE locate the member, REVERSE gives its torage order. and REF must
+ be the reference expression for it. */
static void
scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
@@ -1111,7 +1127,8 @@ scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
}
/* Create a total_scalarization access for VAR as a whole. VAR must be of a
- RECORD_TYPE or ARRAY_TYPE conforming to scalarizable_type_p. */
+ RECORD_TYPE or ARRAY_TYPE conforming to
+ simple_mix_of_records_and_arrays_p. */
static void
create_total_scalarization_access (tree var)
@@ -2803,8 +2820,9 @@ analyze_all_variable_accesses (void)
{
tree var = candidate (i);
- if (VAR_P (var) && scalarizable_type_p (TREE_TYPE (var),
- constant_decl_p (var)))
+ if (VAR_P (var)
+ && simple_mix_of_records_and_arrays_p (TREE_TYPE (var),
+ constant_decl_p (var)))
{
if (tree_to_uhwi (TYPE_SIZE (TREE_TYPE (var)))
<= max_scalarization_size)
diff --git a/gcc/tree-sra.h b/gcc/tree-sra.h
new file mode 100644
index 00000000000..dc901385994
--- /dev/null
+++ b/gcc/tree-sra.h
@@ -0,0 +1,33 @@
+/* tree-sra.h - Run-time parameters.
+ Copyright (C) 2017 Free Software Foundation, Inc.
+
+This file is part of GCC.
+
+GCC is free software; you can redistribute it and/or modify it under
+the terms of the GNU General Public License as published by the Free
+Software Foundation; either version 3, or (at your option) any later
+version.
+
+GCC is distributed in the hope that it will be useful, but WITHOUT ANY
+WARRANTY; without even the implied warranty of MERCHANTABILITY or
+FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
+for more details.
+
+You should have received a copy of the GNU General Public License
+along with GCC; see the file COPYING3. If not see
+<http://www.gnu.org/licenses/>. */
+
+#ifndef TREE_SRA_H
+#define TREE_SRA_H
+
+
+bool simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays);
+bool extract_min_max_idx_from_array (tree type, offset_int *idx,
+ offset_int *max);
+tree build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
+ bool reverse, tree exp_type,
+ gimple_stmt_iterator *gsi, bool insert_after);
+
+
+
+#endif /* TREE_SRA_H */
--
2.14.1
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-10-13 16:31 [RFC, PR 80689] Copy small aggregates element-wise Martin Jambor
@ 2017-10-17 11:44 ` Richard Biener
2017-10-26 12:23 ` Martin Jambor
0 siblings, 1 reply; 26+ messages in thread
From: Richard Biener @ 2017-10-17 11:44 UTC (permalink / raw)
To: GCC Patches, Jan Hubicka
On Fri, Oct 13, 2017 at 6:13 PM, Martin Jambor <mjambor@suse.cz> wrote:
> Hi,
>
> I'd like to request comments to the patch below which aims to fix PR
> 80689, which is an instance of a store-to-load forwarding stall on
> x86_64 CPUs in the Image Magick benchmark, which is responsible for a
> slow down of up to 9% compared to gcc 6, depending on options and HW
> used. (Actually, I have just seen 24% in one specific combination but
> for various reasons can no longer verify it today.)
>
> The revision causing the regression is 237074, which increased the
> size of the mode for copying aggregates "by pieces" to 128 bits,
> incurring big stalls when the values being copied are also still being
> stored in a smaller data type or if the copied values are loaded in a
> smaller types shortly afterwards. Such situations happen in Image
> Magick even across calls, which means that any non-IPA flow-sensitive
> approach would not detect them. Therefore, the patch simply changes
> the way we copy small BLKmode data that are simple combinations of
> records and arrays (meaning no unions, bit-fields but also character
> arrays are disallowed) and simply copies them one field and/or element
> at a time.
>
> "Small" in this RFC patch means up to 35 bytes on x86_64 and i386 CPUs
> (the structure in the benchmark has 32 bytes) but is subject to change
> after more benchmarking and is actually zero - meaning element copying
> never happens - on other architectures. I believe that any
> architecture with a store buffer can benefit but it's probably better
> to leave it to their maintainers to find a different default value. I
> am not sure this is how such HW-dependant decisions should be done and
> is the primary reason, why I am sending this RFC first.
>
> I have decided to implement this change at the expansion level because
> at that point the type information is still readily available and at
> the same time we can also handle various implicit copies, for example
> those passing a parameter. I found I could re-use some bits and
> pieces of tree-SRA and so I did, creating tree-sra.h header file in
> the process.
>
> I am fully aware that in the final patch the new parameter, or indeed
> any new parameters, need to be documented. I have skipped that
> intentionally now and will write the documentation if feedback here is
> generally good.
>
> I have bootstrapped and tested this patch on x86_64-linux, with
> different values of the parameter and only found problems with
> unreasonably high values leading to OOM. I have done the same with a
> previous version of the patch which was equivalent to the limit being
> 64 bytes on aarch64-linux, ppc64le-linux and ia64-linux and only ran
> into failures of tests which assumed that structure padding was copied
> in aggregate copies (mostly gcc.target/aarch64/aapcs64/ stuff but also
> for example gcc.dg/vmx/varargs-4.c).
>
> The patch decreases the SPEC 2017 "rate" run-time of imagick by 9% and
> 8% at -O2 and -Ofast compilation levels respectively on one particular
> new AMD CPU and by 6% and 3% on one particular old Intel machine.
>
> Thanks in advance for any comments,
I wonder if you can at the place you choose to hook this in elide any
copying of padding between fields.
I'd rather have hooked such "high level" optimization in expand_assignment
where you can be reasonably sure you're seeing an actual source-level construct.
35 bytes seems to be much - what is the code-size impact?
IIRC the reason this may be slow isn't loading in smaller types than stored
before by the copy - the store buffers can handle this reasonably well. It's
solely when previous smaller stores are
a1) not mergeabe in the store buffer
a2) not merged because earlier stores are already committed
and
b) loaded afterwards as a type that would access multiple store buffers
a) would be sure to happen in case b) involves accessing padding. Is the
Image Magick case one that involves padding in the structure?
Richard.
> Martin
>
>
> 2017-10-12 Martin Jambor <mjambor@suse.cz>
>
> PR target/80689
> * tree-sra.h: New file.
> * ipa-prop.h: Moved declaration of build_ref_for_offset to
> tree-sra.h.
> * expr.c: Include params.h and tree-sra.h.
> (emit_move_elementwise): New function.
> (store_expr_with_bounds): Optionally use it.
> * ipa-cp.c: Include tree-sra.h.
> * params.def (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY): New.
> * config/i386/i386.c (ix86_option_override_internal): Set
> PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY to 35.
> * tree-sra.c: Include tree-sra.h.
> (scalarizable_type_p): Renamed to
> simple_mix_of_records_and_arrays_p, made public, renamed the
> second parameter to allow_char_arrays.
> (extract_min_max_idx_from_array): New function.
> (completely_scalarize): Moved bits of the function to
> extract_min_max_idx_from_array.
>
> testsuite/
> * gcc.target/i386/pr80689-1.c: New test.
> ---
> gcc/config/i386/i386.c | 4 ++
> gcc/expr.c | 103 ++++++++++++++++++++++++++++--
> gcc/ipa-cp.c | 1 +
> gcc/ipa-prop.h | 4 --
> gcc/params.def | 6 ++
> gcc/testsuite/gcc.target/i386/pr80689-1.c | 38 +++++++++++
> gcc/tree-sra.c | 86 +++++++++++++++----------
> gcc/tree-sra.h | 33 ++++++++++
> 8 files changed, 233 insertions(+), 42 deletions(-)
> create mode 100644 gcc/testsuite/gcc.target/i386/pr80689-1.c
> create mode 100644 gcc/tree-sra.h
>
> diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c
> index 1ee8351c21f..87f602e7ead 100644
> --- a/gcc/config/i386/i386.c
> +++ b/gcc/config/i386/i386.c
> @@ -6511,6 +6511,10 @@ ix86_option_override_internal (bool main_args_p,
> ix86_tune_cost->l2_cache_size,
> opts->x_param_values,
> opts_set->x_param_values);
> + maybe_set_param_value (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
> + 35,
> + opts->x_param_values,
> + opts_set->x_param_values);
>
> /* Enable sw prefetching at -O3 for CPUS that prefetching is helpful. */
> if (opts->x_flag_prefetch_loop_arrays < 0
> diff --git a/gcc/expr.c b/gcc/expr.c
> index 134ee731c29..dff24e7f166 100644
> --- a/gcc/expr.c
> +++ b/gcc/expr.c
> @@ -61,7 +61,8 @@ along with GCC; see the file COPYING3. If not see
> #include "tree-chkp.h"
> #include "rtl-chkp.h"
> #include "ccmp.h"
> -
> +#include "params.h"
> +#include "tree-sra.h"
>
> /* If this is nonzero, we do not bother generating VOLATILE
> around volatile memory references, and we are willing to
> @@ -5340,6 +5341,80 @@ emit_storent_insn (rtx to, rtx from)
> return maybe_expand_insn (code, 2, ops);
> }
>
> +/* Generate code for copying data of type TYPE at SOURCE plus OFFSET to TARGET
> + plus OFFSET, but do so element-wise and/or field-wise for each record and
> + array within TYPE. TYPE must either be a register type or an aggregate
> + complying with scalarizable_type_p.
> +
> + If CALL_PARAM_P is nonzero, this is a store into a call param on the
> + stack, and block moves may need to be treated specially. */
> +
> +static void
> +emit_move_elementwise (tree type, rtx target, rtx source, HOST_WIDE_INT offset,
> + int call_param_p)
> +{
> + switch (TREE_CODE (type))
> + {
> + case RECORD_TYPE:
> + for (tree fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
> + if (TREE_CODE (fld) == FIELD_DECL)
> + {
> + HOST_WIDE_INT fld_offset = offset + int_bit_position (fld);
> + tree ft = TREE_TYPE (fld);
> + emit_move_elementwise (ft, target, source, fld_offset,
> + call_param_p);
> + }
> + break;
> +
> + case ARRAY_TYPE:
> + {
> + tree elem_type = TREE_TYPE (type);
> + HOST_WIDE_INT el_size = tree_to_shwi (TYPE_SIZE (elem_type));
> + gcc_assert (el_size > 0);
> +
> + offset_int idx, max;
> + /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
> + if (extract_min_max_idx_from_array (type, &idx, &max))
> + {
> + HOST_WIDE_INT el_offset = offset;
> + for (; idx <= max; ++idx)
> + {
> + emit_move_elementwise (elem_type, target, source, el_offset,
> + call_param_p);
> + el_offset += el_size;
> + }
> + }
> + }
> + break;
> + default:
> + machine_mode mode = TYPE_MODE (type);
> +
> + rtx ntgt = adjust_address (target, mode, offset / BITS_PER_UNIT);
> + rtx nsrc = adjust_address (source, mode, offset / BITS_PER_UNIT);
> +
> + /* TODO: Figure out whether the following is actually necessary. */
> + if (target == ntgt)
> + ntgt = copy_rtx (target);
> + if (source == nsrc)
> + nsrc = copy_rtx (source);
> +
> + gcc_assert (mode != VOIDmode);
> + if (mode != BLKmode)
> + emit_move_insn (ntgt, nsrc);
> + else
> + {
> + /* For example vector gimple registers can end up here. */
> + rtx size = expand_expr (TYPE_SIZE_UNIT (type), NULL_RTX,
> + TYPE_MODE (sizetype), EXPAND_NORMAL);
> + emit_block_move (ntgt, nsrc, size,
> + (call_param_p
> + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> + }
> + break;
> + }
> + return;
> +}
> +
> /* Generate code for computing expression EXP,
> and storing the value into TARGET.
>
> @@ -5713,9 +5788,29 @@ store_expr_with_bounds (tree exp, rtx target, int call_param_p,
> emit_group_store (target, temp, TREE_TYPE (exp),
> int_size_in_bytes (TREE_TYPE (exp)));
> else if (GET_MODE (temp) == BLKmode)
> - emit_block_move (target, temp, expr_size (exp),
> - (call_param_p
> - ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> + {
> + /* Copying smallish BLKmode structures with emit_block_move and thus
> + by-pieces can result in store-to-load stalls. So copy some simple
> + small aggregates element or field-wise. */
> + if (GET_MODE (target) == BLKmode
> + && AGGREGATE_TYPE_P (TREE_TYPE (exp))
> + && !TREE_ADDRESSABLE (TREE_TYPE (exp))
> + && tree_fits_shwi_p (TYPE_SIZE (TREE_TYPE (exp)))
> + && (tree_to_shwi (TYPE_SIZE (TREE_TYPE (exp)))
> + <= (PARAM_VALUE (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY)
> + * BITS_PER_UNIT))
> + && simple_mix_of_records_and_arrays_p (TREE_TYPE (exp), false))
> + {
> + /* FIXME: Can this happen? What would it mean? */
> + gcc_assert (!reverse);
> + emit_move_elementwise (TREE_TYPE (exp), target, temp, 0,
> + call_param_p);
> + }
> + else
> + emit_block_move (target, temp, expr_size (exp),
> + (call_param_p
> + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> + }
> /* If we emit a nontemporal store, there is nothing else to do. */
> else if (nontemporal && emit_storent_insn (target, temp))
> ;
> diff --git a/gcc/ipa-cp.c b/gcc/ipa-cp.c
> index 6b3d8d7364c..7d6019bbd30 100644
> --- a/gcc/ipa-cp.c
> +++ b/gcc/ipa-cp.c
> @@ -124,6 +124,7 @@ along with GCC; see the file COPYING3. If not see
> #include "tree-ssa-ccp.h"
> #include "stringpool.h"
> #include "attribs.h"
> +#include "tree-sra.h"
>
> template <typename valtype> class ipcp_value;
>
> diff --git a/gcc/ipa-prop.h b/gcc/ipa-prop.h
> index fa5bed49ee0..2313cc884ed 100644
> --- a/gcc/ipa-prop.h
> +++ b/gcc/ipa-prop.h
> @@ -877,10 +877,6 @@ ipa_parm_adjustment *ipa_get_adjustment_candidate (tree **, bool *,
> void ipa_release_body_info (struct ipa_func_body_info *);
> tree ipa_get_callee_param_type (struct cgraph_edge *e, int i);
>
> -/* From tree-sra.c: */
> -tree build_ref_for_offset (location_t, tree, HOST_WIDE_INT, bool, tree,
> - gimple_stmt_iterator *, bool);
> -
> /* In ipa-cp.c */
> void ipa_cp_c_finalize (void);
>
> diff --git a/gcc/params.def b/gcc/params.def
> index e55afc28053..5e19f1414a0 100644
> --- a/gcc/params.def
> +++ b/gcc/params.def
> @@ -1294,6 +1294,12 @@ DEFPARAM (PARAM_VECT_EPILOGUES_NOMASK,
> "Enable loop epilogue vectorization using smaller vector size.",
> 0, 0, 1)
>
> +DEFPARAM (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
> + "max-size-for-elementwise-copy",
> + "Maximum size in bytes of a structure or array to by considered for "
> + "copying by its individual fields or elements",
> + 0, 0, 512)
> +
> /*
>
> Local variables:
> diff --git a/gcc/testsuite/gcc.target/i386/pr80689-1.c b/gcc/testsuite/gcc.target/i386/pr80689-1.c
> new file mode 100644
> index 00000000000..4156d4fba45
> --- /dev/null
> +++ b/gcc/testsuite/gcc.target/i386/pr80689-1.c
> @@ -0,0 +1,38 @@
> +/* { dg-do compile } */
> +/* { dg-options "-O2" } */
> +
> +typedef struct st1
> +{
> + long unsigned int a,b;
> + long int c,d;
> +}R;
> +
> +typedef struct st2
> +{
> + int t;
> + R reg;
> +}N;
> +
> +void Set (const R *region, N *n_info );
> +
> +void test(N *n_obj ,const long unsigned int a, const long unsigned int b, const long int c,const long int d)
> +{
> + R reg;
> +
> + reg.a=a;
> + reg.b=b;
> + reg.c=c;
> + reg.d=d;
> + Set (®, n_obj);
> +
> +}
> +
> +void Set (const R *reg, N *n_obj )
> +{
> + n_obj->reg=(*reg);
> +}
> +
> +
> +/* { dg-final { scan-assembler-not "%(x|y|z)mm\[0-9\]+" } } */
> +/* { dg-final { scan-assembler-not "movdqu" } } */
> +/* { dg-final { scan-assembler-not "movups" } } */
> diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
> index bac593951e7..ade97964205 100644
> --- a/gcc/tree-sra.c
> +++ b/gcc/tree-sra.c
> @@ -104,6 +104,7 @@ along with GCC; see the file COPYING3. If not see
> #include "ipa-fnsummary.h"
> #include "ipa-utils.h"
> #include "builtins.h"
> +#include "tree-sra.h"
>
> /* Enumeration of all aggregate reductions we can do. */
> enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
> @@ -952,14 +953,14 @@ create_access (tree expr, gimple *stmt, bool write)
> }
>
>
> -/* Return true iff TYPE is scalarizable - i.e. a RECORD_TYPE or fixed-length
> - ARRAY_TYPE with fields that are either of gimple register types (excluding
> - bit-fields) or (recursively) scalarizable types. CONST_DECL must be true if
> - we are considering a decl from constant pool. If it is false, char arrays
> - will be refused. */
> +/* Return true if TYPE consists of RECORD_TYPE or fixed-length ARRAY_TYPE with
> + fields/elements that are not bit-fields and are either register types or
> + recursively comply with simple_mix_of_records_and_arrays_p. Furthermore, if
> + ALLOW_CHAR_ARRAYS is false, the function will return false also if TYPE
> + contains an array of elements that only have one byte. */
>
> -static bool
> -scalarizable_type_p (tree type, bool const_decl)
> +bool
> +simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays)
> {
> gcc_assert (!is_gimple_reg_type (type));
> if (type_contains_placeholder_p (type))
> @@ -977,7 +978,7 @@ scalarizable_type_p (tree type, bool const_decl)
> return false;
>
> if (!is_gimple_reg_type (ft)
> - && !scalarizable_type_p (ft, const_decl))
> + && !simple_mix_of_records_and_arrays_p (ft, allow_char_arrays))
> return false;
> }
>
> @@ -986,7 +987,7 @@ scalarizable_type_p (tree type, bool const_decl)
> case ARRAY_TYPE:
> {
> HOST_WIDE_INT min_elem_size;
> - if (const_decl)
> + if (allow_char_arrays)
> min_elem_size = 0;
> else
> min_elem_size = BITS_PER_UNIT;
> @@ -1008,7 +1009,7 @@ scalarizable_type_p (tree type, bool const_decl)
>
> tree elem = TREE_TYPE (type);
> if (!is_gimple_reg_type (elem)
> - && !scalarizable_type_p (elem, const_decl))
> + && !simple_mix_of_records_and_arrays_p (elem, allow_char_arrays))
> return false;
> return true;
> }
> @@ -1017,10 +1018,38 @@ scalarizable_type_p (tree type, bool const_decl)
> }
> }
>
> -static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree, tree);
> +static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree,
> + tree);
> +
> +/* For a given array TYPE, return false if its domain does not have any maximum
> + value. Otherwise calculate MIN and MAX indices of the first and the last
> + element. */
> +
> +bool
> +extract_min_max_idx_from_array (tree type, offset_int *min, offset_int *max)
> +{
> + tree domain = TYPE_DOMAIN (type);
> + tree minidx = TYPE_MIN_VALUE (domain);
> + gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
> + tree maxidx = TYPE_MAX_VALUE (domain);
> + if (!maxidx)
> + return false;
> + gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
> +
> + /* MINIDX and MAXIDX are inclusive, and must be interpreted in
> + DOMAIN (e.g. signed int, whereas min/max may be size_int). */
> + *min = wi::to_offset (minidx);
> + *max = wi::to_offset (maxidx);
> + if (!TYPE_UNSIGNED (domain))
> + {
> + *min = wi::sext (*min, TYPE_PRECISION (domain));
> + *max = wi::sext (*max, TYPE_PRECISION (domain));
> + }
> + return true;
> +}
>
> /* Create total_scalarization accesses for all scalar fields of a member
> - of type DECL_TYPE conforming to scalarizable_type_p. BASE
> + of type DECL_TYPE conforming to simple_mix_of_records_and_arrays_p. BASE
> must be the top-most VAR_DECL representing the variable; within that,
> OFFSET locates the member and REF must be the memory reference expression for
> the member. */
> @@ -1047,27 +1076,14 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
> {
> tree elemtype = TREE_TYPE (decl_type);
> tree elem_size = TYPE_SIZE (elemtype);
> - gcc_assert (elem_size && tree_fits_shwi_p (elem_size));
> HOST_WIDE_INT el_size = tree_to_shwi (elem_size);
> gcc_assert (el_size > 0);
>
> - tree minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (decl_type));
> - gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
> - tree maxidx = TYPE_MAX_VALUE (TYPE_DOMAIN (decl_type));
> + offset_int idx, max;
> /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
> - if (maxidx)
> + if (extract_min_max_idx_from_array (decl_type, &idx, &max))
> {
> - gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
> tree domain = TYPE_DOMAIN (decl_type);
> - /* MINIDX and MAXIDX are inclusive, and must be interpreted in
> - DOMAIN (e.g. signed int, whereas min/max may be size_int). */
> - offset_int idx = wi::to_offset (minidx);
> - offset_int max = wi::to_offset (maxidx);
> - if (!TYPE_UNSIGNED (domain))
> - {
> - idx = wi::sext (idx, TYPE_PRECISION (domain));
> - max = wi::sext (max, TYPE_PRECISION (domain));
> - }
> for (int el_off = offset; idx <= max; ++idx)
> {
> tree nref = build4 (ARRAY_REF, elemtype,
> @@ -1088,10 +1104,10 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
> }
>
> /* Create total_scalarization accesses for a member of type TYPE, which must
> - satisfy either is_gimple_reg_type or scalarizable_type_p. BASE must be the
> - top-most VAR_DECL representing the variable; within that, POS and SIZE locate
> - the member, REVERSE gives its torage order. and REF must be the reference
> - expression for it. */
> + satisfy either is_gimple_reg_type or simple_mix_of_records_and_arrays_p.
> + BASE must be the top-most VAR_DECL representing the variable; within that,
> + POS and SIZE locate the member, REVERSE gives its torage order. and REF must
> + be the reference expression for it. */
>
> static void
> scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
> @@ -1111,7 +1127,8 @@ scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
> }
>
> /* Create a total_scalarization access for VAR as a whole. VAR must be of a
> - RECORD_TYPE or ARRAY_TYPE conforming to scalarizable_type_p. */
> + RECORD_TYPE or ARRAY_TYPE conforming to
> + simple_mix_of_records_and_arrays_p. */
>
> static void
> create_total_scalarization_access (tree var)
> @@ -2803,8 +2820,9 @@ analyze_all_variable_accesses (void)
> {
> tree var = candidate (i);
>
> - if (VAR_P (var) && scalarizable_type_p (TREE_TYPE (var),
> - constant_decl_p (var)))
> + if (VAR_P (var)
> + && simple_mix_of_records_and_arrays_p (TREE_TYPE (var),
> + constant_decl_p (var)))
> {
> if (tree_to_uhwi (TYPE_SIZE (TREE_TYPE (var)))
> <= max_scalarization_size)
> diff --git a/gcc/tree-sra.h b/gcc/tree-sra.h
> new file mode 100644
> index 00000000000..dc901385994
> --- /dev/null
> +++ b/gcc/tree-sra.h
> @@ -0,0 +1,33 @@
> +/* tree-sra.h - Run-time parameters.
> + Copyright (C) 2017 Free Software Foundation, Inc.
> +
> +This file is part of GCC.
> +
> +GCC is free software; you can redistribute it and/or modify it under
> +the terms of the GNU General Public License as published by the Free
> +Software Foundation; either version 3, or (at your option) any later
> +version.
> +
> +GCC is distributed in the hope that it will be useful, but WITHOUT ANY
> +WARRANTY; without even the implied warranty of MERCHANTABILITY or
> +FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
> +for more details.
> +
> +You should have received a copy of the GNU General Public License
> +along with GCC; see the file COPYING3. If not see
> +<http://www.gnu.org/licenses/>. */
> +
> +#ifndef TREE_SRA_H
> +#define TREE_SRA_H
> +
> +
> +bool simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays);
> +bool extract_min_max_idx_from_array (tree type, offset_int *idx,
> + offset_int *max);
> +tree build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
> + bool reverse, tree exp_type,
> + gimple_stmt_iterator *gsi, bool insert_after);
> +
> +
> +
> +#endif /* TREE_SRA_H */
> --
> 2.14.1
>
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-10-17 11:44 ` Richard Biener
@ 2017-10-26 12:23 ` Martin Jambor
2017-10-26 12:44 ` Richard Biener
2017-10-26 14:18 ` Michael Matz
0 siblings, 2 replies; 26+ messages in thread
From: Martin Jambor @ 2017-10-26 12:23 UTC (permalink / raw)
To: Richard Biener; +Cc: GCC Patches, Jan Hubicka
Hi,
On Tue, Oct 17, 2017 at 01:34:54PM +0200, Richard Biener wrote:
> On Fri, Oct 13, 2017 at 6:13 PM, Martin Jambor <mjambor@suse.cz> wrote:
> > Hi,
> >
> > I'd like to request comments to the patch below which aims to fix PR
> > 80689, which is an instance of a store-to-load forwarding stall on
> > x86_64 CPUs in the Image Magick benchmark, which is responsible for a
> > slow down of up to 9% compared to gcc 6, depending on options and HW
> > used. (Actually, I have just seen 24% in one specific combination but
> > for various reasons can no longer verify it today.)
> >
> > The revision causing the regression is 237074, which increased the
> > size of the mode for copying aggregates "by pieces" to 128 bits,
> > incurring big stalls when the values being copied are also still being
> > stored in a smaller data type or if the copied values are loaded in a
> > smaller types shortly afterwards. Such situations happen in Image
> > Magick even across calls, which means that any non-IPA flow-sensitive
> > approach would not detect them. Therefore, the patch simply changes
> > the way we copy small BLKmode data that are simple combinations of
> > records and arrays (meaning no unions, bit-fields but also character
> > arrays are disallowed) and simply copies them one field and/or element
> > at a time.
> >
> > "Small" in this RFC patch means up to 35 bytes on x86_64 and i386 CPUs
> > (the structure in the benchmark has 32 bytes) but is subject to change
> > after more benchmarking and is actually zero - meaning element copying
> > never happens - on other architectures. I believe that any
> > architecture with a store buffer can benefit but it's probably better
> > to leave it to their maintainers to find a different default value. I
> > am not sure this is how such HW-dependant decisions should be done and
> > is the primary reason, why I am sending this RFC first.
> >
> > I have decided to implement this change at the expansion level because
> > at that point the type information is still readily available and at
> > the same time we can also handle various implicit copies, for example
> > those passing a parameter. I found I could re-use some bits and
> > pieces of tree-SRA and so I did, creating tree-sra.h header file in
> > the process.
> >
> > I am fully aware that in the final patch the new parameter, or indeed
> > any new parameters, need to be documented. I have skipped that
> > intentionally now and will write the documentation if feedback here is
> > generally good.
> >
> > I have bootstrapped and tested this patch on x86_64-linux, with
> > different values of the parameter and only found problems with
> > unreasonably high values leading to OOM. I have done the same with a
> > previous version of the patch which was equivalent to the limit being
> > 64 bytes on aarch64-linux, ppc64le-linux and ia64-linux and only ran
> > into failures of tests which assumed that structure padding was copied
> > in aggregate copies (mostly gcc.target/aarch64/aapcs64/ stuff but also
> > for example gcc.dg/vmx/varargs-4.c).
> >
> > The patch decreases the SPEC 2017 "rate" run-time of imagick by 9% and
> > 8% at -O2 and -Ofast compilation levels respectively on one particular
> > new AMD CPU and by 6% and 3% on one particular old Intel machine.
> >
> > Thanks in advance for any comments,
>
> I wonder if you can at the place you choose to hook this in elide any
> copying of padding between fields.
>
> I'd rather have hooked such "high level" optimization in
> expand_assignment where you can be reasonably sure you're seeing an
> actual source-level construct.
I have discussed this with Honza and we eventually decided to make the
elememnt-wise copy an alternative to emit_block_move (which uses the
larger mode for moving since GCC 7) exactly so that we handle not only
source-level assignments but also passing parameters by value and
other situations.
>
> 35 bytes seems to be much - what is the code-size impact?
I will find out and report on that. I need at least 32 bytes (four
long ints) to fix imagemagick, where the problematic structure is:
typedef struct _RectangleInfo
{
size_t
width,
height;
ssize_t
x,
y;
} RectangleInfo;
...so 8 longs, no padding. Since any aggregate having between 33-35
bytes needs to consist of smaller fields/elements, it seemed
reasonable to also copy them element-wise.
Nevertheless, I still intend to experiment with the limit, I sent out
this RFC exactly so that I don't spend a lot of time benchmarking
something that is eventually not deemed acceptable on principle.
>
> IIRC the reason this may be slow isn't loading in smaller types than stored
> before by the copy - the store buffers can handle this reasonably well. It's
> solely when previous smaller stores are
>
> a1) not mergeabe in the store buffer
> a2) not merged because earlier stores are already committed
>
> and
>
> b) loaded afterwards as a type that would access multiple store buffers
>
> a) would be sure to happen in case b) involves accessing padding. Is the
> Image Magick case one that involves padding in the structure?
As I said above, there is no padding.
Basically, what happens is that in a number of places, there is a
variable region of the aforementioned type and it is initialized and
passed to function SetPixelCacheNexusPixels in the following manner:
...
region.width=cache_info->columns;
region.height=1;
region.x=0;
region.y=y;
pixels=SetPixelCacheNexusPixels(cache_info,ReadMode,®ion,MagickTrue,
cache_nexus[id],exception);
...
and the first four statements in SetPixelCacheNexusPixels are:
assert(cache_info != (const CacheInfo *) NULL);
assert(cache_info->signature == MagickSignature);
if (cache_info->type == UndefinedCache)
return((PixelPacket *) NULL);
nexus_info->region=(*region);
with the last one generating the stalls, on both Zen-based machines
and also on 2-3 years old Intel CPUs.
I have had a look at what Agner Fog's micro-architecture document says
about store forwarding stalls and:
- on Broadwells and Haswells, any "write of any size is followed by
a read of a larger size" ioncurs a stall, which fits our example,
- on Skylakes: "A read that is bigger than the write, or a read that
covers both written and unwritten bytes, takes approximately 11
clock cycles extra" seems to apply
- on Intel silvermont, there will also be a stall because "A memory
write can be forwarded to a subsequent read of the same size or a
smaller size..."
- on Zens, Agner Fog says they work perfectly except when crossing a
page or when "A read that has a partial overlap with a preceding
write has a penalty of 6-7 clock cycles," which must be why I see
stalls.
So I guess the pending stores are not really merged even without
padding,
Martin
>
> Richard.
>
> > Martin
> >
> >
> > 2017-10-12 Martin Jambor <mjambor@suse.cz>
> >
> > PR target/80689
> > * tree-sra.h: New file.
> > * ipa-prop.h: Moved declaration of build_ref_for_offset to
> > tree-sra.h.
> > * expr.c: Include params.h and tree-sra.h.
> > (emit_move_elementwise): New function.
> > (store_expr_with_bounds): Optionally use it.
> > * ipa-cp.c: Include tree-sra.h.
> > * params.def (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY): New.
> > * config/i386/i386.c (ix86_option_override_internal): Set
> > PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY to 35.
> > * tree-sra.c: Include tree-sra.h.
> > (scalarizable_type_p): Renamed to
> > simple_mix_of_records_and_arrays_p, made public, renamed the
> > second parameter to allow_char_arrays.
> > (extract_min_max_idx_from_array): New function.
> > (completely_scalarize): Moved bits of the function to
> > extract_min_max_idx_from_array.
> >
> > testsuite/
> > * gcc.target/i386/pr80689-1.c: New test.
> > ---
> > gcc/config/i386/i386.c | 4 ++
> > gcc/expr.c | 103 ++++++++++++++++++++++++++++--
> > gcc/ipa-cp.c | 1 +
> > gcc/ipa-prop.h | 4 --
> > gcc/params.def | 6 ++
> > gcc/testsuite/gcc.target/i386/pr80689-1.c | 38 +++++++++++
> > gcc/tree-sra.c | 86 +++++++++++++++----------
> > gcc/tree-sra.h | 33 ++++++++++
> > 8 files changed, 233 insertions(+), 42 deletions(-)
> > create mode 100644 gcc/testsuite/gcc.target/i386/pr80689-1.c
> > create mode 100644 gcc/tree-sra.h
> >
> > diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c
> > index 1ee8351c21f..87f602e7ead 100644
> > --- a/gcc/config/i386/i386.c
> > +++ b/gcc/config/i386/i386.c
> > @@ -6511,6 +6511,10 @@ ix86_option_override_internal (bool main_args_p,
> > ix86_tune_cost->l2_cache_size,
> > opts->x_param_values,
> > opts_set->x_param_values);
> > + maybe_set_param_value (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
> > + 35,
> > + opts->x_param_values,
> > + opts_set->x_param_values);
> >
> > /* Enable sw prefetching at -O3 for CPUS that prefetching is helpful. */
> > if (opts->x_flag_prefetch_loop_arrays < 0
> > diff --git a/gcc/expr.c b/gcc/expr.c
> > index 134ee731c29..dff24e7f166 100644
> > --- a/gcc/expr.c
> > +++ b/gcc/expr.c
> > @@ -61,7 +61,8 @@ along with GCC; see the file COPYING3. If not see
> > #include "tree-chkp.h"
> > #include "rtl-chkp.h"
> > #include "ccmp.h"
> > -
> > +#include "params.h"
> > +#include "tree-sra.h"
> >
> > /* If this is nonzero, we do not bother generating VOLATILE
> > around volatile memory references, and we are willing to
> > @@ -5340,6 +5341,80 @@ emit_storent_insn (rtx to, rtx from)
> > return maybe_expand_insn (code, 2, ops);
> > }
> >
> > +/* Generate code for copying data of type TYPE at SOURCE plus OFFSET to TARGET
> > + plus OFFSET, but do so element-wise and/or field-wise for each record and
> > + array within TYPE. TYPE must either be a register type or an aggregate
> > + complying with scalarizable_type_p.
> > +
> > + If CALL_PARAM_P is nonzero, this is a store into a call param on the
> > + stack, and block moves may need to be treated specially. */
> > +
> > +static void
> > +emit_move_elementwise (tree type, rtx target, rtx source, HOST_WIDE_INT offset,
> > + int call_param_p)
> > +{
> > + switch (TREE_CODE (type))
> > + {
> > + case RECORD_TYPE:
> > + for (tree fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
> > + if (TREE_CODE (fld) == FIELD_DECL)
> > + {
> > + HOST_WIDE_INT fld_offset = offset + int_bit_position (fld);
> > + tree ft = TREE_TYPE (fld);
> > + emit_move_elementwise (ft, target, source, fld_offset,
> > + call_param_p);
> > + }
> > + break;
> > +
> > + case ARRAY_TYPE:
> > + {
> > + tree elem_type = TREE_TYPE (type);
> > + HOST_WIDE_INT el_size = tree_to_shwi (TYPE_SIZE (elem_type));
> > + gcc_assert (el_size > 0);
> > +
> > + offset_int idx, max;
> > + /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
> > + if (extract_min_max_idx_from_array (type, &idx, &max))
> > + {
> > + HOST_WIDE_INT el_offset = offset;
> > + for (; idx <= max; ++idx)
> > + {
> > + emit_move_elementwise (elem_type, target, source, el_offset,
> > + call_param_p);
> > + el_offset += el_size;
> > + }
> > + }
> > + }
> > + break;
> > + default:
> > + machine_mode mode = TYPE_MODE (type);
> > +
> > + rtx ntgt = adjust_address (target, mode, offset / BITS_PER_UNIT);
> > + rtx nsrc = adjust_address (source, mode, offset / BITS_PER_UNIT);
> > +
> > + /* TODO: Figure out whether the following is actually necessary. */
> > + if (target == ntgt)
> > + ntgt = copy_rtx (target);
> > + if (source == nsrc)
> > + nsrc = copy_rtx (source);
> > +
> > + gcc_assert (mode != VOIDmode);
> > + if (mode != BLKmode)
> > + emit_move_insn (ntgt, nsrc);
> > + else
> > + {
> > + /* For example vector gimple registers can end up here. */
> > + rtx size = expand_expr (TYPE_SIZE_UNIT (type), NULL_RTX,
> > + TYPE_MODE (sizetype), EXPAND_NORMAL);
> > + emit_block_move (ntgt, nsrc, size,
> > + (call_param_p
> > + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> > + }
> > + break;
> > + }
> > + return;
> > +}
> > +
> > /* Generate code for computing expression EXP,
> > and storing the value into TARGET.
> >
> > @@ -5713,9 +5788,29 @@ store_expr_with_bounds (tree exp, rtx target, int call_param_p,
> > emit_group_store (target, temp, TREE_TYPE (exp),
> > int_size_in_bytes (TREE_TYPE (exp)));
> > else if (GET_MODE (temp) == BLKmode)
> > - emit_block_move (target, temp, expr_size (exp),
> > - (call_param_p
> > - ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> > + {
> > + /* Copying smallish BLKmode structures with emit_block_move and thus
> > + by-pieces can result in store-to-load stalls. So copy some simple
> > + small aggregates element or field-wise. */
> > + if (GET_MODE (target) == BLKmode
> > + && AGGREGATE_TYPE_P (TREE_TYPE (exp))
> > + && !TREE_ADDRESSABLE (TREE_TYPE (exp))
> > + && tree_fits_shwi_p (TYPE_SIZE (TREE_TYPE (exp)))
> > + && (tree_to_shwi (TYPE_SIZE (TREE_TYPE (exp)))
> > + <= (PARAM_VALUE (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY)
> > + * BITS_PER_UNIT))
> > + && simple_mix_of_records_and_arrays_p (TREE_TYPE (exp), false))
> > + {
> > + /* FIXME: Can this happen? What would it mean? */
> > + gcc_assert (!reverse);
> > + emit_move_elementwise (TREE_TYPE (exp), target, temp, 0,
> > + call_param_p);
> > + }
> > + else
> > + emit_block_move (target, temp, expr_size (exp),
> > + (call_param_p
> > + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> > + }
> > /* If we emit a nontemporal store, there is nothing else to do. */
> > else if (nontemporal && emit_storent_insn (target, temp))
> > ;
> > diff --git a/gcc/ipa-cp.c b/gcc/ipa-cp.c
> > index 6b3d8d7364c..7d6019bbd30 100644
> > --- a/gcc/ipa-cp.c
> > +++ b/gcc/ipa-cp.c
> > @@ -124,6 +124,7 @@ along with GCC; see the file COPYING3. If not see
> > #include "tree-ssa-ccp.h"
> > #include "stringpool.h"
> > #include "attribs.h"
> > +#include "tree-sra.h"
> >
> > template <typename valtype> class ipcp_value;
> >
> > diff --git a/gcc/ipa-prop.h b/gcc/ipa-prop.h
> > index fa5bed49ee0..2313cc884ed 100644
> > --- a/gcc/ipa-prop.h
> > +++ b/gcc/ipa-prop.h
> > @@ -877,10 +877,6 @@ ipa_parm_adjustment *ipa_get_adjustment_candidate (tree **, bool *,
> > void ipa_release_body_info (struct ipa_func_body_info *);
> > tree ipa_get_callee_param_type (struct cgraph_edge *e, int i);
> >
> > -/* From tree-sra.c: */
> > -tree build_ref_for_offset (location_t, tree, HOST_WIDE_INT, bool, tree,
> > - gimple_stmt_iterator *, bool);
> > -
> > /* In ipa-cp.c */
> > void ipa_cp_c_finalize (void);
> >
> > diff --git a/gcc/params.def b/gcc/params.def
> > index e55afc28053..5e19f1414a0 100644
> > --- a/gcc/params.def
> > +++ b/gcc/params.def
> > @@ -1294,6 +1294,12 @@ DEFPARAM (PARAM_VECT_EPILOGUES_NOMASK,
> > "Enable loop epilogue vectorization using smaller vector size.",
> > 0, 0, 1)
> >
> > +DEFPARAM (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
> > + "max-size-for-elementwise-copy",
> > + "Maximum size in bytes of a structure or array to by considered for "
> > + "copying by its individual fields or elements",
> > + 0, 0, 512)
> > +
> > /*
> >
> > Local variables:
> > diff --git a/gcc/testsuite/gcc.target/i386/pr80689-1.c b/gcc/testsuite/gcc.target/i386/pr80689-1.c
> > new file mode 100644
> > index 00000000000..4156d4fba45
> > --- /dev/null
> > +++ b/gcc/testsuite/gcc.target/i386/pr80689-1.c
> > @@ -0,0 +1,38 @@
> > +/* { dg-do compile } */
> > +/* { dg-options "-O2" } */
> > +
> > +typedef struct st1
> > +{
> > + long unsigned int a,b;
> > + long int c,d;
> > +}R;
> > +
> > +typedef struct st2
> > +{
> > + int t;
> > + R reg;
> > +}N;
> > +
> > +void Set (const R *region, N *n_info );
> > +
> > +void test(N *n_obj ,const long unsigned int a, const long unsigned int b, const long int c,const long int d)
> > +{
> > + R reg;
> > +
> > + reg.a=a;
> > + reg.b=b;
> > + reg.c=c;
> > + reg.d=d;
> > + Set (®, n_obj);
> > +
> > +}
> > +
> > +void Set (const R *reg, N *n_obj )
> > +{
> > + n_obj->reg=(*reg);
> > +}
> > +
> > +
> > +/* { dg-final { scan-assembler-not "%(x|y|z)mm\[0-9\]+" } } */
> > +/* { dg-final { scan-assembler-not "movdqu" } } */
> > +/* { dg-final { scan-assembler-not "movups" } } */
> > diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
> > index bac593951e7..ade97964205 100644
> > --- a/gcc/tree-sra.c
> > +++ b/gcc/tree-sra.c
> > @@ -104,6 +104,7 @@ along with GCC; see the file COPYING3. If not see
> > #include "ipa-fnsummary.h"
> > #include "ipa-utils.h"
> > #include "builtins.h"
> > +#include "tree-sra.h"
> >
> > /* Enumeration of all aggregate reductions we can do. */
> > enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
> > @@ -952,14 +953,14 @@ create_access (tree expr, gimple *stmt, bool write)
> > }
> >
> >
> > -/* Return true iff TYPE is scalarizable - i.e. a RECORD_TYPE or fixed-length
> > - ARRAY_TYPE with fields that are either of gimple register types (excluding
> > - bit-fields) or (recursively) scalarizable types. CONST_DECL must be true if
> > - we are considering a decl from constant pool. If it is false, char arrays
> > - will be refused. */
> > +/* Return true if TYPE consists of RECORD_TYPE or fixed-length ARRAY_TYPE with
> > + fields/elements that are not bit-fields and are either register types or
> > + recursively comply with simple_mix_of_records_and_arrays_p. Furthermore, if
> > + ALLOW_CHAR_ARRAYS is false, the function will return false also if TYPE
> > + contains an array of elements that only have one byte. */
> >
> > -static bool
> > -scalarizable_type_p (tree type, bool const_decl)
> > +bool
> > +simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays)
> > {
> > gcc_assert (!is_gimple_reg_type (type));
> > if (type_contains_placeholder_p (type))
> > @@ -977,7 +978,7 @@ scalarizable_type_p (tree type, bool const_decl)
> > return false;
> >
> > if (!is_gimple_reg_type (ft)
> > - && !scalarizable_type_p (ft, const_decl))
> > + && !simple_mix_of_records_and_arrays_p (ft, allow_char_arrays))
> > return false;
> > }
> >
> > @@ -986,7 +987,7 @@ scalarizable_type_p (tree type, bool const_decl)
> > case ARRAY_TYPE:
> > {
> > HOST_WIDE_INT min_elem_size;
> > - if (const_decl)
> > + if (allow_char_arrays)
> > min_elem_size = 0;
> > else
> > min_elem_size = BITS_PER_UNIT;
> > @@ -1008,7 +1009,7 @@ scalarizable_type_p (tree type, bool const_decl)
> >
> > tree elem = TREE_TYPE (type);
> > if (!is_gimple_reg_type (elem)
> > - && !scalarizable_type_p (elem, const_decl))
> > + && !simple_mix_of_records_and_arrays_p (elem, allow_char_arrays))
> > return false;
> > return true;
> > }
> > @@ -1017,10 +1018,38 @@ scalarizable_type_p (tree type, bool const_decl)
> > }
> > }
> >
> > -static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree, tree);
> > +static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree,
> > + tree);
> > +
> > +/* For a given array TYPE, return false if its domain does not have any maximum
> > + value. Otherwise calculate MIN and MAX indices of the first and the last
> > + element. */
> > +
> > +bool
> > +extract_min_max_idx_from_array (tree type, offset_int *min, offset_int *max)
> > +{
> > + tree domain = TYPE_DOMAIN (type);
> > + tree minidx = TYPE_MIN_VALUE (domain);
> > + gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
> > + tree maxidx = TYPE_MAX_VALUE (domain);
> > + if (!maxidx)
> > + return false;
> > + gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
> > +
> > + /* MINIDX and MAXIDX are inclusive, and must be interpreted in
> > + DOMAIN (e.g. signed int, whereas min/max may be size_int). */
> > + *min = wi::to_offset (minidx);
> > + *max = wi::to_offset (maxidx);
> > + if (!TYPE_UNSIGNED (domain))
> > + {
> > + *min = wi::sext (*min, TYPE_PRECISION (domain));
> > + *max = wi::sext (*max, TYPE_PRECISION (domain));
> > + }
> > + return true;
> > +}
> >
> > /* Create total_scalarization accesses for all scalar fields of a member
> > - of type DECL_TYPE conforming to scalarizable_type_p. BASE
> > + of type DECL_TYPE conforming to simple_mix_of_records_and_arrays_p. BASE
> > must be the top-most VAR_DECL representing the variable; within that,
> > OFFSET locates the member and REF must be the memory reference expression for
> > the member. */
> > @@ -1047,27 +1076,14 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
> > {
> > tree elemtype = TREE_TYPE (decl_type);
> > tree elem_size = TYPE_SIZE (elemtype);
> > - gcc_assert (elem_size && tree_fits_shwi_p (elem_size));
> > HOST_WIDE_INT el_size = tree_to_shwi (elem_size);
> > gcc_assert (el_size > 0);
> >
> > - tree minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (decl_type));
> > - gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
> > - tree maxidx = TYPE_MAX_VALUE (TYPE_DOMAIN (decl_type));
> > + offset_int idx, max;
> > /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
> > - if (maxidx)
> > + if (extract_min_max_idx_from_array (decl_type, &idx, &max))
> > {
> > - gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
> > tree domain = TYPE_DOMAIN (decl_type);
> > - /* MINIDX and MAXIDX are inclusive, and must be interpreted in
> > - DOMAIN (e.g. signed int, whereas min/max may be size_int). */
> > - offset_int idx = wi::to_offset (minidx);
> > - offset_int max = wi::to_offset (maxidx);
> > - if (!TYPE_UNSIGNED (domain))
> > - {
> > - idx = wi::sext (idx, TYPE_PRECISION (domain));
> > - max = wi::sext (max, TYPE_PRECISION (domain));
> > - }
> > for (int el_off = offset; idx <= max; ++idx)
> > {
> > tree nref = build4 (ARRAY_REF, elemtype,
> > @@ -1088,10 +1104,10 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
> > }
> >
> > /* Create total_scalarization accesses for a member of type TYPE, which must
> > - satisfy either is_gimple_reg_type or scalarizable_type_p. BASE must be the
> > - top-most VAR_DECL representing the variable; within that, POS and SIZE locate
> > - the member, REVERSE gives its torage order. and REF must be the reference
> > - expression for it. */
> > + satisfy either is_gimple_reg_type or simple_mix_of_records_and_arrays_p.
> > + BASE must be the top-most VAR_DECL representing the variable; within that,
> > + POS and SIZE locate the member, REVERSE gives its torage order. and REF must
> > + be the reference expression for it. */
> >
> > static void
> > scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
> > @@ -1111,7 +1127,8 @@ scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
> > }
> >
> > /* Create a total_scalarization access for VAR as a whole. VAR must be of a
> > - RECORD_TYPE or ARRAY_TYPE conforming to scalarizable_type_p. */
> > + RECORD_TYPE or ARRAY_TYPE conforming to
> > + simple_mix_of_records_and_arrays_p. */
> >
> > static void
> > create_total_scalarization_access (tree var)
> > @@ -2803,8 +2820,9 @@ analyze_all_variable_accesses (void)
> > {
> > tree var = candidate (i);
> >
> > - if (VAR_P (var) && scalarizable_type_p (TREE_TYPE (var),
> > - constant_decl_p (var)))
> > + if (VAR_P (var)
> > + && simple_mix_of_records_and_arrays_p (TREE_TYPE (var),
> > + constant_decl_p (var)))
> > {
> > if (tree_to_uhwi (TYPE_SIZE (TREE_TYPE (var)))
> > <= max_scalarization_size)
> > diff --git a/gcc/tree-sra.h b/gcc/tree-sra.h
> > new file mode 100644
> > index 00000000000..dc901385994
> > --- /dev/null
> > +++ b/gcc/tree-sra.h
> > @@ -0,0 +1,33 @@
> > +/* tree-sra.h - Run-time parameters.
> > + Copyright (C) 2017 Free Software Foundation, Inc.
> > +
> > +This file is part of GCC.
> > +
> > +GCC is free software; you can redistribute it and/or modify it under
> > +the terms of the GNU General Public License as published by the Free
> > +Software Foundation; either version 3, or (at your option) any later
> > +version.
> > +
> > +GCC is distributed in the hope that it will be useful, but WITHOUT ANY
> > +WARRANTY; without even the implied warranty of MERCHANTABILITY or
> > +FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
> > +for more details.
> > +
> > +You should have received a copy of the GNU General Public License
> > +along with GCC; see the file COPYING3. If not see
> > +<http://www.gnu.org/licenses/>. */
> > +
> > +#ifndef TREE_SRA_H
> > +#define TREE_SRA_H
> > +
> > +
> > +bool simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays);
> > +bool extract_min_max_idx_from_array (tree type, offset_int *idx,
> > + offset_int *max);
> > +tree build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
> > + bool reverse, tree exp_type,
> > + gimple_stmt_iterator *gsi, bool insert_after);
> > +
> > +
> > +
> > +#endif /* TREE_SRA_H */
> > --
> > 2.14.1
> >
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-10-26 12:23 ` Martin Jambor
@ 2017-10-26 12:44 ` Richard Biener
2017-10-26 12:57 ` Jan Hubicka
2017-10-26 14:18 ` Michael Matz
1 sibling, 1 reply; 26+ messages in thread
From: Richard Biener @ 2017-10-26 12:44 UTC (permalink / raw)
To: Richard Biener, GCC Patches, Jan Hubicka
On Thu, Oct 26, 2017 at 2:18 PM, Martin Jambor <mjambor@suse.cz> wrote:
> Hi,
>
> On Tue, Oct 17, 2017 at 01:34:54PM +0200, Richard Biener wrote:
>> On Fri, Oct 13, 2017 at 6:13 PM, Martin Jambor <mjambor@suse.cz> wrote:
>> > Hi,
>> >
>> > I'd like to request comments to the patch below which aims to fix PR
>> > 80689, which is an instance of a store-to-load forwarding stall on
>> > x86_64 CPUs in the Image Magick benchmark, which is responsible for a
>> > slow down of up to 9% compared to gcc 6, depending on options and HW
>> > used. (Actually, I have just seen 24% in one specific combination but
>> > for various reasons can no longer verify it today.)
>> >
>> > The revision causing the regression is 237074, which increased the
>> > size of the mode for copying aggregates "by pieces" to 128 bits,
>> > incurring big stalls when the values being copied are also still being
>> > stored in a smaller data type or if the copied values are loaded in a
>> > smaller types shortly afterwards. Such situations happen in Image
>> > Magick even across calls, which means that any non-IPA flow-sensitive
>> > approach would not detect them. Therefore, the patch simply changes
>> > the way we copy small BLKmode data that are simple combinations of
>> > records and arrays (meaning no unions, bit-fields but also character
>> > arrays are disallowed) and simply copies them one field and/or element
>> > at a time.
>> >
>> > "Small" in this RFC patch means up to 35 bytes on x86_64 and i386 CPUs
>> > (the structure in the benchmark has 32 bytes) but is subject to change
>> > after more benchmarking and is actually zero - meaning element copying
>> > never happens - on other architectures. I believe that any
>> > architecture with a store buffer can benefit but it's probably better
>> > to leave it to their maintainers to find a different default value. I
>> > am not sure this is how such HW-dependant decisions should be done and
>> > is the primary reason, why I am sending this RFC first.
>> >
>> > I have decided to implement this change at the expansion level because
>> > at that point the type information is still readily available and at
>> > the same time we can also handle various implicit copies, for example
>> > those passing a parameter. I found I could re-use some bits and
>> > pieces of tree-SRA and so I did, creating tree-sra.h header file in
>> > the process.
>> >
>> > I am fully aware that in the final patch the new parameter, or indeed
>> > any new parameters, need to be documented. I have skipped that
>> > intentionally now and will write the documentation if feedback here is
>> > generally good.
>> >
>> > I have bootstrapped and tested this patch on x86_64-linux, with
>> > different values of the parameter and only found problems with
>> > unreasonably high values leading to OOM. I have done the same with a
>> > previous version of the patch which was equivalent to the limit being
>> > 64 bytes on aarch64-linux, ppc64le-linux and ia64-linux and only ran
>> > into failures of tests which assumed that structure padding was copied
>> > in aggregate copies (mostly gcc.target/aarch64/aapcs64/ stuff but also
>> > for example gcc.dg/vmx/varargs-4.c).
>> >
>> > The patch decreases the SPEC 2017 "rate" run-time of imagick by 9% and
>> > 8% at -O2 and -Ofast compilation levels respectively on one particular
>> > new AMD CPU and by 6% and 3% on one particular old Intel machine.
>> >
>> > Thanks in advance for any comments,
>>
>> I wonder if you can at the place you choose to hook this in elide any
>> copying of padding between fields.
>>
>> I'd rather have hooked such "high level" optimization in
>> expand_assignment where you can be reasonably sure you're seeing an
>> actual source-level construct.
>
> I have discussed this with Honza and we eventually decided to make the
> elememnt-wise copy an alternative to emit_block_move (which uses the
> larger mode for moving since GCC 7) exactly so that we handle not only
> source-level assignments but also passing parameters by value and
> other situations.
>
>>
>> 35 bytes seems to be much - what is the code-size impact?
>
> I will find out and report on that. I need at least 32 bytes (four
> long ints) to fix imagemagick, where the problematic structure is:
>
> typedef struct _RectangleInfo
> {
> size_t
> width,
> height;
>
> ssize_t
> x,
> y;
> } RectangleInfo;
>
> ...so 8 longs, no padding. Since any aggregate having between 33-35
> bytes needs to consist of smaller fields/elements, it seemed
> reasonable to also copy them element-wise.
>
> Nevertheless, I still intend to experiment with the limit, I sent out
> this RFC exactly so that I don't spend a lot of time benchmarking
> something that is eventually not deemed acceptable on principle.
I think the limit should be on the number of generated copies and not
the overall size of the structure... If the struct were composed of
32 individual chars we wouldn't want to emit 32 loads and 32 stores...
I wonder how rep; movb; interacts with store to load forwarding? Is
that maybe optimized well on some archs? movb should always
forward and wasn't the setup cost for small N reasonable on modern
CPUs?
>>
>> IIRC the reason this may be slow isn't loading in smaller types than stored
>> before by the copy - the store buffers can handle this reasonably well. It's
>> solely when previous smaller stores are
>>
>> a1) not mergeabe in the store buffer
>> a2) not merged because earlier stores are already committed
>>
>> and
>>
>> b) loaded afterwards as a type that would access multiple store buffers
>>
>> a) would be sure to happen in case b) involves accessing padding. Is the
>> Image Magick case one that involves padding in the structure?
>
> As I said above, there is no padding.
>
> Basically, what happens is that in a number of places, there is a
> variable region of the aforementioned type and it is initialized and
> passed to function SetPixelCacheNexusPixels in the following manner:
>
> ...
> region.width=cache_info->columns;
> region.height=1;
> region.x=0;
> region.y=y;
> pixels=SetPixelCacheNexusPixels(cache_info,ReadMode,®ion,MagickTrue,
> cache_nexus[id],exception);
> ...
>
> and the first four statements in SetPixelCacheNexusPixels are:
>
> assert(cache_info != (const CacheInfo *) NULL);
> assert(cache_info->signature == MagickSignature);
> if (cache_info->type == UndefinedCache)
> return((PixelPacket *) NULL);
> nexus_info->region=(*region);
>
> with the last one generating the stalls, on both Zen-based machines
> and also on 2-3 years old Intel CPUs.
>
> I have had a look at what Agner Fog's micro-architecture document says
> about store forwarding stalls and:
>
> - on Broadwells and Haswells, any "write of any size is followed by
> a read of a larger size" ioncurs a stall, which fits our example,
> - on Skylakes: "A read that is bigger than the write, or a read that
> covers both written and unwritten bytes, takes approximately 11
> clock cycles extra" seems to apply
> - on Intel silvermont, there will also be a stall because "A memory
> write can be forwarded to a subsequent read of the same size or a
> smaller size..."
> - on Zens, Agner Fog says they work perfectly except when crossing a
> page or when "A read that has a partial overlap with a preceding
> write has a penalty of 6-7 clock cycles," which must be why I see
> stalls.
>
> So I guess the pending stores are not really merged even without
> padding,
It probably depends on the width of the entries in the store buffer,
if they appear in-order and the alignment of the stores (if they are larger than
8 bytes they are surely aligned). IIRC CPUs had smaller store buffer
entries than cache line size.
Given that load bandwith is usually higher than store bandwith it
might make sense to do the store combining in our copying sequence,
like for the 8 byte entry case use sth like
movq 0(%eax), %xmm0
movhps 8(%eax), %xmm0 // or vpinsert
mov[au]ps %xmm0, 0%(ebx)
...
thus do two loads per store and perform the stores in wider
mode?
As said a general concern was you not copying padding. If you
put this into an even more common place you surely will break
stuff, no?
Richard.
>
> Martin
>
>
>>
>> Richard.
>>
>> > Martin
>> >
>> >
>> > 2017-10-12 Martin Jambor <mjambor@suse.cz>
>> >
>> > PR target/80689
>> > * tree-sra.h: New file.
>> > * ipa-prop.h: Moved declaration of build_ref_for_offset to
>> > tree-sra.h.
>> > * expr.c: Include params.h and tree-sra.h.
>> > (emit_move_elementwise): New function.
>> > (store_expr_with_bounds): Optionally use it.
>> > * ipa-cp.c: Include tree-sra.h.
>> > * params.def (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY): New.
>> > * config/i386/i386.c (ix86_option_override_internal): Set
>> > PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY to 35.
>> > * tree-sra.c: Include tree-sra.h.
>> > (scalarizable_type_p): Renamed to
>> > simple_mix_of_records_and_arrays_p, made public, renamed the
>> > second parameter to allow_char_arrays.
>> > (extract_min_max_idx_from_array): New function.
>> > (completely_scalarize): Moved bits of the function to
>> > extract_min_max_idx_from_array.
>> >
>> > testsuite/
>> > * gcc.target/i386/pr80689-1.c: New test.
>> > ---
>> > gcc/config/i386/i386.c | 4 ++
>> > gcc/expr.c | 103 ++++++++++++++++++++++++++++--
>> > gcc/ipa-cp.c | 1 +
>> > gcc/ipa-prop.h | 4 --
>> > gcc/params.def | 6 ++
>> > gcc/testsuite/gcc.target/i386/pr80689-1.c | 38 +++++++++++
>> > gcc/tree-sra.c | 86 +++++++++++++++----------
>> > gcc/tree-sra.h | 33 ++++++++++
>> > 8 files changed, 233 insertions(+), 42 deletions(-)
>> > create mode 100644 gcc/testsuite/gcc.target/i386/pr80689-1.c
>> > create mode 100644 gcc/tree-sra.h
>> >
>> > diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c
>> > index 1ee8351c21f..87f602e7ead 100644
>> > --- a/gcc/config/i386/i386.c
>> > +++ b/gcc/config/i386/i386.c
>> > @@ -6511,6 +6511,10 @@ ix86_option_override_internal (bool main_args_p,
>> > ix86_tune_cost->l2_cache_size,
>> > opts->x_param_values,
>> > opts_set->x_param_values);
>> > + maybe_set_param_value (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
>> > + 35,
>> > + opts->x_param_values,
>> > + opts_set->x_param_values);
>> >
>> > /* Enable sw prefetching at -O3 for CPUS that prefetching is helpful. */
>> > if (opts->x_flag_prefetch_loop_arrays < 0
>> > diff --git a/gcc/expr.c b/gcc/expr.c
>> > index 134ee731c29..dff24e7f166 100644
>> > --- a/gcc/expr.c
>> > +++ b/gcc/expr.c
>> > @@ -61,7 +61,8 @@ along with GCC; see the file COPYING3. If not see
>> > #include "tree-chkp.h"
>> > #include "rtl-chkp.h"
>> > #include "ccmp.h"
>> > -
>> > +#include "params.h"
>> > +#include "tree-sra.h"
>> >
>> > /* If this is nonzero, we do not bother generating VOLATILE
>> > around volatile memory references, and we are willing to
>> > @@ -5340,6 +5341,80 @@ emit_storent_insn (rtx to, rtx from)
>> > return maybe_expand_insn (code, 2, ops);
>> > }
>> >
>> > +/* Generate code for copying data of type TYPE at SOURCE plus OFFSET to TARGET
>> > + plus OFFSET, but do so element-wise and/or field-wise for each record and
>> > + array within TYPE. TYPE must either be a register type or an aggregate
>> > + complying with scalarizable_type_p.
>> > +
>> > + If CALL_PARAM_P is nonzero, this is a store into a call param on the
>> > + stack, and block moves may need to be treated specially. */
>> > +
>> > +static void
>> > +emit_move_elementwise (tree type, rtx target, rtx source, HOST_WIDE_INT offset,
>> > + int call_param_p)
>> > +{
>> > + switch (TREE_CODE (type))
>> > + {
>> > + case RECORD_TYPE:
>> > + for (tree fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
>> > + if (TREE_CODE (fld) == FIELD_DECL)
>> > + {
>> > + HOST_WIDE_INT fld_offset = offset + int_bit_position (fld);
>> > + tree ft = TREE_TYPE (fld);
>> > + emit_move_elementwise (ft, target, source, fld_offset,
>> > + call_param_p);
>> > + }
>> > + break;
>> > +
>> > + case ARRAY_TYPE:
>> > + {
>> > + tree elem_type = TREE_TYPE (type);
>> > + HOST_WIDE_INT el_size = tree_to_shwi (TYPE_SIZE (elem_type));
>> > + gcc_assert (el_size > 0);
>> > +
>> > + offset_int idx, max;
>> > + /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
>> > + if (extract_min_max_idx_from_array (type, &idx, &max))
>> > + {
>> > + HOST_WIDE_INT el_offset = offset;
>> > + for (; idx <= max; ++idx)
>> > + {
>> > + emit_move_elementwise (elem_type, target, source, el_offset,
>> > + call_param_p);
>> > + el_offset += el_size;
>> > + }
>> > + }
>> > + }
>> > + break;
>> > + default:
>> > + machine_mode mode = TYPE_MODE (type);
>> > +
>> > + rtx ntgt = adjust_address (target, mode, offset / BITS_PER_UNIT);
>> > + rtx nsrc = adjust_address (source, mode, offset / BITS_PER_UNIT);
>> > +
>> > + /* TODO: Figure out whether the following is actually necessary. */
>> > + if (target == ntgt)
>> > + ntgt = copy_rtx (target);
>> > + if (source == nsrc)
>> > + nsrc = copy_rtx (source);
>> > +
>> > + gcc_assert (mode != VOIDmode);
>> > + if (mode != BLKmode)
>> > + emit_move_insn (ntgt, nsrc);
>> > + else
>> > + {
>> > + /* For example vector gimple registers can end up here. */
>> > + rtx size = expand_expr (TYPE_SIZE_UNIT (type), NULL_RTX,
>> > + TYPE_MODE (sizetype), EXPAND_NORMAL);
>> > + emit_block_move (ntgt, nsrc, size,
>> > + (call_param_p
>> > + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
>> > + }
>> > + break;
>> > + }
>> > + return;
>> > +}
>> > +
>> > /* Generate code for computing expression EXP,
>> > and storing the value into TARGET.
>> >
>> > @@ -5713,9 +5788,29 @@ store_expr_with_bounds (tree exp, rtx target, int call_param_p,
>> > emit_group_store (target, temp, TREE_TYPE (exp),
>> > int_size_in_bytes (TREE_TYPE (exp)));
>> > else if (GET_MODE (temp) == BLKmode)
>> > - emit_block_move (target, temp, expr_size (exp),
>> > - (call_param_p
>> > - ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
>> > + {
>> > + /* Copying smallish BLKmode structures with emit_block_move and thus
>> > + by-pieces can result in store-to-load stalls. So copy some simple
>> > + small aggregates element or field-wise. */
>> > + if (GET_MODE (target) == BLKmode
>> > + && AGGREGATE_TYPE_P (TREE_TYPE (exp))
>> > + && !TREE_ADDRESSABLE (TREE_TYPE (exp))
>> > + && tree_fits_shwi_p (TYPE_SIZE (TREE_TYPE (exp)))
>> > + && (tree_to_shwi (TYPE_SIZE (TREE_TYPE (exp)))
>> > + <= (PARAM_VALUE (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY)
>> > + * BITS_PER_UNIT))
>> > + && simple_mix_of_records_and_arrays_p (TREE_TYPE (exp), false))
>> > + {
>> > + /* FIXME: Can this happen? What would it mean? */
>> > + gcc_assert (!reverse);
>> > + emit_move_elementwise (TREE_TYPE (exp), target, temp, 0,
>> > + call_param_p);
>> > + }
>> > + else
>> > + emit_block_move (target, temp, expr_size (exp),
>> > + (call_param_p
>> > + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
>> > + }
>> > /* If we emit a nontemporal store, there is nothing else to do. */
>> > else if (nontemporal && emit_storent_insn (target, temp))
>> > ;
>> > diff --git a/gcc/ipa-cp.c b/gcc/ipa-cp.c
>> > index 6b3d8d7364c..7d6019bbd30 100644
>> > --- a/gcc/ipa-cp.c
>> > +++ b/gcc/ipa-cp.c
>> > @@ -124,6 +124,7 @@ along with GCC; see the file COPYING3. If not see
>> > #include "tree-ssa-ccp.h"
>> > #include "stringpool.h"
>> > #include "attribs.h"
>> > +#include "tree-sra.h"
>> >
>> > template <typename valtype> class ipcp_value;
>> >
>> > diff --git a/gcc/ipa-prop.h b/gcc/ipa-prop.h
>> > index fa5bed49ee0..2313cc884ed 100644
>> > --- a/gcc/ipa-prop.h
>> > +++ b/gcc/ipa-prop.h
>> > @@ -877,10 +877,6 @@ ipa_parm_adjustment *ipa_get_adjustment_candidate (tree **, bool *,
>> > void ipa_release_body_info (struct ipa_func_body_info *);
>> > tree ipa_get_callee_param_type (struct cgraph_edge *e, int i);
>> >
>> > -/* From tree-sra.c: */
>> > -tree build_ref_for_offset (location_t, tree, HOST_WIDE_INT, bool, tree,
>> > - gimple_stmt_iterator *, bool);
>> > -
>> > /* In ipa-cp.c */
>> > void ipa_cp_c_finalize (void);
>> >
>> > diff --git a/gcc/params.def b/gcc/params.def
>> > index e55afc28053..5e19f1414a0 100644
>> > --- a/gcc/params.def
>> > +++ b/gcc/params.def
>> > @@ -1294,6 +1294,12 @@ DEFPARAM (PARAM_VECT_EPILOGUES_NOMASK,
>> > "Enable loop epilogue vectorization using smaller vector size.",
>> > 0, 0, 1)
>> >
>> > +DEFPARAM (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
>> > + "max-size-for-elementwise-copy",
>> > + "Maximum size in bytes of a structure or array to by considered for "
>> > + "copying by its individual fields or elements",
>> > + 0, 0, 512)
>> > +
>> > /*
>> >
>> > Local variables:
>> > diff --git a/gcc/testsuite/gcc.target/i386/pr80689-1.c b/gcc/testsuite/gcc.target/i386/pr80689-1.c
>> > new file mode 100644
>> > index 00000000000..4156d4fba45
>> > --- /dev/null
>> > +++ b/gcc/testsuite/gcc.target/i386/pr80689-1.c
>> > @@ -0,0 +1,38 @@
>> > +/* { dg-do compile } */
>> > +/* { dg-options "-O2" } */
>> > +
>> > +typedef struct st1
>> > +{
>> > + long unsigned int a,b;
>> > + long int c,d;
>> > +}R;
>> > +
>> > +typedef struct st2
>> > +{
>> > + int t;
>> > + R reg;
>> > +}N;
>> > +
>> > +void Set (const R *region, N *n_info );
>> > +
>> > +void test(N *n_obj ,const long unsigned int a, const long unsigned int b, const long int c,const long int d)
>> > +{
>> > + R reg;
>> > +
>> > + reg.a=a;
>> > + reg.b=b;
>> > + reg.c=c;
>> > + reg.d=d;
>> > + Set (®, n_obj);
>> > +
>> > +}
>> > +
>> > +void Set (const R *reg, N *n_obj )
>> > +{
>> > + n_obj->reg=(*reg);
>> > +}
>> > +
>> > +
>> > +/* { dg-final { scan-assembler-not "%(x|y|z)mm\[0-9\]+" } } */
>> > +/* { dg-final { scan-assembler-not "movdqu" } } */
>> > +/* { dg-final { scan-assembler-not "movups" } } */
>> > diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
>> > index bac593951e7..ade97964205 100644
>> > --- a/gcc/tree-sra.c
>> > +++ b/gcc/tree-sra.c
>> > @@ -104,6 +104,7 @@ along with GCC; see the file COPYING3. If not see
>> > #include "ipa-fnsummary.h"
>> > #include "ipa-utils.h"
>> > #include "builtins.h"
>> > +#include "tree-sra.h"
>> >
>> > /* Enumeration of all aggregate reductions we can do. */
>> > enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
>> > @@ -952,14 +953,14 @@ create_access (tree expr, gimple *stmt, bool write)
>> > }
>> >
>> >
>> > -/* Return true iff TYPE is scalarizable - i.e. a RECORD_TYPE or fixed-length
>> > - ARRAY_TYPE with fields that are either of gimple register types (excluding
>> > - bit-fields) or (recursively) scalarizable types. CONST_DECL must be true if
>> > - we are considering a decl from constant pool. If it is false, char arrays
>> > - will be refused. */
>> > +/* Return true if TYPE consists of RECORD_TYPE or fixed-length ARRAY_TYPE with
>> > + fields/elements that are not bit-fields and are either register types or
>> > + recursively comply with simple_mix_of_records_and_arrays_p. Furthermore, if
>> > + ALLOW_CHAR_ARRAYS is false, the function will return false also if TYPE
>> > + contains an array of elements that only have one byte. */
>> >
>> > -static bool
>> > -scalarizable_type_p (tree type, bool const_decl)
>> > +bool
>> > +simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays)
>> > {
>> > gcc_assert (!is_gimple_reg_type (type));
>> > if (type_contains_placeholder_p (type))
>> > @@ -977,7 +978,7 @@ scalarizable_type_p (tree type, bool const_decl)
>> > return false;
>> >
>> > if (!is_gimple_reg_type (ft)
>> > - && !scalarizable_type_p (ft, const_decl))
>> > + && !simple_mix_of_records_and_arrays_p (ft, allow_char_arrays))
>> > return false;
>> > }
>> >
>> > @@ -986,7 +987,7 @@ scalarizable_type_p (tree type, bool const_decl)
>> > case ARRAY_TYPE:
>> > {
>> > HOST_WIDE_INT min_elem_size;
>> > - if (const_decl)
>> > + if (allow_char_arrays)
>> > min_elem_size = 0;
>> > else
>> > min_elem_size = BITS_PER_UNIT;
>> > @@ -1008,7 +1009,7 @@ scalarizable_type_p (tree type, bool const_decl)
>> >
>> > tree elem = TREE_TYPE (type);
>> > if (!is_gimple_reg_type (elem)
>> > - && !scalarizable_type_p (elem, const_decl))
>> > + && !simple_mix_of_records_and_arrays_p (elem, allow_char_arrays))
>> > return false;
>> > return true;
>> > }
>> > @@ -1017,10 +1018,38 @@ scalarizable_type_p (tree type, bool const_decl)
>> > }
>> > }
>> >
>> > -static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree, tree);
>> > +static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree,
>> > + tree);
>> > +
>> > +/* For a given array TYPE, return false if its domain does not have any maximum
>> > + value. Otherwise calculate MIN and MAX indices of the first and the last
>> > + element. */
>> > +
>> > +bool
>> > +extract_min_max_idx_from_array (tree type, offset_int *min, offset_int *max)
>> > +{
>> > + tree domain = TYPE_DOMAIN (type);
>> > + tree minidx = TYPE_MIN_VALUE (domain);
>> > + gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
>> > + tree maxidx = TYPE_MAX_VALUE (domain);
>> > + if (!maxidx)
>> > + return false;
>> > + gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
>> > +
>> > + /* MINIDX and MAXIDX are inclusive, and must be interpreted in
>> > + DOMAIN (e.g. signed int, whereas min/max may be size_int). */
>> > + *min = wi::to_offset (minidx);
>> > + *max = wi::to_offset (maxidx);
>> > + if (!TYPE_UNSIGNED (domain))
>> > + {
>> > + *min = wi::sext (*min, TYPE_PRECISION (domain));
>> > + *max = wi::sext (*max, TYPE_PRECISION (domain));
>> > + }
>> > + return true;
>> > +}
>> >
>> > /* Create total_scalarization accesses for all scalar fields of a member
>> > - of type DECL_TYPE conforming to scalarizable_type_p. BASE
>> > + of type DECL_TYPE conforming to simple_mix_of_records_and_arrays_p. BASE
>> > must be the top-most VAR_DECL representing the variable; within that,
>> > OFFSET locates the member and REF must be the memory reference expression for
>> > the member. */
>> > @@ -1047,27 +1076,14 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
>> > {
>> > tree elemtype = TREE_TYPE (decl_type);
>> > tree elem_size = TYPE_SIZE (elemtype);
>> > - gcc_assert (elem_size && tree_fits_shwi_p (elem_size));
>> > HOST_WIDE_INT el_size = tree_to_shwi (elem_size);
>> > gcc_assert (el_size > 0);
>> >
>> > - tree minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (decl_type));
>> > - gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
>> > - tree maxidx = TYPE_MAX_VALUE (TYPE_DOMAIN (decl_type));
>> > + offset_int idx, max;
>> > /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
>> > - if (maxidx)
>> > + if (extract_min_max_idx_from_array (decl_type, &idx, &max))
>> > {
>> > - gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
>> > tree domain = TYPE_DOMAIN (decl_type);
>> > - /* MINIDX and MAXIDX are inclusive, and must be interpreted in
>> > - DOMAIN (e.g. signed int, whereas min/max may be size_int). */
>> > - offset_int idx = wi::to_offset (minidx);
>> > - offset_int max = wi::to_offset (maxidx);
>> > - if (!TYPE_UNSIGNED (domain))
>> > - {
>> > - idx = wi::sext (idx, TYPE_PRECISION (domain));
>> > - max = wi::sext (max, TYPE_PRECISION (domain));
>> > - }
>> > for (int el_off = offset; idx <= max; ++idx)
>> > {
>> > tree nref = build4 (ARRAY_REF, elemtype,
>> > @@ -1088,10 +1104,10 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
>> > }
>> >
>> > /* Create total_scalarization accesses for a member of type TYPE, which must
>> > - satisfy either is_gimple_reg_type or scalarizable_type_p. BASE must be the
>> > - top-most VAR_DECL representing the variable; within that, POS and SIZE locate
>> > - the member, REVERSE gives its torage order. and REF must be the reference
>> > - expression for it. */
>> > + satisfy either is_gimple_reg_type or simple_mix_of_records_and_arrays_p.
>> > + BASE must be the top-most VAR_DECL representing the variable; within that,
>> > + POS and SIZE locate the member, REVERSE gives its torage order. and REF must
>> > + be the reference expression for it. */
>> >
>> > static void
>> > scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
>> > @@ -1111,7 +1127,8 @@ scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
>> > }
>> >
>> > /* Create a total_scalarization access for VAR as a whole. VAR must be of a
>> > - RECORD_TYPE or ARRAY_TYPE conforming to scalarizable_type_p. */
>> > + RECORD_TYPE or ARRAY_TYPE conforming to
>> > + simple_mix_of_records_and_arrays_p. */
>> >
>> > static void
>> > create_total_scalarization_access (tree var)
>> > @@ -2803,8 +2820,9 @@ analyze_all_variable_accesses (void)
>> > {
>> > tree var = candidate (i);
>> >
>> > - if (VAR_P (var) && scalarizable_type_p (TREE_TYPE (var),
>> > - constant_decl_p (var)))
>> > + if (VAR_P (var)
>> > + && simple_mix_of_records_and_arrays_p (TREE_TYPE (var),
>> > + constant_decl_p (var)))
>> > {
>> > if (tree_to_uhwi (TYPE_SIZE (TREE_TYPE (var)))
>> > <= max_scalarization_size)
>> > diff --git a/gcc/tree-sra.h b/gcc/tree-sra.h
>> > new file mode 100644
>> > index 00000000000..dc901385994
>> > --- /dev/null
>> > +++ b/gcc/tree-sra.h
>> > @@ -0,0 +1,33 @@
>> > +/* tree-sra.h - Run-time parameters.
>> > + Copyright (C) 2017 Free Software Foundation, Inc.
>> > +
>> > +This file is part of GCC.
>> > +
>> > +GCC is free software; you can redistribute it and/or modify it under
>> > +the terms of the GNU General Public License as published by the Free
>> > +Software Foundation; either version 3, or (at your option) any later
>> > +version.
>> > +
>> > +GCC is distributed in the hope that it will be useful, but WITHOUT ANY
>> > +WARRANTY; without even the implied warranty of MERCHANTABILITY or
>> > +FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
>> > +for more details.
>> > +
>> > +You should have received a copy of the GNU General Public License
>> > +along with GCC; see the file COPYING3. If not see
>> > +<http://www.gnu.org/licenses/>. */
>> > +
>> > +#ifndef TREE_SRA_H
>> > +#define TREE_SRA_H
>> > +
>> > +
>> > +bool simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays);
>> > +bool extract_min_max_idx_from_array (tree type, offset_int *idx,
>> > + offset_int *max);
>> > +tree build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
>> > + bool reverse, tree exp_type,
>> > + gimple_stmt_iterator *gsi, bool insert_after);
>> > +
>> > +
>> > +
>> > +#endif /* TREE_SRA_H */
>> > --
>> > 2.14.1
>> >
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-10-26 12:44 ` Richard Biener
@ 2017-10-26 12:57 ` Jan Hubicka
2017-10-26 14:53 ` Richard Biener
0 siblings, 1 reply; 26+ messages in thread
From: Jan Hubicka @ 2017-10-26 12:57 UTC (permalink / raw)
To: Richard Biener; +Cc: GCC Patches
> I think the limit should be on the number of generated copies and not
> the overall size of the structure... If the struct were composed of
> 32 individual chars we wouldn't want to emit 32 loads and 32 stores...
>
> I wonder how rep; movb; interacts with store to load forwarding? Is
> that maybe optimized well on some archs? movb should always
> forward and wasn't the setup cost for small N reasonable on modern
> CPUs?
rep mov is win over loop for blocks over 128bytes on core, for blocks in rage
24-128 on zen. This is w/o store/load forwarding, but I doubt those provide
a cheap way around.
>
> It probably depends on the width of the entries in the store buffer,
> if they appear in-order and the alignment of the stores (if they are larger than
> 8 bytes they are surely aligned). IIRC CPUs had smaller store buffer
> entries than cache line size.
>
> Given that load bandwith is usually higher than store bandwith it
> might make sense to do the store combining in our copying sequence,
> like for the 8 byte entry case use sth like
>
> movq 0(%eax), %xmm0
> movhps 8(%eax), %xmm0 // or vpinsert
> mov[au]ps %xmm0, 0%(ebx)
> ...
>
> thus do two loads per store and perform the stores in wider
> mode?
This may be somewhat faster indeed. I am not sure if store to load
forwarding will work for the later half when read again by halves.
It would not happen on older CPUs :)
Honza
>
> As said a general concern was you not copying padding. If you
> put this into an even more common place you surely will break
> stuff, no?
>
> Richard.
>
> >
> > Martin
> >
> >
> >>
> >> Richard.
> >>
> >> > Martin
> >> >
> >> >
> >> > 2017-10-12 Martin Jambor <mjambor@suse.cz>
> >> >
> >> > PR target/80689
> >> > * tree-sra.h: New file.
> >> > * ipa-prop.h: Moved declaration of build_ref_for_offset to
> >> > tree-sra.h.
> >> > * expr.c: Include params.h and tree-sra.h.
> >> > (emit_move_elementwise): New function.
> >> > (store_expr_with_bounds): Optionally use it.
> >> > * ipa-cp.c: Include tree-sra.h.
> >> > * params.def (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY): New.
> >> > * config/i386/i386.c (ix86_option_override_internal): Set
> >> > PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY to 35.
> >> > * tree-sra.c: Include tree-sra.h.
> >> > (scalarizable_type_p): Renamed to
> >> > simple_mix_of_records_and_arrays_p, made public, renamed the
> >> > second parameter to allow_char_arrays.
> >> > (extract_min_max_idx_from_array): New function.
> >> > (completely_scalarize): Moved bits of the function to
> >> > extract_min_max_idx_from_array.
> >> >
> >> > testsuite/
> >> > * gcc.target/i386/pr80689-1.c: New test.
> >> > ---
> >> > gcc/config/i386/i386.c | 4 ++
> >> > gcc/expr.c | 103 ++++++++++++++++++++++++++++--
> >> > gcc/ipa-cp.c | 1 +
> >> > gcc/ipa-prop.h | 4 --
> >> > gcc/params.def | 6 ++
> >> > gcc/testsuite/gcc.target/i386/pr80689-1.c | 38 +++++++++++
> >> > gcc/tree-sra.c | 86 +++++++++++++++----------
> >> > gcc/tree-sra.h | 33 ++++++++++
> >> > 8 files changed, 233 insertions(+), 42 deletions(-)
> >> > create mode 100644 gcc/testsuite/gcc.target/i386/pr80689-1.c
> >> > create mode 100644 gcc/tree-sra.h
> >> >
> >> > diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c
> >> > index 1ee8351c21f..87f602e7ead 100644
> >> > --- a/gcc/config/i386/i386.c
> >> > +++ b/gcc/config/i386/i386.c
> >> > @@ -6511,6 +6511,10 @@ ix86_option_override_internal (bool main_args_p,
> >> > ix86_tune_cost->l2_cache_size,
> >> > opts->x_param_values,
> >> > opts_set->x_param_values);
> >> > + maybe_set_param_value (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
> >> > + 35,
> >> > + opts->x_param_values,
> >> > + opts_set->x_param_values);
> >> >
> >> > /* Enable sw prefetching at -O3 for CPUS that prefetching is helpful. */
> >> > if (opts->x_flag_prefetch_loop_arrays < 0
> >> > diff --git a/gcc/expr.c b/gcc/expr.c
> >> > index 134ee731c29..dff24e7f166 100644
> >> > --- a/gcc/expr.c
> >> > +++ b/gcc/expr.c
> >> > @@ -61,7 +61,8 @@ along with GCC; see the file COPYING3. If not see
> >> > #include "tree-chkp.h"
> >> > #include "rtl-chkp.h"
> >> > #include "ccmp.h"
> >> > -
> >> > +#include "params.h"
> >> > +#include "tree-sra.h"
> >> >
> >> > /* If this is nonzero, we do not bother generating VOLATILE
> >> > around volatile memory references, and we are willing to
> >> > @@ -5340,6 +5341,80 @@ emit_storent_insn (rtx to, rtx from)
> >> > return maybe_expand_insn (code, 2, ops);
> >> > }
> >> >
> >> > +/* Generate code for copying data of type TYPE at SOURCE plus OFFSET to TARGET
> >> > + plus OFFSET, but do so element-wise and/or field-wise for each record and
> >> > + array within TYPE. TYPE must either be a register type or an aggregate
> >> > + complying with scalarizable_type_p.
> >> > +
> >> > + If CALL_PARAM_P is nonzero, this is a store into a call param on the
> >> > + stack, and block moves may need to be treated specially. */
> >> > +
> >> > +static void
> >> > +emit_move_elementwise (tree type, rtx target, rtx source, HOST_WIDE_INT offset,
> >> > + int call_param_p)
> >> > +{
> >> > + switch (TREE_CODE (type))
> >> > + {
> >> > + case RECORD_TYPE:
> >> > + for (tree fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
> >> > + if (TREE_CODE (fld) == FIELD_DECL)
> >> > + {
> >> > + HOST_WIDE_INT fld_offset = offset + int_bit_position (fld);
> >> > + tree ft = TREE_TYPE (fld);
> >> > + emit_move_elementwise (ft, target, source, fld_offset,
> >> > + call_param_p);
> >> > + }
> >> > + break;
> >> > +
> >> > + case ARRAY_TYPE:
> >> > + {
> >> > + tree elem_type = TREE_TYPE (type);
> >> > + HOST_WIDE_INT el_size = tree_to_shwi (TYPE_SIZE (elem_type));
> >> > + gcc_assert (el_size > 0);
> >> > +
> >> > + offset_int idx, max;
> >> > + /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
> >> > + if (extract_min_max_idx_from_array (type, &idx, &max))
> >> > + {
> >> > + HOST_WIDE_INT el_offset = offset;
> >> > + for (; idx <= max; ++idx)
> >> > + {
> >> > + emit_move_elementwise (elem_type, target, source, el_offset,
> >> > + call_param_p);
> >> > + el_offset += el_size;
> >> > + }
> >> > + }
> >> > + }
> >> > + break;
> >> > + default:
> >> > + machine_mode mode = TYPE_MODE (type);
> >> > +
> >> > + rtx ntgt = adjust_address (target, mode, offset / BITS_PER_UNIT);
> >> > + rtx nsrc = adjust_address (source, mode, offset / BITS_PER_UNIT);
> >> > +
> >> > + /* TODO: Figure out whether the following is actually necessary. */
> >> > + if (target == ntgt)
> >> > + ntgt = copy_rtx (target);
> >> > + if (source == nsrc)
> >> > + nsrc = copy_rtx (source);
> >> > +
> >> > + gcc_assert (mode != VOIDmode);
> >> > + if (mode != BLKmode)
> >> > + emit_move_insn (ntgt, nsrc);
> >> > + else
> >> > + {
> >> > + /* For example vector gimple registers can end up here. */
> >> > + rtx size = expand_expr (TYPE_SIZE_UNIT (type), NULL_RTX,
> >> > + TYPE_MODE (sizetype), EXPAND_NORMAL);
> >> > + emit_block_move (ntgt, nsrc, size,
> >> > + (call_param_p
> >> > + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> >> > + }
> >> > + break;
> >> > + }
> >> > + return;
> >> > +}
> >> > +
> >> > /* Generate code for computing expression EXP,
> >> > and storing the value into TARGET.
> >> >
> >> > @@ -5713,9 +5788,29 @@ store_expr_with_bounds (tree exp, rtx target, int call_param_p,
> >> > emit_group_store (target, temp, TREE_TYPE (exp),
> >> > int_size_in_bytes (TREE_TYPE (exp)));
> >> > else if (GET_MODE (temp) == BLKmode)
> >> > - emit_block_move (target, temp, expr_size (exp),
> >> > - (call_param_p
> >> > - ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> >> > + {
> >> > + /* Copying smallish BLKmode structures with emit_block_move and thus
> >> > + by-pieces can result in store-to-load stalls. So copy some simple
> >> > + small aggregates element or field-wise. */
> >> > + if (GET_MODE (target) == BLKmode
> >> > + && AGGREGATE_TYPE_P (TREE_TYPE (exp))
> >> > + && !TREE_ADDRESSABLE (TREE_TYPE (exp))
> >> > + && tree_fits_shwi_p (TYPE_SIZE (TREE_TYPE (exp)))
> >> > + && (tree_to_shwi (TYPE_SIZE (TREE_TYPE (exp)))
> >> > + <= (PARAM_VALUE (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY)
> >> > + * BITS_PER_UNIT))
> >> > + && simple_mix_of_records_and_arrays_p (TREE_TYPE (exp), false))
> >> > + {
> >> > + /* FIXME: Can this happen? What would it mean? */
> >> > + gcc_assert (!reverse);
> >> > + emit_move_elementwise (TREE_TYPE (exp), target, temp, 0,
> >> > + call_param_p);
> >> > + }
> >> > + else
> >> > + emit_block_move (target, temp, expr_size (exp),
> >> > + (call_param_p
> >> > + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> >> > + }
> >> > /* If we emit a nontemporal store, there is nothing else to do. */
> >> > else if (nontemporal && emit_storent_insn (target, temp))
> >> > ;
> >> > diff --git a/gcc/ipa-cp.c b/gcc/ipa-cp.c
> >> > index 6b3d8d7364c..7d6019bbd30 100644
> >> > --- a/gcc/ipa-cp.c
> >> > +++ b/gcc/ipa-cp.c
> >> > @@ -124,6 +124,7 @@ along with GCC; see the file COPYING3. If not see
> >> > #include "tree-ssa-ccp.h"
> >> > #include "stringpool.h"
> >> > #include "attribs.h"
> >> > +#include "tree-sra.h"
> >> >
> >> > template <typename valtype> class ipcp_value;
> >> >
> >> > diff --git a/gcc/ipa-prop.h b/gcc/ipa-prop.h
> >> > index fa5bed49ee0..2313cc884ed 100644
> >> > --- a/gcc/ipa-prop.h
> >> > +++ b/gcc/ipa-prop.h
> >> > @@ -877,10 +877,6 @@ ipa_parm_adjustment *ipa_get_adjustment_candidate (tree **, bool *,
> >> > void ipa_release_body_info (struct ipa_func_body_info *);
> >> > tree ipa_get_callee_param_type (struct cgraph_edge *e, int i);
> >> >
> >> > -/* From tree-sra.c: */
> >> > -tree build_ref_for_offset (location_t, tree, HOST_WIDE_INT, bool, tree,
> >> > - gimple_stmt_iterator *, bool);
> >> > -
> >> > /* In ipa-cp.c */
> >> > void ipa_cp_c_finalize (void);
> >> >
> >> > diff --git a/gcc/params.def b/gcc/params.def
> >> > index e55afc28053..5e19f1414a0 100644
> >> > --- a/gcc/params.def
> >> > +++ b/gcc/params.def
> >> > @@ -1294,6 +1294,12 @@ DEFPARAM (PARAM_VECT_EPILOGUES_NOMASK,
> >> > "Enable loop epilogue vectorization using smaller vector size.",
> >> > 0, 0, 1)
> >> >
> >> > +DEFPARAM (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
> >> > + "max-size-for-elementwise-copy",
> >> > + "Maximum size in bytes of a structure or array to by considered for "
> >> > + "copying by its individual fields or elements",
> >> > + 0, 0, 512)
> >> > +
> >> > /*
> >> >
> >> > Local variables:
> >> > diff --git a/gcc/testsuite/gcc.target/i386/pr80689-1.c b/gcc/testsuite/gcc.target/i386/pr80689-1.c
> >> > new file mode 100644
> >> > index 00000000000..4156d4fba45
> >> > --- /dev/null
> >> > +++ b/gcc/testsuite/gcc.target/i386/pr80689-1.c
> >> > @@ -0,0 +1,38 @@
> >> > +/* { dg-do compile } */
> >> > +/* { dg-options "-O2" } */
> >> > +
> >> > +typedef struct st1
> >> > +{
> >> > + long unsigned int a,b;
> >> > + long int c,d;
> >> > +}R;
> >> > +
> >> > +typedef struct st2
> >> > +{
> >> > + int t;
> >> > + R reg;
> >> > +}N;
> >> > +
> >> > +void Set (const R *region, N *n_info );
> >> > +
> >> > +void test(N *n_obj ,const long unsigned int a, const long unsigned int b, const long int c,const long int d)
> >> > +{
> >> > + R reg;
> >> > +
> >> > + reg.a=a;
> >> > + reg.b=b;
> >> > + reg.c=c;
> >> > + reg.d=d;
> >> > + Set (®, n_obj);
> >> > +
> >> > +}
> >> > +
> >> > +void Set (const R *reg, N *n_obj )
> >> > +{
> >> > + n_obj->reg=(*reg);
> >> > +}
> >> > +
> >> > +
> >> > +/* { dg-final { scan-assembler-not "%(x|y|z)mm\[0-9\]+" } } */
> >> > +/* { dg-final { scan-assembler-not "movdqu" } } */
> >> > +/* { dg-final { scan-assembler-not "movups" } } */
> >> > diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
> >> > index bac593951e7..ade97964205 100644
> >> > --- a/gcc/tree-sra.c
> >> > +++ b/gcc/tree-sra.c
> >> > @@ -104,6 +104,7 @@ along with GCC; see the file COPYING3. If not see
> >> > #include "ipa-fnsummary.h"
> >> > #include "ipa-utils.h"
> >> > #include "builtins.h"
> >> > +#include "tree-sra.h"
> >> >
> >> > /* Enumeration of all aggregate reductions we can do. */
> >> > enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
> >> > @@ -952,14 +953,14 @@ create_access (tree expr, gimple *stmt, bool write)
> >> > }
> >> >
> >> >
> >> > -/* Return true iff TYPE is scalarizable - i.e. a RECORD_TYPE or fixed-length
> >> > - ARRAY_TYPE with fields that are either of gimple register types (excluding
> >> > - bit-fields) or (recursively) scalarizable types. CONST_DECL must be true if
> >> > - we are considering a decl from constant pool. If it is false, char arrays
> >> > - will be refused. */
> >> > +/* Return true if TYPE consists of RECORD_TYPE or fixed-length ARRAY_TYPE with
> >> > + fields/elements that are not bit-fields and are either register types or
> >> > + recursively comply with simple_mix_of_records_and_arrays_p. Furthermore, if
> >> > + ALLOW_CHAR_ARRAYS is false, the function will return false also if TYPE
> >> > + contains an array of elements that only have one byte. */
> >> >
> >> > -static bool
> >> > -scalarizable_type_p (tree type, bool const_decl)
> >> > +bool
> >> > +simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays)
> >> > {
> >> > gcc_assert (!is_gimple_reg_type (type));
> >> > if (type_contains_placeholder_p (type))
> >> > @@ -977,7 +978,7 @@ scalarizable_type_p (tree type, bool const_decl)
> >> > return false;
> >> >
> >> > if (!is_gimple_reg_type (ft)
> >> > - && !scalarizable_type_p (ft, const_decl))
> >> > + && !simple_mix_of_records_and_arrays_p (ft, allow_char_arrays))
> >> > return false;
> >> > }
> >> >
> >> > @@ -986,7 +987,7 @@ scalarizable_type_p (tree type, bool const_decl)
> >> > case ARRAY_TYPE:
> >> > {
> >> > HOST_WIDE_INT min_elem_size;
> >> > - if (const_decl)
> >> > + if (allow_char_arrays)
> >> > min_elem_size = 0;
> >> > else
> >> > min_elem_size = BITS_PER_UNIT;
> >> > @@ -1008,7 +1009,7 @@ scalarizable_type_p (tree type, bool const_decl)
> >> >
> >> > tree elem = TREE_TYPE (type);
> >> > if (!is_gimple_reg_type (elem)
> >> > - && !scalarizable_type_p (elem, const_decl))
> >> > + && !simple_mix_of_records_and_arrays_p (elem, allow_char_arrays))
> >> > return false;
> >> > return true;
> >> > }
> >> > @@ -1017,10 +1018,38 @@ scalarizable_type_p (tree type, bool const_decl)
> >> > }
> >> > }
> >> >
> >> > -static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree, tree);
> >> > +static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree,
> >> > + tree);
> >> > +
> >> > +/* For a given array TYPE, return false if its domain does not have any maximum
> >> > + value. Otherwise calculate MIN and MAX indices of the first and the last
> >> > + element. */
> >> > +
> >> > +bool
> >> > +extract_min_max_idx_from_array (tree type, offset_int *min, offset_int *max)
> >> > +{
> >> > + tree domain = TYPE_DOMAIN (type);
> >> > + tree minidx = TYPE_MIN_VALUE (domain);
> >> > + gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
> >> > + tree maxidx = TYPE_MAX_VALUE (domain);
> >> > + if (!maxidx)
> >> > + return false;
> >> > + gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
> >> > +
> >> > + /* MINIDX and MAXIDX are inclusive, and must be interpreted in
> >> > + DOMAIN (e.g. signed int, whereas min/max may be size_int). */
> >> > + *min = wi::to_offset (minidx);
> >> > + *max = wi::to_offset (maxidx);
> >> > + if (!TYPE_UNSIGNED (domain))
> >> > + {
> >> > + *min = wi::sext (*min, TYPE_PRECISION (domain));
> >> > + *max = wi::sext (*max, TYPE_PRECISION (domain));
> >> > + }
> >> > + return true;
> >> > +}
> >> >
> >> > /* Create total_scalarization accesses for all scalar fields of a member
> >> > - of type DECL_TYPE conforming to scalarizable_type_p. BASE
> >> > + of type DECL_TYPE conforming to simple_mix_of_records_and_arrays_p. BASE
> >> > must be the top-most VAR_DECL representing the variable; within that,
> >> > OFFSET locates the member and REF must be the memory reference expression for
> >> > the member. */
> >> > @@ -1047,27 +1076,14 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
> >> > {
> >> > tree elemtype = TREE_TYPE (decl_type);
> >> > tree elem_size = TYPE_SIZE (elemtype);
> >> > - gcc_assert (elem_size && tree_fits_shwi_p (elem_size));
> >> > HOST_WIDE_INT el_size = tree_to_shwi (elem_size);
> >> > gcc_assert (el_size > 0);
> >> >
> >> > - tree minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (decl_type));
> >> > - gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
> >> > - tree maxidx = TYPE_MAX_VALUE (TYPE_DOMAIN (decl_type));
> >> > + offset_int idx, max;
> >> > /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
> >> > - if (maxidx)
> >> > + if (extract_min_max_idx_from_array (decl_type, &idx, &max))
> >> > {
> >> > - gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
> >> > tree domain = TYPE_DOMAIN (decl_type);
> >> > - /* MINIDX and MAXIDX are inclusive, and must be interpreted in
> >> > - DOMAIN (e.g. signed int, whereas min/max may be size_int). */
> >> > - offset_int idx = wi::to_offset (minidx);
> >> > - offset_int max = wi::to_offset (maxidx);
> >> > - if (!TYPE_UNSIGNED (domain))
> >> > - {
> >> > - idx = wi::sext (idx, TYPE_PRECISION (domain));
> >> > - max = wi::sext (max, TYPE_PRECISION (domain));
> >> > - }
> >> > for (int el_off = offset; idx <= max; ++idx)
> >> > {
> >> > tree nref = build4 (ARRAY_REF, elemtype,
> >> > @@ -1088,10 +1104,10 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
> >> > }
> >> >
> >> > /* Create total_scalarization accesses for a member of type TYPE, which must
> >> > - satisfy either is_gimple_reg_type or scalarizable_type_p. BASE must be the
> >> > - top-most VAR_DECL representing the variable; within that, POS and SIZE locate
> >> > - the member, REVERSE gives its torage order. and REF must be the reference
> >> > - expression for it. */
> >> > + satisfy either is_gimple_reg_type or simple_mix_of_records_and_arrays_p.
> >> > + BASE must be the top-most VAR_DECL representing the variable; within that,
> >> > + POS and SIZE locate the member, REVERSE gives its torage order. and REF must
> >> > + be the reference expression for it. */
> >> >
> >> > static void
> >> > scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
> >> > @@ -1111,7 +1127,8 @@ scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
> >> > }
> >> >
> >> > /* Create a total_scalarization access for VAR as a whole. VAR must be of a
> >> > - RECORD_TYPE or ARRAY_TYPE conforming to scalarizable_type_p. */
> >> > + RECORD_TYPE or ARRAY_TYPE conforming to
> >> > + simple_mix_of_records_and_arrays_p. */
> >> >
> >> > static void
> >> > create_total_scalarization_access (tree var)
> >> > @@ -2803,8 +2820,9 @@ analyze_all_variable_accesses (void)
> >> > {
> >> > tree var = candidate (i);
> >> >
> >> > - if (VAR_P (var) && scalarizable_type_p (TREE_TYPE (var),
> >> > - constant_decl_p (var)))
> >> > + if (VAR_P (var)
> >> > + && simple_mix_of_records_and_arrays_p (TREE_TYPE (var),
> >> > + constant_decl_p (var)))
> >> > {
> >> > if (tree_to_uhwi (TYPE_SIZE (TREE_TYPE (var)))
> >> > <= max_scalarization_size)
> >> > diff --git a/gcc/tree-sra.h b/gcc/tree-sra.h
> >> > new file mode 100644
> >> > index 00000000000..dc901385994
> >> > --- /dev/null
> >> > +++ b/gcc/tree-sra.h
> >> > @@ -0,0 +1,33 @@
> >> > +/* tree-sra.h - Run-time parameters.
> >> > + Copyright (C) 2017 Free Software Foundation, Inc.
> >> > +
> >> > +This file is part of GCC.
> >> > +
> >> > +GCC is free software; you can redistribute it and/or modify it under
> >> > +the terms of the GNU General Public License as published by the Free
> >> > +Software Foundation; either version 3, or (at your option) any later
> >> > +version.
> >> > +
> >> > +GCC is distributed in the hope that it will be useful, but WITHOUT ANY
> >> > +WARRANTY; without even the implied warranty of MERCHANTABILITY or
> >> > +FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
> >> > +for more details.
> >> > +
> >> > +You should have received a copy of the GNU General Public License
> >> > +along with GCC; see the file COPYING3. If not see
> >> > +<http://www.gnu.org/licenses/>. */
> >> > +
> >> > +#ifndef TREE_SRA_H
> >> > +#define TREE_SRA_H
> >> > +
> >> > +
> >> > +bool simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays);
> >> > +bool extract_min_max_idx_from_array (tree type, offset_int *idx,
> >> > + offset_int *max);
> >> > +tree build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
> >> > + bool reverse, tree exp_type,
> >> > + gimple_stmt_iterator *gsi, bool insert_after);
> >> > +
> >> > +
> >> > +
> >> > +#endif /* TREE_SRA_H */
> >> > --
> >> > 2.14.1
> >> >
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-10-26 12:23 ` Martin Jambor
2017-10-26 12:44 ` Richard Biener
@ 2017-10-26 14:18 ` Michael Matz
1 sibling, 0 replies; 26+ messages in thread
From: Michael Matz @ 2017-10-26 14:18 UTC (permalink / raw)
To: Martin Jambor; +Cc: Richard Biener, GCC Patches, Jan Hubicka
Hi,
On Thu, 26 Oct 2017, Martin Jambor wrote:
> > 35 bytes seems to be much - what is the code-size impact?
>
> I will find out and report on that. I need at least 32 bytes (four
> long ints) to fix imagemagick, where the problematic structure is:
Surely the final heuristic should look at the size and number of elements
of the struct in question, not only on size.
Ciao,
Michael.
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-10-26 12:57 ` Jan Hubicka
@ 2017-10-26 14:53 ` Richard Biener
2017-10-26 15:12 ` Richard Biener
2017-10-27 12:27 ` Jan Hubicka
0 siblings, 2 replies; 26+ messages in thread
From: Richard Biener @ 2017-10-26 14:53 UTC (permalink / raw)
To: Jan Hubicka; +Cc: GCC Patches
On Thu, Oct 26, 2017 at 2:55 PM, Jan Hubicka <hubicka@ucw.cz> wrote:
>> I think the limit should be on the number of generated copies and not
>> the overall size of the structure... If the struct were composed of
>> 32 individual chars we wouldn't want to emit 32 loads and 32 stores...
>>
>> I wonder how rep; movb; interacts with store to load forwarding? Is
>> that maybe optimized well on some archs? movb should always
>> forward and wasn't the setup cost for small N reasonable on modern
>> CPUs?
>
> rep mov is win over loop for blocks over 128bytes on core, for blocks in rage
> 24-128 on zen. This is w/o store/load forwarding, but I doubt those provide
> a cheap way around.
>
>>
>> It probably depends on the width of the entries in the store buffer,
>> if they appear in-order and the alignment of the stores (if they are larger than
>> 8 bytes they are surely aligned). IIRC CPUs had smaller store buffer
>> entries than cache line size.
>>
>> Given that load bandwith is usually higher than store bandwith it
>> might make sense to do the store combining in our copying sequence,
>> like for the 8 byte entry case use sth like
>>
>> movq 0(%eax), %xmm0
>> movhps 8(%eax), %xmm0 // or vpinsert
>> mov[au]ps %xmm0, 0%(ebx)
>> ...
>>
>> thus do two loads per store and perform the stores in wider
>> mode?
>
> This may be somewhat faster indeed. I am not sure if store to load
> forwarding will work for the later half when read again by halves.
> It would not happen on older CPUs :)
Yes, forwarding larger stores to smaller loads generally works fine
since forever with the usual restrictions of alignment/size being
power of two "halves".
The question is of course what to do for 4 byte or smaller elements or
mixed size elements. We can do zero-extending loads
(do we have them for QI, HI mode loads as well?) and
do shift and or's. I'm quite sure the CPUs wouldn't like to
see vpinsert's of different vector mode destinations. So it
would be 8 byte stores from GPRs and values built up via
shift & or.
As said, the important part is that IIRC CPUs can usually
have more loads in flight than stores. Esp. Bulldozer
with the split core was store buffer size limited (but it
could do merging of store buffer entries IIRC).
Richard.
> Honza
>>
>> As said a general concern was you not copying padding. If you
>> put this into an even more common place you surely will break
>> stuff, no?
>>
>> Richard.
>>
>> >
>> > Martin
>> >
>> >
>> >>
>> >> Richard.
>> >>
>> >> > Martin
>> >> >
>> >> >
>> >> > 2017-10-12 Martin Jambor <mjambor@suse.cz>
>> >> >
>> >> > PR target/80689
>> >> > * tree-sra.h: New file.
>> >> > * ipa-prop.h: Moved declaration of build_ref_for_offset to
>> >> > tree-sra.h.
>> >> > * expr.c: Include params.h and tree-sra.h.
>> >> > (emit_move_elementwise): New function.
>> >> > (store_expr_with_bounds): Optionally use it.
>> >> > * ipa-cp.c: Include tree-sra.h.
>> >> > * params.def (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY): New.
>> >> > * config/i386/i386.c (ix86_option_override_internal): Set
>> >> > PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY to 35.
>> >> > * tree-sra.c: Include tree-sra.h.
>> >> > (scalarizable_type_p): Renamed to
>> >> > simple_mix_of_records_and_arrays_p, made public, renamed the
>> >> > second parameter to allow_char_arrays.
>> >> > (extract_min_max_idx_from_array): New function.
>> >> > (completely_scalarize): Moved bits of the function to
>> >> > extract_min_max_idx_from_array.
>> >> >
>> >> > testsuite/
>> >> > * gcc.target/i386/pr80689-1.c: New test.
>> >> > ---
>> >> > gcc/config/i386/i386.c | 4 ++
>> >> > gcc/expr.c | 103 ++++++++++++++++++++++++++++--
>> >> > gcc/ipa-cp.c | 1 +
>> >> > gcc/ipa-prop.h | 4 --
>> >> > gcc/params.def | 6 ++
>> >> > gcc/testsuite/gcc.target/i386/pr80689-1.c | 38 +++++++++++
>> >> > gcc/tree-sra.c | 86 +++++++++++++++----------
>> >> > gcc/tree-sra.h | 33 ++++++++++
>> >> > 8 files changed, 233 insertions(+), 42 deletions(-)
>> >> > create mode 100644 gcc/testsuite/gcc.target/i386/pr80689-1.c
>> >> > create mode 100644 gcc/tree-sra.h
>> >> >
>> >> > diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c
>> >> > index 1ee8351c21f..87f602e7ead 100644
>> >> > --- a/gcc/config/i386/i386.c
>> >> > +++ b/gcc/config/i386/i386.c
>> >> > @@ -6511,6 +6511,10 @@ ix86_option_override_internal (bool main_args_p,
>> >> > ix86_tune_cost->l2_cache_size,
>> >> > opts->x_param_values,
>> >> > opts_set->x_param_values);
>> >> > + maybe_set_param_value (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
>> >> > + 35,
>> >> > + opts->x_param_values,
>> >> > + opts_set->x_param_values);
>> >> >
>> >> > /* Enable sw prefetching at -O3 for CPUS that prefetching is helpful. */
>> >> > if (opts->x_flag_prefetch_loop_arrays < 0
>> >> > diff --git a/gcc/expr.c b/gcc/expr.c
>> >> > index 134ee731c29..dff24e7f166 100644
>> >> > --- a/gcc/expr.c
>> >> > +++ b/gcc/expr.c
>> >> > @@ -61,7 +61,8 @@ along with GCC; see the file COPYING3. If not see
>> >> > #include "tree-chkp.h"
>> >> > #include "rtl-chkp.h"
>> >> > #include "ccmp.h"
>> >> > -
>> >> > +#include "params.h"
>> >> > +#include "tree-sra.h"
>> >> >
>> >> > /* If this is nonzero, we do not bother generating VOLATILE
>> >> > around volatile memory references, and we are willing to
>> >> > @@ -5340,6 +5341,80 @@ emit_storent_insn (rtx to, rtx from)
>> >> > return maybe_expand_insn (code, 2, ops);
>> >> > }
>> >> >
>> >> > +/* Generate code for copying data of type TYPE at SOURCE plus OFFSET to TARGET
>> >> > + plus OFFSET, but do so element-wise and/or field-wise for each record and
>> >> > + array within TYPE. TYPE must either be a register type or an aggregate
>> >> > + complying with scalarizable_type_p.
>> >> > +
>> >> > + If CALL_PARAM_P is nonzero, this is a store into a call param on the
>> >> > + stack, and block moves may need to be treated specially. */
>> >> > +
>> >> > +static void
>> >> > +emit_move_elementwise (tree type, rtx target, rtx source, HOST_WIDE_INT offset,
>> >> > + int call_param_p)
>> >> > +{
>> >> > + switch (TREE_CODE (type))
>> >> > + {
>> >> > + case RECORD_TYPE:
>> >> > + for (tree fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
>> >> > + if (TREE_CODE (fld) == FIELD_DECL)
>> >> > + {
>> >> > + HOST_WIDE_INT fld_offset = offset + int_bit_position (fld);
>> >> > + tree ft = TREE_TYPE (fld);
>> >> > + emit_move_elementwise (ft, target, source, fld_offset,
>> >> > + call_param_p);
>> >> > + }
>> >> > + break;
>> >> > +
>> >> > + case ARRAY_TYPE:
>> >> > + {
>> >> > + tree elem_type = TREE_TYPE (type);
>> >> > + HOST_WIDE_INT el_size = tree_to_shwi (TYPE_SIZE (elem_type));
>> >> > + gcc_assert (el_size > 0);
>> >> > +
>> >> > + offset_int idx, max;
>> >> > + /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
>> >> > + if (extract_min_max_idx_from_array (type, &idx, &max))
>> >> > + {
>> >> > + HOST_WIDE_INT el_offset = offset;
>> >> > + for (; idx <= max; ++idx)
>> >> > + {
>> >> > + emit_move_elementwise (elem_type, target, source, el_offset,
>> >> > + call_param_p);
>> >> > + el_offset += el_size;
>> >> > + }
>> >> > + }
>> >> > + }
>> >> > + break;
>> >> > + default:
>> >> > + machine_mode mode = TYPE_MODE (type);
>> >> > +
>> >> > + rtx ntgt = adjust_address (target, mode, offset / BITS_PER_UNIT);
>> >> > + rtx nsrc = adjust_address (source, mode, offset / BITS_PER_UNIT);
>> >> > +
>> >> > + /* TODO: Figure out whether the following is actually necessary. */
>> >> > + if (target == ntgt)
>> >> > + ntgt = copy_rtx (target);
>> >> > + if (source == nsrc)
>> >> > + nsrc = copy_rtx (source);
>> >> > +
>> >> > + gcc_assert (mode != VOIDmode);
>> >> > + if (mode != BLKmode)
>> >> > + emit_move_insn (ntgt, nsrc);
>> >> > + else
>> >> > + {
>> >> > + /* For example vector gimple registers can end up here. */
>> >> > + rtx size = expand_expr (TYPE_SIZE_UNIT (type), NULL_RTX,
>> >> > + TYPE_MODE (sizetype), EXPAND_NORMAL);
>> >> > + emit_block_move (ntgt, nsrc, size,
>> >> > + (call_param_p
>> >> > + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
>> >> > + }
>> >> > + break;
>> >> > + }
>> >> > + return;
>> >> > +}
>> >> > +
>> >> > /* Generate code for computing expression EXP,
>> >> > and storing the value into TARGET.
>> >> >
>> >> > @@ -5713,9 +5788,29 @@ store_expr_with_bounds (tree exp, rtx target, int call_param_p,
>> >> > emit_group_store (target, temp, TREE_TYPE (exp),
>> >> > int_size_in_bytes (TREE_TYPE (exp)));
>> >> > else if (GET_MODE (temp) == BLKmode)
>> >> > - emit_block_move (target, temp, expr_size (exp),
>> >> > - (call_param_p
>> >> > - ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
>> >> > + {
>> >> > + /* Copying smallish BLKmode structures with emit_block_move and thus
>> >> > + by-pieces can result in store-to-load stalls. So copy some simple
>> >> > + small aggregates element or field-wise. */
>> >> > + if (GET_MODE (target) == BLKmode
>> >> > + && AGGREGATE_TYPE_P (TREE_TYPE (exp))
>> >> > + && !TREE_ADDRESSABLE (TREE_TYPE (exp))
>> >> > + && tree_fits_shwi_p (TYPE_SIZE (TREE_TYPE (exp)))
>> >> > + && (tree_to_shwi (TYPE_SIZE (TREE_TYPE (exp)))
>> >> > + <= (PARAM_VALUE (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY)
>> >> > + * BITS_PER_UNIT))
>> >> > + && simple_mix_of_records_and_arrays_p (TREE_TYPE (exp), false))
>> >> > + {
>> >> > + /* FIXME: Can this happen? What would it mean? */
>> >> > + gcc_assert (!reverse);
>> >> > + emit_move_elementwise (TREE_TYPE (exp), target, temp, 0,
>> >> > + call_param_p);
>> >> > + }
>> >> > + else
>> >> > + emit_block_move (target, temp, expr_size (exp),
>> >> > + (call_param_p
>> >> > + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
>> >> > + }
>> >> > /* If we emit a nontemporal store, there is nothing else to do. */
>> >> > else if (nontemporal && emit_storent_insn (target, temp))
>> >> > ;
>> >> > diff --git a/gcc/ipa-cp.c b/gcc/ipa-cp.c
>> >> > index 6b3d8d7364c..7d6019bbd30 100644
>> >> > --- a/gcc/ipa-cp.c
>> >> > +++ b/gcc/ipa-cp.c
>> >> > @@ -124,6 +124,7 @@ along with GCC; see the file COPYING3. If not see
>> >> > #include "tree-ssa-ccp.h"
>> >> > #include "stringpool.h"
>> >> > #include "attribs.h"
>> >> > +#include "tree-sra.h"
>> >> >
>> >> > template <typename valtype> class ipcp_value;
>> >> >
>> >> > diff --git a/gcc/ipa-prop.h b/gcc/ipa-prop.h
>> >> > index fa5bed49ee0..2313cc884ed 100644
>> >> > --- a/gcc/ipa-prop.h
>> >> > +++ b/gcc/ipa-prop.h
>> >> > @@ -877,10 +877,6 @@ ipa_parm_adjustment *ipa_get_adjustment_candidate (tree **, bool *,
>> >> > void ipa_release_body_info (struct ipa_func_body_info *);
>> >> > tree ipa_get_callee_param_type (struct cgraph_edge *e, int i);
>> >> >
>> >> > -/* From tree-sra.c: */
>> >> > -tree build_ref_for_offset (location_t, tree, HOST_WIDE_INT, bool, tree,
>> >> > - gimple_stmt_iterator *, bool);
>> >> > -
>> >> > /* In ipa-cp.c */
>> >> > void ipa_cp_c_finalize (void);
>> >> >
>> >> > diff --git a/gcc/params.def b/gcc/params.def
>> >> > index e55afc28053..5e19f1414a0 100644
>> >> > --- a/gcc/params.def
>> >> > +++ b/gcc/params.def
>> >> > @@ -1294,6 +1294,12 @@ DEFPARAM (PARAM_VECT_EPILOGUES_NOMASK,
>> >> > "Enable loop epilogue vectorization using smaller vector size.",
>> >> > 0, 0, 1)
>> >> >
>> >> > +DEFPARAM (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
>> >> > + "max-size-for-elementwise-copy",
>> >> > + "Maximum size in bytes of a structure or array to by considered for "
>> >> > + "copying by its individual fields or elements",
>> >> > + 0, 0, 512)
>> >> > +
>> >> > /*
>> >> >
>> >> > Local variables:
>> >> > diff --git a/gcc/testsuite/gcc.target/i386/pr80689-1.c b/gcc/testsuite/gcc.target/i386/pr80689-1.c
>> >> > new file mode 100644
>> >> > index 00000000000..4156d4fba45
>> >> > --- /dev/null
>> >> > +++ b/gcc/testsuite/gcc.target/i386/pr80689-1.c
>> >> > @@ -0,0 +1,38 @@
>> >> > +/* { dg-do compile } */
>> >> > +/* { dg-options "-O2" } */
>> >> > +
>> >> > +typedef struct st1
>> >> > +{
>> >> > + long unsigned int a,b;
>> >> > + long int c,d;
>> >> > +}R;
>> >> > +
>> >> > +typedef struct st2
>> >> > +{
>> >> > + int t;
>> >> > + R reg;
>> >> > +}N;
>> >> > +
>> >> > +void Set (const R *region, N *n_info );
>> >> > +
>> >> > +void test(N *n_obj ,const long unsigned int a, const long unsigned int b, const long int c,const long int d)
>> >> > +{
>> >> > + R reg;
>> >> > +
>> >> > + reg.a=a;
>> >> > + reg.b=b;
>> >> > + reg.c=c;
>> >> > + reg.d=d;
>> >> > + Set (®, n_obj);
>> >> > +
>> >> > +}
>> >> > +
>> >> > +void Set (const R *reg, N *n_obj )
>> >> > +{
>> >> > + n_obj->reg=(*reg);
>> >> > +}
>> >> > +
>> >> > +
>> >> > +/* { dg-final { scan-assembler-not "%(x|y|z)mm\[0-9\]+" } } */
>> >> > +/* { dg-final { scan-assembler-not "movdqu" } } */
>> >> > +/* { dg-final { scan-assembler-not "movups" } } */
>> >> > diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
>> >> > index bac593951e7..ade97964205 100644
>> >> > --- a/gcc/tree-sra.c
>> >> > +++ b/gcc/tree-sra.c
>> >> > @@ -104,6 +104,7 @@ along with GCC; see the file COPYING3. If not see
>> >> > #include "ipa-fnsummary.h"
>> >> > #include "ipa-utils.h"
>> >> > #include "builtins.h"
>> >> > +#include "tree-sra.h"
>> >> >
>> >> > /* Enumeration of all aggregate reductions we can do. */
>> >> > enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
>> >> > @@ -952,14 +953,14 @@ create_access (tree expr, gimple *stmt, bool write)
>> >> > }
>> >> >
>> >> >
>> >> > -/* Return true iff TYPE is scalarizable - i.e. a RECORD_TYPE or fixed-length
>> >> > - ARRAY_TYPE with fields that are either of gimple register types (excluding
>> >> > - bit-fields) or (recursively) scalarizable types. CONST_DECL must be true if
>> >> > - we are considering a decl from constant pool. If it is false, char arrays
>> >> > - will be refused. */
>> >> > +/* Return true if TYPE consists of RECORD_TYPE or fixed-length ARRAY_TYPE with
>> >> > + fields/elements that are not bit-fields and are either register types or
>> >> > + recursively comply with simple_mix_of_records_and_arrays_p. Furthermore, if
>> >> > + ALLOW_CHAR_ARRAYS is false, the function will return false also if TYPE
>> >> > + contains an array of elements that only have one byte. */
>> >> >
>> >> > -static bool
>> >> > -scalarizable_type_p (tree type, bool const_decl)
>> >> > +bool
>> >> > +simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays)
>> >> > {
>> >> > gcc_assert (!is_gimple_reg_type (type));
>> >> > if (type_contains_placeholder_p (type))
>> >> > @@ -977,7 +978,7 @@ scalarizable_type_p (tree type, bool const_decl)
>> >> > return false;
>> >> >
>> >> > if (!is_gimple_reg_type (ft)
>> >> > - && !scalarizable_type_p (ft, const_decl))
>> >> > + && !simple_mix_of_records_and_arrays_p (ft, allow_char_arrays))
>> >> > return false;
>> >> > }
>> >> >
>> >> > @@ -986,7 +987,7 @@ scalarizable_type_p (tree type, bool const_decl)
>> >> > case ARRAY_TYPE:
>> >> > {
>> >> > HOST_WIDE_INT min_elem_size;
>> >> > - if (const_decl)
>> >> > + if (allow_char_arrays)
>> >> > min_elem_size = 0;
>> >> > else
>> >> > min_elem_size = BITS_PER_UNIT;
>> >> > @@ -1008,7 +1009,7 @@ scalarizable_type_p (tree type, bool const_decl)
>> >> >
>> >> > tree elem = TREE_TYPE (type);
>> >> > if (!is_gimple_reg_type (elem)
>> >> > - && !scalarizable_type_p (elem, const_decl))
>> >> > + && !simple_mix_of_records_and_arrays_p (elem, allow_char_arrays))
>> >> > return false;
>> >> > return true;
>> >> > }
>> >> > @@ -1017,10 +1018,38 @@ scalarizable_type_p (tree type, bool const_decl)
>> >> > }
>> >> > }
>> >> >
>> >> > -static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree, tree);
>> >> > +static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree,
>> >> > + tree);
>> >> > +
>> >> > +/* For a given array TYPE, return false if its domain does not have any maximum
>> >> > + value. Otherwise calculate MIN and MAX indices of the first and the last
>> >> > + element. */
>> >> > +
>> >> > +bool
>> >> > +extract_min_max_idx_from_array (tree type, offset_int *min, offset_int *max)
>> >> > +{
>> >> > + tree domain = TYPE_DOMAIN (type);
>> >> > + tree minidx = TYPE_MIN_VALUE (domain);
>> >> > + gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
>> >> > + tree maxidx = TYPE_MAX_VALUE (domain);
>> >> > + if (!maxidx)
>> >> > + return false;
>> >> > + gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
>> >> > +
>> >> > + /* MINIDX and MAXIDX are inclusive, and must be interpreted in
>> >> > + DOMAIN (e.g. signed int, whereas min/max may be size_int). */
>> >> > + *min = wi::to_offset (minidx);
>> >> > + *max = wi::to_offset (maxidx);
>> >> > + if (!TYPE_UNSIGNED (domain))
>> >> > + {
>> >> > + *min = wi::sext (*min, TYPE_PRECISION (domain));
>> >> > + *max = wi::sext (*max, TYPE_PRECISION (domain));
>> >> > + }
>> >> > + return true;
>> >> > +}
>> >> >
>> >> > /* Create total_scalarization accesses for all scalar fields of a member
>> >> > - of type DECL_TYPE conforming to scalarizable_type_p. BASE
>> >> > + of type DECL_TYPE conforming to simple_mix_of_records_and_arrays_p. BASE
>> >> > must be the top-most VAR_DECL representing the variable; within that,
>> >> > OFFSET locates the member and REF must be the memory reference expression for
>> >> > the member. */
>> >> > @@ -1047,27 +1076,14 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
>> >> > {
>> >> > tree elemtype = TREE_TYPE (decl_type);
>> >> > tree elem_size = TYPE_SIZE (elemtype);
>> >> > - gcc_assert (elem_size && tree_fits_shwi_p (elem_size));
>> >> > HOST_WIDE_INT el_size = tree_to_shwi (elem_size);
>> >> > gcc_assert (el_size > 0);
>> >> >
>> >> > - tree minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (decl_type));
>> >> > - gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
>> >> > - tree maxidx = TYPE_MAX_VALUE (TYPE_DOMAIN (decl_type));
>> >> > + offset_int idx, max;
>> >> > /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
>> >> > - if (maxidx)
>> >> > + if (extract_min_max_idx_from_array (decl_type, &idx, &max))
>> >> > {
>> >> > - gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
>> >> > tree domain = TYPE_DOMAIN (decl_type);
>> >> > - /* MINIDX and MAXIDX are inclusive, and must be interpreted in
>> >> > - DOMAIN (e.g. signed int, whereas min/max may be size_int). */
>> >> > - offset_int idx = wi::to_offset (minidx);
>> >> > - offset_int max = wi::to_offset (maxidx);
>> >> > - if (!TYPE_UNSIGNED (domain))
>> >> > - {
>> >> > - idx = wi::sext (idx, TYPE_PRECISION (domain));
>> >> > - max = wi::sext (max, TYPE_PRECISION (domain));
>> >> > - }
>> >> > for (int el_off = offset; idx <= max; ++idx)
>> >> > {
>> >> > tree nref = build4 (ARRAY_REF, elemtype,
>> >> > @@ -1088,10 +1104,10 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
>> >> > }
>> >> >
>> >> > /* Create total_scalarization accesses for a member of type TYPE, which must
>> >> > - satisfy either is_gimple_reg_type or scalarizable_type_p. BASE must be the
>> >> > - top-most VAR_DECL representing the variable; within that, POS and SIZE locate
>> >> > - the member, REVERSE gives its torage order. and REF must be the reference
>> >> > - expression for it. */
>> >> > + satisfy either is_gimple_reg_type or simple_mix_of_records_and_arrays_p.
>> >> > + BASE must be the top-most VAR_DECL representing the variable; within that,
>> >> > + POS and SIZE locate the member, REVERSE gives its torage order. and REF must
>> >> > + be the reference expression for it. */
>> >> >
>> >> > static void
>> >> > scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
>> >> > @@ -1111,7 +1127,8 @@ scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
>> >> > }
>> >> >
>> >> > /* Create a total_scalarization access for VAR as a whole. VAR must be of a
>> >> > - RECORD_TYPE or ARRAY_TYPE conforming to scalarizable_type_p. */
>> >> > + RECORD_TYPE or ARRAY_TYPE conforming to
>> >> > + simple_mix_of_records_and_arrays_p. */
>> >> >
>> >> > static void
>> >> > create_total_scalarization_access (tree var)
>> >> > @@ -2803,8 +2820,9 @@ analyze_all_variable_accesses (void)
>> >> > {
>> >> > tree var = candidate (i);
>> >> >
>> >> > - if (VAR_P (var) && scalarizable_type_p (TREE_TYPE (var),
>> >> > - constant_decl_p (var)))
>> >> > + if (VAR_P (var)
>> >> > + && simple_mix_of_records_and_arrays_p (TREE_TYPE (var),
>> >> > + constant_decl_p (var)))
>> >> > {
>> >> > if (tree_to_uhwi (TYPE_SIZE (TREE_TYPE (var)))
>> >> > <= max_scalarization_size)
>> >> > diff --git a/gcc/tree-sra.h b/gcc/tree-sra.h
>> >> > new file mode 100644
>> >> > index 00000000000..dc901385994
>> >> > --- /dev/null
>> >> > +++ b/gcc/tree-sra.h
>> >> > @@ -0,0 +1,33 @@
>> >> > +/* tree-sra.h - Run-time parameters.
>> >> > + Copyright (C) 2017 Free Software Foundation, Inc.
>> >> > +
>> >> > +This file is part of GCC.
>> >> > +
>> >> > +GCC is free software; you can redistribute it and/or modify it under
>> >> > +the terms of the GNU General Public License as published by the Free
>> >> > +Software Foundation; either version 3, or (at your option) any later
>> >> > +version.
>> >> > +
>> >> > +GCC is distributed in the hope that it will be useful, but WITHOUT ANY
>> >> > +WARRANTY; without even the implied warranty of MERCHANTABILITY or
>> >> > +FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
>> >> > +for more details.
>> >> > +
>> >> > +You should have received a copy of the GNU General Public License
>> >> > +along with GCC; see the file COPYING3. If not see
>> >> > +<http://www.gnu.org/licenses/>. */
>> >> > +
>> >> > +#ifndef TREE_SRA_H
>> >> > +#define TREE_SRA_H
>> >> > +
>> >> > +
>> >> > +bool simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays);
>> >> > +bool extract_min_max_idx_from_array (tree type, offset_int *idx,
>> >> > + offset_int *max);
>> >> > +tree build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
>> >> > + bool reverse, tree exp_type,
>> >> > + gimple_stmt_iterator *gsi, bool insert_after);
>> >> > +
>> >> > +
>> >> > +
>> >> > +#endif /* TREE_SRA_H */
>> >> > --
>> >> > 2.14.1
>> >> >
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-10-26 14:53 ` Richard Biener
@ 2017-10-26 15:12 ` Richard Biener
2017-11-03 16:38 ` Martin Jambor
2017-10-27 12:27 ` Jan Hubicka
1 sibling, 1 reply; 26+ messages in thread
From: Richard Biener @ 2017-10-26 15:12 UTC (permalink / raw)
To: Jan Hubicka; +Cc: GCC Patches
On Thu, Oct 26, 2017 at 4:38 PM, Richard Biener
<richard.guenther@gmail.com> wrote:
> On Thu, Oct 26, 2017 at 2:55 PM, Jan Hubicka <hubicka@ucw.cz> wrote:
>>> I think the limit should be on the number of generated copies and not
>>> the overall size of the structure... If the struct were composed of
>>> 32 individual chars we wouldn't want to emit 32 loads and 32 stores...
>>>
>>> I wonder how rep; movb; interacts with store to load forwarding? Is
>>> that maybe optimized well on some archs? movb should always
>>> forward and wasn't the setup cost for small N reasonable on modern
>>> CPUs?
>>
>> rep mov is win over loop for blocks over 128bytes on core, for blocks in rage
>> 24-128 on zen. This is w/o store/load forwarding, but I doubt those provide
>> a cheap way around.
>>
>>>
>>> It probably depends on the width of the entries in the store buffer,
>>> if they appear in-order and the alignment of the stores (if they are larger than
>>> 8 bytes they are surely aligned). IIRC CPUs had smaller store buffer
>>> entries than cache line size.
>>>
>>> Given that load bandwith is usually higher than store bandwith it
>>> might make sense to do the store combining in our copying sequence,
>>> like for the 8 byte entry case use sth like
>>>
>>> movq 0(%eax), %xmm0
>>> movhps 8(%eax), %xmm0 // or vpinsert
>>> mov[au]ps %xmm0, 0%(ebx)
>>> ...
>>>
>>> thus do two loads per store and perform the stores in wider
>>> mode?
>>
>> This may be somewhat faster indeed. I am not sure if store to load
>> forwarding will work for the later half when read again by halves.
>> It would not happen on older CPUs :)
>
> Yes, forwarding larger stores to smaller loads generally works fine
> since forever with the usual restrictions of alignment/size being
> power of two "halves".
>
> The question is of course what to do for 4 byte or smaller elements or
> mixed size elements. We can do zero-extending loads
> (do we have them for QI, HI mode loads as well?) and
> do shift and or's. I'm quite sure the CPUs wouldn't like to
> see vpinsert's of different vector mode destinations. So it
> would be 8 byte stores from GPRs and values built up via
> shift & or.
Like we generate
foo:
.LFB0:
.cfi_startproc
movl 4(%rdi), %eax
movzwl 2(%rdi), %edx
salq $16, %rax
orq %rdx, %rax
movzbl 1(%rdi), %edx
salq $8, %rax
orq %rdx, %rax
movzbl (%rdi), %edx
salq $8, %rax
orq %rdx, %rax
movq %rax, (%rsi)
ret
for
struct x { char e; char f; short c; int i; } a;
void foo (struct x *p, long *q)
{
*q = (((((((unsigned long)(unsigned int)p->i) << 16)
| (((unsigned long)(unsigned short)p->c))) << 8)
| (((unsigned long)(unsigned char)p->f))) << 8)
| ((unsigned long)(unsigned char)p->e);
}
if you disable the bswap pass. Doing 4 byte stores in this
case would save some prefixes at least. I expected the
ORs and shifts to have smaller encodings...
With 4 byte stores we end up with the same size as with
individual loads & stores.
> As said, the important part is that IIRC CPUs can usually
> have more loads in flight than stores. Esp. Bulldozer
> with the split core was store buffer size limited (but it
> could do merging of store buffer entries IIRC).
Also if we do the stores in smaller chunks we are more
likely hitting the same store-to-load-forwarding issue
elsewhere. Like in case the destination is memcpy'ed
away.
So the proposed change isn't necessarily a win without
a possible similar regression that it tries to fix.
Whole-program analysis of accesses might allow
marking affected objects.
Richard.
> Richard.
>
>> Honza
>>>
>>> As said a general concern was you not copying padding. If you
>>> put this into an even more common place you surely will break
>>> stuff, no?
>>>
>>> Richard.
>>>
>>> >
>>> > Martin
>>> >
>>> >
>>> >>
>>> >> Richard.
>>> >>
>>> >> > Martin
>>> >> >
>>> >> >
>>> >> > 2017-10-12 Martin Jambor <mjambor@suse.cz>
>>> >> >
>>> >> > PR target/80689
>>> >> > * tree-sra.h: New file.
>>> >> > * ipa-prop.h: Moved declaration of build_ref_for_offset to
>>> >> > tree-sra.h.
>>> >> > * expr.c: Include params.h and tree-sra.h.
>>> >> > (emit_move_elementwise): New function.
>>> >> > (store_expr_with_bounds): Optionally use it.
>>> >> > * ipa-cp.c: Include tree-sra.h.
>>> >> > * params.def (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY): New.
>>> >> > * config/i386/i386.c (ix86_option_override_internal): Set
>>> >> > PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY to 35.
>>> >> > * tree-sra.c: Include tree-sra.h.
>>> >> > (scalarizable_type_p): Renamed to
>>> >> > simple_mix_of_records_and_arrays_p, made public, renamed the
>>> >> > second parameter to allow_char_arrays.
>>> >> > (extract_min_max_idx_from_array): New function.
>>> >> > (completely_scalarize): Moved bits of the function to
>>> >> > extract_min_max_idx_from_array.
>>> >> >
>>> >> > testsuite/
>>> >> > * gcc.target/i386/pr80689-1.c: New test.
>>> >> > ---
>>> >> > gcc/config/i386/i386.c | 4 ++
>>> >> > gcc/expr.c | 103 ++++++++++++++++++++++++++++--
>>> >> > gcc/ipa-cp.c | 1 +
>>> >> > gcc/ipa-prop.h | 4 --
>>> >> > gcc/params.def | 6 ++
>>> >> > gcc/testsuite/gcc.target/i386/pr80689-1.c | 38 +++++++++++
>>> >> > gcc/tree-sra.c | 86 +++++++++++++++----------
>>> >> > gcc/tree-sra.h | 33 ++++++++++
>>> >> > 8 files changed, 233 insertions(+), 42 deletions(-)
>>> >> > create mode 100644 gcc/testsuite/gcc.target/i386/pr80689-1.c
>>> >> > create mode 100644 gcc/tree-sra.h
>>> >> >
>>> >> > diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c
>>> >> > index 1ee8351c21f..87f602e7ead 100644
>>> >> > --- a/gcc/config/i386/i386.c
>>> >> > +++ b/gcc/config/i386/i386.c
>>> >> > @@ -6511,6 +6511,10 @@ ix86_option_override_internal (bool main_args_p,
>>> >> > ix86_tune_cost->l2_cache_size,
>>> >> > opts->x_param_values,
>>> >> > opts_set->x_param_values);
>>> >> > + maybe_set_param_value (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
>>> >> > + 35,
>>> >> > + opts->x_param_values,
>>> >> > + opts_set->x_param_values);
>>> >> >
>>> >> > /* Enable sw prefetching at -O3 for CPUS that prefetching is helpful. */
>>> >> > if (opts->x_flag_prefetch_loop_arrays < 0
>>> >> > diff --git a/gcc/expr.c b/gcc/expr.c
>>> >> > index 134ee731c29..dff24e7f166 100644
>>> >> > --- a/gcc/expr.c
>>> >> > +++ b/gcc/expr.c
>>> >> > @@ -61,7 +61,8 @@ along with GCC; see the file COPYING3. If not see
>>> >> > #include "tree-chkp.h"
>>> >> > #include "rtl-chkp.h"
>>> >> > #include "ccmp.h"
>>> >> > -
>>> >> > +#include "params.h"
>>> >> > +#include "tree-sra.h"
>>> >> >
>>> >> > /* If this is nonzero, we do not bother generating VOLATILE
>>> >> > around volatile memory references, and we are willing to
>>> >> > @@ -5340,6 +5341,80 @@ emit_storent_insn (rtx to, rtx from)
>>> >> > return maybe_expand_insn (code, 2, ops);
>>> >> > }
>>> >> >
>>> >> > +/* Generate code for copying data of type TYPE at SOURCE plus OFFSET to TARGET
>>> >> > + plus OFFSET, but do so element-wise and/or field-wise for each record and
>>> >> > + array within TYPE. TYPE must either be a register type or an aggregate
>>> >> > + complying with scalarizable_type_p.
>>> >> > +
>>> >> > + If CALL_PARAM_P is nonzero, this is a store into a call param on the
>>> >> > + stack, and block moves may need to be treated specially. */
>>> >> > +
>>> >> > +static void
>>> >> > +emit_move_elementwise (tree type, rtx target, rtx source, HOST_WIDE_INT offset,
>>> >> > + int call_param_p)
>>> >> > +{
>>> >> > + switch (TREE_CODE (type))
>>> >> > + {
>>> >> > + case RECORD_TYPE:
>>> >> > + for (tree fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
>>> >> > + if (TREE_CODE (fld) == FIELD_DECL)
>>> >> > + {
>>> >> > + HOST_WIDE_INT fld_offset = offset + int_bit_position (fld);
>>> >> > + tree ft = TREE_TYPE (fld);
>>> >> > + emit_move_elementwise (ft, target, source, fld_offset,
>>> >> > + call_param_p);
>>> >> > + }
>>> >> > + break;
>>> >> > +
>>> >> > + case ARRAY_TYPE:
>>> >> > + {
>>> >> > + tree elem_type = TREE_TYPE (type);
>>> >> > + HOST_WIDE_INT el_size = tree_to_shwi (TYPE_SIZE (elem_type));
>>> >> > + gcc_assert (el_size > 0);
>>> >> > +
>>> >> > + offset_int idx, max;
>>> >> > + /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
>>> >> > + if (extract_min_max_idx_from_array (type, &idx, &max))
>>> >> > + {
>>> >> > + HOST_WIDE_INT el_offset = offset;
>>> >> > + for (; idx <= max; ++idx)
>>> >> > + {
>>> >> > + emit_move_elementwise (elem_type, target, source, el_offset,
>>> >> > + call_param_p);
>>> >> > + el_offset += el_size;
>>> >> > + }
>>> >> > + }
>>> >> > + }
>>> >> > + break;
>>> >> > + default:
>>> >> > + machine_mode mode = TYPE_MODE (type);
>>> >> > +
>>> >> > + rtx ntgt = adjust_address (target, mode, offset / BITS_PER_UNIT);
>>> >> > + rtx nsrc = adjust_address (source, mode, offset / BITS_PER_UNIT);
>>> >> > +
>>> >> > + /* TODO: Figure out whether the following is actually necessary. */
>>> >> > + if (target == ntgt)
>>> >> > + ntgt = copy_rtx (target);
>>> >> > + if (source == nsrc)
>>> >> > + nsrc = copy_rtx (source);
>>> >> > +
>>> >> > + gcc_assert (mode != VOIDmode);
>>> >> > + if (mode != BLKmode)
>>> >> > + emit_move_insn (ntgt, nsrc);
>>> >> > + else
>>> >> > + {
>>> >> > + /* For example vector gimple registers can end up here. */
>>> >> > + rtx size = expand_expr (TYPE_SIZE_UNIT (type), NULL_RTX,
>>> >> > + TYPE_MODE (sizetype), EXPAND_NORMAL);
>>> >> > + emit_block_move (ntgt, nsrc, size,
>>> >> > + (call_param_p
>>> >> > + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
>>> >> > + }
>>> >> > + break;
>>> >> > + }
>>> >> > + return;
>>> >> > +}
>>> >> > +
>>> >> > /* Generate code for computing expression EXP,
>>> >> > and storing the value into TARGET.
>>> >> >
>>> >> > @@ -5713,9 +5788,29 @@ store_expr_with_bounds (tree exp, rtx target, int call_param_p,
>>> >> > emit_group_store (target, temp, TREE_TYPE (exp),
>>> >> > int_size_in_bytes (TREE_TYPE (exp)));
>>> >> > else if (GET_MODE (temp) == BLKmode)
>>> >> > - emit_block_move (target, temp, expr_size (exp),
>>> >> > - (call_param_p
>>> >> > - ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
>>> >> > + {
>>> >> > + /* Copying smallish BLKmode structures with emit_block_move and thus
>>> >> > + by-pieces can result in store-to-load stalls. So copy some simple
>>> >> > + small aggregates element or field-wise. */
>>> >> > + if (GET_MODE (target) == BLKmode
>>> >> > + && AGGREGATE_TYPE_P (TREE_TYPE (exp))
>>> >> > + && !TREE_ADDRESSABLE (TREE_TYPE (exp))
>>> >> > + && tree_fits_shwi_p (TYPE_SIZE (TREE_TYPE (exp)))
>>> >> > + && (tree_to_shwi (TYPE_SIZE (TREE_TYPE (exp)))
>>> >> > + <= (PARAM_VALUE (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY)
>>> >> > + * BITS_PER_UNIT))
>>> >> > + && simple_mix_of_records_and_arrays_p (TREE_TYPE (exp), false))
>>> >> > + {
>>> >> > + /* FIXME: Can this happen? What would it mean? */
>>> >> > + gcc_assert (!reverse);
>>> >> > + emit_move_elementwise (TREE_TYPE (exp), target, temp, 0,
>>> >> > + call_param_p);
>>> >> > + }
>>> >> > + else
>>> >> > + emit_block_move (target, temp, expr_size (exp),
>>> >> > + (call_param_p
>>> >> > + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
>>> >> > + }
>>> >> > /* If we emit a nontemporal store, there is nothing else to do. */
>>> >> > else if (nontemporal && emit_storent_insn (target, temp))
>>> >> > ;
>>> >> > diff --git a/gcc/ipa-cp.c b/gcc/ipa-cp.c
>>> >> > index 6b3d8d7364c..7d6019bbd30 100644
>>> >> > --- a/gcc/ipa-cp.c
>>> >> > +++ b/gcc/ipa-cp.c
>>> >> > @@ -124,6 +124,7 @@ along with GCC; see the file COPYING3. If not see
>>> >> > #include "tree-ssa-ccp.h"
>>> >> > #include "stringpool.h"
>>> >> > #include "attribs.h"
>>> >> > +#include "tree-sra.h"
>>> >> >
>>> >> > template <typename valtype> class ipcp_value;
>>> >> >
>>> >> > diff --git a/gcc/ipa-prop.h b/gcc/ipa-prop.h
>>> >> > index fa5bed49ee0..2313cc884ed 100644
>>> >> > --- a/gcc/ipa-prop.h
>>> >> > +++ b/gcc/ipa-prop.h
>>> >> > @@ -877,10 +877,6 @@ ipa_parm_adjustment *ipa_get_adjustment_candidate (tree **, bool *,
>>> >> > void ipa_release_body_info (struct ipa_func_body_info *);
>>> >> > tree ipa_get_callee_param_type (struct cgraph_edge *e, int i);
>>> >> >
>>> >> > -/* From tree-sra.c: */
>>> >> > -tree build_ref_for_offset (location_t, tree, HOST_WIDE_INT, bool, tree,
>>> >> > - gimple_stmt_iterator *, bool);
>>> >> > -
>>> >> > /* In ipa-cp.c */
>>> >> > void ipa_cp_c_finalize (void);
>>> >> >
>>> >> > diff --git a/gcc/params.def b/gcc/params.def
>>> >> > index e55afc28053..5e19f1414a0 100644
>>> >> > --- a/gcc/params.def
>>> >> > +++ b/gcc/params.def
>>> >> > @@ -1294,6 +1294,12 @@ DEFPARAM (PARAM_VECT_EPILOGUES_NOMASK,
>>> >> > "Enable loop epilogue vectorization using smaller vector size.",
>>> >> > 0, 0, 1)
>>> >> >
>>> >> > +DEFPARAM (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
>>> >> > + "max-size-for-elementwise-copy",
>>> >> > + "Maximum size in bytes of a structure or array to by considered for "
>>> >> > + "copying by its individual fields or elements",
>>> >> > + 0, 0, 512)
>>> >> > +
>>> >> > /*
>>> >> >
>>> >> > Local variables:
>>> >> > diff --git a/gcc/testsuite/gcc.target/i386/pr80689-1.c b/gcc/testsuite/gcc.target/i386/pr80689-1.c
>>> >> > new file mode 100644
>>> >> > index 00000000000..4156d4fba45
>>> >> > --- /dev/null
>>> >> > +++ b/gcc/testsuite/gcc.target/i386/pr80689-1.c
>>> >> > @@ -0,0 +1,38 @@
>>> >> > +/* { dg-do compile } */
>>> >> > +/* { dg-options "-O2" } */
>>> >> > +
>>> >> > +typedef struct st1
>>> >> > +{
>>> >> > + long unsigned int a,b;
>>> >> > + long int c,d;
>>> >> > +}R;
>>> >> > +
>>> >> > +typedef struct st2
>>> >> > +{
>>> >> > + int t;
>>> >> > + R reg;
>>> >> > +}N;
>>> >> > +
>>> >> > +void Set (const R *region, N *n_info );
>>> >> > +
>>> >> > +void test(N *n_obj ,const long unsigned int a, const long unsigned int b, const long int c,const long int d)
>>> >> > +{
>>> >> > + R reg;
>>> >> > +
>>> >> > + reg.a=a;
>>> >> > + reg.b=b;
>>> >> > + reg.c=c;
>>> >> > + reg.d=d;
>>> >> > + Set (®, n_obj);
>>> >> > +
>>> >> > +}
>>> >> > +
>>> >> > +void Set (const R *reg, N *n_obj )
>>> >> > +{
>>> >> > + n_obj->reg=(*reg);
>>> >> > +}
>>> >> > +
>>> >> > +
>>> >> > +/* { dg-final { scan-assembler-not "%(x|y|z)mm\[0-9\]+" } } */
>>> >> > +/* { dg-final { scan-assembler-not "movdqu" } } */
>>> >> > +/* { dg-final { scan-assembler-not "movups" } } */
>>> >> > diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
>>> >> > index bac593951e7..ade97964205 100644
>>> >> > --- a/gcc/tree-sra.c
>>> >> > +++ b/gcc/tree-sra.c
>>> >> > @@ -104,6 +104,7 @@ along with GCC; see the file COPYING3. If not see
>>> >> > #include "ipa-fnsummary.h"
>>> >> > #include "ipa-utils.h"
>>> >> > #include "builtins.h"
>>> >> > +#include "tree-sra.h"
>>> >> >
>>> >> > /* Enumeration of all aggregate reductions we can do. */
>>> >> > enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
>>> >> > @@ -952,14 +953,14 @@ create_access (tree expr, gimple *stmt, bool write)
>>> >> > }
>>> >> >
>>> >> >
>>> >> > -/* Return true iff TYPE is scalarizable - i.e. a RECORD_TYPE or fixed-length
>>> >> > - ARRAY_TYPE with fields that are either of gimple register types (excluding
>>> >> > - bit-fields) or (recursively) scalarizable types. CONST_DECL must be true if
>>> >> > - we are considering a decl from constant pool. If it is false, char arrays
>>> >> > - will be refused. */
>>> >> > +/* Return true if TYPE consists of RECORD_TYPE or fixed-length ARRAY_TYPE with
>>> >> > + fields/elements that are not bit-fields and are either register types or
>>> >> > + recursively comply with simple_mix_of_records_and_arrays_p. Furthermore, if
>>> >> > + ALLOW_CHAR_ARRAYS is false, the function will return false also if TYPE
>>> >> > + contains an array of elements that only have one byte. */
>>> >> >
>>> >> > -static bool
>>> >> > -scalarizable_type_p (tree type, bool const_decl)
>>> >> > +bool
>>> >> > +simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays)
>>> >> > {
>>> >> > gcc_assert (!is_gimple_reg_type (type));
>>> >> > if (type_contains_placeholder_p (type))
>>> >> > @@ -977,7 +978,7 @@ scalarizable_type_p (tree type, bool const_decl)
>>> >> > return false;
>>> >> >
>>> >> > if (!is_gimple_reg_type (ft)
>>> >> > - && !scalarizable_type_p (ft, const_decl))
>>> >> > + && !simple_mix_of_records_and_arrays_p (ft, allow_char_arrays))
>>> >> > return false;
>>> >> > }
>>> >> >
>>> >> > @@ -986,7 +987,7 @@ scalarizable_type_p (tree type, bool const_decl)
>>> >> > case ARRAY_TYPE:
>>> >> > {
>>> >> > HOST_WIDE_INT min_elem_size;
>>> >> > - if (const_decl)
>>> >> > + if (allow_char_arrays)
>>> >> > min_elem_size = 0;
>>> >> > else
>>> >> > min_elem_size = BITS_PER_UNIT;
>>> >> > @@ -1008,7 +1009,7 @@ scalarizable_type_p (tree type, bool const_decl)
>>> >> >
>>> >> > tree elem = TREE_TYPE (type);
>>> >> > if (!is_gimple_reg_type (elem)
>>> >> > - && !scalarizable_type_p (elem, const_decl))
>>> >> > + && !simple_mix_of_records_and_arrays_p (elem, allow_char_arrays))
>>> >> > return false;
>>> >> > return true;
>>> >> > }
>>> >> > @@ -1017,10 +1018,38 @@ scalarizable_type_p (tree type, bool const_decl)
>>> >> > }
>>> >> > }
>>> >> >
>>> >> > -static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree, tree);
>>> >> > +static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree,
>>> >> > + tree);
>>> >> > +
>>> >> > +/* For a given array TYPE, return false if its domain does not have any maximum
>>> >> > + value. Otherwise calculate MIN and MAX indices of the first and the last
>>> >> > + element. */
>>> >> > +
>>> >> > +bool
>>> >> > +extract_min_max_idx_from_array (tree type, offset_int *min, offset_int *max)
>>> >> > +{
>>> >> > + tree domain = TYPE_DOMAIN (type);
>>> >> > + tree minidx = TYPE_MIN_VALUE (domain);
>>> >> > + gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
>>> >> > + tree maxidx = TYPE_MAX_VALUE (domain);
>>> >> > + if (!maxidx)
>>> >> > + return false;
>>> >> > + gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
>>> >> > +
>>> >> > + /* MINIDX and MAXIDX are inclusive, and must be interpreted in
>>> >> > + DOMAIN (e.g. signed int, whereas min/max may be size_int). */
>>> >> > + *min = wi::to_offset (minidx);
>>> >> > + *max = wi::to_offset (maxidx);
>>> >> > + if (!TYPE_UNSIGNED (domain))
>>> >> > + {
>>> >> > + *min = wi::sext (*min, TYPE_PRECISION (domain));
>>> >> > + *max = wi::sext (*max, TYPE_PRECISION (domain));
>>> >> > + }
>>> >> > + return true;
>>> >> > +}
>>> >> >
>>> >> > /* Create total_scalarization accesses for all scalar fields of a member
>>> >> > - of type DECL_TYPE conforming to scalarizable_type_p. BASE
>>> >> > + of type DECL_TYPE conforming to simple_mix_of_records_and_arrays_p. BASE
>>> >> > must be the top-most VAR_DECL representing the variable; within that,
>>> >> > OFFSET locates the member and REF must be the memory reference expression for
>>> >> > the member. */
>>> >> > @@ -1047,27 +1076,14 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
>>> >> > {
>>> >> > tree elemtype = TREE_TYPE (decl_type);
>>> >> > tree elem_size = TYPE_SIZE (elemtype);
>>> >> > - gcc_assert (elem_size && tree_fits_shwi_p (elem_size));
>>> >> > HOST_WIDE_INT el_size = tree_to_shwi (elem_size);
>>> >> > gcc_assert (el_size > 0);
>>> >> >
>>> >> > - tree minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (decl_type));
>>> >> > - gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
>>> >> > - tree maxidx = TYPE_MAX_VALUE (TYPE_DOMAIN (decl_type));
>>> >> > + offset_int idx, max;
>>> >> > /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
>>> >> > - if (maxidx)
>>> >> > + if (extract_min_max_idx_from_array (decl_type, &idx, &max))
>>> >> > {
>>> >> > - gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
>>> >> > tree domain = TYPE_DOMAIN (decl_type);
>>> >> > - /* MINIDX and MAXIDX are inclusive, and must be interpreted in
>>> >> > - DOMAIN (e.g. signed int, whereas min/max may be size_int). */
>>> >> > - offset_int idx = wi::to_offset (minidx);
>>> >> > - offset_int max = wi::to_offset (maxidx);
>>> >> > - if (!TYPE_UNSIGNED (domain))
>>> >> > - {
>>> >> > - idx = wi::sext (idx, TYPE_PRECISION (domain));
>>> >> > - max = wi::sext (max, TYPE_PRECISION (domain));
>>> >> > - }
>>> >> > for (int el_off = offset; idx <= max; ++idx)
>>> >> > {
>>> >> > tree nref = build4 (ARRAY_REF, elemtype,
>>> >> > @@ -1088,10 +1104,10 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
>>> >> > }
>>> >> >
>>> >> > /* Create total_scalarization accesses for a member of type TYPE, which must
>>> >> > - satisfy either is_gimple_reg_type or scalarizable_type_p. BASE must be the
>>> >> > - top-most VAR_DECL representing the variable; within that, POS and SIZE locate
>>> >> > - the member, REVERSE gives its torage order. and REF must be the reference
>>> >> > - expression for it. */
>>> >> > + satisfy either is_gimple_reg_type or simple_mix_of_records_and_arrays_p.
>>> >> > + BASE must be the top-most VAR_DECL representing the variable; within that,
>>> >> > + POS and SIZE locate the member, REVERSE gives its torage order. and REF must
>>> >> > + be the reference expression for it. */
>>> >> >
>>> >> > static void
>>> >> > scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
>>> >> > @@ -1111,7 +1127,8 @@ scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
>>> >> > }
>>> >> >
>>> >> > /* Create a total_scalarization access for VAR as a whole. VAR must be of a
>>> >> > - RECORD_TYPE or ARRAY_TYPE conforming to scalarizable_type_p. */
>>> >> > + RECORD_TYPE or ARRAY_TYPE conforming to
>>> >> > + simple_mix_of_records_and_arrays_p. */
>>> >> >
>>> >> > static void
>>> >> > create_total_scalarization_access (tree var)
>>> >> > @@ -2803,8 +2820,9 @@ analyze_all_variable_accesses (void)
>>> >> > {
>>> >> > tree var = candidate (i);
>>> >> >
>>> >> > - if (VAR_P (var) && scalarizable_type_p (TREE_TYPE (var),
>>> >> > - constant_decl_p (var)))
>>> >> > + if (VAR_P (var)
>>> >> > + && simple_mix_of_records_and_arrays_p (TREE_TYPE (var),
>>> >> > + constant_decl_p (var)))
>>> >> > {
>>> >> > if (tree_to_uhwi (TYPE_SIZE (TREE_TYPE (var)))
>>> >> > <= max_scalarization_size)
>>> >> > diff --git a/gcc/tree-sra.h b/gcc/tree-sra.h
>>> >> > new file mode 100644
>>> >> > index 00000000000..dc901385994
>>> >> > --- /dev/null
>>> >> > +++ b/gcc/tree-sra.h
>>> >> > @@ -0,0 +1,33 @@
>>> >> > +/* tree-sra.h - Run-time parameters.
>>> >> > + Copyright (C) 2017 Free Software Foundation, Inc.
>>> >> > +
>>> >> > +This file is part of GCC.
>>> >> > +
>>> >> > +GCC is free software; you can redistribute it and/or modify it under
>>> >> > +the terms of the GNU General Public License as published by the Free
>>> >> > +Software Foundation; either version 3, or (at your option) any later
>>> >> > +version.
>>> >> > +
>>> >> > +GCC is distributed in the hope that it will be useful, but WITHOUT ANY
>>> >> > +WARRANTY; without even the implied warranty of MERCHANTABILITY or
>>> >> > +FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
>>> >> > +for more details.
>>> >> > +
>>> >> > +You should have received a copy of the GNU General Public License
>>> >> > +along with GCC; see the file COPYING3. If not see
>>> >> > +<http://www.gnu.org/licenses/>. */
>>> >> > +
>>> >> > +#ifndef TREE_SRA_H
>>> >> > +#define TREE_SRA_H
>>> >> > +
>>> >> > +
>>> >> > +bool simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays);
>>> >> > +bool extract_min_max_idx_from_array (tree type, offset_int *idx,
>>> >> > + offset_int *max);
>>> >> > +tree build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
>>> >> > + bool reverse, tree exp_type,
>>> >> > + gimple_stmt_iterator *gsi, bool insert_after);
>>> >> > +
>>> >> > +
>>> >> > +
>>> >> > +#endif /* TREE_SRA_H */
>>> >> > --
>>> >> > 2.14.1
>>> >> >
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-10-26 14:53 ` Richard Biener
2017-10-26 15:12 ` Richard Biener
@ 2017-10-27 12:27 ` Jan Hubicka
1 sibling, 0 replies; 26+ messages in thread
From: Jan Hubicka @ 2017-10-27 12:27 UTC (permalink / raw)
To: Richard Biener; +Cc: GCC Patches
> On Thu, Oct 26, 2017 at 2:55 PM, Jan Hubicka <hubicka@ucw.cz> wrote:
> >> I think the limit should be on the number of generated copies and not
> >> the overall size of the structure... If the struct were composed of
> >> 32 individual chars we wouldn't want to emit 32 loads and 32 stores...
> >>
> >> I wonder how rep; movb; interacts with store to load forwarding? Is
> >> that maybe optimized well on some archs? movb should always
> >> forward and wasn't the setup cost for small N reasonable on modern
> >> CPUs?
> >
> > rep mov is win over loop for blocks over 128bytes on core, for blocks in rage
> > 24-128 on zen. This is w/o store/load forwarding, but I doubt those provide
> > a cheap way around.
> >
> >>
> >> It probably depends on the width of the entries in the store buffer,
> >> if they appear in-order and the alignment of the stores (if they are larger than
> >> 8 bytes they are surely aligned). IIRC CPUs had smaller store buffer
> >> entries than cache line size.
> >>
> >> Given that load bandwith is usually higher than store bandwith it
> >> might make sense to do the store combining in our copying sequence,
> >> like for the 8 byte entry case use sth like
> >>
> >> movq 0(%eax), %xmm0
> >> movhps 8(%eax), %xmm0 // or vpinsert
> >> mov[au]ps %xmm0, 0%(ebx)
> >> ...
> >>
> >> thus do two loads per store and perform the stores in wider
> >> mode?
> >
> > This may be somewhat faster indeed. I am not sure if store to load
> > forwarding will work for the later half when read again by halves.
> > It would not happen on older CPUs :)
>
> Yes, forwarding larger stores to smaller loads generally works fine
> since forever with the usual restrictions of alignment/size being
> power of two "halves".
>
> The question is of course what to do for 4 byte or smaller elements or
> mixed size elements. We can do zero-extending loads
> (do we have them for QI, HI mode loads as well?) and
> do shift and or's. I'm quite sure the CPUs wouldn't like to
> see vpinsert's of different vector mode destinations. So it
> would be 8 byte stores from GPRs and values built up via
> shift & or.
>
> As said, the important part is that IIRC CPUs can usually
> have more loads in flight than stores. Esp. Bulldozer
> with the split core was store buffer size limited (but it
> could do merging of store buffer entries IIRC).
In a way this seems an independent optimization to me
(store forwarding) because for sure this can work for user
code which does not originate from copy sequence.
Seems like something bit tricky to implement on top of RTL
though.
Honza
>
> Richard.
>
> > Honza
> >>
> >> As said a general concern was you not copying padding. If you
> >> put this into an even more common place you surely will break
> >> stuff, no?
> >>
> >> Richard.
> >>
> >> >
> >> > Martin
> >> >
> >> >
> >> >>
> >> >> Richard.
> >> >>
> >> >> > Martin
> >> >> >
> >> >> >
> >> >> > 2017-10-12 Martin Jambor <mjambor@suse.cz>
> >> >> >
> >> >> > PR target/80689
> >> >> > * tree-sra.h: New file.
> >> >> > * ipa-prop.h: Moved declaration of build_ref_for_offset to
> >> >> > tree-sra.h.
> >> >> > * expr.c: Include params.h and tree-sra.h.
> >> >> > (emit_move_elementwise): New function.
> >> >> > (store_expr_with_bounds): Optionally use it.
> >> >> > * ipa-cp.c: Include tree-sra.h.
> >> >> > * params.def (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY): New.
> >> >> > * config/i386/i386.c (ix86_option_override_internal): Set
> >> >> > PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY to 35.
> >> >> > * tree-sra.c: Include tree-sra.h.
> >> >> > (scalarizable_type_p): Renamed to
> >> >> > simple_mix_of_records_and_arrays_p, made public, renamed the
> >> >> > second parameter to allow_char_arrays.
> >> >> > (extract_min_max_idx_from_array): New function.
> >> >> > (completely_scalarize): Moved bits of the function to
> >> >> > extract_min_max_idx_from_array.
> >> >> >
> >> >> > testsuite/
> >> >> > * gcc.target/i386/pr80689-1.c: New test.
> >> >> > ---
> >> >> > gcc/config/i386/i386.c | 4 ++
> >> >> > gcc/expr.c | 103 ++++++++++++++++++++++++++++--
> >> >> > gcc/ipa-cp.c | 1 +
> >> >> > gcc/ipa-prop.h | 4 --
> >> >> > gcc/params.def | 6 ++
> >> >> > gcc/testsuite/gcc.target/i386/pr80689-1.c | 38 +++++++++++
> >> >> > gcc/tree-sra.c | 86 +++++++++++++++----------
> >> >> > gcc/tree-sra.h | 33 ++++++++++
> >> >> > 8 files changed, 233 insertions(+), 42 deletions(-)
> >> >> > create mode 100644 gcc/testsuite/gcc.target/i386/pr80689-1.c
> >> >> > create mode 100644 gcc/tree-sra.h
> >> >> >
> >> >> > diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c
> >> >> > index 1ee8351c21f..87f602e7ead 100644
> >> >> > --- a/gcc/config/i386/i386.c
> >> >> > +++ b/gcc/config/i386/i386.c
> >> >> > @@ -6511,6 +6511,10 @@ ix86_option_override_internal (bool main_args_p,
> >> >> > ix86_tune_cost->l2_cache_size,
> >> >> > opts->x_param_values,
> >> >> > opts_set->x_param_values);
> >> >> > + maybe_set_param_value (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
> >> >> > + 35,
> >> >> > + opts->x_param_values,
> >> >> > + opts_set->x_param_values);
> >> >> >
> >> >> > /* Enable sw prefetching at -O3 for CPUS that prefetching is helpful. */
> >> >> > if (opts->x_flag_prefetch_loop_arrays < 0
> >> >> > diff --git a/gcc/expr.c b/gcc/expr.c
> >> >> > index 134ee731c29..dff24e7f166 100644
> >> >> > --- a/gcc/expr.c
> >> >> > +++ b/gcc/expr.c
> >> >> > @@ -61,7 +61,8 @@ along with GCC; see the file COPYING3. If not see
> >> >> > #include "tree-chkp.h"
> >> >> > #include "rtl-chkp.h"
> >> >> > #include "ccmp.h"
> >> >> > -
> >> >> > +#include "params.h"
> >> >> > +#include "tree-sra.h"
> >> >> >
> >> >> > /* If this is nonzero, we do not bother generating VOLATILE
> >> >> > around volatile memory references, and we are willing to
> >> >> > @@ -5340,6 +5341,80 @@ emit_storent_insn (rtx to, rtx from)
> >> >> > return maybe_expand_insn (code, 2, ops);
> >> >> > }
> >> >> >
> >> >> > +/* Generate code for copying data of type TYPE at SOURCE plus OFFSET to TARGET
> >> >> > + plus OFFSET, but do so element-wise and/or field-wise for each record and
> >> >> > + array within TYPE. TYPE must either be a register type or an aggregate
> >> >> > + complying with scalarizable_type_p.
> >> >> > +
> >> >> > + If CALL_PARAM_P is nonzero, this is a store into a call param on the
> >> >> > + stack, and block moves may need to be treated specially. */
> >> >> > +
> >> >> > +static void
> >> >> > +emit_move_elementwise (tree type, rtx target, rtx source, HOST_WIDE_INT offset,
> >> >> > + int call_param_p)
> >> >> > +{
> >> >> > + switch (TREE_CODE (type))
> >> >> > + {
> >> >> > + case RECORD_TYPE:
> >> >> > + for (tree fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
> >> >> > + if (TREE_CODE (fld) == FIELD_DECL)
> >> >> > + {
> >> >> > + HOST_WIDE_INT fld_offset = offset + int_bit_position (fld);
> >> >> > + tree ft = TREE_TYPE (fld);
> >> >> > + emit_move_elementwise (ft, target, source, fld_offset,
> >> >> > + call_param_p);
> >> >> > + }
> >> >> > + break;
> >> >> > +
> >> >> > + case ARRAY_TYPE:
> >> >> > + {
> >> >> > + tree elem_type = TREE_TYPE (type);
> >> >> > + HOST_WIDE_INT el_size = tree_to_shwi (TYPE_SIZE (elem_type));
> >> >> > + gcc_assert (el_size > 0);
> >> >> > +
> >> >> > + offset_int idx, max;
> >> >> > + /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
> >> >> > + if (extract_min_max_idx_from_array (type, &idx, &max))
> >> >> > + {
> >> >> > + HOST_WIDE_INT el_offset = offset;
> >> >> > + for (; idx <= max; ++idx)
> >> >> > + {
> >> >> > + emit_move_elementwise (elem_type, target, source, el_offset,
> >> >> > + call_param_p);
> >> >> > + el_offset += el_size;
> >> >> > + }
> >> >> > + }
> >> >> > + }
> >> >> > + break;
> >> >> > + default:
> >> >> > + machine_mode mode = TYPE_MODE (type);
> >> >> > +
> >> >> > + rtx ntgt = adjust_address (target, mode, offset / BITS_PER_UNIT);
> >> >> > + rtx nsrc = adjust_address (source, mode, offset / BITS_PER_UNIT);
> >> >> > +
> >> >> > + /* TODO: Figure out whether the following is actually necessary. */
> >> >> > + if (target == ntgt)
> >> >> > + ntgt = copy_rtx (target);
> >> >> > + if (source == nsrc)
> >> >> > + nsrc = copy_rtx (source);
> >> >> > +
> >> >> > + gcc_assert (mode != VOIDmode);
> >> >> > + if (mode != BLKmode)
> >> >> > + emit_move_insn (ntgt, nsrc);
> >> >> > + else
> >> >> > + {
> >> >> > + /* For example vector gimple registers can end up here. */
> >> >> > + rtx size = expand_expr (TYPE_SIZE_UNIT (type), NULL_RTX,
> >> >> > + TYPE_MODE (sizetype), EXPAND_NORMAL);
> >> >> > + emit_block_move (ntgt, nsrc, size,
> >> >> > + (call_param_p
> >> >> > + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> >> >> > + }
> >> >> > + break;
> >> >> > + }
> >> >> > + return;
> >> >> > +}
> >> >> > +
> >> >> > /* Generate code for computing expression EXP,
> >> >> > and storing the value into TARGET.
> >> >> >
> >> >> > @@ -5713,9 +5788,29 @@ store_expr_with_bounds (tree exp, rtx target, int call_param_p,
> >> >> > emit_group_store (target, temp, TREE_TYPE (exp),
> >> >> > int_size_in_bytes (TREE_TYPE (exp)));
> >> >> > else if (GET_MODE (temp) == BLKmode)
> >> >> > - emit_block_move (target, temp, expr_size (exp),
> >> >> > - (call_param_p
> >> >> > - ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> >> >> > + {
> >> >> > + /* Copying smallish BLKmode structures with emit_block_move and thus
> >> >> > + by-pieces can result in store-to-load stalls. So copy some simple
> >> >> > + small aggregates element or field-wise. */
> >> >> > + if (GET_MODE (target) == BLKmode
> >> >> > + && AGGREGATE_TYPE_P (TREE_TYPE (exp))
> >> >> > + && !TREE_ADDRESSABLE (TREE_TYPE (exp))
> >> >> > + && tree_fits_shwi_p (TYPE_SIZE (TREE_TYPE (exp)))
> >> >> > + && (tree_to_shwi (TYPE_SIZE (TREE_TYPE (exp)))
> >> >> > + <= (PARAM_VALUE (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY)
> >> >> > + * BITS_PER_UNIT))
> >> >> > + && simple_mix_of_records_and_arrays_p (TREE_TYPE (exp), false))
> >> >> > + {
> >> >> > + /* FIXME: Can this happen? What would it mean? */
> >> >> > + gcc_assert (!reverse);
> >> >> > + emit_move_elementwise (TREE_TYPE (exp), target, temp, 0,
> >> >> > + call_param_p);
> >> >> > + }
> >> >> > + else
> >> >> > + emit_block_move (target, temp, expr_size (exp),
> >> >> > + (call_param_p
> >> >> > + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> >> >> > + }
> >> >> > /* If we emit a nontemporal store, there is nothing else to do. */
> >> >> > else if (nontemporal && emit_storent_insn (target, temp))
> >> >> > ;
> >> >> > diff --git a/gcc/ipa-cp.c b/gcc/ipa-cp.c
> >> >> > index 6b3d8d7364c..7d6019bbd30 100644
> >> >> > --- a/gcc/ipa-cp.c
> >> >> > +++ b/gcc/ipa-cp.c
> >> >> > @@ -124,6 +124,7 @@ along with GCC; see the file COPYING3. If not see
> >> >> > #include "tree-ssa-ccp.h"
> >> >> > #include "stringpool.h"
> >> >> > #include "attribs.h"
> >> >> > +#include "tree-sra.h"
> >> >> >
> >> >> > template <typename valtype> class ipcp_value;
> >> >> >
> >> >> > diff --git a/gcc/ipa-prop.h b/gcc/ipa-prop.h
> >> >> > index fa5bed49ee0..2313cc884ed 100644
> >> >> > --- a/gcc/ipa-prop.h
> >> >> > +++ b/gcc/ipa-prop.h
> >> >> > @@ -877,10 +877,6 @@ ipa_parm_adjustment *ipa_get_adjustment_candidate (tree **, bool *,
> >> >> > void ipa_release_body_info (struct ipa_func_body_info *);
> >> >> > tree ipa_get_callee_param_type (struct cgraph_edge *e, int i);
> >> >> >
> >> >> > -/* From tree-sra.c: */
> >> >> > -tree build_ref_for_offset (location_t, tree, HOST_WIDE_INT, bool, tree,
> >> >> > - gimple_stmt_iterator *, bool);
> >> >> > -
> >> >> > /* In ipa-cp.c */
> >> >> > void ipa_cp_c_finalize (void);
> >> >> >
> >> >> > diff --git a/gcc/params.def b/gcc/params.def
> >> >> > index e55afc28053..5e19f1414a0 100644
> >> >> > --- a/gcc/params.def
> >> >> > +++ b/gcc/params.def
> >> >> > @@ -1294,6 +1294,12 @@ DEFPARAM (PARAM_VECT_EPILOGUES_NOMASK,
> >> >> > "Enable loop epilogue vectorization using smaller vector size.",
> >> >> > 0, 0, 1)
> >> >> >
> >> >> > +DEFPARAM (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
> >> >> > + "max-size-for-elementwise-copy",
> >> >> > + "Maximum size in bytes of a structure or array to by considered for "
> >> >> > + "copying by its individual fields or elements",
> >> >> > + 0, 0, 512)
> >> >> > +
> >> >> > /*
> >> >> >
> >> >> > Local variables:
> >> >> > diff --git a/gcc/testsuite/gcc.target/i386/pr80689-1.c b/gcc/testsuite/gcc.target/i386/pr80689-1.c
> >> >> > new file mode 100644
> >> >> > index 00000000000..4156d4fba45
> >> >> > --- /dev/null
> >> >> > +++ b/gcc/testsuite/gcc.target/i386/pr80689-1.c
> >> >> > @@ -0,0 +1,38 @@
> >> >> > +/* { dg-do compile } */
> >> >> > +/* { dg-options "-O2" } */
> >> >> > +
> >> >> > +typedef struct st1
> >> >> > +{
> >> >> > + long unsigned int a,b;
> >> >> > + long int c,d;
> >> >> > +}R;
> >> >> > +
> >> >> > +typedef struct st2
> >> >> > +{
> >> >> > + int t;
> >> >> > + R reg;
> >> >> > +}N;
> >> >> > +
> >> >> > +void Set (const R *region, N *n_info );
> >> >> > +
> >> >> > +void test(N *n_obj ,const long unsigned int a, const long unsigned int b, const long int c,const long int d)
> >> >> > +{
> >> >> > + R reg;
> >> >> > +
> >> >> > + reg.a=a;
> >> >> > + reg.b=b;
> >> >> > + reg.c=c;
> >> >> > + reg.d=d;
> >> >> > + Set (®, n_obj);
> >> >> > +
> >> >> > +}
> >> >> > +
> >> >> > +void Set (const R *reg, N *n_obj )
> >> >> > +{
> >> >> > + n_obj->reg=(*reg);
> >> >> > +}
> >> >> > +
> >> >> > +
> >> >> > +/* { dg-final { scan-assembler-not "%(x|y|z)mm\[0-9\]+" } } */
> >> >> > +/* { dg-final { scan-assembler-not "movdqu" } } */
> >> >> > +/* { dg-final { scan-assembler-not "movups" } } */
> >> >> > diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
> >> >> > index bac593951e7..ade97964205 100644
> >> >> > --- a/gcc/tree-sra.c
> >> >> > +++ b/gcc/tree-sra.c
> >> >> > @@ -104,6 +104,7 @@ along with GCC; see the file COPYING3. If not see
> >> >> > #include "ipa-fnsummary.h"
> >> >> > #include "ipa-utils.h"
> >> >> > #include "builtins.h"
> >> >> > +#include "tree-sra.h"
> >> >> >
> >> >> > /* Enumeration of all aggregate reductions we can do. */
> >> >> > enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
> >> >> > @@ -952,14 +953,14 @@ create_access (tree expr, gimple *stmt, bool write)
> >> >> > }
> >> >> >
> >> >> >
> >> >> > -/* Return true iff TYPE is scalarizable - i.e. a RECORD_TYPE or fixed-length
> >> >> > - ARRAY_TYPE with fields that are either of gimple register types (excluding
> >> >> > - bit-fields) or (recursively) scalarizable types. CONST_DECL must be true if
> >> >> > - we are considering a decl from constant pool. If it is false, char arrays
> >> >> > - will be refused. */
> >> >> > +/* Return true if TYPE consists of RECORD_TYPE or fixed-length ARRAY_TYPE with
> >> >> > + fields/elements that are not bit-fields and are either register types or
> >> >> > + recursively comply with simple_mix_of_records_and_arrays_p. Furthermore, if
> >> >> > + ALLOW_CHAR_ARRAYS is false, the function will return false also if TYPE
> >> >> > + contains an array of elements that only have one byte. */
> >> >> >
> >> >> > -static bool
> >> >> > -scalarizable_type_p (tree type, bool const_decl)
> >> >> > +bool
> >> >> > +simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays)
> >> >> > {
> >> >> > gcc_assert (!is_gimple_reg_type (type));
> >> >> > if (type_contains_placeholder_p (type))
> >> >> > @@ -977,7 +978,7 @@ scalarizable_type_p (tree type, bool const_decl)
> >> >> > return false;
> >> >> >
> >> >> > if (!is_gimple_reg_type (ft)
> >> >> > - && !scalarizable_type_p (ft, const_decl))
> >> >> > + && !simple_mix_of_records_and_arrays_p (ft, allow_char_arrays))
> >> >> > return false;
> >> >> > }
> >> >> >
> >> >> > @@ -986,7 +987,7 @@ scalarizable_type_p (tree type, bool const_decl)
> >> >> > case ARRAY_TYPE:
> >> >> > {
> >> >> > HOST_WIDE_INT min_elem_size;
> >> >> > - if (const_decl)
> >> >> > + if (allow_char_arrays)
> >> >> > min_elem_size = 0;
> >> >> > else
> >> >> > min_elem_size = BITS_PER_UNIT;
> >> >> > @@ -1008,7 +1009,7 @@ scalarizable_type_p (tree type, bool const_decl)
> >> >> >
> >> >> > tree elem = TREE_TYPE (type);
> >> >> > if (!is_gimple_reg_type (elem)
> >> >> > - && !scalarizable_type_p (elem, const_decl))
> >> >> > + && !simple_mix_of_records_and_arrays_p (elem, allow_char_arrays))
> >> >> > return false;
> >> >> > return true;
> >> >> > }
> >> >> > @@ -1017,10 +1018,38 @@ scalarizable_type_p (tree type, bool const_decl)
> >> >> > }
> >> >> > }
> >> >> >
> >> >> > -static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree, tree);
> >> >> > +static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree,
> >> >> > + tree);
> >> >> > +
> >> >> > +/* For a given array TYPE, return false if its domain does not have any maximum
> >> >> > + value. Otherwise calculate MIN and MAX indices of the first and the last
> >> >> > + element. */
> >> >> > +
> >> >> > +bool
> >> >> > +extract_min_max_idx_from_array (tree type, offset_int *min, offset_int *max)
> >> >> > +{
> >> >> > + tree domain = TYPE_DOMAIN (type);
> >> >> > + tree minidx = TYPE_MIN_VALUE (domain);
> >> >> > + gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
> >> >> > + tree maxidx = TYPE_MAX_VALUE (domain);
> >> >> > + if (!maxidx)
> >> >> > + return false;
> >> >> > + gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
> >> >> > +
> >> >> > + /* MINIDX and MAXIDX are inclusive, and must be interpreted in
> >> >> > + DOMAIN (e.g. signed int, whereas min/max may be size_int). */
> >> >> > + *min = wi::to_offset (minidx);
> >> >> > + *max = wi::to_offset (maxidx);
> >> >> > + if (!TYPE_UNSIGNED (domain))
> >> >> > + {
> >> >> > + *min = wi::sext (*min, TYPE_PRECISION (domain));
> >> >> > + *max = wi::sext (*max, TYPE_PRECISION (domain));
> >> >> > + }
> >> >> > + return true;
> >> >> > +}
> >> >> >
> >> >> > /* Create total_scalarization accesses for all scalar fields of a member
> >> >> > - of type DECL_TYPE conforming to scalarizable_type_p. BASE
> >> >> > + of type DECL_TYPE conforming to simple_mix_of_records_and_arrays_p. BASE
> >> >> > must be the top-most VAR_DECL representing the variable; within that,
> >> >> > OFFSET locates the member and REF must be the memory reference expression for
> >> >> > the member. */
> >> >> > @@ -1047,27 +1076,14 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
> >> >> > {
> >> >> > tree elemtype = TREE_TYPE (decl_type);
> >> >> > tree elem_size = TYPE_SIZE (elemtype);
> >> >> > - gcc_assert (elem_size && tree_fits_shwi_p (elem_size));
> >> >> > HOST_WIDE_INT el_size = tree_to_shwi (elem_size);
> >> >> > gcc_assert (el_size > 0);
> >> >> >
> >> >> > - tree minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (decl_type));
> >> >> > - gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
> >> >> > - tree maxidx = TYPE_MAX_VALUE (TYPE_DOMAIN (decl_type));
> >> >> > + offset_int idx, max;
> >> >> > /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
> >> >> > - if (maxidx)
> >> >> > + if (extract_min_max_idx_from_array (decl_type, &idx, &max))
> >> >> > {
> >> >> > - gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
> >> >> > tree domain = TYPE_DOMAIN (decl_type);
> >> >> > - /* MINIDX and MAXIDX are inclusive, and must be interpreted in
> >> >> > - DOMAIN (e.g. signed int, whereas min/max may be size_int). */
> >> >> > - offset_int idx = wi::to_offset (minidx);
> >> >> > - offset_int max = wi::to_offset (maxidx);
> >> >> > - if (!TYPE_UNSIGNED (domain))
> >> >> > - {
> >> >> > - idx = wi::sext (idx, TYPE_PRECISION (domain));
> >> >> > - max = wi::sext (max, TYPE_PRECISION (domain));
> >> >> > - }
> >> >> > for (int el_off = offset; idx <= max; ++idx)
> >> >> > {
> >> >> > tree nref = build4 (ARRAY_REF, elemtype,
> >> >> > @@ -1088,10 +1104,10 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
> >> >> > }
> >> >> >
> >> >> > /* Create total_scalarization accesses for a member of type TYPE, which must
> >> >> > - satisfy either is_gimple_reg_type or scalarizable_type_p. BASE must be the
> >> >> > - top-most VAR_DECL representing the variable; within that, POS and SIZE locate
> >> >> > - the member, REVERSE gives its torage order. and REF must be the reference
> >> >> > - expression for it. */
> >> >> > + satisfy either is_gimple_reg_type or simple_mix_of_records_and_arrays_p.
> >> >> > + BASE must be the top-most VAR_DECL representing the variable; within that,
> >> >> > + POS and SIZE locate the member, REVERSE gives its torage order. and REF must
> >> >> > + be the reference expression for it. */
> >> >> >
> >> >> > static void
> >> >> > scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
> >> >> > @@ -1111,7 +1127,8 @@ scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
> >> >> > }
> >> >> >
> >> >> > /* Create a total_scalarization access for VAR as a whole. VAR must be of a
> >> >> > - RECORD_TYPE or ARRAY_TYPE conforming to scalarizable_type_p. */
> >> >> > + RECORD_TYPE or ARRAY_TYPE conforming to
> >> >> > + simple_mix_of_records_and_arrays_p. */
> >> >> >
> >> >> > static void
> >> >> > create_total_scalarization_access (tree var)
> >> >> > @@ -2803,8 +2820,9 @@ analyze_all_variable_accesses (void)
> >> >> > {
> >> >> > tree var = candidate (i);
> >> >> >
> >> >> > - if (VAR_P (var) && scalarizable_type_p (TREE_TYPE (var),
> >> >> > - constant_decl_p (var)))
> >> >> > + if (VAR_P (var)
> >> >> > + && simple_mix_of_records_and_arrays_p (TREE_TYPE (var),
> >> >> > + constant_decl_p (var)))
> >> >> > {
> >> >> > if (tree_to_uhwi (TYPE_SIZE (TREE_TYPE (var)))
> >> >> > <= max_scalarization_size)
> >> >> > diff --git a/gcc/tree-sra.h b/gcc/tree-sra.h
> >> >> > new file mode 100644
> >> >> > index 00000000000..dc901385994
> >> >> > --- /dev/null
> >> >> > +++ b/gcc/tree-sra.h
> >> >> > @@ -0,0 +1,33 @@
> >> >> > +/* tree-sra.h - Run-time parameters.
> >> >> > + Copyright (C) 2017 Free Software Foundation, Inc.
> >> >> > +
> >> >> > +This file is part of GCC.
> >> >> > +
> >> >> > +GCC is free software; you can redistribute it and/or modify it under
> >> >> > +the terms of the GNU General Public License as published by the Free
> >> >> > +Software Foundation; either version 3, or (at your option) any later
> >> >> > +version.
> >> >> > +
> >> >> > +GCC is distributed in the hope that it will be useful, but WITHOUT ANY
> >> >> > +WARRANTY; without even the implied warranty of MERCHANTABILITY or
> >> >> > +FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
> >> >> > +for more details.
> >> >> > +
> >> >> > +You should have received a copy of the GNU General Public License
> >> >> > +along with GCC; see the file COPYING3. If not see
> >> >> > +<http://www.gnu.org/licenses/>. */
> >> >> > +
> >> >> > +#ifndef TREE_SRA_H
> >> >> > +#define TREE_SRA_H
> >> >> > +
> >> >> > +
> >> >> > +bool simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays);
> >> >> > +bool extract_min_max_idx_from_array (tree type, offset_int *idx,
> >> >> > + offset_int *max);
> >> >> > +tree build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
> >> >> > + bool reverse, tree exp_type,
> >> >> > + gimple_stmt_iterator *gsi, bool insert_after);
> >> >> > +
> >> >> > +
> >> >> > +
> >> >> > +#endif /* TREE_SRA_H */
> >> >> > --
> >> >> > 2.14.1
> >> >> >
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-10-26 15:12 ` Richard Biener
@ 2017-11-03 16:38 ` Martin Jambor
2017-11-13 13:17 ` Richard Biener
2017-11-14 10:25 ` Martin Jambor
0 siblings, 2 replies; 26+ messages in thread
From: Martin Jambor @ 2017-11-03 16:38 UTC (permalink / raw)
To: Richard Biener; +Cc: GCC Patches, Jan Hubicka, Michael Matz
Hi,
On Thu, Oct 26, 2017 at 02:43:02PM +0200, Richard Biener wrote:
> On Thu, Oct 26, 2017 at 2:18 PM, Martin Jambor <mjambor@suse.cz> wrote:
> >
> > Nevertheless, I still intend to experiment with the limit, I sent out
> > this RFC exactly so that I don't spend a lot of time benchmarking
> > something that is eventually not deemed acceptable on principle.
>
> I think the limit should be on the number of generated copies and not
> the overall size of the structure... If the struct were composed of
> 32 individual chars we wouldn't want to emit 32 loads and 32 stores...
I have added another parameter to also limit the number of generated
element copies. I have kept the size limit so that we don't even
attempt to count them for large structures.
> Given that load bandwith is usually higher than store bandwith it
> might make sense to do the store combining in our copying sequence,
> like for the 8 byte entry case use sth like
>
> movq 0(%eax), %xmm0
> movhps 8(%eax), %xmm0 // or vpinsert
> mov[au]ps %xmm0, 0%(ebx)
I would be concerned about the cost of GPR->XMM moves when the value
being stored is in a GPR, especially with generic tuning which (with
-O2) is the main thing I am targeting here. Wouldn't we actually pass
it through stack with all the associated penalties?
Also, while such store combining might work for ImageMagick, if a
programmer did:
region1->x = x1;
region2->x = x2;
region1->y = 0;
region2->y = 20;
...
SetPixelCacheNexusPixels(cache_info, ReadMode, region1, ...)
The transformation would not work unless it could prove region1 and
region2 are not the same thing.
> As said a general concern was you not copying padding. If you
> put this into an even more common place you surely will break
> stuff, no?
I don't understand, what even more common place do you mean?
I have been testing the patch also on a bunch of other architectures
and those have tests in their testsuite that check that padding is
copied, for example some tests in gcc.target/aarch64/aapcs64/ check
whether a structure passed to a function is binary the same as the
original, and the test fail because of padding. That is the only
"breakage" I know about but I believe that the assumption that padding
must always be is wrong (if it is not than we need to make SRA quite a
bit more conservative).
On Thu, Oct 26, 2017 at 05:09:42PM +0200, Richard Biener wrote:
> Also if we do the stores in smaller chunks we are more
> likely hitting the same store-to-load-forwarding issue
> elsewhere. Like in case the destination is memcpy'ed
> away.
>
> So the proposed change isn't necessarily a win without
> a possible similar regression that it tries to fix.
>
With some encouragement by Honza, I have done some benchmarking anyway
and I did not see anything of that kind.
> Whole-program analysis of accesses might allow
> marking affected objects.
Attempting to save access patterns before IPA and then tracking them
and keep them in sync across inlining and all gimple late passes seems
like a nightmarish task. If this approach is indeed rejected I might
attempt to do the store combining but a WPA analysis seems just too
complex.
Anyway, here are the numbers. They were taken on two different
Zen-based machines. I am also in the process of measuring at least
something on a Haswell machine but I started later and the machine is
quite a bit slower so I will not have the numbers until next week (and
not all equivalents in any way). I found out I do not have access to
any more modern .*Lake intel CPU.
trunk is pristine trunk revision 254205. All benchmarks were run
three times and the median was chosen.
s or strict means the patch with the strictest possible settings to
speed-up ImageMagick, i.e. --param max-size-for-elementwise-copy=32
--param max-insns-for-elementwise-copy=4. Also run three times.
x1 is patched trunk with the parameters having the default values was
going to propose, i.e. --param max-size-for-elementwise-copy=35
--param max-insns-for-elementwise-copy=6. Also run three times.
I then increased the parameter, in search for further missed
opportunities and to see what and how soon will start to regress.
x2 is roughly twice that, --param max-size-for-elementwise-copy=67
--param max-insns-for-elementwise-copy=12. Run twice, outliers
manually checked.
x4 is roughly four times x1, namely --param max-size-for-elementwise-copy=143
--param max-insns-for-elementwise-copy=24. Run only once.
The times below are of course "non-reportable," for a whole bunch of
reasons.
Zen SPECINT 2006 -O2 generic tuning
====================================
Run-time
--------
| Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
|----------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
| 400.perlbench | 237 | 236 | -0.42 | 236 | -0.42 | 238 | +0.42 | 237 | +0.00 |
| 401.bzip2 | 341 | 342 | +0.29 | 341 | +0.00 | 341 | +0.00 | 341 | +0.00 |
| 403.gcc | 217 | 217 | +0.00 | 217 | +0.00 | 216 | -0.46 | 217 | +0.00 |
| 429.mcf | 224 | 218 | -2.68 | 223 | -0.45 | 221 | -1.34 | 226 | +0.89 |
| 445.gobmk | 361 | 361 | +0.00 | 361 | +0.00 | 360 | -0.28 | 363 | +0.55 |
| 456.hmmer | 296 | 296 | +0.00 | 296 | +0.00 | 297 | +0.34 | 296 | +0.00 |
| 458.sjeng | 453 | 452 | -0.22 | 454 | +0.22 | 454 | +0.22 | 460 | +1.55 |
| 462.libquantum | 289 | 289 | +0.00 | 291 | +0.69 | 289 | +0.00 | 291 | +0.69 |
| 464.h264ref | 391 | 391 | +0.00 | 385 | -1.53 | 385 | -1.53 | 385 | -1.53 |
| 471.omnetpp | 269 | 255 | -5.20 | 250 | -7.06 | 247 | -8.18 | 268 | -0.37 |
| 473.astar | 320 | 321 | +0.31 | 317 | -0.94 | 320 | +0.00 | 320 | +0.00 |
| 483.xalancbmk | 187 | 188 | +0.53 | 188 | +0.53 | 187 | +0.00 | 187 | +0.00 |
Although the omnetpp looks like a sizeable improvement I should warn
that this is one of the few slightly jumpy benchmarks. However, I
re-run it a few more times and it seems like it is jumping around a
lower value when compiled with the patched compiler. It might not be
the 5-8% though.
Text size
---------
| Benchmark | trunk | struict | % | x1 | % | x2 | % | x4 | % |
|----------------+---------+---------+-------+---------+-------+---------+-------+---------+-------|
| 400.perlbench | 875874 | 875954 | +0.01 | 875954 | +0.01 | 876018 | +0.02 | 876146 | +0.03 |
| 401.bzip2 | 44754 | 44754 | +0.00 | 44754 | +0.00 | 44754 | +0.00 | 44754 | +0.00 |
| 403.gcc | 2294466 | 2294930 | +0.02 | 2296098 | +0.07 | 2296306 | +0.08 | 2296466 | +0.09 |
| 429.mcf | 8226 | 8226 | +0.00 | 8226 | +0.00 | 8258 | +0.39 | 8258 | +0.39 |
| 445.gobmk | 579778 | 579778 | +0.00 | 579826 | +0.01 | 579826 | +0.01 | 580402 | +0.11 |
| 456.hmmer | 221058 | 221058 | +0.00 | 221058 | +0.00 | 221058 | +0.00 | 221058 | +0.00 |
| 458.sjeng | 93362 | 93362 | +0.00 | 94882 | +1.63 | 94882 | +1.63 | 96066 | +2.90 |
| 462.libquantum | 28314 | 28314 | +0.00 | 28362 | +0.17 | 28362 | +0.17 | 28362 | +0.17 |
| 464.h264ref | 393874 | 393874 | +0.00 | 393922 | +0.01 | 393922 | +0.01 | 394226 | +0.09 |
| 471.omnetpp | 430306 | 430306 | +0.00 | 430418 | +0.03 | 430418 | +0.03 | 430418 | +0.03 |
| 473.astar | 29362 | 29538 | +0.60 | 29538 | +0.60 | 29554 | +0.65 | 29554 | +0.65 |
| 483.xalancbmk | 2361298 | 2361506 | +0.01 | 2361506 | +0.01 | 2361506 | +0.01 | 2361506 | +0.01 |
Zen SPECINT 2006 -Ofast native tuning
======================================
Run-time
--------
| Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
|----------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
| 400.perlbench | 240 | 239 | -0.42 | 239 | -0.42 | 241 | +0.42 | 238 | -0.83 |
| 401.bzip2 | 341 | 341 | +0.00 | 341 | +0.00 | 341 | +0.00 | 340 | -0.29 |
| 403.gcc | 210 | 208 | -0.95 | 207 | -1.43 | 209 | -0.48 | 208 | -0.95 |
| 429.mcf | 225 | 225 | +0.00 | 225 | +0.00 | 228 | +1.33 | 226 | +0.44 |
| 445.gobmk | 352 | 352 | +0.00 | 352 | +0.00 | 351 | -0.28 | 352 | +0.00 |
| 456.hmmer | 131 | 131 | +0.00 | 131 | +0.00 | 131 | +0.00 | 131 | +0.00 |
| 458.sjeng | 442 | 442 | +0.00 | 438 | -0.90 | 438 | -0.90 | 437 | -1.13 |
| 462.libquantum | 291 | 292 | +0.34 | 286 | -1.72 | 287 | -1.37 | 287 | -1.37 |
| 464.h264ref | 364 | 365 | +0.27 | 364 | +0.00 | 364 | +0.00 | 363 | -0.27 |
| 471.omnetpp | 266 | 266 | +0.00 | 265 | -0.38 | 265 | -0.38 | 265 | -0.38 |
| 473.astar | 306 | 307 | +0.33 | 306 | +0.00 | 306 | +0.00 | 309 | +0.98 |
| 483.xalancbmk | 177 | 173 | -2.26 | 170 | -3.95 | 170 | -3.95 | 170 | -3.95 |
Text size
---------
| Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
|----------------+---------+---------+-------+---------+-------+---------+-------+---------+-------|
| 400.perlbench | 1161762 | 1161874 | +0.01 | 1161874 | +0.01 | 1162226 | +0.04 | 1162338 | +0.05 |
| 401.bzip2 | 80834 | 80834 | +0.00 | 80834 | +0.00 | 80834 | +0.00 | 80834 | +0.00 |
| 403.gcc | 3170946 | 3171394 | +0.01 | 3172914 | +0.06 | 3173170 | +0.07 | 3174818 | +0.12 |
| 429.mcf | 10418 | 10418 | +0.00 | 10418 | +0.00 | 10450 | +0.31 | 10450 | +0.31 |
| 445.gobmk | 779778 | 779778 | +0.00 | 779842 | +0.01 | 779842 | +0.01 | 780418 | +0.08 |
| 456.hmmer | 328258 | 328258 | +0.00 | 328258 | +0.00 | 328258 | +0.00 | 328258 | +0.00 |
| 458.sjeng | 146386 | 146386 | +0.00 | 148162 | +1.21 | 148162 | +1.21 | 149330 | +2.01 |
| 462.libquantum | 30666 | 30666 | +0.00 | 30730 | +0.21 | 30730 | +0.21 | 30730 | +0.21 |
| 464.h264ref | 737826 | 737826 | +0.00 | 737890 | +0.01 | 737890 | +0.01 | 739186 | +0.18 |
| 471.omnetpp | 561570 | 561570 | +0.00 | 561826 | +0.05 | 561826 | +0.05 | 561826 | +0.05 |
| 473.astar | 39314 | 39522 | +0.53 | 39522 | +0.53 | 39538 | +0.57 | 39538 | +0.57 |
| 483.xalancbmk | 3319682 | 3319842 | +0.00 | 3319842 | +0.00 | 3319842 | +0.00 | 3319842 | +0.00 |
Zen SPECFP 2006 -O2 generic tuning
==================================
Run-time
--------
| Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
|---------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
| 410.bwaves | 214 | 213 | -0.47 | 214 | +0.00 | 214 | +0.00 | 214 | +0.00 |
| 433.milc | 290 | 291 | +0.34 | 290 | +0.00 | 295 | +1.72 | 289 | -0.34 |
| 434.zeusmp | 182 | 182 | +0.00 | 182 | +0.00 | 184 | +1.10 | 182 | +0.00 |
| 435.gromacs | 218 | 218 | +0.00 | 217 | -0.46 | 216 | -0.92 | 220 | +0.92 |
| 436.cactusADM | 350 | 349 | -0.29 | 349 | -0.29 | 343 | -2.00 | 349 | -0.29 |
| 437.leslie3d | 196 | 195 | -0.51 | 196 | +0.00 | 194 | -1.02 | 196 | +0.00 |
| 444.namd | 273 | 273 | +0.00 | 273 | +0.00 | 273 | +0.00 | 273 | +0.00 |
| 447.dealII | 211 | 211 | +0.00 | 210 | -0.47 | 210 | -0.47 | 211 | +0.00 |
| 450.soplex | 187 | 188 | +0.53 | 188 | +0.53 | 187 | +0.00 | 187 | +0.00 |
| 453.povray | 119 | 118 | -0.84 | 119 | +0.00 | 119 | +0.00 | 118 | -0.84 |
| 454.calculix | 534 | 533 | -0.19 | 531 | -0.56 | 531 | -0.56 | 532 | -0.37 |
| 459.GemsFDTD | 236 | 235 | -0.42 | 235 | -0.42 | 242 | +2.54 | 237 | +0.42 |
| 465.tonto | 366 | 365 | -0.27 | 365 | -0.27 | 364 | -0.55 | 365 | -0.27 |
| 470.lbm | 181 | 180 | -0.55 | 180 | -0.55 | 180 | -0.55 | 180 | -0.55 |
| 481.wrf | 303 | 303 | +0.00 | 302 | -0.33 | 304 | +0.33 | 304 | +0.33 |
| 482.sphinx3 | 362 | 362 | +0.00 | 360 | -0.55 | 361 | -0.28 | 363 | +0.28 |
Text size
---------
| Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
|---------------+---------+---------+-------+---------+-------+---------+-------+---------+-------|
| 410.bwaves | 25954 | 25954 | +0.00 | 25954 | +0.00 | 25954 | +0.00 | 25954 | +0.00 |
| 433.milc | 87922 | 87922 | +0.00 | 87922 | +0.00 | 88610 | +0.78 | 89042 | +1.27 |
| 434.zeusmp | 212034 | 212034 | +0.00 | 212034 | +0.00 | 212034 | +0.00 | 212034 | +0.00 |
| 435.gromacs | 747026 | 747026 | +0.00 | 747026 | +0.00 | 747026 | +0.00 | 747026 | +0.00 |
| 436.cactusADM | 526178 | 526178 | +0.00 | 526178 | +0.00 | 526274 | +0.02 | 526274 | +0.02 |
| 437.leslie3d | 83234 | 83234 | +0.00 | 83234 | +0.00 | 83234 | +0.00 | 83234 | +0.00 |
| 444.namd | 297234 | 297266 | +0.01 | 297266 | +0.01 | 297266 | +0.01 | 297266 | +0.01 |
| 447.dealII | 2165282 | 2167650 | +0.11 | 2172290 | +0.32 | 2174034 | +0.40 | 2174082 | +0.41 |
| 450.soplex | 347122 | 347122 | +0.00 | 347122 | +0.00 | 347122 | +0.00 | 347122 | +0.00 |
| 453.povray | 800914 | 800962 | +0.01 | 801570 | +0.08 | 802002 | +0.14 | 803138 | +0.28 |
| 454.calculix | 1342802 | 1342802 | +0.00 | 1342802 | +0.00 | 1342802 | +0.00 | 1342802 | +0.00 |
| 459.GemsFDTD | 353410 | 354050 | +0.18 | 354050 | +0.18 | 354050 | +0.18 | 354098 | +0.19 |
| 465.tonto | 3464210 | 3465058 | +0.02 | 3465058 | +0.02 | 3468434 | +0.12 | 3476594 | +0.36 |
| 470.lbm | 9202 | 9202 | +0.00 | 9202 | +0.00 | 9202 | +0.00 | 9202 | +0.00 |
| 481.wrf | 3345170 | 3345170 | +0.00 | 3345170 | +0.00 | 3351586 | +0.19 | 3351586 | +0.19 |
| 482.sphinx3 | 125026 | 125026 | +0.00 | 125026 | +0.00 | 125026 | +0.00 | 125026 | +0.00 |
Zen SPECFP 2006 -Ofast native tuning
====================================
Run-time
--------
| Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
|---------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
| 410.bwaves | 151 | 150 | -0.66 | 151 | +0.00 | 151 | +0.00 | 151 | +0.00 |
| 433.milc | 197 | 197 | +0.00 | 197 | +0.00 | 194 | -1.52 | 186 | -5.58 |
| 434.zeusmp | 128 | 128 | +0.00 | 128 | +0.00 | 128 | +0.00 | 128 | +0.00 |
| 435.gromacs | 181 | 181 | +0.00 | 180 | -0.55 | 180 | -0.55 | 181 | +0.00 |
| 436.cactusADM | 139 | 139 | +0.00 | 139 | +0.00 | 132 | -5.04 | 139 | +0.00 |
| 437.leslie3d | 159 | 160 | +0.63 | 160 | +0.63 | 159 | +0.00 | 159 | +0.00 |
| 444.namd | 256 | 256 | +0.00 | 255 | -0.39 | 255 | -0.39 | 256 | +0.00 |
| 447.dealII | 200 | 200 | +0.00 | 199 | -0.50 | 201 | +0.50 | 201 | +0.50 |
| 450.soplex | 184 | 184 | +0.00 | 185 | +0.54 | 184 | +0.00 | 184 | +0.00 |
| 453.povray | 124 | 122 | -1.61 | 123 | -0.81 | 124 | +0.00 | 122 | -1.61 |
| 454.calculix | 192 | 192 | +0.00 | 192 | +0.00 | 193 | +0.52 | 193 | +0.52 |
| 459.GemsFDTD | 208 | 208 | +0.00 | 208 | +0.00 | 214 | +2.88 | 208 | +0.00 |
| 465.tonto | 320 | 320 | +0.00 | 320 | +0.00 | 320 | +0.00 | 320 | +0.00 |
| 470.lbm | 142 | 142 | +0.00 | 142 | +0.00 | 142 | +0.00 | 142 | +0.00 |
| 481.wrf | 195 | 195 | +0.00 | 195 | +0.00 | 195 | +0.00 | 195 | +0.00 |
| 482.sphinx3 | 256 | 258 | +0.78 | 256 | +0.00 | 256 | +0.00 | 257 | +0.39 |
Text size
---------
| Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
|---------------+---------+---------+-------+---------+-------+---------+-------+---------+-------|
| 410.bwaves | 27490 | 27490 | +0.00 | 27490 | +0.00 | 27490 | +0.00 | 27490 | +0.00 |
| 433.milc | 118178 | 118178 | +0.00 | 118178 | +0.00 | 118962 | +0.66 | 119634 | +1.23 |
| 434.zeusmp | 411106 | 411106 | +0.00 | 411106 | +0.00 | 411106 | +0.00 | 411106 | +0.00 |
| 435.gromacs | 935970 | 935970 | +0.00 | 935970 | +0.00 | 935970 | +0.00 | 936162 | +0.02 |
| 436.cactusADM | 750546 | 750546 | +0.00 | 750546 | +0.00 | 750626 | +0.01 | 750626 | +0.01 |
| 437.leslie3d | 123410 | 123410 | +0.00 | 123410 | +0.00 | 123410 | +0.00 | 123410 | +0.00 |
| 444.namd | 284082 | 284114 | +0.01 | 284114 | +0.01 | 284114 | +0.01 | 284114 | +0.01 |
| 447.dealII | 2438610 | 2440946 | +0.10 | 2444978 | +0.26 | 2446882 | +0.34 | 2446930 | +0.34 |
| 450.soplex | 443218 | 443218 | +0.00 | 443218 | +0.00 | 443218 | +0.00 | 443218 | +0.00 |
| 453.povray | 1077778 | 1077890 | +0.01 | 1078658 | +0.08 | 1079026 | +0.12 | 1080370 | +0.24 |
| 454.calculix | 1639138 | 1639138 | +0.00 | 1639138 | +0.00 | 1639474 | +0.02 | 1639474 | +0.02 |
| 459.GemsFDTD | 451202 | 451234 | +0.01 | 451234 | +0.01 | 451234 | +0.01 | 451282 | +0.02 |
| 465.tonto | 4584690 | 4585250 | +0.01 | 4585250 | +0.01 | 4588130 | +0.08 | 4595442 | +0.23 |
| 470.lbm | 9858 | 9858 | +0.00 | 9858 | +0.00 | 9858 | +0.00 | 9858 | +0.00 |
| 481.wrf | 4588002 | 4588002 | +0.00 | 4588290 | +0.01 | 4621010 | +0.72 | 4621922 | +0.74 |
| 482.sphinx3 | 179602 | 179602 | +0.00 | 179602 | +0.00 | 179602 | +0.00 | 179602 | +0.00 |
Zen SPEC INT 2017 -O2 generic tuning
====================================
Run-time
--------
| Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
|-----------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
| 500.perlbench_r | 529 | 529 | +0.00 | 531 | +0.38 | 530 | +0.19 | 534 | +0.95 |
| 502.gcc_r | 338 | 333 | -1.48 | 334 | -1.18 | 339 | +0.30 | 339 | +0.30 |
| 505.mcf_r | 382 | 381 | -0.26 | 382 | +0.00 | 382 | +0.00 | 381 | -0.26 |
| 520.omnetpp_r | 511 | 503 | -1.57 | 497 | -2.74 | 497 | -2.74 | 497 | -2.74 |
| 523.xalancbmk_r | 391 | 388 | -0.77 | 389 | -0.51 | 390 | -0.26 | 391 | +0.00 |
| 525.x264_r | 590 | 590 | +0.00 | 591 | +0.17 | 592 | +0.34 | 593 | +0.51 |
| 531.deepsjeng_r | 427 | 427 | +0.00 | 427 | +0.00 | 428 | +0.23 | 427 | +0.00 |
| 541.leela_r | 716 | 716 | +0.00 | 716 | +0.00 | 719 | +0.42 | 719 | +0.42 |
| 548.exchange2_r | 593 | 593 | +0.00 | 593 | +0.00 | 593 | +0.00 | 593 | +0.00 |
| 557.xz_r | 452 | 452 | +0.00 | 453 | +0.22 | 454 | +0.44 | 452 | +0.00 |
Text size
---------
| Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
|-----------------+---------+---------+-------+---------+-------+---------+-------+---------+-------|
| 500.perlbench_r | 1599442 | 1599522 | +0.01 | 1599522 | +0.01 | 1599522 | +0.01 | 1600082 | +0.04 |
| 502.gcc_r | 6757602 | 6758978 | +0.02 | 6759090 | +0.02 | 6759842 | +0.03 | 6760306 | +0.04 |
| 505.mcf_r | 16098 | 16098 | +0.00 | 16098 | +0.00 | 16098 | +0.00 | 16306 | +1.29 |
| 520.omnetpp_r | 1262498 | 1262562 | +0.01 | 1264034 | +0.12 | 1264034 | +0.12 | 1264034 | +0.12 |
| 523.xalancbmk_r | 3989026 | 3989202 | +0.00 | 3989202 | +0.00 | 3989202 | +0.00 | 3989202 | +0.00 |
| 525.x264_r | 414130 | 414194 | +0.02 | 414194 | +0.02 | 414738 | +0.15 | 415122 | +0.24 |
| 531.deepsjeng_r | 67426 | 67426 | +0.00 | 67458 | +0.05 | 67458 | +0.05 | 67458 | +0.05 |
| 541.leela_r | 219378 | 219378 | +0.00 | 219378 | +0.00 | 224082 | +2.14 | 237026 | +8.04 |
| 548.exchange2_r | 61234 | 61234 | +0.00 | 61234 | +0.00 | 61234 | +0.00 | 61234 | +0.00 |
| 557.xz_r | 111490 | 111490 | +0.00 | 111490 | +0.00 | 111506 | +0.01 | 111890 | +0.36 |
Zen SPEC INT 2017 -Ofast native tuning
======================================
Run-time
---------
| Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
|-----------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
| 500.perlbench_r | 525 | 524 | -0.19 | 525 | +0.00 | 525 | +0.00 | 534 | +1.71 |
| 502.gcc_r | 331 | 329 | -0.60 | 324 | -2.11 | 330 | -0.30 | 324 | -2.11 |
| 505.mcf_r | 380 | 380 | +0.00 | 381 | +0.26 | 380 | +0.00 | 379 | -0.26 |
| 520.omnetpp_r | 487 | 486 | -0.21 | 488 | +0.21 | 489 | +0.41 | 488 | +0.21 |
| 523.xalancbmk_r | 373 | 369 | -1.07 | 367 | -1.61 | 370 | -0.80 | 368 | -1.34 |
| 525.x264_r | 319 | 319 | +0.00 | 320 | +0.31 | 321 | +0.63 | 322 | +0.94 |
| 531.deepsjeng_r | 418 | 418 | +0.00 | 418 | +0.00 | 418 | +0.00 | 419 | +0.24 |
| 541.leela_r | 674 | 674 | +0.00 | 674 | +0.00 | 672 | -0.30 | 672 | -0.30 |
| 548.exchange2_r | 466 | 466 | +0.00 | 466 | +0.00 | 466 | +0.00 | 466 | +0.00 |
| 557.xz_r | 443 | 443 | +0.00 | 443 | +0.00 | 449 | +1.35 | 449 | +1.35 |
Text size
---------
| Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
|-----------------+---------+---------+-------+---------+-------+---------+-------+---------+-------|
| 500.perlbench_r | 2122882 | 2122962 | +0.00 | 2122962 | +0.00 | 2122962 | +0.00 | 2122514 | -0.02 |
| 502.gcc_r | 8566290 | 8567794 | +0.02 | 8569138 | +0.03 | 8570066 | +0.04 | 8570642 | +0.05 |
| 505.mcf_r | 26770 | 26770 | +0.00 | 26770 | +0.00 | 26770 | +0.00 | 26962 | +0.72 |
| 520.omnetpp_r | 1713938 | 1713954 | +0.00 | 1714754 | +0.05 | 1714754 | +0.05 | 1714754 | +0.05 |
| 523.xalancbmk_r | 4881890 | 4882114 | +0.00 | 4882114 | +0.00 | 4882114 | +0.00 | 4882114 | +0.00 |
| 525.x264_r | 601522 | 601602 | +0.01 | 601602 | +0.01 | 602130 | +0.10 | 602834 | +0.22 |
| 531.deepsjeng_r | 90306 | 90306 | +0.00 | 90338 | +0.04 | 90338 | +0.04 | 90338 | +0.04 |
| 541.leela_r | 277634 | 277650 | +0.01 | 277650 | +0.01 | 282386 | +1.71 | 295778 | +6.54 |
| 548.exchange2_r | 109058 | 109058 | +0.00 | 109058 | +0.00 | 109058 | +0.00 | 109058 | +0.00 |
| 557.xz_r | 154594 | 154594 | +0.00 | 154594 | +0.00 | 154610 | +0.01 | 154930 | +0.22 |
Zen SPEC 2017 FP -O2 generic tuning
===================================
Run-time
--------
| Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
|-----------------+-------+-----+--------+-----+--------+-----+--------+-----+--------|
| 503.bwaves_r | 801 | 801 | +0.00 | 801 | +0.00 | 801 | +0.00 | 801 | +0.00 |
| 507.cactuBSSN_r | 303 | 302 | -0.33 | 299 | -1.32 | 302 | -0.33 | 307 | +1.32 |
| 508.namd_r | 306 | 306 | +0.00 | 307 | +0.33 | 306 | +0.00 | 306 | +0.00 |
| 510.parest_r | 558 | 553 | -0.90 | 561 | +0.54 | 554 | -0.72 | 562 | +0.72 |
| 511.povray_r | 679 | 672 | -1.03 | 673 | -0.88 | 680 | +0.15 | 644 | -5.15 |
| 519.lbm_r | 240 | 240 | +0.00 | 240 | +0.00 | 240 | +0.00 | 240 | +0.00 |
| 521.wrf_r | 851 | 827 | -2.82 | 827 | -2.82 | 827 | -2.82 | 828 | -2.70 |
| 526.blender_r | 376 | 376 | +0.00 | 379 | +0.80 | 377 | +0.27 | 376 | +0.00 |
| 527.cam4_r | 529 | 527 | -0.38 | 533 | +0.76 | 536 | +1.32 | 528 | -0.19 |
| 538.imagick_r | 646 | 570 | -11.76 | 570 | -11.76 | 569 | -11.92 | 570 | -11.76 |
| 544.nab_r | 467 | 467 | +0.00 | 467 | +0.00 | 467 | +0.00 | 467 | +0.00 |
| 549.fotonik3d_r | 413 | 413 | +0.00 | 414 | +0.24 | 415 | +0.48 | 413 | +0.00 |
| 554.roms_r | 459 | 455 | -0.87 | 456 | -0.65 | 456 | -0.65 | 456 | -0.65 |
Text size
---------
| Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
|-----------------+----------+----------+-------+----------+-------+----------+-------+----------+-------|
| 503.bwaves_r | 32034 | 32034 | +0.00 | 32034 | +0.00 | 32034 | +0.00 | 32034 | +0.00 |
| 507.cactuBSSN_r | 2951634 | 2951634 | +0.00 | 2951634 | +0.00 | 2951698 | +0.00 | 2951730 | +0.00 |
| 508.namd_r | 837458 | 837490 | +0.00 | 837490 | +0.00 | 837490 | +0.00 | 837490 | +0.00 |
| 510.parest_r | 6540866 | 6545618 | +0.07 | 6546754 | +0.09 | 6561426 | +0.31 | 6569426 | +0.44 |
| 511.povray_r | 803618 | 803666 | +0.01 | 804274 | +0.08 | 804706 | +0.14 | 805842 | +0.28 |
| 519.lbm_r | 12018 | 12018 | +0.00 | 12018 | +0.00 | 12018 | +0.00 | 12018 | +0.00 |
| 521.wrf_r | 16292962 | 16296786 | +0.02 | 16296978 | +0.02 | 16302594 | +0.06 | 16419842 | +0.78 |
| 526.blender_r | 7268224 | 7281264 | +0.18 | 7282608 | +0.20 | 7289168 | +0.29 | 7295296 | +0.37 |
| 527.cam4_r | 5063666 | 5063922 | +0.01 | 5065010 | +0.03 | 5068114 | +0.09 | 5072946 | +0.18 |
| 538.imagick_r | 1608178 | 1609282 | +0.07 | 1609282 | +0.07 | 1613458 | +0.33 | 1613970 | +0.36 |
| 544.nab_r | 156242 | 156242 | +0.00 | 156242 | +0.00 | 156242 | +0.00 | 156242 | +0.00 |
| 549.fotonik3d_r | 326738 | 326738 | +0.00 | 326738 | +0.00 | 326738 | +0.00 | 326738 | +0.00 |
| 554.roms_r | 728546 | 728546 | +0.00 | 728546 | +0.00 | 728546 | +0.00 | 728546 | +0.00 |
Zen SPEC 2017 FP -Ofast native tuning
=====================================
Run-time
--------
| Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
|-----------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
| 503.bwaves_r | 310 | 310 | +0.00 | 310 | +0.00 | 310 | +0.00 | 309 | -0.32 |
| 507.cactuBSSN_r | 269 | 266 | -1.12 | 266 | -1.12 | 268 | -0.37 | 270 | +0.37 |
| 508.namd_r | 270 | 269 | -0.37 | 269 | -0.37 | 268 | -0.74 | 268 | -0.74 |
| 510.parest_r | 607 | 601 | -0.99 | 599 | -1.32 | 599 | -1.32 | 604 | -0.49 |
| 511.povray_r | 662 | 664 | +0.30 | 671 | +1.36 | 680 | +2.72 | 675 | +1.96 |
| 519.lbm_r | 186 | 186 | +0.00 | 186 | +0.00 | 186 | +0.00 | 186 | +0.00 |
| 521.wrf_r | 550 | 554 | +0.73 | 550 | +0.00 | 550 | +0.00 | 549 | -0.18 |
| 526.blender_r | 355 | 354 | -0.28 | 355 | +0.00 | 354 | -0.28 | 354 | -0.28 |
| 527.cam4_r | 434 | 437 | +0.69 | 435 | +0.23 | 437 | +0.69 | 435 | +0.23 |
| 538.imagick_r | 433 | 420 | -3.00 | 420 | -3.00 | 420 | -3.00 | 419 | -3.23 |
| 544.nab_r | 424 | 425 | +0.24 | 425 | +0.24 | 425 | +0.24 | 425 | +0.24 |
| 549.fotonik3d_r | 421 | 422 | +0.24 | 422 | +0.24 | 422 | +0.24 | 422 | +0.24 |
| 554.roms_r | 360 | 361 | +0.28 | 361 | +0.28 | 361 | +0.28 | 361 | +0.28 |
+1.36% for 511.povray_r is the worst regression for the proposed x1
defaults, by the way. I have not investigated it further, however.
Text size
---------
| Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
|-----------------+----------+----------+-------+----------+-------+----------+-------+----------+-------|
| 503.bwaves_r | 34562 | 34562 | +0.00 | 34562 | +0.00 | 34562 | +0.00 | 34562 | +0.00 |
| 507.cactuBSSN_r | 3978402 | 3978402 | +0.00 | 3978402 | +0.00 | 3978514 | +0.00 | 3978546 | +0.00 |
| 508.namd_r | 869106 | 869154 | +0.01 | 869154 | +0.01 | 869154 | +0.01 | 869154 | +0.01 |
| 510.parest_r | 7186258 | 7189298 | +0.04 | 7190370 | +0.06 | 7203890 | +0.25 | 7211202 | +0.35 |
| 511.povray_r | 1063314 | 1063410 | +0.01 | 1064178 | +0.08 | 1064546 | +0.12 | 1065890 | +0.24 |
| 519.lbm_r | 12178 | 12178 | +0.00 | 12178 | +0.00 | 12178 | +0.00 | 12178 | +0.00 |
| 521.wrf_r | 19480946 | 19484146 | +0.02 | 19484466 | +0.02 | 19607538 | +0.65 | 19716178 | +1.21 |
| 526.blender_r | 9708752 | 9719952 | +0.12 | 9722768 | +0.14 | 9730224 | +0.22 | 9737760 | +0.30 |
| 527.cam4_r | 6217970 | 6218162 | +0.00 | 6219570 | +0.03 | 6223362 | +0.09 | 6227762 | +0.16 |
| 538.imagick_r | 2255682 | 2256162 | +0.02 | 2256162 | +0.02 | 2261346 | +0.25 | 2261938 | +0.28 |
| 544.nab_r | 212418 | 212418 | +0.00 | 212418 | +0.00 | 212418 | +0.00 | 212578 | +0.08 |
| 549.fotonik3d_r | 454738 | 454738 | +0.00 | 454738 | +0.00 | 454738 | +0.00 | 454738 | +0.00 |
| 554.roms_r | 910978 | 910978 | +0.00 | 910978 | +0.00 | 910978 | +0.00 | 910978 | +0.00 |
I believe the numbers are good and thus I would like to ask-for
re-consideration of the objection and for approval to commit the patch
below. Needless to say, it has passed bootstrap and testing on
x86_64-linux.
Thanks
Martin
2017-10-27 Martin Jambor <mjambor@suse.cz>
PR target/80689
* tree-sra.h: New file.
* ipa-prop.h: Moved declaration of build_ref_for_offset to
tree-sra.h.
* expr.c: Include params.h and tree-sra.h.
(emit_move_elementwise): New function.
(store_expr_with_bounds): Optionally use it.
* ipa-cp.c: Include tree-sra.h.
* params.def (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY): New.
(PARAM_MAX_INSNS_FOR_ELEMENTWISE_COPY): Likewise.
* config/i386/i386.c (ix86_option_override_internal): Set
PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY to 35.
* tree-sra.c: Include tree-sra.h.
(scalarizable_type_p): Renamed to
simple_mix_of_records_and_arrays_p, made public, renamed the
second parameter to allow_char_arrays, added count_p parameter.
(extract_min_max_idx_from_array): New function.
(completely_scalarize): Moved bits of the function to
extract_min_max_idx_from_array.
testsuite/
* gcc.target/i386/pr80689-1.c: New test.
Added insns count param limit
---
gcc/config/i386/i386.c | 4 +
gcc/expr.c | 106 ++++++++++++++++++++++-
gcc/ipa-cp.c | 1 +
gcc/ipa-prop.h | 4 -
gcc/params.def | 12 +++
gcc/testsuite/gcc.target/i386/pr80689-1.c | 38 +++++++++
gcc/tree-sra.c | 134 +++++++++++++++++++++---------
gcc/tree-sra.h | 34 ++++++++
8 files changed, 288 insertions(+), 45 deletions(-)
create mode 100644 gcc/testsuite/gcc.target/i386/pr80689-1.c
create mode 100644 gcc/tree-sra.h
diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c
index 80c8ce7ecb9..0bff2da72dd 100644
--- a/gcc/config/i386/i386.c
+++ b/gcc/config/i386/i386.c
@@ -4580,6 +4580,10 @@ ix86_option_override_internal (bool main_args_p,
ix86_tune_cost->l2_cache_size,
opts->x_param_values,
opts_set->x_param_values);
+ maybe_set_param_value (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
+ 35,
+ opts->x_param_values,
+ opts_set->x_param_values);
/* Enable sw prefetching at -O3 for CPUS that prefetching is helpful. */
if (opts->x_flag_prefetch_loop_arrays < 0
diff --git a/gcc/expr.c b/gcc/expr.c
index 496d492c9fa..971880b635d 100644
--- a/gcc/expr.c
+++ b/gcc/expr.c
@@ -61,7 +61,8 @@ along with GCC; see the file COPYING3. If not see
#include "tree-chkp.h"
#include "rtl-chkp.h"
#include "ccmp.h"
-
+#include "params.h"
+#include "tree-sra.h"
/* If this is nonzero, we do not bother generating VOLATILE
around volatile memory references, and we are willing to
@@ -5340,6 +5341,80 @@ emit_storent_insn (rtx to, rtx from)
return maybe_expand_insn (code, 2, ops);
}
+/* Generate code for copying data of type TYPE at SOURCE plus OFFSET to TARGET
+ plus OFFSET, but do so element-wise and/or field-wise for each record and
+ array within TYPE. TYPE must either be a register type or an aggregate
+ complying with scalarizable_type_p.
+
+ If CALL_PARAM_P is nonzero, this is a store into a call param on the
+ stack, and block moves may need to be treated specially. */
+
+static void
+emit_move_elementwise (tree type, rtx target, rtx source, HOST_WIDE_INT offset,
+ int call_param_p)
+{
+ switch (TREE_CODE (type))
+ {
+ case RECORD_TYPE:
+ for (tree fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
+ if (TREE_CODE (fld) == FIELD_DECL)
+ {
+ HOST_WIDE_INT fld_offset = offset + int_bit_position (fld);
+ tree ft = TREE_TYPE (fld);
+ emit_move_elementwise (ft, target, source, fld_offset,
+ call_param_p);
+ }
+ break;
+
+ case ARRAY_TYPE:
+ {
+ tree elem_type = TREE_TYPE (type);
+ HOST_WIDE_INT el_size = tree_to_shwi (TYPE_SIZE (elem_type));
+ gcc_assert (el_size > 0);
+
+ offset_int idx, max;
+ /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
+ if (extract_min_max_idx_from_array (type, &idx, &max))
+ {
+ HOST_WIDE_INT el_offset = offset;
+ for (; idx <= max; ++idx)
+ {
+ emit_move_elementwise (elem_type, target, source, el_offset,
+ call_param_p);
+ el_offset += el_size;
+ }
+ }
+ }
+ break;
+ default:
+ machine_mode mode = TYPE_MODE (type);
+
+ rtx ntgt = adjust_address (target, mode, offset / BITS_PER_UNIT);
+ rtx nsrc = adjust_address (source, mode, offset / BITS_PER_UNIT);
+
+ /* TODO: Figure out whether the following is actually necessary. */
+ if (target == ntgt)
+ ntgt = copy_rtx (target);
+ if (source == nsrc)
+ nsrc = copy_rtx (source);
+
+ gcc_assert (mode != VOIDmode);
+ if (mode != BLKmode)
+ emit_move_insn (ntgt, nsrc);
+ else
+ {
+ /* For example vector gimple registers can end up here. */
+ rtx size = expand_expr (TYPE_SIZE_UNIT (type), NULL_RTX,
+ TYPE_MODE (sizetype), EXPAND_NORMAL);
+ emit_block_move (ntgt, nsrc, size,
+ (call_param_p
+ ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
+ }
+ break;
+ }
+ return;
+}
+
/* Generate code for computing expression EXP,
and storing the value into TARGET.
@@ -5713,9 +5788,32 @@ store_expr_with_bounds (tree exp, rtx target, int call_param_p,
emit_group_store (target, temp, TREE_TYPE (exp),
int_size_in_bytes (TREE_TYPE (exp)));
else if (GET_MODE (temp) == BLKmode)
- emit_block_move (target, temp, expr_size (exp),
- (call_param_p
- ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
+ {
+ /* Copying smallish BLKmode structures with emit_block_move and thus
+ by-pieces can result in store-to-load stalls. So copy some simple
+ small aggregates element or field-wise. */
+ int count = 0;
+ if (GET_MODE (target) == BLKmode
+ && AGGREGATE_TYPE_P (TREE_TYPE (exp))
+ && !TREE_ADDRESSABLE (TREE_TYPE (exp))
+ && tree_fits_shwi_p (TYPE_SIZE (TREE_TYPE (exp)))
+ && (tree_to_shwi (TYPE_SIZE (TREE_TYPE (exp)))
+ <= (PARAM_VALUE (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY)
+ * BITS_PER_UNIT))
+ && simple_mix_of_records_and_arrays_p (TREE_TYPE (exp), false,
+ &count)
+ && (count <= PARAM_VALUE (PARAM_MAX_INSNS_FOR_ELEMENTWISE_COPY)))
+ {
+ /* FIXME: Can this happen? What would it mean? */
+ gcc_assert (!reverse);
+ emit_move_elementwise (TREE_TYPE (exp), target, temp, 0,
+ call_param_p);
+ }
+ else
+ emit_block_move (target, temp, expr_size (exp),
+ (call_param_p
+ ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
+ }
/* If we emit a nontemporal store, there is nothing else to do. */
else if (nontemporal && emit_storent_insn (target, temp))
;
diff --git a/gcc/ipa-cp.c b/gcc/ipa-cp.c
index d23c1d8ba3e..30f91e70c22 100644
--- a/gcc/ipa-cp.c
+++ b/gcc/ipa-cp.c
@@ -124,6 +124,7 @@ along with GCC; see the file COPYING3. If not see
#include "tree-ssa-ccp.h"
#include "stringpool.h"
#include "attribs.h"
+#include "tree-sra.h"
template <typename valtype> class ipcp_value;
diff --git a/gcc/ipa-prop.h b/gcc/ipa-prop.h
index fa5bed49ee0..2313cc884ed 100644
--- a/gcc/ipa-prop.h
+++ b/gcc/ipa-prop.h
@@ -877,10 +877,6 @@ ipa_parm_adjustment *ipa_get_adjustment_candidate (tree **, bool *,
void ipa_release_body_info (struct ipa_func_body_info *);
tree ipa_get_callee_param_type (struct cgraph_edge *e, int i);
-/* From tree-sra.c: */
-tree build_ref_for_offset (location_t, tree, HOST_WIDE_INT, bool, tree,
- gimple_stmt_iterator *, bool);
-
/* In ipa-cp.c */
void ipa_cp_c_finalize (void);
diff --git a/gcc/params.def b/gcc/params.def
index 8881f4c403a..9c778f9540a 100644
--- a/gcc/params.def
+++ b/gcc/params.def
@@ -1287,6 +1287,18 @@ DEFPARAM (PARAM_VECT_EPILOGUES_NOMASK,
"Enable loop epilogue vectorization using smaller vector size.",
0, 0, 1)
+DEFPARAM (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
+ "max-size-for-elementwise-copy",
+ "Maximum size in bytes of a structure or an array to by considered "
+ "for copying by its individual fields or elements",
+ 0, 0, 512)
+
+DEFPARAM (PARAM_MAX_INSNS_FOR_ELEMENTWISE_COPY,
+ "max-insns-for-elementwise-copy",
+ "Maximum number of instructions needed to consider copying "
+ "a structure or an array by its individual fields or elements",
+ 6, 0, 64)
+
/*
Local variables:
diff --git a/gcc/testsuite/gcc.target/i386/pr80689-1.c b/gcc/testsuite/gcc.target/i386/pr80689-1.c
new file mode 100644
index 00000000000..4156d4fba45
--- /dev/null
+++ b/gcc/testsuite/gcc.target/i386/pr80689-1.c
@@ -0,0 +1,38 @@
+/* { dg-do compile } */
+/* { dg-options "-O2" } */
+
+typedef struct st1
+{
+ long unsigned int a,b;
+ long int c,d;
+}R;
+
+typedef struct st2
+{
+ int t;
+ R reg;
+}N;
+
+void Set (const R *region, N *n_info );
+
+void test(N *n_obj ,const long unsigned int a, const long unsigned int b, const long int c,const long int d)
+{
+ R reg;
+
+ reg.a=a;
+ reg.b=b;
+ reg.c=c;
+ reg.d=d;
+ Set (®, n_obj);
+
+}
+
+void Set (const R *reg, N *n_obj )
+{
+ n_obj->reg=(*reg);
+}
+
+
+/* { dg-final { scan-assembler-not "%(x|y|z)mm\[0-9\]+" } } */
+/* { dg-final { scan-assembler-not "movdqu" } } */
+/* { dg-final { scan-assembler-not "movups" } } */
diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
index bac593951e7..d06463ce21c 100644
--- a/gcc/tree-sra.c
+++ b/gcc/tree-sra.c
@@ -104,6 +104,7 @@ along with GCC; see the file COPYING3. If not see
#include "ipa-fnsummary.h"
#include "ipa-utils.h"
#include "builtins.h"
+#include "tree-sra.h"
/* Enumeration of all aggregate reductions we can do. */
enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
@@ -952,14 +953,15 @@ create_access (tree expr, gimple *stmt, bool write)
}
-/* Return true iff TYPE is scalarizable - i.e. a RECORD_TYPE or fixed-length
- ARRAY_TYPE with fields that are either of gimple register types (excluding
- bit-fields) or (recursively) scalarizable types. CONST_DECL must be true if
- we are considering a decl from constant pool. If it is false, char arrays
- will be refused. */
+/* Return true if TYPE consists of RECORD_TYPE or fixed-length ARRAY_TYPE with
+ fields/elements that are not bit-fields and are either register types or
+ recursively comply with simple_mix_of_records_and_arrays_p. Furthermore, if
+ ALLOW_CHAR_ARRAYS is false, the function will return false also if TYPE
+ contains an array of elements that only have one byte. */
-static bool
-scalarizable_type_p (tree type, bool const_decl)
+bool
+simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays,
+ int *count_p)
{
gcc_assert (!is_gimple_reg_type (type));
if (type_contains_placeholder_p (type))
@@ -976,8 +978,13 @@ scalarizable_type_p (tree type, bool const_decl)
if (DECL_BIT_FIELD (fld))
return false;
- if (!is_gimple_reg_type (ft)
- && !scalarizable_type_p (ft, const_decl))
+ if (is_gimple_reg_type (ft))
+ {
+ if (count_p)
+ (*count_p)++;
+ }
+ else if (!simple_mix_of_records_and_arrays_p (ft, allow_char_arrays,
+ count_p))
return false;
}
@@ -986,7 +993,7 @@ scalarizable_type_p (tree type, bool const_decl)
case ARRAY_TYPE:
{
HOST_WIDE_INT min_elem_size;
- if (const_decl)
+ if (allow_char_arrays)
min_elem_size = 0;
else
min_elem_size = BITS_PER_UNIT;
@@ -1007,9 +1014,45 @@ scalarizable_type_p (tree type, bool const_decl)
return false;
tree elem = TREE_TYPE (type);
- if (!is_gimple_reg_type (elem)
- && !scalarizable_type_p (elem, const_decl))
- return false;
+ if (!count_p)
+ {
+ if (!is_gimple_reg_type (elem)
+ && !simple_mix_of_records_and_arrays_p (elem, allow_char_arrays,
+ NULL))
+ return false;
+ else
+ return true;
+ }
+
+ offset_int min, max;
+ HOST_WIDE_INT ds;
+ bool nonzero = extract_min_max_idx_from_array (type, &min, &max);
+
+ if (nonzero && (min <= max))
+ {
+ offset_int d = max - min + 1;
+ if (!wi::fits_shwi_p (d))
+ return false;
+ ds = d.to_shwi ();
+ if (ds > INT_MAX)
+ return false;
+ }
+ else
+ ds = 0;
+
+ if (is_gimple_reg_type (elem))
+ *count_p += (int) ds;
+ else
+ {
+ int elc = 0;
+ if (!simple_mix_of_records_and_arrays_p (elem, allow_char_arrays,
+ &elc))
+ return false;
+ ds *= elc;
+ if (ds > INT_MAX)
+ return false;
+ *count_p += (unsigned) ds;
+ }
return true;
}
default:
@@ -1017,10 +1060,38 @@ scalarizable_type_p (tree type, bool const_decl)
}
}
-static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree, tree);
+static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree,
+ tree);
+
+/* For a given array TYPE, return false if its domain does not have any maximum
+ value. Otherwise calculate MIN and MAX indices of the first and the last
+ element. */
+
+bool
+extract_min_max_idx_from_array (tree type, offset_int *min, offset_int *max)
+{
+ tree domain = TYPE_DOMAIN (type);
+ tree minidx = TYPE_MIN_VALUE (domain);
+ gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
+ tree maxidx = TYPE_MAX_VALUE (domain);
+ if (!maxidx)
+ return false;
+ gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
+
+ /* MINIDX and MAXIDX are inclusive, and must be interpreted in
+ DOMAIN (e.g. signed int, whereas min/max may be size_int). */
+ *min = wi::to_offset (minidx);
+ *max = wi::to_offset (maxidx);
+ if (!TYPE_UNSIGNED (domain))
+ {
+ *min = wi::sext (*min, TYPE_PRECISION (domain));
+ *max = wi::sext (*max, TYPE_PRECISION (domain));
+ }
+ return true;
+}
/* Create total_scalarization accesses for all scalar fields of a member
- of type DECL_TYPE conforming to scalarizable_type_p. BASE
+ of type DECL_TYPE conforming to simple_mix_of_records_and_arrays_p. BASE
must be the top-most VAR_DECL representing the variable; within that,
OFFSET locates the member and REF must be the memory reference expression for
the member. */
@@ -1047,27 +1118,14 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
{
tree elemtype = TREE_TYPE (decl_type);
tree elem_size = TYPE_SIZE (elemtype);
- gcc_assert (elem_size && tree_fits_shwi_p (elem_size));
HOST_WIDE_INT el_size = tree_to_shwi (elem_size);
gcc_assert (el_size > 0);
- tree minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (decl_type));
- gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
- tree maxidx = TYPE_MAX_VALUE (TYPE_DOMAIN (decl_type));
+ offset_int idx, max;
/* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
- if (maxidx)
+ if (extract_min_max_idx_from_array (decl_type, &idx, &max))
{
- gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
tree domain = TYPE_DOMAIN (decl_type);
- /* MINIDX and MAXIDX are inclusive, and must be interpreted in
- DOMAIN (e.g. signed int, whereas min/max may be size_int). */
- offset_int idx = wi::to_offset (minidx);
- offset_int max = wi::to_offset (maxidx);
- if (!TYPE_UNSIGNED (domain))
- {
- idx = wi::sext (idx, TYPE_PRECISION (domain));
- max = wi::sext (max, TYPE_PRECISION (domain));
- }
for (int el_off = offset; idx <= max; ++idx)
{
tree nref = build4 (ARRAY_REF, elemtype,
@@ -1088,10 +1146,10 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
}
/* Create total_scalarization accesses for a member of type TYPE, which must
- satisfy either is_gimple_reg_type or scalarizable_type_p. BASE must be the
- top-most VAR_DECL representing the variable; within that, POS and SIZE locate
- the member, REVERSE gives its torage order. and REF must be the reference
- expression for it. */
+ satisfy either is_gimple_reg_type or simple_mix_of_records_and_arrays_p.
+ BASE must be the top-most VAR_DECL representing the variable; within that,
+ POS and SIZE locate the member, REVERSE gives its torage order. and REF must
+ be the reference expression for it. */
static void
scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
@@ -1111,7 +1169,8 @@ scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
}
/* Create a total_scalarization access for VAR as a whole. VAR must be of a
- RECORD_TYPE or ARRAY_TYPE conforming to scalarizable_type_p. */
+ RECORD_TYPE or ARRAY_TYPE conforming to
+ simple_mix_of_records_and_arrays_p. */
static void
create_total_scalarization_access (tree var)
@@ -2803,8 +2862,9 @@ analyze_all_variable_accesses (void)
{
tree var = candidate (i);
- if (VAR_P (var) && scalarizable_type_p (TREE_TYPE (var),
- constant_decl_p (var)))
+ if (VAR_P (var)
+ && simple_mix_of_records_and_arrays_p (TREE_TYPE (var),
+ constant_decl_p (var), NULL))
{
if (tree_to_uhwi (TYPE_SIZE (TREE_TYPE (var)))
<= max_scalarization_size)
diff --git a/gcc/tree-sra.h b/gcc/tree-sra.h
new file mode 100644
index 00000000000..2857688b21e
--- /dev/null
+++ b/gcc/tree-sra.h
@@ -0,0 +1,34 @@
+/* tree-sra.h - Run-time parameters.
+ Copyright (C) 2017 Free Software Foundation, Inc.
+
+This file is part of GCC.
+
+GCC is free software; you can redistribute it and/or modify it under
+the terms of the GNU General Public License as published by the Free
+Software Foundation; either version 3, or (at your option) any later
+version.
+
+GCC is distributed in the hope that it will be useful, but WITHOUT ANY
+WARRANTY; without even the implied warranty of MERCHANTABILITY or
+FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
+for more details.
+
+You should have received a copy of the GNU General Public License
+along with GCC; see the file COPYING3. If not see
+<http://www.gnu.org/licenses/>. */
+
+#ifndef TREE_SRA_H
+#define TREE_SRA_H
+
+
+bool simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays,
+ int *count_pg);
+bool extract_min_max_idx_from_array (tree type, offset_int *idx,
+ offset_int *max);
+tree build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
+ bool reverse, tree exp_type,
+ gimple_stmt_iterator *gsi, bool insert_after);
+
+
+
+#endif /* TREE_SRA_H */
--
2.14.2
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-03 16:38 ` Martin Jambor
@ 2017-11-13 13:17 ` Richard Biener
2017-11-13 13:54 ` Michael Matz
2017-11-23 16:02 ` Martin Jambor
2017-11-14 10:25 ` Martin Jambor
1 sibling, 2 replies; 26+ messages in thread
From: Richard Biener @ 2017-11-13 13:17 UTC (permalink / raw)
To: Richard Biener, GCC Patches, Jan Hubicka, Michael Matz
On Fri, Nov 3, 2017 at 5:38 PM, Martin Jambor <mjambor@suse.cz> wrote:
> Hi,
>
> On Thu, Oct 26, 2017 at 02:43:02PM +0200, Richard Biener wrote:
>> On Thu, Oct 26, 2017 at 2:18 PM, Martin Jambor <mjambor@suse.cz> wrote:
>> >
>> > Nevertheless, I still intend to experiment with the limit, I sent out
>> > this RFC exactly so that I don't spend a lot of time benchmarking
>> > something that is eventually not deemed acceptable on principle.
>>
>> I think the limit should be on the number of generated copies and not
>> the overall size of the structure... If the struct were composed of
>> 32 individual chars we wouldn't want to emit 32 loads and 32 stores...
>
> I have added another parameter to also limit the number of generated
> element copies. I have kept the size limit so that we don't even
> attempt to count them for large structures.
>
>> Given that load bandwith is usually higher than store bandwith it
>> might make sense to do the store combining in our copying sequence,
>> like for the 8 byte entry case use sth like
>>
>> movq 0(%eax), %xmm0
>> movhps 8(%eax), %xmm0 // or vpinsert
>> mov[au]ps %xmm0, 0%(ebx)
>
> I would be concerned about the cost of GPR->XMM moves when the value
> being stored is in a GPR, especially with generic tuning which (with
> -O2) is the main thing I am targeting here. Wouldn't we actually pass
> it through stack with all the associated penalties?
>
> Also, while such store combining might work for ImageMagick, if a
> programmer did:
>
> region1->x = x1;
> region2->x = x2;
> region1->y = 0;
> region2->y = 20;
> ...
> SetPixelCacheNexusPixels(cache_info, ReadMode, region1, ...)
>
> The transformation would not work unless it could prove region1 and
> region2 are not the same thing.
>
>> As said a general concern was you not copying padding. If you
>> put this into an even more common place you surely will break
>> stuff, no?
>
> I don't understand, what even more common place do you mean?
>
> I have been testing the patch also on a bunch of other architectures
> and those have tests in their testsuite that check that padding is
> copied, for example some tests in gcc.target/aarch64/aapcs64/ check
> whether a structure passed to a function is binary the same as the
> original, and the test fail because of padding. That is the only
> "breakage" I know about but I believe that the assumption that padding
> must always be is wrong (if it is not than we need to make SRA quite a
> bit more conservative).
The main concern here is that GIMPLE is not very well defined for
aggregate copies and that gimple-fold.c happily optimizes
memcpy (&a, &b, sizeof (a)) into a = b;
struct A { short s; long i; long j; };
struct A a, b;
void foo ()
{
__builtin_memcpy (&a, &b, sizeof (struct A));
}
gets folded to
MEM[(char * {ref-all})&a] = MEM[(char * {ref-all})&b];
return;
you see we're careful about TBAA but (don't see that above but
can be verified by for example debugging expand_assignment)
TREE_TYPE (MEM[...]) is actually 'struct A'.
And yes, I've been worried about SRA as well here... it _does_
have some early outs when seeing VIEW_CONVERT_EXPR but
appearantly not for the above. Testcase that aborts with SRA but
not without:
struct A { short s; long i; long j; };
struct A a, b;
void foo ()
{
struct A c;
__builtin_memcpy (&c, &b, sizeof (struct A));
__builtin_memcpy (&a, &c, sizeof (struct A));
}
int main()
{
__builtin_memset (&b, 0, sizeof (struct A));
b.s = 1;
__builtin_memcpy ((char *)&b+2, &b, 2);
foo ();
__builtin_memcpy (&a, (char *)&a+2, 2);
if (a.s != 1)
__builtin_abort ();
return 0;
}
> On Thu, Oct 26, 2017 at 05:09:42PM +0200, Richard Biener wrote:
>> Also if we do the stores in smaller chunks we are more
>> likely hitting the same store-to-load-forwarding issue
>> elsewhere. Like in case the destination is memcpy'ed
>> away.
>>
>> So the proposed change isn't necessarily a win without
>> a possible similar regression that it tries to fix.
>>
>
> With some encouragement by Honza, I have done some benchmarking anyway
> and I did not see anything of that kind.
The regression would be visible when the aggregate copy is followed by
SLP vectorized code for example. Then we'd get a vector load from
say v4si mode but had earlier 4 SImode stores -> STLF issue again.
The copying via xmm registers would have made a perfect forwarding
possibility.
I'm not saying you'll hit this in SPEC but just it's easy to construct
a case that didn't have a STLF issue but after the "fix" has.
So the fix is to _not_ split the stores but only the loads ... unless
you can do sophisticated analysis of the context.
That said, splitting the loads is fine if the CPU can handle enough loads
in flight, etc., but splitting stores is dangerous (and CPU resources on
the store side are usually more limited).
>> Whole-program analysis of accesses might allow
>> marking affected objects.
>
> Attempting to save access patterns before IPA and then tracking them
> and keep them in sync across inlining and all gimple late passes seems
> like a nightmarish task. If this approach is indeed rejected I might
> attempt to do the store combining but a WPA analysis seems just too
> complex.
Ok.
> Anyway, here are the numbers. They were taken on two different
> Zen-based machines. I am also in the process of measuring at least
> something on a Haswell machine but I started later and the machine is
> quite a bit slower so I will not have the numbers until next week (and
> not all equivalents in any way). I found out I do not have access to
> any more modern .*Lake intel CPU.
>
> trunk is pristine trunk revision 254205. All benchmarks were run
> three times and the median was chosen.
>
> s or strict means the patch with the strictest possible settings to
> speed-up ImageMagick, i.e. --param max-size-for-elementwise-copy=32
> --param max-insns-for-elementwise-copy=4. Also run three times.
>
> x1 is patched trunk with the parameters having the default values was
> going to propose, i.e. --param max-size-for-elementwise-copy=35
> --param max-insns-for-elementwise-copy=6. Also run three times.
>
> I then increased the parameter, in search for further missed
> opportunities and to see what and how soon will start to regress.
> x2 is roughly twice that, --param max-size-for-elementwise-copy=67
> --param max-insns-for-elementwise-copy=12. Run twice, outliers
> manually checked.
>
> x4 is roughly four times x1, namely --param max-size-for-elementwise-copy=143
> --param max-insns-for-elementwise-copy=24. Run only once.
>
> The times below are of course "non-reportable," for a whole bunch of
> reasons.
>
>
> Zen SPECINT 2006 -O2 generic tuning
> ====================================
>
> Run-time
> --------
>
> | Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
> |----------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
> | 400.perlbench | 237 | 236 | -0.42 | 236 | -0.42 | 238 | +0.42 | 237 | +0.00 |
> | 401.bzip2 | 341 | 342 | +0.29 | 341 | +0.00 | 341 | +0.00 | 341 | +0.00 |
> | 403.gcc | 217 | 217 | +0.00 | 217 | +0.00 | 216 | -0.46 | 217 | +0.00 |
> | 429.mcf | 224 | 218 | -2.68 | 223 | -0.45 | 221 | -1.34 | 226 | +0.89 |
> | 445.gobmk | 361 | 361 | +0.00 | 361 | +0.00 | 360 | -0.28 | 363 | +0.55 |
> | 456.hmmer | 296 | 296 | +0.00 | 296 | +0.00 | 297 | +0.34 | 296 | +0.00 |
> | 458.sjeng | 453 | 452 | -0.22 | 454 | +0.22 | 454 | +0.22 | 460 | +1.55 |
> | 462.libquantum | 289 | 289 | +0.00 | 291 | +0.69 | 289 | +0.00 | 291 | +0.69 |
> | 464.h264ref | 391 | 391 | +0.00 | 385 | -1.53 | 385 | -1.53 | 385 | -1.53 |
> | 471.omnetpp | 269 | 255 | -5.20 | 250 | -7.06 | 247 | -8.18 | 268 | -0.37 |
> | 473.astar | 320 | 321 | +0.31 | 317 | -0.94 | 320 | +0.00 | 320 | +0.00 |
> | 483.xalancbmk | 187 | 188 | +0.53 | 188 | +0.53 | 187 | +0.00 | 187 | +0.00 |
>
> Although the omnetpp looks like a sizeable improvement I should warn
> that this is one of the few slightly jumpy benchmarks. However, I
> re-run it a few more times and it seems like it is jumping around a
> lower value when compiled with the patched compiler. It might not be
> the 5-8% though.
>
> Text size
> ---------
>
> | Benchmark | trunk | struict | % | x1 | % | x2 | % | x4 | % |
> |----------------+---------+---------+-------+---------+-------+---------+-------+---------+-------|
> | 400.perlbench | 875874 | 875954 | +0.01 | 875954 | +0.01 | 876018 | +0.02 | 876146 | +0.03 |
> | 401.bzip2 | 44754 | 44754 | +0.00 | 44754 | +0.00 | 44754 | +0.00 | 44754 | +0.00 |
> | 403.gcc | 2294466 | 2294930 | +0.02 | 2296098 | +0.07 | 2296306 | +0.08 | 2296466 | +0.09 |
> | 429.mcf | 8226 | 8226 | +0.00 | 8226 | +0.00 | 8258 | +0.39 | 8258 | +0.39 |
> | 445.gobmk | 579778 | 579778 | +0.00 | 579826 | +0.01 | 579826 | +0.01 | 580402 | +0.11 |
> | 456.hmmer | 221058 | 221058 | +0.00 | 221058 | +0.00 | 221058 | +0.00 | 221058 | +0.00 |
> | 458.sjeng | 93362 | 93362 | +0.00 | 94882 | +1.63 | 94882 | +1.63 | 96066 | +2.90 |
> | 462.libquantum | 28314 | 28314 | +0.00 | 28362 | +0.17 | 28362 | +0.17 | 28362 | +0.17 |
> | 464.h264ref | 393874 | 393874 | +0.00 | 393922 | +0.01 | 393922 | +0.01 | 394226 | +0.09 |
> | 471.omnetpp | 430306 | 430306 | +0.00 | 430418 | +0.03 | 430418 | +0.03 | 430418 | +0.03 |
> | 473.astar | 29362 | 29538 | +0.60 | 29538 | +0.60 | 29554 | +0.65 | 29554 | +0.65 |
> | 483.xalancbmk | 2361298 | 2361506 | +0.01 | 2361506 | +0.01 | 2361506 | +0.01 | 2361506 | +0.01 |
>
>
>
> Zen SPECINT 2006 -Ofast native tuning
> ======================================
>
> Run-time
> --------
>
> | Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
> |----------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
> | 400.perlbench | 240 | 239 | -0.42 | 239 | -0.42 | 241 | +0.42 | 238 | -0.83 |
> | 401.bzip2 | 341 | 341 | +0.00 | 341 | +0.00 | 341 | +0.00 | 340 | -0.29 |
> | 403.gcc | 210 | 208 | -0.95 | 207 | -1.43 | 209 | -0.48 | 208 | -0.95 |
> | 429.mcf | 225 | 225 | +0.00 | 225 | +0.00 | 228 | +1.33 | 226 | +0.44 |
> | 445.gobmk | 352 | 352 | +0.00 | 352 | +0.00 | 351 | -0.28 | 352 | +0.00 |
> | 456.hmmer | 131 | 131 | +0.00 | 131 | +0.00 | 131 | +0.00 | 131 | +0.00 |
> | 458.sjeng | 442 | 442 | +0.00 | 438 | -0.90 | 438 | -0.90 | 437 | -1.13 |
> | 462.libquantum | 291 | 292 | +0.34 | 286 | -1.72 | 287 | -1.37 | 287 | -1.37 |
> | 464.h264ref | 364 | 365 | +0.27 | 364 | +0.00 | 364 | +0.00 | 363 | -0.27 |
> | 471.omnetpp | 266 | 266 | +0.00 | 265 | -0.38 | 265 | -0.38 | 265 | -0.38 |
> | 473.astar | 306 | 307 | +0.33 | 306 | +0.00 | 306 | +0.00 | 309 | +0.98 |
> | 483.xalancbmk | 177 | 173 | -2.26 | 170 | -3.95 | 170 | -3.95 | 170 | -3.95 |
>
> Text size
> ---------
>
> | Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
> |----------------+---------+---------+-------+---------+-------+---------+-------+---------+-------|
> | 400.perlbench | 1161762 | 1161874 | +0.01 | 1161874 | +0.01 | 1162226 | +0.04 | 1162338 | +0.05 |
> | 401.bzip2 | 80834 | 80834 | +0.00 | 80834 | +0.00 | 80834 | +0.00 | 80834 | +0.00 |
> | 403.gcc | 3170946 | 3171394 | +0.01 | 3172914 | +0.06 | 3173170 | +0.07 | 3174818 | +0.12 |
> | 429.mcf | 10418 | 10418 | +0.00 | 10418 | +0.00 | 10450 | +0.31 | 10450 | +0.31 |
> | 445.gobmk | 779778 | 779778 | +0.00 | 779842 | +0.01 | 779842 | +0.01 | 780418 | +0.08 |
> | 456.hmmer | 328258 | 328258 | +0.00 | 328258 | +0.00 | 328258 | +0.00 | 328258 | +0.00 |
> | 458.sjeng | 146386 | 146386 | +0.00 | 148162 | +1.21 | 148162 | +1.21 | 149330 | +2.01 |
> | 462.libquantum | 30666 | 30666 | +0.00 | 30730 | +0.21 | 30730 | +0.21 | 30730 | +0.21 |
> | 464.h264ref | 737826 | 737826 | +0.00 | 737890 | +0.01 | 737890 | +0.01 | 739186 | +0.18 |
> | 471.omnetpp | 561570 | 561570 | +0.00 | 561826 | +0.05 | 561826 | +0.05 | 561826 | +0.05 |
> | 473.astar | 39314 | 39522 | +0.53 | 39522 | +0.53 | 39538 | +0.57 | 39538 | +0.57 |
> | 483.xalancbmk | 3319682 | 3319842 | +0.00 | 3319842 | +0.00 | 3319842 | +0.00 | 3319842 | +0.00 |
>
>
>
> Zen SPECFP 2006 -O2 generic tuning
> ==================================
>
> Run-time
> --------
>
> | Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
> |---------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
> | 410.bwaves | 214 | 213 | -0.47 | 214 | +0.00 | 214 | +0.00 | 214 | +0.00 |
> | 433.milc | 290 | 291 | +0.34 | 290 | +0.00 | 295 | +1.72 | 289 | -0.34 |
> | 434.zeusmp | 182 | 182 | +0.00 | 182 | +0.00 | 184 | +1.10 | 182 | +0.00 |
> | 435.gromacs | 218 | 218 | +0.00 | 217 | -0.46 | 216 | -0.92 | 220 | +0.92 |
> | 436.cactusADM | 350 | 349 | -0.29 | 349 | -0.29 | 343 | -2.00 | 349 | -0.29 |
> | 437.leslie3d | 196 | 195 | -0.51 | 196 | +0.00 | 194 | -1.02 | 196 | +0.00 |
> | 444.namd | 273 | 273 | +0.00 | 273 | +0.00 | 273 | +0.00 | 273 | +0.00 |
> | 447.dealII | 211 | 211 | +0.00 | 210 | -0.47 | 210 | -0.47 | 211 | +0.00 |
> | 450.soplex | 187 | 188 | +0.53 | 188 | +0.53 | 187 | +0.00 | 187 | +0.00 |
> | 453.povray | 119 | 118 | -0.84 | 119 | +0.00 | 119 | +0.00 | 118 | -0.84 |
> | 454.calculix | 534 | 533 | -0.19 | 531 | -0.56 | 531 | -0.56 | 532 | -0.37 |
> | 459.GemsFDTD | 236 | 235 | -0.42 | 235 | -0.42 | 242 | +2.54 | 237 | +0.42 |
> | 465.tonto | 366 | 365 | -0.27 | 365 | -0.27 | 364 | -0.55 | 365 | -0.27 |
> | 470.lbm | 181 | 180 | -0.55 | 180 | -0.55 | 180 | -0.55 | 180 | -0.55 |
> | 481.wrf | 303 | 303 | +0.00 | 302 | -0.33 | 304 | +0.33 | 304 | +0.33 |
> | 482.sphinx3 | 362 | 362 | +0.00 | 360 | -0.55 | 361 | -0.28 | 363 | +0.28 |
>
> Text size
> ---------
>
> | Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
> |---------------+---------+---------+-------+---------+-------+---------+-------+---------+-------|
> | 410.bwaves | 25954 | 25954 | +0.00 | 25954 | +0.00 | 25954 | +0.00 | 25954 | +0.00 |
> | 433.milc | 87922 | 87922 | +0.00 | 87922 | +0.00 | 88610 | +0.78 | 89042 | +1.27 |
> | 434.zeusmp | 212034 | 212034 | +0.00 | 212034 | +0.00 | 212034 | +0.00 | 212034 | +0.00 |
> | 435.gromacs | 747026 | 747026 | +0.00 | 747026 | +0.00 | 747026 | +0.00 | 747026 | +0.00 |
> | 436.cactusADM | 526178 | 526178 | +0.00 | 526178 | +0.00 | 526274 | +0.02 | 526274 | +0.02 |
> | 437.leslie3d | 83234 | 83234 | +0.00 | 83234 | +0.00 | 83234 | +0.00 | 83234 | +0.00 |
> | 444.namd | 297234 | 297266 | +0.01 | 297266 | +0.01 | 297266 | +0.01 | 297266 | +0.01 |
> | 447.dealII | 2165282 | 2167650 | +0.11 | 2172290 | +0.32 | 2174034 | +0.40 | 2174082 | +0.41 |
> | 450.soplex | 347122 | 347122 | +0.00 | 347122 | +0.00 | 347122 | +0.00 | 347122 | +0.00 |
> | 453.povray | 800914 | 800962 | +0.01 | 801570 | +0.08 | 802002 | +0.14 | 803138 | +0.28 |
> | 454.calculix | 1342802 | 1342802 | +0.00 | 1342802 | +0.00 | 1342802 | +0.00 | 1342802 | +0.00 |
> | 459.GemsFDTD | 353410 | 354050 | +0.18 | 354050 | +0.18 | 354050 | +0.18 | 354098 | +0.19 |
> | 465.tonto | 3464210 | 3465058 | +0.02 | 3465058 | +0.02 | 3468434 | +0.12 | 3476594 | +0.36 |
> | 470.lbm | 9202 | 9202 | +0.00 | 9202 | +0.00 | 9202 | +0.00 | 9202 | +0.00 |
> | 481.wrf | 3345170 | 3345170 | +0.00 | 3345170 | +0.00 | 3351586 | +0.19 | 3351586 | +0.19 |
> | 482.sphinx3 | 125026 | 125026 | +0.00 | 125026 | +0.00 | 125026 | +0.00 | 125026 | +0.00 |
>
>
>
> Zen SPECFP 2006 -Ofast native tuning
> ====================================
>
> Run-time
> --------
>
> | Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
> |---------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
> | 410.bwaves | 151 | 150 | -0.66 | 151 | +0.00 | 151 | +0.00 | 151 | +0.00 |
> | 433.milc | 197 | 197 | +0.00 | 197 | +0.00 | 194 | -1.52 | 186 | -5.58 |
> | 434.zeusmp | 128 | 128 | +0.00 | 128 | +0.00 | 128 | +0.00 | 128 | +0.00 |
> | 435.gromacs | 181 | 181 | +0.00 | 180 | -0.55 | 180 | -0.55 | 181 | +0.00 |
> | 436.cactusADM | 139 | 139 | +0.00 | 139 | +0.00 | 132 | -5.04 | 139 | +0.00 |
> | 437.leslie3d | 159 | 160 | +0.63 | 160 | +0.63 | 159 | +0.00 | 159 | +0.00 |
> | 444.namd | 256 | 256 | +0.00 | 255 | -0.39 | 255 | -0.39 | 256 | +0.00 |
> | 447.dealII | 200 | 200 | +0.00 | 199 | -0.50 | 201 | +0.50 | 201 | +0.50 |
> | 450.soplex | 184 | 184 | +0.00 | 185 | +0.54 | 184 | +0.00 | 184 | +0.00 |
> | 453.povray | 124 | 122 | -1.61 | 123 | -0.81 | 124 | +0.00 | 122 | -1.61 |
> | 454.calculix | 192 | 192 | +0.00 | 192 | +0.00 | 193 | +0.52 | 193 | +0.52 |
> | 459.GemsFDTD | 208 | 208 | +0.00 | 208 | +0.00 | 214 | +2.88 | 208 | +0.00 |
> | 465.tonto | 320 | 320 | +0.00 | 320 | +0.00 | 320 | +0.00 | 320 | +0.00 |
> | 470.lbm | 142 | 142 | +0.00 | 142 | +0.00 | 142 | +0.00 | 142 | +0.00 |
> | 481.wrf | 195 | 195 | +0.00 | 195 | +0.00 | 195 | +0.00 | 195 | +0.00 |
> | 482.sphinx3 | 256 | 258 | +0.78 | 256 | +0.00 | 256 | +0.00 | 257 | +0.39 |
>
> Text size
> ---------
>
> | Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
> |---------------+---------+---------+-------+---------+-------+---------+-------+---------+-------|
> | 410.bwaves | 27490 | 27490 | +0.00 | 27490 | +0.00 | 27490 | +0.00 | 27490 | +0.00 |
> | 433.milc | 118178 | 118178 | +0.00 | 118178 | +0.00 | 118962 | +0.66 | 119634 | +1.23 |
> | 434.zeusmp | 411106 | 411106 | +0.00 | 411106 | +0.00 | 411106 | +0.00 | 411106 | +0.00 |
> | 435.gromacs | 935970 | 935970 | +0.00 | 935970 | +0.00 | 935970 | +0.00 | 936162 | +0.02 |
> | 436.cactusADM | 750546 | 750546 | +0.00 | 750546 | +0.00 | 750626 | +0.01 | 750626 | +0.01 |
> | 437.leslie3d | 123410 | 123410 | +0.00 | 123410 | +0.00 | 123410 | +0.00 | 123410 | +0.00 |
> | 444.namd | 284082 | 284114 | +0.01 | 284114 | +0.01 | 284114 | +0.01 | 284114 | +0.01 |
> | 447.dealII | 2438610 | 2440946 | +0.10 | 2444978 | +0.26 | 2446882 | +0.34 | 2446930 | +0.34 |
> | 450.soplex | 443218 | 443218 | +0.00 | 443218 | +0.00 | 443218 | +0.00 | 443218 | +0.00 |
> | 453.povray | 1077778 | 1077890 | +0.01 | 1078658 | +0.08 | 1079026 | +0.12 | 1080370 | +0.24 |
> | 454.calculix | 1639138 | 1639138 | +0.00 | 1639138 | +0.00 | 1639474 | +0.02 | 1639474 | +0.02 |
> | 459.GemsFDTD | 451202 | 451234 | +0.01 | 451234 | +0.01 | 451234 | +0.01 | 451282 | +0.02 |
> | 465.tonto | 4584690 | 4585250 | +0.01 | 4585250 | +0.01 | 4588130 | +0.08 | 4595442 | +0.23 |
> | 470.lbm | 9858 | 9858 | +0.00 | 9858 | +0.00 | 9858 | +0.00 | 9858 | +0.00 |
> | 481.wrf | 4588002 | 4588002 | +0.00 | 4588290 | +0.01 | 4621010 | +0.72 | 4621922 | +0.74 |
> | 482.sphinx3 | 179602 | 179602 | +0.00 | 179602 | +0.00 | 179602 | +0.00 | 179602 | +0.00 |
>
>
>
> Zen SPEC INT 2017 -O2 generic tuning
> ====================================
>
> Run-time
> --------
>
> | Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
> |-----------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
> | 500.perlbench_r | 529 | 529 | +0.00 | 531 | +0.38 | 530 | +0.19 | 534 | +0.95 |
> | 502.gcc_r | 338 | 333 | -1.48 | 334 | -1.18 | 339 | +0.30 | 339 | +0.30 |
> | 505.mcf_r | 382 | 381 | -0.26 | 382 | +0.00 | 382 | +0.00 | 381 | -0.26 |
> | 520.omnetpp_r | 511 | 503 | -1.57 | 497 | -2.74 | 497 | -2.74 | 497 | -2.74 |
> | 523.xalancbmk_r | 391 | 388 | -0.77 | 389 | -0.51 | 390 | -0.26 | 391 | +0.00 |
> | 525.x264_r | 590 | 590 | +0.00 | 591 | +0.17 | 592 | +0.34 | 593 | +0.51 |
> | 531.deepsjeng_r | 427 | 427 | +0.00 | 427 | +0.00 | 428 | +0.23 | 427 | +0.00 |
> | 541.leela_r | 716 | 716 | +0.00 | 716 | +0.00 | 719 | +0.42 | 719 | +0.42 |
> | 548.exchange2_r | 593 | 593 | +0.00 | 593 | +0.00 | 593 | +0.00 | 593 | +0.00 |
> | 557.xz_r | 452 | 452 | +0.00 | 453 | +0.22 | 454 | +0.44 | 452 | +0.00 |
>
> Text size
> ---------
>
> | Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
> |-----------------+---------+---------+-------+---------+-------+---------+-------+---------+-------|
> | 500.perlbench_r | 1599442 | 1599522 | +0.01 | 1599522 | +0.01 | 1599522 | +0.01 | 1600082 | +0.04 |
> | 502.gcc_r | 6757602 | 6758978 | +0.02 | 6759090 | +0.02 | 6759842 | +0.03 | 6760306 | +0.04 |
> | 505.mcf_r | 16098 | 16098 | +0.00 | 16098 | +0.00 | 16098 | +0.00 | 16306 | +1.29 |
> | 520.omnetpp_r | 1262498 | 1262562 | +0.01 | 1264034 | +0.12 | 1264034 | +0.12 | 1264034 | +0.12 |
> | 523.xalancbmk_r | 3989026 | 3989202 | +0.00 | 3989202 | +0.00 | 3989202 | +0.00 | 3989202 | +0.00 |
> | 525.x264_r | 414130 | 414194 | +0.02 | 414194 | +0.02 | 414738 | +0.15 | 415122 | +0.24 |
> | 531.deepsjeng_r | 67426 | 67426 | +0.00 | 67458 | +0.05 | 67458 | +0.05 | 67458 | +0.05 |
> | 541.leela_r | 219378 | 219378 | +0.00 | 219378 | +0.00 | 224082 | +2.14 | 237026 | +8.04 |
> | 548.exchange2_r | 61234 | 61234 | +0.00 | 61234 | +0.00 | 61234 | +0.00 | 61234 | +0.00 |
> | 557.xz_r | 111490 | 111490 | +0.00 | 111490 | +0.00 | 111506 | +0.01 | 111890 | +0.36 |
>
>
>
> Zen SPEC INT 2017 -Ofast native tuning
> ======================================
>
> Run-time
> ---------
>
> | Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
> |-----------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
> | 500.perlbench_r | 525 | 524 | -0.19 | 525 | +0.00 | 525 | +0.00 | 534 | +1.71 |
> | 502.gcc_r | 331 | 329 | -0.60 | 324 | -2.11 | 330 | -0.30 | 324 | -2.11 |
> | 505.mcf_r | 380 | 380 | +0.00 | 381 | +0.26 | 380 | +0.00 | 379 | -0.26 |
> | 520.omnetpp_r | 487 | 486 | -0.21 | 488 | +0.21 | 489 | +0.41 | 488 | +0.21 |
> | 523.xalancbmk_r | 373 | 369 | -1.07 | 367 | -1.61 | 370 | -0.80 | 368 | -1.34 |
> | 525.x264_r | 319 | 319 | +0.00 | 320 | +0.31 | 321 | +0.63 | 322 | +0.94 |
> | 531.deepsjeng_r | 418 | 418 | +0.00 | 418 | +0.00 | 418 | +0.00 | 419 | +0.24 |
> | 541.leela_r | 674 | 674 | +0.00 | 674 | +0.00 | 672 | -0.30 | 672 | -0.30 |
> | 548.exchange2_r | 466 | 466 | +0.00 | 466 | +0.00 | 466 | +0.00 | 466 | +0.00 |
> | 557.xz_r | 443 | 443 | +0.00 | 443 | +0.00 | 449 | +1.35 | 449 | +1.35 |
>
> Text size
> ---------
>
> | Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
> |-----------------+---------+---------+-------+---------+-------+---------+-------+---------+-------|
> | 500.perlbench_r | 2122882 | 2122962 | +0.00 | 2122962 | +0.00 | 2122962 | +0.00 | 2122514 | -0.02 |
> | 502.gcc_r | 8566290 | 8567794 | +0.02 | 8569138 | +0.03 | 8570066 | +0.04 | 8570642 | +0.05 |
> | 505.mcf_r | 26770 | 26770 | +0.00 | 26770 | +0.00 | 26770 | +0.00 | 26962 | +0.72 |
> | 520.omnetpp_r | 1713938 | 1713954 | +0.00 | 1714754 | +0.05 | 1714754 | +0.05 | 1714754 | +0.05 |
> | 523.xalancbmk_r | 4881890 | 4882114 | +0.00 | 4882114 | +0.00 | 4882114 | +0.00 | 4882114 | +0.00 |
> | 525.x264_r | 601522 | 601602 | +0.01 | 601602 | +0.01 | 602130 | +0.10 | 602834 | +0.22 |
> | 531.deepsjeng_r | 90306 | 90306 | +0.00 | 90338 | +0.04 | 90338 | +0.04 | 90338 | +0.04 |
> | 541.leela_r | 277634 | 277650 | +0.01 | 277650 | +0.01 | 282386 | +1.71 | 295778 | +6.54 |
> | 548.exchange2_r | 109058 | 109058 | +0.00 | 109058 | +0.00 | 109058 | +0.00 | 109058 | +0.00 |
> | 557.xz_r | 154594 | 154594 | +0.00 | 154594 | +0.00 | 154610 | +0.01 | 154930 | +0.22 |
>
>
>
> Zen SPEC 2017 FP -O2 generic tuning
> ===================================
>
> Run-time
> --------
> | Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
> |-----------------+-------+-----+--------+-----+--------+-----+--------+-----+--------|
> | 503.bwaves_r | 801 | 801 | +0.00 | 801 | +0.00 | 801 | +0.00 | 801 | +0.00 |
> | 507.cactuBSSN_r | 303 | 302 | -0.33 | 299 | -1.32 | 302 | -0.33 | 307 | +1.32 |
> | 508.namd_r | 306 | 306 | +0.00 | 307 | +0.33 | 306 | +0.00 | 306 | +0.00 |
> | 510.parest_r | 558 | 553 | -0.90 | 561 | +0.54 | 554 | -0.72 | 562 | +0.72 |
> | 511.povray_r | 679 | 672 | -1.03 | 673 | -0.88 | 680 | +0.15 | 644 | -5.15 |
> | 519.lbm_r | 240 | 240 | +0.00 | 240 | +0.00 | 240 | +0.00 | 240 | +0.00 |
> | 521.wrf_r | 851 | 827 | -2.82 | 827 | -2.82 | 827 | -2.82 | 828 | -2.70 |
> | 526.blender_r | 376 | 376 | +0.00 | 379 | +0.80 | 377 | +0.27 | 376 | +0.00 |
> | 527.cam4_r | 529 | 527 | -0.38 | 533 | +0.76 | 536 | +1.32 | 528 | -0.19 |
> | 538.imagick_r | 646 | 570 | -11.76 | 570 | -11.76 | 569 | -11.92 | 570 | -11.76 |
> | 544.nab_r | 467 | 467 | +0.00 | 467 | +0.00 | 467 | +0.00 | 467 | +0.00 |
> | 549.fotonik3d_r | 413 | 413 | +0.00 | 414 | +0.24 | 415 | +0.48 | 413 | +0.00 |
> | 554.roms_r | 459 | 455 | -0.87 | 456 | -0.65 | 456 | -0.65 | 456 | -0.65 |
>
> Text size
> ---------
>
> | Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
> |-----------------+----------+----------+-------+----------+-------+----------+-------+----------+-------|
> | 503.bwaves_r | 32034 | 32034 | +0.00 | 32034 | +0.00 | 32034 | +0.00 | 32034 | +0.00 |
> | 507.cactuBSSN_r | 2951634 | 2951634 | +0.00 | 2951634 | +0.00 | 2951698 | +0.00 | 2951730 | +0.00 |
> | 508.namd_r | 837458 | 837490 | +0.00 | 837490 | +0.00 | 837490 | +0.00 | 837490 | +0.00 |
> | 510.parest_r | 6540866 | 6545618 | +0.07 | 6546754 | +0.09 | 6561426 | +0.31 | 6569426 | +0.44 |
> | 511.povray_r | 803618 | 803666 | +0.01 | 804274 | +0.08 | 804706 | +0.14 | 805842 | +0.28 |
> | 519.lbm_r | 12018 | 12018 | +0.00 | 12018 | +0.00 | 12018 | +0.00 | 12018 | +0.00 |
> | 521.wrf_r | 16292962 | 16296786 | +0.02 | 16296978 | +0.02 | 16302594 | +0.06 | 16419842 | +0.78 |
> | 526.blender_r | 7268224 | 7281264 | +0.18 | 7282608 | +0.20 | 7289168 | +0.29 | 7295296 | +0.37 |
> | 527.cam4_r | 5063666 | 5063922 | +0.01 | 5065010 | +0.03 | 5068114 | +0.09 | 5072946 | +0.18 |
> | 538.imagick_r | 1608178 | 1609282 | +0.07 | 1609282 | +0.07 | 1613458 | +0.33 | 1613970 | +0.36 |
> | 544.nab_r | 156242 | 156242 | +0.00 | 156242 | +0.00 | 156242 | +0.00 | 156242 | +0.00 |
> | 549.fotonik3d_r | 326738 | 326738 | +0.00 | 326738 | +0.00 | 326738 | +0.00 | 326738 | +0.00 |
> | 554.roms_r | 728546 | 728546 | +0.00 | 728546 | +0.00 | 728546 | +0.00 | 728546 | +0.00 |
>
>
>
> Zen SPEC 2017 FP -Ofast native tuning
> =====================================
>
> Run-time
> --------
>
> | Benchmark | trunk | s | % | x1 | % | x2 | % | x4 | % |
> |-----------------+-------+-----+-------+-----+-------+-----+-------+-----+-------|
> | 503.bwaves_r | 310 | 310 | +0.00 | 310 | +0.00 | 310 | +0.00 | 309 | -0.32 |
> | 507.cactuBSSN_r | 269 | 266 | -1.12 | 266 | -1.12 | 268 | -0.37 | 270 | +0.37 |
> | 508.namd_r | 270 | 269 | -0.37 | 269 | -0.37 | 268 | -0.74 | 268 | -0.74 |
> | 510.parest_r | 607 | 601 | -0.99 | 599 | -1.32 | 599 | -1.32 | 604 | -0.49 |
> | 511.povray_r | 662 | 664 | +0.30 | 671 | +1.36 | 680 | +2.72 | 675 | +1.96 |
> | 519.lbm_r | 186 | 186 | +0.00 | 186 | +0.00 | 186 | +0.00 | 186 | +0.00 |
> | 521.wrf_r | 550 | 554 | +0.73 | 550 | +0.00 | 550 | +0.00 | 549 | -0.18 |
> | 526.blender_r | 355 | 354 | -0.28 | 355 | +0.00 | 354 | -0.28 | 354 | -0.28 |
> | 527.cam4_r | 434 | 437 | +0.69 | 435 | +0.23 | 437 | +0.69 | 435 | +0.23 |
> | 538.imagick_r | 433 | 420 | -3.00 | 420 | -3.00 | 420 | -3.00 | 419 | -3.23 |
> | 544.nab_r | 424 | 425 | +0.24 | 425 | +0.24 | 425 | +0.24 | 425 | +0.24 |
> | 549.fotonik3d_r | 421 | 422 | +0.24 | 422 | +0.24 | 422 | +0.24 | 422 | +0.24 |
> | 554.roms_r | 360 | 361 | +0.28 | 361 | +0.28 | 361 | +0.28 | 361 | +0.28 |
>
> +1.36% for 511.povray_r is the worst regression for the proposed x1
> defaults, by the way. I have not investigated it further, however.
>
> Text size
> ---------
>
> | Benchmark | trunk | strict | % | x1 | % | x2 | % | x4 | % |
> |-----------------+----------+----------+-------+----------+-------+----------+-------+----------+-------|
> | 503.bwaves_r | 34562 | 34562 | +0.00 | 34562 | +0.00 | 34562 | +0.00 | 34562 | +0.00 |
> | 507.cactuBSSN_r | 3978402 | 3978402 | +0.00 | 3978402 | +0.00 | 3978514 | +0.00 | 3978546 | +0.00 |
> | 508.namd_r | 869106 | 869154 | +0.01 | 869154 | +0.01 | 869154 | +0.01 | 869154 | +0.01 |
> | 510.parest_r | 7186258 | 7189298 | +0.04 | 7190370 | +0.06 | 7203890 | +0.25 | 7211202 | +0.35 |
> | 511.povray_r | 1063314 | 1063410 | +0.01 | 1064178 | +0.08 | 1064546 | +0.12 | 1065890 | +0.24 |
> | 519.lbm_r | 12178 | 12178 | +0.00 | 12178 | +0.00 | 12178 | +0.00 | 12178 | +0.00 |
> | 521.wrf_r | 19480946 | 19484146 | +0.02 | 19484466 | +0.02 | 19607538 | +0.65 | 19716178 | +1.21 |
> | 526.blender_r | 9708752 | 9719952 | +0.12 | 9722768 | +0.14 | 9730224 | +0.22 | 9737760 | +0.30 |
> | 527.cam4_r | 6217970 | 6218162 | +0.00 | 6219570 | +0.03 | 6223362 | +0.09 | 6227762 | +0.16 |
> | 538.imagick_r | 2255682 | 2256162 | +0.02 | 2256162 | +0.02 | 2261346 | +0.25 | 2261938 | +0.28 |
> | 544.nab_r | 212418 | 212418 | +0.00 | 212418 | +0.00 | 212418 | +0.00 | 212578 | +0.08 |
> | 549.fotonik3d_r | 454738 | 454738 | +0.00 | 454738 | +0.00 | 454738 | +0.00 | 454738 | +0.00 |
> | 554.roms_r | 910978 | 910978 | +0.00 | 910978 | +0.00 | 910978 | +0.00 | 910978 | +0.00 |
>
>
> I believe the numbers are good and thus I would like to ask-for
> re-consideration of the objection and for approval to commit the patch
> below. Needless to say, it has passed bootstrap and testing on
> x86_64-linux.
>
> Thanks
>
> Martin
>
>
> 2017-10-27 Martin Jambor <mjambor@suse.cz>
>
> PR target/80689
> * tree-sra.h: New file.
> * ipa-prop.h: Moved declaration of build_ref_for_offset to
> tree-sra.h.
> * expr.c: Include params.h and tree-sra.h.
> (emit_move_elementwise): New function.
> (store_expr_with_bounds): Optionally use it.
> * ipa-cp.c: Include tree-sra.h.
> * params.def (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY): New.
> (PARAM_MAX_INSNS_FOR_ELEMENTWISE_COPY): Likewise.
> * config/i386/i386.c (ix86_option_override_internal): Set
> PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY to 35.
> * tree-sra.c: Include tree-sra.h.
> (scalarizable_type_p): Renamed to
> simple_mix_of_records_and_arrays_p, made public, renamed the
> second parameter to allow_char_arrays, added count_p parameter.
> (extract_min_max_idx_from_array): New function.
> (completely_scalarize): Moved bits of the function to
> extract_min_max_idx_from_array.
>
> testsuite/
> * gcc.target/i386/pr80689-1.c: New test.
>
> Added insns count param limit
> ---
> gcc/config/i386/i386.c | 4 +
> gcc/expr.c | 106 ++++++++++++++++++++++-
> gcc/ipa-cp.c | 1 +
> gcc/ipa-prop.h | 4 -
> gcc/params.def | 12 +++
> gcc/testsuite/gcc.target/i386/pr80689-1.c | 38 +++++++++
> gcc/tree-sra.c | 134 +++++++++++++++++++++---------
> gcc/tree-sra.h | 34 ++++++++
> 8 files changed, 288 insertions(+), 45 deletions(-)
> create mode 100644 gcc/testsuite/gcc.target/i386/pr80689-1.c
> create mode 100644 gcc/tree-sra.h
>
> diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c
> index 80c8ce7ecb9..0bff2da72dd 100644
> --- a/gcc/config/i386/i386.c
> +++ b/gcc/config/i386/i386.c
> @@ -4580,6 +4580,10 @@ ix86_option_override_internal (bool main_args_p,
> ix86_tune_cost->l2_cache_size,
> opts->x_param_values,
> opts_set->x_param_values);
> + maybe_set_param_value (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
> + 35,
> + opts->x_param_values,
> + opts_set->x_param_values);
>
> /* Enable sw prefetching at -O3 for CPUS that prefetching is helpful. */
> if (opts->x_flag_prefetch_loop_arrays < 0
> diff --git a/gcc/expr.c b/gcc/expr.c
> index 496d492c9fa..971880b635d 100644
> --- a/gcc/expr.c
> +++ b/gcc/expr.c
> @@ -61,7 +61,8 @@ along with GCC; see the file COPYING3. If not see
> #include "tree-chkp.h"
> #include "rtl-chkp.h"
> #include "ccmp.h"
> -
> +#include "params.h"
> +#include "tree-sra.h"
>
> /* If this is nonzero, we do not bother generating VOLATILE
> around volatile memory references, and we are willing to
> @@ -5340,6 +5341,80 @@ emit_storent_insn (rtx to, rtx from)
> return maybe_expand_insn (code, 2, ops);
> }
>
> +/* Generate code for copying data of type TYPE at SOURCE plus OFFSET to TARGET
> + plus OFFSET, but do so element-wise and/or field-wise for each record and
> + array within TYPE. TYPE must either be a register type or an aggregate
> + complying with scalarizable_type_p.
> +
> + If CALL_PARAM_P is nonzero, this is a store into a call param on the
> + stack, and block moves may need to be treated specially. */
> +
> +static void
> +emit_move_elementwise (tree type, rtx target, rtx source, HOST_WIDE_INT offset,
> + int call_param_p)
> +{
> + switch (TREE_CODE (type))
> + {
> + case RECORD_TYPE:
> + for (tree fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
> + if (TREE_CODE (fld) == FIELD_DECL)
> + {
> + HOST_WIDE_INT fld_offset = offset + int_bit_position (fld);
> + tree ft = TREE_TYPE (fld);
> + emit_move_elementwise (ft, target, source, fld_offset,
> + call_param_p);
> + }
> + break;
> +
> + case ARRAY_TYPE:
> + {
> + tree elem_type = TREE_TYPE (type);
> + HOST_WIDE_INT el_size = tree_to_shwi (TYPE_SIZE (elem_type));
> + gcc_assert (el_size > 0);
> +
> + offset_int idx, max;
> + /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
> + if (extract_min_max_idx_from_array (type, &idx, &max))
> + {
> + HOST_WIDE_INT el_offset = offset;
> + for (; idx <= max; ++idx)
> + {
> + emit_move_elementwise (elem_type, target, source, el_offset,
> + call_param_p);
> + el_offset += el_size;
> + }
> + }
> + }
> + break;
> + default:
> + machine_mode mode = TYPE_MODE (type);
> +
> + rtx ntgt = adjust_address (target, mode, offset / BITS_PER_UNIT);
> + rtx nsrc = adjust_address (source, mode, offset / BITS_PER_UNIT);
> +
> + /* TODO: Figure out whether the following is actually necessary. */
> + if (target == ntgt)
> + ntgt = copy_rtx (target);
> + if (source == nsrc)
> + nsrc = copy_rtx (source);
> +
> + gcc_assert (mode != VOIDmode);
> + if (mode != BLKmode)
> + emit_move_insn (ntgt, nsrc);
> + else
> + {
> + /* For example vector gimple registers can end up here. */
> + rtx size = expand_expr (TYPE_SIZE_UNIT (type), NULL_RTX,
> + TYPE_MODE (sizetype), EXPAND_NORMAL);
> + emit_block_move (ntgt, nsrc, size,
> + (call_param_p
> + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> + }
> + break;
> + }
> + return;
> +}
> +
> /* Generate code for computing expression EXP,
> and storing the value into TARGET.
>
> @@ -5713,9 +5788,32 @@ store_expr_with_bounds (tree exp, rtx target, int call_param_p,
> emit_group_store (target, temp, TREE_TYPE (exp),
> int_size_in_bytes (TREE_TYPE (exp)));
> else if (GET_MODE (temp) == BLKmode)
> - emit_block_move (target, temp, expr_size (exp),
> - (call_param_p
> - ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> + {
> + /* Copying smallish BLKmode structures with emit_block_move and thus
> + by-pieces can result in store-to-load stalls. So copy some simple
> + small aggregates element or field-wise. */
> + int count = 0;
> + if (GET_MODE (target) == BLKmode
> + && AGGREGATE_TYPE_P (TREE_TYPE (exp))
> + && !TREE_ADDRESSABLE (TREE_TYPE (exp))
> + && tree_fits_shwi_p (TYPE_SIZE (TREE_TYPE (exp)))
> + && (tree_to_shwi (TYPE_SIZE (TREE_TYPE (exp)))
> + <= (PARAM_VALUE (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY)
> + * BITS_PER_UNIT))
> + && simple_mix_of_records_and_arrays_p (TREE_TYPE (exp), false,
> + &count)
> + && (count <= PARAM_VALUE (PARAM_MAX_INSNS_FOR_ELEMENTWISE_COPY)))
> + {
> + /* FIXME: Can this happen? What would it mean? */
> + gcc_assert (!reverse);
> + emit_move_elementwise (TREE_TYPE (exp), target, temp, 0,
> + call_param_p);
> + }
> + else
> + emit_block_move (target, temp, expr_size (exp),
> + (call_param_p
> + ? BLOCK_OP_CALL_PARM : BLOCK_OP_NORMAL));
> + }
> /* If we emit a nontemporal store, there is nothing else to do. */
> else if (nontemporal && emit_storent_insn (target, temp))
> ;
> diff --git a/gcc/ipa-cp.c b/gcc/ipa-cp.c
> index d23c1d8ba3e..30f91e70c22 100644
> --- a/gcc/ipa-cp.c
> +++ b/gcc/ipa-cp.c
> @@ -124,6 +124,7 @@ along with GCC; see the file COPYING3. If not see
> #include "tree-ssa-ccp.h"
> #include "stringpool.h"
> #include "attribs.h"
> +#include "tree-sra.h"
>
> template <typename valtype> class ipcp_value;
>
> diff --git a/gcc/ipa-prop.h b/gcc/ipa-prop.h
> index fa5bed49ee0..2313cc884ed 100644
> --- a/gcc/ipa-prop.h
> +++ b/gcc/ipa-prop.h
> @@ -877,10 +877,6 @@ ipa_parm_adjustment *ipa_get_adjustment_candidate (tree **, bool *,
> void ipa_release_body_info (struct ipa_func_body_info *);
> tree ipa_get_callee_param_type (struct cgraph_edge *e, int i);
>
> -/* From tree-sra.c: */
> -tree build_ref_for_offset (location_t, tree, HOST_WIDE_INT, bool, tree,
> - gimple_stmt_iterator *, bool);
> -
> /* In ipa-cp.c */
> void ipa_cp_c_finalize (void);
>
> diff --git a/gcc/params.def b/gcc/params.def
> index 8881f4c403a..9c778f9540a 100644
> --- a/gcc/params.def
> +++ b/gcc/params.def
> @@ -1287,6 +1287,18 @@ DEFPARAM (PARAM_VECT_EPILOGUES_NOMASK,
> "Enable loop epilogue vectorization using smaller vector size.",
> 0, 0, 1)
>
> +DEFPARAM (PARAM_MAX_SIZE_FOR_ELEMENTWISE_COPY,
> + "max-size-for-elementwise-copy",
> + "Maximum size in bytes of a structure or an array to by considered "
> + "for copying by its individual fields or elements",
> + 0, 0, 512)
> +
> +DEFPARAM (PARAM_MAX_INSNS_FOR_ELEMENTWISE_COPY,
> + "max-insns-for-elementwise-copy",
> + "Maximum number of instructions needed to consider copying "
> + "a structure or an array by its individual fields or elements",
> + 6, 0, 64)
> +
> /*
>
> Local variables:
> diff --git a/gcc/testsuite/gcc.target/i386/pr80689-1.c b/gcc/testsuite/gcc.target/i386/pr80689-1.c
> new file mode 100644
> index 00000000000..4156d4fba45
> --- /dev/null
> +++ b/gcc/testsuite/gcc.target/i386/pr80689-1.c
> @@ -0,0 +1,38 @@
> +/* { dg-do compile } */
> +/* { dg-options "-O2" } */
> +
> +typedef struct st1
> +{
> + long unsigned int a,b;
> + long int c,d;
> +}R;
> +
> +typedef struct st2
> +{
> + int t;
> + R reg;
> +}N;
> +
> +void Set (const R *region, N *n_info );
> +
> +void test(N *n_obj ,const long unsigned int a, const long unsigned int b, const long int c,const long int d)
> +{
> + R reg;
> +
> + reg.a=a;
> + reg.b=b;
> + reg.c=c;
> + reg.d=d;
> + Set (®, n_obj);
> +
> +}
> +
> +void Set (const R *reg, N *n_obj )
> +{
> + n_obj->reg=(*reg);
> +}
> +
> +
> +/* { dg-final { scan-assembler-not "%(x|y|z)mm\[0-9\]+" } } */
> +/* { dg-final { scan-assembler-not "movdqu" } } */
> +/* { dg-final { scan-assembler-not "movups" } } */
> diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
> index bac593951e7..d06463ce21c 100644
> --- a/gcc/tree-sra.c
> +++ b/gcc/tree-sra.c
> @@ -104,6 +104,7 @@ along with GCC; see the file COPYING3. If not see
> #include "ipa-fnsummary.h"
> #include "ipa-utils.h"
> #include "builtins.h"
> +#include "tree-sra.h"
>
> /* Enumeration of all aggregate reductions we can do. */
> enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
> @@ -952,14 +953,15 @@ create_access (tree expr, gimple *stmt, bool write)
> }
>
>
> -/* Return true iff TYPE is scalarizable - i.e. a RECORD_TYPE or fixed-length
> - ARRAY_TYPE with fields that are either of gimple register types (excluding
> - bit-fields) or (recursively) scalarizable types. CONST_DECL must be true if
> - we are considering a decl from constant pool. If it is false, char arrays
> - will be refused. */
> +/* Return true if TYPE consists of RECORD_TYPE or fixed-length ARRAY_TYPE with
> + fields/elements that are not bit-fields and are either register types or
> + recursively comply with simple_mix_of_records_and_arrays_p. Furthermore, if
> + ALLOW_CHAR_ARRAYS is false, the function will return false also if TYPE
> + contains an array of elements that only have one byte. */
>
> -static bool
> -scalarizable_type_p (tree type, bool const_decl)
> +bool
> +simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays,
> + int *count_p)
> {
> gcc_assert (!is_gimple_reg_type (type));
> if (type_contains_placeholder_p (type))
> @@ -976,8 +978,13 @@ scalarizable_type_p (tree type, bool const_decl)
> if (DECL_BIT_FIELD (fld))
> return false;
>
> - if (!is_gimple_reg_type (ft)
> - && !scalarizable_type_p (ft, const_decl))
> + if (is_gimple_reg_type (ft))
> + {
> + if (count_p)
> + (*count_p)++;
> + }
> + else if (!simple_mix_of_records_and_arrays_p (ft, allow_char_arrays,
> + count_p))
> return false;
> }
>
> @@ -986,7 +993,7 @@ scalarizable_type_p (tree type, bool const_decl)
> case ARRAY_TYPE:
> {
> HOST_WIDE_INT min_elem_size;
> - if (const_decl)
> + if (allow_char_arrays)
> min_elem_size = 0;
> else
> min_elem_size = BITS_PER_UNIT;
> @@ -1007,9 +1014,45 @@ scalarizable_type_p (tree type, bool const_decl)
> return false;
>
> tree elem = TREE_TYPE (type);
> - if (!is_gimple_reg_type (elem)
> - && !scalarizable_type_p (elem, const_decl))
> - return false;
> + if (!count_p)
> + {
> + if (!is_gimple_reg_type (elem)
> + && !simple_mix_of_records_and_arrays_p (elem, allow_char_arrays,
> + NULL))
> + return false;
> + else
> + return true;
> + }
> +
> + offset_int min, max;
> + HOST_WIDE_INT ds;
> + bool nonzero = extract_min_max_idx_from_array (type, &min, &max);
> +
> + if (nonzero && (min <= max))
> + {
> + offset_int d = max - min + 1;
> + if (!wi::fits_shwi_p (d))
> + return false;
> + ds = d.to_shwi ();
> + if (ds > INT_MAX)
> + return false;
> + }
> + else
> + ds = 0;
> +
> + if (is_gimple_reg_type (elem))
> + *count_p += (int) ds;
> + else
> + {
> + int elc = 0;
> + if (!simple_mix_of_records_and_arrays_p (elem, allow_char_arrays,
> + &elc))
> + return false;
> + ds *= elc;
> + if (ds > INT_MAX)
> + return false;
> + *count_p += (unsigned) ds;
> + }
> return true;
> }
> default:
> @@ -1017,10 +1060,38 @@ scalarizable_type_p (tree type, bool const_decl)
> }
> }
>
> -static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree, tree);
> +static void scalarize_elem (tree, HOST_WIDE_INT, HOST_WIDE_INT, bool, tree,
> + tree);
> +
> +/* For a given array TYPE, return false if its domain does not have any maximum
> + value. Otherwise calculate MIN and MAX indices of the first and the last
> + element. */
> +
> +bool
> +extract_min_max_idx_from_array (tree type, offset_int *min, offset_int *max)
> +{
> + tree domain = TYPE_DOMAIN (type);
> + tree minidx = TYPE_MIN_VALUE (domain);
> + gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
> + tree maxidx = TYPE_MAX_VALUE (domain);
> + if (!maxidx)
> + return false;
> + gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
> +
> + /* MINIDX and MAXIDX are inclusive, and must be interpreted in
> + DOMAIN (e.g. signed int, whereas min/max may be size_int). */
> + *min = wi::to_offset (minidx);
> + *max = wi::to_offset (maxidx);
> + if (!TYPE_UNSIGNED (domain))
> + {
> + *min = wi::sext (*min, TYPE_PRECISION (domain));
> + *max = wi::sext (*max, TYPE_PRECISION (domain));
> + }
> + return true;
> +}
>
> /* Create total_scalarization accesses for all scalar fields of a member
> - of type DECL_TYPE conforming to scalarizable_type_p. BASE
> + of type DECL_TYPE conforming to simple_mix_of_records_and_arrays_p. BASE
> must be the top-most VAR_DECL representing the variable; within that,
> OFFSET locates the member and REF must be the memory reference expression for
> the member. */
> @@ -1047,27 +1118,14 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
> {
> tree elemtype = TREE_TYPE (decl_type);
> tree elem_size = TYPE_SIZE (elemtype);
> - gcc_assert (elem_size && tree_fits_shwi_p (elem_size));
> HOST_WIDE_INT el_size = tree_to_shwi (elem_size);
> gcc_assert (el_size > 0);
>
> - tree minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (decl_type));
> - gcc_assert (TREE_CODE (minidx) == INTEGER_CST);
> - tree maxidx = TYPE_MAX_VALUE (TYPE_DOMAIN (decl_type));
> + offset_int idx, max;
> /* Skip (some) zero-length arrays; others have MAXIDX == MINIDX - 1. */
> - if (maxidx)
> + if (extract_min_max_idx_from_array (decl_type, &idx, &max))
> {
> - gcc_assert (TREE_CODE (maxidx) == INTEGER_CST);
> tree domain = TYPE_DOMAIN (decl_type);
> - /* MINIDX and MAXIDX are inclusive, and must be interpreted in
> - DOMAIN (e.g. signed int, whereas min/max may be size_int). */
> - offset_int idx = wi::to_offset (minidx);
> - offset_int max = wi::to_offset (maxidx);
> - if (!TYPE_UNSIGNED (domain))
> - {
> - idx = wi::sext (idx, TYPE_PRECISION (domain));
> - max = wi::sext (max, TYPE_PRECISION (domain));
> - }
> for (int el_off = offset; idx <= max; ++idx)
> {
> tree nref = build4 (ARRAY_REF, elemtype,
> @@ -1088,10 +1146,10 @@ completely_scalarize (tree base, tree decl_type, HOST_WIDE_INT offset, tree ref)
> }
>
> /* Create total_scalarization accesses for a member of type TYPE, which must
> - satisfy either is_gimple_reg_type or scalarizable_type_p. BASE must be the
> - top-most VAR_DECL representing the variable; within that, POS and SIZE locate
> - the member, REVERSE gives its torage order. and REF must be the reference
> - expression for it. */
> + satisfy either is_gimple_reg_type or simple_mix_of_records_and_arrays_p.
> + BASE must be the top-most VAR_DECL representing the variable; within that,
> + POS and SIZE locate the member, REVERSE gives its torage order. and REF must
> + be the reference expression for it. */
>
> static void
> scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
> @@ -1111,7 +1169,8 @@ scalarize_elem (tree base, HOST_WIDE_INT pos, HOST_WIDE_INT size, bool reverse,
> }
>
> /* Create a total_scalarization access for VAR as a whole. VAR must be of a
> - RECORD_TYPE or ARRAY_TYPE conforming to scalarizable_type_p. */
> + RECORD_TYPE or ARRAY_TYPE conforming to
> + simple_mix_of_records_and_arrays_p. */
>
> static void
> create_total_scalarization_access (tree var)
> @@ -2803,8 +2862,9 @@ analyze_all_variable_accesses (void)
> {
> tree var = candidate (i);
>
> - if (VAR_P (var) && scalarizable_type_p (TREE_TYPE (var),
> - constant_decl_p (var)))
> + if (VAR_P (var)
> + && simple_mix_of_records_and_arrays_p (TREE_TYPE (var),
> + constant_decl_p (var), NULL))
> {
> if (tree_to_uhwi (TYPE_SIZE (TREE_TYPE (var)))
> <= max_scalarization_size)
> diff --git a/gcc/tree-sra.h b/gcc/tree-sra.h
> new file mode 100644
> index 00000000000..2857688b21e
> --- /dev/null
> +++ b/gcc/tree-sra.h
> @@ -0,0 +1,34 @@
> +/* tree-sra.h - Run-time parameters.
> + Copyright (C) 2017 Free Software Foundation, Inc.
> +
> +This file is part of GCC.
> +
> +GCC is free software; you can redistribute it and/or modify it under
> +the terms of the GNU General Public License as published by the Free
> +Software Foundation; either version 3, or (at your option) any later
> +version.
> +
> +GCC is distributed in the hope that it will be useful, but WITHOUT ANY
> +WARRANTY; without even the implied warranty of MERCHANTABILITY or
> +FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
> +for more details.
> +
> +You should have received a copy of the GNU General Public License
> +along with GCC; see the file COPYING3. If not see
> +<http://www.gnu.org/licenses/>. */
> +
> +#ifndef TREE_SRA_H
> +#define TREE_SRA_H
> +
> +
> +bool simple_mix_of_records_and_arrays_p (tree type, bool allow_char_arrays,
> + int *count_pg);
> +bool extract_min_max_idx_from_array (tree type, offset_int *idx,
> + offset_int *max);
> +tree build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
> + bool reverse, tree exp_type,
> + gimple_stmt_iterator *gsi, bool insert_after);
> +
> +
> +
> +#endif /* TREE_SRA_H */
> --
> 2.14.2
>
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-13 13:17 ` Richard Biener
@ 2017-11-13 13:54 ` Michael Matz
2017-11-13 14:19 ` Richard Biener
2017-11-23 16:02 ` Martin Jambor
1 sibling, 1 reply; 26+ messages in thread
From: Michael Matz @ 2017-11-13 13:54 UTC (permalink / raw)
To: Richard Biener; +Cc: GCC Patches, Jan Hubicka
Hi,
On Mon, 13 Nov 2017, Richard Biener wrote:
> The main concern here is that GIMPLE is not very well defined for
> aggregate copies and that gimple-fold.c happily optimizes
> memcpy (&a, &b, sizeof (a)) into a = b;
What you missed to mention is that we then discussed about rectifying this
situation by defining GIMPLE more precisely :) Effectively an aggregate
assignment in GIMPLE (right now) is always defined to be a dumb block
copy. We need a way to describe a member-wise copy as well. That can
either be a flag on the statement or implicit by the alias type (i.e.
block copies always need the ref-all alias type, all others would be
member-wise copies).
Then a user-written memcpy can only be rewritten into a member-wise
assignment when the types contain no padding, and SRA can only look
through member-wise assignments when it doesn't see all accesses to the
destination (when it does then it can look also through block copies).
Then this patch can be restricted to the member-wise assignments (which
still helps imagemagick as the struct in question doesn't contain padding
IIRC).
That, or we must dumb down SRA quite much, which I don't think would be a
good idea.
(I'm not sure if your example would be really valid C as it changes the
dynamic type of a statically typed declaration; but OTOH we shouldn't
care, as in GIMPLE the example should of course be expressible)
Ciao,
Michael.
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-13 13:54 ` Michael Matz
@ 2017-11-13 14:19 ` Richard Biener
2017-11-13 15:33 ` Michael Matz
2017-11-14 0:32 ` Eric Botcazou
0 siblings, 2 replies; 26+ messages in thread
From: Richard Biener @ 2017-11-13 14:19 UTC (permalink / raw)
To: Michael Matz; +Cc: GCC Patches, Jan Hubicka
On Mon, Nov 13, 2017 at 2:46 PM, Michael Matz <matz@suse.de> wrote:
> Hi,
>
> On Mon, 13 Nov 2017, Richard Biener wrote:
>
>> The main concern here is that GIMPLE is not very well defined for
>> aggregate copies and that gimple-fold.c happily optimizes
>> memcpy (&a, &b, sizeof (a)) into a = b;
>
> What you missed to mention is that we then discussed about rectifying this
> situation by defining GIMPLE more precisely :) Effectively an aggregate
> assignment in GIMPLE (right now) is always defined to be a dumb block
> copy. We need a way to describe a member-wise copy as well. That can
> either be a flag on the statement or implicit by the alias type (i.e.
> block copies always need the ref-all alias type, all others would be
> member-wise copies).
Yes. Note that it's already GENERIC that needs to nail down the difference.
For the folding there's the possibility of using a char[n] type with n being
constant, thus a new array type for each size, or a char[] type with
variable size, using a WITH_SIZE_EXPR on the RHS (but support for
WITH_SIZE_EXPR isn't so good in passes so I'd rather avoid this for
constant sizes).
The chance here is, of course (find the PR, it exists...), that SRA then
decomposes the char[] copy bytewise...
That said, memcpy folding is easy to fix. The question is of course
what the semantic of VIEW_CONVERTs is (SRA _does_ contain
bail-outs on those). Like if you have
struct A { short s; int i; } x;
struct B { int i; short s; } y;
void foo ()
{
x = VIEW_CONVERT <struct A> (y);
}
so can you access padding via view-converting its value? Ada uses
VIEW_CONVERT punning on structures a _lot_ (probably the reason
for the SRA bailout).
The above assignment would still be going through that aggregate copy
expansion path.
> Then a user-written memcpy can only be rewritten into a member-wise
> assignment when the types contain no padding, and SRA can only look
> through member-wise assignments when it doesn't see all accesses to the
> destination (when it does then it can look also through block copies).
>
> Then this patch can be restricted to the member-wise assignments (which
> still helps imagemagick as the struct in question doesn't contain padding
> IIRC).
>
> That, or we must dumb down SRA quite much, which I don't think would be a
> good idea.
>
> (I'm not sure if your example would be really valid C as it changes the
> dynamic type of a statically typed declaration; but OTOH we shouldn't
> care, as in GIMPLE the example should of course be expressible)
Yeah, we can equally use allocated storage (but our memcpy folding
then won't apply ...).
Richard.
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-13 14:19 ` Richard Biener
@ 2017-11-13 15:33 ` Michael Matz
2017-11-13 16:30 ` Richard Biener
2017-11-14 0:32 ` Eric Botcazou
1 sibling, 1 reply; 26+ messages in thread
From: Michael Matz @ 2017-11-13 15:33 UTC (permalink / raw)
To: Richard Biener; +Cc: GCC Patches, Jan Hubicka
Hi,
On Mon, 13 Nov 2017, Richard Biener wrote:
> The chance here is, of course (find the PR, it exists...), that SRA then
> decomposes the char[] copy bytewise...
>
> That said, memcpy folding is easy to fix. The question is of course
> what the semantic of VIEW_CONVERTs is (SRA _does_ contain
> bail-outs on those). Like if you have
>
> struct A { short s; int i; } x;
> struct B { int i; short s; } y;
>
> void foo ()
> {
> x = VIEW_CONVERT <struct A> (y);
> }
>
> so can you access padding via view-converting its value? Ada uses
> VIEW_CONVERT punning on structures a _lot_ (probably the reason for the
> SRA bailout).
I would say a VIEW_CONVERT shouldn't be allowed to inspect padding on the
RHS (and expected to clobber padding on the LHS). That is, if you want to
really really access padding on some struct type you can only use memcpy.
(Or view-convert it to some char[N] array, perhaps there it makes sense to
copy padding, i.e. regard that as a block copy).
The above example shows why I'm of this opinion. Both structs have
padding at different place, and it overlaps a member in the other
struct. I don't see how to give that any sane meaning (beyond always
handling it as block copy, and which point we can as well give up and get
rid of VIEW_CONVERT_EXPR in favor of explicit memcpy).
Ciao,
Michael.
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-13 15:33 ` Michael Matz
@ 2017-11-13 16:30 ` Richard Biener
0 siblings, 0 replies; 26+ messages in thread
From: Richard Biener @ 2017-11-13 16:30 UTC (permalink / raw)
To: Michael Matz; +Cc: GCC Patches, Jan Hubicka, ebotcazou
On November 13, 2017 3:20:16 PM GMT+01:00, Michael Matz <matz@suse.de> wrote:
>Hi,
>
>On Mon, 13 Nov 2017, Richard Biener wrote:
>
>> The chance here is, of course (find the PR, it exists...), that SRA
>then
>> decomposes the char[] copy bytewise...
>>
>> That said, memcpy folding is easy to fix. The question is of course
>> what the semantic of VIEW_CONVERTs is (SRA _does_ contain
>> bail-outs on those). Like if you have
>>
>> struct A { short s; int i; } x;
>> struct B { int i; short s; } y;
>>
>> void foo ()
>> {
>> x = VIEW_CONVERT <struct A> (y);
>> }
>>
>> so can you access padding via view-converting its value? Ada uses
>> VIEW_CONVERT punning on structures a _lot_ (probably the reason for
>the
>> SRA bailout).
>
>I would say a VIEW_CONVERT shouldn't be allowed to inspect padding on
>the
>RHS (and expected to clobber padding on the LHS). That is, if you want
>to
>really really access padding on some struct type you can only use
>memcpy.
>(Or view-convert it to some char[N] array, perhaps there it makes sense
>to
>copy padding, i.e. regard that as a block copy).
>
>The above example shows why I'm of this opinion. Both structs have
>padding at different place, and it overlaps a member in the other
>struct. I don't see how to give that any sane meaning (beyond always
>handling it as block copy, and which point we can as well give up and
>get
>rid of VIEW_CONVERT_EXPR in favor of explicit memcpy).
Eric should know constraints important for Ada.
Richard.
>
>Ciao,
>Michael.
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-13 14:19 ` Richard Biener
2017-11-13 15:33 ` Michael Matz
@ 2017-11-14 0:32 ` Eric Botcazou
1 sibling, 0 replies; 26+ messages in thread
From: Eric Botcazou @ 2017-11-14 0:32 UTC (permalink / raw)
To: Richard Biener; +Cc: gcc-patches, Michael Matz, Jan Hubicka
> The chance here is, of course (find the PR, it exists...), that SRA then
> decomposes the char[] copy bytewise...
>
> That said, memcpy folding is easy to fix. The question is of course
> what the semantic of VIEW_CONVERTs is (SRA _does_ contain
> bail-outs on those). Like if you have
>
> struct A { short s; int i; } x;
> struct B { int i; short s; } y;
>
> void foo ()
> {
> x = VIEW_CONVERT <struct A> (y);
> }
>
> so can you access padding via view-converting its value? Ada uses
> VIEW_CONVERT punning on structures a _lot_ (probably the reason
> for the SRA bailout).
Couple of things:
1. We have been trying to get rid of VIEW_CONVERT as much as possible in the
Ada compiler for a number of years now.
2. Padding is garbage in Ada and thus its contents cannot have any effect on
the execution of legal programs (and SRA already killed any hope of preserving
padding a long time ago for not-so-legal programs anyway).
--
Eric Botcazou
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-03 16:38 ` Martin Jambor
2017-11-13 13:17 ` Richard Biener
@ 2017-11-14 10:25 ` Martin Jambor
1 sibling, 0 replies; 26+ messages in thread
From: Martin Jambor @ 2017-11-14 10:25 UTC (permalink / raw)
To: Richard Biener, GCC Patches, Jan Hubicka, Michael Matz
Hi,
I thought I sent the following email last Friday but found it in my
drafts folder right now, so let me send it now so that anybody
interested can see what the patch does on Haswell.
I have only skimmed through new messages in the thread. I am now
looking into something else right now but will get back to this matter
next week at the latest.
On Fri, Nov 03, 2017 at 05:38:30PM +0100, Martin Jambor wrote:
>
...
>
> Anyway, here are the numbers. They were taken on two different
> Zen-based machines. I am also in the process of measuring at least
> something on a Haswell machine but I started later and the machine is
> quite a bit slower so I will not have the numbers until next week (and
> not all equivalents in any way). I found out I do not have access to
> any more modern .*Lake intel CPU.
>
OK, I have the numbers now too. So far I do not know why, in addition
to 416.gamess, also 465.tonto failed to compile, I will investigate
why only later.
Because the machine is quite a bit slower and everything took forever,
I have measured only unpatched trunk three times and then re-run only
those benchmarks which were more than 2% off when compiled with the
patched compiler.
Haswell SPECINT 2006 -O2 generic tuning
=======================================
Run-time
--------
| Benchmark | trunk | x1 | % |
|----------------+-------+------+-------|
| 400.perlbench | 775 | 777 | +0.26 |
| 401.bzip2 | 1200 | 1200 | +0.00 |
| 403.gcc | 655 | 656 | +0.15 |
| 429.mcf | 547 | 517 | -5.48 |
| 445.gobmk | 1140 | 1140 | +0.00 |
| 456.hmmer | 1130 | 1130 | +0.00 |
| 458.sjeng | 1310 | 1300 | -0.76 |
| 462.libquantum | 758 | 751 | -0.92 |
| 464.h264ref | 1370 | 1390 | +1.46 |
| 471.omnetpp | 475 | 471 | -0.84 |
| 473.astar | 870 | 867 | -0.34 |
| 483.xalancbmk | 488 | 486 | -0.41 |
Text size
---------
| Benchmark | trunk | x1 | % |
|----------------+---------+---------+-------|
| 400.perlbench | 875874 | 875954 | +0.01 |
| 401.bzip2 | 44754 | 44754 | +0.00 |
| 403.gcc | 2294466 | 2296098 | +0.07 |
| 429.mcf | 8226 | 8226 | +0.00 |
| 445.gobmk | 579778 | 579826 | +0.01 |
| 456.hmmer | 221058 | 221058 | +0.00 |
| 458.sjeng | 93362 | 94882 | +1.63 |
| 462.libquantum | 28314 | 28362 | +0.17 |
| 464.h264ref | 393874 | 393922 | +0.01 |
| 471.omnetpp | 430306 | 430418 | +0.03 |
| 473.astar | 29362 | 29538 | +0.60 |
| 483.xalancbmk | 2361298 | 2361506 | +0.01 |
Haswell SPECINT 2006 -Ofast native tuning
=========================================
Run-time
--------
| Benchmark | trunk | x1 | % |
|----------------+-------+------+-------|
| 400.perlbench | 802 | 803 | +0.12 |
| 401.bzip2 | 1180 | 1170 | -0.85 |
| 403.gcc | 646 | 647 | +0.15 |
| 429.mcf | 543 | 508 | -6.45 |
| 445.gobmk | 1130 | 1130 | +0.00 |
| 456.hmmer | 529 | 532 | +0.57 |
| 458.sjeng | 1260 | 1260 | +0.00 |
| 462.libquantum | 764 | 761 | -0.39 |
| 464.h264ref | 1280 | 1290 | +0.78 |
| 471.omnetpp | 476 | 464 | -2.52 |
| 473.astar | 844 | 843 | -0.12 |
| 483.xalancbmk | 480 | 476 | -0.83 |
Text size
---------
| Benchmark | trunk | x1 | % |
|----------------+---------+---------+-------|
| 400.perlbench | 1130994 | 1131058 | +0.01 |
| 401.bzip2 | 77346 | 77346 | +0.00 |
| 403.gcc | 3099938 | 3101826 | +0.06 |
| 429.mcf | 10162 | 10162 | +0.00 |
| 445.gobmk | 766706 | 766786 | +0.01 |
| 456.hmmer | 346610 | 346610 | +0.00 |
| 458.sjeng | 143650 | 145522 | +1.30 |
| 462.libquantum | 30986 | 31066 | +0.26 |
| 464.h264ref | 725218 | 725266 | +0.01 |
| 471.omnetpp | 546386 | 546642 | +0.05 |
| 473.astar | 38690 | 38914 | +0.58 |
| 483.xalancbmk | 3313746 | 3313922 | +0.01 |
Haswell SPECFP 2006 -O2 generic tuning
======================================
Run-time
--------
| Benchmark | trunk | x1 | % |
|---------------+-------+------+-------|
| 410.bwaves | 833 | 831 | -0.24 |
| 416.gamess | NR | NR | |
| 433.milc | 820 | 814 | -0.73 |
| 434.zeusmp | 950 | 949 | -0.11 |
| 435.gromacs | 945 | 946 | +0.11 |
| 436.cactusADM | 1380 | 1380 | +0.00 |
| 437.leslie3d | 813 | 812 | -0.12 |
| 444.namd | 983 | 983 | +0.00 |
| 447.dealII | 755 | 759 | +0.53 |
| 450.soplex | 467 | 464 | -0.64 |
| 453.povray | 402 | 395 | -1.74 |
| 454.calculix | 1980 | 1980 | +0.00 |
| 459.GemsFDTD | 765 | 753 | -1.57 |
| 465.tonto | NR | NR | |
| 470.lbm | 806 | 806 | +0.00 |
| 481.wrf | 1330 | 1330 | +0.00 |
| 482.sphinx3 | 1380 | 1380 | +0.00 |
Text size
---------
| Benchmark | trunk | x1 | % |
|---------------+---------+---------+-------|
| 410.bwaves | 25954 | 25954 | +0.00 |
| 433.milc | 87922 | 87922 | +0.00 |
| 434.zeusmp | 212034 | 212034 | +0.00 |
| 435.gromacs | 747026 | 747026 | +0.00 |
| 436.cactusADM | 526178 | 526178 | +0.00 |
| 437.leslie3d | 83234 | 83234 | +0.00 |
| 444.namd | 297234 | 297266 | +0.01 |
| 447.dealII | 2165282 | 2172290 | +0.32 |
| 450.soplex | 347122 | 347122 | +0.00 |
| 453.povray | 800914 | 801570 | +0.08 |
| 454.calculix | 1342802 | 1342802 | +0.00 |
| 459.GemsFDTD | 353410 | 354050 | +0.18 |
| 470.lbm | 9202 | 9202 | +0.00 |
| 481.wrf | 3345170 | 3345170 | +0.00 |
| 482.sphinx3 | 125026 | 125026 | +0.00 |
Haswell SPECFP 2006 -Ofast native tuning
========================================
Run-time
--------
| Benchmark | trunk | x1 | % |
|---------------+-------+------+-------|
| 410.bwaves | 551 | 550 | -0.18 |
| 416.gamess | NR | NR | |
| 433.milc | 773 | 776 | +0.39 |
| 434.zeusmp | 660 | 660 | +0.00 |
| 435.gromacs | 876 | 874 | -0.23 |
| 436.cactusADM | 620 | 619 | -0.16 |
| 437.leslie3d | 501 | 501 | +0.00 |
| 444.namd | 974 | 974 | +0.00 |
| 447.dealII | 722 | 720 | -0.28 |
| 450.soplex | 459 | 457 | -0.44 |
| 453.povray | 416 | 410 | -1.44 |
| 454.calculix | 883 | 882 | -0.11 |
| 459.GemsFDTD | 625 | 614 | -1.76 |
| 465.tonto | NR | NR | |
| 470.lbm | 783 | 781 | -0.26 |
| 481.wrf | 748 | 746 | -0.27 |
| 482.sphinx3 | 1020 | 1020 | +0.00 |
Text size
---------
| Benchmark | trunk | x1 | % |
|---------------+---------+---------+-------|
| 410.bwaves | 30802 | 30802 | +0.00 |
| 433.milc | 122450 | 122450 | +0.00 |
| 434.zeusmp | 613458 | 613458 | +0.00 |
| 435.gromacs | 957922 | 957922 | +0.00 |
| 436.cactusADM | 763794 | 763794 | +0.00 |
| 437.leslie3d | 154690 | 154690 | +0.00 |
| 444.namd | 311282 | 311314 | +0.01 |
| 447.dealII | 2486482 | 2493202 | +0.27 |
| 450.soplex | 436322 | 436322 | +0.00 |
| 453.povray | 1088034 | 1088962 | +0.09 |
| 454.calculix | 1701410 | 1701410 | +0.00 |
| 459.GemsFDTD | 560642 | 560658 | +0.00 |
| 470.lbm | 9458 | 9458 | +0.00 |
| 481.wrf | 5413554 | 5413778 | +0.00 |
| 482.sphinx3 | 190034 | 190034 | +0.00 |
Haswell SPEC INTrate 2017 -O2 generic tuning
============================================
Run-time
--------
| Benchmark | trunk | x1 | % |
|-----------------+-------+------+-------|
| 500.perlbench_r | 1201 | 1204 | +0.25 |
| 502.gcc_r | 798 | 793 | -0.63 |
| 505.mcf_r | 1038 | 1049 | +1.06 |
| 520.omnetpp_r | 825 | 824 | -0.12 |
| 523.xalancbmk_r | 985 | 981 | -0.41 |
| 525.x264_r | 1463 | 1463 | +0.00 |
| 531.deepsjeng_r | 954 | 956 | +0.21 |
| 541.leela_r | 1570 | 1569 | -0.06 |
| 548.exchange2_r | 1266 | 1267 | +0.08 |
| 557.xz_r | 1033 | 1029 | -0.39 |
Test size
---------
| Benchmark | trunk | x1 | % |
|-----------------+---------+---------+-------|
| 500.perlbench_r | 1599442 | 1599522 | +0.01 |
| 502.gcc_r | 6757602 | 6759090 | +0.02 |
| 505.mcf_r | 16098 | 16098 | +0.00 |
| 520.omnetpp_r | 1262498 | 1264034 | +0.12 |
| 523.xalancbmk_r | 3989026 | 3989202 | +0.00 |
| 525.x264_r | 414130 | 414194 | +0.02 |
| 531.deepsjeng_r | 67426 | 67458 | +0.05 |
| 541.leela_r | 219378 | 219378 | +0.00 |
| 548.exchange2_r | 61234 | 61234 | +0.00 |
| 557.xz_r | 111490 | 111490 | +0.00 |
Haswell SPEC INTrate 2017 -Ofast native tuning
==============================================
Run-time
--------
| Benchmark | trunk | x1 | % |
|-----------------+-------+------+--------|
| 500.perlbench_r | 1169 | 1170 | +0.09 |
| 502.gcc_r | 786 | 788 | +0.25 |
| 505.mcf_r | 1034 | 1032 | -0.19 |
| 520.omnetpp_r | 804 | 794 | -1.24 |
| 523.xalancbmk_r | 962 | 971 | +0.94 |
| 525.x264_r | 886 | 887 | +0.11 |
| 531.deepsjeng_r | 939 | 944 | +0.53 |
| 541.leela_r | 1462 | 1461 | -0.07 |
| 548.exchange2_r | 1078 | 1082 | +0.37 |
| 557.xz_r | 960 | 950 | -1.04 |
Text size
---------
| Benchmark | trunk | x1 | % |
|-----------------+---------+---------+-------|
| 500.perlbench_r | 2074450 | 2074498 | +0.00 |
| 502.gcc_r | 8434514 | 8437250 | +0.03 |
| 505.mcf_r | 26322 | 26322 | +0.00 |
| 520.omnetpp_r | 1680082 | 1682130 | +0.12 |
| 523.xalancbmk_r | 4853458 | 4853682 | +0.00 |
| 525.x264_r | 594210 | 594210 | +0.00 |
| 531.deepsjeng_r | 88050 | 88082 | +0.04 |
| 541.leela_r | 269298 | 269314 | +0.01 |
| 548.exchange2_r | 114098 | 114098 | +0.00 |
| 557.xz_r | 152354 | 152354 | +0.00 |
Haswell SPEC FP rate 2017 - generic tuning
==========================================
Run-time
--------
| Benchmark | trunk | x1 | % |
|-----------------+-------+------+--------|
| 503.bwaves_r | 2319 | 2343 | +1.03 |
| 507.cactuBSSN_r | 1023 | 975 | -4.69 |
| 508.namd_r | 934 | 935 | +0.11 |
| 510.parest_r | 1391 | 1413 | +1.58 |
| 511.povray_r | 1544 | 1570 | +1.68 |
| 519.lbm_r | 920 | 920 | +0.00 |
| 521.wrf_r | 2955 | 2958 | +0.10 |
| 526.blender_r | 976 | 974 | -0.20 |
| 527.cam4_r | 1580 | 1586 | +0.38 |
| 538.imagick_r | 1758 | 1581 | -10.07 |
| 544.nab_r | 1357 | 1356 | -0.07 |
| 549.fotonik3d_r | 1063 | 1077 | +1.32 |
| 554.roms_r | 1280 | 1283 | +0.23 |
Text size
---------
| Benchmark | trunk | x1 | % |
|-----------------+----------+----------+-------|
| 503.bwaves_r | 32034 | 32034 | +0.00 |
| 507.cactuBSSN_r | 2951634 | 2951634 | +0.00 |
| 508.namd_r | 837458 | 837490 | +0.00 |
| 510.parest_r | 6540866 | 6546754 | +0.09 |
| 511.povray_r | 803618 | 804274 | +0.08 |
| 519.lbm_r | 12018 | 12018 | +0.00 |
| 521.wrf_r | 16292962 | 16296978 | +0.02 |
| 526.blender_r | 7268224 | 7282608 | +0.20 |
| 527.cam4_r | 5063666 | 5065010 | +0.03 |
| 538.imagick_r | 1608178 | 1609282 | +0.07 |
| 544.nab_r | 156242 | 156242 | +0.00 |
| 549.fotonik3d_r | 326738 | 326738 | +0.00 |
| 554.roms_r | 728546 | 728546 | +0.00 |
Haswell SPEC FP rate 2017 - native tuning
=========================================
Run-time
--------
| Benchmark | trunk | x1 | % |
|-----------------+-------+------+-------|
| 503.bwaves_r | 919 | 919 | +0.00 |
| 507.cactuBSSN_r | 864 | 853 | -1.27 |
| 508.namd_r | 924 | 924 | +0.00 |
| 510.parest_r | 1219 | 1220 | +0.08 |
| 511.povray_r | 1597 | 1624 | +1.69 |
| 519.lbm_r | 851 | 851 | +0.00 |
| 521.wrf_r | 1591 | 1594 | +0.19 |
| 526.blender_r | 912 | 920 | +0.88 |
| 527.cam4_r | 1296 | 1309 | +1.00 |
| 538.imagick_r | 1227 | 1207 | -1.63 |
| 544.nab_r | 1278 | 1278 | +0.00 |
| 549.fotonik3d_r | VE | VE | |
| 554.roms_r | 1036 | 1037 | +0.10 |
Text size
---------
| Benchmark | trunk | x1 | % |
|-----------------+----------+----------+-------|
| 503.bwaves_r | 39426 | 39426 | +0.00 |
| 507.cactuBSSN_r | 3991794 | 3991794 | +0.00 |
| 508.namd_r | 956450 | 956466 | +0.00 |
| 510.parest_r | 7341122 | 7345426 | +0.06 |
| 511.povray_r | 1083010 | 1083938 | +0.09 |
| 519.lbm_r | 11826 | 11826 | +0.00 |
| 521.wrf_r | 22028578 | 22032098 | +0.02 |
| 526.blender_r | 9698768 | 9718544 | +0.20 |
| 527.cam4_r | 6738562 | 6740050 | +0.02 |
| 538.imagick_r | 2246674 | 2247122 | +0.02 |
| 544.nab_r | 211378 | 211378 | +0.00 |
| 549.fotonik3d_r | 582626 | 582626 | +0.00 |
| 554.roms_r | 1085234 | 1085234 | +0.00 |
Martin
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-13 13:17 ` Richard Biener
2017-11-13 13:54 ` Michael Matz
@ 2017-11-23 16:02 ` Martin Jambor
2017-11-23 16:29 ` Jakub Jelinek
2017-11-24 10:47 ` Richard Biener
1 sibling, 2 replies; 26+ messages in thread
From: Martin Jambor @ 2017-11-23 16:02 UTC (permalink / raw)
To: Richard Biener, Richard Biener, GCC Patches, Jan Hubicka, Michael Matz
Hi,
On Mon, Nov 13 2017, Richard Biener wrote:
> The main concern here is that GIMPLE is not very well defined for
> aggregate copies and that gimple-fold.c happily optimizes
> memcpy (&a, &b, sizeof (a)) into a = b;
>
> struct A { short s; long i; long j; };
> struct A a, b;
> void foo ()
> {
> __builtin_memcpy (&a, &b, sizeof (struct A));
> }
>
> gets folded to
>
> MEM[(char * {ref-all})&a] = MEM[(char * {ref-all})&b];
> return;
>
> you see we're careful about TBAA but (don't see that above but
> can be verified by for example debugging expand_assignment)
> TREE_TYPE (MEM[...]) is actually 'struct A'.
>
> And yes, I've been worried about SRA as well here... it _does_
> have some early outs when seeing VIEW_CONVERT_EXPR but
> appearantly not for the above. Testcase that aborts with SRA but
> not without:
>
> struct A { short s; long i; long j; };
> struct A a, b;
> void foo ()
> {
> struct A c;
> __builtin_memcpy (&c, &b, sizeof (struct A));
> __builtin_memcpy (&a, &c, sizeof (struct A));
> }
> int main()
> {
> __builtin_memset (&b, 0, sizeof (struct A));
> b.s = 1;
> __builtin_memcpy ((char *)&b+2, &b, 2);
> foo ();
> __builtin_memcpy (&a, (char *)&a+2, 2);
> if (a.s != 1)
> __builtin_abort ();
> return 0;
> }
Thanks for the testcase, I agree that is a fairly big problem. Do you
think that the following (untested) patch is an appropriate way of
fixing it and generally of extending gimple to capture that a statement
is a bit-copy?
If so, I'll add the testcase, bootstrap it and formally propose it.
Subsequently I will of course make sure that any element-wise copying
patch would test the predicate.
Thanks,
Martin
2017-11-23 Martin Jambor <mjambor@suse.cz>
* gimple.c (gimple_bit_copy_p): New function.
* gimple.h (gimple_bit_copy_p): Declare it.
* tree-sra.c (sra_modify_assign): Use it.
---
gcc/gimple.c | 20 ++++++++++++++++++++
gcc/gimple.h | 1 +
gcc/tree-sra.c | 1 +
3 files changed, 22 insertions(+)
diff --git a/gcc/gimple.c b/gcc/gimple.c
index c986a732004..e1b428d91bb 100644
--- a/gcc/gimple.c
+++ b/gcc/gimple.c
@@ -3087,6 +3087,26 @@ gimple_inexpensive_call_p (gcall *stmt)
return false;
}
+/* Return true if STMT is an assignment performing bit copy and so is also
+ expected to copy any padding. */
+
+bool
+gimple_bit_copy_p (gassign *stmt)
+{
+ if (!gimple_assign_single_p (stmt))
+ return false;
+
+ tree lhs = gimple_assign_lhs (stmt);
+ if (TREE_CODE (lhs) == MEM_REF
+ && TYPE_REF_CAN_ALIAS_ALL (reference_alias_ptr_type (lhs)))
+ return true;
+ tree rhs = gimple_assign_rhs1 (stmt);
+ if (TREE_CODE (rhs) == MEM_REF
+ && TYPE_REF_CAN_ALIAS_ALL (reference_alias_ptr_type (rhs)))
+ return true;
+ return false;
+}
+
#if CHECKING_P
namespace selftest {
diff --git a/gcc/gimple.h b/gcc/gimple.h
index 334def89398..60929473361 100644
--- a/gcc/gimple.h
+++ b/gcc/gimple.h
@@ -1531,6 +1531,7 @@ extern void gimple_seq_discard (gimple_seq);
extern void maybe_remove_unused_call_args (struct function *, gimple *);
extern bool gimple_inexpensive_call_p (gcall *);
extern bool stmt_can_terminate_bb_p (gimple *);
+extern bool gimple_bit_copy_p (gassign *);
/* Formal (expression) temporary table handling: multiple occurrences of
the same scalar expression are evaluated into the same temporary. */
diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
index db490b20c3e..fc0a8fe60bf 100644
--- a/gcc/tree-sra.c
+++ b/gcc/tree-sra.c
@@ -3591,6 +3591,7 @@ sra_modify_assign (gimple *stmt, gimple_stmt_iterator *gsi)
|| gimple_has_volatile_ops (stmt)
|| contains_vce_or_bfcref_p (rhs)
|| contains_vce_or_bfcref_p (lhs)
+ || gimple_bit_copy_p (as_a <gassign *> (stmt))
|| stmt_ends_bb_p (stmt))
{
/* No need to copy into a constant-pool, it comes pre-initialized. */
--
2.15.0
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-23 16:02 ` Martin Jambor
@ 2017-11-23 16:29 ` Jakub Jelinek
2017-11-24 10:47 ` Richard Biener
1 sibling, 0 replies; 26+ messages in thread
From: Jakub Jelinek @ 2017-11-23 16:29 UTC (permalink / raw)
To: Martin Jambor; +Cc: Richard Biener, GCC Patches, Jan Hubicka, Michael Matz
On Thu, Nov 23, 2017 at 04:32:43PM +0100, Martin Jambor wrote:
> > struct A { short s; long i; long j; };
> > struct A a, b;
> > void foo ()
> > {
> > struct A c;
> > __builtin_memcpy (&c, &b, sizeof (struct A));
> > __builtin_memcpy (&a, &c, sizeof (struct A));
> > }
> > int main()
> > {
> > __builtin_memset (&b, 0, sizeof (struct A));
> > b.s = 1;
> > __builtin_memcpy ((char *)&b+2, &b, 2);
> > foo ();
> > __builtin_memcpy (&a, (char *)&a+2, 2);
> > if (a.s != 1)
> > __builtin_abort ();
> > return 0;
> > }
Note the testcase would need to be guarded with sizeof (short) == 2
and offsetof (struct A, i) >= 4.
> Thanks for the testcase, I agree that is a fairly big problem. Do you
> think that the following (untested) patch is an appropriate way of
> fixing it and generally of extending gimple to capture that a statement
> is a bit-copy?
Can you bail out just if the type contains any padding? If there is no
padding, then perhaps SRA still might do its stuff (though, e.g. if it
contains bitfields, we'd need to hope store-merging merges it all back
again).
Jakub
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-23 16:02 ` Martin Jambor
2017-11-23 16:29 ` Jakub Jelinek
@ 2017-11-24 10:47 ` Richard Biener
2017-11-24 11:22 ` Richard Biener
1 sibling, 1 reply; 26+ messages in thread
From: Richard Biener @ 2017-11-24 10:47 UTC (permalink / raw)
To: Martin Jambor; +Cc: GCC Patches, Jan Hubicka, Michael Matz
On Thu, Nov 23, 2017 at 4:32 PM, Martin Jambor <mjambor@suse.cz> wrote:
> Hi,
>
> On Mon, Nov 13 2017, Richard Biener wrote:
>> The main concern here is that GIMPLE is not very well defined for
>> aggregate copies and that gimple-fold.c happily optimizes
>> memcpy (&a, &b, sizeof (a)) into a = b;
>>
>> struct A { short s; long i; long j; };
>> struct A a, b;
>> void foo ()
>> {
>> __builtin_memcpy (&a, &b, sizeof (struct A));
>> }
>>
>> gets folded to
>>
>> MEM[(char * {ref-all})&a] = MEM[(char * {ref-all})&b];
>> return;
>>
>> you see we're careful about TBAA but (don't see that above but
>> can be verified by for example debugging expand_assignment)
>> TREE_TYPE (MEM[...]) is actually 'struct A'.
>>
>> And yes, I've been worried about SRA as well here... it _does_
>> have some early outs when seeing VIEW_CONVERT_EXPR but
>> appearantly not for the above. Testcase that aborts with SRA but
>> not without:
>>
>> struct A { short s; long i; long j; };
>> struct A a, b;
>> void foo ()
>> {
>> struct A c;
>> __builtin_memcpy (&c, &b, sizeof (struct A));
>> __builtin_memcpy (&a, &c, sizeof (struct A));
>> }
>> int main()
>> {
>> __builtin_memset (&b, 0, sizeof (struct A));
>> b.s = 1;
>> __builtin_memcpy ((char *)&b+2, &b, 2);
>> foo ();
>> __builtin_memcpy (&a, (char *)&a+2, 2);
>> if (a.s != 1)
>> __builtin_abort ();
>> return 0;
>> }
>
> Thanks for the testcase, I agree that is a fairly big problem. Do you
> think that the following (untested) patch is an appropriate way of
> fixing it and generally of extending gimple to capture that a statement
> is a bit-copy?
I think the place to fix is the memcpy folding. That is, we'd say that
aggregate assignments are not bit-copies but do element-wise assignments.
For memcpy folding we'd then need to use a type that doesn't contain
padding. Which effectively means char[].
Of course we need to stop SRA from decomposing that copy to
individual characters then ;)
So iff we decide that all aggregate copies are element copies,
maybe only those where TYPE_MAIN_VARIANT of lhs and rhs match
(currently we allow TYPE_CANONICAL compatibility and thus there
might be some mismatches), then we have to fix nothign but
the memcpy folding.
> If so, I'll add the testcase, bootstrap it and formally propose it.
> Subsequently I will of course make sure that any element-wise copying
> patch would test the predicate.
I don't think the alias-set should determine whether a copy is
bit-wise or not.
Richard.
> Thanks,
>
> Martin
>
>
> 2017-11-23 Martin Jambor <mjambor@suse.cz>
>
> * gimple.c (gimple_bit_copy_p): New function.
> * gimple.h (gimple_bit_copy_p): Declare it.
> * tree-sra.c (sra_modify_assign): Use it.
> ---
> gcc/gimple.c | 20 ++++++++++++++++++++
> gcc/gimple.h | 1 +
> gcc/tree-sra.c | 1 +
> 3 files changed, 22 insertions(+)
>
> diff --git a/gcc/gimple.c b/gcc/gimple.c
> index c986a732004..e1b428d91bb 100644
> --- a/gcc/gimple.c
> +++ b/gcc/gimple.c
> @@ -3087,6 +3087,26 @@ gimple_inexpensive_call_p (gcall *stmt)
> return false;
> }
>
> +/* Return true if STMT is an assignment performing bit copy and so is also
> + expected to copy any padding. */
> +
> +bool
> +gimple_bit_copy_p (gassign *stmt)
> +{
> + if (!gimple_assign_single_p (stmt))
> + return false;
> +
> + tree lhs = gimple_assign_lhs (stmt);
> + if (TREE_CODE (lhs) == MEM_REF
> + && TYPE_REF_CAN_ALIAS_ALL (reference_alias_ptr_type (lhs)))
> + return true;
> + tree rhs = gimple_assign_rhs1 (stmt);
> + if (TREE_CODE (rhs) == MEM_REF
> + && TYPE_REF_CAN_ALIAS_ALL (reference_alias_ptr_type (rhs)))
> + return true;
> + return false;
> +}
> +
> #if CHECKING_P
>
> namespace selftest {
> diff --git a/gcc/gimple.h b/gcc/gimple.h
> index 334def89398..60929473361 100644
> --- a/gcc/gimple.h
> +++ b/gcc/gimple.h
> @@ -1531,6 +1531,7 @@ extern void gimple_seq_discard (gimple_seq);
> extern void maybe_remove_unused_call_args (struct function *, gimple *);
> extern bool gimple_inexpensive_call_p (gcall *);
> extern bool stmt_can_terminate_bb_p (gimple *);
> +extern bool gimple_bit_copy_p (gassign *);
>
> /* Formal (expression) temporary table handling: multiple occurrences of
> the same scalar expression are evaluated into the same temporary. */
> diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
> index db490b20c3e..fc0a8fe60bf 100644
> --- a/gcc/tree-sra.c
> +++ b/gcc/tree-sra.c
> @@ -3591,6 +3591,7 @@ sra_modify_assign (gimple *stmt, gimple_stmt_iterator *gsi)
> || gimple_has_volatile_ops (stmt)
> || contains_vce_or_bfcref_p (rhs)
> || contains_vce_or_bfcref_p (lhs)
> + || gimple_bit_copy_p (as_a <gassign *> (stmt))
> || stmt_ends_bb_p (stmt))
> {
> /* No need to copy into a constant-pool, it comes pre-initialized. */
> --
> 2.15.0
>
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-24 10:47 ` Richard Biener
@ 2017-11-24 11:22 ` Richard Biener
2017-11-24 11:30 ` Richard Biener
0 siblings, 1 reply; 26+ messages in thread
From: Richard Biener @ 2017-11-24 11:22 UTC (permalink / raw)
To: Martin Jambor; +Cc: GCC Patches, Jan Hubicka, Michael Matz
[-- Attachment #1: Type: text/plain, Size: 2850 bytes --]
On Fri, Nov 24, 2017 at 11:31 AM, Richard Biener
<richard.guenther@gmail.com> wrote:
> On Thu, Nov 23, 2017 at 4:32 PM, Martin Jambor <mjambor@suse.cz> wrote:
>> Hi,
>>
>> On Mon, Nov 13 2017, Richard Biener wrote:
>>> The main concern here is that GIMPLE is not very well defined for
>>> aggregate copies and that gimple-fold.c happily optimizes
>>> memcpy (&a, &b, sizeof (a)) into a = b;
>>>
>>> struct A { short s; long i; long j; };
>>> struct A a, b;
>>> void foo ()
>>> {
>>> __builtin_memcpy (&a, &b, sizeof (struct A));
>>> }
>>>
>>> gets folded to
>>>
>>> MEM[(char * {ref-all})&a] = MEM[(char * {ref-all})&b];
>>> return;
>>>
>>> you see we're careful about TBAA but (don't see that above but
>>> can be verified by for example debugging expand_assignment)
>>> TREE_TYPE (MEM[...]) is actually 'struct A'.
>>>
>>> And yes, I've been worried about SRA as well here... it _does_
>>> have some early outs when seeing VIEW_CONVERT_EXPR but
>>> appearantly not for the above. Testcase that aborts with SRA but
>>> not without:
>>>
>>> struct A { short s; long i; long j; };
>>> struct A a, b;
>>> void foo ()
>>> {
>>> struct A c;
>>> __builtin_memcpy (&c, &b, sizeof (struct A));
>>> __builtin_memcpy (&a, &c, sizeof (struct A));
>>> }
>>> int main()
>>> {
>>> __builtin_memset (&b, 0, sizeof (struct A));
>>> b.s = 1;
>>> __builtin_memcpy ((char *)&b+2, &b, 2);
>>> foo ();
>>> __builtin_memcpy (&a, (char *)&a+2, 2);
>>> if (a.s != 1)
>>> __builtin_abort ();
>>> return 0;
>>> }
>>
>> Thanks for the testcase, I agree that is a fairly big problem. Do you
>> think that the following (untested) patch is an appropriate way of
>> fixing it and generally of extending gimple to capture that a statement
>> is a bit-copy?
>
> I think the place to fix is the memcpy folding. That is, we'd say that
> aggregate assignments are not bit-copies but do element-wise assignments.
> For memcpy folding we'd then need to use a type that doesn't contain
> padding. Which effectively means char[].
>
> Of course we need to stop SRA from decomposing that copy to
> individual characters then ;)
>
> So iff we decide that all aggregate copies are element copies,
> maybe only those where TYPE_MAIN_VARIANT of lhs and rhs match
> (currently we allow TYPE_CANONICAL compatibility and thus there
> might be some mismatches), then we have to fix nothign but
> the memcpy folding.
>
>> If so, I'll add the testcase, bootstrap it and formally propose it.
>> Subsequently I will of course make sure that any element-wise copying
>> patch would test the predicate.
>
> I don't think the alias-set should determine whether a copy is
> bit-wise or not.
Like the attached. At least FAILs
FAIL: gcc.dg/tree-ssa/ssa-ccp-27.c scan-tree-dump-times ccp1
"memcpy[^\n]*123456" 2 (found 0 times)
not sure why we have this test.
Richard.
[-- Attachment #2: p3-2 --]
[-- Type: application/octet-stream, Size: 7170 bytes --]
Index: gcc/gimple-fold.c
===================================================================
--- gcc/gimple-fold.c (revision 255137)
+++ gcc/gimple-fold.c (working copy)
@@ -925,161 +925,38 @@ gimple_fold_builtin_memory_op (gimple_st
return false;
}
- if (!tree_fits_shwi_p (len))
+ if (!tree_fits_uhwi_p (len))
return false;
- /* FIXME:
- This logic lose for arguments like (type *)malloc (sizeof (type)),
- since we strip the casts of up to VOID return value from malloc.
- Perhaps we ought to inherit type from non-VOID argument here? */
- STRIP_NOPS (src);
- STRIP_NOPS (dest);
+
if (!POINTER_TYPE_P (TREE_TYPE (src))
|| !POINTER_TYPE_P (TREE_TYPE (dest)))
return false;
- /* In the following try to find a type that is most natural to be
- used for the memcpy source and destination and that allows
- the most optimization when memcpy is turned into a plain assignment
- using that type. In theory we could always use a char[len] type
- but that only gains us that the destination and source possibly
- no longer will have their address taken. */
- /* As we fold (void *)(p + CST) to (void *)p + CST undo this here. */
- if (TREE_CODE (src) == POINTER_PLUS_EXPR)
- {
- tree tem = TREE_OPERAND (src, 0);
- STRIP_NOPS (tem);
- if (tem != TREE_OPERAND (src, 0))
- src = build1 (NOP_EXPR, TREE_TYPE (tem), src);
- }
- if (TREE_CODE (dest) == POINTER_PLUS_EXPR)
- {
- tree tem = TREE_OPERAND (dest, 0);
- STRIP_NOPS (tem);
- if (tem != TREE_OPERAND (dest, 0))
- dest = build1 (NOP_EXPR, TREE_TYPE (tem), dest);
- }
- srctype = TREE_TYPE (TREE_TYPE (src));
- if (TREE_CODE (srctype) == ARRAY_TYPE
- && !tree_int_cst_equal (TYPE_SIZE_UNIT (srctype), len))
- {
- srctype = TREE_TYPE (srctype);
- STRIP_NOPS (src);
- src = build1 (NOP_EXPR, build_pointer_type (srctype), src);
- }
- desttype = TREE_TYPE (TREE_TYPE (dest));
- if (TREE_CODE (desttype) == ARRAY_TYPE
- && !tree_int_cst_equal (TYPE_SIZE_UNIT (desttype), len))
- {
- desttype = TREE_TYPE (desttype);
- STRIP_NOPS (dest);
- dest = build1 (NOP_EXPR, build_pointer_type (desttype), dest);
- }
- if (TREE_ADDRESSABLE (srctype)
- || TREE_ADDRESSABLE (desttype))
- return false;
- /* Make sure we are not copying using a floating-point mode or
- a type whose size possibly does not match its precision. */
- if (FLOAT_MODE_P (TYPE_MODE (desttype))
- || TREE_CODE (desttype) == BOOLEAN_TYPE
- || TREE_CODE (desttype) == ENUMERAL_TYPE)
- desttype = bitwise_type_for_mode (TYPE_MODE (desttype));
- if (FLOAT_MODE_P (TYPE_MODE (srctype))
- || TREE_CODE (srctype) == BOOLEAN_TYPE
- || TREE_CODE (srctype) == ENUMERAL_TYPE)
- srctype = bitwise_type_for_mode (TYPE_MODE (srctype));
- if (!srctype)
- srctype = desttype;
- if (!desttype)
- desttype = srctype;
- if (!srctype)
+ /* We want the possibility of a decl becoming non-addressable. */
+ if (! ((TREE_CODE (dest) == ADDR_EXPR
+ && var_decl_component_p (TREE_OPERAND (dest, 0)))
+ || (TREE_CODE (src) == ADDR_EXPR
+ && var_decl_component_p (TREE_OPERAND (src, 0)))))
return false;
+ /* Use an unsigned char[] type to perform the copying to preserve
+ padding and to avoid any issues with TREE_ADDRESSABLE types
+ or float modes behavior on copying. */
+ desttype = build_array_type_nelts (unsigned_char_type_node,
+ tree_to_uhwi (len));
+ srctype = desttype;
+
+ /* Preserve higher pointer alignment. */
src_align = get_pointer_alignment (src);
dest_align = get_pointer_alignment (dest);
- if (dest_align < TYPE_ALIGN (desttype)
- || src_align < TYPE_ALIGN (srctype))
- return false;
-
- destvar = dest;
- STRIP_NOPS (destvar);
- if (TREE_CODE (destvar) == ADDR_EXPR
- && var_decl_component_p (TREE_OPERAND (destvar, 0))
- && tree_int_cst_equal (TYPE_SIZE_UNIT (desttype), len))
- destvar = fold_build2 (MEM_REF, desttype, destvar, off0);
- else
- destvar = NULL_TREE;
-
- srcvar = src;
- STRIP_NOPS (srcvar);
- if (TREE_CODE (srcvar) == ADDR_EXPR
- && var_decl_component_p (TREE_OPERAND (srcvar, 0))
- && tree_int_cst_equal (TYPE_SIZE_UNIT (srctype), len))
- {
- if (!destvar
- || src_align >= TYPE_ALIGN (desttype))
- srcvar = fold_build2 (MEM_REF, destvar ? desttype : srctype,
- srcvar, off0);
- else if (!STRICT_ALIGNMENT)
- {
- srctype = build_aligned_type (TYPE_MAIN_VARIANT (desttype),
- src_align);
- srcvar = fold_build2 (MEM_REF, srctype, srcvar, off0);
- }
- else
- srcvar = NULL_TREE;
- }
- else
- srcvar = NULL_TREE;
-
- if (srcvar == NULL_TREE && destvar == NULL_TREE)
- return false;
-
- if (srcvar == NULL_TREE)
- {
- STRIP_NOPS (src);
- if (src_align >= TYPE_ALIGN (desttype))
- srcvar = fold_build2 (MEM_REF, desttype, src, off0);
- else
- {
- if (STRICT_ALIGNMENT)
- return false;
- srctype = build_aligned_type (TYPE_MAIN_VARIANT (desttype),
- src_align);
- srcvar = fold_build2 (MEM_REF, srctype, src, off0);
- }
- }
- else if (destvar == NULL_TREE)
- {
- STRIP_NOPS (dest);
- if (dest_align >= TYPE_ALIGN (srctype))
- destvar = fold_build2 (MEM_REF, srctype, dest, off0);
- else
- {
- if (STRICT_ALIGNMENT)
- return false;
- desttype = build_aligned_type (TYPE_MAIN_VARIANT (srctype),
- dest_align);
- destvar = fold_build2 (MEM_REF, desttype, dest, off0);
- }
- }
+ if (src_align > TYPE_ALIGN (srctype))
+ srctype = build_aligned_type (srctype, src_align);
+ if (dest_align > TYPE_ALIGN (desttype))
+ desttype = build_aligned_type (desttype, dest_align);
- gimple *new_stmt;
- if (is_gimple_reg_type (TREE_TYPE (srcvar)))
- {
- tree tem = fold_const_aggregate_ref (srcvar);
- if (tem)
- srcvar = tem;
- if (! is_gimple_min_invariant (srcvar))
- {
- new_stmt = gimple_build_assign (NULL_TREE, srcvar);
- srcvar = create_tmp_reg_or_ssa_name (TREE_TYPE (srcvar),
- new_stmt);
- gimple_assign_set_lhs (new_stmt, srcvar);
- gimple_set_vuse (new_stmt, gimple_vuse (stmt));
- gsi_insert_before (gsi, new_stmt, GSI_SAME_STMT);
- }
- }
- new_stmt = gimple_build_assign (destvar, srcvar);
+ gimple *new_stmt
+ = gimple_build_assign (fold_build2 (MEM_REF, desttype, dest, off0),
+ fold_build2 (MEM_REF, srctype, src, off0));
gimple_set_vuse (new_stmt, gimple_vuse (stmt));
gimple_set_vdef (new_stmt, gimple_vdef (stmt));
if (gimple_vdef (new_stmt)
@@ -6121,7 +5998,10 @@ gimple_fold_stmt_to_constant_1 (gimple *
tree op0 = (*valueize) (gimple_assign_rhs1 (stmt));
tree op1 = (*valueize) (gimple_assign_rhs2 (stmt));
if (TREE_CODE (op0) == ADDR_EXPR
- && TREE_CODE (op1) == INTEGER_CST)
+ && TREE_CODE (op1) == INTEGER_CST
+ && (!cfun
+ || cfun->after_inlining
+ || !handled_component_p (TREE_OPERAND (op0, 0))))
{
tree off = fold_convert (ptr_type_node, op1);
return build_fold_addr_expr_loc
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-24 11:22 ` Richard Biener
@ 2017-11-24 11:30 ` Richard Biener
2017-11-24 12:04 ` Martin Jambor
0 siblings, 1 reply; 26+ messages in thread
From: Richard Biener @ 2017-11-24 11:30 UTC (permalink / raw)
To: Martin Jambor; +Cc: GCC Patches, Jan Hubicka, Michael Matz
On Fri, Nov 24, 2017 at 11:57 AM, Richard Biener
<richard.guenther@gmail.com> wrote:
> On Fri, Nov 24, 2017 at 11:31 AM, Richard Biener
> <richard.guenther@gmail.com> wrote:
>> On Thu, Nov 23, 2017 at 4:32 PM, Martin Jambor <mjambor@suse.cz> wrote:
>>> Hi,
>>>
>>> On Mon, Nov 13 2017, Richard Biener wrote:
>>>> The main concern here is that GIMPLE is not very well defined for
>>>> aggregate copies and that gimple-fold.c happily optimizes
>>>> memcpy (&a, &b, sizeof (a)) into a = b;
>>>>
>>>> struct A { short s; long i; long j; };
>>>> struct A a, b;
>>>> void foo ()
>>>> {
>>>> __builtin_memcpy (&a, &b, sizeof (struct A));
>>>> }
>>>>
>>>> gets folded to
>>>>
>>>> MEM[(char * {ref-all})&a] = MEM[(char * {ref-all})&b];
>>>> return;
>>>>
>>>> you see we're careful about TBAA but (don't see that above but
>>>> can be verified by for example debugging expand_assignment)
>>>> TREE_TYPE (MEM[...]) is actually 'struct A'.
>>>>
>>>> And yes, I've been worried about SRA as well here... it _does_
>>>> have some early outs when seeing VIEW_CONVERT_EXPR but
>>>> appearantly not for the above. Testcase that aborts with SRA but
>>>> not without:
>>>>
>>>> struct A { short s; long i; long j; };
>>>> struct A a, b;
>>>> void foo ()
>>>> {
>>>> struct A c;
>>>> __builtin_memcpy (&c, &b, sizeof (struct A));
>>>> __builtin_memcpy (&a, &c, sizeof (struct A));
>>>> }
>>>> int main()
>>>> {
>>>> __builtin_memset (&b, 0, sizeof (struct A));
>>>> b.s = 1;
>>>> __builtin_memcpy ((char *)&b+2, &b, 2);
>>>> foo ();
>>>> __builtin_memcpy (&a, (char *)&a+2, 2);
>>>> if (a.s != 1)
>>>> __builtin_abort ();
>>>> return 0;
>>>> }
>>>
>>> Thanks for the testcase, I agree that is a fairly big problem. Do you
>>> think that the following (untested) patch is an appropriate way of
>>> fixing it and generally of extending gimple to capture that a statement
>>> is a bit-copy?
>>
>> I think the place to fix is the memcpy folding. That is, we'd say that
>> aggregate assignments are not bit-copies but do element-wise assignments.
>> For memcpy folding we'd then need to use a type that doesn't contain
>> padding. Which effectively means char[].
>>
>> Of course we need to stop SRA from decomposing that copy to
>> individual characters then ;)
>>
>> So iff we decide that all aggregate copies are element copies,
>> maybe only those where TYPE_MAIN_VARIANT of lhs and rhs match
>> (currently we allow TYPE_CANONICAL compatibility and thus there
>> might be some mismatches), then we have to fix nothign but
>> the memcpy folding.
>>
>>> If so, I'll add the testcase, bootstrap it and formally propose it.
>>> Subsequently I will of course make sure that any element-wise copying
>>> patch would test the predicate.
>>
>> I don't think the alias-set should determine whether a copy is
>> bit-wise or not.
>
> Like the attached. At least FAILs
>
> FAIL: gcc.dg/tree-ssa/ssa-ccp-27.c scan-tree-dump-times ccp1
> "memcpy[^\n]*123456" 2 (found 0 times)
>
> not sure why we have this test.
Hum. And SRA still decomposes the copy to struct elements w/o padding
even though the access is done using char[]. So somehow it ignores
VIEW_CONVERT_EXPRs
(well, those implicitely present on MEM_REFs).
Looks like this is because total scalarization is done on the decls
and not at all
honoring how the variable is accessed? The following seems to fix
that, otherwise untested.
Index: gcc/tree-sra.c
===================================================================
--- gcc/tree-sra.c (revision 255137)
+++ gcc/tree-sra.c (working copy)
@@ -1338,7 +1338,9 @@ build_accesses_from_assign (gimple *stmt
{
racc->grp_assignment_read = 1;
if (should_scalarize_away_bitmap && !gimple_has_volatile_ops (stmt)
- && !is_gimple_reg_type (racc->type))
+ && !is_gimple_reg_type (racc->type)
+ && (TYPE_MAIN_VARIANT (racc->type)
+ == TYPE_MAIN_VARIANT (TREE_TYPE (racc->base))))
bitmap_set_bit (should_scalarize_away_bitmap, DECL_UID (racc->base));
if (storage_order_barrier_p (lhs))
racc->grp_unscalarizable_region = 1;
I'm giving this full testing with the folding fix.
Richard.
> Richard.
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-24 11:30 ` Richard Biener
@ 2017-11-24 12:04 ` Martin Jambor
2017-11-24 12:06 ` Richard Biener
0 siblings, 1 reply; 26+ messages in thread
From: Martin Jambor @ 2017-11-24 12:04 UTC (permalink / raw)
To: Richard Biener; +Cc: GCC Patches, Jan Hubicka, Michael Matz
Hi Richi,
On Fri, Nov 24 2017, Richard Biener wrote:
> On Fri, Nov 24, 2017 at 11:57 AM, Richard Biener
> <richard.guenther@gmail.com> wrote:
>> On Fri, Nov 24, 2017 at 11:31 AM, Richard Biener
..
>>>>> And yes, I've been worried about SRA as well here... it _does_
>>>>> have some early outs when seeing VIEW_CONVERT_EXPR but
>>>>> appearantly not for the above. Testcase that aborts with SRA but
>>>>> not without:
>>>>>
>>>>> struct A { short s; long i; long j; };
>>>>> struct A a, b;
>>>>> void foo ()
>>>>> {
>>>>> struct A c;
>>>>> __builtin_memcpy (&c, &b, sizeof (struct A));
>>>>> __builtin_memcpy (&a, &c, sizeof (struct A));
>>>>> }
>>>>> int main()
>>>>> {
>>>>> __builtin_memset (&b, 0, sizeof (struct A));
>>>>> b.s = 1;
>>>>> __builtin_memcpy ((char *)&b+2, &b, 2);
>>>>> foo ();
>>>>> __builtin_memcpy (&a, (char *)&a+2, 2);
>>>>> if (a.s != 1)
>>>>> __builtin_abort ();
>>>>> return 0;
>>>>> }
>>>>
>>>> Thanks for the testcase, I agree that is a fairly big problem. Do you
>>>> think that the following (untested) patch is an appropriate way of
>>>> fixing it and generally of extending gimple to capture that a statement
>>>> is a bit-copy?
>>>
>>> I think the place to fix is the memcpy folding. That is, we'd say that
>>> aggregate assignments are not bit-copies but do element-wise assignments.
>>> For memcpy folding we'd then need to use a type that doesn't contain
>>> padding. Which effectively means char[].
>>>
>>> Of course we need to stop SRA from decomposing that copy to
>>> individual characters then ;)
>>>
>>> So iff we decide that all aggregate copies are element copies,
>>> maybe only those where TYPE_MAIN_VARIANT of lhs and rhs match
>>> (currently we allow TYPE_CANONICAL compatibility and thus there
>>> might be some mismatches), then we have to fix nothign but
>>> the memcpy folding.
>>>
>>>> If so, I'll add the testcase, bootstrap it and formally propose it.
>>>> Subsequently I will of course make sure that any element-wise copying
>>>> patch would test the predicate.
>>>
>>> I don't think the alias-set should determine whether a copy is
>>> bit-wise or not.
>>
>> Like the attached. At least FAILs
>>
>> FAIL: gcc.dg/tree-ssa/ssa-ccp-27.c scan-tree-dump-times ccp1
>> "memcpy[^\n]*123456" 2 (found 0 times)
>>
>> not sure why we have this test.
>
> Hum. And SRA still decomposes the copy to struct elements w/o padding
> even though the access is done using char[]. So somehow it ignores
> VIEW_CONVERT_EXPRs
> (well, those implicitely present on MEM_REFs).
Yes. SRA is not even too afraid of top-level V_C_Es. It really bails
out only if they are buried under a handled_component. And it does not
remove aggregate assignments containing them.
>
> Looks like this is because total scalarization is done on the decls
> and not at all
> honoring how the variable is accessed? The following seems to fix
> that, otherwise untested.
>
> Index: gcc/tree-sra.c
> ===================================================================
> --- gcc/tree-sra.c (revision 255137)
> +++ gcc/tree-sra.c (working copy)
> @@ -1338,7 +1338,9 @@ build_accesses_from_assign (gimple *stmt
> {
> racc->grp_assignment_read = 1;
> if (should_scalarize_away_bitmap && !gimple_has_volatile_ops (stmt)
> - && !is_gimple_reg_type (racc->type))
> + && !is_gimple_reg_type (racc->type)
> + && (TYPE_MAIN_VARIANT (racc->type)
> + == TYPE_MAIN_VARIANT (TREE_TYPE (racc->base))))
> bitmap_set_bit (should_scalarize_away_bitmap, DECL_UID (racc->base));
> if (storage_order_barrier_p (lhs))
> racc->grp_unscalarizable_region = 1;
I believe that the added condition is not what you want, this seems
to trigger also for ordinary:
s1 = s2.field
Where racc->type is type of the field but racc->base is s2 and its type
is type of the structure.
I also think you want to be setting a bit in
cannot_scalarize_away_bitmap in order to guarantee that total
scalarization will not happen for the given candidate. Otherwise some
other regular assignment might trigger it ...except if we then also
checked the statement for bit-copying types in sra_modify_assign (in the
condition after the big comment), which I suppose is actually the
correct thing to do.
Thanks a lot for the folding patch, I can take over the SRA bits if you
want to.
Martin
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-24 12:04 ` Martin Jambor
@ 2017-11-24 12:06 ` Richard Biener
2017-11-24 13:31 ` Martin Jambor
0 siblings, 1 reply; 26+ messages in thread
From: Richard Biener @ 2017-11-24 12:06 UTC (permalink / raw)
To: Martin Jambor; +Cc: GCC Patches, Jan Hubicka, Michael Matz
[-- Attachment #1: Type: text/plain, Size: 5457 bytes --]
On Fri, Nov 24, 2017 at 12:53 PM, Martin Jambor <mjambor@suse.cz> wrote:
> Hi Richi,
>
> On Fri, Nov 24 2017, Richard Biener wrote:
>> On Fri, Nov 24, 2017 at 11:57 AM, Richard Biener
>> <richard.guenther@gmail.com> wrote:
>>> On Fri, Nov 24, 2017 at 11:31 AM, Richard Biener
>
> ..
>
>>>>>> And yes, I've been worried about SRA as well here... it _does_
>>>>>> have some early outs when seeing VIEW_CONVERT_EXPR but
>>>>>> appearantly not for the above. Testcase that aborts with SRA but
>>>>>> not without:
>>>>>>
>>>>>> struct A { short s; long i; long j; };
>>>>>> struct A a, b;
>>>>>> void foo ()
>>>>>> {
>>>>>> struct A c;
>>>>>> __builtin_memcpy (&c, &b, sizeof (struct A));
>>>>>> __builtin_memcpy (&a, &c, sizeof (struct A));
>>>>>> }
>>>>>> int main()
>>>>>> {
>>>>>> __builtin_memset (&b, 0, sizeof (struct A));
>>>>>> b.s = 1;
>>>>>> __builtin_memcpy ((char *)&b+2, &b, 2);
>>>>>> foo ();
>>>>>> __builtin_memcpy (&a, (char *)&a+2, 2);
>>>>>> if (a.s != 1)
>>>>>> __builtin_abort ();
>>>>>> return 0;
>>>>>> }
>>>>>
>>>>> Thanks for the testcase, I agree that is a fairly big problem. Do you
>>>>> think that the following (untested) patch is an appropriate way of
>>>>> fixing it and generally of extending gimple to capture that a statement
>>>>> is a bit-copy?
>>>>
>>>> I think the place to fix is the memcpy folding. That is, we'd say that
>>>> aggregate assignments are not bit-copies but do element-wise assignments.
>>>> For memcpy folding we'd then need to use a type that doesn't contain
>>>> padding. Which effectively means char[].
>>>>
>>>> Of course we need to stop SRA from decomposing that copy to
>>>> individual characters then ;)
>>>>
>>>> So iff we decide that all aggregate copies are element copies,
>>>> maybe only those where TYPE_MAIN_VARIANT of lhs and rhs match
>>>> (currently we allow TYPE_CANONICAL compatibility and thus there
>>>> might be some mismatches), then we have to fix nothign but
>>>> the memcpy folding.
>>>>
>>>>> If so, I'll add the testcase, bootstrap it and formally propose it.
>>>>> Subsequently I will of course make sure that any element-wise copying
>>>>> patch would test the predicate.
>>>>
>>>> I don't think the alias-set should determine whether a copy is
>>>> bit-wise or not.
>>>
>>> Like the attached. At least FAILs
>>>
>>> FAIL: gcc.dg/tree-ssa/ssa-ccp-27.c scan-tree-dump-times ccp1
>>> "memcpy[^\n]*123456" 2 (found 0 times)
>>>
>>> not sure why we have this test.
>>
>> Hum. And SRA still decomposes the copy to struct elements w/o padding
>> even though the access is done using char[]. So somehow it ignores
>> VIEW_CONVERT_EXPRs
>> (well, those implicitely present on MEM_REFs).
>
> Yes. SRA is not even too afraid of top-level V_C_Es. It really bails
> out only if they are buried under a handled_component. And it does not
> remove aggregate assignments containing them.
>
>>
>> Looks like this is because total scalarization is done on the decls
>> and not at all
>> honoring how the variable is accessed? The following seems to fix
>> that, otherwise untested.
>
>
>>
>> Index: gcc/tree-sra.c
>> ===================================================================
>> --- gcc/tree-sra.c (revision 255137)
>> +++ gcc/tree-sra.c (working copy)
>> @@ -1338,7 +1338,9 @@ build_accesses_from_assign (gimple *stmt
>> {
>> racc->grp_assignment_read = 1;
>> if (should_scalarize_away_bitmap && !gimple_has_volatile_ops (stmt)
>> - && !is_gimple_reg_type (racc->type))
>> + && !is_gimple_reg_type (racc->type)
>> + && (TYPE_MAIN_VARIANT (racc->type)
>> + == TYPE_MAIN_VARIANT (TREE_TYPE (racc->base))))
>> bitmap_set_bit (should_scalarize_away_bitmap, DECL_UID (racc->base));
>> if (storage_order_barrier_p (lhs))
>> racc->grp_unscalarizable_region = 1;
>
> I believe that the added condition is not what you want, this seems
> to trigger also for ordinary:
>
> s1 = s2.field
>
> Where racc->type is type of the field but racc->base is s2 and its type
> is type of the structure.
Yes. But do we want to totally scalarize s2 in this case? We only
access parts of it. We don't seem to have a testcase that fails
(well, full testing still in progress).
> I also think you want to be setting a bit in
> cannot_scalarize_away_bitmap in order to guarantee that total
> scalarization will not happen for the given candidate. Otherwise some
> other regular assignment might trigger it ...
Yeah, figured that out myself.
> except if we then also
> checked the statement for bit-copying types in sra_modify_assign (in the
> condition after the big comment), which I suppose is actually the
> correct thing to do.
But modification is too late, no?
> Thanks a lot for the folding patch, I can take over the SRA bits if you
> want to.
For reference below is the full patch.
Bootstrapped on x86_64-unknown-linux-gnu, testing in progress.
Ok for the SRA parts?
Thanks,
Richard.
2017-11-24 Richard Biener <rguenther@suse.de>
PR tree-optimization/83141
* gimple-fold.c (gimple_fold_builtin_memory_op): Simplify
aggregate copy generation by always using a unsigned char[]
type to perform the copying.
* tree-sra.c (build_accesses_from_assign): Disqualify
accesses in non-native type for total scalarization.
* gcc.dg/torture/pr83141.c: New testcase.
* gcc.dg/tree-ssa/ssa-ccp-27.c: Adjust.
[-- Attachment #2: p3-2 --]
[-- Type: application/octet-stream, Size: 9390 bytes --]
2017-11-24 Richard Biener <rguenther@suse.de>
PR tree-optimization/83141
* gimple-fold.c (gimple_fold_builtin_memory_op): Simplify
aggregate copy generation by always using a unsigned char[]
type to perform the copying.
* tree-sra.c (build_accesses_from_assign): Disqualify
accesses in non-native type for total scalarization.
* gcc.dg/torture/pr83141.c: New testcase.
* gcc.dg/tree-ssa/ssa-ccp-27.c: Adjust.
Index: gcc/gimple-fold.c
===================================================================
--- gcc/gimple-fold.c (revision 255137)
+++ gcc/gimple-fold.c (working copy)
@@ -925,161 +925,38 @@ gimple_fold_builtin_memory_op (gimple_st
return false;
}
- if (!tree_fits_shwi_p (len))
+ if (!tree_fits_uhwi_p (len))
return false;
- /* FIXME:
- This logic lose for arguments like (type *)malloc (sizeof (type)),
- since we strip the casts of up to VOID return value from malloc.
- Perhaps we ought to inherit type from non-VOID argument here? */
- STRIP_NOPS (src);
- STRIP_NOPS (dest);
+
if (!POINTER_TYPE_P (TREE_TYPE (src))
|| !POINTER_TYPE_P (TREE_TYPE (dest)))
return false;
- /* In the following try to find a type that is most natural to be
- used for the memcpy source and destination and that allows
- the most optimization when memcpy is turned into a plain assignment
- using that type. In theory we could always use a char[len] type
- but that only gains us that the destination and source possibly
- no longer will have their address taken. */
- /* As we fold (void *)(p + CST) to (void *)p + CST undo this here. */
- if (TREE_CODE (src) == POINTER_PLUS_EXPR)
- {
- tree tem = TREE_OPERAND (src, 0);
- STRIP_NOPS (tem);
- if (tem != TREE_OPERAND (src, 0))
- src = build1 (NOP_EXPR, TREE_TYPE (tem), src);
- }
- if (TREE_CODE (dest) == POINTER_PLUS_EXPR)
- {
- tree tem = TREE_OPERAND (dest, 0);
- STRIP_NOPS (tem);
- if (tem != TREE_OPERAND (dest, 0))
- dest = build1 (NOP_EXPR, TREE_TYPE (tem), dest);
- }
- srctype = TREE_TYPE (TREE_TYPE (src));
- if (TREE_CODE (srctype) == ARRAY_TYPE
- && !tree_int_cst_equal (TYPE_SIZE_UNIT (srctype), len))
- {
- srctype = TREE_TYPE (srctype);
- STRIP_NOPS (src);
- src = build1 (NOP_EXPR, build_pointer_type (srctype), src);
- }
- desttype = TREE_TYPE (TREE_TYPE (dest));
- if (TREE_CODE (desttype) == ARRAY_TYPE
- && !tree_int_cst_equal (TYPE_SIZE_UNIT (desttype), len))
- {
- desttype = TREE_TYPE (desttype);
- STRIP_NOPS (dest);
- dest = build1 (NOP_EXPR, build_pointer_type (desttype), dest);
- }
- if (TREE_ADDRESSABLE (srctype)
- || TREE_ADDRESSABLE (desttype))
- return false;
- /* Make sure we are not copying using a floating-point mode or
- a type whose size possibly does not match its precision. */
- if (FLOAT_MODE_P (TYPE_MODE (desttype))
- || TREE_CODE (desttype) == BOOLEAN_TYPE
- || TREE_CODE (desttype) == ENUMERAL_TYPE)
- desttype = bitwise_type_for_mode (TYPE_MODE (desttype));
- if (FLOAT_MODE_P (TYPE_MODE (srctype))
- || TREE_CODE (srctype) == BOOLEAN_TYPE
- || TREE_CODE (srctype) == ENUMERAL_TYPE)
- srctype = bitwise_type_for_mode (TYPE_MODE (srctype));
- if (!srctype)
- srctype = desttype;
- if (!desttype)
- desttype = srctype;
- if (!srctype)
+ /* We want the possibility of a decl becoming non-addressable. */
+ if (! ((TREE_CODE (dest) == ADDR_EXPR
+ && var_decl_component_p (TREE_OPERAND (dest, 0)))
+ || (TREE_CODE (src) == ADDR_EXPR
+ && var_decl_component_p (TREE_OPERAND (src, 0)))))
return false;
+ /* Use an unsigned char[] type to perform the copying to preserve
+ padding and to avoid any issues with TREE_ADDRESSABLE types
+ or float modes behavior on copying. */
+ desttype = build_array_type_nelts (unsigned_char_type_node,
+ tree_to_uhwi (len));
+ srctype = desttype;
+
+ /* Preserve higher pointer alignment. */
src_align = get_pointer_alignment (src);
dest_align = get_pointer_alignment (dest);
- if (dest_align < TYPE_ALIGN (desttype)
- || src_align < TYPE_ALIGN (srctype))
- return false;
-
- destvar = dest;
- STRIP_NOPS (destvar);
- if (TREE_CODE (destvar) == ADDR_EXPR
- && var_decl_component_p (TREE_OPERAND (destvar, 0))
- && tree_int_cst_equal (TYPE_SIZE_UNIT (desttype), len))
- destvar = fold_build2 (MEM_REF, desttype, destvar, off0);
- else
- destvar = NULL_TREE;
-
- srcvar = src;
- STRIP_NOPS (srcvar);
- if (TREE_CODE (srcvar) == ADDR_EXPR
- && var_decl_component_p (TREE_OPERAND (srcvar, 0))
- && tree_int_cst_equal (TYPE_SIZE_UNIT (srctype), len))
- {
- if (!destvar
- || src_align >= TYPE_ALIGN (desttype))
- srcvar = fold_build2 (MEM_REF, destvar ? desttype : srctype,
- srcvar, off0);
- else if (!STRICT_ALIGNMENT)
- {
- srctype = build_aligned_type (TYPE_MAIN_VARIANT (desttype),
- src_align);
- srcvar = fold_build2 (MEM_REF, srctype, srcvar, off0);
- }
- else
- srcvar = NULL_TREE;
- }
- else
- srcvar = NULL_TREE;
-
- if (srcvar == NULL_TREE && destvar == NULL_TREE)
- return false;
-
- if (srcvar == NULL_TREE)
- {
- STRIP_NOPS (src);
- if (src_align >= TYPE_ALIGN (desttype))
- srcvar = fold_build2 (MEM_REF, desttype, src, off0);
- else
- {
- if (STRICT_ALIGNMENT)
- return false;
- srctype = build_aligned_type (TYPE_MAIN_VARIANT (desttype),
- src_align);
- srcvar = fold_build2 (MEM_REF, srctype, src, off0);
- }
- }
- else if (destvar == NULL_TREE)
- {
- STRIP_NOPS (dest);
- if (dest_align >= TYPE_ALIGN (srctype))
- destvar = fold_build2 (MEM_REF, srctype, dest, off0);
- else
- {
- if (STRICT_ALIGNMENT)
- return false;
- desttype = build_aligned_type (TYPE_MAIN_VARIANT (srctype),
- dest_align);
- destvar = fold_build2 (MEM_REF, desttype, dest, off0);
- }
- }
+ if (src_align > TYPE_ALIGN (srctype))
+ srctype = build_aligned_type (srctype, src_align);
+ if (dest_align > TYPE_ALIGN (desttype))
+ desttype = build_aligned_type (desttype, dest_align);
- gimple *new_stmt;
- if (is_gimple_reg_type (TREE_TYPE (srcvar)))
- {
- tree tem = fold_const_aggregate_ref (srcvar);
- if (tem)
- srcvar = tem;
- if (! is_gimple_min_invariant (srcvar))
- {
- new_stmt = gimple_build_assign (NULL_TREE, srcvar);
- srcvar = create_tmp_reg_or_ssa_name (TREE_TYPE (srcvar),
- new_stmt);
- gimple_assign_set_lhs (new_stmt, srcvar);
- gimple_set_vuse (new_stmt, gimple_vuse (stmt));
- gsi_insert_before (gsi, new_stmt, GSI_SAME_STMT);
- }
- }
- new_stmt = gimple_build_assign (destvar, srcvar);
+ gimple *new_stmt
+ = gimple_build_assign (fold_build2 (MEM_REF, desttype, dest, off0),
+ fold_build2 (MEM_REF, srctype, src, off0));
gimple_set_vuse (new_stmt, gimple_vuse (stmt));
gimple_set_vdef (new_stmt, gimple_vdef (stmt));
if (gimple_vdef (new_stmt)
Index: gcc/tree-sra.c
===================================================================
--- gcc/tree-sra.c (revision 255137)
+++ gcc/tree-sra.c (working copy)
@@ -1339,7 +1339,15 @@ build_accesses_from_assign (gimple *stmt
racc->grp_assignment_read = 1;
if (should_scalarize_away_bitmap && !gimple_has_volatile_ops (stmt)
&& !is_gimple_reg_type (racc->type))
- bitmap_set_bit (should_scalarize_away_bitmap, DECL_UID (racc->base));
+ {
+ if (TYPE_MAIN_VARIANT (racc->type)
+ == TYPE_MAIN_VARIANT (TREE_TYPE (racc->base)))
+ bitmap_set_bit (should_scalarize_away_bitmap,
+ DECL_UID (racc->base));
+ else
+ bitmap_set_bit (cannot_scalarize_away_bitmap,
+ DECL_UID (racc->base));
+ }
if (storage_order_barrier_p (lhs))
racc->grp_unscalarizable_region = 1;
}
Index: gcc/testsuite/gcc.dg/torture/pr83141.c
===================================================================
--- gcc/testsuite/gcc.dg/torture/pr83141.c (nonexistent)
+++ gcc/testsuite/gcc.dg/torture/pr83141.c (working copy)
@@ -0,0 +1,21 @@
+/* { dg-do run } */
+
+struct A { short s; long i; long j; };
+struct A a, b;
+void foo ()
+{
+ struct A c;
+ __builtin_memcpy (&c, &b, sizeof (struct A));
+ __builtin_memcpy (&a, &c, sizeof (struct A));
+}
+int main()
+{
+ __builtin_memset (&b, 0, sizeof (struct A));
+ b.s = 1;
+ __builtin_memcpy ((char *)&b + sizeof (short), &b, sizeof (short));
+ foo ();
+ __builtin_memcpy (&a, (char *)&a + sizeof (short), sizeof (short));
+ if (a.s != 1)
+ __builtin_abort ();
+ return 0;
+}
Index: gcc/testsuite/gcc.dg/tree-ssa/ssa-ccp-27.c
===================================================================
--- gcc/testsuite/gcc.dg/tree-ssa/ssa-ccp-27.c (revision 255137)
+++ gcc/testsuite/gcc.dg/tree-ssa/ssa-ccp-27.c (working copy)
@@ -1,3 +1,4 @@
+/* PR/41047 PR/27226 */
/* { dg-do compile } */
/* { dg-options "-O2 -fdump-tree-ccp1" } */
@@ -20,4 +21,7 @@ f2 ()
memcpy (c, p, 6);
}
-/* { dg-final { scan-tree-dump-times "memcpy\[^\n\]*123456" 2 "ccp1" } } */
+/* We have this testcase to verify we're not losing alignment of a string
+ constant when expanding a block-copy. The following test doesn't
+ reflect that. */
+/* { dg-final { scan-tree-dump-times "MEM\[^\n\]*123456" 2 "ccp1" } } */
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-24 12:06 ` Richard Biener
@ 2017-11-24 13:31 ` Martin Jambor
2017-11-24 14:16 ` Richard Biener
0 siblings, 1 reply; 26+ messages in thread
From: Martin Jambor @ 2017-11-24 13:31 UTC (permalink / raw)
To: Richard Biener; +Cc: GCC Patches, Jan Hubicka, Michael Matz
On Fri, Nov 24 2017, Richard Biener wrote:
> On Fri, Nov 24, 2017 at 12:53 PM, Martin Jambor <mjambor@suse.cz> wrote:
>> Hi Richi,
>>
>> On Fri, Nov 24 2017, Richard Biener wrote:
>>> On Fri, Nov 24, 2017 at 11:57 AM, Richard Biener
>>> <richard.guenther@gmail.com> wrote:
>>>> On Fri, Nov 24, 2017 at 11:31 AM, Richard Biener
>>
>> ..
>>
>>>>>>> And yes, I've been worried about SRA as well here... it _does_
>>>>>>> have some early outs when seeing VIEW_CONVERT_EXPR but
>>>>>>> appearantly not for the above. Testcase that aborts with SRA but
>>>>>>> not without:
>>>>>>>
>>>>>>> struct A { short s; long i; long j; };
>>>>>>> struct A a, b;
>>>>>>> void foo ()
>>>>>>> {
>>>>>>> struct A c;
>>>>>>> __builtin_memcpy (&c, &b, sizeof (struct A));
>>>>>>> __builtin_memcpy (&a, &c, sizeof (struct A));
>>>>>>> }
>>>>>>> int main()
>>>>>>> {
>>>>>>> __builtin_memset (&b, 0, sizeof (struct A));
>>>>>>> b.s = 1;
>>>>>>> __builtin_memcpy ((char *)&b+2, &b, 2);
>>>>>>> foo ();
>>>>>>> __builtin_memcpy (&a, (char *)&a+2, 2);
>>>>>>> if (a.s != 1)
>>>>>>> __builtin_abort ();
>>>>>>> return 0;
>>>>>>> }
>>>>>>
>>>>>> Thanks for the testcase, I agree that is a fairly big problem. Do you
>>>>>> think that the following (untested) patch is an appropriate way of
>>>>>> fixing it and generally of extending gimple to capture that a statement
>>>>>> is a bit-copy?
>>>>>
>>>>> I think the place to fix is the memcpy folding. That is, we'd say that
>>>>> aggregate assignments are not bit-copies but do element-wise assignments.
>>>>> For memcpy folding we'd then need to use a type that doesn't contain
>>>>> padding. Which effectively means char[].
>>>>>
>>>>> Of course we need to stop SRA from decomposing that copy to
>>>>> individual characters then ;)
>>>>>
>>>>> So iff we decide that all aggregate copies are element copies,
>>>>> maybe only those where TYPE_MAIN_VARIANT of lhs and rhs match
>>>>> (currently we allow TYPE_CANONICAL compatibility and thus there
>>>>> might be some mismatches), then we have to fix nothign but
>>>>> the memcpy folding.
>>>>>
>>>>>> If so, I'll add the testcase, bootstrap it and formally propose it.
>>>>>> Subsequently I will of course make sure that any element-wise copying
>>>>>> patch would test the predicate.
>>>>>
>>>>> I don't think the alias-set should determine whether a copy is
>>>>> bit-wise or not.
>>>>
>>>> Like the attached. At least FAILs
>>>>
>>>> FAIL: gcc.dg/tree-ssa/ssa-ccp-27.c scan-tree-dump-times ccp1
>>>> "memcpy[^\n]*123456" 2 (found 0 times)
>>>>
>>>> not sure why we have this test.
>>>
>>> Hum. And SRA still decomposes the copy to struct elements w/o padding
>>> even though the access is done using char[]. So somehow it ignores
>>> VIEW_CONVERT_EXPRs
>>> (well, those implicitely present on MEM_REFs).
>>
>> Yes. SRA is not even too afraid of top-level V_C_Es. It really bails
>> out only if they are buried under a handled_component. And it does not
>> remove aggregate assignments containing them.
>>
>>>
>>> Looks like this is because total scalarization is done on the decls
>>> and not at all
>>> honoring how the variable is accessed? The following seems to fix
>>> that, otherwise untested.
>>
>>
>>>
>>> Index: gcc/tree-sra.c
>>> ===================================================================
>>> --- gcc/tree-sra.c (revision 255137)
>>> +++ gcc/tree-sra.c (working copy)
>>> @@ -1338,7 +1338,9 @@ build_accesses_from_assign (gimple *stmt
>>> {
>>> racc->grp_assignment_read = 1;
>>> if (should_scalarize_away_bitmap && !gimple_has_volatile_ops (stmt)
>>> - && !is_gimple_reg_type (racc->type))
>>> + && !is_gimple_reg_type (racc->type)
>>> + && (TYPE_MAIN_VARIANT (racc->type)
>>> + == TYPE_MAIN_VARIANT (TREE_TYPE (racc->base))))
>>> bitmap_set_bit (should_scalarize_away_bitmap, DECL_UID (racc->base));
>>> if (storage_order_barrier_p (lhs))
>>> racc->grp_unscalarizable_region = 1;
>>
>> I believe that the added condition is not what you want, this seems
>> to trigger also for ordinary:
>>
>> s1 = s2.field
>>
>> Where racc->type is type of the field but racc->base is s2 and its type
>> is type of the structure.
>
> Yes. But do we want to totally scalarize s2 in this case? We only
> access parts of it. We don't seem to have a testcase that fails
> (well, full testing still in progress).
If we start with
small_s1 = src;
dst = small_s1->even_smaller_struct_field;
and small_s1 is otherwise unused, I think that we want to facilitate
copy propagation with total scalarization too.
>
>> I also think you want to be setting a bit in
>> cannot_scalarize_away_bitmap in order to guarantee that total
>> scalarization will not happen for the given candidate. Otherwise some
>> other regular assignment might trigger it ...
>
> Yeah, figured that out myself.
>
>> except if we then also
>> checked the statement for bit-copying types in sra_modify_assign (in the
>> condition after the big comment), which I suppose is actually the
>> correct thing to do.
>
> But modification is too late, no?
No, at modification time we still decide whether how to deal with an
aggregate assignment. That can be either pessimistically by storing all
RHS replacement back to its original aggregate, leave the original
assignment in place and then load all LHS replacement from its
aggregate, or optimistically by trying to load LHS replacements either
from RHS replacements or at least from the RHS and storing RHS
replacement directly to LHS, hoping to eliminate the original load. It
is this optimistic approach rather than total scalarization what we need
to disable. In fact, just disabling total scalarization is not enough
if SRA can come across accesses to components from elsewhere in function
body, your patch unfortunately still fails for:
volatile short vs;
volatile long vl;
struct A { short s; long i; long j; };
struct A a, b;
void foo ()
{
struct A c;
__builtin_memcpy (&c, &b, sizeof (struct A));
__builtin_memcpy (&a, &c, sizeof (struct A));
vs = c.s;
vl = c.i;
vl = c.j;
}
int main()
{
__builtin_memset (&b, 0, sizeof (struct A));
b.s = 1;
__builtin_memcpy ((char *)&b+2, &b, 2);
foo ();
__builtin_memcpy (&a, (char *)&a+2, 2);
if (a.s != 1)
__builtin_abort ();
return 0;
}
>
>> Thanks a lot for the folding patch, I can take over the SRA bits if you
>> want to.
>
> For reference below is the full patch.
>
> Bootstrapped on x86_64-unknown-linux-gnu, testing in progress.
>
> Ok for the SRA parts?
>
My preferred SRA part would be:
diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
index db490b20c3e..7a0e4d1ae26 100644
--- a/gcc/tree-sra.c
+++ b/gcc/tree-sra.c
@@ -1302,6 +1302,17 @@ comes_initialized_p (tree base)
return TREE_CODE (base) == PARM_DECL || constant_decl_p (base);
}
+/* Return true if REF is a MEM_REF which changes the type of the data it
+ accesses. */
+
+static bool
+type_changing_mem_ref_p (tree ref)
+{
+ return (TREE_CODE (ref) == MEM_REF
+ && (TYPE_MAIN_VARIANT (TREE_TYPE (ref))
+ != TYPE_MAIN_VARIANT (TREE_TYPE (TREE_OPERAND (ref, 0)))));
+}
+
/* Scan expressions occurring in STMT, create access structures for all accesses
to candidates for scalarization and remove those candidates which occur in
statements or expressions that prevent them from being split apart. Return
@@ -1338,7 +1349,8 @@ build_accesses_from_assign (gimple *stmt)
{
racc->grp_assignment_read = 1;
if (should_scalarize_away_bitmap && !gimple_has_volatile_ops (stmt)
- && !is_gimple_reg_type (racc->type))
+ && !is_gimple_reg_type (racc->type)
+ && !type_changing_mem_ref_p (rhs))
bitmap_set_bit (should_scalarize_away_bitmap, DECL_UID (racc->base));
if (storage_order_barrier_p (lhs))
racc->grp_unscalarizable_region = 1;
@@ -3589,6 +3601,8 @@ sra_modify_assign (gimple *stmt, gimple_stmt_iterator *gsi)
if (modify_this_stmt
|| gimple_has_volatile_ops (stmt)
+ || type_changing_mem_ref_p (lhs)
+ || type_changing_mem_ref_p (rhs)
|| contains_vce_or_bfcref_p (rhs)
|| contains_vce_or_bfcref_p (lhs)
|| stmt_ends_bb_p (stmt))
I kept the condition in build_accesses_from_assign in order not to do
unnecessary total-scalarization work, but it is those in
sra_modify_assign that actually ensure we keep the assignment intact.
It still would not ask what Jakub asked for, i.e. keep the old behavior
if there is no padding. But that should be done at the gimple-fold
level too, I believe.
Thanks,
Martin
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [RFC, PR 80689] Copy small aggregates element-wise
2017-11-24 13:31 ` Martin Jambor
@ 2017-11-24 14:16 ` Richard Biener
0 siblings, 0 replies; 26+ messages in thread
From: Richard Biener @ 2017-11-24 14:16 UTC (permalink / raw)
To: Martin Jambor; +Cc: GCC Patches, Jan Hubicka, Michael Matz
On Fri, Nov 24, 2017 at 2:00 PM, Martin Jambor <mjambor@suse.cz> wrote:
> On Fri, Nov 24 2017, Richard Biener wrote:
>> On Fri, Nov 24, 2017 at 12:53 PM, Martin Jambor <mjambor@suse.cz> wrote:
>>> Hi Richi,
>>>
>>> On Fri, Nov 24 2017, Richard Biener wrote:
>>>> On Fri, Nov 24, 2017 at 11:57 AM, Richard Biener
>>>> <richard.guenther@gmail.com> wrote:
>>>>> On Fri, Nov 24, 2017 at 11:31 AM, Richard Biener
>>>
>>> ..
>>>
>>>>>>>> And yes, I've been worried about SRA as well here... it _does_
>>>>>>>> have some early outs when seeing VIEW_CONVERT_EXPR but
>>>>>>>> appearantly not for the above. Testcase that aborts with SRA but
>>>>>>>> not without:
>>>>>>>>
>>>>>>>> struct A { short s; long i; long j; };
>>>>>>>> struct A a, b;
>>>>>>>> void foo ()
>>>>>>>> {
>>>>>>>> struct A c;
>>>>>>>> __builtin_memcpy (&c, &b, sizeof (struct A));
>>>>>>>> __builtin_memcpy (&a, &c, sizeof (struct A));
>>>>>>>> }
>>>>>>>> int main()
>>>>>>>> {
>>>>>>>> __builtin_memset (&b, 0, sizeof (struct A));
>>>>>>>> b.s = 1;
>>>>>>>> __builtin_memcpy ((char *)&b+2, &b, 2);
>>>>>>>> foo ();
>>>>>>>> __builtin_memcpy (&a, (char *)&a+2, 2);
>>>>>>>> if (a.s != 1)
>>>>>>>> __builtin_abort ();
>>>>>>>> return 0;
>>>>>>>> }
>>>>>>>
>>>>>>> Thanks for the testcase, I agree that is a fairly big problem. Do you
>>>>>>> think that the following (untested) patch is an appropriate way of
>>>>>>> fixing it and generally of extending gimple to capture that a statement
>>>>>>> is a bit-copy?
>>>>>>
>>>>>> I think the place to fix is the memcpy folding. That is, we'd say that
>>>>>> aggregate assignments are not bit-copies but do element-wise assignments.
>>>>>> For memcpy folding we'd then need to use a type that doesn't contain
>>>>>> padding. Which effectively means char[].
>>>>>>
>>>>>> Of course we need to stop SRA from decomposing that copy to
>>>>>> individual characters then ;)
>>>>>>
>>>>>> So iff we decide that all aggregate copies are element copies,
>>>>>> maybe only those where TYPE_MAIN_VARIANT of lhs and rhs match
>>>>>> (currently we allow TYPE_CANONICAL compatibility and thus there
>>>>>> might be some mismatches), then we have to fix nothign but
>>>>>> the memcpy folding.
>>>>>>
>>>>>>> If so, I'll add the testcase, bootstrap it and formally propose it.
>>>>>>> Subsequently I will of course make sure that any element-wise copying
>>>>>>> patch would test the predicate.
>>>>>>
>>>>>> I don't think the alias-set should determine whether a copy is
>>>>>> bit-wise or not.
>>>>>
>>>>> Like the attached. At least FAILs
>>>>>
>>>>> FAIL: gcc.dg/tree-ssa/ssa-ccp-27.c scan-tree-dump-times ccp1
>>>>> "memcpy[^\n]*123456" 2 (found 0 times)
>>>>>
>>>>> not sure why we have this test.
>>>>
>>>> Hum. And SRA still decomposes the copy to struct elements w/o padding
>>>> even though the access is done using char[]. So somehow it ignores
>>>> VIEW_CONVERT_EXPRs
>>>> (well, those implicitely present on MEM_REFs).
>>>
>>> Yes. SRA is not even too afraid of top-level V_C_Es. It really bails
>>> out only if they are buried under a handled_component. And it does not
>>> remove aggregate assignments containing them.
>>>
>>>>
>>>> Looks like this is because total scalarization is done on the decls
>>>> and not at all
>>>> honoring how the variable is accessed? The following seems to fix
>>>> that, otherwise untested.
>>>
>>>
>>>>
>>>> Index: gcc/tree-sra.c
>>>> ===================================================================
>>>> --- gcc/tree-sra.c (revision 255137)
>>>> +++ gcc/tree-sra.c (working copy)
>>>> @@ -1338,7 +1338,9 @@ build_accesses_from_assign (gimple *stmt
>>>> {
>>>> racc->grp_assignment_read = 1;
>>>> if (should_scalarize_away_bitmap && !gimple_has_volatile_ops (stmt)
>>>> - && !is_gimple_reg_type (racc->type))
>>>> + && !is_gimple_reg_type (racc->type)
>>>> + && (TYPE_MAIN_VARIANT (racc->type)
>>>> + == TYPE_MAIN_VARIANT (TREE_TYPE (racc->base))))
>>>> bitmap_set_bit (should_scalarize_away_bitmap, DECL_UID (racc->base));
>>>> if (storage_order_barrier_p (lhs))
>>>> racc->grp_unscalarizable_region = 1;
>>>
>>> I believe that the added condition is not what you want, this seems
>>> to trigger also for ordinary:
>>>
>>> s1 = s2.field
>>>
>>> Where racc->type is type of the field but racc->base is s2 and its type
>>> is type of the structure.
>>
>> Yes. But do we want to totally scalarize s2 in this case? We only
>> access parts of it. We don't seem to have a testcase that fails
>> (well, full testing still in progress).
>
> If we start with
>
> small_s1 = src;
> dst = small_s1->even_smaller_struct_field;
>
> and small_s1 is otherwise unused, I think that we want to facilitate
> copy propagation with total scalarization too.
>
>>
>>> I also think you want to be setting a bit in
>>> cannot_scalarize_away_bitmap in order to guarantee that total
>>> scalarization will not happen for the given candidate. Otherwise some
>>> other regular assignment might trigger it ...
>>
>> Yeah, figured that out myself.
>>
>>> except if we then also
>>> checked the statement for bit-copying types in sra_modify_assign (in the
>>> condition after the big comment), which I suppose is actually the
>>> correct thing to do.
>>
>> But modification is too late, no?
>
> No, at modification time we still decide whether how to deal with an
> aggregate assignment. That can be either pessimistically by storing all
> RHS replacement back to its original aggregate, leave the original
> assignment in place and then load all LHS replacement from its
> aggregate, or optimistically by trying to load LHS replacements either
> from RHS replacements or at least from the RHS and storing RHS
> replacement directly to LHS, hoping to eliminate the original load. It
> is this optimistic approach rather than total scalarization what we need
> to disable. In fact, just disabling total scalarization is not enough
> if SRA can come across accesses to components from elsewhere in function
> body, your patch unfortunately still fails for:
>
> volatile short vs;
> volatile long vl;
>
> struct A { short s; long i; long j; };
> struct A a, b;
> void foo ()
> {
> struct A c;
> __builtin_memcpy (&c, &b, sizeof (struct A));
> __builtin_memcpy (&a, &c, sizeof (struct A));
>
> vs = c.s;
> vl = c.i;
> vl = c.j;
> }
> int main()
> {
> __builtin_memset (&b, 0, sizeof (struct A));
> b.s = 1;
> __builtin_memcpy ((char *)&b+2, &b, 2);
> foo ();
> __builtin_memcpy (&a, (char *)&a+2, 2);
> if (a.s != 1)
> __builtin_abort ();
> return 0;
> }
>
>
>>
>>> Thanks a lot for the folding patch, I can take over the SRA bits if you
>>> want to.
>>
>> For reference below is the full patch.
>>
>> Bootstrapped on x86_64-unknown-linux-gnu, testing in progress.
>>
>> Ok for the SRA parts?
>>
>
> My preferred SRA part would be:
>
> diff --git a/gcc/tree-sra.c b/gcc/tree-sra.c
> index db490b20c3e..7a0e4d1ae26 100644
> --- a/gcc/tree-sra.c
> +++ b/gcc/tree-sra.c
> @@ -1302,6 +1302,17 @@ comes_initialized_p (tree base)
> return TREE_CODE (base) == PARM_DECL || constant_decl_p (base);
> }
>
> +/* Return true if REF is a MEM_REF which changes the type of the data it
> + accesses. */
> +
> +static bool
> +type_changing_mem_ref_p (tree ref)
> +{
> + return (TREE_CODE (ref) == MEM_REF
> + && (TYPE_MAIN_VARIANT (TREE_TYPE (ref))
> + != TYPE_MAIN_VARIANT (TREE_TYPE (TREE_OPERAND (ref, 0)))));
> +}
> +
> /* Scan expressions occurring in STMT, create access structures for all accesses
> to candidates for scalarization and remove those candidates which occur in
> statements or expressions that prevent them from being split apart. Return
> @@ -1338,7 +1349,8 @@ build_accesses_from_assign (gimple *stmt)
> {
> racc->grp_assignment_read = 1;
> if (should_scalarize_away_bitmap && !gimple_has_volatile_ops (stmt)
> - && !is_gimple_reg_type (racc->type))
> + && !is_gimple_reg_type (racc->type)
> + && !type_changing_mem_ref_p (rhs))
> bitmap_set_bit (should_scalarize_away_bitmap, DECL_UID (racc->base));
> if (storage_order_barrier_p (lhs))
> racc->grp_unscalarizable_region = 1;
> @@ -3589,6 +3601,8 @@ sra_modify_assign (gimple *stmt, gimple_stmt_iterator *gsi)
>
> if (modify_this_stmt
> || gimple_has_volatile_ops (stmt)
> + || type_changing_mem_ref_p (lhs)
> + || type_changing_mem_ref_p (rhs)
> || contains_vce_or_bfcref_p (rhs)
> || contains_vce_or_bfcref_p (lhs)
> || stmt_ends_bb_p (stmt))
But a type-changing MEM_REF can appear at each ref base. Maybe it's
only relevant
for toplevel ones though, you should know.
So we really need to handle it like we do VIEW_CONVERT_EXPR and/or we need
to handle VIEW_CONVERT_EXPRs somewhere in the ref chain the same.
So if you want to take over the SRA parts that would be nice.
I have to dumb down the memcpy folding a bit as we get too much for the amount
of fallout I want to fix right now.
Richard.
>
> I kept the condition in build_accesses_from_assign in order not to do
> unnecessary total-scalarization work, but it is those in
> sra_modify_assign that actually ensure we keep the assignment intact.
>
> It still would not ask what Jakub asked for, i.e. keep the old behavior
> if there is no padding. But that should be done at the gimple-fold
> level too, I believe.
>
> Thanks,
>
> Martin
^ permalink raw reply [flat|nested] 26+ messages in thread
end of thread, other threads:[~2017-11-24 13:29 UTC | newest]
Thread overview: 26+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2017-10-13 16:31 [RFC, PR 80689] Copy small aggregates element-wise Martin Jambor
2017-10-17 11:44 ` Richard Biener
2017-10-26 12:23 ` Martin Jambor
2017-10-26 12:44 ` Richard Biener
2017-10-26 12:57 ` Jan Hubicka
2017-10-26 14:53 ` Richard Biener
2017-10-26 15:12 ` Richard Biener
2017-11-03 16:38 ` Martin Jambor
2017-11-13 13:17 ` Richard Biener
2017-11-13 13:54 ` Michael Matz
2017-11-13 14:19 ` Richard Biener
2017-11-13 15:33 ` Michael Matz
2017-11-13 16:30 ` Richard Biener
2017-11-14 0:32 ` Eric Botcazou
2017-11-23 16:02 ` Martin Jambor
2017-11-23 16:29 ` Jakub Jelinek
2017-11-24 10:47 ` Richard Biener
2017-11-24 11:22 ` Richard Biener
2017-11-24 11:30 ` Richard Biener
2017-11-24 12:04 ` Martin Jambor
2017-11-24 12:06 ` Richard Biener
2017-11-24 13:31 ` Martin Jambor
2017-11-24 14:16 ` Richard Biener
2017-11-14 10:25 ` Martin Jambor
2017-10-27 12:27 ` Jan Hubicka
2017-10-26 14:18 ` Michael Matz
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).