* [PATCH 2/4] TLC to vect_check_store_rhs and vect_slp_child_index_for_operand
@ 2023-11-08 15:03 Richard Biener
0 siblings, 0 replies; only message in thread
From: Richard Biener @ 2023-11-08 15:03 UTC (permalink / raw)
To: gcc-patches
This prepares us for the SLP of scatters. We have to tell
vect_slp_child_index_for_operand whether we are dealing with
a scatter/gather stmt so this adds an argument similar to
the one we have for vect_get_operand_map. This also refactors
vect_check_store_rhs to get the actual rhs and the associated
SLP node instead of leaving that to the caller.
Bootstrap and regtest running on x86_64-unknown-linux-gnu.
* tree-vectorizer.h (vect_slp_child_index_for_operand):
Add gatherscatter_p argument.
* tree-vect-slp.cc (vect_slp_child_index_for_operand): Likewise.
Pass it on.
* tree-vect-stmts.cc (vect_check_store_rhs): Turn the rhs
argument into an output, also output the SLP node associated
with it.
(vectorizable_simd_clone_call): Adjust.
(vectorizable_store): Likewise.
(vectorizable_load): Likewise.
---
gcc/tree-vect-slp.cc | 5 ++--
gcc/tree-vect-stmts.cc | 52 ++++++++++++++++++++++--------------------
gcc/tree-vectorizer.h | 2 +-
3 files changed, 31 insertions(+), 28 deletions(-)
diff --git a/gcc/tree-vect-slp.cc b/gcc/tree-vect-slp.cc
index 13137ede8d4..176aaf270f4 100644
--- a/gcc/tree-vect-slp.cc
+++ b/gcc/tree-vect-slp.cc
@@ -589,9 +589,10 @@ vect_get_operand_map (const gimple *stmt, bool gather_scatter_p = false,
/* Return the SLP node child index for operand OP of STMT. */
int
-vect_slp_child_index_for_operand (const gimple *stmt, int op)
+vect_slp_child_index_for_operand (const gimple *stmt, int op,
+ bool gather_scatter_p)
{
- const int *opmap = vect_get_operand_map (stmt);
+ const int *opmap = vect_get_operand_map (stmt, gather_scatter_p);
if (!opmap)
return op;
for (int i = 1; i < 1 + opmap[0]; ++i)
diff --git a/gcc/tree-vect-stmts.cc b/gcc/tree-vect-stmts.cc
index 096a857f2dd..61e23b29516 100644
--- a/gcc/tree-vect-stmts.cc
+++ b/gcc/tree-vect-stmts.cc
@@ -2486,42 +2486,33 @@ vect_check_scalar_mask (vec_info *vinfo, stmt_vec_info stmt_info,
return true;
}
-/* Return true if stored value RHS is suitable for vectorizing store
- statement STMT_INFO. When returning true, store the type of the
- definition in *RHS_DT_OUT, the type of the vectorized store value in
+/* Return true if stored value is suitable for vectorizing store
+ statement STMT_INFO. When returning true, store the scalar stored
+ in *RHS and *RHS_NODE, the type of the definition in *RHS_DT_OUT,
+ the type of the vectorized store value in
*RHS_VECTYPE_OUT and the type of the store in *VLS_TYPE_OUT. */
static bool
vect_check_store_rhs (vec_info *vinfo, stmt_vec_info stmt_info,
- slp_tree slp_node, tree rhs,
+ slp_tree slp_node, tree *rhs, slp_tree *rhs_node,
vect_def_type *rhs_dt_out, tree *rhs_vectype_out,
vec_load_store_type *vls_type_out)
{
- /* In the case this is a store from a constant make sure
- native_encode_expr can handle it. */
- if (CONSTANT_CLASS_P (rhs) && native_encode_expr (rhs, NULL, 64) == 0)
- {
- if (dump_enabled_p ())
- dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
- "cannot encode constant as a byte sequence.\n");
- return false;
- }
-
int op_no = 0;
if (gcall *call = dyn_cast <gcall *> (stmt_info->stmt))
{
if (gimple_call_internal_p (call)
&& internal_store_fn_p (gimple_call_internal_fn (call)))
op_no = internal_fn_stored_value_index (gimple_call_internal_fn (call));
- if (slp_node)
- op_no = vect_slp_child_index_for_operand (call, op_no);
}
+ if (slp_node)
+ op_no = vect_slp_child_index_for_operand
+ (stmt_info->stmt, op_no, STMT_VINFO_GATHER_SCATTER_P (stmt_info));
enum vect_def_type rhs_dt;
tree rhs_vectype;
- slp_tree slp_op;
if (!vect_is_simple_use (vinfo, stmt_info, slp_node, op_no,
- &rhs, &slp_op, &rhs_dt, &rhs_vectype))
+ rhs, rhs_node, &rhs_dt, &rhs_vectype))
{
if (dump_enabled_p ())
dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
@@ -2529,6 +2520,16 @@ vect_check_store_rhs (vec_info *vinfo, stmt_vec_info stmt_info,
return false;
}
+ /* In the case this is a store from a constant make sure
+ native_encode_expr can handle it. */
+ if (CONSTANT_CLASS_P (*rhs) && native_encode_expr (*rhs, NULL, 64) == 0)
+ {
+ if (dump_enabled_p ())
+ dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
+ "cannot encode constant as a byte sequence.\n");
+ return false;
+ }
+
tree vectype = STMT_VINFO_VECTYPE (stmt_info);
if (rhs_vectype && !useless_type_conversion_p (vectype, rhs_vectype))
{
@@ -4052,7 +4053,7 @@ vectorizable_simd_clone_call (vec_info *vinfo, stmt_vec_info stmt_info,
int op_no = i + masked_call_offset;
if (slp_node)
- op_no = vect_slp_child_index_for_operand (stmt, op_no);
+ op_no = vect_slp_child_index_for_operand (stmt, op_no, false);
if (!vect_is_simple_use (vinfo, stmt_info, slp_node,
op_no, &op, &slp_op[i],
&thisarginfo.dt, &thisarginfo.vectype)
@@ -8173,7 +8174,6 @@ vectorizable_store (vec_info *vinfo,
stmt_vector_for_cost *cost_vec)
{
tree data_ref;
- tree op;
tree vec_oprnd = NULL_TREE;
tree elem_type;
loop_vec_info loop_vinfo = dyn_cast <loop_vec_info> (vinfo);
@@ -8236,15 +8236,14 @@ vectorizable_store (vec_info *vinfo,
int mask_index = internal_fn_mask_index (ifn);
if (mask_index >= 0 && slp_node)
- mask_index = vect_slp_child_index_for_operand (call, mask_index);
+ mask_index = vect_slp_child_index_for_operand
+ (call, mask_index, STMT_VINFO_GATHER_SCATTER_P (stmt_info));
if (mask_index >= 0
&& !vect_check_scalar_mask (vinfo, stmt_info, slp_node, mask_index,
&mask, NULL, &mask_dt, &mask_vectype))
return false;
}
- op = vect_get_store_rhs (stmt_info);
-
/* Cannot have hybrid store SLP -- that would mean storing to the
same location twice. */
gcc_assert (slp == PURE_SLP_STMT (stmt_info));
@@ -8279,8 +8278,10 @@ vectorizable_store (vec_info *vinfo,
return false;
}
+ tree op;
+ slp_tree op_node;
if (!vect_check_store_rhs (vinfo, stmt_info, slp_node,
- op, &rhs_dt, &rhs_vectype, &vls_type))
+ &op, &op_node, &rhs_dt, &rhs_vectype, &vls_type))
return false;
elem_type = TREE_TYPE (vectype);
@@ -9855,7 +9856,8 @@ vectorizable_load (vec_info *vinfo,
mask_index = internal_fn_mask_index (ifn);
if (mask_index >= 0 && slp_node)
- mask_index = vect_slp_child_index_for_operand (call, mask_index);
+ mask_index = vect_slp_child_index_for_operand
+ (call, mask_index, STMT_VINFO_GATHER_SCATTER_P (stmt_info));
if (mask_index >= 0
&& !vect_check_scalar_mask (vinfo, stmt_info, slp_node, mask_index,
&mask, &slp_op, &mask_dt, &mask_vectype))
diff --git a/gcc/tree-vectorizer.h b/gcc/tree-vectorizer.h
index d2ddc2e4ad5..e4d7ab4567c 100644
--- a/gcc/tree-vectorizer.h
+++ b/gcc/tree-vectorizer.h
@@ -2462,7 +2462,7 @@ extern int vect_get_place_in_interleaving_chain (stmt_vec_info, stmt_vec_info);
extern slp_tree vect_create_new_slp_node (unsigned, tree_code);
extern void vect_free_slp_tree (slp_tree);
extern bool compatible_calls_p (gcall *, gcall *);
-extern int vect_slp_child_index_for_operand (const gimple *, int op);
+extern int vect_slp_child_index_for_operand (const gimple *, int op, bool);
/* In tree-vect-patterns.cc. */
extern void
--
2.35.3
^ permalink raw reply [flat|nested] only message in thread
only message in thread, other threads:[~2023-11-08 15:03 UTC | newest]
Thread overview: (only message) (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2023-11-08 15:03 [PATCH 2/4] TLC to vect_check_store_rhs and vect_slp_child_index_for_operand Richard Biener
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).