From: "Kewen.Lin" <linkw@linux.ibm.com>
To: richard.sandiford@arm.com
Cc: Richard Biener <richard.guenther@gmail.com>,
GCC Patches <gcc-patches@gcc.gnu.org>
Subject: Re: [PATCH] vect: Move VMAT_GATHER_SCATTER handlings from final loop nest
Date: Mon, 14 Aug 2023 21:01:45 +0800 [thread overview]
Message-ID: <dce2077b-81b6-0b78-eb54-a9339309ae78@linux.ibm.com> (raw)
In-Reply-To: <mptzg2trf1l.fsf@arm.com>
Hi Richard,
on 2023/8/14 20:20, Richard Sandiford wrote:
> Thanks for the clean-ups. But...
>
> "Kewen.Lin" <linkw@linux.ibm.com> writes:
>> Hi,
>>
>> Following Richi's suggestion [1], this patch is to move the
>> handlings on VMAT_GATHER_SCATTER in the final loop nest
>> of function vectorizable_load to its own loop. Basically
>> it duplicates the final loop nest, clean up some useless
>> set up code for the case of VMAT_GATHER_SCATTER, remove some
>> unreachable code. Also remove the corresponding handlings
>> in the final loop nest.
>>
>> Bootstrapped and regtested on x86_64-redhat-linux,
>> aarch64-linux-gnu and powerpc64{,le}-linux-gnu.
>>
>> [1] https://gcc.gnu.org/pipermail/gcc-patches/2023-June/623329.html
>>
>> Is it ok for trunk?
>>
>> BR,
>> Kewen
>> -----
>>
>> gcc/ChangeLog:
>>
>> * tree-vect-stmts.cc (vectorizable_load): Move the handlings on
>> VMAT_GATHER_SCATTER in the final loop nest to its own loop,
>> and update the final nest accordingly.
>> ---
>> gcc/tree-vect-stmts.cc | 361 +++++++++++++++++++++++++----------------
>> 1 file changed, 219 insertions(+), 142 deletions(-)
>
> ...that seems like quite a lot of +s. Is there nothing we can do to
> avoid the cut-&-paste?
Thanks for the comments! I'm not sure if I get your question, if we
want to move out the handlings of VMAT_GATHER_SCATTER, the new +s seem
inevitable? Your concern is mainly about git blame history?
BR,
Kewen
>
> Richard
>
>>
>> diff --git a/gcc/tree-vect-stmts.cc b/gcc/tree-vect-stmts.cc
>> index c361e16cb7b..5e514eca19b 100644
>> --- a/gcc/tree-vect-stmts.cc
>> +++ b/gcc/tree-vect-stmts.cc
>> @@ -10455,6 +10455,218 @@ vectorizable_load (vec_info *vinfo,
>> return true;
>> }
>>
>> + if (memory_access_type == VMAT_GATHER_SCATTER)
>> + {
>> + gcc_assert (alignment_support_scheme == dr_aligned
>> + || alignment_support_scheme == dr_unaligned_supported);
>> + gcc_assert (!grouped_load && !slp_perm);
>> +
>> + unsigned int inside_cost = 0, prologue_cost = 0;
>> + for (j = 0; j < ncopies; j++)
>> + {
>> + /* 1. Create the vector or array pointer update chain. */
>> + if (j == 0 && !costing_p)
>> + {
>> + if (STMT_VINFO_GATHER_SCATTER_P (stmt_info))
>> + vect_get_gather_scatter_ops (loop_vinfo, loop, stmt_info,
>> + slp_node, &gs_info, &dataref_ptr,
>> + &vec_offsets);
>> + else
>> + dataref_ptr
>> + = vect_create_data_ref_ptr (vinfo, first_stmt_info, aggr_type,
>> + at_loop, offset, &dummy, gsi,
>> + &ptr_incr, false, bump);
>> + }
>> + else if (!costing_p)
>> + {
>> + gcc_assert (!LOOP_VINFO_USING_SELECT_VL_P (loop_vinfo));
>> + if (!STMT_VINFO_GATHER_SCATTER_P (stmt_info))
>> + dataref_ptr = bump_vector_ptr (vinfo, dataref_ptr, ptr_incr,
>> + gsi, stmt_info, bump);
>> + }
>> +
>> + if (mask && !costing_p)
>> + vec_mask = vec_masks[j];
>> +
>> + gimple *new_stmt = NULL;
>> + for (i = 0; i < vec_num; i++)
>> + {
>> + tree final_mask = NULL_TREE;
>> + tree final_len = NULL_TREE;
>> + tree bias = NULL_TREE;
>> + if (!costing_p)
>> + {
>> + if (loop_masks)
>> + final_mask
>> + = vect_get_loop_mask (loop_vinfo, gsi, loop_masks,
>> + vec_num * ncopies, vectype,
>> + vec_num * j + i);
>> + if (vec_mask)
>> + final_mask = prepare_vec_mask (loop_vinfo, mask_vectype,
>> + final_mask, vec_mask, gsi);
>> +
>> + if (i > 0 && !STMT_VINFO_GATHER_SCATTER_P (stmt_info))
>> + dataref_ptr = bump_vector_ptr (vinfo, dataref_ptr, ptr_incr,
>> + gsi, stmt_info, bump);
>> + }
>> +
>> + /* 2. Create the vector-load in the loop. */
>> + unsigned HOST_WIDE_INT align;
>> + if (gs_info.ifn != IFN_LAST)
>> + {
>> + if (costing_p)
>> + {
>> + unsigned int cnunits = vect_nunits_for_cost (vectype);
>> + inside_cost
>> + = record_stmt_cost (cost_vec, cnunits, scalar_load,
>> + stmt_info, 0, vect_body);
>> + continue;
>> + }
>> + if (STMT_VINFO_GATHER_SCATTER_P (stmt_info))
>> + vec_offset = vec_offsets[vec_num * j + i];
>> + tree zero = build_zero_cst (vectype);
>> + tree scale = size_int (gs_info.scale);
>> +
>> + if (gs_info.ifn == IFN_MASK_LEN_GATHER_LOAD)
>> + {
>> + if (loop_lens)
>> + final_len
>> + = vect_get_loop_len (loop_vinfo, gsi, loop_lens,
>> + vec_num * ncopies, vectype,
>> + vec_num * j + i, 1);
>> + else
>> + final_len
>> + = build_int_cst (sizetype,
>> + TYPE_VECTOR_SUBPARTS (vectype));
>> + signed char biasval
>> + = LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS (loop_vinfo);
>> + bias = build_int_cst (intQI_type_node, biasval);
>> + if (!final_mask)
>> + {
>> + mask_vectype = truth_type_for (vectype);
>> + final_mask = build_minus_one_cst (mask_vectype);
>> + }
>> + }
>> +
>> + gcall *call;
>> + if (final_len && final_mask)
>> + call
>> + = gimple_build_call_internal (IFN_MASK_LEN_GATHER_LOAD, 7,
>> + dataref_ptr, vec_offset,
>> + scale, zero, final_mask,
>> + final_len, bias);
>> + else if (final_mask)
>> + call = gimple_build_call_internal (IFN_MASK_GATHER_LOAD, 5,
>> + dataref_ptr, vec_offset,
>> + scale, zero, final_mask);
>> + else
>> + call = gimple_build_call_internal (IFN_GATHER_LOAD, 4,
>> + dataref_ptr, vec_offset,
>> + scale, zero);
>> + gimple_call_set_nothrow (call, true);
>> + new_stmt = call;
>> + data_ref = NULL_TREE;
>> + }
>> + else
>> + {
>> + /* Emulated gather-scatter. */
>> + gcc_assert (!final_mask);
>> + unsigned HOST_WIDE_INT const_nunits = nunits.to_constant ();
>> + if (costing_p)
>> + {
>> + /* For emulated gathers N offset vector element
>> + offset add is consumed by the load). */
>> + inside_cost = record_stmt_cost (cost_vec, const_nunits,
>> + vec_to_scalar, stmt_info,
>> + 0, vect_body);
>> + /* N scalar loads plus gathering them into a
>> + vector. */
>> + inside_cost
>> + = record_stmt_cost (cost_vec, const_nunits, scalar_load,
>> + stmt_info, 0, vect_body);
>> + inside_cost
>> + = record_stmt_cost (cost_vec, 1, vec_construct,
>> + stmt_info, 0, vect_body);
>> + continue;
>> + }
>> + unsigned HOST_WIDE_INT const_offset_nunits
>> + = TYPE_VECTOR_SUBPARTS (gs_info.offset_vectype)
>> + .to_constant ();
>> + vec<constructor_elt, va_gc> *ctor_elts;
>> + vec_alloc (ctor_elts, const_nunits);
>> + gimple_seq stmts = NULL;
>> + /* We support offset vectors with more elements
>> + than the data vector for now. */
>> + unsigned HOST_WIDE_INT factor
>> + = const_offset_nunits / const_nunits;
>> + vec_offset = vec_offsets[j / factor];
>> + unsigned elt_offset = (j % factor) * const_nunits;
>> + tree idx_type = TREE_TYPE (TREE_TYPE (vec_offset));
>> + tree scale = size_int (gs_info.scale);
>> + align = get_object_alignment (DR_REF (first_dr_info->dr));
>> + tree ltype = build_aligned_type (TREE_TYPE (vectype), align);
>> + for (unsigned k = 0; k < const_nunits; ++k)
>> + {
>> + tree boff = size_binop (MULT_EXPR, TYPE_SIZE (idx_type),
>> + bitsize_int (k + elt_offset));
>> + tree idx
>> + = gimple_build (&stmts, BIT_FIELD_REF, idx_type,
>> + vec_offset, TYPE_SIZE (idx_type), boff);
>> + idx = gimple_convert (&stmts, sizetype, idx);
>> + idx = gimple_build (&stmts, MULT_EXPR, sizetype, idx,
>> + scale);
>> + tree ptr = gimple_build (&stmts, PLUS_EXPR,
>> + TREE_TYPE (dataref_ptr),
>> + dataref_ptr, idx);
>> + ptr = gimple_convert (&stmts, ptr_type_node, ptr);
>> + tree elt = make_ssa_name (TREE_TYPE (vectype));
>> + tree ref = build2 (MEM_REF, ltype, ptr,
>> + build_int_cst (ref_type, 0));
>> + new_stmt = gimple_build_assign (elt, ref);
>> + gimple_set_vuse (new_stmt, gimple_vuse (gsi_stmt (*gsi)));
>> + gimple_seq_add_stmt (&stmts, new_stmt);
>> + CONSTRUCTOR_APPEND_ELT (ctor_elts, NULL_TREE, elt);
>> + }
>> + gsi_insert_seq_before (gsi, stmts, GSI_SAME_STMT);
>> + new_stmt = gimple_build_assign (
>> + NULL_TREE, build_constructor (vectype, ctor_elts));
>> + data_ref = NULL_TREE;
>> + }
>> +
>> + vec_dest = vect_create_destination_var (scalar_dest, vectype);
>> + /* DATA_REF is null if we've already built the statement. */
>> + if (data_ref)
>> + {
>> + vect_copy_ref_info (data_ref, DR_REF (first_dr_info->dr));
>> + new_stmt = gimple_build_assign (vec_dest, data_ref);
>> + }
>> + new_temp = make_ssa_name (vec_dest, new_stmt);
>> + gimple_set_lhs (new_stmt, new_temp);
>> + vect_finish_stmt_generation (vinfo, stmt_info, new_stmt, gsi);
>> +
>> + /* Store vector loads in the corresponding SLP_NODE. */
>> + if (slp)
>> + slp_node->push_vec_def (new_stmt);
>> + }
>> +
>> + if (!slp && !costing_p)
>> + STMT_VINFO_VEC_STMTS (stmt_info).safe_push (new_stmt);
>> + }
>> +
>> + if (!slp && !costing_p)
>> + *vec_stmt = STMT_VINFO_VEC_STMTS (stmt_info)[0];
>> +
>> + if (costing_p)
>> + {
>> + if (dump_enabled_p ())
>> + dump_printf_loc (MSG_NOTE, vect_location,
>> + "vect_model_load_cost: inside_cost = %u, "
>> + "prologue_cost = %u .\n",
>> + inside_cost, prologue_cost);
>> + }
>> + return true;
>> + }
>> +
>> poly_uint64 group_elt = 0;
>> unsigned int inside_cost = 0, prologue_cost = 0;
>> for (j = 0; j < ncopies; j++)
>> @@ -10504,12 +10716,6 @@ vectorizable_load (vec_info *vinfo,
>> gcc_assert (!compute_in_loop);
>> }
>> }
>> - else if (STMT_VINFO_GATHER_SCATTER_P (stmt_info))
>> - {
>> - vect_get_gather_scatter_ops (loop_vinfo, loop, stmt_info,
>> - slp_node, &gs_info, &dataref_ptr,
>> - &vec_offsets);
>> - }
>> else
>> dataref_ptr
>> = vect_create_data_ref_ptr (vinfo, first_stmt_info, aggr_type,
>> @@ -10525,7 +10731,7 @@ vectorizable_load (vec_info *vinfo,
>> if (dataref_offset)
>> dataref_offset = int_const_binop (PLUS_EXPR, dataref_offset,
>> bump);
>> - else if (!STMT_VINFO_GATHER_SCATTER_P (stmt_info))
>> + else
>> dataref_ptr = bump_vector_ptr (vinfo, dataref_ptr, ptr_incr, gsi,
>> stmt_info, bump);
>> if (mask)
>> @@ -10551,7 +10757,7 @@ vectorizable_load (vec_info *vinfo,
>> final_mask = prepare_vec_mask (loop_vinfo, mask_vectype,
>> final_mask, vec_mask, gsi);
>>
>> - if (i > 0 && !STMT_VINFO_GATHER_SCATTER_P (stmt_info))
>> + if (i > 0)
>> dataref_ptr = bump_vector_ptr (vinfo, dataref_ptr, ptr_incr,
>> gsi, stmt_info, bump);
>> }
>> @@ -10562,139 +10768,11 @@ vectorizable_load (vec_info *vinfo,
>> case dr_aligned:
>> case dr_unaligned_supported:
>> {
>> - unsigned int misalign;
>> - unsigned HOST_WIDE_INT align;
>> -
>> - if (memory_access_type == VMAT_GATHER_SCATTER
>> - && gs_info.ifn != IFN_LAST)
>> - {
>> - if (costing_p)
>> - {
>> - unsigned int cnunits = vect_nunits_for_cost (vectype);
>> - inside_cost
>> - = record_stmt_cost (cost_vec, cnunits, scalar_load,
>> - stmt_info, 0, vect_body);
>> - break;
>> - }
>> - if (STMT_VINFO_GATHER_SCATTER_P (stmt_info))
>> - vec_offset = vec_offsets[vec_num * j + i];
>> - tree zero = build_zero_cst (vectype);
>> - tree scale = size_int (gs_info.scale);
>> -
>> - if (gs_info.ifn == IFN_MASK_LEN_GATHER_LOAD)
>> - {
>> - if (loop_lens)
>> - final_len
>> - = vect_get_loop_len (loop_vinfo, gsi, loop_lens,
>> - vec_num * ncopies, vectype,
>> - vec_num * j + i, 1);
>> - else
>> - final_len
>> - = build_int_cst (sizetype,
>> - TYPE_VECTOR_SUBPARTS (vectype));
>> - signed char biasval
>> - = LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS (loop_vinfo);
>> - bias = build_int_cst (intQI_type_node, biasval);
>> - if (!final_mask)
>> - {
>> - mask_vectype = truth_type_for (vectype);
>> - final_mask = build_minus_one_cst (mask_vectype);
>> - }
>> - }
>> -
>> - gcall *call;
>> - if (final_len && final_mask)
>> - call = gimple_build_call_internal (
>> - IFN_MASK_LEN_GATHER_LOAD, 7, dataref_ptr, vec_offset,
>> - scale, zero, final_mask, final_len, bias);
>> - else if (final_mask)
>> - call
>> - = gimple_build_call_internal (IFN_MASK_GATHER_LOAD, 5,
>> - dataref_ptr, vec_offset,
>> - scale, zero, final_mask);
>> - else
>> - call
>> - = gimple_build_call_internal (IFN_GATHER_LOAD, 4,
>> - dataref_ptr, vec_offset,
>> - scale, zero);
>> - gimple_call_set_nothrow (call, true);
>> - new_stmt = call;
>> - data_ref = NULL_TREE;
>> - break;
>> - }
>> - else if (memory_access_type == VMAT_GATHER_SCATTER)
>> - {
>> - /* Emulated gather-scatter. */
>> - gcc_assert (!final_mask);
>> - unsigned HOST_WIDE_INT const_nunits = nunits.to_constant ();
>> - if (costing_p)
>> - {
>> - /* For emulated gathers N offset vector element
>> - offset add is consumed by the load). */
>> - inside_cost
>> - = record_stmt_cost (cost_vec, const_nunits,
>> - vec_to_scalar, stmt_info, 0,
>> - vect_body);
>> - /* N scalar loads plus gathering them into a
>> - vector. */
>> - inside_cost = record_stmt_cost (cost_vec, const_nunits,
>> - scalar_load, stmt_info,
>> - 0, vect_body);
>> - inside_cost
>> - = record_stmt_cost (cost_vec, 1, vec_construct,
>> - stmt_info, 0, vect_body);
>> - break;
>> - }
>> - unsigned HOST_WIDE_INT const_offset_nunits
>> - = TYPE_VECTOR_SUBPARTS (gs_info.offset_vectype)
>> - .to_constant ();
>> - vec<constructor_elt, va_gc> *ctor_elts;
>> - vec_alloc (ctor_elts, const_nunits);
>> - gimple_seq stmts = NULL;
>> - /* We support offset vectors with more elements
>> - than the data vector for now. */
>> - unsigned HOST_WIDE_INT factor
>> - = const_offset_nunits / const_nunits;
>> - vec_offset = vec_offsets[j / factor];
>> - unsigned elt_offset = (j % factor) * const_nunits;
>> - tree idx_type = TREE_TYPE (TREE_TYPE (vec_offset));
>> - tree scale = size_int (gs_info.scale);
>> - align = get_object_alignment (DR_REF (first_dr_info->dr));
>> - tree ltype
>> - = build_aligned_type (TREE_TYPE (vectype), align);
>> - for (unsigned k = 0; k < const_nunits; ++k)
>> - {
>> - tree boff = size_binop (MULT_EXPR, TYPE_SIZE (idx_type),
>> - bitsize_int (k + elt_offset));
>> - tree idx = gimple_build (&stmts, BIT_FIELD_REF,
>> - idx_type, vec_offset,
>> - TYPE_SIZE (idx_type), boff);
>> - idx = gimple_convert (&stmts, sizetype, idx);
>> - idx = gimple_build (&stmts, MULT_EXPR, sizetype, idx,
>> - scale);
>> - tree ptr = gimple_build (&stmts, PLUS_EXPR,
>> - TREE_TYPE (dataref_ptr),
>> - dataref_ptr, idx);
>> - ptr = gimple_convert (&stmts, ptr_type_node, ptr);
>> - tree elt = make_ssa_name (TREE_TYPE (vectype));
>> - tree ref = build2 (MEM_REF, ltype, ptr,
>> - build_int_cst (ref_type, 0));
>> - new_stmt = gimple_build_assign (elt, ref);
>> - gimple_set_vuse (new_stmt,
>> - gimple_vuse (gsi_stmt (*gsi)));
>> - gimple_seq_add_stmt (&stmts, new_stmt);
>> - CONSTRUCTOR_APPEND_ELT (ctor_elts, NULL_TREE, elt);
>> - }
>> - gsi_insert_seq_before (gsi, stmts, GSI_SAME_STMT);
>> - new_stmt = gimple_build_assign (
>> - NULL_TREE, build_constructor (vectype, ctor_elts));
>> - data_ref = NULL_TREE;
>> - break;
>> - }
>> -
>> if (costing_p)
>> break;
>>
>> + unsigned int misalign;
>> + unsigned HOST_WIDE_INT align;
>> align = known_alignment (DR_TARGET_ALIGNMENT (first_dr_info));
>> if (alignment_support_scheme == dr_aligned)
>> misalign = 0;
>> @@ -11156,10 +11234,9 @@ vectorizable_load (vec_info *vinfo,
>>
>> if (costing_p)
>> {
>> - gcc_assert (memory_access_type != VMAT_INVARIANT
>> - && memory_access_type != VMAT_ELEMENTWISE
>> - && memory_access_type != VMAT_STRIDED_SLP
>> - && memory_access_type != VMAT_LOAD_STORE_LANES);
>> + gcc_assert (memory_access_type == VMAT_CONTIGUOUS
>> + || memory_access_type == VMAT_CONTIGUOUS_REVERSE
>> + || memory_access_type == VMAT_CONTIGUOUS_PERMUTE);
>> if (dump_enabled_p ())
>> dump_printf_loc (MSG_NOTE, vect_location,
>> "vect_model_load_cost: inside_cost = %u, "
>> --
>> 2.39.1
next prev parent reply other threads:[~2023-08-14 13:01 UTC|newest]
Thread overview: 18+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-08-14 8:54 [PATCH] vect: Move VMAT_LOAD_STORE_LANES " Kewen.Lin
2023-08-14 8:59 ` [PATCH] vect: Move VMAT_GATHER_SCATTER " Kewen.Lin
2023-08-14 12:20 ` Richard Sandiford
2023-08-14 13:01 ` Kewen.Lin [this message]
2023-08-14 14:16 ` Richard Sandiford
2023-08-15 2:33 ` Kewen.Lin
2023-08-15 7:53 ` Richard Biener
2023-08-15 8:44 ` Richard Sandiford
2023-08-15 9:04 ` Richard Biener
2023-08-15 9:13 ` Richard Sandiford
2023-08-16 2:38 ` Kewen.Lin
2023-08-17 7:47 ` Richard Biener
2023-08-15 11:47 ` Kewen.Lin
2023-08-15 12:07 ` Richard Biener
2023-08-15 12:13 ` Kewen.Lin
2023-08-14 12:04 ` [PATCH] vect: Move VMAT_LOAD_STORE_LANES " Richard Biener
2023-08-14 12:49 ` Kewen.Lin
2023-08-14 13:47 ` Richard Biener
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=dce2077b-81b6-0b78-eb54-a9339309ae78@linux.ibm.com \
--to=linkw@linux.ibm.com \
--cc=gcc-patches@gcc.gnu.org \
--cc=richard.guenther@gmail.com \
--cc=richard.sandiford@arm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).