Message ID | 20230721022343.314043-1-juzhe.zhong@rivai.ai |
---|---|
State | New |
Headers | show |
Series | cleanup: Change condition order | expand |
On Fri, 21 Jul 2023, Juzhe-Zhong wrote: > Hi, Richard and Richi. > > I have double check the recent codes for len && mask support again. > > Some places code structure: > > if (len_mask_fn) > ... > else if (mask_fn) > ... > > some places code structure: > > if (mask_len_fn) > ... > else if (mask) > > Base on previous review comment from Richi: > https://gcc.gnu.org/pipermail/gcc-patches/2023-July/625067.html > > len mask stuff should be checked before mask. > > So I reorder all condition order to check LEN MASK stuff before MASK. > > This is the last clean up patch. > > Boostrap and Regression is on the way. OK. > gcc/ChangeLog: > > * tree-vect-stmts.cc (check_load_store_for_partial_vectors): Change condition order. > (vectorizable_operation): Ditto. > > --- > gcc/tree-vect-stmts.cc | 24 ++++++++++++------------ > 1 file changed, 12 insertions(+), 12 deletions(-) > > diff --git a/gcc/tree-vect-stmts.cc b/gcc/tree-vect-stmts.cc > index d5b4f020332..2fe856db9ab 100644 > --- a/gcc/tree-vect-stmts.cc > +++ b/gcc/tree-vect-stmts.cc > @@ -1635,17 +1635,17 @@ check_load_store_for_partial_vectors (loop_vec_info loop_vinfo, tree vectype, > internal_fn len_ifn = (is_load > ? IFN_MASK_LEN_GATHER_LOAD > : IFN_MASK_LEN_SCATTER_STORE); > - if (internal_gather_scatter_fn_supported_p (ifn, vectype, > + if (internal_gather_scatter_fn_supported_p (len_ifn, vectype, > gs_info->memory_type, > gs_info->offset_vectype, > gs_info->scale)) > - vect_record_loop_mask (loop_vinfo, masks, nvectors, vectype, > - scalar_mask); > - else if (internal_gather_scatter_fn_supported_p (len_ifn, vectype, > + vect_record_loop_len (loop_vinfo, lens, nvectors, vectype, 1); > + else if (internal_gather_scatter_fn_supported_p (ifn, vectype, > gs_info->memory_type, > gs_info->offset_vectype, > gs_info->scale)) > - vect_record_loop_len (loop_vinfo, lens, nvectors, vectype, 1); > + vect_record_loop_mask (loop_vinfo, masks, nvectors, vectype, > + scalar_mask); > else > { > if (dump_enabled_p ()) > @@ -6596,16 +6596,16 @@ vectorizable_operation (vec_info *vinfo, > && LOOP_VINFO_CAN_USE_PARTIAL_VECTORS_P (loop_vinfo) > && mask_out_inactive) > { > - if (cond_fn != IFN_LAST > - && direct_internal_fn_supported_p (cond_fn, vectype, > + if (cond_len_fn != IFN_LAST > + && direct_internal_fn_supported_p (cond_len_fn, vectype, > OPTIMIZE_FOR_SPEED)) > - vect_record_loop_mask (loop_vinfo, masks, ncopies * vec_num, > - vectype, NULL); > - else if (cond_len_fn != IFN_LAST > - && direct_internal_fn_supported_p (cond_len_fn, vectype, > - OPTIMIZE_FOR_SPEED)) > vect_record_loop_len (loop_vinfo, lens, ncopies * vec_num, vectype, > 1); > + else if (cond_fn != IFN_LAST > + && direct_internal_fn_supported_p (cond_fn, vectype, > + OPTIMIZE_FOR_SPEED)) > + vect_record_loop_mask (loop_vinfo, masks, ncopies * vec_num, > + vectype, NULL); > else > { > if (dump_enabled_p ()) >
Commited, thanks Richard.
Bootstrap and regression passed.
------------------ Original ------------------
From: "Richard Biener" <gcc-patches@gcc.gnu.org>;
Date: Fri, Jul 21, 2023 02:20 PM
To: "Juzhe-Zhong"<juzhe.zhong@rivai.ai>;
Cc: "gcc-patches"<gcc-patches@gcc.gnu.org>;"richard.sandiford"<richard.sandiford@arm.com>;
Subject: Re: [PATCH] cleanup: Change condition order
On Fri, 21 Jul 2023, Juzhe-Zhong wrote:
> Hi, Richard and Richi.
>
> I have double check the recent codes for len && mask support again.
>
> Some places code structure:
>
> if (len_mask_fn)
> ...
> else if (mask_fn)
> ...
>
> some places code structure:
>
> if (mask_len_fn)
> ...
> else if (mask)
>
> Base on previous review comment from Richi:
> https://gcc.gnu.org/pipermail/gcc-patches/2023-July/625067.html
>
> len mask stuff should be checked before mask.
>
> So I reorder all condition order to check LEN MASK stuff before MASK.
>
> This is the last clean up patch.
>
> Boostrap and Regression is on the way.
OK.
> gcc/ChangeLog:
>
> * tree-vect-stmts.cc (check_load_store_for_partial_vectors): Change condition order.
> (vectorizable_operation): Ditto.
>
> ---
> gcc/tree-vect-stmts.cc | 24 ++++++++++++------------
> 1 file changed, 12 insertions(+), 12 deletions(-)
>
> diff --git a/gcc/tree-vect-stmts.cc b/gcc/tree-vect-stmts.cc
> index d5b4f020332..2fe856db9ab 100644
> --- a/gcc/tree-vect-stmts.cc
> +++ b/gcc/tree-vect-stmts.cc
> @@ -1635,17 +1635,17 @@ check_load_store_for_partial_vectors (loop_vec_info loop_vinfo, tree vectype,
> internal_fn len_ifn = (is_load
> ? IFN_MASK_LEN_GATHER_LOAD
> : IFN_MASK_LEN_SCATTER_STORE);
> - if (internal_gather_scatter_fn_supported_p (ifn, vectype,
> + if (internal_gather_scatter_fn_supported_p (len_ifn, vectype,
> gs_info->memory_type,
> gs_info->offset_vectype,
> gs_info->scale))
> - vect_record_loop_mask (loop_vinfo, masks, nvectors, vectype,
> - scalar_mask);
> - else if (internal_gather_scatter_fn_supported_p (len_ifn, vectype,
> + vect_record_loop_len (loop_vinfo, lens, nvectors, vectype, 1);
> + else if (internal_gather_scatter_fn_supported_p (ifn, vectype,
> gs_info->memory_type,
> gs_info->offset_vectype,
> gs_info->scale))
> - vect_record_loop_len (loop_vinfo, lens, nvectors, vectype, 1);
> + vect_record_loop_mask (loop_vinfo, masks, nvectors, vectype,
> + scalar_mask);
> else
> {
> if (dump_enabled_p ())
> @@ -6596,16 +6596,16 @@ vectorizable_operation (vec_info *vinfo,
> && LOOP_VINFO_CAN_USE_PARTIAL_VECTORS_P (loop_vinfo)
> && mask_out_inactive)
> {
> - if (cond_fn != IFN_LAST
> - && direct_internal_fn_supported_p (cond_fn, vectype,
> + if (cond_len_fn != IFN_LAST
> + && direct_internal_fn_supported_p (cond_len_fn, vectype,
> OPTIMIZE_FOR_SPEED))
> - vect_record_loop_mask (loop_vinfo, masks, ncopies * vec_num,
> - vectype, NULL);
> - else if (cond_len_fn != IFN_LAST
> - && direct_internal_fn_supported_p (cond_len_fn, vectype,
> - OPTIMIZE_FOR_SPEED))
> vect_record_loop_len (loop_vinfo, lens, ncopies * vec_num, vectype,
> 1);
> + else if (cond_fn != IFN_LAST
> + && direct_internal_fn_supported_p (cond_fn, vectype,
> + OPTIMIZE_FOR_SPEED))
> + vect_record_loop_mask (loop_vinfo, masks, ncopies * vec_num,
> + vectype, NULL);
> else
> {
> if (dump_enabled_p ())
>
diff --git a/gcc/tree-vect-stmts.cc b/gcc/tree-vect-stmts.cc index d5b4f020332..2fe856db9ab 100644 --- a/gcc/tree-vect-stmts.cc +++ b/gcc/tree-vect-stmts.cc @@ -1635,17 +1635,17 @@ check_load_store_for_partial_vectors (loop_vec_info loop_vinfo, tree vectype, internal_fn len_ifn = (is_load ? IFN_MASK_LEN_GATHER_LOAD : IFN_MASK_LEN_SCATTER_STORE); - if (internal_gather_scatter_fn_supported_p (ifn, vectype, + if (internal_gather_scatter_fn_supported_p (len_ifn, vectype, gs_info->memory_type, gs_info->offset_vectype, gs_info->scale)) - vect_record_loop_mask (loop_vinfo, masks, nvectors, vectype, - scalar_mask); - else if (internal_gather_scatter_fn_supported_p (len_ifn, vectype, + vect_record_loop_len (loop_vinfo, lens, nvectors, vectype, 1); + else if (internal_gather_scatter_fn_supported_p (ifn, vectype, gs_info->memory_type, gs_info->offset_vectype, gs_info->scale)) - vect_record_loop_len (loop_vinfo, lens, nvectors, vectype, 1); + vect_record_loop_mask (loop_vinfo, masks, nvectors, vectype, + scalar_mask); else { if (dump_enabled_p ()) @@ -6596,16 +6596,16 @@ vectorizable_operation (vec_info *vinfo, && LOOP_VINFO_CAN_USE_PARTIAL_VECTORS_P (loop_vinfo) && mask_out_inactive) { - if (cond_fn != IFN_LAST - && direct_internal_fn_supported_p (cond_fn, vectype, + if (cond_len_fn != IFN_LAST + && direct_internal_fn_supported_p (cond_len_fn, vectype, OPTIMIZE_FOR_SPEED)) - vect_record_loop_mask (loop_vinfo, masks, ncopies * vec_num, - vectype, NULL); - else if (cond_len_fn != IFN_LAST - && direct_internal_fn_supported_p (cond_len_fn, vectype, - OPTIMIZE_FOR_SPEED)) vect_record_loop_len (loop_vinfo, lens, ncopies * vec_num, vectype, 1); + else if (cond_fn != IFN_LAST + && direct_internal_fn_supported_p (cond_fn, vectype, + OPTIMIZE_FOR_SPEED)) + vect_record_loop_mask (loop_vinfo, masks, ncopies * vec_num, + vectype, NULL); else { if (dump_enabled_p ())