[PATCH]middle-end Add an RPO pass after successful vectorization
Tamar Christina
Tamar.Christina@arm.com
Tue Nov 2 14:36:15 GMT 2021
> -----Original Message-----
> From: Richard Biener <rguenther@suse.de>
> Sent: Tuesday, November 2, 2021 2:24 PM
> To: Tamar Christina <Tamar.Christina@arm.com>
> Cc: gcc-patches@gcc.gnu.org; nd <nd@arm.com>
> Subject: Re: [PATCH]middle-end Add an RPO pass after successful
> vectorization
>
> On Tue, 2 Nov 2021, Tamar Christina wrote:
>
> > Hi All,
> >
> > Following my current SVE predicate optimization series a problem has
> > presented itself in that the way vector masks are generated for masked
> > operations relies on CSE to share masks efficiently.
> >
> > The issue however is that masking is done using the & operand and & is
> > associative and so reassoc decides to reassociate the masked operations.
>
> But it does this for the purpose of canonicalization and thus CSE.
Yes, but it turns something like
(a & b) & mask into a & (b & mask).
When (a & b) is used somewhere else you now lose the CSE. So it's actually hurting
In this case.
>
> > This makes CSE then unable to CSE an unmasked and a masked operation
> > leading to duplicate operations being performed.
> >
> > To counter this we want to add an RPO pass over the vectorized loop
> > body when vectorization succeeds. This makes it then no longer
> > reliant on the RTL level CSE.
> >
> > I have not added a testcase for this as it requires the changes in my
> > patch series, however the entire series relies on this patch to work
> > so all the tests there cover it.
> >
> > Bootstrapped Regtested on aarch64-none-linux-gnu, x86_64-linux-gnu and
> > no issues.
> >
> > Ok for master?
>
> You are running VN over _all_ loop bodies rather only those vectorized.
> We loop over vectorized loops earlier for optimizing masked store sequences.
> I suppose you could hook in there. I'll also notice that we have
> pass_pre_slp_scalar_cleanup which eventually runs plus we have a late FRE.
> So I don't understand why it doesn't work to CSE later.
>
Atm, say you have the conditions a > b, and a > b & a > c
We generate
mask1 = (a > b) & loop_mask
mask2 = (a > b & a > c) & loop_mask
with the intention that mask1 can be re-used in mask2.
Reassoc changes this to mask2 = a > b & (a > c & loop_mask)
Which has now unmasked (a > b) in mask2, which leaves us unable to combine
the mask1 and mask2. It doesn't generate incorrect code, just inefficient.
> for (i = 1; i < number_of_loops (cfun); i++)
> {
> loop_vec_info loop_vinfo;
> bool has_mask_store;
>
> loop = get_loop (cfun, i);
> if (!loop || !loop->aux)
> continue;
> loop_vinfo = (loop_vec_info) loop->aux;
> has_mask_store = LOOP_VINFO_HAS_MASK_STORE (loop_vinfo);
> delete loop_vinfo;
> if (has_mask_store
> && targetm.vectorize.empty_mask_is_expensive (IFN_MASK_STORE))
> optimize_mask_stores (loop);
> loop->aux = NULL;
> }
>
Ah thanks, I'll make the changes.
Thanks,
Tamar
>
> > Thanks,
> > Tamar
> >
> > gcc/ChangeLog:
> >
> > * tree-vectorizer.c (vectorize_loops): Do local CSE through RPVN
> upon
> > successful vectorization.
> >
> > --- inline copy of patch --
> > diff --git a/gcc/tree-vectorizer.c b/gcc/tree-vectorizer.c index
> >
> 4712dc6e7f907637774482a71036a0bd381c2bd2..1e370d60fb19b03c3b6bce45c
> 660
> > af4b6d32dc51 100644
> > --- a/gcc/tree-vectorizer.c
> > +++ b/gcc/tree-vectorizer.c
> > @@ -81,7 +81,7 @@ along with GCC; see the file COPYING3. If not see
> > #include "gimple-pretty-print.h"
> > #include "opt-problem.h"
> > #include "internal-fn.h"
> > -
> > +#include "tree-ssa-sccvn.h"
> >
> > /* Loop or bb location, with hotness information. */
> > dump_user_location_t vect_location; @@ -1323,6 +1323,27 @@
> > vectorize_loops (void)
> > ??? Also while we try hard to update loop-closed SSA form we fail
> > to properly do this in some corner-cases (see PR56286). */
> > rewrite_into_loop_closed_ssa (NULL,
> > TODO_update_ssa_only_virtuals);
> > +
> > + for (i = 1; i < number_of_loops (cfun); i++)
> > + {
> > + loop = get_loop (cfun, i);
> > + if (!loop || !single_exit (loop))
> > + continue;
> > +
> > + bitmap exit_bbs;
> > + /* Perform local CSE, this esp. helps because we emit code for
> > + predicates that need to be shared for optimal predicate usage.
> > + However reassoc will re-order them and prevent CSE from working
> > + as it should. CSE only the loop body, not the entry. */
> > + exit_bbs = BITMAP_ALLOC (NULL);
> > + bitmap_set_bit (exit_bbs, single_exit (loop)->dest->index);
> > + bitmap_set_bit (exit_bbs, loop->latch->index);
> > +
> > + do_rpo_vn (cfun, loop_preheader_edge (loop), exit_bbs);
> > +
> > + BITMAP_FREE (exit_bbs);
> > + }
> > +
> > return TODO_cleanup_cfg;
> > }
> >
> >
> >
> >
>
> --
> Richard Biener <rguenther@suse.de>
> SUSE Software Solutions Germany GmbH, Maxfeldstrasse 5, 90409
> Nuernberg, Germany; GF: Ivo Totev; HRB 36809 (AG Nuernberg)
More information about the Gcc-patches
mailing list