On Wed, 12 Apr 2023, [email protected] wrote:
>
> >> Thanks for the detailed explanation. Just to clarify - with RVV
> >> there's only a single mask register, v0.t, or did you want to
> >> say an instruction can only specify a single mask register?
>
> RVV has 32 (v0~v31) vector register in total.
> We can store vector data value or mask value in any of them.
> We also have mask-logic instruction for example mask-and between any vector
> register.
>
> However, any vector operation for example like vadd.vv can only predicated
> by v0 (in asm is v0.t) which is the first vector register.
> We can predicate vadd.vv with v1 - v31.
>
> So, you can image every time we want to use a mask to predicate a vector
> operation, we should always first store the mask value
> into v0.
>
> So, we can write intrinsic sequence like this:
>
> vmseq v0,v8,v9 (store mask value to v0)
> vmslt v1,v10,v11 (store mask value to v1)
> vmand v0,v0,v1
> vadd.vv ...v0.t (predicate mask should always be mask).
Ah, I see - that explains it well.
> >> ARM SVE would have a loop control mask and a separate mask
> >> for the if (cond[i]) which would be combined with a mask-and
> >> instruction to a third mask which is then used on the
> >> predicated instructions.
>
> Yeah, I know it. ARM SVE way is a more elegant way than RVV do.
> However, for RVV, we can't follow this flow.
> We don't have a "whilelo" instruction to generate loop control mask.
Yep. Similar for AVX512 where I have to use a vector compare. I'm
currently using
{ 0, 1, 2 ... } < { remaining_len, remaining_len, ... }
and careful updating of remaining_len (we know it will either
be adjusted by the full constant vector length or updated to zero).
> We only can do loop control with length generated by vsetvl.
> And we can only use "v0" to mask predicate vadd.vv, and mask value can only
> generated by comparison or mask logical instructions.
>
> >> PowerPC and s390x might be able to use WHILE_LEN as well (though
> >> they only have LEN variants of loads and stores) - of course
> >> only "simulating it". For the fixed-vector-length ISAs the
> >> predicated vector loop IMHO makes most sense for the epilogue to
> >> handle low-trip loops better.
>
> Yeah, I wonder how they do the flow control (if (cond[i])).
> For RVV, you can image I will need to add a pattern
> LEN_MASK_LOAD/LEN_MASK_STORE (length generated by WHILE_LEN and mask
> generated by comparison)
>
> I think we can CC IBM folks to see whether we can make WHILE_LEN works
> for both IBM and RVV ?
I've CCed them. Adding WHILE_LEN support to rs6000/s390x would be
mainly the "easy" way to get len-masked (epilog) loop support. I've
figured actually implementing WHILE_ULT for AVX512 in the backend
results in some code generation challenges so I'm going to play
(again) with open-coding it as outlined above in the vectorizer itself
so followup passes (mostly IVOPTs) can do a better job.
Richard.
>
> Thanks.
>
>
> [email protected]
>
> From: Richard Biener
> Date: 2023-04-12 16:42
> To: [email protected]
> CC: richard.sandiford; gcc-patches; jeffreyalaw
> Subject: Re: Re: [PATCH] VECT: Add WHILE_LEN pattern for decrement IV support
> for auto-vectorization
> On Wed, 12 Apr 2023, [email protected] wrote:
>
> > Thank you very much for reply.
> >
> > WHILE_LEN is the pattern that calculates the number of the elements of the
> > vector will be updated in each iteration.
> > For RVV, we use vsetvl instruction to calculate the number of the elements
> > of the vector.
> >
> > WHILE_ULT can not work for RVV since WHILE_ULT is generating mask to
> > predicate vector operation, but RVV do not
> > use mask to do the loop strip mining (RVV only use mask for control flow
> > inside the loop).
> >
> > Here is the example WHILE_ULT working in ARM SVE:
> > https://godbolt.org/z/jKsT8E1hP
> >
> > The first example is:
> > void foo (int32_t * __restrict a, int32_t * __restrict b, int n)
> > {
> > for (int i = 0; i < n; i++)
> > a[i] = a[i] + b[i];
> > }
> >
> > ARM SVE:
> > foo:
> > cmp w2, 0
> > ble .L1
> > mov x3, 0
> > cntw x4
> > whilelo p0.s, wzr, w2
> > .L3:
> > ld1w z1.s, p0/z, [x0, x3, lsl 2]
> > ld1w z0.s, p0/z, [x1, x3, lsl 2]
> > add z0.s, z0.s, z1.s
> > st1w z0.s, p0, [x0, x3, lsl 2]
> > add x3, x3, x4
> > whilelo p0.s, w3, w2
> > b.any .L3
> > .L1:
> > ret
> >
> > Here, whilelo will generate the mask according to w3 to w2.
> > So for example, if w3 = 0, and w2 = 3 (Suppose machine vector length > 3).
> > Then it will generate a mask with 0b111 mask to predicate loads and stores.
> >
> > For RVV, we can't do that since RVV doesn't have whilelo instructions to
> > generate predicate mask.
> > Also, we can't use mask as the predicate to do loop strip mining since RVV
> > only has 1 single mask
> > to handle flow control inside the loop.
> >
> > Instead, we use vsetvl to do the strip mining, so base on this, the same C
> > code, RVV ideal asm according RVV ISA should be:
> >
> > preheader:
> > a0 = n (the total number of the scalar should be calculated).
> > .....
> > .L3:
> > vsetvli a5,a0,e32,m1,ta,ma ====> WHILE_LEN pattern generate this
> > instruction, calculate the number of the elements should be updated
> > vle32.v v1,0(a4)
> > sub a0,a0,a5 ============> Decrement the induction
> > variable by the a5 (generated by WHILE_LEN)
> > ....
> >
> > vadd.vv....
> > vse32.v v1,0(a3)
> > add a4,a4,a2
> > add a3,a3,a2
> > bne a0,zero,.L3
> > .L1:
> > ret
> >
> > So you will see, if n = 3 like I said for ARM SVE (Suppose machine vector
> > length > 3), then vsetvli a5,a0,e32,m1,ta,ma will
> > generate a5 = 3, then the vle32.v/vadd.vv/vse32.v are all doing the
> > operation only on the element 0, element 1, element 2.
> >
> > Besides, WHILE_LEN is defined to make sure to be never overflow the input
> > operand which is "a0".
> > That means sub a0,a0,a5 will make a0 never underflow 0.
> >
> > I have tried to return Pmode in TARGET_VECTORIZE_GET_MASK_MODE
> > target hook and then use WHILE_ULT.
> >
> > But there are 2 issues:
> > One is that current GCC is doing the flow from 0-based until the
> > TEST_LIMIT. Wheras the optimal flow of RVV I showed above
> > is from "n" keep decreasing n until 0. Trying to fit the current flow of
> > GCC, RVV needs more instructions to do the loop strip mining.
> >
> > Second is that if we return a Pmode in TARGET_VECTORIZE_GET_MASK_MODE
> > which not only specify the dest mode for WHILE_ULT but also the mask mode
> > of flow control.
> > If we return Pmode which is used as the length for RVV. We can't use mask
> > mode like VNx2BI mode to do the flow control predicate.
> > This another example:
> > void foo2 (int32_t * __restrict a, int32_t * __restrict b, int32_t *
> > restrict cond, int n)
> > {
> > for (int i = 0; i < n; i++)
> > if (cond[i])
> > a[i] = a[i] + b[i];
> > }
> >
> > ARM SVE:
> > ld1w z0.s, p0/z, [x2, x4, lsl 2]
> > cmpne p0.s, p0/z, z0.s, #0
> > ld1w z0.s, p0/z, [x0, x4, lsl 2]
> > ld1w z1.s, p0/z, [x1, x4, lsl 2]
> > add z0.s, z0.s, z1.s
> > st1w z0.s, p0, [x0, x4, lsl 2]
> > add x4, x4, x5
> > whilelo p0.s, w4, w3
> > b.any .L8
> >
> > Here we can see ARM use mask mode for both loop strip minning and flow
> > control.
> >
> > Wheras, RVV use length generated by vsetvl (WHILE_LEN) to do the loop strip
> > minning and mask generated by comparison to do the flow control.
> >
> > So the ASM generated by my downstream LLVM/GCC:
> > .L3:
> > vsetvli a6,a3,e32,m1,ta,mu ==========> generate length to
> > predicate RVV operation.
> > vle32.v v0,(a2)
> > sub a3,a3,a6 ==========> decrease the induction variable
> > until 0.
> > vmsne.vi v0,v0,0 ==========> generate mask to predicate
> > RVV operation.
> > vle32.v v24,(a0),v0.t ===========> here using v0.t is the only
> > mask register to predicate RVV operation
> > vle32.v v25,(a1),v0.t
> > vadd.vv v24,v24,v25
> > vse32.v v24,(a0),v0.t
> > add a2,a2,a4
> > add a0,a0,a4
> > add a1,a1,a4
> > bne a3,zero,.L3
> > .L1:
> > ret
> >
> >
> > This is the how RVV works.
> > Feel free to comment if you have any questions.
>
> Thanks for the detailed explanation. Just to clarify - with RVV
> there's only a single mask register, v0.t, or did you want to
> say an instruction can only specify a single mask register?
> ARM SVE would have a loop control mask and a separate mask
> for the if (cond[i]) which would be combined with a mask-and
> instruction to a third mask which is then used on the
> predicated instructions.
>
> For AVX512 WHILE_ULT is a better match since we need a mask in the
> end (but WHILE_ULT isn't a very good match either, so I'm still
> working on masked loop support there).
>
> PowerPC and s390x might be able to use WHILE_LEN as well (though
> they only have LEN variants of loads and stores) - of course
> only "simulating it". For the fixed-vector-length ISAs the
> predicated vector loop IMHO makes most sense for the epilogue to
> handle low-trip loops better.
>
> Richard.
>
> > Thanks.
> >
> >
> > [email protected]
> >
> > From: Richard Biener
> > Date: 2023-04-12 15:00
> > To: Richard Sandiford
> > CC: [email protected]; gcc-patches; jeffreyalaw
> > Subject: Re: [PATCH] VECT: Add WHILE_LEN pattern for decrement IV support
> > for auto-vectorization
> > On Tue, 11 Apr 2023, Richard Sandiford wrote:
> >
> > > "[email protected]" <[email protected]> writes:
> > > > Hi, Richards.
> > > > Kindly Ping this patch.
> > > > This is the most important patch for RVV auto-vectorization support.
> > > > Bootstraped on X86 has passed.
> > >
> > > Can it wait for GCC 14? It doesn't seem like stage 4 material.
> > >
> > > Also, pinging after 5 days seems a bit soon. It's been a 4-day
> > > holiday weekend for much of Europe.
> >
> > Also can you explain why using WHILE_ULT is not possible? (I've
> > successfully - to some extent - done that for AVX512 for example)
> >
> > The patch lacks the description of what WHILE_LEN actually is.
> >
> > Richard.
> >
> > > Thanks,
> > > Richard
> > >
> > > > Feel free to comments.
> > > >
> > > > Thanks.
> > > >
> > > >
> > > > [email protected]
> > > >
> > > > From: juzhe.zhong
> > > > Date: 2023-04-07 09:47
> > > > To: gcc-patches
> > > > CC: richard.sandiford; rguenther; jeffreyalaw; Juzhe-Zhong
> > > > Subject: [PATCH] VECT: Add WHILE_LEN pattern for decrement IV support
> > > > for auto-vectorization
> > > > From: Juzhe-Zhong <[email protected]>
> > > >
> > > > This patch is to add WHILE_LEN pattern.
> > > > It's inspired by RVV ISA simple "vvaddint32.s" example:
> > > > https://github.com/riscv/riscv-v-spec/blob/master/example/vvaddint32.s
> > > >
> > > > More details are in "vect_set_loop_controls_by_while_len" implementation
> > > > and comments.
> > > >
> > > > Consider such following case:
> > > > #define N 16
> > > > int src[N];
> > > > int dest[N];
> > > >
> > > > void
> > > > foo (int n)
> > > > {
> > > > for (int i = 0; i < n; i++)
> > > > dest[i] = src[i];
> > > > }
> > > >
> > > > -march=rv64gcv -O3 --param riscv-autovec-preference=scalable
> > > > -fno-vect-cost-model -fno-tree-loop-distribute-patterns:
> > > >
> > > > foo:
> > > > ble a0,zero,.L1
> > > > lui a4,%hi(.LANCHOR0)
> > > > addi a4,a4,%lo(.LANCHOR0)
> > > > addi a3,a4,64
> > > > csrr a2,vlenb
> > > > .L3:
> > > > vsetvli a5,a0,e32,m1,ta,ma
> > > > vle32.v v1,0(a4)
> > > > sub a0,a0,a5
> > > > vse32.v v1,0(a3)
> > > > add a4,a4,a2
> > > > add a3,a3,a2
> > > > bne a0,zero,.L3
> > > > .L1:
> > > > ret
> > > >
> > > > gcc/ChangeLog:
> > > >
> > > > * doc/md.texi: Add WHILE_LEN support.
> > > > * internal-fn.cc (while_len_direct): Ditto.
> > > > (expand_while_len_optab_fn): Ditto.
> > > > (direct_while_len_optab_supported_p): Ditto.
> > > > * internal-fn.def (WHILE_LEN): Ditto.
> > > > * optabs.def (OPTAB_D): Ditto.
> > > > * tree-ssa-loop-manip.cc (create_iv): Ditto.
> > > > * tree-ssa-loop-manip.h (create_iv): Ditto.
> > > > * tree-vect-loop-manip.cc
> > > > (vect_set_loop_controls_by_while_len): Ditto.
> > > > (vect_set_loop_condition_partial_vectors): Ditto.
> > > > * tree-vect-loop.cc (vect_get_loop_len): Ditto.
> > > > * tree-vect-stmts.cc (vectorizable_store): Ditto.
> > > > (vectorizable_load): Ditto.
> > > > * tree-vectorizer.h (vect_get_loop_len): Ditto.
> > > >
> > > > ---
> > > > gcc/doc/md.texi | 14 +++
> > > > gcc/internal-fn.cc | 29 ++++++
> > > > gcc/internal-fn.def | 1 +
> > > > gcc/optabs.def | 1 +
> > > > gcc/tree-ssa-loop-manip.cc | 4 +-
> > > > gcc/tree-ssa-loop-manip.h | 2 +-
> > > > gcc/tree-vect-loop-manip.cc | 186 ++++++++++++++++++++++++++++++++++--
> > > > gcc/tree-vect-loop.cc | 35 +++++--
> > > > gcc/tree-vect-stmts.cc | 9 +-
> > > > gcc/tree-vectorizer.h | 4 +-
> > > > 10 files changed, 264 insertions(+), 21 deletions(-)
> > > >
> > > > diff --git a/gcc/doc/md.texi b/gcc/doc/md.texi
> > > > index 8e3113599fd..72178ab014c 100644
> > > > --- a/gcc/doc/md.texi
> > > > +++ b/gcc/doc/md.texi
> > > > @@ -4965,6 +4965,20 @@ for (i = 1; i < operand3; i++)
> > > > operand0[i] = operand0[i - 1] && (operand1 + i < operand2);
> > > > @end smallexample
> > > > +@cindex @code{while_len@var{m}@var{n}} instruction pattern
> > > > +@item @code{while_len@var{m}@var{n}}
> > > > +Set operand 0 to the number of active elements in vector will be
> > > > updated value.
> > > > +operand 1 is the total elements need to be updated value.
> > > > +operand 2 is the vectorization factor.
> > > > +The operation is equivalent to:
> > > > +
> > > > +@smallexample
> > > > +operand0 = MIN (operand1, operand2);
> > > > +operand2 can be const_poly_int or poly_int related to vector mode size.
> > > > +Some target like RISC-V has a standalone instruction to get MIN (n,
> > > > MODE SIZE) so
> > > > +that we can reduce a use of general purpose register.
> > > > +@end smallexample
> > > > +
> > > > @cindex @code{check_raw_ptrs@var{m}} instruction pattern
> > > > @item @samp{check_raw_ptrs@var{m}}
> > > > Check whether, given two pointers @var{a} and @var{b} and a length
> > > > @var{len},
> > > > diff --git a/gcc/internal-fn.cc b/gcc/internal-fn.cc
> > > > index 6e81dc05e0e..5f44def90d3 100644
> > > > --- a/gcc/internal-fn.cc
> > > > +++ b/gcc/internal-fn.cc
> > > > @@ -127,6 +127,7 @@ init_internal_fns ()
> > > > #define cond_binary_direct { 1, 1, true }
> > > > #define cond_ternary_direct { 1, 1, true }
> > > > #define while_direct { 0, 2, false }
> > > > +#define while_len_direct { 0, 0, false }
> > > > #define fold_extract_direct { 2, 2, false }
> > > > #define fold_left_direct { 1, 1, false }
> > > > #define mask_fold_left_direct { 1, 1, false }
> > > > @@ -3702,6 +3703,33 @@ expand_while_optab_fn (internal_fn, gcall *stmt,
> > > > convert_optab optab)
> > > > emit_move_insn (lhs_rtx, ops[0].value);
> > > > }
> > > > +/* Expand WHILE_LEN call STMT using optab OPTAB. */
> > > > +static void
> > > > +expand_while_len_optab_fn (internal_fn, gcall *stmt, convert_optab
> > > > optab)
> > > > +{
> > > > + expand_operand ops[3];
> > > > + tree rhs_type[2];
> > > > +
> > > > + tree lhs = gimple_call_lhs (stmt);
> > > > + tree lhs_type = TREE_TYPE (lhs);
> > > > + rtx lhs_rtx = expand_expr (lhs, NULL_RTX, VOIDmode, EXPAND_WRITE);
> > > > + create_output_operand (&ops[0], lhs_rtx, TYPE_MODE (lhs_type));
> > > > +
> > > > + for (unsigned int i = 0; i < gimple_call_num_args (stmt); ++i)
> > > > + {
> > > > + tree rhs = gimple_call_arg (stmt, i);
> > > > + rhs_type[i] = TREE_TYPE (rhs);
> > > > + rtx rhs_rtx = expand_normal (rhs);
> > > > + create_input_operand (&ops[i + 1], rhs_rtx, TYPE_MODE
> > > > (rhs_type[i]));
> > > > + }
> > > > +
> > > > + insn_code icode = direct_optab_handler (optab, TYPE_MODE
> > > > (rhs_type[0]));
> > > > +
> > > > + expand_insn (icode, 3, ops);
> > > > + if (!rtx_equal_p (lhs_rtx, ops[0].value))
> > > > + emit_move_insn (lhs_rtx, ops[0].value);
> > > > +}
> > > > +
> > > > /* Expand a call to a convert-like optab using the operands in STMT.
> > > > FN has a single output operand and NARGS input operands. */
> > > > @@ -3843,6 +3871,7 @@ multi_vector_optab_supported_p (convert_optab
> > > > optab, tree_pair types,
> > > > #define direct_scatter_store_optab_supported_p convert_optab_supported_p
> > > > #define direct_len_store_optab_supported_p direct_optab_supported_p
> > > > #define direct_while_optab_supported_p convert_optab_supported_p
> > > > +#define direct_while_len_optab_supported_p direct_optab_supported_p
> > > > #define direct_fold_extract_optab_supported_p direct_optab_supported_p
> > > > #define direct_fold_left_optab_supported_p direct_optab_supported_p
> > > > #define direct_mask_fold_left_optab_supported_p direct_optab_supported_p
> > > > diff --git a/gcc/internal-fn.def b/gcc/internal-fn.def
> > > > index 7fe742c2ae7..3a933abff5d 100644
> > > > --- a/gcc/internal-fn.def
> > > > +++ b/gcc/internal-fn.def
> > > > @@ -153,6 +153,7 @@ DEF_INTERNAL_OPTAB_FN (VEC_SET, 0, vec_set, vec_set)
> > > > DEF_INTERNAL_OPTAB_FN (LEN_STORE, 0, len_store, len_store)
> > > > DEF_INTERNAL_OPTAB_FN (WHILE_ULT, ECF_CONST | ECF_NOTHROW, while_ult,
> > > > while)
> > > > +DEF_INTERNAL_OPTAB_FN (WHILE_LEN, ECF_CONST | ECF_NOTHROW, while_len,
> > > > while_len)
> > > > DEF_INTERNAL_OPTAB_FN (CHECK_RAW_PTRS, ECF_CONST | ECF_NOTHROW,
> > > > check_raw_ptrs, check_ptrs)
> > > > DEF_INTERNAL_OPTAB_FN (CHECK_WAR_PTRS, ECF_CONST | ECF_NOTHROW,
> > > > diff --git a/gcc/optabs.def b/gcc/optabs.def
> > > > index 695f5911b30..f5938bd2c24 100644
> > > > --- a/gcc/optabs.def
> > > > +++ b/gcc/optabs.def
> > > > @@ -476,3 +476,4 @@ OPTAB_DC (vec_series_optab, "vec_series$a",
> > > > VEC_SERIES)
> > > > OPTAB_D (vec_shl_insert_optab, "vec_shl_insert_$a")
> > > > OPTAB_D (len_load_optab, "len_load_$a")
> > > > OPTAB_D (len_store_optab, "len_store_$a")
> > > > +OPTAB_D (while_len_optab, "while_len$a")
> > > > diff --git a/gcc/tree-ssa-loop-manip.cc b/gcc/tree-ssa-loop-manip.cc
> > > > index 09acc1c94cc..cdbf280e249 100644
> > > > --- a/gcc/tree-ssa-loop-manip.cc
> > > > +++ b/gcc/tree-ssa-loop-manip.cc
> > > > @@ -59,14 +59,14 @@ static bitmap_obstack loop_renamer_obstack;
> > > > void
> > > > create_iv (tree base, tree step, tree var, class loop *loop,
> > > > gimple_stmt_iterator *incr_pos, bool after,
> > > > - tree *var_before, tree *var_after)
> > > > + tree *var_before, tree *var_after, enum tree_code code)
> > > > {
> > > > gassign *stmt;
> > > > gphi *phi;
> > > > tree initial, step1;
> > > > gimple_seq stmts;
> > > > tree vb, va;
> > > > - enum tree_code incr_op = PLUS_EXPR;
> > > > + enum tree_code incr_op = code;
> > > > edge pe = loop_preheader_edge (loop);
> > > > if (var != NULL_TREE)
> > > > diff --git a/gcc/tree-ssa-loop-manip.h b/gcc/tree-ssa-loop-manip.h
> > > > index d49273a3987..da755320a3a 100644
> > > > --- a/gcc/tree-ssa-loop-manip.h
> > > > +++ b/gcc/tree-ssa-loop-manip.h
> > > > @@ -23,7 +23,7 @@ along with GCC; see the file COPYING3. If not see
> > > > typedef void (*transform_callback)(class loop *, void *);
> > > > extern void create_iv (tree, tree, tree, class loop *,
> > > > gimple_stmt_iterator *,
> > > > - bool, tree *, tree *);
> > > > + bool, tree *, tree *, enum tree_code = PLUS_EXPR);
> > > > extern void rewrite_into_loop_closed_ssa (bitmap, unsigned);
> > > > extern void verify_loop_closed_ssa (bool, class loop * = NULL);
> > > > diff --git a/gcc/tree-vect-loop-manip.cc b/gcc/tree-vect-loop-manip.cc
> > > > index f60fa50e8f4..f3cd6c51d2e 100644
> > > > --- a/gcc/tree-vect-loop-manip.cc
> > > > +++ b/gcc/tree-vect-loop-manip.cc
> > > > @@ -682,6 +682,173 @@ vect_set_loop_controls_directly (class loop
> > > > *loop, loop_vec_info loop_vinfo,
> > > > return next_ctrl;
> > > > }
> > > > +/* Helper for vect_set_loop_condition_partial_vectors. Generate
> > > > definitions
> > > > + for all the rgroup controls in RGC and return a control that is
> > > > nonzero
> > > > + when the loop needs to iterate. Add any new preheader statements to
> > > > + PREHEADER_SEQ. Use LOOP_COND_GSI to insert code before the exit
> > > > gcond.
> > > > +
> > > > + RGC belongs to loop LOOP. The loop originally iterated NITERS
> > > > + times and has been vectorized according to LOOP_VINFO.
> > > > +
> > > > + Unlike vect_set_loop_controls_directly which is iterating from
> > > > 0-based IV
> > > > + to TEST_LIMIT - bias.
> > > > +
> > > > + In vect_set_loop_controls_by_while_len, we are iterating from start
> > > > at
> > > > + IV = TEST_LIMIT - bias and keep subtract IV by the length
> > > > calculated by
> > > > + IFN_WHILE_LEN pattern.
> > > > +
> > > > + Note: the cost of the code generated by this function is modeled
> > > > + by vect_estimate_min_profitable_iters, so changes here may need
> > > > + corresponding changes there.
> > > > +
> > > > + 1. Single rgroup, the Gimple IR should be:
> > > > +
> > > > + <bb 3>
> > > > + _19 = (unsigned long) n_5(D);
> > > > + ...
> > > > +
> > > > + <bb 4>:
> > > > + ...
> > > > + # ivtmp_20 = PHI <ivtmp_21(4), _19(3)>
> > > > + ...
> > > > + _22 = .WHILE_LEN (ivtmp_20, vf);
> > > > + ...
> > > > + vector statement (use _22);
> > > > + ...
> > > > + ivtmp_21 = ivtmp_20 - _22;
> > > > + ...
> > > > + if (ivtmp_21 != 0)
> > > > + goto <bb 4>; [75.00%]
> > > > + else
> > > > + goto <bb 5>; [25.00%]
> > > > +
> > > > + <bb 5>
> > > > + return;
> > > > +
> > > > + Note: IFN_WHILE_LEN will guarantee "ivtmp_21 = ivtmp_20 - _22" never
> > > > + underflow 0.
> > > > +
> > > > + 2. Multiple rgroup, the Gimple IR should be:
> > > > +
> > > > + <bb 3>
> > > > + _70 = (unsigned long) bnd.7_52;
> > > > + _71 = _70 * 2;
> > > > + _72 = MAX_EXPR <_71, 4>;
> > > > + _73 = _72 + 18446744073709551612;
> > > > + ...
> > > > +
> > > > + <bb 4>:
> > > > + ...
> > > > + # ivtmp_74 = PHI <ivtmp_75(6), _73(12)>
> > > > + # ivtmp_77 = PHI <ivtmp_78(6), _71(12)>
> > > > + _76 = .WHILE_LEN (ivtmp_74, vf * nitems_per_ctrl);
> > > > + _79 = .WHILE_LEN (ivtmp_77, vf * nitems_per_ctrl);
> > > > + ...
> > > > + vector statement (use _79);
> > > > + ...
> > > > + vector statement (use _76);
> > > > + ...
> > > > + _65 = _79 / 2;
> > > > + vector statement (use _65);
> > > > + ...
> > > > + _68 = _76 / 2;
> > > > + vector statement (use _68);
> > > > + ...
> > > > + ivtmp_78 = ivtmp_77 - _79;
> > > > + ivtmp_75 = ivtmp_74 - _76;
> > > > + ...
> > > > + if (ivtmp_78 != 0)
> > > > + goto <bb 4>; [75.00%]
> > > > + else
> > > > + goto <bb 5>; [25.00%]
> > > > +
> > > > + <bb 5>
> > > > + return;
> > > > +
> > > > +*/
> > > > +
> > > > +static tree
> > > > +vect_set_loop_controls_by_while_len (class loop *loop, loop_vec_info
> > > > loop_vinfo,
> > > > + gimple_seq *preheader_seq,
> > > > + gimple_seq *header_seq,
> > > > + rgroup_controls *rgc, tree niters)
> > > > +{
> > > > + tree compare_type = LOOP_VINFO_RGROUP_COMPARE_TYPE (loop_vinfo);
> > > > + tree iv_type = LOOP_VINFO_RGROUP_IV_TYPE (loop_vinfo);
> > > > + /* We are not allowing masked approach in WHILE_LEN. */
> > > > + gcc_assert (!LOOP_VINFO_FULLY_MASKED_P (loop_vinfo));
> > > > +
> > > > + tree ctrl_type = rgc->type;
> > > > + unsigned int nitems_per_iter = rgc->max_nscalars_per_iter *
> > > > rgc->factor;
> > > > + poly_uint64 nitems_per_ctrl = TYPE_VECTOR_SUBPARTS (ctrl_type) *
> > > > rgc->factor;
> > > > + poly_uint64 vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo);
> > > > +
> > > > + /* Calculate the maximum number of item values that the rgroup
> > > > + handles in total, the number that it handles for each iteration
> > > > + of the vector loop. */
> > > > + tree nitems_total = niters;
> > > > + if (nitems_per_iter != 1)
> > > > + {
> > > > + /* We checked before setting LOOP_VINFO_USING_PARTIAL_VECTORS_P
> > > > that
> > > > + these multiplications don't overflow. */
> > > > + tree compare_factor = build_int_cst (compare_type,
> > > > nitems_per_iter);
> > > > + nitems_total = gimple_build (preheader_seq, MULT_EXPR,
> > > > compare_type,
> > > > + nitems_total, compare_factor);
> > > > + }
> > > > +
> > > > + /* Convert the comparison value to the IV type (either a no-op or
> > > > + a promotion). */
> > > > + nitems_total = gimple_convert (preheader_seq, iv_type, nitems_total);
> > > > +
> > > > + /* Create an induction variable that counts the number of items
> > > > + processed. */
> > > > + tree index_before_incr, index_after_incr;
> > > > + gimple_stmt_iterator incr_gsi;
> > > > + bool insert_after;
> > > > + standard_iv_increment_position (loop, &incr_gsi, &insert_after);
> > > > +
> > > > + /* Test the decremented IV, which will never underflow 0 since we
> > > > have
> > > > + IFN_WHILE_LEN to gurantee that. */
> > > > + tree test_limit = nitems_total;
> > > > +
> > > > + /* Provide a definition of each control in the group. */
> > > > + tree ctrl;
> > > > + unsigned int i;
> > > > + FOR_EACH_VEC_ELT_REVERSE (rgc->controls, i, ctrl)
> > > > + {
> > > > + /* Previous controls will cover BIAS items. This control covers
> > > > the
> > > > + next batch. */
> > > > + poly_uint64 bias = nitems_per_ctrl * i;
> > > > + tree bias_tree = build_int_cst (iv_type, bias);
> > > > +
> > > > + /* Rather than have a new IV that starts at TEST_LIMIT and goes
> > > > down to
> > > > + BIAS, prefer to use the same TEST_LIMIT - BIAS based IV for each
> > > > + control and adjust the bound down by BIAS. */
> > > > + tree this_test_limit = test_limit;
> > > > + if (i != 0)
> > > > + {
> > > > + this_test_limit = gimple_build (preheader_seq, MAX_EXPR, iv_type,
> > > > + this_test_limit, bias_tree);
> > > > + this_test_limit = gimple_build (preheader_seq, MINUS_EXPR, iv_type,
> > > > + this_test_limit, bias_tree);
> > > > + }
> > > > +
> > > > + /* Create decrement IV. */
> > > > + create_iv (this_test_limit, ctrl, NULL_TREE, loop, &incr_gsi,
> > > > + insert_after, &index_before_incr, &index_after_incr,
> > > > + MINUS_EXPR);
> > > > +
> > > > + poly_uint64 final_vf = vf * nitems_per_iter;
> > > > + tree vf_step = build_int_cst (iv_type, final_vf);
> > > > + tree res_len = gimple_build (header_seq, IFN_WHILE_LEN, iv_type,
> > > > + index_before_incr, vf_step);
> > > > + gassign *assign = gimple_build_assign (ctrl, res_len);
> > > > + gimple_seq_add_stmt (header_seq, assign);
> > > > + }
> > > > +
> > > > + return index_after_incr;
> > > > +}
> > > > +
> > > > /* Set up the iteration condition and rgroup controls for LOOP, given
> > > > that LOOP_VINFO_USING_PARTIAL_VECTORS_P is true for the vectorized
> > > > loop. LOOP_VINFO describes the vectorization of LOOP. NITERS is
> > > > @@ -703,6 +870,7 @@ vect_set_loop_condition_partial_vectors (class loop
> > > > *loop,
> > > > bool use_masks_p = LOOP_VINFO_FULLY_MASKED_P (loop_vinfo);
> > > > tree compare_type = LOOP_VINFO_RGROUP_COMPARE_TYPE (loop_vinfo);
> > > > + tree iv_type = LOOP_VINFO_RGROUP_IV_TYPE (loop_vinfo);
> > > > unsigned int compare_precision = TYPE_PRECISION (compare_type);
> > > > tree orig_niters = niters;
> > > > @@ -757,12 +925,18 @@ vect_set_loop_condition_partial_vectors (class
> > > > loop *loop,
> > > > bool might_wrap_p = vect_rgroup_iv_might_wrap_p (loop_vinfo, rgc);
> > > > /* Set up all controls for this group. */
> > > > - test_ctrl = vect_set_loop_controls_directly (loop, loop_vinfo,
> > > > - &preheader_seq,
> > > > - &header_seq,
> > > > - loop_cond_gsi, rgc,
> > > > - niters, niters_skip,
> > > > - might_wrap_p);
> > > > + if (direct_internal_fn_supported_p (IFN_WHILE_LEN, iv_type,
> > > > + OPTIMIZE_FOR_SPEED))
> > > > + test_ctrl
> > > > + = vect_set_loop_controls_by_while_len (loop, loop_vinfo,
> > > > + &preheader_seq, &header_seq,
> > > > + rgc, niters);
> > > > + else
> > > > + test_ctrl
> > > > + = vect_set_loop_controls_directly (loop, loop_vinfo,
> > > > &preheader_seq,
> > > > + &header_seq, loop_cond_gsi, rgc,
> > > > + niters, niters_skip,
> > > > + might_wrap_p);
> > > > }
> > > > /* Emit all accumulated statements. */
> > > > diff --git a/gcc/tree-vect-loop.cc b/gcc/tree-vect-loop.cc
> > > > index 1ba9f18d73e..5bffd9a6322 100644
> > > > --- a/gcc/tree-vect-loop.cc
> > > > +++ b/gcc/tree-vect-loop.cc
> > > > @@ -10360,12 +10360,14 @@ vect_record_loop_len (loop_vec_info
> > > > loop_vinfo, vec_loop_lens *lens,
> > > > rgroup that operates on NVECTORS vectors, where 0 <= INDEX <
> > > > NVECTORS. */
> > > > tree
> > > > -vect_get_loop_len (loop_vec_info loop_vinfo, vec_loop_lens *lens,
> > > > - unsigned int nvectors, unsigned int index)
> > > > +vect_get_loop_len (gimple_stmt_iterator *gsi, loop_vec_info loop_vinfo,
> > > > + vec_loop_lens *lens, unsigned int nvectors, tree vectype,
> > > > + unsigned int index)
> > > > {
> > > > rgroup_controls *rgl = &(*lens)[nvectors - 1];
> > > > - bool use_bias_adjusted_len =
> > > > - LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS (loop_vinfo) != 0;
> > > > + bool use_bias_adjusted_len
> > > > + = LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS (loop_vinfo) != 0;
> > > > + tree iv_type = LOOP_VINFO_RGROUP_IV_TYPE (loop_vinfo);
> > > > /* Populate the rgroup's len array, if this is the first time we've
> > > > used it. */
> > > > @@ -10386,8 +10388,8 @@ vect_get_loop_len (loop_vec_info loop_vinfo,
> > > > vec_loop_lens *lens,
> > > > if (use_bias_adjusted_len)
> > > > {
> > > > gcc_assert (i == 0);
> > > > - tree adjusted_len =
> > > > - make_temp_ssa_name (len_type, NULL, "adjusted_loop_len");
> > > > + tree adjusted_len
> > > > + = make_temp_ssa_name (len_type, NULL, "adjusted_loop_len");
> > > > SSA_NAME_DEF_STMT (adjusted_len) = gimple_build_nop ();
> > > > rgl->bias_adjusted_ctrl = adjusted_len;
> > > > }
> > > > @@ -10396,6 +10398,27 @@ vect_get_loop_len (loop_vec_info loop_vinfo,
> > > > vec_loop_lens *lens,
> > > > if (use_bias_adjusted_len)
> > > > return rgl->bias_adjusted_ctrl;
> > > > + else if (direct_internal_fn_supported_p (IFN_WHILE_LEN, iv_type,
> > > > + OPTIMIZE_FOR_SPEED))
> > > > + {
> > > > + tree loop_len = rgl->controls[index];
> > > > + poly_int64 nunits1 = TYPE_VECTOR_SUBPARTS (rgl->type);
> > > > + poly_int64 nunits2 = TYPE_VECTOR_SUBPARTS (vectype);
> > > > + if (maybe_ne (nunits1, nunits2))
> > > > + {
> > > > + /* A loop len for data type X can be reused for data type Y
> > > > + if X has N times more elements than Y and if Y's elements
> > > > + are N times bigger than X's. */
> > > > + gcc_assert (multiple_p (nunits1, nunits2));
> > > > + unsigned int factor = exact_div (nunits1, nunits2).to_constant ();
> > > > + gimple_seq seq = NULL;
> > > > + loop_len = gimple_build (&seq, RDIV_EXPR, iv_type, loop_len,
> > > > + build_int_cst (iv_type, factor));
> > > > + if (seq)
> > > > + gsi_insert_seq_before (gsi, seq, GSI_SAME_STMT);
> > > > + }
> > > > + return loop_len;
> > > > + }
> > > > else
> > > > return rgl->controls[index];
> > > > }
> > > > diff --git a/gcc/tree-vect-stmts.cc b/gcc/tree-vect-stmts.cc
> > > > index efa2d0daa52..708c8a1d806 100644
> > > > --- a/gcc/tree-vect-stmts.cc
> > > > +++ b/gcc/tree-vect-stmts.cc
> > > > @@ -8653,8 +8653,9 @@ vectorizable_store (vec_info *vinfo,
> > > > else if (loop_lens)
> > > > {
> > > > tree final_len
> > > > - = vect_get_loop_len (loop_vinfo, loop_lens,
> > > > - vec_num * ncopies, vec_num * j + i);
> > > > + = vect_get_loop_len (gsi, loop_vinfo, loop_lens,
> > > > + vec_num * ncopies, vectype,
> > > > + vec_num * j + i);
> > > > tree ptr = build_int_cst (ref_type, align * BITS_PER_UNIT);
> > > > machine_mode vmode = TYPE_MODE (vectype);
> > > > opt_machine_mode new_ovmode
> > > > @@ -10009,8 +10010,8 @@ vectorizable_load (vec_info *vinfo,
> > > > else if (loop_lens && memory_access_type != VMAT_INVARIANT)
> > > > {
> > > > tree final_len
> > > > - = vect_get_loop_len (loop_vinfo, loop_lens,
> > > > - vec_num * ncopies,
> > > > + = vect_get_loop_len (gsi, loop_vinfo, loop_lens,
> > > > + vec_num * ncopies, vectype,
> > > > vec_num * j + i);
> > > > tree ptr = build_int_cst (ref_type,
> > > > align * BITS_PER_UNIT);
> > > > diff --git a/gcc/tree-vectorizer.h b/gcc/tree-vectorizer.h
> > > > index 9cf2fb23fe3..e5cf38caf4b 100644
> > > > --- a/gcc/tree-vectorizer.h
> > > > +++ b/gcc/tree-vectorizer.h
> > > > @@ -2293,8 +2293,8 @@ extern tree vect_get_loop_mask
> > > > (gimple_stmt_iterator *, vec_loop_masks *,
> > > > unsigned int, tree, unsigned int);
> > > > extern void vect_record_loop_len (loop_vec_info, vec_loop_lens *,
> > > > unsigned int,
> > > > tree, unsigned int);
> > > > -extern tree vect_get_loop_len (loop_vec_info, vec_loop_lens *,
> > > > unsigned int,
> > > > - unsigned int);
> > > > +extern tree vect_get_loop_len (gimple_stmt_iterator *, loop_vec_info,
> > > > + vec_loop_lens *, unsigned int, tree, unsigned int);
> > > > extern gimple_seq vect_gen_len (tree, tree, tree, tree);
> > > > extern stmt_vec_info info_for_reduction (vec_info *, stmt_vec_info);
> > > > extern bool reduction_fn_for_scalar_code (code_helper, internal_fn *);
> > >
> >
> >
>
>
--
Richard Biener <[email protected]>
SUSE Software Solutions Germany GmbH, Frankenstrasse 146, 90461 Nuernberg,
Germany; GF: Ivo Totev, Andrew Myers, Andrew McDonald, Boudien Moerman;
HRB 36809 (AG Nuernberg)