The following applies the earlier fix for conditional SIMD clone
calls to the loop masked case.

Bootstrapped and tested on x86_64-unknown-linux-gnu, pushed.

        * tree-vect-stmts.cc (vectorizable_simd_clone_call): Handle
        AVX512 masking for loop masked SIMD clone call.
---
 gcc/tree-vect-stmts.cc | 5 ++---
 1 file changed, 2 insertions(+), 3 deletions(-)

diff --git a/gcc/tree-vect-stmts.cc b/gcc/tree-vect-stmts.cc
index 42b1bf11668..189847b8fe0 100644
--- a/gcc/tree-vect-stmts.cc
+++ b/gcc/tree-vect-stmts.cc
@@ -4835,9 +4835,8 @@ vectorizable_simd_clone_call (vec_info *vinfo, 
stmt_vec_info stmt_info,
 
          tree masktype = bestn->simdclone->args[mask_i].vector_type;
          if (SCALAR_INT_MODE_P (bestn->simdclone->mask_mode))
-           /* Guess the number of lanes represented by masktype.  */
            callee_nelements = exact_div (bestn->simdclone->simdlen,
-                                         bestn->simdclone->nargs - nargs);
+                                         
bestn->simdclone->args[i].linear_step);
          else
            callee_nelements = TYPE_VECTOR_SUBPARTS (masktype);
          o = vector_unroll_factor (nunits, callee_nelements);
@@ -4847,7 +4846,7 @@ vectorizable_simd_clone_call (vec_info *vinfo, 
stmt_vec_info stmt_info,
                {
                  vec_loop_masks *loop_masks = &LOOP_VINFO_MASKS (loop_vinfo);
                  mask = vect_get_loop_mask (loop_vinfo, gsi, loop_masks,
-                                            ncopies, masktype, j);
+                                            ncopies_in, vectype, j);
                }
              else
                mask = vect_build_all_ones_mask (vinfo, stmt_info, masktype);
-- 
2.51.0

Reply via email to