GCC Middle and Back End API Reference
tree-vect-loop.cc File Reference
#include "config.h"
#include "system.h"
#include "coretypes.h"
#include "backend.h"
#include "target.h"
#include "rtl.h"
#include "tree.h"
#include "gimple.h"
#include "cfghooks.h"
#include "tree-pass.h"
#include "ssa.h"
#include "optabs-tree.h"
#include "memmodel.h"
#include "optabs.h"
#include "diagnostic-core.h"
#include "fold-const.h"
#include "stor-layout.h"
#include "cfganal.h"
#include "gimplify.h"
#include "gimple-iterator.h"
#include "gimplify-me.h"
#include "tree-ssa-loop-ivopts.h"
#include "tree-ssa-loop-manip.h"
#include "tree-ssa-loop-niter.h"
#include "tree-ssa-loop.h"
#include "cfgloop.h"
#include "tree-scalar-evolution.h"
#include "tree-vectorizer.h"
#include "gimple-fold.h"
#include "cgraph.h"
#include "tree-cfg.h"
#include "tree-if-conv.h"
#include "internal-fn.h"
#include "tree-vector-builder.h"
#include "vec-perm-indices.h"
#include "tree-eh.h"
#include "case-cfn-macros.h"
#include "langhooks.h"
Include dependency graph for tree-vect-loop.cc:

Macros

#define INCLUDE_ALGORITHM
 

Functions

static void vect_estimate_min_profitable_iters (loop_vec_info, int *, int *, unsigned *)
 
static stmt_vec_info vect_is_simple_reduction (loop_vec_info, stmt_vec_info, bool *, bool *, bool)
 
static opt_result vect_determine_vf_for_stmt_1 (vec_info *vinfo, stmt_vec_info stmt_info, bool vectype_maybe_set_p, poly_uint64 *vf)
 
static opt_result vect_determine_vf_for_stmt (vec_info *vinfo, stmt_vec_info stmt_info, poly_uint64 *vf)
 
static opt_result vect_determine_vectorization_factor (loop_vec_info loop_vinfo)
 
static bool vect_is_simple_iv_evolution (unsigned loop_nb, tree access_fn, tree *init, tree *step)
 
static bool vect_is_nonlinear_iv_evolution (class loop *loop, stmt_vec_info stmt_info, gphi *loop_phi_node, tree *init, tree *step)
 
static bool vect_inner_phi_in_double_reduction_p (loop_vec_info loop_vinfo, gphi *phi)
 
static bool vect_phi_first_order_recurrence_p (loop_vec_info loop_vinfo, class loop *loop, gphi *phi)
 
static void vect_analyze_scalar_cycles_1 (loop_vec_info loop_vinfo, class loop *loop, bool slp)
 
static void vect_analyze_scalar_cycles (loop_vec_info loop_vinfo, bool slp)
 
static void vect_fixup_reduc_chain (stmt_vec_info stmt_info)
 
static void vect_fixup_scalar_cycles_with_patterns (loop_vec_info loop_vinfo)
 
static vec< gcond * > vect_get_loop_niters (class loop *loop, const_edge main_exit, tree *assumptions, tree *number_of_iterations, tree *number_of_iterationsm1)
 
edge vec_init_loop_exit_info (class loop *loop)
 
static bool bb_in_loop_p (const_basic_block bb, const void *data)
 
void release_vec_loop_controls (vec< rgroup_controls > *controls)
 
tree cse_and_gimplify_to_preheader (loop_vec_info loop_vinfo, tree expr)
 
static bool can_produce_all_loop_masks_p (loop_vec_info loop_vinfo, tree cmp_type)
 
static unsigned int vect_get_max_nscalars_per_iter (loop_vec_info loop_vinfo)
 
static unsigned vect_min_prec_for_max_niters (loop_vec_info loop_vinfo, unsigned int factor)
 
static bool vect_need_peeling_or_partial_vectors_p (loop_vec_info loop_vinfo)
 
static bool vect_verify_full_masking (loop_vec_info loop_vinfo)
 
static bool vect_verify_full_masking_avx512 (loop_vec_info loop_vinfo)
 
static bool vect_verify_loop_lens (loop_vec_info loop_vinfo)
 
static void vect_compute_single_scalar_iteration_cost (loop_vec_info loop_vinfo)
 
opt_result vect_analyze_loop_form (class loop *loop, vect_loop_form_info *info)
 
loop_vec_info vect_create_loop_vinfo (class loop *loop, vec_info_shared *shared, const vect_loop_form_info *info, loop_vec_info main_loop_info)
 
static void vect_update_vf_for_slp (loop_vec_info loop_vinfo)
 
static bool vect_active_double_reduction_p (stmt_vec_info stmt_info)
 
static opt_result vect_analyze_loop_operations (loop_vec_info loop_vinfo)
 
static bool vect_known_niters_smaller_than_vf (loop_vec_info loop_vinfo)
 
static int vect_analyze_loop_costing (loop_vec_info loop_vinfo, unsigned *suggested_unroll_factor)
 
static opt_result vect_get_datarefs_in_loop (loop_p loop, basic_block *bbs, vec< data_reference_p > *datarefs, unsigned int *n_stmts)
 
static void vect_dissolve_slp_only_groups (loop_vec_info loop_vinfo)
 
opt_result vect_determine_partial_vectors_and_peeling (loop_vec_info loop_vinfo)
 
static opt_result vect_analyze_loop_2 (loop_vec_info loop_vinfo, bool &fatal, unsigned *suggested_unroll_factor, bool &slp_done_for_suggested_uf)
 
static bool vect_better_loop_vinfo_p (loop_vec_info new_loop_vinfo, loop_vec_info old_loop_vinfo)
 
static bool vect_joust_loop_vinfos (loop_vec_info new_loop_vinfo, loop_vec_info old_loop_vinfo)
 
static opt_loop_vec_info vect_analyze_loop_1 (class loop *loop, vec_info_shared *shared, const vect_loop_form_info *loop_form_info, loop_vec_info main_loop_vinfo, const vector_modes &vector_modes, unsigned &mode_i, machine_mode &autodetected_vector_mode, bool &fatal)
 
opt_loop_vec_info vect_analyze_loop (class loop *loop, vec_info_shared *shared)
 
static bool fold_left_reduction_fn (code_helper code, internal_fn *reduc_fn)
 
bool reduction_fn_for_scalar_code (code_helper code, internal_fn *reduc_fn)
 
tree neutral_op_for_reduction (tree scalar_type, code_helper code, tree initial_value, bool as_initial)
 
static void report_vect_op (dump_flags_t msg_type, gimple *stmt, const char *msg)
 
bool needs_fold_left_reduction_p (tree type, code_helper code)
 
static bool check_reduction_path (dump_user_location_t loc, loop_p loop, gphi *phi, tree loop_arg, code_helper *code, vec< std::pair< ssa_op_iter, use_operand_p > > &path)
 
bool check_reduction_path (dump_user_location_t loc, loop_p loop, gphi *phi, tree loop_arg, enum tree_code code)
 
static int vect_get_peel_iters_epilogue (loop_vec_info loop_vinfo, int peel_iters_prologue)
 
int vect_get_known_peeling_cost (loop_vec_info loop_vinfo, int peel_iters_prologue, int *peel_iters_epilogue, stmt_vector_for_cost *scalar_cost_vec, stmt_vector_for_cost *prologue_cost_vec, stmt_vector_for_cost *epilogue_cost_vec)
 
static void calc_vec_perm_mask_for_shift (unsigned int offset, unsigned int nelt, vec_perm_builder *sel)
 
static bool have_whole_vector_shift (machine_mode mode)
 
static bool vect_is_emulated_mixed_dot_prod (loop_vec_info loop_vinfo, stmt_vec_info stmt_info)
 
static void vect_model_reduction_cost (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, internal_fn reduc_fn, vect_reduction_type reduction_type, int ncopies, stmt_vector_for_cost *cost_vec)
 
static void vect_emit_reduction_init_stmts (loop_vec_info loop_vinfo, stmt_vec_info reduc_info, gimple *seq)
 
static tree get_initial_def_for_reduction (loop_vec_info loop_vinfo, stmt_vec_info reduc_info, tree init_val, tree neutral_op)
 
static void get_initial_defs_for_reduction (loop_vec_info loop_vinfo, stmt_vec_info reduc_info, vec< tree > *vec_oprnds, unsigned int number_of_vectors, unsigned int group_size, tree neutral_op)
 
stmt_vec_info info_for_reduction (vec_info *vinfo, stmt_vec_info stmt_info)
 
static bool vect_find_reusable_accumulator (loop_vec_info loop_vinfo, stmt_vec_info reduc_info)
 
static tree vect_create_partial_epilog (tree vec_def, tree vectype, code_helper code, gimple_seq *seq)
 
static void vect_create_epilog_for_reduction (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, slp_tree slp_node, slp_instance slp_node_instance, edge loop_exit)
 
static tree merge_with_identity (gimple_stmt_iterator *gsi, tree mask, tree vectype, tree vec, tree identity)
 
static tree vect_expand_fold_left (gimple_stmt_iterator *gsi, tree scalar_dest, tree_code code, tree lhs, tree vector_rhs, tree mask)
 
static internal_fn get_masked_reduction_fn (internal_fn reduc_fn, tree vectype_in)
 
static bool vectorize_fold_left_reduction (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, gimple_stmt_iterator *gsi, gimple **vec_stmt, slp_tree slp_node, gimple *reduc_def_stmt, code_helper code, internal_fn reduc_fn, tree *ops, int num_ops, tree vectype_in, int reduc_index, vec_loop_masks *masks, vec_loop_lens *lens)
 
static bool is_nonwrapping_integer_induction (stmt_vec_info stmt_vinfo, class loop *loop)
 
static bool use_mask_by_cond_expr_p (code_helper code, internal_fn cond_fn, tree vectype_in)
 
static void build_vect_cond_expr (code_helper code, tree vop[3], tree mask, gimple_stmt_iterator *gsi)
 
bool vectorizable_reduction (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, slp_tree slp_node, slp_instance slp_node_instance, stmt_vector_for_cost *cost_vec)
 
static gassignvect_emulate_mixed_dot_prod (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, gimple_stmt_iterator *gsi, tree vec_dest, tree vop[3])
 
bool vect_transform_reduction (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, gimple_stmt_iterator *gsi, gimple **vec_stmt, slp_tree slp_node)
 
bool vect_transform_cycle_phi (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, gimple **vec_stmt, slp_tree slp_node, slp_instance slp_node_instance)
 
bool vectorizable_lc_phi (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, gimple **vec_stmt, slp_tree slp_node)
 
bool vectorizable_phi (vec_info *, stmt_vec_info stmt_info, gimple **vec_stmt, slp_tree slp_node, stmt_vector_for_cost *cost_vec)
 
bool vectorizable_recurr (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, gimple **vec_stmt, slp_tree slp_node, stmt_vector_for_cost *cost_vec)
 
bool vect_emulated_vector_p (tree vectype)
 
bool vect_can_vectorize_without_simd_p (tree_code code)
 
bool vect_can_vectorize_without_simd_p (code_helper code)
 
static tree vect_create_nonlinear_iv_init (gimple_seq *stmts, tree init_expr, tree step_expr, poly_uint64 nunits, tree vectype, enum vect_induction_op_type induction_type)
 
tree vect_peel_nonlinear_iv_init (gimple_seq *stmts, tree init_expr, tree skip_niters, tree step_expr, enum vect_induction_op_type induction_type)
 
static tree vect_create_nonlinear_iv_step (gimple_seq *stmts, tree step_expr, poly_uint64 vf, enum vect_induction_op_type induction_type)
 
static tree vect_create_nonlinear_iv_vec_step (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, tree new_name, tree vectype, enum vect_induction_op_type induction_type)
 
static tree vect_update_nonlinear_iv (gimple_seq *stmts, tree vectype, tree induc_def, tree vec_step, enum vect_induction_op_type induction_type)
 
static bool vectorizable_nonlinear_induction (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, gimple **vec_stmt, slp_tree slp_node, stmt_vector_for_cost *cost_vec)
 
bool vectorizable_induction (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, gimple **vec_stmt, slp_tree slp_node, stmt_vector_for_cost *cost_vec)
 
static tree vectorizable_live_operation_1 (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, basic_block exit_bb, tree vectype, int ncopies, slp_tree slp_node, tree bitsize, tree bitstart, tree vec_lhs, tree lhs_type, gimple_stmt_iterator *exit_gsi)
 
bool vectorizable_live_operation (vec_info *vinfo, stmt_vec_info stmt_info, slp_tree slp_node, slp_instance slp_node_instance, int slp_index, bool vec_stmt_p, stmt_vector_for_cost *cost_vec)
 
static void vect_loop_kill_debug_uses (class loop *loop, stmt_vec_info stmt_info)
 
static bool loop_niters_no_overflow (loop_vec_info loop_vinfo)
 
tree vect_halve_mask_nunits (tree old_type, machine_mode new_mode)
 
tree vect_double_mask_nunits (tree old_type, machine_mode new_mode)
 
void vect_record_loop_mask (loop_vec_info loop_vinfo, vec_loop_masks *masks, unsigned int nvectors, tree vectype, tree scalar_mask)
 
tree vect_get_loop_mask (loop_vec_info loop_vinfo, gimple_stmt_iterator *gsi, vec_loop_masks *masks, unsigned int nvectors, tree vectype, unsigned int index)
 
void vect_record_loop_len (loop_vec_info loop_vinfo, vec_loop_lens *lens, unsigned int nvectors, tree vectype, unsigned int factor)
 
tree vect_get_loop_len (loop_vec_info loop_vinfo, gimple_stmt_iterator *gsi, vec_loop_lens *lens, unsigned int nvectors, tree vectype, unsigned int index, unsigned int factor)
 
static void scale_profile_for_vect_loop (class loop *loop, edge exit_e, unsigned vf, bool flat)
 
static void maybe_set_vectorized_backedge_value (loop_vec_info loop_vinfo, stmt_vec_info def_stmt_info)
 
static bool vect_transform_loop_stmt (loop_vec_info loop_vinfo, stmt_vec_info stmt_info, gimple_stmt_iterator *gsi, stmt_vec_info *seen_store)
 
static tree find_in_mapping (tree t, void *context)
 
static void update_epilogue_loop_vinfo (class loop *epilogue, tree advance)
 
static void move_early_exit_stmts (loop_vec_info loop_vinfo)
 
class loopvect_transform_loop (loop_vec_info loop_vinfo, gimple *loop_vectorized_call)
 
void optimize_mask_stores (class loop *loop)
 
widest_int vect_iv_limit_for_partial_vectors (loop_vec_info loop_vinfo)
 
bool vect_rgroup_iv_might_wrap_p (loop_vec_info loop_vinfo, rgroup_controls *rgc)
 

Macro Definition Documentation

◆ INCLUDE_ALGORITHM

#define INCLUDE_ALGORITHM
Loop Vectorization
   Copyright (C) 2003-2024 Free Software Foundation, Inc.
   Contributed by Dorit Naishlos <dorit@il.ibm.com> and
   Ira Rosen <irar@il.ibm.com>

This file is part of GCC.

GCC is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free
Software Foundation; either version 3, or (at your option) any later
version.

GCC is distributed in the hope that it will be useful, but WITHOUT ANY
WARRANTY; without even the implied warranty of MERCHANTABILITY or
FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
for more details.

You should have received a copy of the GNU General Public License
along with GCC; see the file COPYING3.  If not see
<http://www.gnu.org/licenses/>.   

Function Documentation

◆ bb_in_loop_p()

static bool bb_in_loop_p ( const_basic_block bb,
const void * data )
static
Function bb_in_loop_p

Used as predicate for dfs order traversal of the loop bbs.   

References flow_bb_inside_loop_p().

◆ build_vect_cond_expr()

static void build_vect_cond_expr ( code_helper code,
tree vop[3],
tree mask,
gimple_stmt_iterator * gsi )
static
Insert a conditional expression to enable masked vectorization.  CODE is the
code for the operation.  VOP is the array of operands.  MASK is the loop
mask.  GSI is a statement iterator used to place the new conditional
expression.   

References build_zero_cst(), gcc_unreachable, ggc_alloc(), gimple_build_assign(), gsi_insert_before(), GSI_SAME_STMT, make_temp_ssa_name(), NULL, and TREE_TYPE.

Referenced by vect_transform_reduction().

◆ calc_vec_perm_mask_for_shift()

static void calc_vec_perm_mask_for_shift ( unsigned int offset,
unsigned int nelt,
vec_perm_builder * sel )
static
Writes into SEL a mask for a vec_perm, equivalent to a vec_shr by OFFSET
vector elements (not bits) for a vector with NELT elements.   

References ggc_alloc(), i, and offset.

Referenced by have_whole_vector_shift(), and vect_create_epilog_for_reduction().

◆ can_produce_all_loop_masks_p()

static bool can_produce_all_loop_masks_p ( loop_vec_info loop_vinfo,
tree cmp_type )
static
Return true if we can use CMP_TYPE as the comparison type to produce
all masks required to mask LOOP_VINFO.   

References direct_internal_fn_supported_p(), FOR_EACH_VEC_ELT, ggc_alloc(), i, LOOP_VINFO_MASKS, NULL_TREE, and OPTIMIZE_FOR_SPEED.

Referenced by vect_verify_full_masking().

◆ check_reduction_path() [1/2]

◆ check_reduction_path() [2/2]

bool check_reduction_path ( dump_user_location_t loc,
loop_p loop,
gphi * phi,
tree loop_arg,
enum tree_code )
Used in gimple-loop-interchange.c and tree-parloops.cc.   

References check_reduction_path(), ggc_alloc(), and path.

◆ cse_and_gimplify_to_preheader()

tree cse_and_gimplify_to_preheader ( loop_vec_info loop_vinfo,
tree expr )

◆ find_in_mapping()

static tree find_in_mapping ( tree t,
void * context )
static
Helper function to pass to simplify_replace_tree to enable replacing tree's
in the hash_map with its corresponding values.   

Referenced by update_epilogue_loop_vinfo().

◆ fold_left_reduction_fn()

static bool fold_left_reduction_fn ( code_helper code,
internal_fn * reduc_fn )
static
Return true if there is an in-order reduction function for CODE, storing
it in *REDUC_FN if so.   

References ggc_alloc().

Referenced by vectorizable_reduction().

◆ get_initial_def_for_reduction()

static tree get_initial_def_for_reduction ( loop_vec_info loop_vinfo,
stmt_vec_info reduc_info,
tree init_val,
tree neutral_op )
static
Function get_initial_def_for_reduction

Input:
REDUC_INFO - the info_for_reduction
INIT_VAL - the initial value of the reduction variable
NEUTRAL_OP - a value that has no effect on the reduction, as per
        neutral_op_for_reduction

Output:
Return a vector variable, initialized according to the operation that
     STMT_VINFO performs. This vector will be used as the initial value
     of the vector of partial results.

The value we need is a vector in which element 0 has value INIT_VAL
and every other element has value NEUTRAL_OP.   

References gcc_assert, get_vectype_for_scalar_type(), ggc_alloc(), gimple_bb(), gimple_build(), gimple_build_vector(), gimple_build_vector_from_val(), gimple_convert(), INTEGRAL_TYPE_P, LOOP_VINFO_LOOP, nested_in_vect_loop_p(), NULL, operand_equal_p(), POINTER_TYPE_P, SCALAR_FLOAT_TYPE_P, TREE_TYPE, TYPE_VECTOR_SUBPARTS(), and vect_emit_reduction_init_stmts().

Referenced by vect_transform_cycle_phi().

◆ get_initial_defs_for_reduction()

static void get_initial_defs_for_reduction ( loop_vec_info loop_vinfo,
stmt_vec_info reduc_info,
vec< tree > * vec_oprnds,
unsigned int number_of_vectors,
unsigned int group_size,
tree neutral_op )
static
Get at the initial defs for the reduction PHIs for REDUC_INFO,
which performs a reduction involving GROUP_SIZE scalar statements.
NUMBER_OF_VECTORS is the number of vector defs to create.  If NEUTRAL_OP
is nonnull, introducing extra elements of that value will not change the
result.   

References CONSTANT_CLASS_P, duplicate_and_interleave(), gcc_assert, ggc_alloc(), gimple_build(), gimple_build_vector(), gimple_build_vector_from_val(), i, known_eq, tree_vector_builder::new_vector(), NULL, STMT_VINFO_VECTYPE, TYPE_VECTOR_SUBPARTS(), and vect_emit_reduction_init_stmts().

Referenced by vect_transform_cycle_phi().

◆ get_masked_reduction_fn()

static internal_fn get_masked_reduction_fn ( internal_fn reduc_fn,
tree vectype_in )
static
Get a masked internal function equivalent to REDUC_FN.  VECTYPE_IN is the
type of the vector input.   

References direct_internal_fn_supported_p(), ggc_alloc(), IFN_LAST, and OPTIMIZE_FOR_SPEED.

Referenced by vectorizable_reduction(), and vectorize_fold_left_reduction().

◆ have_whole_vector_shift()

static bool have_whole_vector_shift ( machine_mode mode)
static
Checks whether the target supports whole-vector shifts for vectors of mode
MODE.  This is the case if _either_ the platform handles vec_shr_optab, _or_
it supports vec_perm_const with masks for all necessary shift amounts.   

References calc_vec_perm_mask_for_shift(), can_vec_perm_const_p(), GET_MODE_NUNITS(), ggc_alloc(), i, and optab_handler().

Referenced by vect_create_epilog_for_reduction(), and vect_model_reduction_cost().

◆ info_for_reduction()

◆ is_nonwrapping_integer_induction()

static bool is_nonwrapping_integer_induction ( stmt_vec_info stmt_vinfo,
class loop * loop )
static
Function is_nonwrapping_integer_induction.

Check if STMT_VINO (which is part of loop LOOP) both increments and
does not cause overflow.   

References wi::add(), ggc_alloc(), gimple_phi_result(), max_stmt_executions(), wi::min_precision(), wi::mul(), wi::OVF_NONE, STMT_VINFO_LOOP_PHI_EVOLUTION_BASE_UNCHANGED, STMT_VINFO_LOOP_PHI_EVOLUTION_PART, wi::to_widest(), TREE_CODE, TREE_TYPE, TYPE_OVERFLOW_UNDEFINED, TYPE_PRECISION, and TYPE_SIGN.

Referenced by vectorizable_reduction().

◆ loop_niters_no_overflow()

static bool loop_niters_no_overflow ( loop_vec_info loop_vinfo)
static
Given loop represented by LOOP_VINFO, return true if computation of
LOOP_VINFO_NITERS (= LOOP_VINFO_NITERSM1 + 1) doesn't overflow, false
otherwise.   

References gcc_assert, get_max_loop_iterations(), ggc_alloc(), LOOP_VINFO_LOOP, LOOP_VINFO_NITERS, LOOP_VINFO_NITERS_KNOWN_P, LOOP_VINFO_NITERSM1, wi::max_value(), wi::to_widest(), TREE_CODE, TREE_TYPE, and TYPE_SIGN.

Referenced by vect_transform_loop().

◆ maybe_set_vectorized_backedge_value()

◆ merge_with_identity()

static tree merge_with_identity ( gimple_stmt_iterator * gsi,
tree mask,
tree vectype,
tree vec,
tree identity )
static
Return a vector of type VECTYPE that is equal to the vector select
operation "MASK ? VEC : IDENTITY".  Insert the select statements
before GSI.   

References ggc_alloc(), gimple_build_assign(), gsi_insert_before(), GSI_SAME_STMT, make_temp_ssa_name(), and NULL.

Referenced by vectorize_fold_left_reduction().

◆ move_early_exit_stmts()

static void move_early_exit_stmts ( loop_vec_info loop_vinfo)
static
When vectorizing early break statements instructions that happen before
the early break in the current BB need to be moved to after the early
break.  This function deals with that and assumes that any validity
checks has already been performed.

While moving the instructions if it encounters a VUSE or VDEF it then
corrects the VUSES as it moves the statements along.  GDEST is the location
in which to insert the new statements.   

References CDI_DOMINATORS, dominated_by_p(), dump_enabled_p(), dump_printf_loc(), DUMP_VECT_SCOPE, FOR_EACH_IMM_USE_ON_STMT, FOR_EACH_IMM_USE_STMT, get_loop_exit_edges(), get_virtual_phi(), ggc_alloc(), gimple_phi_arg_def(), gimple_phi_result(), gimple_set_vuse(), gimple_vuse(), gsi_after_labels(), gsi_for_stmt(), gsi_move_before(), GSI_NEW_STMT, is_empty(), vec_info::lookup_stmt(), LOOP_VINFO_EARLY_BRK_DEST_BB, LOOP_VINFO_EARLY_BRK_STORES, LOOP_VINFO_EARLY_BRK_VUSES, LOOP_VINFO_LOOP, MSG_NOTE, NULL_TREE, remove_phi_node(), SET_PHI_ARG_DEF_ON_EDGE, SET_USE, data_reference::stmt, update_stmt(), and vect_location.

Referenced by vect_transform_loop().

◆ needs_fold_left_reduction_p()

bool needs_fold_left_reduction_p ( tree type,
code_helper code )
Return true if we need an in-order reduction for operation CODE
on type TYPE.  NEED_WRAPPING_INTEGRAL_OVERFLOW is true if integer
overflow must wrap.   

References ggc_alloc(), INTEGRAL_TYPE_P, code_helper::is_tree_code(), operation_no_trapping_overflow(), SAT_FIXED_POINT_TYPE_P, and SCALAR_FLOAT_TYPE_P.

Referenced by vect_optimize_slp_pass::start_choosing_layouts(), vect_reassociating_reduction_p(), vect_slp_check_for_roots(), and vectorizable_reduction().

◆ neutral_op_for_reduction()

tree neutral_op_for_reduction ( tree scalar_type,
code_helper code,
tree initial_value,
bool as_initial )
If there is a neutral value X such that a reduction would not be affected
by the introduction of additional X elements, return that X, otherwise
return null.  CODE is the code of the reduction and SCALAR_TYPE is type
of the scalar elements.  If the reduction has just a single initial value
then INITIAL_VALUE is that value, otherwise it is null.
If AS_INITIAL is TRUE the value is supposed to be used as initial value.
In that case no signed zero is returned.   

References build_all_ones_cst(), build_one_cst(), build_real(), build_zero_cst(), dconstm0, ggc_alloc(), HONOR_SIGNED_ZEROS(), code_helper::is_tree_code(), and NULL_TREE.

Referenced by convert_scalar_cond_reduction(), vect_create_epilog_for_reduction(), vect_expand_fold_left(), vect_find_reusable_accumulator(), vect_transform_cycle_phi(), and vectorizable_reduction().

◆ optimize_mask_stores()

void optimize_mask_stores ( class loop * loop)
The code below is trying to perform simple optimization - revert
  if-conversion for masked stores, i.e. if the mask of a store is zero
  do not perform it and all stored value producers also if possible.
  For example,
    for (i=0; i<n; i++)
      if (c[i])
       {
         p1[i] += 1;
         p2[i] = p3[i] +2;
       }
  this transformation will produce the following semi-hammock:

  if (!mask__ifc__42.18_165 == { 0, 0, 0, 0, 0, 0, 0, 0 })
    {
      vect__11.19_170 = MASK_LOAD (vectp_p1.20_168, 0B, mask__ifc__42.18_165);
      vect__12.22_172 = vect__11.19_170 + vect_cst__171;
      MASK_STORE (vectp_p1.23_175, 0B, mask__ifc__42.18_165, vect__12.22_172);
      vect__18.25_182 = MASK_LOAD (vectp_p3.26_180, 0B, mask__ifc__42.18_165);
      vect__19.28_184 = vect__18.25_182 + vect_cst__183;
      MASK_STORE (vectp_p2.29_187, 0B, mask__ifc__42.18_165, vect__19.28_184);
    }

References add_bb_to_loop(), add_phi_arg(), build_zero_cst(), CDI_DOMINATORS, cfun, basic_block_def::count, create_empty_bb(), create_phi_node(), dom_info_available_p(), dump_enabled_p(), dump_printf_loc(), EDGE_SUCC, find_loop_location(), flow_loop_nested_p(), FOR_EACH_IMM_USE_FAST, free(), gcc_assert, get_loop_body(), ggc_alloc(), gimple_bb(), gimple_build_cond(), gimple_call_arg(), gimple_call_internal_p(), gimple_get_lhs(), gimple_has_volatile_ops(), gimple_set_vdef(), gimple_vdef(), gimple_vop(), gimple_vuse(), gsi_end_p(), gsi_for_stmt(), gsi_insert_after(), gsi_last_bb(), gsi_move_before(), gsi_next(), gsi_prev(), gsi_remove(), GSI_SAME_STMT, gsi_start_bb(), gsi_stmt(), has_zero_uses(), i, basic_block_def::index, is_gimple_debug(), last, profile_probability::likely(), basic_block_def::loop_father, make_edge(), make_single_succ_edge(), make_ssa_name(), MSG_NOTE, NULL, NULL_TREE, loop::num_nodes, set_immediate_dominator(), split_block(), TREE_CODE, TREE_TYPE, UNKNOWN_LOCATION, USE_STMT, vect_location, VECTOR_TYPE_P, and worklist.

◆ reduction_fn_for_scalar_code()

bool reduction_fn_for_scalar_code ( code_helper code,
internal_fn * reduc_fn )
Function reduction_fn_for_scalar_code

Input:
CODE - tree_code of a reduction operations.

Output:
REDUC_FN - the corresponding internal function to be used to reduce the
   vector of partial results into a single scalar result, or IFN_LAST
   if the operation is a supported reduction operation, but does not have
   such an internal function.

Return FALSE if CODE currently cannot be vectorized as reduction.   

References ggc_alloc(), IFN_LAST, and code_helper::is_tree_code().

Referenced by vect_slp_check_for_roots(), vectorizable_bb_reduc_epilogue(), vectorizable_reduction(), and vectorize_slp_instance_root_stmt().

◆ release_vec_loop_controls()

void release_vec_loop_controls ( vec< rgroup_controls > * controls)
Free all levels of rgroup CONTROLS.   

References FOR_EACH_VEC_ELT, ggc_alloc(), and i.

Referenced by vect_analyze_loop_2(), vect_verify_full_masking_avx512(), and _loop_vec_info::~_loop_vec_info().

◆ report_vect_op()

static void report_vect_op ( dump_flags_t msg_type,
gimple * stmt,
const char * msg )
static
Error reporting helper for vect_is_simple_reduction below.  GIMPLE statement
STMT is printed with a message MSG.  

References dump_printf_loc(), ggc_alloc(), msg, and vect_location.

Referenced by vect_is_simple_reduction().

◆ scale_profile_for_vect_loop()

static void scale_profile_for_vect_loop ( class loop * loop,
edge exit_e,
unsigned vf,
bool flat )
static
Scale profiling counters by estimation for LOOP which is vectorized
by factor VF.
If FLAT is true, the loop we started with had unrealistically flat
profile.   

References profile_probability::always(), basic_block_def::count, dump_file, dump_flags, get_likely_max_loop_iterations_int(), ggc_alloc(), loop::header, loop::latch, loop_preheader_edge(), scale_loop_profile(), set_edge_probability_and_rescale_others(), single_pred_edge(), and TDF_DETAILS.

Referenced by vect_transform_loop().

◆ update_epilogue_loop_vinfo()

static void update_epilogue_loop_vinfo ( class loop * epilogue,
tree advance )
static
Update EPILOGUE's loop_vec_info.  EPILOGUE was constructed as a copy of the
original loop that has now been vectorized.

The inits of the data_references need to be advanced with the number of
iterations of the main loop.  This has been computed in vect_do_peeling and
is stored in parameter ADVANCE.  We first restore the data_references
initial offset with the values recored in ORIG_DRS_INIT.

Since the loop_vec_info of this EPILOGUE was constructed for the original
loop, its stmt_vec_infos all point to the original statements.  These need
to be updated to point to their corresponding copies as well as the SSA_NAMES
in their PATTERN_DEF_SEQs and RELATED_STMTs.

The data_reference's connections also need to be updated.  Their
corresponding dr_vec_info need to be reconnected to the EPILOGUE's
stmt_vec_infos, their statements need to point to their corresponding copy,
if they are gather loads or scatter stores then their reference needs to be
updated to point to its corresponding copy.   

References advance(), DR_BASE_ADDRESS, DR_REF, DR_STMT, find_in_mapping(), FOR_EACH_VEC_ELT, free(), gcc_assert, get_loop_body(), ggc_alloc(), gimple_bb(), gimple_get_lhs(), gimple_num_ops(), gimple_op(), gimple_phi_result(), gimple_set_bb(), gimple_set_op(), gimple_uid(), gsi_end_p(), gsi_next(), gsi_start(), gsi_start_bb(), gsi_start_phis(), gsi_stmt(), i, is_gimple_debug(), loop_vec_info_for_loop(), LOOP_VINFO_BBS, LOOP_VINFO_DATAREFS, NULL, NULL_TREE, simplify_replace_tree(), STMT_VINFO_GATHER_SCATTER_P, STMT_VINFO_MEMORY_ACCESS_TYPE, STMT_VINFO_PATTERN_DEF_SEQ, STMT_VINFO_RELATED_STMT, STMT_VINFO_STMT, vect_stmt_to_vectorize(), vect_update_inits_of_drs(), and VMAT_GATHER_SCATTER.

Referenced by vect_transform_loop().

◆ use_mask_by_cond_expr_p()

static bool use_mask_by_cond_expr_p ( code_helper code,
internal_fn cond_fn,
tree vectype_in )
static
Check if masking can be supported by inserting a conditional expression.
CODE is the code for the operation.  COND_FN is the conditional internal
function, if it exists.  VECTYPE_IN is the type of the vector input.   

References direct_internal_fn_supported_p(), ggc_alloc(), IFN_LAST, code_helper::is_tree_code(), and OPTIMIZE_FOR_SPEED.

Referenced by vect_transform_reduction(), and vectorizable_reduction().

◆ vec_init_loop_exit_info()

◆ vect_active_double_reduction_p()

static bool vect_active_double_reduction_p ( stmt_vec_info stmt_info)
static
Return true if STMT_INFO describes a double reduction phi and if
the other phi in the reduction is also relevant for vectorization.
This rejects cases such as:

   outer1:
     x_1 = PHI <x_3(outer2), ...>;
     ...

   inner:
     x_2 = ...;
     ...

   outer2:
     x_3 = PHI <x_2(inner)>;

if nothing in x_2 or elsewhere makes x_1 relevant.   

References STMT_VINFO_DEF_TYPE, STMT_VINFO_REDUC_DEF, STMT_VINFO_RELEVANT_P, and vect_double_reduction_def.

Referenced by vect_analyze_loop_operations().

◆ vect_analyze_loop()

◆ vect_analyze_loop_1()

static opt_loop_vec_info vect_analyze_loop_1 ( class loop * loop,
vec_info_shared * shared,
const vect_loop_form_info * loop_form_info,
loop_vec_info main_loop_vinfo,
const vector_modes & vector_modes,
unsigned & mode_i,
machine_mode & autodetected_vector_mode,
bool & fatal )
static
Analyze LOOP with VECTOR_MODES[MODE_I] and as epilogue if MAIN_LOOP_VINFO is
not NULL.  Set AUTODETECTED_VECTOR_MODE if VOIDmode and advance
MODE_I to the next mode useful to analyze.
Return the loop_vinfo on success and wrapped null on failure.   

References dump_enabled_p(), dump_printf_loc(), fatal(), gcc_checking_assert, GET_MODE_INNER, GET_MODE_NAME, ggc_alloc(), MSG_NOTE, NULL, opt_pointer_wrapper< PtrType_t >::propagate_failure(), related_vector_mode(), opt_pointer_wrapper< PtrType_t >::success(), vect_analyze_loop_2(), vect_chooses_same_modes_p(), vect_create_loop_vinfo(), vect_location, vec_info::vector_mode, and VECTOR_MODE_P.

Referenced by vect_analyze_loop().

◆ vect_analyze_loop_2()

static opt_result vect_analyze_loop_2 ( loop_vec_info loop_vinfo,
bool & fatal,
unsigned * suggested_unroll_factor,
bool & slp_done_for_suggested_uf )
static
Function vect_analyze_loop_2.

Apply a set of analyses on LOOP specified by LOOP_VINFO, the different
analyses will record information in some members of LOOP_VINFO.  FATAL
indicates if some analysis meets fatal error.  If one non-NULL pointer
SUGGESTED_UNROLL_FACTOR is provided, it's intent to be filled with one
worked out suggested unroll factor, while one NULL pointer shows it's
going to apply the suggested unroll factor.  SLP_DONE_FOR_SUGGESTED_UF
is to hold the slp decision when the suggested unroll factor is worked
out.   

References vec_info_shared::check_datarefs(), direct_internal_fn_supported_p(), DR_GROUP_FIRST_ELEMENT, DR_GROUP_NEXT_ELEMENT, DR_GROUP_SIZE, dump_dec(), dump_enabled_p(), dump_printf(), dump_printf_loc(), opt_result::failure_at(), fatal(), FOR_EACH_VEC_ELT, gcc_assert, ggc_alloc(), gsi_end_p(), gsi_next(), gsi_start(), gsi_start_bb(), gsi_start_phis(), gsi_stmt(), i, IFN_LAST, init_cost, integer_zerop(), is_empty(), is_gimple_debug(), known_eq, known_le, known_ne, vec_info::lookup_stmt(), LOOP_REQUIRES_VERSIONING, loop_vect, LOOP_VINFO_BBS, LOOP_VINFO_CAN_USE_PARTIAL_VECTORS_P, LOOP_VINFO_CHECK_UNEQUAL_ADDRS, LOOP_VINFO_COMP_ALIAS_DDRS, LOOP_VINFO_COST_MODEL_THRESHOLD, LOOP_VINFO_DATAREFS, LOOP_VINFO_EARLY_BREAKS, LOOP_VINFO_EPILOGUE_P, LOOP_VINFO_INT_NITERS, LOOP_VINFO_IV_EXIT, LOOP_VINFO_LENS, LOOP_VINFO_LOOP, LOOP_VINFO_LOWER_BOUNDS, LOOP_VINFO_MASKS, LOOP_VINFO_MAX_VECT_FACTOR, LOOP_VINFO_N_STMTS, LOOP_VINFO_NITERS_KNOWN_P, LOOP_VINFO_ORIG_LOOP_INFO, LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS, LOOP_VINFO_PARTIAL_VECTORS_STYLE, LOOP_VINFO_PEELING_FOR_ALIGNMENT, LOOP_VINFO_PEELING_FOR_GAPS, LOOP_VINFO_PEELING_FOR_NITER, LOOP_VINFO_REDUCTION_CHAINS, LOOP_VINFO_RGROUP_IV_TYPE, LOOP_VINFO_SIMD_IF_COND, LOOP_VINFO_SLP_INSTANCES, LOOP_VINFO_UNALIGNED_DR, LOOP_VINFO_USING_DECREMENTING_IV_P, LOOP_VINFO_USING_PARTIAL_VECTORS_P, LOOP_VINFO_USING_SELECT_VL_P, LOOP_VINFO_VECT_FACTOR, LOOP_VINFO_VECTORIZABLE_P, LOOP_VINFO_VERSIONING_THRESHOLD, MAX_VECTORIZATION_FACTOR, maybe_ge, MSG_MISSED_OPTIMIZATION, MSG_NOTE, NULL, OPTIMIZE_FOR_SPEED, release_vec_loop_controls(), vec_info_shared::save_datarefs(), vec_info::shared, si, slp_inst_kind_store, SLP_INSTANCE_KIND, SLP_INSTANCE_LOADS, SLP_INSTANCE_TREE, SLP_TREE_LANES, SLP_TREE_LOAD_PERMUTATION, SLP_TREE_REPRESENTATIVE, SLP_TREE_SCALAR_STMTS, SLP_TREE_VECTYPE, slpeel_can_duplicate_loop_p(), STMT_SLP_TYPE, STMT_VINFO_DEF_TYPE, STMT_VINFO_GROUPED_ACCESS, STMT_VINFO_IN_PATTERN_P, STMT_VINFO_PATTERN_DEF_SEQ, STMT_VINFO_REDUC_DEF, STMT_VINFO_RELATED_STMT, STMT_VINFO_SLP_VECT_ONLY_PATTERN, STMT_VINFO_STRIDED_P, STMT_VINFO_VECTYPE, opt_result::success(), _loop_vec_info::suggested_unroll_factor, TYPE_VECTOR_SUBPARTS(), vect_analyze_data_ref_accesses(), vect_analyze_data_ref_dependences(), vect_analyze_data_refs(), vect_analyze_data_refs_alignment(), vect_analyze_loop_costing(), vect_analyze_loop_operations(), vect_analyze_scalar_cycles(), vect_analyze_slp(), vect_apply_runtime_profitability_check_p(), vect_can_advance_ivs_p(), vect_compute_single_scalar_iteration_cost(), vect_detect_hybrid_slp(), vect_determine_partial_vectors_and_peeling(), vect_determine_vectorization_factor(), vect_dissolve_slp_only_groups(), vect_double_reduction_def, vect_enhance_data_refs_alignment(), vect_fixup_scalar_cycles_with_patterns(), vect_free_slp_instance(), vect_gather_slp_loads(), vect_get_datarefs_in_loop(), vect_grouped_load_supported(), vect_grouped_store_supported(), vect_load_lanes_supported(), vect_location, vect_make_slp_decision(), vect_mark_stmts_to_be_vectorized(), vect_optimize_slp(), vect_partial_vectors_len, vect_pattern_recog(), vect_prune_runtime_alias_test_list(), vect_reduction_def, vect_slp_analyze_operations(), vect_stmt_to_vectorize(), vect_store_lanes_supported(), vect_update_vf_for_slp(), vect_use_loop_mask_for_alignment_p(), vect_verify_full_masking(), vect_verify_full_masking_avx512(), vect_verify_loop_lens(), and _loop_vec_info::vector_costs.

Referenced by vect_analyze_loop_1().

◆ vect_analyze_loop_costing()

◆ vect_analyze_loop_form()

◆ vect_analyze_loop_operations()

◆ vect_analyze_scalar_cycles()

static void vect_analyze_scalar_cycles ( loop_vec_info loop_vinfo,
bool slp )
static
Function vect_analyze_scalar_cycles.

Examine the cross iteration def-use cycles of scalar variables, by
analyzing the loop-header PHIs of scalar variables.  Classify each
cycle as one of the following: invariant, induction, reduction, unknown.
We do that for the loop represented by LOOP_VINFO, and also to its
inner-loop, if exists.
Examples for scalar cycles:

Example1: reduction:

           loop1:
           for (i=0; i<N; i++)
              sum += a[i];

Example2: induction:

           loop2:
           for (i=0; i<N; i++)
              a[i] = i;   

References ggc_alloc(), loop::inner, LOOP_VINFO_LOOP, and vect_analyze_scalar_cycles_1().

Referenced by vect_analyze_loop_2().

◆ vect_analyze_scalar_cycles_1()

◆ vect_better_loop_vinfo_p()

static bool vect_better_loop_vinfo_p ( loop_vec_info new_loop_vinfo,
loop_vec_info old_loop_vinfo )
static
Return true if vectorizing a loop using NEW_LOOP_VINFO appears
to be better than vectorizing it using OLD_LOOP_VINFO.  Assume that
OLD_LOOP_VINFO is better unless something specifically indicates
otherwise.

Note that this deliberately isn't a partial order.   

References gcc_assert, ggc_alloc(), known_eq, LOOP_VINFO_LOOP, LOOP_VINFO_ORIG_LOOP_INFO, LOOP_VINFO_VECT_FACTOR, and loop::simdlen.

Referenced by vect_joust_loop_vinfos().

◆ vect_can_vectorize_without_simd_p() [1/2]

bool vect_can_vectorize_without_simd_p ( code_helper code)
Likewise, but taking a code_helper.   

References code_helper::is_tree_code(), and vect_can_vectorize_without_simd_p().

◆ vect_can_vectorize_without_simd_p() [2/2]

bool vect_can_vectorize_without_simd_p ( tree_code code)
Return true if we can emulate CODE on an integer mode representation
of a vector.   

References ggc_alloc().

Referenced by vect_can_vectorize_without_simd_p(), vectorizable_operation(), and vectorizable_reduction().

◆ vect_compute_single_scalar_iteration_cost()

◆ vect_create_epilog_for_reduction()

static void vect_create_epilog_for_reduction ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info,
slp_tree slp_node,
slp_instance slp_node_instance,
edge loop_exit )
static
Function vect_create_epilog_for_reduction

  Create code at the loop-epilog to finalize the result of a reduction
  computation. 
 
  STMT_INFO is the scalar reduction stmt that is being vectorized.
  SLP_NODE is an SLP node containing a group of reduction statements. The 
    first one in this group is STMT_INFO.
  SLP_NODE_INSTANCE is the SLP node instance containing SLP_NODE
  REDUC_INDEX says which rhs operand of the STMT_INFO is the reduction phi
    (counting from 0)
  LOOP_EXIT is the edge to update in the merge block.  In the case of a single
    exit this edge is always the main loop exit.

  This function:
  1. Completes the reduction def-use cycles.
  2. "Reduces" each vector of partial results VECT_DEFS into a single result,
     by calling the function specified by REDUC_FN if available, or by
     other means (whole-vector shifts or a scalar loop).
     The function also creates a new phi node at the loop exit to preserve
     loop-closed form, as illustrated below.

    The flow at the entry to this function:

       loop:
         vec_def = phi <vec_init, null>        # REDUCTION_PHI
         VECT_DEF = vector_stmt                # vectorized form of STMT_INFO
         s_loop = scalar_stmt                  # (scalar) STMT_INFO
       loop_exit:
         s_out0 = phi <s_loop>                 # (scalar) EXIT_PHI
         use <s_out0>
         use <s_out0>

    The above is transformed by this function into:

       loop:
         vec_def = phi <vec_init, VECT_DEF>    # REDUCTION_PHI
         VECT_DEF = vector_stmt                # vectorized form of STMT_INFO
         s_loop = scalar_stmt                  # (scalar) STMT_INFO
       loop_exit:
         s_out0 = phi <s_loop>                 # (scalar) EXIT_PHI
         v_out1 = phi <VECT_DEF>               # NEW_EXIT_PHI
         v_out2 = reduce <v_out1>
         s_out3 = extract_field <v_out2, 0>
         s_out4 = adjust_result <s_out3>
         use <s_out4>
         use <s_out4>

References add_phi_arg(), as_combined_fn(), bitsize_int, bitsize_zero_node, boolean_type_node, build1(), build3(), build_index_vector(), build_int_cst(), build_vector_from_val(), build_zero_cst(), calc_vec_perm_mask_for_shift(), COND_REDUCTION, copy_ssa_name(), create_iv(), create_phi_node(), directly_supported_p(), dump_enabled_p(), dump_printf_loc(), exact_log2(), flow_bb_inside_loop_p(), FOR_EACH_IMM_USE_FAST, FOR_EACH_IMM_USE_ON_STMT, FOR_EACH_IMM_USE_STMT, FOR_EACH_VEC_ELT, gcc_assert, gcc_checking_assert, gcc_unreachable, GET_MODE_NUNITS(), GET_MODE_PRECISION(), get_related_vectype_for_scalar_type(), get_same_sized_vectype(), ggc_alloc(), gimple_assign_rhs1(), gimple_assign_rhs_code(), gimple_assign_set_lhs(), gimple_bb(), gimple_build(), gimple_build_assign(), gimple_build_call_internal(), gimple_build_vector_from_val(), gimple_call_set_lhs(), gimple_convert(), gimple_get_lhs(), gimple_op(), gimple_phi_arg_def(), gimple_phi_num_args(), gimple_seq_last_stmt(), gsi_after_labels(), gsi_insert_before(), gsi_insert_seq_before(), GSI_SAME_STMT, have_whole_vector_shift(), loop::header, i, IFN_LAST, info_for_reduction(), loop::inner, INTEGER_INDUC_COND_REDUCTION, poly_int< N, C >::is_constant(), is_gimple_debug(), least_common_multiple(), vec_info::lookup_def(), loop_latch_edge(), loop_preheader_edge(), LOOP_VINFO_IV_EXIT, LOOP_VINFO_LOOP, make_ssa_name(), make_unsigned_type(), MSG_NOTE, nested_in_vect_loop_p(), neutral_op_for_reduction(), NULL, NULL_TREE, PHI_RESULT, phis, pow2p_hwi(), hash_map< KeyId, Value, Traits >::put(), REDUC_GROUP_FIRST_ELEMENT, _loop_vec_info::reusable_accumulators, SCALAR_TYPE_MODE, SET_PHI_ARG_DEF, SET_USE, single_imm_use(), single_succ(), single_succ_edge(), single_succ_p(), _loop_vec_info::skip_this_loop_edge, SLP_TREE_LANES, SLP_TREE_SCALAR_STMTS, SLP_TREE_VEC_DEFS, STMT_VINFO_DEF_TYPE, STMT_VINFO_IN_PATTERN_P, STMT_VINFO_REDUC_CODE, STMT_VINFO_REDUC_DEF, STMT_VINFO_REDUC_EPILOGUE_ADJUSTMENT, STMT_VINFO_REDUC_FN, STMT_VINFO_REDUC_IDX, STMT_VINFO_REDUC_TYPE, STMT_VINFO_REDUC_VECTYPE, STMT_VINFO_RELATED_STMT, STMT_VINFO_VEC_INDUC_COND_INITIAL_VAL, STMT_VINFO_VEC_STMTS, targetm, poly_int< N, C >::to_constant(), TREE_CODE, TREE_CODE_REDUCTION, tree_to_uhwi(), TREE_TYPE, truth_type_for(), TYPE_MODE, TYPE_SIZE, TYPE_UNSIGNED, TYPE_VECTOR_SUBPARTS(), UNKNOWN_LOCATION, unshare_expr(), update_stmt(), USE_STMT, vect_create_destination_var(), vect_create_partial_epilog(), vect_double_reduction_def, vect_gen_perm_mask_any(), vect_get_slp_vect_def(), vect_iv_increment_position(), vect_location, vect_orig_stmt(), vect_stmt_to_vectorize(), VECTOR_MODE_P, and VECTOR_TYPE_P.

Referenced by vectorizable_live_operation().

◆ vect_create_loop_vinfo()

◆ vect_create_nonlinear_iv_init()

◆ vect_create_nonlinear_iv_step()

◆ vect_create_nonlinear_iv_vec_step()

static tree vect_create_nonlinear_iv_vec_step ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info,
tree new_name,
tree vectype,
enum vect_induction_op_type induction_type )
static

◆ vect_create_partial_epilog()

◆ vect_determine_partial_vectors_and_peeling()

opt_result vect_determine_partial_vectors_and_peeling ( loop_vec_info loop_vinfo)
Determine if operating on full vectors for LOOP_VINFO might leave
 some scalar iterations still to do.  If so, decide how we should
 handle those scalar iterations.  The possibilities are:

 (1) Make LOOP_VINFO operate on partial vectors instead of full vectors.
     In this case:

       LOOP_VINFO_USING_PARTIAL_VECTORS_P == true
       LOOP_VINFO_EPIL_USING_PARTIAL_VECTORS_P == false
       LOOP_VINFO_PEELING_FOR_NITER == false

 (2) Make LOOP_VINFO operate on full vectors and use an epilogue loop
     to handle the remaining scalar iterations.  In this case:

       LOOP_VINFO_USING_PARTIAL_VECTORS_P == false
       LOOP_VINFO_PEELING_FOR_NITER == true

     There are two choices:

     (2a) Consider vectorizing the epilogue loop at the same VF as the
          main loop, but using partial vectors instead of full vectors.
          In this case:

            LOOP_VINFO_EPIL_USING_PARTIAL_VECTORS_P == true

     (2b) Consider vectorizing the epilogue loop at lower VFs only.
          In this case:

            LOOP_VINFO_EPIL_USING_PARTIAL_VECTORS_P == false

References dump_enabled_p(), dump_printf_loc(), ggc_alloc(), LOOP_VINFO_CAN_USE_PARTIAL_VECTORS_P, LOOP_VINFO_EPIL_USING_PARTIAL_VECTORS_P, LOOP_VINFO_EPILOGUE_P, LOOP_VINFO_PEELING_FOR_NITER, LOOP_VINFO_USING_PARTIAL_VECTORS_P, LOOP_VINFO_USING_SELECT_VL_P, MSG_NOTE, opt_result::success(), _loop_vec_info::suggested_unroll_factor, vect_known_niters_smaller_than_vf(), vect_location, and vect_need_peeling_or_partial_vectors_p().

Referenced by vect_analyze_loop_2(), and vect_do_peeling().

◆ vect_determine_vectorization_factor()

static opt_result vect_determine_vectorization_factor ( loop_vec_info loop_vinfo)
static
Function vect_determine_vectorization_factor

  Determine the vectorization factor (VF).  VF is the number of data elements
  that are operated upon in parallel in a single iteration of the vectorized
  loop.  For example, when vectorizing a loop that operates on 4byte elements,
  on a target with vector size (VS) 16byte, the VF is set to 4, since 4
  elements can fit in a single vector register.

  We currently support vectorization of loops in which all types operated upon
  are of the same size.  Therefore this function currently sets VF according to
  the size of the types operated upon, and fails if there are multiple sizes
  in the loop.

  VF is also the factor by which the loop iterations are strip-mined, e.g.:
  original loop:
       for (i=0; i<N; i++){
         a[i] = b[i] + c[i];
       }

  vectorized loop:
       for (i=0; i<N; i+=VF){
         a[i:VF] = b[i:VF] + c[i:VF];
       }

References dump_dec(), dump_enabled_p(), dump_printf(), dump_printf_loc(), DUMP_VECT_SCOPE, opt_result::failure_at(), gcc_assert, get_vectype_for_scalar_type(), ggc_alloc(), gsi_end_p(), gsi_next(), gsi_start_bb(), gsi_start_phis(), gsi_stmt(), i, is_gimple_debug(), known_le, vec_info::lookup_stmt(), LOOP_VINFO_BBS, LOOP_VINFO_LOOP, LOOP_VINFO_VECT_FACTOR, MSG_NOTE, NULL_TREE, loop::num_nodes, PHI_RESULT, si, STMT_VINFO_LIVE_P, STMT_VINFO_RELEVANT_P, STMT_VINFO_VECTYPE, opt_result::success(), TREE_TYPE, TYPE_VECTOR_SUBPARTS(), vect_determine_vf_for_stmt(), vect_location, and vect_update_max_nunits().

Referenced by vect_analyze_loop_2().

◆ vect_determine_vf_for_stmt()

static opt_result vect_determine_vf_for_stmt ( vec_info * vinfo,
stmt_vec_info stmt_info,
poly_uint64 * vf )
static
Subroutine of vect_determine_vectorization_factor.  Set the vector
types of STMT_INFO and all attached pattern statements and update
the vectorization factor VF accordingly.  Return true on success
or false if something prevented vectorization.   

References dump_enabled_p(), dump_printf_loc(), ggc_alloc(), gsi_end_p(), gsi_next(), gsi_start(), gsi_stmt(), vec_info::lookup_stmt(), MSG_NOTE, si, STMT_VINFO_IN_PATTERN_P, STMT_VINFO_PATTERN_DEF_SEQ, STMT_VINFO_RELATED_STMT, opt_result::success(), vect_determine_vf_for_stmt_1(), and vect_location.

Referenced by vect_determine_vectorization_factor().

◆ vect_determine_vf_for_stmt_1()

static opt_result vect_determine_vf_for_stmt_1 ( vec_info * vinfo,
stmt_vec_info stmt_info,
bool vectype_maybe_set_p,
poly_uint64 * vf )
static
Subroutine of vect_determine_vf_for_stmt that handles only one
statement.  VECTYPE_MAYBE_SET_P is true if STMT_VINFO_VECTYPE
may already be set for general statements (not just data refs).   

References dump_enabled_p(), dump_printf_loc(), gcc_assert, ggc_alloc(), gimple_clobber_p(), MSG_NOTE, stmt_vectype(), STMT_VINFO_DATA_REF, STMT_VINFO_LIVE_P, STMT_VINFO_RELEVANT_P, STMT_VINFO_VECTYPE, opt_result::success(), vect_get_vector_types_for_stmt(), vect_location, and vect_update_max_nunits().

Referenced by vect_determine_vf_for_stmt().

◆ vect_dissolve_slp_only_groups()

◆ vect_double_mask_nunits()

tree vect_double_mask_nunits ( tree old_type,
machine_mode new_mode )
Return a mask type with twice as many elements as OLD_TYPE,
given that it should have mode NEW_MODE.   

References build_truth_vector_type_for_mode(), ggc_alloc(), new_mode(), and TYPE_VECTOR_SUBPARTS().

Referenced by supportable_narrowing_operation().

◆ vect_emit_reduction_init_stmts()

static void vect_emit_reduction_init_stmts ( loop_vec_info loop_vinfo,
stmt_vec_info reduc_info,
gimple * seq )
static
SEQ is a sequence of instructions that initialize the reduction
described by REDUC_INFO.  Emit them in the appropriate place.   

References gcc_assert, ggc_alloc(), gsi_insert_seq_before(), gsi_insert_seq_on_edge_immediate(), gsi_last_bb(), GSI_SAME_STMT, loop_preheader_edge(), LOOP_VINFO_LOOP, and _loop_vec_info::skip_main_loop_edge.

Referenced by get_initial_def_for_reduction(), and get_initial_defs_for_reduction().

◆ vect_emulate_mixed_dot_prod()

static gassign * vect_emulate_mixed_dot_prod ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info,
gimple_stmt_iterator * gsi,
tree vec_dest,
tree vop[3] )
static
STMT_INFO is a dot-product reduction whose multiplication operands
have different signs.  Emit a sequence to emulate the operation
using a series of signed DOT_PROD_EXPRs and return the last
statement generated.  VEC_DEST is the result of the vector operation
and VOP lists its inputs.   

References build_vector_from_val(), ggc_alloc(), gimple_build_assign(), i, wi::lrshift(), make_ssa_name(), signed_type_for(), wi::to_wide(), TREE_TYPE, TYPE_MIN_VALUE, TYPE_UNSIGNED, vect_finish_stmt_generation(), and wide_int_to_tree().

Referenced by vect_transform_reduction().

◆ vect_emulated_vector_p()

bool vect_emulated_vector_p ( tree vectype)
Return true if VECTYPE represents a vector that requires lowering
by the vector lowering pass.   

References TREE_TYPE, TYPE_MODE, TYPE_PRECISION, VECTOR_BOOLEAN_TYPE_P, and VECTOR_MODE_P.

Referenced by vectorizable_call(), vectorizable_operation(), vectorizable_reduction(), and vectorizable_shift().

◆ vect_estimate_min_profitable_iters()

static void vect_estimate_min_profitable_iters ( loop_vec_info loop_vinfo,
int * ret_min_profitable_niters,
int * ret_min_profitable_estimate,
unsigned * suggested_unroll_factor )
static
Loop Vectorization Pass.

  This pass tries to vectorize loops.

  For example, the vectorizer transforms the following simple loop:

       short a[N]; short b[N]; short c[N]; int i;

       for (i=0; i<N; i++){
         a[i] = b[i] + c[i];
       }

  as if it was manually vectorized by rewriting the source code into:

       typedef int __attribute__((mode(V8HI))) v8hi;
       short a[N];  short b[N]; short c[N];   int i;
       v8hi *pa = (v8hi*)a, *pb = (v8hi*)b, *pc = (v8hi*)c;
       v8hi va, vb, vc;

       for (i=0; i<N/8; i++){
         vb = pb[i];
         vc = pc[i];
         va = vb + vc;
         pa[i] = va;
       }

       The main entry to this pass is vectorize_loops(), in which
  the vectorizer applies a set of analyses on a given set of loops,
  followed by the actual vectorization transformation for the loops that
  had successfully passed the analysis phase.
       Throughout this pass we make a distinction between two types of
  data: scalars (which are represented by SSA_NAMES), and memory references
  ("data-refs").  These two types of data require different handling both
  during analysis and transformation. The types of data-refs that the
  vectorizer currently supports are ARRAY_REFS which base is an array DECL
  (not a pointer), and INDIRECT_REFS through pointers; both array and pointer
  accesses are required to have a simple (consecutive) access pattern.

  Analysis phase:
  ===============
       The driver for the analysis phase is vect_analyze_loop().
  It applies a set of analyses, some of which rely on the scalar evolution
  analyzer (scev) developed by Sebastian Pop.

       During the analysis phase the vectorizer records some information
  per stmt in a "stmt_vec_info" struct which is attached to each stmt in the
  loop, as well as general information about the loop as a whole, which is
  recorded in a "loop_vec_info" struct attached to each loop.

  Transformation phase:
  =====================
       The loop transformation phase scans all the stmts in the loop, and
  creates a vector stmt (or a sequence of stmts) for each scalar stmt S in
  the loop that needs to be vectorized.  It inserts the vector code sequence
  just before the scalar stmt S, and records a pointer to the vector code
  in STMT_VINFO_VEC_STMT (stmt_info) (stmt_info is the stmt_vec_info struct
  attached to S).  This pointer will be used for the vectorization of following
  stmts which use the def of stmt S. Stmt S is removed if it writes to memory;
  otherwise, we rely on dead code elimination for removing it.

       For example, say stmt S1 was vectorized into stmt VS1:

  VS1: vb = px[i];
  S1:  b = x[i];    STMT_VINFO_VEC_STMT (stmt_info (S1)) = VS1
  S2:  a = b;

  To vectorize stmt S2, the vectorizer first finds the stmt that defines
  the operand 'b' (S1), and gets the relevant vector def 'vb' from the
  vector stmt VS1 pointed to by STMT_VINFO_VEC_STMT (stmt_info (S1)).  The
  resulting sequence would be:

  VS1: vb = px[i];
  S1:  b = x[i];       STMT_VINFO_VEC_STMT (stmt_info (S1)) = VS1
  VS2: va = vb;
  S2:  a = b;          STMT_VINFO_VEC_STMT (stmt_info (S2)) = VS2

       Operands that are not SSA_NAMEs, are data-refs that appear in
  load/store operations (like 'x[i]' in S1), and are handled differently.

  Target modeling:
  =================
       Currently the only target specific information that is used is the
  size of the vector (in bytes) - "TARGET_VECTORIZE_UNITS_PER_SIMD_WORD".
  Targets that can support different sizes of vectors, for now will need
  to specify one value for "TARGET_VECTORIZE_UNITS_PER_SIMD_WORD".  More
  flexibility will be added in the future.

       Since we only vectorize operations which vector form can be
  expressed using existing tree codes, to verify that an operation is
  supported, the vectorizer checks the relevant optab at the relevant
  machine_mode (e.g, optab_handler (add_optab, V8HImode)).  If
  the value found is CODE_FOR_nothing, then there's no target support, and
  we can't vectorize the stmt.

  For additional information on this project see:
  http://gcc.gnu.org/projects/tree-ssa/vectorization.html
Function vect_estimate_min_profitable_iters

Return the number of iterations required for the vector version of the
loop to be profitable relative to the cost of the scalar version of the
loop.

*RET_MIN_PROFITABLE_NITERS is a cost model profitability threshold
of iterations for vectorization.  -1 value means loop vectorization
is not profitable.  This returned value may be used for dynamic
profitability check.

*RET_MIN_PROFITABLE_ESTIMATE is a profitability threshold to be used
for static check against estimated number of iterations.   

References add_stmt_cost(), cond_branch_not_taken, cond_branch_taken, dump_enabled_p(), dump_printf(), dump_printf_loc(), finish_cost(), FOR_EACH_VEC_ELT, gcc_assert, dump_user_location_t::get_location_t(), ggc_alloc(), HOST_WIDE_INT_PRINT_UNSIGNED, i, known_le, LOOP_REQUIRES_VERSIONING, LOOP_REQUIRES_VERSIONING_FOR_ALIAS, LOOP_REQUIRES_VERSIONING_FOR_ALIGNMENT, LOOP_REQUIRES_VERSIONING_FOR_NITERS, LOOP_VINFO_CHECK_UNEQUAL_ADDRS, LOOP_VINFO_COMP_ALIAS_DDRS, LOOP_VINFO_EPILOGUE_P, LOOP_VINFO_FULLY_MASKED_P, LOOP_VINFO_FULLY_WITH_LENGTH_P, LOOP_VINFO_LENS, LOOP_VINFO_LOOP, LOOP_VINFO_LOWER_BOUNDS, LOOP_VINFO_MASKS, LOOP_VINFO_MAX_VECT_FACTOR, LOOP_VINFO_MAY_MISALIGN_STMTS, LOOP_VINFO_NITERS_KNOWN_P, LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS, LOOP_VINFO_PARTIAL_VECTORS_STYLE, LOOP_VINFO_PEELING_FOR_ALIGNMENT, LOOP_VINFO_PEELING_FOR_GAPS, LOOP_VINFO_RGROUP_IV_TYPE, LOOP_VINFO_SCALAR_ITERATION_COST, LOOP_VINFO_USING_DECREMENTING_IV_P, LOOP_VINFO_USING_PARTIAL_VECTORS_P, LOOP_VINFO_VECT_FACTOR, MAX, MAX_VECTORIZATION_FACTOR, MSG_MISSED_OPTIMIZATION, MSG_NOTE, NULL, NULL_TREE, _loop_vec_info::scalar_costs, scalar_stmt, si, vector_costs::total_cost(), TREE_TYPE, TYPE_PRECISION, unlimited_cost_model(), vect_body, vect_epilogue, vect_get_peel_iters_epilogue(), vect_get_stmt_cost(), vect_known_niters_smaller_than_vf(), vect_location, vect_partial_vectors_avx512, vect_partial_vectors_while_ult, vect_prologue, vect_rgroup_iv_might_wrap_p(), vect_use_loop_mask_for_alignment_p(), vect_vf_for_cost(), _loop_vec_info::vector_costs, vector_stmt, and warning_at().

Referenced by vect_analyze_loop_costing().

◆ vect_expand_fold_left()

static tree vect_expand_fold_left ( gimple_stmt_iterator * gsi,
tree scalar_dest,
tree_code code,
tree lhs,
tree vector_rhs,
tree mask )
static
Successively apply CODE to each element of VECTOR_RHS, in left-to-right
order, starting with LHS.  Insert the extraction statements before GSI and
associate the new scalar SSA names with variable SCALAR_DEST.
If MASK is nonzero mask the input and then operate on it unconditionally.
Return the SSA name for the result.   

References bitsize_int, build3(), build_vector_from_val(), ggc_alloc(), gimple_assign_set_lhs(), gimple_build_assign(), gsi_insert_before(), GSI_SAME_STMT, make_ssa_name(), make_temp_ssa_name(), neutral_op_for_reduction(), NULL, NULL_TREE, tree_to_uhwi(), TREE_TYPE, and TYPE_SIZE.

Referenced by vectorize_fold_left_reduction().

◆ vect_find_reusable_accumulator()

◆ vect_fixup_reduc_chain()

static void vect_fixup_reduc_chain ( stmt_vec_info stmt_info)
static
Transfer group and reduction information from STMT_INFO to its
pattern stmt.   

References gcc_assert, gcc_checking_assert, ggc_alloc(), REDUC_GROUP_FIRST_ELEMENT, REDUC_GROUP_NEXT_ELEMENT, REDUC_GROUP_SIZE, STMT_VINFO_DEF_TYPE, and STMT_VINFO_RELATED_STMT.

Referenced by vect_fixup_scalar_cycles_with_patterns().

◆ vect_fixup_scalar_cycles_with_patterns()

◆ vect_get_datarefs_in_loop()

◆ vect_get_known_peeling_cost()

int vect_get_known_peeling_cost ( loop_vec_info loop_vinfo,
int peel_iters_prologue,
int * peel_iters_epilogue,
stmt_vector_for_cost * scalar_cost_vec,
stmt_vector_for_cost * prologue_cost_vec,
stmt_vector_for_cost * epilogue_cost_vec )

◆ vect_get_loop_len()

tree vect_get_loop_len ( loop_vec_info loop_vinfo,
gimple_stmt_iterator * gsi,
vec_loop_lens * lens,
unsigned int nvectors,
tree vectype,
unsigned int index,
unsigned int factor )
Given a complete set of lengths LENS, extract length number INDEX
for an rgroup that operates on NVECTORS vectors of type VECTYPE,
where 0 <= INDEX < NVECTORS.  Return a value that contains FACTOR
multipled by the number of elements that should be processed.
Insert any set-up statements before GSI.   

References build_int_cst(), gcc_assert, ggc_alloc(), gimple_build(), gimple_build_nop(), gsi_insert_seq_before(), GSI_SAME_STMT, i, LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS, LOOP_VINFO_RGROUP_COMPARE_TYPE, LOOP_VINFO_RGROUP_IV_TYPE, make_temp_ssa_name(), NULL, NULL_TREE, SSA_NAME_DEF_STMT, and TYPE_VECTOR_SUBPARTS().

Referenced by vect_get_loop_variant_data_ptr_increment(), vect_get_strided_load_store_ops(), vectorizable_call(), vectorizable_condition(), vectorizable_induction(), vectorizable_live_operation_1(), vectorizable_load(), vectorizable_operation(), vectorizable_store(), and vectorize_fold_left_reduction().

◆ vect_get_loop_mask()

◆ vect_get_loop_niters()

static vec< gcond * > vect_get_loop_niters ( class loop * loop,
const_edge main_exit,
tree * assumptions,
tree * number_of_iterations,
tree * number_of_iterationsm1 )
static

◆ vect_get_max_nscalars_per_iter()

static unsigned int vect_get_max_nscalars_per_iter ( loop_vec_info loop_vinfo)
static
Calculate the maximum number of scalars per iteration for every
rgroup in LOOP_VINFO.   

References FOR_EACH_VEC_ELT, ggc_alloc(), i, LOOP_VINFO_MASKS, and MAX.

Referenced by vect_verify_full_masking().

◆ vect_get_peel_iters_epilogue()

static int vect_get_peel_iters_epilogue ( loop_vec_info loop_vinfo,
int peel_iters_prologue )
static
Estimate the number of peeled epilogue iterations for LOOP_VINFO.
PEEL_ITERS_PROLOGUE is the number of peeled prologue iterations,
or -1 if not known.   

References dump_enabled_p(), dump_printf_loc(), ggc_alloc(), LOOP_VINFO_INT_NITERS, LOOP_VINFO_NITERS_KNOWN_P, LOOP_VINFO_PEELING_FOR_GAPS, MIN, MSG_NOTE, vect_location, and vect_vf_for_cost().

Referenced by vect_estimate_min_profitable_iters(), and vect_get_known_peeling_cost().

◆ vect_halve_mask_nunits()

tree vect_halve_mask_nunits ( tree old_type,
machine_mode new_mode )
Return a mask type with half the number of elements as OLD_TYPE,
given that it should have mode NEW_MODE.   

References build_truth_vector_type_for_mode(), ggc_alloc(), new_mode(), and TYPE_VECTOR_SUBPARTS().

Referenced by supportable_widening_operation(), and vect_maybe_permute_loop_masks().

◆ vect_inner_phi_in_double_reduction_p()

static bool vect_inner_phi_in_double_reduction_p ( loop_vec_info loop_vinfo,
gphi * phi )
static
Return true if PHI, described by STMT_INFO, is the inner PHI in
what we are assuming is a double reduction.  For example, given
a structure like this:

   outer1:
     x_1 = PHI <x_4(outer2), ...>;
     ...

   inner:
     x_2 = PHI <x_1(outer1), ...>;
     ...
     x_3 = ...;
     ...

   outer2:
     x_4 = PHI <x_3(inner)>;
     ...

outer loop analysis would treat x_1 as a double reduction phi and
this function would then return true for x_2.   

References FOR_EACH_PHI_ARG, ggc_alloc(), vec_info::lookup_def(), SSA_OP_USE, STMT_VINFO_DEF_TYPE, USE_FROM_PTR, and vect_double_reduction_def.

Referenced by vect_analyze_scalar_cycles_1().

◆ vect_is_emulated_mixed_dot_prod()

static bool vect_is_emulated_mixed_dot_prod ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info )
static
Return true if (a) STMT_INFO is a DOT_PROD_EXPR reduction whose
multiplication operands have differing signs and (b) we intend
to emulate the operation using a series of signed DOT_PROD_EXPRs.
See vect_emulate_mixed_dot_prod for the actual sequence used.   

References directly_supported_p(), gcc_assert, ggc_alloc(), gimple_assign_rhs1(), gimple_assign_rhs2(), gimple_assign_rhs_code(), info_for_reduction(), optab_vector_mixed_sign, STMT_VINFO_REDUC_VECTYPE_IN, TREE_TYPE, and TYPE_SIGN.

Referenced by vect_model_reduction_cost(), vect_transform_reduction(), and vectorizable_reduction().

◆ vect_is_nonlinear_iv_evolution()

static bool vect_is_nonlinear_iv_evolution ( class loop * loop,
stmt_vec_info stmt_info,
gphi * loop_phi_node,
tree * init,
tree * step )
static

◆ vect_is_simple_iv_evolution()

static bool vect_is_simple_iv_evolution ( unsigned loop_nb,
tree access_fn,
tree * init,
tree * step )
static

◆ vect_is_simple_reduction()

static stmt_vec_info vect_is_simple_reduction ( loop_vec_info loop_info,
stmt_vec_info phi_info,
bool * double_reduc,
bool * reduc_chain_p,
bool slp )
static
Function vect_is_simple_reduction

  (1) Detect a cross-iteration def-use cycle that represents a simple
  reduction computation.  We look for the following pattern:

  loop_header:
    a1 = phi < a0, a2 >
    a3 = ...
    a2 = operation (a3, a1)

  or

  a3 = ...
  loop_header:
    a1 = phi < a0, a2 >
    a2 = operation (a3, a1)

  such that:
  1. operation is commutative and associative and it is safe to
     change the order of the computation
  2. no uses for a2 in the loop (a2 is used out of the loop)
  3. no uses of a1 in the loop besides the reduction operation
  4. no uses of a1 outside the loop.

  Conditions 1,4 are tested here.
  Conditions 2,3 are tested in vect_mark_stmts_to_be_vectorized.

  (2) Detect a cross-iteration def-use cycle in nested loops, i.e.,
  nested cycles.

  (3) Detect cycles of phi nodes in outer-loop vectorization, i.e., double
  reductions:

    a1 = phi < a0, a2 >
    inner loop (def of a3)
    a2 = phi < a3 >

  (4) Detect condition expressions, ie:
    for (int i = 0; i < N; i++)
      if (a[i] < val)
       ret_val = a[i];

References check_reduction_path(), gimple_match_op::code, COND_REDUCTION, CONVERT_EXPR_CODE_P, dump_enabled_p(), dump_printf_loc(), flow_bb_inside_loop_p(), flow_loop_nested_p(), FOR_EACH_IMM_USE_FAST, gcc_unreachable, ggc_alloc(), gimple_assign_rhs1_ptr(), gimple_bb(), gimple_call_arg_ptr(), gimple_extract_op(), gimple_get_lhs(), gimple_phi_num_args(), has_single_use(), has_zero_uses(), i, loop::inner, is_gimple_assign(), is_gimple_call(), is_gimple_debug(), loop_latch_edge(), LOOP_VINFO_LOOP, LOOP_VINFO_REDUCTION_CHAINS, MSG_MISSED_OPTIMIZATION, MSG_NOTE, NULL, path, PHI_ARG_DEF, PHI_ARG_DEF_FROM_EDGE, PHI_RESULT, REDUC_GROUP_FIRST_ELEMENT, REDUC_GROUP_NEXT_ELEMENT, REDUC_GROUP_SIZE, report_vect_op(), SSA_NAME_DEF_STMT, STMT_VINFO_DEF_TYPE, STMT_VINFO_REDUC_CODE, STMT_VINFO_REDUC_IDX, STMT_VINFO_REDUC_TYPE, TREE_CODE, TREE_CODE_REDUCTION, USE_STMT, vect_double_reduction_def, and vect_location.

Referenced by vect_analyze_scalar_cycles_1().

◆ vect_iv_limit_for_partial_vectors()

widest_int vect_iv_limit_for_partial_vectors ( loop_vec_info loop_vinfo)
Decide whether it is possible to use a zero-based induction variable
when vectorizing LOOP_VINFO with partial vectors.  If it is, return
the value that the induction variable must be able to hold in order
to ensure that the rgroups eventually have no active vector elements.
Return -1 otherwise.   

References ggc_alloc(), LOOP_VINFO_LOOP, LOOP_VINFO_MASK_SKIP_NITERS, LOOP_VINFO_PEELING_FOR_ALIGNMENT, LOOP_VINFO_VECT_FACTOR, max_loop_iterations(), wi::to_widest(), TREE_CODE, and vect_max_vf().

Referenced by vect_rgroup_iv_might_wrap_p(), vect_verify_full_masking(), and vect_verify_full_masking_avx512().

◆ vect_joust_loop_vinfos()

static bool vect_joust_loop_vinfos ( loop_vec_info new_loop_vinfo,
loop_vec_info old_loop_vinfo )
static
Decide whether to replace OLD_LOOP_VINFO with NEW_LOOP_VINFO.  Return
true if we should.   

References dump_enabled_p(), dump_printf_loc(), GET_MODE_NAME, ggc_alloc(), MSG_NOTE, vect_better_loop_vinfo_p(), and vect_location.

Referenced by vect_analyze_loop().

◆ vect_known_niters_smaller_than_vf()

static bool vect_known_niters_smaller_than_vf ( loop_vec_info loop_vinfo)
static
Return true if we know that the iteration count is smaller than the
vectorization factor.  Return false if it isn't, or if we can't be sure
either way.   

References ggc_alloc(), LOOP_VINFO_INT_NITERS, LOOP_VINFO_LOOP, LOOP_VINFO_NITERS_KNOWN_P, max_stmt_executions_int(), and vect_vf_for_cost().

Referenced by vect_analyze_loop_costing(), vect_determine_partial_vectors_and_peeling(), and vect_estimate_min_profitable_iters().

◆ vect_loop_kill_debug_uses()

◆ vect_min_prec_for_max_niters()

static unsigned vect_min_prec_for_max_niters ( loop_vec_info loop_vinfo,
unsigned int factor )
static
Calculate the minimum precision necessary to represent:

   MAX_NITERS * FACTOR

as an unsigned integer, where MAX_NITERS is the maximum number of
loop header iterations for the original scalar form of LOOP_VINFO.   

References ggc_alloc(), LOOP_VINFO_LOOP, LOOP_VINFO_NITERSM1, max_loop_iterations(), wi::min_precision(), wi::smin(), wi::to_widest(), TREE_TYPE, TYPE_MAX_VALUE, and UNSIGNED.

Referenced by vect_verify_full_masking(), and vect_verify_loop_lens().

◆ vect_model_reduction_cost()

static void vect_model_reduction_cost ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info,
internal_fn reduc_fn,
vect_reduction_type reduction_type,
int ncopies,
stmt_vector_for_cost * cost_vec )
static
TODO: Close dependency between vect_model_*_cost and vectorizable_*
functions. Design better to avoid maintenance issues.   
Function vect_model_reduction_cost.

Models cost for a reduction operation, including the vector ops
generated within the strip-mine loop in some cases, the initial
definition before the loop, and the epilogue code that must be generated.   

References gimple_match_op::code, COND_REDUCTION, directly_supported_p(), dump_enabled_p(), dump_printf(), exact_log2(), EXTRACT_LAST_REDUCTION, FOLD_LEFT_REDUCTION, gcc_unreachable, ggc_alloc(), gimple_extract_op(), have_whole_vector_shift(), IFN_LAST, LOOP_VINFO_LOOP, MSG_NOTE, nested_in_vect_loop_p(), NULL, record_stmt_cost(), scalar_stmt, scalar_to_vec, STMT_VINFO_VECTYPE, tree_to_uhwi(), gimple_match_op::type, TYPE_MODE, TYPE_SIZE, vec_to_scalar, vect_body, vect_epilogue, vect_is_emulated_mixed_dot_prod(), vect_nunits_for_cost(), vect_orig_stmt(), vect_prologue, VECTOR_MODE_P, and vector_stmt.

Referenced by vectorizable_reduction().

◆ vect_need_peeling_or_partial_vectors_p()

◆ vect_peel_nonlinear_iv_init()

◆ vect_phi_first_order_recurrence_p()

static bool vect_phi_first_order_recurrence_p ( loop_vec_info loop_vinfo,
class loop * loop,
gphi * phi )
static
Returns true if Phi is a first-order recurrence. A first-order
recurrence is a non-reduction recurrence relation in which the value of
the recurrence in the current loop iteration equals a value defined in
the previous iteration.   

References flow_bb_inside_loop_p(), FOR_EACH_IMM_USE_FAST, get_vectype_for_scalar_type(), ggc_alloc(), gimple_bb(), gimple_phi_result(), is_gimple_debug(), loop::latch, loop_latch_edge(), LOOP_VINFO_LOOP, PHI_ARG_DEF_FROM_EDGE, SSA_NAME_DEF_STMT, SSA_NAME_IS_DEFAULT_DEF, TREE_CODE, TREE_TYPE, USE_STMT, and vect_stmt_dominates_stmt_p().

Referenced by vect_analyze_scalar_cycles_1().

◆ vect_record_loop_len()

void vect_record_loop_len ( loop_vec_info loop_vinfo,
vec_loop_lens * lens,
unsigned int nvectors,
tree vectype,
unsigned int factor )
Record that LOOP_VINFO would need LENS to contain a sequence of NVECTORS
lengths for controlling an operation on VECTYPE.  The operation splits
each element of VECTYPE into FACTOR separate subelements, measuring the
length as a number of these subelements.   

References gcc_assert, ggc_alloc(), LOOP_VINFO_VECT_FACTOR, and TYPE_VECTOR_SUBPARTS().

Referenced by check_load_store_for_partial_vectors(), vectorizable_call(), vectorizable_condition(), vectorizable_live_operation(), vectorizable_operation(), and vectorizable_reduction().

◆ vect_record_loop_mask()

void vect_record_loop_mask ( loop_vec_info loop_vinfo,
vec_loop_masks * masks,
unsigned int nvectors,
tree vectype,
tree scalar_mask )
Record that a fully-masked version of LOOP_VINFO would need MASKS to
contain a sequence of NVECTORS masks that each control a vector of type
VECTYPE.  If SCALAR_MASK is nonnull, the fully-masked loop would AND
these vector masks with the vector version of SCALAR_MASK.   

References hash_set< KeyId, Lazy, Traits >::add(), gcc_assert, ggc_alloc(), vec_loop_masks::mask_set, and _loop_vec_info::scalar_cond_masked_set.

Referenced by check_load_store_for_partial_vectors(), vectorizable_call(), vectorizable_condition(), vectorizable_early_exit(), vectorizable_live_operation(), vectorizable_operation(), vectorizable_reduction(), and vectorizable_simd_clone_call().

◆ vect_rgroup_iv_might_wrap_p()

bool vect_rgroup_iv_might_wrap_p ( loop_vec_info loop_vinfo,
rgroup_controls * rgc )
For the given rgroup_controls RGC, check whether an induction variable
would ever hit a value that produces a set of all-false masks or zero
lengths before wrapping around.  Return true if it's possible to wrap
around before hitting the desirable value, otherwise return false.   

References ggc_alloc(), LOOP_VINFO_RGROUP_COMPARE_TYPE, wi::min_precision(), TYPE_PRECISION, UNSIGNED, and vect_iv_limit_for_partial_vectors().

Referenced by vect_estimate_min_profitable_iters(), and vect_set_loop_condition_partial_vectors().

◆ vect_transform_cycle_phi()

bool vect_transform_cycle_phi ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info,
gimple ** vec_stmt,
slp_tree slp_node,
slp_instance slp_node_instance )
Transform phase of a cycle PHI.   

References add_phi_arg(), build_vector_from_val(), build_vector_type(), build_vector_type_for_mode(), COND_REDUCTION, CONST_COND_REDUCTION, create_phi_node(), EXTRACT_LAST_REDUCTION, FOLD_LEFT_REDUCTION, gcc_assert, get_initial_def_for_reduction(), get_initial_defs_for_reduction(), ggc_alloc(), gimple_convert(), gimple_phi_result(), GSI_CONTINUE_LINKING, gsi_end_p(), gsi_insert_seq_after(), gsi_insert_seq_on_edge_immediate(), gsi_last_bb(), gsi_prev(), gsi_stmt(), loop::header, i, info_for_reduction(), loop::inner, INTEGER_INDUC_COND_REDUCTION, integer_zerop(), loop_preheader_edge(), LOOP_VINFO_LOOP, LOOP_VINFO_VECT_FACTOR, _loop_vec_info::main_loop_edge, nested_in_vect_loop_p(), neutral_op_for_reduction(), NULL, NULL_TREE, num_phis(), operand_equal_p(), REDUC_GROUP_FIRST_ELEMENT, SLP_TREE_CHILDREN, SLP_TREE_LANES, SLP_TREE_SCALAR_STMTS, stmt_ends_bb_p(), STMT_VINFO_DEF_TYPE, STMT_VINFO_FORCE_SINGLE_CYCLE, STMT_VINFO_REDUC_CODE, STMT_VINFO_REDUC_DEF, STMT_VINFO_REDUC_EPILOGUE_ADJUSTMENT, STMT_VINFO_REDUC_TYPE, STMT_VINFO_REDUC_VECTYPE_IN, STMT_VINFO_VEC_INDUC_COND_INITIAL_VAL, STMT_VINFO_VEC_STMTS, STMT_VINFO_VECTYPE, TREE_CODE, tree_int_cst_lt(), TREE_TYPE, TYPE_MODE, TYPE_VECTOR_SUBPARTS(), UNKNOWN_LOCATION, useless_type_conversion_p(), vect_create_destination_var(), vect_create_partial_epilog(), vect_find_reusable_accumulator(), vect_get_main_loop_result(), vect_get_num_copies(), vect_get_num_vectors(), vect_get_slp_defs(), vect_get_vec_defs_for_operand(), vect_phi_initial_value(), vect_reduction_def, and vect_stmt_to_vectorize().

Referenced by vect_transform_stmt().

◆ vect_transform_loop()

class loop * vect_transform_loop ( loop_vec_info loop_vinfo,
gimple * loop_vectorized_call )
Function vect_transform_loop.

The analysis phase has determined that the loop is vectorizable.
Vectorize the loop - created vectorized stmts to replace the scalar
stmts in the loop, and update the loop exit condition.
Returns scalar epilogue loop if any.   

References advance(), loop::any_estimate, loop::any_likely_upper_bound, loop::any_upper_bound, build_int_cst(), build_one_cst(), build_zero_cst(), vec_info_shared::check_datarefs(), conditional_internal_fn_code(), DR_GROUP_FIRST_ELEMENT, dump_enabled_p(), dump_printf(), dump_printf_loc(), DUMP_VECT_SCOPE, EDGE_COUNT, FOR_EACH_VEC_ELT, loop::force_vectorize, gcc_assert, GET_MODE_NAME, ggc_alloc(), gimple_build_assign(), gimple_call_arg(), gimple_call_internal_fn(), gimple_call_internal_p(), gimple_call_num_args(), gimple_clobber_p(), gimple_get_lhs(), gsi_end_p(), gsi_next(), gsi_remove(), gsi_replace(), gsi_start(), gsi_start_bb(), gsi_start_phis(), gsi_stmt(), loop::header, i, loop::inner, integer_onep(), poly_int< N, C >::is_constant(), known_eq, vec_info::lookup_stmt(), loop_niters_no_overflow(), loop_preheader_edge(), LOOP_REQUIRES_VERSIONING, LOOP_VINFO_BBS, LOOP_VINFO_COST_MODEL_THRESHOLD, LOOP_VINFO_EARLY_BREAKS, LOOP_VINFO_EPILOGUE_P, LOOP_VINFO_INT_NITERS, LOOP_VINFO_IV_EXIT, LOOP_VINFO_LOOP, LOOP_VINFO_NITERS, LOOP_VINFO_NITERS_KNOWN_P, LOOP_VINFO_NITERS_UNCHANGED, LOOP_VINFO_NITERSM1, LOOP_VINFO_ORIG_LOOP_INFO, LOOP_VINFO_PEELING_FOR_ALIGNMENT, LOOP_VINFO_PEELING_FOR_GAPS, LOOP_VINFO_SCALAR_IV_EXIT, LOOP_VINFO_SCALAR_LOOP, LOOP_VINFO_SCALAR_LOOP_SCALING, LOOP_VINFO_SLP_INSTANCES, LOOP_VINFO_USING_PARTIAL_VECTORS_P, LOOP_VINFO_VECT_FACTOR, LOOP_VINFO_VERSIONING_THRESHOLD, MAY_HAVE_DEBUG_BIND_STMTS, maybe_flat_loop_profile(), maybe_set_vectorized_backedge_value(), move_early_exit_stmts(), MSG_NOTE, loop::nb_iterations_estimate, loop::nb_iterations_likely_upper_bound, loop::nb_iterations_upper_bound, NULL, NULL_TREE, loop::num_nodes, basic_block_def::preds, PURE_SLP_STMT, release_defs(), vec_info::remove_stmt(), loop::safelen, scale_loop_frequencies(), scale_profile_for_vect_loop(), vec_info::shared, si, loop::simduid, single_pred_p(), vec_info::slp_instances, split_edge(), split_loop_exit_edge(), STMT_VINFO_DEF_TYPE, STMT_VINFO_GROUPED_ACCESS, STMT_VINFO_IN_PATTERN_P, STMT_VINFO_LIVE_P, STMT_VINFO_PATTERN_DEF_SEQ, STMT_VINFO_RELATED_STMT, STMT_VINFO_RELEVANT_P, STMT_VINFO_VECTYPE, TREE_TYPE, TYPE_VECTOR_SUBPARTS(), wi::udiv_ceil(), wi::udiv_floor(), wi::umin(), unlink_stmt_vdef(), loop::unroll, unshare_expr(), update_epilogue_loop_vinfo(), vect_apply_runtime_profitability_check_p(), vect_build_loop_niters(), vect_do_peeling(), vect_double_reduction_def, vect_first_order_recurrence, vect_free_slp_instance(), vect_gen_vector_loop_niters(), vect_induction_def, vect_internal_def, vect_location, vect_loop_kill_debug_uses(), vect_loop_versioning(), vect_nested_cycle, vect_prepare_for_masked_peels(), vect_reduction_def, vect_remove_stores(), vect_schedule_slp(), vect_set_loop_condition(), vect_transform_loop_stmt(), vect_transform_stmt(), vect_use_loop_mask_for_alignment_p(), vect_vf_for_cost(), vec_info::vector_mode, and VECTOR_TYPE_P.

Referenced by vect_transform_loops().

◆ vect_transform_loop_stmt()

static bool vect_transform_loop_stmt ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info,
gimple_stmt_iterator * gsi,
stmt_vec_info * seen_store )
static

◆ vect_transform_reduction()

bool vect_transform_reduction ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info,
gimple_stmt_iterator * gsi,
gimple ** vec_stmt,
slp_tree slp_node )
Transform the definition stmt STMT_INFO of a reduction PHI backedge
value.   

References build_vect_cond_expr(), canonicalize_code(), gimple_match_op::code, commutative_binary_op_p(), conditional_internal_fn_code(), dump_enabled_p(), dump_printf_loc(), FOLD_LEFT_REDUCTION, FOR_EACH_VEC_ELT, gcc_assert, gcc_unreachable, get_conditional_internal_fn(), ggc_alloc(), gimple_build_assign(), gimple_build_call_internal(), gimple_call_set_lhs(), gimple_call_set_nothrow(), gimple_extract_op(), gimple_get_lhs(), gimple_set_lhs(), i, info_for_reduction(), loop::inner, internal_fn_else_index(), code_helper::is_internal_fn(), code_helper::is_tree_code(), LOOP_VINFO_FULLY_MASKED_P, LOOP_VINFO_LENS, LOOP_VINFO_LOOP, LOOP_VINFO_MASKS, make_ssa_name(), MSG_NOTE, nested_in_vect_loop_p(), NULL_TREE, gimple_match_op::num_ops, gimple_match_op::ops, SLP_TREE_NUMBER_OF_VEC_STMTS, STMT_VINFO_DEF_TYPE, STMT_VINFO_FORCE_SINGLE_CYCLE, STMT_VINFO_REDUC_DEF, STMT_VINFO_REDUC_FN, STMT_VINFO_REDUC_IDX, STMT_VINFO_REDUC_TYPE, STMT_VINFO_REDUC_VECTYPE_IN, STMT_VINFO_VEC_STMTS, STMT_VINFO_VECTYPE, truth_type_for(), gimple_match_op::type, use_mask_by_cond_expr_p(), vect_create_destination_var(), vect_double_reduction_def, vect_emulate_mixed_dot_prod(), vect_finish_stmt_generation(), vect_get_loop_mask(), vect_get_num_copies(), vect_get_vec_defs(), vect_get_vec_defs_for_operand(), vect_is_emulated_mixed_dot_prod(), vect_location, vect_orig_stmt(), and vectorize_fold_left_reduction().

Referenced by vect_transform_stmt().

◆ vect_update_nonlinear_iv()

static tree vect_update_nonlinear_iv ( gimple_seq * stmts,
tree vectype,
tree induc_def,
tree vec_step,
enum vect_induction_op_type induction_type )
static

◆ vect_update_vf_for_slp()

◆ vect_verify_full_masking()

◆ vect_verify_full_masking_avx512()

◆ vect_verify_loop_lens()

◆ vectorizable_induction()

bool vectorizable_induction ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info,
gimple ** vec_stmt,
slp_tree slp_node,
stmt_vector_for_cost * cost_vec )
Function vectorizable_induction

Check if STMT_INFO performs an induction computation that can be vectorized.
If VEC_STMT is also passed, vectorize the induction PHI: create a vectorized
phi to replace it, put it in VEC_STMT, and add it to the same basic block.
Return true if STMT_INFO is vectorizable in this way.   

References add_phi_arg(), build1(), build_index_vector(), build_int_cst(), build_int_cstu(), build_real_from_wide(), build_vector_from_val(), CONSTANT_CLASS_P, create_phi_node(), directly_supported_p(), dump_enabled_p(), dump_printf_loc(), DUMP_VECT_SCOPE, expr, FLOAT_TYPE_P, flow_bb_inside_loop_p(), fold_convert, FOR_EACH_IMM_USE_FAST, FOR_EACH_VEC_ELT, force_gimple_operand(), gcc_assert, get_same_sized_vectype(), ggc_alloc(), gimple_assign_lhs(), gimple_bb(), gimple_build(), gimple_build_assign(), gimple_build_vector(), gimple_build_vector_from_val(), gimple_convert(), gimple_get_lhs(), gimple_phi_arg_def(), gsi_after_labels(), GSI_CONTINUE_LINKING, gsi_for_stmt(), gsi_insert_on_edge_immediate(), gsi_insert_seq_after(), gsi_insert_seq_before(), gsi_insert_seq_on_edge_immediate(), GSI_SAME_STMT, i, induc_vec_info_type, init_expr(), loop::inner, integer_type_node, integer_zerop(), INTEGRAL_TYPE_P, is_gimple_debug(), least_common_multiple(), vec_info::lookup_stmt(), loop_latch_edge(), loop_preheader_edge(), LOOP_VINFO_LENS, LOOP_VINFO_LOOP, LOOP_VINFO_MASK_SKIP_NITERS, LOOP_VINFO_USING_SELECT_VL_P, LOOP_VINFO_VECT_FACTOR, MSG_MISSED_OPTIMIZATION, MSG_NOTE, nested_in_vect_loop_p(), NULL, NULL_TREE, PHI_ARG_DEF_FROM_EDGE, PHI_RESULT, record_stmt_cost(), SCALAR_FLOAT_TYPE_P, scalar_to_vec, si, SLP_TREE_CHILDREN, SLP_TREE_LANES, SLP_TREE_NUMBER_OF_VEC_STMTS, SLP_TREE_SCALAR_STMTS, SLP_TREE_VEC_DEFS, SLP_TREE_VECTYPE, SSA_NAME_DEF_STMT, STMT_VINFO_DEF_TYPE, STMT_VINFO_LIVE_P, STMT_VINFO_LOOP_PHI_EVOLUTION_PART, STMT_VINFO_LOOP_PHI_EVOLUTION_TYPE, STMT_VINFO_RELEVANT_P, STMT_VINFO_TYPE, STMT_VINFO_VEC_STMTS, STMT_VINFO_VECTYPE, TREE_CODE, TREE_TYPE, type_has_mode_precision_p(), TYPE_VECTOR_SUBPARTS(), UNKNOWN_LOCATION, unshare_expr(), UNSIGNED, USE_STMT, useless_type_conversion_p(), vect_body, vect_get_loop_len(), vect_get_new_ssa_name(), vect_get_new_vect_var(), vect_get_num_copies(), vect_get_slp_vect_def(), vect_get_vec_defs_for_operand(), vect_induction_def, vect_init_vector(), vect_location, vect_maybe_update_slp_op_vectype(), vect_phi_initial_value(), vect_prologue, vect_simple_var, vect_step_op_add, vector_stmt, and vectorizable_nonlinear_induction().

Referenced by vect_analyze_loop_operations(), vect_analyze_stmt(), and vect_transform_stmt().

◆ vectorizable_lc_phi()

◆ vectorizable_live_operation()

bool vectorizable_live_operation ( vec_info * vinfo,
stmt_vec_info stmt_info,
slp_tree slp_node,
slp_instance slp_node_instance,
int slp_index,
bool vec_stmt_p,
stmt_vector_for_cost * cost_vec )
Function vectorizable_live_operation.

STMT_INFO computes a value that is used outside the loop.  Check if
it can be supported.   

References bitsize_int, build3(), build_nonstandard_integer_type(), build_zero_cst(), can_vec_extract_var_idx_p(), direct_internal_fn_supported_p(), dump_enabled_p(), dump_printf_loc(), EXTRACT_LAST_REDUCTION, flow_bb_inside_loop_p(), flow_loop_nested_p(), fold_convert, FOLD_LEFT_REDUCTION, FOR_EACH_IMM_USE_ON_STMT, FOR_EACH_IMM_USE_STMT, force_gimple_operand(), gcc_assert, get_loop_exit_edges(), ggc_alloc(), gimple_bb(), gimple_build_assign(), gimple_get_lhs(), gimple_phi_arg_edge(), gimple_phi_result(), gsi_after_labels(), gsi_for_stmt(), gsi_insert_before(), gsi_insert_seq_after(), gsi_insert_seq_before(), GSI_SAME_STMT, info_for_reduction(), int_const_binop(), is_gimple_debug(), is_simple_and_all_uses_invariant(), vec_info::lookup_stmt(), loop_exit_edge_p(), LOOP_VINFO_CAN_USE_PARTIAL_VECTORS_P, LOOP_VINFO_EARLY_BREAKS, LOOP_VINFO_EARLY_BREAKS_VECT_PEELED, LOOP_VINFO_FULLY_MASKED_P, LOOP_VINFO_FULLY_WITH_LENGTH_P, LOOP_VINFO_IV_EXIT, LOOP_VINFO_LENS, LOOP_VINFO_LOOP, LOOP_VINFO_MASKS, MSG_MISSED_OPTIMIZATION, MSG_NOTE, NULL, NULL_TREE, OPTIMIZE_FOR_SPEED, phi_arg_index_from_use(), PURE_SLP_STMT, record_stmt_cost(), REDUC_GROUP_FIRST_ELEMENT, remove_phi_node(), SET_USE, si, SLP_TREE_LANES, SLP_TREE_NUMBER_OF_VEC_STMTS, SLP_TREE_VEC_DEFS, SLP_TREE_VECTYPE, SSA_NAME_DEF_STMT, SSA_NAME_IS_DEFAULT_DEF, SSA_NAME_OCCURS_IN_ABNORMAL_PHI, STMT_VINFO_DEF_TYPE, STMT_VINFO_LIVE_P, STMT_VINFO_REDUC_DEF, STMT_VINFO_REDUC_TYPE, STMT_VINFO_RELEVANT_P, STMT_VINFO_VEC_STMTS, STMT_VINFO_VECTYPE, TREE_CODE, tree_to_uhwi(), TREE_TYPE, TYPE_MODE, TYPE_VECTOR_SUBPARTS(), update_stmt(), vec_to_scalar, vect_create_epilog_for_reduction(), vect_epilogue, vect_get_num_copies(), vect_induction_def, vect_location, vect_orig_stmt(), vect_record_loop_len(), vect_record_loop_mask(), vect_stmt_dominates_stmt_p(), vect_stmt_to_vectorize(), VECTOR_BOOLEAN_TYPE_P, vector_element_bits_tree(), and vectorizable_live_operation_1().

Referenced by can_vectorize_live_stmts(), vect_analyze_loop_operations(), vect_bb_slp_mark_live_stmts(), vect_schedule_slp_node(), and vect_slp_analyze_node_operations_1().

◆ vectorizable_live_operation_1()

◆ vectorizable_nonlinear_induction()

static bool vectorizable_nonlinear_induction ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info,
gimple ** vec_stmt,
slp_tree slp_node,
stmt_vector_for_cost * cost_vec )
static
Function vectorizable_induction

Check if STMT_INFO performs an nonlinear induction computation that can be
vectorized. If VEC_STMT is also passed, vectorize the induction PHI: create
a vectorized phi to replace it, put it in VEC_STMT, and add it to the same
basic block.
Return true if STMT_INFO is vectorizable in this way.   

References add_phi_arg(), can_vec_perm_const_p(), create_phi_node(), directly_supported_p(), dump_enabled_p(), dump_printf_loc(), DUMP_VECT_SCOPE, fold_convert, gcc_assert, gcc_unreachable, ggc_alloc(), gimple_bb(), gsi_after_labels(), gsi_insert_seq_before(), gsi_insert_seq_on_edge_immediate(), GSI_SAME_STMT, i, induc_vec_info_type, init_expr(), INTEGRAL_TYPE_P, poly_int< N, C >::is_constant(), loop_latch_edge(), loop_preheader_edge(), LOOP_VINFO_LOOP, LOOP_VINFO_MASK_SKIP_NITERS, LOOP_VINFO_VECT_FACTOR, maybe_ge, MSG_MISSED_OPTIMIZATION, MSG_NOTE, nested_in_vect_loop_p(), NULL, NULL_TREE, optab_vector, PHI_RESULT, record_stmt_cost(), scalar_to_vec, si, SSA_NAME_DEF_STMT, STMT_VINFO_LOOP_PHI_EVOLUTION_PART, STMT_VINFO_LOOP_PHI_EVOLUTION_TYPE, STMT_VINFO_TYPE, STMT_VINFO_VEC_STMTS, STMT_VINFO_VECTYPE, TREE_CODE, tree_fits_uhwi_p(), tree_nop_conversion_p(), tree_to_uhwi(), TREE_TYPE, TYPE_MODE, TYPE_PRECISION, TYPE_VECTOR_SUBPARTS(), UNKNOWN_LOCATION, vect_body, vect_create_nonlinear_iv_init(), vect_create_nonlinear_iv_step(), vect_create_nonlinear_iv_vec_step(), vect_get_new_vect_var(), vect_get_num_copies(), vect_location, vect_peel_nonlinear_iv_init(), vect_phi_initial_value(), vect_prologue, vect_simple_var, vect_step_op_add, vect_step_op_mul, vect_step_op_neg, vect_step_op_shl, vect_step_op_shr, vect_update_nonlinear_iv(), and vector_stmt.

Referenced by vectorizable_induction().

◆ vectorizable_phi()

◆ vectorizable_recurr()

bool vectorizable_recurr ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info,
gimple ** vec_stmt,
slp_tree slp_node,
stmt_vector_for_cost * cost_vec )
Vectorizes first order recurrences.  An overview of the transformation
is described below. Suppose we have the following loop.

  int t = 0;
  for (int i = 0; i < n; ++i)
    {
      b[i] = a[i] - t;
      t = a[i];
    }

There is a first-order recurrence on 'a'. For this loop, the scalar IR
looks (simplified) like:

 scalar.preheader:
   init = 0;

 scalar.body:
   i = PHI <0(scalar.preheader), i+1(scalar.body)>
   _2 = PHI <(init(scalar.preheader), <_1(scalar.body)>
   _1 = a[i]
   b[i] = _1 - _2
   if (i < n) goto scalar.body

In this example, _2 is a recurrence because it's value depends on the
previous iteration.  We vectorize this as (VF = 4)

 vector.preheader:
   vect_init = vect_cst(..., ..., ..., 0)

 vector.body
   i = PHI <0(vector.preheader), i+4(vector.body)>
   vect_1 = PHI <vect_init(vector.preheader), v2(vector.body)>
   vect_2 = a[i, i+1, i+2, i+3];
   vect_3 = vec_perm (vect_1, vect_2, { 3, 4, 5, 6 })
   b[i, i+1, i+2, i+3] = vect_2 - vect_3
   if (..) goto vector.body

In this function, vectorizable_recurr, we code generate both the
vector PHI node and the permute since those together compute the
vectorized value of the scalar PHI.  We do not yet have the
backedge value to fill in there nor into the vec_perm.  Those
are filled in maybe_set_vectorized_backedge_value and
vect_schedule_scc.

TODO:  Since the scalar loop does not have a use of the recurrence
outside of the loop the natural way to implement peeling via
vectorizing the live value doesn't work.  For now peeling of loops
with a recurrence is not implemented.  For SLP the supported cases
are restricted to those requiring a single vector recurrence PHI.   

References add_phi_arg(), build_vector_from_val(), can_vec_perm_const_p(), create_phi_node(), dump_enabled_p(), dump_printf_loc(), FOR_EACH_VEC_ELT, ggc_alloc(), gimple_bb(), gimple_build_assign(), gimple_convert(), gimple_phi_result(), gsi_for_stmt(), gsi_insert_seq_on_edge_immediate(), gsi_next(), i, vec_info::lookup_def(), loop_latch_edge(), loop_preheader_edge(), LOOP_VINFO_LOOP, make_ssa_name(), maybe_gt, MSG_MISSED_OPTIMIZATION, MSG_NOTE, NULL, NULL_TREE, PHI_ARG_DEF_FROM_EDGE, record_stmt_cost(), recurr_info_type, scalar_to_vec, SLP_TREE_CHILDREN, SLP_TREE_LANES, SLP_TREE_NUMBER_OF_VEC_STMTS, SLP_TREE_VECTYPE, SSA_NAME_DEF_STMT, STMT_VINFO_DEF_TYPE, STMT_VINFO_TYPE, STMT_VINFO_VEC_STMTS, STMT_VINFO_VECTYPE, TREE_CODE, TREE_TYPE, TYPE_MODE, TYPE_VECTOR_SUBPARTS(), types_compatible_p(), UNKNOWN_LOCATION, useless_type_conversion_p(), vect_body, vect_finish_stmt_generation(), vect_first_order_recurrence, vect_gen_perm_mask_checked(), vect_get_new_vect_var(), vect_get_num_copies(), vect_init_vector(), vect_location, vect_maybe_update_slp_op_vectype(), vect_prologue, vect_simple_var, vect_stmt_to_vectorize(), and vector_stmt.

Referenced by vect_analyze_loop_operations(), vect_analyze_stmt(), and vect_transform_stmt().

◆ vectorizable_reduction()

bool vectorizable_reduction ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info,
slp_tree slp_node,
slp_instance slp_node_instance,
stmt_vector_for_cost * cost_vec )
Function vectorizable_reduction.

Check if STMT_INFO performs a reduction operation that can be vectorized.
If VEC_STMT is also passed, vectorize STMT_INFO: create a vectorized
stmt to replace it, put it in VEC_STMT, and insert it at GSI.
Return true if STMT_INFO is vectorizable in this way.

This function also handles reduction idioms (patterns) that have been
recognized in advance during vect_pattern_recog.  In this case, STMT_INFO
may be of this form:
  X = pattern_expr (arg0, arg1, ..., X)
and its STMT_VINFO_RELATED_STMT points to the last stmt in the original
sequence that had been detected and replaced by the pattern-stmt
(STMT_INFO).

This function also handles reduction of condition expressions, for example:
  for (int i = 0; i < N; i++)
    if (a[i] < value)
      last = a[i];
This is handled by vectorising the loop and creating an additional vector
containing the loop indexes for which "a[i] < value" was true.  In the
function epilogue this is reduced to a single max value and then used to
index into the vector of results.

In some cases of reduction patterns, the type of the reduction variable X is
different than the type of the other arguments of STMT_INFO.
In such cases, the vectype that is used when transforming STMT_INFO into
a vector stmt is different than the vectype that is used to determine the
vectorization factor, because it consists of a different number of elements
than the actual number of elements that are being operated upon in parallel.

For example, consider an accumulation of shorts into an int accumulator.
On some targets it's possible to vectorize this pattern operating on 8
shorts at a time (hence, the vectype for purposes of determining the
vectorization factor should be V8HI); on the other hand, the vectype that
is used to create the vector form is actually V4SI (the type of the result).

Upon entry to this function, STMT_VINFO_VECTYPE records the vectype that
indicates what is the actual level of parallelism (V8HI in the example), so
that the right vectorization factor would be derived.  This vectype
corresponds to the type of arguments to the reduction stmt, and should *NOT*
be used to create the vectorized stmt.  The right vectype for the vectorized
stmt is obtained from the type of the result X:
   get_vectype_for_scalar_type (vinfo, TREE_TYPE (X))

This means that, contrary to "regular" reductions (or "regular" stmts in
general), the following equation:
   STMT_VINFO_VECTYPE == get_vectype_for_scalar_type (vinfo, TREE_TYPE (X))
does *NOT* necessarily hold for reduction patterns.   

References associative_binary_op_p(), boolean_type_node, build_int_cst(), can_duplicate_and_interleave_p(), gimple_match_op::code, commutative_binary_op_p(), COMPARISON_CLASS_P, COND_REDUCTION, conditional_internal_fn_code(), CONST_COND_REDUCTION, CONVERT_EXPR_CODE_P, cycle_phi_info_type, direct_internal_fn_supported_p(), directly_supported_p(), dump_enabled_p(), dump_printf(), dump_printf_loc(), expand_vec_cond_expr_p(), EXTRACT_LAST_REDUCTION, FLOAT_TYPE_P, fold_binary, FOLD_LEFT_REDUCTION, fold_left_reduction_fn(), FOR_EACH_VEC_ELT, gcc_assert, gcc_unreachable, get_conditional_internal_fn(), get_masked_reduction_fn(), GET_MODE_PRECISION(), GET_MODE_SIZE(), get_same_sized_vectype(), get_vectype_for_scalar_type(), wi::geu_p(), ggc_alloc(), gimple_bb(), gimple_extract_op(), gimple_phi_result(), HONOR_SIGN_DEPENDENT_ROUNDING(), i, IFN_LAST, info_for_reduction(), loop::inner, int_const_binop(), INTEGER_INDUC_COND_REDUCTION, integer_one_node, integer_onep(), integer_zerop(), INTEGRAL_TYPE_P, internal_fn_mask_index(), is_nonwrapping_integer_induction(), known_eq, vec_info::lookup_def(), vec_info::lookup_stmt(), loop_latch_edge(), LOOP_VINFO_CAN_USE_PARTIAL_VECTORS_P, LOOP_VINFO_LENS, LOOP_VINFO_LOOP, LOOP_VINFO_MASKS, LOOP_VINFO_VECT_FACTOR, make_unsigned_type(), max_loop_iterations(), MSG_MISSED_OPTIMIZATION, MSG_NOTE, needs_fold_left_reduction_p(), nested_in_vect_loop_p(), neutral_op_for_reduction(), NULL, NULL_TREE, gimple_match_op::num_ops, gimple_match_op::ops, optab_vector, OPTIMIZE_FOR_SPEED, PHI_ARG_DEF_FROM_EDGE, PHI_RESULT, POINTER_TYPE_P, record_stmt_cost(), REDUC_GROUP_FIRST_ELEMENT, REDUC_GROUP_NEXT_ELEMENT, reduc_vec_info_type, reduction_fn_for_scalar_code(), SCALAR_FLOAT_TYPE_P, SCALAR_TYPE_MODE, single_imm_use(), SLP_TREE_CHILDREN, SLP_TREE_LANES, SLP_TREE_NUMBER_OF_VEC_STMTS, SLP_TREE_SCALAR_STMTS, SLP_TREE_VECTYPE, STMT_VINFO_DEF_TYPE, STMT_VINFO_FORCE_SINGLE_CYCLE, STMT_VINFO_IN_PATTERN_P, STMT_VINFO_LIVE_P, STMT_VINFO_LOOP_PHI_EVOLUTION_BASE_UNCHANGED, STMT_VINFO_LOOP_PHI_EVOLUTION_PART, STMT_VINFO_REDUC_CODE, STMT_VINFO_REDUC_DEF, STMT_VINFO_REDUC_FN, STMT_VINFO_REDUC_IDX, STMT_VINFO_REDUC_TYPE, STMT_VINFO_REDUC_VECTYPE, STMT_VINFO_REDUC_VECTYPE_IN, STMT_VINFO_RELATED_STMT, STMT_VINFO_RELEVANT, STMT_VINFO_STMT, STMT_VINFO_TYPE, STMT_VINFO_VEC_INDUC_COND_INITIAL_VAL, STMT_VINFO_VECTYPE, _loop_vec_info::suggested_unroll_factor, wi::to_widest(), TREE_CODE, TREE_CODE_REDUCTION, tree_int_cst_lt(), tree_int_cst_sgn(), tree_nop_conversion_p(), TREE_TYPE, truth_type_for(), gimple_match_op::type, type_has_mode_precision_p(), TYPE_MAX_VALUE, TYPE_MIN_VALUE, TYPE_MODE, TYPE_VECTOR_SUBPARTS(), types_compatible_p(), use_mask_by_cond_expr_p(), vect_body, vect_can_vectorize_without_simd_p(), vect_constant_def, vect_double_reduction_def, vect_emulated_vector_p(), vect_get_num_copies(), vect_induction_def, vect_internal_def, vect_is_emulated_mixed_dot_prod(), vect_is_simple_use(), vect_location, vect_maybe_update_slp_op_vectype(), vect_model_reduction_cost(), vect_nested_cycle, vect_orig_stmt(), vect_phi_initial_value(), vect_record_loop_len(), vect_record_loop_mask(), vect_reduction_def, vect_stmt_to_vectorize(), vect_unknown_def_type, vect_unused_in_scope, vect_used_in_outer, vect_used_only_live, vector_stmt, and VECTORIZABLE_CYCLE_DEF.

Referenced by vect_analyze_loop_operations(), and vect_analyze_stmt().

◆ vectorize_fold_left_reduction()

static bool vectorize_fold_left_reduction ( loop_vec_info loop_vinfo,
stmt_vec_info stmt_info,
gimple_stmt_iterator * gsi,
gimple ** vec_stmt,
slp_tree slp_node,
gimple * reduc_def_stmt,
code_helper code,
internal_fn reduc_fn,
tree * ops,
int num_ops,
tree vectype_in,
int reduc_index,
vec_loop_masks * masks,
vec_loop_lens * lens )
static
Perform an in-order reduction (FOLD_LEFT_REDUCTION).  STMT_INFO is the
statement that sets the live-out value.  REDUC_DEF_STMT is the phi
statement.  CODE is the operation performed by STMT_INFO and OPS are
its scalar operands.  REDUC_INDEX is the index of the operand in
OPS that is set by REDUC_DEF_STMT.  REDUC_FN is the function that
implements in-order reduction, or IFN_LAST if we should open-code it.
VECTYPE_IN is the type of the vector input.  MASKS specifies the masks
that should be used to control the operation in a fully-masked loop.   

References binary_op, build_int_cst(), build_minus_one_cst(), build_zero_cst(), conditional_internal_fn_code(), const_unop(), dump_enabled_p(), dump_printf_loc(), FOR_EACH_VEC_ELT, gcc_assert, gcc_checking_assert, get_masked_reduction_fn(), ggc_alloc(), gimple_build_assign(), gimple_build_call_internal(), gimple_get_lhs(), gimple_phi_result(), gimple_set_lhs(), gsi_for_stmt(), gsi_insert_before(), gsi_remove(), GSI_SAME_STMT, HONOR_SIGN_DEPENDENT_ROUNDING(), HONOR_SIGNED_ZEROS(), i, IFN_LAST, intQI_type_node, code_helper::is_tree_code(), known_eq, LOOP_VINFO_FULLY_MASKED_P, LOOP_VINFO_FULLY_WITH_LENGTH_P, LOOP_VINFO_LOOP, LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS, make_ssa_name(), merge_with_identity(), MSG_MISSED_OPTIMIZATION, nested_in_vect_loop_p(), NULL, NULL_TREE, SLP_TREE_SCALAR_STMTS, SSA_NAME_DEF_STMT, STMT_VINFO_VEC_STMTS, STMT_VINFO_VECTYPE, TREE_CODE_LENGTH, TREE_TYPE, truth_type_for(), TYPE_VECTOR_SUBPARTS(), useless_type_conversion_p(), vect_create_destination_var(), vect_expand_fold_left(), vect_finish_replace_stmt(), vect_finish_stmt_generation(), vect_get_loop_len(), vect_get_loop_mask(), vect_get_num_copies(), vect_get_slp_defs(), vect_get_vec_defs_for_operand(), vect_location, and vect_orig_stmt().

Referenced by vect_transform_reduction().