LCOV - code coverage report
Current view: top level - gcc - tree-vectorizer.h (source / functions) Coverage Total Hit
Test: gcc.info Lines: 97.3 % 263 256
Test Date: 2026-05-11 19:44:49 Functions: 90.5 % 42 38
Legend: Lines:     hit not hit

            Line data    Source code
       1              : /* Vectorizer
       2              :    Copyright (C) 2003-2026 Free Software Foundation, Inc.
       3              :    Contributed by Dorit Naishlos <dorit@il.ibm.com>
       4              : 
       5              : This file is part of GCC.
       6              : 
       7              : GCC is free software; you can redistribute it and/or modify it under
       8              : the terms of the GNU General Public License as published by the Free
       9              : Software Foundation; either version 3, or (at your option) any later
      10              : version.
      11              : 
      12              : GCC is distributed in the hope that it will be useful, but WITHOUT ANY
      13              : WARRANTY; without even the implied warranty of MERCHANTABILITY or
      14              : FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
      15              : for more details.
      16              : 
      17              : You should have received a copy of the GNU General Public License
      18              : along with GCC; see the file COPYING3.  If not see
      19              : <http://www.gnu.org/licenses/>.  */
      20              : 
      21              : #ifndef GCC_TREE_VECTORIZER_H
      22              : #define GCC_TREE_VECTORIZER_H
      23              : 
      24              : typedef class _stmt_vec_info *stmt_vec_info;
      25              : typedef struct _slp_tree *slp_tree;
      26              : 
      27              : #include "tree-data-ref.h"
      28              : #include "tree-hash-traits.h"
      29              : #include "target.h"
      30              : #include "internal-fn.h"
      31              : #include "tree-ssa-operands.h"
      32              : #include "gimple-match.h"
      33              : #include "dominance.h"
      34              : 
      35              : /* Used for naming of new temporaries.  */
      36              : enum vect_var_kind {
      37              :   vect_simple_var,
      38              :   vect_pointer_var,
      39              :   vect_scalar_var,
      40              :   vect_mask_var
      41              : };
      42              : 
      43              : /* Defines type of operation.  */
      44              : enum operation_type {
      45              :   unary_op = 1,
      46              :   binary_op,
      47              :   ternary_op
      48              : };
      49              : 
      50              : /* Define type of available alignment support.  */
      51              : enum dr_alignment_support {
      52              :   dr_unaligned_unsupported,
      53              :   dr_unaligned_supported,
      54              :   dr_explicit_realign,
      55              :   dr_explicit_realign_optimized,
      56              :   dr_aligned
      57              : };
      58              : 
      59              : /* Define type of peeling support to indicate how peeling for alignment can help
      60              :    make vectorization supported.  */
      61              : enum peeling_support {
      62              :   peeling_known_supported,
      63              :   peeling_maybe_supported,
      64              :   peeling_unsupported
      65              : };
      66              : 
      67              : /* Define type of def-use cross-iteration cycle.  */
      68              : enum vect_def_type {
      69              :   vect_uninitialized_def = 0,
      70              :   vect_constant_def = 1,
      71              :   vect_external_def,
      72              :   vect_internal_def,
      73              :   vect_induction_def,
      74              :   vect_reduction_def,
      75              :   vect_double_reduction_def,
      76              :   vect_nested_cycle,
      77              :   vect_first_order_recurrence,
      78              :   vect_condition_def,
      79              :   vect_unknown_def_type
      80              : };
      81              : 
      82              : /* Define operation type of linear/non-linear induction variable.  */
      83              : enum vect_induction_op_type {
      84              :    vect_step_op_add = 0,
      85              :    vect_step_op_neg,
      86              :    vect_step_op_mul,
      87              :    vect_step_op_shl,
      88              :    vect_step_op_shr
      89              : };
      90              : 
      91              : /* Define type of reduction.  */
      92              : enum vect_reduction_type {
      93              :   TREE_CODE_REDUCTION,
      94              :   COND_REDUCTION,
      95              :   INTEGER_INDUC_COND_REDUCTION,
      96              :   CONST_COND_REDUCTION,
      97              : 
      98              :   /* Retain a scalar phi and use a FOLD_EXTRACT_LAST within the loop
      99              :      to implement:
     100              : 
     101              :        for (int i = 0; i < VF; ++i)
     102              :          res = cond[i] ? val[i] : res;  */
     103              :   EXTRACT_LAST_REDUCTION,
     104              : 
     105              :   /* Use a folding reduction within the loop to implement:
     106              : 
     107              :        for (int i = 0; i < VF; ++i)
     108              :          res = res OP val[i];
     109              : 
     110              :      (with no reassocation).  */
     111              :   FOLD_LEFT_REDUCTION
     112              : };
     113              : 
     114              : #define VECTORIZABLE_CYCLE_DEF(D) (((D) == vect_reduction_def)           \
     115              :                                    || ((D) == vect_double_reduction_def) \
     116              :                                    || ((D) == vect_nested_cycle))
     117              : 
     118              : /* Structure to encapsulate information about a group of like
     119              :    instructions to be presented to the target cost model.  */
     120              : struct stmt_info_for_cost {
     121              :   int count;
     122              :   enum vect_cost_for_stmt kind;
     123              :   enum vect_cost_model_location where;
     124              :   stmt_vec_info stmt_info;
     125              :   slp_tree node;
     126              :   tree vectype;
     127              :   int misalign;
     128              : };
     129              : 
     130              : typedef vec<stmt_info_for_cost> stmt_vector_for_cost;
     131              : 
     132              : /* Maps base addresses to an innermost_loop_behavior and the stmt it was
     133              :    derived from that gives the maximum known alignment for that base.  */
     134              : typedef hash_map<tree_operand_hash,
     135              :                  std::pair<stmt_vec_info, innermost_loop_behavior *> >
     136              :           vec_base_alignments;
     137              : 
     138              : /* Represents elements [START, START + LENGTH) of cyclical array OPS*
     139              :    (i.e. OPS repeated to give at least START + LENGTH elements)  */
     140              : struct vect_scalar_ops_slice
     141              : {
     142              :   tree op (unsigned int i) const;
     143              :   bool all_same_p () const;
     144              : 
     145              :   vec<tree> *ops;
     146              :   unsigned int start;
     147              :   unsigned int length;
     148              : };
     149              : 
     150              : /* Return element I of the slice.  */
     151              : inline tree
     152      2711904 : vect_scalar_ops_slice::op (unsigned int i) const
     153              : {
     154      5423808 :   return (*ops)[(i + start) % ops->length ()];
     155              : }
     156              : 
     157              : /* Hash traits for vect_scalar_ops_slice.  */
     158              : struct vect_scalar_ops_slice_hash : typed_noop_remove<vect_scalar_ops_slice>
     159              : {
     160              :   typedef vect_scalar_ops_slice value_type;
     161              :   typedef vect_scalar_ops_slice compare_type;
     162              : 
     163              :   static const bool empty_zero_p = true;
     164              : 
     165              :   static void mark_deleted (value_type &s) { s.length = ~0U; }
     166            0 :   static void mark_empty (value_type &s) { s.length = 0; }
     167       426092 :   static bool is_deleted (const value_type &s) { return s.length == ~0U; }
     168      4018329 :   static bool is_empty (const value_type &s) { return s.length == 0; }
     169              :   static hashval_t hash (const value_type &);
     170              :   static bool equal (const value_type &, const compare_type &);
     171              : };
     172              : 
     173              : /* Describes how we're going to vectorize an individual load or store,
     174              :    or a group of loads or stores.  */
     175              : enum vect_memory_access_type {
     176              :   VMAT_UNINITIALIZED,
     177              : 
     178              :   /* An access to an invariant address.  This is used only for loads.  */
     179              :   VMAT_INVARIANT,
     180              : 
     181              :   /* A simple contiguous access.  */
     182              :   VMAT_CONTIGUOUS,
     183              : 
     184              :   /* A contiguous access that goes down in memory rather than up,
     185              :      with no additional permutation.  This is used only for stores
     186              :      of invariants.  */
     187              :   VMAT_CONTIGUOUS_DOWN,
     188              : 
     189              :   /* A simple contiguous access in which the elements need to be reversed
     190              :      after loading or before storing.  */
     191              :   VMAT_CONTIGUOUS_REVERSE,
     192              : 
     193              :   /* An access that uses IFN_LOAD_LANES or IFN_STORE_LANES.  */
     194              :   VMAT_LOAD_STORE_LANES,
     195              : 
     196              :   /* An access in which each scalar element is loaded or stored
     197              :      individually.  */
     198              :   VMAT_ELEMENTWISE,
     199              : 
     200              :   /* A hybrid of VMAT_CONTIGUOUS and VMAT_ELEMENTWISE, used for grouped
     201              :      SLP accesses.  Each unrolled iteration uses a contiguous load
     202              :      or store for the whole group, but the groups from separate iterations
     203              :      are combined in the same way as for VMAT_ELEMENTWISE.  */
     204              :   VMAT_STRIDED_SLP,
     205              : 
     206              :   /* The access uses gather loads or scatter stores.  */
     207              :   VMAT_GATHER_SCATTER_LEGACY,
     208              :   VMAT_GATHER_SCATTER_IFN,
     209              :   VMAT_GATHER_SCATTER_EMULATED
     210              : };
     211              : 
     212              : /* Returns whether MAT is any of the VMAT_GATHER_SCATTER_* kinds.  */
     213              : 
     214              : inline bool
     215      6344942 : mat_gather_scatter_p (vect_memory_access_type mat)
     216              : {
     217      6344942 :   return (mat == VMAT_GATHER_SCATTER_LEGACY
     218              :           || mat == VMAT_GATHER_SCATTER_IFN
     219      6344942 :           || mat == VMAT_GATHER_SCATTER_EMULATED);
     220              : }
     221              : 
     222              : /*-----------------------------------------------------------------*/
     223              : /* Info on vectorized defs.                                        */
     224              : /*-----------------------------------------------------------------*/
     225              : enum stmt_vec_info_type {
     226              :   undef_vec_info_type = 0,
     227              :   load_vec_info_type,
     228              :   store_vec_info_type,
     229              :   shift_vec_info_type,
     230              :   op_vec_info_type,
     231              :   call_vec_info_type,
     232              :   call_simd_clone_vec_info_type,
     233              :   assignment_vec_info_type,
     234              :   condition_vec_info_type,
     235              :   comparison_vec_info_type,
     236              :   reduc_vec_info_type,
     237              :   induc_vec_info_type,
     238              :   type_promotion_vec_info_type,
     239              :   type_demotion_vec_info_type,
     240              :   type_conversion_vec_info_type,
     241              :   cycle_phi_info_type,
     242              :   lc_phi_info_type,
     243              :   phi_info_type,
     244              :   recurr_info_type,
     245              :   loop_exit_ctrl_vec_info_type,
     246              :   permute_info_type
     247              : };
     248              : 
     249              : /************************************************************************
     250              :   SLP
     251              :  ************************************************************************/
     252              : typedef vec<std::pair<unsigned, unsigned> > lane_permutation_t;
     253              : typedef auto_vec<std::pair<unsigned, unsigned>, 16> auto_lane_permutation_t;
     254              : typedef vec<unsigned> load_permutation_t;
     255              : typedef auto_vec<unsigned, 16> auto_load_permutation_t;
     256              : 
     257      3311228 : struct vect_data {
     258      2062446 :   virtual ~vect_data () = default;
     259              : };
     260              : 
     261              : /* Analysis data from vectorizable_simd_clone_call for
     262              :    call_simd_clone_vec_info_type.  */
     263              : struct vect_simd_clone_data : vect_data {
     264         1903 :   virtual ~vect_simd_clone_data () = default;
     265         1436 :   vect_simd_clone_data () = default;
     266          467 :   vect_simd_clone_data (vect_simd_clone_data &&other) = default;
     267              : 
     268              :   /* Selected SIMD clone and clone for in-branch.  */
     269              :   cgraph_node *clone;
     270              :   cgraph_node *clone_inbranch;
     271              : 
     272              :   /* Selected SIMD clone's function info.  First vector element
     273              :      is NULL_TREE, followed by a pair of trees (base + step)
     274              :      for linear arguments (pair of NULLs for other arguments).  */
     275              :   auto_vec<tree> simd_clone_info;
     276              : };
     277              : 
     278              : /* Analysis data from vectorizable_load and vectorizable_store for
     279              :    load_vec_info_type and store_vec_info_type.  */
     280              : struct vect_load_store_data : vect_data {
     281      1246879 :   vect_load_store_data (vect_load_store_data &&other) = default;
     282      2062446 :   vect_load_store_data () = default;
     283      3308684 :   virtual ~vect_load_store_data () = default;
     284              : 
     285              :   vect_memory_access_type memory_access_type;
     286              :   dr_alignment_support alignment_support_scheme;
     287              :   int misalignment;
     288              :   internal_fn lanes_ifn; // VMAT_LOAD_STORE_LANES
     289              :   poly_int64 poffset;
     290              :   union {
     291              :       internal_fn ifn;  // VMAT_GATHER_SCATTER_IFN
     292              :       tree decl;        // VMAT_GATHER_SCATTER_DECL
     293              :   } gs;
     294              :   tree strided_offset_vectype; // VMAT_GATHER_SCATTER_IFN, originally strided
     295              :   /* Load/store type with larger element mode used for punning the vectype.  */
     296              :   tree ls_type; // VMAT_GATHER_SCATTER_IFN
     297              :   /* This is set to a supported offset vector type if we don't support the
     298              :      originally requested offset type, otherwise NULL.
     299              :      If nonzero there will be an additional offset conversion before
     300              :      the gather/scatter.  */
     301              :   tree supported_offset_vectype; // VMAT_GATHER_SCATTER_IFN
     302              :   /* Similar for scale.  Only nonzero if we don't support the requested
     303              :      scale.  Then we need to multiply the offset vector before the
     304              :      gather/scatter.  */
     305              :   int supported_scale; // VMAT_GATHER_SCATTER_IFN
     306              :   auto_vec<int> elsvals;
     307              :   /* True if the load requires a load permutation.  */
     308              :   bool slp_perm;    // SLP_TREE_LOAD_PERMUTATION
     309              :   unsigned n_perms; // SLP_TREE_LOAD_PERMUTATION
     310              :   unsigned n_loads; // SLP_TREE_LOAD_PERMUTATION
     311              :   /* Whether the load permutation is consecutive and simple.  */
     312              :   bool subchain_p; // VMAT_STRIDED_SLP and VMAT_GATHER_SCATTER
     313              : };
     314              : 
     315              : /* A computation tree of an SLP instance.  Each node corresponds to a group of
     316              :    stmts to be packed in a SIMD stmt.  */
     317              : struct _slp_tree {
     318              :   _slp_tree ();
     319              :   ~_slp_tree ();
     320              : 
     321              :   void push_vec_def (gimple *def);
     322         8465 :   void push_vec_def (tree def) { vec_defs.quick_push (def); }
     323              : 
     324              :   /* Nodes that contain def-stmts of this node statements operands.  */
     325              :   vec<slp_tree> children;
     326              : 
     327              :   /* A group of scalar stmts to be vectorized together.  */
     328              :   vec<stmt_vec_info> stmts;
     329              :   /* A group of scalar operands to be vectorized together.  */
     330              :   vec<tree> ops;
     331              :   /* A set of lane indices that are live and to be code-generated from
     332              :      this SLP node.  */
     333              :   vec<unsigned> live_lanes;
     334              : 
     335              :   /* The representative that should be used for analysis and
     336              :      code generation.  */
     337              :   stmt_vec_info representative;
     338              : 
     339              :   struct {
     340              :       /* SLP cycle the node resides in, or -1.  */
     341              :       int id;
     342              :       /* The SLP operand index with the edge on the SLP cycle, or -1.  */
     343              :       int reduc_idx;
     344              :   } cycle_info;
     345              : 
     346              :   /* Load permutation relative to the stores, NULL if there is no
     347              :      permutation.  */
     348              :   load_permutation_t load_permutation;
     349              :   /* Lane permutation of the operands scalar lanes encoded as pairs
     350              :      of { operand number, lane number }.  The number of elements
     351              :      denotes the number of output lanes.  */
     352              :   lane_permutation_t lane_permutation;
     353              : 
     354              :   tree vectype;
     355              :   /* Vectorized defs.  */
     356              :   vec<tree> vec_defs;
     357              : 
     358              :   /* Reference count in the SLP graph.  */
     359              :   unsigned int refcnt;
     360              :   /* The maximum number of vector elements for the subtree rooted
     361              :      at this node.  */
     362              :   poly_uint64 max_nunits;
     363              :   /* The DEF type of this node.  */
     364              :   enum vect_def_type def_type;
     365              :   /* The number of scalar lanes produced by this node.  */
     366              :   unsigned int lanes;
     367              :   /* The operation of this node.  */
     368              :   enum tree_code code;
     369              :   /* For gather/scatter memory operations the scale each offset element
     370              :      should be multiplied by before being added to the base.  */
     371              :   int gs_scale;
     372              :   /* For gather/scatter memory operations the loop-invariant base value.  */
     373              :   tree gs_base;
     374              :   /* Whether uses of this load or feeders of this store are suitable
     375              :      for load/store-lanes.  */
     376              :   bool ldst_lanes;
     377              :   /* For BB vect, flag to indicate this load node should be vectorized
     378              :      as to avoid STLF fails because of related stores.  */
     379              :   bool avoid_stlf_fail;
     380              : 
     381              :   int vertex;
     382              : 
     383              :   /* The kind of operation as determined by analysis and optional
     384              :      kind specific data.  */
     385              :   enum stmt_vec_info_type type;
     386              :   vect_data *data;
     387              : 
     388              :   template <class T>
     389      2063882 :   T& get_data (T& else_) { return data ? *static_cast <T *> (data) : else_; }
     390              : 
     391              :   /* If not NULL this is a cached failed SLP discovery attempt with
     392              :      the lanes that failed during SLP discovery as 'false'.  This is
     393              :      a copy of the matches array.  */
     394              :   bool *failed;
     395              : 
     396              :   /* Allocate from slp_tree_pool.  */
     397              :   static void *operator new (size_t);
     398              : 
     399              :   /* Return memory to slp_tree_pool.  */
     400              :   static void operator delete (void *, size_t);
     401              : 
     402              :   /* Linked list of nodes to release when we free the slp_tree_pool.  */
     403              :   slp_tree next_node;
     404              :   slp_tree prev_node;
     405              : };
     406              : 
     407              : /* The enum describes the type of operations that an SLP instance
     408              :    can perform. */
     409              : 
     410              : enum slp_instance_kind {
     411              :     slp_inst_kind_store,
     412              :     slp_inst_kind_reduc_group,
     413              :     slp_inst_kind_reduc_chain,
     414              :     slp_inst_kind_bb_reduc,
     415              :     slp_inst_kind_ctor,
     416              :     slp_inst_kind_gcond
     417              : };
     418              : 
     419              : /* SLP instance is a sequence of stmts in a loop that can be packed into
     420              :    SIMD stmts.  */
     421              : typedef class _slp_instance {
     422              : public:
     423              :   /* The root of SLP tree.  */
     424              :   slp_tree root;
     425              : 
     426              :   /* For vector constructors, the constructor stmt that the SLP tree is built
     427              :      from, NULL otherwise.  */
     428              :   vec<stmt_vec_info> root_stmts;
     429              : 
     430              :   /* For slp_inst_kind_bb_reduc the defs that were not vectorized, NULL
     431              :      otherwise.  */
     432              :   vec<tree> remain_defs;
     433              : 
     434              :   /* The group of nodes that contain loads of this SLP instance.  */
     435              :   vec<slp_tree> loads;
     436              : 
     437              :   /* The SLP node containing the reduction PHIs.  */
     438              :   slp_tree reduc_phis;
     439              : 
     440              :   /* Vector cost of this entry to the SLP graph.  */
     441              :   stmt_vector_for_cost cost_vec;
     442              : 
     443              :   /* If this instance is the main entry of a subgraph the set of
     444              :      entries into the same subgraph, including itself.  */
     445              :   vec<_slp_instance *> subgraph_entries;
     446              : 
     447              :   /* The type of operation the SLP instance is performing.  */
     448              :   slp_instance_kind kind;
     449              : 
     450              :   dump_user_location_t location () const;
     451              : } *slp_instance;
     452              : 
     453              : 
     454              : /* Access Functions.  */
     455              : #define SLP_INSTANCE_TREE(S)                     (S)->root
     456              : #define SLP_INSTANCE_LOADS(S)                    (S)->loads
     457              : #define SLP_INSTANCE_ROOT_STMTS(S)               (S)->root_stmts
     458              : #define SLP_INSTANCE_REMAIN_DEFS(S)              (S)->remain_defs
     459              : #define SLP_INSTANCE_KIND(S)                     (S)->kind
     460              : 
     461              : #define SLP_TREE_CHILDREN(S)                     (S)->children
     462              : #define SLP_TREE_SCALAR_STMTS(S)                 (S)->stmts
     463              : #define SLP_TREE_SCALAR_OPS(S)                   (S)->ops
     464              : #define SLP_TREE_LIVE_LANES(S)                   (S)->live_lanes
     465              : #define SLP_TREE_REF_COUNT(S)                    (S)->refcnt
     466              : #define SLP_TREE_VEC_DEFS(S)                     (S)->vec_defs
     467              : #define SLP_TREE_LOAD_PERMUTATION(S)             (S)->load_permutation
     468              : #define SLP_TREE_LANE_PERMUTATION(S)             (S)->lane_permutation
     469              : #define SLP_TREE_DEF_TYPE(S)                     (S)->def_type
     470              : #define SLP_TREE_VECTYPE(S)                      (S)->vectype
     471              : #define SLP_TREE_REPRESENTATIVE(S)               (S)->representative
     472              : #define SLP_TREE_LANES(S)                        (S)->lanes
     473              : #define SLP_TREE_CODE(S)                         (S)->code
     474              : #define SLP_TREE_TYPE(S)                         (S)->type
     475              : #define SLP_TREE_GS_SCALE(S)                     (S)->gs_scale
     476              : #define SLP_TREE_GS_BASE(S)                      (S)->gs_base
     477              : #define SLP_TREE_REDUC_IDX(S)                    (S)->cycle_info.reduc_idx
     478              : #define SLP_TREE_PERMUTE_P(S)                    ((S)->code == VEC_PERM_EXPR)
     479              : 
     480              : inline vect_memory_access_type
     481      1247793 : SLP_TREE_MEMORY_ACCESS_TYPE (slp_tree node)
     482              : {
     483       482764 :   if (SLP_TREE_TYPE (node) == load_vec_info_type
     484       425834 :       || SLP_TREE_TYPE (node) == store_vec_info_type)
     485       230208 :     return static_cast<vect_load_store_data *> (node->data)->memory_access_type;
     486              :   return VMAT_UNINITIALIZED;
     487              : }
     488              : 
     489              : enum vect_partial_vector_style {
     490              :     vect_partial_vectors_none,
     491              :     vect_partial_vectors_while_ult,
     492              :     vect_partial_vectors_avx512,
     493              :     vect_partial_vectors_len
     494              : };
     495              : 
     496              : /* Key for map that records association between
     497              :    scalar conditions and corresponding loop mask, and
     498              :    is populated by vect_record_loop_mask.  */
     499              : 
     500              : struct scalar_cond_masked_key
     501              : {
     502        63264 :   scalar_cond_masked_key (tree t, unsigned ncopies_)
     503        63264 :     : ncopies (ncopies_)
     504              :   {
     505        63264 :     get_cond_ops_from_tree (t);
     506              :   }
     507              : 
     508              :   void get_cond_ops_from_tree (tree);
     509              : 
     510              :   unsigned ncopies;
     511              :   bool inverted_p;
     512              :   tree_code code;
     513              :   tree op0;
     514              :   tree op1;
     515              : };
     516              : 
     517              : template<>
     518              : struct default_hash_traits<scalar_cond_masked_key>
     519              : {
     520              :   typedef scalar_cond_masked_key compare_type;
     521              :   typedef scalar_cond_masked_key value_type;
     522              : 
     523              :   static inline hashval_t
     524        71851 :   hash (value_type v)
     525              :   {
     526        71851 :     inchash::hash h;
     527        71851 :     h.add_int (v.code);
     528        71851 :     inchash::add_expr (v.op0, h, 0);
     529        71851 :     inchash::add_expr (v.op1, h, 0);
     530        71851 :     h.add_int (v.ncopies);
     531        71851 :     h.add_flag (v.inverted_p);
     532        71851 :     return h.end ();
     533              :   }
     534              : 
     535              :   static inline bool
     536        10315 :   equal (value_type existing, value_type candidate)
     537              :   {
     538        10315 :     return (existing.ncopies == candidate.ncopies
     539        10099 :             && existing.code == candidate.code
     540         6517 :             && existing.inverted_p == candidate.inverted_p
     541         5044 :             && operand_equal_p (existing.op0, candidate.op0, 0)
     542        13449 :             && operand_equal_p (existing.op1, candidate.op1, 0));
     543              :   }
     544              : 
     545              :   static const bool empty_zero_p = true;
     546              : 
     547              :   static inline void
     548            0 :   mark_empty (value_type &v)
     549              :   {
     550            0 :     v.ncopies = 0;
     551            0 :     v.inverted_p = false;
     552              :   }
     553              : 
     554              :   static inline bool
     555      9117502 :   is_empty (value_type v)
     556              :   {
     557      9055411 :     return v.ncopies == 0;
     558              :   }
     559              : 
     560              :   static inline void mark_deleted (value_type &) {}
     561              : 
     562              :   static inline bool is_deleted (const value_type &)
     563              :   {
     564              :     return false;
     565              :   }
     566              : 
     567        54783 :   static inline void remove (value_type &) {}
     568              : };
     569              : 
     570              : typedef hash_set<scalar_cond_masked_key> scalar_cond_masked_set_type;
     571              : 
     572              : /* Key and map that records association between vector conditions and
     573              :    corresponding loop mask, and is populated by prepare_vec_mask.  */
     574              : 
     575              : typedef pair_hash<tree_operand_hash, tree_operand_hash> tree_cond_mask_hash;
     576              : typedef hash_set<tree_cond_mask_hash> vec_cond_masked_set_type;
     577              : 
     578              : /* Describes two objects whose addresses must be unequal for the vectorized
     579              :    loop to be valid.  */
     580              : typedef std::pair<tree, tree> vec_object_pair;
     581              : 
     582              : /* Records that vectorization is only possible if abs (EXPR) >= MIN_VALUE.
     583              :    UNSIGNED_P is true if we can assume that abs (EXPR) == EXPR.  */
     584              : class vec_lower_bound {
     585              : public:
     586              :   vec_lower_bound () {}
     587         1622 :   vec_lower_bound (tree e, bool u, poly_uint64 m)
     588         1622 :     : expr (e), unsigned_p (u), min_value (m) {}
     589              : 
     590              :   tree expr;
     591              :   bool unsigned_p;
     592              :   poly_uint64 min_value;
     593              : };
     594              : 
     595              : /* Vectorizer state shared between different analyses like vector sizes
     596              :    of the same CFG region.  */
     597              : class vec_info_shared {
     598              : public:
     599              :   vec_info_shared();
     600              :   ~vec_info_shared();
     601              : 
     602              :   void save_datarefs();
     603              :   void check_datarefs();
     604              : 
     605              :   /* All data references.  Freed by free_data_refs, so not an auto_vec.  */
     606              :   vec<data_reference_p> datarefs;
     607              :   vec<data_reference> datarefs_copy;
     608              : 
     609              :   /* The loop nest in which the data dependences are computed.  */
     610              :   auto_vec<loop_p> loop_nest;
     611              : 
     612              :   /* All data dependences.  Freed by free_dependence_relations, so not
     613              :      an auto_vec.  */
     614              :   vec<ddr_p> ddrs;
     615              : };
     616              : 
     617              : /* Vectorizer state common between loop and basic-block vectorization.  */
     618              : class vec_info {
     619              : public:
     620              :   typedef hash_set<int_hash<machine_mode, E_VOIDmode, E_BLKmode> > mode_set;
     621              :   enum vec_kind { bb, loop };
     622              : 
     623              :   vec_info (vec_kind, vec_info_shared *);
     624              :   ~vec_info ();
     625              : 
     626              :   stmt_vec_info add_stmt (gimple *);
     627              :   stmt_vec_info add_pattern_stmt (gimple *, stmt_vec_info);
     628              :   stmt_vec_info resync_stmt_addr (gimple *);
     629              :   stmt_vec_info lookup_stmt (gimple *);
     630              :   stmt_vec_info lookup_def (tree);
     631              :   stmt_vec_info lookup_single_use (tree);
     632              :   class dr_vec_info *lookup_dr (data_reference *);
     633              :   void move_dr (stmt_vec_info, stmt_vec_info);
     634              :   void remove_stmt (stmt_vec_info);
     635              :   void replace_stmt (gimple_stmt_iterator *, stmt_vec_info, gimple *);
     636              :   void insert_on_entry (stmt_vec_info, gimple *);
     637              :   void insert_seq_on_entry (stmt_vec_info, gimple_seq);
     638              : 
     639              :   /* The type of vectorization.  */
     640              :   vec_kind kind;
     641              : 
     642              :   /* Shared vectorizer state.  */
     643              :   vec_info_shared *shared;
     644              : 
     645              :   /* The mapping of GIMPLE UID to stmt_vec_info.  */
     646              :   vec<stmt_vec_info> stmt_vec_infos;
     647              :   /* Whether the above mapping is complete.  */
     648              :   bool stmt_vec_info_ro;
     649              : 
     650              :   /* Whether we've done a transform we think OK to not update virtual
     651              :      SSA form.  */
     652              :   bool any_known_not_updated_vssa;
     653              : 
     654              :   /* The SLP graph.  */
     655              :   auto_vec<slp_instance> slp_instances;
     656              : 
     657              :   /* Maps base addresses to an innermost_loop_behavior that gives the maximum
     658              :      known alignment for that base.  */
     659              :   vec_base_alignments base_alignments;
     660              : 
     661              :   /* All interleaving chains of stores, represented by the first
     662              :      stmt in the chain.  */
     663              :   auto_vec<stmt_vec_info> grouped_stores;
     664              : 
     665              :   /* The set of vector modes used in the vectorized region.  */
     666              :   mode_set used_vector_modes;
     667              : 
     668              :   /* The argument we should pass to related_vector_mode when looking up
     669              :      the vector mode for a scalar mode, or VOIDmode if we haven't yet
     670              :      made any decisions about which vector modes to use.  */
     671              :   machine_mode vector_mode;
     672              : 
     673              :   /* The basic blocks in the vectorization region.  For _loop_vec_info,
     674              :      the memory is internally managed, while for _bb_vec_info, it points
     675              :      to element space of an external auto_vec<>.  This inconsistency is
     676              :      not a good class design pattern.  TODO: improve it with an unified
     677              :      auto_vec<> whose lifetime is confined to vec_info object.  */
     678              :   basic_block *bbs;
     679              : 
     680              :   /* The count of the basic blocks in the vectorization region.  */
     681              :   unsigned int nbbs;
     682              : 
     683              :   /* Used to keep a sequence of def stmts of a pattern stmt that are loop
     684              :     invariant if they exists.
     685              :     The sequence is emitted in the loop preheader should the loop be vectorized
     686              :     and are reset when undoing patterns.  */
     687              :   gimple_seq inv_pattern_def_seq;
     688              : 
     689              : private:
     690              :   stmt_vec_info new_stmt_vec_info (gimple *stmt);
     691              :   void set_vinfo_for_stmt (gimple *, stmt_vec_info, bool = true);
     692              :   void free_stmt_vec_infos ();
     693              :   void free_stmt_vec_info (stmt_vec_info);
     694              : };
     695              : 
     696              : class _loop_vec_info;
     697              : class _bb_vec_info;
     698              : 
     699              : template<>
     700              : template<>
     701              : inline bool
     702    368826845 : is_a_helper <_loop_vec_info *>::test (vec_info *i)
     703              : {
     704    368216745 :   return i->kind == vec_info::loop;
     705              : }
     706              : 
     707              : template<>
     708              : template<>
     709              : inline bool
     710     70238763 : is_a_helper <_bb_vec_info *>::test (vec_info *i)
     711              : {
     712     70238763 :   return i->kind == vec_info::bb;
     713              : }
     714              : 
     715              : /* In general, we can divide the vector statements in a vectorized loop
     716              :    into related groups ("rgroups") and say that for each rgroup there is
     717              :    some nS such that the rgroup operates on nS values from one scalar
     718              :    iteration followed by nS values from the next.  That is, if VF is the
     719              :    vectorization factor of the loop, the rgroup operates on a sequence:
     720              : 
     721              :      (1,1) (1,2) ... (1,nS) (2,1) ... (2,nS) ... (VF,1) ... (VF,nS)
     722              : 
     723              :    where (i,j) represents a scalar value with index j in a scalar
     724              :    iteration with index i.
     725              : 
     726              :    [ We use the term "rgroup" to emphasise that this grouping isn't
     727              :      necessarily the same as the grouping of statements used elsewhere.
     728              :      For example, if we implement a group of scalar loads using gather
     729              :      loads, we'll use a separate gather load for each scalar load, and
     730              :      thus each gather load will belong to its own rgroup. ]
     731              : 
     732              :    In general this sequence will occupy nV vectors concatenated
     733              :    together.  If these vectors have nL lanes each, the total number
     734              :    of scalar values N is given by:
     735              : 
     736              :        N = nS * VF = nV * nL
     737              : 
     738              :    None of nS, VF, nV and nL are required to be a power of 2.  nS and nV
     739              :    are compile-time constants but VF and nL can be variable (if the target
     740              :    supports variable-length vectors).
     741              : 
     742              :    In classical vectorization, each iteration of the vector loop would
     743              :    handle exactly VF iterations of the original scalar loop.  However,
     744              :    in vector loops that are able to operate on partial vectors, a
     745              :    particular iteration of the vector loop might handle fewer than VF
     746              :    iterations of the scalar loop.  The vector lanes that correspond to
     747              :    iterations of the scalar loop are said to be "active" and the other
     748              :    lanes are said to be "inactive".
     749              : 
     750              :    In such vector loops, many rgroups need to be controlled to ensure
     751              :    that they have no effect for the inactive lanes.  Conceptually, each
     752              :    such rgroup needs a sequence of booleans in the same order as above,
     753              :    but with each (i,j) replaced by a boolean that indicates whether
     754              :    iteration i is active.  This sequence occupies nV vector controls
     755              :    that again have nL lanes each.  Thus the control sequence as a whole
     756              :    consists of VF independent booleans that are each repeated nS times.
     757              : 
     758              :    Taking mask-based approach as a partially-populated vectors example.
     759              :    We make the simplifying assumption that if a sequence of nV masks is
     760              :    suitable for one (nS,nL) pair, we can reuse it for (nS/2,nL/2) by
     761              :    VIEW_CONVERTing it.  This holds for all current targets that support
     762              :    fully-masked loops.  For example, suppose the scalar loop is:
     763              : 
     764              :      float *f;
     765              :      double *d;
     766              :      for (int i = 0; i < n; ++i)
     767              :        {
     768              :          f[i * 2 + 0] += 1.0f;
     769              :          f[i * 2 + 1] += 2.0f;
     770              :          d[i] += 3.0;
     771              :        }
     772              : 
     773              :    and suppose that vectors have 256 bits.  The vectorized f accesses
     774              :    will belong to one rgroup and the vectorized d access to another:
     775              : 
     776              :      f rgroup: nS = 2, nV = 1, nL = 8
     777              :      d rgroup: nS = 1, nV = 1, nL = 4
     778              :                VF = 4
     779              : 
     780              :      [ In this simple example the rgroups do correspond to the normal
     781              :        SLP grouping scheme. ]
     782              : 
     783              :    If only the first three lanes are active, the masks we need are:
     784              : 
     785              :      f rgroup: 1 1 | 1 1 | 1 1 | 0 0
     786              :      d rgroup:  1  |  1  |  1  |  0
     787              : 
     788              :    Here we can use a mask calculated for f's rgroup for d's, but not
     789              :    vice versa.
     790              : 
     791              :    Thus for each value of nV, it is enough to provide nV masks, with the
     792              :    mask being calculated based on the highest nL (or, equivalently, based
     793              :    on the highest nS) required by any rgroup with that nV.  We therefore
     794              :    represent the entire collection of masks as a two-level table, with the
     795              :    first level being indexed by nV - 1 (since nV == 0 doesn't exist) and
     796              :    the second being indexed by the mask index 0 <= i < nV.  */
     797              : 
     798              : /* The controls (like masks or lengths) needed by rgroups with nV vectors,
     799              :    according to the description above.  */
     800              : struct rgroup_controls {
     801              :   /* The largest nS for all rgroups that use these controls.
     802              :      For vect_partial_vectors_avx512 this is the constant nscalars_per_iter
     803              :      for all members of the group.  */
     804              :   unsigned int max_nscalars_per_iter;
     805              : 
     806              :   /* For the largest nS recorded above, the loop controls divide each scalar
     807              :      into FACTOR equal-sized pieces.  This is useful if we need to split
     808              :      element-based accesses into byte-based accesses.
     809              :      For vect_partial_vectors_avx512 this records nV instead.  */
     810              :   unsigned int factor;
     811              : 
     812              :   /* This is a vector type with MAX_NSCALARS_PER_ITER * VF / nV elements.
     813              :      For mask-based controls, it is the type of the masks in CONTROLS.
     814              :      For length-based controls, it can be any vector type that has the
     815              :      specified number of elements; the type of the elements doesn't matter.  */
     816              :   tree type;
     817              : 
     818              :   /* When there is no uniformly used LOOP_VINFO_RGROUP_COMPARE_TYPE this
     819              :      is the rgroup specific type used.  */
     820              :   tree compare_type;
     821              : 
     822              :   /* A vector of nV controls, in iteration order.  */
     823              :   vec<tree> controls;
     824              : 
     825              :   /* In case of len_load and len_store with a bias there is only one
     826              :      rgroup.  This holds the adjusted loop length for the this rgroup.  */
     827              :   tree bias_adjusted_ctrl;
     828              : };
     829              : 
     830       573065 : struct vec_loop_masks
     831              : {
     832       503788 :   bool is_empty () const { return mask_set.is_empty (); }
     833              : 
     834              :   /* Set to record vectype, nvector pairs.  */
     835              :   hash_set<pair_hash <nofree_ptr_hash <tree_node>,
     836              :                       int_hash<unsigned, 0>>> mask_set;
     837              : 
     838              :   /* rgroup_controls used for the partial vector scheme.  */
     839              :   auto_vec<rgroup_controls> rgc_vec;
     840              : };
     841              : 
     842              : typedef auto_vec<rgroup_controls> vec_loop_lens;
     843              : 
     844              : typedef auto_vec<std::pair<data_reference*, tree> > drs_init_vec;
     845              : 
     846              : /* Abstraction around info on reductions which is still in stmt_vec_info
     847              :    but will be duplicated or moved elsewhere.  */
     848       202932 : class vect_reduc_info_s
     849              : {
     850              : public:
     851              :   /* The def type of the main reduction PHI, vect_reduction_def or
     852              :      vect_double_reduction_def.  */
     853              :   enum vect_def_type def_type;
     854              : 
     855              :   /* The reduction type as detected by
     856              :      vect_is_simple_reduction and vectorizable_reduction.  */
     857              :   enum vect_reduction_type reduc_type;
     858              : 
     859              :   /* The original scalar reduction code, to be used in the epilogue.  */
     860              :   code_helper reduc_code;
     861              : 
     862              :   /* A vector internal function we should use in the epilogue.  */
     863              :   internal_fn reduc_fn;
     864              : 
     865              :   /* For loop reduction with multiple vectorized results (ncopies > 1), a
     866              :      lane-reducing operation participating in it may not use all of those
     867              :      results, this field specifies result index starting from which any
     868              :      following land-reducing operation would be assigned to.  */
     869              :   unsigned int reduc_result_pos;
     870              : 
     871              :   /* Whether this represents a reduction chain.  */
     872              :   bool is_reduc_chain;
     873              : 
     874              :   /* Whether we force a single cycle PHI during reduction vectorization.  */
     875              :   bool force_single_cycle;
     876              : 
     877              :   /* The vector type for performing the actual reduction operation.  */
     878              :   tree reduc_vectype;
     879              : 
     880              :   /* The vector type we should use for the final reduction in the epilogue
     881              :      when we reduce a mask.  */
     882              :   tree reduc_vectype_for_mask;
     883              : 
     884              :   /* The neutral operand to use, if any.  */
     885              :   tree neutral_op;
     886              : 
     887              :   /* For INTEGER_INDUC_COND_REDUCTION, the initial value to be used.  */
     888              :   tree induc_cond_initial_val;
     889              : 
     890              :   /* If not NULL the value to be added to compute final reduction value.  */
     891              :   tree reduc_epilogue_adjustment;
     892              : 
     893              :   /* If non-null, the reduction is being performed by an epilogue loop
     894              :      and we have decided to reuse this accumulator from the main loop.  */
     895              :   struct vect_reusable_accumulator *reused_accumulator;
     896              : 
     897              :   /* If the vector code is performing N scalar reductions in parallel,
     898              :      this variable gives the initial scalar values of those N reductions.  */
     899              :   auto_vec<tree> reduc_initial_values;
     900              : 
     901              :   /* If the vector code is performing N scalar reductions in parallel, this
     902              :      variable gives the vectorized code's final (scalar) result for each of
     903              :      those N reductions.  In other words, REDUC_SCALAR_RESULTS[I] replaces
     904              :      the original scalar code's loop-closed SSA PHI for reduction number I.  */
     905              :   auto_vec<tree> reduc_scalar_results;
     906              : };
     907              : 
     908              : typedef class vect_reduc_info_s *vect_reduc_info;
     909              : 
     910              : #define VECT_REDUC_INFO_DEF_TYPE(I) ((I)->def_type)
     911              : #define VECT_REDUC_INFO_TYPE(I) ((I)->reduc_type)
     912              : #define VECT_REDUC_INFO_CODE(I) ((I)->reduc_code)
     913              : #define VECT_REDUC_INFO_FN(I) ((I)->reduc_fn)
     914              : #define VECT_REDUC_INFO_SCALAR_RESULTS(I) ((I)->reduc_scalar_results)
     915              : #define VECT_REDUC_INFO_INITIAL_VALUES(I) ((I)->reduc_initial_values)
     916              : #define VECT_REDUC_INFO_REUSED_ACCUMULATOR(I) ((I)->reused_accumulator)
     917              : #define VECT_REDUC_INFO_INDUC_COND_INITIAL_VAL(I) ((I)->induc_cond_initial_val)
     918              : #define VECT_REDUC_INFO_EPILOGUE_ADJUSTMENT(I) ((I)->reduc_epilogue_adjustment)
     919              : #define VECT_REDUC_INFO_VECTYPE(I) ((I)->reduc_vectype)
     920              : #define VECT_REDUC_INFO_VECTYPE_FOR_MASK(I) ((I)->reduc_vectype_for_mask)
     921              : #define VECT_REDUC_INFO_FORCE_SINGLE_CYCLE(I) ((I)->force_single_cycle)
     922              : #define VECT_REDUC_INFO_RESULT_POS(I) ((I)->reduc_result_pos)
     923              : #define VECT_REDUC_INFO_NEUTRAL_OP(I) ((I)->neutral_op)
     924              : 
     925              : /* Information about a reduction accumulator from the main loop that could
     926              :    conceivably be reused as the input to a reduction in an epilogue loop.  */
     927              : struct vect_reusable_accumulator {
     928              :   /* The final value of the accumulator, which forms the input to the
     929              :      reduction operation.  */
     930              :   tree reduc_input;
     931              : 
     932              :   /* The stmt_vec_info that describes the reduction (i.e. the one for
     933              :      which is_reduc_info is true).  */
     934              :   vect_reduc_info reduc_info;
     935              : };
     936              : 
     937              : /*-----------------------------------------------------------------*/
     938              : /* Info on vectorized loops.                                       */
     939              : /*-----------------------------------------------------------------*/
     940              : typedef class _loop_vec_info : public vec_info {
     941              : public:
     942              :   _loop_vec_info (class loop *, vec_info_shared *);
     943              :   ~_loop_vec_info ();
     944              : 
     945              :   /* The loop to which this info struct refers to.  */
     946              :   class loop *loop;
     947              : 
     948              :   /* Number of latch executions.  */
     949              :   tree num_itersm1;
     950              :   /* Number of iterations.  */
     951              :   tree num_iters;
     952              :   /* Number of iterations of the original loop.  */
     953              :   tree num_iters_unchanged;
     954              :   /* Condition under which this loop is analyzed and versioned.  */
     955              :   tree num_iters_assumptions;
     956              : 
     957              :   /* The cost of the vector code.  */
     958              :   class vector_costs *vector_costs;
     959              : 
     960              :   /* The cost of the scalar code.  */
     961              :   class vector_costs *scalar_costs;
     962              : 
     963              :   /* Threshold of number of iterations below which vectorization will not be
     964              :      performed. It is calculated from MIN_PROFITABLE_ITERS and
     965              :      param_min_vect_loop_bound.  */
     966              :   unsigned int th;
     967              : 
     968              :   /* When applying loop versioning, the vector form should only be used
     969              :      if the number of scalar iterations is >= this value, on top of all
     970              :      the other requirements.  Ignored when loop versioning is not being
     971              :      used.  */
     972              :   poly_uint64 versioning_threshold;
     973              : 
     974              :   /* Unrolling factor.  In case of suitable super-word parallelism
     975              :      it can be that no unrolling is needed, and thus this is 1.  */
     976              :   poly_uint64 vectorization_factor;
     977              : 
     978              :   /* If this loop is an epilogue loop whose main loop can be skipped,
     979              :      MAIN_LOOP_EDGE is the edge from the main loop to this loop's
     980              :      preheader.  SKIP_MAIN_LOOP_EDGE is then the edge that skips the
     981              :      main loop and goes straight to this loop's preheader.
     982              : 
     983              :      Both fields are null otherwise.  */
     984              :   edge main_loop_edge;
     985              :   edge skip_main_loop_edge;
     986              : 
     987              :   /* If this loop is an epilogue loop that might be skipped after executing
     988              :      the main loop, this edge is the one that skips the epilogue.  */
     989              :   edge skip_this_loop_edge;
     990              : 
     991              :   /* Reduction descriptors of this loop.  Referenced to from SLP nodes
     992              :      by index.  */
     993              :   auto_vec<vect_reduc_info> reduc_infos;
     994              : 
     995              :   /* The vectorized form of a standard reduction replaces the original
     996              :      scalar code's final result (a loop-closed SSA PHI) with the result
     997              :      of a vector-to-scalar reduction operation.  After vectorization,
     998              :      this variable maps these vector-to-scalar results to information
     999              :      about the reductions that generated them.  */
    1000              :   hash_map<tree, vect_reusable_accumulator> reusable_accumulators;
    1001              : 
    1002              :   /* The number of times that the target suggested we unroll the vector loop
    1003              :      in order to promote more ILP.  This value will be used to re-analyze the
    1004              :      loop for vectorization and if successful the value will be folded into
    1005              :      vectorization_factor (and therefore exactly divides
    1006              :      vectorization_factor).  */
    1007              :   unsigned int suggested_unroll_factor;
    1008              : 
    1009              :   /* Maximum runtime vectorization factor, or MAX_VECTORIZATION_FACTOR
    1010              :      if there is no particular limit.  */
    1011              :   unsigned HOST_WIDE_INT max_vectorization_factor;
    1012              : 
    1013              :   /* The masks that a fully-masked loop should use to avoid operating
    1014              :      on inactive scalars.  */
    1015              :   vec_loop_masks masks;
    1016              : 
    1017              :   /* The lengths that a loop with length should use to avoid operating
    1018              :      on inactive scalars.  */
    1019              :   vec_loop_lens lens;
    1020              : 
    1021              :   /* Set of scalar conditions that have loop mask applied.  */
    1022              :   scalar_cond_masked_set_type scalar_cond_masked_set;
    1023              : 
    1024              :   /* Set of vector conditions that have loop mask applied.  */
    1025              :   vec_cond_masked_set_type vec_cond_masked_set;
    1026              : 
    1027              :   /* If we are using a loop mask to align memory addresses, this variable
    1028              :      contains the number of vector elements that we should skip in the
    1029              :      first iteration of the vector loop (i.e. the number of leading
    1030              :      elements that should be false in the first mask).  */
    1031              :   tree mask_skip_niters;
    1032              : 
    1033              :   /* If we are using a loop mask to align memory addresses and we're in an
    1034              :      early break loop then this variable contains the number of elements that
    1035              :      were skipped during the initial iteration of the loop. */
    1036              :   tree mask_skip_niters_pfa_offset;
    1037              : 
    1038              :   /* The type that the loop control IV should be converted to before
    1039              :      testing which of the VF scalars are active and inactive.
    1040              :      Only meaningful if LOOP_VINFO_USING_PARTIAL_VECTORS_P.  */
    1041              :   tree rgroup_compare_type;
    1042              : 
    1043              :   /* For #pragma omp simd if (x) loops the x expression.  If constant 0,
    1044              :      the loop should not be vectorized, if constant non-zero, simd_if_cond
    1045              :      shouldn't be set and loop vectorized normally, if SSA_NAME, the loop
    1046              :      should be versioned on that condition, using scalar loop if the condition
    1047              :      is false and vectorized loop otherwise.  */
    1048              :   tree simd_if_cond;
    1049              : 
    1050              :   /* The type that the vector loop control IV should have when
    1051              :      LOOP_VINFO_USING_PARTIAL_VECTORS_P is true.  */
    1052              :   tree rgroup_iv_type;
    1053              : 
    1054              :   /* The style used for implementing partial vectors when
    1055              :      LOOP_VINFO_USING_PARTIAL_VECTORS_P is true.  */
    1056              :   vect_partial_vector_style partial_vector_style;
    1057              : 
    1058              :   /* Unknown DRs according to which loop was peeled.  */
    1059              :   class dr_vec_info *unaligned_dr;
    1060              : 
    1061              :   /* peeling_for_alignment indicates whether peeling for alignment will take
    1062              :      place, and what the peeling factor should be:
    1063              :      peeling_for_alignment = X means:
    1064              :         If X=0: Peeling for alignment will not be applied.
    1065              :         If X>0: Peel first X iterations.
    1066              :         If X=-1: Generate a runtime test to calculate the number of iterations
    1067              :                  to be peeled, using the dataref recorded in the field
    1068              :                  unaligned_dr.  */
    1069              :   int peeling_for_alignment;
    1070              : 
    1071              :   /* The mask used to check the alignment of pointers or arrays.  */
    1072              :   poly_uint64 ptr_mask;
    1073              : 
    1074              :   /* The maximum speculative read amount in VLA modes for runtime check.  */
    1075              :   poly_uint64 max_spec_read_amount;
    1076              : 
    1077              :   /* Indicates whether the loop has any non-linear IV.  */
    1078              :   bool nonlinear_iv;
    1079              : 
    1080              :   /* Data Dependence Relations defining address ranges that are candidates
    1081              :      for a run-time aliasing check.  */
    1082              :   auto_vec<ddr_p> may_alias_ddrs;
    1083              : 
    1084              :   /* Data Dependence Relations defining address ranges together with segment
    1085              :      lengths from which the run-time aliasing check is built.  */
    1086              :   auto_vec<dr_with_seg_len_pair_t> comp_alias_ddrs;
    1087              : 
    1088              :   /* Check that the addresses of each pair of objects is unequal.  */
    1089              :   auto_vec<vec_object_pair> check_unequal_addrs;
    1090              : 
    1091              :   /* List of values that are required to be nonzero.  This is used to check
    1092              :      whether things like "x[i * n] += 1;" are safe and eventually gets added
    1093              :      to the checks for lower bounds below.  */
    1094              :   auto_vec<tree> check_nonzero;
    1095              : 
    1096              :   /* List of values that need to be checked for a minimum value.  */
    1097              :   auto_vec<vec_lower_bound> lower_bounds;
    1098              : 
    1099              :   /* Statements in the loop that have data references that are candidates for a
    1100              :      runtime (loop versioning) misalignment check.  */
    1101              :   auto_vec<stmt_vec_info> may_misalign_stmts;
    1102              : 
    1103              :   /* Reduction cycles detected in the loop. Used in loop-aware SLP.  */
    1104              :   auto_vec<stmt_vec_info> reductions;
    1105              : 
    1106              :   /* Defs that could not be analyzed such as OMP SIMD calls without
    1107              :      a LHS.  */
    1108              :   auto_vec<stmt_vec_info> alternate_defs;
    1109              : 
    1110              :   /* Cost vector for a single scalar iteration.  */
    1111              :   auto_vec<stmt_info_for_cost> scalar_cost_vec;
    1112              : 
    1113              :   /* Map of IV base/step expressions to inserted name in the preheader.  */
    1114              :   hash_map<tree_operand_hash, tree> *ivexpr_map;
    1115              : 
    1116              :   /* Map of OpenMP "omp simd array" scan variables to corresponding
    1117              :      rhs of the store of the initializer.  */
    1118              :   hash_map<tree, tree> *scan_map;
    1119              : 
    1120              :   /* The factor used to over weight those statements in an inner loop
    1121              :      relative to the loop being vectorized.  */
    1122              :   unsigned int inner_loop_cost_factor;
    1123              : 
    1124              :   /* Is the loop vectorizable? */
    1125              :   bool vectorizable;
    1126              : 
    1127              :   /* Records whether we still have the option of vectorizing this loop
    1128              :      using partially-populated vectors; in other words, whether it is
    1129              :      still possible for one iteration of the vector loop to handle
    1130              :      fewer than VF scalars.  */
    1131              :   bool can_use_partial_vectors_p;
    1132              : 
    1133              :   /* Records whether we must use niter masking for correctness reasons.  */
    1134              :   bool must_use_partial_vectors_p;
    1135              : 
    1136              :   /* True if we've decided to use partially-populated vectors, so that
    1137              :      the vector loop can handle fewer than VF scalars.  */
    1138              :   bool using_partial_vectors_p;
    1139              : 
    1140              :   /* True if we've decided to use a decrementing loop control IV that counts
    1141              :      scalars. This can be done for any loop that:
    1142              : 
    1143              :         (a) uses length "controls"; and
    1144              :         (b) can iterate more than once.  */
    1145              :   bool using_decrementing_iv_p;
    1146              : 
    1147              :   /* True if we've decided to use output of select_vl to adjust IV of
    1148              :      both loop control and data reference pointer. This is only true
    1149              :      for single-rgroup control.  */
    1150              :   bool using_select_vl_p;
    1151              : 
    1152              :   /* True if we've decided to use peeling with versioning together, which allows
    1153              :      unaligned unsupported data refs to be uniformly aligned after a certain
    1154              :      amount of peeling (mutual alignment).  Otherwise, we use versioning alone
    1155              :      so these data refs must be already aligned to a power-of-two boundary
    1156              :      without peeling.  */
    1157              :   bool allow_mutual_alignment;
    1158              : 
    1159              :   /* The bias for len_load and len_store.  For now, only 0 and -1 are
    1160              :      supported.  -1 must be used when a backend does not support
    1161              :      len_load/len_store with a length of zero.  */
    1162              :   signed char partial_load_store_bias;
    1163              : 
    1164              :   /* When we have grouped data accesses with gaps, we may introduce invalid
    1165              :      memory accesses.  We peel the last iteration of the loop to prevent
    1166              :      this.  */
    1167              :   bool peeling_for_gaps;
    1168              : 
    1169              :   /* When the number of iterations is not a multiple of the vector size
    1170              :      we need to peel off iterations at the end to form an epilogue loop.  */
    1171              :   bool peeling_for_niter;
    1172              : 
    1173              :   /* When the loop has early breaks that we can vectorize we need to peel
    1174              :      the loop for the break finding loop.  */
    1175              :   bool early_breaks;
    1176              : 
    1177              :   /* List of loop additional IV conditionals found in the loop.  */
    1178              :   auto_vec<gcond *> conds;
    1179              : 
    1180              :   /* Main loop IV cond.  */
    1181              :   gcond* loop_iv_cond;
    1182              : 
    1183              :   /* True if we have an unroll factor requested by the user through pragma GCC
    1184              :      unroll.  */
    1185              :   bool user_unroll;
    1186              : 
    1187              :   /* True if there are no loop carried data dependencies in the loop.
    1188              :      If loop->safelen <= 1, then this is always true, either the loop
    1189              :      didn't have any loop carried data dependencies, or the loop is being
    1190              :      vectorized guarded with some runtime alias checks, or couldn't
    1191              :      be vectorized at all, but then this field shouldn't be used.
    1192              :      For loop->safelen >= 2, the user has asserted that there are no
    1193              :      backward dependencies, but there still could be loop carried forward
    1194              :      dependencies in such loops.  This flag will be false if normal
    1195              :      vectorizer data dependency analysis would fail or require versioning
    1196              :      for alias, but because of loop->safelen >= 2 it has been vectorized
    1197              :      even without versioning for alias.  E.g. in:
    1198              :      #pragma omp simd
    1199              :      for (int i = 0; i < m; i++)
    1200              :        a[i] = a[i + k] * c;
    1201              :      (or #pragma simd or #pragma ivdep) we can vectorize this and it will
    1202              :      DTRT even for k > 0 && k < m, but without safelen we would not
    1203              :      vectorize this, so this field would be false.  */
    1204              :   bool no_data_dependencies;
    1205              : 
    1206              :   /* Mark loops having masked stores.  */
    1207              :   bool has_mask_store;
    1208              : 
    1209              :   /* Queued scaling factor for the scalar loop.  */
    1210              :   profile_probability scalar_loop_scaling;
    1211              : 
    1212              :   /* If if-conversion versioned this loop before conversion, this is the
    1213              :      loop version without if-conversion.  */
    1214              :   class loop *scalar_loop;
    1215              : 
    1216              :   /* For loops being epilogues of already vectorized loops
    1217              :      this points to the main vectorized loop.  Otherwise NULL.  */
    1218              :   _loop_vec_info *main_loop_info;
    1219              : 
    1220              :   /* For loops being epilogues of already vectorized loops
    1221              :      this points to the preceeding vectorized (possibly epilogue) loop.
    1222              :      Otherwise NULL.  */
    1223              :   _loop_vec_info *orig_loop_info;
    1224              : 
    1225              :   /* Used to store loop_vec_infos of the epilogue of this loop during
    1226              :      analysis.  */
    1227              :   _loop_vec_info *epilogue_vinfo;
    1228              : 
    1229              :   /* If this is an epilogue loop the DR advancement applied.  */
    1230              :   tree drs_advanced_by;
    1231              : 
    1232              :   /* The controlling loop exit for the current loop when vectorizing.
    1233              :      For counted loops, this IV controls the natural exits of the loop.  */
    1234              :   edge vec_loop_main_exit;
    1235              : 
    1236              :   /* The controlling loop exit for the epilogue loop when vectorizing.
    1237              :      For counted loops, this IV controls the natural exits of the loop.  */
    1238              :   edge vec_epilogue_loop_main_exit;
    1239              : 
    1240              :   /* The controlling loop exit for the scalar loop being vectorized.
    1241              :      For counted loops, this IV controls the natural exits of the loop.  */
    1242              :   edge scalar_loop_main_exit;
    1243              : 
    1244              :   /* Used to store the list of stores needing to be moved if doing early
    1245              :      break vectorization as they would violate the scalar loop semantics if
    1246              :      vectorized in their current location.  These are stored in order that they
    1247              :      need to be moved.  */
    1248              :   auto_vec<gimple *> early_break_stores;
    1249              : 
    1250              :   /* The final basic block where to move statements to.  In the case of
    1251              :      multiple exits this could be pretty far away.  */
    1252              :   basic_block early_break_dest_bb;
    1253              : 
    1254              :   /* Statements whose VUSES need updating if early break vectorization is to
    1255              :      happen.  */
    1256              :   auto_vec<gimple*> early_break_vuses;
    1257              : 
    1258              :   /* The IV adjustment value for inductions that needs to be materialized
    1259              :      inside the relavent exit blocks in order to adjust for early break.  */
    1260              :   tree early_break_niters_var;
    1261              : 
    1262              :   /* The type of the variable to be used to create the scalar IV for early break
    1263              :      loops.  */
    1264              :   tree early_break_iv_type;
    1265              : 
    1266              :   /* Record statements that are needed to be live for early break vectorization
    1267              :      but may not have an LC PHI node materialized yet in the exits.  */
    1268              :   auto_vec<stmt_vec_info> early_break_live_ivs;
    1269              : } *loop_vec_info;
    1270              : 
    1271              : /* Access Functions.  */
    1272              : #define LOOP_VINFO_LOOP(L)                 (L)->loop
    1273              : #define LOOP_VINFO_MAIN_EXIT(L)              (L)->vec_loop_main_exit
    1274              : #define LOOP_VINFO_EPILOGUE_MAIN_EXIT(L)     (L)->vec_epilogue_loop_main_exit
    1275              : #define LOOP_VINFO_SCALAR_MAIN_EXIT(L)       (L)->scalar_loop_main_exit
    1276              : #define LOOP_VINFO_BBS(L)                  (L)->bbs
    1277              : #define LOOP_VINFO_NBBS(L)                 (L)->nbbs
    1278              : #define LOOP_VINFO_NITERSM1(L)             (L)->num_itersm1
    1279              : #define LOOP_VINFO_NITERS(L)               (L)->num_iters
    1280              : #define LOOP_VINFO_NITERS_UNCOUNTED_P(L)   (LOOP_VINFO_NITERS (L) \
    1281              :                                             == chrec_dont_know)
    1282              : /* Since LOOP_VINFO_NITERS and LOOP_VINFO_NITERSM1 can change after
    1283              :    prologue peeling retain total unchanged scalar loop iterations for
    1284              :    cost model.  */
    1285              : #define LOOP_VINFO_NITERS_UNCHANGED(L)     (L)->num_iters_unchanged
    1286              : #define LOOP_VINFO_NITERS_ASSUMPTIONS(L)   (L)->num_iters_assumptions
    1287              : #define LOOP_VINFO_COST_MODEL_THRESHOLD(L) (L)->th
    1288              : #define LOOP_VINFO_VERSIONING_THRESHOLD(L) (L)->versioning_threshold
    1289              : #define LOOP_VINFO_VECTORIZABLE_P(L)       (L)->vectorizable
    1290              : #define LOOP_VINFO_CAN_USE_PARTIAL_VECTORS_P(L) (L)->can_use_partial_vectors_p
    1291              : #define LOOP_VINFO_MUST_USE_PARTIAL_VECTORS_P(L) (L)->must_use_partial_vectors_p
    1292              : #define LOOP_VINFO_USING_PARTIAL_VECTORS_P(L) (L)->using_partial_vectors_p
    1293              : #define LOOP_VINFO_USING_DECREMENTING_IV_P(L) (L)->using_decrementing_iv_p
    1294              : #define LOOP_VINFO_USING_SELECT_VL_P(L) (L)->using_select_vl_p
    1295              : #define LOOP_VINFO_ALLOW_MUTUAL_ALIGNMENT(L) (L)->allow_mutual_alignment
    1296              : #define LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS(L) (L)->partial_load_store_bias
    1297              : #define LOOP_VINFO_VECT_FACTOR(L)          (L)->vectorization_factor
    1298              : #define LOOP_VINFO_MAX_VECT_FACTOR(L)      (L)->max_vectorization_factor
    1299              : #define LOOP_VINFO_MASKS(L)                (L)->masks
    1300              : #define LOOP_VINFO_LENS(L)                 (L)->lens
    1301              : #define LOOP_VINFO_MASK_SKIP_NITERS(L)     (L)->mask_skip_niters
    1302              : #define LOOP_VINFO_MASK_NITERS_PFA_OFFSET(L) (L)->mask_skip_niters_pfa_offset
    1303              : #define LOOP_VINFO_RGROUP_COMPARE_TYPE(L)  (L)->rgroup_compare_type
    1304              : #define LOOP_VINFO_RGROUP_IV_TYPE(L)       (L)->rgroup_iv_type
    1305              : #define LOOP_VINFO_PARTIAL_VECTORS_STYLE(L) (L)->partial_vector_style
    1306              : #define LOOP_VINFO_PTR_MASK(L)             (L)->ptr_mask
    1307              : #define LOOP_VINFO_MAX_SPEC_READ_AMOUNT(L) (L)->max_spec_read_amount
    1308              : #define LOOP_VINFO_LOOP_NEST(L)            (L)->shared->loop_nest
    1309              : #define LOOP_VINFO_DATAREFS(L)             (L)->shared->datarefs
    1310              : #define LOOP_VINFO_DDRS(L)                 (L)->shared->ddrs
    1311              : #define LOOP_VINFO_INT_NITERS(L)           (TREE_INT_CST_LOW ((L)->num_iters))
    1312              : #define LOOP_VINFO_PEELING_FOR_ALIGNMENT(L) (L)->peeling_for_alignment
    1313              : #define LOOP_VINFO_NON_LINEAR_IV(L)        (L)->nonlinear_iv
    1314              : #define LOOP_VINFO_UNALIGNED_DR(L)         (L)->unaligned_dr
    1315              : #define LOOP_VINFO_MAY_MISALIGN_STMTS(L)   (L)->may_misalign_stmts
    1316              : #define LOOP_VINFO_MAY_ALIAS_DDRS(L)       (L)->may_alias_ddrs
    1317              : #define LOOP_VINFO_COMP_ALIAS_DDRS(L)      (L)->comp_alias_ddrs
    1318              : #define LOOP_VINFO_CHECK_UNEQUAL_ADDRS(L)  (L)->check_unequal_addrs
    1319              : #define LOOP_VINFO_CHECK_NONZERO(L)        (L)->check_nonzero
    1320              : #define LOOP_VINFO_LOWER_BOUNDS(L)         (L)->lower_bounds
    1321              : #define LOOP_VINFO_USER_UNROLL(L)          (L)->user_unroll
    1322              : #define LOOP_VINFO_GROUPED_STORES(L)       (L)->grouped_stores
    1323              : #define LOOP_VINFO_SLP_INSTANCES(L)        (L)->slp_instances
    1324              : #define LOOP_VINFO_REDUCTIONS(L)           (L)->reductions
    1325              : #define LOOP_VINFO_PEELING_FOR_GAPS(L)     (L)->peeling_for_gaps
    1326              : #define LOOP_VINFO_PEELING_FOR_NITER(L)    (L)->peeling_for_niter
    1327              : #define LOOP_VINFO_EARLY_BREAKS(L)         (L)->early_breaks
    1328              : #define LOOP_VINFO_EARLY_BRK_STORES(L)     (L)->early_break_stores
    1329              : #define LOOP_VINFO_EARLY_BREAKS_VECT_PEELED(L)  \
    1330              :   ((single_pred ((L)->loop->latch) != (L)->vec_loop_main_exit->src) \
    1331              :    || LOOP_VINFO_NITERS_UNCOUNTED_P (L))
    1332              : #define LOOP_VINFO_EARLY_BREAKS_LIVE_IVS(L)  \
    1333              :   (L)->early_break_live_ivs
    1334              : #define LOOP_VINFO_EARLY_BRK_DEST_BB(L)    (L)->early_break_dest_bb
    1335              : #define LOOP_VINFO_EARLY_BRK_VUSES(L)      (L)->early_break_vuses
    1336              : #define LOOP_VINFO_EARLY_BRK_NITERS_VAR(L) (L)->early_break_niters_var
    1337              : #define LOOP_VINFO_EARLY_BRK_IV_TYPE(L)    (L)->early_break_iv_type
    1338              : #define LOOP_VINFO_LOOP_CONDS(L)           (L)->conds
    1339              : #define LOOP_VINFO_LOOP_IV_COND(L)         (L)->loop_iv_cond
    1340              : #define LOOP_VINFO_NO_DATA_DEPENDENCIES(L) (L)->no_data_dependencies
    1341              : #define LOOP_VINFO_SCALAR_LOOP(L)          (L)->scalar_loop
    1342              : #define LOOP_VINFO_SCALAR_LOOP_SCALING(L)  (L)->scalar_loop_scaling
    1343              : #define LOOP_VINFO_HAS_MASK_STORE(L)       (L)->has_mask_store
    1344              : #define LOOP_VINFO_SCALAR_ITERATION_COST(L) (L)->scalar_cost_vec
    1345              : #define LOOP_VINFO_MAIN_LOOP_INFO(L)       (L)->main_loop_info
    1346              : #define LOOP_VINFO_ORIG_LOOP_INFO(L)       (L)->orig_loop_info
    1347              : #define LOOP_VINFO_SIMD_IF_COND(L)         (L)->simd_if_cond
    1348              : #define LOOP_VINFO_INNER_LOOP_COST_FACTOR(L) (L)->inner_loop_cost_factor
    1349              : #define LOOP_VINFO_INV_PATTERN_DEF_SEQ(L)  (L)->inv_pattern_def_seq
    1350              : #define LOOP_VINFO_DRS_ADVANCED_BY(L)      (L)->drs_advanced_by
    1351              : #define LOOP_VINFO_ALTERNATE_DEFS(L)       (L)->alternate_defs
    1352              : 
    1353              : #define LOOP_VINFO_FULLY_MASKED_P(L)            \
    1354              :   (LOOP_VINFO_USING_PARTIAL_VECTORS_P (L)       \
    1355              :    && !LOOP_VINFO_MASKS (L).is_empty ())
    1356              : 
    1357              : #define LOOP_VINFO_FULLY_WITH_LENGTH_P(L)       \
    1358              :   (LOOP_VINFO_USING_PARTIAL_VECTORS_P (L)       \
    1359              :    && !LOOP_VINFO_LENS (L).is_empty ())
    1360              : 
    1361              : #define LOOP_REQUIRES_VERSIONING_FOR_ALIGNMENT(L)       \
    1362              :   ((L)->may_misalign_stmts.length () > 0)
    1363              : #define LOOP_REQUIRES_VERSIONING_FOR_SPEC_READ(L)       \
    1364              :   (maybe_gt ((L)->max_spec_read_amount, 0U))
    1365              : #define LOOP_REQUIRES_VERSIONING_FOR_ALIAS(L)           \
    1366              :   ((L)->comp_alias_ddrs.length () > 0 \
    1367              :    || (L)->check_unequal_addrs.length () > 0 \
    1368              :    || (L)->lower_bounds.length () > 0)
    1369              : #define LOOP_REQUIRES_VERSIONING_FOR_NITERS(L)          \
    1370              :   (LOOP_VINFO_NITERS_ASSUMPTIONS (L))
    1371              : #define LOOP_REQUIRES_VERSIONING_FOR_SIMD_IF_COND(L)    \
    1372              :   (LOOP_VINFO_SIMD_IF_COND (L))
    1373              : #define LOOP_REQUIRES_VERSIONING(L)                     \
    1374              :   (LOOP_REQUIRES_VERSIONING_FOR_ALIGNMENT (L)           \
    1375              :    || LOOP_REQUIRES_VERSIONING_FOR_SPEC_READ (L)        \
    1376              :    || LOOP_REQUIRES_VERSIONING_FOR_ALIAS (L)            \
    1377              :    || LOOP_REQUIRES_VERSIONING_FOR_NITERS (L)           \
    1378              :    || LOOP_REQUIRES_VERSIONING_FOR_SIMD_IF_COND (L))
    1379              : 
    1380              : #define LOOP_VINFO_USE_VERSIONING_WITHOUT_PEELING(L)    \
    1381              :   ((L)->may_misalign_stmts.length () > 0          \
    1382              :    && !LOOP_VINFO_ALLOW_MUTUAL_ALIGNMENT (L))
    1383              : 
    1384              : #define LOOP_VINFO_NITERS_KNOWN_P(L)          \
    1385              :   (tree_fits_shwi_p ((L)->num_iters) && tree_to_shwi ((L)->num_iters) > 0)
    1386              : 
    1387              : #define LOOP_VINFO_EPILOGUE_P(L) \
    1388              :   (LOOP_VINFO_ORIG_LOOP_INFO (L) != NULL)
    1389              : 
    1390              : #define LOOP_VINFO_ORIG_MAX_VECT_FACTOR(L) \
    1391              :   (LOOP_VINFO_MAX_VECT_FACTOR (LOOP_VINFO_ORIG_LOOP_INFO (L)))
    1392              : 
    1393              : /* Wrapper for loop_vec_info, for tracking success/failure, where a non-NULL
    1394              :    value signifies success, and a NULL value signifies failure, supporting
    1395              :    propagating an opt_problem * describing the failure back up the call
    1396              :    stack.  */
    1397              : typedef opt_pointer_wrapper <loop_vec_info> opt_loop_vec_info;
    1398              : 
    1399              : inline loop_vec_info
    1400       532247 : loop_vec_info_for_loop (class loop *loop)
    1401              : {
    1402       532247 :   return (loop_vec_info) loop->aux;
    1403              : }
    1404              : 
    1405              : struct slp_root
    1406              : {
    1407      1219231 :   slp_root (slp_instance_kind kind_, vec<stmt_vec_info> stmts_,
    1408        12824 :             vec<stmt_vec_info> roots_, vec<tree> remain_ = vNULL)
    1409      1219231 :     : kind(kind_), stmts(stmts_), roots(roots_), remain(remain_) {}
    1410              :   slp_instance_kind kind;
    1411              :   vec<stmt_vec_info> stmts;
    1412              :   vec<stmt_vec_info> roots;
    1413              :   vec<tree> remain;
    1414              : };
    1415              : 
    1416              : typedef class _bb_vec_info : public vec_info
    1417              : {
    1418              : public:
    1419              :   _bb_vec_info (vec<basic_block> bbs, vec_info_shared *);
    1420              :   ~_bb_vec_info ();
    1421              : 
    1422              :   vec<slp_root> roots;
    1423              : } *bb_vec_info;
    1424              : 
    1425              : #define BB_VINFO_BBS(B)              (B)->bbs
    1426              : #define BB_VINFO_NBBS(B)             (B)->nbbs
    1427              : #define BB_VINFO_GROUPED_STORES(B)   (B)->grouped_stores
    1428              : #define BB_VINFO_SLP_INSTANCES(B)    (B)->slp_instances
    1429              : #define BB_VINFO_DATAREFS(B)         (B)->shared->datarefs
    1430              : #define BB_VINFO_DDRS(B)             (B)->shared->ddrs
    1431              : 
    1432              : /* Indicates whether/how a variable is used in the scope of loop/basic
    1433              :    block.  */
    1434              : enum vect_relevant {
    1435              :   vect_unused_in_scope = 0,
    1436              : 
    1437              :   /* The def is only used outside the loop.  */
    1438              :   vect_used_only_live,
    1439              :   /* The def is in the inner loop, and the use is in the outer loop, and the
    1440              :      use is a reduction stmt.  */
    1441              :   vect_used_in_outer_by_reduction,
    1442              :   /* The def is in the inner loop, and the use is in the outer loop (and is
    1443              :      not part of reduction).  */
    1444              :   vect_used_in_outer,
    1445              : 
    1446              :   /* defs that feed computations that end up (only) in a reduction. These
    1447              :      defs may be used by non-reduction stmts, but eventually, any
    1448              :      computations/values that are affected by these defs are used to compute
    1449              :      a reduction (i.e. don't get stored to memory, for example). We use this
    1450              :      to identify computations that we can change the order in which they are
    1451              :      computed.  */
    1452              :   vect_used_by_reduction,
    1453              : 
    1454              :   vect_used_in_scope
    1455              : };
    1456              : 
    1457              : /* The type of vectorization.  pure_slp means the stmt is covered by the
    1458              :    SLP graph, not_vect means it is not.  This is mostly used by BB
    1459              :    vectorization.  */
    1460              : enum slp_vect_type {
    1461              :   not_vect = 0,
    1462              :   pure_slp,
    1463              : };
    1464              : 
    1465              : /* Says whether a statement is a load, a store of a vectorized statement
    1466              :    result, or a store of an invariant value.  */
    1467              : enum vec_load_store_type {
    1468              :   VLS_LOAD,
    1469              :   VLS_STORE,
    1470              :   VLS_STORE_INVARIANT
    1471              : };
    1472              : 
    1473              : class dr_vec_info {
    1474              : public:
    1475              :   /* The data reference itself.  */
    1476              :   data_reference *dr;
    1477              :   /* The statement that contains the data reference.  */
    1478              :   stmt_vec_info stmt;
    1479              :   /* The analysis group this DR belongs to when doing BB vectorization.
    1480              :      DRs of the same group belong to the same conditional execution context.  */
    1481              :   unsigned group;
    1482              :   /* The misalignment in bytes of the reference, or -1 if not known.  */
    1483              :   int misalignment;
    1484              :   /* The byte alignment that we'd ideally like the reference to have,
    1485              :      and the value that misalignment is measured against.  */
    1486              :   poly_uint64 target_alignment;
    1487              :   /* If true the alignment of base_decl needs to be increased.  */
    1488              :   bool base_misaligned;
    1489              : 
    1490              :   /* Set by early break vectorization when this DR needs peeling for alignment
    1491              :      for correctness.  */
    1492              :   bool safe_speculative_read_required;
    1493              : 
    1494              :   /* Set by early break vectorization when this DR's scalar accesses are known
    1495              :      to be inbounds of a known bounds loop.  */
    1496              :   bool scalar_access_known_in_bounds;
    1497              : 
    1498              :   tree base_decl;
    1499              : 
    1500              :   /* Stores current vectorized loop's offset.  To be added to the DR's
    1501              :      offset to calculate current offset of data reference.  */
    1502              :   tree offset;
    1503              : };
    1504              : 
    1505              : typedef struct data_reference *dr_p;
    1506              : 
    1507              : class _stmt_vec_info {
    1508              : public:
    1509              : 
    1510              :   /* Indicates whether this stmts is part of a computation whose result is
    1511              :      used outside the loop.  */
    1512              :   bool live;
    1513              : 
    1514              :   /* Stmt is part of some pattern (computation idiom)  */
    1515              :   bool in_pattern_p;
    1516              : 
    1517              :   /* True if the statement was created during pattern recognition as
    1518              :      part of the replacement for RELATED_STMT.  This implies that the
    1519              :      statement isn't part of any basic block, although for convenience
    1520              :      its gimple_bb is the same as for RELATED_STMT.  */
    1521              :   bool pattern_stmt_p;
    1522              : 
    1523              :   /* Is this statement vectorizable or should it be skipped in (partial)
    1524              :      vectorization.  */
    1525              :   bool vectorizable;
    1526              : 
    1527              :   /* The stmt to which this info struct refers to.  */
    1528              :   gimple *stmt;
    1529              : 
    1530              :   /* The vector type to be used for the LHS of this statement.  */
    1531              :   tree vectype;
    1532              : 
    1533              :   /* The following is relevant only for stmts that contain a non-scalar
    1534              :      data-ref (array/pointer/struct access). A GIMPLE stmt is expected to have
    1535              :      at most one such data-ref.  */
    1536              : 
    1537              :   dr_vec_info dr_aux;
    1538              : 
    1539              :   /* Information about the data-ref relative to this loop
    1540              :      nest (the loop that is being considered for vectorization).  */
    1541              :   innermost_loop_behavior dr_wrt_vec_loop;
    1542              : 
    1543              :   /* For loop PHI nodes, the base and evolution part of it.  This makes sure
    1544              :      this information is still available in vect_update_ivs_after_vectorizer
    1545              :      where we may not be able to re-analyze the PHI nodes evolution as
    1546              :      peeling for the prologue loop can make it unanalyzable.  The evolution
    1547              :      part is still correct after peeling, but the base may have changed from
    1548              :      the version here.  */
    1549              :   tree loop_phi_evolution_base_unchanged;
    1550              :   tree loop_phi_evolution_part;
    1551              :   enum vect_induction_op_type loop_phi_evolution_type;
    1552              : 
    1553              :   /* Used for various bookkeeping purposes, generally holding a pointer to
    1554              :      some other stmt S that is in some way "related" to this stmt.
    1555              :      Current use of this field is:
    1556              :         If this stmt is part of a pattern (i.e. the field 'in_pattern_p' is
    1557              :         true): S is the "pattern stmt" that represents (and replaces) the
    1558              :         sequence of stmts that constitutes the pattern.  Similarly, the
    1559              :         related_stmt of the "pattern stmt" points back to this stmt (which is
    1560              :         the last stmt in the original sequence of stmts that constitutes the
    1561              :         pattern).  */
    1562              :   stmt_vec_info related_stmt;
    1563              : 
    1564              :   /* Used to keep a sequence of def stmts of a pattern stmt if such exists.
    1565              :      The sequence is attached to the original statement rather than the
    1566              :      pattern statement.  */
    1567              :   gimple_seq pattern_def_seq;
    1568              : 
    1569              :   /* Classify the def of this stmt.  */
    1570              :   enum vect_def_type def_type;
    1571              : 
    1572              :   /*  Whether the stmt is SLPed, loop-based vectorized, or both.  */
    1573              :   enum slp_vect_type slp_type;
    1574              : 
    1575              :   /* Interleaving chains info.  */
    1576              :   /* First element in the group.  */
    1577              :   stmt_vec_info first_element;
    1578              :   /* Pointer to the next element in the group.  */
    1579              :   stmt_vec_info next_element;
    1580              :   /* The size of the group.  */
    1581              :   unsigned int size;
    1582              :   /* For loads only, the gap from the previous load. For consecutive loads, GAP
    1583              :      is 1.  */
    1584              :   unsigned int gap;
    1585              : 
    1586              :   /* The minimum negative dependence distance this stmt participates in
    1587              :      or zero if none.  */
    1588              :   unsigned int min_neg_dist;
    1589              : 
    1590              :   /* Not all stmts in the loop need to be vectorized. e.g, the increment
    1591              :      of the loop induction variable and computation of array indexes. relevant
    1592              :      indicates whether the stmt needs to be vectorized.  */
    1593              :   enum vect_relevant relevant;
    1594              : 
    1595              :   /* For loads if this is a gather, for stores if this is a scatter.  */
    1596              :   bool gather_scatter_p;
    1597              : 
    1598              :   /* True if this is an access with loop-invariant stride.  */
    1599              :   bool strided_p;
    1600              : 
    1601              :   /* For both loads and stores.  */
    1602              :   unsigned simd_lane_access_p : 3;
    1603              : 
    1604              :   /* On a reduction PHI the reduction type as detected by
    1605              :      vect_is_simple_reduction.  */
    1606              :   enum vect_reduction_type reduc_type;
    1607              : 
    1608              :   /* On a reduction PHI, the original reduction code as detected by
    1609              :      vect_is_simple_reduction.  */
    1610              :   code_helper reduc_code;
    1611              : 
    1612              :   /* On a stmt participating in a reduction the index of the operand
    1613              :      on the reduction SSA cycle.  */
    1614              :   int reduc_idx;
    1615              : 
    1616              :   /* On a reduction PHI the def returned by vect_is_simple_reduction.
    1617              :      On the def returned by vect_is_simple_reduction the corresponding PHI.  */
    1618              :   stmt_vec_info reduc_def;
    1619              : 
    1620              :   /* If nonzero, the lhs of the statement could be truncated to this
    1621              :      many bits without affecting any users of the result.  */
    1622              :   unsigned int min_output_precision;
    1623              : 
    1624              :   /* If nonzero, all non-boolean input operands have the same precision,
    1625              :      and they could each be truncated to this many bits without changing
    1626              :      the result.  */
    1627              :   unsigned int min_input_precision;
    1628              : 
    1629              :   /* If OPERATION_BITS is nonzero, the statement could be performed on
    1630              :      an integer with the sign and number of bits given by OPERATION_SIGN
    1631              :      and OPERATION_BITS without changing the result.  */
    1632              :   unsigned int operation_precision;
    1633              :   signop operation_sign;
    1634              : 
    1635              :   /* If the statement produces a boolean result, this value describes
    1636              :      how we should choose the associated vector type.  The possible
    1637              :      values are:
    1638              : 
    1639              :      - an integer precision N if we should use the vector mask type
    1640              :        associated with N-bit integers.  This is only used if all relevant
    1641              :        input booleans also want the vector mask type for N-bit integers,
    1642              :        or if we can convert them into that form by pattern-matching.
    1643              : 
    1644              :      - ~0U if we considered choosing a vector mask type but decided
    1645              :        to treat the boolean as a normal integer type instead.
    1646              : 
    1647              :      - 0 otherwise.  This means either that the operation isn't one that
    1648              :        could have a vector mask type (and so should have a normal vector
    1649              :        type instead) or that we simply haven't made a choice either way.  */
    1650              :   unsigned int mask_precision;
    1651              : 
    1652              :   /* True if this is only suitable for SLP vectorization.  */
    1653              :   bool slp_vect_only_p;
    1654              : };
    1655              : 
    1656              : /* Information about a gather/scatter call.  */
    1657              : struct gather_scatter_info {
    1658              :   /* The internal function to use for the gather/scatter operation,
    1659              :      or IFN_LAST if a built-in function should be used instead.  */
    1660              :   internal_fn ifn;
    1661              : 
    1662              :   /* The FUNCTION_DECL for the built-in gather/scatter function,
    1663              :      or null if an internal function should be used instead.  */
    1664              :   tree decl;
    1665              : 
    1666              :   /* The loop-invariant base value.  */
    1667              :   tree base;
    1668              : 
    1669              :   /* The TBBA alias pointer the value of which determines the alignment
    1670              :      of the scalar accesses.  */
    1671              :   tree alias_ptr;
    1672              : 
    1673              :   /* The original scalar offset, which is a non-loop-invariant SSA_NAME.  */
    1674              :   tree offset;
    1675              : 
    1676              :   /* Each offset element should be multiplied by this amount before
    1677              :      being added to the base.  */
    1678              :   int scale;
    1679              : 
    1680              :   /* The type of the vectorized offset.  */
    1681              :   tree offset_vectype;
    1682              : 
    1683              :   /* The type of the scalar elements after loading or before storing.  */
    1684              :   tree element_type;
    1685              : 
    1686              :   /* The type of the scalar elements being loaded or stored.  */
    1687              :   tree memory_type;
    1688              : };
    1689              : 
    1690              : /* Access Functions.  */
    1691              : #define STMT_VINFO_STMT(S)                 (S)->stmt
    1692              : #define STMT_VINFO_RELEVANT(S)             (S)->relevant
    1693              : #define STMT_VINFO_LIVE_P(S)               (S)->live
    1694              : #define STMT_VINFO_VECTYPE(S)              (S)->vectype
    1695              : #define STMT_VINFO_VECTORIZABLE(S)         (S)->vectorizable
    1696              : #define STMT_VINFO_DATA_REF(S)             ((S)->dr_aux.dr + 0)
    1697              : #define STMT_VINFO_GATHER_SCATTER_P(S)     (S)->gather_scatter_p
    1698              : #define STMT_VINFO_STRIDED_P(S)            (S)->strided_p
    1699              : #define STMT_VINFO_SIMD_LANE_ACCESS_P(S)   (S)->simd_lane_access_p
    1700              : #define STMT_VINFO_REDUC_IDX(S)            (S)->reduc_idx
    1701              : 
    1702              : #define STMT_VINFO_DR_WRT_VEC_LOOP(S)      (S)->dr_wrt_vec_loop
    1703              : #define STMT_VINFO_DR_BASE_ADDRESS(S)      (S)->dr_wrt_vec_loop.base_address
    1704              : #define STMT_VINFO_DR_INIT(S)              (S)->dr_wrt_vec_loop.init
    1705              : #define STMT_VINFO_DR_OFFSET(S)            (S)->dr_wrt_vec_loop.offset
    1706              : #define STMT_VINFO_DR_STEP(S)              (S)->dr_wrt_vec_loop.step
    1707              : #define STMT_VINFO_DR_BASE_ALIGNMENT(S)    (S)->dr_wrt_vec_loop.base_alignment
    1708              : #define STMT_VINFO_DR_BASE_MISALIGNMENT(S) \
    1709              :   (S)->dr_wrt_vec_loop.base_misalignment
    1710              : #define STMT_VINFO_DR_OFFSET_ALIGNMENT(S) \
    1711              :   (S)->dr_wrt_vec_loop.offset_alignment
    1712              : #define STMT_VINFO_DR_STEP_ALIGNMENT(S) \
    1713              :   (S)->dr_wrt_vec_loop.step_alignment
    1714              : 
    1715              : #define STMT_VINFO_DR_INFO(S) \
    1716              :   (gcc_checking_assert ((S)->dr_aux.stmt == (S)), &(S)->dr_aux)
    1717              : 
    1718              : #define STMT_VINFO_IN_PATTERN_P(S)         (S)->in_pattern_p
    1719              : #define STMT_VINFO_RELATED_STMT(S)         (S)->related_stmt
    1720              : #define STMT_VINFO_PATTERN_DEF_SEQ(S)      (S)->pattern_def_seq
    1721              : #define STMT_VINFO_DEF_TYPE(S)             (S)->def_type
    1722              : #define STMT_VINFO_GROUPED_ACCESS(S) \
    1723              :   ((S)->dr_aux.dr && DR_GROUP_FIRST_ELEMENT(S))
    1724              : #define STMT_VINFO_LOOP_PHI_EVOLUTION_BASE_UNCHANGED(S) (S)->loop_phi_evolution_base_unchanged
    1725              : #define STMT_VINFO_LOOP_PHI_EVOLUTION_PART(S) (S)->loop_phi_evolution_part
    1726              : #define STMT_VINFO_LOOP_PHI_EVOLUTION_TYPE(S) (S)->loop_phi_evolution_type
    1727              : #define STMT_VINFO_MIN_NEG_DIST(S)      (S)->min_neg_dist
    1728              : #define STMT_VINFO_REDUC_TYPE(S)        (S)->reduc_type
    1729              : #define STMT_VINFO_REDUC_CODE(S)        (S)->reduc_code
    1730              : #define STMT_VINFO_REDUC_DEF(S)         (S)->reduc_def
    1731              : #define STMT_VINFO_SLP_VECT_ONLY(S)     (S)->slp_vect_only_p
    1732              : #define STMT_VINFO_REDUC_VECTYPE_IN(S)  (S)->reduc_vectype_in
    1733              : 
    1734              : #define DR_GROUP_FIRST_ELEMENT(S) \
    1735              :   (gcc_checking_assert ((S)->dr_aux.dr), (S)->first_element)
    1736              : #define DR_GROUP_NEXT_ELEMENT(S) \
    1737              :   (gcc_checking_assert ((S)->dr_aux.dr), (S)->next_element)
    1738              : #define DR_GROUP_SIZE(S) \
    1739              :   (gcc_checking_assert ((S)->dr_aux.dr), (S)->size)
    1740              : #define DR_GROUP_GAP(S) \
    1741              :   (gcc_checking_assert ((S)->dr_aux.dr), (S)->gap)
    1742              : 
    1743              : #define STMT_VINFO_RELEVANT_P(S)          ((S)->relevant != vect_unused_in_scope)
    1744              : 
    1745              : #define PURE_SLP_STMT(S)                  ((S)->slp_type == pure_slp)
    1746              : #define STMT_SLP_TYPE(S)                   (S)->slp_type
    1747              : 
    1748              : 
    1749              : /* Contains the scalar or vector costs for a vec_info.  */
    1750              : class vector_costs
    1751              : {
    1752              : public:
    1753              :   vector_costs (vec_info *, bool);
    1754            0 :   virtual ~vector_costs () {}
    1755              : 
    1756              :   /* Update the costs in response to adding COUNT copies of a statement.
    1757              : 
    1758              :      - WHERE specifies whether the cost occurs in the loop prologue,
    1759              :        the loop body, or the loop epilogue.
    1760              :      - KIND is the kind of statement, which is always meaningful.
    1761              :      - STMT_INFO or NODE, if nonnull, describe the statement that will be
    1762              :        vectorized.
    1763              :      - VECTYPE, if nonnull, is the vector type that the vectorized
    1764              :        statement will operate on.  Note that this should be used in
    1765              :        preference to STMT_VINFO_VECTYPE (STMT_INFO) since the latter
    1766              :        is not correct for SLP.
    1767              :      - for unaligned_load and unaligned_store statements, MISALIGN is
    1768              :        the byte misalignment of the load or store relative to the target's
    1769              :        preferred alignment for VECTYPE, or DR_MISALIGNMENT_UNKNOWN
    1770              :        if the misalignment is not known.
    1771              : 
    1772              :      Return the calculated cost as well as recording it.  The return
    1773              :      value is used for dumping purposes.  */
    1774              :   virtual unsigned int add_stmt_cost (int count, vect_cost_for_stmt kind,
    1775              :                                       stmt_vec_info stmt_info,
    1776              :                                       slp_tree node,
    1777              :                                       tree vectype, int misalign,
    1778              :                                       vect_cost_model_location where);
    1779              : 
    1780              :   /* Finish calculating the cost of the code.  The results can be
    1781              :      read back using the functions below.
    1782              : 
    1783              :      If the costs describe vector code, SCALAR_COSTS gives the costs
    1784              :      of the corresponding scalar code, otherwise it is null.  */
    1785              :   virtual void finish_cost (const vector_costs *scalar_costs);
    1786              : 
    1787              :   /* The costs in THIS and OTHER both describe ways of vectorizing
    1788              :      a main loop.  Return true if the costs described by THIS are
    1789              :      cheaper than the costs described by OTHER.  Return false if any
    1790              :      of the following are true:
    1791              : 
    1792              :      - THIS and OTHER are of equal cost
    1793              :      - OTHER is better than THIS
    1794              :      - we can't be sure about the relative costs of THIS and OTHER.  */
    1795              :   virtual bool better_main_loop_than_p (const vector_costs *other) const;
    1796              : 
    1797              :   /* Likewise, but the costs in THIS and OTHER both describe ways of
    1798              :      vectorizing an epilogue loop of MAIN_LOOP.  */
    1799              :   virtual bool better_epilogue_loop_than_p (const vector_costs *other,
    1800              :                                             loop_vec_info main_loop) const;
    1801              : 
    1802              :   unsigned int prologue_cost () const;
    1803              :   unsigned int body_cost () const;
    1804              :   unsigned int epilogue_cost () const;
    1805              :   unsigned int outside_cost () const;
    1806              :   unsigned int total_cost () const;
    1807              : 
    1808              :   unsigned int suggested_unroll_factor () const;
    1809              :   machine_mode suggested_epilogue_mode (int &masked) const;
    1810              : 
    1811        32297 :   vec_info *vinfo () const { return m_vinfo; }
    1812      7447795 :   bool costing_for_scalar () const { return m_costing_for_scalar; }
    1813              : 
    1814              : protected:
    1815              :   unsigned int record_stmt_cost (stmt_vec_info, vect_cost_model_location,
    1816              :                                  unsigned int);
    1817              :   unsigned int adjust_cost_for_freq (stmt_vec_info, vect_cost_model_location,
    1818              :                                      unsigned int);
    1819              :   int compare_inside_loop_cost (const vector_costs *) const;
    1820              :   int compare_outside_loop_cost (const vector_costs *) const;
    1821              : 
    1822              :   /* The region of code that we're considering vectorizing.  */
    1823              :   vec_info *m_vinfo;
    1824              : 
    1825              :   /* True if we're costing the scalar code, false if we're costing
    1826              :      the vector code.  */
    1827              :   bool m_costing_for_scalar;
    1828              : 
    1829              :   /* The costs of the three regions, indexed by vect_cost_model_location.  */
    1830              :   unsigned int m_costs[3];
    1831              : 
    1832              :   /* The suggested unrolling factor determined at finish_cost.  */
    1833              :   unsigned int m_suggested_unroll_factor;
    1834              : 
    1835              :   /* The suggested mode to be used for a vectorized epilogue or VOIDmode,
    1836              :      determined at finish_cost.  m_masked_epilogue specifies whether the
    1837              :      epilogue should use masked vectorization, regardless of the
    1838              :      --param vect-partial-vector-usage default.  If -1 then the
    1839              :      --param setting takes precedence.  If the user explicitly specified
    1840              :      --param vect-partial-vector-usage then that takes precedence.  */
    1841              :   machine_mode m_suggested_epilogue_mode;
    1842              :   int m_masked_epilogue;
    1843              : 
    1844              :   /* True if finish_cost has been called.  */
    1845              :   bool m_finished;
    1846              : };
    1847              : 
    1848              : /* Create costs for VINFO.  COSTING_FOR_SCALAR is true if the costs
    1849              :    are for scalar code, false if they are for vector code.  */
    1850              : 
    1851              : inline
    1852      2086803 : vector_costs::vector_costs (vec_info *vinfo, bool costing_for_scalar)
    1853      2086803 :   : m_vinfo (vinfo),
    1854      2086803 :     m_costing_for_scalar (costing_for_scalar),
    1855      2086803 :     m_costs (),
    1856      2086803 :     m_suggested_unroll_factor(1),
    1857      2086803 :     m_suggested_epilogue_mode(VOIDmode),
    1858      2086803 :     m_masked_epilogue (-1),
    1859      2086803 :     m_finished (false)
    1860              : {
    1861              : }
    1862              : 
    1863              : /* Return the cost of the prologue code (in abstract units).  */
    1864              : 
    1865              : inline unsigned int
    1866      1265206 : vector_costs::prologue_cost () const
    1867              : {
    1868      1265206 :   gcc_checking_assert (m_finished);
    1869      1265206 :   return m_costs[vect_prologue];
    1870              : }
    1871              : 
    1872              : /* Return the cost of the body code (in abstract units).  */
    1873              : 
    1874              : inline unsigned int
    1875      1984332 : vector_costs::body_cost () const
    1876              : {
    1877      1984332 :   gcc_checking_assert (m_finished);
    1878      1984332 :   return m_costs[vect_body];
    1879              : }
    1880              : 
    1881              : /* Return the cost of the epilogue code (in abstract units).  */
    1882              : 
    1883              : inline unsigned int
    1884      1265206 : vector_costs::epilogue_cost () const
    1885              : {
    1886      1265206 :   gcc_checking_assert (m_finished);
    1887      1265206 :   return m_costs[vect_epilogue];
    1888              : }
    1889              : 
    1890              : /* Return the cost of the prologue and epilogue code (in abstract units).  */
    1891              : 
    1892              : inline unsigned int
    1893       482988 : vector_costs::outside_cost () const
    1894              : {
    1895       482988 :   return prologue_cost () + epilogue_cost ();
    1896              : }
    1897              : 
    1898              : /* Return the cost of the prologue, body and epilogue code
    1899              :    (in abstract units).  */
    1900              : 
    1901              : inline unsigned int
    1902       124472 : vector_costs::total_cost () const
    1903              : {
    1904       124472 :   return body_cost () + outside_cost ();
    1905              : }
    1906              : 
    1907              : /* Return the suggested unroll factor.  */
    1908              : 
    1909              : inline unsigned int
    1910       124104 : vector_costs::suggested_unroll_factor () const
    1911              : {
    1912       124104 :   gcc_checking_assert (m_finished);
    1913       124104 :   return m_suggested_unroll_factor;
    1914              : }
    1915              : 
    1916              : /* Return the suggested epilogue mode.  */
    1917              : 
    1918              : inline machine_mode
    1919        14266 : vector_costs::suggested_epilogue_mode (int &masked_p) const
    1920              : {
    1921        14266 :   gcc_checking_assert (m_finished);
    1922        14266 :   masked_p = m_masked_epilogue;
    1923        14266 :   return m_suggested_epilogue_mode;
    1924              : }
    1925              : 
    1926              : #define VECT_MAX_COST 1000
    1927              : 
    1928              : /* The maximum number of intermediate steps required in multi-step type
    1929              :    conversion.  */
    1930              : #define MAX_INTERM_CVT_STEPS         3
    1931              : 
    1932              : #define MAX_VECTORIZATION_FACTOR INT_MAX
    1933              : 
    1934              : /* Nonzero if TYPE represents a (scalar) boolean type or type
    1935              :    in the middle-end compatible with it (unsigned precision 1 integral
    1936              :    types).  Used to determine which types should be vectorized as
    1937              :    VECTOR_BOOLEAN_TYPE_P.  */
    1938              : 
    1939              : #define VECT_SCALAR_BOOLEAN_TYPE_P(TYPE) \
    1940              :   (TREE_CODE (TYPE) == BOOLEAN_TYPE             \
    1941              :    || ((TREE_CODE (TYPE) == INTEGER_TYPE        \
    1942              :         || TREE_CODE (TYPE) == ENUMERAL_TYPE)   \
    1943              :        && TYPE_PRECISION (TYPE) == 1            \
    1944              :        && TYPE_UNSIGNED (TYPE)))
    1945              : 
    1946              : inline bool
    1947     11422193 : nested_in_vect_loop_p (class loop *loop, stmt_vec_info stmt_info)
    1948              : {
    1949     11422193 :   return (loop->inner
    1950      9279050 :           && (loop->inner == (gimple_bb (stmt_info->stmt))->loop_father));
    1951              : }
    1952              : 
    1953              : /* PHI is either a scalar reduction phi or a scalar induction phi.
    1954              :    Return the initial value of the variable on entry to the containing
    1955              :    loop.  */
    1956              : 
    1957              : inline tree
    1958        33842 : vect_phi_initial_value (gphi *phi)
    1959              : {
    1960        33842 :   basic_block bb = gimple_bb (phi);
    1961        33842 :   edge pe = loop_preheader_edge (bb->loop_father);
    1962        33842 :   gcc_assert (pe->dest == bb);
    1963        33842 :   return PHI_ARG_DEF_FROM_EDGE (phi, pe);
    1964              : }
    1965              : 
    1966              : /* Return true if STMT_INFO should produce a vector mask type rather than
    1967              :    a normal nonmask type.  */
    1968              : 
    1969              : inline bool
    1970      7076980 : vect_use_mask_type_p (stmt_vec_info stmt_info)
    1971              : {
    1972      7076980 :   return stmt_info->mask_precision && stmt_info->mask_precision != ~0U;
    1973              : }
    1974              : 
    1975              : /* Return TRUE if a statement represented by STMT_INFO is a part of a
    1976              :    pattern.  */
    1977              : 
    1978              : inline bool
    1979    127180196 : is_pattern_stmt_p (stmt_vec_info stmt_info)
    1980              : {
    1981     81153847 :   return stmt_info->pattern_stmt_p;
    1982              : }
    1983              : 
    1984              : /* If STMT_INFO is a pattern statement, return the statement that it
    1985              :    replaces, otherwise return STMT_INFO itself.  */
    1986              : 
    1987              : inline stmt_vec_info
    1988     50312438 : vect_orig_stmt (stmt_vec_info stmt_info)
    1989              : {
    1990     37905035 :   if (is_pattern_stmt_p (stmt_info))
    1991      3430556 :     return STMT_VINFO_RELATED_STMT (stmt_info);
    1992              :   return stmt_info;
    1993              : }
    1994              : 
    1995              : /* Return the later statement between STMT1_INFO and STMT2_INFO.  */
    1996              : 
    1997              : inline stmt_vec_info
    1998      6064605 : get_later_stmt (stmt_vec_info stmt1_info, stmt_vec_info stmt2_info)
    1999              : {
    2000      6064605 :   gimple *stmt1 = vect_orig_stmt (stmt1_info)->stmt;
    2001      6064605 :   gimple *stmt2 = vect_orig_stmt (stmt2_info)->stmt;
    2002      6064605 :   if (gimple_bb (stmt1) == gimple_bb (stmt2))
    2003              :     {
    2004      6038988 :       if (gimple_uid (stmt1) > gimple_uid (stmt2))
    2005              :         return stmt1_info;
    2006              :       else
    2007              :         return stmt2_info;
    2008              :     }
    2009              :   /* ???  We should be really calling this function only with stmts
    2010              :      in the same BB but we can recover if there's a domination
    2011              :      relationship between them.  */
    2012        25617 :   else if (dominated_by_p (CDI_DOMINATORS,
    2013        25617 :                            gimple_bb (stmt1), gimple_bb (stmt2)))
    2014              :     return stmt1_info;
    2015         8922 :   else if (dominated_by_p (CDI_DOMINATORS,
    2016         8922 :                            gimple_bb (stmt2), gimple_bb (stmt1)))
    2017              :     return stmt2_info;
    2018            0 :   gcc_unreachable ();
    2019              : }
    2020              : 
    2021              : /* If STMT_INFO has been replaced by a pattern statement, return the
    2022              :    replacement statement, otherwise return STMT_INFO itself.  */
    2023              : 
    2024              : inline stmt_vec_info
    2025     51601174 : vect_stmt_to_vectorize (stmt_vec_info stmt_info)
    2026              : {
    2027     51601174 :   if (STMT_VINFO_IN_PATTERN_P (stmt_info))
    2028      1547449 :     return STMT_VINFO_RELATED_STMT (stmt_info);
    2029              :   return stmt_info;
    2030              : }
    2031              : 
    2032              : /* Return true if BB is a loop header.  */
    2033              : 
    2034              : inline bool
    2035      1515808 : is_loop_header_bb_p (basic_block bb)
    2036              : {
    2037      1515808 :   if (bb == (bb->loop_father)->header)
    2038      1505126 :     return true;
    2039              : 
    2040              :   return false;
    2041              : }
    2042              : 
    2043              : /* Return pow2 (X).  */
    2044              : 
    2045              : inline int
    2046              : vect_pow2 (int x)
    2047              : {
    2048              :   int i, res = 1;
    2049              : 
    2050              :   for (i = 0; i < x; i++)
    2051              :     res *= 2;
    2052              : 
    2053              :   return res;
    2054              : }
    2055              : 
    2056              : /* Alias targetm.vectorize.builtin_vectorization_cost.  */
    2057              : 
    2058              : inline int
    2059      9122185 : builtin_vectorization_cost (enum vect_cost_for_stmt type_of_cost,
    2060              :                             tree vectype, int misalign)
    2061              : {
    2062      9055038 :   return targetm.vectorize.builtin_vectorization_cost (type_of_cost,
    2063              :                                                        vectype, misalign);
    2064              : }
    2065              : 
    2066              : /* Get cost by calling cost target builtin.  */
    2067              : 
    2068              : inline
    2069          152 : int vect_get_stmt_cost (enum vect_cost_for_stmt type_of_cost)
    2070              : {
    2071        66995 :   return builtin_vectorization_cost (type_of_cost, NULL, 0);
    2072              : }
    2073              : 
    2074              : /* Alias targetm.vectorize.init_cost.  */
    2075              : 
    2076              : inline vector_costs *
    2077      2086803 : init_cost (vec_info *vinfo, bool costing_for_scalar)
    2078              : {
    2079      2086803 :   return targetm.vectorize.create_costs (vinfo, costing_for_scalar);
    2080              : }
    2081              : 
    2082              : extern void dump_stmt_cost (FILE *, int, enum vect_cost_for_stmt,
    2083              :                             stmt_vec_info, slp_tree, tree, int, unsigned,
    2084              :                             enum vect_cost_model_location);
    2085              : 
    2086              : /* Dump and add costs.  */
    2087              : 
    2088              : inline unsigned
    2089      7447795 : add_stmt_cost (vector_costs *costs, int count,
    2090              :                enum vect_cost_for_stmt kind,
    2091              :                stmt_vec_info stmt_info, slp_tree node,
    2092              :                tree vectype, int misalign,
    2093              :                enum vect_cost_model_location where)
    2094              : {
    2095              :   /* Even though a vector type might be set on stmt do not pass that on when
    2096              :      costing the scalar IL.  A SLP node shouldn't have been recorded.  */
    2097      7447795 :   if (costs->costing_for_scalar ())
    2098              :     {
    2099      3861690 :       vectype = NULL_TREE;
    2100      3861690 :       gcc_checking_assert (node == NULL);
    2101              :     }
    2102      7447795 :   unsigned cost = costs->add_stmt_cost (count, kind, stmt_info, node, vectype,
    2103              :                                         misalign, where);
    2104      7447795 :   if (dump_file && (dump_flags & TDF_DETAILS))
    2105       218578 :     dump_stmt_cost (dump_file, count, kind, stmt_info, node, vectype, misalign,
    2106              :                     cost, where);
    2107      7447795 :   return cost;
    2108              : }
    2109              : 
    2110              : inline unsigned
    2111        81866 : add_stmt_cost (vector_costs *costs, int count, enum vect_cost_for_stmt kind,
    2112              :                enum vect_cost_model_location where)
    2113              : {
    2114        81866 :   gcc_assert (kind == cond_branch_taken || kind == cond_branch_not_taken
    2115              :               || kind == scalar_stmt);
    2116        81866 :   return add_stmt_cost (costs, count, kind, NULL, NULL, NULL_TREE, 0, where);
    2117              : }
    2118              : 
    2119              : inline unsigned
    2120      3834780 : add_stmt_cost (vector_costs *costs, stmt_info_for_cost *i)
    2121              : {
    2122      3834780 :   return add_stmt_cost (costs, i->count, i->kind, i->stmt_info, i->node,
    2123      3834780 :                         i->vectype, i->misalign, i->where);
    2124              : }
    2125              : 
    2126              : inline void
    2127       642041 : add_stmt_costs (vector_costs *costs, stmt_vector_for_cost *cost_vec)
    2128              : {
    2129       642041 :   stmt_info_for_cost *cost;
    2130       642041 :   unsigned i;
    2131      3860817 :   FOR_EACH_VEC_ELT (*cost_vec, i, cost)
    2132      3218776 :     add_stmt_cost (costs, cost->count, cost->kind, cost->stmt_info,
    2133              :                    cost->node, cost->vectype, cost->misalign, cost->where);
    2134       642041 : }
    2135              : 
    2136              : /*-----------------------------------------------------------------*/
    2137              : /* Info on data references alignment.                              */
    2138              : /*-----------------------------------------------------------------*/
    2139              : #define DR_MISALIGNMENT_UNKNOWN (-1)
    2140              : #define DR_MISALIGNMENT_UNINITIALIZED (-2)
    2141              : 
    2142              : inline void
    2143      2561891 : set_dr_misalignment (dr_vec_info *dr_info, int val)
    2144              : {
    2145      2561891 :   dr_info->misalignment = val;
    2146              : }
    2147              : 
    2148              : extern int dr_misalignment (dr_vec_info *dr_info, tree vectype,
    2149              :                             poly_int64 offset = 0);
    2150              : 
    2151              : #define SET_DR_MISALIGNMENT(DR, VAL) set_dr_misalignment (DR, VAL)
    2152              : 
    2153              : /* Only defined once DR_MISALIGNMENT is defined.  */
    2154              : inline const poly_uint64
    2155      7999700 : dr_target_alignment (dr_vec_info *dr_info)
    2156              : {
    2157      7999700 :   if (STMT_VINFO_GROUPED_ACCESS (dr_info->stmt))
    2158      5905013 :     dr_info = STMT_VINFO_DR_INFO (DR_GROUP_FIRST_ELEMENT (dr_info->stmt));
    2159      7999700 :   return dr_info->target_alignment;
    2160              : }
    2161              : #define DR_TARGET_ALIGNMENT(DR) dr_target_alignment (DR)
    2162              : #define DR_SCALAR_KNOWN_BOUNDS(DR) (DR)->scalar_access_known_in_bounds
    2163              : 
    2164              : /* Return if the stmt_vec_info requires peeling for alignment.  */
    2165              : inline bool
    2166      4457034 : dr_safe_speculative_read_required (stmt_vec_info stmt_info)
    2167              : {
    2168      4457034 :   dr_vec_info *dr_info;
    2169      4457034 :   if (STMT_VINFO_GROUPED_ACCESS (stmt_info))
    2170      1663860 :     dr_info = STMT_VINFO_DR_INFO (DR_GROUP_FIRST_ELEMENT (stmt_info));
    2171              :   else
    2172      2793174 :     dr_info = STMT_VINFO_DR_INFO (stmt_info);
    2173              : 
    2174      4457034 :   return dr_info->safe_speculative_read_required;
    2175              : }
    2176              : 
    2177              : /* Set the safe_speculative_read_required for the stmt_vec_info, if group
    2178              :    access then set on the fist element otherwise set on DR directly.  */
    2179              : inline void
    2180       225460 : dr_set_safe_speculative_read_required (stmt_vec_info stmt_info,
    2181              :                                        bool requires_alignment)
    2182              : {
    2183       225460 :   dr_vec_info *dr_info;
    2184       225460 :   if (STMT_VINFO_GROUPED_ACCESS (stmt_info))
    2185        67717 :     dr_info = STMT_VINFO_DR_INFO (DR_GROUP_FIRST_ELEMENT (stmt_info));
    2186              :   else
    2187       157743 :     dr_info = STMT_VINFO_DR_INFO (stmt_info);
    2188              : 
    2189       225460 :   dr_info->safe_speculative_read_required = requires_alignment;
    2190       225460 : }
    2191              : 
    2192              : inline void
    2193      1577343 : set_dr_target_alignment (dr_vec_info *dr_info, poly_uint64 val)
    2194              : {
    2195      1577343 :   dr_info->target_alignment = val;
    2196              : }
    2197              : #define SET_DR_TARGET_ALIGNMENT(DR, VAL) set_dr_target_alignment (DR, VAL)
    2198              : 
    2199              : /* Return true if data access DR_INFO is aligned to the targets
    2200              :    preferred alignment for VECTYPE (which may be less than a full vector).  */
    2201              : 
    2202              : inline bool
    2203       377013 : aligned_access_p (dr_vec_info *dr_info, tree vectype)
    2204              : {
    2205       377013 :   return (dr_misalignment (dr_info, vectype) == 0);
    2206              : }
    2207              : 
    2208              : /* Return TRUE if the (mis-)alignment of the data access is known with
    2209              :    respect to the targets preferred alignment for VECTYPE, and FALSE
    2210              :    otherwise.  */
    2211              : 
    2212              : inline bool
    2213      2247248 : known_alignment_for_access_p (dr_vec_info *dr_info, tree vectype)
    2214              : {
    2215      2011208 :   return (dr_misalignment (dr_info, vectype) != DR_MISALIGNMENT_UNKNOWN);
    2216              : }
    2217              : 
    2218              : /* Return the minimum alignment in bytes that the vectorized version
    2219              :    of DR_INFO is guaranteed to have.  */
    2220              : 
    2221              : inline unsigned int
    2222       273347 : vect_known_alignment_in_bytes (dr_vec_info *dr_info, tree vectype,
    2223              :                                poly_int64 offset = 0)
    2224              : {
    2225       273347 :   int misalignment = dr_misalignment (dr_info, vectype, offset);
    2226       273347 :   if (misalignment == DR_MISALIGNMENT_UNKNOWN)
    2227       133205 :     return TYPE_ALIGN_UNIT (TREE_TYPE (DR_REF (dr_info->dr)));
    2228       140142 :   else if (misalignment == 0)
    2229        97655 :     return known_alignment (DR_TARGET_ALIGNMENT (dr_info));
    2230        42487 :   return misalignment & -misalignment;
    2231              : }
    2232              : 
    2233              : /* Return the behavior of DR_INFO with respect to the vectorization context
    2234              :    (which for outer loop vectorization might not be the behavior recorded
    2235              :    in DR_INFO itself).  */
    2236              : 
    2237              : inline innermost_loop_behavior *
    2238      5603469 : vect_dr_behavior (vec_info *vinfo, dr_vec_info *dr_info)
    2239              : {
    2240      5603469 :   stmt_vec_info stmt_info = dr_info->stmt;
    2241      5603469 :   loop_vec_info loop_vinfo = dyn_cast<loop_vec_info> (vinfo);
    2242      2150334 :   if (loop_vinfo == NULL
    2243      2150334 :       || !nested_in_vect_loop_p (LOOP_VINFO_LOOP (loop_vinfo), stmt_info))
    2244      5599532 :     return &DR_INNERMOST (dr_info->dr);
    2245              :   else
    2246         3937 :     return &STMT_VINFO_DR_WRT_VEC_LOOP (stmt_info);
    2247              : }
    2248              : 
    2249              : /* Return the offset calculated by adding the offset of this DR_INFO to the
    2250              :    corresponding data_reference's offset.  If CHECK_OUTER then use
    2251              :    vect_dr_behavior to select the appropriate data_reference to use.  */
    2252              : 
    2253              : inline tree
    2254       733401 : get_dr_vinfo_offset (vec_info *vinfo,
    2255              :                      dr_vec_info *dr_info, bool check_outer = false)
    2256              : {
    2257       733401 :   innermost_loop_behavior *base;
    2258       733401 :   if (check_outer)
    2259       693615 :     base = vect_dr_behavior (vinfo, dr_info);
    2260              :   else
    2261        39786 :     base = &dr_info->dr->innermost;
    2262              : 
    2263       733401 :   tree offset = base->offset;
    2264              : 
    2265       733401 :   if (!dr_info->offset)
    2266              :     return offset;
    2267              : 
    2268        18958 :   offset = fold_convert (sizetype, offset);
    2269        18958 :   return fold_build2 (PLUS_EXPR, TREE_TYPE (dr_info->offset), offset,
    2270              :                       dr_info->offset);
    2271              : }
    2272              : 
    2273              : 
    2274              : /* Return the vect cost model for LOOP.  */
    2275              : inline enum vect_cost_model
    2276      2377073 : loop_cost_model (loop_p loop)
    2277              : {
    2278      2377073 :   if (loop != NULL
    2279      1719520 :       && loop->force_vectorize
    2280        77344 :       && flag_simd_cost_model != VECT_COST_MODEL_DEFAULT)
    2281              :     return flag_simd_cost_model;
    2282      2299729 :   return flag_vect_cost_model;
    2283              : }
    2284              : 
    2285              : /* Return true if the vect cost model is unlimited.  */
    2286              : inline bool
    2287      1650794 : unlimited_cost_model (loop_p loop)
    2288              : {
    2289      1650794 :   return loop_cost_model (loop) == VECT_COST_MODEL_UNLIMITED;
    2290              : }
    2291              : 
    2292              : /* Return true if the loop described by LOOP_VINFO is fully-masked and
    2293              :    if the first iteration should use a partial mask in order to achieve
    2294              :    alignment.  */
    2295              : 
    2296              : inline bool
    2297       264538 : vect_use_loop_mask_for_alignment_p (loop_vec_info loop_vinfo)
    2298              : {
    2299              :   /* With early break vectorization we don't know whether the accesses will stay
    2300              :      inside the loop or not.  TODO: The early break adjustment code can be
    2301              :      implemented the same way as vectorizable_linear_induction.  However we
    2302              :      can't test this today so reject it.  */
    2303           85 :   return (LOOP_VINFO_FULLY_MASKED_P (loop_vinfo)
    2304           85 :           && LOOP_VINFO_PEELING_FOR_ALIGNMENT (loop_vinfo)
    2305       264542 :           && !(LOOP_VINFO_NON_LINEAR_IV (loop_vinfo)
    2306            0 :                && LOOP_VINFO_EARLY_BREAKS (loop_vinfo)));
    2307              : }
    2308              : 
    2309              : /* Return the number of vectors of type VECTYPE that are needed to get
    2310              :    NUNITS elements.  NUNITS should be based on the vectorization factor,
    2311              :    so it is always a known multiple of the number of elements in VECTYPE.  */
    2312              : 
    2313              : inline unsigned int
    2314      7268395 : vect_get_num_vectors (poly_uint64 nunits, tree vectype)
    2315              : {
    2316      7268395 :   return exact_div (nunits, TYPE_VECTOR_SUBPARTS (vectype)).to_constant ();
    2317              : }
    2318              : 
    2319              : /* Return the number of vectors in the context of vectorization region VINFO,
    2320              :    needed for a group of statements and a vector type as specified by NODE.  */
    2321              : 
    2322              : inline unsigned int
    2323      7267566 : vect_get_num_copies (vec_info *vinfo, slp_tree node)
    2324              : {
    2325      7267566 :   poly_uint64 vf;
    2326              : 
    2327      7267566 :   if (loop_vec_info loop_vinfo = dyn_cast <loop_vec_info> (vinfo))
    2328      3305249 :     vf = LOOP_VINFO_VECT_FACTOR (loop_vinfo);
    2329              :   else
    2330              :     vf = 1;
    2331              : 
    2332      7267566 :   vf *= SLP_TREE_LANES (node);
    2333      7267566 :   tree vectype = SLP_TREE_VECTYPE (node);
    2334              : 
    2335      7267566 :   return vect_get_num_vectors (vf, vectype);
    2336              : }
    2337              : 
    2338              : /* Update maximum unit count *MAX_NUNITS so that it accounts for
    2339              :    NUNITS.  *MAX_NUNITS can be 1 if we haven't yet recorded anything.  */
    2340              : 
    2341              : inline void
    2342      9497957 : vect_update_max_nunits (poly_uint64 *max_nunits, poly_uint64 nunits)
    2343              : {
    2344              :   /* All unit counts have the form vec_info::vector_size * X for some
    2345              :      rational X, so two unit sizes must have a common multiple.
    2346              :      Everything is a multiple of the initial value of 1.  */
    2347      4184753 :   *max_nunits = force_common_multiple (*max_nunits, nunits);
    2348              : }
    2349              : 
    2350              : /* Update maximum unit count *MAX_NUNITS so that it accounts for
    2351              :    the number of units in vector type VECTYPE.  *MAX_NUNITS can be 1
    2352              :    if we haven't yet recorded any vector types.  */
    2353              : 
    2354              : inline void
    2355      5313204 : vect_update_max_nunits (poly_uint64 *max_nunits, tree vectype)
    2356              : {
    2357      5313204 :   vect_update_max_nunits (max_nunits, TYPE_VECTOR_SUBPARTS (vectype));
    2358      5313204 : }
    2359              : 
    2360              : /* Return the vectorization factor that should be used for costing
    2361              :    purposes while vectorizing the loop described by LOOP_VINFO.
    2362              :    Pick a reasonable estimate if the vectorization factor isn't
    2363              :    known at compile time.  */
    2364              : 
    2365              : inline unsigned int
    2366      1275150 : vect_vf_for_cost (loop_vec_info loop_vinfo)
    2367              : {
    2368      1275150 :   return estimated_poly_value (LOOP_VINFO_VECT_FACTOR (loop_vinfo));
    2369              : }
    2370              : 
    2371              : /* Estimate the number of elements in VEC_TYPE for costing purposes.
    2372              :    Pick a reasonable estimate if the exact number isn't known at
    2373              :    compile time.  */
    2374              : 
    2375              : inline unsigned int
    2376        31285 : vect_nunits_for_cost (tree vec_type)
    2377              : {
    2378        31285 :   return estimated_poly_value (TYPE_VECTOR_SUBPARTS (vec_type));
    2379              : }
    2380              : 
    2381              : /* Return the maximum possible vectorization factor for LOOP_VINFO.  */
    2382              : 
    2383              : inline unsigned HOST_WIDE_INT
    2384       105598 : vect_max_vf (loop_vec_info loop_vinfo)
    2385              : {
    2386       105598 :   unsigned HOST_WIDE_INT vf;
    2387       105598 :   if (LOOP_VINFO_VECT_FACTOR (loop_vinfo).is_constant (&vf))
    2388       105598 :     return vf;
    2389              :   return MAX_VECTORIZATION_FACTOR;
    2390              : }
    2391              : 
    2392              : /* Return the size of the value accessed by unvectorized data reference
    2393              :    DR_INFO.  This is only valid once STMT_VINFO_VECTYPE has been calculated
    2394              :    for the associated gimple statement, since that guarantees that DR_INFO
    2395              :    accesses either a scalar or a scalar equivalent.  ("Scalar equivalent"
    2396              :    here includes things like V1SI, which can be vectorized in the same way
    2397              :    as a plain SI.)  */
    2398              : 
    2399              : inline unsigned int
    2400      1950308 : vect_get_scalar_dr_size (dr_vec_info *dr_info)
    2401              : {
    2402      1950308 :   return tree_to_uhwi (TYPE_SIZE_UNIT (TREE_TYPE (DR_REF (dr_info->dr))));
    2403              : }
    2404              : 
    2405              : /* Return true if LOOP_VINFO requires a runtime check for whether the
    2406              :    vector loop is profitable.  */
    2407              : 
    2408              : inline bool
    2409        71005 : vect_apply_runtime_profitability_check_p (loop_vec_info loop_vinfo)
    2410              : {
    2411        71005 :   unsigned int th = LOOP_VINFO_COST_MODEL_THRESHOLD (loop_vinfo);
    2412        37492 :   return (!LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo)
    2413        71005 :           && th >= vect_vf_for_cost (loop_vinfo));
    2414              : }
    2415              : 
    2416              : /* Return true if CODE is a lane-reducing opcode.  */
    2417              : 
    2418              : inline bool
    2419       381193 : lane_reducing_op_p (code_helper code)
    2420              : {
    2421       381193 :   return code == DOT_PROD_EXPR || code == WIDEN_SUM_EXPR || code == SAD_EXPR;
    2422              : }
    2423              : 
    2424              : /* Return true if STMT is a lane-reducing statement.  */
    2425              : 
    2426              : inline bool
    2427       459804 : lane_reducing_stmt_p (gimple *stmt)
    2428              : {
    2429       459804 :   if (auto *assign = dyn_cast <gassign *> (stmt))
    2430       318534 :     return lane_reducing_op_p (gimple_assign_rhs_code (assign));
    2431              :   return false;
    2432              : }
    2433              : 
    2434              : /* Source location + hotness information. */
    2435              : extern dump_user_location_t vect_location;
    2436              : 
    2437              : /* A macro for calling:
    2438              :      dump_begin_scope (MSG, vect_location);
    2439              :    via an RAII object, thus printing "=== MSG ===\n" to the dumpfile etc,
    2440              :    and then calling
    2441              :      dump_end_scope ();
    2442              :    once the object goes out of scope, thus capturing the nesting of
    2443              :    the scopes.
    2444              : 
    2445              :    These scopes affect dump messages within them: dump messages at the
    2446              :    top level implicitly default to MSG_PRIORITY_USER_FACING, whereas those
    2447              :    in a nested scope implicitly default to MSG_PRIORITY_INTERNALS.  */
    2448              : 
    2449              : #define DUMP_VECT_SCOPE(MSG) \
    2450              :   AUTO_DUMP_SCOPE (MSG, vect_location)
    2451              : 
    2452              : /* A sentinel class for ensuring that the "vect_location" global gets
    2453              :    reset at the end of a scope.
    2454              : 
    2455              :    The "vect_location" global is used during dumping and contains a
    2456              :    location_t, which could contain references to a tree block via the
    2457              :    ad-hoc data.  This data is used for tracking inlining information,
    2458              :    but it's not a GC root; it's simply assumed that such locations never
    2459              :    get accessed if the blocks are optimized away.
    2460              : 
    2461              :    Hence we need to ensure that such locations are purged at the end
    2462              :    of any operations using them (e.g. via this class).  */
    2463              : 
    2464              : class auto_purge_vect_location
    2465              : {
    2466              :  public:
    2467              :   ~auto_purge_vect_location ();
    2468              : };
    2469              : 
    2470              : /*-----------------------------------------------------------------*/
    2471              : /* Function prototypes.                                            */
    2472              : /*-----------------------------------------------------------------*/
    2473              : 
    2474              : /* Simple loop peeling and versioning utilities for vectorizer's purposes -
    2475              :    in tree-vect-loop-manip.cc.  */
    2476              : extern void vect_set_loop_condition (class loop *, edge, loop_vec_info,
    2477              :                                      tree, tree, tree, bool);
    2478              : extern bool slpeel_can_duplicate_loop_p (const class loop *, const_edge,
    2479              :                                          const_edge);
    2480              : class loop *slpeel_tree_duplicate_loop_to_edge_cfg (class loop *, edge,
    2481              :                                                     class loop *, edge,
    2482              :                                                     edge, edge *, bool = true,
    2483              :                                                     vec<basic_block> * = NULL,
    2484              :                                                     bool = false, bool = false);
    2485              : class loop *vect_loop_versioning (loop_vec_info, gimple *);
    2486              : extern class loop *vect_do_peeling (loop_vec_info, tree, tree,
    2487              :                                     tree *, tree *, tree *, int, bool, bool,
    2488              :                                     tree *);
    2489              : extern tree vect_get_main_loop_result (loop_vec_info, tree, tree);
    2490              : extern void vect_prepare_for_masked_peels (loop_vec_info);
    2491              : extern dump_user_location_t find_loop_location (class loop *);
    2492              : extern bool vect_can_advance_ivs_p (loop_vec_info);
    2493              : extern void vect_update_inits_of_drs (loop_vec_info, tree, tree_code);
    2494              : extern edge vec_init_loop_exit_info (class loop *);
    2495              : extern void vect_iv_increment_position (edge, gimple_stmt_iterator *, bool *);
    2496              : 
    2497              : /* In tree-vect-stmts.cc.  */
    2498              : extern tree get_related_vectype_for_scalar_type (machine_mode, tree,
    2499              :                                                  poly_uint64 = 0);
    2500              : extern tree get_vectype_for_scalar_type (vec_info *, tree, unsigned int = 0);
    2501              : extern tree get_vectype_for_scalar_type (vec_info *, tree, slp_tree);
    2502              : extern tree get_mask_type_for_scalar_type (vec_info *, tree, unsigned int = 0);
    2503              : extern tree get_mask_type_for_scalar_type (vec_info *, tree, slp_tree);
    2504              : extern tree get_same_sized_vectype (tree, tree);
    2505              : extern bool vect_chooses_same_modes_p (vec_info *, machine_mode);
    2506              : extern bool vect_chooses_same_modes_p (machine_mode, machine_mode);
    2507              : extern bool vect_get_loop_mask_type (loop_vec_info);
    2508              : extern bool vect_is_simple_use (tree, vec_info *, enum vect_def_type *,
    2509              :                                 stmt_vec_info * = NULL, gimple ** = NULL);
    2510              : extern bool vect_is_simple_use (vec_info *, slp_tree,
    2511              :                                 unsigned, tree *, slp_tree *,
    2512              :                                 enum vect_def_type *,
    2513              :                                 tree *, stmt_vec_info * = NULL);
    2514              : extern bool vect_maybe_update_slp_op_vectype (slp_tree, tree);
    2515              : extern tree perm_mask_for_reverse (tree);
    2516              : extern bool supportable_widening_operation (code_helper, tree, tree, bool,
    2517              :                                             code_helper*, code_helper*,
    2518              :                                             int*, vec<tree> *);
    2519              : extern bool supportable_narrowing_operation (code_helper, tree, tree,
    2520              :                                              code_helper *, int *,
    2521              :                                              vec<tree> *);
    2522              : extern bool supportable_indirect_convert_operation (code_helper,
    2523              :                                                     tree, tree,
    2524              :                                                     vec<std::pair<tree, tree_code> > &,
    2525              :                                                     tree = NULL_TREE,
    2526              :                                                     slp_tree = NULL);
    2527              : extern int compare_step_with_zero (vec_info *, stmt_vec_info);
    2528              : 
    2529              : extern unsigned record_stmt_cost (stmt_vector_for_cost *, int,
    2530              :                                   enum vect_cost_for_stmt, stmt_vec_info,
    2531              :                                   tree, int, enum vect_cost_model_location);
    2532              : extern unsigned record_stmt_cost (stmt_vector_for_cost *, int,
    2533              :                                   enum vect_cost_for_stmt, slp_tree,
    2534              :                                   tree, int, enum vect_cost_model_location);
    2535              : extern unsigned record_stmt_cost (stmt_vector_for_cost *, int,
    2536              :                                   enum vect_cost_for_stmt,
    2537              :                                   enum vect_cost_model_location);
    2538              : extern unsigned record_stmt_cost (stmt_vector_for_cost *, int,
    2539              :                                   enum vect_cost_for_stmt, stmt_vec_info,
    2540              :                                   slp_tree, tree, int,
    2541              :                                   enum vect_cost_model_location);
    2542              : 
    2543              : /* Overload of record_stmt_cost with VECTYPE derived from STMT_INFO.  */
    2544              : 
    2545              : inline unsigned
    2546      1771387 : record_stmt_cost (stmt_vector_for_cost *body_cost_vec, int count,
    2547              :                   enum vect_cost_for_stmt kind, stmt_vec_info stmt_info,
    2548              :                   int misalign, enum vect_cost_model_location where)
    2549              : {
    2550      1770858 :   return record_stmt_cost (body_cost_vec, count, kind, stmt_info,
    2551      1770858 :                            STMT_VINFO_VECTYPE (stmt_info), misalign, where);
    2552              : }
    2553              : 
    2554              : /* Overload of record_stmt_cost with VECTYPE derived from SLP node.  */
    2555              : 
    2556              : inline unsigned
    2557      1638827 : record_stmt_cost (stmt_vector_for_cost *body_cost_vec, int count,
    2558              :                   enum vect_cost_for_stmt kind, slp_tree node,
    2559              :                   int misalign, enum vect_cost_model_location where)
    2560              : {
    2561      1436630 :   return record_stmt_cost (body_cost_vec, count, kind, node,
    2562        68830 :                            SLP_TREE_VECTYPE (node), misalign, where);
    2563              : }
    2564              : 
    2565              : extern void vect_finish_replace_stmt (vec_info *, stmt_vec_info, gimple *);
    2566              : extern void vect_finish_stmt_generation (vec_info *, stmt_vec_info, gimple *,
    2567              :                                          gimple_stmt_iterator *);
    2568              : extern opt_result vect_mark_stmts_to_be_vectorized (loop_vec_info, bool *);
    2569              : extern tree vect_get_store_rhs (stmt_vec_info);
    2570              : void vect_get_vec_defs (vec_info *, slp_tree,
    2571              :                         tree, vec<tree> *,
    2572              :                         tree = NULL, vec<tree> * = NULL,
    2573              :                         tree = NULL, vec<tree> * = NULL,
    2574              :                         tree = NULL, vec<tree> * = NULL);
    2575              : extern tree vect_init_vector (vec_info *, stmt_vec_info, tree, tree,
    2576              :                               gimple_stmt_iterator *);
    2577              : extern tree vect_get_slp_vect_def (slp_tree, unsigned);
    2578              : extern bool vect_transform_stmt (vec_info *, stmt_vec_info,
    2579              :                                  gimple_stmt_iterator *,
    2580              :                                  slp_tree, slp_instance);
    2581              : extern void vect_remove_stores (vec_info *, stmt_vec_info);
    2582              : extern bool vect_nop_conversion_p (stmt_vec_info);
    2583              : extern opt_result vect_analyze_stmt (vec_info *, slp_tree,
    2584              :                                      slp_instance, stmt_vector_for_cost *);
    2585              : extern void vect_get_load_cost (vec_info *, stmt_vec_info, slp_tree, int,
    2586              :                                 dr_alignment_support, int, bool,
    2587              :                                 unsigned int *, unsigned int *,
    2588              :                                 stmt_vector_for_cost *,
    2589              :                                 stmt_vector_for_cost *, bool);
    2590              : extern void vect_get_store_cost (vec_info *, stmt_vec_info, slp_tree, int,
    2591              :                                  dr_alignment_support, int,
    2592              :                                  unsigned int *, stmt_vector_for_cost *);
    2593              : extern bool vect_supportable_shift (vec_info *, enum tree_code, tree);
    2594              : extern tree vect_gen_perm_mask_any (tree, const vec_perm_indices &);
    2595              : extern tree vect_gen_perm_mask_checked (tree, const vec_perm_indices &);
    2596              : extern void optimize_mask_stores (class loop*);
    2597              : extern tree vect_gen_while (gimple_seq *, tree, tree, tree,
    2598              :                             const char * = nullptr);
    2599              : extern tree vect_gen_while_not (gimple_seq *, tree, tree, tree);
    2600              : extern opt_result vect_get_vector_types_for_stmt (vec_info *,
    2601              :                                                   stmt_vec_info, tree *,
    2602              :                                                   tree *, unsigned int = 0);
    2603              : extern opt_tree vect_get_mask_type_for_stmt (stmt_vec_info, unsigned int = 0);
    2604              : 
    2605              : /* In tree-if-conv.cc.  */
    2606              : extern bool ref_within_array_bound (gimple *, tree);
    2607              : 
    2608              : /* In tree-vect-data-refs.cc.  */
    2609              : extern bool vect_can_force_dr_alignment_p (const_tree, poly_uint64);
    2610              : extern enum dr_alignment_support vect_supportable_dr_alignment
    2611              :                                    (vec_info *, dr_vec_info *, tree, int,
    2612              :                                     bool = false);
    2613              : extern tree vect_get_smallest_scalar_type (stmt_vec_info, tree);
    2614              : extern opt_result vect_analyze_data_ref_dependences (loop_vec_info, unsigned int *);
    2615              : extern bool vect_slp_analyze_instance_dependence (vec_info *, slp_instance);
    2616              : extern opt_result vect_enhance_data_refs_alignment (loop_vec_info);
    2617              : extern void vect_analyze_data_refs_alignment (loop_vec_info);
    2618              : extern bool vect_slp_analyze_instance_alignment (vec_info *, slp_instance);
    2619              : extern opt_result vect_analyze_data_ref_accesses (vec_info *, vec<int> *);
    2620              : extern opt_result vect_prune_runtime_alias_test_list (loop_vec_info);
    2621              : extern bool vect_gather_scatter_fn_p (vec_info *, bool, bool, tree, tree,
    2622              :                                       tree, int, int *, internal_fn *, tree *,
    2623              :                                       tree *, vec<int> * = nullptr);
    2624              : extern bool vect_check_gather_scatter (stmt_vec_info, tree,
    2625              :                                        loop_vec_info, gather_scatter_info *,
    2626              :                                        vec<int> * = nullptr);
    2627              : extern void vect_describe_gather_scatter_call (stmt_vec_info,
    2628              :                                                gather_scatter_info *);
    2629              : extern opt_result vect_find_stmt_data_reference (loop_p, gimple *,
    2630              :                                                  vec<data_reference_p> *,
    2631              :                                                  vec<int> *, int);
    2632              : extern opt_result vect_analyze_data_refs (vec_info *, bool *);
    2633              : extern void vect_record_base_alignments (vec_info *);
    2634              : extern tree vect_create_data_ref_ptr (vec_info *,
    2635              :                                       stmt_vec_info, tree, class loop *, tree,
    2636              :                                       tree *, gimple_stmt_iterator *,
    2637              :                                       gimple **, bool,
    2638              :                                       tree = NULL_TREE);
    2639              : extern tree bump_vector_ptr (vec_info *, tree, gimple *, gimple_stmt_iterator *,
    2640              :                              stmt_vec_info, tree);
    2641              : extern void vect_copy_ref_info (tree, tree);
    2642              : extern tree vect_create_destination_var (tree, tree);
    2643              : extern bool vect_grouped_store_supported (tree, unsigned HOST_WIDE_INT);
    2644              : extern internal_fn vect_store_lanes_supported (tree, unsigned HOST_WIDE_INT, bool);
    2645              : extern bool vect_grouped_load_supported (tree, bool, unsigned HOST_WIDE_INT);
    2646              : extern internal_fn vect_load_lanes_supported (tree, unsigned HOST_WIDE_INT,
    2647              :                                               bool, vec<int> * = nullptr);
    2648              : extern tree vect_setup_realignment (vec_info *,
    2649              :                                     stmt_vec_info, tree, gimple_stmt_iterator *,
    2650              :                                     tree *, enum dr_alignment_support, tree,
    2651              :                                     class loop **);
    2652              : extern tree vect_get_new_vect_var (tree, enum vect_var_kind, const char *);
    2653              : extern tree vect_get_new_ssa_name (tree, enum vect_var_kind,
    2654              :                                    const char * = NULL);
    2655              : extern tree vect_create_addr_base_for_vector_ref (vec_info *,
    2656              :                                                   stmt_vec_info, gimple_seq *,
    2657              :                                                   tree);
    2658              : 
    2659              : /* In tree-vect-loop.cc.  */
    2660              : extern tree neutral_op_for_reduction (tree, code_helper, tree, bool = true);
    2661              : extern widest_int vect_iv_limit_for_partial_vectors (loop_vec_info loop_vinfo);
    2662              : bool vect_rgroup_iv_might_wrap_p (loop_vec_info, rgroup_controls *);
    2663              : /* Used in gimple-loop-interchange.c and tree-parloops.cc.  */
    2664              : extern bool check_reduction_path (dump_user_location_t, loop_p, gphi *, tree,
    2665              :                                   enum tree_code);
    2666              : extern bool needs_fold_left_reduction_p (tree, code_helper);
    2667              : /* Drive for loop analysis stage.  */
    2668              : extern opt_loop_vec_info vect_analyze_loop (class loop *, gimple *,
    2669              :                                             vec_info_shared *);
    2670              : extern tree vect_build_loop_niters (loop_vec_info, bool * = NULL);
    2671              : extern void vect_gen_vector_loop_niters (loop_vec_info, tree, tree *,
    2672              :                                          tree *, bool);
    2673              : extern tree vect_halve_mask_nunits (tree, machine_mode);
    2674              : extern tree vect_double_mask_nunits (tree, machine_mode);
    2675              : extern void vect_record_loop_mask (loop_vec_info, vec_loop_masks *,
    2676              :                                    unsigned int, tree, tree);
    2677              : extern tree vect_get_loop_mask (loop_vec_info, gimple_stmt_iterator *,
    2678              :                                 vec_loop_masks *,
    2679              :                                 unsigned int, tree, unsigned int);
    2680              : extern void vect_record_loop_len (loop_vec_info, vec_loop_lens *, unsigned int,
    2681              :                                   tree, unsigned int);
    2682              : extern tree vect_get_loop_len (loop_vec_info, gimple_stmt_iterator *,
    2683              :                                vec_loop_lens *, unsigned int, tree,
    2684              :                                unsigned int, unsigned int, bool);
    2685              : extern tree vect_gen_loop_len_mask (loop_vec_info, gimple_stmt_iterator *,
    2686              :                                     gimple_stmt_iterator *, vec_loop_lens *,
    2687              :                                     unsigned int, tree, tree, unsigned int,
    2688              :                                     unsigned int);
    2689              : extern gimple_seq vect_gen_len (tree, tree, tree, tree);
    2690              : extern vect_reduc_info info_for_reduction (loop_vec_info, slp_tree);
    2691              : extern bool reduction_fn_for_scalar_code (code_helper, internal_fn *);
    2692              : extern unsigned vect_min_prec_for_max_niters (loop_vec_info, unsigned int);
    2693              : /* Drive for loop transformation stage.  */
    2694              : extern class loop *vect_transform_loop (loop_vec_info, gimple *);
    2695       907048 : struct vect_loop_form_info
    2696              : {
    2697              :   tree number_of_iterations;
    2698              :   tree number_of_iterationsm1;
    2699              :   tree assumptions;
    2700              :   auto_vec<gcond *> conds;
    2701              :   gcond *inner_loop_cond;
    2702              :   edge loop_exit;
    2703              : };
    2704              : extern opt_result vect_analyze_loop_form (class loop *, gimple *,
    2705              :                                           vect_loop_form_info *);
    2706              : extern loop_vec_info vect_create_loop_vinfo (class loop *, vec_info_shared *,
    2707              :                                              const vect_loop_form_info *,
    2708              :                                              loop_vec_info = nullptr);
    2709              : extern bool vectorizable_live_operation (vec_info *, stmt_vec_info,
    2710              :                                          slp_tree, slp_instance, int,
    2711              :                                          bool, stmt_vector_for_cost *);
    2712              : extern bool vectorizable_lane_reducing (loop_vec_info, stmt_vec_info,
    2713              :                                         slp_tree, stmt_vector_for_cost *);
    2714              : extern bool vectorizable_reduction (loop_vec_info, stmt_vec_info,
    2715              :                                     slp_tree, slp_instance,
    2716              :                                     stmt_vector_for_cost *);
    2717              : extern bool vectorizable_induction (loop_vec_info, stmt_vec_info,
    2718              :                                     slp_tree, stmt_vector_for_cost *);
    2719              : extern bool vect_transform_reduction (loop_vec_info, stmt_vec_info,
    2720              :                                       gimple_stmt_iterator *,
    2721              :                                       slp_tree);
    2722              : extern bool vect_transform_cycle_phi (loop_vec_info, stmt_vec_info,
    2723              :                                       slp_tree, slp_instance);
    2724              : extern bool vectorizable_lc_phi (loop_vec_info, stmt_vec_info, slp_tree);
    2725              : extern bool vect_transform_lc_phi (loop_vec_info, stmt_vec_info, slp_tree);
    2726              : extern bool vectorizable_phi (bb_vec_info, stmt_vec_info, slp_tree,
    2727              :                               stmt_vector_for_cost *);
    2728              : extern bool vectorizable_recurr (loop_vec_info, stmt_vec_info,
    2729              :                                   slp_tree, stmt_vector_for_cost *);
    2730              : extern bool vectorizable_early_exit (loop_vec_info, stmt_vec_info,
    2731              :                                      gimple_stmt_iterator *,
    2732              :                                      slp_tree, stmt_vector_for_cost *);
    2733              : extern bool vect_emulated_vector_p (tree);
    2734              : extern bool vect_can_vectorize_without_simd_p (tree_code);
    2735              : extern bool vect_can_vectorize_without_simd_p (code_helper);
    2736              : extern int vect_get_known_peeling_cost (loop_vec_info, int);
    2737              : extern tree cse_and_gimplify_to_preheader (loop_vec_info, tree);
    2738              : 
    2739              : /* Nonlinear induction.  */
    2740              : extern tree vect_peel_nonlinear_iv_init (gimple_seq*, tree, tree,
    2741              :                                          tree, enum vect_induction_op_type,
    2742              :                                          bool);
    2743              : 
    2744              : /* In tree-vect-slp.cc.  */
    2745              : extern void vect_slp_init (void);
    2746              : extern void vect_slp_fini (void);
    2747              : extern void vect_free_slp_instance (slp_instance);
    2748              : extern bool vect_transform_slp_perm_load (vec_info *, slp_tree, const vec<tree> &,
    2749              :                                           gimple_stmt_iterator *, poly_uint64,
    2750              :                                           bool, unsigned *,
    2751              :                                           unsigned * = nullptr, bool = false);
    2752              : extern bool vectorizable_slp_permutation (vec_info *, gimple_stmt_iterator *,
    2753              :                                           slp_tree, stmt_vector_for_cost *);
    2754              : extern bool vect_slp_analyze_operations (vec_info *);
    2755              : extern void vect_schedule_slp (vec_info *, const vec<slp_instance> &);
    2756              : extern opt_result vect_analyze_slp (vec_info *, unsigned, bool);
    2757              : extern bool vect_make_slp_decision (loop_vec_info);
    2758              : extern bool vect_detect_hybrid_slp (loop_vec_info);
    2759              : extern void vect_optimize_slp (vec_info *);
    2760              : extern void vect_gather_slp_loads (vec_info *);
    2761              : extern tree vect_get_slp_scalar_def (slp_tree, unsigned);
    2762              : extern void vect_get_slp_defs (slp_tree, vec<tree> *);
    2763              : extern void vect_get_slp_defs (vec_info *, slp_tree, vec<vec<tree> > *,
    2764              :                                unsigned n = -1U);
    2765              : extern bool vect_slp_if_converted_bb (basic_block bb, loop_p orig_loop);
    2766              : extern bool vect_slp_function (function *);
    2767              : extern stmt_vec_info vect_find_last_scalar_stmt_in_slp (slp_tree);
    2768              : extern stmt_vec_info vect_find_first_scalar_stmt_in_slp (slp_tree);
    2769              : extern bool is_simple_and_all_uses_invariant (stmt_vec_info, loop_vec_info);
    2770              : extern bool can_duplicate_and_interleave_p (vec_info *, unsigned int, tree,
    2771              :                                             unsigned int * = NULL,
    2772              :                                             tree * = NULL, tree * = NULL);
    2773              : extern void duplicate_and_interleave (vec_info *, gimple_seq *, tree,
    2774              :                                       const vec<tree> &, unsigned int, vec<tree> &);
    2775              : extern int vect_get_place_in_interleaving_chain (stmt_vec_info, stmt_vec_info);
    2776              : extern slp_tree vect_create_new_slp_node (unsigned, tree_code);
    2777              : extern void vect_free_slp_tree (slp_tree);
    2778              : extern bool compatible_calls_p (gcall *, gcall *, bool);
    2779              : extern int vect_slp_child_index_for_operand (const stmt_vec_info, int op);
    2780              : 
    2781              : extern tree prepare_vec_mask (loop_vec_info, tree, tree, tree,
    2782              :                               gimple_stmt_iterator *);
    2783              : extern tree vect_get_mask_load_else (int, tree);
    2784              : extern bool vect_load_perm_consecutive_p (slp_tree, unsigned = UINT_MAX);
    2785              : 
    2786              : /* In tree-vect-patterns.cc.  */
    2787              : extern void
    2788              : vect_mark_pattern_stmts (vec_info *, stmt_vec_info, gimple *, tree);
    2789              : extern bool vect_get_range_info (tree, wide_int*, wide_int*);
    2790              : 
    2791              : /* Pattern recognition functions.
    2792              :    Additional pattern recognition functions can (and will) be added
    2793              :    in the future.  */
    2794              : void vect_pattern_recog (vec_info *);
    2795              : 
    2796              : /* In tree-vectorizer.cc.  */
    2797              : unsigned vectorize_loops (void);
    2798              : void vect_free_loop_info_assumptions (class loop *);
    2799              : gimple *vect_loop_vectorized_call (class loop *, gcond **cond = NULL);
    2800              : bool vect_stmt_dominates_stmt_p (gimple *, gimple *);
    2801              : 
    2802              : /* SLP Pattern matcher types, tree-vect-slp-patterns.cc.  */
    2803              : 
    2804              : /* Forward declaration of possible two operands operation that can be matched
    2805              :    by the complex numbers pattern matchers.  */
    2806              : enum _complex_operation : unsigned;
    2807              : 
    2808              : /* All possible load permute values that could result from the partial data-flow
    2809              :    analysis.  */
    2810              : typedef enum _complex_perm_kinds {
    2811              :    PERM_UNKNOWN,
    2812              :    PERM_EVENODD,
    2813              :    PERM_ODDEVEN,
    2814              :    PERM_ODDODD,
    2815              :    PERM_EVENEVEN,
    2816              :    /* Can be combined with any other PERM values.  */
    2817              :    PERM_TOP
    2818              : } complex_perm_kinds_t;
    2819              : 
    2820              : /* Cache from nodes to the load permutation they represent.  */
    2821              : typedef hash_map <slp_tree, complex_perm_kinds_t>
    2822              :   slp_tree_to_load_perm_map_t;
    2823              : 
    2824              : /* Cache from nodes pair to being compatible or not.  */
    2825              : typedef pair_hash <nofree_ptr_hash <_slp_tree>,
    2826              :                    nofree_ptr_hash <_slp_tree>> slp_node_hash;
    2827              : typedef hash_map <slp_node_hash, bool> slp_compat_nodes_map_t;
    2828              : 
    2829              : 
    2830              : /* Vector pattern matcher base class.  All SLP pattern matchers must inherit
    2831              :    from this type.  */
    2832              : 
    2833              : class vect_pattern
    2834              : {
    2835              :   protected:
    2836              :     /* The number of arguments that the IFN requires.  */
    2837              :     unsigned m_num_args;
    2838              : 
    2839              :     /* The internal function that will be used when a pattern is created.  */
    2840              :     internal_fn m_ifn;
    2841              : 
    2842              :     /* The current node being inspected.  */
    2843              :     slp_tree *m_node;
    2844              : 
    2845              :     /* The list of operands to be the children for the node produced when the
    2846              :        internal function is created.  */
    2847              :     vec<slp_tree> m_ops;
    2848              : 
    2849              :     /* Default constructor where NODE is the root of the tree to inspect.  */
    2850         1171 :     vect_pattern (slp_tree *node, vec<slp_tree> *m_ops, internal_fn ifn)
    2851         1171 :     {
    2852         1171 :       this->m_ifn = ifn;
    2853         1171 :       this->m_node = node;
    2854         1171 :       this->m_ops.create (0);
    2855         1171 :       if (m_ops)
    2856           32 :         this->m_ops.safe_splice (*m_ops);
    2857              :     }
    2858              : 
    2859              :   public:
    2860              : 
    2861              :     /* Create a new instance of the pattern matcher class of the given type.  */
    2862              :     static vect_pattern* recognize (slp_tree_to_load_perm_map_t *,
    2863              :                                     slp_compat_nodes_map_t *, slp_tree *);
    2864              : 
    2865              :     /* Build the pattern from the data collected so far.  */
    2866              :     virtual void build (vec_info *) = 0;
    2867              : 
    2868              :     /* Default destructor.  */
    2869              :     virtual ~vect_pattern ()
    2870              :     {
    2871              :         this->m_ops.release ();
    2872              :     }
    2873              : };
    2874              : 
    2875              : /* Function pointer to create a new pattern matcher from a generic type.  */
    2876              : typedef vect_pattern* (*vect_pattern_decl_t) (slp_tree_to_load_perm_map_t *,
    2877              :                                               slp_compat_nodes_map_t *,
    2878              :                                               slp_tree *);
    2879              : 
    2880              : /* List of supported pattern matchers.  */
    2881              : extern vect_pattern_decl_t slp_patterns[];
    2882              : 
    2883              : /* Number of supported pattern matchers.  */
    2884              : extern size_t num__slp_patterns;
    2885              : 
    2886              : /* ----------------------------------------------------------------------
    2887              :    Target support routines
    2888              :    -----------------------------------------------------------------------
    2889              :    The following routines are provided to simplify costing decisions in
    2890              :    target code.  Please add more as needed.  */
    2891              : 
    2892              : /* Return true if an operaton of kind KIND for STMT_INFO represents
    2893              :    the extraction of an element from a vector in preparation for
    2894              :    storing the element to memory.  */
    2895              : inline bool
    2896              : vect_is_store_elt_extraction (vect_cost_for_stmt kind, stmt_vec_info stmt_info)
    2897              : {
    2898              :   return (kind == vec_to_scalar
    2899              :           && STMT_VINFO_DATA_REF (stmt_info)
    2900              :           && DR_IS_WRITE (STMT_VINFO_DATA_REF (stmt_info)));
    2901              : }
    2902              : 
    2903              : /* Return true if STMT_INFO represents part of a reduction.  */
    2904              : inline bool
    2905     47129752 : vect_is_reduction (stmt_vec_info stmt_info)
    2906              : {
    2907     47129752 :   return STMT_VINFO_REDUC_IDX (stmt_info) != -1;
    2908              : }
    2909              : 
    2910              : /* Return true if SLP_NODE represents part of a reduction.  */
    2911              : inline bool
    2912       259740 : vect_is_reduction (slp_tree slp_node)
    2913              : {
    2914       259740 :   return SLP_TREE_REDUC_IDX (slp_node) != -1;
    2915              : }
    2916              : 
    2917              : /* If STMT_INFO describes a reduction, return the vect_reduction_type
    2918              :    of the reduction it describes, otherwise return -1.  */
    2919              : inline int
    2920           40 : vect_reduc_type (vec_info *vinfo, slp_tree node)
    2921              : {
    2922           40 :   if (loop_vec_info loop_vinfo = dyn_cast<loop_vec_info> (vinfo))
    2923              :     {
    2924           40 :       vect_reduc_info reduc_info = info_for_reduction (loop_vinfo, node);
    2925           40 :       if (reduc_info)
    2926           40 :         return int (VECT_REDUC_INFO_TYPE (reduc_info));
    2927              :     }
    2928              :   return -1;
    2929              : }
    2930              : 
    2931              : /* If STMT_INFO is a COND_EXPR that includes an embedded comparison, return the
    2932              :    scalar type of the values being compared.  Return null otherwise.  */
    2933              : inline tree
    2934              : vect_embedded_comparison_type (stmt_vec_info stmt_info)
    2935              : {
    2936              :   if (auto *assign = dyn_cast<gassign *> (stmt_info->stmt))
    2937              :     if (gimple_assign_rhs_code (assign) == COND_EXPR)
    2938              :       {
    2939              :         tree cond = gimple_assign_rhs1 (assign);
    2940              :         if (COMPARISON_CLASS_P (cond))
    2941              :           return TREE_TYPE (TREE_OPERAND (cond, 0));
    2942              :       }
    2943              :   return NULL_TREE;
    2944              : }
    2945              : 
    2946              : /* If STMT_INFO is a comparison or contains an embedded comparison, return the
    2947              :    scalar type of the values being compared.  Return null otherwise.  */
    2948              : inline tree
    2949              : vect_comparison_type (stmt_vec_info stmt_info)
    2950              : {
    2951              :   if (auto *assign = dyn_cast<gassign *> (stmt_info->stmt))
    2952              :     if (TREE_CODE_CLASS (gimple_assign_rhs_code (assign)) == tcc_comparison)
    2953              :       return TREE_TYPE (gimple_assign_rhs1 (assign));
    2954              :   return vect_embedded_comparison_type (stmt_info);
    2955              : }
    2956              : 
    2957              : /* Return true if STMT_INFO extends the result of a load.  */
    2958              : inline bool
    2959              : vect_is_extending_load (class vec_info *vinfo, stmt_vec_info stmt_info)
    2960              : {
    2961              :   /* Although this is quite large for an inline function, this part
    2962              :      at least should be inline.  */
    2963              :   gassign *assign = dyn_cast <gassign *> (stmt_info->stmt);
    2964              :   if (!assign || !CONVERT_EXPR_CODE_P (gimple_assign_rhs_code (assign)))
    2965              :     return false;
    2966              : 
    2967              :   tree rhs = gimple_assign_rhs1 (stmt_info->stmt);
    2968              :   tree lhs_type = TREE_TYPE (gimple_assign_lhs (assign));
    2969              :   tree rhs_type = TREE_TYPE (rhs);
    2970              :   if (!INTEGRAL_TYPE_P (lhs_type)
    2971              :       || !INTEGRAL_TYPE_P (rhs_type)
    2972              :       || TYPE_PRECISION (lhs_type) <= TYPE_PRECISION (rhs_type))
    2973              :     return false;
    2974              : 
    2975              :   stmt_vec_info def_stmt_info = vinfo->lookup_def (rhs);
    2976              :   return (def_stmt_info
    2977              :           && STMT_VINFO_DATA_REF (def_stmt_info)
    2978              :           && DR_IS_READ (STMT_VINFO_DATA_REF (def_stmt_info)));
    2979              : }
    2980              : 
    2981              : /* Return true if STMT_INFO is an integer truncation.  */
    2982              : inline bool
    2983              : vect_is_integer_truncation (stmt_vec_info stmt_info)
    2984              : {
    2985              :   gassign *assign = dyn_cast <gassign *> (stmt_info->stmt);
    2986              :   if (!assign || !CONVERT_EXPR_CODE_P (gimple_assign_rhs_code (assign)))
    2987              :     return false;
    2988              : 
    2989              :   tree lhs_type = TREE_TYPE (gimple_assign_lhs (assign));
    2990              :   tree rhs_type = TREE_TYPE (gimple_assign_rhs1 (assign));
    2991              :   return (INTEGRAL_TYPE_P (lhs_type)
    2992              :           && INTEGRAL_TYPE_P (rhs_type)
    2993              :           && TYPE_PRECISION (lhs_type) < TYPE_PRECISION (rhs_type));
    2994              : }
    2995              : 
    2996              : /* Build a GIMPLE_ASSIGN or GIMPLE_CALL with the tree_code,
    2997              :    or internal_fn contained in ch, respectively.  */
    2998              : gimple * vect_gimple_build (tree, code_helper, tree, tree = NULL_TREE);
    2999              : #endif  /* GCC_TREE_VECTORIZER_H  */
        

Generated by: LCOV version 2.4-beta

LCOV profile is generated on x86_64 machine using following configure options: configure --disable-bootstrap --enable-coverage=opt --enable-languages=c,c++,fortran,go,jit,lto,rust,m2 --enable-host-shared. GCC test suite is run with the built compiler.