LCOV - code coverage report
Current view: top level - gcc - gimple-range-cache.cc (source / functions) Coverage Total Hit
Test: gcc.info Lines: 84.1 % 816 686
Test Date: 2026-02-28 14:20:25 Functions: 90.8 % 76 69
Legend: Lines:     hit not hit

            Line data    Source code
       1              : /* Gimple ranger SSA cache implementation.
       2              :    Copyright (C) 2017-2026 Free Software Foundation, Inc.
       3              :    Contributed by Andrew MacLeod <amacleod@redhat.com>.
       4              : 
       5              : This file is part of GCC.
       6              : 
       7              : GCC is free software; you can redistribute it and/or modify
       8              : it under the terms of the GNU General Public License as published by
       9              : the Free Software Foundation; either version 3, or (at your option)
      10              : any later version.
      11              : 
      12              : GCC is distributed in the hope that it will be useful,
      13              : but WITHOUT ANY WARRANTY; without even the implied warranty of
      14              : MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
      15              : GNU General Public License for more details.
      16              : 
      17              : You should have received a copy of the GNU General Public License
      18              : along with GCC; see the file COPYING3.  If not see
      19              : <http://www.gnu.org/licenses/>.  */
      20              : 
      21              : #include "config.h"
      22              : #include "system.h"
      23              : #include "coretypes.h"
      24              : #include "backend.h"
      25              : #include "insn-codes.h"
      26              : #include "tree.h"
      27              : #include "gimple.h"
      28              : #include "ssa.h"
      29              : #include "gimple-pretty-print.h"
      30              : #include "gimple-range.h"
      31              : #include "value-range-storage.h"
      32              : #include "tree-cfg.h"
      33              : #include "target.h"
      34              : #include "attribs.h"
      35              : #include "gimple-iterator.h"
      36              : #include "gimple-walk.h"
      37              : #include "cfganal.h"
      38              : 
      39              : #define DEBUG_RANGE_CACHE (dump_file                                    \
      40              :                            && (param_ranger_debug & RANGER_DEBUG_CACHE))
      41              : 
      42              : // This class represents the API into a cache of ranges for an SSA_NAME.
      43              : // Routines must be implemented to set, get, and query if a value is set.
      44              : 
      45              : class ssa_block_ranges
      46              : {
      47              : public:
      48     28431593 :   ssa_block_ranges (tree t) : m_type (t) { }
      49              :   virtual bool set_bb_range (const_basic_block bb, const vrange &r) = 0;
      50              :   virtual bool get_bb_range (vrange &r, const_basic_block bb) = 0;
      51              :   virtual bool bb_range_p (const_basic_block bb) = 0;
      52              : 
      53              :   void dump(FILE *f);
      54              : private:
      55              :   tree m_type;
      56              : };
      57              : 
      58              : // Print the list of known ranges for file F in a nice format.
      59              : 
      60              : void
      61            0 : ssa_block_ranges::dump (FILE *f)
      62              : {
      63            0 :   basic_block bb;
      64            0 :   value_range r (m_type);
      65              : 
      66            0 :   FOR_EACH_BB_FN (bb, cfun)
      67            0 :     if (get_bb_range (r, bb))
      68              :       {
      69            0 :         fprintf (f, "BB%d  -> ", bb->index);
      70            0 :         r.dump (f);
      71            0 :         fprintf (f, "\n");
      72              :       }
      73            0 : }
      74              : 
      75              : // This class implements the range cache as a linear vector, indexed by BB.
      76              : // It caches a varying and undefined range which are used instead of
      77              : // allocating new ones each time.
      78              : 
      79              : class sbr_vector : public ssa_block_ranges
      80              : {
      81              : public:
      82              :   sbr_vector (tree t, vrange_allocator *allocator, bool zero_p = true);
      83              : 
      84              :   virtual bool set_bb_range (const_basic_block bb, const vrange &r) override;
      85              :   virtual bool get_bb_range (vrange &r, const_basic_block bb) override;
      86              :   virtual bool bb_range_p (const_basic_block bb) override;
      87              : protected:
      88              :   vrange_storage **m_tab;       // Non growing vector.
      89              :   int m_tab_size;
      90              :   vrange_storage *m_varying;
      91              :   vrange_storage *m_undefined;
      92              :   tree m_type;
      93              :   vrange_allocator *m_range_allocator;
      94              :   bool m_zero_p;
      95              :   void grow ();
      96              : };
      97              : 
      98              : 
      99              : // Initialize a block cache for an ssa_name of type T.
     100              : 
     101     28345346 : sbr_vector::sbr_vector (tree t, vrange_allocator *allocator, bool zero_p)
     102     28345346 :   : ssa_block_ranges (t)
     103              : {
     104     28345346 :   gcc_checking_assert (TYPE_P (t));
     105     28345346 :   m_type = t;
     106     28345346 :   m_zero_p = zero_p;
     107     28345346 :   m_range_allocator = allocator;
     108     28345346 :   m_tab_size = last_basic_block_for_fn (cfun) + 1;
     109     56690692 :   m_tab = static_cast <vrange_storage **>
     110     28345346 :     (allocator->alloc (m_tab_size * sizeof (vrange_storage *)));
     111     28345346 :   if (zero_p)
     112     24875341 :     memset (m_tab, 0, m_tab_size * sizeof (vrange *));
     113              : 
     114              :   // Create the cached type range.
     115     28345346 :   m_varying = m_range_allocator->clone_varying (t);
     116     28345346 :   m_undefined = m_range_allocator->clone_undefined (t);
     117     28345346 : }
     118              : 
     119              : // Grow the vector when the CFG has increased in size.
     120              : 
     121              : void
     122        10129 : sbr_vector::grow ()
     123              : {
     124        10129 :   int curr_bb_size = last_basic_block_for_fn (cfun);
     125        10129 :   gcc_checking_assert (curr_bb_size > m_tab_size);
     126              : 
     127              :   // Increase the max of a)128, b)needed increase * 2, c)10% of current_size.
     128        10129 :   int inc = MAX ((curr_bb_size - m_tab_size) * 2, 128);
     129        10129 :   inc = MAX (inc, curr_bb_size / 10);
     130        10129 :   int new_size = inc + curr_bb_size;
     131              : 
     132              :   // Allocate new memory, copy the old vector and clear the new space.
     133        10129 :   vrange_storage **t = static_cast <vrange_storage **>
     134        10129 :     (m_range_allocator->alloc (new_size * sizeof (vrange_storage *)));
     135        10129 :   memcpy (t, m_tab, m_tab_size * sizeof (vrange_storage *));
     136        10129 :   if (m_zero_p)
     137         7898 :     memset (t + m_tab_size, 0, (new_size - m_tab_size) * sizeof (vrange_storage *));
     138              : 
     139        10129 :   m_tab = t;
     140        10129 :   m_tab_size = new_size;
     141        10129 : }
     142              : 
     143              : // Set the range for block BB to be R.
     144              : 
     145              : bool
     146     74059595 : sbr_vector::set_bb_range (const_basic_block bb, const vrange &r)
     147              : {
     148     74059595 :   vrange_storage *m;
     149     74059595 :   if (bb->index >= m_tab_size)
     150        10129 :     grow ();
     151     74059595 :   if (r.varying_p ())
     152     23213039 :     m = m_varying;
     153     50846556 :   else if (r.undefined_p ())
     154      5453788 :     m = m_undefined;
     155              :   else
     156     45392768 :     m = m_range_allocator->clone (r);
     157     74059595 :   m_tab[bb->index] = m;
     158     74059595 :   return true;
     159              : }
     160              : 
     161              : // Return the range associated with block BB in R.  Return false if
     162              : // there is no range.
     163              : 
     164              : bool
     165    318782921 : sbr_vector::get_bb_range (vrange &r, const_basic_block bb)
     166              : {
     167    318782921 :   if (bb->index >= m_tab_size)
     168              :     return false;
     169    318775197 :   vrange_storage *m = m_tab[bb->index];
     170    318775197 :   if (m)
     171              :     {
     172    241278068 :       m->get_vrange (r, m_type);
     173    241278068 :       return true;
     174              :     }
     175              :   return false;
     176              : }
     177              : 
     178              : // Return true if a range is present.
     179              : 
     180              : bool
     181    236945270 : sbr_vector::bb_range_p (const_basic_block bb)
     182              : {
     183    236945270 :   if (bb->index < m_tab_size)
     184    236933996 :     return m_tab[bb->index] != NULL;
     185              :   return false;
     186              : }
     187              : 
     188              : // Like an sbr_vector, except it uses a bitmap to manage whetehr  vale is set
     189              : // or not rather than cleared memory.
     190              : 
     191              : class sbr_lazy_vector : public sbr_vector
     192              : {
     193              : public:
     194              :   sbr_lazy_vector (tree t, vrange_allocator *allocator, bitmap_obstack *bm);
     195              : 
     196              :   virtual bool set_bb_range (const_basic_block bb, const vrange &r) override;
     197              :   virtual bool get_bb_range (vrange &r, const_basic_block bb) override;
     198              :   virtual bool bb_range_p (const_basic_block bb) override;
     199              : protected:
     200              :   bitmap m_has_value;
     201              : };
     202              : 
     203      3470005 : sbr_lazy_vector::sbr_lazy_vector (tree t, vrange_allocator *allocator,
     204      3470005 :                                   bitmap_obstack *bm)
     205      3470005 :   : sbr_vector (t, allocator, false)
     206              : {
     207      3470005 :   m_has_value = BITMAP_ALLOC (bm);
     208      3470005 : }
     209              : 
     210              : bool
     211     11674047 : sbr_lazy_vector::set_bb_range (const_basic_block bb, const vrange &r)
     212              : {
     213     11674047 :   sbr_vector::set_bb_range (bb, r);
     214     11674047 :   bitmap_set_bit (m_has_value, bb->index);
     215     11674047 :   return true;
     216              : }
     217              : 
     218              : bool
     219    239651740 : sbr_lazy_vector::get_bb_range (vrange &r, const_basic_block bb)
     220              : {
     221    239651740 :   if (bitmap_bit_p (m_has_value, bb->index))
     222     39630835 :     return sbr_vector::get_bb_range (r, bb);
     223              :   return false;
     224              : }
     225              : 
     226              : bool
     227     42917073 : sbr_lazy_vector::bb_range_p (const_basic_block bb)
     228              : {
     229     42917073 :   return bitmap_bit_p (m_has_value, bb->index);
     230              : }
     231              : 
     232              : // This class implements the on entry cache via a sparse bitmap.
     233              : // It uses the quad bit routines to access 4 bits at a time.
     234              : // A value of 0 (the default) means there is no entry, and a value of
     235              : // 1 thru SBR_NUM represents an element in the m_range vector.
     236              : // Varying is given the first value (1) and pre-cached.
     237              : // SBR_NUM + 1 represents the value of UNDEFINED, and is never stored.
     238              : // SBR_NUM is the number of values that can be cached.
     239              : // Indexes are 1..SBR_NUM and are stored locally at m_range[0..SBR_NUM-1]
     240              : 
     241              : #define SBR_NUM         14
     242              : #define SBR_UNDEF       SBR_NUM + 1
     243              : #define SBR_VARYING     1
     244              : 
     245              : class sbr_sparse_bitmap : public ssa_block_ranges
     246              : {
     247              : public:
     248              :   sbr_sparse_bitmap (tree t, vrange_allocator *allocator, bitmap_obstack *bm);
     249              :   virtual bool set_bb_range (const_basic_block bb, const vrange &r) override;
     250              :   virtual bool get_bb_range (vrange &r, const_basic_block bb) override;
     251              :   virtual bool bb_range_p (const_basic_block bb) override;
     252              : private:
     253              :   void bitmap_set_quad (bitmap head, int quad, int quad_value);
     254              :   int bitmap_get_quad (const_bitmap head, int quad);
     255              :   vrange_allocator *m_range_allocator;
     256              :   vrange_storage *m_range[SBR_NUM];
     257              :   bitmap_head bitvec;
     258              :   tree m_type;
     259              : };
     260              : 
     261              : // Initialize a block cache for an ssa_name of type T.
     262              : 
     263        86247 : sbr_sparse_bitmap::sbr_sparse_bitmap (tree t, vrange_allocator *allocator,
     264        86247 :                                       bitmap_obstack *bm)
     265        86247 :   : ssa_block_ranges (t)
     266              : {
     267        86247 :   gcc_checking_assert (TYPE_P (t));
     268        86247 :   m_type = t;
     269        86247 :   bitmap_initialize (&bitvec, bm);
     270        86247 :   bitmap_tree_view (&bitvec);
     271        86247 :   m_range_allocator = allocator;
     272              :   // Pre-cache varying.
     273        86247 :   m_range[0] = m_range_allocator->clone_varying (t);
     274              :   // Pre-cache zero and non-zero values for pointers.
     275        86247 :   if (POINTER_TYPE_P (t))
     276              :     {
     277         1188 :       prange nonzero;
     278         1188 :       nonzero.set_nonzero (t);
     279         1188 :       m_range[1] = m_range_allocator->clone (nonzero);
     280         1188 :       prange zero;
     281         1188 :       zero.set_zero (t);
     282         1188 :       m_range[2] = m_range_allocator->clone (zero);
     283         1188 :     }
     284              :   else
     285        85059 :     m_range[1] = m_range[2] = NULL;
     286              :   // Clear SBR_NUM entries.
     287      1034964 :   for (int x = 3; x < SBR_NUM; x++)
     288       948717 :     m_range[x] = 0;
     289        86247 : }
     290              : 
     291              : // Set 4 bit values in a sparse bitmap. This allows a bitmap to
     292              : // function as a sparse array of 4 bit values.
     293              : // QUAD is the index, QUAD_VALUE is the 4 bit value to set.
     294              : 
     295              : inline void
     296       445818 : sbr_sparse_bitmap::bitmap_set_quad (bitmap head, int quad, int quad_value)
     297              : {
     298       445818 :   bitmap_set_aligned_chunk (head, quad, 4, (BITMAP_WORD) quad_value);
     299              : }
     300              : 
     301              : // Get a 4 bit value from a sparse bitmap. This allows a bitmap to
     302              : // function as a sparse array of 4 bit values.
     303              : // QUAD is the index.
     304              : inline int
     305     14256466 : sbr_sparse_bitmap::bitmap_get_quad (const_bitmap head, int quad)
     306              : {
     307     28512932 :   return (int) bitmap_get_aligned_chunk (head, quad, 4);
     308              : }
     309              : 
     310              : // Set the range on entry to basic block BB to R.
     311              : 
     312              : bool
     313       445818 : sbr_sparse_bitmap::set_bb_range (const_basic_block bb, const vrange &r)
     314              : {
     315       445818 :   if (r.undefined_p ())
     316              :     {
     317        18275 :       bitmap_set_quad (&bitvec, bb->index, SBR_UNDEF);
     318        18275 :       return true;
     319              :     }
     320              : 
     321              :   // Loop thru the values to see if R is already present.
     322       786123 :   for (int x = 0; x < SBR_NUM; x++)
     323       775108 :     if (!m_range[x] || m_range[x]->equal_p (r))
     324              :       {
     325       416528 :         if (!m_range[x])
     326       100374 :           m_range[x] = m_range_allocator->clone (r);
     327       416528 :         bitmap_set_quad (&bitvec, bb->index, x + 1);
     328       416528 :         return true;
     329              :       }
     330              :   // All values are taken, default to VARYING.
     331        11015 :   bitmap_set_quad (&bitvec, bb->index, SBR_VARYING);
     332        11015 :   return false;
     333              : }
     334              : 
     335              : // Return the range associated with block BB in R.  Return false if
     336              : // there is no range.
     337              : 
     338              : bool
     339     11875988 : sbr_sparse_bitmap::get_bb_range (vrange &r, const_basic_block bb)
     340              : {
     341     11875988 :   int value = bitmap_get_quad (&bitvec, bb->index);
     342              : 
     343     11875988 :   if (!value)
     344              :     return false;
     345              : 
     346      1833559 :   gcc_checking_assert (value <= SBR_UNDEF);
     347      1833559 :   if (value == SBR_UNDEF)
     348        36338 :     r.set_undefined ();
     349              :   else
     350      1797221 :     m_range[value - 1]->get_vrange (r, m_type);
     351              :   return true;
     352              : }
     353              : 
     354              : // Return true if a range is present.
     355              : 
     356              : bool
     357      2380478 : sbr_sparse_bitmap::bb_range_p (const_basic_block bb)
     358              : {
     359      2380478 :   return (bitmap_get_quad (&bitvec, bb->index) != 0);
     360              : }
     361              : 
     362              : // -------------------------------------------------------------------------
     363              : 
     364              : // Initialize the block cache.
     365              : 
     366     27881050 : block_range_cache::block_range_cache ()
     367              : {
     368     27881050 :   bitmap_obstack_initialize (&m_bitmaps);
     369     27881050 :   m_ssa_ranges.create (0);
     370     55762100 :   m_ssa_ranges.safe_grow_cleared (num_ssa_names);
     371     27881050 :   m_range_allocator = new vrange_allocator;
     372     27881050 : }
     373              : 
     374              : // Remove any m_block_caches which have been created.
     375              : 
     376     27881049 : block_range_cache::~block_range_cache ()
     377              : {
     378     27881049 :   delete m_range_allocator;
     379              :   // Release the vector itself.
     380     27881049 :   m_ssa_ranges.release ();
     381     27881049 :   bitmap_obstack_release (&m_bitmaps);
     382     27881049 : }
     383              : 
     384              : // Set the range for NAME on entry to block BB to R.
     385              : // If it has not been accessed yet, allocate it first.
     386              : 
     387              : bool
     388     74505413 : block_range_cache::set_bb_range (tree name, const_basic_block bb,
     389              :                                  const vrange &r)
     390              : {
     391     74505413 :   unsigned v = SSA_NAME_VERSION (name);
     392     74505413 :   if (v >= m_ssa_ranges.length ())
     393            2 :     m_ssa_ranges.safe_grow_cleared (num_ssa_names);
     394              : 
     395     74505413 :   if (!m_ssa_ranges[v])
     396              :     {
     397              :       // Use sparse bitmap representation if there are too many basic blocks.
     398     28431593 :       if (last_basic_block_for_fn (cfun) > param_vrp_sparse_threshold)
     399              :         {
     400        86247 :           void *r = m_range_allocator->alloc (sizeof (sbr_sparse_bitmap));
     401        86247 :           m_ssa_ranges[v] = new (r) sbr_sparse_bitmap (TREE_TYPE (name),
     402              :                                                        m_range_allocator,
     403        86247 :                                                        &m_bitmaps);
     404              :         }
     405     28345346 :       else if (last_basic_block_for_fn (cfun) < param_vrp_vector_threshold)
     406              :         {
     407              :           // For small CFGs use the basic vector implemntation.
     408     24875341 :           void *r = m_range_allocator->alloc (sizeof (sbr_vector));
     409     24875341 :           m_ssa_ranges[v] = new (r) sbr_vector (TREE_TYPE (name),
     410     24875341 :                                                 m_range_allocator);
     411              :         }
     412              :       else
     413              :         {
     414              :           // Otherwise use the sparse vector implementation.
     415      3470005 :           void *r = m_range_allocator->alloc (sizeof (sbr_lazy_vector));
     416      3470005 :           m_ssa_ranges[v] = new (r) sbr_lazy_vector (TREE_TYPE (name),
     417              :                                                      m_range_allocator,
     418      3470005 :                                                      &m_bitmaps);
     419              :         }
     420              :     }
     421     74505413 :   return m_ssa_ranges[v]->set_bb_range (bb, r);
     422              : }
     423              : 
     424              : 
     425              : // Return a pointer to the ssa_block_cache for NAME.  If it has not been
     426              : // accessed yet, return NULL.
     427              : 
     428              : inline ssa_block_ranges *
     429   1100956008 : block_range_cache::query_block_ranges (tree name)
     430              : {
     431   1100956008 :   unsigned v = SSA_NAME_VERSION (name);
     432   1100956008 :   if (v >= m_ssa_ranges.length () || !m_ssa_ranges[v])
     433              :     return NULL;
     434              :   return m_ssa_ranges[v];
     435              : }
     436              : 
     437              : 
     438              : 
     439              : // Return the range for NAME on entry to BB in R.  Return true if there
     440              : // is one.
     441              : 
     442              : bool
     443    723942351 : block_range_cache::get_bb_range (vrange &r, tree name, const_basic_block bb)
     444              : {
     445    723942351 :   ssa_block_ranges *ptr = query_block_ranges (name);
     446    723942351 :   if (ptr)
     447    530678559 :     return ptr->get_bb_range (r, bb);
     448              :   return false;
     449              : }
     450              : 
     451              : // Return true if NAME has a range set in block BB.
     452              : 
     453              : bool
     454    377013657 : block_range_cache::bb_range_p (tree name, const_basic_block bb)
     455              : {
     456    377013657 :   ssa_block_ranges *ptr = query_block_ranges (name);
     457    377013657 :   if (ptr)
     458    282242821 :     return ptr->bb_range_p (bb);
     459              :   return false;
     460              : }
     461              : 
     462              : // Print all known block caches to file F.
     463              : 
     464              : void
     465            0 : block_range_cache::dump (FILE *f)
     466              : {
     467            0 :   unsigned x;
     468            0 :   for (x = 1; x < m_ssa_ranges.length (); ++x)
     469              :     {
     470            0 :       if (m_ssa_ranges[x])
     471              :         {
     472            0 :           fprintf (f, " Ranges for ");
     473            0 :           print_generic_expr (f, ssa_name (x), TDF_NONE);
     474            0 :           fprintf (f, ":\n");
     475            0 :           m_ssa_ranges[x]->dump (f);
     476            0 :           fprintf (f, "\n");
     477              :         }
     478              :     }
     479            0 : }
     480              : 
     481              : // Print all known ranges on entry to block BB to file F.
     482              : 
     483              : void
     484          248 : block_range_cache::dump (FILE *f, basic_block bb, bool print_varying)
     485              : {
     486          248 :   unsigned x;
     487          248 :   bool summarize_varying = false;
     488        12593 :   for (x = 1; x < m_ssa_ranges.length (); ++x)
     489              :     {
     490        12345 :       if (!m_ssa_ranges[x])
     491        22180 :         continue;
     492              : 
     493         1255 :       if (!gimple_range_ssa_p (ssa_name (x)))
     494            0 :         continue;
     495              : 
     496         1255 :       value_range r (TREE_TYPE (ssa_name (x)));
     497         1255 :       if (m_ssa_ranges[x]->get_bb_range (r, bb))
     498              :         {
     499          220 :           if (!print_varying && r.varying_p ())
     500              :             {
     501            0 :               summarize_varying = true;
     502            0 :               continue;
     503              :             }
     504          220 :           print_generic_expr (f, ssa_name (x), TDF_NONE);
     505          220 :           fprintf (f, "\t");
     506          220 :           r.dump(f);
     507          220 :           fprintf (f, "\n");
     508              :         }
     509         1255 :     }
     510              :   // If there were any varying entries, lump them all together.
     511          248 :   if (summarize_varying)
     512              :     {
     513            0 :       fprintf (f, "VARYING_P on entry : ");
     514            0 :       for (x = 1; x < m_ssa_ranges.length (); ++x)
     515              :         {
     516            0 :           if (!m_ssa_ranges[x])
     517            0 :             continue;
     518              : 
     519            0 :           if (!gimple_range_ssa_p (ssa_name (x)))
     520            0 :             continue;
     521              : 
     522            0 :           value_range r (TREE_TYPE (ssa_name (x)));
     523            0 :           if (m_ssa_ranges[x]->get_bb_range (r, bb))
     524              :             {
     525            0 :               if (r.varying_p ())
     526              :                 {
     527            0 :                   print_generic_expr (f, ssa_name (x), TDF_NONE);
     528            0 :                   fprintf (f, "  ");
     529              :                 }
     530              :             }
     531            0 :         }
     532            0 :       fprintf (f, "\n");
     533              :     }
     534          248 : }
     535              : 
     536              : // -------------------------------------------------------------------------
     537              : 
     538              : // Initialize an ssa cache.
     539              : 
     540     56880826 : ssa_cache::ssa_cache ()
     541              : {
     542     56880826 :   m_tab.create (0);
     543     56880826 :   m_range_allocator = new vrange_allocator;
     544     56880826 : }
     545              : 
     546              : // Deconstruct an ssa cache.
     547              : 
     548     56880816 : ssa_cache::~ssa_cache ()
     549              : {
     550     56880816 :   m_tab.release ();
     551     56880816 :   delete m_range_allocator;
     552     56880816 : }
     553              : 
     554              : // Enable a query to evaluate staements/ramnges based on picking up ranges
     555              : // from just an ssa-cache.
     556              : 
     557              : bool
     558          532 : ssa_cache::range_of_expr (vrange &r, tree expr, gimple *stmt)
     559              : {
     560          532 :   if (!gimple_range_ssa_p (expr))
     561            0 :     return get_tree_range (r, expr, stmt);
     562              : 
     563          532 :   if (!get_range (r, expr))
     564           20 :     gimple_range_global (r, expr, cfun);
     565              :   return true;
     566              : }
     567              : 
     568              : // Return TRUE if the global range of NAME has a cache entry.
     569              : 
     570              : bool
     571          236 : ssa_cache::has_range (tree name) const
     572              : {
     573          236 :   unsigned v = SSA_NAME_VERSION (name);
     574          236 :   if (v >= m_tab.length ())
     575              :     return false;
     576          230 :   return m_tab[v] != NULL;
     577              : }
     578              : 
     579              : // Retrieve the global range of NAME from cache memory if it exists.
     580              : // Return the value in R.
     581              : 
     582              : bool
     583   1175657701 : ssa_cache::get_range (vrange &r, tree name) const
     584              : {
     585   1175657701 :   unsigned v = SSA_NAME_VERSION (name);
     586   1175657701 :   if (v >= m_tab.length ())
     587              :     return false;
     588              : 
     589   1163640835 :   vrange_storage *stow = m_tab[v];
     590   1163640835 :   if (!stow)
     591              :     return false;
     592    949810359 :   stow->get_vrange (r, TREE_TYPE (name));
     593    949810359 :   return true;
     594              : }
     595              : 
     596              : // Set the range for NAME to R in the ssa cache.
     597              : // Return TRUE if there was already a range set, otherwise false.
     598              : 
     599              : bool
     600    155110784 : ssa_cache::set_range (tree name, const vrange &r)
     601              : {
     602    155110784 :   unsigned v = SSA_NAME_VERSION (name);
     603    155110784 :   if (v >= m_tab.length ())
     604     15806738 :     m_tab.safe_grow_cleared (num_ssa_names + 1);
     605              : 
     606    155110784 :   vrange_storage *m = m_tab[v];
     607    155110784 :   if (m && m->fits_p (r))
     608     23566548 :     m->set_vrange (r);
     609              :   else
     610    131544236 :     m_tab[v] = m_range_allocator->clone (r);
     611    155110784 :   return m != NULL;
     612              : }
     613              : 
     614              : // If NAME has a range, intersect it with R, otherwise set it to R.
     615              : // Return TRUE if the range is new or changes.
     616              : 
     617              : bool
     618          122 : ssa_cache::merge_range (tree name, const vrange &r)
     619              : {
     620          122 :   unsigned v = SSA_NAME_VERSION (name);
     621          122 :   if (v >= m_tab.length ())
     622           12 :     m_tab.safe_grow_cleared (num_ssa_names + 1);
     623              : 
     624          122 :   vrange_storage *m = m_tab[v];
     625              :   // Check if this is a new value.
     626          122 :   if (!m)
     627          121 :     m_tab[v] = m_range_allocator->clone (r);
     628              :   else
     629              :     {
     630            1 :       value_range curr (TREE_TYPE (name));
     631            1 :       m->get_vrange (curr, TREE_TYPE (name));
     632              :       // If there is no change, return false.
     633            1 :       if (!curr.intersect (r))
     634            1 :         return false;
     635              : 
     636            0 :       if (m->fits_p (curr))
     637            0 :         m->set_vrange (curr);
     638              :       else
     639            0 :         m_tab[v] = m_range_allocator->clone (curr);
     640            1 :     }
     641              :   return true;
     642              : }
     643              : 
     644              : // Set the range for NAME to R in the ssa cache.
     645              : 
     646              : void
     647            0 : ssa_cache::clear_range (tree name)
     648              : {
     649            0 :   unsigned v = SSA_NAME_VERSION (name);
     650            0 :   if (v >= m_tab.length ())
     651              :     return;
     652            0 :   m_tab[v] = NULL;
     653              : }
     654              : 
     655              : // Clear the ssa cache.
     656              : 
     657              : void
     658            0 : ssa_cache::clear ()
     659              : {
     660            0 :   if (m_tab.address ())
     661            0 :     memset (m_tab.address(), 0, m_tab.length () * sizeof (vrange *));
     662            0 : }
     663              : 
     664              : // Dump the contents of the ssa cache to F.
     665              : 
     666              : void
     667           61 : ssa_cache::dump (FILE *f)
     668              : {
     669         3138 :   for (unsigned x = 1; x < num_ssa_names; x++)
     670              :     {
     671         3077 :       if (!gimple_range_ssa_p (ssa_name (x)))
     672         1262 :         continue;
     673         1815 :       value_range r (TREE_TYPE (ssa_name (x)));
     674              :       // Dump all non-varying ranges.
     675         1815 :       if (get_range (r, ssa_name (x)) && !r.varying_p ())
     676              :         {
     677          298 :           print_generic_expr (f, ssa_name (x), TDF_NONE);
     678          298 :           fprintf (f, "  : ");
     679          298 :           r.dump (f);
     680          298 :           fprintf (f, "\n");
     681              :         }
     682         1815 :     }
     683              : 
     684           61 : }
     685              : 
     686              : // Construct an ssa_lazy_cache. If OB is specified, us it, otherwise use
     687              : // a local bitmap obstack.
     688              : 
     689     28999770 : ssa_lazy_cache::ssa_lazy_cache (bitmap_obstack *ob)
     690              : {
     691     28999770 :   if (!ob)
     692              :     {
     693     28999761 :       bitmap_obstack_initialize (&m_bitmaps);
     694     28999761 :       m_ob = &m_bitmaps;
     695              :     }
     696              :   else
     697            9 :     m_ob = ob;
     698     28999770 :   active_p = BITMAP_ALLOC (m_ob);
     699     28999770 : }
     700              : 
     701              : // Destruct an sa_lazy_cache.  Free the bitmap if it came from a different
     702              : // obstack, or release the obstack if it was a local one.
     703              : 
     704     28999761 : ssa_lazy_cache::~ssa_lazy_cache ()
     705              : {
     706     28999761 :   if (m_ob == &m_bitmaps)
     707     28999761 :     bitmap_obstack_release (&m_bitmaps);
     708              :   else
     709            0 :     BITMAP_FREE (active_p);
     710     28999761 : }
     711              : 
     712              : // Return true if NAME has an active range in the cache.
     713              : 
     714              : bool
     715          259 : ssa_lazy_cache::has_range (tree name) const
     716              : {
     717          259 :   return bitmap_bit_p (active_p, SSA_NAME_VERSION (name));
     718              : }
     719              : 
     720              : // Set range of NAME to R in a lazy cache.  Return FALSE if it did not already
     721              : // have a range.
     722              : 
     723              : bool
     724    109722943 : ssa_lazy_cache::set_range (tree name, const vrange &r)
     725              : {
     726    109722943 :   unsigned v = SSA_NAME_VERSION (name);
     727    109722943 :   if (!bitmap_set_bit (active_p, v))
     728              :     {
     729              :       // There is already an entry, simply set it.
     730     14997060 :       gcc_checking_assert (v < m_tab.length ());
     731     14997060 :       return ssa_cache::set_range (name, r);
     732              :     }
     733     94725883 :   if (v >= m_tab.length ())
     734     49435336 :     m_tab.safe_grow (num_ssa_names + 1);
     735     94725883 :   m_tab[v] = m_range_allocator->clone (r);
     736     94725883 :   return false;
     737              : }
     738              : 
     739              : // If NAME has a range, intersect it with R, otherwise set it to R.
     740              : // Return TRUE if the range is new or changes.
     741              : 
     742              : bool
     743          210 : ssa_lazy_cache::merge_range (tree name, const vrange &r)
     744              : {
     745          210 :   unsigned v = SSA_NAME_VERSION (name);
     746          210 :   if (!bitmap_set_bit (active_p, v))
     747              :     {
     748              :       // There is already an entry, simply merge it.
     749            1 :       gcc_checking_assert (v < m_tab.length ());
     750            1 :       return ssa_cache::merge_range (name, r);
     751              :     }
     752          209 :   if (v >= m_tab.length ())
     753          156 :     m_tab.safe_grow (num_ssa_names + 1);
     754          209 :   m_tab[v] = m_range_allocator->clone (r);
     755          209 :   return true;
     756              : }
     757              : 
     758              : // Merge all elements of CACHE with this cache.
     759              : // Any names in CACHE that are not in this one are added.
     760              : // Any names in both are merged via merge_range..
     761              : 
     762              : void
     763            7 : ssa_lazy_cache::merge (const ssa_lazy_cache &cache)
     764              : {
     765            7 :   unsigned x;
     766            7 :   bitmap_iterator bi;
     767           57 :   EXECUTE_IF_SET_IN_BITMAP (cache.active_p, 0, x, bi)
     768              :     {
     769           50 :       tree name = ssa_name (x);
     770           50 :       value_range r(TREE_TYPE (name));
     771           50 :       cache.get_range (r, name);
     772           50 :       merge_range (ssa_name (x), r);
     773           50 :     }
     774            7 : }
     775              : 
     776              : // Return TRUE if NAME has a range, and return it in R.
     777              : 
     778              : bool
     779    285153401 : ssa_lazy_cache::get_range (vrange &r, tree name) const
     780              : {
     781    285153401 :   if (!bitmap_bit_p (active_p, SSA_NAME_VERSION (name)))
     782              :     return false;
     783    123694430 :   return ssa_cache::get_range (r, name);
     784              : }
     785              : 
     786              : // Remove NAME from the active range list.
     787              : 
     788              : void
     789     55369024 : ssa_lazy_cache::clear_range (tree name)
     790              : {
     791     55369024 :   bitmap_clear_bit (active_p, SSA_NAME_VERSION (name));
     792     55369024 : }
     793              : 
     794              : // Remove all ranges from the active range list.
     795              : 
     796              : void
     797     35597546 : ssa_lazy_cache::clear ()
     798              : {
     799     35597546 :   bitmap_clear (active_p);
     800     35597546 : }
     801              : 
     802              : // --------------------------------------------------------------------------
     803              : 
     804              : 
     805              : // This class will manage the timestamps for each ssa_name.
     806              : // When a value is calculated, the timestamp is set to the current time.
     807              : // Current time is then incremented.  Any dependencies will already have
     808              : // been calculated, and will thus have older timestamps.
     809              : // If one of those values is ever calculated again, it will get a newer
     810              : // timestamp, and the "current_p" check will fail.
     811              : 
     812              : class temporal_cache
     813              : {
     814              : public:
     815              :   temporal_cache ();
     816              :   ~temporal_cache ();
     817              :   bool current_p (tree name, tree dep1, tree dep2) const;
     818              :   void set_timestamp (tree name);
     819              :   void set_always_current (tree name, bool value);
     820              :   bool always_current_p (tree name) const;
     821              : private:
     822              :   int temporal_value (unsigned ssa) const;
     823              :   int m_current_time;
     824              :   vec <int> m_timestamp;
     825              : };
     826              : 
     827              : inline
     828     27881050 : temporal_cache::temporal_cache ()
     829              : {
     830     27881050 :   m_current_time = 1;
     831     27881050 :   m_timestamp.create (0);
     832     55762100 :   m_timestamp.safe_grow_cleared (num_ssa_names);
     833     27881050 : }
     834              : 
     835              : inline
     836     27881049 : temporal_cache::~temporal_cache ()
     837              : {
     838     27881049 :   m_timestamp.release ();
     839     27881049 : }
     840              : 
     841              : // Return the timestamp value for SSA, or 0 if there isn't one.
     842              : 
     843              : inline int
     844    570215815 : temporal_cache::temporal_value (unsigned ssa) const
     845              : {
     846    570215815 :   if (ssa >= m_timestamp.length ())
     847              :     return 0;
     848    570215815 :   return abs (m_timestamp[ssa]);
     849              : }
     850              : 
     851              : // Return TRUE if the timestamp for NAME is newer than any of its dependents.
     852              : // Up to 2 dependencies can be checked.
     853              : 
     854              : bool
     855    342558414 : temporal_cache::current_p (tree name, tree dep1, tree dep2) const
     856              : {
     857    342558414 :   if (always_current_p (name))
     858              :     return true;
     859              : 
     860              :   // Any non-registered dependencies will have a value of 0 and thus be older.
     861              :   // Return true if time is newer than either dependent.
     862    335876124 :   int ts = temporal_value (SSA_NAME_VERSION (name));
     863    517846857 :   if (dep1 && ts < temporal_value (SSA_NAME_VERSION (dep1)))
     864              :     return false;
     865    336167659 :   if (dep2 && ts < temporal_value (SSA_NAME_VERSION (dep2)))
     866      6467084 :     return false;
     867              : 
     868              :   return true;
     869              : }
     870              : 
     871              : // This increments the global timer and sets the timestamp for NAME.
     872              : 
     873              : inline void
     874    122458108 : temporal_cache::set_timestamp (tree name)
     875              : {
     876    122458108 :   unsigned v = SSA_NAME_VERSION (name);
     877    122458108 :   if (v >= m_timestamp.length ())
     878            0 :     m_timestamp.safe_grow_cleared (num_ssa_names + 20);
     879    122458108 :   m_timestamp[v] = ++m_current_time;
     880    122458108 : }
     881              : 
     882              : // Set the timestamp to 0, marking it as "always up to date".
     883              : 
     884              : inline void
     885    277307048 : temporal_cache::set_always_current (tree name, bool value)
     886              : {
     887    277307048 :   unsigned v = SSA_NAME_VERSION (name);
     888    277307048 :   if (v >= m_timestamp.length ())
     889         1340 :     m_timestamp.safe_grow_cleared (num_ssa_names + 20);
     890              : 
     891    277307048 :   int ts = abs (m_timestamp[v]);
     892              :   // If this does not have a timestamp, create one.
     893    277307048 :   if (ts == 0)
     894    127085400 :     ts = ++m_current_time;
     895    277307048 :   m_timestamp[v] = value ? -ts : ts;
     896    277307048 : }
     897              : 
     898              : // Return true if NAME is always current.
     899              : 
     900              : inline bool
     901    342558414 : temporal_cache::always_current_p (tree name) const
     902              : {
     903    342558414 :   unsigned v = SSA_NAME_VERSION (name);
     904    342558414 :   if (v >= m_timestamp.length ())
     905              :     return false;
     906    342558414 :   return m_timestamp[v] <= 0;
     907              : }
     908              : 
     909              : // --------------------------------------------------------------------------
     910              : 
     911              : // This class provides an abstraction of a list of blocks to be updated
     912              : // by the cache.  It is currently a stack but could be changed.  It also
     913              : // maintains a list of blocks which have failed propagation, and does not
     914              : // enter any of those blocks into the list.
     915              : 
     916              : // A vector over the BBs is maintained, and an entry of 0 means it is not in
     917              : // a list.  Otherwise, the entry is the next block in the list. -1 terminates
     918              : // the list.  m_head points to the top of the list, -1 if the list is empty.
     919              : 
     920              : class update_list
     921              : {
     922              : public:
     923              :   update_list ();
     924              :   ~update_list ();
     925              :   void add (basic_block bb);
     926              :   basic_block pop ();
     927    156831553 :   inline bool empty_p () { return m_update_head == -1; }
     928      5968140 :   inline void clear_failures () { bitmap_clear (m_propfail); }
     929            6 :   inline void propagation_failed (basic_block bb)
     930            6 :                                   { bitmap_set_bit (m_propfail, bb->index); }
     931              : private:
     932              :   vec<int> m_update_list;
     933              :   int m_update_head;
     934              :   bitmap m_propfail;
     935              :   bitmap_obstack m_bitmaps;
     936              : };
     937              : 
     938              : // Create an update list.
     939              : 
     940     27881050 : update_list::update_list ()
     941              : {
     942     27881050 :   m_update_list.create (0);
     943     27881050 :   m_update_list.safe_grow_cleared (last_basic_block_for_fn (cfun) + 64);
     944     27881050 :   m_update_head = -1;
     945     27881050 :   bitmap_obstack_initialize (&m_bitmaps);
     946     27881050 :   m_propfail = BITMAP_ALLOC (&m_bitmaps);
     947     27881050 : }
     948              : 
     949              : // Destroy an update list.
     950              : 
     951     27881049 : update_list::~update_list ()
     952              : {
     953     27881049 :   m_update_list.release ();
     954     27881049 :   bitmap_obstack_release (&m_bitmaps);
     955     27881049 : }
     956              : 
     957              : // Add BB to the list of blocks to update, unless it's already in the list.
     958              : 
     959              : void
     960     13675393 : update_list::add (basic_block bb)
     961              : {
     962     13675393 :   int i = bb->index;
     963              :   // If propagation has failed for BB, or its already in the list, don't
     964              :   // add it again.
     965     13675393 :   if ((unsigned)i >= m_update_list.length ())
     966           68 :     m_update_list.safe_grow_cleared (i + 64);
     967     13675393 :   if (!m_update_list[i] && !bitmap_bit_p (m_propfail, i))
     968              :     {
     969     12944746 :       if (empty_p ())
     970              :         {
     971      7382184 :           m_update_head = i;
     972      7382184 :           m_update_list[i] = -1;
     973              :         }
     974              :       else
     975              :         {
     976      5562562 :           gcc_checking_assert (m_update_head > 0);
     977      5562562 :           m_update_list[i] = m_update_head;
     978      5562562 :           m_update_head = i;
     979              :         }
     980              :     }
     981     13675393 : }
     982              : 
     983              : // Remove a block from the list.
     984              : 
     985              : basic_block
     986     12944746 : update_list::pop ()
     987              : {
     988     12944746 :   gcc_checking_assert (!empty_p ());
     989     12944746 :   basic_block bb = BASIC_BLOCK_FOR_FN (cfun, m_update_head);
     990     12944746 :   int pop = m_update_head;
     991     12944746 :   m_update_head = m_update_list[pop];
     992     12944746 :   m_update_list[pop] = 0;
     993     12944746 :   return bb;
     994              : }
     995              : 
     996              : // --------------------------------------------------------------------------
     997              : 
     998     27881050 : ranger_cache::ranger_cache (int not_executable_flag, bool use_imm_uses)
     999              : {
    1000     27881050 :   m_workback = vNULL;
    1001     27881050 :   m_temporal = new temporal_cache;
    1002              : 
    1003              :   // If DOM info is available, spawn an oracle as well.
    1004     27881050 :   create_relation_oracle ();
    1005              :   // Create an infer oracle using this cache as the range query.  The cache
    1006              :   // version acts as a read-only query, and will spawn no additional lookups.
    1007              :   // It just ues what is already known.
    1008     27881050 :   create_infer_oracle (this, use_imm_uses);
    1009     27881050 :   create_gori (not_executable_flag, param_vrp_switch_limit);
    1010              : 
    1011     27881050 :   unsigned x, lim = last_basic_block_for_fn (cfun);
    1012              :   // Calculate outgoing range info upfront.  This will fully populate the
    1013              :   // m_maybe_variant bitmap which will help eliminate processing of names
    1014              :   // which never have their ranges adjusted.
    1015    358246214 :   for (x = 0; x < lim ; x++)
    1016              :     {
    1017    330365164 :       basic_block bb = BASIC_BLOCK_FOR_FN (cfun, x);
    1018    330365164 :       if (bb)
    1019    311434578 :         gori_ssa ()->exports (bb);
    1020              :     }
    1021     27881050 :   m_update = new update_list ();
    1022     27881050 : }
    1023              : 
    1024     27881049 : ranger_cache::~ranger_cache ()
    1025              : {
    1026     27881049 :   delete m_update;
    1027     27881049 :   destroy_infer_oracle ();
    1028     27881049 :   destroy_relation_oracle ();
    1029     55762098 :   delete m_temporal;
    1030     27881049 :   m_workback.release ();
    1031     27881049 : }
    1032              : 
    1033              : // Dump the global caches to file F.  if GORI_DUMP is true, dump the
    1034              : // gori map as well.
    1035              : 
    1036              : void
    1037           45 : ranger_cache::dump (FILE *f)
    1038              : {
    1039           45 :   fprintf (f, "Non-varying global ranges:\n");
    1040           45 :   fprintf (f, "=========================:\n");
    1041           45 :   m_globals.dump (f);
    1042           45 :   fprintf (f, "\n");
    1043           45 : }
    1044              : 
    1045              : // Dump the caches for basic block BB to file F.
    1046              : 
    1047              : void
    1048          248 : ranger_cache::dump_bb (FILE *f, basic_block bb)
    1049              : {
    1050          248 :   gori_ssa ()->dump (f, bb, false);
    1051          248 :   m_on_entry.dump (f, bb);
    1052          248 :   m_relation->dump (f, bb);
    1053          248 : }
    1054              : 
    1055              : // Get the global range for NAME, and return in R.  Return false if the
    1056              : // global range is not set, and return the legacy global value in R.
    1057              : 
    1058              : bool
    1059    835914117 : ranger_cache::get_global_range (vrange &r, tree name) const
    1060              : {
    1061    835914117 :   if (m_globals.get_range (r, name))
    1062              :     return true;
    1063    195902212 :   gimple_range_global (r, name);
    1064    195902212 :   return false;
    1065              : }
    1066              : 
    1067              : // Get the global range for NAME, and return in R.  Return false if the
    1068              : // global range is not set, and R will contain the legacy global value.
    1069              : // CURRENT_P is set to true if the value was in cache and not stale.
    1070              : // Otherwise, set CURRENT_P to false and mark as it always current.
    1071              : // If the global cache did not have a value, initialize it as well.
    1072              : // After this call, the global cache will have a value.
    1073              : 
    1074              : bool
    1075    344128230 : ranger_cache::get_global_range (vrange &r, tree name, bool &current_p)
    1076              : {
    1077    344128230 :   bool had_global = get_global_range (r, name);
    1078              : 
    1079              :   // If there was a global value, set current flag, otherwise set a value.
    1080    344128230 :   current_p = false;
    1081    344128230 :   if (had_global)
    1082    434199360 :     current_p = r.singleton_p ()
    1083    433973161 :                 || m_temporal->current_p (name, gori_ssa ()->depend1 (name),
    1084    216873481 :                                           gori_ssa ()->depend2 (name));
    1085              :   else
    1086              :     {
    1087              :       // If no global value has been set and value is VARYING, fold the stmt
    1088              :       // using just global ranges to get a better initial value.
    1089              :       // After inlining we tend to decide some things are constant, so
    1090              :       // so not do this evaluation after inlining.
    1091    127028550 :       if (r.varying_p () && !cfun->after_inlining)
    1092              :         {
    1093     20375164 :           gimple *s = SSA_NAME_DEF_STMT (name);
    1094              :           // Do not process PHIs as SCEV may be in use and it can
    1095              :           // spawn cyclic lookups.
    1096     20375164 :           if (gimple_get_lhs (s) == name && !is_a<gphi *> (s))
    1097              :             {
    1098     15924850 :               if (!fold_range (r, s, get_global_range_query ()))
    1099            0 :                 gimple_range_global (r, name);
    1100              :             }
    1101              :         }
    1102    127028550 :       m_globals.set_range (name, r);
    1103              :     }
    1104              : 
    1105              :   // If the existing value was not current, mark it as always current.
    1106    344128230 :   if (!current_p)
    1107    138536941 :     m_temporal->set_always_current (name, true);
    1108    344128230 :   return had_global;
    1109              : }
    1110              : 
    1111              : // Consumers of NAME that have already calculated values should recalculate.
    1112              : // Accomplished by updating the timestamp.
    1113              : 
    1114              : void
    1115     62336818 : ranger_cache::update_consumers (tree name)
    1116              : {
    1117     62336818 :   m_temporal->set_timestamp (name);
    1118     62336818 : }
    1119              : 
    1120              : //  Set the global range of NAME to R and give it a timestamp.
    1121              : 
    1122              : void
    1123    138770107 : ranger_cache::set_global_range (tree name, const vrange &r, bool changed)
    1124              : {
    1125              :   // Setting a range always clears the always_current flag.
    1126    138770107 :   m_temporal->set_always_current (name, false);
    1127    138770107 :   if (!changed)
    1128              :     {
    1129              :       // If there are dependencies, make sure this is not out of date.
    1130    125684933 :       if (!m_temporal->current_p (name, gori_ssa ()->depend1 (name),
    1131    125684933 :                                  gori_ssa ()->depend2 (name)))
    1132     47036116 :         m_temporal->set_timestamp (name);
    1133    125684933 :       return;
    1134              :     }
    1135     13085174 :   if (m_globals.set_range (name, r))
    1136              :     {
    1137              :       // If there was already a range set, propagate the new value.
    1138     13028323 :       basic_block bb = gimple_bb (SSA_NAME_DEF_STMT (name));
    1139     13028323 :       if (!bb)
    1140         1225 :         bb = ENTRY_BLOCK_PTR_FOR_FN (cfun);
    1141              : 
    1142     13028323 :       if (DEBUG_RANGE_CACHE)
    1143            0 :         fprintf (dump_file, "   GLOBAL :");
    1144              : 
    1145     13028323 :       propagate_updated_value (name, bb);
    1146              :     }
    1147              :   // Constants no longer need to tracked.  Any further refinement has to be
    1148              :   // undefined. Propagation works better with constants. PR 100512.
    1149              :   // Pointers which resolve to non-zero also do not need
    1150              :   // tracking in the cache as they will never change.  See PR 98866.
    1151              :   // Timestamp must always be updated, or dependent calculations may
    1152              :   // not include this latest value. PR 100774.
    1153              : 
    1154     13085174 :   if (r.singleton_p ()
    1155     13085174 :       || (POINTER_TYPE_P (TREE_TYPE (name)) && r.nonzero_p ()))
    1156      2335497 :     gori_ssa ()->set_range_invariant (name);
    1157     13085174 :   m_temporal->set_timestamp (name);
    1158              : }
    1159              : 
    1160              : //  Provide lookup for the gori-computes class to access the best known range
    1161              : //  of an ssa_name in any given basic block.  Note, this does no additional
    1162              : //  lookups, just accesses the data that is already known.
    1163              : 
    1164              : // Get the range of NAME when the def occurs in block BB.  If BB is NULL
    1165              : // get the best global value available.
    1166              : 
    1167              : void
    1168    216048165 : ranger_cache::range_of_def (vrange &r, tree name, basic_block bb)
    1169              : {
    1170    216048165 :   gcc_checking_assert (gimple_range_ssa_p (name));
    1171    362939583 :   gcc_checking_assert (!bb || bb == gimple_bb (SSA_NAME_DEF_STMT (name)));
    1172              : 
    1173              :   // Pick up the best global range available.
    1174    216048165 :   if (!m_globals.get_range (r, name))
    1175              :     {
    1176              :       // If that fails, try to calculate the range using just global values.
    1177     29945109 :       gimple *s = SSA_NAME_DEF_STMT (name);
    1178     29945109 :       if (gimple_get_lhs (s) == name)
    1179     26721887 :         fold_range (r, s, get_global_range_query ());
    1180              :       else
    1181      3223222 :         gimple_range_global (r, name);
    1182              :     }
    1183    216048165 : }
    1184              : 
    1185              : // Get the range of NAME as it occurs on entry to block BB.  Use MODE for
    1186              : // lookups.
    1187              : 
    1188              : void
    1189    152593336 : ranger_cache::entry_range (vrange &r, tree name, basic_block bb,
    1190              :                            enum rfd_mode mode)
    1191              : {
    1192    152593336 :   if (bb == ENTRY_BLOCK_PTR_FOR_FN (cfun))
    1193              :     {
    1194            0 :       gimple_range_global (r, name);
    1195            0 :       return;
    1196              :     }
    1197              : 
    1198              :   // If NAME is invariant, simply return the defining range.
    1199    152593336 :   if (!gori ().has_edge_range_p (name))
    1200              :     {
    1201     33351468 :       range_of_def (r, name);
    1202     33351468 :       return;
    1203              :     }
    1204              : 
    1205              :   // Look for the on-entry value of name in BB from the cache.
    1206              :   // Otherwise pick up the best available global value.
    1207    119241868 :   if (!m_on_entry.get_bb_range (r, name, bb))
    1208     42256032 :     if (!range_from_dom (r, name, bb, mode))
    1209     35805279 :       range_of_def (r, name);
    1210              : }
    1211              : 
    1212              : // Get the range of NAME as it occurs on exit from block BB.  Use MODE for
    1213              : // lookups.
    1214              : 
    1215              : void
    1216    111941557 : ranger_cache::exit_range (vrange &r, tree name, basic_block bb,
    1217              :                           enum rfd_mode mode)
    1218              : {
    1219    111941557 :   if (bb == ENTRY_BLOCK_PTR_FOR_FN (cfun))
    1220              :     {
    1221        64521 :       gimple_range_global (r, name);
    1222        64521 :       return;
    1223              :     }
    1224              : 
    1225    111877036 :   gimple *s = SSA_NAME_DEF_STMT (name);
    1226    111877036 :   basic_block def_bb = gimple_bb (s);
    1227    111877036 :   if (def_bb == bb)
    1228     46786494 :     range_of_def (r, name, bb);
    1229              :   else
    1230     65090542 :     entry_range (r, name, bb, mode);
    1231              : }
    1232              : 
    1233              : // Get the range of NAME on edge E using MODE, return the result in R.
    1234              : // Always returns a range and true.
    1235              : 
    1236              : bool
    1237    101697285 : ranger_cache::edge_range (vrange &r, edge e, tree name, enum rfd_mode mode)
    1238              : {
    1239    101697285 :   exit_range (r, name, e->src, mode);
    1240              :   // If this is not an abnormal edge, check for inferred ranges on exit.
    1241    101697285 :   if ((e->flags & (EDGE_EH | EDGE_ABNORMAL)) == 0)
    1242    101391816 :     infer_oracle ().maybe_adjust_range (r, name, e->src);
    1243    101697285 :   value_range er (TREE_TYPE (name));
    1244    101697285 :   if (gori ().edge_range_p (er, e, name, *this))
    1245     22834539 :     r.intersect (er);
    1246    203394570 :   return true;
    1247    101697285 : }
    1248              : 
    1249              : 
    1250              : 
    1251              : // Implement range_of_expr.
    1252              : 
    1253              : bool
    1254    228034946 : ranger_cache::range_of_expr (vrange &r, tree name, gimple *stmt)
    1255              : {
    1256    228034946 :   if (!gimple_range_ssa_p (name))
    1257              :     {
    1258     40427228 :       get_tree_range (r, name, stmt);
    1259     40427228 :       return true;
    1260              :     }
    1261              : 
    1262    187607718 :   basic_block bb = gimple_bb (stmt);
    1263    187607718 :   gimple *def_stmt = SSA_NAME_DEF_STMT (name);
    1264    187607718 :   basic_block def_bb = gimple_bb (def_stmt);
    1265              : 
    1266    187607718 :   if (bb == def_bb)
    1267    100104924 :     range_of_def (r, name, bb);
    1268              :   else
    1269     87502794 :     entry_range (r, name, bb, RFD_NONE);
    1270              :   return true;
    1271              : }
    1272              : 
    1273              : 
    1274              : // Implement range_on_edge.  Always return the best available range using
    1275              : // the current cache values.
    1276              : 
    1277              : bool
    1278     76218520 : ranger_cache::range_on_edge (vrange &r, edge e, tree expr)
    1279              : {
    1280     76218520 :   if (gimple_range_ssa_p (expr))
    1281     73353830 :     return edge_range (r, e, expr, RFD_NONE);
    1282      2864690 :   return get_tree_range (r, expr, NULL);
    1283              : }
    1284              : 
    1285              : // Return a static range for NAME on entry to basic block BB in R.  If
    1286              : // calc is true, fill any cache entries required between BB and the
    1287              : // def block for NAME.  Otherwise, return false if the cache is empty.
    1288              : 
    1289              : bool
    1290    383989562 : ranger_cache::block_range (vrange &r, basic_block bb, tree name, bool calc)
    1291              : {
    1292    383989562 :   gcc_checking_assert (gimple_range_ssa_p (name));
    1293              : 
    1294              :   // If there are no range calculations anywhere in the IL, global range
    1295              :   // applies everywhere, so don't bother caching it.
    1296    383989562 :   if (!gori ().has_edge_range_p (name))
    1297              :     return false;
    1298              : 
    1299    241920804 :   if (calc)
    1300              :     {
    1301    119382442 :       gimple *def_stmt = SSA_NAME_DEF_STMT (name);
    1302    119382442 :       basic_block def_bb = NULL;
    1303    119382442 :       if (def_stmt)
    1304    119382442 :         def_bb = gimple_bb (def_stmt);
    1305    119382442 :       if (!def_bb)
    1306              :         {
    1307              :           // If we get to the entry block, this better be a default def
    1308              :           // or range_on_entry was called for a block not dominated by
    1309              :           // the def.  But it could be also SSA_NAME defined by a statement
    1310              :           // not yet in the IL (such as queued edge insertion), in that case
    1311              :           // just punt.
    1312     17340612 :           if (!SSA_NAME_IS_DEFAULT_DEF (name))
    1313              :             return false;
    1314     17340611 :           def_bb = ENTRY_BLOCK_PTR_FOR_FN (cfun);
    1315              :         }
    1316              : 
    1317              :       // There is no range on entry for the definition block.
    1318    119382441 :       if (def_bb == bb)
    1319              :         return false;
    1320              : 
    1321              :       // Otherwise, go figure out what is known in predecessor blocks.
    1322    119020279 :       fill_block_cache (name, bb, def_bb);
    1323    119020279 :       gcc_checking_assert (m_on_entry.bb_range_p (name, bb));
    1324              :     }
    1325    241558641 :   return m_on_entry.get_bb_range (r, name, bb);
    1326              : }
    1327              : 
    1328              : // If there is anything in the propagation update_list, continue
    1329              : // processing NAME until the list of blocks is empty.
    1330              : 
    1331              : void
    1332      5968140 : ranger_cache::propagate_cache (tree name)
    1333              : {
    1334      5968140 :   basic_block bb;
    1335      5968140 :   edge_iterator ei;
    1336      5968140 :   edge e;
    1337      5968140 :   tree type = TREE_TYPE (name);
    1338      5968140 :   value_range new_range (type);
    1339      5968140 :   value_range current_range (type);
    1340      5968140 :   value_range e_range (type);
    1341              : 
    1342              :   // Process each block by seeing if its calculated range on entry is
    1343              :   // the same as its cached value. If there is a difference, update
    1344              :   // the cache to reflect the new value, and check to see if any
    1345              :   // successors have cache entries which may need to be checked for
    1346              :   // updates.
    1347              : 
    1348     24881026 :   while (!m_update->empty_p ())
    1349              :     {
    1350     12944746 :       bb = m_update->pop ();
    1351     12944746 :       gcc_checking_assert (m_on_entry.bb_range_p (name, bb));
    1352     12944746 :       m_on_entry.get_bb_range (current_range, name, bb);
    1353              : 
    1354     12944746 :       if (DEBUG_RANGE_CACHE)
    1355              :         {
    1356            0 :           fprintf (dump_file, "FWD visiting block %d for ", bb->index);
    1357            0 :           print_generic_expr (dump_file, name, TDF_SLIM);
    1358            0 :           fprintf (dump_file, "  starting range : ");
    1359            0 :           current_range.dump (dump_file);
    1360            0 :           fprintf (dump_file, "\n");
    1361              :         }
    1362              : 
    1363              :       // Calculate the "new" range on entry by unioning the pred edges.
    1364     12944746 :       new_range.set_undefined ();
    1365     27438513 :       FOR_EACH_EDGE (e, ei, bb->preds)
    1366              :         {
    1367     18122577 :           edge_range (e_range, e, name, RFD_READ_ONLY);
    1368     18122577 :           if (DEBUG_RANGE_CACHE)
    1369              :             {
    1370            0 :               fprintf (dump_file, "   edge %d->%d :", e->src->index, bb->index);
    1371            0 :               e_range.dump (dump_file);
    1372            0 :               fprintf (dump_file, "\n");
    1373              :             }
    1374     18122577 :           new_range.union_ (e_range);
    1375     18122577 :           if (new_range.varying_p ())
    1376              :             break;
    1377              :         }
    1378              : 
    1379              :       // If the range on entry has changed, update it.
    1380     12944746 :       if (new_range != current_range)
    1381              :         {
    1382      7480903 :           bool ok_p = m_on_entry.set_bb_range (name, bb, new_range);
    1383              :           // If the cache couldn't set the value, mark it as failed.
    1384      7480903 :           if (!ok_p)
    1385            6 :             m_update->propagation_failed (bb);
    1386      7480903 :           if (DEBUG_RANGE_CACHE)
    1387              :             {
    1388            0 :               if (!ok_p)
    1389              :                 {
    1390            0 :                   fprintf (dump_file, "   Cache failure to store value:");
    1391            0 :                   print_generic_expr (dump_file, name, TDF_SLIM);
    1392            0 :                   fprintf (dump_file, "  ");
    1393              :                 }
    1394              :               else
    1395              :                 {
    1396            0 :                   fprintf (dump_file, "      Updating range to ");
    1397            0 :                   new_range.dump (dump_file);
    1398              :                 }
    1399            0 :               fprintf (dump_file, "\n      Updating blocks :");
    1400              :             }
    1401              :           // Mark each successor that has a range to re-check its range
    1402     19163267 :           FOR_EACH_EDGE (e, ei, bb->succs)
    1403     11682364 :             if (m_on_entry.bb_range_p (name, e->dest))
    1404              :               {
    1405      7089526 :                 if (DEBUG_RANGE_CACHE)
    1406            0 :                   fprintf (dump_file, " bb%d",e->dest->index);
    1407      7089526 :                 m_update->add (e->dest);
    1408              :               }
    1409      7480903 :           if (DEBUG_RANGE_CACHE)
    1410            0 :             fprintf (dump_file, "\n");
    1411              :         }
    1412              :     }
    1413      5968140 :   if (DEBUG_RANGE_CACHE)
    1414              :     {
    1415            0 :       fprintf (dump_file, "DONE visiting blocks for ");
    1416            0 :       print_generic_expr (dump_file, name, TDF_SLIM);
    1417            0 :       fprintf (dump_file, "\n");
    1418              :     }
    1419      5968140 :   m_update->clear_failures ();
    1420      5968140 : }
    1421              : 
    1422              : // Check to see if an update to the value for NAME in BB has any effect
    1423              : // on values already in the on-entry cache for successor blocks.
    1424              : // If it does, update them.  Don't visit any blocks which don't have a cache
    1425              : // entry.
    1426              : 
    1427              : void
    1428     55539163 : ranger_cache::propagate_updated_value (tree name, basic_block bb)
    1429              : {
    1430     55539163 :   edge e;
    1431     55539163 :   edge_iterator ei;
    1432              : 
    1433              :   // The update work list should be empty at this point.
    1434     55539163 :   gcc_checking_assert (m_update->empty_p ());
    1435     55539163 :   gcc_checking_assert (bb);
    1436              : 
    1437     55539163 :   if (DEBUG_RANGE_CACHE)
    1438              :     {
    1439            0 :       fprintf (dump_file, " UPDATE cache for ");
    1440            0 :       print_generic_expr (dump_file, name, TDF_SLIM);
    1441            0 :       fprintf (dump_file, " in BB %d : successors : ", bb->index);
    1442              :     }
    1443    161096992 :   FOR_EACH_EDGE (e, ei, bb->succs)
    1444              :     {
    1445              :       // Only update active cache entries.
    1446    105557829 :       if (m_on_entry.bb_range_p (name, e->dest))
    1447              :         {
    1448      5085471 :           m_update->add (e->dest);
    1449      5085471 :           if (DEBUG_RANGE_CACHE)
    1450            0 :             fprintf (dump_file, " UPDATE: bb%d", e->dest->index);
    1451              :         }
    1452              :     }
    1453     55539163 :     if (!m_update->empty_p ())
    1454              :       {
    1455      5017291 :         if (DEBUG_RANGE_CACHE)
    1456            0 :           fprintf (dump_file, "\n");
    1457      5017291 :         propagate_cache (name);
    1458              :       }
    1459              :     else
    1460              :       {
    1461     50521872 :         if (DEBUG_RANGE_CACHE)
    1462            0 :           fprintf (dump_file, "  : No updates!\n");
    1463              :       }
    1464     55539163 : }
    1465              : 
    1466              : // Make sure that the range-on-entry cache for NAME is set for block BB.
    1467              : // Work back through the CFG to DEF_BB ensuring the range is calculated
    1468              : // on the block/edges leading back to that point.
    1469              : 
    1470              : void
    1471    119020279 : ranger_cache::fill_block_cache (tree name, basic_block bb, basic_block def_bb)
    1472              : {
    1473    119020279 :   edge_iterator ei;
    1474    119020279 :   edge e;
    1475    119020279 :   tree type = TREE_TYPE (name);
    1476    119020279 :   value_range block_result (type);
    1477    119020279 :   value_range undefined (type);
    1478              : 
    1479              :   // At this point we shouldn't be looking at the def, entry block.
    1480    119020279 :   gcc_checking_assert (bb != def_bb && bb != ENTRY_BLOCK_PTR_FOR_FN (cfun));
    1481    119020279 :   unsigned start_length = m_workback.length ();
    1482              : 
    1483              :   // If the block cache is set, then we've already visited this block.
    1484    119020279 :   if (m_on_entry.bb_range_p (name, bb))
    1485              :     return;
    1486              : 
    1487     52245836 :   if (DEBUG_RANGE_CACHE)
    1488              :     {
    1489            0 :       fprintf (dump_file, "\n");
    1490            0 :       print_generic_expr (dump_file, name, TDF_SLIM);
    1491            0 :       fprintf (dump_file, " : ");
    1492              :     }
    1493              : 
    1494              :   // Check if a dominators can supply the range.
    1495     52245836 :   if (range_from_dom (block_result, name, bb, RFD_FILL))
    1496              :     {
    1497     51294987 :       if (DEBUG_RANGE_CACHE)
    1498              :         {
    1499            0 :           fprintf (dump_file, "Filled from dominator! :  ");
    1500            0 :           block_result.dump (dump_file);
    1501            0 :           fprintf (dump_file, "\n");
    1502              :         }
    1503              :       // See if any equivalences can refine it.
    1504              :       // PR 109462, like 108139 below, a one way equivalence introduced
    1505              :       // by a PHI node can also be through the definition side.  Disallow it.
    1506     51294987 :       tree equiv_name;
    1507     51294987 :       relation_kind rel;
    1508     51294987 :       int prec = TYPE_PRECISION (type);
    1509              :       // If there are too many basic blocks, do not attempt to process
    1510              :       // equivalencies.
    1511     51294987 :       if (last_basic_block_for_fn (cfun) > param_vrp_sparse_threshold)
    1512              :         {
    1513       387582 :           m_on_entry.set_bb_range (name, bb, block_result);
    1514       775134 :           gcc_checking_assert (m_workback.length () == start_length);
    1515              :           return;
    1516              :         }
    1517     60386060 :       FOR_EACH_PARTIAL_AND_FULL_EQUIV (m_relation, bb, name, equiv_name, rel)
    1518              :         {
    1519      9478655 :           basic_block equiv_bb = gimple_bb (SSA_NAME_DEF_STMT (equiv_name));
    1520              : 
    1521              :           // Ignore partial equivs that are smaller than this object.
    1522     16960427 :           if (rel != VREL_EQ && prec > pe_to_bits (rel))
    1523      3685762 :             continue;
    1524              : 
    1525              :           // Check if the equiv has any ranges calculated.
    1526      8457976 :           if (!gori ().has_edge_range_p (equiv_name))
    1527       362491 :             continue;
    1528              : 
    1529              :           // Check if the equiv definition dominates this block
    1530      8095485 :           if (equiv_bb == bb ||
    1531      7870137 :               (equiv_bb && !dominated_by_p (CDI_DOMINATORS, bb, equiv_bb)))
    1532      2302592 :             continue;
    1533              : 
    1534      5792893 :           if (DEBUG_RANGE_CACHE)
    1535              :             {
    1536            0 :               if (rel == VREL_EQ)
    1537            0 :                 fprintf (dump_file, "Checking Equivalence (");
    1538              :               else
    1539            0 :                 fprintf (dump_file, "Checking Partial equiv (");
    1540            0 :               print_relation (dump_file, rel);
    1541            0 :               fprintf (dump_file, ") ");
    1542            0 :               print_generic_expr (dump_file, equiv_name, TDF_SLIM);
    1543            0 :               fprintf (dump_file, "\n");
    1544              :             }
    1545      5792893 :           value_range equiv_range (TREE_TYPE (equiv_name));
    1546      5792893 :           if (range_from_dom (equiv_range, equiv_name, bb, RFD_READ_ONLY))
    1547              :             {
    1548      5792893 :               if (rel != VREL_EQ)
    1549      4055358 :                 range_cast (equiv_range, type);
    1550              :               else
    1551      1737535 :                 adjust_equivalence_range (equiv_range);
    1552              : 
    1553      5792893 :               if (block_result.intersect (equiv_range))
    1554              :                 {
    1555       321418 :                   if (DEBUG_RANGE_CACHE)
    1556              :                     {
    1557            0 :                       if (rel == VREL_EQ)
    1558            0 :                         fprintf (dump_file, "Equivalence update! :  ");
    1559              :                       else
    1560            0 :                         fprintf (dump_file, "Partial equiv update! :  ");
    1561            0 :                       print_generic_expr (dump_file, equiv_name, TDF_SLIM);
    1562            0 :                       fprintf (dump_file, " has range  :  ");
    1563            0 :                       equiv_range.dump (dump_file);
    1564            0 :                       fprintf (dump_file, " refining range to :");
    1565            0 :                       block_result.dump (dump_file);
    1566            0 :                       fprintf (dump_file, "\n");
    1567              :                     }
    1568              :                 }
    1569              :             }
    1570      5792893 :         }
    1571              : 
    1572     50907405 :       m_on_entry.set_bb_range (name, bb, block_result);
    1573     99220770 :       gcc_checking_assert (m_workback.length () == start_length);
    1574              :       return;
    1575              :     }
    1576              : 
    1577              :   // Visit each block back to the DEF.  Initialize each one to UNDEFINED.
    1578              :   // m_visited at the end will contain all the blocks that we needed to set
    1579              :   // the range_on_entry cache for.
    1580       950849 :   m_workback.safe_push (bb);
    1581       950849 :   undefined.set_undefined ();
    1582       950849 :   m_on_entry.set_bb_range (name, bb, undefined);
    1583       950849 :   gcc_checking_assert (m_update->empty_p ());
    1584              : 
    1585      6356922 :   while (m_workback.length () > start_length)
    1586              :     {
    1587      5406073 :       basic_block node = m_workback.pop ();
    1588      5406073 :       if (DEBUG_RANGE_CACHE)
    1589              :         {
    1590            0 :           fprintf (dump_file, "BACK visiting block %d for ", node->index);
    1591            0 :           print_generic_expr (dump_file, name, TDF_SLIM);
    1592            0 :           fprintf (dump_file, "\n");
    1593              :         }
    1594              : 
    1595     12946583 :       FOR_EACH_EDGE (e, ei, node->preds)
    1596              :         {
    1597      7540510 :           basic_block pred = e->src;
    1598      7540510 :           value_range r (TREE_TYPE (name));
    1599              : 
    1600      7540510 :           if (DEBUG_RANGE_CACHE)
    1601            0 :             fprintf (dump_file, "  %d->%d ",e->src->index, e->dest->index);
    1602              : 
    1603              :           // If the pred block is the def block add this BB to update list.
    1604      7540510 :           if (pred == def_bb)
    1605              :             {
    1606       901533 :               m_update->add (node);
    1607       901533 :               continue;
    1608              :             }
    1609              : 
    1610              :           // If the pred is entry but NOT def, then it is used before
    1611              :           // defined, it'll get set to [] and no need to update it.
    1612      6638977 :           if (pred == ENTRY_BLOCK_PTR_FOR_FN (cfun))
    1613              :             {
    1614            0 :               if (DEBUG_RANGE_CACHE)
    1615            0 :                 fprintf (dump_file, "entry: bail.");
    1616            0 :               continue;
    1617              :             }
    1618              : 
    1619              :           // Regardless of whether we have visited pred or not, if the
    1620              :           // pred has inferred ranges, revisit this block.
    1621              :           // Don't search the DOM tree.
    1622      6638977 :           if (infer_oracle ().has_range_p (pred, name))
    1623              :             {
    1624        12714 :               if (DEBUG_RANGE_CACHE)
    1625            0 :                 fprintf (dump_file, "Inferred range: update ");
    1626        12714 :               m_update->add (node);
    1627              :             }
    1628              : 
    1629              :           // If the pred block already has a range, or if it can contribute
    1630              :           // something new. Ie, the edge generates a range of some sort.
    1631      6638977 :           if (m_on_entry.get_bb_range (r, name, pred))
    1632              :             {
    1633      2183753 :               if (DEBUG_RANGE_CACHE)
    1634              :                 {
    1635            0 :                   fprintf (dump_file, "has cache, ");
    1636            0 :                   r.dump (dump_file);
    1637            0 :                   fprintf (dump_file, ", ");
    1638              :                 }
    1639      2183753 :               if (!r.undefined_p () || gori ().has_edge_range_p (name, e))
    1640              :                 {
    1641       586149 :                   m_update->add (node);
    1642       586149 :                   if (DEBUG_RANGE_CACHE)
    1643            0 :                     fprintf (dump_file, "update. ");
    1644              :                 }
    1645      2183753 :               continue;
    1646              :             }
    1647              : 
    1648      4455224 :           if (DEBUG_RANGE_CACHE)
    1649            0 :             fprintf (dump_file, "pushing undefined pred block.\n");
    1650              :           // If the pred hasn't been visited (has no range), add it to
    1651              :           // the list.
    1652      4455224 :           gcc_checking_assert (!m_on_entry.bb_range_p (name, pred));
    1653      4455224 :           m_on_entry.set_bb_range (name, pred, undefined);
    1654      4455224 :           m_workback.safe_push (pred);
    1655      7540510 :         }
    1656              :     }
    1657              : 
    1658       950849 :   if (DEBUG_RANGE_CACHE)
    1659            0 :     fprintf (dump_file, "\n");
    1660              : 
    1661              :   // Now fill in the marked blocks with values.
    1662       950849 :   propagate_cache (name);
    1663       950849 :   if (DEBUG_RANGE_CACHE)
    1664            0 :     fprintf (dump_file, "  Propagation update done.\n");
    1665    119020279 : }
    1666              : 
    1667              : // Resolve the range of BB if the dominators range is R by calculating incoming
    1668              : // edges to this block.  All lead back to the dominator so should be cheap.
    1669              : // The range for BB is set and returned in R.
    1670              : 
    1671              : void
    1672      4332936 : ranger_cache::resolve_dom (vrange &r, tree name, basic_block bb)
    1673              : {
    1674      4332936 :   basic_block def_bb = gimple_bb (SSA_NAME_DEF_STMT (name));
    1675      4332936 :   basic_block dom_bb = get_immediate_dominator (CDI_DOMINATORS, bb);
    1676              : 
    1677              :   // if it doesn't already have a value, store the incoming range.
    1678      4332936 :   if (!m_on_entry.bb_range_p (name, dom_bb) && def_bb != dom_bb)
    1679              :     {
    1680              :       // If the range can't be store, don't try to accumulate
    1681              :       // the range in PREV_BB due to excessive recalculations.
    1682      1142594 :       if (!m_on_entry.set_bb_range (name, dom_bb, r))
    1683            0 :         return;
    1684              :     }
    1685              :   // With the dominator set, we should be able to cheaply query
    1686              :   // each incoming edge now and accumulate the results.
    1687      4332936 :   r.set_undefined ();
    1688      4332936 :   edge e;
    1689      4332936 :   edge_iterator ei;
    1690      4332936 :   value_range er (TREE_TYPE (name));
    1691     14577712 :   FOR_EACH_EDGE (e, ei, bb->preds)
    1692              :     {
    1693              :       // If the predecessor is dominated by this block, then there is a back
    1694              :       // edge, and won't provide anything useful.  We'll actually end up with
    1695              :       // VARYING as we will not resolve this node.
    1696     10244776 :       if (dominated_by_p (CDI_DOMINATORS, e->src, bb))
    1697        23898 :         continue;
    1698     10220878 :       edge_range (er, e, name, RFD_READ_ONLY);
    1699     10220878 :       r.union_ (er);
    1700              :     }
    1701              :   // Set the cache in PREV_BB so it is not calculated again.
    1702      4332936 :   m_on_entry.set_bb_range (name, bb, r);
    1703      4332936 : }
    1704              : 
    1705              : // Get the range of NAME from dominators of BB and return it in R.  Search the
    1706              : // dominator tree based on MODE.
    1707              : 
    1708              : bool
    1709    100294761 : ranger_cache::range_from_dom (vrange &r, tree name, basic_block start_bb,
    1710              :                               enum rfd_mode mode)
    1711              : {
    1712    100294761 :   if (mode == RFD_NONE || !dom_info_available_p (CDI_DOMINATORS))
    1713     36756128 :     return false;
    1714              : 
    1715              :   // Search back to the definition block or entry block.
    1716     63538633 :   basic_block def_bb = gimple_bb (SSA_NAME_DEF_STMT (name));
    1717     63538633 :   if (def_bb == NULL)
    1718      8199624 :     def_bb = ENTRY_BLOCK_PTR_FOR_FN (cfun);
    1719              : 
    1720     63538633 :   basic_block bb;
    1721     63538633 :   basic_block prev_bb = start_bb;
    1722              : 
    1723              :   // Track any inferred ranges seen.
    1724     63538633 :   value_range infer (TREE_TYPE (name));
    1725     63538633 :   infer.set_varying (TREE_TYPE (name));
    1726              : 
    1727              :   // Range on entry to the DEF block should not be queried.
    1728     63538633 :   gcc_checking_assert (start_bb != def_bb);
    1729     63538633 :   unsigned start_limit = m_workback.length ();
    1730              : 
    1731              :   // Default value is global range.
    1732     63538633 :   get_global_range (r, name);
    1733              : 
    1734              :   // The dominator of EXIT_BLOCK doesn't seem to be set, so at least handle
    1735              :   // the common single exit cases.
    1736     63679731 :   if (start_bb == EXIT_BLOCK_PTR_FOR_FN (cfun) && single_pred_p (start_bb))
    1737       140880 :     bb = single_pred_edge (start_bb)->src;
    1738              :   else
    1739     63397753 :     bb = get_immediate_dominator (CDI_DOMINATORS, start_bb);
    1740              : 
    1741              :   // Search until a value is found, pushing blocks which may need calculating.
    1742    367365954 :   for ( ; bb; prev_bb = bb, bb = get_immediate_dominator (CDI_DOMINATORS, bb))
    1743              :     {
    1744              :       // Accumulate any block exit inferred ranges.
    1745    366666760 :       infer_oracle ().maybe_adjust_range (infer, name, bb);
    1746              : 
    1747              :       // This block has an outgoing range.
    1748    366666760 :       if (gori ().has_edge_range_p (name, bb))
    1749     44843131 :         m_workback.safe_push (prev_bb);
    1750              :       else
    1751              :         {
    1752              :           // Normally join blocks don't carry any new range information on
    1753              :           // incoming edges.  If the first incoming edge to this block does
    1754              :           // generate a range, calculate the ranges if all incoming edges
    1755              :           // are also dominated by the dominator.  (Avoids backedges which
    1756              :           // will break the rule of moving only upward in the dominator tree).
    1757              :           // If the first pred does not generate a range, then we will be
    1758              :           // using the dominator range anyway, so that's all the check needed.
    1759    321823629 :           if (EDGE_COUNT (prev_bb->preds) > 1
    1760    321823629 :               && gori ().has_edge_range_p (name, EDGE_PRED (prev_bb, 0)->src))
    1761              :             {
    1762       751541 :               edge e;
    1763       751541 :               edge_iterator ei;
    1764       751541 :               bool all_dom = true;
    1765      2542679 :               FOR_EACH_EDGE (e, ei, prev_bb->preds)
    1766      1791138 :                 if (e->src != bb
    1767      1791138 :                     && !dominated_by_p (CDI_DOMINATORS, e->src, bb))
    1768              :                   {
    1769              :                     all_dom = false;
    1770              :                     break;
    1771              :                   }
    1772       751541 :               if (all_dom)
    1773       751541 :                 m_workback.safe_push (prev_bb);
    1774              :             }
    1775              :         }
    1776              : 
    1777    366666760 :       if (def_bb == bb)
    1778              :         break;
    1779              : 
    1780    327208423 :       if (m_on_entry.get_bb_range (r, name, bb))
    1781              :         break;
    1782              :     }
    1783              : 
    1784     63538633 :   if (DEBUG_RANGE_CACHE)
    1785              :     {
    1786            0 :       fprintf (dump_file, "CACHE: BB %d DOM query for ", start_bb->index);
    1787            0 :       print_generic_expr (dump_file, name, TDF_SLIM);
    1788            0 :       fprintf (dump_file, ", found ");
    1789            0 :       r.dump (dump_file);
    1790            0 :       if (bb)
    1791            0 :         fprintf (dump_file, " at BB%d\n", bb->index);
    1792              :       else
    1793            0 :         fprintf (dump_file, " at function top\n");
    1794              :     }
    1795              : 
    1796              :   // Now process any blocks wit incoming edges that nay have adjustments.
    1797    109133305 :   while (m_workback.length () > start_limit)
    1798              :     {
    1799     45594672 :       value_range er (TREE_TYPE (name));
    1800     45594672 :       prev_bb = m_workback.pop ();
    1801     45594672 :       if (!single_pred_p (prev_bb))
    1802              :         {
    1803              :           // Non single pred means we need to cache a value in the dominator
    1804              :           // so we can cheaply calculate incoming edges to this block, and
    1805              :           // then store the resulting value.  If processing mode is not
    1806              :           // RFD_FILL, then the cache cant be stored to, so don't try.
    1807              :           // Otherwise this becomes a quadratic timed calculation.
    1808      6777881 :           if (mode == RFD_FILL)
    1809      4332936 :             resolve_dom (r, name, prev_bb);
    1810      6777881 :           continue;
    1811              :         }
    1812              : 
    1813     38816791 :       edge e = single_pred_edge (prev_bb);
    1814     38816791 :       bb = e->src;
    1815     38816791 :       if (gori ().edge_range_p (er, e, name, *this))
    1816              :         {
    1817     35183990 :           r.intersect (er);
    1818              :           // If this is a normal edge, apply any inferred ranges.
    1819     35183990 :           if ((e->flags & (EDGE_EH | EDGE_ABNORMAL)) == 0)
    1820     35183990 :             infer_oracle ().maybe_adjust_range (r, name, bb);
    1821              : 
    1822     35183990 :           if (DEBUG_RANGE_CACHE)
    1823              :             {
    1824            0 :               fprintf (dump_file, "CACHE: Adjusted edge range for %d->%d : ",
    1825              :                        bb->index, prev_bb->index);
    1826            0 :               r.dump (dump_file);
    1827            0 :               fprintf (dump_file, "\n");
    1828              :             }
    1829              :         }
    1830     45594672 :     }
    1831              : 
    1832              :   // Apply non-null if appropriate.
    1833     63538633 :   if (!has_abnormal_call_or_eh_pred_edge_p (start_bb))
    1834     63358816 :     r.intersect (infer);
    1835              : 
    1836     63538633 :   if (DEBUG_RANGE_CACHE)
    1837              :     {
    1838            0 :       fprintf (dump_file, "CACHE: Range for DOM returns : ");
    1839            0 :       r.dump (dump_file);
    1840            0 :       fprintf (dump_file, "\n");
    1841              :     }
    1842     63538633 :   return true;
    1843     63538633 : }
    1844              : 
    1845              : // This routine will register an inferred value in block BB, and possibly
    1846              : // update the on-entry cache if appropriate.
    1847              : 
    1848              : void
    1849     16349696 : ranger_cache::register_inferred_value (const vrange &ir, tree name,
    1850              :                                        basic_block bb)
    1851              : {
    1852     16349696 :   value_range r (TREE_TYPE (name));
    1853     16349696 :   if (!m_on_entry.get_bb_range (r, name, bb))
    1854     10244272 :     exit_range (r, name, bb, RFD_READ_ONLY);
    1855     16349696 :   if (r.intersect (ir))
    1856              :     {
    1857      4847920 :       m_on_entry.set_bb_range (name, bb, r);
    1858              :       // If this range was invariant before, remove invariant.
    1859      4847920 :       if (!gori ().has_edge_range_p (name))
    1860      4056590 :         gori_ssa ()->set_range_invariant (name, false);
    1861              :     }
    1862     16349696 : }
    1863              : 
    1864              : // This routine is used during a block walk to adjust any inferred ranges
    1865              : // of operands on stmt S.
    1866              : 
    1867              : void
    1868    252911694 : ranger_cache::apply_inferred_ranges (gimple *s)
    1869              : {
    1870    252911694 :   bool update = true;
    1871              : 
    1872    252911694 :   basic_block bb = gimple_bb (s);
    1873    252911694 :   gimple_infer_range infer(s, this);
    1874    252911694 :   if (infer.num () == 0)
    1875              :     return;
    1876              : 
    1877              :   // Do not update the on-entry cache for block ending stmts.
    1878     16041678 :   if (stmt_ends_bb_p (s))
    1879              :     {
    1880      1136130 :       edge_iterator ei;
    1881      1136130 :       edge e;
    1882      2060740 :       FOR_EACH_EDGE (e, ei, gimple_bb (s)->succs)
    1883      2055244 :         if (!(e->flags & (EDGE_ABNORMAL|EDGE_EH)))
    1884              :           break;
    1885      1136130 :       if (e == NULL)
    1886         5496 :         update = false;
    1887              :     }
    1888              : 
    1889     16041678 :   infer_oracle ().add_ranges (s, infer);
    1890     16041678 :   if (update)
    1891     32366003 :     for (unsigned x = 0; x < infer.num (); x++)
    1892     16329821 :       register_inferred_value (infer.range (x), infer.name (x), bb);
    1893              : }
        

Generated by: LCOV version 2.4-beta

LCOV profile is generated on x86_64 machine using following configure options: configure --disable-bootstrap --enable-coverage=opt --enable-languages=c,c++,fortran,go,jit,lto,rust,m2 --enable-host-shared. GCC test suite is run with the built compiler.