ira.c 175 KB
Newer Older
Vladimir Makarov committed
1
/* Integrated Register Allocator (IRA) entry point.
2
   Copyright (C) 2006-2019 Free Software Foundation, Inc.
Vladimir Makarov committed
3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39
   Contributed by Vladimir Makarov <vmakarov@redhat.com>.

This file is part of GCC.

GCC is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free
Software Foundation; either version 3, or (at your option) any later
version.

GCC is distributed in the hope that it will be useful, but WITHOUT ANY
WARRANTY; without even the implied warranty of MERCHANTABILITY or
FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
for more details.

You should have received a copy of the GNU General Public License
along with GCC; see the file COPYING3.  If not see
<http://www.gnu.org/licenses/>.  */

/* The integrated register allocator (IRA) is a
   regional register allocator performing graph coloring on a top-down
   traversal of nested regions.  Graph coloring in a region is based
   on Chaitin-Briggs algorithm.  It is called integrated because
   register coalescing, register live range splitting, and choosing a
   better hard register are done on-the-fly during coloring.  Register
   coalescing and choosing a cheaper hard register is done by hard
   register preferencing during hard register assigning.  The live
   range splitting is a byproduct of the regional register allocation.

   Major IRA notions are:

     o *Region* is a part of CFG where graph coloring based on
       Chaitin-Briggs algorithm is done.  IRA can work on any set of
       nested CFG regions forming a tree.  Currently the regions are
       the entire function for the root region and natural loops for
       the other regions.  Therefore data structure representing a
       region is called loop_tree_node.

Vladimir Makarov committed
40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57
     o *Allocno class* is a register class used for allocation of
       given allocno.  It means that only hard register of given
       register class can be assigned to given allocno.  In reality,
       even smaller subset of (*profitable*) hard registers can be
       assigned.  In rare cases, the subset can be even smaller
       because our modification of Chaitin-Briggs algorithm requires
       that sets of hard registers can be assigned to allocnos forms a
       forest, i.e. the sets can be ordered in a way where any
       previous set is not intersected with given set or is a superset
       of given set.

     o *Pressure class* is a register class belonging to a set of
       register classes containing all of the hard-registers available
       for register allocation.  The set of all pressure classes for a
       target is defined in the corresponding machine-description file
       according some criteria.  Register pressure is calculated only
       for pressure classes and it affects some IRA decisions as
       forming allocation regions.
Vladimir Makarov committed
58 59 60

     o *Allocno* represents the live range of a pseudo-register in a
       region.  Besides the obvious attributes like the corresponding
Vladimir Makarov committed
61
       pseudo-register number, allocno class, conflicting allocnos and
Vladimir Makarov committed
62 63 64
       conflicting hard-registers, there are a few allocno attributes
       which are important for understanding the allocation algorithm:

Vladimir Makarov committed
65 66 67
       - *Live ranges*.  This is a list of ranges of *program points*
         where the allocno lives.  Program points represent places
         where a pseudo can be born or become dead (there are
Vladimir Makarov committed
68 69
         approximately two times more program points than the insns)
         and they are represented by integers starting with 0.  The
Vladimir Makarov committed
70 71 72 73 74 75
         live ranges are used to find conflicts between allocnos.
         They also play very important role for the transformation of
         the IRA internal representation of several regions into a one
         region representation.  The later is used during the reload
         pass work because each allocno represents all of the
         corresponding pseudo-registers.
Vladimir Makarov committed
76 77

       - *Hard-register costs*.  This is a vector of size equal to the
Vladimir Makarov committed
78 79 80 81 82 83 84
         number of available hard-registers of the allocno class.  The
         cost of a callee-clobbered hard-register for an allocno is
         increased by the cost of save/restore code around the calls
         through the given allocno's life.  If the allocno is a move
         instruction operand and another operand is a hard-register of
         the allocno class, the cost of the hard-register is decreased
         by the move cost.
Vladimir Makarov committed
85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151

         When an allocno is assigned, the hard-register with minimal
         full cost is used.  Initially, a hard-register's full cost is
         the corresponding value from the hard-register's cost vector.
         If the allocno is connected by a *copy* (see below) to
         another allocno which has just received a hard-register, the
         cost of the hard-register is decreased.  Before choosing a
         hard-register for an allocno, the allocno's current costs of
         the hard-registers are modified by the conflict hard-register
         costs of all of the conflicting allocnos which are not
         assigned yet.

       - *Conflict hard-register costs*.  This is a vector of the same
         size as the hard-register costs vector.  To permit an
         unassigned allocno to get a better hard-register, IRA uses
         this vector to calculate the final full cost of the
         available hard-registers.  Conflict hard-register costs of an
         unassigned allocno are also changed with a change of the
         hard-register cost of the allocno when a copy involving the
         allocno is processed as described above.  This is done to
         show other unassigned allocnos that a given allocno prefers
         some hard-registers in order to remove the move instruction
         corresponding to the copy.

     o *Cap*.  If a pseudo-register does not live in a region but
       lives in a nested region, IRA creates a special allocno called
       a cap in the outer region.  A region cap is also created for a
       subregion cap.

     o *Copy*.  Allocnos can be connected by copies.  Copies are used
       to modify hard-register costs for allocnos during coloring.
       Such modifications reflects a preference to use the same
       hard-register for the allocnos connected by copies.  Usually
       copies are created for move insns (in this case it results in
       register coalescing).  But IRA also creates copies for operands
       of an insn which should be assigned to the same hard-register
       due to constraints in the machine description (it usually
       results in removing a move generated in reload to satisfy
       the constraints) and copies referring to the allocno which is
       the output operand of an instruction and the allocno which is
       an input operand dying in the instruction (creation of such
       copies results in less register shuffling).  IRA *does not*
       create copies between the same register allocnos from different
       regions because we use another technique for propagating
       hard-register preference on the borders of regions.

   Allocnos (including caps) for the upper region in the region tree
   *accumulate* information important for coloring from allocnos with
   the same pseudo-register from nested regions.  This includes
   hard-register and memory costs, conflicts with hard-registers,
   allocno conflicts, allocno copies and more.  *Thus, attributes for
   allocnos in a region have the same values as if the region had no
   subregions*.  It means that attributes for allocnos in the
   outermost region corresponding to the function have the same values
   as though the allocation used only one region which is the entire
   function.  It also means that we can look at IRA work as if the
   first IRA did allocation for all function then it improved the
   allocation for loops then their subloops and so on.

   IRA major passes are:

     o Building IRA internal representation which consists of the
       following subpasses:

       * First, IRA builds regions and creates allocnos (file
         ira-build.c) and initializes most of their attributes.

Vladimir Makarov committed
152 153 154
       * Then IRA finds an allocno class for each allocno and
         calculates its initial (non-accumulated) cost of memory and
         each hard-register of its allocno class (file ira-cost.c).
Vladimir Makarov committed
155

Kito Cheng committed
156
       * IRA creates live ranges of each allocno, calculates register
Vladimir Makarov committed
157
         pressure for each pressure class in each region, sets up
Vladimir Makarov committed
158 159 160 161 162 163 164 165 166 167 168 169
         conflict hard registers for each allocno and info about calls
         the allocno lives through (file ira-lives.c).

       * IRA removes low register pressure loops from the regions
         mostly to speed IRA up (file ira-build.c).

       * IRA propagates accumulated allocno info from lower region
         allocnos to corresponding upper region allocnos (file
         ira-build.c).

       * IRA creates all caps (file ira-build.c).

Vladimir Makarov committed
170 171 172 173 174
       * Having live-ranges of allocnos and their classes, IRA creates
         conflicting allocnos for each allocno.  Conflicting allocnos
         are stored as a bit vector or array of pointers to the
         conflicting allocnos whatever is more profitable (file
         ira-conflicts.c).  At this point IRA creates allocno copies.
Vladimir Makarov committed
175 176 177 178

     o Coloring.  Now IRA has all necessary info to start graph coloring
       process.  It is done in each region on top-down traverse of the
       region tree (file ira-color.c).  There are following subpasses:
H.J. Lu committed
179

Vladimir Makarov committed
180 181 182 183 184 185 186 187 188
       * Finding profitable hard registers of corresponding allocno
         class for each allocno.  For example, only callee-saved hard
         registers are frequently profitable for allocnos living
         through colors.  If the profitable hard register set of
         allocno does not form a tree based on subset relation, we use
         some approximation to form the tree.  This approximation is
         used to figure out trivial colorability of allocnos.  The
         approximation is a pretty rare case.

Vladimir Makarov committed
189 190 191 192 193
       * Putting allocnos onto the coloring stack.  IRA uses Briggs
         optimistic coloring which is a major improvement over
         Chaitin's coloring.  Therefore IRA does not spill allocnos at
         this point.  There is some freedom in the order of putting
         allocnos on the stack which can affect the final result of
Vladimir Makarov committed
194
         the allocation.  IRA uses some heuristics to improve the
195 196 197 198 199 200 201 202
         order.  The major one is to form *threads* from colorable
         allocnos and push them on the stack by threads.  Thread is a
         set of non-conflicting colorable allocnos connected by
         copies.  The thread contains allocnos from the colorable
         bucket or colorable allocnos already pushed onto the coloring
         stack.  Pushing thread allocnos one after another onto the
         stack increases chances of removing copies when the allocnos
         get the same hard reg.
Vladimir Makarov committed
203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233
	 
	 We also use a modification of Chaitin-Briggs algorithm which
         works for intersected register classes of allocnos.  To
         figure out trivial colorability of allocnos, the mentioned
         above tree of hard register sets is used.  To get an idea how
         the algorithm works in i386 example, let us consider an
         allocno to which any general hard register can be assigned.
         If the allocno conflicts with eight allocnos to which only
         EAX register can be assigned, given allocno is still
         trivially colorable because all conflicting allocnos might be
         assigned only to EAX and all other general hard registers are
         still free.

	 To get an idea of the used trivial colorability criterion, it
	 is also useful to read article "Graph-Coloring Register
	 Allocation for Irregular Architectures" by Michael D. Smith
	 and Glen Holloway.  Major difference between the article
	 approach and approach used in IRA is that Smith's approach
	 takes register classes only from machine description and IRA
	 calculate register classes from intermediate code too
	 (e.g. an explicit usage of hard registers in RTL code for
	 parameter passing can result in creation of additional
	 register classes which contain or exclude the hard
	 registers).  That makes IRA approach useful for improving
	 coloring even for architectures with regular register files
	 and in fact some benchmarking shows the improvement for
	 regular class architectures is even bigger than for irregular
	 ones.  Another difference is that Smith's approach chooses
	 intersection of classes of all insn operands in which a given
	 pseudo occurs.  IRA can use bigger classes if it is still
	 more profitable than memory usage.
Vladimir Makarov committed
234 235 236 237 238 239 240 241 242 243 244 245 246

       * Popping the allocnos from the stack and assigning them hard
         registers.  If IRA can not assign a hard register to an
         allocno and the allocno is coalesced, IRA undoes the
         coalescing and puts the uncoalesced allocnos onto the stack in
         the hope that some such allocnos will get a hard register
         separately.  If IRA fails to assign hard register or memory
         is more profitable for it, IRA spills the allocno.  IRA
         assigns the allocno the hard-register with minimal full
         allocation cost which reflects the cost of usage of the
         hard-register for the allocno and cost of usage of the
         hard-register for allocnos conflicting with given allocno.

Vladimir Makarov committed
247
       * Chaitin-Briggs coloring assigns as many pseudos as possible
Kito Cheng committed
248
         to hard registers.  After coloring we try to improve
Vladimir Makarov committed
249 250 251 252 253
         allocation with cost point of view.  We improve the
         allocation by spilling some allocnos and assigning the freed
         hard registers to other allocnos if it decreases the overall
         allocation cost.

254
       * After allocno assigning in the region, IRA modifies the hard
Vladimir Makarov committed
255 256 257 258 259 260
         register and memory costs for the corresponding allocnos in
         the subregions to reflect the cost of possible loads, stores,
         or moves on the border of the region and its subregions.
         When default regional allocation algorithm is used
         (-fira-algorithm=mixed), IRA just propagates the assignment
         for allocnos if the register pressure in the region for the
Vladimir Makarov committed
261 262
         corresponding pressure class is less than number of available
         hard registers for given pressure class.
Vladimir Makarov committed
263 264 265 266 267 268 269 270 271 272 273 274 275 276

     o Spill/restore code moving.  When IRA performs an allocation
       by traversing regions in top-down order, it does not know what
       happens below in the region tree.  Therefore, sometimes IRA
       misses opportunities to perform a better allocation.  A simple
       optimization tries to improve allocation in a region having
       subregions and containing in another region.  If the
       corresponding allocnos in the subregion are spilled, it spills
       the region allocno if it is profitable.  The optimization
       implements a simple iterative algorithm performing profitable
       transformations while they are still possible.  It is fast in
       practice, so there is no real need for a better time complexity
       algorithm.

Vladimir Makarov committed
277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295
     o Code change.  After coloring, two allocnos representing the
       same pseudo-register outside and inside a region respectively
       may be assigned to different locations (hard-registers or
       memory).  In this case IRA creates and uses a new
       pseudo-register inside the region and adds code to move allocno
       values on the region's borders.  This is done during top-down
       traversal of the regions (file ira-emit.c).  In some
       complicated cases IRA can create a new allocno to move allocno
       values (e.g. when a swap of values stored in two hard-registers
       is needed).  At this stage, the new allocno is marked as
       spilled.  IRA still creates the pseudo-register and the moves
       on the region borders even when both allocnos were assigned to
       the same hard-register.  If the reload pass spills a
       pseudo-register for some reason, the effect will be smaller
       because another allocno will still be in the hard-register.  In
       most cases, this is better then spilling both allocnos.  If
       reload does not change the allocation for the two
       pseudo-registers, the trivial move will be removed by
       post-reload optimizations.  IRA does not generate moves for
Vladimir Makarov committed
296 297
       allocnos assigned to the same hard register when the default
       regional allocation algorithm is used and the register pressure
Vladimir Makarov committed
298 299
       in the region for the corresponding pressure class is less than
       number of available hard registers for given pressure class.
Vladimir Makarov committed
300 301 302 303 304 305 306 307 308 309
       IRA also does some optimizations to remove redundant stores and
       to reduce code duplication on the region borders.

     o Flattening internal representation.  After changing code, IRA
       transforms its internal representation for several regions into
       one region representation (file ira-build.c).  This process is
       called IR flattening.  Such process is more complicated than IR
       rebuilding would be, but is much faster.

     o After IR flattening, IRA tries to assign hard registers to all
Kito Cheng committed
310
       spilled allocnos.  This is implemented by a simple and fast
Vladimir Makarov committed
311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330
       priority coloring algorithm (see function
       ira_reassign_conflict_allocnos::ira-color.c).  Here new allocnos
       created during the code change pass can be assigned to hard
       registers.

     o At the end IRA calls the reload pass.  The reload pass
       communicates with IRA through several functions in file
       ira-color.c to improve its decisions in

       * sharing stack slots for the spilled pseudos based on IRA info
         about pseudo-register conflicts.

       * reassigning hard-registers to all spilled pseudos at the end
         of each reload iteration.

       * choosing a better hard-register to spill based on IRA info
         about pseudo-register live ranges and the register pressure
         in places where the pseudo-register lives.

   IRA uses a lot of data representing the target processors.  These
Kito Cheng committed
331
   data are initialized in file ira.c.
Vladimir Makarov committed
332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354

   If function has no loops (or the loops are ignored when
   -fira-algorithm=CB is used), we have classic Chaitin-Briggs
   coloring (only instead of separate pass of coalescing, we use hard
   register preferencing).  In such case, IRA works much faster
   because many things are not made (like IR flattening, the
   spill/restore optimization, and the code change).

   Literature is worth to read for better understanding the code:

   o Preston Briggs, Keith D. Cooper, Linda Torczon.  Improvements to
     Graph Coloring Register Allocation.

   o David Callahan, Brian Koblenz.  Register allocation via
     hierarchical graph coloring.

   o Keith Cooper, Anshuman Dasgupta, Jason Eckhardt. Revisiting Graph
     Coloring Register Allocation: A Study of the Chaitin-Briggs and
     Callahan-Koblenz Algorithms.

   o Guei-Yuan Lueh, Thomas Gross, and Ali-Reza Adl-Tabatabai. Global
     Register Allocation Based on Graph Fusion.

Vladimir Makarov committed
355 356 357
   o Michael D. Smith and Glenn Holloway.  Graph-Coloring Register
     Allocation for Irregular Architectures

Vladimir Makarov committed
358 359 360 361 362 363 364 365 366 367 368
   o Vladimir Makarov. The Integrated Register Allocator for GCC.

   o Vladimir Makarov.  The top-down register allocator for irregular
     register file architectures.

*/


#include "config.h"
#include "system.h"
#include "coretypes.h"
369
#include "backend.h"
370
#include "target.h"
Vladimir Makarov committed
371
#include "rtl.h"
372
#include "tree.h"
373
#include "df.h"
374
#include "memmodel.h"
375 376
#include "tm_p.h"
#include "insn-config.h"
377
#include "regs.h"
378 379 380
#include "ira.h"
#include "ira-int.h"
#include "diagnostic-core.h"
381 382 383
#include "cfgrtl.h"
#include "cfgbuild.h"
#include "cfgcleanup.h"
Vladimir Makarov committed
384 385 386 387
#include "expr.h"
#include "tree-pass.h"
#include "output.h"
#include "reload.h"
388
#include "cfgloop.h"
389
#include "lra.h"
390
#include "dce.h"
391
#include "dbgcnt.h"
392
#include "rtl-iter.h"
393
#include "shrink-wrap.h"
394
#include "print-rtl.h"
Vladimir Makarov committed
395

396 397 398 399 400 401 402
struct target_ira default_target_ira;
struct target_ira_int default_target_ira_int;
#if SWITCHABLE_TARGET
struct target_ira *this_target_ira = &default_target_ira;
struct target_ira_int *this_target_ira_int = &default_target_ira_int;
#endif

Vladimir Makarov committed
403 404 405 406 407 408 409 410 411 412 413 414 415
/* A modified value of flag `-fira-verbose' used internally.  */
int internal_flag_ira_verbose;

/* Dump file of the allocator if it is not NULL.  */
FILE *ira_dump_file;

/* The number of elements in the following array.  */
int ira_spilled_reg_stack_slots_num;

/* The following array contains info about spilled pseudo-registers
   stack slots used in current function so far.  */
struct ira_spilled_reg_stack_slot *ira_spilled_reg_stack_slots;

416 417 418 419 420
/* Correspondingly overall cost of the allocation, overall cost before
   reload, cost of the allocnos assigned to hard-registers, cost of
   the allocnos assigned to memory, cost of loads, stores and register
   move insns generated for pseudo-register live range splitting (see
   ira-emit.c).  */
421 422 423
int64_t ira_overall_cost, overall_cost_before;
int64_t ira_reg_cost, ira_mem_cost;
int64_t ira_load_cost, ira_store_cost, ira_shuffle_cost;
Vladimir Makarov committed
424 425
int ira_move_loops_num, ira_additional_jumps_num;

426 427 428 429
/* All registers that can be eliminated.  */

HARD_REG_SET eliminable_regset;

430 431 432 433 434
/* Value of max_reg_num () before IRA work start.  This value helps
   us to recognize a situation when new pseudos were created during
   IRA work.  */
static int max_regno_before_ira;

Vladimir Makarov committed
435 436 437
/* Temporary hard reg set used for a different calculation.  */
static HARD_REG_SET temp_hard_regset;

438 439
#define last_mode_for_init_move_cost \
  (this_target_ira_int->x_last_mode_for_init_move_cost)
Vladimir Makarov committed
440 441 442 443 444 445 446 447 448 449 450 451


/* The function sets up the map IRA_REG_MODE_HARD_REGSET.  */
static void
setup_reg_mode_hard_regset (void)
{
  int i, m, hard_regno;

  for (m = 0; m < NUM_MACHINE_MODES; m++)
    for (hard_regno = 0; hard_regno < FIRST_PSEUDO_REGISTER; hard_regno++)
      {
	CLEAR_HARD_REG_SET (ira_reg_mode_hard_regset[hard_regno][m]);
452 453
	for (i = hard_regno_nregs (hard_regno, (machine_mode) m) - 1;
	     i >= 0; i--)
Vladimir Makarov committed
454 455 456 457 458 459 460
	  if (hard_regno + i < FIRST_PSEUDO_REGISTER)
	    SET_HARD_REG_BIT (ira_reg_mode_hard_regset[hard_regno][m],
			      hard_regno + i);
      }
}


461 462
#define no_unit_alloc_regs \
  (this_target_ira_int->x_no_unit_alloc_regs)
Vladimir Makarov committed
463 464 465 466 467 468 469 470 471 472 473 474 475 476

/* The function sets up the three arrays declared above.  */
static void
setup_class_hard_regs (void)
{
  int cl, i, hard_regno, n;
  HARD_REG_SET processed_hard_reg_set;

  ira_assert (SHRT_MAX >= FIRST_PSEUDO_REGISTER);
  for (cl = (int) N_REG_CLASSES - 1; cl >= 0; cl--)
    {
      COPY_HARD_REG_SET (temp_hard_regset, reg_class_contents[cl]);
      AND_COMPL_HARD_REG_SET (temp_hard_regset, no_unit_alloc_regs);
      CLEAR_HARD_REG_SET (processed_hard_reg_set);
477
      for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
478
	{
479 480
	  ira_non_ordered_class_hard_regs[cl][i] = -1;
	  ira_class_hard_reg_index[cl][i] = -1;
481
	}
Vladimir Makarov committed
482 483 484 485 486 487
      for (n = 0, i = 0; i < FIRST_PSEUDO_REGISTER; i++)
	{
#ifdef REG_ALLOC_ORDER
	  hard_regno = reg_alloc_order[i];
#else
	  hard_regno = i;
H.J. Lu committed
488
#endif
Vladimir Makarov committed
489 490 491 492 493 494 495 496 497 498 499 500
	  if (TEST_HARD_REG_BIT (processed_hard_reg_set, hard_regno))
	    continue;
	  SET_HARD_REG_BIT (processed_hard_reg_set, hard_regno);
      	  if (! TEST_HARD_REG_BIT (temp_hard_regset, hard_regno))
	    ira_class_hard_reg_index[cl][hard_regno] = -1;
	  else
	    {
	      ira_class_hard_reg_index[cl][hard_regno] = n;
	      ira_class_hard_regs[cl][n++] = hard_regno;
	    }
	}
      ira_class_hard_regs_num[cl] = n;
501 502 503 504
      for (n = 0, i = 0; i < FIRST_PSEUDO_REGISTER; i++)
	if (TEST_HARD_REG_BIT (temp_hard_regset, i))
	  ira_non_ordered_class_hard_regs[cl][n++] = i;
      ira_assert (ira_class_hard_regs_num[cl] == n);
Vladimir Makarov committed
505 506 507 508 509 510 511 512 513
    }
}

/* Set up global variables defining info about hard registers for the
   allocation.  These depend on USE_HARD_FRAME_P whose TRUE value means
   that we can use the hard frame pointer for the allocation.  */
static void
setup_alloc_regs (bool use_hard_frame_p)
{
514 515 516
#ifdef ADJUST_REG_ALLOC_ORDER
  ADJUST_REG_ALLOC_ORDER;
#endif
517
  COPY_HARD_REG_SET (no_unit_alloc_regs, fixed_nonglobal_reg_set);
Vladimir Makarov committed
518 519 520 521 522 523 524
  if (! use_hard_frame_p)
    SET_HARD_REG_BIT (no_unit_alloc_regs, HARD_FRAME_POINTER_REGNUM);
  setup_class_hard_regs ();
}



Vladimir Makarov committed
525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567
#define alloc_reg_class_subclasses \
  (this_target_ira_int->x_alloc_reg_class_subclasses)

/* Initialize the table of subclasses of each reg class.  */
static void
setup_reg_subclasses (void)
{
  int i, j;
  HARD_REG_SET temp_hard_regset2;

  for (i = 0; i < N_REG_CLASSES; i++)
    for (j = 0; j < N_REG_CLASSES; j++)
      alloc_reg_class_subclasses[i][j] = LIM_REG_CLASSES;

  for (i = 0; i < N_REG_CLASSES; i++)
    {
      if (i == (int) NO_REGS)
	continue;

      COPY_HARD_REG_SET (temp_hard_regset, reg_class_contents[i]);
      AND_COMPL_HARD_REG_SET (temp_hard_regset, no_unit_alloc_regs);
      if (hard_reg_set_empty_p (temp_hard_regset))
	continue;
      for (j = 0; j < N_REG_CLASSES; j++)
	if (i != j)
	  {
	    enum reg_class *p;

	    COPY_HARD_REG_SET (temp_hard_regset2, reg_class_contents[j]);
	    AND_COMPL_HARD_REG_SET (temp_hard_regset2, no_unit_alloc_regs);
	    if (! hard_reg_set_subset_p (temp_hard_regset,
					 temp_hard_regset2))
	      continue;
	    p = &alloc_reg_class_subclasses[j][0];
	    while (*p != LIM_REG_CLASSES) p++;
	    *p = (enum reg_class) i;
	  }
    }
}



/* Set up IRA_MEMORY_MOVE_COST and IRA_MAX_MEMORY_MOVE_COST.  */
Vladimir Makarov committed
568 569 570
static void
setup_class_subset_and_memory_move_costs (void)
{
Vladimir Makarov committed
571
  int cl, cl2, mode, cost;
Vladimir Makarov committed
572 573 574 575 576 577 578 579 580 581
  HARD_REG_SET temp_hard_regset2;

  for (mode = 0; mode < MAX_MACHINE_MODE; mode++)
    ira_memory_move_cost[mode][NO_REGS][0]
      = ira_memory_move_cost[mode][NO_REGS][1] = SHRT_MAX;
  for (cl = (int) N_REG_CLASSES - 1; cl >= 0; cl--)
    {
      if (cl != (int) NO_REGS)
	for (mode = 0; mode < MAX_MACHINE_MODE; mode++)
	  {
Vladimir Makarov committed
582 583
	    ira_max_memory_move_cost[mode][cl][0]
	      = ira_memory_move_cost[mode][cl][0]
584
	      = memory_move_cost ((machine_mode) mode,
585
				  (reg_class_t) cl, false);
Vladimir Makarov committed
586 587
	    ira_max_memory_move_cost[mode][cl][1]
	      = ira_memory_move_cost[mode][cl][1]
588
	      = memory_move_cost ((machine_mode) mode,
589
				  (reg_class_t) cl, true);
Vladimir Makarov committed
590 591 592 593 594
	    /* Costs for NO_REGS are used in cost calculation on the
	       1st pass when the preferred register classes are not
	       known yet.  In this case we take the best scenario.  */
	    if (ira_memory_move_cost[mode][NO_REGS][0]
		> ira_memory_move_cost[mode][cl][0])
Vladimir Makarov committed
595 596
	      ira_max_memory_move_cost[mode][NO_REGS][0]
		= ira_memory_move_cost[mode][NO_REGS][0]
Vladimir Makarov committed
597 598 599
		= ira_memory_move_cost[mode][cl][0];
	    if (ira_memory_move_cost[mode][NO_REGS][1]
		> ira_memory_move_cost[mode][cl][1])
Vladimir Makarov committed
600 601
	      ira_max_memory_move_cost[mode][NO_REGS][1]
		= ira_memory_move_cost[mode][NO_REGS][1]
Vladimir Makarov committed
602 603 604
		= ira_memory_move_cost[mode][cl][1];
	  }
    }
Vladimir Makarov committed
605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635
  for (cl = (int) N_REG_CLASSES - 1; cl >= 0; cl--)
    for (cl2 = (int) N_REG_CLASSES - 1; cl2 >= 0; cl2--)
      {
	COPY_HARD_REG_SET (temp_hard_regset, reg_class_contents[cl]);
	AND_COMPL_HARD_REG_SET (temp_hard_regset, no_unit_alloc_regs);
	COPY_HARD_REG_SET (temp_hard_regset2, reg_class_contents[cl2]);
	AND_COMPL_HARD_REG_SET (temp_hard_regset2, no_unit_alloc_regs);
	ira_class_subset_p[cl][cl2]
	  = hard_reg_set_subset_p (temp_hard_regset, temp_hard_regset2);
	if (! hard_reg_set_empty_p (temp_hard_regset2)
	    && hard_reg_set_subset_p (reg_class_contents[cl2],
				      reg_class_contents[cl]))
	  for (mode = 0; mode < MAX_MACHINE_MODE; mode++)
	    {
	      cost = ira_memory_move_cost[mode][cl2][0];
	      if (cost > ira_max_memory_move_cost[mode][cl][0])
		ira_max_memory_move_cost[mode][cl][0] = cost;
	      cost = ira_memory_move_cost[mode][cl2][1];
	      if (cost > ira_max_memory_move_cost[mode][cl][1])
		ira_max_memory_move_cost[mode][cl][1] = cost;
	    }
      }
  for (cl = (int) N_REG_CLASSES - 1; cl >= 0; cl--)
    for (mode = 0; mode < MAX_MACHINE_MODE; mode++)
      {
	ira_memory_move_cost[mode][cl][0]
	  = ira_max_memory_move_cost[mode][cl][0];
	ira_memory_move_cost[mode][cl][1]
	  = ira_max_memory_move_cost[mode][cl][1];
      }
  setup_reg_subclasses ();
Vladimir Makarov committed
636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717
}



/* Define the following macro if allocation through malloc if
   preferable.  */
#define IRA_NO_OBSTACK

#ifndef IRA_NO_OBSTACK
/* Obstack used for storing all dynamic data (except bitmaps) of the
   IRA.  */
static struct obstack ira_obstack;
#endif

/* Obstack used for storing all bitmaps of the IRA.  */
static struct bitmap_obstack ira_bitmap_obstack;

/* Allocate memory of size LEN for IRA data.  */
void *
ira_allocate (size_t len)
{
  void *res;

#ifndef IRA_NO_OBSTACK
  res = obstack_alloc (&ira_obstack, len);
#else
  res = xmalloc (len);
#endif
  return res;
}

/* Free memory ADDR allocated for IRA data.  */
void
ira_free (void *addr ATTRIBUTE_UNUSED)
{
#ifndef IRA_NO_OBSTACK
  /* do nothing */
#else
  free (addr);
#endif
}


/* Allocate and returns bitmap for IRA.  */
bitmap
ira_allocate_bitmap (void)
{
  return BITMAP_ALLOC (&ira_bitmap_obstack);
}

/* Free bitmap B allocated for IRA.  */
void
ira_free_bitmap (bitmap b ATTRIBUTE_UNUSED)
{
  /* do nothing */
}



/* Output information about allocation of all allocnos (except for
   caps) into file F.  */
void
ira_print_disposition (FILE *f)
{
  int i, n, max_regno;
  ira_allocno_t a;
  basic_block bb;

  fprintf (f, "Disposition:");
  max_regno = max_reg_num ();
  for (n = 0, i = FIRST_PSEUDO_REGISTER; i < max_regno; i++)
    for (a = ira_regno_allocno_map[i];
	 a != NULL;
	 a = ALLOCNO_NEXT_REGNO_ALLOCNO (a))
      {
	if (n % 4 == 0)
	  fprintf (f, "\n");
	n++;
	fprintf (f, " %4d:r%-4d", ALLOCNO_NUM (a), ALLOCNO_REGNO (a));
	if ((bb = ALLOCNO_LOOP_TREE_NODE (a)->bb) != NULL)
	  fprintf (f, "b%-3d", bb->index);
	else
718
	  fprintf (f, "l%-3d", ALLOCNO_LOOP_TREE_NODE (a)->loop_num);
Vladimir Makarov committed
719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736
	if (ALLOCNO_HARD_REGNO (a) >= 0)
	  fprintf (f, " %3d", ALLOCNO_HARD_REGNO (a));
	else
	  fprintf (f, " mem");
      }
  fprintf (f, "\n");
}

/* Outputs information about allocation of all allocnos into
   stderr.  */
void
ira_debug_disposition (void)
{
  ira_print_disposition (stderr);
}



Vladimir Makarov committed
737 738 739 740 741 742
/* Set up ira_stack_reg_pressure_class which is the biggest pressure
   register class containing stack registers or NO_REGS if there are
   no stack registers.  To find this class, we iterate through all
   register pressure classes and choose the first register pressure
   class containing all the stack registers and having the biggest
   size.  */
Vladimir Makarov committed
743
static void
Vladimir Makarov committed
744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787
setup_stack_reg_pressure_class (void)
{
  ira_stack_reg_pressure_class = NO_REGS;
#ifdef STACK_REGS
  {
    int i, best, size;
    enum reg_class cl;
    HARD_REG_SET temp_hard_regset2;

    CLEAR_HARD_REG_SET (temp_hard_regset);
    for (i = FIRST_STACK_REG; i <= LAST_STACK_REG; i++)
      SET_HARD_REG_BIT (temp_hard_regset, i);
    best = 0;
    for (i = 0; i < ira_pressure_classes_num; i++)
      {
	cl = ira_pressure_classes[i];
	COPY_HARD_REG_SET (temp_hard_regset2, temp_hard_regset);
	AND_HARD_REG_SET (temp_hard_regset2, reg_class_contents[cl]);
	size = hard_reg_set_size (temp_hard_regset2);
	if (best < size)
	  {
	    best = size;
	    ira_stack_reg_pressure_class = cl;
	  }
      }
  }
#endif
}

/* Find pressure classes which are register classes for which we
   calculate register pressure in IRA, register pressure sensitive
   insn scheduling, and register pressure sensitive loop invariant
   motion.

   To make register pressure calculation easy, we always use
   non-intersected register pressure classes.  A move of hard
   registers from one register pressure class is not more expensive
   than load and store of the hard registers.  Most likely an allocno
   class will be a subset of a register pressure class and in many
   cases a register pressure class.  That makes usage of register
   pressure classes a good approximation to find a high register
   pressure.  */
static void
setup_pressure_classes (void)
Vladimir Makarov committed
788
{
Vladimir Makarov committed
789 790 791 792
  int cost, i, n, curr;
  int cl, cl2;
  enum reg_class pressure_classes[N_REG_CLASSES];
  int m;
Vladimir Makarov committed
793
  HARD_REG_SET temp_hard_regset2;
Vladimir Makarov committed
794
  bool insert_p;
Vladimir Makarov committed
795

796 797 798 799 800 801
  if (targetm.compute_pressure_classes)
    n = targetm.compute_pressure_classes (pressure_classes);
  else
    { 
      n = 0;
      for (cl = 0; cl < N_REG_CLASSES; cl++)
Vladimir Makarov committed
802
	{
803 804 805 806 807 808 809 810
	  if (ira_class_hard_regs_num[cl] == 0)
	    continue;
	  if (ira_class_hard_regs_num[cl] != 1
	      /* A register class without subclasses may contain a few
		 hard registers and movement between them is costly
		 (e.g. SPARC FPCC registers).  We still should consider it
		 as a candidate for a pressure class.  */
	      && alloc_reg_class_subclasses[cl][0] < cl)
811
	    {
812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831
	      /* Check that the moves between any hard registers of the
		 current class are not more expensive for a legal mode
		 than load/store of the hard registers of the current
		 class.  Such class is a potential candidate to be a
		 register pressure class.  */
	      for (m = 0; m < NUM_MACHINE_MODES; m++)
		{
		  COPY_HARD_REG_SET (temp_hard_regset, reg_class_contents[cl]);
		  AND_COMPL_HARD_REG_SET (temp_hard_regset, no_unit_alloc_regs);
		  AND_COMPL_HARD_REG_SET (temp_hard_regset,
					  ira_prohibited_class_mode_regs[cl][m]);
		  if (hard_reg_set_empty_p (temp_hard_regset))
		    continue;
		  ira_init_register_move_cost_if_necessary ((machine_mode) m);
		  cost = ira_register_move_cost[m][cl][cl];
		  if (cost <= ira_max_memory_move_cost[m][cl][1]
		      || cost <= ira_max_memory_move_cost[m][cl][0])
		    break;
		}
	      if (m >= NUM_MACHINE_MODES)
832 833
		continue;
	    }
834 835 836 837 838 839 840 841 842 843 844 845
	  curr = 0;
	  insert_p = true;
	  COPY_HARD_REG_SET (temp_hard_regset, reg_class_contents[cl]);
	  AND_COMPL_HARD_REG_SET (temp_hard_regset, no_unit_alloc_regs);
	  /* Remove so far added pressure classes which are subset of the
	     current candidate class.  Prefer GENERAL_REGS as a pressure
	     register class to another class containing the same
	     allocatable hard registers.  We do this because machine
	     dependent cost hooks might give wrong costs for the latter
	     class but always give the right cost for the former class
	     (GENERAL_REGS).  */
	  for (i = 0; i < n; i++)
Vladimir Makarov committed
846
	    {
847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865
	      cl2 = pressure_classes[i];
	      COPY_HARD_REG_SET (temp_hard_regset2, reg_class_contents[cl2]);
	      AND_COMPL_HARD_REG_SET (temp_hard_regset2, no_unit_alloc_regs);
	      if (hard_reg_set_subset_p (temp_hard_regset, temp_hard_regset2)
		  && (! hard_reg_set_equal_p (temp_hard_regset,
					      temp_hard_regset2)
		      || cl2 == (int) GENERAL_REGS))
		{
		  pressure_classes[curr++] = (enum reg_class) cl2;
		  insert_p = false;
		  continue;
		}
	      if (hard_reg_set_subset_p (temp_hard_regset2, temp_hard_regset)
		  && (! hard_reg_set_equal_p (temp_hard_regset2,
					      temp_hard_regset)
		      || cl == (int) GENERAL_REGS))
		continue;
	      if (hard_reg_set_equal_p (temp_hard_regset2, temp_hard_regset))
		insert_p = false;
Vladimir Makarov committed
866 867
	      pressure_classes[curr++] = (enum reg_class) cl2;
	    }
868 869 870 871 872 873
	  /* If the current candidate is a subset of a so far added
	     pressure class, don't add it to the list of the pressure
	     classes.  */
	  if (insert_p)
	    pressure_classes[curr++] = (enum reg_class) cl;
	  n = curr;
Vladimir Makarov committed
874
	}
Vladimir Makarov committed
875
    }
Vladimir Makarov committed
876
#ifdef ENABLE_IRA_CHECKING
877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906
  {
    HARD_REG_SET ignore_hard_regs;

    /* Check pressure classes correctness: here we check that hard
       registers from all register pressure classes contains all hard
       registers available for the allocation.  */
    CLEAR_HARD_REG_SET (temp_hard_regset);
    CLEAR_HARD_REG_SET (temp_hard_regset2);
    COPY_HARD_REG_SET (ignore_hard_regs, no_unit_alloc_regs);
    for (cl = 0; cl < LIM_REG_CLASSES; cl++)
      {
	/* For some targets (like MIPS with MD_REGS), there are some
	   classes with hard registers available for allocation but
	   not able to hold value of any mode.  */
	for (m = 0; m < NUM_MACHINE_MODES; m++)
	  if (contains_reg_of_mode[cl][m])
	    break;
	if (m >= NUM_MACHINE_MODES)
	  {
	    IOR_HARD_REG_SET (ignore_hard_regs, reg_class_contents[cl]);
	    continue;
	  }
	for (i = 0; i < n; i++)
	  if ((int) pressure_classes[i] == cl)
	    break;
	IOR_HARD_REG_SET (temp_hard_regset2, reg_class_contents[cl]);
	if (i < n)
	  IOR_HARD_REG_SET (temp_hard_regset, reg_class_contents[cl]);
      }
    for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
Kito Cheng committed
907
      /* Some targets (like SPARC with ICC reg) have allocatable regs
908 909 910 911 912 913 914
	 for which no reg class is defined.  */
      if (REGNO_REG_CLASS (i) == NO_REGS)
	SET_HARD_REG_BIT (ignore_hard_regs, i);
    AND_COMPL_HARD_REG_SET (temp_hard_regset, ignore_hard_regs);
    AND_COMPL_HARD_REG_SET (temp_hard_regset2, ignore_hard_regs);
    ira_assert (hard_reg_set_subset_p (temp_hard_regset2, temp_hard_regset));
  }
Vladimir Makarov committed
915 916 917 918 919 920 921 922 923
#endif
  ira_pressure_classes_num = 0;
  for (i = 0; i < n; i++)
    {
      cl = (int) pressure_classes[i];
      ira_reg_pressure_class_p[cl] = true;
      ira_pressure_classes[ira_pressure_classes_num++] = (enum reg_class) cl;
    }
  setup_stack_reg_pressure_class ();
Vladimir Makarov committed
924 925
}

926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951
/* Set up IRA_UNIFORM_CLASS_P.  Uniform class is a register class
   whose register move cost between any registers of the class is the
   same as for all its subclasses.  We use the data to speed up the
   2nd pass of calculations of allocno costs.  */
static void
setup_uniform_class_p (void)
{
  int i, cl, cl2, m;

  for (cl = 0; cl < N_REG_CLASSES; cl++)
    {
      ira_uniform_class_p[cl] = false;
      if (ira_class_hard_regs_num[cl] == 0)
	continue;
      /* We can not use alloc_reg_class_subclasses here because move
	 cost hooks does not take into account that some registers are
	 unavailable for the subtarget.  E.g. for i686, INT_SSE_REGS
	 is element of alloc_reg_class_subclasses for GENERAL_REGS
	 because SSE regs are unavailable.  */
      for (i = 0; (cl2 = reg_class_subclasses[cl][i]) != LIM_REG_CLASSES; i++)
	{
	  if (ira_class_hard_regs_num[cl2] == 0)
	    continue;
      	  for (m = 0; m < NUM_MACHINE_MODES; m++)
	    if (contains_reg_of_mode[cl][m] && contains_reg_of_mode[cl2][m])
	      {
952
		ira_init_register_move_cost_if_necessary ((machine_mode) m);
953 954 955 956 957 958 959 960 961 962 963 964
		if (ira_register_move_cost[m][cl][cl]
		    != ira_register_move_cost[m][cl2][cl2])
		  break;
	      }
	  if (m < NUM_MACHINE_MODES)
	    break;
	}
      if (cl2 == LIM_REG_CLASSES)
	ira_uniform_class_p[cl] = true;
    }
}

Vladimir Makarov committed
965 966 967
/* Set up IRA_ALLOCNO_CLASSES, IRA_ALLOCNO_CLASSES_NUM,
   IRA_IMPORTANT_CLASSES, and IRA_IMPORTANT_CLASSES_NUM.

Kito Cheng committed
968
   Target may have many subtargets and not all target hard registers can
Vladimir Makarov committed
969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989
   be used for allocation, e.g. x86 port in 32-bit mode can not use
   hard registers introduced in x86-64 like r8-r15).  Some classes
   might have the same allocatable hard registers, e.g.  INDEX_REGS
   and GENERAL_REGS in x86 port in 32-bit mode.  To decrease different
   calculations efforts we introduce allocno classes which contain
   unique non-empty sets of allocatable hard-registers.

   Pseudo class cost calculation in ira-costs.c is very expensive.
   Therefore we are trying to decrease number of classes involved in
   such calculation.  Register classes used in the cost calculation
   are called important classes.  They are allocno classes and other
   non-empty classes whose allocatable hard register sets are inside
   of an allocno class hard register set.  From the first sight, it
   looks like that they are just allocno classes.  It is not true.  In
   example of x86-port in 32-bit mode, allocno classes will contain
   GENERAL_REGS but not LEGACY_REGS (because allocatable hard
   registers are the same for the both classes).  The important
   classes will contain GENERAL_REGS and LEGACY_REGS.  It is done
   because a machine description insn constraint may refers for
   LEGACY_REGS and code in ira-costs.c is mostly base on investigation
   of the insn constraints.  */
Vladimir Makarov committed
990
static void
Vladimir Makarov committed
991
setup_allocno_and_important_classes (void)
Vladimir Makarov committed
992
{
993
  int i, j, n, cl;
994
  bool set_p;
Vladimir Makarov committed
995
  HARD_REG_SET temp_hard_regset2;
996 997
  static enum reg_class classes[LIM_REG_CLASSES + 1];

Vladimir Makarov committed
998 999 1000 1001
  n = 0;
  /* Collect classes which contain unique sets of allocatable hard
     registers.  Prefer GENERAL_REGS to other classes containing the
     same set of hard registers.  */
1002
  for (i = 0; i < LIM_REG_CLASSES; i++)
1003
    {
Vladimir Makarov committed
1004 1005 1006
      COPY_HARD_REG_SET (temp_hard_regset, reg_class_contents[i]);
      AND_COMPL_HARD_REG_SET (temp_hard_regset, no_unit_alloc_regs);
      for (j = 0; j < n; j++)
1007
	{
Vladimir Makarov committed
1008 1009 1010 1011 1012 1013 1014
	  cl = classes[j];
	  COPY_HARD_REG_SET (temp_hard_regset2, reg_class_contents[cl]);
	  AND_COMPL_HARD_REG_SET (temp_hard_regset2,
				  no_unit_alloc_regs);
	  if (hard_reg_set_equal_p (temp_hard_regset,
				    temp_hard_regset2))
	    break;
1015
	}
1016
      if (j >= n || targetm.additional_allocno_class_p (i))
Vladimir Makarov committed
1017 1018 1019 1020 1021 1022 1023
	classes[n++] = (enum reg_class) i;
      else if (i == GENERAL_REGS)
	/* Prefer general regs.  For i386 example, it means that
	   we prefer GENERAL_REGS over INDEX_REGS or LEGACY_REGS
	   (all of them consists of the same available hard
	   registers).  */
	classes[j] = (enum reg_class) i;
1024
    }
Vladimir Makarov committed
1025
  classes[n] = LIM_REG_CLASSES;
Vladimir Makarov committed
1026

Vladimir Makarov committed
1027
  /* Set up classes which can be used for allocnos as classes
Kito Cheng committed
1028
     containing non-empty unique sets of allocatable hard
Vladimir Makarov committed
1029 1030
     registers.  */
  ira_allocno_classes_num = 0;
Vladimir Makarov committed
1031
  for (i = 0; (cl = classes[i]) != LIM_REG_CLASSES; i++)
1032
    if (ira_class_hard_regs_num[cl] > 0)
Vladimir Makarov committed
1033
      ira_allocno_classes[ira_allocno_classes_num++] = (enum reg_class) cl;
Vladimir Makarov committed
1034
  ira_important_classes_num = 0;
Vladimir Makarov committed
1035 1036
  /* Add non-allocno classes containing to non-empty set of
     allocatable hard regs.  */
Vladimir Makarov committed
1037
  for (cl = 0; cl < N_REG_CLASSES; cl++)
1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057
    if (ira_class_hard_regs_num[cl] > 0)
      {
	COPY_HARD_REG_SET (temp_hard_regset, reg_class_contents[cl]);
	AND_COMPL_HARD_REG_SET (temp_hard_regset, no_unit_alloc_regs);
	set_p = false;
	for (j = 0; j < ira_allocno_classes_num; j++)
	  {
	    COPY_HARD_REG_SET (temp_hard_regset2,
			       reg_class_contents[ira_allocno_classes[j]]);
	    AND_COMPL_HARD_REG_SET (temp_hard_regset2, no_unit_alloc_regs);
	    if ((enum reg_class) cl == ira_allocno_classes[j])
	      break;
	    else if (hard_reg_set_subset_p (temp_hard_regset,
					    temp_hard_regset2))
	      set_p = true;
	  }
	if (set_p && j >= ira_allocno_classes_num)
	  ira_important_classes[ira_important_classes_num++]
	    = (enum reg_class) cl;
      }
Vladimir Makarov committed
1058 1059
  /* Now add allocno classes to the important classes.  */
  for (j = 0; j < ira_allocno_classes_num; j++)
1060
    ira_important_classes[ira_important_classes_num++]
Vladimir Makarov committed
1061 1062 1063 1064 1065 1066 1067 1068 1069
      = ira_allocno_classes[j];
  for (cl = 0; cl < N_REG_CLASSES; cl++)
    {
      ira_reg_allocno_class_p[cl] = false;
      ira_reg_pressure_class_p[cl] = false;
    }
  for (j = 0; j < ira_allocno_classes_num; j++)
    ira_reg_allocno_class_p[ira_allocno_classes[j]] = true;
  setup_pressure_classes ();
1070
  setup_uniform_class_p ();
Vladimir Makarov committed
1071 1072
}

Vladimir Makarov committed
1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086
/* Setup translation in CLASS_TRANSLATE of all classes into a class
   given by array CLASSES of length CLASSES_NUM.  The function is used
   make translation any reg class to an allocno class or to an
   pressure class.  This translation is necessary for some
   calculations when we can use only allocno or pressure classes and
   such translation represents an approximate representation of all
   classes.

   The translation in case when allocatable hard register set of a
   given class is subset of allocatable hard register set of a class
   in CLASSES is pretty simple.  We use smallest classes from CLASSES
   containing a given class.  If allocatable hard register set of a
   given class is not a subset of any corresponding set of a class
   from CLASSES, we use the cheapest (with load/store point of view)
1087
   class from CLASSES whose set intersects with given class set.  */
Vladimir Makarov committed
1088
static void
Vladimir Makarov committed
1089 1090
setup_class_translate_array (enum reg_class *class_translate,
			     int classes_num, enum reg_class *classes)
Vladimir Makarov committed
1091
{
1092
  int cl, mode;
Vladimir Makarov committed
1093
  enum reg_class aclass, best_class, *cl_ptr;
Vladimir Makarov committed
1094 1095 1096
  int i, cost, min_cost, best_cost;

  for (cl = 0; cl < N_REG_CLASSES; cl++)
Vladimir Makarov committed
1097
    class_translate[cl] = NO_REGS;
H.J. Lu committed
1098

Vladimir Makarov committed
1099
  for (i = 0; i < classes_num; i++)
Vladimir Makarov committed
1100
    {
Vladimir Makarov committed
1101 1102 1103 1104 1105 1106 1107
      aclass = classes[i];
      for (cl_ptr = &alloc_reg_class_subclasses[aclass][0];
	   (cl = *cl_ptr) != LIM_REG_CLASSES;
	   cl_ptr++)
	if (class_translate[cl] == NO_REGS)
	  class_translate[cl] = aclass;
      class_translate[aclass] = aclass;
Vladimir Makarov committed
1108
    }
Vladimir Makarov committed
1109 1110 1111
  /* For classes which are not fully covered by one of given classes
     (in other words covered by more one given class), use the
     cheapest class.  */
Vladimir Makarov committed
1112 1113
  for (cl = 0; cl < N_REG_CLASSES; cl++)
    {
Vladimir Makarov committed
1114
      if (cl == NO_REGS || class_translate[cl] != NO_REGS)
Vladimir Makarov committed
1115 1116 1117
	continue;
      best_class = NO_REGS;
      best_cost = INT_MAX;
Vladimir Makarov committed
1118
      for (i = 0; i < classes_num; i++)
Vladimir Makarov committed
1119
	{
Vladimir Makarov committed
1120
	  aclass = classes[i];
Vladimir Makarov committed
1121
	  COPY_HARD_REG_SET (temp_hard_regset,
Vladimir Makarov committed
1122
			     reg_class_contents[aclass]);
Vladimir Makarov committed
1123 1124
	  AND_HARD_REG_SET (temp_hard_regset, reg_class_contents[cl]);
	  AND_COMPL_HARD_REG_SET (temp_hard_regset, no_unit_alloc_regs);
1125
	  if (! hard_reg_set_empty_p (temp_hard_regset))
Vladimir Makarov committed
1126 1127 1128 1129
	    {
	      min_cost = INT_MAX;
	      for (mode = 0; mode < MAX_MACHINE_MODE; mode++)
		{
1130 1131
		  cost = (ira_memory_move_cost[mode][aclass][0]
			  + ira_memory_move_cost[mode][aclass][1]);
Vladimir Makarov committed
1132 1133 1134 1135 1136
		  if (min_cost > cost)
		    min_cost = cost;
		}
	      if (best_class == NO_REGS || best_cost > min_cost)
		{
Vladimir Makarov committed
1137
		  best_class = aclass;
Vladimir Makarov committed
1138 1139 1140 1141
		  best_cost = min_cost;
		}
	    }
	}
Vladimir Makarov committed
1142
      class_translate[cl] = best_class;
Vladimir Makarov committed
1143 1144 1145
    }
}

Vladimir Makarov committed
1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159
/* Set up array IRA_ALLOCNO_CLASS_TRANSLATE and
   IRA_PRESSURE_CLASS_TRANSLATE.  */
static void
setup_class_translate (void)
{
  setup_class_translate_array (ira_allocno_class_translate,
			       ira_allocno_classes_num, ira_allocno_classes);
  setup_class_translate_array (ira_pressure_class_translate,
			       ira_pressure_classes_num, ira_pressure_classes);
}

/* Order numbers of allocno classes in original target allocno class
   array, -1 for non-allocno classes.  */
static int allocno_class_order[N_REG_CLASSES];
1160 1161 1162 1163 1164 1165 1166

/* The function used to sort the important classes.  */
static int
comp_reg_classes_func (const void *v1p, const void *v2p)
{
  enum reg_class cl1 = *(const enum reg_class *) v1p;
  enum reg_class cl2 = *(const enum reg_class *) v2p;
Vladimir Makarov committed
1167
  enum reg_class tcl1, tcl2;
1168 1169
  int diff;

Vladimir Makarov committed
1170 1171 1172 1173
  tcl1 = ira_allocno_class_translate[cl1];
  tcl2 = ira_allocno_class_translate[cl2];
  if (tcl1 != NO_REGS && tcl2 != NO_REGS
      && (diff = allocno_class_order[tcl1] - allocno_class_order[tcl2]) != 0)
1174 1175 1176 1177
    return diff;
  return (int) cl1 - (int) cl2;
}

Vladimir Makarov committed
1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188
/* For correct work of function setup_reg_class_relation we need to
   reorder important classes according to the order of their allocno
   classes.  It places important classes containing the same
   allocatable hard register set adjacent to each other and allocno
   class with the allocatable hard register set right after the other
   important classes with the same set.

   In example from comments of function
   setup_allocno_and_important_classes, it places LEGACY_REGS and
   GENERAL_REGS close to each other and GENERAL_REGS is after
   LEGACY_REGS.  */
1189 1190 1191 1192 1193 1194
static void
reorder_important_classes (void)
{
  int i;

  for (i = 0; i < N_REG_CLASSES; i++)
Vladimir Makarov committed
1195 1196 1197
    allocno_class_order[i] = -1;
  for (i = 0; i < ira_allocno_classes_num; i++)
    allocno_class_order[ira_allocno_classes[i]] = i;
1198 1199
  qsort (ira_important_classes, ira_important_classes_num,
	 sizeof (enum reg_class), comp_reg_classes_func);
Vladimir Makarov committed
1200 1201
  for (i = 0; i < ira_important_classes_num; i++)
    ira_important_class_nums[ira_important_classes[i]] = i;
1202 1203
}

Vladimir Makarov committed
1204 1205 1206 1207
/* Set up IRA_REG_CLASS_SUBUNION, IRA_REG_CLASS_SUPERUNION,
   IRA_REG_CLASS_SUPER_CLASSES, IRA_REG_CLASSES_INTERSECT, and
   IRA_REG_CLASSES_INTERSECT_P.  For the meaning of the relations,
   please see corresponding comments in ira-int.h.  */
Vladimir Makarov committed
1208
static void
1209
setup_reg_class_relations (void)
Vladimir Makarov committed
1210 1211 1212
{
  int i, cl1, cl2, cl3;
  HARD_REG_SET intersection_set, union_set, temp_set2;
1213
  bool important_class_p[N_REG_CLASSES];
Vladimir Makarov committed
1214

1215 1216 1217
  memset (important_class_p, 0, sizeof (important_class_p));
  for (i = 0; i < ira_important_classes_num; i++)
    important_class_p[ira_important_classes[i]] = true;
Vladimir Makarov committed
1218 1219
  for (cl1 = 0; cl1 < N_REG_CLASSES; cl1++)
    {
1220
      ira_reg_class_super_classes[cl1][0] = LIM_REG_CLASSES;
Vladimir Makarov committed
1221 1222
      for (cl2 = 0; cl2 < N_REG_CLASSES; cl2++)
	{
1223
	  ira_reg_classes_intersect_p[cl1][cl2] = false;
Vladimir Makarov committed
1224
	  ira_reg_class_intersect[cl1][cl2] = NO_REGS;
1225
	  ira_reg_class_subset[cl1][cl2] = NO_REGS;
Vladimir Makarov committed
1226 1227 1228 1229
	  COPY_HARD_REG_SET (temp_hard_regset, reg_class_contents[cl1]);
	  AND_COMPL_HARD_REG_SET (temp_hard_regset, no_unit_alloc_regs);
	  COPY_HARD_REG_SET (temp_set2, reg_class_contents[cl2]);
	  AND_COMPL_HARD_REG_SET (temp_set2, no_unit_alloc_regs);
1230 1231
	  if (hard_reg_set_empty_p (temp_hard_regset)
	      && hard_reg_set_empty_p (temp_set2))
Vladimir Makarov committed
1232
	    {
Vladimir Makarov committed
1233 1234 1235
	      /* The both classes have no allocatable hard registers
		 -- take all class hard registers into account and use
		 reg_class_subunion and reg_class_superunion.  */
Vladimir Makarov committed
1236 1237 1238 1239 1240 1241
	      for (i = 0;; i++)
		{
		  cl3 = reg_class_subclasses[cl1][i];
		  if (cl3 == LIM_REG_CLASSES)
		    break;
		  if (reg_class_subset_p (ira_reg_class_intersect[cl1][cl2],
1242 1243
					  (enum reg_class) cl3))
		    ira_reg_class_intersect[cl1][cl2] = (enum reg_class) cl3;
Vladimir Makarov committed
1244
		}
Vladimir Makarov committed
1245 1246
	      ira_reg_class_subunion[cl1][cl2] = reg_class_subunion[cl1][cl2];
	      ira_reg_class_superunion[cl1][cl2] = reg_class_superunion[cl1][cl2];
Vladimir Makarov committed
1247 1248
	      continue;
	    }
1249 1250 1251 1252 1253
	  ira_reg_classes_intersect_p[cl1][cl2]
	    = hard_reg_set_intersect_p (temp_hard_regset, temp_set2);
	  if (important_class_p[cl1] && important_class_p[cl2]
	      && hard_reg_set_subset_p (temp_hard_regset, temp_set2))
	    {
Vladimir Makarov committed
1254 1255 1256
	      /* CL1 and CL2 are important classes and CL1 allocatable
		 hard register set is inside of CL2 allocatable hard
		 registers -- make CL1 a superset of CL2.  */
1257 1258 1259 1260 1261 1262 1263 1264
	      enum reg_class *p;

	      p = &ira_reg_class_super_classes[cl1][0];
	      while (*p != LIM_REG_CLASSES)
		p++;
	      *p++ = (enum reg_class) cl2;
	      *p = LIM_REG_CLASSES;
	    }
Vladimir Makarov committed
1265 1266
	  ira_reg_class_subunion[cl1][cl2] = NO_REGS;
	  ira_reg_class_superunion[cl1][cl2] = NO_REGS;
Vladimir Makarov committed
1267 1268 1269 1270 1271 1272
	  COPY_HARD_REG_SET (intersection_set, reg_class_contents[cl1]);
	  AND_HARD_REG_SET (intersection_set, reg_class_contents[cl2]);
	  AND_COMPL_HARD_REG_SET (intersection_set, no_unit_alloc_regs);
	  COPY_HARD_REG_SET (union_set, reg_class_contents[cl1]);
	  IOR_HARD_REG_SET (union_set, reg_class_contents[cl2]);
	  AND_COMPL_HARD_REG_SET (union_set, no_unit_alloc_regs);
1273
	  for (cl3 = 0; cl3 < N_REG_CLASSES; cl3++)
Vladimir Makarov committed
1274 1275 1276 1277 1278
	    {
	      COPY_HARD_REG_SET (temp_hard_regset, reg_class_contents[cl3]);
	      AND_COMPL_HARD_REG_SET (temp_hard_regset, no_unit_alloc_regs);
	      if (hard_reg_set_subset_p (temp_hard_regset, intersection_set))
		{
Vladimir Makarov committed
1279 1280 1281
		  /* CL3 allocatable hard register set is inside of
		     intersection of allocatable hard register sets
		     of CL1 and CL2.  */
1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304
		  if (important_class_p[cl3])
		    {
		      COPY_HARD_REG_SET
			(temp_set2,
			 reg_class_contents
			 [(int) ira_reg_class_intersect[cl1][cl2]]);
		      AND_COMPL_HARD_REG_SET (temp_set2, no_unit_alloc_regs);
		      if (! hard_reg_set_subset_p (temp_hard_regset, temp_set2)
			  /* If the allocatable hard register sets are
			     the same, prefer GENERAL_REGS or the
			     smallest class for debugging
			     purposes.  */
			  || (hard_reg_set_equal_p (temp_hard_regset, temp_set2)
			      && (cl3 == GENERAL_REGS
				  || ((ira_reg_class_intersect[cl1][cl2]
				       != GENERAL_REGS)
				      && hard_reg_set_subset_p
				         (reg_class_contents[cl3],
					  reg_class_contents
					  [(int)
					   ira_reg_class_intersect[cl1][cl2]])))))
			ira_reg_class_intersect[cl1][cl2] = (enum reg_class) cl3;
		    }
Vladimir Makarov committed
1305 1306
		  COPY_HARD_REG_SET
		    (temp_set2,
1307
		     reg_class_contents[(int) ira_reg_class_subset[cl1][cl2]]);
Vladimir Makarov committed
1308
		  AND_COMPL_HARD_REG_SET (temp_set2, no_unit_alloc_regs);
1309 1310 1311
		  if (! hard_reg_set_subset_p (temp_hard_regset, temp_set2)
		      /* Ignore unavailable hard registers and prefer
			 smallest class for debugging purposes.  */
Vladimir Makarov committed
1312
		      || (hard_reg_set_equal_p (temp_hard_regset, temp_set2)
1313 1314 1315 1316 1317
			  && hard_reg_set_subset_p
			     (reg_class_contents[cl3],
			      reg_class_contents
			      [(int) ira_reg_class_subset[cl1][cl2]])))
		    ira_reg_class_subset[cl1][cl2] = (enum reg_class) cl3;
Vladimir Makarov committed
1318
		}
1319 1320
	      if (important_class_p[cl3]
		  && hard_reg_set_subset_p (temp_hard_regset, union_set))
Vladimir Makarov committed
1321
		{
Kito Cheng committed
1322
		  /* CL3 allocatable hard register set is inside of
Vladimir Makarov committed
1323 1324
		     union of allocatable hard register sets of CL1
		     and CL2.  */
Vladimir Makarov committed
1325 1326
		  COPY_HARD_REG_SET
		    (temp_set2,
Vladimir Makarov committed
1327
		     reg_class_contents[(int) ira_reg_class_subunion[cl1][cl2]]);
Vladimir Makarov committed
1328
		  AND_COMPL_HARD_REG_SET (temp_set2, no_unit_alloc_regs);
Vladimir Makarov committed
1329
	 	  if (ira_reg_class_subunion[cl1][cl2] == NO_REGS
Vladimir Makarov committed
1330
		      || (hard_reg_set_subset_p (temp_set2, temp_hard_regset)
Vladimir Makarov committed
1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355
			  
			  && (! hard_reg_set_equal_p (temp_set2,
						      temp_hard_regset)
			      || cl3 == GENERAL_REGS
			      /* If the allocatable hard register sets are the
				 same, prefer GENERAL_REGS or the smallest
				 class for debugging purposes.  */
			      || (ira_reg_class_subunion[cl1][cl2] != GENERAL_REGS
				  && hard_reg_set_subset_p
				     (reg_class_contents[cl3],
				      reg_class_contents
				      [(int) ira_reg_class_subunion[cl1][cl2]])))))
		    ira_reg_class_subunion[cl1][cl2] = (enum reg_class) cl3;
		}
	      if (hard_reg_set_subset_p (union_set, temp_hard_regset))
		{
		  /* CL3 allocatable hard register set contains union
		     of allocatable hard register sets of CL1 and
		     CL2.  */
		  COPY_HARD_REG_SET
		    (temp_set2,
		     reg_class_contents[(int) ira_reg_class_superunion[cl1][cl2]]);
		  AND_COMPL_HARD_REG_SET (temp_set2, no_unit_alloc_regs);
	 	  if (ira_reg_class_superunion[cl1][cl2] == NO_REGS
		      || (hard_reg_set_subset_p (temp_hard_regset, temp_set2)
H.J. Lu committed
1356

Vladimir Makarov committed
1357 1358
			  && (! hard_reg_set_equal_p (temp_set2,
						      temp_hard_regset)
Vladimir Makarov committed
1359 1360 1361 1362 1363 1364 1365 1366 1367 1368
			      || cl3 == GENERAL_REGS
			      /* If the allocatable hard register sets are the
				 same, prefer GENERAL_REGS or the smallest
				 class for debugging purposes.  */
			      || (ira_reg_class_superunion[cl1][cl2] != GENERAL_REGS
				  && hard_reg_set_subset_p
				     (reg_class_contents[cl3],
				      reg_class_contents
				      [(int) ira_reg_class_superunion[cl1][cl2]])))))
		    ira_reg_class_superunion[cl1][cl2] = (enum reg_class) cl3;
Vladimir Makarov committed
1369 1370 1371 1372 1373 1374
		}
	    }
	}
    }
}

Kito Cheng committed
1375
/* Output all uniform and important classes into file F.  */
1376
static void
1377
print_uniform_and_important_classes (FILE *f)
1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392
{
  int i, cl;

  fprintf (f, "Uniform classes:\n");
  for (cl = 0; cl < N_REG_CLASSES; cl++)
    if (ira_uniform_class_p[cl])
      fprintf (f, " %s", reg_class_names[cl]);
  fprintf (f, "\nImportant classes:\n");
  for (i = 0; i < ira_important_classes_num; i++)
    fprintf (f, " %s", reg_class_names[ira_important_classes[i]]);
  fprintf (f, "\n");
}

/* Output all possible allocno or pressure classes and their
   translation map into file F.  */
Vladimir Makarov committed
1393
static void
1394
print_translated_classes (FILE *f, bool pressure_p)
Vladimir Makarov committed
1395 1396 1397 1398 1399 1400 1401 1402
{
  int classes_num = (pressure_p
		     ? ira_pressure_classes_num : ira_allocno_classes_num);
  enum reg_class *classes = (pressure_p
			     ? ira_pressure_classes : ira_allocno_classes);
  enum reg_class *class_translate = (pressure_p
				     ? ira_pressure_class_translate
				     : ira_allocno_class_translate);
Vladimir Makarov committed
1403 1404
  int i;

Vladimir Makarov committed
1405 1406 1407
  fprintf (f, "%s classes:\n", pressure_p ? "Pressure" : "Allocno");
  for (i = 0; i < classes_num; i++)
    fprintf (f, " %s", reg_class_names[classes[i]]);
Vladimir Makarov committed
1408 1409 1410
  fprintf (f, "\nClass translation:\n");
  for (i = 0; i < N_REG_CLASSES; i++)
    fprintf (f, " %s -> %s\n", reg_class_names[i],
Vladimir Makarov committed
1411
	     reg_class_names[class_translate[i]]);
Vladimir Makarov committed
1412 1413
}

Vladimir Makarov committed
1414 1415
/* Output all possible allocno and translation classes and the
   translation maps into stderr.  */
Vladimir Makarov committed
1416
void
Vladimir Makarov committed
1417
ira_debug_allocno_classes (void)
Vladimir Makarov committed
1418
{
1419
  print_uniform_and_important_classes (stderr);
1420 1421
  print_translated_classes (stderr, false);
  print_translated_classes (stderr, true);
Vladimir Makarov committed
1422 1423
}

Vladimir Makarov committed
1424
/* Set up different arrays concerning class subsets, allocno and
Vladimir Makarov committed
1425 1426
   important classes.  */
static void
Vladimir Makarov committed
1427
find_reg_classes (void)
Vladimir Makarov committed
1428
{
Vladimir Makarov committed
1429
  setup_allocno_and_important_classes ();
1430
  setup_class_translate ();
1431
  reorder_important_classes ();
1432
  setup_reg_class_relations ();
Vladimir Makarov committed
1433 1434 1435 1436
}



1437 1438
/* Set up the array above.  */
static void
Vladimir Makarov committed
1439
setup_hard_regno_aclass (void)
1440
{
1441
  int i;
1442 1443 1444

  for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
    {
Vladimir Makarov committed
1445 1446
#if 1
      ira_hard_regno_allocno_class[i]
1447 1448
	= (TEST_HARD_REG_BIT (no_unit_alloc_regs, i)
	   ? NO_REGS
Vladimir Makarov committed
1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463
	   : ira_allocno_class_translate[REGNO_REG_CLASS (i)]);
#else
      int j;
      enum reg_class cl;
      ira_hard_regno_allocno_class[i] = NO_REGS;
      for (j = 0; j < ira_allocno_classes_num; j++)
 	{
	  cl = ira_allocno_classes[j];
 	  if (ira_class_hard_reg_index[cl][i] >= 0)
 	    {
	      ira_hard_regno_allocno_class[i] = cl;
 	      break;
 	    }
 	}
#endif
1464 1465 1466 1467 1468
    }
}



Vladimir Makarov committed
1469
/* Form IRA_REG_CLASS_MAX_NREGS and IRA_REG_CLASS_MIN_NREGS maps.  */
Vladimir Makarov committed
1470 1471 1472
static void
setup_reg_class_nregs (void)
{
Vladimir Makarov committed
1473
  int i, cl, cl2, m;
Vladimir Makarov committed
1474

Vladimir Makarov committed
1475 1476 1477 1478 1479
  for (m = 0; m < MAX_MACHINE_MODE; m++)
    {
      for (cl = 0; cl < N_REG_CLASSES; cl++)
	ira_reg_class_max_nregs[cl][m]
	  = ira_reg_class_min_nregs[cl][m]
1480
	  = targetm.class_max_nregs ((reg_class_t) cl, (machine_mode) m);
Vladimir Makarov committed
1481 1482 1483 1484 1485 1486 1487 1488
      for (cl = 0; cl < N_REG_CLASSES; cl++)
	for (i = 0;
	     (cl2 = alloc_reg_class_subclasses[cl][i]) != LIM_REG_CLASSES;
	     i++)
	  if (ira_reg_class_min_nregs[cl2][m]
	      < ira_reg_class_min_nregs[cl][m])
	    ira_reg_class_min_nregs[cl][m] = ira_reg_class_min_nregs[cl2][m];
    }
Vladimir Makarov committed
1489 1490 1491 1492
}



1493 1494
/* Set up IRA_PROHIBITED_CLASS_MODE_REGS and IRA_CLASS_SINGLETON.
   This function is called once IRA_CLASS_HARD_REGS has been initialized.  */
Vladimir Makarov committed
1495 1496 1497
static void
setup_prohibited_class_mode_regs (void)
{
1498
  int j, k, hard_regno, cl, last_hard_regno, count;
Vladimir Makarov committed
1499

Vladimir Makarov committed
1500
  for (cl = (int) N_REG_CLASSES - 1; cl >= 0; cl--)
Vladimir Makarov committed
1501
    {
1502 1503
      COPY_HARD_REG_SET (temp_hard_regset, reg_class_contents[cl]);
      AND_COMPL_HARD_REG_SET (temp_hard_regset, no_unit_alloc_regs);
Vladimir Makarov committed
1504 1505
      for (j = 0; j < NUM_MACHINE_MODES; j++)
	{
1506 1507
	  count = 0;
	  last_hard_regno = -1;
Vladimir Makarov committed
1508
	  CLEAR_HARD_REG_SET (ira_prohibited_class_mode_regs[cl][j]);
Vladimir Makarov committed
1509 1510 1511
	  for (k = ira_class_hard_regs_num[cl] - 1; k >= 0; k--)
	    {
	      hard_regno = ira_class_hard_regs[cl][k];
1512
	      if (!targetm.hard_regno_mode_ok (hard_regno, (machine_mode) j))
Vladimir Makarov committed
1513
		SET_HARD_REG_BIT (ira_prohibited_class_mode_regs[cl][j],
Vladimir Makarov committed
1514
				  hard_regno);
1515
	      else if (in_hard_reg_set_p (temp_hard_regset,
1516
					  (machine_mode) j, hard_regno))
1517 1518 1519 1520
		{
		  last_hard_regno = hard_regno;
		  count++;
		}
Vladimir Makarov committed
1521
	    }
1522
	  ira_class_singleton[cl][j] = (count == 1 ? last_hard_regno : -1);
Vladimir Makarov committed
1523 1524 1525 1526
	}
    }
}

Vladimir Makarov committed
1527 1528 1529 1530 1531 1532 1533 1534 1535 1536
/* Clarify IRA_PROHIBITED_CLASS_MODE_REGS by excluding hard registers
   spanning from one register pressure class to another one.  It is
   called after defining the pressure classes.  */
static void
clarify_prohibited_class_mode_regs (void)
{
  int j, k, hard_regno, cl, pclass, nregs;

  for (cl = (int) N_REG_CLASSES - 1; cl >= 0; cl--)
    for (j = 0; j < NUM_MACHINE_MODES; j++)
1537 1538 1539 1540 1541 1542 1543
      {
	CLEAR_HARD_REG_SET (ira_useful_class_mode_regs[cl][j]);
	for (k = ira_class_hard_regs_num[cl] - 1; k >= 0; k--)
	  {
	    hard_regno = ira_class_hard_regs[cl][k];
	    if (TEST_HARD_REG_BIT (ira_prohibited_class_mode_regs[cl][j], hard_regno))
	      continue;
1544
	    nregs = hard_regno_nregs (hard_regno, (machine_mode) j);
1545
	    if (hard_regno + nregs > FIRST_PSEUDO_REGISTER)
Vladimir Makarov committed
1546 1547 1548
	      {
		SET_HARD_REG_BIT (ira_prohibited_class_mode_regs[cl][j],
				  hard_regno);
1549
		 continue;
Vladimir Makarov committed
1550
	      }
1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563
	    pclass = ira_pressure_class_translate[REGNO_REG_CLASS (hard_regno)];
	    for (nregs-- ;nregs >= 0; nregs--)
	      if (((enum reg_class) pclass
		   != ira_pressure_class_translate[REGNO_REG_CLASS
						   (hard_regno + nregs)]))
		{
		  SET_HARD_REG_BIT (ira_prohibited_class_mode_regs[cl][j],
				    hard_regno);
		  break;
		}
	    if (!TEST_HARD_REG_BIT (ira_prohibited_class_mode_regs[cl][j],
				    hard_regno))
	      add_to_hard_reg_set (&ira_useful_class_mode_regs[cl][j],
1564
				   (machine_mode) j, hard_regno);
1565 1566
	  }
      }
Vladimir Makarov committed
1567
}
Vladimir Makarov committed
1568

1569 1570 1571
/* Allocate and initialize IRA_REGISTER_MOVE_COST, IRA_MAY_MOVE_IN_COST
   and IRA_MAY_MOVE_OUT_COST for MODE.  */
void
1572
ira_init_register_move_cost (machine_mode mode)
1573 1574 1575
{
  static unsigned short last_move_cost[N_REG_CLASSES][N_REG_CLASSES];
  bool all_match = true;
1576 1577
  unsigned int i, cl1, cl2;
  HARD_REG_SET ok_regs;
1578

1579 1580 1581
  ira_assert (ira_register_move_cost[mode] == NULL
	      && ira_may_move_in_cost[mode] == NULL
	      && ira_may_move_out_cost[mode] == NULL);
1582 1583 1584 1585 1586
  CLEAR_HARD_REG_SET (ok_regs);
  for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
    if (targetm.hard_regno_mode_ok (i, mode))
      SET_HARD_REG_BIT (ok_regs, i);

1587 1588 1589 1590
  /* Note that we might be asked about the move costs of modes that
     cannot be stored in any hard register, for example if an inline
     asm tries to create a register operand with an impossible mode.
     We therefore can't assert have_regs_of_mode[mode] here.  */
1591
  for (cl1 = 0; cl1 < N_REG_CLASSES; cl1++)
1592 1593 1594
    for (cl2 = 0; cl2 < N_REG_CLASSES; cl2++)
      {
	int cost;
1595 1596
	if (!hard_reg_set_intersect_p (ok_regs, reg_class_contents[cl1])
	    || !hard_reg_set_intersect_p (ok_regs, reg_class_contents[cl2]))
1597 1598 1599 1600 1601 1602 1603 1604
	  {
	    if ((ira_reg_class_max_nregs[cl1][mode]
		 > ira_class_hard_regs_num[cl1])
		|| (ira_reg_class_max_nregs[cl2][mode]
		    > ira_class_hard_regs_num[cl2]))
	      cost = 65535;
	    else
	      cost = (ira_memory_move_cost[mode][cl1][0]
1605
		      + ira_memory_move_cost[mode][cl2][1]) * 2;
1606 1607 1608 1609 1610 1611 1612 1613 1614 1615
	  }
	else
	  {
	    cost = register_move_cost (mode, (enum reg_class) cl1,
				       (enum reg_class) cl2);
	    ira_assert (cost < 65535);
	  }
	all_match &= (last_move_cost[cl1][cl2] == cost);
	last_move_cost[cl1][cl2] = cost;
      }
1616 1617
  if (all_match && last_mode_for_init_move_cost != -1)
    {
1618 1619 1620 1621 1622 1623
      ira_register_move_cost[mode]
	= ira_register_move_cost[last_mode_for_init_move_cost];
      ira_may_move_in_cost[mode]
	= ira_may_move_in_cost[last_mode_for_init_move_cost];
      ira_may_move_out_cost[mode]
	= ira_may_move_out_cost[last_mode_for_init_move_cost];
1624 1625
      return;
    }
1626
  last_mode_for_init_move_cost = mode;
1627 1628 1629
  ira_register_move_cost[mode] = XNEWVEC (move_table, N_REG_CLASSES);
  ira_may_move_in_cost[mode] = XNEWVEC (move_table, N_REG_CLASSES);
  ira_may_move_out_cost[mode] = XNEWVEC (move_table, N_REG_CLASSES);
1630
  for (cl1 = 0; cl1 < N_REG_CLASSES; cl1++)
1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671 1672 1673
    for (cl2 = 0; cl2 < N_REG_CLASSES; cl2++)
      {
	int cost;
	enum reg_class *p1, *p2;
	
	if (last_move_cost[cl1][cl2] == 65535)
	  {
	    ira_register_move_cost[mode][cl1][cl2] = 65535;
	    ira_may_move_in_cost[mode][cl1][cl2] = 65535;
	    ira_may_move_out_cost[mode][cl1][cl2] = 65535;
	  }
	else
	  {
	    cost = last_move_cost[cl1][cl2];
	    
	    for (p2 = &reg_class_subclasses[cl2][0];
		 *p2 != LIM_REG_CLASSES; p2++)
	      if (ira_class_hard_regs_num[*p2] > 0
		  && (ira_reg_class_max_nregs[*p2][mode]
		      <= ira_class_hard_regs_num[*p2]))
		cost = MAX (cost, ira_register_move_cost[mode][cl1][*p2]);
	    
	    for (p1 = &reg_class_subclasses[cl1][0];
		 *p1 != LIM_REG_CLASSES; p1++)
	      if (ira_class_hard_regs_num[*p1] > 0
		  && (ira_reg_class_max_nregs[*p1][mode]
		      <= ira_class_hard_regs_num[*p1]))
		cost = MAX (cost, ira_register_move_cost[mode][*p1][cl2]);
	    
	    ira_assert (cost <= 65535);
	    ira_register_move_cost[mode][cl1][cl2] = cost;
	    
	    if (ira_class_subset_p[cl1][cl2])
	      ira_may_move_in_cost[mode][cl1][cl2] = 0;
	    else
	      ira_may_move_in_cost[mode][cl1][cl2] = cost;
	    
	    if (ira_class_subset_p[cl2][cl1])
	      ira_may_move_out_cost[mode][cl1][cl2] = 0;
	    else
	      ira_may_move_out_cost[mode][cl1][cl2] = cost;
	  }
      }
Vladimir Makarov committed
1674
}
1675

Vladimir Makarov committed
1676 1677 1678 1679 1680 1681 1682 1683 1684


/* This is called once during compiler work.  It sets up
   different arrays whose values don't depend on the compiled
   function.  */
void
ira_init_once (void)
{
  ira_init_costs_once ();
1685
  lra_init_once ();
1686 1687

  ira_use_lra_p = targetm.lra_p ();
Vladimir Makarov committed
1688 1689
}

1690 1691
/* Free ira_max_register_move_cost, ira_may_move_in_cost and
   ira_may_move_out_cost for each mode.  */
1692 1693
void
target_ira_int::free_register_move_costs (void)
Vladimir Makarov committed
1694
{
1695
  int mode, i;
Vladimir Makarov committed
1696

1697 1698 1699
  /* Reset move_cost and friends, making sure we only free shared
     table entries once.  */
  for (mode = 0; mode < MAX_MACHINE_MODE; mode++)
1700
    if (x_ira_register_move_cost[mode])
1701
      {
1702
	for (i = 0;
1703 1704
	     i < mode && (x_ira_register_move_cost[i]
			  != x_ira_register_move_cost[mode]);
1705
	     i++)
1706 1707 1708
	  ;
	if (i == mode)
	  {
1709 1710 1711
	    free (x_ira_register_move_cost[mode]);
	    free (x_ira_may_move_in_cost[mode]);
	    free (x_ira_may_move_out_cost[mode]);
1712 1713
	  }
      }
1714 1715 1716
  memset (x_ira_register_move_cost, 0, sizeof x_ira_register_move_cost);
  memset (x_ira_may_move_in_cost, 0, sizeof x_ira_may_move_in_cost);
  memset (x_ira_may_move_out_cost, 0, sizeof x_ira_may_move_out_cost);
1717
  last_mode_for_init_move_cost = -1;
Vladimir Makarov committed
1718 1719
}

1720 1721 1722 1723 1724 1725
target_ira_int::~target_ira_int ()
{
  free_ira_costs ();
  free_register_move_costs ();
}

Vladimir Makarov committed
1726 1727 1728 1729 1730
/* This is called every time when register related information is
   changed.  */
void
ira_init (void)
{
1731
  this_target_ira_int->free_register_move_costs ();
Vladimir Makarov committed
1732 1733 1734 1735 1736
  setup_reg_mode_hard_regset ();
  setup_alloc_regs (flag_omit_frame_pointer != 0);
  setup_class_subset_and_memory_move_costs ();
  setup_reg_class_nregs ();
  setup_prohibited_class_mode_regs ();
Vladimir Makarov committed
1737 1738 1739
  find_reg_classes ();
  clarify_prohibited_class_mode_regs ();
  setup_hard_regno_aclass ();
Vladimir Makarov committed
1740 1741 1742 1743
  ira_init_costs ();
}


1744 1745
#define ira_prohibited_mode_move_regs_initialized_p \
  (this_target_ira_int->x_ira_prohibited_mode_move_regs_initialized_p)
Vladimir Makarov committed
1746 1747 1748 1749 1750 1751

/* Set up IRA_PROHIBITED_MODE_MOVE_REGS.  */
static void
setup_prohibited_mode_move_regs (void)
{
  int i, j;
1752 1753
  rtx test_reg1, test_reg2, move_pat;
  rtx_insn *move_insn;
Vladimir Makarov committed
1754 1755 1756 1757

  if (ira_prohibited_mode_move_regs_initialized_p)
    return;
  ira_prohibited_mode_move_regs_initialized_p = true;
1758 1759
  test_reg1 = gen_rtx_REG (word_mode, LAST_VIRTUAL_REGISTER + 1);
  test_reg2 = gen_rtx_REG (word_mode, LAST_VIRTUAL_REGISTER + 2);
1760
  move_pat = gen_rtx_SET (test_reg1, test_reg2);
Richard Sandiford committed
1761
  move_insn = gen_rtx_INSN (VOIDmode, 0, 0, 0, move_pat, 0, -1, 0);
Vladimir Makarov committed
1762 1763 1764 1765 1766
  for (i = 0; i < NUM_MACHINE_MODES; i++)
    {
      SET_HARD_REG_SET (ira_prohibited_mode_move_regs[i]);
      for (j = 0; j < FIRST_PSEUDO_REGISTER; j++)
	{
1767
	  if (!targetm.hard_regno_mode_ok (j, (machine_mode) i))
Vladimir Makarov committed
1768
	    continue;
1769 1770
	  set_mode_and_regno (test_reg1, (machine_mode) i, j);
	  set_mode_and_regno (test_reg2, (machine_mode) i, j);
Vladimir Makarov committed
1771 1772 1773 1774 1775
	  INSN_CODE (move_insn) = -1;
	  recog_memoized (move_insn);
	  if (INSN_CODE (move_insn) < 0)
	    continue;
	  extract_insn (move_insn);
1776 1777 1778
	  /* We don't know whether the move will be in code that is optimized
	     for size or speed, so consider all enabled alternatives.  */
	  if (! constrain_operands (1, get_enabled_alternatives (move_insn)))
Vladimir Makarov committed
1779 1780 1781 1782 1783 1784 1785 1786
	    continue;
	  CLEAR_HARD_REG_BIT (ira_prohibited_mode_move_regs[i], j);
	}
    }
}



Vladimir Makarov committed
1787 1788
/* Setup possible alternatives in ALTS for INSN.  */
void
1789
ira_setup_alts (rtx_insn *insn, HARD_REG_SET &alts)
Vladimir Makarov committed
1790 1791
{
  /* MAP nalt * nop -> start of constraints for given operand and
1792
     alternative.  */
Vladimir Makarov committed
1793 1794 1795 1796 1797 1798 1799
  static vec<const char *> insn_constraints;
  int nop, nalt;
  bool curr_swapped;
  const char *p;
  int commutative = -1;

  extract_insn (insn);
1800
  alternative_mask preferred = get_preferred_alternatives (insn);
Vladimir Makarov committed
1801 1802 1803 1804 1805 1806 1807 1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819 1820 1821 1822
  CLEAR_HARD_REG_SET (alts);
  insn_constraints.release ();
  insn_constraints.safe_grow_cleared (recog_data.n_operands
				      * recog_data.n_alternatives + 1);
  /* Check that the hard reg set is enough for holding all
     alternatives.  It is hard to imagine the situation when the
     assertion is wrong.  */
  ira_assert (recog_data.n_alternatives
	      <= (int) MAX (sizeof (HARD_REG_ELT_TYPE) * CHAR_BIT,
			    FIRST_PSEUDO_REGISTER));
  for (curr_swapped = false;; curr_swapped = true)
    {
      /* Calculate some data common for all alternatives to speed up the
	 function.  */
      for (nop = 0; nop < recog_data.n_operands; nop++)
	{
	  for (nalt = 0, p = recog_data.constraints[nop];
	       nalt < recog_data.n_alternatives;
	       nalt++)
	    {
	      insn_constraints[nop * recog_data.n_alternatives + nalt] = p;
	      while (*p && *p != ',')
1823 1824 1825 1826 1827 1828 1829
		{
		  /* We only support one commutative marker, the first
		     one.  We already set commutative above.  */
		  if (*p == '%' && commutative < 0)
		    commutative = nop;
		  p++;
		}
Vladimir Makarov committed
1830 1831 1832 1833 1834 1835
	      if (*p)
		p++;
	    }
	}
      for (nalt = 0; nalt < recog_data.n_alternatives; nalt++)
	{
1836
	  if (!TEST_BIT (preferred, nalt)
1837
	      || TEST_HARD_REG_BIT (alts, nalt))
Vladimir Makarov committed
1838 1839 1840 1841 1842 1843
	    continue;

	  for (nop = 0; nop < recog_data.n_operands; nop++)
	    {
	      int c, len;

1844
	      rtx op = recog_data.operand[nop];
Vladimir Makarov committed
1845 1846 1847 1848 1849 1850 1851 1852 1853 1854
	      p = insn_constraints[nop * recog_data.n_alternatives + nalt];
	      if (*p == 0 || *p == ',')
		continue;
	      
	      do
		switch (c = *p, len = CONSTRAINT_LEN (c, p), c)
		  {
		  case '#':
		  case ',':
		    c = '\0';
1855
		    /* FALLTHRU */
Vladimir Makarov committed
1856 1857 1858 1859 1860
		  case '\0':
		    len = 0;
		    break;
		  
		  case '%':
1861
		    /* The commutative modifier is handled above.  */
Vladimir Makarov committed
1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874
		    break;

		  case '0':  case '1':  case '2':  case '3':  case '4':
		  case '5':  case '6':  case '7':  case '8':  case '9':
		    goto op_success;
		    break;
		    
		  case 'g':
		    goto op_success;
		    break;
		    
		  default:
		    {
1875 1876 1877 1878 1879 1880 1881 1882
		      enum constraint_num cn = lookup_constraint (p);
		      switch (get_constraint_type (cn))
			{
			case CT_REGISTER:
			  if (reg_class_for_constraint (cn) != NO_REGS)
			    goto op_success;
			  break;

1883 1884 1885 1886 1887 1888 1889
			case CT_CONST_INT:
			  if (CONST_INT_P (op)
			      && (insn_const_int_ok_for_constraint
				  (INTVAL (op), cn)))
			    goto op_success;
			  break;

1890 1891
			case CT_ADDRESS:
			case CT_MEMORY:
1892
			case CT_SPECIAL_MEMORY:
1893 1894 1895 1896 1897 1898 1899
			  goto op_success;

			case CT_FIXED_FORM:
			  if (constraint_satisfied_p (op, cn))
			    goto op_success;
			  break;
			}
Vladimir Makarov committed
1900 1901 1902 1903 1904 1905 1906 1907 1908 1909 1910 1911 1912
		      break;
		    }
		  }
	      while (p += len, c);
	      break;
	    op_success:
	      ;
	    }
	  if (nop >= recog_data.n_operands)
	    SET_HARD_REG_BIT (alts, nalt);
	}
      if (commutative < 0)
	break;
1913
      /* Swap forth and back to avoid changing recog_data.  */
1914 1915
      std::swap (recog_data.operand[commutative],
		 recog_data.operand[commutative + 1]);
1916 1917
      if (curr_swapped)
	break;
Vladimir Makarov committed
1918 1919 1920 1921 1922 1923 1924 1925 1926 1927 1928 1929 1930 1931 1932 1933
    }
}

/* Return the number of the output non-early clobber operand which
   should be the same in any case as operand with number OP_NUM (or
   negative value if there is no such operand).  The function takes
   only really possible alternatives into consideration.  */
int
ira_get_dup_out_num (int op_num, HARD_REG_SET &alts)
{
  int curr_alt, c, original, dup;
  bool ignore_p, use_commut_op_p;
  const char *str;

  if (op_num < 0 || recog_data.n_alternatives == 0)
    return -1;
1934 1935 1936
  /* We should find duplications only for input operands.  */
  if (recog_data.operand_type[op_num] != OP_IN)
    return -1;
Vladimir Makarov committed
1937
  str = recog_data.constraints[op_num];
1938
  use_commut_op_p = false;
Vladimir Makarov committed
1939 1940
  for (;;)
    {
1941
      rtx op = recog_data.operand[op_num];
Vladimir Makarov committed
1942
      
1943 1944
      for (curr_alt = 0, ignore_p = !TEST_HARD_REG_BIT (alts, curr_alt),
	   original = -1;;)
Vladimir Makarov committed
1945 1946 1947 1948
	{
	  c = *str;
	  if (c == '\0')
	    break;
1949
	  if (c == '#')
Vladimir Makarov committed
1950 1951 1952 1953
	    ignore_p = true;
	  else if (c == ',')
	    {
	      curr_alt++;
1954
	      ignore_p = !TEST_HARD_REG_BIT (alts, curr_alt);
Vladimir Makarov committed
1955 1956 1957 1958 1959 1960
	    }
	  else if (! ignore_p)
	    switch (c)
	      {
	      case 'g':
		goto fail;
Richard Sandiford committed
1961
	      default:
Vladimir Makarov committed
1962
		{
1963 1964 1965 1966 1967 1968
		  enum constraint_num cn = lookup_constraint (str);
		  enum reg_class cl = reg_class_for_constraint (cn);
		  if (cl != NO_REGS
		      && !targetm.class_likely_spilled_p (cl))
		    goto fail;
		  if (constraint_satisfied_p (op, cn))
Vladimir Makarov committed
1969 1970 1971 1972 1973 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008
		    goto fail;
		  break;
		}
		
	      case '0': case '1': case '2': case '3': case '4':
	      case '5': case '6': case '7': case '8': case '9':
		if (original != -1 && original != c)
		  goto fail;
		original = c;
		break;
	      }
	  str += CONSTRAINT_LEN (c, str);
	}
      if (original == -1)
	goto fail;
      dup = -1;
      for (ignore_p = false, str = recog_data.constraints[original - '0'];
	   *str != 0;
	   str++)
	if (ignore_p)
	  {
	    if (*str == ',')
	      ignore_p = false;
	  }
	else if (*str == '#')
	  ignore_p = true;
	else if (! ignore_p)
	  {
	    if (*str == '=')
	      dup = original - '0';
	    /* It is better ignore an alternative with early clobber.  */
	    else if (*str == '&')
	      goto fail;
	  }
      if (dup >= 0)
	return dup;
    fail:
      if (use_commut_op_p)
	break;
      use_commut_op_p = true;
2009
      if (recog_data.constraints[op_num][0] == '%')
Vladimir Makarov committed
2010
	str = recog_data.constraints[op_num + 1];
2011
      else if (op_num > 0 && recog_data.constraints[op_num - 1][0] == '%')
Vladimir Makarov committed
2012 2013 2014 2015 2016 2017 2018 2019 2020 2021 2022 2023 2024 2025 2026 2027 2028 2029 2030 2031 2032 2033 2034
	str = recog_data.constraints[op_num - 1];
      else
	break;
    }
  return -1;
}



/* Search forward to see if the source register of a copy insn dies
   before either it or the destination register is modified, but don't
   scan past the end of the basic block.  If so, we can replace the
   source with the destination and let the source die in the copy
   insn.

   This will reduce the number of registers live in that range and may
   enable the destination and the source coalescing, thus often saving
   one register in addition to a register-register copy.  */

static void
decrease_live_ranges_number (void)
{
  basic_block bb;
2035
  rtx_insn *insn;
2036 2037
  rtx set, src, dest, dest_death, note;
  rtx_insn *p, *q;
Vladimir Makarov committed
2038 2039 2040 2041 2042 2043 2044 2045
  int sregno, dregno;

  if (! flag_expensive_optimizations)
    return;

  if (ira_dump_file)
    fprintf (ira_dump_file, "Starting decreasing number of live ranges...\n");

2046
  FOR_EACH_BB_FN (bb, cfun)
Vladimir Makarov committed
2047 2048 2049 2050 2051 2052 2053 2054 2055 2056 2057 2058 2059 2060 2061 2062 2063 2064 2065 2066 2067 2068 2069 2070 2071 2072 2073 2074 2075 2076 2077 2078 2079 2080 2081 2082 2083 2084 2085 2086 2087 2088 2089 2090 2091 2092 2093 2094 2095 2096 2097 2098 2099 2100 2101 2102 2103 2104 2105 2106 2107 2108 2109 2110 2111 2112 2113 2114 2115 2116 2117 2118 2119 2120 2121 2122 2123 2124 2125 2126 2127 2128 2129 2130 2131 2132 2133 2134 2135 2136 2137 2138 2139 2140 2141 2142 2143 2144 2145 2146 2147 2148 2149 2150 2151 2152 2153 2154 2155 2156 2157 2158 2159 2160 2161 2162 2163 2164 2165 2166 2167 2168 2169 2170 2171 2172 2173 2174 2175 2176 2177 2178 2179 2180 2181 2182 2183 2184 2185 2186 2187 2188 2189 2190 2191 2192 2193 2194 2195 2196 2197 2198 2199
    FOR_BB_INSNS (bb, insn)
      {
	set = single_set (insn);
	if (! set)
	  continue;
	src = SET_SRC (set);
	dest = SET_DEST (set);
	if (! REG_P (src) || ! REG_P (dest)
	    || find_reg_note (insn, REG_DEAD, src))
	  continue;
	sregno = REGNO (src);
	dregno = REGNO (dest);
	
	/* We don't want to mess with hard regs if register classes
	   are small.  */
	if (sregno == dregno
	    || (targetm.small_register_classes_for_mode_p (GET_MODE (src))
		&& (sregno < FIRST_PSEUDO_REGISTER
		    || dregno < FIRST_PSEUDO_REGISTER))
	    /* We don't see all updates to SP if they are in an
	       auto-inc memory reference, so we must disallow this
	       optimization on them.  */
	    || sregno == STACK_POINTER_REGNUM
	    || dregno == STACK_POINTER_REGNUM)
	  continue;
	
	dest_death = NULL_RTX;

	for (p = NEXT_INSN (insn); p; p = NEXT_INSN (p))
	  {
	    if (! INSN_P (p))
	      continue;
	    if (BLOCK_FOR_INSN (p) != bb)
	      break;
	    
	    if (reg_set_p (src, p) || reg_set_p (dest, p)
		/* If SRC is an asm-declared register, it must not be
		   replaced in any asm.  Unfortunately, the REG_EXPR
		   tree for the asm variable may be absent in the SRC
		   rtx, so we can't check the actual register
		   declaration easily (the asm operand will have it,
		   though).  To avoid complicating the test for a rare
		   case, we just don't perform register replacement
		   for a hard reg mentioned in an asm.  */
		|| (sregno < FIRST_PSEUDO_REGISTER
		    && asm_noperands (PATTERN (p)) >= 0
		    && reg_overlap_mentioned_p (src, PATTERN (p)))
		/* Don't change hard registers used by a call.  */
		|| (CALL_P (p) && sregno < FIRST_PSEUDO_REGISTER
		    && find_reg_fusage (p, USE, src))
		/* Don't change a USE of a register.  */
		|| (GET_CODE (PATTERN (p)) == USE
		    && reg_overlap_mentioned_p (src, XEXP (PATTERN (p), 0))))
	      break;
	    
	    /* See if all of SRC dies in P.  This test is slightly
	       more conservative than it needs to be.  */
	    if ((note = find_regno_note (p, REG_DEAD, sregno))
		&& GET_MODE (XEXP (note, 0)) == GET_MODE (src))
	      {
		int failed = 0;
		
		/* We can do the optimization.  Scan forward from INSN
		   again, replacing regs as we go.  Set FAILED if a
		   replacement can't be done.  In that case, we can't
		   move the death note for SRC.  This should be
		   rare.  */
		
		/* Set to stop at next insn.  */
		for (q = next_real_insn (insn);
		     q != next_real_insn (p);
		     q = next_real_insn (q))
		  {
		    if (reg_overlap_mentioned_p (src, PATTERN (q)))
		      {
			/* If SRC is a hard register, we might miss
			   some overlapping registers with
			   validate_replace_rtx, so we would have to
			   undo it.  We can't if DEST is present in
			   the insn, so fail in that combination of
			   cases.  */
			if (sregno < FIRST_PSEUDO_REGISTER
			    && reg_mentioned_p (dest, PATTERN (q)))
			  failed = 1;
			
			/* Attempt to replace all uses.  */
			else if (!validate_replace_rtx (src, dest, q))
			  failed = 1;
			
			/* If this succeeded, but some part of the
			   register is still present, undo the
			   replacement.  */
			else if (sregno < FIRST_PSEUDO_REGISTER
				 && reg_overlap_mentioned_p (src, PATTERN (q)))
			  {
			    validate_replace_rtx (dest, src, q);
			    failed = 1;
			  }
		      }
		    
		    /* If DEST dies here, remove the death note and
		       save it for later.  Make sure ALL of DEST dies
		       here; again, this is overly conservative.  */
		    if (! dest_death
			&& (dest_death = find_regno_note (q, REG_DEAD, dregno)))
		      {
			if (GET_MODE (XEXP (dest_death, 0)) == GET_MODE (dest))
			  remove_note (q, dest_death);
			else
			  {
			    failed = 1;
			    dest_death = 0;
			  }
		      }
		  }
		
		if (! failed)
		  {
		    /* Move death note of SRC from P to INSN.  */
		    remove_note (p, note);
		    XEXP (note, 1) = REG_NOTES (insn);
		    REG_NOTES (insn) = note;
		  }
		
		/* DEST is also dead if INSN has a REG_UNUSED note for
		   DEST.  */
		if (! dest_death
		    && (dest_death
			= find_regno_note (insn, REG_UNUSED, dregno)))
		  {
		    PUT_REG_NOTE_KIND (dest_death, REG_DEAD);
		    remove_note (insn, dest_death);
		  }
		
		/* Put death note of DEST on P if we saw it die.  */
		if (dest_death)
		  {
		    XEXP (dest_death, 1) = REG_NOTES (p);
		    REG_NOTES (p) = dest_death;
		  }
		break;
	      }
	    
	    /* If SRC is a hard register which is set or killed in
	       some other way, we can't do this optimization.  */
	    else if (sregno < FIRST_PSEUDO_REGISTER && dead_or_set_p (p, src))
	      break;
	  }
      }
}



2200 2201 2202 2203
/* Return nonzero if REGNO is a particularly bad choice for reloading X.  */
static bool
ira_bad_reload_regno_1 (int regno, rtx x)
{
2204
  int x_regno, n, i;
2205 2206 2207 2208 2209 2210 2211 2212 2213 2214 2215 2216 2217 2218 2219 2220 2221 2222 2223 2224
  ira_allocno_t a;
  enum reg_class pref;

  /* We only deal with pseudo regs.  */
  if (! x || GET_CODE (x) != REG)
    return false;

  x_regno = REGNO (x);
  if (x_regno < FIRST_PSEUDO_REGISTER)
    return false;

  /* If the pseudo prefers REGNO explicitly, then do not consider
     REGNO a bad spill choice.  */
  pref = reg_preferred_class (x_regno);
  if (reg_class_size[pref] == 1)
    return !TEST_HARD_REG_BIT (reg_class_contents[pref], regno);

  /* If the pseudo conflicts with REGNO, then we consider REGNO a
     poor choice for a reload regno.  */
  a = ira_regno_allocno_map[x_regno];
2225 2226 2227 2228 2229 2230 2231
  n = ALLOCNO_NUM_OBJECTS (a);
  for (i = 0; i < n; i++)
    {
      ira_object_t obj = ALLOCNO_OBJECT (a, i);
      if (TEST_HARD_REG_BIT (OBJECT_TOTAL_CONFLICT_HARD_REGS (obj), regno))
	return true;
    }
2232 2233 2234 2235 2236 2237 2238 2239 2240 2241 2242 2243
  return false;
}

/* Return nonzero if REGNO is a particularly bad choice for reloading
   IN or OUT.  */
bool
ira_bad_reload_regno (int regno, rtx in, rtx out)
{
  return (ira_bad_reload_regno_1 (regno, in)
	  || ira_bad_reload_regno_1 (regno, out));
}

2244
/* Add register clobbers from asm statements.  */
Vladimir Makarov committed
2245
static void
2246
compute_regs_asm_clobbered (void)
Vladimir Makarov committed
2247 2248 2249
{
  basic_block bb;

2250
  FOR_EACH_BB_FN (bb, cfun)
Vladimir Makarov committed
2251
    {
2252
      rtx_insn *insn;
Vladimir Makarov committed
2253 2254
      FOR_BB_INSNS_REVERSE (bb, insn)
	{
2255
	  df_ref def;
Vladimir Makarov committed
2256

2257
	  if (NONDEBUG_INSN_P (insn) && asm_noperands (PATTERN (insn)) >= 0)
2258
	    FOR_EACH_INSN_DEF (def, insn)
Vladimir Makarov committed
2259 2260
	      {
		unsigned int dregno = DF_REF_REGNO (def);
2261 2262 2263 2264
		if (HARD_REGISTER_NUM_P (dregno))
		  add_to_hard_reg_set (&crtl->asm_clobbers,
				       GET_MODE (DF_REF_REAL_REG (def)),
				       dregno);
Vladimir Makarov committed
2265 2266 2267 2268 2269 2270
	      }
	}
    }
}


2271 2272
/* Set up ELIMINABLE_REGSET, IRA_NO_ALLOC_REGS, and
   REGS_EVER_LIVE.  */
2273
void
2274
ira_setup_eliminable_regset (void)
Vladimir Makarov committed
2275
{
2276
  int i;
Vladimir Makarov committed
2277
  static const struct {const int from, to; } eliminables[] = ELIMINABLE_REGS;
2278

2279 2280 2281 2282
  /* Setup is_leaf as frame_pointer_required may use it.  This function
     is called by sched_init before ira if scheduling is enabled.  */
  crtl->is_leaf = leaf_function_p ();

Vladimir Makarov committed
2283 2284 2285 2286
  /* FIXME: If EXIT_IGNORE_STACK is set, we will not save and restore
     sp for alloca.  So we can't eliminate the frame pointer in that
     case.  At some point, we should improve this by emitting the
     sp-adjusting insns for this case.  */
2287
  frame_pointer_needed
Vladimir Makarov committed
2288 2289
    = (! flag_omit_frame_pointer
       || (cfun->calls_alloca && EXIT_IGNORE_STACK)
2290 2291 2292 2293 2294 2295
       /* We need the frame pointer to catch stack overflow exceptions if
	  the stack pointer is moving (as for the alloca case just above).  */
       || (STACK_CHECK_MOVING_SP
	   && flag_stack_check
	   && flag_exceptions
	   && cfun->can_throw_non_call_exceptions)
Vladimir Makarov committed
2296
       || crtl->accesses_prior_frames
2297
       || (SUPPORTS_STACK_ALIGNMENT && crtl->stack_realign_needed)
2298
       || targetm.frame_pointer_required ());
Vladimir Makarov committed
2299

2300 2301 2302 2303
    /* The chance that FRAME_POINTER_NEEDED is changed from inspecting
       RTL is very small.  So if we use frame pointer for RA and RTL
       actually prevents this, we will spill pseudos assigned to the
       frame pointer in LRA.  */
Vladimir Makarov committed
2304

2305 2306 2307
  if (frame_pointer_needed)
    df_set_regs_ever_live (HARD_FRAME_POINTER_REGNUM, true);
    
Vladimir Makarov committed
2308 2309 2310
  COPY_HARD_REG_SET (ira_no_alloc_regs, no_unit_alloc_regs);
  CLEAR_HARD_REG_SET (eliminable_regset);

2311 2312
  compute_regs_asm_clobbered ();

Vladimir Makarov committed
2313 2314 2315 2316 2317
  /* Build the regset of all eliminable registers and show we can't
     use those that we already know won't be eliminated.  */
  for (i = 0; i < (int) ARRAY_SIZE (eliminables); i++)
    {
      bool cannot_elim
2318
	= (! targetm.can_eliminate (eliminables[i].from, eliminables[i].to)
2319
	   || (eliminables[i].to == STACK_POINTER_REGNUM && frame_pointer_needed));
Vladimir Makarov committed
2320

2321
      if (!TEST_HARD_REG_BIT (crtl->asm_clobbers, eliminables[i].from))
Vladimir Makarov committed
2322 2323 2324 2325 2326 2327 2328 2329 2330 2331 2332 2333
	{
	    SET_HARD_REG_BIT (eliminable_regset, eliminables[i].from);

	    if (cannot_elim)
	      SET_HARD_REG_BIT (ira_no_alloc_regs, eliminables[i].from);
	}
      else if (cannot_elim)
	error ("%s cannot be used in asm here",
	       reg_names[eliminables[i].from]);
      else
	df_set_regs_ever_live (eliminables[i].from, true);
    }
2334
  if (!HARD_FRAME_POINTER_IS_FRAME_POINTER)
Vladimir Makarov committed
2335
    {
2336 2337 2338 2339 2340 2341 2342 2343 2344 2345 2346
      if (!TEST_HARD_REG_BIT (crtl->asm_clobbers, HARD_FRAME_POINTER_REGNUM))
	{
	  SET_HARD_REG_BIT (eliminable_regset, HARD_FRAME_POINTER_REGNUM);
	  if (frame_pointer_needed)
	    SET_HARD_REG_BIT (ira_no_alloc_regs, HARD_FRAME_POINTER_REGNUM);
	}
      else if (frame_pointer_needed)
	error ("%s cannot be used in asm here",
	       reg_names[HARD_FRAME_POINTER_REGNUM]);
      else
	df_set_regs_ever_live (HARD_FRAME_POINTER_REGNUM, true);
Vladimir Makarov committed
2347 2348 2349 2350 2351
    }
}



2352 2353 2354 2355 2356 2357 2358 2359
/* Vector of substitutions of register numbers,
   used to map pseudo regs into hardware regs.
   This is set up as a result of register allocation.
   Element N is the hard reg assigned to pseudo reg N,
   or is -1 if no hard reg was assigned.
   If N is a hard reg number, element N is N.  */
short *reg_renumber;

Vladimir Makarov committed
2360 2361 2362 2363 2364 2365 2366 2367 2368 2369 2370 2371
/* Set up REG_RENUMBER and CALLER_SAVE_NEEDED (used by reload) from
   the allocation found by IRA.  */
static void
setup_reg_renumber (void)
{
  int regno, hard_regno;
  ira_allocno_t a;
  ira_allocno_iterator ai;

  caller_save_needed = 0;
  FOR_EACH_ALLOCNO (a, ai)
    {
2372 2373
      if (ira_use_lra_p && ALLOCNO_CAP_MEMBER (a) != NULL)
	continue;
Vladimir Makarov committed
2374 2375 2376 2377 2378 2379 2380 2381
      /* There are no caps at this point.  */
      ira_assert (ALLOCNO_CAP_MEMBER (a) == NULL);
      if (! ALLOCNO_ASSIGNED_P (a))
	/* It can happen if A is not referenced but partially anticipated
	   somewhere in a region.  */
	ALLOCNO_ASSIGNED_P (a) = true;
      ira_free_allocno_updated_costs (a);
      hard_regno = ALLOCNO_HARD_REGNO (a);
Vladimir Makarov committed
2382
      regno = ALLOCNO_REGNO (a);
Vladimir Makarov committed
2383
      reg_renumber[regno] = (hard_regno < 0 ? -1 : hard_regno);
Vladimir Makarov committed
2384
      if (hard_regno >= 0)
Vladimir Makarov committed
2385
	{
Vladimir Makarov committed
2386 2387 2388 2389 2390 2391 2392 2393 2394 2395 2396 2397 2398
	  int i, nwords;
	  enum reg_class pclass;
	  ira_object_t obj;
	  
	  pclass = ira_pressure_class_translate[REGNO_REG_CLASS (hard_regno)];
	  nwords = ALLOCNO_NUM_OBJECTS (a);
	  for (i = 0; i < nwords; i++)
	    {
	      obj = ALLOCNO_OBJECT (a, i);
	      IOR_COMPL_HARD_REG_SET (OBJECT_TOTAL_CONFLICT_HARD_REGS (obj),
				      reg_class_contents[pclass]);
	    }
	  if (ALLOCNO_CALLS_CROSSED_NUM (a) != 0
2399 2400
	      && ira_hard_reg_set_intersection_p (hard_regno, ALLOCNO_MODE (a),
						  call_used_reg_set))
Vladimir Makarov committed
2401 2402
	    {
	      ira_assert (!optimize || flag_caller_saves
2403 2404
			  || (ALLOCNO_CALLS_CROSSED_NUM (a)
			      == ALLOCNO_CHEAP_CALLS_CROSSED_NUM (a))
2405
			  || regno >= ira_reg_equiv_len
2406
			  || ira_equiv_no_lvalue_p (regno));
Vladimir Makarov committed
2407 2408
	      caller_save_needed = 1;
	    }
Vladimir Makarov committed
2409 2410 2411 2412 2413 2414 2415 2416 2417 2418 2419 2420 2421 2422 2423 2424 2425 2426 2427 2428 2429 2430 2431 2432 2433 2434 2435
	}
    }
}

/* Set up allocno assignment flags for further allocation
   improvements.  */
static void
setup_allocno_assignment_flags (void)
{
  int hard_regno;
  ira_allocno_t a;
  ira_allocno_iterator ai;

  FOR_EACH_ALLOCNO (a, ai)
    {
      if (! ALLOCNO_ASSIGNED_P (a))
	/* It can happen if A is not referenced but partially anticipated
	   somewhere in a region.  */
	ira_free_allocno_updated_costs (a);
      hard_regno = ALLOCNO_HARD_REGNO (a);
      /* Don't assign hard registers to allocnos which are destination
	 of removed store at the end of loop.  It has no sense to keep
	 the same value in different hard registers.  It is also
	 impossible to assign hard registers correctly to such
	 allocnos because the cost info and info about intersected
	 calls are incorrect for them.  */
      ALLOCNO_ASSIGNED_P (a) = (hard_regno >= 0
Vladimir Makarov committed
2436
				|| ALLOCNO_EMIT_DATA (a)->mem_optimized_dest_p
Vladimir Makarov committed
2437
				|| (ALLOCNO_MEMORY_COST (a)
Vladimir Makarov committed
2438
				    - ALLOCNO_CLASS_COST (a)) < 0);
2439 2440 2441 2442
      ira_assert
	(hard_regno < 0
	 || ira_hard_reg_in_set_p (hard_regno, ALLOCNO_MODE (a),
				   reg_class_contents[ALLOCNO_CLASS (a)]));
Vladimir Makarov committed
2443 2444 2445 2446 2447 2448 2449 2450 2451 2452 2453 2454 2455 2456 2457 2458 2459
    }
}

/* Evaluate overall allocation cost and the costs for using hard
   registers and memory for allocnos.  */
static void
calculate_allocation_cost (void)
{
  int hard_regno, cost;
  ira_allocno_t a;
  ira_allocno_iterator ai;

  ira_overall_cost = ira_reg_cost = ira_mem_cost = 0;
  FOR_EACH_ALLOCNO (a, ai)
    {
      hard_regno = ALLOCNO_HARD_REGNO (a);
      ira_assert (hard_regno < 0
2460 2461 2462
		  || (ira_hard_reg_in_set_p
		      (hard_regno, ALLOCNO_MODE (a),
		       reg_class_contents[ALLOCNO_CLASS (a)])));
Vladimir Makarov committed
2463 2464 2465 2466 2467 2468 2469 2470 2471
      if (hard_regno < 0)
	{
	  cost = ALLOCNO_MEMORY_COST (a);
	  ira_mem_cost += cost;
	}
      else if (ALLOCNO_HARD_REG_COSTS (a) != NULL)
	{
	  cost = (ALLOCNO_HARD_REG_COSTS (a)
		  [ira_class_hard_reg_index
Vladimir Makarov committed
2472
		   [ALLOCNO_CLASS (a)][hard_regno]]);
Vladimir Makarov committed
2473 2474 2475 2476
	  ira_reg_cost += cost;
	}
      else
	{
Vladimir Makarov committed
2477
	  cost = ALLOCNO_CLASS_COST (a);
Vladimir Makarov committed
2478 2479 2480 2481 2482 2483 2484 2485
	  ira_reg_cost += cost;
	}
      ira_overall_cost += cost;
    }

  if (internal_flag_ira_verbose > 0 && ira_dump_file != NULL)
    {
      fprintf (ira_dump_file,
2486 2487 2488 2489 2490 2491
	       "+++Costs: overall %" PRId64
	       ", reg %" PRId64
	       ", mem %" PRId64
	       ", ld %" PRId64
	       ", st %" PRId64
	       ", move %" PRId64,
Vladimir Makarov committed
2492 2493
	       ira_overall_cost, ira_reg_cost, ira_mem_cost,
	       ira_load_cost, ira_store_cost, ira_shuffle_cost);
2494
      fprintf (ira_dump_file, "\n+++       move loops %d, new jumps %d\n",
Vladimir Makarov committed
2495 2496 2497 2498 2499 2500 2501 2502 2503 2504 2505 2506
	       ira_move_loops_num, ira_additional_jumps_num);
    }

}

#ifdef ENABLE_IRA_CHECKING
/* Check the correctness of the allocation.  We do need this because
   of complicated code to transform more one region internal
   representation into one region representation.  */
static void
check_allocation (void)
{
2507
  ira_allocno_t a;
2508
  int hard_regno, nregs, conflict_nregs;
Vladimir Makarov committed
2509 2510 2511 2512
  ira_allocno_iterator ai;

  FOR_EACH_ALLOCNO (a, ai)
    {
2513 2514
      int n = ALLOCNO_NUM_OBJECTS (a);
      int i;
2515

Vladimir Makarov committed
2516 2517 2518
      if (ALLOCNO_CAP_MEMBER (a) != NULL
	  || (hard_regno = ALLOCNO_HARD_REGNO (a)) < 0)
	continue;
2519
      nregs = hard_regno_nregs (hard_regno, ALLOCNO_MODE (a));
2520 2521 2522 2523 2524 2525 2526 2527
      if (nregs == 1)
	/* We allocated a single hard register.  */
	n = 1;
      else if (n > 1)
	/* We allocated multiple hard registers, and we will test
	   conflicts in a granularity of single hard regs.  */
	nregs = 1;

2528 2529 2530 2531 2532 2533 2534
      for (i = 0; i < n; i++)
	{
	  ira_object_t obj = ALLOCNO_OBJECT (a, i);
	  ira_object_t conflict_obj;
	  ira_object_conflict_iterator oci;
	  int this_regno = hard_regno;
	  if (n > 1)
2535
	    {
2536
	      if (REG_WORDS_BIG_ENDIAN)
2537 2538 2539 2540 2541 2542 2543 2544 2545 2546
		this_regno += n - i - 1;
	      else
		this_regno += i;
	    }
	  FOR_EACH_OBJECT_CONFLICT (obj, conflict_obj, oci)
	    {
	      ira_allocno_t conflict_a = OBJECT_ALLOCNO (conflict_obj);
	      int conflict_hard_regno = ALLOCNO_HARD_REGNO (conflict_a);
	      if (conflict_hard_regno < 0)
		continue;
2547

2548 2549
	      conflict_nregs = hard_regno_nregs (conflict_hard_regno,
						 ALLOCNO_MODE (conflict_a));
2550 2551 2552

	      if (ALLOCNO_NUM_OBJECTS (conflict_a) > 1
		  && conflict_nregs == ALLOCNO_NUM_OBJECTS (conflict_a))
2553
		{
2554
		  if (REG_WORDS_BIG_ENDIAN)
2555 2556 2557 2558 2559 2560 2561 2562 2563 2564 2565
		    conflict_hard_regno += (ALLOCNO_NUM_OBJECTS (conflict_a)
					    - OBJECT_SUBWORD (conflict_obj) - 1);
		  else
		    conflict_hard_regno += OBJECT_SUBWORD (conflict_obj);
		  conflict_nregs = 1;
		}

	      if ((conflict_hard_regno <= this_regno
		 && this_regno < conflict_hard_regno + conflict_nregs)
		|| (this_regno <= conflict_hard_regno
		    && conflict_hard_regno < this_regno + nregs))
2566 2567 2568 2569 2570 2571 2572
		{
		  fprintf (stderr, "bad allocation for %d and %d\n",
			   ALLOCNO_REGNO (a), ALLOCNO_REGNO (conflict_a));
		  gcc_unreachable ();
		}
	    }
	}
Vladimir Makarov committed
2573 2574 2575 2576
    }
}
#endif

2577 2578 2579 2580 2581 2582 2583 2584 2585 2586 2587 2588 2589 2590 2591 2592 2593 2594 2595
/* Allocate REG_EQUIV_INIT.  Set up it from IRA_REG_EQUIV which should
   be already calculated.  */
static void
setup_reg_equiv_init (void)
{
  int i;
  int max_regno = max_reg_num ();

  for (i = 0; i < max_regno; i++)
    reg_equiv_init (i) = ira_reg_equiv[i].init_insns;
}

/* Update equiv regno from movement of FROM_REGNO to TO_REGNO.  INSNS
   are insns which were generated for such movement.  It is assumed
   that FROM_REGNO and TO_REGNO always have the same value at the
   point of any move containing such registers. This function is used
   to update equiv info for register shuffles on the region borders
   and for caller save/restore insns.  */
void
2596
ira_update_equiv_info_by_shuffle_insn (int to_regno, int from_regno, rtx_insn *insns)
2597
{
2598 2599
  rtx_insn *insn;
  rtx x, note;
2600 2601 2602 2603 2604

  if (! ira_reg_equiv[from_regno].defined_p
      && (! ira_reg_equiv[to_regno].defined_p
	  || ((x = ira_reg_equiv[to_regno].memory) != NULL_RTX
	      && ! MEM_READONLY_P (x))))
2605
    return;
2606 2607 2608 2609 2610 2611 2612 2613 2614 2615 2616 2617
  insn = insns;
  if (NEXT_INSN (insn) != NULL_RTX)
    {
      if (! ira_reg_equiv[to_regno].defined_p)
	{
	  ira_assert (ira_reg_equiv[to_regno].init_insns == NULL_RTX);
	  return;
	}
      ira_reg_equiv[to_regno].defined_p = false;
      ira_reg_equiv[to_regno].memory
	= ira_reg_equiv[to_regno].constant
	= ira_reg_equiv[to_regno].invariant
2618
	= ira_reg_equiv[to_regno].init_insns = NULL;
2619 2620 2621 2622 2623 2624 2625 2626 2627 2628 2629 2630 2631 2632 2633 2634 2635 2636 2637 2638 2639 2640 2641 2642 2643 2644 2645 2646 2647 2648 2649 2650 2651 2652 2653 2654 2655 2656 2657 2658 2659
      if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
	fprintf (ira_dump_file,
		 "      Invalidating equiv info for reg %d\n", to_regno);
      return;
    }
  /* It is possible that FROM_REGNO still has no equivalence because
     in shuffles to_regno<-from_regno and from_regno<-to_regno the 2nd
     insn was not processed yet.  */
  if (ira_reg_equiv[from_regno].defined_p)
    {
      ira_reg_equiv[to_regno].defined_p = true;
      if ((x = ira_reg_equiv[from_regno].memory) != NULL_RTX)
	{
	  ira_assert (ira_reg_equiv[from_regno].invariant == NULL_RTX
		      && ira_reg_equiv[from_regno].constant == NULL_RTX);
	  ira_assert (ira_reg_equiv[to_regno].memory == NULL_RTX
		      || rtx_equal_p (ira_reg_equiv[to_regno].memory, x));
	  ira_reg_equiv[to_regno].memory = x;
	  if (! MEM_READONLY_P (x))
	    /* We don't add the insn to insn init list because memory
	       equivalence is just to say what memory is better to use
	       when the pseudo is spilled.  */
	    return;
	}
      else if ((x = ira_reg_equiv[from_regno].constant) != NULL_RTX)
	{
	  ira_assert (ira_reg_equiv[from_regno].invariant == NULL_RTX);
	  ira_assert (ira_reg_equiv[to_regno].constant == NULL_RTX
		      || rtx_equal_p (ira_reg_equiv[to_regno].constant, x));
	  ira_reg_equiv[to_regno].constant = x;
	}
      else
	{
	  x = ira_reg_equiv[from_regno].invariant;
	  ira_assert (x != NULL_RTX);
	  ira_assert (ira_reg_equiv[to_regno].invariant == NULL_RTX
		      || rtx_equal_p (ira_reg_equiv[to_regno].invariant, x));
	  ira_reg_equiv[to_regno].invariant = x;
	}
      if (find_reg_note (insn, REG_EQUIV, x) == NULL_RTX)
	{
2660
	  note = set_unique_reg_note (insn, REG_EQUIV, copy_rtx (x));
2661 2662 2663 2664 2665 2666
	  gcc_assert (note != NULL_RTX);
	  if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
	    {
	      fprintf (ira_dump_file,
		       "      Adding equiv note to insn %u for reg %d ",
		       INSN_UID (insn), to_regno);
2667
	      dump_value_slim (ira_dump_file, x, 1);
2668 2669 2670 2671 2672 2673 2674 2675 2676 2677 2678 2679 2680
	      fprintf (ira_dump_file, "\n");
	    }
	}
    }
  ira_reg_equiv[to_regno].init_insns
    = gen_rtx_INSN_LIST (VOIDmode, insn,
			 ira_reg_equiv[to_regno].init_insns);
  if (internal_flag_ira_verbose > 3 && ira_dump_file != NULL)
    fprintf (ira_dump_file,
	     "      Adding equiv init move insn %u to reg %d\n",
	     INSN_UID (insn), to_regno);
}

Vladimir Makarov committed
2681 2682 2683 2684 2685
/* Fix values of array REG_EQUIV_INIT after live range splitting done
   by IRA.  */
static void
fix_reg_equiv_init (void)
{
2686
  int max_regno = max_reg_num ();
2687
  int i, new_regno, max;
2688 2689 2690
  rtx set;
  rtx_insn_list *x, *next, *prev;
  rtx_insn *insn;
H.J. Lu committed
2691

2692
  if (max_regno_before_ira < max_regno)
Vladimir Makarov committed
2693
    {
2694
      max = vec_safe_length (reg_equivs);
2695 2696
      grow_reg_equivs ();
      for (i = FIRST_PSEUDO_REGISTER; i < max; i++)
2697
	for (prev = NULL, x = reg_equiv_init (i);
2698 2699
	     x != NULL_RTX;
	     x = next)
Vladimir Makarov committed
2700
	  {
2701 2702 2703
	    next = x->next ();
	    insn = x->insn ();
	    set = single_set (insn);
Vladimir Makarov committed
2704 2705 2706 2707 2708 2709 2710 2711 2712 2713 2714 2715 2716 2717 2718 2719
	    ira_assert (set != NULL_RTX
			&& (REG_P (SET_DEST (set)) || REG_P (SET_SRC (set))));
	    if (REG_P (SET_DEST (set))
		&& ((int) REGNO (SET_DEST (set)) == i
		    || (int) ORIGINAL_REGNO (SET_DEST (set)) == i))
	      new_regno = REGNO (SET_DEST (set));
	    else if (REG_P (SET_SRC (set))
		     && ((int) REGNO (SET_SRC (set)) == i
			 || (int) ORIGINAL_REGNO (SET_SRC (set)) == i))
	      new_regno = REGNO (SET_SRC (set));
	    else
 	      gcc_unreachable ();
	    if (new_regno == i)
	      prev = x;
	    else
	      {
2720
		/* Remove the wrong list element.  */
Vladimir Makarov committed
2721
		if (prev == NULL_RTX)
2722
		  reg_equiv_init (i) = next;
Vladimir Makarov committed
2723 2724
		else
		  XEXP (prev, 1) = next;
2725 2726
		XEXP (x, 1) = reg_equiv_init (new_regno);
		reg_equiv_init (new_regno) = x;
Vladimir Makarov committed
2727 2728 2729 2730 2731 2732 2733 2734 2735 2736 2737 2738 2739 2740
	      }
	  }
    }
}

#ifdef ENABLE_IRA_CHECKING
/* Print redundant memory-memory copies.  */
static void
print_redundant_copies (void)
{
  int hard_regno;
  ira_allocno_t a;
  ira_copy_t cp, next_cp;
  ira_allocno_iterator ai;
H.J. Lu committed
2741

Vladimir Makarov committed
2742 2743 2744
  FOR_EACH_ALLOCNO (a, ai)
    {
      if (ALLOCNO_CAP_MEMBER (a) != NULL)
2745
	/* It is a cap.  */
Vladimir Makarov committed
2746 2747 2748 2749 2750 2751 2752 2753 2754 2755 2756 2757 2758 2759 2760 2761 2762 2763 2764 2765 2766 2767 2768 2769 2770 2771 2772 2773 2774 2775 2776 2777
	continue;
      hard_regno = ALLOCNO_HARD_REGNO (a);
      if (hard_regno >= 0)
	continue;
      for (cp = ALLOCNO_COPIES (a); cp != NULL; cp = next_cp)
	if (cp->first == a)
	  next_cp = cp->next_first_allocno_copy;
	else
	  {
	    next_cp = cp->next_second_allocno_copy;
	    if (internal_flag_ira_verbose > 4 && ira_dump_file != NULL
		&& cp->insn != NULL_RTX
		&& ALLOCNO_HARD_REGNO (cp->first) == hard_regno)
	      fprintf (ira_dump_file,
		       "        Redundant move from %d(freq %d):%d\n",
		       INSN_UID (cp->insn), cp->freq, hard_regno);
	  }
    }
}
#endif

/* Setup preferred and alternative classes for new pseudo-registers
   created by IRA starting with START.  */
static void
setup_preferred_alternate_classes_for_new_pseudos (int start)
{
  int i, old_regno;
  int max_regno = max_reg_num ();

  for (i = start; i < max_regno; i++)
    {
      old_regno = ORIGINAL_REGNO (regno_reg_rtx[i]);
H.J. Lu committed
2778
      ira_assert (i != old_regno);
Vladimir Makarov committed
2779
      setup_reg_classes (i, reg_preferred_class (old_regno),
2780
			 reg_alternate_class (old_regno),
Vladimir Makarov committed
2781
			 reg_allocno_class (old_regno));
Vladimir Makarov committed
2782 2783 2784 2785 2786 2787 2788 2789 2790
      if (internal_flag_ira_verbose > 2 && ira_dump_file != NULL)
	fprintf (ira_dump_file,
		 "    New r%d: setting preferred %s, alternative %s\n",
		 i, reg_class_names[reg_preferred_class (old_regno)],
		 reg_class_names[reg_alternate_class (old_regno)]);
    }
}


Kito Cheng committed
2791
/* The number of entries allocated in reg_info.  */
2792
static int allocated_reg_info_size;
Vladimir Makarov committed
2793 2794 2795 2796

/* Regional allocation can create new pseudo-registers.  This function
   expands some arrays for pseudo-registers.  */
static void
2797
expand_reg_info (void)
Vladimir Makarov committed
2798 2799 2800 2801 2802
{
  int i;
  int size = max_reg_num ();

  resize_reg_info ();
2803
  for (i = allocated_reg_info_size; i < size; i++)
2804
    setup_reg_classes (i, GENERAL_REGS, ALL_REGS, GENERAL_REGS);
2805 2806
  setup_preferred_alternate_classes_for_new_pseudos (allocated_reg_info_size);
  allocated_reg_info_size = size;
Vladimir Makarov committed
2807 2808
}

2809 2810 2811 2812 2813 2814
/* Return TRUE if there is too high register pressure in the function.
   It is used to decide when stack slot sharing is worth to do.  */
static bool
too_high_register_pressure_p (void)
{
  int i;
Vladimir Makarov committed
2815
  enum reg_class pclass;
H.J. Lu committed
2816

Vladimir Makarov committed
2817
  for (i = 0; i < ira_pressure_classes_num; i++)
2818
    {
Vladimir Makarov committed
2819 2820
      pclass = ira_pressure_classes[i];
      if (ira_loop_tree_root->reg_pressure[pclass] > 10000)
2821 2822 2823 2824 2825
	return true;
    }
  return false;
}

Vladimir Makarov committed
2826 2827


2828 2829 2830 2831 2832 2833 2834 2835 2836
/* Indicate that hard register number FROM was eliminated and replaced with
   an offset from hard register number TO.  The status of hard registers live
   at the start of a basic block is updated by replacing a use of FROM with
   a use of TO.  */

void
mark_elimination (int from, int to)
{
  basic_block bb;
2837
  bitmap r;
2838

2839
  FOR_EACH_BB_FN (bb, cfun)
2840
    {
2841 2842 2843 2844 2845 2846 2847 2848 2849 2850
      r = DF_LR_IN (bb);
      if (bitmap_bit_p (r, from))
	{
	  bitmap_clear_bit (r, from);
	  bitmap_set_bit (r, to);
	}
      if (! df_live)
        continue;
      r = DF_LIVE_IN (bb);
      if (bitmap_bit_p (r, from))
2851
	{
2852 2853
	  bitmap_clear_bit (r, from);
	  bitmap_set_bit (r, to);
2854 2855 2856 2857 2858 2859
	}
    }
}



2860 2861 2862 2863
/* The length of the following array.  */
int ira_reg_equiv_len;

/* Info about equiv. info for each register.  */
2864
struct ira_reg_equiv_s *ira_reg_equiv;
2865 2866 2867 2868 2869 2870 2871 2872 2873 2874 2875

/* Expand ira_reg_equiv if necessary.  */
void
ira_expand_reg_equiv (void)
{
  int old = ira_reg_equiv_len;

  if (ira_reg_equiv_len > max_reg_num ())
    return;
  ira_reg_equiv_len = max_reg_num () * 3 / 2 + 1;
  ira_reg_equiv
2876
    = (struct ira_reg_equiv_s *) xrealloc (ira_reg_equiv,
2877
					 ira_reg_equiv_len
2878
					 * sizeof (struct ira_reg_equiv_s));
2879 2880
  gcc_assert (old < ira_reg_equiv_len);
  memset (ira_reg_equiv + old, 0,
2881
	  sizeof (struct ira_reg_equiv_s) * (ira_reg_equiv_len - old));
2882 2883 2884 2885 2886 2887 2888 2889 2890 2891 2892 2893 2894 2895 2896 2897 2898 2899
}

static void
init_reg_equiv (void)
{
  ira_reg_equiv_len = 0;
  ira_reg_equiv = NULL;
  ira_expand_reg_equiv ();
}

static void
finish_reg_equiv (void)
{
  free (ira_reg_equiv);
}



2900 2901 2902 2903 2904 2905 2906
struct equivalence
{
  /* Set when a REG_EQUIV note is found or created.  Use to
     keep track of what memory accesses might be created later,
     e.g. by reload.  */
  rtx replacement;
  rtx *src_p;
2907 2908 2909 2910 2911 2912 2913 2914 2915 2916

  /* The list of each instruction which initializes this register.

     NULL indicates we know nothing about this register's equivalence
     properties.

     An INSN_LIST with a NULL insn indicates this pseudo is already
     known to not have a valid equivalence.  */
  rtx_insn_list *init_insns;

2917 2918
  /* Loop depth is used to recognize equivalences which appear
     to be present within the same loop (or in an inner loop).  */
2919
  short loop_depth;
2920
  /* Nonzero if this had a preexisting REG_EQUIV note.  */
2921
  unsigned char is_arg_equivalence : 1;
2922 2923
  /* Set when an attempt should be made to replace a register
     with the associated src_p entry.  */
2924 2925 2926
  unsigned char replace : 1;
  /* Set if this register has no known equivalence.  */
  unsigned char no_equiv : 1;
2927 2928
  /* Set if this register is mentioned in a paradoxical subreg.  */
  unsigned char pdx_subregs : 1;
2929 2930 2931 2932 2933 2934
};

/* reg_equiv[N] (where N is a pseudo reg number) is the equivalence
   structure for that register.  */
static struct equivalence *reg_equiv;

2935 2936 2937 2938 2939
/* Used for communication between the following two functions.  */
struct equiv_mem_data
{
  /* A MEM that we wish to ensure remains unchanged.  */
  rtx equiv_mem;
2940

2941 2942 2943
  /* Set true if EQUIV_MEM is modified.  */
  bool equiv_mem_modified;
};
2944 2945 2946 2947 2948

/* If EQUIV_MEM is modified by modifying DEST, indicate that it is modified.
   Called via note_stores.  */
static void
validate_equiv_mem_from_store (rtx dest, const_rtx set ATTRIBUTE_UNUSED,
2949
			       void *data)
2950
{
2951 2952
  struct equiv_mem_data *info = (struct equiv_mem_data *) data;

2953
  if ((REG_P (dest)
2954
       && reg_overlap_mentioned_p (dest, info->equiv_mem))
2955
      || (MEM_P (dest)
2956 2957
	  && anti_dependence (info->equiv_mem, dest)))
    info->equiv_mem_modified = true;
2958 2959
}

Alan Modra committed
2960 2961
enum valid_equiv { valid_none, valid_combine, valid_reload };

2962 2963 2964 2965 2966
/* Verify that no store between START and the death of REG invalidates
   MEMREF.  MEMREF is invalidated by modifying a register used in MEMREF,
   by storing into an overlapping memory location, or with a non-const
   CALL_INSN.

Alan Modra committed
2967 2968 2969 2970
   Return VALID_RELOAD if MEMREF remains valid for both reload and
   combine_and_move insns, VALID_COMBINE if only valid for
   combine_and_move_insns, and VALID_NONE otherwise.  */
static enum valid_equiv
2971
validate_equiv_mem (rtx_insn *start, rtx reg, rtx memref)
2972
{
2973
  rtx_insn *insn;
2974
  rtx note;
2975
  struct equiv_mem_data info = { memref, false };
Alan Modra committed
2976
  enum valid_equiv ret = valid_reload;
2977 2978 2979 2980

  /* If the memory reference has side effects or is volatile, it isn't a
     valid equivalence.  */
  if (side_effects_p (memref))
Alan Modra committed
2981
    return valid_none;
2982

2983
  for (insn = start; insn; insn = NEXT_INSN (insn))
2984
    {
Alan Modra committed
2985
      if (!INSN_P (insn))
2986 2987 2988
	continue;

      if (find_reg_note (insn, REG_DEAD, reg))
Alan Modra committed
2989
	return ret;
2990

2991
      if (CALL_P (insn))
Alan Modra committed
2992 2993 2994 2995 2996 2997 2998 2999 3000 3001 3002 3003 3004 3005
	{
	  /* We can combine a reg def from one insn into a reg use in
	     another over a call if the memory is readonly or the call
	     const/pure.  However, we can't set reg_equiv notes up for
	     reload over any call.  The problem is the equivalent form
	     may reference a pseudo which gets assigned a call
	     clobbered hard reg.  When we later replace REG with its
	     equivalent form, the value in the call-clobbered reg has
	     been changed and all hell breaks loose.  */
	  ret = valid_combine;
	  if (!MEM_READONLY_P (memref)
	      && !RTL_CONST_OR_PURE_CALL_P (insn))
	    return valid_none;
	}
3006

3007 3008
      note_stores (PATTERN (insn), validate_equiv_mem_from_store, &info);
      if (info.equiv_mem_modified)
Alan Modra committed
3009
	return valid_none;
3010 3011 3012 3013 3014 3015 3016 3017 3018 3019 3020

      /* If a register mentioned in MEMREF is modified via an
	 auto-increment, we lose the equivalence.  Do the same if one
	 dies; although we could extend the life, it doesn't seem worth
	 the trouble.  */

      for (note = REG_NOTES (insn); note; note = XEXP (note, 1))
	if ((REG_NOTE_KIND (note) == REG_INC
	     || REG_NOTE_KIND (note) == REG_DEAD)
	    && REG_P (XEXP (note, 0))
	    && reg_overlap_mentioned_p (XEXP (note, 0), memref))
Alan Modra committed
3021
	  return valid_none;
3022 3023
    }

Alan Modra committed
3024
  return valid_none;
3025 3026 3027 3028 3029 3030 3031 3032 3033 3034 3035 3036 3037 3038 3039 3040
}

/* Returns zero if X is known to be invariant.  */
static int
equiv_init_varies_p (rtx x)
{
  RTX_CODE code = GET_CODE (x);
  int i;
  const char *fmt;

  switch (code)
    {
    case MEM:
      return !MEM_READONLY_P (x) || equiv_init_varies_p (XEXP (x, 0));

    case CONST:
3041
    CASE_CONST_ANY:
3042 3043 3044 3045 3046 3047 3048 3049 3050 3051 3052 3053 3054 3055 3056 3057 3058 3059 3060 3061 3062 3063 3064 3065 3066 3067 3068 3069 3070 3071 3072 3073 3074 3075 3076 3077 3078 3079 3080 3081 3082 3083 3084 3085 3086 3087 3088 3089 3090 3091 3092 3093 3094
    case SYMBOL_REF:
    case LABEL_REF:
      return 0;

    case REG:
      return reg_equiv[REGNO (x)].replace == 0 && rtx_varies_p (x, 0);

    case ASM_OPERANDS:
      if (MEM_VOLATILE_P (x))
	return 1;

      /* Fall through.  */

    default:
      break;
    }

  fmt = GET_RTX_FORMAT (code);
  for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
    if (fmt[i] == 'e')
      {
	if (equiv_init_varies_p (XEXP (x, i)))
	  return 1;
      }
    else if (fmt[i] == 'E')
      {
	int j;
	for (j = 0; j < XVECLEN (x, i); j++)
	  if (equiv_init_varies_p (XVECEXP (x, i, j)))
	    return 1;
      }

  return 0;
}

/* Returns nonzero if X (used to initialize register REGNO) is movable.
   X is only movable if the registers it uses have equivalent initializations
   which appear to be within the same loop (or in an inner loop) and movable
   or if they are not candidates for local_alloc and don't vary.  */
static int
equiv_init_movable_p (rtx x, int regno)
{
  int i, j;
  const char *fmt;
  enum rtx_code code = GET_CODE (x);

  switch (code)
    {
    case SET:
      return equiv_init_movable_p (SET_SRC (x), regno);

    case CC0:
    case CLOBBER:
3095
    case CLOBBER_HIGH:
3096 3097 3098 3099 3100 3101 3102 3103 3104 3105 3106
      return 0;

    case PRE_INC:
    case PRE_DEC:
    case POST_INC:
    case POST_DEC:
    case PRE_MODIFY:
    case POST_MODIFY:
      return 0;

    case REG:
Vladimir Makarov committed
3107 3108 3109 3110
      return ((reg_equiv[REGNO (x)].loop_depth >= reg_equiv[regno].loop_depth
	       && reg_equiv[REGNO (x)].replace)
	      || (REG_BASIC_BLOCK (REGNO (x)) < NUM_FIXED_BLOCKS
		  && ! rtx_varies_p (x, 0)));
3111 3112 3113 3114 3115 3116 3117 3118 3119 3120 3121 3122 3123 3124 3125 3126 3127 3128 3129 3130 3131 3132 3133 3134 3135 3136 3137 3138 3139 3140 3141 3142 3143 3144 3145 3146 3147 3148 3149 3150 3151 3152 3153 3154 3155 3156

    case UNSPEC_VOLATILE:
      return 0;

    case ASM_OPERANDS:
      if (MEM_VOLATILE_P (x))
	return 0;

      /* Fall through.  */

    default:
      break;
    }

  fmt = GET_RTX_FORMAT (code);
  for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
    switch (fmt[i])
      {
      case 'e':
	if (! equiv_init_movable_p (XEXP (x, i), regno))
	  return 0;
	break;
      case 'E':
	for (j = XVECLEN (x, i) - 1; j >= 0; j--)
	  if (! equiv_init_movable_p (XVECEXP (x, i, j), regno))
	    return 0;
	break;
      }

  return 1;
}

/* TRUE if X references a memory location that would be affected by a store
   to MEMREF.  */
static int
memref_referenced_p (rtx memref, rtx x)
{
  int i, j;
  const char *fmt;
  enum rtx_code code = GET_CODE (x);

  switch (code)
    {
    case CONST:
    case LABEL_REF:
    case SYMBOL_REF:
3157
    CASE_CONST_ANY:
3158 3159 3160 3161 3162 3163 3164 3165 3166 3167 3168 3169
    case PC:
    case CC0:
    case HIGH:
    case LO_SUM:
      return 0;

    case REG:
      return (reg_equiv[REGNO (x)].replacement
	      && memref_referenced_p (memref,
				      reg_equiv[REGNO (x)].replacement));

    case MEM:
3170
      if (true_dependence (memref, VOIDmode, x))
3171 3172 3173 3174 3175 3176 3177 3178 3179 3180 3181 3182 3183 3184 3185 3186 3187 3188 3189 3190 3191 3192 3193 3194 3195 3196 3197 3198 3199 3200 3201 3202 3203 3204 3205 3206 3207 3208 3209
	return 1;
      break;

    case SET:
      /* If we are setting a MEM, it doesn't count (its address does), but any
	 other SET_DEST that has a MEM in it is referencing the MEM.  */
      if (MEM_P (SET_DEST (x)))
	{
	  if (memref_referenced_p (memref, XEXP (SET_DEST (x), 0)))
	    return 1;
	}
      else if (memref_referenced_p (memref, SET_DEST (x)))
	return 1;

      return memref_referenced_p (memref, SET_SRC (x));

    default:
      break;
    }

  fmt = GET_RTX_FORMAT (code);
  for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
    switch (fmt[i])
      {
      case 'e':
	if (memref_referenced_p (memref, XEXP (x, i)))
	  return 1;
	break;
      case 'E':
	for (j = XVECLEN (x, i) - 1; j >= 0; j--)
	  if (memref_referenced_p (memref, XVECEXP (x, i, j)))
	    return 1;
	break;
      }

  return 0;
}

/* TRUE if some insn in the range (START, END] references a memory location
3210 3211 3212 3213 3214
   that would be affected by a store to MEMREF.

   Callers should not call this routine if START is after END in the
   RTL chain.  */

3215
static int
3216
memref_used_between_p (rtx memref, rtx_insn *start, rtx_insn *end)
3217
{
3218
  rtx_insn *insn;
3219

3220 3221
  for (insn = NEXT_INSN (start);
       insn && insn != NEXT_INSN (end);
3222 3223
       insn = NEXT_INSN (insn))
    {
3224
      if (!NONDEBUG_INSN_P (insn))
3225
	continue;
H.J. Lu committed
3226

3227 3228 3229 3230 3231 3232 3233 3234
      if (memref_referenced_p (memref, PATTERN (insn)))
	return 1;

      /* Nonconst functions may access memory.  */
      if (CALL_P (insn) && (! RTL_CONST_CALL_P (insn)))
	return 1;
    }

3235
  gcc_assert (insn == NEXT_INSN (end));
3236 3237 3238 3239 3240 3241 3242 3243 3244 3245 3246
  return 0;
}

/* Mark REG as having no known equivalence.
   Some instructions might have been processed before and furnished
   with REG_EQUIV notes for this register; these notes will have to be
   removed.
   STORE is the piece of RTL that does the non-constant / conflicting
   assignment - a SET, CLOBBER or REG_INC note.  It is currently not used,
   but needs to be there because this function is called from note_stores.  */
static void
Vladimir Makarov committed
3247 3248
no_equiv (rtx reg, const_rtx store ATTRIBUTE_UNUSED,
	  void *data ATTRIBUTE_UNUSED)
3249 3250
{
  int regno;
3251
  rtx_insn_list *list;
3252 3253 3254 3255

  if (!REG_P (reg))
    return;
  regno = REGNO (reg);
3256
  reg_equiv[regno].no_equiv = 1;
3257
  list = reg_equiv[regno].init_insns;
3258
  if (list && list->insn () == NULL)
3259
    return;
3260
  reg_equiv[regno].init_insns = gen_rtx_INSN_LIST (VOIDmode, NULL_RTX, NULL);
3261 3262 3263 3264 3265
  reg_equiv[regno].replacement = NULL_RTX;
  /* This doesn't matter for equivalences made for argument registers, we
     should keep their initialization insns.  */
  if (reg_equiv[regno].is_arg_equivalence)
    return;
3266
  ira_reg_equiv[regno].defined_p = false;
3267
  ira_reg_equiv[regno].init_insns = NULL;
3268
  for (; list; list = list->next ())
3269
    {
3270
      rtx_insn *insn = list->insn ();
3271 3272 3273 3274
      remove_note (insn, find_reg_note (insn, REG_EQUIV, NULL_RTX));
    }
}

3275 3276 3277
/* Check whether the SUBREG is a paradoxical subreg and set the result
   in PDX_SUBREGS.  */

3278
static void
3279
set_paradoxical_subreg (rtx_insn *insn)
3280
{
3281 3282 3283 3284 3285 3286 3287 3288
  subrtx_iterator::array_type array;
  FOR_EACH_SUBRTX (iter, array, PATTERN (insn), NONCONST)
    {
      const_rtx subreg = *iter;
      if (GET_CODE (subreg) == SUBREG)
	{
	  const_rtx reg = SUBREG_REG (subreg);
	  if (REG_P (reg) && paradoxical_subreg_p (subreg))
3289
	    reg_equiv[REGNO (reg)].pdx_subregs = true;
3290 3291
	}
    }
3292 3293
}

3294 3295 3296 3297 3298 3299 3300 3301 3302 3303
/* In DEBUG_INSN location adjust REGs from CLEARED_REGS bitmap to the
   equivalent replacement.  */

static rtx
adjust_cleared_regs (rtx loc, const_rtx old_rtx ATTRIBUTE_UNUSED, void *data)
{
  if (REG_P (loc))
    {
      bitmap cleared_regs = (bitmap) data;
      if (bitmap_bit_p (cleared_regs, REGNO (loc)))
3304
	return simplify_replace_fn_rtx (copy_rtx (*reg_equiv[REGNO (loc)].src_p),
3305 3306 3307 3308 3309
					NULL_RTX, adjust_cleared_regs, data);
    }
  return NULL_RTX;
}

3310 3311 3312 3313 3314 3315 3316 3317 3318 3319 3320 3321 3322 3323 3324 3325 3326 3327 3328 3329 3330 3331 3332 3333 3334 3335 3336 3337 3338 3339 3340 3341 3342 3343 3344 3345 3346 3347 3348 3349 3350 3351 3352
/* Given register REGNO is set only once, return true if the defining
   insn dominates all uses.  */

static bool
def_dominates_uses (int regno)
{
  df_ref def = DF_REG_DEF_CHAIN (regno);

  struct df_insn_info *def_info = DF_REF_INSN_INFO (def);
  /* If this is an artificial def (eh handler regs, hard frame pointer
     for non-local goto, regs defined on function entry) then def_info
     is NULL and the reg is always live before any use.  We might
     reasonably return true in that case, but since the only call
     of this function is currently here in ira.c when we are looking
     at a defining insn we can't have an artificial def as that would
     bump DF_REG_DEF_COUNT.  */
  gcc_assert (DF_REG_DEF_COUNT (regno) == 1 && def_info != NULL);

  rtx_insn *def_insn = DF_REF_INSN (def);
  basic_block def_bb = BLOCK_FOR_INSN (def_insn);

  for (df_ref use = DF_REG_USE_CHAIN (regno);
       use;
       use = DF_REF_NEXT_REG (use))
    {
      struct df_insn_info *use_info = DF_REF_INSN_INFO (use);
      /* Only check real uses, not artificial ones.  */
      if (use_info)
	{
	  rtx_insn *use_insn = DF_REF_INSN (use);
	  if (!DEBUG_INSN_P (use_insn))
	    {
	      basic_block use_bb = BLOCK_FOR_INSN (use_insn);
	      if (use_bb != def_bb
		  ? !dominated_by_p (CDI_DOMINATORS, use_bb, def_bb)
		  : DF_INSN_INFO_LUID (use_info) < DF_INSN_INFO_LUID (def_info))
		return false;
	    }
	}
    }
  return true;
}

3353
/* Find registers that are equivalent to a single value throughout the
Vladimir Makarov committed
3354 3355 3356
   compilation (either because they can be referenced in memory or are
   set once from a single constant).  Lower their priority for a
   register.
3357

Vladimir Makarov committed
3358 3359 3360
   If such a register is only referenced once, try substituting its
   value into the using insn.  If it succeeds, we can eliminate the
   register completely.
3361

3362 3363
   Initialize init_insns in ira_reg_equiv array.  */
static void
3364 3365
update_equiv_regs (void)
{
3366
  rtx_insn *insn;
3367 3368
  basic_block bb;

3369 3370
  /* Scan insns and set pdx_subregs if the reg is used in a
     paradoxical subreg.  Don't set such reg equivalent to a mem,
3371 3372
     because lra will not substitute such equiv memory in order to
     prevent access beyond allocated memory for paradoxical memory subreg.  */
3373
  FOR_EACH_BB_FN (bb, cfun)
3374
    FOR_BB_INSNS (bb, insn)
3375
      if (NONDEBUG_INSN_P (insn))
3376
	set_paradoxical_subreg (insn);
3377

3378 3379 3380
  /* Scan the insns and find which registers have equivalences.  Do this
     in a separate scan of the insns because (due to -fcse-follow-jumps)
     a register can be set below its use.  */
Alan Modra committed
3381
  bitmap setjmp_crosses = regstat_get_setjmp_crosses ();
3382
  FOR_EACH_BB_FN (bb, cfun)
3383
    {
Alan Modra committed
3384
      int loop_depth = bb_loop_depth (bb);
3385 3386 3387 3388 3389 3390 3391 3392 3393 3394 3395 3396 3397 3398 3399 3400 3401 3402 3403 3404 3405

      for (insn = BB_HEAD (bb);
	   insn != NEXT_INSN (BB_END (bb));
	   insn = NEXT_INSN (insn))
	{
	  rtx note;
	  rtx set;
	  rtx dest, src;
	  int regno;

	  if (! INSN_P (insn))
	    continue;

	  for (note = REG_NOTES (insn); note; note = XEXP (note, 1))
	    if (REG_NOTE_KIND (note) == REG_INC)
	      no_equiv (XEXP (note, 0), note, NULL);

	  set = single_set (insn);

	  /* If this insn contains more (or less) than a single SET,
	     only mark all destinations as having no known equivalence.  */
3406 3407
	  if (set == NULL_RTX
	      || side_effects_p (SET_SRC (set)))
3408 3409 3410 3411 3412 3413 3414 3415 3416 3417 3418 3419 3420 3421 3422 3423 3424 3425 3426 3427 3428 3429 3430 3431 3432 3433 3434
	    {
	      note_stores (PATTERN (insn), no_equiv, NULL);
	      continue;
	    }
	  else if (GET_CODE (PATTERN (insn)) == PARALLEL)
	    {
	      int i;

	      for (i = XVECLEN (PATTERN (insn), 0) - 1; i >= 0; i--)
		{
		  rtx part = XVECEXP (PATTERN (insn), 0, i);
		  if (part != set)
		    note_stores (part, no_equiv, NULL);
		}
	    }

	  dest = SET_DEST (set);
	  src = SET_SRC (set);

	  /* See if this is setting up the equivalence between an argument
	     register and its stack slot.  */
	  note = find_reg_note (insn, REG_EQUIV, NULL_RTX);
	  if (note)
	    {
	      gcc_assert (REG_P (dest));
	      regno = REGNO (dest);

3435 3436 3437
	      /* Note that we don't want to clear init_insns in
		 ira_reg_equiv even if there are multiple sets of this
		 register.  */
3438 3439
	      reg_equiv[regno].is_arg_equivalence = 1;

3440 3441 3442 3443 3444 3445
	      /* The insn result can have equivalence memory although
		 the equivalence is not set up by the insn.  We add
		 this insn to init insns as it is a flag for now that
		 regno has an equivalence.  We will remove the insn
		 from init insn list later.  */
	      if (rtx_equal_p (src, XEXP (note, 0)) || MEM_P (XEXP (note, 0)))
3446 3447 3448
		ira_reg_equiv[regno].init_insns
		  = gen_rtx_INSN_LIST (VOIDmode, insn,
				       ira_reg_equiv[regno].init_insns);
3449 3450 3451 3452 3453 3454 3455 3456 3457 3458

	      /* Continue normally in case this is a candidate for
		 replacements.  */
	    }

	  if (!optimize)
	    continue;

	  /* We only handle the case of a pseudo register being set
	     once, or always to the same value.  */
3459 3460 3461 3462 3463 3464 3465 3466 3467 3468
	  /* ??? The mn10200 port breaks if we add equivalences for
	     values that need an ADDRESS_REGS register and set them equivalent
	     to a MEM of a pseudo.  The actual problem is in the over-conservative
	     handling of INPADDR_ADDRESS / INPUT_ADDRESS / INPUT triples in
	     calculate_needs, but we traditionally work around this problem
	     here by rejecting equivalences when the destination is in a register
	     that's likely spilled.  This is fragile, of course, since the
	     preferred class of a pseudo depends on all instructions that set
	     or use it.  */

3469 3470
	  if (!REG_P (dest)
	      || (regno = REGNO (dest)) < FIRST_PSEUDO_REGISTER
3471 3472
	      || (reg_equiv[regno].init_insns
		  && reg_equiv[regno].init_insns->insn () == NULL)
3473
	      || (targetm.class_likely_spilled_p (reg_preferred_class (regno))
3474
		  && MEM_P (src) && ! reg_equiv[regno].is_arg_equivalence))
3475 3476 3477 3478 3479 3480 3481
	    {
	      /* This might be setting a SUBREG of a pseudo, a pseudo that is
		 also set somewhere else to a constant.  */
	      note_stores (set, no_equiv, NULL);
	      continue;
	    }

3482 3483 3484
	  /* Don't set reg mentioned in a paradoxical subreg
	     equivalent to a mem.  */
	  if (MEM_P (src) && reg_equiv[regno].pdx_subregs)
3485 3486 3487 3488 3489
	    {
	      note_stores (set, no_equiv, NULL);
	      continue;
	    }

3490 3491 3492 3493 3494 3495 3496 3497 3498
	  note = find_reg_note (insn, REG_EQUAL, NULL_RTX);

	  /* cse sometimes generates function invariants, but doesn't put a
	     REG_EQUAL note on the insn.  Since this note would be redundant,
	     there's no point creating it earlier than here.  */
	  if (! note && ! rtx_varies_p (src, 0))
	    note = set_unique_reg_note (insn, REG_EQUAL, copy_rtx (src));

	  /* Don't bother considering a REG_EQUAL note containing an EXPR_LIST
3499
	     since it represents a function call.  */
3500 3501 3502
	  if (note && GET_CODE (XEXP (note, 0)) == EXPR_LIST)
	    note = NULL_RTX;

3503 3504 3505 3506 3507 3508 3509 3510 3511 3512 3513
	  if (DF_REG_DEF_COUNT (regno) != 1)
	    {
	      bool equal_p = true;
	      rtx_insn_list *list;

	      /* If we have already processed this pseudo and determined it
		 can not have an equivalence, then honor that decision.  */
	      if (reg_equiv[regno].no_equiv)
		continue;

	      if (! note
3514 3515 3516
		  || rtx_varies_p (XEXP (note, 0), 0)
		  || (reg_equiv[regno].replacement
		      && ! rtx_equal_p (XEXP (note, 0),
3517 3518 3519 3520 3521 3522 3523 3524 3525 3526 3527 3528 3529 3530 3531 3532 3533 3534 3535 3536 3537 3538 3539 3540 3541 3542 3543
					reg_equiv[regno].replacement)))
		{
		  no_equiv (dest, set, NULL);
		  continue;
		}

	      list = reg_equiv[regno].init_insns;
	      for (; list; list = list->next ())
		{
		  rtx note_tmp;
		  rtx_insn *insn_tmp;

		  insn_tmp = list->insn ();
		  note_tmp = find_reg_note (insn_tmp, REG_EQUAL, NULL_RTX);
		  gcc_assert (note_tmp);
		  if (! rtx_equal_p (XEXP (note, 0), XEXP (note_tmp, 0)))
		    {
		      equal_p = false;
		      break;
		    }
		}

	      if (! equal_p)
		{
		  no_equiv (dest, set, NULL);
		  continue;
		}
3544
	    }
3545

3546 3547 3548 3549 3550
	  /* Record this insn as initializing this register.  */
	  reg_equiv[regno].init_insns
	    = gen_rtx_INSN_LIST (VOIDmode, insn, reg_equiv[regno].init_insns);

	  /* If this register is known to be equal to a constant, record that
3551 3552 3553 3554 3555 3556 3557
	     it is always equivalent to the constant.
	     Note that it is possible to have a register use before
	     the def in loops (see gcc.c-torture/execute/pr79286.c)
	     where the reg is undefined on first use.  If the def insn
	     won't trap we can use it as an equivalence, effectively
	     choosing the "undefined" value for the reg to be the
	     same as the value set by the def.  */
3558
	  if (DF_REG_DEF_COUNT (regno) == 1
3559 3560
	      && note
	      && !rtx_varies_p (XEXP (note, 0), 0)
3561 3562
	      && (!may_trap_or_fault_p (XEXP (note, 0))
		  || def_dominates_uses (regno)))
3563 3564 3565 3566 3567 3568 3569 3570 3571 3572 3573 3574 3575 3576 3577 3578 3579 3580 3581 3582 3583 3584
	    {
	      rtx note_value = XEXP (note, 0);
	      remove_note (insn, note);
	      set_unique_reg_note (insn, REG_EQUIV, note_value);
	    }

	  /* If this insn introduces a "constant" register, decrease the priority
	     of that register.  Record this insn if the register is only used once
	     more and the equivalence value is the same as our source.

	     The latter condition is checked for two reasons:  First, it is an
	     indication that it may be more efficient to actually emit the insn
	     as written (if no registers are available, reload will substitute
	     the equivalence).  Secondly, it avoids problems with any registers
	     dying in this insn whose death notes would be missed.

	     If we don't have a REG_EQUIV note, see if this insn is loading
	     a register used only in one basic block from a MEM.  If so, and the
	     MEM remains unchanged for the life of the register, add a REG_EQUIV
	     note.  */
	  note = find_reg_note (insn, REG_EQUIV, NULL_RTX);

Alan Modra committed
3585
	  rtx replacement = NULL_RTX;
3586
	  if (note)
Alan Modra committed
3587 3588 3589
	    replacement = XEXP (note, 0);
	  else if (REG_BASIC_BLOCK (regno) >= NUM_FIXED_BLOCKS
		   && MEM_P (SET_SRC (set)))
3590
	    {
Alan Modra committed
3591 3592 3593 3594 3595 3596 3597 3598 3599
	      enum valid_equiv validity;
	      validity = validate_equiv_mem (insn, dest, SET_SRC (set));
	      if (validity != valid_none)
		{
		  replacement = copy_rtx (SET_SRC (set));
		  if (validity == valid_reload)
		    note = set_unique_reg_note (insn, REG_EQUIV, replacement);
		}
	    }
3600

Alan Modra committed
3601 3602 3603 3604 3605 3606
	  /* If we haven't done so, record for reload that this is an
	     equivalencing insn.  */
	  if (note && !reg_equiv[regno].is_arg_equivalence)
	    ira_reg_equiv[regno].init_insns
	      = gen_rtx_INSN_LIST (VOIDmode, insn,
				   ira_reg_equiv[regno].init_insns);
3607

Alan Modra committed
3608 3609 3610
	  if (replacement)
	    {
	      reg_equiv[regno].replacement = replacement;
3611
	      reg_equiv[regno].src_p = &SET_SRC (set);
3612
	      reg_equiv[regno].loop_depth = (short) loop_depth;
3613 3614

	      /* Don't mess with things live during setjmp.  */
Alan Modra committed
3615
	      if (optimize && !bitmap_bit_p (setjmp_crosses, regno))
3616 3617 3618 3619 3620 3621 3622 3623 3624 3625 3626
		{
		  /* If the register is referenced exactly twice, meaning it is
		     set once and used once, indicate that the reference may be
		     replaced by the equivalence we computed above.  Do this
		     even if the register is only used in one block so that
		     dependencies can be handled where the last register is
		     used in a different block (i.e. HIGH / LO_SUM sequences)
		     and to reduce the number of registers alive across
		     calls.  */

		  if (REG_N_REFS (regno) == 2
Alan Modra committed
3627
		      && (rtx_equal_p (replacement, src)
3628 3629 3630 3631 3632 3633 3634 3635
			  || ! equiv_init_varies_p (src))
		      && NONJUMP_INSN_P (insn)
		      && equiv_init_movable_p (PATTERN (insn), regno))
		    reg_equiv[regno].replace = 1;
		}
	    }
	}
    }
3636
}
3637

3638 3639 3640 3641 3642 3643 3644 3645
/* For insns that set a MEM to the contents of a REG that is only used
   in a single basic block, see if the register is always equivalent
   to that memory location and if moving the store from INSN to the
   insn that sets REG is safe.  If so, put a REG_EQUIV note on the
   initializing insn.  */
static void
add_store_equivs (void)
{
3646
  auto_bitmap seen_insns;
3647

3648
  for (rtx_insn *insn = get_insns (); insn; insn = NEXT_INSN (insn))
3649 3650 3651
    {
      rtx set, src, dest;
      unsigned regno;
3652
      rtx_insn *init_insn;
3653

3654
      bitmap_set_bit (seen_insns, INSN_UID (insn));
3655

3656 3657 3658 3659 3660 3661 3662 3663 3664 3665
      if (! INSN_P (insn))
	continue;

      set = single_set (insn);
      if (! set)
	continue;

      dest = SET_DEST (set);
      src = SET_SRC (set);

3666
      /* Don't add a REG_EQUIV note if the insn already has one.  The existing
3667
	 REG_EQUIV is likely more useful than the one we are adding.  */
3668 3669 3670 3671
      if (MEM_P (dest) && REG_P (src)
	  && (regno = REGNO (src)) >= FIRST_PSEUDO_REGISTER
	  && REG_BASIC_BLOCK (regno) >= NUM_FIXED_BLOCKS
	  && DF_REG_DEF_COUNT (regno) == 1
3672
	  && ! reg_equiv[regno].pdx_subregs
3673
	  && reg_equiv[regno].init_insns != NULL
3674
	  && (init_insn = reg_equiv[regno].init_insns->insn ()) != 0
3675
	  && bitmap_bit_p (seen_insns, INSN_UID (init_insn))
3676
	  && ! find_reg_note (init_insn, REG_EQUIV, NULL_RTX)
Alan Modra committed
3677
	  && validate_equiv_mem (init_insn, src, dest) == valid_reload
3678 3679 3680 3681
	  && ! memref_used_between_p (dest, init_insn, insn)
	  /* Attaching a REG_EQUIV note will fail if INIT_INSN has
	     multiple sets.  */
	  && set_unique_reg_note (init_insn, REG_EQUIV, copy_rtx (dest)))
3682
	{
3683 3684 3685 3686 3687 3688 3689 3690 3691 3692
	  /* This insn makes the equivalence, not the one initializing
	     the register.  */
	  ira_reg_equiv[regno].init_insns
	    = gen_rtx_INSN_LIST (VOIDmode, insn, NULL_RTX);
	  df_notes_rescan (init_insn);
	  if (dump_file)
	    fprintf (dump_file,
		     "Adding REG_EQUIV to insn %d for source of insn %d\n",
		     INSN_UID (init_insn),
		     INSN_UID (insn));
3693 3694
	}
    }
3695 3696 3697 3698 3699 3700 3701 3702 3703 3704 3705 3706
}

/* Scan all regs killed in an insn to see if any of them are registers
   only used that once.  If so, see if we can replace the reference
   with the equivalent form.  If we can, delete the initializing
   reference and this register will go away.  If we can't replace the
   reference, and the initializing reference is within the same loop
   (or in an inner loop), then move the register initialization just
   before the use, so that they are in the same basic block.  */
static void
combine_and_move_insns (void)
{
Trevor Saunders committed
3707
  auto_bitmap cleared_regs;
3708
  int max = max_reg_num ();
3709

3710
  for (int regno = FIRST_PSEUDO_REGISTER; regno < max; regno++)
3711
    {
3712 3713
      if (!reg_equiv[regno].replace)
	continue;
3714

3715 3716 3717 3718 3719 3720 3721 3722 3723 3724 3725 3726
      rtx_insn *use_insn = 0;
      for (df_ref use = DF_REG_USE_CHAIN (regno);
	   use;
	   use = DF_REF_NEXT_REG (use))
	if (DF_REF_INSN_INFO (use))
	  {
	    if (DEBUG_INSN_P (DF_REF_INSN (use)))
	      continue;
	    gcc_assert (!use_insn);
	    use_insn = DF_REF_INSN (use);
	  }
      gcc_assert (use_insn);
3727

3728 3729 3730 3731 3732
      /* Don't substitute into jumps.  indirect_jump_optimize does
	 this for anything we are prepared to handle.  */
      if (JUMP_P (use_insn))
	continue;

3733 3734 3735 3736 3737
      /* Also don't substitute into a conditional trap insn -- it can become
	 an unconditional trap, and that is a flow control insn.  */
      if (GET_CODE (PATTERN (use_insn)) == TRAP_IF)
	continue;

3738 3739 3740 3741 3742 3743 3744 3745 3746 3747 3748 3749 3750 3751
      df_ref def = DF_REG_DEF_CHAIN (regno);
      gcc_assert (DF_REG_DEF_COUNT (regno) == 1 && DF_REF_INSN_INFO (def));
      rtx_insn *def_insn = DF_REF_INSN (def);

      /* We may not move instructions that can throw, since that
	 changes basic block boundaries and we are not prepared to
	 adjust the CFG to match.  */
      if (can_throw_internal (def_insn))
	continue;

      basic_block use_bb = BLOCK_FOR_INSN (use_insn);
      basic_block def_bb = BLOCK_FOR_INSN (def_insn);
      if (bb_loop_depth (use_bb) > bb_loop_depth (def_bb))
	continue;
3752

3753 3754 3755 3756 3757 3758 3759
      if (asm_noperands (PATTERN (def_insn)) < 0
	  && validate_replace_rtx (regno_reg_rtx[regno],
				   *reg_equiv[regno].src_p, use_insn))
	{
	  rtx link;
	  /* Append the REG_DEAD notes from def_insn.  */
	  for (rtx *p = &REG_NOTES (def_insn); (link = *p) != 0; )
3760
	    {
3761
	      if (REG_NOTE_KIND (XEXP (link, 0)) == REG_DEAD)
3762
		{
3763 3764 3765 3766 3767 3768 3769
		  *p = XEXP (link, 1);
		  XEXP (link, 1) = REG_NOTES (use_insn);
		  REG_NOTES (use_insn) = link;
		}
	      else
		p = &XEXP (link, 1);
	    }
3770

3771 3772 3773
	  remove_death (regno, use_insn);
	  SET_REG_N_REFS (regno, 0);
	  REG_FREQ (regno) = 0;
3774 3775 3776 3777 3778 3779 3780 3781
	  df_ref use;
	  FOR_EACH_INSN_USE (use, def_insn)
	    {
	      unsigned int use_regno = DF_REF_REGNO (use);
	      if (!HARD_REGISTER_NUM_P (use_regno))
		reg_equiv[use_regno].replace = 0;
	    }

3782
	  delete_insn (def_insn);
3783

3784 3785 3786 3787
	  reg_equiv[regno].init_insns = NULL;
	  ira_reg_equiv[regno].init_insns = NULL;
	  bitmap_set_bit (cleared_regs, regno);
	}
3788

3789 3790 3791 3792 3793
      /* Move the initialization of the register to just before
	 USE_INSN.  Update the flow information.  */
      else if (prev_nondebug_insn (use_insn) != def_insn)
	{
	  rtx_insn *new_insn;
3794

3795 3796 3797 3798 3799
	  new_insn = emit_insn_before (PATTERN (def_insn), use_insn);
	  REG_NOTES (new_insn) = REG_NOTES (def_insn);
	  REG_NOTES (def_insn) = 0;
	  /* Rescan it to process the notes.  */
	  df_insn_rescan (new_insn);
3800

3801 3802 3803
	  /* Make sure this insn is recognized before reload begins,
	     otherwise eliminate_regs_in_insn will die.  */
	  INSN_CODE (new_insn) = INSN_CODE (def_insn);
3804

3805
	  delete_insn (def_insn);
3806

3807
	  XEXP (reg_equiv[regno].init_insns, 0) = new_insn;
3808

3809 3810
	  REG_BASIC_BLOCK (regno) = use_bb->index;
	  REG_N_CALLS_CROSSED (regno) = 0;
3811

3812 3813
	  if (use_insn == BB_HEAD (use_bb))
	    BB_HEAD (use_bb) = new_insn;
3814

3815 3816 3817 3818 3819 3820 3821 3822 3823
	  /* We know regno dies in use_insn, but inside a loop
	     REG_DEAD notes might be missing when def_insn was in
	     another basic block.  However, when we move def_insn into
	     this bb we'll definitely get a REG_DEAD note and reload
	     will see the death.  It's possible that update_equiv_regs
	     set up an equivalence referencing regno for a reg set by
	     use_insn, when regno was seen as non-local.  Now that
	     regno is local to this block, and dies, such an
	     equivalence is invalid.  */
Alan Modra committed
3824
	  if (find_reg_note (use_insn, REG_EQUIV, regno_reg_rtx[regno]))
3825 3826 3827 3828 3829 3830
	    {
	      rtx set = single_set (use_insn);
	      if (set && REG_P (SET_DEST (set)))
		no_equiv (SET_DEST (set), set, NULL);
	    }

3831 3832 3833
	  ira_reg_equiv[regno].init_insns
	    = gen_rtx_INSN_LIST (VOIDmode, new_insn, NULL_RTX);
	  bitmap_set_bit (cleared_regs, regno);
3834 3835 3836 3837
	}
    }

  if (!bitmap_empty_p (cleared_regs))
3838
    {
3839 3840
      basic_block bb;

3841
      FOR_EACH_BB_FN (bb, cfun)
3842 3843 3844
	{
	  bitmap_and_compl_into (DF_LR_IN (bb), cleared_regs);
	  bitmap_and_compl_into (DF_LR_OUT (bb), cleared_regs);
3845
	  if (!df_live)
3846 3847 3848
	    continue;
	  bitmap_and_compl_into (DF_LIVE_IN (bb), cleared_regs);
	  bitmap_and_compl_into (DF_LIVE_OUT (bb), cleared_regs);
3849 3850 3851
	}

      /* Last pass - adjust debug insns referencing cleared regs.  */
3852
      if (MAY_HAVE_DEBUG_BIND_INSNS)
3853
	for (rtx_insn *insn = get_insns (); insn; insn = NEXT_INSN (insn))
3854
	  if (DEBUG_BIND_INSN_P (insn))
3855 3856 3857 3858 3859 3860 3861 3862 3863 3864
	    {
	      rtx old_loc = INSN_VAR_LOCATION_LOC (insn);
	      INSN_VAR_LOCATION_LOC (insn)
		= simplify_replace_fn_rtx (old_loc, NULL_RTX,
					   adjust_cleared_regs,
					   (void *) cleared_regs);
	      if (old_loc != INSN_VAR_LOCATION_LOC (insn))
		df_insn_rescan (insn);
	    }
    }
3865 3866
}

3867 3868
/* A pass over indirect jumps, converting simple cases to direct jumps.
   Combine does this optimization too, but only within a basic block.  */
3869 3870 3871 3872 3873 3874 3875 3876 3877
static void
indirect_jump_optimize (void)
{
  basic_block bb;
  bool rebuild_p = false;

  FOR_EACH_BB_REVERSE_FN (bb, cfun)
    {
      rtx_insn *insn = BB_END (bb);
3878 3879
      if (!JUMP_P (insn)
	  || find_reg_note (insn, REG_NON_LOCAL_GOTO, NULL_RTX))
3880 3881 3882 3883 3884 3885 3886 3887 3888
	continue;

      rtx x = pc_set (insn);
      if (!x || !REG_P (SET_SRC (x)))
	continue;

      int regno = REGNO (SET_SRC (x));
      if (DF_REG_DEF_COUNT (regno) == 1)
	{
3889 3890
	  df_ref def = DF_REG_DEF_CHAIN (regno);
	  if (!DF_REF_IS_ARTIFICIAL (def))
3891
	    {
3892
	      rtx_insn *def_insn = DF_REF_INSN (def);
3893 3894 3895 3896 3897
	      rtx lab = NULL_RTX;
	      rtx set = single_set (def_insn);
	      if (set && GET_CODE (SET_SRC (set)) == LABEL_REF)
		lab = SET_SRC (set);
	      else
3898
		{
3899 3900 3901
		  rtx eqnote = find_reg_note (def_insn, REG_EQUAL, NULL_RTX);
		  if (eqnote && GET_CODE (XEXP (eqnote, 0)) == LABEL_REF)
		    lab = XEXP (eqnote, 0);
3902
		}
3903 3904
	      if (lab && validate_replace_rtx (SET_SRC (x), lab, insn))
		rebuild_p = true;
3905 3906 3907
	    }
	}
    }
3908

3909 3910 3911 3912 3913 3914 3915 3916 3917 3918
  if (rebuild_p)
    {
      timevar_push (TV_JUMP);
      rebuild_jump_labels (get_insns ());
      if (purge_all_dead_edges ())
	delete_unreachable_blocks ();
      timevar_pop (TV_JUMP);
    }
}

3919 3920 3921 3922 3923 3924
/* Set up fields memory, constant, and invariant from init_insns in
   the structures of array ira_reg_equiv.  */
static void
setup_reg_equiv (void)
{
  int i;
3925 3926 3927
  rtx_insn_list *elem, *prev_elem, *next_elem;
  rtx_insn *insn;
  rtx set, x;
3928 3929

  for (i = FIRST_PSEUDO_REGISTER; i < ira_reg_equiv_len; i++)
3930 3931 3932
    for (prev_elem = NULL, elem = ira_reg_equiv[i].init_insns;
	 elem;
	 prev_elem = elem, elem = next_elem)
3933
      {
3934 3935
	next_elem = elem->next ();
	insn = elem->insn ();
3936 3937 3938 3939 3940 3941 3942
	set = single_set (insn);
	
	/* Init insns can set up equivalence when the reg is a destination or
	   a source (in this case the destination is memory).  */
	if (set != 0 && (REG_P (SET_DEST (set)) || REG_P (SET_SRC (set))))
	  {
	    if ((x = find_reg_note (insn, REG_EQUIV, NULL_RTX)) != NULL)
3943 3944 3945 3946 3947 3948 3949 3950 3951 3952 3953 3954 3955 3956 3957 3958
	      {
		x = XEXP (x, 0);
		if (REG_P (SET_DEST (set))
		    && REGNO (SET_DEST (set)) == (unsigned int) i
		    && ! rtx_equal_p (SET_SRC (set), x) && MEM_P (x))
		  {
		    /* This insn reporting the equivalence but
		       actually not setting it.  Remove it from the
		       list.  */
		    if (prev_elem == NULL)
		      ira_reg_equiv[i].init_insns = next_elem;
		    else
		      XEXP (prev_elem, 1) = next_elem;
		    elem = prev_elem;
		  }
	      }
3959 3960 3961 3962 3963 3964 3965 3966 3967 3968 3969 3970 3971 3972 3973 3974 3975 3976 3977 3978 3979 3980 3981 3982 3983 3984 3985 3986 3987
	    else if (REG_P (SET_DEST (set))
		     && REGNO (SET_DEST (set)) == (unsigned int) i)
	      x = SET_SRC (set);
	    else
	      {      
		gcc_assert (REG_P (SET_SRC (set))
			    && REGNO (SET_SRC (set)) == (unsigned int) i);
		x = SET_DEST (set);
	      }
	    if (! function_invariant_p (x)
		|| ! flag_pic
		/* A function invariant is often CONSTANT_P but may
		   include a register.  We promise to only pass
		   CONSTANT_P objects to LEGITIMATE_PIC_OPERAND_P.  */
		|| (CONSTANT_P (x) && LEGITIMATE_PIC_OPERAND_P (x)))
	      {
		/* It can happen that a REG_EQUIV note contains a MEM
		   that is not a legitimate memory operand.  As later
		   stages of reload assume that all addresses found in
		   the lra_regno_equiv_* arrays were originally
		   legitimate, we ignore such REG_EQUIV notes.  */
		if (memory_operand (x, VOIDmode))
		  {
		    ira_reg_equiv[i].defined_p = true;
		    ira_reg_equiv[i].memory = x;
		    continue;
		  }
		else if (function_invariant_p (x))
		  {
3988
		    machine_mode mode;
3989 3990 3991 3992 3993 3994 3995 3996 3997 3998 3999 4000 4001 4002 4003
		    
		    mode = GET_MODE (SET_DEST (set));
		    if (GET_CODE (x) == PLUS
			|| x == frame_pointer_rtx || x == arg_pointer_rtx)
		      /* This is PLUS of frame pointer and a constant,
			 or fp, or argp.  */
		      ira_reg_equiv[i].invariant = x;
		    else if (targetm.legitimate_constant_p (mode, x))
		      ira_reg_equiv[i].constant = x;
		    else
		      {
			ira_reg_equiv[i].memory = force_const_mem (mode, x);
			if (ira_reg_equiv[i].memory == NULL_RTX)
			  {
			    ira_reg_equiv[i].defined_p = false;
4004
			    ira_reg_equiv[i].init_insns = NULL;
4005 4006 4007 4008 4009 4010 4011 4012 4013
			    break;
			  }
		      }
		    ira_reg_equiv[i].defined_p = true;
		    continue;
		  }
	      }
	  }
	ira_reg_equiv[i].defined_p = false;
4014
	ira_reg_equiv[i].init_insns = NULL;
4015 4016 4017 4018 4019 4020
	break;
      }
}



4021 4022 4023 4024
/* Print chain C to FILE.  */
static void
print_insn_chain (FILE *file, struct insn_chain *c)
{
4025
  fprintf (file, "insn=%d, ", INSN_UID (c->insn));
4026 4027 4028 4029 4030 4031 4032 4033 4034 4035 4036 4037 4038 4039 4040 4041 4042 4043 4044 4045 4046
  bitmap_print (file, &c->live_throughout, "live_throughout: ", ", ");
  bitmap_print (file, &c->dead_or_set, "dead_or_set: ", "\n");
}


/* Print all reload_insn_chains to FILE.  */
static void
print_insn_chains (FILE *file)
{
  struct insn_chain *c;
  for (c = reload_insn_chain; c ; c = c->next)
    print_insn_chain (file, c);
}

/* Return true if pseudo REGNO should be added to set live_throughout
   or dead_or_set of the insn chains for reload consideration.  */
static bool
pseudo_for_reload_consideration_p (int regno)
{
  /* Consider spilled pseudos too for IRA because they still have a
     chance to get hard-registers in the reload when IRA is used.  */
4047
  return (reg_renumber[regno] >= 0 || ira_conflicts_p);
4048 4049
}

4050 4051 4052 4053 4054 4055 4056 4057 4058
/* Return true if we can track the individual bytes of subreg X.
   When returning true, set *OUTER_SIZE to the number of bytes in
   X itself, *INNER_SIZE to the number of bytes in the inner register
   and *START to the offset of the first byte.  */
static bool
get_subreg_tracking_sizes (rtx x, HOST_WIDE_INT *outer_size,
			   HOST_WIDE_INT *inner_size, HOST_WIDE_INT *start)
{
  rtx reg = regno_reg_rtx[REGNO (SUBREG_REG (x))];
4059 4060 4061
  return (GET_MODE_SIZE (GET_MODE (x)).is_constant (outer_size)
	  && GET_MODE_SIZE (GET_MODE (reg)).is_constant (inner_size)
	  && SUBREG_BYTE (x).is_constant (start));
4062 4063 4064 4065
}

/* Init LIVE_SUBREGS[ALLOCNUM] and LIVE_SUBREGS_USED[ALLOCNUM] for
   a register with SIZE bytes, making the register live if INIT_VALUE.  */
4066 4067
static void
init_live_subregs (bool init_value, sbitmap *live_subregs,
4068
		   bitmap live_subregs_used, int allocnum, int size)
4069 4070 4071 4072
{
  gcc_assert (size > 0);

  /* Been there, done that.  */
4073
  if (bitmap_bit_p (live_subregs_used, allocnum))
4074 4075
    return;

4076
  /* Create a new one.  */
4077 4078 4079 4080 4081 4082
  if (live_subregs[allocnum] == NULL)
    live_subregs[allocnum] = sbitmap_alloc (size);

  /* If the entire reg was live before blasting into subregs, we need
     to init all of the subregs to ones else init to 0.  */
  if (init_value)
4083
    bitmap_ones (live_subregs[allocnum]);
H.J. Lu committed
4084
  else
4085
    bitmap_clear (live_subregs[allocnum]);
4086

4087
  bitmap_set_bit (live_subregs_used, allocnum);
4088 4089 4090 4091 4092 4093 4094 4095 4096 4097 4098 4099
}

/* Walk the insns of the current function and build reload_insn_chain,
   and record register life information.  */
static void
build_insn_chain (void)
{
  unsigned int i;
  struct insn_chain **p = &reload_insn_chain;
  basic_block bb;
  struct insn_chain *c = NULL;
  struct insn_chain *next = NULL;
Trevor Saunders committed
4100 4101
  auto_bitmap live_relevant_regs;
  auto_bitmap elim_regset;
4102 4103 4104 4105
  /* live_subregs is a vector used to keep accurate information about
     which hardregs are live in multiword pseudos.  live_subregs and
     live_subregs_used are indexed by pseudo number.  The live_subreg
     entry for a particular pseudo is only used if the corresponding
4106 4107
     element is non zero in live_subregs_used.  The sbitmap size of
     live_subreg[allocno] is number of bytes that the pseudo can
4108 4109
     occupy.  */
  sbitmap *live_subregs = XCNEWVEC (sbitmap, max_regno);
Trevor Saunders committed
4110
  auto_bitmap live_subregs_used;
4111 4112 4113 4114

  for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
    if (TEST_HARD_REG_BIT (eliminable_regset, i))
      bitmap_set_bit (elim_regset, i);
4115
  FOR_EACH_BB_REVERSE_FN (bb, cfun)
4116 4117
    {
      bitmap_iterator bi;
4118
      rtx_insn *insn;
H.J. Lu committed
4119

4120
      CLEAR_REG_SET (live_relevant_regs);
4121
      bitmap_clear (live_subregs_used);
H.J. Lu committed
4122

4123
      EXECUTE_IF_SET_IN_BITMAP (df_get_live_out (bb), 0, i, bi)
4124 4125 4126 4127 4128 4129
	{
	  if (i >= FIRST_PSEUDO_REGISTER)
	    break;
	  bitmap_set_bit (live_relevant_regs, i);
	}

4130
      EXECUTE_IF_SET_IN_BITMAP (df_get_live_out (bb),
4131 4132 4133 4134 4135 4136 4137 4138 4139 4140
				FIRST_PSEUDO_REGISTER, i, bi)
	{
	  if (pseudo_for_reload_consideration_p (i))
	    bitmap_set_bit (live_relevant_regs, i);
	}

      FOR_BB_INSNS_REVERSE (bb, insn)
	{
	  if (!NOTE_P (insn) && !BARRIER_P (insn))
	    {
4141 4142
	      struct df_insn_info *insn_info = DF_INSN_INFO_GET (insn);
	      df_ref def, use;
4143 4144 4145 4146 4147 4148

	      c = new_insn_chain ();
	      c->next = next;
	      next = c;
	      *p = c;
	      p = &c->prev;
H.J. Lu committed
4149

4150 4151 4152
	      c->insn = insn;
	      c->block = bb->index;

4153
	      if (NONDEBUG_INSN_P (insn))
4154
		FOR_EACH_INSN_INFO_DEF (def, insn_info)
4155 4156
		  {
		    unsigned int regno = DF_REF_REGNO (def);
H.J. Lu committed
4157

4158 4159 4160 4161 4162 4163 4164 4165 4166 4167 4168 4169 4170 4171 4172 4173 4174 4175 4176 4177
		    /* Ignore may clobbers because these are generated
		       from calls. However, every other kind of def is
		       added to dead_or_set.  */
		    if (!DF_REF_FLAGS_IS_SET (def, DF_REF_MAY_CLOBBER))
		      {
			if (regno < FIRST_PSEUDO_REGISTER)
			  {
			    if (!fixed_regs[regno])
			      bitmap_set_bit (&c->dead_or_set, regno);
			  }
			else if (pseudo_for_reload_consideration_p (regno))
			  bitmap_set_bit (&c->dead_or_set, regno);
		      }

		    if ((regno < FIRST_PSEUDO_REGISTER
			 || reg_renumber[regno] >= 0
			 || ira_conflicts_p)
			&& (!DF_REF_FLAGS_IS_SET (def, DF_REF_CONDITIONAL)))
		      {
			rtx reg = DF_REF_REG (def);
4178 4179 4180 4181 4182 4183 4184 4185 4186
			HOST_WIDE_INT outer_size, inner_size, start;

			/* We can usually track the liveness of individual
			   bytes within a subreg.  The only exceptions are
			   subregs wrapped in ZERO_EXTRACTs and subregs whose
			   size is not known; in those cases we need to be
			   conservative and treat the definition as a partial
			   definition of the full register rather than a full
			   definition of a specific part of the register.  */
4187
			if (GET_CODE (reg) == SUBREG
4188 4189 4190
			    && !DF_REF_FLAGS_IS_SET (def, DF_REF_ZERO_EXTRACT)
			    && get_subreg_tracking_sizes (reg, &outer_size,
							  &inner_size, &start))
4191
			  {
4192
			    HOST_WIDE_INT last = start + outer_size;
4193 4194

			    init_live_subregs
H.J. Lu committed
4195
			      (bitmap_bit_p (live_relevant_regs, regno),
4196 4197
			       live_subregs, live_subregs_used, regno,
			       inner_size);
4198 4199 4200 4201 4202 4203 4204 4205 4206 4207 4208 4209 4210

			    if (!DF_REF_FLAGS_IS_SET
				(def, DF_REF_STRICT_LOW_PART))
			      {
				/* Expand the range to cover entire words.
				   Bytes added here are "don't care".  */
				start
				  = start / UNITS_PER_WORD * UNITS_PER_WORD;
				last = ((last + UNITS_PER_WORD - 1)
					/ UNITS_PER_WORD * UNITS_PER_WORD);
			      }

			    /* Ignore the paradoxical bits.  */
4211 4212
			    if (last > SBITMAP_SIZE (live_subregs[regno]))
			      last = SBITMAP_SIZE (live_subregs[regno]);
4213 4214 4215

			    while (start < last)
			      {
4216
				bitmap_clear_bit (live_subregs[regno], start);
4217 4218
				start++;
			      }
H.J. Lu committed
4219

4220
			    if (bitmap_empty_p (live_subregs[regno]))
4221
			      {
4222
				bitmap_clear_bit (live_subregs_used, regno);
4223 4224 4225 4226 4227 4228 4229 4230 4231 4232 4233 4234 4235 4236 4237 4238 4239
				bitmap_clear_bit (live_relevant_regs, regno);
			      }
			    else
			      /* Set live_relevant_regs here because
				 that bit has to be true to get us to
				 look at the live_subregs fields.  */
			      bitmap_set_bit (live_relevant_regs, regno);
			  }
			else
			  {
			    /* DF_REF_PARTIAL is generated for
			       subregs, STRICT_LOW_PART, and
			       ZERO_EXTRACT.  We handle the subreg
			       case above so here we have to keep from
			       modeling the def as a killing def.  */
			    if (!DF_REF_FLAGS_IS_SET (def, DF_REF_PARTIAL))
			      {
4240
				bitmap_clear_bit (live_subregs_used, regno);
4241 4242 4243 4244 4245
				bitmap_clear_bit (live_relevant_regs, regno);
			      }
			  }
		      }
		  }
H.J. Lu committed
4246

4247 4248 4249
	      bitmap_and_compl_into (live_relevant_regs, elim_regset);
	      bitmap_copy (&c->live_throughout, live_relevant_regs);

4250
	      if (NONDEBUG_INSN_P (insn))
4251
		FOR_EACH_INSN_INFO_USE (use, insn_info)
4252 4253 4254
		  {
		    unsigned int regno = DF_REF_REGNO (use);
		    rtx reg = DF_REF_REG (use);
H.J. Lu committed
4255

4256 4257 4258 4259 4260
		    /* DF_REF_READ_WRITE on a use means that this use
		       is fabricated from a def that is a partial set
		       to a multiword reg.  Here, we only model the
		       subreg case that is not wrapped in ZERO_EXTRACT
		       precisely so we do not need to look at the
4261
		       fabricated use.  */
H.J. Lu committed
4262 4263
		    if (DF_REF_FLAGS_IS_SET (use, DF_REF_READ_WRITE)
			&& !DF_REF_FLAGS_IS_SET (use, DF_REF_ZERO_EXTRACT)
4264 4265
			&& DF_REF_FLAGS_IS_SET (use, DF_REF_SUBREG))
		      continue;
H.J. Lu committed
4266

4267 4268 4269 4270 4271 4272 4273 4274 4275 4276 4277
		    /* Add the last use of each var to dead_or_set.  */
		    if (!bitmap_bit_p (live_relevant_regs, regno))
		      {
			if (regno < FIRST_PSEUDO_REGISTER)
			  {
			    if (!fixed_regs[regno])
			      bitmap_set_bit (&c->dead_or_set, regno);
			  }
			else if (pseudo_for_reload_consideration_p (regno))
			  bitmap_set_bit (&c->dead_or_set, regno);
		      }
H.J. Lu committed
4278

4279 4280 4281
		    if (regno < FIRST_PSEUDO_REGISTER
			|| pseudo_for_reload_consideration_p (regno))
		      {
4282
			HOST_WIDE_INT outer_size, inner_size, start;
4283 4284 4285
			if (GET_CODE (reg) == SUBREG
			    && !DF_REF_FLAGS_IS_SET (use,
						     DF_REF_SIGN_EXTRACT
4286 4287 4288
						     | DF_REF_ZERO_EXTRACT)
			    && get_subreg_tracking_sizes (reg, &outer_size,
							  &inner_size, &start))
4289
			  {
4290
			    HOST_WIDE_INT last = start + outer_size;
H.J. Lu committed
4291

4292
			    init_live_subregs
H.J. Lu committed
4293
			      (bitmap_bit_p (live_relevant_regs, regno),
4294 4295
			       live_subregs, live_subregs_used, regno,
			       inner_size);
H.J. Lu committed
4296

4297
			    /* Ignore the paradoxical bits.  */
4298 4299
			    if (last > SBITMAP_SIZE (live_subregs[regno]))
			      last = SBITMAP_SIZE (live_subregs[regno]);
4300 4301 4302

			    while (start < last)
			      {
4303
				bitmap_set_bit (live_subregs[regno], start);
4304 4305 4306 4307 4308 4309 4310 4311
				start++;
			      }
			  }
			else
			  /* Resetting the live_subregs_used is
			     effectively saying do not use the subregs
			     because we are reading the whole
			     pseudo.  */
4312
			  bitmap_clear_bit (live_subregs_used, regno);
4313 4314 4315 4316 4317 4318 4319 4320 4321 4322
			bitmap_set_bit (live_relevant_regs, regno);
		      }
		  }
	    }
	}

      /* FIXME!! The following code is a disaster.  Reload needs to see the
	 labels and jump tables that are just hanging out in between
	 the basic blocks.  See pr33676.  */
      insn = BB_HEAD (bb);
H.J. Lu committed
4323

4324
      /* Skip over the barriers and cruft.  */
H.J. Lu committed
4325
      while (insn && (BARRIER_P (insn) || NOTE_P (insn)
4326 4327
		      || BLOCK_FOR_INSN (insn) == bb))
	insn = PREV_INSN (insn);
H.J. Lu committed
4328

4329 4330 4331 4332 4333 4334 4335 4336 4337
      /* While we add anything except barriers and notes, the focus is
	 to get the labels and jump tables into the
	 reload_insn_chain.  */
      while (insn)
	{
	  if (!NOTE_P (insn) && !BARRIER_P (insn))
	    {
	      if (BLOCK_FOR_INSN (insn))
		break;
H.J. Lu committed
4338

4339 4340 4341 4342 4343
	      c = new_insn_chain ();
	      c->next = next;
	      next = c;
	      *p = c;
	      p = &c->prev;
H.J. Lu committed
4344

4345 4346 4347 4348 4349
	      /* The block makes no sense here, but it is what the old
		 code did.  */
	      c->block = bb->index;
	      c->insn = insn;
	      bitmap_copy (&c->live_throughout, live_relevant_regs);
H.J. Lu committed
4350
	    }
4351 4352 4353 4354 4355 4356 4357
	  insn = PREV_INSN (insn);
	}
    }

  reload_insn_chain = c;
  *p = NULL;

4358 4359 4360
  for (i = 0; i < (unsigned int) max_regno; i++)
    if (live_subregs[i] != NULL)
      sbitmap_free (live_subregs[i]);
4361 4362 4363 4364 4365
  free (live_subregs);

  if (dump_file)
    print_insn_chains (dump_file);
}
4366 4367 4368 4369 4370 4371 4372 4373 4374 4375 4376 4377 4378 4379 4380 4381 4382
 
/* Examine the rtx found in *LOC, which is read or written to as determined
   by TYPE.  Return false if we find a reason why an insn containing this
   rtx should not be moved (such as accesses to non-constant memory), true
   otherwise.  */
static bool
rtx_moveable_p (rtx *loc, enum op_type type)
{
  const char *fmt;
  rtx x = *loc;
  enum rtx_code code = GET_CODE (x);
  int i, j;

  code = GET_CODE (x);
  switch (code)
    {
    case CONST:
4383
    CASE_CONST_ANY:
4384 4385 4386 4387 4388 4389 4390 4391 4392 4393 4394 4395 4396 4397 4398 4399 4400 4401 4402 4403 4404 4405 4406 4407 4408 4409 4410 4411 4412 4413 4414 4415 4416 4417 4418 4419 4420
    case SYMBOL_REF:
    case LABEL_REF:
      return true;

    case PC:
      return type == OP_IN;

    case CC0:
      return false;

    case REG:
      if (x == frame_pointer_rtx)
	return true;
      if (HARD_REGISTER_P (x))
	return false;
      
      return true;

    case MEM:
      if (type == OP_IN && MEM_READONLY_P (x))
	return rtx_moveable_p (&XEXP (x, 0), OP_IN);
      return false;

    case SET:
      return (rtx_moveable_p (&SET_SRC (x), OP_IN)
	      && rtx_moveable_p (&SET_DEST (x), OP_OUT));

    case STRICT_LOW_PART:
      return rtx_moveable_p (&XEXP (x, 0), OP_OUT);

    case ZERO_EXTRACT:
    case SIGN_EXTRACT:
      return (rtx_moveable_p (&XEXP (x, 0), type)
	      && rtx_moveable_p (&XEXP (x, 1), OP_IN)
	      && rtx_moveable_p (&XEXP (x, 2), OP_IN));

    case CLOBBER:
4421
    case CLOBBER_HIGH:
4422 4423
      return rtx_moveable_p (&SET_DEST (x), OP_OUT);

4424
    case UNSPEC_VOLATILE:
4425
      /* It is a bad idea to consider insns with such rtl
4426 4427 4428 4429
	 as moveable ones.  The insn scheduler also considers them as barrier
	 for a reason.  */
      return false;

4430 4431 4432 4433 4434 4435
    case ASM_OPERANDS:
      /* The same is true for volatile asm: it has unknown side effects, it
         cannot be moved at will.  */
      if (MEM_VOLATILE_P (x))
	return false;

4436 4437 4438 4439 4440 4441 4442 4443 4444 4445 4446 4447 4448 4449 4450 4451 4452 4453 4454 4455 4456 4457 4458 4459 4460 4461 4462 4463 4464 4465 4466 4467 4468 4469 4470 4471 4472 4473 4474 4475 4476 4477
    default:
      break;
    }

  fmt = GET_RTX_FORMAT (code);
  for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
    {
      if (fmt[i] == 'e')
	{
	  if (!rtx_moveable_p (&XEXP (x, i), type))
	    return false;
	}
      else if (fmt[i] == 'E')
	for (j = XVECLEN (x, i) - 1; j >= 0; j--)
	  {
	    if (!rtx_moveable_p (&XVECEXP (x, i, j), type))
	      return false;
	  }
    }
  return true;
}

/* A wrapper around dominated_by_p, which uses the information in UID_LUID
   to give dominance relationships between two insns I1 and I2.  */
static bool
insn_dominated_by_p (rtx i1, rtx i2, int *uid_luid)
{
  basic_block bb1 = BLOCK_FOR_INSN (i1);
  basic_block bb2 = BLOCK_FOR_INSN (i2);

  if (bb1 == bb2)
    return uid_luid[INSN_UID (i2)] < uid_luid[INSN_UID (i1)];
  return dominated_by_p (CDI_DOMINATORS, bb1, bb2);
}

/* Record the range of register numbers added by find_moveable_pseudos.  */
int first_moveable_pseudo, last_moveable_pseudo;

/* These two vectors hold data for every register added by
   find_movable_pseudos, with index 0 holding data for the
   first_moveable_pseudo.  */
/* The original home register.  */
4478
static vec<rtx> pseudo_replaced_reg;
4479 4480 4481 4482 4483 4484 4485 4486 4487 4488 4489 4490 4491 4492 4493 4494 4495 4496 4497 4498 4499 4500 4501 4502 4503 4504 4505 4506 4507 4508

/* Look for instances where we have an instruction that is known to increase
   register pressure, and whose result is not used immediately.  If it is
   possible to move the instruction downwards to just before its first use,
   split its lifetime into two ranges.  We create a new pseudo to compute the
   value, and emit a move instruction just before the first use.  If, after
   register allocation, the new pseudo remains unallocated, the function
   move_unallocated_pseudos then deletes the move instruction and places
   the computation just before the first use.

   Such a move is safe and profitable if all the input registers remain live
   and unchanged between the original computation and its first use.  In such
   a situation, the computation is known to increase register pressure, and
   moving it is known to at least not worsen it.

   We restrict moves to only those cases where a register remains unallocated,
   in order to avoid interfering too much with the instruction schedule.  As
   an exception, we may move insns which only modify their input register
   (typically induction variables), as this increases the freedom for our
   intended transformation, and does not limit the second instruction
   scheduler pass.  */
   
static void
find_moveable_pseudos (void)
{
  unsigned i;
  int max_regs = max_reg_num ();
  int max_uid = get_max_uid ();
  basic_block bb;
  int *uid_luid = XNEWVEC (int, max_uid);
4509
  rtx_insn **closest_uses = XNEWVEC (rtx_insn *, max_regs);
4510
  /* A set of registers which are live but not modified throughout a block.  */
4511 4512
  bitmap_head *bb_transp_live = XNEWVEC (bitmap_head,
					 last_basic_block_for_fn (cfun));
4513
  /* A set of registers which only exist in a given basic block.  */
4514 4515
  bitmap_head *bb_local = XNEWVEC (bitmap_head,
				   last_basic_block_for_fn (cfun));
4516 4517
  /* A set of registers which are set once, in an instruction that can be
     moved freely downwards, but are otherwise transparent to a block.  */
4518 4519
  bitmap_head *bb_moveable_reg_sets = XNEWVEC (bitmap_head,
					       last_basic_block_for_fn (cfun));
4520
  auto_bitmap live, used, set, interesting, unusable_as_input;
4521 4522 4523
  bitmap_iterator bi;

  first_moveable_pseudo = max_regs;
4524 4525
  pseudo_replaced_reg.release ();
  pseudo_replaced_reg.safe_grow_cleared (max_regs);
4526

4527 4528 4529
  df_analyze ();
  calculate_dominance_info (CDI_DOMINATORS);

4530
  i = 0;
4531
  FOR_EACH_BB_FN (bb, cfun)
4532
    {
4533
      rtx_insn *insn;
4534 4535 4536 4537 4538 4539 4540
      bitmap transp = bb_transp_live + bb->index;
      bitmap moveable = bb_moveable_reg_sets + bb->index;
      bitmap local = bb_local + bb->index;

      bitmap_initialize (local, 0);
      bitmap_initialize (transp, 0);
      bitmap_initialize (moveable, 0);
4541 4542 4543
      bitmap_copy (live, df_get_live_out (bb));
      bitmap_and_into (live, df_get_live_in (bb));
      bitmap_copy (transp, live);
4544
      bitmap_clear (moveable);
4545 4546 4547
      bitmap_clear (live);
      bitmap_clear (used);
      bitmap_clear (set);
4548 4549 4550
      FOR_BB_INSNS (bb, insn)
	if (NONDEBUG_INSN_P (insn))
	  {
4551 4552
	    df_insn_info *insn_info = DF_INSN_INFO_GET (insn);
	    df_ref def, use;
4553 4554 4555

	    uid_luid[INSN_UID (insn)] = i++;
	    
4556 4557 4558 4559 4560
	    def = df_single_def (insn_info);
	    use = df_single_use (insn_info);
	    if (use
		&& def
		&& DF_REF_REGNO (use) == DF_REF_REGNO (def)
4561
		&& !bitmap_bit_p (set, DF_REF_REGNO (use))
4562 4563
		&& rtx_moveable_p (&PATTERN (insn), OP_IN))
	      {
4564
		unsigned regno = DF_REF_REGNO (use);
4565
		bitmap_set_bit (moveable, regno);
4566 4567
		bitmap_set_bit (set, regno);
		bitmap_set_bit (used, regno);
4568 4569 4570
		bitmap_clear_bit (transp, regno);
		continue;
	      }
4571
	    FOR_EACH_INSN_INFO_USE (use, insn_info)
4572
	      {
4573
		unsigned regno = DF_REF_REGNO (use);
4574
		bitmap_set_bit (used, regno);
4575 4576 4577 4578
		if (bitmap_clear_bit (moveable, regno))
		  bitmap_clear_bit (transp, regno);
	      }

4579
	    FOR_EACH_INSN_INFO_DEF (def, insn_info)
4580
	      {
4581
		unsigned regno = DF_REF_REGNO (def);
4582
		bitmap_set_bit (set, regno);
4583 4584 4585 4586 4587 4588
		bitmap_clear_bit (transp, regno);
		bitmap_clear_bit (moveable, regno);
	      }
	  }
    }

4589
  FOR_EACH_BB_FN (bb, cfun)
4590 4591
    {
      bitmap local = bb_local + bb->index;
4592
      rtx_insn *insn;
4593 4594 4595 4596

      FOR_BB_INSNS (bb, insn)
	if (NONDEBUG_INSN_P (insn))
	  {
4597
	    df_insn_info *insn_info = DF_INSN_INFO_GET (insn);
4598 4599
	    rtx_insn *def_insn;
	    rtx closest_use, note;
4600
	    df_ref def, use;
4601 4602
	    unsigned regno;
	    bool all_dominated, all_local;
4603
	    machine_mode mode;
4604

4605
	    def = df_single_def (insn_info);
4606
	    /* There must be exactly one def in this insn.  */
4607
	    if (!def || !single_set (insn))
4608 4609 4610 4611 4612 4613 4614 4615 4616
	      continue;
	    /* This must be the only definition of the reg.  We also limit
	       which modes we deal with so that we can assume we can generate
	       move instructions.  */
	    regno = DF_REF_REGNO (def);
	    mode = GET_MODE (DF_REF_REG (def));
	    if (DF_REG_DEF_COUNT (regno) != 1
		|| !DF_REF_INSN_INFO (def)
		|| HARD_REGISTER_NUM_P (regno)
4617
		|| DF_REG_EQ_USE_COUNT (regno) > 0
4618 4619 4620 4621 4622 4623 4624 4625 4626 4627 4628 4629 4630
		|| (!INTEGRAL_MODE_P (mode) && !FLOAT_MODE_P (mode)))
	      continue;
	    def_insn = DF_REF_INSN (def);

	    for (note = REG_NOTES (def_insn); note; note = XEXP (note, 1))
	      if (REG_NOTE_KIND (note) == REG_EQUIV && MEM_P (XEXP (note, 0)))
		break;
		
	    if (note)
	      {
		if (dump_file)
		  fprintf (dump_file, "Ignoring reg %d, has equiv memory\n",
			   regno);
4631
		bitmap_set_bit (unusable_as_input, regno);
4632 4633 4634 4635 4636 4637 4638 4639 4640
		continue;
	      }

	    use = DF_REG_USE_CHAIN (regno);
	    all_dominated = true;
	    all_local = true;
	    closest_use = NULL_RTX;
	    for (; use; use = DF_REF_NEXT_REG (use))
	      {
4641
		rtx_insn *insn;
4642 4643 4644 4645 4646 4647 4648 4649 4650 4651 4652 4653 4654 4655 4656 4657 4658 4659 4660 4661 4662 4663 4664 4665 4666 4667 4668 4669 4670 4671 4672 4673 4674 4675 4676 4677 4678 4679 4680 4681 4682
		if (!DF_REF_INSN_INFO (use))
		  {
		    all_dominated = false;
		    all_local = false;
		    break;
		  }
		insn = DF_REF_INSN (use);
		if (DEBUG_INSN_P (insn))
		  continue;
		if (BLOCK_FOR_INSN (insn) != BLOCK_FOR_INSN (def_insn))
		  all_local = false;
		if (!insn_dominated_by_p (insn, def_insn, uid_luid))
		  all_dominated = false;
		if (closest_use != insn && closest_use != const0_rtx)
		  {
		    if (closest_use == NULL_RTX)
		      closest_use = insn;
		    else if (insn_dominated_by_p (closest_use, insn, uid_luid))
		      closest_use = insn;
		    else if (!insn_dominated_by_p (insn, closest_use, uid_luid))
		      closest_use = const0_rtx;
		  }
	      }
	    if (!all_dominated)
	      {
		if (dump_file)
		  fprintf (dump_file, "Reg %d not all uses dominated by set\n",
			   regno);
		continue;
	      }
	    if (all_local)
	      bitmap_set_bit (local, regno);
	    if (closest_use == const0_rtx || closest_use == NULL
		|| next_nonnote_nondebug_insn (def_insn) == closest_use)
	      {
		if (dump_file)
		  fprintf (dump_file, "Reg %d uninteresting%s\n", regno,
			   closest_use == const0_rtx || closest_use == NULL
			   ? " (no unique first use)" : "");
		continue;
	      }
4683
	    if (HAVE_cc0 && reg_referenced_p (cc0_rtx, PATTERN (closest_use)))
4684 4685 4686 4687 4688 4689
	      {
		if (dump_file)
		  fprintf (dump_file, "Reg %d: closest user uses cc0\n",
			   regno);
		continue;
	      }
4690

4691
	    bitmap_set_bit (interesting, regno);
4692 4693 4694
	    /* If we get here, we know closest_use is a non-NULL insn
	       (as opposed to const_0_rtx).  */
	    closest_uses[regno] = as_a <rtx_insn *> (closest_use);
4695 4696 4697 4698 4699 4700 4701 4702 4703 4704 4705 4706 4707 4708 4709

	    if (dump_file && (all_local || all_dominated))
	      {
		fprintf (dump_file, "Reg %u:", regno);
		if (all_local)
		  fprintf (dump_file, " local to bb %d", bb->index);
		if (all_dominated)
		  fprintf (dump_file, " def dominates all uses");
		if (closest_use != const0_rtx)
		  fprintf (dump_file, " has unique first use");
		fputs ("\n", dump_file);
	      }
	  }
    }

4710
  EXECUTE_IF_SET_IN_BITMAP (interesting, 0, i, bi)
4711 4712
    {
      df_ref def = DF_REG_DEF_CHAIN (i);
4713
      rtx_insn *def_insn = DF_REF_INSN (def);
4714 4715 4716 4717 4718
      basic_block def_block = BLOCK_FOR_INSN (def_insn);
      bitmap def_bb_local = bb_local + def_block->index;
      bitmap def_bb_moveable = bb_moveable_reg_sets + def_block->index;
      bitmap def_bb_transp = bb_transp_live + def_block->index;
      bool local_to_bb_p = bitmap_bit_p (def_bb_local, i);
4719
      rtx_insn *use_insn = closest_uses[i];
4720
      df_ref use;
4721 4722 4723 4724 4725 4726 4727 4728 4729 4730 4731 4732 4733 4734 4735 4736 4737 4738 4739 4740 4741 4742 4743 4744 4745 4746 4747 4748 4749 4750
      bool all_ok = true;
      bool all_transp = true;

      if (!REG_P (DF_REF_REG (def)))
	continue;

      if (!local_to_bb_p)
	{
	  if (dump_file)
	    fprintf (dump_file, "Reg %u not local to one basic block\n",
		     i);
	  continue;
	}
      if (reg_equiv_init (i) != NULL_RTX)
	{
	  if (dump_file)
	    fprintf (dump_file, "Ignoring reg %u with equiv init insn\n",
		     i);
	  continue;
	}
      if (!rtx_moveable_p (&PATTERN (def_insn), OP_IN))
	{
	  if (dump_file)
	    fprintf (dump_file, "Found def insn %d for %d to be not moveable\n",
		     INSN_UID (def_insn), i);
	  continue;
	}
      if (dump_file)
	fprintf (dump_file, "Examining insn %d, def for %d\n",
		 INSN_UID (def_insn), i);
4751
      FOR_EACH_INSN_USE (use, def_insn)
4752 4753
	{
	  unsigned regno = DF_REF_REGNO (use);
4754
	  if (bitmap_bit_p (unusable_as_input, regno))
4755 4756 4757 4758 4759 4760 4761 4762 4763 4764
	    {
	      all_ok = false;
	      if (dump_file)
		fprintf (dump_file, "  found unusable input reg %u.\n", regno);
	      break;
	    }
	  if (!bitmap_bit_p (def_bb_transp, regno))
	    {
	      if (bitmap_bit_p (def_bb_moveable, regno)
		  && !control_flow_insn_p (use_insn)
4765
		  && (!HAVE_cc0 || !sets_cc0_p (use_insn)))
4766 4767 4768
		{
		  if (modified_between_p (DF_REF_REG (use), def_insn, use_insn))
		    {
4769
		      rtx_insn *x = NEXT_INSN (def_insn);
4770 4771 4772 4773 4774 4775 4776 4777 4778 4779 4780 4781 4782 4783 4784 4785 4786 4787 4788 4789 4790 4791 4792 4793 4794 4795 4796 4797 4798 4799 4800 4801 4802 4803
		      while (!modified_in_p (DF_REF_REG (use), x))
			{
			  gcc_assert (x != use_insn);
			  x = NEXT_INSN (x);
			}
		      if (dump_file)
			fprintf (dump_file, "  input reg %u modified but insn %d moveable\n",
				 regno, INSN_UID (x));
		      emit_insn_after (PATTERN (x), use_insn);
		      set_insn_deleted (x);
		    }
		  else
		    {
		      if (dump_file)
			fprintf (dump_file, "  input reg %u modified between def and use\n",
				 regno);
		      all_transp = false;
		    }
		}
	      else
		all_transp = false;
	    }
	}
      if (!all_ok)
	continue;
      if (!dbg_cnt (ira_move))
	break;
      if (dump_file)
	fprintf (dump_file, "  all ok%s\n", all_transp ? " and transp" : "");

      if (all_transp)
	{
	  rtx def_reg = DF_REF_REG (def);
	  rtx newreg = ira_create_new_reg (def_reg);
4804
	  if (validate_change (def_insn, DF_REF_REAL_LOC (def), newreg, 0))
4805 4806
	    {
	      unsigned nregno = REGNO (newreg);
4807
	      emit_insn_before (gen_move_insn (def_reg, newreg), use_insn);
4808
	      nregno -= max_regs;
4809
	      pseudo_replaced_reg[nregno] = def_reg;
4810 4811 4812 4813
	    }
	}
    }
  
4814
  FOR_EACH_BB_FN (bb, cfun)
4815 4816 4817 4818 4819 4820 4821 4822 4823 4824 4825 4826
    {
      bitmap_clear (bb_local + bb->index);
      bitmap_clear (bb_transp_live + bb->index);
      bitmap_clear (bb_moveable_reg_sets + bb->index);
    }
  free (uid_luid);
  free (closest_uses);
  free (bb_local);
  free (bb_transp_live);
  free (bb_moveable_reg_sets);

  last_moveable_pseudo = max_reg_num ();
4827 4828 4829 4830 4831 4832 4833 4834

  fix_reg_equiv_init ();
  expand_reg_info ();
  regstat_free_n_sets_and_refs ();
  regstat_free_ri ();
  regstat_init_n_sets_and_refs ();
  regstat_compute_ri ();
  free_dominance_info (CDI_DOMINATORS);
4835
}
4836

4837 4838 4839
/* If SET pattern SET is an assignment from a hard register to a pseudo which
   is live at CALL_DOM (if non-NULL, otherwise this check is omitted), return
   the destination.  Otherwise return NULL.  */
4840 4841

static rtx
4842
interesting_dest_for_shprep_1 (rtx set, basic_block call_dom)
4843 4844 4845 4846 4847 4848 4849 4850 4851 4852
{
  rtx src = SET_SRC (set);
  rtx dest = SET_DEST (set);
  if (!REG_P (src) || !HARD_REGISTER_P (src)
      || !REG_P (dest) || HARD_REGISTER_P (dest)
      || (call_dom && !bitmap_bit_p (df_get_live_in (call_dom), REGNO (dest))))
    return NULL;
  return dest;
}

Kito Cheng committed
4853
/* If insn is interesting for parameter range-splitting shrink-wrapping
4854 4855 4856 4857 4858 4859
   preparation, i.e. it is a single set from a hard register to a pseudo, which
   is live at CALL_DOM (if non-NULL, otherwise this check is omitted), or a
   parallel statement with only one such statement, return the destination.
   Otherwise return NULL.  */

static rtx
4860
interesting_dest_for_shprep (rtx_insn *insn, basic_block call_dom)
4861 4862 4863 4864 4865 4866 4867 4868 4869 4870 4871 4872 4873
{
  if (!INSN_P (insn))
    return NULL;
  rtx pat = PATTERN (insn);
  if (GET_CODE (pat) == SET)
    return interesting_dest_for_shprep_1 (pat, call_dom);

  if (GET_CODE (pat) != PARALLEL)
    return NULL;
  rtx ret = NULL;
  for (int i = 0; i < XVECLEN (pat, 0); i++)
    {
      rtx sub = XVECEXP (pat, 0, i);
4874 4875 4876
      if (GET_CODE (sub) == USE
	  || GET_CODE (sub) == CLOBBER
	  || GET_CODE (sub) == CLOBBER_HIGH)
4877 4878 4879 4880 4881 4882 4883 4884 4885 4886 4887 4888 4889
	continue;
      if (GET_CODE (sub) != SET
	  || side_effects_p (sub))
	return NULL;
      rtx dest = interesting_dest_for_shprep_1 (sub, call_dom);
      if (dest && ret)
	return NULL;
      if (dest)
	ret = dest;
    }
  return ret;
}

4890 4891 4892 4893 4894 4895 4896 4897 4898
/* Split live ranges of pseudos that are loaded from hard registers in the
   first BB in a BB that dominates all non-sibling call if such a BB can be
   found and is not in a loop.  Return true if the function has made any
   changes.  */

static bool
split_live_ranges_for_shrink_wrap (void)
{
  basic_block bb, call_dom = NULL;
4899
  basic_block first = single_succ (ENTRY_BLOCK_PTR_FOR_FN (cfun));
4900
  rtx_insn *insn, *last_interesting_insn = NULL;
4901
  auto_bitmap need_new, reachable;
4902 4903
  vec<basic_block> queue;

4904
  if (!SHRINK_WRAPPING_ENABLED)
4905 4906
    return false;

4907
  queue.create (n_basic_blocks_for_fn (cfun));
4908

4909
  FOR_EACH_BB_FN (bb, cfun)
4910 4911 4912 4913 4914 4915 4916 4917 4918
    FOR_BB_INSNS (bb, insn)
      if (CALL_P (insn) && !SIBLING_CALL_P (insn))
	{
	  if (bb == first)
	    {
	      queue.release ();
	      return false;
	    }

4919 4920
	  bitmap_set_bit (need_new, bb->index);
	  bitmap_set_bit (reachable, bb->index);
4921 4922 4923 4924 4925 4926 4927 4928 4929 4930 4931 4932 4933 4934 4935 4936 4937
	  queue.quick_push (bb);
	  break;
	}

  if (queue.is_empty ())
    {
      queue.release ();
      return false;
    }

  while (!queue.is_empty ())
    {
      edge e;
      edge_iterator ei;

      bb = queue.pop ();
      FOR_EACH_EDGE (e, ei, bb->succs)
4938
	if (e->dest != EXIT_BLOCK_PTR_FOR_FN (cfun)
4939
	    && bitmap_set_bit (reachable, e->dest->index))
4940 4941 4942 4943 4944 4945 4946 4947 4948 4949 4950
	  queue.quick_push (e->dest);
    }
  queue.release ();

  FOR_BB_INSNS (first, insn)
    {
      rtx dest = interesting_dest_for_shprep (insn, NULL);
      if (!dest)
	continue;

      if (DF_REG_DEF_COUNT (REGNO (dest)) > 1)
4951
	return false;
4952 4953 4954 4955 4956 4957

      for (df_ref use = DF_REG_USE_CHAIN (REGNO(dest));
	   use;
	   use = DF_REF_NEXT_REG (use))
	{
	  int ubbi = DF_REF_BB (use)->index;
4958 4959
	  if (bitmap_bit_p (reachable, ubbi))
	    bitmap_set_bit (need_new, ubbi);
4960 4961 4962 4963 4964
	}
      last_interesting_insn = insn;
    }

  if (!last_interesting_insn)
4965
    return false;
4966

4967
  call_dom = nearest_common_dominator_for_set (CDI_DOMINATORS, need_new);
4968 4969 4970 4971 4972 4973 4974 4975 4976 4977 4978 4979 4980 4981 4982 4983 4984 4985 4986 4987 4988 4989 4990 4991 4992 4993 4994
  if (call_dom == first)
    return false;

  loop_optimizer_init (AVOID_CFG_MODIFICATIONS);
  while (bb_loop_depth (call_dom) > 0)
    call_dom = get_immediate_dominator (CDI_DOMINATORS, call_dom);
  loop_optimizer_finalize ();

  if (call_dom == first)
    return false;

  calculate_dominance_info (CDI_POST_DOMINATORS);
  if (dominated_by_p (CDI_POST_DOMINATORS, first, call_dom))
    {
      free_dominance_info (CDI_POST_DOMINATORS);
      return false;
    }
  free_dominance_info (CDI_POST_DOMINATORS);

  if (dump_file)
    fprintf (dump_file, "Will split live ranges of parameters at BB %i\n",
	     call_dom->index);

  bool ret = false;
  FOR_BB_INSNS (first, insn)
    {
      rtx dest = interesting_dest_for_shprep (insn, call_dom);
4995
      if (!dest || dest == pic_offset_table_rtx)
4996 4997
	continue;

4998
      bool need_newreg = false;
4999
      df_ref use, next;
5000
      for (use = DF_REG_USE_CHAIN (REGNO (dest)); use; use = next)
5001
	{
5002
	  rtx_insn *uin = DF_REF_INSN (use);
5003 5004
	  next = DF_REF_NEXT_REG (use);

5005 5006 5007
	  if (DEBUG_INSN_P (uin))
	    continue;

5008 5009 5010 5011
	  basic_block ubb = BLOCK_FOR_INSN (uin);
	  if (ubb == call_dom
	      || dominated_by_p (CDI_DOMINATORS, ubb, call_dom))
	    {
5012 5013
	      need_newreg = true;
	      break;
5014 5015 5016
	    }
	}

5017
      if (need_newreg)
5018
	{
5019 5020 5021 5022 5023 5024 5025 5026 5027 5028 5029 5030 5031
	  rtx newreg = ira_create_new_reg (dest);

	  for (use = DF_REG_USE_CHAIN (REGNO (dest)); use; use = next)
	    {
	      rtx_insn *uin = DF_REF_INSN (use);
	      next = DF_REF_NEXT_REG (use);

	      basic_block ubb = BLOCK_FOR_INSN (uin);
	      if (ubb == call_dom
		  || dominated_by_p (CDI_DOMINATORS, ubb, call_dom))
		validate_change (uin, DF_REF_REAL_LOC (use), newreg, true);
	    }

5032
	  rtx_insn *new_move = gen_move_insn (newreg, dest);
5033 5034 5035 5036 5037 5038 5039 5040 5041 5042 5043 5044 5045 5046
	  emit_insn_after (new_move, bb_note (call_dom));
	  if (dump_file)
	    {
	      fprintf (dump_file, "Split live-range of register ");
	      print_rtl_single (dump_file, dest);
	    }
	  ret = true;
	}

      if (insn == last_interesting_insn)
	break;
    }
  apply_change_group ();
  return ret;
5047
}
5048

5049 5050 5051 5052 5053 5054 5055 5056 5057 5058 5059 5060 5061
/* Perform the second half of the transformation started in
   find_moveable_pseudos.  We look for instances where the newly introduced
   pseudo remains unallocated, and remove it by moving the definition to
   just before its use, replacing the move instruction generated by
   find_moveable_pseudos.  */
static void
move_unallocated_pseudos (void)
{
  int i;
  for (i = first_moveable_pseudo; i < last_moveable_pseudo; i++)
    if (reg_renumber[i] < 0)
      {
	int idx = i - first_moveable_pseudo;
5062
	rtx other_reg = pseudo_replaced_reg[idx];
5063
	rtx_insn *def_insn = DF_REF_INSN (DF_REG_DEF_CHAIN (i));
5064 5065 5066
	/* The use must follow all definitions of OTHER_REG, so we can
	   insert the new definition immediately after any of them.  */
	df_ref other_def = DF_REG_DEF_CHAIN (REGNO (other_reg));
5067 5068
	rtx_insn *move_insn = DF_REF_INSN (other_def);
	rtx_insn *newinsn = emit_insn_after (PATTERN (def_insn), move_insn);
5069
	rtx set;
5070 5071 5072 5073 5074 5075
	int success;

	if (dump_file)
	  fprintf (dump_file, "moving def of %d (insn %d now) ",
		   REGNO (other_reg), INSN_UID (def_insn));

5076 5077 5078 5079 5080
	delete_insn (move_insn);
	while ((other_def = DF_REG_DEF_CHAIN (REGNO (other_reg))))
	  delete_insn (DF_REF_INSN (other_def));
	delete_insn (def_insn);

5081 5082 5083 5084 5085 5086 5087 5088 5089
	set = single_set (newinsn);
	success = validate_change (newinsn, &SET_DEST (set), other_reg, 0);
	gcc_assert (success);
	if (dump_file)
	  fprintf (dump_file, " %d) rather than keep unallocated replacement %d\n",
		   INSN_UID (newinsn), i);
	SET_REG_N_REFS (i, 0);
      }
}
5090

Steven Bosscher committed
5091 5092
/* If the backend knows where to allocate pseudos for hard
   register initial values, register these allocations now.  */
5093
static void
Steven Bosscher committed
5094 5095 5096 5097 5098 5099 5100 5101 5102 5103 5104 5105 5106 5107 5108 5109 5110 5111 5112 5113 5114 5115 5116 5117 5118 5119 5120 5121 5122 5123
allocate_initial_values (void)
{
  if (targetm.allocate_initial_value)
    {
      rtx hreg, preg, x;
      int i, regno;

      for (i = 0; HARD_REGISTER_NUM_P (i); i++)
	{
	  if (! initial_value_entry (i, &hreg, &preg))
	    break;

	  x = targetm.allocate_initial_value (hreg);
	  regno = REGNO (preg);
	  if (x && REG_N_SETS (regno) <= 1)
	    {
	      if (MEM_P (x))
		reg_equiv_memory_loc (regno) = x;
	      else
		{
		  basic_block bb;
		  int new_regno;

		  gcc_assert (REG_P (x));
		  new_regno = REGNO (x);
		  reg_renumber[regno] = new_regno;
		  /* Poke the regno right into regno_reg_rtx so that even
		     fixed regs are accepted.  */
		  SET_REGNO (preg, new_regno);
		  /* Update global register liveness information.  */
5124
		  FOR_EACH_BB_FN (bb, cfun)
Steven Bosscher committed
5125
		    {
5126
		      if (REGNO_REG_SET_P (df_get_live_in (bb), regno))
Steven Bosscher committed
5127
			SET_REGNO_REG_SET (df_get_live_in (bb), new_regno);
5128
		      if (REGNO_REG_SET_P (df_get_live_out (bb), regno))
Steven Bosscher committed
5129 5130 5131 5132 5133
			SET_REGNO_REG_SET (df_get_live_out (bb), new_regno);
		    }
		}
	    }
	}
5134

Steven Bosscher committed
5135 5136 5137 5138 5139
      gcc_checking_assert (! initial_value_entry (FIRST_PSEUDO_REGISTER,
						  &hreg, &preg));
    }
}

5140 5141 5142 5143 5144

/* True when we use LRA instead of reload pass for the current
   function.  */
bool ira_use_lra_p;

5145 5146 5147 5148
/* True if we have allocno conflicts.  It is false for non-optimized
   mode or when the conflict table is too big.  */
bool ira_conflicts_p;

5149 5150 5151
/* Saved between IRA and reload.  */
static int saved_flag_ira_share_spill_slots;

Vladimir Makarov committed
5152 5153 5154 5155 5156
/* This is the main entry of IRA.  */
static void
ira (FILE *f)
{
  bool loops_p;
5157
  int ira_max_point_before_emit;
5158 5159 5160
  bool saved_flag_caller_saves = flag_caller_saves;
  enum ira_region saved_flag_ira_region = flag_ira_region;

5161 5162
  clear_bb_flags ();

5163 5164 5165 5166 5167 5168 5169
  /* Determine if the current function is a leaf before running IRA
     since this can impact optimizations done by the prologue and
     epilogue thus changing register elimination offsets.
     Other target callbacks may use crtl->is_leaf too, including
     SHRINK_WRAPPING_ENABLED, so initialize as early as possible.  */
  crtl->is_leaf = leaf_function_p ();

5170 5171 5172
  /* Perform target specific PIC register initialization.  */
  targetm.init_pic_reg ();

5173 5174 5175 5176 5177 5178
  ira_conflicts_p = optimize > 0;

  /* If there are too many pseudos and/or basic blocks (e.g. 10K
     pseudos and 10K blocks or 100K pseudos and 1K blocks), we will
     use simplified and faster algorithms in LRA.  */
  lra_simple_p
5179 5180
    = (ira_use_lra_p
       && max_reg_num () >= (1 << 26) / last_basic_block_for_fn (cfun));
5181 5182 5183 5184 5185 5186 5187 5188 5189 5190 5191 5192 5193 5194
  if (lra_simple_p)
    {
      /* It permits to skip live range splitting in LRA.  */
      flag_caller_saves = false;
      /* There is no sense to do regional allocation when we use
	 simplified LRA.  */
      flag_ira_region = IRA_REGION_ONE;
      ira_conflicts_p = false;
    }

#ifndef IRA_NO_OBSTACK
  gcc_obstack_init (&ira_obstack);
#endif
  bitmap_obstack_initialize (&ira_bitmap_obstack);
Vladimir Makarov committed
5195

5196 5197
  /* LRA uses its own infrastructure to handle caller save registers.  */
  if (flag_caller_saves && !ira_use_lra_p)
5198 5199
    init_caller_save ();

Vladimir Makarov committed
5200 5201 5202 5203 5204 5205 5206 5207 5208 5209 5210 5211
  if (flag_ira_verbose < 10)
    {
      internal_flag_ira_verbose = flag_ira_verbose;
      ira_dump_file = f;
    }
  else
    {
      internal_flag_ira_verbose = flag_ira_verbose - 10;
      ira_dump_file = stderr;
    }

  setup_prohibited_mode_move_regs ();
Vladimir Makarov committed
5212
  decrease_live_ranges_number ();
Vladimir Makarov committed
5213
  df_note_add_problem ();
5214 5215 5216 5217 5218 5219 5220 5221 5222 5223

  /* DF_LIVE can't be used in the register allocator, too many other
     parts of the compiler depend on using the "classic" liveness
     interpretation of the DF_LR problem.  See PR38711.
     Remove the problem, so that we don't spend time updating it in
     any of the df_analyze() calls during IRA/LRA.  */
  if (optimize > 1)
    df_remove_problem (df_live);
  gcc_checking_assert (df_live == NULL);

5224 5225 5226
  if (flag_checking)
    df->changeable_flags |= DF_VERIFY_SCHEDULED;

Vladimir Makarov committed
5227
  df_analyze ();
Vladimir Makarov committed
5228

5229 5230 5231 5232 5233 5234 5235 5236 5237 5238 5239
  init_reg_equiv ();
  if (ira_conflicts_p)
    {
      calculate_dominance_info (CDI_DOMINATORS);

      if (split_live_ranges_for_shrink_wrap ())
	df_analyze ();

      free_dominance_info (CDI_DOMINATORS);
    }

Vladimir Makarov committed
5240
  df_clear_flags (DF_NO_INSN_RESCAN);
5241

5242 5243 5244 5245
  indirect_jump_optimize ();
  if (delete_trivially_dead_insns (get_insns (), max_reg_num ()))
    df_analyze ();

Vladimir Makarov committed
5246 5247 5248 5249 5250 5251 5252 5253 5254
  regstat_init_n_sets_and_refs ();
  regstat_compute_ri ();

  /* If we are not optimizing, then this is the only place before
     register allocation where dataflow is done.  And that is needed
     to generate these warnings.  */
  if (warn_clobbered)
    generate_setjmp_warnings ();

5255
  if (resize_reg_info () && flag_ira_loop_pressure)
5256
    ira_set_pseudo_classes (true, ira_dump_file);
5257

5258
  init_alias_analysis ();
Alan Modra committed
5259
  loop_optimizer_init (AVOID_CFG_MODIFICATIONS);
5260
  reg_equiv = XCNEWVEC (struct equivalence, max_reg_num ());
5261
  update_equiv_regs ();
5262 5263 5264 5265 5266 5267 5268 5269 5270 5271

  /* Don't move insns if live range shrinkage or register
     pressure-sensitive scheduling were done because it will not
     improve allocation but likely worsen insn scheduling.  */
  if (optimize
      && !flag_live_range_shrinkage
      && !(flag_sched_pressure && flag_schedule_insns))
    combine_and_move_insns ();

  /* Gather additional equivalences with memory.  */
5272
  if (optimize)
5273 5274
    add_store_equivs ();

Alan Modra committed
5275
  loop_optimizer_finalize ();
Alan Modra committed
5276
  free_dominance_info (CDI_DOMINATORS);
5277 5278 5279
  end_alias_analysis ();
  free (reg_equiv);

5280
  setup_reg_equiv ();
5281
  grow_reg_equivs ();
5282
  setup_reg_equiv_init ();
Vladimir Makarov committed
5283

5284
  allocated_reg_info_size = max_reg_num ();
5285 5286 5287 5288 5289

  /* It is not worth to do such improvement when we use a simple
     allocation because of -O0 usage or because the function is too
     big.  */
  if (ira_conflicts_p)
5290
    find_moveable_pseudos ();
5291

5292
  max_regno_before_ira = max_reg_num ();
5293
  ira_setup_eliminable_regset ();
H.J. Lu committed
5294

Vladimir Makarov committed
5295 5296 5297
  ira_overall_cost = ira_reg_cost = ira_mem_cost = 0;
  ira_load_cost = ira_store_cost = ira_shuffle_cost = 0;
  ira_move_loops_num = ira_additional_jumps_num = 0;
H.J. Lu committed
5298

Vladimir Makarov committed
5299
  ira_assert (current_loops == NULL);
5300
  if (flag_ira_region == IRA_REGION_ALL || flag_ira_region == IRA_REGION_MIXED)
5301
    loop_optimizer_init (AVOID_CFG_MODIFICATIONS | LOOPS_HAVE_RECORDED_EXITS);
H.J. Lu committed
5302

Vladimir Makarov committed
5303 5304
  if (internal_flag_ira_verbose > 0 && ira_dump_file != NULL)
    fprintf (ira_dump_file, "Building IRA IR\n");
5305
  loops_p = ira_build ();
H.J. Lu committed
5306

5307
  ira_assert (ira_conflicts_p || !loops_p);
5308 5309

  saved_flag_ira_share_spill_slots = flag_ira_share_spill_slots;
5310
  if (too_high_register_pressure_p () || cfun->calls_setjmp)
5311
    /* It is just wasting compiler's time to pack spilled pseudos into
5312 5313 5314 5315
       stack slots in this case -- prohibit it.  We also do this if
       there is setjmp call because a variable not modified between
       setjmp and longjmp the compiler is required to preserve its
       value and sharing slots does not guarantee it.  */
5316 5317
    flag_ira_share_spill_slots = FALSE;

5318
  ira_color ();
H.J. Lu committed
5319

Vladimir Makarov committed
5320
  ira_max_point_before_emit = ira_max_point;
H.J. Lu committed
5321

Vladimir Makarov committed
5322 5323
  ira_initiate_emit_data ();

Vladimir Makarov committed
5324
  ira_emit (loops_p);
H.J. Lu committed
5325

5326
  max_regno = max_reg_num ();
5327
  if (ira_conflicts_p)
Vladimir Makarov committed
5328 5329
    {
      if (! loops_p)
5330 5331 5332 5333
	{
	  if (! ira_use_lra_p)
	    ira_initiate_assign ();
	}
Vladimir Makarov committed
5334 5335
      else
	{
5336
	  expand_reg_info ();
H.J. Lu committed
5337

5338 5339 5340 5341 5342 5343
	  if (ira_use_lra_p)
	    {
	      ira_allocno_t a;
	      ira_allocno_iterator ai;

	      FOR_EACH_ALLOCNO (a, ai)
5344 5345 5346 5347 5348 5349 5350 5351 5352 5353 5354
                {
                  int old_regno = ALLOCNO_REGNO (a);
                  int new_regno = REGNO (ALLOCNO_EMIT_DATA (a)->reg);

                  ALLOCNO_REGNO (a) = new_regno;

                  if (old_regno != new_regno)
                    setup_reg_classes (new_regno, reg_preferred_class (old_regno),
                                       reg_alternate_class (old_regno),
                                       reg_allocno_class (old_regno));
                }
5355 5356 5357 5358 5359 5360 5361
	    }
	  else
	    {
	      if (internal_flag_ira_verbose > 0 && ira_dump_file != NULL)
		fprintf (ira_dump_file, "Flattening IR\n");
	      ira_flattening (max_regno_before_ira, ira_max_point_before_emit);
	    }
Vladimir Makarov committed
5362 5363 5364
	  /* New insns were generated: add notes and recalculate live
	     info.  */
	  df_analyze ();
H.J. Lu committed
5365

5366 5367 5368
	  /* ??? Rebuild the loop tree, but why?  Does the loop tree
	     change if new insns were generated?  Can that be handled
	     by updating the loop tree incrementally?  */
5369
	  loop_optimizer_finalize ();
5370
	  free_dominance_info (CDI_DOMINATORS);
5371 5372
	  loop_optimizer_init (AVOID_CFG_MODIFICATIONS
			       | LOOPS_HAVE_RECORDED_EXITS);
Vladimir Makarov committed
5373

5374 5375 5376 5377 5378 5379
	  if (! ira_use_lra_p)
	    {
	      setup_allocno_assignment_flags ();
	      ira_initiate_assign ();
	      ira_reassign_conflict_allocnos (max_regno);
	    }
Vladimir Makarov committed
5380 5381 5382
	}
    }

Vladimir Makarov committed
5383 5384
  ira_finish_emit_data ();

Vladimir Makarov committed
5385
  setup_reg_renumber ();
H.J. Lu committed
5386

Vladimir Makarov committed
5387
  calculate_allocation_cost ();
H.J. Lu committed
5388

Vladimir Makarov committed
5389
#ifdef ENABLE_IRA_CHECKING
5390 5391 5392 5393 5394 5395 5396 5397
  if (ira_conflicts_p && ! ira_use_lra_p)
    /* Opposite to reload pass, LRA does not use any conflict info
       from IRA.  We don't rebuild conflict info for LRA (through
       ira_flattening call) and can not use the check here.  We could
       rebuild this info for LRA in the check mode but there is a risk
       that code generated with the check and without it will be a bit
       different.  Calling ira_flattening in any mode would be a
       wasting CPU time.  So do not check the allocation for LRA.  */
Vladimir Makarov committed
5398 5399
    check_allocation ();
#endif
H.J. Lu committed
5400

5401
  if (max_regno != max_regno_before_ira)
Vladimir Makarov committed
5402 5403 5404 5405 5406 5407 5408 5409
    {
      regstat_free_n_sets_and_refs ();
      regstat_free_ri ();
      regstat_init_n_sets_and_refs ();
      regstat_compute_ri ();
    }

  overall_cost_before = ira_overall_cost;
5410 5411 5412
  if (! ira_conflicts_p)
    grow_reg_equivs ();
  else
Vladimir Makarov committed
5413 5414
    {
      fix_reg_equiv_init ();
H.J. Lu committed
5415

Vladimir Makarov committed
5416 5417 5418
#ifdef ENABLE_IRA_CHECKING
      print_redundant_copies ();
#endif
5419 5420 5421 5422 5423 5424 5425
      if (! ira_use_lra_p)
	{
	  ira_spilled_reg_stack_slots_num = 0;
	  ira_spilled_reg_stack_slots
	    = ((struct ira_spilled_reg_stack_slot *)
	       ira_allocate (max_regno
			     * sizeof (struct ira_spilled_reg_stack_slot)));
5426
	  memset ((void *)ira_spilled_reg_stack_slots, 0,
5427 5428
		  max_regno * sizeof (struct ira_spilled_reg_stack_slot));
	}
Vladimir Makarov committed
5429
    }
Steven Bosscher committed
5430
  allocate_initial_values ();
5431 5432 5433 5434

  /* See comment for find_moveable_pseudos call.  */
  if (ira_conflicts_p)
    move_unallocated_pseudos ();
5435 5436 5437 5438 5439 5440 5441

  /* Restore original values.  */
  if (lra_simple_p)
    {
      flag_caller_saves = saved_flag_caller_saves;
      flag_ira_region = saved_flag_ira_region;
    }
5442 5443 5444 5445 5446 5447 5448
}

static void
do_reload (void)
{
  basic_block bb;
  bool need_dce;
5449
  unsigned pic_offset_table_regno = INVALID_REGNUM;
5450

5451
  if (flag_ira_verbose < 10)
5452
    ira_dump_file = dump_file;
Vladimir Makarov committed
5453

5454 5455 5456 5457 5458 5459 5460
  /* If pic_offset_table_rtx is a pseudo register, then keep it so
     after reload to avoid possible wrong usages of hard reg assigned
     to it.  */
  if (pic_offset_table_rtx
      && REGNO (pic_offset_table_rtx) >= FIRST_PSEUDO_REGISTER)
    pic_offset_table_regno = REGNO (pic_offset_table_rtx);

5461 5462 5463 5464 5465
  timevar_push (TV_RELOAD);
  if (ira_use_lra_p)
    {
      if (current_loops != NULL)
	{
5466
	  loop_optimizer_finalize ();
5467 5468
	  free_dominance_info (CDI_DOMINATORS);
	}
5469
      FOR_ALL_BB_FN (bb, cfun)
5470 5471 5472 5473
	bb->loop_father = NULL;
      current_loops = NULL;

      ira_destroy ();
Vladimir Makarov committed
5474

5475 5476 5477
      lra (ira_dump_file);
      /* ???!!! Move it before lra () when we use ira_reg_equiv in
	 LRA.  */
5478
      vec_free (reg_equivs);
5479 5480 5481 5482 5483 5484 5485 5486
      reg_equivs = NULL;
      need_dce = false;
    }
  else
    {
      df_set_flags (DF_NO_INSN_RESCAN);
      build_insn_chain ();

5487
      need_dce = reload (get_insns (), ira_conflicts_p);
5488 5489 5490
    }

  timevar_pop (TV_RELOAD);
Vladimir Makarov committed
5491

5492 5493
  timevar_push (TV_IRA);

5494
  if (ira_conflicts_p && ! ira_use_lra_p)
Vladimir Makarov committed
5495 5496 5497
    {
      ira_free (ira_spilled_reg_stack_slots);
      ira_finish_assign ();
H.J. Lu committed
5498
    }
5499

Vladimir Makarov committed
5500 5501
  if (internal_flag_ira_verbose > 0 && ira_dump_file != NULL
      && overall_cost_before != ira_overall_cost)
5502
    fprintf (ira_dump_file, "+++Overall after reload %" PRId64 "\n",
5503
	     ira_overall_cost);
H.J. Lu committed
5504

5505 5506
  flag_ira_share_spill_slots = saved_flag_ira_share_spill_slots;

5507
  if (! ira_use_lra_p)
5508
    {
5509 5510 5511
      ira_destroy ();
      if (current_loops != NULL)
	{
5512
	  loop_optimizer_finalize ();
5513 5514
	  free_dominance_info (CDI_DOMINATORS);
	}
5515
      FOR_ALL_BB_FN (bb, cfun)
5516 5517 5518 5519 5520
	bb->loop_father = NULL;
      current_loops = NULL;
      
      regstat_free_ri ();
      regstat_free_n_sets_and_refs ();
5521
    }
H.J. Lu committed
5522

Vladimir Makarov committed
5523
  if (optimize)
5524
    cleanup_cfg (CLEANUP_EXPENSIVE);
H.J. Lu committed
5525

5526
  finish_reg_equiv ();
Vladimir Makarov committed
5527 5528 5529 5530 5531 5532 5533

  bitmap_obstack_release (&ira_bitmap_obstack);
#ifndef IRA_NO_OBSTACK
  obstack_free (&ira_obstack, NULL);
#endif

  /* The code after the reload has changed so much that at this point
5534
     we might as well just rescan everything.  Note that
Vladimir Makarov committed
5535 5536 5537 5538 5539 5540
     df_rescan_all_insns is not going to help here because it does not
     touch the artificial uses and defs.  */
  df_finish_pass (true);
  df_scan_alloc (NULL);
  df_scan_blocks ();

5541 5542 5543 5544 5545 5546
  if (optimize > 1)
    {
      df_live_add_problem ();
      df_live_set_all_dirty ();
    }

Vladimir Makarov committed
5547 5548 5549
  if (optimize)
    df_analyze ();

5550 5551
  if (need_dce && optimize)
    run_fast_dce ();
5552

5553 5554 5555 5556 5557 5558 5559 5560 5561 5562 5563 5564
  /* Diagnose uses of the hard frame pointer when it is used as a global
     register.  Often we can get away with letting the user appropriate
     the frame pointer, but we should let them know when code generation
     makes that impossible.  */
  if (global_regs[HARD_FRAME_POINTER_REGNUM] && frame_pointer_needed)
    {
      tree decl = global_regs_decl[HARD_FRAME_POINTER_REGNUM];
      error_at (DECL_SOURCE_LOCATION (current_function_decl),
                "frame pointer required, but reserved");
      inform (DECL_SOURCE_LOCATION (decl), "for %qD", decl);
    }

5565 5566 5567 5568
  /* If we are doing generic stack checking, give a warning if this
     function's frame size is larger than we expect.  */
  if (flag_stack_check == GENERIC_STACK_CHECK)
    {
5569
      poly_int64 size = get_frame_size () + STACK_CHECK_FIXED_FRAME_SIZE;
5570 5571 5572 5573 5574

      for (int i = 0; i < FIRST_PSEUDO_REGISTER; i++)
	if (df_regs_ever_live_p (i) && !fixed_regs[i] && call_used_regs[i])
	  size += UNITS_PER_WORD;

5575
      if (constant_lower_bound (size) > STACK_CHECK_MAX_FRAME_SIZE)
5576 5577 5578
	warning (0, "frame size too large for reliable stack checking");
    }

5579 5580 5581
  if (pic_offset_table_regno != INVALID_REGNUM)
    pic_offset_table_rtx = gen_rtx_REG (Pmode, pic_offset_table_regno);

5582
  timevar_pop (TV_IRA);
Vladimir Makarov committed
5583 5584 5585 5586
}

/* Run the integrated register allocator.  */

5587 5588 5589
namespace {

const pass_data pass_data_ira =
Vladimir Makarov committed
5590
{
5591 5592 5593 5594 5595 5596 5597 5598 5599
  RTL_PASS, /* type */
  "ira", /* name */
  OPTGROUP_NONE, /* optinfo_flags */
  TV_IRA, /* tv_id */
  0, /* properties_required */
  0, /* properties_provided */
  0, /* properties_destroyed */
  0, /* todo_flags_start */
  TODO_do_not_ggc_collect, /* todo_flags_finish */
5600 5601
};

5602 5603 5604
class pass_ira : public rtl_opt_pass
{
public:
5605 5606
  pass_ira (gcc::context *ctxt)
    : rtl_opt_pass (pass_data_ira, ctxt)
5607 5608 5609
  {}

  /* opt_pass methods: */
5610 5611 5612 5613
  virtual bool gate (function *)
    {
      return !targetm.no_register_allocation;
    }
5614 5615 5616 5617 5618
  virtual unsigned int execute (function *)
    {
      ira (dump_file);
      return 0;
    }
5619 5620 5621 5622 5623 5624 5625 5626 5627 5628 5629 5630 5631 5632

}; // class pass_ira

} // anon namespace

rtl_opt_pass *
make_pass_ira (gcc::context *ctxt)
{
  return new pass_ira (ctxt);
}

namespace {

const pass_data pass_data_reload =
5633
{
5634 5635 5636 5637 5638 5639 5640 5641 5642
  RTL_PASS, /* type */
  "reload", /* name */
  OPTGROUP_NONE, /* optinfo_flags */
  TV_RELOAD, /* tv_id */
  0, /* properties_required */
  0, /* properties_provided */
  0, /* properties_destroyed */
  0, /* todo_flags_start */
  0, /* todo_flags_finish */
Vladimir Makarov committed
5643
};
5644 5645 5646 5647

class pass_reload : public rtl_opt_pass
{
public:
5648 5649
  pass_reload (gcc::context *ctxt)
    : rtl_opt_pass (pass_data_reload, ctxt)
5650 5651 5652
  {}

  /* opt_pass methods: */
5653 5654 5655 5656
  virtual bool gate (function *)
    {
      return !targetm.no_register_allocation;
    }
5657 5658 5659 5660 5661
  virtual unsigned int execute (function *)
    {
      do_reload ();
      return 0;
    }
5662 5663 5664 5665 5666 5667 5668 5669 5670 5671

}; // class pass_reload

} // anon namespace

rtl_opt_pass *
make_pass_reload (gcc::context *ctxt)
{
  return new pass_reload (ctxt);
}