lcm.c 40.1 KB
Newer Older
1
/* Generic partial redundancy elimination with lazy code motion support.
2
   Copyright (C) 1998, 1999, 2000, 2001, 2002, 2003, 2004
3
   Free Software Foundation, Inc.
Jeffrey A Law committed
4

5
This file is part of GCC.
Jeffrey A Law committed
6

7 8 9 10
GCC is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free
Software Foundation; either version 2, or (at your option) any later
version.
Jeffrey A Law committed
11

12 13 14 15
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
WARRANTY; without even the implied warranty of MERCHANTABILITY or
FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
for more details.
Jeffrey A Law committed
16 17

You should have received a copy of the GNU General Public License
18 19 20
along with GCC; see the file COPYING.  If not, write to the Free
Software Foundation, 59 Temple Place - Suite 330, Boston, MA
02111-1307, USA.  */
Jeffrey A Law committed
21 22

/* These routines are meant to be used by various optimization
23
   passes which can be modeled as lazy code motion problems.
Jeffrey A Law committed
24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54
   Including, but not limited to:

	* Traditional partial redundancy elimination.

	* Placement of caller/caller register save/restores.

	* Load/store motion.

	* Copy motion.

	* Conversion of flat register files to a stacked register
	model.

	* Dead load/store elimination.

  These routines accept as input:

	* Basic block information (number of blocks, lists of
	predecessors and successors).  Note the granularity
	does not need to be basic block, they could be statements
	or functions.

	* Bitmaps of local properties (computed, transparent and
	anticipatable expressions).

  The output of these routines is bitmap of redundant computations
  and a bitmap of optimal placement points.  */


#include "config.h"
#include "system.h"
55 56
#include "coretypes.h"
#include "tm.h"
Jeffrey A Law committed
57 58 59 60 61 62 63 64
#include "rtl.h"
#include "regs.h"
#include "hard-reg-set.h"
#include "flags.h"
#include "real.h"
#include "insn-config.h"
#include "recog.h"
#include "basic-block.h"
65
#include "output.h"
66
#include "tm_p.h"
67
#include "function.h"
68

69 70 71
/* We want target macros for the mode switching code to be able to refer
   to instruction attribute values.  */
#include "insn-attr.h"
Jeffrey A Law committed
72

73
/* Edge based LCM routines.  */
74 75 76 77 78 79 80
static void compute_antinout_edge (sbitmap *, sbitmap *, sbitmap *, sbitmap *);
static void compute_earliest (struct edge_list *, int, sbitmap *, sbitmap *,
			      sbitmap *, sbitmap *, sbitmap *);
static void compute_laterin (struct edge_list *, sbitmap *, sbitmap *,
			     sbitmap *, sbitmap *);
static void compute_insert_delete (struct edge_list *edge_list, sbitmap *,
				   sbitmap *, sbitmap *, sbitmap *, sbitmap *);
81 82

/* Edge based LCM routines on a reverse flowgraph.  */
83 84 85 86 87 88 89
static void compute_farthest (struct edge_list *, int, sbitmap *, sbitmap *,
			      sbitmap*, sbitmap *, sbitmap *);
static void compute_nearerout (struct edge_list *, sbitmap *, sbitmap *,
			       sbitmap *, sbitmap *);
static void compute_rev_insert_delete (struct edge_list *edge_list, sbitmap *,
				       sbitmap *, sbitmap *, sbitmap *,
				       sbitmap *);
90 91

/* Edge based lcm routines.  */
Daniel Berlin committed
92

93 94
/* Compute expression anticipatability at entrance and exit of each block.
   This is done based on the flow graph, and not on the pred-succ lists.
95
   Other than that, its pretty much identical to compute_antinout.  */
Jeffrey A Law committed
96 97

static void
98 99
compute_antinout_edge (sbitmap *antloc, sbitmap *transp, sbitmap *antin,
		       sbitmap *antout)
Jeffrey A Law committed
100
{
101
  basic_block bb;
102
  edge e;
103 104
  basic_block *worklist, *qin, *qout, *qend;
  unsigned int qlen;
Daniel Berlin committed
105

106 107 108
  /* Allocate a worklist array/queue.  Entries are only added to the
     list if they were not already on the list.  So the size is
     bounded by the number of basic blocks.  */
109
  qin = qout = worklist = xmalloc (sizeof (basic_block) * n_basic_blocks);
Jeffrey A Law committed
110

111 112
  /* We want a maximal solution, so make an optimistic initialization of
     ANTIN.  */
113
  sbitmap_vector_ones (antin, last_basic_block);
Jeffrey A Law committed
114

115 116
  /* Put every block on the worklist; this is necessary because of the
     optimistic initialization of ANTIN above.  */
117
  FOR_EACH_BB_REVERSE (bb)
Jeffrey A Law committed
118
    {
119
      *qin++ = bb;
120
      bb->aux = bb;
121
    }
122

123
  qin = worklist;
124 125
  qend = &worklist[n_basic_blocks];
  qlen = n_basic_blocks;
Jeffrey A Law committed
126

127 128 129 130 131
  /* Mark blocks which are predecessors of the exit block so that we
     can easily identify them below.  */
  for (e = EXIT_BLOCK_PTR->pred; e; e = e->pred_next)
    e->src->aux = EXIT_BLOCK_PTR;

132
  /* Iterate until the worklist is empty.  */
133
  while (qlen)
134 135
    {
      /* Take the first entry off the worklist.  */
136
      bb = *qout++;
137
      qlen--;
Daniel Berlin committed
138

139
      if (qout >= qend)
140
	qout = worklist;
Jeffrey A Law committed
141

142
      if (bb->aux == EXIT_BLOCK_PTR)
143 144 145
	/* Do not clear the aux field for blocks which are predecessors of
	   the EXIT block.  That way we never add then to the worklist
	   again.  */
146
	sbitmap_zero (antout[bb->index]);
147 148 149 150
      else
	{
	  /* Clear the aux field of this block so that it can be added to
	     the worklist again if necessary.  */
151 152
	  bb->aux = NULL;
	  sbitmap_intersection_of_succs (antout[bb->index], antin, bb->index);
153
	}
154

155 156
      if (sbitmap_a_or_b_and_c_cg (antin[bb->index], antloc[bb->index],
				   transp[bb->index], antout[bb->index]))
157 158 159
	/* If the in state of this block changed, then we need
	   to add the predecessors of this block to the worklist
	   if they are not already on the worklist.  */
160
	for (e = bb->pred; e; e = e->pred_next)
161
	  if (!e->src->aux && e->src != ENTRY_BLOCK_PTR)
Jeffrey A Law committed
162
	    {
163
	      *qin++ = e->src;
164
	      e->src->aux = e;
165 166
	      qlen++;
	      if (qin >= qend)
167
		qin = worklist;
Jeffrey A Law committed
168 169
	    }
    }
170

171 172
  clear_aux_for_edges ();
  clear_aux_for_blocks ();
173
  free (worklist);
Jeffrey A Law committed
174 175
}

176
/* Compute the earliest vector for edge based lcm.  */
177

Jeffrey A Law committed
178
static void
179 180 181
compute_earliest (struct edge_list *edge_list, int n_exprs, sbitmap *antin,
		  sbitmap *antout, sbitmap *avout, sbitmap *kill,
		  sbitmap *earliest)
Jeffrey A Law committed
182
{
183
  sbitmap difference, temp_bitmap;
184
  int x, num_edges;
185
  basic_block pred, succ;
Jeffrey A Law committed
186

187
  num_edges = NUM_EDGES (edge_list);
Jeffrey A Law committed
188

189 190
  difference = sbitmap_alloc (n_exprs);
  temp_bitmap = sbitmap_alloc (n_exprs);
Jeffrey A Law committed
191

192
  for (x = 0; x < num_edges; x++)
Jeffrey A Law committed
193
    {
194 195 196
      pred = INDEX_EDGE_PRED_BB (edge_list, x);
      succ = INDEX_EDGE_SUCC_BB (edge_list, x);
      if (pred == ENTRY_BLOCK_PTR)
197
	sbitmap_copy (earliest[x], antin[succ->index]);
198
      else
199
	{
200
	  if (succ == EXIT_BLOCK_PTR)
201
	    sbitmap_zero (earliest[x]);
202
	  else
Jeffrey A Law committed
203
	    {
204 205 206
	      sbitmap_difference (difference, antin[succ->index],
				  avout[pred->index]);
	      sbitmap_not (temp_bitmap, antout[pred->index]);
207
	      sbitmap_a_and_b_or_c (earliest[x], difference,
208
				    kill[pred->index], temp_bitmap);
Jeffrey A Law committed
209 210 211
	    }
	}
    }
212

213 214
  sbitmap_free (temp_bitmap);
  sbitmap_free (difference);
Jeffrey A Law committed
215 216
}

217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244
/* later(p,s) is dependent on the calculation of laterin(p).
   laterin(p) is dependent on the calculation of later(p2,p).

     laterin(ENTRY) is defined as all 0's
     later(ENTRY, succs(ENTRY)) are defined using laterin(ENTRY)
     laterin(succs(ENTRY)) is defined by later(ENTRY, succs(ENTRY)).

   If we progress in this manner, starting with all basic blocks
   in the work list, anytime we change later(bb), we need to add
   succs(bb) to the worklist if they are not already on the worklist.

   Boundary conditions:

     We prime the worklist all the normal basic blocks.   The ENTRY block can
     never be added to the worklist since it is never the successor of any
     block.  We explicitly prevent the EXIT block from being added to the
     worklist.

     We optimistically initialize LATER.  That is the only time this routine
     will compute LATER for an edge out of the entry block since the entry
     block is never on the worklist.  Thus, LATERIN is neither used nor
     computed for the ENTRY block.

     Since the EXIT block is never added to the worklist, we will neither
     use nor compute LATERIN for the exit block.  Edges which reach the
     EXIT block are handled in the normal fashion inside the loop.  However,
     the insertion/deletion computation needs LATERIN(EXIT), so we have
     to compute it.  */
245

Jeffrey A Law committed
246
static void
247 248
compute_laterin (struct edge_list *edge_list, sbitmap *earliest,
		 sbitmap *antloc, sbitmap *later, sbitmap *laterin)
Jeffrey A Law committed
249
{
250
  int num_edges, i;
251
  edge e;
252
  basic_block *worklist, *qin, *qout, *qend, bb;
253
  unsigned int qlen;
Jeffrey A Law committed
254

255
  num_edges = NUM_EDGES (edge_list);
Jeffrey A Law committed
256

257 258 259
  /* Allocate a worklist array/queue.  Entries are only added to the
     list if they were not already on the list.  So the size is
     bounded by the number of basic blocks.  */
260
  qin = qout = worklist
261
    = xmalloc (sizeof (basic_block) * (n_basic_blocks + 1));
262 263 264

  /* Initialize a mapping from each edge to its index.  */
  for (i = 0; i < num_edges; i++)
265
    INDEX_EDGE (edge_list, i)->aux = (void *) (size_t) i;
266 267 268 269 270 271 272 273 274 275 276 277 278

  /* We want a maximal solution, so initially consider LATER true for
     all edges.  This allows propagation through a loop since the incoming
     loop edge will have LATER set, so if all the other incoming edges
     to the loop are set, then LATERIN will be set for the head of the
     loop.

     If the optimistic setting of LATER on that edge was incorrect (for
     example the expression is ANTLOC in a block within the loop) then
     this algorithm will detect it when we process the block at the head
     of the optimistic edge.  That will requeue the affected blocks.  */
  sbitmap_vector_ones (later, num_edges);

279 280 281 282 283
  /* Note that even though we want an optimistic setting of LATER, we
     do not want to be overly optimistic.  Consider an outgoing edge from
     the entry block.  That edge should always have a LATER value the
     same as EARLIEST for that edge.  */
  for (e = ENTRY_BLOCK_PTR->succ; e; e = e->succ_next)
284
    sbitmap_copy (later[(size_t) e->aux], earliest[(size_t) e->aux]);
285

286 287
  /* Add all the blocks to the worklist.  This prevents an early exit from
     the loop given our optimistic initialization of LATER above.  */
288
  FOR_EACH_BB (bb)
Jeffrey A Law committed
289
    {
290 291
      *qin++ = bb;
      bb->aux = bb;
292
    }
293 294 295
  qin = worklist;
  /* Note that we do not use the last allocated element for our queue,
     as EXIT_BLOCK is never inserted into it. In fact the above allocation
296
     of n_basic_blocks + 1 elements is not necessary.  */
297 298
  qend = &worklist[n_basic_blocks];
  qlen = n_basic_blocks;
299

300
  /* Iterate until the worklist is empty.  */
301
  while (qlen)
302
    {
303
      /* Take the first entry off the worklist.  */
304 305
      bb = *qout++;
      bb->aux = NULL;
306 307
      qlen--;
      if (qout >= qend)
308
	qout = worklist;
309 310

      /* Compute the intersection of LATERIN for each incoming edge to B.  */
311 312 313
      sbitmap_ones (laterin[bb->index]);
      for (e = bb->pred; e != NULL; e = e->pred_next)
	sbitmap_a_and_b (laterin[bb->index], laterin[bb->index], later[(size_t)e->aux]);
314 315

      /* Calculate LATER for all outgoing edges.  */
316
      for (e = bb->succ; e != NULL; e = e->succ_next)
317
	if (sbitmap_union_of_diff_cg (later[(size_t) e->aux],
318 319 320
				      earliest[(size_t) e->aux],
				      laterin[e->src->index],
				      antloc[e->src->index])
321 322 323 324
	    /* If LATER for an outgoing edge was changed, then we need
	       to add the target of the outgoing edge to the worklist.  */
	    && e->dest != EXIT_BLOCK_PTR && e->dest->aux == 0)
	  {
325
	    *qin++ = e->dest;
326
	    e->dest->aux = e;
327 328 329
	    qlen++;
	    if (qin >= qend)
	      qin = worklist;
330
	  }
Jeffrey A Law committed
331 332
    }

333 334 335
  /* Computation of insertion and deletion points requires computing LATERIN
     for the EXIT block.  We allocated an extra entry in the LATERIN array
     for just this purpose.  */
336
  sbitmap_ones (laterin[last_basic_block]);
337
  for (e = EXIT_BLOCK_PTR->pred; e != NULL; e = e->pred_next)
338 339
    sbitmap_a_and_b (laterin[last_basic_block],
		     laterin[last_basic_block],
340
		     later[(size_t) e->aux]);
341

342
  clear_aux_for_edges ();
343
  free (worklist);
Jeffrey A Law committed
344 345
}

346
/* Compute the insertion and deletion points for edge based LCM.  */
347

348
static void
349 350 351
compute_insert_delete (struct edge_list *edge_list, sbitmap *antloc,
		       sbitmap *later, sbitmap *laterin, sbitmap *insert,
		       sbitmap *delete)
352 353
{
  int x;
354
  basic_block bb;
Jeffrey A Law committed
355

356 357
  FOR_EACH_BB (bb)
    sbitmap_difference (delete[bb->index], antloc[bb->index], laterin[bb->index]);
358

359 360 361
  for (x = 0; x < NUM_EDGES (edge_list); x++)
    {
      basic_block b = INDEX_EDGE_SUCC_BB (edge_list, x);
362

363
      if (b == EXIT_BLOCK_PTR)
364
	sbitmap_difference (insert[x], later[x], laterin[last_basic_block]);
365
      else
366
	sbitmap_difference (insert[x], later[x], laterin[b->index]);
367 368
    }
}
Jeffrey A Law committed
369

370 371 372
/* Given local properties TRANSP, ANTLOC, AVOUT, KILL return the insert and
   delete vectors for edge based LCM.  Returns an edgelist which is used to
   map the insert vector to what edge an expression should be inserted on.  */
Jeffrey A Law committed
373

374
struct edge_list *
375 376 377
pre_edge_lcm (FILE *file ATTRIBUTE_UNUSED, int n_exprs, sbitmap *transp,
	      sbitmap *avloc, sbitmap *antloc, sbitmap *kill,
	      sbitmap **insert, sbitmap **delete)
Jeffrey A Law committed
378
{
379 380 381 382 383
  sbitmap *antin, *antout, *earliest;
  sbitmap *avin, *avout;
  sbitmap *later, *laterin;
  struct edge_list *edge_list;
  int num_edges;
Jeffrey A Law committed
384

385 386
  edge_list = create_edge_list ();
  num_edges = NUM_EDGES (edge_list);
Jeffrey A Law committed
387

388 389
#ifdef LCM_DEBUG_INFO
  if (file)
Jeffrey A Law committed
390
    {
391 392 393
      fprintf (file, "Edge List:\n");
      verify_edge_list (file, edge_list);
      print_edge_list (file, edge_list);
394 395 396 397
      dump_sbitmap_vector (file, "transp", "", transp, last_basic_block);
      dump_sbitmap_vector (file, "antloc", "", antloc, last_basic_block);
      dump_sbitmap_vector (file, "avloc", "", avloc, last_basic_block);
      dump_sbitmap_vector (file, "kill", "", kill, last_basic_block);
Jeffrey A Law committed
398
    }
399
#endif
Jeffrey A Law committed
400

401
  /* Compute global availability.  */
402 403
  avin = sbitmap_vector_alloc (last_basic_block, n_exprs);
  avout = sbitmap_vector_alloc (last_basic_block, n_exprs);
404
  compute_available (avloc, kill, avout, avin);
405
  sbitmap_vector_free (avin);
Jeffrey A Law committed
406

407
  /* Compute global anticipatability.  */
408 409
  antin = sbitmap_vector_alloc (last_basic_block, n_exprs);
  antout = sbitmap_vector_alloc (last_basic_block, n_exprs);
410
  compute_antinout_edge (antloc, transp, antin, antout);
Jeffrey A Law committed
411

412 413
#ifdef LCM_DEBUG_INFO
  if (file)
Jeffrey A Law committed
414
    {
415 416
      dump_sbitmap_vector (file, "antin", "", antin, last_basic_block);
      dump_sbitmap_vector (file, "antout", "", antout, last_basic_block);
Jeffrey A Law committed
417
    }
418
#endif
Jeffrey A Law committed
419

420 421 422
  /* Compute earliestness.  */
  earliest = sbitmap_vector_alloc (num_edges, n_exprs);
  compute_earliest (edge_list, n_exprs, antin, antout, avout, kill, earliest);
Jeffrey A Law committed
423

424 425 426 427
#ifdef LCM_DEBUG_INFO
  if (file)
    dump_sbitmap_vector (file, "earliest", "", earliest, num_edges);
#endif
Jeffrey A Law committed
428

429 430 431
  sbitmap_vector_free (antout);
  sbitmap_vector_free (antin);
  sbitmap_vector_free (avout);
Jeffrey A Law committed
432

433
  later = sbitmap_vector_alloc (num_edges, n_exprs);
434

435
  /* Allocate an extra element for the exit block in the laterin vector.  */
436
  laterin = sbitmap_vector_alloc (last_basic_block + 1, n_exprs);
437 438
  compute_laterin (edge_list, earliest, antloc, later, laterin);

439 440 441
#ifdef LCM_DEBUG_INFO
  if (file)
    {
442
      dump_sbitmap_vector (file, "laterin", "", laterin, last_basic_block + 1);
443 444 445
      dump_sbitmap_vector (file, "later", "", later, num_edges);
    }
#endif
Jeffrey A Law committed
446

447
  sbitmap_vector_free (earliest);
448 449

  *insert = sbitmap_vector_alloc (num_edges, n_exprs);
450
  *delete = sbitmap_vector_alloc (last_basic_block, n_exprs);
451
  compute_insert_delete (edge_list, antloc, later, laterin, *insert, *delete);
Jeffrey A Law committed
452

453 454
  sbitmap_vector_free (laterin);
  sbitmap_vector_free (later);
455 456 457

#ifdef LCM_DEBUG_INFO
  if (file)
Jeffrey A Law committed
458
    {
459
      dump_sbitmap_vector (file, "pre_insert_map", "", *insert, num_edges);
460
      dump_sbitmap_vector (file, "pre_delete_map", "", *delete,
461
			   last_basic_block);
Jeffrey A Law committed
462
    }
463
#endif
Jeffrey A Law committed
464

465 466
  return edge_list;
}
Daniel Berlin committed
467 468 469 470

/* Compute the AVIN and AVOUT vectors from the AVLOC and KILL vectors.
   Return the number of passes we performed to iterate to a solution.  */

471
void
472 473
compute_available (sbitmap *avloc, sbitmap *kill, sbitmap *avout,
		   sbitmap *avin)
Jeffrey A Law committed
474
{
Daniel Berlin committed
475
  edge e;
476
  basic_block *worklist, *qin, *qout, *qend, bb;
Daniel Berlin committed
477 478 479 480 481
  unsigned int qlen;

  /* Allocate a worklist array/queue.  Entries are only added to the
     list if they were not already on the list.  So the size is
     bounded by the number of basic blocks.  */
482
  qin = qout = worklist = xmalloc (sizeof (basic_block) * n_basic_blocks);
Daniel Berlin committed
483 484

  /* We want a maximal solution.  */
485
  sbitmap_vector_ones (avout, last_basic_block);
Daniel Berlin committed
486 487 488

  /* Put every block on the worklist; this is necessary because of the
     optimistic initialization of AVOUT above.  */
489
  FOR_EACH_BB (bb)
Daniel Berlin committed
490
    {
491 492
      *qin++ = bb;
      bb->aux = bb;
Daniel Berlin committed
493 494 495
    }

  qin = worklist;
496 497
  qend = &worklist[n_basic_blocks];
  qlen = n_basic_blocks;
Daniel Berlin committed
498 499 500 501 502 503 504 505 506 507

  /* Mark blocks which are successors of the entry block so that we
     can easily identify them below.  */
  for (e = ENTRY_BLOCK_PTR->succ; e; e = e->succ_next)
    e->dest->aux = ENTRY_BLOCK_PTR;

  /* Iterate until the worklist is empty.  */
  while (qlen)
    {
      /* Take the first entry off the worklist.  */
508
      bb = *qout++;
Daniel Berlin committed
509 510 511
      qlen--;

      if (qout >= qend)
512
	qout = worklist;
Daniel Berlin committed
513 514 515 516

      /* If one of the predecessor blocks is the ENTRY block, then the
	 intersection of avouts is the null set.  We can identify such blocks
	 by the special value in the AUX field in the block structure.  */
517
      if (bb->aux == ENTRY_BLOCK_PTR)
Daniel Berlin committed
518 519
	/* Do not clear the aux field for blocks which are successors of the
	   ENTRY block.  That way we never add then to the worklist again.  */
520
	sbitmap_zero (avin[bb->index]);
Daniel Berlin committed
521 522 523 524
      else
	{
	  /* Clear the aux field of this block so that it can be added to
	     the worklist again if necessary.  */
525 526
	  bb->aux = NULL;
	  sbitmap_intersection_of_preds (avin[bb->index], avout, bb->index);
Daniel Berlin committed
527 528
	}

529
      if (sbitmap_union_of_diff_cg (avout[bb->index], avloc[bb->index], avin[bb->index], kill[bb->index]))
Daniel Berlin committed
530 531 532
	/* If the out state of this block changed, then we need
	   to add the successors of this block to the worklist
	   if they are not already on the worklist.  */
533
	for (e = bb->succ; e; e = e->succ_next)
Daniel Berlin committed
534 535 536 537 538 539 540
	  if (!e->dest->aux && e->dest != EXIT_BLOCK_PTR)
	    {
	      *qin++ = e->dest;
	      e->dest->aux = e;
	      qlen++;

	      if (qin >= qend)
541
		qin = worklist;
Daniel Berlin committed
542 543 544 545 546 547
	    }
    }

  clear_aux_for_edges ();
  clear_aux_for_blocks ();
  free (worklist);
Jeffrey A Law committed
548 549
}

550
/* Compute the farthest vector for edge based lcm.  */
551

Jeffrey A Law committed
552
static void
553 554 555
compute_farthest (struct edge_list *edge_list, int n_exprs,
		  sbitmap *st_avout, sbitmap *st_avin, sbitmap *st_antin,
		  sbitmap *kill, sbitmap *farthest)
Jeffrey A Law committed
556
{
557
  sbitmap difference, temp_bitmap;
558
  int x, num_edges;
559
  basic_block pred, succ;
Jeffrey A Law committed
560

561
  num_edges = NUM_EDGES (edge_list);
Jeffrey A Law committed
562

563 564
  difference = sbitmap_alloc (n_exprs);
  temp_bitmap = sbitmap_alloc (n_exprs);
Jeffrey A Law committed
565

566
  for (x = 0; x < num_edges; x++)
Jeffrey A Law committed
567
    {
568 569 570
      pred = INDEX_EDGE_PRED_BB (edge_list, x);
      succ = INDEX_EDGE_SUCC_BB (edge_list, x);
      if (succ == EXIT_BLOCK_PTR)
571
	sbitmap_copy (farthest[x], st_avout[pred->index]);
572
      else
Jeffrey A Law committed
573
	{
574
	  if (pred == ENTRY_BLOCK_PTR)
575
	    sbitmap_zero (farthest[x]);
576 577
	  else
	    {
578 579 580
	      sbitmap_difference (difference, st_avout[pred->index],
				  st_antin[succ->index]);
	      sbitmap_not (temp_bitmap, st_avin[succ->index]);
581
	      sbitmap_a_and_b_or_c (farthest[x], difference,
582
				    kill[succ->index], temp_bitmap);
583
	    }
Jeffrey A Law committed
584 585
	}
    }
586

587 588
  sbitmap_free (temp_bitmap);
  sbitmap_free (difference);
Jeffrey A Law committed
589 590
}

591 592 593 594 595
/* Compute nearer and nearerout vectors for edge based lcm.

   This is the mirror of compute_laterin, additional comments on the
   implementation can be found before compute_laterin.  */

Jeffrey A Law committed
596
static void
597 598
compute_nearerout (struct edge_list *edge_list, sbitmap *farthest,
		   sbitmap *st_avloc, sbitmap *nearer, sbitmap *nearerout)
Jeffrey A Law committed
599
{
600
  int num_edges, i;
601
  edge e;
602
  basic_block *worklist, *tos, bb;
Jeffrey A Law committed
603

604
  num_edges = NUM_EDGES (edge_list);
Jeffrey A Law committed
605

606 607 608
  /* Allocate a worklist array/queue.  Entries are only added to the
     list if they were not already on the list.  So the size is
     bounded by the number of basic blocks.  */
609
  tos = worklist = xmalloc (sizeof (basic_block) * (n_basic_blocks + 1));
Jeffrey A Law committed
610

611 612 613
  /* Initialize NEARER for each edge and build a mapping from an edge to
     its index.  */
  for (i = 0; i < num_edges; i++)
614
    INDEX_EDGE (edge_list, i)->aux = (void *) (size_t) i;
615

616 617 618
  /* We want a maximal solution.  */
  sbitmap_vector_ones (nearer, num_edges);

619 620 621 622 623
  /* Note that even though we want an optimistic setting of NEARER, we
     do not want to be overly optimistic.  Consider an incoming edge to
     the exit block.  That edge should always have a NEARER value the
     same as FARTHEST for that edge.  */
  for (e = EXIT_BLOCK_PTR->pred; e; e = e->pred_next)
624
    sbitmap_copy (nearer[(size_t)e->aux], farthest[(size_t)e->aux]);
625

626 627
  /* Add all the blocks to the worklist.  This prevents an early exit
     from the loop given our optimistic initialization of NEARER.  */
628
  FOR_EACH_BB (bb)
Jeffrey A Law committed
629
    {
630 631
      *tos++ = bb;
      bb->aux = bb;
632
    }
633

634 635
  /* Iterate until the worklist is empty.  */
  while (tos != worklist)
636
    {
637
      /* Take the first entry off the worklist.  */
638 639
      bb = *--tos;
      bb->aux = NULL;
640 641

      /* Compute the intersection of NEARER for each outgoing edge from B.  */
642 643 644
      sbitmap_ones (nearerout[bb->index]);
      for (e = bb->succ; e != NULL; e = e->succ_next)
	sbitmap_a_and_b (nearerout[bb->index], nearerout[bb->index],
645
			 nearer[(size_t) e->aux]);
646 647

      /* Calculate NEARER for all incoming edges.  */
648
      for (e = bb->pred; e != NULL; e = e->pred_next)
649
	if (sbitmap_union_of_diff_cg (nearer[(size_t) e->aux],
650 651 652
				      farthest[(size_t) e->aux],
				      nearerout[e->dest->index],
				      st_avloc[e->dest->index])
653 654 655 656 657 658 659
	    /* If NEARER for an incoming edge was changed, then we need
	       to add the source of the incoming edge to the worklist.  */
	    && e->src != ENTRY_BLOCK_PTR && e->src->aux == 0)
	  {
	    *tos++ = e->src;
	    e->src->aux = e;
	  }
660
    }
Jeffrey A Law committed
661

662 663 664
  /* Computation of insertion and deletion points requires computing NEAREROUT
     for the ENTRY block.  We allocated an extra entry in the NEAREROUT array
     for just this purpose.  */
665
  sbitmap_ones (nearerout[last_basic_block]);
666
  for (e = ENTRY_BLOCK_PTR->succ; e != NULL; e = e->succ_next)
667 668
    sbitmap_a_and_b (nearerout[last_basic_block],
		     nearerout[last_basic_block],
669
		     nearer[(size_t) e->aux]);
670

671
  clear_aux_for_edges ();
672
  free (tos);
673
}
Jeffrey A Law committed
674

675
/* Compute the insertion and deletion points for edge based LCM.  */
676

Jeffrey A Law committed
677
static void
678 679 680
compute_rev_insert_delete (struct edge_list *edge_list, sbitmap *st_avloc,
			   sbitmap *nearer, sbitmap *nearerout,
			   sbitmap *insert, sbitmap *delete)
Jeffrey A Law committed
681
{
682
  int x;
683
  basic_block bb;
Jeffrey A Law committed
684

685 686
  FOR_EACH_BB (bb)
    sbitmap_difference (delete[bb->index], st_avloc[bb->index], nearerout[bb->index]);
687

688
  for (x = 0; x < NUM_EDGES (edge_list); x++)
Jeffrey A Law committed
689
    {
690 691
      basic_block b = INDEX_EDGE_PRED_BB (edge_list, x);
      if (b == ENTRY_BLOCK_PTR)
692
	sbitmap_difference (insert[x], nearer[x], nearerout[last_basic_block]);
Jeffrey A Law committed
693
      else
694
	sbitmap_difference (insert[x], nearer[x], nearerout[b->index]);
Jeffrey A Law committed
695 696 697
    }
}

698
/* Given local properties TRANSP, ST_AVLOC, ST_ANTLOC, KILL return the
699 700 701
   insert and delete vectors for edge based reverse LCM.  Returns an
   edgelist which is used to map the insert vector to what edge
   an expression should be inserted on.  */
Jeffrey A Law committed
702

703
struct edge_list *
704 705 706
pre_edge_rev_lcm (FILE *file ATTRIBUTE_UNUSED, int n_exprs, sbitmap *transp,
		  sbitmap *st_avloc, sbitmap *st_antloc, sbitmap *kill,
		  sbitmap **insert, sbitmap **delete)
Jeffrey A Law committed
707
{
708 709 710 711
  sbitmap *st_antin, *st_antout;
  sbitmap *st_avout, *st_avin, *farthest;
  sbitmap *nearer, *nearerout;
  struct edge_list *edge_list;
712
  int num_edges;
713 714 715 716

  edge_list = create_edge_list ();
  num_edges = NUM_EDGES (edge_list);

717 718
  st_antin = sbitmap_vector_alloc (last_basic_block, n_exprs);
  st_antout = sbitmap_vector_alloc (last_basic_block, n_exprs);
719 720
  sbitmap_vector_zero (st_antin, last_basic_block);
  sbitmap_vector_zero (st_antout, last_basic_block);
721 722 723
  compute_antinout_edge (st_antloc, transp, st_antin, st_antout);

  /* Compute global anticipatability.  */
724 725
  st_avout = sbitmap_vector_alloc (last_basic_block, n_exprs);
  st_avin = sbitmap_vector_alloc (last_basic_block, n_exprs);
726 727 728 729 730 731 732 733
  compute_available (st_avloc, kill, st_avout, st_avin);

#ifdef LCM_DEBUG_INFO
  if (file)
    {
      fprintf (file, "Edge List:\n");
      verify_edge_list (file, edge_list);
      print_edge_list (file, edge_list);
734 735 736 737 738 739
      dump_sbitmap_vector (file, "transp", "", transp, last_basic_block);
      dump_sbitmap_vector (file, "st_avloc", "", st_avloc, last_basic_block);
      dump_sbitmap_vector (file, "st_antloc", "", st_antloc, last_basic_block);
      dump_sbitmap_vector (file, "st_antin", "", st_antin, last_basic_block);
      dump_sbitmap_vector (file, "st_antout", "", st_antout, last_basic_block);
      dump_sbitmap_vector (file, "st_kill", "", kill, last_basic_block);
740 741
    }
#endif
Jeffrey A Law committed
742

743 744 745
#ifdef LCM_DEBUG_INFO
  if (file)
    {
746 747
      dump_sbitmap_vector (file, "st_avout", "", st_avout, last_basic_block);
      dump_sbitmap_vector (file, "st_avin", "", st_avin, last_basic_block);
748 749
    }
#endif
Jeffrey A Law committed
750

751 752
  /* Compute farthestness.  */
  farthest = sbitmap_vector_alloc (num_edges, n_exprs);
753
  compute_farthest (edge_list, n_exprs, st_avout, st_avin, st_antin,
754 755 756 757 758 759 760
		    kill, farthest);

#ifdef LCM_DEBUG_INFO
  if (file)
    dump_sbitmap_vector (file, "farthest", "", farthest, num_edges);
#endif

761 762 763 764 765
  sbitmap_vector_free (st_antin);
  sbitmap_vector_free (st_antout);

  sbitmap_vector_free (st_avin);
  sbitmap_vector_free (st_avout);
766 767

  nearer = sbitmap_vector_alloc (num_edges, n_exprs);
768

769
  /* Allocate an extra element for the entry block.  */
770
  nearerout = sbitmap_vector_alloc (last_basic_block + 1, n_exprs);
771
  compute_nearerout (edge_list, farthest, st_avloc, nearer, nearerout);
772 773 774

#ifdef LCM_DEBUG_INFO
  if (file)
Jeffrey A Law committed
775
    {
776
      dump_sbitmap_vector (file, "nearerout", "", nearerout,
777
			   last_basic_block + 1);
778
      dump_sbitmap_vector (file, "nearer", "", nearer, num_edges);
Jeffrey A Law committed
779
    }
780 781
#endif

782
  sbitmap_vector_free (farthest);
783 784

  *insert = sbitmap_vector_alloc (num_edges, n_exprs);
785
  *delete = sbitmap_vector_alloc (last_basic_block, n_exprs);
786 787
  compute_rev_insert_delete (edge_list, st_avloc, nearer, nearerout,
			     *insert, *delete);
788

789 790
  sbitmap_vector_free (nearerout);
  sbitmap_vector_free (nearer);
791 792 793 794 795

#ifdef LCM_DEBUG_INFO
  if (file)
    {
      dump_sbitmap_vector (file, "pre_insert_map", "", *insert, num_edges);
796
      dump_sbitmap_vector (file, "pre_delete_map", "", *delete,
797
			   last_basic_block);
798 799 800
    }
#endif
  return edge_list;
Jeffrey A Law committed
801
}
802

803 804 805
/* Mode switching:

   The algorithm for setting the modes consists of scanning the insn list
806 807 808 809 810 811 812 813 814 815 816 817
   and finding all the insns which require a specific mode.  Each insn gets
   a unique struct seginfo element.  These structures are inserted into a list
   for each basic block.  For each entity, there is an array of bb_info over
   the flow graph basic blocks (local var 'bb_info'), and contains a list
   of all insns within that basic block, in the order they are encountered.

   For each entity, any basic block WITHOUT any insns requiring a specific
   mode are given a single entry, without a mode.  (Each basic block
   in the flow graph must have at least one entry in the segment table.)

   The LCM algorithm is then run over the flow graph to determine where to
   place the sets to the highest-priority value in respect of first the first
818
   insn in any one block.  Any adjustments required to the transparency
819
   vectors are made, then the next iteration starts for the next-lower
820
   priority mode, till for each entity all modes are exhausted.
821 822 823 824

   More details are located in the code for optimize_mode_switching().  */

/* This structure contains the information for each insn which requires
825
   either single or double mode to be set.
826
   MODE is the mode this insn must be executed in.
827 828
   INSN_PTR is the insn to be executed (may be the note that marks the
   beginning of a basic block).
829 830
   BBNUM is the flow graph basic block this insn occurs in.
   NEXT is the next insn in the same basic block.  */
831
struct seginfo
832 833 834 835 836 837 838 839
{
  int mode;
  rtx insn_ptr;
  int bbnum;
  struct seginfo *next;
  HARD_REG_SET regs_live;
};

Daniel Berlin committed
840
struct bb_info
841 842 843 844 845 846 847
{
  struct seginfo *seginfo;
  int computing;
};

/* These bitmaps are used for the LCM algorithm.  */

848
#ifdef OPTIMIZE_MODE_SWITCHING
849 850 851 852 853 854
static sbitmap *antic;
static sbitmap *transp;
static sbitmap *comp;
static sbitmap *delete;
static sbitmap *insert;

855 856 857 858 859
static struct seginfo * new_seginfo (int, rtx, int, HARD_REG_SET);
static void add_seginfo (struct bb_info *, struct seginfo *);
static void reg_dies (rtx, HARD_REG_SET);
static void reg_becomes_live (rtx, rtx, void *);
static void make_preds_opaque (basic_block, int);
860 861 862
#endif

#ifdef OPTIMIZE_MODE_SWITCHING
863 864

/* This function will allocate a new BBINFO structure, initialized
865
   with the MODE, INSN, and basic block BB parameters.  */
866

867
static struct seginfo *
868
new_seginfo (int mode, rtx insn, int bb, HARD_REG_SET regs_live)
869 870 871 872 873 874 875 876 877 878 879
{
  struct seginfo *ptr;
  ptr = xmalloc (sizeof (struct seginfo));
  ptr->mode = mode;
  ptr->insn_ptr = insn;
  ptr->bbnum = bb;
  ptr->next = NULL;
  COPY_HARD_REG_SET (ptr->regs_live, regs_live);
  return ptr;
}

880
/* Add a seginfo element to the end of a list.
881 882
   HEAD is a pointer to the list beginning.
   INFO is the structure to be linked in.  */
883

884
static void
885
add_seginfo (struct bb_info *head, struct seginfo *info)
886 887 888 889 890 891 892 893 894
{
  struct seginfo *ptr;

  if (head->seginfo == NULL)
    head->seginfo = info;
  else
    {
      ptr = head->seginfo;
      while (ptr->next != NULL)
895
	ptr = ptr->next;
896 897 898 899 900 901 902 903 904
      ptr->next = info;
    }
}

/* Make all predecessors of basic block B opaque, recursively, till we hit
   some that are already non-transparent, or an edge where aux is set; that
   denotes that a mode set is to be done on that edge.
   J is the bit number in the bitmaps that corresponds to the entity that
   we are currently handling mode-switching for.  */
905

906
static void
907
make_preds_opaque (basic_block b, int j)
908 909 910 911 912 913
{
  edge e;

  for (e = b->pred; e; e = e->pred_next)
    {
      basic_block pb = e->src;
914

915
      if (e->aux || ! TEST_BIT (transp[pb->index], j))
916
	continue;
917

918
      RESET_BIT (transp[pb->index], j);
919 920 921 922 923
      make_preds_opaque (pb, j);
    }
}

/* Record in LIVE that register REG died.  */
924

925
static void
926
reg_dies (rtx reg, HARD_REG_SET live)
927
{
928
  int regno, nregs;
929 930 931

  if (GET_CODE (reg) != REG)
    return;
932

933 934
  regno = REGNO (reg);
  if (regno < FIRST_PSEUDO_REGISTER)
935
    for (nregs = hard_regno_nregs[regno][GET_MODE (reg)] - 1; nregs >= 0;
936 937
	 nregs--)
      CLEAR_HARD_REG_BIT (live, regno + nregs);
938 939 940 941
}

/* Record in LIVE that register REG became live.
   This is called via note_stores.  */
942

943
static void
944
reg_becomes_live (rtx reg, rtx setter ATTRIBUTE_UNUSED, void *live)
945
{
946
  int regno, nregs;
947 948 949 950 951 952 953 954 955

  if (GET_CODE (reg) == SUBREG)
    reg = SUBREG_REG (reg);

  if (GET_CODE (reg) != REG)
    return;

  regno = REGNO (reg);
  if (regno < FIRST_PSEUDO_REGISTER)
956
    for (nregs = hard_regno_nregs[regno][GET_MODE (reg)] - 1; nregs >= 0;
957 958
	 nregs--)
      SET_HARD_REG_BIT (* (HARD_REG_SET *) live, regno + nregs);
959 960
}

961 962 963 964 965 966
/* Make sure if MODE_ENTRY is defined the MODE_EXIT is defined
   and vice versa.  */
#if defined (MODE_ENTRY) != defined (MODE_EXIT)
 #error "Both MODE_ENTRY and MODE_EXIT must be defined"
#endif

967 968
/* Find all insns that need a particular mode setting, and insert the
   necessary mode switches.  Return true if we did work.  */
969

970
int
971
optimize_mode_switching (FILE *file)
972 973
{
  rtx insn;
974 975
  int e;
  basic_block bb;
976 977 978
  int need_commit = 0;
  sbitmap *kill;
  struct edge_list *edge_list;
979
  static const int num_modes[] = NUM_MODES_FOR_MODE_SWITCHING;
980
#define N_ENTITIES ARRAY_SIZE (num_modes)
981
  int entity_map[N_ENTITIES];
Daniel Berlin committed
982
  struct bb_info *bb_info[N_ENTITIES];
983 984 985
  int i, j;
  int n_entities;
  int max_num_modes = 0;
986
  bool emited = false;
987
  basic_block post_entry ATTRIBUTE_UNUSED, pre_exit ATTRIBUTE_UNUSED;
988

989
  clear_bb_flags ();
990

991
  for (e = N_ENTITIES - 1, n_entities = 0; e >= 0; e--)
992 993
    if (OPTIMIZE_MODE_SWITCHING (e))
      {
994 995 996 997
	int entry_exit_extra = 0;

	/* Create the list of segments within each basic block.
	   If NORMAL_MODE is defined, allow for two extra
998
	   blocks split from the entry and exit block.  */
999
#if defined (MODE_ENTRY) && defined (MODE_EXIT)
1000 1001
	entry_exit_extra = 2;
#endif
1002
	bb_info[n_entities]
1003
	  = xcalloc (last_basic_block + entry_exit_extra, sizeof **bb_info);
1004 1005 1006 1007 1008
	entity_map[n_entities++] = e;
	if (num_modes[e] > max_num_modes)
	  max_num_modes = num_modes[e];
      }

1009
  if (! n_entities)
1010
    return 0;
1011

1012
#if defined (MODE_ENTRY) && defined (MODE_EXIT)
1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033
  {
    /* Split the edge from the entry block and the fallthrough edge to the
       exit block, so that we can note that there NORMAL_MODE is supplied /
       required.  */
    edge eg;
    post_entry = split_edge (ENTRY_BLOCK_PTR->succ);
    /* The only non-call predecessor at this stage is a block with a
       fallthrough edge; there can be at most one, but there could be
       none at all, e.g. when exit is called.  */
    for (pre_exit = 0, eg = EXIT_BLOCK_PTR->pred; eg; eg = eg->pred_next)
      if (eg->flags & EDGE_FALLTHRU)
	{
	  regset live_at_end = eg->src->global_live_at_end;

	  if (pre_exit)
	    abort ();
	  pre_exit = split_edge (eg);
	  COPY_REG_SET (pre_exit->global_live_at_start, live_at_end);
	  COPY_REG_SET (pre_exit->global_live_at_end, live_at_end);
	}
  }
1034 1035
#endif

1036 1037
  /* Create the bitmap vectors.  */

1038 1039 1040
  antic = sbitmap_vector_alloc (last_basic_block, n_entities);
  transp = sbitmap_vector_alloc (last_basic_block, n_entities);
  comp = sbitmap_vector_alloc (last_basic_block, n_entities);
1041

1042
  sbitmap_vector_ones (transp, last_basic_block);
1043 1044 1045 1046 1047

  for (j = n_entities - 1; j >= 0; j--)
    {
      int e = entity_map[j];
      int no_mode = num_modes[e];
Daniel Berlin committed
1048
      struct bb_info *info = bb_info[j];
1049 1050

      /* Determine what the first use (if any) need for a mode of entity E is.
1051
	 This will be the mode that is anticipatable for this block.
1052
	 Also compute the initial transparency settings.  */
1053
      FOR_EACH_BB (bb)
1054 1055 1056 1057 1058 1059
	{
	  struct seginfo *ptr;
	  int last_mode = no_mode;
	  HARD_REG_SET live_now;

	  REG_SET_TO_HARD_REG_SET (live_now,
1060
				   bb->global_live_at_start);
1061 1062
	  for (insn = BB_HEAD (bb);
	       insn != NULL && insn != NEXT_INSN (BB_END (bb));
1063 1064
	       insn = NEXT_INSN (insn))
	    {
1065
	      if (INSN_P (insn))
1066 1067 1068 1069 1070 1071 1072
		{
		  int mode = MODE_NEEDED (e, insn);
		  rtx link;

		  if (mode != no_mode && mode != last_mode)
		    {
		      last_mode = mode;
1073 1074 1075
		      ptr = new_seginfo (mode, insn, bb->index, live_now);
		      add_seginfo (info + bb->index, ptr);
		      RESET_BIT (transp[bb->index], j);
1076
		    }
1077 1078 1079
#ifdef MODE_AFTER
		  last_mode = MODE_AFTER (last_mode, insn);
#endif
1080 1081 1082 1083
		  /* Update LIVE_NOW.  */
		  for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
		    if (REG_NOTE_KIND (link) == REG_DEAD)
		      reg_dies (XEXP (link, 0), live_now);
1084

1085 1086 1087 1088 1089 1090
		  note_stores (PATTERN (insn), reg_becomes_live, &live_now);
		  for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
		    if (REG_NOTE_KIND (link) == REG_UNUSED)
		      reg_dies (XEXP (link, 0), live_now);
		}
	    }
1091

1092
	  info[bb->index].computing = last_mode;
1093 1094 1095
	  /* Check for blocks without ANY mode requirements.  */
	  if (last_mode == no_mode)
	    {
1096
	      ptr = new_seginfo (no_mode, BB_END (bb), bb->index, live_now);
1097
	      add_seginfo (info + bb->index, ptr);
1098 1099
	    }
	}
1100
#if defined (MODE_ENTRY) && defined (MODE_EXIT)
1101
      {
1102
	int mode = MODE_ENTRY (e);
1103

1104 1105
	if (mode != no_mode)
	  {
1106
	    bb = post_entry;
1107

1108 1109 1110 1111 1112
	    /* By always making this nontransparent, we save
	       an extra check in make_preds_opaque.  We also
	       need this to avoid confusing pre_edge_lcm when
	       antic is cleared but transp and comp are set.  */
	    RESET_BIT (transp[bb->index], j);
1113

1114 1115 1116 1117 1118 1119
	    /* Insert a fake computing definition of MODE into entry
	       blocks which compute no mode. This represents the mode on
	       entry.  */
	    info[bb->index].computing = mode;

	    if (pre_exit)
1120
	      info[pre_exit->index].seginfo->mode = MODE_EXIT (e);
1121 1122
	  }
      }
1123
#endif /* NORMAL_MODE */
1124 1125
    }

1126
  kill = sbitmap_vector_alloc (last_basic_block, n_entities);
1127 1128 1129 1130 1131
  for (i = 0; i < max_num_modes; i++)
    {
      int current_mode[N_ENTITIES];

      /* Set the anticipatable and computing arrays.  */
1132 1133
      sbitmap_vector_zero (antic, last_basic_block);
      sbitmap_vector_zero (comp, last_basic_block);
1134 1135 1136
      for (j = n_entities - 1; j >= 0; j--)
	{
	  int m = current_mode[j] = MODE_PRIORITY_TO_MODE (entity_map[j], i);
Daniel Berlin committed
1137
	  struct bb_info *info = bb_info[j];
1138

1139
	  FOR_EACH_BB (bb)
1140
	    {
1141 1142
	      if (info[bb->index].seginfo->mode == m)
		SET_BIT (antic[bb->index], j);
1143

1144 1145
	      if (info[bb->index].computing == m)
		SET_BIT (comp[bb->index], j);
1146 1147 1148 1149 1150 1151
	    }
	}

      /* Calculate the optimal locations for the
	 placement mode switches to modes with priority I.  */

1152 1153
      FOR_EACH_BB (bb)
	sbitmap_not (kill[bb->index], transp[bb->index]);
1154 1155 1156
      edge_list = pre_edge_lcm (file, 1, transp, comp, antic,
				kill, &insert, &delete);

1157
      for (j = n_entities - 1; j >= 0; j--)
1158 1159 1160
	{
	  /* Insert all mode sets that have been inserted by lcm.  */
	  int no_mode = num_modes[entity_map[j]];
1161

1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186
	  /* Wherever we have moved a mode setting upwards in the flow graph,
	     the blocks between the new setting site and the now redundant
	     computation ceases to be transparent for any lower-priority
	     mode of the same entity.  First set the aux field of each
	     insertion site edge non-transparent, then propagate the new
	     non-transparency from the redundant computation upwards till
	     we hit an insertion site or an already non-transparent block.  */
	  for (e = NUM_EDGES (edge_list) - 1; e >= 0; e--)
	    {
	      edge eg = INDEX_EDGE (edge_list, e);
	      int mode;
	      basic_block src_bb;
	      HARD_REG_SET live_at_edge;
	      rtx mode_set;

	      eg->aux = 0;

	      if (! TEST_BIT (insert[e], j))
		continue;

	      eg->aux = (void *)1;

	      mode = current_mode[j];
	      src_bb = eg->src;

1187 1188 1189
	      REG_SET_TO_HARD_REG_SET (live_at_edge,
				       src_bb->global_live_at_end);

1190 1191
	      start_sequence ();
	      EMIT_MODE_SET (entity_map[j], mode, live_at_edge);
1192
	      mode_set = get_insns ();
1193 1194
	      end_sequence ();

1195
	      /* Do not bother to insert empty sequence.  */
1196
	      if (mode_set == NULL_RTX)
1197 1198
		continue;

1199 1200
	      /* If this is an abnormal edge, we'll insert at the end
		 of the previous block.  */
1201 1202
	      if (eg->flags & EDGE_ABNORMAL)
		{
1203
		  emited = true;
1204 1205
		  if (GET_CODE (BB_END (src_bb)) == JUMP_INSN)
		    emit_insn_before (mode_set, BB_END (src_bb));
1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216
		  /* It doesn't make sense to switch to normal mode
		     after a CALL_INSN, so we're going to abort if we
		     find one.  The cases in which a CALL_INSN may
		     have an abnormal edge are sibcalls and EH edges.
		     In the case of sibcalls, the dest basic-block is
		     the EXIT_BLOCK, that runs in normal mode; it is
		     assumed that a sibcall insn requires normal mode
		     itself, so no mode switch would be required after
		     the call (it wouldn't make sense, anyway).  In
		     the case of EH edges, EH entry points also start
		     in normal mode, so a similar reasoning applies.  */
1217 1218
		  else if (GET_CODE (BB_END (src_bb)) == INSN)
		    emit_insn_after (mode_set, BB_END (src_bb));
1219 1220
		  else
		    abort ();
1221 1222
		  bb_info[j][src_bb->index].computing = mode;
		  RESET_BIT (transp[src_bb->index], j);
1223 1224 1225 1226 1227 1228 1229 1230
		}
	      else
		{
		  need_commit = 1;
		  insert_insn_on_edge (mode_set, eg);
		}
	    }

1231 1232
	  FOR_EACH_BB_REVERSE (bb)
	    if (TEST_BIT (delete[bb->index], j))
1233
	      {
1234
		make_preds_opaque (bb, j);
1235
		/* Cancel the 'deleted' mode set.  */
1236
		bb_info[j][bb->index].seginfo->mode = no_mode;
1237
	      }
1238
	}
1239

1240
      clear_aux_for_edges ();
1241 1242 1243 1244 1245 1246
      free_edge_list (edge_list);
    }

  /* Now output the remaining mode sets in all the segments.  */
  for (j = n_entities - 1; j >= 0; j--)
    {
1247 1248
      int no_mode = num_modes[entity_map[j]];

1249
      FOR_EACH_BB_REVERSE (bb)
1250 1251
	{
	  struct seginfo *ptr, *next;
1252
	  for (ptr = bb_info[j][bb->index].seginfo; ptr; ptr = next)
1253 1254
	    {
	      next = ptr->next;
1255
	      if (ptr->mode != no_mode)
1256 1257 1258 1259 1260
		{
		  rtx mode_set;

		  start_sequence ();
		  EMIT_MODE_SET (entity_map[j], ptr->mode, ptr->regs_live);
1261
		  mode_set = get_insns ();
1262 1263
		  end_sequence ();

1264
		  /* Do not bother to insert empty sequence.  */
1265
		  if (mode_set == NULL_RTX)
1266 1267 1268
		    continue;

		  emited = true;
1269 1270 1271
		  if (GET_CODE (ptr->insn_ptr) == NOTE
		      && (NOTE_LINE_NUMBER (ptr->insn_ptr)
			  == NOTE_INSN_BASIC_BLOCK))
1272
		    emit_insn_after (mode_set, ptr->insn_ptr);
1273
		  else
1274
		    emit_insn_before (mode_set, ptr->insn_ptr);
1275
		}
1276

1277 1278 1279
	      free (ptr);
	    }
	}
1280

1281 1282 1283 1284
      free (bb_info[j]);
    }

  /* Finished. Free up all the things we've allocated.  */
1285

1286 1287 1288 1289 1290 1291 1292 1293 1294
  sbitmap_vector_free (kill);
  sbitmap_vector_free (antic);
  sbitmap_vector_free (transp);
  sbitmap_vector_free (comp);
  sbitmap_vector_free (delete);
  sbitmap_vector_free (insert);

  if (need_commit)
    commit_edge_insertions ();
1295

1296
#if defined (MODE_ENTRY) && defined (MODE_EXIT)
1297 1298
  cleanup_cfg (CLEANUP_NO_INSN_DEL);
#else
1299 1300
  if (!need_commit && !emited)
    return 0;
1301
#endif
1302

1303 1304 1305 1306 1307
  max_regno = max_reg_num ();
  allocate_reg_info (max_regno, FALSE, FALSE);
  update_life_info_in_dirty_blocks (UPDATE_LIFE_GLOBAL_RM_NOTES,
				    (PROP_DEATH_NOTES | PROP_KILL_DEAD_CODE
				     | PROP_SCAN_DEAD_CODE));
1308 1309

  return 1;
1310
}
1311
#endif /* OPTIMIZE_MODE_SWITCHING */