ipa.c 41.7 KB
Newer Older
1
/* Basic IPA optimizations and utilities.
Jakub Jelinek committed
2
   Copyright (C) 2003-2015 Free Software Foundation, Inc.
3 4 5 6 7

This file is part of GCC.

GCC is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free
8
Software Foundation; either version 3, or (at your option) any later
9 10 11 12 13 14 15 16
version.

GCC is distributed in the hope that it will be useful, but WITHOUT ANY
WARRANTY; without even the implied warranty of MERCHANTABILITY or
FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
for more details.

You should have received a copy of the GNU General Public License
17 18
along with GCC; see the file COPYING3.  If not see
<http://www.gnu.org/licenses/>.  */
19 20 21 22

#include "config.h"
#include "system.h"
#include "coretypes.h"
23 24 25 26
#include "backend.h"
#include "tree.h"
#include "gimple.h"
#include "hard-reg-set.h"
27 28 29
#include "alias.h"
#include "options.h"
#include "fold-const.h"
30 31
#include "calls.h"
#include "stringpool.h"
Andrew MacLeod committed
32 33
#include "cgraph.h"
#include "tree-pass.h"
34
#include "gimplify.h"
35
#include "flags.h"
36 37
#include "target.h"
#include "tree-iterator.h"
38
#include "ipa-utils.h"
Andrew MacLeod committed
39
#include "alloc-pool.h"
40
#include "symbol-summary.h"
Andrew MacLeod committed
41
#include "ipa-prop.h"
42
#include "ipa-inline.h"
43 44 45
#include "tree-inline.h"
#include "profile.h"
#include "params.h"
46 47
#include "internal-fn.h"
#include "dbgcnt.h"
48

49 50 51 52 53 54 55 56

/* Return true when NODE has ADDR reference.  */

static bool
has_addr_references_p (struct cgraph_node *node,
		       void *data ATTRIBUTE_UNUSED)
{
  int i;
Martin Liska committed
57
  struct ipa_ref *ref = NULL;
58

Martin Liska committed
59
  for (i = 0; node->iterate_referring (i, ref); i++)
60 61 62 63 64
    if (ref->use == IPA_REF_ADDR)
      return true;
  return false;
}

65 66 67 68 69 70 71 72 73 74 75 76 77 78 79
/* Look for all functions inlined to NODE and update their inlined_to pointers
   to INLINED_TO.  */

static void
update_inlined_to_pointer (struct cgraph_node *node, struct cgraph_node *inlined_to)
{
  struct cgraph_edge *e;
  for (e = node->callees; e; e = e->next_callee)
    if (e->callee->global.inlined_to)
      {
        e->callee->global.inlined_to = inlined_to;
	update_inlined_to_pointer (e->callee, inlined_to);
      }
}

80
/* Add symtab NODE to queue starting at FIRST.
81 82 83 84 85 86

   The queue is linked via AUX pointers and terminated by pointer to 1.
   We enqueue nodes at two occasions: when we find them reachable or when we find
   their bodies needed for further clonning.  In the second case we mark them
   by pointer to 2 after processing so they are re-queue when they become
   reachable.  */
87 88

static void
89
enqueue_node (symtab_node *node, symtab_node **first,
90
	      hash_set<symtab_node *> *reachable)
91
{
92
  /* Node is still in queue; do nothing.  */
93
  if (node->aux && node->aux != (void *) 2)
94 95 96
    return;
  /* Node was already processed as unreachable, re-enqueue
     only if it became reachable now.  */
97
  if (node->aux == (void *)2 && !reachable->contains (node))
98
    return;
99
  node->aux = *first;
100 101 102 103 104 105
  *first = node;
}

/* Process references.  */

static void
Martin Liska committed
106
process_references (symtab_node *snode,
107
		    symtab_node **first,
108
		    bool before_inlining_p,
109
		    hash_set<symtab_node *> *reachable)
110 111
{
  int i;
Martin Liska committed
112 113
  struct ipa_ref *ref = NULL;
  for (i = 0; snode->iterate_reference (i, ref); i++)
114
    {
115
      symtab_node *node = ref->referred;
116
      symtab_node *body = node->ultimate_alias_target ();
117

118 119
      if (node->definition && !node->in_other_partition
	  && ((!DECL_EXTERNAL (node->decl) || node->alias)
120
	      || (((before_inlining_p
121 122 123 124
		    && ((TREE_CODE (node->decl) != FUNCTION_DECL
			 && optimize)
			|| (TREE_CODE (node->decl) == FUNCTION_DECL
			    && opt_for_fn (body->decl, optimize))
125 126 127 128 129
		        || (symtab->state < IPA_SSA
		            && lookup_attribute
				 ("always_inline",
			          DECL_ATTRIBUTES (body->decl))))))
		  /* We use variable constructors during late compilation for
130 131
		     constant folding.  Keep references alive so partitioning
		     knows about potential references.  */
132
		  || (TREE_CODE (node->decl) == VAR_DECL
133
		      && flag_wpa
134
		      && ctor_for_folding (node->decl)
135
		         != error_mark_node))))
136 137 138 139 140 141 142 143 144
	{
	  /* Be sure that we will not optimize out alias target
	     body.  */
	  if (DECL_EXTERNAL (node->decl)
	      && node->alias
	      && before_inlining_p)
	    reachable->add (body);
	  reachable->add (node);
	}
145
      enqueue_node (node, first, reachable);
146 147 148
    }
}

149 150 151 152 153 154 155 156 157
/* EDGE is an polymorphic call.  If BEFORE_INLINING_P is set, mark
   all its potential targets as reachable to permit later inlining if
   devirtualization happens.  After inlining still keep their declarations
   around, so we can devirtualize to a direct call.

   Also try to make trivial devirutalization when no or only one target is
   possible.  */

static void
158
walk_polymorphic_call_targets (hash_set<void *> *reachable_call_targets,
159
			       struct cgraph_edge *edge,
160
			       symtab_node **first,
161 162
			       hash_set<symtab_node *> *reachable,
			       bool before_inlining_p)
163 164 165 166 167 168 169 170
{
  unsigned int i;
  void *cache_token;
  bool final;
  vec <cgraph_node *>targets
    = possible_polymorphic_call_targets
	(edge, &final, &cache_token);

171
  if (!reachable_call_targets->add (cache_token))
172
    {
173
      for (i = 0; i < targets.length (); i++)
174 175 176 177 178 179
	{
	  struct cgraph_node *n = targets[i];

	  /* Do not bother to mark virtual methods in anonymous namespace;
	     either we will find use of virtual table defining it, or it is
	     unused.  */
180
	  if (TREE_CODE (TREE_TYPE (n->decl)) == METHOD_TYPE
181
	      && type_in_anonymous_namespace_p
182
		    (TYPE_METHOD_BASETYPE (TREE_TYPE (n->decl))))
183 184
	    continue;

185 186
	   symtab_node *body = n->function_symbol ();

187 188
	  /* Prior inlining, keep alive bodies of possible targets for
	     devirtualization.  */
189
	   if (n->definition
190
	       && (before_inlining_p
191 192 193 194 195 196 197 198 199 200 201
		   && opt_for_fn (body->decl, optimize)
		   && opt_for_fn (body->decl, flag_devirtualize)))
	      {
		 /* Be sure that we will not optimize out alias target
		    body.  */
		 if (DECL_EXTERNAL (n->decl)
		     && n->alias
		     && before_inlining_p)
		   reachable->add (body);
		reachable->add (n);
	      }
202 203 204
	  /* Even after inlining we want to keep the possible targets in the
	     boundary, so late passes can still produce direct call even if
	     the chance for inlining is lost.  */
205
	  enqueue_node (n, first, reachable);
206 207 208 209 210 211 212 213 214
	}
    }

  /* Very trivial devirtualization; when the type is
     final or anonymous (so we know all its derivation)
     and there is only one possible virtual call target,
     make the edge direct.  */
  if (final)
    {
215
      if (targets.length () <= 1 && dbg_cnt (devirt))
216
	{
217
	  cgraph_node *target, *node = edge->caller;
218 219 220
	  if (targets.length () == 1)
	    target = targets[0];
	  else
Martin Liska committed
221
	    target = cgraph_node::get_create
222 223
		       (builtin_decl_implicit (BUILT_IN_UNREACHABLE));

224 225
	  if (dump_enabled_p ())
            {
226 227 228 229 230
	      location_t locus;
	      if (edge->call_stmt)
		locus = gimple_location (edge->call_stmt);
	      else
		locus = UNKNOWN_LOCATION;
Martin Liska committed
231
	      dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, locus,
232 233 234 235 236
                               "devirtualizing call in %s/%i to %s/%i\n",
                               edge->caller->name (), edge->caller->order,
                               target->name (),
                               target->order);
	    }
Martin Liska committed
237
	  edge = edge->make_direct (target);
238
	  if (inline_summaries)
239
	    inline_update_overall_summary (node);
240
	  else if (edge->call_stmt)
Ilya Enkovich committed
241 242 243 244 245 246 247
	    {
	      edge->redirect_call_stmt_to_callee ();

	      /* Call to __builtin_unreachable shouldn't be instrumented.  */
	      if (!targets.length ())
		gimple_call_set_with_bounds (edge->call_stmt, false);
	    }
248 249 250
	}
    }
}
251

252
/* Perform reachability analysis and reclaim all unreachable nodes.
253 254 255 256 257 258 259 260 261 262 263 264

   The algorithm is basically mark&sweep but with some extra refinements:

   - reachable extern inline functions needs special handling; the bodies needs
     to stay in memory until inlining in hope that they will be inlined.
     After inlining we release their bodies and turn them into unanalyzed
     nodes even when they are reachable.

   - virtual functions are kept in callgraph even if they seem unreachable in
     hope calls to them will be devirtualized. 

     Again we remove them after inlining.  In late optimization some
265
     devirtualization may happen, but it is not important since we won't inline
266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281
     the call. In theory early opts and IPA should work out all important cases.

   - virtual clones needs bodies of their origins for later materialization;
     this means that we want to keep the body even if the origin is unreachable
     otherwise.  To avoid origin from sitting in the callgraph and being
     walked by IPA passes, we turn them into unanalyzed nodes with body
     defined.

     We maintain set of function declaration where body needs to stay in
     body_needed_for_clonning

     Inline clones represent special case: their declaration match the
     declaration of origin and cgraph_remove_node already knows how to
     reshape callgraph and preserve body when offline copy of function or
     inline clone is being removed.

282 283 284 285 286 287
   - C++ virtual tables keyed to other unit are represented as DECL_EXTERNAL
     variables with DECL_INITIAL set.  We finalize these and keep reachable
     ones around for constant folding purposes.  After inlining we however
     stop walking their references to let everything static referneced by them
     to be removed when it is otherwise unreachable.

288 289 290 291 292
   We maintain queue of both reachable symbols (i.e. defined symbols that needs
   to stay) and symbols that are in boundary (i.e. external symbols referenced
   by reachable symbols or origins of clones).  The queue is represented
   as linked list by AUX pointer terminated by 1.

293
   At the end we keep all reachable symbols. For symbols in boundary we always
294 295 296 297 298 299 300 301 302 303
   turn definition into a declaration, but we may keep function body around
   based on body_needed_for_clonning

   All symbols that enter the queue have AUX pointer non-zero and are in the
   boundary.  Pointer set REACHABLE is used to track reachable symbols.

   Every symbol can be visited twice - once as part of boundary and once
   as real reachable symbol. enqueue_node needs to decide whether the
   node needs to be re-queued for second processing.  For this purpose
   we set AUX pointer of processed symbols in the boundary to constant 2.  */
304 305

bool
306
symbol_table::remove_unreachable_nodes (FILE *file)
307
{
308
  symtab_node *first = (symtab_node *) (void *) 1;
309
  struct cgraph_node *node, *next;
310
  varpool_node *vnode, *vnext;
311
  bool changed = false;
312 313 314
  hash_set<symtab_node *> reachable;
  hash_set<tree> body_needed_for_clonning;
  hash_set<void *> reachable_call_targets;
315 316
  bool before_inlining_p = symtab->state < (!optimize ? IPA_SSA
					    : IPA_SSA_AFTER_INLINING);
317

318
  timevar_push (TV_IPA_UNREACHABLE);
319
  build_type_inheritance_graph ();
320 321
  if (file)
    fprintf (file, "\nReclaiming functions:");
322
#ifdef ENABLE_CHECKING
323
  FOR_EACH_FUNCTION (node)
324
    gcc_assert (!node->aux);
325
  FOR_EACH_VARIABLE (vnode)
326
    gcc_assert (!vnode->aux);
327
#endif
328 329 330 331
  /* Mark functions whose bodies are obviously needed.
     This is mostly when they can be referenced externally.  Inline clones
     are special since their declarations are shared with master clone and thus
     cgraph_can_remove_if_no_direct_calls_and_refs_p should not be called on them.  */
332 333 334
  FOR_EACH_FUNCTION (node)
    {
      node->used_as_abstract_origin = false;
335
      if (node->definition
336
	  && !node->global.inlined_to
337
	  && !node->in_other_partition
Martin Liska committed
338
	  && !node->can_remove_if_no_direct_calls_and_refs_p ())
339 340
	{
	  gcc_assert (!node->global.inlined_to);
341 342
	  reachable.add (node);
	  enqueue_node (node, &first, &reachable);
343 344
	}
      else
345
	gcc_assert (!node->aux);
346
     }
347 348

  /* Mark variables that are obviously needed.  */
349
  FOR_EACH_DEFINED_VARIABLE (vnode)
Martin Liska committed
350
    if (!vnode->can_remove_if_no_refs_p()
351
	&& !vnode->in_other_partition)
352
      {
353 354
	reachable.add (vnode);
	enqueue_node (vnode, &first, &reachable);
355 356 357
      }

  /* Perform reachability analysis.  */
358
  while (first != (symtab_node *) (void *) 1)
359
    {
360
      bool in_boundary_p = !reachable.contains (first);
361
      symtab_node *node = first;
362

363
      first = (symtab_node *)first->aux;
364

365 366 367
      /* If we are processing symbol in boundary, mark its AUX pointer for
	 possible later re-processing in enqueue_node.  */
      if (in_boundary_p)
368 369 370 371 372
	{
	  node->aux = (void *)2;
	  if (node->alias && node->analyzed)
	    enqueue_node (node->get_alias_target (), &first, &reachable);
	}
373 374
      else
	{
375 376
	  if (TREE_CODE (node->decl) == FUNCTION_DECL
	      && DECL_ABSTRACT_ORIGIN (node->decl))
377 378
	    {
	      struct cgraph_node *origin_node
379 380 381 382 383 384 385 386 387 388 389
	      = cgraph_node::get (DECL_ABSTRACT_ORIGIN (node->decl));
	      if (origin_node && !origin_node->used_as_abstract_origin)
		{
	          origin_node->used_as_abstract_origin = true;
		  gcc_assert (!origin_node->prev_sibling_clone);
		  gcc_assert (!origin_node->next_sibling_clone);
		  for (cgraph_node *n = origin_node->clones; n;
		       n = n->next_sibling_clone)
		    if (n->decl == DECL_ABSTRACT_ORIGIN (node->decl))
		      n->used_as_abstract_origin = true;
		}
390
	    }
391
	  /* If any symbol in a comdat group is reachable, force
392 393 394
	     all externally visible symbols in the same comdat
	     group to be reachable as well.  Comdat-local symbols
	     can be discarded if all uses were inlined.  */
395
	  if (node->same_comdat_group)
396
	    {
397
	      symtab_node *next;
398
	      for (next = node->same_comdat_group;
399
		   next != node;
400
		   next = next->same_comdat_group)
Martin Liska committed
401
		if (!next->comdat_local_p ()
402 403
		    && !reachable.add (next))
		  enqueue_node (next, &first, &reachable);
404 405
	    }
	  /* Mark references as reachable.  */
406
	  process_references (node, &first, before_inlining_p, &reachable);
407
	}
408

409
      if (cgraph_node *cnode = dyn_cast <cgraph_node *> (node))
410
	{
411 412 413
	  /* Mark the callees reachable unless they are direct calls to extern
 	     inline functions we decided to not inline.  */
	  if (!in_boundary_p)
414
	    {
415
	      struct cgraph_edge *e;
416
	      /* Keep alive possible targets for devirtualization.  */
417 418
	      if (opt_for_fn (cnode->decl, optimize)
		  && opt_for_fn (cnode->decl, flag_devirtualize))
419 420 421 422 423 424
		{
		  struct cgraph_edge *next;
		  for (e = cnode->indirect_calls; e; e = next)
		    {
		      next = e->next_callee;
		      if (e->indirect_info->polymorphic)
425 426
			walk_polymorphic_call_targets (&reachable_call_targets,
						       e, &first, &reachable,
427 428 429
						       before_inlining_p);
		    }
		}
430
	      for (e = cnode->callees; e; e = e->next_callee)
431
		{
432
	          symtab_node *body = e->callee->function_symbol ();
433 434
		  if (e->callee->definition
		      && !e->callee->in_other_partition
435
		      && (!e->inline_failed
436 437
			  || !DECL_EXTERNAL (e->callee->decl)
			  || e->callee->alias
438 439 440 441 442 443
			  || (before_inlining_p
			      && (opt_for_fn (body->decl, optimize)
		                  || (symtab->state < IPA_SSA
		                      && lookup_attribute
				          ("always_inline",
				           DECL_ATTRIBUTES (body->decl)))))))
444 445 446 447 448 449
		    {
		      /* Be sure that we will not optimize out alias target
			 body.  */
		      if (DECL_EXTERNAL (e->callee->decl)
			  && e->callee->alias
			  && before_inlining_p)
450
			reachable.add (body);
451
		      reachable.add (e->callee);
452
		    }
453
		  enqueue_node (e->callee, &first, &reachable);
454
		}
455 456 457

	      /* When inline clone exists, mark body to be preserved so when removing
		 offline copy of the function we don't kill it.  */
458
	      if (cnode->global.inlined_to)
459
	        body_needed_for_clonning.add (cnode->decl);
460

461 462 463 464 465 466 467 468 469 470 471 472 473 474
	      /* For instrumentation clones we always need original
		 function node for proper LTO privatization.  */
	      if (cnode->instrumentation_clone
		  && cnode->definition)
		{
		  gcc_assert (cnode->instrumented_version || in_lto_p);
		  if (cnode->instrumented_version)
		    {
		      enqueue_node (cnode->instrumented_version, &first,
				    &reachable);
		      reachable.add (cnode->instrumented_version);
		    }
		}

475 476 477 478
	      /* For non-inline clones, force their origins to the boundary and ensure
		 that body is not removed.  */
	      while (cnode->clone_of)
		{
479
		  bool noninline = cnode->clone_of->decl != cnode->decl;
480 481 482
		  cnode = cnode->clone_of;
		  if (noninline)
		    {
483 484
		      body_needed_for_clonning.add (cnode->decl);
		      enqueue_node (cnode, &first, &reachable);
485
		    }
486
		}
487 488

	    }
489 490
	  else if (cnode->thunk.thunk_p)
	    enqueue_node (cnode->callees->callee, &first, &reachable);
491

492 493 494 495 496 497 498 499 500
	  /* If any reachable function has simd clones, mark them as
	     reachable as well.  */
	  if (cnode->simd_clones)
	    {
	      cgraph_node *next;
	      for (next = cnode->simd_clones;
		   next;
		   next = next->simdclone->next_clone)
		if (in_boundary_p
501 502
		    || !reachable.add (next))
		  enqueue_node (next, &first, &reachable);
503
	    }
504
	}
505
      /* When we see constructor of external variable, keep referred nodes in the
506 507
	boundary.  This will also hold initializers of the external vars NODE
	refers to.  */
508
      varpool_node *vnode = dyn_cast <varpool_node *> (node);
509
      if (vnode
510 511
	  && DECL_EXTERNAL (node->decl)
	  && !vnode->alias
512
	  && in_boundary_p)
513
	{
Martin Liska committed
514 515
	  struct ipa_ref *ref = NULL;
	  for (int i = 0; node->iterate_reference (i, ref); i++)
516
	    enqueue_node (ref->referred, &first, &reachable);
517
	}
518 519
    }

520
  /* Remove unreachable functions.   */
Martin Liska committed
521
  for (node = first_function (); node; node = next)
522
    {
Martin Liska committed
523
      next = next_function (node);
524 525

      /* If node is not needed at all, remove it.  */
526
      if (!node->aux)
527
	{
528
	  if (file)
529
	    fprintf (file, " %s/%i", node->name (), node->order);
Martin Liska committed
530
	  node->remove ();
531 532
	  changed = true;
	}
533
      /* If node is unreachable, remove its body.  */
534
      else if (!reachable.contains (node))
535
        {
536 537 538 539 540 541 542
	  /* We keep definitions of thunks and aliases in the boundary so
	     we can walk to the ultimate alias targets and function symbols
	     reliably.  */
	  if (node->alias || node->thunk.thunk_p)
	    ;
	  else if (!body_needed_for_clonning.contains (node->decl)
	      && !node->alias && !node->thunk.thunk_p)
Martin Liska committed
543
	    node->release_body ();
544
	  else if (!node->clone_of)
545
	    gcc_assert (in_lto_p || DECL_RESULT (node->decl));
546
	  if (node->definition && !node->alias && !node->thunk.thunk_p)
547
	    {
548
	      if (file)
549
		fprintf (file, " %s/%i", node->name (), node->order);
550
	      node->body_removed = true;
551 552 553 554
	      node->analyzed = false;
	      node->definition = false;
	      node->cpp_implicit_alias = false;
	      node->alias = false;
555
	      node->thunk.thunk_p = false;
556
	      node->weakref = false;
557 558 559 560 561 562
	      /* After early inlining we drop always_inline attributes on
		 bodies of functions that are still referenced (have their
		 address taken).  */
	      DECL_ATTRIBUTES (node->decl)
		= remove_attribute ("always_inline",
				    DECL_ATTRIBUTES (node->decl));
563
	      if (!node->in_other_partition)
564
		node->local.local = false;
Martin Liska committed
565
	      node->remove_callees ();
Martin Liska committed
566
	      node->remove_all_references ();
567
	      changed = true;
Ilya Enkovich committed
568 569 570 571 572 573
	      if (node->thunk.thunk_p
		  && node->thunk.add_pointer_bounds_args)
		{
		  node->thunk.thunk_p = false;
		  node->thunk.add_pointer_bounds_args = false;
		}
574
	    }
575
	}
576
      else
Martin Liska committed
577
	gcc_assert (node->clone_of || !node->has_gimple_body_p ()
578
		    || in_lto_p || DECL_RESULT (node->decl));
579
    }
580 581 582 583

  /* Inline clones might be kept around so their materializing allows further
     cloning.  If the function the clone is inlined into is removed, we need
     to turn it into normal cone.  */
584
  FOR_EACH_FUNCTION (node)
585 586 587 588 589
    {
      if (node->global.inlined_to
	  && !node->callers)
	{
	  gcc_assert (node->clones);
590 591
	  node->global.inlined_to = NULL;
	  update_inlined_to_pointer (node, node);
592
	}
593
      node->aux = NULL;
594
    }
595

596
  /* Remove unreachable variables.  */
597
  if (file)
598
    fprintf (file, "\nReclaiming variables:");
Martin Liska committed
599
  for (vnode = first_variable (); vnode; vnode = vnext)
600
    {
Martin Liska committed
601
      vnext = next_variable (vnode);
602
      if (!vnode->aux
603 604 605
	  /* For can_refer_decl_in_current_unit_p we want to track for
	     all external variables if they are defined in other partition
	     or not.  */
606
	  && (!flag_ltrans || !DECL_EXTERNAL (vnode->decl)))
607
	{
608 609 610 611 612 613 614 615 616 617 618
	  struct ipa_ref *ref = NULL;

	  /* First remove the aliases, so varpool::remove can possibly lookup
	     the constructor and save it for future use.  */
	  while (vnode->iterate_direct_aliases (0, ref))
	    {
	      if (file)
		fprintf (file, " %s/%i", ref->referred->name (),
			 ref->referred->order);
	      ref->referring->remove ();
	    }
619
	  if (file)
620
	    fprintf (file, " %s/%i", vnode->name (), vnode->order);
621
          vnext = next_variable (vnode);
Martin Liska committed
622
	  vnode->remove ();
623
	  changed = true;
624
	}
625
      else if (!reachable.contains (vnode) && !vnode->alias)
626
        {
627
	  tree init;
628
	  if (vnode->definition)
629 630
	    {
	      if (file)
631
		fprintf (file, " %s", vnode->name ());
632 633
	      changed = true;
	    }
634
	  /* Keep body if it may be useful for constant folding.  */
Ilya Enkovich committed
635 636
	  if ((init = ctor_for_folding (vnode->decl)) == error_mark_node
	      && !POINTER_BOUNDS_P (vnode->decl))
637 638 639
	    vnode->remove_initializer ();
	  else
	    DECL_INITIAL (vnode->decl) = init;
640
	  vnode->body_removed = true;
641 642 643
	  vnode->definition = false;
	  vnode->analyzed = false;
	  vnode->aux = NULL;
644

Martin Liska committed
645
	  vnode->remove_from_same_comdat_group ();
646

Martin Liska committed
647
	  vnode->remove_all_references ();
648 649
	}
      else
650
	vnode->aux = NULL;
651
    }
652

653
  /* Now update address_taken flags and try to promote functions to be local.  */
654 655
  if (file)
    fprintf (file, "\nClearing address taken flags:");
656
  FOR_EACH_DEFINED_FUNCTION (node)
657 658
    if (node->address_taken
	&& !node->used_from_other_partition)
659
      {
660
	if (!node->call_for_symbol_and_aliases
Ilya Enkovich committed
661 662 663 664
	    (has_addr_references_p, NULL, true)
	    && (!node->instrumentation_clone
		|| !node->instrumented_version
		|| !node->instrumented_version->address_taken))
665 666
	  {
	    if (file)
667
	      fprintf (file, " %s", node->name ());
668
	    node->address_taken = false;
669
	    changed = true;
Martin Liska committed
670
	    if (node->local_p ())
671 672 673 674 675
	      {
		node->local.local = true;
		if (file)
		  fprintf (file, " (local)");
	      }
676 677
	  }
      }
678 679
  if (file)
    fprintf (file, "\n");
680

681
#ifdef ENABLE_CHECKING
Martin Liska committed
682
  symtab_node::verify_symtab_nodes ();
683
#endif
Diego Novillo committed
684

685
  /* If we removed something, perhaps profile could be improved.  */
686
  if (changed && optimize && inline_edge_summary_vec.exists ())
687
    FOR_EACH_DEFINED_FUNCTION (node)
688
      ipa_propagate_frequency (node);
689

690
  timevar_pop (TV_IPA_UNREACHABLE);
691 692
  return changed;
}
693

694 695 696 697 698 699 700 701 702 703 704 705
/* Process references to VNODE and set flags WRITTEN, ADDRESS_TAKEN, READ
   as needed, also clear EXPLICIT_REFS if the references to given variable
   do not need to be explicit.  */

void
process_references (varpool_node *vnode,
		    bool *written, bool *address_taken,
		    bool *read, bool *explicit_refs)
{
  int i;
  struct ipa_ref *ref;

Martin Liska committed
706
  if (!vnode->all_refs_explicit_p ()
707 708 709
      || TREE_THIS_VOLATILE (vnode->decl))
    *explicit_refs = false;

Martin Liska committed
710
  for (i = 0; vnode->iterate_referring (i, ref)
711 712 713 714 715 716 717 718 719 720 721 722 723
	      && *explicit_refs && (!*written || !*address_taken || !*read); i++)
    switch (ref->use)
      {
      case IPA_REF_ADDR:
	*address_taken = true;
	break;
      case IPA_REF_LOAD:
	*read = true;
	break;
      case IPA_REF_STORE:
	*written = true;
	break;
      case IPA_REF_ALIAS:
Martin Liska committed
724 725
	process_references (dyn_cast<varpool_node *> (ref->referring), written,
			    address_taken, read, explicit_refs);
726
	break;
Ilya Enkovich committed
727 728
      case IPA_REF_CHKP:
	gcc_unreachable ();
729 730 731 732 733 734 735 736 737 738 739 740 741 742 743
      }
}

/* Set TREE_READONLY bit.  */

bool
set_readonly_bit (varpool_node *vnode, void *data ATTRIBUTE_UNUSED)
{
  TREE_READONLY (vnode->decl) = true;
  return false;
}

/* Set writeonly bit and clear the initalizer, since it will not be needed.  */

bool
744
set_writeonly_bit (varpool_node *vnode, void *data)
745 746 747 748 749 750
{
  vnode->writeonly = true;
  if (optimize)
    {
      DECL_INITIAL (vnode->decl) = NULL;
      if (!vnode->alias)
751 752 753 754 755
	{
	  if (vnode->num_references ())
	    *(bool *)data = true;
	  vnode->remove_all_references ();
	}
756 757 758 759 760 761 762 763 764 765 766 767 768 769
    }
  return false;
}

/* Clear addressale bit of VNODE.  */

bool
clear_addressable_bit (varpool_node *vnode, void *data ATTRIBUTE_UNUSED)
{
  vnode->address_taken = false;
  TREE_ADDRESSABLE (vnode->decl) = 0;
  return false;
}

770 771 772
/* Discover variables that have no longer address taken or that are read only
   and update their flags.

773 774
   Return true when unreachable symbol removan should be done.

775 776
   FIXME: This can not be done in between gimplify and omp_expand since
   readonly flag plays role on what is shared and what is not.  Currently we do
777 778 779
   this transformation as part of whole program visibility and re-do at
   ipa-reference pass (to take into account clonning), but it would
   make sense to do it before early optimizations.  */
780

781
bool
782 783
ipa_discover_readonly_nonaddressable_vars (void)
{
784
  bool remove_p = false;
785
  varpool_node *vnode;
786 787
  if (dump_file)
    fprintf (dump_file, "Clearing variable flags:");
788
  FOR_EACH_VARIABLE (vnode)
789
    if (!vnode->alias
790
	&& (TREE_ADDRESSABLE (vnode->decl)
791
	    || !vnode->writeonly
792
	    || !TREE_READONLY (vnode->decl)))
793 794 795
      {
	bool written = false;
	bool address_taken = false;
796 797 798
	bool read = false;
	bool explicit_refs = true;

799 800
	process_references (vnode, &written, &address_taken, &read,
			    &explicit_refs);
801 802 803
	if (!explicit_refs)
	  continue;
	if (!address_taken)
804
	  {
805
	    if (TREE_ADDRESSABLE (vnode->decl) && dump_file)
806
	      fprintf (dump_file, " %s (non-addressable)", vnode->name ());
807 808
	    vnode->call_for_symbol_and_aliases (clear_addressable_bit, NULL,
					        true);
809
	  }
810
	if (!address_taken && !written
811 812 813
	    /* Making variable in explicit section readonly can cause section
	       type conflict. 
	       See e.g. gcc.c-torture/compile/pr23237.c */
814
	    && vnode->get_section () == NULL)
815
	  {
816
	    if (!TREE_READONLY (vnode->decl) && dump_file)
817
	      fprintf (dump_file, " %s (read-only)", vnode->name ());
818
	    vnode->call_for_symbol_and_aliases (set_readonly_bit, NULL, true);
819
	  }
820
	if (!vnode->writeonly && !read && !address_taken && written)
821 822 823
	  {
	    if (dump_file)
	      fprintf (dump_file, " %s (write-only)", vnode->name ());
824 825
	    vnode->call_for_symbol_and_aliases (set_writeonly_bit, &remove_p, 
					        true);
826 827 828 829
	  }
      }
  if (dump_file)
    fprintf (dump_file, "\n");
830
  return remove_p;
831 832
}

833 834
/* Free inline summary.  */

835 836 837
namespace {

const pass_data pass_data_ipa_free_inline_summary =
838
{
839
  SIMPLE_IPA_PASS, /* type */
840
  "free-inline-summary", /* name */
841 842 843 844 845 846
  OPTGROUP_NONE, /* optinfo_flags */
  TV_IPA_FREE_INLINE_SUMMARY, /* tv_id */
  0, /* properties_required */
  0, /* properties_provided */
  0, /* properties_destroyed */
  0, /* todo_flags_start */
847 848 849 850
  /* Early optimizations may make function unreachable.  We can not
     remove unreachable functions as part of the ealry opts pass because
     TODOs are run before subpasses.  Do it here.  */
  ( TODO_remove_functions | TODO_dump_symtab ), /* todo_flags_finish */
851 852
};

853
class pass_ipa_free_inline_summary : public simple_ipa_opt_pass
854 855
{
public:
856 857
  pass_ipa_free_inline_summary (gcc::context *ctxt)
    : simple_ipa_opt_pass (pass_data_ipa_free_inline_summary, ctxt)
858 859 860
  {}

  /* opt_pass methods: */
861 862 863 864 865
  virtual unsigned int execute (function *)
    {
      inline_free_summary ();
      return 0;
    }
866 867 868

}; // class pass_ipa_free_inline_summary

869 870
} // anon namespace

871 872 873 874 875 876
simple_ipa_opt_pass *
make_pass_ipa_free_inline_summary (gcc::context *ctxt)
{
  return new pass_ipa_free_inline_summary (ctxt);
}

877
/* Generate and emit a static constructor or destructor.  WHICH must
Ilya Enkovich committed
878 879 880 881 882
   be one of 'I' (for a constructor), 'D' (for a destructor), 'P'
   (for chp static vars constructor) or 'B' (for chkp static bounds
   constructor).  BODY is a STATEMENT_LIST containing GENERIC
   statements.  PRIORITY is the initialization priority for this
   constructor or destructor.
883

884 885 886 887 888
   FINAL specify whether the externally visible name for collect2 should
   be produced. */

static void
cgraph_build_static_cdtor_1 (char which, tree body, int priority, bool final)
889 890 891 892 893 894 895 896
{
  static int counter = 0;
  char which_buf[16];
  tree decl, name, resdecl;

  /* The priority is encoded in the constructor or destructor name.
     collect2 will sort the names and arrange that they are called at
     program startup.  */
897 898 899 900 901 902
  if (final)
    sprintf (which_buf, "%c_%.5d_%d", which, priority, counter++);
  else
  /* Proudce sane name but one not recognizable by collect2, just for the
     case we fail to inline the function.  */
    sprintf (which_buf, "sub_%c_%.5d_%d", which, priority, counter++);
903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919
  name = get_file_function_name (which_buf);

  decl = build_decl (input_location, FUNCTION_DECL, name,
		     build_function_type_list (void_type_node, NULL_TREE));
  current_function_decl = decl;

  resdecl = build_decl (input_location,
			RESULT_DECL, NULL_TREE, void_type_node);
  DECL_ARTIFICIAL (resdecl) = 1;
  DECL_RESULT (decl) = resdecl;
  DECL_CONTEXT (resdecl) = decl;

  allocate_struct_function (decl, false);

  TREE_STATIC (decl) = 1;
  TREE_USED (decl) = 1;
  DECL_ARTIFICIAL (decl) = 1;
920
  DECL_IGNORED_P (decl) = 1;
921 922
  DECL_NO_INSTRUMENT_FUNCTION_ENTRY_EXIT (decl) = 1;
  DECL_SAVED_TREE (decl) = body;
923
  if (!targetm.have_ctors_dtors && final)
924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941
    {
      TREE_PUBLIC (decl) = 1;
      DECL_PRESERVE_P (decl) = 1;
    }
  DECL_UNINLINABLE (decl) = 1;

  DECL_INITIAL (decl) = make_node (BLOCK);
  TREE_USED (DECL_INITIAL (decl)) = 1;

  DECL_SOURCE_LOCATION (decl) = input_location;
  cfun->function_end_locus = input_location;

  switch (which)
    {
    case 'I':
      DECL_STATIC_CONSTRUCTOR (decl) = 1;
      decl_init_priority_insert (decl, priority);
      break;
Ilya Enkovich committed
942 943 944 945 946 947 948 949 950 951 952 953 954 955
    case 'P':
      DECL_STATIC_CONSTRUCTOR (decl) = 1;
      DECL_ATTRIBUTES (decl) = tree_cons (get_identifier ("chkp ctor"),
					  NULL,
					  NULL_TREE);
      decl_init_priority_insert (decl, priority);
      break;
    case 'B':
      DECL_STATIC_CONSTRUCTOR (decl) = 1;
      DECL_ATTRIBUTES (decl) = tree_cons (get_identifier ("bnd_legacy"),
					  NULL,
					  NULL_TREE);
      decl_init_priority_insert (decl, priority);
      break;
956 957 958 959 960 961 962 963 964 965
    case 'D':
      DECL_STATIC_DESTRUCTOR (decl) = 1;
      decl_fini_priority_insert (decl, priority);
      break;
    default:
      gcc_unreachable ();
    }

  gimplify_function_tree (decl);

Martin Liska committed
966
  cgraph_node::add_new_function (decl, false);
967 968 969 970 971

  set_cfun (NULL);
  current_function_decl = NULL;
}

972
/* Generate and emit a static constructor or destructor.  WHICH must
Ilya Enkovich committed
973 974 975 976 977
   be one of 'I' (for a constructor), 'D' (for a destructor), 'P'
   (for chkp static vars constructor) or 'B' (for chkp static bounds
   constructor).  BODY is a STATEMENT_LIST containing GENERIC
   statements.  PRIORITY is the initialization priority for this
   constructor or destructor.  */
978 979 980 981 982 983

void
cgraph_build_static_cdtor (char which, tree body, int priority)
{
  cgraph_build_static_cdtor_1 (which, body, priority, false);
}
984 985

/* A vector of FUNCTION_DECLs declared as static constructors.  */
986
static vec<tree> static_ctors;
987
/* A vector of FUNCTION_DECLs declared as static destructors.  */
988
static vec<tree> static_dtors;
989 990 991 992 993 994 995 996 997 998 999

/* When target does not have ctors and dtors, we call all constructor
   and destructor by special initialization/destruction function
   recognized by collect2.

   When we are going to build this function, collect all constructors and
   destructors and turn them into normal functions.  */

static void
record_cdtor_fn (struct cgraph_node *node)
{
1000 1001 1002 1003
  if (DECL_STATIC_CONSTRUCTOR (node->decl))
    static_ctors.safe_push (node->decl);
  if (DECL_STATIC_DESTRUCTOR (node->decl))
    static_dtors.safe_push (node->decl);
Martin Liska committed
1004
  node = cgraph_node::get (node->decl);
1005
  DECL_DISREGARD_INLINE_LIMITS (node->decl) = 1;
1006 1007 1008 1009 1010 1011 1012 1013
}

/* Define global constructors/destructor functions for the CDTORS, of
   which they are LEN.  The CDTORS are sorted by initialization
   priority.  If CTOR_P is true, these are constructors; otherwise,
   they are destructors.  */

static void
1014
build_cdtor (bool ctor_p, vec<tree> cdtors)
1015 1016
{
  size_t i,j;
1017
  size_t len = cdtors.length ();
1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031

  i = 0;
  while (i < len)
    {
      tree body;
      tree fn;
      priority_type priority;

      priority = 0;
      body = NULL_TREE;
      j = i;
      do
	{
	  priority_type p;
1032
	  fn = cdtors[j];
1033 1034 1035 1036 1037 1038 1039 1040 1041
	  p = ctor_p ? DECL_INIT_PRIORITY (fn) : DECL_FINI_PRIORITY (fn);
	  if (j == i)
	    priority = p;
	  else if (p != priority)
	    break;
	  j++;
	}
      while (j < len);

1042
      /* When there is only one cdtor and target supports them, do nothing.  */
1043 1044 1045 1046 1047 1048 1049 1050
      if (j == i + 1
	  && targetm.have_ctors_dtors)
	{
	  i++;
	  continue;
	}
      /* Find the next batch of constructors/destructors with the same
	 initialization priority.  */
1051
      for (;i < j; i++)
1052 1053
	{
	  tree call;
1054
	  fn = cdtors[i];
1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068
	  call = build_call_expr (fn, 0);
	  if (ctor_p)
	    DECL_STATIC_CONSTRUCTOR (fn) = 0;
	  else
	    DECL_STATIC_DESTRUCTOR (fn) = 0;
	  /* We do not want to optimize away pure/const calls here.
	     When optimizing, these should be already removed, when not
	     optimizing, we want user to be able to breakpoint in them.  */
	  TREE_SIDE_EFFECTS (call) = 1;
	  append_to_statement_list (call, &body);
	}
      gcc_assert (body != NULL_TREE);
      /* Generate a function to call all the function of like
	 priority.  */
1069
      cgraph_build_static_cdtor_1 (ctor_p ? 'I' : 'D', body, priority, true);
1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132
    }
}

/* Comparison function for qsort.  P1 and P2 are actually of type
   "tree *" and point to static constructors.  DECL_INIT_PRIORITY is
   used to determine the sort order.  */

static int
compare_ctor (const void *p1, const void *p2)
{
  tree f1;
  tree f2;
  int priority1;
  int priority2;

  f1 = *(const tree *)p1;
  f2 = *(const tree *)p2;
  priority1 = DECL_INIT_PRIORITY (f1);
  priority2 = DECL_INIT_PRIORITY (f2);

  if (priority1 < priority2)
    return -1;
  else if (priority1 > priority2)
    return 1;
  else
    /* Ensure a stable sort.  Constructors are executed in backwarding
       order to make LTO initialize braries first.  */
    return DECL_UID (f2) - DECL_UID (f1);
}

/* Comparison function for qsort.  P1 and P2 are actually of type
   "tree *" and point to static destructors.  DECL_FINI_PRIORITY is
   used to determine the sort order.  */

static int
compare_dtor (const void *p1, const void *p2)
{
  tree f1;
  tree f2;
  int priority1;
  int priority2;

  f1 = *(const tree *)p1;
  f2 = *(const tree *)p2;
  priority1 = DECL_FINI_PRIORITY (f1);
  priority2 = DECL_FINI_PRIORITY (f2);

  if (priority1 < priority2)
    return -1;
  else if (priority1 > priority2)
    return 1;
  else
    /* Ensure a stable sort.  */
    return DECL_UID (f1) - DECL_UID (f2);
}

/* Generate functions to call static constructors and destructors
   for targets that do not support .ctors/.dtors sections.  These
   functions have magic names which are detected by collect2.  */

static void
build_cdtor_fns (void)
{
1133
  if (!static_ctors.is_empty ())
1134 1135
    {
      gcc_assert (!targetm.have_ctors_dtors || in_lto_p);
1136
      static_ctors.qsort (compare_ctor);
1137
      build_cdtor (/*ctor_p=*/true, static_ctors);
1138 1139
    }

1140
  if (!static_dtors.is_empty ())
1141 1142
    {
      gcc_assert (!targetm.have_ctors_dtors || in_lto_p);
1143
      static_dtors.qsort (compare_dtor);
1144
      build_cdtor (/*ctor_p=*/false, static_dtors);
1145 1146 1147 1148 1149
    }
}

/* Look for constructors and destructors and produce function calling them.
   This is needed for targets not supporting ctors or dtors, but we perform the
Joseph Myers committed
1150
   transformation also at linktime to merge possibly numerous
1151 1152 1153 1154 1155 1156 1157
   constructors/destructors into single function to improve code locality and
   reduce size.  */

static unsigned int
ipa_cdtor_merge (void)
{
  struct cgraph_node *node;
1158
  FOR_EACH_DEFINED_FUNCTION (node)
1159 1160
    if (DECL_STATIC_CONSTRUCTOR (node->decl)
	|| DECL_STATIC_DESTRUCTOR (node->decl))
1161 1162
       record_cdtor_fn (node);
  build_cdtor_fns ();
1163 1164
  static_ctors.release ();
  static_dtors.release ();
1165 1166 1167
  return 0;
}

1168 1169 1170
namespace {

const pass_data pass_data_ipa_cdtor_merge =
1171
{
1172 1173 1174 1175 1176 1177 1178 1179 1180
  IPA_PASS, /* type */
  "cdtor", /* name */
  OPTGROUP_NONE, /* optinfo_flags */
  TV_CGRAPHOPT, /* tv_id */
  0, /* properties_required */
  0, /* properties_provided */
  0, /* properties_destroyed */
  0, /* todo_flags_start */
  0, /* todo_flags_finish */
1181
};
1182

1183
class pass_ipa_cdtor_merge : public ipa_opt_pass_d
1184 1185
{
public:
1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196
  pass_ipa_cdtor_merge (gcc::context *ctxt)
    : ipa_opt_pass_d (pass_data_ipa_cdtor_merge, ctxt,
		      NULL, /* generate_summary */
		      NULL, /* write_summary */
		      NULL, /* read_summary */
		      NULL, /* write_optimization_summary */
		      NULL, /* read_optimization_summary */
		      NULL, /* stmt_fixup */
		      0, /* function_transform_todo_flags_start */
		      NULL, /* function_transform */
		      NULL) /* variable_transform */
1197 1198 1199
  {}

  /* opt_pass methods: */
1200
  virtual bool gate (function *);
1201
  virtual unsigned int execute (function *) { return ipa_cdtor_merge (); }
1202 1203 1204

}; // class pass_ipa_cdtor_merge

1205 1206 1207 1208 1209 1210 1211 1212 1213
bool
pass_ipa_cdtor_merge::gate (function *)
{
  /* Perform the pass when we have no ctors/dtors support
     or at LTO time to merge multiple constructors into single
     function.  */
  return !targetm.have_ctors_dtors || (optimize && in_lto_p);
}

1214 1215
} // anon namespace

1216 1217 1218 1219 1220
ipa_opt_pass_d *
make_pass_ipa_cdtor_merge (gcc::context *ctxt)
{
  return new pass_ipa_cdtor_merge (ctxt);
}
1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237

/* Invalid pointer representing BOTTOM for single user dataflow.  */
#define BOTTOM ((cgraph_node *)(size_t) 2)

/* Meet operation for single user dataflow.
   Here we want to associate variables with sigle function that may access it.

   FUNCTION is current single user of a variable, VAR is variable that uses it.
   Latttice is stored in SINGLE_USER_MAP.

   We represent: 
    - TOP by no entry in SIGNLE_USER_MAP
    - BOTTOM by BOTTOM in AUX pointer (to save lookups)
    - known single user by cgraph pointer in SINGLE_USER_MAP.  */

cgraph_node *
meet (cgraph_node *function, varpool_node *var,
Trevor Saunders committed
1238
       hash_map<varpool_node *, cgraph_node *> &single_user_map)
1239 1240 1241 1242 1243 1244
{
  struct cgraph_node *user, **f;

  if (var->aux == BOTTOM)
    return BOTTOM;

Trevor Saunders committed
1245
  f = single_user_map.get (var);
1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263
  if (!f)
    return function;
  user = *f;
  if (!function)
    return user;
  else if (function != user)
    return BOTTOM;
  else
    return function;
}

/* Propagation step of single-use dataflow.

   Check all uses of VNODE and see if they are used by single function FUNCTION.
   SINGLE_USER_MAP represents the dataflow lattice.  */

cgraph_node *
propagate_single_user (varpool_node *vnode, cgraph_node *function,
Trevor Saunders committed
1264
		       hash_map<varpool_node *, cgraph_node *> &single_user_map)
1265 1266 1267 1268 1269 1270 1271 1272
{
  int i;
  struct ipa_ref *ref;

  gcc_assert (!vnode->externally_visible);

  /* If node is an alias, first meet with its target.  */
  if (vnode->alias)
Martin Liska committed
1273
    function = meet (function, vnode->get_alias_target (), single_user_map);
1274 1275

  /* Check all users and see if they correspond to a single function.  */
Martin Liska committed
1276
  for (i = 0; vnode->iterate_referring (i, ref) && function != BOTTOM; i++)
1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288
    {
      struct cgraph_node *cnode = dyn_cast <cgraph_node *> (ref->referring);
      if (cnode)
	{
	  if (cnode->global.inlined_to)
	    cnode = cnode->global.inlined_to;
	  if (!function)
	    function = cnode;
	  else if (function != cnode)
	    function = BOTTOM;
	}
      else
1289 1290
	function = meet (function, dyn_cast <varpool_node *> (ref->referring),
			 single_user_map);
1291 1292 1293 1294 1295
    }
  return function;
}

/* Pass setting used_by_single_function flag.
1296 1297
   This flag is set on variable when there is only one function that may
   possibly referr to it.  */
1298 1299 1300 1301 1302 1303

static unsigned int
ipa_single_use (void)
{
  varpool_node *first = (varpool_node *) (void *) 1;
  varpool_node *var;
Trevor Saunders committed
1304
  hash_map<varpool_node *, cgraph_node *> single_user_map;
1305 1306

  FOR_EACH_DEFINED_VARIABLE (var)
Martin Liska committed
1307
    if (!var->all_refs_explicit_p ())
1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324
      var->aux = BOTTOM;
    else
      {
	/* Enqueue symbol for dataflow.  */
        var->aux = first;
	first = var;
      }

  /* The actual dataflow.  */

  while (first != (void *) 1)
    {
      cgraph_node *user, *orig_user, **f;

      var = first;
      first = (varpool_node *)first->aux;

Trevor Saunders committed
1325
      f = single_user_map.get (var);
1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339
      if (f)
	orig_user = *f;
      else
	orig_user = NULL;
      user = propagate_single_user (var, orig_user, single_user_map);

      gcc_checking_assert (var->aux != BOTTOM);

      /* If user differs, enqueue all references.  */
      if (user != orig_user)
	{
	  unsigned int i;
	  ipa_ref *ref;

Trevor Saunders committed
1340
	  single_user_map.put (var, user);
1341 1342

	  /* Enqueue all aliases for re-processing.  */
1343 1344
	  for (i = 0; var->iterate_direct_aliases (i, ref); i++)
	    if (!ref->referring->aux)
1345 1346 1347 1348 1349
	      {
		ref->referring->aux = first;
		first = dyn_cast <varpool_node *> (ref->referring);
	      }
	  /* Enqueue all users for re-processing.  */
Martin Liska committed
1350
	  for (i = 0; var->iterate_reference (i, ref); i++)
1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373
	    if (!ref->referred->aux
	        && ref->referred->definition
		&& is_a <varpool_node *> (ref->referred))
	      {
		ref->referred->aux = first;
		first = dyn_cast <varpool_node *> (ref->referred);
	      }

	  /* If user is BOTTOM, just punt on this var.  */
	  if (user == BOTTOM)
	    var->aux = BOTTOM;
	  else
	    var->aux = NULL;
	}
      else
	var->aux = NULL;
    }

  FOR_EACH_DEFINED_VARIABLE (var)
    {
      if (var->aux != BOTTOM)
	{
#ifdef ENABLE_CHECKING
1374 1375 1376 1377
	  /* Not having the single user known means that the VAR is
	     unreachable.  Either someone forgot to remove unreachable
	     variables or the reachability here is wrong.  */

Trevor Saunders committed
1378
          gcc_assert (single_user_map.get (var));
1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391
#endif
	  if (dump_file)
	    {
	      fprintf (dump_file, "Variable %s/%i is used by single function\n",
		       var->name (), var->order);
	    }
	  var->used_by_single_function = true;
	}
      var->aux = NULL;
    }
  return 0;
}

1392 1393 1394
namespace {

const pass_data pass_data_ipa_single_use =
1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406
{
  IPA_PASS, /* type */
  "single-use", /* name */
  OPTGROUP_NONE, /* optinfo_flags */
  TV_CGRAPHOPT, /* tv_id */
  0, /* properties_required */
  0, /* properties_provided */
  0, /* properties_destroyed */
  0, /* todo_flags_start */
  0, /* todo_flags_finish */
};

1407
class pass_ipa_single_use : public ipa_opt_pass_d
1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434
{
public:
  pass_ipa_single_use (gcc::context *ctxt)
    : ipa_opt_pass_d (pass_data_ipa_single_use, ctxt,
		      NULL, /* generate_summary */
		      NULL, /* write_summary */
		      NULL, /* read_summary */
		      NULL, /* write_optimization_summary */
		      NULL, /* read_optimization_summary */
		      NULL, /* stmt_fixup */
		      0, /* function_transform_todo_flags_start */
		      NULL, /* function_transform */
		      NULL) /* variable_transform */
  {}

  /* opt_pass methods: */
  virtual bool gate (function *);
  virtual unsigned int execute (function *) { return ipa_single_use (); }

}; // class pass_ipa_single_use

bool
pass_ipa_single_use::gate (function *)
{
  return optimize;
}

1435 1436
} // anon namespace

1437 1438 1439 1440 1441
ipa_opt_pass_d *
make_pass_ipa_single_use (gcc::context *ctxt)
{
  return new pass_ipa_single_use (ctxt);
}