tree-ssa-loop-prefetch.c 63.1 KB
Newer Older
Zdenek Dvorak committed
1
/* Array prefetching.
2
   Copyright (C) 2005-2017 Free Software Foundation, Inc.
H.J. Lu committed
3

Zdenek Dvorak committed
4
This file is part of GCC.
H.J. Lu committed
5

Zdenek Dvorak committed
6 7
GCC is free software; you can redistribute it and/or modify it
under the terms of the GNU General Public License as published by the
8
Free Software Foundation; either version 3, or (at your option) any
Zdenek Dvorak committed
9
later version.
H.J. Lu committed
10

Zdenek Dvorak committed
11 12 13 14
GCC is distributed in the hope that it will be useful, but WITHOUT
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
for more details.
H.J. Lu committed
15

Zdenek Dvorak committed
16
You should have received a copy of the GNU General Public License
17 18
along with GCC; see the file COPYING3.  If not see
<http://www.gnu.org/licenses/>.  */
Zdenek Dvorak committed
19 20 21 22

#include "config.h"
#include "system.h"
#include "coretypes.h"
23
#include "backend.h"
24 25
#include "target.h"
#include "rtl.h"
Zdenek Dvorak committed
26
#include "tree.h"
27
#include "gimple.h"
28 29 30 31 32
#include "predict.h"
#include "tree-pass.h"
#include "gimple-ssa.h"
#include "optabs-query.h"
#include "tree-pretty-print.h"
33
#include "fold-const.h"
34
#include "stor-layout.h"
35
#include "gimplify.h"
36
#include "gimple-iterator.h"
37
#include "gimplify-me.h"
38 39 40
#include "tree-ssa-loop-ivopts.h"
#include "tree-ssa-loop-manip.h"
#include "tree-ssa-loop-niter.h"
41
#include "tree-ssa-loop.h"
42
#include "ssa.h"
43
#include "tree-into-ssa.h"
Zdenek Dvorak committed
44 45 46 47
#include "cfgloop.h"
#include "tree-scalar-evolution.h"
#include "params.h"
#include "langhooks.h"
48
#include "tree-inline.h"
49
#include "tree-data-ref.h"
50
#include "diagnostic-core.h"
51
#include "dbgcnt.h"
52

Zdenek Dvorak committed
53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83
/* This pass inserts prefetch instructions to optimize cache usage during
   accesses to arrays in loops.  It processes loops sequentially and:

   1) Gathers all memory references in the single loop.
   2) For each of the references it decides when it is profitable to prefetch
      it.  To do it, we evaluate the reuse among the accesses, and determines
      two values: PREFETCH_BEFORE (meaning that it only makes sense to do
      prefetching in the first PREFETCH_BEFORE iterations of the loop) and
      PREFETCH_MOD (meaning that it only makes sense to prefetch in the
      iterations of the loop that are zero modulo PREFETCH_MOD).  For example
      (assuming cache line size is 64 bytes, char has size 1 byte and there
      is no hardware sequential prefetch):

      char *a;
      for (i = 0; i < max; i++)
	{
	  a[255] = ...;		(0)
	  a[i] = ...;		(1)
	  a[i + 64] = ...;	(2)
	  a[16*i] = ...;	(3)
	  a[187*i] = ...;	(4)
	  a[187*i + 50] = ...;	(5)
	}

       (0) obviously has PREFETCH_BEFORE 1
       (1) has PREFETCH_BEFORE 64, since (2) accesses the same memory
           location 64 iterations before it, and PREFETCH_MOD 64 (since
	   it hits the same cache line otherwise).
       (2) has PREFETCH_MOD 64
       (3) has PREFETCH_MOD 4
       (4) has PREFETCH_MOD 1.  We do not set PREFETCH_BEFORE here, since
84
           the cache line accessed by (5) is the same with probability only
Zdenek Dvorak committed
85 86 87
	   7/32.
       (5) has PREFETCH_MOD 1 as well.

88 89 90 91
      Additionally, we use data dependence analysis to determine for each
      reference the distance till the first reuse; this information is used
      to determine the temporality of the issued prefetch instruction.

Zdenek Dvorak committed
92 93 94 95 96 97 98 99 100 101 102 103
   3) We determine how much ahead we need to prefetch.  The number of
      iterations needed is time to fetch / time spent in one iteration of
      the loop.  The problem is that we do not know either of these values,
      so we just make a heuristic guess based on a magic (possibly)
      target-specific constant and size of the loop.

   4) Determine which of the references we prefetch.  We take into account
      that there is a maximum number of simultaneous prefetches (provided
      by machine description).  We prefetch as many prefetches as possible
      while still within this bound (starting with those with lowest
      prefetch_mod, since they are responsible for most of the cache
      misses).
H.J. Lu committed
104

Zdenek Dvorak committed
105 106 107 108
   5) We unroll and peel loops so that we are able to satisfy PREFETCH_MOD
      and PREFETCH_BEFORE requirements (within some bounds), and to avoid
      prefetching nonaccessed memory.
      TODO -- actually implement peeling.
H.J. Lu committed
109

Zdenek Dvorak committed
110 111 112 113
   6) We actually emit the prefetch instructions.  ??? Perhaps emit the
      prefetch instructions with guards in cases where 5) was not sufficient
      to satisfy the constraints?

114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136
   A cost model is implemented to determine whether or not prefetching is
   profitable for a given loop.  The cost model has three heuristics:

   1. Function trip_count_to_ahead_ratio_too_small_p implements a
      heuristic that determines whether or not the loop has too few
      iterations (compared to ahead).  Prefetching is not likely to be
      beneficial if the trip count to ahead ratio is below a certain
      minimum.

   2. Function mem_ref_count_reasonable_p implements a heuristic that
      determines whether the given loop has enough CPU ops that can be
      overlapped with cache missing memory ops.  If not, the loop
      won't benefit from prefetching.  In the implementation,
      prefetching is not considered beneficial if the ratio between
      the instruction count and the mem ref count is below a certain
      minimum.

   3. Function insn_to_prefetch_ratio_too_small_p implements a
      heuristic that disables prefetching in a loop if the prefetching
      cost is above a certain limit.  The relative prefetching cost is
      estimated by taking the ratio between the prefetch count and the
      total intruction count (this models the I-cache cost).

137
   The limits used in these heuristics are defined as parameters with
H.J. Lu committed
138
   reasonable default values. Machine-specific default values will be
139
   added later.
H.J. Lu committed
140

Zdenek Dvorak committed
141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164
   Some other TODO:
      -- write and use more general reuse analysis (that could be also used
	 in other cache aimed loop optimizations)
      -- make it behave sanely together with the prefetches given by user
	 (now we just ignore them; at the very least we should avoid
	 optimizing loops in that user put his own prefetches)
      -- we assume cache line size alignment of arrays; this could be
	 improved.  */

/* Magic constants follow.  These should be replaced by machine specific
   numbers.  */

/* True if write can be prefetched by a read prefetch.  */

#ifndef WRITE_CAN_USE_READ_PREFETCH
#define WRITE_CAN_USE_READ_PREFETCH 1
#endif

/* True if read can be prefetched by a write prefetch. */

#ifndef READ_CAN_USE_WRITE_PREFETCH
#define READ_CAN_USE_WRITE_PREFETCH 0
#endif

165 166 167
/* The size of the block loaded by a single prefetch.  Usually, this is
   the same as cache line size (at the moment, we only consider one level
   of cache hierarchy).  */
Zdenek Dvorak committed
168 169

#ifndef PREFETCH_BLOCK
170
#define PREFETCH_BLOCK L1_CACHE_LINE_SIZE
Zdenek Dvorak committed
171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187
#endif

/* Do we have a forward hardware sequential prefetching?  */

#ifndef HAVE_FORWARD_PREFETCH
#define HAVE_FORWARD_PREFETCH 0
#endif

/* Do we have a backward hardware sequential prefetching?  */

#ifndef HAVE_BACKWARD_PREFETCH
#define HAVE_BACKWARD_PREFETCH 0
#endif

/* In some cases we are only able to determine that there is a certain
   probability that the two accesses hit the same cache line.  In this
   case, we issue the prefetches for both of them if this probability
188
   is less then (1000 - ACCEPTABLE_MISS_RATE) per thousand.  */
Zdenek Dvorak committed
189 190 191 192 193

#ifndef ACCEPTABLE_MISS_RATE
#define ACCEPTABLE_MISS_RATE 50
#endif

194 195
#define L1_CACHE_SIZE_BYTES ((unsigned) (L1_CACHE_SIZE * 1024))
#define L2_CACHE_SIZE_BYTES ((unsigned) (L2_CACHE_SIZE * 1024))
196 197 198 199 200 201 202 203

/* We consider a memory access nontemporal if it is not reused sooner than
   after L2_CACHE_SIZE_BYTES of memory are accessed.  However, we ignore
   accesses closer than L1_CACHE_SIZE_BYTES / NONTEMPORAL_FRACTION,
   so that we use nontemporal prefetches e.g. if single memory location
   is accessed several times in a single iteration of the loop.  */
#define NONTEMPORAL_FRACTION 16

204 205 206 207 208 209 210
/* In case we have to emit a memory fence instruction after the loop that
   uses nontemporal stores, this defines the builtin to use.  */

#ifndef FENCE_FOLLOWING_MOVNT
#define FENCE_FOLLOWING_MOVNT NULL_TREE
#endif

211 212 213 214 215 216 217 218 219 220 221 222
/* It is not profitable to prefetch when the trip count is not at
   least TRIP_COUNT_TO_AHEAD_RATIO times the prefetch ahead distance.
   For example, in a loop with a prefetch ahead distance of 10,
   supposing that TRIP_COUNT_TO_AHEAD_RATIO is equal to 4, it is
   profitable to prefetch when the trip count is greater or equal to
   40.  In that case, 30 out of the 40 iterations will benefit from
   prefetching.  */

#ifndef TRIP_COUNT_TO_AHEAD_RATIO
#define TRIP_COUNT_TO_AHEAD_RATIO 4
#endif

Zdenek Dvorak committed
223 224 225 226 227
/* The group of references between that reuse may occur.  */

struct mem_ref_group
{
  tree base;			/* Base of the reference.  */
228
  tree step;			/* Step of the reference.  */
Zdenek Dvorak committed
229 230
  struct mem_ref *refs;		/* References in the group.  */
  struct mem_ref_group *next;	/* Next group of references.  */
231
  unsigned int uid;		/* Group UID, used only for debugging.  */
Zdenek Dvorak committed
232 233 234 235
};

/* Assigned to PREFETCH_BEFORE when all iterations are to be prefetched.  */

236
#define PREFETCH_ALL		HOST_WIDE_INT_M1U
Zdenek Dvorak committed
237

238 239
/* Do not generate a prefetch if the unroll factor is significantly less
   than what is required by the prefetch.  This is to avoid redundant
240 241 242
   prefetches.  For example, when prefetch_mod is 16 and unroll_factor is
   2, prefetching requires unrolling the loop 16 times, but
   the loop is actually unrolled twice.  In this case (ratio = 8),
243 244 245
   prefetching is not likely to be beneficial.  */

#ifndef PREFETCH_MOD_TO_UNROLL_FACTOR_RATIO
246
#define PREFETCH_MOD_TO_UNROLL_FACTOR_RATIO 4
247 248
#endif

249 250 251 252 253 254 255 256
/* Some of the prefetch computations have quadratic complexity.  We want to
   avoid huge compile times and, therefore, want to limit the amount of
   memory references per loop where we consider prefetching.  */

#ifndef PREFETCH_MAX_MEM_REFS_PER_LOOP
#define PREFETCH_MAX_MEM_REFS_PER_LOOP 200
#endif

Zdenek Dvorak committed
257 258 259 260
/* The memory reference.  */

struct mem_ref
{
261
  gimple *stmt;			/* Statement in that the reference appears.  */
Zdenek Dvorak committed
262 263 264 265 266 267 268 269 270
  tree mem;			/* The reference.  */
  HOST_WIDE_INT delta;		/* Constant offset of the reference.  */
  struct mem_ref_group *group;	/* The group of references it belongs to.  */
  unsigned HOST_WIDE_INT prefetch_mod;
				/* Prefetch only each PREFETCH_MOD-th
				   iteration.  */
  unsigned HOST_WIDE_INT prefetch_before;
				/* Prefetch only first PREFETCH_BEFORE
				   iterations.  */
271 272
  unsigned reuse_distance;	/* The amount of data accessed before the first
				   reuse of this value.  */
Zdenek Dvorak committed
273
  struct mem_ref *next;		/* The next reference in the group.  */
274
  unsigned int uid;		/* Ref UID, used only for debugging.  */
275 276 277 278 279 280
  unsigned write_p : 1;		/* Is it a write?  */
  unsigned independent_p : 1;	/* True if the reference is independent on
				   all other references inside the loop.  */
  unsigned issue_prefetch_p : 1;	/* Should we really issue the prefetch?  */
  unsigned storent_p : 1;	/* True if we changed the store to a
				   nontemporal one.  */
Zdenek Dvorak committed
281 282
};

283
/* Dumps information about memory reference */
Zdenek Dvorak committed
284
static void
285 286
dump_mem_details (FILE *file, tree base, tree step,
	    HOST_WIDE_INT delta, bool write_p) 
Zdenek Dvorak committed
287
{
288 289
  fprintf (file, "(base ");
  print_generic_expr (file, base, TDF_SLIM);
Zdenek Dvorak committed
290
  fprintf (file, ", step ");
291 292
  if (cst_and_fits_in_hwi (step))
    fprintf (file, HOST_WIDE_INT_PRINT_DEC, int_cst_value (step));
293
  else
294
    print_generic_expr (file, step, TDF_SLIM);
Zdenek Dvorak committed
295
  fprintf (file, ")\n");
296 297
  fprintf (file, "  delta " HOST_WIDE_INT_PRINT_DEC "\n", delta);
  fprintf (file, "  %s\n\n", write_p ? "write" : "read");
298
}
Zdenek Dvorak committed
299

300
/* Dumps information about reference REF to FILE.  */
Zdenek Dvorak committed
301

302 303 304
static void
dump_mem_ref (FILE *file, struct mem_ref *ref)
{
305 306 307
  fprintf (file, "reference %u:%u (", ref->group->uid, ref->uid);
  print_generic_expr (file, ref->mem, TDF_SLIM);
  fprintf (file, ")\n");
Zdenek Dvorak committed
308 309 310 311 312 313
}

/* Finds a group with BASE and STEP in GROUPS, or creates one if it does not
   exist.  */

static struct mem_ref_group *
314
find_or_create_group (struct mem_ref_group **groups, tree base, tree step)
Zdenek Dvorak committed
315
{
316 317 318
  /* Global count for setting struct mem_ref_group->uid.  */
  static unsigned int last_mem_ref_group_uid = 0;

Zdenek Dvorak committed
319 320 321 322
  struct mem_ref_group *group;

  for (; *groups; groups = &(*groups)->next)
    {
323
      if (operand_equal_p ((*groups)->step, step, 0)
Zdenek Dvorak committed
324 325 326
	  && operand_equal_p ((*groups)->base, base, 0))
	return *groups;

327 328
      /* If step is an integer constant, keep the list of groups sorted
         by decreasing step.  */
329 330
      if (cst_and_fits_in_hwi ((*groups)->step) && cst_and_fits_in_hwi (step)
	  && int_cst_value ((*groups)->step) < int_cst_value (step))
Zdenek Dvorak committed
331 332 333
	break;
    }

334
  group = XNEW (struct mem_ref_group);
Zdenek Dvorak committed
335 336 337
  group->base = base;
  group->step = step;
  group->refs = NULL;
338
  group->uid = ++last_mem_ref_group_uid;
Zdenek Dvorak committed
339 340 341 342 343 344 345 346 347 348
  group->next = *groups;
  *groups = group;

  return group;
}

/* Records a memory reference MEM in GROUP with offset DELTA and write status
   WRITE_P.  The reference occurs in statement STMT.  */

static void
349
record_ref (struct mem_ref_group *group, gimple *stmt, tree mem,
Zdenek Dvorak committed
350 351
	    HOST_WIDE_INT delta, bool write_p)
{
352
  unsigned int last_mem_ref_uid = 0;
Zdenek Dvorak committed
353 354 355 356 357
  struct mem_ref **aref;

  /* Do not record the same address twice.  */
  for (aref = &group->refs; *aref; aref = &(*aref)->next)
    {
358 359
      last_mem_ref_uid = (*aref)->uid;

Zdenek Dvorak committed
360 361 362 363 364 365 366 367 368 369 370 371 372 373 374
      /* It does not have to be possible for write reference to reuse the read
	 prefetch, or vice versa.  */
      if (!WRITE_CAN_USE_READ_PREFETCH
	  && write_p
	  && !(*aref)->write_p)
	continue;
      if (!READ_CAN_USE_WRITE_PREFETCH
	  && !write_p
	  && (*aref)->write_p)
	continue;

      if ((*aref)->delta == delta)
	return;
    }

375
  (*aref) = XNEW (struct mem_ref);
Zdenek Dvorak committed
376 377 378 379 380 381
  (*aref)->stmt = stmt;
  (*aref)->mem = mem;
  (*aref)->delta = delta;
  (*aref)->write_p = write_p;
  (*aref)->prefetch_before = PREFETCH_ALL;
  (*aref)->prefetch_mod = 1;
382
  (*aref)->reuse_distance = 0;
Zdenek Dvorak committed
383 384 385
  (*aref)->issue_prefetch_p = false;
  (*aref)->group = group;
  (*aref)->next = NULL;
386 387
  (*aref)->independent_p = false;
  (*aref)->storent_p = false;
388
  (*aref)->uid = last_mem_ref_uid + 1;
Zdenek Dvorak committed
389 390

  if (dump_file && (dump_flags & TDF_DETAILS))
391 392 393 394 395 396 397
    {
      dump_mem_ref (dump_file, *aref);

      fprintf (dump_file, "  group %u ", group->uid);
      dump_mem_details (dump_file, group->base, group->step, delta,
			write_p);
    }
Zdenek Dvorak committed
398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424
}

/* Release memory references in GROUPS.  */

static void
release_mem_refs (struct mem_ref_group *groups)
{
  struct mem_ref_group *next_g;
  struct mem_ref *ref, *next_r;

  for (; groups; groups = next_g)
    {
      next_g = groups->next;
      for (ref = groups->refs; ref; ref = next_r)
	{
	  next_r = ref->next;
	  free (ref);
	}
      free (groups);
    }
}

/* A structure used to pass arguments to idx_analyze_ref.  */

struct ar_data
{
  struct loop *loop;			/* Loop of the reference.  */
425
  gimple *stmt;				/* Statement of the reference.  */
426
  tree *step;				/* Step of the memory reference.  */
Zdenek Dvorak committed
427 428 429 430 431 432 433 434 435
  HOST_WIDE_INT *delta;			/* Offset of the memory reference.  */
};

/* Analyzes a single INDEX of a memory reference to obtain information
   described at analyze_ref.  Callback for for_each_index.  */

static bool
idx_analyze_ref (tree base, tree *index, void *data)
{
436
  struct ar_data *ar_data = (struct ar_data *) data;
Zdenek Dvorak committed
437
  tree ibase, step, stepsize;
438
  HOST_WIDE_INT idelta = 0, imult = 1;
Zdenek Dvorak committed
439 440
  affine_iv iv;

441
  if (!simple_iv (ar_data->loop, loop_containing_stmt (ar_data->stmt),
442
		  *index, &iv, true))
Zdenek Dvorak committed
443 444 445 446
    return false;
  ibase = iv.base;
  step = iv.step;

Andrew Pinski committed
447
  if (TREE_CODE (ibase) == POINTER_PLUS_EXPR
Zdenek Dvorak committed
448 449 450 451 452 453 454 455
      && cst_and_fits_in_hwi (TREE_OPERAND (ibase, 1)))
    {
      idelta = int_cst_value (TREE_OPERAND (ibase, 1));
      ibase = TREE_OPERAND (ibase, 0);
    }
  if (cst_and_fits_in_hwi (ibase))
    {
      idelta += int_cst_value (ibase);
456
      ibase = build_int_cst (TREE_TYPE (ibase), 0);
Zdenek Dvorak committed
457 458 459 460 461 462 463 464
    }

  if (TREE_CODE (base) == ARRAY_REF)
    {
      stepsize = array_ref_element_size (base);
      if (!cst_and_fits_in_hwi (stepsize))
	return false;
      imult = int_cst_value (stepsize);
465 466 467
      step = fold_build2 (MULT_EXPR, sizetype,
			  fold_convert (sizetype, step),
			  fold_convert (sizetype, stepsize));
Zdenek Dvorak committed
468 469 470
      idelta *= imult;
    }

471 472 473 474 475 476
  if (*ar_data->step == NULL_TREE)
    *ar_data->step = step;
  else
    *ar_data->step = fold_build2 (PLUS_EXPR, sizetype,
				  fold_convert (sizetype, *ar_data->step),
				  fold_convert (sizetype, step));
Zdenek Dvorak committed
477 478 479 480 481 482
  *ar_data->delta += idelta;
  *index = ibase;

  return true;
}

483
/* Tries to express REF_P in shape &BASE + STEP * iter + DELTA, where DELTA and
Zdenek Dvorak committed
484
   STEP are integer constants and iter is number of iterations of LOOP.  The
485 486
   reference occurs in statement STMT.  Strips nonaddressable component
   references from REF_P.  */
Zdenek Dvorak committed
487 488

static bool
489
analyze_ref (struct loop *loop, tree *ref_p, tree *base,
490
	     tree *step, HOST_WIDE_INT *delta,
491
	     gimple *stmt)
Zdenek Dvorak committed
492 493 494 495
{
  struct ar_data ar_data;
  tree off;
  HOST_WIDE_INT bit_offset;
496
  tree ref = *ref_p;
Zdenek Dvorak committed
497

498
  *step = NULL_TREE;
Zdenek Dvorak committed
499 500
  *delta = 0;

501 502 503 504 505 506 507 508 509 510 511 512
  /* First strip off the component references.  Ignore bitfields.
     Also strip off the real and imagine parts of a complex, so that
     they can have the same base.  */
  if (TREE_CODE (ref) == REALPART_EXPR
      || TREE_CODE (ref) == IMAGPART_EXPR
      || (TREE_CODE (ref) == COMPONENT_REF
          && DECL_NONADDRESSABLE_P (TREE_OPERAND (ref, 1))))
    {
      if (TREE_CODE (ref) == IMAGPART_EXPR)
        *delta += int_size_in_bytes (TREE_TYPE (ref));
      ref = TREE_OPERAND (ref, 0);
    }
Zdenek Dvorak committed
513

514 515
  *ref_p = ref;

Zdenek Dvorak committed
516 517 518 519 520
  for (; TREE_CODE (ref) == COMPONENT_REF; ref = TREE_OPERAND (ref, 0))
    {
      off = DECL_FIELD_BIT_OFFSET (TREE_OPERAND (ref, 1));
      bit_offset = TREE_INT_CST_LOW (off);
      gcc_assert (bit_offset % BITS_PER_UNIT == 0);
H.J. Lu committed
521

Zdenek Dvorak committed
522 523 524 525 526 527 528 529 530 531 532 533
      *delta += bit_offset / BITS_PER_UNIT;
    }

  *base = unshare_expr (ref);
  ar_data.loop = loop;
  ar_data.stmt = stmt;
  ar_data.step = step;
  ar_data.delta = delta;
  return for_each_index (base, idx_analyze_ref, &ar_data);
}

/* Record a memory reference REF to the list REFS.  The reference occurs in
534 535
   LOOP in statement STMT and it is write if WRITE_P.  Returns true if the
   reference was recorded, false otherwise.  */
Zdenek Dvorak committed
536

537
static bool
Zdenek Dvorak committed
538
gather_memory_references_ref (struct loop *loop, struct mem_ref_group **refs,
539
			      tree ref, bool write_p, gimple *stmt)
Zdenek Dvorak committed
540
{
541 542
  tree base, step;
  HOST_WIDE_INT delta;
Zdenek Dvorak committed
543 544
  struct mem_ref_group *agrp;

545 546 547
  if (get_base_address (ref) == NULL)
    return false;

548
  if (!analyze_ref (loop, &ref, &base, &step, &delta, stmt))
549
    return false;
550 551 552
  /* If analyze_ref fails the default is a NULL_TREE.  We can stop here.  */
  if (step == NULL_TREE)
    return false;
Zdenek Dvorak committed
553

554
  /* Stop if the address of BASE could not be taken.  */
555 556 557
  if (may_be_nonaddressable_p (base))
    return false;

558 559 560 561 562 563 564 565 566
  /* Limit non-constant step prefetching only to the innermost loops and 
     only when the step is loop invariant in the entire loop nest. */
  if (!cst_and_fits_in_hwi (step))
    {
      if (loop->inner != NULL)
        {
          if (dump_file && (dump_flags & TDF_DETAILS))
            {
              fprintf (dump_file, "Memory expression %p\n",(void *) ref ); 
567 568
	      print_generic_expr (dump_file, ref, TDF_SLIM);
	      fprintf (dump_file,":");
569
              dump_mem_details (dump_file, base, step, delta, write_p);
570 571 572 573 574 575 576 577 578 579 580 581 582 583
              fprintf (dump_file, 
                       "Ignoring %p, non-constant step prefetching is "
                       "limited to inner most loops \n", 
                       (void *) ref);
            }
            return false;    
         }
      else
        {
          if (!expr_invariant_in_loop_p (loop_outermost (loop), step))
          {
            if (dump_file && (dump_flags & TDF_DETAILS))
              {
                fprintf (dump_file, "Memory expression %p\n",(void *) ref );
584
		print_generic_expr (dump_file, ref, TDF_SLIM);
585
                fprintf (dump_file,":");
586
                dump_mem_details (dump_file, base, step, delta, write_p);
587 588 589 590 591 592 593 594 595
                fprintf (dump_file, 
                         "Not prefetching, ignoring %p due to "
                         "loop variant step\n",
                         (void *) ref);
              }
              return false;                 
            }
        }
    }
596

Zdenek Dvorak committed
597 598 599 600
  /* Now we know that REF = &BASE + STEP * iter + DELTA, where DELTA and STEP
     are integer constants.  */
  agrp = find_or_create_group (refs, base, step);
  record_ref (agrp, stmt, ref, delta, write_p);
601 602

  return true;
Zdenek Dvorak committed
603 604
}

605 606
/* Record the suitable memory references in LOOP.  NO_OTHER_REFS is set to
   true if there are no other memory references inside the loop.  */
Zdenek Dvorak committed
607 608

static struct mem_ref_group *
609
gather_memory_references (struct loop *loop, bool *no_other_refs, unsigned *ref_count)
Zdenek Dvorak committed
610 611 612 613
{
  basic_block *body = get_loop_body_in_dom_order (loop);
  basic_block bb;
  unsigned i;
614
  gimple_stmt_iterator bsi;
615
  gimple *stmt;
616
  tree lhs, rhs;
Zdenek Dvorak committed
617 618
  struct mem_ref_group *refs = NULL;

619
  *no_other_refs = true;
620
  *ref_count = 0;
621

Zdenek Dvorak committed
622 623 624 625 626 627 628 629
  /* Scan the loop body in order, so that the former references precede the
     later ones.  */
  for (i = 0; i < loop->num_nodes; i++)
    {
      bb = body[i];
      if (bb->loop_father != loop)
	continue;

630
      for (bsi = gsi_start_bb (bb); !gsi_end_p (bsi); gsi_next (&bsi))
Zdenek Dvorak committed
631
	{
632
	  stmt = gsi_stmt (bsi);
633

634
	  if (gimple_code (stmt) != GIMPLE_ASSIGN)
635
	    {
636
	      if (gimple_vuse (stmt)
637 638
		  || (is_gimple_call (stmt)
		      && !(gimple_call_flags (stmt) & ECF_CONST)))
639 640 641
		*no_other_refs = false;
	      continue;
	    }
Zdenek Dvorak committed
642

643 644 645
	  if (! gimple_vuse (stmt))
	    continue;

646 647
	  lhs = gimple_assign_lhs (stmt);
	  rhs = gimple_assign_rhs1 (stmt);
Zdenek Dvorak committed
648 649

	  if (REFERENCE_CLASS_P (rhs))
650
	    {
651 652
	    *no_other_refs &= gather_memory_references_ref (loop, &refs,
							    rhs, false, stmt);
653 654
	    *ref_count += 1;
	    }
Zdenek Dvorak committed
655
	  if (REFERENCE_CLASS_P (lhs))
656
	    {
657 658
	    *no_other_refs &= gather_memory_references_ref (loop, &refs,
							    lhs, true, stmt);
659 660
	    *ref_count += 1;
	    }
Zdenek Dvorak committed
661 662 663 664 665 666 667 668 669 670 671 672
	}
    }
  free (body);

  return refs;
}

/* Prune the prefetch candidate REF using the self-reuse.  */

static void
prune_ref_by_self_reuse (struct mem_ref *ref)
{
673 674 675 676 677 678 679 680 681 682
  HOST_WIDE_INT step;
  bool backward;

  /* If the step size is non constant, we cannot calculate prefetch_mod.  */
  if (!cst_and_fits_in_hwi (ref->group->step))
    return;

  step = int_cst_value (ref->group->step);

  backward = step < 0;
Zdenek Dvorak committed
683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714

  if (step == 0)
    {
      /* Prefetch references to invariant address just once.  */
      ref->prefetch_before = 1;
      return;
    }

  if (backward)
    step = -step;

  if (step > PREFETCH_BLOCK)
    return;

  if ((backward && HAVE_BACKWARD_PREFETCH)
      || (!backward && HAVE_FORWARD_PREFETCH))
    {
      ref->prefetch_before = 1;
      return;
    }

  ref->prefetch_mod = PREFETCH_BLOCK / step;
}

/* Divides X by BY, rounding down.  */

static HOST_WIDE_INT
ddown (HOST_WIDE_INT x, unsigned HOST_WIDE_INT by)
{
  gcc_assert (by > 0);

  if (x >= 0)
715
    return x / (HOST_WIDE_INT) by;
Zdenek Dvorak committed
716
  else
717
    return (x + (HOST_WIDE_INT) by - 1) / (HOST_WIDE_INT) by;
Zdenek Dvorak committed
718 719
}

H.J. Lu committed
720 721 722
/* Given a CACHE_LINE_SIZE and two inductive memory references
   with a common STEP greater than CACHE_LINE_SIZE and an address
   difference DELTA, compute the probability that they will fall
723 724 725
   in different cache lines.  Return true if the computed miss rate
   is not greater than the ACCEPTABLE_MISS_RATE.  DISTINCT_ITERS is the
   number of distinct iterations after which the pattern repeats itself.
726 727
   ALIGN_UNIT is the unit of alignment in bytes.  */

728 729
static bool
is_miss_rate_acceptable (unsigned HOST_WIDE_INT cache_line_size,
730 731 732 733 734
		   HOST_WIDE_INT step, HOST_WIDE_INT delta,
		   unsigned HOST_WIDE_INT distinct_iters,
		   int align_unit)
{
  unsigned align, iter;
735
  int total_positions, miss_positions, max_allowed_miss_positions;
736 737
  int address1, address2, cache_line1, cache_line2;

738 739
  /* It always misses if delta is greater than or equal to the cache
     line size.  */
740 741
  if (delta >= (HOST_WIDE_INT) cache_line_size)
    return false;
742

743
  miss_positions = 0;
744 745
  total_positions = (cache_line_size / align_unit) * distinct_iters;
  max_allowed_miss_positions = (ACCEPTABLE_MISS_RATE * total_positions) / 1000;
H.J. Lu committed
746

747 748 749 750 751 752 753 754 755 756 757 758
  /* Iterate through all possible alignments of the first
     memory reference within its cache line.  */
  for (align = 0; align < cache_line_size; align += align_unit)

    /* Iterate through all distinct iterations.  */
    for (iter = 0; iter < distinct_iters; iter++)
      {
	address1 = align + step * iter;
	address2 = address1 + delta;
	cache_line1 = address1 / cache_line_size;
	cache_line2 = address2 / cache_line_size;
	if (cache_line1 != cache_line2)
759 760 761 762 763
	  {
	    miss_positions += 1;
            if (miss_positions > max_allowed_miss_positions)
	      return false;
          }
764
      }
765
  return true;
766 767
}

Zdenek Dvorak committed
768 769 770 771 772 773 774
/* Prune the prefetch candidate REF using the reuse with BY.
   If BY_IS_BEFORE is true, BY is before REF in the loop.  */

static void
prune_ref_by_group_reuse (struct mem_ref *ref, struct mem_ref *by,
			  bool by_is_before)
{
775 776
  HOST_WIDE_INT step;
  bool backward;
Zdenek Dvorak committed
777 778 779 780
  HOST_WIDE_INT delta_r = ref->delta, delta_b = by->delta;
  HOST_WIDE_INT delta = delta_b - delta_r;
  HOST_WIDE_INT hit_from;
  unsigned HOST_WIDE_INT prefetch_before, prefetch_block;
781 782 783 784
  HOST_WIDE_INT reduced_step;
  unsigned HOST_WIDE_INT reduced_prefetch_block;
  tree ref_type;
  int align_unit;
Zdenek Dvorak committed
785

786 787 788 789 790 791 792 793 794 795
  /* If the step is non constant we cannot calculate prefetch_before.  */
  if (!cst_and_fits_in_hwi (ref->group->step)) {
    return;
  }

  step = int_cst_value (ref->group->step);

  backward = step < 0;


Zdenek Dvorak committed
796 797 798 799 800 801
  if (delta == 0)
    {
      /* If the references has the same address, only prefetch the
	 former.  */
      if (by_is_before)
	ref->prefetch_before = 0;
H.J. Lu committed
802

Zdenek Dvorak committed
803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849
      return;
    }

  if (!step)
    {
      /* If the reference addresses are invariant and fall into the
	 same cache line, prefetch just the first one.  */
      if (!by_is_before)
	return;

      if (ddown (ref->delta, PREFETCH_BLOCK)
	  != ddown (by->delta, PREFETCH_BLOCK))
	return;

      ref->prefetch_before = 0;
      return;
    }

  /* Only prune the reference that is behind in the array.  */
  if (backward)
    {
      if (delta > 0)
	return;

      /* Transform the data so that we may assume that the accesses
	 are forward.  */
      delta = - delta;
      step = -step;
      delta_r = PREFETCH_BLOCK - 1 - delta_r;
      delta_b = PREFETCH_BLOCK - 1 - delta_b;
    }
  else
    {
      if (delta < 0)
	return;
    }

  /* Check whether the two references are likely to hit the same cache
     line, and how distant the iterations in that it occurs are from
     each other.  */

  if (step <= PREFETCH_BLOCK)
    {
      /* The accesses are sure to meet.  Let us check when.  */
      hit_from = ddown (delta_b, PREFETCH_BLOCK) * PREFETCH_BLOCK;
      prefetch_before = (hit_from - delta_r + step - 1) / step;

850
      /* Do not reduce prefetch_before if we meet beyond cache size.  */
851
      if (prefetch_before > absu_hwi (L2_CACHE_SIZE_BYTES / step))
852
        prefetch_before = PREFETCH_ALL;
Zdenek Dvorak committed
853 854 855 856 857 858
      if (prefetch_before < ref->prefetch_before)
	ref->prefetch_before = prefetch_before;

      return;
    }

H.J. Lu committed
859
  /* A more complicated case with step > prefetch_block.  First reduce
860
     the ratio between the step and the cache line size to its simplest
H.J. Lu committed
861 862 863
     terms.  The resulting denominator will then represent the number of
     distinct iterations after which each address will go back to its
     initial location within the cache line.  This computation assumes
864
     that PREFETCH_BLOCK is a power of two.  */
Zdenek Dvorak committed
865
  prefetch_block = PREFETCH_BLOCK;
866 867 868 869
  reduced_prefetch_block = prefetch_block;
  reduced_step = step;
  while ((reduced_step & 1) == 0
	 && reduced_prefetch_block > 1)
Zdenek Dvorak committed
870
    {
871 872
      reduced_step >>= 1;
      reduced_prefetch_block >>= 1;
Zdenek Dvorak committed
873 874 875 876
    }

  prefetch_before = delta / step;
  delta %= step;
877 878
  ref_type = TREE_TYPE (ref->mem);
  align_unit = TYPE_ALIGN (ref_type) / 8;
879 880
  if (is_miss_rate_acceptable (prefetch_block, step, delta,
			       reduced_prefetch_block, align_unit))
Zdenek Dvorak committed
881
    {
882 883 884
      /* Do not reduce prefetch_before if we meet beyond cache size.  */
      if (prefetch_before > L2_CACHE_SIZE_BYTES / PREFETCH_BLOCK)
        prefetch_before = PREFETCH_ALL;
Zdenek Dvorak committed
885 886 887 888 889 890 891 892 893
      if (prefetch_before < ref->prefetch_before)
	ref->prefetch_before = prefetch_before;

      return;
    }

  /* Try also the following iteration.  */
  prefetch_before++;
  delta = step - delta;
894 895
  if (is_miss_rate_acceptable (prefetch_block, step, delta,
			       reduced_prefetch_block, align_unit))
Zdenek Dvorak committed
896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951
    {
      if (prefetch_before < ref->prefetch_before)
	ref->prefetch_before = prefetch_before;

      return;
    }

  /* The ref probably does not reuse by.  */
  return;
}

/* Prune the prefetch candidate REF using the reuses with other references
   in REFS.  */

static void
prune_ref_by_reuse (struct mem_ref *ref, struct mem_ref *refs)
{
  struct mem_ref *prune_by;
  bool before = true;

  prune_ref_by_self_reuse (ref);

  for (prune_by = refs; prune_by; prune_by = prune_by->next)
    {
      if (prune_by == ref)
	{
	  before = false;
	  continue;
	}

      if (!WRITE_CAN_USE_READ_PREFETCH
	  && ref->write_p
	  && !prune_by->write_p)
	continue;
      if (!READ_CAN_USE_WRITE_PREFETCH
	  && !ref->write_p
	  && prune_by->write_p)
	continue;

      prune_ref_by_group_reuse (ref, prune_by, before);
    }
}

/* Prune the prefetch candidates in GROUP using the reuse analysis.  */

static void
prune_group_by_reuse (struct mem_ref_group *group)
{
  struct mem_ref *ref_pruned;

  for (ref_pruned = group->refs; ref_pruned; ref_pruned = ref_pruned->next)
    {
      prune_ref_by_reuse (ref_pruned, group->refs);

      if (dump_file && (dump_flags & TDF_DETAILS))
	{
952
	  dump_mem_ref (dump_file, ref_pruned);
Zdenek Dvorak committed
953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997

	  if (ref_pruned->prefetch_before == PREFETCH_ALL
	      && ref_pruned->prefetch_mod == 1)
	    fprintf (dump_file, " no restrictions");
	  else if (ref_pruned->prefetch_before == 0)
	    fprintf (dump_file, " do not prefetch");
	  else if (ref_pruned->prefetch_before <= ref_pruned->prefetch_mod)
	    fprintf (dump_file, " prefetch once");
	  else
	    {
	      if (ref_pruned->prefetch_before != PREFETCH_ALL)
		{
		  fprintf (dump_file, " prefetch before ");
		  fprintf (dump_file, HOST_WIDE_INT_PRINT_DEC,
			   ref_pruned->prefetch_before);
		}
	      if (ref_pruned->prefetch_mod != 1)
		{
		  fprintf (dump_file, " prefetch mod ");
		  fprintf (dump_file, HOST_WIDE_INT_PRINT_DEC,
			   ref_pruned->prefetch_mod);
		}
	    }
	  fprintf (dump_file, "\n");
	}
    }
}

/* Prune the list of prefetch candidates GROUPS using the reuse analysis.  */

static void
prune_by_reuse (struct mem_ref_group *groups)
{
  for (; groups; groups = groups->next)
    prune_group_by_reuse (groups);
}

/* Returns true if we should issue prefetch for REF.  */

static bool
should_issue_prefetch_p (struct mem_ref *ref)
{
  /* For now do not issue prefetches for only first few of the
     iterations.  */
  if (ref->prefetch_before != PREFETCH_ALL)
998 999
    {
      if (dump_file && (dump_flags & TDF_DETAILS))
1000 1001
        fprintf (dump_file, "Ignoring reference %u:%u due to prefetch_before\n",
		 ref->group->uid, ref->uid);
1002 1003
      return false;
    }
Zdenek Dvorak committed
1004

1005 1006
  /* Do not prefetch nontemporal stores.  */
  if (ref->storent_p)
1007 1008
    {
      if (dump_file && (dump_flags & TDF_DETAILS))
1009
        fprintf (dump_file, "Ignoring nontemporal store reference %u:%u\n", ref->group->uid, ref->uid);
1010 1011
      return false;
    }
1012

Zdenek Dvorak committed
1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025
  return true;
}

/* Decide which of the prefetch candidates in GROUPS to prefetch.
   AHEAD is the number of iterations to prefetch ahead (which corresponds
   to the number of simultaneous instances of one prefetch running at a
   time).  UNROLL_FACTOR is the factor by that the loop is going to be
   unrolled.  Returns true if there is anything to prefetch.  */

static bool
schedule_prefetches (struct mem_ref_group *groups, unsigned unroll_factor,
		     unsigned ahead)
{
1026 1027
  unsigned remaining_prefetch_slots, n_prefetches, prefetch_slots;
  unsigned slots_per_prefetch;
Zdenek Dvorak committed
1028 1029 1030
  struct mem_ref *ref;
  bool any = false;

1031 1032
  /* At most SIMULTANEOUS_PREFETCHES should be running at the same time.  */
  remaining_prefetch_slots = SIMULTANEOUS_PREFETCHES;
Zdenek Dvorak committed
1033

1034 1035 1036 1037
  /* The prefetch will run for AHEAD iterations of the original loop, i.e.,
     AHEAD / UNROLL_FACTOR iterations of the unrolled loop.  In each iteration,
     it will need a prefetch slot.  */
  slots_per_prefetch = (ahead + unroll_factor / 2) / unroll_factor;
Zdenek Dvorak committed
1038
  if (dump_file && (dump_flags & TDF_DETAILS))
1039 1040
    fprintf (dump_file, "Each prefetch instruction takes %u prefetch slots.\n",
	     slots_per_prefetch);
Zdenek Dvorak committed
1041 1042 1043 1044

  /* For now we just take memory references one by one and issue
     prefetches for as many as possible.  The groups are sorted
     starting with the largest step, since the references with
1045
     large step are more likely to cause many cache misses.  */
Zdenek Dvorak committed
1046 1047 1048 1049 1050 1051 1052

  for (; groups; groups = groups->next)
    for (ref = groups->refs; ref; ref = ref->next)
      {
	if (!should_issue_prefetch_p (ref))
	  continue;

1053 1054 1055 1056 1057 1058
        /* The loop is far from being sufficiently unrolled for this
           prefetch.  Do not generate prefetch to avoid many redudant
           prefetches.  */
        if (ref->prefetch_mod / unroll_factor > PREFETCH_MOD_TO_UNROLL_FACTOR_RATIO)
          continue;

1059 1060 1061 1062
	/* If we need to prefetch the reference each PREFETCH_MOD iterations,
	   and we unroll the loop UNROLL_FACTOR times, we need to insert
	   ceil (UNROLL_FACTOR / PREFETCH_MOD) instructions in each
	   iteration.  */
Zdenek Dvorak committed
1063 1064
	n_prefetches = ((unroll_factor + ref->prefetch_mod - 1)
			/ ref->prefetch_mod);
1065 1066 1067 1068 1069 1070 1071
	prefetch_slots = n_prefetches * slots_per_prefetch;

	/* If more than half of the prefetches would be lost anyway, do not
	   issue the prefetch.  */
	if (2 * remaining_prefetch_slots < prefetch_slots)
	  continue;

1072 1073 1074 1075
	/* Stop prefetching if debug counter is activated.  */
	if (!dbg_cnt (prefetch))
	  continue;

1076
	ref->issue_prefetch_p = true;
1077 1078 1079
	if (dump_file && (dump_flags & TDF_DETAILS))
	  fprintf (dump_file, "Decided to issue prefetch for reference %u:%u\n",
		   ref->group->uid, ref->uid);
Zdenek Dvorak committed
1080

1081 1082 1083
	if (remaining_prefetch_slots <= prefetch_slots)
	  return true;
	remaining_prefetch_slots -= prefetch_slots;
Zdenek Dvorak committed
1084 1085 1086 1087 1088 1089
	any = true;
      }

  return any;
}

1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107
/* Return TRUE if no prefetch is going to be generated in the given
   GROUPS.  */

static bool
nothing_to_prefetch_p (struct mem_ref_group *groups)
{
  struct mem_ref *ref;

  for (; groups; groups = groups->next)
    for (ref = groups->refs; ref; ref = ref->next)
      if (should_issue_prefetch_p (ref))
	return false;

  return true;
}

/* Estimate the number of prefetches in the given GROUPS.
   UNROLL_FACTOR is the factor by which LOOP was unrolled.  */
Zdenek Dvorak committed
1108

1109
static int
1110
estimate_prefetch_count (struct mem_ref_group *groups, unsigned unroll_factor)
Zdenek Dvorak committed
1111 1112
{
  struct mem_ref *ref;
1113
  unsigned n_prefetches;
1114
  int prefetch_count = 0;
Zdenek Dvorak committed
1115 1116 1117 1118

  for (; groups; groups = groups->next)
    for (ref = groups->refs; ref; ref = ref->next)
      if (should_issue_prefetch_p (ref))
1119 1120 1121 1122 1123
	{
	  n_prefetches = ((unroll_factor + ref->prefetch_mod - 1)
			  / ref->prefetch_mod);
	  prefetch_count += n_prefetches;
	}
Zdenek Dvorak committed
1124

1125
  return prefetch_count;
Zdenek Dvorak committed
1126 1127 1128 1129
}

/* Issue prefetches for the reference REF into loop as decided before.
   HEAD is the number of iterations to prefetch ahead.  UNROLL_FACTOR
1130
   is the factor by which LOOP was unrolled.  */
Zdenek Dvorak committed
1131 1132 1133 1134 1135

static void
issue_prefetch_ref (struct mem_ref *ref, unsigned unroll_factor, unsigned ahead)
{
  HOST_WIDE_INT delta;
1136
  tree addr, addr_base, write_p, local, forward;
1137
  gcall *prefetch;
1138
  gimple_stmt_iterator bsi;
Zdenek Dvorak committed
1139
  unsigned n_prefetches, ap;
1140
  bool nontemporal = ref->reuse_distance >= L2_CACHE_SIZE_BYTES;
Zdenek Dvorak committed
1141 1142

  if (dump_file && (dump_flags & TDF_DETAILS))
1143
    fprintf (dump_file, "Issued%s prefetch for reference %u:%u.\n",
1144
	     nontemporal ? " nontemporal" : "",
1145
	     ref->group->uid, ref->uid);
Zdenek Dvorak committed
1146

1147
  bsi = gsi_for_stmt (ref->stmt);
Zdenek Dvorak committed
1148 1149 1150 1151

  n_prefetches = ((unroll_factor + ref->prefetch_mod - 1)
		  / ref->prefetch_mod);
  addr_base = build_fold_addr_expr_with_type (ref->mem, ptr_type_node);
1152 1153
  addr_base = force_gimple_operand_gsi (&bsi, unshare_expr (addr_base),
					true, NULL, true, GSI_SAME_STMT);
1154
  write_p = ref->write_p ? integer_one_node : integer_zero_node;
1155
  local = nontemporal ? integer_zero_node : integer_three_node;
Zdenek Dvorak committed
1156 1157 1158

  for (ap = 0; ap < n_prefetches; ap++)
    {
1159 1160 1161 1162 1163
      if (cst_and_fits_in_hwi (ref->group->step))
        {
          /* Determine the address to prefetch.  */
          delta = (ahead + ap * ref->prefetch_mod) *
		   int_cst_value (ref->group->step);
1164
          addr = fold_build_pointer_plus_hwi (addr_base, delta);
1165 1166
          addr = force_gimple_operand_gsi (&bsi, unshare_expr (addr), true,
					   NULL, true, GSI_SAME_STMT);
1167 1168 1169 1170 1171 1172 1173 1174
        }
      else
        {
          /* The step size is non-constant but loop-invariant.  We use the
             heuristic to simply prefetch ahead iterations ahead.  */
          forward = fold_build2 (MULT_EXPR, sizetype,
                                 fold_convert (sizetype, ref->group->step),
                                 fold_convert (sizetype, size_int (ahead)));
1175
          addr = fold_build_pointer_plus (addr_base, forward);
1176 1177 1178
          addr = force_gimple_operand_gsi (&bsi, unshare_expr (addr), true,
					   NULL, true, GSI_SAME_STMT);
      }
1179

1180 1181
      if (addr_base != addr
	  && TREE_CODE (addr_base) == SSA_NAME
1182 1183 1184 1185 1186 1187 1188 1189 1190
	  && TREE_CODE (addr) == SSA_NAME)
	{
	  duplicate_ssa_name_ptr_info (addr, SSA_NAME_PTR_INFO (addr_base));
	  /* As this isn't a plain copy we have to reset alignment
	     information.  */
	  if (SSA_NAME_PTR_INFO (addr))
	    mark_ptr_info_alignment_unknown (SSA_NAME_PTR_INFO (addr));
	}

Zdenek Dvorak committed
1191
      /* Create the prefetch instruction.  */
1192
      prefetch = gimple_build_call (builtin_decl_explicit (BUILT_IN_PREFETCH),
1193 1194
				    3, addr, write_p, local);
      gsi_insert_before (&bsi, prefetch, GSI_SAME_STMT);
Zdenek Dvorak committed
1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213
    }
}

/* Issue prefetches for the references in GROUPS into loop as decided before.
   HEAD is the number of iterations to prefetch ahead.  UNROLL_FACTOR is the
   factor by that LOOP was unrolled.  */

static void
issue_prefetches (struct mem_ref_group *groups,
		  unsigned unroll_factor, unsigned ahead)
{
  struct mem_ref *ref;

  for (; groups; groups = groups->next)
    for (ref = groups->refs; ref; ref = ref->next)
      if (ref->issue_prefetch_p)
	issue_prefetch_ref (ref, unroll_factor, ahead);
}

1214 1215 1216 1217 1218 1219
/* Returns true if REF is a memory write for that a nontemporal store insn
   can be used.  */

static bool
nontemporal_store_p (struct mem_ref *ref)
{
1220
  machine_mode mode;
1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235
  enum insn_code code;

  /* REF must be a write that is not reused.  We require it to be independent
     on all other memory references in the loop, as the nontemporal stores may
     be reordered with respect to other memory references.  */
  if (!ref->write_p
      || !ref->independent_p
      || ref->reuse_distance < L2_CACHE_SIZE_BYTES)
    return false;

  /* Check that we have the storent instruction for the mode.  */
  mode = TYPE_MODE (TREE_TYPE (ref->mem));
  if (mode == BLKmode)
    return false;

1236
  code = optab_handler (storent_optab, mode);
1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249
  return code != CODE_FOR_nothing;
}

/* If REF is a nontemporal store, we mark the corresponding modify statement
   and return true.  Otherwise, we return false.  */

static bool
mark_nontemporal_store (struct mem_ref *ref)
{
  if (!nontemporal_store_p (ref))
    return false;

  if (dump_file && (dump_flags & TDF_DETAILS))
1250 1251
    fprintf (dump_file, "Marked reference %u:%u as a nontemporal store.\n",
	     ref->group->uid, ref->uid);
1252

1253
  gimple_assign_set_nontemporal_move (ref->stmt, true);
1254 1255 1256 1257 1258 1259 1260 1261 1262 1263
  ref->storent_p = true;

  return true;
}

/* Issue a memory fence instruction after LOOP.  */

static void
emit_mfence_after_loop (struct loop *loop)
{
1264
  vec<edge> exits = get_loop_exit_edges (loop);
1265
  edge exit;
1266
  gcall *call;
1267
  gimple_stmt_iterator bsi;
1268 1269
  unsigned i;

1270
  FOR_EACH_VEC_ELT (exits, i, exit)
1271
    {
1272
      call = gimple_build_call (FENCE_FOLLOWING_MOVNT, 0);
1273 1274 1275 1276 1277 1278

      if (!single_pred_p (exit->dest)
	  /* If possible, we prefer not to insert the fence on other paths
	     in cfg.  */
	  && !(exit->flags & EDGE_ABNORMAL))
	split_loop_exit_edge (exit);
1279
      bsi = gsi_after_labels (exit->dest);
1280

1281
      gsi_insert_before (&bsi, call, GSI_NEW_STMT);
1282 1283
    }

1284
  exits.release ();
1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301
  update_ssa (TODO_update_ssa_only_virtuals);
}

/* Returns true if we can use storent in loop, false otherwise.  */

static bool
may_use_storent_in_loop_p (struct loop *loop)
{
  bool ret = true;

  if (loop->inner != NULL)
    return false;

  /* If we must issue a mfence insn after using storent, check that there
     is a suitable place for it at each of the loop exits.  */
  if (FENCE_FOLLOWING_MOVNT != NULL_TREE)
    {
1302
      vec<edge> exits = get_loop_exit_edges (loop);
1303 1304 1305
      unsigned i;
      edge exit;

1306
      FOR_EACH_VEC_ELT (exits, i, exit)
1307
	if ((exit->flags & EDGE_ABNORMAL)
1308
	    && exit->dest == EXIT_BLOCK_PTR_FOR_FN (cfun))
1309 1310
	  ret = false;

1311
      exits.release ();
1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336
    }

  return ret;
}

/* Marks nontemporal stores in LOOP.  GROUPS contains the description of memory
   references in the loop.  */

static void
mark_nontemporal_stores (struct loop *loop, struct mem_ref_group *groups)
{
  struct mem_ref *ref;
  bool any = false;

  if (!may_use_storent_in_loop_p (loop))
    return;

  for (; groups; groups = groups->next)
    for (ref = groups->refs; ref; ref = ref->next)
      any |= mark_nontemporal_store (ref);

  if (any && FENCE_FOLLOWING_MOVNT != NULL_TREE)
    emit_mfence_after_loop (loop);
}

Zdenek Dvorak committed
1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360
/* Determines whether we can profitably unroll LOOP FACTOR times, and if
   this is the case, fill in DESC by the description of number of
   iterations.  */

static bool
should_unroll_loop_p (struct loop *loop, struct tree_niter_desc *desc,
		      unsigned factor)
{
  if (!can_unroll_loop_p (loop, factor, desc))
    return false;

  /* We only consider loops without control flow for unrolling.  This is not
     a hard restriction -- tree_unroll_loop works with arbitrary loops
     as well; but the unrolling/prefetching is usually more profitable for
     loops consisting of a single basic block, and we want to limit the
     code growth.  */
  if (loop->num_nodes > 2)
    return false;

  return true;
}

/* Determine the coefficient by that unroll LOOP, from the information
   contained in the list of memory references REFS.  Description of
1361
   number of iterations of LOOP is stored to DESC.  NINSNS is the number of
1362 1363
   insns of the LOOP.  EST_NITER is the estimated number of iterations of
   the loop, or -1 if no estimate is available.  */
Zdenek Dvorak committed
1364 1365 1366

static unsigned
determine_unroll_factor (struct loop *loop, struct mem_ref_group *refs,
1367 1368
			 unsigned ninsns, struct tree_niter_desc *desc,
			 HOST_WIDE_INT est_niter)
Zdenek Dvorak committed
1369
{
1370 1371
  unsigned upper_bound;
  unsigned nfactor, factor, mod_constraint;
Zdenek Dvorak committed
1372 1373 1374
  struct mem_ref_group *agp;
  struct mem_ref *ref;

1375 1376 1377 1378 1379 1380 1381 1382
  /* First check whether the loop is not too large to unroll.  We ignore
     PARAM_MAX_UNROLL_TIMES, because for small loops, it prevented us
     from unrolling them enough to make exactly one cache line covered by each
     iteration.  Also, the goal of PARAM_MAX_UNROLL_TIMES is to prevent
     us from unrolling the loops too many times in cases where we only expect
     gains from better scheduling and decreasing loop overhead, which is not
     the case here.  */
  upper_bound = PARAM_VALUE (PARAM_MAX_UNROLLED_INSNS) / ninsns;
1383 1384 1385 1386 1387 1388

  /* If we unrolled the loop more times than it iterates, the unrolled version
     of the loop would be never entered.  */
  if (est_niter >= 0 && est_niter < (HOST_WIDE_INT) upper_bound)
    upper_bound = est_niter;

1389
  if (upper_bound <= 1)
Zdenek Dvorak committed
1390 1391
    return 1;

1392 1393 1394
  /* Choose the factor so that we may prefetch each cache just once,
     but bound the unrolling by UPPER_BOUND.  */
  factor = 1;
Zdenek Dvorak committed
1395 1396
  for (agp = refs; agp; agp = agp->next)
    for (ref = agp->refs; ref; ref = ref->next)
1397 1398 1399 1400 1401 1402 1403
      if (should_issue_prefetch_p (ref))
	{
	  mod_constraint = ref->prefetch_mod;
	  nfactor = least_common_multiple (mod_constraint, factor);
	  if (nfactor <= upper_bound)
	    factor = nfactor;
	}
Zdenek Dvorak committed
1404 1405 1406 1407 1408 1409 1410

  if (!should_unroll_loop_p (loop, desc, factor))
    return 1;

  return factor;
}

1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484
/* Returns the total volume of the memory references REFS, taking into account
   reuses in the innermost loop and cache line size.  TODO -- we should also
   take into account reuses across the iterations of the loops in the loop
   nest.  */

static unsigned
volume_of_references (struct mem_ref_group *refs)
{
  unsigned volume = 0;
  struct mem_ref_group *gr;
  struct mem_ref *ref;

  for (gr = refs; gr; gr = gr->next)
    for (ref = gr->refs; ref; ref = ref->next)
      {
	/* Almost always reuses another value?  */
	if (ref->prefetch_before != PREFETCH_ALL)
	  continue;

	/* If several iterations access the same cache line, use the size of
	   the line divided by this number.  Otherwise, a cache line is
	   accessed in each iteration.  TODO -- in the latter case, we should
	   take the size of the reference into account, rounding it up on cache
	   line size multiple.  */
	volume += L1_CACHE_LINE_SIZE / ref->prefetch_mod;
      }
  return volume;
}

/* Returns the volume of memory references accessed across VEC iterations of
   loops, whose sizes are described in the LOOP_SIZES array.  N is the number
   of the loops in the nest (length of VEC and LOOP_SIZES vectors).  */

static unsigned
volume_of_dist_vector (lambda_vector vec, unsigned *loop_sizes, unsigned n)
{
  unsigned i;

  for (i = 0; i < n; i++)
    if (vec[i] != 0)
      break;

  if (i == n)
    return 0;

  gcc_assert (vec[i] > 0);

  /* We ignore the parts of the distance vector in subloops, since usually
     the numbers of iterations are much smaller.  */
  return loop_sizes[i] * vec[i];
}

/* Add the steps of ACCESS_FN multiplied by STRIDE to the array STRIDE
   at the position corresponding to the loop of the step.  N is the depth
   of the considered loop nest, and, LOOP is its innermost loop.  */

static void
add_subscript_strides (tree access_fn, unsigned stride,
		       HOST_WIDE_INT *strides, unsigned n, struct loop *loop)
{
  struct loop *aloop;
  tree step;
  HOST_WIDE_INT astep;
  unsigned min_depth = loop_depth (loop) - n;

  while (TREE_CODE (access_fn) == POLYNOMIAL_CHREC)
    {
      aloop = get_chrec_loop (access_fn);
      step = CHREC_RIGHT (access_fn);
      access_fn = CHREC_LEFT (access_fn);

      if ((unsigned) loop_depth (aloop) <= min_depth)
	continue;

1485
      if (tree_fits_shwi_p (step))
1486
	astep = tree_to_shwi (step);
1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505
      else
	astep = L1_CACHE_LINE_SIZE;

      strides[n - 1 - loop_depth (loop) + loop_depth (aloop)] += astep * stride;

    }
}

/* Returns the volume of memory references accessed between two consecutive
   self-reuses of the reference DR.  We consider the subscripts of DR in N
   loops, and LOOP_SIZES contains the volumes of accesses in each of the
   loops.  LOOP is the innermost loop of the current loop nest.  */

static unsigned
self_reuse_distance (data_reference_p dr, unsigned *loop_sizes, unsigned n,
		     struct loop *loop)
{
  tree stride, access_fn;
  HOST_WIDE_INT *strides, astride;
1506
  vec<tree> access_fns;
1507 1508 1509 1510 1511 1512 1513 1514 1515 1516 1517 1518 1519 1520 1521 1522 1523 1524
  tree ref = DR_REF (dr);
  unsigned i, ret = ~0u;

  /* In the following example:

     for (i = 0; i < N; i++)
       for (j = 0; j < N; j++)
         use (a[j][i]);
     the same cache line is accessed each N steps (except if the change from
     i to i + 1 crosses the boundary of the cache line).  Thus, for self-reuse,
     we cannot rely purely on the results of the data dependence analysis.

     Instead, we compute the stride of the reference in each loop, and consider
     the innermost loop in that the stride is less than cache size.  */

  strides = XCNEWVEC (HOST_WIDE_INT, n);
  access_fns = DR_ACCESS_FNS (dr);

1525
  FOR_EACH_VEC_ELT (access_fns, i, access_fn)
1526 1527 1528 1529 1530
    {
      /* Keep track of the reference corresponding to the subscript, so that we
	 know its stride.  */
      while (handled_component_p (ref) && TREE_CODE (ref) != ARRAY_REF)
	ref = TREE_OPERAND (ref, 0);
H.J. Lu committed
1531

1532 1533 1534
      if (TREE_CODE (ref) == ARRAY_REF)
	{
	  stride = TYPE_SIZE_UNIT (TREE_TYPE (ref));
1535
	  if (tree_fits_uhwi_p (stride))
1536
	    astride = tree_to_uhwi (stride);
1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567
	  else
	    astride = L1_CACHE_LINE_SIZE;

	  ref = TREE_OPERAND (ref, 0);
	}
      else
	astride = 1;

      add_subscript_strides (access_fn, astride, strides, n, loop);
    }

  for (i = n; i-- > 0; )
    {
      unsigned HOST_WIDE_INT s;

      s = strides[i] < 0 ?  -strides[i] : strides[i];

      if (s < (unsigned) L1_CACHE_LINE_SIZE
	  && (loop_sizes[i]
	      > (unsigned) (L1_CACHE_SIZE_BYTES / NONTEMPORAL_FRACTION)))
	{
	  ret = loop_sizes[i];
	  break;
	}
    }

  free (strides);
  return ret;
}

/* Determines the distance till the first reuse of each reference in REFS
1568
   in the loop nest of LOOP.  NO_OTHER_REFS is true if there are no other
1569
   memory references in the loop.  Return false if the analysis fails.  */
1570

1571
static bool
1572 1573
determine_loop_nest_reuse (struct loop *loop, struct mem_ref_group *refs,
			   bool no_other_refs)
1574 1575
{
  struct loop *nest, *aloop;
1576 1577
  vec<data_reference_p> datarefs = vNULL;
  vec<ddr_p> dependences = vNULL;
1578
  struct mem_ref_group *gr;
1579
  struct mem_ref *ref, *refb;
1580
  auto_vec<loop_p> vloops;
1581 1582 1583 1584 1585 1586 1587 1588
  unsigned *loop_data_size;
  unsigned i, j, n;
  unsigned volume, dist, adist;
  HOST_WIDE_INT vol;
  data_reference_p dr;
  ddr_p dep;

  if (loop->inner)
1589
    return true;
1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608

  /* Find the outermost loop of the loop nest of loop (we require that
     there are no sibling loops inside the nest).  */
  nest = loop;
  while (1)
    {
      aloop = loop_outer (nest);

      if (aloop == current_loops->tree_root
	  || aloop->inner->next)
	break;

      nest = aloop;
    }

  /* For each loop, determine the amount of data accessed in each iteration.
     We use this to estimate whether the reference is evicted from the
     cache before its reuse.  */
  find_loop_nest (nest, &vloops);
1609
  n = vloops.length ();
1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620
  loop_data_size = XNEWVEC (unsigned, n);
  volume = volume_of_references (refs);
  i = n;
  while (i-- != 0)
    {
      loop_data_size[i] = volume;
      /* Bound the volume by the L2 cache size, since above this bound,
	 all dependence distances are equivalent.  */
      if (volume > L2_CACHE_SIZE_BYTES)
	continue;

1621
      aloop = vloops[i];
1622
      vol = estimated_stmt_executions_int (aloop);
1623
      if (vol == -1)
1624 1625 1626 1627 1628
	vol = expected_loop_iterations (aloop);
      volume *= vol;
    }

  /* Prepare the references in the form suitable for data dependence
1629
     analysis.  We ignore unanalyzable data references (the results
1630 1631 1632 1633 1634
     are used just as a heuristics to estimate temporality of the
     references, hence we do not need to worry about correctness).  */
  for (gr = refs; gr; gr = gr->next)
    for (ref = gr->refs; ref; ref = ref->next)
      {
1635
	dr = create_data_ref (nest, loop_containing_stmt (ref->stmt),
1636
			      ref->mem, ref->stmt, !ref->write_p, false);
1637 1638 1639 1640 1641

	if (dr)
	  {
	    ref->reuse_distance = volume;
	    dr->aux = ref;
1642
	    datarefs.safe_push (dr);
1643
	  }
1644 1645
	else
	  no_other_refs = false;
1646 1647
      }

1648
  FOR_EACH_VEC_ELT (datarefs, i, dr)
1649 1650
    {
      dist = self_reuse_distance (dr, loop_data_size, n, loop);
1651
      ref = (struct mem_ref *) dr->aux;
1652 1653
      if (ref->reuse_distance > dist)
	ref->reuse_distance = dist;
1654 1655 1656

      if (no_other_refs)
	ref->independent_p = true;
1657 1658
    }

1659 1660
  if (!compute_all_dependences (datarefs, &dependences, vloops, true))
    return false;
1661

1662
  FOR_EACH_VEC_ELT (dependences, i, dep)
1663 1664 1665 1666
    {
      if (DDR_ARE_DEPENDENT (dep) == chrec_known)
	continue;

1667 1668
      ref = (struct mem_ref *) DDR_A (dep)->aux;
      refb = (struct mem_ref *) DDR_B (dep)->aux;
1669

1670
      if (DDR_ARE_DEPENDENT (dep) == chrec_dont_know
1671
	  || DDR_COULD_BE_INDEPENDENT_P (dep)
1672 1673
	  || DDR_NUM_DIST_VECTS (dep) == 0)
	{
1674
	  /* If the dependence cannot be analyzed, assume that there might be
1675 1676
	     a reuse.  */
	  dist = 0;
H.J. Lu committed
1677

1678 1679
	  ref->independent_p = false;
	  refb->independent_p = false;
1680 1681 1682
	}
      else
	{
1683
	  /* The distance vectors are normalized to be always lexicographically
1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696
	     positive, hence we cannot tell just from them whether DDR_A comes
	     before DDR_B or vice versa.  However, it is not important,
	     anyway -- if DDR_A is close to DDR_B, then it is either reused in
	     DDR_B (and it is not nontemporal), or it reuses the value of DDR_B
	     in cache (and marking it as nontemporal would not affect
	     anything).  */

	  dist = volume;
	  for (j = 0; j < DDR_NUM_DIST_VECTS (dep); j++)
	    {
	      adist = volume_of_dist_vector (DDR_DIST_VECT (dep, j),
					     loop_data_size, n);

1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708
	      /* If this is a dependence in the innermost loop (i.e., the
		 distances in all superloops are zero) and it is not
		 the trivial self-dependence with distance zero, record that
		 the references are not completely independent.  */
	      if (lambda_vector_zerop (DDR_DIST_VECT (dep, j), n - 1)
		  && (ref != refb
		      || DDR_DIST_VECT (dep, j)[n-1] != 0))
		{
		  ref->independent_p = false;
		  refb->independent_p = false;
		}

1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723
	      /* Ignore accesses closer than
		 L1_CACHE_SIZE_BYTES / NONTEMPORAL_FRACTION,
	      	 so that we use nontemporal prefetches e.g. if single memory
		 location is accessed several times in a single iteration of
		 the loop.  */
	      if (adist < L1_CACHE_SIZE_BYTES / NONTEMPORAL_FRACTION)
		continue;

	      if (adist < dist)
		dist = adist;
	    }
	}

      if (ref->reuse_distance > dist)
	ref->reuse_distance = dist;
1724 1725
      if (refb->reuse_distance > dist)
	refb->reuse_distance = dist;
1726 1727 1728 1729 1730 1731 1732 1733 1734 1735 1736
    }

  free_dependence_relations (dependences);
  free_data_refs (datarefs);
  free (loop_data_size);

  if (dump_file && (dump_flags & TDF_DETAILS))
    {
      fprintf (dump_file, "Reuse distances:\n");
      for (gr = refs; gr; gr = gr->next)
	for (ref = gr->refs; ref; ref = ref->next)
1737 1738
	  fprintf (dump_file, " reference %u:%u distance %u\n",
		   ref->group->uid, ref->uid, ref->reuse_distance);
1739
    }
1740 1741

  return true;
1742 1743
}

1744 1745
/* Determine whether or not the trip count to ahead ratio is too small based
   on prefitablility consideration.
1746
   AHEAD: the iteration ahead distance,
1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770
   EST_NITER: the estimated trip count.  */

static bool
trip_count_to_ahead_ratio_too_small_p (unsigned ahead, HOST_WIDE_INT est_niter)
{
  /* Assume trip count to ahead ratio is big enough if the trip count could not
     be estimated at compile time.  */
  if (est_niter < 0)
    return false;

  if (est_niter < (HOST_WIDE_INT) (TRIP_COUNT_TO_AHEAD_RATIO * ahead))
    {
      if (dump_file && (dump_flags & TDF_DETAILS))
	fprintf (dump_file,
		 "Not prefetching -- loop estimated to roll only %d times\n",
		 (int) est_niter);
      return true;
    }

  return false;
}

/* Determine whether or not the number of memory references in the loop is
   reasonable based on the profitablity and compilation time considerations.
1771 1772 1773
   NINSNS: estimated number of instructions in the loop,
   MEM_REF_COUNT: total number of memory references in the loop.  */

H.J. Lu committed
1774
static bool
1775
mem_ref_count_reasonable_p (unsigned ninsns, unsigned mem_ref_count)
1776
{
1777
  int insn_to_mem_ratio;
1778 1779 1780 1781

  if (mem_ref_count == 0)
    return false;

1782 1783 1784 1785 1786 1787 1788
  /* Miss rate computation (is_miss_rate_acceptable) and dependence analysis
     (compute_all_dependences) have high costs based on quadratic complexity.
     To avoid huge compilation time, we give up prefetching if mem_ref_count
     is too large.  */
  if (mem_ref_count > PREFETCH_MAX_MEM_REFS_PER_LOOP)
    return false;

H.J. Lu committed
1789 1790 1791 1792 1793
  /* Prefetching improves performance by overlapping cache missing
     memory accesses with CPU operations.  If the loop does not have
     enough CPU operations to overlap with memory operations, prefetching
     won't give a significant benefit.  One approximate way of checking
     this is to require the ratio of instructions to memory references to
1794 1795 1796 1797
     be above a certain limit.  This approximation works well in practice.
     TODO: Implement a more precise computation by estimating the time
     for each CPU or memory op in the loop. Time estimates for memory ops
     should account for cache misses.  */
H.J. Lu committed
1798
  insn_to_mem_ratio = ninsns / mem_ref_count;
1799 1800

  if (insn_to_mem_ratio < PREFETCH_MIN_INSN_TO_MEM_RATIO)
1801 1802 1803 1804 1805 1806 1807
    {
      if (dump_file && (dump_flags & TDF_DETAILS))
        fprintf (dump_file,
		 "Not prefetching -- instruction to memory reference ratio (%d) too small\n",
		 insn_to_mem_ratio);
      return false;
    }
1808

1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819 1820 1821 1822 1823
  return true;
}

/* Determine whether or not the instruction to prefetch ratio in the loop is
   too small based on the profitablity consideration.
   NINSNS: estimated number of instructions in the loop,
   PREFETCH_COUNT: an estimate of the number of prefetches,
   UNROLL_FACTOR:  the factor to unroll the loop if prefetching.  */

static bool
insn_to_prefetch_ratio_too_small_p (unsigned ninsns, unsigned prefetch_count,
                                     unsigned unroll_factor)
{
  int insn_to_prefetch_ratio;

1824 1825 1826
  /* Prefetching most likely causes performance degradation when the instruction
     to prefetch ratio is too small.  Too many prefetch instructions in a loop
     may reduce the I-cache performance.
1827 1828 1829 1830 1831 1832 1833 1834 1835
     (unroll_factor * ninsns) is used to estimate the number of instructions in
     the unrolled loop.  This implementation is a bit simplistic -- the number
     of issued prefetch instructions is also affected by unrolling.  So,
     prefetch_mod and the unroll factor should be taken into account when
     determining prefetch_count.  Also, the number of insns of the unrolled
     loop will usually be significantly smaller than the number of insns of the
     original loop * unroll_factor (at least the induction variable increases
     and the exit branches will get eliminated), so it might be better to use
     tree_estimate_loop_size + estimated_unrolled_size.  */
1836 1837
  insn_to_prefetch_ratio = (unroll_factor * ninsns) / prefetch_count;
  if (insn_to_prefetch_ratio < MIN_INSN_TO_PREFETCH_RATIO)
1838
    {
1839 1840 1841 1842
      if (dump_file && (dump_flags & TDF_DETAILS))
        fprintf (dump_file,
		 "Not prefetching -- instruction to prefetch ratio (%d) too small\n",
		 insn_to_prefetch_ratio);
1843
      return true;
1844
    }
H.J. Lu committed
1845

1846
  return false;
1847 1848 1849
}


Zdenek Dvorak committed
1850
/* Issue prefetch instructions for array references in LOOP.  Returns
1851
   true if the LOOP was unrolled.  */
Zdenek Dvorak committed
1852 1853

static bool
1854
loop_prefetch_arrays (struct loop *loop)
Zdenek Dvorak committed
1855 1856
{
  struct mem_ref_group *refs;
1857 1858
  unsigned ahead, ninsns, time, unroll_factor;
  HOST_WIDE_INT est_niter;
Zdenek Dvorak committed
1859
  struct tree_niter_desc desc;
1860
  bool unrolled = false, no_other_refs;
1861 1862
  unsigned prefetch_count;
  unsigned mem_ref_count;
Zdenek Dvorak committed
1863

1864
  if (optimize_loop_nest_for_size_p (loop))
1865 1866 1867 1868 1869 1870
    {
      if (dump_file && (dump_flags & TDF_DETAILS))
	fprintf (dump_file, "  ignored (cold area)\n");
      return false;
    }

1871 1872 1873 1874 1875 1876 1877
  /* FIXME: the time should be weighted by the probabilities of the blocks in
     the loop body.  */
  time = tree_num_loop_insns (loop, &eni_time_weights);
  if (time == 0)
    return false;

  ahead = (PREFETCH_LATENCY + time - 1) / time;
1878
  est_niter = estimated_stmt_executions_int (loop);
1879
  if (est_niter == -1)
1880
    est_niter = likely_max_stmt_executions_int (loop);
1881 1882 1883 1884 1885 1886 1887 1888

  /* Prefetching is not likely to be profitable if the trip count to ahead
     ratio is too small.  */
  if (trip_count_to_ahead_ratio_too_small_p (ahead, est_niter))
    return false;

  ninsns = tree_num_loop_insns (loop, &eni_size_weights);

Zdenek Dvorak committed
1889
  /* Step 1: gather the memory references.  */
1890
  refs = gather_memory_references (loop, &no_other_refs, &mem_ref_count);
Zdenek Dvorak committed
1891

1892 1893 1894 1895 1896 1897
  /* Give up prefetching if the number of memory references in the
     loop is not reasonable based on profitablity and compilation time
     considerations.  */
  if (!mem_ref_count_reasonable_p (ninsns, mem_ref_count))
    goto fail;

Zdenek Dvorak committed
1898 1899 1900
  /* Step 2: estimate the reuse effects.  */
  prune_by_reuse (refs);

1901
  if (nothing_to_prefetch_p (refs))
Zdenek Dvorak committed
1902 1903
    goto fail;

1904 1905
  if (!determine_loop_nest_reuse (loop, refs, no_other_refs))
    goto fail;
1906

1907
  /* Step 3: determine unroll factor.  */
1908 1909
  unroll_factor = determine_unroll_factor (loop, refs, ninsns, &desc,
					   est_niter);
1910 1911 1912 1913 1914 1915

  /* Estimate prefetch count for the unrolled loop.  */
  prefetch_count = estimate_prefetch_count (refs, unroll_factor);
  if (prefetch_count == 0)
    goto fail;

1916
  if (dump_file && (dump_flags & TDF_DETAILS))
H.J. Lu committed
1917
    fprintf (dump_file, "Ahead %d, unroll factor %d, trip count "
1918
	     HOST_WIDE_INT_PRINT_DEC "\n"
H.J. Lu committed
1919 1920 1921
	     "insn count %d, mem ref count %d, prefetch count %d\n",
	     ahead, unroll_factor, est_niter,
	     ninsns, mem_ref_count, prefetch_count);
1922

1923 1924 1925 1926
  /* Prefetching is not likely to be profitable if the instruction to prefetch
     ratio is too small.  */
  if (insn_to_prefetch_ratio_too_small_p (ninsns, prefetch_count,
					  unroll_factor))
1927 1928 1929
    goto fail;

  mark_nontemporal_stores (loop, refs);
1930

Zdenek Dvorak committed
1931 1932 1933 1934 1935 1936 1937 1938
  /* Step 4: what to prefetch?  */
  if (!schedule_prefetches (refs, unroll_factor, ahead))
    goto fail;

  /* Step 5: unroll the loop.  TODO -- peeling of first and last few
     iterations so that we do not issue superfluous prefetches.  */
  if (unroll_factor != 1)
    {
1939
      tree_unroll_loop (loop, unroll_factor,
Zdenek Dvorak committed
1940 1941 1942 1943 1944 1945 1946 1947 1948 1949 1950 1951
			single_dom_exit (loop), &desc);
      unrolled = true;
    }

  /* Step 6: issue the prefetches.  */
  issue_prefetches (refs, unroll_factor, ahead);

fail:
  release_mem_refs (refs);
  return unrolled;
}

1952
/* Issue prefetch instructions for array references in loops.  */
Zdenek Dvorak committed
1953

1954
unsigned int
1955
tree_ssa_prefetch_arrays (void)
Zdenek Dvorak committed
1956 1957 1958
{
  struct loop *loop;
  bool unrolled = false;
1959
  int todo_flags = 0;
Zdenek Dvorak committed
1960

1961
  if (!targetm.have_prefetch ()
Zdenek Dvorak committed
1962 1963 1964
      /* It is possible to ask compiler for say -mtune=i486 -march=pentium4.
	 -mtune=i486 causes us having PREFETCH_BLOCK 0, since this is part
	 of processor costs and i486 does not have prefetch, but
1965
	 -march=pentium4 causes targetm.have_prefetch to be true.  Ugh.  */
Zdenek Dvorak committed
1966
      || PREFETCH_BLOCK == 0)
1967
    return 0;
Zdenek Dvorak committed
1968

1969 1970 1971 1972 1973 1974 1975
  if (dump_file && (dump_flags & TDF_DETAILS))
    {
      fprintf (dump_file, "Prefetching parameters:\n");
      fprintf (dump_file, "    simultaneous prefetches: %d\n",
	       SIMULTANEOUS_PREFETCHES);
      fprintf (dump_file, "    prefetch latency: %d\n", PREFETCH_LATENCY);
      fprintf (dump_file, "    prefetch block size: %d\n", PREFETCH_BLOCK);
1976 1977
      fprintf (dump_file, "    L1 cache size: %d lines, %d kB\n",
	       L1_CACHE_SIZE_BYTES / L1_CACHE_LINE_SIZE, L1_CACHE_SIZE);
1978
      fprintf (dump_file, "    L1 cache line size: %d\n", L1_CACHE_LINE_SIZE);
H.J. Lu committed
1979 1980
      fprintf (dump_file, "    L2 cache size: %d kB\n", L2_CACHE_SIZE);
      fprintf (dump_file, "    min insn-to-prefetch ratio: %d \n",
1981
	       MIN_INSN_TO_PREFETCH_RATIO);
H.J. Lu committed
1982
      fprintf (dump_file, "    min insn-to-mem ratio: %d \n",
1983
	       PREFETCH_MIN_INSN_TO_MEM_RATIO);
1984 1985 1986
      fprintf (dump_file, "\n");
    }

Zdenek Dvorak committed
1987 1988
  initialize_original_copy_tables ();

1989
  if (!builtin_decl_explicit_p (BUILT_IN_PREFETCH))
Zdenek Dvorak committed
1990
    {
1991 1992
      tree type = build_function_type_list (void_type_node,
					    const_ptr_type_node, NULL_TREE);
1993 1994 1995
      tree decl = add_builtin_function ("__builtin_prefetch", type,
					BUILT_IN_PREFETCH, BUILT_IN_NORMAL,
					NULL, NULL_TREE);
Zdenek Dvorak committed
1996
      DECL_IS_NOVOPS (decl) = true;
1997
      set_builtin_decl (BUILT_IN_PREFETCH, decl, false);
Zdenek Dvorak committed
1998 1999
    }

2000
  FOR_EACH_LOOP (loop, LI_FROM_INNERMOST)
Zdenek Dvorak committed
2001 2002 2003 2004
    {
      if (dump_file && (dump_flags & TDF_DETAILS))
	fprintf (dump_file, "Processing loop %d:\n", loop->num);

2005
      unrolled |= loop_prefetch_arrays (loop);
Zdenek Dvorak committed
2006 2007 2008 2009 2010 2011 2012 2013

      if (dump_file && (dump_flags & TDF_DETAILS))
	fprintf (dump_file, "\n\n");
    }

  if (unrolled)
    {
      scev_reset ();
2014
      todo_flags |= TODO_cleanup_cfg;
Zdenek Dvorak committed
2015 2016 2017
    }

  free_original_copy_tables ();
2018
  return todo_flags;
Zdenek Dvorak committed
2019
}
2020 2021 2022 2023 2024 2025 2026 2027 2028 2029 2030 2031 2032 2033 2034 2035 2036 2037 2038 2039 2040 2041 2042 2043 2044 2045

/* Prefetching.  */

namespace {

const pass_data pass_data_loop_prefetch =
{
  GIMPLE_PASS, /* type */
  "aprefetch", /* name */
  OPTGROUP_LOOP, /* optinfo_flags */
  TV_TREE_PREFETCH, /* tv_id */
  ( PROP_cfg | PROP_ssa ), /* properties_required */
  0, /* properties_provided */
  0, /* properties_destroyed */
  0, /* todo_flags_start */
  0, /* todo_flags_finish */
};

class pass_loop_prefetch : public gimple_opt_pass
{
public:
  pass_loop_prefetch (gcc::context *ctxt)
    : gimple_opt_pass (pass_data_loop_prefetch, ctxt)
  {}

  /* opt_pass methods: */
2046
  virtual bool gate (function *) { return flag_prefetch_loop_arrays > 0; }
2047
  virtual unsigned int execute (function *);
2048 2049 2050

}; // class pass_loop_prefetch

2051 2052 2053 2054 2055 2056
unsigned int
pass_loop_prefetch::execute (function *fun)
{
  if (number_of_loops (fun) <= 1)
    return 0;

2057 2058 2059 2060 2061 2062 2063 2064 2065 2066 2067 2068 2069 2070
  if ((PREFETCH_BLOCK & (PREFETCH_BLOCK - 1)) != 0)
    {
      static bool warned = false;

      if (!warned)
	{
	  warning (OPT_Wdisabled_optimization,
		   "%<l1-cache-size%> parameter is not a power of two %d",
		   PREFETCH_BLOCK);
	  warned = true;
	}
      return 0;
    }

2071 2072 2073
  return tree_ssa_prefetch_arrays ();
}

2074 2075 2076 2077 2078 2079 2080 2081 2082
} // anon namespace

gimple_opt_pass *
make_pass_loop_prefetch (gcc::context *ctxt)
{
  return new pass_loop_prefetch (ctxt);
}