explow.c 68.9 KB
Newer Older
Richard Kenner committed
1
/* Subroutines for manipulating rtx's in semantically interesting ways.
2
   Copyright (C) 1987-2019 Free Software Foundation, Inc.
Richard Kenner committed
3

4
This file is part of GCC.
Richard Kenner committed
5

6 7
GCC is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free
8
Software Foundation; either version 3, or (at your option) any later
9
version.
Richard Kenner committed
10

11 12 13 14
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
WARRANTY; without even the implied warranty of MERCHANTABILITY or
FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
for more details.
Richard Kenner committed
15 16

You should have received a copy of the GNU General Public License
17 18
along with GCC; see the file COPYING3.  If not see
<http://www.gnu.org/licenses/>.  */
Richard Kenner committed
19 20 21


#include "config.h"
22
#include "system.h"
23
#include "coretypes.h"
24 25
#include "target.h"
#include "function.h"
Richard Kenner committed
26 27
#include "rtl.h"
#include "tree.h"
28
#include "memmodel.h"
29
#include "tm_p.h"
30
#include "expmed.h"
31
#include "profile-count.h"
32 33 34 35 36
#include "optabs.h"
#include "emit-rtl.h"
#include "recog.h"
#include "diagnostic-core.h"
#include "stor-layout.h"
37
#include "except.h"
38 39
#include "dojump.h"
#include "explow.h"
Richard Kenner committed
40
#include "expr.h"
41
#include "stringpool.h"
42
#include "common/common-target.h"
43
#include "output.h"
Richard Kenner committed
44

45
static rtx break_out_memory_refs (rtx);
46
static void anti_adjust_stack_and_probe_stack_clash (rtx);
47 48 49 50 51


/* Truncate and perhaps sign-extend C as appropriate for MODE.  */

HOST_WIDE_INT
52
trunc_int_for_mode (HOST_WIDE_INT c, machine_mode mode)
53
{
54 55 56
  /* Not scalar_int_mode because we also allow pointer bound modes.  */
  scalar_mode smode = as_a <scalar_mode> (mode);
  int width = GET_MODE_PRECISION (smode);
57

58
  /* You want to truncate to a _what_?  */
59
  gcc_assert (SCALAR_INT_MODE_P (mode));
60

61
  /* Canonicalize BImode to 0 and STORE_FLAG_VALUE.  */
62
  if (smode == BImode)
63 64
    return c & 1 ? STORE_FLAG_VALUE : 0;

65 66 67 68 69 70 71 72 73 74
  /* Sign-extend for the requested mode.  */

  if (width < HOST_BITS_PER_WIDE_INT)
    {
      HOST_WIDE_INT sign = 1;
      sign <<= width - 1;
      c &= (sign << 1) - 1;
      c ^= sign;
      c -= sign;
    }
75 76 77 78

  return c;
}

79 80 81 82 83 84 85 86 87 88 89
/* Likewise for polynomial values, using the sign-extended representation
   for each individual coefficient.  */

poly_int64
trunc_int_for_mode (poly_int64 x, machine_mode mode)
{
  for (unsigned int i = 0; i < NUM_POLY_INT_COEFFS; ++i)
    x.coeffs[i] = trunc_int_for_mode (x.coeffs[i], mode);
  return x;
}

90
/* Return an rtx for the sum of X and the integer C, given that X has
91 92
   mode MODE.  INPLACE is true if X can be modified inplace or false
   if it must be treated as immutable.  */
Richard Kenner committed
93 94

rtx
95
plus_constant (machine_mode mode, rtx x, poly_int64 c, bool inplace)
Richard Kenner committed
96
{
97
  RTX_CODE code;
98
  rtx y;
99
  rtx tem;
Richard Kenner committed
100 101
  int all_constant = 0;

102 103
  gcc_assert (GET_MODE (x) == VOIDmode || GET_MODE (x) == mode);

104
  if (known_eq (c, 0))
Richard Kenner committed
105 106 107 108 109
    return x;

 restart:

  code = GET_CODE (x);
110 111
  y = x;

Richard Kenner committed
112 113
  switch (code)
    {
Kenneth Zadeck committed
114
    CASE_CONST_SCALAR_INT:
115
      return immed_wide_int_const (wi::add (rtx_mode_t (x, mode), c), mode);
Richard Kenner committed
116 117 118 119 120 121 122
    case MEM:
      /* If this is a reference to the constant pool, try replacing it with
	 a reference to a new constant.  If the resulting address isn't
	 valid, don't return it because we have no way to validize it.  */
      if (GET_CODE (XEXP (x, 0)) == SYMBOL_REF
	  && CONSTANT_POOL_ADDRESS_P (XEXP (x, 0)))
	{
123 124 125 126 127 128 129 130
	  rtx cst = get_pool_constant (XEXP (x, 0));

	  if (GET_CODE (cst) == CONST_VECTOR
	      && GET_MODE_INNER (GET_MODE (cst)) == mode)
	    {
	      cst = gen_lowpart (mode, cst);
	      gcc_assert (cst);
	    }
131 132 133 134 135 136 137 138 139
	  if (GET_MODE (cst) == VOIDmode || GET_MODE (cst) == mode)
	    {
	      tem = plus_constant (mode, cst, c);
	      tem = force_const_mem (GET_MODE (x), tem);
	      /* Targets may disallow some constants in the constant pool, thus
		 force_const_mem may return NULL_RTX.  */
	      if (tem && memory_address_p (GET_MODE (tem), XEXP (tem, 0)))
		return tem;
	    }
Richard Kenner committed
140 141 142 143 144 145
	}
      break;

    case CONST:
      /* If adding to something entirely constant, set a flag
	 so that we can add a CONST around the result.  */
146 147
      if (inplace && shared_const_p (x))
	inplace = false;
Richard Kenner committed
148 149 150 151 152 153 154 155 156 157
      x = XEXP (x, 0);
      all_constant = 1;
      goto restart;

    case SYMBOL_REF:
    case LABEL_REF:
      all_constant = 1;
      break;

    case PLUS:
158 159 160 161
      /* The interesting case is adding the integer to a sum.  Look
	 for constant term in the sum and combine with C.  For an
	 integer constant term or a constant term that is not an
	 explicit integer, we combine or group them together anyway.
162 163 164

	 We may not immediately return from the recursive call here, lest
	 all_constant gets lost.  */
165

166
      if (CONSTANT_P (XEXP (x, 1)))
167
	{
168 169 170 171 172 173 174
	  rtx term = plus_constant (mode, XEXP (x, 1), c, inplace);
	  if (term == const0_rtx)
	    x = XEXP (x, 0);
	  else if (inplace)
	    XEXP (x, 1) = term;
	  else
	    x = gen_rtx_PLUS (mode, XEXP (x, 0), term);
175 176
	  c = 0;
	}
177
      else if (rtx *const_loc = find_constant_term_loc (&y))
178
	{
179 180 181 182 183 184 185 186
	  if (!inplace)
	    {
	      /* We need to be careful since X may be shared and we can't
		 modify it in place.  */
	      x = copy_rtx (x);
	      const_loc = find_constant_term_loc (&x);
	    }
	  *const_loc = plus_constant (mode, *const_loc, c, true);
187 188
	  c = 0;
	}
189
      break;
190

191
    default:
192 193
      if (CONST_POLY_INT_P (x))
	return immed_wide_int_const (const_poly_int_value (x) + c, mode);
194
      break;
Richard Kenner committed
195 196
    }

197
  if (maybe_ne (c, 0))
198
    x = gen_rtx_PLUS (mode, x, gen_int_mode (c, mode));
Richard Kenner committed
199 200 201 202

  if (GET_CODE (x) == SYMBOL_REF || GET_CODE (x) == LABEL_REF)
    return x;
  else if (all_constant)
203
    return gen_rtx_CONST (mode, x);
Richard Kenner committed
204 205 206 207 208 209 210 211 212 213
  else
    return x;
}

/* If X is a sum, return a new sum like X but lacking any constant terms.
   Add all the removed constant terms into *CONSTPTR.
   X itself is not altered.  The result != X if and only if
   it is not isomorphic to X.  */

rtx
214
eliminate_constant_term (rtx x, rtx *constptr)
Richard Kenner committed
215
{
216
  rtx x0, x1;
Richard Kenner committed
217 218 219 220 221 222
  rtx tem;

  if (GET_CODE (x) != PLUS)
    return x;

  /* First handle constants appearing at this level explicitly.  */
Shujing Zhao committed
223
  if (CONST_INT_P (XEXP (x, 1))
224 225
      && (tem = simplify_binary_operation (PLUS, GET_MODE (x), *constptr,
					   XEXP (x, 1))) != 0
Shujing Zhao committed
226
      && CONST_INT_P (tem))
Richard Kenner committed
227 228 229 230 231 232 233 234 235
    {
      *constptr = tem;
      return eliminate_constant_term (XEXP (x, 0), constptr);
    }

  tem = const0_rtx;
  x0 = eliminate_constant_term (XEXP (x, 0), &tem);
  x1 = eliminate_constant_term (XEXP (x, 1), &tem);
  if ((x1 != XEXP (x, 1) || x0 != XEXP (x, 0))
236 237
      && (tem = simplify_binary_operation (PLUS, GET_MODE (x),
					   *constptr, tem)) != 0
Shujing Zhao committed
238
      && CONST_INT_P (tem))
Richard Kenner committed
239 240
    {
      *constptr = tem;
241
      return gen_rtx_PLUS (GET_MODE (x), x0, x1);
Richard Kenner committed
242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264
    }

  return x;
}


/* Return a copy of X in which all memory references
   and all constants that involve symbol refs
   have been replaced with new temporary registers.
   Also emit code to load the memory locations and constants
   into those registers.

   If X contains no such constants or memory references,
   X itself (not a copy) is returned.

   If a constant is found in the address that is not a legitimate constant
   in an insn, it is left alone in the hope that it might be valid in the
   address.

   X may contain no arithmetic except addition, subtraction and multiplication.
   Values returned by expand_expr with 1 for sum_ok fit this constraint.  */

static rtx
265
break_out_memory_refs (rtx x)
Richard Kenner committed
266
{
267
  if (MEM_P (x)
268
      || (CONSTANT_P (x) && CONSTANT_ADDRESS_P (x)
Richard Kenner committed
269
	  && GET_MODE (x) != VOIDmode))
270
    x = force_reg (GET_MODE (x), x);
Richard Kenner committed
271 272 273
  else if (GET_CODE (x) == PLUS || GET_CODE (x) == MINUS
	   || GET_CODE (x) == MULT)
    {
274 275
      rtx op0 = break_out_memory_refs (XEXP (x, 0));
      rtx op1 = break_out_memory_refs (XEXP (x, 1));
276

Richard Kenner committed
277
      if (op0 != XEXP (x, 0) || op1 != XEXP (x, 1))
278
	x = simplify_gen_binary (GET_CODE (x), GET_MODE (x), op0, op1);
Richard Kenner committed
279
    }
280

Richard Kenner committed
281 282 283
  return x;
}

284 285 286 287
/* Given X, a memory address in address space AS' pointer mode, convert it to
   an address in the address space's address mode, or vice versa (TO_MODE says
   which way).  We take advantage of the fact that pointers are not allowed to
   overflow by commuting arithmetic operations over conversions so that address
288
   arithmetic insns can be used. IN_CONST is true if this conversion is inside
289 290
   a CONST. NO_EMIT is true if no insns should be emitted, and instead
   it should return NULL if it can't be simplified without emitting insns.  */
291

292
rtx
293
convert_memory_address_addr_space_1 (scalar_int_mode to_mode ATTRIBUTE_UNUSED,
294
				     rtx x, addr_space_t as ATTRIBUTE_UNUSED,
295 296
				     bool in_const ATTRIBUTE_UNUSED,
				     bool no_emit ATTRIBUTE_UNUSED)
297
{
298
#ifndef POINTERS_EXTEND_UNSIGNED
299
  gcc_assert (GET_MODE (x) == to_mode || GET_MODE (x) == VOIDmode);
300 301
  return x;
#else /* defined(POINTERS_EXTEND_UNSIGNED) */
302
  scalar_int_mode pointer_mode, address_mode, from_mode;
303
  rtx temp;
304
  enum rtx_code code;
305

306 307 308 309
  /* If X already has the right mode, just return it.  */
  if (GET_MODE (x) == to_mode)
    return x;

310 311 312
  pointer_mode = targetm.addr_space.pointer_mode (as);
  address_mode = targetm.addr_space.address_mode (as);
  from_mode = to_mode == pointer_mode ? address_mode : pointer_mode;
313

314 315
  /* Here we handle some special cases.  If none of them apply, fall through
     to the default case.  */
316 317
  switch (GET_CODE (x))
    {
318
    CASE_CONST_SCALAR_INT:
319 320 321 322 323 324 325 326 327 328 329 330
      if (GET_MODE_SIZE (to_mode) < GET_MODE_SIZE (from_mode))
	code = TRUNCATE;
      else if (POINTERS_EXTEND_UNSIGNED < 0)
	break;
      else if (POINTERS_EXTEND_UNSIGNED > 0)
	code = ZERO_EXTEND;
      else
	code = SIGN_EXTEND;
      temp = simplify_unary_operation (code, to_mode, x, from_mode);
      if (temp)
	return temp;
      break;
331

332
    case SUBREG:
333
      if ((SUBREG_PROMOTED_VAR_P (x) || REG_POINTER (SUBREG_REG (x)))
334
	  && GET_MODE (SUBREG_REG (x)) == to_mode)
335 336 337
	return SUBREG_REG (x);
      break;

338
    case LABEL_REF:
339
      temp = gen_rtx_LABEL_REF (to_mode, label_ref_label (x));
340 341
      LABEL_REF_NONLOCAL_P (temp) = LABEL_REF_NONLOCAL_P (x);
      return temp;
342

343
    case SYMBOL_REF:
344 345
      temp = shallow_copy_rtx (x);
      PUT_MODE (temp, to_mode);
346
      return temp;
347

348
    case CONST:
349 350 351
      temp = convert_memory_address_addr_space_1 (to_mode, XEXP (x, 0), as,
						  true, no_emit);
      return temp ? gen_rtx_CONST (to_mode, temp) : temp;
352

353 354
    case PLUS:
    case MULT:
355 356 357 358
      /* For addition we can safely permute the conversion and addition
	 operation if one operand is a constant and converting the constant
	 does not change it or if one operand is a constant and we are
	 using a ptr_extend instruction  (POINTERS_EXTEND_UNSIGNED < 0).
359
	 We can always safely permute them if we are making the address
360 361
	 narrower. Inside a CONST RTL, this is safe for both pointers
	 zero or sign extended as pointers cannot wrap. */
362 363
      if (GET_MODE_SIZE (to_mode) < GET_MODE_SIZE (from_mode)
	  || (GET_CODE (x) == PLUS
Shujing Zhao committed
364
	      && CONST_INT_P (XEXP (x, 1))
365 366
	      && ((in_const && POINTERS_EXTEND_UNSIGNED != 0)
		  || XEXP (x, 1) == convert_memory_address_addr_space_1
367 368
				     (to_mode, XEXP (x, 1), as, in_const,
				      no_emit)
369
                  || POINTERS_EXTEND_UNSIGNED < 0)))
370 371 372
	{
	  temp = convert_memory_address_addr_space_1 (to_mode, XEXP (x, 0),
						      as, in_const, no_emit);
373 374 375
	  return (temp ? gen_rtx_fmt_ee (GET_CODE (x), to_mode,
					 temp, XEXP (x, 1))
		       : temp);
376
	}
377
      break;
378

379 380
    default:
      break;
381
    }
382

383 384 385
  if (no_emit)
    return NULL_RTX;

386 387
  return convert_modes (to_mode, from_mode,
			x, POINTERS_EXTEND_UNSIGNED);
388
#endif /* defined(POINTERS_EXTEND_UNSIGNED) */
389
}
390 391 392 393 394 395 396 397

/* Given X, a memory address in address space AS' pointer mode, convert it to
   an address in the address space's address mode, or vice versa (TO_MODE says
   which way).  We take advantage of the fact that pointers are not allowed to
   overflow by commuting arithmetic operations over conversions so that address
   arithmetic insns can be used.  */

rtx
398 399
convert_memory_address_addr_space (scalar_int_mode to_mode, rtx x,
				   addr_space_t as)
400
{
401
  return convert_memory_address_addr_space_1 (to_mode, x, as, false, false);
402
}
Richard Kenner committed
403

404

405 406 407
/* Return something equivalent to X but valid as a memory address for something
   of mode MODE in the named address space AS.  When X is not itself valid,
   this works by copying X or subexpressions of it into registers.  */
Richard Kenner committed
408 409

rtx
410
memory_address_addr_space (machine_mode mode, rtx x, addr_space_t as)
Richard Kenner committed
411
{
412
  rtx oldx = x;
413
  scalar_int_mode address_mode = targetm.addr_space.address_mode (as);
Richard Kenner committed
414

415
  x = convert_memory_address_addr_space (address_mode, x, as);
416

417
  /* By passing constant addresses through registers
Richard Kenner committed
418
     we get a chance to cse them.  */
419
  if (! cse_not_expected && CONSTANT_P (x) && CONSTANT_ADDRESS_P (x))
420
    x = force_reg (address_mode, x);
Richard Kenner committed
421 422 423 424 425

  /* We get better cse by rejecting indirect addressing at this stage.
     Let the combiner create indirect addresses where appropriate.
     For now, generate the code so that the subexpressions useful to share
     are visible.  But not if cse won't be done!  */
426
  else
Richard Kenner committed
427
    {
428
      if (! cse_not_expected && !REG_P (x))
429 430 431
	x = break_out_memory_refs (x);

      /* At this point, any valid address is accepted.  */
432
      if (memory_address_addr_space_p (mode, x, as))
433
	goto done;
434 435 436

      /* If it was valid before but breaking out memory refs invalidated it,
	 use it the old way.  */
437
      if (memory_address_addr_space_p (mode, oldx, as))
438 439 440 441
	{
	  x = oldx;
	  goto done;
	}
442 443 444 445 446

      /* Perform machine-dependent transformations on X
	 in certain cases.  This is not necessary since the code
	 below can handle all possible cases, but machine-dependent
	 transformations can make better code.  */
447
      {
448 449 450
	rtx orig_x = x;
	x = targetm.addr_space.legitimize_address (x, oldx, mode, as);
	if (orig_x != x && memory_address_addr_space_p (mode, x, as))
451 452
	  goto done;
      }
453 454 455 456 457 458 459 460 461 462 463 464 465 466 467

      /* PLUS and MULT can appear in special ways
	 as the result of attempts to make an address usable for indexing.
	 Usually they are dealt with by calling force_operand, below.
	 But a sum containing constant terms is special
	 if removing them makes the sum a valid address:
	 then we generate that address in a register
	 and index off of it.  We do this because it often makes
	 shorter code, and because the addresses thus generated
	 in registers often become common subexpressions.  */
      if (GET_CODE (x) == PLUS)
	{
	  rtx constant_term = const0_rtx;
	  rtx y = eliminate_constant_term (x, &constant_term);
	  if (constant_term == const0_rtx
468
	      || ! memory_address_addr_space_p (mode, y, as))
469 470 471
	    x = force_operand (x, NULL_RTX);
	  else
	    {
472
	      y = gen_rtx_PLUS (GET_MODE (x), copy_to_reg (y), constant_term);
473
	      if (! memory_address_addr_space_p (mode, y, as))
474 475 476 477 478
		x = force_operand (x, NULL_RTX);
	      else
		x = y;
	    }
	}
Richard Kenner committed
479

480
      else if (GET_CODE (x) == MULT || GET_CODE (x) == MINUS)
481
	x = force_operand (x, NULL_RTX);
Richard Kenner committed
482

483 484
      /* If we have a register that's an invalid address,
	 it must be a hard reg of the wrong class.  Copy it to a pseudo.  */
485
      else if (REG_P (x))
486 487 488 489 490
	x = copy_to_reg (x);

      /* Last resort: copy the value to a register, since
	 the register is a valid address.  */
      else
491
	x = force_reg (address_mode, x);
Richard Kenner committed
492
    }
493 494 495

 done:

496
  gcc_assert (memory_address_addr_space_p (mode, x, as));
497 498 499 500
  /* If we didn't change the address, we are done.  Otherwise, mark
     a reg as a pointer if we have REG or REG + CONST_INT.  */
  if (oldx == x)
    return x;
501
  else if (REG_P (x))
502
    mark_reg_pointer (x, BITS_PER_UNIT);
503
  else if (GET_CODE (x) == PLUS
504
	   && REG_P (XEXP (x, 0))
Shujing Zhao committed
505
	   && CONST_INT_P (XEXP (x, 1)))
506
    mark_reg_pointer (XEXP (x, 0), BITS_PER_UNIT);
507

508 509 510 511
  /* OLDX may have been the address on a temporary.  Update the address
     to indicate that X is now used.  */
  update_temp_slot_address (oldx, x);

Richard Kenner committed
512 513 514
  return x;
}

515 516
/* Convert a mem ref into one with a valid memory address.
   Pass through anything else unchanged.  */
Richard Kenner committed
517 518

rtx
519
validize_mem (rtx ref)
Richard Kenner committed
520
{
521
  if (!MEM_P (ref))
Richard Kenner committed
522
    return ref;
523
  ref = use_anchored_address (ref);
524 525
  if (memory_address_addr_space_p (GET_MODE (ref), XEXP (ref, 0),
				   MEM_ADDR_SPACE (ref)))
Richard Kenner committed
526
    return ref;
527

528 529
  /* Don't alter REF itself, since that is probably a stack slot.  */
  return replace_equiv_address (ref, XEXP (ref, 0));
Richard Kenner committed
530
}
531 532 533 534 535 536 537 538 539 540

/* If X is a memory reference to a member of an object block, try rewriting
   it to use an anchor instead.  Return the new memory reference on success
   and the old one on failure.  */

rtx
use_anchored_address (rtx x)
{
  rtx base;
  HOST_WIDE_INT offset;
541
  machine_mode mode;
542 543 544 545 546 547 548 549 550 551 552 553

  if (!flag_section_anchors)
    return x;

  if (!MEM_P (x))
    return x;

  /* Split the address into a base and offset.  */
  base = XEXP (x, 0);
  offset = 0;
  if (GET_CODE (base) == CONST
      && GET_CODE (XEXP (base, 0)) == PLUS
Shujing Zhao committed
554
      && CONST_INT_P (XEXP (XEXP (base, 0), 1)))
555 556 557 558 559 560 561
    {
      offset += INTVAL (XEXP (XEXP (base, 0), 1));
      base = XEXP (XEXP (base, 0), 0);
    }

  /* Check whether BASE is suitable for anchors.  */
  if (GET_CODE (base) != SYMBOL_REF
562
      || !SYMBOL_REF_HAS_BLOCK_INFO_P (base)
563
      || SYMBOL_REF_ANCHOR_P (base)
564
      || SYMBOL_REF_BLOCK (base) == NULL
565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581
      || !targetm.use_anchors_for_symbol_p (base))
    return x;

  /* Decide where BASE is going to be.  */
  place_block_symbol (base);

  /* Get the anchor we need to use.  */
  offset += SYMBOL_REF_BLOCK_OFFSET (base);
  base = get_section_anchor (SYMBOL_REF_BLOCK (base), offset,
			     SYMBOL_REF_TLS_MODEL (base));

  /* Work out the offset from the anchor.  */
  offset -= SYMBOL_REF_BLOCK_OFFSET (base);

  /* If we're going to run a CSE pass, force the anchor into a register.
     We will then be able to reuse registers for several accesses, if the
     target costs say that that's worthwhile.  */
582
  mode = GET_MODE (base);
583
  if (!cse_not_expected)
584
    base = force_reg (mode, base);
585

586
  return replace_equiv_address (x, plus_constant (mode, base, offset));
587
}
Richard Kenner committed
588 589 590 591

/* Copy the value or contents of X to a new temp reg and return that reg.  */

rtx
592
copy_to_reg (rtx x)
Richard Kenner committed
593
{
594
  rtx temp = gen_reg_rtx (GET_MODE (x));
595

Richard Kenner committed
596
  /* If not an operand, must be an address with PLUS and MULT so
597
     do the computation.  */
Richard Kenner committed
598 599
  if (! general_operand (x, VOIDmode))
    x = force_operand (x, temp);
600

Richard Kenner committed
601 602 603 604 605 606 607 608 609 610
  if (x != temp)
    emit_move_insn (temp, x);

  return temp;
}

/* Like copy_to_reg but always give the new register mode Pmode
   in case X is a constant.  */

rtx
611
copy_addr_to_reg (rtx x)
Richard Kenner committed
612 613 614 615 616 617 618 619
{
  return copy_to_mode_reg (Pmode, x);
}

/* Like copy_to_reg but always give the new register mode MODE
   in case X is a constant.  */

rtx
620
copy_to_mode_reg (machine_mode mode, rtx x)
Richard Kenner committed
621
{
622
  rtx temp = gen_reg_rtx (mode);
623

Richard Kenner committed
624
  /* If not an operand, must be an address with PLUS and MULT so
625
     do the computation.  */
Richard Kenner committed
626 627 628
  if (! general_operand (x, VOIDmode))
    x = force_operand (x, temp);

629
  gcc_assert (GET_MODE (x) == mode || GET_MODE (x) == VOIDmode);
Richard Kenner committed
630 631 632 633 634 635 636 637 638 639 640 641 642 643
  if (x != temp)
    emit_move_insn (temp, x);
  return temp;
}

/* Load X into a register if it is not already one.
   Use mode MODE for the register.
   X should be valid for mode MODE, but it may be a constant which
   is valid for all integer modes; that's why caller must specify MODE.

   The caller must not alter the value in the register we return,
   since we mark it as a "constant" register.  */

rtx
644
force_reg (machine_mode mode, rtx x)
Richard Kenner committed
645
{
646 647
  rtx temp, set;
  rtx_insn *insn;
Richard Kenner committed
648

649
  if (REG_P (x))
Richard Kenner committed
650
    return x;
651

652 653 654 655 656 657 658 659
  if (general_operand (x, mode))
    {
      temp = gen_reg_rtx (mode);
      insn = emit_move_insn (temp, x);
    }
  else
    {
      temp = force_operand (x, NULL_RTX);
660
      if (REG_P (temp))
661 662 663 664 665 666 667 668
	insn = get_last_insn ();
      else
	{
	  rtx temp2 = gen_reg_rtx (mode);
	  insn = emit_move_insn (temp2, temp);
	  temp = temp2;
	}
    }
669

Richard Kenner committed
670
  /* Let optimizers know that TEMP's value never changes
671 672 673 674
     and that X can be substituted for it.  Don't get confused
     if INSN set something else (such as a SUBREG of TEMP).  */
  if (CONSTANT_P (x)
      && (set = single_set (insn)) != 0
675 676
      && SET_DEST (set) == temp
      && ! rtx_equal_p (x, SET_SRC (set)))
677
    set_unique_reg_note (insn, REG_EQUAL, x);
678

679 680 681 682 683 684 685 686 687 688 689 690 691 692 693
  /* Let optimizers know that TEMP is a pointer, and if so, the
     known alignment of that pointer.  */
  {
    unsigned align = 0;
    if (GET_CODE (x) == SYMBOL_REF)
      {
        align = BITS_PER_UNIT;
	if (SYMBOL_REF_DECL (x) && DECL_P (SYMBOL_REF_DECL (x)))
	  align = DECL_ALIGN (SYMBOL_REF_DECL (x));
      }
    else if (GET_CODE (x) == LABEL_REF)
      align = BITS_PER_UNIT;
    else if (GET_CODE (x) == CONST
	     && GET_CODE (XEXP (x, 0)) == PLUS
	     && GET_CODE (XEXP (XEXP (x, 0), 0)) == SYMBOL_REF
Shujing Zhao committed
694
	     && CONST_INT_P (XEXP (XEXP (x, 0), 1)))
695 696 697 698 699 700 701 702 703
      {
	rtx s = XEXP (XEXP (x, 0), 0);
	rtx c = XEXP (XEXP (x, 0), 1);
	unsigned sa, ca;

	sa = BITS_PER_UNIT;
	if (SYMBOL_REF_DECL (s) && DECL_P (SYMBOL_REF_DECL (s)))
	  sa = DECL_ALIGN (SYMBOL_REF_DECL (s));

704 705 706 707 708 709 710
	if (INTVAL (c) == 0)
	  align = sa;
	else
	  {
	    ca = ctz_hwi (INTVAL (c)) * BITS_PER_UNIT;
	    align = MIN (sa, ca);
	  }
711 712
      }

713
    if (align || (MEM_P (x) && MEM_POINTER (x)))
714 715 716
      mark_reg_pointer (temp, align);
  }

Richard Kenner committed
717 718 719 720 721 722 723
  return temp;
}

/* If X is a memory ref, copy its contents to a new temp reg and return
   that reg.  Otherwise, return X.  */

rtx
724
force_not_mem (rtx x)
Richard Kenner committed
725
{
726
  rtx temp;
Aldy Hernandez committed
727

728
  if (!MEM_P (x) || GET_MODE (x) == BLKmode)
Richard Kenner committed
729
    return x;
Aldy Hernandez committed
730

Richard Kenner committed
731
  temp = gen_reg_rtx (GET_MODE (x));
732 733 734 735

  if (MEM_POINTER (x))
    REG_POINTER (temp) = 1;

Richard Kenner committed
736 737 738 739 740 741 742 743 744
  emit_move_insn (temp, x);
  return temp;
}

/* Copy X to TARGET (if it's nonzero and a reg)
   or to a new temp reg and return that reg.
   MODE is the mode to use for X in case it is a constant.  */

rtx
745
copy_to_suggested_reg (rtx x, rtx target, machine_mode mode)
Richard Kenner committed
746
{
747
  rtx temp;
Richard Kenner committed
748

749
  if (target && REG_P (target))
Richard Kenner committed
750 751 752 753 754 755 756 757
    temp = target;
  else
    temp = gen_reg_rtx (mode);

  emit_move_insn (temp, x);
  return temp;
}

758
/* Return the mode to use to pass or return a scalar of TYPE and MODE.
759 760 761
   PUNSIGNEDP points to the signedness of the type and may be adjusted
   to show what signedness to use on extension operations.

762 763
   FOR_RETURN is nonzero if the caller is promoting the return value
   of FNDECL, else it is for promoting args.  */
764

765 766
machine_mode
promote_function_mode (const_tree type, machine_mode mode, int *punsignedp,
767 768
		       const_tree funtype, int for_return)
{
769 770 771 772 773 774 775 776 777 778 779
  /* Called without a type node for a libcall.  */
  if (type == NULL_TREE)
    {
      if (INTEGRAL_MODE_P (mode))
	return targetm.calls.promote_function_mode (NULL_TREE, mode,
						    punsignedp, funtype,
						    for_return);
      else
	return mode;
    }

780 781 782 783 784 785 786 787 788 789 790 791 792 793 794
  switch (TREE_CODE (type))
    {
    case INTEGER_TYPE:   case ENUMERAL_TYPE:   case BOOLEAN_TYPE:
    case REAL_TYPE:      case OFFSET_TYPE:     case FIXED_POINT_TYPE:
    case POINTER_TYPE:   case REFERENCE_TYPE:
      return targetm.calls.promote_function_mode (type, mode, punsignedp, funtype,
						  for_return);

    default:
      return mode;
    }
}
/* Return the mode to use to store a scalar of TYPE and MODE.
   PUNSIGNEDP points to the signedness of the type and may be adjusted
   to show what signedness to use on extension operations.  */
795

796 797
machine_mode
promote_mode (const_tree type ATTRIBUTE_UNUSED, machine_mode mode,
798
	      int *punsignedp ATTRIBUTE_UNUSED)
799
{
800 801 802
#ifdef PROMOTE_MODE
  enum tree_code code;
  int unsignedp;
803
  scalar_mode smode;
804 805
#endif

806 807 808 809 810 811
  /* For libcalls this is invoked without TYPE from the backends
     TARGET_PROMOTE_FUNCTION_MODE hooks.  Don't do anything in that
     case.  */
  if (type == NULL_TREE)
    return mode;

812 813 814 815
  /* FIXME: this is the same logic that was there until GCC 4.4, but we
     probably want to test POINTERS_EXTEND_UNSIGNED even if PROMOTE_MODE
     is not defined.  The affected targets are M32C, S390, SPARC.  */
#ifdef PROMOTE_MODE
816 817
  code = TREE_CODE (type);
  unsignedp = *punsignedp;
818 819 820 821

  switch (code)
    {
    case INTEGER_TYPE:   case ENUMERAL_TYPE:   case BOOLEAN_TYPE:
822
    case REAL_TYPE:      case OFFSET_TYPE:     case FIXED_POINT_TYPE:
823 824 825
      /* Values of these types always have scalar mode.  */
      smode = as_a <scalar_mode> (mode);
      PROMOTE_MODE (smode, unsignedp, type);
826
      *punsignedp = unsignedp;
827
      return smode;
828

829
#ifdef POINTERS_EXTEND_UNSIGNED
830
    case REFERENCE_TYPE:
831
    case POINTER_TYPE:
832
      *punsignedp = POINTERS_EXTEND_UNSIGNED;
833 834
      return targetm.addr_space.address_mode
	       (TYPE_ADDR_SPACE (TREE_TYPE (type)));
835
#endif
836

837
    default:
838
      return mode;
839
    }
840
#else
841
  return mode;
842
#endif
843
}
844 845 846 847 848 849


/* Use one of promote_mode or promote_function_mode to find the promoted
   mode of DECL.  If PUNSIGNEDP is not NULL, store there the unsignedness
   of DECL after promotion.  */

850
machine_mode
851 852 853 854
promote_decl_mode (const_tree decl, int *punsignedp)
{
  tree type = TREE_TYPE (decl);
  int unsignedp = TYPE_UNSIGNED (type);
855 856
  machine_mode mode = DECL_MODE (decl);
  machine_mode pmode;
857

858 859 860 861
  if (TREE_CODE (decl) == RESULT_DECL && !DECL_BY_REFERENCE (decl))
    pmode = promote_function_mode (type, mode, &unsignedp,
                                   TREE_TYPE (current_function_decl), 1);
  else if (TREE_CODE (decl) == RESULT_DECL || TREE_CODE (decl) == PARM_DECL)
862
    pmode = promote_function_mode (type, mode, &unsignedp,
863
                                   TREE_TYPE (current_function_decl), 2);
864 865 866 867 868 869 870 871
  else
    pmode = promote_mode (type, mode, &unsignedp);

  if (punsignedp)
    *punsignedp = unsignedp;
  return pmode;
}

872 873 874 875 876 877 878 879 880 881 882 883 884 885 886
/* Return the promoted mode for name.  If it is a named SSA_NAME, it
   is the same as promote_decl_mode.  Otherwise, it is the promoted
   mode of a temp decl of same type as the SSA_NAME, if we had created
   one.  */

machine_mode
promote_ssa_mode (const_tree name, int *punsignedp)
{
  gcc_assert (TREE_CODE (name) == SSA_NAME);

  /* Partitions holding parms and results must be promoted as expected
     by function.c.  */
  if (SSA_NAME_VAR (name)
      && (TREE_CODE (SSA_NAME_VAR (name)) == PARM_DECL
	  || TREE_CODE (SSA_NAME_VAR (name)) == RESULT_DECL))
887 888 889 890 891
    {
      machine_mode mode = promote_decl_mode (SSA_NAME_VAR (name), punsignedp);
      if (mode != BLKmode)
	return mode;
    }
892 893 894

  tree type = TREE_TYPE (name);
  int unsignedp = TYPE_UNSIGNED (type);
895
  machine_mode pmode = promote_mode (type, TYPE_MODE (type), &unsignedp);
896 897 898 899 900 901 902
  if (punsignedp)
    *punsignedp = unsignedp;

  return pmode;
}


903

904
/* Controls the behavior of {anti_,}adjust_stack.  */
905 906 907 908 909 910 911
static bool suppress_reg_args_size;

/* A helper for adjust_stack and anti_adjust_stack.  */

static void
adjust_stack_1 (rtx adjust, bool anti_p)
{
912 913
  rtx temp;
  rtx_insn *insn;
914 915

  /* Hereafter anti_p means subtract_p.  */
916 917
  if (!STACK_GROWS_DOWNWARD)
    anti_p = !anti_p;
918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933

  temp = expand_binop (Pmode,
		       anti_p ? sub_optab : add_optab,
		       stack_pointer_rtx, adjust, stack_pointer_rtx, 0,
		       OPTAB_LIB_WIDEN);

  if (temp != stack_pointer_rtx)
    insn = emit_move_insn (stack_pointer_rtx, temp);
  else
    {
      insn = get_last_insn ();
      temp = single_set (insn);
      gcc_assert (temp != NULL && SET_DEST (temp) == stack_pointer_rtx);
    }

  if (!suppress_reg_args_size)
934
    add_args_size_note (insn, stack_pointer_delta);
935 936
}

Richard Kenner committed
937 938 939 940
/* Adjust the stack pointer by ADJUST (an rtx for a number of bytes).
   This pops when ADJUST is positive.  ADJUST need not be constant.  */

void
941
adjust_stack (rtx adjust)
Richard Kenner committed
942 943 944 945
{
  if (adjust == const0_rtx)
    return;

946 947
  /* We expect all variable sized adjustments to be multiple of
     PREFERRED_STACK_BOUNDARY.  */
948 949 950
  poly_int64 const_adjust;
  if (poly_int_rtx_p (adjust, &const_adjust))
    stack_pointer_delta -= const_adjust;
951

952
  adjust_stack_1 (adjust, false);
Richard Kenner committed
953 954 955 956 957 958
}

/* Adjust the stack pointer by minus ADJUST (an rtx for a number of bytes).
   This pushes when ADJUST is positive.  ADJUST need not be constant.  */

void
959
anti_adjust_stack (rtx adjust)
Richard Kenner committed
960 961 962 963
{
  if (adjust == const0_rtx)
    return;

964 965
  /* We expect all variable sized adjustments to be multiple of
     PREFERRED_STACK_BOUNDARY.  */
966 967 968
  poly_int64 const_adjust;
  if (poly_int_rtx_p (adjust, &const_adjust))
    stack_pointer_delta += const_adjust;
969

970
  adjust_stack_1 (adjust, true);
Richard Kenner committed
971 972 973 974 975
}

/* Round the size of a block to be pushed up to the boundary required
   by this machine.  SIZE is the desired size, which need not be constant.  */

976
static rtx
977
round_push (rtx size)
Richard Kenner committed
978
{
979
  rtx align_rtx, alignm1_rtx;
980

981 982
  if (!SUPPORTS_STACK_ALIGNMENT
      || crtl->preferred_stack_boundary == MAX_SUPPORTED_STACK_ALIGNMENT)
Richard Kenner committed
983
    {
984 985 986 987 988 989 990 991
      int align = crtl->preferred_stack_boundary / BITS_PER_UNIT;

      if (align == 1)
	return size;

      if (CONST_INT_P (size))
	{
	  HOST_WIDE_INT new_size = (INTVAL (size) + align - 1) / align * align;
992

993 994 995 996 997 998 999
	  if (INTVAL (size) != new_size)
	    size = GEN_INT (new_size);
	  return size;
	}

      align_rtx = GEN_INT (align);
      alignm1_rtx = GEN_INT (align - 1);
Richard Kenner committed
1000 1001 1002
    }
  else
    {
1003 1004 1005 1006 1007
      /* If crtl->preferred_stack_boundary might still grow, use
	 virtual_preferred_stack_boundary_rtx instead.  This will be
	 substituted by the right value in vregs pass and optimized
	 during combine.  */
      align_rtx = virtual_preferred_stack_boundary_rtx;
1008 1009
      alignm1_rtx = force_operand (plus_constant (Pmode, align_rtx, -1),
				   NULL_RTX);
Richard Kenner committed
1010
    }
1011

1012 1013 1014 1015 1016 1017 1018 1019 1020
  /* CEIL_DIV_EXPR needs to worry about the addition overflowing,
     but we know it can't.  So add ourselves and then do
     TRUNC_DIV_EXPR.  */
  size = expand_binop (Pmode, add_optab, size, alignm1_rtx,
		       NULL_RTX, 1, OPTAB_LIB_WIDEN);
  size = expand_divmod (0, TRUNC_DIV_EXPR, Pmode, size, align_rtx,
			NULL_RTX, 1);
  size = expand_mult (Pmode, size, align_rtx, NULL_RTX, 1);

Richard Kenner committed
1021 1022 1023
  return size;
}

1024 1025 1026
/* Save the stack pointer for the purpose in SAVE_LEVEL.  PSAVE is a pointer
   to a previously-created save area.  If no save area has been allocated,
   this function will allocate one.  If a save area is specified, it
1027
   must be of the proper mode.  */
1028 1029

void
1030
emit_stack_save (enum save_level save_level, rtx *psave)
1031 1032 1033
{
  rtx sa = *psave;
  /* The default is that we use a move insn and save in a Pmode object.  */
Richard Sandiford committed
1034
  rtx_insn *(*fcn) (rtx, rtx) = gen_move_insn;
1035
  machine_mode mode = STACK_SAVEAREA_MODE (save_level);
1036 1037 1038 1039 1040

  /* See if this machine has anything special to do for this kind of save.  */
  switch (save_level)
    {
    case SAVE_BLOCK:
Richard Sandiford committed
1041 1042
      if (targetm.have_save_stack_block ())
	fcn = targetm.gen_save_stack_block;
1043 1044
      break;
    case SAVE_FUNCTION:
Richard Sandiford committed
1045 1046
      if (targetm.have_save_stack_function ())
	fcn = targetm.gen_save_stack_function;
1047 1048
      break;
    case SAVE_NONLOCAL:
Richard Sandiford committed
1049 1050
      if (targetm.have_save_stack_nonlocal ())
	fcn = targetm.gen_save_stack_nonlocal;
1051
      break;
1052 1053
    default:
      break;
1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069
    }

  /* If there is no save area and we have to allocate one, do so.  Otherwise
     verify the save area is the proper mode.  */

  if (sa == 0)
    {
      if (mode != VOIDmode)
	{
	  if (save_level == SAVE_NONLOCAL)
	    *psave = sa = assign_stack_local (mode, GET_MODE_SIZE (mode), 0);
	  else
	    *psave = sa = gen_reg_rtx (mode);
	}
    }

1070 1071 1072 1073
  do_pending_stack_adjust ();
  if (sa != 0)
    sa = validize_mem (sa);
  emit_insn (fcn (sa, stack_pointer_rtx));
1074 1075 1076
}

/* Restore the stack pointer for the purpose in SAVE_LEVEL.  SA is the save
1077
   area made by emit_stack_save.  If it is zero, we have nothing to do.  */
1078 1079

void
1080
emit_stack_restore (enum save_level save_level, rtx sa)
1081 1082
{
  /* The default is that we use a move insn.  */
Richard Sandiford committed
1083
  rtx_insn *(*fcn) (rtx, rtx) = gen_move_insn;
1084

1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098
  /* If stack_realign_drap, the x86 backend emits a prologue that aligns both
     STACK_POINTER and HARD_FRAME_POINTER.
     If stack_realign_fp, the x86 backend emits a prologue that aligns only
     STACK_POINTER. This renders the HARD_FRAME_POINTER unusable for accessing
     aligned variables, which is reflected in ix86_can_eliminate.
     We normally still have the realigned STACK_POINTER that we can use.
     But if there is a stack restore still present at reload, it can trigger 
     mark_not_eliminable for the STACK_POINTER, leaving no way to eliminate
     FRAME_POINTER into a hard reg.
     To prevent this situation, we force need_drap if we emit a stack
     restore.  */
  if (SUPPORTS_STACK_ALIGNMENT)
    crtl->need_drap = true;

1099 1100 1101 1102
  /* See if this machine has anything special to do for this kind of save.  */
  switch (save_level)
    {
    case SAVE_BLOCK:
Richard Sandiford committed
1103 1104
      if (targetm.have_restore_stack_block ())
	fcn = targetm.gen_restore_stack_block;
1105 1106
      break;
    case SAVE_FUNCTION:
Richard Sandiford committed
1107 1108
      if (targetm.have_restore_stack_function ())
	fcn = targetm.gen_restore_stack_function;
1109 1110
      break;
    case SAVE_NONLOCAL:
Richard Sandiford committed
1111 1112
      if (targetm.have_restore_stack_nonlocal ())
	fcn = targetm.gen_restore_stack_nonlocal;
1113
      break;
1114 1115
    default:
      break;
1116 1117
    }

1118
  if (sa != 0)
1119 1120 1121 1122
    {
      sa = validize_mem (sa);
      /* These clobbers prevent the scheduler from moving
	 references to variable arrays below the code
1123
	 that deletes (pops) the arrays.  */
1124 1125
      emit_clobber (gen_rtx_MEM (BLKmode, gen_rtx_SCRATCH (VOIDmode)));
      emit_clobber (gen_rtx_MEM (BLKmode, stack_pointer_rtx));
1126
    }
1127

1128 1129
  discard_pending_stack_adjust ();

1130
  emit_insn (fcn (stack_pointer_rtx, sa));
1131
}
1132 1133

/* Invoke emit_stack_save on the nonlocal_goto_save_area for the current
1134 1135
   function.  This should be called whenever we allocate or deallocate
   dynamic stack space.  */
1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146

void
update_nonlocal_goto_save_area (void)
{
  tree t_save;
  rtx r_save;

  /* The nonlocal_goto_save_area object is an array of N pointers.  The
     first one is used for the frame pointer save; the rest are sized by
     STACK_SAVEAREA_MODE.  Create a reference to array index 1, the first
     of the stack save area slots.  */
1147 1148 1149
  t_save = build4 (ARRAY_REF,
		   TREE_TYPE (TREE_TYPE (cfun->nonlocal_goto_save_area)),
		   cfun->nonlocal_goto_save_area,
1150
		   integer_one_node, NULL_TREE, NULL_TREE);
1151 1152
  r_save = expand_expr (t_save, NULL_RTX, VOIDmode, EXPAND_WRITE);

1153
  emit_stack_save (SAVE_NONLOCAL, &r_save);
1154
}
1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169

/* Record a new stack level for the current function.  This should be called
   whenever we allocate or deallocate dynamic stack space.  */

void
record_new_stack_level (void)
{
  /* Record the new stack level for nonlocal gotos.  */
  if (cfun->nonlocal_goto_save_area)
    update_nonlocal_goto_save_area ();
 
  /* Record the new stack level for SJLJ exceptions.  */
  if (targetm_common.except_unwind_info (&global_options) == UI_SJLJ)
    update_sjlj_context ();
}
1170

1171
/* Return an rtx doing runtime alignment to REQUIRED_ALIGN on TARGET.  */
1172 1173

rtx
1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190
align_dynamic_address (rtx target, unsigned required_align)
{
  /* CEIL_DIV_EXPR needs to worry about the addition overflowing,
     but we know it can't.  So add ourselves and then do
     TRUNC_DIV_EXPR.  */
  target = expand_binop (Pmode, add_optab, target,
			 gen_int_mode (required_align / BITS_PER_UNIT - 1,
				       Pmode),
			 NULL_RTX, 1, OPTAB_LIB_WIDEN);
  target = expand_divmod (0, TRUNC_DIV_EXPR, Pmode, target,
			  gen_int_mode (required_align / BITS_PER_UNIT,
					Pmode),
			  NULL_RTX, 1);
  target = expand_mult (Pmode, target,
			gen_int_mode (required_align / BITS_PER_UNIT,
				      Pmode),
			NULL_RTX, 1);
Richard Kenner committed
1191

1192 1193
  return target;
}
Richard Kenner committed
1194

1195 1196 1197 1198
/* Return an rtx through *PSIZE, representing the size of an area of memory to
   be dynamically pushed on the stack.

   *PSIZE is an rtx representing the size of the area.
1199

1200
   SIZE_ALIGN is the alignment (in bits) that we know SIZE has.  This
1201
   parameter may be zero.  If so, a proper value will be extracted
1202 1203 1204 1205
   from SIZE if it is constant, otherwise BITS_PER_UNIT will be assumed.

   REQUIRED_ALIGN is the alignment (in bits) required for the region
   of memory.
Eric Botcazou committed
1206

1207 1208 1209 1210 1211 1212
   If PSTACK_USAGE_SIZE is not NULL it points to a value that is increased for
   the additional size returned.  */
void
get_dynamic_stack_size (rtx *psize, unsigned size_align,
			unsigned required_align,
			HOST_WIDE_INT *pstack_usage_size)
Richard Kenner committed
1213
{
1214
  rtx size = *psize;
Eric Botcazou committed
1215

Richard Kenner committed
1216 1217 1218 1219
  /* Ensure the size is in the proper mode.  */
  if (GET_MODE (size) != VOIDmode && GET_MODE (size) != Pmode)
    size = convert_to_mode (Pmode, size, 1);

1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235
  if (CONST_INT_P (size))
    {
      unsigned HOST_WIDE_INT lsb;

      lsb = INTVAL (size);
      lsb &= -lsb;

      /* Watch out for overflow truncating to "unsigned".  */
      if (lsb > UINT_MAX / BITS_PER_UNIT)
	size_align = 1u << (HOST_BITS_PER_INT - 1);
      else
	size_align = (unsigned)lsb * BITS_PER_UNIT;
    }
  else if (size_align < BITS_PER_UNIT)
    size_align = BITS_PER_UNIT;

Richard Henderson committed
1236 1237 1238 1239 1240 1241
  /* We can't attempt to minimize alignment necessary, because we don't
     know the final value of preferred_stack_boundary yet while executing
     this code.  */
  if (crtl->preferred_stack_boundary < PREFERRED_STACK_BOUNDARY)
    crtl->preferred_stack_boundary = PREFERRED_STACK_BOUNDARY;

Richard Kenner committed
1242
  /* We will need to ensure that the address we return is aligned to
1243 1244 1245 1246 1247
     REQUIRED_ALIGN.  At this point in the compilation, we don't always
     know the final value of the STACK_DYNAMIC_OFFSET used in function.c
     (it might depend on the size of the outgoing parameter lists, for
     example), so we must preventively align the value.  We leave space
     in SIZE for the hole that might result from the alignment operation.  */
1248

1249 1250 1251 1252
  unsigned known_align = REGNO_POINTER_ALIGN (VIRTUAL_STACK_DYNAMIC_REGNUM);
  if (known_align == 0)
    known_align = BITS_PER_UNIT;
  if (required_align > known_align)
1253
    {
1254
      unsigned extra = (required_align - known_align) / BITS_PER_UNIT;
1255 1256
      size = plus_constant (Pmode, size, extra);
      size = force_operand (size, NULL_RTX);
1257 1258
      if (size_align > known_align)
	size_align = known_align;
1259 1260 1261 1262

      if (flag_stack_usage_info && pstack_usage_size)
	*pstack_usage_size += extra;
    }
1263

Richard Kenner committed
1264
  /* Round the size to a multiple of the required stack alignment.
1265
     Since the stack is presumed to be rounded before this allocation,
Richard Kenner committed
1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276
     this will maintain the required alignment.

     If the stack grows downward, we could save an insn by subtracting
     SIZE from the stack pointer and then aligning the stack pointer.
     The problem with this is that the stack pointer may be unaligned
     between the execution of the subtraction and alignment insns and
     some machines do not allow this.  Even on those that do, some
     signal handlers malfunction if a signal should occur between those
     insns.  Since this is an extremely rare event, we have no reliable
     way of knowing which systems have this problem.  So we avoid even
     momentarily mis-aligning the stack.  */
1277
  if (size_align % MAX_SUPPORTED_STACK_ALIGNMENT != 0)
Eric Botcazou committed
1278 1279
    {
      size = round_push (size);
Richard Kenner committed
1280

1281
      if (flag_stack_usage_info && pstack_usage_size)
Eric Botcazou committed
1282
	{
1283
	  int align = crtl->preferred_stack_boundary / BITS_PER_UNIT;
1284 1285
	  *pstack_usage_size =
	    (*pstack_usage_size + align - 1) / align * align;
Eric Botcazou committed
1286 1287 1288
	}
    }

1289 1290 1291
  *psize = size;
}

1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314
/* Return the number of bytes to "protect" on the stack for -fstack-check.

   "protect" in the context of -fstack-check means how many bytes we
   should always ensure are available on the stack.  More importantly
   this is how many bytes are skipped when probing the stack.

   On some targets we want to reuse the -fstack-check prologue support
   to give a degree of protection against stack clashing style attacks.

   In that scenario we do not want to skip bytes before probing as that
   would render the stack clash protections useless.

   So we never use STACK_CHECK_PROTECT directly.  Instead we indirect though
   this helper which allows us to provide different values for
   -fstack-check and -fstack-clash-protection.  */
HOST_WIDE_INT
get_stack_check_protect (void)
{
  if (flag_stack_clash_protection)
    return 0;
 return STACK_CHECK_PROTECT;
}

1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328
/* Return an rtx representing the address of an area of memory dynamically
   pushed on the stack.

   Any required stack pointer alignment is preserved.

   SIZE is an rtx representing the size of the area.

   SIZE_ALIGN is the alignment (in bits) that we know SIZE has.  This
   parameter may be zero.  If so, a proper value will be extracted
   from SIZE if it is constant, otherwise BITS_PER_UNIT will be assumed.

   REQUIRED_ALIGN is the alignment (in bits) required for the region
   of memory.

1329 1330 1331
   MAX_SIZE is an upper bound for SIZE, if SIZE is not constant, or -1 if
   no such upper bound is known.

1332 1333 1334 1335 1336 1337 1338 1339 1340
   If CANNOT_ACCUMULATE is set to TRUE, the caller guarantees that the
   stack space allocated by the generated code cannot be added with itself
   in the course of the execution of the function.  It is always safe to
   pass FALSE here and the following criterion is sufficient in order to
   pass TRUE: every path in the CFG that starts at the allocation point and
   loops to it executes the associated deallocation code.  */

rtx
allocate_dynamic_stack_space (rtx size, unsigned size_align,
1341 1342 1343
			      unsigned required_align,
			      HOST_WIDE_INT max_size,
			      bool cannot_accumulate)
1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381
{
  HOST_WIDE_INT stack_usage_size = -1;
  rtx_code_label *final_label;
  rtx final_target, target;

  /* If we're asking for zero bytes, it doesn't matter what we point
     to since we can't dereference it.  But return a reasonable
     address anyway.  */
  if (size == const0_rtx)
    return virtual_stack_dynamic_rtx;

  /* Otherwise, show we're calling alloca or equivalent.  */
  cfun->calls_alloca = 1;

  /* If stack usage info is requested, look into the size we are passed.
     We need to do so this early to avoid the obfuscation that may be
     introduced later by the various alignment operations.  */
  if (flag_stack_usage_info)
    {
      if (CONST_INT_P (size))
	stack_usage_size = INTVAL (size);
      else if (REG_P (size))
        {
	  /* Look into the last emitted insn and see if we can deduce
	     something for the register.  */
	  rtx_insn *insn;
	  rtx set, note;
	  insn = get_last_insn ();
	  if ((set = single_set (insn)) && rtx_equal_p (SET_DEST (set), size))
	    {
	      if (CONST_INT_P (SET_SRC (set)))
		stack_usage_size = INTVAL (SET_SRC (set));
	      else if ((note = find_reg_equal_equiv_note (insn))
		       && CONST_INT_P (XEXP (note, 0)))
		stack_usage_size = INTVAL (XEXP (note, 0));
	    }
	}

1382 1383 1384 1385 1386 1387
      /* If the size is not constant, try the maximum size.  */
      if (stack_usage_size < 0)
	stack_usage_size = max_size;

      /* If the size is still not constant, we can't say anything.  */
      if (stack_usage_size < 0)
1388 1389 1390 1391 1392 1393 1394 1395
	{
	  current_function_has_unbounded_dynamic_stack_size = 1;
	  stack_usage_size = 0;
	}
    }

  get_dynamic_stack_size (&size, size_align, required_align, &stack_usage_size);

1396
  target = gen_reg_rtx (Pmode);
1397

Eric Botcazou committed
1398 1399
  /* The size is supposed to be fully adjusted at this point so record it
     if stack usage info is requested.  */
1400
  if (flag_stack_usage_info)
Eric Botcazou committed
1401 1402 1403 1404 1405 1406 1407 1408
    {
      current_function_dynamic_stack_size += stack_usage_size;

      /* ??? This is gross but the only safe stance in the absence
	 of stack usage oriented flow analysis.  */
      if (!cannot_accumulate)
	current_function_has_unbounded_dynamic_stack_size = 1;
    }
Richard Kenner committed
1409

1410 1411
  do_pending_stack_adjust ();

1412
  final_label = NULL;
1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423
  final_target = NULL_RTX;

  /* If we are splitting the stack, we need to ask the backend whether
     there is enough room on the current stack.  If there isn't, or if
     the backend doesn't know how to tell is, then we need to call a
     function to allocate memory in some other way.  This memory will
     be released when we release the current stack segment.  The
     effect is that stack allocation becomes less efficient, but at
     least it doesn't cause a stack overflow.  */
  if (flag_split_stack)
    {
1424 1425
      rtx_code_label *available_label;
      rtx ask, space, func;
1426

1427
      available_label = NULL;
1428

Richard Sandiford committed
1429
      if (targetm.have_split_stack_space_check ())
1430 1431 1432 1433 1434
	{
	  available_label = gen_label_rtx ();

	  /* This instruction will branch to AVAILABLE_LABEL if there
	     are SIZE bytes available on the stack.  */
Richard Sandiford committed
1435 1436
	  emit_insn (targetm.gen_split_stack_space_check
		     (size, available_label));
1437 1438
	}

1439
      /* The __morestack_allocate_stack_space function will allocate
1440 1441 1442 1443 1444 1445
	 memory using malloc.  If the alignment of the memory returned
	 by malloc does not meet REQUIRED_ALIGN, we increase SIZE to
	 make sure we allocate enough space.  */
      if (MALLOC_ABI_ALIGNMENT >= required_align)
	ask = size;
      else
1446 1447 1448 1449
	ask = expand_binop (Pmode, add_optab, size,
			    gen_int_mode (required_align / BITS_PER_UNIT - 1,
					  Pmode),
			    NULL_RTX, 1, OPTAB_LIB_WIDEN);
1450

1451 1452 1453
      func = init_one_libfunc ("__morestack_allocate_stack_space");

      space = emit_library_call_value (func, target, LCT_NORMAL, Pmode,
1454
				       ask, Pmode);
1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468

      if (available_label == NULL_RTX)
	return space;

      final_target = gen_reg_rtx (Pmode);

      emit_move_insn (final_target, space);

      final_label = gen_label_rtx ();
      emit_jump (final_label);

      emit_label (available_label);
    }

1469
 /* We ought to be called always on the toplevel and stack ought to be aligned
1470
    properly.  */
1471 1472
  gcc_assert (multiple_p (stack_pointer_delta,
			  PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT));
1473

1474 1475 1476 1477 1478
  /* If needed, check that we have the required amount of stack.  Take into
     account what has already been checked.  */
  if (STACK_CHECK_MOVING_SP)
    ;
  else if (flag_stack_check == GENERIC_STACK_CHECK)
1479 1480 1481
    probe_stack_range (STACK_OLD_CHECK_PROTECT + STACK_CHECK_MAX_FRAME_SIZE,
		       size);
  else if (flag_stack_check == STATIC_BUILTIN_STACK_CHECK)
1482
    probe_stack_range (get_stack_check_protect (), size);
1483

1484 1485 1486
  /* Don't let anti_adjust_stack emit notes.  */
  suppress_reg_args_size = true;

Richard Kenner committed
1487 1488
  /* Perform the required allocation from the stack.  Some systems do
     this differently than simply incrementing/decrementing from the
1489
     stack pointer, such as acquiring the space by calling malloc().  */
Richard Sandiford committed
1490
  if (targetm.have_allocate_stack ())
Richard Kenner committed
1491
    {
1492
      class expand_operand ops[2];
1493 1494
      /* We don't have to check against the predicate for operand 0 since
	 TARGET is known to be a pseudo of the proper mode, which must
1495 1496 1497
	 be valid for the operand.  */
      create_fixed_operand (&ops[0], target);
      create_convert_operand_to (&ops[1], size, STACK_SIZE_MODE, true);
Richard Sandiford committed
1498
      expand_insn (targetm.code_for_allocate_stack, 2, ops);
Richard Kenner committed
1499 1500
    }
  else
1501
    {
1502
      poly_int64 saved_stack_pointer_delta;
1503

1504 1505
      if (!STACK_GROWS_DOWNWARD)
	emit_move_insn (target, virtual_stack_dynamic_rtx);
1506 1507

      /* Check stack bounds if necessary.  */
1508
      if (crtl->limit_stack)
1509 1510
	{
	  rtx available;
1511
	  rtx_code_label *space_available = gen_label_rtx ();
1512 1513 1514 1515 1516 1517 1518 1519 1520
	  if (STACK_GROWS_DOWNWARD)
	    available = expand_binop (Pmode, sub_optab,
				      stack_pointer_rtx, stack_limit_rtx,
				      NULL_RTX, 1, OPTAB_WIDEN);
	  else
	    available = expand_binop (Pmode, sub_optab,
				      stack_limit_rtx, stack_pointer_rtx,
				      NULL_RTX, 1, OPTAB_WIDEN);

1521
	  emit_cmp_and_jump_insns (available, size, GEU, NULL_RTX, Pmode, 1,
1522
				   space_available);
1523 1524
	  if (targetm.have_trap ())
	    emit_insn (targetm.gen_trap ());
1525 1526 1527 1528 1529 1530
	  else
	    error ("stack limits not supported on this target");
	  emit_barrier ();
	  emit_label (space_available);
	}

1531
      saved_stack_pointer_delta = stack_pointer_delta;
1532

1533
      if (flag_stack_check && STACK_CHECK_MOVING_SP)
1534
	anti_adjust_stack_and_probe (size, false);
1535 1536
      else if (flag_stack_clash_protection)
	anti_adjust_stack_and_probe_stack_clash (size);
1537 1538
      else
	anti_adjust_stack (size);
1539

1540 1541 1542 1543
      /* Even if size is constant, don't modify stack_pointer_delta.
	 The constant size alloca should preserve
	 crtl->preferred_stack_boundary alignment.  */
      stack_pointer_delta = saved_stack_pointer_delta;
1544

1545 1546
      if (STACK_GROWS_DOWNWARD)
	emit_move_insn (target, virtual_stack_dynamic_rtx);
1547
    }
Richard Kenner committed
1548

1549 1550
  suppress_reg_args_size = false;

1551 1552 1553 1554 1555 1556 1557 1558 1559
  /* Finish up the split stack handling.  */
  if (final_label != NULL_RTX)
    {
      gcc_assert (flag_split_stack);
      emit_move_insn (final_target, target);
      emit_label (final_label);
      target = final_target;
    }

1560
  target = align_dynamic_address (target, required_align);
1561

1562 1563 1564
  /* Now that we've committed to a return value, mark its alignment.  */
  mark_reg_pointer (target, required_align);

1565 1566
  /* Record the new stack level.  */
  record_new_stack_level ();
1567

Richard Kenner committed
1568 1569
  return target;
}
1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582

/* Return an rtx representing the address of an area of memory already
   statically pushed onto the stack in the virtual stack vars area.  (It is
   assumed that the area is allocated in the function prologue.)

   Any required stack pointer alignment is preserved.

   OFFSET is the offset of the area into the virtual stack vars area.

   REQUIRED_ALIGN is the alignment (in bits) required for the region
   of memory.  */

rtx
1583
get_dynamic_stack_base (poly_int64 offset, unsigned required_align)
1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601
{
  rtx target;

  if (crtl->preferred_stack_boundary < PREFERRED_STACK_BOUNDARY)
    crtl->preferred_stack_boundary = PREFERRED_STACK_BOUNDARY;

  target = gen_reg_rtx (Pmode);
  emit_move_insn (target, virtual_stack_vars_rtx);
  target = expand_binop (Pmode, add_optab, target,
			 gen_int_mode (offset, Pmode),
			 NULL_RTX, 1, OPTAB_LIB_WIDEN);
  target = align_dynamic_address (target, required_align);

  /* Now that we've committed to a return value, mark its alignment.  */
  mark_reg_pointer (target, required_align);

  return target;
}
Richard Kenner committed
1602

1603
/* A front end may want to override GCC's stack checking by providing a
Richard Kenner committed
1604 1605 1606
   run-time routine to call to check the stack, so provide a mechanism for
   calling that routine.  */

1607
static GTY(()) rtx stack_check_libfunc;
Richard Kenner committed
1608 1609

void
1610
set_stack_check_libfunc (const char *libfunc_name)
Richard Kenner committed
1611
{
1612 1613
  gcc_assert (stack_check_libfunc == NULL_RTX);
  stack_check_libfunc = gen_rtx_SYMBOL_REF (Pmode, libfunc_name);
1614 1615 1616 1617
  tree decl = build_decl (UNKNOWN_LOCATION, FUNCTION_DECL,
			  get_identifier (libfunc_name), void_type_node);
  DECL_EXTERNAL (decl) = 1;
  SET_SYMBOL_REF_DECL (stack_check_libfunc, decl);
Richard Kenner committed
1618 1619
}

1620 1621
/* Emit one stack probe at ADDRESS, an address within the stack.  */

1622
void
1623
emit_stack_probe (rtx address)
1624
{
Richard Sandiford committed
1625
  if (targetm.have_probe_stack_address ())
1626
    {
1627
      class expand_operand ops[1];
1628 1629 1630 1631 1632
      insn_code icode = targetm.code_for_probe_stack_address;
      create_address_operand (ops, address);
      maybe_legitimize_operands (icode, 0, 1, ops);
      expand_insn (icode, 1, ops);
    }
1633 1634 1635
  else
    {
      rtx memref = gen_rtx_MEM (word_mode, address);
1636

1637
      MEM_VOLATILE_P (memref) = 1;
1638
      memref = validize_mem (memref);
1639

1640
      /* See if we have an insn to probe the stack.  */
Richard Sandiford committed
1641
      if (targetm.have_probe_stack ())
1642
	emit_insn (targetm.gen_probe_stack (memref));
1643
      else
1644
	emit_move_insn (memref, const0_rtx);
1645
    }
1646 1647
}

1648
/* Probe a range of stack addresses from FIRST to FIRST+SIZE, inclusive.
1649 1650 1651 1652 1653
   FIRST is a constant and size is a Pmode RTX.  These are offsets from
   the current stack pointer.  STACK_GROWS_DOWNWARD says whether to add
   or subtract them from the stack pointer.  */

#define PROBE_INTERVAL (1 << STACK_CHECK_PROBE_INTERVAL_EXP)
1654

1655
#if STACK_GROWS_DOWNWARD
1656
#define STACK_GROW_OP MINUS
1657 1658
#define STACK_GROW_OPTAB sub_optab
#define STACK_GROW_OFF(off) -(off)
1659 1660
#else
#define STACK_GROW_OP PLUS
1661 1662
#define STACK_GROW_OPTAB add_optab
#define STACK_GROW_OFF(off) (off)
1663 1664 1665
#endif

void
1666
probe_stack_range (HOST_WIDE_INT first, rtx size)
1667
{
1668 1669 1670 1671
  /* First ensure SIZE is Pmode.  */
  if (GET_MODE (size) != VOIDmode && GET_MODE (size) != Pmode)
    size = convert_to_mode (Pmode, size, 1);

1672 1673
  /* Next see if we have a function to check the stack.  */
  if (stack_check_libfunc)
1674
    {
1675
      rtx addr = memory_address (Pmode,
1676 1677
				 gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
					         stack_pointer_rtx,
1678 1679
					         plus_constant (Pmode,
								size, first)));
1680 1681
      emit_library_call (stack_check_libfunc, LCT_THROW, VOIDmode,
			 addr, Pmode);
1682
    }
Richard Kenner committed
1683

1684
  /* Next see if we have an insn to check the stack.  */
Richard Sandiford committed
1685
  else if (targetm.have_check_stack ())
1686
    {
1687
      class expand_operand ops[1];
1688 1689 1690
      rtx addr = memory_address (Pmode,
				 gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
					         stack_pointer_rtx,
1691 1692
					         plus_constant (Pmode,
								size, first)));
1693
      bool success;
1694
      create_input_operand (&ops[0], addr, Pmode);
Richard Sandiford committed
1695
      success = maybe_expand_insn (targetm.code_for_check_stack, 1, ops);
1696
      gcc_assert (success);
1697 1698
    }

1699 1700 1701
  /* Otherwise we have to generate explicit probes.  If we have a constant
     small number of them to generate, that's the easy case.  */
  else if (CONST_INT_P (size) && INTVAL (size) < 7 * PROBE_INTERVAL)
1702
    {
1703 1704 1705 1706 1707 1708 1709 1710 1711
      HOST_WIDE_INT isize = INTVAL (size), i;
      rtx addr;

      /* Probe at FIRST + N * PROBE_INTERVAL for values of N from 1 until
	 it exceeds SIZE.  If only one probe is needed, this will not
	 generate any code.  Then probe at FIRST + SIZE.  */
      for (i = PROBE_INTERVAL; i < isize; i += PROBE_INTERVAL)
	{
	  addr = memory_address (Pmode,
1712
				 plus_constant (Pmode, stack_pointer_rtx,
1713 1714 1715 1716 1717
				 		STACK_GROW_OFF (first + i)));
	  emit_stack_probe (addr);
	}

      addr = memory_address (Pmode,
1718
			     plus_constant (Pmode, stack_pointer_rtx,
1719 1720
					    STACK_GROW_OFF (first + isize)));
      emit_stack_probe (addr);
1721 1722
    }

1723 1724 1725 1726 1727
  /* In the variable case, do the same as above, but in a loop.  Note that we
     must be extra careful with variables wrapping around because we might be
     at the very top (or the very bottom) of the address space and we have to
     be able to handle this case properly; in particular, we use an equality
     test for the loop condition.  */
1728 1729
  else
    {
1730
      rtx rounded_size, rounded_size_op, test_addr, last_addr, temp;
1731 1732
      rtx_code_label *loop_lab = gen_label_rtx ();
      rtx_code_label *end_lab = gen_label_rtx ();
1733

1734 1735 1736 1737
      /* Step 1: round SIZE to the previous multiple of the interval.  */

      /* ROUNDED_SIZE = SIZE & -PROBE_INTERVAL  */
      rounded_size
Richard Sandiford committed
1738 1739
	= simplify_gen_binary (AND, Pmode, size,
			       gen_int_mode (-PROBE_INTERVAL, Pmode));
1740 1741 1742 1743 1744 1745 1746 1747
      rounded_size_op = force_operand (rounded_size, NULL_RTX);


      /* Step 2: compute initial and final value of the loop counter.  */

      /* TEST_ADDR = SP + FIRST.  */
      test_addr = force_operand (gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
					 	 stack_pointer_rtx,
1748 1749
						 gen_int_mode (first, Pmode)),
				 NULL_RTX);
1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766

      /* LAST_ADDR = SP + FIRST + ROUNDED_SIZE.  */
      last_addr = force_operand (gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
						 test_addr,
						 rounded_size_op), NULL_RTX);


      /* Step 3: the loop

	 while (TEST_ADDR != LAST_ADDR)
	   {
	     TEST_ADDR = TEST_ADDR + PROBE_INTERVAL
	     probe at TEST_ADDR
	   }

	 probes at FIRST + N * PROBE_INTERVAL for values of N from 1
	 until it is equal to ROUNDED_SIZE.  */
1767 1768 1769

      emit_label (loop_lab);

1770 1771 1772 1773 1774 1775
      /* Jump to END_LAB if TEST_ADDR == LAST_ADDR.  */
      emit_cmp_and_jump_insns (test_addr, last_addr, EQ, NULL_RTX, Pmode, 1,
			       end_lab);

      /* TEST_ADDR = TEST_ADDR + PROBE_INTERVAL.  */
      temp = expand_binop (Pmode, STACK_GROW_OPTAB, test_addr,
1776
			   gen_int_mode (PROBE_INTERVAL, Pmode), test_addr,
1777 1778
			   1, OPTAB_WIDEN);

1779
      gcc_assert (temp == test_addr);
1780

1781 1782 1783 1784 1785
      /* Probe at TEST_ADDR.  */
      emit_stack_probe (test_addr);

      emit_jump (loop_lab);

1786 1787
      emit_label (end_lab);

1788 1789 1790 1791 1792 1793 1794 1795 1796 1797

      /* Step 4: probe at FIRST + SIZE if we cannot assert at compile-time
	 that SIZE is equal to ROUNDED_SIZE.  */

      /* TEMP = SIZE - ROUNDED_SIZE.  */
      temp = simplify_gen_binary (MINUS, Pmode, size, rounded_size);
      if (temp != const0_rtx)
	{
	  rtx addr;

1798
	  if (CONST_INT_P (temp))
1799 1800 1801 1802
	    {
	      /* Use [base + disp} addressing mode if supported.  */
	      HOST_WIDE_INT offset = INTVAL (temp);
	      addr = memory_address (Pmode,
1803
				     plus_constant (Pmode, last_addr,
1804 1805 1806 1807 1808 1809 1810 1811 1812 1813 1814 1815 1816
						    STACK_GROW_OFF (offset)));
	    }
	  else
	    {
	      /* Manual CSE if the difference is not known at compile-time.  */
	      temp = gen_rtx_MINUS (Pmode, size, rounded_size_op);
	      addr = memory_address (Pmode,
				     gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
						     last_addr, temp));
	    }

	  emit_stack_probe (addr);
	}
1817
    }
1818 1819 1820

  /* Make sure nothing is scheduled before we are done.  */
  emit_insn (gen_blockage ());
1821
}
1822

1823 1824 1825 1826 1827 1828 1829 1830 1831 1832 1833 1834 1835 1836 1837 1838
/* Compute parameters for stack clash probing a dynamic stack
   allocation of SIZE bytes.

   We compute ROUNDED_SIZE, LAST_ADDR, RESIDUAL and PROBE_INTERVAL.

   Additionally we conditionally dump the type of probing that will
   be needed given the values computed.  */

void
compute_stack_clash_protection_loop_data (rtx *rounded_size, rtx *last_addr,
					  rtx *residual,
					  HOST_WIDE_INT *probe_interval,
					  rtx size)
{
  /* Round SIZE down to STACK_CLASH_PROTECTION_PROBE_INTERVAL */
  *probe_interval
1839
    = 1 << param_stack_clash_protection_probe_interval;
1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850 1851 1852 1853 1854 1855 1856 1857 1858 1859 1860
  *rounded_size = simplify_gen_binary (AND, Pmode, size,
				        GEN_INT (-*probe_interval));

  /* Compute the value of the stack pointer for the last iteration.
     It's just SP + ROUNDED_SIZE.  */
  rtx rounded_size_op = force_operand (*rounded_size, NULL_RTX);
  *last_addr = force_operand (gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
					      stack_pointer_rtx,
					      rounded_size_op),
			      NULL_RTX);

  /* Compute any residuals not allocated by the loop above.  Residuals
     are just the ROUNDED_SIZE - SIZE.  */
  *residual = simplify_gen_binary (MINUS, Pmode, size, *rounded_size);

  /* Dump key information to make writing tests easy.  */
  if (dump_file)
    {
      if (*rounded_size == CONST0_RTX (Pmode))
	fprintf (dump_file,
		 "Stack clash skipped dynamic allocation and probing loop.\n");
1861
      else if (CONST_INT_P (*rounded_size)
1862 1863 1864
	       && INTVAL (*rounded_size) <= 4 * *probe_interval)
	fprintf (dump_file,
		 "Stack clash dynamic allocation and probing inline.\n");
1865
      else if (CONST_INT_P (*rounded_size))
1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883 1884 1885 1886 1887 1888 1889 1890 1891 1892 1893 1894 1895 1896 1897 1898 1899 1900 1901 1902 1903 1904 1905 1906 1907 1908 1909 1910 1911 1912 1913 1914 1915 1916 1917 1918 1919 1920 1921 1922 1923 1924 1925 1926 1927 1928 1929 1930 1931 1932 1933 1934 1935 1936 1937 1938 1939 1940 1941 1942 1943 1944 1945 1946 1947 1948 1949 1950 1951 1952 1953 1954 1955 1956
	fprintf (dump_file,
		 "Stack clash dynamic allocation and probing in "
		 "rotated loop.\n");
      else
	fprintf (dump_file,
		 "Stack clash dynamic allocation and probing in loop.\n");

      if (*residual != CONST0_RTX (Pmode))
	fprintf (dump_file,
		 "Stack clash dynamic allocation and probing residuals.\n");
      else
	fprintf (dump_file,
		 "Stack clash skipped dynamic allocation and "
		 "probing residuals.\n");
    }
}

/* Emit the start of an allocate/probe loop for stack
   clash protection.

   LOOP_LAB and END_LAB are returned for use when we emit the
   end of the loop.

   LAST addr is the value for SP which stops the loop.  */
void
emit_stack_clash_protection_probe_loop_start (rtx *loop_lab,
					      rtx *end_lab,
					      rtx last_addr,
					      bool rotated)
{
  /* Essentially we want to emit any setup code, the top of loop
     label and the comparison at the top of the loop.  */
  *loop_lab = gen_label_rtx ();
  *end_lab = gen_label_rtx ();

  emit_label (*loop_lab);
  if (!rotated)
    emit_cmp_and_jump_insns (stack_pointer_rtx, last_addr, EQ, NULL_RTX,
			     Pmode, 1, *end_lab);
}

/* Emit the end of a stack clash probing loop.

   This consists of just the jump back to LOOP_LAB and
   emitting END_LOOP after the loop.  */

void
emit_stack_clash_protection_probe_loop_end (rtx loop_lab, rtx end_loop,
					    rtx last_addr, bool rotated)
{
  if (rotated)
    emit_cmp_and_jump_insns (stack_pointer_rtx, last_addr, NE, NULL_RTX,
			     Pmode, 1, loop_lab);
  else
    emit_jump (loop_lab);

  emit_label (end_loop);

}

/* Adjust the stack pointer by minus SIZE (an rtx for a number of bytes)
   while probing it.  This pushes when SIZE is positive.  SIZE need not
   be constant.

   This is subtly different than anti_adjust_stack_and_probe to try and
   prevent stack-clash attacks

     1. It must assume no knowledge of the probing state, any allocation
	must probe.

	Consider the case of a 1 byte alloca in a loop.  If the sum of the
	allocations is large, then this could be used to jump the guard if
	probes were not emitted.

     2. It never skips probes, whereas anti_adjust_stack_and_probe will
	skip probes on the first couple PROBE_INTERVALs on the assumption
	they're done elsewhere.

     3. It only allocates and probes SIZE bytes, it does not need to
	allocate/probe beyond that because this probing style does not
	guarantee signal handling capability if the guard is hit.  */

static void
anti_adjust_stack_and_probe_stack_clash (rtx size)
{
  /* First ensure SIZE is Pmode.  */
  if (GET_MODE (size) != VOIDmode && GET_MODE (size) != Pmode)
    size = convert_to_mode (Pmode, size, 1);

  /* We can get here with a constant size on some targets.  */
  rtx rounded_size, last_addr, residual;
1957 1958
  HOST_WIDE_INT probe_interval, probe_range;
  bool target_probe_range_p = false;
1959 1960 1961
  compute_stack_clash_protection_loop_data (&rounded_size, &last_addr,
					    &residual, &probe_interval, size);

1962 1963 1964 1965 1966 1967 1968 1969 1970 1971
  /* Get the back-end specific probe ranges.  */
  probe_range = targetm.stack_clash_protection_alloca_probe_range ();
  target_probe_range_p = probe_range != 0;
  gcc_assert (probe_range >= 0);

  /* If no back-end specific range defined, default to the top of the newly
     allocated range.  */
  if (probe_range == 0)
    probe_range = probe_interval - GET_MODE_SIZE (word_mode);

1972 1973
  if (rounded_size != CONST0_RTX (Pmode))
    {
1974 1975
      if (CONST_INT_P (rounded_size)
	  && INTVAL (rounded_size) <= 4 * probe_interval)
1976 1977 1978 1979 1980 1981 1982 1983 1984 1985
	{
	  for (HOST_WIDE_INT i = 0;
	       i < INTVAL (rounded_size);
	       i += probe_interval)
	    {
	      anti_adjust_stack (GEN_INT (probe_interval));
	      /* The prologue does not probe residuals.  Thus the offset
		 here to probe just beyond what the prologue had already
		 allocated.  */
	      emit_stack_probe (plus_constant (Pmode, stack_pointer_rtx,
1986 1987
					       probe_range));

1988 1989 1990 1991 1992 1993
	      emit_insn (gen_blockage ());
	    }
	}
      else
	{
	  rtx loop_lab, end_loop;
1994
	  bool rotate_loop = CONST_INT_P (rounded_size);
1995 1996 1997 1998 1999 2000
	  emit_stack_clash_protection_probe_loop_start (&loop_lab, &end_loop,
							last_addr, rotate_loop);

	  anti_adjust_stack (GEN_INT (probe_interval));

	  /* The prologue does not probe residuals.  Thus the offset here
2001 2002
	     to probe just beyond what the prologue had already
	     allocated.  */
2003
	  emit_stack_probe (plus_constant (Pmode, stack_pointer_rtx,
2004
					   probe_range));
2005 2006 2007 2008 2009 2010 2011 2012 2013

	  emit_stack_clash_protection_probe_loop_end (loop_lab, end_loop,
						      last_addr, rotate_loop);
	  emit_insn (gen_blockage ());
	}
    }

  if (residual != CONST0_RTX (Pmode))
    {
2014 2015 2016 2017 2018
      rtx label = NULL_RTX;
      /* RESIDUAL could be zero at runtime and in that case *sp could
	 hold live data.  Furthermore, we do not want to probe into the
	 red zone.

2019 2020 2021 2022 2023 2024 2025 2026 2027 2028
	 If TARGET_PROBE_RANGE_P then the target has promised it's safe to
	 probe at offset 0.  In which case we no longer have to check for
	 RESIDUAL == 0.  However we still need to probe at the right offset
	 when RESIDUAL > PROBE_RANGE, in which case we probe at PROBE_RANGE.

	 If !TARGET_PROBE_RANGE_P then go ahead and just guard the probe at *sp
	 on RESIDUAL != 0 at runtime if RESIDUAL is not a compile time constant.
	 */
      anti_adjust_stack (residual);

2029 2030 2031
      if (!CONST_INT_P (residual))
	{
	  label = gen_label_rtx ();
2032 2033 2034 2035
	  rtx_code op = target_probe_range_p ? LT : EQ;
	  rtx probe_cmp_value = target_probe_range_p
	    ? gen_rtx_CONST_INT (GET_MODE (residual), probe_range)
	    : CONST0_RTX (GET_MODE (residual));
2036

2037 2038
	  if (target_probe_range_p)
	    emit_stack_probe (stack_pointer_rtx);
2039

2040 2041 2042
	  emit_cmp_and_jump_insns (residual, probe_cmp_value,
				   op, NULL_RTX, Pmode, 1, label);
	}
2043

2044 2045 2046 2047 2048 2049 2050 2051 2052
      rtx x = NULL_RTX;

      /* If RESIDUAL isn't a constant and TARGET_PROBE_RANGE_P then we probe up
	 by the ABI defined safe value.  */
      if (!CONST_INT_P (residual) && target_probe_range_p)
	x = GEN_INT (probe_range);
      /* If RESIDUAL is a constant but smaller than the ABI defined safe value,
	 we still want to probe up, but the safest amount if a word.  */
      else if (target_probe_range_p)
2053
	{
2054 2055 2056 2057
	  if (INTVAL (residual) <= probe_range)
	    x = GEN_INT (GET_MODE_SIZE (word_mode));
	  else
	    x = GEN_INT (probe_range);
2058
	}
2059 2060 2061 2062 2063
      else
      /* If nothing else, probe at the top of the new allocation.  */
	x = plus_constant (Pmode, residual, -GET_MODE_SIZE (word_mode));

      emit_stack_probe (gen_rtx_PLUS (Pmode, stack_pointer_rtx, x));
2064 2065

      emit_insn (gen_blockage ());
2066 2067
      if (!CONST_INT_P (residual))
	  emit_label (label);
2068 2069 2070 2071
    }
}


2072 2073 2074 2075
/* Adjust the stack pointer by minus SIZE (an rtx for a number of bytes)
   while probing it.  This pushes when SIZE is positive.  SIZE need not
   be constant.  If ADJUST_BACK is true, adjust back the stack pointer
   by plus SIZE at the end.  */
2076

2077 2078
void
anti_adjust_stack_and_probe (rtx size, bool adjust_back)
2079
{
2080 2081 2082
  /* We skip the probe for the first interval + a small dope of 4 words and
     probe that many bytes past the specified size to maintain a protection
     area at the botton of the stack.  */
2083 2084 2085 2086 2087 2088 2089 2090
  const int dope = 4 * UNITS_PER_WORD;

  /* First ensure SIZE is Pmode.  */
  if (GET_MODE (size) != VOIDmode && GET_MODE (size) != Pmode)
    size = convert_to_mode (Pmode, size, 1);

  /* If we have a constant small number of probes to generate, that's the
     easy case.  */
2091
  if (CONST_INT_P (size) && INTVAL (size) < 7 * PROBE_INTERVAL)
2092 2093 2094 2095
    {
      HOST_WIDE_INT isize = INTVAL (size), i;
      bool first_probe = true;

2096
      /* Adjust SP and probe at PROBE_INTERVAL + N * PROBE_INTERVAL for
2097 2098 2099 2100 2101 2102 2103 2104 2105 2106 2107 2108 2109 2110 2111 2112
	 values of N from 1 until it exceeds SIZE.  If only one probe is
	 needed, this will not generate any code.  Then adjust and probe
	 to PROBE_INTERVAL + SIZE.  */
      for (i = PROBE_INTERVAL; i < isize; i += PROBE_INTERVAL)
	{
	  if (first_probe)
	    {
	      anti_adjust_stack (GEN_INT (2 * PROBE_INTERVAL + dope));
	      first_probe = false;
	    }
	  else
	    anti_adjust_stack (GEN_INT (PROBE_INTERVAL));
	  emit_stack_probe (stack_pointer_rtx);
	}

      if (first_probe)
2113
	anti_adjust_stack (plus_constant (Pmode, size, PROBE_INTERVAL + dope));
2114
      else
2115
	anti_adjust_stack (plus_constant (Pmode, size, PROBE_INTERVAL - i));
2116 2117 2118 2119 2120 2121 2122 2123 2124 2125 2126
      emit_stack_probe (stack_pointer_rtx);
    }

  /* In the variable case, do the same as above, but in a loop.  Note that we
     must be extra careful with variables wrapping around because we might be
     at the very top (or the very bottom) of the address space and we have to
     be able to handle this case properly; in particular, we use an equality
     test for the loop condition.  */
  else
    {
      rtx rounded_size, rounded_size_op, last_addr, temp;
2127 2128
      rtx_code_label *loop_lab = gen_label_rtx ();
      rtx_code_label *end_lab = gen_label_rtx ();
2129 2130 2131 2132 2133 2134


      /* Step 1: round SIZE to the previous multiple of the interval.  */

      /* ROUNDED_SIZE = SIZE & -PROBE_INTERVAL  */
      rounded_size
Richard Sandiford committed
2135 2136
	= simplify_gen_binary (AND, Pmode, size,
			       gen_int_mode (-PROBE_INTERVAL, Pmode));
2137 2138 2139 2140 2141 2142 2143 2144 2145 2146 2147 2148 2149 2150 2151 2152
      rounded_size_op = force_operand (rounded_size, NULL_RTX);


      /* Step 2: compute initial and final value of the loop counter.  */

      /* SP = SP_0 + PROBE_INTERVAL.  */
      anti_adjust_stack (GEN_INT (PROBE_INTERVAL + dope));

      /* LAST_ADDR = SP_0 + PROBE_INTERVAL + ROUNDED_SIZE.  */
      last_addr = force_operand (gen_rtx_fmt_ee (STACK_GROW_OP, Pmode,
						 stack_pointer_rtx,
						 rounded_size_op), NULL_RTX);


      /* Step 3: the loop

2153 2154 2155 2156 2157
	 while (SP != LAST_ADDR)
	   {
	     SP = SP + PROBE_INTERVAL
	     probe at SP
	   }
2158

2159
	 adjusts SP and probes at PROBE_INTERVAL + N * PROBE_INTERVAL for
2160 2161 2162 2163 2164 2165 2166 2167 2168 2169 2170 2171 2172 2173 2174 2175 2176
	 values of N from 1 until it is equal to ROUNDED_SIZE.  */

      emit_label (loop_lab);

      /* Jump to END_LAB if SP == LAST_ADDR.  */
      emit_cmp_and_jump_insns (stack_pointer_rtx, last_addr, EQ, NULL_RTX,
			       Pmode, 1, end_lab);

      /* SP = SP + PROBE_INTERVAL and probe at SP.  */
      anti_adjust_stack (GEN_INT (PROBE_INTERVAL));
      emit_stack_probe (stack_pointer_rtx);

      emit_jump (loop_lab);

      emit_label (end_lab);


2177
      /* Step 4: adjust SP and probe at PROBE_INTERVAL + SIZE if we cannot
2178 2179 2180 2181 2182 2183 2184 2185 2186 2187 2188 2189 2190 2191
	 assert at compile-time that SIZE is equal to ROUNDED_SIZE.  */

      /* TEMP = SIZE - ROUNDED_SIZE.  */
      temp = simplify_gen_binary (MINUS, Pmode, size, rounded_size);
      if (temp != const0_rtx)
	{
	  /* Manual CSE if the difference is not known at compile-time.  */
	  if (GET_CODE (temp) != CONST_INT)
	    temp = gen_rtx_MINUS (Pmode, size, rounded_size_op);
	  anti_adjust_stack (temp);
	  emit_stack_probe (stack_pointer_rtx);
	}
    }

2192 2193
  /* Adjust back and account for the additional first interval.  */
  if (adjust_back)
2194
    adjust_stack (plus_constant (Pmode, size, PROBE_INTERVAL + dope));
2195 2196
  else
    adjust_stack (GEN_INT (PROBE_INTERVAL + dope));
2197 2198
}

Richard Kenner committed
2199 2200 2201
/* Return an rtx representing the register or memory location
   in which a scalar value of data type VALTYPE
   was returned by a function call to function FUNC.
2202 2203
   FUNC is a FUNCTION_DECL, FNTYPE a FUNCTION_TYPE node if the precise
   function is known, otherwise 0.
2204 2205
   OUTGOING is 1 if on a machine with register windows this function
   should return the register in which the function will put its result
2206
   and 0 otherwise.  */
Richard Kenner committed
2207 2208

rtx
2209
hard_function_value (const_tree valtype, const_tree func, const_tree fntype,
2210
		     int outgoing ATTRIBUTE_UNUSED)
Richard Kenner committed
2211
{
2212
  rtx val;
2213

2214
  val = targetm.calls.function_value (valtype, func ? func : fntype, outgoing);
2215

2216
  if (REG_P (val)
2217 2218
      && GET_MODE (val) == BLKmode)
    {
2219
      unsigned HOST_WIDE_INT bytes = arg_int_size_in_bytes (valtype);
2220
      opt_scalar_int_mode tmpmode;
2221

2222
      /* int_size_in_bytes can return -1.  We don't need a check here
2223 2224
	 since the value of bytes will then be large enough that no
	 mode will match anyway.  */
2225

2226
      FOR_EACH_MODE_IN_CLASS (tmpmode, MODE_INT)
Kazu Hirata committed
2227 2228
	{
	  /* Have we found a large enough mode?  */
2229
	  if (GET_MODE_SIZE (tmpmode.require ()) >= bytes)
Kazu Hirata committed
2230 2231
	    break;
	}
2232

2233
      PUT_MODE (val, tmpmode.require ());
2234
    }
2235
  return val;
Richard Kenner committed
2236 2237 2238 2239 2240 2241
}

/* Return an rtx representing the register or memory location
   in which a scalar value of mode MODE was returned by a library call.  */

rtx
2242
hard_libcall_value (machine_mode mode, rtx fun)
Richard Kenner committed
2243
{
2244
  return targetm.calls.libcall_value (mode, fun);
Richard Kenner committed
2245
}
2246 2247

/* Look up the tree code for a given rtx code
2248
   to provide the arithmetic operation for real_arithmetic.
2249 2250 2251 2252
   The function returns an int because the caller may not know
   what `enum tree_code' means.  */

int
2253
rtx_to_tree_code (enum rtx_code code)
2254 2255 2256 2257 2258 2259 2260 2261 2262 2263 2264 2265 2266 2267 2268 2269 2270 2271 2272 2273 2274 2275 2276 2277 2278 2279 2280 2281 2282
{
  enum tree_code tcode;

  switch (code)
    {
    case PLUS:
      tcode = PLUS_EXPR;
      break;
    case MINUS:
      tcode = MINUS_EXPR;
      break;
    case MULT:
      tcode = MULT_EXPR;
      break;
    case DIV:
      tcode = RDIV_EXPR;
      break;
    case SMIN:
      tcode = MIN_EXPR;
      break;
    case SMAX:
      tcode = MAX_EXPR;
      break;
    default:
      tcode = LAST_AND_UNUSED_TREE_CODE;
      break;
    }
  return ((int) tcode);
}
2283 2284

#include "gt-explow.h"