Skip to content
Projects
Groups
Snippets
Help
This project
Loading...
Sign in / Register
Toggle navigation
R
riscv-gcc-1
Overview
Overview
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
lvzhengyang
riscv-gcc-1
Commits
a29c7ea6
Commit
a29c7ea6
authored
Feb 26, 2003
by
Zdenek Dvorak
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
* loop-unroll.c: New.
From-SVN: r63470
parent
4c8826cd
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
1195 additions
and
0 deletions
+1195
-0
gcc/loop-unroll.c
+1195
-0
No files found.
gcc/loop-unroll.c
0 → 100644
View file @
a29c7ea6
/* Loop unrolling and peeling.
Copyright (C) 2002 Free Software Foundation, Inc.
This file is part of GCC.
GCC is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free
Software Foundation; either version 2, or (at your option) any later
version.
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
WARRANTY; without even the implied warranty of MERCHANTABILITY or
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
for more details.
You should have received a copy of the GNU General Public License
along with GCC; see the file COPYING. If not, write to the Free
Software Foundation, 59 Temple Place - Suite 330, Boston, MA
02111-1307, USA. */
#include "config.h"
#include "system.h"
#include "coretypes.h"
#include "tm.h"
#include "rtl.h"
#include "hard-reg-set.h"
#include "basic-block.h"
#include "cfgloop.h"
#include "cfglayout.h"
#include "params.h"
#include "output.h"
#include "expr.h"
/* This pass performs loop unrolling and peeling. We only perform these
optimalizations on innermost loops (with single exception) because
the impact on performance is greatest here, and we want to avoid
unnecessary code size growth. The gain is caused by greater sequentiality
of code, better code to optimize for futher passes and in some cases
by fewer testings of exit conditions. The main problem is code growth,
that impacts performance negatively due to effect of caches.
What we do:
-- complete peeling of once-rolling loops; this is the above mentioned
exception, as this causes loop to be cancelled completely and
does not cause code growth
-- complete peeling of loops that roll (small) constant times.
-- simple peeling of first iterations of loops that do not roll much
(according to profile feedback)
-- unrolling of loops that roll constant times; this is almost always
win, as we get rid of exit condition tests.
-- unrolling of loops that roll number of times that we can compute
in runtime; we also get rid of exit condition tests here, but there
is the extra expense for calculating the number of iterations
-- simple unrolling of remaining loops; this is performed only if we
are asked to, as the gain is questionable in this case and often
it may even slow down the code
For more detailed descriptions of each of those, see comments at
appropriate function below.
There is a lot of parameters (defined and described in params.def) that
control how much we unroll/peel.
??? A great problem is that we don't have a good way how to determine
how many times we should unroll the loop; the experiments I have made
showed that this choice may affect performance in order of several %.
*/
static
void
decide_unrolling_and_peeling
PARAMS
((
struct
loops
*
,
int
));
static
void
peel_loops_completely
PARAMS
((
struct
loops
*
,
int
));
static
void
decide_peel_simple
PARAMS
((
struct
loops
*
,
struct
loop
*
,
int
));
static
void
decide_peel_once_rolling
PARAMS
((
struct
loops
*
,
struct
loop
*
,
int
));
static
void
decide_peel_completely
PARAMS
((
struct
loops
*
,
struct
loop
*
,
int
));
static
void
decide_unroll_stupid
PARAMS
((
struct
loops
*
,
struct
loop
*
,
int
));
static
void
decide_unroll_constant_iterations
PARAMS
((
struct
loops
*
,
struct
loop
*
,
int
));
static
void
decide_unroll_runtime_iterations
PARAMS
((
struct
loops
*
,
struct
loop
*
,
int
));
static
void
peel_loop_simple
PARAMS
((
struct
loops
*
,
struct
loop
*
));
static
void
peel_loop_completely
PARAMS
((
struct
loops
*
,
struct
loop
*
));
static
void
unroll_loop_stupid
PARAMS
((
struct
loops
*
,
struct
loop
*
));
static
void
unroll_loop_constant_iterations
PARAMS
((
struct
loops
*
,
struct
loop
*
));
static
void
unroll_loop_runtime_iterations
PARAMS
((
struct
loops
*
,
struct
loop
*
));
/* Unroll and/or peel (depending on FLAGS) LOOPS. */
void
unroll_and_peel_loops
(
loops
,
flags
)
struct
loops
*
loops
;
int
flags
;
{
struct
loop
*
loop
,
*
next
;
int
check
;
/* First perform complete loop peeling (it is almost surely a win,
and affects parameters for further decision a lot). */
peel_loops_completely
(
loops
,
flags
);
/* Now decide rest of unrolling and peeling. */
decide_unrolling_and_peeling
(
loops
,
flags
);
loop
=
loops
->
tree_root
;
while
(
loop
->
inner
)
loop
=
loop
->
inner
;
/* Scan the loops, inner ones first. */
while
(
loop
!=
loops
->
tree_root
)
{
if
(
loop
->
next
)
{
next
=
loop
->
next
;
while
(
next
->
inner
)
next
=
next
->
inner
;
}
else
next
=
loop
->
outer
;
check
=
1
;
/* And perform the appropriate transformations. */
switch
(
loop
->
lpt_decision
.
decision
)
{
case
LPT_PEEL_COMPLETELY
:
/* Already done. */
abort
();
case
LPT_PEEL_SIMPLE
:
peel_loop_simple
(
loops
,
loop
);
break
;
case
LPT_UNROLL_CONSTANT
:
unroll_loop_constant_iterations
(
loops
,
loop
);
break
;
case
LPT_UNROLL_RUNTIME
:
unroll_loop_runtime_iterations
(
loops
,
loop
);
break
;
case
LPT_UNROLL_STUPID
:
unroll_loop_stupid
(
loops
,
loop
);
break
;
case
LPT_NONE
:
check
=
0
;
break
;
default
:
abort
();
}
if
(
check
)
{
#ifdef ENABLE_CHECKING
verify_dominators
(
loops
->
cfg
.
dom
);
verify_loop_structure
(
loops
);
#endif
}
loop
=
next
;
}
}
/* Check whether to peel LOOPS (depending on FLAGS) completely and do so. */
static
void
peel_loops_completely
(
loops
,
flags
)
struct
loops
*
loops
;
int
flags
;
{
struct
loop
*
loop
,
*
next
;
loop
=
loops
->
tree_root
;
while
(
loop
->
inner
)
loop
=
loop
->
inner
;
while
(
loop
!=
loops
->
tree_root
)
{
if
(
loop
->
next
)
{
next
=
loop
->
next
;
while
(
next
->
inner
)
next
=
next
->
inner
;
}
else
next
=
loop
->
outer
;
loop
->
lpt_decision
.
decision
=
LPT_NONE
;
loop
->
has_desc
=
0
;
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Considering loop %d for complete peeling
\n
"
,
loop
->
num
);
/* Do not peel cold areas. */
if
(
!
maybe_hot_bb_p
(
loop
->
header
))
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not considering loop, cold area
\n
"
);
loop
=
next
;
continue
;
}
/* Can the loop be manipulated? */
if
(
!
can_duplicate_loop_p
(
loop
))
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not considering loop, cannot duplicate
\n
"
);
loop
=
next
;
continue
;
}
loop
->
ninsns
=
num_loop_insns
(
loop
);
decide_peel_once_rolling
(
loops
,
loop
,
flags
);
if
(
loop
->
lpt_decision
.
decision
==
LPT_NONE
)
decide_peel_completely
(
loops
,
loop
,
flags
);
if
(
loop
->
lpt_decision
.
decision
==
LPT_PEEL_COMPLETELY
)
{
peel_loop_completely
(
loops
,
loop
);
#ifdef ENABLE_CHECKING
verify_dominators
(
loops
->
cfg
.
dom
);
verify_loop_structure
(
loops
);
#endif
}
loop
=
next
;
}
}
/* Decide whether unroll or peel LOOPS (depending on FLAGS) and how much. */
static
void
decide_unrolling_and_peeling
(
loops
,
flags
)
struct
loops
*
loops
;
int
flags
;
{
struct
loop
*
loop
=
loops
->
tree_root
,
*
next
;
while
(
loop
->
inner
)
loop
=
loop
->
inner
;
/* Scan the loops, inner ones first. */
while
(
loop
!=
loops
->
tree_root
)
{
if
(
loop
->
next
)
{
next
=
loop
->
next
;
while
(
next
->
inner
)
next
=
next
->
inner
;
}
else
next
=
loop
->
outer
;
loop
->
lpt_decision
.
decision
=
LPT_NONE
;
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Considering loop %d
\n
"
,
loop
->
num
);
/* Do not peel cold areas. */
if
(
!
maybe_hot_bb_p
(
loop
->
header
))
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not considering loop, cold area
\n
"
);
loop
=
next
;
continue
;
}
/* Can the loop be manipulated? */
if
(
!
can_duplicate_loop_p
(
loop
))
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not considering loop, cannot duplicate
\n
"
);
loop
=
next
;
continue
;
}
/* Skip non-innermost loops. */
if
(
loop
->
inner
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not considering loop, is not innermost
\n
"
);
loop
=
next
;
continue
;
}
loop
->
ninsns
=
num_loop_insns
(
loop
);
loop
->
av_ninsns
=
average_num_loop_insns
(
loop
);
/* Try transformations one by one in decreasing order of
priority. */
decide_unroll_constant_iterations
(
loops
,
loop
,
flags
);
if
(
loop
->
lpt_decision
.
decision
==
LPT_NONE
)
decide_unroll_runtime_iterations
(
loops
,
loop
,
flags
);
if
(
loop
->
lpt_decision
.
decision
==
LPT_NONE
)
decide_unroll_stupid
(
loops
,
loop
,
flags
);
if
(
loop
->
lpt_decision
.
decision
==
LPT_NONE
)
decide_peel_simple
(
loops
,
loop
,
flags
);
loop
=
next
;
}
}
/* Decide whether the LOOP is once rolling and suitable for complete
peeling. */
static
void
decide_peel_once_rolling
(
loops
,
loop
,
flags
)
struct
loops
*
loops
;
struct
loop
*
loop
;
int
flags
ATTRIBUTE_UNUSED
;
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Considering peeling once rolling loop
\n
"
);
/* Is the loop small enough? */
if
((
unsigned
)
PARAM_VALUE
(
PARAM_MAX_ONCE_PEELED_INSNS
)
<
loop
->
ninsns
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not considering loop, is too big
\n
"
);
return
;
}
/* Check for simple loops. */
loop
->
simple
=
simple_loop_p
(
loops
,
loop
,
&
loop
->
desc
);
loop
->
has_desc
=
1
;
/* Check number of iterations. */
if
(
!
loop
->
simple
||
!
loop
->
desc
.
const_iter
||
loop
->
desc
.
niter
!=
0
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Unable to prove that the loop rolls exactly once
\n
"
);
return
;
}
/* Success. */
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Decided to peel exactly once rolling loop
\n
"
);
loop
->
lpt_decision
.
decision
=
LPT_PEEL_COMPLETELY
;
}
/* Decide whether the LOOP is suitable for complete peeling. */
static
void
decide_peel_completely
(
loops
,
loop
,
flags
)
struct
loops
*
loops
;
struct
loop
*
loop
;
int
flags
ATTRIBUTE_UNUSED
;
{
unsigned
npeel
;
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Considering peeling completely
\n
"
);
/* Skip non-innermost loops. */
if
(
loop
->
inner
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not considering loop, is not innermost
\n
"
);
return
;
}
/* npeel = number of iterations to peel. */
npeel
=
PARAM_VALUE
(
PARAM_MAX_COMPLETELY_PEELED_INSNS
)
/
loop
->
ninsns
;
if
(
npeel
>
(
unsigned
)
PARAM_VALUE
(
PARAM_MAX_COMPLETELY_PEEL_TIMES
))
npeel
=
PARAM_VALUE
(
PARAM_MAX_COMPLETELY_PEEL_TIMES
);
/* Is the loop small enough? */
if
(
!
npeel
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not considering loop, is too big
\n
"
);
return
;
}
/* Check for simple loops. */
if
(
!
loop
->
has_desc
)
loop
->
simple
=
simple_loop_p
(
loops
,
loop
,
&
loop
->
desc
);
/* Check number of iterations. */
if
(
!
loop
->
simple
||
!
loop
->
desc
.
const_iter
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Unable to prove that the loop iterates constant times
\n
"
);
return
;
}
if
(
loop
->
desc
.
niter
>
npeel
-
1
)
{
if
(
rtl_dump_file
)
{
fprintf
(
rtl_dump_file
,
";; Not peeling loop completely, rolls too much ("
);
fprintf
(
rtl_dump_file
,
HOST_WIDEST_INT_PRINT_DEC
,(
HOST_WIDEST_INT
)
loop
->
desc
.
niter
);
fprintf
(
rtl_dump_file
,
"iterations > %d [maximum peelings])
\n
"
,
npeel
);
}
return
;
}
/* Success. */
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Decided to peel loop completely
\n
"
);
loop
->
lpt_decision
.
decision
=
LPT_PEEL_COMPLETELY
;
}
/* Peel all iterations of LOOP, remove exit edges and cancel the loop
completely. The transformation done:
for (i = 0; i < 4; i++)
body;
==>
i = 0;
body; i++;
body; i++;
body; i++;
body; i++;
*/
static
void
peel_loop_completely
(
loops
,
loop
)
struct
loops
*
loops
;
struct
loop
*
loop
;
{
sbitmap
wont_exit
;
unsigned
HOST_WIDE_INT
npeel
;
edge
e
;
unsigned
n_remove_edges
,
i
;
edge
*
remove_edges
;
struct
loop_desc
*
desc
=
&
loop
->
desc
;
npeel
=
desc
->
niter
;
wont_exit
=
sbitmap_alloc
(
npeel
+
2
);
sbitmap_ones
(
wont_exit
);
RESET_BIT
(
wont_exit
,
0
);
RESET_BIT
(
wont_exit
,
npeel
+
1
);
if
(
desc
->
may_be_zero
)
RESET_BIT
(
wont_exit
,
1
);
remove_edges
=
xcalloc
(
npeel
,
sizeof
(
edge
));
n_remove_edges
=
0
;
if
(
!
duplicate_loop_to_header_edge
(
loop
,
loop_preheader_edge
(
loop
),
loops
,
npeel
+
1
,
wont_exit
,
desc
->
out_edge
,
remove_edges
,
&
n_remove_edges
,
DLTHE_FLAG_UPDATE_FREQ
))
abort
();
free
(
wont_exit
);
/* Remove the exit edges. */
for
(
i
=
0
;
i
<
n_remove_edges
;
i
++
)
remove_path
(
loops
,
remove_edges
[
i
]);
free
(
remove_edges
);
/* Now remove the loop. */
for
(
e
=
RBI
(
desc
->
in_edge
->
src
)
->
copy
->
succ
;
e
&&
e
->
dest
!=
RBI
(
desc
->
in_edge
->
dest
)
->
copy
;
e
=
e
->
succ_next
);
if
(
!
e
)
abort
();
remove_path
(
loops
,
e
);
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Peeled loop completely, %d times
\n
"
,
(
int
)
npeel
);
}
/* Decide whether to unroll LOOP iterating constant number of times and how much. */
static
void
decide_unroll_constant_iterations
(
loops
,
loop
,
flags
)
struct
loops
*
loops
;
struct
loop
*
loop
;
int
flags
;
{
unsigned
nunroll
,
nunroll_by_av
,
best_copies
,
best_unroll
=
-
1
,
n_copies
,
i
;
if
(
!
(
flags
&
UAP_UNROLL
))
{
/* We were not asked to, just return back silently. */
return
;
}
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Considering unrolling loop with constant number of iterations
\n
"
);
/* nunroll = total number of copies of the original loop body in
unrolled loop (i.e. if it is 2, we have to duplicate loop body once. */
nunroll
=
PARAM_VALUE
(
PARAM_MAX_UNROLLED_INSNS
)
/
loop
->
ninsns
;
nunroll_by_av
=
PARAM_VALUE
(
PARAM_MAX_AVERAGE_UNROLLED_INSNS
)
/
loop
->
av_ninsns
;
if
(
nunroll
>
nunroll_by_av
)
nunroll
=
nunroll_by_av
;
if
(
nunroll
>
(
unsigned
)
PARAM_VALUE
(
PARAM_MAX_UNROLL_TIMES
))
nunroll
=
PARAM_VALUE
(
PARAM_MAX_UNROLL_TIMES
);
/* Skip big loops. */
if
(
nunroll
<=
1
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not considering loop, is too big
\n
"
);
return
;
}
/* Check for simple loops. */
if
(
!
loop
->
has_desc
)
loop
->
simple
=
simple_loop_p
(
loops
,
loop
,
&
loop
->
desc
);
/* Check number of iterations. */
if
(
!
loop
->
simple
||
!
loop
->
desc
.
const_iter
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Unable to prove that the loop iterates constant times
\n
"
);
return
;
}
/* Check whether the loop rolls enough to consider. */
if
(
loop
->
desc
.
niter
<
2
*
nunroll
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not unrolling loop, doesn't roll
\n
"
);
return
;
}
/* Success; now compute number of iterations to unroll. We alter
nunroll so that as few as possible copies of loop body are
neccesary, while still not decreasing the number of unrollings
too much (at most by 1). */
best_copies
=
2
*
nunroll
+
10
;
i
=
2
*
nunroll
+
2
;
if
((
unsigned
)
i
-
1
>=
loop
->
desc
.
niter
)
i
=
loop
->
desc
.
niter
-
2
;
for
(;
i
>=
nunroll
-
1
;
i
--
)
{
unsigned
exit_mod
=
loop
->
desc
.
niter
%
(
i
+
1
);
if
(
loop
->
desc
.
postincr
)
n_copies
=
exit_mod
+
i
+
1
;
else
if
(
exit_mod
!=
(
unsigned
)
i
||
loop
->
desc
.
may_be_zero
)
n_copies
=
exit_mod
+
i
+
2
;
else
n_copies
=
i
+
1
;
if
(
n_copies
<
best_copies
)
{
best_copies
=
n_copies
;
best_unroll
=
i
;
}
}
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; max_unroll %d (%d copies, initial %d).
\n
"
,
best_unroll
+
1
,
best_copies
,
nunroll
);
loop
->
lpt_decision
.
decision
=
LPT_UNROLL_CONSTANT
;
loop
->
lpt_decision
.
times
=
best_unroll
;
}
/* Unroll LOOP with constant number of iterations LOOP->LPT_DECISION.TIMES + 1
times. The transformation does this:
for (i = 0; i < 102; i++)
body;
==>
i = 0;
body; i++;
body; i++;
while (i < 102)
{
body; i++;
body; i++;
body; i++;
body; i++;
}
*/
static
void
unroll_loop_constant_iterations
(
loops
,
loop
)
struct
loops
*
loops
;
struct
loop
*
loop
;
{
unsigned
HOST_WIDE_INT
niter
;
unsigned
exit_mod
;
sbitmap
wont_exit
;
unsigned
n_remove_edges
,
i
;
edge
*
remove_edges
;
unsigned
max_unroll
=
loop
->
lpt_decision
.
times
;
struct
loop_desc
*
desc
=
&
loop
->
desc
;
niter
=
desc
->
niter
;
if
(
niter
<=
(
unsigned
)
max_unroll
+
1
)
abort
();
/* Should not get here (such loop should be peeled instead). */
exit_mod
=
niter
%
(
max_unroll
+
1
);
wont_exit
=
sbitmap_alloc
(
max_unroll
+
1
);
sbitmap_ones
(
wont_exit
);
remove_edges
=
xcalloc
(
max_unroll
+
exit_mod
+
1
,
sizeof
(
edge
));
n_remove_edges
=
0
;
if
(
desc
->
postincr
)
{
/* Counter is incremented after the exit test; leave exit test
in the first copy, so that the loops that start with test
of exit condition have continuous body after unrolling. */
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Condition on beginning of loop.
\n
"
);
/* Peel exit_mod iterations. */
RESET_BIT
(
wont_exit
,
0
);
if
(
desc
->
may_be_zero
)
RESET_BIT
(
wont_exit
,
1
);
if
(
exit_mod
&&
!
duplicate_loop_to_header_edge
(
loop
,
loop_preheader_edge
(
loop
),
loops
,
exit_mod
,
wont_exit
,
desc
->
out_edge
,
remove_edges
,
&
n_remove_edges
,
DLTHE_FLAG_UPDATE_FREQ
))
abort
();
SET_BIT
(
wont_exit
,
1
);
}
else
{
/* Leave exit test in last copy, for the same reason as above if
the loop tests the condition at the end of loop body. */
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Condition on end of loop.
\n
"
);
/* We know that niter >= max_unroll + 2; so we do not need to care of
case when we would exit before reaching the loop. So just peel
exit_mod + 1 iterations.
*/
if
(
exit_mod
!=
(
unsigned
)
max_unroll
||
desc
->
may_be_zero
)
{
RESET_BIT
(
wont_exit
,
0
);
if
(
desc
->
may_be_zero
)
RESET_BIT
(
wont_exit
,
1
);
if
(
!
duplicate_loop_to_header_edge
(
loop
,
loop_preheader_edge
(
loop
),
loops
,
exit_mod
+
1
,
wont_exit
,
desc
->
out_edge
,
remove_edges
,
&
n_remove_edges
,
DLTHE_FLAG_UPDATE_FREQ
))
abort
();
SET_BIT
(
wont_exit
,
0
);
SET_BIT
(
wont_exit
,
1
);
}
RESET_BIT
(
wont_exit
,
max_unroll
);
}
/* Now unroll the loop. */
if
(
!
duplicate_loop_to_header_edge
(
loop
,
loop_latch_edge
(
loop
),
loops
,
max_unroll
,
wont_exit
,
desc
->
out_edge
,
remove_edges
,
&
n_remove_edges
,
DLTHE_FLAG_UPDATE_FREQ
))
abort
();
free
(
wont_exit
);
/* Remove the edges. */
for
(
i
=
0
;
i
<
n_remove_edges
;
i
++
)
remove_path
(
loops
,
remove_edges
[
i
]);
free
(
remove_edges
);
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Unrolled loop %d times, constant # of iterations %i insns
\n
"
,
max_unroll
,
num_loop_insns
(
loop
));
}
/* Decide whether to unroll LOOP iterating runtime computable number of times
and how much. */
static
void
decide_unroll_runtime_iterations
(
loops
,
loop
,
flags
)
struct
loops
*
loops
;
struct
loop
*
loop
;
int
flags
;
{
unsigned
nunroll
,
nunroll_by_av
,
i
;
if
(
!
(
flags
&
UAP_UNROLL
))
{
/* We were not asked to, just return back silently. */
return
;
}
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Considering unrolling loop with runtime computable number of iterations
\n
"
);
/* nunroll = total number of copies of the original loop body in
unrolled loop (i.e. if it is 2, we have to duplicate loop body once. */
nunroll
=
PARAM_VALUE
(
PARAM_MAX_UNROLLED_INSNS
)
/
loop
->
ninsns
;
nunroll_by_av
=
PARAM_VALUE
(
PARAM_MAX_AVERAGE_UNROLLED_INSNS
)
/
loop
->
av_ninsns
;
if
(
nunroll
>
nunroll_by_av
)
nunroll
=
nunroll_by_av
;
if
(
nunroll
>
(
unsigned
)
PARAM_VALUE
(
PARAM_MAX_UNROLL_TIMES
))
nunroll
=
PARAM_VALUE
(
PARAM_MAX_UNROLL_TIMES
);
/* Skip big loops. */
if
(
nunroll
<=
1
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not considering loop, is too big
\n
"
);
return
;
}
/* Check for simple loops. */
if
(
!
loop
->
has_desc
)
loop
->
simple
=
simple_loop_p
(
loops
,
loop
,
&
loop
->
desc
);
/* Check simpleness. */
if
(
!
loop
->
simple
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Unable to prove that the number of iterations can be counted in runtime
\n
"
);
return
;
}
if
(
loop
->
desc
.
const_iter
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Loop iterates constant times
\n
"
);
return
;
}
/* If we have profile feedback, check whether the loop rolls. */
if
(
loop
->
header
->
count
&&
expected_loop_iterations
(
loop
)
<
2
*
nunroll
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not unrolling loop, doesn't roll
\n
"
);
return
;
}
/* Success; now force nunroll to be power of 2, as we are unable to
cope with overflows in computation of number of iterations. */
for
(
i
=
1
;
2
*
i
<=
nunroll
;
i
*=
2
);
loop
->
lpt_decision
.
decision
=
LPT_UNROLL_RUNTIME
;
loop
->
lpt_decision
.
times
=
i
-
1
;
}
/* Unroll LOOP for that we are able to count number of iterations in runtime
LOOP->LPT_DECISION.TIMES + 1 times. The transformation does this (with some
extra care for case n < 0):
for (i = 0; i < n; i++)
body;
==>
i = 0;
mod = n % 4;
switch (mod)
{
case 3:
body; i++;
case 2:
body; i++;
case 1:
body; i++;
case 0: ;
}
while (i < n)
{
body; i++;
body; i++;
body; i++;
body; i++;
}
*/
static
void
unroll_loop_runtime_iterations
(
loops
,
loop
)
struct
loops
*
loops
;
struct
loop
*
loop
;
{
rtx
niter
,
init_code
,
branch_code
,
jump
,
label
;
unsigned
i
,
j
,
p
;
basic_block
preheader
,
*
body
,
*
dom_bbs
,
swtch
,
ezc_swtch
;
unsigned
n_dom_bbs
;
sbitmap
wont_exit
;
int
may_exit_copy
;
unsigned
n_peel
,
n_remove_edges
;
edge
*
remove_edges
,
e
;
bool
extra_zero_check
,
last_may_exit
;
unsigned
max_unroll
=
loop
->
lpt_decision
.
times
;
struct
loop_desc
*
desc
=
&
loop
->
desc
;
/* Remember blocks whose dominators will have to be updated. */
dom_bbs
=
xcalloc
(
n_basic_blocks
,
sizeof
(
basic_block
));
n_dom_bbs
=
0
;
body
=
get_loop_body
(
loop
);
for
(
i
=
0
;
i
<
loop
->
num_nodes
;
i
++
)
{
unsigned
nldom
;
basic_block
*
ldom
;
nldom
=
get_dominated_by
(
loops
->
cfg
.
dom
,
body
[
i
],
&
ldom
);
for
(
j
=
0
;
j
<
nldom
;
j
++
)
if
(
!
flow_bb_inside_loop_p
(
loop
,
ldom
[
j
]))
dom_bbs
[
n_dom_bbs
++
]
=
ldom
[
j
];
free
(
ldom
);
}
free
(
body
);
if
(
desc
->
postincr
)
{
/* Leave exit in first copy (for explanation why see comment in
unroll_loop_constant_iterations). */
may_exit_copy
=
0
;
n_peel
=
max_unroll
-
1
;
extra_zero_check
=
true
;
last_may_exit
=
false
;
}
else
{
/* Leave exit in last copy (for explanation why see comment in
unroll_loop_constant_iterations). */
may_exit_copy
=
max_unroll
;
n_peel
=
max_unroll
;
extra_zero_check
=
false
;
last_may_exit
=
true
;
}
/* Get expression for number of iterations. */
start_sequence
();
niter
=
count_loop_iterations
(
desc
,
NULL
,
NULL
);
if
(
!
niter
)
abort
();
niter
=
force_operand
(
niter
,
NULL
);
/* Count modulo by ANDing it with max_unroll; we use the fact that
the number of unrollings is a power of two, and thus this is correct
even if there is overflow in the computation. */
niter
=
expand_simple_binop
(
GET_MODE
(
desc
->
var
),
AND
,
niter
,
GEN_INT
(
max_unroll
),
NULL_RTX
,
0
,
OPTAB_LIB_WIDEN
);
init_code
=
get_insns
();
end_sequence
();
/* Precondition the loop. */
loop_split_edge_with
(
loop_preheader_edge
(
loop
),
init_code
,
loops
);
remove_edges
=
xcalloc
(
max_unroll
+
n_peel
+
1
,
sizeof
(
edge
));
n_remove_edges
=
0
;
wont_exit
=
sbitmap_alloc
(
max_unroll
+
2
);
/* Peel the first copy of loop body (almost always we must leave exit test
here; the only exception is when we have extra zero check and the number
of iterations is reliable (i.e. comes out of NE condition). Also record
the place of (possible) extra zero check. */
sbitmap_zero
(
wont_exit
);
if
(
extra_zero_check
&&
desc
->
cond
==
NE
)
SET_BIT
(
wont_exit
,
1
);
ezc_swtch
=
loop_preheader_edge
(
loop
)
->
src
;
if
(
!
duplicate_loop_to_header_edge
(
loop
,
loop_preheader_edge
(
loop
),
loops
,
1
,
wont_exit
,
desc
->
out_edge
,
remove_edges
,
&
n_remove_edges
,
DLTHE_FLAG_UPDATE_FREQ
))
abort
();
/* Record the place where switch will be built for preconditioning. */
swtch
=
loop_split_edge_with
(
loop_preheader_edge
(
loop
),
NULL_RTX
,
loops
);
for
(
i
=
0
;
i
<
n_peel
;
i
++
)
{
/* Peel the copy. */
sbitmap_zero
(
wont_exit
);
if
(
i
!=
n_peel
-
1
||
!
last_may_exit
)
SET_BIT
(
wont_exit
,
1
);
if
(
!
duplicate_loop_to_header_edge
(
loop
,
loop_preheader_edge
(
loop
),
loops
,
1
,
wont_exit
,
desc
->
out_edge
,
remove_edges
,
&
n_remove_edges
,
DLTHE_FLAG_UPDATE_FREQ
))
abort
();
if
(
i
!=
n_peel
)
{
/* Create item for switch. */
j
=
n_peel
-
i
-
(
extra_zero_check
?
0
:
1
);
p
=
REG_BR_PROB_BASE
/
(
i
+
2
);
preheader
=
loop_split_edge_with
(
loop_preheader_edge
(
loop
),
NULL_RTX
,
loops
);
label
=
block_label
(
preheader
);
start_sequence
();
do_compare_rtx_and_jump
(
copy_rtx
(
niter
),
GEN_INT
(
j
),
EQ
,
0
,
GET_MODE
(
desc
->
var
),
NULL_RTX
,
NULL_RTX
,
label
);
jump
=
get_last_insn
();
JUMP_LABEL
(
jump
)
=
label
;
REG_NOTES
(
jump
)
=
gen_rtx_EXPR_LIST
(
REG_BR_PROB
,
GEN_INT
(
p
),
REG_NOTES
(
jump
));
LABEL_NUSES
(
label
)
++
;
branch_code
=
get_insns
();
end_sequence
();
swtch
=
loop_split_edge_with
(
swtch
->
pred
,
branch_code
,
loops
);
set_immediate_dominator
(
loops
->
cfg
.
dom
,
preheader
,
swtch
);
swtch
->
succ
->
probability
=
REG_BR_PROB_BASE
-
p
;
e
=
make_edge
(
swtch
,
preheader
,
0
);
e
->
probability
=
p
;
}
}
if
(
extra_zero_check
)
{
/* Add branch for zero iterations. */
p
=
REG_BR_PROB_BASE
/
(
max_unroll
+
1
);
swtch
=
ezc_swtch
;
preheader
=
loop_split_edge_with
(
loop_preheader_edge
(
loop
),
NULL_RTX
,
loops
);
label
=
block_label
(
preheader
);
start_sequence
();
do_compare_rtx_and_jump
(
copy_rtx
(
niter
),
const0_rtx
,
EQ
,
0
,
GET_MODE
(
desc
->
var
),
NULL_RTX
,
NULL_RTX
,
label
);
jump
=
get_last_insn
();
JUMP_LABEL
(
jump
)
=
label
;
REG_NOTES
(
jump
)
=
gen_rtx_EXPR_LIST
(
REG_BR_PROB
,
GEN_INT
(
p
),
REG_NOTES
(
jump
));
LABEL_NUSES
(
label
)
++
;
branch_code
=
get_insns
();
end_sequence
();
swtch
=
loop_split_edge_with
(
swtch
->
succ
,
branch_code
,
loops
);
set_immediate_dominator
(
loops
->
cfg
.
dom
,
preheader
,
swtch
);
swtch
->
succ
->
probability
=
REG_BR_PROB_BASE
-
p
;
e
=
make_edge
(
swtch
,
preheader
,
0
);
e
->
probability
=
p
;
}
/* Recount dominators for outer blocks. */
iterate_fix_dominators
(
loops
->
cfg
.
dom
,
dom_bbs
,
n_dom_bbs
);
/* And unroll loop. */
sbitmap_ones
(
wont_exit
);
RESET_BIT
(
wont_exit
,
may_exit_copy
);
if
(
!
duplicate_loop_to_header_edge
(
loop
,
loop_latch_edge
(
loop
),
loops
,
max_unroll
,
wont_exit
,
desc
->
out_edge
,
remove_edges
,
&
n_remove_edges
,
DLTHE_FLAG_UPDATE_FREQ
))
abort
();
free
(
wont_exit
);
/* Remove the edges. */
for
(
i
=
0
;
i
<
n_remove_edges
;
i
++
)
remove_path
(
loops
,
remove_edges
[
i
]);
free
(
remove_edges
);
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Unrolled loop %d times, counting # of iterations in runtime, %i insns
\n
"
,
max_unroll
,
num_loop_insns
(
loop
));
}
/* Decide whether to simply peel LOOP and how much. */
static
void
decide_peel_simple
(
loops
,
loop
,
flags
)
struct
loops
*
loops
;
struct
loop
*
loop
;
int
flags
;
{
unsigned
npeel
;
if
(
!
(
flags
&
UAP_PEEL
))
{
/* We were not asked to, just return back silently. */
return
;
}
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Considering simply peeling loop
\n
"
);
/* npeel = number of iterations to peel. */
npeel
=
PARAM_VALUE
(
PARAM_MAX_PEELED_INSNS
)
/
loop
->
ninsns
;
if
(
npeel
>
(
unsigned
)
PARAM_VALUE
(
PARAM_MAX_PEEL_TIMES
))
npeel
=
PARAM_VALUE
(
PARAM_MAX_PEEL_TIMES
);
/* Skip big loops. */
if
(
!
npeel
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not considering loop, is too big
\n
"
);
return
;
}
/* Check for simple loops. */
if
(
!
loop
->
has_desc
)
loop
->
simple
=
simple_loop_p
(
loops
,
loop
,
&
loop
->
desc
);
/* Check number of iterations. */
if
(
loop
->
simple
&&
loop
->
desc
.
const_iter
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Loop iterates constant times
\n
"
);
return
;
}
/* Do not simply peel loops with branches inside -- it increases number
of mispredicts. */
if
(
loop
->
desc
.
n_branches
>
1
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not peeling, contains branches
\n
"
);
return
;
}
if
(
loop
->
header
->
count
)
{
unsigned
niter
=
expected_loop_iterations
(
loop
);
if
(
niter
+
1
>
npeel
)
{
if
(
rtl_dump_file
)
{
fprintf
(
rtl_dump_file
,
";; Not peeling loop, rolls too much ("
);
fprintf
(
rtl_dump_file
,
HOST_WIDEST_INT_PRINT_DEC
,
(
HOST_WIDEST_INT
)
(
niter
+
1
));
fprintf
(
rtl_dump_file
,
" iterations > %d [maximum peelings])
\n
"
,
npeel
);
}
return
;
}
npeel
=
niter
+
1
;
}
else
{
/* For now we have no good heuristics to decide whether loop peeling
will be effective, so disable it. */
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not peeling loop, no evidence it will be profitable
\n
"
);
return
;
}
/* Success. */
loop
->
lpt_decision
.
decision
=
LPT_PEEL_SIMPLE
;
loop
->
lpt_decision
.
times
=
npeel
;
}
/* Peel a LOOP LOOP->LPT_DECISION.TIMES times. The transformation:
while (cond)
body;
==>
if (!cond) goto end;
body;
if (!cond) goto end;
body;
while (cond)
body;
end: ;
*/
static
void
peel_loop_simple
(
loops
,
loop
)
struct
loops
*
loops
;
struct
loop
*
loop
;
{
sbitmap
wont_exit
;
unsigned
npeel
=
loop
->
lpt_decision
.
times
;
wont_exit
=
sbitmap_alloc
(
npeel
+
1
);
sbitmap_zero
(
wont_exit
);
if
(
!
duplicate_loop_to_header_edge
(
loop
,
loop_preheader_edge
(
loop
),
loops
,
npeel
,
wont_exit
,
NULL
,
NULL
,
NULL
,
DLTHE_FLAG_UPDATE_FREQ
))
abort
();
free
(
wont_exit
);
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Peeling loop %d times
\n
"
,
npeel
);
}
/* Decide whether to unroll LOOP stupidly and how much. */
static
void
decide_unroll_stupid
(
loops
,
loop
,
flags
)
struct
loops
*
loops
;
struct
loop
*
loop
;
int
flags
;
{
unsigned
nunroll
,
nunroll_by_av
,
i
;
if
(
!
(
flags
&
UAP_UNROLL_ALL
))
{
/* We were not asked to, just return back silently. */
return
;
}
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Considering unrolling loop stupidly
\n
"
);
/* nunroll = total number of copies of the original loop body in
unrolled loop (i.e. if it is 2, we have to duplicate loop body once. */
nunroll
=
PARAM_VALUE
(
PARAM_MAX_UNROLLED_INSNS
)
/
loop
->
ninsns
;
nunroll_by_av
=
PARAM_VALUE
(
PARAM_MAX_AVERAGE_UNROLLED_INSNS
)
/
loop
->
av_ninsns
;
if
(
nunroll
>
nunroll_by_av
)
nunroll
=
nunroll_by_av
;
if
(
nunroll
>
(
unsigned
)
PARAM_VALUE
(
PARAM_MAX_UNROLL_TIMES
))
nunroll
=
PARAM_VALUE
(
PARAM_MAX_UNROLL_TIMES
);
/* Skip big loops. */
if
(
nunroll
<=
1
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not considering loop, is too big
\n
"
);
return
;
}
/* Check for simple loops. */
if
(
!
loop
->
has_desc
)
loop
->
simple
=
simple_loop_p
(
loops
,
loop
,
&
loop
->
desc
);
/* Check simpleness. */
if
(
loop
->
simple
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; The loop is simple
\n
"
);
return
;
}
/* Do not unroll loops with branches inside -- it increases number
of mispredicts. */
if
(
loop
->
desc
.
n_branches
>
1
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not unrolling, contains branches
\n
"
);
return
;
}
/* If we have profile feedback, check whether the loop rolls. */
if
(
loop
->
header
->
count
&&
expected_loop_iterations
(
loop
)
<
2
*
nunroll
)
{
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Not unrolling loop, doesn't roll
\n
"
);
return
;
}
/* Success. Now force nunroll to be power of 2, as it seems that this
improves results (partially because of better aligments, partially
because of some dark magic). */
for
(
i
=
1
;
2
*
i
<=
nunroll
;
i
*=
2
);
loop
->
lpt_decision
.
decision
=
LPT_UNROLL_STUPID
;
loop
->
lpt_decision
.
times
=
i
-
1
;
}
/* Unroll a LOOP LOOP->LPT_DECISION.TIMES times. The transformation:
while (cond)
body;
==>
while (cond)
{
body;
if (!cond) break;
body;
if (!cond) break;
body;
if (!cond) break;
body;
}
*/
static
void
unroll_loop_stupid
(
loops
,
loop
)
struct
loops
*
loops
;
struct
loop
*
loop
;
{
sbitmap
wont_exit
;
unsigned
nunroll
=
loop
->
lpt_decision
.
times
;
wont_exit
=
sbitmap_alloc
(
nunroll
+
1
);
sbitmap_zero
(
wont_exit
);
if
(
!
duplicate_loop_to_header_edge
(
loop
,
loop_latch_edge
(
loop
),
loops
,
nunroll
,
wont_exit
,
NULL
,
NULL
,
NULL
,
DLTHE_FLAG_UPDATE_FREQ
))
abort
();
free
(
wont_exit
);
if
(
rtl_dump_file
)
fprintf
(
rtl_dump_file
,
";; Unrolled loop %d times, %i insns
\n
"
,
nunroll
,
num_loop_insns
(
loop
));
}
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment