blob: 4230f62051b27b1990988aa49c71c64da06b99b7 [file] [log] [blame]
/* Conversion of SESE regions to Polyhedra.
Copyright (C) 2009-2015 Free Software Foundation, Inc.
Contributed by Sebastian Pop <sebastian.pop@amd.com>.
This file is part of GCC.
GCC is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3, or (at your option)
any later version.
GCC is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with GCC; see the file COPYING3. If not see
<http://www.gnu.org/licenses/>. */
#include "config.h"
#ifdef HAVE_isl
#include <isl/constraint.h>
#include <isl/set.h>
#include <isl/map.h>
#include <isl/union_map.h>
#include <isl/constraint.h>
#include <isl/aff.h>
#include <isl/val.h>
/* Since ISL-0.13, the extern is in val_gmp.h. */
#if !defined(HAVE_ISL_SCHED_CONSTRAINTS_COMPUTE_SCHEDULE) && defined(__cplusplus)
extern "C" {
#endif
#include <isl/val_gmp.h>
#if !defined(HAVE_ISL_SCHED_CONSTRAINTS_COMPUTE_SCHEDULE) && defined(__cplusplus)
}
#endif
#endif
#include "system.h"
#include "coretypes.h"
#include "hash-set.h"
#include "machmode.h"
#include "vec.h"
#include "double-int.h"
#include "input.h"
#include "alias.h"
#include "symtab.h"
#include "options.h"
#include "wide-int.h"
#include "inchash.h"
#include "tree.h"
#include "fold-const.h"
#include "predict.h"
#include "tm.h"
#include "hard-reg-set.h"
#include "function.h"
#include "dominance.h"
#include "cfg.h"
#include "basic-block.h"
#include "tree-ssa-alias.h"
#include "internal-fn.h"
#include "gimple-expr.h"
#include "is-a.h"
#include "gimple.h"
#include "gimple-iterator.h"
#include "gimplify.h"
#include "gimplify-me.h"
#include "gimple-ssa.h"
#include "tree-cfg.h"
#include "tree-phinodes.h"
#include "ssa-iterators.h"
#include "stringpool.h"
#include "tree-ssanames.h"
#include "tree-ssa-loop-manip.h"
#include "tree-ssa-loop-niter.h"
#include "tree-ssa-loop.h"
#include "tree-into-ssa.h"
#include "tree-pass.h"
#include "cfgloop.h"
#include "tree-chrec.h"
#include "tree-data-ref.h"
#include "tree-scalar-evolution.h"
#include "domwalk.h"
#include "sese.h"
#include "tree-ssa-propagate.h"
#ifdef HAVE_isl
#include "hashtab.h"
#include "rtl.h"
#include "flags.h"
#include "statistics.h"
#include "real.h"
#include "fixed-value.h"
#include "insn-config.h"
#include "expmed.h"
#include "dojump.h"
#include "explow.h"
#include "calls.h"
#include "emit-rtl.h"
#include "varasm.h"
#include "stmt.h"
#include "expr.h"
#include "graphite-poly.h"
#include "graphite-sese-to-poly.h"
/* Assigns to RES the value of the INTEGER_CST T. */
static inline void
tree_int_to_gmp (tree t, mpz_t res)
{
wi::to_mpz (t, res, TYPE_SIGN (TREE_TYPE (t)));
}
/* Returns the index of the PHI argument defined in the outermost
loop. */
static size_t
phi_arg_in_outermost_loop (gphi *phi)
{
loop_p loop = gimple_bb (phi)->loop_father;
size_t i, res = 0;
for (i = 0; i < gimple_phi_num_args (phi); i++)
if (!flow_bb_inside_loop_p (loop, gimple_phi_arg_edge (phi, i)->src))
{
loop = gimple_phi_arg_edge (phi, i)->src->loop_father;
res = i;
}
return res;
}
/* Removes a simple copy phi node "RES = phi (INIT, RES)" at position
PSI by inserting on the loop ENTRY edge assignment "RES = INIT". */
static void
remove_simple_copy_phi (gphi_iterator *psi)
{
gphi *phi = psi->phi ();
tree res = gimple_phi_result (phi);
size_t entry = phi_arg_in_outermost_loop (phi);
tree init = gimple_phi_arg_def (phi, entry);
gassign *stmt = gimple_build_assign (res, init);
edge e = gimple_phi_arg_edge (phi, entry);
remove_phi_node (psi, false);
gsi_insert_on_edge_immediate (e, stmt);
}
/* Removes an invariant phi node at position PSI by inserting on the
loop ENTRY edge the assignment RES = INIT. */
static void
remove_invariant_phi (sese region, gphi_iterator *psi)
{
gphi *phi = psi->phi ();
loop_p loop = loop_containing_stmt (phi);
tree res = gimple_phi_result (phi);
tree scev = scalar_evolution_in_region (region, loop, res);
size_t entry = phi_arg_in_outermost_loop (phi);
edge e = gimple_phi_arg_edge (phi, entry);
tree var;
gassign *stmt;
gimple_seq stmts = NULL;
if (tree_contains_chrecs (scev, NULL))
scev = gimple_phi_arg_def (phi, entry);
var = force_gimple_operand (scev, &stmts, true, NULL_TREE);
stmt = gimple_build_assign (res, var);
remove_phi_node (psi, false);
gimple_seq_add_stmt (&stmts, stmt);
gsi_insert_seq_on_edge (e, stmts);
gsi_commit_edge_inserts ();
SSA_NAME_DEF_STMT (res) = stmt;
}
/* Returns true when the phi node at PSI is of the form "a = phi (a, x)". */
static inline bool
simple_copy_phi_p (gphi *phi)
{
tree res;
if (gimple_phi_num_args (phi) != 2)
return false;
res = gimple_phi_result (phi);
return (res == gimple_phi_arg_def (phi, 0)
|| res == gimple_phi_arg_def (phi, 1));
}
/* Returns true when the phi node at position PSI is a reduction phi
node in REGION. Otherwise moves the pointer PSI to the next phi to
be considered. */
static bool
reduction_phi_p (sese region, gphi_iterator *psi)
{
loop_p loop;
gphi *phi = psi->phi ();
tree res = gimple_phi_result (phi);
loop = loop_containing_stmt (phi);
if (simple_copy_phi_p (phi))
{
/* PRE introduces phi nodes like these, for an example,
see id-5.f in the fortran graphite testsuite:
# prephitmp.85_265 = PHI <prephitmp.85_258(33), prephitmp.85_265(18)>
*/
remove_simple_copy_phi (psi);
return false;
}
if (scev_analyzable_p (res, region))
{
tree scev = scalar_evolution_in_region (region, loop, res);
if (evolution_function_is_invariant_p (scev, loop->num))
remove_invariant_phi (region, psi);
else
gsi_next (psi);
return false;
}
/* All the other cases are considered reductions. */
return true;
}
/* Store the GRAPHITE representation of BB. */
static gimple_bb_p
new_gimple_bb (basic_block bb, vec<data_reference_p> drs)
{
struct gimple_bb *gbb;
gbb = XNEW (struct gimple_bb);
bb->aux = gbb;
GBB_BB (gbb) = bb;
GBB_DATA_REFS (gbb) = drs;
GBB_CONDITIONS (gbb).create (0);
GBB_CONDITION_CASES (gbb).create (0);
return gbb;
}
static void
free_data_refs_aux (vec<data_reference_p> datarefs)
{
unsigned int i;
struct data_reference *dr;
FOR_EACH_VEC_ELT (datarefs, i, dr)
if (dr->aux)
{
base_alias_pair *bap = (base_alias_pair *)(dr->aux);
free (bap->alias_set);
free (bap);
dr->aux = NULL;
}
}
/* Frees GBB. */
static void
free_gimple_bb (struct gimple_bb *gbb)
{
free_data_refs_aux (GBB_DATA_REFS (gbb));
free_data_refs (GBB_DATA_REFS (gbb));
GBB_CONDITIONS (gbb).release ();
GBB_CONDITION_CASES (gbb).release ();
GBB_BB (gbb)->aux = 0;
XDELETE (gbb);
}
/* Deletes all gimple bbs in SCOP. */
static void
remove_gbbs_in_scop (scop_p scop)
{
int i;
poly_bb_p pbb;
FOR_EACH_VEC_ELT (SCOP_BBS (scop), i, pbb)
free_gimple_bb (PBB_BLACK_BOX (pbb));
}
/* Deletes all scops in SCOPS. */
void
free_scops (vec<scop_p> scops)
{
int i;
scop_p scop;
FOR_EACH_VEC_ELT (scops, i, scop)
{
remove_gbbs_in_scop (scop);
free_sese (SCOP_REGION (scop));
free_scop (scop);
}
scops.release ();
}
/* Same as outermost_loop_in_sese, returns the outermost loop
containing BB in REGION, but makes sure that the returned loop
belongs to the REGION, and so this returns the first loop in the
REGION when the loop containing BB does not belong to REGION. */
static loop_p
outermost_loop_in_sese_1 (sese region, basic_block bb)
{
loop_p nest = outermost_loop_in_sese (region, bb);
if (loop_in_sese_p (nest, region))
return nest;
/* When the basic block BB does not belong to a loop in the region,
return the first loop in the region. */
nest = nest->inner;
while (nest)
if (loop_in_sese_p (nest, region))
break;
else
nest = nest->next;
gcc_assert (nest);
return nest;
}
/* Generates a polyhedral black box only if the bb contains interesting
information. */
static gimple_bb_p
try_generate_gimple_bb (scop_p scop, basic_block bb)
{
vec<data_reference_p> drs;
drs.create (5);
sese region = SCOP_REGION (scop);
loop_p nest = outermost_loop_in_sese_1 (region, bb);
gimple_stmt_iterator gsi;
for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
{
gimple stmt = gsi_stmt (gsi);
loop_p loop;
if (is_gimple_debug (stmt))
continue;
loop = loop_containing_stmt (stmt);
if (!loop_in_sese_p (loop, region))
loop = nest;
graphite_find_data_references_in_stmt (nest, loop, stmt, &drs);
}
return new_gimple_bb (bb, drs);
}
/* Returns true if all predecessors of BB, that are not dominated by BB, are
marked in MAP. The predecessors dominated by BB are loop latches and will
be handled after BB. */
static bool
all_non_dominated_preds_marked_p (basic_block bb, sbitmap map)
{
edge e;
edge_iterator ei;
FOR_EACH_EDGE (e, ei, bb->preds)
if (!bitmap_bit_p (map, e->src->index)
&& !dominated_by_p (CDI_DOMINATORS, e->src, bb))
return false;
return true;
}
/* Compare the depth of two basic_block's P1 and P2. */
static int
compare_bb_depths (const void *p1, const void *p2)
{
const_basic_block const bb1 = *(const_basic_block const*)p1;
const_basic_block const bb2 = *(const_basic_block const*)p2;
int d1 = loop_depth (bb1->loop_father);
int d2 = loop_depth (bb2->loop_father);
if (d1 < d2)
return 1;
if (d1 > d2)
return -1;
return 0;
}
/* Sort the basic blocks from DOM such that the first are the ones at
a deepest loop level. */
static void
graphite_sort_dominated_info (vec<basic_block> dom)
{
dom.qsort (compare_bb_depths);
}
/* Recursive helper function for build_scops_bbs. */
static void
build_scop_bbs_1 (scop_p scop, sbitmap visited, basic_block bb)
{
sese region = SCOP_REGION (scop);
vec<basic_block> dom;
poly_bb_p pbb;
if (bitmap_bit_p (visited, bb->index)
|| !bb_in_sese_p (bb, region))
return;
pbb = new_poly_bb (scop, try_generate_gimple_bb (scop, bb));
SCOP_BBS (scop).safe_push (pbb);
bitmap_set_bit (visited, bb->index);
dom = get_dominated_by (CDI_DOMINATORS, bb);
if (!dom.exists ())
return;
graphite_sort_dominated_info (dom);
while (!dom.is_empty ())
{
int i;
basic_block dom_bb;
FOR_EACH_VEC_ELT (dom, i, dom_bb)
if (all_non_dominated_preds_marked_p (dom_bb, visited))
{
build_scop_bbs_1 (scop, visited, dom_bb);
dom.unordered_remove (i);
break;
}
}
dom.release ();
}
/* Gather the basic blocks belonging to the SCOP. */
static void
build_scop_bbs (scop_p scop)
{
sbitmap visited = sbitmap_alloc (last_basic_block_for_fn (cfun));
sese region = SCOP_REGION (scop);
bitmap_clear (visited);
build_scop_bbs_1 (scop, visited, SESE_ENTRY_BB (region));
sbitmap_free (visited);
}
/* Return an ISL identifier for the polyhedral basic block PBB. */
static isl_id *
isl_id_for_pbb (scop_p s, poly_bb_p pbb)
{
char name[50];
snprintf (name, sizeof (name), "S_%d", pbb_index (pbb));
return isl_id_alloc (s->ctx, name, pbb);
}
/* Converts the STATIC_SCHEDULE of PBB into a scattering polyhedron.
We generate SCATTERING_DIMENSIONS scattering dimensions.
CLooG 0.15.0 and previous versions require, that all
scattering functions of one CloogProgram have the same number of
scattering dimensions, therefore we allow to specify it. This
should be removed in future versions of CLooG.
The scattering polyhedron consists of these dimensions: scattering,
loop_iterators, parameters.
Example:
| scattering_dimensions = 5
| used_scattering_dimensions = 3
| nb_iterators = 1
| scop_nb_params = 2
|
| Schedule:
| i
| 4 5
|
| Scattering polyhedron:
|
| scattering: {s1, s2, s3, s4, s5}
| loop_iterators: {i}
| parameters: {p1, p2}
|
| s1 s2 s3 s4 s5 i p1 p2 1
| 1 0 0 0 0 0 0 0 -4 = 0
| 0 1 0 0 0 -1 0 0 0 = 0
| 0 0 1 0 0 0 0 0 -5 = 0 */
static void
build_pbb_scattering_polyhedrons (isl_aff *static_sched,
poly_bb_p pbb, int scattering_dimensions)
{
int i;
int nb_iterators = pbb_dim_iter_domain (pbb);
int used_scattering_dimensions = nb_iterators * 2 + 1;
isl_val *val;
isl_space *dc, *dm;
gcc_assert (scattering_dimensions >= used_scattering_dimensions);
dc = isl_set_get_space (pbb->domain);
dm = isl_space_add_dims (isl_space_from_domain (dc),
isl_dim_out, scattering_dimensions);
pbb->schedule = isl_map_universe (dm);
for (i = 0; i < scattering_dimensions; i++)
{
/* Textual order inside this loop. */
if ((i % 2) == 0)
{
isl_constraint *c = isl_equality_alloc
(isl_local_space_from_space (isl_map_get_space (pbb->schedule)));
val = isl_aff_get_coefficient_val (static_sched, isl_dim_in, i / 2);
val = isl_val_neg (val);
c = isl_constraint_set_constant_val (c, val);
c = isl_constraint_set_coefficient_si (c, isl_dim_out, i, 1);
pbb->schedule = isl_map_add_constraint (pbb->schedule, c);
}
/* Iterations of this loop. */
else /* if ((i % 2) == 1) */
{
int loop = (i - 1) / 2;
pbb->schedule = isl_map_equate (pbb->schedule, isl_dim_in, loop,
isl_dim_out, i);
}
}
pbb->transformed = isl_map_copy (pbb->schedule);
}
/* Build for BB the static schedule.
The static schedule is a Dewey numbering of the abstract syntax
tree: http://en.wikipedia.org/wiki/Dewey_Decimal_Classification
The following example informally defines the static schedule:
A
for (i: ...)
{
for (j: ...)
{
B
C
}
for (k: ...)
{
D
E
}
}
F
Static schedules for A to F:
DEPTH
0 1 2
A 0
B 1 0 0
C 1 0 1
D 1 1 0
E 1 1 1
F 2
*/
static void
build_scop_scattering (scop_p scop)
{
int i;
poly_bb_p pbb;
gimple_bb_p previous_gbb = NULL;
isl_space *dc = isl_set_get_space (scop->context);
isl_aff *static_sched;
dc = isl_space_add_dims (dc, isl_dim_set, number_of_loops (cfun));
static_sched = isl_aff_zero_on_domain (isl_local_space_from_space (dc));
/* We have to start schedules at 0 on the first component and
because we cannot compare_prefix_loops against a previous loop,
prefix will be equal to zero, and that index will be
incremented before copying. */
static_sched = isl_aff_add_coefficient_si (static_sched, isl_dim_in, 0, -1);
FOR_EACH_VEC_ELT (SCOP_BBS (scop), i, pbb)
{
gimple_bb_p gbb = PBB_BLACK_BOX (pbb);
int prefix;
int nb_scat_dims = pbb_dim_iter_domain (pbb) * 2 + 1;
if (previous_gbb)
prefix = nb_common_loops (SCOP_REGION (scop), previous_gbb, gbb);
else
prefix = 0;
previous_gbb = gbb;
static_sched = isl_aff_add_coefficient_si (static_sched, isl_dim_in,
prefix, 1);
build_pbb_scattering_polyhedrons (static_sched, pbb, nb_scat_dims);
}
isl_aff_free (static_sched);
}
static isl_pw_aff *extract_affine (scop_p, tree, __isl_take isl_space *space);
/* Extract an affine expression from the chain of recurrence E. */
static isl_pw_aff *
extract_affine_chrec (scop_p s, tree e, __isl_take isl_space *space)
{
isl_pw_aff *lhs = extract_affine (s, CHREC_LEFT (e), isl_space_copy (space));
isl_pw_aff *rhs = extract_affine (s, CHREC_RIGHT (e), isl_space_copy (space));
isl_local_space *ls = isl_local_space_from_space (space);
unsigned pos = sese_loop_depth ((sese) s->region, get_chrec_loop (e)) - 1;
isl_aff *loop = isl_aff_set_coefficient_si
(isl_aff_zero_on_domain (ls), isl_dim_in, pos, 1);
isl_pw_aff *l = isl_pw_aff_from_aff (loop);
/* Before multiplying, make sure that the result is affine. */
gcc_assert (isl_pw_aff_is_cst (rhs)
|| isl_pw_aff_is_cst (l));
return isl_pw_aff_add (lhs, isl_pw_aff_mul (rhs, l));
}
/* Extract an affine expression from the mult_expr E. */
static isl_pw_aff *
extract_affine_mul (scop_p s, tree e, __isl_take isl_space *space)
{
isl_pw_aff *lhs = extract_affine (s, TREE_OPERAND (e, 0),
isl_space_copy (space));
isl_pw_aff *rhs = extract_affine (s, TREE_OPERAND (e, 1), space);
if (!isl_pw_aff_is_cst (lhs)
&& !isl_pw_aff_is_cst (rhs))
{
isl_pw_aff_free (lhs);
isl_pw_aff_free (rhs);
return NULL;
}
return isl_pw_aff_mul (lhs, rhs);
}
/* Return an ISL identifier from the name of the ssa_name E. */
static isl_id *
isl_id_for_ssa_name (scop_p s, tree e)
{
const char *name = get_name (e);
isl_id *id;
if (name)
id = isl_id_alloc (s->ctx, name, e);
else
{
char name1[50];
snprintf (name1, sizeof (name1), "P_%d", SSA_NAME_VERSION (e));
id = isl_id_alloc (s->ctx, name1, e);
}
return id;
}
/* Return an ISL identifier for the data reference DR. */
static isl_id *
isl_id_for_dr (scop_p s, data_reference_p dr ATTRIBUTE_UNUSED)
{
/* Data references all get the same isl_id. They need to be comparable
and are distinguished through the first dimension, which contains the
alias set number. */
return isl_id_alloc (s->ctx, "", 0);
}
/* Extract an affine expression from the ssa_name E. */
static isl_pw_aff *
extract_affine_name (scop_p s, tree e, __isl_take isl_space *space)
{
isl_aff *aff;
isl_set *dom;
isl_id *id;
int dimension;
id = isl_id_for_ssa_name (s, e);
dimension = isl_space_find_dim_by_id (space, isl_dim_param, id);
isl_id_free (id);
dom = isl_set_universe (isl_space_copy (space));
aff = isl_aff_zero_on_domain (isl_local_space_from_space (space));
aff = isl_aff_add_coefficient_si (aff, isl_dim_param, dimension, 1);
return isl_pw_aff_alloc (dom, aff);
}
/* Extract an affine expression from the gmp constant G. */
static isl_pw_aff *
extract_affine_gmp (mpz_t g, __isl_take isl_space *space)
{
isl_local_space *ls = isl_local_space_from_space (isl_space_copy (space));
isl_aff *aff = isl_aff_zero_on_domain (ls);
isl_set *dom = isl_set_universe (space);
isl_val *v;
isl_ctx *ct;
ct = isl_aff_get_ctx (aff);
v = isl_val_int_from_gmp (ct, g);
aff = isl_aff_add_constant_val (aff, v);
return isl_pw_aff_alloc (dom, aff);
}
/* Extract an affine expression from the integer_cst E. */
static isl_pw_aff *
extract_affine_int (tree e, __isl_take isl_space *space)
{
isl_pw_aff *res;
mpz_t g;
mpz_init (g);
tree_int_to_gmp (e, g);
res = extract_affine_gmp (g, space);
mpz_clear (g);
return res;
}
/* Compute pwaff mod 2^width. */
extern isl_ctx *the_isl_ctx;
static isl_pw_aff *
wrap (isl_pw_aff *pwaff, unsigned width)
{
isl_val *mod;
mod = isl_val_int_from_ui(the_isl_ctx, width);
mod = isl_val_2exp (mod);
pwaff = isl_pw_aff_mod_val (pwaff, mod);
return pwaff;
}
/* When parameter NAME is in REGION, returns its index in SESE_PARAMS.
Otherwise returns -1. */
static inline int
parameter_index_in_region_1 (tree name, sese region)
{
int i;
tree p;
gcc_assert (TREE_CODE (name) == SSA_NAME);
FOR_EACH_VEC_ELT (SESE_PARAMS (region), i, p)
if (p == name)
return i;
return -1;
}
/* When the parameter NAME is in REGION, returns its index in
SESE_PARAMS. Otherwise this function inserts NAME in SESE_PARAMS
and returns the index of NAME. */
static int
parameter_index_in_region (tree name, sese region)
{
int i;
gcc_assert (TREE_CODE (name) == SSA_NAME);
i = parameter_index_in_region_1 (name, region);
if (i != -1)
return i;
gcc_assert (SESE_ADD_PARAMS (region));
i = SESE_PARAMS (region).length ();
SESE_PARAMS (region).safe_push (name);
return i;
}
/* Extract an affine expression from the tree E in the scop S. */
static isl_pw_aff *
extract_affine (scop_p s, tree e, __isl_take isl_space *space)
{
isl_pw_aff *lhs, *rhs, *res;
tree type;
if (e == chrec_dont_know) {
isl_space_free (space);
return NULL;
}
switch (TREE_CODE (e))
{
case POLYNOMIAL_CHREC:
res = extract_affine_chrec (s, e, space);
break;
case MULT_EXPR:
res = extract_affine_mul (s, e, space);
break;
case PLUS_EXPR:
case POINTER_PLUS_EXPR:
lhs = extract_affine (s, TREE_OPERAND (e, 0), isl_space_copy (space));
rhs = extract_affine (s, TREE_OPERAND (e, 1), space);
res = isl_pw_aff_add (lhs, rhs);
break;
case MINUS_EXPR:
lhs = extract_affine (s, TREE_OPERAND (e, 0), isl_space_copy (space));
rhs = extract_affine (s, TREE_OPERAND (e, 1), space);
res = isl_pw_aff_sub (lhs, rhs);
break;
case NEGATE_EXPR:
case BIT_NOT_EXPR:
lhs = extract_affine (s, TREE_OPERAND (e, 0), isl_space_copy (space));
rhs = extract_affine (s, integer_minus_one_node, space);
res = isl_pw_aff_mul (lhs, rhs);
break;
case SSA_NAME:
gcc_assert (-1 != parameter_index_in_region_1 (e, SCOP_REGION (s)));
res = extract_affine_name (s, e, space);
break;
case INTEGER_CST:
res = extract_affine_int (e, space);
/* No need to wrap a single integer. */
return res;
CASE_CONVERT:
case NON_LVALUE_EXPR:
res = extract_affine (s, TREE_OPERAND (e, 0), space);
break;
default:
gcc_unreachable ();
break;
}
type = TREE_TYPE (e);
if (TYPE_UNSIGNED (type))
res = wrap (res, TYPE_PRECISION (type));
return res;
}
/* In the context of sese S, scan the expression E and translate it to
a linear expression C. When parsing a symbolic multiplication, K
represents the constant multiplier of an expression containing
parameters. */
static void
scan_tree_for_params (sese s, tree e)
{
if (e == chrec_dont_know)
return;
switch (TREE_CODE (e))
{
case POLYNOMIAL_CHREC:
scan_tree_for_params (s, CHREC_LEFT (e));
break;
case MULT_EXPR:
if (chrec_contains_symbols (TREE_OPERAND (e, 0)))
scan_tree_for_params (s, TREE_OPERAND (e, 0));
else
scan_tree_for_params (s, TREE_OPERAND (e, 1));
break;
case PLUS_EXPR:
case POINTER_PLUS_EXPR:
case MINUS_EXPR:
scan_tree_for_params (s, TREE_OPERAND (e, 0));
scan_tree_for_params (s, TREE_OPERAND (e, 1));
break;
case NEGATE_EXPR:
case BIT_NOT_EXPR:
CASE_CONVERT:
case NON_LVALUE_EXPR:
scan_tree_for_params (s, TREE_OPERAND (e, 0));
break;
case SSA_NAME:
parameter_index_in_region (e, s);
break;
case INTEGER_CST:
case ADDR_EXPR:
break;
default:
gcc_unreachable ();
break;
}
}
/* Find parameters with respect to REGION in BB. We are looking in memory
access functions, conditions and loop bounds. */
static void
find_params_in_bb (sese region, gimple_bb_p gbb)
{
int i;
unsigned j;
data_reference_p dr;
gimple stmt;
loop_p loop = GBB_BB (gbb)->loop_father;
/* Find parameters in the access functions of data references. */
FOR_EACH_VEC_ELT (GBB_DATA_REFS (gbb), i, dr)
for (j = 0; j < DR_NUM_DIMENSIONS (dr); j++)
scan_tree_for_params (region, DR_ACCESS_FN (dr, j));
/* Find parameters in conditional statements. */
FOR_EACH_VEC_ELT (GBB_CONDITIONS (gbb), i, stmt)
{
tree lhs = scalar_evolution_in_region (region, loop,
gimple_cond_lhs (stmt));
tree rhs = scalar_evolution_in_region (region, loop,
gimple_cond_rhs (stmt));
scan_tree_for_params (region, lhs);
scan_tree_for_params (region, rhs);
}
}
/* Record the parameters used in the SCOP. A variable is a parameter
in a scop if it does not vary during the execution of that scop. */
static void
find_scop_parameters (scop_p scop)
{
poly_bb_p pbb;
unsigned i;
sese region = SCOP_REGION (scop);
struct loop *loop;
int nbp;
/* Find the parameters used in the loop bounds. */
FOR_EACH_VEC_ELT (SESE_LOOP_NEST (region), i, loop)
{
tree nb_iters = number_of_latch_executions (loop);
if (!chrec_contains_symbols (nb_iters))
continue;
nb_iters = scalar_evolution_in_region (region, loop, nb_iters);
scan_tree_for_params (region, nb_iters);
}
/* Find the parameters used in data accesses. */
FOR_EACH_VEC_ELT (SCOP_BBS (scop), i, pbb)
find_params_in_bb (region, PBB_BLACK_BOX (pbb));
nbp = sese_nb_params (region);
scop_set_nb_params (scop, nbp);
SESE_ADD_PARAMS (region) = false;
{
tree e;
isl_space *space = isl_space_set_alloc (scop->ctx, nbp, 0);
FOR_EACH_VEC_ELT (SESE_PARAMS (region), i, e)
space = isl_space_set_dim_id (space, isl_dim_param, i,
isl_id_for_ssa_name (scop, e));
scop->context = isl_set_universe (space);
}
}
/* Builds the constraint polyhedra for LOOP in SCOP. OUTER_PH gives
the constraints for the surrounding loops. */
static void
build_loop_iteration_domains (scop_p scop, struct loop *loop,
int nb,
isl_set *outer, isl_set **doms)
{
tree nb_iters = number_of_latch_executions (loop);
sese region = SCOP_REGION (scop);
isl_set *inner = isl_set_copy (outer);
isl_space *space;
isl_constraint *c;
int pos = isl_set_dim (outer, isl_dim_set);
isl_val *v;
mpz_t g;
mpz_init (g);
inner = isl_set_add_dims (inner, isl_dim_set, 1);
space = isl_set_get_space (inner);
/* 0 <= loop_i */
c = isl_inequality_alloc
(isl_local_space_from_space (isl_space_copy (space)));
c = isl_constraint_set_coefficient_si (c, isl_dim_set, pos, 1);
inner = isl_set_add_constraint (inner, c);
/* loop_i <= cst_nb_iters */
if (TREE_CODE (nb_iters) == INTEGER_CST)
{
c = isl_inequality_alloc
(isl_local_space_from_space (isl_space_copy (space)));
c = isl_constraint_set_coefficient_si (c, isl_dim_set, pos, -1);
tree_int_to_gmp (nb_iters, g);
v = isl_val_int_from_gmp (the_isl_ctx, g);
c = isl_constraint_set_constant_val (c, v);
inner = isl_set_add_constraint (inner, c);
}
/* loop_i <= expr_nb_iters */
else if (!chrec_contains_undetermined (nb_iters))
{
widest_int nit;
isl_pw_aff *aff;
isl_set *valid;
isl_local_space *ls;
isl_aff *al;
isl_set *le;
nb_iters = scalar_evolution_in_region (region, loop, nb_iters);
aff = extract_affine (scop, nb_iters, isl_set_get_space (inner));
valid = isl_pw_aff_nonneg_set (isl_pw_aff_copy (aff));
valid = isl_set_project_out (valid, isl_dim_set, 0,
isl_set_dim (valid, isl_dim_set));
scop->context = isl_set_intersect (scop->context, valid);
ls = isl_local_space_from_space (isl_space_copy (space));
al = isl_aff_set_coefficient_si (isl_aff_zero_on_domain (ls),
isl_dim_in, pos, 1);
le = isl_pw_aff_le_set (isl_pw_aff_from_aff (al),
isl_pw_aff_copy (aff));
inner = isl_set_intersect (inner, le);
if (max_stmt_executions (loop, &nit))
{
/* Insert in the context the constraints from the
estimation of the number of iterations NIT and the
symbolic number of iterations (involving parameter
names) NB_ITERS. First, build the affine expression
"NIT - NB_ITERS" and then say that it is positive,
i.e., NIT approximates NB_ITERS: "NIT >= NB_ITERS". */
isl_pw_aff *approx;
mpz_t g;
isl_set *x;
isl_constraint *c;
mpz_init (g);
wi::to_mpz (nit, g, SIGNED);
mpz_sub_ui (g, g, 1);
approx = extract_affine_gmp (g, isl_set_get_space (inner));
x = isl_pw_aff_ge_set (approx, aff);
x = isl_set_project_out (x, isl_dim_set, 0,
isl_set_dim (x, isl_dim_set));
scop->context = isl_set_intersect (scop->context, x);
c = isl_inequality_alloc
(isl_local_space_from_space (isl_space_copy (space)));
c = isl_constraint_set_coefficient_si (c, isl_dim_set, pos, -1);
v = isl_val_int_from_gmp (the_isl_ctx, g);
mpz_clear (g);
c = isl_constraint_set_constant_val (c, v);
inner = isl_set_add_constraint (inner, c);
}
else
isl_pw_aff_free (aff);
}
else
gcc_unreachable ();
if (loop->inner && loop_in_sese_p (loop->inner, region))
build_loop_iteration_domains (scop, loop->inner, nb + 1,
isl_set_copy (inner), doms);
if (nb != 0
&& loop->next
&& loop_in_sese_p (loop->next, region))
build_loop_iteration_domains (scop, loop->next, nb,
isl_set_copy (outer), doms);
doms[loop->num] = inner;
isl_set_free (outer);
isl_space_free (space);
mpz_clear (g);
}
/* Returns a linear expression for tree T evaluated in PBB. */
static isl_pw_aff *
create_pw_aff_from_tree (poly_bb_p pbb, tree t)
{
scop_p scop = PBB_SCOP (pbb);
t = scalar_evolution_in_region (SCOP_REGION (scop), pbb_loop (pbb), t);
gcc_assert (!automatically_generated_chrec_p (t));
return extract_affine (scop, t, isl_set_get_space (pbb->domain));
}
/* Add conditional statement STMT to pbb. CODE is used as the comparison
operator. This allows us to invert the condition or to handle
inequalities. */
static void
add_condition_to_pbb (poly_bb_p pbb, gcond *stmt, enum tree_code code)
{
isl_pw_aff *lhs = create_pw_aff_from_tree (pbb, gimple_cond_lhs (stmt));
isl_pw_aff *rhs = create_pw_aff_from_tree (pbb, gimple_cond_rhs (stmt));
isl_set *cond;
switch (code)
{
case LT_EXPR:
cond = isl_pw_aff_lt_set (lhs, rhs);
break;
case GT_EXPR:
cond = isl_pw_aff_gt_set (lhs, rhs);
break;
case LE_EXPR:
cond = isl_pw_aff_le_set (lhs, rhs);
break;
case GE_EXPR:
cond = isl_pw_aff_ge_set (lhs, rhs);
break;
case EQ_EXPR:
cond = isl_pw_aff_eq_set (lhs, rhs);
break;
case NE_EXPR:
cond = isl_pw_aff_ne_set (lhs, rhs);
break;
default:
isl_pw_aff_free (lhs);
isl_pw_aff_free (rhs);
return;
}
cond = isl_set_coalesce (cond);
cond = isl_set_set_tuple_id (cond, isl_set_get_tuple_id (pbb->domain));
pbb->domain = isl_set_intersect (pbb->domain, cond);
}
/* Add conditions to the domain of PBB. */
static void
add_conditions_to_domain (poly_bb_p pbb)
{
unsigned int i;
gimple stmt;
gimple_bb_p gbb = PBB_BLACK_BOX (pbb);
if (GBB_CONDITIONS (gbb).is_empty ())
return;
FOR_EACH_VEC_ELT (GBB_CONDITIONS (gbb), i, stmt)
switch (gimple_code (stmt))
{
case GIMPLE_COND:
{
gcond *cond_stmt = as_a <gcond *> (stmt);
enum tree_code code = gimple_cond_code (cond_stmt);
/* The conditions for ELSE-branches are inverted. */
if (!GBB_CONDITION_CASES (gbb)[i])
code = invert_tree_comparison (code, false);
add_condition_to_pbb (pbb, cond_stmt, code);
break;
}
case GIMPLE_SWITCH:
/* Switch statements are not supported right now - fall through. */
default:
gcc_unreachable ();
break;
}
}
/* Traverses all the GBBs of the SCOP and add their constraints to the
iteration domains. */
static void
add_conditions_to_constraints (scop_p scop)
{
int i;
poly_bb_p pbb;
FOR_EACH_VEC_ELT (SCOP_BBS (scop), i, pbb)
add_conditions_to_domain (pbb);
}
/* Returns a COND_EXPR statement when BB has a single predecessor, the
edge between BB and its predecessor is not a loop exit edge, and
the last statement of the single predecessor is a COND_EXPR. */
static gcond *
single_pred_cond_non_loop_exit (basic_block bb)
{
if (single_pred_p (bb))
{
edge e = single_pred_edge (bb);
basic_block pred = e->src;
gimple stmt;
if (loop_depth (pred->loop_father) > loop_depth (bb->loop_father))
return NULL;
stmt = last_stmt (pred);
if (stmt && gimple_code (stmt) == GIMPLE_COND)
return as_a <gcond *> (stmt);
}
return NULL;
}
class sese_dom_walker : public dom_walker
{
public:
sese_dom_walker (cdi_direction, sese);
virtual void before_dom_children (basic_block);
virtual void after_dom_children (basic_block);
private:
auto_vec<gimple, 3> m_conditions, m_cases;
sese m_region;
};
sese_dom_walker::sese_dom_walker (cdi_direction direction, sese region)
: dom_walker (direction), m_region (region)
{
}
/* Call-back for dom_walk executed before visiting the dominated
blocks. */
void
sese_dom_walker::before_dom_children (basic_block bb)
{
gimple_bb_p gbb;
gcond *stmt;
if (!bb_in_sese_p (bb, m_region))
return;
stmt = single_pred_cond_non_loop_exit (bb);
if (stmt)
{
edge e = single_pred_edge (bb);
m_conditions.safe_push (stmt);
if (e->flags & EDGE_TRUE_VALUE)
m_cases.safe_push (stmt);
else
m_cases.safe_push (NULL);
}
gbb = gbb_from_bb (bb);
if (gbb)
{
GBB_CONDITIONS (gbb) = m_conditions.copy ();
GBB_CONDITION_CASES (gbb) = m_cases.copy ();
}
}
/* Call-back for dom_walk executed after visiting the dominated
blocks. */
void
sese_dom_walker::after_dom_children (basic_block bb)
{
if (!bb_in_sese_p (bb, m_region))
return;
if (single_pred_cond_non_loop_exit (bb))
{
m_conditions.pop ();
m_cases.pop ();
}
}
/* Add constraints on the possible values of parameter P from the type
of P. */
static void
add_param_constraints (scop_p scop, graphite_dim_t p)
{
tree parameter = SESE_PARAMS (SCOP_REGION (scop))[p];
tree type = TREE_TYPE (parameter);
tree lb = NULL_TREE;
tree ub = NULL_TREE;
if (POINTER_TYPE_P (type) || !TYPE_MIN_VALUE (type))
lb = lower_bound_in_type (type, type);
else
lb = TYPE_MIN_VALUE (type);
if (POINTER_TYPE_P (type) || !TYPE_MAX_VALUE (type))
ub = upper_bound_in_type (type, type);
else
ub = TYPE_MAX_VALUE (type);
if (lb)
{
isl_space *space = isl_set_get_space (scop->context);
isl_constraint *c;
mpz_t g;
isl_val *v;
c = isl_inequality_alloc (isl_local_space_from_space (space));
mpz_init (g);
tree_int_to_gmp (lb, g);
v = isl_val_int_from_gmp (the_isl_ctx, g);
v = isl_val_neg (v);
mpz_clear (g);
c = isl_constraint_set_constant_val (c, v);
c = isl_constraint_set_coefficient_si (c, isl_dim_param, p, 1);
scop->context = isl_set_add_constraint (scop->context, c);
}
if (ub)
{
isl_space *space = isl_set_get_space (scop->context);
isl_constraint *c;
mpz_t g;
isl_val *v;
c = isl_inequality_alloc (isl_local_space_from_space (space));
mpz_init (g);
tree_int_to_gmp (ub, g);
v = isl_val_int_from_gmp (the_isl_ctx, g);
mpz_clear (g);
c = isl_constraint_set_constant_val (c, v);
c = isl_constraint_set_coefficient_si (c, isl_dim_param, p, -1);
scop->context = isl_set_add_constraint (scop->context, c);
}
}
/* Build the context of the SCOP. The context usually contains extra
constraints that are added to the iteration domains that constrain
some parameters. */
static void
build_scop_context (scop_p scop)
{
graphite_dim_t p, n = scop_nb_params (scop);
for (p = 0; p < n; p++)
add_param_constraints (scop, p);
}
/* Build the iteration domains: the loops belonging to the current
SCOP, and that vary for the execution of the current basic block.
Returns false if there is no loop in SCOP. */
static void
build_scop_iteration_domain (scop_p scop)
{
struct loop *loop;
sese region = SCOP_REGION (scop);
int i;
poly_bb_p pbb;
int nb_loops = number_of_loops (cfun);
isl_set **doms = XCNEWVEC (isl_set *, nb_loops);
FOR_EACH_VEC_ELT (SESE_LOOP_NEST (region), i, loop)
if (!loop_in_sese_p (loop_outer (loop), region))
build_loop_iteration_domains (scop, loop, 0,
isl_set_copy (scop->context), doms);
FOR_EACH_VEC_ELT (SCOP_BBS (scop), i, pbb)
{
loop = pbb_loop (pbb);
if (doms[loop->num])
pbb->domain = isl_set_copy (doms[loop->num]);
else
pbb->domain = isl_set_copy (scop->context);
pbb->domain = isl_set_set_tuple_id (pbb->domain,
isl_id_for_pbb (scop, pbb));
}
for (i = 0; i < nb_loops; i++)
if (doms[i])
isl_set_free (doms[i]);
free (doms);
}
/* Add a constrain to the ACCESSES polyhedron for the alias set of
data reference DR. ACCESSP_NB_DIMS is the dimension of the
ACCESSES polyhedron, DOM_NB_DIMS is the dimension of the iteration
domain. */
static isl_map *
pdr_add_alias_set (isl_map *acc, data_reference_p dr)
{
isl_constraint *c;
int alias_set_num = 0;
base_alias_pair *bap = (base_alias_pair *)(dr->aux);
if (bap && bap->alias_set)
alias_set_num = *(bap->alias_set);
c = isl_equality_alloc
(isl_local_space_from_space (isl_map_get_space (acc)));
c = isl_constraint_set_constant_si (c, -alias_set_num);
c = isl_constraint_set_coefficient_si (c, isl_dim_out, 0, 1);
return isl_map_add_constraint (acc, c);
}
/* Assign the affine expression INDEX to the output dimension POS of
MAP and return the result. */
static isl_map *
set_index (isl_map *map, int pos, isl_pw_aff *index)
{
isl_map *index_map;
int len = isl_map_dim (map, isl_dim_out);
isl_id *id;
index_map = isl_map_from_pw_aff (index);
index_map = isl_map_insert_dims (index_map, isl_dim_out, 0, pos);
index_map = isl_map_add_dims (index_map, isl_dim_out, len - pos - 1);
id = isl_map_get_tuple_id (map, isl_dim_out);
index_map = isl_map_set_tuple_id (index_map, isl_dim_out, id);
id = isl_map_get_tuple_id (map, isl_dim_in);
index_map = isl_map_set_tuple_id (index_map, isl_dim_in, id);
return isl_map_intersect (map, index_map);
}
/* Add to ACCESSES polyhedron equalities defining the access functions
to the memory. ACCESSP_NB_DIMS is the dimension of the ACCESSES
polyhedron, DOM_NB_DIMS is the dimension of the iteration domain.
PBB is the poly_bb_p that contains the data reference DR. */
static isl_map *
pdr_add_memory_accesses (isl_map *acc, data_reference_p dr, poly_bb_p pbb)
{
int i, nb_subscripts = DR_NUM_DIMENSIONS (dr);
scop_p scop = PBB_SCOP (pbb);
for (i = 0; i < nb_subscripts; i++)
{
isl_pw_aff *aff;
tree afn = DR_ACCESS_FN (dr, nb_subscripts - 1 - i);
aff = extract_affine (scop, afn,
isl_space_domain (isl_map_get_space (acc)));
acc = set_index (acc, i + 1, aff);
}
return acc;
}
/* Add constrains representing the size of the accessed data to the
ACCESSES polyhedron. ACCESSP_NB_DIMS is the dimension of the
ACCESSES polyhedron, DOM_NB_DIMS is the dimension of the iteration
domain. */
static isl_set *
pdr_add_data_dimensions (isl_set *extent, scop_p scop, data_reference_p dr)
{
tree ref = DR_REF (dr);
int i, nb_subscripts = DR_NUM_DIMENSIONS (dr);
for (i = nb_subscripts - 1; i >= 0; i--, ref = TREE_OPERAND (ref, 0))
{
tree low, high;
if (TREE_CODE (ref) != ARRAY_REF)
break;
low = array_ref_low_bound (ref);
high = array_ref_up_bound (ref);
/* XXX The PPL code dealt separately with
subscript - low >= 0 and high - subscript >= 0 in case one of
the two bounds isn't known. Do the same here? */
if (tree_fits_shwi_p (low)
&& high
&& tree_fits_shwi_p (high)
/* 1-element arrays at end of structures may extend over
their declared size. */
&& !(array_at_struct_end_p (ref)
&& operand_equal_p (low, high, 0)))
{
isl_id *id;
isl_aff *aff;
isl_set *univ, *lbs, *ubs;
isl_pw_aff *index;
isl_space *space;
isl_set *valid;
isl_pw_aff *lb = extract_affine_int (low, isl_set_get_space (extent));
isl_pw_aff *ub = extract_affine_int (high, isl_set_get_space (extent));
/* high >= 0 */
valid = isl_pw_aff_nonneg_set (isl_pw_aff_copy (ub));
valid = isl_set_project_out (valid, isl_dim_set, 0,
isl_set_dim (valid, isl_dim_set));
scop->context = isl_set_intersect (scop->context, valid);
space = isl_set_get_space (extent);
aff = isl_aff_zero_on_domain (isl_local_space_from_space (space));
aff = isl_aff_add_coefficient_si (aff, isl_dim_in, i + 1, 1);
univ = isl_set_universe (isl_space_domain (isl_aff_get_space (aff)));
index = isl_pw_aff_alloc (univ, aff);
id = isl_set_get_tuple_id (extent);
lb = isl_pw_aff_set_tuple_id (lb, isl_dim_in, isl_id_copy (id));
ub = isl_pw_aff_set_tuple_id (ub, isl_dim_in, id);
/* low <= sub_i <= high */
lbs = isl_pw_aff_ge_set (isl_pw_aff_copy (index), lb);
ubs = isl_pw_aff_le_set (index, ub);
extent = isl_set_intersect (extent, lbs);
extent = isl_set_intersect (extent, ubs);
}
}
return extent;
}
/* Build data accesses for DR in PBB. */
static void
build_poly_dr (data_reference_p dr, poly_bb_p pbb)
{
int dr_base_object_set;
isl_map *acc;
isl_set *extent;
scop_p scop = PBB_SCOP (pbb);
{
isl_space *dc = isl_set_get_space (pbb->domain);
int nb_out = 1 + DR_NUM_DIMENSIONS (dr);
isl_space *space = isl_space_add_dims (isl_space_from_domain (dc),
isl_dim_out, nb_out);
acc = isl_map_universe (space);
acc = isl_map_set_tuple_id (acc, isl_dim_out, isl_id_for_dr (scop, dr));
}
acc = pdr_add_alias_set (acc, dr);
acc = pdr_add_memory_accesses (acc, dr, pbb);
{
isl_id *id = isl_id_for_dr (scop, dr);
int nb = 1 + DR_NUM_DIMENSIONS (dr);
isl_space *space = isl_space_set_alloc (scop->ctx, 0, nb);
int alias_set_num = 0;
base_alias_pair *bap = (base_alias_pair *)(dr->aux);
if (bap && bap->alias_set)
alias_set_num = *(bap->alias_set);
space = isl_space_set_tuple_id (space, isl_dim_set, id);
extent = isl_set_nat_universe (space);
extent = isl_set_fix_si (extent, isl_dim_set, 0, alias_set_num);
extent = pdr_add_data_dimensions (extent, scop, dr);
}
gcc_assert (dr->aux);
dr_base_object_set = ((base_alias_pair *)(dr->aux))->base_obj_set;
new_poly_dr (pbb, dr_base_object_set,
DR_IS_READ (dr) ? PDR_READ : PDR_WRITE,
dr, DR_NUM_DIMENSIONS (dr), acc, extent);
}
/* Write to FILE the alias graph of data references in DIMACS format. */
static inline bool
write_alias_graph_to_ascii_dimacs (FILE *file, char *comment,
vec<data_reference_p> drs)
{
int num_vertex = drs.length ();
int edge_num = 0;
data_reference_p dr1, dr2;
int i, j;
if (num_vertex == 0)
return true;
FOR_EACH_VEC_ELT (drs, i, dr1)
for (j = i + 1; drs.iterate (j, &dr2); j++)
if (dr_may_alias_p (dr1, dr2, true))
edge_num++;
fprintf (file, "$\n");
if (comment)
fprintf (file, "c %s\n", comment);
fprintf (file, "p edge %d %d\n", num_vertex, edge_num);
FOR_EACH_VEC_ELT (drs, i, dr1)
for (j = i + 1; drs.iterate (j, &dr2); j++)
if (dr_may_alias_p (dr1, dr2, true))
fprintf (file, "e %d %d\n", i + 1, j + 1);
return true;
}
/* Write to FILE the alias graph of data references in DOT format. */
static inline bool
write_alias_graph_to_ascii_dot (FILE *file, char *comment,
vec<data_reference_p> drs)
{
int num_vertex = drs.length ();
data_reference_p dr1, dr2;
int i, j;
if (num_vertex == 0)
return true;
fprintf (file, "$\n");
if (comment)
fprintf (file, "c %s\n", comment);
/* First print all the vertices. */
FOR_EACH_VEC_ELT (drs, i, dr1)
fprintf (file, "n%d;\n", i);
FOR_EACH_VEC_ELT (drs, i, dr1)
for (j = i + 1; drs.iterate (j, &dr2); j++)
if (dr_may_alias_p (dr1, dr2, true))
fprintf (file, "n%d n%d\n", i, j);
return true;
}
/* Write to FILE the alias graph of data references in ECC format. */
static inline bool
write_alias_graph_to_ascii_ecc (FILE *file, char *comment,
vec<data_reference_p> drs)
{
int num_vertex = drs.length ();
data_reference_p dr1, dr2;
int i, j;
if (num_vertex == 0)
return true;
fprintf (file, "$\n");
if (comment)
fprintf (file, "c %s\n", comment);
FOR_EACH_VEC_ELT (drs, i, dr1)
for (j = i + 1; drs.iterate (j, &dr2); j++)
if (dr_may_alias_p (dr1, dr2, true))
fprintf (file, "%d %d\n", i, j);
return true;
}
/* Check if DR1 and DR2 are in the same object set. */
static bool
dr_same_base_object_p (const struct data_reference *dr1,
const struct data_reference *dr2)
{
return operand_equal_p (DR_BASE_OBJECT (dr1), DR_BASE_OBJECT (dr2), 0);
}
/* Uses DFS component number as representative of alias-sets. Also tests for
optimality by verifying if every connected component is a clique. Returns
true (1) if the above test is true, and false (0) otherwise. */
static int
build_alias_set_optimal_p (vec<data_reference_p> drs)
{
int num_vertices = drs.length ();
struct graph *g = new_graph (num_vertices);
data_reference_p dr1, dr2;
int i, j;
int num_connected_components;
int v_indx1, v_indx2, num_vertices_in_component;
int *all_vertices;
int *vertices;
struct graph_edge *e;
int this_component_is_clique;
int all_components_are_cliques = 1;
FOR_EACH_VEC_ELT (drs, i, dr1)
for (j = i+1; drs.iterate (j, &dr2); j++)
if (dr_may_alias_p (dr1, dr2, true))
{
add_edge (g, i, j);
add_edge (g, j, i);
}
all_vertices = XNEWVEC (int, num_vertices);
vertices = XNEWVEC (int, num_vertices);
for (i = 0; i < num_vertices; i++)
all_vertices[i] = i;
num_connected_components = graphds_dfs (g, all_vertices, num_vertices,
NULL, true, NULL);
for (i = 0; i < g->n_vertices; i++)
{
data_reference_p dr = drs[i];
base_alias_pair *bap;
gcc_assert (dr->aux);
bap = (base_alias_pair *)(dr->aux);
bap->alias_set = XNEW (int);
*(bap->alias_set) = g->vertices[i].component + 1;
}
/* Verify if the DFS numbering results in optimal solution. */
for (i = 0; i < num_connected_components; i++)
{
num_vertices_in_component = 0;
/* Get all vertices whose DFS component number is the same as i. */
for (j = 0; j < num_vertices; j++)
if (g->vertices[j].component == i)
vertices[num_vertices_in_component++] = j;
/* Now test if the vertices in 'vertices' form a clique, by testing
for edges among each pair. */
this_component_is_clique = 1;
for (v_indx1 = 0; v_indx1 < num_vertices_in_component; v_indx1++)
{
for (v_indx2 = v_indx1+1; v_indx2 < num_vertices_in_component; v_indx2++)
{
/* Check if the two vertices are connected by iterating
through all the edges which have one of these are source. */
e = g->vertices[vertices[v_indx2]].pred;
while (e)
{
if (e->src == vertices[v_indx1])
break;
e = e->pred_next;
}
if (!e)
{
this_component_is_clique = 0;
break;
}
}
if (!this_component_is_clique)
all_components_are_cliques = 0;
}
}
free (all_vertices);
free (vertices);
free_graph (g);
return all_components_are_cliques;
}
/* Group each data reference in DRS with its base object set num. */
static void
build_base_obj_set_for_drs (vec<data_reference_p> drs)
{
int num_vertex = drs.length ();
struct graph *g = new_graph (num_vertex);
data_reference_p dr1, dr2;
int i, j;
int *queue;
FOR_EACH_VEC_ELT (drs, i, dr1)
for (j = i + 1; drs.iterate (j, &dr2); j++)
if (dr_same_base_object_p (dr1, dr2))
{
add_edge (g, i, j);
add_edge (g, j, i);
}
queue = XNEWVEC (int, num_vertex);
for (i = 0; i < num_vertex; i++)
queue[i] = i;
graphds_dfs (g, queue, num_vertex, NULL, true, NULL);
for (i = 0; i < g->n_vertices; i++)
{
data_reference_p dr = drs[i];
base_alias_pair *bap;
gcc_assert (dr->aux);
bap = (base_alias_pair *)(dr->aux);
bap->base_obj_set = g->vertices[i].component + 1;
}
free (queue);
free_graph (g);
}
/* Build the data references for PBB. */
static void
build_pbb_drs (poly_bb_p pbb)
{
int j;
data_reference_p dr;
vec<data_reference_p> gbb_drs = GBB_DATA_REFS (PBB_BLACK_BOX (pbb));
FOR_EACH_VEC_ELT (gbb_drs, j, dr)
build_poly_dr (dr, pbb);
}
/* Dump to file the alias graphs for the data references in DRS. */
static void
dump_alias_graphs (vec<data_reference_p> drs)
{
char comment[100];
FILE *file_dimacs, *file_ecc, *file_dot;
file_dimacs = fopen ("/tmp/dr_alias_graph_dimacs", "ab");
if (file_dimacs)
{
snprintf (comment, sizeof (comment), "%s %s", main_input_filename,
current_function_name ());
write_alias_graph_to_ascii_dimacs (file_dimacs, comment, drs);
fclose (file_dimacs);
}
file_ecc = fopen ("/tmp/dr_alias_graph_ecc", "ab");
if (file_ecc)
{
snprintf (comment, sizeof (comment), "%s %s", main_input_filename,
current_function_name ());
write_alias_graph_to_ascii_ecc (file_ecc, comment, drs);
fclose (file_ecc);
}
file_dot = fopen ("/tmp/dr_alias_graph_dot", "ab");
if (file_dot)
{
snprintf (comment, sizeof (comment), "%s %s", main_input_filename,
current_function_name ());
write_alias_graph_to_ascii_dot (file_dot, comment, drs);
fclose (file_dot);
}
}
/* Build data references in SCOP. */
static void
build_scop_drs (scop_p scop)
{
int i, j;
poly_bb_p pbb;
data_reference_p dr;
auto_vec<data_reference_p, 3> drs;
/* Remove all the PBBs that do not have data references: these basic
blocks are not handled in the polyhedral representation. */
for (i = 0; SCOP_BBS (scop).iterate (i, &pbb); i++)
if (GBB_DATA_REFS (PBB_BLACK_BOX (pbb)).is_empty ())
{
free_gimple_bb (PBB_BLACK_BOX (pbb));
free_poly_bb (pbb);
SCOP_BBS (scop).ordered_remove (i);
i--;
}
FOR_EACH_VEC_ELT (SCOP_BBS (scop), i, pbb)
for (j = 0; GBB_DATA_REFS (PBB_BLACK_BOX (pbb)).iterate (j, &dr); j++)
drs.safe_push (dr);
FOR_EACH_VEC_ELT (drs, i, dr)
dr->aux = XNEW (base_alias_pair);
if (!build_alias_set_optimal_p (drs))
{
/* TODO: Add support when building alias set is not optimal. */
;
}
build_base_obj_set_for_drs (drs);
/* When debugging, enable the following code. This cannot be used
in production compilers. */
if (0)
dump_alias_graphs (drs);
drs.release ();
FOR_EACH_VEC_ELT (SCOP_BBS (scop), i, pbb)
build_pbb_drs (pbb);
}
/* Return a gsi at the position of the phi node STMT. */
static gphi_iterator
gsi_for_phi_node (gphi *stmt)
{
gphi_iterator psi;
basic_block bb = gimple_bb (stmt);
for (psi = gsi_start_phis (bb); !gsi_end_p (psi); gsi_next (&psi))
if (stmt == psi.phi ())
return psi;
gcc_unreachable ();
return psi;
}
/* Analyze all the data references of STMTS and add them to the
GBB_DATA_REFS vector of BB. */
static void
analyze_drs_in_stmts (scop_p scop, basic_block bb, vec<gimple> stmts)
{
loop_p nest;
gimple_bb_p gbb;
gimple stmt;
int i;
sese region = SCOP_REGION (scop);
if (!bb_in_sese_p (bb, region))
return;
nest = outermost_loop_in_sese_1 (region, bb);
gbb = gbb_from_bb (bb);
FOR_EACH_VEC_ELT (stmts, i, stmt)
{
loop_p loop;
if (is_gimple_debug (stmt))
continue;
loop = loop_containing_stmt (stmt);
if (!loop_in_sese_p (loop, region))
loop = nest;
graphite_find_data_references_in_stmt (nest, loop, stmt,
&GBB_DATA_REFS (gbb));
}
}
/* Insert STMT at the end of the STMTS sequence and then insert the
statements from STMTS at INSERT_GSI and call analyze_drs_in_stmts
on STMTS. */
static void
insert_stmts (scop_p scop, gimple stmt, gimple_seq stmts,
gimple_stmt_iterator insert_gsi)
{
gimple_stmt_iterator gsi;
auto_vec<gimple, 3> x;
gimple_seq_add_stmt (&stmts, stmt);
for (gsi = gsi_start (stmts); !gsi_end_p (gsi); gsi_next (&gsi))
x.safe_push (gsi_stmt (gsi));
gsi_insert_seq_before (&insert_gsi, stmts, GSI_SAME_STMT);
analyze_drs_in_stmts (scop, gsi_bb (insert_gsi), x);
}
/* Insert the assignment "RES := EXPR" just after AFTER_STMT. */
static void
insert_out_of_ssa_copy (scop_p scop, tree res, tree expr, gimple after_stmt)
{
gimple_seq stmts;
gimple_stmt_iterator gsi;
tree var = force_gimple_operand (expr, &stmts, true, NULL_TREE);
gassign *stmt = gimple_build_assign (unshare_expr (res), var);
auto_vec<gimple, 3> x;
gimple_seq_add_stmt (&stmts, stmt);
for (gsi = gsi_start (stmts); !gsi_end_p (gsi); gsi_next (&gsi))
x.safe_push (gsi_stmt (gsi));
if (gimple_code (after_stmt) == GIMPLE_PHI)
{
gsi = gsi_after_labels (gimple_bb (after_stmt));
gsi_insert_seq_before (&gsi, stmts, GSI_NEW_STMT);
}
else
{
gsi = gsi_for_stmt (after_stmt);
gsi_insert_seq_after (&gsi, stmts, GSI_NEW_STMT);
}
analyze_drs_in_stmts (scop, gimple_bb (after_stmt), x);
}
/* Creates a poly_bb_p for basic_block BB from the existing PBB. */
static void
new_pbb_from_pbb (scop_p scop, poly_bb_p pbb, basic_block bb)
{
vec<data_reference_p> drs;
drs.create (3);
gimple_bb_p gbb = PBB_BLACK_BOX (pbb);
gimple_bb_p gbb1 = new_gimple_bb (bb, drs);
poly_bb_p pbb1 = new_poly_bb (scop, gbb1);
int index, n = SCOP_BBS (scop).length ();
/* The INDEX of PBB in SCOP_BBS. */
for (index = 0; index < n; index++)
if (SCOP_BBS (scop)[index] == pbb)
break;
pbb1->domain = isl_set_copy (pbb->domain);
pbb1->domain = isl_set_set_tuple_id (pbb1->domain,
isl_id_for_pbb (scop, pbb1));
GBB_PBB (gbb1) = pbb1;
GBB_CONDITIONS (gbb1) = GBB_CONDITIONS (gbb).copy ();
GBB_CONDITION_CASES (gbb1) = GBB_CONDITION_CASES (gbb).copy ();
SCOP_BBS (scop).safe_insert (index + 1, pbb1);
}
/* Insert on edge E the assignment "RES := EXPR". */
static void
insert_out_of_ssa_copy_on_edge (scop_p scop, edge e, tree res, tree expr)
{
gimple_stmt_iterator gsi;
gimple_seq stmts = NULL;
tree var = force_gimple_operand (expr, &stmts, true, NULL_TREE);
gimple stmt = gimple_build_assign (unshare_expr (res), var);
basic_block bb;
auto_vec<gimple, 3> x;
gimple_seq_add_stmt (&stmts, stmt);
for (gsi = gsi_start (stmts); !gsi_end_p (gsi); gsi_next (&gsi))
x.safe_push (gsi_stmt (gsi));
gsi_insert_seq_on_edge (e, stmts);
gsi_commit_edge_inserts ();
bb = gimple_bb (stmt);
if (!bb_in_sese_p (bb, SCOP_REGION (scop)))
return;
if (!gbb_from_bb (bb))
new_pbb_from_pbb (scop, pbb_from_bb (e->src), bb);
analyze_drs_in_stmts (scop, bb, x);
}
/* Creates a zero dimension array of the same type as VAR. */
static tree
create_zero_dim_array (tree var, const char *base_name)
{
tree index_type = build_index_type (integer_zero_node);
tree elt_type = TREE_TYPE (var);
tree array_type = build_array_type (elt_type, index_type);
tree base = create_tmp_var (array_type, base_name);
return build4 (ARRAY_REF, elt_type, base, integer_zero_node, NULL_TREE,
NULL_TREE);
}
/* Returns true when PHI is a loop close phi node. */
static bool
scalar_close_phi_node_p (gimple phi)
{
if (gimple_code (phi) != GIMPLE_PHI
|| virtual_operand_p (gimple_phi_result (phi)))
return false;
/* Note that loop close phi nodes should have a single argument
because we translated the representation into a canonical form
before Graphite: see canonicalize_loop_closed_ssa_form. */
return (gimple_phi_num_args (phi) == 1);
}
/* For a definition DEF in REGION, propagates the expression EXPR in
all the uses of DEF outside REGION. */
static void
propagate_expr_outside_region (tree def, tree expr, sese region)
{
imm_use_iterator imm_iter;
gimple use_stmt;
gimple_seq stmts;
bool replaced_once = false;
gcc_assert (TREE_CODE (def) == SSA_NAME);
expr = force_gimple_operand (unshare_expr (expr), &stmts, true,
NULL_TREE);
FOR_EACH_IMM_USE_STMT (use_stmt, imm_iter, def)
if (!is_gimple_debug (use_stmt)
&& !bb_in_sese_p (gimple_bb (use_stmt), region))
{
ssa_op_iter iter;
use_operand_p use_p;
FOR_EACH_PHI_OR_STMT_USE (use_p, use_stmt, iter, SSA_OP_ALL_USES)
if (operand_equal_p (def, USE_FROM_PTR (use_p), 0)
&& (replaced_once = true))
replace_exp (use_p, expr);
update_stmt (use_stmt);
}
if (replaced_once)
{
gsi_insert_seq_on_edge (SESE_ENTRY (region), stmts);
gsi_commit_edge_inserts ();
}
}
/* Rewrite out of SSA the reduction phi node at PSI by creating a zero
dimension array for it. */
static void
rewrite_close_phi_out_of_ssa (scop_p scop, gimple_stmt_iterator *psi)
{
sese region = SCOP_REGION (scop);
gimple phi = gsi_stmt (*psi);
tree res = gimple_phi_result (phi);
basic_block bb = gimple_bb (phi);
gimple_stmt_iterator gsi = gsi_after_labels (bb);
tree arg = gimple_phi_arg_def (phi, 0);
gimple stmt;
/* Note that loop close phi nodes should have a single argument
because we translated the representation into a canonical form
before Graphite: see canonicalize_loop_closed_ssa_form. */
gcc_assert (gimple_phi_num_args (phi) == 1);
/* The phi node can be a non close phi node, when its argument is
invariant, or a default definition. */
if (is_gimple_min_invariant (arg)
|| SSA_NAME_IS_DEFAULT_DEF (arg))
{
propagate_expr_outside_region (res, arg, region);
gsi_next (psi);
return;
}
else if (gimple_bb (SSA_NAME_DEF_STMT (arg))->loop_father == bb->loop_father)
{
propagate_expr_outside_region (res, arg, region);
stmt = gimple_build_assign (res, arg);
remove_phi_node (psi, false);
gsi_insert_before (&gsi, stmt, GSI_NEW_STMT);
return;
}
/* If res is scev analyzable and is not a scalar value, it is safe
to ignore the close phi node: it will be code generated in the
out of Graphite pass. */
else if (scev_analyzable_p (res, region))
{
loop_p loop = loop_containing_stmt (SSA_NAME_DEF_STMT (res));
tree scev;
if (!loop_in_sese_p (loop, region))
{
loop = loop_containing_stmt (SSA_NAME_DEF_STMT (arg));
scev = scalar_evolution_in_region (region, loop, arg);
scev = compute_overall_effect_of_inner_loop (loop, scev);
}
else
scev = scalar_evolution_in_region (region, loop, res);
if (tree_does_not_contain_chrecs (scev))
propagate_expr_outside_region (res, scev, region);
gsi_next (psi);
return;
}
else
{
tree zero_dim_array = create_zero_dim_array (res, "Close_Phi");
stmt = gimple_build_assign (res, unshare_expr (zero_dim_array));
if (TREE_CODE (arg) == SSA_NAME)
insert_out_of_ssa_copy (scop, zero_dim_array, arg,
SSA_NAME_DEF_STMT (arg));
else
insert_out_of_ssa_copy_on_edge (scop, single_pred_edge (bb),
zero_dim_array, arg);
}
remove_phi_node (psi, false);
SSA_NAME_DEF_STMT (res) = stmt;
insert_stmts (scop, stmt, NULL, gsi_after_labels (bb));
}
/* Rewrite out of SSA the reduction phi node at PSI by creating a zero
dimension array for it. */
static void
rewrite_phi_out_of_ssa (scop_p scop, gphi_iterator *psi)
{
size_t i;
gphi *phi = psi->phi ();
basic_block bb = gimple_bb (phi);
tree res = gimple_phi_result (phi);
tree zero_dim_array = create_zero_dim_array (res, "phi_out_of_ssa");
gimple stmt;
for (i = 0; i < gimple_phi_num_args (phi); i++)
{
tree arg = gimple_phi_arg_def (phi, i);
edge e = gimple_phi_arg_edge (phi, i);
/* Avoid the insertion of code in the loop latch to please the
pattern matching of the vectorizer. */
if (TREE_CODE (arg) == SSA_NAME
&& !SSA_NAME_IS_DEFAULT_DEF (arg)
&& e->src == bb->loop_father->latch)
insert_out_of_ssa_copy (scop, zero_dim_array, arg,
SSA_NAME_DEF_STMT (arg));
else
insert_out_of_ssa_copy_on_edge (scop, e, zero_dim_array, arg);
}
stmt = gimple_build_assign (res, unshare_expr (zero_dim_array));
remove_phi_node (psi, false);
insert_stmts (scop, stmt, NULL, gsi_after_labels (bb));
}
/* Rewrite the degenerate phi node at position PSI from the degenerate
form "x = phi (y, y, ..., y)" to "x = y". */
static void
rewrite_degenerate_phi (gphi_iterator *psi)
{
tree rhs;
gimple stmt;
gimple_stmt_iterator gsi;
gphi *phi = psi->phi ();
tree res = gimple_phi_result (phi);
basic_block bb;
bb = gimple_bb (phi);
rhs = degenerate_phi_result (phi);
gcc_assert (rhs);
stmt = gimple_build_assign (res, rhs);
remove_phi_node (psi, false);
gsi = gsi_after_labels (bb);
gsi_insert_before (&gsi, stmt, GSI_NEW_STMT);
}
/* Rewrite out of SSA all the reduction phi nodes of SCOP. */
static void
rewrite_reductions_out_of_ssa (scop_p scop)
{
basic_block bb;
gphi_iterator psi;
sese region = SCOP_REGION (scop);
FOR_EACH_BB_FN (bb, cfun)
if (bb_in_sese_p (bb, region))
for (psi = gsi_start_phis (bb); !gsi_end_p (psi);)
{
gphi *phi = psi.phi ();
if (virtual_operand_p (gimple_phi_result (phi)))
{
gsi_next (&psi);
continue;
}
if (gimple_phi_num_args (phi) > 1
&& degenerate_phi_result (phi))
rewrite_degenerate_phi (&psi);
else if (scalar_close_phi_node_p (phi))
rewrite_close_phi_out_of_ssa (scop, &psi);
else if (reduction_phi_p (region, &psi))
rewrite_phi_out_of_ssa (scop, &psi);
}
update_ssa (TODO_update_ssa);
#ifdef ENABLE_CHECKING
verify_loop_closed_ssa (true);
#endif
}
/* Rewrite the scalar dependence of DEF used in USE_STMT with a memory
read from ZERO_DIM_ARRAY. */
static void
rewrite_cross_bb_scalar_dependence (scop_p scop, tree zero_dim_array,
tree def, gimple use_stmt)
{
gimple name_stmt;
tree name;
ssa_op_iter iter;
use_operand_p use_p;
gcc_assert (gimple_code (use_stmt) != GIMPLE_PHI);
name = copy_ssa_name (def);
name_stmt = gimple_build_assign (name, zero_dim_array);
gimple_assign_set_lhs (name_stmt, name);
insert_stmts (scop, name_stmt, NULL, gsi_for_stmt (use_stmt));
FOR_EACH_SSA_USE_OPERAND (use_p, use_stmt, iter, SSA_OP_ALL_USES)
if (operand_equal_p (def, USE_FROM_PTR (use_p), 0))
replace_exp (use_p, name);
update_stmt (use_stmt);
}
/* For every definition DEF in the SCOP that is used outside the scop,
insert a closing-scop definition in the basic block just after this
SCOP. */
static void
handle_scalar_deps_crossing_scop_limits (scop_p scop, tree def, gimple stmt)
{
tree var = create_tmp_reg (TREE_TYPE (def));
tree new_name = make_ssa_name (var, stmt);
bool needs_copy = false;
use_operand_p use_p;
imm_use_iterator imm_iter;
gimple use_stmt;
sese region = SCOP_REGION (scop);
FOR_EACH_IMM_USE_STMT (use_stmt, imm_iter, def)
{
if (!bb_in_sese_p (gimple_bb (use_stmt), region))
{
FOR_EACH_IMM_USE_ON_STMT (use_p, imm_iter)
{
SET_USE (use_p, new_name);
}
update_stmt (use_stmt);
needs_copy = true;
}
}
/* Insert in the empty BB just after the scop a use of DEF such
that the rewrite of cross_bb_scalar_dependences won't insert
arrays everywhere else. */
if (needs_copy)
{
gimple assign = gimple_build_assign (new_name, def);
gimple_stmt_iterator psi = gsi_after_labels (SESE_EXIT (region)->dest);
update_stmt (assign);
gsi_insert_before (&psi, assign, GSI_SAME_STMT);
}
}
/* Rewrite the scalar dependences crossing the boundary of the BB
containing STMT with an array. Return true when something has been
changed. */
static bool
rewrite_cross_bb_scalar_deps (scop_p scop, gimple_stmt_iterator *gsi)
{
sese region = SCOP_REGION (scop);
gimple stmt = gsi_stmt (*gsi);
imm_use_iterator imm_iter;
tree def;
basic_block def_bb;
tree zero_dim_array = NULL_TREE;
gimple use_stmt;
bool res = false;
switch (gimple_code (stmt))
{
case GIMPLE_ASSIGN:
def = gimple_assign_lhs (stmt);
break;
case GIMPLE_CALL:
def = gimple_call_lhs (stmt);
break;
default:
return false;
}
if (!def
|| !is_gimple_reg (def))
return false;
if (scev_analyzable_p (def, region))
{
loop_p loop = loop_containing_stmt (SSA_NAME_DEF_STMT (def));
tree scev = scalar_evolution_in_region (region, loop, def);
if (tree_contains_chrecs (scev, NULL))
return false;
propagate_expr_outside_region (def, scev, region);
return true;
}
def_bb = gimple_bb (stmt);
handle_scalar_deps_crossing_scop_limits (scop, def, stmt);
FOR_EACH_IMM_USE_STMT (use_stmt, imm_iter, def)
if (gimple_code (use_stmt) == GIMPLE_PHI
&& (res = true))
{
gphi_iterator psi = gsi_start_phis (gimple_bb (use_stmt));
if (scalar_close_phi_node_p (gsi_stmt (psi)))
rewrite_close_phi_out_of_ssa (scop, &psi);
else
rewrite_phi_out_of_ssa (scop, &psi);
}
FOR_EACH_IMM_USE_STMT (use_stmt, imm_iter, def)
if (gimple_code (use_stmt) != GIMPLE_PHI
&& def_bb != gimple_bb (use_stmt)
&& !is_gimple_debug (use_stmt)
&& (res = true))
{
if (!zero_dim_array)
{
zero_dim_array = create_zero_dim_array
(def, "Cross_BB_scalar_dependence");
insert_out_of_ssa_copy (scop, zero_dim_array, def,
SSA_NAME_DEF_STMT (def));
gsi_next (gsi);
}
rewrite_cross_bb_scalar_dependence (scop, unshare_expr (zero_dim_array),
def, use_stmt);
}
return res;
}
/* Rewrite out of SSA all the reduction phi nodes of SCOP. */
static void
rewrite_cross_bb_scalar_deps_out_of_ssa (scop_p scop)
{
basic_block bb;
gimple_stmt_iterator psi;
sese region = SCOP_REGION (scop);
bool changed = false;
/* Create an extra empty BB after the scop. */
split_edge (SESE_EXIT (region));
FOR_EACH_BB_FN (bb, cfun)
if (bb_in_sese_p (bb, region))
for (psi = gsi_start_bb (bb); !gsi_end_p (psi); gsi_next (&psi))
changed |= rewrite_cross_bb_scalar_deps (scop, &psi);
if (changed)
{
scev_reset_htab ();
update_ssa (TODO_update_ssa);
#ifdef ENABLE_CHECKING
verify_loop_closed_ssa (true);
#endif
}
}
/* Returns the number of pbbs that are in loops contained in SCOP. */
static int
nb_pbbs_in_loops (scop_p scop)
{
int i;
poly_bb_p pbb;
int res = 0;
FOR_EACH_VEC_ELT (SCOP_BBS (scop), i, pbb)
if (loop_in_sese_p (gbb_loop (PBB_BLACK_BOX (pbb)), SCOP_REGION (scop)))
res++;
return res;
}
/* Return the number of data references in BB that write in
memory. */
static int
nb_data_writes_in_bb (basic_block bb)
{
int res = 0;
gimple_stmt_iterator gsi;
for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
if (gimple_vdef (gsi_stmt (gsi)))
res++;
return res;
}
/* Splits at STMT the basic block BB represented as PBB in the
polyhedral form. */
static edge
split_pbb (scop_p scop, poly_bb_p pbb, basic_block bb, gimple stmt)
{
edge e1 = split_block (bb, stmt);
new_pbb_from_pbb (scop, pbb, e1->dest);
return e1;
}
/* Splits STMT out of its current BB. This is done for reduction
statements for which we want to ignore data dependences. */
static basic_block
split_reduction_stmt (scop_p scop, gimple stmt)
{
basic_block bb = gimple_bb (stmt);
poly_bb_p pbb = pbb_from_bb (bb);
gimple_bb_p gbb = gbb_from_bb (bb);
edge e1;
int i;
data_reference_p dr;
/* Do not split basic blocks with no writes to memory: the reduction
will be the only write to memory. */
if (nb_data_writes_in_bb (bb) == 0
/* Or if we have already marked BB as a reduction. */
|| PBB_IS_REDUCTION (pbb_from_bb (bb)))
return bb;
e1 = split_pbb (scop, pbb, bb, stmt);
/* Split once more only when the reduction stmt is not the only one
left in the original BB. */
if (!gsi_one_before_end_p (gsi_start_nondebug_bb (bb)))
{
gimple_stmt_iterator gsi = gsi_last_bb (bb);
gsi_prev (&gsi);
e1 = split_pbb (scop, pbb, bb, gsi_stmt (gsi));
}
/* A part of the data references will end in a different basic block
after the split: move the DRs from the original GBB to the newly
created GBB1. */
FOR_EACH_VEC_ELT (GBB_DATA_REFS (gbb), i, dr)
{
basic_block bb1 = gimple_bb (DR_STMT (dr));
if (bb1 != bb)
{
gimple_bb_p gbb1 = gbb_from_bb (bb1);
GBB_DATA_REFS (gbb1).safe_push (dr);
GBB_DATA_REFS (gbb).ordered_remove (i);
i--;
}
}
return e1->dest;
}
/* Return true when stmt is a reduction operation. */
static inline bool
is_reduction_operation_p (gimple stmt)
{
enum tree_code code;
gcc_assert (is_gimple_assign (stmt));
code = gimple_assign_rhs_code (stmt);
if (!commutative_tree_code (code)
|| !associative_tree_code (code))
return false;
tree type = TREE_TYPE (gimple_assign_lhs (stmt));
if (FLOAT_TYPE_P (type))
return flag_associative_math;
return (INTEGRAL_TYPE_P (type)
&& TYPE_OVERFLOW_WRAPS (type));
}
/* Returns true when PHI contains an argument ARG. */
static bool
phi_contains_arg (gphi *phi, tree arg)
{
size_t i;
for (i = 0; i < gimple_phi_num_args (phi); i++)
if (operand_equal_p (arg, gimple_phi_arg_def (phi, i), 0))
return true;
return false;
}
/* Return a loop phi node that corresponds to a reduction containing LHS. */
static gphi *
follow_ssa_with_commutative_ops (tree arg, tree lhs)
{
gimple stmt;
if (TREE_CODE (arg) != SSA_NAME)
return NULL;
stmt = SSA_NAME_DEF_STMT (arg);
if (gimple_code (stmt) == GIMPLE_NOP
|| gimple_code (stmt) == GIMPLE_CALL)
return NULL;
if (gphi *phi = dyn_cast <gphi *> (stmt))
{
if (phi_contains_arg (phi, lhs))
return phi;
return NULL;
}
if (!is_gimple_assign (stmt))
return NULL;
if (gimple_num_ops (stmt) == 2)
return follow_ssa_with_commutative_ops (gimple_assign_rhs1 (stmt), lhs);
if (is_reduction_operation_p (stmt))
{
gphi *res
= follow_ssa_with_commutative_ops (gimple_assign_rhs1 (stmt), lhs);
return res ? res :
follow_ssa_with_commutative_ops (gimple_assign_rhs2 (stmt), lhs);
}
return NULL;
}
/* Detect commutative and associative scalar reductions starting at
the STMT. Return the phi node of the reduction cycle, or NULL. */
static gphi *
detect_commutative_reduction_arg (tree lhs, gimple stmt, tree arg,
vec<gimple> *in,
vec<gimple> *out)
{
gphi *phi = follow_ssa_with_commutative_ops (arg, lhs);
if (!phi)
return NULL;
in->safe_push (stmt);
out->safe_push (stmt);
return phi;
}
/* Detect commutative and associative scalar reductions starting at
STMT. Return the phi node of the reduction cycle, or NULL. */
static gphi *
detect_commutative_reduction_assign (gimple stmt, vec<gimple> *in,
vec<gimple> *out)
{
tree lhs = gimple_assign_lhs (stmt);
if (gimple_num_ops (stmt) == 2)
return detect_commutative_reduction_arg (lhs, stmt,
gimple_assign_rhs1 (stmt),
in, out);
if (is_reduction_operation_p (stmt))
{
gphi *res = detect_commutative_reduction_arg (lhs, stmt,
gimple_assign_rhs1 (stmt),
in, out);
return res ? res
: detect_commutative_reduction_arg (lhs, stmt,
gimple_assign_rhs2 (stmt),
in, out);
}
return NULL;
}
/* Return a loop phi node that corresponds to a reduction containing LHS. */
static gphi *
follow_inital_value_to_phi (tree arg, tree lhs)
{
gimple stmt;
if (!arg || TREE_CODE (arg) != SSA_NAME)
return NULL;
stmt = SSA_NAME_DEF_STMT (arg);
if (gphi *phi = dyn_cast <gphi *> (stmt))
if (phi_contains_arg (phi, lhs))
return phi;
return NULL;
}
/* Return the argument of the loop PHI that is the initial value coming
from outside the loop. */
static edge
edge_initial_value_for_loop_phi (gphi *phi)
{
size_t i;
for (i = 0; i < gimple_phi_num_args (phi); i++)
{
edge e = gimple_phi_arg_edge (phi, i);
if (loop_depth (e->src->loop_father)
< loop_depth (e->dest->loop_father))
return e;
}
return NULL;
}
/* Return the argument of the loop PHI that is the initial value coming
from outside the loop. */
static tree
initial_value_for_loop_phi (gphi *phi)
{
size_t i;
for (i = 0; i < gimple_phi_num_args (phi); i++)
{
edge e = gimple_phi_arg_edge (phi, i);
if (loop_depth (e->src->loop_father)
< loop_depth (e->dest->loop_father))
return gimple_phi_arg_def (phi, i);
}
return NULL_TREE;
}
/* Returns true when DEF is used outside the reduction cycle of
LOOP_PHI. */
static bool
used_outside_reduction (tree def, gimple loop_phi)
{
use_operand_p use_p;
imm_use_iterator imm_iter;
loop_p loop = loop_containing_stmt (loop_phi);
/* In LOOP, DEF should be used only in LOOP_PHI. */
FOR_EACH_IMM_USE_FAST (use_p, imm_iter, def)
{
gimple stmt = USE_STMT (use_p);
if (stmt != loop_phi
&& !is_gimple_debug (stmt)
&& flow_bb_inside_loop_p (loop, gimple_bb (stmt)))
return true;
}
return false;
}
/* Detect commutative and associative scalar reductions belonging to
the SCOP starting at the loop closed phi node STMT. Return the phi
node of the reduction cycle, or NULL. */
static gphi *
detect_commutative_reduction (scop_p scop, gimple stmt, vec<gimple> *in,
vec<gimple> *out)
{
if (scalar_close_phi_node_p (stmt))
{
gimple def;
gphi *loop_phi, *phi, *close_phi = as_a <gphi *> (stmt);
tree init, lhs, arg = gimple_phi_arg_def (close_phi, 0);
if (TREE_CODE (arg) != SSA_NAME)
return NULL;
/* Note that loop close phi nodes should have a single argument
because we translated the representation into a canonical form
before Graphite: see canonicalize_loop_closed_ssa_form. */
gcc_assert (gimple_phi_num_args (close_phi) == 1);
def = SSA_NAME_DEF_STMT (arg);
if (!stmt_in_sese_p (def, SCOP_REGION (scop))
|| !(loop_phi = detect_commutative_reduction (scop, def, in, out)))
return NULL;
lhs = gimple_phi_result (close_phi);
init = initial_value_for_loop_phi (loop_phi);
phi = follow_inital_value_to_phi (init, lhs);
if (phi && (used_outside_reduction (lhs, phi)
|| !has_single_use (gimple_phi_result (phi))))
return NULL;
in->safe_push (loop_phi);
out->safe_push (close_phi);
return phi;
}
if (gimple_code (stmt) == GIMPLE_ASSIGN)
return detect_commutative_reduction_assign (stmt, in, out);
return NULL;
}
/* Translate the scalar reduction statement STMT to an array RED
knowing that its recursive phi node is LOOP_PHI. */
static void
translate_scalar_reduction_to_array_for_stmt (scop_p scop, tree red,
gimple stmt, gphi *loop_phi)
{
tree res = gimple_phi_result (loop_phi);
gassign *assign = gimple_build_assign (res, unshare_expr (red));
gimple_stmt_iterator gsi;
insert_stmts (scop, assign, NULL, gsi_after_labels (gimple_bb (loop_phi)));
assign = gimple_build_assign (unshare_expr (red), gimple_assign_lhs (stmt));
gsi = gsi_for_stmt (stmt);
gsi_next (&gsi);
insert_stmts (scop, assign, NULL, gsi);
}
/* Removes the PHI node and resets all the debug stmts that are using
the PHI_RESULT. */
static void
remove_phi (gphi *phi)
{
imm_use_iterator imm_iter;
tree def;
use_operand_p use_p;
gimple_stmt_iterator gsi;
auto_vec<gimple, 3> update;
unsigned int i;
gimple stmt;
def = PHI_RESULT (phi);
FOR_EACH_IMM_USE_FAST (use_p, imm_iter, def)
{
stmt = USE_STMT (use_p);
if (is_gimple_debug (stmt))
{
gimple_debug_bind_reset_value (stmt);
update.safe_push (stmt);
}
}
FOR_EACH_VEC_ELT (update, i, stmt)
update_stmt (stmt);
gsi = gsi_for_phi_node (phi);
remove_phi_node (&gsi, false);
}
/* Helper function for for_each_index. For each INDEX of the data
reference REF, returns true when its indices are valid in the loop
nest LOOP passed in as DATA. */
static bool
dr_indices_valid_in_loop (tree ref ATTRIBUTE_UNUSED, tree *index, void *data)
{
loop_p loop;
basic_block header, def_bb;
gimple stmt;
if (TREE_CODE (*index) != SSA_NAME)
return true;
loop = *((loop_p *) data);
header = loop->header;
stmt = SSA_NAME_DEF_STMT (*index);
if (!stmt)
return true;
def_bb = gimple_bb (stmt);
if (!def_bb)
return true;
return dominated_by_p (CDI_DOMINATORS, header, def_bb);
}
/* When the result of a CLOSE_PHI is written to a memory location,
return a pointer to that memory reference, otherwise return
NULL_TREE. */
static tree
close_phi_written_to_memory (gphi *close_phi)
{
imm_use_iterator imm_iter;
use_operand_p use_p;
gimple stmt;
tree res, def = gimple_phi_result (close_phi);
FOR_EACH_IMM_USE_FAST (use_p, imm_iter, def)
if ((stmt = USE_STMT (use_p))
&& gimple_code (stmt) == GIMPLE_ASSIGN
&& (res = gimple_assign_lhs (stmt)))
{
switch (TREE_CODE (res))
{
case VAR_DECL:
case PARM_DECL:
case RESULT_DECL:
return res;
case ARRAY_REF:
case MEM_REF:
{
tree arg = gimple_phi_arg_def (close_phi, 0);
loop_p nest = loop_containing_stmt (SSA_NAME_DEF_STMT (arg));
/* FIXME: this restriction is for id-{24,25}.f and
could be handled by duplicating the computation of
array indices before the loop of the close_phi. */
if (for_each_index (&res, dr_indices_valid_in_loop, &nest))
return res;
}
/* Fallthru. */
default:
continue;
}
}
return NULL_TREE;
}
/* Rewrite out of SSA the reduction described by the loop phi nodes
IN, and the close phi nodes OUT. IN and OUT are structured by loop
levels like this:
IN: stmt, loop_n, ..., loop_0
OUT: stmt, close_n, ..., close_0
the first element is the reduction statement, and the next elements
are the loop and close phi nodes of each of the outer loops. */
static void
translate_scalar_reduction_to_array (scop_p scop,
vec<gimple> in,
vec<gimple> out)
{
gimple loop_stmt;
unsigned int i = out.length () - 1;
tree red = close_phi_written_to_memory (as_a <gphi *> (out[i]));
FOR_EACH_VEC_ELT (in, i, loop_stmt)
{
gimple close_stmt = out[i];
if (i == 0)
{
basic_block bb = split_reduction_stmt (scop, loop_stmt);
poly_bb_p pbb = pbb_from_bb (bb);
PBB_IS_REDUCTION (pbb) = true;
gcc_assert (close_stmt == loop_stmt);
if (!red)
red = create_zero_dim_array
(gimple_assign_lhs (loop_stmt), "Commutative_Associative_Reduction");
translate_scalar_reduction_to_array_for_stmt (scop, red, loop_stmt,
as_a <gphi *> (in[1]));
continue;
}
gphi *loop_phi = as_a <gphi *> (loop_stmt);
gphi *close_phi = as_a <gphi *> (close_stmt);
if (i == in.length () - 1)
{
insert_out_of_ssa_copy (scop, gimple_phi_result (close_phi),
unshare_expr (red), close_phi);
insert_out_of_ssa_copy_on_edge
(scop, edge_initial_value_for_loop_phi (loop_phi),
unshare_expr (red), initial_value_for_loop_phi (loop_phi));
}
remove_phi (loop_phi);
remove_phi (close_phi);
}
}
/* Rewrites out of SSA a commutative reduction at CLOSE_PHI. Returns
true when something has been changed. */
static bool
rewrite_commutative_reductions_out_of_ssa_close_phi (scop_p scop,
gphi *close_phi)
{
bool res;
auto_vec<gimple, 10> in;
auto_vec<gimple, 10> out;
detect_commutative_reduction (scop, close_phi, &in, &out);
res = in.length () > 1;
if (res)
translate_scalar_reduction_to_array (scop, in, out);
return res;
}
/* Rewrites all the commutative reductions from LOOP out of SSA.
Returns true when something has been changed. */
static bool
rewrite_commutative_reductions_out_of_ssa_loop (scop_p scop,
loop_p loop)
{
gphi_iterator gsi;
edge exit = single_exit (loop);
tree res;
bool changed = false;
if (!exit)
return false;
for (gsi = gsi_start_phis (exit->dest); !gsi_end_p (gsi); gsi_next (&gsi))
if ((res = gimple_phi_result (gsi.phi ()))
&& !virtual_operand_p (res)
&& !scev_analyzable_p (res, SCOP_REGION (scop)))
changed |= rewrite_commutative_reductions_out_of_ssa_close_phi
(scop, gsi.phi ());
return changed;
}
/* Rewrites all the commutative reductions from SCOP out of SSA. */
static void
rewrite_commutative_reductions_out_of_ssa (scop_p scop)
{
loop_p loop;
bool changed = false;
sese region = SCOP_REGION (scop);
FOR_EACH_LOOP (loop, 0)
if (loop_in_sese_p (loop, region))
changed |= rewrite_commutative_reductions_out_of_ssa_loop (scop, loop);
if (changed)
{
scev_reset_htab ();
gsi_commit_edge_inserts ();
update_ssa (TODO_update_ssa);
#ifdef ENABLE_CHECKING
verify_loop_closed_ssa (true);
#endif
}
}
/* Can all ivs be represented by a signed integer?
As CLooG might generate negative values in its expressions, signed loop ivs
are required in the backend. */
static bool
scop_ivs_can_be_represented (scop_p scop)
{
loop_p loop;
gphi_iterator psi;
bool result = true;
FOR_EACH_LOOP (loop, 0)
{
if (!loop_in_sese_p (loop, SCOP_REGION (scop)))
continue;
for (psi = gsi_start_phis (loop->header);
!gsi_end_p (psi); gsi_next (&psi))
{
gphi *phi = psi.phi ();
tree res = PHI_RESULT (phi);
tree type = TREE_TYPE (res);
if (TYPE_UNSIGNED (type)
&& TYPE_PRECISION (type) >= TYPE_PRECISION (long_long_integer_type_node))
{
result = false;
break;
}
}
if (!result)
break;
}
return result;
}
/* Builds the polyhedral representation for a SESE region. */
void
build_poly_scop (scop_p scop)
{
sese region = SCOP_REGION (scop);
graphite_dim_t max_dim;
build_scop_bbs (scop);
/* FIXME: This restriction is needed to avoid a problem in CLooG.
Once CLooG is fixed, remove this guard. Anyways, it makes no
sense to optimize a scop containing only PBBs that do not belong
to any loops. */
if (nb_pbbs_in_loops (scop) == 0)
return;
if (!scop_ivs_can_be_represented (scop))
return;
if (flag_associative_math)
rewrite_commutative_reductions_out_of_ssa (scop);
build_sese_loop_nests (region);
/* Record all conditions in REGION. */
sese_dom_walker (CDI_DOMINATORS, region).walk (cfun->cfg->x_entry_block_ptr);
find_scop_parameters (scop);
max_dim = PARAM_VALUE (PARAM_GRAPHITE_MAX_NB_SCOP_PARAMS);
if (scop_nb_params (scop) > max_dim)
return;
build_scop_iteration_domain (scop);
build_scop_context (scop);
add_conditions_to_constraints (scop);
/* Rewrite out of SSA only after having translated the
representation to the polyhedral representation to avoid scev
analysis failures. That means that these functions will insert
new data references that they create in the right place. */
rewrite_reductions_out_of_ssa (scop);
rewrite_cross_bb_scalar_deps_out_of_ssa (scop);
build_scop_drs (scop);
scop_to_lst (scop);
build_scop_scattering (scop);
/* This SCoP has been translated to the polyhedral
representation. */
POLY_SCOP_P (scop) = true;
}
#endif