blob: ec883126ad86d86a2c2dafee4592b8d83e2ed917 [file] [log] [blame]
/* Convert a program in SSA form into Normal form.
Copyright (C) 2004-2022 Free Software Foundation, Inc.
Contributed by Andrew Macleod <amacleod@redhat.com>
This file is part of GCC.
GCC is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3, or (at your option)
any later version.
GCC is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with GCC; see the file COPYING3. If not see
<http://www.gnu.org/licenses/>. */
#include "config.h"
#include "system.h"
#include "coretypes.h"
#include "backend.h"
#include "rtl.h"
#include "tree.h"
#include "gimple.h"
#include "cfghooks.h"
#include "ssa.h"
#include "tree-ssa.h"
#include "memmodel.h"
#include "emit-rtl.h"
#include "gimple-pretty-print.h"
#include "diagnostic-core.h"
#include "tree-dfa.h"
#include "stor-layout.h"
#include "cfgrtl.h"
#include "cfganal.h"
#include "tree-eh.h"
#include "gimple-iterator.h"
#include "tree-cfg.h"
#include "dumpfile.h"
#include "tree-ssa-live.h"
#include "tree-ssa-ter.h"
#include "tree-ssa-coalesce.h"
#include "tree-outof-ssa.h"
#include "dojump.h"
/* FIXME: A lot of code here deals with expanding to RTL. All that code
should be in cfgexpand.cc. */
#include "explow.h"
#include "expr.h"
/* Return TRUE if expression STMT is suitable for replacement. */
bool
ssa_is_replaceable_p (gimple *stmt)
{
use_operand_p use_p;
tree def;
gimple *use_stmt;
/* Only consider modify stmts. */
if (!is_gimple_assign (stmt))
return false;
/* If the statement may throw an exception, it cannot be replaced. */
if (stmt_could_throw_p (cfun, stmt))
return false;
/* Punt if there is more than 1 def. */
def = SINGLE_SSA_TREE_OPERAND (stmt, SSA_OP_DEF);
if (!def)
return false;
/* Only consider definitions which have a single use. */
if (!single_imm_use (def, &use_p, &use_stmt))
return false;
/* Used in this block, but at the TOP of the block, not the end. */
if (gimple_code (use_stmt) == GIMPLE_PHI)
return false;
/* There must be no VDEFs. */
if (gimple_vdef (stmt))
return false;
/* Float expressions must go through memory if float-store is on. */
if (flag_float_store
&& FLOAT_TYPE_P (TREE_TYPE (def)))
return false;
/* An assignment with a register variable on the RHS is not
replaceable. */
if (gimple_assign_rhs_code (stmt) == VAR_DECL
&& DECL_HARD_REGISTER (gimple_assign_rhs1 (stmt)))
return false;
/* No function calls can be replaced. */
if (is_gimple_call (stmt))
return false;
/* Leave any stmt with volatile operands alone as well. */
if (gimple_has_volatile_ops (stmt))
return false;
return true;
}
/* Used to hold all the components required to do SSA PHI elimination.
The node and pred/succ list is a simple linear list of nodes and
edges represented as pairs of nodes.
The predecessor and successor list: Nodes are entered in pairs, where
[0] ->PRED, [1]->SUCC. All the even indexes in the array represent
predecessors, all the odd elements are successors.
Rationale:
When implemented as bitmaps, very large programs SSA->Normal times were
being dominated by clearing the interference graph.
Typically this list of edges is extremely small since it only includes
PHI results and uses from a single edge which have not coalesced with
each other. This means that no virtual PHI nodes are included, and
empirical evidence suggests that the number of edges rarely exceed
3, and in a bootstrap of GCC, the maximum size encountered was 7.
This also limits the number of possible nodes that are involved to
rarely more than 6, and in the bootstrap of gcc, the maximum number
of nodes encountered was 12. */
class elim_graph
{
public:
elim_graph (var_map map);
/* Size of the elimination vectors. */
int size;
/* List of nodes in the elimination graph. */
auto_vec<int> nodes;
/* The predecessor and successor edge list. */
auto_vec<int> edge_list;
/* Source locus on each edge */
auto_vec<location_t> edge_locus;
/* Visited vector. */
auto_sbitmap visited;
/* Stack for visited nodes. */
auto_vec<int> stack;
/* The variable partition map. */
var_map map;
/* Edge being eliminated by this graph. */
edge e;
/* List of constant copies to emit. These are pushed on in pairs. */
auto_vec<int> const_dests;
auto_vec<tree> const_copies;
/* Source locations for any constant copies. */
auto_vec<location_t> copy_locus;
};
/* For an edge E find out a good source location to associate with
instructions inserted on edge E. If E has an implicit goto set,
use its location. Otherwise search instructions in predecessors
of E for a location, and use that one. That makes sense because
we insert on edges for PHI nodes, and effects of PHIs happen on
the end of the predecessor conceptually. An exception is made
for EH edges because we don't want to drag the source location
of unrelated statements at the beginning of handlers; they would
be further reused for various EH constructs, which would damage
the coverage information. */
static void
set_location_for_edge (edge e)
{
if (e->goto_locus)
set_curr_insn_location (e->goto_locus);
else if (e->flags & EDGE_EH)
{
basic_block bb = e->dest;
gimple_stmt_iterator gsi;
do
{
for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
{
gimple *stmt = gsi_stmt (gsi);
if (is_gimple_debug (stmt))
continue;
if (gimple_has_location (stmt) || gimple_block (stmt))
{
set_curr_insn_location (gimple_location (stmt));
return;
}
}
/* Nothing found in this basic block. Make a half-assed attempt
to continue with another block. */
if (single_succ_p (bb))
bb = single_succ (bb);
else
bb = e->dest;
}
while (bb != e->dest);
}
else
{
basic_block bb = e->src;
gimple_stmt_iterator gsi;
do
{
for (gsi = gsi_last_bb (bb); !gsi_end_p (gsi); gsi_prev (&gsi))
{
gimple *stmt = gsi_stmt (gsi);
if (is_gimple_debug (stmt))
continue;
if (gimple_has_location (stmt) || gimple_block (stmt))
{
set_curr_insn_location (gimple_location (stmt));
return;
}
}
/* Nothing found in this basic block. Make a half-assed attempt
to continue with another block. */
if (single_pred_p (bb))
bb = single_pred (bb);
else
bb = e->src;
}
while (bb != e->src);
}
}
/* Emit insns to copy SRC into DEST converting SRC if necessary. As
SRC/DEST might be BLKmode memory locations SIZEEXP is a tree from
which we deduce the size to copy in that case. */
static inline rtx_insn *
emit_partition_copy (rtx dest, rtx src, int unsignedsrcp, tree sizeexp)
{
start_sequence ();
if (GET_MODE (src) != VOIDmode && GET_MODE (src) != GET_MODE (dest))
src = convert_to_mode (GET_MODE (dest), src, unsignedsrcp);
if (GET_MODE (src) == BLKmode)
{
gcc_assert (GET_MODE (dest) == BLKmode);
emit_block_move (dest, src, expr_size (sizeexp), BLOCK_OP_NORMAL);
}
else
emit_move_insn (dest, src);
do_pending_stack_adjust ();
rtx_insn *seq = get_insns ();
end_sequence ();
return seq;
}
/* Insert a copy instruction from partition SRC to DEST onto edge E. */
static void
insert_partition_copy_on_edge (edge e, int dest, int src, location_t locus)
{
tree var;
if (dump_file && (dump_flags & TDF_DETAILS))
{
fprintf (dump_file,
"Inserting a partition copy on edge BB%d->BB%d : "
"PART.%d = PART.%d",
e->src->index,
e->dest->index, dest, src);
fprintf (dump_file, "\n");
}
gcc_assert (SA.partition_to_pseudo[dest]);
gcc_assert (SA.partition_to_pseudo[src]);
set_location_for_edge (e);
/* If a locus is provided, override the default. */
if (locus)
set_curr_insn_location (locus);
var = partition_to_var (SA.map, src);
rtx_insn *seq = emit_partition_copy (copy_rtx (SA.partition_to_pseudo[dest]),
copy_rtx (SA.partition_to_pseudo[src]),
TYPE_UNSIGNED (TREE_TYPE (var)),
var);
insert_insn_on_edge (seq, e);
}
/* Insert a copy instruction from expression SRC to partition DEST
onto edge E. */
static void
insert_value_copy_on_edge (edge e, int dest, tree src, location_t locus)
{
rtx dest_rtx, seq, x;
machine_mode dest_mode, src_mode;
int unsignedp;
if (dump_file && (dump_flags & TDF_DETAILS))
{
fprintf (dump_file,
"Inserting a value copy on edge BB%d->BB%d : PART.%d = ",
e->src->index,
e->dest->index, dest);
print_generic_expr (dump_file, src, TDF_SLIM);
fprintf (dump_file, "\n");
}
dest_rtx = copy_rtx (SA.partition_to_pseudo[dest]);
gcc_assert (dest_rtx);
set_location_for_edge (e);
/* If a locus is provided, override the default. */
if (locus)
set_curr_insn_location (locus);
start_sequence ();
tree name = partition_to_var (SA.map, dest);
src_mode = TYPE_MODE (TREE_TYPE (src));
dest_mode = GET_MODE (dest_rtx);
gcc_assert (src_mode == TYPE_MODE (TREE_TYPE (name)));
gcc_assert (!REG_P (dest_rtx)
|| dest_mode == promote_ssa_mode (name, &unsignedp));
if (src_mode != dest_mode)
{
x = expand_expr (src, NULL, src_mode, EXPAND_NORMAL);
x = convert_modes (dest_mode, src_mode, x, unsignedp);
}
else if (src_mode == BLKmode)
{
x = dest_rtx;
store_expr (src, x, 0, false, false);
}
else
x = expand_expr (src, dest_rtx, dest_mode, EXPAND_NORMAL);
if (x != dest_rtx)
emit_move_insn (dest_rtx, x);
do_pending_stack_adjust ();
seq = get_insns ();
end_sequence ();
insert_insn_on_edge (seq, e);
}
/* Insert a copy instruction from RTL expression SRC to partition DEST
onto edge E. */
static void
insert_rtx_to_part_on_edge (edge e, int dest, rtx src, int unsignedsrcp,
location_t locus)
{
if (dump_file && (dump_flags & TDF_DETAILS))
{
fprintf (dump_file,
"Inserting a temp copy on edge BB%d->BB%d : PART.%d = ",
e->src->index,
e->dest->index, dest);
print_simple_rtl (dump_file, src);
fprintf (dump_file, "\n");
}
gcc_assert (SA.partition_to_pseudo[dest]);
set_location_for_edge (e);
/* If a locus is provided, override the default. */
if (locus)
set_curr_insn_location (locus);
/* We give the destination as sizeexp in case src/dest are BLKmode
mems. Usually we give the source. As we result from SSA names
the left and right size should be the same (and no WITH_SIZE_EXPR
involved), so it doesn't matter. */
rtx_insn *seq = emit_partition_copy (copy_rtx (SA.partition_to_pseudo[dest]),
src, unsignedsrcp,
partition_to_var (SA.map, dest));
insert_insn_on_edge (seq, e);
}
/* Insert a copy instruction from partition SRC to RTL lvalue DEST
onto edge E. */
static void
insert_part_to_rtx_on_edge (edge e, rtx dest, int src, location_t locus)
{
tree var;
if (dump_file && (dump_flags & TDF_DETAILS))
{
fprintf (dump_file,
"Inserting a temp copy on edge BB%d->BB%d : ",
e->src->index,
e->dest->index);
print_simple_rtl (dump_file, dest);
fprintf (dump_file, "= PART.%d\n", src);
}
gcc_assert (SA.partition_to_pseudo[src]);
set_location_for_edge (e);
/* If a locus is provided, override the default. */
if (locus)
set_curr_insn_location (locus);
var = partition_to_var (SA.map, src);
rtx_insn *seq = emit_partition_copy (dest,
copy_rtx (SA.partition_to_pseudo[src]),
TYPE_UNSIGNED (TREE_TYPE (var)),
var);
insert_insn_on_edge (seq, e);
}
/* Create an elimination graph for map. */
elim_graph::elim_graph (var_map map) :
nodes (30), edge_list (20), edge_locus (10), visited (map->num_partitions),
stack (30), map (map), const_dests (20), const_copies (20), copy_locus (10)
{
}
/* Empty elimination graph G. */
static inline void
clear_elim_graph (elim_graph *g)
{
g->nodes.truncate (0);
g->edge_list.truncate (0);
g->edge_locus.truncate (0);
}
/* Return the number of nodes in graph G. */
static inline int
elim_graph_size (elim_graph *g)
{
return g->nodes.length ();
}
/* Add NODE to graph G, if it doesn't exist already. */
static inline void
elim_graph_add_node (elim_graph *g, int node)
{
int x;
int t;
FOR_EACH_VEC_ELT (g->nodes, x, t)
if (t == node)
return;
g->nodes.safe_push (node);
}
/* Add the edge PRED->SUCC to graph G. */
static inline void
elim_graph_add_edge (elim_graph *g, int pred, int succ, location_t locus)
{
g->edge_list.safe_push (pred);
g->edge_list.safe_push (succ);
g->edge_locus.safe_push (locus);
}
/* Remove an edge from graph G for which NODE is the predecessor, and
return the successor node. -1 is returned if there is no such edge. */
static inline int
elim_graph_remove_succ_edge (elim_graph *g, int node, location_t *locus)
{
int y;
unsigned x;
for (x = 0; x < g->edge_list.length (); x += 2)
if (g->edge_list[x] == node)
{
g->edge_list[x] = -1;
y = g->edge_list[x + 1];
g->edge_list[x + 1] = -1;
*locus = g->edge_locus[x / 2];
g->edge_locus[x / 2] = UNKNOWN_LOCATION;
return y;
}
*locus = UNKNOWN_LOCATION;
return -1;
}
/* Find all the nodes in GRAPH which are successors to NODE in the
edge list. VAR will hold the partition number found. CODE is the
code fragment executed for every node found. */
#define FOR_EACH_ELIM_GRAPH_SUCC(GRAPH, NODE, VAR, LOCUS, CODE) \
do { \
unsigned x_; \
int y_; \
for (x_ = 0; x_ < (GRAPH)->edge_list.length (); x_ += 2) \
{ \
y_ = (GRAPH)->edge_list[x_]; \
if (y_ != (NODE)) \
continue; \
(void) ((VAR) = (GRAPH)->edge_list[x_ + 1]); \
(void) ((LOCUS) = (GRAPH)->edge_locus[x_ / 2]); \
CODE; \
} \
} while (0)
/* Find all the nodes which are predecessors of NODE in the edge list for
GRAPH. VAR will hold the partition number found. CODE is the
code fragment executed for every node found. */
#define FOR_EACH_ELIM_GRAPH_PRED(GRAPH, NODE, VAR, LOCUS, CODE) \
do { \
unsigned x_; \
int y_; \
for (x_ = 0; x_ < (GRAPH)->edge_list.length (); x_ += 2) \
{ \
y_ = (GRAPH)->edge_list[x_ + 1]; \
if (y_ != (NODE)) \
continue; \
(void) ((VAR) = (GRAPH)->edge_list[x_]); \
(void) ((LOCUS) = (GRAPH)->edge_locus[x_ / 2]); \
CODE; \
} \
} while (0)
/* Add T to elimination graph G. */
static inline void
eliminate_name (elim_graph *g, int T)
{
elim_graph_add_node (g, T);
}
/* Return true if this phi argument T should have a copy queued when using
var_map MAP. PHI nodes should contain only ssa_names and invariants. A
test for ssa_name is definitely simpler, but don't let invalid contents
slip through in the meantime. */
static inline bool
queue_phi_copy_p (var_map map, tree t)
{
if (TREE_CODE (t) == SSA_NAME)
{
if (var_to_partition (map, t) == NO_PARTITION)
return true;
return false;
}
gcc_checking_assert (is_gimple_min_invariant (t));
return true;
}
/* Build elimination graph G for basic block BB on incoming PHI edge
G->e. */
static void
eliminate_build (elim_graph *g)
{
tree Ti;
int p0, pi;
gphi_iterator gsi;
clear_elim_graph (g);
for (gsi = gsi_start_phis (g->e->dest); !gsi_end_p (gsi); gsi_next (&gsi))
{
gphi *phi = gsi.phi ();
location_t locus;
p0 = var_to_partition (g->map, gimple_phi_result (phi));
/* Ignore results which are not in partitions. */
if (p0 == NO_PARTITION)
continue;
Ti = PHI_ARG_DEF (phi, g->e->dest_idx);
/* See set_location_for_edge for the rationale. */
if (g->e->flags & EDGE_EH)
locus = UNKNOWN_LOCATION;
else
locus = gimple_phi_arg_location_from_edge (phi, g->e);
/* If this argument is a constant, or a SSA_NAME which is being
left in SSA form, just queue a copy to be emitted on this
edge. */
if (queue_phi_copy_p (g->map, Ti))
{
/* Save constant copies until all other copies have been emitted
on this edge. */
g->const_dests.safe_push (p0);
g->const_copies.safe_push (Ti);
g->copy_locus.safe_push (locus);
}
else
{
pi = var_to_partition (g->map, Ti);
if (p0 != pi)
{
eliminate_name (g, p0);
eliminate_name (g, pi);
elim_graph_add_edge (g, p0, pi, locus);
}
}
}
}
/* Push successors of T onto the elimination stack for G. */
static void
elim_forward (elim_graph *g, int T)
{
int S;
location_t locus;
bitmap_set_bit (g->visited, T);
FOR_EACH_ELIM_GRAPH_SUCC (g, T, S, locus,
{
if (!bitmap_bit_p (g->visited, S))
elim_forward (g, S);
});
g->stack.safe_push (T);
}
/* Return 1 if there unvisited predecessors of T in graph G. */
static int
elim_unvisited_predecessor (elim_graph *g, int T)
{
int P;
location_t locus;
FOR_EACH_ELIM_GRAPH_PRED (g, T, P, locus,
{
if (!bitmap_bit_p (g->visited, P))
return 1;
});
return 0;
}
/* Process predecessors first, and insert a copy. */
static void
elim_backward (elim_graph *g, int T)
{
int P;
location_t locus;
bitmap_set_bit (g->visited, T);
FOR_EACH_ELIM_GRAPH_PRED (g, T, P, locus,
{
if (!bitmap_bit_p (g->visited, P))
{
elim_backward (g, P);
insert_partition_copy_on_edge (g->e, P, T, locus);
}
});
}
/* Allocate a new pseudo register usable for storing values sitting
in NAME (a decl or SSA name), i.e. with matching mode and attributes. */
static rtx
get_temp_reg (tree name)
{
tree type = TREE_TYPE (name);
int unsignedp;
machine_mode reg_mode = promote_ssa_mode (name, &unsignedp);
if (reg_mode == BLKmode)
return assign_temp (type, 0, 0);
rtx x = gen_reg_rtx (reg_mode);
if (POINTER_TYPE_P (type))
mark_reg_pointer (x, TYPE_ALIGN (TREE_TYPE (type)));
return x;
}
/* Insert required copies for T in graph G. Check for a strongly connected
region, and create a temporary to break the cycle if one is found. */
static void
elim_create (elim_graph *g, int T)
{
int P, S;
location_t locus;
if (elim_unvisited_predecessor (g, T))
{
tree var = partition_to_var (g->map, T);
rtx U = get_temp_reg (var);
int unsignedsrcp = TYPE_UNSIGNED (TREE_TYPE (var));
insert_part_to_rtx_on_edge (g->e, U, T, UNKNOWN_LOCATION);
FOR_EACH_ELIM_GRAPH_PRED (g, T, P, locus,
{
if (!bitmap_bit_p (g->visited, P))
{
elim_backward (g, P);
insert_rtx_to_part_on_edge (g->e, P, U, unsignedsrcp, locus);
}
});
}
else
{
S = elim_graph_remove_succ_edge (g, T, &locus);
if (S != -1)
{
bitmap_set_bit (g->visited, T);
insert_partition_copy_on_edge (g->e, T, S, locus);
}
}
}
/* Eliminate all the phi nodes on edge E in graph G. */
static void
eliminate_phi (edge e, elim_graph *g)
{
int x;
gcc_assert (g->const_copies.length () == 0);
gcc_assert (g->copy_locus.length () == 0);
/* Abnormal edges already have everything coalesced. */
if (e->flags & EDGE_ABNORMAL)
return;
g->e = e;
eliminate_build (g);
if (elim_graph_size (g) != 0)
{
int part;
bitmap_clear (g->visited);
g->stack.truncate (0);
FOR_EACH_VEC_ELT (g->nodes, x, part)
{
if (!bitmap_bit_p (g->visited, part))
elim_forward (g, part);
}
bitmap_clear (g->visited);
while (g->stack.length () > 0)
{
x = g->stack.pop ();
if (!bitmap_bit_p (g->visited, x))
elim_create (g, x);
}
}
/* If there are any pending constant copies, issue them now. */
while (g->const_copies.length () > 0)
{
int dest;
tree src;
location_t locus;
src = g->const_copies.pop ();
dest = g->const_dests.pop ();
locus = g->copy_locus.pop ();
insert_value_copy_on_edge (e, dest, src, locus);
}
}
/* Remove each argument from PHI. If an arg was the last use of an SSA_NAME,
check to see if this allows another PHI node to be removed. */
static void
remove_gimple_phi_args (gphi *phi)
{
use_operand_p arg_p;
ssa_op_iter iter;
if (dump_file && (dump_flags & TDF_DETAILS))
{
fprintf (dump_file, "Removing Dead PHI definition: ");
print_gimple_stmt (dump_file, phi, 0, TDF_SLIM);
}
FOR_EACH_PHI_ARG (arg_p, phi, iter, SSA_OP_USE)
{
tree arg = USE_FROM_PTR (arg_p);
if (TREE_CODE (arg) == SSA_NAME)
{
/* Remove the reference to the existing argument. */
SET_USE (arg_p, NULL_TREE);
if (has_zero_uses (arg))
{
gimple *stmt;
gimple_stmt_iterator gsi;
stmt = SSA_NAME_DEF_STMT (arg);
/* Also remove the def if it is a PHI node. */
if (gimple_code (stmt) == GIMPLE_PHI)
{
remove_gimple_phi_args (as_a <gphi *> (stmt));
gsi = gsi_for_stmt (stmt);
remove_phi_node (&gsi, true);
}
}
}
}
}
/* Remove any PHI node which is a virtual PHI, or a PHI with no uses. */
static void
eliminate_useless_phis (void)
{
basic_block bb;
gphi_iterator gsi;
tree result;
FOR_EACH_BB_FN (bb, cfun)
{
for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); )
{
gphi *phi = gsi.phi ();
result = gimple_phi_result (phi);
if (virtual_operand_p (result))
remove_phi_node (&gsi, true);
else
{
/* Also remove real PHIs with no uses. */
if (has_zero_uses (result))
{
remove_gimple_phi_args (phi);
remove_phi_node (&gsi, true);
}
else
gsi_next (&gsi);
}
}
}
}
/* This function will rewrite the current program using the variable mapping
found in MAP. If the replacement vector VALUES is provided, any
occurrences of partitions with non-null entries in the vector will be
replaced with the expression in the vector instead of its mapped
variable. */
static void
rewrite_trees (var_map map)
{
if (!flag_checking)
return;
basic_block bb;
/* Search for PHIs where the destination has no partition, but one
or more arguments has a partition. This should not happen and can
create incorrect code. */
FOR_EACH_BB_FN (bb, cfun)
{
gphi_iterator gsi;
for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi))
{
gphi *phi = gsi.phi ();
tree T0 = var_to_partition_to_var (map, gimple_phi_result (phi));
if (T0 == NULL_TREE)
{
size_t i;
for (i = 0; i < gimple_phi_num_args (phi); i++)
{
tree arg = PHI_ARG_DEF (phi, i);
if (TREE_CODE (arg) == SSA_NAME
&& var_to_partition (map, arg) != NO_PARTITION)
{
fprintf (stderr, "Argument of PHI is in a partition :(");
print_generic_expr (stderr, arg, TDF_SLIM);
fprintf (stderr, "), but the result is not :");
print_gimple_stmt (stderr, phi, 0, TDF_SLIM);
internal_error ("SSA corruption");
}
}
}
}
}
}
/* Create a default def for VAR. */
static void
create_default_def (tree var, void *arg ATTRIBUTE_UNUSED)
{
if (!is_gimple_reg (var))
return;
tree ssa = get_or_create_ssa_default_def (cfun, var);
gcc_assert (ssa);
}
/* Call CALLBACK for all PARM_DECLs and RESULT_DECLs for which
assign_parms may ask for a default partition. */
static void
for_all_parms (void (*callback)(tree var, void *arg), void *arg)
{
for (tree var = DECL_ARGUMENTS (current_function_decl); var;
var = DECL_CHAIN (var))
callback (var, arg);
if (!VOID_TYPE_P (TREE_TYPE (DECL_RESULT (current_function_decl))))
callback (DECL_RESULT (current_function_decl), arg);
if (cfun->static_chain_decl)
callback (cfun->static_chain_decl, arg);
}
/* We need to pass two arguments to set_parm_default_def_partition,
but for_all_parms only supports one. Use a pair. */
typedef std::pair<var_map, bitmap> parm_default_def_partition_arg;
/* Set in ARG's PARTS bitmap the bit corresponding to the partition in
ARG's MAP containing VAR's default def. */
static void
set_parm_default_def_partition (tree var, void *arg_)
{
parm_default_def_partition_arg *arg = (parm_default_def_partition_arg *)arg_;
var_map map = arg->first;
bitmap parts = arg->second;
if (!is_gimple_reg (var))
return;
tree ssa = ssa_default_def (cfun, var);
gcc_assert (ssa);
int version = var_to_partition (map, ssa);
gcc_assert (version != NO_PARTITION);
bool changed = bitmap_set_bit (parts, version);
gcc_assert (changed);
}
/* Allocate and return a bitmap that has a bit set for each partition
that contains a default def for a parameter. */
static bitmap
get_parm_default_def_partitions (var_map map)
{
bitmap parm_default_def_parts = BITMAP_ALLOC (NULL);
parm_default_def_partition_arg
arg = std::make_pair (map, parm_default_def_parts);
for_all_parms (set_parm_default_def_partition, &arg);
return parm_default_def_parts;
}
/* Allocate and return a bitmap that has a bit set for each partition
that contains an undefined value. */
static bitmap
get_undefined_value_partitions (var_map map)
{
bitmap undefined_value_parts = BITMAP_ALLOC (NULL);
for (unsigned int i = 1; i < num_ssa_names; i++)
{
tree var = ssa_name (i);
if (var
&& !virtual_operand_p (var)
&& !has_zero_uses (var)
&& ssa_undefined_value_p (var))
{
const int p = var_to_partition (map, var);
if (p != NO_PARTITION)
bitmap_set_bit (undefined_value_parts, p);
}
}
return undefined_value_parts;
}
/* Given the out-of-ssa info object SA (with prepared partitions)
eliminate all phi nodes in all basic blocks. Afterwards no
basic block will have phi nodes anymore and there are possibly
some RTL instructions inserted on edges. */
void
expand_phi_nodes (struct ssaexpand *sa)
{
basic_block bb;
elim_graph g (sa->map);
FOR_BB_BETWEEN (bb, ENTRY_BLOCK_PTR_FOR_FN (cfun)->next_bb,
EXIT_BLOCK_PTR_FOR_FN (cfun), next_bb)
if (!gimple_seq_empty_p (phi_nodes (bb)))
{
edge e;
edge_iterator ei;
FOR_EACH_EDGE (e, ei, bb->preds)
eliminate_phi (e, &g);
set_phi_nodes (bb, NULL);
/* We can't redirect EH edges in RTL land, so we need to do this
here. Redirection happens only when splitting is necessary,
which it is only for critical edges, normally. For EH edges
it might also be necessary when the successor has more than
one predecessor. In that case the edge is either required to
be fallthru (which EH edges aren't), or the predecessor needs
to end with a jump (which again, isn't the case with EH edges).
Hence, split all EH edges on which we inserted instructions
and whose successor has multiple predecessors. */
for (ei = ei_start (bb->preds); (e = ei_safe_edge (ei)); )
{
if (e->insns.r && (e->flags & EDGE_EH)
&& !single_pred_p (e->dest))
{
rtx_insn *insns = e->insns.r;
basic_block bb;
e->insns.r = NULL;
bb = split_edge (e);
single_pred_edge (bb)->insns.r = insns;
}
else
ei_next (&ei);
}
}
}
/* Remove the ssa-names in the current function and translate them into normal
compiler variables. PERFORM_TER is true if Temporary Expression Replacement
should also be used. */
static void
remove_ssa_form (bool perform_ter, struct ssaexpand *sa)
{
bitmap values = NULL;
var_map map;
for_all_parms (create_default_def, NULL);
map = init_var_map (num_ssa_names);
coalesce_ssa_name (map);
/* Return to viewing the variable list as just all reference variables after
coalescing has been performed. */
partition_view_normal (map);
if (dump_file && (dump_flags & TDF_DETAILS))
{
fprintf (dump_file, "After Coalescing:\n");
dump_var_map (dump_file, map);
}
if (perform_ter)
{
values = find_replaceable_exprs (map);
if (values && dump_file && (dump_flags & TDF_DETAILS))
dump_replaceable_exprs (dump_file, values);
}
rewrite_trees (map);
sa->map = map;
sa->values = values;
sa->partitions_for_parm_default_defs = get_parm_default_def_partitions (map);
sa->partitions_for_undefined_values = get_undefined_value_partitions (map);
}
/* If not already done so for basic block BB, assign increasing uids
to each of its instructions. */
static void
maybe_renumber_stmts_bb (basic_block bb)
{
unsigned i = 0;
gimple_stmt_iterator gsi;
if (!bb->aux)
return;
bb->aux = NULL;
for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
{
gimple *stmt = gsi_stmt (gsi);
gimple_set_uid (stmt, i);
i++;
}
}
/* Return true if we can determine that the SSA_NAMEs RESULT (a result
of a PHI node) and ARG (one of its arguments) conflict. Return false
otherwise, also when we simply aren't sure. */
static bool
trivially_conflicts_p (basic_block bb, tree result, tree arg)
{
use_operand_p use;
imm_use_iterator imm_iter;
gimple *defa = SSA_NAME_DEF_STMT (arg);
/* If ARG isn't defined in the same block it's too complicated for
our little mind. */
if (gimple_bb (defa) != bb)
return false;
FOR_EACH_IMM_USE_FAST (use, imm_iter, result)
{
gimple *use_stmt = USE_STMT (use);
if (is_gimple_debug (use_stmt))
continue;
/* Now, if there's a use of RESULT that lies outside this basic block,
then there surely is a conflict with ARG. */
if (gimple_bb (use_stmt) != bb)
return true;
if (gimple_code (use_stmt) == GIMPLE_PHI)
continue;
/* The use now is in a real stmt of BB, so if ARG was defined
in a PHI node (like RESULT) both conflict. */
if (gimple_code (defa) == GIMPLE_PHI)
return true;
maybe_renumber_stmts_bb (bb);
/* If the use of RESULT occurs after the definition of ARG,
the two conflict too. */
if (gimple_uid (defa) < gimple_uid (use_stmt))
return true;
}
return false;
}
/* Search every PHI node for arguments associated with backedges which
we can trivially determine will need a copy (the argument is either
not an SSA_NAME or the argument has a different underlying variable
than the PHI result).
Insert a copy from the PHI argument to a new destination at the
end of the block with the backedge to the top of the loop. Update
the PHI argument to reference this new destination. */
static void
insert_backedge_copies (void)
{
basic_block bb;
gphi_iterator gsi;
mark_dfs_back_edges ();
FOR_EACH_BB_FN (bb, cfun)
{
/* Mark block as possibly needing calculation of UIDs. */
bb->aux = &bb->aux;
for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi))
{
gphi *phi = gsi.phi ();
tree result = gimple_phi_result (phi);
size_t i;
if (virtual_operand_p (result))
continue;
for (i = 0; i < gimple_phi_num_args (phi); i++)
{
tree arg = gimple_phi_arg_def (phi, i);
edge e = gimple_phi_arg_edge (phi, i);
/* We are only interested in copies emitted on critical
backedges. */
if (!(e->flags & EDGE_DFS_BACK)
|| !EDGE_CRITICAL_P (e))
continue;
/* If the argument is not an SSA_NAME, then we will need a
constant initialization. If the argument is an SSA_NAME then
a copy statement may be needed. First handle the case
where we cannot insert before the argument definition. */
if (TREE_CODE (arg) != SSA_NAME
|| (gimple_code (SSA_NAME_DEF_STMT (arg)) == GIMPLE_PHI
&& trivially_conflicts_p (bb, result, arg)))
{
tree name;
gassign *stmt;
gimple *last = NULL;
gimple_stmt_iterator gsi2;
gsi2 = gsi_last_bb (gimple_phi_arg_edge (phi, i)->src);
if (!gsi_end_p (gsi2))
last = gsi_stmt (gsi2);
/* In theory the only way we ought to get back to the
start of a loop should be with a COND_EXPR or GOTO_EXPR.
However, better safe than sorry.
If the block ends with a control statement or
something that might throw, then we have to
insert this assignment before the last
statement. Else insert it after the last statement. */
if (last && stmt_ends_bb_p (last))
{
/* If the last statement in the block is the definition
site of the PHI argument, then we can't insert
anything after it. */
if (TREE_CODE (arg) == SSA_NAME
&& SSA_NAME_DEF_STMT (arg) == last)
continue;
}
/* Create a new instance of the underlying variable of the
PHI result. */
name = copy_ssa_name (result);
stmt = gimple_build_assign (name,
gimple_phi_arg_def (phi, i));
/* copy location if present. */
if (gimple_phi_arg_has_location (phi, i))
gimple_set_location (stmt,
gimple_phi_arg_location (phi, i));
/* Insert the new statement into the block and update
the PHI node. */
if (last && stmt_ends_bb_p (last))
gsi_insert_before (&gsi2, stmt, GSI_NEW_STMT);
else
gsi_insert_after (&gsi2, stmt, GSI_NEW_STMT);
SET_PHI_ARG_DEF (phi, i, name);
}
/* Insert a copy before the definition of the backedge value
and adjust all conflicting uses. */
else if (trivially_conflicts_p (bb, result, arg))
{
gimple *def = SSA_NAME_DEF_STMT (arg);
if (gimple_nop_p (def)
|| gimple_code (def) == GIMPLE_PHI)
continue;
tree name = copy_ssa_name (result);
gimple *stmt = gimple_build_assign (name, result);
imm_use_iterator imm_iter;
gimple *use_stmt;
/* The following matches trivially_conflicts_p. */
FOR_EACH_IMM_USE_STMT (use_stmt, imm_iter, result)
{
if (gimple_bb (use_stmt) != bb
|| (gimple_code (use_stmt) != GIMPLE_PHI
&& (maybe_renumber_stmts_bb (bb), true)
&& gimple_uid (use_stmt) > gimple_uid (def)))
{
use_operand_p use;
FOR_EACH_IMM_USE_ON_STMT (use, imm_iter)
SET_USE (use, name);
}
}
gimple_stmt_iterator gsi = gsi_for_stmt (def);
gsi_insert_before (&gsi, stmt, GSI_SAME_STMT);
}
}
}
/* Unmark this block again. */
bb->aux = NULL;
}
}
/* Free all memory associated with going out of SSA form. SA is
the outof-SSA info object. */
void
finish_out_of_ssa (struct ssaexpand *sa)
{
free (sa->partition_to_pseudo);
if (sa->values)
BITMAP_FREE (sa->values);
delete_var_map (sa->map);
BITMAP_FREE (sa->partitions_for_parm_default_defs);
BITMAP_FREE (sa->partitions_for_undefined_values);
memset (sa, 0, sizeof *sa);
}
/* Take the current function out of SSA form, translating PHIs as described in
R. Morgan, ``Building an Optimizing Compiler'',
Butterworth-Heinemann, Boston, MA, 1998. pp 176-186. */
unsigned int
rewrite_out_of_ssa (struct ssaexpand *sa)
{
/* If elimination of a PHI requires inserting a copy on a backedge,
then we will have to split the backedge which has numerous
undesirable performance effects.
A significant number of such cases can be handled here by inserting
copies into the loop itself. */
insert_backedge_copies ();
/* Eliminate PHIs which are of no use, such as virtual or dead phis. */
eliminate_useless_phis ();
if (dump_file && (dump_flags & TDF_DETAILS))
gimple_dump_cfg (dump_file, dump_flags & ~TDF_DETAILS);
remove_ssa_form (flag_tree_ter, sa);
if (dump_file && (dump_flags & TDF_DETAILS))
gimple_dump_cfg (dump_file, dump_flags & ~TDF_DETAILS);
return 0;
}