Commit 37d7ef61 authored by Matthias Braun's avatar Matthias Braun
Browse files

simplify unreachable code elimination:

only set block/phi inputs to bad to kill unreachable code, all other
code will be dead and removed anyway
parent 40c5823f
...@@ -98,6 +98,14 @@ void local_optimize_node(ir_node *n) ...@@ -98,6 +98,14 @@ void local_optimize_node(ir_node *n)
current_ir_graph = rem; current_ir_graph = rem;
} }
static void enqueue_node(ir_node *node, pdeq *waitq)
{
if (get_irn_link(node) == waitq)
return;
pdeq_putr(waitq, node);
set_irn_link(node, waitq);
}
/** /**
* Enqueue all users of a node to a wait queue. * Enqueue all users of a node to a wait queue.
* Handles mode_T nodes. * Handles mode_T nodes.
...@@ -109,10 +117,7 @@ static void enqueue_users(ir_node *n, pdeq *waitq) ...@@ -109,10 +117,7 @@ static void enqueue_users(ir_node *n, pdeq *waitq)
foreach_out_edge(n, edge) { foreach_out_edge(n, edge) {
ir_node *succ = get_edge_src_irn(edge); ir_node *succ = get_edge_src_irn(edge);
if (get_irn_link(succ) != waitq) { enqueue_node(succ, waitq);
pdeq_putr(waitq, succ);
set_irn_link(succ, waitq);
}
if (get_irn_mode(succ) == mode_T) { if (get_irn_mode(succ) == mode_T) {
/* A mode_T node has Proj's. Because most optimizations /* A mode_T node has Proj's. Because most optimizations
run on the Proj's we have to enqueue them also. */ run on the Proj's we have to enqueue them also. */
...@@ -124,18 +129,26 @@ static void enqueue_users(ir_node *n, pdeq *waitq) ...@@ -124,18 +129,26 @@ static void enqueue_users(ir_node *n, pdeq *waitq)
/** /**
* Block-Walker: uses dominance depth to mark dead blocks. * Block-Walker: uses dominance depth to mark dead blocks.
*/ */
static void kill_dead_blocks(ir_node *block, void *env) static void find_unreachable_blocks(ir_node *block, void *env)
{ {
pdeq *waitq = (pdeq*) env; pdeq *waitq = (pdeq*) env;
if (get_Block_dom_depth(block) < 0) { if (get_Block_dom_depth(block) < 0) {
/*
* Note that the new dominance code correctly handles
* the End block, i.e. it is always reachable from Start
*/
ir_graph *irg = get_irn_irg(block); ir_graph *irg = get_irn_irg(block);
enqueue_users(block, waitq); ir_node *end = get_irg_end(irg);
exchange(block, new_r_Bad(irg, mode_BB));
const ir_edge_t *edge;
foreach_block_succ(block, edge) {
const ir_edge_t *edge2;
ir_node *succ_block = get_edge_src_irn(edge);
enqueue_node(succ_block, waitq);
foreach_out_edge(succ_block, edge2) {
ir_node *succ = get_edge_src_irn(edge2);
if (is_Phi(succ))
enqueue_node(succ, waitq);
}
}
enqueue_node(end, waitq);
} }
} }
...@@ -298,6 +311,9 @@ int optimize_graph_df(ir_graph *irg) ...@@ -298,6 +311,9 @@ int optimize_graph_df(ir_graph *irg)
ir_reserve_resources(irg, IR_RESOURCE_IRN_LINK); ir_reserve_resources(irg, IR_RESOURCE_IRN_LINK);
/* Calculate dominance so we can kill unreachable code */
assure_doms(irg);
/* walk over the graph, but don't touch keep-alives */ /* walk over the graph, but don't touch keep-alives */
irg_walk_graph(irg, NULL, opt_walker, waitq); irg_walk_graph(irg, NULL, opt_walker, waitq);
...@@ -305,16 +321,17 @@ int optimize_graph_df(ir_graph *irg) ...@@ -305,16 +321,17 @@ int optimize_graph_df(ir_graph *irg)
* so if it's not empty, the graph has been changed */ * so if it's not empty, the graph has been changed */
changed = !pdeq_empty(waitq); changed = !pdeq_empty(waitq);
do { while (!pdeq_empty(waitq)) {
/* finish the wait queue */ /* finish the wait queue */
while (! pdeq_empty(waitq)) { while (! pdeq_empty(waitq)) {
ir_node *n = (ir_node*)pdeq_getl(waitq); ir_node *n = (ir_node*)pdeq_getl(waitq);
opt_walker(n, waitq); opt_walker(n, waitq);
} }
/* kill newly generated unreachable code */ /* Calculate dominance so we can kill unreachable code */
compute_doms(irg); compute_doms(irg);
irg_block_walk_graph(irg, NULL, kill_dead_blocks, waitq); irg_block_walk_graph(irg, NULL, find_unreachable_blocks, waitq);
} while (! pdeq_empty(waitq)); }
set_irg_doms_inconsistent(irg);
del_pdeq(waitq); del_pdeq(waitq);
......
...@@ -3977,19 +3977,6 @@ static ir_node *transform_node_Cmp(ir_node *n) ...@@ -3977,19 +3977,6 @@ static ir_node *transform_node_Cmp(ir_node *n)
/* the following reassociations work only for == and != */ /* the following reassociations work only for == and != */
if (relation == ir_relation_equal || relation == ir_relation_less_greater) { if (relation == ir_relation_equal || relation == ir_relation_less_greater) {
#if 0 /* Might be not that good in general */
/* a-b == 0 ==> a == b, a-b != 0 ==> a != b */
if (tarval_is_null(tv) && is_Sub(left)) {
right = get_Sub_right(left);
left = get_Sub_left(left);
tv = value_of(right);
changed = 1;
DBG_OPT_ALGSIM0(n, n, FS_OPT_CMP_OP_C);
}
#endif
if (tv != tarval_bad) { if (tv != tarval_bad) {
/* a-c1 == c2 ==> a == c2+c1, a-c1 != c2 ==> a != c2+c1 */ /* a-c1 == c2 ==> a == c2+c1, a-c1 != c2 ==> a != c2+c1 */
if (is_Sub(left)) { if (is_Sub(left)) {
...@@ -4353,35 +4340,34 @@ static ir_node *transform_node_Proj(ir_node *proj) ...@@ -4353,35 +4340,34 @@ static ir_node *transform_node_Proj(ir_node *proj)
return proj; return proj;
} /* transform_node_Proj */ } /* transform_node_Proj */
static bool is_block_unreachable(const ir_node *block)
{
const ir_graph *irg = get_irn_irg(block);
if (!is_irg_state(irg, IR_GRAPH_STATE_BAD_BLOCK))
return false;
return get_Block_dom_depth(block) < 0;
}
static ir_node *transform_node_Block(ir_node *block) static ir_node *transform_node_Block(ir_node *block)
{ {
ir_graph *irg = get_irn_irg(block); ir_graph *irg = get_irn_irg(block);
int arity = get_irn_arity(block);
ir_node *bad = NULL;
int i;
if (!is_irg_state(irg, IR_GRAPH_STATE_BAD_BLOCK)) if (!is_irg_state(irg, IR_GRAPH_STATE_BAD_BLOCK))
return block; return block;
/* don't optimize labeled blocks */
if (has_Block_entity(block))
return block;
if (!get_Block_matured(block))
return block;
/* remove blocks with only Bad inputs (or no inputs) */
{
int i;
int n_cfgpreds = get_Block_n_cfgpreds(block);
for (i = 0; i < n_cfgpreds; ++i) { for (i = 0; i < arity; ++i) {
ir_node *pred = get_Block_cfgpred(block, i); ir_node *pred = get_Block_cfgpred(block, i);
if (!is_Bad(pred)) ir_node *pred_block = get_nodes_block(pred);
break; if (!is_Bad(pred) && !is_block_unreachable(pred_block))
} continue;
/* only bad unreachable inputs? It's unreachable code (unless it is the if (bad == NULL)
* start or end block) */ bad = new_r_Bad(irg, mode_X);
if (i >= n_cfgpreds && block != get_irg_start_block(irg) set_irn_n(block, i, bad);
&& block != get_irg_end_block(irg)) {
return new_r_Bad(irg, mode_BB);
}
} }
return block; return block;
} }
...@@ -4397,12 +4383,12 @@ static ir_node *transform_node_Phi(ir_node *phi) ...@@ -4397,12 +4383,12 @@ static ir_node *transform_node_Phi(ir_node *phi)
/* Set phi-operands for bad-block inputs to bad */ /* Set phi-operands for bad-block inputs to bad */
for (i = 0; i < n; ++i) { for (i = 0; i < n; ++i) {
ir_node *pred = get_Block_cfgpred(block, i); ir_node *pred = get_Block_cfgpred(block, i);
if (!is_Bad(pred)) if (is_Bad(pred) || is_block_unreachable(get_nodes_block(pred))) {
continue;
if (bad == NULL) if (bad == NULL)
bad = new_r_Bad(irg, mode); bad = new_r_Bad(irg, mode);
set_irn_n(phi, i, bad); set_irn_n(phi, i, bad);
} }
}
/* Move Confirms down through Phi nodes. */ /* Move Confirms down through Phi nodes. */
if (mode_is_reference(mode)) { if (mode_is_reference(mode)) {
...@@ -5216,10 +5202,14 @@ static ir_node *transform_node_End(ir_node *n) ...@@ -5216,10 +5202,14 @@ static ir_node *transform_node_End(ir_node *n)
for (i = j = 0; i < n_keepalives; ++i) { for (i = j = 0; i < n_keepalives; ++i) {
ir_node *ka = get_End_keepalive(n, i); ir_node *ka = get_End_keepalive(n, i);
if (is_Bad(ka)) { ir_node *block;
/* no need to keep Bad */ /* no need to keep Bad */
if (is_Bad(ka))
continue;
/* dont keep unreachable code */
block = is_Block(ka) ? ka : get_nodes_block(ka);
if (is_block_unreachable(block))
continue; continue;
}
in[j++] = ka; in[j++] = ka;
} }
if (j != n_keepalives) if (j != n_keepalives)
...@@ -6194,18 +6184,6 @@ void visit_all_identities(ir_graph *irg, irg_walk_func visit, void *env) ...@@ -6194,18 +6184,6 @@ void visit_all_identities(ir_graph *irg, irg_walk_func visit, void *env)
current_ir_graph = rem; current_ir_graph = rem;
} /* visit_all_identities */ } /* visit_all_identities */
static bool is_unreachable(ir_node *node)
{
ir_op *op = get_irn_op(node);
/* Code in "Bad" blocks is unreachable and can be replaced by Bad */
if (op != op_Block && is_Bad(get_nodes_block(node))) {
return true;
}
return false;
}
/** /**
* These optimizations deallocate nodes from the obstack. * These optimizations deallocate nodes from the obstack.
* It can only be called if it is guaranteed that no other nodes * It can only be called if it is guaranteed that no other nodes
...@@ -6223,15 +6201,6 @@ ir_node *optimize_node(ir_node *n) ...@@ -6223,15 +6201,6 @@ ir_node *optimize_node(ir_node *n)
/* Always optimize Phi nodes: part of the construction. */ /* Always optimize Phi nodes: part of the construction. */
if ((!get_opt_optimize()) && (iro != iro_Phi)) return n; if ((!get_opt_optimize()) && (iro != iro_Phi)) return n;
/* Remove nodes with dead (Bad) input.
Run always for transformation induced Bads. */
if (is_unreachable(n)) {
ir_mode *mode = get_irn_mode(n);
edges_node_deleted(n);
irg_kill_node(irg, n);
return new_r_Bad(irg, mode);
}
/* constant expression evaluation / constant folding */ /* constant expression evaluation / constant folding */
if (get_opt_constant_folding()) { if (get_opt_constant_folding()) {
/* neither constants nor Tuple values can be evaluated */ /* neither constants nor Tuple values can be evaluated */
...@@ -6330,14 +6299,6 @@ ir_node *optimize_in_place_2(ir_node *n) ...@@ -6330,14 +6299,6 @@ ir_node *optimize_in_place_2(ir_node *n)
if (iro == iro_Deleted) if (iro == iro_Deleted)
return n; return n;
/* Remove nodes with dead (Bad) input.
Run always for transformation induced Bads. */
if (is_unreachable(n)) {
ir_graph *irg = get_irn_irg(n);
ir_mode *mode = get_irn_mode(n);
return new_r_Bad(irg, mode);
}
/* constant expression evaluation / constant folding */ /* constant expression evaluation / constant folding */
if (get_opt_constant_folding()) { if (get_opt_constant_folding()) {
/* neither constants nor Tuple values can be evaluated */ /* neither constants nor Tuple values can be evaluated */
......
Supports Markdown
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment