ldstopt.c 55.9 KB
Newer Older
Christian Würdig's avatar
Christian Würdig committed
1
2
/*
 * This file is part of libFirm.
3
 * Copyright (C) 2012 University of Karlsruhe.
Christian Würdig's avatar
Christian Würdig committed
4
5
 */

Michael Beck's avatar
Michael Beck committed
6
7
8
9
/**
 * @file
 * @brief   Load/Store optimizations.
 * @author  Michael Beck
Michael Beck's avatar
Michael Beck committed
10
 */
Matthias Braun's avatar
Matthias Braun committed
11
#include "config.h"
Michael Beck's avatar
Michael Beck committed
12

13
#include <string.h>
Michael Beck's avatar
Michael Beck committed
14

15
#include "iroptimize.h"
16
17
18
19
20
21
22
#include "irnode_t.h"
#include "irgraph_t.h"
#include "irmode_t.h"
#include "iropt_t.h"
#include "ircons_t.h"
#include "irgmod.h"
#include "irgwalk.h"
23
#include "irtools.h"
24
25
26
27
#include "tv_t.h"
#include "dbginfo_t.h"
#include "iropt_dbg.h"
#include "irflag_t.h"
28
#include "array_t.h"
29
#include "irhooks.h"
Michael Beck's avatar
BugFix:    
Michael Beck committed
30
#include "iredges.h"
31
#include "irpass.h"
32
#include "irmemory.h"
33
#include "irnodehashmap.h"
34
#include "irgopt.h"
35
#include "set.h"
Matthias Braun's avatar
Matthias Braun committed
36
#include "be.h"
37
38
39
40
#include "debug.h"

/** The debug handle. */
DEBUG_ONLY(static firm_dbg_module_t *dbg;)
41

42
#define MAX_PROJ MAX(MAX((long)pn_Load_max, (long)pn_Store_max), (long)pn_Call_max)
Michael Beck's avatar
Michael Beck committed
43

44
enum changes_t {
45
46
	DF_CHANGED = 1,       /**< data flow changed */
	CF_CHANGED = 2,       /**< control flow changed */
47
48
};

Michael Beck's avatar
Michael Beck committed
49
50
51
/**
 * walker environment
 */
52
typedef struct walk_env_t {
53
54
	struct obstack obst;          /**< list of all stores */
	unsigned changes;             /**< a bitmask of graph changes */
Michael Beck's avatar
Michael Beck committed
55
56
} walk_env_t;

57
/** A Load/Store info. */
58
typedef struct ldst_info_t {
59
	ir_node  *projs[MAX_PROJ+1];  /**< list of Proj's of this node */
60
61
62
	ir_node  *exc_block;          /**< the exception block if available */
	int      exc_idx;             /**< predecessor index in the exception block */
	unsigned visited;             /**< visited counter for breaking loops */
Michael Beck's avatar
Michael Beck committed
63
64
} ldst_info_t;

65
/**
66
 * flags for control flow.
67
68
 */
enum block_flags_t {
69
70
	BLOCK_HAS_COND = 1,      /**< Block has conditional control flow */
	BLOCK_HAS_EXC  = 2       /**< Block has exceptional control flow */
71
72
73
};

/**
74
 * a Block info.
75
 */
76
typedef struct block_info_t {
77
	unsigned flags;               /**< flags for the block */
78
79
} block_info_t;

80
81
82
83
84
85
86
/** the master visited flag for loop detection. */
static unsigned master_visited = 0;

#define INC_MASTER()       ++master_visited
#define MARK_NODE(info)    (info)->visited = master_visited
#define NODE_VISITED(info) (info)->visited >= master_visited

Michael Beck's avatar
Michael Beck committed
87
88
89
/**
 * get the Load/Store info of a node
 */
90
91
static ldst_info_t *get_ldst_info(ir_node *node, struct obstack *obst)
{
92
	ldst_info_t *info = (ldst_info_t*)get_irn_link(node);
Michael Beck's avatar
Michael Beck committed
93

94
	if (! info) {
95
		info = OALLOCZ(obst, ldst_info_t);
96
97
98
		set_irn_link(node, info);
	}
	return info;
99
}
Michael Beck's avatar
Michael Beck committed
100

101
102
103
/**
 * get the Block info of a node
 */
104
105
static block_info_t *get_block_info(ir_node *node, struct obstack *obst)
{
106
	block_info_t *info = (block_info_t*)get_irn_link(node);
107

108
	if (! info) {
109
		info = OALLOCZ(obst, block_info_t);
110
111
112
		set_irn_link(node, info);
	}
	return info;
113
}
114

Michael Beck's avatar
Michael Beck committed
115
/**
Michael Beck's avatar
Michael Beck committed
116
 * update the projection info for a Load/Store
Michael Beck's avatar
Michael Beck committed
117
 */
118
static unsigned update_projs(ldst_info_t *info, ir_node *proj)
Michael Beck's avatar
Michael Beck committed
119
{
120
	long nr = get_Proj_proj(proj);
Michael Beck's avatar
Michael Beck committed
121

122
	assert(0 <= nr && nr <= MAX_PROJ && "Wrong proj from LoadStore");
Michael Beck's avatar
Michael Beck committed
123

124
125
126
127
128
129
130
131
132
	if (info->projs[nr]) {
		/* there is already one, do CSE */
		exchange(proj, info->projs[nr]);
		return DF_CHANGED;
	}
	else {
		info->projs[nr] = proj;
		return 0;
	}
133
}
Michael Beck's avatar
Michael Beck committed
134
135

/**
136
137
138
139
140
 * update the exception block info for a Load/Store node.
 *
 * @param info   the load/store info struct
 * @param block  the exception handler block for this load/store
 * @param pos    the control flow input of the block
Michael Beck's avatar
Michael Beck committed
141
 */
142
static unsigned update_exc(ldst_info_t *info, ir_node *block, int pos)
Michael Beck's avatar
Michael Beck committed
143
{
144
	assert(info->exc_block == NULL && "more than one exception block found");
Michael Beck's avatar
Michael Beck committed
145

146
147
148
	info->exc_block = block;
	info->exc_idx   = pos;
	return 0;
149
}
Michael Beck's avatar
Michael Beck committed
150
151
152

/**
 * walker, collects all Load/Store/Proj nodes
153
 *
154
 * walks from Start -> End
Michael Beck's avatar
Michael Beck committed
155
 */
Michael Beck's avatar
Michael Beck committed
156
static void collect_nodes(ir_node *node, void *env)
Michael Beck's avatar
Michael Beck committed
157
{
158
159
	walk_env_t  *wenv   = (walk_env_t *)env;
	unsigned     opcode = get_irn_opcode(node);
160
161
162
	ir_node     *pred, *blk, *pred_blk;
	ldst_info_t *ldst_info;

163
164
165
	if (opcode == iro_Proj) {
		pred   = get_Proj_pred(node);
		opcode = get_irn_opcode(pred);
166

167
		if (opcode == iro_Load || opcode == iro_Store || opcode == iro_Call) {
168
			ldst_info = get_ldst_info(pred, &wenv->obst);
169
170
171
172

			wenv->changes |= update_projs(ldst_info, node);

			/*
173
174
175
176
177
			 * Place the Proj's to the same block as the
			 * predecessor Load. This is always ok and prevents
			 * "non-SSA" form after optimizations if the Proj
			 * is in a wrong block.
			 */
178
179
180
181
182
183
184
			blk      = get_nodes_block(node);
			pred_blk = get_nodes_block(pred);
			if (blk != pred_blk) {
				wenv->changes |= DF_CHANGED;
				set_nodes_block(node, pred_blk);
			}
		}
185
	} else if (opcode == iro_Block) {
186
187
188
		int i;

		for (i = get_Block_n_cfgpreds(node) - 1; i >= 0; --i) {
189
			ir_node      *pred_block, *proj;
190
			block_info_t *bl_info;
191
192
193
			int          is_exc = 0;

			pred = proj = get_Block_cfgpred(node, i);
194

195
196
			if (is_Proj(proj)) {
				pred   = get_Proj_pred(proj);
Matthias Braun's avatar
Matthias Braun committed
197
				is_exc = is_x_except_Proj(proj);
198
			}
199
200
201
202
203
204

			/* ignore Bad predecessors, they will be removed later */
			if (is_Bad(pred))
				continue;

			pred_block = get_nodes_block(pred);
205
			bl_info    = get_block_info(pred_block, &wenv->obst);
206

207
			if (is_fragile_op(pred) && is_exc)
208
209
210
211
				bl_info->flags |= BLOCK_HAS_EXC;
			else if (is_irn_forking(pred))
				bl_info->flags |= BLOCK_HAS_COND;

212
213
			opcode = get_irn_opcode(pred);
			if (is_exc && (opcode == iro_Load || opcode == iro_Store || opcode == iro_Call)) {
214
				ldst_info = get_ldst_info(pred, &wenv->obst);
215
216
217
218
219

				wenv->changes |= update_exc(ldst_info, node, i);
			}
		}
	}
220
}
Michael Beck's avatar
Michael Beck committed
221

Michael Beck's avatar
Michael Beck committed
222
/**
223
 * Returns an entity if the address ptr points to a constant one.
224
225
226
227
 *
 * @param ptr  the address
 *
 * @return an entity or NULL
Michael Beck's avatar
Michael Beck committed
228
 */
229
static ir_entity *find_constant_entity(ir_node *ptr)
Michael Beck's avatar
Michael Beck committed
230
{
231
	for (;;) {
232
		if (is_SymConst(ptr) && get_SymConst_kind(ptr) == symconst_addr_ent) {
233
			return get_SymConst_entity(ptr);
234
		} else if (is_Sel(ptr)) {
235
236
237
238
			ir_entity *ent = get_Sel_entity(ptr);
			ir_type   *tp  = get_entity_owner(ent);

			/* Do not fiddle with polymorphism. */
239
			if (is_Class_type(tp) &&
240
241
242
243
244
245
246
247
248
				((get_entity_n_overwrites(ent)    != 0) ||
				(get_entity_n_overwrittenby(ent) != 0)   ) )
				return NULL;

			if (is_Array_type(tp)) {
				/* check bounds */
				int i, n;

				for (i = 0, n = get_Sel_n_indexs(ptr); i < n; ++i) {
Matthias Braun's avatar
Matthias Braun committed
249
250
251
252
					ir_node   *bound;
					ir_tarval *tlower, *tupper;
					ir_node   *index = get_Sel_index(ptr, i);
					ir_tarval *tv    = computed_value(index);
253
254
255
256
257
258
259
260
261
262
263
264
265

					/* check if the index is constant */
					if (tv == tarval_bad)
						return NULL;

					bound  = get_array_lower_bound(tp, i);
					tlower = computed_value(bound);
					bound  = get_array_upper_bound(tp, i);
					tupper = computed_value(bound);

					if (tlower == tarval_bad || tupper == tarval_bad)
						return NULL;

266
					if (tarval_cmp(tv, tlower) == ir_relation_less)
267
						return NULL;
268
					if (tarval_cmp(tupper, tv) == ir_relation_less)
269
270
271
272
273
274
						return NULL;

					/* ok, bounds check finished */
				}
			}

Matthias Braun's avatar
Matthias Braun committed
275
			if (get_entity_linkage(ent) & IR_LINKAGE_CONSTANT)
276
277
278
279
				return ent;

			/* try next */
			ptr = get_Sel_ptr(ptr);
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
		} else if (is_Add(ptr)) {
			ir_node *l = get_Add_left(ptr);
			ir_node *r = get_Add_right(ptr);

			if (get_irn_mode(l) == get_irn_mode(ptr) && is_Const(r))
				ptr = l;
			else if (get_irn_mode(r) == get_irn_mode(ptr) && is_Const(l))
				ptr = r;
			else
				return NULL;

			/* for now, we support only one addition, reassoc should fold all others */
			if (! is_SymConst(ptr) && !is_Sel(ptr))
				return NULL;
		} else if (is_Sub(ptr)) {
			ir_node *l = get_Sub_left(ptr);
			ir_node *r = get_Sub_right(ptr);

298
			if (get_irn_mode(l) == get_irn_mode(ptr) && is_Const(r))
299
300
301
302
303
304
				ptr = l;
			else
				return NULL;
			/* for now, we support only one substraction, reassoc should fold all others */
			if (! is_SymConst(ptr) && !is_Sel(ptr))
				return NULL;
305
306
307
		} else
			return NULL;
	}
308
}
Michael Beck's avatar
Michael Beck committed
309

Michael Beck's avatar
Michael Beck committed
310
311
312
/**
 * Return the Selection index of a Sel node from dimension n
 */
313
314
static long get_Sel_array_index_long(ir_node *n, int dim)
{
315
316
	ir_node *index = get_Sel_index(n, dim);
	return get_tarval_long(get_Const_tarval(index));
317
}
318

319
320
321
typedef struct path_entry {
	ir_entity         *ent;
	struct path_entry *next;
322
	size_t            index;
323
324
} path_entry;

325
326
static ir_node *rec_find_compound_ent_value(ir_node *ptr, path_entry *next)
{
327
	path_entry       entry, *p;
328
	ir_entity        *ent, *field;
329
	ir_initializer_t *initializer;
Matthias Braun's avatar
Matthias Braun committed
330
	ir_tarval        *tv;
331
	ir_type          *tp;
332
	size_t           n;
333

334
	entry.next = next;
335
336
	if (is_SymConst(ptr)) {
		/* found the root */
337
		ent         = get_SymConst_entity(ptr);
338
		initializer = get_entity_initializer(ent);
339
		for (p = next; p != NULL;) {
340
341
			if (initializer->kind != IR_INITIALIZER_COMPOUND)
				return NULL;
342
343
			n  = get_initializer_compound_n_entries(initializer);
			tp = get_entity_type(ent);
344

345
346
347
348
349
350
351
352
353
354
			if (is_Array_type(tp)) {
				ent = get_array_element_entity(tp);
				if (ent != p->ent) {
					/* a missing [0] */
					if (0 >= n)
						return NULL;
					initializer = get_initializer_compound_value(initializer, 0);
					continue;
				}
			}
355
			if (p->index >= n)
356
357
				return NULL;
			initializer = get_initializer_compound_value(initializer, p->index);
358
359
360
361
362
363
364
365
366
367
368
369
370

			ent = p->ent;
			p   = p->next;
		}
		tp = get_entity_type(ent);
		while (is_Array_type(tp)) {
			ent = get_array_element_entity(tp);
			tp = get_entity_type(ent);
			/* a missing [0] */
			n  = get_initializer_compound_n_entries(initializer);
			if (0 >= n)
				return NULL;
			initializer = get_initializer_compound_value(initializer, 0);
371
		}
372

373
374
375
376
377
378
379
380
381
		switch (initializer->kind) {
		case IR_INITIALIZER_CONST:
			return get_initializer_const_value(initializer);
		case IR_INITIALIZER_TARVAL:
		case IR_INITIALIZER_NULL:
		default:
			return NULL;
		}
	} else if (is_Sel(ptr)) {
382
383
384
385
		entry.ent = field = get_Sel_entity(ptr);
		tp = get_entity_owner(field);
		if (is_Array_type(tp)) {
			assert(get_Sel_n_indexs(ptr) == 1 && "multi dim arrays not implemented");
386
			entry.index = get_Sel_array_index_long(ptr, 0) - get_array_lower_bound_int(tp, 0);
387
		} else {
388
			size_t i, n_members = get_compound_n_members(tp);
389
390
391
392
393
394
395
396
397
398
399
			for (i = 0; i < n_members; ++i) {
				if (get_compound_member(tp, i) == field)
					break;
			}
			if (i >= n_members) {
				/* not found: should NOT happen */
				return NULL;
			}
			entry.index = i;
		}
		return rec_find_compound_ent_value(get_Sel_ptr(ptr), &entry);
400
401
402
	}  else if (is_Add(ptr)) {
		ir_mode  *mode;
		unsigned pos;
403

404
405
406
407
408
409
410
411
412
413
		{
			ir_node *l = get_Add_left(ptr);
			ir_node *r = get_Add_right(ptr);
			if (is_Const(r)) {
				ptr = l;
				tv  = get_Const_tarval(r);
			} else {
				ptr = r;
				tv  = get_Const_tarval(l);
			}
414
415
416
		}
ptr_arith:
		mode = get_tarval_mode(tv);
417

418
419
420
421
422
423
		/* ptr must be a Sel or a SymConst, this was checked in find_constant_entity() */
		if (is_Sel(ptr)) {
			field = get_Sel_entity(ptr);
		} else {
			field = get_SymConst_entity(ptr);
		}
424

425
426
427
428
429
430
431
432
433
434
435
		/* count needed entries */
		pos = 0;
		for (ent = field;;) {
			tp = get_entity_type(ent);
			if (! is_Array_type(tp))
				break;
			ent = get_array_element_entity(tp);
			++pos;
		}
		/* should be at least ONE entry */
		if (pos == 0)
436
437
			return NULL;

438
439
440
441
442
443
		/* allocate the right number of entries */
		NEW_ARR_A(path_entry, p, pos);

		/* fill them up */
		pos = 0;
		for (ent = field;;) {
Matthias Braun's avatar
Matthias Braun committed
444
445
446
447
			unsigned   size;
			ir_tarval *sz, *tv_index, *tlower, *tupper;
			long       index;
			ir_node   *bound;
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473

			tp = get_entity_type(ent);
			if (! is_Array_type(tp))
				break;
			ent = get_array_element_entity(tp);
			p[pos].ent  = ent;
			p[pos].next = &p[pos + 1];

			size = get_type_size_bytes(get_entity_type(ent));
			sz   = new_tarval_from_long(size, mode);

			tv_index = tarval_div(tv, sz);
			tv       = tarval_mod(tv, sz);

			if (tv_index == tarval_bad || tv == tarval_bad)
				return NULL;

			assert(get_array_n_dimensions(tp) == 1 && "multiarrays not implemented");
			bound  = get_array_lower_bound(tp, 0);
			tlower = computed_value(bound);
			bound  = get_array_upper_bound(tp, 0);
			tupper = computed_value(bound);

			if (tlower == tarval_bad || tupper == tarval_bad)
				return NULL;

474
			if (tarval_cmp(tv_index, tlower) == ir_relation_less)
475
				return NULL;
476
			if (tarval_cmp(tupper, tv_index) == ir_relation_less)
477
478
479
480
481
482
483
484
485
				return NULL;

			/* ok, bounds check finished */
			index = get_tarval_long(tv_index);
			p[pos].index = index;
			++pos;
		}
		if (! tarval_is_null(tv)) {
			/* hmm, wrong access */
486
			return NULL;
487
488
489
490
491
492
		}
		p[pos - 1].next = next;
		return rec_find_compound_ent_value(ptr, p);
	} else if (is_Sub(ptr)) {
		ir_node *l = get_Sub_left(ptr);
		ir_node *r = get_Sub_right(ptr);
493

494
495
496
497
		ptr = l;
		tv  = get_Const_tarval(r);
		tv  = tarval_neg(tv);
		goto ptr_arith;
498
	}
499
	return NULL;
500
501
}

502
503
static ir_node *find_compound_ent_value(ir_node *ptr)
{
504
505
506
	return rec_find_compound_ent_value(ptr, NULL);
}

507
508
509
510
/* forward */
static void reduce_adr_usage(ir_node *ptr);

/**
Christoph Mallon's avatar
Christoph Mallon committed
511
 * Update a Load that may have lost its users.
512
 */
513
514
static void handle_load_update(ir_node *load)
{
515
	ldst_info_t *info = (ldst_info_t*)get_irn_link(load);
516

517
518
519
	/* do NOT touch volatile loads for now */
	if (get_Load_volatility(load) == volatility_is_volatile)
		return;
520

521
522
523
	if (! info->projs[pn_Load_res] && ! info->projs[pn_Load_X_except]) {
		ir_node *ptr = get_Load_ptr(load);
		ir_node *mem = get_Load_mem(load);
524

Christoph Mallon's avatar
Christoph Mallon committed
525
		/* a Load whose value is neither used nor exception checked, remove it */
Andreas Zwinkau's avatar
Andreas Zwinkau committed
526
		exchange(info->projs[pn_Load_M], mem);
527
		if (info->projs[pn_Load_X_regular])
528
			exchange(info->projs[pn_Load_X_regular], new_r_Jmp(get_nodes_block(load)));
529
		kill_node(load);
530
531
		reduce_adr_usage(ptr);
	}
532
}
533
534

/**
Christoph Mallon's avatar
Christoph Mallon committed
535
 * A use of an address node has vanished. Check if this was a Proj
536
537
 * node and update the counters.
 */
538
539
static void reduce_adr_usage(ir_node *ptr)
{
540
541
542
543
544
	ir_node *pred;
	if (!is_Proj(ptr))
		return;
	if (get_irn_n_edges(ptr) > 0)
		return;
545

546
547
548
	/* this Proj is dead now */
	pred = get_Proj_pred(ptr);
	if (is_Load(pred)) {
549
		ldst_info_t *info = (ldst_info_t*)get_irn_link(pred);
550
		info->projs[get_Proj_proj(ptr)] = NULL;
551

552
553
		/* this node lost its result proj, handle that */
		handle_load_update(pred);
554
	}
555
}
556

557
558
559
560
/**
 * Check, if an already existing value of mode old_mode can be converted
 * into the needed one new_mode without loss.
 */
561
562
static int can_use_stored_value(ir_mode *old_mode, ir_mode *new_mode)
{
563
564
	unsigned old_size;
	unsigned new_size;
565
	if (old_mode == new_mode)
566
567
568
569
		return true;

	old_size = get_mode_size_bits(old_mode);
	new_size = get_mode_size_bits(new_mode);
570
571

	/* if both modes are two-complement ones, we can always convert the
572
573
574
	   Stored value into the needed one. (on big endian machines we currently
	   only support this for modes of same size) */
	if (old_size >= new_size &&
575
		  get_mode_arithmetic(old_mode) == irma_twos_complement &&
576
577
578
579
580
581
582
		  get_mode_arithmetic(new_mode) == irma_twos_complement &&
		  (!be_get_backend_param()->byte_order_big_endian
	        || old_size == new_size)) {
		return true;
	}
	return false;
}
583

Michael Beck's avatar
Michael Beck committed
584
/**
585
 * Check whether a Call is at least pure, i.e. does only read memory.
586
 */
587
588
static unsigned is_Call_pure(ir_node *call)
{
589
590
591
592
593
594
595
596
	ir_type *call_tp = get_Call_type(call);
	unsigned prop = get_method_additional_properties(call_tp);

	/* check first the call type */
	if ((prop & (mtp_property_const|mtp_property_pure)) == 0) {
		/* try the called entity */
		ir_node *ptr = get_Call_ptr(call);

597
598
		if (is_SymConst_addr_ent(ptr)) {
			ir_entity *ent = get_SymConst_entity(ptr);
599
600
601
602
603

			prop = get_entity_additional_properties(ent);
		}
	}
	return (prop & (mtp_property_const|mtp_property_pure)) != 0;
604
}
605

Michael Beck's avatar
Michael Beck committed
606
static ir_node *get_base_and_offset(ir_node *ptr, long *pOffset)
607
{
Michael Beck's avatar
Michael Beck committed
608
609
	ir_mode *mode  = get_irn_mode(ptr);
	long    offset = 0;
610
611

	/* TODO: long might not be enough, we should probably use some tarval thingy... */
Michael Beck's avatar
Michael Beck committed
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
	for (;;) {
		if (is_Add(ptr)) {
			ir_node *l = get_Add_left(ptr);
			ir_node *r = get_Add_right(ptr);

			if (get_irn_mode(l) != mode || !is_Const(r))
				break;

			offset += get_tarval_long(get_Const_tarval(r));
			ptr     = l;
		} else if (is_Sub(ptr)) {
			ir_node *l = get_Sub_left(ptr);
			ir_node *r = get_Sub_right(ptr);

			if (get_irn_mode(l) != mode || !is_Const(r))
				break;

			offset -= get_tarval_long(get_Const_tarval(r));
			ptr     = l;
		} else if (is_Sel(ptr)) {
			ir_entity *ent = get_Sel_entity(ptr);
			ir_type   *tp  = get_entity_owner(ent);

			if (is_Array_type(tp)) {
				int     size;
				ir_node *index;

				/* only one dimensional arrays yet */
				if (get_Sel_n_indexs(ptr) != 1)
					break;
				index = get_Sel_index(ptr, 0);
				if (! is_Const(index))
					break;

				tp = get_entity_type(ent);
				if (get_type_state(tp) != layout_fixed)
					break;

				size    = get_type_size_bytes(tp);
				offset += size * get_tarval_long(get_Const_tarval(index));
			} else {
				if (get_type_state(tp) != layout_fixed)
					break;
				offset += get_entity_offset(ent);
			}
			ptr = get_Sel_ptr(ptr);
		} else
659
660
661
			break;
	}

Michael Beck's avatar
Michael Beck committed
662
663
	*pOffset = offset;
	return ptr;
664
665
}

Michael Beck's avatar
Michael Beck committed
666
static int try_load_after_store(ir_node *load,
667
668
669
670
		ir_node *load_base_ptr, long load_offset, ir_node *store)
{
	ldst_info_t *info;
	ir_node *store_ptr      = get_Store_ptr(store);
Michael Beck's avatar
Michael Beck committed
671
672
	long     store_offset;
	ir_node *store_base_ptr = get_base_and_offset(store_ptr, &store_offset);
673
674
675
676
677
678
679
680
681
682
683
684
	ir_node *store_value;
	ir_mode *store_mode;
	ir_node *load_ptr;
	ir_mode *load_mode;
	long     load_mode_len;
	long     store_mode_len;
	long     delta;
	int      res;

	if (load_base_ptr != store_base_ptr)
		return 0;

Michael Beck's avatar
Michael Beck committed
685
686
	load_mode      = get_Load_mode(load);
	load_mode_len  = get_mode_size_bytes(load_mode);
687
688
	store_mode     = get_irn_mode(get_Store_value(store));
	store_mode_len = get_mode_size_bytes(store_mode);
Michael Beck's avatar
Michael Beck committed
689
	delta          = load_offset - store_offset;
690
	store_value    = get_Store_value(store);
691

692
693
	if (delta < 0 || delta+load_mode_len > store_mode_len)
		return 0;
694

695
696
697
	if (store_mode != load_mode &&
	    get_mode_arithmetic(store_mode) == irma_twos_complement &&
	    get_mode_arithmetic(load_mode)  == irma_twos_complement) {
Michael Beck's avatar
Michael Beck committed
698

699
		/* produce a shift to adjust offset delta */
700
701
702
703
704
705
		unsigned const shift = be_get_backend_param()->byte_order_big_endian
			? store_mode_len - load_mode_len - delta
			: delta;
		if (shift != 0) {
			ir_graph *const irg  = get_irn_irg(load);
			ir_node  *const cnst = new_r_Const_long(irg, mode_Iu, shift * 8);
706
			store_value = new_r_Shr(get_nodes_block(load),
707
708
709
									store_value, cnst, store_mode);
		}

710
711
712
713
		store_value = new_r_Conv(get_nodes_block(load), store_value, load_mode);
	} else {
		/* we would need some kind of bitcast node here */
		return 0;
714
715
	}

Michael Beck's avatar
Michael Beck committed
716
717
	DBG_OPT_RAW(load, store_value);

718
	info = (ldst_info_t*)get_irn_link(load);
719
720
721
722
723
724
	if (info->projs[pn_Load_M])
		exchange(info->projs[pn_Load_M], get_Load_mem(load));

	res = 0;
	/* no exception */
	if (info->projs[pn_Load_X_except]) {
725
		ir_graph *irg = get_irn_irg(load);
Matthias Braun's avatar
Matthias Braun committed
726
		exchange( info->projs[pn_Load_X_except], new_r_Bad(irg, mode_X));
727
728
729
		res |= CF_CHANGED;
	}
	if (info->projs[pn_Load_X_regular]) {
730
		exchange( info->projs[pn_Load_X_regular], new_r_Jmp(get_nodes_block(load)));
731
732
733
734
735
736
737
738
739
740
741
742
		res |= CF_CHANGED;
	}

	if (info->projs[pn_Load_res])
		exchange(info->projs[pn_Load_res], store_value);

	load_ptr = get_Load_ptr(load);
	kill_node(load);
	reduce_adr_usage(load_ptr);
	return res | DF_CHANGED;
}

743
744
745
746
/**
 * Follow the memory chain as long as there are only Loads,
 * alias free Stores, and constant Calls and try to replace the
 * current Load by a previous ones.
Michael Beck's avatar
Michael Beck committed
747
748
749
750
751
752
 * Note that in unreachable loops it might happen that we reach
 * load again, as well as we can fall into a cycle.
 * We break such cycles using a special visited flag.
 *
 * INC_MASTER() must be called before dive into
 */
753
754
static unsigned follow_Mem_chain(ir_node *load, ir_node *curr)
{
Michael Beck's avatar
Michael Beck committed
755
	unsigned    res = 0;
756
	ldst_info_t *info = (ldst_info_t*)get_irn_link(load);
Michael Beck's avatar
Michael Beck committed
757
758
759
760
	ir_node     *pred;
	ir_node     *ptr       = get_Load_ptr(load);
	ir_node     *mem       = get_Load_mem(load);
	ir_mode     *load_mode = get_Load_mode(load);
761
762

	for (pred = curr; load != pred; ) {
763
		ldst_info_t *pred_info = (ldst_info_t*)get_irn_link(pred);
764
765

		/*
766
767
		 * a Load immediately after a Store -- a read after write.
		 * We may remove the Load, if both Load & Store does not have an
Matthias Braun's avatar
Matthias Braun committed
768
		 * exception handler OR they are in the same Block. In the latter
769
770
771
772
773
		 * case the Load cannot throw an exception when the previous Store was
		 * quiet.
		 *
		 * Why we need to check for Store Exception? If the Store cannot
		 * be executed (ROM) the exception handler might simply jump into
Matthias Braun's avatar
Matthias Braun committed
774
		 * the load Block :-(
775
776
		 * We could make it a little bit better if we would know that the
		 * exception handler of the Store jumps directly to the end...
777
		 */
778
779
		if (is_Store(pred) && ((pred_info->projs[pn_Store_X_except] == NULL
				&& info->projs[pn_Load_X_except] == NULL)
Matthias Braun's avatar
Matthias Braun committed
780
				|| get_nodes_block(load) == get_nodes_block(pred)))
781
		{
Michael Beck's avatar
Michael Beck committed
782
783
784
785
786
			long    load_offset;
			ir_node *base_ptr = get_base_and_offset(ptr, &load_offset);
			int     changes   = try_load_after_store(load, base_ptr, load_offset, pred);

			if (changes != 0)
787
				return res | changes;
788
		} else if (is_Load(pred) && get_Load_ptr(pred) == ptr &&
789
		           can_use_stored_value(get_Load_mode(pred), load_mode)) {
790
791
			/*
			 * a Load after a Load -- a read after read.
Matthias Braun's avatar
Matthias Braun committed
792
793
794
795
			 * We may remove the second Load, if it does not have an exception
			 * handler OR they are in the same Block. In the later case
			 * the Load cannot throw an exception when the previous Load was
			 * quiet.
796
			 *
Matthias Braun's avatar
Matthias Braun committed
797
798
799
			 * Here, there is no need to check if the previous Load has an
			 * exception hander because they would have exact the same
			 * exception...
800
801
802
			 *
			 * TODO: implement load-after-load with different mode for big
			 *       endian
803
			 */
Matthias Braun's avatar
Matthias Braun committed
804
805
			if (info->projs[pn_Load_X_except] == NULL
					|| get_nodes_block(load) == get_nodes_block(pred)) {
806
807
				ir_node *value;

808
809
				DBG_OPT_RAR(load, pred);

810
811
812
813
				/* the result is used */
				if (info->projs[pn_Load_res]) {
					if (pred_info->projs[pn_Load_res] == NULL) {
						/* create a new Proj again */
814
						pred_info->projs[pn_Load_res] = new_r_Proj(pred, get_Load_mode(pred), pn_Load_res);
815
816
					}
					value = pred_info->projs[pn_Load_res];
817
818
819

					/* add an convert if needed */
					if (get_Load_mode(pred) != load_mode) {
820
						value = new_r_Conv(get_nodes_block(load), value, load_mode);
821
822
					}

823
					exchange(info->projs[pn_Load_res], value);
824
825
				}

826
827
828
				if (info->projs[pn_Load_M])
					exchange(info->projs[pn_Load_M], mem);

829
830
				/* no exception */
				if (info->projs[pn_Load_X_except]) {
831
					ir_graph *irg = get_irn_irg(load);
Matthias Braun's avatar
Matthias Braun committed
832
					exchange(info->projs[pn_Load_X_except], new_r_Bad(irg, mode_X));
833
834
					res |= CF_CHANGED;
				}
835
				if (info->projs[pn_Load_X_regular]) {
836
					exchange( info->projs[pn_Load_X_regular], new_r_Jmp(get_nodes_block(load)));
837
838
					res |= CF_CHANGED;
				}
839

840
				kill_node(load);
841
842
843
844
845
				reduce_adr_usage(ptr);
				return res |= DF_CHANGED;
			}
		}

846
		if (is_Store(pred)) {
847
			/* check if we can pass through this store */
848
849
850
			ir_alias_relation rel = get_alias_relation(
				get_Store_ptr(pred),
				get_irn_mode(get_Store_value(pred)),
851
				ptr, load_mode);
852
			/* if the might be an alias, we cannot pass this Store */
Michael Beck's avatar
Michael Beck committed
853
			if (rel != ir_no_alias)
854
855
				break;
			pred = skip_Proj(get_Store_mem(pred));
856
		} else if (is_Load(pred)) {
857
			pred = skip_Proj(get_Load_mem(pred));
858
859
860
861
862
863
864
865
866
		} else if (is_Call(pred)) {
			if (is_Call_pure(pred)) {
				/* The called graph is at least pure, so there are no Store's
				   in it. We can handle it like a Load and skip it. */
				pred = skip_Proj(get_Call_mem(pred));
			} else {
				/* there might be Store's in the graph, stop here */
				break;
			}
867
868
869
870
871
872
873
874
875
876
877
		} else {
			/* follow only Load chains */
			break;
		}

		/* check for cycles */
		if (NODE_VISITED(pred_info))
			break;
		MARK_NODE(pred_info);
	}

878
	if (is_Sync(pred)) {
879
880
881
882
883
884
		int i;

		/* handle all Sync predecessors */
		for (i = get_Sync_n_preds(pred) - 1; i >= 0; --i) {
			res |= follow_Mem_chain(load, skip_Proj(get_Sync_pred(pred, i)));
			if (res)
885
				return res;
886
887
888
889
		}
	}

	return res;
890
}
Michael Beck's avatar
Michael Beck committed
891

892
893
ir_node *can_replace_load_by_const(const ir_node *load, ir_node *c)
{
894
895
896
897
898
	ir_mode  *c_mode = get_irn_mode(c);
	ir_mode  *l_mode = get_Load_mode(load);
	ir_node  *block  = get_nodes_block(load);
	dbg_info *dbgi   = get_irn_dbg_info(load);
	ir_node  *res    = copy_const_value(dbgi, c, block);
899
900
901
902
903

	if (c_mode != l_mode) {
		/* check, if the mode matches OR can be easily converted info */
		if (is_reinterpret_cast(c_mode, l_mode)) {
			/* copy the value from the const code irg and cast it */
904
			res = new_rd_Conv(dbgi, block, res, l_mode);
Andreas Zwinkau's avatar
Andreas Zwinkau committed
905
906
		} else {
			return NULL;
907
908
909
		}
	}
	return res;
910
}
911

Michael Beck's avatar
Michael Beck committed
912
913
/**
 * optimize a Load
914
915
 *
 * @param load  the Load node
Michael Beck's avatar
Michael Beck committed
916
 */
917
static unsigned optimize_load(ir_node *load)
Michael Beck's avatar
Michael Beck committed
918
{
919
	ldst_info_t *info = (ldst_info_t*)get_irn_link(load);
Michael Beck's avatar
Michael Beck committed
920
921
922
923
	ir_node     *mem, *ptr, *value;
	ir_entity   *ent;
	long        dummy;
	unsigned    res = 0;
924
925
926
927
928
929
930
931
932

	/* do NOT touch volatile loads for now */
	if (get_Load_volatility(load) == volatility_is_volatile)
		return 0;

	/* the address of the load to be optimized */
	ptr = get_Load_ptr(load);

	/* The mem of the Load. Must still be returned after optimization. */
Christoph Mallon's avatar
Christoph Mallon committed
933
	mem = get_Load_mem(load);
934

935
936
937
	if (info->projs[pn_Load_res] == NULL
			&& info->projs[pn_Load_X_except] == NULL) {
		/* the value is never used and we don't care about exceptions, remove */
938
939
		exchange(info->projs[pn_Load_M], mem);

940
941
		if (info->projs[pn_Load_X_regular]) {
			/* should not happen, but if it does, remove it */
942
			exchange(info->projs[pn_Load_X_regular], new_r_Jmp(get_nodes_block(load)));
943
944
			res |= CF_CHANGED;
		}
945
		kill_node(load);
946
947
948
949
		reduce_adr_usage(ptr);
		return res | DF_CHANGED;
	}

Matthias Braun's avatar
Matthias Braun committed
950
951
952
953
954
955
956
957
	value = NULL;
	/* check if we can determine the entity that will be loaded */
	ent = find_constant_entity(ptr);
	if (ent != NULL
			&& get_entity_visibility(ent) != ir_visibility_external) {
		/* a static allocation that is not external: there should be NO
		 * exception when loading even if we cannot replace the load itself.
		 */
Christoph Mallon's avatar
Christoph Mallon committed
958

Matthias Braun's avatar
Matthias Braun committed
959
960
961
962
963
964
965
966
967
968
969
970
		/* no exception, clear the info field as it might be checked later again */
		if (info->projs[pn_Load_X_except]) {
			ir_graph *irg = get_irn_irg(load);
			exchange(info->projs[pn_Load_X_except], new_r_Bad(irg, mode_X));
			info->projs[pn_Load_X_except] = NULL;
			res |= CF_CHANGED;
		}
		if (info->projs[pn_Load_X_regular]) {
			exchange(info->projs[pn_Load_X_regular], new_r_Jmp(get_nodes_block(load)));
			info->projs[pn_Load_X_regular] = NULL;
			res |= CF_CHANGED;
		}
Michael Beck's avatar
Michael Beck committed
971

Matthias Braun's avatar
Matthias Braun committed
972
973
974
975
976
977
978
979
		if (get_entity_linkage(ent) & IR_LINKAGE_CONSTANT) {
			if (has_entity_initializer(ent)) {
				/* new style initializer */
				value = find_compound_ent_value(ptr);
			}
			if (value != NULL) {
				ir_graph *irg = get_irn_irg(load);
				value = can_replace_load_by_const(load, value);
980
				if (value != NULL && is_Sel(ptr)) {
Matthias Braun's avatar
Matthias Braun committed
981
982
983
					/* frontend has inserted masking operations after bitfield accesses,
					 * so we might have to shift the const. */
					unsigned char bit_offset = get_entity_offset_bits_remainder(get_Sel_entity(ptr));
984
985
986
987
988
989
990
991
992
993
					if (bit_offset != 0) {
						if (is_Const(value)) {
							ir_tarval *tv_old = get_Const_tarval(value);
							ir_tarval *tv_offset = new_tarval_from_long(bit_offset, mode_Bu);
							ir_tarval *tv_new = tarval_shl(tv_old, tv_offset);
							value = new_r_Const(irg, tv_new);
						} else {
							value = NULL;
						}
					}
Andreas Zwinkau's avatar
Andreas Zwinkau committed
994
				}
Michael Beck's avatar
Michael Beck committed
995
			}
996
		}
Michael Beck's avatar
Michael Beck committed
997
998
999
	}
	if (value != NULL) {
		/* we completely replace the load by this value */
1000
		if (info->projs[pn_Load_X_except]) {
1001
			ir_graph *irg = get_irn_irg(load);
Matthias Braun's avatar
Matthias Braun committed
1002
			exchange(info->projs[pn_Load_X_except], new_r_Bad(irg, mode_X));
1003
			info->projs[pn_Load_X_except] = NULL;
1004
1005
1006
			res |= CF_CHANGED;
		}
		if (info->projs[pn_Load_X_regular]) {
1007
			exchange(info->projs[pn_Load_X_regular], new_r_Jmp(get_nodes_block(load)));
1008
1009
			info->projs[pn_Load_X_regular] = NULL;
			res |= CF_CHANGED;
1010
		}
Michael Beck's avatar
Michael Beck committed
1011
1012
1013
1014
1015
1016
1017
1018
		if (info->projs[pn_Load_M]) {
			exchange(info->projs[pn_Load_M], mem);
			res |= DF_CHANGED;
		}
		if (info->projs[pn_Load_res]) {
			exchange(info->projs[pn_Load_res], value);
			res |= DF_CHANGED;
		}
1019
		kill_node(load);
1020
		reduce_adr_usage(ptr);
Michael Beck's avatar
Michael Beck committed
1021
		return res;
1022
1023
1024
	}

	/* Check, if the address of this load is used more than once.
Michael Beck's avatar
Michael Beck committed
1025
	 * If not, more load cannot be removed in any case. */
1026
	if (get_irn_n_edges(ptr) <= 1 && get_irn_n_edges(get_base_and_offset(ptr, &dummy)) <= 1)
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
		return res;

	/*
	 * follow the memory chain as long as there are only Loads
	 * and try to replace current Load or Store by a previous one.
	 * Note that in unreachable loops it might happen that we reach
	 * load again, as well as we can fall into a cycle.
	 * We break such cycles using a special visited flag.
	 */
	INC_MASTER();
	res = follow_Mem_chain(load, skip_Proj(mem));
	return res;
1039
}
Michael Beck's avatar
Michael Beck committed
1040

1041
1042
1043
1044
1045
1046
1047
/**
 * Check whether a value of mode new_mode would completely overwrite a value
 * of mode old_mode in memory.
 */
static int is_completely_overwritten(ir_mode *old_mode, ir_mode *new_mode)
{
	return get_mode_size_bits(new_mode) >= get_mode_size_bits(old_mode);
1048
}
1049