1 /*
2  * Copyright (C) 1995-2011 University of Karlsruhe.  All right reserved.
3  *
4  * This file is part of libFirm.
5  *
6  * This file may be distributed and/or modified under the terms of the
7  * GNU General Public License version 2 as published by the Free Software
8  * Foundation and appearing in the file LICENSE.GPL included in the
9  * packaging of this file.
10  *
11  * Licensees holding valid libFirm Professional Edition licenses may use
12  * this file in accordance with the libFirm Commercial License.
13  * Agreement provided with the Software.
14  *
15  * This file is provided AS IS with NO WARRANTY OF ANY KIND, INCLUDING THE
16  * WARRANTY OF DESIGN, MERCHANTABILITY AND FITNESS FOR A PARTICULAR
17  * PURPOSE.
18  */
19 
20 /**
21  * @file
22  * @brief   Lower double word operations, i.e. 64bit -> 32bit, 32bit -> 16bit etc.
23  * @date    8.10.2004
24  * @author  Michael Beck
25  */
26 #include "config.h"
27 
28 #include <string.h>
29 #include <stdlib.h>
30 #include <stdbool.h>
31 #include <assert.h>
32 
33 #include "be.h"
34 #include "error.h"
35 #include "lowering.h"
36 #include "irnode_t.h"
37 #include "irnodeset.h"
38 #include "irgraph_t.h"
39 #include "irmode_t.h"
40 #include "iropt_t.h"
41 #include "irgmod.h"
42 #include "tv_t.h"
43 #include "dbginfo_t.h"
44 #include "iropt_dbg.h"
45 #include "irflag_t.h"
46 #include "firmstat.h"
47 #include "irgwalk.h"
48 #include "ircons.h"
49 #include "irflag.h"
50 #include "iroptimize.h"
51 #include "irtools.h"
52 #include "debug.h"
53 #include "set.h"
54 #include "pmap.h"
55 #include "pdeq.h"
56 #include "irdump.h"
57 #include "array_t.h"
58 #include "irpass_t.h"
59 #include "lower_dw.h"
60 
61 /** A map from (op, imode, omode) to Intrinsic functions entities. */
62 static set *intrinsic_fkt;
63 
64 /** A map from (imode, omode) to conv function types. */
65 static set *conv_types;
66 
67 /** A map from a method type to its lowered type. */
68 static pmap *lowered_type;
69 
70 /** A map from a builtin type to its lower and higher type. */
71 static pmap *lowered_builtin_type_high;
72 static pmap *lowered_builtin_type_low;
73 
74 /** The types for the binop and unop intrinsics. */
75 static ir_type *binop_tp_u, *binop_tp_s, *unop_tp_u, *unop_tp_s, *tp_s, *tp_u;
76 
77 static ir_nodeset_t created_mux_nodes;
78 
79 /** the debug handle */
80 DEBUG_ONLY(static firm_dbg_module_t *dbg = NULL;)
81 
82 /**
83  * An entry in the (op, imode, omode) -> entity map.
84  */
85 typedef struct op_mode_entry {
86 	const ir_op   *op;    /**< the op */
87 	const ir_mode *imode; /**< the input mode */
88 	const ir_mode *omode; /**< the output mode */
89 	ir_entity     *ent;   /**< the associated entity of this (op, imode, omode) triple */
90 } op_mode_entry_t;
91 
92 /**
93  * An entry in the (imode, omode) -> tp map.
94  */
95 typedef struct conv_tp_entry {
96 	const ir_mode *imode; /**< the input mode */
97 	const ir_mode *omode; /**< the output mode */
98 	ir_type       *mtd;   /**< the associated method type of this (imode, omode) pair */
99 } conv_tp_entry_t;
100 
101 enum lower_flags {
102 	MUST_BE_LOWERED = 1,  /**< graph must be lowered */
103 	CF_CHANGED      = 2,  /**< control flow was changed */
104 };
105 
106 /**
107  * The lower environment.
108  */
109 typedef struct lower_dw_env_t {
110 	lower64_entry_t **entries;     /**< entries per node */
111 	ir_graph      *irg;
112 	struct obstack obst;           /**< an obstack holding the temporary data */
113 	ir_tarval *tv_mode_bytes;      /**< a tarval containing the number of bytes in the lowered modes */
114 	pdeq      *waitq;              /**< a wait queue of all nodes that must be handled later */
115 	ir_node  **lowered_phis;       /**< list of lowered phis */
116 	ir_mode   *high_signed;        /**< doubleword signed type */
117 	ir_mode   *high_unsigned;      /**< doubleword unsigned type */
118 	ir_mode   *low_signed;         /**< word signed type */
119 	ir_mode   *low_unsigned;       /**< word unsigned type */
120 	ident     *first_id;           /**< .l for little and .h for big endian */
121 	ident     *next_id;            /**< .h for little and .l for big endian */
122 	const lwrdw_param_t *params;   /**< transformation parameter */
123 	unsigned flags;                /**< some flags */
124 	unsigned n_entries;            /**< number of entries */
125 } lower_dw_env_t;
126 
127 static lower_dw_env_t *env;
128 
129 static void lower_node(ir_node *node);
130 
131 /**
132  * Create a method type for a Conv emulation from imode to omode.
133  */
get_conv_type(ir_mode * imode,ir_mode * omode)134 static ir_type *get_conv_type(ir_mode *imode, ir_mode *omode)
135 {
136 	conv_tp_entry_t key, *entry;
137 	ir_type *mtd;
138 
139 	key.imode = imode;
140 	key.omode = omode;
141 	key.mtd   = NULL;
142 
143 	entry = set_insert(conv_tp_entry_t, conv_types, &key, sizeof(key), hash_ptr(imode) ^ hash_ptr(omode));
144 	if (! entry->mtd) {
145 		int n_param = 1, n_res = 1;
146 
147 		if (imode == env->high_signed || imode == env->high_unsigned)
148 			n_param = 2;
149 		if (omode == env->high_signed || omode == env->high_unsigned)
150 			n_res = 2;
151 
152 		/* create a new one */
153 		mtd = new_type_method(n_param, n_res);
154 
155 		/* set param types and result types */
156 		n_param = 0;
157 		if (imode == env->high_signed) {
158 			if (env->params->little_endian) {
159 				set_method_param_type(mtd, n_param++, tp_u);
160 				set_method_param_type(mtd, n_param++, tp_s);
161 			} else {
162 				set_method_param_type(mtd, n_param++, tp_s);
163 				set_method_param_type(mtd, n_param++, tp_u);
164 			}
165 		} else if (imode == env->high_unsigned) {
166 			set_method_param_type(mtd, n_param++, tp_u);
167 			set_method_param_type(mtd, n_param++, tp_u);
168 		} else {
169 			ir_type *tp = get_type_for_mode(imode);
170 			set_method_param_type(mtd, n_param++, tp);
171 		}
172 
173 		n_res = 0;
174 		if (omode == env->high_signed) {
175 			if (env->params->little_endian) {
176 				set_method_res_type(mtd, n_res++, tp_u);
177 				set_method_res_type(mtd, n_res++, tp_s);
178 			} else {
179 				set_method_res_type(mtd, n_res++, tp_s);
180 				set_method_res_type(mtd, n_res++, tp_u);
181 			}
182 		} else if (omode == env->high_unsigned) {
183 			set_method_res_type(mtd, n_res++, tp_u);
184 			set_method_res_type(mtd, n_res++, tp_u);
185 		} else {
186 			ir_type *tp = get_type_for_mode(omode);
187 			set_method_res_type(mtd, n_res++, tp);
188 		}
189 		entry->mtd = mtd;
190 	} else {
191 		mtd = entry->mtd;
192 	}
193 	return mtd;
194 }
195 
196 /**
197  * Add an additional control flow input to a block.
198  * Patch all Phi nodes. The new Phi inputs are copied from
199  * old input number nr.
200  */
add_block_cf_input_nr(ir_node * block,int nr,ir_node * cf)201 static void add_block_cf_input_nr(ir_node *block, int nr, ir_node *cf)
202 {
203 	int i, arity = get_irn_arity(block);
204 	ir_node **in;
205 
206 	assert(nr < arity);
207 
208 	NEW_ARR_A(ir_node *, in, arity + 1);
209 	for (i = 0; i < arity; ++i)
210 		in[i] = get_irn_n(block, i);
211 	in[i] = cf;
212 
213 	set_irn_in(block, i + 1, in);
214 
215 	foreach_out_edge(block, edge) {
216 		ir_node *phi = get_edge_src_irn(edge);
217 		if (!is_Phi(phi))
218 			continue;
219 
220 		for (i = 0; i < arity; ++i)
221 			in[i] = get_irn_n(phi, i);
222 		in[i] = in[nr];
223 		set_irn_in(phi, i + 1, in);
224 	}
225 }
226 
227 /**
228  * Add an additional control flow input to a block.
229  * Patch all Phi nodes. The new Phi inputs are copied from
230  * old input from cf tmpl.
231  */
add_block_cf_input(ir_node * block,ir_node * tmpl,ir_node * cf)232 static void add_block_cf_input(ir_node *block, ir_node *tmpl, ir_node *cf)
233 {
234 	int i, arity = get_irn_arity(block);
235 	int nr = 0;
236 
237 	for (i = 0; i < arity; ++i) {
238 		if (get_irn_n(block, i) == tmpl) {
239 			nr = i;
240 			break;
241 		}
242 	}
243 	assert(i < arity);
244 	add_block_cf_input_nr(block, nr, cf);
245 }
246 
247 /**
248  * Return the "operational" mode of a Firm node.
249  */
get_irn_op_mode(ir_node * node)250 static ir_mode *get_irn_op_mode(ir_node *node)
251 {
252 	switch (get_irn_opcode(node)) {
253 	case iro_Load:
254 		return get_Load_mode(node);
255 	case iro_Store:
256 		return get_irn_mode(get_Store_value(node));
257 	case iro_Div:
258 		return get_irn_mode(get_Div_left(node));
259 	case iro_Mod:
260 		return get_irn_mode(get_Mod_left(node));
261 	case iro_Cmp:
262 		return get_irn_mode(get_Cmp_left(node));
263 	default:
264 		return get_irn_mode(node);
265 	}
266 }
267 
268 /**
269  * Walker, prepare the node links and determine which nodes need to be lowered
270  * at all.
271  */
prepare_links(ir_node * node)272 static void prepare_links(ir_node *node)
273 {
274 	ir_mode         *mode = get_irn_op_mode(node);
275 	lower64_entry_t *link;
276 
277 	if (mode == env->high_signed || mode == env->high_unsigned) {
278 		unsigned idx = get_irn_idx(node);
279 		/* ok, found a node that will be lowered */
280 		link = OALLOCZ(&env->obst, lower64_entry_t);
281 
282 		if (idx >= env->n_entries) {
283 			/* enlarge: this happens only for Rotl nodes which is RARELY */
284 			unsigned old   = env->n_entries;
285 			unsigned n_idx = idx + (idx >> 3);
286 
287 			ARR_RESIZE(lower64_entry_t *, env->entries, n_idx);
288 			memset(&env->entries[old], 0, (n_idx - old) * sizeof(env->entries[0]));
289 			env->n_entries = n_idx;
290 		}
291 		env->entries[idx] = link;
292 		env->flags |= MUST_BE_LOWERED;
293 	} else if (is_Conv(node)) {
294 		/* Conv nodes have two modes */
295 		ir_node *pred = get_Conv_op(node);
296 		mode = get_irn_mode(pred);
297 
298 		if (mode == env->high_signed || mode == env->high_unsigned) {
299 			/* must lower this node either but don't need a link */
300 			env->flags |= MUST_BE_LOWERED;
301 		}
302 		return;
303 	} else if (is_Call(node)) {
304 		/* Special case:  If the result of the Call is never used, we won't
305 		 * find a Proj with a mode that potentially triggers MUST_BE_LOWERED
306 		 * to be set.  Thus, if we see a call, we check its result types and
307 		 * decide whether MUST_BE_LOWERED has to be set.
308 		 */
309 		ir_type *tp = get_Call_type(node);
310 		size_t   n_res, i;
311 
312 		n_res = get_method_n_ress(tp);
313 		for (i = 0; i < n_res; ++i) {
314 			ir_type *rtp = get_method_res_type(tp, i);
315 
316 			if (is_Primitive_type(rtp)) {
317 				ir_mode *rmode = get_type_mode(rtp);
318 
319 				if (rmode == env->high_signed || rmode == env->high_unsigned) {
320 					env->flags |= MUST_BE_LOWERED;
321 				}
322 			}
323 		}
324 	}
325 }
326 
get_node_entry(ir_node * node)327 lower64_entry_t *get_node_entry(ir_node *node)
328 {
329 	unsigned idx = get_irn_idx(node);
330 	assert(idx < env->n_entries);
331 	return env->entries[idx];
332 }
333 
ir_set_dw_lowered(ir_node * old,ir_node * new_low,ir_node * new_high)334 void ir_set_dw_lowered(ir_node *old, ir_node *new_low, ir_node *new_high)
335 {
336 	lower64_entry_t *entry = get_node_entry(old);
337 	entry->low_word  = new_low;
338 	entry->high_word = new_high;
339 }
340 
ir_get_low_unsigned_mode(void)341 ir_mode *ir_get_low_unsigned_mode(void)
342 {
343 	return env->low_unsigned;
344 }
345 
346 /**
347  * Translate a Constant: create two.
348  */
lower_Const(ir_node * node,ir_mode * mode)349 static void lower_Const(ir_node *node, ir_mode *mode)
350 {
351 	ir_graph  *irg      = get_irn_irg(node);
352 	dbg_info  *dbg      = get_irn_dbg_info(node);
353 	ir_mode   *low_mode = env->low_unsigned;
354 	ir_tarval *tv       = get_Const_tarval(node);
355 	ir_tarval *tv_l     = tarval_convert_to(tv, low_mode);
356 	ir_node   *res_low  = new_rd_Const(dbg, irg, tv_l);
357 	ir_tarval *tv_shrs  = tarval_shrs_unsigned(tv, get_mode_size_bits(low_mode));
358 	ir_tarval *tv_h     = tarval_convert_to(tv_shrs, mode);
359 	ir_node   *res_high = new_rd_Const(dbg, irg, tv_h);
360 
361 	ir_set_dw_lowered(node, res_low, res_high);
362 }
363 
364 /**
365  * Translate a Load: create two.
366  */
lower_Load(ir_node * node,ir_mode * mode)367 static void lower_Load(ir_node *node, ir_mode *mode)
368 {
369 	ir_mode    *low_mode = env->low_unsigned;
370 	ir_graph   *irg = get_irn_irg(node);
371 	ir_node    *adr = get_Load_ptr(node);
372 	ir_node    *mem = get_Load_mem(node);
373 	ir_node    *low;
374 	ir_node    *high;
375 	ir_node    *proj_m;
376 	dbg_info   *dbg;
377 	ir_node    *block = get_nodes_block(node);
378 	ir_cons_flags volatility = get_Load_volatility(node) == volatility_is_volatile
379 	                         ? cons_volatile : cons_none;
380 
381 	if (env->params->little_endian) {
382 		low  = adr;
383 		high = new_r_Add(block, adr, new_r_Const(irg, env->tv_mode_bytes), get_irn_mode(adr));
384 	} else {
385 		low  = new_r_Add(block, adr, new_r_Const(irg, env->tv_mode_bytes), get_irn_mode(adr));
386 		high = adr;
387 	}
388 
389 	/* create two loads */
390 	dbg    = get_irn_dbg_info(node);
391 	low    = new_rd_Load(dbg, block, mem,  low,  low_mode, volatility);
392 	proj_m = new_r_Proj(low, mode_M, pn_Load_M);
393 	high   = new_rd_Load(dbg, block, proj_m, high, mode, volatility);
394 
395 	foreach_out_edge_safe(node, edge) {
396 		ir_node *proj = get_edge_src_irn(edge);
397 		if (!is_Proj(proj))
398 			continue;
399 
400 		switch (get_Proj_proj(proj)) {
401 		case pn_Load_M:         /* Memory result. */
402 			/* put it to the second one */
403 			set_Proj_pred(proj, high);
404 			break;
405 		case pn_Load_X_except:  /* Execution result if exception occurred. */
406 			/* put it to the first one */
407 			set_Proj_pred(proj, low);
408 			break;
409 		case pn_Load_res: {       /* Result of load operation. */
410 			ir_node *res_low  = new_r_Proj(low,  low_mode, pn_Load_res);
411 			ir_node *res_high = new_r_Proj(high, mode,     pn_Load_res);
412 			ir_set_dw_lowered(proj, res_low, res_high);
413 			break;
414 		}
415 		default:
416 			assert(0 && "unexpected Proj number");
417 		}
418 		/* mark this proj: we have handled it already, otherwise we might fall
419 		 * into out new nodes. */
420 		mark_irn_visited(proj);
421 	}
422 }
423 
424 /**
425  * Translate a Store: create two.
426  */
lower_Store(ir_node * node,ir_mode * mode)427 static void lower_Store(ir_node *node, ir_mode *mode)
428 {
429 	ir_graph              *irg;
430 	ir_node               *block, *adr, *mem;
431 	ir_node               *low, *high, *proj_m;
432 	dbg_info              *dbg;
433 	ir_node               *value = get_Store_value(node);
434 	const lower64_entry_t *entry = get_node_entry(value);
435 	ir_cons_flags volatility = get_Store_volatility(node) == volatility_is_volatile
436 	                           ? cons_volatile : cons_none;
437 	(void) mode;
438 
439 	assert(entry);
440 
441 	if (! entry->low_word) {
442 		/* not ready yet, wait */
443 		pdeq_putr(env->waitq, node);
444 		return;
445 	}
446 
447 	irg = get_irn_irg(node);
448 	adr = get_Store_ptr(node);
449 	mem = get_Store_mem(node);
450 	block = get_nodes_block(node);
451 
452 	if (env->params->little_endian) {
453 		low  = adr;
454 		high = new_r_Add(block, adr, new_r_Const(irg, env->tv_mode_bytes), get_irn_mode(adr));
455 	} else {
456 		low  = new_r_Add(block, adr, new_r_Const(irg, env->tv_mode_bytes), get_irn_mode(adr));
457 		high = adr;
458 	}
459 
460 	/* create two Stores */
461 	dbg    = get_irn_dbg_info(node);
462 	low    = new_rd_Store(dbg, block, mem, low,  entry->low_word, volatility);
463 	proj_m = new_r_Proj(low, mode_M, pn_Store_M);
464 	high   = new_rd_Store(dbg, block, proj_m, high, entry->high_word, volatility);
465 
466 	foreach_out_edge_safe(node, edge) {
467 		ir_node *proj = get_edge_src_irn(edge);
468 		if (!is_Proj(proj))
469 			continue;
470 
471 		switch (get_Proj_proj(proj)) {
472 		case pn_Store_M:         /* Memory result. */
473 			/* put it to the second one */
474 			set_Proj_pred(proj, high);
475 			break;
476 		case pn_Store_X_except:  /* Execution result if exception occurred. */
477 			/* put it to the first one */
478 			set_Proj_pred(proj, low);
479 			break;
480 		default:
481 			assert(0 && "unexpected Proj number");
482 		}
483 		/* mark this proj: we have handled it already, otherwise we might fall into
484 		 * out new nodes. */
485 		mark_irn_visited(proj);
486 	}
487 }
488 
489 /**
490  * Return a node containing the address of the intrinsic emulation function.
491  *
492  * @param method  the method type of the emulation function
493  * @param op      the emulated ir_op
494  * @param imode   the input mode of the emulated opcode
495  * @param omode   the output mode of the emulated opcode
496  * @param env     the lower environment
497  */
get_intrinsic_address(ir_type * method,ir_op * op,ir_mode * imode,ir_mode * omode)498 static ir_node *get_intrinsic_address(ir_type *method, ir_op *op,
499                                       ir_mode *imode, ir_mode *omode)
500 {
501 	symconst_symbol sym;
502 	ir_entity *ent;
503 	op_mode_entry_t key, *entry;
504 
505 	key.op    = op;
506 	key.imode = imode;
507 	key.omode = omode;
508 	key.ent   = NULL;
509 
510 	entry = set_insert(op_mode_entry_t, intrinsic_fkt, &key, sizeof(key),
511 				hash_ptr(op) ^ hash_ptr(imode) ^ (hash_ptr(omode) << 8));
512 	if (! entry->ent) {
513 		/* create a new one */
514 		ent = env->params->create_intrinsic(method, op, imode, omode, env->params->ctx);
515 
516 		assert(ent && "Intrinsic creator must return an entity");
517 		entry->ent = ent;
518 	} else {
519 		ent = entry->ent;
520 	}
521 	sym.entity_p = ent;
522 	return new_r_SymConst(env->irg, mode_P_code, sym, symconst_addr_ent);
523 }
524 
525 /**
526  * Translate a Div.
527  *
528  * Create an intrinsic Call.
529  */
lower_Div(ir_node * node,ir_mode * mode)530 static void lower_Div(ir_node *node, ir_mode *mode)
531 {
532 	ir_node  *left   = get_Div_left(node);
533 	ir_node  *right  = get_Div_right(node);
534 	ir_node  *block  = get_nodes_block(node);
535 	dbg_info *dbgi   = get_irn_dbg_info(node);
536 	ir_type  *mtp    = mode_is_signed(mode) ? binop_tp_s : binop_tp_u;
537 	ir_mode  *opmode = get_irn_op_mode(node);
538 	ir_node  *addr   = get_intrinsic_address(mtp, get_irn_op(node), opmode, opmode);
539 	ir_node  *in[4];
540 	ir_node  *call;
541 	ir_node  *resproj;
542 
543 	if (env->params->little_endian) {
544 		in[0] = get_lowered_low(left);
545 		in[1] = get_lowered_high(left);
546 		in[2] = get_lowered_low(right);
547 		in[3] = get_lowered_high(right);
548 	} else {
549 		in[0] = get_lowered_high(left);
550 		in[1] = get_lowered_low(left);
551 		in[2] = get_lowered_high(right);
552 		in[3] = get_lowered_low(right);
553 	}
554 	call    = new_rd_Call(dbgi, block, get_Div_mem(node), addr, 4, in, mtp);
555 	resproj = new_r_Proj(call, mode_T, pn_Call_T_result);
556 	set_irn_pinned(call, get_irn_pinned(node));
557 
558 	foreach_out_edge_safe(node, edge) {
559 		ir_node *proj = get_edge_src_irn(edge);
560 		if (!is_Proj(proj))
561 			continue;
562 
563 		switch (get_Proj_proj(proj)) {
564 		case pn_Div_M:         /* Memory result. */
565 			/* reroute to the call */
566 			set_Proj_pred(proj, call);
567 			set_Proj_proj(proj, pn_Call_M);
568 			break;
569 		case pn_Div_X_regular:
570 			set_Proj_pred(proj, call);
571 			set_Proj_proj(proj, pn_Call_X_regular);
572 			break;
573 		case pn_Div_X_except:
574 			set_Proj_pred(proj, call);
575 			set_Proj_proj(proj, pn_Call_X_except);
576 			break;
577 		case pn_Div_res:
578 			if (env->params->little_endian) {
579 				ir_node *res_low  = new_r_Proj(resproj, env->low_unsigned, 0);
580 				ir_node *res_high = new_r_Proj(resproj, mode,              1);
581 				ir_set_dw_lowered(proj, res_low, res_high);
582 			} else {
583 				ir_node *res_low  = new_r_Proj(resproj, env->low_unsigned, 1);
584 				ir_node *res_high = new_r_Proj(resproj, mode,              0);
585 				ir_set_dw_lowered(proj, res_low, res_high);
586 			}
587 			break;
588 		default:
589 			assert(0 && "unexpected Proj number");
590 		}
591 		/* mark this proj: we have handled it already, otherwise we might fall into
592 		 * out new nodes. */
593 		mark_irn_visited(proj);
594 	}
595 }
596 
597 /**
598  * Translate a Mod.
599  *
600  * Create an intrinsic Call.
601  */
lower_Mod(ir_node * node,ir_mode * mode)602 static void lower_Mod(ir_node *node, ir_mode *mode)
603 {
604 	ir_node  *left   = get_Mod_left(node);
605 	ir_node  *right  = get_Mod_right(node);
606 	dbg_info *dbgi   = get_irn_dbg_info(node);
607 	ir_node  *block  = get_nodes_block(node);
608 	ir_type  *mtp    = mode_is_signed(mode) ? binop_tp_s : binop_tp_u;
609 	ir_mode  *opmode = get_irn_op_mode(node);
610 	ir_node  *addr   = get_intrinsic_address(mtp, get_irn_op(node), opmode, opmode);
611 	ir_node  *in[4];
612 	ir_node  *call;
613 	ir_node  *resproj;
614 
615 	if (env->params->little_endian) {
616 		in[0] = get_lowered_low(left);
617 		in[1] = get_lowered_high(left);
618 		in[2] = get_lowered_low(right);
619 		in[3] = get_lowered_high(right);
620 	} else {
621 		in[0] = get_lowered_high(left);
622 		in[1] = get_lowered_low(left);
623 		in[2] = get_lowered_high(right);
624 		in[3] = get_lowered_low(right);
625 	}
626 	call    = new_rd_Call(dbgi, block, get_Mod_mem(node), addr, 4, in, mtp);
627 	resproj = new_r_Proj(call, mode_T, pn_Call_T_result);
628 	set_irn_pinned(call, get_irn_pinned(node));
629 
630 	foreach_out_edge_safe(node, edge) {
631 		ir_node *proj = get_edge_src_irn(edge);
632 		if (!is_Proj(proj))
633 			continue;
634 
635 		switch (get_Proj_proj(proj)) {
636 		case pn_Mod_M:         /* Memory result. */
637 			/* reroute to the call */
638 			set_Proj_pred(proj, call);
639 			set_Proj_proj(proj, pn_Call_M);
640 			break;
641 		case pn_Div_X_regular:
642 			set_Proj_pred(proj, call);
643 			set_Proj_proj(proj, pn_Call_X_regular);
644 			break;
645 		case pn_Mod_X_except:
646 			set_Proj_pred(proj, call);
647 			set_Proj_proj(proj, pn_Call_X_except);
648 			break;
649 		case pn_Mod_res:
650 			if (env->params->little_endian) {
651 				ir_node *res_low  = new_r_Proj(resproj, env->low_unsigned, 0);
652 				ir_node *res_high = new_r_Proj(resproj, mode,              1);
653 				ir_set_dw_lowered(proj, res_low, res_high);
654 			} else {
655 				ir_node *res_low  = new_r_Proj(resproj, env->low_unsigned, 1);
656 				ir_node *res_high = new_r_Proj(resproj, mode,              0);
657 				ir_set_dw_lowered(proj, res_low, res_high);
658 			}
659 			break;
660 		default:
661 			assert(0 && "unexpected Proj number");
662 		}
663 		/* mark this proj: we have handled it already, otherwise we might fall
664 		 * into out new nodes. */
665 		mark_irn_visited(proj);
666 	}
667 }
668 
669 /**
670  * Translate a binop.
671  *
672  * Create an intrinsic Call.
673  */
lower_binop(ir_node * node,ir_mode * mode)674 static void lower_binop(ir_node *node, ir_mode *mode)
675 {
676 	ir_node  *left  = get_binop_left(node);
677 	ir_node  *right = get_binop_right(node);
678 	dbg_info *dbgi  = get_irn_dbg_info(node);
679 	ir_node  *block = get_nodes_block(node);
680 	ir_graph *irg   = get_irn_irg(block);
681 	ir_type  *mtp   = mode_is_signed(mode) ? binop_tp_s : binop_tp_u;
682 	ir_node  *addr  = get_intrinsic_address(mtp, get_irn_op(node), mode, mode);
683 	ir_node  *in[4];
684 	ir_node  *call;
685 	ir_node  *resproj;
686 
687 	if (env->params->little_endian) {
688 		in[0] = get_lowered_low(left);
689 		in[1] = get_lowered_high(left);
690 		in[2] = get_lowered_low(right);
691 		in[3] = get_lowered_high(right);
692 	} else {
693 		in[0] = get_lowered_high(left);
694 		in[1] = get_lowered_low(left);
695 		in[2] = get_lowered_high(right);
696 		in[3] = get_lowered_low(right);
697 	}
698 	call    = new_rd_Call(dbgi, block, get_irg_no_mem(irg), addr, 4, in, mtp);
699 	resproj = new_r_Proj(call, mode_T, pn_Call_T_result);
700 	set_irn_pinned(call, get_irn_pinned(node));
701 
702 	if (env->params->little_endian) {
703 		ir_node *res_low  = new_r_Proj(resproj, env->low_unsigned, 0);
704 		ir_node *res_high = new_r_Proj(resproj, mode,              1);
705 		ir_set_dw_lowered(node, res_low, res_high);
706 	} else {
707 		ir_node *res_low  = new_r_Proj(resproj, env->low_unsigned, 1);
708 		ir_node *res_high = new_r_Proj(resproj, mode,              0);
709 		ir_set_dw_lowered(node, res_low, res_high);
710 	}
711 }
712 
create_conv(ir_node * block,ir_node * node,ir_mode * dest_mode)713 static ir_node *create_conv(ir_node *block, ir_node *node, ir_mode *dest_mode)
714 {
715 	if (get_irn_mode(node) == dest_mode)
716 		return node;
717 	return new_r_Conv(block, node, dest_mode);
718 }
719 
720 /**
721  * Moves node and all predecessors of node from from_bl to to_bl.
722  * Does not move predecessors of Phi nodes (or block nodes).
723  */
move(ir_node * node,ir_node * from_bl,ir_node * to_bl)724 static void move(ir_node *node, ir_node *from_bl, ir_node *to_bl)
725 {
726 	int i, arity;
727 
728 	/* move this node */
729 	set_nodes_block(node, to_bl);
730 
731 	/* move its Projs */
732 	if (get_irn_mode(node) == mode_T) {
733 		foreach_out_edge(node, edge) {
734 			ir_node *proj = get_edge_src_irn(edge);
735 			if (!is_Proj(proj))
736 				continue;
737 			move(proj, from_bl, to_bl);
738 		}
739 	}
740 
741 	/* We must not move predecessors of Phi nodes, even if they are in
742 	 * from_bl. (because these are values from an earlier loop iteration
743 	 * which are not predecessors of node here)
744 	 */
745 	if (is_Phi(node))
746 		return;
747 
748 	/* recursion ... */
749 	arity = get_irn_arity(node);
750 	for (i = 0; i < arity; i++) {
751 		ir_node *pred      = get_irn_n(node, i);
752 		ir_mode *pred_mode = get_irn_mode(pred);
753 		if (get_nodes_block(pred) == from_bl)
754 			move(pred, from_bl, to_bl);
755 		if (pred_mode == env->high_signed || pred_mode == env->high_unsigned) {
756 			ir_node *pred_low  = get_lowered_low(pred);
757 			ir_node *pred_high = get_lowered_high(pred);
758 			if (get_nodes_block(pred_low) == from_bl)
759 				move(pred_low, from_bl, to_bl);
760 			if (pred_high != NULL && get_nodes_block(pred_high) == from_bl)
761 				move(pred_high, from_bl, to_bl);
762 		}
763 	}
764 }
765 
766 /**
767  * We need a custom version of part_block_edges because during transformation
768  * not all data-dependencies are explicit yet if a lowered nodes users are not
769  * lowered yet.
770  * We can fix this by modifying move to look for such implicit dependencies.
771  * Additionally we have to keep the proj_2_block map updated
772  */
part_block_dw(ir_node * node)773 static ir_node *part_block_dw(ir_node *node)
774 {
775 	ir_graph *irg        = get_irn_irg(node);
776 	ir_node  *old_block  = get_nodes_block(node);
777 	int       n_cfgpreds = get_Block_n_cfgpreds(old_block);
778 	ir_node **cfgpreds   = get_Block_cfgpred_arr(old_block);
779 	ir_node  *new_block  = new_r_Block(irg, n_cfgpreds, cfgpreds);
780 
781 	/* old_block has no predecessors anymore for now */
782 	set_irn_in(old_block, 0, NULL);
783 
784 	/* move node and its predecessors to new_block */
785 	move(node, old_block, new_block);
786 
787 	/* move Phi nodes to new_block */
788 	foreach_out_edge_safe(old_block, edge) {
789 		ir_node *phi = get_edge_src_irn(edge);
790 		if (!is_Phi(phi))
791 			continue;
792 		set_nodes_block(phi, new_block);
793 	}
794 	return old_block;
795 }
796 
797 typedef ir_node* (*new_rd_shr_func)(dbg_info *dbgi, ir_node *block,
798                                     ir_node *left, ir_node *right,
799                                     ir_mode *mode);
800 
lower_shr_helper(ir_node * node,ir_mode * mode,new_rd_shr_func new_rd_shrs)801 static void lower_shr_helper(ir_node *node, ir_mode *mode,
802                              new_rd_shr_func new_rd_shrs)
803 {
804 	ir_node  *right         = get_binop_right(node);
805 	ir_node  *left          = get_binop_left(node);
806 	ir_mode  *shr_mode      = get_irn_mode(node);
807 	unsigned  modulo_shift  = get_mode_modulo_shift(shr_mode);
808 	ir_mode  *low_unsigned  = env->low_unsigned;
809 	unsigned  modulo_shift2 = get_mode_modulo_shift(mode);
810 	ir_graph *irg           = get_irn_irg(node);
811 	ir_node  *left_low      = get_lowered_low(left);
812 	ir_node  *left_high     = get_lowered_high(left);
813 	dbg_info *dbgi          = get_irn_dbg_info(node);
814 	ir_node  *lower_block;
815 	ir_node  *block;
816 	ir_node  *cnst;
817 	ir_node  *andn;
818 	ir_node  *cmp;
819 	ir_node  *cond;
820 	ir_node  *proj_true;
821 	ir_node  *proj_false;
822 	ir_node  *phi_low;
823 	ir_node  *phi_high;
824 	ir_node  *lower_in[2];
825 	ir_node  *phi_low_in[2];
826 	ir_node  *phi_high_in[2];
827 
828 	/* this version is optimized for modulo shift architectures
829 	 * (and can't handle anything else) */
830 	if (modulo_shift != get_mode_size_bits(shr_mode)
831 			|| modulo_shift2<<1 != modulo_shift) {
832 		panic("Shr lowering only implemented for modulo shift shr operations");
833 	}
834 	if (!is_po2(modulo_shift) || !is_po2(modulo_shift2)) {
835 		panic("Shr lowering only implemented for power-of-2 modes");
836 	}
837 	/* without 2-complement the -x instead of (bit_width-x) trick won't work */
838 	if (get_mode_arithmetic(shr_mode) != irma_twos_complement) {
839 		panic("Shr lowering only implemented for two-complement modes");
840 	}
841 
842 	block = get_nodes_block(node);
843 
844 	/* if the right operand is a 64bit value, we're only interested in the
845 	 * lower word */
846 	if (get_irn_mode(right) == env->high_unsigned) {
847 		right = get_lowered_low(right);
848 	} else {
849 		/* shift should never have signed mode on the right */
850 		assert(get_irn_mode(right) != env->high_signed);
851 		right = create_conv(block, right, low_unsigned);
852 	}
853 
854 	lower_block = part_block_dw(node);
855 	env->flags |= CF_CHANGED;
856 	block = get_nodes_block(node);
857 
858 	/* add a Cmp to test if highest bit is set <=> whether we shift more
859 	 * than half the word width */
860 	cnst       = new_r_Const_long(irg, low_unsigned, modulo_shift2);
861 	andn       = new_r_And(block, right, cnst, low_unsigned);
862 	cnst       = new_r_Const(irg, get_mode_null(low_unsigned));
863 	cmp        = new_rd_Cmp(dbgi, block, andn, cnst, ir_relation_equal);
864 	cond       = new_rd_Cond(dbgi, block, cmp);
865 	proj_true  = new_r_Proj(cond, mode_X, pn_Cond_true);
866 	proj_false = new_r_Proj(cond, mode_X, pn_Cond_false);
867 
868 	/* the true block => shift_width < 1word */
869 	{
870 		/* In theory the low value (for 64bit shifts) is:
871 		 *    Or(High << (32-x)), Low >> x)
872 		 * In practice High << 32-x will fail when x is zero (since we have
873 		 * modulo shift and 32 will be 0). So instead we use:
874 		 *    Or(High<<1<<~x, Low >> x)
875 		 */
876 		ir_node *in[1]        = { proj_true };
877 		ir_node *block_true   = new_r_Block(irg, ARRAY_SIZE(in), in);
878 		ir_node *res_high     = new_rd_shrs(dbgi, block_true, left_high,
879 		                                    right, mode);
880 		ir_node *shift_low    = new_rd_Shr(dbgi, block_true, left_low, right,
881 		                                   low_unsigned);
882 		ir_node *not_shiftval = new_rd_Not(dbgi, block_true, right,
883 		                                   low_unsigned);
884 		ir_node *conv         = create_conv(block_true, left_high,
885 		                                    low_unsigned);
886 		ir_node *one          = new_r_Const(irg, get_mode_one(low_unsigned));
887 		ir_node *carry0       = new_rd_Shl(dbgi, block_true, conv, one,
888 		                                   low_unsigned);
889 		ir_node *carry1       = new_rd_Shl(dbgi, block_true, carry0,
890 		                                   not_shiftval, low_unsigned);
891 		ir_node *res_low      = new_rd_Or(dbgi, block_true, shift_low, carry1,
892 		                                  low_unsigned);
893 		lower_in[0]           = new_r_Jmp(block_true);
894 		phi_low_in[0]         = res_low;
895 		phi_high_in[0]        = res_high;
896 	}
897 
898 	/* false block => shift_width > 1word */
899 	{
900 		ir_node *in[1]       = { proj_false };
901 		ir_node *block_false = new_r_Block(irg, ARRAY_SIZE(in), in);
902 		ir_node *conv        = create_conv(block_false, left_high, low_unsigned);
903 		ir_node *res_low     = new_rd_shrs(dbgi, block_false, conv, right,
904 		                                   low_unsigned);
905 		int      cnsti       = modulo_shift2-1;
906 		ir_node *cnst2       = new_r_Const_long(irg, low_unsigned, cnsti);
907 		ir_node *res_high;
908 		if (new_rd_shrs == new_rd_Shrs) {
909 			res_high = new_rd_shrs(dbgi, block_false, left_high, cnst2, mode);
910 		} else {
911 			res_high = new_r_Const(irg, get_mode_null(mode));
912 		}
913 		lower_in[1]          = new_r_Jmp(block_false);
914 		phi_low_in[1]        = res_low;
915 		phi_high_in[1]       = res_high;
916 	}
917 
918 	/* patch lower block */
919 	set_irn_in(lower_block, ARRAY_SIZE(lower_in), lower_in);
920 	phi_low  = new_r_Phi(lower_block, ARRAY_SIZE(phi_low_in), phi_low_in,
921 	                     low_unsigned);
922 	phi_high = new_r_Phi(lower_block, ARRAY_SIZE(phi_high_in), phi_high_in,
923 	                     mode);
924 	ir_set_dw_lowered(node, phi_low, phi_high);
925 }
926 
lower_Shr(ir_node * node,ir_mode * mode)927 static void lower_Shr(ir_node *node, ir_mode *mode)
928 {
929 	lower_shr_helper(node, mode, new_rd_Shr);
930 }
931 
lower_Shrs(ir_node * node,ir_mode * mode)932 static void lower_Shrs(ir_node *node, ir_mode *mode)
933 {
934 	lower_shr_helper(node, mode, new_rd_Shrs);
935 }
936 
lower_Shl(ir_node * node,ir_mode * mode)937 static void lower_Shl(ir_node *node, ir_mode *mode)
938 {
939 	ir_node  *right         = get_binop_right(node);
940 	ir_node  *left          = get_binop_left(node);
941 	ir_mode  *shr_mode      = get_irn_mode(node);
942 	unsigned  modulo_shift  = get_mode_modulo_shift(shr_mode);
943 	ir_mode  *low_unsigned  = env->low_unsigned;
944 	unsigned  modulo_shift2 = get_mode_modulo_shift(mode);
945 	ir_graph *irg           = get_irn_irg(node);
946 	ir_node  *left_low      = get_lowered_low(left);
947 	ir_node  *left_high     = get_lowered_high(left);
948 	dbg_info *dbgi          = get_irn_dbg_info(node);
949 	ir_node  *lower_block   = get_nodes_block(node);
950 	ir_node  *block;
951 	ir_node  *cnst;
952 	ir_node  *andn;
953 	ir_node  *cmp;
954 	ir_node  *cond;
955 	ir_node  *proj_true;
956 	ir_node  *proj_false;
957 	ir_node  *phi_low;
958 	ir_node  *phi_high;
959 	ir_node  *lower_in[2];
960 	ir_node  *phi_low_in[2];
961 	ir_node  *phi_high_in[2];
962 
963 	/* this version is optimized for modulo shift architectures
964 	 * (and can't handle anything else) */
965 	if (modulo_shift != get_mode_size_bits(shr_mode)
966 			|| modulo_shift2<<1 != modulo_shift) {
967 		panic("Shl lowering only implemented for modulo shift shl operations");
968 	}
969 	if (!is_po2(modulo_shift) || !is_po2(modulo_shift2)) {
970 		panic("Shl lowering only implemented for power-of-2 modes");
971 	}
972 	/* without 2-complement the -x instead of (bit_width-x) trick won't work */
973 	if (get_mode_arithmetic(shr_mode) != irma_twos_complement) {
974 		panic("Shl lowering only implemented for two-complement modes");
975 	}
976 
977 	/* if the right operand is a 64bit value, we're only interested in the
978 	 * lower word */
979 	if (get_irn_mode(right) == env->high_unsigned) {
980 		right = get_lowered_low(right);
981 	} else {
982 		/* shift should never have signed mode on the right */
983 		assert(get_irn_mode(right) != env->high_signed);
984 		right = create_conv(lower_block, right, low_unsigned);
985 	}
986 
987 	part_block_dw(node);
988 	env->flags |= CF_CHANGED;
989 	block = get_nodes_block(node);
990 
991 	/* add a Cmp to test if highest bit is set <=> whether we shift more
992 	 * than half the word width */
993 	cnst       = new_r_Const_long(irg, low_unsigned, modulo_shift2);
994 	andn       = new_r_And(block, right, cnst, low_unsigned);
995 	cnst       = new_r_Const(irg, get_mode_null(low_unsigned));
996 	cmp        = new_rd_Cmp(dbgi, block, andn, cnst, ir_relation_equal);
997 	cond       = new_rd_Cond(dbgi, block, cmp);
998 	proj_true  = new_r_Proj(cond, mode_X, pn_Cond_true);
999 	proj_false = new_r_Proj(cond, mode_X, pn_Cond_false);
1000 
1001 	/* the true block => shift_width < 1word */
1002 	{
1003 		ir_node *in[1]        = { proj_true };
1004 		ir_node *block_true   = new_r_Block(irg, ARRAY_SIZE(in), in);
1005 
1006 		ir_node *res_low      = new_rd_Shl(dbgi, block_true, left_low,
1007 		                                   right, low_unsigned);
1008 		ir_node *shift_high   = new_rd_Shl(dbgi, block_true, left_high, right,
1009 		                                   mode);
1010 		ir_node *not_shiftval = new_rd_Not(dbgi, block_true, right,
1011 		                                   low_unsigned);
1012 		ir_node *conv         = create_conv(block_true, left_low, mode);
1013 		ir_node *one          = new_r_Const(irg, get_mode_one(low_unsigned));
1014 		ir_node *carry0       = new_rd_Shr(dbgi, block_true, conv, one, mode);
1015 		ir_node *carry1       = new_rd_Shr(dbgi, block_true, carry0,
1016 		                                   not_shiftval, mode);
1017 		ir_node *res_high     = new_rd_Or(dbgi, block_true, shift_high, carry1,
1018 		                                  mode);
1019 		lower_in[0]           = new_r_Jmp(block_true);
1020 		phi_low_in[0]         = res_low;
1021 		phi_high_in[0]        = res_high;
1022 	}
1023 
1024 	/* false block => shift_width > 1word */
1025 	{
1026 		ir_node *in[1]       = { proj_false };
1027 		ir_node *block_false = new_r_Block(irg, ARRAY_SIZE(in), in);
1028 		ir_node *res_low     = new_r_Const(irg, get_mode_null(low_unsigned));
1029 		ir_node *conv        = create_conv(block_false, left_low, mode);
1030 		ir_node *res_high    = new_rd_Shl(dbgi, block_false, conv, right, mode);
1031 		lower_in[1]          = new_r_Jmp(block_false);
1032 		phi_low_in[1]        = res_low;
1033 		phi_high_in[1]       = res_high;
1034 	}
1035 
1036 	/* patch lower block */
1037 	set_irn_in(lower_block, ARRAY_SIZE(lower_in), lower_in);
1038 	phi_low  = new_r_Phi(lower_block, ARRAY_SIZE(phi_low_in), phi_low_in,
1039 	                     low_unsigned);
1040 	phi_high = new_r_Phi(lower_block, ARRAY_SIZE(phi_high_in), phi_high_in,
1041 	                     mode);
1042 	ir_set_dw_lowered(node, phi_low, phi_high);
1043 }
1044 
1045 /**
1046  * Rebuild Rotl nodes into Or(Shl, Shr) and prepare all nodes.
1047  */
prepare_links_and_handle_rotl(ir_node * node,void * data)1048 static void prepare_links_and_handle_rotl(ir_node *node, void *data)
1049 {
1050 	(void) data;
1051 	if (is_Rotl(node)) {
1052 		ir_mode  *mode = get_irn_op_mode(node);
1053 		ir_node  *right;
1054 		ir_node  *left, *shl, *shr, *ornode, *block, *sub, *c;
1055 		ir_mode  *omode, *rmode;
1056 		ir_graph *irg;
1057 		dbg_info *dbg;
1058 		optimization_state_t state;
1059 
1060 		if (mode != env->high_signed && mode != env->high_unsigned) {
1061 			prepare_links(node);
1062 			return;
1063 		}
1064 
1065 		/* replace the Rotl(x,y) by an Or(Shl(x,y), Shr(x,64-y)) */
1066 		right = get_Rotl_right(node);
1067 		irg   = get_irn_irg(node);
1068 		dbg   = get_irn_dbg_info(node);
1069 		omode = get_irn_mode(node);
1070 		left  = get_Rotl_left(node);
1071 		block = get_nodes_block(node);
1072 		shl   = new_rd_Shl(dbg, block, left, right, omode);
1073 		rmode = get_irn_mode(right);
1074 		c     = new_r_Const_long(irg, rmode, get_mode_size_bits(omode));
1075 		sub   = new_rd_Sub(dbg, block, c, right, rmode);
1076 		shr   = new_rd_Shr(dbg, block, left, sub, omode);
1077 
1078 		/* switch optimization off here, or we will get the Rotl back */
1079 		save_optimization_state(&state);
1080 		set_opt_algebraic_simplification(0);
1081 		ornode = new_rd_Or(dbg, block, shl, shr, omode);
1082 		restore_optimization_state(&state);
1083 
1084 		exchange(node, ornode);
1085 
1086 		/* do lowering on the new nodes */
1087 		prepare_links(shl);
1088 		prepare_links(c);
1089 		prepare_links(sub);
1090 		prepare_links(shr);
1091 		prepare_links(ornode);
1092 		return;
1093 	}
1094 
1095 	prepare_links(node);
1096 }
1097 
1098 /**
1099  * Translate an Unop.
1100  *
1101  * Create an intrinsic Call.
1102  */
lower_unop(ir_node * node,ir_mode * mode)1103 static void lower_unop(ir_node *node, ir_mode *mode)
1104 {
1105 	ir_node  *op       = get_unop_op(node);
1106 	dbg_info *dbgi     = get_irn_dbg_info(node);
1107 	ir_node  *block    = get_nodes_block(node);
1108 	ir_graph *irg      = get_irn_irg(block);
1109 	ir_type  *mtp      = mode_is_signed(mode) ? unop_tp_s : unop_tp_u;
1110 	ir_op    *irop     = get_irn_op(node);
1111 	ir_node  *addr     = get_intrinsic_address(mtp, irop, mode, mode);
1112 	ir_node  *nomem    = get_irg_no_mem(irg);
1113 	ir_node  *in[2];
1114 	ir_node  *call;
1115 	ir_node  *resproj;
1116 
1117 	if (env->params->little_endian) {
1118 		in[0] = get_lowered_low(op);
1119 		in[1] = get_lowered_high(op);
1120 	} else {
1121 		in[0] = get_lowered_high(op);
1122 		in[1] = get_lowered_low(op);
1123 	}
1124 	call    = new_rd_Call(dbgi, block, nomem, addr, 2, in, mtp);
1125 	resproj = new_r_Proj(call, mode_T, pn_Call_T_result);
1126 	set_irn_pinned(call, get_irn_pinned(node));
1127 
1128 	if (env->params->little_endian) {
1129 		ir_node *res_low  = new_r_Proj(resproj, env->low_unsigned, 0);
1130 		ir_node *res_high = new_r_Proj(resproj, mode,              1);
1131 		ir_set_dw_lowered(node, res_low, res_high);
1132 	} else {
1133 		ir_node *res_low  = new_r_Proj(resproj, env->low_unsigned, 1);
1134 		ir_node *res_high = new_r_Proj(resproj, mode,              0);
1135 		ir_set_dw_lowered(node, res_low, res_high);
1136 	}
1137 }
1138 
1139 /**
1140  * Translate a logical binop.
1141  *
1142  * Create two logical binops.
1143  */
lower_binop_logical(ir_node * node,ir_mode * mode,ir_node * (* constr_rd)(dbg_info * db,ir_node * block,ir_node * op1,ir_node * op2,ir_mode * mode))1144 static void lower_binop_logical(ir_node *node, ir_mode *mode,
1145 								ir_node *(*constr_rd)(dbg_info *db, ir_node *block, ir_node *op1, ir_node *op2, ir_mode *mode) )
1146 {
1147 	ir_node               *left        = get_binop_left(node);
1148 	ir_node               *right       = get_binop_right(node);
1149 	const lower64_entry_t *left_entry  = get_node_entry(left);
1150 	const lower64_entry_t *right_entry = get_node_entry(right);
1151 	dbg_info              *dbgi        = get_irn_dbg_info(node);
1152 	ir_node               *block       = get_nodes_block(node);
1153 	ir_node               *res_low
1154 		= constr_rd(dbgi, block, left_entry->low_word, right_entry->low_word,
1155 		            env->low_unsigned);
1156 	ir_node               *res_high
1157 		= constr_rd(dbgi, block, left_entry->high_word, right_entry->high_word,
1158 		            mode);
1159 	ir_set_dw_lowered(node, res_low, res_high);
1160 }
1161 
lower_And(ir_node * node,ir_mode * mode)1162 static void lower_And(ir_node *node, ir_mode *mode)
1163 {
1164 	lower_binop_logical(node, mode, new_rd_And);
1165 }
1166 
lower_Or(ir_node * node,ir_mode * mode)1167 static void lower_Or(ir_node *node, ir_mode *mode)
1168 {
1169 	lower_binop_logical(node, mode, new_rd_Or);
1170 }
1171 
lower_Eor(ir_node * node,ir_mode * mode)1172 static void lower_Eor(ir_node *node, ir_mode *mode)
1173 {
1174 	lower_binop_logical(node, mode, new_rd_Eor);
1175 }
1176 
1177 /**
1178  * Translate a Not.
1179  *
1180  * Create two logical Nots.
1181  */
lower_Not(ir_node * node,ir_mode * mode)1182 static void lower_Not(ir_node *node, ir_mode *mode)
1183 {
1184 	ir_node               *op       = get_Not_op(node);
1185 	const lower64_entry_t *op_entry = get_node_entry(op);
1186 	dbg_info              *dbgi     = get_irn_dbg_info(node);
1187 	ir_node               *block    = get_nodes_block(node);
1188 	ir_node               *res_low
1189 		= new_rd_Not(dbgi, block, op_entry->low_word, env->low_unsigned);
1190 	ir_node               *res_high
1191 		= new_rd_Not(dbgi, block, op_entry->high_word, mode);
1192 	ir_set_dw_lowered(node, res_low, res_high);
1193 }
1194 
lower_Proj(ir_node * node,ir_mode * op_mode)1195 static void lower_Proj(ir_node *node, ir_mode *op_mode)
1196 {
1197 	ir_mode *mode = get_irn_mode(node);
1198 	ir_node *pred;
1199 	(void)op_mode;
1200 	if (mode != env->high_signed && mode != env->high_unsigned)
1201 		return;
1202 	/* skip tuples */
1203 	pred = get_Proj_pred(node);
1204 	if (is_Tuple(pred)) {
1205 		long                   pn    = get_Proj_proj(node);
1206 		ir_node               *op    = get_irn_n(pred, pn);
1207 		const lower64_entry_t *entry = get_node_entry(op);
1208 		ir_set_dw_lowered(node, entry->low_word, entry->high_word);
1209 	}
1210 }
1211 
is_equality_cmp(const ir_node * node)1212 static bool is_equality_cmp(const ir_node *node)
1213 {
1214 	ir_relation relation = get_Cmp_relation(node);
1215 	ir_node    *left     = get_Cmp_left(node);
1216 	ir_node    *right    = get_Cmp_right(node);
1217 	ir_mode    *mode     = get_irn_mode(left);
1218 
1219 	/* this probably makes no sense if unordered is involved */
1220 	assert(!mode_is_float(mode));
1221 
1222 	if (relation == ir_relation_equal || relation == ir_relation_less_greater)
1223 		return true;
1224 
1225 	if (!is_Const(right) || !is_Const_null(right))
1226 		return false;
1227 	if (mode_is_signed(mode)) {
1228 		return relation == ir_relation_less_greater;
1229 	} else {
1230 		return relation == ir_relation_greater;
1231 	}
1232 }
1233 
get_cfop_destination(const ir_node * cfop)1234 static ir_node *get_cfop_destination(const ir_node *cfop)
1235 {
1236 	const ir_edge_t *first = get_irn_out_edge_first(cfop);
1237 	/* we should only have 1 destination */
1238 	assert(get_irn_n_edges(cfop) == 1);
1239 	return get_edge_src_irn(first);
1240 }
1241 
lower_Switch(ir_node * node,ir_mode * high_mode)1242 static void lower_Switch(ir_node *node, ir_mode *high_mode)
1243 {
1244 	ir_node *selector = get_Switch_selector(node);
1245 	ir_mode *mode     = get_irn_mode(selector);
1246 	(void)high_mode;
1247 	if (mode == env->high_signed || mode == env->high_unsigned) {
1248 		/* we can't really handle Switch with 64bit offsets */
1249 		panic("Switch with 64bit jumptable not supported");
1250 	}
1251 	lower_node(selector);
1252 }
1253 
1254 /**
1255  * Translate a Cond.
1256  */
lower_Cond(ir_node * node,ir_mode * high_mode)1257 static void lower_Cond(ir_node *node, ir_mode *high_mode)
1258 {
1259 	ir_node *left, *right, *block;
1260 	ir_node *sel = get_Cond_selector(node);
1261 	ir_mode *cmp_mode;
1262 	const lower64_entry_t *lentry, *rentry;
1263 	ir_node  *projT = NULL, *projF = NULL;
1264 	ir_node  *new_bl, *irn;
1265 	ir_node  *projHF, *projHT;
1266 	ir_node  *dst_blk;
1267 	ir_relation relation;
1268 	ir_graph *irg;
1269 	dbg_info *dbg;
1270 
1271 	(void) high_mode;
1272 
1273 	if (!is_Cmp(sel)) {
1274 		lower_node(sel);
1275 		return;
1276 	}
1277 
1278 	left     = get_Cmp_left(sel);
1279 	cmp_mode = get_irn_mode(left);
1280 	if (cmp_mode != env->high_signed && cmp_mode != env->high_unsigned) {
1281 		lower_node(sel);
1282 		return;
1283 	}
1284 
1285 	right  = get_Cmp_right(sel);
1286 	lower_node(left);
1287 	lower_node(right);
1288 	lentry = get_node_entry(left);
1289 	rentry = get_node_entry(right);
1290 
1291 	/* all right, build the code */
1292 	foreach_out_edge_safe(node, edge) {
1293 		ir_node *proj    = get_edge_src_irn(edge);
1294 		long     proj_nr;
1295 		if (!is_Proj(proj))
1296 			continue;
1297 		proj_nr = get_Proj_proj(proj);
1298 
1299 		if (proj_nr == pn_Cond_true) {
1300 			assert(projT == NULL && "more than one Proj(true)");
1301 			projT = proj;
1302 		} else {
1303 			assert(proj_nr == pn_Cond_false);
1304 			assert(projF == NULL && "more than one Proj(false)");
1305 			projF = proj;
1306 		}
1307 		mark_irn_visited(proj);
1308 	}
1309 	assert(projT && projF);
1310 
1311 	/* create a new high compare */
1312 	block    = get_nodes_block(node);
1313 	irg      = get_Block_irg(block);
1314 	dbg      = get_irn_dbg_info(sel);
1315 	relation = get_Cmp_relation(sel);
1316 
1317 	if (is_equality_cmp(sel)) {
1318 		/* x ==/!= y ==> or(x_low^y_low,x_high^y_high) ==/!= 0 */
1319 		ir_mode *mode       = env->low_unsigned;
1320 		ir_node *low_left   = new_rd_Conv(dbg, block, lentry->low_word, mode);
1321 		ir_node *high_left  = new_rd_Conv(dbg, block, lentry->high_word, mode);
1322 		ir_node *low_right  = new_rd_Conv(dbg, block, rentry->low_word, mode);
1323 		ir_node *high_right = new_rd_Conv(dbg, block, rentry->high_word, mode);
1324 		ir_node *xor_low    = new_rd_Eor(dbg, block, low_left, low_right, mode);
1325 		ir_node *xor_high   = new_rd_Eor(dbg, block, high_left, high_right, mode);
1326 		ir_node *ornode = new_rd_Or(dbg, block, xor_low, xor_high, mode);
1327 		ir_node *cmp    = new_rd_Cmp(dbg, block, ornode, new_r_Const(irg, get_mode_null(mode)), relation);
1328 		set_Cond_selector(node, cmp);
1329 		return;
1330 	}
1331 
1332 	if (relation == ir_relation_equal) {
1333 		ir_node *proj;
1334 		/* simple case:a == b <==> a_h == b_h && a_l == b_l */
1335 		dst_blk = get_cfop_destination(projF);
1336 
1337 		irn = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1338 		                 ir_relation_equal);
1339 		dbg = get_irn_dbg_info(node);
1340 		irn = new_rd_Cond(dbg, block, irn);
1341 
1342 		projHF = new_r_Proj(irn, mode_X, pn_Cond_false);
1343 		mark_irn_visited(projHF);
1344 		exchange(projF, projHF);
1345 
1346 		projHT = new_r_Proj(irn, mode_X, pn_Cond_true);
1347 		mark_irn_visited(projHT);
1348 
1349 		new_bl = new_r_Block(irg, 1, &projHT);
1350 
1351 		dbg = get_irn_dbg_info(sel);
1352 		irn = new_rd_Cmp(dbg, new_bl, lentry->low_word, rentry->low_word,
1353 		                  ir_relation_equal);
1354 		dbg = get_irn_dbg_info(node);
1355 		irn = new_rd_Cond(dbg, new_bl, irn);
1356 
1357 		proj = new_r_Proj(irn, mode_X, pn_Cond_false);
1358 		mark_irn_visited(proj);
1359 		add_block_cf_input(dst_blk, projHF, proj);
1360 
1361 		proj = new_r_Proj(irn, mode_X, pn_Cond_true);
1362 		mark_irn_visited(proj);
1363 		exchange(projT, proj);
1364 	} else if (relation == ir_relation_less_greater) {
1365 		ir_node *proj;
1366 		/* simple case:a != b <==> a_h != b_h || a_l != b_l */
1367 		dst_blk = get_cfop_destination(projT);
1368 
1369 		irn = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1370 		                 ir_relation_less_greater);
1371 		dbg = get_irn_dbg_info(node);
1372 		irn = new_rd_Cond(dbg, block, irn);
1373 
1374 		projHT = new_r_Proj(irn, mode_X, pn_Cond_true);
1375 		mark_irn_visited(projHT);
1376 		exchange(projT, projHT);
1377 
1378 		projHF = new_r_Proj(irn, mode_X, pn_Cond_false);
1379 		mark_irn_visited(projHF);
1380 
1381 		new_bl = new_r_Block(irg, 1, &projHF);
1382 
1383 		dbg = get_irn_dbg_info(sel);
1384 		irn = new_rd_Cmp(dbg, new_bl, lentry->low_word, rentry->low_word,
1385 		                 ir_relation_less_greater);
1386 		dbg = get_irn_dbg_info(node);
1387 		irn = new_rd_Cond(dbg, new_bl, irn);
1388 
1389 		proj = new_r_Proj(irn, mode_X, pn_Cond_true);
1390 		mark_irn_visited(proj);
1391 		add_block_cf_input(dst_blk, projHT, proj);
1392 
1393 		proj = new_r_Proj(irn, mode_X, pn_Cond_false);
1394 		mark_irn_visited(proj);
1395 		exchange(projF, proj);
1396 	} else {
1397 		ir_node *proj;
1398 		/* a rel b <==> a_h REL b_h || (a_h == b_h && a_l rel b_l) */
1399 		ir_node *dstT, *dstF, *newbl_eq, *newbl_l;
1400 		ir_node *projEqF;
1401 
1402 		dstT = get_cfop_destination(projT);
1403 		dstF = get_cfop_destination(projF);
1404 
1405 		irn = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1406 		                 relation & ~ir_relation_equal);
1407 		dbg = get_irn_dbg_info(node);
1408 		irn = new_rd_Cond(dbg, block, irn);
1409 
1410 		projHT = new_r_Proj(irn, mode_X, pn_Cond_true);
1411 		mark_irn_visited(projHT);
1412 
1413 		projHF = new_r_Proj(irn, mode_X, pn_Cond_false);
1414 		mark_irn_visited(projHF);
1415 
1416 		newbl_eq = new_r_Block(irg, 1, &projHF);
1417 
1418 		irn = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1419 		                 ir_relation_equal);
1420 		irn = new_rd_Cond(dbg, newbl_eq, irn);
1421 
1422 		projEqF = new_r_Proj(irn, mode_X, pn_Cond_false);
1423 		mark_irn_visited(projEqF);
1424 
1425 		proj = new_r_Proj(irn, mode_X, pn_Cond_true);
1426 		mark_irn_visited(proj);
1427 
1428 		newbl_l = new_r_Block(irg, 1, &proj);
1429 
1430 		dbg = get_irn_dbg_info(sel);
1431 		irn = new_rd_Cmp(dbg, newbl_l, lentry->low_word, rentry->low_word,
1432 		                 relation);
1433 		dbg = get_irn_dbg_info(node);
1434 		irn = new_rd_Cond(dbg, newbl_l, irn);
1435 
1436 		proj = new_r_Proj(irn, mode_X, pn_Cond_true);
1437 		mark_irn_visited(proj);
1438 		add_block_cf_input(dstT, projT, proj);
1439 
1440 		proj = new_r_Proj(irn, mode_X, pn_Cond_false);
1441 		mark_irn_visited(proj);
1442 		add_block_cf_input(dstF, projF, proj);
1443 
1444 		exchange(projT, projHT);
1445 		exchange(projF, projEqF);
1446 	}
1447 
1448 	/* we have changed the control flow */
1449 	env->flags |= CF_CHANGED;
1450 }
1451 
1452 /**
1453  * Translate a Conv to higher_signed
1454  */
lower_Conv_to_Ll(ir_node * node)1455 static void lower_Conv_to_Ll(ir_node *node)
1456 {
1457 	ir_mode  *omode        = get_irn_mode(node);
1458 	ir_node  *op           = get_Conv_op(node);
1459 	ir_mode  *imode        = get_irn_mode(op);
1460 	ir_graph *irg          = get_irn_irg(node);
1461 	ir_node  *block        = get_nodes_block(node);
1462 	dbg_info *dbg          = get_irn_dbg_info(node);
1463 	ir_node  *res_low;
1464 	ir_node  *res_high;
1465 
1466 	ir_mode  *low_unsigned = env->low_unsigned;
1467 	ir_mode  *low_signed
1468 		= mode_is_signed(omode) ? env->low_signed : low_unsigned;
1469 
1470 	if (mode_is_int(imode) || mode_is_reference(imode)) {
1471 		if (imode == env->high_signed || imode == env->high_unsigned) {
1472 			/* a Conv from Lu to Ls or Ls to Lu */
1473 			const lower64_entry_t *op_entry = get_node_entry(op);
1474 			res_low  = op_entry->low_word;
1475 			res_high = new_rd_Conv(dbg, block, op_entry->high_word, low_signed);
1476 		} else {
1477 			/* simple case: create a high word */
1478 			if (imode != low_unsigned)
1479 				op = new_rd_Conv(dbg, block, op, low_unsigned);
1480 
1481 			res_low = op;
1482 
1483 			if (mode_is_signed(imode)) {
1484 				int      c       = get_mode_size_bits(low_signed) - 1;
1485 				ir_node *cnst    = new_r_Const_long(irg, low_unsigned, c);
1486 				if (get_irn_mode(op) != low_signed)
1487 					op = new_rd_Conv(dbg, block, op, low_signed);
1488 				res_high = new_rd_Shrs(dbg, block, op, cnst, low_signed);
1489 			} else {
1490 				res_high = new_r_Const(irg, get_mode_null(low_signed));
1491 			}
1492 		}
1493 	} else if (imode == mode_b) {
1494 		res_low  = new_rd_Conv(dbg, block, op, low_unsigned);
1495 		res_high = new_r_Const(irg, get_mode_null(low_signed));
1496 	} else {
1497 		ir_node *irn, *call;
1498 		ir_type *mtp = get_conv_type(imode, omode);
1499 
1500 		irn = get_intrinsic_address(mtp, get_irn_op(node), imode, omode);
1501 		call = new_rd_Call(dbg, block, get_irg_no_mem(irg), irn, 1, &op, mtp);
1502 		set_irn_pinned(call, get_irn_pinned(node));
1503 		irn = new_r_Proj(call, mode_T, pn_Call_T_result);
1504 
1505 		if (env->params->little_endian) {
1506 			res_low  = new_r_Proj(irn, low_unsigned, 0);
1507 			res_high = new_r_Proj(irn, low_signed, 1);
1508 		} else {
1509 			res_low  = new_r_Proj(irn, low_unsigned, 1);
1510 			res_high = new_r_Proj(irn, low_signed,   0);
1511 		}
1512 	}
1513 	ir_set_dw_lowered(node, res_low, res_high);
1514 }
1515 
1516 /**
1517  * Translate a Conv from higher_unsigned
1518  */
lower_Conv_from_Ll(ir_node * node)1519 static void lower_Conv_from_Ll(ir_node *node)
1520 {
1521 	ir_node               *op    = get_Conv_op(node);
1522 	ir_mode               *omode = get_irn_mode(node);
1523 	ir_node               *block = get_nodes_block(node);
1524 	dbg_info              *dbg   = get_irn_dbg_info(node);
1525 	ir_graph              *irg   = get_irn_irg(node);
1526 	const lower64_entry_t *entry = get_node_entry(op);
1527 
1528 	if (mode_is_int(omode) || mode_is_reference(omode)) {
1529 		op = entry->low_word;
1530 
1531 		/* simple case: create a high word */
1532 		if (omode != env->low_unsigned)
1533 			op = new_rd_Conv(dbg, block, op, omode);
1534 
1535 		set_Conv_op(node, op);
1536 	} else if (omode == mode_b) {
1537 		/* llu ? true : false  <=> (low|high) ? true : false */
1538 		ir_mode *mode   = env->low_unsigned;
1539 		ir_node *ornode = new_rd_Or(dbg, block, entry->low_word,
1540 		                            entry->high_word, mode);
1541 		set_Conv_op(node, ornode);
1542 	} else {
1543 		ir_node *irn, *call, *in[2];
1544 		ir_mode *imode = get_irn_mode(op);
1545 		ir_type *mtp   = get_conv_type(imode, omode);
1546 		ir_node *res;
1547 
1548 		irn   = get_intrinsic_address(mtp, get_irn_op(node), imode, omode);
1549 		if (env->params->little_endian) {
1550 			in[0] = entry->low_word;
1551 			in[1] = entry->high_word;
1552 		} else {
1553 			in[0] = entry->high_word;
1554 			in[1] = entry->low_word;
1555 		}
1556 
1557 		call = new_rd_Call(dbg, block, get_irg_no_mem(irg), irn, 2, in, mtp);
1558 		set_irn_pinned(call, get_irn_pinned(node));
1559 		irn = new_r_Proj(call, mode_T, pn_Call_T_result);
1560 		res = new_r_Proj(irn, omode, 0);
1561 
1562 		exchange(node, res);
1563 	}
1564 }
1565 
1566 /**
1567  * lower Cmp
1568  */
lower_Cmp(ir_node * cmp,ir_mode * m)1569 static void lower_Cmp(ir_node *cmp, ir_mode *m)
1570 {
1571 	ir_node  *l        = get_Cmp_left(cmp);
1572 	ir_mode  *cmp_mode = get_irn_mode(l);
1573 	ir_node  *r, *low, *high, *t, *res;
1574 	ir_relation relation;
1575 	ir_node  *block;
1576 	dbg_info *dbg;
1577 	const lower64_entry_t *lentry;
1578 	const lower64_entry_t *rentry;
1579 	(void) m;
1580 
1581 	if (cmp_mode != env->high_signed && cmp_mode != env->high_unsigned)
1582 		return;
1583 
1584 	r        = get_Cmp_right(cmp);
1585 	lentry   = get_node_entry(l);
1586 	rentry   = get_node_entry(r);
1587 	relation = get_Cmp_relation(cmp);
1588 	block    = get_nodes_block(cmp);
1589 	dbg      = get_irn_dbg_info(cmp);
1590 
1591 	/* easy case for x ==/!= 0 (see lower_Cond for details) */
1592 	if (is_equality_cmp(cmp)) {
1593 		ir_graph *irg        = get_irn_irg(cmp);
1594 		ir_mode  *mode       = env->low_unsigned;
1595 		ir_node  *low_left   = new_rd_Conv(dbg, block, lentry->low_word, mode);
1596 		ir_node  *high_left  = new_rd_Conv(dbg, block, lentry->high_word, mode);
1597 		ir_node  *low_right  = new_rd_Conv(dbg, block, rentry->low_word, mode);
1598 		ir_node  *high_right = new_rd_Conv(dbg, block, rentry->high_word, mode);
1599 		ir_node  *xor_low    = new_rd_Eor(dbg, block, low_left, low_right, mode);
1600 		ir_node  *xor_high   = new_rd_Eor(dbg, block, high_left, high_right, mode);
1601 		ir_node  *ornode     = new_rd_Or(dbg, block, xor_low, xor_high, mode);
1602 		ir_node  *new_cmp    = new_rd_Cmp(dbg, block, ornode, new_r_Const(irg, get_mode_null(mode)), relation);
1603 		exchange(cmp, new_cmp);
1604 		return;
1605 	}
1606 
1607 	if (relation == ir_relation_equal) {
1608 		/* simple case:a == b <==> a_h == b_h && a_l == b_l */
1609 		low  = new_rd_Cmp(dbg, block, lentry->low_word, rentry->low_word,
1610 		                  relation);
1611 		high = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1612 		                  relation);
1613 		res  = new_rd_And(dbg, block, low, high, mode_b);
1614 	} else if (relation == ir_relation_less_greater) {
1615 		/* simple case:a != b <==> a_h != b_h || a_l != b_l */
1616 		low  = new_rd_Cmp(dbg, block, lentry->low_word, rentry->low_word,
1617 		                  relation);
1618 		high = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1619 		                  relation);
1620 		res = new_rd_Or(dbg, block, low, high, mode_b);
1621 	} else {
1622 		/* a rel b <==> a_h REL b_h || (a_h == b_h && a_l rel b_l) */
1623 		ir_node *high1 = new_rd_Cmp(dbg, block, lentry->high_word,
1624 			rentry->high_word, relation & ~ir_relation_equal);
1625 		low  = new_rd_Cmp(dbg, block, lentry->low_word, rentry->low_word,
1626 		                  relation);
1627 		high = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1628 		                  ir_relation_equal);
1629 		t = new_rd_And(dbg, block, low, high, mode_b);
1630 		res = new_rd_Or(dbg, block, high1, t, mode_b);
1631 	}
1632 	exchange(cmp, res);
1633 }
1634 
1635 /**
1636  * Translate a Conv.
1637  */
lower_Conv(ir_node * node,ir_mode * mode)1638 static void lower_Conv(ir_node *node, ir_mode *mode)
1639 {
1640 	mode = get_irn_mode(node);
1641 
1642 	if (mode == env->high_signed || mode == env->high_unsigned) {
1643 		lower_Conv_to_Ll(node);
1644 	} else {
1645 		ir_mode *op_mode = get_irn_mode(get_Conv_op(node));
1646 
1647 		if (op_mode == env->high_signed || op_mode == env->high_unsigned) {
1648 			lower_Conv_from_Ll(node);
1649 		}
1650 	}
1651 }
1652 
fix_parameter_entities(ir_graph * irg,ir_type * orig_mtp)1653 static void fix_parameter_entities(ir_graph *irg, ir_type *orig_mtp)
1654 {
1655 	size_t      orig_n_params      = get_method_n_params(orig_mtp);
1656 	ir_entity **parameter_entities;
1657 
1658 	parameter_entities = ALLOCANZ(ir_entity*, orig_n_params);
1659 
1660 	ir_type *frame_type = get_irg_frame_type(irg);
1661 	size_t   n          = get_compound_n_members(frame_type);
1662 	size_t   i;
1663 	size_t   n_param;
1664 
1665 	/* collect parameter entities */
1666 	for (i = 0; i < n; ++i) {
1667 		ir_entity *entity = get_compound_member(frame_type, i);
1668 		size_t     p;
1669 		if (!is_parameter_entity(entity))
1670 			continue;
1671 		p = get_entity_parameter_number(entity);
1672 		if (p == IR_VA_START_PARAMETER_NUMBER)
1673 			continue;
1674 		assert(p < orig_n_params);
1675 		assert(parameter_entities[p] == NULL);
1676 		parameter_entities[p] = entity;
1677 	}
1678 
1679 	/* adjust indices */
1680 	n_param = 0;
1681 	for (i = 0; i < orig_n_params; ++i, ++n_param) {
1682 		ir_entity *entity = parameter_entities[i];
1683 		ir_type   *tp;
1684 
1685 		if (entity != NULL)
1686 			set_entity_parameter_number(entity, n_param);
1687 
1688 		tp = get_method_param_type(orig_mtp, i);
1689 		if (is_Primitive_type(tp)) {
1690 			ir_mode *mode = get_type_mode(tp);
1691 			if (mode == env->high_signed || mode == env->high_unsigned) {
1692 				++n_param;
1693 				/* note that we do not change the type of the parameter
1694 				 * entities, as calling convention fixup later still needs to
1695 				 * know which is/was a lowered doubleword.
1696 				 * So we just mark/remember it for later */
1697 				if (entity != NULL) {
1698 					assert(entity->attr.parameter.doubleword_low_mode == NULL);
1699 					entity->attr.parameter.doubleword_low_mode
1700 						= env->low_unsigned;
1701 				}
1702 			}
1703 		}
1704 	}
1705 }
1706 
1707 /**
1708  * Lower the method type.
1709  *
1710  * @param env  the lower environment
1711  * @param mtp  the method type to lower
1712  *
1713  * @return the lowered type
1714  */
lower_mtp(ir_type * mtp)1715 static ir_type *lower_mtp(ir_type *mtp)
1716 {
1717 	ir_type *res;
1718 	size_t   i;
1719 	size_t   orig_n_params;
1720 	size_t   orig_n_res;
1721 	size_t   n_param;
1722 	size_t   n_res;
1723 	bool     must_be_lowered;
1724 
1725 	res = pmap_get(ir_type, lowered_type, mtp);
1726 	if (res != NULL)
1727 		return res;
1728 	if (type_visited(mtp))
1729 		return mtp;
1730 	mark_type_visited(mtp);
1731 
1732 	orig_n_params   = get_method_n_params(mtp);
1733 	orig_n_res      = get_method_n_ress(mtp);
1734 	n_param         = orig_n_params;
1735 	n_res           = orig_n_res;
1736 	must_be_lowered = false;
1737 
1738 	/* count new number of params */
1739 	for (i = orig_n_params; i > 0;) {
1740 		ir_type *tp = get_method_param_type(mtp, --i);
1741 
1742 		if (is_Primitive_type(tp)) {
1743 			ir_mode *mode = get_type_mode(tp);
1744 
1745 			if (mode == env->high_signed || mode == env->high_unsigned) {
1746 				++n_param;
1747 				must_be_lowered = true;
1748 			}
1749 		}
1750 	}
1751 
1752 	/* count new number of results */
1753 	for (i = orig_n_res; i > 0;) {
1754 		ir_type *tp = get_method_res_type(mtp, --i);
1755 
1756 		if (is_Primitive_type(tp)) {
1757 			ir_mode *mode = get_type_mode(tp);
1758 
1759 			if (mode == env->high_signed || mode == env->high_unsigned) {
1760 				++n_res;
1761 				must_be_lowered = true;
1762 			}
1763 		}
1764 	}
1765 	if (!must_be_lowered) {
1766 		set_type_link(mtp, NULL);
1767 		return mtp;
1768 	}
1769 
1770 	res = new_d_type_method(n_param, n_res, get_type_dbg_info(mtp));
1771 
1772 	/* set param types and result types */
1773 	for (i = n_param = 0; i < orig_n_params; ++i) {
1774 		ir_type *tp = get_method_param_type(mtp, i);
1775 
1776 		if (is_Primitive_type(tp)) {
1777 			ir_mode *mode = get_type_mode(tp);
1778 
1779 			if (mode == env->high_signed) {
1780 				if (env->params->little_endian) {
1781 					set_method_param_type(res, n_param++, tp_u);
1782 					set_method_param_type(res, n_param++, tp_s);
1783 				} else {
1784 					set_method_param_type(res, n_param++, tp_s);
1785 					set_method_param_type(res, n_param++, tp_u);
1786 				}
1787 			} else if (mode == env->high_unsigned) {
1788 				set_method_param_type(res, n_param++, tp_u);
1789 				set_method_param_type(res, n_param++, tp_u);
1790 			} else {
1791 				set_method_param_type(res, n_param, tp);
1792 				++n_param;
1793 			}
1794 		} else {
1795 			set_method_param_type(res, n_param, tp);
1796 			++n_param;
1797 		}
1798 	}
1799 	for (i = n_res = 0; i < orig_n_res; ++i) {
1800 		ir_type *tp = get_method_res_type(mtp, i);
1801 
1802 		if (is_Primitive_type(tp)) {
1803 			ir_mode *mode = get_type_mode(tp);
1804 
1805 			if (mode == env->high_signed) {
1806 				if (env->params->little_endian) {
1807 					set_method_res_type(res, n_res++, tp_u);
1808 					set_method_res_type(res, n_res++, tp_s);
1809 				} else {
1810 					set_method_res_type(res, n_res++, tp_s);
1811 					set_method_res_type(res, n_res++, tp_u);
1812 				}
1813 			} else if (mode == env->high_unsigned) {
1814 				set_method_res_type(res, n_res++, tp_u);
1815 				set_method_res_type(res, n_res++, tp_u);
1816 			} else {
1817 				set_method_res_type(res, n_res++, tp);
1818 			}
1819 		} else {
1820 			set_method_res_type(res, n_res++, tp);
1821 		}
1822 	}
1823 
1824 	set_method_variadicity(res, get_method_variadicity(mtp));
1825 	set_method_calling_convention(res, get_method_calling_convention(mtp));
1826 	set_method_additional_properties(res, get_method_additional_properties(mtp));
1827 
1828 	set_higher_type(res, mtp);
1829 	set_type_link(res, mtp);
1830 
1831 	mark_type_visited(res);
1832 	pmap_insert(lowered_type, mtp, res);
1833 	return res;
1834 }
1835 
1836 /**
1837  * Translate a Return.
1838  */
lower_Return(ir_node * node,ir_mode * mode)1839 static void lower_Return(ir_node *node, ir_mode *mode)
1840 {
1841 	ir_node  **in;
1842 	size_t     i, j, n;
1843 	int        need_conv = 0;
1844 	(void) mode;
1845 
1846 	/* check if this return must be lowered */
1847 	for (i = 0, n = get_Return_n_ress(node); i < n; ++i) {
1848 		ir_node *pred  = get_Return_res(node, i);
1849 		ir_mode *rmode = get_irn_op_mode(pred);
1850 
1851 		if (rmode == env->high_signed || rmode == env->high_unsigned)
1852 			need_conv = 1;
1853 	}
1854 	if (! need_conv)
1855 		return;
1856 
1857 	ir_graph  *irg = get_irn_irg(node);
1858 	ir_entity *ent = get_irg_entity(irg);
1859 	ir_type   *mtp = get_entity_type(ent);
1860 
1861 	/* create a new in array */
1862 	NEW_ARR_A(ir_node *, in, get_method_n_ress(mtp) + 1);
1863 	j = 0;
1864 	in[j++] = get_Return_mem(node);
1865 
1866 	for (i = 0, n = get_Return_n_ress(node); i < n; ++i) {
1867 		ir_node *pred      = get_Return_res(node, i);
1868 		ir_mode *pred_mode = get_irn_mode(pred);
1869 
1870 		if (pred_mode == env->high_signed || pred_mode == env->high_unsigned) {
1871 			const lower64_entry_t *entry = get_node_entry(pred);
1872 			if (env->params->little_endian) {
1873 				in[j++] = entry->low_word;
1874 				in[j++] = entry->high_word;
1875 			} else {
1876 				in[j++] = entry->high_word;
1877 				in[j++] = entry->low_word;
1878 			}
1879 		} else {
1880 			in[j++] = pred;
1881 		}
1882 	}
1883 	assert(j == get_method_n_ress(mtp)+1);
1884 
1885 	set_irn_in(node, j, in);
1886 }
1887 
1888 /**
1889  * Translate the parameters.
1890  */
lower_Start(ir_node * node,ir_mode * high_mode)1891 static void lower_Start(ir_node *node, ir_mode *high_mode)
1892 {
1893 	ir_graph  *irg      = get_irn_irg(node);
1894 	ir_entity *ent      = get_irg_entity(irg);
1895 	ir_type   *mtp      = get_entity_type(ent);
1896 	ir_type   *orig_mtp = (ir_type*)get_type_link(mtp);
1897 	ir_node   *args;
1898 	long      *new_projs;
1899 	size_t    i, j, n_params;
1900 	(void) high_mode;
1901 
1902 	/* if type link is NULL then the type was not lowered, hence no changes
1903 	 * at Start necessary */
1904 	if (orig_mtp == NULL)
1905 		return;
1906 
1907 	n_params = get_method_n_params(orig_mtp);
1908 
1909 	NEW_ARR_A(long, new_projs, n_params);
1910 
1911 	/* Calculate mapping of proj numbers in new_projs */
1912 	for (i = j = 0; i < n_params; ++i, ++j) {
1913 		ir_type *ptp = get_method_param_type(orig_mtp, i);
1914 
1915 		new_projs[i] = j;
1916 		if (is_Primitive_type(ptp)) {
1917 			ir_mode *amode = get_type_mode(ptp);
1918 			if (amode == env->high_signed || amode == env->high_unsigned)
1919 				++j;
1920 		}
1921 	}
1922 
1923 	/* find args Proj */
1924 	args = NULL;
1925 	foreach_out_edge(node, edge) {
1926 		ir_node *proj = get_edge_src_irn(edge);
1927 		if (!is_Proj(proj))
1928 			continue;
1929 		if (get_Proj_proj(proj) == pn_Start_T_args) {
1930 			args = proj;
1931 			break;
1932 		}
1933 	}
1934 	if (args == NULL)
1935 		return;
1936 
1937 	/* fix all Proj's and create new ones */
1938 	foreach_out_edge_safe(args, edge) {
1939 		ir_node *proj   = get_edge_src_irn(edge);
1940 		ir_mode *mode   = get_irn_mode(proj);
1941 		ir_mode *mode_l = env->low_unsigned;
1942 		ir_node *pred;
1943 		long     proj_nr;
1944 		ir_mode *mode_h;
1945 		ir_node *res_low;
1946 		ir_node *res_high;
1947 		int      old_cse;
1948 		dbg_info *dbg;
1949 
1950 		if (!is_Proj(proj))
1951 			continue;
1952 		pred    = get_Proj_pred(proj);
1953 		proj_nr = get_Proj_proj(proj);
1954 
1955 		if (mode == env->high_signed) {
1956 			mode_h = env->low_signed;
1957 		} else if (mode == env->high_unsigned) {
1958 			mode_h = env->low_unsigned;
1959 		} else {
1960 			long new_pn = new_projs[proj_nr];
1961 			set_Proj_proj(proj, new_pn);
1962 			continue;
1963 		}
1964 
1965 		/* Switch off CSE or we might get an already existing Proj. */
1966 		old_cse = get_opt_cse();
1967 		set_opt_cse(0);
1968 		dbg = get_irn_dbg_info(proj);
1969 		if (env->params->little_endian) {
1970 			res_low  = new_rd_Proj(dbg, pred, mode_l, new_projs[proj_nr]);
1971 			res_high = new_rd_Proj(dbg, pred, mode_h, new_projs[proj_nr] + 1);
1972 		} else {
1973 			res_high = new_rd_Proj(dbg, pred, mode_h, new_projs[proj_nr]);
1974 			res_low  = new_rd_Proj(dbg, pred, mode_l, new_projs[proj_nr] + 1);
1975 		}
1976 		set_opt_cse(old_cse);
1977 		ir_set_dw_lowered(proj, res_low, res_high);
1978 	}
1979 }
1980 
1981 /**
1982  * Translate a Call.
1983  */
lower_Call(ir_node * node,ir_mode * mode)1984 static void lower_Call(ir_node *node, ir_mode *mode)
1985 {
1986 	ir_type  *tp = get_Call_type(node);
1987 	ir_node  **in;
1988 	size_t   n_params, n_res;
1989 	bool     need_lower = false;
1990 	size_t   i, j;
1991 	size_t   p;
1992 	long     *res_numbers = NULL;
1993 	ir_node  *resproj;
1994 	(void) mode;
1995 
1996 	n_params = get_method_n_params(tp);
1997 	for (p = 0; p < n_params; ++p) {
1998 		ir_type *ptp = get_method_param_type(tp, p);
1999 
2000 		if (is_Primitive_type(ptp)) {
2001 			ir_mode *pmode = get_type_mode(ptp);
2002 			if (pmode == env->high_signed || pmode == env->high_unsigned) {
2003 				need_lower = true;
2004 				break;
2005 			}
2006 		}
2007 	}
2008 	n_res = get_method_n_ress(tp);
2009 	if (n_res > 0) {
2010 		NEW_ARR_A(long, res_numbers, n_res);
2011 
2012 		for (i = j = 0; i < n_res; ++i, ++j) {
2013 			ir_type *ptp = get_method_res_type(tp, i);
2014 
2015 			res_numbers[i] = j;
2016 			if (is_Primitive_type(ptp)) {
2017 				ir_mode *rmode = get_type_mode(ptp);
2018 				if (rmode == env->high_signed || rmode == env->high_unsigned) {
2019 					need_lower = true;
2020 					++j;
2021 				}
2022 			}
2023 		}
2024 	}
2025 
2026 	if (! need_lower)
2027 		return;
2028 
2029 	/* let's lower it */
2030 	tp = lower_mtp(tp);
2031 	set_Call_type(node, tp);
2032 
2033 	NEW_ARR_A(ir_node *, in, get_method_n_params(tp) + 2);
2034 
2035 	in[0] = get_Call_mem(node);
2036 	in[1] = get_Call_ptr(node);
2037 
2038 	for (j = 2, i = 0; i < n_params; ++i) {
2039 		ir_node *pred      = get_Call_param(node, i);
2040 		ir_mode *pred_mode = get_irn_mode(pred);
2041 
2042 		if (pred_mode == env->high_signed || pred_mode == env->high_unsigned) {
2043 			const lower64_entry_t *pred_entry = get_node_entry(pred);
2044 			if (env->params->little_endian) {
2045 				in[j++] = pred_entry->low_word;
2046 				in[j++] = pred_entry->high_word;
2047 			} else {
2048 				in[j++] = pred_entry->high_word;
2049 				in[j++] = pred_entry->low_word;
2050 			}
2051 		} else {
2052 			in[j++] = pred;
2053 		}
2054 	}
2055 
2056 	set_irn_in(node, j, in);
2057 
2058 	/* find results T */
2059 	resproj = NULL;
2060 	foreach_out_edge(node, edge) {
2061 		ir_node *proj = get_edge_src_irn(edge);
2062 		if (!is_Proj(proj))
2063 			continue;
2064 		if (get_Proj_proj(proj) == pn_Call_T_result) {
2065 			resproj = proj;
2066 			break;
2067 		}
2068 	}
2069 	if (resproj == NULL)
2070 		return;
2071 
2072 	/* fix the results */
2073 	foreach_out_edge_safe(resproj, edge) {
2074 		ir_node  *proj      = get_edge_src_irn(edge);
2075 		ir_mode  *proj_mode = get_irn_mode(proj);
2076 		ir_mode  *mode_l    = env->low_unsigned;
2077 		ir_node  *pred;
2078 		long      proj_nr;
2079 		ir_mode  *mode_h;
2080 		ir_node  *res_low;
2081 		ir_node  *res_high;
2082 		dbg_info *dbg;
2083 
2084 		if (!is_Proj(proj))
2085 			continue;
2086 		pred    = get_Proj_pred(proj);
2087 		proj_nr = get_Proj_proj(proj);
2088 
2089 		if (proj_mode == env->high_signed) {
2090 			mode_h = env->low_signed;
2091 		} else if (proj_mode == env->high_unsigned) {
2092 			mode_h = env->low_unsigned;
2093 		} else {
2094 			long new_nr = res_numbers[proj_nr];
2095 			set_Proj_proj(proj, new_nr);
2096 			continue;
2097 		}
2098 
2099 		dbg = get_irn_dbg_info(proj);
2100 		if (env->params->little_endian) {
2101 			res_low  = new_rd_Proj(dbg, pred, mode_l, res_numbers[proj_nr]);
2102 			res_high = new_rd_Proj(dbg, pred, mode_h, res_numbers[proj_nr] + 1);
2103 		} else {
2104 			res_high = new_rd_Proj(dbg, pred, mode_h, res_numbers[proj_nr]);
2105 			res_low  = new_rd_Proj(dbg, pred, mode_l, res_numbers[proj_nr] + 1);
2106 		}
2107 		ir_set_dw_lowered(proj, res_low, res_high);
2108 	}
2109 }
2110 
2111 /**
2112  * Translate an Unknown into two.
2113  */
lower_Unknown(ir_node * node,ir_mode * mode)2114 static void lower_Unknown(ir_node *node, ir_mode *mode)
2115 {
2116 	ir_mode  *low_mode = env->low_unsigned;
2117 	ir_graph *irg      = get_irn_irg(node);
2118 	ir_node  *res_low  = new_r_Unknown(irg, low_mode);
2119 	ir_node  *res_high = new_r_Unknown(irg, mode);
2120 	ir_set_dw_lowered(node, res_low, res_high);
2121 }
2122 
2123 /**
2124  * Translate a Bad into two.
2125  */
lower_Bad(ir_node * node,ir_mode * mode)2126 static void lower_Bad(ir_node *node, ir_mode *mode)
2127 {
2128 	ir_mode  *low_mode = env->low_unsigned;
2129 	ir_graph *irg      = get_irn_irg(node);
2130 	ir_node  *res_low  = new_r_Bad(irg, low_mode);
2131 	ir_node  *res_high = new_r_Bad(irg, mode);
2132 	ir_set_dw_lowered(node, res_low, res_high);
2133 }
2134 
2135 /**
2136  * Translate a Phi.
2137  *
2138  * First step: just create two templates
2139  */
lower_Phi(ir_node * phi)2140 static void lower_Phi(ir_node *phi)
2141 {
2142 	ir_mode  *mode = get_irn_mode(phi);
2143 	int       i;
2144 	int       arity;
2145 	ir_node **in_l;
2146 	ir_node **in_h;
2147 	ir_node  *unk_l;
2148 	ir_node  *unk_h;
2149 	ir_node  *phi_l;
2150 	ir_node  *phi_h;
2151 	dbg_info *dbg;
2152 	ir_node  *block;
2153 	ir_graph *irg;
2154 	ir_mode  *mode_l;
2155 	ir_mode  *mode_h;
2156 
2157 	/* enqueue predecessors */
2158 	arity = get_Phi_n_preds(phi);
2159 	for (i = 0; i < arity; ++i) {
2160 		ir_node *pred = get_Phi_pred(phi, i);
2161 		pdeq_putr(env->waitq, pred);
2162 	}
2163 
2164 	if (mode != env->high_signed && mode != env->high_unsigned)
2165 		return;
2166 
2167 	/* first create a new in array */
2168 	NEW_ARR_A(ir_node *, in_l, arity);
2169 	NEW_ARR_A(ir_node *, in_h, arity);
2170 	irg    = get_irn_irg(phi);
2171 	mode_l = env->low_unsigned;
2172 	mode_h = mode == env->high_signed ? env->low_signed : env->low_unsigned;
2173 	unk_l  = new_r_Dummy(irg, mode_l);
2174 	unk_h  = new_r_Dummy(irg, mode_h);
2175 	for (i = 0; i < arity; ++i) {
2176 		in_l[i] = unk_l;
2177 		in_h[i] = unk_h;
2178 	}
2179 
2180 	dbg   = get_irn_dbg_info(phi);
2181 	block = get_nodes_block(phi);
2182 	phi_l = new_rd_Phi(dbg, block, arity, in_l, mode_l);
2183 	phi_h = new_rd_Phi(dbg, block, arity, in_h, mode_h);
2184 
2185 	ir_set_dw_lowered(phi, phi_l, phi_h);
2186 
2187 	/* remember that we need to fixup the predecessors later */
2188 	ARR_APP1(ir_node*, env->lowered_phis, phi);
2189 }
2190 
fixup_phi(ir_node * phi)2191 static void fixup_phi(ir_node *phi)
2192 {
2193 	const lower64_entry_t *entry = get_node_entry(phi);
2194 	ir_node               *phi_l = entry->low_word;
2195 	ir_node               *phi_h = entry->high_word;
2196 	int                    arity = get_Phi_n_preds(phi);
2197 	int                    i;
2198 
2199 	/* exchange phi predecessors which are lowered by now */
2200 	for (i = 0; i < arity; ++i) {
2201 		ir_node               *pred       = get_Phi_pred(phi, i);
2202 		const lower64_entry_t *pred_entry = get_node_entry(pred);
2203 
2204 		set_Phi_pred(phi_l, i, pred_entry->low_word);
2205 		set_Phi_pred(phi_h, i, pred_entry->high_word);
2206 	}
2207 }
2208 
2209 /**
2210  * Translate a Mux.
2211  */
lower_Mux(ir_node * mux,ir_mode * mode)2212 static void lower_Mux(ir_node *mux, ir_mode *mode)
2213 {
2214 	ir_node               *truen       = get_Mux_true(mux);
2215 	ir_node               *falsen      = get_Mux_false(mux);
2216 	ir_node               *sel         = get_Mux_sel(mux);
2217 	const lower64_entry_t *true_entry  = get_node_entry(truen);
2218 	const lower64_entry_t *false_entry = get_node_entry(falsen);
2219 	ir_node               *true_l      = true_entry->low_word;
2220 	ir_node               *true_h      = true_entry->high_word;
2221 	ir_node               *false_l     = false_entry->low_word;
2222 	ir_node               *false_h     = false_entry->high_word;
2223 	dbg_info              *dbgi        = get_irn_dbg_info(mux);
2224 	ir_node               *block       = get_nodes_block(mux);
2225 	ir_node               *res_low
2226 		= new_rd_Mux(dbgi, block, sel, false_l, true_l, env->low_unsigned);
2227 	ir_node               *res_high
2228 		= new_rd_Mux(dbgi, block, sel, false_h, true_h, mode);
2229 	ir_set_dw_lowered(mux, res_low, res_high);
2230 }
2231 
2232 /**
2233  * Translate an ASM node.
2234  */
lower_ASM(ir_node * asmn,ir_mode * mode)2235 static void lower_ASM(ir_node *asmn, ir_mode *mode)
2236 {
2237 	ir_mode           *high_signed        = env->high_signed;
2238 	ir_mode           *high_unsigned      = env->high_unsigned;
2239 	int                n_outs             = get_ASM_n_output_constraints(asmn);
2240 	ir_asm_constraint *output_constraints = get_ASM_output_constraints(asmn);
2241 	ir_asm_constraint *input_constraints  = get_ASM_input_constraints(asmn);
2242 	unsigned           n_64bit_outs       = 0;
2243 
2244 	(void)mode;
2245 
2246 	for (int i = get_irn_arity(asmn) - 1; i >= 0; --i) {
2247 		ir_node *op      = get_irn_n(asmn, i);
2248 		ir_mode *op_mode = get_irn_mode(op);
2249 		if (op_mode == high_signed || op_mode == high_unsigned) {
2250 			panic("lowering ASM 64bit input unimplemented");
2251 		}
2252 	}
2253 
2254 	for (int o = 0; o < n_outs; ++o) {
2255 		const ir_asm_constraint *constraint = &output_constraints[o];
2256 		if (constraint->mode == high_signed || constraint->mode == high_unsigned) {
2257 			const char *constr = get_id_str(constraint->constraint);
2258 			++n_64bit_outs;
2259 			/* TODO: How to do this architecture neutral? This is very
2260 			 * i386 specific... */
2261 			if (constr[0] != '=' || constr[1] != 'A') {
2262 				panic("lowering ASM 64bit output only supports '=A' currently");
2263 			}
2264 		}
2265 	}
2266 
2267 	if (n_64bit_outs == 0)
2268 		return;
2269 
2270 	dbg_info          *dbgi       = get_irn_dbg_info(asmn);
2271 	ir_node           *block      = get_nodes_block(asmn);
2272 	ir_node           *mem        = get_ASM_mem(asmn);
2273 	int                new_n_outs = 0;
2274 	int                n_clobber  = get_ASM_n_clobbers(asmn);
2275 	long              *proj_map   = ALLOCAN(long, n_outs);
2276 	ident            **clobbers   = get_ASM_clobbers(asmn);
2277 	ident             *asm_text   = get_ASM_text(asmn);
2278 	ir_asm_constraint *new_outputs
2279 		= ALLOCAN(ir_asm_constraint, n_outs+n_64bit_outs);
2280 	ir_node           *new_asm;
2281 
2282 	for (int o = 0; o < n_outs; ++o) {
2283 		const ir_asm_constraint *constraint = &output_constraints[o];
2284 		if (constraint->mode == high_signed || constraint->mode == high_unsigned) {
2285 			new_outputs[new_n_outs].pos        = constraint->pos;
2286 			new_outputs[new_n_outs].constraint = new_id_from_str("=a");
2287 			new_outputs[new_n_outs].mode       = env->low_unsigned;
2288 			proj_map[o] = new_n_outs;
2289 			++new_n_outs;
2290 			new_outputs[new_n_outs].pos        = constraint->pos;
2291 			new_outputs[new_n_outs].constraint = new_id_from_str("=d");
2292 			if (constraint->mode == high_signed)
2293 				new_outputs[new_n_outs].mode = env->low_signed;
2294 			else
2295 				new_outputs[new_n_outs].mode = env->low_unsigned;
2296 			++new_n_outs;
2297 		} else {
2298 			new_outputs[new_n_outs] = *constraint;
2299 			proj_map[o] = new_n_outs;
2300 			++new_n_outs;
2301 		}
2302 	}
2303 	assert(new_n_outs == n_outs+(int)n_64bit_outs);
2304 
2305 	int       n_inputs = get_ASM_n_inputs(asmn);
2306 	ir_node **new_ins  = ALLOCAN(ir_node*, n_inputs);
2307 	for (int i = 0; i < n_inputs; ++i)
2308 		new_ins[i] = get_ASM_input(asmn, i);
2309 
2310 	new_asm = new_rd_ASM(dbgi, block, mem, n_inputs, new_ins, input_constraints,
2311 						 new_n_outs, new_outputs, n_clobber, clobbers,
2312 						 asm_text);
2313 
2314 	foreach_out_edge_safe(asmn, edge) {
2315 		ir_node *proj      = get_edge_src_irn(edge);
2316 		ir_mode *proj_mode = get_irn_mode(proj);
2317 		long     pn;
2318 
2319 		if (!is_Proj(proj))
2320 			continue;
2321 		pn = get_Proj_proj(proj);
2322 
2323 		if (pn < n_outs)
2324 			pn = proj_map[pn];
2325 		else
2326 			pn = new_n_outs + pn - n_outs;
2327 
2328 		if (proj_mode == high_signed || proj_mode == high_unsigned) {
2329 			ir_mode *high_mode
2330 				= proj_mode == high_signed ? env->low_signed : env->low_unsigned;
2331 			ir_node *np_low  = new_r_Proj(new_asm, env->low_unsigned, pn);
2332 			ir_node *np_high = new_r_Proj(new_asm, high_mode, pn+1);
2333 			ir_set_dw_lowered(proj, np_low, np_high);
2334 		} else {
2335 			ir_node *np = new_r_Proj(new_asm, proj_mode, pn);
2336 			exchange(proj, np);
2337 		}
2338 	}
2339 }
2340 
2341 /**
2342  * Lower the builtin type to its higher part.
2343  *
2344  * @param mtp  the builtin type to lower
2345  *
2346  * @return the lowered type
2347  */
lower_Builtin_type_high(ir_type * mtp)2348 static ir_type *lower_Builtin_type_high(ir_type *mtp)
2349 {
2350 	ir_type *res;
2351 	size_t   i;
2352 	size_t   n_params;
2353 	size_t   n_results;
2354 	bool     must_be_lowered;
2355 
2356 	res = pmap_get(ir_type, lowered_builtin_type_high, mtp);
2357 	if (res != NULL)
2358 		return res;
2359 
2360 	n_params        = get_method_n_params(mtp);
2361 	n_results       = get_method_n_ress(mtp);
2362 	must_be_lowered = false;
2363 
2364 	/* check for double word parameter */
2365 	for (i = n_params; i > 0;) {
2366 		ir_type *tp = get_method_param_type(mtp, --i);
2367 
2368 		if (is_Primitive_type(tp)) {
2369 			ir_mode *mode = get_type_mode(tp);
2370 
2371 			if (mode == env->high_signed || mode == env->high_unsigned) {
2372 				must_be_lowered = true;
2373 				break;
2374 			}
2375 		}
2376 	}
2377 
2378 	if (!must_be_lowered) {
2379 		set_type_link(mtp, NULL);
2380 		return mtp;
2381 	}
2382 
2383 	res = new_d_type_method(n_params, n_results, get_type_dbg_info(mtp));
2384 
2385 	/* set param types and result types */
2386 	for (i = 0; i < n_params; ++i) {
2387 		ir_type *tp = get_method_param_type(mtp, i);
2388 
2389 		if (is_Primitive_type(tp)) {
2390 			ir_mode *mode = get_type_mode(tp);
2391 
2392 			if (mode == env->high_signed) {
2393 				if (env->params->little_endian) {
2394 					set_method_param_type(res, i, tp_u);
2395 				} else {
2396 					set_method_param_type(res, i, tp_s);
2397 				}
2398 			} else if (mode == env->high_unsigned) {
2399 				set_method_param_type(res, i, tp_u);
2400 			} else {
2401 				set_method_param_type(res, i, tp);
2402 			}
2403 		} else {
2404 			set_method_param_type(res, i, tp);
2405 		}
2406 	}
2407 	for (i = n_results = 0; i < n_results; ++i) {
2408 		ir_type *tp = get_method_res_type(mtp, i);
2409 
2410 		if (is_Primitive_type(tp)) {
2411 			ir_mode *mode = get_type_mode(tp);
2412 
2413 			if (mode == env->high_signed) {
2414 				if (env->params->little_endian) {
2415 					set_method_res_type(res, i, tp_u);
2416 				} else {
2417 					set_method_res_type(res, i, tp_s);
2418 				}
2419 			} else if (mode == env->high_unsigned) {
2420 				set_method_res_type(res, i, tp_u);
2421 			} else {
2422 				set_method_res_type(res, i, tp);
2423 			}
2424 		} else {
2425 			set_method_res_type(res, i, tp);
2426 		}
2427 	}
2428 
2429 	set_method_variadicity(res, get_method_variadicity(mtp));
2430 	set_method_calling_convention(res, get_method_calling_convention(mtp));
2431 	set_method_additional_properties(res, get_method_additional_properties(mtp));
2432 
2433 	pmap_insert(lowered_builtin_type_high, mtp, res);
2434 	return res;
2435 }
2436 
2437 /**
2438  * Lower the builtin type to its lower part.
2439  *
2440  * @param mtp  the builtin type to lower
2441  *
2442  * @return the lowered type
2443  */
lower_Builtin_type_low(ir_type * mtp)2444 static ir_type *lower_Builtin_type_low(ir_type *mtp)
2445 {
2446 	ir_type *res;
2447 	size_t   i;
2448 	size_t   n_params;
2449 	size_t   n_results;
2450 	bool     must_be_lowered;
2451 
2452 	res = pmap_get(ir_type, lowered_builtin_type_low, mtp);
2453 	if (res != NULL)
2454 		return res;
2455 
2456 	n_params        = get_method_n_params(mtp);
2457 	n_results       = get_method_n_ress(mtp);
2458 	must_be_lowered = false;
2459 
2460 	/* check for double word parameter */
2461 	for (i = n_params; i > 0;) {
2462 		ir_type *tp = get_method_param_type(mtp, --i);
2463 
2464 		if (is_Primitive_type(tp)) {
2465 			ir_mode *mode = get_type_mode(tp);
2466 
2467 			if (mode == env->high_signed || mode == env->high_unsigned) {
2468 				must_be_lowered = true;
2469 				break;
2470 			}
2471 		}
2472 	}
2473 
2474 	if (!must_be_lowered) {
2475 		set_type_link(mtp, NULL);
2476 		return mtp;
2477 	}
2478 
2479 	res = new_d_type_method(n_params, n_results, get_type_dbg_info(mtp));
2480 
2481 	/* set param types and result types */
2482 	for (i = 0; i < n_params; ++i) {
2483 		ir_type *tp = get_method_param_type(mtp, i);
2484 
2485 		if (is_Primitive_type(tp)) {
2486 			ir_mode *mode = get_type_mode(tp);
2487 
2488 			if (mode == env->high_signed) {
2489 				if (env->params->little_endian) {
2490 					set_method_param_type(res, i, tp_s);
2491 				} else {
2492 					set_method_param_type(res, i, tp_u);
2493 				}
2494 			} else if (mode == env->high_unsigned) {
2495 				set_method_param_type(res, i, tp_u);
2496 			} else {
2497 				set_method_param_type(res, i, tp);
2498 			}
2499 		} else {
2500 			set_method_param_type(res, i, tp);
2501 		}
2502 	}
2503 	for (i = 0; i < n_results; ++i) {
2504 		ir_type *tp = get_method_res_type(mtp, i);
2505 
2506 		if (is_Primitive_type(tp)) {
2507 			ir_mode *mode = get_type_mode(tp);
2508 
2509 			if (mode == env->high_signed) {
2510 				if (env->params->little_endian) {
2511 					set_method_res_type(res, i, tp_s);
2512 				} else {
2513 					set_method_res_type(res, i, tp_u);
2514 				}
2515 			} else if (mode == env->high_unsigned) {
2516 				set_method_res_type(res, i, tp_u);
2517 			} else {
2518 				set_method_res_type(res, i, tp);
2519 			}
2520 		} else {
2521 			set_method_res_type(res, i, tp);
2522 		}
2523 	}
2524 
2525 	set_method_variadicity(res, get_method_variadicity(mtp));
2526 	set_method_calling_convention(res, get_method_calling_convention(mtp));
2527 	set_method_additional_properties(res, get_method_additional_properties(mtp));
2528 
2529 	pmap_insert(lowered_builtin_type_low, mtp, res);
2530 	return res;
2531 }
2532 
2533 /**
2534  * lowers a builtin which reduces a 64bit value to a simple summary value
2535  * (popcount, ffs, ...)
2536  */
lower_reduce_builtin(ir_node * builtin,ir_mode * mode)2537 static void lower_reduce_builtin(ir_node *builtin, ir_mode *mode)
2538 {
2539 	ir_builtin_kind  kind         = get_Builtin_kind(builtin);
2540 	ir_node         *operand      = get_Builtin_param(builtin, 0);
2541 	ir_mode         *operand_mode = get_irn_mode(operand);
2542 	if (operand_mode != env->high_signed && operand_mode != env->high_unsigned)
2543 		return;
2544 
2545 	{
2546 	arch_allow_ifconv_func  allow_ifconv      = be_get_backend_param()->allow_ifconv;
2547 	int                     arity             = get_irn_arity(builtin);
2548 	dbg_info               *dbgi              = get_irn_dbg_info(builtin);
2549 	ir_graph               *irg               = get_irn_irg(builtin);
2550 	ir_type                *type              = get_Builtin_type(builtin);
2551 	ir_type                *lowered_type_high = lower_Builtin_type_high(type);
2552 	ir_type                *lowered_type_low  = lower_Builtin_type_low(type);
2553 	ir_type                *result_type       = get_method_res_type(lowered_type_low, 0);
2554 	ir_mode                *result_mode       = get_type_mode(result_type);
2555 	ir_node                *block             = get_nodes_block(builtin);
2556 	ir_node                *mem               = get_Builtin_mem(builtin);
2557 	const lower64_entry_t  *entry             = get_node_entry(operand);
2558 	ir_mode                *high_mode         = get_irn_mode(entry->high_word);
2559 	ir_node                *in_high[1]        = {entry->high_word};
2560 	ir_node                *in_low[1]         = {entry->low_word};
2561 	ir_node                *res;
2562 
2563 	assert(is_NoMem(mem));
2564 	assert(arity == 2);
2565 
2566 	switch (kind) {
2567 	case ir_bk_ffs: {
2568 		ir_node               *number_of_bits = new_r_Const_long(irg, result_mode, get_mode_size_bits(env->low_unsigned));
2569 		ir_node               *zero_high      = new_rd_Const(dbgi, irg, get_mode_null(high_mode));
2570 		ir_node               *zero_unsigned  = new_rd_Const(dbgi, irg, get_mode_null(env->low_unsigned));
2571 		ir_node               *zero_result    = new_rd_Const(dbgi, irg, get_mode_null(result_mode));
2572 		ir_node               *cmp_low        = new_rd_Cmp(dbgi, block, entry->low_word, zero_unsigned, ir_relation_equal);
2573 		ir_node               *cmp_high       = new_rd_Cmp(dbgi, block, entry->high_word, zero_high, ir_relation_equal);
2574 		ir_node               *ffs_high       = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2575 		ir_node               *high_proj      = new_r_Proj(ffs_high, result_mode, pn_Builtin_max+1);
2576 		ir_node               *high           = new_rd_Add(dbgi, block, high_proj, number_of_bits, result_mode);
2577 		ir_node               *ffs_low        = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2578 		ir_node               *low            = new_r_Proj(ffs_low, result_mode, pn_Builtin_max+1);
2579 		ir_node               *mux_high       = new_rd_Mux(dbgi, block, cmp_high, high, zero_result, result_mode);
2580 
2581 		if (! allow_ifconv(cmp_high, high, zero_result))
2582 			ir_nodeset_insert(&created_mux_nodes, mux_high);
2583 
2584 		res = new_rd_Mux(dbgi, block, cmp_low, low, mux_high, result_mode);
2585 
2586 		if (! allow_ifconv(cmp_low, low, mux_high))
2587 			ir_nodeset_insert(&created_mux_nodes, res);
2588 		break;
2589 	}
2590 	case ir_bk_clz: {
2591 		ir_node               *zero           = new_rd_Const(dbgi, irg, get_mode_null(high_mode));
2592 		ir_node               *cmp_high       = new_rd_Cmp(dbgi, block, entry->high_word, zero, ir_relation_equal);
2593 		ir_node               *clz_high       = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2594 		ir_node               *high           = new_r_Proj(clz_high, result_mode, pn_Builtin_max+1);
2595 		ir_node               *clz_low        = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2596 		ir_node               *low_proj       = new_r_Proj(clz_low, result_mode, pn_Builtin_max+1);
2597 		ir_node               *number_of_bits = new_r_Const_long(irg, result_mode, get_mode_size_bits(mode));
2598 		ir_node               *low            = new_rd_Add(dbgi, block, low_proj, number_of_bits, result_mode);
2599 
2600 		res = new_rd_Mux(dbgi, block, cmp_high, high, low, result_mode);
2601 
2602 		if (! allow_ifconv(cmp_high, high, low))
2603 			ir_nodeset_insert(&created_mux_nodes, res);
2604 		break;
2605 	}
2606 	case ir_bk_ctz: {
2607 		ir_node               *zero_unsigned  = new_rd_Const(dbgi, irg, get_mode_null(env->low_unsigned));
2608 		ir_node               *cmp_low        = new_rd_Cmp(dbgi, block, entry->low_word, zero_unsigned, ir_relation_equal);
2609 		ir_node               *ffs_high       = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2610 		ir_node               *high_proj      = new_r_Proj(ffs_high, result_mode, pn_Builtin_max+1);
2611 		ir_node               *number_of_bits = new_r_Const_long(irg, result_mode, get_mode_size_bits(env->low_unsigned));
2612 		ir_node               *high           = new_rd_Add(dbgi, block, high_proj, number_of_bits, result_mode);
2613 		ir_node               *ffs_low        = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2614 		ir_node               *low            = new_r_Proj(ffs_low, result_mode, pn_Builtin_max+1);
2615 
2616 		res = new_rd_Mux(dbgi, block, cmp_low, low, high, result_mode);
2617 
2618 		if (! allow_ifconv(cmp_low, low, high))
2619 			ir_nodeset_insert(&created_mux_nodes, res);
2620 		break;
2621 	}
2622 	case ir_bk_popcount: {
2623 		ir_node               *popcount_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2624 		ir_node               *popcount_low  = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2625 		ir_node               *high          = new_r_Proj(popcount_high, result_mode, pn_Builtin_max+1);
2626 		ir_node               *low           = new_r_Proj(popcount_low, result_mode, pn_Builtin_max+1);
2627 
2628 		res = new_rd_Add(dbgi, block, high, low, result_mode);
2629 		break;
2630 	}
2631 	case ir_bk_parity: {
2632 		ir_node  *parity_high;
2633 		ir_node  *parity_low;
2634 		ir_node  *high;
2635 		ir_node  *low;
2636 
2637 		assert(arity == 2);
2638 
2639 		parity_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2640 		high        = new_r_Proj(parity_high, result_mode, pn_Builtin_max+1);
2641 		parity_low  = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2642 		low         = new_r_Proj(parity_low, result_mode, pn_Builtin_max+1);
2643 		res         = new_rd_Eor(dbgi, block, high, low, result_mode);
2644 		break;
2645 	}
2646 	default:
2647 		panic("unexpected builtin");
2648 	}
2649 
2650 	turn_into_tuple(builtin, 2);
2651 	set_irn_n(builtin, pn_Builtin_M, mem);
2652 	set_irn_n(builtin, pn_Builtin_max+1, res);
2653 	}
2654 }
2655 
2656 /**
2657  * lowers builtins performing arithmetic (bswap)
2658  */
lower_arithmetic_builtin(ir_node * builtin,ir_mode * mode)2659 static void lower_arithmetic_builtin(ir_node *builtin, ir_mode *mode)
2660 {
2661 	ir_builtin_kind  kind         = get_Builtin_kind(builtin);
2662 	ir_node         *operand      = get_Builtin_param(builtin, 0);
2663 	ir_mode         *operand_mode = get_irn_mode(operand);
2664 	(void) mode;
2665 	if (operand_mode != env->high_signed && operand_mode != env->high_unsigned)
2666 		return;
2667 
2668 	{
2669 	dbg_info              *dbgi              = get_irn_dbg_info(builtin);
2670 	ir_type               *type              = get_Builtin_type(builtin);
2671 	ir_type               *lowered_type_high = lower_Builtin_type_high(type);
2672 	ir_type               *lowered_type_low  = lower_Builtin_type_low(type);
2673 	ir_node               *block             = get_nodes_block(builtin);
2674 	ir_node               *mem               = get_Builtin_mem(builtin);
2675 	const lower64_entry_t *entry             = get_node_entry(operand);
2676 	ir_mode               *mode_high         = get_irn_mode(entry->high_word);
2677 	ir_node               *res_high;
2678 	ir_node               *res_low;
2679 
2680 	switch (kind) {
2681 	case ir_bk_bswap: {
2682 		ir_node               *in_high[1] = { entry->high_word };
2683 		ir_node               *in_low[1]  = { entry->low_word };
2684 		ir_node               *swap_high  = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2685 		ir_node               *swap_low   = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2686 		ir_node               *high       = new_r_Proj(swap_high, mode_high, pn_Builtin_max+1);
2687 		ir_node               *low        = new_r_Proj(swap_low, env->low_unsigned, pn_Builtin_max+1);
2688 		if (mode_high == env->low_signed) {
2689 			res_high = new_rd_Conv(dbgi, block, low, env->low_signed);
2690 			res_low  = new_rd_Conv(dbgi, block, high, env->low_unsigned);
2691 		} else {
2692 			res_high = low;
2693 			res_low  = high;
2694 		}
2695 		break;
2696 	}
2697 	default:
2698 		panic("unexpected builtin");
2699 	}
2700 
2701 	/* search result Proj */
2702 	foreach_out_edge_safe(builtin, edge) {
2703 		ir_node *proj = get_edge_src_irn(edge);
2704 		if (!is_Proj(proj))
2705 			continue;
2706 
2707 		if (get_Proj_proj(proj) == pn_Builtin_max+1) {
2708 			ir_set_dw_lowered(proj, res_low, res_high);
2709 		}
2710 	}
2711 	}
2712 }
2713 
2714 /**
2715  * Lower double word builtins.
2716  */
lower_Builtin(ir_node * builtin,ir_mode * mode)2717 static void lower_Builtin(ir_node *builtin, ir_mode *mode)
2718 {
2719 	ir_builtin_kind kind = get_Builtin_kind(builtin);
2720 
2721 	switch (kind) {
2722 	case ir_bk_trap:
2723 	case ir_bk_debugbreak:
2724 	case ir_bk_return_address:
2725 	case ir_bk_frame_address:
2726 	case ir_bk_prefetch:
2727 	case ir_bk_inport:
2728 	case ir_bk_outport:
2729 	case ir_bk_inner_trampoline:
2730 		/* Nothing to do. */
2731 		return;
2732 	case ir_bk_bswap:
2733 		lower_arithmetic_builtin(builtin, mode);
2734 		return;
2735 	case ir_bk_ffs:
2736 	case ir_bk_clz:
2737 	case ir_bk_ctz:
2738 	case ir_bk_popcount:
2739 	case ir_bk_parity:
2740 		lower_reduce_builtin(builtin, mode);
2741 		return;
2742 	}
2743 	panic("unknown builtin");
2744 }
2745 
2746 /**
2747  * check for opcodes that must always be lowered.
2748  */
always_lower(unsigned code)2749 static bool always_lower(unsigned code)
2750 {
2751 	switch (code) {
2752 	case iro_ASM:
2753 	case iro_Builtin:
2754 	case iro_Proj:
2755 	case iro_Start:
2756 	case iro_Call:
2757 	case iro_Return:
2758 	case iro_Cond:
2759 	case iro_Switch:
2760 	case iro_Conv:
2761 	case iro_Sel:
2762 		return true;
2763 	default:
2764 		return false;
2765 	}
2766 }
2767 
2768 /**
2769  * Compare two op_mode_entry_t's.
2770  */
cmp_op_mode(const void * elt,const void * key,size_t size)2771 static int cmp_op_mode(const void *elt, const void *key, size_t size)
2772 {
2773 	const op_mode_entry_t *e1 = (const op_mode_entry_t*)elt;
2774 	const op_mode_entry_t *e2 = (const op_mode_entry_t*)key;
2775 	(void) size;
2776 
2777 	return (e1->op != e2->op) | (e1->imode != e2->imode) | (e1->omode != e2->omode);
2778 }
2779 
2780 /**
2781  * Compare two conv_tp_entry_t's.
2782  */
cmp_conv_tp(const void * elt,const void * key,size_t size)2783 static int cmp_conv_tp(const void *elt, const void *key, size_t size)
2784 {
2785 	const conv_tp_entry_t *e1 = (const conv_tp_entry_t*)elt;
2786 	const conv_tp_entry_t *e2 = (const conv_tp_entry_t*)key;
2787 	(void) size;
2788 
2789 	return (e1->imode != e2->imode) | (e1->omode != e2->omode);
2790 }
2791 
2792 /**
2793  * Enter a lowering function into an ir_op.
2794  */
ir_register_dw_lower_function(ir_op * op,lower_dw_func func)2795 void ir_register_dw_lower_function(ir_op *op, lower_dw_func func)
2796 {
2797 	op->ops.generic = (op_func)func;
2798 }
2799 
2800 /* Determine which modes need to be lowered */
setup_modes(void)2801 static void setup_modes(void)
2802 {
2803 	unsigned           size_bits           = env->params->doubleword_size;
2804 	ir_mode           *doubleword_signed   = NULL;
2805 	ir_mode           *doubleword_unsigned = NULL;
2806 	size_t             n_modes             = ir_get_n_modes();
2807 	ir_mode_arithmetic arithmetic;
2808 	unsigned           modulo_shift;
2809 	size_t             i;
2810 
2811 	/* search for doubleword modes... */
2812 	for (i = 0; i < n_modes; ++i) {
2813 		ir_mode *mode = ir_get_mode(i);
2814 		if (!mode_is_int(mode))
2815 			continue;
2816 		if (get_mode_size_bits(mode) != size_bits)
2817 			continue;
2818 		if (mode_is_signed(mode)) {
2819 			if (doubleword_signed != NULL) {
2820 				/* sigh - the lowerer should really just lower all mode with
2821 				 * size_bits it finds. Unfortunately this required a bigger
2822 				 * rewrite. */
2823 				panic("multiple double word signed modes found");
2824 			}
2825 			doubleword_signed = mode;
2826 		} else {
2827 			if (doubleword_unsigned != NULL) {
2828 				/* sigh - the lowerer should really just lower all mode with
2829 				 * size_bits it finds. Unfortunately this required a bigger
2830 				 * rewrite. */
2831 				panic("multiple double word unsigned modes found");
2832 			}
2833 			doubleword_unsigned = mode;
2834 		}
2835 	}
2836 	if (doubleword_signed == NULL || doubleword_unsigned == NULL) {
2837 		panic("Couldn't find doubleword modes");
2838 	}
2839 
2840 	arithmetic   = get_mode_arithmetic(doubleword_signed);
2841 	modulo_shift = get_mode_modulo_shift(doubleword_signed);
2842 
2843 	assert(get_mode_size_bits(doubleword_unsigned) == size_bits);
2844 	assert(size_bits % 2 == 0);
2845 	assert(get_mode_sign(doubleword_signed) == 1);
2846 	assert(get_mode_sign(doubleword_unsigned) == 0);
2847 	assert(get_mode_sort(doubleword_signed) == irms_int_number);
2848 	assert(get_mode_sort(doubleword_unsigned) == irms_int_number);
2849 	assert(get_mode_arithmetic(doubleword_unsigned) == arithmetic);
2850 	assert(get_mode_modulo_shift(doubleword_unsigned) == modulo_shift);
2851 
2852 	/* try to guess a sensible modulo shift for the new mode.
2853 	 * (This is IMO another indication that this should really be a node
2854 	 *  attribute instead of a mode thing) */
2855 	if (modulo_shift == size_bits) {
2856 		modulo_shift = modulo_shift / 2;
2857 	} else if (modulo_shift == 0) {
2858 		/* fine */
2859 	} else {
2860 		panic("Don't know what new modulo shift to use for lowered doubleword mode");
2861 	}
2862 	size_bits /= 2;
2863 
2864 	/* produce lowered modes */
2865 	env->high_signed   = doubleword_signed;
2866 	env->high_unsigned = doubleword_unsigned;
2867 	env->low_signed    = new_int_mode("WS", arithmetic, size_bits, 1,
2868 	                                  modulo_shift);
2869 	env->low_unsigned  = new_int_mode("WU", arithmetic, size_bits, 0,
2870 	                                  modulo_shift);
2871 }
2872 
enqueue_preds(ir_node * node)2873 static void enqueue_preds(ir_node *node)
2874 {
2875 	int arity = get_irn_arity(node);
2876 	int i;
2877 
2878 	for (i = 0; i < arity; ++i) {
2879 		ir_node *pred = get_irn_n(node, i);
2880 		pdeq_putr(env->waitq, pred);
2881 	}
2882 }
2883 
lower_node(ir_node * node)2884 static void lower_node(ir_node *node)
2885 {
2886 	int              arity;
2887 	int              i;
2888 	lower_dw_func    func;
2889 	ir_op           *op;
2890 	ir_mode         *mode;
2891 	unsigned         idx;
2892 	lower64_entry_t *entry;
2893 
2894 	if (irn_visited_else_mark(node))
2895 		return;
2896 
2897 	/* cycles are always broken at Phi and Block nodes. So we don't need special
2898 	 * magic in all the other lower functions */
2899 	if (is_Block(node)) {
2900 		enqueue_preds(node);
2901 		return;
2902 	} else if (is_Phi(node)) {
2903 		lower_Phi(node);
2904 		return;
2905 	}
2906 
2907 	/* depth-first: descend into operands */
2908 	if (!is_Block(node)) {
2909 		ir_node *block = get_nodes_block(node);
2910 		lower_node(block);
2911 	}
2912 
2913 	if (!is_Cond(node)) {
2914 		arity = get_irn_arity(node);
2915 		for (i = 0; i < arity; ++i) {
2916 			ir_node *pred = get_irn_n(node, i);
2917 			lower_node(pred);
2918 		}
2919 	}
2920 
2921 	op   = get_irn_op(node);
2922 	func = (lower_dw_func) op->ops.generic;
2923 	if (func == NULL)
2924 		return;
2925 
2926 	idx   = get_irn_idx(node);
2927 	entry = idx < env->n_entries ? env->entries[idx] : NULL;
2928 	if (entry != NULL || always_lower(get_irn_opcode(node))) {
2929 		mode = get_irn_op_mode(node);
2930 		if (mode == env->high_signed) {
2931 			mode = env->low_signed;
2932 		} else {
2933 			mode = env->low_unsigned;
2934 		}
2935 		DB((dbg, LEVEL_1, "  %+F\n", node));
2936 		func(node, mode);
2937 	}
2938 }
2939 
clear_node_and_phi_links(ir_node * node,void * data)2940 static void clear_node_and_phi_links(ir_node *node, void *data)
2941 {
2942 	(void) data;
2943 	if (get_irn_mode(node) == mode_T) {
2944 		set_irn_link(node, node);
2945 	} else {
2946 		set_irn_link(node, NULL);
2947 	}
2948 	if (is_Block(node))
2949 		set_Block_phis(node, NULL);
2950 	else if (is_Phi(node))
2951 		set_Phi_next(node, NULL);
2952 }
2953 
lower_irg(ir_graph * irg)2954 static void lower_irg(ir_graph *irg)
2955 {
2956 	ir_entity *ent;
2957 	ir_type   *mtp;
2958 	ir_type   *lowered_mtp;
2959 	unsigned   n_idx;
2960 
2961 	obstack_init(&env->obst);
2962 
2963 	/* just here for debugging */
2964 	current_ir_graph = irg;
2965 	assure_edges(irg);
2966 
2967 	n_idx = get_irg_last_idx(irg);
2968 	n_idx = n_idx + (n_idx >> 2);  /* add 25% */
2969 	env->n_entries = n_idx;
2970 	env->entries   = NEW_ARR_F(lower64_entry_t*, n_idx);
2971 	memset(env->entries, 0, sizeof(env->entries[0]) * n_idx);
2972 
2973 	env->irg            = irg;
2974 	env->flags          = 0;
2975 
2976 	ent = get_irg_entity(irg);
2977 	mtp = get_entity_type(ent);
2978 	lowered_mtp = lower_mtp(mtp);
2979 
2980 	if (lowered_mtp != mtp) {
2981 		set_entity_type(ent, lowered_mtp);
2982 		env->flags |= MUST_BE_LOWERED;
2983 
2984 		fix_parameter_entities(irg, mtp);
2985 	}
2986 
2987 	/* first step: link all nodes and allocate data */
2988 	ir_reserve_resources(irg, IR_RESOURCE_PHI_LIST | IR_RESOURCE_IRN_LINK);
2989 	visit_all_identities(irg, clear_node_and_phi_links, NULL);
2990 	irg_walk_graph(irg, NULL, prepare_links_and_handle_rotl, env);
2991 
2992 	if (env->flags & MUST_BE_LOWERED) {
2993 		size_t i;
2994 		ir_reserve_resources(irg, IR_RESOURCE_IRN_VISITED);
2995 		inc_irg_visited(irg);
2996 
2997 		assert(pdeq_empty(env->waitq));
2998 		pdeq_putr(env->waitq, get_irg_end(irg));
2999 
3000 		env->lowered_phis = NEW_ARR_F(ir_node*, 0);
3001 		while (!pdeq_empty(env->waitq)) {
3002 			ir_node *node = (ir_node*)pdeq_getl(env->waitq);
3003 			lower_node(node);
3004 		}
3005 
3006 		/* we need to fixup phis */
3007 		for (i = 0; i < ARR_LEN(env->lowered_phis); ++i) {
3008 			ir_node *phi = env->lowered_phis[i];
3009 			fixup_phi(phi);
3010 		}
3011 		DEL_ARR_F(env->lowered_phis);
3012 
3013 
3014 		ir_free_resources(irg, IR_RESOURCE_IRN_VISITED);
3015 
3016 		if (env->flags & CF_CHANGED) {
3017 			/* control flow changed, dominance info is invalid */
3018 			clear_irg_properties(irg, IR_GRAPH_PROPERTY_CONSISTENT_DOMINANCE);
3019 		}
3020 		edges_deactivate(irg);
3021 	}
3022 
3023 	ir_free_resources(irg, IR_RESOURCE_PHI_LIST | IR_RESOURCE_IRN_LINK);
3024 
3025 	DEL_ARR_F(env->entries);
3026 	obstack_free(&env->obst, NULL);
3027 }
3028 
3029 static const lwrdw_param_t *param;
3030 
ir_prepare_dw_lowering(const lwrdw_param_t * new_param)3031 void ir_prepare_dw_lowering(const lwrdw_param_t *new_param)
3032 {
3033 	assert(new_param != NULL);
3034 	FIRM_DBG_REGISTER(dbg, "firm.lower.dw");
3035 
3036 	param = new_param;
3037 
3038 	ir_clear_opcodes_generic_func();
3039 	ir_register_dw_lower_function(op_ASM,     lower_ASM);
3040 	ir_register_dw_lower_function(op_Add,     lower_binop);
3041 	ir_register_dw_lower_function(op_And,     lower_And);
3042 	ir_register_dw_lower_function(op_Bad,     lower_Bad);
3043 	ir_register_dw_lower_function(op_Builtin, lower_Builtin);
3044 	ir_register_dw_lower_function(op_Call,    lower_Call);
3045 	ir_register_dw_lower_function(op_Cmp,     lower_Cmp);
3046 	ir_register_dw_lower_function(op_Cond,    lower_Cond);
3047 	ir_register_dw_lower_function(op_Const,   lower_Const);
3048 	ir_register_dw_lower_function(op_Conv,    lower_Conv);
3049 	ir_register_dw_lower_function(op_Div,     lower_Div);
3050 	ir_register_dw_lower_function(op_Eor,     lower_Eor);
3051 	ir_register_dw_lower_function(op_Load,    lower_Load);
3052 	ir_register_dw_lower_function(op_Minus,   lower_unop);
3053 	ir_register_dw_lower_function(op_Mod,     lower_Mod);
3054 	ir_register_dw_lower_function(op_Mul,     lower_binop);
3055 	ir_register_dw_lower_function(op_Mux,     lower_Mux);
3056 	ir_register_dw_lower_function(op_Not,     lower_Not);
3057 	ir_register_dw_lower_function(op_Or,      lower_Or);
3058 	ir_register_dw_lower_function(op_Proj,    lower_Proj);
3059 	ir_register_dw_lower_function(op_Return,  lower_Return);
3060 	ir_register_dw_lower_function(op_Shl,     lower_Shl);
3061 	ir_register_dw_lower_function(op_Shr,     lower_Shr);
3062 	ir_register_dw_lower_function(op_Shrs,    lower_Shrs);
3063 	ir_register_dw_lower_function(op_Start,   lower_Start);
3064 	ir_register_dw_lower_function(op_Store,   lower_Store);
3065 	ir_register_dw_lower_function(op_Sub,     lower_binop);
3066 	ir_register_dw_lower_function(op_Switch,  lower_Switch);
3067 	ir_register_dw_lower_function(op_Unknown, lower_Unknown);
3068 }
3069 
3070 /**
3071  * Callback to lower only the Mux nodes we created.
3072  */
lower_mux_cb(ir_node * mux)3073 static int lower_mux_cb(ir_node *mux)
3074 {
3075 	return ir_nodeset_contains(&created_mux_nodes, mux);
3076 }
3077 
3078 /*
3079  * Do the lowering.
3080  */
ir_lower_dw_ops(void)3081 void ir_lower_dw_ops(void)
3082 {
3083 	lower_dw_env_t lenv;
3084 	size_t      i, n;
3085 
3086 	memset(&lenv, 0, sizeof(lenv));
3087 	lenv.params = param;
3088 	env = &lenv;
3089 
3090 	setup_modes();
3091 
3092 	/* create the necessary maps */
3093 	if (! intrinsic_fkt)
3094 		intrinsic_fkt = new_set(cmp_op_mode, iro_Last + 1);
3095 	if (! conv_types)
3096 		conv_types = new_set(cmp_conv_tp, 16);
3097 	if (! lowered_type)
3098 		lowered_type = pmap_create();
3099 	if (! lowered_builtin_type_low)
3100 		lowered_builtin_type_low = pmap_create();
3101 	if (! lowered_builtin_type_high)
3102 		lowered_builtin_type_high = pmap_create();
3103 
3104 	/* create a primitive unsigned and signed type */
3105 	if (! tp_u)
3106 		tp_u = get_type_for_mode(lenv.low_unsigned);
3107 	if (! tp_s)
3108 		tp_s = get_type_for_mode(lenv.low_signed);
3109 
3110 	/* create method types for the created binop calls */
3111 	if (! binop_tp_u) {
3112 		binop_tp_u = new_type_method(4, 2);
3113 		set_method_param_type(binop_tp_u, 0, tp_u);
3114 		set_method_param_type(binop_tp_u, 1, tp_u);
3115 		set_method_param_type(binop_tp_u, 2, tp_u);
3116 		set_method_param_type(binop_tp_u, 3, tp_u);
3117 		set_method_res_type(binop_tp_u, 0, tp_u);
3118 		set_method_res_type(binop_tp_u, 1, tp_u);
3119 	}
3120 	if (! binop_tp_s) {
3121 		binop_tp_s = new_type_method(4, 2);
3122 		if (env->params->little_endian) {
3123 			set_method_param_type(binop_tp_s, 0, tp_u);
3124 			set_method_param_type(binop_tp_s, 1, tp_s);
3125 			set_method_param_type(binop_tp_s, 2, tp_u);
3126 			set_method_param_type(binop_tp_s, 3, tp_s);
3127 			set_method_res_type(binop_tp_s, 0, tp_u);
3128 			set_method_res_type(binop_tp_s, 1, tp_s);
3129 		} else {
3130 			set_method_param_type(binop_tp_s, 0, tp_s);
3131 			set_method_param_type(binop_tp_s, 1, tp_u);
3132 			set_method_param_type(binop_tp_s, 2, tp_s);
3133 			set_method_param_type(binop_tp_s, 3, tp_u);
3134 			set_method_res_type(binop_tp_s, 0, tp_s);
3135 			set_method_res_type(binop_tp_s, 1, tp_u);
3136 		}
3137 	}
3138 	if (! unop_tp_u) {
3139 		unop_tp_u = new_type_method(2, 2);
3140 		set_method_param_type(unop_tp_u, 0, tp_u);
3141 		set_method_param_type(unop_tp_u, 1, tp_u);
3142 		set_method_res_type(unop_tp_u, 0, tp_u);
3143 		set_method_res_type(unop_tp_u, 1, tp_u);
3144 	}
3145 	if (! unop_tp_s) {
3146 		unop_tp_s = new_type_method(2, 2);
3147 		if (env->params->little_endian) {
3148 			set_method_param_type(unop_tp_s, 0, tp_u);
3149 			set_method_param_type(unop_tp_s, 1, tp_s);
3150 			set_method_res_type(unop_tp_s, 0, tp_u);
3151 			set_method_res_type(unop_tp_s, 1, tp_s);
3152 		} else {
3153 			set_method_param_type(unop_tp_s, 0, tp_s);
3154 			set_method_param_type(unop_tp_s, 1, tp_u);
3155 			set_method_res_type(unop_tp_s, 0, tp_s);
3156 			set_method_res_type(unop_tp_s, 1, tp_u);
3157 		}
3158 	}
3159 
3160 	lenv.tv_mode_bytes = new_tarval_from_long(param->doubleword_size/(2*8), lenv.low_unsigned);
3161 	lenv.waitq         = new_pdeq();
3162 	lenv.first_id      = new_id_from_chars(param->little_endian ? ".l" : ".h", 2);
3163 	lenv.next_id       = new_id_from_chars(param->little_endian ? ".h" : ".l", 2);
3164 
3165 	irp_reserve_resources(irp, IRP_RESOURCE_TYPE_LINK | IRP_RESOURCE_TYPE_VISITED);
3166 	inc_master_type_visited();
3167 	/* transform all graphs */
3168 	for (i = 0, n = get_irp_n_irgs(); i < n; ++i) {
3169 		ir_graph *irg = get_irp_irg(i);
3170 
3171 		ir_nodeset_init(&created_mux_nodes);
3172 
3173 		lower_irg(irg);
3174 
3175 		if (ir_nodeset_size(&created_mux_nodes) > 0)
3176 			lower_mux(irg, lower_mux_cb);
3177 
3178 		ir_nodeset_destroy(&created_mux_nodes);
3179 	}
3180 	irp_free_resources(irp, IRP_RESOURCE_TYPE_LINK | IRP_RESOURCE_TYPE_VISITED);
3181 	del_pdeq(lenv.waitq);
3182 
3183 	env = NULL;
3184 }
3185 
3186 /* Default implementation. */
def_create_intrinsic_fkt(ir_type * method,const ir_op * op,const ir_mode * imode,const ir_mode * omode,void * context)3187 ir_entity *def_create_intrinsic_fkt(ir_type *method, const ir_op *op,
3188                                     const ir_mode *imode, const ir_mode *omode,
3189                                     void *context)
3190 {
3191 	char buf[64];
3192 	ident *id;
3193 	ir_entity *ent;
3194 	(void) context;
3195 
3196 	if (imode == omode) {
3197 		snprintf(buf, sizeof(buf), "__l%s%s", get_op_name(op), get_mode_name(imode));
3198 	} else {
3199 		snprintf(buf, sizeof(buf), "__l%s%s%s", get_op_name(op),
3200 			get_mode_name(imode), get_mode_name(omode));
3201 	}
3202 	id = new_id_from_str(buf);
3203 
3204 	ent = new_entity(get_glob_type(), id, method);
3205 	set_entity_ld_ident(ent, get_entity_ident(ent));
3206 	set_entity_visibility(ent, ir_visibility_external);
3207 	return ent;
3208 }
3209