1 /*
2 * Copyright (C) 1995-2011 University of Karlsruhe. All right reserved.
3 *
4 * This file is part of libFirm.
5 *
6 * This file may be distributed and/or modified under the terms of the
7 * GNU General Public License version 2 as published by the Free Software
8 * Foundation and appearing in the file LICENSE.GPL included in the
9 * packaging of this file.
10 *
11 * Licensees holding valid libFirm Professional Edition licenses may use
12 * this file in accordance with the libFirm Commercial License.
13 * Agreement provided with the Software.
14 *
15 * This file is provided AS IS with NO WARRANTY OF ANY KIND, INCLUDING THE
16 * WARRANTY OF DESIGN, MERCHANTABILITY AND FITNESS FOR A PARTICULAR
17 * PURPOSE.
18 */
19
20 /**
21 * @file
22 * @brief Lower double word operations, i.e. 64bit -> 32bit, 32bit -> 16bit etc.
23 * @date 8.10.2004
24 * @author Michael Beck
25 */
26 #include "config.h"
27
28 #include <string.h>
29 #include <stdlib.h>
30 #include <stdbool.h>
31 #include <assert.h>
32
33 #include "be.h"
34 #include "error.h"
35 #include "lowering.h"
36 #include "irnode_t.h"
37 #include "irnodeset.h"
38 #include "irgraph_t.h"
39 #include "irmode_t.h"
40 #include "iropt_t.h"
41 #include "irgmod.h"
42 #include "tv_t.h"
43 #include "dbginfo_t.h"
44 #include "iropt_dbg.h"
45 #include "irflag_t.h"
46 #include "firmstat.h"
47 #include "irgwalk.h"
48 #include "ircons.h"
49 #include "irflag.h"
50 #include "iroptimize.h"
51 #include "irtools.h"
52 #include "debug.h"
53 #include "set.h"
54 #include "pmap.h"
55 #include "pdeq.h"
56 #include "irdump.h"
57 #include "array_t.h"
58 #include "irpass_t.h"
59 #include "lower_dw.h"
60
61 /** A map from (op, imode, omode) to Intrinsic functions entities. */
62 static set *intrinsic_fkt;
63
64 /** A map from (imode, omode) to conv function types. */
65 static set *conv_types;
66
67 /** A map from a method type to its lowered type. */
68 static pmap *lowered_type;
69
70 /** A map from a builtin type to its lower and higher type. */
71 static pmap *lowered_builtin_type_high;
72 static pmap *lowered_builtin_type_low;
73
74 /** The types for the binop and unop intrinsics. */
75 static ir_type *binop_tp_u, *binop_tp_s, *unop_tp_u, *unop_tp_s, *tp_s, *tp_u;
76
77 static ir_nodeset_t created_mux_nodes;
78
79 /** the debug handle */
80 DEBUG_ONLY(static firm_dbg_module_t *dbg = NULL;)
81
82 /**
83 * An entry in the (op, imode, omode) -> entity map.
84 */
85 typedef struct op_mode_entry {
86 const ir_op *op; /**< the op */
87 const ir_mode *imode; /**< the input mode */
88 const ir_mode *omode; /**< the output mode */
89 ir_entity *ent; /**< the associated entity of this (op, imode, omode) triple */
90 } op_mode_entry_t;
91
92 /**
93 * An entry in the (imode, omode) -> tp map.
94 */
95 typedef struct conv_tp_entry {
96 const ir_mode *imode; /**< the input mode */
97 const ir_mode *omode; /**< the output mode */
98 ir_type *mtd; /**< the associated method type of this (imode, omode) pair */
99 } conv_tp_entry_t;
100
101 enum lower_flags {
102 MUST_BE_LOWERED = 1, /**< graph must be lowered */
103 CF_CHANGED = 2, /**< control flow was changed */
104 };
105
106 /**
107 * The lower environment.
108 */
109 typedef struct lower_dw_env_t {
110 lower64_entry_t **entries; /**< entries per node */
111 ir_graph *irg;
112 struct obstack obst; /**< an obstack holding the temporary data */
113 ir_tarval *tv_mode_bytes; /**< a tarval containing the number of bytes in the lowered modes */
114 pdeq *waitq; /**< a wait queue of all nodes that must be handled later */
115 ir_node **lowered_phis; /**< list of lowered phis */
116 ir_mode *high_signed; /**< doubleword signed type */
117 ir_mode *high_unsigned; /**< doubleword unsigned type */
118 ir_mode *low_signed; /**< word signed type */
119 ir_mode *low_unsigned; /**< word unsigned type */
120 ident *first_id; /**< .l for little and .h for big endian */
121 ident *next_id; /**< .h for little and .l for big endian */
122 const lwrdw_param_t *params; /**< transformation parameter */
123 unsigned flags; /**< some flags */
124 unsigned n_entries; /**< number of entries */
125 } lower_dw_env_t;
126
127 static lower_dw_env_t *env;
128
129 static void lower_node(ir_node *node);
130
131 /**
132 * Create a method type for a Conv emulation from imode to omode.
133 */
get_conv_type(ir_mode * imode,ir_mode * omode)134 static ir_type *get_conv_type(ir_mode *imode, ir_mode *omode)
135 {
136 conv_tp_entry_t key, *entry;
137 ir_type *mtd;
138
139 key.imode = imode;
140 key.omode = omode;
141 key.mtd = NULL;
142
143 entry = set_insert(conv_tp_entry_t, conv_types, &key, sizeof(key), hash_ptr(imode) ^ hash_ptr(omode));
144 if (! entry->mtd) {
145 int n_param = 1, n_res = 1;
146
147 if (imode == env->high_signed || imode == env->high_unsigned)
148 n_param = 2;
149 if (omode == env->high_signed || omode == env->high_unsigned)
150 n_res = 2;
151
152 /* create a new one */
153 mtd = new_type_method(n_param, n_res);
154
155 /* set param types and result types */
156 n_param = 0;
157 if (imode == env->high_signed) {
158 if (env->params->little_endian) {
159 set_method_param_type(mtd, n_param++, tp_u);
160 set_method_param_type(mtd, n_param++, tp_s);
161 } else {
162 set_method_param_type(mtd, n_param++, tp_s);
163 set_method_param_type(mtd, n_param++, tp_u);
164 }
165 } else if (imode == env->high_unsigned) {
166 set_method_param_type(mtd, n_param++, tp_u);
167 set_method_param_type(mtd, n_param++, tp_u);
168 } else {
169 ir_type *tp = get_type_for_mode(imode);
170 set_method_param_type(mtd, n_param++, tp);
171 }
172
173 n_res = 0;
174 if (omode == env->high_signed) {
175 if (env->params->little_endian) {
176 set_method_res_type(mtd, n_res++, tp_u);
177 set_method_res_type(mtd, n_res++, tp_s);
178 } else {
179 set_method_res_type(mtd, n_res++, tp_s);
180 set_method_res_type(mtd, n_res++, tp_u);
181 }
182 } else if (omode == env->high_unsigned) {
183 set_method_res_type(mtd, n_res++, tp_u);
184 set_method_res_type(mtd, n_res++, tp_u);
185 } else {
186 ir_type *tp = get_type_for_mode(omode);
187 set_method_res_type(mtd, n_res++, tp);
188 }
189 entry->mtd = mtd;
190 } else {
191 mtd = entry->mtd;
192 }
193 return mtd;
194 }
195
196 /**
197 * Add an additional control flow input to a block.
198 * Patch all Phi nodes. The new Phi inputs are copied from
199 * old input number nr.
200 */
add_block_cf_input_nr(ir_node * block,int nr,ir_node * cf)201 static void add_block_cf_input_nr(ir_node *block, int nr, ir_node *cf)
202 {
203 int i, arity = get_irn_arity(block);
204 ir_node **in;
205
206 assert(nr < arity);
207
208 NEW_ARR_A(ir_node *, in, arity + 1);
209 for (i = 0; i < arity; ++i)
210 in[i] = get_irn_n(block, i);
211 in[i] = cf;
212
213 set_irn_in(block, i + 1, in);
214
215 foreach_out_edge(block, edge) {
216 ir_node *phi = get_edge_src_irn(edge);
217 if (!is_Phi(phi))
218 continue;
219
220 for (i = 0; i < arity; ++i)
221 in[i] = get_irn_n(phi, i);
222 in[i] = in[nr];
223 set_irn_in(phi, i + 1, in);
224 }
225 }
226
227 /**
228 * Add an additional control flow input to a block.
229 * Patch all Phi nodes. The new Phi inputs are copied from
230 * old input from cf tmpl.
231 */
add_block_cf_input(ir_node * block,ir_node * tmpl,ir_node * cf)232 static void add_block_cf_input(ir_node *block, ir_node *tmpl, ir_node *cf)
233 {
234 int i, arity = get_irn_arity(block);
235 int nr = 0;
236
237 for (i = 0; i < arity; ++i) {
238 if (get_irn_n(block, i) == tmpl) {
239 nr = i;
240 break;
241 }
242 }
243 assert(i < arity);
244 add_block_cf_input_nr(block, nr, cf);
245 }
246
247 /**
248 * Return the "operational" mode of a Firm node.
249 */
get_irn_op_mode(ir_node * node)250 static ir_mode *get_irn_op_mode(ir_node *node)
251 {
252 switch (get_irn_opcode(node)) {
253 case iro_Load:
254 return get_Load_mode(node);
255 case iro_Store:
256 return get_irn_mode(get_Store_value(node));
257 case iro_Div:
258 return get_irn_mode(get_Div_left(node));
259 case iro_Mod:
260 return get_irn_mode(get_Mod_left(node));
261 case iro_Cmp:
262 return get_irn_mode(get_Cmp_left(node));
263 default:
264 return get_irn_mode(node);
265 }
266 }
267
268 /**
269 * Walker, prepare the node links and determine which nodes need to be lowered
270 * at all.
271 */
prepare_links(ir_node * node)272 static void prepare_links(ir_node *node)
273 {
274 ir_mode *mode = get_irn_op_mode(node);
275 lower64_entry_t *link;
276
277 if (mode == env->high_signed || mode == env->high_unsigned) {
278 unsigned idx = get_irn_idx(node);
279 /* ok, found a node that will be lowered */
280 link = OALLOCZ(&env->obst, lower64_entry_t);
281
282 if (idx >= env->n_entries) {
283 /* enlarge: this happens only for Rotl nodes which is RARELY */
284 unsigned old = env->n_entries;
285 unsigned n_idx = idx + (idx >> 3);
286
287 ARR_RESIZE(lower64_entry_t *, env->entries, n_idx);
288 memset(&env->entries[old], 0, (n_idx - old) * sizeof(env->entries[0]));
289 env->n_entries = n_idx;
290 }
291 env->entries[idx] = link;
292 env->flags |= MUST_BE_LOWERED;
293 } else if (is_Conv(node)) {
294 /* Conv nodes have two modes */
295 ir_node *pred = get_Conv_op(node);
296 mode = get_irn_mode(pred);
297
298 if (mode == env->high_signed || mode == env->high_unsigned) {
299 /* must lower this node either but don't need a link */
300 env->flags |= MUST_BE_LOWERED;
301 }
302 return;
303 } else if (is_Call(node)) {
304 /* Special case: If the result of the Call is never used, we won't
305 * find a Proj with a mode that potentially triggers MUST_BE_LOWERED
306 * to be set. Thus, if we see a call, we check its result types and
307 * decide whether MUST_BE_LOWERED has to be set.
308 */
309 ir_type *tp = get_Call_type(node);
310 size_t n_res, i;
311
312 n_res = get_method_n_ress(tp);
313 for (i = 0; i < n_res; ++i) {
314 ir_type *rtp = get_method_res_type(tp, i);
315
316 if (is_Primitive_type(rtp)) {
317 ir_mode *rmode = get_type_mode(rtp);
318
319 if (rmode == env->high_signed || rmode == env->high_unsigned) {
320 env->flags |= MUST_BE_LOWERED;
321 }
322 }
323 }
324 }
325 }
326
get_node_entry(ir_node * node)327 lower64_entry_t *get_node_entry(ir_node *node)
328 {
329 unsigned idx = get_irn_idx(node);
330 assert(idx < env->n_entries);
331 return env->entries[idx];
332 }
333
ir_set_dw_lowered(ir_node * old,ir_node * new_low,ir_node * new_high)334 void ir_set_dw_lowered(ir_node *old, ir_node *new_low, ir_node *new_high)
335 {
336 lower64_entry_t *entry = get_node_entry(old);
337 entry->low_word = new_low;
338 entry->high_word = new_high;
339 }
340
ir_get_low_unsigned_mode(void)341 ir_mode *ir_get_low_unsigned_mode(void)
342 {
343 return env->low_unsigned;
344 }
345
346 /**
347 * Translate a Constant: create two.
348 */
lower_Const(ir_node * node,ir_mode * mode)349 static void lower_Const(ir_node *node, ir_mode *mode)
350 {
351 ir_graph *irg = get_irn_irg(node);
352 dbg_info *dbg = get_irn_dbg_info(node);
353 ir_mode *low_mode = env->low_unsigned;
354 ir_tarval *tv = get_Const_tarval(node);
355 ir_tarval *tv_l = tarval_convert_to(tv, low_mode);
356 ir_node *res_low = new_rd_Const(dbg, irg, tv_l);
357 ir_tarval *tv_shrs = tarval_shrs_unsigned(tv, get_mode_size_bits(low_mode));
358 ir_tarval *tv_h = tarval_convert_to(tv_shrs, mode);
359 ir_node *res_high = new_rd_Const(dbg, irg, tv_h);
360
361 ir_set_dw_lowered(node, res_low, res_high);
362 }
363
364 /**
365 * Translate a Load: create two.
366 */
lower_Load(ir_node * node,ir_mode * mode)367 static void lower_Load(ir_node *node, ir_mode *mode)
368 {
369 ir_mode *low_mode = env->low_unsigned;
370 ir_graph *irg = get_irn_irg(node);
371 ir_node *adr = get_Load_ptr(node);
372 ir_node *mem = get_Load_mem(node);
373 ir_node *low;
374 ir_node *high;
375 ir_node *proj_m;
376 dbg_info *dbg;
377 ir_node *block = get_nodes_block(node);
378 ir_cons_flags volatility = get_Load_volatility(node) == volatility_is_volatile
379 ? cons_volatile : cons_none;
380
381 if (env->params->little_endian) {
382 low = adr;
383 high = new_r_Add(block, adr, new_r_Const(irg, env->tv_mode_bytes), get_irn_mode(adr));
384 } else {
385 low = new_r_Add(block, adr, new_r_Const(irg, env->tv_mode_bytes), get_irn_mode(adr));
386 high = adr;
387 }
388
389 /* create two loads */
390 dbg = get_irn_dbg_info(node);
391 low = new_rd_Load(dbg, block, mem, low, low_mode, volatility);
392 proj_m = new_r_Proj(low, mode_M, pn_Load_M);
393 high = new_rd_Load(dbg, block, proj_m, high, mode, volatility);
394
395 foreach_out_edge_safe(node, edge) {
396 ir_node *proj = get_edge_src_irn(edge);
397 if (!is_Proj(proj))
398 continue;
399
400 switch (get_Proj_proj(proj)) {
401 case pn_Load_M: /* Memory result. */
402 /* put it to the second one */
403 set_Proj_pred(proj, high);
404 break;
405 case pn_Load_X_except: /* Execution result if exception occurred. */
406 /* put it to the first one */
407 set_Proj_pred(proj, low);
408 break;
409 case pn_Load_res: { /* Result of load operation. */
410 ir_node *res_low = new_r_Proj(low, low_mode, pn_Load_res);
411 ir_node *res_high = new_r_Proj(high, mode, pn_Load_res);
412 ir_set_dw_lowered(proj, res_low, res_high);
413 break;
414 }
415 default:
416 assert(0 && "unexpected Proj number");
417 }
418 /* mark this proj: we have handled it already, otherwise we might fall
419 * into out new nodes. */
420 mark_irn_visited(proj);
421 }
422 }
423
424 /**
425 * Translate a Store: create two.
426 */
lower_Store(ir_node * node,ir_mode * mode)427 static void lower_Store(ir_node *node, ir_mode *mode)
428 {
429 ir_graph *irg;
430 ir_node *block, *adr, *mem;
431 ir_node *low, *high, *proj_m;
432 dbg_info *dbg;
433 ir_node *value = get_Store_value(node);
434 const lower64_entry_t *entry = get_node_entry(value);
435 ir_cons_flags volatility = get_Store_volatility(node) == volatility_is_volatile
436 ? cons_volatile : cons_none;
437 (void) mode;
438
439 assert(entry);
440
441 if (! entry->low_word) {
442 /* not ready yet, wait */
443 pdeq_putr(env->waitq, node);
444 return;
445 }
446
447 irg = get_irn_irg(node);
448 adr = get_Store_ptr(node);
449 mem = get_Store_mem(node);
450 block = get_nodes_block(node);
451
452 if (env->params->little_endian) {
453 low = adr;
454 high = new_r_Add(block, adr, new_r_Const(irg, env->tv_mode_bytes), get_irn_mode(adr));
455 } else {
456 low = new_r_Add(block, adr, new_r_Const(irg, env->tv_mode_bytes), get_irn_mode(adr));
457 high = adr;
458 }
459
460 /* create two Stores */
461 dbg = get_irn_dbg_info(node);
462 low = new_rd_Store(dbg, block, mem, low, entry->low_word, volatility);
463 proj_m = new_r_Proj(low, mode_M, pn_Store_M);
464 high = new_rd_Store(dbg, block, proj_m, high, entry->high_word, volatility);
465
466 foreach_out_edge_safe(node, edge) {
467 ir_node *proj = get_edge_src_irn(edge);
468 if (!is_Proj(proj))
469 continue;
470
471 switch (get_Proj_proj(proj)) {
472 case pn_Store_M: /* Memory result. */
473 /* put it to the second one */
474 set_Proj_pred(proj, high);
475 break;
476 case pn_Store_X_except: /* Execution result if exception occurred. */
477 /* put it to the first one */
478 set_Proj_pred(proj, low);
479 break;
480 default:
481 assert(0 && "unexpected Proj number");
482 }
483 /* mark this proj: we have handled it already, otherwise we might fall into
484 * out new nodes. */
485 mark_irn_visited(proj);
486 }
487 }
488
489 /**
490 * Return a node containing the address of the intrinsic emulation function.
491 *
492 * @param method the method type of the emulation function
493 * @param op the emulated ir_op
494 * @param imode the input mode of the emulated opcode
495 * @param omode the output mode of the emulated opcode
496 * @param env the lower environment
497 */
get_intrinsic_address(ir_type * method,ir_op * op,ir_mode * imode,ir_mode * omode)498 static ir_node *get_intrinsic_address(ir_type *method, ir_op *op,
499 ir_mode *imode, ir_mode *omode)
500 {
501 symconst_symbol sym;
502 ir_entity *ent;
503 op_mode_entry_t key, *entry;
504
505 key.op = op;
506 key.imode = imode;
507 key.omode = omode;
508 key.ent = NULL;
509
510 entry = set_insert(op_mode_entry_t, intrinsic_fkt, &key, sizeof(key),
511 hash_ptr(op) ^ hash_ptr(imode) ^ (hash_ptr(omode) << 8));
512 if (! entry->ent) {
513 /* create a new one */
514 ent = env->params->create_intrinsic(method, op, imode, omode, env->params->ctx);
515
516 assert(ent && "Intrinsic creator must return an entity");
517 entry->ent = ent;
518 } else {
519 ent = entry->ent;
520 }
521 sym.entity_p = ent;
522 return new_r_SymConst(env->irg, mode_P_code, sym, symconst_addr_ent);
523 }
524
525 /**
526 * Translate a Div.
527 *
528 * Create an intrinsic Call.
529 */
lower_Div(ir_node * node,ir_mode * mode)530 static void lower_Div(ir_node *node, ir_mode *mode)
531 {
532 ir_node *left = get_Div_left(node);
533 ir_node *right = get_Div_right(node);
534 ir_node *block = get_nodes_block(node);
535 dbg_info *dbgi = get_irn_dbg_info(node);
536 ir_type *mtp = mode_is_signed(mode) ? binop_tp_s : binop_tp_u;
537 ir_mode *opmode = get_irn_op_mode(node);
538 ir_node *addr = get_intrinsic_address(mtp, get_irn_op(node), opmode, opmode);
539 ir_node *in[4];
540 ir_node *call;
541 ir_node *resproj;
542
543 if (env->params->little_endian) {
544 in[0] = get_lowered_low(left);
545 in[1] = get_lowered_high(left);
546 in[2] = get_lowered_low(right);
547 in[3] = get_lowered_high(right);
548 } else {
549 in[0] = get_lowered_high(left);
550 in[1] = get_lowered_low(left);
551 in[2] = get_lowered_high(right);
552 in[3] = get_lowered_low(right);
553 }
554 call = new_rd_Call(dbgi, block, get_Div_mem(node), addr, 4, in, mtp);
555 resproj = new_r_Proj(call, mode_T, pn_Call_T_result);
556 set_irn_pinned(call, get_irn_pinned(node));
557
558 foreach_out_edge_safe(node, edge) {
559 ir_node *proj = get_edge_src_irn(edge);
560 if (!is_Proj(proj))
561 continue;
562
563 switch (get_Proj_proj(proj)) {
564 case pn_Div_M: /* Memory result. */
565 /* reroute to the call */
566 set_Proj_pred(proj, call);
567 set_Proj_proj(proj, pn_Call_M);
568 break;
569 case pn_Div_X_regular:
570 set_Proj_pred(proj, call);
571 set_Proj_proj(proj, pn_Call_X_regular);
572 break;
573 case pn_Div_X_except:
574 set_Proj_pred(proj, call);
575 set_Proj_proj(proj, pn_Call_X_except);
576 break;
577 case pn_Div_res:
578 if (env->params->little_endian) {
579 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 0);
580 ir_node *res_high = new_r_Proj(resproj, mode, 1);
581 ir_set_dw_lowered(proj, res_low, res_high);
582 } else {
583 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 1);
584 ir_node *res_high = new_r_Proj(resproj, mode, 0);
585 ir_set_dw_lowered(proj, res_low, res_high);
586 }
587 break;
588 default:
589 assert(0 && "unexpected Proj number");
590 }
591 /* mark this proj: we have handled it already, otherwise we might fall into
592 * out new nodes. */
593 mark_irn_visited(proj);
594 }
595 }
596
597 /**
598 * Translate a Mod.
599 *
600 * Create an intrinsic Call.
601 */
lower_Mod(ir_node * node,ir_mode * mode)602 static void lower_Mod(ir_node *node, ir_mode *mode)
603 {
604 ir_node *left = get_Mod_left(node);
605 ir_node *right = get_Mod_right(node);
606 dbg_info *dbgi = get_irn_dbg_info(node);
607 ir_node *block = get_nodes_block(node);
608 ir_type *mtp = mode_is_signed(mode) ? binop_tp_s : binop_tp_u;
609 ir_mode *opmode = get_irn_op_mode(node);
610 ir_node *addr = get_intrinsic_address(mtp, get_irn_op(node), opmode, opmode);
611 ir_node *in[4];
612 ir_node *call;
613 ir_node *resproj;
614
615 if (env->params->little_endian) {
616 in[0] = get_lowered_low(left);
617 in[1] = get_lowered_high(left);
618 in[2] = get_lowered_low(right);
619 in[3] = get_lowered_high(right);
620 } else {
621 in[0] = get_lowered_high(left);
622 in[1] = get_lowered_low(left);
623 in[2] = get_lowered_high(right);
624 in[3] = get_lowered_low(right);
625 }
626 call = new_rd_Call(dbgi, block, get_Mod_mem(node), addr, 4, in, mtp);
627 resproj = new_r_Proj(call, mode_T, pn_Call_T_result);
628 set_irn_pinned(call, get_irn_pinned(node));
629
630 foreach_out_edge_safe(node, edge) {
631 ir_node *proj = get_edge_src_irn(edge);
632 if (!is_Proj(proj))
633 continue;
634
635 switch (get_Proj_proj(proj)) {
636 case pn_Mod_M: /* Memory result. */
637 /* reroute to the call */
638 set_Proj_pred(proj, call);
639 set_Proj_proj(proj, pn_Call_M);
640 break;
641 case pn_Div_X_regular:
642 set_Proj_pred(proj, call);
643 set_Proj_proj(proj, pn_Call_X_regular);
644 break;
645 case pn_Mod_X_except:
646 set_Proj_pred(proj, call);
647 set_Proj_proj(proj, pn_Call_X_except);
648 break;
649 case pn_Mod_res:
650 if (env->params->little_endian) {
651 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 0);
652 ir_node *res_high = new_r_Proj(resproj, mode, 1);
653 ir_set_dw_lowered(proj, res_low, res_high);
654 } else {
655 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 1);
656 ir_node *res_high = new_r_Proj(resproj, mode, 0);
657 ir_set_dw_lowered(proj, res_low, res_high);
658 }
659 break;
660 default:
661 assert(0 && "unexpected Proj number");
662 }
663 /* mark this proj: we have handled it already, otherwise we might fall
664 * into out new nodes. */
665 mark_irn_visited(proj);
666 }
667 }
668
669 /**
670 * Translate a binop.
671 *
672 * Create an intrinsic Call.
673 */
lower_binop(ir_node * node,ir_mode * mode)674 static void lower_binop(ir_node *node, ir_mode *mode)
675 {
676 ir_node *left = get_binop_left(node);
677 ir_node *right = get_binop_right(node);
678 dbg_info *dbgi = get_irn_dbg_info(node);
679 ir_node *block = get_nodes_block(node);
680 ir_graph *irg = get_irn_irg(block);
681 ir_type *mtp = mode_is_signed(mode) ? binop_tp_s : binop_tp_u;
682 ir_node *addr = get_intrinsic_address(mtp, get_irn_op(node), mode, mode);
683 ir_node *in[4];
684 ir_node *call;
685 ir_node *resproj;
686
687 if (env->params->little_endian) {
688 in[0] = get_lowered_low(left);
689 in[1] = get_lowered_high(left);
690 in[2] = get_lowered_low(right);
691 in[3] = get_lowered_high(right);
692 } else {
693 in[0] = get_lowered_high(left);
694 in[1] = get_lowered_low(left);
695 in[2] = get_lowered_high(right);
696 in[3] = get_lowered_low(right);
697 }
698 call = new_rd_Call(dbgi, block, get_irg_no_mem(irg), addr, 4, in, mtp);
699 resproj = new_r_Proj(call, mode_T, pn_Call_T_result);
700 set_irn_pinned(call, get_irn_pinned(node));
701
702 if (env->params->little_endian) {
703 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 0);
704 ir_node *res_high = new_r_Proj(resproj, mode, 1);
705 ir_set_dw_lowered(node, res_low, res_high);
706 } else {
707 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 1);
708 ir_node *res_high = new_r_Proj(resproj, mode, 0);
709 ir_set_dw_lowered(node, res_low, res_high);
710 }
711 }
712
create_conv(ir_node * block,ir_node * node,ir_mode * dest_mode)713 static ir_node *create_conv(ir_node *block, ir_node *node, ir_mode *dest_mode)
714 {
715 if (get_irn_mode(node) == dest_mode)
716 return node;
717 return new_r_Conv(block, node, dest_mode);
718 }
719
720 /**
721 * Moves node and all predecessors of node from from_bl to to_bl.
722 * Does not move predecessors of Phi nodes (or block nodes).
723 */
move(ir_node * node,ir_node * from_bl,ir_node * to_bl)724 static void move(ir_node *node, ir_node *from_bl, ir_node *to_bl)
725 {
726 int i, arity;
727
728 /* move this node */
729 set_nodes_block(node, to_bl);
730
731 /* move its Projs */
732 if (get_irn_mode(node) == mode_T) {
733 foreach_out_edge(node, edge) {
734 ir_node *proj = get_edge_src_irn(edge);
735 if (!is_Proj(proj))
736 continue;
737 move(proj, from_bl, to_bl);
738 }
739 }
740
741 /* We must not move predecessors of Phi nodes, even if they are in
742 * from_bl. (because these are values from an earlier loop iteration
743 * which are not predecessors of node here)
744 */
745 if (is_Phi(node))
746 return;
747
748 /* recursion ... */
749 arity = get_irn_arity(node);
750 for (i = 0; i < arity; i++) {
751 ir_node *pred = get_irn_n(node, i);
752 ir_mode *pred_mode = get_irn_mode(pred);
753 if (get_nodes_block(pred) == from_bl)
754 move(pred, from_bl, to_bl);
755 if (pred_mode == env->high_signed || pred_mode == env->high_unsigned) {
756 ir_node *pred_low = get_lowered_low(pred);
757 ir_node *pred_high = get_lowered_high(pred);
758 if (get_nodes_block(pred_low) == from_bl)
759 move(pred_low, from_bl, to_bl);
760 if (pred_high != NULL && get_nodes_block(pred_high) == from_bl)
761 move(pred_high, from_bl, to_bl);
762 }
763 }
764 }
765
766 /**
767 * We need a custom version of part_block_edges because during transformation
768 * not all data-dependencies are explicit yet if a lowered nodes users are not
769 * lowered yet.
770 * We can fix this by modifying move to look for such implicit dependencies.
771 * Additionally we have to keep the proj_2_block map updated
772 */
part_block_dw(ir_node * node)773 static ir_node *part_block_dw(ir_node *node)
774 {
775 ir_graph *irg = get_irn_irg(node);
776 ir_node *old_block = get_nodes_block(node);
777 int n_cfgpreds = get_Block_n_cfgpreds(old_block);
778 ir_node **cfgpreds = get_Block_cfgpred_arr(old_block);
779 ir_node *new_block = new_r_Block(irg, n_cfgpreds, cfgpreds);
780
781 /* old_block has no predecessors anymore for now */
782 set_irn_in(old_block, 0, NULL);
783
784 /* move node and its predecessors to new_block */
785 move(node, old_block, new_block);
786
787 /* move Phi nodes to new_block */
788 foreach_out_edge_safe(old_block, edge) {
789 ir_node *phi = get_edge_src_irn(edge);
790 if (!is_Phi(phi))
791 continue;
792 set_nodes_block(phi, new_block);
793 }
794 return old_block;
795 }
796
797 typedef ir_node* (*new_rd_shr_func)(dbg_info *dbgi, ir_node *block,
798 ir_node *left, ir_node *right,
799 ir_mode *mode);
800
lower_shr_helper(ir_node * node,ir_mode * mode,new_rd_shr_func new_rd_shrs)801 static void lower_shr_helper(ir_node *node, ir_mode *mode,
802 new_rd_shr_func new_rd_shrs)
803 {
804 ir_node *right = get_binop_right(node);
805 ir_node *left = get_binop_left(node);
806 ir_mode *shr_mode = get_irn_mode(node);
807 unsigned modulo_shift = get_mode_modulo_shift(shr_mode);
808 ir_mode *low_unsigned = env->low_unsigned;
809 unsigned modulo_shift2 = get_mode_modulo_shift(mode);
810 ir_graph *irg = get_irn_irg(node);
811 ir_node *left_low = get_lowered_low(left);
812 ir_node *left_high = get_lowered_high(left);
813 dbg_info *dbgi = get_irn_dbg_info(node);
814 ir_node *lower_block;
815 ir_node *block;
816 ir_node *cnst;
817 ir_node *andn;
818 ir_node *cmp;
819 ir_node *cond;
820 ir_node *proj_true;
821 ir_node *proj_false;
822 ir_node *phi_low;
823 ir_node *phi_high;
824 ir_node *lower_in[2];
825 ir_node *phi_low_in[2];
826 ir_node *phi_high_in[2];
827
828 /* this version is optimized for modulo shift architectures
829 * (and can't handle anything else) */
830 if (modulo_shift != get_mode_size_bits(shr_mode)
831 || modulo_shift2<<1 != modulo_shift) {
832 panic("Shr lowering only implemented for modulo shift shr operations");
833 }
834 if (!is_po2(modulo_shift) || !is_po2(modulo_shift2)) {
835 panic("Shr lowering only implemented for power-of-2 modes");
836 }
837 /* without 2-complement the -x instead of (bit_width-x) trick won't work */
838 if (get_mode_arithmetic(shr_mode) != irma_twos_complement) {
839 panic("Shr lowering only implemented for two-complement modes");
840 }
841
842 block = get_nodes_block(node);
843
844 /* if the right operand is a 64bit value, we're only interested in the
845 * lower word */
846 if (get_irn_mode(right) == env->high_unsigned) {
847 right = get_lowered_low(right);
848 } else {
849 /* shift should never have signed mode on the right */
850 assert(get_irn_mode(right) != env->high_signed);
851 right = create_conv(block, right, low_unsigned);
852 }
853
854 lower_block = part_block_dw(node);
855 env->flags |= CF_CHANGED;
856 block = get_nodes_block(node);
857
858 /* add a Cmp to test if highest bit is set <=> whether we shift more
859 * than half the word width */
860 cnst = new_r_Const_long(irg, low_unsigned, modulo_shift2);
861 andn = new_r_And(block, right, cnst, low_unsigned);
862 cnst = new_r_Const(irg, get_mode_null(low_unsigned));
863 cmp = new_rd_Cmp(dbgi, block, andn, cnst, ir_relation_equal);
864 cond = new_rd_Cond(dbgi, block, cmp);
865 proj_true = new_r_Proj(cond, mode_X, pn_Cond_true);
866 proj_false = new_r_Proj(cond, mode_X, pn_Cond_false);
867
868 /* the true block => shift_width < 1word */
869 {
870 /* In theory the low value (for 64bit shifts) is:
871 * Or(High << (32-x)), Low >> x)
872 * In practice High << 32-x will fail when x is zero (since we have
873 * modulo shift and 32 will be 0). So instead we use:
874 * Or(High<<1<<~x, Low >> x)
875 */
876 ir_node *in[1] = { proj_true };
877 ir_node *block_true = new_r_Block(irg, ARRAY_SIZE(in), in);
878 ir_node *res_high = new_rd_shrs(dbgi, block_true, left_high,
879 right, mode);
880 ir_node *shift_low = new_rd_Shr(dbgi, block_true, left_low, right,
881 low_unsigned);
882 ir_node *not_shiftval = new_rd_Not(dbgi, block_true, right,
883 low_unsigned);
884 ir_node *conv = create_conv(block_true, left_high,
885 low_unsigned);
886 ir_node *one = new_r_Const(irg, get_mode_one(low_unsigned));
887 ir_node *carry0 = new_rd_Shl(dbgi, block_true, conv, one,
888 low_unsigned);
889 ir_node *carry1 = new_rd_Shl(dbgi, block_true, carry0,
890 not_shiftval, low_unsigned);
891 ir_node *res_low = new_rd_Or(dbgi, block_true, shift_low, carry1,
892 low_unsigned);
893 lower_in[0] = new_r_Jmp(block_true);
894 phi_low_in[0] = res_low;
895 phi_high_in[0] = res_high;
896 }
897
898 /* false block => shift_width > 1word */
899 {
900 ir_node *in[1] = { proj_false };
901 ir_node *block_false = new_r_Block(irg, ARRAY_SIZE(in), in);
902 ir_node *conv = create_conv(block_false, left_high, low_unsigned);
903 ir_node *res_low = new_rd_shrs(dbgi, block_false, conv, right,
904 low_unsigned);
905 int cnsti = modulo_shift2-1;
906 ir_node *cnst2 = new_r_Const_long(irg, low_unsigned, cnsti);
907 ir_node *res_high;
908 if (new_rd_shrs == new_rd_Shrs) {
909 res_high = new_rd_shrs(dbgi, block_false, left_high, cnst2, mode);
910 } else {
911 res_high = new_r_Const(irg, get_mode_null(mode));
912 }
913 lower_in[1] = new_r_Jmp(block_false);
914 phi_low_in[1] = res_low;
915 phi_high_in[1] = res_high;
916 }
917
918 /* patch lower block */
919 set_irn_in(lower_block, ARRAY_SIZE(lower_in), lower_in);
920 phi_low = new_r_Phi(lower_block, ARRAY_SIZE(phi_low_in), phi_low_in,
921 low_unsigned);
922 phi_high = new_r_Phi(lower_block, ARRAY_SIZE(phi_high_in), phi_high_in,
923 mode);
924 ir_set_dw_lowered(node, phi_low, phi_high);
925 }
926
lower_Shr(ir_node * node,ir_mode * mode)927 static void lower_Shr(ir_node *node, ir_mode *mode)
928 {
929 lower_shr_helper(node, mode, new_rd_Shr);
930 }
931
lower_Shrs(ir_node * node,ir_mode * mode)932 static void lower_Shrs(ir_node *node, ir_mode *mode)
933 {
934 lower_shr_helper(node, mode, new_rd_Shrs);
935 }
936
lower_Shl(ir_node * node,ir_mode * mode)937 static void lower_Shl(ir_node *node, ir_mode *mode)
938 {
939 ir_node *right = get_binop_right(node);
940 ir_node *left = get_binop_left(node);
941 ir_mode *shr_mode = get_irn_mode(node);
942 unsigned modulo_shift = get_mode_modulo_shift(shr_mode);
943 ir_mode *low_unsigned = env->low_unsigned;
944 unsigned modulo_shift2 = get_mode_modulo_shift(mode);
945 ir_graph *irg = get_irn_irg(node);
946 ir_node *left_low = get_lowered_low(left);
947 ir_node *left_high = get_lowered_high(left);
948 dbg_info *dbgi = get_irn_dbg_info(node);
949 ir_node *lower_block = get_nodes_block(node);
950 ir_node *block;
951 ir_node *cnst;
952 ir_node *andn;
953 ir_node *cmp;
954 ir_node *cond;
955 ir_node *proj_true;
956 ir_node *proj_false;
957 ir_node *phi_low;
958 ir_node *phi_high;
959 ir_node *lower_in[2];
960 ir_node *phi_low_in[2];
961 ir_node *phi_high_in[2];
962
963 /* this version is optimized for modulo shift architectures
964 * (and can't handle anything else) */
965 if (modulo_shift != get_mode_size_bits(shr_mode)
966 || modulo_shift2<<1 != modulo_shift) {
967 panic("Shl lowering only implemented for modulo shift shl operations");
968 }
969 if (!is_po2(modulo_shift) || !is_po2(modulo_shift2)) {
970 panic("Shl lowering only implemented for power-of-2 modes");
971 }
972 /* without 2-complement the -x instead of (bit_width-x) trick won't work */
973 if (get_mode_arithmetic(shr_mode) != irma_twos_complement) {
974 panic("Shl lowering only implemented for two-complement modes");
975 }
976
977 /* if the right operand is a 64bit value, we're only interested in the
978 * lower word */
979 if (get_irn_mode(right) == env->high_unsigned) {
980 right = get_lowered_low(right);
981 } else {
982 /* shift should never have signed mode on the right */
983 assert(get_irn_mode(right) != env->high_signed);
984 right = create_conv(lower_block, right, low_unsigned);
985 }
986
987 part_block_dw(node);
988 env->flags |= CF_CHANGED;
989 block = get_nodes_block(node);
990
991 /* add a Cmp to test if highest bit is set <=> whether we shift more
992 * than half the word width */
993 cnst = new_r_Const_long(irg, low_unsigned, modulo_shift2);
994 andn = new_r_And(block, right, cnst, low_unsigned);
995 cnst = new_r_Const(irg, get_mode_null(low_unsigned));
996 cmp = new_rd_Cmp(dbgi, block, andn, cnst, ir_relation_equal);
997 cond = new_rd_Cond(dbgi, block, cmp);
998 proj_true = new_r_Proj(cond, mode_X, pn_Cond_true);
999 proj_false = new_r_Proj(cond, mode_X, pn_Cond_false);
1000
1001 /* the true block => shift_width < 1word */
1002 {
1003 ir_node *in[1] = { proj_true };
1004 ir_node *block_true = new_r_Block(irg, ARRAY_SIZE(in), in);
1005
1006 ir_node *res_low = new_rd_Shl(dbgi, block_true, left_low,
1007 right, low_unsigned);
1008 ir_node *shift_high = new_rd_Shl(dbgi, block_true, left_high, right,
1009 mode);
1010 ir_node *not_shiftval = new_rd_Not(dbgi, block_true, right,
1011 low_unsigned);
1012 ir_node *conv = create_conv(block_true, left_low, mode);
1013 ir_node *one = new_r_Const(irg, get_mode_one(low_unsigned));
1014 ir_node *carry0 = new_rd_Shr(dbgi, block_true, conv, one, mode);
1015 ir_node *carry1 = new_rd_Shr(dbgi, block_true, carry0,
1016 not_shiftval, mode);
1017 ir_node *res_high = new_rd_Or(dbgi, block_true, shift_high, carry1,
1018 mode);
1019 lower_in[0] = new_r_Jmp(block_true);
1020 phi_low_in[0] = res_low;
1021 phi_high_in[0] = res_high;
1022 }
1023
1024 /* false block => shift_width > 1word */
1025 {
1026 ir_node *in[1] = { proj_false };
1027 ir_node *block_false = new_r_Block(irg, ARRAY_SIZE(in), in);
1028 ir_node *res_low = new_r_Const(irg, get_mode_null(low_unsigned));
1029 ir_node *conv = create_conv(block_false, left_low, mode);
1030 ir_node *res_high = new_rd_Shl(dbgi, block_false, conv, right, mode);
1031 lower_in[1] = new_r_Jmp(block_false);
1032 phi_low_in[1] = res_low;
1033 phi_high_in[1] = res_high;
1034 }
1035
1036 /* patch lower block */
1037 set_irn_in(lower_block, ARRAY_SIZE(lower_in), lower_in);
1038 phi_low = new_r_Phi(lower_block, ARRAY_SIZE(phi_low_in), phi_low_in,
1039 low_unsigned);
1040 phi_high = new_r_Phi(lower_block, ARRAY_SIZE(phi_high_in), phi_high_in,
1041 mode);
1042 ir_set_dw_lowered(node, phi_low, phi_high);
1043 }
1044
1045 /**
1046 * Rebuild Rotl nodes into Or(Shl, Shr) and prepare all nodes.
1047 */
prepare_links_and_handle_rotl(ir_node * node,void * data)1048 static void prepare_links_and_handle_rotl(ir_node *node, void *data)
1049 {
1050 (void) data;
1051 if (is_Rotl(node)) {
1052 ir_mode *mode = get_irn_op_mode(node);
1053 ir_node *right;
1054 ir_node *left, *shl, *shr, *ornode, *block, *sub, *c;
1055 ir_mode *omode, *rmode;
1056 ir_graph *irg;
1057 dbg_info *dbg;
1058 optimization_state_t state;
1059
1060 if (mode != env->high_signed && mode != env->high_unsigned) {
1061 prepare_links(node);
1062 return;
1063 }
1064
1065 /* replace the Rotl(x,y) by an Or(Shl(x,y), Shr(x,64-y)) */
1066 right = get_Rotl_right(node);
1067 irg = get_irn_irg(node);
1068 dbg = get_irn_dbg_info(node);
1069 omode = get_irn_mode(node);
1070 left = get_Rotl_left(node);
1071 block = get_nodes_block(node);
1072 shl = new_rd_Shl(dbg, block, left, right, omode);
1073 rmode = get_irn_mode(right);
1074 c = new_r_Const_long(irg, rmode, get_mode_size_bits(omode));
1075 sub = new_rd_Sub(dbg, block, c, right, rmode);
1076 shr = new_rd_Shr(dbg, block, left, sub, omode);
1077
1078 /* switch optimization off here, or we will get the Rotl back */
1079 save_optimization_state(&state);
1080 set_opt_algebraic_simplification(0);
1081 ornode = new_rd_Or(dbg, block, shl, shr, omode);
1082 restore_optimization_state(&state);
1083
1084 exchange(node, ornode);
1085
1086 /* do lowering on the new nodes */
1087 prepare_links(shl);
1088 prepare_links(c);
1089 prepare_links(sub);
1090 prepare_links(shr);
1091 prepare_links(ornode);
1092 return;
1093 }
1094
1095 prepare_links(node);
1096 }
1097
1098 /**
1099 * Translate an Unop.
1100 *
1101 * Create an intrinsic Call.
1102 */
lower_unop(ir_node * node,ir_mode * mode)1103 static void lower_unop(ir_node *node, ir_mode *mode)
1104 {
1105 ir_node *op = get_unop_op(node);
1106 dbg_info *dbgi = get_irn_dbg_info(node);
1107 ir_node *block = get_nodes_block(node);
1108 ir_graph *irg = get_irn_irg(block);
1109 ir_type *mtp = mode_is_signed(mode) ? unop_tp_s : unop_tp_u;
1110 ir_op *irop = get_irn_op(node);
1111 ir_node *addr = get_intrinsic_address(mtp, irop, mode, mode);
1112 ir_node *nomem = get_irg_no_mem(irg);
1113 ir_node *in[2];
1114 ir_node *call;
1115 ir_node *resproj;
1116
1117 if (env->params->little_endian) {
1118 in[0] = get_lowered_low(op);
1119 in[1] = get_lowered_high(op);
1120 } else {
1121 in[0] = get_lowered_high(op);
1122 in[1] = get_lowered_low(op);
1123 }
1124 call = new_rd_Call(dbgi, block, nomem, addr, 2, in, mtp);
1125 resproj = new_r_Proj(call, mode_T, pn_Call_T_result);
1126 set_irn_pinned(call, get_irn_pinned(node));
1127
1128 if (env->params->little_endian) {
1129 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 0);
1130 ir_node *res_high = new_r_Proj(resproj, mode, 1);
1131 ir_set_dw_lowered(node, res_low, res_high);
1132 } else {
1133 ir_node *res_low = new_r_Proj(resproj, env->low_unsigned, 1);
1134 ir_node *res_high = new_r_Proj(resproj, mode, 0);
1135 ir_set_dw_lowered(node, res_low, res_high);
1136 }
1137 }
1138
1139 /**
1140 * Translate a logical binop.
1141 *
1142 * Create two logical binops.
1143 */
lower_binop_logical(ir_node * node,ir_mode * mode,ir_node * (* constr_rd)(dbg_info * db,ir_node * block,ir_node * op1,ir_node * op2,ir_mode * mode))1144 static void lower_binop_logical(ir_node *node, ir_mode *mode,
1145 ir_node *(*constr_rd)(dbg_info *db, ir_node *block, ir_node *op1, ir_node *op2, ir_mode *mode) )
1146 {
1147 ir_node *left = get_binop_left(node);
1148 ir_node *right = get_binop_right(node);
1149 const lower64_entry_t *left_entry = get_node_entry(left);
1150 const lower64_entry_t *right_entry = get_node_entry(right);
1151 dbg_info *dbgi = get_irn_dbg_info(node);
1152 ir_node *block = get_nodes_block(node);
1153 ir_node *res_low
1154 = constr_rd(dbgi, block, left_entry->low_word, right_entry->low_word,
1155 env->low_unsigned);
1156 ir_node *res_high
1157 = constr_rd(dbgi, block, left_entry->high_word, right_entry->high_word,
1158 mode);
1159 ir_set_dw_lowered(node, res_low, res_high);
1160 }
1161
lower_And(ir_node * node,ir_mode * mode)1162 static void lower_And(ir_node *node, ir_mode *mode)
1163 {
1164 lower_binop_logical(node, mode, new_rd_And);
1165 }
1166
lower_Or(ir_node * node,ir_mode * mode)1167 static void lower_Or(ir_node *node, ir_mode *mode)
1168 {
1169 lower_binop_logical(node, mode, new_rd_Or);
1170 }
1171
lower_Eor(ir_node * node,ir_mode * mode)1172 static void lower_Eor(ir_node *node, ir_mode *mode)
1173 {
1174 lower_binop_logical(node, mode, new_rd_Eor);
1175 }
1176
1177 /**
1178 * Translate a Not.
1179 *
1180 * Create two logical Nots.
1181 */
lower_Not(ir_node * node,ir_mode * mode)1182 static void lower_Not(ir_node *node, ir_mode *mode)
1183 {
1184 ir_node *op = get_Not_op(node);
1185 const lower64_entry_t *op_entry = get_node_entry(op);
1186 dbg_info *dbgi = get_irn_dbg_info(node);
1187 ir_node *block = get_nodes_block(node);
1188 ir_node *res_low
1189 = new_rd_Not(dbgi, block, op_entry->low_word, env->low_unsigned);
1190 ir_node *res_high
1191 = new_rd_Not(dbgi, block, op_entry->high_word, mode);
1192 ir_set_dw_lowered(node, res_low, res_high);
1193 }
1194
lower_Proj(ir_node * node,ir_mode * op_mode)1195 static void lower_Proj(ir_node *node, ir_mode *op_mode)
1196 {
1197 ir_mode *mode = get_irn_mode(node);
1198 ir_node *pred;
1199 (void)op_mode;
1200 if (mode != env->high_signed && mode != env->high_unsigned)
1201 return;
1202 /* skip tuples */
1203 pred = get_Proj_pred(node);
1204 if (is_Tuple(pred)) {
1205 long pn = get_Proj_proj(node);
1206 ir_node *op = get_irn_n(pred, pn);
1207 const lower64_entry_t *entry = get_node_entry(op);
1208 ir_set_dw_lowered(node, entry->low_word, entry->high_word);
1209 }
1210 }
1211
is_equality_cmp(const ir_node * node)1212 static bool is_equality_cmp(const ir_node *node)
1213 {
1214 ir_relation relation = get_Cmp_relation(node);
1215 ir_node *left = get_Cmp_left(node);
1216 ir_node *right = get_Cmp_right(node);
1217 ir_mode *mode = get_irn_mode(left);
1218
1219 /* this probably makes no sense if unordered is involved */
1220 assert(!mode_is_float(mode));
1221
1222 if (relation == ir_relation_equal || relation == ir_relation_less_greater)
1223 return true;
1224
1225 if (!is_Const(right) || !is_Const_null(right))
1226 return false;
1227 if (mode_is_signed(mode)) {
1228 return relation == ir_relation_less_greater;
1229 } else {
1230 return relation == ir_relation_greater;
1231 }
1232 }
1233
get_cfop_destination(const ir_node * cfop)1234 static ir_node *get_cfop_destination(const ir_node *cfop)
1235 {
1236 const ir_edge_t *first = get_irn_out_edge_first(cfop);
1237 /* we should only have 1 destination */
1238 assert(get_irn_n_edges(cfop) == 1);
1239 return get_edge_src_irn(first);
1240 }
1241
lower_Switch(ir_node * node,ir_mode * high_mode)1242 static void lower_Switch(ir_node *node, ir_mode *high_mode)
1243 {
1244 ir_node *selector = get_Switch_selector(node);
1245 ir_mode *mode = get_irn_mode(selector);
1246 (void)high_mode;
1247 if (mode == env->high_signed || mode == env->high_unsigned) {
1248 /* we can't really handle Switch with 64bit offsets */
1249 panic("Switch with 64bit jumptable not supported");
1250 }
1251 lower_node(selector);
1252 }
1253
1254 /**
1255 * Translate a Cond.
1256 */
lower_Cond(ir_node * node,ir_mode * high_mode)1257 static void lower_Cond(ir_node *node, ir_mode *high_mode)
1258 {
1259 ir_node *left, *right, *block;
1260 ir_node *sel = get_Cond_selector(node);
1261 ir_mode *cmp_mode;
1262 const lower64_entry_t *lentry, *rentry;
1263 ir_node *projT = NULL, *projF = NULL;
1264 ir_node *new_bl, *irn;
1265 ir_node *projHF, *projHT;
1266 ir_node *dst_blk;
1267 ir_relation relation;
1268 ir_graph *irg;
1269 dbg_info *dbg;
1270
1271 (void) high_mode;
1272
1273 if (!is_Cmp(sel)) {
1274 lower_node(sel);
1275 return;
1276 }
1277
1278 left = get_Cmp_left(sel);
1279 cmp_mode = get_irn_mode(left);
1280 if (cmp_mode != env->high_signed && cmp_mode != env->high_unsigned) {
1281 lower_node(sel);
1282 return;
1283 }
1284
1285 right = get_Cmp_right(sel);
1286 lower_node(left);
1287 lower_node(right);
1288 lentry = get_node_entry(left);
1289 rentry = get_node_entry(right);
1290
1291 /* all right, build the code */
1292 foreach_out_edge_safe(node, edge) {
1293 ir_node *proj = get_edge_src_irn(edge);
1294 long proj_nr;
1295 if (!is_Proj(proj))
1296 continue;
1297 proj_nr = get_Proj_proj(proj);
1298
1299 if (proj_nr == pn_Cond_true) {
1300 assert(projT == NULL && "more than one Proj(true)");
1301 projT = proj;
1302 } else {
1303 assert(proj_nr == pn_Cond_false);
1304 assert(projF == NULL && "more than one Proj(false)");
1305 projF = proj;
1306 }
1307 mark_irn_visited(proj);
1308 }
1309 assert(projT && projF);
1310
1311 /* create a new high compare */
1312 block = get_nodes_block(node);
1313 irg = get_Block_irg(block);
1314 dbg = get_irn_dbg_info(sel);
1315 relation = get_Cmp_relation(sel);
1316
1317 if (is_equality_cmp(sel)) {
1318 /* x ==/!= y ==> or(x_low^y_low,x_high^y_high) ==/!= 0 */
1319 ir_mode *mode = env->low_unsigned;
1320 ir_node *low_left = new_rd_Conv(dbg, block, lentry->low_word, mode);
1321 ir_node *high_left = new_rd_Conv(dbg, block, lentry->high_word, mode);
1322 ir_node *low_right = new_rd_Conv(dbg, block, rentry->low_word, mode);
1323 ir_node *high_right = new_rd_Conv(dbg, block, rentry->high_word, mode);
1324 ir_node *xor_low = new_rd_Eor(dbg, block, low_left, low_right, mode);
1325 ir_node *xor_high = new_rd_Eor(dbg, block, high_left, high_right, mode);
1326 ir_node *ornode = new_rd_Or(dbg, block, xor_low, xor_high, mode);
1327 ir_node *cmp = new_rd_Cmp(dbg, block, ornode, new_r_Const(irg, get_mode_null(mode)), relation);
1328 set_Cond_selector(node, cmp);
1329 return;
1330 }
1331
1332 if (relation == ir_relation_equal) {
1333 ir_node *proj;
1334 /* simple case:a == b <==> a_h == b_h && a_l == b_l */
1335 dst_blk = get_cfop_destination(projF);
1336
1337 irn = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1338 ir_relation_equal);
1339 dbg = get_irn_dbg_info(node);
1340 irn = new_rd_Cond(dbg, block, irn);
1341
1342 projHF = new_r_Proj(irn, mode_X, pn_Cond_false);
1343 mark_irn_visited(projHF);
1344 exchange(projF, projHF);
1345
1346 projHT = new_r_Proj(irn, mode_X, pn_Cond_true);
1347 mark_irn_visited(projHT);
1348
1349 new_bl = new_r_Block(irg, 1, &projHT);
1350
1351 dbg = get_irn_dbg_info(sel);
1352 irn = new_rd_Cmp(dbg, new_bl, lentry->low_word, rentry->low_word,
1353 ir_relation_equal);
1354 dbg = get_irn_dbg_info(node);
1355 irn = new_rd_Cond(dbg, new_bl, irn);
1356
1357 proj = new_r_Proj(irn, mode_X, pn_Cond_false);
1358 mark_irn_visited(proj);
1359 add_block_cf_input(dst_blk, projHF, proj);
1360
1361 proj = new_r_Proj(irn, mode_X, pn_Cond_true);
1362 mark_irn_visited(proj);
1363 exchange(projT, proj);
1364 } else if (relation == ir_relation_less_greater) {
1365 ir_node *proj;
1366 /* simple case:a != b <==> a_h != b_h || a_l != b_l */
1367 dst_blk = get_cfop_destination(projT);
1368
1369 irn = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1370 ir_relation_less_greater);
1371 dbg = get_irn_dbg_info(node);
1372 irn = new_rd_Cond(dbg, block, irn);
1373
1374 projHT = new_r_Proj(irn, mode_X, pn_Cond_true);
1375 mark_irn_visited(projHT);
1376 exchange(projT, projHT);
1377
1378 projHF = new_r_Proj(irn, mode_X, pn_Cond_false);
1379 mark_irn_visited(projHF);
1380
1381 new_bl = new_r_Block(irg, 1, &projHF);
1382
1383 dbg = get_irn_dbg_info(sel);
1384 irn = new_rd_Cmp(dbg, new_bl, lentry->low_word, rentry->low_word,
1385 ir_relation_less_greater);
1386 dbg = get_irn_dbg_info(node);
1387 irn = new_rd_Cond(dbg, new_bl, irn);
1388
1389 proj = new_r_Proj(irn, mode_X, pn_Cond_true);
1390 mark_irn_visited(proj);
1391 add_block_cf_input(dst_blk, projHT, proj);
1392
1393 proj = new_r_Proj(irn, mode_X, pn_Cond_false);
1394 mark_irn_visited(proj);
1395 exchange(projF, proj);
1396 } else {
1397 ir_node *proj;
1398 /* a rel b <==> a_h REL b_h || (a_h == b_h && a_l rel b_l) */
1399 ir_node *dstT, *dstF, *newbl_eq, *newbl_l;
1400 ir_node *projEqF;
1401
1402 dstT = get_cfop_destination(projT);
1403 dstF = get_cfop_destination(projF);
1404
1405 irn = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1406 relation & ~ir_relation_equal);
1407 dbg = get_irn_dbg_info(node);
1408 irn = new_rd_Cond(dbg, block, irn);
1409
1410 projHT = new_r_Proj(irn, mode_X, pn_Cond_true);
1411 mark_irn_visited(projHT);
1412
1413 projHF = new_r_Proj(irn, mode_X, pn_Cond_false);
1414 mark_irn_visited(projHF);
1415
1416 newbl_eq = new_r_Block(irg, 1, &projHF);
1417
1418 irn = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1419 ir_relation_equal);
1420 irn = new_rd_Cond(dbg, newbl_eq, irn);
1421
1422 projEqF = new_r_Proj(irn, mode_X, pn_Cond_false);
1423 mark_irn_visited(projEqF);
1424
1425 proj = new_r_Proj(irn, mode_X, pn_Cond_true);
1426 mark_irn_visited(proj);
1427
1428 newbl_l = new_r_Block(irg, 1, &proj);
1429
1430 dbg = get_irn_dbg_info(sel);
1431 irn = new_rd_Cmp(dbg, newbl_l, lentry->low_word, rentry->low_word,
1432 relation);
1433 dbg = get_irn_dbg_info(node);
1434 irn = new_rd_Cond(dbg, newbl_l, irn);
1435
1436 proj = new_r_Proj(irn, mode_X, pn_Cond_true);
1437 mark_irn_visited(proj);
1438 add_block_cf_input(dstT, projT, proj);
1439
1440 proj = new_r_Proj(irn, mode_X, pn_Cond_false);
1441 mark_irn_visited(proj);
1442 add_block_cf_input(dstF, projF, proj);
1443
1444 exchange(projT, projHT);
1445 exchange(projF, projEqF);
1446 }
1447
1448 /* we have changed the control flow */
1449 env->flags |= CF_CHANGED;
1450 }
1451
1452 /**
1453 * Translate a Conv to higher_signed
1454 */
lower_Conv_to_Ll(ir_node * node)1455 static void lower_Conv_to_Ll(ir_node *node)
1456 {
1457 ir_mode *omode = get_irn_mode(node);
1458 ir_node *op = get_Conv_op(node);
1459 ir_mode *imode = get_irn_mode(op);
1460 ir_graph *irg = get_irn_irg(node);
1461 ir_node *block = get_nodes_block(node);
1462 dbg_info *dbg = get_irn_dbg_info(node);
1463 ir_node *res_low;
1464 ir_node *res_high;
1465
1466 ir_mode *low_unsigned = env->low_unsigned;
1467 ir_mode *low_signed
1468 = mode_is_signed(omode) ? env->low_signed : low_unsigned;
1469
1470 if (mode_is_int(imode) || mode_is_reference(imode)) {
1471 if (imode == env->high_signed || imode == env->high_unsigned) {
1472 /* a Conv from Lu to Ls or Ls to Lu */
1473 const lower64_entry_t *op_entry = get_node_entry(op);
1474 res_low = op_entry->low_word;
1475 res_high = new_rd_Conv(dbg, block, op_entry->high_word, low_signed);
1476 } else {
1477 /* simple case: create a high word */
1478 if (imode != low_unsigned)
1479 op = new_rd_Conv(dbg, block, op, low_unsigned);
1480
1481 res_low = op;
1482
1483 if (mode_is_signed(imode)) {
1484 int c = get_mode_size_bits(low_signed) - 1;
1485 ir_node *cnst = new_r_Const_long(irg, low_unsigned, c);
1486 if (get_irn_mode(op) != low_signed)
1487 op = new_rd_Conv(dbg, block, op, low_signed);
1488 res_high = new_rd_Shrs(dbg, block, op, cnst, low_signed);
1489 } else {
1490 res_high = new_r_Const(irg, get_mode_null(low_signed));
1491 }
1492 }
1493 } else if (imode == mode_b) {
1494 res_low = new_rd_Conv(dbg, block, op, low_unsigned);
1495 res_high = new_r_Const(irg, get_mode_null(low_signed));
1496 } else {
1497 ir_node *irn, *call;
1498 ir_type *mtp = get_conv_type(imode, omode);
1499
1500 irn = get_intrinsic_address(mtp, get_irn_op(node), imode, omode);
1501 call = new_rd_Call(dbg, block, get_irg_no_mem(irg), irn, 1, &op, mtp);
1502 set_irn_pinned(call, get_irn_pinned(node));
1503 irn = new_r_Proj(call, mode_T, pn_Call_T_result);
1504
1505 if (env->params->little_endian) {
1506 res_low = new_r_Proj(irn, low_unsigned, 0);
1507 res_high = new_r_Proj(irn, low_signed, 1);
1508 } else {
1509 res_low = new_r_Proj(irn, low_unsigned, 1);
1510 res_high = new_r_Proj(irn, low_signed, 0);
1511 }
1512 }
1513 ir_set_dw_lowered(node, res_low, res_high);
1514 }
1515
1516 /**
1517 * Translate a Conv from higher_unsigned
1518 */
lower_Conv_from_Ll(ir_node * node)1519 static void lower_Conv_from_Ll(ir_node *node)
1520 {
1521 ir_node *op = get_Conv_op(node);
1522 ir_mode *omode = get_irn_mode(node);
1523 ir_node *block = get_nodes_block(node);
1524 dbg_info *dbg = get_irn_dbg_info(node);
1525 ir_graph *irg = get_irn_irg(node);
1526 const lower64_entry_t *entry = get_node_entry(op);
1527
1528 if (mode_is_int(omode) || mode_is_reference(omode)) {
1529 op = entry->low_word;
1530
1531 /* simple case: create a high word */
1532 if (omode != env->low_unsigned)
1533 op = new_rd_Conv(dbg, block, op, omode);
1534
1535 set_Conv_op(node, op);
1536 } else if (omode == mode_b) {
1537 /* llu ? true : false <=> (low|high) ? true : false */
1538 ir_mode *mode = env->low_unsigned;
1539 ir_node *ornode = new_rd_Or(dbg, block, entry->low_word,
1540 entry->high_word, mode);
1541 set_Conv_op(node, ornode);
1542 } else {
1543 ir_node *irn, *call, *in[2];
1544 ir_mode *imode = get_irn_mode(op);
1545 ir_type *mtp = get_conv_type(imode, omode);
1546 ir_node *res;
1547
1548 irn = get_intrinsic_address(mtp, get_irn_op(node), imode, omode);
1549 if (env->params->little_endian) {
1550 in[0] = entry->low_word;
1551 in[1] = entry->high_word;
1552 } else {
1553 in[0] = entry->high_word;
1554 in[1] = entry->low_word;
1555 }
1556
1557 call = new_rd_Call(dbg, block, get_irg_no_mem(irg), irn, 2, in, mtp);
1558 set_irn_pinned(call, get_irn_pinned(node));
1559 irn = new_r_Proj(call, mode_T, pn_Call_T_result);
1560 res = new_r_Proj(irn, omode, 0);
1561
1562 exchange(node, res);
1563 }
1564 }
1565
1566 /**
1567 * lower Cmp
1568 */
lower_Cmp(ir_node * cmp,ir_mode * m)1569 static void lower_Cmp(ir_node *cmp, ir_mode *m)
1570 {
1571 ir_node *l = get_Cmp_left(cmp);
1572 ir_mode *cmp_mode = get_irn_mode(l);
1573 ir_node *r, *low, *high, *t, *res;
1574 ir_relation relation;
1575 ir_node *block;
1576 dbg_info *dbg;
1577 const lower64_entry_t *lentry;
1578 const lower64_entry_t *rentry;
1579 (void) m;
1580
1581 if (cmp_mode != env->high_signed && cmp_mode != env->high_unsigned)
1582 return;
1583
1584 r = get_Cmp_right(cmp);
1585 lentry = get_node_entry(l);
1586 rentry = get_node_entry(r);
1587 relation = get_Cmp_relation(cmp);
1588 block = get_nodes_block(cmp);
1589 dbg = get_irn_dbg_info(cmp);
1590
1591 /* easy case for x ==/!= 0 (see lower_Cond for details) */
1592 if (is_equality_cmp(cmp)) {
1593 ir_graph *irg = get_irn_irg(cmp);
1594 ir_mode *mode = env->low_unsigned;
1595 ir_node *low_left = new_rd_Conv(dbg, block, lentry->low_word, mode);
1596 ir_node *high_left = new_rd_Conv(dbg, block, lentry->high_word, mode);
1597 ir_node *low_right = new_rd_Conv(dbg, block, rentry->low_word, mode);
1598 ir_node *high_right = new_rd_Conv(dbg, block, rentry->high_word, mode);
1599 ir_node *xor_low = new_rd_Eor(dbg, block, low_left, low_right, mode);
1600 ir_node *xor_high = new_rd_Eor(dbg, block, high_left, high_right, mode);
1601 ir_node *ornode = new_rd_Or(dbg, block, xor_low, xor_high, mode);
1602 ir_node *new_cmp = new_rd_Cmp(dbg, block, ornode, new_r_Const(irg, get_mode_null(mode)), relation);
1603 exchange(cmp, new_cmp);
1604 return;
1605 }
1606
1607 if (relation == ir_relation_equal) {
1608 /* simple case:a == b <==> a_h == b_h && a_l == b_l */
1609 low = new_rd_Cmp(dbg, block, lentry->low_word, rentry->low_word,
1610 relation);
1611 high = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1612 relation);
1613 res = new_rd_And(dbg, block, low, high, mode_b);
1614 } else if (relation == ir_relation_less_greater) {
1615 /* simple case:a != b <==> a_h != b_h || a_l != b_l */
1616 low = new_rd_Cmp(dbg, block, lentry->low_word, rentry->low_word,
1617 relation);
1618 high = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1619 relation);
1620 res = new_rd_Or(dbg, block, low, high, mode_b);
1621 } else {
1622 /* a rel b <==> a_h REL b_h || (a_h == b_h && a_l rel b_l) */
1623 ir_node *high1 = new_rd_Cmp(dbg, block, lentry->high_word,
1624 rentry->high_word, relation & ~ir_relation_equal);
1625 low = new_rd_Cmp(dbg, block, lentry->low_word, rentry->low_word,
1626 relation);
1627 high = new_rd_Cmp(dbg, block, lentry->high_word, rentry->high_word,
1628 ir_relation_equal);
1629 t = new_rd_And(dbg, block, low, high, mode_b);
1630 res = new_rd_Or(dbg, block, high1, t, mode_b);
1631 }
1632 exchange(cmp, res);
1633 }
1634
1635 /**
1636 * Translate a Conv.
1637 */
lower_Conv(ir_node * node,ir_mode * mode)1638 static void lower_Conv(ir_node *node, ir_mode *mode)
1639 {
1640 mode = get_irn_mode(node);
1641
1642 if (mode == env->high_signed || mode == env->high_unsigned) {
1643 lower_Conv_to_Ll(node);
1644 } else {
1645 ir_mode *op_mode = get_irn_mode(get_Conv_op(node));
1646
1647 if (op_mode == env->high_signed || op_mode == env->high_unsigned) {
1648 lower_Conv_from_Ll(node);
1649 }
1650 }
1651 }
1652
fix_parameter_entities(ir_graph * irg,ir_type * orig_mtp)1653 static void fix_parameter_entities(ir_graph *irg, ir_type *orig_mtp)
1654 {
1655 size_t orig_n_params = get_method_n_params(orig_mtp);
1656 ir_entity **parameter_entities;
1657
1658 parameter_entities = ALLOCANZ(ir_entity*, orig_n_params);
1659
1660 ir_type *frame_type = get_irg_frame_type(irg);
1661 size_t n = get_compound_n_members(frame_type);
1662 size_t i;
1663 size_t n_param;
1664
1665 /* collect parameter entities */
1666 for (i = 0; i < n; ++i) {
1667 ir_entity *entity = get_compound_member(frame_type, i);
1668 size_t p;
1669 if (!is_parameter_entity(entity))
1670 continue;
1671 p = get_entity_parameter_number(entity);
1672 if (p == IR_VA_START_PARAMETER_NUMBER)
1673 continue;
1674 assert(p < orig_n_params);
1675 assert(parameter_entities[p] == NULL);
1676 parameter_entities[p] = entity;
1677 }
1678
1679 /* adjust indices */
1680 n_param = 0;
1681 for (i = 0; i < orig_n_params; ++i, ++n_param) {
1682 ir_entity *entity = parameter_entities[i];
1683 ir_type *tp;
1684
1685 if (entity != NULL)
1686 set_entity_parameter_number(entity, n_param);
1687
1688 tp = get_method_param_type(orig_mtp, i);
1689 if (is_Primitive_type(tp)) {
1690 ir_mode *mode = get_type_mode(tp);
1691 if (mode == env->high_signed || mode == env->high_unsigned) {
1692 ++n_param;
1693 /* note that we do not change the type of the parameter
1694 * entities, as calling convention fixup later still needs to
1695 * know which is/was a lowered doubleword.
1696 * So we just mark/remember it for later */
1697 if (entity != NULL) {
1698 assert(entity->attr.parameter.doubleword_low_mode == NULL);
1699 entity->attr.parameter.doubleword_low_mode
1700 = env->low_unsigned;
1701 }
1702 }
1703 }
1704 }
1705 }
1706
1707 /**
1708 * Lower the method type.
1709 *
1710 * @param env the lower environment
1711 * @param mtp the method type to lower
1712 *
1713 * @return the lowered type
1714 */
lower_mtp(ir_type * mtp)1715 static ir_type *lower_mtp(ir_type *mtp)
1716 {
1717 ir_type *res;
1718 size_t i;
1719 size_t orig_n_params;
1720 size_t orig_n_res;
1721 size_t n_param;
1722 size_t n_res;
1723 bool must_be_lowered;
1724
1725 res = pmap_get(ir_type, lowered_type, mtp);
1726 if (res != NULL)
1727 return res;
1728 if (type_visited(mtp))
1729 return mtp;
1730 mark_type_visited(mtp);
1731
1732 orig_n_params = get_method_n_params(mtp);
1733 orig_n_res = get_method_n_ress(mtp);
1734 n_param = orig_n_params;
1735 n_res = orig_n_res;
1736 must_be_lowered = false;
1737
1738 /* count new number of params */
1739 for (i = orig_n_params; i > 0;) {
1740 ir_type *tp = get_method_param_type(mtp, --i);
1741
1742 if (is_Primitive_type(tp)) {
1743 ir_mode *mode = get_type_mode(tp);
1744
1745 if (mode == env->high_signed || mode == env->high_unsigned) {
1746 ++n_param;
1747 must_be_lowered = true;
1748 }
1749 }
1750 }
1751
1752 /* count new number of results */
1753 for (i = orig_n_res; i > 0;) {
1754 ir_type *tp = get_method_res_type(mtp, --i);
1755
1756 if (is_Primitive_type(tp)) {
1757 ir_mode *mode = get_type_mode(tp);
1758
1759 if (mode == env->high_signed || mode == env->high_unsigned) {
1760 ++n_res;
1761 must_be_lowered = true;
1762 }
1763 }
1764 }
1765 if (!must_be_lowered) {
1766 set_type_link(mtp, NULL);
1767 return mtp;
1768 }
1769
1770 res = new_d_type_method(n_param, n_res, get_type_dbg_info(mtp));
1771
1772 /* set param types and result types */
1773 for (i = n_param = 0; i < orig_n_params; ++i) {
1774 ir_type *tp = get_method_param_type(mtp, i);
1775
1776 if (is_Primitive_type(tp)) {
1777 ir_mode *mode = get_type_mode(tp);
1778
1779 if (mode == env->high_signed) {
1780 if (env->params->little_endian) {
1781 set_method_param_type(res, n_param++, tp_u);
1782 set_method_param_type(res, n_param++, tp_s);
1783 } else {
1784 set_method_param_type(res, n_param++, tp_s);
1785 set_method_param_type(res, n_param++, tp_u);
1786 }
1787 } else if (mode == env->high_unsigned) {
1788 set_method_param_type(res, n_param++, tp_u);
1789 set_method_param_type(res, n_param++, tp_u);
1790 } else {
1791 set_method_param_type(res, n_param, tp);
1792 ++n_param;
1793 }
1794 } else {
1795 set_method_param_type(res, n_param, tp);
1796 ++n_param;
1797 }
1798 }
1799 for (i = n_res = 0; i < orig_n_res; ++i) {
1800 ir_type *tp = get_method_res_type(mtp, i);
1801
1802 if (is_Primitive_type(tp)) {
1803 ir_mode *mode = get_type_mode(tp);
1804
1805 if (mode == env->high_signed) {
1806 if (env->params->little_endian) {
1807 set_method_res_type(res, n_res++, tp_u);
1808 set_method_res_type(res, n_res++, tp_s);
1809 } else {
1810 set_method_res_type(res, n_res++, tp_s);
1811 set_method_res_type(res, n_res++, tp_u);
1812 }
1813 } else if (mode == env->high_unsigned) {
1814 set_method_res_type(res, n_res++, tp_u);
1815 set_method_res_type(res, n_res++, tp_u);
1816 } else {
1817 set_method_res_type(res, n_res++, tp);
1818 }
1819 } else {
1820 set_method_res_type(res, n_res++, tp);
1821 }
1822 }
1823
1824 set_method_variadicity(res, get_method_variadicity(mtp));
1825 set_method_calling_convention(res, get_method_calling_convention(mtp));
1826 set_method_additional_properties(res, get_method_additional_properties(mtp));
1827
1828 set_higher_type(res, mtp);
1829 set_type_link(res, mtp);
1830
1831 mark_type_visited(res);
1832 pmap_insert(lowered_type, mtp, res);
1833 return res;
1834 }
1835
1836 /**
1837 * Translate a Return.
1838 */
lower_Return(ir_node * node,ir_mode * mode)1839 static void lower_Return(ir_node *node, ir_mode *mode)
1840 {
1841 ir_node **in;
1842 size_t i, j, n;
1843 int need_conv = 0;
1844 (void) mode;
1845
1846 /* check if this return must be lowered */
1847 for (i = 0, n = get_Return_n_ress(node); i < n; ++i) {
1848 ir_node *pred = get_Return_res(node, i);
1849 ir_mode *rmode = get_irn_op_mode(pred);
1850
1851 if (rmode == env->high_signed || rmode == env->high_unsigned)
1852 need_conv = 1;
1853 }
1854 if (! need_conv)
1855 return;
1856
1857 ir_graph *irg = get_irn_irg(node);
1858 ir_entity *ent = get_irg_entity(irg);
1859 ir_type *mtp = get_entity_type(ent);
1860
1861 /* create a new in array */
1862 NEW_ARR_A(ir_node *, in, get_method_n_ress(mtp) + 1);
1863 j = 0;
1864 in[j++] = get_Return_mem(node);
1865
1866 for (i = 0, n = get_Return_n_ress(node); i < n; ++i) {
1867 ir_node *pred = get_Return_res(node, i);
1868 ir_mode *pred_mode = get_irn_mode(pred);
1869
1870 if (pred_mode == env->high_signed || pred_mode == env->high_unsigned) {
1871 const lower64_entry_t *entry = get_node_entry(pred);
1872 if (env->params->little_endian) {
1873 in[j++] = entry->low_word;
1874 in[j++] = entry->high_word;
1875 } else {
1876 in[j++] = entry->high_word;
1877 in[j++] = entry->low_word;
1878 }
1879 } else {
1880 in[j++] = pred;
1881 }
1882 }
1883 assert(j == get_method_n_ress(mtp)+1);
1884
1885 set_irn_in(node, j, in);
1886 }
1887
1888 /**
1889 * Translate the parameters.
1890 */
lower_Start(ir_node * node,ir_mode * high_mode)1891 static void lower_Start(ir_node *node, ir_mode *high_mode)
1892 {
1893 ir_graph *irg = get_irn_irg(node);
1894 ir_entity *ent = get_irg_entity(irg);
1895 ir_type *mtp = get_entity_type(ent);
1896 ir_type *orig_mtp = (ir_type*)get_type_link(mtp);
1897 ir_node *args;
1898 long *new_projs;
1899 size_t i, j, n_params;
1900 (void) high_mode;
1901
1902 /* if type link is NULL then the type was not lowered, hence no changes
1903 * at Start necessary */
1904 if (orig_mtp == NULL)
1905 return;
1906
1907 n_params = get_method_n_params(orig_mtp);
1908
1909 NEW_ARR_A(long, new_projs, n_params);
1910
1911 /* Calculate mapping of proj numbers in new_projs */
1912 for (i = j = 0; i < n_params; ++i, ++j) {
1913 ir_type *ptp = get_method_param_type(orig_mtp, i);
1914
1915 new_projs[i] = j;
1916 if (is_Primitive_type(ptp)) {
1917 ir_mode *amode = get_type_mode(ptp);
1918 if (amode == env->high_signed || amode == env->high_unsigned)
1919 ++j;
1920 }
1921 }
1922
1923 /* find args Proj */
1924 args = NULL;
1925 foreach_out_edge(node, edge) {
1926 ir_node *proj = get_edge_src_irn(edge);
1927 if (!is_Proj(proj))
1928 continue;
1929 if (get_Proj_proj(proj) == pn_Start_T_args) {
1930 args = proj;
1931 break;
1932 }
1933 }
1934 if (args == NULL)
1935 return;
1936
1937 /* fix all Proj's and create new ones */
1938 foreach_out_edge_safe(args, edge) {
1939 ir_node *proj = get_edge_src_irn(edge);
1940 ir_mode *mode = get_irn_mode(proj);
1941 ir_mode *mode_l = env->low_unsigned;
1942 ir_node *pred;
1943 long proj_nr;
1944 ir_mode *mode_h;
1945 ir_node *res_low;
1946 ir_node *res_high;
1947 int old_cse;
1948 dbg_info *dbg;
1949
1950 if (!is_Proj(proj))
1951 continue;
1952 pred = get_Proj_pred(proj);
1953 proj_nr = get_Proj_proj(proj);
1954
1955 if (mode == env->high_signed) {
1956 mode_h = env->low_signed;
1957 } else if (mode == env->high_unsigned) {
1958 mode_h = env->low_unsigned;
1959 } else {
1960 long new_pn = new_projs[proj_nr];
1961 set_Proj_proj(proj, new_pn);
1962 continue;
1963 }
1964
1965 /* Switch off CSE or we might get an already existing Proj. */
1966 old_cse = get_opt_cse();
1967 set_opt_cse(0);
1968 dbg = get_irn_dbg_info(proj);
1969 if (env->params->little_endian) {
1970 res_low = new_rd_Proj(dbg, pred, mode_l, new_projs[proj_nr]);
1971 res_high = new_rd_Proj(dbg, pred, mode_h, new_projs[proj_nr] + 1);
1972 } else {
1973 res_high = new_rd_Proj(dbg, pred, mode_h, new_projs[proj_nr]);
1974 res_low = new_rd_Proj(dbg, pred, mode_l, new_projs[proj_nr] + 1);
1975 }
1976 set_opt_cse(old_cse);
1977 ir_set_dw_lowered(proj, res_low, res_high);
1978 }
1979 }
1980
1981 /**
1982 * Translate a Call.
1983 */
lower_Call(ir_node * node,ir_mode * mode)1984 static void lower_Call(ir_node *node, ir_mode *mode)
1985 {
1986 ir_type *tp = get_Call_type(node);
1987 ir_node **in;
1988 size_t n_params, n_res;
1989 bool need_lower = false;
1990 size_t i, j;
1991 size_t p;
1992 long *res_numbers = NULL;
1993 ir_node *resproj;
1994 (void) mode;
1995
1996 n_params = get_method_n_params(tp);
1997 for (p = 0; p < n_params; ++p) {
1998 ir_type *ptp = get_method_param_type(tp, p);
1999
2000 if (is_Primitive_type(ptp)) {
2001 ir_mode *pmode = get_type_mode(ptp);
2002 if (pmode == env->high_signed || pmode == env->high_unsigned) {
2003 need_lower = true;
2004 break;
2005 }
2006 }
2007 }
2008 n_res = get_method_n_ress(tp);
2009 if (n_res > 0) {
2010 NEW_ARR_A(long, res_numbers, n_res);
2011
2012 for (i = j = 0; i < n_res; ++i, ++j) {
2013 ir_type *ptp = get_method_res_type(tp, i);
2014
2015 res_numbers[i] = j;
2016 if (is_Primitive_type(ptp)) {
2017 ir_mode *rmode = get_type_mode(ptp);
2018 if (rmode == env->high_signed || rmode == env->high_unsigned) {
2019 need_lower = true;
2020 ++j;
2021 }
2022 }
2023 }
2024 }
2025
2026 if (! need_lower)
2027 return;
2028
2029 /* let's lower it */
2030 tp = lower_mtp(tp);
2031 set_Call_type(node, tp);
2032
2033 NEW_ARR_A(ir_node *, in, get_method_n_params(tp) + 2);
2034
2035 in[0] = get_Call_mem(node);
2036 in[1] = get_Call_ptr(node);
2037
2038 for (j = 2, i = 0; i < n_params; ++i) {
2039 ir_node *pred = get_Call_param(node, i);
2040 ir_mode *pred_mode = get_irn_mode(pred);
2041
2042 if (pred_mode == env->high_signed || pred_mode == env->high_unsigned) {
2043 const lower64_entry_t *pred_entry = get_node_entry(pred);
2044 if (env->params->little_endian) {
2045 in[j++] = pred_entry->low_word;
2046 in[j++] = pred_entry->high_word;
2047 } else {
2048 in[j++] = pred_entry->high_word;
2049 in[j++] = pred_entry->low_word;
2050 }
2051 } else {
2052 in[j++] = pred;
2053 }
2054 }
2055
2056 set_irn_in(node, j, in);
2057
2058 /* find results T */
2059 resproj = NULL;
2060 foreach_out_edge(node, edge) {
2061 ir_node *proj = get_edge_src_irn(edge);
2062 if (!is_Proj(proj))
2063 continue;
2064 if (get_Proj_proj(proj) == pn_Call_T_result) {
2065 resproj = proj;
2066 break;
2067 }
2068 }
2069 if (resproj == NULL)
2070 return;
2071
2072 /* fix the results */
2073 foreach_out_edge_safe(resproj, edge) {
2074 ir_node *proj = get_edge_src_irn(edge);
2075 ir_mode *proj_mode = get_irn_mode(proj);
2076 ir_mode *mode_l = env->low_unsigned;
2077 ir_node *pred;
2078 long proj_nr;
2079 ir_mode *mode_h;
2080 ir_node *res_low;
2081 ir_node *res_high;
2082 dbg_info *dbg;
2083
2084 if (!is_Proj(proj))
2085 continue;
2086 pred = get_Proj_pred(proj);
2087 proj_nr = get_Proj_proj(proj);
2088
2089 if (proj_mode == env->high_signed) {
2090 mode_h = env->low_signed;
2091 } else if (proj_mode == env->high_unsigned) {
2092 mode_h = env->low_unsigned;
2093 } else {
2094 long new_nr = res_numbers[proj_nr];
2095 set_Proj_proj(proj, new_nr);
2096 continue;
2097 }
2098
2099 dbg = get_irn_dbg_info(proj);
2100 if (env->params->little_endian) {
2101 res_low = new_rd_Proj(dbg, pred, mode_l, res_numbers[proj_nr]);
2102 res_high = new_rd_Proj(dbg, pred, mode_h, res_numbers[proj_nr] + 1);
2103 } else {
2104 res_high = new_rd_Proj(dbg, pred, mode_h, res_numbers[proj_nr]);
2105 res_low = new_rd_Proj(dbg, pred, mode_l, res_numbers[proj_nr] + 1);
2106 }
2107 ir_set_dw_lowered(proj, res_low, res_high);
2108 }
2109 }
2110
2111 /**
2112 * Translate an Unknown into two.
2113 */
lower_Unknown(ir_node * node,ir_mode * mode)2114 static void lower_Unknown(ir_node *node, ir_mode *mode)
2115 {
2116 ir_mode *low_mode = env->low_unsigned;
2117 ir_graph *irg = get_irn_irg(node);
2118 ir_node *res_low = new_r_Unknown(irg, low_mode);
2119 ir_node *res_high = new_r_Unknown(irg, mode);
2120 ir_set_dw_lowered(node, res_low, res_high);
2121 }
2122
2123 /**
2124 * Translate a Bad into two.
2125 */
lower_Bad(ir_node * node,ir_mode * mode)2126 static void lower_Bad(ir_node *node, ir_mode *mode)
2127 {
2128 ir_mode *low_mode = env->low_unsigned;
2129 ir_graph *irg = get_irn_irg(node);
2130 ir_node *res_low = new_r_Bad(irg, low_mode);
2131 ir_node *res_high = new_r_Bad(irg, mode);
2132 ir_set_dw_lowered(node, res_low, res_high);
2133 }
2134
2135 /**
2136 * Translate a Phi.
2137 *
2138 * First step: just create two templates
2139 */
lower_Phi(ir_node * phi)2140 static void lower_Phi(ir_node *phi)
2141 {
2142 ir_mode *mode = get_irn_mode(phi);
2143 int i;
2144 int arity;
2145 ir_node **in_l;
2146 ir_node **in_h;
2147 ir_node *unk_l;
2148 ir_node *unk_h;
2149 ir_node *phi_l;
2150 ir_node *phi_h;
2151 dbg_info *dbg;
2152 ir_node *block;
2153 ir_graph *irg;
2154 ir_mode *mode_l;
2155 ir_mode *mode_h;
2156
2157 /* enqueue predecessors */
2158 arity = get_Phi_n_preds(phi);
2159 for (i = 0; i < arity; ++i) {
2160 ir_node *pred = get_Phi_pred(phi, i);
2161 pdeq_putr(env->waitq, pred);
2162 }
2163
2164 if (mode != env->high_signed && mode != env->high_unsigned)
2165 return;
2166
2167 /* first create a new in array */
2168 NEW_ARR_A(ir_node *, in_l, arity);
2169 NEW_ARR_A(ir_node *, in_h, arity);
2170 irg = get_irn_irg(phi);
2171 mode_l = env->low_unsigned;
2172 mode_h = mode == env->high_signed ? env->low_signed : env->low_unsigned;
2173 unk_l = new_r_Dummy(irg, mode_l);
2174 unk_h = new_r_Dummy(irg, mode_h);
2175 for (i = 0; i < arity; ++i) {
2176 in_l[i] = unk_l;
2177 in_h[i] = unk_h;
2178 }
2179
2180 dbg = get_irn_dbg_info(phi);
2181 block = get_nodes_block(phi);
2182 phi_l = new_rd_Phi(dbg, block, arity, in_l, mode_l);
2183 phi_h = new_rd_Phi(dbg, block, arity, in_h, mode_h);
2184
2185 ir_set_dw_lowered(phi, phi_l, phi_h);
2186
2187 /* remember that we need to fixup the predecessors later */
2188 ARR_APP1(ir_node*, env->lowered_phis, phi);
2189 }
2190
fixup_phi(ir_node * phi)2191 static void fixup_phi(ir_node *phi)
2192 {
2193 const lower64_entry_t *entry = get_node_entry(phi);
2194 ir_node *phi_l = entry->low_word;
2195 ir_node *phi_h = entry->high_word;
2196 int arity = get_Phi_n_preds(phi);
2197 int i;
2198
2199 /* exchange phi predecessors which are lowered by now */
2200 for (i = 0; i < arity; ++i) {
2201 ir_node *pred = get_Phi_pred(phi, i);
2202 const lower64_entry_t *pred_entry = get_node_entry(pred);
2203
2204 set_Phi_pred(phi_l, i, pred_entry->low_word);
2205 set_Phi_pred(phi_h, i, pred_entry->high_word);
2206 }
2207 }
2208
2209 /**
2210 * Translate a Mux.
2211 */
lower_Mux(ir_node * mux,ir_mode * mode)2212 static void lower_Mux(ir_node *mux, ir_mode *mode)
2213 {
2214 ir_node *truen = get_Mux_true(mux);
2215 ir_node *falsen = get_Mux_false(mux);
2216 ir_node *sel = get_Mux_sel(mux);
2217 const lower64_entry_t *true_entry = get_node_entry(truen);
2218 const lower64_entry_t *false_entry = get_node_entry(falsen);
2219 ir_node *true_l = true_entry->low_word;
2220 ir_node *true_h = true_entry->high_word;
2221 ir_node *false_l = false_entry->low_word;
2222 ir_node *false_h = false_entry->high_word;
2223 dbg_info *dbgi = get_irn_dbg_info(mux);
2224 ir_node *block = get_nodes_block(mux);
2225 ir_node *res_low
2226 = new_rd_Mux(dbgi, block, sel, false_l, true_l, env->low_unsigned);
2227 ir_node *res_high
2228 = new_rd_Mux(dbgi, block, sel, false_h, true_h, mode);
2229 ir_set_dw_lowered(mux, res_low, res_high);
2230 }
2231
2232 /**
2233 * Translate an ASM node.
2234 */
lower_ASM(ir_node * asmn,ir_mode * mode)2235 static void lower_ASM(ir_node *asmn, ir_mode *mode)
2236 {
2237 ir_mode *high_signed = env->high_signed;
2238 ir_mode *high_unsigned = env->high_unsigned;
2239 int n_outs = get_ASM_n_output_constraints(asmn);
2240 ir_asm_constraint *output_constraints = get_ASM_output_constraints(asmn);
2241 ir_asm_constraint *input_constraints = get_ASM_input_constraints(asmn);
2242 unsigned n_64bit_outs = 0;
2243
2244 (void)mode;
2245
2246 for (int i = get_irn_arity(asmn) - 1; i >= 0; --i) {
2247 ir_node *op = get_irn_n(asmn, i);
2248 ir_mode *op_mode = get_irn_mode(op);
2249 if (op_mode == high_signed || op_mode == high_unsigned) {
2250 panic("lowering ASM 64bit input unimplemented");
2251 }
2252 }
2253
2254 for (int o = 0; o < n_outs; ++o) {
2255 const ir_asm_constraint *constraint = &output_constraints[o];
2256 if (constraint->mode == high_signed || constraint->mode == high_unsigned) {
2257 const char *constr = get_id_str(constraint->constraint);
2258 ++n_64bit_outs;
2259 /* TODO: How to do this architecture neutral? This is very
2260 * i386 specific... */
2261 if (constr[0] != '=' || constr[1] != 'A') {
2262 panic("lowering ASM 64bit output only supports '=A' currently");
2263 }
2264 }
2265 }
2266
2267 if (n_64bit_outs == 0)
2268 return;
2269
2270 dbg_info *dbgi = get_irn_dbg_info(asmn);
2271 ir_node *block = get_nodes_block(asmn);
2272 ir_node *mem = get_ASM_mem(asmn);
2273 int new_n_outs = 0;
2274 int n_clobber = get_ASM_n_clobbers(asmn);
2275 long *proj_map = ALLOCAN(long, n_outs);
2276 ident **clobbers = get_ASM_clobbers(asmn);
2277 ident *asm_text = get_ASM_text(asmn);
2278 ir_asm_constraint *new_outputs
2279 = ALLOCAN(ir_asm_constraint, n_outs+n_64bit_outs);
2280 ir_node *new_asm;
2281
2282 for (int o = 0; o < n_outs; ++o) {
2283 const ir_asm_constraint *constraint = &output_constraints[o];
2284 if (constraint->mode == high_signed || constraint->mode == high_unsigned) {
2285 new_outputs[new_n_outs].pos = constraint->pos;
2286 new_outputs[new_n_outs].constraint = new_id_from_str("=a");
2287 new_outputs[new_n_outs].mode = env->low_unsigned;
2288 proj_map[o] = new_n_outs;
2289 ++new_n_outs;
2290 new_outputs[new_n_outs].pos = constraint->pos;
2291 new_outputs[new_n_outs].constraint = new_id_from_str("=d");
2292 if (constraint->mode == high_signed)
2293 new_outputs[new_n_outs].mode = env->low_signed;
2294 else
2295 new_outputs[new_n_outs].mode = env->low_unsigned;
2296 ++new_n_outs;
2297 } else {
2298 new_outputs[new_n_outs] = *constraint;
2299 proj_map[o] = new_n_outs;
2300 ++new_n_outs;
2301 }
2302 }
2303 assert(new_n_outs == n_outs+(int)n_64bit_outs);
2304
2305 int n_inputs = get_ASM_n_inputs(asmn);
2306 ir_node **new_ins = ALLOCAN(ir_node*, n_inputs);
2307 for (int i = 0; i < n_inputs; ++i)
2308 new_ins[i] = get_ASM_input(asmn, i);
2309
2310 new_asm = new_rd_ASM(dbgi, block, mem, n_inputs, new_ins, input_constraints,
2311 new_n_outs, new_outputs, n_clobber, clobbers,
2312 asm_text);
2313
2314 foreach_out_edge_safe(asmn, edge) {
2315 ir_node *proj = get_edge_src_irn(edge);
2316 ir_mode *proj_mode = get_irn_mode(proj);
2317 long pn;
2318
2319 if (!is_Proj(proj))
2320 continue;
2321 pn = get_Proj_proj(proj);
2322
2323 if (pn < n_outs)
2324 pn = proj_map[pn];
2325 else
2326 pn = new_n_outs + pn - n_outs;
2327
2328 if (proj_mode == high_signed || proj_mode == high_unsigned) {
2329 ir_mode *high_mode
2330 = proj_mode == high_signed ? env->low_signed : env->low_unsigned;
2331 ir_node *np_low = new_r_Proj(new_asm, env->low_unsigned, pn);
2332 ir_node *np_high = new_r_Proj(new_asm, high_mode, pn+1);
2333 ir_set_dw_lowered(proj, np_low, np_high);
2334 } else {
2335 ir_node *np = new_r_Proj(new_asm, proj_mode, pn);
2336 exchange(proj, np);
2337 }
2338 }
2339 }
2340
2341 /**
2342 * Lower the builtin type to its higher part.
2343 *
2344 * @param mtp the builtin type to lower
2345 *
2346 * @return the lowered type
2347 */
lower_Builtin_type_high(ir_type * mtp)2348 static ir_type *lower_Builtin_type_high(ir_type *mtp)
2349 {
2350 ir_type *res;
2351 size_t i;
2352 size_t n_params;
2353 size_t n_results;
2354 bool must_be_lowered;
2355
2356 res = pmap_get(ir_type, lowered_builtin_type_high, mtp);
2357 if (res != NULL)
2358 return res;
2359
2360 n_params = get_method_n_params(mtp);
2361 n_results = get_method_n_ress(mtp);
2362 must_be_lowered = false;
2363
2364 /* check for double word parameter */
2365 for (i = n_params; i > 0;) {
2366 ir_type *tp = get_method_param_type(mtp, --i);
2367
2368 if (is_Primitive_type(tp)) {
2369 ir_mode *mode = get_type_mode(tp);
2370
2371 if (mode == env->high_signed || mode == env->high_unsigned) {
2372 must_be_lowered = true;
2373 break;
2374 }
2375 }
2376 }
2377
2378 if (!must_be_lowered) {
2379 set_type_link(mtp, NULL);
2380 return mtp;
2381 }
2382
2383 res = new_d_type_method(n_params, n_results, get_type_dbg_info(mtp));
2384
2385 /* set param types and result types */
2386 for (i = 0; i < n_params; ++i) {
2387 ir_type *tp = get_method_param_type(mtp, i);
2388
2389 if (is_Primitive_type(tp)) {
2390 ir_mode *mode = get_type_mode(tp);
2391
2392 if (mode == env->high_signed) {
2393 if (env->params->little_endian) {
2394 set_method_param_type(res, i, tp_u);
2395 } else {
2396 set_method_param_type(res, i, tp_s);
2397 }
2398 } else if (mode == env->high_unsigned) {
2399 set_method_param_type(res, i, tp_u);
2400 } else {
2401 set_method_param_type(res, i, tp);
2402 }
2403 } else {
2404 set_method_param_type(res, i, tp);
2405 }
2406 }
2407 for (i = n_results = 0; i < n_results; ++i) {
2408 ir_type *tp = get_method_res_type(mtp, i);
2409
2410 if (is_Primitive_type(tp)) {
2411 ir_mode *mode = get_type_mode(tp);
2412
2413 if (mode == env->high_signed) {
2414 if (env->params->little_endian) {
2415 set_method_res_type(res, i, tp_u);
2416 } else {
2417 set_method_res_type(res, i, tp_s);
2418 }
2419 } else if (mode == env->high_unsigned) {
2420 set_method_res_type(res, i, tp_u);
2421 } else {
2422 set_method_res_type(res, i, tp);
2423 }
2424 } else {
2425 set_method_res_type(res, i, tp);
2426 }
2427 }
2428
2429 set_method_variadicity(res, get_method_variadicity(mtp));
2430 set_method_calling_convention(res, get_method_calling_convention(mtp));
2431 set_method_additional_properties(res, get_method_additional_properties(mtp));
2432
2433 pmap_insert(lowered_builtin_type_high, mtp, res);
2434 return res;
2435 }
2436
2437 /**
2438 * Lower the builtin type to its lower part.
2439 *
2440 * @param mtp the builtin type to lower
2441 *
2442 * @return the lowered type
2443 */
lower_Builtin_type_low(ir_type * mtp)2444 static ir_type *lower_Builtin_type_low(ir_type *mtp)
2445 {
2446 ir_type *res;
2447 size_t i;
2448 size_t n_params;
2449 size_t n_results;
2450 bool must_be_lowered;
2451
2452 res = pmap_get(ir_type, lowered_builtin_type_low, mtp);
2453 if (res != NULL)
2454 return res;
2455
2456 n_params = get_method_n_params(mtp);
2457 n_results = get_method_n_ress(mtp);
2458 must_be_lowered = false;
2459
2460 /* check for double word parameter */
2461 for (i = n_params; i > 0;) {
2462 ir_type *tp = get_method_param_type(mtp, --i);
2463
2464 if (is_Primitive_type(tp)) {
2465 ir_mode *mode = get_type_mode(tp);
2466
2467 if (mode == env->high_signed || mode == env->high_unsigned) {
2468 must_be_lowered = true;
2469 break;
2470 }
2471 }
2472 }
2473
2474 if (!must_be_lowered) {
2475 set_type_link(mtp, NULL);
2476 return mtp;
2477 }
2478
2479 res = new_d_type_method(n_params, n_results, get_type_dbg_info(mtp));
2480
2481 /* set param types and result types */
2482 for (i = 0; i < n_params; ++i) {
2483 ir_type *tp = get_method_param_type(mtp, i);
2484
2485 if (is_Primitive_type(tp)) {
2486 ir_mode *mode = get_type_mode(tp);
2487
2488 if (mode == env->high_signed) {
2489 if (env->params->little_endian) {
2490 set_method_param_type(res, i, tp_s);
2491 } else {
2492 set_method_param_type(res, i, tp_u);
2493 }
2494 } else if (mode == env->high_unsigned) {
2495 set_method_param_type(res, i, tp_u);
2496 } else {
2497 set_method_param_type(res, i, tp);
2498 }
2499 } else {
2500 set_method_param_type(res, i, tp);
2501 }
2502 }
2503 for (i = 0; i < n_results; ++i) {
2504 ir_type *tp = get_method_res_type(mtp, i);
2505
2506 if (is_Primitive_type(tp)) {
2507 ir_mode *mode = get_type_mode(tp);
2508
2509 if (mode == env->high_signed) {
2510 if (env->params->little_endian) {
2511 set_method_res_type(res, i, tp_s);
2512 } else {
2513 set_method_res_type(res, i, tp_u);
2514 }
2515 } else if (mode == env->high_unsigned) {
2516 set_method_res_type(res, i, tp_u);
2517 } else {
2518 set_method_res_type(res, i, tp);
2519 }
2520 } else {
2521 set_method_res_type(res, i, tp);
2522 }
2523 }
2524
2525 set_method_variadicity(res, get_method_variadicity(mtp));
2526 set_method_calling_convention(res, get_method_calling_convention(mtp));
2527 set_method_additional_properties(res, get_method_additional_properties(mtp));
2528
2529 pmap_insert(lowered_builtin_type_low, mtp, res);
2530 return res;
2531 }
2532
2533 /**
2534 * lowers a builtin which reduces a 64bit value to a simple summary value
2535 * (popcount, ffs, ...)
2536 */
lower_reduce_builtin(ir_node * builtin,ir_mode * mode)2537 static void lower_reduce_builtin(ir_node *builtin, ir_mode *mode)
2538 {
2539 ir_builtin_kind kind = get_Builtin_kind(builtin);
2540 ir_node *operand = get_Builtin_param(builtin, 0);
2541 ir_mode *operand_mode = get_irn_mode(operand);
2542 if (operand_mode != env->high_signed && operand_mode != env->high_unsigned)
2543 return;
2544
2545 {
2546 arch_allow_ifconv_func allow_ifconv = be_get_backend_param()->allow_ifconv;
2547 int arity = get_irn_arity(builtin);
2548 dbg_info *dbgi = get_irn_dbg_info(builtin);
2549 ir_graph *irg = get_irn_irg(builtin);
2550 ir_type *type = get_Builtin_type(builtin);
2551 ir_type *lowered_type_high = lower_Builtin_type_high(type);
2552 ir_type *lowered_type_low = lower_Builtin_type_low(type);
2553 ir_type *result_type = get_method_res_type(lowered_type_low, 0);
2554 ir_mode *result_mode = get_type_mode(result_type);
2555 ir_node *block = get_nodes_block(builtin);
2556 ir_node *mem = get_Builtin_mem(builtin);
2557 const lower64_entry_t *entry = get_node_entry(operand);
2558 ir_mode *high_mode = get_irn_mode(entry->high_word);
2559 ir_node *in_high[1] = {entry->high_word};
2560 ir_node *in_low[1] = {entry->low_word};
2561 ir_node *res;
2562
2563 assert(is_NoMem(mem));
2564 assert(arity == 2);
2565
2566 switch (kind) {
2567 case ir_bk_ffs: {
2568 ir_node *number_of_bits = new_r_Const_long(irg, result_mode, get_mode_size_bits(env->low_unsigned));
2569 ir_node *zero_high = new_rd_Const(dbgi, irg, get_mode_null(high_mode));
2570 ir_node *zero_unsigned = new_rd_Const(dbgi, irg, get_mode_null(env->low_unsigned));
2571 ir_node *zero_result = new_rd_Const(dbgi, irg, get_mode_null(result_mode));
2572 ir_node *cmp_low = new_rd_Cmp(dbgi, block, entry->low_word, zero_unsigned, ir_relation_equal);
2573 ir_node *cmp_high = new_rd_Cmp(dbgi, block, entry->high_word, zero_high, ir_relation_equal);
2574 ir_node *ffs_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2575 ir_node *high_proj = new_r_Proj(ffs_high, result_mode, pn_Builtin_max+1);
2576 ir_node *high = new_rd_Add(dbgi, block, high_proj, number_of_bits, result_mode);
2577 ir_node *ffs_low = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2578 ir_node *low = new_r_Proj(ffs_low, result_mode, pn_Builtin_max+1);
2579 ir_node *mux_high = new_rd_Mux(dbgi, block, cmp_high, high, zero_result, result_mode);
2580
2581 if (! allow_ifconv(cmp_high, high, zero_result))
2582 ir_nodeset_insert(&created_mux_nodes, mux_high);
2583
2584 res = new_rd_Mux(dbgi, block, cmp_low, low, mux_high, result_mode);
2585
2586 if (! allow_ifconv(cmp_low, low, mux_high))
2587 ir_nodeset_insert(&created_mux_nodes, res);
2588 break;
2589 }
2590 case ir_bk_clz: {
2591 ir_node *zero = new_rd_Const(dbgi, irg, get_mode_null(high_mode));
2592 ir_node *cmp_high = new_rd_Cmp(dbgi, block, entry->high_word, zero, ir_relation_equal);
2593 ir_node *clz_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2594 ir_node *high = new_r_Proj(clz_high, result_mode, pn_Builtin_max+1);
2595 ir_node *clz_low = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2596 ir_node *low_proj = new_r_Proj(clz_low, result_mode, pn_Builtin_max+1);
2597 ir_node *number_of_bits = new_r_Const_long(irg, result_mode, get_mode_size_bits(mode));
2598 ir_node *low = new_rd_Add(dbgi, block, low_proj, number_of_bits, result_mode);
2599
2600 res = new_rd_Mux(dbgi, block, cmp_high, high, low, result_mode);
2601
2602 if (! allow_ifconv(cmp_high, high, low))
2603 ir_nodeset_insert(&created_mux_nodes, res);
2604 break;
2605 }
2606 case ir_bk_ctz: {
2607 ir_node *zero_unsigned = new_rd_Const(dbgi, irg, get_mode_null(env->low_unsigned));
2608 ir_node *cmp_low = new_rd_Cmp(dbgi, block, entry->low_word, zero_unsigned, ir_relation_equal);
2609 ir_node *ffs_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2610 ir_node *high_proj = new_r_Proj(ffs_high, result_mode, pn_Builtin_max+1);
2611 ir_node *number_of_bits = new_r_Const_long(irg, result_mode, get_mode_size_bits(env->low_unsigned));
2612 ir_node *high = new_rd_Add(dbgi, block, high_proj, number_of_bits, result_mode);
2613 ir_node *ffs_low = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2614 ir_node *low = new_r_Proj(ffs_low, result_mode, pn_Builtin_max+1);
2615
2616 res = new_rd_Mux(dbgi, block, cmp_low, low, high, result_mode);
2617
2618 if (! allow_ifconv(cmp_low, low, high))
2619 ir_nodeset_insert(&created_mux_nodes, res);
2620 break;
2621 }
2622 case ir_bk_popcount: {
2623 ir_node *popcount_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2624 ir_node *popcount_low = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2625 ir_node *high = new_r_Proj(popcount_high, result_mode, pn_Builtin_max+1);
2626 ir_node *low = new_r_Proj(popcount_low, result_mode, pn_Builtin_max+1);
2627
2628 res = new_rd_Add(dbgi, block, high, low, result_mode);
2629 break;
2630 }
2631 case ir_bk_parity: {
2632 ir_node *parity_high;
2633 ir_node *parity_low;
2634 ir_node *high;
2635 ir_node *low;
2636
2637 assert(arity == 2);
2638
2639 parity_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2640 high = new_r_Proj(parity_high, result_mode, pn_Builtin_max+1);
2641 parity_low = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2642 low = new_r_Proj(parity_low, result_mode, pn_Builtin_max+1);
2643 res = new_rd_Eor(dbgi, block, high, low, result_mode);
2644 break;
2645 }
2646 default:
2647 panic("unexpected builtin");
2648 }
2649
2650 turn_into_tuple(builtin, 2);
2651 set_irn_n(builtin, pn_Builtin_M, mem);
2652 set_irn_n(builtin, pn_Builtin_max+1, res);
2653 }
2654 }
2655
2656 /**
2657 * lowers builtins performing arithmetic (bswap)
2658 */
lower_arithmetic_builtin(ir_node * builtin,ir_mode * mode)2659 static void lower_arithmetic_builtin(ir_node *builtin, ir_mode *mode)
2660 {
2661 ir_builtin_kind kind = get_Builtin_kind(builtin);
2662 ir_node *operand = get_Builtin_param(builtin, 0);
2663 ir_mode *operand_mode = get_irn_mode(operand);
2664 (void) mode;
2665 if (operand_mode != env->high_signed && operand_mode != env->high_unsigned)
2666 return;
2667
2668 {
2669 dbg_info *dbgi = get_irn_dbg_info(builtin);
2670 ir_type *type = get_Builtin_type(builtin);
2671 ir_type *lowered_type_high = lower_Builtin_type_high(type);
2672 ir_type *lowered_type_low = lower_Builtin_type_low(type);
2673 ir_node *block = get_nodes_block(builtin);
2674 ir_node *mem = get_Builtin_mem(builtin);
2675 const lower64_entry_t *entry = get_node_entry(operand);
2676 ir_mode *mode_high = get_irn_mode(entry->high_word);
2677 ir_node *res_high;
2678 ir_node *res_low;
2679
2680 switch (kind) {
2681 case ir_bk_bswap: {
2682 ir_node *in_high[1] = { entry->high_word };
2683 ir_node *in_low[1] = { entry->low_word };
2684 ir_node *swap_high = new_rd_Builtin(dbgi, block, mem, 1, in_high, kind, lowered_type_high);
2685 ir_node *swap_low = new_rd_Builtin(dbgi, block, mem, 1, in_low, kind, lowered_type_low);
2686 ir_node *high = new_r_Proj(swap_high, mode_high, pn_Builtin_max+1);
2687 ir_node *low = new_r_Proj(swap_low, env->low_unsigned, pn_Builtin_max+1);
2688 if (mode_high == env->low_signed) {
2689 res_high = new_rd_Conv(dbgi, block, low, env->low_signed);
2690 res_low = new_rd_Conv(dbgi, block, high, env->low_unsigned);
2691 } else {
2692 res_high = low;
2693 res_low = high;
2694 }
2695 break;
2696 }
2697 default:
2698 panic("unexpected builtin");
2699 }
2700
2701 /* search result Proj */
2702 foreach_out_edge_safe(builtin, edge) {
2703 ir_node *proj = get_edge_src_irn(edge);
2704 if (!is_Proj(proj))
2705 continue;
2706
2707 if (get_Proj_proj(proj) == pn_Builtin_max+1) {
2708 ir_set_dw_lowered(proj, res_low, res_high);
2709 }
2710 }
2711 }
2712 }
2713
2714 /**
2715 * Lower double word builtins.
2716 */
lower_Builtin(ir_node * builtin,ir_mode * mode)2717 static void lower_Builtin(ir_node *builtin, ir_mode *mode)
2718 {
2719 ir_builtin_kind kind = get_Builtin_kind(builtin);
2720
2721 switch (kind) {
2722 case ir_bk_trap:
2723 case ir_bk_debugbreak:
2724 case ir_bk_return_address:
2725 case ir_bk_frame_address:
2726 case ir_bk_prefetch:
2727 case ir_bk_inport:
2728 case ir_bk_outport:
2729 case ir_bk_inner_trampoline:
2730 /* Nothing to do. */
2731 return;
2732 case ir_bk_bswap:
2733 lower_arithmetic_builtin(builtin, mode);
2734 return;
2735 case ir_bk_ffs:
2736 case ir_bk_clz:
2737 case ir_bk_ctz:
2738 case ir_bk_popcount:
2739 case ir_bk_parity:
2740 lower_reduce_builtin(builtin, mode);
2741 return;
2742 }
2743 panic("unknown builtin");
2744 }
2745
2746 /**
2747 * check for opcodes that must always be lowered.
2748 */
always_lower(unsigned code)2749 static bool always_lower(unsigned code)
2750 {
2751 switch (code) {
2752 case iro_ASM:
2753 case iro_Builtin:
2754 case iro_Proj:
2755 case iro_Start:
2756 case iro_Call:
2757 case iro_Return:
2758 case iro_Cond:
2759 case iro_Switch:
2760 case iro_Conv:
2761 case iro_Sel:
2762 return true;
2763 default:
2764 return false;
2765 }
2766 }
2767
2768 /**
2769 * Compare two op_mode_entry_t's.
2770 */
cmp_op_mode(const void * elt,const void * key,size_t size)2771 static int cmp_op_mode(const void *elt, const void *key, size_t size)
2772 {
2773 const op_mode_entry_t *e1 = (const op_mode_entry_t*)elt;
2774 const op_mode_entry_t *e2 = (const op_mode_entry_t*)key;
2775 (void) size;
2776
2777 return (e1->op != e2->op) | (e1->imode != e2->imode) | (e1->omode != e2->omode);
2778 }
2779
2780 /**
2781 * Compare two conv_tp_entry_t's.
2782 */
cmp_conv_tp(const void * elt,const void * key,size_t size)2783 static int cmp_conv_tp(const void *elt, const void *key, size_t size)
2784 {
2785 const conv_tp_entry_t *e1 = (const conv_tp_entry_t*)elt;
2786 const conv_tp_entry_t *e2 = (const conv_tp_entry_t*)key;
2787 (void) size;
2788
2789 return (e1->imode != e2->imode) | (e1->omode != e2->omode);
2790 }
2791
2792 /**
2793 * Enter a lowering function into an ir_op.
2794 */
ir_register_dw_lower_function(ir_op * op,lower_dw_func func)2795 void ir_register_dw_lower_function(ir_op *op, lower_dw_func func)
2796 {
2797 op->ops.generic = (op_func)func;
2798 }
2799
2800 /* Determine which modes need to be lowered */
setup_modes(void)2801 static void setup_modes(void)
2802 {
2803 unsigned size_bits = env->params->doubleword_size;
2804 ir_mode *doubleword_signed = NULL;
2805 ir_mode *doubleword_unsigned = NULL;
2806 size_t n_modes = ir_get_n_modes();
2807 ir_mode_arithmetic arithmetic;
2808 unsigned modulo_shift;
2809 size_t i;
2810
2811 /* search for doubleword modes... */
2812 for (i = 0; i < n_modes; ++i) {
2813 ir_mode *mode = ir_get_mode(i);
2814 if (!mode_is_int(mode))
2815 continue;
2816 if (get_mode_size_bits(mode) != size_bits)
2817 continue;
2818 if (mode_is_signed(mode)) {
2819 if (doubleword_signed != NULL) {
2820 /* sigh - the lowerer should really just lower all mode with
2821 * size_bits it finds. Unfortunately this required a bigger
2822 * rewrite. */
2823 panic("multiple double word signed modes found");
2824 }
2825 doubleword_signed = mode;
2826 } else {
2827 if (doubleword_unsigned != NULL) {
2828 /* sigh - the lowerer should really just lower all mode with
2829 * size_bits it finds. Unfortunately this required a bigger
2830 * rewrite. */
2831 panic("multiple double word unsigned modes found");
2832 }
2833 doubleword_unsigned = mode;
2834 }
2835 }
2836 if (doubleword_signed == NULL || doubleword_unsigned == NULL) {
2837 panic("Couldn't find doubleword modes");
2838 }
2839
2840 arithmetic = get_mode_arithmetic(doubleword_signed);
2841 modulo_shift = get_mode_modulo_shift(doubleword_signed);
2842
2843 assert(get_mode_size_bits(doubleword_unsigned) == size_bits);
2844 assert(size_bits % 2 == 0);
2845 assert(get_mode_sign(doubleword_signed) == 1);
2846 assert(get_mode_sign(doubleword_unsigned) == 0);
2847 assert(get_mode_sort(doubleword_signed) == irms_int_number);
2848 assert(get_mode_sort(doubleword_unsigned) == irms_int_number);
2849 assert(get_mode_arithmetic(doubleword_unsigned) == arithmetic);
2850 assert(get_mode_modulo_shift(doubleword_unsigned) == modulo_shift);
2851
2852 /* try to guess a sensible modulo shift for the new mode.
2853 * (This is IMO another indication that this should really be a node
2854 * attribute instead of a mode thing) */
2855 if (modulo_shift == size_bits) {
2856 modulo_shift = modulo_shift / 2;
2857 } else if (modulo_shift == 0) {
2858 /* fine */
2859 } else {
2860 panic("Don't know what new modulo shift to use for lowered doubleword mode");
2861 }
2862 size_bits /= 2;
2863
2864 /* produce lowered modes */
2865 env->high_signed = doubleword_signed;
2866 env->high_unsigned = doubleword_unsigned;
2867 env->low_signed = new_int_mode("WS", arithmetic, size_bits, 1,
2868 modulo_shift);
2869 env->low_unsigned = new_int_mode("WU", arithmetic, size_bits, 0,
2870 modulo_shift);
2871 }
2872
enqueue_preds(ir_node * node)2873 static void enqueue_preds(ir_node *node)
2874 {
2875 int arity = get_irn_arity(node);
2876 int i;
2877
2878 for (i = 0; i < arity; ++i) {
2879 ir_node *pred = get_irn_n(node, i);
2880 pdeq_putr(env->waitq, pred);
2881 }
2882 }
2883
lower_node(ir_node * node)2884 static void lower_node(ir_node *node)
2885 {
2886 int arity;
2887 int i;
2888 lower_dw_func func;
2889 ir_op *op;
2890 ir_mode *mode;
2891 unsigned idx;
2892 lower64_entry_t *entry;
2893
2894 if (irn_visited_else_mark(node))
2895 return;
2896
2897 /* cycles are always broken at Phi and Block nodes. So we don't need special
2898 * magic in all the other lower functions */
2899 if (is_Block(node)) {
2900 enqueue_preds(node);
2901 return;
2902 } else if (is_Phi(node)) {
2903 lower_Phi(node);
2904 return;
2905 }
2906
2907 /* depth-first: descend into operands */
2908 if (!is_Block(node)) {
2909 ir_node *block = get_nodes_block(node);
2910 lower_node(block);
2911 }
2912
2913 if (!is_Cond(node)) {
2914 arity = get_irn_arity(node);
2915 for (i = 0; i < arity; ++i) {
2916 ir_node *pred = get_irn_n(node, i);
2917 lower_node(pred);
2918 }
2919 }
2920
2921 op = get_irn_op(node);
2922 func = (lower_dw_func) op->ops.generic;
2923 if (func == NULL)
2924 return;
2925
2926 idx = get_irn_idx(node);
2927 entry = idx < env->n_entries ? env->entries[idx] : NULL;
2928 if (entry != NULL || always_lower(get_irn_opcode(node))) {
2929 mode = get_irn_op_mode(node);
2930 if (mode == env->high_signed) {
2931 mode = env->low_signed;
2932 } else {
2933 mode = env->low_unsigned;
2934 }
2935 DB((dbg, LEVEL_1, " %+F\n", node));
2936 func(node, mode);
2937 }
2938 }
2939
clear_node_and_phi_links(ir_node * node,void * data)2940 static void clear_node_and_phi_links(ir_node *node, void *data)
2941 {
2942 (void) data;
2943 if (get_irn_mode(node) == mode_T) {
2944 set_irn_link(node, node);
2945 } else {
2946 set_irn_link(node, NULL);
2947 }
2948 if (is_Block(node))
2949 set_Block_phis(node, NULL);
2950 else if (is_Phi(node))
2951 set_Phi_next(node, NULL);
2952 }
2953
lower_irg(ir_graph * irg)2954 static void lower_irg(ir_graph *irg)
2955 {
2956 ir_entity *ent;
2957 ir_type *mtp;
2958 ir_type *lowered_mtp;
2959 unsigned n_idx;
2960
2961 obstack_init(&env->obst);
2962
2963 /* just here for debugging */
2964 current_ir_graph = irg;
2965 assure_edges(irg);
2966
2967 n_idx = get_irg_last_idx(irg);
2968 n_idx = n_idx + (n_idx >> 2); /* add 25% */
2969 env->n_entries = n_idx;
2970 env->entries = NEW_ARR_F(lower64_entry_t*, n_idx);
2971 memset(env->entries, 0, sizeof(env->entries[0]) * n_idx);
2972
2973 env->irg = irg;
2974 env->flags = 0;
2975
2976 ent = get_irg_entity(irg);
2977 mtp = get_entity_type(ent);
2978 lowered_mtp = lower_mtp(mtp);
2979
2980 if (lowered_mtp != mtp) {
2981 set_entity_type(ent, lowered_mtp);
2982 env->flags |= MUST_BE_LOWERED;
2983
2984 fix_parameter_entities(irg, mtp);
2985 }
2986
2987 /* first step: link all nodes and allocate data */
2988 ir_reserve_resources(irg, IR_RESOURCE_PHI_LIST | IR_RESOURCE_IRN_LINK);
2989 visit_all_identities(irg, clear_node_and_phi_links, NULL);
2990 irg_walk_graph(irg, NULL, prepare_links_and_handle_rotl, env);
2991
2992 if (env->flags & MUST_BE_LOWERED) {
2993 size_t i;
2994 ir_reserve_resources(irg, IR_RESOURCE_IRN_VISITED);
2995 inc_irg_visited(irg);
2996
2997 assert(pdeq_empty(env->waitq));
2998 pdeq_putr(env->waitq, get_irg_end(irg));
2999
3000 env->lowered_phis = NEW_ARR_F(ir_node*, 0);
3001 while (!pdeq_empty(env->waitq)) {
3002 ir_node *node = (ir_node*)pdeq_getl(env->waitq);
3003 lower_node(node);
3004 }
3005
3006 /* we need to fixup phis */
3007 for (i = 0; i < ARR_LEN(env->lowered_phis); ++i) {
3008 ir_node *phi = env->lowered_phis[i];
3009 fixup_phi(phi);
3010 }
3011 DEL_ARR_F(env->lowered_phis);
3012
3013
3014 ir_free_resources(irg, IR_RESOURCE_IRN_VISITED);
3015
3016 if (env->flags & CF_CHANGED) {
3017 /* control flow changed, dominance info is invalid */
3018 clear_irg_properties(irg, IR_GRAPH_PROPERTY_CONSISTENT_DOMINANCE);
3019 }
3020 edges_deactivate(irg);
3021 }
3022
3023 ir_free_resources(irg, IR_RESOURCE_PHI_LIST | IR_RESOURCE_IRN_LINK);
3024
3025 DEL_ARR_F(env->entries);
3026 obstack_free(&env->obst, NULL);
3027 }
3028
3029 static const lwrdw_param_t *param;
3030
ir_prepare_dw_lowering(const lwrdw_param_t * new_param)3031 void ir_prepare_dw_lowering(const lwrdw_param_t *new_param)
3032 {
3033 assert(new_param != NULL);
3034 FIRM_DBG_REGISTER(dbg, "firm.lower.dw");
3035
3036 param = new_param;
3037
3038 ir_clear_opcodes_generic_func();
3039 ir_register_dw_lower_function(op_ASM, lower_ASM);
3040 ir_register_dw_lower_function(op_Add, lower_binop);
3041 ir_register_dw_lower_function(op_And, lower_And);
3042 ir_register_dw_lower_function(op_Bad, lower_Bad);
3043 ir_register_dw_lower_function(op_Builtin, lower_Builtin);
3044 ir_register_dw_lower_function(op_Call, lower_Call);
3045 ir_register_dw_lower_function(op_Cmp, lower_Cmp);
3046 ir_register_dw_lower_function(op_Cond, lower_Cond);
3047 ir_register_dw_lower_function(op_Const, lower_Const);
3048 ir_register_dw_lower_function(op_Conv, lower_Conv);
3049 ir_register_dw_lower_function(op_Div, lower_Div);
3050 ir_register_dw_lower_function(op_Eor, lower_Eor);
3051 ir_register_dw_lower_function(op_Load, lower_Load);
3052 ir_register_dw_lower_function(op_Minus, lower_unop);
3053 ir_register_dw_lower_function(op_Mod, lower_Mod);
3054 ir_register_dw_lower_function(op_Mul, lower_binop);
3055 ir_register_dw_lower_function(op_Mux, lower_Mux);
3056 ir_register_dw_lower_function(op_Not, lower_Not);
3057 ir_register_dw_lower_function(op_Or, lower_Or);
3058 ir_register_dw_lower_function(op_Proj, lower_Proj);
3059 ir_register_dw_lower_function(op_Return, lower_Return);
3060 ir_register_dw_lower_function(op_Shl, lower_Shl);
3061 ir_register_dw_lower_function(op_Shr, lower_Shr);
3062 ir_register_dw_lower_function(op_Shrs, lower_Shrs);
3063 ir_register_dw_lower_function(op_Start, lower_Start);
3064 ir_register_dw_lower_function(op_Store, lower_Store);
3065 ir_register_dw_lower_function(op_Sub, lower_binop);
3066 ir_register_dw_lower_function(op_Switch, lower_Switch);
3067 ir_register_dw_lower_function(op_Unknown, lower_Unknown);
3068 }
3069
3070 /**
3071 * Callback to lower only the Mux nodes we created.
3072 */
lower_mux_cb(ir_node * mux)3073 static int lower_mux_cb(ir_node *mux)
3074 {
3075 return ir_nodeset_contains(&created_mux_nodes, mux);
3076 }
3077
3078 /*
3079 * Do the lowering.
3080 */
ir_lower_dw_ops(void)3081 void ir_lower_dw_ops(void)
3082 {
3083 lower_dw_env_t lenv;
3084 size_t i, n;
3085
3086 memset(&lenv, 0, sizeof(lenv));
3087 lenv.params = param;
3088 env = &lenv;
3089
3090 setup_modes();
3091
3092 /* create the necessary maps */
3093 if (! intrinsic_fkt)
3094 intrinsic_fkt = new_set(cmp_op_mode, iro_Last + 1);
3095 if (! conv_types)
3096 conv_types = new_set(cmp_conv_tp, 16);
3097 if (! lowered_type)
3098 lowered_type = pmap_create();
3099 if (! lowered_builtin_type_low)
3100 lowered_builtin_type_low = pmap_create();
3101 if (! lowered_builtin_type_high)
3102 lowered_builtin_type_high = pmap_create();
3103
3104 /* create a primitive unsigned and signed type */
3105 if (! tp_u)
3106 tp_u = get_type_for_mode(lenv.low_unsigned);
3107 if (! tp_s)
3108 tp_s = get_type_for_mode(lenv.low_signed);
3109
3110 /* create method types for the created binop calls */
3111 if (! binop_tp_u) {
3112 binop_tp_u = new_type_method(4, 2);
3113 set_method_param_type(binop_tp_u, 0, tp_u);
3114 set_method_param_type(binop_tp_u, 1, tp_u);
3115 set_method_param_type(binop_tp_u, 2, tp_u);
3116 set_method_param_type(binop_tp_u, 3, tp_u);
3117 set_method_res_type(binop_tp_u, 0, tp_u);
3118 set_method_res_type(binop_tp_u, 1, tp_u);
3119 }
3120 if (! binop_tp_s) {
3121 binop_tp_s = new_type_method(4, 2);
3122 if (env->params->little_endian) {
3123 set_method_param_type(binop_tp_s, 0, tp_u);
3124 set_method_param_type(binop_tp_s, 1, tp_s);
3125 set_method_param_type(binop_tp_s, 2, tp_u);
3126 set_method_param_type(binop_tp_s, 3, tp_s);
3127 set_method_res_type(binop_tp_s, 0, tp_u);
3128 set_method_res_type(binop_tp_s, 1, tp_s);
3129 } else {
3130 set_method_param_type(binop_tp_s, 0, tp_s);
3131 set_method_param_type(binop_tp_s, 1, tp_u);
3132 set_method_param_type(binop_tp_s, 2, tp_s);
3133 set_method_param_type(binop_tp_s, 3, tp_u);
3134 set_method_res_type(binop_tp_s, 0, tp_s);
3135 set_method_res_type(binop_tp_s, 1, tp_u);
3136 }
3137 }
3138 if (! unop_tp_u) {
3139 unop_tp_u = new_type_method(2, 2);
3140 set_method_param_type(unop_tp_u, 0, tp_u);
3141 set_method_param_type(unop_tp_u, 1, tp_u);
3142 set_method_res_type(unop_tp_u, 0, tp_u);
3143 set_method_res_type(unop_tp_u, 1, tp_u);
3144 }
3145 if (! unop_tp_s) {
3146 unop_tp_s = new_type_method(2, 2);
3147 if (env->params->little_endian) {
3148 set_method_param_type(unop_tp_s, 0, tp_u);
3149 set_method_param_type(unop_tp_s, 1, tp_s);
3150 set_method_res_type(unop_tp_s, 0, tp_u);
3151 set_method_res_type(unop_tp_s, 1, tp_s);
3152 } else {
3153 set_method_param_type(unop_tp_s, 0, tp_s);
3154 set_method_param_type(unop_tp_s, 1, tp_u);
3155 set_method_res_type(unop_tp_s, 0, tp_s);
3156 set_method_res_type(unop_tp_s, 1, tp_u);
3157 }
3158 }
3159
3160 lenv.tv_mode_bytes = new_tarval_from_long(param->doubleword_size/(2*8), lenv.low_unsigned);
3161 lenv.waitq = new_pdeq();
3162 lenv.first_id = new_id_from_chars(param->little_endian ? ".l" : ".h", 2);
3163 lenv.next_id = new_id_from_chars(param->little_endian ? ".h" : ".l", 2);
3164
3165 irp_reserve_resources(irp, IRP_RESOURCE_TYPE_LINK | IRP_RESOURCE_TYPE_VISITED);
3166 inc_master_type_visited();
3167 /* transform all graphs */
3168 for (i = 0, n = get_irp_n_irgs(); i < n; ++i) {
3169 ir_graph *irg = get_irp_irg(i);
3170
3171 ir_nodeset_init(&created_mux_nodes);
3172
3173 lower_irg(irg);
3174
3175 if (ir_nodeset_size(&created_mux_nodes) > 0)
3176 lower_mux(irg, lower_mux_cb);
3177
3178 ir_nodeset_destroy(&created_mux_nodes);
3179 }
3180 irp_free_resources(irp, IRP_RESOURCE_TYPE_LINK | IRP_RESOURCE_TYPE_VISITED);
3181 del_pdeq(lenv.waitq);
3182
3183 env = NULL;
3184 }
3185
3186 /* Default implementation. */
def_create_intrinsic_fkt(ir_type * method,const ir_op * op,const ir_mode * imode,const ir_mode * omode,void * context)3187 ir_entity *def_create_intrinsic_fkt(ir_type *method, const ir_op *op,
3188 const ir_mode *imode, const ir_mode *omode,
3189 void *context)
3190 {
3191 char buf[64];
3192 ident *id;
3193 ir_entity *ent;
3194 (void) context;
3195
3196 if (imode == omode) {
3197 snprintf(buf, sizeof(buf), "__l%s%s", get_op_name(op), get_mode_name(imode));
3198 } else {
3199 snprintf(buf, sizeof(buf), "__l%s%s%s", get_op_name(op),
3200 get_mode_name(imode), get_mode_name(omode));
3201 }
3202 id = new_id_from_str(buf);
3203
3204 ent = new_entity(get_glob_type(), id, method);
3205 set_entity_ld_ident(ent, get_entity_ident(ent));
3206 set_entity_visibility(ent, ir_visibility_external);
3207 return ent;
3208 }
3209