xref: /qemu/target/hexagon/gen_tcg.h (revision 370ed600)
1 /*
2  *  Copyright(c) 2019-2023 Qualcomm Innovation Center, Inc. All Rights Reserved.
3  *
4  *  This program is free software; you can redistribute it and/or modify
5  *  it under the terms of the GNU General Public License as published by
6  *  the Free Software Foundation; either version 2 of the License, or
7  *  (at your option) any later version.
8  *
9  *  This program is distributed in the hope that it will be useful,
10  *  but WITHOUT ANY WARRANTY; without even the implied warranty of
11  *  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
12  *  GNU General Public License for more details.
13  *
14  *  You should have received a copy of the GNU General Public License
15  *  along with this program; if not, see <http://www.gnu.org/licenses/>.
16  */
17 
18 #ifndef HEXAGON_GEN_TCG_H
19 #define HEXAGON_GEN_TCG_H
20 
21 /*
22  * Here is a primer to understand the tag names for load/store instructions
23  *
24  * Data types
25  *      b        signed byte                       r0 = memb(r2+#0)
26  *     ub        unsigned byte                     r0 = memub(r2+#0)
27  *      h        signed half word (16 bits)        r0 = memh(r2+#0)
28  *     uh        unsigned half word                r0 = memuh(r2+#0)
29  *      i        integer (32 bits)                 r0 = memw(r2+#0)
30  *      d        double word (64 bits)             r1:0 = memd(r2+#0)
31  *
32  * Addressing modes
33  *     _io       indirect with offset              r0 = memw(r1+#4)
34  *     _ur       absolute with register offset     r0 = memw(r1<<#4+##variable)
35  *     _rr       indirect with register offset     r0 = memw(r1+r4<<#2)
36  *     gp        global pointer relative           r0 = memw(gp+#200)
37  *     _sp       stack pointer relative            r0 = memw(r29+#12)
38  *     _ap       absolute set                      r0 = memw(r1=##variable)
39  *     _pr       post increment register           r0 = memw(r1++m1)
40  *     _pbr      post increment bit reverse        r0 = memw(r1++m1:brev)
41  *     _pi       post increment immediate          r0 = memb(r1++#1)
42  *     _pci      post increment circular immediate r0 = memw(r1++#4:circ(m0))
43  *     _pcr      post increment circular register  r0 = memw(r1++I:circ(m0))
44  */
45 
46 /* Macros for complex addressing modes */
47 #define GET_EA_ap \
48     do { \
49         fEA_IMM(UiV); \
50         tcg_gen_movi_tl(ReV, UiV); \
51     } while (0)
52 #define GET_EA_pr \
53     do { \
54         fEA_REG(RxV); \
55         fPM_M(RxV, MuV); \
56     } while (0)
57 #define GET_EA_pbr \
58     do { \
59         gen_helper_fbrev(EA, RxV); \
60         tcg_gen_add_tl(RxV, RxV, MuV); \
61     } while (0)
62 #define GET_EA_pi \
63     do { \
64         fEA_REG(RxV); \
65         fPM_I(RxV, siV); \
66     } while (0)
67 #define GET_EA_pci \
68     do { \
69         TCGv tcgv_siV = tcg_constant_tl(siV); \
70         tcg_gen_mov_tl(EA, RxV); \
71         gen_helper_fcircadd(RxV, RxV, tcgv_siV, MuV, \
72                             hex_gpr[HEX_REG_CS0 + MuN]); \
73     } while (0)
74 #define GET_EA_pcr(SHIFT) \
75     do { \
76         TCGv ireg = tcg_temp_new(); \
77         tcg_gen_mov_tl(EA, RxV); \
78         gen_read_ireg(ireg, MuV, (SHIFT)); \
79         gen_helper_fcircadd(RxV, RxV, ireg, MuV, hex_gpr[HEX_REG_CS0 + MuN]); \
80     } while (0)
81 
82 /* Instructions with multiple definitions */
83 #define fGEN_TCG_LOAD_AP(RES, SIZE, SIGN) \
84     do { \
85         fMUST_IMMEXT(UiV); \
86         fEA_IMM(UiV); \
87         fLOAD(1, SIZE, SIGN, EA, RES); \
88         tcg_gen_movi_tl(ReV, UiV); \
89     } while (0)
90 
91 #define fGEN_TCG_L4_loadrub_ap(SHORTCODE) \
92     fGEN_TCG_LOAD_AP(RdV, 1, u)
93 #define fGEN_TCG_L4_loadrb_ap(SHORTCODE) \
94     fGEN_TCG_LOAD_AP(RdV, 1, s)
95 #define fGEN_TCG_L4_loadruh_ap(SHORTCODE) \
96     fGEN_TCG_LOAD_AP(RdV, 2, u)
97 #define fGEN_TCG_L4_loadrh_ap(SHORTCODE) \
98     fGEN_TCG_LOAD_AP(RdV, 2, s)
99 #define fGEN_TCG_L4_loadri_ap(SHORTCODE) \
100     fGEN_TCG_LOAD_AP(RdV, 4, u)
101 #define fGEN_TCG_L4_loadrd_ap(SHORTCODE) \
102     fGEN_TCG_LOAD_AP(RddV, 8, u)
103 
104 #define fGEN_TCG_L2_loadrub_pci(SHORTCODE)    SHORTCODE
105 #define fGEN_TCG_L2_loadrb_pci(SHORTCODE)     SHORTCODE
106 #define fGEN_TCG_L2_loadruh_pci(SHORTCODE)    SHORTCODE
107 #define fGEN_TCG_L2_loadrh_pci(SHORTCODE)     SHORTCODE
108 #define fGEN_TCG_L2_loadri_pci(SHORTCODE)     SHORTCODE
109 #define fGEN_TCG_L2_loadrd_pci(SHORTCODE)     SHORTCODE
110 
111 #define fGEN_TCG_LOAD_pcr(SHIFT, LOAD) \
112     do { \
113         TCGv ireg = tcg_temp_new(); \
114         tcg_gen_mov_tl(EA, RxV); \
115         gen_read_ireg(ireg, MuV, SHIFT); \
116         gen_helper_fcircadd(RxV, RxV, ireg, MuV, hex_gpr[HEX_REG_CS0 + MuN]); \
117         LOAD; \
118     } while (0)
119 
120 #define fGEN_TCG_L2_loadrub_pcr(SHORTCODE) \
121       fGEN_TCG_LOAD_pcr(0, fLOAD(1, 1, u, EA, RdV))
122 #define fGEN_TCG_L2_loadrb_pcr(SHORTCODE) \
123       fGEN_TCG_LOAD_pcr(0, fLOAD(1, 1, s, EA, RdV))
124 #define fGEN_TCG_L2_loadruh_pcr(SHORTCODE) \
125       fGEN_TCG_LOAD_pcr(1, fLOAD(1, 2, u, EA, RdV))
126 #define fGEN_TCG_L2_loadrh_pcr(SHORTCODE) \
127       fGEN_TCG_LOAD_pcr(1, fLOAD(1, 2, s, EA, RdV))
128 #define fGEN_TCG_L2_loadri_pcr(SHORTCODE) \
129       fGEN_TCG_LOAD_pcr(2, fLOAD(1, 4, u, EA, RdV))
130 #define fGEN_TCG_L2_loadrd_pcr(SHORTCODE) \
131       fGEN_TCG_LOAD_pcr(3, fLOAD(1, 8, u, EA, RddV))
132 
133 #define fGEN_TCG_L2_loadrub_pr(SHORTCODE)      SHORTCODE
134 #define fGEN_TCG_L2_loadrub_pbr(SHORTCODE)     SHORTCODE
135 #define fGEN_TCG_L2_loadrub_pi(SHORTCODE)      SHORTCODE
136 #define fGEN_TCG_L2_loadrb_pr(SHORTCODE)       SHORTCODE
137 #define fGEN_TCG_L2_loadrb_pbr(SHORTCODE)      SHORTCODE
138 #define fGEN_TCG_L2_loadrb_pi(SHORTCODE)       SHORTCODE
139 #define fGEN_TCG_L2_loadruh_pr(SHORTCODE)      SHORTCODE
140 #define fGEN_TCG_L2_loadruh_pbr(SHORTCODE)     SHORTCODE
141 #define fGEN_TCG_L2_loadruh_pi(SHORTCODE)      SHORTCODE
142 #define fGEN_TCG_L2_loadrh_pr(SHORTCODE)       SHORTCODE
143 #define fGEN_TCG_L2_loadrh_pbr(SHORTCODE)      SHORTCODE
144 #define fGEN_TCG_L2_loadrh_pi(SHORTCODE)       SHORTCODE
145 #define fGEN_TCG_L2_loadri_pr(SHORTCODE)       SHORTCODE
146 #define fGEN_TCG_L2_loadri_pbr(SHORTCODE)      SHORTCODE
147 #define fGEN_TCG_L2_loadri_pi(SHORTCODE)       SHORTCODE
148 #define fGEN_TCG_L2_loadrd_pr(SHORTCODE)       SHORTCODE
149 #define fGEN_TCG_L2_loadrd_pbr(SHORTCODE)      SHORTCODE
150 #define fGEN_TCG_L2_loadrd_pi(SHORTCODE)       SHORTCODE
151 
152 /*
153  * These instructions load 2 bytes and places them in
154  * two halves of the destination register.
155  * The GET_EA macro determines the addressing mode.
156  * The SIGN argument determines whether to zero-extend or
157  * sign-extend.
158  */
159 #define fGEN_TCG_loadbXw2(GET_EA, SIGN) \
160     do { \
161         TCGv tmp = tcg_temp_new(); \
162         TCGv byte = tcg_temp_new(); \
163         GET_EA; \
164         fLOAD(1, 2, u, EA, tmp); \
165         tcg_gen_movi_tl(RdV, 0); \
166         for (int i = 0; i < 2; i++) { \
167             gen_set_half(i, RdV, gen_get_byte(byte, i, tmp, (SIGN))); \
168         } \
169     } while (0)
170 
171 #define fGEN_TCG_L2_loadbzw2_io(SHORTCODE) \
172     fGEN_TCG_loadbXw2(fEA_RI(RsV, siV), false)
173 #define fGEN_TCG_L4_loadbzw2_ur(SHORTCODE) \
174     fGEN_TCG_loadbXw2(fEA_IRs(UiV, RtV, uiV), false)
175 #define fGEN_TCG_L2_loadbsw2_io(SHORTCODE) \
176     fGEN_TCG_loadbXw2(fEA_RI(RsV, siV), true)
177 #define fGEN_TCG_L4_loadbsw2_ur(SHORTCODE) \
178     fGEN_TCG_loadbXw2(fEA_IRs(UiV, RtV, uiV), true)
179 #define fGEN_TCG_L4_loadbzw2_ap(SHORTCODE) \
180     fGEN_TCG_loadbXw2(GET_EA_ap, false)
181 #define fGEN_TCG_L2_loadbzw2_pr(SHORTCODE) \
182     fGEN_TCG_loadbXw2(GET_EA_pr, false)
183 #define fGEN_TCG_L2_loadbzw2_pbr(SHORTCODE) \
184     fGEN_TCG_loadbXw2(GET_EA_pbr, false)
185 #define fGEN_TCG_L2_loadbzw2_pi(SHORTCODE) \
186     fGEN_TCG_loadbXw2(GET_EA_pi, false)
187 #define fGEN_TCG_L4_loadbsw2_ap(SHORTCODE) \
188     fGEN_TCG_loadbXw2(GET_EA_ap, true)
189 #define fGEN_TCG_L2_loadbsw2_pr(SHORTCODE) \
190     fGEN_TCG_loadbXw2(GET_EA_pr, true)
191 #define fGEN_TCG_L2_loadbsw2_pbr(SHORTCODE) \
192     fGEN_TCG_loadbXw2(GET_EA_pbr, true)
193 #define fGEN_TCG_L2_loadbsw2_pi(SHORTCODE) \
194     fGEN_TCG_loadbXw2(GET_EA_pi, true)
195 #define fGEN_TCG_L2_loadbzw2_pci(SHORTCODE) \
196     fGEN_TCG_loadbXw2(GET_EA_pci, false)
197 #define fGEN_TCG_L2_loadbsw2_pci(SHORTCODE) \
198     fGEN_TCG_loadbXw2(GET_EA_pci, true)
199 #define fGEN_TCG_L2_loadbzw2_pcr(SHORTCODE) \
200     fGEN_TCG_loadbXw2(GET_EA_pcr(1), false)
201 #define fGEN_TCG_L2_loadbsw2_pcr(SHORTCODE) \
202     fGEN_TCG_loadbXw2(GET_EA_pcr(1), true)
203 
204 /*
205  * These instructions load 4 bytes and places them in
206  * four halves of the destination register pair.
207  * The GET_EA macro determines the addressing mode.
208  * The SIGN argument determines whether to zero-extend or
209  * sign-extend.
210  */
211 #define fGEN_TCG_loadbXw4(GET_EA, SIGN) \
212     do { \
213         TCGv tmp = tcg_temp_new(); \
214         TCGv byte = tcg_temp_new(); \
215         GET_EA; \
216         fLOAD(1, 4, u, EA, tmp);  \
217         tcg_gen_movi_i64(RddV, 0); \
218         for (int i = 0; i < 4; i++) { \
219             gen_set_half_i64(i, RddV, gen_get_byte(byte, i, tmp, (SIGN)));  \
220         }  \
221     } while (0)
222 
223 #define fGEN_TCG_L2_loadbzw4_io(SHORTCODE) \
224     fGEN_TCG_loadbXw4(fEA_RI(RsV, siV), false)
225 #define fGEN_TCG_L4_loadbzw4_ur(SHORTCODE) \
226     fGEN_TCG_loadbXw4(fEA_IRs(UiV, RtV, uiV), false)
227 #define fGEN_TCG_L2_loadbsw4_io(SHORTCODE) \
228     fGEN_TCG_loadbXw4(fEA_RI(RsV, siV), true)
229 #define fGEN_TCG_L4_loadbsw4_ur(SHORTCODE) \
230     fGEN_TCG_loadbXw4(fEA_IRs(UiV, RtV, uiV), true)
231 #define fGEN_TCG_L2_loadbzw4_pci(SHORTCODE) \
232     fGEN_TCG_loadbXw4(GET_EA_pci, false)
233 #define fGEN_TCG_L2_loadbsw4_pci(SHORTCODE) \
234     fGEN_TCG_loadbXw4(GET_EA_pci, true)
235 #define fGEN_TCG_L2_loadbzw4_pcr(SHORTCODE) \
236     fGEN_TCG_loadbXw4(GET_EA_pcr(2), false)
237 #define fGEN_TCG_L2_loadbsw4_pcr(SHORTCODE) \
238     fGEN_TCG_loadbXw4(GET_EA_pcr(2), true)
239 #define fGEN_TCG_L4_loadbzw4_ap(SHORTCODE) \
240     fGEN_TCG_loadbXw4(GET_EA_ap, false)
241 #define fGEN_TCG_L2_loadbzw4_pr(SHORTCODE) \
242     fGEN_TCG_loadbXw4(GET_EA_pr, false)
243 #define fGEN_TCG_L2_loadbzw4_pbr(SHORTCODE) \
244     fGEN_TCG_loadbXw4(GET_EA_pbr, false)
245 #define fGEN_TCG_L2_loadbzw4_pi(SHORTCODE) \
246     fGEN_TCG_loadbXw4(GET_EA_pi, false)
247 #define fGEN_TCG_L4_loadbsw4_ap(SHORTCODE) \
248     fGEN_TCG_loadbXw4(GET_EA_ap, true)
249 #define fGEN_TCG_L2_loadbsw4_pr(SHORTCODE) \
250     fGEN_TCG_loadbXw4(GET_EA_pr, true)
251 #define fGEN_TCG_L2_loadbsw4_pbr(SHORTCODE) \
252     fGEN_TCG_loadbXw4(GET_EA_pbr, true)
253 #define fGEN_TCG_L2_loadbsw4_pi(SHORTCODE) \
254     fGEN_TCG_loadbXw4(GET_EA_pi, true)
255 
256 /*
257  * These instructions load a half word, shift the destination right by 16 bits
258  * and place the loaded value in the high half word of the destination pair.
259  * The GET_EA macro determines the addressing mode.
260  */
261 #define fGEN_TCG_loadalignh(GET_EA) \
262     do { \
263         TCGv tmp = tcg_temp_new(); \
264         TCGv_i64 tmp_i64 = tcg_temp_new_i64(); \
265         GET_EA;  \
266         fLOAD(1, 2, u, EA, tmp);  \
267         tcg_gen_extu_i32_i64(tmp_i64, tmp); \
268         tcg_gen_shri_i64(RyyV, RyyV, 16); \
269         tcg_gen_deposit_i64(RyyV, RyyV, tmp_i64, 48, 16); \
270     } while (0)
271 
272 #define fGEN_TCG_L4_loadalignh_ur(SHORTCODE) \
273     fGEN_TCG_loadalignh(fEA_IRs(UiV, RtV, uiV))
274 #define fGEN_TCG_L2_loadalignh_io(SHORTCODE) \
275     fGEN_TCG_loadalignh(fEA_RI(RsV, siV))
276 #define fGEN_TCG_L2_loadalignh_pci(SHORTCODE) \
277     fGEN_TCG_loadalignh(GET_EA_pci)
278 #define fGEN_TCG_L2_loadalignh_pcr(SHORTCODE) \
279     fGEN_TCG_loadalignh(GET_EA_pcr(1))
280 #define fGEN_TCG_L4_loadalignh_ap(SHORTCODE) \
281     fGEN_TCG_loadalignh(GET_EA_ap)
282 #define fGEN_TCG_L2_loadalignh_pr(SHORTCODE) \
283     fGEN_TCG_loadalignh(GET_EA_pr)
284 #define fGEN_TCG_L2_loadalignh_pbr(SHORTCODE) \
285     fGEN_TCG_loadalignh(GET_EA_pbr)
286 #define fGEN_TCG_L2_loadalignh_pi(SHORTCODE) \
287     fGEN_TCG_loadalignh(GET_EA_pi)
288 
289 /* Same as above, but loads a byte instead of half word */
290 #define fGEN_TCG_loadalignb(GET_EA) \
291     do { \
292         TCGv tmp = tcg_temp_new(); \
293         TCGv_i64 tmp_i64 = tcg_temp_new_i64(); \
294         GET_EA;  \
295         fLOAD(1, 1, u, EA, tmp);  \
296         tcg_gen_extu_i32_i64(tmp_i64, tmp); \
297         tcg_gen_shri_i64(RyyV, RyyV, 8); \
298         tcg_gen_deposit_i64(RyyV, RyyV, tmp_i64, 56, 8); \
299     } while (0)
300 
301 #define fGEN_TCG_L2_loadalignb_io(SHORTCODE) \
302     fGEN_TCG_loadalignb(fEA_RI(RsV, siV))
303 #define fGEN_TCG_L4_loadalignb_ur(SHORTCODE) \
304     fGEN_TCG_loadalignb(fEA_IRs(UiV, RtV, uiV))
305 #define fGEN_TCG_L2_loadalignb_pci(SHORTCODE) \
306     fGEN_TCG_loadalignb(GET_EA_pci)
307 #define fGEN_TCG_L2_loadalignb_pcr(SHORTCODE) \
308     fGEN_TCG_loadalignb(GET_EA_pcr(0))
309 #define fGEN_TCG_L4_loadalignb_ap(SHORTCODE) \
310     fGEN_TCG_loadalignb(GET_EA_ap)
311 #define fGEN_TCG_L2_loadalignb_pr(SHORTCODE) \
312     fGEN_TCG_loadalignb(GET_EA_pr)
313 #define fGEN_TCG_L2_loadalignb_pbr(SHORTCODE) \
314     fGEN_TCG_loadalignb(GET_EA_pbr)
315 #define fGEN_TCG_L2_loadalignb_pi(SHORTCODE) \
316     fGEN_TCG_loadalignb(GET_EA_pi)
317 
318 /*
319  * Predicated loads
320  * Here is a primer to understand the tag names
321  *
322  * Predicate used
323  *      t        true "old" value                  if (p0) r0 = memb(r2+#0)
324  *      f        false "old" value                 if (!p0) r0 = memb(r2+#0)
325  *      tnew     true "new" value                  if (p0.new) r0 = memb(r2+#0)
326  *      fnew     false "new" value                 if (!p0.new) r0 = memb(r2+#0)
327  */
328 #define fGEN_TCG_PRED_LOAD(GET_EA, PRED, SIZE, SIGN) \
329     do { \
330         TCGv LSB = tcg_temp_new(); \
331         TCGLabel *label = gen_new_label(); \
332         tcg_gen_movi_tl(EA, 0); \
333         PRED;  \
334         CHECK_NOSHUF_PRED(GET_EA, SIZE, LSB); \
335         tcg_gen_brcondi_tl(TCG_COND_EQ, LSB, 0, label); \
336         fLOAD(1, SIZE, SIGN, EA, RdV); \
337         gen_set_label(label); \
338     } while (0)
339 
340 #define fGEN_TCG_L2_ploadrubt_pi(SHORTCODE) \
341     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBOLD(PtV), 1, u)
342 #define fGEN_TCG_L2_ploadrubf_pi(SHORTCODE) \
343     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBOLDNOT(PtV), 1, u)
344 #define fGEN_TCG_L2_ploadrubtnew_pi(SHORTCODE) \
345     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBNEW(PtN), 1, u)
346 #define fGEN_TCG_L2_ploadrubfnew_pi(SHORTCODE) \
347     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBNEWNOT(PtN), 1, u)
348 #define fGEN_TCG_L2_ploadrbt_pi(SHORTCODE) \
349     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBOLD(PtV), 1, s)
350 #define fGEN_TCG_L2_ploadrbf_pi(SHORTCODE) \
351     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBOLDNOT(PtV), 1, s)
352 #define fGEN_TCG_L2_ploadrbtnew_pi(SHORTCODE) \
353     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBNEW(PtN), 1, s)
354 #define fGEN_TCG_L2_ploadrbfnew_pi(SHORTCODE) \
355     fGEN_TCG_PRED_LOAD({ fEA_REG(RxV); fPM_I(RxV, siV); }, \
356                        fLSBNEWNOT(PtN), 1, s)
357 
358 #define fGEN_TCG_L2_ploadruht_pi(SHORTCODE) \
359     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBOLD(PtV), 2, u)
360 #define fGEN_TCG_L2_ploadruhf_pi(SHORTCODE) \
361     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBOLDNOT(PtV), 2, u)
362 #define fGEN_TCG_L2_ploadruhtnew_pi(SHORTCODE) \
363     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBNEW(PtN), 2, u)
364 #define fGEN_TCG_L2_ploadruhfnew_pi(SHORTCODE) \
365     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBNEWNOT(PtN), 2, u)
366 #define fGEN_TCG_L2_ploadrht_pi(SHORTCODE) \
367     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBOLD(PtV), 2, s)
368 #define fGEN_TCG_L2_ploadrhf_pi(SHORTCODE) \
369     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBOLDNOT(PtV), 2, s)
370 #define fGEN_TCG_L2_ploadrhtnew_pi(SHORTCODE) \
371     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBNEW(PtN), 2, s)
372 #define fGEN_TCG_L2_ploadrhfnew_pi(SHORTCODE) \
373     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBNEWNOT(PtN), 2, s)
374 
375 #define fGEN_TCG_L2_ploadrit_pi(SHORTCODE) \
376     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBOLD(PtV), 4, u)
377 #define fGEN_TCG_L2_ploadrif_pi(SHORTCODE) \
378     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBOLDNOT(PtV), 4, u)
379 #define fGEN_TCG_L2_ploadritnew_pi(SHORTCODE) \
380     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBNEW(PtN), 4, u)
381 #define fGEN_TCG_L2_ploadrifnew_pi(SHORTCODE) \
382     fGEN_TCG_PRED_LOAD(GET_EA_pi, fLSBNEWNOT(PtN), 4, u)
383 
384 /* Predicated loads into a register pair */
385 #define fGEN_TCG_PRED_LOAD_PAIR(GET_EA, PRED) \
386     do { \
387         TCGv LSB = tcg_temp_new(); \
388         TCGLabel *label = gen_new_label(); \
389         tcg_gen_movi_tl(EA, 0); \
390         PRED;  \
391         CHECK_NOSHUF_PRED(GET_EA, 8, LSB); \
392         tcg_gen_brcondi_tl(TCG_COND_EQ, LSB, 0, label); \
393         fLOAD(1, 8, u, EA, RddV); \
394         gen_set_label(label); \
395     } while (0)
396 
397 #define fGEN_TCG_L2_ploadrdt_pi(SHORTCODE) \
398     fGEN_TCG_PRED_LOAD_PAIR(GET_EA_pi, fLSBOLD(PtV))
399 #define fGEN_TCG_L2_ploadrdf_pi(SHORTCODE) \
400     fGEN_TCG_PRED_LOAD_PAIR(GET_EA_pi, fLSBOLDNOT(PtV))
401 #define fGEN_TCG_L2_ploadrdtnew_pi(SHORTCODE) \
402     fGEN_TCG_PRED_LOAD_PAIR(GET_EA_pi, fLSBNEW(PtN))
403 #define fGEN_TCG_L2_ploadrdfnew_pi(SHORTCODE) \
404     fGEN_TCG_PRED_LOAD_PAIR(GET_EA_pi, fLSBNEWNOT(PtN))
405 
406 /* load-locked and store-locked */
407 #define fGEN_TCG_L2_loadw_locked(SHORTCODE) \
408     SHORTCODE
409 #define fGEN_TCG_L4_loadd_locked(SHORTCODE) \
410     SHORTCODE
411 #define fGEN_TCG_S2_storew_locked(SHORTCODE) \
412     SHORTCODE
413 #define fGEN_TCG_S4_stored_locked(SHORTCODE) \
414     SHORTCODE
415 
416 #define fGEN_TCG_STORE(SHORTCODE) \
417     do { \
418         TCGv HALF G_GNUC_UNUSED = tcg_temp_new(); \
419         TCGv BYTE G_GNUC_UNUSED = tcg_temp_new(); \
420         SHORTCODE; \
421     } while (0)
422 
423 #define fGEN_TCG_STORE_pcr(SHIFT, STORE) \
424     do { \
425         TCGv ireg = tcg_temp_new(); \
426         TCGv HALF G_GNUC_UNUSED = tcg_temp_new(); \
427         TCGv BYTE G_GNUC_UNUSED = tcg_temp_new(); \
428         tcg_gen_mov_tl(EA, RxV); \
429         gen_read_ireg(ireg, MuV, SHIFT); \
430         gen_helper_fcircadd(RxV, RxV, ireg, MuV, hex_gpr[HEX_REG_CS0 + MuN]); \
431         STORE; \
432     } while (0)
433 
434 #define fGEN_TCG_S2_storerb_pbr(SHORTCODE) \
435     fGEN_TCG_STORE(SHORTCODE)
436 #define fGEN_TCG_S2_storerb_pci(SHORTCODE) \
437     fGEN_TCG_STORE(SHORTCODE)
438 #define fGEN_TCG_S2_storerb_pcr(SHORTCODE) \
439     fGEN_TCG_STORE_pcr(0, fSTORE(1, 1, EA, fGETBYTE(0, RtV)))
440 
441 #define fGEN_TCG_S2_storerh_pbr(SHORTCODE) \
442     fGEN_TCG_STORE(SHORTCODE)
443 #define fGEN_TCG_S2_storerh_pci(SHORTCODE) \
444     fGEN_TCG_STORE(SHORTCODE)
445 #define fGEN_TCG_S2_storerh_pcr(SHORTCODE) \
446     fGEN_TCG_STORE_pcr(1, fSTORE(1, 2, EA, fGETHALF(0, RtV)))
447 
448 #define fGEN_TCG_S2_storerf_pbr(SHORTCODE) \
449     fGEN_TCG_STORE(SHORTCODE)
450 #define fGEN_TCG_S2_storerf_pci(SHORTCODE) \
451     fGEN_TCG_STORE(SHORTCODE)
452 #define fGEN_TCG_S2_storerf_pcr(SHORTCODE) \
453     fGEN_TCG_STORE_pcr(1, fSTORE(1, 2, EA, fGETHALF(1, RtV)))
454 
455 #define fGEN_TCG_S2_storeri_pbr(SHORTCODE) \
456     fGEN_TCG_STORE(SHORTCODE)
457 #define fGEN_TCG_S2_storeri_pci(SHORTCODE) \
458     fGEN_TCG_STORE(SHORTCODE)
459 #define fGEN_TCG_S2_storeri_pcr(SHORTCODE) \
460     fGEN_TCG_STORE_pcr(2, fSTORE(1, 4, EA, RtV))
461 
462 #define fGEN_TCG_S2_storerd_pbr(SHORTCODE) \
463     fGEN_TCG_STORE(SHORTCODE)
464 #define fGEN_TCG_S2_storerd_pci(SHORTCODE) \
465     fGEN_TCG_STORE(SHORTCODE)
466 #define fGEN_TCG_S2_storerd_pcr(SHORTCODE) \
467     fGEN_TCG_STORE_pcr(3, fSTORE(1, 8, EA, RttV))
468 
469 #define fGEN_TCG_S2_storerbnew_pbr(SHORTCODE) \
470     fGEN_TCG_STORE(SHORTCODE)
471 #define fGEN_TCG_S2_storerbnew_pci(SHORTCODE) \
472     fGEN_TCG_STORE(SHORTCODE)
473 #define fGEN_TCG_S2_storerbnew_pcr(SHORTCODE) \
474     fGEN_TCG_STORE_pcr(0, fSTORE(1, 1, EA, fGETBYTE(0, NtN)))
475 
476 #define fGEN_TCG_S2_storerhnew_pbr(SHORTCODE) \
477     fGEN_TCG_STORE(SHORTCODE)
478 #define fGEN_TCG_S2_storerhnew_pci(SHORTCODE) \
479     fGEN_TCG_STORE(SHORTCODE)
480 #define fGEN_TCG_S2_storerhnew_pcr(SHORTCODE) \
481     fGEN_TCG_STORE_pcr(1, fSTORE(1, 2, EA, fGETHALF(0, NtN)))
482 
483 #define fGEN_TCG_S2_storerinew_pbr(SHORTCODE) \
484     fGEN_TCG_STORE(SHORTCODE)
485 #define fGEN_TCG_S2_storerinew_pci(SHORTCODE) \
486     fGEN_TCG_STORE(SHORTCODE)
487 #define fGEN_TCG_S2_storerinew_pcr(SHORTCODE) \
488     fGEN_TCG_STORE_pcr(2, fSTORE(1, 4, EA, NtN))
489 
490 /* dczeroa clears the 32 byte cache line at the address given */
491 #define fGEN_TCG_Y2_dczeroa(SHORTCODE) SHORTCODE
492 
493 /* In linux-user mode, these are not modelled, suppress compiler warning */
494 #define fGEN_TCG_Y2_dcinva(SHORTCODE) \
495     do { RsV = RsV; } while (0)
496 #define fGEN_TCG_Y2_dccleaninva(SHORTCODE) \
497     do { RsV = RsV; } while (0)
498 #define fGEN_TCG_Y2_dccleana(SHORTCODE) \
499     do { RsV = RsV; } while (0)
500 #define fGEN_TCG_Y2_icinva(SHORTCODE) \
501     do { RsV = RsV; } while (0)
502 
503 /*
504  * dealloc_return
505  * Assembler mapped to
506  * r31:30 = dealloc_return(r30):raw
507  */
508 #define fGEN_TCG_L4_return(SHORTCODE) \
509     gen_return(ctx, RddV, RsV)
510 
511 /*
512  * sub-instruction version (no RddV, so handle it manually)
513  */
514 #define fGEN_TCG_SL2_return(SHORTCODE) \
515     do { \
516         TCGv_i64 RddV = get_result_gpr_pair(ctx, HEX_REG_FP); \
517         gen_return(ctx, RddV, hex_gpr[HEX_REG_FP]); \
518         gen_log_reg_write_pair(HEX_REG_FP, RddV); \
519     } while (0)
520 
521 /*
522  * Conditional returns follow this naming convention
523  *     _t                 predicate true
524  *     _f                 predicate false
525  *     _tnew_pt           predicate.new true predict taken
526  *     _fnew_pt           predicate.new false predict taken
527  *     _tnew_pnt          predicate.new true predict not taken
528  *     _fnew_pnt          predicate.new false predict not taken
529  * Predictions are not modelled in QEMU
530  *
531  * Example:
532  *     if (p1) r31:30 = dealloc_return(r30):raw
533  */
534 #define fGEN_TCG_L4_return_t(SHORTCODE) \
535     gen_cond_return(ctx, RddV, RsV, PvV, TCG_COND_EQ);
536 #define fGEN_TCG_L4_return_f(SHORTCODE) \
537     gen_cond_return(ctx, RddV, RsV, PvV, TCG_COND_NE)
538 #define fGEN_TCG_L4_return_tnew_pt(SHORTCODE) \
539     gen_cond_return(ctx, RddV, RsV, PvN, TCG_COND_EQ)
540 #define fGEN_TCG_L4_return_fnew_pt(SHORTCODE) \
541     gen_cond_return(ctx, RddV, RsV, PvN, TCG_COND_NE)
542 #define fGEN_TCG_L4_return_tnew_pnt(SHORTCODE) \
543     gen_cond_return(ctx, RddV, RsV, PvN, TCG_COND_EQ)
544 #define fGEN_TCG_L4_return_fnew_pnt(SHORTCODE) \
545     gen_cond_return(ctx, RddV, RsV, PvN, TCG_COND_NE)
546 
547 #define fGEN_TCG_SL2_return_t(SHORTCODE) \
548     gen_cond_return_subinsn(ctx, TCG_COND_EQ, hex_pred[0])
549 #define fGEN_TCG_SL2_return_f(SHORTCODE) \
550     gen_cond_return_subinsn(ctx, TCG_COND_NE, hex_pred[0])
551 #define fGEN_TCG_SL2_return_tnew(SHORTCODE) \
552     gen_cond_return_subinsn(ctx, TCG_COND_EQ, hex_new_pred_value[0])
553 #define fGEN_TCG_SL2_return_fnew(SHORTCODE) \
554     gen_cond_return_subinsn(ctx, TCG_COND_NE, hex_new_pred_value[0])
555 
556 /*
557  * Mathematical operations with more than one definition require
558  * special handling
559  */
560 #define fGEN_TCG_A5_ACS(SHORTCODE) \
561     do { \
562         gen_helper_vacsh_pred(PeV, cpu_env, RxxV, RssV, RttV); \
563         gen_helper_vacsh_val(RxxV, cpu_env, RxxV, RssV, RttV); \
564     } while (0)
565 
566 /*
567  * Approximate reciprocal
568  * r3,p1 = sfrecipa(r0, r1)
569  *
570  * The helper packs the 2 32-bit results into a 64-bit value,
571  * so unpack them into the proper results.
572  */
573 #define fGEN_TCG_F2_sfrecipa(SHORTCODE) \
574     do { \
575         TCGv_i64 tmp = tcg_temp_new_i64(); \
576         gen_helper_sfrecipa(tmp, cpu_env, RsV, RtV);  \
577         tcg_gen_extrh_i64_i32(RdV, tmp); \
578         tcg_gen_extrl_i64_i32(PeV, tmp); \
579     } while (0)
580 
581 /*
582  * Approximation of the reciprocal square root
583  * r1,p0 = sfinvsqrta(r0)
584  *
585  * The helper packs the 2 32-bit results into a 64-bit value,
586  * so unpack them into the proper results.
587  */
588 #define fGEN_TCG_F2_sfinvsqrta(SHORTCODE) \
589     do { \
590         TCGv_i64 tmp = tcg_temp_new_i64(); \
591         gen_helper_sfinvsqrta(tmp, cpu_env, RsV); \
592         tcg_gen_extrh_i64_i32(RdV, tmp); \
593         tcg_gen_extrl_i64_i32(PeV, tmp); \
594     } while (0)
595 
596 /*
597  * Add or subtract with carry.
598  * Predicate register is used as an extra input and output.
599  * r5:4 = add(r1:0, r3:2, p1):carry
600  */
601 #define fGEN_TCG_A4_addp_c(SHORTCODE) \
602     do { \
603         TCGv_i64 carry = tcg_temp_new_i64(); \
604         TCGv_i64 zero = tcg_constant_i64(0); \
605         tcg_gen_extu_i32_i64(carry, PxV); \
606         tcg_gen_andi_i64(carry, carry, 1); \
607         tcg_gen_add2_i64(RddV, carry, RssV, zero, carry, zero); \
608         tcg_gen_add2_i64(RddV, carry, RddV, carry, RttV, zero); \
609         tcg_gen_extrl_i64_i32(PxV, carry); \
610         gen_8bitsof(PxV, PxV); \
611     } while (0)
612 
613 /* r5:4 = sub(r1:0, r3:2, p1):carry */
614 #define fGEN_TCG_A4_subp_c(SHORTCODE) \
615     do { \
616         TCGv_i64 carry = tcg_temp_new_i64(); \
617         TCGv_i64 zero = tcg_constant_i64(0); \
618         TCGv_i64 not_RttV = tcg_temp_new_i64(); \
619         tcg_gen_extu_i32_i64(carry, PxV); \
620         tcg_gen_andi_i64(carry, carry, 1); \
621         tcg_gen_not_i64(not_RttV, RttV); \
622         tcg_gen_add2_i64(RddV, carry, RssV, zero, carry, zero); \
623         tcg_gen_add2_i64(RddV, carry, RddV, carry, not_RttV, zero); \
624         tcg_gen_extrl_i64_i32(PxV, carry); \
625         gen_8bitsof(PxV, PxV); \
626     } while (0)
627 
628 /*
629  * Compare each of the 8 unsigned bytes
630  * The minimum is placed in each byte of the destination.
631  * Each bit of the predicate is set true if the bit from the first operand
632  * is greater than the bit from the second operand.
633  * r5:4,p1 = vminub(r1:0, r3:2)
634  */
635 #define fGEN_TCG_A6_vminub_RdP(SHORTCODE) \
636     do { \
637         TCGv left = tcg_temp_new(); \
638         TCGv right = tcg_temp_new(); \
639         TCGv tmp = tcg_temp_new(); \
640         tcg_gen_movi_tl(PeV, 0); \
641         tcg_gen_movi_i64(RddV, 0); \
642         for (int i = 0; i < 8; i++) { \
643             gen_get_byte_i64(left, i, RttV, false); \
644             gen_get_byte_i64(right, i, RssV, false); \
645             tcg_gen_setcond_tl(TCG_COND_GT, tmp, left, right); \
646             tcg_gen_deposit_tl(PeV, PeV, tmp, i, 1); \
647             tcg_gen_umin_tl(tmp, left, right); \
648             gen_set_byte_i64(i, RddV, tmp); \
649         } \
650     } while (0)
651 
652 #define fGEN_TCG_J2_call(SHORTCODE) \
653     gen_call(ctx, riV)
654 #define fGEN_TCG_J2_callr(SHORTCODE) \
655     gen_callr(ctx, RsV)
656 
657 #define fGEN_TCG_J2_callt(SHORTCODE) \
658     gen_cond_call(ctx, PuV, TCG_COND_EQ, riV)
659 #define fGEN_TCG_J2_callf(SHORTCODE) \
660     gen_cond_call(ctx, PuV, TCG_COND_NE, riV)
661 #define fGEN_TCG_J2_callrt(SHORTCODE) \
662     gen_cond_callr(ctx, TCG_COND_EQ, PuV, RsV)
663 #define fGEN_TCG_J2_callrf(SHORTCODE) \
664     gen_cond_callr(ctx, TCG_COND_NE, PuV, RsV)
665 
666 #define fGEN_TCG_J2_endloop0(SHORTCODE) \
667     gen_endloop0(ctx)
668 #define fGEN_TCG_J2_endloop1(SHORTCODE) \
669     gen_endloop1(ctx)
670 #define fGEN_TCG_J2_endloop01(SHORTCODE) \
671     gen_endloop01(ctx)
672 
673 /*
674  * Compound compare and jump instructions
675  * Here is a primer to understand the tag names
676  *
677  * Comparison
678  *      cmpeqi   compare equal to an immediate
679  *      cmpgti   compare greater than an immediate
680  *      cmpgtiu  compare greater than an unsigned immediate
681  *      cmpeqn1  compare equal to negative 1
682  *      cmpgtn1  compare greater than negative 1
683  *      cmpeq    compare equal (two registers)
684  *      cmpgtu   compare greater than unsigned (two registers)
685  *      tstbit0  test bit zero
686  *
687  * Condition
688  *      tp0      p0 is true     p0 = cmp.eq(r0,#5); if (p0.new) jump:nt address
689  *      fp0      p0 is false    p0 = cmp.eq(r0,#5); if (!p0.new) jump:nt address
690  *      tp1      p1 is true     p1 = cmp.eq(r0,#5); if (p1.new) jump:nt address
691  *      fp1      p1 is false    p1 = cmp.eq(r0,#5); if (!p1.new) jump:nt address
692  *
693  * Prediction (not modelled in qemu)
694  *      _nt      not taken
695  *      _t       taken
696  */
697 #define fGEN_TCG_J4_cmpeq_tp0_jump_t(SHORTCODE) \
698     gen_cmpnd_cmp_jmp_t(ctx, 0, TCG_COND_EQ, RsV, RtV, riV)
699 #define fGEN_TCG_J4_cmpeq_tp0_jump_nt(SHORTCODE) \
700     gen_cmpnd_cmp_jmp_t(ctx, 0, TCG_COND_EQ, RsV, RtV, riV)
701 #define fGEN_TCG_J4_cmpeq_fp0_jump_t(SHORTCODE) \
702     gen_cmpnd_cmp_jmp_f(ctx, 0, TCG_COND_EQ, RsV, RtV, riV)
703 #define fGEN_TCG_J4_cmpeq_fp0_jump_nt(SHORTCODE) \
704     gen_cmpnd_cmp_jmp_f(ctx, 0, TCG_COND_EQ, RsV, RtV, riV)
705 #define fGEN_TCG_J4_cmpeq_tp1_jump_t(SHORTCODE) \
706     gen_cmpnd_cmp_jmp_t(ctx, 1, TCG_COND_EQ, RsV, RtV, riV)
707 #define fGEN_TCG_J4_cmpeq_tp1_jump_nt(SHORTCODE) \
708     gen_cmpnd_cmp_jmp_t(ctx, 1, TCG_COND_EQ, RsV, RtV, riV)
709 #define fGEN_TCG_J4_cmpeq_fp1_jump_t(SHORTCODE) \
710     gen_cmpnd_cmp_jmp_f(ctx, 1, TCG_COND_EQ, RsV, RtV, riV)
711 #define fGEN_TCG_J4_cmpeq_fp1_jump_nt(SHORTCODE) \
712     gen_cmpnd_cmp_jmp_f(ctx, 1, TCG_COND_EQ, RsV, RtV, riV)
713 
714 #define fGEN_TCG_J4_cmpgt_tp0_jump_t(SHORTCODE) \
715     gen_cmpnd_cmp_jmp_t(ctx, 0, TCG_COND_GT, RsV, RtV, riV)
716 #define fGEN_TCG_J4_cmpgt_tp0_jump_nt(SHORTCODE) \
717     gen_cmpnd_cmp_jmp_t(ctx, 0, TCG_COND_GT, RsV, RtV, riV)
718 #define fGEN_TCG_J4_cmpgt_fp0_jump_t(SHORTCODE) \
719     gen_cmpnd_cmp_jmp_f(ctx, 0, TCG_COND_GT, RsV, RtV, riV)
720 #define fGEN_TCG_J4_cmpgt_fp0_jump_nt(SHORTCODE) \
721     gen_cmpnd_cmp_jmp_f(ctx, 0, TCG_COND_GT, RsV, RtV, riV)
722 #define fGEN_TCG_J4_cmpgt_tp1_jump_t(SHORTCODE) \
723     gen_cmpnd_cmp_jmp_t(ctx, 1, TCG_COND_GT, RsV, RtV, riV)
724 #define fGEN_TCG_J4_cmpgt_tp1_jump_nt(SHORTCODE) \
725     gen_cmpnd_cmp_jmp_t(ctx, 1, TCG_COND_GT, RsV, RtV, riV)
726 #define fGEN_TCG_J4_cmpgt_fp1_jump_t(SHORTCODE) \
727     gen_cmpnd_cmp_jmp_f(ctx, 1, TCG_COND_GT, RsV, RtV, riV)
728 #define fGEN_TCG_J4_cmpgt_fp1_jump_nt(SHORTCODE) \
729     gen_cmpnd_cmp_jmp_f(ctx, 1, TCG_COND_GT, RsV, RtV, riV)
730 
731 #define fGEN_TCG_J4_cmpgtu_tp0_jump_t(SHORTCODE) \
732     gen_cmpnd_cmp_jmp_t(ctx, 0, TCG_COND_GTU, RsV, RtV, riV)
733 #define fGEN_TCG_J4_cmpgtu_tp0_jump_nt(SHORTCODE) \
734     gen_cmpnd_cmp_jmp_t(ctx, 0, TCG_COND_GTU, RsV, RtV, riV)
735 #define fGEN_TCG_J4_cmpgtu_fp0_jump_t(SHORTCODE) \
736     gen_cmpnd_cmp_jmp_f(ctx, 0, TCG_COND_GTU, RsV, RtV, riV)
737 #define fGEN_TCG_J4_cmpgtu_fp0_jump_nt(SHORTCODE) \
738     gen_cmpnd_cmp_jmp_f(ctx, 0, TCG_COND_GTU, RsV, RtV, riV)
739 #define fGEN_TCG_J4_cmpgtu_tp1_jump_t(SHORTCODE) \
740     gen_cmpnd_cmp_jmp_t(ctx, 1, TCG_COND_GTU, RsV, RtV, riV)
741 #define fGEN_TCG_J4_cmpgtu_tp1_jump_nt(SHORTCODE) \
742     gen_cmpnd_cmp_jmp_t(ctx, 1, TCG_COND_GTU, RsV, RtV, riV)
743 #define fGEN_TCG_J4_cmpgtu_fp1_jump_t(SHORTCODE) \
744     gen_cmpnd_cmp_jmp_f(ctx, 1, TCG_COND_GTU, RsV, RtV, riV)
745 #define fGEN_TCG_J4_cmpgtu_fp1_jump_nt(SHORTCODE) \
746     gen_cmpnd_cmp_jmp_f(ctx, 1, TCG_COND_GTU, RsV, RtV, riV)
747 
748 #define fGEN_TCG_J4_cmpeqi_tp0_jump_t(SHORTCODE) \
749     gen_cmpnd_cmpi_jmp_t(ctx, 0, TCG_COND_EQ, RsV, UiV, riV)
750 #define fGEN_TCG_J4_cmpeqi_tp0_jump_nt(SHORTCODE) \
751     gen_cmpnd_cmpi_jmp_t(ctx, 0, TCG_COND_EQ, RsV, UiV, riV)
752 #define fGEN_TCG_J4_cmpeqi_fp0_jump_t(SHORTCODE) \
753     gen_cmpnd_cmpi_jmp_f(ctx, 0, TCG_COND_EQ, RsV, UiV, riV)
754 #define fGEN_TCG_J4_cmpeqi_fp0_jump_nt(SHORTCODE) \
755     gen_cmpnd_cmpi_jmp_f(ctx, 0, TCG_COND_EQ, RsV, UiV, riV)
756 #define fGEN_TCG_J4_cmpeqi_tp1_jump_t(SHORTCODE) \
757     gen_cmpnd_cmpi_jmp_t(ctx, 1, TCG_COND_EQ, RsV, UiV, riV)
758 #define fGEN_TCG_J4_cmpeqi_tp1_jump_nt(SHORTCODE) \
759     gen_cmpnd_cmpi_jmp_t(ctx, 1, TCG_COND_EQ, RsV, UiV, riV)
760 #define fGEN_TCG_J4_cmpeqi_fp1_jump_t(SHORTCODE) \
761     gen_cmpnd_cmpi_jmp_f(ctx, 1, TCG_COND_EQ, RsV, UiV, riV)
762 #define fGEN_TCG_J4_cmpeqi_fp1_jump_nt(SHORTCODE) \
763     gen_cmpnd_cmpi_jmp_f(ctx, 1, TCG_COND_EQ, RsV, UiV, riV)
764 
765 #define fGEN_TCG_J4_cmpgti_tp0_jump_t(SHORTCODE) \
766     gen_cmpnd_cmpi_jmp_t(ctx, 0, TCG_COND_GT, RsV, UiV, riV)
767 #define fGEN_TCG_J4_cmpgti_tp0_jump_nt(SHORTCODE) \
768     gen_cmpnd_cmpi_jmp_t(ctx, 0, TCG_COND_GT, RsV, UiV, riV)
769 #define fGEN_TCG_J4_cmpgti_fp0_jump_t(SHORTCODE) \
770     gen_cmpnd_cmpi_jmp_f(ctx, 0, TCG_COND_GT, RsV, UiV, riV)
771 #define fGEN_TCG_J4_cmpgti_fp0_jump_nt(SHORTCODE) \
772     gen_cmpnd_cmpi_jmp_f(ctx, 0, TCG_COND_GT, RsV, UiV, riV)
773 #define fGEN_TCG_J4_cmpgti_tp1_jump_t(SHORTCODE) \
774     gen_cmpnd_cmpi_jmp_t(ctx, 1, TCG_COND_GT, RsV, UiV, riV)
775 #define fGEN_TCG_J4_cmpgti_tp1_jump_nt(SHORTCODE) \
776     gen_cmpnd_cmpi_jmp_t(ctx, 1, TCG_COND_GT, RsV, UiV, riV)
777 #define fGEN_TCG_J4_cmpgti_fp1_jump_t(SHORTCODE) \
778     gen_cmpnd_cmpi_jmp_f(ctx, 1, TCG_COND_GT, RsV, UiV, riV)
779 #define fGEN_TCG_J4_cmpgti_fp1_jump_nt(SHORTCODE) \
780     gen_cmpnd_cmpi_jmp_f(ctx, 1, TCG_COND_GT, RsV, UiV, riV)
781 
782 #define fGEN_TCG_J4_cmpgtui_tp0_jump_t(SHORTCODE) \
783     gen_cmpnd_cmpi_jmp_t(ctx, 0, TCG_COND_GTU, RsV, UiV, riV)
784 #define fGEN_TCG_J4_cmpgtui_tp0_jump_nt(SHORTCODE) \
785     gen_cmpnd_cmpi_jmp_t(ctx, 0, TCG_COND_GTU, RsV, UiV, riV)
786 #define fGEN_TCG_J4_cmpgtui_fp0_jump_t(SHORTCODE) \
787     gen_cmpnd_cmpi_jmp_f(ctx, 0, TCG_COND_GTU, RsV, UiV, riV)
788 #define fGEN_TCG_J4_cmpgtui_fp0_jump_nt(SHORTCODE) \
789     gen_cmpnd_cmpi_jmp_f(ctx, 0, TCG_COND_GTU, RsV, UiV, riV)
790 #define fGEN_TCG_J4_cmpgtui_tp1_jump_t(SHORTCODE) \
791     gen_cmpnd_cmpi_jmp_t(ctx, 1, TCG_COND_GTU, RsV, UiV, riV)
792 #define fGEN_TCG_J4_cmpgtui_tp1_jump_nt(SHORTCODE) \
793     gen_cmpnd_cmpi_jmp_t(ctx, 1, TCG_COND_GTU, RsV, UiV, riV)
794 #define fGEN_TCG_J4_cmpgtui_fp1_jump_t(SHORTCODE) \
795     gen_cmpnd_cmpi_jmp_f(ctx, 1, TCG_COND_GTU, RsV, UiV, riV)
796 #define fGEN_TCG_J4_cmpgtui_fp1_jump_nt(SHORTCODE) \
797     gen_cmpnd_cmpi_jmp_f(ctx, 1, TCG_COND_GTU, RsV, UiV, riV)
798 
799 #define fGEN_TCG_J4_cmpeqn1_tp0_jump_t(SHORTCODE) \
800     gen_cmpnd_cmp_n1_jmp_t(ctx, 0, TCG_COND_EQ, RsV, riV)
801 #define fGEN_TCG_J4_cmpeqn1_tp0_jump_nt(SHORTCODE) \
802     gen_cmpnd_cmp_n1_jmp_t(ctx, 0, TCG_COND_EQ, RsV, riV)
803 #define fGEN_TCG_J4_cmpeqn1_fp0_jump_t(SHORTCODE) \
804     gen_cmpnd_cmp_n1_jmp_f(ctx, 0, TCG_COND_EQ, RsV, riV)
805 #define fGEN_TCG_J4_cmpeqn1_fp0_jump_nt(SHORTCODE) \
806     gen_cmpnd_cmp_n1_jmp_f(ctx, 0, TCG_COND_EQ, RsV, riV)
807 #define fGEN_TCG_J4_cmpeqn1_tp1_jump_t(SHORTCODE) \
808     gen_cmpnd_cmp_n1_jmp_t(ctx, 1, TCG_COND_EQ, RsV, riV)
809 #define fGEN_TCG_J4_cmpeqn1_tp1_jump_nt(SHORTCODE) \
810     gen_cmpnd_cmp_n1_jmp_t(ctx, 1, TCG_COND_EQ, RsV, riV)
811 #define fGEN_TCG_J4_cmpeqn1_fp1_jump_t(SHORTCODE) \
812     gen_cmpnd_cmp_n1_jmp_f(ctx, 1, TCG_COND_EQ, RsV, riV)
813 #define fGEN_TCG_J4_cmpeqn1_fp1_jump_nt(SHORTCODE) \
814     gen_cmpnd_cmp_n1_jmp_f(ctx, 1, TCG_COND_EQ, RsV, riV)
815 
816 #define fGEN_TCG_J4_cmpgtn1_tp0_jump_t(SHORTCODE) \
817     gen_cmpnd_cmp_n1_jmp_t(ctx, 0, TCG_COND_GT, RsV, riV)
818 #define fGEN_TCG_J4_cmpgtn1_tp0_jump_nt(SHORTCODE) \
819     gen_cmpnd_cmp_n1_jmp_t(ctx, 0, TCG_COND_GT, RsV, riV)
820 #define fGEN_TCG_J4_cmpgtn1_fp0_jump_t(SHORTCODE) \
821     gen_cmpnd_cmp_n1_jmp_f(ctx, 0, TCG_COND_GT, RsV, riV)
822 #define fGEN_TCG_J4_cmpgtn1_fp0_jump_nt(SHORTCODE) \
823     gen_cmpnd_cmp_n1_jmp_f(ctx, 0, TCG_COND_GT, RsV, riV)
824 #define fGEN_TCG_J4_cmpgtn1_tp1_jump_t(SHORTCODE) \
825     gen_cmpnd_cmp_n1_jmp_t(ctx, 1, TCG_COND_GT, RsV, riV)
826 #define fGEN_TCG_J4_cmpgtn1_tp1_jump_nt(SHORTCODE) \
827     gen_cmpnd_cmp_n1_jmp_t(ctx, 1, TCG_COND_GT, RsV, riV)
828 #define fGEN_TCG_J4_cmpgtn1_fp1_jump_t(SHORTCODE) \
829     gen_cmpnd_cmp_n1_jmp_f(ctx, 1, TCG_COND_GT, RsV, riV)
830 #define fGEN_TCG_J4_cmpgtn1_fp1_jump_nt(SHORTCODE) \
831     gen_cmpnd_cmp_n1_jmp_f(ctx, 1, TCG_COND_GT, RsV, riV)
832 
833 #define fGEN_TCG_J4_tstbit0_tp0_jump_nt(SHORTCODE) \
834     gen_cmpnd_tstbit0_jmp(ctx, 0, RsV, TCG_COND_EQ, riV)
835 #define fGEN_TCG_J4_tstbit0_tp0_jump_t(SHORTCODE) \
836     gen_cmpnd_tstbit0_jmp(ctx, 0, RsV, TCG_COND_EQ, riV)
837 #define fGEN_TCG_J4_tstbit0_fp0_jump_nt(SHORTCODE) \
838     gen_cmpnd_tstbit0_jmp(ctx, 0, RsV, TCG_COND_NE, riV)
839 #define fGEN_TCG_J4_tstbit0_fp0_jump_t(SHORTCODE) \
840     gen_cmpnd_tstbit0_jmp(ctx, 0, RsV, TCG_COND_NE, riV)
841 #define fGEN_TCG_J4_tstbit0_tp1_jump_nt(SHORTCODE) \
842     gen_cmpnd_tstbit0_jmp(ctx, 1, RsV, TCG_COND_EQ, riV)
843 #define fGEN_TCG_J4_tstbit0_tp1_jump_t(SHORTCODE) \
844     gen_cmpnd_tstbit0_jmp(ctx, 1, RsV, TCG_COND_EQ, riV)
845 #define fGEN_TCG_J4_tstbit0_fp1_jump_nt(SHORTCODE) \
846     gen_cmpnd_tstbit0_jmp(ctx, 1, RsV, TCG_COND_NE, riV)
847 #define fGEN_TCG_J4_tstbit0_fp1_jump_t(SHORTCODE) \
848     gen_cmpnd_tstbit0_jmp(ctx, 1, RsV, TCG_COND_NE, riV)
849 
850 #define fGEN_TCG_J2_jump(SHORTCODE) \
851     gen_jump(ctx, riV)
852 #define fGEN_TCG_J2_jumpr(SHORTCODE) \
853     gen_jumpr(ctx, RsV)
854 #define fGEN_TCG_J4_jumpseti(SHORTCODE) \
855     do { \
856         tcg_gen_movi_tl(RdV, UiV); \
857         gen_jump(ctx, riV); \
858     } while (0)
859 
860 #define fGEN_TCG_cond_jumpt(COND) \
861     do { \
862         TCGv LSB = tcg_temp_new(); \
863         COND; \
864         gen_cond_jump(ctx, TCG_COND_EQ, LSB, riV); \
865     } while (0)
866 #define fGEN_TCG_cond_jumpf(COND) \
867     do { \
868         TCGv LSB = tcg_temp_new(); \
869         COND; \
870         gen_cond_jump(ctx, TCG_COND_NE, LSB, riV); \
871     } while (0)
872 
873 #define fGEN_TCG_J2_jumpt(SHORTCODE) \
874     fGEN_TCG_cond_jumpt(fLSBOLD(PuV))
875 #define fGEN_TCG_J2_jumptpt(SHORTCODE) \
876     fGEN_TCG_cond_jumpt(fLSBOLD(PuV))
877 #define fGEN_TCG_J2_jumpf(SHORTCODE) \
878     fGEN_TCG_cond_jumpf(fLSBOLD(PuV))
879 #define fGEN_TCG_J2_jumpfpt(SHORTCODE) \
880     fGEN_TCG_cond_jumpf(fLSBOLD(PuV))
881 #define fGEN_TCG_J2_jumptnew(SHORTCODE) \
882     gen_cond_jump(ctx, TCG_COND_EQ, PuN, riV)
883 #define fGEN_TCG_J2_jumptnewpt(SHORTCODE) \
884     gen_cond_jump(ctx, TCG_COND_EQ, PuN, riV)
885 #define fGEN_TCG_J2_jumpfnewpt(SHORTCODE) \
886     fGEN_TCG_cond_jumpf(fLSBNEW(PuN))
887 #define fGEN_TCG_J2_jumpfnew(SHORTCODE) \
888     fGEN_TCG_cond_jumpf(fLSBNEW(PuN))
889 #define fGEN_TCG_J2_jumprz(SHORTCODE) \
890     fGEN_TCG_cond_jumpt(tcg_gen_setcondi_tl(TCG_COND_NE, LSB, RsV, 0))
891 #define fGEN_TCG_J2_jumprzpt(SHORTCODE) \
892     fGEN_TCG_cond_jumpt(tcg_gen_setcondi_tl(TCG_COND_NE, LSB, RsV, 0))
893 #define fGEN_TCG_J2_jumprnz(SHORTCODE) \
894     fGEN_TCG_cond_jumpt(tcg_gen_setcondi_tl(TCG_COND_EQ, LSB, RsV, 0))
895 #define fGEN_TCG_J2_jumprnzpt(SHORTCODE) \
896     fGEN_TCG_cond_jumpt(tcg_gen_setcondi_tl(TCG_COND_EQ, LSB, RsV, 0))
897 #define fGEN_TCG_J2_jumprgtez(SHORTCODE) \
898     fGEN_TCG_cond_jumpt(tcg_gen_setcondi_tl(TCG_COND_GE, LSB, RsV, 0))
899 #define fGEN_TCG_J2_jumprgtezpt(SHORTCODE) \
900     fGEN_TCG_cond_jumpt(tcg_gen_setcondi_tl(TCG_COND_GE, LSB, RsV, 0))
901 #define fGEN_TCG_J2_jumprltez(SHORTCODE) \
902     fGEN_TCG_cond_jumpt(tcg_gen_setcondi_tl(TCG_COND_LE, LSB, RsV, 0))
903 #define fGEN_TCG_J2_jumprltezpt(SHORTCODE) \
904     fGEN_TCG_cond_jumpt(tcg_gen_setcondi_tl(TCG_COND_LE, LSB, RsV, 0))
905 
906 #define fGEN_TCG_cond_jumprt(COND) \
907     do { \
908         TCGv LSB = tcg_temp_new(); \
909         COND; \
910         gen_cond_jumpr(ctx, RsV, TCG_COND_EQ, LSB); \
911     } while (0)
912 #define fGEN_TCG_cond_jumprf(COND) \
913     do { \
914         TCGv LSB = tcg_temp_new(); \
915         COND; \
916         gen_cond_jumpr(ctx, RsV, TCG_COND_NE, LSB); \
917     } while (0)
918 
919 #define fGEN_TCG_J2_jumprt(SHORTCODE) \
920     fGEN_TCG_cond_jumprt(fLSBOLD(PuV))
921 #define fGEN_TCG_J2_jumprtpt(SHORTCODE) \
922     fGEN_TCG_cond_jumprt(fLSBOLD(PuV))
923 #define fGEN_TCG_J2_jumprf(SHORTCODE) \
924     fGEN_TCG_cond_jumprf(fLSBOLD(PuV))
925 #define fGEN_TCG_J2_jumprfpt(SHORTCODE) \
926     fGEN_TCG_cond_jumprf(fLSBOLD(PuV))
927 #define fGEN_TCG_J2_jumprtnew(SHORTCODE) \
928     fGEN_TCG_cond_jumprt(fLSBNEW(PuN))
929 #define fGEN_TCG_J2_jumprtnewpt(SHORTCODE) \
930     fGEN_TCG_cond_jumprt(fLSBNEW(PuN))
931 #define fGEN_TCG_J2_jumprfnew(SHORTCODE) \
932     fGEN_TCG_cond_jumprf(fLSBNEW(PuN))
933 #define fGEN_TCG_J2_jumprfnewpt(SHORTCODE) \
934     fGEN_TCG_cond_jumprf(fLSBNEW(PuN))
935 
936 /*
937  * New value compare & jump instructions
938  * if ([!]COND(r0.new, r1) jump:t address
939  * if ([!]COND(r0.new, #7) jump:t address
940  */
941 #define fGEN_TCG_J4_cmpgt_t_jumpnv_t(SHORTCODE) \
942     gen_cmp_jumpnv(ctx, TCG_COND_GT, NsN, RtV, riV)
943 #define fGEN_TCG_J4_cmpgt_t_jumpnv_nt(SHORTCODE) \
944     gen_cmp_jumpnv(ctx, TCG_COND_GT, NsN, RtV, riV)
945 #define fGEN_TCG_J4_cmpgt_f_jumpnv_t(SHORTCODE) \
946     gen_cmp_jumpnv(ctx, TCG_COND_LE, NsN, RtV, riV)
947 #define fGEN_TCG_J4_cmpgt_f_jumpnv_nt(SHORTCODE) \
948     gen_cmp_jumpnv(ctx, TCG_COND_LE, NsN, RtV, riV)
949 
950 #define fGEN_TCG_J4_cmpeq_t_jumpnv_t(SHORTCODE) \
951     gen_cmp_jumpnv(ctx, TCG_COND_EQ, NsN, RtV, riV)
952 #define fGEN_TCG_J4_cmpeq_t_jumpnv_nt(SHORTCODE) \
953     gen_cmp_jumpnv(ctx, TCG_COND_EQ, NsN, RtV, riV)
954 #define fGEN_TCG_J4_cmpeq_f_jumpnv_t(SHORTCODE) \
955     gen_cmp_jumpnv(ctx, TCG_COND_NE, NsN, RtV, riV)
956 #define fGEN_TCG_J4_cmpeq_f_jumpnv_nt(SHORTCODE) \
957     gen_cmp_jumpnv(ctx, TCG_COND_NE, NsN, RtV, riV)
958 
959 #define fGEN_TCG_J4_cmplt_t_jumpnv_t(SHORTCODE) \
960     gen_cmp_jumpnv(ctx, TCG_COND_LT, NsN, RtV, riV)
961 #define fGEN_TCG_J4_cmplt_t_jumpnv_nt(SHORTCODE) \
962     gen_cmp_jumpnv(ctx, TCG_COND_LT, NsN, RtV, riV)
963 #define fGEN_TCG_J4_cmplt_f_jumpnv_t(SHORTCODE) \
964     gen_cmp_jumpnv(ctx, TCG_COND_GE, NsN, RtV, riV)
965 #define fGEN_TCG_J4_cmplt_f_jumpnv_nt(SHORTCODE) \
966     gen_cmp_jumpnv(ctx, TCG_COND_GE, NsN, RtV, riV)
967 
968 #define fGEN_TCG_J4_cmpeqi_t_jumpnv_t(SHORTCODE) \
969     gen_cmpi_jumpnv(ctx, TCG_COND_EQ, NsN, UiV, riV)
970 #define fGEN_TCG_J4_cmpeqi_t_jumpnv_nt(SHORTCODE) \
971     gen_cmpi_jumpnv(ctx, TCG_COND_EQ, NsN, UiV, riV)
972 #define fGEN_TCG_J4_cmpeqi_f_jumpnv_t(SHORTCODE) \
973     gen_cmpi_jumpnv(ctx, TCG_COND_NE, NsN, UiV, riV)
974 #define fGEN_TCG_J4_cmpeqi_f_jumpnv_nt(SHORTCODE) \
975     gen_cmpi_jumpnv(ctx, TCG_COND_NE, NsN, UiV, riV)
976 
977 #define fGEN_TCG_J4_cmpgti_t_jumpnv_t(SHORTCODE) \
978     gen_cmpi_jumpnv(ctx, TCG_COND_GT, NsN, UiV, riV)
979 #define fGEN_TCG_J4_cmpgti_t_jumpnv_nt(SHORTCODE) \
980     gen_cmpi_jumpnv(ctx, TCG_COND_GT, NsN, UiV, riV)
981 #define fGEN_TCG_J4_cmpgti_f_jumpnv_t(SHORTCODE) \
982     gen_cmpi_jumpnv(ctx, TCG_COND_LE, NsN, UiV, riV)
983 #define fGEN_TCG_J4_cmpgti_f_jumpnv_nt(SHORTCODE) \
984     gen_cmpi_jumpnv(ctx, TCG_COND_LE, NsN, UiV, riV)
985 
986 #define fGEN_TCG_J4_cmpltu_t_jumpnv_t(SHORTCODE) \
987     gen_cmp_jumpnv(ctx, TCG_COND_LTU, NsN, RtV, riV)
988 #define fGEN_TCG_J4_cmpltu_t_jumpnv_nt(SHORTCODE) \
989     gen_cmp_jumpnv(ctx, TCG_COND_LTU, NsN, RtV, riV)
990 #define fGEN_TCG_J4_cmpltu_f_jumpnv_t(SHORTCODE) \
991     gen_cmp_jumpnv(ctx, TCG_COND_GEU, NsN, RtV, riV)
992 #define fGEN_TCG_J4_cmpltu_f_jumpnv_nt(SHORTCODE) \
993     gen_cmp_jumpnv(ctx, TCG_COND_GEU, NsN, RtV, riV)
994 
995 #define fGEN_TCG_J4_cmpgtui_t_jumpnv_t(SHORTCODE) \
996     gen_cmpi_jumpnv(ctx, TCG_COND_GTU, NsN, UiV, riV)
997 #define fGEN_TCG_J4_cmpgtui_t_jumpnv_nt(SHORTCODE) \
998     gen_cmpi_jumpnv(ctx, TCG_COND_GTU, NsN, UiV, riV)
999 #define fGEN_TCG_J4_cmpgtui_f_jumpnv_t(SHORTCODE) \
1000     gen_cmpi_jumpnv(ctx, TCG_COND_LEU, NsN, UiV, riV)
1001 #define fGEN_TCG_J4_cmpgtui_f_jumpnv_nt(SHORTCODE) \
1002     gen_cmpi_jumpnv(ctx, TCG_COND_LEU, NsN, UiV, riV)
1003 
1004 #define fGEN_TCG_J4_cmpgtu_t_jumpnv_t(SHORTCODE) \
1005     gen_cmp_jumpnv(ctx, TCG_COND_GTU, NsN, RtV, riV)
1006 #define fGEN_TCG_J4_cmpgtu_t_jumpnv_nt(SHORTCODE) \
1007     gen_cmp_jumpnv(ctx, TCG_COND_GTU, NsN, RtV, riV)
1008 #define fGEN_TCG_J4_cmpgtu_f_jumpnv_t(SHORTCODE) \
1009     gen_cmp_jumpnv(ctx, TCG_COND_LEU, NsN, RtV, riV)
1010 #define fGEN_TCG_J4_cmpgtu_f_jumpnv_nt(SHORTCODE) \
1011     gen_cmp_jumpnv(ctx, TCG_COND_LEU, NsN, RtV, riV)
1012 
1013 #define fGEN_TCG_J4_cmpeqn1_t_jumpnv_t(SHORTCODE) \
1014     gen_cmpi_jumpnv(ctx, TCG_COND_EQ, NsN, -1, riV)
1015 #define fGEN_TCG_J4_cmpeqn1_t_jumpnv_nt(SHORTCODE) \
1016     gen_cmpi_jumpnv(ctx, TCG_COND_EQ, NsN, -1, riV)
1017 #define fGEN_TCG_J4_cmpeqn1_f_jumpnv_t(SHORTCODE) \
1018     gen_cmpi_jumpnv(ctx, TCG_COND_NE, NsN, -1, riV)
1019 #define fGEN_TCG_J4_cmpeqn1_f_jumpnv_nt(SHORTCODE) \
1020     gen_cmpi_jumpnv(ctx, TCG_COND_NE, NsN, -1, riV)
1021 
1022 #define fGEN_TCG_J4_cmpgtn1_t_jumpnv_t(SHORTCODE) \
1023     gen_cmpi_jumpnv(ctx, TCG_COND_GT, NsN, -1, riV)
1024 #define fGEN_TCG_J4_cmpgtn1_t_jumpnv_nt(SHORTCODE) \
1025     gen_cmpi_jumpnv(ctx, TCG_COND_GT, NsN, -1, riV)
1026 #define fGEN_TCG_J4_cmpgtn1_f_jumpnv_t(SHORTCODE) \
1027     gen_cmpi_jumpnv(ctx, TCG_COND_LE, NsN, -1, riV)
1028 #define fGEN_TCG_J4_cmpgtn1_f_jumpnv_nt(SHORTCODE) \
1029     gen_cmpi_jumpnv(ctx, TCG_COND_LE, NsN, -1, riV)
1030 
1031 #define fGEN_TCG_J4_tstbit0_t_jumpnv_t(SHORTCODE) \
1032     gen_testbit0_jumpnv(ctx, NsN, TCG_COND_EQ, riV)
1033 #define fGEN_TCG_J4_tstbit0_t_jumpnv_nt(SHORTCODE) \
1034     gen_testbit0_jumpnv(ctx, NsN, TCG_COND_EQ, riV)
1035 #define fGEN_TCG_J4_tstbit0_f_jumpnv_t(SHORTCODE) \
1036     gen_testbit0_jumpnv(ctx, NsN, TCG_COND_NE, riV)
1037 #define fGEN_TCG_J4_tstbit0_f_jumpnv_nt(SHORTCODE) \
1038     gen_testbit0_jumpnv(ctx, NsN, TCG_COND_NE, riV)
1039 
1040 /* r0 = r1 ; jump address */
1041 #define fGEN_TCG_J4_jumpsetr(SHORTCODE) \
1042     do { \
1043         tcg_gen_mov_tl(RdV, RsV); \
1044         gen_jump(ctx, riV); \
1045     } while (0)
1046 
1047 #define fGEN_TCG_J2_pause(SHORTCODE) \
1048     do { \
1049         uiV = uiV; \
1050         tcg_gen_movi_tl(hex_gpr[HEX_REG_PC], ctx->next_PC); \
1051     } while (0)
1052 
1053 /* r0 = asr(r1, r2):sat */
1054 #define fGEN_TCG_S2_asr_r_r_sat(SHORTCODE) \
1055     gen_asr_r_r_sat(ctx, RdV, RsV, RtV)
1056 
1057 /* r0 = asl(r1, r2):sat */
1058 #define fGEN_TCG_S2_asl_r_r_sat(SHORTCODE) \
1059     gen_asl_r_r_sat(ctx, RdV, RsV, RtV)
1060 
1061 #define fGEN_TCG_SL2_jumpr31(SHORTCODE) \
1062     gen_jumpr(ctx, hex_gpr[HEX_REG_LR])
1063 
1064 #define fGEN_TCG_SL2_jumpr31_t(SHORTCODE) \
1065     gen_cond_jumpr31(ctx, TCG_COND_EQ, hex_pred[0])
1066 #define fGEN_TCG_SL2_jumpr31_f(SHORTCODE) \
1067     gen_cond_jumpr31(ctx, TCG_COND_NE, hex_pred[0])
1068 
1069 #define fGEN_TCG_SL2_jumpr31_tnew(SHORTCODE) \
1070     gen_cond_jumpr31(ctx, TCG_COND_EQ, hex_new_pred_value[0])
1071 #define fGEN_TCG_SL2_jumpr31_fnew(SHORTCODE) \
1072     gen_cond_jumpr31(ctx, TCG_COND_NE, hex_new_pred_value[0])
1073 
1074 /* Count trailing zeros/ones */
1075 #define fGEN_TCG_S2_ct0(SHORTCODE) \
1076     do { \
1077         tcg_gen_ctzi_tl(RdV, RsV, 32); \
1078     } while (0)
1079 #define fGEN_TCG_S2_ct1(SHORTCODE) \
1080     do { \
1081         tcg_gen_not_tl(RdV, RsV); \
1082         tcg_gen_ctzi_tl(RdV, RdV, 32); \
1083     } while (0)
1084 #define fGEN_TCG_S2_ct0p(SHORTCODE) \
1085     do { \
1086         TCGv_i64 tmp = tcg_temp_new_i64(); \
1087         tcg_gen_ctzi_i64(tmp, RssV, 64); \
1088         tcg_gen_extrl_i64_i32(RdV, tmp); \
1089     } while (0)
1090 #define fGEN_TCG_S2_ct1p(SHORTCODE) \
1091     do { \
1092         TCGv_i64 tmp = tcg_temp_new_i64(); \
1093         tcg_gen_not_i64(tmp, RssV); \
1094         tcg_gen_ctzi_i64(tmp, tmp, 64); \
1095         tcg_gen_extrl_i64_i32(RdV, tmp); \
1096     } while (0)
1097 
1098 /* Floating point */
1099 #define fGEN_TCG_F2_conv_sf2df(SHORTCODE) \
1100     gen_helper_conv_sf2df(RddV, cpu_env, RsV)
1101 #define fGEN_TCG_F2_conv_df2sf(SHORTCODE) \
1102     gen_helper_conv_df2sf(RdV, cpu_env, RssV)
1103 #define fGEN_TCG_F2_conv_uw2sf(SHORTCODE) \
1104     gen_helper_conv_uw2sf(RdV, cpu_env, RsV)
1105 #define fGEN_TCG_F2_conv_uw2df(SHORTCODE) \
1106     gen_helper_conv_uw2df(RddV, cpu_env, RsV)
1107 #define fGEN_TCG_F2_conv_w2sf(SHORTCODE) \
1108     gen_helper_conv_w2sf(RdV, cpu_env, RsV)
1109 #define fGEN_TCG_F2_conv_w2df(SHORTCODE) \
1110     gen_helper_conv_w2df(RddV, cpu_env, RsV)
1111 #define fGEN_TCG_F2_conv_ud2sf(SHORTCODE) \
1112     gen_helper_conv_ud2sf(RdV, cpu_env, RssV)
1113 #define fGEN_TCG_F2_conv_ud2df(SHORTCODE) \
1114     gen_helper_conv_ud2df(RddV, cpu_env, RssV)
1115 #define fGEN_TCG_F2_conv_d2sf(SHORTCODE) \
1116     gen_helper_conv_d2sf(RdV, cpu_env, RssV)
1117 #define fGEN_TCG_F2_conv_d2df(SHORTCODE) \
1118     gen_helper_conv_d2df(RddV, cpu_env, RssV)
1119 #define fGEN_TCG_F2_conv_sf2uw(SHORTCODE) \
1120     gen_helper_conv_sf2uw(RdV, cpu_env, RsV)
1121 #define fGEN_TCG_F2_conv_sf2w(SHORTCODE) \
1122     gen_helper_conv_sf2w(RdV, cpu_env, RsV)
1123 #define fGEN_TCG_F2_conv_sf2ud(SHORTCODE) \
1124     gen_helper_conv_sf2ud(RddV, cpu_env, RsV)
1125 #define fGEN_TCG_F2_conv_sf2d(SHORTCODE) \
1126     gen_helper_conv_sf2d(RddV, cpu_env, RsV)
1127 #define fGEN_TCG_F2_conv_df2uw(SHORTCODE) \
1128     gen_helper_conv_df2uw(RdV, cpu_env, RssV)
1129 #define fGEN_TCG_F2_conv_df2w(SHORTCODE) \
1130     gen_helper_conv_df2w(RdV, cpu_env, RssV)
1131 #define fGEN_TCG_F2_conv_df2ud(SHORTCODE) \
1132     gen_helper_conv_df2ud(RddV, cpu_env, RssV)
1133 #define fGEN_TCG_F2_conv_df2d(SHORTCODE) \
1134     gen_helper_conv_df2d(RddV, cpu_env, RssV)
1135 #define fGEN_TCG_F2_conv_sf2uw_chop(SHORTCODE) \
1136     gen_helper_conv_sf2uw_chop(RdV, cpu_env, RsV)
1137 #define fGEN_TCG_F2_conv_sf2w_chop(SHORTCODE) \
1138     gen_helper_conv_sf2w_chop(RdV, cpu_env, RsV)
1139 #define fGEN_TCG_F2_conv_sf2ud_chop(SHORTCODE) \
1140     gen_helper_conv_sf2ud_chop(RddV, cpu_env, RsV)
1141 #define fGEN_TCG_F2_conv_sf2d_chop(SHORTCODE) \
1142     gen_helper_conv_sf2d_chop(RddV, cpu_env, RsV)
1143 #define fGEN_TCG_F2_conv_df2uw_chop(SHORTCODE) \
1144     gen_helper_conv_df2uw_chop(RdV, cpu_env, RssV)
1145 #define fGEN_TCG_F2_conv_df2w_chop(SHORTCODE) \
1146     gen_helper_conv_df2w_chop(RdV, cpu_env, RssV)
1147 #define fGEN_TCG_F2_conv_df2ud_chop(SHORTCODE) \
1148     gen_helper_conv_df2ud_chop(RddV, cpu_env, RssV)
1149 #define fGEN_TCG_F2_conv_df2d_chop(SHORTCODE) \
1150     gen_helper_conv_df2d_chop(RddV, cpu_env, RssV)
1151 #define fGEN_TCG_F2_sfadd(SHORTCODE) \
1152     gen_helper_sfadd(RdV, cpu_env, RsV, RtV)
1153 #define fGEN_TCG_F2_sfsub(SHORTCODE) \
1154     gen_helper_sfsub(RdV, cpu_env, RsV, RtV)
1155 #define fGEN_TCG_F2_sfcmpeq(SHORTCODE) \
1156     gen_helper_sfcmpeq(PdV, cpu_env, RsV, RtV)
1157 #define fGEN_TCG_F2_sfcmpgt(SHORTCODE) \
1158     gen_helper_sfcmpgt(PdV, cpu_env, RsV, RtV)
1159 #define fGEN_TCG_F2_sfcmpge(SHORTCODE) \
1160     gen_helper_sfcmpge(PdV, cpu_env, RsV, RtV)
1161 #define fGEN_TCG_F2_sfcmpuo(SHORTCODE) \
1162     gen_helper_sfcmpuo(PdV, cpu_env, RsV, RtV)
1163 #define fGEN_TCG_F2_sfmax(SHORTCODE) \
1164     gen_helper_sfmax(RdV, cpu_env, RsV, RtV)
1165 #define fGEN_TCG_F2_sfmin(SHORTCODE) \
1166     gen_helper_sfmin(RdV, cpu_env, RsV, RtV)
1167 #define fGEN_TCG_F2_sfclass(SHORTCODE) \
1168     do { \
1169         TCGv imm = tcg_constant_tl(uiV); \
1170         gen_helper_sfclass(PdV, cpu_env, RsV, imm); \
1171     } while (0)
1172 #define fGEN_TCG_F2_sffixupn(SHORTCODE) \
1173     gen_helper_sffixupn(RdV, cpu_env, RsV, RtV)
1174 #define fGEN_TCG_F2_sffixupd(SHORTCODE) \
1175     gen_helper_sffixupd(RdV, cpu_env, RsV, RtV)
1176 #define fGEN_TCG_F2_sffixupr(SHORTCODE) \
1177     gen_helper_sffixupr(RdV, cpu_env, RsV)
1178 #define fGEN_TCG_F2_dfadd(SHORTCODE) \
1179     gen_helper_dfadd(RddV, cpu_env, RssV, RttV)
1180 #define fGEN_TCG_F2_dfsub(SHORTCODE) \
1181     gen_helper_dfsub(RddV, cpu_env, RssV, RttV)
1182 #define fGEN_TCG_F2_dfmax(SHORTCODE) \
1183     gen_helper_dfmax(RddV, cpu_env, RssV, RttV)
1184 #define fGEN_TCG_F2_dfmin(SHORTCODE) \
1185     gen_helper_dfmin(RddV, cpu_env, RssV, RttV)
1186 #define fGEN_TCG_F2_dfcmpeq(SHORTCODE) \
1187     gen_helper_dfcmpeq(PdV, cpu_env, RssV, RttV)
1188 #define fGEN_TCG_F2_dfcmpgt(SHORTCODE) \
1189     gen_helper_dfcmpgt(PdV, cpu_env, RssV, RttV)
1190 #define fGEN_TCG_F2_dfcmpge(SHORTCODE) \
1191     gen_helper_dfcmpge(PdV, cpu_env, RssV, RttV)
1192 #define fGEN_TCG_F2_dfcmpuo(SHORTCODE) \
1193     gen_helper_dfcmpuo(PdV, cpu_env, RssV, RttV)
1194 #define fGEN_TCG_F2_dfclass(SHORTCODE) \
1195     do { \
1196         TCGv imm = tcg_constant_tl(uiV); \
1197         gen_helper_dfclass(PdV, cpu_env, RssV, imm); \
1198     } while (0)
1199 #define fGEN_TCG_F2_sfmpy(SHORTCODE) \
1200     gen_helper_sfmpy(RdV, cpu_env, RsV, RtV)
1201 #define fGEN_TCG_F2_sffma(SHORTCODE) \
1202     gen_helper_sffma(RxV, cpu_env, RxV, RsV, RtV)
1203 #define fGEN_TCG_F2_sffma_sc(SHORTCODE) \
1204     gen_helper_sffma_sc(RxV, cpu_env, RxV, RsV, RtV, PuV)
1205 #define fGEN_TCG_F2_sffms(SHORTCODE) \
1206     gen_helper_sffms(RxV, cpu_env, RxV, RsV, RtV)
1207 #define fGEN_TCG_F2_sffma_lib(SHORTCODE) \
1208     gen_helper_sffma_lib(RxV, cpu_env, RxV, RsV, RtV)
1209 #define fGEN_TCG_F2_sffms_lib(SHORTCODE) \
1210     gen_helper_sffms_lib(RxV, cpu_env, RxV, RsV, RtV)
1211 
1212 #define fGEN_TCG_F2_dfmpyfix(SHORTCODE) \
1213     gen_helper_dfmpyfix(RddV, cpu_env, RssV, RttV)
1214 #define fGEN_TCG_F2_dfmpyhh(SHORTCODE) \
1215     gen_helper_dfmpyhh(RxxV, cpu_env, RxxV, RssV, RttV)
1216 
1217 /* Nothing to do for these in qemu, need to suppress compiler warnings */
1218 #define fGEN_TCG_Y4_l2fetch(SHORTCODE) \
1219     do { \
1220         RsV = RsV; \
1221         RtV = RtV; \
1222     } while (0)
1223 #define fGEN_TCG_Y5_l2fetch(SHORTCODE) \
1224     do { \
1225         RsV = RsV; \
1226     } while (0)
1227 #define fGEN_TCG_Y2_isync(SHORTCODE) \
1228     do { } while (0)
1229 #define fGEN_TCG_Y2_barrier(SHORTCODE) \
1230     do { } while (0)
1231 #define fGEN_TCG_Y2_syncht(SHORTCODE) \
1232     do { } while (0)
1233 #define fGEN_TCG_Y2_dcfetchbo(SHORTCODE) \
1234     do { \
1235         RsV = RsV; \
1236         uiV = uiV; \
1237     } while (0)
1238 
1239 #define fGEN_TCG_J2_trap0(SHORTCODE) \
1240     do { \
1241         uiV = uiV; \
1242         tcg_gen_movi_tl(hex_gpr[HEX_REG_PC], ctx->pkt->pc); \
1243         TCGv excp = tcg_constant_tl(HEX_EXCP_TRAP0); \
1244         gen_helper_raise_exception(cpu_env, excp); \
1245     } while (0)
1246 #endif
1247