1 /*	$NetBSD: sljitNativeMIPS_common.c,v 1.4 2019/01/20 23:14:16 alnsn Exp $	*/
2 
3 /*
4  *    Stack-less Just-In-Time compiler
5  *
6  *    Copyright Zoltan Herczeg (hzmester@freemail.hu). All rights reserved.
7  *
8  * Redistribution and use in source and binary forms, with or without modification, are
9  * permitted provided that the following conditions are met:
10  *
11  *   1. Redistributions of source code must retain the above copyright notice, this list of
12  *      conditions and the following disclaimer.
13  *
14  *   2. Redistributions in binary form must reproduce the above copyright notice, this list
15  *      of conditions and the following disclaimer in the documentation and/or other materials
16  *      provided with the distribution.
17  *
18  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDER(S) AND CONTRIBUTORS ``AS IS'' AND ANY
19  * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
20  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT
21  * SHALL THE COPYRIGHT HOLDER(S) OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
22  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
23  * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
24  * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
25  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
26  * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27  */
28 
29 /* Latest MIPS architecture. */
30 /* Automatically detect SLJIT_MIPS_R1 */
31 
sljit_get_platform_name(void)32 SLJIT_API_FUNC_ATTRIBUTE const char* sljit_get_platform_name(void)
33 {
34 #if (defined SLJIT_MIPS_R1 && SLJIT_MIPS_R1)
35 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
36 	return "MIPS32-R1" SLJIT_CPUINFO;
37 #else
38 	return "MIPS64-R1" SLJIT_CPUINFO;
39 #endif
40 #else /* SLJIT_MIPS_R1 */
41 	return "MIPS III" SLJIT_CPUINFO;
42 #endif
43 }
44 
45 /* Length of an instruction word
46    Both for mips-32 and mips-64 */
47 typedef sljit_u32 sljit_ins;
48 
49 #define TMP_REG1	(SLJIT_NUMBER_OF_REGISTERS + 2)
50 #define TMP_REG2	(SLJIT_NUMBER_OF_REGISTERS + 3)
51 #define TMP_REG3	(SLJIT_NUMBER_OF_REGISTERS + 4)
52 
53 /* For position independent code, t9 must contain the function address. */
54 #define PIC_ADDR_REG	TMP_REG2
55 
56 /* Floating point status register. */
57 #define FCSR_REG	31
58 /* Return address register. */
59 #define RETURN_ADDR_REG	31
60 
61 /* Flags are kept in volatile registers. */
62 #define EQUAL_FLAG	31
63 #define OTHER_FLAG	1
64 
65 #define TMP_FREG1	(0)
66 #define TMP_FREG2	((SLJIT_NUMBER_OF_FLOAT_REGISTERS + 1) << 1)
67 
68 static const sljit_u8 reg_map[SLJIT_NUMBER_OF_REGISTERS + 5] = {
69 	0, 2, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 24, 23, 22, 21, 20, 19, 18, 17, 16, 29, 3, 25, 4
70 };
71 
72 /* --------------------------------------------------------------------- */
73 /*  Instrucion forms                                                     */
74 /* --------------------------------------------------------------------- */
75 
76 #define S(s)		(reg_map[s] << 21)
77 #define T(t)		(reg_map[t] << 16)
78 #define D(d)		(reg_map[d] << 11)
79 /* Absolute registers. */
80 #define SA(s)		((s) << 21)
81 #define TA(t)		((t) << 16)
82 #define DA(d)		((d) << 11)
83 #define FT(t)		((t) << 16)
84 #define FS(s)		((s) << 11)
85 #define FD(d)		((d) << 6)
86 #define IMM(imm)	((imm) & 0xffff)
87 #define SH_IMM(imm)	((imm) << 6)
88 
89 #define DR(dr)		(reg_map[dr])
90 #define HI(opcode)	((opcode) << 26)
91 #define LO(opcode)	(opcode)
92 /* S = (16 << 21) D = (17 << 21) */
93 #define FMT_S		(16 << 21)
94 
95 #define ABS_S		(HI(17) | FMT_S | LO(5))
96 #define ADD_S		(HI(17) | FMT_S | LO(0))
97 #define ADDIU		(HI(9))
98 #define ADDU		(HI(0) | LO(33))
99 #define AND		(HI(0) | LO(36))
100 #define ANDI		(HI(12))
101 #define B		(HI(4))
102 #define BAL		(HI(1) | (17 << 16))
103 #define BC1F		(HI(17) | (8 << 21))
104 #define BC1T		(HI(17) | (8 << 21) | (1 << 16))
105 #define BEQ		(HI(4))
106 #define BGEZ		(HI(1) | (1 << 16))
107 #define BGTZ		(HI(7))
108 #define BLEZ		(HI(6))
109 #define BLTZ		(HI(1) | (0 << 16))
110 #define BNE		(HI(5))
111 #define BREAK		(HI(0) | LO(13))
112 #define CFC1		(HI(17) | (2 << 21))
113 #define C_UN_S		(HI(17) | FMT_S | LO(49))
114 #define C_UEQ_S		(HI(17) | FMT_S | LO(51))
115 #define C_ULE_S		(HI(17) | FMT_S | LO(55))
116 #define C_ULT_S		(HI(17) | FMT_S | LO(53))
117 #define CVT_S_S		(HI(17) | FMT_S | LO(32))
118 #define DADDIU		(HI(25))
119 #define DADDU		(HI(0) | LO(45))
120 #define DDIV		(HI(0) | LO(30))
121 #define DDIVU		(HI(0) | LO(31))
122 #define DIV		(HI(0) | LO(26))
123 #define DIVU		(HI(0) | LO(27))
124 #define DIV_S		(HI(17) | FMT_S | LO(3))
125 #define DMULT		(HI(0) | LO(28))
126 #define DMULTU		(HI(0) | LO(29))
127 #define DSLL		(HI(0) | LO(56))
128 #define DSLL32		(HI(0) | LO(60))
129 #define DSLLV		(HI(0) | LO(20))
130 #define DSRA		(HI(0) | LO(59))
131 #define DSRA32		(HI(0) | LO(63))
132 #define DSRAV		(HI(0) | LO(23))
133 #define DSRL		(HI(0) | LO(58))
134 #define DSRL32		(HI(0) | LO(62))
135 #define DSRLV		(HI(0) | LO(22))
136 #define DSUBU		(HI(0) | LO(47))
137 #define J		(HI(2))
138 #define JAL		(HI(3))
139 #define JALR		(HI(0) | LO(9))
140 #define JR		(HI(0) | LO(8))
141 #define LD		(HI(55))
142 #define LUI		(HI(15))
143 #define LW		(HI(35))
144 #define MFC1		(HI(17))
145 #define MFHI		(HI(0) | LO(16))
146 #define MFLO		(HI(0) | LO(18))
147 #define MOV_S		(HI(17) | FMT_S | LO(6))
148 #define MTC1		(HI(17) | (4 << 21))
149 #define MUL_S		(HI(17) | FMT_S | LO(2))
150 #define MULT		(HI(0) | LO(24))
151 #define MULTU		(HI(0) | LO(25))
152 #define NEG_S		(HI(17) | FMT_S | LO(7))
153 #define NOP		(HI(0) | LO(0))
154 #define NOR		(HI(0) | LO(39))
155 #define OR		(HI(0) | LO(37))
156 #define ORI		(HI(13))
157 #define SD		(HI(63))
158 #define SLT		(HI(0) | LO(42))
159 #define SLTI		(HI(10))
160 #define SLTIU		(HI(11))
161 #define SLTU		(HI(0) | LO(43))
162 #define SLL		(HI(0) | LO(0))
163 #define SLLV		(HI(0) | LO(4))
164 #define SRL		(HI(0) | LO(2))
165 #define SRLV		(HI(0) | LO(6))
166 #define SRA		(HI(0) | LO(3))
167 #define SRAV		(HI(0) | LO(7))
168 #define SUB_S		(HI(17) | FMT_S | LO(1))
169 #define SUBU		(HI(0) | LO(35))
170 #define SW		(HI(43))
171 #define TRUNC_W_S	(HI(17) | FMT_S | LO(13))
172 #define XOR		(HI(0) | LO(38))
173 #define XORI		(HI(14))
174 
175 #if (defined SLJIT_MIPS_R1 && SLJIT_MIPS_R1)
176 #define CLZ		(HI(28) | LO(32))
177 #define DCLZ		(HI(28) | LO(36))
178 #define MUL		(HI(28) | LO(2))
179 #define SEB		(HI(31) | (16 << 6) | LO(32))
180 #define SEH		(HI(31) | (24 << 6) | LO(32))
181 #endif
182 
183 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
184 #define ADDU_W		ADDU
185 #define ADDIU_W		ADDIU
186 #define SLL_W		SLL
187 #define SUBU_W		SUBU
188 #else
189 #define ADDU_W		DADDU
190 #define ADDIU_W		DADDIU
191 #define SLL_W		DSLL
192 #define SUBU_W		DSUBU
193 #endif
194 
195 #define SIMM_MAX	(0x7fff)
196 #define SIMM_MIN	(-0x8000)
197 #define UIMM_MAX	(0xffff)
198 
199 /* dest_reg is the absolute name of the register
200    Useful for reordering instructions in the delay slot. */
push_inst(struct sljit_compiler * compiler,sljit_ins ins,sljit_s32 delay_slot)201 static sljit_s32 push_inst(struct sljit_compiler *compiler, sljit_ins ins, sljit_s32 delay_slot)
202 {
203 	SLJIT_ASSERT(delay_slot == MOVABLE_INS || delay_slot >= UNMOVABLE_INS
204 		|| delay_slot == ((ins >> 11) & 0x1f) || delay_slot == ((ins >> 16) & 0x1f));
205 	sljit_ins *ptr = (sljit_ins*)ensure_buf(compiler, sizeof(sljit_ins));
206 	FAIL_IF(!ptr);
207 	*ptr = ins;
208 	compiler->size++;
209 	compiler->delay_slot = delay_slot;
210 	return SLJIT_SUCCESS;
211 }
212 
invert_branch(sljit_s32 flags)213 static SLJIT_INLINE sljit_ins invert_branch(sljit_s32 flags)
214 {
215 	return (flags & IS_BIT26_COND) ? (1 << 26) : (1 << 16);
216 }
217 
detect_jump_type(struct sljit_jump * jump,sljit_ins * code_ptr,sljit_ins * code,sljit_sw executable_offset)218 static SLJIT_INLINE sljit_ins* detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_ins *code, sljit_sw executable_offset)
219 {
220 	sljit_sw diff;
221 	sljit_uw target_addr;
222 	sljit_ins *inst;
223 	sljit_ins saved_inst;
224 
225 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
226 	if (jump->flags & (SLJIT_REWRITABLE_JUMP | IS_CALL))
227 		return code_ptr;
228 #else
229 	if (jump->flags & SLJIT_REWRITABLE_JUMP)
230 		return code_ptr;
231 #endif
232 
233 	if (jump->flags & JUMP_ADDR)
234 		target_addr = jump->u.target;
235 	else {
236 		SLJIT_ASSERT(jump->flags & JUMP_LABEL);
237 		target_addr = (sljit_uw)(code + jump->u.label->size) + (sljit_uw)executable_offset;
238 	}
239 
240 	inst = (sljit_ins *)jump->addr;
241 	if (jump->flags & IS_COND)
242 		inst--;
243 
244 #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
245 	if (jump->flags & IS_CALL)
246 		goto keep_address;
247 #endif
248 
249 	/* B instructions. */
250 	if (jump->flags & IS_MOVABLE) {
251 		diff = ((sljit_sw)target_addr - (sljit_sw)inst - executable_offset) >> 2;
252 		if (diff <= SIMM_MAX && diff >= SIMM_MIN) {
253 			jump->flags |= PATCH_B;
254 
255 			if (!(jump->flags & IS_COND)) {
256 				inst[0] = inst[-1];
257 				inst[-1] = (jump->flags & IS_JAL) ? BAL : B;
258 				jump->addr -= sizeof(sljit_ins);
259 				return inst;
260 			}
261 			saved_inst = inst[0];
262 			inst[0] = inst[-1];
263 			inst[-1] = saved_inst ^ invert_branch(jump->flags);
264 			jump->addr -= 2 * sizeof(sljit_ins);
265 			return inst;
266 		}
267 	}
268 	else {
269 		diff = ((sljit_sw)target_addr - (sljit_sw)(inst + 1) - executable_offset) >> 2;
270 		if (diff <= SIMM_MAX && diff >= SIMM_MIN) {
271 			jump->flags |= PATCH_B;
272 
273 			if (!(jump->flags & IS_COND)) {
274 				inst[0] = (jump->flags & IS_JAL) ? BAL : B;
275 				inst[1] = NOP;
276 				return inst + 1;
277 			}
278 			inst[0] = inst[0] ^ invert_branch(jump->flags);
279 			inst[1] = NOP;
280 			jump->addr -= sizeof(sljit_ins);
281 			return inst + 1;
282 		}
283 	}
284 
285 	if (jump->flags & IS_COND) {
286 		if ((jump->flags & IS_MOVABLE) && (target_addr & ~0xfffffff) == ((jump->addr + 2 * sizeof(sljit_ins)) & ~0xfffffff)) {
287 			jump->flags |= PATCH_J;
288 			saved_inst = inst[0];
289 			inst[0] = inst[-1];
290 			inst[-1] = (saved_inst & 0xffff0000) | 3;
291 			inst[1] = J;
292 			inst[2] = NOP;
293 			return inst + 2;
294 		}
295 		else if ((target_addr & ~0xfffffff) == ((jump->addr + 3 * sizeof(sljit_ins)) & ~0xfffffff)) {
296 			jump->flags |= PATCH_J;
297 			inst[0] = (inst[0] & 0xffff0000) | 3;
298 			inst[1] = NOP;
299 			inst[2] = J;
300 			inst[3] = NOP;
301 			jump->addr += sizeof(sljit_ins);
302 			return inst + 3;
303 		}
304 	}
305 	else {
306 		/* J instuctions. */
307 		if ((jump->flags & IS_MOVABLE) && (target_addr & ~0xfffffff) == (jump->addr & ~0xfffffff)) {
308 			jump->flags |= PATCH_J;
309 			inst[0] = inst[-1];
310 			inst[-1] = (jump->flags & IS_JAL) ? JAL : J;
311 			jump->addr -= sizeof(sljit_ins);
312 			return inst;
313 		}
314 
315 		if ((target_addr & ~0xfffffff) == ((jump->addr + sizeof(sljit_ins)) & ~0xfffffff)) {
316 			jump->flags |= PATCH_J;
317 			inst[0] = (jump->flags & IS_JAL) ? JAL : J;
318 			inst[1] = NOP;
319 			return inst + 1;
320 		}
321 	}
322 
323 #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
324 keep_address:
325 	if (target_addr <= 0x7fffffff) {
326 		jump->flags |= PATCH_ABS32;
327 		if (jump->flags & IS_COND) {
328 			inst[0] -= 4;
329 			inst++;
330 		}
331 		inst[2] = inst[6];
332 		inst[3] = inst[7];
333 		return inst + 3;
334 	}
335 	if (target_addr <= 0x7fffffffffffl) {
336 		jump->flags |= PATCH_ABS48;
337 		if (jump->flags & IS_COND) {
338 			inst[0] -= 2;
339 			inst++;
340 		}
341 		inst[4] = inst[6];
342 		inst[5] = inst[7];
343 		return inst + 5;
344 	}
345 #endif
346 
347 	return code_ptr;
348 }
349 
350 #ifdef __GNUC__
sljit_cache_flush(void * code,void * code_ptr)351 static __attribute__ ((noinline)) void sljit_cache_flush(void* code, void* code_ptr)
352 {
353 	SLJIT_CACHE_FLUSH(code, code_ptr);
354 }
355 #endif
356 
sljit_generate_code(struct sljit_compiler * compiler)357 SLJIT_API_FUNC_ATTRIBUTE void* sljit_generate_code(struct sljit_compiler *compiler)
358 {
359 	struct sljit_memory_fragment *buf;
360 	sljit_ins *code;
361 	sljit_ins *code_ptr;
362 	sljit_ins *buf_ptr;
363 	sljit_ins *buf_end;
364 	sljit_uw word_count;
365 	sljit_sw executable_offset;
366 	sljit_uw addr;
367 
368 	struct sljit_label *label;
369 	struct sljit_jump *jump;
370 	struct sljit_const *const_;
371 
372 	CHECK_ERROR_PTR();
373 	CHECK_PTR(check_sljit_generate_code(compiler));
374 	reverse_buf(compiler);
375 
376 	code = (sljit_ins*)SLJIT_MALLOC_EXEC(compiler->size * sizeof(sljit_ins));
377 	PTR_FAIL_WITH_EXEC_IF(code);
378 	buf = compiler->buf;
379 
380 	code_ptr = code;
381 	word_count = 0;
382 	executable_offset = SLJIT_EXEC_OFFSET(code);
383 
384 	label = compiler->labels;
385 	jump = compiler->jumps;
386 	const_ = compiler->consts;
387 
388 	do {
389 		buf_ptr = (sljit_ins*)buf->memory;
390 		buf_end = buf_ptr + (buf->used_size >> 2);
391 		do {
392 			*code_ptr = *buf_ptr++;
393 			SLJIT_ASSERT(!label || label->size >= word_count);
394 			SLJIT_ASSERT(!jump || jump->addr >= word_count);
395 			SLJIT_ASSERT(!const_ || const_->addr >= word_count);
396 			/* These structures are ordered by their address. */
397 			if (label && label->size == word_count) {
398 				label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset);
399 				label->size = code_ptr - code;
400 				label = label->next;
401 			}
402 			if (jump && jump->addr == word_count) {
403 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
404 				jump->addr = (sljit_uw)(code_ptr - 3);
405 #else
406 				jump->addr = (sljit_uw)(code_ptr - 7);
407 #endif
408 				code_ptr = detect_jump_type(jump, code_ptr, code, executable_offset);
409 				jump = jump->next;
410 			}
411 			if (const_ && const_->addr == word_count) {
412 				/* Just recording the address. */
413 				const_->addr = (sljit_uw)code_ptr;
414 				const_ = const_->next;
415 			}
416 			code_ptr ++;
417 			word_count ++;
418 		} while (buf_ptr < buf_end);
419 
420 		buf = buf->next;
421 	} while (buf);
422 
423 	if (label && label->size == word_count) {
424 		label->addr = (sljit_uw)code_ptr;
425 		label->size = code_ptr - code;
426 		label = label->next;
427 	}
428 
429 	SLJIT_ASSERT(!label);
430 	SLJIT_ASSERT(!jump);
431 	SLJIT_ASSERT(!const_);
432 	SLJIT_ASSERT(code_ptr - code <= (sljit_sw)compiler->size);
433 
434 	jump = compiler->jumps;
435 	while (jump) {
436 		do {
437 			addr = (jump->flags & JUMP_LABEL) ? jump->u.label->addr : jump->u.target;
438 			buf_ptr = (sljit_ins *)jump->addr;
439 
440 			if (jump->flags & PATCH_B) {
441 				addr = (sljit_sw)(addr - ((sljit_uw)SLJIT_ADD_EXEC_OFFSET(buf_ptr, executable_offset) + sizeof(sljit_ins))) >> 2;
442 				SLJIT_ASSERT((sljit_sw)addr <= SIMM_MAX && (sljit_sw)addr >= SIMM_MIN);
443 				buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | (addr & 0xffff);
444 				break;
445 			}
446 			if (jump->flags & PATCH_J) {
447 				SLJIT_ASSERT((addr & ~0xfffffff) == (((sljit_uw)SLJIT_ADD_EXEC_OFFSET(buf_ptr, executable_offset) + sizeof(sljit_ins)) & ~0xfffffff));
448 				buf_ptr[0] |= (addr >> 2) & 0x03ffffff;
449 				break;
450 			}
451 
452 			/* Set the fields of immediate loads. */
453 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
454 			buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | ((addr >> 16) & 0xffff);
455 			buf_ptr[1] = (buf_ptr[1] & 0xffff0000) | (addr & 0xffff);
456 #else
457 			if (jump->flags & PATCH_ABS32) {
458 				SLJIT_ASSERT(addr <= 0x7fffffff);
459 				buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | ((addr >> 16) & 0xffff);
460 				buf_ptr[1] = (buf_ptr[1] & 0xffff0000) | (addr & 0xffff);
461 			}
462 			else if (jump->flags & PATCH_ABS48) {
463 				SLJIT_ASSERT(addr <= 0x7fffffffffffl);
464 				buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | ((addr >> 32) & 0xffff);
465 				buf_ptr[1] = (buf_ptr[1] & 0xffff0000) | ((addr >> 16) & 0xffff);
466 				buf_ptr[3] = (buf_ptr[3] & 0xffff0000) | (addr & 0xffff);
467 			}
468 			else {
469 				buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | ((addr >> 48) & 0xffff);
470 				buf_ptr[1] = (buf_ptr[1] & 0xffff0000) | ((addr >> 32) & 0xffff);
471 				buf_ptr[3] = (buf_ptr[3] & 0xffff0000) | ((addr >> 16) & 0xffff);
472 				buf_ptr[5] = (buf_ptr[5] & 0xffff0000) | (addr & 0xffff);
473 			}
474 #endif
475 		} while (0);
476 		jump = jump->next;
477 	}
478 
479 	compiler->error = SLJIT_ERR_COMPILED;
480 	compiler->executable_offset = executable_offset;
481 	compiler->executable_size = (code_ptr - code) * sizeof(sljit_ins);
482 
483 	code = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(code, executable_offset);
484 	code_ptr = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset);
485 
486 #ifndef __GNUC__
487 	SLJIT_CACHE_FLUSH(code, code_ptr);
488 #else
489 	/* GCC workaround for invalid code generation with -O2. */
490 	sljit_cache_flush(code, code_ptr);
491 #endif
492 	return code;
493 }
494 
495 /* --------------------------------------------------------------------- */
496 /*  Entry, exit                                                          */
497 /* --------------------------------------------------------------------- */
498 
499 /* Creates an index in data_transfer_insts array. */
500 #define LOAD_DATA	0x01
501 #define WORD_DATA	0x00
502 #define BYTE_DATA	0x02
503 #define HALF_DATA	0x04
504 #define INT_DATA	0x06
505 #define SIGNED_DATA	0x08
506 /* Separates integer and floating point registers */
507 #define GPR_REG		0x0f
508 #define DOUBLE_DATA	0x10
509 #define SINGLE_DATA	0x12
510 
511 #define MEM_MASK	0x1f
512 
513 #define WRITE_BACK	0x00020
514 #define ARG_TEST	0x00040
515 #define ALT_KEEP_CACHE	0x00080
516 #define CUMULATIVE_OP	0x00100
517 #define LOGICAL_OP	0x00200
518 #define IMM_OP		0x00400
519 #define SRC2_IMM	0x00800
520 
521 #define UNUSED_DEST	0x01000
522 #define REG_DEST	0x02000
523 #define REG1_SOURCE	0x04000
524 #define REG2_SOURCE	0x08000
525 #define SLOW_SRC1	0x10000
526 #define SLOW_SRC2	0x20000
527 #define SLOW_DEST	0x40000
528 
529 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
530 #define STACK_STORE	SW
531 #define STACK_LOAD	LW
532 #else
533 #define STACK_STORE	SD
534 #define STACK_LOAD	LD
535 #endif
536 
537 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
538 #include "sljitNativeMIPS_32.c"
539 #else
540 #include "sljitNativeMIPS_64.c"
541 #endif
542 
sljit_emit_enter(struct sljit_compiler * compiler,sljit_s32 options,sljit_s32 args,sljit_s32 scratches,sljit_s32 saveds,sljit_s32 fscratches,sljit_s32 fsaveds,sljit_s32 local_size)543 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_enter(struct sljit_compiler *compiler,
544 	sljit_s32 options, sljit_s32 args, sljit_s32 scratches, sljit_s32 saveds,
545 	sljit_s32 fscratches, sljit_s32 fsaveds, sljit_s32 local_size)
546 {
547 	sljit_ins base;
548 	sljit_s32 i, tmp, offs;
549 
550 	CHECK_ERROR();
551 	CHECK(check_sljit_emit_enter(compiler, options, args, scratches, saveds, fscratches, fsaveds, local_size));
552 	set_emit_enter(compiler, options, args, scratches, saveds, fscratches, fsaveds, local_size);
553 
554 	local_size += GET_SAVED_REGISTERS_SIZE(scratches, saveds, 1) + SLJIT_LOCALS_OFFSET;
555 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
556 	local_size = (local_size + 15) & ~0xf;
557 #else
558 	local_size = (local_size + 31) & ~0x1f;
559 #endif
560 	compiler->local_size = local_size;
561 
562 	if (local_size <= SIMM_MAX) {
563 		/* Frequent case. */
564 		FAIL_IF(push_inst(compiler, ADDIU_W | S(SLJIT_SP) | T(SLJIT_SP) | IMM(-local_size), DR(SLJIT_SP)));
565 		base = S(SLJIT_SP);
566 	}
567 	else {
568 		FAIL_IF(load_immediate(compiler, DR(TMP_REG1), local_size));
569 		FAIL_IF(push_inst(compiler, ADDU_W | S(SLJIT_SP) | TA(0) | D(TMP_REG2), DR(TMP_REG2)));
570 		FAIL_IF(push_inst(compiler, SUBU_W | S(SLJIT_SP) | T(TMP_REG1) | D(SLJIT_SP), DR(SLJIT_SP)));
571 		base = S(TMP_REG2);
572 		local_size = 0;
573 	}
574 
575 	offs = local_size - (sljit_sw)(sizeof(sljit_sw));
576 	FAIL_IF(push_inst(compiler, STACK_STORE | base | TA(RETURN_ADDR_REG) | IMM(offs), MOVABLE_INS));
577 
578 	tmp = saveds < SLJIT_NUMBER_OF_SAVED_REGISTERS ? (SLJIT_S0 + 1 - saveds) : SLJIT_FIRST_SAVED_REG;
579 	for (i = SLJIT_S0; i >= tmp; i--) {
580 		offs -= (sljit_s32)(sizeof(sljit_sw));
581 		FAIL_IF(push_inst(compiler, STACK_STORE | base | T(i) | IMM(offs), MOVABLE_INS));
582 	}
583 
584 	for (i = scratches; i >= SLJIT_FIRST_SAVED_REG; i--) {
585 		offs -= (sljit_s32)(sizeof(sljit_sw));
586 		FAIL_IF(push_inst(compiler, STACK_STORE | base | T(i) | IMM(offs), MOVABLE_INS));
587 	}
588 
589 	if (args >= 1)
590 		FAIL_IF(push_inst(compiler, ADDU_W | SA(4) | TA(0) | D(SLJIT_S0), DR(SLJIT_S0)));
591 	if (args >= 2)
592 		FAIL_IF(push_inst(compiler, ADDU_W | SA(5) | TA(0) | D(SLJIT_S1), DR(SLJIT_S1)));
593 	if (args >= 3)
594 		FAIL_IF(push_inst(compiler, ADDU_W | SA(6) | TA(0) | D(SLJIT_S2), DR(SLJIT_S2)));
595 
596 	return SLJIT_SUCCESS;
597 }
598 
sljit_set_context(struct sljit_compiler * compiler,sljit_s32 options,sljit_s32 args,sljit_s32 scratches,sljit_s32 saveds,sljit_s32 fscratches,sljit_s32 fsaveds,sljit_s32 local_size)599 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_set_context(struct sljit_compiler *compiler,
600 	sljit_s32 options, sljit_s32 args, sljit_s32 scratches, sljit_s32 saveds,
601 	sljit_s32 fscratches, sljit_s32 fsaveds, sljit_s32 local_size)
602 {
603 	CHECK_ERROR();
604 	CHECK(check_sljit_set_context(compiler, options, args, scratches, saveds, fscratches, fsaveds, local_size));
605 	set_set_context(compiler, options, args, scratches, saveds, fscratches, fsaveds, local_size);
606 
607 	local_size += GET_SAVED_REGISTERS_SIZE(scratches, saveds, 1) + SLJIT_LOCALS_OFFSET;
608 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
609 	compiler->local_size = (local_size + 15) & ~0xf;
610 #else
611 	compiler->local_size = (local_size + 31) & ~0x1f;
612 #endif
613 	return SLJIT_SUCCESS;
614 }
615 
sljit_emit_return(struct sljit_compiler * compiler,sljit_s32 op,sljit_s32 src,sljit_sw srcw)616 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_return(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 src, sljit_sw srcw)
617 {
618 	sljit_s32 local_size, i, tmp, offs;
619 	sljit_ins base;
620 
621 	CHECK_ERROR();
622 	CHECK(check_sljit_emit_return(compiler, op, src, srcw));
623 
624 	FAIL_IF(emit_mov_before_return(compiler, op, src, srcw));
625 
626 	local_size = compiler->local_size;
627 	if (local_size <= SIMM_MAX)
628 		base = S(SLJIT_SP);
629 	else {
630 		FAIL_IF(load_immediate(compiler, DR(TMP_REG1), local_size));
631 		FAIL_IF(push_inst(compiler, ADDU_W | S(SLJIT_SP) | T(TMP_REG1) | D(TMP_REG1), DR(TMP_REG1)));
632 		base = S(TMP_REG1);
633 		local_size = 0;
634 	}
635 
636 	FAIL_IF(push_inst(compiler, STACK_LOAD | base | TA(RETURN_ADDR_REG) | IMM(local_size - (sljit_s32)sizeof(sljit_sw)), RETURN_ADDR_REG));
637 	offs = local_size - (sljit_s32)GET_SAVED_REGISTERS_SIZE(compiler->scratches, compiler->saveds, 1);
638 
639 	tmp = compiler->scratches;
640 	for (i = SLJIT_FIRST_SAVED_REG; i <= tmp; i++) {
641 		FAIL_IF(push_inst(compiler, STACK_LOAD | base | T(i) | IMM(offs), DR(i)));
642 		offs += (sljit_s32)(sizeof(sljit_sw));
643 	}
644 
645 	tmp = compiler->saveds < SLJIT_NUMBER_OF_SAVED_REGISTERS ? (SLJIT_S0 + 1 - compiler->saveds) : SLJIT_FIRST_SAVED_REG;
646 	for (i = tmp; i <= SLJIT_S0; i++) {
647 		FAIL_IF(push_inst(compiler, STACK_LOAD | base | T(i) | IMM(offs), DR(i)));
648 		offs += (sljit_s32)(sizeof(sljit_sw));
649 	}
650 
651 	SLJIT_ASSERT(offs == local_size - (sljit_sw)(sizeof(sljit_sw)));
652 
653 	FAIL_IF(push_inst(compiler, JR | SA(RETURN_ADDR_REG), UNMOVABLE_INS));
654 	if (compiler->local_size <= SIMM_MAX)
655 		return push_inst(compiler, ADDIU_W | S(SLJIT_SP) | T(SLJIT_SP) | IMM(compiler->local_size), UNMOVABLE_INS);
656 	else
657 		return push_inst(compiler, ADDU_W | S(TMP_REG1) | TA(0) | D(SLJIT_SP), UNMOVABLE_INS);
658 }
659 
660 #undef STACK_STORE
661 #undef STACK_LOAD
662 
663 /* --------------------------------------------------------------------- */
664 /*  Operators                                                            */
665 /* --------------------------------------------------------------------- */
666 
667 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
668 #define ARCH_32_64(a, b)	a
669 #else
670 #define ARCH_32_64(a, b)	b
671 #endif
672 
673 static const sljit_ins data_transfer_insts[16 + 4] = {
674 /* u w s */ ARCH_32_64(HI(43) /* sw */, HI(63) /* sd */),
675 /* u w l */ ARCH_32_64(HI(35) /* lw */, HI(55) /* ld */),
676 /* u b s */ HI(40) /* sb */,
677 /* u b l */ HI(36) /* lbu */,
678 /* u h s */ HI(41) /* sh */,
679 /* u h l */ HI(37) /* lhu */,
680 /* u i s */ HI(43) /* sw */,
681 /* u i l */ ARCH_32_64(HI(35) /* lw */, HI(39) /* lwu */),
682 
683 /* s w s */ ARCH_32_64(HI(43) /* sw */, HI(63) /* sd */),
684 /* s w l */ ARCH_32_64(HI(35) /* lw */, HI(55) /* ld */),
685 /* s b s */ HI(40) /* sb */,
686 /* s b l */ HI(32) /* lb */,
687 /* s h s */ HI(41) /* sh */,
688 /* s h l */ HI(33) /* lh */,
689 /* s i s */ HI(43) /* sw */,
690 /* s i l */ HI(35) /* lw */,
691 
692 /* d   s */ HI(61) /* sdc1 */,
693 /* d   l */ HI(53) /* ldc1 */,
694 /* s   s */ HI(57) /* swc1 */,
695 /* s   l */ HI(49) /* lwc1 */,
696 };
697 
698 #undef ARCH_32_64
699 
700 /* reg_ar is an absoulute register! */
701 
702 /* Can perform an operation using at most 1 instruction. */
getput_arg_fast(struct sljit_compiler * compiler,sljit_s32 flags,sljit_s32 reg_ar,sljit_s32 arg,sljit_sw argw)703 static sljit_s32 getput_arg_fast(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg_ar, sljit_s32 arg, sljit_sw argw)
704 {
705 	SLJIT_ASSERT(arg & SLJIT_MEM);
706 
707 	if ((!(flags & WRITE_BACK) || !(arg & REG_MASK)) && !(arg & OFFS_REG_MASK) && argw <= SIMM_MAX && argw >= SIMM_MIN) {
708 		/* Works for both absoulte and relative addresses. */
709 		if (SLJIT_UNLIKELY(flags & ARG_TEST))
710 			return 1;
711 		FAIL_IF(push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(arg & REG_MASK)
712 			| TA(reg_ar) | IMM(argw), ((flags & MEM_MASK) <= GPR_REG && (flags & LOAD_DATA)) ? reg_ar : MOVABLE_INS));
713 		return -1;
714 	}
715 	return 0;
716 }
717 
718 /* See getput_arg below.
719    Note: can_cache is called only for binary operators. Those
720    operators always uses word arguments without write back. */
can_cache(sljit_s32 arg,sljit_sw argw,sljit_s32 next_arg,sljit_sw next_argw)721 static sljit_s32 can_cache(sljit_s32 arg, sljit_sw argw, sljit_s32 next_arg, sljit_sw next_argw)
722 {
723 	SLJIT_ASSERT((arg & SLJIT_MEM) && (next_arg & SLJIT_MEM));
724 
725 	/* Simple operation except for updates. */
726 	if (arg & OFFS_REG_MASK) {
727 		argw &= 0x3;
728 		next_argw &= 0x3;
729 		if (argw && argw == next_argw && (arg == next_arg || (arg & OFFS_REG_MASK) == (next_arg & OFFS_REG_MASK)))
730 			return 1;
731 		return 0;
732 	}
733 
734 	if (arg == next_arg) {
735 		if (((next_argw - argw) <= SIMM_MAX && (next_argw - argw) >= SIMM_MIN))
736 			return 1;
737 		return 0;
738 	}
739 
740 	return 0;
741 }
742 
743 /* Emit the necessary instructions. See can_cache above. */
getput_arg(struct sljit_compiler * compiler,sljit_s32 flags,sljit_s32 reg_ar,sljit_s32 arg,sljit_sw argw,sljit_s32 next_arg,sljit_sw next_argw)744 static sljit_s32 getput_arg(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg_ar, sljit_s32 arg, sljit_sw argw, sljit_s32 next_arg, sljit_sw next_argw)
745 {
746 	sljit_s32 tmp_ar, base, delay_slot;
747 
748 	SLJIT_ASSERT(arg & SLJIT_MEM);
749 	if (!(next_arg & SLJIT_MEM)) {
750 		next_arg = 0;
751 		next_argw = 0;
752 	}
753 
754 	if ((flags & MEM_MASK) <= GPR_REG && (flags & LOAD_DATA)) {
755 		tmp_ar = reg_ar;
756 		delay_slot = reg_ar;
757 	} else {
758 		tmp_ar = DR(TMP_REG1);
759 		delay_slot = MOVABLE_INS;
760 	}
761 	base = arg & REG_MASK;
762 
763 	if (SLJIT_UNLIKELY(arg & OFFS_REG_MASK)) {
764 		if (SLJIT_UNLIKELY(flags & WRITE_BACK)) {
765 			SLJIT_ASSERT(argw == 0);
766 			FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(OFFS_REG(arg)) | D(base), DR(base)));
767 			return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(base) | TA(reg_ar), delay_slot);
768 		}
769 
770 		argw &= 0x3;
771 
772 		/* Using the cache. */
773 		if (argw == compiler->cache_argw) {
774 			if (arg == compiler->cache_arg)
775 				return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
776 
777 			if ((SLJIT_MEM | (arg & OFFS_REG_MASK)) == compiler->cache_arg) {
778 				if (arg == next_arg && argw == (next_argw & 0x3)) {
779 					compiler->cache_arg = arg;
780 					compiler->cache_argw = argw;
781 					FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(TMP_REG3) | D(TMP_REG3), DR(TMP_REG3)));
782 					return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
783 				}
784 				FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(TMP_REG3) | DA(tmp_ar), tmp_ar));
785 				return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
786 			}
787 		}
788 
789 		if (SLJIT_UNLIKELY(argw)) {
790 			compiler->cache_arg = SLJIT_MEM | (arg & OFFS_REG_MASK);
791 			compiler->cache_argw = argw;
792 			FAIL_IF(push_inst(compiler, SLL_W | T(OFFS_REG(arg)) | D(TMP_REG3) | SH_IMM(argw), DR(TMP_REG3)));
793 		}
794 
795 		if (arg == next_arg && argw == (next_argw & 0x3)) {
796 			compiler->cache_arg = arg;
797 			compiler->cache_argw = argw;
798 			FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(!argw ? OFFS_REG(arg) : TMP_REG3) | D(TMP_REG3), DR(TMP_REG3)));
799 			tmp_ar = DR(TMP_REG3);
800 		}
801 		else
802 			FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(!argw ? OFFS_REG(arg) : TMP_REG3) | DA(tmp_ar), tmp_ar));
803 		return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
804 	}
805 
806 	if (SLJIT_UNLIKELY(flags & WRITE_BACK) && base) {
807 		if (argw <= SIMM_MAX && argw >= SIMM_MIN) {
808 			if (argw)
809 				FAIL_IF(push_inst(compiler, ADDIU_W | S(base) | T(base) | IMM(argw), DR(base)));
810 		}
811 		else {
812 			if (compiler->cache_arg == SLJIT_MEM && argw - compiler->cache_argw <= SIMM_MAX && argw - compiler->cache_argw >= SIMM_MIN) {
813 				if (argw != compiler->cache_argw) {
814 					FAIL_IF(push_inst(compiler, ADDIU_W | S(TMP_REG3) | T(TMP_REG3) | IMM(argw - compiler->cache_argw), DR(TMP_REG3)));
815 					compiler->cache_argw = argw;
816 				}
817 				FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(TMP_REG3) | D(base), DR(base)));
818 			}
819 			else {
820 				compiler->cache_arg = SLJIT_MEM;
821 				compiler->cache_argw = argw;
822 				FAIL_IF(load_immediate(compiler, DR(TMP_REG3), argw));
823 				FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(TMP_REG3) | D(base), DR(base)));
824 			}
825 		}
826 		return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(base) | TA(reg_ar), delay_slot);
827 	}
828 
829 	if (compiler->cache_arg == arg && argw - compiler->cache_argw <= SIMM_MAX && argw - compiler->cache_argw >= SIMM_MIN) {
830 		if (argw != compiler->cache_argw) {
831 			FAIL_IF(push_inst(compiler, ADDIU_W | S(TMP_REG3) | T(TMP_REG3) | IMM(argw - compiler->cache_argw), DR(TMP_REG3)));
832 			compiler->cache_argw = argw;
833 		}
834 		return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
835 	}
836 
837 	if (compiler->cache_arg == SLJIT_MEM && argw - compiler->cache_argw <= SIMM_MAX && argw - compiler->cache_argw >= SIMM_MIN) {
838 		if (argw != compiler->cache_argw)
839 			FAIL_IF(push_inst(compiler, ADDIU_W | S(TMP_REG3) | T(TMP_REG3) | IMM(argw - compiler->cache_argw), DR(TMP_REG3)));
840 	}
841 	else {
842 		compiler->cache_arg = SLJIT_MEM;
843 		FAIL_IF(load_immediate(compiler, DR(TMP_REG3), argw));
844 	}
845 	compiler->cache_argw = argw;
846 
847 	if (!base)
848 		return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
849 
850 	if (arg == next_arg && next_argw - argw <= SIMM_MAX && next_argw - argw >= SIMM_MIN) {
851 		compiler->cache_arg = arg;
852 		FAIL_IF(push_inst(compiler, ADDU_W | S(TMP_REG3) | T(base) | D(TMP_REG3), DR(TMP_REG3)));
853 		return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
854 	}
855 
856 	FAIL_IF(push_inst(compiler, ADDU_W | S(TMP_REG3) | T(base) | DA(tmp_ar), tmp_ar));
857 	return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
858 }
859 
emit_op_mem(struct sljit_compiler * compiler,sljit_s32 flags,sljit_s32 reg_ar,sljit_s32 arg,sljit_sw argw)860 static SLJIT_INLINE sljit_s32 emit_op_mem(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg_ar, sljit_s32 arg, sljit_sw argw)
861 {
862 	if (getput_arg_fast(compiler, flags, reg_ar, arg, argw))
863 		return compiler->error;
864 	compiler->cache_arg = 0;
865 	compiler->cache_argw = 0;
866 	return getput_arg(compiler, flags, reg_ar, arg, argw, 0, 0);
867 }
868 
emit_op_mem2(struct sljit_compiler * compiler,sljit_s32 flags,sljit_s32 reg,sljit_s32 arg1,sljit_sw arg1w,sljit_s32 arg2,sljit_sw arg2w)869 static SLJIT_INLINE sljit_s32 emit_op_mem2(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, sljit_s32 arg1, sljit_sw arg1w, sljit_s32 arg2, sljit_sw arg2w)
870 {
871 	if (getput_arg_fast(compiler, flags, reg, arg1, arg1w))
872 		return compiler->error;
873 	return getput_arg(compiler, flags, reg, arg1, arg1w, arg2, arg2w);
874 }
875 
emit_op(struct sljit_compiler * compiler,sljit_s32 op,sljit_s32 flags,sljit_s32 dst,sljit_sw dstw,sljit_s32 src1,sljit_sw src1w,sljit_s32 src2,sljit_sw src2w)876 static sljit_s32 emit_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 flags,
877 	sljit_s32 dst, sljit_sw dstw,
878 	sljit_s32 src1, sljit_sw src1w,
879 	sljit_s32 src2, sljit_sw src2w)
880 {
881 	/* arg1 goes to TMP_REG1 or src reg
882 	   arg2 goes to TMP_REG2, imm or src reg
883 	   TMP_REG3 can be used for caching
884 	   result goes to TMP_REG2, so put result can use TMP_REG1 and TMP_REG3. */
885 	sljit_s32 dst_r = TMP_REG2;
886 	sljit_s32 src1_r;
887 	sljit_sw src2_r = 0;
888 	sljit_s32 sugg_src2_r = TMP_REG2;
889 
890 	if (!(flags & ALT_KEEP_CACHE)) {
891 		compiler->cache_arg = 0;
892 		compiler->cache_argw = 0;
893 	}
894 
895 	if (SLJIT_UNLIKELY(dst == SLJIT_UNUSED)) {
896 		if (op >= SLJIT_MOV && op <= SLJIT_MOVU_S32 && !(src2 & SLJIT_MEM))
897 			return SLJIT_SUCCESS;
898 		if (HAS_FLAGS(op))
899 			flags |= UNUSED_DEST;
900 	}
901 	else if (FAST_IS_REG(dst)) {
902 		dst_r = dst;
903 		flags |= REG_DEST;
904 		if (op >= SLJIT_MOV && op <= SLJIT_MOVU_S32)
905 			sugg_src2_r = dst_r;
906 	}
907 	else if ((dst & SLJIT_MEM) && !getput_arg_fast(compiler, flags | ARG_TEST, DR(TMP_REG1), dst, dstw))
908 		flags |= SLOW_DEST;
909 
910 	if (flags & IMM_OP) {
911 		if ((src2 & SLJIT_IMM) && src2w) {
912 			if ((!(flags & LOGICAL_OP) && (src2w <= SIMM_MAX && src2w >= SIMM_MIN))
913 				|| ((flags & LOGICAL_OP) && !(src2w & ~UIMM_MAX))) {
914 				flags |= SRC2_IMM;
915 				src2_r = src2w;
916 			}
917 		}
918 		if (!(flags & SRC2_IMM) && (flags & CUMULATIVE_OP) && (src1 & SLJIT_IMM) && src1w) {
919 			if ((!(flags & LOGICAL_OP) && (src1w <= SIMM_MAX && src1w >= SIMM_MIN))
920 				|| ((flags & LOGICAL_OP) && !(src1w & ~UIMM_MAX))) {
921 				flags |= SRC2_IMM;
922 				src2_r = src1w;
923 
924 				/* And swap arguments. */
925 				src1 = src2;
926 				src1w = src2w;
927 				src2 = SLJIT_IMM;
928 				/* src2w = src2_r unneeded. */
929 			}
930 		}
931 	}
932 
933 	/* Source 1. */
934 	if (FAST_IS_REG(src1)) {
935 		src1_r = src1;
936 		flags |= REG1_SOURCE;
937 	}
938 	else if (src1 & SLJIT_IMM) {
939 		if (src1w) {
940 			FAIL_IF(load_immediate(compiler, DR(TMP_REG1), src1w));
941 			src1_r = TMP_REG1;
942 		}
943 		else
944 			src1_r = 0;
945 	}
946 	else {
947 		if (getput_arg_fast(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w))
948 			FAIL_IF(compiler->error);
949 		else
950 			flags |= SLOW_SRC1;
951 		src1_r = TMP_REG1;
952 	}
953 
954 	/* Source 2. */
955 	if (FAST_IS_REG(src2)) {
956 		src2_r = src2;
957 		flags |= REG2_SOURCE;
958 		if (!(flags & REG_DEST) && op >= SLJIT_MOV && op <= SLJIT_MOVU_S32)
959 			dst_r = src2_r;
960 	}
961 	else if (src2 & SLJIT_IMM) {
962 		if (!(flags & SRC2_IMM)) {
963 			if (src2w) {
964 				FAIL_IF(load_immediate(compiler, DR(sugg_src2_r), src2w));
965 				src2_r = sugg_src2_r;
966 			}
967 			else {
968 				src2_r = 0;
969 				if ((op >= SLJIT_MOV && op <= SLJIT_MOVU_S32) && (dst & SLJIT_MEM))
970 					dst_r = 0;
971 			}
972 		}
973 	}
974 	else {
975 		if (getput_arg_fast(compiler, flags | LOAD_DATA, DR(sugg_src2_r), src2, src2w))
976 			FAIL_IF(compiler->error);
977 		else
978 			flags |= SLOW_SRC2;
979 		src2_r = sugg_src2_r;
980 	}
981 
982 	if ((flags & (SLOW_SRC1 | SLOW_SRC2)) == (SLOW_SRC1 | SLOW_SRC2)) {
983 		SLJIT_ASSERT(src2_r == TMP_REG2);
984 		if (!can_cache(src1, src1w, src2, src2w) && can_cache(src1, src1w, dst, dstw)) {
985 			FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG2), src2, src2w, src1, src1w));
986 			FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w, dst, dstw));
987 		}
988 		else {
989 			FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w, src2, src2w));
990 			FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG2), src2, src2w, dst, dstw));
991 		}
992 	}
993 	else if (flags & SLOW_SRC1)
994 		FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w, dst, dstw));
995 	else if (flags & SLOW_SRC2)
996 		FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(sugg_src2_r), src2, src2w, dst, dstw));
997 
998 	FAIL_IF(emit_single_op(compiler, op, flags, dst_r, src1_r, src2_r));
999 
1000 	if (dst & SLJIT_MEM) {
1001 		if (!(flags & SLOW_DEST)) {
1002 			getput_arg_fast(compiler, flags, DR(dst_r), dst, dstw);
1003 			return compiler->error;
1004 		}
1005 		return getput_arg(compiler, flags, DR(dst_r), dst, dstw, 0, 0);
1006 	}
1007 
1008 	return SLJIT_SUCCESS;
1009 }
1010 
sljit_emit_op0(struct sljit_compiler * compiler,sljit_s32 op)1011 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op0(struct sljit_compiler *compiler, sljit_s32 op)
1012 {
1013 #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
1014 	sljit_s32 int_op = op & SLJIT_I32_OP;
1015 #endif
1016 
1017 	CHECK_ERROR();
1018 	CHECK(check_sljit_emit_op0(compiler, op));
1019 
1020 	op = GET_OPCODE(op);
1021 	switch (op) {
1022 	case SLJIT_BREAKPOINT:
1023 		return push_inst(compiler, BREAK, UNMOVABLE_INS);
1024 	case SLJIT_NOP:
1025 		return push_inst(compiler, NOP, UNMOVABLE_INS);
1026 	case SLJIT_LMUL_UW:
1027 	case SLJIT_LMUL_SW:
1028 #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
1029 		FAIL_IF(push_inst(compiler, (op == SLJIT_LMUL_UW ? DMULTU : DMULT) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
1030 #else
1031 		FAIL_IF(push_inst(compiler, (op == SLJIT_LMUL_UW ? MULTU : MULT) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
1032 #endif
1033 		FAIL_IF(push_inst(compiler, MFLO | D(SLJIT_R0), DR(SLJIT_R0)));
1034 		return push_inst(compiler, MFHI | D(SLJIT_R1), DR(SLJIT_R1));
1035 	case SLJIT_DIVMOD_UW:
1036 	case SLJIT_DIVMOD_SW:
1037 	case SLJIT_DIV_UW:
1038 	case SLJIT_DIV_SW:
1039 		SLJIT_COMPILE_ASSERT((SLJIT_DIVMOD_UW & 0x2) == 0 && SLJIT_DIV_UW - 0x2 == SLJIT_DIVMOD_UW, bad_div_opcode_assignments);
1040 #if !(defined SLJIT_MIPS_R1 && SLJIT_MIPS_R1)
1041 		FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
1042 		FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
1043 #endif
1044 
1045 #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
1046 		if (int_op)
1047 			FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DIVU : DIV) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
1048 		else
1049 			FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DDIVU : DDIV) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
1050 #else
1051 		FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DIVU : DIV) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
1052 #endif
1053 
1054 		FAIL_IF(push_inst(compiler, MFLO | D(SLJIT_R0), DR(SLJIT_R0)));
1055 		return (op >= SLJIT_DIV_UW) ? SLJIT_SUCCESS : push_inst(compiler, MFHI | D(SLJIT_R1), DR(SLJIT_R1));
1056 	}
1057 
1058 	return SLJIT_SUCCESS;
1059 }
1060 
sljit_emit_op1(struct sljit_compiler * compiler,sljit_s32 op,sljit_s32 dst,sljit_sw dstw,sljit_s32 src,sljit_sw srcw)1061 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op1(struct sljit_compiler *compiler, sljit_s32 op,
1062 	sljit_s32 dst, sljit_sw dstw,
1063 	sljit_s32 src, sljit_sw srcw)
1064 {
1065 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1066 #	define flags 0
1067 #else
1068 	sljit_s32 flags = 0;
1069 #endif
1070 
1071 	CHECK_ERROR();
1072 	CHECK(check_sljit_emit_op1(compiler, op, dst, dstw, src, srcw));
1073 	ADJUST_LOCAL_OFFSET(dst, dstw);
1074 	ADJUST_LOCAL_OFFSET(src, srcw);
1075 
1076 #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
1077 	if ((op & SLJIT_I32_OP) && GET_OPCODE(op) >= SLJIT_NOT) {
1078 		flags |= INT_DATA | SIGNED_DATA;
1079 		if (src & SLJIT_IMM)
1080 			srcw = (sljit_s32)srcw;
1081 	}
1082 #endif
1083 
1084 	switch (GET_OPCODE(op)) {
1085 	case SLJIT_MOV:
1086 	case SLJIT_MOV_P:
1087 		return emit_op(compiler, SLJIT_MOV, WORD_DATA, dst, dstw, TMP_REG1, 0, src, srcw);
1088 
1089 	case SLJIT_MOV_U32:
1090 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1091 		return emit_op(compiler, SLJIT_MOV_U32, INT_DATA, dst, dstw, TMP_REG1, 0, src, srcw);
1092 #else
1093 		return emit_op(compiler, SLJIT_MOV_U32, INT_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_u32)srcw : srcw);
1094 #endif
1095 
1096 	case SLJIT_MOV_S32:
1097 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1098 		return emit_op(compiler, SLJIT_MOV_S32, INT_DATA | SIGNED_DATA, dst, dstw, TMP_REG1, 0, src, srcw);
1099 #else
1100 		return emit_op(compiler, SLJIT_MOV_S32, INT_DATA | SIGNED_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_s32)srcw : srcw);
1101 #endif
1102 
1103 	case SLJIT_MOV_U8:
1104 		return emit_op(compiler, SLJIT_MOV_U8, BYTE_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_u8)srcw : srcw);
1105 
1106 	case SLJIT_MOV_S8:
1107 		return emit_op(compiler, SLJIT_MOV_S8, BYTE_DATA | SIGNED_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_s8)srcw : srcw);
1108 
1109 	case SLJIT_MOV_U16:
1110 		return emit_op(compiler, SLJIT_MOV_U16, HALF_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_u16)srcw : srcw);
1111 
1112 	case SLJIT_MOV_S16:
1113 		return emit_op(compiler, SLJIT_MOV_S16, HALF_DATA | SIGNED_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_s16)srcw : srcw);
1114 
1115 	case SLJIT_MOVU:
1116 	case SLJIT_MOVU_P:
1117 		return emit_op(compiler, SLJIT_MOV, WORD_DATA | WRITE_BACK, dst, dstw, TMP_REG1, 0, src, srcw);
1118 
1119 	case SLJIT_MOVU_U32:
1120 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1121 		return emit_op(compiler, SLJIT_MOV_U32, INT_DATA | WRITE_BACK, dst, dstw, TMP_REG1, 0, src, srcw);
1122 #else
1123 		return emit_op(compiler, SLJIT_MOV_U32, INT_DATA | WRITE_BACK, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_u32)srcw : srcw);
1124 #endif
1125 
1126 	case SLJIT_MOVU_S32:
1127 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1128 		return emit_op(compiler, SLJIT_MOV_S32, INT_DATA | SIGNED_DATA | WRITE_BACK, dst, dstw, TMP_REG1, 0, src, srcw);
1129 #else
1130 		return emit_op(compiler, SLJIT_MOV_S32, INT_DATA | SIGNED_DATA | WRITE_BACK, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_s32)srcw : srcw);
1131 #endif
1132 
1133 	case SLJIT_MOVU_U8:
1134 		return emit_op(compiler, SLJIT_MOV_U8, BYTE_DATA | WRITE_BACK, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_u8)srcw : srcw);
1135 
1136 	case SLJIT_MOVU_S8:
1137 		return emit_op(compiler, SLJIT_MOV_S8, BYTE_DATA | SIGNED_DATA | WRITE_BACK, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_s8)srcw : srcw);
1138 
1139 	case SLJIT_MOVU_U16:
1140 		return emit_op(compiler, SLJIT_MOV_U16, HALF_DATA | WRITE_BACK, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_u16)srcw : srcw);
1141 
1142 	case SLJIT_MOVU_S16:
1143 		return emit_op(compiler, SLJIT_MOV_S16, HALF_DATA | SIGNED_DATA | WRITE_BACK, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_s16)srcw : srcw);
1144 
1145 	case SLJIT_NOT:
1146 		return emit_op(compiler, op, flags, dst, dstw, TMP_REG1, 0, src, srcw);
1147 
1148 	case SLJIT_NEG:
1149 		return emit_op(compiler, SLJIT_SUB | GET_ALL_FLAGS(op), flags | IMM_OP, dst, dstw, SLJIT_IMM, 0, src, srcw);
1150 
1151 	case SLJIT_CLZ:
1152 		return emit_op(compiler, op, flags, dst, dstw, TMP_REG1, 0, src, srcw);
1153 	}
1154 
1155 	return SLJIT_SUCCESS;
1156 
1157 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1158 #	undef flags
1159 #endif
1160 }
1161 
sljit_emit_op2(struct sljit_compiler * compiler,sljit_s32 op,sljit_s32 dst,sljit_sw dstw,sljit_s32 src1,sljit_sw src1w,sljit_s32 src2,sljit_sw src2w)1162 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op2(struct sljit_compiler *compiler, sljit_s32 op,
1163 	sljit_s32 dst, sljit_sw dstw,
1164 	sljit_s32 src1, sljit_sw src1w,
1165 	sljit_s32 src2, sljit_sw src2w)
1166 {
1167 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1168 #	define flags 0
1169 #else
1170 	sljit_s32 flags = 0;
1171 #endif
1172 
1173 	CHECK_ERROR();
1174 	CHECK(check_sljit_emit_op2(compiler, op, dst, dstw, src1, src1w, src2, src2w));
1175 	ADJUST_LOCAL_OFFSET(dst, dstw);
1176 	ADJUST_LOCAL_OFFSET(src1, src1w);
1177 	ADJUST_LOCAL_OFFSET(src2, src2w);
1178 
1179 #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
1180 	if (op & SLJIT_I32_OP) {
1181 		flags |= INT_DATA | SIGNED_DATA;
1182 		if (src1 & SLJIT_IMM)
1183 			src1w = (sljit_s32)src1w;
1184 		if (src2 & SLJIT_IMM)
1185 			src2w = (sljit_s32)src2w;
1186 	}
1187 #endif
1188 
1189 	switch (GET_OPCODE(op)) {
1190 	case SLJIT_ADD:
1191 	case SLJIT_ADDC:
1192 		return emit_op(compiler, op, flags | CUMULATIVE_OP | IMM_OP, dst, dstw, src1, src1w, src2, src2w);
1193 
1194 	case SLJIT_SUB:
1195 	case SLJIT_SUBC:
1196 		return emit_op(compiler, op, flags | IMM_OP, dst, dstw, src1, src1w, src2, src2w);
1197 
1198 	case SLJIT_MUL:
1199 		return emit_op(compiler, op, flags | CUMULATIVE_OP, dst, dstw, src1, src1w, src2, src2w);
1200 
1201 	case SLJIT_AND:
1202 	case SLJIT_OR:
1203 	case SLJIT_XOR:
1204 		return emit_op(compiler, op, flags | CUMULATIVE_OP | LOGICAL_OP | IMM_OP, dst, dstw, src1, src1w, src2, src2w);
1205 
1206 	case SLJIT_SHL:
1207 	case SLJIT_LSHR:
1208 	case SLJIT_ASHR:
1209 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1210 		if (src2 & SLJIT_IMM)
1211 			src2w &= 0x1f;
1212 #else
1213 		if (src2 & SLJIT_IMM) {
1214 			if (op & SLJIT_I32_OP)
1215 				src2w &= 0x1f;
1216 			else
1217 				src2w &= 0x3f;
1218 		}
1219 #endif
1220 		return emit_op(compiler, op, flags | IMM_OP, dst, dstw, src1, src1w, src2, src2w);
1221 	}
1222 
1223 	return SLJIT_SUCCESS;
1224 
1225 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1226 #	undef flags
1227 #endif
1228 }
1229 
sljit_get_register_index(sljit_s32 reg)1230 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_register_index(sljit_s32 reg)
1231 {
1232 	CHECK_REG_INDEX(check_sljit_get_register_index(reg));
1233 	return reg_map[reg];
1234 }
1235 
sljit_get_float_register_index(sljit_s32 reg)1236 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_float_register_index(sljit_s32 reg)
1237 {
1238 	CHECK_REG_INDEX(check_sljit_get_float_register_index(reg));
1239 	return reg << 1;
1240 }
1241 
sljit_emit_op_custom(struct sljit_compiler * compiler,void * instruction,sljit_s32 size)1242 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_custom(struct sljit_compiler *compiler,
1243 	void *instruction, sljit_s32 size)
1244 {
1245 	CHECK_ERROR();
1246 	CHECK(check_sljit_emit_op_custom(compiler, instruction, size));
1247 
1248 	return push_inst(compiler, *(sljit_ins*)instruction, UNMOVABLE_INS);
1249 }
1250 
1251 /* --------------------------------------------------------------------- */
1252 /*  Floating point operators                                             */
1253 /* --------------------------------------------------------------------- */
1254 
sljit_is_fpu_available(void)1255 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_is_fpu_available(void)
1256 {
1257 #ifdef SLJIT_IS_FPU_AVAILABLE
1258 	return SLJIT_IS_FPU_AVAILABLE;
1259 #elif defined(__GNUC__)
1260 	sljit_sw fir;
1261 	asm ("cfc1 %0, $0" : "=r"(fir));
1262 	return (fir >> 22) & 0x1;
1263 #else
1264 #error "FIR check is not implemented for this architecture"
1265 #endif
1266 }
1267 
1268 #define FLOAT_DATA(op) (DOUBLE_DATA | ((op & SLJIT_F32_OP) >> 7))
1269 #define FMT(op) (((op & SLJIT_F32_OP) ^ SLJIT_F32_OP) << (21 - 8))
1270 
sljit_emit_fop1_conv_sw_from_f64(struct sljit_compiler * compiler,sljit_s32 op,sljit_s32 dst,sljit_sw dstw,sljit_s32 src,sljit_sw srcw)1271 static SLJIT_INLINE sljit_s32 sljit_emit_fop1_conv_sw_from_f64(struct sljit_compiler *compiler, sljit_s32 op,
1272 	sljit_s32 dst, sljit_sw dstw,
1273 	sljit_s32 src, sljit_sw srcw)
1274 {
1275 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1276 #	define flags 0
1277 #else
1278 	sljit_s32 flags = (GET_OPCODE(op) == SLJIT_CONV_SW_FROM_F64) << 21;
1279 #endif
1280 
1281 	if (src & SLJIT_MEM) {
1282 		FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op) | LOAD_DATA, TMP_FREG1, src, srcw, dst, dstw));
1283 		src = TMP_FREG1;
1284 	}
1285 	else
1286 		src <<= 1;
1287 
1288 	FAIL_IF(push_inst(compiler, (TRUNC_W_S ^ (flags >> 19)) | FMT(op) | FS(src) | FD(TMP_FREG1), MOVABLE_INS));
1289 
1290 	if (dst == SLJIT_UNUSED)
1291 		return SLJIT_SUCCESS;
1292 
1293 	if (FAST_IS_REG(dst))
1294 		return push_inst(compiler, MFC1 | flags | T(dst) | FS(TMP_FREG1), MOVABLE_INS);
1295 
1296 	/* Store the integer value from a VFP register. */
1297 	return emit_op_mem2(compiler, flags ? DOUBLE_DATA : SINGLE_DATA, TMP_FREG1, dst, dstw, 0, 0);
1298 
1299 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1300 #	undef is_long
1301 #endif
1302 }
1303 
sljit_emit_fop1_conv_f64_from_sw(struct sljit_compiler * compiler,sljit_s32 op,sljit_s32 dst,sljit_sw dstw,sljit_s32 src,sljit_sw srcw)1304 static SLJIT_INLINE sljit_s32 sljit_emit_fop1_conv_f64_from_sw(struct sljit_compiler *compiler, sljit_s32 op,
1305 	sljit_s32 dst, sljit_sw dstw,
1306 	sljit_s32 src, sljit_sw srcw)
1307 {
1308 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1309 #	define flags 0
1310 #else
1311 	sljit_s32 flags = (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_SW) << 21;
1312 #endif
1313 
1314 	sljit_s32 dst_r = FAST_IS_REG(dst) ? (dst << 1) : TMP_FREG1;
1315 
1316 	if (FAST_IS_REG(src))
1317 		FAIL_IF(push_inst(compiler, MTC1 | flags | T(src) | FS(TMP_FREG1), MOVABLE_INS));
1318 	else if (src & SLJIT_MEM) {
1319 		/* Load the integer value into a VFP register. */
1320 		FAIL_IF(emit_op_mem2(compiler, ((flags) ? DOUBLE_DATA : SINGLE_DATA) | LOAD_DATA, TMP_FREG1, src, srcw, dst, dstw));
1321 	}
1322 	else {
1323 #if (defined SLJIT_CONFIG_X86_64 && SLJIT_CONFIG_X86_64)
1324 		if (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_S32)
1325 			srcw = (sljit_s32)srcw;
1326 #endif
1327 		FAIL_IF(load_immediate(compiler, DR(TMP_REG1), srcw));
1328 		FAIL_IF(push_inst(compiler, MTC1 | flags | T(TMP_REG1) | FS(TMP_FREG1), MOVABLE_INS));
1329 	}
1330 
1331 	FAIL_IF(push_inst(compiler, CVT_S_S | flags | (4 << 21) | (((op & SLJIT_F32_OP) ^ SLJIT_F32_OP) >> 8) | FS(TMP_FREG1) | FD(dst_r), MOVABLE_INS));
1332 
1333 	if (dst & SLJIT_MEM)
1334 		return emit_op_mem2(compiler, FLOAT_DATA(op), TMP_FREG1, dst, dstw, 0, 0);
1335 	return SLJIT_SUCCESS;
1336 
1337 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1338 #	undef flags
1339 #endif
1340 }
1341 
sljit_emit_fop1_cmp(struct sljit_compiler * compiler,sljit_s32 op,sljit_s32 src1,sljit_sw src1w,sljit_s32 src2,sljit_sw src2w)1342 static SLJIT_INLINE sljit_s32 sljit_emit_fop1_cmp(struct sljit_compiler *compiler, sljit_s32 op,
1343 	sljit_s32 src1, sljit_sw src1w,
1344 	sljit_s32 src2, sljit_sw src2w)
1345 {
1346 	sljit_ins inst;
1347 
1348 	if (src1 & SLJIT_MEM) {
1349 		FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op) | LOAD_DATA, TMP_FREG1, src1, src1w, src2, src2w));
1350 		src1 = TMP_FREG1;
1351 	}
1352 	else
1353 		src1 <<= 1;
1354 
1355 	if (src2 & SLJIT_MEM) {
1356 		FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op) | LOAD_DATA, TMP_FREG2, src2, src2w, 0, 0));
1357 		src2 = TMP_FREG2;
1358 	}
1359 	else
1360 		src2 <<= 1;
1361 
1362 	switch (GET_FLAG_TYPE(op)) {
1363 	case SLJIT_EQUAL_F64:
1364 	case SLJIT_NOT_EQUAL_F64:
1365 		inst = C_UEQ_S;
1366 		break;
1367 	case SLJIT_LESS_F64:
1368 	case SLJIT_GREATER_EQUAL_F64:
1369 		inst = C_ULT_S;
1370 		break;
1371 	case SLJIT_GREATER_F64:
1372 	case SLJIT_LESS_EQUAL_F64:
1373 		inst = C_ULE_S;
1374 		break;
1375 	default:
1376 		SLJIT_ASSERT(GET_FLAG_TYPE(op) == SLJIT_UNORDERED_F64 || GET_FLAG_TYPE(op) == SLJIT_ORDERED_F64);
1377 		inst = C_UN_S;
1378 		break;
1379 	}
1380 
1381 	return push_inst(compiler, inst | FMT(op) | FT(src2) | FS(src1), UNMOVABLE_INS);
1382 }
1383 
sljit_emit_fop1(struct sljit_compiler * compiler,sljit_s32 op,sljit_s32 dst,sljit_sw dstw,sljit_s32 src,sljit_sw srcw)1384 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fop1(struct sljit_compiler *compiler, sljit_s32 op,
1385 	sljit_s32 dst, sljit_sw dstw,
1386 	sljit_s32 src, sljit_sw srcw)
1387 {
1388 	sljit_s32 dst_r;
1389 
1390 	CHECK_ERROR();
1391 	compiler->cache_arg = 0;
1392 	compiler->cache_argw = 0;
1393 
1394 	SLJIT_COMPILE_ASSERT((SLJIT_F32_OP == 0x100) && !(DOUBLE_DATA & 0x2), float_transfer_bit_error);
1395 	SELECT_FOP1_OPERATION_WITH_CHECKS(compiler, op, dst, dstw, src, srcw);
1396 
1397 	if (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_F32)
1398 		op ^= SLJIT_F32_OP;
1399 
1400 	dst_r = FAST_IS_REG(dst) ? (dst << 1) : TMP_FREG1;
1401 
1402 	if (src & SLJIT_MEM) {
1403 		FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op) | LOAD_DATA, dst_r, src, srcw, dst, dstw));
1404 		src = dst_r;
1405 	}
1406 	else
1407 		src <<= 1;
1408 
1409 	switch (GET_OPCODE(op)) {
1410 	case SLJIT_MOV_F64:
1411 		if (src != dst_r) {
1412 			if (dst_r != TMP_FREG1)
1413 				FAIL_IF(push_inst(compiler, MOV_S | FMT(op) | FS(src) | FD(dst_r), MOVABLE_INS));
1414 			else
1415 				dst_r = src;
1416 		}
1417 		break;
1418 	case SLJIT_NEG_F64:
1419 		FAIL_IF(push_inst(compiler, NEG_S | FMT(op) | FS(src) | FD(dst_r), MOVABLE_INS));
1420 		break;
1421 	case SLJIT_ABS_F64:
1422 		FAIL_IF(push_inst(compiler, ABS_S | FMT(op) | FS(src) | FD(dst_r), MOVABLE_INS));
1423 		break;
1424 	case SLJIT_CONV_F64_FROM_F32:
1425 		FAIL_IF(push_inst(compiler, CVT_S_S | ((op & SLJIT_F32_OP) ? 1 : (1 << 21)) | FS(src) | FD(dst_r), MOVABLE_INS));
1426 		op ^= SLJIT_F32_OP;
1427 		break;
1428 	}
1429 
1430 	if (dst & SLJIT_MEM)
1431 		return emit_op_mem2(compiler, FLOAT_DATA(op), dst_r, dst, dstw, 0, 0);
1432 	return SLJIT_SUCCESS;
1433 }
1434 
sljit_emit_fop2(struct sljit_compiler * compiler,sljit_s32 op,sljit_s32 dst,sljit_sw dstw,sljit_s32 src1,sljit_sw src1w,sljit_s32 src2,sljit_sw src2w)1435 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fop2(struct sljit_compiler *compiler, sljit_s32 op,
1436 	sljit_s32 dst, sljit_sw dstw,
1437 	sljit_s32 src1, sljit_sw src1w,
1438 	sljit_s32 src2, sljit_sw src2w)
1439 {
1440 	sljit_s32 dst_r, flags = 0;
1441 
1442 	CHECK_ERROR();
1443 	CHECK(check_sljit_emit_fop2(compiler, op, dst, dstw, src1, src1w, src2, src2w));
1444 	ADJUST_LOCAL_OFFSET(dst, dstw);
1445 	ADJUST_LOCAL_OFFSET(src1, src1w);
1446 	ADJUST_LOCAL_OFFSET(src2, src2w);
1447 
1448 	compiler->cache_arg = 0;
1449 	compiler->cache_argw = 0;
1450 
1451 	dst_r = FAST_IS_REG(dst) ? (dst << 1) : TMP_FREG2;
1452 
1453 	if (src1 & SLJIT_MEM) {
1454 		if (getput_arg_fast(compiler, FLOAT_DATA(op) | LOAD_DATA, TMP_FREG1, src1, src1w)) {
1455 			FAIL_IF(compiler->error);
1456 			src1 = TMP_FREG1;
1457 		} else
1458 			flags |= SLOW_SRC1;
1459 	}
1460 	else
1461 		src1 <<= 1;
1462 
1463 	if (src2 & SLJIT_MEM) {
1464 		if (getput_arg_fast(compiler, FLOAT_DATA(op) | LOAD_DATA, TMP_FREG2, src2, src2w)) {
1465 			FAIL_IF(compiler->error);
1466 			src2 = TMP_FREG2;
1467 		} else
1468 			flags |= SLOW_SRC2;
1469 	}
1470 	else
1471 		src2 <<= 1;
1472 
1473 	if ((flags & (SLOW_SRC1 | SLOW_SRC2)) == (SLOW_SRC1 | SLOW_SRC2)) {
1474 		if (!can_cache(src1, src1w, src2, src2w) && can_cache(src1, src1w, dst, dstw)) {
1475 			FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, TMP_FREG2, src2, src2w, src1, src1w));
1476 			FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, TMP_FREG1, src1, src1w, dst, dstw));
1477 		}
1478 		else {
1479 			FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, TMP_FREG1, src1, src1w, src2, src2w));
1480 			FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, TMP_FREG2, src2, src2w, dst, dstw));
1481 		}
1482 	}
1483 	else if (flags & SLOW_SRC1)
1484 		FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, TMP_FREG1, src1, src1w, dst, dstw));
1485 	else if (flags & SLOW_SRC2)
1486 		FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, TMP_FREG2, src2, src2w, dst, dstw));
1487 
1488 	if (flags & SLOW_SRC1)
1489 		src1 = TMP_FREG1;
1490 	if (flags & SLOW_SRC2)
1491 		src2 = TMP_FREG2;
1492 
1493 	switch (GET_OPCODE(op)) {
1494 	case SLJIT_ADD_F64:
1495 		FAIL_IF(push_inst(compiler, ADD_S | FMT(op) | FT(src2) | FS(src1) | FD(dst_r), MOVABLE_INS));
1496 		break;
1497 
1498 	case SLJIT_SUB_F64:
1499 		FAIL_IF(push_inst(compiler, SUB_S | FMT(op) | FT(src2) | FS(src1) | FD(dst_r), MOVABLE_INS));
1500 		break;
1501 
1502 	case SLJIT_MUL_F64:
1503 		FAIL_IF(push_inst(compiler, MUL_S | FMT(op) | FT(src2) | FS(src1) | FD(dst_r), MOVABLE_INS));
1504 		break;
1505 
1506 	case SLJIT_DIV_F64:
1507 		FAIL_IF(push_inst(compiler, DIV_S | FMT(op) | FT(src2) | FS(src1) | FD(dst_r), MOVABLE_INS));
1508 		break;
1509 	}
1510 
1511 	if (dst_r == TMP_FREG2)
1512 		FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op), TMP_FREG2, dst, dstw, 0, 0));
1513 
1514 	return SLJIT_SUCCESS;
1515 }
1516 
1517 /* --------------------------------------------------------------------- */
1518 /*  Other instructions                                                   */
1519 /* --------------------------------------------------------------------- */
1520 
sljit_emit_fast_enter(struct sljit_compiler * compiler,sljit_s32 dst,sljit_sw dstw)1521 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fast_enter(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw)
1522 {
1523 	CHECK_ERROR();
1524 	CHECK(check_sljit_emit_fast_enter(compiler, dst, dstw));
1525 	ADJUST_LOCAL_OFFSET(dst, dstw);
1526 
1527 	/* For UNUSED dst. Uncommon, but possible. */
1528 	if (dst == SLJIT_UNUSED)
1529 		return SLJIT_SUCCESS;
1530 
1531 	if (FAST_IS_REG(dst))
1532 		return push_inst(compiler, ADDU_W | SA(RETURN_ADDR_REG) | TA(0) | D(dst), DR(dst));
1533 
1534 	/* Memory. */
1535 	return emit_op_mem(compiler, WORD_DATA, RETURN_ADDR_REG, dst, dstw);
1536 }
1537 
sljit_emit_fast_return(struct sljit_compiler * compiler,sljit_s32 src,sljit_sw srcw)1538 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fast_return(struct sljit_compiler *compiler, sljit_s32 src, sljit_sw srcw)
1539 {
1540 	CHECK_ERROR();
1541 	CHECK(check_sljit_emit_fast_return(compiler, src, srcw));
1542 	ADJUST_LOCAL_OFFSET(src, srcw);
1543 
1544 	if (FAST_IS_REG(src))
1545 		FAIL_IF(push_inst(compiler, ADDU_W | S(src) | TA(0) | DA(RETURN_ADDR_REG), RETURN_ADDR_REG));
1546 	else if (src & SLJIT_MEM)
1547 		FAIL_IF(emit_op_mem(compiler, WORD_DATA | LOAD_DATA, RETURN_ADDR_REG, src, srcw));
1548 	else if (src & SLJIT_IMM)
1549 		FAIL_IF(load_immediate(compiler, RETURN_ADDR_REG, srcw));
1550 
1551 	FAIL_IF(push_inst(compiler, JR | SA(RETURN_ADDR_REG), UNMOVABLE_INS));
1552 	return push_inst(compiler, NOP, UNMOVABLE_INS);
1553 }
1554 
1555 /* --------------------------------------------------------------------- */
1556 /*  Conditional instructions                                             */
1557 /* --------------------------------------------------------------------- */
1558 
sljit_emit_label(struct sljit_compiler * compiler)1559 SLJIT_API_FUNC_ATTRIBUTE struct sljit_label* sljit_emit_label(struct sljit_compiler *compiler)
1560 {
1561 	struct sljit_label *label;
1562 
1563 	CHECK_ERROR_PTR();
1564 	CHECK_PTR(check_sljit_emit_label(compiler));
1565 
1566 	if (compiler->last_label && compiler->last_label->size == compiler->size)
1567 		return compiler->last_label;
1568 
1569 	label = (struct sljit_label*)ensure_abuf(compiler, sizeof(struct sljit_label));
1570 	PTR_FAIL_IF(!label);
1571 	set_label(label, compiler);
1572 	compiler->delay_slot = UNMOVABLE_INS;
1573 	return label;
1574 }
1575 
1576 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1577 #define JUMP_LENGTH	4
1578 #else
1579 #define JUMP_LENGTH	8
1580 #endif
1581 
1582 #define BR_Z(src) \
1583 	inst = BEQ | SA(src) | TA(0) | JUMP_LENGTH; \
1584 	flags = IS_BIT26_COND; \
1585 	delay_check = src;
1586 
1587 #define BR_NZ(src) \
1588 	inst = BNE | SA(src) | TA(0) | JUMP_LENGTH; \
1589 	flags = IS_BIT26_COND; \
1590 	delay_check = src;
1591 
1592 #define BR_T() \
1593 	inst = BC1T | JUMP_LENGTH; \
1594 	flags = IS_BIT16_COND; \
1595 	delay_check = FCSR_FCC;
1596 
1597 #define BR_F() \
1598 	inst = BC1F | JUMP_LENGTH; \
1599 	flags = IS_BIT16_COND; \
1600 	delay_check = FCSR_FCC;
1601 
sljit_emit_jump(struct sljit_compiler * compiler,sljit_s32 type)1602 SLJIT_API_FUNC_ATTRIBUTE struct sljit_jump* sljit_emit_jump(struct sljit_compiler *compiler, sljit_s32 type)
1603 {
1604 	struct sljit_jump *jump;
1605 	sljit_ins inst;
1606 	sljit_s32 flags = 0;
1607 	sljit_s32 delay_check = UNMOVABLE_INS;
1608 
1609 	CHECK_ERROR_PTR();
1610 	CHECK_PTR(check_sljit_emit_jump(compiler, type));
1611 
1612 	jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump));
1613 	PTR_FAIL_IF(!jump);
1614 	set_jump(jump, compiler, type & SLJIT_REWRITABLE_JUMP);
1615 	type &= 0xff;
1616 
1617 	switch (type) {
1618 	case SLJIT_EQUAL:
1619 		BR_NZ(EQUAL_FLAG);
1620 		break;
1621 	case SLJIT_NOT_EQUAL:
1622 		BR_Z(EQUAL_FLAG);
1623 		break;
1624 	case SLJIT_LESS:
1625 	case SLJIT_GREATER:
1626 	case SLJIT_SIG_LESS:
1627 	case SLJIT_SIG_GREATER:
1628 	case SLJIT_OVERFLOW:
1629 	case SLJIT_MUL_OVERFLOW:
1630 		BR_Z(OTHER_FLAG);
1631 		break;
1632 	case SLJIT_GREATER_EQUAL:
1633 	case SLJIT_LESS_EQUAL:
1634 	case SLJIT_SIG_GREATER_EQUAL:
1635 	case SLJIT_SIG_LESS_EQUAL:
1636 	case SLJIT_NOT_OVERFLOW:
1637 	case SLJIT_MUL_NOT_OVERFLOW:
1638 		BR_NZ(OTHER_FLAG);
1639 		break;
1640 	case SLJIT_NOT_EQUAL_F64:
1641 	case SLJIT_GREATER_EQUAL_F64:
1642 	case SLJIT_GREATER_F64:
1643 	case SLJIT_ORDERED_F64:
1644 		BR_T();
1645 		break;
1646 	case SLJIT_EQUAL_F64:
1647 	case SLJIT_LESS_F64:
1648 	case SLJIT_LESS_EQUAL_F64:
1649 	case SLJIT_UNORDERED_F64:
1650 		BR_F();
1651 		break;
1652 	default:
1653 		/* Not conditional branch. */
1654 		inst = 0;
1655 		break;
1656 	}
1657 
1658 	jump->flags |= flags;
1659 	if (compiler->delay_slot == MOVABLE_INS || (compiler->delay_slot != UNMOVABLE_INS && compiler->delay_slot != delay_check))
1660 		jump->flags |= IS_MOVABLE;
1661 
1662 	if (inst)
1663 		PTR_FAIL_IF(push_inst(compiler, inst, UNMOVABLE_INS));
1664 
1665 	PTR_FAIL_IF(emit_const(compiler, TMP_REG2, 0));
1666 	if (type <= SLJIT_JUMP) {
1667 		PTR_FAIL_IF(push_inst(compiler, JR | S(TMP_REG2), UNMOVABLE_INS));
1668 		jump->addr = compiler->size;
1669 		PTR_FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
1670 	} else {
1671 		SLJIT_ASSERT(DR(PIC_ADDR_REG) == 25 && PIC_ADDR_REG == TMP_REG2);
1672 		/* Cannot be optimized out if type is >= CALL0. */
1673 		jump->flags |= IS_JAL | (type >= SLJIT_CALL0 ? IS_CALL : 0);
1674 		PTR_FAIL_IF(push_inst(compiler, JALR | S(TMP_REG2) | DA(RETURN_ADDR_REG), UNMOVABLE_INS));
1675 		jump->addr = compiler->size;
1676 		/* A NOP if type < CALL1. */
1677 		PTR_FAIL_IF(push_inst(compiler, ADDU_W | S(SLJIT_R0) | TA(0) | DA(4), UNMOVABLE_INS));
1678 	}
1679 	return jump;
1680 }
1681 
1682 #define RESOLVE_IMM1() \
1683 	if (src1 & SLJIT_IMM) { \
1684 		if (src1w) { \
1685 			PTR_FAIL_IF(load_immediate(compiler, DR(TMP_REG1), src1w)); \
1686 			src1 = TMP_REG1; \
1687 		} \
1688 		else \
1689 			src1 = 0; \
1690 	}
1691 
1692 #define RESOLVE_IMM2() \
1693 	if (src2 & SLJIT_IMM) { \
1694 		if (src2w) { \
1695 			PTR_FAIL_IF(load_immediate(compiler, DR(TMP_REG2), src2w)); \
1696 			src2 = TMP_REG2; \
1697 		} \
1698 		else \
1699 			src2 = 0; \
1700 	}
1701 
sljit_emit_cmp(struct sljit_compiler * compiler,sljit_s32 type,sljit_s32 src1,sljit_sw src1w,sljit_s32 src2,sljit_sw src2w)1702 SLJIT_API_FUNC_ATTRIBUTE struct sljit_jump* sljit_emit_cmp(struct sljit_compiler *compiler, sljit_s32 type,
1703 	sljit_s32 src1, sljit_sw src1w,
1704 	sljit_s32 src2, sljit_sw src2w)
1705 {
1706 	struct sljit_jump *jump;
1707 	sljit_s32 flags;
1708 	sljit_ins inst;
1709 
1710 	CHECK_ERROR_PTR();
1711 	CHECK_PTR(check_sljit_emit_cmp(compiler, type, src1, src1w, src2, src2w));
1712 	ADJUST_LOCAL_OFFSET(src1, src1w);
1713 	ADJUST_LOCAL_OFFSET(src2, src2w);
1714 
1715 	compiler->cache_arg = 0;
1716 	compiler->cache_argw = 0;
1717 	flags = ((type & SLJIT_I32_OP) ? INT_DATA : WORD_DATA) | LOAD_DATA;
1718 	if (src1 & SLJIT_MEM) {
1719 		PTR_FAIL_IF(emit_op_mem2(compiler, flags, DR(TMP_REG1), src1, src1w, src2, src2w));
1720 		src1 = TMP_REG1;
1721 	}
1722 	if (src2 & SLJIT_MEM) {
1723 		PTR_FAIL_IF(emit_op_mem2(compiler, flags, DR(TMP_REG2), src2, src2w, 0, 0));
1724 		src2 = TMP_REG2;
1725 	}
1726 
1727 	jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump));
1728 	PTR_FAIL_IF(!jump);
1729 	set_jump(jump, compiler, type & SLJIT_REWRITABLE_JUMP);
1730 	type &= 0xff;
1731 
1732 	if (type <= SLJIT_NOT_EQUAL) {
1733 		RESOLVE_IMM1();
1734 		RESOLVE_IMM2();
1735 		jump->flags |= IS_BIT26_COND;
1736 		if (compiler->delay_slot == MOVABLE_INS || (compiler->delay_slot != UNMOVABLE_INS && compiler->delay_slot != DR(src1) && compiler->delay_slot != DR(src2)))
1737 			jump->flags |= IS_MOVABLE;
1738 		PTR_FAIL_IF(push_inst(compiler, (type == SLJIT_EQUAL ? BNE : BEQ) | S(src1) | T(src2) | JUMP_LENGTH, UNMOVABLE_INS));
1739 	}
1740 	else if (type >= SLJIT_SIG_LESS && (((src1 & SLJIT_IMM) && (src1w == 0)) || ((src2 & SLJIT_IMM) && (src2w == 0)))) {
1741 		inst = NOP;
1742 		if ((src1 & SLJIT_IMM) && (src1w == 0)) {
1743 			RESOLVE_IMM2();
1744 			switch (type) {
1745 			case SLJIT_SIG_LESS:
1746 				inst = BLEZ;
1747 				jump->flags |= IS_BIT26_COND;
1748 				break;
1749 			case SLJIT_SIG_GREATER_EQUAL:
1750 				inst = BGTZ;
1751 				jump->flags |= IS_BIT26_COND;
1752 				break;
1753 			case SLJIT_SIG_GREATER:
1754 				inst = BGEZ;
1755 				jump->flags |= IS_BIT16_COND;
1756 				break;
1757 			case SLJIT_SIG_LESS_EQUAL:
1758 				inst = BLTZ;
1759 				jump->flags |= IS_BIT16_COND;
1760 				break;
1761 			}
1762 			src1 = src2;
1763 		}
1764 		else {
1765 			RESOLVE_IMM1();
1766 			switch (type) {
1767 			case SLJIT_SIG_LESS:
1768 				inst = BGEZ;
1769 				jump->flags |= IS_BIT16_COND;
1770 				break;
1771 			case SLJIT_SIG_GREATER_EQUAL:
1772 				inst = BLTZ;
1773 				jump->flags |= IS_BIT16_COND;
1774 				break;
1775 			case SLJIT_SIG_GREATER:
1776 				inst = BLEZ;
1777 				jump->flags |= IS_BIT26_COND;
1778 				break;
1779 			case SLJIT_SIG_LESS_EQUAL:
1780 				inst = BGTZ;
1781 				jump->flags |= IS_BIT26_COND;
1782 				break;
1783 			}
1784 		}
1785 		PTR_FAIL_IF(push_inst(compiler, inst | S(src1) | JUMP_LENGTH, UNMOVABLE_INS));
1786 	}
1787 	else {
1788 		if (type == SLJIT_LESS || type == SLJIT_GREATER_EQUAL || type == SLJIT_SIG_LESS || type == SLJIT_SIG_GREATER_EQUAL) {
1789 			RESOLVE_IMM1();
1790 			if ((src2 & SLJIT_IMM) && src2w <= SIMM_MAX && src2w >= SIMM_MIN)
1791 				PTR_FAIL_IF(push_inst(compiler, (type <= SLJIT_LESS_EQUAL ? SLTIU : SLTI) | S(src1) | T(TMP_REG1) | IMM(src2w), DR(TMP_REG1)));
1792 			else {
1793 				RESOLVE_IMM2();
1794 				PTR_FAIL_IF(push_inst(compiler, (type <= SLJIT_LESS_EQUAL ? SLTU : SLT) | S(src1) | T(src2) | D(TMP_REG1), DR(TMP_REG1)));
1795 			}
1796 			type = (type == SLJIT_LESS || type == SLJIT_SIG_LESS) ? SLJIT_NOT_EQUAL : SLJIT_EQUAL;
1797 		}
1798 		else {
1799 			RESOLVE_IMM2();
1800 			if ((src1 & SLJIT_IMM) && src1w <= SIMM_MAX && src1w >= SIMM_MIN)
1801 				PTR_FAIL_IF(push_inst(compiler, (type <= SLJIT_LESS_EQUAL ? SLTIU : SLTI) | S(src2) | T(TMP_REG1) | IMM(src1w), DR(TMP_REG1)));
1802 			else {
1803 				RESOLVE_IMM1();
1804 				PTR_FAIL_IF(push_inst(compiler, (type <= SLJIT_LESS_EQUAL ? SLTU : SLT) | S(src2) | T(src1) | D(TMP_REG1), DR(TMP_REG1)));
1805 			}
1806 			type = (type == SLJIT_GREATER || type == SLJIT_SIG_GREATER) ? SLJIT_NOT_EQUAL : SLJIT_EQUAL;
1807 		}
1808 
1809 		jump->flags |= IS_BIT26_COND;
1810 		PTR_FAIL_IF(push_inst(compiler, (type == SLJIT_EQUAL ? BNE : BEQ) | S(TMP_REG1) | TA(0) | JUMP_LENGTH, UNMOVABLE_INS));
1811 	}
1812 
1813 	PTR_FAIL_IF(emit_const(compiler, TMP_REG2, 0));
1814 	PTR_FAIL_IF(push_inst(compiler, JR | S(TMP_REG2), UNMOVABLE_INS));
1815 	jump->addr = compiler->size;
1816 	PTR_FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
1817 	return jump;
1818 }
1819 
1820 #undef RESOLVE_IMM1
1821 #undef RESOLVE_IMM2
1822 
1823 #undef JUMP_LENGTH
1824 #undef BR_Z
1825 #undef BR_NZ
1826 #undef BR_T
1827 #undef BR_F
1828 
1829 #undef FLOAT_DATA
1830 #undef FMT
1831 
sljit_emit_ijump(struct sljit_compiler * compiler,sljit_s32 type,sljit_s32 src,sljit_sw srcw)1832 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_ijump(struct sljit_compiler *compiler, sljit_s32 type, sljit_s32 src, sljit_sw srcw)
1833 {
1834 	sljit_s32 src_r = TMP_REG2;
1835 	struct sljit_jump *jump = NULL;
1836 
1837 	CHECK_ERROR();
1838 	CHECK(check_sljit_emit_ijump(compiler, type, src, srcw));
1839 	ADJUST_LOCAL_OFFSET(src, srcw);
1840 
1841 	if (FAST_IS_REG(src)) {
1842 		if (DR(src) != 4)
1843 			src_r = src;
1844 		else
1845 			FAIL_IF(push_inst(compiler, ADDU_W | S(src) | TA(0) | D(TMP_REG2), DR(TMP_REG2)));
1846 	}
1847 
1848 	if (type >= SLJIT_CALL0) {
1849 		SLJIT_ASSERT(DR(PIC_ADDR_REG) == 25 && PIC_ADDR_REG == TMP_REG2);
1850 		if (src & (SLJIT_IMM | SLJIT_MEM)) {
1851 			if (src & SLJIT_IMM)
1852 				FAIL_IF(load_immediate(compiler, DR(PIC_ADDR_REG), srcw));
1853 			else {
1854 				SLJIT_ASSERT(src_r == TMP_REG2 && (src & SLJIT_MEM));
1855 				FAIL_IF(emit_op(compiler, SLJIT_MOV, WORD_DATA, TMP_REG2, 0, TMP_REG1, 0, src, srcw));
1856 			}
1857 			FAIL_IF(push_inst(compiler, JALR | S(PIC_ADDR_REG) | DA(RETURN_ADDR_REG), UNMOVABLE_INS));
1858 			/* We need an extra instruction in any case. */
1859 			return push_inst(compiler, ADDU_W | S(SLJIT_R0) | TA(0) | DA(4), UNMOVABLE_INS);
1860 		}
1861 
1862 		/* Register input. */
1863 		if (type >= SLJIT_CALL1)
1864 			FAIL_IF(push_inst(compiler, ADDU_W | S(SLJIT_R0) | TA(0) | DA(4), 4));
1865 		FAIL_IF(push_inst(compiler, JALR | S(src_r) | DA(RETURN_ADDR_REG), UNMOVABLE_INS));
1866 		return push_inst(compiler, ADDU_W | S(src_r) | TA(0) | D(PIC_ADDR_REG), UNMOVABLE_INS);
1867 	}
1868 
1869 	if (src & SLJIT_IMM) {
1870 		jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump));
1871 		FAIL_IF(!jump);
1872 		set_jump(jump, compiler, JUMP_ADDR | ((type >= SLJIT_FAST_CALL) ? IS_JAL : 0));
1873 		jump->u.target = srcw;
1874 
1875 		if (compiler->delay_slot != UNMOVABLE_INS)
1876 			jump->flags |= IS_MOVABLE;
1877 
1878 		FAIL_IF(emit_const(compiler, TMP_REG2, 0));
1879 	}
1880 	else if (src & SLJIT_MEM)
1881 		FAIL_IF(emit_op(compiler, SLJIT_MOV, WORD_DATA, TMP_REG2, 0, TMP_REG1, 0, src, srcw));
1882 
1883 	FAIL_IF(push_inst(compiler, JR | S(src_r), UNMOVABLE_INS));
1884 	if (jump)
1885 		jump->addr = compiler->size;
1886 	FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
1887 	return SLJIT_SUCCESS;
1888 }
1889 
sljit_emit_op_flags(struct sljit_compiler * compiler,sljit_s32 op,sljit_s32 dst,sljit_sw dstw,sljit_s32 src,sljit_sw srcw,sljit_s32 type)1890 SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_flags(struct sljit_compiler *compiler, sljit_s32 op,
1891 	sljit_s32 dst, sljit_sw dstw,
1892 	sljit_s32 src, sljit_sw srcw,
1893 	sljit_s32 type)
1894 {
1895 	sljit_s32 sugg_dst_ar, dst_ar;
1896 	sljit_s32 flags = GET_ALL_FLAGS(op);
1897 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1898 #	define mem_type WORD_DATA
1899 #else
1900 	sljit_s32 mem_type = (op & SLJIT_I32_OP) ? (INT_DATA | SIGNED_DATA) : WORD_DATA;
1901 #endif
1902 
1903 	CHECK_ERROR();
1904 	CHECK(check_sljit_emit_op_flags(compiler, op, dst, dstw, src, srcw, type));
1905 	ADJUST_LOCAL_OFFSET(dst, dstw);
1906 
1907 	if (dst == SLJIT_UNUSED)
1908 		return SLJIT_SUCCESS;
1909 
1910 	op = GET_OPCODE(op);
1911 #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
1912 	if (op == SLJIT_MOV_S32 || op == SLJIT_MOV_U32)
1913 		mem_type = INT_DATA | SIGNED_DATA;
1914 #endif
1915 	sugg_dst_ar = DR((op < SLJIT_ADD && FAST_IS_REG(dst)) ? dst : TMP_REG2);
1916 
1917 	compiler->cache_arg = 0;
1918 	compiler->cache_argw = 0;
1919 	if (op >= SLJIT_ADD && (src & SLJIT_MEM)) {
1920 		ADJUST_LOCAL_OFFSET(src, srcw);
1921 		FAIL_IF(emit_op_mem2(compiler, mem_type | LOAD_DATA, DR(TMP_REG1), src, srcw, dst, dstw));
1922 		src = TMP_REG1;
1923 		srcw = 0;
1924 	}
1925 
1926 	switch (type & 0xff) {
1927 	case SLJIT_EQUAL:
1928 	case SLJIT_NOT_EQUAL:
1929 		FAIL_IF(push_inst(compiler, SLTIU | SA(EQUAL_FLAG) | TA(sugg_dst_ar) | IMM(1), sugg_dst_ar));
1930 		dst_ar = sugg_dst_ar;
1931 		break;
1932 	case SLJIT_MUL_OVERFLOW:
1933 	case SLJIT_MUL_NOT_OVERFLOW:
1934 		FAIL_IF(push_inst(compiler, SLTIU | SA(OTHER_FLAG) | TA(sugg_dst_ar) | IMM(1), sugg_dst_ar));
1935 		dst_ar = sugg_dst_ar;
1936 		type ^= 0x1; /* Flip type bit for the XORI below. */
1937 		break;
1938 	case SLJIT_GREATER_F64:
1939 	case SLJIT_LESS_EQUAL_F64:
1940 		type ^= 0x1; /* Flip type bit for the XORI below. */
1941 	case SLJIT_EQUAL_F64:
1942 	case SLJIT_NOT_EQUAL_F64:
1943 	case SLJIT_LESS_F64:
1944 	case SLJIT_GREATER_EQUAL_F64:
1945 	case SLJIT_UNORDERED_F64:
1946 	case SLJIT_ORDERED_F64:
1947 		FAIL_IF(push_inst(compiler, CFC1 | TA(sugg_dst_ar) | DA(FCSR_REG), sugg_dst_ar));
1948 		FAIL_IF(push_inst(compiler, SRL | TA(sugg_dst_ar) | DA(sugg_dst_ar) | SH_IMM(23), sugg_dst_ar));
1949 		FAIL_IF(push_inst(compiler, ANDI | SA(sugg_dst_ar) | TA(sugg_dst_ar) | IMM(1), sugg_dst_ar));
1950 		dst_ar = sugg_dst_ar;
1951 		break;
1952 
1953 	default:
1954 		dst_ar = OTHER_FLAG;
1955 		break;
1956 	}
1957 
1958 	if (type & 0x1) {
1959 		FAIL_IF(push_inst(compiler, XORI | SA(dst_ar) | TA(sugg_dst_ar) | IMM(1), sugg_dst_ar));
1960 		dst_ar = sugg_dst_ar;
1961 	}
1962 
1963 	if (op >= SLJIT_ADD) {
1964 		if (DR(TMP_REG2) != dst_ar)
1965 			FAIL_IF(push_inst(compiler, ADDU_W | SA(dst_ar) | TA(0) | D(TMP_REG2), DR(TMP_REG2)));
1966 		return emit_op(compiler, op | flags, mem_type | CUMULATIVE_OP | LOGICAL_OP | IMM_OP | ALT_KEEP_CACHE, dst, dstw, src, srcw, TMP_REG2, 0);
1967 	}
1968 
1969 	if (dst & SLJIT_MEM)
1970 		return emit_op_mem(compiler, mem_type, dst_ar, dst, dstw);
1971 
1972 	if (sugg_dst_ar != dst_ar)
1973 		return push_inst(compiler, ADDU_W | SA(dst_ar) | TA(0) | DA(sugg_dst_ar), sugg_dst_ar);
1974 	return SLJIT_SUCCESS;
1975 
1976 #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
1977 #	undef mem_type
1978 #endif
1979 }
1980 
sljit_emit_const(struct sljit_compiler * compiler,sljit_s32 dst,sljit_sw dstw,sljit_sw init_value)1981 SLJIT_API_FUNC_ATTRIBUTE struct sljit_const* sljit_emit_const(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw, sljit_sw init_value)
1982 {
1983 	struct sljit_const *const_;
1984 	sljit_s32 reg;
1985 
1986 	CHECK_ERROR_PTR();
1987 	CHECK_PTR(check_sljit_emit_const(compiler, dst, dstw, init_value));
1988 	ADJUST_LOCAL_OFFSET(dst, dstw);
1989 
1990 	const_ = (struct sljit_const*)ensure_abuf(compiler, sizeof(struct sljit_const));
1991 	PTR_FAIL_IF(!const_);
1992 	set_const(const_, compiler);
1993 
1994 	reg = SLOW_IS_REG(dst) ? dst : TMP_REG2;
1995 
1996 	PTR_FAIL_IF(emit_const(compiler, reg, init_value));
1997 
1998 	if (dst & SLJIT_MEM)
1999 		PTR_FAIL_IF(emit_op(compiler, SLJIT_MOV, WORD_DATA, dst, dstw, TMP_REG1, 0, TMP_REG2, 0));
2000 	return const_;
2001 }
2002