/qemu/target/mips/tcg/ |
H A D | mxu_translate.c | 770 tcg_gen_shli_tl(t1, t1, 8); in gen_mxu_s8ldd() 778 tcg_gen_shli_tl(t1, t1, 16); in gen_mxu_s8ldd() 784 tcg_gen_deposit_tl(t1, t1, t1, 8, 8); in gen_mxu_s8ldd() 1344 tcg_gen_add_tl(t1, t1, t0); in gen_mxu_s16mad() 1347 tcg_gen_sub_tl(t1, t1, t0); in gen_mxu_s16mad() 1405 tcg_gen_mul_tl(t1, t1, t5); in gen_mxu_q8mul_mac() 1478 tcg_gen_mul_tl(t1, t1, t5); in gen_mxu_q8madl() 1487 tcg_gen_sub_tl(t1, t7, t1); in gen_mxu_q8madl() 1490 tcg_gen_add_tl(t1, t7, t1); in gen_mxu_q8madl() 1604 tcg_gen_shli_tl(t1, t1, strd2); in gen_mxu_s32ldxvx() [all …]
|
H A D | translate.c | 2017 tcg_gen_shli_tl(t1, t1, 3); in gen_lxl() 2022 tcg_gen_andc_tl(t1, reg, t1); in gen_lxl() 2044 tcg_gen_shli_tl(t1, t1, 3); in gen_lxr() 2050 tcg_gen_and_tl(t1, reg, t1); in gen_lxr() 3093 tcg_gen_movcond_tl(TCG_COND_EQ, t1, t1, t2, t3, t1); in gen_r6_muldiv() 3104 tcg_gen_movcond_tl(TCG_COND_EQ, t1, t1, t2, t3, t1); in gen_r6_muldiv() 3182 tcg_gen_movcond_tl(TCG_COND_EQ, t1, t1, t2, t3, t1); in gen_r6_muldiv() 3190 tcg_gen_movcond_tl(TCG_COND_EQ, t1, t1, t2, t3, t1); in gen_r6_muldiv() 3257 tcg_gen_movcond_tl(TCG_COND_EQ, t1, t1, t2, t3, t1); in gen_div1_tx79() 3313 tcg_gen_movcond_tl(TCG_COND_EQ, t1, t1, t2, t3, t1); in gen_muldiv() [all …]
|
H A D | octeon_translate.c | 46 TCGv t0, t1; in trans_BADDU() local 54 t1 = tcg_temp_new(); in trans_BADDU() 56 gen_load_gpr(t1, a->rt); in trans_BADDU() 58 tcg_gen_add_tl(t0, t0, t1); in trans_BADDU() 65 TCGv t0, t1; in trans_DMUL() local 73 t1 = tcg_temp_new(); in trans_DMUL() 75 gen_load_gpr(t1, a->rt); in trans_DMUL() 77 tcg_gen_mul_i64(cpu_gpr[a->rd], t0, t1); in trans_DMUL() 134 TCGv t0, t1; in trans_SEQNE() local 142 t1 = tcg_temp_new(); in trans_SEQNE() [all …]
|
H A D | translate_addr_const.c | 19 TCGv t1; in gen_lsa() local 26 t1 = tcg_temp_new(); in gen_lsa() 28 gen_load_gpr(t1, rt); in gen_lsa() 30 tcg_gen_add_tl(cpu_gpr[rd], t0, t1); in gen_lsa() 38 TCGv t1; in gen_dlsa() local 47 t1 = tcg_temp_new(); in gen_dlsa() 49 gen_load_gpr(t1, rt); in gen_dlsa() 51 tcg_gen_add_tl(cpu_gpr[rd], t0, t1); in gen_dlsa()
|
/qemu/target/ppc/translate/ |
H A D | fp-impl.c.inc | 191 TCGv_i64 t1; 310 TCGv_i64 t1; 343 TCGv_i64 t1; 363 TCGv_i64 t1; 384 TCGv_i64 t1; 421 TCGv_i64 t1; 441 TCGv_i64 t1; 461 TCGv_i64 t1; 483 TCGv_i64 t1; 619 tcg_gen_andi_i64(t1, t1, FP_RN); [all …]
|
H A D | vmx-impl.c.inc | 776 tcg_gen_and_vec(vece, t1, t1, t2); 784 tcg_gen_shri_vec(vece, t1, t1, 1); 981 tcg_gen_shl_i64(t1, t1, t0); 982 tcg_gen_shli_i64(t1, t1, 1); 992 tcg_gen_shl_i64(t1, t1, t0); 993 tcg_gen_shli_i64(t1, t1, 1); 1047 tcg_gen_or_i64(t1, ah, t1); 1056 tcg_gen_and_i64(t1, t1, al); 1341 tcg_gen_xor_i64(t1, t0, t1); 1343 tcg_gen_or_i64(t1, t1, t2); [all …]
|
/qemu/docs/devel/ |
H A D | tcg-ops.rst | 62 add_i32 t0, t1, t2 /* (t0 <- t1 + t2) */ 207 add_i32 t0, t1, t2 266 - | *t0* = *t1* + *t2* 270 - | *t0* = *t1* - *t2* 320 - | *t0* = ~\ *t1* 328 - | *t0* = ~(*t1* ^ *t2*), or equivalently, *t0* = *t1* ^ ~\ *t2* 344 - | *t0* = *t1* ? clz(*t1*) : *t2* 348 - | *t0* = *t1* ? ctz(*t1*) : *t2* 396 - | *t0* = *t1* 724 ld_vec *v0*, *t1* [all …]
|
/qemu/include/exec/ |
H A D | helper-gen.h.inc | 21 #define DEF_HELPER_FLAGS_1(name, flags, ret, t1) \ 24 dh_arg_decl(t1, 1)) \ 28 dh_arg(t1, 1)); \ 31 #define DEF_HELPER_FLAGS_2(name, flags, ret, t1, t2) \ 34 dh_arg_decl(t1, 1), dh_arg_decl(t2, 2)) \ 38 dh_arg(t1, 1), dh_arg(t2, 2)); \ 41 #define DEF_HELPER_FLAGS_3(name, flags, ret, t1, t2, t3) \ 44 dh_arg_decl(t1, 1), dh_arg_decl(t2, 2), dh_arg_decl(t3, 3)) \ 48 dh_arg(t1, 1), dh_arg(t2, 2), dh_arg(t3, 3)); \ 51 #define DEF_HELPER_FLAGS_4(name, flags, ret, t1, t2, t3, t4) \ [all …]
|
H A D | helper-proto.h.inc | 26 #define DEF_HELPER_FLAGS_1(name, flags, ret, t1) \ 27 dh_ctype(ret) HELPER(name) (dh_ctype(t1)) DEF_HELPER_ATTR; 29 #define DEF_HELPER_FLAGS_2(name, flags, ret, t1, t2) \ 30 dh_ctype(ret) HELPER(name) (dh_ctype(t1), dh_ctype(t2)) DEF_HELPER_ATTR; 32 #define DEF_HELPER_FLAGS_3(name, flags, ret, t1, t2, t3) \ 33 dh_ctype(ret) HELPER(name) (dh_ctype(t1), dh_ctype(t2), \ 36 #define DEF_HELPER_FLAGS_4(name, flags, ret, t1, t2, t3, t4) \ 37 dh_ctype(ret) HELPER(name) (dh_ctype(t1), dh_ctype(t2), dh_ctype(t3), \ 40 #define DEF_HELPER_FLAGS_5(name, flags, ret, t1, t2, t3, t4, t5) \ 41 dh_ctype(ret) HELPER(name) (dh_ctype(t1), dh_ctype(t2), dh_ctype(t3), \ [all …]
|
H A D | helper-head.h.inc | 118 #define DEF_HELPER_1(name, ret, t1) \ 119 DEF_HELPER_FLAGS_1(name, 0, ret, t1) 120 #define DEF_HELPER_2(name, ret, t1, t2) \ 121 DEF_HELPER_FLAGS_2(name, 0, ret, t1, t2) 122 #define DEF_HELPER_3(name, ret, t1, t2, t3) \ 123 DEF_HELPER_FLAGS_3(name, 0, ret, t1, t2, t3) 124 #define DEF_HELPER_4(name, ret, t1, t2, t3, t4) \ 125 DEF_HELPER_FLAGS_4(name, 0, ret, t1, t2, t3, t4) 126 #define DEF_HELPER_5(name, ret, t1, t2, t3, t4, t5) \ 127 DEF_HELPER_FLAGS_5(name, 0, ret, t1, t2, t3, t4, t5) [all …]
|
/qemu/crypto/ |
H A D | aes.c | 1451 u32 s0, s1, s2, s3, t0, t1, t2, t3; in AES_encrypt() local 1550 t1 = in AES_encrypt() 1576 AES_Te1[(t1 >> 16) & 0xff] ^ in AES_encrypt() 1581 AES_Te0[(t1 >> 24) ] ^ in AES_encrypt() 1590 AES_Te3[(t1 ) & 0xff] ^ in AES_encrypt() 1595 AES_Te2[(t1 >> 8) & 0xff] ^ in AES_encrypt() 1741 t1 = in AES_decrypt() 1769 AES_Td3[(t1 ) & 0xff] ^ in AES_decrypt() 1772 AES_Td0[(t1 >> 24) ] ^ in AES_decrypt() 1779 AES_Td1[(t1 >> 16) & 0xff] ^ in AES_decrypt() [all …]
|
/qemu/tcg/ |
H A D | tcg-op.c | 711 tcg_gen_clz_i64(t1, t1, t2); in tcg_gen_clz_i32() 735 tcg_gen_ctz_i64(t1, t1, t2); in tcg_gen_ctz_i32() 818 tcg_gen_shr_i32(t1, arg1, t1); in tcg_gen_rotl_i32() 856 tcg_gen_shl_i32(t1, arg1, t1); in tcg_gen_rotr_i32() 912 tcg_gen_shli_i32(t1, t1, ofs); in tcg_gen_deposit_i32() 1330 tcg_gen_sari_i32(t1, t1, 16); /* t1 = ssb. */ in tcg_gen_bswap16_i32() 1333 tcg_gen_shli_i32(t1, t1, 8); /* t1 = ..b. */ in tcg_gen_bswap16_i32() 1364 tcg_gen_shli_i32(t1, t1, 8); /* t1 = b.d. */ in tcg_gen_bswap32_i32() 2556 tcg_gen_shr_i64(t1, arg1, t1); in tcg_gen_rotl_i64() 2593 tcg_gen_shl_i64(t1, arg1, t1); in tcg_gen_rotr_i64() [all …]
|
H A D | tcg-op-gvec.c | 715 fni(t1, t0); in expand_2_i32() 735 fni(t1, t0, c); in expand_2i_i32() 779 fni(t2, t0, t1); in expand_3_i32() 874 fni(t1, t0); in expand_2_i64() 894 fni(t1, t0, c); in expand_2i_i64() 938 fni(t2, t0, t1); in expand_3_i64() 1883 tcg_gen_add_i32(t1, t1, b); in tcg_gen_vec_add16_i32() 1897 tcg_gen_add_i64(t1, t1, b); in tcg_gen_vec_add32_i64() 2066 tcg_gen_sub_i32(t1, a, t1); in tcg_gen_vec_sub16_i32() 2080 tcg_gen_sub_i64(t1, a, t1); in tcg_gen_vec_sub32_i64() [all …]
|
/qemu/target/riscv/insn_trans/ |
H A D | trans_rvk.c.inc | 158 tcg_gen_xor_i32(t1, t1, t2); 160 tcg_gen_xor_i32(t1, t1, t2); 206 tcg_gen_xor_i64(t1, t1, t2); 208 tcg_gen_xor_i64(t1, t1, t2); 259 func(t1, t0, num1); 262 tcg_gen_xor_i64(t1, t1, t2); 264 tcg_gen_xor_i64(t1, t1, t2); 298 tcg_gen_xor_i64(t1, t1, t2); 300 tcg_gen_xor_i64(t1, t1, t2); 345 tcg_gen_xor_i32(t1, t0, t1); [all …]
|
/qemu/tests/tcg/xtensa/ |
H A D | test_fp0_sqrt.S | 8 .macro sqrt_seq r, a, y, t1, hn, h2, t5, h 10 const.s \t1, 0 11 maddn.s \t1, \y, \y 15 maddn.s \r, \t1, \hn 16 nexp01.s \t1, \a 17 neg.s \h2, \t1 26 maddn.s \t1, \r, \r 29 maddn.s \r, \t1, \y 34 neg.s \t1, \h 36 addexp.s \t1, \y [all …]
|
/qemu/target/sh4/ |
H A D | op_helper.c | 258 t0 = float32_add(t0, t1, &env->fp_status); in helper_fadd_FT() 266 t0 = float64_add(t0, t1, &env->fp_status); in helper_fadd_DT() 276 relation = float32_compare(t0, t1, &env->fp_status); in helper_fcmp_eq_FT() 286 relation = float64_compare(t0, t1, &env->fp_status); in helper_fcmp_eq_DT() 296 relation = float32_compare(t0, t1, &env->fp_status); in helper_fcmp_gt_FT() 332 t0 = float32_div(t0, t1, &env->fp_status); in helper_fdiv_FT() 340 t0 = float64_div(t0, t1, &env->fp_status); in helper_fdiv_DT() 374 t0 = float32_mul(t0, t1, &env->fp_status); in helper_fmul_FT() 382 t0 = float64_mul(t0, t1, &env->fp_status); in helper_fmul_DT() 424 t0 = float32_sub(t0, t1, &env->fp_status); in helper_fsub_FT() [all …]
|
/qemu/target/loongarch/tcg/insn_trans/ |
H A D | trans_bit.c.inc | 118 TCGv t1 = tcg_temp_new(); 123 tcg_gen_shli_tl(t1, t1, 8); 124 tcg_gen_or_tl(dest, t0, t1); 131 TCGv t1 = tcg_temp_new(); 136 tcg_gen_shli_tl(t1, t1, 8); 137 tcg_gen_or_tl(dest, t0, t1); 149 tcg_gen_shli_i64(t1, t1, 16); 150 tcg_gen_or_i64(dest, t1, t0); 156 TCGv t1 = tcg_temp_new(); 160 tcg_gen_and_tl(t1, t1, mask); [all …]
|
H A D | trans_vec.c.inc | 517 tcg_gen_sari_vec(vece, t1, t1, halfbits); 688 tcg_gen_sari_vec(vece, t1, t1, halfbits); 2060 tcg_gen_sari_vec(vece, t1, t1, halfbits); 2558 TCGv_vec t1; 2567 TCGv_i32 t1; 2576 TCGv_i64 t1; 2636 TCGv_vec t1; 2721 tcg_gen_sari_vec(vece, t1, t1, halfbits); 2932 tcg_gen_mul_vec(vece, t1, t1, t2); 3081 tcg_gen_mul_vec(vece, t1, t1, t2); [all …]
|
/qemu/target/ppc/ |
H A D | translate.c | 482 tcg_gen_or_tl(t1, t1, t0); in spr_write_CTRL_ST() 1894 tcg_gen_movcond_i32(TCG_COND_EQ, t1, t1, t3, t2, t1); in gen_op_arith_modw() 1934 tcg_gen_movcond_i64(TCG_COND_EQ, t1, t1, t3, t2, t1); in gen_op_arith_modd() 3421 tcg_gen_andi_i32(t1, t1, 0x7F); in gen_stswx() 4594 tcg_gen_shli_i32(t1, t1, 2); in gen_mcrxr() 5381 tcg_gen_ext16s_tl(t1, t1); in gen_405_mulladd_insn() 5389 tcg_gen_ext16u_tl(t1, t1); in gen_405_mulladd_insn() 5400 tcg_gen_ext16s_tl(t1, t1); in gen_405_mulladd_insn() 5409 tcg_gen_ext16u_tl(t1, t1); in gen_405_mulladd_insn() 5430 tcg_gen_mul_tl(t1, t0, t1); in gen_405_mulladd_insn() [all …]
|
/qemu/target/openrisc/ |
H A D | translate.c | 310 tcg_gen_negsetcond_i64(TCG_COND_NE, t1, t1, high); in gen_muld() 345 tcg_gen_mul_i64(t1, t1, t2); in gen_mac() 350 tcg_gen_xor_i64(t1, t1, cpu_mac); in gen_mac() 351 tcg_gen_andc_i64(t1, t1, t2); in gen_mac() 369 tcg_gen_mul_i64(t1, t1, t2); in gen_macu() 373 tcg_gen_setcond_i64(TCG_COND_LTU, t1, cpu_mac, t1); in gen_macu() 386 tcg_gen_mul_i64(t1, t1, t2); in gen_msb() 391 tcg_gen_xor_i64(t1, t1, cpu_mac); in gen_msb() 392 tcg_gen_and_i64(t1, t1, t2); in gen_msb() 410 tcg_gen_mul_i64(t1, t1, t2); in gen_msbu() [all …]
|
/qemu/tests/tcg/loongarch64/system/ |
H A D | boot.S | 25 li.w t1, 0x100e001c 26 st.b t0, t1, 0 36 li.d t1, 1000000 44 addi.w t1, t1, -1 45 bnez t1, loop
|
/qemu/target/tricore/ |
H A D | op_helper.c | 637 if (t1 > t2) { in helper_absdif_ssov() 653 if (t1 > t2) { in helper_absdif_h_ssov() 661 if (t1 > t2) { in helper_absdif_h_ssov() 702 mul = t1 * t3; in helper_madd64_ssov() 707 env->PSW_USB_AV = t1 ^ t1 * 2u; in helper_madd64_ssov() 859 mul = t1 * t3; in helper_madd64_suov() 863 env->PSW_USB_AV = t1 ^ t1 * 2u; in helper_madd64_suov() 924 mul = t1 * t3; in helper_msub64_ssov() 929 env->PSW_USB_AV = t1 ^ t1 * 2u; in helper_msub64_ssov() 959 env->PSW_USB_AV = t1 ^ t1 * 2u; in helper_msub64_suov() [all …]
|
/qemu/target/avr/ |
H A D | translate.c | 215 TCGv t1 = tcg_temp_new_i32(); in gen_add_CHf() local 222 tcg_gen_or_tl(t1, t1, t2); /* t1 = t1 | t2 | t3 */ in gen_add_CHf() 223 tcg_gen_or_tl(t1, t1, t3); in gen_add_CHf() 232 TCGv t1 = tcg_temp_new_i32(); in gen_add_Vf() local 237 tcg_gen_xor_tl(t1, Rd, R); in gen_add_Vf() 239 tcg_gen_andc_tl(t1, t1, t2); in gen_add_Vf() 246 TCGv t1 = tcg_temp_new_i32(); in gen_sub_CHf() local 263 TCGv t1 = tcg_temp_new_i32(); in gen_sub_Vf() local 268 tcg_gen_xor_tl(t1, Rd, R); in gen_sub_Vf() 270 tcg_gen_and_tl(t1, t1, t2); in gen_sub_Vf() [all …]
|
/qemu/common-user/host/riscv/ |
H A D | safe-syscall.inc.S | 42 mv t1, a1 /* syscall number */ 49 mv a7, t1 61 lw t1, 0(t0) 62 bnez t1, 2f
|
/qemu/common-user/host/loongarch64/ |
H A D | safe-syscall.inc.S | 45 move $t1, $a1 /* syscall number */ 52 move $a7, $t1 73 ld.w $t1, $a6, 0 74 bnez $t1, 2f
|