/qemu/tcg/loongarch64/ |
H A D | tcg-insn-defs.c.inc | 1269 tcg_debug_assert(vk >= 0x20 && vk <= 0x3f); 1279 tcg_debug_assert(vk >= 0x20 && vk <= 0x3f); 2482 /* Emits the `vseq.b vd, vj, vk` instruction. */ 2489 /* Emits the `vseq.h vd, vj, vk` instruction. */ 2496 /* Emits the `vseq.w vd, vj, vk` instruction. */ 2503 /* Emits the `vseq.d vd, vj, vk` instruction. */ 2510 /* Emits the `vsle.b vd, vj, vk` instruction. */ 2517 /* Emits the `vsle.h vd, vj, vk` instruction. */ 2524 /* Emits the `vsle.w vd, vj, vk` instruction. */ 2531 /* Emits the `vsle.d vd, vj, vk` instruction. */ [all …]
|
H A D | tcg-target.c.inc | 1902 * vandn vd, vj, vk: vd = vk & ~vj 1903 * andc_vec vd, vj, vk: vd = vj & ~vk 1904 * vk and vk are swapped 2048 /* vbitsel vd, vj, vk, va = bitsel_vec vd, va, vk, vj */
|
/qemu/target/loongarch/tcg/ |
H A D | vec_helper.c | 42 VReg *Vk = (VReg *)vk; in HELPER() 60 VReg *Vk = (VReg *)vk; in HELPER() 78 VReg *Vk = (VReg *)vk; in HELPER() 96 VReg *Vk = (VReg *)vk; in HELPER() 140 VReg *Vk = (VReg *)vk; in HELPER() 158 VReg *Vk = (VReg *)vk; in HELPER() 176 VReg *Vk = (VReg *)vk; in HELPER() 194 VReg *Vk = (VReg *)vk; in HELPER() 212 VReg *Vk = (VReg *)vk; in HELPER() 230 VReg *Vk = (VReg *)vk; in HELPER() [all …]
|
/qemu/tests/tcg/i386/ |
H A D | test-aes.c | 25 __m128i vk = _mm_loadu_si128((const __m128i_u *)k); in test_SB_SR_MC_AK() local 27 vi = _mm_aesenc_si128(vi, vk); in test_SB_SR_MC_AK() 62 __m128i vk = _mm_loadu_si128((const __m128i_u *)k); in test_ISB_ISR_IMC_AK() local 64 vi = _mm_aesdec_si128(vi, vk); in test_ISB_ISR_IMC_AK()
|
/qemu/target/loongarch/ |
H A D | insns.decode | 509 &vvv vd vj vk 511 &vvvv vd vj vk va 512 &vvv_fcond vd vj vk fcond 526 @vvv .... ........ ..... vk:5 vj:5 vd:5 &vvv 536 @vvvv .... ........ va:5 vk:5 vj:5 vd:5 &vvvv 537 @vvv_fcond .... ........ fcond:5 vk:5 vj:5 vd:5 &vvv_fcond
|
H A D | disas.c | 825 output(ctx, mnemonic, "v%d, v%d, v%d", a->vd, a->vj, a->vk); in output_vvv() 840 output(ctx, mnemonic, "v%d, v%d, v%d, v%d", a->vd, a->vj, a->vk, a->va); in output_vvvv() 1520 a->vj, a->vk); \ in INSN_LSX() 1730 output(ctx, mnemonic, "x%d, x%d, x%d, x%d", a->vd, a->vj, a->vk, a->va); in output_vvvv_x() 1735 output(ctx, mnemonic, "x%d, x%d, x%d", a->vd, a->vj, a->vk); in output_vvv_x() 2434 a->vj, a->vk); \ in INSN_LASX()
|
/qemu/target/loongarch/tcg/insn_trans/ |
H A D | trans_vec.c.inc | 31 vec_full_offset(a->vk), 59 vec_full_offset(a->vk), 85 vec_full_offset(a->vk), 112 vec_full_offset(a->vk), 234 uint32_t vk_ofs = vec_full_offset(a->vk); 373 get_vreg64(bh, a->vk, 1 + i * 2); 374 get_vreg64(bl, a->vk, i * 2); 3616 vk_ofs = vec_full_offset(a->vk); 4527 vk_ofs = vec_full_offset(a->vk); 4662 TCGv_i32 vk = tcg_constant_i32(a->vk); [all …]
|
/qemu/target/arm/tcg/ |
H A D | sve_helper.c | 7087 void HELPER(sve2_eor3)(void *vd, void *vn, void *vm, void *vk, uint32_t desc) in DO_ST1_ZPZ_S() 7090 uint64_t *d = vd, *n = vn, *m = vm, *k = vk; in DO_ST1_ZPZ_S() 7097 void HELPER(sve2_bcax)(void *vd, void *vn, void *vm, void *vk, uint32_t desc) in HELPER() 7100 uint64_t *d = vd, *n = vn, *m = vm, *k = vk; in HELPER() 7107 void HELPER(sve2_bsl1n)(void *vd, void *vn, void *vm, void *vk, uint32_t desc) in HELPER() 7110 uint64_t *d = vd, *n = vn, *m = vm, *k = vk; in HELPER() 7117 void HELPER(sve2_bsl2n)(void *vd, void *vn, void *vm, void *vk, uint32_t desc) in HELPER() 7120 uint64_t *d = vd, *n = vn, *m = vm, *k = vk; in HELPER() 7127 void HELPER(sve2_nbsl)(void *vd, void *vn, void *vm, void *vk, uint32_t desc) in HELPER() 7130 uint64_t *d = vd, *n = vn, *m = vm, *k = vk; in HELPER()
|