/qemu/target/riscv/ |
H A D | vector_helper.c | 690 RVVCALL(OPIVV2, vadd_vv_w, OP_SSS_W, H4, H4, H4, DO_ADD) 694 RVVCALL(OPIVV2, vsub_vv_w, OP_SSS_W, H4, H4, H4, DO_SUB) 1043 RVVCALL(OPIVV2, vand_vv_w, OP_SSS_W, H4, H4, H4, DO_AND) 1047 RVVCALL(OPIVV2, vor_vv_w, OP_SSS_W, H4, H4, H4, DO_OR) 1051 RVVCALL(OPIVV2, vxor_vv_w, OP_SSS_W, H4, H4, H4, DO_XOR) 1357 RVVCALL(OPIVV2, vminu_vv_w, OP_UUU_W, H4, H4, H4, DO_MIN) 1361 RVVCALL(OPIVV2, vmin_vv_w, OP_SSS_W, H4, H4, H4, DO_MIN) 1369 RVVCALL(OPIVV2, vmax_vv_w, OP_SSS_W, H4, H4, H4, DO_MAX) 1425 RVVCALL(OPIVV2, vmul_vv_w, OP_SSS_W, H4, H4, H4, DO_MUL) 1594 RVVCALL(OPIVV2, vdiv_vv_w, OP_SSS_W, H4, H4, H4, DO_DIV) [all …]
|
H A D | vcrypto_helper.c | 67 RVVCALL(OPIVV2, vror_vv_w, OP_UUU_W, H4, H4, H4, ror32) in RVVCALL() 76 RVVCALL(OPIVX2, vror_vx_w, OP_UUU_W, H4, H4, ror32) in RVVCALL() 85 RVVCALL(OPIVV2, vrol_vv_w, OP_UUU_W, H4, H4, H4, rol32) in RVVCALL() 94 RVVCALL(OPIVX2, vrol_vx_w, OP_UUU_W, H4, H4, rol32) in RVVCALL() 135 RVVCALL(OPIVV2, vandn_vv_w, OP_UUU_W, H4, H4, H4, DO_ANDN) in RVVCALL() 433 res[0] = sig1_sha256(vs1[H4(2)]) + vs2[H4(1)] + sig0_sha256(vd[H4(1)]) + in vsha2ms_e32() 435 res[1] = sig1_sha256(vs1[H4(3)]) + vs2[H4(2)] + sig0_sha256(vd[H4(2)]) + in vsha2ms_e32() 438 sig1_sha256(res[0]) + vs2[H4(3)] + sig0_sha256(vd[H4(3)]) + vd[H4(2)]; in vsha2ms_e32() 440 sig1_sha256(res[1]) + vs1[H4(0)] + sig0_sha256(vs2[H4(0)]) + vd[H4(3)]; in vsha2ms_e32() 546 uint32_t a = vs2[H4(3)], b = vs2[H4(2)], e = vs2[H4(1)], f = vs2[H4(0)]; in vsha2c_32() [all …]
|
H A D | vector_internals.h | 48 #define H4(x) ((x) ^ 1) macro 55 #define H4(x) (x) macro
|
/qemu/target/arm/tcg/ |
H A D | vec_helper.c | 1042 d[H4(i)] = float32_muladd(e2, e1, a[H4(i)], 0, fpst); in HELPER() 1043 d[H4(i + 1)] = float32_muladd(e4, e3, a[H4(i + 1)], 0, fpst); in HELPER() 1076 d[H4(j)] = float32_muladd(e2, e1, a[H4(j)], 0, fpst); in HELPER() 1077 d[H4(j + 1)] = float32_muladd(e4, e3, a[H4(j + 1)], 0, fpst); in HELPER() 2059 d[H4(i)] = float32_muladd(n_1, m_1, d[H4(i)], 0, fpst); in do_fmlal() 2124 d[H4(i)] = float32_muladd(n_1, m_1, d[H4(i)], 0, fpst); in do_fmlal_idx() 2780 d[H4(0)] = sum0; in do_mmla_b() 2781 d[H4(1)] = sum1; in do_mmla_b() 2782 d[H4(2)] = sum2; in do_mmla_b() 3025 d[H4(i)] = float32_muladd(nn, mm, a[H4(i)], 0, stat); in HELPER() [all …]
|
H A D | vec_internal.h | 37 #define H4(x) ((x) ^ 1) macro 43 #define H4(x) (x) macro
|
H A D | mve_helper.c | 783 mergemask(&d[H4(e)], val, mask); in HELPER() 1108 r += n[H4(e)]; in DO_2OP_S() 1109 r += m[H4(e)] ^ inv; in DO_2OP_S() 1113 mergemask(&d[H4(e)], r, mask); in DO_2OP_S() 1844 mul = (LTYPE)n[H4(e - 1 * XCHG)] * m[H4(e)]; \ 1849 mul = (LTYPE)n[H4(e + 1 * XCHG)] * m[H4(e)]; \ 2384 rdm = d[H4(e)]; 2386 mergemask(&d[H4(e)], r, mask); 2394 rdm = d[H4(e)] >> (32 - shift); 2396 mergemask(&d[H4(e)], r, mask); [all …]
|
H A D | sme_helper.c | 831 zda[tile_vslice_index(row) + H4(col)] += zn[H4(col)]; in DO_ST() 871 uint32_t zn_row = zn[H4(row)]; in HELPER() 876 zda[tile_vslice_index(row) + H4(col)] += zn_row; in HELPER() 1158 uint32_t n = zn[H4(row)]; in do_imopa_s() 1162 uint32_t *a = &za_row[H4(col)]; in do_imopa_s() 1164 *a = fn(n, zm[H4(col)], *a, pa & pb, neg); in do_imopa_s()
|
H A D | sve_helper.c | 3026 DO_TB(s, uint32_t, H4) 3456 d[H4(j)] = n[H4(i)]; 3461 d[H4(j)] = 0; 7405 float32 n00 = n[H4(0)], n01 = n[H4(1)]; in HELPER() 7406 float32 n10 = n[H4(2)], n11 = n[H4(3)]; in HELPER() 7407 float32 m00 = m[H4(0)], m01 = m[H4(1)]; in HELPER() 7408 float32 m10 = m[H4(2)], m11 = m[H4(3)]; in HELPER() 7414 d[H4(0)] = float32_add(a[H4(0)], float32_add(p0, p1, status), status); in HELPER() 7419 d[H4(1)] = float32_add(a[H4(1)], float32_add(p0, p1, status), status); in HELPER() 7424 d[H4(2)] = float32_add(a[H4(2)], float32_add(p0, p1, status), status); in HELPER() [all …]
|
/qemu/target/s390x/tcg/ |
H A D | vec.h | 44 #define H4(x) ((x) ^ 1) macro 48 #define H4(x) (x) macro 66 return v->word[H4(enr)]; in s390_vec_read_element32() 110 v->word[H4(enr)] = data; in s390_vec_write_element32()
|