/qemu/crypto/ |
H A D | random-platform.c | 79 if (likely(fd < 0)) { in qcrypto_random_bytes() 82 if (likely(got == buflen)) { in qcrypto_random_bytes() 98 if (likely(got == buflen)) { in qcrypto_random_bytes()
|
/qemu/include/exec/ |
H A D | memory_ldst_cached.h.inc | 32 if (likely(cache->ptr)) { 44 if (likely(cache->ptr)) { 56 if (likely(cache->ptr)) { 78 if (likely(cache->ptr)) { 89 if (likely(cache->ptr)) { 100 if (likely(cache->ptr)) {
|
/qemu/target/ppc/ |
H A D | mem_helper.c | 65 if (likely(nb <= nb_pg1)) { in probe_contiguous() 90 if (likely(host)) { in helper_lmw() 112 if (likely(host)) { in helper_stmw() 141 if (likely(host)) { in do_lsw() 202 if (likely(xer_bc != 0)) { in helper_lswx() 230 if (likely(host)) { in helper_stsw() 352 if (likely(reg != rb && (ra == 0 || reg != ra))) { in helper_lscbx() 358 if (likely(d != 0)) { in helper_lscbx()
|
H A D | internal.h | 40 if (likely(start == 0)) { \ 42 } else if (likely(end == max_bit)) { \
|
/qemu/accel/ |
H A D | accel-blocker.c | 44 if (likely(bql_locked())) { in accel_ioctl_begin() 54 if (likely(bql_locked())) { in accel_ioctl_end()
|
/qemu/target/riscv/ |
H A D | internals.h | 106 if (likely((f & mask) == mask)) { in check_nanbox_s() 132 if (likely((f & mask) == mask)) { in check_nanbox_h()
|
/qemu/target/i386/tcg/ |
H A D | access.c | 65 if (likely(offset <= ac->size1 - len)) { in access_ptr() 74 if (likely(offset >= ac->size1)) { in access_ptr() 84 # define test_ptr(p) likely(p)
|
/qemu/target/i386/tcg/sysemu/ |
H A D | excp_helper.c | 88 if (likely(in->haddr)) { in ptw_ldl() 96 if (likely(in->haddr)) { in ptw_ldq() 125 if (likely(in->haddr)) { in ptw_setl() 545 if (likely(use_stage2)) { in get_physical_address() 565 if (likely(env->cr[0] & CR0_PG_MASK)) { in get_physical_address()
|
/qemu/subprojects/libvhost-user/include/ |
H A D | compiler.h | 43 #ifndef likely 44 #define likely(x) __builtin_expect(!!(x), 1) macro
|
/qemu/subprojects/libvduse/include/ |
H A D | compiler.h | 43 #ifndef likely 44 #define likely(x) __builtin_expect(!!(x), 1) macro
|
/qemu/include/qemu/ |
H A D | compiler.h | 43 #ifndef likely 44 #define likely(x) __builtin_expect(!!(x), 1) macro
|
/qemu/target/arm/tcg/ |
H A D | mte_helper.c | 336 if (likely(cmp == old)) { in store_tag1_parallel() 439 if (likely(in_page >= 2 * TAG_GRANULE)) { in HELPER() 839 if (likely(tag_last - prev_page < TARGET_PAGE_SIZE)) { in mte_probe_int() 874 if (likely(n == tag_count)) { in mte_probe_int() 1032 if (likely(mem_tag == ptr_tag)) { in HELPER() 1079 if (likely(n == tag_count)) { in mte_mops_probe() 1137 if (likely(n == tag_count)) { in mte_mops_probe_rev()
|
/qemu/target/openrisc/ |
H A D | mmu.c | 85 if (likely(need & right)) { in get_phys_mmu() 128 if (likely(excp == 0)) { in openrisc_cpu_tlb_fill()
|
/qemu/util/ |
H A D | qht.c | 358 if (likely(!qht_map_is_stale__locked(ht, map))) { in qht_map_lock_buckets__no_stale() 392 if (likely(!qht_map_is_stale__locked(ht, map))) { in qht_bucket_lock__no_stale() 571 if (likely(p) && likely(func(p, userp))) { in qht_do_lookup() 609 if (likely(!seqlock_read_retry(&b->sequence, version))) { in qht_lookup_custom() 713 if (likely(prev == NULL)) { in qht_insert()
|
/qemu/host/include/i386/host/crypto/ |
H A D | clmul.h | 16 # define HAVE_CLMUL_ACCEL likely(cpuinfo & CPUINFO_PCLMUL)
|
/qemu/accel/tcg/ |
H A D | ldst_atomicity.c.inc | 403 if (likely((pi & 1) == 0)) { 408 if (likely(left_in_page > 8)) { 445 if (likely((pi & 3) == 0)) { 450 if (likely(left_in_page > 8)) { 494 if (HAVE_al8 && likely((pi & 7) == 0)) { 547 if (HAVE_ATOMIC128_RO && likely((pi & 15) == 0)) { 866 if (likely((pi & 1) == 0)) { 917 if (likely((pi & 3) == 0)) { 984 if (HAVE_al8 && likely((pi & 7) == 0)) { 1050 if (HAVE_ATOMIC128_RW && likely((pi & 15) == 0)) {
|
H A D | translator.c | 260 if (likely(((base ^ last) & TARGET_PAGE_MASK) == 0)) { in translator_ld() 315 if (likely(new_page1 != old_page1)) { in translator_ld()
|
/qemu/host/include/aarch64/host/crypto/ |
H A D | clmul.h | 19 # define HAVE_CLMUL_ACCEL likely(cpuinfo & CPUINFO_PMULL)
|
/qemu/host/include/loongarch64/host/ |
H A D | atomic128-ldst.h | 15 #define HAVE_ATOMIC128_RO likely(cpuinfo & CPUINFO_LSX)
|
/qemu/host/include/x86_64/host/ |
H A D | atomic128-ldst.h | 30 #define HAVE_ATOMIC128_RO likely(cpuinfo & CPUINFO_ATOMIC_VMOVDQA)
|
/qemu/target/s390x/tcg/ |
H A D | vec_helper.c | 46 if (likely(bytes >= 16)) { in HELPER() 198 if (likely(bytes >= 16)) { in HELPER()
|
/qemu/target/microblaze/ |
H A D | helper.c | 62 if (likely(hit)) { in mb_cpu_tlb_fill() 287 if (likely(iflags & ESR_ESS_FLAG)) { in mb_cpu_do_unaligned_access()
|
/qemu/fpu/ |
H A D | softfloat-parts.c.inc | 112 if (likely(frac_eqz(p))) { 124 } else if (likely(p->exp < fmt->exp_max) || fmt->arm_althp) { 129 } else if (likely(frac_eqz(p))) { 197 if (likely(exp > 0)) { 308 if (likely(p->cls == float_class_normal)) { 348 if (likely(ab_mask == float_cmask_normal)) { 381 if (likely(ab_mask == float_cmask_normal)) { 426 if (likely(ab_mask == float_cmask_normal)) { 604 if (likely(ab_mask == float_cmask_normal)) { 662 if (likely(ab_mask == float_cmask_normal)) { [all …]
|
H A D | softfloat.c | 1176 if (likely(a0)) { in frac256_normalize() 1285 if (likely(a0)) { in frac64_modrem() 1288 } else if (likely(a1)) { in frac64_modrem() 1374 if (likely(a0)) { in frac128_modrem() 1377 } else if (likely(a1)) { in frac128_modrem() 1382 } else if (likely(a2)) { in frac128_modrem() 1453 if (likely(c != 0)) { in frac64_shrjam() 1454 if (likely(c < 64)) { in frac64_shrjam() 1470 } else if (likely(c < 64)) { in frac128_shrjam() 1472 } else if (likely(c < 128)) { in frac128_shrjam() [all …]
|
H A D | softfloat-parts-addsub.c.inc | 56 if (likely(shift < N)) {
|