/qemu/tests/tcg/cris/bare/ |
H A D | check_neg.s | 8 neg.d r4,r3 14 neg.d r3,r3 20 neg.d r3,r3 25 neg.d r3,r3 30 neg.d r3,r3 35 neg.d r3,r3 41 neg.w r4,r3 47 neg.w r3,r3 52 neg.w r3,r3 57 neg.w r3,r3 [all …]
|
/qemu/tests/qemu-iotests/ |
H A D | 083.out | 87 === Check disconnect before neg-classic === 91 === Check disconnect 8 neg-classic === 95 === Check disconnect 16 neg-classic === 99 === Check disconnect 24 neg-classic === 103 === Check disconnect 28 neg-classic === 107 === Check disconnect after neg-classic === 200 === Check disconnect 8 neg-classic === 204 === Check disconnect 16 neg-classic === 208 === Check disconnect 24 neg-classic === 212 === Check disconnect 28 neg-classic === [all …]
|
/qemu/accel/tcg/ |
H A D | tcg-accel-ops-icount.c | 114 g_assert(cpu->neg.icount_decr.u16.low == 0); in icount_prepare_for_run() 121 cpu->neg.icount_decr.u16.low = insns_left; in icount_prepare_for_run() 141 cpu->neg.icount_decr.u16.low = 0; in icount_process_data() 156 !cpu->neg.can_do_io in icount_handle_interrupt()
|
H A D | cputlb.c | 334 qemu_spin_init(&cpu->neg.tlb.c.lock); in tlb_init() 337 cpu->neg.tlb.c.dirty = 0; in tlb_init() 340 tlb_mmu_init(&cpu->neg.tlb.d[i], &cpu->neg.tlb.f[i], now); in tlb_init() 387 qemu_spin_lock(&cpu->neg.tlb.c.lock); in tlb_flush_by_mmuidx_async_work() 389 all_dirty = cpu->neg.tlb.c.dirty; in tlb_flush_by_mmuidx_async_work() 392 cpu->neg.tlb.c.dirty = all_dirty; in tlb_flush_by_mmuidx_async_work() 548 qemu_spin_lock(&cpu->neg.tlb.c.lock); in tlb_flush_page_by_mmuidx_async_0() 738 qemu_spin_lock(&cpu->neg.tlb.c.lock); in tlb_flush_range_by_mmuidx_async_0() 938 qemu_spin_lock(&cpu->neg.tlb.c.lock); in tlb_reset_dirty() 1041 CPUTLB *tlb = &cpu->neg.tlb; in tlb_set_page_full() [all …]
|
H A D | cpu-exec.c | 77 cpu_icount = cpu->icount_extra + cpu->neg.icount_decr.u16.low; in align_clocks() 128 = cpu->icount_extra + cpu->neg.icount_decr.u16.low; in init_delay_params() 419 cpu->neg.can_do_io = true; in HELPER() 462 cpu->neg.can_do_io = true; in cpu_tb_exec() 724 && cpu->neg.icount_decr.u16.low + cpu->icount_extra == 0) { in cpu_handle_exception() 792 return cpu->neg.icount_decr.u16.low + cpu->icount_extra == 0; in icount_exit_request() 812 qatomic_set_mb(&cpu->neg.icount_decr.u16.high, 0); in cpu_handle_interrupt() 942 cpu->neg.icount_decr.u16.low = insns_left; in cpu_loop_exec_tb()
|
H A D | monitor.c | 135 full += qatomic_read(&cpu->neg.tlb.c.full_flush_count); in tlb_flush_counts() 136 part += qatomic_read(&cpu->neg.tlb.c.part_flush_count); in tlb_flush_counts() 137 elide += qatomic_read(&cpu->neg.tlb.c.elide_flush_count); in tlb_flush_counts()
|
H A D | translator.c | 24 QEMU_BUILD_BUG_ON(sizeof_field(CPUState, neg.can_do_io) != 1); in set_can_do_io() 26 offsetof(ArchCPU, parent_obj.neg.can_do_io) - in set_can_do_io() 50 offsetof(ArchCPU, parent_obj.neg.icount_decr.u32) in gen_tb_start() 80 offsetof(ArchCPU, parent_obj.neg.icount_decr.u16.low) in gen_tb_start()
|
H A D | tcg-accel-ops-rr.c | 195 cpu->neg.can_do_io = true; in rr_cpu_thread_fn() 337 cpu->neg.can_do_io = 1; in rr_start_vcpu_thread()
|
H A D | cpu-exec-common.c | 38 cpu->neg.can_do_io = true; in cpu_loop_exit()
|
H A D | internal-common.h | 40 return !!cpu->neg.plugin_mem_cbs; in cpu_plugin_mem_cbs_enabled()
|
H A D | icount-common.c | 75 (cpu->neg.icount_decr.u16.low + cpu->icount_extra)); in icount_get_executed() 111 if (!cpu->neg.can_do_io) { in icount_get_raw_locked()
|
H A D | tcg-accel-ops-mttcg.c | 83 cpu->neg.can_do_io = true; in mttcg_cpu_thread_fn()
|
H A D | watchpoint.c | 101 if (!cpu->neg.can_do_io) { in cpu_check_watchpoint()
|
H A D | plugin-gen.c | 93 offsetof(CPUState, neg.plugin_mem_cbs) - in gen_enable_mem_helper() 100 offsetof(CPUState, neg.plugin_mem_cbs) - in gen_disable_mem_helper()
|
/qemu/tests/tcg/xtensa/ |
H A D | test_fp0_sqrt.S | 17 neg.s \h2, \t1 28 neg.s \y, \h 34 neg.s \t1, \h
|
H A D | test_fp0_div.S | 19 neg.s \r, \an 28 neg.s \r, \an
|
/qemu/target/arm/tcg/ |
H A D | sme_helper.c | 894 uint32_t neg = simd_data(desc) << 31; in HELPER() local 936 uint64_t neg = (uint64_t)simd_data(desc) << 63; in HELPER() local 946 uint64_t n = zn[row] ^ neg; in HELPER() 968 pair ^= neg; in f16mop_adj_pair() 1009 uint32_t neg = simd_data(desc) * 0x80008000u; in HELPER() local 1029 n = f16mop_adj_pair(n, prow, neg); in HELPER() 1055 uint32_t neg = simd_data(desc) * 0x80008000u; in HELPER() local 1064 n = f16mop_adj_pair(n, prow, neg); in HELPER() 1092 bool neg = simd_data(desc); in do_imopa_s() local 1114 bool neg = simd_data(desc); in do_imopa_d() local [all …]
|
/qemu/hw/core/ |
H A D | cpu-common.c | 88 qatomic_set(&cpu->neg.icount_decr.u16.high, -1); in cpu_exit() 132 qatomic_set(&cpu->neg.icount_decr.u32, 0); in cpu_common_reset_hold() 133 cpu->neg.can_do_io = true; in cpu_common_reset_hold()
|
/qemu/target/s390x/tcg/ |
H A D | fpu_helper.c | 818 static inline uint16_t dcmask(int bit, bool neg) in dcmask() argument 820 return 1 << (11 - bit - neg); in dcmask() 826 const bool neg = _TYPE##_is_neg(f1); \ 830 return dcmask(2, neg); \ 832 return dcmask(0, neg); \ 834 return dcmask(4, neg); \ 836 return dcmask(6, neg); \ 838 return dcmask(8, neg); \ 841 return dcmask(10, neg); \
|
/qemu/common-user/host/riscv/ |
H A D | safe-syscall.inc.S | 71 0: neg a0, a0
|
/qemu/common-user/host/aarch64/ |
H A D | safe-syscall.inc.S | 79 0: neg w0, w0
|
/qemu/common-user/host/x86_64/ |
H A D | safe-syscall.inc.S | 91 0: neg %eax
|
/qemu/util/ |
H A D | cutils.c | 509 bool neg; in qemu_strtoui() local 534 neg = memchr(nptr, '-', ep - nptr) != NULL; in qemu_strtoui() 535 if (neg) { in qemu_strtoui() 542 *result = neg ? -lresult : lresult; in qemu_strtoui()
|
/qemu/common-user/host/i386/ |
H A D | safe-syscall.inc.S | 103 0: neg %eax
|
/qemu/tcg/ |
H A D | optimize.c | 495 CASE_OP_32_64(neg): in do_constant_folding_2() 2102 static bool fold_setcond_zmask(OptContext *ctx, TCGOp *op, bool neg) in fold_setcond_zmask() argument 2135 return tcg_opt_gen_movi(ctx, op, op->args[0], neg ? -inv : inv); in fold_setcond_zmask() 2169 if (!inv && !neg) { in fold_setcond_zmask() 2190 } else if (neg) { in fold_setcond_zmask() 2204 static void fold_setcond_tst_pow2(OptContext *ctx, TCGOp *op, bool neg) in fold_setcond_tst_pow2() argument 2257 if (sh && sext_opc && neg && !inv) { in fold_setcond_tst_pow2() 2281 if (neg && inv) { in fold_setcond_tst_pow2() 2291 } else if (neg) { in fold_setcond_tst_pow2() 2873 CASE_OP_32_64(neg): in tcg_optimize()
|