/qemu/block/ |
H A D | aio_task.c | 34 bool waiting; member 55 if (pool->waiting) { in aio_task_co() 56 pool->waiting = false; in aio_task_co() 66 pool->waiting = true; in aio_task_pool_wait_one() 69 assert(!pool->waiting); in aio_task_pool_wait_one()
|
/qemu/include/qemu/ |
H A D | rcu.h | 60 bool waiting; member 115 if (unlikely(qatomic_read(&p_rcu_reader->waiting))) { in rcu_read_unlock() 116 qatomic_set(&p_rcu_reader->waiting, false); in rcu_read_unlock()
|
/qemu/hw/scsi/ |
H A D | lsi53c895a.c | 235 int waiting; member 356 s->waiting = LSI_NOWAIT; in lsi_soft_reset() 601 if (s->waiting != 2) { in lsi_resume_script() 602 s->waiting = LSI_NOWAIT; in lsi_resume_script() 605 s->waiting = LSI_NOWAIT; in lsi_resume_script() 808 s->waiting = 0; in lsi_command_complete() 844 if (s->waiting) { in lsi_transfer_data() 1163 s->waiting = LSI_NOWAIT; in lsi_execute_script() 1284 if (s->waiting) in lsi_execute_script() 1290 if (s->waiting) in lsi_execute_script() [all …]
|
/qemu/tests/qemu-iotests/ |
H A D | 161.out | 27 …oseconds": TIMESTAMP}, "event": "JOB_STATUS_CHANGE", "data": {"status": "waiting", "id": "none0"}} 49 …oseconds": TIMESTAMP}, "event": "JOB_STATUS_CHANGE", "data": {"status": "waiting", "id": "none0"}}
|
H A D | 173.out | 36 …oseconds": TIMESTAMP}, "event": "JOB_STATUS_CHANGE", "data": {"status": "waiting", "id": "disk2"}}
|
H A D | 247.out | 15 …roseconds": TIMESTAMP}, "event": "JOB_STATUS_CHANGE", "data": {"status": "waiting", "id": "job0"}}
|
H A D | 094.out | 20 …croseconds": TIMESTAMP}, "event": "JOB_STATUS_CHANGE", "data": {"status": "waiting", "id": "src"}}
|
H A D | 095.out | 21 …roseconds": TIMESTAMP}, "event": "JOB_STATUS_CHANGE", "data": {"status": "waiting", "id": "test"}}
|
H A D | 127.out | 25 …seconds": TIMESTAMP}, "event": "JOB_STATUS_CHANGE", "data": {"status": "waiting", "id": "mirror"}}
|
H A D | 144.out | 39 …seconds": TIMESTAMP}, "event": "JOB_STATUS_CHANGE", "data": {"status": "waiting", "id": "virtio0"…
|
H A D | common.qemu | 144 echo "Timeout waiting for ${success_match} on handle ${h}"
|
H A D | 249.out | 36 …roseconds": TIMESTAMP}, "event": "JOB_STATUS_CHANGE", "data": {"status": "waiting", "id": "job0"}}
|
H A D | 219.out | 87 {"data": {"id": "job0", "status": "waiting"}, "event": "JOB_STATUS_CHANGE", "timestamp": {"microsec… 140 {"data": {"id": "job0", "status": "waiting"}, "event": "JOB_STATUS_CHANGE", "timestamp": {"microsec… 193 {"data": {"id": "job0", "status": "waiting"}, "event": "JOB_STATUS_CHANGE", "timestamp": {"microsec… 254 {"data": {"id": "job0", "status": "waiting"}, "event": "JOB_STATUS_CHANGE", "timestamp": {"microsec… 315 {"data": {"id": "job0", "status": "waiting"}, "event": "JOB_STATUS_CHANGE", "timestamp": {"microsec…
|
H A D | 312.out | 69 …seconds": TIMESTAMP}, "event": "JOB_STATUS_CHANGE", "data": {"status": "waiting", "id": "virtio0"…
|
H A D | 156.out | 43 …seconds": TIMESTAMP}, "event": "JOB_STATUS_CHANGE", "data": {"status": "waiting", "id": "source"}}
|
H A D | 191.out | 57 "status": "waiting", 464 "status": "waiting",
|
H A D | 108.out | 169 …seconds": TIMESTAMP}, "event": "JOB_STATUS_CHANGE", "data": {"status": "waiting", "id": "create"}}
|
/qemu/util/ |
H A D | rcu.c | 87 qatomic_set(&index->waiting, true); in wait_for_readers() 108 qatomic_set(&index->waiting, false); in wait_for_readers()
|
H A D | trace-events | 70 lockcnt_futex_wait(const void *lockcnt, int val) "lockcnt %p waiting on %d" 80 qemu_mutex_lock(void *mutex, const char *file, const int line) "waiting on mutex %p (%s:%d)"
|
/qemu/qapi/ |
H A D | job.json | 62 # @waiting: The job is waiting for other jobs in the transaction to 63 # converge to the waiting state. This status will likely not be 87 'waiting', 'pending', 'aborting', 'concluded', 'null' ] }
|
H A D | dump.json | 78 # @detach: if true, QMP will return immediately rather than waiting
|
H A D | transaction.json | 29 # waiting on others. This is the default.
|
/qemu/docs/specs/ |
H A D | pvpanic.rst | 8 The management app has the option of waiting for GUEST_PANICKED events,
|
/qemu/docs/devel/migration/ |
H A D | postcopy.rst | 53 the destination is waiting for). 240 thread will be halted waiting for the page to be migrated, it means it can 303 threads waiting for that lock will also be blocked.
|
/qemu/docs/devel/ |
H A D | multi-process.rst | 667 MMIOs may be waiting to be consumed by an emulation program and multiple 668 threads may be waiting for MMIO replies. The scoreboard would contain a 760 pending queue to the sent queue, and if there are threads waiting for 768 queue, and any threads waiting for the reply are woken. If a store is 771 waiting for posted stores to complete, the load is continued. 792 to determine if there are MMIO requests waiting to be read. It will 816 driver putting the QEMU thread to sleep waiting for the emulation 826 queue is full. In this case, the QEMU thread must sleep waiting for
|