/qemu/block/ |
H A D | qcow.c | 639 if (qiov->niov > 1) { in qcow_co_preadv() 716 if (qiov->niov > 1) { in qcow_co_preadv() 739 if (bs->encrypted || qiov->niov > 1) { in qcow_co_pwritev()
|
H A D | blkverify.c | 242 qemu_iovec_init(&raw_qiov, qiov->niov); in blkverify_co_preadv()
|
H A D | nvme.c | 1066 for (i = 0; i < qiov->niov; ++i) { in nvme_cmd_map_qiov() 1208 trace_nvme_prw_aligned(s, is_write, offset, bytes, flags, qiov->niov); in nvme_co_prw_aligned() 1244 for (i = 0; i < qiov->niov; ++i) { in nvme_qiov_aligned() 1274 trace_nvme_prw_buffered(s, offset, bytes, qiov->niov, is_write); in nvme_co_prw()
|
H A D | blkio.c | 371 int iovcnt = qiov->niov; in blkio_co_preadv() 415 int iovcnt = qiov->niov; in blkio_co_pwritev()
|
H A D | io.c | 1608 struct iovec *iov, int niov, in bdrv_create_padded_qiov() argument 1614 assert(niov <= IOV_MAX); in bdrv_create_padded_qiov() 1628 padded_niov = !!pad->head + niov + !!pad->tail; in bdrv_create_padded_qiov() 1661 niov -= collapse_count; in bdrv_create_padded_qiov() 1679 qemu_iovec_concat_iov(&pad->local_qiov, iov, niov, iov_offset, bytes); in bdrv_create_padded_qiov() 1686 assert(pad->local_qiov.niov == MIN(padded_niov, IOV_MAX)); in bdrv_create_padded_qiov()
|
H A D | crypto.c | 487 qemu_iovec_init(&hd_qiov, qiov->niov); in block_crypto_co_preadv() 551 qemu_iovec_init(&hd_qiov, qiov->niov); in block_crypto_co_pwritev()
|
H A D | vdi.c | 563 qemu_iovec_init(&local_qiov, qiov->niov); in vdi_co_preadv() 623 qemu_iovec_init(&local_qiov, qiov->niov); in vdi_co_pwritev()
|
H A D | vpc.c | 631 qemu_iovec_init(&local_qiov, qiov->niov); in vpc_co_preadv() 681 qemu_iovec_init(&local_qiov, qiov->niov); in vpc_co_pwritev()
|
H A D | parallels.c | 451 qemu_iovec_init(&hd_qiov, qiov->niov); in parallels_co_writev() 494 qemu_iovec_init(&hd_qiov, qiov->niov); in parallels_co_readv()
|
H A D | gluster.c | 1195 ret = glfs_pwritev_async(s->fd, qiov->iov, qiov->niov, offset, 0, in qemu_gluster_co_rw() 1198 ret = glfs_preadv_async(s->fd, qiov->iov, qiov->niov, offset, 0, in qemu_gluster_co_rw()
|
H A D | raw-format.c | 261 qemu_iovec_init(&local_qiov, qiov->niov + 1); in raw_co_pwritev()
|
H A D | replication.c | 275 qemu_iovec_init(&hd_qiov, qiov->niov); in replication_co_writev()
|
H A D | rbd.c | 1312 r = rbd_aio_readv(s->image, qiov->iov, qiov->niov, offset, c); in qemu_rbd_start_co() 1315 r = rbd_aio_writev(s->image, qiov->iov, qiov->niov, offset, c); in qemu_rbd_start_co()
|
H A D | vhdx.c | 1191 qemu_iovec_init(&hd_qiov, qiov->niov); in vhdx_co_readv() 1353 qemu_iovec_init(&hd_qiov, qiov->niov); in vhdx_co_writev()
|
H A D | mirror.c | 209 for (i = 0; i < op->qiov.niov; i++) { in mirror_iteration_done() 1595 iov_to_buf_full(qiov->iov, qiov->niov, 0, bounce_buf, bytes); in bdrv_mirror_top_pwritev()
|
H A D | file-win32.c | 163 acb->aio_niov = qiov->niov; in paio_submit()
|
H A D | nfs.c | 308 if (iov->niov != 1) { in nfs_co_pwritev()
|
H A D | vmdk.c | 1880 qemu_iovec_init(&local_qiov, qiov->niov); in vmdk_write_extent() 1989 qemu_iovec_init(&local_qiov, qiov->niov); in vmdk_co_preadv()
|
H A D | qed.c | 1403 qemu_iovec_init(&acb.cur_qiov, qiov->niov); in qed_co_request()
|
/qemu/include/hw/virtio/ |
H A D | virtio-gpu.h | 310 uint32_t *niov);
|
/qemu/hw/usb/ |
H A D | dev-serial.c | 492 for (i = 0; i < p->iov.niov; i++) { in usb_serial_handle_data()
|
/qemu/nbd/ |
H A D | server.c | 2025 unsigned niov, Error **errp) in nbd_co_send_iov() argument 2033 ret = qio_channel_writev_all(client->ioc, iov, niov, errp) < 0 ? -EIO : 0; in nbd_co_send_iov() 2084 size_t niov, uint16_t flags, uint16_t type, in set_be_chunk() argument 2089 for (i = 1; i < niov; i++) { in set_be_chunk()
|
/qemu/migration/ |
H A D | rdma.c | 2747 size_t niov, in qio_channel_rdma_writev() argument 2783 for (int i = 0; i < niov; i++) { in qio_channel_rdma_writev() 2835 size_t niov, in qio_channel_rdma_readv() argument 2862 for (int i = 0; i < niov; i++) { in qio_channel_rdma_readv()
|
/qemu/hw/display/ |
H A D | virtio-gpu.c | 810 uint32_t *niov) in virtio_gpu_create_mapping_iov() argument 880 *niov = v; in virtio_gpu_create_mapping_iov()
|
/qemu/hw/ide/ |
H A D | core.c | 432 iocb->j = iocb->qiov->niov - 1; in trim_aio_cancel() 477 while (iocb->j < iocb->qiov->niov) { in ide_issue_trim_cb()
|