/qemu/block/ |
H A D | blkio.c | 371 int iovcnt = qiov->niov; in blkio_co_preadv() 415 int iovcnt = qiov->niov; in blkio_co_pwritev()
|
H A D | blklogwrites.c | 476 size_t niov = qiov ? qiov->niov : 0; in blk_log_writes_co_log() local 504 qemu_iovec_init(&log_qiov, niov + 2); in blk_log_writes_co_log()
|
H A D | blkverify.c | 242 qemu_iovec_init(&raw_qiov, qiov->niov); in blkverify_co_preadv()
|
H A D | bochs.c | 261 qemu_iovec_init(&local_qiov, qiov->niov); in bochs_co_preadv()
|
H A D | crypto.c | 487 qemu_iovec_init(&hd_qiov, qiov->niov); in block_crypto_co_preadv() 551 qemu_iovec_init(&hd_qiov, qiov->niov); in block_crypto_co_pwritev()
|
H A D | file-posix.c | 208 int niov; member 1692 aiocb->io.niov, in handle_aiocb_rw_vector() 1696 aiocb->io.niov, in handle_aiocb_rw_vector() 1763 if (aiocb->io.niov == 1) { in handle_aiocb_rw() 1803 for (i = 0; i < aiocb->io.niov; ++i) { in handle_aiocb_rw() 1816 for (i = 0; i < aiocb->io.niov && count; ++i) { in handle_aiocb_rw() 1837 iov_memset(aiocb->io.iov, aiocb->io.niov, nbytes, in handle_aiocb_rw() 2422 for (i = 0; i < qiov->niov; i++) { in bdrv_qiov_is_aligned() 2528 .niov = qiov->niov, in raw_co_prw() 3524 for (int i = 0; i < qiov->niov; i++) { in raw_co_zone_append()
|
H A D | file-win32.c | 163 acb->aio_niov = qiov->niov; in paio_submit()
|
H A D | gluster.c | 1195 ret = glfs_pwritev_async(s->fd, qiov->iov, qiov->niov, offset, 0, in qemu_gluster_co_rw() 1198 ret = glfs_preadv_async(s->fd, qiov->iov, qiov->niov, offset, 0, in qemu_gluster_co_rw()
|
H A D | io.c | 1608 struct iovec *iov, int niov, in bdrv_create_padded_qiov() argument 1614 assert(niov <= IOV_MAX); in bdrv_create_padded_qiov() 1628 padded_niov = !!pad->head + niov + !!pad->tail; in bdrv_create_padded_qiov() 1661 niov -= collapse_count; in bdrv_create_padded_qiov() 1679 qemu_iovec_concat_iov(&pad->local_qiov, iov, niov, iov_offset, bytes); in bdrv_create_padded_qiov() 1686 assert(pad->local_qiov.niov == MIN(padded_niov, IOV_MAX)); in bdrv_create_padded_qiov()
|
H A D | io_uring.c | 96 qemu_iovec_init(resubmit_qiov, luringcb->qiov->niov); in luring_resubmit_short_read() 106 luringcb->sqeq.len = luringcb->resubmit_qiov.niov; in luring_resubmit_short_read() 346 luringcb->qiov->niov, offset); in luring_do_submit() 350 luringcb->qiov->niov, offset); in luring_do_submit() 354 luringcb->qiov->niov, offset); in luring_do_submit()
|
H A D | iscsi.c | 636 (struct scsi_iovec *)iov->iov, iov->niov); in iscsi_co_writev() 642 (struct scsi_iovec *)iov->iov, iov->niov); in iscsi_co_writev() 662 iov->niov); 868 (struct scsi_iovec *)iov->iov, iov->niov); in iscsi_co_readv() 875 (struct scsi_iovec *)iov->iov, iov->niov); in iscsi_co_readv() 895 scsi_task_set_iov_in(iTask.task, (struct scsi_iovec *) iov->iov, iov->niov);
|
H A D | linux-aio.c | 379 io_prep_pwritev(iocbs, fd, qiov->iov, qiov->niov, offset); in laio_do_submit() 382 io_prep_pwritev(iocbs, fd, qiov->iov, qiov->niov, offset); in laio_do_submit() 385 io_prep_preadv(iocbs, fd, qiov->iov, qiov->niov, offset); in laio_do_submit()
|
H A D | mirror.c | 208 for (i = 0; i < op->qiov.niov; i++) { in mirror_iteration_done() 1590 iov_to_buf_full(qiov->iov, qiov->niov, 0, bounce_buf, bytes); in bdrv_mirror_top_pwritev()
|
H A D | nbd.c | 536 if (rc >= 0 && qio_channel_writev_all(s->ioc, qiov->iov, qiov->niov, in nbd_co_send_request() 791 qemu_iovec_init(&sub_qiov, qiov->niov); in nbd_co_receive_offset_data_payload() 793 ret = qio_channel_readv_all(s->ioc, sub_qiov.iov, sub_qiov.niov, errp); in nbd_co_receive_offset_data_payload() 886 return qio_channel_readv_all(s->ioc, qiov->iov, qiov->niov, in nbd_co_do_receive_one_chunk() 1216 assert(request->len == iov_size(write_qiov->iov, write_qiov->niov)); in nbd_co_request()
|
H A D | nfs.c | 308 if (iov->niov != 1) { in nfs_co_pwritev()
|
H A D | nvme.c | 1066 for (i = 0; i < qiov->niov; ++i) { in nvme_cmd_map_qiov() 1208 trace_nvme_prw_aligned(s, is_write, offset, bytes, flags, qiov->niov); in nvme_co_prw_aligned() 1244 for (i = 0; i < qiov->niov; ++i) { in nvme_qiov_aligned() 1274 trace_nvme_prw_buffered(s, offset, bytes, qiov->niov, is_write); in nvme_co_prw()
|
H A D | parallels.c | 451 qemu_iovec_init(&hd_qiov, qiov->niov); in parallels_co_writev() 494 qemu_iovec_init(&hd_qiov, qiov->niov); in parallels_co_readv()
|
H A D | qcow.c | 639 if (qiov->niov > 1) { in qcow_co_preadv() 716 if (qiov->niov > 1) { in qcow_co_preadv() 739 if (bs->encrypted || qiov->niov > 1) { in qcow_co_pwritev()
|
H A D | qed.c | 1403 qemu_iovec_init(&acb.cur_qiov, qiov->niov); in qed_co_request()
|
H A D | quorum.c | 238 assert(dest->niov == source->niov); in quorum_copy_qiov() 240 for (i = 0; i < source->niov; i++) { in quorum_copy_qiov() 397 qiov->iov, qiov->niov, in quorum_compute_hash() 432 assert(a->niov == b->niov); in quorum_iovec_compare() 433 for (i = 0; i < a->niov; i++) { in quorum_iovec_compare() 618 qemu_iovec_init(&acb->qcrs[i].qiov, acb->qiov->niov); in read_quorum_children()
|
H A D | raw-format.c | 261 qemu_iovec_init(&local_qiov, qiov->niov + 1); in raw_co_pwritev()
|
H A D | rbd.c | 1312 r = rbd_aio_readv(s->image, qiov->iov, qiov->niov, offset, c); in qemu_rbd_start_co() 1315 r = rbd_aio_writev(s->image, qiov->iov, qiov->niov, offset, c); in qemu_rbd_start_co()
|
H A D | replication.c | 275 qemu_iovec_init(&hd_qiov, qiov->niov); in replication_co_writev()
|
H A D | trace-events | 148 … offset, uint64_t bytes, int flags, int niov) "s %p is_write %d offset 0x%"PRIx64" bytes %"PRId64"… 151 …id *s, uint64_t offset, uint64_t bytes, int niov, int is_write) "s %p offset 0x%"PRIx64" bytes %"P…
|
H A D | vdi.c | 563 qemu_iovec_init(&local_qiov, qiov->niov); in vdi_co_preadv() 623 qemu_iovec_init(&local_qiov, qiov->niov); in vdi_co_pwritev()
|