/qemu/migration/ |
H A D | channel-block.c | 52 const struct iovec *iov, in qio_channel_block_readv() argument 63 qemu_iovec_init_external(&qiov, (struct iovec *)iov, niov); in qio_channel_block_readv() 77 const struct iovec *iov, in qio_channel_block_writev() argument 88 qemu_iovec_init_external(&qiov, (struct iovec *)iov, niov); in qio_channel_block_writev()
|
H A D | multifd.h | 152 struct iovec *iov; member 212 struct iovec *iov; member 248 p->iov[0].iov_len = p->packet_len; in multifd_send_prepare_header() 249 p->iov[0].iov_base = p->packet; in multifd_send_prepare_header()
|
/qemu/tests/qtest/ |
H A D | test-filter-redirector.c | 100 struct iovec iov[] = { in test_redirector_tx() local 110 ret = iov_send(backend_sock[0], iov, 2, 0, sizeof(size) + sizeof(send_buf)); in test_redirector_tx() 164 struct iovec iov[] = { in test_redirector_rx() local 179 ret = iov_send(send_sock, iov, 2, 0, sizeof(size) + sizeof(send_buf)); in test_redirector_rx()
|
/qemu/crypto/ |
H A D | hash-gnutls.c | 58 const struct iovec *iov, in qcrypto_gnutls_hash_bytesv() argument 94 gnutls_hash(hash, iov[i].iov_base, iov[i].iov_len); in qcrypto_gnutls_hash_bytesv()
|
H A D | hash-gcrypt.c | 50 const struct iovec *iov, in qcrypto_gcrypt_hash_bytesv() argument 77 gcry_md_write(md, iov[i].iov_base, iov[i].iov_len); in qcrypto_gcrypt_hash_bytesv()
|
H A D | hmac-glib.c | 86 const struct iovec *iov, in qcrypto_glib_hmac_bytesv() argument 98 g_hmac_update(ctx->ghmac, iov[i].iov_base, iov[i].iov_len); in qcrypto_glib_hmac_bytesv()
|
H A D | hmac-gnutls.c | 99 const struct iovec *iov, in qcrypto_gnutls_hmac_bytesv() argument 112 gnutls_hmac(ctx->handle, iov[i].iov_base, iov[i].iov_len); in qcrypto_gnutls_hmac_bytesv()
|
H A D | hmac-gcrypt.c | 97 const struct iovec *iov, in qcrypto_gcrypt_hmac_bytesv() argument 111 gcry_mac_write(ctx->handle, iov[i].iov_base, iov[i].iov_len); in qcrypto_gcrypt_hmac_bytesv()
|
H A D | cipher-afalg.c | 152 struct iovec iov; in qcrypto_afalg_cipher_op() local 162 afalg->msg->msg_iov = &iov; in qcrypto_afalg_cipher_op() 172 iov.iov_base = (void *)in + done; in qcrypto_afalg_cipher_op() 173 iov.iov_len = len - done; in qcrypto_afalg_cipher_op()
|
H A D | hash-nettle.c | 109 const struct iovec *iov, in qcrypto_nettle_hash_bytesv() argument 133 size_t len = iov[i].iov_len; in qcrypto_nettle_hash_bytesv() 134 uint8_t *base = iov[i].iov_base; in qcrypto_nettle_hash_bytesv()
|
/qemu/net/ |
H A D | vmnet-common.m | 64 * vmnet.framework supports iov, but writing more than 65 * one iov into vmnet interface fails with 77 struct iovec iov; 88 iov.iov_base = (char *) buf; 89 iov.iov_len = size; 94 packet.vm_pkt_iov = &iov; 236 struct iovec *iov = s->iov_buf; 240 iov[i].iov_len = s->max_packet_size; 241 iov[i].iov_base = g_malloc0(iov[i].iov_len); 242 packets[i].vm_pkt_iov = iov + i;
|
H A D | dump.c | 64 static ssize_t dump_receive_iov(DumpState *s, const struct iovec *iov, int cnt, in dump_receive_iov() argument 70 size_t size = iov_size(iov, cnt) - offset; in dump_receive_iov() 88 cnt = iov_copy(&dumpiov[1], cnt, iov, cnt, offset, caplen); in dump_receive_iov() 153 unsigned flags, const struct iovec *iov, in filter_dump_receive_iov() argument 160 dump_receive_iov(&nfds->ds, iov, iovcnt, offset); in filter_dump_receive_iov()
|
H A D | net.c | 635 const struct iovec *iov, in filter_receive_iov() argument 671 struct iovec iov = { in filter_receive() local 795 buffer = iov[0].iov_base; in nc_sendv_compat() 796 offset = iov[0].iov_len; in nc_sendv_compat() 798 offset = iov_size(iov, iovcnt); in nc_sendv_compat() 804 offset = iov_to_buf(iov, iovcnt, 0, buf, offset); in nc_sendv_compat() 829 return iov_size(iov, iovcnt); in qemu_deliver_packet_iov() 845 ret = nc->info->receive_iov(nc, iov, iovcnt); in qemu_deliver_packet_iov() 847 ret = nc_sendv_compat(nc, iov, iovcnt, flags); in qemu_deliver_packet_iov() 866 size_t size = iov_size(iov, iovcnt); in qemu_sendv_packet_async() [all …]
|
H A D | filter.c | 32 const struct iovec *iov, in qemu_netfilter_receive() argument 42 nf, sender, flags, iov, iovcnt, sent_cb); in qemu_netfilter_receive() 66 const struct iovec *iov, in qemu_netfilter_pass_to_next() argument 99 ret = qemu_netfilter_receive(next, direction, sender, flags, iov, in qemu_netfilter_pass_to_next() 114 sender, flags, iov, iovcnt, NULL); in qemu_netfilter_pass_to_next() 119 return iov_size(iov, iovcnt); in qemu_netfilter_pass_to_next()
|
H A D | hub.c | 64 const struct iovec *iov, int iovcnt) in net_hub_receive_iov() argument 67 ssize_t len = iov_size(iov, iovcnt); in net_hub_receive_iov() 74 qemu_sendv_packet(&port->nc, iov, iovcnt); in net_hub_receive_iov() 121 const struct iovec *iov, int iovcnt) in net_hub_port_receive_iov() argument 125 return net_hub_receive_iov(port->hub, port, iov, iovcnt); in net_hub_port_receive_iov()
|
/qemu/hw/usb/ |
H A D | pcap.c | 175 .length = p->iov.size, in do_usb_pcap_data() 177 int data_len = p->iov.size; in do_usb_pcap_data() 212 iov_to_buf(p->iov.iov, p->iov.niov, 0, buf, data_len); in do_usb_pcap_data()
|
H A D | dev-serial.c | 430 packet_len = p->iov.size; in usb_serial_token_in() 484 struct iovec *iov; in usb_serial_handle_data() local 492 for (i = 0; i < p->iov.niov; i++) { in usb_serial_handle_data() 493 iov = p->iov.iov + i; in usb_serial_handle_data() 498 qemu_chr_fe_write_all(&s->cs, iov->iov_base, iov->iov_len); in usb_serial_handle_data() 500 p->actual_length = p->iov.size; in usb_serial_handle_data()
|
H A D | canokey.c | 199 trace_canokey_handle_data_out(ep_out, p->iov.size); in canokey_handle_data() 200 usb_packet_copy(p, key->ep_out_buffer[ep_out], p->iov.size); in canokey_handle_data() 202 while (out_pos != p->iov.size) { in canokey_handle_data() 209 out_len = MIN(p->iov.size - out_pos, key->ep_out_size[ep_out]); in canokey_handle_data() 250 p->iov.size); in canokey_handle_data()
|
/qemu/nbd/ |
H A D | server.c | 839 struct iovec iov[] = { in nbd_negotiate_send_meta_context() local 2058 struct iovec iov[] = { in nbd_co_send_simple_reply() local 2090 length += iov[i].iov_len; in set_be_chunk() 2121 struct iovec iov[] = { in nbd_co_send_chunk_done() local 2141 struct iovec iov[] = { in nbd_co_send_chunk_read() local 2165 struct iovec iov[] = { in nbd_co_send_chunk_error() local 2219 struct iovec iov[] = { in nbd_co_send_sparse_read() local 2439 iov[1].iov_base = &meta_ext; in nbd_co_send_extents() 2445 iov[2].iov_base = ea->extents; in nbd_co_send_extents() 2450 iov[1].iov_base = &meta; in nbd_co_send_extents() [all …]
|
/qemu/hw/net/rocker/ |
H A D | rocker_tlv.h | 152 int type, const struct iovec *iov, in rocker_tlv_put_iov() argument 155 size_t len = iov_size(iov, iovcnt); in rocker_tlv_put_iov() 163 iov_to_buf(iov, iovcnt, 0, rocker_tlv_data(tlv), len); in rocker_tlv_put_iov() 170 struct iovec iov = { in rocker_tlv_put() local 175 rocker_tlv_put_iov(buf, buf_pos, type, &iov, 1); in rocker_tlv_put()
|
/qemu/hw/char/ |
H A D | sclpconsole.c | 39 uint8_t iov[SIZE_BUFFER_VT220]; member 76 memcpy(&scon->iov[scon->iov_bs], buf, size); in chr_read() 115 memcpy(buf, &cons->iov[cons->iov_sclp], cons->iov_sclp_rest); in get_console_data() 125 memcpy(buf, &cons->iov[cons->iov_sclp], avail); in get_console_data() 211 VMSTATE_UINT8_ARRAY(iov, SCLPConsole, SIZE_BUFFER_VT220),
|
/qemu/contrib/vhost-user-gpu/ |
H A D | vhost-user-gpu.c | 37 struct iovec *iov; member 147 struct iovec iov = { in vg_sock_fd_write() local 152 .msg_iov = &iov, in vg_sock_fd_write() 482 struct iovec **iov) in vg_create_mapping_iov() argument 513 g_free(*iov); in vg_create_mapping_iov() 515 *iov = NULL; in vg_create_mapping_iov() 542 if (res->iov) { in vg_resource_attach_backing() 560 g_free(iov); in vg_cleanup_mapping_iov() 568 res->iov = NULL; in vg_cleanup_mapping() 583 if (!res || !res->iov) { in vg_resource_detach_backing() [all …]
|
/qemu/io/ |
H A D | channel-socket.c | 502 const struct iovec *iov, in qio_channel_socket_readv() argument 517 msg.msg_iov = (struct iovec *)iov; in qio_channel_socket_readv() 555 const struct iovec *iov, in qio_channel_socket_writev() argument 572 msg.msg_iov = (struct iovec *)iov; in qio_channel_socket_writev() 635 const struct iovec *iov, in qio_channel_socket_readv() argument 655 iov[i].iov_base, in qio_channel_socket_readv() 656 iov[i].iov_len, in qio_channel_socket_readv() 674 if (ret < iov[i].iov_len) { in qio_channel_socket_readv() 698 iov[i].iov_base, in qio_channel_socket_writev() 699 iov[i].iov_len, in qio_channel_socket_writev() [all …]
|
/qemu/include/hw/hyperv/ |
H A D | vmbus.h | 209 const struct iovec *iov, size_t iov_cnt); 217 int vmbus_map_sgl(VMBusChanReq *req, DMADirection dir, struct iovec *iov, 223 void vmbus_unmap_sgl(VMBusChanReq *req, DMADirection dir, struct iovec *iov,
|
/qemu/hw/display/ |
H A D | virtio-gpu-udmabuf.c | 46 rb = qemu_ram_block_from_host(res->iov[i].iov_base, false, &offset); in virtio_gpu_create_udmabuf() 56 list->list[i].size = res->iov[i].iov_len; in virtio_gpu_create_udmabuf() 136 res->iov[0].iov_len < 4096) { in virtio_gpu_init_udmabuf() 137 pdata = res->iov[0].iov_base; in virtio_gpu_init_udmabuf()
|