/qemu/hw/misc/ |
H A D | ivshmem.c | 110 Peer *peers; member 201 if (vector < s->peers[dest].nb_eventfds) { in ivshmem_io_write() 370 &s->peers[posn].eventfds[i]); in ivshmem_add_eventfd() 388 n = s->peers[posn].nb_eventfds; in close_peer_eventfds() 402 g_free(s->peers[posn].eventfds); in close_peer_eventfds() 403 s->peers[posn].nb_eventfds = 0; in close_peer_eventfds() 414 s->peers = g_renew(Peer, s->peers, nb_peers); in resize_peers() 419 s->peers[i].nb_eventfds = 0; in resize_peers() 521 Peer *peer = &s->peers[posn]; in process_msg_connect() 946 if (s->peers) { in ivshmem_exit() [all …]
|
/qemu/backends/ |
H A D | cryptodev-vhost-user.c | 107 options.cc = b->conf.peers.ccs[i]; in cryptodev_vhost_user_start() 158 int queues = b->conf.peers.queues; in cryptodev_vhost_user_event() 184 int queues = backend->conf.peers.queues; in cryptodev_vhost_user_init() 207 backend->conf.peers.ccs[i] = cc; in cryptodev_vhost_user_init() 241 backend->conf.peers.ccs[queue_index]; in cryptodev_vhost_user_crypto_create_session() 312 backend->conf.peers.ccs[queue_index]; in cryptodev_vhost_user_close_session() 342 int queues = backend->conf.peers.queues; in cryptodev_vhost_user_cleanup() 348 cc = backend->conf.peers.ccs[i]; in cryptodev_vhost_user_cleanup() 351 backend->conf.peers.ccs[i] = NULL; in cryptodev_vhost_user_cleanup()
|
H A D | cryptodev-vhost.c | 189 cc = b->conf.peers.ccs[i]; in cryptodev_vhost_start() 210 cc = b->conf.peers.ccs[i]; in cryptodev_vhost_start() 233 cc = b->conf.peers.ccs[i]; in cryptodev_vhost_start() 258 cc = b->conf.peers.ccs[i]; in cryptodev_vhost_stop() 282 cc = b->conf.peers.ccs[queue]; in cryptodev_vhost_virtqueue_mask() 298 cc = b->conf.peers.ccs[queue]; in cryptodev_vhost_virtqueue_pending()
|
H A D | cryptodev-builtin.c | 80 int queues = backend->conf.peers.queues; in cryptodev_builtin_init() 93 backend->conf.peers.ccs[0] = cc; in cryptodev_builtin_init() 590 int queues = backend->conf.peers.queues; in cryptodev_builtin_cleanup() 600 cc = backend->conf.peers.ccs[i]; in cryptodev_builtin_cleanup() 603 backend->conf.peers.ccs[i] = NULL; in cryptodev_builtin_cleanup()
|
H A D | cryptodev-lkcf.c | 210 int queues = backend->conf.peers.queues, i; in cryptodev_lkcf_init() 230 backend->conf.peers.ccs[0] = cc; in cryptodev_lkcf_init() 255 int queues = backend->conf.peers.queues; in cryptodev_lkcf_cleanup() 294 cc = backend->conf.peers.ccs[i]; in cryptodev_lkcf_cleanup() 297 backend->conf.peers.ccs[i] = NULL; in cryptodev_lkcf_cleanup()
|
H A D | cryptodev.c | 83 for (i = 0; i < backend->conf.peers.queues; i++) { in qmp_query_cryptodev_foreach() 84 CryptoDevBackendClient *cc = backend->conf.peers.ccs[i]; in qmp_query_cryptodev_foreach() 297 uint32_t value = backend->conf.peers.queues; in cryptodev_backend_get_queues() 317 backend->conf.peers.queues = value; in cryptodev_backend_set_queues()
|
/qemu/docs/specs/ |
H A D | ivshmem-spec.rst | 14 In the latter case, the device can additionally interrupt its peers, and 15 get interrupted by its peers. 41 - If you additionally need the capability for peers to interrupt each 111 capable to tell guest software what peers are connected, or how many 130 When configured for interrupts, the peers share eventfd objects in 209 for receiving interrupts from peers using vector 0,..,N-1, in 234 descriptors for interrupting peers and getting interrupted by peers
|
/qemu/hw/net/rocker/ |
H A D | rocker_fp.c | 227 NICPeers *peers) in fp_port_alloc() argument 242 port->conf.peers = *peers; in fp_port_alloc()
|
H A D | rocker_fp.h | 50 NICPeers *peers);
|
/qemu/hw/net/ |
H A D | igb.c | 330 s->core.max_queue_num = s->conf.peers.queues ? s->conf.peers.queues - 1 : 0; in igb_init_net_peer() 338 for (i = 0; i < s->conf.peers.queues; i++) { in igb_init_net_peer() 349 for (i = 0; i < s->conf.peers.queues; i++) { in igb_init_net_peer()
|
H A D | e1000e.c | 325 s->core.max_queue_num = s->conf.peers.queues ? s->conf.peers.queues - 1 : 0; in e1000e_init_net_peer() 341 for (i = 0; i < s->conf.peers.queues; i++) { in e1000e_init_net_peer() 352 for (i = 0; i < s->conf.peers.queues; i++) { in e1000e_init_net_peer()
|
H A D | virtio-net.c | 671 NetClientState *peer = n->nic_conf.peers.ncs[0]; in virtio_net_max_tx_queue_size() 3665 n->max_ncs = MAX(n->nic_conf.peers.queues, 1); in virtio_net_device_realize() 3671 if (n->nic_conf.peers.queues) { in virtio_net_device_realize() 3673 if (n->nic_conf.peers.ncs[i]->is_datapath) { in virtio_net_device_realize()
|
/qemu/hw/core/ |
H A D | qdev-properties-system.c | 397 NetClientState *peers[MAX_QUEUE_NUM]; in set_netdev() local 405 queues = qemu_find_net_clients_except(str, peers, in set_netdev() 420 if (peers[i]->peer) { in set_netdev() 433 if (peers[i]->info->check_peer_type) { in set_netdev() 434 if (!peers[i]->info->check_peer_type(peers[i], obj->class, errp)) { in set_netdev() 439 ncs[i] = peers[i]; in set_netdev()
|
/qemu/docs/interop/ |
H A D | dbus.rst | 37 security between peers once the connection is established. Peers may 41 The daemon can control which peers can send/recv messages using
|
/qemu/include/net/ |
H A D | net.h | 34 NICPeers peers; member 40 DEFINE_PROP_NETDEV("netdev", _state, _conf.peers)
|
/qemu/hw/virtio/ |
H A D | virtio-net-pci.c | 57 vpci_dev->nvectors = 2 * MAX(net->nic_conf.peers.queues, 1) in virtio_net_pci_realize()
|
H A D | virtio-crypto.c | 1068 vcrypto->max_queues = MAX(vcrypto->cryptodev->conf.peers.queues, 1); in virtio_crypto_device_realize() 1175 CryptoDevBackendClient *cc = b->conf.peers.ccs[0]; in virtio_crypto_vhost_status() 1259 cc = b->conf.peers.ccs[0]; in virtio_crypto_get_vhost()
|
/qemu/net/ |
H A D | net.c | 331 NetClientState **peers = conf->peers.ncs; in qemu_new_nic() local 333 int i, queues = MAX(1, conf->peers.queues); in qemu_new_nic() 345 qemu_net_client_setup(&nic->ncs[i], info, peers[i], model, name, in qemu_new_nic() 459 int i, queues = MAX(nic->conf->peers.queues, 1); in qemu_del_nic()
|
/qemu/include/sysemu/ |
H A D | cryptodev.h | 230 CryptoDevBackendPeers peers; member
|