/qemu/tests/bench/ |
H A D | benchmark-crypto-cipher.c | 175 #define ADD_TEST(mode, cipher, keysize, chunk) \ in main() argument 177 (!size || g_str_equal(size, #chunk))) \ in main() 190 #define ADD_TESTS(chunk) \ in main() argument 192 ADD_TEST(ecb, aes, 128, chunk); \ in main() 193 ADD_TEST(ecb, aes, 256, chunk); \ in main() 194 ADD_TEST(cbc, aes, 128, chunk); \ in main() 195 ADD_TEST(cbc, aes, 256, chunk); \ in main() 196 ADD_TEST(ctr, aes, 128, chunk); \ in main() 197 ADD_TEST(ctr, aes, 256, chunk); \ in main() 198 ADD_TEST(xts, aes, 128, chunk); \ in main() [all …]
|
/qemu/block/ |
H A D | dmg.c | 123 switch (s->types[chunk]) { in update_max_chunk_size() 127 compressed_size = s->lengths[chunk]; in update_max_chunk_size() 128 uncompressed_sectors = s->sectorcounts[chunk]; in update_max_chunk_size() 637 uint32_t chunk = search_chunk(s, sector_num); in dmg_read_chunk() local 639 if (chunk >= s->n_chunks) { in dmg_read_chunk() 648 ret = bdrv_co_pread(bs->file, s->offsets[chunk], s->lengths[chunk], in dmg_read_chunk() 655 s->zstream.avail_in = s->lengths[chunk]; in dmg_read_chunk() 674 ret = bdrv_co_pread(bs->file, s->offsets[chunk], s->lengths[chunk], in dmg_read_chunk() 695 ret = bdrv_co_pread(bs->file, s->offsets[chunk], s->lengths[chunk], in dmg_read_chunk() 711 ret = bdrv_co_pread(bs->file, s->offsets[chunk], s->lengths[chunk], in dmg_read_chunk() [all …]
|
H A D | nbd.c | 628 if (chunk->length < pay_len) { in nbd_parse_blockstatus_payload() 725 assert(chunk->type & (1 << 15)); in nbd_parse_error_payload() 858 NBDStructuredReplyChunk *chunk; in nbd_co_do_receive_one_chunk() local 892 chunk = &s->reply.structured; in nbd_co_do_receive_one_chunk() 900 if (chunk->length) { in nbd_co_do_receive_one_chunk() 1017 NBDStructuredReplyChunk *chunk; in nbd_reply_chunk_iter_receive() local 1043 chunk = &reply->structured; in nbd_reply_chunk_iter_receive() 1104 switch (chunk->type) { in nbd_co_receive_cmdread_reply() 1125 chunk->type, nbd_reply_type_lookup(chunk->type)); in nbd_co_receive_cmdread_reply() 1158 switch (chunk->type) { in nbd_co_receive_blockstatus_reply() [all …]
|
H A D | trace-events | 168 nbd_structured_read_compliance(const char *type) "server sent non-compliant unaligned read %s chunk" 169 nbd_extended_headers_compliance(const char *type) "server sent non-compliant %s chunk not matching …
|
/qemu/ui/ |
H A D | vdagent.c | 44 VDIChunkHeader chunk; member 163 VDIChunkHeader chunk; in vdagent_send_msg() local 175 chunk.port = VDP_CLIENT_PORT; in vdagent_send_msg() 176 chunk.size = msgsize - msgoff; in vdagent_send_msg() 177 if (chunk.size > 1024) { in vdagent_send_msg() 178 chunk.size = 1024; in vdagent_send_msg() 180 buffer_reserve(&vd->outbuf, sizeof(chunk) + chunk.size); in vdagent_send_msg() 181 buffer_append(&vd->outbuf, &chunk, sizeof(chunk)); in vdagent_send_msg() 183 msgoff += chunk.size; in vdagent_send_msg() 799 memset(&vd->chunk, 0, sizeof(vd->chunk)); in vdagent_reset_bufs() [all …]
|
H A D | spice-display.c | 283 cursor->chunk.data_size = size; in qemu_spice_create_cursor_update() 284 memcpy(cursor->chunk.data, c->data, size); in qemu_spice_create_cursor_update()
|
/qemu/migration/ |
H A D | rdma.c | 1286 if (!block->pmr[chunk]) { in qemu_rdma_register_and_get_keys() 1300 if (!block->pmr[chunk] && in qemu_rdma_register_and_get_keys() 1315 if (!block->pmr[chunk]) { in qemu_rdma_register_and_get_keys() 1363 uint64_t chunk = in qemu_rdma_unregister_waiting() local 1405 block->pmr[chunk] = NULL; in qemu_rdma_unregister_waiting() 1415 reg.key.chunk = chunk; in qemu_rdma_unregister_waiting() 1478 uint64_t chunk = in qemu_rdma_poll() local 2002 uint64_t chunk, chunks; in qemu_rdma_write_one() local 2111 reg.key.chunk = chunk; in qemu_rdma_write_one() 3669 uint64_t chunk; in rdma_registration_handle() local [all …]
|
H A D | trace-events | 236 qemu_rdma_signal_unregister_append(uint64_t chunk, int pos) "Appending unregister chunk %" PRIu64 "… 237 qemu_rdma_signal_unregister_already(uint64_t chunk) "Unregister chunk %" PRIu64 " already in queue" 238 qemu_rdma_unregister_waiting_inflight(uint64_t chunk) "Cannot unregister inflight chunk: %" PRIu64 239 qemu_rdma_unregister_waiting_proc(uint64_t chunk, int pos) "Processing unregister for chunk: %" PRI… 240 qemu_rdma_unregister_waiting_send(uint64_t chunk) "Sending unregister for chunk: %" PRIu64 241 qemu_rdma_unregister_waiting_complete(uint64_t chunk) "Unregister for chunk: %" PRIu64 " complete." 244 qemu_rdma_write_one_post(uint64_t chunk, long addr, long remote, uint32_t len) "Posting chunk: %" P… 246 …egres(int mykey, int theirkey, uint64_t chunk) "Received registration result: my key: 0x%x their k… 247 qemu_rdma_write_one_sendreg(uint64_t chunk, int len, int index, int64_t offset) "Sending registrati… 261 …le_unregister_loop(int count, int index, uint64_t chunk) "Unregistration request (%d): index %d, c… [all …]
|
/qemu/hw/display/ |
H A D | qxl-render.c | 220 QXLDataChunk *chunk, uint32_t group_id) in qxl_unpack_chunks() argument 227 bytes = MIN(size - offset, chunk->data_size); in qxl_unpack_chunks() 228 memcpy(dest + offset, chunk->data, bytes); in qxl_unpack_chunks() 233 chunk = qxl_phys2virt(qxl, chunk->next_chunk, group_id, in qxl_unpack_chunks() 234 sizeof(QXLDataChunk) + chunk->data_size); in qxl_unpack_chunks() 235 if (!chunk) { in qxl_unpack_chunks() 271 and_mask = cursor->chunk.data; in qxl_cursor() 280 qxl_unpack_chunks(c->data, size, qxl, &cursor->chunk, group_id); in qxl_cursor() 329 sizeof(QXLCursor) + cursor->chunk.data_size); in qxl_render_cursor()
|
/qemu/nbd/ |
H A D | client.c | 1440 if (chunk->magic == NBD_STRUCTURED_REPLY_MAGIC) { in nbd_receive_reply_chunk_header() 1441 len = sizeof(chunk->structured); in nbd_receive_reply_chunk_header() 1444 len = sizeof(chunk->extended); in nbd_receive_reply_chunk_header() 1447 ret = nbd_read(ioc, (uint8_t *)chunk + sizeof(chunk->magic), in nbd_receive_reply_chunk_header() 1455 chunk->structured.flags = be16_to_cpu(chunk->structured.flags); in nbd_receive_reply_chunk_header() 1456 chunk->structured.type = be16_to_cpu(chunk->structured.type); in nbd_receive_reply_chunk_header() 1457 chunk->structured.cookie = be64_to_cpu(chunk->structured.cookie); in nbd_receive_reply_chunk_header() 1465 if (chunk->magic == NBD_STRUCTURED_REPLY_MAGIC) { in nbd_receive_reply_chunk_header() 1470 chunk->magic = NBD_STRUCTURED_REPLY_MAGIC; in nbd_receive_reply_chunk_header() 1474 chunk->structured.type, in nbd_receive_reply_chunk_header() [all …]
|
H A D | server.c | 2099 stw_be_p(&chunk->flags, flags); in set_be_chunk() 2100 stw_be_p(&chunk->type, type); in set_be_chunk() 2109 stw_be_p(&chunk->flags, flags); in set_be_chunk() 2110 stw_be_p(&chunk->type, type); in set_be_chunk() 2140 NBDStructuredReadData chunk; in nbd_co_send_chunk_read() local 2143 {.iov_base = &chunk, .iov_len = sizeof(chunk)}, in nbd_co_send_chunk_read() 2151 stq_be_p(&chunk.offset, offset); in nbd_co_send_chunk_read() 2163 NBDStructuredError chunk; in nbd_co_send_chunk_error() local 2167 {.iov_base = &chunk, .iov_len = sizeof(chunk)}, in nbd_co_send_chunk_error() 2176 stl_be_p(&chunk.error, nbd_err); in nbd_co_send_chunk_error() [all …]
|
H A D | trace-events | 36 …16_t type, const char *name, uint64_t cookie, uint32_t length) "Got reply chunk header: { flags = … 71 …ookie = %" PRIu64 ", extents = %u, context = %d (extents cover %" PRIu64 " bytes, last chunk = %d)"
|
/qemu/tests/qemu-iotests/ |
H A D | 303 | 32 chunk = 1024 * 1024 variable 55 write_to_disk((i) * chunk, chunk)
|
H A D | 242 | 33 chunk = 256 * 1024 variable 85 write_to_disk((num-1) * chunk, chunk)
|
H A D | nbd-fault-injector.py | 83 chunk = sock.recv(bufsize - received) 84 if len(chunk) == 0: 86 chunks.append(chunk) 87 received += len(chunk)
|
/qemu/hw/audio/ |
H A D | hda-codec.c | 248 uint32_t chunk = MIN(B_SIZE - start, to_transfer); in hda_audio_input_timer() local 254 rpos += chunk; in hda_audio_input_timer() 255 to_transfer -= chunk; in hda_audio_input_timer() 256 st->rpos += chunk; in hda_audio_input_timer() 277 uint32_t chunk = (uint32_t) MIN(B_SIZE - start, to_transfer); in hda_audio_input_cb() local 282 if (chunk != read) { in hda_audio_input_cb() 318 uint32_t chunk = MIN(B_SIZE - start, to_transfer); in hda_audio_output_timer() local 324 wpos += chunk; in hda_audio_output_timer() 325 to_transfer -= chunk; in hda_audio_output_timer() 326 st->wpos += chunk; in hda_audio_output_timer() [all …]
|
/qemu/tests/tcg/multiarch/ |
H A D | sha512.c | 271 static void Transform(uint64_t *s, const uint64_t *chunk) in Transform() argument 276 Round(a, b, c, &d, e, f, g, &h, 0x428a2f98d728ae22ull, w0 = be64_to_cpu(chunk[0])); in Transform() 277 Round(h, a, b, &c, d, e, f, &g, 0x7137449123ef65cdull, w1 = be64_to_cpu(chunk[1])); in Transform() 278 Round(g, h, a, &b, c, d, e, &f, 0xb5c0fbcfec4d3b2full, w2 = be64_to_cpu(chunk[2])); in Transform() 279 Round(f, g, h, &a, b, c, d, &e, 0xe9b5dba58189dbbcull, w3 = be64_to_cpu(chunk[3])); in Transform() 280 Round(e, f, g, &h, a, b, c, &d, 0x3956c25bf348b538ull, w4 = be64_to_cpu(chunk[4])); in Transform() 281 Round(d, e, f, &g, h, a, b, &c, 0x59f111f1b605d019ull, w5 = be64_to_cpu(chunk[5])); in Transform() 282 Round(c, d, e, &f, g, h, a, &b, 0x923f82a4af194f9bull, w6 = be64_to_cpu(chunk[6])); in Transform() 283 Round(b, c, d, &e, f, g, h, &a, 0xab1c5ed5da6d8118ull, w7 = be64_to_cpu(chunk[7])); in Transform() 284 Round(a, b, c, &d, e, f, g, &h, 0xd807aa98a3030242ull, w8 = be64_to_cpu(chunk[8])); in Transform() [all …]
|
/qemu/scripts/ |
H A D | qemu-stamp.py | 16 for chunk in iter(lambda: f.read(65536), b''): 17 sha.update(chunk)
|
H A D | dump-guest-memory.py | 542 chunk = qemu_core.read_memory(cur, chunk_size) 543 vmcore.write(chunk)
|
/qemu/docs/ |
H A D | rdma.txt | 213 8. Register request (dynamic chunk registration) 272 2. During runtime, once a 'chunk' becomes full of pages ready to 274 other side to register the memory for this chunk and respond 280 using chunk registration (or not checked at all and unconditionally 281 written if chunk registration is disabled. This is accomplished using 283 then we check the entire chunk for zero. Only if the entire chunk is 378 When a chunk is full (or a flush() occurs), the memory backed by 379 the chunk is registered with librdmacm is pinned in memory on 382 for the entire chunk. 386 for the completion of *every* chunk. The current batch size [all …]
|
/qemu/audio/ |
H A D | pwaudio.c | 136 buf->datas[0].chunk->offset = 0; in playback_on_process() 137 buf->datas[0].chunk->stride = v->frame_size; in playback_on_process() 138 buf->datas[0].chunk->size = n_bytes; in playback_on_process() 170 offs = SPA_MIN(buf->datas[0].chunk->offset, buf->datas[0].maxsize); in capture_on_process() 171 n_bytes = SPA_MIN(buf->datas[0].chunk->size, buf->datas[0].maxsize - offs); in capture_on_process()
|
/qemu/ |
H A D | qemu-img.c | 1373 int64_t chunk; in img_compare() local 1535 chunk = MIN(pnum1, pnum2); in img_compare() 1551 chunk = MIN(chunk, IO_BUF_SIZE); in img_compare() 1552 ret = blk_pread(blk1, offset, chunk, buf1, 0); in img_compare() 1560 ret = blk_pread(blk2, offset, chunk, buf2, 0); in img_compare() 1569 if (ret || pnum != chunk) { in img_compare() 1577 chunk = MIN(chunk, IO_BUF_SIZE); in img_compare() 1579 ret = check_empty_sectors(blk1, offset, chunk, in img_compare() 1589 offset += chunk; in img_compare() 1618 chunk = MIN(chunk, IO_BUF_SIZE); in img_compare() [all …]
|
/qemu/hw/usb/ |
H A D | hcd-xhci.c | 1448 unsigned int chunk = 0; in xhci_xfer_create_sgl() local 1464 chunk = trb->status & 0x1ffff; in xhci_xfer_create_sgl() 1506 unsigned int chunk = 0; in xhci_xfer_report() local 1510 chunk = trb->status & 0x1ffff; in xhci_xfer_report() 1511 if (chunk > 8) { in xhci_xfer_report() 1512 chunk = 8; in xhci_xfer_report() 1518 chunk = trb->status & 0x1ffff; in xhci_xfer_report() 1519 if (chunk > left) { in xhci_xfer_report() 1520 chunk = left; in xhci_xfer_report() 1525 left -= chunk; in xhci_xfer_report() [all …]
|
/qemu/docs/interop/ |
H A D | parallels.txt | 22 Sector A 512-byte data chunk. 24 Cluster A data chunk of the size specified in the image header.
|
H A D | prl-xml.txt | 58 Sector a 512-byte data chunk.
|