/qemu/tcg/loongarch64/ |
H A D | tcg-target.c.inc | 1165 * If address is aligned to 16-bytes, the 128-bit load/store is atomic.
|
/qemu/target/s390x/tcg/ |
H A D | translate.c | 235 const uint8_t bytes = 1 << es; in vec_reg_offset() local 236 int offs = enr * bytes; in vec_reg_offset() 262 offs ^= (8 - bytes); in vec_reg_offset()
|
/qemu/ |
H A D | qemu-options.hx | 154 interleave. Default 256 (bytes). Only 256, 512, 1k, 2k, 186 " kvm-shadow-mem=size of KVM shadow MMU in bytes\n" 522 belongs. size is the size of memory side cache in bytes. level is 527 is the write policy. line is the cache Line size in bytes. 536 policy is write-back, the cache Line size is 8 bytes: 1441 caches in bytes (default: the sum of l2-cache-size and 1452 The maximum size of the refcount block cache in bytes 2771 Handle 0x0E00, DMI type 11, 5 bytes 5469 filename. At most len bytes (64k by default) per packet are 5608 key should be 32-bytes long and be base64 encoded. The iv [all …]
|
H A D | hmp-commands.hx | 883 Read and print up to *size* bytes from ring buffer character 1087 "length: the memory size, in bytes.", 1118 the memory size, in bytes. It's optional, and should be specified
|
/qemu/block/ |
H A D | nbd.c | 1253 .len = bytes, in nbd_client_co_preadv() 1256 assert(bytes <= NBD_MAX_BUFFER_SIZE); in nbd_client_co_preadv() 1258 if (!bytes) { in nbd_client_co_preadv() 1272 if (offset + bytes > s->info.size) { in nbd_client_co_preadv() 1309 .len = bytes, in nbd_client_co_pwritev() 1320 if (!bytes) { in nbd_client_co_pwritev() 1334 .len = bytes, in nbd_client_co_pwrite_zeroes() 1357 if (!bytes) { in nbd_client_co_pwrite_zeroes() 1385 .len = bytes, in nbd_client_co_pdiscard() 1416 *pnum = bytes; in nbd_client_co_block_status() [all …]
|
H A D | nfs.c | 264 int64_t bytes, QEMUIOVector *iov, in nfs_co_preadv() argument 275 offset, bytes, nfs_co_generic_cb, &task) != 0) { in nfs_co_preadv() 298 int64_t bytes, QEMUIOVector *iov, in nfs_co_pwritev() argument 309 buf = g_try_malloc(bytes); in nfs_co_pwritev() 310 if (bytes && buf == NULL) { in nfs_co_pwritev() 313 qemu_iovec_to_buf(iov, 0, buf, bytes); in nfs_co_pwritev() 321 offset, bytes, buf, in nfs_co_pwritev() 339 if (task.ret != bytes) { in nfs_co_pwritev()
|
H A D | gluster.c | 1004 int64_t bytes, in qemu_gluster_co_pwrite_zeroes() argument 1011 acb.size = bytes; in qemu_gluster_co_pwrite_zeroes() 1016 ret = glfs_zerofill_async(s->fd, offset, bytes, gluster_finish_aiocb, &acb); in qemu_gluster_co_pwrite_zeroes() 1296 int64_t offset, int64_t bytes) in qemu_gluster_co_pdiscard() argument 1302 assert(bytes <= SIZE_MAX); /* rely on max_pdiscard */ in qemu_gluster_co_pdiscard() 1309 ret = glfs_discard_async(s->fd, offset, bytes, gluster_finish_aiocb, &acb); in qemu_gluster_co_pdiscard() 1471 int64_t bytes, in qemu_gluster_co_block_status() argument 1480 assert(QEMU_IS_ALIGNED(offset | bytes, bs->bl.request_alignment)); in qemu_gluster_co_block_status() 1487 *pnum = bytes; in qemu_gluster_co_block_status() 1496 *pnum = bytes; in qemu_gluster_co_block_status() [all …]
|
/qemu/target/i386/tcg/ |
H A D | decode-new.c.inc | 2357 case 0x01: /* Implied 0f leading opcode bytes. */ 2360 case 0x02: /* Implied 0f 38 leading opcode bytes. */ 2363 case 0x03: /* Implied 0f 3a leading opcode bytes. */
|
/qemu/qapi/ |
H A D | migration.json | 19 # @remaining: amount of bytes remaining to be transferred to the 28 # @normal-bytes: number of normal bytes sent (since 1.2) 44 # @multifd-bytes: The number of bytes sent through multifd (since 3.0) 49 # @precopy-bytes: The number of bytes sent in the pre-copy phase 52 # @downtime-bytes: The number of bytes sent while the guest is paused 55 # @postcopy-bytes: The number of bytes sent during the post-copy phase 73 'precopy-bytes': 'uint64', 'downtime-bytes': 'uint64', 74 'postcopy-bytes': 'uint64', 84 # @bytes: amount of bytes already transferred to the target VM 92 # @encoding-rate: rate of encoded bytes (since 5.1) [all …]
|
/qemu/migration/ |
H A D | trace-events | 226 qemu_rdma_fill(size_t control_len, size_t size) "RDMA %zd of %zd bytes already in buffer" 232 qemu_rdma_register_and_get_keys(uint64_t len, void *start) "Registering %" PRIu64 " bytes @ %p" 244 … long remote, uint32_t len) "Posting chunk: %" PRIu64 ", addr: 0x%lx remote: 0x%lx, bytes %" PRIu32 247 …x, int64_t offset) "Sending registration request chunk %" PRIu64 " for %d bytes, index: %d, offset… 249 …t64_t offset) "Entire chunk is zero, sending compress: %" PRIu64 " for %d bytes, index: %d, offset… 253 …t64_t length, int index, int64_t offset) "Zapping zero chunk: %" PRId64 " bytes, index %d, offset …
|
H A D | ram.c | 459 void ram_transferred_add(uint64_t bytes) in ram_transferred_add() argument 462 stat64_add(&mig_stats.precopy_bytes, bytes); in ram_transferred_add() 464 stat64_add(&mig_stats.postcopy_bytes, bytes); in ram_transferred_add() 466 stat64_add(&mig_stats.downtime_bytes, bytes); in ram_transferred_add() 680 xbzrle_counters.bytes += TARGET_PAGE_SIZE; in save_xbzrle_page() 695 xbzrle_counters.bytes += bytes_xbzrle - 8; in save_xbzrle_page() 974 encoded_size = xbzrle_counters.bytes - rs->xbzrle_bytes_prev; in migration_update_rates() 981 rs->xbzrle_bytes_prev = xbzrle_counters.bytes; in migration_update_rates()
|
H A D | migration.c | 1165 info->xbzrle_cache->bytes = xbzrle_counters.bytes; in populate_ram_info() 3060 uint64_t bytes = migration_transferred_bytes(); in migration_completion_end() local 3074 s->mbps = ((double) bytes * 8.0) / transfer_time / 1000; in migration_completion_end()
|
H A D | migration-hmp-cmds.c | 150 info->xbzrle_cache->bytes >> 10); in hmp_info_migrate()
|
/qemu/system/ |
H A D | memory.c | 1166 size_t bytes = 0; in memory_region_escape_name() local 1169 bytes += memory_region_need_escape(*p) ? 4 : 1; in memory_region_escape_name() 1171 if (bytes == p - name) { in memory_region_escape_name() 1172 return g_memdup(name, bytes + 1); in memory_region_escape_name() 1175 escaped = g_malloc(bytes + 1); in memory_region_escape_name()
|
/qemu/hw/remote/ |
H A D | vfio-user-obj.c | 270 size_t bytes = count; in vfu_object_cfg_access() local 281 while (bytes > 0) { in vfu_object_cfg_access() 282 len = (bytes > pci_access_width) ? pci_access_width : bytes; in vfu_object_cfg_access() 297 bytes -= len; in vfu_object_cfg_access()
|
/qemu/tcg/i386/ |
H A D | tcg-target.c.inc | 1348 * ??? While INC is 2 bytes shorter than ADDL $1, they also induce 1889 This is 6 + 8 = 14 bytes, as compared to using an 1890 immediate load 10 + 6 = 16 bytes, plus we may 3937 * Unpack v1 bytes to words, 0 | x. 3938 * Unpack v2 bytes to words, y | 0. 3940 * Shift logical right by 8 bits to clear the high 8 bytes before
|
/qemu/target/hexagon/ |
H A D | attribs_def.h.inc | 44 DEF_ATTRIB(MEMSIZE_2B, "Memory width is 2 bytes", "", "") 45 DEF_ATTRIB(MEMSIZE_4B, "Memory width is 4 bytes", "", "") 46 DEF_ATTRIB(MEMSIZE_8B, "Memory width is 8 bytes", "", "") 50 DEF_ATTRIB(REGWRSIZE_2B, "Memory width is 2 bytes", "", "") 51 DEF_ATTRIB(REGWRSIZE_4B, "Memory width is 4 bytes", "", "") 52 DEF_ATTRIB(REGWRSIZE_8B, "Memory width is 8 bytes", "", "")
|
/qemu/include/qemu/ |
H A D | coroutine.h | 295 size_t bytes, bool do_send); 296 #define qemu_co_recvv(sockfd, iov, iov_cnt, offset, bytes) \ argument 297 qemu_co_sendv_recvv(sockfd, iov, iov_cnt, offset, bytes, false) 298 #define qemu_co_sendv(sockfd, iov, iov_cnt, offset, bytes) \ argument 299 qemu_co_sendv_recvv(sockfd, iov, iov_cnt, offset, bytes, true) 304 ssize_t coroutine_fn qemu_co_send_recv(int sockfd, void *buf, size_t bytes, 306 #define qemu_co_recv(sockfd, buf, bytes) \ argument 307 qemu_co_send_recv(sockfd, buf, bytes, false) 308 #define qemu_co_send(sockfd, buf, bytes) \ argument 309 qemu_co_send_recv(sockfd, buf, bytes, true)
|
/qemu/tests/qtest/ |
H A D | ide-test.c | 555 static void string_cpu_to_be16(uint16_t *s, size_t bytes) in string_cpu_to_be16() argument 557 g_assert((bytes & 1) == 0); in string_cpu_to_be16() 558 bytes /= 2; in string_cpu_to_be16() 560 while (bytes--) { in string_cpu_to_be16()
|
/qemu/ui/ |
H A D | cocoa.m | 1886 [text length], [text bytes], true);
|
/qemu/target/s390x/kvm/ |
H A D | kvm.c | 2061 int32_t bytes; in kvm_s390_vcpu_interrupt_pre_save() local 2067 bytes = kvm_vcpu_ioctl(cs, KVM_S390_GET_IRQ_STATE, &irq_state); in kvm_s390_vcpu_interrupt_pre_save() 2068 if (bytes < 0) { in kvm_s390_vcpu_interrupt_pre_save() 2074 cpu->irqstate_saved_size = bytes; in kvm_s390_vcpu_interrupt_pre_save()
|
/qemu/target/arm/ |
H A D | helper.c | 12395 uint32_t HELPER(crc32)(uint32_t acc, uint32_t val, uint32_t bytes) in HELPER() 12402 return crc32(acc ^ 0xffffffff, buf, bytes) ^ 0xffffffff; in HELPER() 12405 uint32_t HELPER(crc32c)(uint32_t acc, uint32_t val, uint32_t bytes) in HELPER() 12412 return crc32c(acc, buf, bytes) ^ 0xffffffff; in HELPER()
|
/qemu/qga/ |
H A D | qapi-schema.json | 259 # @buf-b64: base64-encoded bytes read 293 # @count: number of bytes written (note: count is actual bytes 312 # @count: bytes to write (actual bytes, after base64-decode), default 674 # @rx-bytes: total bytes received 682 # @tx-bytes: total bytes transmitted 693 'data': {'rx-bytes': 'uint64', 697 'tx-bytes': 'uint64', 1027 # @used-bytes: file system used bytes (since 3.0) 1029 # @total-bytes: filesystem capacity in bytes for unprivileged users (since 3.0) 1031 # @total-bytes-privileged: filesystem capacity in bytes for privileged users [all …]
|
/qemu/host/include/x86_64/host/ |
H A D | load-extract-al16-al8.h.inc | 17 * @s: object size in bytes, @s <= 8. 19 * Load @s bytes from @pv, when pv % s != 0. If [p, p+s-1] does not
|
/qemu/host/include/loongarch64/host/ |
H A D | load-extract-al16-al8.h.inc | 17 * @s: object size in bytes, @s <= 8. 19 * Load @s bytes from @pv, when pv % s != 0. If [p, p+s-1] does not
|