/qemu/include/qemu/ |
H A D | range.h | 41 assert(range->lob <= range->upb || range->lob == range->upb + 1); in range_invariant() 51 return range->lob > range->upb; in range_is_empty() 57 return val >= range->lob && val <= range->upb; in range_contains() 63 *range = range_empty; in range_make_empty() 75 range->lob = lob; in range_set_bounds() 76 range->upb = upb; in range_set_bounds() 102 return range->lob; in range_lob() 109 return range->upb; in range_upb() 123 range->lob = lob; in range_init() 135 range->lob = lob; in range_init_nofail() [all …]
|
/qemu/hw/hyperv/ |
H A D | hv-balloon-page_range_tree.c | 41 PageRange *range = g_malloc(sizeof(*range)); in page_range_tree_insert_new() local 46 range->count = count; in page_range_tree_insert_new() 58 PageRange *range; in hvb_page_range_tree_insert() local 74 assert(range); in hvb_page_range_tree_insert() 91 assert(range); in hvb_page_range_tree_insert() 102 range->count = MAX(range->count, count); in hvb_page_range_tree_insert() 115 range->start, range->count); in hvb_page_range_tree_insert() 126 range->count = MAX(range->count, count); in hvb_page_range_tree_insert() 140 PageRange *range; in hvb_page_range_tree_pop() local 148 assert(range); in hvb_page_range_tree_pop() [all …]
|
H A D | hv-balloon-page_range_tree.h | 24 uint64_t endr = range->start + range->count; in page_range_part_before() 27 out->start = range->start; in page_range_part_before() 40 uint64_t end = range->start + range->count; in page_range_part_after() 43 out->start = MAX(range->start, ends); in page_range_part_after() 55 uint64_t end1 = range->start + range->count; in page_range_intersect() 59 out->start = MAX(range->start, start); in page_range_intersect() 68 page_range_intersect(range, start, count, &trange); in page_range_intersection_size() 75 return start + count == range->start; in page_range_joinable_left() 81 return range->start + range->count == start; in page_range_joinable_right() 87 return page_range_joinable_left(range, start, count) || in page_range_joinable() [all …]
|
H A D | hv-balloon.c | 371 PageRange range; in hv_balloon_unballoon_posting() local 401 *dctr -= range.count; in hv_balloon_unballoon_posting() 410 range.count, range.start, in hv_balloon_unballoon_posting() 720 hvb_page_range_tree_insert(tree, range->start, range->count, &dupcount); in hv_balloon_remove_response_insert_range() 748 trace_hv_balloon_remove_response(range->count, range->start, both); in hv_balloon_remove_response_handle_range() 769 range->count, range->start, in hv_balloon_remove_response_handle_range() 782 range->count, range->start, in hv_balloon_remove_response_handle_range() 795 if (range->count > 0) { in hv_balloon_remove_response_handle_range() 798 trace_hv_balloon_remove_response_remainder(range->count, range->start, in hv_balloon_remove_response_handle_range() 800 range->count = 0; in hv_balloon_remove_response_handle_range() [all …]
|
H A D | hv-balloon-our_range_memslots.h | 23 PageRange range; member 37 return our_range->range.start + our_range->added; in our_range_get_remaining_start() 42 return our_range->range.count - our_range->added - our_range->unusable_tail; in our_range_get_remaining_size() 49 our_range->unusable_tail = our_range->range.count - our_range->added; in our_range_mark_remaining_unusable() 88 OurRange range; member
|
H A D | trace-events | 30 hv_balloon_our_range_add(uint64_t count, uint64_t start) "adding our range %"PRIu64" @ 0x%"PRIx64 31 …4_t count, uint64_t start, unsigned int both) "processing remove response range %"PRIu64" @ 0x%"PR… 32 …ange, uint64_t starthpr, unsigned int both) "response range hole %"PRIu64" @ 0x%"PRIx64" from rang… 33 …, uint64_t removed, unsigned int both) "response common range %"PRIu64" @ 0x%"PRIx64" from range %… 34 …64_t count, uint64_t start, unsigned int both) "remove response remaining range %"PRIu64" @ 0x%"PR…
|
H A D | hv-balloon-our_range_memslots.c | 19 our_range->range.start = start; in our_range_init() 20 our_range->range.count = count; in our_range_init() 52 our_range->range.count); in hvb_our_range_mark_added() 109 our_range_init(&our_range->range, in hvb_our_range_memslots_new() 156 our_range_destroy(&our_range->range); in hvb_our_range_memslots_free() 168 total_map_size = (our_range->range.added + additional_map_size) * in hvb_our_range_memslots_ensure_mapped_additional()
|
/qemu/tests/qemu-iotests/ |
H A D | 093 | 45 for i in range(0, self.max_drives): 56 for i in range(0, ndrives): 90 for i in range(rd_nr): 96 for i in range(wr_nr): 113 for i in range(0, ndrives): 121 for i in range(0, ndrives): 127 for i in range(0, ndrives): 312 for i in range(self.max_drives): 317 for i in range(self.max_drives): 325 for i in range(self.max_drives): [all …]
|
H A D | 044 | 57 for i in range(0, h.refcount_table_clusters): 60 for j in range(0, 64)) 65 sector = b''.join(struct.pack('>H', 1) for j in range(0, 64 * 256)) 66 for block in range(0, h.refcount_table_clusters): 73 for j in range(0, h.l1_size)) 84 for j in range(0, 1024)) 91 for j in range(0, remaining // 512))
|
H A D | 151 | 76 for offset in range(1 * self.image_len // 8, 3 * self.image_len // 8, 1024 * 1024): 78 for offset in range(2 * self.image_len // 8, 3 * self.image_len // 8, 1024 * 1024): 91 for offset in range(3 * self.image_len // 8, 5 * self.image_len // 8, 1024 * 1024): 93 for offset in range(4 * self.image_len // 8, 5 * self.image_len // 8, 1024 * 1024): 103 for offset in range(5 * self.image_len // 8, 7 * self.image_len // 8, 1024 * 1024): 105 for offset in range(6 * self.image_len // 8, 7 * self.image_len // 8, 1024 * 1024): 323 for _ in range(0, 5 * self.iops): 337 for _ in range(0, 4 * self.iops): 347 for _ in range(0, 10 * self.iops): 356 for _ in range(0, 5 * self.iops):
|
H A D | 148 | 55 for i in range(len(imgs)): 69 for i in range(len(imgs)): 92 for i in range(3): 104 for i in range(3): 112 for i in range(len(imgs)): 124 for i in range(len(imgs)):
|
H A D | 136 | 224 for i in range(rd_ops): 227 for i in range(wr_ops): 230 for i in range(flush_ops): 235 for i in range(invalid_rd_ops): 238 for i in range(invalid_wr_ops): 241 for i in range(failed_rd_ops): 244 for i in range(failed_wr_ops):
|
/qemu/tests/bench/ |
H A D | atomic_add-bench.c | 22 static unsigned int range = 1024; variable 66 index = info->r & (range - 1); in thread_func() 101 counts = qemu_memalign(64, sizeof(*counts) * range); in create_threads() 102 memset(counts, 0, sizeof(*counts) * range); in create_threads() 103 for (i = 0; i < range; i++) { in create_threads() 121 printf(" ops' range: %u\n", range); in pr_params() 130 for (i = 0; i < range; i++) { in pr_stats() 167 range = pow2ceil(atoi(optarg)); in parse_args()
|
H A D | atomic64-bench.c | 27 static unsigned int range = 1024; variable 68 index = info->r & (range - 1); in thread_func() 98 counts = g_malloc0_n(range, sizeof(*counts)); in create_threads() 115 printf(" ops' range: %u\n", range); in pr_params() 155 range = pow2ceil(atoi(optarg)); in parse_args()
|
/qemu/hw/pci/ |
H A D | pci-qmp-cmds.c | 72 PciMemoryRange *range; in qmp_query_pci_bridge() local 81 range = info->bus->io_range = g_new0(PciMemoryRange, 1); in qmp_query_pci_bridge() 82 range->base = pci_bridge_get_base(dev, PCI_BASE_ADDRESS_SPACE_IO); in qmp_query_pci_bridge() 83 range->limit = pci_bridge_get_limit(dev, PCI_BASE_ADDRESS_SPACE_IO); in qmp_query_pci_bridge() 85 range = info->bus->memory_range = g_new0(PciMemoryRange, 1); in qmp_query_pci_bridge() 86 range->base = pci_bridge_get_base(dev, PCI_BASE_ADDRESS_SPACE_MEMORY); in qmp_query_pci_bridge() 87 range->limit = pci_bridge_get_limit(dev, PCI_BASE_ADDRESS_SPACE_MEMORY); in qmp_query_pci_bridge() 89 range = info->bus->prefetchable_range = g_new0(PciMemoryRange, 1); in qmp_query_pci_bridge() 90 range->base = pci_bridge_get_base(dev, PCI_BASE_ADDRESS_MEM_PREFETCH); in qmp_query_pci_bridge() 91 range->limit = pci_bridge_get_limit(dev, PCI_BASE_ADDRESS_MEM_PREFETCH); in qmp_query_pci_bridge()
|
/qemu/tests/tcg/aarch64/gdbstub/ |
H A D | test-sve.py | 25 for i in range(0, 32): 29 for j in range(0, 4): 33 for j in range(0, 4): 37 for j in range(0, 8): 41 for j in range(0, 8):
|
H A D | test-sve-ioctl.py | 25 for i in range(0, int(val_i)): 28 for i in range(i + 1, initial_vlen): 37 for i in range(0, 16):
|
/qemu/util/ |
H A D | userfaultfd.c | 163 uffd_register.range.start = (uintptr_t) addr; in uffd_register_memory() 164 uffd_register.range.len = length; in uffd_register_memory() 218 uffd_writeprotect.range.start = (uintptr_t) addr; in uffd_change_protection() 219 uffd_writeprotect.range.len = length; in uffd_change_protection() 287 uffd_zeropage.range.start = (uintptr_t) addr; in uffd_zero_page() 288 uffd_zeropage.range.len = length; in uffd_zero_page()
|
H A D | reserved-region.c | 27 Range *r = ®->range; in resv_region_list_insert() 33 range_iter = &resv_iter->range; in resv_region_list_insert() 73 range_set_bounds(&new_reg->range, in resv_region_list_insert()
|
/qemu/scripts/ |
H A D | cpu-x86-uarch-abi.py | 128 for level in range(len(levels)): 147 for level in range(len(levels)): 152 for level in range(len(abi_models)): 185 for level in range(len(levels)):
|
/qemu/linux-headers/linux/ |
H A D | userfaultfd.h | 262 struct uffdio_range range; member 297 struct uffdio_range range; member 309 struct uffdio_range range; member 329 struct uffdio_range range; member 348 struct uffdio_range range; member
|
/qemu/tests/image-fuzzer/qcow2/ |
H A D | layout.py | 300 guest_clusters = random.sample(range(self.image_size // 309 meta_data |= set(range(l1_start, l1_start + l1_size)) 368 table_clusters = set(range(table_start, table_start + table_size)) 371 table_clusters_allocated = set(range(table_start, table_start + 423 block_clusters = set([random.choice(list(set(range(1, 4)) - 426 table_clusters = set([random.choice(list(set(range(1, 4)) - 542 free = set(range(1, append_id)) - used 546 return free | set(range(append_id, append_id + number - len(free))) 564 for i in range(1, len(lst)): 578 free = list(set(range(1, append_id)) - used) [all …]
|
/qemu/scripts/simplebench/ |
H A D | results_to_text.py | 72 tab.append([''] + [chr(ord('A') + i) for i in range(n_columns)]) 87 for i in range(1, n_columns): 97 for j in range(0, i):
|
/qemu/hw/vfio/ |
H A D | container.c | 244 struct vfio_iommu_type1_dirty_bitmap_get *range; in vfio_legacy_query_dirty_bitmap() local 247 dbitmap = g_malloc0(sizeof(*dbitmap) + sizeof(*range)); in vfio_legacy_query_dirty_bitmap() 249 dbitmap->argsz = sizeof(*dbitmap) + sizeof(*range); in vfio_legacy_query_dirty_bitmap() 252 range->iova = iova; in vfio_legacy_query_dirty_bitmap() 253 range->size = size; in vfio_legacy_query_dirty_bitmap() 260 range->bitmap.pgsize = qemu_real_host_page_size(); in vfio_legacy_query_dirty_bitmap() 261 range->bitmap.size = vbmap->size; in vfio_legacy_query_dirty_bitmap() 262 range->bitmap.data = (__u64 *)vbmap->bitmap; in vfio_legacy_query_dirty_bitmap() 269 (uint64_t)range->size, errno); in vfio_legacy_query_dirty_bitmap() 323 Range *range = g_new(Range, 1); in vfio_get_info_iova_range() local [all …]
|
/qemu/tests/qemu-iotests/tests/ |
H A D | migrate-bitmaps-postcopy-test | 120 for i in range(nb_bitmaps): 141 for i in range(1, nb_bitmaps): 152 for i in range(2, nb_bitmaps, 2): 211 for i in range(0, nb_bitmaps, 5): 224 for i in range(0, nb_bitmaps):
|