Searched refs:BITS_PER_LONG (Results 1 – 15 of 15) sorted by last modified time
/qemu/target/s390x/ |
H A D | cpu_features.h | 84 #define BE_BIT_NR(BIT) (BIT ^ (BITS_PER_LONG - 1))
|
/qemu/migration/ |
H A D | ram.c | 306 le_bitmap = bitmap_new(nbits + BITS_PER_LONG); in ramblock_recv_bitmap_send() 4393 le_bitmap = bitmap_new(nbits + BITS_PER_LONG); in ram_dirty_bitmap_reload()
|
/qemu/system/ |
H A D | physmem.c | 1468 candidate = ROUND_UP(candidate, BITS_PER_LONG << TARGET_PAGE_BITS); in find_ram_offset()
|
/qemu/include/qemu/ |
H A D | bitmap.h | 66 #define BITMAP_FIRST_WORD_MASK(start) (~0UL << ((start) & (BITS_PER_LONG - 1))) 67 #define BITMAP_LAST_WORD_MASK(nbits) (~0UL >> (-(nbits) & (BITS_PER_LONG - 1))) 73 ((nbits) <= BITS_PER_LONG) 239 long aligned_offset = QEMU_ALIGN_DOWN(offset, BITS_PER_LONG); in bitmap_count_one_with_offset() 243 aligned_offset / BITS_PER_LONG; in bitmap_count_one_with_offset()
|
H A D | bitops.h | 20 #define BITS_PER_LONG (sizeof (unsigned long) * BITS_PER_BYTE) macro 24 #define BIT_MASK(nr) (1UL << ((nr) % BITS_PER_LONG)) 25 #define BIT_WORD(nr) ((nr) / BITS_PER_LONG) 148 return 1UL & (addr[BIT_WORD(nr)] >> (nr & (BITS_PER_LONG-1))); in test_bit() 202 for (result = 0; result < size; result += BITS_PER_LONG) { in find_first_bit()
|
H A D | hbitmap.h | 21 #define BITS_PER_LEVEL (BITS_PER_LONG == 32 ? 5 : 6) 27 #define HBITMAP_LOG_MAX_SIZE (BITS_PER_LONG == 32 ? 34 : 41)
|
/qemu/accel/kvm/ |
H A D | kvm-all.c | 943 assert(bmap_start % BITS_PER_LONG == 0); in kvm_log_clear_one_slot()
|
/qemu/target/riscv/kvm/ |
H A D | kvm-cpu.c | 1583 socket_bits = find_last_bit(&socket_count, BITS_PER_LONG) + 1; in kvm_riscv_aia_create() 1602 find_last_bit(&guest_num, BITS_PER_LONG) + 1; in kvm_riscv_aia_create() 1640 hart_bits = find_last_bit(&max_hart_per_socket, BITS_PER_LONG) + 1; in kvm_riscv_aia_create()
|
/qemu/include/exec/ |
H A D | ram_addr.h | 361 if ((((page * BITS_PER_LONG) << TARGET_PAGE_BITS) == start) && in cpu_physical_memory_set_dirty_lebitmap() 488 if (((word * BITS_PER_LONG) << TARGET_PAGE_BITS) == in cpu_physical_memory_sync_dirty_bitmap() 490 !(length & ((BITS_PER_LONG << TARGET_PAGE_BITS) - 1))) { in cpu_physical_memory_sync_dirty_bitmap() 494 unsigned long idx = (word * BITS_PER_LONG) / DIRTY_MEMORY_BLOCK_SIZE; in cpu_physical_memory_sync_dirty_bitmap() 495 unsigned long offset = BIT_WORD((word * BITS_PER_LONG) % in cpu_physical_memory_sync_dirty_bitmap()
|
/qemu/util/ |
H A D | bitops.c | 31 offset %= BITS_PER_LONG; in find_next_bit() 41 size -= BITS_PER_LONG; in find_next_bit() 42 result += BITS_PER_LONG; in find_next_bit() 58 size -= 4*BITS_PER_LONG; in find_next_bit() 64 result += BITS_PER_LONG; in find_next_bit() 65 size -= BITS_PER_LONG; in find_next_bit() 96 offset %= BITS_PER_LONG; in find_next_zero_bit() 106 size -= BITS_PER_LONG; in find_next_zero_bit() 107 result += BITS_PER_LONG; in find_next_zero_bit() 114 size -= BITS_PER_LONG; in find_next_zero_bit() [all …]
|
H A D | hbitmap.c | 181 bit = pos & (BITS_PER_LONG - 1); in hbitmap_iter_init() 403 int bit = end & (BITS_PER_LONG - 1); in hb_count_between() 422 mask = 2UL << (last & (BITS_PER_LONG - 1)); in hb_set_elem() 423 mask -= 1UL << (start & (BITS_PER_LONG - 1)); in hb_set_elem() 445 next += BITS_PER_LONG; in hb_set_between() 500 mask = 2UL << (last & (BITS_PER_LONG - 1)); in hb_reset_elem() 501 mask -= 1UL << (start & (BITS_PER_LONG - 1)); in hb_reset_elem() 534 next += BITS_PER_LONG; in hb_reset_between() 595 hb->levels[0][0] = 1UL << (BITS_PER_LONG - 1); in hbitmap_reset_all() 711 if (BITS_PER_LONG == 32) { in hbitmap_deserialize_part() [all …]
|
H A D | bitmap.c | 50 if (bits % BITS_PER_LONG) { in slow_bitmap_empty() 69 if (bits % BITS_PER_LONG) { in slow_bitmap_full() 89 if (bits % BITS_PER_LONG) { in slow_bitmap_equal() 164 int bits_to_set = BITS_PER_LONG - (start % BITS_PER_LONG); in bitmap_set() 186 int bits_to_set = BITS_PER_LONG - (start % BITS_PER_LONG); in bitmap_set_atomic() 225 int bits_to_clear = BITS_PER_LONG - (start % BITS_PER_LONG); in bitmap_clear() 247 int bits_to_clear = BITS_PER_LONG - (start % BITS_PER_LONG); in bitmap_test_and_clear() 292 int bits_to_clear = BITS_PER_LONG - (start % BITS_PER_LONG); in bitmap_test_and_clear_atomic() 342 nr -= BITS_PER_LONG; in bitmap_copy_and_clear_atomic() 463 shift %= BITS_PER_LONG; in bitmap_copy_with_src_offset() [all …]
|
/qemu/tests/unit/ |
H A D | test-hbitmap.c | 19 #define L1 BITS_PER_LONG 20 #define L2 (BITS_PER_LONG * L1) 21 #define L3 (BITS_PER_LONG * L2) 55 bit = i & (BITS_PER_LONG - 1); in hbitmap_test_check() 65 bit = i & (BITS_PER_LONG - 1); in hbitmap_test_check() 84 n = DIV_ROUND_UP(size, BITS_PER_LONG); in hbitmap_test_init() 98 size_t n = DIV_ROUND_UP(bits, BITS_PER_LONG); in hbitmap_test_array_size() 154 int bit = first & (BITS_PER_LONG - 1); in hbitmap_test_set() 173 int bit = first & (BITS_PER_LONG - 1); in hbitmap_test_reset() 190 n = DIV_ROUND_UP(data->size, BITS_PER_LONG); in hbitmap_test_reset_all() [all …]
|
H A D | test-bitmap.c | 28 total = BITS_PER_LONG * 4; in check_bitmap_copy_with_offset() 37 g_assert_cmpmem(bmap1, total / BITS_PER_LONG, in check_bitmap_copy_with_offset() 38 bmap2, total / BITS_PER_LONG); in check_bitmap_copy_with_offset() 82 set_func(bmap, BITS_PER_LONG, 2 * BITS_PER_LONG); in bitmap_set_case() 85 g_assert_cmpint(find_first_bit(bmap, BITS_PER_LONG), ==, BITS_PER_LONG); in bitmap_set_case() 86 g_assert_cmpint(find_next_zero_bit(bmap, 3 * BITS_PER_LONG, BITS_PER_LONG), in bitmap_set_case() 87 ==, 3 * BITS_PER_LONG); in bitmap_set_case() 92 set_func(bmap, BITS_PER_LONG - offset, 2 * BITS_PER_LONG + offset); in bitmap_set_case() 100 ==, 3 * BITS_PER_LONG); in bitmap_set_case() 106 set_func(bmap, BITS_PER_LONG, 2 * BITS_PER_LONG + offset); in bitmap_set_case() [all …]
|
/qemu/tests/qtest/ |
H A D | npcm7xx_rng-test.c | 129 g_assert(nr_bits % BITS_PER_LONG == 0); in calc_runs_p() 131 for (j = 0; j < nr_bits / BITS_PER_LONG; j++) { in calc_runs_p() 209 unsigned long l[TEST_INPUT_BITS / BITS_PER_LONG]; in test_continuous_runs() 254 unsigned long l[TEST_INPUT_BITS / BITS_PER_LONG]; in test_first_byte_runs()
|