/qemu/block/ |
H A D | parallels.c | 214 payload_bytes -= s->data_start * BDRV_SECTOR_SIZE; in parallels_fill_used_bitmap() 283 bytes += s->prealloc_size * BDRV_SECTOR_SIZE; in allocate_clusters() 285 host_off = s->data_end * BDRV_SECTOR_SIZE; in allocate_clusters() 321 host_off = s->data_start * BDRV_SECTOR_SIZE; in allocate_clusters() 378 if (host_off > s->data_end * BDRV_SECTOR_SIZE) { in allocate_clusters() 379 s->data_end = host_off / BDRV_SECTOR_SIZE; in allocate_clusters() 432 *pnum = count * BDRV_SECTOR_SIZE; in parallels_co_block_status() 437 *map = offset * BDRV_SECTOR_SIZE; in parallels_co_block_status() 1004 uint8_t tmp[BDRV_SECTOR_SIZE]; in parallels_co_create() 1034 if (!QEMU_IS_ALIGNED(cl_size, BDRV_SECTOR_SIZE)) { in parallels_co_create() [all …]
|
H A D | bochs.c | 174 if (s->extent_size < BDRV_SECTOR_SIZE) { in bochs_open() 192 s->extent_size / BDRV_SECTOR_SIZE)) in bochs_open() 209 bs->bl.request_alignment = BDRV_SECTOR_SIZE; /* No sub-sector I/O */ in bochs_refresh_limits() 258 assert(QEMU_IS_ALIGNED(offset, BDRV_SECTOR_SIZE)); in bochs_co_preadv() 259 assert(QEMU_IS_ALIGNED(bytes, BDRV_SECTOR_SIZE)); in bochs_co_preadv()
|
H A D | replication.c | 241 ret = bdrv_co_preadv(bs->file, sector_num * BDRV_SECTOR_SIZE, in replication_co_readv() 242 remaining_sectors * BDRV_SECTOR_SIZE, qiov, 0); in replication_co_readv() 266 ret = bdrv_co_pwritev(top, sector_num * BDRV_SECTOR_SIZE, in replication_co_writev() 267 remaining_sectors * BDRV_SECTOR_SIZE, qiov, 0); in replication_co_writev() 280 sector_num * BDRV_SECTOR_SIZE, in replication_co_writev() 281 remaining_sectors * BDRV_SECTOR_SIZE, in replication_co_writev() 287 assert(QEMU_IS_ALIGNED(count, BDRV_SECTOR_SIZE)); in replication_co_writev() 293 ret = bdrv_co_pwritev(target, sector_num * BDRV_SECTOR_SIZE, in replication_co_writev() 294 n * BDRV_SECTOR_SIZE, &hd_qiov, 0); in replication_co_writev()
|
H A D | qcow.c | 443 assert(QEMU_IS_ALIGNED(n_start | n_end, BDRV_SECTOR_SIZE)); in get_cluster_offset() 489 for (i = 0; i < s->cluster_size; i += BDRV_SECTOR_SIZE) { in get_cluster_offset() 491 memset(s->cluster_data, 0x00, BDRV_SECTOR_SIZE); in get_cluster_offset() 495 BDRV_SECTOR_SIZE, in get_cluster_offset() 501 BDRV_SECTOR_SIZE, in get_cluster_offset() 625 bs->bl.request_alignment = BDRV_SECTOR_SIZE; in qcow_refresh_limits() 911 tmp = g_malloc0(BDRV_SECTOR_SIZE); in qcow_co_create() 912 for (i = 0; i < DIV_ROUND_UP(sizeof(uint64_t) * l1_size, BDRV_SECTOR_SIZE); in qcow_co_create() 914 ret = blk_co_pwrite(qcow_blk, header_size + BDRV_SECTOR_SIZE * i, in qcow_co_create() 915 BDRV_SECTOR_SIZE, tmp, 0); in qcow_co_create() [all …]
|
H A D | iscsi.c | 424 return sector * iscsilun->block_size / BDRV_SECTOR_SIZE; in sector_lun2qemu() 429 return sector * BDRV_SECTOR_SIZE / iscsilun->block_size; in sector_qemu2lun() 678 nb_sectors * BDRV_SECTOR_SIZE); 686 nb_sectors * BDRV_SECTOR_SIZE); 820 if (iscsi_allocmap_is_valid(iscsilun, sector_num * BDRV_SECTOR_SIZE, in iscsi_co_readv() 821 nb_sectors * BDRV_SECTOR_SIZE) && in iscsi_co_readv() 823 nb_sectors * BDRV_SECTOR_SIZE)) { in iscsi_co_readv() 830 nb_sectors * BDRV_SECTOR_SIZE) && in iscsi_co_readv() 832 nb_sectors * BDRV_SECTOR_SIZE)) { in iscsi_co_readv() 850 pnum >= nb_sectors * BDRV_SECTOR_SIZE + head) { in iscsi_co_readv() [all …]
|
H A D | vmdk.c | 1055 * BDRV_SECTOR_SIZE)); in vmdk_open_vmdk4() 1774 (extent->end_sector - extent->sectors) * BDRV_SECTOR_SIZE; in vmdk_find_offset_in_cluster() 2071 if (DIV_ROUND_UP(offset, BDRV_SECTOR_SIZE) > bs->total_sectors) { in vmdk_pwritev() 2185 length = QEMU_ALIGN_UP(length, BDRV_SECTOR_SIZE); in vmdk_co_pwritev_compressed() 2242 header.capacity = filesize / BDRV_SECTOR_SIZE; in vmdk_init_extent() 2244 header.num_gtes_per_gt = BDRV_SECTOR_SIZE; in vmdk_init_extent() 2248 BDRV_SECTOR_SIZE); in vmdk_init_extent() 2296 gd_buf_size = gd_sectors * BDRV_SECTOR_SIZE; in vmdk_init_extent() 2740 BDRV_SECTOR_SIZE); in vmdk_co_create_opts() 2868 if (!QEMU_IS_ALIGNED(opts->size, BDRV_SECTOR_SIZE)) { in vmdk_co_create() [all …]
|
H A D | vpc.c | 340 BDRV_SECTOR_SIZE; in vpc_open() 364 if (!is_power_of_2(s->block_size) || s->block_size < BDRV_SECTOR_SIZE) { in vpc_open() 566 if ((offset < 0) || (offset > bs->total_sectors * BDRV_SECTOR_SIZE)) { in alloc_block() 945 total_sectors = MIN(VHD_MAX_GEOMETRY, total_size / BDRV_SECTOR_SIZE); in calculate_rounded_image_size() 952 total_sectors = total_size / BDRV_SECTOR_SIZE; in calculate_rounded_image_size() 1030 if (total_size != total_sectors * BDRV_SECTOR_SIZE) { in vpc_co_create() 1036 total_sectors * BDRV_SECTOR_SIZE); in vpc_co_create() 1148 ROUND_UP(create_options->u.vpc.size, BDRV_SECTOR_SIZE); in vpc_co_create_opts() 1158 create_options->u.vpc.size = total_sectors * BDRV_SECTOR_SIZE; in vpc_co_create_opts()
|
H A D | cloop.c | 218 bs->bl.request_alignment = BDRV_SECTOR_SIZE; /* No sub-sector I/O */ in cloop_refresh_limits() 263 assert(QEMU_IS_ALIGNED(offset, BDRV_SECTOR_SIZE)); in cloop_co_preadv() 264 assert(QEMU_IS_ALIGNED(bytes, BDRV_SECTOR_SIZE)); in cloop_co_preadv()
|
H A D | raw-format.c | 128 if (has_size && !QEMU_IS_ALIGNED(size, BDRV_SECTOR_SIZE)) { in raw_apply_options() 130 BDRV_SECTOR_SIZE); in raw_apply_options() 238 QEMU_BUILD_BUG_ON(BDRV_SECTOR_SIZE != 512); in raw_co_pwritev() 385 BDRV_SECTOR_SIZE); in raw_measure() 410 bs->bl.request_alignment = BDRV_SECTOR_SIZE; in raw_refresh_limits()
|
H A D | preallocate.c | 123 if (!QEMU_IS_ALIGNED(dest->prealloc_align, BDRV_SECTOR_SIZE)) { in preallocate_absorb_opts() 125 "is not aligned to %llu", BDRV_SECTOR_SIZE); in preallocate_absorb_opts() 566 bs->file->bs->total_sectors * BDRV_SECTOR_SIZE; in preallocate_set_perm()
|
H A D | vvfat.c | 1301 bs->bl.request_alignment = BDRV_SECTOR_SIZE; /* No sub-sector I/O */ in vvfat_refresh_limits() 1493 if (bdrv_co_pread(s->qcow, sector_num * BDRV_SECTOR_SIZE, n, in vvfat_read() 1545 assert(QEMU_IS_ALIGNED(offset, BDRV_SECTOR_SIZE)); in vvfat_co_preadv() 1546 assert(QEMU_IS_ALIGNED(bytes, BDRV_SECTOR_SIZE)); in vvfat_co_preadv() 1812 i) * BDRV_SECTOR_SIZE, in cluster_was_modified() 1813 BDRV_SECTOR_SIZE, NULL); in cluster_was_modified() 1973 BDRV_SECTOR_SIZE, NULL); in get_cluster_count_for_direntry() 3089 ret = bdrv_co_pwrite(s->qcow, sector_num * BDRV_SECTOR_SIZE, in vvfat_write() 3090 nb_sectors * BDRV_SECTOR_SIZE, buf, 0); in vvfat_write() 3120 assert(QEMU_IS_ALIGNED(offset, BDRV_SECTOR_SIZE)); in vvfat_co_pwritev() [all …]
|
H A D | ssh.c | 971 BDRV_SECTOR_SIZE); in ssh_co_create_opts() 1131 ret = ssh_read(s, bs, sector_num * BDRV_SECTOR_SIZE, in ssh_co_readv() 1132 nb_sectors * BDRV_SECTOR_SIZE, qiov); in ssh_co_readv() 1206 ret = ssh_write(s, bs, sector_num * BDRV_SECTOR_SIZE, in ssh_co_writev() 1207 nb_sectors * BDRV_SECTOR_SIZE, qiov); in ssh_co_writev()
|
H A D | qed.c | 111 int nsectors = DIV_ROUND_UP(sizeof(QEDHeader), BDRV_SECTOR_SIZE); in qed_write_header() 112 size_t len = nsectors * BDRV_SECTOR_SIZE; in qed_write_header() 177 if (image_size % BDRV_SECTOR_SIZE != 0) { in qed_is_image_size_valid() 822 ROUND_UP(create_options->u.qed.size, BDRV_SECTOR_SIZE); in bdrv_qed_co_create_opts() 1398 .cur_pos = (uint64_t) sector_num * BDRV_SECTOR_SIZE, in qed_co_request() 1399 .end_pos = (sector_num + nb_sectors) * BDRV_SECTOR_SIZE, in qed_co_request()
|
H A D | dmg.c | 595 bs->bl.request_alignment = BDRV_SECTOR_SIZE; /* No sub-sector I/O */ in dmg_refresh_limits() 737 assert(QEMU_IS_ALIGNED(offset, BDRV_SECTOR_SIZE)); in dmg_co_preadv() 738 assert(QEMU_IS_ALIGNED(bytes, BDRV_SECTOR_SIZE)); in dmg_co_preadv()
|
H A D | qcow2.c | 1530 bs->total_sectors = header.size / BDRV_SECTOR_SIZE; in qcow2_do_open() 2929 total_size = bs->total_sectors * BDRV_SECTOR_SIZE; in qcow2_update_header() 3521 if (!QEMU_IS_ALIGNED(qcow2_opts->size, BDRV_SECTOR_SIZE)) { in qcow2_co_create() 3523 (unsigned) BDRV_SECTOR_SIZE); in qcow2_co_create() 3994 BDRV_SECTOR_SIZE); in qcow2_co_create_opts() 4024 bytes = bs->total_sectors * BDRV_SECTOR_SIZE - offset; in is_zero() 4301 if (!QEMU_IS_ALIGNED(offset, BDRV_SECTOR_SIZE)) { in qcow2_co_truncate() 4303 (unsigned) BDRV_SECTOR_SIZE); in qcow2_co_truncate() 4325 old_length = bs->total_sectors * BDRV_SECTOR_SIZE; in qcow2_co_truncate() 4633 bs->total_sectors = offset / BDRV_SECTOR_SIZE; in qcow2_co_truncate() [all …]
|
/qemu/hw/scsi/ |
H A D | scsi-disk.c | 76 /* Both sector and sector_count are in terms of BDRV_SECTOR_SIZE bytes. */ 144 r->iov.iov_len = MIN(r->sector_count * BDRV_SECTOR_SIZE, r->buflen); in scsi_init_iovec() 410 n = r->qiov.size / BDRV_SECTOR_SIZE; in scsi_read_complete() 455 BDRV_SECTOR_SIZE, in scsi_do_read_cb() 541 n = r->qiov.size / BDRV_SECTOR_SIZE; in scsi_write_complete() 616 BDRV_SECTOR_SIZE, in scsi_disk_emulate_vpd_page() 1372 nb_sectors /= (s->qdev.blocksize / BDRV_SECTOR_SIZE); in scsi_disk_emulate_mode_sense() 1430 nb_sectors /= s->qdev.blocksize / BDRV_SECTOR_SIZE; in scsi_disk_emulate_read_toc() 1655 * size of 256 since we cannot handle anything below BDRV_SECTOR_SIZE. in scsi_disk_emulate_mode_select() 1732 r->sector = sector_num * (s->qdev.blocksize / BDRV_SECTOR_SIZE); in scsi_unmap_complete_noio() [all...] |
/qemu/hw/ppc/ |
H A D | pnv_pnor.c | 44 offset = QEMU_ALIGN_DOWN(offset, BDRV_SECTOR_SIZE); in pnv_pnor_update() 45 offset_end = QEMU_ALIGN_UP(offset_end, BDRV_SECTOR_SIZE); in pnv_pnor_update()
|
/qemu/hw/block/ |
H A D | hd-geometry.c | 58 uint8_t buf[BDRV_SECTOR_SIZE]; in guess_disk_lchs() 66 if (blk_pread(blk, 0, BDRV_SECTOR_SIZE, buf, 0) < 0) { in guess_disk_lchs()
|
H A D | onenand.c | 311 if (blk_pread(s->blk_cur, offset, BDRV_SECTOR_SIZE, buf, 0) < 0) { in onenand_load_spare() 336 || blk_pread(s->blk_cur, offset, BDRV_SECTOR_SIZE, dp, 0) < 0) { in onenand_prog_spare() 354 result = blk_pwrite(s->blk_cur, offset, BDRV_SECTOR_SIZE, dp, in onenand_prog_spare() 374 BDRV_SECTOR_SIZE, blankbuf, 0) < 0) { in onenand_erase() 378 BDRV_SECTOR_SIZE, tmpbuf, 0) < 0) { in onenand_erase() 383 BDRV_SECTOR_SIZE, tmpbuf, 0) < 0) { in onenand_erase()
|
H A D | nand.c | 742 BDRV_SECTOR_SIZE, iobuf, 0) < 0) { 749 BDRV_SECTOR_SIZE, iobuf, 0) < 0) { 754 BDRV_SECTOR_SIZE, iobuf, 0) < 0) { 762 if (blk_pwrite(s->blk, i, BDRV_SECTOR_SIZE, iobuf, 0) < 0) { 770 BDRV_SECTOR_SIZE, iobuf, 0) < 0) { 775 BDRV_SECTOR_SIZE, iobuf, 0) < 0) {
|
H A D | block.c | 137 conf->physical_block_size = BDRV_SECTOR_SIZE; in blkconf_blocksizes() 144 conf->logical_block_size = BDRV_SECTOR_SIZE; in blkconf_blocksizes()
|
/qemu/ |
H A D | qemu-img.c | 1158 for (i = 0; i < end; i += BDRV_SECTOR_SIZE) { in find_nonzero() 1191 buf += BDRV_SECTOR_SIZE; in is_allocated_sectors() 1253 buf += BDRV_SECTOR_SIZE * *pnum; in is_allocated_sectors_min() 1261 buf += BDRV_SECTOR_SIZE * *pnum; in is_allocated_sectors_min() 1293 chsize = BDRV_SECTOR_SIZE; in compare_buffers() 1752 count = n * BDRV_SECTOR_SIZE; in convert_iteration_sectors() 1767 count = BDRV_SECTOR_SIZE; in convert_iteration_sectors() 1780 n = DIV_ROUND_UP(count, BDRV_SECTOR_SIZE); in convert_iteration_sectors() 1870 buf += n * BDRV_SECTOR_SIZE; in convert_co_read() 1935 buf += n * BDRV_SECTOR_SIZE; in convert_co_write() [all …]
|
/qemu/hw/ide/ |
H A D | core.c | 789 ide_transfer_start(s, s->io_buffer, n * BDRV_SECTOR_SIZE, ide_sector_read); in ide_sector_read_cb() 822 qemu_iovec_init_buf(&s->qiov, s->io_buffer, n * BDRV_SECTOR_SIZE); in ide_sector_read() 825 n * BDRV_SECTOR_SIZE, BLOCK_ACCT_READ); in ide_sector_read() 964 BDRV_SECTOR_SIZE, ide_dma_cb, s); in ide_dma_cb() 968 BDRV_SECTOR_SIZE, ide_dma_cb, s); in ide_dma_cb() 972 &s->sg, offset, BDRV_SECTOR_SIZE, in ide_dma_cb() 997 s->nsector * BDRV_SECTOR_SIZE, BLOCK_ACCT_READ); in ide_sector_start_dma() 1001 s->nsector * BDRV_SECTOR_SIZE, BLOCK_ACCT_WRITE); in ide_sector_start_dma() 1058 ide_transfer_start(s, s->io_buffer, n1 * BDRV_SECTOR_SIZE, in ide_sector_write_cb() 1097 qemu_iovec_init_buf(&s->qiov, s->io_buffer, n * BDRV_SECTOR_SIZE); in ide_sector_write() [all …]
|
H A D | atapi.c | 835 if (max_len > BDRV_SECTOR_SIZE) { in cmd_get_configuration() 837 max_len = BDRV_SECTOR_SIZE; in cmd_get_configuration() 1211 memset(buf, 0, max_len > IDE_DMA_BUF_SECTORS * BDRV_SECTOR_SIZE + 4 ? in cmd_read_dvd_structure() 1212 IDE_DMA_BUF_SECTORS * BDRV_SECTOR_SIZE + 4 : max_len); in cmd_read_dvd_structure()
|
/qemu/include/block/ |
H A D | block-common.h | 263 #define BDRV_SECTOR_SIZE (1ULL << BDRV_SECTOR_BITS) macro
|