/linux/drivers/md/dm-vdo/ |
H A D | io-submitter.c | 148 vio->bios_merged.head->bi_iter.bi_sector); in get_bio_list() 150 vio->bios_merged.tail->bi_iter.bi_sector); in get_bio_list() 193 sector_t merge_sector = bio->bi_iter.bi_sector; in get_mergeable_locked() 216 return (vio_merge->bios_merged.tail->bi_iter.bi_sector == merge_sector ? in get_mergeable_locked() 220 return (vio_merge->bios_merged.head->bi_iter.bi_sector == merge_sector ? in get_mergeable_locked() 229 bio_sector = vio->bios_merged.head->bi_iter.bi_sector; in map_merged_vio() 234 bio_sector = vio->bios_merged.tail->bi_iter.bi_sector; in map_merged_vio() 241 vdo_int_map_remove(bio_map, prev_vio->bios_merged.tail->bi_iter.bi_sector); in merge_to_prev_tail() 254 vdo_int_map_remove(bio_map, next_vio->bios_merged.head->bi_iter.bi_sector); in merge_to_next_head() 291 bio->bi_iter.bi_sector, in try_bio_map_merge()
|
/linux/drivers/md/ |
H A D | dm-ebs-target.c | 50 sector_t end_sector = __block_mod(bio->bi_iter.bi_sector, ec->u_bs) + bio_sectors(bio); in __nr_blocks() 72 unsigned int buf_off = to_bytes(__block_mod(iter->bi_sector, ec->u_bs)); in __ebs_rw_bvec() 73 sector_t block = __sector_to_block(ec, iter->bi_sector); in __ebs_rw_bvec() 145 sector_t block, blocks, sector = bio->bi_iter.bi_sector; in __ebs_discard_bio() 169 sector_t blocks, sector = bio->bi_iter.bi_sector; in __ebs_forget_bio() 195 block1 = __sector_to_block(ec, bio->bi_iter.bi_sector); in __ebs_process_bios() 200 if (__block_mod(bio->bi_iter.bi_sector, ec->u_bs)) in __ebs_process_bios() 366 bio->bi_iter.bi_sector = ec->start + dm_target_offset(ti, bio->bi_iter.bi_sector); in ebs_map() 375 if (likely(__block_mod(bio->bi_iter.bi_sector, ec->u_bs) || in ebs_map()
|
H A D | dm-linear.c | 81 static sector_t linear_map_sector(struct dm_target *ti, sector_t bi_sector) in linear_map_sector() argument 85 return lc->start + dm_target_offset(ti, bi_sector); in linear_map_sector() 93 bio->bi_iter.bi_sector = linear_map_sector(ti, bio->bi_iter.bi_sector); in linear_map()
|
H A D | dm-flakey.c | 326 static sector_t flakey_map_sector(struct dm_target *ti, sector_t bi_sector) in flakey_map_sector() argument 330 return fc->start + dm_target_offset(ti, bi_sector); in flakey_map_sector() 338 bio->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); in flakey_map_bio() 360 (unsigned long long)bio->bi_iter.bi_sector, in corrupt_bio_common() 431 clone->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); in clone_bio()
|
H A D | dm-stripe.c | 254 stripe_map_range_sector(sc, bio->bi_iter.bi_sector, in stripe_map_range() 260 bio->bi_iter.bi_sector = begin + in stripe_map_range() 291 stripe_map_sector(sc, bio->bi_iter.bi_sector, in stripe_map() 292 &stripe, &bio->bi_iter.bi_sector); in stripe_map() 294 bio->bi_iter.bi_sector += sc->stripe[stripe].physical_start; in stripe_map()
|
H A D | dm-io-rewind.c | 62 bip->bip_iter.bi_sector -= bio_integrity_intervals(bi, bytes_done >> 9); in dm_bio_integrity_rewind() 113 iter->bi_sector -= bytes >> 9; in dm_bio_rewind_iter()
|
H A D | dm-stats.c | 635 sector_t bi_sector, sector_t end_sector, in __dm_stat_bio() argument 642 if (end_sector <= s->start || bi_sector >= s->end) in __dm_stat_bio() 644 if (unlikely(bi_sector < s->start)) { in __dm_stat_bio() 648 rel_sector = bi_sector - s->start; in __dm_stat_bio() 649 todo = end_sector - bi_sector; in __dm_stat_bio() 673 sector_t bi_sector, unsigned int bi_sectors, bool end, in dm_stats_account_io() argument 686 end_sector = bi_sector + bi_sectors; in dm_stats_account_io() 695 (bi_sector == (READ_ONCE(last->last_sector) && in dm_stats_account_io() 714 __dm_stat_bio(s, bi_rw, bi_sector, end_sector, end, duration_jiffies, stats_aux); in dm_stats_account_io()
|
H A D | dm-unstripe.c | 120 sector_t sector = bio->bi_iter.bi_sector; in map_to_core() 140 bio->bi_iter.bi_sector = map_to_core(ti, bio) + uc->physical_start; in unstripe_map()
|
H A D | dm-dust.c | 230 bio->bi_iter.bi_sector = dd->start + dm_target_offset(ti, bio->bi_iter.bi_sector); in dust_map() 233 r = dust_map_read(dd, bio->bi_iter.bi_sector, dd->fail_read_on_bb); in dust_map() 235 r = dust_map_write(dd, bio->bi_iter.bi_sector, dd->fail_read_on_bb); in dust_map()
|
H A D | dm-stats.h | 33 sector_t bi_sector, unsigned int bi_sectors, bool end,
|
/linux/drivers/md/bcache/ |
H A D | request.c | 123 bio->bi_iter.bi_sector += sectors; in bch_data_invalidate() 128 bio->bi_iter.bi_sector, in bch_data_invalidate() 220 SET_KEY_OFFSET(k, bio->bi_iter.bi_sector); in CLOSURE_CALLBACK() 431 if (i->last == bio->bi_iter.bi_sector && in check_should_bypass() 538 KEY_START(k) > bio->bi_iter.bi_sector) { in cache_lookup_fn() 542 KEY_START(k) - bio->bi_iter.bi_sector) in cache_lookup_fn() 848 s->iop.bio->bi_iter.bi_sector = in CLOSURE_CALLBACK() 849 s->cache_miss->bi_iter.bi_sector; in CLOSURE_CALLBACK() 935 cache_bio->bi_iter.bi_sector = miss->bi_iter.bi_sector; in cached_dev_cache_miss() 1218 bio->bi_iter.bi_sector += dc->sb.data_offset; in cached_dev_submit_bio() [all …]
|
H A D | debug.c | 54 bio->bi_iter.bi_sector = PTR_OFFSET(&b->key, 0); in bch_btree_verify() 120 check->bi_iter.bi_sector = bio->bi_iter.bi_sector; in bch_data_verify() 141 (uint64_t) bio->bi_iter.bi_sector); in bch_data_verify()
|
/linux/fs/btrfs/ |
H A D | bio.c | 222 repair_bbio->saved_iter.bi_sector << SECTOR_SHIFT, in btrfs_end_repair_bio() 273 repair_bio->bi_iter.bi_sector = failed_bbio->saved_iter.bi_sector; in repair_one_sector() 425 stripe->physical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_orig_write_end_io() 439 stripe->physical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_clone_write_end_io() 464 u64 physical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_submit_dev_bio() 468 bio->bi_iter.bi_sector = zone_start >> SECTOR_SHIFT; in btrfs_submit_dev_bio() 472 __func__, bio_op(bio), bio->bi_opf, bio->bi_iter.bi_sector, in btrfs_submit_dev_bio() 511 bio->bi_iter.bi_sector = smap->physical >> SECTOR_SHIFT; in __btrfs_submit_bio() 673 u64 logical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_submit_chunk() 816 bio.bi_iter.bi_sector = smap.physical >> SECTOR_SHIFT; in btrfs_repair_io_failure() [all …]
|
/linux/fs/xfs/ |
H A D | xfs_bio_io.c | 31 bio->bi_iter.bi_sector = sector; in xfs_rw_bdev() 43 bio->bi_iter.bi_sector = bio_end_sector(prev); in xfs_rw_bdev()
|
/linux/block/ |
H A D | blk-lib.c | 50 bio->bi_iter.bi_sector = *sector; in blk_alloc_discard_bio() 126 bio->bi_iter.bi_sector = sector; in __blkdev_issue_write_zeroes() 167 bio->bi_iter.bi_sector = sector; in __blkdev_issue_zero_pages() 314 bio->bi_iter.bi_sector = sector; in blkdev_issue_secure_erase()
|
H A D | blk-merge.c | 135 tmp = bio->bi_iter.bi_sector + split_sectors - in bio_split_discard() 174 blk_chunk_sectors_left(bio->bi_iter.bi_sector, in get_max_io_size() 178 start = bio->bi_iter.bi_sector & (pbs - 1); in get_max_io_size() 379 trace_block_split(split, bio->bi_iter.bi_sector); in __bio_split_to_limits() 662 blk_rq_get_max_sectors(req, bio->bi_iter.bi_sector)) { in ll_front_merge_fn() 961 else if (blk_rq_pos(rq) + blk_rq_sectors(rq) == bio->bi_iter.bi_sector) in blk_try_merge() 963 else if (blk_rq_pos(rq) - bio_sectors(bio) == bio->bi_iter.bi_sector) in blk_try_merge() 1034 req->__sector = bio->bi_iter.bi_sector; in bio_attempt_front_merge()
|
H A D | bio-integrity.c | 256 bip->bip_iter.bi_sector = seed; in bio_integrity_copy_user() 276 bip->bip_iter.bi_sector = seed; in bio_integrity_init_user() 396 iter.seed = proc_iter->bi_sector; in bio_integrity_process() 480 bip_set_seed(bip, bio->bi_iter.bi_sector); in bio_integrity_prep() 588 bip->bip_iter.bi_sector += bio_integrity_intervals(bi, bytes_done >> 9); in bio_integrity_advance()
|
/linux/mm/ |
H A D | page_io.c | 46 (unsigned long long)bio->bi_iter.bi_sector); in __end_swap_bio_write() 65 (unsigned long long)bio->bi_iter.bi_sector); in __end_swap_bio_read() 343 bio.bi_iter.bi_sector = swap_folio_sector(folio); in swap_writepage_bdev_sync() 364 bio->bi_iter.bi_sector = swap_folio_sector(folio); in swap_writepage_bdev_async() 470 bio.bi_iter.bi_sector = swap_folio_sector(folio); in swap_read_folio_bdev_sync() 489 bio->bi_iter.bi_sector = swap_folio_sector(folio); in swap_read_folio_bdev_async()
|
/linux/include/trace/events/ |
H A D | bcache.h | 28 __entry->sector = bio->bi_iter.bi_sector; 29 __entry->orig_sector = bio->bi_iter.bi_sector - 16; 103 __entry->sector = bio->bi_iter.bi_sector; 138 __entry->sector = bio->bi_iter.bi_sector; 169 __entry->sector = bio->bi_iter.bi_sector; 238 __entry->sector = bio->bi_iter.bi_sector;
|
H A D | block.h | 300 __entry->sector = bio->bi_iter.bi_sector; 328 __entry->sector = bio->bi_iter.bi_sector; 485 __entry->sector = bio->bi_iter.bi_sector; 524 __entry->sector = bio->bi_iter.bi_sector;
|
/linux/drivers/nvdimm/ |
H A D | nd_virtio.c | 107 if (bio && bio->bi_iter.bi_sector != -1) { in async_pmem_flush() 115 child->bi_iter.bi_sector = -1; in async_pmem_flush()
|
/linux/fs/ext4/ |
H A D | page-io.c | 326 sector_t bi_sector = bio->bi_iter.bi_sector; in ext4_end_bio() local 330 (long long) bio->bi_iter.bi_sector, in ext4_end_bio() 346 bi_sector >> (inode->i_blkbits - 9)); in ext4_end_bio() 401 bio->bi_iter.bi_sector = bh->b_blocknr * (bh->b_size >> 9); in io_submit_init_bio()
|
/linux/fs/crypto/ |
H A D | bio.c | 68 bio->bi_iter.bi_sector = in fscrypt_zeroout_range_inline_crypt() 161 bio->bi_iter.bi_sector = sector; in fscrypt_zeroout_range()
|
/linux/include/linux/ |
H A D | bio.h | 40 #define bvec_iter_end_sector(iter) ((iter).bi_sector + bvec_iter_sectors((iter))) 101 iter->bi_sector += bytes >> 9; in bio_advance_iter() 115 iter->bi_sector += bytes >> 9; in bio_advance_iter_single() 376 return bip->bip_iter.bi_sector; in bip_get_seed() 382 bip->bip_iter.bi_sector = seed; in bip_set_seed()
|
H A D | bvec.h | 78 sector_t bi_sector; /* device address in 512 byte member 190 .bi_sector = 0, \
|