/linux/drivers/md/dm-vdo/ |
H A D | vio.c | 174 blk_opf_t bi_opf, physical_block_number_t pbn) in vdo_set_bio_properties() argument 185 bio->bi_opf = bi_opf; in vdo_set_bio_properties() 195 blk_opf_t bi_opf, physical_block_number_t pbn) in vio_reset_bio() argument 200 bio_reset(bio, bio->bi_bdev, bi_opf); in vio_reset_bio() 201 vdo_set_bio_properties(bio, vio, callback, bi_opf, pbn); in vio_reset_bio() 284 } else if ((vio->bio->bi_opf & REQ_PREFLUSH) == REQ_PREFLUSH) { in vio_record_metadata_io_error() 285 description = (((vio->bio->bi_opf & REQ_FUA) == REQ_FUA) ? in vio_record_metadata_io_error() 288 } else if ((vio->bio->bi_opf & REQ_FUA) == REQ_FUA) { in vio_record_metadata_io_error() 447 if (((bio->bi_opf & REQ_PREFLUSH) != 0) && (bio->bi_iter.bi_size == 0)) { in vdo_count_bios() 472 if ((bio->bi_opf & REQ_PREFLUSH) != 0) in vdo_count_bios() [all …]
|
H A D | vio.h | 122 blk_opf_t bi_opf, physical_block_number_t pbn); 125 blk_opf_t bi_opf, physical_block_number_t pbn);
|
/linux/fs/btrfs/ |
H A D | bio.c | 341 else if (!(bio->bi_opf & REQ_RAHEAD)) in btrfs_log_dev_io_error() 343 if (bio->bi_opf & REQ_PREFLUSH) in btrfs_log_dev_io_error() 350 if (bio->bi_opf & REQ_META) in btrfs_end_io_wq() 472 __func__, bio_op(bio), bio->bi_opf, bio->bi_iter.bi_sector, in btrfs_submit_dev_bio() 476 if (bio->bi_opf & REQ_BTRFS_CGROUP_PUNT) in btrfs_submit_dev_bio() 537 if (bbio->bio.bi_opf & REQ_META) in btrfs_bio_csum() 605 bio->bi_opf |= REQ_BTRFS_CGROUP_PUNT; in run_one_async_done() 631 if (op_is_sync(bbio->bio.bi_opf)) in should_async_write() 635 if ((bbio->bio.bi_opf & REQ_META) && btrfs_is_zoned(bbio->fs_info)) in should_async_write() 714 bio->bi_opf &= ~REQ_OP_WRITE; in btrfs_submit_chunk() [all …]
|
/linux/block/ |
H A D | blk-merge.c | 312 if (bio->bi_opf & REQ_NOWAIT) { in bio_split_rw() 375 split->bi_opf |= REQ_NOMERGE; in __bio_split_to_limits() 744 WARN_ON_ONCE((bio->bi_opf & REQ_FAILFAST_MASK) && in blk_rq_set_mixed_merge() 745 (bio->bi_opf & REQ_FAILFAST_MASK) != ff); in blk_rq_set_mixed_merge() 746 bio->bi_opf |= ff; in blk_rq_set_mixed_merge() 753 if (bio->bi_opf & REQ_RAHEAD) in bio_failfast() 756 return bio->bi_opf & REQ_FAILFAST_MASK; in bio_failfast() 768 if (bio->bi_opf & REQ_RAHEAD) in blk_update_mixed_merge() 769 bio->bi_opf |= REQ_FAILFAST_MASK; in blk_update_mixed_merge() 773 req->cmd_flags |= bio->bi_opf & REQ_FAILFAST_MASK; in blk_update_mixed_merge()
|
H A D | blk-throttle.h | 179 blkg_rwstat_add(&tg->stat_bytes, bio->bi_opf, in blk_should_throtl() 182 blkg_rwstat_add(&tg->stat_ios, bio->bi_opf, 1); in blk_should_throtl()
|
H A D | blk-core.c | 344 if (bio->bi_opf & REQ_NOWAIT) { in __bio_queue_enter() 517 if (op_is_flush(bio->bi_opf) && !bio_sectors(bio)) in bio_check_ro() 557 current->comm, bio->bi_bdev, bio->bi_opf, in bio_check_eod() 611 bio->bi_opf |= REQ_NOMERGE; in blk_check_zone_append() 763 if ((bio->bi_opf & REQ_NOWAIT) && !bdev_nowait(bdev)) in submit_bio_noacct() 781 if (op_is_flush(bio->bi_opf)) { in submit_bio_noacct() 786 bio->bi_opf &= ~(REQ_PREFLUSH | REQ_FUA); in submit_bio_noacct() 1245 sizeof_field(struct bio, bi_opf)); in blk_dev_init()
|
H A D | blk-mq-sched.h | 33 return !(bio->bi_opf & REQ_NOMERGE_FLAGS); in bio_mergeable()
|
H A D | fops.c | 85 bio.bi_opf |= REQ_NOWAIT; in __blkdev_direct_IO_simple() 227 bio->bi_opf |= REQ_NOWAIT; in __blkdev_direct_IO() 347 bio->bi_opf |= REQ_NOWAIT; in __blkdev_direct_IO_async() 350 bio->bi_opf |= REQ_POLLED; in __blkdev_direct_IO_async()
|
/linux/drivers/md/bcache/ |
H A D | writeback.h | 123 return (op_is_sync(bio->bi_opf) || in should_writeback() 124 bio->bi_opf & (REQ_META|REQ_PRIO) || in should_writeback()
|
H A D | request.c | 202 bio->bi_opf &= ~(REQ_PREFLUSH|REQ_FUA); in CLOSURE_CALLBACK() 247 n->bi_opf = REQ_OP_WRITE; in CLOSURE_CALLBACK() 405 if ((bio->bi_opf & (REQ_RAHEAD|REQ_BACKGROUND))) { in check_should_bypass() 406 if (!(bio->bi_opf & (REQ_META|REQ_PRIO)) && in check_should_bypass() 667 bio->bi_opf & REQ_PREFLUSH)) { in backing_request_endio() 758 s->iop.flush_journal = op_is_flush(bio->bi_opf); in search_alloc() 1033 if (bio->bi_opf & REQ_PREFLUSH) { in cached_dev_write()
|
/linux/include/trace/events/ |
H A D | bcache.h | 31 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 105 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 140 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 171 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 241 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
|
H A D | block.h | 303 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 330 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 487 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 528 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
|
/linux/drivers/md/ |
H A D | dm-raid1.c | 268 .bi_opf = REQ_OP_WRITE | REQ_PREFLUSH | REQ_SYNC, in mirror_flush() 547 .bi_opf = REQ_OP_READ, in read_async_bio() 659 blk_opf_t op_flags = bio->bi_opf & (REQ_FUA | REQ_PREFLUSH); in do_write() 661 .bi_opf = REQ_OP_WRITE | op_flags, in do_write() 670 io_req.bi_opf = REQ_OP_DISCARD | op_flags; in do_write() 708 if ((bio->bi_opf & REQ_PREFLUSH) || in do_writes() 1222 if (bio->bi_opf & REQ_RAHEAD) in mirror_map() 1259 if (!(bio->bi_opf & REQ_PREFLUSH) && in mirror_end_io() 1268 if (bio->bi_opf & REQ_RAHEAD) in mirror_end_io()
|
H A D | raid1.c | 1343 bio->bi_opf & REQ_NOWAIT)) { in raid1_read_request() 1644 mbio->bi_opf = bio_op(bio) | (bio->bi_opf & (REQ_SYNC | REQ_FUA)); in raid1_write_request() 1648 mbio->bi_opf |= MD_FAILFAST; in raid1_write_request() 2355 wbio->bi_opf = REQ_OP_WRITE; in sync_request_write() 2357 wbio->bi_opf |= MD_FAILFAST; in sync_request_write() 2516 wbio->bi_opf = REQ_OP_WRITE; in narrow_write_error() 2867 bio->bi_opf = REQ_OP_WRITE; in raid1_sync_request() 2894 bio->bi_opf = REQ_OP_READ; in raid1_sync_request() 2906 bio->bi_opf = REQ_OP_WRITE; in raid1_sync_request() 2916 bio->bi_opf |= MD_FAILFAST; in raid1_sync_request() [all …]
|
H A D | dm.c | 633 clone->bi_opf &= ~REQ_DM_POLL_LIST; in alloc_tio() 900 (bio->bi_opf & REQ_POLLED)); in dm_handle_requeue() 907 if (bio->bi_opf & REQ_POLLED) { in dm_handle_requeue() 981 bio->bi_opf &= ~REQ_PREFLUSH; in __dm_io_complete() 1681 bio->bi_opf |= REQ_DM_POLL_LIST; in dm_queue_poll_io() 1839 if (bio->bi_opf & REQ_PREFLUSH) { in dm_split_and_process_bio() 1887 if (bio->bi_opf & REQ_NOWAIT) in dm_submit_bio() 1889 else if (bio->bi_opf & REQ_RAHEAD) in dm_submit_bio() 1923 if (!(bio->bi_opf & REQ_DM_POLL_LIST)) in dm_poll_bio() 1936 bio->bi_opf &= ~REQ_DM_POLL_LIST; in dm_poll_bio() [all …]
|
H A D | dm-zero.c | 42 if (bio->bi_opf & REQ_RAHEAD) { in zero_map()
|
H A D | raid10.c | 1134 if (bio->bi_opf & REQ_NOWAIT) { in regular_request_wait() 1269 mbio->bi_opf |= MD_FAILFAST; in raid10_write_one_disk() 2402 tbio->bi_opf |= MD_FAILFAST; in sync_request_write() 2797 wbio->bi_opf = REQ_OP_WRITE; in narrow_write_error() 3409 bio->bi_opf = REQ_OP_READ; in raid10_sync_request() 3455 bio->bi_opf = REQ_OP_WRITE; in raid10_sync_request() 3609 bio->bi_opf = REQ_OP_READ; in raid10_sync_request() 3611 bio->bi_opf |= MD_FAILFAST; in raid10_sync_request() 3630 bio->bi_opf = REQ_OP_WRITE; in raid10_sync_request() 3632 bio->bi_opf |= MD_FAILFAST; in raid10_sync_request() [all …]
|
H A D | dm-io.c | 507 if ((io_req->bi_opf & REQ_OP_MASK) == REQ_OP_READ) { in dp_init() 538 io_req->bi_opf, &dp, sync_error_bits, ioprio); in dm_io() 541 io_req->bi_opf, &dp, io_req->notify.fn, in dm_io()
|
H A D | dm-flakey.c | 22 (((bio)->bi_opf & (fc)->corrupt_bio_flags) == (fc)->corrupt_bio_flags) 359 (bio_data_dir(bio) == WRITE) ? 'w' : 'r', bio->bi_opf, in corrupt_bio_common() 429 bio_init(clone, fc->dev->bdev, bio->bi_inline_vecs, nr_iovecs, bio->bi_opf); in clone_bio()
|
/linux/include/linux/ |
H A D | dm-io.h | 62 blk_opf_t bi_opf; /* Request type and flags */ member
|
H A D | blk_types.h | 209 blk_opf_t bi_opf; /* bottom bits REQ_OP, top bits member 425 return bio->bi_opf & REQ_OP_MASK; in bio_op()
|
H A D | bio.h | 358 if (bio->bi_opf & REQ_INTEGRITY) in bio_integrity() 826 bio->bi_opf |= REQ_POLLED; 828 bio->bi_opf |= REQ_NOWAIT; in bio_clear_polled() 833 bio->bi_opf &= ~REQ_POLLED;
|
/linux/drivers/block/ |
H A D | brd.c | 249 if (unlikely(op_is_discard(bio->bi_opf))) { in brd_submit_bio() 264 bio->bi_opf, sector); in brd_submit_bio() 266 if (err == -ENOMEM && bio->bi_opf & REQ_NOWAIT) { in brd_submit_bio()
|
/linux/fs/xfs/ |
H A D | xfs_bio_io.c | 42 prev->bi_opf, GFP_KERNEL); in xfs_rw_bdev()
|
/linux/fs/bcachefs/ |
H A D | io_write.c | 423 n->bio.bi_opf = wbio->bio.bi_opf; in bch2_submit_wbio_replicas() 736 wbio->bio.bi_opf = src->bi_opf; in bch2_write_bio_alloc() 1076 dst->bi_opf = src->bi_opf; in bch2_write_extent() 1317 bio->bi_opf = op->wbio.bio.bi_opf; in bch2_nocow_write() 1327 bio->bi_opf |= REQ_OP_WRITE; in bch2_nocow_write() 1483 bio->bi_opf | in __bch2_write() [all...] |