/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/md/ |
H A D | raid0.c | 88 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones() 153 mddev->chunk_sectors << 9, blksize); in create_strip_zones() 354 ~(sector_t)(mddev->chunk_sectors-1)); in raid0_size() 366 if (mddev->chunk_sectors == 0) { in raid0_run() 468 mddev->chunk_sectors; in raid0_handle_discard() 470 mddev->chunk_sectors) + in raid0_handle_discard() 473 mddev->chunk_sectors; in raid0_handle_discard() 475 mddev->chunk_sectors) + in raid0_handle_discard() 484 mddev->chunk_sectors; in raid0_handle_discard() 531 chunk_sects = mddev->chunk_sectors; in raid0_make_request() [all …]
|
H A D | md-linear.c | 88 if (mddev->chunk_sectors) { in linear_conf() 90 sector_div(sectors, mddev->chunk_sectors); in linear_conf() 91 rdev->sectors = sectors * mddev->chunk_sectors; in linear_conf() 281 seq_printf(seq, " %dk rounding", mddev->chunk_sectors / 2); in linear_status()
|
H A D | dm-zoned-target.c | 999 unsigned int chunk_sectors = dmz_zone_nr_sectors(dmz->metadata); in dmz_io_hints() local 1009 limits->max_discard_sectors = chunk_sectors; in dmz_io_hints() 1010 limits->max_hw_discard_sectors = chunk_sectors; in dmz_io_hints() 1011 limits->max_write_zeroes_sectors = chunk_sectors; in dmz_io_hints() 1014 limits->chunk_sectors = chunk_sectors; in dmz_io_hints() 1015 limits->max_sectors = chunk_sectors; in dmz_io_hints()
|
H A D | raid5.c | 2928 : conf->chunk_sectors; in raid5_compute_sector() 3124 : conf->chunk_sectors; in raid5_compute_blocknr() 5311 unsigned int chunk_sectors; in in_chunk_boundary() local 5316 chunk_sectors = min(conf->chunk_sectors, conf->prev_chunk_sectors); in in_chunk_boundary() 5317 return chunk_sectors >= in in_chunk_boundary() 7026 max(conf->chunk_sectors, in alloc_scratch_buffer() 7482 int chunk_sectors; in raid5_run() local 7505 chunk_sectors = max(mddev->chunk_sectors, mddev->new_chunk_sectors); in raid5_run() 7536 here_old * chunk_sectors) in raid5_run() 8119 mddev->chunk_sectors) in check_reshape() [all …]
|
H A D | dm-raid.c | 716 mddev->new_chunk_sectors = mddev->chunk_sectors; in rs_set_cur() 729 mddev->chunk_sectors = mddev->new_chunk_sectors; in rs_set_new() 978 if (region_size < rs->md.chunk_sectors) { in validate_region_size() 1157 rs->md.new_chunk_sectors = rs->md.chunk_sectors = value; in parse_raid_params() 1487 if (rs->md.chunk_sectors) in parse_raid_params() 1488 max_io_len = rs->md.chunk_sectors; in parse_raid_params() 1877 rs->md.new_chunk_sectors != rs->md.chunk_sectors; in rs_is_layout_change() 2119 sb->stripe_sectors = cpu_to_le32(mddev->chunk_sectors); in super_sync() 2231 mddev->chunk_sectors = le32_to_cpu(sb->stripe_sectors); in super_init_validation() 2286 if (mddev->chunk_sectors != mddev->new_chunk_sectors) in super_init_validation() [all …]
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/md/ |
H A D | raid0.c | 88 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones() 153 mddev->chunk_sectors << 9, blksize); in create_strip_zones() 354 ~(sector_t)(mddev->chunk_sectors-1)); in raid0_size() 366 if (mddev->chunk_sectors == 0) { in raid0_run() 468 mddev->chunk_sectors; in raid0_handle_discard() 470 mddev->chunk_sectors) + in raid0_handle_discard() 473 mddev->chunk_sectors; in raid0_handle_discard() 475 mddev->chunk_sectors) + in raid0_handle_discard() 484 mddev->chunk_sectors; in raid0_handle_discard() 531 chunk_sects = mddev->chunk_sectors; in raid0_make_request() [all …]
|
H A D | md-linear.c | 88 if (mddev->chunk_sectors) { in linear_conf() 90 sector_div(sectors, mddev->chunk_sectors); in linear_conf() 91 rdev->sectors = sectors * mddev->chunk_sectors; in linear_conf() 281 seq_printf(seq, " %dk rounding", mddev->chunk_sectors / 2); in linear_status()
|
H A D | dm-zoned-target.c | 999 unsigned int chunk_sectors = dmz_zone_nr_sectors(dmz->metadata); in dmz_io_hints() local 1009 limits->max_discard_sectors = chunk_sectors; in dmz_io_hints() 1010 limits->max_hw_discard_sectors = chunk_sectors; in dmz_io_hints() 1011 limits->max_write_zeroes_sectors = chunk_sectors; in dmz_io_hints() 1014 limits->chunk_sectors = chunk_sectors; in dmz_io_hints() 1015 limits->max_sectors = chunk_sectors; in dmz_io_hints()
|
H A D | raid5.c | 2928 : conf->chunk_sectors; in raid5_compute_sector() 3124 : conf->chunk_sectors; in raid5_compute_blocknr() 5311 unsigned int chunk_sectors; in in_chunk_boundary() local 5316 chunk_sectors = min(conf->chunk_sectors, conf->prev_chunk_sectors); in in_chunk_boundary() 5317 return chunk_sectors >= in in_chunk_boundary() 7026 max(conf->chunk_sectors, in alloc_scratch_buffer() 7482 int chunk_sectors; in raid5_run() local 7505 chunk_sectors = max(mddev->chunk_sectors, mddev->new_chunk_sectors); in raid5_run() 7536 here_old * chunk_sectors) in raid5_run() 8119 mddev->chunk_sectors) in check_reshape() [all …]
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/md/ |
H A D | raid0.c | 88 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones() 153 mddev->chunk_sectors << 9, blksize); in create_strip_zones() 354 ~(sector_t)(mddev->chunk_sectors-1)); in raid0_size() 366 if (mddev->chunk_sectors == 0) { in raid0_run() 468 mddev->chunk_sectors; in raid0_handle_discard() 470 mddev->chunk_sectors) + in raid0_handle_discard() 473 mddev->chunk_sectors; in raid0_handle_discard() 475 mddev->chunk_sectors) + in raid0_handle_discard() 484 mddev->chunk_sectors; in raid0_handle_discard() 531 chunk_sects = mddev->chunk_sectors; in raid0_make_request() [all …]
|
H A D | md-linear.c | 88 if (mddev->chunk_sectors) { in linear_conf() 90 sector_div(sectors, mddev->chunk_sectors); in linear_conf() 91 rdev->sectors = sectors * mddev->chunk_sectors; in linear_conf() 281 seq_printf(seq, " %dk rounding", mddev->chunk_sectors / 2); in linear_status()
|
H A D | dm-zoned-target.c | 999 unsigned int chunk_sectors = dmz_zone_nr_sectors(dmz->metadata); in dmz_io_hints() local 1009 limits->max_discard_sectors = chunk_sectors; in dmz_io_hints() 1010 limits->max_hw_discard_sectors = chunk_sectors; in dmz_io_hints() 1011 limits->max_write_zeroes_sectors = chunk_sectors; in dmz_io_hints() 1014 limits->chunk_sectors = chunk_sectors; in dmz_io_hints() 1015 limits->max_sectors = chunk_sectors; in dmz_io_hints()
|
H A D | raid5.c | 2928 : conf->chunk_sectors; in raid5_compute_sector() 3124 : conf->chunk_sectors; in raid5_compute_blocknr() 5311 unsigned int chunk_sectors; in in_chunk_boundary() local 5316 chunk_sectors = min(conf->chunk_sectors, conf->prev_chunk_sectors); in in_chunk_boundary() 5317 return chunk_sectors >= in in_chunk_boundary() 7026 max(conf->chunk_sectors, in alloc_scratch_buffer() 7482 int chunk_sectors; in raid5_run() local 7505 chunk_sectors = max(mddev->chunk_sectors, mddev->new_chunk_sectors); in raid5_run() 7536 here_old * chunk_sectors) in raid5_run() 8119 mddev->chunk_sectors) in check_reshape() [all …]
|
H A D | dm-raid.c | 716 mddev->new_chunk_sectors = mddev->chunk_sectors; in rs_set_cur() 729 mddev->chunk_sectors = mddev->new_chunk_sectors; in rs_set_new() 978 if (region_size < rs->md.chunk_sectors) { in validate_region_size() 1157 rs->md.new_chunk_sectors = rs->md.chunk_sectors = value; in parse_raid_params() 1487 if (rs->md.chunk_sectors) in parse_raid_params() 1488 max_io_len = rs->md.chunk_sectors; in parse_raid_params() 1877 rs->md.new_chunk_sectors != rs->md.chunk_sectors; in rs_is_layout_change() 2119 sb->stripe_sectors = cpu_to_le32(mddev->chunk_sectors); in super_sync() 2231 mddev->chunk_sectors = le32_to_cpu(sb->stripe_sectors); in super_init_validation() 2286 if (mddev->chunk_sectors != mddev->new_chunk_sectors) in super_init_validation() [all …]
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/block/ |
H A D | blk-settings.c | 43 lim->chunk_sectors = 0; in blk_set_default_limits() 159 void blk_queue_chunk_sectors(struct request_queue *q, unsigned int chunk_sectors) in blk_queue_chunk_sectors() argument 161 q->limits.chunk_sectors = chunk_sectors; in blk_queue_chunk_sectors() 217 max_sectors = min(q->limits.chunk_sectors, max_sectors); in blk_queue_max_zone_append_sectors() 568 if (b->chunk_sectors) in blk_stack_limits() 569 t->chunk_sectors = gcd(t->chunk_sectors, b->chunk_sectors); in blk_stack_limits() 593 if ((t->chunk_sectors << 9) & (t->physical_block_size - 1)) { in blk_stack_limits() 594 t->chunk_sectors = 0; in blk_stack_limits()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/block/ |
H A D | blk-settings.c | 43 lim->chunk_sectors = 0; in blk_set_default_limits() 159 void blk_queue_chunk_sectors(struct request_queue *q, unsigned int chunk_sectors) in blk_queue_chunk_sectors() argument 161 q->limits.chunk_sectors = chunk_sectors; in blk_queue_chunk_sectors() 217 max_sectors = min(q->limits.chunk_sectors, max_sectors); in blk_queue_max_zone_append_sectors() 568 if (b->chunk_sectors) in blk_stack_limits() 569 t->chunk_sectors = gcd(t->chunk_sectors, b->chunk_sectors); in blk_stack_limits() 593 if ((t->chunk_sectors << 9) & (t->physical_block_size - 1)) { in blk_stack_limits() 594 t->chunk_sectors = 0; in blk_stack_limits()
|
/dports/multimedia/libv4l/linux-5.13-rc2/block/ |
H A D | blk-settings.c | 43 lim->chunk_sectors = 0; in blk_set_default_limits() 159 void blk_queue_chunk_sectors(struct request_queue *q, unsigned int chunk_sectors) in blk_queue_chunk_sectors() argument 161 q->limits.chunk_sectors = chunk_sectors; in blk_queue_chunk_sectors() 217 max_sectors = min(q->limits.chunk_sectors, max_sectors); in blk_queue_max_zone_append_sectors() 568 if (b->chunk_sectors) in blk_stack_limits() 569 t->chunk_sectors = gcd(t->chunk_sectors, b->chunk_sectors); in blk_stack_limits() 593 if ((t->chunk_sectors << 9) & (t->physical_block_size - 1)) { in blk_stack_limits() 594 t->chunk_sectors = 0; in blk_stack_limits()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/char/ |
H A D | ps3flash.c | 26 u64 chunk_sectors; member 38 start_sector, priv->chunk_sectors, in ps3flash_read_write_sectors() 118 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_read() 151 sector += priv->chunk_sectors; in ps3flash_read() 187 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_write() 226 sector += priv->chunk_sectors; in ps3flash_write() 376 priv->chunk_sectors = dev->bounce_size / dev->blk_size; in ps3flash_probe()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/char/ |
H A D | ps3flash.c | 26 u64 chunk_sectors; member 38 start_sector, priv->chunk_sectors, in ps3flash_read_write_sectors() 118 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_read() 151 sector += priv->chunk_sectors; in ps3flash_read() 187 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_write() 226 sector += priv->chunk_sectors; in ps3flash_write() 376 priv->chunk_sectors = dev->bounce_size / dev->blk_size; in ps3flash_probe()
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/char/ |
H A D | ps3flash.c | 26 u64 chunk_sectors; member 38 start_sector, priv->chunk_sectors, in ps3flash_read_write_sectors() 118 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_read() 151 sector += priv->chunk_sectors; in ps3flash_read() 187 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_write() 226 sector += priv->chunk_sectors; in ps3flash_write() 376 priv->chunk_sectors = dev->bounce_size / dev->blk_size; in ps3flash_probe()
|
/dports/devel/android-tools-fastboot/platform_system_core-platform-tools-29.0.5/fs_mgr/libsnapshot/ |
H A D | dm_snapshot_internals.h | 24 DmSnapCowSizeCalculator(unsigned int sector_bytes, unsigned int chunk_sectors) in DmSnapCowSizeCalculator() argument 26 chunk_sectors_(chunk_sectors), in DmSnapCowSizeCalculator()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/include/linux/ |
H A D | blkdev.h | 335 unsigned int chunk_sectors; member 744 return sector >> ilog2(q->limits.chunk_sectors); in blk_queue_zone_no() 1068 unsigned int chunk_sectors) in blk_max_size_offset() argument 1070 if (!chunk_sectors) { in blk_max_size_offset() 1071 if (q->limits.chunk_sectors) in blk_max_size_offset() 1072 chunk_sectors = q->limits.chunk_sectors; in blk_max_size_offset() 1077 if (likely(is_power_of_2(chunk_sectors))) in blk_max_size_offset() 1078 chunk_sectors -= offset & (chunk_sectors - 1); in blk_max_size_offset() 1080 chunk_sectors -= sector_div(offset, chunk_sectors); in blk_max_size_offset() 1082 return min(q->limits.max_sectors, chunk_sectors); in blk_max_size_offset() [all …]
|
/dports/multimedia/libv4l/linux-5.13-rc2/include/linux/ |
H A D | blkdev.h | 335 unsigned int chunk_sectors; member 744 return sector >> ilog2(q->limits.chunk_sectors); in blk_queue_zone_no() 1068 unsigned int chunk_sectors) in blk_max_size_offset() argument 1070 if (!chunk_sectors) { in blk_max_size_offset() 1071 if (q->limits.chunk_sectors) in blk_max_size_offset() 1072 chunk_sectors = q->limits.chunk_sectors; in blk_max_size_offset() 1077 if (likely(is_power_of_2(chunk_sectors))) in blk_max_size_offset() 1078 chunk_sectors -= offset & (chunk_sectors - 1); in blk_max_size_offset() 1080 chunk_sectors -= sector_div(offset, chunk_sectors); in blk_max_size_offset() 1082 return min(q->limits.max_sectors, chunk_sectors); in blk_max_size_offset() [all …]
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/include/linux/ |
H A D | blkdev.h | 335 unsigned int chunk_sectors; member 744 return sector >> ilog2(q->limits.chunk_sectors); in blk_queue_zone_no() 1068 unsigned int chunk_sectors) in blk_max_size_offset() argument 1070 if (!chunk_sectors) { in blk_max_size_offset() 1071 if (q->limits.chunk_sectors) in blk_max_size_offset() 1072 chunk_sectors = q->limits.chunk_sectors; in blk_max_size_offset() 1077 if (likely(is_power_of_2(chunk_sectors))) in blk_max_size_offset() 1078 chunk_sectors -= offset & (chunk_sectors - 1); in blk_max_size_offset() 1080 chunk_sectors -= sector_div(offset, chunk_sectors); in blk_max_size_offset() 1082 return min(q->limits.max_sectors, chunk_sectors); in blk_max_size_offset() [all …]
|
/dports/sysutils/wiimms/wiimms-iso-tools/src/ |
H A D | lib-wia.c | 142 wia->chunk_sectors = chunk_groups * WII_GROUP_SECTORS; in AllocBufferWIA() 162 wia->chunk_groups, wia->chunk_sectors ); in AllocBufferWIA() 780 hashtab0,wia->gdata,wia->chunk_sectors); in read_part_gdata() 866 const int base_group = sector / wia->chunk_sectors; in ReadWIA() 1617 wia->gdata,wia->gdata,hashtab0,wia->chunk_sectors); in write_part_data() 1921 const int base_group = sector / wia->chunk_sectors; in WriteWIA() 2121 const u32 n_grp = ( n_sect + wia->chunk_sectors - 1 ) / wia->chunk_sectors; in need_raw_data() 2471 pd->n_groups = ( pd->n_sectors + wia->chunk_sectors - 1 ) in SetupWriteWIA() 2472 / wia->chunk_sectors; in SetupWriteWIA() 2480 pd->n_groups = ( pd->n_sectors + wia->chunk_sectors - 1 ) in SetupWriteWIA() [all …]
|