Home
last modified time | relevance | path

Searched refs:chunk_sectors (Results 1 – 25 of 118) sorted by relevance

12345

/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/md/
H A Draid0.c88 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones()
153 mddev->chunk_sectors << 9, blksize); in create_strip_zones()
354 ~(sector_t)(mddev->chunk_sectors-1)); in raid0_size()
366 if (mddev->chunk_sectors == 0) { in raid0_run()
468 mddev->chunk_sectors; in raid0_handle_discard()
470 mddev->chunk_sectors) + in raid0_handle_discard()
473 mddev->chunk_sectors; in raid0_handle_discard()
475 mddev->chunk_sectors) + in raid0_handle_discard()
484 mddev->chunk_sectors; in raid0_handle_discard()
531 chunk_sects = mddev->chunk_sectors; in raid0_make_request()
[all …]
H A Dmd-linear.c88 if (mddev->chunk_sectors) { in linear_conf()
90 sector_div(sectors, mddev->chunk_sectors); in linear_conf()
91 rdev->sectors = sectors * mddev->chunk_sectors; in linear_conf()
281 seq_printf(seq, " %dk rounding", mddev->chunk_sectors / 2); in linear_status()
H A Ddm-zoned-target.c999 unsigned int chunk_sectors = dmz_zone_nr_sectors(dmz->metadata); in dmz_io_hints() local
1009 limits->max_discard_sectors = chunk_sectors; in dmz_io_hints()
1010 limits->max_hw_discard_sectors = chunk_sectors; in dmz_io_hints()
1011 limits->max_write_zeroes_sectors = chunk_sectors; in dmz_io_hints()
1014 limits->chunk_sectors = chunk_sectors; in dmz_io_hints()
1015 limits->max_sectors = chunk_sectors; in dmz_io_hints()
H A Draid5.c2928 : conf->chunk_sectors; in raid5_compute_sector()
3124 : conf->chunk_sectors; in raid5_compute_blocknr()
5311 unsigned int chunk_sectors; in in_chunk_boundary() local
5316 chunk_sectors = min(conf->chunk_sectors, conf->prev_chunk_sectors); in in_chunk_boundary()
5317 return chunk_sectors >= in in_chunk_boundary()
7026 max(conf->chunk_sectors, in alloc_scratch_buffer()
7482 int chunk_sectors; in raid5_run() local
7505 chunk_sectors = max(mddev->chunk_sectors, mddev->new_chunk_sectors); in raid5_run()
7536 here_old * chunk_sectors) in raid5_run()
8119 mddev->chunk_sectors) in check_reshape()
[all …]
H A Ddm-raid.c716 mddev->new_chunk_sectors = mddev->chunk_sectors; in rs_set_cur()
729 mddev->chunk_sectors = mddev->new_chunk_sectors; in rs_set_new()
978 if (region_size < rs->md.chunk_sectors) { in validate_region_size()
1157 rs->md.new_chunk_sectors = rs->md.chunk_sectors = value; in parse_raid_params()
1487 if (rs->md.chunk_sectors) in parse_raid_params()
1488 max_io_len = rs->md.chunk_sectors; in parse_raid_params()
1877 rs->md.new_chunk_sectors != rs->md.chunk_sectors; in rs_is_layout_change()
2119 sb->stripe_sectors = cpu_to_le32(mddev->chunk_sectors); in super_sync()
2231 mddev->chunk_sectors = le32_to_cpu(sb->stripe_sectors); in super_init_validation()
2286 if (mddev->chunk_sectors != mddev->new_chunk_sectors) in super_init_validation()
[all …]
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/md/
H A Draid0.c88 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones()
153 mddev->chunk_sectors << 9, blksize); in create_strip_zones()
354 ~(sector_t)(mddev->chunk_sectors-1)); in raid0_size()
366 if (mddev->chunk_sectors == 0) { in raid0_run()
468 mddev->chunk_sectors; in raid0_handle_discard()
470 mddev->chunk_sectors) + in raid0_handle_discard()
473 mddev->chunk_sectors; in raid0_handle_discard()
475 mddev->chunk_sectors) + in raid0_handle_discard()
484 mddev->chunk_sectors; in raid0_handle_discard()
531 chunk_sects = mddev->chunk_sectors; in raid0_make_request()
[all …]
H A Dmd-linear.c88 if (mddev->chunk_sectors) { in linear_conf()
90 sector_div(sectors, mddev->chunk_sectors); in linear_conf()
91 rdev->sectors = sectors * mddev->chunk_sectors; in linear_conf()
281 seq_printf(seq, " %dk rounding", mddev->chunk_sectors / 2); in linear_status()
H A Ddm-zoned-target.c999 unsigned int chunk_sectors = dmz_zone_nr_sectors(dmz->metadata); in dmz_io_hints() local
1009 limits->max_discard_sectors = chunk_sectors; in dmz_io_hints()
1010 limits->max_hw_discard_sectors = chunk_sectors; in dmz_io_hints()
1011 limits->max_write_zeroes_sectors = chunk_sectors; in dmz_io_hints()
1014 limits->chunk_sectors = chunk_sectors; in dmz_io_hints()
1015 limits->max_sectors = chunk_sectors; in dmz_io_hints()
H A Draid5.c2928 : conf->chunk_sectors; in raid5_compute_sector()
3124 : conf->chunk_sectors; in raid5_compute_blocknr()
5311 unsigned int chunk_sectors; in in_chunk_boundary() local
5316 chunk_sectors = min(conf->chunk_sectors, conf->prev_chunk_sectors); in in_chunk_boundary()
5317 return chunk_sectors >= in in_chunk_boundary()
7026 max(conf->chunk_sectors, in alloc_scratch_buffer()
7482 int chunk_sectors; in raid5_run() local
7505 chunk_sectors = max(mddev->chunk_sectors, mddev->new_chunk_sectors); in raid5_run()
7536 here_old * chunk_sectors) in raid5_run()
8119 mddev->chunk_sectors) in check_reshape()
[all …]
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/md/
H A Draid0.c88 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones()
153 mddev->chunk_sectors << 9, blksize); in create_strip_zones()
354 ~(sector_t)(mddev->chunk_sectors-1)); in raid0_size()
366 if (mddev->chunk_sectors == 0) { in raid0_run()
468 mddev->chunk_sectors; in raid0_handle_discard()
470 mddev->chunk_sectors) + in raid0_handle_discard()
473 mddev->chunk_sectors; in raid0_handle_discard()
475 mddev->chunk_sectors) + in raid0_handle_discard()
484 mddev->chunk_sectors; in raid0_handle_discard()
531 chunk_sects = mddev->chunk_sectors; in raid0_make_request()
[all …]
H A Dmd-linear.c88 if (mddev->chunk_sectors) { in linear_conf()
90 sector_div(sectors, mddev->chunk_sectors); in linear_conf()
91 rdev->sectors = sectors * mddev->chunk_sectors; in linear_conf()
281 seq_printf(seq, " %dk rounding", mddev->chunk_sectors / 2); in linear_status()
H A Ddm-zoned-target.c999 unsigned int chunk_sectors = dmz_zone_nr_sectors(dmz->metadata); in dmz_io_hints() local
1009 limits->max_discard_sectors = chunk_sectors; in dmz_io_hints()
1010 limits->max_hw_discard_sectors = chunk_sectors; in dmz_io_hints()
1011 limits->max_write_zeroes_sectors = chunk_sectors; in dmz_io_hints()
1014 limits->chunk_sectors = chunk_sectors; in dmz_io_hints()
1015 limits->max_sectors = chunk_sectors; in dmz_io_hints()
H A Draid5.c2928 : conf->chunk_sectors; in raid5_compute_sector()
3124 : conf->chunk_sectors; in raid5_compute_blocknr()
5311 unsigned int chunk_sectors; in in_chunk_boundary() local
5316 chunk_sectors = min(conf->chunk_sectors, conf->prev_chunk_sectors); in in_chunk_boundary()
5317 return chunk_sectors >= in in_chunk_boundary()
7026 max(conf->chunk_sectors, in alloc_scratch_buffer()
7482 int chunk_sectors; in raid5_run() local
7505 chunk_sectors = max(mddev->chunk_sectors, mddev->new_chunk_sectors); in raid5_run()
7536 here_old * chunk_sectors) in raid5_run()
8119 mddev->chunk_sectors) in check_reshape()
[all …]
H A Ddm-raid.c716 mddev->new_chunk_sectors = mddev->chunk_sectors; in rs_set_cur()
729 mddev->chunk_sectors = mddev->new_chunk_sectors; in rs_set_new()
978 if (region_size < rs->md.chunk_sectors) { in validate_region_size()
1157 rs->md.new_chunk_sectors = rs->md.chunk_sectors = value; in parse_raid_params()
1487 if (rs->md.chunk_sectors) in parse_raid_params()
1488 max_io_len = rs->md.chunk_sectors; in parse_raid_params()
1877 rs->md.new_chunk_sectors != rs->md.chunk_sectors; in rs_is_layout_change()
2119 sb->stripe_sectors = cpu_to_le32(mddev->chunk_sectors); in super_sync()
2231 mddev->chunk_sectors = le32_to_cpu(sb->stripe_sectors); in super_init_validation()
2286 if (mddev->chunk_sectors != mddev->new_chunk_sectors) in super_init_validation()
[all …]
/dports/multimedia/v4l-utils/linux-5.13-rc2/block/
H A Dblk-settings.c43 lim->chunk_sectors = 0; in blk_set_default_limits()
159 void blk_queue_chunk_sectors(struct request_queue *q, unsigned int chunk_sectors) in blk_queue_chunk_sectors() argument
161 q->limits.chunk_sectors = chunk_sectors; in blk_queue_chunk_sectors()
217 max_sectors = min(q->limits.chunk_sectors, max_sectors); in blk_queue_max_zone_append_sectors()
568 if (b->chunk_sectors) in blk_stack_limits()
569 t->chunk_sectors = gcd(t->chunk_sectors, b->chunk_sectors); in blk_stack_limits()
593 if ((t->chunk_sectors << 9) & (t->physical_block_size - 1)) { in blk_stack_limits()
594 t->chunk_sectors = 0; in blk_stack_limits()
/dports/multimedia/v4l_compat/linux-5.13-rc2/block/
H A Dblk-settings.c43 lim->chunk_sectors = 0; in blk_set_default_limits()
159 void blk_queue_chunk_sectors(struct request_queue *q, unsigned int chunk_sectors) in blk_queue_chunk_sectors() argument
161 q->limits.chunk_sectors = chunk_sectors; in blk_queue_chunk_sectors()
217 max_sectors = min(q->limits.chunk_sectors, max_sectors); in blk_queue_max_zone_append_sectors()
568 if (b->chunk_sectors) in blk_stack_limits()
569 t->chunk_sectors = gcd(t->chunk_sectors, b->chunk_sectors); in blk_stack_limits()
593 if ((t->chunk_sectors << 9) & (t->physical_block_size - 1)) { in blk_stack_limits()
594 t->chunk_sectors = 0; in blk_stack_limits()
/dports/multimedia/libv4l/linux-5.13-rc2/block/
H A Dblk-settings.c43 lim->chunk_sectors = 0; in blk_set_default_limits()
159 void blk_queue_chunk_sectors(struct request_queue *q, unsigned int chunk_sectors) in blk_queue_chunk_sectors() argument
161 q->limits.chunk_sectors = chunk_sectors; in blk_queue_chunk_sectors()
217 max_sectors = min(q->limits.chunk_sectors, max_sectors); in blk_queue_max_zone_append_sectors()
568 if (b->chunk_sectors) in blk_stack_limits()
569 t->chunk_sectors = gcd(t->chunk_sectors, b->chunk_sectors); in blk_stack_limits()
593 if ((t->chunk_sectors << 9) & (t->physical_block_size - 1)) { in blk_stack_limits()
594 t->chunk_sectors = 0; in blk_stack_limits()
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/char/
H A Dps3flash.c26 u64 chunk_sectors; member
38 start_sector, priv->chunk_sectors, in ps3flash_read_write_sectors()
118 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_read()
151 sector += priv->chunk_sectors; in ps3flash_read()
187 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_write()
226 sector += priv->chunk_sectors; in ps3flash_write()
376 priv->chunk_sectors = dev->bounce_size / dev->blk_size; in ps3flash_probe()
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/char/
H A Dps3flash.c26 u64 chunk_sectors; member
38 start_sector, priv->chunk_sectors, in ps3flash_read_write_sectors()
118 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_read()
151 sector += priv->chunk_sectors; in ps3flash_read()
187 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_write()
226 sector += priv->chunk_sectors; in ps3flash_write()
376 priv->chunk_sectors = dev->bounce_size / dev->blk_size; in ps3flash_probe()
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/char/
H A Dps3flash.c26 u64 chunk_sectors; member
38 start_sector, priv->chunk_sectors, in ps3flash_read_write_sectors()
118 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_read()
151 sector += priv->chunk_sectors; in ps3flash_read()
187 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_write()
226 sector += priv->chunk_sectors; in ps3flash_write()
376 priv->chunk_sectors = dev->bounce_size / dev->blk_size; in ps3flash_probe()
/dports/devel/android-tools-fastboot/platform_system_core-platform-tools-29.0.5/fs_mgr/libsnapshot/
H A Ddm_snapshot_internals.h24 DmSnapCowSizeCalculator(unsigned int sector_bytes, unsigned int chunk_sectors) in DmSnapCowSizeCalculator() argument
26 chunk_sectors_(chunk_sectors), in DmSnapCowSizeCalculator()
/dports/multimedia/v4l_compat/linux-5.13-rc2/include/linux/
H A Dblkdev.h335 unsigned int chunk_sectors; member
744 return sector >> ilog2(q->limits.chunk_sectors); in blk_queue_zone_no()
1068 unsigned int chunk_sectors) in blk_max_size_offset() argument
1070 if (!chunk_sectors) { in blk_max_size_offset()
1071 if (q->limits.chunk_sectors) in blk_max_size_offset()
1072 chunk_sectors = q->limits.chunk_sectors; in blk_max_size_offset()
1077 if (likely(is_power_of_2(chunk_sectors))) in blk_max_size_offset()
1078 chunk_sectors -= offset & (chunk_sectors - 1); in blk_max_size_offset()
1080 chunk_sectors -= sector_div(offset, chunk_sectors); in blk_max_size_offset()
1082 return min(q->limits.max_sectors, chunk_sectors); in blk_max_size_offset()
[all …]
/dports/multimedia/libv4l/linux-5.13-rc2/include/linux/
H A Dblkdev.h335 unsigned int chunk_sectors; member
744 return sector >> ilog2(q->limits.chunk_sectors); in blk_queue_zone_no()
1068 unsigned int chunk_sectors) in blk_max_size_offset() argument
1070 if (!chunk_sectors) { in blk_max_size_offset()
1071 if (q->limits.chunk_sectors) in blk_max_size_offset()
1072 chunk_sectors = q->limits.chunk_sectors; in blk_max_size_offset()
1077 if (likely(is_power_of_2(chunk_sectors))) in blk_max_size_offset()
1078 chunk_sectors -= offset & (chunk_sectors - 1); in blk_max_size_offset()
1080 chunk_sectors -= sector_div(offset, chunk_sectors); in blk_max_size_offset()
1082 return min(q->limits.max_sectors, chunk_sectors); in blk_max_size_offset()
[all …]
/dports/multimedia/v4l-utils/linux-5.13-rc2/include/linux/
H A Dblkdev.h335 unsigned int chunk_sectors; member
744 return sector >> ilog2(q->limits.chunk_sectors); in blk_queue_zone_no()
1068 unsigned int chunk_sectors) in blk_max_size_offset() argument
1070 if (!chunk_sectors) { in blk_max_size_offset()
1071 if (q->limits.chunk_sectors) in blk_max_size_offset()
1072 chunk_sectors = q->limits.chunk_sectors; in blk_max_size_offset()
1077 if (likely(is_power_of_2(chunk_sectors))) in blk_max_size_offset()
1078 chunk_sectors -= offset & (chunk_sectors - 1); in blk_max_size_offset()
1080 chunk_sectors -= sector_div(offset, chunk_sectors); in blk_max_size_offset()
1082 return min(q->limits.max_sectors, chunk_sectors); in blk_max_size_offset()
[all …]
/dports/sysutils/wiimms/wiimms-iso-tools/src/
H A Dlib-wia.c142 wia->chunk_sectors = chunk_groups * WII_GROUP_SECTORS; in AllocBufferWIA()
162 wia->chunk_groups, wia->chunk_sectors ); in AllocBufferWIA()
780 hashtab0,wia->gdata,wia->chunk_sectors); in read_part_gdata()
866 const int base_group = sector / wia->chunk_sectors; in ReadWIA()
1617 wia->gdata,wia->gdata,hashtab0,wia->chunk_sectors); in write_part_data()
1921 const int base_group = sector / wia->chunk_sectors; in WriteWIA()
2121 const u32 n_grp = ( n_sect + wia->chunk_sectors - 1 ) / wia->chunk_sectors; in need_raw_data()
2471 pd->n_groups = ( pd->n_sectors + wia->chunk_sectors - 1 ) in SetupWriteWIA()
2472 / wia->chunk_sectors; in SetupWriteWIA()
2480 pd->n_groups = ( pd->n_sectors + wia->chunk_sectors - 1 ) in SetupWriteWIA()
[all …]

12345