Lines Matching refs:queue_depth
3625 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping() argument
3636 for (i = 0; i < queue_depth; i++) in blk_mq_clear_flush_rq_mapping()
3661 set->queue_depth, flush_rq); in blk_mq_exit_hctx()
3852 set->queue_depth); in __blk_mq_alloc_map_and_rqs()
4322 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue()
4355 set->queue_depth); in __blk_mq_alloc_rq_maps()
4390 depth = set->queue_depth; in blk_mq_alloc_set_map_and_rqs()
4396 set->queue_depth >>= 1; in blk_mq_alloc_set_map_and_rqs()
4397 if (set->queue_depth < set->reserved_tags + BLK_MQ_TAG_MIN) { in blk_mq_alloc_set_map_and_rqs()
4401 } while (set->queue_depth); in blk_mq_alloc_set_map_and_rqs()
4403 if (!set->queue_depth || err) { in blk_mq_alloc_set_map_and_rqs()
4408 if (depth != set->queue_depth) in blk_mq_alloc_set_map_and_rqs()
4410 depth, set->queue_depth); in blk_mq_alloc_set_map_and_rqs()
4500 if (!set->queue_depth) in blk_mq_alloc_tag_set()
4502 if (set->queue_depth < set->reserved_tags + BLK_MQ_TAG_MIN) in blk_mq_alloc_tag_set()
4511 if (set->queue_depth > BLK_MQ_MAX_DEPTH) { in blk_mq_alloc_tag_set()
4514 set->queue_depth = BLK_MQ_MAX_DEPTH; in blk_mq_alloc_tag_set()
4528 set->queue_depth = min(64U, set->queue_depth); in blk_mq_alloc_tag_set()
4592 const struct blk_mq_ops *ops, unsigned int queue_depth, in blk_mq_alloc_sq_tag_set() argument
4599 set->queue_depth = queue_depth; in blk_mq_alloc_sq_tag_set()