/linux/drivers/target/ |
H A D | target_core_tpg.c | 158 struct se_node_acl *acl, u32 queue_depth) in target_set_nacl_queue_depth() argument 160 acl->queue_depth = queue_depth; in target_set_nacl_queue_depth() 162 if (!acl->queue_depth) { in target_set_nacl_queue_depth() 166 acl->queue_depth = 1; in target_set_nacl_queue_depth() 174 u32 queue_depth; in target_alloc_node_acl() local 193 queue_depth = 1; in target_alloc_node_acl() 194 target_set_nacl_queue_depth(tpg, acl, queue_depth); in target_alloc_node_acl() 218 acl->queue_depth, in target_add_node_acl() 381 u32 queue_depth) in core_tpg_set_initiator_node_queue_depth() argument 390 if (acl->queue_depth == queue_depth) in core_tpg_set_initiator_node_queue_depth() [all …]
|
H A D | target_core_pscsi.c | 290 if (!sd->queue_depth) { in pscsi_add_device_to_list() 291 sd->queue_depth = PSCSI_DEFAULT_QUEUEDEPTH; in pscsi_add_device_to_list() 295 sd->lun, sd->queue_depth); in pscsi_add_device_to_list() 302 dev->dev_attrib.hw_queue_depth = sd->queue_depth; in pscsi_add_device_to_list()
|
/linux/arch/um/drivers/ |
H A D | vector_kern.c | 250 int queue_depth; in vector_advancehead() local 268 queue_depth = qi->queue_depth; in vector_advancehead() 270 return queue_depth; in vector_advancehead() 280 int queue_depth; in vector_advancetail() local 287 queue_depth = qi->queue_depth; in vector_advancetail() 289 return queue_depth; in vector_advancetail() 336 int queue_depth; in vector_enqueue() local 343 queue_depth = qi->queue_depth; in vector_enqueue() 367 return queue_depth; in vector_enqueue() 416 queue_depth = qi->queue_depth; in vector_send() [all …]
|
H A D | vector_kern.h | 47 int queue_depth, head, tail, max_depth, max_iov_frags; member
|
/linux/drivers/s390/block/ |
H A D | dasd_genhd.c | 23 static unsigned int queue_depth = 32; variable 26 module_param(queue_depth, uint, 0444); 27 MODULE_PARM_DESC(queue_depth, "Default queue depth for new DASD devices"); 59 block->tag_set.queue_depth = queue_depth; in dasd_gendisk_alloc()
|
/linux/drivers/target/iscsi/ |
H A D | iscsi_target_device.c | 41 sess->cmdsn_window = se_nacl->queue_depth; in iscsit_determine_maxcmdsn() 42 atomic_add(se_nacl->queue_depth - 1, &sess->max_cmd_sn); in iscsit_determine_maxcmdsn()
|
/linux/drivers/ata/ |
H A D | libata-sata.c | 1139 int queue_depth) in ata_change_queue_depth() argument 1148 if (!dev || queue_depth < 1 || queue_depth == sdev->queue_depth) { in ata_change_queue_depth() 1150 return sdev->queue_depth; in ata_change_queue_depth() 1159 if (queue_depth > max_queue_depth) { in ata_change_queue_depth() 1168 if (queue_depth == 1 || !ata_ncq_supported(dev)) { in ata_change_queue_depth() 1170 queue_depth = 1; in ata_change_queue_depth() 1177 if (queue_depth == sdev->queue_depth) in ata_change_queue_depth() 1178 return sdev->queue_depth; in ata_change_queue_depth() 1180 return scsi_change_queue_depth(sdev, queue_depth); in ata_change_queue_depth() 1199 int ata_scsi_change_queue_depth(struct scsi_device *sdev, int queue_depth) in ata_scsi_change_queue_depth() argument [all …]
|
/linux/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs-clt.c | 71 size_t max_depth = clt->queue_depth; in __rtrs_get_permit() 725 q_size = clt_path->queue_depth; in post_recv_path() 1694 if (WARN_ON(!clt_path->queue_depth)) in create_con_cq_qp() 1702 clt_path->queue_depth * 4 + 1); in create_con_cq_qp() 1704 clt_path->queue_depth * 3 + 1); in create_con_cq_qp() 1830 u16 version, queue_depth; in rtrs_rdma_conn_established() local 1857 queue_depth = le16_to_cpu(msg->queue_depth); in rtrs_rdma_conn_established() 1859 if (clt_path->queue_depth > 0 && queue_depth != clt_path->queue_depth) { in rtrs_rdma_conn_established() 1878 clt_path->queue_depth = queue_depth; in rtrs_rdma_conn_established() 1895 clt->queue_depth = clt_path->queue_depth; in rtrs_rdma_conn_established() [all …]
|
H A D | rtrs-srv.c | 109 for (i = 0; i < srv->queue_depth; i++) in rtrs_srv_free_ops_ids() 144 for (i = 0; i < srv->queue_depth; ++i) { in rtrs_srv_alloc_ops_ids() 585 mrs_num = srv->queue_depth; in map_cont_bufs() 607 srv->queue_depth - chunks); in map_cont_bufs() 989 q_size = srv->queue_depth; in post_recv_path() 1329 return srv->queue_depth; in rtrs_srv_get_queue_depth() 1364 for (i = 0; i < srv->queue_depth; i++) in free_srv() 1407 srv->queue_depth = sess_queue_depth; in get_or_create_srv() 1417 for (i = 0; i < srv->queue_depth; i++) { in get_or_create_srv() 1599 .queue_depth = cpu_to_le16(srv->queue_depth), in rtrs_rdma_do_accept() [all …]
|
H A D | rtrs-clt.h | 144 size_t queue_depth; member 173 size_t queue_depth; member
|
H A D | rtrs.h | 111 u32 queue_depth; member
|
H A D | rtrs-srv.h | 110 size_t queue_depth; member
|
/linux/include/net/mana/ |
H A D | hw_channel.h | 139 u16 queue_depth; member 153 u16 queue_depth; member
|
/linux/block/ |
H A D | blk-rq-qos.c | 122 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth() 138 rqd->queue_depth); in rq_depth_calc_max_depth() 142 unsigned int maxd = 3 * rqd->queue_depth / 4; in rq_depth_calc_max_depth()
|
H A D | blk-mq.c | 3645 set->queue_depth, flush_rq); in blk_mq_exit_hctx() 3836 set->queue_depth); in __blk_mq_alloc_map_and_rqs() 4339 set->queue_depth); in __blk_mq_alloc_rq_maps() 4374 depth = set->queue_depth; in blk_mq_alloc_set_map_and_rqs() 4380 set->queue_depth >>= 1; in blk_mq_alloc_set_map_and_rqs() 4385 } while (set->queue_depth); in blk_mq_alloc_set_map_and_rqs() 4392 if (depth != set->queue_depth) in blk_mq_alloc_set_map_and_rqs() 4394 depth, set->queue_depth); in blk_mq_alloc_set_map_and_rqs() 4484 if (!set->queue_depth) in blk_mq_alloc_tag_set() 4512 set->queue_depth = min(64U, set->queue_depth); in blk_mq_alloc_tag_set() [all …]
|
/linux/drivers/block/rnbd/ |
H A D | rnbd-srv.h | 25 int queue_depth; member
|
H A D | rnbd-clt.h | 86 size_t queue_depth; member
|
/linux/drivers/scsi/snic/ |
H A D | snic_main.c | 91 if (qsz < sdev->queue_depth) in snic_change_queue_depth() 93 else if (qsz > sdev->queue_depth) in snic_change_queue_depth() 96 atomic64_set(&snic->s_stats.misc.last_qsz, sdev->queue_depth); in snic_change_queue_depth() 100 return sdev->queue_depth; in snic_change_queue_depth()
|
/linux/drivers/block/ |
H A D | virtio_blk.c | 1244 module_param_named(queue_depth, virtblk_queue_depth, uint, 0444); 1453 unsigned int queue_depth; in virtblk_probe() local 1485 queue_depth = vblk->vqs[0].vq->num_free; in virtblk_probe() 1488 queue_depth /= 2; in virtblk_probe() 1490 queue_depth = virtblk_queue_depth; in virtblk_probe() 1495 vblk->tag_set.queue_depth = queue_depth; in virtblk_probe()
|
/linux/include/target/ |
H A D | target_core_base.h | 579 u32 queue_depth; member 723 u32 queue_depth; member 815 u32 queue_depth; member
|
/linux/drivers/scsi/ |
H A D | scsi.c | 222 sdev->queue_depth = depth; in scsi_change_queue_depth() 229 sbitmap_resize(&sdev->budget_map, sdev->queue_depth); in scsi_change_queue_depth() 231 return sdev->queue_depth; in scsi_change_queue_depth()
|
/linux/drivers/net/ethernet/microsoft/mana/ |
H A D | hw_channel.c | 318 comp_read = mana_gd_poll_cq(q_self, completions, hwc_cq->queue_depth); in mana_hwc_comp_event() 319 WARN_ON_ONCE(comp_read <= 0 || comp_read > hwc_cq->queue_depth); in mana_hwc_comp_event() 398 hwc_cq->queue_depth = q_depth; in mana_hwc_create_cq() 512 hwc_wq->queue_depth = q_depth; in mana_hwc_create_wq()
|
/linux/drivers/scsi/bfa/ |
H A D | bfad_im.c | 90 (bfa_lun_queue_depth > cmnd->device->queue_depth)) { in bfa_cb_ioim_done() 118 if (bfa_lun_queue_depth > cmnd->device->queue_depth) { in bfa_cb_ioim_good_comp() 878 if (bfa_lun_queue_depth > tmp_sdev->queue_depth) { in bfad_ramp_up_qdepth() 882 tmp_sdev->queue_depth + 1); in bfad_ramp_up_qdepth() 900 scsi_track_queue_full(tmp_sdev, tmp_sdev->queue_depth - 1); in bfad_handle_qfull()
|
/linux/include/uapi/linux/ |
H A D | ublk_cmd.h | 219 __u16 queue_depth; member
|
/linux/drivers/mmc/core/ |
H A D | queue.c | 436 mq->tag_set.queue_depth = in mmc_init_queue() 439 mq->tag_set.queue_depth = MMC_QUEUE_DEPTH; in mmc_init_queue()
|