/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/nvme/target/ |
H A D | loop.c | 155 if (blk_rq_nr_phys_segments(req)) { in nvme_loop_queue_rq() 158 blk_rq_nr_phys_segments(req), in nvme_loop_queue_rq()
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/nvme/target/ |
H A D | loop.c | 155 if (blk_rq_nr_phys_segments(req)) { in nvme_loop_queue_rq() 158 blk_rq_nr_phys_segments(req), in nvme_loop_queue_rq()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/nvme/target/ |
H A D | loop.c | 155 if (blk_rq_nr_phys_segments(req)) { in nvme_loop_queue_rq() 158 blk_rq_nr_phys_segments(req), in nvme_loop_queue_rq()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/nvme/host/ |
H A D | pci.c | 528 return (void **)(iod->sg + blk_rq_nr_phys_segments(req)); in nvme_pci_iod_list() 534 int nseg = blk_rq_nr_phys_segments(req); in nvme_pci_use_sgls() 849 if (blk_rq_nr_phys_segments(req) == 1) { in nvme_map_data() 868 sg_init_table(iod->sg, blk_rq_nr_phys_segments(req)); in nvme_map_data() 943 if (blk_rq_nr_phys_segments(req)) { in nvme_queue_rq() 973 if (blk_rq_nr_phys_segments(req)) in nvme_pci_complete_rq()
|
H A D | fc.c | 2572 if (!blk_rq_nr_phys_segments(rq)) in nvme_fc_map_data() 2577 blk_rq_nr_phys_segments(rq), freq->sg_table.sgl, in nvme_fc_map_data() 2583 WARN_ON(op->nents > blk_rq_nr_phys_segments(rq)); in nvme_fc_map_data() 2784 if (blk_rq_nr_phys_segments(rq)) { in nvme_fc_queue_rq()
|
H A D | rdma.c | 1284 if (!blk_rq_nr_phys_segments(rq)) in nvme_rdma_unmap_data() 1528 if (!blk_rq_nr_phys_segments(rq)) in nvme_rdma_map_data() 1533 blk_rq_nr_phys_segments(rq), req->data_sgl.sg_table.sgl, in nvme_rdma_map_data()
|
H A D | tcp.c | 2260 if (!blk_rq_nr_phys_segments(rq)) in nvme_tcp_map_data() 2289 req->data_len = blk_rq_nr_phys_segments(rq) ? in nvme_tcp_setup_cmd_pdu()
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/nvme/host/ |
H A D | pci.c | 528 return (void **)(iod->sg + blk_rq_nr_phys_segments(req)); in nvme_pci_iod_list() 534 int nseg = blk_rq_nr_phys_segments(req); in nvme_pci_use_sgls() 849 if (blk_rq_nr_phys_segments(req) == 1) { in nvme_map_data() 868 sg_init_table(iod->sg, blk_rq_nr_phys_segments(req)); in nvme_map_data() 943 if (blk_rq_nr_phys_segments(req)) { in nvme_queue_rq() 973 if (blk_rq_nr_phys_segments(req)) in nvme_pci_complete_rq()
|
H A D | fc.c | 2572 if (!blk_rq_nr_phys_segments(rq)) in nvme_fc_map_data() 2577 blk_rq_nr_phys_segments(rq), freq->sg_table.sgl, in nvme_fc_map_data() 2583 WARN_ON(op->nents > blk_rq_nr_phys_segments(rq)); in nvme_fc_map_data() 2784 if (blk_rq_nr_phys_segments(rq)) { in nvme_fc_queue_rq()
|
H A D | rdma.c | 1284 if (!blk_rq_nr_phys_segments(rq)) in nvme_rdma_unmap_data() 1528 if (!blk_rq_nr_phys_segments(rq)) in nvme_rdma_map_data() 1533 blk_rq_nr_phys_segments(rq), req->data_sgl.sg_table.sgl, in nvme_rdma_map_data()
|
H A D | tcp.c | 2260 if (!blk_rq_nr_phys_segments(rq)) in nvme_tcp_map_data() 2289 req->data_len = blk_rq_nr_phys_segments(rq) ? in nvme_tcp_setup_cmd_pdu()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/nvme/host/ |
H A D | pci.c | 528 return (void **)(iod->sg + blk_rq_nr_phys_segments(req)); in nvme_pci_iod_list() 534 int nseg = blk_rq_nr_phys_segments(req); in nvme_pci_use_sgls() 849 if (blk_rq_nr_phys_segments(req) == 1) { in nvme_map_data() 868 sg_init_table(iod->sg, blk_rq_nr_phys_segments(req)); in nvme_map_data() 943 if (blk_rq_nr_phys_segments(req)) { in nvme_queue_rq() 973 if (blk_rq_nr_phys_segments(req)) in nvme_pci_complete_rq()
|
H A D | fc.c | 2572 if (!blk_rq_nr_phys_segments(rq)) in nvme_fc_map_data() 2577 blk_rq_nr_phys_segments(rq), freq->sg_table.sgl, in nvme_fc_map_data() 2583 WARN_ON(op->nents > blk_rq_nr_phys_segments(rq)); in nvme_fc_map_data() 2784 if (blk_rq_nr_phys_segments(rq)) { in nvme_fc_queue_rq()
|
H A D | rdma.c | 1284 if (!blk_rq_nr_phys_segments(rq)) in nvme_rdma_unmap_data() 1528 if (!blk_rq_nr_phys_segments(rq)) in nvme_rdma_map_data() 1533 blk_rq_nr_phys_segments(rq), req->data_sgl.sg_table.sgl, in nvme_rdma_map_data()
|
H A D | tcp.c | 2260 if (!blk_rq_nr_phys_segments(rq)) in nvme_tcp_map_data() 2289 req->data_len = blk_rq_nr_phys_segments(rq) ? in nvme_tcp_setup_cmd_pdu()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/block/ |
H A D | blk-merge.c | 546 WARN_ON(nsegs > blk_rq_nr_phys_segments(rq)); in __blk_rq_map_sg()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/block/ |
H A D | blk-merge.c | 546 WARN_ON(nsegs > blk_rq_nr_phys_segments(rq)); in __blk_rq_map_sg()
|
/dports/multimedia/libv4l/linux-5.13-rc2/block/ |
H A D | blk-merge.c | 546 WARN_ON(nsegs > blk_rq_nr_phys_segments(rq)); in __blk_rq_map_sg()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/include/linux/ |
H A D | blkdev.h | 1184 static inline unsigned short blk_rq_nr_phys_segments(struct request *rq) in blk_rq_nr_phys_segments() function
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/arch/um/drivers/ |
H A D | ubd_kern.c | 1442 segs = blk_rq_nr_phys_segments(req); in ubd_submit_request()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/arch/um/drivers/ |
H A D | ubd_kern.c | 1442 segs = blk_rq_nr_phys_segments(req); in ubd_submit_request()
|
/dports/multimedia/libv4l/linux-5.13-rc2/include/linux/ |
H A D | blkdev.h | 1184 static inline unsigned short blk_rq_nr_phys_segments(struct request *rq) in blk_rq_nr_phys_segments() function
|
/dports/multimedia/libv4l/linux-5.13-rc2/arch/um/drivers/ |
H A D | ubd_kern.c | 1442 segs = blk_rq_nr_phys_segments(req); in ubd_submit_request()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/include/linux/ |
H A D | blkdev.h | 1184 static inline unsigned short blk_rq_nr_phys_segments(struct request *rq) in blk_rq_nr_phys_segments() function
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/block/rnbd/ |
H A D | rnbd-clt.c | 1156 blk_rq_nr_phys_segments(rq) ? : 1, in rnbd_queue_rq()
|