/linux/drivers/interconnect/ |
H A D | bulk.c | 20 for (i = 0; i < num_paths; i++) { in of_icc_bulk_get() 48 while (--num_paths >= 0) { in icc_bulk_put() 49 icc_put(paths[num_paths].path); in icc_bulk_put() 50 paths[num_paths].path = NULL; in icc_bulk_put() 67 for (i = 0; i < num_paths; i++) { in icc_bulk_set_bw() 90 for (i = 0; i < num_paths; i++) { in icc_bulk_enable() 114 while (--num_paths >= 0) in icc_bulk_disable() 115 icc_disable(paths[num_paths].path); in icc_bulk_disable() 121 int num_paths; member 148 ret = of_icc_bulk_get(dev, num_paths, paths); in devm_of_icc_bulk_get() [all …]
|
/linux/include/linux/ |
H A D | interconnect.h | 45 int devm_of_icc_bulk_get(struct device *dev, int num_paths, struct icc_bulk_data *paths); 53 int __must_check of_icc_bulk_get(struct device *dev, int num_paths, 55 void icc_bulk_put(int num_paths, struct icc_bulk_data *paths); 56 int icc_bulk_set_bw(int num_paths, const struct icc_bulk_data *paths); 57 int icc_bulk_enable(int num_paths, const struct icc_bulk_data *paths); 58 void icc_bulk_disable(int num_paths, const struct icc_bulk_data *paths); 112 static inline int devm_of_icc_bulk_get(struct device *dev, int num_paths, in devm_of_icc_bulk_get() argument 118 static inline void icc_bulk_put(int num_paths, struct icc_bulk_data *paths) in icc_bulk_put() argument 122 static inline int icc_bulk_set_bw(int num_paths, const struct icc_bulk_data *paths) in icc_bulk_set_bw() argument 127 static inline int icc_bulk_enable(int num_paths, const struct icc_bulk_data *paths) in icc_bulk_enable() argument [all …]
|
H A D | netdevice.h | 872 int num_paths; member
|
/linux/samples/landlock/ |
H A D | sandboxer.c | 62 int i, num_paths = 0; in parse_path() local 65 num_paths++; in parse_path() 68 num_paths++; in parse_path() 71 *path_list = malloc(num_paths * sizeof(**path_list)); in parse_path() 72 for (i = 0; i < num_paths; i++) in parse_path() 75 return num_paths; in parse_path() 92 int num_paths, i, ret = 1; in populate_ruleset_fs() local 107 num_paths = parse_path(env_path_name, &path_list); in populate_ruleset_fs() 108 if (num_paths == 1 && path_list[0][0] == '\0') { in populate_ruleset_fs() 117 for (i = 0; i < num_paths; i++) { in populate_ruleset_fs()
|
/linux/drivers/pmdomain/imx/ |
H A D | imx8m-blk-ctrl.c | 45 int num_paths; member 69 int num_paths; member 230 domain->num_paths = data->num_paths; in imx8m_blk_ctrl_probe() 493 .num_paths = 1, 503 .num_paths = 1, 513 .num_paths = 1, 734 .num_paths = 2, 744 .num_paths = 3, 763 .num_paths = 2, 773 .num_paths = 2, [all …]
|
H A D | imx8mp-blk-ctrl.c | 54 int num_paths; member 68 int num_paths; member 262 .num_paths = 2, 278 .num_paths = 2, 464 .num_paths = 1, 502 .num_paths = 1, 510 .num_paths = 1, 554 ret = icc_bulk_set_bw(domain->num_paths, domain->paths); in imx8mp_blk_ctrl_power_on() 659 domain->num_paths = data->num_paths; in imx8mp_blk_ctrl_probe() 664 for (j = 0; j < data->num_paths; j++) { in imx8mp_blk_ctrl_probe() [all …]
|
/linux/drivers/gpu/drm/msm/disp/dpu1/ |
H A D | dpu_core_perf.c | 218 if (!kms->num_paths) in _dpu_core_perf_crtc_update_bus() 234 dpu_cstate->new_perf.bw_ctl, kms->num_paths); in _dpu_core_perf_crtc_update_bus() 239 do_div(avg_bw, (kms->num_paths * 1000)); /*Bps_to_icc*/ in _dpu_core_perf_crtc_update_bus() 241 for (i = 0; i < kms->num_paths; i++) in _dpu_core_perf_crtc_update_bus()
|
H A D | dpu_kms.h | 111 u32 num_paths; member
|
H A D | dpu_kms.c | 400 dpu_kms->num_paths = 1; in dpu_kms_parse_data_bus_icc_path() 404 dpu_kms->num_paths++; in dpu_kms_parse_data_bus_icc_path() 1354 for (i = 0; i < dpu_kms->num_paths; i++) in dpu_runtime_suspend()
|
/linux/include/uapi/rdma/ |
H A D | rdma_user_cm.h | 164 __u32 num_paths; member 185 __u32 num_paths; member
|
/linux/drivers/net/ethernet/brocade/bna/ |
H A D | bna_tx_rx.c | 1265 (qcfg)->num_paths : ((qcfg)->num_paths * 2)) 2170 mem_info->num = q_cfg->num_paths; in bna_rx_res_req() 2182 mem_info->num = q_cfg->num_paths; in bna_rx_res_req() 2188 mem_info->num = q_cfg->num_paths; in bna_rx_res_req() 2194 mem_info->num = q_cfg->num_paths; in bna_rx_res_req() 2200 mem_info->num = q_cfg->num_paths; in bna_rx_res_req() 2206 mem_info->num = q_cfg->num_paths; in bna_rx_res_req() 2212 mem_info->num = q_cfg->num_paths; in bna_rx_res_req() 2236 mem_info->num = q_cfg->num_paths; in bna_rx_res_req() 2338 rx->num_paths = rx_cfg->num_paths; in bna_rx_create() [all …]
|
H A D | bna_types.h | 666 int num_paths; member 797 int num_paths; member
|
H A D | bnad.c | 2028 rx_config->num_paths = bnad->num_rxp_per_rx; in bnad_init_rx_config() 2171 bnad_rx_msix_unregister(bnad, rx_info, rx_config->num_paths); in bnad_destroy_rx() 2219 rx_config->num_paths, in bnad_setup_rx() 2226 rx_config->num_paths, in bnad_setup_rx() 2261 rx_config->num_paths); in bnad_setup_rx()
|
/linux/Documentation/admin-guide/device-mapper/ |
H A D | switch.rst | 71 <num_paths> <region_size> <num_optional_args> [<optional_args>...] [<dev_path> <offset>]+ 72 <num_paths> 105 The path number in the range 0 ... (<num_paths> - 1).
|
/linux/drivers/opp/ |
H A D | of.c | 508 int ret, i, count, num_paths; in dev_pm_opp_of_find_icc_paths() local 535 num_paths = count / 2; in dev_pm_opp_of_find_icc_paths() 536 paths = kcalloc(num_paths, sizeof(*paths), GFP_KERNEL); in dev_pm_opp_of_find_icc_paths() 540 for (i = 0; i < num_paths; i++) { in dev_pm_opp_of_find_icc_paths() 550 opp_table->path_count = num_paths; in dev_pm_opp_of_find_icc_paths()
|
/linux/drivers/message/fusion/ |
H A D | mptscsih.c | 2181 int num_paths; in mptscsih_is_phys_disk() local 2200 num_paths = mpt_raid_phys_disk_get_num_paths(ioc, in mptscsih_is_phys_disk() 2202 if (num_paths < 2) in mptscsih_is_phys_disk() 2205 (num_paths * sizeof(RAID_PHYS_DISK1_PATH)), GFP_KERNEL); in mptscsih_is_phys_disk() 2214 for (j = 0; j < num_paths; j++) { in mptscsih_is_phys_disk() 2259 int num_paths; in mptscsih_raid_id_to_num() local 2278 num_paths = mpt_raid_phys_disk_get_num_paths(ioc, in mptscsih_raid_id_to_num() 2280 if (num_paths < 2) in mptscsih_raid_id_to_num() 2283 (num_paths * sizeof(RAID_PHYS_DISK1_PATH)), GFP_KERNEL); in mptscsih_raid_id_to_num() 2292 for (j = 0; j < num_paths; j++) { in mptscsih_raid_id_to_num()
|
H A D | mptsas.c | 4172 int num_paths; in mptsas_find_phyinfo_by_phys_disk_num() local 4180 num_paths = mpt_raid_phys_disk_get_num_paths(ioc, phys_disk_num); in mptsas_find_phyinfo_by_phys_disk_num() 4181 if (!num_paths) in mptsas_find_phyinfo_by_phys_disk_num() 4184 (num_paths * sizeof(RAID_PHYS_DISK1_PATH)), GFP_KERNEL); in mptsas_find_phyinfo_by_phys_disk_num() 4188 for (i = 0; i < num_paths; i++) { in mptsas_find_phyinfo_by_phys_disk_num()
|
/linux/net/netfilter/ |
H A D | nft_flow_offload.c | 103 for (i = 0; i < stack->num_paths; i++) { in nft_dev_path_info() 117 i = stack->num_paths; in nft_dev_path_info()
|
/linux/drivers/infiniband/core/ |
H A D | ucma.c | 756 resp->num_paths = route->num_pri_alt_paths; in ucma_copy_ib_route() 783 resp->num_paths = route->num_pri_alt_paths; in ucma_copy_iboe_route() 923 resp->num_paths = ctx->cm_id->route.num_pri_alt_paths; in ucma_query_path() 925 i < resp->num_paths && out_len > sizeof(struct ib_path_rec_data); in ucma_query_path()
|
H A D | sa_query.c | 129 unsigned int num_paths, void *context); 1523 unsigned int num_paths, void *context), in ib_sa_path_rec_get() argument
|
/linux/drivers/net/ethernet/mediatek/ |
H A D | mtk_ppe_offload.c | 106 path = &stack.path[stack.num_paths - 1]; in mtk_flow_get_wdma_info()
|
/linux/drivers/platform/x86/ |
H A D | thinkpad_acpi.c | 742 char **paths, const int num_paths) in drv_acpi_handle_init() argument 750 for (i = 0; i < num_paths; i++) { in drv_acpi_handle_init()
|
/linux/drivers/infiniband/ulp/srp/ |
H A D | ib_srp.c | 697 unsigned int num_paths, void *ch_ptr) in srp_path_rec_completion() argument
|
/linux/net/core/ |
H A D | dev.c | 704 int k = stack->num_paths++; in dev_fwd_path() 723 stack->num_paths = 0; in dev_fill_forward_path()
|