/linux/drivers/net/ethernet/intel/ice/ |
H A D | ice_vf_lib.c | 61 vf->vf_ops->free(vf); in ice_release_vf() 128 return vf->pf->vsi[vf->lan_vsi_idx]; in ice_get_vf_vsi() 216 vf->vf_ops->clear_mbx_register(vf); in ice_trigger_vf_reset() 244 vf->vf_ops->irq_close(vf); in ice_vf_pre_vsi_rebuild() 247 vf->vf_ops->clear_reset_trigger(vf); in ice_vf_pre_vsi_rebuild() 376 vf->min_tx_rate, vf->vf_id, err); in ice_vf_rebuild_host_tx_rate_cfg() 385 vf->max_tx_rate, vf->vf_id, err); in ice_vf_rebuild_host_tx_rate_cfg() 437 &vf->hw_lan_addr[0], vf->vf_id, in ice_vf_rebuild_host_mac_cfg() 557 vf->vf_ops->post_vsi_rebuild(vf); in ice_vf_post_vsi_rebuild() 1245 params.vf = vf; in ice_vf_ctrl_vsi_setup() [all …]
|
H A D | ice_sriov.c | 229 params.vf = vf; in ice_vf_vsi_setup() 592 vf->vf_ops->clear_reset_trigger(vf); in ice_start_vfs() 812 vf = kzalloc(sizeof(*vf), GFP_KERNEL); in ice_create_vf_entries() 1095 if (!vf) in ice_sriov_set_msix_vec_count() 1130 vf->vf_id, vf->num_msix, vf->num_vf_qs); in ice_sriov_set_msix_vec_count() 1140 vf->num_msix, vf->vf_id, prev_msix); in ice_sriov_set_msix_vec_count() 1303 if (!vf) in ice_vf_lan_overflow_event() 1330 if (!vf) in ice_set_vf_spoofchk() 1340 vf->lan_vsi_idx, vf->vf_id); in ice_set_vf_spoofchk() 1386 if (!vf) in ice_get_vf_cfg() [all …]
|
H A D | ice_virtchnl.h | 32 int (*get_ver_msg)(struct ice_vf *vf, u8 *msg); 33 int (*get_vf_res_msg)(struct ice_vf *vf, u8 *msg); 34 void (*reset_vf)(struct ice_vf *vf); 37 int (*cfg_qs_msg)(struct ice_vf *vf, u8 *msg); 38 int (*ena_qs_msg)(struct ice_vf *vf, u8 *msg); 39 int (*dis_qs_msg)(struct ice_vf *vf, u8 *msg); 47 int (*add_vlan_msg)(struct ice_vf *vf, u8 *msg); 49 int (*query_rxdid)(struct ice_vf *vf); 50 int (*get_rss_hena)(struct ice_vf *vf); 52 int (*ena_vlan_stripping)(struct ice_vf *vf); [all …]
|
H A D | ice_virtchnl.c | 302 pf = vf->pf; in ice_vc_send_msg_to_vf() 777 vf->vf_id); in ice_vc_handle_rss_cfg() 784 vf->vf_id); in ice_vc_handle_rss_cfg() 798 vf->vf_id); in ice_vc_handle_rss_cfg() 1073 vf->vf_id); in ice_vc_cfg_promiscuous_mode_msg() 1831 if (vf->pf_set_mac && !ice_is_vf_trusted(vf)) in ice_can_vf_change_mac() 3855 pf = vf->pf; in ice_vc_repr_add_mac() 3911 vf->vf_id); in ice_vc_repr_cfg_promiscuous_mode() 3977 pf = vf->pf; in ice_is_malicious_vf() 3988 vf->vf_id, vf->dev_lan_addr, status); in ice_is_malicious_vf() [all …]
|
H A D | ice_vf_lib.h | 58 void (*free)(struct ice_vf *vf); 59 void (*clear_reset_state)(struct ice_vf *vf); 64 void (*irq_close)(struct ice_vf *vf); 65 void (*post_vsi_rebuild)(struct ice_vf *vf); 146 return vf->port_vlan_info.vid; in ice_vf_get_port_vlan_id() 151 return vf->port_vlan_info.prio; in ice_vf_get_port_vlan_prio() 156 return (ice_vf_get_port_vlan_id(vf) || ice_vf_get_port_vlan_prio(vf)); in ice_vf_is_port_vlan_ena() 161 return vf->port_vlan_info.tpid; in ice_vf_get_port_vlan_tpid() 189 #define ice_for_each_vf(pf, bkt, vf) \ argument 206 #define ice_for_each_vf_rcu(pf, bkt, vf) \ argument [all …]
|
H A D | ice_virtchnl_fdir.c | 128 vf->vf_id); in ice_vf_start_ctrl_vsi() 135 vf->vf_id); in ice_vf_start_ctrl_vsi() 634 pf = vf->pf; in ice_vc_fdir_write_flow_prof() 744 vf->vf_id); in ice_vc_fdir_config_input_set() 776 vf->vf_id); in ice_vc_fdir_config_input_set() 1273 pf = vf->pf; in ice_vc_fdir_write_fltr() 1330 pf = vf->pf; in ice_vf_fdir_timer() 1360 struct ice_vf *vf = ctrl_vsi->vf; in ice_vc_fdir_irq_handler() local 1412 pf = vf->pf; in ice_vf_fdir_dump_info() 1799 pf = vf->pf; in ice_vc_add_fdir_fltr() [all …]
|
H A D | ice_vf_lib_private.h | 26 int ice_vf_reconfig_vsi(struct ice_vf *vf); 27 void ice_initialize_vf_entry(struct ice_vf *vf); 28 void ice_dis_vf_qs(struct ice_vf *vf); 29 int ice_check_vf_init(struct ice_vf *vf); 33 bool ice_is_vf_trusted(struct ice_vf *vf); 34 bool ice_vf_has_no_qs_ena(struct ice_vf *vf); 35 bool ice_is_vf_link_up(struct ice_vf *vf); 36 void ice_vf_ctrl_invalidate_vsi(struct ice_vf *vf); 37 void ice_vf_ctrl_vsi_release(struct ice_vf *vf); 40 void ice_vf_invalidate_vsi(struct ice_vf *vf); [all …]
|
/linux/drivers/vdpa/ifcvf/ |
H A D | ifcvf_main.c | 25 return vf->config_cb.callback(vf->config_cb.private); in ifcvf_config_changed() 90 devm_free_irq(&pdev->dev, vf->vqs_reused_irq, vf); in ifcvf_free_vqs_reused_irq() 116 devm_free_irq(&pdev->dev, vf->config_irq, vf); in ifcvf_free_config_irq() 202 vf->vring[0].msix_name, vf); in ifcvf_request_vqs_reused_irq() 235 vf->vring[0].msix_name, vf); in ifcvf_request_dev_irq() 299 vf->config_msix_name, vf); in ifcvf_request_config_irq() 705 vf = &ifcvf_mgmt_dev->vf; in ifcvf_vdpa_dev_add() 718 adapter->vf = vf; in ifcvf_vdpa_dev_add() 801 vf = &ifcvf_mgmt_dev->vf; in ifcvf_probe() 815 vf->hw_features = ifcvf_get_hw_features(vf); in ifcvf_probe() [all …]
|
/linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_vf.c | 41 dev_err(vf->dev, in otx2vf_process_vfaf_mbox_msg() 47 dev_err(vf->dev, in otx2vf_process_vfaf_mbox_msg() 54 dev_err(vf->dev, in otx2vf_process_vfaf_mbox_msg() 78 dev_err(vf->dev, in otx2vf_process_vfaf_mbox_msg() 209 queue_work(vf->mbox_wq, &vf->mbox.mbox_wrk); in otx2vf_vfaf_mbox_intr_handler() 226 queue_work(vf->mbox_wq, &vf->mbox.mbox_up_wrk); in otx2vf_vfaf_mbox_intr_handler() 417 queue_work(vf->otx2_wq, &vf->rx_mode_work); in otx2vf_set_rx_mode() 576 vf->dev = dev; in otx2vf_probe() 580 hw = &vf->hw; in otx2vf_probe() 727 qmem_free(vf->dev, vf->dync_lmt); in otx2vf_probe() [all …]
|
/linux/drivers/net/ethernet/sfc/siena/ |
H A D | siena_sriov.c | 1106 vf = nic_data->vf + pos; in efx_siena_sriov_peer_work() 1162 vf = nic_data->vf + pos; in efx_siena_sriov_peer_work() 1206 vf = nic_data->vf + index; in efx_siena_sriov_vf_alloc() 1230 vf = nic_data->vf + pos; in efx_siena_sriov_vfs_fini() 1387 vf = nic_data->vf + pos; in efx_siena_sriov_fini() 1467 vf = nic_data->vf + vf_i; in efx_siena_sriov_flr() 1571 vf = nic_data->vf + vf_i; in efx_siena_sriov_reset() 1602 vf = nic_data->vf + vf_i; in efx_siena_sriov_set_vf_mac() 1621 vf = nic_data->vf + vf_i; in efx_siena_sriov_set_vf_vlan() 1641 vf = nic_data->vf + vf_i; in efx_siena_sriov_set_vf_spoofchk() [all …]
|
/linux/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_virtchnl_pf.c | 26 struct i40e_vf *vf = pf->vf; in i40e_vc_vf_broadcast() local 195 if (!vf || vf->vf_id >= vf->pf->num_alloc_vfs) in i40e_vc_notify_vf_reset() 1639 for (vf = &pf->vf[0]; vf < &pf->vf[pf->num_alloc_vfs]; ++vf) { in i40e_reset_all_vfs() 1685 for (vf = &pf->vf[0]; vf < &pf->vf[pf->num_alloc_vfs]; ++vf) { in i40e_reset_all_vfs() 1700 for (vf = &pf->vf[0]; vf < &pf->vf[pf->num_alloc_vfs]; ++vf) { in i40e_reset_all_vfs() 1718 for (vf = &pf->vf[0]; vf < &pf->vf[pf->num_alloc_vfs]; ++vf) { in i40e_reset_all_vfs() 1983 if (!vf || vf->vf_id >= vf->pf->num_alloc_vfs) in i40e_vc_send_msg_to_vf() 4356 vf = &pf->vf[vf_id]; in i40e_validate_vf() 4425 vf = &pf->vf[vf_id]; in i40e_ndo_set_vf_mac() 4527 vf = &pf->vf[vf_id]; in i40e_ndo_set_vf_port_vlan() [all …]
|
/linux/drivers/net/ethernet/intel/ixgbe/ |
H A D | ixgbe_sriov.c | 129 int vf = 0; in ixgbe_get_vfs() local 148 ++vf; in ixgbe_get_vfs() 218 for (vf = 0; vf < num_vfs; ++vf) { in ixgbe_disable_sriov() 647 if (entry->vf == vf) { in ixgbe_set_vf_macvlan() 688 entry->vf = vf; in ixgbe_set_vf_macvlan() 961 vf); in ixgbe_set_vf_mac_addr() 1359 u32 vf; in ixgbe_msg_task() local 1362 for (vf = 0; vf < adapter->num_vfs; vf++) { in ixgbe_msg_task() 1508 adapter->vfinfo[vf].pf_vlan, vf); in ixgbe_disable_port_vlan() 1762 if (vf < 0 || vf >= adapter->num_vfs) { in ixgbe_ndo_set_vf_link_state() [all …]
|
/linux/drivers/net/ethernet/broadcom/bnx2x/ |
H A D | bnx2x_sriov.c | 268 bnx2x_vf_igu_ack_sb(bp, vf, vf_igu_sb(vf, bnx2x_vfq(vf, qid, sb_idx)), in bnx2x_vf_queue_create() 319 if (vf) { in bnx2x_vf_set_igu_info() 805 dev = pci_get_domain_bus_and_slot(vf->domain, vf->bus, vf->devfn); in bnx2x_vf_is_pcie_pending() 1620 vf->abs_vfid, vf->bus, vf->devfn, in bnx2x_iov_nic_init() 1921 vf->fw_stat_map + j * vf->stats_stride; in bnx2x_iov_adjust_stats_req() 2038 vf->abs_vfid, vf->state); in bnx2x_vf_acquire() 2060 vf_sb_count(vf), vf_rxq_count(vf), in bnx2x_vf_acquire() 2061 vf_txq_count(vf), vf_mac_rules_cnt(vf), in bnx2x_vf_acquire() 2102 vf_igu_sb(vf, i), vf_igu_sb(vf, i)); in bnx2x_vf_init() 2107 vf->abs_vfid, vf->state); in bnx2x_vf_init() [all …]
|
H A D | bnx2x_sriov.h | 166 #define vf_rxq_count(vf) ((vf)->alloc_resc.num_rxqs) argument 167 #define vf_txq_count(vf) ((vf)->alloc_resc.num_txqs) argument 168 #define vf_sb_count(vf) ((vf)->alloc_resc.num_sbs) argument 169 #define vf_mac_rules_cnt(vf) ((vf)->alloc_resc.num_mac_filters) argument 170 #define vf_vlan_rules_cnt(vf) ((vf)->alloc_resc.num_vlan_filters) argument 171 #define vf_mc_rules_cnt(vf) ((vf)->alloc_resc.num_mc_filters) argument 179 #define bnx2x_vfq(vf, nr, var) ((vf)->vfqs[(nr)].var) argument 180 #define bnx2x_leading_vfq(vf, var) ((vf)->vfqs[LEADING_IDX].var) argument 230 #define is_vf_multi(vf) (vf_rxq_count(vf) > 1) argument 343 + (vf)) [all …]
|
H A D | bnx2x_vfpf.c | 1174 vf->abs_vfid); in bnx2x_vf_mbx_resp_send_msg() 1196 vf->abs_vfid); in bnx2x_vf_mbx_resp_send_msg() 1299 vfq_qzone_id(vf, vfq_get(vf, i)); in bnx2x_vf_mbx_acquire_resp() 1316 vf->abs_vfid, in bnx2x_vf_mbx_acquire_resp() 1422 vf->abs_vfid); in bnx2x_vf_mbx_acquire() 1450 vf->abs_vfid); in bnx2x_vf_mbx_acquire() 1458 vf->abs_vfid); in bnx2x_vf_mbx_acquire() 1935 vf->abs_vfid, in bnx2x_vf_mbx_set_q_filters() 1991 vf->index); in bnx2x_vf_mbx_update_rss() 2171 vf->state); in bnx2x_vf_mbx_request() [all …]
|
/linux/drivers/crypto/cavium/cpt/ |
H A D | cptpf_mbox.c | 25 cpt_send_msg_to_vf(cpt, vf, mbx); in cpt_mbox_send_ack() 103 cpt_mbox_send_ack(cpt, vf, &mbx); in cpt_handle_mbox_intr() 107 mbx.data = vf; in cpt_handle_mbox_intr() 125 vf, mbx.data); in cpt_handle_mbox_intr() 128 vf, mbx.data); in cpt_handle_mbox_intr() 141 vf, mbx.msg); in cpt_handle_mbox_intr() 149 u8 vf; in cpt_mbox_intr_handler() local 153 for (vf = 0; vf < CPT_MAX_VF_NUM; vf++) { in cpt_mbox_intr_handler() 154 if (intr & (1ULL << vf)) { in cpt_mbox_intr_handler() 156 cpt_handle_mbox_intr(cpt, vf); in cpt_mbox_intr_handler() [all …]
|
/linux/drivers/net/ethernet/sfc/ |
H A D | ef10_sriov.c | 122 struct ef10_vf *vf = nic_data->vf + i; in efx_ef10_sriov_free_vf_vports() local 125 if (vf->pci_dev && pci_is_dev_assigned(vf->pci_dev)) in efx_ef10_sriov_free_vf_vports() 160 struct ef10_vf *vf = nic_data->vf + vf_i; in efx_ef10_sriov_assign_vf_vport() local 168 vf->vlan, &vf->vport_id); in efx_ef10_sriov_assign_vf_vport() 498 vf = nic_data->vf + vf_i; in efx_ef10_sriov_set_vf_mac() 504 vf->efx->type->filter_table_remove(vf->efx); in efx_ef10_sriov_set_vf_mac() 541 vf->efx->type->filter_table_probe(vf->efx); in efx_ef10_sriov_set_vf_mac() 566 vf = nic_data->vf + vf_i; in efx_ef10_sriov_set_vf_vlan() 577 vf->efx->type->filter_table_remove(vf->efx); in efx_ef10_sriov_set_vf_vlan() 617 vf->vlan, &vf->vport_id); in efx_ef10_sriov_set_vf_vlan() [all …]
|
/linux/drivers/crypto/marvell/octeontx/ |
H A D | otx_cptpf_mbox.c | 91 otx_cpt_send_msg_to_vf(cpt, vf, mbx); in otx_cpt_mbox_send_ack() 100 otx_cpt_send_msg_to_vf(cpt, vf, mbx); in otx_cptpf_mbox_send_nack() 190 dump_mbox_msg(&mbx, vf); in otx_cpt_handle_mbox_intr() 200 mbx.data = vf; in otx_cpt_handle_mbox_intr() 216 vf, mbx.data); in otx_cpt_handle_mbox_intr() 235 vf, mbx.msg); in otx_cpt_handle_mbox_intr() 243 u8 vf; in otx_cpt_mbox_intr_handler() local 247 for (vf = 0; vf < cpt->max_vfs; vf++) { in otx_cpt_mbox_intr_handler() 248 if (intr & (1ULL << vf)) { in otx_cpt_mbox_intr_handler() 249 otx_cpt_handle_mbox_intr(cpt, vf); in otx_cpt_mbox_intr_handler() [all …]
|
/linux/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_sriov_pf.c | 743 struct qlcnic_vf_info *vf = trans->vf; in qlcnic_sriov_pf_channel_cfg_cmd() local 887 struct qlcnic_vf_info *vf = tran->vf; in qlcnic_sriov_pf_create_rx_ctx_cmd() local 915 struct qlcnic_vf_info *vf = trans->vf; in qlcnic_sriov_pf_mac_address_cmd() local 947 struct qlcnic_vf_info *vf = trans->vf; in qlcnic_sriov_pf_create_tx_ctx_cmd() local 985 struct qlcnic_vf_info *vf = trans->vf; in qlcnic_sriov_pf_del_rx_ctx_cmd() local 1020 struct qlcnic_vf_info *vf = trans->vf; in qlcnic_sriov_pf_del_tx_ctx_cmd() local 1051 struct qlcnic_vf_info *vf = trans->vf; in qlcnic_sriov_pf_cfg_lro_cmd() local 1068 struct qlcnic_vf_info *vf = trans->vf; in qlcnic_sriov_pf_cfg_ip_cmd() local 1097 struct qlcnic_vf_info *vf = trans->vf; in qlcnic_sriov_pf_cfg_intrpt_cmd() local 1240 struct qlcnic_vf_info *vf = tran->vf; in qlcnic_sriov_pf_cfg_intrcoal_cmd() local [all …]
|
/linux/drivers/net/ethernet/cisco/enic/ |
H A D | enic_pp.c | 25 if (vf != PORT_SELF_VF) { in enic_is_valid_pp_vf() 28 if (vf < 0 || vf >= enic->num_vfs) { in enic_is_valid_pp_vf() 66 ENIC_PP_BY_INDEX(enic, vf, pp, &err); in enic_set_port_profile() 84 } else if (vf == PORT_SELF_VF) { in enic_set_port_profile() 88 "for VF %d\n", vf); in enic_set_port_profile() 138 if (vf == PORT_SELF_VF) in enic_unset_port_profile() 187 ENIC_PP_BY_INDEX(enic, vf, pp, &err); in enic_pp_disassociate() 209 ENIC_PP_BY_INDEX(enic, vf, pp, &err); in enic_pp_preassociate_rr() 249 ENIC_PP_BY_INDEX(enic, vf, pp, &err); in enic_pp_associate() 258 enic, vf, prev_pp, restore_pp); in enic_pp_associate() [all …]
|
/linux/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt_sriov.c | 38 if (vf) in bnxt_hwrm_fwd_async_event_cmpl() 85 vf = &bp->pf.vf[vf_id]; in bnxt_set_vf_spoofchk() 169 vf = &bp->pf.vf[vf_id]; in bnxt_set_vf_trust() 191 vf = &bp->pf.vf[vf_id]; in bnxt_get_vf_config() 233 vf = &bp->pf.vf[vf_id]; in bnxt_set_vf_mac() 272 vf = &bp->pf.vf[vf_id]; in bnxt_set_vf_vlan() 302 vf = &bp->pf.vf[vf_id]; in bnxt_set_vf_bw() 343 vf = &bp->pf.vf[vf_id]; in bnxt_set_vf_link_state() 373 vf = &bp->pf.vf[i]; in bnxt_set_vf_attr() 374 memset(vf, 0, sizeof(*vf)); in bnxt_set_vf_attr() [all …]
|
/linux/drivers/net/ethernet/netronome/nfp/ |
H A D | nfp_net_sriov.c | 32 if (vf < 0 || vf >= app->pf->num_vfs) { in nfp_net_sriov_check() 34 nfp_warn(app->pf->cpp, "invalid VF id %d\n", vf); in nfp_net_sriov_check() 48 writeb(vf, app->pf->vfcfg_tbl2 + NFP_NET_VF_CFG_MB_VF_NUM); in nfp_net_sriov_update() 77 mac, vf); in nfp_app_set_vf_mac() 82 vf_offset = NFP_NET_VF_CFG_MB_SZ + vf * NFP_NET_VF_CFG_SZ; in nfp_app_set_vf_mac() 91 mac, vf); in nfp_app_set_vf_mac() 115 "invalid vlan id or qos for VF id %d\n", vf); in nfp_app_set_vf_vlan() 144 return nfp_net_sriov_update(app, vf, update, "vlan"); in nfp_app_set_vf_vlan() 147 int nfp_app_set_vf_rate(struct net_device *netdev, int vf, in nfp_app_set_vf_rate() argument 270 err = nfp_net_sriov_check(app, vf, 0, "", true); in nfp_app_get_vf_config() [all …]
|
/linux/drivers/crypto/intel/qat/qat_common/ |
H A D | adf_pfvf_pf_msg.c | 15 struct adf_accel_vf_info *vf; in adf_pf2vf_notify_restarting() local 20 for (i = 0, vf = accel_dev->pf.vf_info; i < num_vfs; i++, vf++) { in adf_pf2vf_notify_restarting() 21 vf->restarting = false; in adf_pf2vf_notify_restarting() 22 if (!vf->init) in adf_pf2vf_notify_restarting() 28 vf->restarting = true; in adf_pf2vf_notify_restarting() 42 for (i = 0, vf = accel_dev->pf.vf_info; i < num_vfs; i++, vf++) in adf_pf2vf_wait_for_restarting_complete() 43 if (vf->restarting) in adf_pf2vf_wait_for_restarting_complete() 61 for (i = 0, vf = accel_dev->pf.vf_info; i < num_vfs; i++, vf++) { in adf_pf2vf_notify_restarted() 62 if (vf->init && vf->vf_compat_ver >= ADF_PFVF_COMPAT_FALLBACK && in adf_pf2vf_notify_restarted() 76 for (i = 0, vf = accel_dev->pf.vf_info; i < num_vfs; i++, vf++) { in adf_pf2vf_notify_fatal_error() [all …]
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | sriov.c | 75 int err, vf, num_msix_count; in mlx5_device_enable_sriov() local 86 for (vf = 0; vf < num_vfs; vf++) { in mlx5_device_enable_sriov() 102 vf, err); in mlx5_device_enable_sriov() 106 sriov->vfs_ctx[vf].enabled = 1; in mlx5_device_enable_sriov() 110 : vf + 1; in mlx5_device_enable_sriov() 115 vf, err); in mlx5_device_enable_sriov() 132 int vf; in mlx5_device_disable_sriov() local 134 for (vf = num_vfs - 1; vf >= 0; vf--) { in mlx5_device_disable_sriov() 135 if (!sriov->vfs_ctx[vf].enabled) in mlx5_device_disable_sriov() 147 sriov->vfs_ctx[vf].enabled = 0; in mlx5_device_disable_sriov() [all …]
|
/linux/drivers/infiniband/hw/mlx5/ |
H A D | ib_virt.c | 65 vf, err); in mlx5_ib_get_vf_config() 113 vfs_ctx[vf].policy = in->policy; in mlx5_ib_set_vf_link_state() 120 int mlx5_ib_get_vf_stats(struct ib_device *device, int vf, in mlx5_ib_get_vf_stats() argument 168 vfs_ctx[vf].node_guid = guid; in set_vf_node_guid() 169 vfs_ctx[vf].node_guid_valid = 1; in set_vf_node_guid() 192 vfs_ctx[vf].port_guid = guid; in set_vf_port_guid() 193 vfs_ctx[vf].port_guid_valid = 1; in set_vf_port_guid() 203 return set_vf_node_guid(device, vf, port, guid); in mlx5_ib_set_vf_guid() 205 return set_vf_port_guid(device, vf, port, guid); in mlx5_ib_set_vf_guid() 219 vfs_ctx[vf].node_guid_valid ? vfs_ctx[vf].node_guid : 0; in mlx5_ib_get_vf_guid() [all …]
|