1 /* 2 * Copyright (c) 2017-2018 Cavium, Inc. 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 9 * 1. Redistributions of source code must retain the above copyright 10 * notice, this list of conditions and the following disclaimer. 11 * 2. Redistributions in binary form must reproduce the above copyright 12 * notice, this list of conditions and the following disclaimer in the 13 * documentation and/or other materials provided with the distribution. 14 * 15 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" 16 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 17 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 18 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE 19 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 20 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 21 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 22 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 23 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 24 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 25 * POSSIBILITY OF SUCH DAMAGE. 26 * 27 * $FreeBSD$ 28 * 29 */ 30 31 #ifndef __ECORE_SRIOV_API_H__ 32 #define __ECORE_SRIOV_API_H__ 33 34 #include "common_hsi.h" 35 #include "ecore_status.h" 36 37 #define ECORE_ETH_VF_NUM_MAC_FILTERS 1 38 #define ECORE_ETH_VF_NUM_VLAN_FILTERS 2 39 #define ECORE_VF_ARRAY_LENGTH (3) 40 41 #define IS_VF(p_dev) ((p_dev)->b_is_vf) 42 #define IS_PF(p_dev) (!((p_dev)->b_is_vf)) 43 #ifdef CONFIG_ECORE_SRIOV 44 #define IS_PF_SRIOV(p_hwfn) (!!((p_hwfn)->p_dev->p_iov_info)) 45 #else 46 #define IS_PF_SRIOV(p_hwfn) (0) 47 #endif 48 #define IS_PF_SRIOV_ALLOC(p_hwfn) (!!((p_hwfn)->pf_iov_info)) 49 #define IS_PF_PDA(p_hwfn) 0 /* @@TBD Michalk */ 50 51 /* @@@ TBD MichalK - what should this number be*/ 52 #define ECORE_MAX_VF_CHAINS_PER_PF 16 53 54 /* vport update extended feature tlvs flags */ 55 enum ecore_iov_vport_update_flag { 56 ECORE_IOV_VP_UPDATE_ACTIVATE = 0, 57 ECORE_IOV_VP_UPDATE_VLAN_STRIP = 1, 58 ECORE_IOV_VP_UPDATE_TX_SWITCH = 2, 59 ECORE_IOV_VP_UPDATE_MCAST = 3, 60 ECORE_IOV_VP_UPDATE_ACCEPT_PARAM = 4, 61 ECORE_IOV_VP_UPDATE_RSS = 5, 62 ECORE_IOV_VP_UPDATE_ACCEPT_ANY_VLAN = 6, 63 ECORE_IOV_VP_UPDATE_SGE_TPA = 7, 64 ECORE_IOV_VP_UPDATE_MAX = 8, 65 }; 66 67 /*PF to VF STATUS is part of vfpf-channel API 68 and must be forward compatible */ 69 enum ecore_iov_pf_to_vf_status { 70 PFVF_STATUS_WAITING = 0, 71 PFVF_STATUS_SUCCESS, 72 PFVF_STATUS_FAILURE, 73 PFVF_STATUS_NOT_SUPPORTED, 74 PFVF_STATUS_NO_RESOURCE, 75 PFVF_STATUS_FORCED, 76 PFVF_STATUS_MALICIOUS, 77 }; 78 79 struct ecore_mcp_link_params; 80 struct ecore_mcp_link_state; 81 struct ecore_mcp_link_capabilities; 82 83 /* These defines are used by the hw-channel; should never change order */ 84 #define VFPF_ACQUIRE_OS_LINUX (0) 85 #define VFPF_ACQUIRE_OS_WINDOWS (1) 86 #define VFPF_ACQUIRE_OS_ESX (2) 87 #define VFPF_ACQUIRE_OS_SOLARIS (3) 88 #define VFPF_ACQUIRE_OS_LINUX_USERSPACE (4) 89 #define VFPF_ACQUIRE_OS_FREEBSD (5) 90 91 struct ecore_vf_acquire_sw_info { 92 u32 driver_version; 93 u8 os_type; 94 }; 95 96 struct ecore_public_vf_info { 97 /* These copies will later be reflected in the bulletin board, 98 * but this copy should be newer. 99 */ 100 u8 forced_mac[ETH_ALEN]; 101 u16 forced_vlan; 102 }; 103 104 struct ecore_iov_vf_init_params { 105 u16 rel_vf_id; 106 107 /* Number of requested Queues; Currently, don't support different 108 * number of Rx/Tx queues. 109 */ 110 /* TODO - remove this limitation */ 111 u16 num_queues; 112 113 /* Allow the client to choose which qzones to use for Rx/Tx, 114 * and which queue_base to use for Tx queues on a per-queue basis. 115 * Notice values should be relative to the PF resources. 116 */ 117 u16 req_rx_queue[ECORE_MAX_VF_CHAINS_PER_PF]; 118 u16 req_tx_queue[ECORE_MAX_VF_CHAINS_PER_PF]; 119 120 u8 vport_id; 121 122 /* Should be set in case RSS is going to be used for VF */ 123 u8 rss_eng_id; 124 }; 125 126 #ifdef CONFIG_ECORE_SW_CHANNEL 127 /* This is SW channel related only... */ 128 enum mbx_state { 129 VF_PF_UNKNOWN_STATE = 0, 130 VF_PF_WAIT_FOR_START_REQUEST = 1, 131 VF_PF_WAIT_FOR_NEXT_CHUNK_OF_REQUEST = 2, 132 VF_PF_REQUEST_IN_PROCESSING = 3, 133 VF_PF_RESPONSE_READY = 4, 134 }; 135 136 struct ecore_iov_sw_mbx { 137 enum mbx_state mbx_state; 138 139 u32 request_size; 140 u32 request_offset; 141 142 u32 response_size; 143 u32 response_offset; 144 }; 145 146 /** 147 * @brief Get the vf sw mailbox params 148 * 149 * @param p_hwfn 150 * @param rel_vf_id 151 * 152 * @return struct ecore_iov_sw_mbx* 153 */ 154 struct ecore_iov_sw_mbx* 155 ecore_iov_get_vf_sw_mbx(struct ecore_hwfn *p_hwfn, 156 u16 rel_vf_id); 157 #endif 158 159 /* This struct is part of ecore_dev and contains data relevant to all hwfns; 160 * Initialized only if SR-IOV cpabability is exposed in PCIe config space. 161 */ 162 struct ecore_hw_sriov_info { 163 /* standard SRIOV capability fields, mostly for debugging */ 164 int pos; /* capability position */ 165 int nres; /* number of resources */ 166 u32 cap; /* SR-IOV Capabilities */ 167 u16 ctrl; /* SR-IOV Control */ 168 u16 total_vfs; /* total VFs associated with the PF */ 169 u16 num_vfs; /* number of vfs that have been started */ 170 u16 initial_vfs; /* initial VFs associated with the PF */ 171 u16 nr_virtfn; /* number of VFs available */ 172 u16 offset; /* first VF Routing ID offset */ 173 u16 stride; /* following VF stride */ 174 u16 vf_device_id; /* VF device id */ 175 u32 pgsz; /* page size for BAR alignment */ 176 u8 link; /* Function Dependency Link */ 177 178 u32 first_vf_in_pf; 179 }; 180 181 #ifdef CONFIG_ECORE_SRIOV 182 #ifndef LINUX_REMOVE 183 /** 184 * @brief mark/clear all VFs before/after an incoming PCIe sriov 185 * disable. 186 * 187 * @param p_dev 188 * @param to_disable 189 */ 190 void ecore_iov_set_vfs_to_disable(struct ecore_dev *p_dev, 191 u8 to_disable); 192 193 /** 194 * @brief mark/clear chosen VF before/after an incoming PCIe 195 * sriov disable. 196 * 197 * @param p_dev 198 * @param rel_vf_id 199 * @param to_disable 200 */ 201 void ecore_iov_set_vf_to_disable(struct ecore_dev *p_dev, 202 u16 rel_vf_id, 203 u8 to_disable); 204 205 /** 206 * @brief ecore_iov_init_hw_for_vf - initialize the HW for 207 * enabling access of a VF. Also includes preparing the 208 * IGU for VF access. This needs to be called AFTER hw is 209 * initialized and BEFORE VF is loaded inside the VM. 210 * 211 * @param p_hwfn 212 * @param p_ptt 213 * @param p_params 214 * 215 * @return enum _ecore_status_t 216 */ 217 enum _ecore_status_t 218 ecore_iov_init_hw_for_vf(struct ecore_hwfn *p_hwfn, 219 struct ecore_ptt *p_ptt, 220 struct ecore_iov_vf_init_params *p_params); 221 222 /** 223 * @brief ecore_iov_process_mbx_req - process a request received 224 * from the VF 225 * 226 * @param p_hwfn 227 * @param p_ptt 228 * @param vfid 229 */ 230 void ecore_iov_process_mbx_req(struct ecore_hwfn *p_hwfn, 231 struct ecore_ptt *p_ptt, 232 int vfid); 233 234 /** 235 * @brief ecore_iov_release_hw_for_vf - called once upper layer 236 * knows VF is done with - can release any resources 237 * allocated for VF at this point. this must be done once 238 * we know VF is no longer loaded in VM. 239 * 240 * @param p_hwfn 241 * @param p_ptt 242 * @param rel_vf_id 243 * 244 * @return enum _ecore_status_t 245 */ 246 enum _ecore_status_t ecore_iov_release_hw_for_vf(struct ecore_hwfn *p_hwfn, 247 struct ecore_ptt *p_ptt, 248 u16 rel_vf_id); 249 250 /** 251 * @brief ecore_iov_set_vf_ctx - set a context for a given VF 252 * 253 * @param p_hwfn 254 * @param vf_id 255 * @param ctx 256 * 257 * @return enum _ecore_status_t 258 */ 259 enum _ecore_status_t ecore_iov_set_vf_ctx(struct ecore_hwfn *p_hwfn, 260 u16 vf_id, 261 void *ctx); 262 263 /** 264 * @brief FLR cleanup for all VFs 265 * 266 * @param p_hwfn 267 * @param p_ptt 268 * 269 * @return enum _ecore_status_t 270 */ 271 enum _ecore_status_t ecore_iov_vf_flr_cleanup(struct ecore_hwfn *p_hwfn, 272 struct ecore_ptt *p_ptt); 273 274 /** 275 * @brief FLR cleanup for single VF 276 * 277 * @param p_hwfn 278 * @param p_ptt 279 * @param rel_vf_id 280 * 281 * @return enum _ecore_status_t 282 */ 283 enum _ecore_status_t 284 ecore_iov_single_vf_flr_cleanup(struct ecore_hwfn *p_hwfn, 285 struct ecore_ptt *p_ptt, 286 u16 rel_vf_id); 287 288 /** 289 * @brief Update the bulletin with link information. Notice this does NOT 290 * send a bulletin update, only updates the PF's bulletin. 291 * 292 * @param p_hwfn 293 * @param p_vf 294 * @param params - the link params to use for the VF link configuration 295 * @param link - the link output to use for the VF link configuration 296 * @param p_caps - the link default capabilities. 297 */ 298 void ecore_iov_set_link(struct ecore_hwfn *p_hwfn, 299 u16 vfid, 300 struct ecore_mcp_link_params *params, 301 struct ecore_mcp_link_state *link, 302 struct ecore_mcp_link_capabilities *p_caps); 303 304 /** 305 * @brief Returns link information as perceived by VF. 306 * 307 * @param p_hwfn 308 * @param p_vf 309 * @param p_params - the link params visible to vf. 310 * @param p_link - the link state visible to vf. 311 * @param p_caps - the link default capabilities visible to vf. 312 */ 313 void ecore_iov_get_link(struct ecore_hwfn *p_hwfn, 314 u16 vfid, 315 struct ecore_mcp_link_params *params, 316 struct ecore_mcp_link_state *link, 317 struct ecore_mcp_link_capabilities *p_caps); 318 319 /** 320 * @brief return if the VF is pending FLR 321 * 322 * @param p_hwfn 323 * @param rel_vf_id 324 * 325 * @return bool 326 */ 327 bool ecore_iov_is_vf_pending_flr(struct ecore_hwfn *p_hwfn, 328 u16 rel_vf_id); 329 #endif 330 331 /** 332 * @brief Check if given VF ID @vfid is valid 333 * w.r.t. @b_enabled_only value 334 * if b_enabled_only = true - only enabled VF id is valid 335 * else any VF id less than max_vfs is valid 336 * 337 * @param p_hwfn 338 * @param rel_vf_id - Relative VF ID 339 * @param b_enabled_only - consider only enabled VF 340 * @param b_non_malicious - true iff we want to validate vf isn't malicious. 341 * 342 * @return bool - true for valid VF ID 343 */ 344 bool ecore_iov_is_valid_vfid(struct ecore_hwfn *p_hwfn, 345 int rel_vf_id, 346 bool b_enabled_only, bool b_non_malicious); 347 348 #ifndef LINUX_REMOVE 349 /** 350 * @brief Get VF's public info structure 351 * 352 * @param p_hwfn 353 * @param vfid - Relative VF ID 354 * @param b_enabled_only - false if want to access even if vf is disabled 355 * 356 * @return struct ecore_public_vf_info * 357 */ 358 struct ecore_public_vf_info* 359 ecore_iov_get_public_vf_info(struct ecore_hwfn *p_hwfn, 360 u16 vfid, bool b_enabled_only); 361 362 /** 363 * @brief fills a bitmask of all VFs which have pending unhandled 364 * messages. 365 * 366 * @param p_hwfn 367 */ 368 void ecore_iov_pf_get_pending_events(struct ecore_hwfn *p_hwfn, 369 u64 *events); 370 371 /** 372 * @brief Copy VF's message to PF's buffer 373 * 374 * @param p_hwfn 375 * @param ptt 376 * @param vfid 377 * 378 * @return enum _ecore_status_t 379 */ 380 enum _ecore_status_t ecore_iov_copy_vf_msg(struct ecore_hwfn *p_hwfn, 381 struct ecore_ptt *ptt, 382 int vfid); 383 /** 384 * @brief Set forced MAC address in PFs copy of bulletin board 385 * and configures FW/HW to support the configuration. 386 * 387 * @param p_hwfn 388 * @param mac 389 * @param vfid 390 */ 391 void ecore_iov_bulletin_set_forced_mac(struct ecore_hwfn *p_hwfn, 392 u8 *mac, int vfid); 393 394 /** 395 * @brief Set MAC address in PFs copy of bulletin board without 396 * configuring FW/HW. 397 * 398 * @param p_hwfn 399 * @param mac 400 * @param vfid 401 */ 402 enum _ecore_status_t ecore_iov_bulletin_set_mac(struct ecore_hwfn *p_hwfn, 403 u8 *mac, int vfid); 404 405 /** 406 * @brief Set default behaviour of VF in case no vlans are configured for it 407 * whether to accept only untagged traffic or all. 408 * Must be called prior to the VF vport-start. 409 * 410 * @param p_hwfn 411 * @param b_untagged_only 412 * @param vfid 413 * 414 * @return ECORE_SUCCESS if configuration would stick. 415 */ 416 enum _ecore_status_t 417 ecore_iov_bulletin_set_forced_untagged_default(struct ecore_hwfn *p_hwfn, 418 bool b_untagged_only, 419 int vfid); 420 421 /** 422 * @brief Get VFs opaque fid. 423 * 424 * @param p_hwfn 425 * @param vfid 426 * @param opaque_fid 427 */ 428 void ecore_iov_get_vfs_opaque_fid(struct ecore_hwfn *p_hwfn, int vfid, 429 u16 *opaque_fid); 430 431 /** 432 * @brief Set forced VLAN [pvid] in PFs copy of bulletin board 433 * and configures FW/HW to support the configuration. 434 * Setting of pvid 0 would clear the feature. 435 * @param p_hwfn 436 * @param pvid 437 * @param vfid 438 */ 439 void ecore_iov_bulletin_set_forced_vlan(struct ecore_hwfn *p_hwfn, 440 u16 pvid, int vfid); 441 442 /** 443 * @brief Check if VF has VPORT instance. This can be used 444 * to check if VPORT is active. 445 * 446 * @param p_hwfn 447 */ 448 bool ecore_iov_vf_has_vport_instance(struct ecore_hwfn *p_hwfn, int vfid); 449 450 /** 451 * @brief PF posts the bulletin to the VF 452 * 453 * @param p_hwfn 454 * @param p_vf 455 * @param p_ptt 456 * 457 * @return enum _ecore_status_t 458 */ 459 enum _ecore_status_t ecore_iov_post_vf_bulletin(struct ecore_hwfn *p_hwfn, 460 int vfid, 461 struct ecore_ptt *p_ptt); 462 463 /** 464 * @brief Check if given VF (@vfid) is marked as stopped 465 * 466 * @param p_hwfn 467 * @param vfid 468 * 469 * @return bool : true if stopped 470 */ 471 bool ecore_iov_is_vf_stopped(struct ecore_hwfn *p_hwfn, int vfid); 472 473 /** 474 * @brief Configure VF anti spoofing 475 * 476 * @param p_hwfn 477 * @param vfid 478 * @param val - spoofchk value - true/false 479 * 480 * @return enum _ecore_status_t 481 */ 482 enum _ecore_status_t ecore_iov_spoofchk_set(struct ecore_hwfn *p_hwfn, 483 int vfid, bool val); 484 485 /** 486 * @brief Get VF's configured spoof value. 487 * 488 * @param p_hwfn 489 * @param vfid 490 * 491 * @return bool - spoofchk value - true/false 492 */ 493 bool ecore_iov_spoofchk_get(struct ecore_hwfn *p_hwfn, int vfid); 494 495 /** 496 * @brief Check for SRIOV sanity by PF. 497 * 498 * @param p_hwfn 499 * @param vfid 500 * 501 * @return bool - true if sanity checks passes, else false 502 */ 503 bool ecore_iov_pf_sanity_check(struct ecore_hwfn *p_hwfn, int vfid); 504 505 /** 506 * @brief Get the num of VF chains. 507 * 508 * @param p_hwfn 509 * 510 * @return u8 511 */ 512 u8 ecore_iov_vf_chains_per_pf(struct ecore_hwfn *p_hwfn); 513 514 /** 515 * @brief Get vf request mailbox params 516 * 517 * @param p_hwfn 518 * @param rel_vf_id 519 * @param pp_req_virt_addr 520 * @param p_req_virt_size 521 */ 522 void ecore_iov_get_vf_req_virt_mbx_params(struct ecore_hwfn *p_hwfn, 523 u16 rel_vf_id, 524 void **pp_req_virt_addr, 525 u16 *p_req_virt_size); 526 527 /** 528 * @brief Get vf mailbox params 529 * 530 * @param p_hwfn 531 * @param rel_vf_id 532 * @param pp_reply_virt_addr 533 * @param p_reply_virt_size 534 */ 535 void ecore_iov_get_vf_reply_virt_mbx_params(struct ecore_hwfn *p_hwfn, 536 u16 rel_vf_id, 537 void **pp_reply_virt_addr, 538 u16 *p_reply_virt_size); 539 540 /** 541 * @brief Validate if the given length is a valid vfpf message 542 * length 543 * 544 * @param length 545 * 546 * @return bool 547 */ 548 bool ecore_iov_is_valid_vfpf_msg_length(u32 length); 549 550 /** 551 * @brief Return the max pfvf message length 552 * 553 * @return u32 554 */ 555 u32 ecore_iov_pfvf_msg_length(void); 556 557 /** 558 * @brief Returns forced MAC address if one is configured 559 * 560 * @parm p_hwfn 561 * @parm rel_vf_id 562 * 563 * @return OSAL_NULL if mac isn't forced; Otherwise, returns MAC. 564 */ 565 u8 *ecore_iov_bulletin_get_forced_mac(struct ecore_hwfn *p_hwfn, 566 u16 rel_vf_id); 567 568 /** 569 * @brief Returns pvid if one is configured 570 * 571 * @parm p_hwfn 572 * @parm rel_vf_id 573 * 574 * @return 0 if no pvid is configured, otherwise the pvid. 575 */ 576 u16 ecore_iov_bulletin_get_forced_vlan(struct ecore_hwfn *p_hwfn, 577 u16 rel_vf_id); 578 /** 579 * @brief Configure VFs tx rate 580 * 581 * @param p_hwfn 582 * @param p_ptt 583 * @param vfid 584 * @param val - tx rate value in Mb/sec. 585 * 586 * @return enum _ecore_status_t 587 */ 588 enum _ecore_status_t ecore_iov_configure_tx_rate(struct ecore_hwfn *p_hwfn, 589 struct ecore_ptt *p_ptt, 590 int vfid, int val); 591 592 /** 593 * @brief - Retrieves the statistics associated with a VF 594 * 595 * @param p_hwfn 596 * @param p_ptt 597 * @param vfid 598 * @param p_stats - this will be filled with the VF statistics 599 * 600 * @return ECORE_SUCCESS iff statistics were retrieved. Error otherwise. 601 */ 602 enum _ecore_status_t ecore_iov_get_vf_stats(struct ecore_hwfn *p_hwfn, 603 struct ecore_ptt *p_ptt, 604 int vfid, 605 struct ecore_eth_stats *p_stats); 606 607 /** 608 * @brief - Retrieves num of rxqs chains 609 * 610 * @param p_hwfn 611 * @param rel_vf_id 612 * 613 * @return num of rxqs chains. 614 */ 615 u8 ecore_iov_get_vf_num_rxqs(struct ecore_hwfn *p_hwfn, 616 u16 rel_vf_id); 617 618 /** 619 * @brief - Retrieves num of active rxqs chains 620 * 621 * @param p_hwfn 622 * @param rel_vf_id 623 * 624 * @return 625 */ 626 u8 ecore_iov_get_vf_num_active_rxqs(struct ecore_hwfn *p_hwfn, 627 u16 rel_vf_id); 628 629 /** 630 * @brief - Retrieves ctx pointer 631 * 632 * @param p_hwfn 633 * @param rel_vf_id 634 * 635 * @return 636 */ 637 void *ecore_iov_get_vf_ctx(struct ecore_hwfn *p_hwfn, 638 u16 rel_vf_id); 639 640 /** 641 * @brief - Retrieves VF`s num sbs 642 * 643 * @param p_hwfn 644 * @param rel_vf_id 645 * 646 * @return 647 */ 648 u8 ecore_iov_get_vf_num_sbs(struct ecore_hwfn *p_hwfn, 649 u16 rel_vf_id); 650 651 /** 652 * @brief - Returm true if VF is waiting for acquire 653 * 654 * @param p_hwfn 655 * @param rel_vf_id 656 * 657 * @return 658 */ 659 bool ecore_iov_is_vf_wait_for_acquire(struct ecore_hwfn *p_hwfn, 660 u16 rel_vf_id); 661 662 /** 663 * @brief - Returm true if VF is acquired but not initialized 664 * 665 * @param p_hwfn 666 * @param rel_vf_id 667 * 668 * @return 669 */ 670 bool ecore_iov_is_vf_acquired_not_initialized(struct ecore_hwfn *p_hwfn, 671 u16 rel_vf_id); 672 673 /** 674 * @brief - Returm true if VF is acquired and initialized 675 * 676 * @param p_hwfn 677 * @param rel_vf_id 678 * 679 * @return 680 */ 681 bool ecore_iov_is_vf_initialized(struct ecore_hwfn *p_hwfn, 682 u16 rel_vf_id); 683 684 /** 685 * @brief - Returm true if VF has started in FW 686 * 687 * @param p_hwfn 688 * @param rel_vf_id 689 * 690 * @return 691 */ 692 bool ecore_iov_is_vf_started(struct ecore_hwfn *p_hwfn, 693 u16 rel_vf_id); 694 695 /** 696 * @brief - Get VF's vport min rate configured. 697 * @param p_hwfn 698 * @param rel_vf_id 699 * 700 * @return - rate in Mbps 701 */ 702 int ecore_iov_get_vf_min_rate(struct ecore_hwfn *p_hwfn, int vfid); 703 704 /** 705 * @brief - Configure min rate for VF's vport. 706 * @param p_dev 707 * @param vfid 708 * @param - rate in Mbps 709 * 710 * @return 711 */ 712 enum _ecore_status_t ecore_iov_configure_min_tx_rate(struct ecore_dev *p_dev, 713 int vfid, u32 rate); 714 715 #endif 716 717 /** 718 * @brief ecore_pf_configure_vf_queue_coalesce - PF configure coalesce parameters 719 * of VFs for Rx and Tx queue. 720 * While the API allows setting coalescing per-qid, all queues sharing a SB 721 * should be in same range [i.e., either 0-0x7f, 0x80-0xff or 0x100-0x1ff] 722 * otherwise configuration would break. 723 * 724 * @param p_hwfn 725 * @param rx_coal - Rx Coalesce value in micro seconds. 726 * @param tx_coal - TX Coalesce value in micro seconds. 727 * @param vf_id 728 * @param qid 729 * 730 * @return int 731 **/ 732 enum _ecore_status_t 733 ecore_iov_pf_configure_vf_queue_coalesce(struct ecore_hwfn *p_hwfn, 734 u16 rx_coal, u16 tx_coal, 735 u16 vf_id, u16 qid); 736 737 /** 738 * @brief - Given a VF index, return index of next [including that] active VF. 739 * 740 * @param p_hwfn 741 * @param rel_vf_id 742 * 743 * @return MAX_NUM_VFS_E4 in case no further active VFs, otherwise index. 744 */ 745 u16 ecore_iov_get_next_active_vf(struct ecore_hwfn *p_hwfn, u16 rel_vf_id); 746 void ecore_iov_bulletin_set_udp_ports(struct ecore_hwfn *p_hwfn, int vfid, 747 u16 vxlan_port, u16 geneve_port); 748 749 #ifdef CONFIG_ECORE_SW_CHANNEL 750 /** 751 * @brief Set whether PF should communicate with VF using SW/HW channel 752 * Needs to be called for an enabled VF before acquire is over 753 * [latest good point for doing that is OSAL_IOV_VF_ACQUIRE()] 754 * 755 * @param p_hwfn 756 * @param vfid - relative vf index 757 * @param b_is_hw - true iff PF is to use HW channel for communication 758 */ 759 void ecore_iov_set_vf_hw_channel(struct ecore_hwfn *p_hwfn, int vfid, 760 bool b_is_hw); 761 #endif 762 #else 763 #ifndef LINUX_REMOVE 764 static OSAL_INLINE void ecore_iov_set_vfs_to_disable(struct ecore_dev OSAL_UNUSED *p_dev, u8 OSAL_UNUSED to_disable) {} 765 static OSAL_INLINE void ecore_iov_set_vf_to_disable(struct ecore_dev OSAL_UNUSED *p_dev, u16 OSAL_UNUSED rel_vf_id, u8 OSAL_UNUSED to_disable) {} 766 static OSAL_INLINE enum _ecore_status_t ecore_iov_init_hw_for_vf(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, struct ecore_iov_vf_init_params OSAL_UNUSED *p_params) {return ECORE_INVAL;} 767 static OSAL_INLINE void ecore_iov_process_mbx_req(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, int OSAL_UNUSED vfid) {} 768 static OSAL_INLINE enum _ecore_status_t ecore_iov_release_hw_for_vf(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, u16 OSAL_UNUSED rel_vf_id) {return ECORE_SUCCESS;} 769 static OSAL_INLINE enum _ecore_status_t ecore_iov_set_vf_ctx(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED vf_id, OSAL_UNUSED void *ctx) {return ECORE_INVAL;} 770 static OSAL_INLINE enum _ecore_status_t ecore_iov_vf_flr_cleanup(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt) {return ECORE_INVAL;} 771 static OSAL_INLINE enum _ecore_status_t ecore_iov_single_vf_flr_cleanup(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, u16 OSAL_UNUSED rel_vf_id) {return ECORE_INVAL;} 772 static OSAL_INLINE void ecore_iov_set_link(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED vfid, struct ecore_mcp_link_params OSAL_UNUSED *params, struct ecore_mcp_link_state OSAL_UNUSED *link, struct ecore_mcp_link_capabilities OSAL_UNUSED *p_caps) {} 773 static OSAL_INLINE void ecore_iov_get_link(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED vfid, struct ecore_mcp_link_params OSAL_UNUSED *params, struct ecore_mcp_link_state OSAL_UNUSED *link, struct ecore_mcp_link_capabilities OSAL_UNUSED *p_caps) {} 774 static OSAL_INLINE bool ecore_iov_is_vf_pending_flr(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return false;} 775 #endif 776 static OSAL_INLINE bool 777 ecore_iov_is_valid_vfid(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED rel_vf_id, 778 bool OSAL_UNUSED b_enabled_only, 779 bool OSAL_UNUSED b_non_malicious) 780 { 781 return false; 782 } 783 #ifndef LINUX_REMOVE 784 static OSAL_INLINE struct ecore_public_vf_info* ecore_iov_get_public_vf_info(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED vfid, bool OSAL_UNUSED b_enabled_only) {return OSAL_NULL;} 785 static OSAL_INLINE void ecore_iov_pf_add_pending_events(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u8 OSAL_UNUSED vfid) {} 786 static OSAL_INLINE void ecore_iov_pf_get_and_clear_pending_events(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u64 OSAL_UNUSED *events) {} 787 static OSAL_INLINE enum _ecore_status_t ecore_iov_copy_vf_msg(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *ptt, int OSAL_UNUSED vfid) {return ECORE_INVAL;} 788 static OSAL_INLINE void ecore_iov_bulletin_set_forced_mac(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u8 OSAL_UNUSED *mac, int OSAL_UNUSED vfid) {} 789 static OSAL_INLINE enum _ecore_status_t ecore_iov_bulletin_set_mac(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u8 OSAL_UNUSED *mac, OSAL_UNUSED int vfid) {return ECORE_INVAL;} 790 static OSAL_INLINE enum _ecore_status_t ecore_iov_bulletin_set_forced_untagged_default(struct ecore_hwfn OSAL_UNUSED *p_hwfn, bool OSAL_UNUSED b_untagged_only, int OSAL_UNUSED vfid) {return ECORE_INVAL;} 791 static OSAL_INLINE void ecore_iov_get_vfs_opaque_fid(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid, u16 OSAL_UNUSED *opaque_fid) {} 792 static OSAL_INLINE void ecore_iov_bulletin_set_forced_vlan(struct ecore_hwfn OSAL_UNUSED p_hwfn, u16 OSAL_UNUSED pvid, int OSAL_UNUSED vfid) {} 793 794 static OSAL_INLINE bool ecore_iov_vf_has_vport_instance(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) {return false;} 795 static OSAL_INLINE enum _ecore_status_t ecore_iov_post_vf_bulletin(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid, struct ecore_ptt OSAL_UNUSED *p_ptt) {return ECORE_INVAL;} 796 static OSAL_INLINE bool ecore_iov_is_vf_stopped(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) {return false;} 797 static OSAL_INLINE enum _ecore_status_t ecore_iov_spoofchk_set(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid, bool OSAL_UNUSED val) {return ECORE_INVAL;} 798 static OSAL_INLINE bool ecore_iov_spoofchk_get(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) {return false;} 799 static OSAL_INLINE bool ecore_iov_pf_sanity_check(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) {return false;} 800 static OSAL_INLINE u8 ecore_iov_vf_chains_per_pf(struct ecore_hwfn OSAL_UNUSED *p_hwfn) {return 0;} 801 static OSAL_INLINE void ecore_iov_get_vf_req_virt_mbx_params(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id, void OSAL_UNUSED **pp_req_virt_addr, u16 OSAL_UNUSED *p_req_virt_size) {} 802 static OSAL_INLINE void ecore_iov_get_vf_reply_virt_mbx_params(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id, void OSAL_UNUSED **pp_reply_virt_addr, u16 OSAL_UNUSED *p_reply_virt_size) {} 803 static OSAL_INLINE bool ecore_iov_is_valid_vfpf_msg_length(u32 OSAL_UNUSED length) {return false;} 804 static OSAL_INLINE u32 ecore_iov_pfvf_msg_length(void) {return 0;} 805 static OSAL_INLINE u8 *ecore_iov_bulletin_get_forced_mac(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return OSAL_NULL;} 806 static OSAL_INLINE u16 ecore_iov_bulletin_get_forced_vlan(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return 0;} 807 static OSAL_INLINE enum _ecore_status_t ecore_iov_configure_tx_rate(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, int OSAL_UNUSED vfid, int OSAL_UNUSED val) { return ECORE_INVAL; } 808 static OSAL_INLINE enum _ecore_status_t ecore_iov_get_vf_stats(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, int OSAL_UNUSED vfid, struct ecore_eth_stats OSAL_UNUSED *p_stats) { return ECORE_INVAL; } 809 810 static OSAL_INLINE u8 ecore_iov_get_vf_num_rxqs(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return 0;} 811 static OSAL_INLINE u8 ecore_iov_get_vf_num_active_rxqs(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return 0;} 812 static OSAL_INLINE void *ecore_iov_get_vf_ctx(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return OSAL_NULL;} 813 static OSAL_INLINE u8 ecore_iov_get_vf_num_sbs(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return 0;} 814 static OSAL_INLINE bool ecore_iov_is_vf_wait_for_acquire(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return false;} 815 static OSAL_INLINE bool ecore_iov_is_vf_acquired_not_initialized(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return false;} 816 static OSAL_INLINE bool ecore_iov_is_vf_initialized(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return false;} 817 static OSAL_INLINE int ecore_iov_get_vf_min_rate(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) { return 0; } 818 static OSAL_INLINE enum _ecore_status_t ecore_iov_configure_min_tx_rate(struct ecore_dev OSAL_UNUSED *p_dev, int OSAL_UNUSED vfid, OSAL_UNUSED u32 rate) { return ECORE_INVAL; } 819 #endif 820 static OSAL_INLINE void ecore_iov_bulletin_set_udp_ports(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid, u16 OSAL_UNUSED vxlan_port, u16 OSAL_UNUSED geneve_port) { return; } 821 static OSAL_INLINE u16 ecore_iov_get_next_active_vf(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) { return MAX_NUM_VFS_E4; } 822 823 #ifdef CONFIG_ECORE_SW_CHANNEL 824 static OSAL_INLINE void 825 ecore_iov_set_vf_hw_channel(struct ecore_hwfn OSAL_UNUSED *p_hwfn, 826 int OSAL_UNUSED vfid, bool OSAL_UNUSED b_is_hw) {} 827 #endif 828 #endif 829 830 #define ecore_for_each_vf(_p_hwfn, _i) \ 831 for (_i = ecore_iov_get_next_active_vf(_p_hwfn, 0); \ 832 _i < MAX_NUM_VFS_E4; \ 833 _i = ecore_iov_get_next_active_vf(_p_hwfn, _i + 1)) 834 835 #endif 836