xref: /freebsd/sys/dev/qlnx/qlnxe/ecore_iov_api.h (revision 95ee2897)
1 /*
2  * Copyright (c) 2017-2018 Cavium, Inc.
3  * All rights reserved.
4  *
5  *  Redistribution and use in source and binary forms, with or without
6  *  modification, are permitted provided that the following conditions
7  *  are met:
8  *
9  *  1. Redistributions of source code must retain the above copyright
10  *     notice, this list of conditions and the following disclaimer.
11  *  2. Redistributions in binary form must reproduce the above copyright
12  *     notice, this list of conditions and the following disclaimer in the
13  *     documentation and/or other materials provided with the distribution.
14  *
15  *  THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
16  *  AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17  *  IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18  *  ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
19  *  LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
20  *  CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
21  *  SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
22  *  INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
23  *  CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
24  *  ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
25  *  POSSIBILITY OF SUCH DAMAGE.
26  *
27  */
28 
29 #ifndef __ECORE_SRIOV_API_H__
30 #define __ECORE_SRIOV_API_H__
31 
32 #include "common_hsi.h"
33 #include "ecore_status.h"
34 
35 #define ECORE_ETH_VF_NUM_MAC_FILTERS 1
36 #define ECORE_ETH_VF_NUM_VLAN_FILTERS 2
37 #define ECORE_VF_ARRAY_LENGTH (3)
38 
39 #define IS_VF(p_dev)		((p_dev)->b_is_vf)
40 #define IS_PF(p_dev)		(!((p_dev)->b_is_vf))
41 #ifdef CONFIG_ECORE_SRIOV
42 #define IS_PF_SRIOV(p_hwfn)	(!!((p_hwfn)->p_dev->p_iov_info))
43 #else
44 #define IS_PF_SRIOV(p_hwfn)	(0)
45 #endif
46 #define IS_PF_SRIOV_ALLOC(p_hwfn)	(!!((p_hwfn)->pf_iov_info))
47 #define IS_PF_PDA(p_hwfn)	0 /* @@TBD Michalk */
48 
49 /* @@@ TBD MichalK - what should this number be*/
50 #define ECORE_MAX_VF_CHAINS_PER_PF 16
51 
52 /* vport update extended feature tlvs flags */
53 enum ecore_iov_vport_update_flag {
54 	ECORE_IOV_VP_UPDATE_ACTIVATE		= 0,
55 	ECORE_IOV_VP_UPDATE_VLAN_STRIP		= 1,
56 	ECORE_IOV_VP_UPDATE_TX_SWITCH		= 2,
57 	ECORE_IOV_VP_UPDATE_MCAST		= 3,
58 	ECORE_IOV_VP_UPDATE_ACCEPT_PARAM	= 4,
59 	ECORE_IOV_VP_UPDATE_RSS			= 5,
60 	ECORE_IOV_VP_UPDATE_ACCEPT_ANY_VLAN	= 6,
61 	ECORE_IOV_VP_UPDATE_SGE_TPA		= 7,
62 	ECORE_IOV_VP_UPDATE_MAX			= 8,
63 };
64 
65 /*PF to VF STATUS is part of vfpf-channel API
66 and must be forward compatible */
67 enum ecore_iov_pf_to_vf_status {
68         PFVF_STATUS_WAITING = 0,
69         PFVF_STATUS_SUCCESS,
70         PFVF_STATUS_FAILURE,
71         PFVF_STATUS_NOT_SUPPORTED,
72         PFVF_STATUS_NO_RESOURCE,
73         PFVF_STATUS_FORCED,
74 	PFVF_STATUS_MALICIOUS,
75 };
76 
77 struct ecore_mcp_link_params;
78 struct ecore_mcp_link_state;
79 struct ecore_mcp_link_capabilities;
80 
81 /* These defines are used by the hw-channel; should never change order */
82 #define VFPF_ACQUIRE_OS_LINUX (0)
83 #define VFPF_ACQUIRE_OS_WINDOWS (1)
84 #define VFPF_ACQUIRE_OS_ESX (2)
85 #define VFPF_ACQUIRE_OS_SOLARIS (3)
86 #define VFPF_ACQUIRE_OS_LINUX_USERSPACE (4)
87 #define VFPF_ACQUIRE_OS_FREEBSD (5)
88 
89 struct ecore_vf_acquire_sw_info {
90 	u32 driver_version;
91 	u8 os_type;
92 };
93 
94 struct ecore_public_vf_info {
95 	/* These copies will later be reflected in the bulletin board,
96 	 * but this copy should be newer.
97 	 */
98 	u8 forced_mac[ETH_ALEN];
99 	u16 forced_vlan;
100 };
101 
102 struct ecore_iov_vf_init_params {
103 	u16 rel_vf_id;
104 
105 	/* Number of requested Queues; Currently, don't support different
106 	 * number of Rx/Tx queues.
107 	 */
108 	/* TODO - remove this limitation */
109 	u16 num_queues;
110 
111 	/* Allow the client to choose which qzones to use for Rx/Tx,
112 	 * and which queue_base to use for Tx queues on a per-queue basis.
113 	 * Notice values should be relative to the PF resources.
114 	 */
115 	u16 req_rx_queue[ECORE_MAX_VF_CHAINS_PER_PF];
116 	u16 req_tx_queue[ECORE_MAX_VF_CHAINS_PER_PF];
117 
118 	u8 vport_id;
119 
120 	/* Should be set in case RSS is going to be used for VF */
121 	u8 rss_eng_id;
122 };
123 
124 #ifdef CONFIG_ECORE_SW_CHANNEL
125 /* This is SW channel related only... */
126 enum mbx_state {
127 	VF_PF_UNKNOWN_STATE			= 0,
128 	VF_PF_WAIT_FOR_START_REQUEST		= 1,
129 	VF_PF_WAIT_FOR_NEXT_CHUNK_OF_REQUEST	= 2,
130 	VF_PF_REQUEST_IN_PROCESSING		= 3,
131 	VF_PF_RESPONSE_READY			= 4,
132 };
133 
134 struct ecore_iov_sw_mbx {
135 	enum mbx_state		mbx_state;
136 
137 	u32			request_size;
138 	u32			request_offset;
139 
140 	u32			response_size;
141 	u32			response_offset;
142 };
143 
144 /**
145  * @brief Get the vf sw mailbox params
146  *
147  * @param p_hwfn
148  * @param rel_vf_id
149  *
150  * @return struct ecore_iov_sw_mbx*
151  */
152 struct ecore_iov_sw_mbx*
153 ecore_iov_get_vf_sw_mbx(struct ecore_hwfn *p_hwfn,
154 			u16 rel_vf_id);
155 #endif
156 
157 /* This struct is part of ecore_dev and contains data relevant to all hwfns;
158  * Initialized only if SR-IOV cpabability is exposed in PCIe config space.
159  */
160 struct ecore_hw_sriov_info {
161 	/* standard SRIOV capability fields, mostly for debugging */
162 	int	pos;		/* capability position */
163 	int	nres;		/* number of resources */
164 	u32	cap;		/* SR-IOV Capabilities */
165 	u16	ctrl;		/* SR-IOV Control */
166 	u16	total_vfs;	/* total VFs associated with the PF */
167 	u16	num_vfs;        /* number of vfs that have been started */
168 	u16	initial_vfs;    /* initial VFs associated with the PF */
169 	u16	nr_virtfn;	/* number of VFs available */
170 	u16	offset;		/* first VF Routing ID offset */
171 	u16	stride;		/* following VF stride */
172 	u16	vf_device_id;	/* VF device id */
173 	u32	pgsz;		/* page size for BAR alignment */
174 	u8	link;		/* Function Dependency Link */
175 
176 	u32	first_vf_in_pf;
177 };
178 
179 #ifdef CONFIG_ECORE_SRIOV
180 #ifndef LINUX_REMOVE
181 /**
182  * @brief mark/clear all VFs before/after an incoming PCIe sriov
183  *        disable.
184  *
185  * @param p_dev
186  * @param to_disable
187  */
188 void ecore_iov_set_vfs_to_disable(struct ecore_dev *p_dev,
189 				  u8 to_disable);
190 
191 /**
192  * @brief mark/clear chosen VF before/after an incoming PCIe
193  *        sriov disable.
194  *
195  * @param p_dev
196  * @param rel_vf_id
197  * @param to_disable
198  */
199 void ecore_iov_set_vf_to_disable(struct ecore_dev *p_dev,
200 				 u16 rel_vf_id,
201 				 u8 to_disable);
202 
203 /**
204  * @brief ecore_iov_init_hw_for_vf - initialize the HW for
205  *        enabling access of a VF. Also includes preparing the
206  *        IGU for VF access. This needs to be called AFTER hw is
207  *        initialized and BEFORE VF is loaded inside the VM.
208  *
209  * @param p_hwfn
210  * @param p_ptt
211  * @param p_params
212  *
213  * @return enum _ecore_status_t
214  */
215 enum _ecore_status_t
216 ecore_iov_init_hw_for_vf(struct ecore_hwfn *p_hwfn,
217 			 struct ecore_ptt *p_ptt,
218 			 struct ecore_iov_vf_init_params *p_params);
219 
220 /**
221  * @brief ecore_iov_process_mbx_req - process a request received
222  *        from the VF
223  *
224  * @param p_hwfn
225  * @param p_ptt
226  * @param vfid
227  */
228 void ecore_iov_process_mbx_req(struct ecore_hwfn *p_hwfn,
229 			       struct ecore_ptt *p_ptt,
230 			       int vfid);
231 
232 /**
233  * @brief ecore_iov_release_hw_for_vf - called once upper layer
234  *        knows VF is done with - can release any resources
235  *        allocated for VF at this point. this must be done once
236  *        we know VF is no longer loaded in VM.
237  *
238  * @param p_hwfn
239  * @param p_ptt
240  * @param rel_vf_id
241  *
242  * @return enum _ecore_status_t
243  */
244 enum _ecore_status_t ecore_iov_release_hw_for_vf(struct ecore_hwfn *p_hwfn,
245 						 struct ecore_ptt *p_ptt,
246 						 u16 rel_vf_id);
247 
248 /**
249  * @brief ecore_iov_set_vf_ctx - set a context for a given VF
250  *
251  * @param p_hwfn
252  * @param vf_id
253  * @param ctx
254  *
255  * @return enum _ecore_status_t
256  */
257 enum _ecore_status_t ecore_iov_set_vf_ctx(struct ecore_hwfn *p_hwfn,
258 					  u16 vf_id,
259 					  void *ctx);
260 
261 /**
262  * @brief FLR cleanup for all VFs
263  *
264  * @param p_hwfn
265  * @param p_ptt
266  *
267  * @return enum _ecore_status_t
268  */
269 enum _ecore_status_t ecore_iov_vf_flr_cleanup(struct ecore_hwfn *p_hwfn,
270 					      struct ecore_ptt *p_ptt);
271 
272 /**
273  * @brief FLR cleanup for single VF
274  *
275  * @param p_hwfn
276  * @param p_ptt
277  * @param rel_vf_id
278  *
279  * @return enum _ecore_status_t
280  */
281 enum _ecore_status_t
282 ecore_iov_single_vf_flr_cleanup(struct ecore_hwfn *p_hwfn,
283 				struct ecore_ptt *p_ptt,
284 				u16 rel_vf_id);
285 
286 /**
287  * @brief Update the bulletin with link information. Notice this does NOT
288  *        send a bulletin update, only updates the PF's bulletin.
289  *
290  * @param p_hwfn
291  * @param p_vf
292  * @param params - the link params to use for the VF link configuration
293  * @param link - the link output to use for the VF link configuration
294  * @param p_caps - the link default capabilities.
295  */
296 void ecore_iov_set_link(struct ecore_hwfn *p_hwfn,
297 			u16 vfid,
298 			struct ecore_mcp_link_params *params,
299 			struct ecore_mcp_link_state *link,
300 			struct ecore_mcp_link_capabilities *p_caps);
301 
302 /**
303  * @brief Returns link information as perceived by VF.
304  *
305  * @param p_hwfn
306  * @param p_vf
307  * @param p_params - the link params visible to vf.
308  * @param p_link - the link state visible to vf.
309  * @param p_caps - the link default capabilities visible to vf.
310  */
311 void ecore_iov_get_link(struct ecore_hwfn *p_hwfn,
312 			u16 vfid,
313 			struct ecore_mcp_link_params *params,
314 			struct ecore_mcp_link_state *link,
315 			struct ecore_mcp_link_capabilities *p_caps);
316 
317 /**
318  * @brief return if the VF is pending FLR
319  *
320  * @param p_hwfn
321  * @param rel_vf_id
322  *
323  * @return bool
324  */
325 bool ecore_iov_is_vf_pending_flr(struct ecore_hwfn *p_hwfn,
326 				 u16 rel_vf_id);
327 #endif
328 
329 /**
330  * @brief Check if given VF ID @vfid is valid
331  *        w.r.t. @b_enabled_only value
332  *        if b_enabled_only = true - only enabled VF id is valid
333  *        else any VF id less than max_vfs is valid
334  *
335  * @param p_hwfn
336  * @param rel_vf_id - Relative VF ID
337  * @param b_enabled_only - consider only enabled VF
338  * @param b_non_malicious - true iff we want to validate vf isn't malicious.
339  *
340  * @return bool - true for valid VF ID
341  */
342 bool ecore_iov_is_valid_vfid(struct ecore_hwfn *p_hwfn,
343 			     int rel_vf_id,
344 			     bool b_enabled_only, bool b_non_malicious);
345 
346 #ifndef LINUX_REMOVE
347 /**
348  * @brief Get VF's public info structure
349  *
350  * @param p_hwfn
351  * @param vfid - Relative VF ID
352  * @param b_enabled_only - false if want to access even if vf is disabled
353  *
354  * @return struct ecore_public_vf_info *
355  */
356 struct ecore_public_vf_info*
357 ecore_iov_get_public_vf_info(struct ecore_hwfn *p_hwfn,
358 			     u16 vfid, bool b_enabled_only);
359 
360 /**
361  * @brief fills a bitmask of all VFs which have pending unhandled
362  *        messages.
363  *
364  * @param p_hwfn
365  */
366 void ecore_iov_pf_get_pending_events(struct ecore_hwfn *p_hwfn,
367 				     u64 *events);
368 
369 /**
370  * @brief Copy VF's message to PF's buffer
371  *
372  * @param p_hwfn
373  * @param ptt
374  * @param vfid
375  *
376  * @return enum _ecore_status_t
377  */
378 enum _ecore_status_t ecore_iov_copy_vf_msg(struct ecore_hwfn *p_hwfn,
379 					   struct ecore_ptt *ptt,
380 					   int vfid);
381 /**
382  * @brief Set forced MAC address in PFs copy of bulletin board
383  *        and configures FW/HW to support the configuration.
384  *
385  * @param p_hwfn
386  * @param mac
387  * @param vfid
388  */
389 void ecore_iov_bulletin_set_forced_mac(struct ecore_hwfn *p_hwfn,
390 				       u8 *mac, int vfid);
391 
392 /**
393  * @brief Set MAC address in PFs copy of bulletin board without
394  *        configuring FW/HW.
395  *
396  * @param p_hwfn
397  * @param mac
398  * @param vfid
399  */
400 enum _ecore_status_t ecore_iov_bulletin_set_mac(struct ecore_hwfn *p_hwfn,
401 						u8 *mac, int vfid);
402 
403 /**
404  * @brief Set default behaviour of VF in case no vlans are configured for it
405  *        whether to accept only untagged traffic or all.
406  *        Must be called prior to the VF vport-start.
407  *
408  * @param p_hwfn
409  * @param b_untagged_only
410  * @param vfid
411  *
412  * @return ECORE_SUCCESS if configuration would stick.
413  */
414 enum _ecore_status_t
415 ecore_iov_bulletin_set_forced_untagged_default(struct ecore_hwfn *p_hwfn,
416 					       bool b_untagged_only,
417 					       int vfid);
418 
419 /**
420  * @brief Get VFs opaque fid.
421  *
422  * @param p_hwfn
423  * @param vfid
424  * @param opaque_fid
425  */
426 void ecore_iov_get_vfs_opaque_fid(struct ecore_hwfn *p_hwfn, int vfid,
427 				  u16 *opaque_fid);
428 
429 /**
430  * @brief Set forced VLAN [pvid] in PFs copy of bulletin board
431  *        and configures FW/HW to support the configuration.
432  *        Setting of pvid 0 would clear the feature.
433  * @param p_hwfn
434  * @param pvid
435  * @param vfid
436  */
437 void ecore_iov_bulletin_set_forced_vlan(struct ecore_hwfn *p_hwfn,
438 					u16 pvid, int vfid);
439 
440 /**
441  * @brief Check if VF has VPORT instance. This can be used
442  *	  to check if VPORT is active.
443  *
444  * @param p_hwfn
445  */
446 bool ecore_iov_vf_has_vport_instance(struct ecore_hwfn *p_hwfn, int vfid);
447 
448 /**
449  * @brief PF posts the bulletin to the VF
450  *
451  * @param p_hwfn
452  * @param p_vf
453  * @param p_ptt
454  *
455  * @return enum _ecore_status_t
456  */
457 enum _ecore_status_t ecore_iov_post_vf_bulletin(struct ecore_hwfn *p_hwfn,
458 						int vfid,
459 						struct ecore_ptt *p_ptt);
460 
461 /**
462  * @brief Check if given VF (@vfid) is marked as stopped
463  *
464  * @param p_hwfn
465  * @param vfid
466  *
467  * @return bool : true if stopped
468  */
469 bool ecore_iov_is_vf_stopped(struct ecore_hwfn *p_hwfn, int vfid);
470 
471 /**
472  * @brief Configure VF anti spoofing
473  *
474  * @param p_hwfn
475  * @param vfid
476  * @param val - spoofchk value - true/false
477  *
478  * @return enum _ecore_status_t
479  */
480 enum _ecore_status_t ecore_iov_spoofchk_set(struct ecore_hwfn *p_hwfn,
481 					    int vfid, bool val);
482 
483 /**
484  * @brief Get VF's configured spoof value.
485  *
486  * @param p_hwfn
487  * @param vfid
488  *
489  * @return bool - spoofchk value - true/false
490  */
491 bool ecore_iov_spoofchk_get(struct ecore_hwfn *p_hwfn, int vfid);
492 
493 /**
494  * @brief Check for SRIOV sanity by PF.
495  *
496  * @param p_hwfn
497  * @param vfid
498  *
499  * @return bool - true if sanity checks passes, else false
500  */
501 bool ecore_iov_pf_sanity_check(struct ecore_hwfn *p_hwfn, int vfid);
502 
503 /**
504  * @brief Get the num of VF chains.
505  *
506  * @param p_hwfn
507  *
508  * @return u8
509  */
510 u8 ecore_iov_vf_chains_per_pf(struct ecore_hwfn *p_hwfn);
511 
512 /**
513  * @brief Get vf request mailbox params
514  *
515  * @param p_hwfn
516  * @param rel_vf_id
517  * @param pp_req_virt_addr
518  * @param p_req_virt_size
519  */
520 void ecore_iov_get_vf_req_virt_mbx_params(struct ecore_hwfn *p_hwfn,
521 					  u16 rel_vf_id,
522 					  void **pp_req_virt_addr,
523 					  u16 *p_req_virt_size);
524 
525 /**
526  * @brief Get vf mailbox params
527  *
528  * @param p_hwfn
529  * @param rel_vf_id
530  * @param pp_reply_virt_addr
531  * @param p_reply_virt_size
532  */
533 void ecore_iov_get_vf_reply_virt_mbx_params(struct ecore_hwfn *p_hwfn,
534 					    u16	rel_vf_id,
535 					    void **pp_reply_virt_addr,
536 					    u16	*p_reply_virt_size);
537 
538 /**
539  * @brief Validate if the given length is a valid vfpf message
540  *        length
541  *
542  * @param length
543  *
544  * @return bool
545  */
546 bool ecore_iov_is_valid_vfpf_msg_length(u32 length);
547 
548 /**
549  * @brief Return the max pfvf message length
550  *
551  * @return u32
552  */
553 u32 ecore_iov_pfvf_msg_length(void);
554 
555 /**
556  * @brief Returns forced MAC address if one is configured
557  *
558  * @parm p_hwfn
559  * @parm rel_vf_id
560  *
561  * @return OSAL_NULL if mac isn't forced; Otherwise, returns MAC.
562  */
563 u8 *ecore_iov_bulletin_get_forced_mac(struct ecore_hwfn *p_hwfn,
564 				      u16 rel_vf_id);
565 
566 /**
567  * @brief Returns pvid if one is configured
568  *
569  * @parm p_hwfn
570  * @parm rel_vf_id
571  *
572  * @return 0 if no pvid is configured, otherwise the pvid.
573  */
574 u16 ecore_iov_bulletin_get_forced_vlan(struct ecore_hwfn *p_hwfn,
575 				       u16 rel_vf_id);
576 /**
577  * @brief Configure VFs tx rate
578  *
579  * @param p_hwfn
580  * @param p_ptt
581  * @param vfid
582  * @param val - tx rate value in Mb/sec.
583  *
584  * @return enum _ecore_status_t
585  */
586 enum _ecore_status_t ecore_iov_configure_tx_rate(struct ecore_hwfn *p_hwfn,
587 						 struct ecore_ptt *p_ptt,
588 						 int vfid, int val);
589 
590 /**
591  * @brief - Retrieves the statistics associated with a VF
592  *
593  * @param p_hwfn
594  * @param p_ptt
595  * @param vfid
596  * @param p_stats - this will be filled with the VF statistics
597  *
598  * @return ECORE_SUCCESS iff statistics were retrieved. Error otherwise.
599  */
600 enum _ecore_status_t ecore_iov_get_vf_stats(struct ecore_hwfn *p_hwfn,
601 					    struct ecore_ptt *p_ptt,
602 					    int vfid,
603 					    struct ecore_eth_stats *p_stats);
604 
605 /**
606  * @brief - Retrieves num of rxqs chains
607  *
608  * @param p_hwfn
609  * @param rel_vf_id
610  *
611  * @return num of rxqs chains.
612  */
613 u8 ecore_iov_get_vf_num_rxqs(struct ecore_hwfn *p_hwfn,
614 			     u16 rel_vf_id);
615 
616 /**
617  * @brief - Retrieves num of active rxqs chains
618  *
619  * @param p_hwfn
620  * @param rel_vf_id
621  *
622  * @return
623  */
624 u8 ecore_iov_get_vf_num_active_rxqs(struct ecore_hwfn *p_hwfn,
625 				    u16 rel_vf_id);
626 
627 /**
628  * @brief - Retrieves ctx pointer
629  *
630  * @param p_hwfn
631  * @param rel_vf_id
632  *
633  * @return
634  */
635 void *ecore_iov_get_vf_ctx(struct ecore_hwfn *p_hwfn,
636 			   u16 rel_vf_id);
637 
638 /**
639  * @brief - Retrieves VF`s num sbs
640  *
641  * @param p_hwfn
642  * @param rel_vf_id
643  *
644  * @return
645  */
646 u8 ecore_iov_get_vf_num_sbs(struct ecore_hwfn *p_hwfn,
647 			    u16 rel_vf_id);
648 
649 /**
650  * @brief - Returm true if VF is waiting for acquire
651  *
652  * @param p_hwfn
653  * @param rel_vf_id
654  *
655  * @return
656  */
657 bool ecore_iov_is_vf_wait_for_acquire(struct ecore_hwfn *p_hwfn,
658 				      u16 rel_vf_id);
659 
660 /**
661  * @brief - Returm true if VF is acquired but not initialized
662  *
663  * @param p_hwfn
664  * @param rel_vf_id
665  *
666  * @return
667  */
668 bool ecore_iov_is_vf_acquired_not_initialized(struct ecore_hwfn *p_hwfn,
669 					      u16 rel_vf_id);
670 
671 /**
672  * @brief - Returm true if VF is acquired and initialized
673  *
674  * @param p_hwfn
675  * @param rel_vf_id
676  *
677  * @return
678  */
679 bool ecore_iov_is_vf_initialized(struct ecore_hwfn *p_hwfn,
680 				 u16 rel_vf_id);
681 
682 /**
683  * @brief - Returm true if VF has started in FW
684  *
685  * @param p_hwfn
686  * @param rel_vf_id
687  *
688  * @return
689  */
690 bool ecore_iov_is_vf_started(struct ecore_hwfn *p_hwfn,
691 			     u16 rel_vf_id);
692 
693 /**
694  * @brief - Get VF's vport min rate configured.
695  * @param p_hwfn
696  * @param rel_vf_id
697  *
698  * @return - rate in Mbps
699  */
700 int ecore_iov_get_vf_min_rate(struct ecore_hwfn *p_hwfn, int vfid);
701 
702 /**
703  * @brief - Configure min rate for VF's vport.
704  * @param p_dev
705  * @param vfid
706  * @param - rate in Mbps
707  *
708  * @return
709  */
710 enum _ecore_status_t ecore_iov_configure_min_tx_rate(struct ecore_dev *p_dev,
711 						     int vfid, u32 rate);
712 
713 #endif
714 
715 /**
716  * @brief ecore_pf_configure_vf_queue_coalesce - PF configure coalesce parameters
717  *    of VFs for Rx and Tx queue.
718  *    While the API allows setting coalescing per-qid, all queues sharing a SB
719  *    should be in same range [i.e., either 0-0x7f, 0x80-0xff or 0x100-0x1ff]
720  *    otherwise configuration would break.
721  *
722  * @param p_hwfn
723  * @param rx_coal - Rx Coalesce value in micro seconds.
724  * @param tx_coal - TX Coalesce value in micro seconds.
725  * @param vf_id
726  * @param qid
727  *
728  * @return int
729  **/
730 enum _ecore_status_t
731 ecore_iov_pf_configure_vf_queue_coalesce(struct ecore_hwfn *p_hwfn,
732 					 u16 rx_coal, u16 tx_coal,
733 					 u16 vf_id, u16 qid);
734 
735 /**
736  * @brief - Given a VF index, return index of next [including that] active VF.
737  *
738  * @param p_hwfn
739  * @param rel_vf_id
740  *
741  * @return MAX_NUM_VFS_E4 in case no further active VFs, otherwise index.
742  */
743 u16 ecore_iov_get_next_active_vf(struct ecore_hwfn *p_hwfn, u16 rel_vf_id);
744 void ecore_iov_bulletin_set_udp_ports(struct ecore_hwfn *p_hwfn, int vfid,
745 				      u16 vxlan_port, u16 geneve_port);
746 
747 #ifdef CONFIG_ECORE_SW_CHANNEL
748 /**
749  * @brief Set whether PF should communicate with VF using SW/HW channel
750  *        Needs to be called for an enabled VF before acquire is over
751  *        [latest good point for doing that is OSAL_IOV_VF_ACQUIRE()]
752  *
753  * @param p_hwfn
754  * @param vfid - relative vf index
755  * @param b_is_hw - true iff PF is to use HW channel for communication
756  */
757 void ecore_iov_set_vf_hw_channel(struct ecore_hwfn *p_hwfn, int vfid,
758 				 bool b_is_hw);
759 #endif
760 #else
761 #ifndef LINUX_REMOVE
ecore_iov_set_vfs_to_disable(struct ecore_dev OSAL_UNUSED * p_dev,u8 OSAL_UNUSED to_disable)762 static OSAL_INLINE void ecore_iov_set_vfs_to_disable(struct ecore_dev OSAL_UNUSED *p_dev, u8 OSAL_UNUSED to_disable) {}
ecore_iov_set_vf_to_disable(struct ecore_dev OSAL_UNUSED * p_dev,u16 OSAL_UNUSED rel_vf_id,u8 OSAL_UNUSED to_disable)763 static OSAL_INLINE void ecore_iov_set_vf_to_disable(struct ecore_dev OSAL_UNUSED *p_dev, u16 OSAL_UNUSED rel_vf_id, u8 OSAL_UNUSED to_disable) {}
ecore_iov_init_hw_for_vf(struct ecore_hwfn OSAL_UNUSED * p_hwfn,struct ecore_ptt OSAL_UNUSED * p_ptt,struct ecore_iov_vf_init_params OSAL_UNUSED * p_params)764 static OSAL_INLINE enum _ecore_status_t ecore_iov_init_hw_for_vf(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, struct ecore_iov_vf_init_params OSAL_UNUSED *p_params) {return ECORE_INVAL;}
ecore_iov_process_mbx_req(struct ecore_hwfn OSAL_UNUSED * p_hwfn,struct ecore_ptt OSAL_UNUSED * p_ptt,int OSAL_UNUSED vfid)765 static OSAL_INLINE void ecore_iov_process_mbx_req(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, int OSAL_UNUSED vfid) {}
ecore_iov_release_hw_for_vf(struct ecore_hwfn OSAL_UNUSED * p_hwfn,struct ecore_ptt OSAL_UNUSED * p_ptt,u16 OSAL_UNUSED rel_vf_id)766 static OSAL_INLINE enum _ecore_status_t ecore_iov_release_hw_for_vf(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, u16 OSAL_UNUSED rel_vf_id) {return ECORE_SUCCESS;}
ecore_iov_set_vf_ctx(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED vf_id,OSAL_UNUSED void * ctx)767 static OSAL_INLINE enum _ecore_status_t ecore_iov_set_vf_ctx(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED vf_id, OSAL_UNUSED void *ctx) {return ECORE_INVAL;}
ecore_iov_vf_flr_cleanup(struct ecore_hwfn OSAL_UNUSED * p_hwfn,struct ecore_ptt OSAL_UNUSED * p_ptt)768 static OSAL_INLINE enum _ecore_status_t ecore_iov_vf_flr_cleanup(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt) {return ECORE_INVAL;}
ecore_iov_single_vf_flr_cleanup(struct ecore_hwfn OSAL_UNUSED * p_hwfn,struct ecore_ptt OSAL_UNUSED * p_ptt,u16 OSAL_UNUSED rel_vf_id)769 static OSAL_INLINE enum _ecore_status_t ecore_iov_single_vf_flr_cleanup(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, u16 OSAL_UNUSED rel_vf_id) {return ECORE_INVAL;}
ecore_iov_set_link(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED vfid,struct ecore_mcp_link_params OSAL_UNUSED * params,struct ecore_mcp_link_state OSAL_UNUSED * link,struct ecore_mcp_link_capabilities OSAL_UNUSED * p_caps)770 static OSAL_INLINE void ecore_iov_set_link(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED vfid, struct ecore_mcp_link_params OSAL_UNUSED *params, struct ecore_mcp_link_state OSAL_UNUSED *link, struct ecore_mcp_link_capabilities OSAL_UNUSED *p_caps) {}
ecore_iov_get_link(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED vfid,struct ecore_mcp_link_params OSAL_UNUSED * params,struct ecore_mcp_link_state OSAL_UNUSED * link,struct ecore_mcp_link_capabilities OSAL_UNUSED * p_caps)771 static OSAL_INLINE void ecore_iov_get_link(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED vfid, struct ecore_mcp_link_params OSAL_UNUSED *params, struct ecore_mcp_link_state OSAL_UNUSED *link, struct ecore_mcp_link_capabilities OSAL_UNUSED *p_caps) {}
ecore_iov_is_vf_pending_flr(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id)772 static OSAL_INLINE bool ecore_iov_is_vf_pending_flr(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return false;}
773 #endif
774 static OSAL_INLINE bool
ecore_iov_is_valid_vfid(struct ecore_hwfn OSAL_UNUSED * p_hwfn,int OSAL_UNUSED rel_vf_id,bool OSAL_UNUSED b_enabled_only,bool OSAL_UNUSED b_non_malicious)775 ecore_iov_is_valid_vfid(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED rel_vf_id,
776 			bool OSAL_UNUSED b_enabled_only,
777 			bool OSAL_UNUSED b_non_malicious)
778 {
779 	return false;
780 }
781 #ifndef LINUX_REMOVE
ecore_iov_get_public_vf_info(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED vfid,bool OSAL_UNUSED b_enabled_only)782 static OSAL_INLINE struct ecore_public_vf_info* ecore_iov_get_public_vf_info(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED vfid, bool OSAL_UNUSED b_enabled_only) {return OSAL_NULL;}
ecore_iov_pf_add_pending_events(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u8 OSAL_UNUSED vfid)783 static OSAL_INLINE void ecore_iov_pf_add_pending_events(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u8 OSAL_UNUSED vfid) {}
ecore_iov_pf_get_and_clear_pending_events(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u64 OSAL_UNUSED * events)784 static OSAL_INLINE void ecore_iov_pf_get_and_clear_pending_events(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u64 OSAL_UNUSED *events) {}
ecore_iov_copy_vf_msg(struct ecore_hwfn OSAL_UNUSED * p_hwfn,struct ecore_ptt OSAL_UNUSED * ptt,int OSAL_UNUSED vfid)785 static OSAL_INLINE enum _ecore_status_t ecore_iov_copy_vf_msg(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *ptt, int OSAL_UNUSED vfid) {return ECORE_INVAL;}
ecore_iov_bulletin_set_forced_mac(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u8 OSAL_UNUSED * mac,int OSAL_UNUSED vfid)786 static OSAL_INLINE void ecore_iov_bulletin_set_forced_mac(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u8 OSAL_UNUSED *mac, int OSAL_UNUSED vfid) {}
ecore_iov_bulletin_set_mac(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u8 OSAL_UNUSED * mac,OSAL_UNUSED int vfid)787 static OSAL_INLINE enum _ecore_status_t ecore_iov_bulletin_set_mac(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u8 OSAL_UNUSED *mac, OSAL_UNUSED int vfid) {return ECORE_INVAL;}
ecore_iov_bulletin_set_forced_untagged_default(struct ecore_hwfn OSAL_UNUSED * p_hwfn,bool OSAL_UNUSED b_untagged_only,int OSAL_UNUSED vfid)788 static OSAL_INLINE enum _ecore_status_t ecore_iov_bulletin_set_forced_untagged_default(struct ecore_hwfn OSAL_UNUSED *p_hwfn, bool OSAL_UNUSED b_untagged_only, int OSAL_UNUSED vfid) {return ECORE_INVAL;}
ecore_iov_get_vfs_opaque_fid(struct ecore_hwfn OSAL_UNUSED * p_hwfn,int OSAL_UNUSED vfid,u16 OSAL_UNUSED * opaque_fid)789 static OSAL_INLINE void ecore_iov_get_vfs_opaque_fid(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid, u16 OSAL_UNUSED *opaque_fid) {}
ecore_iov_bulletin_set_forced_vlan(struct ecore_hwfn OSAL_UNUSED p_hwfn,u16 OSAL_UNUSED pvid,int OSAL_UNUSED vfid)790 static OSAL_INLINE void ecore_iov_bulletin_set_forced_vlan(struct ecore_hwfn OSAL_UNUSED p_hwfn, u16 OSAL_UNUSED pvid, int OSAL_UNUSED vfid) {}
ecore_iov_vf_has_vport_instance(struct ecore_hwfn OSAL_UNUSED * p_hwfn,int OSAL_UNUSED vfid)791 static OSAL_INLINE bool ecore_iov_vf_has_vport_instance(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) {return false;}
ecore_iov_post_vf_bulletin(struct ecore_hwfn OSAL_UNUSED * p_hwfn,int OSAL_UNUSED vfid,struct ecore_ptt OSAL_UNUSED * p_ptt)792 static OSAL_INLINE enum _ecore_status_t ecore_iov_post_vf_bulletin(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid, struct ecore_ptt OSAL_UNUSED *p_ptt) {return ECORE_INVAL;}
ecore_iov_is_vf_stopped(struct ecore_hwfn OSAL_UNUSED * p_hwfn,int OSAL_UNUSED vfid)793 static OSAL_INLINE bool ecore_iov_is_vf_stopped(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) {return false;}
ecore_iov_spoofchk_set(struct ecore_hwfn OSAL_UNUSED * p_hwfn,int OSAL_UNUSED vfid,bool OSAL_UNUSED val)794 static OSAL_INLINE enum _ecore_status_t ecore_iov_spoofchk_set(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid, bool OSAL_UNUSED val) {return ECORE_INVAL;}
ecore_iov_spoofchk_get(struct ecore_hwfn OSAL_UNUSED * p_hwfn,int OSAL_UNUSED vfid)795 static OSAL_INLINE bool ecore_iov_spoofchk_get(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) {return false;}
ecore_iov_pf_sanity_check(struct ecore_hwfn OSAL_UNUSED * p_hwfn,int OSAL_UNUSED vfid)796 static OSAL_INLINE bool ecore_iov_pf_sanity_check(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) {return false;}
ecore_iov_vf_chains_per_pf(struct ecore_hwfn OSAL_UNUSED * p_hwfn)797 static OSAL_INLINE u8 ecore_iov_vf_chains_per_pf(struct ecore_hwfn OSAL_UNUSED *p_hwfn) {return 0;}
ecore_iov_get_vf_req_virt_mbx_params(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id,void OSAL_UNUSED ** pp_req_virt_addr,u16 OSAL_UNUSED * p_req_virt_size)798 static OSAL_INLINE void ecore_iov_get_vf_req_virt_mbx_params(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id, void OSAL_UNUSED **pp_req_virt_addr, u16 OSAL_UNUSED *p_req_virt_size) {}
ecore_iov_get_vf_reply_virt_mbx_params(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id,void OSAL_UNUSED ** pp_reply_virt_addr,u16 OSAL_UNUSED * p_reply_virt_size)799 static OSAL_INLINE void ecore_iov_get_vf_reply_virt_mbx_params(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id, void OSAL_UNUSED **pp_reply_virt_addr, u16 OSAL_UNUSED *p_reply_virt_size) {}
ecore_iov_is_valid_vfpf_msg_length(u32 OSAL_UNUSED length)800 static OSAL_INLINE bool ecore_iov_is_valid_vfpf_msg_length(u32 OSAL_UNUSED length) {return false;}
ecore_iov_pfvf_msg_length(void)801 static OSAL_INLINE u32 ecore_iov_pfvf_msg_length(void) {return 0;}
ecore_iov_bulletin_get_forced_mac(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id)802 static OSAL_INLINE u8 *ecore_iov_bulletin_get_forced_mac(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return OSAL_NULL;}
ecore_iov_bulletin_get_forced_vlan(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id)803 static OSAL_INLINE u16 ecore_iov_bulletin_get_forced_vlan(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return 0;}
ecore_iov_configure_tx_rate(struct ecore_hwfn OSAL_UNUSED * p_hwfn,struct ecore_ptt OSAL_UNUSED * p_ptt,int OSAL_UNUSED vfid,int OSAL_UNUSED val)804 static OSAL_INLINE enum _ecore_status_t ecore_iov_configure_tx_rate(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, int OSAL_UNUSED vfid, int OSAL_UNUSED val) { return ECORE_INVAL; }
ecore_iov_get_vf_stats(struct ecore_hwfn OSAL_UNUSED * p_hwfn,struct ecore_ptt OSAL_UNUSED * p_ptt,int OSAL_UNUSED vfid,struct ecore_eth_stats OSAL_UNUSED * p_stats)805 static OSAL_INLINE enum _ecore_status_t ecore_iov_get_vf_stats(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, int OSAL_UNUSED vfid, struct ecore_eth_stats OSAL_UNUSED *p_stats) { return ECORE_INVAL; }
ecore_iov_get_vf_num_rxqs(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id)806 static OSAL_INLINE u8 ecore_iov_get_vf_num_rxqs(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return 0;}
ecore_iov_get_vf_num_active_rxqs(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id)807 static OSAL_INLINE u8 ecore_iov_get_vf_num_active_rxqs(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return 0;}
ecore_iov_get_vf_ctx(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id)808 static OSAL_INLINE void *ecore_iov_get_vf_ctx(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return OSAL_NULL;}
ecore_iov_get_vf_num_sbs(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id)809 static OSAL_INLINE u8 ecore_iov_get_vf_num_sbs(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return 0;}
ecore_iov_is_vf_wait_for_acquire(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id)810 static OSAL_INLINE bool ecore_iov_is_vf_wait_for_acquire(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return false;}
ecore_iov_is_vf_acquired_not_initialized(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id)811 static OSAL_INLINE bool ecore_iov_is_vf_acquired_not_initialized(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return false;}
ecore_iov_is_vf_initialized(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id)812 static OSAL_INLINE bool ecore_iov_is_vf_initialized(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return false;}
ecore_iov_get_vf_min_rate(struct ecore_hwfn OSAL_UNUSED * p_hwfn,int OSAL_UNUSED vfid)813 static OSAL_INLINE int ecore_iov_get_vf_min_rate(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) { return 0; }
ecore_iov_configure_min_tx_rate(struct ecore_dev OSAL_UNUSED * p_dev,int OSAL_UNUSED vfid,OSAL_UNUSED u32 rate)814 static OSAL_INLINE enum _ecore_status_t ecore_iov_configure_min_tx_rate(struct ecore_dev OSAL_UNUSED *p_dev, int OSAL_UNUSED vfid, OSAL_UNUSED u32 rate) { return ECORE_INVAL; }
815 #endif
ecore_iov_bulletin_set_udp_ports(struct ecore_hwfn OSAL_UNUSED * p_hwfn,int OSAL_UNUSED vfid,u16 OSAL_UNUSED vxlan_port,u16 OSAL_UNUSED geneve_port)816 static OSAL_INLINE void ecore_iov_bulletin_set_udp_ports(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid, u16 OSAL_UNUSED vxlan_port, u16 OSAL_UNUSED geneve_port) { return; }
ecore_iov_get_next_active_vf(struct ecore_hwfn OSAL_UNUSED * p_hwfn,u16 OSAL_UNUSED rel_vf_id)817 static OSAL_INLINE u16 ecore_iov_get_next_active_vf(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) { return MAX_NUM_VFS_E4; }
818 
819 #ifdef CONFIG_ECORE_SW_CHANNEL
820 static OSAL_INLINE void
ecore_iov_set_vf_hw_channel(struct ecore_hwfn OSAL_UNUSED * p_hwfn,int OSAL_UNUSED vfid,bool OSAL_UNUSED b_is_hw)821 ecore_iov_set_vf_hw_channel(struct ecore_hwfn OSAL_UNUSED *p_hwfn,
822 			    int OSAL_UNUSED vfid, bool OSAL_UNUSED b_is_hw) {}
823 #endif
824 #endif
825 
826 #define ecore_for_each_vf(_p_hwfn, _i)					\
827 	for (_i = ecore_iov_get_next_active_vf(_p_hwfn, 0);		\
828 	     _i < MAX_NUM_VFS_E4;					\
829 	     _i = ecore_iov_get_next_active_vf(_p_hwfn, _i + 1))
830 
831 #endif
832