xref: /freebsd/sys/dev/oce/oce_if.c (revision c03c5b1c)
1 /*-
2  * SPDX-License-Identifier: BSD-3-Clause
3  *
4  * Copyright (C) 2013 Emulex
5  * All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions are met:
9  *
10  * 1. Redistributions of source code must retain the above copyright notice,
11  *    this list of conditions and the following disclaimer.
12  *
13  * 2. Redistributions in binary form must reproduce the above copyright
14  *    notice, this list of conditions and the following disclaimer in the
15  *    documentation and/or other materials provided with the distribution.
16  *
17  * 3. Neither the name of the Emulex Corporation nor the names of its
18  *    contributors may be used to endorse or promote products derived from
19  *    this software without specific prior written permission.
20  *
21  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
22  * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
23  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
24  * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
25  * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
26  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
27  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
28  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
29  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
30  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
31  * POSSIBILITY OF SUCH DAMAGE.
32  *
33  * Contact Information:
34  * freebsd-drivers@emulex.com
35  *
36  * Emulex
37  * 3333 Susan Street
38  * Costa Mesa, CA 92626
39  */
40 
41 /* $FreeBSD$ */
42 
43 #include "opt_inet6.h"
44 #include "opt_inet.h"
45 
46 #include "oce_if.h"
47 #include "oce_user.h"
48 
49 #define is_tso_pkt(m) (m->m_pkthdr.csum_flags & CSUM_TSO)
50 
51 /* UE Status Low CSR */
52 static char *ue_status_low_desc[] = {
53         "CEV",
54         "CTX",
55         "DBUF",
56         "ERX",
57         "Host",
58         "MPU",
59         "NDMA",
60         "PTC ",
61         "RDMA ",
62         "RXF ",
63         "RXIPS ",
64         "RXULP0 ",
65         "RXULP1 ",
66         "RXULP2 ",
67         "TIM ",
68         "TPOST ",
69         "TPRE ",
70         "TXIPS ",
71         "TXULP0 ",
72         "TXULP1 ",
73         "UC ",
74         "WDMA ",
75         "TXULP2 ",
76         "HOST1 ",
77         "P0_OB_LINK ",
78         "P1_OB_LINK ",
79         "HOST_GPIO ",
80         "MBOX ",
81         "AXGMAC0",
82         "AXGMAC1",
83         "JTAG",
84         "MPU_INTPEND"
85 };
86 
87 /* UE Status High CSR */
88 static char *ue_status_hi_desc[] = {
89         "LPCMEMHOST",
90         "MGMT_MAC",
91         "PCS0ONLINE",
92         "MPU_IRAM",
93         "PCS1ONLINE",
94         "PCTL0",
95         "PCTL1",
96         "PMEM",
97         "RR",
98         "TXPB",
99         "RXPP",
100         "XAUI",
101         "TXP",
102         "ARM",
103         "IPC",
104         "HOST2",
105         "HOST3",
106         "HOST4",
107         "HOST5",
108         "HOST6",
109         "HOST7",
110         "HOST8",
111         "HOST9",
112         "NETC",
113         "Unknown",
114         "Unknown",
115         "Unknown",
116         "Unknown",
117         "Unknown",
118         "Unknown",
119         "Unknown",
120         "Unknown"
121 };
122 
123 struct oce_common_cqe_info{
124         uint8_t vtp:1;
125         uint8_t l4_cksum_pass:1;
126         uint8_t ip_cksum_pass:1;
127         uint8_t ipv6_frame:1;
128         uint8_t qnq:1;
129         uint8_t rsvd:3;
130         uint8_t num_frags;
131         uint16_t pkt_size;
132         uint16_t vtag;
133 };
134 
135 /* Driver entry points prototypes */
136 static int  oce_probe(device_t dev);
137 static int  oce_attach(device_t dev);
138 static int  oce_detach(device_t dev);
139 static int  oce_shutdown(device_t dev);
140 static int  oce_ioctl(struct ifnet *ifp, u_long command, caddr_t data);
141 static void oce_init(void *xsc);
142 static int  oce_multiq_start(struct ifnet *ifp, struct mbuf *m);
143 static void oce_multiq_flush(struct ifnet *ifp);
144 
145 /* Driver interrupt routines protypes */
146 static void oce_intr(void *arg, int pending);
147 static int  oce_setup_intr(POCE_SOFTC sc);
148 static int  oce_fast_isr(void *arg);
149 static int  oce_alloc_intr(POCE_SOFTC sc, int vector,
150 			  void (*isr) (void *arg, int pending));
151 
152 /* Media callbacks prototypes */
153 static void oce_media_status(struct ifnet *ifp, struct ifmediareq *req);
154 static int  oce_media_change(struct ifnet *ifp);
155 
156 /* Transmit routines prototypes */
157 static int  oce_tx(POCE_SOFTC sc, struct mbuf **mpp, int wq_index);
158 static void oce_tx_restart(POCE_SOFTC sc, struct oce_wq *wq);
159 static void oce_process_tx_completion(struct oce_wq *wq);
160 static int  oce_multiq_transmit(struct ifnet *ifp, struct mbuf *m,
161 				 struct oce_wq *wq);
162 
163 /* Receive routines prototypes */
164 static int  oce_cqe_vtp_valid(POCE_SOFTC sc, struct oce_nic_rx_cqe *cqe);
165 static int  oce_cqe_portid_valid(POCE_SOFTC sc, struct oce_nic_rx_cqe *cqe);
166 static void oce_rx(struct oce_rq *rq, struct oce_nic_rx_cqe *cqe);
167 static void oce_check_rx_bufs(POCE_SOFTC sc, uint32_t num_cqes, struct oce_rq *rq);
168 static uint16_t oce_rq_handler_lro(void *arg);
169 static void oce_correct_header(struct mbuf *m, struct nic_hwlro_cqe_part1 *cqe1, struct nic_hwlro_cqe_part2 *cqe2);
170 static void oce_rx_lro(struct oce_rq *rq, struct nic_hwlro_singleton_cqe *cqe, struct nic_hwlro_cqe_part2 *cqe2);
171 static void oce_rx_mbuf_chain(struct oce_rq *rq, struct oce_common_cqe_info *cqe_info, struct mbuf **m);
172 
173 /* Helper function prototypes in this file */
174 static int  oce_attach_ifp(POCE_SOFTC sc);
175 static void oce_add_vlan(void *arg, struct ifnet *ifp, uint16_t vtag);
176 static void oce_del_vlan(void *arg, struct ifnet *ifp, uint16_t vtag);
177 static int  oce_vid_config(POCE_SOFTC sc);
178 static void oce_mac_addr_set(POCE_SOFTC sc);
179 static int  oce_handle_passthrough(struct ifnet *ifp, caddr_t data);
180 static void oce_local_timer(void *arg);
181 static void oce_if_deactivate(POCE_SOFTC sc);
182 static void oce_if_activate(POCE_SOFTC sc);
183 static void setup_max_queues_want(POCE_SOFTC sc);
184 static void update_queues_got(POCE_SOFTC sc);
185 static void process_link_state(POCE_SOFTC sc,
186 		 struct oce_async_cqe_link_state *acqe);
187 static int oce_tx_asic_stall_verify(POCE_SOFTC sc, struct mbuf *m);
188 static void oce_get_config(POCE_SOFTC sc);
189 static struct mbuf *oce_insert_vlan_tag(POCE_SOFTC sc, struct mbuf *m, boolean_t *complete);
190 static void oce_read_env_variables(POCE_SOFTC sc);
191 
192 /* IP specific */
193 #if defined(INET6) || defined(INET)
194 static int  oce_init_lro(POCE_SOFTC sc);
195 static struct mbuf * oce_tso_setup(POCE_SOFTC sc, struct mbuf **mpp);
196 #endif
197 
198 static device_method_t oce_dispatch[] = {
199 	DEVMETHOD(device_probe, oce_probe),
200 	DEVMETHOD(device_attach, oce_attach),
201 	DEVMETHOD(device_detach, oce_detach),
202 	DEVMETHOD(device_shutdown, oce_shutdown),
203 
204 	DEVMETHOD_END
205 };
206 
207 static driver_t oce_driver = {
208 	"oce",
209 	oce_dispatch,
210 	sizeof(OCE_SOFTC)
211 };
212 static devclass_t oce_devclass;
213 
214 /* global vars */
215 const char component_revision[32] = {"///" COMPONENT_REVISION "///"};
216 
217 /* Module capabilites and parameters */
218 uint32_t oce_max_rsp_handled = OCE_MAX_RSP_HANDLED;
219 uint32_t oce_enable_rss = OCE_MODCAP_RSS;
220 uint32_t oce_rq_buf_size = 2048;
221 
222 TUNABLE_INT("hw.oce.max_rsp_handled", &oce_max_rsp_handled);
223 TUNABLE_INT("hw.oce.enable_rss", &oce_enable_rss);
224 
225 /* Supported devices table */
226 static uint32_t supportedDevices[] =  {
227 	(PCI_VENDOR_SERVERENGINES << 16) | PCI_PRODUCT_BE2,
228 	(PCI_VENDOR_SERVERENGINES << 16) | PCI_PRODUCT_BE3,
229 	(PCI_VENDOR_EMULEX << 16) | PCI_PRODUCT_BE3,
230 	(PCI_VENDOR_EMULEX << 16) | PCI_PRODUCT_XE201,
231 	(PCI_VENDOR_EMULEX << 16) | PCI_PRODUCT_XE201_VF,
232 	(PCI_VENDOR_EMULEX << 16) | PCI_PRODUCT_SH
233 };
234 
235 DRIVER_MODULE(oce, pci, oce_driver, oce_devclass, 0, 0);
236 MODULE_PNP_INFO("W32:vendor/device", pci, oce, supportedDevices,
237     nitems(supportedDevices));
238 MODULE_DEPEND(oce, pci, 1, 1, 1);
239 MODULE_DEPEND(oce, ether, 1, 1, 1);
240 MODULE_VERSION(oce, 1);
241 
242 POCE_SOFTC softc_head = NULL;
243 POCE_SOFTC softc_tail = NULL;
244 
245 struct oce_rdma_if *oce_rdma_if = NULL;
246 
247 /*****************************************************************************
248  *			Driver entry points functions                        *
249  *****************************************************************************/
250 
251 static int
252 oce_probe(device_t dev)
253 {
254 	uint16_t vendor = 0;
255 	uint16_t device = 0;
256 	int i = 0;
257 	char str[256] = {0};
258 	POCE_SOFTC sc;
259 
260 	sc = device_get_softc(dev);
261 	bzero(sc, sizeof(OCE_SOFTC));
262 	sc->dev = dev;
263 
264 	vendor = pci_get_vendor(dev);
265 	device = pci_get_device(dev);
266 
267 	for (i = 0; i < (sizeof(supportedDevices) / sizeof(uint32_t)); i++) {
268 		if (vendor == ((supportedDevices[i] >> 16) & 0xffff)) {
269 			if (device == (supportedDevices[i] & 0xffff)) {
270 				sprintf(str, "%s:%s", "Emulex CNA NIC function",
271 					component_revision);
272 				device_set_desc_copy(dev, str);
273 
274 				switch (device) {
275 				case PCI_PRODUCT_BE2:
276 					sc->flags |= OCE_FLAGS_BE2;
277 					break;
278 				case PCI_PRODUCT_BE3:
279 					sc->flags |= OCE_FLAGS_BE3;
280 					break;
281 				case PCI_PRODUCT_XE201:
282 				case PCI_PRODUCT_XE201_VF:
283 					sc->flags |= OCE_FLAGS_XE201;
284 					break;
285 				case PCI_PRODUCT_SH:
286 					sc->flags |= OCE_FLAGS_SH;
287 					break;
288 				default:
289 					return ENXIO;
290 				}
291 				return BUS_PROBE_DEFAULT;
292 			}
293 		}
294 	}
295 
296 	return ENXIO;
297 }
298 
299 static int
300 oce_attach(device_t dev)
301 {
302 	POCE_SOFTC sc;
303 	int rc = 0;
304 
305 	sc = device_get_softc(dev);
306 
307 	rc = oce_hw_pci_alloc(sc);
308 	if (rc)
309 		return rc;
310 
311 	sc->tx_ring_size = OCE_TX_RING_SIZE;
312 	sc->rx_ring_size = OCE_RX_RING_SIZE;
313 	/* receive fragment size should be multiple of 2K */
314 	sc->rq_frag_size = ((oce_rq_buf_size / 2048) * 2048);
315 	sc->flow_control = OCE_DEFAULT_FLOW_CONTROL;
316 	sc->promisc	 = OCE_DEFAULT_PROMISCUOUS;
317 
318 	LOCK_CREATE(&sc->bmbx_lock, "Mailbox_lock");
319 	LOCK_CREATE(&sc->dev_lock,  "Device_lock");
320 
321 	/* initialise the hardware */
322 	rc = oce_hw_init(sc);
323 	if (rc)
324 		goto pci_res_free;
325 
326 	oce_read_env_variables(sc);
327 
328 	oce_get_config(sc);
329 
330 	setup_max_queues_want(sc);
331 
332 	rc = oce_setup_intr(sc);
333 	if (rc)
334 		goto mbox_free;
335 
336 	rc = oce_queue_init_all(sc);
337 	if (rc)
338 		goto intr_free;
339 
340 	rc = oce_attach_ifp(sc);
341 	if (rc)
342 		goto queues_free;
343 
344 #if defined(INET6) || defined(INET)
345 	rc = oce_init_lro(sc);
346 	if (rc)
347 		goto ifp_free;
348 #endif
349 
350 	rc = oce_hw_start(sc);
351 	if (rc)
352 		goto lro_free;
353 
354 	sc->vlan_attach = EVENTHANDLER_REGISTER(vlan_config,
355 				oce_add_vlan, sc, EVENTHANDLER_PRI_FIRST);
356 	sc->vlan_detach = EVENTHANDLER_REGISTER(vlan_unconfig,
357 				oce_del_vlan, sc, EVENTHANDLER_PRI_FIRST);
358 
359 	rc = oce_stats_init(sc);
360 	if (rc)
361 		goto vlan_free;
362 
363 	oce_add_sysctls(sc);
364 
365 	callout_init(&sc->timer, CALLOUT_MPSAFE);
366 	rc = callout_reset(&sc->timer, 2 * hz, oce_local_timer, sc);
367 	if (rc)
368 		goto stats_free;
369 
370 	sc->next =NULL;
371 	if (softc_tail != NULL) {
372 	  softc_tail->next = sc;
373 	} else {
374 	  softc_head = sc;
375 	}
376 	softc_tail = sc;
377 
378 	return 0;
379 
380 stats_free:
381 	callout_drain(&sc->timer);
382 	oce_stats_free(sc);
383 vlan_free:
384 	if (sc->vlan_attach)
385 		EVENTHANDLER_DEREGISTER(vlan_config, sc->vlan_attach);
386 	if (sc->vlan_detach)
387 		EVENTHANDLER_DEREGISTER(vlan_unconfig, sc->vlan_detach);
388 	oce_hw_intr_disable(sc);
389 lro_free:
390 #if defined(INET6) || defined(INET)
391 	oce_free_lro(sc);
392 ifp_free:
393 #endif
394 	ether_ifdetach(sc->ifp);
395 	if_free(sc->ifp);
396 queues_free:
397 	oce_queue_release_all(sc);
398 intr_free:
399 	oce_intr_free(sc);
400 mbox_free:
401 	oce_dma_free(sc, &sc->bsmbx);
402 pci_res_free:
403 	oce_hw_pci_free(sc);
404 	LOCK_DESTROY(&sc->dev_lock);
405 	LOCK_DESTROY(&sc->bmbx_lock);
406 	return rc;
407 
408 }
409 
410 static int
411 oce_detach(device_t dev)
412 {
413 	POCE_SOFTC sc = device_get_softc(dev);
414 	POCE_SOFTC poce_sc_tmp, *ppoce_sc_tmp1, poce_sc_tmp2 = NULL;
415 
416         poce_sc_tmp = softc_head;
417         ppoce_sc_tmp1 = &softc_head;
418         while (poce_sc_tmp != NULL) {
419           if (poce_sc_tmp == sc) {
420             *ppoce_sc_tmp1 = sc->next;
421             if (sc->next == NULL) {
422               softc_tail = poce_sc_tmp2;
423             }
424             break;
425           }
426           poce_sc_tmp2 = poce_sc_tmp;
427           ppoce_sc_tmp1 = &poce_sc_tmp->next;
428           poce_sc_tmp = poce_sc_tmp->next;
429         }
430 
431 	LOCK(&sc->dev_lock);
432 	oce_if_deactivate(sc);
433 	UNLOCK(&sc->dev_lock);
434 
435 	callout_drain(&sc->timer);
436 
437 	if (sc->vlan_attach != NULL)
438 		EVENTHANDLER_DEREGISTER(vlan_config, sc->vlan_attach);
439 	if (sc->vlan_detach != NULL)
440 		EVENTHANDLER_DEREGISTER(vlan_unconfig, sc->vlan_detach);
441 
442 	ether_ifdetach(sc->ifp);
443 
444 	if_free(sc->ifp);
445 
446 	oce_hw_shutdown(sc);
447 
448 	bus_generic_detach(dev);
449 
450 	return 0;
451 }
452 
453 static int
454 oce_shutdown(device_t dev)
455 {
456 	int rc;
457 
458 	rc = oce_detach(dev);
459 
460 	return rc;
461 }
462 
463 static int
464 oce_ioctl(struct ifnet *ifp, u_long command, caddr_t data)
465 {
466 	struct ifreq *ifr = (struct ifreq *)data;
467 	POCE_SOFTC sc = ifp->if_softc;
468 	struct ifi2creq i2c;
469 	uint8_t	offset = 0;
470 	int rc = 0;
471 	uint32_t u;
472 
473 	switch (command) {
474 	case SIOCGIFMEDIA:
475 		rc = ifmedia_ioctl(ifp, ifr, &sc->media, command);
476 		break;
477 
478 	case SIOCSIFMTU:
479 		if (ifr->ifr_mtu > OCE_MAX_MTU)
480 			rc = EINVAL;
481 		else
482 			ifp->if_mtu = ifr->ifr_mtu;
483 		break;
484 
485 	case SIOCSIFFLAGS:
486 		if (ifp->if_flags & IFF_UP) {
487 			if (!(ifp->if_drv_flags & IFF_DRV_RUNNING)) {
488 				sc->ifp->if_drv_flags |= IFF_DRV_RUNNING;
489 				oce_init(sc);
490 			}
491 			device_printf(sc->dev, "Interface Up\n");
492 		} else {
493 			LOCK(&sc->dev_lock);
494 
495 			sc->ifp->if_drv_flags &=
496 			    ~(IFF_DRV_RUNNING | IFF_DRV_OACTIVE);
497 			oce_if_deactivate(sc);
498 
499 			UNLOCK(&sc->dev_lock);
500 
501 			device_printf(sc->dev, "Interface Down\n");
502 		}
503 
504 		if ((ifp->if_flags & IFF_PROMISC) && !sc->promisc) {
505 			if (!oce_rxf_set_promiscuous(sc, (1 | (1 << 1))))
506 				sc->promisc = TRUE;
507 		} else if (!(ifp->if_flags & IFF_PROMISC) && sc->promisc) {
508 			if (!oce_rxf_set_promiscuous(sc, 0))
509 				sc->promisc = FALSE;
510 		}
511 
512 		break;
513 
514 	case SIOCADDMULTI:
515 	case SIOCDELMULTI:
516 		rc = oce_hw_update_multicast(sc);
517 		if (rc)
518 			device_printf(sc->dev,
519 				"Update multicast address failed\n");
520 		break;
521 
522 	case SIOCSIFCAP:
523 		u = ifr->ifr_reqcap ^ ifp->if_capenable;
524 
525 		if (u & IFCAP_TXCSUM) {
526 			ifp->if_capenable ^= IFCAP_TXCSUM;
527 			ifp->if_hwassist ^= (CSUM_TCP | CSUM_UDP | CSUM_IP);
528 
529 			if (IFCAP_TSO & ifp->if_capenable &&
530 			    !(IFCAP_TXCSUM & ifp->if_capenable)) {
531 				u &= ~IFCAP_TSO;
532 				ifp->if_capenable &= ~IFCAP_TSO;
533 				ifp->if_hwassist &= ~CSUM_TSO;
534 				if_printf(ifp,
535 					 "TSO disabled due to -txcsum.\n");
536 			}
537 		}
538 
539 		if (u & IFCAP_RXCSUM)
540 			ifp->if_capenable ^= IFCAP_RXCSUM;
541 
542 		if (u & IFCAP_TSO4) {
543 			ifp->if_capenable ^= IFCAP_TSO4;
544 
545 			if (IFCAP_TSO & ifp->if_capenable) {
546 				if (IFCAP_TXCSUM & ifp->if_capenable)
547 					ifp->if_hwassist |= CSUM_TSO;
548 				else {
549 					ifp->if_capenable &= ~IFCAP_TSO;
550 					ifp->if_hwassist &= ~CSUM_TSO;
551 					if_printf(ifp,
552 					    "Enable txcsum first.\n");
553 					rc = EAGAIN;
554 				}
555 			} else
556 				ifp->if_hwassist &= ~CSUM_TSO;
557 		}
558 
559 		if (u & IFCAP_VLAN_HWTAGGING)
560 			ifp->if_capenable ^= IFCAP_VLAN_HWTAGGING;
561 
562 		if (u & IFCAP_VLAN_HWFILTER) {
563 			ifp->if_capenable ^= IFCAP_VLAN_HWFILTER;
564 			oce_vid_config(sc);
565 		}
566 #if defined(INET6) || defined(INET)
567 		if (u & IFCAP_LRO) {
568 			ifp->if_capenable ^= IFCAP_LRO;
569 			if(sc->enable_hwlro) {
570 				if(ifp->if_capenable & IFCAP_LRO) {
571 					rc = oce_mbox_nic_set_iface_lro_config(sc, 1);
572 				}else {
573 					rc = oce_mbox_nic_set_iface_lro_config(sc, 0);
574 				}
575 			}
576 		}
577 #endif
578 
579 		break;
580 
581 	case SIOCGI2C:
582 		rc = copyin(ifr_data_get_ptr(ifr), &i2c, sizeof(i2c));
583 		if (rc)
584 			break;
585 
586 		if (i2c.dev_addr == PAGE_NUM_A0) {
587 			offset = i2c.offset;
588 		} else if (i2c.dev_addr == PAGE_NUM_A2) {
589 			offset = TRANSCEIVER_A0_SIZE + i2c.offset;
590 		} else {
591 			rc = EINVAL;
592 			break;
593 		}
594 
595 		if (i2c.len > sizeof(i2c.data) ||
596 		    i2c.len + offset > sizeof(sfp_vpd_dump_buffer)) {
597 			rc = EINVAL;
598 			break;
599 		}
600 
601 		rc = oce_mbox_read_transrecv_data(sc, i2c.dev_addr);
602 		if (rc) {
603 			rc = -rc;
604 			break;
605 		}
606 
607 		memcpy(&i2c.data[0], &sfp_vpd_dump_buffer[offset], i2c.len);
608 
609 		rc = copyout(&i2c, ifr_data_get_ptr(ifr), sizeof(i2c));
610 		break;
611 
612 	case SIOCGPRIVATE_0:
613 		rc = priv_check(curthread, PRIV_DRIVER);
614 		if (rc != 0)
615 			break;
616 		rc = oce_handle_passthrough(ifp, data);
617 		break;
618 	default:
619 		rc = ether_ioctl(ifp, command, data);
620 		break;
621 	}
622 
623 	return rc;
624 }
625 
626 static void
627 oce_init(void *arg)
628 {
629 	POCE_SOFTC sc = arg;
630 
631 	LOCK(&sc->dev_lock);
632 
633 	if (sc->ifp->if_flags & IFF_UP) {
634 		oce_if_deactivate(sc);
635 		oce_if_activate(sc);
636 	}
637 
638 	UNLOCK(&sc->dev_lock);
639 
640 }
641 
642 static int
643 oce_multiq_start(struct ifnet *ifp, struct mbuf *m)
644 {
645 	POCE_SOFTC sc = ifp->if_softc;
646 	struct oce_wq *wq = NULL;
647 	int queue_index = 0;
648 	int status = 0;
649 
650 	if (!sc->link_status)
651 		return ENXIO;
652 
653 	if (M_HASHTYPE_GET(m) != M_HASHTYPE_NONE)
654 		queue_index = m->m_pkthdr.flowid % sc->nwqs;
655 
656 	wq = sc->wq[queue_index];
657 
658 	LOCK(&wq->tx_lock);
659 	status = oce_multiq_transmit(ifp, m, wq);
660 	UNLOCK(&wq->tx_lock);
661 
662 	return status;
663 
664 }
665 
666 static void
667 oce_multiq_flush(struct ifnet *ifp)
668 {
669 	POCE_SOFTC sc = ifp->if_softc;
670 	struct mbuf     *m;
671 	int i = 0;
672 
673 	for (i = 0; i < sc->nwqs; i++) {
674 		while ((m = buf_ring_dequeue_sc(sc->wq[i]->br)) != NULL)
675 			m_freem(m);
676 	}
677 	if_qflush(ifp);
678 }
679 
680 /*****************************************************************************
681  *                   Driver interrupt routines functions                     *
682  *****************************************************************************/
683 
684 static void
685 oce_intr(void *arg, int pending)
686 {
687 
688 	POCE_INTR_INFO ii = (POCE_INTR_INFO) arg;
689 	POCE_SOFTC sc = ii->sc;
690 	struct oce_eq *eq = ii->eq;
691 	struct oce_eqe *eqe;
692 	struct oce_cq *cq = NULL;
693 	int i, num_eqes = 0;
694 
695 	bus_dmamap_sync(eq->ring->dma.tag, eq->ring->dma.map,
696 				 BUS_DMASYNC_POSTWRITE);
697 	do {
698 		eqe = RING_GET_CONSUMER_ITEM_VA(eq->ring, struct oce_eqe);
699 		if (eqe->evnt == 0)
700 			break;
701 		eqe->evnt = 0;
702 		bus_dmamap_sync(eq->ring->dma.tag, eq->ring->dma.map,
703 					BUS_DMASYNC_POSTWRITE);
704 		RING_GET(eq->ring, 1);
705 		num_eqes++;
706 
707 	} while (TRUE);
708 
709 	if (!num_eqes)
710 		goto eq_arm; /* Spurious */
711 
712  	/* Clear EQ entries, but dont arm */
713 	oce_arm_eq(sc, eq->eq_id, num_eqes, FALSE, FALSE);
714 
715 	/* Process TX, RX and MCC. But dont arm CQ*/
716 	for (i = 0; i < eq->cq_valid; i++) {
717 		cq = eq->cq[i];
718 		(*cq->cq_handler)(cq->cb_arg);
719 	}
720 
721 	/* Arm all cqs connected to this EQ */
722 	for (i = 0; i < eq->cq_valid; i++) {
723 		cq = eq->cq[i];
724 		oce_arm_cq(sc, cq->cq_id, 0, TRUE);
725 	}
726 
727 eq_arm:
728 	oce_arm_eq(sc, eq->eq_id, 0, TRUE, FALSE);
729 
730 	return;
731 }
732 
733 static int
734 oce_setup_intr(POCE_SOFTC sc)
735 {
736 	int rc = 0, use_intx = 0;
737 	int vector = 0, req_vectors = 0;
738 	int tot_req_vectors, tot_vectors;
739 
740 	if (is_rss_enabled(sc))
741 		req_vectors = MAX((sc->nrqs - 1), sc->nwqs);
742 	else
743 		req_vectors = 1;
744 
745 	tot_req_vectors = req_vectors;
746 	if (sc->rdma_flags & OCE_RDMA_FLAG_SUPPORTED) {
747 	  if (req_vectors > 1) {
748 	    tot_req_vectors += OCE_RDMA_VECTORS;
749 	    sc->roce_intr_count = OCE_RDMA_VECTORS;
750 	  }
751 	}
752 
753         if (sc->flags & OCE_FLAGS_MSIX_CAPABLE) {
754 		sc->intr_count = req_vectors;
755                 tot_vectors = tot_req_vectors;
756 		rc = pci_alloc_msix(sc->dev, &tot_vectors);
757 		if (rc != 0) {
758 			use_intx = 1;
759 			pci_release_msi(sc->dev);
760 		} else {
761 		  if (sc->rdma_flags & OCE_RDMA_FLAG_SUPPORTED) {
762 		    if (tot_vectors < tot_req_vectors) {
763 		      if (sc->intr_count < (2 * OCE_RDMA_VECTORS)) {
764 			sc->roce_intr_count = (tot_vectors / 2);
765 		      }
766 		      sc->intr_count = tot_vectors - sc->roce_intr_count;
767 		    }
768 		  } else {
769 		    sc->intr_count = tot_vectors;
770 		  }
771     		  sc->flags |= OCE_FLAGS_USING_MSIX;
772 		}
773 	} else
774 		use_intx = 1;
775 
776 	if (use_intx)
777 		sc->intr_count = 1;
778 
779 	/* Scale number of queues based on intr we got */
780 	update_queues_got(sc);
781 
782 	if (use_intx) {
783 		device_printf(sc->dev, "Using legacy interrupt\n");
784 		rc = oce_alloc_intr(sc, vector, oce_intr);
785 		if (rc)
786 			goto error;
787 	} else {
788 		for (; vector < sc->intr_count; vector++) {
789 			rc = oce_alloc_intr(sc, vector, oce_intr);
790 			if (rc)
791 				goto error;
792 		}
793 	}
794 
795 	return 0;
796 error:
797 	oce_intr_free(sc);
798 	return rc;
799 }
800 
801 static int
802 oce_fast_isr(void *arg)
803 {
804 	POCE_INTR_INFO ii = (POCE_INTR_INFO) arg;
805 	POCE_SOFTC sc = ii->sc;
806 
807 	if (ii->eq == NULL)
808 		return FILTER_STRAY;
809 
810 	oce_arm_eq(sc, ii->eq->eq_id, 0, FALSE, TRUE);
811 
812 	taskqueue_enqueue(ii->tq, &ii->task);
813 
814  	ii->eq->intr++;
815 
816 	return FILTER_HANDLED;
817 }
818 
819 static int
820 oce_alloc_intr(POCE_SOFTC sc, int vector, void (*isr) (void *arg, int pending))
821 {
822 	POCE_INTR_INFO ii;
823 	int rc = 0, rr;
824 
825 	if (vector >= OCE_MAX_EQ)
826 		return (EINVAL);
827 
828 	ii = &sc->intrs[vector];
829 
830 	/* Set the resource id for the interrupt.
831 	 * MSIx is vector + 1 for the resource id,
832 	 * INTx is 0 for the resource id.
833 	 */
834 	if (sc->flags & OCE_FLAGS_USING_MSIX)
835 		rr = vector + 1;
836 	else
837 		rr = 0;
838 	ii->intr_res = bus_alloc_resource_any(sc->dev,
839 					      SYS_RES_IRQ,
840 					      &rr, RF_ACTIVE|RF_SHAREABLE);
841 	ii->irq_rr = rr;
842 	if (ii->intr_res == NULL) {
843 		device_printf(sc->dev,
844 			  "Could not allocate interrupt\n");
845 		rc = ENXIO;
846 		return rc;
847 	}
848 
849 	TASK_INIT(&ii->task, 0, isr, ii);
850 	ii->vector = vector;
851 	sprintf(ii->task_name, "oce_task[%d]", ii->vector);
852 	ii->tq = taskqueue_create_fast(ii->task_name,
853 			M_NOWAIT,
854 			taskqueue_thread_enqueue,
855 			&ii->tq);
856 	taskqueue_start_threads(&ii->tq, 1, PI_NET, "%s taskq",
857 			device_get_nameunit(sc->dev));
858 
859 	ii->sc = sc;
860 	rc = bus_setup_intr(sc->dev,
861 			ii->intr_res,
862 			INTR_TYPE_NET,
863 			oce_fast_isr, NULL, ii, &ii->tag);
864 	return rc;
865 
866 }
867 
868 void
869 oce_intr_free(POCE_SOFTC sc)
870 {
871 	int i = 0;
872 
873 	for (i = 0; i < sc->intr_count; i++) {
874 
875 		if (sc->intrs[i].tag != NULL)
876 			bus_teardown_intr(sc->dev, sc->intrs[i].intr_res,
877 						sc->intrs[i].tag);
878 		if (sc->intrs[i].tq != NULL)
879 			taskqueue_free(sc->intrs[i].tq);
880 
881 		if (sc->intrs[i].intr_res != NULL)
882 			bus_release_resource(sc->dev, SYS_RES_IRQ,
883 						sc->intrs[i].irq_rr,
884 						sc->intrs[i].intr_res);
885 		sc->intrs[i].tag = NULL;
886 		sc->intrs[i].intr_res = NULL;
887 	}
888 
889 	if (sc->flags & OCE_FLAGS_USING_MSIX)
890 		pci_release_msi(sc->dev);
891 
892 }
893 
894 /******************************************************************************
895 *			  Media callbacks functions 			      *
896 ******************************************************************************/
897 
898 static void
899 oce_media_status(struct ifnet *ifp, struct ifmediareq *req)
900 {
901 	POCE_SOFTC sc = (POCE_SOFTC) ifp->if_softc;
902 
903 	req->ifm_status = IFM_AVALID;
904 	req->ifm_active = IFM_ETHER;
905 
906 	if (sc->link_status == 1)
907 		req->ifm_status |= IFM_ACTIVE;
908 	else
909 		return;
910 
911 	switch (sc->link_speed) {
912 	case 1: /* 10 Mbps */
913 		req->ifm_active |= IFM_10_T | IFM_FDX;
914 		sc->speed = 10;
915 		break;
916 	case 2: /* 100 Mbps */
917 		req->ifm_active |= IFM_100_TX | IFM_FDX;
918 		sc->speed = 100;
919 		break;
920 	case 3: /* 1 Gbps */
921 		req->ifm_active |= IFM_1000_T | IFM_FDX;
922 		sc->speed = 1000;
923 		break;
924 	case 4: /* 10 Gbps */
925 		req->ifm_active |= IFM_10G_SR | IFM_FDX;
926 		sc->speed = 10000;
927 		break;
928 	case 5: /* 20 Gbps */
929 		req->ifm_active |= IFM_10G_SR | IFM_FDX;
930 		sc->speed = 20000;
931 		break;
932 	case 6: /* 25 Gbps */
933 		req->ifm_active |= IFM_10G_SR | IFM_FDX;
934 		sc->speed = 25000;
935 		break;
936 	case 7: /* 40 Gbps */
937 		req->ifm_active |= IFM_40G_SR4 | IFM_FDX;
938 		sc->speed = 40000;
939 		break;
940 	default:
941 		sc->speed = 0;
942 		break;
943 	}
944 
945 	return;
946 }
947 
948 int
949 oce_media_change(struct ifnet *ifp)
950 {
951 	return 0;
952 }
953 
954 static void oce_is_pkt_dest_bmc(POCE_SOFTC sc,
955 				struct mbuf *m, boolean_t *os2bmc,
956 				struct mbuf **m_new)
957 {
958 	struct ether_header *eh = NULL;
959 
960 	eh = mtod(m, struct ether_header *);
961 
962 	if (!is_os2bmc_enabled(sc) || *os2bmc) {
963 		*os2bmc = FALSE;
964 		goto done;
965 	}
966 	if (!ETHER_IS_MULTICAST(eh->ether_dhost))
967 		goto done;
968 
969 	if (is_mc_allowed_on_bmc(sc, eh) ||
970 	    is_bc_allowed_on_bmc(sc, eh) ||
971 	    is_arp_allowed_on_bmc(sc, ntohs(eh->ether_type))) {
972 		*os2bmc = TRUE;
973 		goto done;
974 	}
975 
976 	if (mtod(m, struct ip *)->ip_p == IPPROTO_IPV6) {
977 		struct ip6_hdr *ip6 = mtod(m, struct ip6_hdr *);
978 		uint8_t nexthdr = ip6->ip6_nxt;
979 		if (nexthdr == IPPROTO_ICMPV6) {
980 			struct icmp6_hdr *icmp6 = (struct icmp6_hdr *)(ip6 + 1);
981 			switch (icmp6->icmp6_type) {
982 			case ND_ROUTER_ADVERT:
983 				*os2bmc = is_ipv6_ra_filt_enabled(sc);
984 				goto done;
985 			case ND_NEIGHBOR_ADVERT:
986 				*os2bmc = is_ipv6_na_filt_enabled(sc);
987 				goto done;
988 			default:
989 				break;
990 			}
991 		}
992 	}
993 
994 	if (mtod(m, struct ip *)->ip_p == IPPROTO_UDP) {
995 		struct ip *ip = mtod(m, struct ip *);
996 		int iphlen = ip->ip_hl << 2;
997 		struct udphdr *uh = (struct udphdr *)((caddr_t)ip + iphlen);
998 		switch (uh->uh_dport) {
999 		case DHCP_CLIENT_PORT:
1000 			*os2bmc = is_dhcp_client_filt_enabled(sc);
1001 			goto done;
1002 		case DHCP_SERVER_PORT:
1003 			*os2bmc = is_dhcp_srvr_filt_enabled(sc);
1004 			goto done;
1005 		case NET_BIOS_PORT1:
1006 		case NET_BIOS_PORT2:
1007 			*os2bmc = is_nbios_filt_enabled(sc);
1008 			goto done;
1009 		case DHCPV6_RAS_PORT:
1010 			*os2bmc = is_ipv6_ras_filt_enabled(sc);
1011 			goto done;
1012 		default:
1013 			break;
1014 		}
1015 	}
1016 done:
1017 	if (*os2bmc) {
1018 		*m_new = m_dup(m, M_NOWAIT);
1019 		if (!*m_new) {
1020 			*os2bmc = FALSE;
1021 			return;
1022 		}
1023 		*m_new = oce_insert_vlan_tag(sc, *m_new, NULL);
1024 	}
1025 }
1026 
1027 /*****************************************************************************
1028  *			  Transmit routines functions			     *
1029  *****************************************************************************/
1030 
1031 static int
1032 oce_tx(POCE_SOFTC sc, struct mbuf **mpp, int wq_index)
1033 {
1034 	int rc = 0, i, retry_cnt = 0;
1035 	bus_dma_segment_t segs[OCE_MAX_TX_ELEMENTS];
1036 	struct mbuf *m, *m_temp, *m_new = NULL;
1037 	struct oce_wq *wq = sc->wq[wq_index];
1038 	struct oce_packet_desc *pd;
1039 	struct oce_nic_hdr_wqe *nichdr;
1040 	struct oce_nic_frag_wqe *nicfrag;
1041 	struct ether_header *eh = NULL;
1042 	int num_wqes;
1043 	uint32_t reg_value;
1044 	boolean_t complete = TRUE;
1045 	boolean_t os2bmc = FALSE;
1046 
1047 	m = *mpp;
1048 	if (!m)
1049 		return EINVAL;
1050 
1051 	if (!(m->m_flags & M_PKTHDR)) {
1052 		rc = ENXIO;
1053 		goto free_ret;
1054 	}
1055 
1056 	/* Don't allow non-TSO packets longer than MTU */
1057 	if (!is_tso_pkt(m)) {
1058 		eh = mtod(m, struct ether_header *);
1059 		if(m->m_pkthdr.len > ETHER_MAX_FRAME(sc->ifp, eh->ether_type, FALSE))
1060 			 goto free_ret;
1061 	}
1062 
1063 	if(oce_tx_asic_stall_verify(sc, m)) {
1064 		m = oce_insert_vlan_tag(sc, m, &complete);
1065 		if(!m) {
1066 			device_printf(sc->dev, "Insertion unsuccessful\n");
1067 			return 0;
1068 		}
1069 	}
1070 
1071 	/* Lancer, SH ASIC has a bug wherein Packets that are 32 bytes or less
1072 	 * may cause a transmit stall on that port. So the work-around is to
1073 	 * pad short packets (<= 32 bytes) to a 36-byte length.
1074 	*/
1075 	if(IS_SH(sc) || IS_XE201(sc) ) {
1076 		if(m->m_pkthdr.len <= 32) {
1077 			char buf[36];
1078 			bzero((void *)buf, 36);
1079 			m_append(m, (36 - m->m_pkthdr.len), buf);
1080 		}
1081 	}
1082 
1083 tx_start:
1084 	if (m->m_pkthdr.csum_flags & CSUM_TSO) {
1085 		/* consolidate packet buffers for TSO/LSO segment offload */
1086 #if defined(INET6) || defined(INET)
1087 		m = oce_tso_setup(sc, mpp);
1088 #else
1089 		m = NULL;
1090 #endif
1091 		if (m == NULL) {
1092 			rc = ENXIO;
1093 			goto free_ret;
1094 		}
1095 	}
1096 
1097 	pd = &wq->pckts[wq->pkt_desc_head];
1098 
1099 retry:
1100 	rc = bus_dmamap_load_mbuf_sg(wq->tag,
1101 				     pd->map,
1102 				     m, segs, &pd->nsegs, BUS_DMA_NOWAIT);
1103 	if (rc == 0) {
1104 		num_wqes = pd->nsegs + 1;
1105 		if (IS_BE(sc) || IS_SH(sc)) {
1106 			/*Dummy required only for BE3.*/
1107 			if (num_wqes & 1)
1108 				num_wqes++;
1109 		}
1110 		if (num_wqes >= RING_NUM_FREE(wq->ring)) {
1111 			bus_dmamap_unload(wq->tag, pd->map);
1112 			return EBUSY;
1113 		}
1114 		atomic_store_rel_int(&wq->pkt_desc_head,
1115 				     (wq->pkt_desc_head + 1) % \
1116 				      OCE_WQ_PACKET_ARRAY_SIZE);
1117 		bus_dmamap_sync(wq->tag, pd->map, BUS_DMASYNC_PREWRITE);
1118 		pd->mbuf = m;
1119 
1120 		nichdr =
1121 		    RING_GET_PRODUCER_ITEM_VA(wq->ring, struct oce_nic_hdr_wqe);
1122 		nichdr->u0.dw[0] = 0;
1123 		nichdr->u0.dw[1] = 0;
1124 		nichdr->u0.dw[2] = 0;
1125 		nichdr->u0.dw[3] = 0;
1126 
1127 		nichdr->u0.s.complete = complete;
1128 		nichdr->u0.s.mgmt = os2bmc;
1129 		nichdr->u0.s.event = 1;
1130 		nichdr->u0.s.crc = 1;
1131 		nichdr->u0.s.forward = 0;
1132 		nichdr->u0.s.ipcs = (m->m_pkthdr.csum_flags & CSUM_IP) ? 1 : 0;
1133 		nichdr->u0.s.udpcs =
1134 			(m->m_pkthdr.csum_flags & CSUM_UDP) ? 1 : 0;
1135 		nichdr->u0.s.tcpcs =
1136 			(m->m_pkthdr.csum_flags & CSUM_TCP) ? 1 : 0;
1137 		nichdr->u0.s.num_wqe = num_wqes;
1138 		nichdr->u0.s.total_length = m->m_pkthdr.len;
1139 
1140 		if (m->m_flags & M_VLANTAG) {
1141 			nichdr->u0.s.vlan = 1; /*Vlan present*/
1142 			nichdr->u0.s.vlan_tag = m->m_pkthdr.ether_vtag;
1143 		}
1144 
1145 		if (m->m_pkthdr.csum_flags & CSUM_TSO) {
1146 			if (m->m_pkthdr.tso_segsz) {
1147 				nichdr->u0.s.lso = 1;
1148 				nichdr->u0.s.lso_mss  = m->m_pkthdr.tso_segsz;
1149 			}
1150 			if (!IS_BE(sc) || !IS_SH(sc))
1151 				nichdr->u0.s.ipcs = 1;
1152 		}
1153 
1154 		RING_PUT(wq->ring, 1);
1155 		atomic_add_int(&wq->ring->num_used, 1);
1156 
1157 		for (i = 0; i < pd->nsegs; i++) {
1158 			nicfrag =
1159 			    RING_GET_PRODUCER_ITEM_VA(wq->ring,
1160 						      struct oce_nic_frag_wqe);
1161 			nicfrag->u0.s.rsvd0 = 0;
1162 			nicfrag->u0.s.frag_pa_hi = ADDR_HI(segs[i].ds_addr);
1163 			nicfrag->u0.s.frag_pa_lo = ADDR_LO(segs[i].ds_addr);
1164 			nicfrag->u0.s.frag_len = segs[i].ds_len;
1165 			pd->wqe_idx = wq->ring->pidx;
1166 			RING_PUT(wq->ring, 1);
1167 			atomic_add_int(&wq->ring->num_used, 1);
1168 		}
1169 		if (num_wqes > (pd->nsegs + 1)) {
1170 			nicfrag =
1171 			    RING_GET_PRODUCER_ITEM_VA(wq->ring,
1172 						      struct oce_nic_frag_wqe);
1173 			nicfrag->u0.dw[0] = 0;
1174 			nicfrag->u0.dw[1] = 0;
1175 			nicfrag->u0.dw[2] = 0;
1176 			nicfrag->u0.dw[3] = 0;
1177 			pd->wqe_idx = wq->ring->pidx;
1178 			RING_PUT(wq->ring, 1);
1179 			atomic_add_int(&wq->ring->num_used, 1);
1180 			pd->nsegs++;
1181 		}
1182 
1183 		if_inc_counter(sc->ifp, IFCOUNTER_OPACKETS, 1);
1184 		wq->tx_stats.tx_reqs++;
1185 		wq->tx_stats.tx_wrbs += num_wqes;
1186 		wq->tx_stats.tx_bytes += m->m_pkthdr.len;
1187 		wq->tx_stats.tx_pkts++;
1188 
1189 		bus_dmamap_sync(wq->ring->dma.tag, wq->ring->dma.map,
1190 				BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
1191 		reg_value = (num_wqes << 16) | wq->wq_id;
1192 
1193 		/* if os2bmc is not enabled or if the pkt is already tagged as
1194 		   bmc, do nothing
1195 		 */
1196 		oce_is_pkt_dest_bmc(sc, m, &os2bmc, &m_new);
1197 
1198 		if_inc_counter(sc->ifp, IFCOUNTER_OBYTES, m->m_pkthdr.len);
1199 		if (m->m_flags & M_MCAST)
1200 			if_inc_counter(sc->ifp, IFCOUNTER_OMCASTS, 1);
1201 		ETHER_BPF_MTAP(sc->ifp, m);
1202 
1203 		OCE_WRITE_REG32(sc, db, wq->db_offset, reg_value);
1204 
1205 	} else if (rc == EFBIG)	{
1206 		if (retry_cnt == 0) {
1207 			m_temp = m_defrag(m, M_NOWAIT);
1208 			if (m_temp == NULL)
1209 				goto free_ret;
1210 			m = m_temp;
1211 			*mpp = m_temp;
1212 			retry_cnt = retry_cnt + 1;
1213 			goto retry;
1214 		} else
1215 			goto free_ret;
1216 	} else if (rc == ENOMEM)
1217 		return rc;
1218 	else
1219 		goto free_ret;
1220 
1221 	if (os2bmc) {
1222 		m = m_new;
1223 		goto tx_start;
1224 	}
1225 
1226 	return 0;
1227 
1228 free_ret:
1229 	m_freem(*mpp);
1230 	*mpp = NULL;
1231 	return rc;
1232 }
1233 
1234 static void
1235 oce_process_tx_completion(struct oce_wq *wq)
1236 {
1237 	struct oce_packet_desc *pd;
1238 	POCE_SOFTC sc = (POCE_SOFTC) wq->parent;
1239 	struct mbuf *m;
1240 
1241 	pd = &wq->pckts[wq->pkt_desc_tail];
1242 	atomic_store_rel_int(&wq->pkt_desc_tail,
1243 			     (wq->pkt_desc_tail + 1) % OCE_WQ_PACKET_ARRAY_SIZE);
1244 	atomic_subtract_int(&wq->ring->num_used, pd->nsegs + 1);
1245 	bus_dmamap_sync(wq->tag, pd->map, BUS_DMASYNC_POSTWRITE);
1246 	bus_dmamap_unload(wq->tag, pd->map);
1247 
1248 	m = pd->mbuf;
1249 	m_freem(m);
1250 	pd->mbuf = NULL;
1251 
1252 	if (sc->ifp->if_drv_flags & IFF_DRV_OACTIVE) {
1253 		if (wq->ring->num_used < (wq->ring->num_items / 2)) {
1254 			sc->ifp->if_drv_flags &= ~(IFF_DRV_OACTIVE);
1255 			oce_tx_restart(sc, wq);
1256 		}
1257 	}
1258 }
1259 
1260 static void
1261 oce_tx_restart(POCE_SOFTC sc, struct oce_wq *wq)
1262 {
1263 
1264 	if ((sc->ifp->if_drv_flags & IFF_DRV_RUNNING) != IFF_DRV_RUNNING)
1265 		return;
1266 
1267 	if (!drbr_empty(sc->ifp, wq->br))
1268 		taskqueue_enqueue(taskqueue_swi, &wq->txtask);
1269 
1270 }
1271 
1272 #if defined(INET6) || defined(INET)
1273 static struct mbuf *
1274 oce_tso_setup(POCE_SOFTC sc, struct mbuf **mpp)
1275 {
1276 	struct mbuf *m;
1277 #ifdef INET
1278 	struct ip *ip;
1279 #endif
1280 #ifdef INET6
1281 	struct ip6_hdr *ip6;
1282 #endif
1283 	struct ether_vlan_header *eh;
1284 	struct tcphdr *th;
1285 	uint16_t etype;
1286 	int total_len = 0, ehdrlen = 0;
1287 
1288 	m = *mpp;
1289 
1290 	if (M_WRITABLE(m) == 0) {
1291 		m = m_dup(*mpp, M_NOWAIT);
1292 		if (!m)
1293 			return NULL;
1294 		m_freem(*mpp);
1295 		*mpp = m;
1296 	}
1297 
1298 	eh = mtod(m, struct ether_vlan_header *);
1299 	if (eh->evl_encap_proto == htons(ETHERTYPE_VLAN)) {
1300 		etype = ntohs(eh->evl_proto);
1301 		ehdrlen = ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN;
1302 	} else {
1303 		etype = ntohs(eh->evl_encap_proto);
1304 		ehdrlen = ETHER_HDR_LEN;
1305 	}
1306 
1307 	switch (etype) {
1308 #ifdef INET
1309 	case ETHERTYPE_IP:
1310 		ip = (struct ip *)(m->m_data + ehdrlen);
1311 		if (ip->ip_p != IPPROTO_TCP)
1312 			return NULL;
1313 		th = (struct tcphdr *)((caddr_t)ip + (ip->ip_hl << 2));
1314 
1315 		total_len = ehdrlen + (ip->ip_hl << 2) + (th->th_off << 2);
1316 		break;
1317 #endif
1318 #ifdef INET6
1319 	case ETHERTYPE_IPV6:
1320 		ip6 = (struct ip6_hdr *)(m->m_data + ehdrlen);
1321 		if (ip6->ip6_nxt != IPPROTO_TCP)
1322 			return NULL;
1323 		th = (struct tcphdr *)((caddr_t)ip6 + sizeof(struct ip6_hdr));
1324 
1325 		total_len = ehdrlen + sizeof(struct ip6_hdr) + (th->th_off << 2);
1326 		break;
1327 #endif
1328 	default:
1329 		return NULL;
1330 	}
1331 
1332 	m = m_pullup(m, total_len);
1333 	*mpp = m;
1334 	return m;
1335 }
1336 #endif /* INET6 || INET */
1337 
1338 void
1339 oce_tx_task(void *arg, int npending)
1340 {
1341 	struct oce_wq *wq = arg;
1342 	POCE_SOFTC sc = wq->parent;
1343 	struct ifnet *ifp = sc->ifp;
1344 	int rc = 0;
1345 
1346 	LOCK(&wq->tx_lock);
1347 	rc = oce_multiq_transmit(ifp, NULL, wq);
1348 	if (rc) {
1349 		device_printf(sc->dev,
1350 				"TX[%d] restart failed\n", wq->queue_index);
1351 	}
1352 	UNLOCK(&wq->tx_lock);
1353 }
1354 
1355 void
1356 oce_start(struct ifnet *ifp)
1357 {
1358 	POCE_SOFTC sc = ifp->if_softc;
1359 	struct mbuf *m;
1360 	int rc = 0;
1361 	int def_q = 0; /* Defualt tx queue is 0*/
1362 
1363 	if ((ifp->if_drv_flags & (IFF_DRV_RUNNING | IFF_DRV_OACTIVE)) !=
1364 			IFF_DRV_RUNNING)
1365 		return;
1366 
1367 	if (!sc->link_status)
1368 		return;
1369 
1370 	while (true) {
1371 		IF_DEQUEUE(&sc->ifp->if_snd, m);
1372 		if (m == NULL)
1373 			break;
1374 
1375 		LOCK(&sc->wq[def_q]->tx_lock);
1376 		rc = oce_tx(sc, &m, def_q);
1377 		UNLOCK(&sc->wq[def_q]->tx_lock);
1378 		if (rc) {
1379 			if (m != NULL) {
1380 				sc->wq[def_q]->tx_stats.tx_stops ++;
1381 				ifp->if_drv_flags |= IFF_DRV_OACTIVE;
1382 				IFQ_DRV_PREPEND(&ifp->if_snd, m);
1383 				m = NULL;
1384 			}
1385 			break;
1386 		}
1387 	}
1388 }
1389 
1390 /* Handle the Completion Queue for transmit */
1391 uint16_t
1392 oce_wq_handler(void *arg)
1393 {
1394 	struct oce_wq *wq = (struct oce_wq *)arg;
1395 	POCE_SOFTC sc = wq->parent;
1396 	struct oce_cq *cq = wq->cq;
1397 	struct oce_nic_tx_cqe *cqe;
1398 	int num_cqes = 0;
1399 
1400 	LOCK(&wq->tx_compl_lock);
1401 	bus_dmamap_sync(cq->ring->dma.tag,
1402 			cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
1403 	cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_nic_tx_cqe);
1404 	while (cqe->u0.dw[3]) {
1405 		DW_SWAP((uint32_t *) cqe, sizeof(oce_wq_cqe));
1406 
1407 		wq->ring->cidx = cqe->u0.s.wqe_index + 1;
1408 		if (wq->ring->cidx >= wq->ring->num_items)
1409 			wq->ring->cidx -= wq->ring->num_items;
1410 
1411 		oce_process_tx_completion(wq);
1412 		wq->tx_stats.tx_compl++;
1413 		cqe->u0.dw[3] = 0;
1414 		RING_GET(cq->ring, 1);
1415 		bus_dmamap_sync(cq->ring->dma.tag,
1416 				cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
1417 		cqe =
1418 		    RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_nic_tx_cqe);
1419 		num_cqes++;
1420 	}
1421 
1422 	if (num_cqes)
1423 		oce_arm_cq(sc, cq->cq_id, num_cqes, FALSE);
1424 
1425 	UNLOCK(&wq->tx_compl_lock);
1426 	return num_cqes;
1427 }
1428 
1429 static int
1430 oce_multiq_transmit(struct ifnet *ifp, struct mbuf *m, struct oce_wq *wq)
1431 {
1432 	POCE_SOFTC sc = ifp->if_softc;
1433 	int status = 0, queue_index = 0;
1434 	struct mbuf *next = NULL;
1435 	struct buf_ring *br = NULL;
1436 
1437 	br  = wq->br;
1438 	queue_index = wq->queue_index;
1439 
1440 	if ((ifp->if_drv_flags & (IFF_DRV_RUNNING | IFF_DRV_OACTIVE)) !=
1441 		IFF_DRV_RUNNING) {
1442 		if (m != NULL)
1443 			status = drbr_enqueue(ifp, br, m);
1444 		return status;
1445 	}
1446 
1447 	if (m != NULL) {
1448 		if ((status = drbr_enqueue(ifp, br, m)) != 0)
1449 			return status;
1450 	}
1451 	while ((next = drbr_peek(ifp, br)) != NULL) {
1452 		if (oce_tx(sc, &next, queue_index)) {
1453 			if (next == NULL) {
1454 				drbr_advance(ifp, br);
1455 			} else {
1456 				drbr_putback(ifp, br, next);
1457 				wq->tx_stats.tx_stops ++;
1458 				ifp->if_drv_flags |= IFF_DRV_OACTIVE;
1459 			}
1460 			break;
1461 		}
1462 		drbr_advance(ifp, br);
1463 	}
1464 
1465 	return 0;
1466 }
1467 
1468 /*****************************************************************************
1469  *			    Receive  routines functions 		     *
1470  *****************************************************************************/
1471 
1472 static void
1473 oce_correct_header(struct mbuf *m, struct nic_hwlro_cqe_part1 *cqe1, struct nic_hwlro_cqe_part2 *cqe2)
1474 {
1475 	uint32_t *p;
1476         struct ether_header *eh = NULL;
1477         struct tcphdr *tcp_hdr = NULL;
1478         struct ip *ip4_hdr = NULL;
1479         struct ip6_hdr *ip6 = NULL;
1480         uint32_t payload_len = 0;
1481 
1482         eh = mtod(m, struct ether_header *);
1483         /* correct IP header */
1484         if(!cqe2->ipv6_frame) {
1485 		ip4_hdr = (struct ip *)((char*)eh + sizeof(struct ether_header));
1486                 ip4_hdr->ip_ttl = cqe2->frame_lifespan;
1487                 ip4_hdr->ip_len = htons(cqe2->coalesced_size - sizeof(struct ether_header));
1488                 tcp_hdr = (struct tcphdr *)((char*)ip4_hdr + sizeof(struct ip));
1489         }else {
1490         	ip6 = (struct ip6_hdr *)((char*)eh + sizeof(struct ether_header));
1491                 ip6->ip6_ctlun.ip6_un1.ip6_un1_hlim = cqe2->frame_lifespan;
1492                 payload_len = cqe2->coalesced_size - sizeof(struct ether_header)
1493                                                 - sizeof(struct ip6_hdr);
1494                 ip6->ip6_ctlun.ip6_un1.ip6_un1_plen = htons(payload_len);
1495                 tcp_hdr = (struct tcphdr *)((char*)ip6 + sizeof(struct ip6_hdr));
1496         }
1497 
1498         /* correct tcp header */
1499         tcp_hdr->th_ack = htonl(cqe2->tcp_ack_num);
1500         if(cqe2->push) {
1501         	tcp_hdr->th_flags |= TH_PUSH;
1502         }
1503         tcp_hdr->th_win = htons(cqe2->tcp_window);
1504         tcp_hdr->th_sum = 0xffff;
1505         if(cqe2->ts_opt) {
1506                 p = (uint32_t *)((char*)tcp_hdr + sizeof(struct tcphdr) + 2);
1507                 *p = cqe1->tcp_timestamp_val;
1508                 *(p+1) = cqe1->tcp_timestamp_ecr;
1509         }
1510 
1511 	return;
1512 }
1513 
1514 static void
1515 oce_rx_mbuf_chain(struct oce_rq *rq, struct oce_common_cqe_info *cqe_info, struct mbuf **m)
1516 {
1517 	POCE_SOFTC sc = (POCE_SOFTC) rq->parent;
1518         uint32_t i = 0, frag_len = 0;
1519 	uint32_t len = cqe_info->pkt_size;
1520         struct oce_packet_desc *pd;
1521         struct mbuf *tail = NULL;
1522 
1523         for (i = 0; i < cqe_info->num_frags; i++) {
1524                 if (rq->ring->cidx == rq->ring->pidx) {
1525                         device_printf(sc->dev,
1526                                   "oce_rx_mbuf_chain: Invalid RX completion - Queue is empty\n");
1527                         return;
1528                 }
1529                 pd = &rq->pckts[rq->ring->cidx];
1530 
1531                 bus_dmamap_sync(rq->tag, pd->map, BUS_DMASYNC_POSTWRITE);
1532                 bus_dmamap_unload(rq->tag, pd->map);
1533 		RING_GET(rq->ring, 1);
1534                 rq->pending--;
1535 
1536                 frag_len = (len > rq->cfg.frag_size) ? rq->cfg.frag_size : len;
1537                 pd->mbuf->m_len = frag_len;
1538 
1539                 if (tail != NULL) {
1540                         /* additional fragments */
1541                         pd->mbuf->m_flags &= ~M_PKTHDR;
1542                         tail->m_next = pd->mbuf;
1543 			if(rq->islro)
1544                         	tail->m_nextpkt = NULL;
1545                         tail = pd->mbuf;
1546                 } else {
1547                         /* first fragment, fill out much of the packet header */
1548                         pd->mbuf->m_pkthdr.len = len;
1549 			if(rq->islro)
1550                         	pd->mbuf->m_nextpkt = NULL;
1551                         pd->mbuf->m_pkthdr.csum_flags = 0;
1552                         if (IF_CSUM_ENABLED(sc)) {
1553                                 if (cqe_info->l4_cksum_pass) {
1554                                         if(!cqe_info->ipv6_frame) { /* IPV4 */
1555                                                 pd->mbuf->m_pkthdr.csum_flags |=
1556                                                         (CSUM_DATA_VALID | CSUM_PSEUDO_HDR);
1557                                         }else { /* IPV6 frame */
1558 						if(rq->islro) {
1559                                                 	pd->mbuf->m_pkthdr.csum_flags |=
1560                                                         (CSUM_DATA_VALID | CSUM_PSEUDO_HDR);
1561 						}
1562                                         }
1563                                         pd->mbuf->m_pkthdr.csum_data = 0xffff;
1564                                 }
1565                                 if (cqe_info->ip_cksum_pass) {
1566                                         pd->mbuf->m_pkthdr.csum_flags |=
1567                                                (CSUM_IP_CHECKED|CSUM_IP_VALID);
1568                                 }
1569                         }
1570                         *m = tail = pd->mbuf;
1571                }
1572                 pd->mbuf = NULL;
1573                 len -= frag_len;
1574         }
1575 
1576         return;
1577 }
1578 
1579 static void
1580 oce_rx_lro(struct oce_rq *rq, struct nic_hwlro_singleton_cqe *cqe, struct nic_hwlro_cqe_part2 *cqe2)
1581 {
1582         POCE_SOFTC sc = (POCE_SOFTC) rq->parent;
1583         struct nic_hwlro_cqe_part1 *cqe1 = NULL;
1584         struct mbuf *m = NULL;
1585 	struct oce_common_cqe_info cq_info;
1586 
1587 	/* parse cqe */
1588         if(cqe2 == NULL) {
1589                 cq_info.pkt_size =  cqe->pkt_size;
1590                 cq_info.vtag = cqe->vlan_tag;
1591                 cq_info.l4_cksum_pass = cqe->l4_cksum_pass;
1592                 cq_info.ip_cksum_pass = cqe->ip_cksum_pass;
1593                 cq_info.ipv6_frame = cqe->ipv6_frame;
1594                 cq_info.vtp = cqe->vtp;
1595                 cq_info.qnq = cqe->qnq;
1596         }else {
1597                 cqe1 = (struct nic_hwlro_cqe_part1 *)cqe;
1598                 cq_info.pkt_size =  cqe2->coalesced_size;
1599                 cq_info.vtag = cqe2->vlan_tag;
1600                 cq_info.l4_cksum_pass = cqe2->l4_cksum_pass;
1601                 cq_info.ip_cksum_pass = cqe2->ip_cksum_pass;
1602                 cq_info.ipv6_frame = cqe2->ipv6_frame;
1603                 cq_info.vtp = cqe2->vtp;
1604                 cq_info.qnq = cqe1->qnq;
1605         }
1606 
1607 	cq_info.vtag = BSWAP_16(cq_info.vtag);
1608 
1609         cq_info.num_frags = cq_info.pkt_size / rq->cfg.frag_size;
1610         if(cq_info.pkt_size % rq->cfg.frag_size)
1611                 cq_info.num_frags++;
1612 
1613 	oce_rx_mbuf_chain(rq, &cq_info, &m);
1614 
1615 	if (m) {
1616 		if(cqe2) {
1617 			//assert(cqe2->valid != 0);
1618 
1619 			//assert(cqe2->cqe_type != 2);
1620 			oce_correct_header(m, cqe1, cqe2);
1621 		}
1622 
1623 		m->m_pkthdr.rcvif = sc->ifp;
1624 		if (rq->queue_index)
1625 			m->m_pkthdr.flowid = (rq->queue_index - 1);
1626 		else
1627 			m->m_pkthdr.flowid = rq->queue_index;
1628 		M_HASHTYPE_SET(m, M_HASHTYPE_OPAQUE);
1629 
1630 		/* This deternies if vlan tag is Valid */
1631 		if (cq_info.vtp) {
1632 			if (sc->function_mode & FNM_FLEX10_MODE) {
1633 				/* FLEX10. If QnQ is not set, neglect VLAN */
1634 				if (cq_info.qnq) {
1635 					m->m_pkthdr.ether_vtag = cq_info.vtag;
1636 					m->m_flags |= M_VLANTAG;
1637 				}
1638 			} else if (sc->pvid != (cq_info.vtag & VLAN_VID_MASK))  {
1639 				/* In UMC mode generally pvid will be striped by
1640 				   hw. But in some cases we have seen it comes
1641 				   with pvid. So if pvid == vlan, neglect vlan.
1642 				 */
1643 				m->m_pkthdr.ether_vtag = cq_info.vtag;
1644 				m->m_flags |= M_VLANTAG;
1645 			}
1646 		}
1647 		if_inc_counter(sc->ifp, IFCOUNTER_IPACKETS, 1);
1648 
1649 		(*sc->ifp->if_input) (sc->ifp, m);
1650 
1651 		/* Update rx stats per queue */
1652 		rq->rx_stats.rx_pkts++;
1653 		rq->rx_stats.rx_bytes += cq_info.pkt_size;
1654 		rq->rx_stats.rx_frags += cq_info.num_frags;
1655 		rq->rx_stats.rx_ucast_pkts++;
1656 	}
1657         return;
1658 }
1659 
1660 static void
1661 oce_rx(struct oce_rq *rq, struct oce_nic_rx_cqe *cqe)
1662 {
1663 	POCE_SOFTC sc = (POCE_SOFTC) rq->parent;
1664 	int len;
1665 	struct mbuf *m = NULL;
1666 	struct oce_common_cqe_info cq_info;
1667 	uint16_t vtag = 0;
1668 
1669 	/* Is it a flush compl that has no data */
1670 	if(!cqe->u0.s.num_fragments)
1671 		goto exit;
1672 
1673 	len = cqe->u0.s.pkt_size;
1674 	if (!len) {
1675 		/*partial DMA workaround for Lancer*/
1676 		oce_discard_rx_comp(rq, cqe->u0.s.num_fragments);
1677 		goto exit;
1678 	}
1679 
1680 	if (!oce_cqe_portid_valid(sc, cqe)) {
1681 		oce_discard_rx_comp(rq, cqe->u0.s.num_fragments);
1682 		goto exit;
1683 	}
1684 
1685 	 /* Get vlan_tag value */
1686 	if(IS_BE(sc) || IS_SH(sc))
1687 		vtag = BSWAP_16(cqe->u0.s.vlan_tag);
1688 	else
1689 		vtag = cqe->u0.s.vlan_tag;
1690 
1691 	cq_info.l4_cksum_pass = cqe->u0.s.l4_cksum_pass;
1692 	cq_info.ip_cksum_pass = cqe->u0.s.ip_cksum_pass;
1693 	cq_info.ipv6_frame = cqe->u0.s.ip_ver;
1694 	cq_info.num_frags = cqe->u0.s.num_fragments;
1695 	cq_info.pkt_size = cqe->u0.s.pkt_size;
1696 
1697 	oce_rx_mbuf_chain(rq, &cq_info, &m);
1698 
1699 	if (m) {
1700 		m->m_pkthdr.rcvif = sc->ifp;
1701 		if (rq->queue_index)
1702 			m->m_pkthdr.flowid = (rq->queue_index - 1);
1703 		else
1704 			m->m_pkthdr.flowid = rq->queue_index;
1705 		M_HASHTYPE_SET(m, M_HASHTYPE_OPAQUE);
1706 
1707 		/* This deternies if vlan tag is Valid */
1708 		if (oce_cqe_vtp_valid(sc, cqe)) {
1709 			if (sc->function_mode & FNM_FLEX10_MODE) {
1710 				/* FLEX10. If QnQ is not set, neglect VLAN */
1711 				if (cqe->u0.s.qnq) {
1712 					m->m_pkthdr.ether_vtag = vtag;
1713 					m->m_flags |= M_VLANTAG;
1714 				}
1715 			} else if (sc->pvid != (vtag & VLAN_VID_MASK))  {
1716 				/* In UMC mode generally pvid will be striped by
1717 				   hw. But in some cases we have seen it comes
1718 				   with pvid. So if pvid == vlan, neglect vlan.
1719 				*/
1720 				m->m_pkthdr.ether_vtag = vtag;
1721 				m->m_flags |= M_VLANTAG;
1722 			}
1723 		}
1724 
1725 		if_inc_counter(sc->ifp, IFCOUNTER_IPACKETS, 1);
1726 #if defined(INET6) || defined(INET)
1727 		/* Try to queue to LRO */
1728 		if (IF_LRO_ENABLED(sc) &&
1729 		    (cqe->u0.s.ip_cksum_pass) &&
1730 		    (cqe->u0.s.l4_cksum_pass) &&
1731 		    (!cqe->u0.s.ip_ver)       &&
1732 		    (rq->lro.lro_cnt != 0)) {
1733 			if (tcp_lro_rx(&rq->lro, m, 0) == 0) {
1734 				rq->lro_pkts_queued ++;
1735 				goto post_done;
1736 			}
1737 			/* If LRO posting fails then try to post to STACK */
1738 		}
1739 #endif
1740 
1741 		(*sc->ifp->if_input) (sc->ifp, m);
1742 #if defined(INET6) || defined(INET)
1743 post_done:
1744 #endif
1745 		/* Update rx stats per queue */
1746 		rq->rx_stats.rx_pkts++;
1747 		rq->rx_stats.rx_bytes += cqe->u0.s.pkt_size;
1748 		rq->rx_stats.rx_frags += cqe->u0.s.num_fragments;
1749 		if (cqe->u0.s.pkt_type == OCE_MULTICAST_PACKET)
1750 			rq->rx_stats.rx_mcast_pkts++;
1751 		if (cqe->u0.s.pkt_type == OCE_UNICAST_PACKET)
1752 			rq->rx_stats.rx_ucast_pkts++;
1753 	}
1754 exit:
1755 	return;
1756 }
1757 
1758 void
1759 oce_discard_rx_comp(struct oce_rq *rq, int num_frags)
1760 {
1761 	uint32_t i = 0;
1762 	struct oce_packet_desc *pd;
1763 	POCE_SOFTC sc = (POCE_SOFTC) rq->parent;
1764 
1765 	for (i = 0; i < num_frags; i++) {
1766                 if (rq->ring->cidx == rq->ring->pidx) {
1767                         device_printf(sc->dev,
1768                                 "oce_discard_rx_comp: Invalid RX completion - Queue is empty\n");
1769                         return;
1770                 }
1771                 pd = &rq->pckts[rq->ring->cidx];
1772                 bus_dmamap_sync(rq->tag, pd->map, BUS_DMASYNC_POSTWRITE);
1773                 bus_dmamap_unload(rq->tag, pd->map);
1774                 if (pd->mbuf != NULL) {
1775                         m_freem(pd->mbuf);
1776                         pd->mbuf = NULL;
1777                 }
1778 
1779 		RING_GET(rq->ring, 1);
1780                 rq->pending--;
1781 	}
1782 }
1783 
1784 static int
1785 oce_cqe_vtp_valid(POCE_SOFTC sc, struct oce_nic_rx_cqe *cqe)
1786 {
1787 	struct oce_nic_rx_cqe_v1 *cqe_v1;
1788 	int vtp = 0;
1789 
1790 	if (sc->be3_native) {
1791 		cqe_v1 = (struct oce_nic_rx_cqe_v1 *)cqe;
1792 		vtp =  cqe_v1->u0.s.vlan_tag_present;
1793 	} else
1794 		vtp = cqe->u0.s.vlan_tag_present;
1795 
1796 	return vtp;
1797 
1798 }
1799 
1800 static int
1801 oce_cqe_portid_valid(POCE_SOFTC sc, struct oce_nic_rx_cqe *cqe)
1802 {
1803 	struct oce_nic_rx_cqe_v1 *cqe_v1;
1804 	int port_id = 0;
1805 
1806 	if (sc->be3_native && (IS_BE(sc) || IS_SH(sc))) {
1807 		cqe_v1 = (struct oce_nic_rx_cqe_v1 *)cqe;
1808 		port_id =  cqe_v1->u0.s.port;
1809 		if (sc->port_id != port_id)
1810 			return 0;
1811 	} else
1812 		;/* For BE3 legacy and Lancer this is dummy */
1813 
1814 	return 1;
1815 
1816 }
1817 
1818 #if defined(INET6) || defined(INET)
1819 void
1820 oce_rx_flush_lro(struct oce_rq *rq)
1821 {
1822 	struct lro_ctrl	*lro = &rq->lro;
1823 	POCE_SOFTC sc = (POCE_SOFTC) rq->parent;
1824 
1825 	if (!IF_LRO_ENABLED(sc))
1826 		return;
1827 
1828 	tcp_lro_flush_all(lro);
1829 	rq->lro_pkts_queued = 0;
1830 
1831 	return;
1832 }
1833 
1834 static int
1835 oce_init_lro(POCE_SOFTC sc)
1836 {
1837 	struct lro_ctrl *lro = NULL;
1838 	int i = 0, rc = 0;
1839 
1840 	for (i = 0; i < sc->nrqs; i++) {
1841 		lro = &sc->rq[i]->lro;
1842 		rc = tcp_lro_init(lro);
1843 		if (rc != 0) {
1844 			device_printf(sc->dev, "LRO init failed\n");
1845 			return rc;
1846 		}
1847 		lro->ifp = sc->ifp;
1848 	}
1849 
1850 	return rc;
1851 }
1852 
1853 void
1854 oce_free_lro(POCE_SOFTC sc)
1855 {
1856 	struct lro_ctrl *lro = NULL;
1857 	int i = 0;
1858 
1859 	for (i = 0; i < sc->nrqs; i++) {
1860 		lro = &sc->rq[i]->lro;
1861 		if (lro)
1862 			tcp_lro_free(lro);
1863 	}
1864 }
1865 #endif
1866 
1867 int
1868 oce_alloc_rx_bufs(struct oce_rq *rq, int count)
1869 {
1870 	POCE_SOFTC sc = (POCE_SOFTC) rq->parent;
1871 	int i, in, rc;
1872 	struct oce_packet_desc *pd;
1873 	bus_dma_segment_t segs[6];
1874 	int nsegs, added = 0;
1875 	struct oce_nic_rqe *rqe;
1876 	pd_rxulp_db_t rxdb_reg;
1877 	uint32_t val = 0;
1878 	uint32_t oce_max_rq_posts = 64;
1879 
1880 	bzero(&rxdb_reg, sizeof(pd_rxulp_db_t));
1881 	for (i = 0; i < count; i++) {
1882 		in = (rq->ring->pidx + 1) % OCE_RQ_PACKET_ARRAY_SIZE;
1883 
1884 		pd = &rq->pckts[rq->ring->pidx];
1885 		pd->mbuf = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR, oce_rq_buf_size);
1886 		if (pd->mbuf == NULL) {
1887 			device_printf(sc->dev, "mbuf allocation failed, size = %d\n",oce_rq_buf_size);
1888 			break;
1889 		}
1890 		pd->mbuf->m_nextpkt = NULL;
1891 
1892 		pd->mbuf->m_len = pd->mbuf->m_pkthdr.len = rq->cfg.frag_size;
1893 
1894 		rc = bus_dmamap_load_mbuf_sg(rq->tag,
1895 					     pd->map,
1896 					     pd->mbuf,
1897 					     segs, &nsegs, BUS_DMA_NOWAIT);
1898 		if (rc) {
1899 			m_free(pd->mbuf);
1900 			device_printf(sc->dev, "bus_dmamap_load_mbuf_sg failed rc = %d\n", rc);
1901 			break;
1902 		}
1903 
1904 		if (nsegs != 1) {
1905 			i--;
1906 			continue;
1907 		}
1908 
1909 		bus_dmamap_sync(rq->tag, pd->map, BUS_DMASYNC_PREREAD);
1910 
1911 		rqe = RING_GET_PRODUCER_ITEM_VA(rq->ring, struct oce_nic_rqe);
1912 		rqe->u0.s.frag_pa_hi = ADDR_HI(segs[0].ds_addr);
1913 		rqe->u0.s.frag_pa_lo = ADDR_LO(segs[0].ds_addr);
1914 		DW_SWAP(u32ptr(rqe), sizeof(struct oce_nic_rqe));
1915 		RING_PUT(rq->ring, 1);
1916 		added++;
1917 		rq->pending++;
1918 	}
1919 	oce_max_rq_posts = sc->enable_hwlro ? OCE_HWLRO_MAX_RQ_POSTS : OCE_MAX_RQ_POSTS;
1920 	if (added != 0) {
1921 		for (i = added / oce_max_rq_posts; i > 0; i--) {
1922 			rxdb_reg.bits.num_posted = oce_max_rq_posts;
1923 			rxdb_reg.bits.qid = rq->rq_id;
1924 			if(rq->islro) {
1925                                 val |= rq->rq_id & DB_LRO_RQ_ID_MASK;
1926                                 val |= oce_max_rq_posts << 16;
1927                                 OCE_WRITE_REG32(sc, db, DB_OFFSET, val);
1928 			}else {
1929 				OCE_WRITE_REG32(sc, db, PD_RXULP_DB, rxdb_reg.dw0);
1930 			}
1931 			added -= oce_max_rq_posts;
1932 		}
1933 		if (added > 0) {
1934 			rxdb_reg.bits.qid = rq->rq_id;
1935 			rxdb_reg.bits.num_posted = added;
1936 			if(rq->islro) {
1937                                 val |= rq->rq_id & DB_LRO_RQ_ID_MASK;
1938                                 val |= added << 16;
1939                                 OCE_WRITE_REG32(sc, db, DB_OFFSET, val);
1940 			}else {
1941 				OCE_WRITE_REG32(sc, db, PD_RXULP_DB, rxdb_reg.dw0);
1942 			}
1943 		}
1944 	}
1945 
1946 	return 0;
1947 }
1948 
1949 static void
1950 oce_check_rx_bufs(POCE_SOFTC sc, uint32_t num_cqes, struct oce_rq *rq)
1951 {
1952         if (num_cqes) {
1953                 oce_arm_cq(sc, rq->cq->cq_id, num_cqes, FALSE);
1954 		if(!sc->enable_hwlro) {
1955 			if((OCE_RQ_PACKET_ARRAY_SIZE - rq->pending) > 1)
1956 				oce_alloc_rx_bufs(rq, ((OCE_RQ_PACKET_ARRAY_SIZE - rq->pending) - 1));
1957 		}else {
1958                 	if ((OCE_RQ_PACKET_ARRAY_SIZE -1 - rq->pending) > 64)
1959                         	oce_alloc_rx_bufs(rq, 64);
1960         	}
1961 	}
1962 
1963         return;
1964 }
1965 
1966 uint16_t
1967 oce_rq_handler_lro(void *arg)
1968 {
1969         struct oce_rq *rq = (struct oce_rq *)arg;
1970         struct oce_cq *cq = rq->cq;
1971         POCE_SOFTC sc = rq->parent;
1972         struct nic_hwlro_singleton_cqe *cqe;
1973         struct nic_hwlro_cqe_part2 *cqe2;
1974         int num_cqes = 0;
1975 
1976 	LOCK(&rq->rx_lock);
1977         bus_dmamap_sync(cq->ring->dma.tag,cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
1978         cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct nic_hwlro_singleton_cqe);
1979         while (cqe->valid) {
1980                 if(cqe->cqe_type == 0) { /* singleton cqe */
1981 			/* we should not get singleton cqe after cqe1 on same rq */
1982 			if(rq->cqe_firstpart != NULL) {
1983 				device_printf(sc->dev, "Got singleton cqe after cqe1 \n");
1984 				goto exit_rq_handler_lro;
1985 			}
1986                         if(cqe->error != 0) {
1987                                 rq->rx_stats.rxcp_err++;
1988 				if_inc_counter(sc->ifp, IFCOUNTER_IERRORS, 1);
1989                         }
1990                         oce_rx_lro(rq, cqe, NULL);
1991                         rq->rx_stats.rx_compl++;
1992                         cqe->valid = 0;
1993                         RING_GET(cq->ring, 1);
1994                         num_cqes++;
1995                         if (num_cqes >= (IS_XE201(sc) ? 8 : oce_max_rsp_handled))
1996                                 break;
1997                 }else if(cqe->cqe_type == 0x1) { /* first part */
1998 			/* we should not get cqe1 after cqe1 on same rq */
1999 			if(rq->cqe_firstpart != NULL) {
2000 				device_printf(sc->dev, "Got cqe1 after cqe1 \n");
2001 				goto exit_rq_handler_lro;
2002 			}
2003 			rq->cqe_firstpart = (struct nic_hwlro_cqe_part1 *)cqe;
2004                         RING_GET(cq->ring, 1);
2005                 }else if(cqe->cqe_type == 0x2) { /* second part */
2006 			cqe2 = (struct nic_hwlro_cqe_part2 *)cqe;
2007                         if(cqe2->error != 0) {
2008                                 rq->rx_stats.rxcp_err++;
2009 				if_inc_counter(sc->ifp, IFCOUNTER_IERRORS, 1);
2010                         }
2011 			/* We should not get cqe2 without cqe1 */
2012 			if(rq->cqe_firstpart == NULL) {
2013 				device_printf(sc->dev, "Got cqe2 without cqe1 \n");
2014 				goto exit_rq_handler_lro;
2015 			}
2016                         oce_rx_lro(rq, (struct nic_hwlro_singleton_cqe *)rq->cqe_firstpart, cqe2);
2017 
2018                         rq->rx_stats.rx_compl++;
2019                         rq->cqe_firstpart->valid = 0;
2020                         cqe2->valid = 0;
2021 			rq->cqe_firstpart = NULL;
2022 
2023                         RING_GET(cq->ring, 1);
2024                         num_cqes += 2;
2025                         if (num_cqes >= (IS_XE201(sc) ? 8 : oce_max_rsp_handled))
2026                                 break;
2027 		}
2028 
2029                 bus_dmamap_sync(cq->ring->dma.tag,cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
2030                 cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct nic_hwlro_singleton_cqe);
2031         }
2032 	oce_check_rx_bufs(sc, num_cqes, rq);
2033 exit_rq_handler_lro:
2034 	UNLOCK(&rq->rx_lock);
2035 	return 0;
2036 }
2037 
2038 /* Handle the Completion Queue for receive */
2039 uint16_t
2040 oce_rq_handler(void *arg)
2041 {
2042 	struct epoch_tracker et;
2043 	struct oce_rq *rq = (struct oce_rq *)arg;
2044 	struct oce_cq *cq = rq->cq;
2045 	POCE_SOFTC sc = rq->parent;
2046 	struct oce_nic_rx_cqe *cqe;
2047 	int num_cqes = 0;
2048 
2049 	NET_EPOCH_ENTER(et);
2050 	if(rq->islro) {
2051 		oce_rq_handler_lro(arg);
2052 		NET_EPOCH_EXIT(et);
2053 		return 0;
2054 	}
2055 	LOCK(&rq->rx_lock);
2056 	bus_dmamap_sync(cq->ring->dma.tag,
2057 			cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
2058 	cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_nic_rx_cqe);
2059 	while (cqe->u0.dw[2]) {
2060 		DW_SWAP((uint32_t *) cqe, sizeof(oce_rq_cqe));
2061 
2062 		if (cqe->u0.s.error == 0) {
2063 			oce_rx(rq, cqe);
2064 		} else {
2065 			rq->rx_stats.rxcp_err++;
2066 			if_inc_counter(sc->ifp, IFCOUNTER_IERRORS, 1);
2067 			/* Post L3/L4 errors to stack.*/
2068 			oce_rx(rq, cqe);
2069 		}
2070 		rq->rx_stats.rx_compl++;
2071 		cqe->u0.dw[2] = 0;
2072 
2073 #if defined(INET6) || defined(INET)
2074 		if (IF_LRO_ENABLED(sc) && rq->lro_pkts_queued >= 16) {
2075 			oce_rx_flush_lro(rq);
2076 		}
2077 #endif
2078 
2079 		RING_GET(cq->ring, 1);
2080 		bus_dmamap_sync(cq->ring->dma.tag,
2081 				cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
2082 		cqe =
2083 		    RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_nic_rx_cqe);
2084 		num_cqes++;
2085 		if (num_cqes >= (IS_XE201(sc) ? 8 : oce_max_rsp_handled))
2086 			break;
2087 	}
2088 
2089 #if defined(INET6) || defined(INET)
2090         if (IF_LRO_ENABLED(sc))
2091                 oce_rx_flush_lro(rq);
2092 #endif
2093 
2094 	oce_check_rx_bufs(sc, num_cqes, rq);
2095 	UNLOCK(&rq->rx_lock);
2096 	NET_EPOCH_EXIT(et);
2097 	return 0;
2098 
2099 }
2100 
2101 /*****************************************************************************
2102  *		   Helper function prototypes in this file 		     *
2103  *****************************************************************************/
2104 
2105 static int
2106 oce_attach_ifp(POCE_SOFTC sc)
2107 {
2108 
2109 	sc->ifp = if_alloc(IFT_ETHER);
2110 	if (!sc->ifp)
2111 		return ENOMEM;
2112 
2113 	ifmedia_init(&sc->media, IFM_IMASK, oce_media_change, oce_media_status);
2114 	ifmedia_add(&sc->media, IFM_ETHER | IFM_AUTO, 0, NULL);
2115 	ifmedia_set(&sc->media, IFM_ETHER | IFM_AUTO);
2116 
2117 	sc->ifp->if_flags = IFF_BROADCAST | IFF_MULTICAST | IFF_KNOWSEPOCH;
2118 	sc->ifp->if_ioctl = oce_ioctl;
2119 	sc->ifp->if_start = oce_start;
2120 	sc->ifp->if_init = oce_init;
2121 	sc->ifp->if_mtu = ETHERMTU;
2122 	sc->ifp->if_softc = sc;
2123 	sc->ifp->if_transmit = oce_multiq_start;
2124 	sc->ifp->if_qflush = oce_multiq_flush;
2125 
2126 	if_initname(sc->ifp,
2127 		    device_get_name(sc->dev), device_get_unit(sc->dev));
2128 
2129 	sc->ifp->if_snd.ifq_drv_maxlen = OCE_MAX_TX_DESC - 1;
2130 	IFQ_SET_MAXLEN(&sc->ifp->if_snd, sc->ifp->if_snd.ifq_drv_maxlen);
2131 	IFQ_SET_READY(&sc->ifp->if_snd);
2132 
2133 	sc->ifp->if_hwassist = OCE_IF_HWASSIST;
2134 	sc->ifp->if_hwassist |= CSUM_TSO;
2135 	sc->ifp->if_hwassist |= (CSUM_IP | CSUM_TCP | CSUM_UDP);
2136 
2137 	sc->ifp->if_capabilities = OCE_IF_CAPABILITIES;
2138 	sc->ifp->if_capabilities |= IFCAP_HWCSUM;
2139 	sc->ifp->if_capabilities |= IFCAP_VLAN_HWFILTER;
2140 
2141 #if defined(INET6) || defined(INET)
2142 	sc->ifp->if_capabilities |= IFCAP_TSO;
2143 	sc->ifp->if_capabilities |= IFCAP_LRO;
2144 	sc->ifp->if_capabilities |= IFCAP_VLAN_HWTSO;
2145 #endif
2146 
2147 	sc->ifp->if_capenable = sc->ifp->if_capabilities;
2148 	sc->ifp->if_baudrate = IF_Gbps(10);
2149 
2150 	sc->ifp->if_hw_tsomax = 65536 - (ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN);
2151 	sc->ifp->if_hw_tsomaxsegcount = OCE_MAX_TX_ELEMENTS;
2152 	sc->ifp->if_hw_tsomaxsegsize = 4096;
2153 
2154 	ether_ifattach(sc->ifp, sc->macaddr.mac_addr);
2155 
2156 	return 0;
2157 }
2158 
2159 static void
2160 oce_add_vlan(void *arg, struct ifnet *ifp, uint16_t vtag)
2161 {
2162 	POCE_SOFTC sc = ifp->if_softc;
2163 
2164 	if (ifp->if_softc !=  arg)
2165 		return;
2166 	if ((vtag == 0) || (vtag > 4095))
2167 		return;
2168 
2169 	sc->vlan_tag[vtag] = 1;
2170 	sc->vlans_added++;
2171 	if (sc->vlans_added <= (sc->max_vlans + 1))
2172 		oce_vid_config(sc);
2173 }
2174 
2175 static void
2176 oce_del_vlan(void *arg, struct ifnet *ifp, uint16_t vtag)
2177 {
2178 	POCE_SOFTC sc = ifp->if_softc;
2179 
2180 	if (ifp->if_softc !=  arg)
2181 		return;
2182 	if ((vtag == 0) || (vtag > 4095))
2183 		return;
2184 
2185 	sc->vlan_tag[vtag] = 0;
2186 	sc->vlans_added--;
2187 	oce_vid_config(sc);
2188 }
2189 
2190 /*
2191  * A max of 64 vlans can be configured in BE. If the user configures
2192  * more, place the card in vlan promiscuous mode.
2193  */
2194 static int
2195 oce_vid_config(POCE_SOFTC sc)
2196 {
2197 	struct normal_vlan vtags[MAX_VLANFILTER_SIZE];
2198 	uint16_t ntags = 0, i;
2199 	int status = 0;
2200 
2201 	if ((sc->vlans_added <= MAX_VLANFILTER_SIZE) &&
2202 			(sc->ifp->if_capenable & IFCAP_VLAN_HWFILTER)) {
2203 		for (i = 0; i < MAX_VLANS; i++) {
2204 			if (sc->vlan_tag[i]) {
2205 				vtags[ntags].vtag = i;
2206 				ntags++;
2207 			}
2208 		}
2209 		if (ntags)
2210 			status = oce_config_vlan(sc, (uint8_t) sc->if_id,
2211 						vtags, ntags, 1, 0);
2212 	} else
2213 		status = oce_config_vlan(sc, (uint8_t) sc->if_id,
2214 					 	NULL, 0, 1, 1);
2215 	return status;
2216 }
2217 
2218 static void
2219 oce_mac_addr_set(POCE_SOFTC sc)
2220 {
2221 	uint32_t old_pmac_id = sc->pmac_id;
2222 	int status = 0;
2223 
2224 	status = bcmp((IF_LLADDR(sc->ifp)), sc->macaddr.mac_addr,
2225 			 sc->macaddr.size_of_struct);
2226 	if (!status)
2227 		return;
2228 
2229 	status = oce_mbox_macaddr_add(sc, (uint8_t *)(IF_LLADDR(sc->ifp)),
2230 					sc->if_id, &sc->pmac_id);
2231 	if (!status) {
2232 		status = oce_mbox_macaddr_del(sc, sc->if_id, old_pmac_id);
2233 		bcopy((IF_LLADDR(sc->ifp)), sc->macaddr.mac_addr,
2234 				 sc->macaddr.size_of_struct);
2235 	}
2236 	if (status)
2237 		device_printf(sc->dev, "Failed update macaddress\n");
2238 
2239 }
2240 
2241 static int
2242 oce_handle_passthrough(struct ifnet *ifp, caddr_t data)
2243 {
2244 	POCE_SOFTC sc = ifp->if_softc;
2245 	struct ifreq *ifr = (struct ifreq *)data;
2246 	int rc = ENXIO;
2247 	char cookie[32] = {0};
2248 	void *priv_data = ifr_data_get_ptr(ifr);
2249 	void *ioctl_ptr;
2250 	uint32_t req_size;
2251 	struct mbx_hdr req;
2252 	OCE_DMA_MEM dma_mem;
2253 	struct mbx_common_get_cntl_attr *fw_cmd;
2254 
2255 	if (copyin(priv_data, cookie, strlen(IOCTL_COOKIE)))
2256 		return EFAULT;
2257 
2258 	if (memcmp(cookie, IOCTL_COOKIE, strlen(IOCTL_COOKIE)))
2259 		return EINVAL;
2260 
2261 	ioctl_ptr = (char *)priv_data + strlen(IOCTL_COOKIE);
2262 	if (copyin(ioctl_ptr, &req, sizeof(struct mbx_hdr)))
2263 		return EFAULT;
2264 
2265 	req_size = le32toh(req.u0.req.request_length);
2266 	if (req_size > 65536)
2267 		return EINVAL;
2268 
2269 	req_size += sizeof(struct mbx_hdr);
2270 	rc = oce_dma_alloc(sc, req_size, &dma_mem, 0);
2271 	if (rc)
2272 		return ENOMEM;
2273 
2274 	if (copyin(ioctl_ptr, OCE_DMAPTR(&dma_mem,char), req_size)) {
2275 		rc = EFAULT;
2276 		goto dma_free;
2277 	}
2278 
2279 	rc = oce_pass_through_mbox(sc, &dma_mem, req_size);
2280 	if (rc) {
2281 		rc = EIO;
2282 		goto dma_free;
2283 	}
2284 
2285 	if (copyout(OCE_DMAPTR(&dma_mem,char), ioctl_ptr, req_size))
2286 		rc =  EFAULT;
2287 
2288 	/*
2289 	   firmware is filling all the attributes for this ioctl except
2290 	   the driver version..so fill it
2291 	 */
2292 	if(req.u0.rsp.opcode == OPCODE_COMMON_GET_CNTL_ATTRIBUTES) {
2293 		fw_cmd = (struct mbx_common_get_cntl_attr *) ioctl_ptr;
2294 		strncpy(fw_cmd->params.rsp.cntl_attr_info.hba_attr.drv_ver_str,
2295 			COMPONENT_REVISION, strlen(COMPONENT_REVISION));
2296 	}
2297 
2298 dma_free:
2299 	oce_dma_free(sc, &dma_mem);
2300 	return rc;
2301 
2302 }
2303 
2304 static void
2305 oce_eqd_set_periodic(POCE_SOFTC sc)
2306 {
2307 	struct oce_set_eqd set_eqd[OCE_MAX_EQ];
2308 	struct oce_aic_obj *aic;
2309 	struct oce_eq *eqo;
2310 	uint64_t now = 0, delta;
2311 	int eqd, i, num = 0;
2312 	uint32_t tx_reqs = 0, rxpkts = 0, pps;
2313 	struct oce_wq *wq;
2314 	struct oce_rq *rq;
2315 
2316 	#define ticks_to_msecs(t)       (1000 * (t) / hz)
2317 
2318 	for (i = 0 ; i < sc->neqs; i++) {
2319 		eqo = sc->eq[i];
2320 		aic = &sc->aic_obj[i];
2321 		/* When setting the static eq delay from the user space */
2322 		if (!aic->enable) {
2323 			if (aic->ticks)
2324 				aic->ticks = 0;
2325 			eqd = aic->et_eqd;
2326 			goto modify_eqd;
2327 		}
2328 
2329 		if (i == 0) {
2330 			rq = sc->rq[0];
2331 			rxpkts = rq->rx_stats.rx_pkts;
2332 		} else
2333 			rxpkts = 0;
2334 		if (i + 1 < sc->nrqs) {
2335 			rq = sc->rq[i + 1];
2336 			rxpkts += rq->rx_stats.rx_pkts;
2337 		}
2338 		if (i < sc->nwqs) {
2339 			wq = sc->wq[i];
2340 			tx_reqs = wq->tx_stats.tx_reqs;
2341 		} else
2342 			tx_reqs = 0;
2343 		now = ticks;
2344 
2345 		if (!aic->ticks || now < aic->ticks ||
2346 		    rxpkts < aic->prev_rxpkts || tx_reqs < aic->prev_txreqs) {
2347 			aic->prev_rxpkts = rxpkts;
2348 			aic->prev_txreqs = tx_reqs;
2349 			aic->ticks = now;
2350 			continue;
2351 		}
2352 
2353 		delta = ticks_to_msecs(now - aic->ticks);
2354 
2355 		pps = (((uint32_t)(rxpkts - aic->prev_rxpkts) * 1000) / delta) +
2356 		      (((uint32_t)(tx_reqs - aic->prev_txreqs) * 1000) / delta);
2357 		eqd = (pps / 15000) << 2;
2358 		if (eqd < 8)
2359 			eqd = 0;
2360 
2361 		/* Make sure that the eq delay is in the known range */
2362 		eqd = min(eqd, aic->max_eqd);
2363 		eqd = max(eqd, aic->min_eqd);
2364 
2365 		aic->prev_rxpkts = rxpkts;
2366 		aic->prev_txreqs = tx_reqs;
2367 		aic->ticks = now;
2368 
2369 modify_eqd:
2370 		if (eqd != aic->cur_eqd) {
2371 			set_eqd[num].delay_multiplier = (eqd * 65)/100;
2372 			set_eqd[num].eq_id = eqo->eq_id;
2373 			aic->cur_eqd = eqd;
2374 			num++;
2375 		}
2376 	}
2377 
2378 	/* Is there atleast one eq that needs to be modified? */
2379         for(i = 0; i < num; i += 8) {
2380                 if((num - i) >=8 )
2381                         oce_mbox_eqd_modify_periodic(sc, &set_eqd[i], 8);
2382                 else
2383                         oce_mbox_eqd_modify_periodic(sc, &set_eqd[i], (num - i));
2384         }
2385 
2386 }
2387 
2388 static void oce_detect_hw_error(POCE_SOFTC sc)
2389 {
2390 
2391 	uint32_t ue_low = 0, ue_high = 0, ue_low_mask = 0, ue_high_mask = 0;
2392 	uint32_t sliport_status = 0, sliport_err1 = 0, sliport_err2 = 0;
2393 	uint32_t i;
2394 
2395 	if (sc->hw_error)
2396 		return;
2397 
2398 	if (IS_XE201(sc)) {
2399 		sliport_status = OCE_READ_REG32(sc, db, SLIPORT_STATUS_OFFSET);
2400 		if (sliport_status & SLIPORT_STATUS_ERR_MASK) {
2401 			sliport_err1 = OCE_READ_REG32(sc, db, SLIPORT_ERROR1_OFFSET);
2402 			sliport_err2 = OCE_READ_REG32(sc, db, SLIPORT_ERROR2_OFFSET);
2403 		}
2404 	} else {
2405 		ue_low = OCE_READ_REG32(sc, devcfg, PCICFG_UE_STATUS_LOW);
2406 		ue_high = OCE_READ_REG32(sc, devcfg, PCICFG_UE_STATUS_HIGH);
2407 		ue_low_mask = OCE_READ_REG32(sc, devcfg, PCICFG_UE_STATUS_LOW_MASK);
2408 		ue_high_mask = OCE_READ_REG32(sc, devcfg, PCICFG_UE_STATUS_HI_MASK);
2409 
2410 		ue_low = (ue_low & ~ue_low_mask);
2411 		ue_high = (ue_high & ~ue_high_mask);
2412 	}
2413 
2414 	/* On certain platforms BE hardware can indicate spurious UEs.
2415 	 * Allow the h/w to stop working completely in case of a real UE.
2416 	 * Hence not setting the hw_error for UE detection.
2417 	 */
2418 	if (sliport_status & SLIPORT_STATUS_ERR_MASK) {
2419 		sc->hw_error = TRUE;
2420 		device_printf(sc->dev, "Error detected in the card\n");
2421 	}
2422 
2423 	if (sliport_status & SLIPORT_STATUS_ERR_MASK) {
2424 		device_printf(sc->dev,
2425 				"ERR: sliport status 0x%x\n", sliport_status);
2426 		device_printf(sc->dev,
2427 				"ERR: sliport error1 0x%x\n", sliport_err1);
2428 		device_printf(sc->dev,
2429 				"ERR: sliport error2 0x%x\n", sliport_err2);
2430 	}
2431 
2432 	if (ue_low) {
2433 		for (i = 0; ue_low; ue_low >>= 1, i++) {
2434 			if (ue_low & 1)
2435 				device_printf(sc->dev, "UE: %s bit set\n",
2436 							ue_status_low_desc[i]);
2437 		}
2438 	}
2439 
2440 	if (ue_high) {
2441 		for (i = 0; ue_high; ue_high >>= 1, i++) {
2442 			if (ue_high & 1)
2443 				device_printf(sc->dev, "UE: %s bit set\n",
2444 							ue_status_hi_desc[i]);
2445 		}
2446 	}
2447 
2448 }
2449 
2450 static void
2451 oce_local_timer(void *arg)
2452 {
2453 	POCE_SOFTC sc = arg;
2454 	int i = 0;
2455 
2456 	oce_detect_hw_error(sc);
2457 	oce_refresh_nic_stats(sc);
2458 	oce_refresh_queue_stats(sc);
2459 	oce_mac_addr_set(sc);
2460 
2461 	/* TX Watch Dog*/
2462 	for (i = 0; i < sc->nwqs; i++)
2463 		oce_tx_restart(sc, sc->wq[i]);
2464 
2465 	/* calculate and set the eq delay for optimal interrupt rate */
2466 	if (IS_BE(sc) || IS_SH(sc))
2467 		oce_eqd_set_periodic(sc);
2468 
2469 	callout_reset(&sc->timer, hz, oce_local_timer, sc);
2470 }
2471 
2472 static void
2473 oce_tx_compl_clean(POCE_SOFTC sc)
2474 {
2475 	struct oce_wq *wq;
2476 	int i = 0, timeo = 0, num_wqes = 0;
2477 	int pending_txqs = sc->nwqs;
2478 
2479 	/* Stop polling for compls when HW has been silent for 10ms or
2480 	 * hw_error or no outstanding completions expected
2481 	 */
2482 	do {
2483 		pending_txqs = sc->nwqs;
2484 
2485 		for_all_wq_queues(sc, wq, i) {
2486 			num_wqes = oce_wq_handler(wq);
2487 
2488 			if(num_wqes)
2489 				timeo = 0;
2490 
2491 			if(!wq->ring->num_used)
2492 				pending_txqs--;
2493 		}
2494 
2495 		if (pending_txqs == 0 || ++timeo > 10 || sc->hw_error)
2496 			break;
2497 
2498 		DELAY(1000);
2499 	} while (TRUE);
2500 
2501 	for_all_wq_queues(sc, wq, i) {
2502 		while(wq->ring->num_used) {
2503 			LOCK(&wq->tx_compl_lock);
2504 			oce_process_tx_completion(wq);
2505 			UNLOCK(&wq->tx_compl_lock);
2506 		}
2507 	}
2508 
2509 }
2510 
2511 /* NOTE : This should only be called holding
2512  *        DEVICE_LOCK.
2513  */
2514 static void
2515 oce_if_deactivate(POCE_SOFTC sc)
2516 {
2517 	int i;
2518 	struct oce_rq *rq;
2519 	struct oce_wq *wq;
2520 	struct oce_eq *eq;
2521 
2522 	sc->ifp->if_drv_flags &= ~(IFF_DRV_RUNNING | IFF_DRV_OACTIVE);
2523 
2524 	oce_tx_compl_clean(sc);
2525 
2526 	/* Stop intrs and finish any bottom halves pending */
2527 	oce_hw_intr_disable(sc);
2528 
2529 	/* Since taskqueue_drain takes a Gaint Lock, We should not acquire
2530 	   any other lock. So unlock device lock and require after
2531 	   completing taskqueue_drain.
2532 	*/
2533 	UNLOCK(&sc->dev_lock);
2534 	for (i = 0; i < sc->intr_count; i++) {
2535 		if (sc->intrs[i].tq != NULL) {
2536 			taskqueue_drain(sc->intrs[i].tq, &sc->intrs[i].task);
2537 		}
2538 	}
2539 	LOCK(&sc->dev_lock);
2540 
2541 	/* Delete RX queue in card with flush param */
2542 	oce_stop_rx(sc);
2543 
2544 	/* Invalidate any pending cq and eq entries*/
2545 	for_all_evnt_queues(sc, eq, i)
2546 		oce_drain_eq(eq);
2547 	for_all_rq_queues(sc, rq, i)
2548 		oce_drain_rq_cq(rq);
2549 	for_all_wq_queues(sc, wq, i)
2550 		oce_drain_wq_cq(wq);
2551 
2552 	/* But still we need to get MCC aync events.
2553 	   So enable intrs and also arm first EQ
2554 	*/
2555 	oce_hw_intr_enable(sc);
2556 	oce_arm_eq(sc, sc->eq[0]->eq_id, 0, TRUE, FALSE);
2557 
2558 	DELAY(10);
2559 }
2560 
2561 static void
2562 oce_if_activate(POCE_SOFTC sc)
2563 {
2564 	struct oce_eq *eq;
2565 	struct oce_rq *rq;
2566 	struct oce_wq *wq;
2567 	int i, rc = 0;
2568 
2569 	sc->ifp->if_drv_flags |= IFF_DRV_RUNNING;
2570 
2571 	oce_hw_intr_disable(sc);
2572 
2573 	oce_start_rx(sc);
2574 
2575 	for_all_rq_queues(sc, rq, i) {
2576 		rc = oce_start_rq(rq);
2577 		if (rc)
2578 			device_printf(sc->dev, "Unable to start RX\n");
2579 	}
2580 
2581 	for_all_wq_queues(sc, wq, i) {
2582 		rc = oce_start_wq(wq);
2583 		if (rc)
2584 			device_printf(sc->dev, "Unable to start TX\n");
2585 	}
2586 
2587 	for_all_evnt_queues(sc, eq, i)
2588 		oce_arm_eq(sc, eq->eq_id, 0, TRUE, FALSE);
2589 
2590 	oce_hw_intr_enable(sc);
2591 
2592 }
2593 
2594 static void
2595 process_link_state(POCE_SOFTC sc, struct oce_async_cqe_link_state *acqe)
2596 {
2597 	/* Update Link status */
2598 	if ((acqe->u0.s.link_status & ~ASYNC_EVENT_LOGICAL) ==
2599 	     ASYNC_EVENT_LINK_UP) {
2600 		sc->link_status = ASYNC_EVENT_LINK_UP;
2601 		if_link_state_change(sc->ifp, LINK_STATE_UP);
2602 	} else {
2603 		sc->link_status = ASYNC_EVENT_LINK_DOWN;
2604 		if_link_state_change(sc->ifp, LINK_STATE_DOWN);
2605 	}
2606 }
2607 
2608 static void oce_async_grp5_osbmc_process(POCE_SOFTC sc,
2609 					 struct oce_async_evt_grp5_os2bmc *evt)
2610 {
2611 	DW_SWAP(evt, sizeof(struct oce_async_evt_grp5_os2bmc));
2612 	if (evt->u.s.mgmt_enable)
2613 		sc->flags |= OCE_FLAGS_OS2BMC;
2614 	else
2615 		return;
2616 
2617 	sc->bmc_filt_mask = evt->u.s.arp_filter;
2618 	sc->bmc_filt_mask |= (evt->u.s.dhcp_client_filt << 1);
2619 	sc->bmc_filt_mask |= (evt->u.s.dhcp_server_filt << 2);
2620 	sc->bmc_filt_mask |= (evt->u.s.net_bios_filt << 3);
2621 	sc->bmc_filt_mask |= (evt->u.s.bcast_filt << 4);
2622 	sc->bmc_filt_mask |= (evt->u.s.ipv6_nbr_filt << 5);
2623 	sc->bmc_filt_mask |= (evt->u.s.ipv6_ra_filt << 6);
2624 	sc->bmc_filt_mask |= (evt->u.s.ipv6_ras_filt << 7);
2625 	sc->bmc_filt_mask |= (evt->u.s.mcast_filt << 8);
2626 }
2627 
2628 static void oce_process_grp5_events(POCE_SOFTC sc, struct oce_mq_cqe *cqe)
2629 {
2630 	struct oce_async_event_grp5_pvid_state *gcqe;
2631 	struct oce_async_evt_grp5_os2bmc *bmccqe;
2632 
2633 	switch (cqe->u0.s.async_type) {
2634 	case ASYNC_EVENT_PVID_STATE:
2635 		/* GRP5 PVID */
2636 		gcqe = (struct oce_async_event_grp5_pvid_state *)cqe;
2637 		if (gcqe->enabled)
2638 			sc->pvid = gcqe->tag & VLAN_VID_MASK;
2639 		else
2640 			sc->pvid = 0;
2641 		break;
2642 	case ASYNC_EVENT_OS2BMC:
2643 		bmccqe = (struct oce_async_evt_grp5_os2bmc *)cqe;
2644 		oce_async_grp5_osbmc_process(sc, bmccqe);
2645 		break;
2646 	default:
2647 		break;
2648 	}
2649 }
2650 
2651 /* Handle the Completion Queue for the Mailbox/Async notifications */
2652 uint16_t
2653 oce_mq_handler(void *arg)
2654 {
2655 	struct oce_mq *mq = (struct oce_mq *)arg;
2656 	POCE_SOFTC sc = mq->parent;
2657 	struct oce_cq *cq = mq->cq;
2658 	int num_cqes = 0, evt_type = 0, optype = 0;
2659 	struct oce_mq_cqe *cqe;
2660 	struct oce_async_cqe_link_state *acqe;
2661 	struct oce_async_event_qnq *dbgcqe;
2662 
2663 	bus_dmamap_sync(cq->ring->dma.tag,
2664 			cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
2665 	cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_mq_cqe);
2666 
2667 	while (cqe->u0.dw[3]) {
2668 		DW_SWAP((uint32_t *) cqe, sizeof(oce_mq_cqe));
2669 		if (cqe->u0.s.async_event) {
2670 			evt_type = cqe->u0.s.event_type;
2671 			optype = cqe->u0.s.async_type;
2672 			if (evt_type  == ASYNC_EVENT_CODE_LINK_STATE) {
2673 				/* Link status evt */
2674 				acqe = (struct oce_async_cqe_link_state *)cqe;
2675 				process_link_state(sc, acqe);
2676 			} else if (evt_type == ASYNC_EVENT_GRP5) {
2677 				oce_process_grp5_events(sc, cqe);
2678 			} else if (evt_type == ASYNC_EVENT_CODE_DEBUG &&
2679 					optype == ASYNC_EVENT_DEBUG_QNQ) {
2680 				dbgcqe =  (struct oce_async_event_qnq *)cqe;
2681 				if(dbgcqe->valid)
2682 					sc->qnqid = dbgcqe->vlan_tag;
2683 				sc->qnq_debug_event = TRUE;
2684 			}
2685 		}
2686 		cqe->u0.dw[3] = 0;
2687 		RING_GET(cq->ring, 1);
2688 		bus_dmamap_sync(cq->ring->dma.tag,
2689 				cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
2690 		cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_mq_cqe);
2691 		num_cqes++;
2692 	}
2693 
2694 	if (num_cqes)
2695 		oce_arm_cq(sc, cq->cq_id, num_cqes, FALSE);
2696 
2697 	return 0;
2698 }
2699 
2700 static void
2701 setup_max_queues_want(POCE_SOFTC sc)
2702 {
2703 	/* Check if it is FLEX machine. Is so dont use RSS */
2704 	if ((sc->function_mode & FNM_FLEX10_MODE) ||
2705 	    (sc->function_mode & FNM_UMC_MODE)    ||
2706 	    (sc->function_mode & FNM_VNIC_MODE)	  ||
2707 	    (!is_rss_enabled(sc))		  ||
2708 	    IS_BE2(sc)) {
2709 		sc->nrqs = 1;
2710 		sc->nwqs = 1;
2711 	} else {
2712 		sc->nrqs = MIN(OCE_NCPUS, sc->nrssqs) + 1;
2713 		sc->nwqs = MIN(OCE_NCPUS, sc->nrssqs);
2714 	}
2715 
2716 	if (IS_BE2(sc) && is_rss_enabled(sc))
2717 		sc->nrqs = MIN(OCE_NCPUS, sc->nrssqs) + 1;
2718 }
2719 
2720 static void
2721 update_queues_got(POCE_SOFTC sc)
2722 {
2723 	if (is_rss_enabled(sc)) {
2724 		sc->nrqs = sc->intr_count + 1;
2725 		sc->nwqs = sc->intr_count;
2726 	} else {
2727 		sc->nrqs = 1;
2728 		sc->nwqs = 1;
2729 	}
2730 
2731 	if (IS_BE2(sc))
2732 		sc->nwqs = 1;
2733 }
2734 
2735 static int
2736 oce_check_ipv6_ext_hdr(struct mbuf *m)
2737 {
2738 	struct ether_header *eh = mtod(m, struct ether_header *);
2739 	caddr_t m_datatemp = m->m_data;
2740 
2741 	if (eh->ether_type == htons(ETHERTYPE_IPV6)) {
2742 		m->m_data += sizeof(struct ether_header);
2743 		struct ip6_hdr *ip6 = mtod(m, struct ip6_hdr *);
2744 
2745 		if((ip6->ip6_nxt != IPPROTO_TCP) && \
2746 				(ip6->ip6_nxt != IPPROTO_UDP)){
2747 			struct ip6_ext *ip6e = NULL;
2748 			m->m_data += sizeof(struct ip6_hdr);
2749 
2750 			ip6e = (struct ip6_ext *) mtod(m, struct ip6_ext *);
2751 			if(ip6e->ip6e_len == 0xff) {
2752 				m->m_data = m_datatemp;
2753 				return TRUE;
2754 			}
2755 		}
2756 		m->m_data = m_datatemp;
2757 	}
2758 	return FALSE;
2759 }
2760 
2761 static int
2762 is_be3_a1(POCE_SOFTC sc)
2763 {
2764 	if((sc->flags & OCE_FLAGS_BE3)  && ((sc->asic_revision & 0xFF) < 2)) {
2765 		return TRUE;
2766 	}
2767 	return FALSE;
2768 }
2769 
2770 static struct mbuf *
2771 oce_insert_vlan_tag(POCE_SOFTC sc, struct mbuf *m, boolean_t *complete)
2772 {
2773 	uint16_t vlan_tag = 0;
2774 
2775 	if(!M_WRITABLE(m))
2776 		return NULL;
2777 
2778 	/* Embed vlan tag in the packet if it is not part of it */
2779 	if(m->m_flags & M_VLANTAG) {
2780 		vlan_tag = EVL_VLANOFTAG(m->m_pkthdr.ether_vtag);
2781 		m->m_flags &= ~M_VLANTAG;
2782 	}
2783 
2784 	/* if UMC, ignore vlan tag insertion and instead insert pvid */
2785 	if(sc->pvid) {
2786 		if(!vlan_tag)
2787 			vlan_tag = sc->pvid;
2788 		if (complete)
2789 			*complete = FALSE;
2790 	}
2791 
2792 	if(vlan_tag) {
2793 		m = ether_vlanencap(m, vlan_tag);
2794 	}
2795 
2796 	if(sc->qnqid) {
2797 		m = ether_vlanencap(m, sc->qnqid);
2798 
2799 		if (complete)
2800 			*complete = FALSE;
2801 	}
2802 	return m;
2803 }
2804 
2805 static int
2806 oce_tx_asic_stall_verify(POCE_SOFTC sc, struct mbuf *m)
2807 {
2808 	if(is_be3_a1(sc) && IS_QNQ_OR_UMC(sc) && \
2809 			oce_check_ipv6_ext_hdr(m)) {
2810 		return TRUE;
2811 	}
2812 	return FALSE;
2813 }
2814 
2815 static void
2816 oce_get_config(POCE_SOFTC sc)
2817 {
2818 	int rc = 0;
2819 	uint32_t max_rss = 0;
2820 
2821 	if ((IS_BE(sc) || IS_SH(sc)) && (!sc->be3_native))
2822 		max_rss = OCE_LEGACY_MODE_RSS;
2823 	else
2824 		max_rss = OCE_MAX_RSS;
2825 
2826 	if (!IS_BE(sc)) {
2827 		rc = oce_get_profile_config(sc, max_rss);
2828 		if (rc) {
2829 			sc->nwqs = OCE_MAX_WQ;
2830 			sc->nrssqs = max_rss;
2831 			sc->nrqs = sc->nrssqs + 1;
2832 		}
2833 	}
2834 	else { /* For BE3 don't rely on fw for determining the resources */
2835 		sc->nrssqs = max_rss;
2836 		sc->nrqs = sc->nrssqs + 1;
2837 		sc->nwqs = OCE_MAX_WQ;
2838 		sc->max_vlans = MAX_VLANFILTER_SIZE;
2839 	}
2840 }
2841 
2842 static void
2843 oce_rdma_close(void)
2844 {
2845   if (oce_rdma_if != NULL) {
2846     oce_rdma_if = NULL;
2847   }
2848 }
2849 
2850 static void
2851 oce_get_mac_addr(POCE_SOFTC sc, uint8_t *macaddr)
2852 {
2853   memcpy(macaddr, sc->macaddr.mac_addr, 6);
2854 }
2855 
2856 int
2857 oce_register_rdma(POCE_RDMA_INFO rdma_info, POCE_RDMA_IF rdma_if)
2858 {
2859   POCE_SOFTC sc;
2860   struct oce_dev_info di;
2861   int i;
2862 
2863   if ((rdma_info == NULL) || (rdma_if == NULL)) {
2864     return -EINVAL;
2865   }
2866 
2867   if ((rdma_info->size != OCE_RDMA_INFO_SIZE) ||
2868       (rdma_if->size != OCE_RDMA_IF_SIZE)) {
2869     return -ENXIO;
2870   }
2871 
2872   rdma_info->close = oce_rdma_close;
2873   rdma_info->mbox_post = oce_mbox_post;
2874   rdma_info->common_req_hdr_init = mbx_common_req_hdr_init;
2875   rdma_info->get_mac_addr = oce_get_mac_addr;
2876 
2877   oce_rdma_if = rdma_if;
2878 
2879   sc = softc_head;
2880   while (sc != NULL) {
2881     if (oce_rdma_if->announce != NULL) {
2882       memset(&di, 0, sizeof(di));
2883       di.dev = sc->dev;
2884       di.softc = sc;
2885       di.ifp = sc->ifp;
2886       di.db_bhandle = sc->db_bhandle;
2887       di.db_btag = sc->db_btag;
2888       di.db_page_size = 4096;
2889       if (sc->flags & OCE_FLAGS_USING_MSIX) {
2890         di.intr_mode = OCE_INTERRUPT_MODE_MSIX;
2891       } else if (sc->flags & OCE_FLAGS_USING_MSI) {
2892         di.intr_mode = OCE_INTERRUPT_MODE_MSI;
2893       } else {
2894         di.intr_mode = OCE_INTERRUPT_MODE_INTX;
2895       }
2896       di.dev_family = OCE_GEN2_FAMILY; // fixme: must detect skyhawk
2897       if (di.intr_mode != OCE_INTERRUPT_MODE_INTX) {
2898         di.msix.num_vectors = sc->intr_count + sc->roce_intr_count;
2899         di.msix.start_vector = sc->intr_count;
2900         for (i=0; i<di.msix.num_vectors; i++) {
2901           di.msix.vector_list[i] = sc->intrs[i].vector;
2902         }
2903       } else {
2904       }
2905       memcpy(di.mac_addr, sc->macaddr.mac_addr, 6);
2906       di.vendor_id = pci_get_vendor(sc->dev);
2907       di.dev_id = pci_get_device(sc->dev);
2908 
2909       if (sc->rdma_flags & OCE_RDMA_FLAG_SUPPORTED) {
2910           di.flags  |= OCE_RDMA_INFO_RDMA_SUPPORTED;
2911       }
2912 
2913       rdma_if->announce(&di);
2914       sc = sc->next;
2915     }
2916   }
2917 
2918   return 0;
2919 }
2920 
2921 static void
2922 oce_read_env_variables( POCE_SOFTC sc )
2923 {
2924 	char *value = NULL;
2925 	int rc = 0;
2926 
2927         /* read if user wants to enable hwlro or swlro */
2928         //value = getenv("oce_enable_hwlro");
2929         if(value && IS_SH(sc)) {
2930                 sc->enable_hwlro = strtol(value, NULL, 10);
2931                 if(sc->enable_hwlro) {
2932                         rc = oce_mbox_nic_query_lro_capabilities(sc, NULL, NULL);
2933                         if(rc) {
2934                                 device_printf(sc->dev, "no hardware lro support\n");
2935                 		device_printf(sc->dev, "software lro enabled\n");
2936                                 sc->enable_hwlro = 0;
2937                         }else {
2938                                 device_printf(sc->dev, "hardware lro enabled\n");
2939 				oce_max_rsp_handled = 32;
2940                         }
2941                 }else {
2942                         device_printf(sc->dev, "software lro enabled\n");
2943                 }
2944         }else {
2945                 sc->enable_hwlro = 0;
2946         }
2947 
2948         /* read mbuf size */
2949         //value = getenv("oce_rq_buf_size");
2950         if(value && IS_SH(sc)) {
2951                 oce_rq_buf_size = strtol(value, NULL, 10);
2952                 switch(oce_rq_buf_size) {
2953                 case 2048:
2954                 case 4096:
2955                 case 9216:
2956                 case 16384:
2957                         break;
2958 
2959                 default:
2960                         device_printf(sc->dev, " Supported oce_rq_buf_size values are 2K, 4K, 9K, 16K \n");
2961                         oce_rq_buf_size = 2048;
2962                 }
2963         }
2964 
2965 	return;
2966 }
2967