xref: /freebsd/sys/dev/qlxge/qls_hw.c (revision d6b92ffa)
1 /*
2  * Copyright (c) 2013-2014 Qlogic Corporation
3  * All rights reserved.
4  *
5  *  Redistribution and use in source and binary forms, with or without
6  *  modification, are permitted provided that the following conditions
7  *  are met:
8  *
9  *  1. Redistributions of source code must retain the above copyright
10  *     notice, this list of conditions and the following disclaimer.
11  *  2. Redistributions in binary form must reproduce the above copyright
12  *     notice, this list of conditions and the following disclaimer in the
13  *     documentation and/or other materials provided with the distribution.
14  *
15  *  THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
16  *  AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17  *  IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18  *  ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
19  *  LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
20  *  CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
21  *  SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
22  *  INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
23  *  CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
24  *  ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
25  *  POSSIBILITY OF SUCH DAMAGE.
26  */
27 
28 /*
29  * File: qls_hw.c
30  * Author : David C Somayajulu, Qlogic Corporation, Aliso Viejo, CA 92656.
31  * Content: Contains Hardware dependent functions
32  */
33 #include <sys/cdefs.h>
34 __FBSDID("$FreeBSD$");
35 
36 
37 
38 #include "qls_os.h"
39 #include "qls_hw.h"
40 #include "qls_def.h"
41 #include "qls_inline.h"
42 #include "qls_ver.h"
43 #include "qls_glbl.h"
44 #include "qls_dbg.h"
45 
46 /*
47  * Static Functions
48  */
49 static int qls_wait_for_mac_proto_idx_ready(qla_host_t *ha, uint32_t op);
50 static int qls_config_unicast_mac_addr(qla_host_t *ha, uint32_t add_mac);
51 static int qls_config_mcast_mac_addr(qla_host_t *ha, uint8_t *mac_addr,
52                 uint32_t add_mac, uint32_t index);
53 
54 static int qls_init_rss(qla_host_t *ha);
55 static int qls_init_comp_queue(qla_host_t *ha, int cid);
56 static int qls_init_work_queue(qla_host_t *ha, int wid);
57 static int qls_init_fw_routing_table(qla_host_t *ha);
58 static int qls_hw_add_all_mcast(qla_host_t *ha);
59 static int qls_hw_add_mcast(qla_host_t *ha, uint8_t *mta);
60 static int qls_hw_del_mcast(qla_host_t *ha, uint8_t *mta);
61 static int qls_wait_for_flash_ready(qla_host_t *ha);
62 
63 static int qls_sem_lock(qla_host_t *ha, uint32_t mask, uint32_t value);
64 static void qls_sem_unlock(qla_host_t *ha, uint32_t mask);
65 
66 static void qls_free_tx_dma(qla_host_t *ha);
67 static int qls_alloc_tx_dma(qla_host_t *ha);
68 static void qls_free_rx_dma(qla_host_t *ha);
69 static int qls_alloc_rx_dma(qla_host_t *ha);
70 static void qls_free_mpi_dma(qla_host_t *ha);
71 static int qls_alloc_mpi_dma(qla_host_t *ha);
72 static void qls_free_rss_dma(qla_host_t *ha);
73 static int qls_alloc_rss_dma(qla_host_t *ha);
74 
75 static int qls_flash_validate(qla_host_t *ha, const char *signature);
76 
77 
78 static int qls_wait_for_proc_addr_ready(qla_host_t *ha);
79 static int qls_proc_addr_rd_reg(qla_host_t *ha, uint32_t addr_module,
80 		uint32_t reg, uint32_t *data);
81 static int qls_proc_addr_wr_reg(qla_host_t *ha, uint32_t addr_module,
82 		uint32_t reg, uint32_t data);
83 
84 static int qls_hw_reset(qla_host_t *ha);
85 
86 /*
87  * MPI Related Functions
88  */
89 static int qls_mbx_cmd(qla_host_t *ha, uint32_t *in_mbx, uint32_t i_count,
90 		uint32_t *out_mbx, uint32_t o_count);
91 static int qls_mbx_set_mgmt_ctrl(qla_host_t *ha, uint32_t t_ctrl);
92 static int qls_mbx_get_mgmt_ctrl(qla_host_t *ha, uint32_t *t_status);
93 static void qls_mbx_get_link_status(qla_host_t *ha);
94 static void qls_mbx_about_fw(qla_host_t *ha);
95 
96 int
97 qls_get_msix_count(qla_host_t *ha)
98 {
99 	return (ha->num_rx_rings);
100 }
101 
102 static int
103 qls_syctl_mpi_dump(SYSCTL_HANDLER_ARGS)
104 {
105         int err = 0, ret;
106         qla_host_t *ha;
107 
108         err = sysctl_handle_int(oidp, &ret, 0, req);
109 
110         if (err || !req->newptr)
111                 return (err);
112 
113 
114         if (ret == 1) {
115                 ha = (qla_host_t *)arg1;
116 		qls_mpi_core_dump(ha);
117         }
118         return (err);
119 }
120 
121 static int
122 qls_syctl_link_status(SYSCTL_HANDLER_ARGS)
123 {
124         int err = 0, ret;
125         qla_host_t *ha;
126 
127         err = sysctl_handle_int(oidp, &ret, 0, req);
128 
129         if (err || !req->newptr)
130                 return (err);
131 
132 
133         if (ret == 1) {
134                 ha = (qla_host_t *)arg1;
135 		qls_mbx_get_link_status(ha);
136 		qls_mbx_about_fw(ha);
137         }
138         return (err);
139 }
140 
141 void
142 qls_hw_add_sysctls(qla_host_t *ha)
143 {
144         device_t	dev;
145 
146         dev = ha->pci_dev;
147 
148 	ha->num_rx_rings = MAX_RX_RINGS; ha->num_tx_rings = MAX_TX_RINGS;
149 
150 	SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
151 		SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
152 		OID_AUTO, "num_rx_rings", CTLFLAG_RD, &ha->num_rx_rings,
153 		ha->num_rx_rings, "Number of Completion Queues");
154 
155         SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
156                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
157                 OID_AUTO, "num_tx_rings", CTLFLAG_RD, &ha->num_tx_rings,
158 		ha->num_tx_rings, "Number of Transmit Rings");
159 
160         SYSCTL_ADD_PROC(device_get_sysctl_ctx(dev),
161                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
162                 OID_AUTO, "mpi_dump", CTLTYPE_INT | CTLFLAG_RW,
163                 (void *)ha, 0,
164                 qls_syctl_mpi_dump, "I", "MPI Dump");
165 
166         SYSCTL_ADD_PROC(device_get_sysctl_ctx(dev),
167                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
168                 OID_AUTO, "link_status", CTLTYPE_INT | CTLFLAG_RW,
169                 (void *)ha, 0,
170                 qls_syctl_link_status, "I", "Link Status");
171 }
172 
173 /*
174  * Name: qls_free_dma
175  * Function: Frees the DMA'able memory allocated in qls_alloc_dma()
176  */
177 void
178 qls_free_dma(qla_host_t *ha)
179 {
180 	qls_free_rss_dma(ha);
181 	qls_free_mpi_dma(ha);
182 	qls_free_tx_dma(ha);
183 	qls_free_rx_dma(ha);
184 	return;
185 }
186 
187 /*
188  * Name: qls_alloc_dma
189  * Function: Allocates DMA'able memory for Tx/Rx Rings, Tx/Rx Contexts.
190  */
191 int
192 qls_alloc_dma(qla_host_t *ha)
193 {
194 	if (qls_alloc_rx_dma(ha))
195 		return (-1);
196 
197 	if (qls_alloc_tx_dma(ha)) {
198 		qls_free_rx_dma(ha);
199 		return (-1);
200 	}
201 
202 	if (qls_alloc_mpi_dma(ha)) {
203 		qls_free_tx_dma(ha);
204 		qls_free_rx_dma(ha);
205 		return (-1);
206 	}
207 
208 	if (qls_alloc_rss_dma(ha)) {
209 		qls_free_mpi_dma(ha);
210 		qls_free_tx_dma(ha);
211 		qls_free_rx_dma(ha);
212 		return (-1);
213 	}
214 
215 	return (0);
216 }
217 
218 
219 static int
220 qls_wait_for_mac_proto_idx_ready(qla_host_t *ha, uint32_t op)
221 {
222 	uint32_t data32;
223 	uint32_t count = 3;
224 
225 	while (count--) {
226 		data32 = READ_REG32(ha, Q81_CTL_MAC_PROTO_ADDR_INDEX);
227 
228 		if (data32 & op)
229 			return (0);
230 
231 		QLA_USEC_DELAY(100);
232 	}
233 	ha->qla_initiate_recovery = 1;
234 	return (-1);
235 }
236 
237 /*
238  * Name: qls_config_unicast_mac_addr
239  * Function: binds/unbinds a unicast MAC address to the interface.
240  */
241 static int
242 qls_config_unicast_mac_addr(qla_host_t *ha, uint32_t add_mac)
243 {
244 	int ret = 0;
245 	uint32_t mac_upper = 0;
246 	uint32_t mac_lower = 0;
247 	uint32_t value = 0, index;
248 
249 	if (qls_sem_lock(ha, Q81_CTL_SEM_MASK_MAC_SERDES,
250 		Q81_CTL_SEM_SET_MAC_SERDES)) {
251 		QL_DPRINT1((ha->pci_dev, "%s: semlock failed\n", __func__));
252 		return(-1);
253 	}
254 
255 	if (add_mac) {
256 		mac_upper = (ha->mac_addr[0] << 8) | ha->mac_addr[1];
257 		mac_lower = (ha->mac_addr[2] << 24) | (ha->mac_addr[3] << 16) |
258 				(ha->mac_addr[4] << 8) | ha->mac_addr[5];
259 	}
260 	ret = qls_wait_for_mac_proto_idx_ready(ha, Q81_CTL_MAC_PROTO_AI_MW);
261 	if (ret)
262 		goto qls_config_unicast_mac_addr_exit;
263 
264 	index = 128 * (ha->pci_func & 0x1); /* index */
265 
266 	value = (index << Q81_CTL_MAC_PROTO_AI_IDX_SHIFT) |
267 		Q81_CTL_MAC_PROTO_AI_TYPE_CAM_MAC;
268 
269 	WRITE_REG32(ha, Q81_CTL_MAC_PROTO_ADDR_INDEX, value);
270 	WRITE_REG32(ha, Q81_CTL_MAC_PROTO_ADDR_DATA, mac_lower);
271 
272 	ret = qls_wait_for_mac_proto_idx_ready(ha, Q81_CTL_MAC_PROTO_AI_MW);
273 	if (ret)
274 		goto qls_config_unicast_mac_addr_exit;
275 
276 	value = (index << Q81_CTL_MAC_PROTO_AI_IDX_SHIFT) |
277 		Q81_CTL_MAC_PROTO_AI_TYPE_CAM_MAC | 0x1;
278 
279 	WRITE_REG32(ha, Q81_CTL_MAC_PROTO_ADDR_INDEX, value);
280 	WRITE_REG32(ha, Q81_CTL_MAC_PROTO_ADDR_DATA, mac_upper);
281 
282 	ret = qls_wait_for_mac_proto_idx_ready(ha, Q81_CTL_MAC_PROTO_AI_MW);
283 	if (ret)
284 		goto qls_config_unicast_mac_addr_exit;
285 
286 	value = (index << Q81_CTL_MAC_PROTO_AI_IDX_SHIFT) |
287 		Q81_CTL_MAC_PROTO_AI_TYPE_CAM_MAC | 0x2;
288 
289 	WRITE_REG32(ha, Q81_CTL_MAC_PROTO_ADDR_INDEX, value);
290 
291 	value = Q81_CAM_MAC_OFF2_ROUTE_NIC |
292 			((ha->pci_func & 0x1) << Q81_CAM_MAC_OFF2_FUNC_SHIFT) |
293 			(0 << Q81_CAM_MAC_OFF2_CQID_SHIFT);
294 
295 	WRITE_REG32(ha, Q81_CTL_MAC_PROTO_ADDR_DATA, value);
296 
297 qls_config_unicast_mac_addr_exit:
298 	qls_sem_unlock(ha, Q81_CTL_SEM_MASK_MAC_SERDES);
299 	return (ret);
300 }
301 
302 /*
303  * Name: qls_config_mcast_mac_addr
304  * Function: binds/unbinds a multicast MAC address to the interface.
305  */
306 static int
307 qls_config_mcast_mac_addr(qla_host_t *ha, uint8_t *mac_addr, uint32_t add_mac,
308 	uint32_t index)
309 {
310 	int ret = 0;
311 	uint32_t mac_upper = 0;
312 	uint32_t mac_lower = 0;
313 	uint32_t value = 0;
314 
315 	if (qls_sem_lock(ha, Q81_CTL_SEM_MASK_MAC_SERDES,
316 		Q81_CTL_SEM_SET_MAC_SERDES)) {
317 		QL_DPRINT1((ha->pci_dev, "%s: semlock failed\n", __func__));
318 		return(-1);
319 	}
320 
321 	if (add_mac) {
322 		mac_upper = (mac_addr[0] << 8) | mac_addr[1];
323 		mac_lower = (mac_addr[2] << 24) | (mac_addr[3] << 16) |
324 				(mac_addr[4] << 8) | mac_addr[5];
325 	}
326 	ret = qls_wait_for_mac_proto_idx_ready(ha, Q81_CTL_MAC_PROTO_AI_MW);
327 	if (ret)
328 		goto qls_config_mcast_mac_addr_exit;
329 
330 	value = Q81_CTL_MAC_PROTO_AI_E |
331 			(index << Q81_CTL_MAC_PROTO_AI_IDX_SHIFT) |
332 			Q81_CTL_MAC_PROTO_AI_TYPE_MCAST ;
333 
334 	WRITE_REG32(ha, Q81_CTL_MAC_PROTO_ADDR_INDEX, value);
335 	WRITE_REG32(ha, Q81_CTL_MAC_PROTO_ADDR_DATA, mac_lower);
336 
337 	ret = qls_wait_for_mac_proto_idx_ready(ha, Q81_CTL_MAC_PROTO_AI_MW);
338 	if (ret)
339 		goto qls_config_mcast_mac_addr_exit;
340 
341 	value = Q81_CTL_MAC_PROTO_AI_E |
342 			(index << Q81_CTL_MAC_PROTO_AI_IDX_SHIFT) |
343 			Q81_CTL_MAC_PROTO_AI_TYPE_MCAST | 0x1;
344 
345 	WRITE_REG32(ha, Q81_CTL_MAC_PROTO_ADDR_INDEX, value);
346 	WRITE_REG32(ha, Q81_CTL_MAC_PROTO_ADDR_DATA, mac_upper);
347 
348 qls_config_mcast_mac_addr_exit:
349 	qls_sem_unlock(ha, Q81_CTL_SEM_MASK_MAC_SERDES);
350 
351 	return (ret);
352 }
353 
354 /*
355  * Name: qls_set_mac_rcv_mode
356  * Function: Enable/Disable AllMulticast and Promiscuous Modes.
357  */
358 static int
359 qls_wait_for_route_idx_ready(qla_host_t *ha, uint32_t op)
360 {
361 	uint32_t data32;
362 	uint32_t count = 3;
363 
364 	while (count--) {
365 		data32 = READ_REG32(ha, Q81_CTL_ROUTING_INDEX);
366 
367 		if (data32 & op)
368 			return (0);
369 
370 		QLA_USEC_DELAY(100);
371 	}
372 	ha->qla_initiate_recovery = 1;
373 	return (-1);
374 }
375 
376 static int
377 qls_load_route_idx_reg(qla_host_t *ha, uint32_t index, uint32_t data)
378 {
379 	int ret = 0;
380 
381 	ret = qls_wait_for_route_idx_ready(ha, Q81_CTL_RI_MW);
382 
383 	if (ret) {
384 		device_printf(ha->pci_dev, "%s: [0x%08x, 0x%08x] failed\n",
385 			__func__, index, data);
386 		goto qls_load_route_idx_reg_exit;
387 	}
388 
389 
390 	WRITE_REG32(ha, Q81_CTL_ROUTING_INDEX, index);
391 	WRITE_REG32(ha, Q81_CTL_ROUTING_DATA, data);
392 
393 qls_load_route_idx_reg_exit:
394 	return (ret);
395 }
396 
397 static int
398 qls_load_route_idx_reg_locked(qla_host_t *ha, uint32_t index, uint32_t data)
399 {
400 	int ret = 0;
401 
402 	if (qls_sem_lock(ha, Q81_CTL_SEM_MASK_RIDX_DATAREG,
403 		Q81_CTL_SEM_SET_RIDX_DATAREG)) {
404 		QL_DPRINT1((ha->pci_dev, "%s: semlock failed\n", __func__));
405 		return(-1);
406 	}
407 
408 	ret = qls_load_route_idx_reg(ha, index, data);
409 
410 	qls_sem_unlock(ha, Q81_CTL_SEM_MASK_RIDX_DATAREG);
411 
412 	return (ret);
413 }
414 
415 static int
416 qls_clear_routing_table(qla_host_t *ha)
417 {
418 	int i, ret = 0;
419 
420 	if (qls_sem_lock(ha, Q81_CTL_SEM_MASK_RIDX_DATAREG,
421 		Q81_CTL_SEM_SET_RIDX_DATAREG)) {
422 		QL_DPRINT1((ha->pci_dev, "%s: semlock failed\n", __func__));
423 		return(-1);
424 	}
425 
426 	for (i = 0; i < 16; i++) {
427 		ret = qls_load_route_idx_reg(ha, (Q81_CTL_RI_TYPE_NICQMASK|
428 			(i << 8) | Q81_CTL_RI_DST_DFLTQ), 0);
429 		if (ret)
430 			break;
431 	}
432 
433 	qls_sem_unlock(ha, Q81_CTL_SEM_MASK_RIDX_DATAREG);
434 
435 	return (ret);
436 }
437 
438 int
439 qls_set_promisc(qla_host_t *ha)
440 {
441 	int ret;
442 
443 	ret = qls_load_route_idx_reg_locked(ha,
444 			(Q81_CTL_RI_E | Q81_CTL_RI_TYPE_NICQMASK |
445 			Q81_CTL_RI_IDX_PROMISCUOUS | Q81_CTL_RI_DST_DFLTQ),
446 			Q81_CTL_RD_VALID_PKT);
447 	return (ret);
448 }
449 
450 void
451 qls_reset_promisc(qla_host_t *ha)
452 {
453 	int ret;
454 
455 	ret = qls_load_route_idx_reg_locked(ha, (Q81_CTL_RI_TYPE_NICQMASK |
456 			Q81_CTL_RI_IDX_PROMISCUOUS | Q81_CTL_RI_DST_DFLTQ), 0);
457 	return;
458 }
459 
460 int
461 qls_set_allmulti(qla_host_t *ha)
462 {
463 	int ret;
464 
465 	ret = qls_load_route_idx_reg_locked(ha,
466 			(Q81_CTL_RI_E | Q81_CTL_RI_TYPE_NICQMASK |
467 			Q81_CTL_RI_IDX_ALLMULTI | Q81_CTL_RI_DST_DFLTQ),
468 			Q81_CTL_RD_MCAST);
469 	return (ret);
470 }
471 
472 void
473 qls_reset_allmulti(qla_host_t *ha)
474 {
475 	int ret;
476 
477 	ret = qls_load_route_idx_reg_locked(ha, (Q81_CTL_RI_TYPE_NICQMASK |
478 			Q81_CTL_RI_IDX_ALLMULTI | Q81_CTL_RI_DST_DFLTQ), 0);
479 	return;
480 }
481 
482 
483 static int
484 qls_init_fw_routing_table(qla_host_t *ha)
485 {
486 	int ret = 0;
487 
488 	ret = qls_clear_routing_table(ha);
489 	if (ret)
490 		return (-1);
491 
492 	if (qls_sem_lock(ha, Q81_CTL_SEM_MASK_RIDX_DATAREG,
493 		Q81_CTL_SEM_SET_RIDX_DATAREG)) {
494 		QL_DPRINT1((ha->pci_dev, "%s: semlock failed\n", __func__));
495 		return(-1);
496 	}
497 
498 	ret = qls_load_route_idx_reg(ha, (Q81_CTL_RI_E | Q81_CTL_RI_DST_DROP |
499 			Q81_CTL_RI_TYPE_NICQMASK | Q81_CTL_RI_IDX_ALL_ERROR),
500 			Q81_CTL_RD_ERROR_PKT);
501 	if (ret)
502 		goto qls_init_fw_routing_table_exit;
503 
504 	ret = qls_load_route_idx_reg(ha, (Q81_CTL_RI_E | Q81_CTL_RI_DST_DFLTQ |
505 			Q81_CTL_RI_TYPE_NICQMASK | Q81_CTL_RI_IDX_BCAST),
506 			Q81_CTL_RD_BCAST);
507 	if (ret)
508 		goto qls_init_fw_routing_table_exit;
509 
510 	if (ha->num_rx_rings > 1 ) {
511 		ret = qls_load_route_idx_reg(ha,
512 				(Q81_CTL_RI_E | Q81_CTL_RI_DST_RSS |
513 				Q81_CTL_RI_TYPE_NICQMASK |
514 				Q81_CTL_RI_IDX_RSS_MATCH),
515 				Q81_CTL_RD_RSS_MATCH);
516 		if (ret)
517 			goto qls_init_fw_routing_table_exit;
518 	}
519 
520 	ret = qls_load_route_idx_reg(ha, (Q81_CTL_RI_E | Q81_CTL_RI_DST_DFLTQ |
521 			Q81_CTL_RI_TYPE_NICQMASK | Q81_CTL_RI_IDX_MCAST_MATCH),
522 			Q81_CTL_RD_MCAST_REG_MATCH);
523 	if (ret)
524 		goto qls_init_fw_routing_table_exit;
525 
526 	ret = qls_load_route_idx_reg(ha, (Q81_CTL_RI_E | Q81_CTL_RI_DST_DFLTQ |
527 			Q81_CTL_RI_TYPE_NICQMASK | Q81_CTL_RI_IDX_CAM_HIT),
528 			Q81_CTL_RD_CAM_HIT);
529 	if (ret)
530 		goto qls_init_fw_routing_table_exit;
531 
532 qls_init_fw_routing_table_exit:
533 	qls_sem_unlock(ha, Q81_CTL_SEM_MASK_RIDX_DATAREG);
534 	return (ret);
535 }
536 
537 static int
538 qls_tx_tso_chksum(qla_host_t *ha, struct mbuf *mp, q81_tx_tso_t *tx_mac)
539 {
540         struct ether_vlan_header *eh;
541         struct ip *ip;
542         struct ip6_hdr *ip6;
543 	struct tcphdr *th;
544         uint32_t ehdrlen, ip_hlen;
545 	int ret = 0;
546         uint16_t etype;
547         device_t dev;
548         uint8_t buf[sizeof(struct ip6_hdr)];
549 
550         dev = ha->pci_dev;
551 
552         eh = mtod(mp, struct ether_vlan_header *);
553 
554         if (eh->evl_encap_proto == htons(ETHERTYPE_VLAN)) {
555                 ehdrlen = ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN;
556                 etype = ntohs(eh->evl_proto);
557         } else {
558                 ehdrlen = ETHER_HDR_LEN;
559                 etype = ntohs(eh->evl_encap_proto);
560         }
561 
562         switch (etype) {
563                 case ETHERTYPE_IP:
564                         ip = (struct ip *)(mp->m_data + ehdrlen);
565 
566                         ip_hlen = sizeof (struct ip);
567 
568                         if (mp->m_len < (ehdrlen + ip_hlen)) {
569                                 m_copydata(mp, ehdrlen, sizeof(struct ip), buf);
570                                 ip = (struct ip *)buf;
571                         }
572 			tx_mac->opcode = Q81_IOCB_TX_TSO;
573 			tx_mac->flags |= Q81_TX_TSO_FLAGS_IPV4 ;
574 
575 			tx_mac->phdr_offsets = ehdrlen;
576 
577 			tx_mac->phdr_offsets |= ((ehdrlen + ip_hlen) <<
578 							Q81_TX_TSO_PHDR_SHIFT);
579 
580 			ip->ip_sum = 0;
581 
582 			if (mp->m_pkthdr.csum_flags & CSUM_TSO) {
583 				tx_mac->flags |= Q81_TX_TSO_FLAGS_LSO;
584 
585 				th = (struct tcphdr *)(ip + 1);
586 
587 				th->th_sum = in_pseudo(ip->ip_src.s_addr,
588 						ip->ip_dst.s_addr,
589 						htons(IPPROTO_TCP));
590 				tx_mac->mss = mp->m_pkthdr.tso_segsz;
591 				tx_mac->phdr_length = ip_hlen + ehdrlen +
592 							(th->th_off << 2);
593 				break;
594 			}
595 			tx_mac->vlan_off |= Q81_TX_TSO_VLAN_OFF_IC ;
596 
597 
598                         if (ip->ip_p == IPPROTO_TCP) {
599 				tx_mac->flags |= Q81_TX_TSO_FLAGS_TC;
600                         } else if (ip->ip_p == IPPROTO_UDP) {
601 				tx_mac->flags |= Q81_TX_TSO_FLAGS_UC;
602                         }
603                 break;
604 
605                 case ETHERTYPE_IPV6:
606                         ip6 = (struct ip6_hdr *)(mp->m_data + ehdrlen);
607 
608                         ip_hlen = sizeof(struct ip6_hdr);
609 
610                         if (mp->m_len < (ehdrlen + ip_hlen)) {
611                                 m_copydata(mp, ehdrlen, sizeof (struct ip6_hdr),
612                                         buf);
613                                 ip6 = (struct ip6_hdr *)buf;
614                         }
615 
616 			tx_mac->opcode = Q81_IOCB_TX_TSO;
617 			tx_mac->flags |= Q81_TX_TSO_FLAGS_IPV6 ;
618 			tx_mac->vlan_off |= Q81_TX_TSO_VLAN_OFF_IC ;
619 
620 			tx_mac->phdr_offsets = ehdrlen;
621 			tx_mac->phdr_offsets |= ((ehdrlen + ip_hlen) <<
622 							Q81_TX_TSO_PHDR_SHIFT);
623 
624                         if (ip6->ip6_nxt == IPPROTO_TCP) {
625 				tx_mac->flags |= Q81_TX_TSO_FLAGS_TC;
626                         } else if (ip6->ip6_nxt == IPPROTO_UDP) {
627 				tx_mac->flags |= Q81_TX_TSO_FLAGS_UC;
628                         }
629                 break;
630 
631                 default:
632                         ret = -1;
633                 break;
634         }
635 
636         return (ret);
637 }
638 
639 #define QLA_TX_MIN_FREE 2
640 int
641 qls_hw_tx_done(qla_host_t *ha, uint32_t txr_idx)
642 {
643 	uint32_t txr_done, txr_next;
644 
645 	txr_done = ha->tx_ring[txr_idx].txr_done;
646 	txr_next = ha->tx_ring[txr_idx].txr_next;
647 
648 	if (txr_done == txr_next) {
649 		ha->tx_ring[txr_idx].txr_free = NUM_TX_DESCRIPTORS;
650 	} else if (txr_done > txr_next) {
651 		ha->tx_ring[txr_idx].txr_free = txr_done - txr_next;
652 	} else {
653 		ha->tx_ring[txr_idx].txr_free = NUM_TX_DESCRIPTORS +
654 			txr_done - txr_next;
655 	}
656 
657 	if (ha->tx_ring[txr_idx].txr_free <= QLA_TX_MIN_FREE)
658 		return (-1);
659 
660 	return (0);
661 }
662 
663 /*
664  * Name: qls_hw_send
665  * Function: Transmits a packet. It first checks if the packet is a
666  *	candidate for Large TCP Segment Offload and then for UDP/TCP checksum
667  *	offload. If either of these creteria are not met, it is transmitted
668  *	as a regular ethernet frame.
669  */
670 int
671 qls_hw_send(qla_host_t *ha, bus_dma_segment_t *segs, int nsegs,
672 	uint32_t txr_next,  struct mbuf *mp, uint32_t txr_idx)
673 {
674         q81_tx_mac_t *tx_mac;
675 	q81_txb_desc_t *tx_desc;
676         uint32_t total_length = 0;
677         uint32_t i;
678         device_t dev;
679 	int ret = 0;
680 
681 	dev = ha->pci_dev;
682 
683         total_length = mp->m_pkthdr.len;
684 
685         if (total_length > QLA_MAX_TSO_FRAME_SIZE) {
686                 device_printf(dev, "%s: total length exceeds maxlen(%d)\n",
687                         __func__, total_length);
688                 return (-1);
689         }
690 
691 	if (ha->tx_ring[txr_idx].txr_free <= (NUM_TX_DESCRIPTORS >> 2)) {
692 		if (qls_hw_tx_done(ha, txr_idx)) {
693 			device_printf(dev, "%s: tx_free[%d] = %d\n",
694 				__func__, txr_idx,
695 				ha->tx_ring[txr_idx].txr_free);
696 			return (-1);
697 		}
698 	}
699 
700 	tx_mac = (q81_tx_mac_t *)&ha->tx_ring[txr_idx].wq_vaddr[txr_next];
701 
702 	bzero(tx_mac, sizeof(q81_tx_mac_t));
703 
704 	if ((mp->m_pkthdr.csum_flags &
705 			(CSUM_TCP | CSUM_UDP | CSUM_IP | CSUM_TSO)) != 0) {
706 
707 		ret = qls_tx_tso_chksum(ha, mp, (q81_tx_tso_t *)tx_mac);
708 		if (ret)
709 			return (EINVAL);
710 
711 		if (mp->m_pkthdr.csum_flags & CSUM_TSO)
712 			ha->tx_ring[txr_idx].tx_tso_frames++;
713 		else
714 			ha->tx_ring[txr_idx].tx_frames++;
715 
716 	} else {
717 		tx_mac->opcode = Q81_IOCB_TX_MAC;
718 	}
719 
720 	if (mp->m_flags & M_VLANTAG) {
721 
722 		tx_mac->vlan_tci = mp->m_pkthdr.ether_vtag;
723 		tx_mac->vlan_off |= Q81_TX_MAC_VLAN_OFF_V;
724 
725 		ha->tx_ring[txr_idx].tx_vlan_frames++;
726 	}
727 
728 	tx_mac->frame_length = total_length;
729 
730 	tx_mac->tid_lo = txr_next;
731 
732 	if (nsegs <= MAX_TX_MAC_DESC) {
733 
734 		QL_DPRINT2((dev, "%s: 1 [%d, %d]\n", __func__, total_length,
735 			tx_mac->tid_lo));
736 
737 		for (i = 0; i < nsegs; i++) {
738 			tx_mac->txd[i].baddr = segs->ds_addr;
739 			tx_mac->txd[i].length = segs->ds_len;
740 			segs++;
741 		}
742 		tx_mac->txd[(nsegs - 1)].flags = Q81_RXB_DESC_FLAGS_E;
743 
744 	} else {
745 		QL_DPRINT2((dev, "%s: 2 [%d, %d]\n", __func__, total_length,
746 			tx_mac->tid_lo));
747 
748 		tx_mac->txd[0].baddr =
749 			ha->tx_ring[txr_idx].tx_buf[txr_next].oal_paddr;
750 		tx_mac->txd[0].length =
751 			nsegs * (sizeof(q81_txb_desc_t));
752 		tx_mac->txd[0].flags = Q81_RXB_DESC_FLAGS_C;
753 
754 		tx_desc = ha->tx_ring[txr_idx].tx_buf[txr_next].oal_vaddr;
755 
756 		for (i = 0; i < nsegs; i++) {
757 			tx_desc->baddr = segs->ds_addr;
758 			tx_desc->length = segs->ds_len;
759 
760 			if (i == (nsegs -1))
761 				tx_desc->flags = Q81_RXB_DESC_FLAGS_E;
762 			else
763 				tx_desc->flags = 0;
764 
765 			segs++;
766 			tx_desc++;
767 		}
768 	}
769 	txr_next = (txr_next + 1) & (NUM_TX_DESCRIPTORS - 1);
770 	ha->tx_ring[txr_idx].txr_next = txr_next;
771 
772 	ha->tx_ring[txr_idx].txr_free--;
773 
774 	Q81_WR_WQ_PROD_IDX(txr_idx, txr_next);
775 
776 	return (0);
777 }
778 
779 /*
780  * Name: qls_del_hw_if
781  * Function: Destroys the hardware specific entities corresponding to an
782  *	Ethernet Interface
783  */
784 void
785 qls_del_hw_if(qla_host_t *ha)
786 {
787 	uint32_t value;
788 	int i;
789 	//int  count;
790 
791 	if (ha->hw_init == 0) {
792 		qls_hw_reset(ha);
793 		return;
794 	}
795 
796 	for (i = 0;  i < ha->num_tx_rings; i++) {
797 		Q81_SET_WQ_INVALID(i);
798 	}
799 	for (i = 0;  i < ha->num_rx_rings; i++) {
800 		Q81_SET_CQ_INVALID(i);
801 	}
802 
803 	for (i = 0; i < ha->num_rx_rings; i++) {
804 		Q81_DISABLE_INTR(ha, i); /* MSI-x i */
805 	}
806 
807 	value = (Q81_CTL_INTRE_IHD << Q81_CTL_INTRE_MASK_SHIFT);
808 	WRITE_REG32(ha, Q81_CTL_INTR_ENABLE, value);
809 
810 	value = (Q81_CTL_INTRE_EI << Q81_CTL_INTRE_MASK_SHIFT);
811 	WRITE_REG32(ha, Q81_CTL_INTR_ENABLE, value);
812 	ha->flags.intr_enable = 0;
813 
814 	qls_hw_reset(ha);
815 
816 	return;
817 }
818 
819 /*
820  * Name: qls_init_hw_if
821  * Function: Creates the hardware specific entities corresponding to an
822  *	Ethernet Interface - Transmit and Receive Contexts. Sets the MAC Address
823  *	corresponding to the interface. Enables LRO if allowed.
824  */
825 int
826 qls_init_hw_if(qla_host_t *ha)
827 {
828 	device_t	dev;
829 	uint32_t	value;
830 	int		ret = 0;
831 	int		i;
832 
833 
834 	QL_DPRINT2((ha->pci_dev, "%s:enter\n", __func__));
835 
836 	dev = ha->pci_dev;
837 
838 	ret = qls_hw_reset(ha);
839 	if (ret)
840 		goto qls_init_hw_if_exit;
841 
842 	ha->vm_pgsize = 4096;
843 
844 	/* Enable FAE and EFE bits in System Register */
845 	value = Q81_CTL_SYSTEM_ENABLE_FAE | Q81_CTL_SYSTEM_ENABLE_EFE;
846 	value = (value << Q81_CTL_SYSTEM_MASK_SHIFT) | value;
847 
848 	WRITE_REG32(ha, Q81_CTL_SYSTEM, value);
849 
850 	/* Set Default Completion Queue_ID in NIC Rcv Configuration Register */
851 	value = (Q81_CTL_NIC_RCVC_DCQ_MASK << Q81_CTL_NIC_RCVC_MASK_SHIFT);
852 	WRITE_REG32(ha, Q81_CTL_NIC_RCV_CONFIG, value);
853 
854 	/* Function Specific Control Register - Set Page Size and Enable NIC */
855 	value = Q81_CTL_FUNC_SPECIFIC_FE |
856 		Q81_CTL_FUNC_SPECIFIC_VM_PGSIZE_MASK |
857 		Q81_CTL_FUNC_SPECIFIC_EPC_O |
858 		Q81_CTL_FUNC_SPECIFIC_EPC_I |
859 		Q81_CTL_FUNC_SPECIFIC_EC;
860 	value = (value << Q81_CTL_FUNC_SPECIFIC_MASK_SHIFT) |
861                         Q81_CTL_FUNC_SPECIFIC_FE |
862 			Q81_CTL_FUNC_SPECIFIC_VM_PGSIZE_4K |
863 			Q81_CTL_FUNC_SPECIFIC_EPC_O |
864 			Q81_CTL_FUNC_SPECIFIC_EPC_I |
865 			Q81_CTL_FUNC_SPECIFIC_EC;
866 
867 	WRITE_REG32(ha, Q81_CTL_FUNC_SPECIFIC, value);
868 
869 	/* Interrupt Mask Register */
870 	value = Q81_CTL_INTRM_PI;
871 	value = (value << Q81_CTL_INTRM_MASK_SHIFT) | value;
872 
873 	WRITE_REG32(ha, Q81_CTL_INTR_MASK, value);
874 
875 	/* Initialiatize Completion Queue */
876 	for (i = 0; i < ha->num_rx_rings; i++) {
877 		ret = qls_init_comp_queue(ha, i);
878 		if (ret)
879 			goto qls_init_hw_if_exit;
880 	}
881 
882 	if (ha->num_rx_rings > 1 ) {
883 		ret = qls_init_rss(ha);
884 		if (ret)
885 			goto qls_init_hw_if_exit;
886 	}
887 
888 	/* Initialize Work Queue */
889 
890 	for (i = 0; i < ha->num_tx_rings; i++) {
891 		ret = qls_init_work_queue(ha, i);
892 		if (ret)
893 			goto qls_init_hw_if_exit;
894 	}
895 
896 	if (ret)
897 		goto qls_init_hw_if_exit;
898 
899 	/* Set up CAM RAM with MAC Address */
900 	ret = qls_config_unicast_mac_addr(ha, 1);
901 	if (ret)
902 		goto qls_init_hw_if_exit;
903 
904 	ret = qls_hw_add_all_mcast(ha);
905 	if (ret)
906 		goto qls_init_hw_if_exit;
907 
908 	/* Initialize Firmware Routing Table */
909 	ret = qls_init_fw_routing_table(ha);
910 	if (ret)
911 		goto qls_init_hw_if_exit;
912 
913 	/* Get Chip Revision ID */
914 	ha->rev_id = READ_REG32(ha, Q81_CTL_REV_ID);
915 
916 	/* Enable Global Interrupt */
917 	value = Q81_CTL_INTRE_EI;
918 	value = (value << Q81_CTL_INTRE_MASK_SHIFT) | value;
919 
920 	WRITE_REG32(ha, Q81_CTL_INTR_ENABLE, value);
921 
922 	/* Enable Interrupt Handshake Disable */
923 	value = Q81_CTL_INTRE_IHD;
924 	value = (value << Q81_CTL_INTRE_MASK_SHIFT) | value;
925 
926 	WRITE_REG32(ha, Q81_CTL_INTR_ENABLE, value);
927 
928 	/* Enable Completion Interrupt */
929 
930 	ha->flags.intr_enable = 1;
931 
932 	for (i = 0; i < ha->num_rx_rings; i++) {
933 		Q81_ENABLE_INTR(ha, i); /* MSI-x i */
934 	}
935 
936 	ha->hw_init = 1;
937 
938 	qls_mbx_get_link_status(ha);
939 
940 	QL_DPRINT2((ha->pci_dev, "%s:rxr [0x%08x]\n", __func__,
941 		ha->rx_ring[0].cq_db_offset));
942 	QL_DPRINT2((ha->pci_dev, "%s:txr [0x%08x]\n", __func__,
943 		ha->tx_ring[0].wq_db_offset));
944 
945 	for (i = 0; i < ha->num_rx_rings; i++) {
946 
947 		Q81_WR_CQ_CONS_IDX(i, 0);
948 		Q81_WR_LBQ_PROD_IDX(i, ha->rx_ring[i].lbq_in);
949 		Q81_WR_SBQ_PROD_IDX(i, ha->rx_ring[i].sbq_in);
950 
951 		QL_DPRINT2((dev, "%s: [wq_idx, cq_idx, lbq_idx, sbq_idx]"
952 			"[0x%08x, 0x%08x, 0x%08x, 0x%08x]\n", __func__,
953 			Q81_RD_WQ_IDX(i), Q81_RD_CQ_IDX(i), Q81_RD_LBQ_IDX(i),
954 			Q81_RD_SBQ_IDX(i)));
955 	}
956 
957 	for (i = 0; i < ha->num_rx_rings; i++) {
958 		Q81_SET_CQ_VALID(i);
959 	}
960 
961 qls_init_hw_if_exit:
962 	QL_DPRINT2((ha->pci_dev, "%s:exit\n", __func__));
963 	return (ret);
964 }
965 
966 static int
967 qls_wait_for_config_reg_bits(qla_host_t *ha, uint32_t bits, uint32_t value)
968 {
969 	uint32_t data32;
970 	uint32_t count = 3;
971 
972 	while (count--) {
973 
974 		data32 = READ_REG32(ha, Q81_CTL_CONFIG);
975 
976 		if ((data32 & bits) == value)
977 			return (0);
978 
979 		QLA_USEC_DELAY(100);
980 	}
981 	ha->qla_initiate_recovery = 1;
982 	device_printf(ha->pci_dev, "%s: failed\n", __func__);
983 	return (-1);
984 }
985 
986 static uint8_t q81_hash_key[] = {
987 			0xda, 0x56, 0x5a, 0x6d,
988 			0xc2, 0x0e, 0x5b, 0x25,
989 			0x3d, 0x25, 0x67, 0x41,
990 			0xb0, 0x8f, 0xa3, 0x43,
991 			0xcb, 0x2b, 0xca, 0xd0,
992 			0xb4, 0x30, 0x7b, 0xae,
993 			0xa3, 0x2d, 0xcb, 0x77,
994 			0x0c, 0xf2, 0x30, 0x80,
995 			0x3b, 0xb7, 0x42, 0x6a,
996 			0xfa, 0x01, 0xac, 0xbe };
997 
998 static int
999 qls_init_rss(qla_host_t *ha)
1000 {
1001 	q81_rss_icb_t	*rss_icb;
1002 	int		ret = 0;
1003 	int		i;
1004 	uint32_t	value;
1005 
1006 	rss_icb = ha->rss_dma.dma_b;
1007 
1008 	bzero(rss_icb, sizeof (q81_rss_icb_t));
1009 
1010 	rss_icb->flags_base_cq_num = Q81_RSS_ICB_FLAGS_L4K |
1011 				Q81_RSS_ICB_FLAGS_L6K | Q81_RSS_ICB_FLAGS_LI |
1012 				Q81_RSS_ICB_FLAGS_LB | Q81_RSS_ICB_FLAGS_LM |
1013 				Q81_RSS_ICB_FLAGS_RT4 | Q81_RSS_ICB_FLAGS_RT6;
1014 
1015 	rss_icb->mask = 0x3FF;
1016 
1017 	for (i = 0; i < Q81_RSS_ICB_NUM_INDTBL_ENTRIES; i++) {
1018 		rss_icb->cq_id[i] = (i & (ha->num_rx_rings - 1));
1019 	}
1020 
1021 	memcpy(rss_icb->ipv6_rss_hash_key, q81_hash_key, 40);
1022 	memcpy(rss_icb->ipv4_rss_hash_key, q81_hash_key, 16);
1023 
1024 	ret = qls_wait_for_config_reg_bits(ha, Q81_CTL_CONFIG_LR, 0);
1025 
1026 	if (ret)
1027 		goto qls_init_rss_exit;
1028 
1029 	ret = qls_sem_lock(ha, Q81_CTL_SEM_MASK_ICB, Q81_CTL_SEM_SET_ICB);
1030 
1031 	if (ret) {
1032 		QL_DPRINT1((ha->pci_dev, "%s: semlock failed\n", __func__));
1033 		goto qls_init_rss_exit;
1034 	}
1035 
1036 	value = (uint32_t)ha->rss_dma.dma_addr;
1037 	WRITE_REG32(ha, Q81_CTL_ICB_ACCESS_ADDR_LO, value);
1038 
1039 	value = (uint32_t)(ha->rss_dma.dma_addr >> 32);
1040 	WRITE_REG32(ha, Q81_CTL_ICB_ACCESS_ADDR_HI, value);
1041 
1042 	qls_sem_unlock(ha, Q81_CTL_SEM_MASK_ICB);
1043 
1044 	value = (Q81_CTL_CONFIG_LR << Q81_CTL_CONFIG_MASK_SHIFT) |
1045 			Q81_CTL_CONFIG_LR;
1046 
1047 	WRITE_REG32(ha, Q81_CTL_CONFIG, value);
1048 
1049 	ret = qls_wait_for_config_reg_bits(ha, Q81_CTL_CONFIG_LR, 0);
1050 
1051 qls_init_rss_exit:
1052 	return (ret);
1053 }
1054 
1055 static int
1056 qls_init_comp_queue(qla_host_t *ha, int cid)
1057 {
1058 	q81_cq_icb_t	*cq_icb;
1059 	qla_rx_ring_t	*rxr;
1060 	int		ret = 0;
1061 	uint32_t	value;
1062 
1063 	rxr = &ha->rx_ring[cid];
1064 
1065 	rxr->cq_db_offset = ha->vm_pgsize * (128 + cid);
1066 
1067 	cq_icb = rxr->cq_icb_vaddr;
1068 
1069 	bzero(cq_icb, sizeof (q81_cq_icb_t));
1070 
1071 	cq_icb->msix_vector = cid;
1072 	cq_icb->flags = Q81_CQ_ICB_FLAGS_LC |
1073 			Q81_CQ_ICB_FLAGS_LI |
1074 			Q81_CQ_ICB_FLAGS_LL |
1075 			Q81_CQ_ICB_FLAGS_LS |
1076 			Q81_CQ_ICB_FLAGS_LV;
1077 
1078 	cq_icb->length_v = NUM_CQ_ENTRIES;
1079 
1080 	cq_icb->cq_baddr_lo = (rxr->cq_base_paddr & 0xFFFFFFFF);
1081 	cq_icb->cq_baddr_hi = (rxr->cq_base_paddr >> 32) & 0xFFFFFFFF;
1082 
1083 	cq_icb->cqi_addr_lo = (rxr->cqi_paddr & 0xFFFFFFFF);
1084 	cq_icb->cqi_addr_hi = (rxr->cqi_paddr >> 32) & 0xFFFFFFFF;
1085 
1086 	cq_icb->pkt_idelay = 10;
1087 	cq_icb->idelay = 100;
1088 
1089 	cq_icb->lbq_baddr_lo = (rxr->lbq_addr_tbl_paddr & 0xFFFFFFFF);
1090 	cq_icb->lbq_baddr_hi = (rxr->lbq_addr_tbl_paddr >> 32) & 0xFFFFFFFF;
1091 
1092 	cq_icb->lbq_bsize = QLA_LGB_SIZE;
1093 	cq_icb->lbq_length = QLA_NUM_LGB_ENTRIES;
1094 
1095 	cq_icb->sbq_baddr_lo = (rxr->sbq_addr_tbl_paddr & 0xFFFFFFFF);
1096 	cq_icb->sbq_baddr_hi = (rxr->sbq_addr_tbl_paddr >> 32) & 0xFFFFFFFF;
1097 
1098 	cq_icb->sbq_bsize = (uint16_t)ha->msize;
1099 	cq_icb->sbq_length = QLA_NUM_SMB_ENTRIES;
1100 
1101 	QL_DUMP_CQ(ha);
1102 
1103 	ret = qls_wait_for_config_reg_bits(ha, Q81_CTL_CONFIG_LCQ, 0);
1104 
1105 	if (ret)
1106 		goto qls_init_comp_queue_exit;
1107 
1108 	ret = qls_sem_lock(ha, Q81_CTL_SEM_MASK_ICB, Q81_CTL_SEM_SET_ICB);
1109 
1110 	if (ret) {
1111 		QL_DPRINT1((ha->pci_dev, "%s: semlock failed\n", __func__));
1112 		goto qls_init_comp_queue_exit;
1113 	}
1114 
1115 	value = (uint32_t)rxr->cq_icb_paddr;
1116 	WRITE_REG32(ha, Q81_CTL_ICB_ACCESS_ADDR_LO, value);
1117 
1118 	value = (uint32_t)(rxr->cq_icb_paddr >> 32);
1119 	WRITE_REG32(ha, Q81_CTL_ICB_ACCESS_ADDR_HI, value);
1120 
1121 	qls_sem_unlock(ha, Q81_CTL_SEM_MASK_ICB);
1122 
1123 	value = Q81_CTL_CONFIG_LCQ | Q81_CTL_CONFIG_Q_NUM_MASK;
1124 	value = (value << Q81_CTL_CONFIG_MASK_SHIFT) | Q81_CTL_CONFIG_LCQ;
1125 	value |= (cid << Q81_CTL_CONFIG_Q_NUM_SHIFT);
1126 	WRITE_REG32(ha, Q81_CTL_CONFIG, value);
1127 
1128 	ret = qls_wait_for_config_reg_bits(ha, Q81_CTL_CONFIG_LCQ, 0);
1129 
1130 	rxr->cq_next = 0;
1131 	rxr->lbq_next = rxr->lbq_free = 0;
1132 	rxr->sbq_next = rxr->sbq_free = 0;
1133 	rxr->rx_free = rxr->rx_next = 0;
1134 	rxr->lbq_in = (QLA_NUM_LGB_ENTRIES - 1) & ~0xF;
1135 	rxr->sbq_in = (QLA_NUM_SMB_ENTRIES - 1) & ~0xF;
1136 
1137 qls_init_comp_queue_exit:
1138 	return (ret);
1139 }
1140 
1141 static int
1142 qls_init_work_queue(qla_host_t *ha, int wid)
1143 {
1144 	q81_wq_icb_t	*wq_icb;
1145 	qla_tx_ring_t	*txr;
1146 	int		ret = 0;
1147 	uint32_t	value;
1148 
1149 	txr = &ha->tx_ring[wid];
1150 
1151 	txr->wq_db_addr = (struct resource *)((uint8_t *)ha->pci_reg1
1152 						+ (ha->vm_pgsize * wid));
1153 
1154 	txr->wq_db_offset = (ha->vm_pgsize * wid);
1155 
1156 	wq_icb = txr->wq_icb_vaddr;
1157 	bzero(wq_icb, sizeof (q81_wq_icb_t));
1158 
1159 	wq_icb->length_v = NUM_TX_DESCRIPTORS  |
1160 				Q81_WQ_ICB_VALID;
1161 
1162 	wq_icb->flags = Q81_WQ_ICB_FLAGS_LO | Q81_WQ_ICB_FLAGS_LI |
1163 			Q81_WQ_ICB_FLAGS_LB | Q81_WQ_ICB_FLAGS_LC;
1164 
1165 	wq_icb->wqcqid_rss = wid;
1166 
1167 	wq_icb->baddr_lo = txr->wq_paddr & 0xFFFFFFFF;
1168 	wq_icb->baddr_hi = (txr->wq_paddr >> 32)& 0xFFFFFFFF;
1169 
1170 	wq_icb->ci_addr_lo = txr->txr_cons_paddr & 0xFFFFFFFF;
1171 	wq_icb->ci_addr_hi = (txr->txr_cons_paddr >> 32)& 0xFFFFFFFF;
1172 
1173 	ret = qls_wait_for_config_reg_bits(ha, Q81_CTL_CONFIG_LRQ, 0);
1174 
1175 	if (ret)
1176 		goto qls_init_wq_exit;
1177 
1178 	ret = qls_sem_lock(ha, Q81_CTL_SEM_MASK_ICB, Q81_CTL_SEM_SET_ICB);
1179 
1180 	if (ret) {
1181 		QL_DPRINT1((ha->pci_dev, "%s: semlock failed\n", __func__));
1182 		goto qls_init_wq_exit;
1183 	}
1184 
1185 	value = (uint32_t)txr->wq_icb_paddr;
1186 	WRITE_REG32(ha, Q81_CTL_ICB_ACCESS_ADDR_LO, value);
1187 
1188 	value = (uint32_t)(txr->wq_icb_paddr >> 32);
1189 	WRITE_REG32(ha, Q81_CTL_ICB_ACCESS_ADDR_HI, value);
1190 
1191 	qls_sem_unlock(ha, Q81_CTL_SEM_MASK_ICB);
1192 
1193 	value = Q81_CTL_CONFIG_LRQ | Q81_CTL_CONFIG_Q_NUM_MASK;
1194 	value = (value << Q81_CTL_CONFIG_MASK_SHIFT) | Q81_CTL_CONFIG_LRQ;
1195 	value |= (wid << Q81_CTL_CONFIG_Q_NUM_SHIFT);
1196 	WRITE_REG32(ha, Q81_CTL_CONFIG, value);
1197 
1198 	ret = qls_wait_for_config_reg_bits(ha, Q81_CTL_CONFIG_LRQ, 0);
1199 
1200 	txr->txr_free = NUM_TX_DESCRIPTORS;
1201 	txr->txr_next = 0;
1202 	txr->txr_done = 0;
1203 
1204 qls_init_wq_exit:
1205 	return (ret);
1206 }
1207 
1208 static int
1209 qls_hw_add_all_mcast(qla_host_t *ha)
1210 {
1211 	int i, nmcast;
1212 
1213 	nmcast = ha->nmcast;
1214 
1215 	for (i = 0 ; ((i < Q8_MAX_NUM_MULTICAST_ADDRS) && nmcast); i++) {
1216 		if ((ha->mcast[i].addr[0] != 0) ||
1217 			(ha->mcast[i].addr[1] != 0) ||
1218 			(ha->mcast[i].addr[2] != 0) ||
1219 			(ha->mcast[i].addr[3] != 0) ||
1220 			(ha->mcast[i].addr[4] != 0) ||
1221 			(ha->mcast[i].addr[5] != 0)) {
1222 
1223 			if (qls_config_mcast_mac_addr(ha, ha->mcast[i].addr,
1224 				1, i)) {
1225                 		device_printf(ha->pci_dev, "%s: failed\n",
1226 					__func__);
1227 				return (-1);
1228 			}
1229 
1230 			nmcast--;
1231 		}
1232 	}
1233 	return 0;
1234 }
1235 
1236 static int
1237 qls_hw_add_mcast(qla_host_t *ha, uint8_t *mta)
1238 {
1239 	int i;
1240 
1241 	for (i = 0; i < Q8_MAX_NUM_MULTICAST_ADDRS; i++) {
1242 
1243 		if (QL_MAC_CMP(ha->mcast[i].addr, mta) == 0)
1244 			return 0; /* its been already added */
1245 	}
1246 
1247 	for (i = 0; i < Q8_MAX_NUM_MULTICAST_ADDRS; i++) {
1248 
1249 		if ((ha->mcast[i].addr[0] == 0) &&
1250 			(ha->mcast[i].addr[1] == 0) &&
1251 			(ha->mcast[i].addr[2] == 0) &&
1252 			(ha->mcast[i].addr[3] == 0) &&
1253 			(ha->mcast[i].addr[4] == 0) &&
1254 			(ha->mcast[i].addr[5] == 0)) {
1255 
1256 			if (qls_config_mcast_mac_addr(ha, mta, 1, i))
1257 				return (-1);
1258 
1259 			bcopy(mta, ha->mcast[i].addr, Q8_MAC_ADDR_LEN);
1260 			ha->nmcast++;
1261 
1262 			return 0;
1263 		}
1264 	}
1265 	return 0;
1266 }
1267 
1268 static int
1269 qls_hw_del_mcast(qla_host_t *ha, uint8_t *mta)
1270 {
1271 	int i;
1272 
1273 	for (i = 0; i < Q8_MAX_NUM_MULTICAST_ADDRS; i++) {
1274 		if (QL_MAC_CMP(ha->mcast[i].addr, mta) == 0) {
1275 
1276 			if (qls_config_mcast_mac_addr(ha, mta, 0, i))
1277 				return (-1);
1278 
1279 			ha->mcast[i].addr[0] = 0;
1280 			ha->mcast[i].addr[1] = 0;
1281 			ha->mcast[i].addr[2] = 0;
1282 			ha->mcast[i].addr[3] = 0;
1283 			ha->mcast[i].addr[4] = 0;
1284 			ha->mcast[i].addr[5] = 0;
1285 
1286 			ha->nmcast--;
1287 
1288 			return 0;
1289 		}
1290 	}
1291 	return 0;
1292 }
1293 
1294 /*
1295  * Name: qls_hw_set_multi
1296  * Function: Sets the Multicast Addresses provided the host O.S into the
1297  *	hardware (for the given interface)
1298  */
1299 void
1300 qls_hw_set_multi(qla_host_t *ha, uint8_t *mta, uint32_t mcnt,
1301 	uint32_t add_mac)
1302 {
1303 	int i;
1304 
1305 	for (i = 0; i < mcnt; i++) {
1306 		if (add_mac) {
1307 			if (qls_hw_add_mcast(ha, mta))
1308 				break;
1309 		} else {
1310 			if (qls_hw_del_mcast(ha, mta))
1311 				break;
1312 		}
1313 
1314 		mta += Q8_MAC_ADDR_LEN;
1315 	}
1316 	return;
1317 }
1318 
1319 void
1320 qls_update_link_state(qla_host_t *ha)
1321 {
1322 	uint32_t link_state;
1323 	uint32_t prev_link_state;
1324 
1325 	if (!(ha->ifp->if_drv_flags & IFF_DRV_RUNNING)) {
1326 		ha->link_up = 0;
1327 		return;
1328 	}
1329 	link_state = READ_REG32(ha, Q81_CTL_STATUS);
1330 
1331 	prev_link_state =  ha->link_up;
1332 
1333 	if ((ha->pci_func & 0x1) == 0)
1334 		ha->link_up = ((link_state & Q81_CTL_STATUS_PL0)? 1 : 0);
1335 	else
1336 		ha->link_up = ((link_state & Q81_CTL_STATUS_PL1)? 1 : 0);
1337 
1338 	if (prev_link_state !=  ha->link_up) {
1339 
1340 
1341 		if (ha->link_up) {
1342 			if_link_state_change(ha->ifp, LINK_STATE_UP);
1343 		} else {
1344 			if_link_state_change(ha->ifp, LINK_STATE_DOWN);
1345 		}
1346 	}
1347 	return;
1348 }
1349 
1350 static void
1351 qls_free_tx_ring_dma(qla_host_t *ha, int r_idx)
1352 {
1353 	if (ha->tx_ring[r_idx].flags.wq_dma) {
1354 		qls_free_dmabuf(ha, &ha->tx_ring[r_idx].wq_dma);
1355 		ha->tx_ring[r_idx].flags.wq_dma = 0;
1356 	}
1357 
1358 	if (ha->tx_ring[r_idx].flags.privb_dma) {
1359 		qls_free_dmabuf(ha, &ha->tx_ring[r_idx].privb_dma);
1360 		ha->tx_ring[r_idx].flags.privb_dma = 0;
1361 	}
1362 	return;
1363 }
1364 
1365 static void
1366 qls_free_tx_dma(qla_host_t *ha)
1367 {
1368 	int i, j;
1369 	qla_tx_buf_t *txb;
1370 
1371 	for (i = 0; i < ha->num_tx_rings; i++) {
1372 
1373 		qls_free_tx_ring_dma(ha, i);
1374 
1375 		for (j = 0; j < NUM_TX_DESCRIPTORS; j++) {
1376 
1377 			txb = &ha->tx_ring[i].tx_buf[j];
1378 
1379 			if (txb->map) {
1380 				bus_dmamap_destroy(ha->tx_tag, txb->map);
1381 			}
1382 		}
1383 	}
1384 
1385         if (ha->tx_tag != NULL) {
1386                 bus_dma_tag_destroy(ha->tx_tag);
1387                 ha->tx_tag = NULL;
1388         }
1389 
1390 	return;
1391 }
1392 
1393 static int
1394 qls_alloc_tx_ring_dma(qla_host_t *ha, int ridx)
1395 {
1396 	int		ret = 0, i;
1397 	uint8_t		*v_addr;
1398 	bus_addr_t	p_addr;
1399 	qla_tx_buf_t	*txb;
1400 	device_t	dev = ha->pci_dev;
1401 
1402 	ha->tx_ring[ridx].wq_dma.alignment = 8;
1403 	ha->tx_ring[ridx].wq_dma.size =
1404 		NUM_TX_DESCRIPTORS * (sizeof (q81_tx_cmd_t));
1405 
1406 	ret = qls_alloc_dmabuf(ha, &ha->tx_ring[ridx].wq_dma);
1407 
1408 	if (ret) {
1409 		device_printf(dev, "%s: [%d] txr failed\n", __func__, ridx);
1410 		goto qls_alloc_tx_ring_dma_exit;
1411 	}
1412 	ha->tx_ring[ridx].flags.wq_dma = 1;
1413 
1414 	ha->tx_ring[ridx].privb_dma.alignment = 8;
1415 	ha->tx_ring[ridx].privb_dma.size = QLA_TX_PRIVATE_BSIZE;
1416 
1417 	ret = qls_alloc_dmabuf(ha, &ha->tx_ring[ridx].privb_dma);
1418 
1419 	if (ret) {
1420 		device_printf(dev, "%s: [%d] oalb failed\n", __func__, ridx);
1421 		goto qls_alloc_tx_ring_dma_exit;
1422 	}
1423 
1424 	ha->tx_ring[ridx].flags.privb_dma = 1;
1425 
1426 	ha->tx_ring[ridx].wq_vaddr = ha->tx_ring[ridx].wq_dma.dma_b;
1427 	ha->tx_ring[ridx].wq_paddr = ha->tx_ring[ridx].wq_dma.dma_addr;
1428 
1429 	v_addr = ha->tx_ring[ridx].privb_dma.dma_b;
1430 	p_addr = ha->tx_ring[ridx].privb_dma.dma_addr;
1431 
1432 	ha->tx_ring[ridx].wq_icb_vaddr = v_addr;
1433 	ha->tx_ring[ridx].wq_icb_paddr = p_addr;
1434 
1435 	ha->tx_ring[ridx].txr_cons_vaddr =
1436 		(uint32_t *)(v_addr + (PAGE_SIZE >> 1));
1437 	ha->tx_ring[ridx].txr_cons_paddr = p_addr + (PAGE_SIZE >> 1);
1438 
1439 	v_addr = v_addr + (PAGE_SIZE >> 1);
1440 	p_addr = p_addr + (PAGE_SIZE >> 1);
1441 
1442 	txb = ha->tx_ring[ridx].tx_buf;
1443 
1444 	for (i = 0; i < NUM_TX_DESCRIPTORS; i++) {
1445 
1446 		txb[i].oal_vaddr = v_addr;
1447 		txb[i].oal_paddr = p_addr;
1448 
1449 		v_addr = v_addr + QLA_OAL_BLK_SIZE;
1450 		p_addr = p_addr + QLA_OAL_BLK_SIZE;
1451 	}
1452 
1453 qls_alloc_tx_ring_dma_exit:
1454 	return (ret);
1455 }
1456 
1457 static int
1458 qls_alloc_tx_dma(qla_host_t *ha)
1459 {
1460 	int	i, j;
1461 	int	ret = 0;
1462 	qla_tx_buf_t *txb;
1463 
1464         if (bus_dma_tag_create(NULL,    /* parent */
1465                 1, 0,    /* alignment, bounds */
1466                 BUS_SPACE_MAXADDR,       /* lowaddr */
1467                 BUS_SPACE_MAXADDR,       /* highaddr */
1468                 NULL, NULL,      /* filter, filterarg */
1469                 QLA_MAX_TSO_FRAME_SIZE,     /* maxsize */
1470                 QLA_MAX_SEGMENTS,        /* nsegments */
1471                 PAGE_SIZE,        /* maxsegsize */
1472                 BUS_DMA_ALLOCNOW,        /* flags */
1473                 NULL,    /* lockfunc */
1474                 NULL,    /* lockfuncarg */
1475                 &ha->tx_tag)) {
1476                 device_printf(ha->pci_dev, "%s: tx_tag alloc failed\n",
1477                         __func__);
1478                 return (ENOMEM);
1479         }
1480 
1481 	for (i = 0; i < ha->num_tx_rings; i++) {
1482 
1483 		ret = qls_alloc_tx_ring_dma(ha, i);
1484 
1485 		if (ret) {
1486 			qls_free_tx_dma(ha);
1487 			break;
1488 		}
1489 
1490 		for (j = 0; j < NUM_TX_DESCRIPTORS; j++) {
1491 
1492 			txb = &ha->tx_ring[i].tx_buf[j];
1493 
1494 			ret = bus_dmamap_create(ha->tx_tag,
1495 				BUS_DMA_NOWAIT, &txb->map);
1496 			if (ret) {
1497 				ha->err_tx_dmamap_create++;
1498 				device_printf(ha->pci_dev,
1499 				"%s: bus_dmamap_create failed[%d, %d, %d]\n",
1500 				__func__, ret, i, j);
1501 
1502 				qls_free_tx_dma(ha);
1503 
1504                 		return (ret);
1505        			}
1506 		}
1507 	}
1508 
1509 	return (ret);
1510 }
1511 
1512 static void
1513 qls_free_rss_dma(qla_host_t *ha)
1514 {
1515 	qls_free_dmabuf(ha, &ha->rss_dma);
1516 	ha->flags.rss_dma = 0;
1517 }
1518 
1519 static int
1520 qls_alloc_rss_dma(qla_host_t *ha)
1521 {
1522 	int ret = 0;
1523 
1524 	ha->rss_dma.alignment = 4;
1525 	ha->rss_dma.size = PAGE_SIZE;
1526 
1527 	ret = qls_alloc_dmabuf(ha, &ha->rss_dma);
1528 
1529 	if (ret)
1530 		device_printf(ha->pci_dev, "%s: failed\n", __func__);
1531 	else
1532 		ha->flags.rss_dma = 1;
1533 
1534 	return (ret);
1535 }
1536 
1537 static void
1538 qls_free_mpi_dma(qla_host_t *ha)
1539 {
1540 	qls_free_dmabuf(ha, &ha->mpi_dma);
1541 	ha->flags.mpi_dma = 0;
1542 }
1543 
1544 static int
1545 qls_alloc_mpi_dma(qla_host_t *ha)
1546 {
1547 	int ret = 0;
1548 
1549 	ha->mpi_dma.alignment = 4;
1550 	ha->mpi_dma.size = (0x4000 * 4);
1551 
1552 	ret = qls_alloc_dmabuf(ha, &ha->mpi_dma);
1553 	if (ret)
1554 		device_printf(ha->pci_dev, "%s: failed\n", __func__);
1555 	else
1556 		ha->flags.mpi_dma = 1;
1557 
1558 	return (ret);
1559 }
1560 
1561 static void
1562 qls_free_rx_ring_dma(qla_host_t *ha, int ridx)
1563 {
1564 	if (ha->rx_ring[ridx].flags.cq_dma) {
1565 		qls_free_dmabuf(ha, &ha->rx_ring[ridx].cq_dma);
1566 		ha->rx_ring[ridx].flags.cq_dma = 0;
1567 	}
1568 
1569 	if (ha->rx_ring[ridx].flags.lbq_dma) {
1570 		qls_free_dmabuf(ha, &ha->rx_ring[ridx].lbq_dma);
1571 		ha->rx_ring[ridx].flags.lbq_dma = 0;
1572 	}
1573 
1574 	if (ha->rx_ring[ridx].flags.sbq_dma) {
1575 		qls_free_dmabuf(ha, &ha->rx_ring[ridx].sbq_dma);
1576 		ha->rx_ring[ridx].flags.sbq_dma = 0;
1577 	}
1578 
1579 	if (ha->rx_ring[ridx].flags.lb_dma) {
1580 		qls_free_dmabuf(ha, &ha->rx_ring[ridx].lb_dma);
1581 		ha->rx_ring[ridx].flags.lb_dma = 0;
1582 	}
1583 	return;
1584 }
1585 
1586 static void
1587 qls_free_rx_dma(qla_host_t *ha)
1588 {
1589 	int i;
1590 
1591 	for (i = 0; i < ha->num_rx_rings; i++) {
1592 		qls_free_rx_ring_dma(ha, i);
1593 	}
1594 
1595         if (ha->rx_tag != NULL) {
1596                 bus_dma_tag_destroy(ha->rx_tag);
1597                 ha->rx_tag = NULL;
1598         }
1599 
1600 	return;
1601 }
1602 
1603 static int
1604 qls_alloc_rx_ring_dma(qla_host_t *ha, int ridx)
1605 {
1606 	int				i, ret = 0;
1607 	uint8_t				*v_addr;
1608 	bus_addr_t			p_addr;
1609 	volatile q81_bq_addr_e_t	*bq_e;
1610 	device_t			dev = ha->pci_dev;
1611 
1612 	ha->rx_ring[ridx].cq_dma.alignment = 128;
1613 	ha->rx_ring[ridx].cq_dma.size =
1614 		(NUM_CQ_ENTRIES * (sizeof (q81_cq_e_t))) + PAGE_SIZE;
1615 
1616 	ret = qls_alloc_dmabuf(ha, &ha->rx_ring[ridx].cq_dma);
1617 
1618 	if (ret) {
1619 		device_printf(dev, "%s: [%d] cq failed\n", __func__, ridx);
1620 		goto qls_alloc_rx_ring_dma_exit;
1621 	}
1622 	ha->rx_ring[ridx].flags.cq_dma = 1;
1623 
1624 	ha->rx_ring[ridx].lbq_dma.alignment = 8;
1625 	ha->rx_ring[ridx].lbq_dma.size = QLA_LGBQ_AND_TABLE_SIZE;
1626 
1627 	ret = qls_alloc_dmabuf(ha, &ha->rx_ring[ridx].lbq_dma);
1628 
1629 	if (ret) {
1630 		device_printf(dev, "%s: [%d] lbq failed\n", __func__, ridx);
1631 		goto qls_alloc_rx_ring_dma_exit;
1632 	}
1633 	ha->rx_ring[ridx].flags.lbq_dma = 1;
1634 
1635 	ha->rx_ring[ridx].sbq_dma.alignment = 8;
1636 	ha->rx_ring[ridx].sbq_dma.size = QLA_SMBQ_AND_TABLE_SIZE;
1637 
1638 	ret = qls_alloc_dmabuf(ha, &ha->rx_ring[ridx].sbq_dma);
1639 
1640 	if (ret) {
1641 		device_printf(dev, "%s: [%d] sbq failed\n", __func__, ridx);
1642 		goto qls_alloc_rx_ring_dma_exit;
1643 	}
1644 	ha->rx_ring[ridx].flags.sbq_dma = 1;
1645 
1646 	ha->rx_ring[ridx].lb_dma.alignment = 8;
1647 	ha->rx_ring[ridx].lb_dma.size = (QLA_LGB_SIZE * QLA_NUM_LGB_ENTRIES);
1648 
1649 	ret = qls_alloc_dmabuf(ha, &ha->rx_ring[ridx].lb_dma);
1650 	if (ret) {
1651 		device_printf(dev, "%s: [%d] lb failed\n", __func__, ridx);
1652 		goto qls_alloc_rx_ring_dma_exit;
1653 	}
1654 	ha->rx_ring[ridx].flags.lb_dma = 1;
1655 
1656 	bzero(ha->rx_ring[ridx].cq_dma.dma_b, ha->rx_ring[ridx].cq_dma.size);
1657 	bzero(ha->rx_ring[ridx].lbq_dma.dma_b, ha->rx_ring[ridx].lbq_dma.size);
1658 	bzero(ha->rx_ring[ridx].sbq_dma.dma_b, ha->rx_ring[ridx].sbq_dma.size);
1659 	bzero(ha->rx_ring[ridx].lb_dma.dma_b, ha->rx_ring[ridx].lb_dma.size);
1660 
1661 	/* completion queue */
1662 	ha->rx_ring[ridx].cq_base_vaddr = ha->rx_ring[ridx].cq_dma.dma_b;
1663 	ha->rx_ring[ridx].cq_base_paddr = ha->rx_ring[ridx].cq_dma.dma_addr;
1664 
1665 	v_addr = ha->rx_ring[ridx].cq_dma.dma_b;
1666 	p_addr = ha->rx_ring[ridx].cq_dma.dma_addr;
1667 
1668 	v_addr = v_addr + (NUM_CQ_ENTRIES * (sizeof (q81_cq_e_t)));
1669 	p_addr = p_addr + (NUM_CQ_ENTRIES * (sizeof (q81_cq_e_t)));
1670 
1671 	/* completion queue icb */
1672 	ha->rx_ring[ridx].cq_icb_vaddr = v_addr;
1673 	ha->rx_ring[ridx].cq_icb_paddr = p_addr;
1674 
1675 	v_addr = v_addr + (PAGE_SIZE >> 2);
1676 	p_addr = p_addr + (PAGE_SIZE >> 2);
1677 
1678 	/* completion queue index register */
1679 	ha->rx_ring[ridx].cqi_vaddr = (uint32_t *)v_addr;
1680 	ha->rx_ring[ridx].cqi_paddr = p_addr;
1681 
1682 	v_addr = ha->rx_ring[ridx].lbq_dma.dma_b;
1683 	p_addr = ha->rx_ring[ridx].lbq_dma.dma_addr;
1684 
1685 	/* large buffer queue address table */
1686 	ha->rx_ring[ridx].lbq_addr_tbl_vaddr = v_addr;
1687 	ha->rx_ring[ridx].lbq_addr_tbl_paddr = p_addr;
1688 
1689 	/* large buffer queue */
1690 	ha->rx_ring[ridx].lbq_vaddr = v_addr + PAGE_SIZE;
1691 	ha->rx_ring[ridx].lbq_paddr = p_addr + PAGE_SIZE;
1692 
1693 	v_addr = ha->rx_ring[ridx].sbq_dma.dma_b;
1694 	p_addr = ha->rx_ring[ridx].sbq_dma.dma_addr;
1695 
1696 	/* small buffer queue address table */
1697 	ha->rx_ring[ridx].sbq_addr_tbl_vaddr = v_addr;
1698 	ha->rx_ring[ridx].sbq_addr_tbl_paddr = p_addr;
1699 
1700 	/* small buffer queue */
1701 	ha->rx_ring[ridx].sbq_vaddr = v_addr + PAGE_SIZE;
1702 	ha->rx_ring[ridx].sbq_paddr = p_addr + PAGE_SIZE;
1703 
1704 	ha->rx_ring[ridx].lb_vaddr = ha->rx_ring[ridx].lb_dma.dma_b;
1705 	ha->rx_ring[ridx].lb_paddr = ha->rx_ring[ridx].lb_dma.dma_addr;
1706 
1707 	/* Initialize Large Buffer Queue Table */
1708 
1709 	p_addr = ha->rx_ring[ridx].lbq_paddr;
1710 	bq_e = ha->rx_ring[ridx].lbq_addr_tbl_vaddr;
1711 
1712 	bq_e->addr_lo = p_addr & 0xFFFFFFFF;
1713 	bq_e->addr_hi = (p_addr >> 32) & 0xFFFFFFFF;
1714 
1715 	p_addr = ha->rx_ring[ridx].lb_paddr;
1716 	bq_e = ha->rx_ring[ridx].lbq_vaddr;
1717 
1718 	for (i = 0; i < QLA_NUM_LGB_ENTRIES; i++) {
1719 		bq_e->addr_lo = p_addr & 0xFFFFFFFF;
1720 		bq_e->addr_hi = (p_addr >> 32) & 0xFFFFFFFF;
1721 
1722 		p_addr = p_addr + QLA_LGB_SIZE;
1723 		bq_e++;
1724 	}
1725 
1726 	/* Initialize Small Buffer Queue Table */
1727 
1728 	p_addr = ha->rx_ring[ridx].sbq_paddr;
1729 	bq_e = ha->rx_ring[ridx].sbq_addr_tbl_vaddr;
1730 
1731 	for (i =0; i < (QLA_SBQ_SIZE/QLA_PAGE_SIZE); i++) {
1732 		bq_e->addr_lo = p_addr & 0xFFFFFFFF;
1733 		bq_e->addr_hi = (p_addr >> 32) & 0xFFFFFFFF;
1734 
1735 		p_addr = p_addr + QLA_PAGE_SIZE;
1736 		bq_e++;
1737 	}
1738 
1739 qls_alloc_rx_ring_dma_exit:
1740 	return (ret);
1741 }
1742 
1743 static int
1744 qls_alloc_rx_dma(qla_host_t *ha)
1745 {
1746 	int	i;
1747 	int	ret = 0;
1748 
1749         if (bus_dma_tag_create(NULL,    /* parent */
1750                         1, 0,    /* alignment, bounds */
1751                         BUS_SPACE_MAXADDR,       /* lowaddr */
1752                         BUS_SPACE_MAXADDR,       /* highaddr */
1753                         NULL, NULL,      /* filter, filterarg */
1754                         MJUM9BYTES,     /* maxsize */
1755                         1,        /* nsegments */
1756                         MJUM9BYTES,        /* maxsegsize */
1757                         BUS_DMA_ALLOCNOW,        /* flags */
1758                         NULL,    /* lockfunc */
1759                         NULL,    /* lockfuncarg */
1760                         &ha->rx_tag)) {
1761 
1762                 device_printf(ha->pci_dev, "%s: rx_tag alloc failed\n",
1763                         __func__);
1764 
1765                 return (ENOMEM);
1766         }
1767 
1768 	for (i = 0; i < ha->num_rx_rings; i++) {
1769 		ret = qls_alloc_rx_ring_dma(ha, i);
1770 
1771 		if (ret) {
1772 			qls_free_rx_dma(ha);
1773 			break;
1774 		}
1775 	}
1776 
1777 	return (ret);
1778 }
1779 
1780 static int
1781 qls_wait_for_flash_ready(qla_host_t *ha)
1782 {
1783 	uint32_t data32;
1784 	uint32_t count = 3;
1785 
1786 	while (count--) {
1787 
1788 		data32 = READ_REG32(ha, Q81_CTL_FLASH_ADDR);
1789 
1790 		if (data32 & Q81_CTL_FLASH_ADDR_ERR)
1791 			goto qls_wait_for_flash_ready_exit;
1792 
1793 		if (data32 & Q81_CTL_FLASH_ADDR_RDY)
1794 			return (0);
1795 
1796 		QLA_USEC_DELAY(100);
1797 	}
1798 
1799 qls_wait_for_flash_ready_exit:
1800 	QL_DPRINT1((ha->pci_dev, "%s: failed\n", __func__));
1801 
1802 	return (-1);
1803 }
1804 
1805 /*
1806  * Name: qls_rd_flash32
1807  * Function: Read Flash Memory
1808  */
1809 int
1810 qls_rd_flash32(qla_host_t *ha, uint32_t addr, uint32_t *data)
1811 {
1812 	int ret;
1813 
1814 	ret = qls_wait_for_flash_ready(ha);
1815 
1816 	if (ret)
1817 		return (ret);
1818 
1819 	WRITE_REG32(ha, Q81_CTL_FLASH_ADDR, (addr | Q81_CTL_FLASH_ADDR_R));
1820 
1821 	ret = qls_wait_for_flash_ready(ha);
1822 
1823 	if (ret)
1824 		return (ret);
1825 
1826 	*data = READ_REG32(ha, Q81_CTL_FLASH_DATA);
1827 
1828 	return 0;
1829 }
1830 
1831 static int
1832 qls_flash_validate(qla_host_t *ha, const char *signature)
1833 {
1834 	uint16_t csum16 = 0;
1835 	uint16_t *data16;
1836 	int i;
1837 
1838 	if (bcmp(ha->flash.id, signature, 4)) {
1839 		QL_DPRINT1((ha->pci_dev, "%s: invalid signature "
1840 			"%x:%x:%x:%x %s\n", __func__, ha->flash.id[0],
1841 			ha->flash.id[1], ha->flash.id[2], ha->flash.id[3],
1842 			signature));
1843 		return(-1);
1844 	}
1845 
1846 	data16 = (uint16_t *)&ha->flash;
1847 
1848 	for (i = 0; i < (sizeof (q81_flash_t) >> 1); i++) {
1849 		csum16 += *data16++;
1850 	}
1851 
1852 	if (csum16) {
1853 		QL_DPRINT1((ha->pci_dev, "%s: invalid checksum\n", __func__));
1854 		return(-1);
1855 	}
1856 	return(0);
1857 }
1858 
1859 int
1860 qls_rd_nic_params(qla_host_t *ha)
1861 {
1862 	int		i, ret = 0;
1863 	uint32_t	faddr;
1864 	uint32_t	*qflash;
1865 
1866 	if (qls_sem_lock(ha, Q81_CTL_SEM_MASK_FLASH, Q81_CTL_SEM_SET_FLASH)) {
1867 		QL_DPRINT1((ha->pci_dev, "%s: semlock failed\n", __func__));
1868 		return(-1);
1869 	}
1870 
1871 	if ((ha->pci_func & 0x1) == 0)
1872 		faddr = Q81_F0_FLASH_OFFSET >> 2;
1873 	else
1874 		faddr = Q81_F1_FLASH_OFFSET >> 2;
1875 
1876 	qflash = (uint32_t *)&ha->flash;
1877 
1878 	for (i = 0; i < (sizeof(q81_flash_t) >> 2) ; i++) {
1879 
1880 		ret = qls_rd_flash32(ha, faddr, qflash);
1881 
1882 		if (ret)
1883 			goto qls_rd_flash_data_exit;
1884 
1885 		faddr++;
1886 		qflash++;
1887 	}
1888 
1889 	QL_DUMP_BUFFER8(ha, __func__, (&ha->flash), (sizeof (q81_flash_t)));
1890 
1891 	ret = qls_flash_validate(ha, Q81_FLASH_ID);
1892 
1893 	if (ret)
1894 		goto qls_rd_flash_data_exit;
1895 
1896 	bcopy(ha->flash.mac_addr0, ha->mac_addr, ETHER_ADDR_LEN);
1897 
1898 	QL_DPRINT1((ha->pci_dev, "%s: mac %02x:%02x:%02x:%02x:%02x:%02x\n",
1899 		__func__, ha->mac_addr[0],  ha->mac_addr[1], ha->mac_addr[2],
1900 		ha->mac_addr[3], ha->mac_addr[4],  ha->mac_addr[5]));
1901 
1902 qls_rd_flash_data_exit:
1903 
1904 	qls_sem_unlock(ha, Q81_CTL_SEM_MASK_FLASH);
1905 
1906 	return(ret);
1907 }
1908 
1909 static int
1910 qls_sem_lock(qla_host_t *ha, uint32_t mask, uint32_t value)
1911 {
1912 	uint32_t count = 30;
1913 	uint32_t data;
1914 
1915 	while (count--) {
1916 		WRITE_REG32(ha, Q81_CTL_SEMAPHORE, (mask|value));
1917 
1918 		data = READ_REG32(ha, Q81_CTL_SEMAPHORE);
1919 
1920 		if (data & value) {
1921 			return (0);
1922 		} else {
1923 			QLA_USEC_DELAY(100);
1924 		}
1925 	}
1926 	ha->qla_initiate_recovery = 1;
1927 	return (-1);
1928 }
1929 
1930 static void
1931 qls_sem_unlock(qla_host_t *ha, uint32_t mask)
1932 {
1933 	WRITE_REG32(ha, Q81_CTL_SEMAPHORE, mask);
1934 }
1935 
1936 static int
1937 qls_wait_for_proc_addr_ready(qla_host_t *ha)
1938 {
1939 	uint32_t data32;
1940 	uint32_t count = 3;
1941 
1942 	while (count--) {
1943 
1944 		data32 = READ_REG32(ha, Q81_CTL_PROC_ADDR);
1945 
1946 		if (data32 & Q81_CTL_PROC_ADDR_ERR)
1947 			goto qls_wait_for_proc_addr_ready_exit;
1948 
1949 		if (data32 & Q81_CTL_PROC_ADDR_RDY)
1950 			return (0);
1951 
1952 		QLA_USEC_DELAY(100);
1953 	}
1954 
1955 qls_wait_for_proc_addr_ready_exit:
1956 	QL_DPRINT1((ha->pci_dev, "%s: failed\n", __func__));
1957 
1958 	ha->qla_initiate_recovery = 1;
1959 	return (-1);
1960 }
1961 
1962 static int
1963 qls_proc_addr_rd_reg(qla_host_t *ha, uint32_t addr_module, uint32_t reg,
1964 	uint32_t *data)
1965 {
1966 	int ret;
1967 	uint32_t value;
1968 
1969 	ret = qls_wait_for_proc_addr_ready(ha);
1970 
1971 	if (ret)
1972 		goto qls_proc_addr_rd_reg_exit;
1973 
1974 	value = addr_module | reg | Q81_CTL_PROC_ADDR_READ;
1975 
1976 	WRITE_REG32(ha, Q81_CTL_PROC_ADDR, value);
1977 
1978 	ret = qls_wait_for_proc_addr_ready(ha);
1979 
1980 	if (ret)
1981 		goto qls_proc_addr_rd_reg_exit;
1982 
1983 	*data = READ_REG32(ha, Q81_CTL_PROC_DATA);
1984 
1985 qls_proc_addr_rd_reg_exit:
1986 	return (ret);
1987 }
1988 
1989 static int
1990 qls_proc_addr_wr_reg(qla_host_t *ha, uint32_t addr_module, uint32_t reg,
1991 	uint32_t data)
1992 {
1993 	int ret;
1994 	uint32_t value;
1995 
1996 	ret = qls_wait_for_proc_addr_ready(ha);
1997 
1998 	if (ret)
1999 		goto qls_proc_addr_wr_reg_exit;
2000 
2001 	WRITE_REG32(ha, Q81_CTL_PROC_DATA, data);
2002 
2003 	value = addr_module | reg;
2004 
2005 	WRITE_REG32(ha, Q81_CTL_PROC_ADDR, value);
2006 
2007 	ret = qls_wait_for_proc_addr_ready(ha);
2008 
2009 qls_proc_addr_wr_reg_exit:
2010 	return (ret);
2011 }
2012 
2013 static int
2014 qls_hw_nic_reset(qla_host_t *ha)
2015 {
2016 	int		count;
2017 	uint32_t	data;
2018 	device_t	dev = ha->pci_dev;
2019 
2020 	ha->hw_init = 0;
2021 
2022 	data = (Q81_CTL_RESET_FUNC << Q81_CTL_RESET_MASK_SHIFT) |
2023 			Q81_CTL_RESET_FUNC;
2024 	WRITE_REG32(ha, Q81_CTL_RESET, data);
2025 
2026 	count = 10;
2027 	while (count--) {
2028 		data = READ_REG32(ha, Q81_CTL_RESET);
2029 		if ((data & Q81_CTL_RESET_FUNC) == 0)
2030 			break;
2031 		QLA_USEC_DELAY(10);
2032 	}
2033 	if (count == 0) {
2034 		device_printf(dev, "%s: Bit 15 not cleared after Reset\n",
2035 			__func__);
2036 		return (-1);
2037 	}
2038 	return (0);
2039 }
2040 
2041 static int
2042 qls_hw_reset(qla_host_t *ha)
2043 {
2044 	device_t	dev = ha->pci_dev;
2045 	int		ret;
2046 	int		count;
2047 	uint32_t	data;
2048 
2049 	QL_DPRINT2((ha->pci_dev, "%s:enter[%d]\n", __func__, ha->hw_init));
2050 
2051 	if (ha->hw_init == 0) {
2052 		ret = qls_hw_nic_reset(ha);
2053 		goto qls_hw_reset_exit;
2054 	}
2055 
2056 	ret = qls_clear_routing_table(ha);
2057 	if (ret)
2058 		goto qls_hw_reset_exit;
2059 
2060 	ret = qls_mbx_set_mgmt_ctrl(ha, Q81_MBX_SET_MGMT_CTL_STOP);
2061 	if (ret)
2062 		goto qls_hw_reset_exit;
2063 
2064 	/*
2065 	 * Wait for FIFO to empty
2066 	 */
2067 	count = 5;
2068 	while (count--) {
2069 		data = READ_REG32(ha, Q81_CTL_STATUS);
2070 		if (data & Q81_CTL_STATUS_NFE)
2071 			break;
2072 		qls_mdelay(__func__, 100);
2073 	}
2074 	if (count == 0) {
2075 		device_printf(dev, "%s: NFE bit not set\n", __func__);
2076 		goto qls_hw_reset_exit;
2077 	}
2078 
2079 	count = 5;
2080 	while (count--) {
2081 		(void)qls_mbx_get_mgmt_ctrl(ha, &data);
2082 
2083 		if ((data & Q81_MBX_GET_MGMT_CTL_FIFO_EMPTY) &&
2084 			(data & Q81_MBX_GET_MGMT_CTL_SET_MGMT))
2085 			break;
2086 		qls_mdelay(__func__, 100);
2087 	}
2088 	if (count == 0)
2089 		goto qls_hw_reset_exit;
2090 
2091 	/*
2092 	 * Reset the NIC function
2093 	 */
2094 	ret = qls_hw_nic_reset(ha);
2095 	if (ret)
2096 		goto qls_hw_reset_exit;
2097 
2098 	ret = qls_mbx_set_mgmt_ctrl(ha, Q81_MBX_SET_MGMT_CTL_RESUME);
2099 
2100 qls_hw_reset_exit:
2101 	if (ret)
2102 		device_printf(dev, "%s: failed\n", __func__);
2103 
2104 	return (ret);
2105 }
2106 
2107 /*
2108  * MPI Related Functions
2109  */
2110 int
2111 qls_mpi_risc_rd_reg(qla_host_t *ha, uint32_t reg, uint32_t *data)
2112 {
2113 	int ret;
2114 
2115 	ret = qls_proc_addr_rd_reg(ha, Q81_CTL_PROC_ADDR_MPI_RISC,
2116 			reg, data);
2117 	return (ret);
2118 }
2119 
2120 int
2121 qls_mpi_risc_wr_reg(qla_host_t *ha, uint32_t reg, uint32_t data)
2122 {
2123 	int ret;
2124 
2125 	ret = qls_proc_addr_wr_reg(ha, Q81_CTL_PROC_ADDR_MPI_RISC,
2126 			reg, data);
2127 	return (ret);
2128 }
2129 
2130 int
2131 qls_mbx_rd_reg(qla_host_t *ha, uint32_t reg, uint32_t *data)
2132 {
2133 	int ret;
2134 
2135 	if ((ha->pci_func & 0x1) == 0)
2136 		reg += Q81_FUNC0_MBX_OUT_REG0;
2137 	else
2138 		reg += Q81_FUNC1_MBX_OUT_REG0;
2139 
2140 	ret = qls_mpi_risc_rd_reg(ha, reg, data);
2141 
2142 	return (ret);
2143 }
2144 
2145 int
2146 qls_mbx_wr_reg(qla_host_t *ha, uint32_t reg, uint32_t data)
2147 {
2148 	int ret;
2149 
2150 	if ((ha->pci_func & 0x1) == 0)
2151 		reg += Q81_FUNC0_MBX_IN_REG0;
2152 	else
2153 		reg += Q81_FUNC1_MBX_IN_REG0;
2154 
2155 	ret = qls_mpi_risc_wr_reg(ha, reg, data);
2156 
2157 	return (ret);
2158 }
2159 
2160 
2161 static int
2162 qls_mbx_cmd(qla_host_t *ha, uint32_t *in_mbx, uint32_t i_count,
2163 	uint32_t *out_mbx, uint32_t o_count)
2164 {
2165 	int i, ret = -1;
2166 	uint32_t data32, mbx_cmd = 0;
2167 	uint32_t count = 50;
2168 
2169 	QL_DPRINT2((ha->pci_dev, "%s: enter[0x%08x 0x%08x 0x%08x]\n",
2170 		__func__, *in_mbx, *(in_mbx + 1), *(in_mbx + 2)));
2171 
2172 	data32 = READ_REG32(ha, Q81_CTL_HOST_CMD_STATUS);
2173 
2174 	if (data32 & Q81_CTL_HCS_HTR_INTR) {
2175 		device_printf(ha->pci_dev, "%s: cmd_status[0x%08x]\n",
2176 			__func__, data32);
2177 		goto qls_mbx_cmd_exit;
2178 	}
2179 
2180 	if (qls_sem_lock(ha, Q81_CTL_SEM_MASK_PROC_ADDR_NIC_RCV,
2181 		Q81_CTL_SEM_SET_PROC_ADDR_NIC_RCV)) {
2182 		device_printf(ha->pci_dev, "%s: semlock failed\n", __func__);
2183 		goto qls_mbx_cmd_exit;
2184 	}
2185 
2186 	ha->mbx_done = 0;
2187 
2188 	mbx_cmd = *in_mbx;
2189 
2190 	for (i = 0; i < i_count; i++) {
2191 
2192 		ret = qls_mbx_wr_reg(ha, i, *in_mbx);
2193 
2194 		if (ret) {
2195 			device_printf(ha->pci_dev,
2196 				"%s: mbx_wr[%d, 0x%08x] failed\n", __func__,
2197 				i, *in_mbx);
2198 			qls_sem_unlock(ha, Q81_CTL_SEM_MASK_PROC_ADDR_NIC_RCV);
2199 			goto qls_mbx_cmd_exit;
2200 		}
2201 
2202 		in_mbx++;
2203 	}
2204 	WRITE_REG32(ha, Q81_CTL_HOST_CMD_STATUS, Q81_CTL_HCS_CMD_SET_HTR_INTR);
2205 
2206 	qls_sem_unlock(ha, Q81_CTL_SEM_MASK_PROC_ADDR_NIC_RCV);
2207 
2208 	ret = -1;
2209 	ha->mbx_done = 0;
2210 
2211 	while (count--) {
2212 
2213 		if (ha->flags.intr_enable == 0) {
2214 			data32 = READ_REG32(ha, Q81_CTL_STATUS);
2215 
2216 			if (!(data32 & Q81_CTL_STATUS_PI)) {
2217 				qls_mdelay(__func__, 100);
2218 				continue;
2219 			}
2220 
2221 			ret = qls_mbx_rd_reg(ha, 0, &data32);
2222 
2223 			if (ret == 0 ) {
2224 				if ((data32 & 0xF000) == 0x4000) {
2225 
2226 					out_mbx[0] = data32;
2227 
2228 					for (i = 1; i < o_count; i++) {
2229 						ret = qls_mbx_rd_reg(ha, i,
2230 								&data32);
2231 						if (ret) {
2232 							device_printf(
2233 								ha->pci_dev,
2234 								"%s: mbx_rd[%d]"
2235 								" failed\n",
2236 								__func__, i);
2237 							break;
2238 						}
2239 						out_mbx[i] = data32;
2240 					}
2241 					break;
2242 				} else if ((data32 & 0xF000) == 0x8000) {
2243 					count = 50;
2244 					WRITE_REG32(ha,\
2245 						Q81_CTL_HOST_CMD_STATUS,\
2246 						Q81_CTL_HCS_CMD_CLR_RTH_INTR);
2247 				}
2248 			}
2249 		} else {
2250 			if (ha->mbx_done) {
2251 				for (i = 1; i < o_count; i++) {
2252 					out_mbx[i] = ha->mbox[i];
2253 				}
2254 				ret = 0;
2255 				break;
2256 			}
2257 		}
2258 		qls_mdelay(__func__, 1000);
2259 	}
2260 
2261 qls_mbx_cmd_exit:
2262 
2263 	if (ha->flags.intr_enable == 0) {
2264 		WRITE_REG32(ha, Q81_CTL_HOST_CMD_STATUS,\
2265 			Q81_CTL_HCS_CMD_CLR_RTH_INTR);
2266 	}
2267 
2268 	if (ret) {
2269 		ha->qla_initiate_recovery = 1;
2270 	}
2271 
2272 	QL_DPRINT2((ha->pci_dev, "%s: exit[%d]\n", __func__, ret));
2273 	return (ret);
2274 }
2275 
2276 static int
2277 qls_mbx_set_mgmt_ctrl(qla_host_t *ha, uint32_t t_ctrl)
2278 {
2279 	uint32_t *mbox;
2280 	device_t dev = ha->pci_dev;
2281 
2282 	mbox = ha->mbox;
2283 	bzero(mbox, (sizeof (uint32_t) * Q81_NUM_MBX_REGISTERS));
2284 
2285 	mbox[0] = Q81_MBX_SET_MGMT_CTL;
2286 	mbox[1] = t_ctrl;
2287 
2288 	if (qls_mbx_cmd(ha, mbox, 2, mbox, 1)) {
2289 		device_printf(dev, "%s failed\n", __func__);
2290 		return (-1);
2291 	}
2292 
2293 	if ((mbox[0] == Q81_MBX_CMD_COMPLETE) ||
2294 		((t_ctrl == Q81_MBX_SET_MGMT_CTL_STOP) &&
2295 			(mbox[0] == Q81_MBX_CMD_ERROR))){
2296 		return (0);
2297 	}
2298 	device_printf(dev, "%s failed [0x%08x]\n", __func__, mbox[0]);
2299 	return (-1);
2300 
2301 }
2302 
2303 static int
2304 qls_mbx_get_mgmt_ctrl(qla_host_t *ha, uint32_t *t_status)
2305 {
2306 	uint32_t *mbox;
2307 	device_t dev = ha->pci_dev;
2308 
2309 	*t_status = 0;
2310 
2311 	mbox = ha->mbox;
2312 	bzero(mbox, (sizeof (uint32_t) * Q81_NUM_MBX_REGISTERS));
2313 
2314 	mbox[0] = Q81_MBX_GET_MGMT_CTL;
2315 
2316 	if (qls_mbx_cmd(ha, mbox, 1, mbox, 2)) {
2317 		device_printf(dev, "%s failed\n", __func__);
2318 		return (-1);
2319 	}
2320 
2321 	*t_status = mbox[1];
2322 
2323 	return (0);
2324 }
2325 
2326 static void
2327 qls_mbx_get_link_status(qla_host_t *ha)
2328 {
2329 	uint32_t *mbox;
2330 	device_t dev = ha->pci_dev;
2331 
2332 	mbox = ha->mbox;
2333 	bzero(mbox, (sizeof (uint32_t) * Q81_NUM_MBX_REGISTERS));
2334 
2335 	mbox[0] = Q81_MBX_GET_LNK_STATUS;
2336 
2337 	if (qls_mbx_cmd(ha, mbox, 1, mbox, 6)) {
2338 		device_printf(dev, "%s failed\n", __func__);
2339 		return;
2340 	}
2341 
2342 	ha->link_status			= mbox[1];
2343 	ha->link_down_info		= mbox[2];
2344 	ha->link_hw_info		= mbox[3];
2345 	ha->link_dcbx_counters		= mbox[4];
2346 	ha->link_change_counters	= mbox[5];
2347 
2348 	device_printf(dev, "%s 0x%08x 0x%08x 0x%08x 0x%08x 0x%08x 0x%08x\n",
2349 		__func__, mbox[0],mbox[1],mbox[2],mbox[3],mbox[4],mbox[5]);
2350 
2351 	return;
2352 }
2353 
2354 static void
2355 qls_mbx_about_fw(qla_host_t *ha)
2356 {
2357 	uint32_t *mbox;
2358 	device_t dev = ha->pci_dev;
2359 
2360 	mbox = ha->mbox;
2361 	bzero(mbox, (sizeof (uint32_t) * Q81_NUM_MBX_REGISTERS));
2362 
2363 	mbox[0] = Q81_MBX_ABOUT_FW;
2364 
2365 	if (qls_mbx_cmd(ha, mbox, 1, mbox, 6)) {
2366 		device_printf(dev, "%s failed\n", __func__);
2367 		return;
2368 	}
2369 
2370 	device_printf(dev, "%s 0x%08x 0x%08x 0x%08x 0x%08x 0x%08x 0x%08x\n",
2371 		__func__, mbox[0],mbox[1],mbox[2],mbox[3],mbox[4],mbox[5]);
2372 }
2373 
2374 int
2375 qls_mbx_dump_risc_ram(qla_host_t *ha, void *buf, uint32_t r_addr,
2376 	uint32_t r_size)
2377 {
2378 	bus_addr_t b_paddr;
2379 	uint32_t *mbox;
2380 	device_t dev = ha->pci_dev;
2381 
2382 	mbox = ha->mbox;
2383 	bzero(mbox, (sizeof (uint32_t) * Q81_NUM_MBX_REGISTERS));
2384 
2385 	bzero(ha->mpi_dma.dma_b,(r_size << 2));
2386 	b_paddr = ha->mpi_dma.dma_addr;
2387 
2388 	mbox[0] = Q81_MBX_DUMP_RISC_RAM;
2389 	mbox[1] = r_addr & 0xFFFF;
2390 	mbox[2] = ((uint32_t)(b_paddr >> 16)) & 0xFFFF;
2391 	mbox[3] = ((uint32_t)b_paddr) & 0xFFFF;
2392 	mbox[4] = (r_size >> 16) & 0xFFFF;
2393 	mbox[5] = r_size & 0xFFFF;
2394 	mbox[6] = ((uint32_t)(b_paddr >> 48)) & 0xFFFF;
2395 	mbox[7] = ((uint32_t)(b_paddr >> 32)) & 0xFFFF;
2396 	mbox[8] = (r_addr >> 16) & 0xFFFF;
2397 
2398 	bus_dmamap_sync(ha->mpi_dma.dma_tag, ha->mpi_dma.dma_map,
2399 		BUS_DMASYNC_PREREAD);
2400 
2401 	if (qls_mbx_cmd(ha, mbox, 9, mbox, 1)) {
2402 		device_printf(dev, "%s failed\n", __func__);
2403 		return (-1);
2404 	}
2405         if (mbox[0] != 0x4000) {
2406                 device_printf(ha->pci_dev, "%s: failed!\n", __func__);
2407 		return (-1);
2408         } else {
2409                 bus_dmamap_sync(ha->mpi_dma.dma_tag, ha->mpi_dma.dma_map,
2410                         BUS_DMASYNC_POSTREAD);
2411                 bcopy(ha->mpi_dma.dma_b, buf, (r_size << 2));
2412         }
2413 
2414 	return (0);
2415 }
2416 
2417 int
2418 qls_mpi_reset(qla_host_t *ha)
2419 {
2420 	int		count;
2421 	uint32_t	data;
2422 	device_t	dev = ha->pci_dev;
2423 
2424 	WRITE_REG32(ha, Q81_CTL_HOST_CMD_STATUS,\
2425 		Q81_CTL_HCS_CMD_SET_RISC_RESET);
2426 
2427 	count = 10;
2428 	while (count--) {
2429 		data = READ_REG32(ha, Q81_CTL_HOST_CMD_STATUS);
2430 		if (data & Q81_CTL_HCS_RISC_RESET) {
2431 			WRITE_REG32(ha, Q81_CTL_HOST_CMD_STATUS,\
2432 				Q81_CTL_HCS_CMD_CLR_RISC_RESET);
2433 			break;
2434 		}
2435 		qls_mdelay(__func__, 10);
2436 	}
2437 	if (count == 0) {
2438 		device_printf(dev, "%s: failed\n", __func__);
2439 		return (-1);
2440 	}
2441 	return (0);
2442 }
2443 
2444