xref: /freebsd/sys/dev/mwl/if_mwl.c (revision 1d386b48)
1 /*-
2  * SPDX-License-Identifier: BSD-2-Clause
3  *
4  * Copyright (c) 2007-2009 Sam Leffler, Errno Consulting
5  * Copyright (c) 2007-2008 Marvell Semiconductor, Inc.
6  * All rights reserved.
7  *
8  * Redistribution and use in source and binary forms, with or without
9  * modification, are permitted provided that the following conditions
10  * are met:
11  * 1. Redistributions of source code must retain the above copyright
12  *    notice, this list of conditions and the following disclaimer,
13  *    without modification.
14  * 2. Redistributions in binary form must reproduce at minimum a disclaimer
15  *    similar to the "NO WARRANTY" disclaimer below ("Disclaimer") and any
16  *    redistribution must be conditioned upon including a substantially
17  *    similar Disclaimer requirement for further binary redistribution.
18  *
19  * NO WARRANTY
20  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
21  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
22  * LIMITED TO, THE IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTIBILITY
23  * AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL
24  * THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY,
25  * OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
28  * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
30  * THE POSSIBILITY OF SUCH DAMAGES.
31  */
32 
33 #include <sys/cdefs.h>
34 /*
35  * Driver for the Marvell 88W8363 Wireless LAN controller.
36  */
37 
38 #include "opt_inet.h"
39 #include "opt_mwl.h"
40 #include "opt_wlan.h"
41 
42 #include <sys/param.h>
43 #include <sys/systm.h>
44 #include <sys/sysctl.h>
45 #include <sys/mbuf.h>
46 #include <sys/malloc.h>
47 #include <sys/lock.h>
48 #include <sys/mutex.h>
49 #include <sys/kernel.h>
50 #include <sys/socket.h>
51 #include <sys/sockio.h>
52 #include <sys/errno.h>
53 #include <sys/callout.h>
54 #include <sys/bus.h>
55 #include <sys/endian.h>
56 #include <sys/kthread.h>
57 #include <sys/taskqueue.h>
58 
59 #include <machine/bus.h>
60 
61 #include <net/if.h>
62 #include <net/if_var.h>
63 #include <net/if_dl.h>
64 #include <net/if_media.h>
65 #include <net/if_types.h>
66 #include <net/if_arp.h>
67 #include <net/ethernet.h>
68 #include <net/if_llc.h>
69 
70 #include <net/bpf.h>
71 
72 #include <net80211/ieee80211_var.h>
73 #include <net80211/ieee80211_input.h>
74 #include <net80211/ieee80211_regdomain.h>
75 
76 #ifdef INET
77 #include <netinet/in.h>
78 #include <netinet/if_ether.h>
79 #endif /* INET */
80 
81 #include <dev/mwl/if_mwlvar.h>
82 #include <dev/mwl/mwldiag.h>
83 
84 static struct ieee80211vap *mwl_vap_create(struct ieee80211com *,
85 		    const char [IFNAMSIZ], int, enum ieee80211_opmode, int,
86 		    const uint8_t [IEEE80211_ADDR_LEN],
87 		    const uint8_t [IEEE80211_ADDR_LEN]);
88 static void	mwl_vap_delete(struct ieee80211vap *);
89 static int	mwl_setupdma(struct mwl_softc *);
90 static int	mwl_hal_reset(struct mwl_softc *sc);
91 static int	mwl_init(struct mwl_softc *);
92 static void	mwl_parent(struct ieee80211com *);
93 static int	mwl_reset(struct ieee80211vap *, u_long);
94 static void	mwl_stop(struct mwl_softc *);
95 static void	mwl_start(struct mwl_softc *);
96 static int	mwl_transmit(struct ieee80211com *, struct mbuf *);
97 static int	mwl_raw_xmit(struct ieee80211_node *, struct mbuf *,
98 			const struct ieee80211_bpf_params *);
99 static int	mwl_media_change(if_t);
100 static void	mwl_watchdog(void *);
101 static int	mwl_ioctl(struct ieee80211com *, u_long, void *);
102 static void	mwl_radar_proc(void *, int);
103 static void	mwl_chanswitch_proc(void *, int);
104 static void	mwl_bawatchdog_proc(void *, int);
105 static int	mwl_key_alloc(struct ieee80211vap *,
106 			struct ieee80211_key *,
107 			ieee80211_keyix *, ieee80211_keyix *);
108 static int	mwl_key_delete(struct ieee80211vap *,
109 			const struct ieee80211_key *);
110 static int	mwl_key_set(struct ieee80211vap *,
111 			const struct ieee80211_key *);
112 static int	_mwl_key_set(struct ieee80211vap *,
113 			const struct ieee80211_key *,
114 			const uint8_t mac[IEEE80211_ADDR_LEN]);
115 static int	mwl_mode_init(struct mwl_softc *);
116 static void	mwl_update_mcast(struct ieee80211com *);
117 static void	mwl_update_promisc(struct ieee80211com *);
118 static void	mwl_updateslot(struct ieee80211com *);
119 static int	mwl_beacon_setup(struct ieee80211vap *);
120 static void	mwl_beacon_update(struct ieee80211vap *, int);
121 #ifdef MWL_HOST_PS_SUPPORT
122 static void	mwl_update_ps(struct ieee80211vap *, int);
123 static int	mwl_set_tim(struct ieee80211_node *, int);
124 #endif
125 static int	mwl_dma_setup(struct mwl_softc *);
126 static void	mwl_dma_cleanup(struct mwl_softc *);
127 static struct ieee80211_node *mwl_node_alloc(struct ieee80211vap *,
128 		    const uint8_t [IEEE80211_ADDR_LEN]);
129 static void	mwl_node_cleanup(struct ieee80211_node *);
130 static void	mwl_node_drain(struct ieee80211_node *);
131 static void	mwl_node_getsignal(const struct ieee80211_node *,
132 			int8_t *, int8_t *);
133 static void	mwl_node_getmimoinfo(const struct ieee80211_node *,
134 			struct ieee80211_mimo_info *);
135 static int	mwl_rxbuf_init(struct mwl_softc *, struct mwl_rxbuf *);
136 static void	mwl_rx_proc(void *, int);
137 static void	mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *, int);
138 static int	mwl_tx_setup(struct mwl_softc *, int, int);
139 static int	mwl_wme_update(struct ieee80211com *);
140 static void	mwl_tx_cleanupq(struct mwl_softc *, struct mwl_txq *);
141 static void	mwl_tx_cleanup(struct mwl_softc *);
142 static uint16_t	mwl_calcformat(uint8_t rate, const struct ieee80211_node *);
143 static int	mwl_tx_start(struct mwl_softc *, struct ieee80211_node *,
144 			     struct mwl_txbuf *, struct mbuf *);
145 static void	mwl_tx_proc(void *, int);
146 static int	mwl_chan_set(struct mwl_softc *, struct ieee80211_channel *);
147 static void	mwl_draintxq(struct mwl_softc *);
148 static void	mwl_cleartxq(struct mwl_softc *, struct ieee80211vap *);
149 static int	mwl_recv_action(struct ieee80211_node *,
150 			const struct ieee80211_frame *,
151 			const uint8_t *, const uint8_t *);
152 static int	mwl_addba_request(struct ieee80211_node *,
153 			struct ieee80211_tx_ampdu *, int dialogtoken,
154 			int baparamset, int batimeout);
155 static int	mwl_addba_response(struct ieee80211_node *,
156 			struct ieee80211_tx_ampdu *, int status,
157 			int baparamset, int batimeout);
158 static void	mwl_addba_stop(struct ieee80211_node *,
159 			struct ieee80211_tx_ampdu *);
160 static int	mwl_startrecv(struct mwl_softc *);
161 static MWL_HAL_APMODE mwl_getapmode(const struct ieee80211vap *,
162 			struct ieee80211_channel *);
163 static int	mwl_setapmode(struct ieee80211vap *, struct ieee80211_channel*);
164 static void	mwl_scan_start(struct ieee80211com *);
165 static void	mwl_scan_end(struct ieee80211com *);
166 static void	mwl_set_channel(struct ieee80211com *);
167 static int	mwl_peerstadb(struct ieee80211_node *,
168 			int aid, int staid, MWL_HAL_PEERINFO *pi);
169 static int	mwl_localstadb(struct ieee80211vap *);
170 static int	mwl_newstate(struct ieee80211vap *, enum ieee80211_state, int);
171 static int	allocstaid(struct mwl_softc *sc, int aid);
172 static void	delstaid(struct mwl_softc *sc, int staid);
173 static void	mwl_newassoc(struct ieee80211_node *, int);
174 static void	mwl_agestations(void *);
175 static int	mwl_setregdomain(struct ieee80211com *,
176 			struct ieee80211_regdomain *, int,
177 			struct ieee80211_channel []);
178 static void	mwl_getradiocaps(struct ieee80211com *, int, int *,
179 			struct ieee80211_channel []);
180 static int	mwl_getchannels(struct mwl_softc *);
181 
182 static void	mwl_sysctlattach(struct mwl_softc *);
183 static void	mwl_announce(struct mwl_softc *);
184 
185 SYSCTL_NODE(_hw, OID_AUTO, mwl, CTLFLAG_RD | CTLFLAG_MPSAFE, 0,
186     "Marvell driver parameters");
187 
188 static	int mwl_rxdesc = MWL_RXDESC;		/* # rx desc's to allocate */
189 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdesc, CTLFLAG_RW, &mwl_rxdesc,
190 	    0, "rx descriptors allocated");
191 static	int mwl_rxbuf = MWL_RXBUF;		/* # rx buffers to allocate */
192 SYSCTL_INT(_hw_mwl, OID_AUTO, rxbuf, CTLFLAG_RWTUN, &mwl_rxbuf,
193 	    0, "rx buffers allocated");
194 static	int mwl_txbuf = MWL_TXBUF;		/* # tx buffers to allocate */
195 SYSCTL_INT(_hw_mwl, OID_AUTO, txbuf, CTLFLAG_RWTUN, &mwl_txbuf,
196 	    0, "tx buffers allocated");
197 static	int mwl_txcoalesce = 8;		/* # tx packets to q before poking f/w*/
198 SYSCTL_INT(_hw_mwl, OID_AUTO, txcoalesce, CTLFLAG_RWTUN, &mwl_txcoalesce,
199 	    0, "tx buffers to send at once");
200 static	int mwl_rxquota = MWL_RXBUF;		/* # max buffers to process */
201 SYSCTL_INT(_hw_mwl, OID_AUTO, rxquota, CTLFLAG_RWTUN, &mwl_rxquota,
202 	    0, "max rx buffers to process per interrupt");
203 static	int mwl_rxdmalow = 3;			/* # min buffers for wakeup */
204 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdmalow, CTLFLAG_RWTUN, &mwl_rxdmalow,
205 	    0, "min free rx buffers before restarting traffic");
206 
207 #ifdef MWL_DEBUG
208 static	int mwl_debug = 0;
209 SYSCTL_INT(_hw_mwl, OID_AUTO, debug, CTLFLAG_RWTUN, &mwl_debug,
210 	    0, "control debugging printfs");
211 enum {
212 	MWL_DEBUG_XMIT		= 0x00000001,	/* basic xmit operation */
213 	MWL_DEBUG_XMIT_DESC	= 0x00000002,	/* xmit descriptors */
214 	MWL_DEBUG_RECV		= 0x00000004,	/* basic recv operation */
215 	MWL_DEBUG_RECV_DESC	= 0x00000008,	/* recv descriptors */
216 	MWL_DEBUG_RESET		= 0x00000010,	/* reset processing */
217 	MWL_DEBUG_BEACON 	= 0x00000020,	/* beacon handling */
218 	MWL_DEBUG_INTR		= 0x00000040,	/* ISR */
219 	MWL_DEBUG_TX_PROC	= 0x00000080,	/* tx ISR proc */
220 	MWL_DEBUG_RX_PROC	= 0x00000100,	/* rx ISR proc */
221 	MWL_DEBUG_KEYCACHE	= 0x00000200,	/* key cache management */
222 	MWL_DEBUG_STATE		= 0x00000400,	/* 802.11 state transitions */
223 	MWL_DEBUG_NODE		= 0x00000800,	/* node management */
224 	MWL_DEBUG_RECV_ALL	= 0x00001000,	/* trace all frames (beacons) */
225 	MWL_DEBUG_TSO		= 0x00002000,	/* TSO processing */
226 	MWL_DEBUG_AMPDU		= 0x00004000,	/* BA stream handling */
227 	MWL_DEBUG_ANY		= 0xffffffff
228 };
229 #define	IS_BEACON(wh) \
230     ((wh->i_fc[0] & (IEEE80211_FC0_TYPE_MASK|IEEE80211_FC0_SUBTYPE_MASK)) == \
231 	 (IEEE80211_FC0_TYPE_MGT|IEEE80211_FC0_SUBTYPE_BEACON))
232 #define	IFF_DUMPPKTS_RECV(sc, wh) \
233     ((sc->sc_debug & MWL_DEBUG_RECV) && \
234       ((sc->sc_debug & MWL_DEBUG_RECV_ALL) || !IS_BEACON(wh)))
235 #define	IFF_DUMPPKTS_XMIT(sc) \
236 	(sc->sc_debug & MWL_DEBUG_XMIT)
237 
238 #define	DPRINTF(sc, m, fmt, ...) do {				\
239 	if (sc->sc_debug & (m))					\
240 		printf(fmt, __VA_ARGS__);			\
241 } while (0)
242 #define	KEYPRINTF(sc, hk, mac) do {				\
243 	if (sc->sc_debug & MWL_DEBUG_KEYCACHE)			\
244 		mwl_keyprint(sc, __func__, hk, mac);		\
245 } while (0)
246 static	void mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix);
247 static	void mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix);
248 #else
249 #define	IFF_DUMPPKTS_RECV(sc, wh)	0
250 #define	IFF_DUMPPKTS_XMIT(sc)		0
251 #define	DPRINTF(sc, m, fmt, ...)	do { (void )sc; } while (0)
252 #define	KEYPRINTF(sc, k, mac)		do { (void )sc; } while (0)
253 #endif
254 
255 static MALLOC_DEFINE(M_MWLDEV, "mwldev", "mwl driver dma buffers");
256 
257 /*
258  * Each packet has fixed front matter: a 2-byte length
259  * of the payload, followed by a 4-address 802.11 header
260  * (regardless of the actual header and always w/o any
261  * QoS header).  The payload then follows.
262  */
263 struct mwltxrec {
264 	uint16_t fwlen;
265 	struct ieee80211_frame_addr4 wh;
266 } __packed;
267 
268 /*
269  * Read/Write shorthands for accesses to BAR 0.  Note
270  * that all BAR 1 operations are done in the "hal" and
271  * there should be no reference to them here.
272  */
273 #ifdef MWL_DEBUG
274 static __inline uint32_t
275 RD4(struct mwl_softc *sc, bus_size_t off)
276 {
277 	return bus_space_read_4(sc->sc_io0t, sc->sc_io0h, off);
278 }
279 #endif
280 
281 static __inline void
282 WR4(struct mwl_softc *sc, bus_size_t off, uint32_t val)
283 {
284 	bus_space_write_4(sc->sc_io0t, sc->sc_io0h, off, val);
285 }
286 
287 int
288 mwl_attach(uint16_t devid, struct mwl_softc *sc)
289 {
290 	struct ieee80211com *ic = &sc->sc_ic;
291 	struct mwl_hal *mh;
292 	int error = 0;
293 
294 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: devid 0x%x\n", __func__, devid);
295 
296 	/*
297 	 * Setup the RX free list lock early, so it can be consistently
298 	 * removed.
299 	 */
300 	MWL_RXFREE_INIT(sc);
301 
302 	mh = mwl_hal_attach(sc->sc_dev, devid,
303 	    sc->sc_io1h, sc->sc_io1t, sc->sc_dmat);
304 	if (mh == NULL) {
305 		device_printf(sc->sc_dev, "unable to attach HAL\n");
306 		error = EIO;
307 		goto bad;
308 	}
309 	sc->sc_mh = mh;
310 	/*
311 	 * Load firmware so we can get setup.  We arbitrarily
312 	 * pick station firmware; we'll re-load firmware as
313 	 * needed so setting up the wrong mode isn't a big deal.
314 	 */
315 	if (mwl_hal_fwload(mh, NULL) != 0) {
316 		device_printf(sc->sc_dev, "unable to setup builtin firmware\n");
317 		error = EIO;
318 		goto bad1;
319 	}
320 	if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
321 		device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
322 		error = EIO;
323 		goto bad1;
324 	}
325 	error = mwl_getchannels(sc);
326 	if (error != 0)
327 		goto bad1;
328 
329 	sc->sc_txantenna = 0;		/* h/w default */
330 	sc->sc_rxantenna = 0;		/* h/w default */
331 	sc->sc_invalid = 0;		/* ready to go, enable int handling */
332 	sc->sc_ageinterval = MWL_AGEINTERVAL;
333 
334 	/*
335 	 * Allocate tx+rx descriptors and populate the lists.
336 	 * We immediately push the information to the firmware
337 	 * as otherwise it gets upset.
338 	 */
339 	error = mwl_dma_setup(sc);
340 	if (error != 0) {
341 		device_printf(sc->sc_dev, "failed to setup descriptors: %d\n",
342 		    error);
343 		goto bad1;
344 	}
345 	error = mwl_setupdma(sc);	/* push to firmware */
346 	if (error != 0)			/* NB: mwl_setupdma prints msg */
347 		goto bad1;
348 
349 	callout_init(&sc->sc_timer, 1);
350 	callout_init_mtx(&sc->sc_watchdog, &sc->sc_mtx, 0);
351 	mbufq_init(&sc->sc_snd, ifqmaxlen);
352 
353 	sc->sc_tq = taskqueue_create("mwl_taskq", M_NOWAIT,
354 		taskqueue_thread_enqueue, &sc->sc_tq);
355 	taskqueue_start_threads(&sc->sc_tq, 1, PI_NET,
356 		"%s taskq", device_get_nameunit(sc->sc_dev));
357 
358 	NET_TASK_INIT(&sc->sc_rxtask, 0, mwl_rx_proc, sc);
359 	TASK_INIT(&sc->sc_radartask, 0, mwl_radar_proc, sc);
360 	TASK_INIT(&sc->sc_chanswitchtask, 0, mwl_chanswitch_proc, sc);
361 	TASK_INIT(&sc->sc_bawatchdogtask, 0, mwl_bawatchdog_proc, sc);
362 
363 	/* NB: insure BK queue is the lowest priority h/w queue */
364 	if (!mwl_tx_setup(sc, WME_AC_BK, MWL_WME_AC_BK)) {
365 		device_printf(sc->sc_dev,
366 		    "unable to setup xmit queue for %s traffic!\n",
367 		     ieee80211_wme_acnames[WME_AC_BK]);
368 		error = EIO;
369 		goto bad2;
370 	}
371 	if (!mwl_tx_setup(sc, WME_AC_BE, MWL_WME_AC_BE) ||
372 	    !mwl_tx_setup(sc, WME_AC_VI, MWL_WME_AC_VI) ||
373 	    !mwl_tx_setup(sc, WME_AC_VO, MWL_WME_AC_VO)) {
374 		/*
375 		 * Not enough hardware tx queues to properly do WME;
376 		 * just punt and assign them all to the same h/w queue.
377 		 * We could do a better job of this if, for example,
378 		 * we allocate queues when we switch from station to
379 		 * AP mode.
380 		 */
381 		if (sc->sc_ac2q[WME_AC_VI] != NULL)
382 			mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_VI]);
383 		if (sc->sc_ac2q[WME_AC_BE] != NULL)
384 			mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_BE]);
385 		sc->sc_ac2q[WME_AC_BE] = sc->sc_ac2q[WME_AC_BK];
386 		sc->sc_ac2q[WME_AC_VI] = sc->sc_ac2q[WME_AC_BK];
387 		sc->sc_ac2q[WME_AC_VO] = sc->sc_ac2q[WME_AC_BK];
388 	}
389 	TASK_INIT(&sc->sc_txtask, 0, mwl_tx_proc, sc);
390 
391 	ic->ic_softc = sc;
392 	ic->ic_name = device_get_nameunit(sc->sc_dev);
393 	/* XXX not right but it's not used anywhere important */
394 	ic->ic_phytype = IEEE80211_T_OFDM;
395 	ic->ic_opmode = IEEE80211_M_STA;
396 	ic->ic_caps =
397 		  IEEE80211_C_STA		/* station mode supported */
398 		| IEEE80211_C_HOSTAP		/* hostap mode */
399 		| IEEE80211_C_MONITOR		/* monitor mode */
400 #if 0
401 		| IEEE80211_C_IBSS		/* ibss, nee adhoc, mode */
402 		| IEEE80211_C_AHDEMO		/* adhoc demo mode */
403 #endif
404 		| IEEE80211_C_MBSS		/* mesh point link mode */
405 		| IEEE80211_C_WDS		/* WDS supported */
406 		| IEEE80211_C_SHPREAMBLE	/* short preamble supported */
407 		| IEEE80211_C_SHSLOT		/* short slot time supported */
408 		| IEEE80211_C_WME		/* WME/WMM supported */
409 		| IEEE80211_C_BURST		/* xmit bursting supported */
410 		| IEEE80211_C_WPA		/* capable of WPA1+WPA2 */
411 		| IEEE80211_C_BGSCAN		/* capable of bg scanning */
412 		| IEEE80211_C_TXFRAG		/* handle tx frags */
413 		| IEEE80211_C_TXPMGT		/* capable of txpow mgt */
414 		| IEEE80211_C_DFS		/* DFS supported */
415 		;
416 
417 	ic->ic_htcaps =
418 		  IEEE80211_HTCAP_SMPS_ENA	/* SM PS mode enabled */
419 		| IEEE80211_HTCAP_CHWIDTH40	/* 40MHz channel width */
420 		| IEEE80211_HTCAP_SHORTGI20	/* short GI in 20MHz */
421 		| IEEE80211_HTCAP_SHORTGI40	/* short GI in 40MHz */
422 		| IEEE80211_HTCAP_RXSTBC_2STREAM/* 1-2 spatial streams */
423 #if MWL_AGGR_SIZE == 7935
424 		| IEEE80211_HTCAP_MAXAMSDU_7935	/* max A-MSDU length */
425 #else
426 		| IEEE80211_HTCAP_MAXAMSDU_3839	/* max A-MSDU length */
427 #endif
428 #if 0
429 		| IEEE80211_HTCAP_PSMP		/* PSMP supported */
430 		| IEEE80211_HTCAP_40INTOLERANT	/* 40MHz intolerant */
431 #endif
432 		/* s/w capabilities */
433 		| IEEE80211_HTC_HT		/* HT operation */
434 		| IEEE80211_HTC_AMPDU		/* tx A-MPDU */
435 		| IEEE80211_HTC_AMSDU		/* tx A-MSDU */
436 		| IEEE80211_HTC_SMPS		/* SMPS available */
437 		;
438 
439 	/*
440 	 * Mark h/w crypto support.
441 	 * XXX no way to query h/w support.
442 	 */
443 	ic->ic_cryptocaps |= IEEE80211_CRYPTO_WEP
444 			  |  IEEE80211_CRYPTO_AES_CCM
445 			  |  IEEE80211_CRYPTO_TKIP
446 			  |  IEEE80211_CRYPTO_TKIPMIC
447 			  ;
448 	/*
449 	 * Transmit requires space in the packet for a special
450 	 * format transmit record and optional padding between
451 	 * this record and the payload.  Ask the net80211 layer
452 	 * to arrange this when encapsulating packets so we can
453 	 * add it efficiently.
454 	 */
455 	ic->ic_headroom = sizeof(struct mwltxrec) -
456 		sizeof(struct ieee80211_frame);
457 
458 	IEEE80211_ADDR_COPY(ic->ic_macaddr, sc->sc_hwspecs.macAddr);
459 
460 	/* call MI attach routine. */
461 	ieee80211_ifattach(ic);
462 	ic->ic_setregdomain = mwl_setregdomain;
463 	ic->ic_getradiocaps = mwl_getradiocaps;
464 	/* override default methods */
465 	ic->ic_raw_xmit = mwl_raw_xmit;
466 	ic->ic_newassoc = mwl_newassoc;
467 	ic->ic_updateslot = mwl_updateslot;
468 	ic->ic_update_mcast = mwl_update_mcast;
469 	ic->ic_update_promisc = mwl_update_promisc;
470 	ic->ic_wme.wme_update = mwl_wme_update;
471 	ic->ic_transmit = mwl_transmit;
472 	ic->ic_ioctl = mwl_ioctl;
473 	ic->ic_parent = mwl_parent;
474 
475 	ic->ic_node_alloc = mwl_node_alloc;
476 	sc->sc_node_cleanup = ic->ic_node_cleanup;
477 	ic->ic_node_cleanup = mwl_node_cleanup;
478 	sc->sc_node_drain = ic->ic_node_drain;
479 	ic->ic_node_drain = mwl_node_drain;
480 	ic->ic_node_getsignal = mwl_node_getsignal;
481 	ic->ic_node_getmimoinfo = mwl_node_getmimoinfo;
482 
483 	ic->ic_scan_start = mwl_scan_start;
484 	ic->ic_scan_end = mwl_scan_end;
485 	ic->ic_set_channel = mwl_set_channel;
486 
487 	sc->sc_recv_action = ic->ic_recv_action;
488 	ic->ic_recv_action = mwl_recv_action;
489 	sc->sc_addba_request = ic->ic_addba_request;
490 	ic->ic_addba_request = mwl_addba_request;
491 	sc->sc_addba_response = ic->ic_addba_response;
492 	ic->ic_addba_response = mwl_addba_response;
493 	sc->sc_addba_stop = ic->ic_addba_stop;
494 	ic->ic_addba_stop = mwl_addba_stop;
495 
496 	ic->ic_vap_create = mwl_vap_create;
497 	ic->ic_vap_delete = mwl_vap_delete;
498 
499 	ieee80211_radiotap_attach(ic,
500 	    &sc->sc_tx_th.wt_ihdr, sizeof(sc->sc_tx_th),
501 		MWL_TX_RADIOTAP_PRESENT,
502 	    &sc->sc_rx_th.wr_ihdr, sizeof(sc->sc_rx_th),
503 		MWL_RX_RADIOTAP_PRESENT);
504 	/*
505 	 * Setup dynamic sysctl's now that country code and
506 	 * regdomain are available from the hal.
507 	 */
508 	mwl_sysctlattach(sc);
509 
510 	if (bootverbose)
511 		ieee80211_announce(ic);
512 	mwl_announce(sc);
513 	return 0;
514 bad2:
515 	mwl_dma_cleanup(sc);
516 bad1:
517 	mwl_hal_detach(mh);
518 bad:
519 	MWL_RXFREE_DESTROY(sc);
520 	sc->sc_invalid = 1;
521 	return error;
522 }
523 
524 int
525 mwl_detach(struct mwl_softc *sc)
526 {
527 	struct ieee80211com *ic = &sc->sc_ic;
528 
529 	MWL_LOCK(sc);
530 	mwl_stop(sc);
531 	MWL_UNLOCK(sc);
532 	/*
533 	 * NB: the order of these is important:
534 	 * o call the 802.11 layer before detaching the hal to
535 	 *   insure callbacks into the driver to delete global
536 	 *   key cache entries can be handled
537 	 * o reclaim the tx queue data structures after calling
538 	 *   the 802.11 layer as we'll get called back to reclaim
539 	 *   node state and potentially want to use them
540 	 * o to cleanup the tx queues the hal is called, so detach
541 	 *   it last
542 	 * Other than that, it's straightforward...
543 	 */
544 	ieee80211_ifdetach(ic);
545 	callout_drain(&sc->sc_watchdog);
546 	mwl_dma_cleanup(sc);
547 	MWL_RXFREE_DESTROY(sc);
548 	mwl_tx_cleanup(sc);
549 	mwl_hal_detach(sc->sc_mh);
550 	mbufq_drain(&sc->sc_snd);
551 
552 	return 0;
553 }
554 
555 /*
556  * MAC address handling for multiple BSS on the same radio.
557  * The first vap uses the MAC address from the EEPROM.  For
558  * subsequent vap's we set the U/L bit (bit 1) in the MAC
559  * address and use the next six bits as an index.
560  */
561 static void
562 assign_address(struct mwl_softc *sc, uint8_t mac[IEEE80211_ADDR_LEN], int clone)
563 {
564 	int i;
565 
566 	if (clone && mwl_hal_ismbsscapable(sc->sc_mh)) {
567 		/* NB: we only do this if h/w supports multiple bssid */
568 		for (i = 0; i < 32; i++)
569 			if ((sc->sc_bssidmask & (1<<i)) == 0)
570 				break;
571 		if (i != 0)
572 			mac[0] |= (i << 2)|0x2;
573 	} else
574 		i = 0;
575 	sc->sc_bssidmask |= 1<<i;
576 	if (i == 0)
577 		sc->sc_nbssid0++;
578 }
579 
580 static void
581 reclaim_address(struct mwl_softc *sc, const uint8_t mac[IEEE80211_ADDR_LEN])
582 {
583 	int i = mac[0] >> 2;
584 	if (i != 0 || --sc->sc_nbssid0 == 0)
585 		sc->sc_bssidmask &= ~(1<<i);
586 }
587 
588 static struct ieee80211vap *
589 mwl_vap_create(struct ieee80211com *ic, const char name[IFNAMSIZ], int unit,
590     enum ieee80211_opmode opmode, int flags,
591     const uint8_t bssid[IEEE80211_ADDR_LEN],
592     const uint8_t mac0[IEEE80211_ADDR_LEN])
593 {
594 	struct mwl_softc *sc = ic->ic_softc;
595 	struct mwl_hal *mh = sc->sc_mh;
596 	struct ieee80211vap *vap, *apvap;
597 	struct mwl_hal_vap *hvap;
598 	struct mwl_vap *mvp;
599 	uint8_t mac[IEEE80211_ADDR_LEN];
600 
601 	IEEE80211_ADDR_COPY(mac, mac0);
602 	switch (opmode) {
603 	case IEEE80211_M_HOSTAP:
604 	case IEEE80211_M_MBSS:
605 		if ((flags & IEEE80211_CLONE_MACADDR) == 0)
606 			assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
607 		hvap = mwl_hal_newvap(mh, MWL_HAL_AP, mac);
608 		if (hvap == NULL) {
609 			if ((flags & IEEE80211_CLONE_MACADDR) == 0)
610 				reclaim_address(sc, mac);
611 			return NULL;
612 		}
613 		break;
614 	case IEEE80211_M_STA:
615 		if ((flags & IEEE80211_CLONE_MACADDR) == 0)
616 			assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
617 		hvap = mwl_hal_newvap(mh, MWL_HAL_STA, mac);
618 		if (hvap == NULL) {
619 			if ((flags & IEEE80211_CLONE_MACADDR) == 0)
620 				reclaim_address(sc, mac);
621 			return NULL;
622 		}
623 		/* no h/w beacon miss support; always use s/w */
624 		flags |= IEEE80211_CLONE_NOBEACONS;
625 		break;
626 	case IEEE80211_M_WDS:
627 		hvap = NULL;		/* NB: we use associated AP vap */
628 		if (sc->sc_napvaps == 0)
629 			return NULL;	/* no existing AP vap */
630 		break;
631 	case IEEE80211_M_MONITOR:
632 		hvap = NULL;
633 		break;
634 	case IEEE80211_M_IBSS:
635 	case IEEE80211_M_AHDEMO:
636 	default:
637 		return NULL;
638 	}
639 
640 	mvp = malloc(sizeof(struct mwl_vap), M_80211_VAP, M_WAITOK | M_ZERO);
641 	mvp->mv_hvap = hvap;
642 	if (opmode == IEEE80211_M_WDS) {
643 		/*
644 		 * WDS vaps must have an associated AP vap; find one.
645 		 * XXX not right.
646 		 */
647 		TAILQ_FOREACH(apvap, &ic->ic_vaps, iv_next)
648 			if (apvap->iv_opmode == IEEE80211_M_HOSTAP) {
649 				mvp->mv_ap_hvap = MWL_VAP(apvap)->mv_hvap;
650 				break;
651 			}
652 		KASSERT(mvp->mv_ap_hvap != NULL, ("no ap vap"));
653 	}
654 	vap = &mvp->mv_vap;
655 	ieee80211_vap_setup(ic, vap, name, unit, opmode, flags, bssid);
656 	/* override with driver methods */
657 	mvp->mv_newstate = vap->iv_newstate;
658 	vap->iv_newstate = mwl_newstate;
659 	vap->iv_max_keyix = 0;	/* XXX */
660 	vap->iv_key_alloc = mwl_key_alloc;
661 	vap->iv_key_delete = mwl_key_delete;
662 	vap->iv_key_set = mwl_key_set;
663 #ifdef MWL_HOST_PS_SUPPORT
664 	if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS) {
665 		vap->iv_update_ps = mwl_update_ps;
666 		mvp->mv_set_tim = vap->iv_set_tim;
667 		vap->iv_set_tim = mwl_set_tim;
668 	}
669 #endif
670 	vap->iv_reset = mwl_reset;
671 	vap->iv_update_beacon = mwl_beacon_update;
672 
673 	/* override max aid so sta's cannot assoc when we're out of sta id's */
674 	vap->iv_max_aid = MWL_MAXSTAID;
675 	/* override default A-MPDU rx parameters */
676 	vap->iv_ampdu_rxmax = IEEE80211_HTCAP_MAXRXAMPDU_64K;
677 	vap->iv_ampdu_density = IEEE80211_HTCAP_MPDUDENSITY_4;
678 
679 	/* complete setup */
680 	ieee80211_vap_attach(vap, mwl_media_change, ieee80211_media_status,
681 	    mac);
682 
683 	switch (vap->iv_opmode) {
684 	case IEEE80211_M_HOSTAP:
685 	case IEEE80211_M_MBSS:
686 	case IEEE80211_M_STA:
687 		/*
688 		 * Setup sta db entry for local address.
689 		 */
690 		mwl_localstadb(vap);
691 		if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
692 		    vap->iv_opmode == IEEE80211_M_MBSS)
693 			sc->sc_napvaps++;
694 		else
695 			sc->sc_nstavaps++;
696 		break;
697 	case IEEE80211_M_WDS:
698 		sc->sc_nwdsvaps++;
699 		break;
700 	default:
701 		break;
702 	}
703 	/*
704 	 * Setup overall operating mode.
705 	 */
706 	if (sc->sc_napvaps)
707 		ic->ic_opmode = IEEE80211_M_HOSTAP;
708 	else if (sc->sc_nstavaps)
709 		ic->ic_opmode = IEEE80211_M_STA;
710 	else
711 		ic->ic_opmode = opmode;
712 
713 	return vap;
714 }
715 
716 static void
717 mwl_vap_delete(struct ieee80211vap *vap)
718 {
719 	struct mwl_vap *mvp = MWL_VAP(vap);
720 	struct mwl_softc *sc = vap->iv_ic->ic_softc;
721 	struct mwl_hal *mh = sc->sc_mh;
722 	struct mwl_hal_vap *hvap = mvp->mv_hvap;
723 	enum ieee80211_opmode opmode = vap->iv_opmode;
724 
725 	/* XXX disallow ap vap delete if WDS still present */
726 	if (sc->sc_running) {
727 		/* quiesce h/w while we remove the vap */
728 		mwl_hal_intrset(mh, 0);		/* disable interrupts */
729 	}
730 	ieee80211_vap_detach(vap);
731 	switch (opmode) {
732 	case IEEE80211_M_HOSTAP:
733 	case IEEE80211_M_MBSS:
734 	case IEEE80211_M_STA:
735 		KASSERT(hvap != NULL, ("no hal vap handle"));
736 		(void) mwl_hal_delstation(hvap, vap->iv_myaddr);
737 		mwl_hal_delvap(hvap);
738 		if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS)
739 			sc->sc_napvaps--;
740 		else
741 			sc->sc_nstavaps--;
742 		/* XXX don't do it for IEEE80211_CLONE_MACADDR */
743 		reclaim_address(sc, vap->iv_myaddr);
744 		break;
745 	case IEEE80211_M_WDS:
746 		sc->sc_nwdsvaps--;
747 		break;
748 	default:
749 		break;
750 	}
751 	mwl_cleartxq(sc, vap);
752 	free(mvp, M_80211_VAP);
753 	if (sc->sc_running)
754 		mwl_hal_intrset(mh, sc->sc_imask);
755 }
756 
757 void
758 mwl_suspend(struct mwl_softc *sc)
759 {
760 
761 	MWL_LOCK(sc);
762 	mwl_stop(sc);
763 	MWL_UNLOCK(sc);
764 }
765 
766 void
767 mwl_resume(struct mwl_softc *sc)
768 {
769 	int error = EDOOFUS;
770 
771 	MWL_LOCK(sc);
772 	if (sc->sc_ic.ic_nrunning > 0)
773 		error = mwl_init(sc);
774 	MWL_UNLOCK(sc);
775 
776 	if (error == 0)
777 		ieee80211_start_all(&sc->sc_ic);	/* start all vap's */
778 }
779 
780 void
781 mwl_shutdown(void *arg)
782 {
783 	struct mwl_softc *sc = arg;
784 
785 	MWL_LOCK(sc);
786 	mwl_stop(sc);
787 	MWL_UNLOCK(sc);
788 }
789 
790 /*
791  * Interrupt handler.  Most of the actual processing is deferred.
792  */
793 void
794 mwl_intr(void *arg)
795 {
796 	struct mwl_softc *sc = arg;
797 	struct mwl_hal *mh = sc->sc_mh;
798 	uint32_t status;
799 
800 	if (sc->sc_invalid) {
801 		/*
802 		 * The hardware is not ready/present, don't touch anything.
803 		 * Note this can happen early on if the IRQ is shared.
804 		 */
805 		DPRINTF(sc, MWL_DEBUG_ANY, "%s: invalid; ignored\n", __func__);
806 		return;
807 	}
808 	/*
809 	 * Figure out the reason(s) for the interrupt.
810 	 */
811 	mwl_hal_getisr(mh, &status);		/* NB: clears ISR too */
812 	if (status == 0)			/* must be a shared irq */
813 		return;
814 
815 	DPRINTF(sc, MWL_DEBUG_INTR, "%s: status 0x%x imask 0x%x\n",
816 	    __func__, status, sc->sc_imask);
817 	if (status & MACREG_A2HRIC_BIT_RX_RDY)
818 		taskqueue_enqueue(sc->sc_tq, &sc->sc_rxtask);
819 	if (status & MACREG_A2HRIC_BIT_TX_DONE)
820 		taskqueue_enqueue(sc->sc_tq, &sc->sc_txtask);
821 	if (status & MACREG_A2HRIC_BIT_BA_WATCHDOG)
822 		taskqueue_enqueue(sc->sc_tq, &sc->sc_bawatchdogtask);
823 	if (status & MACREG_A2HRIC_BIT_OPC_DONE)
824 		mwl_hal_cmddone(mh);
825 	if (status & MACREG_A2HRIC_BIT_MAC_EVENT) {
826 		;
827 	}
828 	if (status & MACREG_A2HRIC_BIT_ICV_ERROR) {
829 		/* TKIP ICV error */
830 		sc->sc_stats.mst_rx_badtkipicv++;
831 	}
832 	if (status & MACREG_A2HRIC_BIT_QUEUE_EMPTY) {
833 		/* 11n aggregation queue is empty, re-fill */
834 		;
835 	}
836 	if (status & MACREG_A2HRIC_BIT_QUEUE_FULL) {
837 		;
838 	}
839 	if (status & MACREG_A2HRIC_BIT_RADAR_DETECT) {
840 		/* radar detected, process event */
841 		taskqueue_enqueue(sc->sc_tq, &sc->sc_radartask);
842 	}
843 	if (status & MACREG_A2HRIC_BIT_CHAN_SWITCH) {
844 		/* DFS channel switch */
845 		taskqueue_enqueue(sc->sc_tq, &sc->sc_chanswitchtask);
846 	}
847 }
848 
849 static void
850 mwl_radar_proc(void *arg, int pending)
851 {
852 	struct mwl_softc *sc = arg;
853 	struct ieee80211com *ic = &sc->sc_ic;
854 
855 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: radar detected, pending %u\n",
856 	    __func__, pending);
857 
858 	sc->sc_stats.mst_radardetect++;
859 	/* XXX stop h/w BA streams? */
860 
861 	IEEE80211_LOCK(ic);
862 	ieee80211_dfs_notify_radar(ic, ic->ic_curchan);
863 	IEEE80211_UNLOCK(ic);
864 }
865 
866 static void
867 mwl_chanswitch_proc(void *arg, int pending)
868 {
869 	struct mwl_softc *sc = arg;
870 	struct ieee80211com *ic = &sc->sc_ic;
871 
872 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: channel switch notice, pending %u\n",
873 	    __func__, pending);
874 
875 	IEEE80211_LOCK(ic);
876 	sc->sc_csapending = 0;
877 	ieee80211_csa_completeswitch(ic);
878 	IEEE80211_UNLOCK(ic);
879 }
880 
881 static void
882 mwl_bawatchdog(const MWL_HAL_BASTREAM *sp)
883 {
884 	struct ieee80211_node *ni = sp->data[0];
885 
886 	/* send DELBA and drop the stream */
887 	ieee80211_ampdu_stop(ni, sp->data[1], IEEE80211_REASON_UNSPECIFIED);
888 }
889 
890 static void
891 mwl_bawatchdog_proc(void *arg, int pending)
892 {
893 	struct mwl_softc *sc = arg;
894 	struct mwl_hal *mh = sc->sc_mh;
895 	const MWL_HAL_BASTREAM *sp;
896 	uint8_t bitmap, n;
897 
898 	sc->sc_stats.mst_bawatchdog++;
899 
900 	if (mwl_hal_getwatchdogbitmap(mh, &bitmap) != 0) {
901 		DPRINTF(sc, MWL_DEBUG_AMPDU,
902 		    "%s: could not get bitmap\n", __func__);
903 		sc->sc_stats.mst_bawatchdog_failed++;
904 		return;
905 	}
906 	DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: bitmap 0x%x\n", __func__, bitmap);
907 	if (bitmap == 0xff) {
908 		n = 0;
909 		/* disable all ba streams */
910 		for (bitmap = 0; bitmap < 8; bitmap++) {
911 			sp = mwl_hal_bastream_lookup(mh, bitmap);
912 			if (sp != NULL) {
913 				mwl_bawatchdog(sp);
914 				n++;
915 			}
916 		}
917 		if (n == 0) {
918 			DPRINTF(sc, MWL_DEBUG_AMPDU,
919 			    "%s: no BA streams found\n", __func__);
920 			sc->sc_stats.mst_bawatchdog_empty++;
921 		}
922 	} else if (bitmap != 0xaa) {
923 		/* disable a single ba stream */
924 		sp = mwl_hal_bastream_lookup(mh, bitmap);
925 		if (sp != NULL) {
926 			mwl_bawatchdog(sp);
927 		} else {
928 			DPRINTF(sc, MWL_DEBUG_AMPDU,
929 			    "%s: no BA stream %d\n", __func__, bitmap);
930 			sc->sc_stats.mst_bawatchdog_notfound++;
931 		}
932 	}
933 }
934 
935 /*
936  * Convert net80211 channel to a HAL channel.
937  */
938 static void
939 mwl_mapchan(MWL_HAL_CHANNEL *hc, const struct ieee80211_channel *chan)
940 {
941 	hc->channel = chan->ic_ieee;
942 
943 	*(uint32_t *)&hc->channelFlags = 0;
944 	if (IEEE80211_IS_CHAN_2GHZ(chan))
945 		hc->channelFlags.FreqBand = MWL_FREQ_BAND_2DOT4GHZ;
946 	else if (IEEE80211_IS_CHAN_5GHZ(chan))
947 		hc->channelFlags.FreqBand = MWL_FREQ_BAND_5GHZ;
948 	if (IEEE80211_IS_CHAN_HT40(chan)) {
949 		hc->channelFlags.ChnlWidth = MWL_CH_40_MHz_WIDTH;
950 		if (IEEE80211_IS_CHAN_HT40U(chan))
951 			hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_ABOVE_CTRL_CH;
952 		else
953 			hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_BELOW_CTRL_CH;
954 	} else
955 		hc->channelFlags.ChnlWidth = MWL_CH_20_MHz_WIDTH;
956 	/* XXX 10MHz channels */
957 }
958 
959 /*
960  * Inform firmware of our tx/rx dma setup.  The BAR 0
961  * writes below are for compatibility with older firmware.
962  * For current firmware we send this information with a
963  * cmd block via mwl_hal_sethwdma.
964  */
965 static int
966 mwl_setupdma(struct mwl_softc *sc)
967 {
968 	int error, i;
969 
970 	sc->sc_hwdma.rxDescRead = sc->sc_rxdma.dd_desc_paddr;
971 	WR4(sc, sc->sc_hwspecs.rxDescRead, sc->sc_hwdma.rxDescRead);
972 	WR4(sc, sc->sc_hwspecs.rxDescWrite, sc->sc_hwdma.rxDescRead);
973 
974 	for (i = 0; i < MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES; i++) {
975 		struct mwl_txq *txq = &sc->sc_txq[i];
976 		sc->sc_hwdma.wcbBase[i] = txq->dma.dd_desc_paddr;
977 		WR4(sc, sc->sc_hwspecs.wcbBase[i], sc->sc_hwdma.wcbBase[i]);
978 	}
979 	sc->sc_hwdma.maxNumTxWcb = mwl_txbuf;
980 	sc->sc_hwdma.maxNumWCB = MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES;
981 
982 	error = mwl_hal_sethwdma(sc->sc_mh, &sc->sc_hwdma);
983 	if (error != 0) {
984 		device_printf(sc->sc_dev,
985 		    "unable to setup tx/rx dma; hal status %u\n", error);
986 		/* XXX */
987 	}
988 	return error;
989 }
990 
991 /*
992  * Inform firmware of tx rate parameters.
993  * Called after a channel change.
994  */
995 static int
996 mwl_setcurchanrates(struct mwl_softc *sc)
997 {
998 	struct ieee80211com *ic = &sc->sc_ic;
999 	const struct ieee80211_rateset *rs;
1000 	MWL_HAL_TXRATE rates;
1001 
1002 	memset(&rates, 0, sizeof(rates));
1003 	rs = ieee80211_get_suprates(ic, ic->ic_curchan);
1004 	/* rate used to send management frames */
1005 	rates.MgtRate = rs->rs_rates[0] & IEEE80211_RATE_VAL;
1006 	/* rate used to send multicast frames */
1007 	rates.McastRate = rates.MgtRate;
1008 
1009 	return mwl_hal_settxrate_auto(sc->sc_mh, &rates);
1010 }
1011 
1012 /*
1013  * Inform firmware of tx rate parameters.  Called whenever
1014  * user-settable params change and after a channel change.
1015  */
1016 static int
1017 mwl_setrates(struct ieee80211vap *vap)
1018 {
1019 	struct mwl_vap *mvp = MWL_VAP(vap);
1020 	struct ieee80211_node *ni = vap->iv_bss;
1021 	const struct ieee80211_txparam *tp = ni->ni_txparms;
1022 	MWL_HAL_TXRATE rates;
1023 
1024 	KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1025 
1026 	/*
1027 	 * Update the h/w rate map.
1028 	 * NB: 0x80 for MCS is passed through unchanged
1029 	 */
1030 	memset(&rates, 0, sizeof(rates));
1031 	/* rate used to send management frames */
1032 	rates.MgtRate = tp->mgmtrate;
1033 	/* rate used to send multicast frames */
1034 	rates.McastRate = tp->mcastrate;
1035 
1036 	/* while here calculate EAPOL fixed rate cookie */
1037 	mvp->mv_eapolformat = htole16(mwl_calcformat(rates.MgtRate, ni));
1038 
1039 	return mwl_hal_settxrate(mvp->mv_hvap,
1040 	    tp->ucastrate != IEEE80211_FIXED_RATE_NONE ?
1041 		RATE_FIXED : RATE_AUTO, &rates);
1042 }
1043 
1044 /*
1045  * Setup a fixed xmit rate cookie for EAPOL frames.
1046  */
1047 static void
1048 mwl_seteapolformat(struct ieee80211vap *vap)
1049 {
1050 	struct mwl_vap *mvp = MWL_VAP(vap);
1051 	struct ieee80211_node *ni = vap->iv_bss;
1052 	enum ieee80211_phymode mode;
1053 	uint8_t rate;
1054 
1055 	KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1056 
1057 	mode = ieee80211_chan2mode(ni->ni_chan);
1058 	/*
1059 	 * Use legacy rates when operating a mixed HT+non-HT bss.
1060 	 * NB: this may violate POLA for sta and wds vap's.
1061 	 */
1062 	if (mode == IEEE80211_MODE_11NA &&
1063 	    (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1064 		rate = vap->iv_txparms[IEEE80211_MODE_11A].mgmtrate;
1065 	else if (mode == IEEE80211_MODE_11NG &&
1066 	    (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1067 		rate = vap->iv_txparms[IEEE80211_MODE_11G].mgmtrate;
1068 	else
1069 		rate = vap->iv_txparms[mode].mgmtrate;
1070 
1071 	mvp->mv_eapolformat = htole16(mwl_calcformat(rate, ni));
1072 }
1073 
1074 /*
1075  * Map SKU+country code to region code for radar bin'ing.
1076  */
1077 static int
1078 mwl_map2regioncode(const struct ieee80211_regdomain *rd)
1079 {
1080 	switch (rd->regdomain) {
1081 	case SKU_FCC:
1082 	case SKU_FCC3:
1083 		return DOMAIN_CODE_FCC;
1084 	case SKU_CA:
1085 		return DOMAIN_CODE_IC;
1086 	case SKU_ETSI:
1087 	case SKU_ETSI2:
1088 	case SKU_ETSI3:
1089 		if (rd->country == CTRY_SPAIN)
1090 			return DOMAIN_CODE_SPAIN;
1091 		if (rd->country == CTRY_FRANCE || rd->country == CTRY_FRANCE2)
1092 			return DOMAIN_CODE_FRANCE;
1093 		/* XXX force 1.3.1 radar type */
1094 		return DOMAIN_CODE_ETSI_131;
1095 	case SKU_JAPAN:
1096 		return DOMAIN_CODE_MKK;
1097 	case SKU_ROW:
1098 		return DOMAIN_CODE_DGT;	/* Taiwan */
1099 	case SKU_APAC:
1100 	case SKU_APAC2:
1101 	case SKU_APAC3:
1102 		return DOMAIN_CODE_AUS;	/* Australia */
1103 	}
1104 	/* XXX KOREA? */
1105 	return DOMAIN_CODE_FCC;			/* XXX? */
1106 }
1107 
1108 static int
1109 mwl_hal_reset(struct mwl_softc *sc)
1110 {
1111 	struct ieee80211com *ic = &sc->sc_ic;
1112 	struct mwl_hal *mh = sc->sc_mh;
1113 
1114 	mwl_hal_setantenna(mh, WL_ANTENNATYPE_RX, sc->sc_rxantenna);
1115 	mwl_hal_setantenna(mh, WL_ANTENNATYPE_TX, sc->sc_txantenna);
1116 	mwl_hal_setradio(mh, 1, WL_AUTO_PREAMBLE);
1117 	mwl_hal_setwmm(sc->sc_mh, (ic->ic_flags & IEEE80211_F_WME) != 0);
1118 	mwl_chan_set(sc, ic->ic_curchan);
1119 	/* NB: RF/RA performance tuned for indoor mode */
1120 	mwl_hal_setrateadaptmode(mh, 0);
1121 	mwl_hal_setoptimizationlevel(mh,
1122 	    (ic->ic_flags & IEEE80211_F_BURST) != 0);
1123 
1124 	mwl_hal_setregioncode(mh, mwl_map2regioncode(&ic->ic_regdomain));
1125 
1126 	mwl_hal_setaggampduratemode(mh, 1, 80);		/* XXX */
1127 	mwl_hal_setcfend(mh, 0);			/* XXX */
1128 
1129 	return 1;
1130 }
1131 
1132 static int
1133 mwl_init(struct mwl_softc *sc)
1134 {
1135 	struct mwl_hal *mh = sc->sc_mh;
1136 	int error = 0;
1137 
1138 	MWL_LOCK_ASSERT(sc);
1139 
1140 	/*
1141 	 * Stop anything previously setup.  This is safe
1142 	 * whether this is the first time through or not.
1143 	 */
1144 	mwl_stop(sc);
1145 
1146 	/*
1147 	 * Push vap-independent state to the firmware.
1148 	 */
1149 	if (!mwl_hal_reset(sc)) {
1150 		device_printf(sc->sc_dev, "unable to reset hardware\n");
1151 		return EIO;
1152 	}
1153 
1154 	/*
1155 	 * Setup recv (once); transmit is already good to go.
1156 	 */
1157 	error = mwl_startrecv(sc);
1158 	if (error != 0) {
1159 		device_printf(sc->sc_dev, "unable to start recv logic\n");
1160 		return error;
1161 	}
1162 
1163 	/*
1164 	 * Enable interrupts.
1165 	 */
1166 	sc->sc_imask = MACREG_A2HRIC_BIT_RX_RDY
1167 		     | MACREG_A2HRIC_BIT_TX_DONE
1168 		     | MACREG_A2HRIC_BIT_OPC_DONE
1169 #if 0
1170 		     | MACREG_A2HRIC_BIT_MAC_EVENT
1171 #endif
1172 		     | MACREG_A2HRIC_BIT_ICV_ERROR
1173 		     | MACREG_A2HRIC_BIT_RADAR_DETECT
1174 		     | MACREG_A2HRIC_BIT_CHAN_SWITCH
1175 #if 0
1176 		     | MACREG_A2HRIC_BIT_QUEUE_EMPTY
1177 #endif
1178 		     | MACREG_A2HRIC_BIT_BA_WATCHDOG
1179 		     | MACREQ_A2HRIC_BIT_TX_ACK
1180 		     ;
1181 
1182 	sc->sc_running = 1;
1183 	mwl_hal_intrset(mh, sc->sc_imask);
1184 	callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
1185 
1186 	return 0;
1187 }
1188 
1189 static void
1190 mwl_stop(struct mwl_softc *sc)
1191 {
1192 
1193 	MWL_LOCK_ASSERT(sc);
1194 	if (sc->sc_running) {
1195 		/*
1196 		 * Shutdown the hardware and driver.
1197 		 */
1198 		sc->sc_running = 0;
1199 		callout_stop(&sc->sc_watchdog);
1200 		sc->sc_tx_timer = 0;
1201 		mwl_draintxq(sc);
1202 	}
1203 }
1204 
1205 static int
1206 mwl_reset_vap(struct ieee80211vap *vap, int state)
1207 {
1208 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1209 	struct ieee80211com *ic = vap->iv_ic;
1210 
1211 	if (state == IEEE80211_S_RUN)
1212 		mwl_setrates(vap);
1213 	/* XXX off by 1? */
1214 	mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
1215 	/* XXX auto? 20/40 split? */
1216 	mwl_hal_sethtgi(hvap, (vap->iv_flags_ht &
1217 	    (IEEE80211_FHT_SHORTGI20|IEEE80211_FHT_SHORTGI40)) ? 1 : 0);
1218 	mwl_hal_setnprot(hvap, ic->ic_htprotmode == IEEE80211_PROT_NONE ?
1219 	    HTPROTECT_NONE : HTPROTECT_AUTO);
1220 	/* XXX txpower cap */
1221 
1222 	/* re-setup beacons */
1223 	if (state == IEEE80211_S_RUN &&
1224 	    (vap->iv_opmode == IEEE80211_M_HOSTAP ||
1225 	     vap->iv_opmode == IEEE80211_M_MBSS ||
1226 	     vap->iv_opmode == IEEE80211_M_IBSS)) {
1227 		mwl_setapmode(vap, vap->iv_bss->ni_chan);
1228 		mwl_hal_setnprotmode(hvap, _IEEE80211_MASKSHIFT(
1229 		    ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1230 		return mwl_beacon_setup(vap);
1231 	}
1232 	return 0;
1233 }
1234 
1235 /*
1236  * Reset the hardware w/o losing operational state.
1237  * Used to reset or reload hardware state for a vap.
1238  */
1239 static int
1240 mwl_reset(struct ieee80211vap *vap, u_long cmd)
1241 {
1242 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1243 	int error = 0;
1244 
1245 	if (hvap != NULL) {			/* WDS, MONITOR, etc. */
1246 		struct ieee80211com *ic = vap->iv_ic;
1247 		struct mwl_softc *sc = ic->ic_softc;
1248 		struct mwl_hal *mh = sc->sc_mh;
1249 
1250 		/* XXX handle DWDS sta vap change */
1251 		/* XXX do we need to disable interrupts? */
1252 		mwl_hal_intrset(mh, 0);		/* disable interrupts */
1253 		error = mwl_reset_vap(vap, vap->iv_state);
1254 		mwl_hal_intrset(mh, sc->sc_imask);
1255 	}
1256 	return error;
1257 }
1258 
1259 /*
1260  * Allocate a tx buffer for sending a frame.  The
1261  * packet is assumed to have the WME AC stored so
1262  * we can use it to select the appropriate h/w queue.
1263  */
1264 static struct mwl_txbuf *
1265 mwl_gettxbuf(struct mwl_softc *sc, struct mwl_txq *txq)
1266 {
1267 	struct mwl_txbuf *bf;
1268 
1269 	/*
1270 	 * Grab a TX buffer and associated resources.
1271 	 */
1272 	MWL_TXQ_LOCK(txq);
1273 	bf = STAILQ_FIRST(&txq->free);
1274 	if (bf != NULL) {
1275 		STAILQ_REMOVE_HEAD(&txq->free, bf_list);
1276 		txq->nfree--;
1277 	}
1278 	MWL_TXQ_UNLOCK(txq);
1279 	if (bf == NULL)
1280 		DPRINTF(sc, MWL_DEBUG_XMIT,
1281 		    "%s: out of xmit buffers on q %d\n", __func__, txq->qnum);
1282 	return bf;
1283 }
1284 
1285 /*
1286  * Return a tx buffer to the queue it came from.  Note there
1287  * are two cases because we must preserve the order of buffers
1288  * as it reflects the fixed order of descriptors in memory
1289  * (the firmware pre-fetches descriptors so we cannot reorder).
1290  */
1291 static void
1292 mwl_puttxbuf_head(struct mwl_txq *txq, struct mwl_txbuf *bf)
1293 {
1294 	bf->bf_m = NULL;
1295 	bf->bf_node = NULL;
1296 	MWL_TXQ_LOCK(txq);
1297 	STAILQ_INSERT_HEAD(&txq->free, bf, bf_list);
1298 	txq->nfree++;
1299 	MWL_TXQ_UNLOCK(txq);
1300 }
1301 
1302 static void
1303 mwl_puttxbuf_tail(struct mwl_txq *txq, struct mwl_txbuf *bf)
1304 {
1305 	bf->bf_m = NULL;
1306 	bf->bf_node = NULL;
1307 	MWL_TXQ_LOCK(txq);
1308 	STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
1309 	txq->nfree++;
1310 	MWL_TXQ_UNLOCK(txq);
1311 }
1312 
1313 static int
1314 mwl_transmit(struct ieee80211com *ic, struct mbuf *m)
1315 {
1316 	struct mwl_softc *sc = ic->ic_softc;
1317 	int error;
1318 
1319 	MWL_LOCK(sc);
1320 	if (!sc->sc_running) {
1321 		MWL_UNLOCK(sc);
1322 		return (ENXIO);
1323 	}
1324 	error = mbufq_enqueue(&sc->sc_snd, m);
1325 	if (error) {
1326 		MWL_UNLOCK(sc);
1327 		return (error);
1328 	}
1329 	mwl_start(sc);
1330 	MWL_UNLOCK(sc);
1331 	return (0);
1332 }
1333 
1334 static void
1335 mwl_start(struct mwl_softc *sc)
1336 {
1337 	struct ieee80211_node *ni;
1338 	struct mwl_txbuf *bf;
1339 	struct mbuf *m;
1340 	struct mwl_txq *txq = NULL;	/* XXX silence gcc */
1341 	int nqueued;
1342 
1343 	MWL_LOCK_ASSERT(sc);
1344 	if (!sc->sc_running || sc->sc_invalid)
1345 		return;
1346 	nqueued = 0;
1347 	while ((m = mbufq_dequeue(&sc->sc_snd)) != NULL) {
1348 		/*
1349 		 * Grab the node for the destination.
1350 		 */
1351 		ni = (struct ieee80211_node *) m->m_pkthdr.rcvif;
1352 		KASSERT(ni != NULL, ("no node"));
1353 		m->m_pkthdr.rcvif = NULL;	/* committed, clear ref */
1354 		/*
1355 		 * Grab a TX buffer and associated resources.
1356 		 * We honor the classification by the 802.11 layer.
1357 		 */
1358 		txq = sc->sc_ac2q[M_WME_GETAC(m)];
1359 		bf = mwl_gettxbuf(sc, txq);
1360 		if (bf == NULL) {
1361 			m_freem(m);
1362 			ieee80211_free_node(ni);
1363 #ifdef MWL_TX_NODROP
1364 			sc->sc_stats.mst_tx_qstop++;
1365 			break;
1366 #else
1367 			DPRINTF(sc, MWL_DEBUG_XMIT,
1368 			    "%s: tail drop on q %d\n", __func__, txq->qnum);
1369 			sc->sc_stats.mst_tx_qdrop++;
1370 			continue;
1371 #endif /* MWL_TX_NODROP */
1372 		}
1373 
1374 		/*
1375 		 * Pass the frame to the h/w for transmission.
1376 		 */
1377 		if (mwl_tx_start(sc, ni, bf, m)) {
1378 			if_inc_counter(ni->ni_vap->iv_ifp,
1379 			    IFCOUNTER_OERRORS, 1);
1380 			mwl_puttxbuf_head(txq, bf);
1381 			ieee80211_free_node(ni);
1382 			continue;
1383 		}
1384 		nqueued++;
1385 		if (nqueued >= mwl_txcoalesce) {
1386 			/*
1387 			 * Poke the firmware to process queued frames;
1388 			 * see below about (lack of) locking.
1389 			 */
1390 			nqueued = 0;
1391 			mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1392 		}
1393 	}
1394 	if (nqueued) {
1395 		/*
1396 		 * NB: We don't need to lock against tx done because
1397 		 * this just prods the firmware to check the transmit
1398 		 * descriptors.  The firmware will also start fetching
1399 		 * descriptors by itself if it notices new ones are
1400 		 * present when it goes to deliver a tx done interrupt
1401 		 * to the host. So if we race with tx done processing
1402 		 * it's ok.  Delivering the kick here rather than in
1403 		 * mwl_tx_start is an optimization to avoid poking the
1404 		 * firmware for each packet.
1405 		 *
1406 		 * NB: the queue id isn't used so 0 is ok.
1407 		 */
1408 		mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1409 	}
1410 }
1411 
1412 static int
1413 mwl_raw_xmit(struct ieee80211_node *ni, struct mbuf *m,
1414 	const struct ieee80211_bpf_params *params)
1415 {
1416 	struct ieee80211com *ic = ni->ni_ic;
1417 	struct mwl_softc *sc = ic->ic_softc;
1418 	struct mwl_txbuf *bf;
1419 	struct mwl_txq *txq;
1420 
1421 	if (!sc->sc_running || sc->sc_invalid) {
1422 		m_freem(m);
1423 		return ENETDOWN;
1424 	}
1425 	/*
1426 	 * Grab a TX buffer and associated resources.
1427 	 * Note that we depend on the classification
1428 	 * by the 802.11 layer to get to the right h/w
1429 	 * queue.  Management frames must ALWAYS go on
1430 	 * queue 1 but we cannot just force that here
1431 	 * because we may receive non-mgt frames.
1432 	 */
1433 	txq = sc->sc_ac2q[M_WME_GETAC(m)];
1434 	bf = mwl_gettxbuf(sc, txq);
1435 	if (bf == NULL) {
1436 		sc->sc_stats.mst_tx_qstop++;
1437 		m_freem(m);
1438 		return ENOBUFS;
1439 	}
1440 	/*
1441 	 * Pass the frame to the h/w for transmission.
1442 	 */
1443 	if (mwl_tx_start(sc, ni, bf, m)) {
1444 		mwl_puttxbuf_head(txq, bf);
1445 
1446 		return EIO;		/* XXX */
1447 	}
1448 	/*
1449 	 * NB: We don't need to lock against tx done because
1450 	 * this just prods the firmware to check the transmit
1451 	 * descriptors.  The firmware will also start fetching
1452 	 * descriptors by itself if it notices new ones are
1453 	 * present when it goes to deliver a tx done interrupt
1454 	 * to the host. So if we race with tx done processing
1455 	 * it's ok.  Delivering the kick here rather than in
1456 	 * mwl_tx_start is an optimization to avoid poking the
1457 	 * firmware for each packet.
1458 	 *
1459 	 * NB: the queue id isn't used so 0 is ok.
1460 	 */
1461 	mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1462 	return 0;
1463 }
1464 
1465 static int
1466 mwl_media_change(if_t ifp)
1467 {
1468 	struct ieee80211vap *vap;
1469 	int error;
1470 
1471 	/* NB: only the fixed rate can change and that doesn't need a reset */
1472 	error = ieee80211_media_change(ifp);
1473 	if (error != 0)
1474 		return (error);
1475 
1476 	vap = if_getsoftc(ifp);
1477 	mwl_setrates(vap);
1478 	return (0);
1479 }
1480 
1481 #ifdef MWL_DEBUG
1482 static void
1483 mwl_keyprint(struct mwl_softc *sc, const char *tag,
1484 	const MWL_HAL_KEYVAL *hk, const uint8_t mac[IEEE80211_ADDR_LEN])
1485 {
1486 	static const char *ciphers[] = {
1487 		"WEP",
1488 		"TKIP",
1489 		"AES-CCM",
1490 	};
1491 	int i, n;
1492 
1493 	printf("%s: [%u] %-7s", tag, hk->keyIndex, ciphers[hk->keyTypeId]);
1494 	for (i = 0, n = hk->keyLen; i < n; i++)
1495 		printf(" %02x", hk->key.aes[i]);
1496 	printf(" mac %s", ether_sprintf(mac));
1497 	if (hk->keyTypeId == KEY_TYPE_ID_TKIP) {
1498 		printf(" %s", "rxmic");
1499 		for (i = 0; i < sizeof(hk->key.tkip.rxMic); i++)
1500 			printf(" %02x", hk->key.tkip.rxMic[i]);
1501 		printf(" txmic");
1502 		for (i = 0; i < sizeof(hk->key.tkip.txMic); i++)
1503 			printf(" %02x", hk->key.tkip.txMic[i]);
1504 	}
1505 	printf(" flags 0x%x\n", hk->keyFlags);
1506 }
1507 #endif
1508 
1509 /*
1510  * Allocate a key cache slot for a unicast key.  The
1511  * firmware handles key allocation and every station is
1512  * guaranteed key space so we are always successful.
1513  */
1514 static int
1515 mwl_key_alloc(struct ieee80211vap *vap, struct ieee80211_key *k,
1516 	ieee80211_keyix *keyix, ieee80211_keyix *rxkeyix)
1517 {
1518 	struct mwl_softc *sc = vap->iv_ic->ic_softc;
1519 
1520 	if (k->wk_keyix != IEEE80211_KEYIX_NONE ||
1521 	    (k->wk_flags & IEEE80211_KEY_GROUP)) {
1522 		if (!(&vap->iv_nw_keys[0] <= k &&
1523 		      k < &vap->iv_nw_keys[IEEE80211_WEP_NKID])) {
1524 			/* should not happen */
1525 			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1526 				"%s: bogus group key\n", __func__);
1527 			return 0;
1528 		}
1529 		/* give the caller what they requested */
1530 		*keyix = *rxkeyix = ieee80211_crypto_get_key_wepidx(vap, k);
1531 	} else {
1532 		/*
1533 		 * Firmware handles key allocation.
1534 		 */
1535 		*keyix = *rxkeyix = 0;
1536 	}
1537 	return 1;
1538 }
1539 
1540 /*
1541  * Delete a key entry allocated by mwl_key_alloc.
1542  */
1543 static int
1544 mwl_key_delete(struct ieee80211vap *vap, const struct ieee80211_key *k)
1545 {
1546 	struct mwl_softc *sc = vap->iv_ic->ic_softc;
1547 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1548 	MWL_HAL_KEYVAL hk;
1549 	const uint8_t bcastaddr[IEEE80211_ADDR_LEN] =
1550 	    { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff };
1551 
1552 	if (hvap == NULL) {
1553 		if (vap->iv_opmode != IEEE80211_M_WDS) {
1554 			/* XXX monitor mode? */
1555 			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1556 			    "%s: no hvap for opmode %d\n", __func__,
1557 			    vap->iv_opmode);
1558 			return 0;
1559 		}
1560 		hvap = MWL_VAP(vap)->mv_ap_hvap;
1561 	}
1562 
1563 	DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: delete key %u\n",
1564 	    __func__, k->wk_keyix);
1565 
1566 	memset(&hk, 0, sizeof(hk));
1567 	hk.keyIndex = k->wk_keyix;
1568 	switch (k->wk_cipher->ic_cipher) {
1569 	case IEEE80211_CIPHER_WEP:
1570 		hk.keyTypeId = KEY_TYPE_ID_WEP;
1571 		break;
1572 	case IEEE80211_CIPHER_TKIP:
1573 		hk.keyTypeId = KEY_TYPE_ID_TKIP;
1574 		break;
1575 	case IEEE80211_CIPHER_AES_CCM:
1576 		hk.keyTypeId = KEY_TYPE_ID_AES;
1577 		break;
1578 	default:
1579 		/* XXX should not happen */
1580 		DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1581 		    __func__, k->wk_cipher->ic_cipher);
1582 		return 0;
1583 	}
1584 	return (mwl_hal_keyreset(hvap, &hk, bcastaddr) == 0);	/*XXX*/
1585 }
1586 
1587 static __inline int
1588 addgroupflags(MWL_HAL_KEYVAL *hk, const struct ieee80211_key *k)
1589 {
1590 	if (k->wk_flags & IEEE80211_KEY_GROUP) {
1591 		if (k->wk_flags & IEEE80211_KEY_XMIT)
1592 			hk->keyFlags |= KEY_FLAG_TXGROUPKEY;
1593 		if (k->wk_flags & IEEE80211_KEY_RECV)
1594 			hk->keyFlags |= KEY_FLAG_RXGROUPKEY;
1595 		return 1;
1596 	} else
1597 		return 0;
1598 }
1599 
1600 /*
1601  * Set the key cache contents for the specified key.  Key cache
1602  * slot(s) must already have been allocated by mwl_key_alloc.
1603  */
1604 static int
1605 mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k)
1606 {
1607 	return (_mwl_key_set(vap, k, k->wk_macaddr));
1608 }
1609 
1610 static int
1611 _mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k,
1612 	const uint8_t mac[IEEE80211_ADDR_LEN])
1613 {
1614 #define	GRPXMIT	(IEEE80211_KEY_XMIT | IEEE80211_KEY_GROUP)
1615 /* NB: static wep keys are marked GROUP+tx/rx; GTK will be tx or rx */
1616 #define	IEEE80211_IS_STATICKEY(k) \
1617 	(((k)->wk_flags & (GRPXMIT|IEEE80211_KEY_RECV)) == \
1618 	 (GRPXMIT|IEEE80211_KEY_RECV))
1619 	struct mwl_softc *sc = vap->iv_ic->ic_softc;
1620 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1621 	const struct ieee80211_cipher *cip = k->wk_cipher;
1622 	const uint8_t *macaddr;
1623 	MWL_HAL_KEYVAL hk;
1624 
1625 	KASSERT((k->wk_flags & IEEE80211_KEY_SWCRYPT) == 0,
1626 		("s/w crypto set?"));
1627 
1628 	if (hvap == NULL) {
1629 		if (vap->iv_opmode != IEEE80211_M_WDS) {
1630 			/* XXX monitor mode? */
1631 			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1632 			    "%s: no hvap for opmode %d\n", __func__,
1633 			    vap->iv_opmode);
1634 			return 0;
1635 		}
1636 		hvap = MWL_VAP(vap)->mv_ap_hvap;
1637 	}
1638 	memset(&hk, 0, sizeof(hk));
1639 	hk.keyIndex = k->wk_keyix;
1640 	switch (cip->ic_cipher) {
1641 	case IEEE80211_CIPHER_WEP:
1642 		hk.keyTypeId = KEY_TYPE_ID_WEP;
1643 		hk.keyLen = k->wk_keylen;
1644 		if (k->wk_keyix == vap->iv_def_txkey)
1645 			hk.keyFlags = KEY_FLAG_WEP_TXKEY;
1646 		if (!IEEE80211_IS_STATICKEY(k)) {
1647 			/* NB: WEP is never used for the PTK */
1648 			(void) addgroupflags(&hk, k);
1649 		}
1650 		break;
1651 	case IEEE80211_CIPHER_TKIP:
1652 		hk.keyTypeId = KEY_TYPE_ID_TKIP;
1653 		hk.key.tkip.tsc.high = (uint32_t)(k->wk_keytsc >> 16);
1654 		hk.key.tkip.tsc.low = (uint16_t)k->wk_keytsc;
1655 		hk.keyFlags = KEY_FLAG_TSC_VALID | KEY_FLAG_MICKEY_VALID;
1656 		hk.keyLen = k->wk_keylen + IEEE80211_MICBUF_SIZE;
1657 		if (!addgroupflags(&hk, k))
1658 			hk.keyFlags |= KEY_FLAG_PAIRWISE;
1659 		break;
1660 	case IEEE80211_CIPHER_AES_CCM:
1661 		hk.keyTypeId = KEY_TYPE_ID_AES;
1662 		hk.keyLen = k->wk_keylen;
1663 		if (!addgroupflags(&hk, k))
1664 			hk.keyFlags |= KEY_FLAG_PAIRWISE;
1665 		break;
1666 	default:
1667 		/* XXX should not happen */
1668 		DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1669 		    __func__, k->wk_cipher->ic_cipher);
1670 		return 0;
1671 	}
1672 	/*
1673 	 * NB: tkip mic keys get copied here too; the layout
1674 	 *     just happens to match that in ieee80211_key.
1675 	 */
1676 	memcpy(hk.key.aes, k->wk_key, hk.keyLen);
1677 
1678 	/*
1679 	 * Locate address of sta db entry for writing key;
1680 	 * the convention unfortunately is somewhat different
1681 	 * than how net80211, hostapd, and wpa_supplicant think.
1682 	 */
1683 	if (vap->iv_opmode == IEEE80211_M_STA) {
1684 		/*
1685 		 * NB: keys plumbed before the sta reaches AUTH state
1686 		 * will be discarded or written to the wrong sta db
1687 		 * entry because iv_bss is meaningless.  This is ok
1688 		 * (right now) because we handle deferred plumbing of
1689 		 * WEP keys when the sta reaches AUTH state.
1690 		 */
1691 		macaddr = vap->iv_bss->ni_bssid;
1692 		if ((k->wk_flags & IEEE80211_KEY_GROUP) == 0) {
1693 			/* XXX plumb to local sta db too for static key wep */
1694 			mwl_hal_keyset(hvap, &hk, vap->iv_myaddr);
1695 		}
1696 	} else if (vap->iv_opmode == IEEE80211_M_WDS &&
1697 	    vap->iv_state != IEEE80211_S_RUN) {
1698 		/*
1699 		 * Prior to RUN state a WDS vap will not it's BSS node
1700 		 * setup so we will plumb the key to the wrong mac
1701 		 * address (it'll be our local address).  Workaround
1702 		 * this for the moment by grabbing the correct address.
1703 		 */
1704 		macaddr = vap->iv_des_bssid;
1705 	} else if ((k->wk_flags & GRPXMIT) == GRPXMIT)
1706 		macaddr = vap->iv_myaddr;
1707 	else
1708 		macaddr = mac;
1709 	KEYPRINTF(sc, &hk, macaddr);
1710 	return (mwl_hal_keyset(hvap, &hk, macaddr) == 0);
1711 #undef IEEE80211_IS_STATICKEY
1712 #undef GRPXMIT
1713 }
1714 
1715 /*
1716  * Set the multicast filter contents into the hardware.
1717  * XXX f/w has no support; just defer to the os.
1718  */
1719 static void
1720 mwl_setmcastfilter(struct mwl_softc *sc)
1721 {
1722 #if 0
1723 	struct ether_multi *enm;
1724 	struct ether_multistep estep;
1725 	uint8_t macs[IEEE80211_ADDR_LEN*MWL_HAL_MCAST_MAX];/* XXX stack use */
1726 	uint8_t *mp;
1727 	int nmc;
1728 
1729 	mp = macs;
1730 	nmc = 0;
1731 	ETHER_FIRST_MULTI(estep, &sc->sc_ec, enm);
1732 	while (enm != NULL) {
1733 		/* XXX Punt on ranges. */
1734 		if (nmc == MWL_HAL_MCAST_MAX ||
1735 		    !IEEE80211_ADDR_EQ(enm->enm_addrlo, enm->enm_addrhi)) {
1736 			if_setflagsbit(ifp, IFF_ALLMULTI, 0);
1737 			return;
1738 		}
1739 		IEEE80211_ADDR_COPY(mp, enm->enm_addrlo);
1740 		mp += IEEE80211_ADDR_LEN, nmc++;
1741 		ETHER_NEXT_MULTI(estep, enm);
1742 	}
1743 	if_setflagsbit(ifp, 0, IFF_ALLMULTI);
1744 	mwl_hal_setmcast(sc->sc_mh, nmc, macs);
1745 #endif
1746 }
1747 
1748 static int
1749 mwl_mode_init(struct mwl_softc *sc)
1750 {
1751 	struct ieee80211com *ic = &sc->sc_ic;
1752 	struct mwl_hal *mh = sc->sc_mh;
1753 
1754 	mwl_hal_setpromisc(mh, ic->ic_promisc > 0);
1755 	mwl_setmcastfilter(sc);
1756 
1757 	return 0;
1758 }
1759 
1760 /*
1761  * Callback from the 802.11 layer after a multicast state change.
1762  */
1763 static void
1764 mwl_update_mcast(struct ieee80211com *ic)
1765 {
1766 	struct mwl_softc *sc = ic->ic_softc;
1767 
1768 	mwl_setmcastfilter(sc);
1769 }
1770 
1771 /*
1772  * Callback from the 802.11 layer after a promiscuous mode change.
1773  * Note this interface does not check the operating mode as this
1774  * is an internal callback and we are expected to honor the current
1775  * state (e.g. this is used for setting the interface in promiscuous
1776  * mode when operating in hostap mode to do ACS).
1777  */
1778 static void
1779 mwl_update_promisc(struct ieee80211com *ic)
1780 {
1781 	struct mwl_softc *sc = ic->ic_softc;
1782 
1783 	mwl_hal_setpromisc(sc->sc_mh, ic->ic_promisc > 0);
1784 }
1785 
1786 /*
1787  * Callback from the 802.11 layer to update the slot time
1788  * based on the current setting.  We use it to notify the
1789  * firmware of ERP changes and the f/w takes care of things
1790  * like slot time and preamble.
1791  */
1792 static void
1793 mwl_updateslot(struct ieee80211com *ic)
1794 {
1795 	struct mwl_softc *sc = ic->ic_softc;
1796 	struct mwl_hal *mh = sc->sc_mh;
1797 	int prot;
1798 
1799 	/* NB: can be called early; suppress needless cmds */
1800 	if (!sc->sc_running)
1801 		return;
1802 
1803 	/*
1804 	 * Calculate the ERP flags.  The firwmare will use
1805 	 * this to carry out the appropriate measures.
1806 	 */
1807 	prot = 0;
1808 	if (IEEE80211_IS_CHAN_ANYG(ic->ic_curchan)) {
1809 		if ((ic->ic_flags & IEEE80211_F_SHSLOT) == 0)
1810 			prot |= IEEE80211_ERP_NON_ERP_PRESENT;
1811 		if (ic->ic_flags & IEEE80211_F_USEPROT)
1812 			prot |= IEEE80211_ERP_USE_PROTECTION;
1813 		if (ic->ic_flags & IEEE80211_F_USEBARKER)
1814 			prot |= IEEE80211_ERP_LONG_PREAMBLE;
1815 	}
1816 
1817 	DPRINTF(sc, MWL_DEBUG_RESET,
1818 	    "%s: chan %u MHz/flags 0x%x %s slot, (prot 0x%x ic_flags 0x%x)\n",
1819 	    __func__, ic->ic_curchan->ic_freq, ic->ic_curchan->ic_flags,
1820 	    ic->ic_flags & IEEE80211_F_SHSLOT ? "short" : "long", prot,
1821 	    ic->ic_flags);
1822 
1823 	mwl_hal_setgprot(mh, prot);
1824 }
1825 
1826 /*
1827  * Setup the beacon frame.
1828  */
1829 static int
1830 mwl_beacon_setup(struct ieee80211vap *vap)
1831 {
1832 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1833 	struct ieee80211_node *ni = vap->iv_bss;
1834 	struct mbuf *m;
1835 
1836 	m = ieee80211_beacon_alloc(ni);
1837 	if (m == NULL)
1838 		return ENOBUFS;
1839 	mwl_hal_setbeacon(hvap, mtod(m, const void *), m->m_len);
1840 	m_free(m);
1841 
1842 	return 0;
1843 }
1844 
1845 /*
1846  * Update the beacon frame in response to a change.
1847  */
1848 static void
1849 mwl_beacon_update(struct ieee80211vap *vap, int item)
1850 {
1851 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1852 	struct ieee80211com *ic = vap->iv_ic;
1853 
1854 	KASSERT(hvap != NULL, ("no beacon"));
1855 	switch (item) {
1856 	case IEEE80211_BEACON_ERP:
1857 		mwl_updateslot(ic);
1858 		break;
1859 	case IEEE80211_BEACON_HTINFO:
1860 		mwl_hal_setnprotmode(hvap, _IEEE80211_MASKSHIFT(
1861 		    ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1862 		break;
1863 	case IEEE80211_BEACON_CAPS:
1864 	case IEEE80211_BEACON_WME:
1865 	case IEEE80211_BEACON_APPIE:
1866 	case IEEE80211_BEACON_CSA:
1867 		break;
1868 	case IEEE80211_BEACON_TIM:
1869 		/* NB: firmware always forms TIM */
1870 		return;
1871 	}
1872 	/* XXX retain beacon frame and update */
1873 	mwl_beacon_setup(vap);
1874 }
1875 
1876 static void
1877 mwl_load_cb(void *arg, bus_dma_segment_t *segs, int nsegs, int error)
1878 {
1879 	bus_addr_t *paddr = (bus_addr_t*) arg;
1880 	KASSERT(error == 0, ("error %u on bus_dma callback", error));
1881 	*paddr = segs->ds_addr;
1882 }
1883 
1884 #ifdef MWL_HOST_PS_SUPPORT
1885 /*
1886  * Handle power save station occupancy changes.
1887  */
1888 static void
1889 mwl_update_ps(struct ieee80211vap *vap, int nsta)
1890 {
1891 	struct mwl_vap *mvp = MWL_VAP(vap);
1892 
1893 	if (nsta == 0 || mvp->mv_last_ps_sta == 0)
1894 		mwl_hal_setpowersave_bss(mvp->mv_hvap, nsta);
1895 	mvp->mv_last_ps_sta = nsta;
1896 }
1897 
1898 /*
1899  * Handle associated station power save state changes.
1900  */
1901 static int
1902 mwl_set_tim(struct ieee80211_node *ni, int set)
1903 {
1904 	struct ieee80211vap *vap = ni->ni_vap;
1905 	struct mwl_vap *mvp = MWL_VAP(vap);
1906 
1907 	if (mvp->mv_set_tim(ni, set)) {		/* NB: state change */
1908 		mwl_hal_setpowersave_sta(mvp->mv_hvap,
1909 		    IEEE80211_AID(ni->ni_associd), set);
1910 		return 1;
1911 	} else
1912 		return 0;
1913 }
1914 #endif /* MWL_HOST_PS_SUPPORT */
1915 
1916 static int
1917 mwl_desc_setup(struct mwl_softc *sc, const char *name,
1918 	struct mwl_descdma *dd,
1919 	int nbuf, size_t bufsize, int ndesc, size_t descsize)
1920 {
1921 	uint8_t *ds;
1922 	int error;
1923 
1924 	DPRINTF(sc, MWL_DEBUG_RESET,
1925 	    "%s: %s DMA: %u bufs (%ju) %u desc/buf (%ju)\n",
1926 	    __func__, name, nbuf, (uintmax_t) bufsize,
1927 	    ndesc, (uintmax_t) descsize);
1928 
1929 	dd->dd_name = name;
1930 	dd->dd_desc_len = nbuf * ndesc * descsize;
1931 
1932 	/*
1933 	 * Setup DMA descriptor area.
1934 	 */
1935 	error = bus_dma_tag_create(bus_get_dma_tag(sc->sc_dev),	/* parent */
1936 		       PAGE_SIZE, 0,		/* alignment, bounds */
1937 		       BUS_SPACE_MAXADDR_32BIT,	/* lowaddr */
1938 		       BUS_SPACE_MAXADDR,	/* highaddr */
1939 		       NULL, NULL,		/* filter, filterarg */
1940 		       dd->dd_desc_len,		/* maxsize */
1941 		       1,			/* nsegments */
1942 		       dd->dd_desc_len,		/* maxsegsize */
1943 		       BUS_DMA_ALLOCNOW,	/* flags */
1944 		       NULL,			/* lockfunc */
1945 		       NULL,			/* lockarg */
1946 		       &dd->dd_dmat);
1947 	if (error != 0) {
1948 		device_printf(sc->sc_dev, "cannot allocate %s DMA tag\n", dd->dd_name);
1949 		return error;
1950 	}
1951 
1952 	/* allocate descriptors */
1953 	error = bus_dmamem_alloc(dd->dd_dmat, (void**) &dd->dd_desc,
1954 				 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
1955 				 &dd->dd_dmamap);
1956 	if (error != 0) {
1957 		device_printf(sc->sc_dev, "unable to alloc memory for %u %s descriptors, "
1958 			"error %u\n", nbuf * ndesc, dd->dd_name, error);
1959 		goto fail1;
1960 	}
1961 
1962 	error = bus_dmamap_load(dd->dd_dmat, dd->dd_dmamap,
1963 				dd->dd_desc, dd->dd_desc_len,
1964 				mwl_load_cb, &dd->dd_desc_paddr,
1965 				BUS_DMA_NOWAIT);
1966 	if (error != 0) {
1967 		device_printf(sc->sc_dev, "unable to map %s descriptors, error %u\n",
1968 			dd->dd_name, error);
1969 		goto fail2;
1970 	}
1971 
1972 	ds = dd->dd_desc;
1973 	memset(ds, 0, dd->dd_desc_len);
1974 	DPRINTF(sc, MWL_DEBUG_RESET,
1975 	    "%s: %s DMA map: %p (%lu) -> 0x%jx (%lu)\n",
1976 	    __func__, dd->dd_name, ds, (u_long) dd->dd_desc_len,
1977 	    (uintmax_t) dd->dd_desc_paddr, /*XXX*/ (u_long) dd->dd_desc_len);
1978 
1979 	return 0;
1980 fail2:
1981 	bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
1982 fail1:
1983 	bus_dma_tag_destroy(dd->dd_dmat);
1984 	memset(dd, 0, sizeof(*dd));
1985 	return error;
1986 #undef DS2PHYS
1987 }
1988 
1989 static void
1990 mwl_desc_cleanup(struct mwl_softc *sc, struct mwl_descdma *dd)
1991 {
1992 	bus_dmamap_unload(dd->dd_dmat, dd->dd_dmamap);
1993 	bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
1994 	bus_dma_tag_destroy(dd->dd_dmat);
1995 
1996 	memset(dd, 0, sizeof(*dd));
1997 }
1998 
1999 /*
2000  * Construct a tx q's free list.  The order of entries on
2001  * the list must reflect the physical layout of tx descriptors
2002  * because the firmware pre-fetches descriptors.
2003  *
2004  * XXX might be better to use indices into the buffer array.
2005  */
2006 static void
2007 mwl_txq_reset(struct mwl_softc *sc, struct mwl_txq *txq)
2008 {
2009 	struct mwl_txbuf *bf;
2010 	int i;
2011 
2012 	bf = txq->dma.dd_bufptr;
2013 	STAILQ_INIT(&txq->free);
2014 	for (i = 0; i < mwl_txbuf; i++, bf++)
2015 		STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
2016 	txq->nfree = i;
2017 }
2018 
2019 #define	DS2PHYS(_dd, _ds) \
2020 	((_dd)->dd_desc_paddr + ((caddr_t)(_ds) - (caddr_t)(_dd)->dd_desc))
2021 
2022 static int
2023 mwl_txdma_setup(struct mwl_softc *sc, struct mwl_txq *txq)
2024 {
2025 	int error, bsize, i;
2026 	struct mwl_txbuf *bf;
2027 	struct mwl_txdesc *ds;
2028 
2029 	error = mwl_desc_setup(sc, "tx", &txq->dma,
2030 			mwl_txbuf, sizeof(struct mwl_txbuf),
2031 			MWL_TXDESC, sizeof(struct mwl_txdesc));
2032 	if (error != 0)
2033 		return error;
2034 
2035 	/* allocate and setup tx buffers */
2036 	bsize = mwl_txbuf * sizeof(struct mwl_txbuf);
2037 	bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2038 	if (bf == NULL) {
2039 		device_printf(sc->sc_dev, "malloc of %u tx buffers failed\n",
2040 			mwl_txbuf);
2041 		return ENOMEM;
2042 	}
2043 	txq->dma.dd_bufptr = bf;
2044 
2045 	ds = txq->dma.dd_desc;
2046 	for (i = 0; i < mwl_txbuf; i++, bf++, ds += MWL_TXDESC) {
2047 		bf->bf_desc = ds;
2048 		bf->bf_daddr = DS2PHYS(&txq->dma, ds);
2049 		error = bus_dmamap_create(sc->sc_dmat, BUS_DMA_NOWAIT,
2050 				&bf->bf_dmamap);
2051 		if (error != 0) {
2052 			device_printf(sc->sc_dev, "unable to create dmamap for tx "
2053 				"buffer %u, error %u\n", i, error);
2054 			return error;
2055 		}
2056 	}
2057 	mwl_txq_reset(sc, txq);
2058 	return 0;
2059 }
2060 
2061 static void
2062 mwl_txdma_cleanup(struct mwl_softc *sc, struct mwl_txq *txq)
2063 {
2064 	struct mwl_txbuf *bf;
2065 	int i;
2066 
2067 	bf = txq->dma.dd_bufptr;
2068 	for (i = 0; i < mwl_txbuf; i++, bf++) {
2069 		KASSERT(bf->bf_m == NULL, ("mbuf on free list"));
2070 		KASSERT(bf->bf_node == NULL, ("node on free list"));
2071 		if (bf->bf_dmamap != NULL)
2072 			bus_dmamap_destroy(sc->sc_dmat, bf->bf_dmamap);
2073 	}
2074 	STAILQ_INIT(&txq->free);
2075 	txq->nfree = 0;
2076 	if (txq->dma.dd_bufptr != NULL) {
2077 		free(txq->dma.dd_bufptr, M_MWLDEV);
2078 		txq->dma.dd_bufptr = NULL;
2079 	}
2080 	if (txq->dma.dd_desc_len != 0)
2081 		mwl_desc_cleanup(sc, &txq->dma);
2082 }
2083 
2084 static int
2085 mwl_rxdma_setup(struct mwl_softc *sc)
2086 {
2087 	int error, jumbosize, bsize, i;
2088 	struct mwl_rxbuf *bf;
2089 	struct mwl_jumbo *rbuf;
2090 	struct mwl_rxdesc *ds;
2091 	caddr_t data;
2092 
2093 	error = mwl_desc_setup(sc, "rx", &sc->sc_rxdma,
2094 			mwl_rxdesc, sizeof(struct mwl_rxbuf),
2095 			1, sizeof(struct mwl_rxdesc));
2096 	if (error != 0)
2097 		return error;
2098 
2099 	/*
2100 	 * Receive is done to a private pool of jumbo buffers.
2101 	 * This allows us to attach to mbuf's and avoid re-mapping
2102 	 * memory on each rx we post.  We allocate a large chunk
2103 	 * of memory and manage it in the driver.  The mbuf free
2104 	 * callback method is used to reclaim frames after sending
2105 	 * them up the stack.  By default we allocate 2x the number of
2106 	 * rx descriptors configured so we have some slop to hold
2107 	 * us while frames are processed.
2108 	 */
2109 	if (mwl_rxbuf < 2*mwl_rxdesc) {
2110 		device_printf(sc->sc_dev,
2111 		    "too few rx dma buffers (%d); increasing to %d\n",
2112 		    mwl_rxbuf, 2*mwl_rxdesc);
2113 		mwl_rxbuf = 2*mwl_rxdesc;
2114 	}
2115 	jumbosize = roundup(MWL_AGGR_SIZE, PAGE_SIZE);
2116 	sc->sc_rxmemsize = mwl_rxbuf*jumbosize;
2117 
2118 	error = bus_dma_tag_create(sc->sc_dmat,	/* parent */
2119 		       PAGE_SIZE, 0,		/* alignment, bounds */
2120 		       BUS_SPACE_MAXADDR_32BIT,	/* lowaddr */
2121 		       BUS_SPACE_MAXADDR,	/* highaddr */
2122 		       NULL, NULL,		/* filter, filterarg */
2123 		       sc->sc_rxmemsize,	/* maxsize */
2124 		       1,			/* nsegments */
2125 		       sc->sc_rxmemsize,	/* maxsegsize */
2126 		       BUS_DMA_ALLOCNOW,	/* flags */
2127 		       NULL,			/* lockfunc */
2128 		       NULL,			/* lockarg */
2129 		       &sc->sc_rxdmat);
2130 	if (error != 0) {
2131 		device_printf(sc->sc_dev, "could not create rx DMA tag\n");
2132 		return error;
2133 	}
2134 
2135 	error = bus_dmamem_alloc(sc->sc_rxdmat, (void**) &sc->sc_rxmem,
2136 				 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
2137 				 &sc->sc_rxmap);
2138 	if (error != 0) {
2139 		device_printf(sc->sc_dev, "could not alloc %ju bytes of rx DMA memory\n",
2140 		    (uintmax_t) sc->sc_rxmemsize);
2141 		return error;
2142 	}
2143 
2144 	error = bus_dmamap_load(sc->sc_rxdmat, sc->sc_rxmap,
2145 				sc->sc_rxmem, sc->sc_rxmemsize,
2146 				mwl_load_cb, &sc->sc_rxmem_paddr,
2147 				BUS_DMA_NOWAIT);
2148 	if (error != 0) {
2149 		device_printf(sc->sc_dev, "could not load rx DMA map\n");
2150 		return error;
2151 	}
2152 
2153 	/*
2154 	 * Allocate rx buffers and set them up.
2155 	 */
2156 	bsize = mwl_rxdesc * sizeof(struct mwl_rxbuf);
2157 	bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2158 	if (bf == NULL) {
2159 		device_printf(sc->sc_dev, "malloc of %u rx buffers failed\n", bsize);
2160 		return error;
2161 	}
2162 	sc->sc_rxdma.dd_bufptr = bf;
2163 
2164 	STAILQ_INIT(&sc->sc_rxbuf);
2165 	ds = sc->sc_rxdma.dd_desc;
2166 	for (i = 0; i < mwl_rxdesc; i++, bf++, ds++) {
2167 		bf->bf_desc = ds;
2168 		bf->bf_daddr = DS2PHYS(&sc->sc_rxdma, ds);
2169 		/* pre-assign dma buffer */
2170 		bf->bf_data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2171 		/* NB: tail is intentional to preserve descriptor order */
2172 		STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list);
2173 	}
2174 
2175 	/*
2176 	 * Place remainder of dma memory buffers on the free list.
2177 	 */
2178 	SLIST_INIT(&sc->sc_rxfree);
2179 	for (; i < mwl_rxbuf; i++) {
2180 		data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2181 		rbuf = MWL_JUMBO_DATA2BUF(data);
2182 		SLIST_INSERT_HEAD(&sc->sc_rxfree, rbuf, next);
2183 		sc->sc_nrxfree++;
2184 	}
2185 	return 0;
2186 }
2187 #undef DS2PHYS
2188 
2189 static void
2190 mwl_rxdma_cleanup(struct mwl_softc *sc)
2191 {
2192 	if (sc->sc_rxmem_paddr != 0) {
2193 		bus_dmamap_unload(sc->sc_rxdmat, sc->sc_rxmap);
2194 		sc->sc_rxmem_paddr = 0;
2195 	}
2196 	if (sc->sc_rxmem != NULL) {
2197 		bus_dmamem_free(sc->sc_rxdmat, sc->sc_rxmem, sc->sc_rxmap);
2198 		sc->sc_rxmem = NULL;
2199 	}
2200 	if (sc->sc_rxdma.dd_bufptr != NULL) {
2201 		free(sc->sc_rxdma.dd_bufptr, M_MWLDEV);
2202 		sc->sc_rxdma.dd_bufptr = NULL;
2203 	}
2204 	if (sc->sc_rxdma.dd_desc_len != 0)
2205 		mwl_desc_cleanup(sc, &sc->sc_rxdma);
2206 }
2207 
2208 static int
2209 mwl_dma_setup(struct mwl_softc *sc)
2210 {
2211 	int error, i;
2212 
2213 	error = mwl_rxdma_setup(sc);
2214 	if (error != 0) {
2215 		mwl_rxdma_cleanup(sc);
2216 		return error;
2217 	}
2218 
2219 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
2220 		error = mwl_txdma_setup(sc, &sc->sc_txq[i]);
2221 		if (error != 0) {
2222 			mwl_dma_cleanup(sc);
2223 			return error;
2224 		}
2225 	}
2226 	return 0;
2227 }
2228 
2229 static void
2230 mwl_dma_cleanup(struct mwl_softc *sc)
2231 {
2232 	int i;
2233 
2234 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2235 		mwl_txdma_cleanup(sc, &sc->sc_txq[i]);
2236 	mwl_rxdma_cleanup(sc);
2237 }
2238 
2239 static struct ieee80211_node *
2240 mwl_node_alloc(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
2241 {
2242 	struct ieee80211com *ic = vap->iv_ic;
2243 	struct mwl_softc *sc = ic->ic_softc;
2244 	const size_t space = sizeof(struct mwl_node);
2245 	struct mwl_node *mn;
2246 
2247 	mn = malloc(space, M_80211_NODE, M_NOWAIT|M_ZERO);
2248 	if (mn == NULL) {
2249 		/* XXX stat+msg */
2250 		return NULL;
2251 	}
2252 	DPRINTF(sc, MWL_DEBUG_NODE, "%s: mn %p\n", __func__, mn);
2253 	return &mn->mn_node;
2254 }
2255 
2256 static void
2257 mwl_node_cleanup(struct ieee80211_node *ni)
2258 {
2259 	struct ieee80211com *ic = ni->ni_ic;
2260         struct mwl_softc *sc = ic->ic_softc;
2261 	struct mwl_node *mn = MWL_NODE(ni);
2262 
2263 	DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p ic %p staid %d\n",
2264 	    __func__, ni, ni->ni_ic, mn->mn_staid);
2265 
2266 	if (mn->mn_staid != 0) {
2267 		struct ieee80211vap *vap = ni->ni_vap;
2268 
2269 		if (mn->mn_hvap != NULL) {
2270 			if (vap->iv_opmode == IEEE80211_M_STA)
2271 				mwl_hal_delstation(mn->mn_hvap, vap->iv_myaddr);
2272 			else
2273 				mwl_hal_delstation(mn->mn_hvap, ni->ni_macaddr);
2274 		}
2275 		/*
2276 		 * NB: legacy WDS peer sta db entry is installed using
2277 		 * the associate ap's hvap; use it again to delete it.
2278 		 * XXX can vap be NULL?
2279 		 */
2280 		else if (vap->iv_opmode == IEEE80211_M_WDS &&
2281 		    MWL_VAP(vap)->mv_ap_hvap != NULL)
2282 			mwl_hal_delstation(MWL_VAP(vap)->mv_ap_hvap,
2283 			    ni->ni_macaddr);
2284 		delstaid(sc, mn->mn_staid);
2285 		mn->mn_staid = 0;
2286 	}
2287 	sc->sc_node_cleanup(ni);
2288 }
2289 
2290 /*
2291  * Reclaim rx dma buffers from packets sitting on the ampdu
2292  * reorder queue for a station.  We replace buffers with a
2293  * system cluster (if available).
2294  */
2295 static void
2296 mwl_ampdu_rxdma_reclaim(struct ieee80211_rx_ampdu *rap)
2297 {
2298 #if 0
2299 	int i, n, off;
2300 	struct mbuf *m;
2301 	void *cl;
2302 
2303 	n = rap->rxa_qframes;
2304 	for (i = 0; i < rap->rxa_wnd && n > 0; i++) {
2305 		m = rap->rxa_m[i];
2306 		if (m == NULL)
2307 			continue;
2308 		n--;
2309 		/* our dma buffers have a well-known free routine */
2310 		if ((m->m_flags & M_EXT) == 0 ||
2311 		    m->m_ext.ext_free != mwl_ext_free)
2312 			continue;
2313 		/*
2314 		 * Try to allocate a cluster and move the data.
2315 		 */
2316 		off = m->m_data - m->m_ext.ext_buf;
2317 		if (off + m->m_pkthdr.len > MCLBYTES) {
2318 			/* XXX no AMSDU for now */
2319 			continue;
2320 		}
2321 		cl = pool_cache_get_paddr(&mclpool_cache, 0,
2322 		    &m->m_ext.ext_paddr);
2323 		if (cl != NULL) {
2324 			/*
2325 			 * Copy the existing data to the cluster, remove
2326 			 * the rx dma buffer, and attach the cluster in
2327 			 * its place.  Note we preserve the offset to the
2328 			 * data so frames being bridged can still prepend
2329 			 * their headers without adding another mbuf.
2330 			 */
2331 			memcpy((caddr_t) cl + off, m->m_data, m->m_pkthdr.len);
2332 			MEXTREMOVE(m);
2333 			MEXTADD(m, cl, MCLBYTES, 0, NULL, &mclpool_cache);
2334 			/* setup mbuf like _MCLGET does */
2335 			m->m_flags |= M_CLUSTER | M_EXT_RW;
2336 			_MOWNERREF(m, M_EXT | M_CLUSTER);
2337 			/* NB: m_data is clobbered by MEXTADDR, adjust */
2338 			m->m_data += off;
2339 		}
2340 	}
2341 #endif
2342 }
2343 
2344 /*
2345  * Callback to reclaim resources.  We first let the
2346  * net80211 layer do it's thing, then if we are still
2347  * blocked by a lack of rx dma buffers we walk the ampdu
2348  * reorder q's to reclaim buffers by copying to a system
2349  * cluster.
2350  */
2351 static void
2352 mwl_node_drain(struct ieee80211_node *ni)
2353 {
2354 	struct ieee80211com *ic = ni->ni_ic;
2355         struct mwl_softc *sc = ic->ic_softc;
2356 	struct mwl_node *mn = MWL_NODE(ni);
2357 
2358 	DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p vap %p staid %d\n",
2359 	    __func__, ni, ni->ni_vap, mn->mn_staid);
2360 
2361 	/* NB: call up first to age out ampdu q's */
2362 	sc->sc_node_drain(ni);
2363 
2364 	/* XXX better to not check low water mark? */
2365 	if (sc->sc_rxblocked && mn->mn_staid != 0 &&
2366 	    (ni->ni_flags & IEEE80211_NODE_HT)) {
2367 		uint8_t tid;
2368 		/*
2369 		 * Walk the reorder q and reclaim rx dma buffers by copying
2370 		 * the packet contents into clusters.
2371 		 */
2372 		for (tid = 0; tid < WME_NUM_TID; tid++) {
2373 			struct ieee80211_rx_ampdu *rap;
2374 
2375 			rap = &ni->ni_rx_ampdu[tid];
2376 			if ((rap->rxa_flags & IEEE80211_AGGR_XCHGPEND) == 0)
2377 				continue;
2378 			if (rap->rxa_qframes)
2379 				mwl_ampdu_rxdma_reclaim(rap);
2380 		}
2381 	}
2382 }
2383 
2384 static void
2385 mwl_node_getsignal(const struct ieee80211_node *ni, int8_t *rssi, int8_t *noise)
2386 {
2387 	*rssi = ni->ni_ic->ic_node_getrssi(ni);
2388 #ifdef MWL_ANT_INFO_SUPPORT
2389 #if 0
2390 	/* XXX need to smooth data */
2391 	*noise = -MWL_NODE_CONST(ni)->mn_ai.nf;
2392 #else
2393 	*noise = -95;		/* XXX */
2394 #endif
2395 #else
2396 	*noise = -95;		/* XXX */
2397 #endif
2398 }
2399 
2400 /*
2401  * Convert Hardware per-antenna rssi info to common format:
2402  * Let a1, a2, a3 represent the amplitudes per chain
2403  * Let amax represent max[a1, a2, a3]
2404  * Rssi1_dBm = RSSI_dBm + 20*log10(a1/amax)
2405  * Rssi1_dBm = RSSI_dBm + 20*log10(a1) - 20*log10(amax)
2406  * We store a table that is 4*20*log10(idx) - the extra 4 is to store or
2407  * maintain some extra precision.
2408  *
2409  * Values are stored in .5 db format capped at 127.
2410  */
2411 static void
2412 mwl_node_getmimoinfo(const struct ieee80211_node *ni,
2413 	struct ieee80211_mimo_info *mi)
2414 {
2415 #define	CVT(_dst, _src) do {						\
2416 	(_dst) = rssi + ((logdbtbl[_src] - logdbtbl[rssi_max]) >> 2);	\
2417 	(_dst) = (_dst) > 64 ? 127 : ((_dst) << 1);			\
2418 } while (0)
2419 	static const int8_t logdbtbl[32] = {
2420 	       0,   0,  24,  38,  48,  56,  62,  68,
2421 	      72,  76,  80,  83,  86,  89,  92,  94,
2422 	      96,  98, 100, 102, 104, 106, 107, 109,
2423 	     110, 112, 113, 115, 116, 117, 118, 119
2424 	};
2425 	const struct mwl_node *mn = MWL_NODE_CONST(ni);
2426 	uint8_t rssi = mn->mn_ai.rsvd1/2;		/* XXX */
2427 	uint32_t rssi_max;
2428 
2429 	rssi_max = mn->mn_ai.rssi_a;
2430 	if (mn->mn_ai.rssi_b > rssi_max)
2431 		rssi_max = mn->mn_ai.rssi_b;
2432 	if (mn->mn_ai.rssi_c > rssi_max)
2433 		rssi_max = mn->mn_ai.rssi_c;
2434 
2435 	CVT(mi->ch[0].rssi[0], mn->mn_ai.rssi_a);
2436 	CVT(mi->ch[1].rssi[0], mn->mn_ai.rssi_b);
2437 	CVT(mi->ch[2].rssi[0], mn->mn_ai.rssi_c);
2438 
2439 	mi->ch[0].noise[0] = mn->mn_ai.nf_a;
2440 	mi->ch[1].noise[0] = mn->mn_ai.nf_b;
2441 	mi->ch[2].noise[0] = mn->mn_ai.nf_c;
2442 #undef CVT
2443 }
2444 
2445 static __inline void *
2446 mwl_getrxdma(struct mwl_softc *sc)
2447 {
2448 	struct mwl_jumbo *buf;
2449 	void *data;
2450 
2451 	/*
2452 	 * Allocate from jumbo pool.
2453 	 */
2454 	MWL_RXFREE_LOCK(sc);
2455 	buf = SLIST_FIRST(&sc->sc_rxfree);
2456 	if (buf == NULL) {
2457 		DPRINTF(sc, MWL_DEBUG_ANY,
2458 		    "%s: out of rx dma buffers\n", __func__);
2459 		sc->sc_stats.mst_rx_nodmabuf++;
2460 		data = NULL;
2461 	} else {
2462 		SLIST_REMOVE_HEAD(&sc->sc_rxfree, next);
2463 		sc->sc_nrxfree--;
2464 		data = MWL_JUMBO_BUF2DATA(buf);
2465 	}
2466 	MWL_RXFREE_UNLOCK(sc);
2467 	return data;
2468 }
2469 
2470 static __inline void
2471 mwl_putrxdma(struct mwl_softc *sc, void *data)
2472 {
2473 	struct mwl_jumbo *buf;
2474 
2475 	/* XXX bounds check data */
2476 	MWL_RXFREE_LOCK(sc);
2477 	buf = MWL_JUMBO_DATA2BUF(data);
2478 	SLIST_INSERT_HEAD(&sc->sc_rxfree, buf, next);
2479 	sc->sc_nrxfree++;
2480 	MWL_RXFREE_UNLOCK(sc);
2481 }
2482 
2483 static int
2484 mwl_rxbuf_init(struct mwl_softc *sc, struct mwl_rxbuf *bf)
2485 {
2486 	struct mwl_rxdesc *ds;
2487 
2488 	ds = bf->bf_desc;
2489 	if (bf->bf_data == NULL) {
2490 		bf->bf_data = mwl_getrxdma(sc);
2491 		if (bf->bf_data == NULL) {
2492 			/* mark descriptor to be skipped */
2493 			ds->RxControl = EAGLE_RXD_CTRL_OS_OWN;
2494 			/* NB: don't need PREREAD */
2495 			MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREWRITE);
2496 			sc->sc_stats.mst_rxbuf_failed++;
2497 			return ENOMEM;
2498 		}
2499 	}
2500 	/*
2501 	 * NB: DMA buffer contents is known to be unmodified
2502 	 *     so there's no need to flush the data cache.
2503 	 */
2504 
2505 	/*
2506 	 * Setup descriptor.
2507 	 */
2508 	ds->QosCtrl = 0;
2509 	ds->RSSI = 0;
2510 	ds->Status = EAGLE_RXD_STATUS_IDLE;
2511 	ds->Channel = 0;
2512 	ds->PktLen = htole16(MWL_AGGR_SIZE);
2513 	ds->SQ2 = 0;
2514 	ds->pPhysBuffData = htole32(MWL_JUMBO_DMA_ADDR(sc, bf->bf_data));
2515 	/* NB: don't touch pPhysNext, set once */
2516 	ds->RxControl = EAGLE_RXD_CTRL_DRIVER_OWN;
2517 	MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
2518 
2519 	return 0;
2520 }
2521 
2522 static void
2523 mwl_ext_free(struct mbuf *m)
2524 {
2525 	struct mwl_softc *sc = m->m_ext.ext_arg1;
2526 
2527 	/* XXX bounds check data */
2528 	mwl_putrxdma(sc, m->m_ext.ext_buf);
2529 	/*
2530 	 * If we were previously blocked by a lack of rx dma buffers
2531 	 * check if we now have enough to restart rx interrupt handling.
2532 	 * NB: we know we are called at splvm which is above splnet.
2533 	 */
2534 	if (sc->sc_rxblocked && sc->sc_nrxfree > mwl_rxdmalow) {
2535 		sc->sc_rxblocked = 0;
2536 		mwl_hal_intrset(sc->sc_mh, sc->sc_imask);
2537 	}
2538 }
2539 
2540 struct mwl_frame_bar {
2541 	u_int8_t	i_fc[2];
2542 	u_int8_t	i_dur[2];
2543 	u_int8_t	i_ra[IEEE80211_ADDR_LEN];
2544 	u_int8_t	i_ta[IEEE80211_ADDR_LEN];
2545 	/* ctl, seq, FCS */
2546 } __packed;
2547 
2548 /*
2549  * Like ieee80211_anyhdrsize, but handles BAR frames
2550  * specially so the logic below to piece the 802.11
2551  * header together works.
2552  */
2553 static __inline int
2554 mwl_anyhdrsize(const void *data)
2555 {
2556 	const struct ieee80211_frame *wh = data;
2557 
2558 	if ((wh->i_fc[0]&IEEE80211_FC0_TYPE_MASK) == IEEE80211_FC0_TYPE_CTL) {
2559 		switch (wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK) {
2560 		case IEEE80211_FC0_SUBTYPE_CTS:
2561 		case IEEE80211_FC0_SUBTYPE_ACK:
2562 			return sizeof(struct ieee80211_frame_ack);
2563 		case IEEE80211_FC0_SUBTYPE_BAR:
2564 			return sizeof(struct mwl_frame_bar);
2565 		}
2566 		return sizeof(struct ieee80211_frame_min);
2567 	} else
2568 		return ieee80211_hdrsize(data);
2569 }
2570 
2571 static void
2572 mwl_handlemicerror(struct ieee80211com *ic, const uint8_t *data)
2573 {
2574 	const struct ieee80211_frame *wh;
2575 	struct ieee80211_node *ni;
2576 
2577 	wh = (const struct ieee80211_frame *)(data + sizeof(uint16_t));
2578 	ni = ieee80211_find_rxnode(ic, (const struct ieee80211_frame_min *) wh);
2579 	if (ni != NULL) {
2580 		ieee80211_notify_michael_failure(ni->ni_vap, wh, 0);
2581 		ieee80211_free_node(ni);
2582 	}
2583 }
2584 
2585 /*
2586  * Convert hardware signal strength to rssi.  The value
2587  * provided by the device has the noise floor added in;
2588  * we need to compensate for this but we don't have that
2589  * so we use a fixed value.
2590  *
2591  * The offset of 8 is good for both 2.4 and 5GHz.  The LNA
2592  * offset is already set as part of the initial gain.  This
2593  * will give at least +/- 3dB for 2.4GHz and +/- 5dB for 5GHz.
2594  */
2595 static __inline int
2596 cvtrssi(uint8_t ssi)
2597 {
2598 	int rssi = (int) ssi + 8;
2599 	/* XXX hack guess until we have a real noise floor */
2600 	rssi = 2*(87 - rssi);	/* NB: .5 dBm units */
2601 	return (rssi < 0 ? 0 : rssi > 127 ? 127 : rssi);
2602 }
2603 
2604 static void
2605 mwl_rx_proc(void *arg, int npending)
2606 {
2607 	struct epoch_tracker et;
2608 	struct mwl_softc *sc = arg;
2609 	struct ieee80211com *ic = &sc->sc_ic;
2610 	struct mwl_rxbuf *bf;
2611 	struct mwl_rxdesc *ds;
2612 	struct mbuf *m;
2613 	struct ieee80211_qosframe *wh;
2614 	struct ieee80211_node *ni;
2615 	struct mwl_node *mn;
2616 	int off, len, hdrlen, pktlen, rssi, ntodo;
2617 	uint8_t *data, status;
2618 	void *newdata;
2619 	int16_t nf;
2620 
2621 	DPRINTF(sc, MWL_DEBUG_RX_PROC, "%s: pending %u rdptr 0x%x wrptr 0x%x\n",
2622 	    __func__, npending, RD4(sc, sc->sc_hwspecs.rxDescRead),
2623 	    RD4(sc, sc->sc_hwspecs.rxDescWrite));
2624 	nf = -96;			/* XXX */
2625 	bf = sc->sc_rxnext;
2626 	for (ntodo = mwl_rxquota; ntodo > 0; ntodo--) {
2627 		if (bf == NULL)
2628 			bf = STAILQ_FIRST(&sc->sc_rxbuf);
2629 		ds = bf->bf_desc;
2630 		data = bf->bf_data;
2631 		if (data == NULL) {
2632 			/*
2633 			 * If data allocation failed previously there
2634 			 * will be no buffer; try again to re-populate it.
2635 			 * Note the firmware will not advance to the next
2636 			 * descriptor with a dma buffer so we must mimic
2637 			 * this or we'll get out of sync.
2638 			 */
2639 			DPRINTF(sc, MWL_DEBUG_ANY,
2640 			    "%s: rx buf w/o dma memory\n", __func__);
2641 			(void) mwl_rxbuf_init(sc, bf);
2642 			sc->sc_stats.mst_rx_dmabufmissing++;
2643 			break;
2644 		}
2645 		MWL_RXDESC_SYNC(sc, ds,
2646 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
2647 		if (ds->RxControl != EAGLE_RXD_CTRL_DMA_OWN)
2648 			break;
2649 #ifdef MWL_DEBUG
2650 		if (sc->sc_debug & MWL_DEBUG_RECV_DESC)
2651 			mwl_printrxbuf(bf, 0);
2652 #endif
2653 		status = ds->Status;
2654 		if (status & EAGLE_RXD_STATUS_DECRYPT_ERR_MASK) {
2655 			counter_u64_add(ic->ic_ierrors, 1);
2656 			sc->sc_stats.mst_rx_crypto++;
2657 			/*
2658 			 * NB: Check EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR
2659 			 *     for backwards compatibility.
2660 			 */
2661 			if (status != EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR &&
2662 			    (status & EAGLE_RXD_STATUS_TKIP_MIC_DECRYPT_ERR)) {
2663 				/*
2664 				 * MIC error, notify upper layers.
2665 				 */
2666 				bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap,
2667 				    BUS_DMASYNC_POSTREAD);
2668 				mwl_handlemicerror(ic, data);
2669 				sc->sc_stats.mst_rx_tkipmic++;
2670 			}
2671 			/* XXX too painful to tap packets */
2672 			goto rx_next;
2673 		}
2674 		/*
2675 		 * Sync the data buffer.
2676 		 */
2677 		len = le16toh(ds->PktLen);
2678 		bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap, BUS_DMASYNC_POSTREAD);
2679 		/*
2680 		 * The 802.11 header is provided all or in part at the front;
2681 		 * use it to calculate the true size of the header that we'll
2682 		 * construct below.  We use this to figure out where to copy
2683 		 * payload prior to constructing the header.
2684 		 */
2685 		hdrlen = mwl_anyhdrsize(data + sizeof(uint16_t));
2686 		off = sizeof(uint16_t) + sizeof(struct ieee80211_frame_addr4);
2687 
2688 		/* calculate rssi early so we can re-use for each aggregate */
2689 		rssi = cvtrssi(ds->RSSI);
2690 
2691 		pktlen = hdrlen + (len - off);
2692 		/*
2693 		 * NB: we know our frame is at least as large as
2694 		 * IEEE80211_MIN_LEN because there is a 4-address
2695 		 * frame at the front.  Hence there's no need to
2696 		 * vet the packet length.  If the frame in fact
2697 		 * is too small it should be discarded at the
2698 		 * net80211 layer.
2699 		 */
2700 
2701 		/*
2702 		 * Attach dma buffer to an mbuf.  We tried
2703 		 * doing this based on the packet size (i.e.
2704 		 * copying small packets) but it turns out to
2705 		 * be a net loss.  The tradeoff might be system
2706 		 * dependent (cache architecture is important).
2707 		 */
2708 		MGETHDR(m, M_NOWAIT, MT_DATA);
2709 		if (m == NULL) {
2710 			DPRINTF(sc, MWL_DEBUG_ANY,
2711 			    "%s: no rx mbuf\n", __func__);
2712 			sc->sc_stats.mst_rx_nombuf++;
2713 			goto rx_next;
2714 		}
2715 		/*
2716 		 * Acquire the replacement dma buffer before
2717 		 * processing the frame.  If we're out of dma
2718 		 * buffers we disable rx interrupts and wait
2719 		 * for the free pool to reach mlw_rxdmalow buffers
2720 		 * before starting to do work again.  If the firmware
2721 		 * runs out of descriptors then it will toss frames
2722 		 * which is better than our doing it as that can
2723 		 * starve our processing.  It is also important that
2724 		 * we always process rx'd frames in case they are
2725 		 * A-MPDU as otherwise the host's view of the BA
2726 		 * window may get out of sync with the firmware.
2727 		 */
2728 		newdata = mwl_getrxdma(sc);
2729 		if (newdata == NULL) {
2730 			/* NB: stat+msg in mwl_getrxdma */
2731 			m_free(m);
2732 			/* disable RX interrupt and mark state */
2733 			mwl_hal_intrset(sc->sc_mh,
2734 			    sc->sc_imask &~ MACREG_A2HRIC_BIT_RX_RDY);
2735 			sc->sc_rxblocked = 1;
2736 			ieee80211_drain(ic);
2737 			/* XXX check rxblocked and immediately start again? */
2738 			goto rx_stop;
2739 		}
2740 		bf->bf_data = newdata;
2741 		/*
2742 		 * Attach the dma buffer to the mbuf;
2743 		 * mwl_rxbuf_init will re-setup the rx
2744 		 * descriptor using the replacement dma
2745 		 * buffer we just installed above.
2746 		 */
2747 		m_extadd(m, data, MWL_AGGR_SIZE, mwl_ext_free, sc, NULL, 0,
2748 		    EXT_NET_DRV);
2749 		m->m_data += off - hdrlen;
2750 		m->m_pkthdr.len = m->m_len = pktlen;
2751 		/* NB: dma buffer assumed read-only */
2752 
2753 		/*
2754 		 * Piece 802.11 header together.
2755 		 */
2756 		wh = mtod(m, struct ieee80211_qosframe *);
2757 		/* NB: don't need to do this sometimes but ... */
2758 		/* XXX special case so we can memcpy after m_devget? */
2759 		ovbcopy(data + sizeof(uint16_t), wh, hdrlen);
2760 		if (IEEE80211_QOS_HAS_SEQ(wh))
2761 			*(uint16_t *)ieee80211_getqos(wh) = ds->QosCtrl;
2762 		/*
2763 		 * The f/w strips WEP header but doesn't clear
2764 		 * the WEP bit; mark the packet with M_WEP so
2765 		 * net80211 will treat the data as decrypted.
2766 		 * While here also clear the PWR_MGT bit since
2767 		 * power save is handled by the firmware and
2768 		 * passing this up will potentially cause the
2769 		 * upper layer to put a station in power save
2770 		 * (except when configured with MWL_HOST_PS_SUPPORT).
2771 		 */
2772 		if (wh->i_fc[1] & IEEE80211_FC1_PROTECTED)
2773 			m->m_flags |= M_WEP;
2774 #ifdef MWL_HOST_PS_SUPPORT
2775 		wh->i_fc[1] &= ~IEEE80211_FC1_PROTECTED;
2776 #else
2777 		wh->i_fc[1] &= ~(IEEE80211_FC1_PROTECTED |
2778 		    IEEE80211_FC1_PWR_MGT);
2779 #endif
2780 
2781 		if (ieee80211_radiotap_active(ic)) {
2782 			struct mwl_rx_radiotap_header *tap = &sc->sc_rx_th;
2783 
2784 			tap->wr_flags = 0;
2785 			tap->wr_rate = ds->Rate;
2786 			tap->wr_antsignal = rssi + nf;
2787 			tap->wr_antnoise = nf;
2788 		}
2789 		if (IFF_DUMPPKTS_RECV(sc, wh)) {
2790 			ieee80211_dump_pkt(ic, mtod(m, caddr_t),
2791 			    len, ds->Rate, rssi);
2792 		}
2793 		/* dispatch */
2794 		ni = ieee80211_find_rxnode(ic,
2795 		    (const struct ieee80211_frame_min *) wh);
2796 
2797 		NET_EPOCH_ENTER(et);
2798 		if (ni != NULL) {
2799 			mn = MWL_NODE(ni);
2800 #ifdef MWL_ANT_INFO_SUPPORT
2801 			mn->mn_ai.rssi_a = ds->ai.rssi_a;
2802 			mn->mn_ai.rssi_b = ds->ai.rssi_b;
2803 			mn->mn_ai.rssi_c = ds->ai.rssi_c;
2804 			mn->mn_ai.rsvd1 = rssi;
2805 #endif
2806 			/* tag AMPDU aggregates for reorder processing */
2807 			if (ni->ni_flags & IEEE80211_NODE_HT)
2808 				m->m_flags |= M_AMPDU;
2809 			(void) ieee80211_input(ni, m, rssi, nf);
2810 			ieee80211_free_node(ni);
2811 		} else
2812 			(void) ieee80211_input_all(ic, m, rssi, nf);
2813 		NET_EPOCH_EXIT(et);
2814 rx_next:
2815 		/* NB: ignore ENOMEM so we process more descriptors */
2816 		(void) mwl_rxbuf_init(sc, bf);
2817 		bf = STAILQ_NEXT(bf, bf_list);
2818 	}
2819 rx_stop:
2820 	sc->sc_rxnext = bf;
2821 
2822 	if (mbufq_first(&sc->sc_snd) != NULL) {
2823 		/* NB: kick fw; the tx thread may have been preempted */
2824 		mwl_hal_txstart(sc->sc_mh, 0);
2825 		mwl_start(sc);
2826 	}
2827 }
2828 
2829 static void
2830 mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *txq, int qnum)
2831 {
2832 	struct mwl_txbuf *bf, *bn;
2833 	struct mwl_txdesc *ds;
2834 
2835 	MWL_TXQ_LOCK_INIT(sc, txq);
2836 	txq->qnum = qnum;
2837 	txq->txpri = 0;	/* XXX */
2838 #if 0
2839 	/* NB: q setup by mwl_txdma_setup XXX */
2840 	STAILQ_INIT(&txq->free);
2841 #endif
2842 	STAILQ_FOREACH(bf, &txq->free, bf_list) {
2843 		bf->bf_txq = txq;
2844 
2845 		ds = bf->bf_desc;
2846 		bn = STAILQ_NEXT(bf, bf_list);
2847 		if (bn == NULL)
2848 			bn = STAILQ_FIRST(&txq->free);
2849 		ds->pPhysNext = htole32(bn->bf_daddr);
2850 	}
2851 	STAILQ_INIT(&txq->active);
2852 }
2853 
2854 /*
2855  * Setup a hardware data transmit queue for the specified
2856  * access control.  We record the mapping from ac's
2857  * to h/w queues for use by mwl_tx_start.
2858  */
2859 static int
2860 mwl_tx_setup(struct mwl_softc *sc, int ac, int mvtype)
2861 {
2862 	struct mwl_txq *txq;
2863 
2864 	if (ac >= nitems(sc->sc_ac2q)) {
2865 		device_printf(sc->sc_dev, "AC %u out of range, max %zu!\n",
2866 			ac, nitems(sc->sc_ac2q));
2867 		return 0;
2868 	}
2869 	if (mvtype >= MWL_NUM_TX_QUEUES) {
2870 		device_printf(sc->sc_dev, "mvtype %u out of range, max %u!\n",
2871 			mvtype, MWL_NUM_TX_QUEUES);
2872 		return 0;
2873 	}
2874 	txq = &sc->sc_txq[mvtype];
2875 	mwl_txq_init(sc, txq, mvtype);
2876 	sc->sc_ac2q[ac] = txq;
2877 	return 1;
2878 }
2879 
2880 /*
2881  * Update WME parameters for a transmit queue.
2882  */
2883 static int
2884 mwl_txq_update(struct mwl_softc *sc, int ac)
2885 {
2886 #define	MWL_EXPONENT_TO_VALUE(v)	((1<<v)-1)
2887 	struct ieee80211com *ic = &sc->sc_ic;
2888 	struct chanAccParams chp;
2889 	struct mwl_txq *txq = sc->sc_ac2q[ac];
2890 	struct wmeParams *wmep;
2891 	struct mwl_hal *mh = sc->sc_mh;
2892 	int aifs, cwmin, cwmax, txoplim;
2893 
2894 	ieee80211_wme_ic_getparams(ic, &chp);
2895 	wmep = &chp.cap_wmeParams[ac];
2896 
2897 	aifs = wmep->wmep_aifsn;
2898 	/* XXX in sta mode need to pass log values for cwmin/max */
2899 	cwmin = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmin);
2900 	cwmax = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmax);
2901 	txoplim = wmep->wmep_txopLimit;		/* NB: units of 32us */
2902 
2903 	if (mwl_hal_setedcaparams(mh, txq->qnum, cwmin, cwmax, aifs, txoplim)) {
2904 		device_printf(sc->sc_dev, "unable to update hardware queue "
2905 			"parameters for %s traffic!\n",
2906 			ieee80211_wme_acnames[ac]);
2907 		return 0;
2908 	}
2909 	return 1;
2910 #undef MWL_EXPONENT_TO_VALUE
2911 }
2912 
2913 /*
2914  * Callback from the 802.11 layer to update WME parameters.
2915  */
2916 static int
2917 mwl_wme_update(struct ieee80211com *ic)
2918 {
2919 	struct mwl_softc *sc = ic->ic_softc;
2920 
2921 	return !mwl_txq_update(sc, WME_AC_BE) ||
2922 	    !mwl_txq_update(sc, WME_AC_BK) ||
2923 	    !mwl_txq_update(sc, WME_AC_VI) ||
2924 	    !mwl_txq_update(sc, WME_AC_VO) ? EIO : 0;
2925 }
2926 
2927 /*
2928  * Reclaim resources for a setup queue.
2929  */
2930 static void
2931 mwl_tx_cleanupq(struct mwl_softc *sc, struct mwl_txq *txq)
2932 {
2933 	/* XXX hal work? */
2934 	MWL_TXQ_LOCK_DESTROY(txq);
2935 }
2936 
2937 /*
2938  * Reclaim all tx queue resources.
2939  */
2940 static void
2941 mwl_tx_cleanup(struct mwl_softc *sc)
2942 {
2943 	int i;
2944 
2945 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2946 		mwl_tx_cleanupq(sc, &sc->sc_txq[i]);
2947 }
2948 
2949 static int
2950 mwl_tx_dmasetup(struct mwl_softc *sc, struct mwl_txbuf *bf, struct mbuf *m0)
2951 {
2952 	struct mbuf *m;
2953 	int error;
2954 
2955 	/*
2956 	 * Load the DMA map so any coalescing is done.  This
2957 	 * also calculates the number of descriptors we need.
2958 	 */
2959 	error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
2960 				     bf->bf_segs, &bf->bf_nseg,
2961 				     BUS_DMA_NOWAIT);
2962 	if (error == EFBIG) {
2963 		/* XXX packet requires too many descriptors */
2964 		bf->bf_nseg = MWL_TXDESC+1;
2965 	} else if (error != 0) {
2966 		sc->sc_stats.mst_tx_busdma++;
2967 		m_freem(m0);
2968 		return error;
2969 	}
2970 	/*
2971 	 * Discard null packets and check for packets that
2972 	 * require too many TX descriptors.  We try to convert
2973 	 * the latter to a cluster.
2974 	 */
2975 	if (error == EFBIG) {		/* too many desc's, linearize */
2976 		sc->sc_stats.mst_tx_linear++;
2977 #if MWL_TXDESC > 1
2978 		m = m_collapse(m0, M_NOWAIT, MWL_TXDESC);
2979 #else
2980 		m = m_defrag(m0, M_NOWAIT);
2981 #endif
2982 		if (m == NULL) {
2983 			m_freem(m0);
2984 			sc->sc_stats.mst_tx_nombuf++;
2985 			return ENOMEM;
2986 		}
2987 		m0 = m;
2988 		error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
2989 					     bf->bf_segs, &bf->bf_nseg,
2990 					     BUS_DMA_NOWAIT);
2991 		if (error != 0) {
2992 			sc->sc_stats.mst_tx_busdma++;
2993 			m_freem(m0);
2994 			return error;
2995 		}
2996 		KASSERT(bf->bf_nseg <= MWL_TXDESC,
2997 		    ("too many segments after defrag; nseg %u", bf->bf_nseg));
2998 	} else if (bf->bf_nseg == 0) {		/* null packet, discard */
2999 		sc->sc_stats.mst_tx_nodata++;
3000 		m_freem(m0);
3001 		return EIO;
3002 	}
3003 	DPRINTF(sc, MWL_DEBUG_XMIT, "%s: m %p len %u\n",
3004 		__func__, m0, m0->m_pkthdr.len);
3005 	bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap, BUS_DMASYNC_PREWRITE);
3006 	bf->bf_m = m0;
3007 
3008 	return 0;
3009 }
3010 
3011 static __inline int
3012 mwl_cvtlegacyrate(int rate)
3013 {
3014 	switch (rate) {
3015 	case 2:	 return 0;
3016 	case 4:	 return 1;
3017 	case 11: return 2;
3018 	case 22: return 3;
3019 	case 44: return 4;
3020 	case 12: return 5;
3021 	case 18: return 6;
3022 	case 24: return 7;
3023 	case 36: return 8;
3024 	case 48: return 9;
3025 	case 72: return 10;
3026 	case 96: return 11;
3027 	case 108:return 12;
3028 	}
3029 	return 0;
3030 }
3031 
3032 /*
3033  * Calculate fixed tx rate information per client state;
3034  * this value is suitable for writing to the Format field
3035  * of a tx descriptor.
3036  */
3037 static uint16_t
3038 mwl_calcformat(uint8_t rate, const struct ieee80211_node *ni)
3039 {
3040 	uint16_t fmt;
3041 
3042 	fmt = _IEEE80211_SHIFTMASK(3, EAGLE_TXD_ANTENNA)
3043 	    | (IEEE80211_IS_CHAN_HT40D(ni->ni_chan) ?
3044 		EAGLE_TXD_EXTCHAN_LO : EAGLE_TXD_EXTCHAN_HI);
3045 	if (rate & IEEE80211_RATE_MCS) {	/* HT MCS */
3046 		fmt |= EAGLE_TXD_FORMAT_HT
3047 		    /* NB: 0x80 implicitly stripped from ucastrate */
3048 		    | _IEEE80211_SHIFTMASK(rate, EAGLE_TXD_RATE);
3049 		/* XXX short/long GI may be wrong; re-check */
3050 		if (IEEE80211_IS_CHAN_HT40(ni->ni_chan)) {
3051 			fmt |= EAGLE_TXD_CHW_40
3052 			    | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI40 ?
3053 			        EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3054 		} else {
3055 			fmt |= EAGLE_TXD_CHW_20
3056 			    | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI20 ?
3057 			        EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3058 		}
3059 	} else {			/* legacy rate */
3060 		fmt |= EAGLE_TXD_FORMAT_LEGACY
3061 		    | _IEEE80211_SHIFTMASK(mwl_cvtlegacyrate(rate),
3062 			EAGLE_TXD_RATE)
3063 		    | EAGLE_TXD_CHW_20
3064 		    /* XXX iv_flags & IEEE80211_F_SHPREAMBLE? */
3065 		    | (ni->ni_capinfo & IEEE80211_CAPINFO_SHORT_PREAMBLE ?
3066 			EAGLE_TXD_PREAMBLE_SHORT : EAGLE_TXD_PREAMBLE_LONG);
3067 	}
3068 	return fmt;
3069 }
3070 
3071 static int
3072 mwl_tx_start(struct mwl_softc *sc, struct ieee80211_node *ni, struct mwl_txbuf *bf,
3073     struct mbuf *m0)
3074 {
3075 	struct ieee80211com *ic = &sc->sc_ic;
3076 	struct ieee80211vap *vap = ni->ni_vap;
3077 	int error, iswep, ismcast;
3078 	int hdrlen, pktlen;
3079 	struct mwl_txdesc *ds;
3080 	struct mwl_txq *txq;
3081 	struct ieee80211_frame *wh;
3082 	struct mwltxrec *tr;
3083 	struct mwl_node *mn;
3084 	uint16_t qos;
3085 #if MWL_TXDESC > 1
3086 	int i;
3087 #endif
3088 
3089 	wh = mtod(m0, struct ieee80211_frame *);
3090 	iswep = wh->i_fc[1] & IEEE80211_FC1_PROTECTED;
3091 	ismcast = IEEE80211_IS_MULTICAST(wh->i_addr1);
3092 	hdrlen = ieee80211_anyhdrsize(wh);
3093 	pktlen = m0->m_pkthdr.len;
3094 	if (IEEE80211_QOS_HAS_SEQ(wh)) {
3095 		qos = *(uint16_t *)ieee80211_getqos(wh);
3096 	} else
3097 		qos = 0;
3098 
3099 	if (iswep) {
3100 		const struct ieee80211_cipher *cip;
3101 		struct ieee80211_key *k;
3102 
3103 		/*
3104 		 * Construct the 802.11 header+trailer for an encrypted
3105 		 * frame. The only reason this can fail is because of an
3106 		 * unknown or unsupported cipher/key type.
3107 		 *
3108 		 * NB: we do this even though the firmware will ignore
3109 		 *     what we've done for WEP and TKIP as we need the
3110 		 *     ExtIV filled in for CCMP and this also adjusts
3111 		 *     the headers which simplifies our work below.
3112 		 */
3113 		k = ieee80211_crypto_encap(ni, m0);
3114 		if (k == NULL) {
3115 			/*
3116 			 * This can happen when the key is yanked after the
3117 			 * frame was queued.  Just discard the frame; the
3118 			 * 802.11 layer counts failures and provides
3119 			 * debugging/diagnostics.
3120 			 */
3121 			m_freem(m0);
3122 			return EIO;
3123 		}
3124 		/*
3125 		 * Adjust the packet length for the crypto additions
3126 		 * done during encap and any other bits that the f/w
3127 		 * will add later on.
3128 		 */
3129 		cip = k->wk_cipher;
3130 		pktlen += cip->ic_header + cip->ic_miclen + cip->ic_trailer;
3131 
3132 		/* packet header may have moved, reset our local pointer */
3133 		wh = mtod(m0, struct ieee80211_frame *);
3134 	}
3135 
3136 	if (ieee80211_radiotap_active_vap(vap)) {
3137 		sc->sc_tx_th.wt_flags = 0;	/* XXX */
3138 		if (iswep)
3139 			sc->sc_tx_th.wt_flags |= IEEE80211_RADIOTAP_F_WEP;
3140 #if 0
3141 		sc->sc_tx_th.wt_rate = ds->DataRate;
3142 #endif
3143 		sc->sc_tx_th.wt_txpower = ni->ni_txpower;
3144 		sc->sc_tx_th.wt_antenna = sc->sc_txantenna;
3145 
3146 		ieee80211_radiotap_tx(vap, m0);
3147 	}
3148 	/*
3149 	 * Copy up/down the 802.11 header; the firmware requires
3150 	 * we present a 2-byte payload length followed by a
3151 	 * 4-address header (w/o QoS), followed (optionally) by
3152 	 * any WEP/ExtIV header (but only filled in for CCMP).
3153 	 * We are assured the mbuf has sufficient headroom to
3154 	 * prepend in-place by the setup of ic_headroom in
3155 	 * mwl_attach.
3156 	 */
3157 	if (hdrlen < sizeof(struct mwltxrec)) {
3158 		const int space = sizeof(struct mwltxrec) - hdrlen;
3159 		if (M_LEADINGSPACE(m0) < space) {
3160 			/* NB: should never happen */
3161 			device_printf(sc->sc_dev,
3162 			    "not enough headroom, need %d found %zd, "
3163 			    "m_flags 0x%x m_len %d\n",
3164 			    space, M_LEADINGSPACE(m0), m0->m_flags, m0->m_len);
3165 			ieee80211_dump_pkt(ic,
3166 			    mtod(m0, const uint8_t *), m0->m_len, 0, -1);
3167 			m_freem(m0);
3168 			sc->sc_stats.mst_tx_noheadroom++;
3169 			return EIO;
3170 		}
3171 		M_PREPEND(m0, space, M_NOWAIT);
3172 	}
3173 	tr = mtod(m0, struct mwltxrec *);
3174 	if (wh != (struct ieee80211_frame *) &tr->wh)
3175 		ovbcopy(wh, &tr->wh, hdrlen);
3176 	/*
3177 	 * Note: the "firmware length" is actually the length
3178 	 * of the fully formed "802.11 payload".  That is, it's
3179 	 * everything except for the 802.11 header.  In particular
3180 	 * this includes all crypto material including the MIC!
3181 	 */
3182 	tr->fwlen = htole16(pktlen - hdrlen);
3183 
3184 	/*
3185 	 * Load the DMA map so any coalescing is done.  This
3186 	 * also calculates the number of descriptors we need.
3187 	 */
3188 	error = mwl_tx_dmasetup(sc, bf, m0);
3189 	if (error != 0) {
3190 		/* NB: stat collected in mwl_tx_dmasetup */
3191 		DPRINTF(sc, MWL_DEBUG_XMIT,
3192 		    "%s: unable to setup dma\n", __func__);
3193 		return error;
3194 	}
3195 	bf->bf_node = ni;			/* NB: held reference */
3196 	m0 = bf->bf_m;				/* NB: may have changed */
3197 	tr = mtod(m0, struct mwltxrec *);
3198 	wh = (struct ieee80211_frame *)&tr->wh;
3199 
3200 	/*
3201 	 * Formulate tx descriptor.
3202 	 */
3203 	ds = bf->bf_desc;
3204 	txq = bf->bf_txq;
3205 
3206 	ds->QosCtrl = qos;			/* NB: already little-endian */
3207 #if MWL_TXDESC == 1
3208 	/*
3209 	 * NB: multiframes should be zero because the descriptors
3210 	 *     are initialized to zero.  This should handle the case
3211 	 *     where the driver is built with MWL_TXDESC=1 but we are
3212 	 *     using firmware with multi-segment support.
3213 	 */
3214 	ds->PktPtr = htole32(bf->bf_segs[0].ds_addr);
3215 	ds->PktLen = htole16(bf->bf_segs[0].ds_len);
3216 #else
3217 	ds->multiframes = htole32(bf->bf_nseg);
3218 	ds->PktLen = htole16(m0->m_pkthdr.len);
3219 	for (i = 0; i < bf->bf_nseg; i++) {
3220 		ds->PktPtrArray[i] = htole32(bf->bf_segs[i].ds_addr);
3221 		ds->PktLenArray[i] = htole16(bf->bf_segs[i].ds_len);
3222 	}
3223 #endif
3224 	/* NB: pPhysNext, DataRate, and SapPktInfo setup once, don't touch */
3225 	ds->Format = 0;
3226 	ds->pad = 0;
3227 	ds->ack_wcb_addr = 0;
3228 
3229 	mn = MWL_NODE(ni);
3230 	/*
3231 	 * Select transmit rate.
3232 	 */
3233 	switch (wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK) {
3234 	case IEEE80211_FC0_TYPE_MGT:
3235 		sc->sc_stats.mst_tx_mgmt++;
3236 		/* fall thru... */
3237 	case IEEE80211_FC0_TYPE_CTL:
3238 		/* NB: assign to BE q to avoid bursting */
3239 		ds->TxPriority = MWL_WME_AC_BE;
3240 		break;
3241 	case IEEE80211_FC0_TYPE_DATA:
3242 		if (!ismcast) {
3243 			const struct ieee80211_txparam *tp = ni->ni_txparms;
3244 			/*
3245 			 * EAPOL frames get forced to a fixed rate and w/o
3246 			 * aggregation; otherwise check for any fixed rate
3247 			 * for the client (may depend on association state).
3248 			 */
3249 			if (m0->m_flags & M_EAPOL) {
3250 				const struct mwl_vap *mvp = MWL_VAP_CONST(vap);
3251 				ds->Format = mvp->mv_eapolformat;
3252 				ds->pad = htole16(
3253 				    EAGLE_TXD_FIXED_RATE | EAGLE_TXD_DONT_AGGR);
3254 			} else if (tp->ucastrate != IEEE80211_FIXED_RATE_NONE) {
3255 				/* XXX pre-calculate per node */
3256 				ds->Format = htole16(
3257 				    mwl_calcformat(tp->ucastrate, ni));
3258 				ds->pad = htole16(EAGLE_TXD_FIXED_RATE);
3259 			}
3260 			/* NB: EAPOL frames will never have qos set */
3261 			if (qos == 0)
3262 				ds->TxPriority = txq->qnum;
3263 #if MWL_MAXBA > 3
3264 			else if (mwl_bastream_match(&mn->mn_ba[3], qos))
3265 				ds->TxPriority = mn->mn_ba[3].txq;
3266 #endif
3267 #if MWL_MAXBA > 2
3268 			else if (mwl_bastream_match(&mn->mn_ba[2], qos))
3269 				ds->TxPriority = mn->mn_ba[2].txq;
3270 #endif
3271 #if MWL_MAXBA > 1
3272 			else if (mwl_bastream_match(&mn->mn_ba[1], qos))
3273 				ds->TxPriority = mn->mn_ba[1].txq;
3274 #endif
3275 #if MWL_MAXBA > 0
3276 			else if (mwl_bastream_match(&mn->mn_ba[0], qos))
3277 				ds->TxPriority = mn->mn_ba[0].txq;
3278 #endif
3279 			else
3280 				ds->TxPriority = txq->qnum;
3281 		} else
3282 			ds->TxPriority = txq->qnum;
3283 		break;
3284 	default:
3285 		device_printf(sc->sc_dev, "bogus frame type 0x%x (%s)\n",
3286 			wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK, __func__);
3287 		sc->sc_stats.mst_tx_badframetype++;
3288 		m_freem(m0);
3289 		return EIO;
3290 	}
3291 
3292 	if (IFF_DUMPPKTS_XMIT(sc))
3293 		ieee80211_dump_pkt(ic,
3294 		    mtod(m0, const uint8_t *)+sizeof(uint16_t),
3295 		    m0->m_len - sizeof(uint16_t), ds->DataRate, -1);
3296 
3297 	MWL_TXQ_LOCK(txq);
3298 	ds->Status = htole32(EAGLE_TXD_STATUS_FW_OWNED);
3299 	STAILQ_INSERT_TAIL(&txq->active, bf, bf_list);
3300 	MWL_TXDESC_SYNC(txq, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
3301 
3302 	sc->sc_tx_timer = 5;
3303 	MWL_TXQ_UNLOCK(txq);
3304 
3305 	return 0;
3306 }
3307 
3308 static __inline int
3309 mwl_cvtlegacyrix(int rix)
3310 {
3311 	static const int ieeerates[] =
3312 	    { 2, 4, 11, 22, 44, 12, 18, 24, 36, 48, 72, 96, 108 };
3313 	return (rix < nitems(ieeerates) ? ieeerates[rix] : 0);
3314 }
3315 
3316 /*
3317  * Process completed xmit descriptors from the specified queue.
3318  */
3319 static int
3320 mwl_tx_processq(struct mwl_softc *sc, struct mwl_txq *txq)
3321 {
3322 #define	EAGLE_TXD_STATUS_MCAST \
3323 	(EAGLE_TXD_STATUS_MULTICAST_TX | EAGLE_TXD_STATUS_BROADCAST_TX)
3324 	struct ieee80211com *ic = &sc->sc_ic;
3325 	struct mwl_txbuf *bf;
3326 	struct mwl_txdesc *ds;
3327 	struct ieee80211_node *ni;
3328 	int nreaped;
3329 	uint32_t status;
3330 
3331 	DPRINTF(sc, MWL_DEBUG_TX_PROC, "%s: tx queue %u\n", __func__, txq->qnum);
3332 	for (nreaped = 0;; nreaped++) {
3333 		MWL_TXQ_LOCK(txq);
3334 		bf = STAILQ_FIRST(&txq->active);
3335 		if (bf == NULL) {
3336 			MWL_TXQ_UNLOCK(txq);
3337 			break;
3338 		}
3339 		ds = bf->bf_desc;
3340 		MWL_TXDESC_SYNC(txq, ds,
3341 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
3342 		if (ds->Status & htole32(EAGLE_TXD_STATUS_FW_OWNED)) {
3343 			MWL_TXQ_UNLOCK(txq);
3344 			break;
3345 		}
3346 		STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3347 		MWL_TXQ_UNLOCK(txq);
3348 
3349 #ifdef MWL_DEBUG
3350 		if (sc->sc_debug & MWL_DEBUG_XMIT_DESC)
3351 			mwl_printtxbuf(bf, txq->qnum, nreaped);
3352 #endif
3353 		ni = bf->bf_node;
3354 		if (ni != NULL) {
3355 			status = le32toh(ds->Status);
3356 			if (status & EAGLE_TXD_STATUS_OK) {
3357 				uint16_t Format = le16toh(ds->Format);
3358 				uint8_t txant = _IEEE80211_MASKSHIFT(Format,
3359 				    EAGLE_TXD_ANTENNA);
3360 
3361 				sc->sc_stats.mst_ant_tx[txant]++;
3362 				if (status & EAGLE_TXD_STATUS_OK_RETRY)
3363 					sc->sc_stats.mst_tx_retries++;
3364 				if (status & EAGLE_TXD_STATUS_OK_MORE_RETRY)
3365 					sc->sc_stats.mst_tx_mretries++;
3366 				if (txq->qnum >= MWL_WME_AC_VO)
3367 					ic->ic_wme.wme_hipri_traffic++;
3368 				ni->ni_txrate = _IEEE80211_MASKSHIFT(Format,
3369 				    EAGLE_TXD_RATE);
3370 				if ((Format & EAGLE_TXD_FORMAT_HT) == 0) {
3371 					ni->ni_txrate = mwl_cvtlegacyrix(
3372 					    ni->ni_txrate);
3373 				} else
3374 					ni->ni_txrate |= IEEE80211_RATE_MCS;
3375 				sc->sc_stats.mst_tx_rate = ni->ni_txrate;
3376 			} else {
3377 				if (status & EAGLE_TXD_STATUS_FAILED_LINK_ERROR)
3378 					sc->sc_stats.mst_tx_linkerror++;
3379 				if (status & EAGLE_TXD_STATUS_FAILED_XRETRY)
3380 					sc->sc_stats.mst_tx_xretries++;
3381 				if (status & EAGLE_TXD_STATUS_FAILED_AGING)
3382 					sc->sc_stats.mst_tx_aging++;
3383 				if (bf->bf_m->m_flags & M_FF)
3384 					sc->sc_stats.mst_ff_txerr++;
3385 			}
3386 			if (bf->bf_m->m_flags & M_TXCB)
3387 				/* XXX strip fw len in case header inspected */
3388 				m_adj(bf->bf_m, sizeof(uint16_t));
3389 			ieee80211_tx_complete(ni, bf->bf_m,
3390 			    (status & EAGLE_TXD_STATUS_OK) == 0);
3391 		} else
3392 			m_freem(bf->bf_m);
3393 		ds->Status = htole32(EAGLE_TXD_STATUS_IDLE);
3394 
3395 		bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap,
3396 		    BUS_DMASYNC_POSTWRITE);
3397 		bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3398 
3399 		mwl_puttxbuf_tail(txq, bf);
3400 	}
3401 	return nreaped;
3402 #undef EAGLE_TXD_STATUS_MCAST
3403 }
3404 
3405 /*
3406  * Deferred processing of transmit interrupt; special-cased
3407  * for four hardware queues, 0-3.
3408  */
3409 static void
3410 mwl_tx_proc(void *arg, int npending)
3411 {
3412 	struct mwl_softc *sc = arg;
3413 	int nreaped;
3414 
3415 	/*
3416 	 * Process each active queue.
3417 	 */
3418 	nreaped = 0;
3419 	if (!STAILQ_EMPTY(&sc->sc_txq[0].active))
3420 		nreaped += mwl_tx_processq(sc, &sc->sc_txq[0]);
3421 	if (!STAILQ_EMPTY(&sc->sc_txq[1].active))
3422 		nreaped += mwl_tx_processq(sc, &sc->sc_txq[1]);
3423 	if (!STAILQ_EMPTY(&sc->sc_txq[2].active))
3424 		nreaped += mwl_tx_processq(sc, &sc->sc_txq[2]);
3425 	if (!STAILQ_EMPTY(&sc->sc_txq[3].active))
3426 		nreaped += mwl_tx_processq(sc, &sc->sc_txq[3]);
3427 
3428 	if (nreaped != 0) {
3429 		sc->sc_tx_timer = 0;
3430 		if (mbufq_first(&sc->sc_snd) != NULL) {
3431 			/* NB: kick fw; the tx thread may have been preempted */
3432 			mwl_hal_txstart(sc->sc_mh, 0);
3433 			mwl_start(sc);
3434 		}
3435 	}
3436 }
3437 
3438 static void
3439 mwl_tx_draintxq(struct mwl_softc *sc, struct mwl_txq *txq)
3440 {
3441 	struct ieee80211_node *ni;
3442 	struct mwl_txbuf *bf;
3443 	u_int ix __unused;
3444 
3445 	/*
3446 	 * NB: this assumes output has been stopped and
3447 	 *     we do not need to block mwl_tx_tasklet
3448 	 */
3449 	for (ix = 0;; ix++) {
3450 		MWL_TXQ_LOCK(txq);
3451 		bf = STAILQ_FIRST(&txq->active);
3452 		if (bf == NULL) {
3453 			MWL_TXQ_UNLOCK(txq);
3454 			break;
3455 		}
3456 		STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3457 		MWL_TXQ_UNLOCK(txq);
3458 #ifdef MWL_DEBUG
3459 		if (sc->sc_debug & MWL_DEBUG_RESET) {
3460 			struct ieee80211com *ic = &sc->sc_ic;
3461 			const struct mwltxrec *tr =
3462 			    mtod(bf->bf_m, const struct mwltxrec *);
3463 			mwl_printtxbuf(bf, txq->qnum, ix);
3464 			ieee80211_dump_pkt(ic, (const uint8_t *)&tr->wh,
3465 				bf->bf_m->m_len - sizeof(tr->fwlen), 0, -1);
3466 		}
3467 #endif /* MWL_DEBUG */
3468 		bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3469 		ni = bf->bf_node;
3470 		if (ni != NULL) {
3471 			/*
3472 			 * Reclaim node reference.
3473 			 */
3474 			ieee80211_free_node(ni);
3475 		}
3476 		m_freem(bf->bf_m);
3477 
3478 		mwl_puttxbuf_tail(txq, bf);
3479 	}
3480 }
3481 
3482 /*
3483  * Drain the transmit queues and reclaim resources.
3484  */
3485 static void
3486 mwl_draintxq(struct mwl_softc *sc)
3487 {
3488 	int i;
3489 
3490 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3491 		mwl_tx_draintxq(sc, &sc->sc_txq[i]);
3492 	sc->sc_tx_timer = 0;
3493 }
3494 
3495 #ifdef MWL_DIAGAPI
3496 /*
3497  * Reset the transmit queues to a pristine state after a fw download.
3498  */
3499 static void
3500 mwl_resettxq(struct mwl_softc *sc)
3501 {
3502 	int i;
3503 
3504 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3505 		mwl_txq_reset(sc, &sc->sc_txq[i]);
3506 }
3507 #endif /* MWL_DIAGAPI */
3508 
3509 /*
3510  * Clear the transmit queues of any frames submitted for the
3511  * specified vap.  This is done when the vap is deleted so we
3512  * don't potentially reference the vap after it is gone.
3513  * Note we cannot remove the frames; we only reclaim the node
3514  * reference.
3515  */
3516 static void
3517 mwl_cleartxq(struct mwl_softc *sc, struct ieee80211vap *vap)
3518 {
3519 	struct mwl_txq *txq;
3520 	struct mwl_txbuf *bf;
3521 	int i;
3522 
3523 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
3524 		txq = &sc->sc_txq[i];
3525 		MWL_TXQ_LOCK(txq);
3526 		STAILQ_FOREACH(bf, &txq->active, bf_list) {
3527 			struct ieee80211_node *ni = bf->bf_node;
3528 			if (ni != NULL && ni->ni_vap == vap) {
3529 				bf->bf_node = NULL;
3530 				ieee80211_free_node(ni);
3531 			}
3532 		}
3533 		MWL_TXQ_UNLOCK(txq);
3534 	}
3535 }
3536 
3537 static int
3538 mwl_recv_action(struct ieee80211_node *ni, const struct ieee80211_frame *wh,
3539 	const uint8_t *frm, const uint8_t *efrm)
3540 {
3541 	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3542 	const struct ieee80211_action *ia;
3543 
3544 	ia = (const struct ieee80211_action *) frm;
3545 	if (ia->ia_category == IEEE80211_ACTION_CAT_HT &&
3546 	    ia->ia_action == IEEE80211_ACTION_HT_MIMOPWRSAVE) {
3547 		const struct ieee80211_action_ht_mimopowersave *mps =
3548 		    (const struct ieee80211_action_ht_mimopowersave *) ia;
3549 
3550 		mwl_hal_setmimops(sc->sc_mh, ni->ni_macaddr,
3551 		    mps->am_control & IEEE80211_A_HT_MIMOPWRSAVE_ENA,
3552 		    _IEEE80211_MASKSHIFT(mps->am_control,
3553 			IEEE80211_A_HT_MIMOPWRSAVE_MODE));
3554 		return 0;
3555 	} else
3556 		return sc->sc_recv_action(ni, wh, frm, efrm);
3557 }
3558 
3559 static int
3560 mwl_addba_request(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3561 	int dialogtoken, int baparamset, int batimeout)
3562 {
3563 	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3564 	struct ieee80211vap *vap = ni->ni_vap;
3565 	struct mwl_node *mn = MWL_NODE(ni);
3566 	struct mwl_bastate *bas;
3567 
3568 	bas = tap->txa_private;
3569 	if (bas == NULL) {
3570 		const MWL_HAL_BASTREAM *sp;
3571 		/*
3572 		 * Check for a free BA stream slot.
3573 		 */
3574 #if MWL_MAXBA > 3
3575 		if (mn->mn_ba[3].bastream == NULL)
3576 			bas = &mn->mn_ba[3];
3577 		else
3578 #endif
3579 #if MWL_MAXBA > 2
3580 		if (mn->mn_ba[2].bastream == NULL)
3581 			bas = &mn->mn_ba[2];
3582 		else
3583 #endif
3584 #if MWL_MAXBA > 1
3585 		if (mn->mn_ba[1].bastream == NULL)
3586 			bas = &mn->mn_ba[1];
3587 		else
3588 #endif
3589 #if MWL_MAXBA > 0
3590 		if (mn->mn_ba[0].bastream == NULL)
3591 			bas = &mn->mn_ba[0];
3592 		else
3593 #endif
3594 		{
3595 			/* sta already has max BA streams */
3596 			/* XXX assign BA stream to highest priority tid */
3597 			DPRINTF(sc, MWL_DEBUG_AMPDU,
3598 			    "%s: already has max bastreams\n", __func__);
3599 			sc->sc_stats.mst_ampdu_reject++;
3600 			return 0;
3601 		}
3602 		/* NB: no held reference to ni */
3603 		sp = mwl_hal_bastream_alloc(MWL_VAP(vap)->mv_hvap,
3604 		    (baparamset & IEEE80211_BAPS_POLICY_IMMEDIATE) != 0,
3605 		    ni->ni_macaddr, tap->txa_tid, ni->ni_htparam,
3606 		    ni, tap);
3607 		if (sp == NULL) {
3608 			/*
3609 			 * No available stream, return 0 so no
3610 			 * a-mpdu aggregation will be done.
3611 			 */
3612 			DPRINTF(sc, MWL_DEBUG_AMPDU,
3613 			    "%s: no bastream available\n", __func__);
3614 			sc->sc_stats.mst_ampdu_nostream++;
3615 			return 0;
3616 		}
3617 		DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: alloc bastream %p\n",
3618 		    __func__, sp);
3619 		/* NB: qos is left zero so we won't match in mwl_tx_start */
3620 		bas->bastream = sp;
3621 		tap->txa_private = bas;
3622 	}
3623 	/* fetch current seq# from the firmware; if available */
3624 	if (mwl_hal_bastream_get_seqno(sc->sc_mh, bas->bastream,
3625 	    vap->iv_opmode == IEEE80211_M_STA ? vap->iv_myaddr : ni->ni_macaddr,
3626 	    &tap->txa_start) != 0)
3627 		tap->txa_start = 0;
3628 	return sc->sc_addba_request(ni, tap, dialogtoken, baparamset, batimeout);
3629 }
3630 
3631 static int
3632 mwl_addba_response(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3633 	int code, int baparamset, int batimeout)
3634 {
3635 	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3636 	struct mwl_bastate *bas;
3637 
3638 	bas = tap->txa_private;
3639 	if (bas == NULL) {
3640 		/* XXX should not happen */
3641 		DPRINTF(sc, MWL_DEBUG_AMPDU,
3642 		    "%s: no BA stream allocated, TID %d\n",
3643 		    __func__, tap->txa_tid);
3644 		sc->sc_stats.mst_addba_nostream++;
3645 		return 0;
3646 	}
3647 	if (code == IEEE80211_STATUS_SUCCESS) {
3648 		struct ieee80211vap *vap = ni->ni_vap;
3649 		int bufsiz, error;
3650 
3651 		/*
3652 		 * Tell the firmware to setup the BA stream;
3653 		 * we know resources are available because we
3654 		 * pre-allocated one before forming the request.
3655 		 */
3656 		bufsiz = _IEEE80211_MASKSHIFT(baparamset, IEEE80211_BAPS_BUFSIZ);
3657 		if (bufsiz == 0)
3658 			bufsiz = IEEE80211_AGGR_BAWMAX;
3659 		error = mwl_hal_bastream_create(MWL_VAP(vap)->mv_hvap,
3660 		    bas->bastream, bufsiz, bufsiz, tap->txa_start);
3661 		if (error != 0) {
3662 			/*
3663 			 * Setup failed, return immediately so no a-mpdu
3664 			 * aggregation will be done.
3665 			 */
3666 			mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3667 			mwl_bastream_free(bas);
3668 			tap->txa_private = NULL;
3669 
3670 			DPRINTF(sc, MWL_DEBUG_AMPDU,
3671 			    "%s: create failed, error %d, bufsiz %d TID %d "
3672 			    "htparam 0x%x\n", __func__, error, bufsiz,
3673 			    tap->txa_tid, ni->ni_htparam);
3674 			sc->sc_stats.mst_bacreate_failed++;
3675 			return 0;
3676 		}
3677 		/* NB: cache txq to avoid ptr indirect */
3678 		mwl_bastream_setup(bas, tap->txa_tid, bas->bastream->txq);
3679 		DPRINTF(sc, MWL_DEBUG_AMPDU,
3680 		    "%s: bastream %p assigned to txq %d TID %d bufsiz %d "
3681 		    "htparam 0x%x\n", __func__, bas->bastream,
3682 		    bas->txq, tap->txa_tid, bufsiz, ni->ni_htparam);
3683 	} else {
3684 		/*
3685 		 * Other side NAK'd us; return the resources.
3686 		 */
3687 		DPRINTF(sc, MWL_DEBUG_AMPDU,
3688 		    "%s: request failed with code %d, destroy bastream %p\n",
3689 		    __func__, code, bas->bastream);
3690 		mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3691 		mwl_bastream_free(bas);
3692 		tap->txa_private = NULL;
3693 	}
3694 	/* NB: firmware sends BAR so we don't need to */
3695 	return sc->sc_addba_response(ni, tap, code, baparamset, batimeout);
3696 }
3697 
3698 static void
3699 mwl_addba_stop(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap)
3700 {
3701 	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3702 	struct mwl_bastate *bas;
3703 
3704 	bas = tap->txa_private;
3705 	if (bas != NULL) {
3706 		DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: destroy bastream %p\n",
3707 		    __func__, bas->bastream);
3708 		mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3709 		mwl_bastream_free(bas);
3710 		tap->txa_private = NULL;
3711 	}
3712 	sc->sc_addba_stop(ni, tap);
3713 }
3714 
3715 /*
3716  * Setup the rx data structures.  This should only be
3717  * done once or we may get out of sync with the firmware.
3718  */
3719 static int
3720 mwl_startrecv(struct mwl_softc *sc)
3721 {
3722 	if (!sc->sc_recvsetup) {
3723 		struct mwl_rxbuf *bf, *prev;
3724 		struct mwl_rxdesc *ds;
3725 
3726 		prev = NULL;
3727 		STAILQ_FOREACH(bf, &sc->sc_rxbuf, bf_list) {
3728 			int error = mwl_rxbuf_init(sc, bf);
3729 			if (error != 0) {
3730 				DPRINTF(sc, MWL_DEBUG_RECV,
3731 					"%s: mwl_rxbuf_init failed %d\n",
3732 					__func__, error);
3733 				return error;
3734 			}
3735 			if (prev != NULL) {
3736 				ds = prev->bf_desc;
3737 				ds->pPhysNext = htole32(bf->bf_daddr);
3738 			}
3739 			prev = bf;
3740 		}
3741 		if (prev != NULL) {
3742 			ds = prev->bf_desc;
3743 			ds->pPhysNext =
3744 			    htole32(STAILQ_FIRST(&sc->sc_rxbuf)->bf_daddr);
3745 		}
3746 		sc->sc_recvsetup = 1;
3747 	}
3748 	mwl_mode_init(sc);		/* set filters, etc. */
3749 	return 0;
3750 }
3751 
3752 static MWL_HAL_APMODE
3753 mwl_getapmode(const struct ieee80211vap *vap, struct ieee80211_channel *chan)
3754 {
3755 	MWL_HAL_APMODE mode;
3756 
3757 	if (IEEE80211_IS_CHAN_HT(chan)) {
3758 		if (vap->iv_flags_ht & IEEE80211_FHT_PUREN)
3759 			mode = AP_MODE_N_ONLY;
3760 		else if (IEEE80211_IS_CHAN_5GHZ(chan))
3761 			mode = AP_MODE_AandN;
3762 		else if (vap->iv_flags & IEEE80211_F_PUREG)
3763 			mode = AP_MODE_GandN;
3764 		else
3765 			mode = AP_MODE_BandGandN;
3766 	} else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3767 		if (vap->iv_flags & IEEE80211_F_PUREG)
3768 			mode = AP_MODE_G_ONLY;
3769 		else
3770 			mode = AP_MODE_MIXED;
3771 	} else if (IEEE80211_IS_CHAN_B(chan))
3772 		mode = AP_MODE_B_ONLY;
3773 	else if (IEEE80211_IS_CHAN_A(chan))
3774 		mode = AP_MODE_A_ONLY;
3775 	else
3776 		mode = AP_MODE_MIXED;		/* XXX should not happen? */
3777 	return mode;
3778 }
3779 
3780 static int
3781 mwl_setapmode(struct ieee80211vap *vap, struct ieee80211_channel *chan)
3782 {
3783 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
3784 	return mwl_hal_setapmode(hvap, mwl_getapmode(vap, chan));
3785 }
3786 
3787 /*
3788  * Set/change channels.
3789  */
3790 static int
3791 mwl_chan_set(struct mwl_softc *sc, struct ieee80211_channel *chan)
3792 {
3793 	struct mwl_hal *mh = sc->sc_mh;
3794 	struct ieee80211com *ic = &sc->sc_ic;
3795 	MWL_HAL_CHANNEL hchan;
3796 	int maxtxpow;
3797 
3798 	DPRINTF(sc, MWL_DEBUG_RESET, "%s: chan %u MHz/flags 0x%x\n",
3799 	    __func__, chan->ic_freq, chan->ic_flags);
3800 
3801 	/*
3802 	 * Convert to a HAL channel description with
3803 	 * the flags constrained to reflect the current
3804 	 * operating mode.
3805 	 */
3806 	mwl_mapchan(&hchan, chan);
3807 	mwl_hal_intrset(mh, 0);		/* disable interrupts */
3808 #if 0
3809 	mwl_draintxq(sc);		/* clear pending tx frames */
3810 #endif
3811 	mwl_hal_setchannel(mh, &hchan);
3812 	/*
3813 	 * Tx power is cap'd by the regulatory setting and
3814 	 * possibly a user-set limit.  We pass the min of
3815 	 * these to the hal to apply them to the cal data
3816 	 * for this channel.
3817 	 * XXX min bound?
3818 	 */
3819 	maxtxpow = 2*chan->ic_maxregpower;
3820 	if (maxtxpow > ic->ic_txpowlimit)
3821 		maxtxpow = ic->ic_txpowlimit;
3822 	mwl_hal_settxpower(mh, &hchan, maxtxpow / 2);
3823 	/* NB: potentially change mcast/mgt rates */
3824 	mwl_setcurchanrates(sc);
3825 
3826 	/*
3827 	 * Update internal state.
3828 	 */
3829 	sc->sc_tx_th.wt_chan_freq = htole16(chan->ic_freq);
3830 	sc->sc_rx_th.wr_chan_freq = htole16(chan->ic_freq);
3831 	if (IEEE80211_IS_CHAN_A(chan)) {
3832 		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_A);
3833 		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_A);
3834 	} else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3835 		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_G);
3836 		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_G);
3837 	} else {
3838 		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_B);
3839 		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_B);
3840 	}
3841 	sc->sc_curchan = hchan;
3842 	mwl_hal_intrset(mh, sc->sc_imask);
3843 
3844 	return 0;
3845 }
3846 
3847 static void
3848 mwl_scan_start(struct ieee80211com *ic)
3849 {
3850 	struct mwl_softc *sc = ic->ic_softc;
3851 
3852 	DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3853 }
3854 
3855 static void
3856 mwl_scan_end(struct ieee80211com *ic)
3857 {
3858 	struct mwl_softc *sc = ic->ic_softc;
3859 
3860 	DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3861 }
3862 
3863 static void
3864 mwl_set_channel(struct ieee80211com *ic)
3865 {
3866 	struct mwl_softc *sc = ic->ic_softc;
3867 
3868 	(void) mwl_chan_set(sc, ic->ic_curchan);
3869 }
3870 
3871 /*
3872  * Handle a channel switch request.  We inform the firmware
3873  * and mark the global state to suppress various actions.
3874  * NB: we issue only one request to the fw; we may be called
3875  * multiple times if there are multiple vap's.
3876  */
3877 static void
3878 mwl_startcsa(struct ieee80211vap *vap)
3879 {
3880 	struct ieee80211com *ic = vap->iv_ic;
3881 	struct mwl_softc *sc = ic->ic_softc;
3882 	MWL_HAL_CHANNEL hchan;
3883 
3884 	if (sc->sc_csapending)
3885 		return;
3886 
3887 	mwl_mapchan(&hchan, ic->ic_csa_newchan);
3888 	/* 1 =>'s quiet channel */
3889 	mwl_hal_setchannelswitchie(sc->sc_mh, &hchan, 1, ic->ic_csa_count);
3890 	sc->sc_csapending = 1;
3891 }
3892 
3893 /*
3894  * Plumb any static WEP key for the station.  This is
3895  * necessary as we must propagate the key from the
3896  * global key table of the vap to each sta db entry.
3897  */
3898 static void
3899 mwl_setanywepkey(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
3900 {
3901 	if ((vap->iv_flags & (IEEE80211_F_PRIVACY|IEEE80211_F_WPA)) ==
3902 		IEEE80211_F_PRIVACY &&
3903 	    vap->iv_def_txkey != IEEE80211_KEYIX_NONE &&
3904 	    vap->iv_nw_keys[vap->iv_def_txkey].wk_keyix != IEEE80211_KEYIX_NONE)
3905 		(void) _mwl_key_set(vap, &vap->iv_nw_keys[vap->iv_def_txkey],
3906 				    mac);
3907 }
3908 
3909 static int
3910 mwl_peerstadb(struct ieee80211_node *ni, int aid, int staid, MWL_HAL_PEERINFO *pi)
3911 {
3912 #define	WME(ie) ((const struct ieee80211_wme_info *) ie)
3913 	struct ieee80211vap *vap = ni->ni_vap;
3914 	struct mwl_hal_vap *hvap;
3915 	int error;
3916 
3917 	if (vap->iv_opmode == IEEE80211_M_WDS) {
3918 		/*
3919 		 * WDS vap's do not have a f/w vap; instead they piggyback
3920 		 * on an AP vap and we must install the sta db entry and
3921 		 * crypto state using that AP's handle (the WDS vap has none).
3922 		 */
3923 		hvap = MWL_VAP(vap)->mv_ap_hvap;
3924 	} else
3925 		hvap = MWL_VAP(vap)->mv_hvap;
3926 	error = mwl_hal_newstation(hvap, ni->ni_macaddr,
3927 	    aid, staid, pi,
3928 	    ni->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT),
3929 	    ni->ni_ies.wme_ie != NULL ? WME(ni->ni_ies.wme_ie)->wme_info : 0);
3930 	if (error == 0) {
3931 		/*
3932 		 * Setup security for this station.  For sta mode this is
3933 		 * needed even though do the same thing on transition to
3934 		 * AUTH state because the call to mwl_hal_newstation
3935 		 * clobbers the crypto state we setup.
3936 		 */
3937 		mwl_setanywepkey(vap, ni->ni_macaddr);
3938 	}
3939 	return error;
3940 #undef WME
3941 }
3942 
3943 static void
3944 mwl_setglobalkeys(struct ieee80211vap *vap)
3945 {
3946 	struct ieee80211_key *wk;
3947 
3948 	wk = &vap->iv_nw_keys[0];
3949 	for (; wk < &vap->iv_nw_keys[IEEE80211_WEP_NKID]; wk++)
3950 		if (wk->wk_keyix != IEEE80211_KEYIX_NONE)
3951 			(void) _mwl_key_set(vap, wk, vap->iv_myaddr);
3952 }
3953 
3954 /*
3955  * Convert a legacy rate set to a firmware bitmask.
3956  */
3957 static uint32_t
3958 get_rate_bitmap(const struct ieee80211_rateset *rs)
3959 {
3960 	uint32_t rates;
3961 	int i;
3962 
3963 	rates = 0;
3964 	for (i = 0; i < rs->rs_nrates; i++)
3965 		switch (rs->rs_rates[i] & IEEE80211_RATE_VAL) {
3966 		case 2:	  rates |= 0x001; break;
3967 		case 4:	  rates |= 0x002; break;
3968 		case 11:  rates |= 0x004; break;
3969 		case 22:  rates |= 0x008; break;
3970 		case 44:  rates |= 0x010; break;
3971 		case 12:  rates |= 0x020; break;
3972 		case 18:  rates |= 0x040; break;
3973 		case 24:  rates |= 0x080; break;
3974 		case 36:  rates |= 0x100; break;
3975 		case 48:  rates |= 0x200; break;
3976 		case 72:  rates |= 0x400; break;
3977 		case 96:  rates |= 0x800; break;
3978 		case 108: rates |= 0x1000; break;
3979 		}
3980 	return rates;
3981 }
3982 
3983 /*
3984  * Construct an HT firmware bitmask from an HT rate set.
3985  */
3986 static uint32_t
3987 get_htrate_bitmap(const struct ieee80211_htrateset *rs)
3988 {
3989 	uint32_t rates;
3990 	int i;
3991 
3992 	rates = 0;
3993 	for (i = 0; i < rs->rs_nrates; i++) {
3994 		if (rs->rs_rates[i] < 16)
3995 			rates |= 1<<rs->rs_rates[i];
3996 	}
3997 	return rates;
3998 }
3999 
4000 /*
4001  * Craft station database entry for station.
4002  * NB: use host byte order here, the hal handles byte swapping.
4003  */
4004 static MWL_HAL_PEERINFO *
4005 mkpeerinfo(MWL_HAL_PEERINFO *pi, const struct ieee80211_node *ni)
4006 {
4007 	const struct ieee80211vap *vap = ni->ni_vap;
4008 
4009 	memset(pi, 0, sizeof(*pi));
4010 	pi->LegacyRateBitMap = get_rate_bitmap(&ni->ni_rates);
4011 	pi->CapInfo = ni->ni_capinfo;
4012 	if (ni->ni_flags & IEEE80211_NODE_HT) {
4013 		/* HT capabilities, etc */
4014 		pi->HTCapabilitiesInfo = ni->ni_htcap;
4015 		/* XXX pi.HTCapabilitiesInfo */
4016 	        pi->MacHTParamInfo = ni->ni_htparam;
4017 		pi->HTRateBitMap = get_htrate_bitmap(&ni->ni_htrates);
4018 		pi->AddHtInfo.ControlChan = ni->ni_htctlchan;
4019 		pi->AddHtInfo.AddChan = ni->ni_ht2ndchan;
4020 		pi->AddHtInfo.OpMode = ni->ni_htopmode;
4021 		pi->AddHtInfo.stbc = ni->ni_htstbc;
4022 
4023 		/* constrain according to local configuration */
4024 		if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI40) == 0)
4025 			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI40;
4026 		if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI20) == 0)
4027 			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI20;
4028 		if (ni->ni_chw != 40)
4029 			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_CHWIDTH40;
4030 	}
4031 	return pi;
4032 }
4033 
4034 /*
4035  * Re-create the local sta db entry for a vap to ensure
4036  * up to date WME state is pushed to the firmware.  Because
4037  * this resets crypto state this must be followed by a
4038  * reload of any keys in the global key table.
4039  */
4040 static int
4041 mwl_localstadb(struct ieee80211vap *vap)
4042 {
4043 #define	WME(ie) ((const struct ieee80211_wme_info *) ie)
4044 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
4045 	struct ieee80211_node *bss;
4046 	MWL_HAL_PEERINFO pi;
4047 	int error;
4048 
4049 	switch (vap->iv_opmode) {
4050 	case IEEE80211_M_STA:
4051 		bss = vap->iv_bss;
4052 		error = mwl_hal_newstation(hvap, vap->iv_myaddr, 0, 0,
4053 		    vap->iv_state == IEEE80211_S_RUN ?
4054 			mkpeerinfo(&pi, bss) : NULL,
4055 		    (bss->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT)),
4056 		    bss->ni_ies.wme_ie != NULL ?
4057 			WME(bss->ni_ies.wme_ie)->wme_info : 0);
4058 		if (error == 0)
4059 			mwl_setglobalkeys(vap);
4060 		break;
4061 	case IEEE80211_M_HOSTAP:
4062 	case IEEE80211_M_MBSS:
4063 		error = mwl_hal_newstation(hvap, vap->iv_myaddr,
4064 		    0, 0, NULL, vap->iv_flags & IEEE80211_F_WME, 0);
4065 		if (error == 0)
4066 			mwl_setglobalkeys(vap);
4067 		break;
4068 	default:
4069 		error = 0;
4070 		break;
4071 	}
4072 	return error;
4073 #undef WME
4074 }
4075 
4076 static int
4077 mwl_newstate(struct ieee80211vap *vap, enum ieee80211_state nstate, int arg)
4078 {
4079 	struct mwl_vap *mvp = MWL_VAP(vap);
4080 	struct mwl_hal_vap *hvap = mvp->mv_hvap;
4081 	struct ieee80211com *ic = vap->iv_ic;
4082 	struct ieee80211_node *ni = NULL;
4083 	struct mwl_softc *sc = ic->ic_softc;
4084 	struct mwl_hal *mh = sc->sc_mh;
4085 	enum ieee80211_state ostate = vap->iv_state;
4086 	int error;
4087 
4088 	DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: %s -> %s\n",
4089 	    if_name(vap->iv_ifp), __func__,
4090 	    ieee80211_state_name[ostate], ieee80211_state_name[nstate]);
4091 
4092 	callout_stop(&sc->sc_timer);
4093 	/*
4094 	 * Clear current radar detection state.
4095 	 */
4096 	if (ostate == IEEE80211_S_CAC) {
4097 		/* stop quiet mode radar detection */
4098 		mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_STOP);
4099 	} else if (sc->sc_radarena) {
4100 		/* stop in-service radar detection */
4101 		mwl_hal_setradardetection(mh, DR_DFS_DISABLE);
4102 		sc->sc_radarena = 0;
4103 	}
4104 	/*
4105 	 * Carry out per-state actions before doing net80211 work.
4106 	 */
4107 	if (nstate == IEEE80211_S_INIT) {
4108 		/* NB: only ap+sta vap's have a fw entity */
4109 		if (hvap != NULL)
4110 			mwl_hal_stop(hvap);
4111 	} else if (nstate == IEEE80211_S_SCAN) {
4112 		mwl_hal_start(hvap);
4113 		/* NB: this disables beacon frames */
4114 		mwl_hal_setinframode(hvap);
4115 	} else if (nstate == IEEE80211_S_AUTH) {
4116 		/*
4117 		 * Must create a sta db entry in case a WEP key needs to
4118 		 * be plumbed.  This entry will be overwritten if we
4119 		 * associate; otherwise it will be reclaimed on node free.
4120 		 */
4121 		ni = vap->iv_bss;
4122 		MWL_NODE(ni)->mn_hvap = hvap;
4123 		(void) mwl_peerstadb(ni, 0, 0, NULL);
4124 	} else if (nstate == IEEE80211_S_CSA) {
4125 		/* XXX move to below? */
4126 		if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
4127 		    vap->iv_opmode == IEEE80211_M_MBSS)
4128 			mwl_startcsa(vap);
4129 	} else if (nstate == IEEE80211_S_CAC) {
4130 		/* XXX move to below? */
4131 		/* stop ap xmit and enable quiet mode radar detection */
4132 		mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_START);
4133 	}
4134 
4135 	/*
4136 	 * Invoke the parent method to do net80211 work.
4137 	 */
4138 	error = mvp->mv_newstate(vap, nstate, arg);
4139 
4140 	/*
4141 	 * Carry out work that must be done after net80211 runs;
4142 	 * this work requires up to date state (e.g. iv_bss).
4143 	 */
4144 	if (error == 0 && nstate == IEEE80211_S_RUN) {
4145 		/* NB: collect bss node again, it may have changed */
4146 		ni = vap->iv_bss;
4147 
4148 		DPRINTF(sc, MWL_DEBUG_STATE,
4149 		    "%s: %s(RUN): iv_flags 0x%08x bintvl %d bssid %s "
4150 		    "capinfo 0x%04x chan %d\n",
4151 		    if_name(vap->iv_ifp), __func__, vap->iv_flags,
4152 		    ni->ni_intval, ether_sprintf(ni->ni_bssid), ni->ni_capinfo,
4153 		    ieee80211_chan2ieee(ic, ic->ic_curchan));
4154 
4155 		/*
4156 		 * Recreate local sta db entry to update WME/HT state.
4157 		 */
4158 		mwl_localstadb(vap);
4159 		switch (vap->iv_opmode) {
4160 		case IEEE80211_M_HOSTAP:
4161 		case IEEE80211_M_MBSS:
4162 			if (ostate == IEEE80211_S_CAC) {
4163 				/* enable in-service radar detection */
4164 				mwl_hal_setradardetection(mh,
4165 				    DR_IN_SERVICE_MONITOR_START);
4166 				sc->sc_radarena = 1;
4167 			}
4168 			/*
4169 			 * Allocate and setup the beacon frame
4170 			 * (and related state).
4171 			 */
4172 			error = mwl_reset_vap(vap, IEEE80211_S_RUN);
4173 			if (error != 0) {
4174 				DPRINTF(sc, MWL_DEBUG_STATE,
4175 				    "%s: beacon setup failed, error %d\n",
4176 				    __func__, error);
4177 				goto bad;
4178 			}
4179 			/* NB: must be after setting up beacon */
4180 			mwl_hal_start(hvap);
4181 			break;
4182 		case IEEE80211_M_STA:
4183 			DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: aid 0x%x\n",
4184 			    if_name(vap->iv_ifp), __func__, ni->ni_associd);
4185 			/*
4186 			 * Set state now that we're associated.
4187 			 */
4188 			mwl_hal_setassocid(hvap, ni->ni_bssid, ni->ni_associd);
4189 			mwl_setrates(vap);
4190 			mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
4191 			if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4192 			    sc->sc_ndwdsvaps++ == 0)
4193 				mwl_hal_setdwds(mh, 1);
4194 			break;
4195 		case IEEE80211_M_WDS:
4196 			DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: bssid %s\n",
4197 			    if_name(vap->iv_ifp), __func__,
4198 			    ether_sprintf(ni->ni_bssid));
4199 			mwl_seteapolformat(vap);
4200 			break;
4201 		default:
4202 			break;
4203 		}
4204 		/*
4205 		 * Set CS mode according to operating channel;
4206 		 * this mostly an optimization for 5GHz.
4207 		 *
4208 		 * NB: must follow mwl_hal_start which resets csmode
4209 		 */
4210 		if (IEEE80211_IS_CHAN_5GHZ(ic->ic_bsschan))
4211 			mwl_hal_setcsmode(mh, CSMODE_AGGRESSIVE);
4212 		else
4213 			mwl_hal_setcsmode(mh, CSMODE_AUTO_ENA);
4214 		/*
4215 		 * Start timer to prod firmware.
4216 		 */
4217 		if (sc->sc_ageinterval != 0)
4218 			callout_reset(&sc->sc_timer, sc->sc_ageinterval*hz,
4219 			    mwl_agestations, sc);
4220 	} else if (nstate == IEEE80211_S_SLEEP) {
4221 		/* XXX set chip in power save */
4222 	} else if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4223 	    --sc->sc_ndwdsvaps == 0)
4224 		mwl_hal_setdwds(mh, 0);
4225 bad:
4226 	return error;
4227 }
4228 
4229 /*
4230  * Manage station id's; these are separate from AID's
4231  * as AID's may have values out of the range of possible
4232  * station id's acceptable to the firmware.
4233  */
4234 static int
4235 allocstaid(struct mwl_softc *sc, int aid)
4236 {
4237 	int staid;
4238 
4239 	if (!(0 < aid && aid < MWL_MAXSTAID) || isset(sc->sc_staid, aid)) {
4240 		/* NB: don't use 0 */
4241 		for (staid = 1; staid < MWL_MAXSTAID; staid++)
4242 			if (isclr(sc->sc_staid, staid))
4243 				break;
4244 	} else
4245 		staid = aid;
4246 	setbit(sc->sc_staid, staid);
4247 	return staid;
4248 }
4249 
4250 static void
4251 delstaid(struct mwl_softc *sc, int staid)
4252 {
4253 	clrbit(sc->sc_staid, staid);
4254 }
4255 
4256 /*
4257  * Setup driver-specific state for a newly associated node.
4258  * Note that we're called also on a re-associate, the isnew
4259  * param tells us if this is the first time or not.
4260  */
4261 static void
4262 mwl_newassoc(struct ieee80211_node *ni, int isnew)
4263 {
4264 	struct ieee80211vap *vap = ni->ni_vap;
4265         struct mwl_softc *sc = vap->iv_ic->ic_softc;
4266 	struct mwl_node *mn = MWL_NODE(ni);
4267 	MWL_HAL_PEERINFO pi;
4268 	uint16_t aid;
4269 	int error;
4270 
4271 	aid = IEEE80211_AID(ni->ni_associd);
4272 	if (isnew) {
4273 		mn->mn_staid = allocstaid(sc, aid);
4274 		mn->mn_hvap = MWL_VAP(vap)->mv_hvap;
4275 	} else {
4276 		mn = MWL_NODE(ni);
4277 		/* XXX reset BA stream? */
4278 	}
4279 	DPRINTF(sc, MWL_DEBUG_NODE, "%s: mac %s isnew %d aid %d staid %d\n",
4280 	    __func__, ether_sprintf(ni->ni_macaddr), isnew, aid, mn->mn_staid);
4281 	error = mwl_peerstadb(ni, aid, mn->mn_staid, mkpeerinfo(&pi, ni));
4282 	if (error != 0) {
4283 		DPRINTF(sc, MWL_DEBUG_NODE,
4284 		    "%s: error %d creating sta db entry\n",
4285 		    __func__, error);
4286 		/* XXX how to deal with error? */
4287 	}
4288 }
4289 
4290 /*
4291  * Periodically poke the firmware to age out station state
4292  * (power save queues, pending tx aggregates).
4293  */
4294 static void
4295 mwl_agestations(void *arg)
4296 {
4297 	struct mwl_softc *sc = arg;
4298 
4299 	mwl_hal_setkeepalive(sc->sc_mh);
4300 	if (sc->sc_ageinterval != 0)		/* NB: catch dynamic changes */
4301 		callout_schedule(&sc->sc_timer, sc->sc_ageinterval*hz);
4302 }
4303 
4304 static const struct mwl_hal_channel *
4305 findhalchannel(const MWL_HAL_CHANNELINFO *ci, int ieee)
4306 {
4307 	int i;
4308 
4309 	for (i = 0; i < ci->nchannels; i++) {
4310 		const struct mwl_hal_channel *hc = &ci->channels[i];
4311 		if (hc->ieee == ieee)
4312 			return hc;
4313 	}
4314 	return NULL;
4315 }
4316 
4317 static int
4318 mwl_setregdomain(struct ieee80211com *ic, struct ieee80211_regdomain *rd,
4319 	int nchan, struct ieee80211_channel chans[])
4320 {
4321 	struct mwl_softc *sc = ic->ic_softc;
4322 	struct mwl_hal *mh = sc->sc_mh;
4323 	const MWL_HAL_CHANNELINFO *ci;
4324 	int i;
4325 
4326 	for (i = 0; i < nchan; i++) {
4327 		struct ieee80211_channel *c = &chans[i];
4328 		const struct mwl_hal_channel *hc;
4329 
4330 		if (IEEE80211_IS_CHAN_2GHZ(c)) {
4331 			mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_2DOT4GHZ,
4332 			    IEEE80211_IS_CHAN_HT40(c) ?
4333 				MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4334 		} else if (IEEE80211_IS_CHAN_5GHZ(c)) {
4335 			mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_5GHZ,
4336 			    IEEE80211_IS_CHAN_HT40(c) ?
4337 				MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4338 		} else {
4339 			device_printf(sc->sc_dev,
4340 			    "%s: channel %u freq %u/0x%x not 2.4/5GHz\n",
4341 			    __func__, c->ic_ieee, c->ic_freq, c->ic_flags);
4342 			return EINVAL;
4343 		}
4344 		/*
4345 		 * Verify channel has cal data and cap tx power.
4346 		 */
4347 		hc = findhalchannel(ci, c->ic_ieee);
4348 		if (hc != NULL) {
4349 			if (c->ic_maxpower > 2*hc->maxTxPow)
4350 				c->ic_maxpower = 2*hc->maxTxPow;
4351 			goto next;
4352 		}
4353 		if (IEEE80211_IS_CHAN_HT40(c)) {
4354 			/*
4355 			 * Look for the extension channel since the
4356 			 * hal table only has the primary channel.
4357 			 */
4358 			hc = findhalchannel(ci, c->ic_extieee);
4359 			if (hc != NULL) {
4360 				if (c->ic_maxpower > 2*hc->maxTxPow)
4361 					c->ic_maxpower = 2*hc->maxTxPow;
4362 				goto next;
4363 			}
4364 		}
4365 		device_printf(sc->sc_dev,
4366 		    "%s: no cal data for channel %u ext %u freq %u/0x%x\n",
4367 		    __func__, c->ic_ieee, c->ic_extieee,
4368 		    c->ic_freq, c->ic_flags);
4369 		return EINVAL;
4370 	next:
4371 		;
4372 	}
4373 	return 0;
4374 }
4375 
4376 #define	IEEE80211_CHAN_HTG	(IEEE80211_CHAN_HT|IEEE80211_CHAN_G)
4377 #define	IEEE80211_CHAN_HTA	(IEEE80211_CHAN_HT|IEEE80211_CHAN_A)
4378 
4379 static void
4380 addht40channels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4381 	const MWL_HAL_CHANNELINFO *ci, int flags)
4382 {
4383 	int i, error;
4384 
4385 	for (i = 0; i < ci->nchannels; i++) {
4386 		const struct mwl_hal_channel *hc = &ci->channels[i];
4387 
4388 		error = ieee80211_add_channel_ht40(chans, maxchans, nchans,
4389 		    hc->ieee, hc->maxTxPow, flags);
4390 		if (error != 0 && error != ENOENT)
4391 			break;
4392 	}
4393 }
4394 
4395 static void
4396 addchannels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4397 	const MWL_HAL_CHANNELINFO *ci, const uint8_t bands[])
4398 {
4399 	int i, error;
4400 
4401 	error = 0;
4402 	for (i = 0; i < ci->nchannels && error == 0; i++) {
4403 		const struct mwl_hal_channel *hc = &ci->channels[i];
4404 
4405 		error = ieee80211_add_channel(chans, maxchans, nchans,
4406 		    hc->ieee, hc->freq, hc->maxTxPow, 0, bands);
4407 	}
4408 }
4409 
4410 static void
4411 getchannels(struct mwl_softc *sc, int maxchans, int *nchans,
4412 	struct ieee80211_channel chans[])
4413 {
4414 	const MWL_HAL_CHANNELINFO *ci;
4415 	uint8_t bands[IEEE80211_MODE_BYTES];
4416 
4417 	/*
4418 	 * Use the channel info from the hal to craft the
4419 	 * channel list.  Note that we pass back an unsorted
4420 	 * list; the caller is required to sort it for us
4421 	 * (if desired).
4422 	 */
4423 	*nchans = 0;
4424 	if (mwl_hal_getchannelinfo(sc->sc_mh,
4425 	    MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0) {
4426 		memset(bands, 0, sizeof(bands));
4427 		setbit(bands, IEEE80211_MODE_11B);
4428 		setbit(bands, IEEE80211_MODE_11G);
4429 		setbit(bands, IEEE80211_MODE_11NG);
4430 		addchannels(chans, maxchans, nchans, ci, bands);
4431 	}
4432 	if (mwl_hal_getchannelinfo(sc->sc_mh,
4433 	    MWL_FREQ_BAND_5GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0) {
4434 		memset(bands, 0, sizeof(bands));
4435 		setbit(bands, IEEE80211_MODE_11A);
4436 		setbit(bands, IEEE80211_MODE_11NA);
4437 		addchannels(chans, maxchans, nchans, ci, bands);
4438 	}
4439 	if (mwl_hal_getchannelinfo(sc->sc_mh,
4440 	    MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4441 		addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG);
4442 	if (mwl_hal_getchannelinfo(sc->sc_mh,
4443 	    MWL_FREQ_BAND_5GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4444 		addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA);
4445 }
4446 
4447 static void
4448 mwl_getradiocaps(struct ieee80211com *ic,
4449 	int maxchans, int *nchans, struct ieee80211_channel chans[])
4450 {
4451 	struct mwl_softc *sc = ic->ic_softc;
4452 
4453 	getchannels(sc, maxchans, nchans, chans);
4454 }
4455 
4456 static int
4457 mwl_getchannels(struct mwl_softc *sc)
4458 {
4459 	struct ieee80211com *ic = &sc->sc_ic;
4460 
4461 	/*
4462 	 * Use the channel info from the hal to craft the
4463 	 * channel list for net80211.  Note that we pass up
4464 	 * an unsorted list; net80211 will sort it for us.
4465 	 */
4466 	memset(ic->ic_channels, 0, sizeof(ic->ic_channels));
4467 	ic->ic_nchans = 0;
4468 	getchannels(sc, IEEE80211_CHAN_MAX, &ic->ic_nchans, ic->ic_channels);
4469 
4470 	ic->ic_regdomain.regdomain = SKU_DEBUG;
4471 	ic->ic_regdomain.country = CTRY_DEFAULT;
4472 	ic->ic_regdomain.location = 'I';
4473 	ic->ic_regdomain.isocc[0] = ' ';	/* XXX? */
4474 	ic->ic_regdomain.isocc[1] = ' ';
4475 	return (ic->ic_nchans == 0 ? EIO : 0);
4476 }
4477 #undef IEEE80211_CHAN_HTA
4478 #undef IEEE80211_CHAN_HTG
4479 
4480 #ifdef MWL_DEBUG
4481 static void
4482 mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix)
4483 {
4484 	const struct mwl_rxdesc *ds = bf->bf_desc;
4485 	uint32_t status = le32toh(ds->Status);
4486 
4487 	printf("R[%2u] (DS.V:%p DS.P:0x%jx) NEXT:%08x DATA:%08x RC:%02x%s\n"
4488 	       "      STAT:%02x LEN:%04x RSSI:%02x CHAN:%02x RATE:%02x QOS:%04x HT:%04x\n",
4489 	    ix, ds, (uintmax_t)bf->bf_daddr, le32toh(ds->pPhysNext),
4490 	    le32toh(ds->pPhysBuffData), ds->RxControl,
4491 	    ds->RxControl != EAGLE_RXD_CTRL_DRIVER_OWN ?
4492 	        "" : (status & EAGLE_RXD_STATUS_OK) ? " *" : " !",
4493 	    ds->Status, le16toh(ds->PktLen), ds->RSSI, ds->Channel,
4494 	    ds->Rate, le16toh(ds->QosCtrl), le16toh(ds->HtSig2));
4495 }
4496 
4497 static void
4498 mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix)
4499 {
4500 	const struct mwl_txdesc *ds = bf->bf_desc;
4501 	uint32_t status = le32toh(ds->Status);
4502 
4503 	printf("Q%u[%3u]", qnum, ix);
4504 	printf(" (DS.V:%p DS.P:0x%jx)\n", ds, (uintmax_t)bf->bf_daddr);
4505 	printf("    NEXT:%08x DATA:%08x LEN:%04x STAT:%08x%s\n",
4506 	    le32toh(ds->pPhysNext),
4507 	    le32toh(ds->PktPtr), le16toh(ds->PktLen), status,
4508 	    status & EAGLE_TXD_STATUS_USED ?
4509 		"" : (status & 3) != 0 ? " *" : " !");
4510 	printf("    RATE:%02x PRI:%x QOS:%04x SAP:%08x FORMAT:%04x\n",
4511 	    ds->DataRate, ds->TxPriority, le16toh(ds->QosCtrl),
4512 	    le32toh(ds->SapPktInfo), le16toh(ds->Format));
4513 #if MWL_TXDESC > 1
4514 	printf("    MULTIFRAMES:%u LEN:%04x %04x %04x %04x %04x %04x\n"
4515 	    , le32toh(ds->multiframes)
4516 	    , le16toh(ds->PktLenArray[0]), le16toh(ds->PktLenArray[1])
4517 	    , le16toh(ds->PktLenArray[2]), le16toh(ds->PktLenArray[3])
4518 	    , le16toh(ds->PktLenArray[4]), le16toh(ds->PktLenArray[5])
4519 	);
4520 	printf("    DATA:%08x %08x %08x %08x %08x %08x\n"
4521 	    , le32toh(ds->PktPtrArray[0]), le32toh(ds->PktPtrArray[1])
4522 	    , le32toh(ds->PktPtrArray[2]), le32toh(ds->PktPtrArray[3])
4523 	    , le32toh(ds->PktPtrArray[4]), le32toh(ds->PktPtrArray[5])
4524 	);
4525 #endif
4526 #if 0
4527 { const uint8_t *cp = (const uint8_t *) ds;
4528   int i;
4529   for (i = 0; i < sizeof(struct mwl_txdesc); i++) {
4530 	printf("%02x ", cp[i]);
4531 	if (((i+1) % 16) == 0)
4532 		printf("\n");
4533   }
4534   printf("\n");
4535 }
4536 #endif
4537 }
4538 #endif /* MWL_DEBUG */
4539 
4540 #if 0
4541 static void
4542 mwl_txq_dump(struct mwl_txq *txq)
4543 {
4544 	struct mwl_txbuf *bf;
4545 	int i = 0;
4546 
4547 	MWL_TXQ_LOCK(txq);
4548 	STAILQ_FOREACH(bf, &txq->active, bf_list) {
4549 		struct mwl_txdesc *ds = bf->bf_desc;
4550 		MWL_TXDESC_SYNC(txq, ds,
4551 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
4552 #ifdef MWL_DEBUG
4553 		mwl_printtxbuf(bf, txq->qnum, i);
4554 #endif
4555 		i++;
4556 	}
4557 	MWL_TXQ_UNLOCK(txq);
4558 }
4559 #endif
4560 
4561 static void
4562 mwl_watchdog(void *arg)
4563 {
4564 	struct mwl_softc *sc = arg;
4565 
4566 	callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
4567 	if (sc->sc_tx_timer == 0 || --sc->sc_tx_timer > 0)
4568 		return;
4569 
4570 	if (sc->sc_running && !sc->sc_invalid) {
4571 		if (mwl_hal_setkeepalive(sc->sc_mh))
4572 			device_printf(sc->sc_dev,
4573 			    "transmit timeout (firmware hung?)\n");
4574 		else
4575 			device_printf(sc->sc_dev,
4576 			    "transmit timeout\n");
4577 #if 0
4578 		mwl_reset(sc);
4579 mwl_txq_dump(&sc->sc_txq[0]);/*XXX*/
4580 #endif
4581 		counter_u64_add(sc->sc_ic.ic_oerrors, 1);
4582 		sc->sc_stats.mst_watchdog++;
4583 	}
4584 }
4585 
4586 #ifdef MWL_DIAGAPI
4587 /*
4588  * Diagnostic interface to the HAL.  This is used by various
4589  * tools to do things like retrieve register contents for
4590  * debugging.  The mechanism is intentionally opaque so that
4591  * it can change frequently w/o concern for compatibility.
4592  */
4593 static int
4594 mwl_ioctl_diag(struct mwl_softc *sc, struct mwl_diag *md)
4595 {
4596 	struct mwl_hal *mh = sc->sc_mh;
4597 	u_int id = md->md_id & MWL_DIAG_ID;
4598 	void *indata = NULL;
4599 	void *outdata = NULL;
4600 	u_int32_t insize = md->md_in_size;
4601 	u_int32_t outsize = md->md_out_size;
4602 	int error = 0;
4603 
4604 	if (md->md_id & MWL_DIAG_IN) {
4605 		/*
4606 		 * Copy in data.
4607 		 */
4608 		indata = malloc(insize, M_TEMP, M_NOWAIT);
4609 		if (indata == NULL) {
4610 			error = ENOMEM;
4611 			goto bad;
4612 		}
4613 		error = copyin(md->md_in_data, indata, insize);
4614 		if (error)
4615 			goto bad;
4616 	}
4617 	if (md->md_id & MWL_DIAG_DYN) {
4618 		/*
4619 		 * Allocate a buffer for the results (otherwise the HAL
4620 		 * returns a pointer to a buffer where we can read the
4621 		 * results).  Note that we depend on the HAL leaving this
4622 		 * pointer for us to use below in reclaiming the buffer;
4623 		 * may want to be more defensive.
4624 		 */
4625 		outdata = malloc(outsize, M_TEMP, M_NOWAIT);
4626 		if (outdata == NULL) {
4627 			error = ENOMEM;
4628 			goto bad;
4629 		}
4630 	}
4631 	if (mwl_hal_getdiagstate(mh, id, indata, insize, &outdata, &outsize)) {
4632 		if (outsize < md->md_out_size)
4633 			md->md_out_size = outsize;
4634 		if (outdata != NULL)
4635 			error = copyout(outdata, md->md_out_data,
4636 					md->md_out_size);
4637 	} else {
4638 		error = EINVAL;
4639 	}
4640 bad:
4641 	if ((md->md_id & MWL_DIAG_IN) && indata != NULL)
4642 		free(indata, M_TEMP);
4643 	if ((md->md_id & MWL_DIAG_DYN) && outdata != NULL)
4644 		free(outdata, M_TEMP);
4645 	return error;
4646 }
4647 
4648 static int
4649 mwl_ioctl_reset(struct mwl_softc *sc, struct mwl_diag *md)
4650 {
4651 	struct mwl_hal *mh = sc->sc_mh;
4652 	int error;
4653 
4654 	MWL_LOCK_ASSERT(sc);
4655 
4656 	if (md->md_id == 0 && mwl_hal_fwload(mh, NULL) != 0) {
4657 		device_printf(sc->sc_dev, "unable to load firmware\n");
4658 		return EIO;
4659 	}
4660 	if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
4661 		device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
4662 		return EIO;
4663 	}
4664 	error = mwl_setupdma(sc);
4665 	if (error != 0) {
4666 		/* NB: mwl_setupdma prints a msg */
4667 		return error;
4668 	}
4669 	/*
4670 	 * Reset tx/rx data structures; after reload we must
4671 	 * re-start the driver's notion of the next xmit/recv.
4672 	 */
4673 	mwl_draintxq(sc);		/* clear pending frames */
4674 	mwl_resettxq(sc);		/* rebuild tx q lists */
4675 	sc->sc_rxnext = NULL;		/* force rx to start at the list head */
4676 	return 0;
4677 }
4678 #endif /* MWL_DIAGAPI */
4679 
4680 static void
4681 mwl_parent(struct ieee80211com *ic)
4682 {
4683 	struct mwl_softc *sc = ic->ic_softc;
4684 	int startall = 0;
4685 
4686 	MWL_LOCK(sc);
4687 	if (ic->ic_nrunning > 0) {
4688 		if (sc->sc_running) {
4689 			/*
4690 			 * To avoid rescanning another access point,
4691 			 * do not call mwl_init() here.  Instead,
4692 			 * only reflect promisc mode settings.
4693 			 */
4694 			mwl_mode_init(sc);
4695 		} else {
4696 			/*
4697 			 * Beware of being called during attach/detach
4698 			 * to reset promiscuous mode.  In that case we
4699 			 * will still be marked UP but not RUNNING.
4700 			 * However trying to re-init the interface
4701 			 * is the wrong thing to do as we've already
4702 			 * torn down much of our state.  There's
4703 			 * probably a better way to deal with this.
4704 			 */
4705 			if (!sc->sc_invalid) {
4706 				mwl_init(sc);	/* XXX lose error */
4707 				startall = 1;
4708 			}
4709 		}
4710 	} else
4711 		mwl_stop(sc);
4712 	MWL_UNLOCK(sc);
4713 	if (startall)
4714 		ieee80211_start_all(ic);
4715 }
4716 
4717 static int
4718 mwl_ioctl(struct ieee80211com *ic, u_long cmd, void *data)
4719 {
4720 	struct mwl_softc *sc = ic->ic_softc;
4721 	struct ifreq *ifr = data;
4722 	int error = 0;
4723 
4724 	switch (cmd) {
4725 	case SIOCGMVSTATS:
4726 		mwl_hal_gethwstats(sc->sc_mh, &sc->sc_stats.hw_stats);
4727 #if 0
4728 		/* NB: embed these numbers to get a consistent view */
4729 		sc->sc_stats.mst_tx_packets =
4730 		    if_get_counter(ifp, IFCOUNTER_OPACKETS);
4731 		sc->sc_stats.mst_rx_packets =
4732 		    if_get_counter(ifp, IFCOUNTER_IPACKETS);
4733 #endif
4734 		/*
4735 		 * NB: Drop the softc lock in case of a page fault;
4736 		 * we'll accept any potential inconsisentcy in the
4737 		 * statistics.  The alternative is to copy the data
4738 		 * to a local structure.
4739 		 */
4740 		return (copyout(&sc->sc_stats, ifr_data_get_ptr(ifr),
4741 		    sizeof (sc->sc_stats)));
4742 #ifdef MWL_DIAGAPI
4743 	case SIOCGMVDIAG:
4744 		/* XXX check privs */
4745 		return mwl_ioctl_diag(sc, (struct mwl_diag *) ifr);
4746 	case SIOCGMVRESET:
4747 		/* XXX check privs */
4748 		MWL_LOCK(sc);
4749 		error = mwl_ioctl_reset(sc,(struct mwl_diag *) ifr);
4750 		MWL_UNLOCK(sc);
4751 		break;
4752 #endif /* MWL_DIAGAPI */
4753 	default:
4754 		error = ENOTTY;
4755 		break;
4756 	}
4757 	return (error);
4758 }
4759 
4760 #ifdef	MWL_DEBUG
4761 static int
4762 mwl_sysctl_debug(SYSCTL_HANDLER_ARGS)
4763 {
4764 	struct mwl_softc *sc = arg1;
4765 	int debug, error;
4766 
4767 	debug = sc->sc_debug | (mwl_hal_getdebug(sc->sc_mh) << 24);
4768 	error = sysctl_handle_int(oidp, &debug, 0, req);
4769 	if (error || !req->newptr)
4770 		return error;
4771 	mwl_hal_setdebug(sc->sc_mh, debug >> 24);
4772 	sc->sc_debug = debug & 0x00ffffff;
4773 	return 0;
4774 }
4775 #endif /* MWL_DEBUG */
4776 
4777 static void
4778 mwl_sysctlattach(struct mwl_softc *sc)
4779 {
4780 #ifdef	MWL_DEBUG
4781 	struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(sc->sc_dev);
4782 	struct sysctl_oid *tree = device_get_sysctl_tree(sc->sc_dev);
4783 
4784 	sc->sc_debug = mwl_debug;
4785 	SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(tree), OID_AUTO, "debug",
4786 	    CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_NEEDGIANT, sc, 0,
4787 	    mwl_sysctl_debug, "I", "control debugging printfs");
4788 #endif
4789 }
4790 
4791 /*
4792  * Announce various information on device/driver attach.
4793  */
4794 static void
4795 mwl_announce(struct mwl_softc *sc)
4796 {
4797 
4798 	device_printf(sc->sc_dev, "Rev A%d hardware, v%d.%d.%d.%d firmware (regioncode %d)\n",
4799 		sc->sc_hwspecs.hwVersion,
4800 		(sc->sc_hwspecs.fwReleaseNumber>>24) & 0xff,
4801 		(sc->sc_hwspecs.fwReleaseNumber>>16) & 0xff,
4802 		(sc->sc_hwspecs.fwReleaseNumber>>8) & 0xff,
4803 		(sc->sc_hwspecs.fwReleaseNumber>>0) & 0xff,
4804 		sc->sc_hwspecs.regionCode);
4805 	sc->sc_fwrelease = sc->sc_hwspecs.fwReleaseNumber;
4806 
4807 	if (bootverbose) {
4808 		int i;
4809 		for (i = 0; i <= WME_AC_VO; i++) {
4810 			struct mwl_txq *txq = sc->sc_ac2q[i];
4811 			device_printf(sc->sc_dev, "Use hw queue %u for %s traffic\n",
4812 				txq->qnum, ieee80211_wme_acnames[i]);
4813 		}
4814 	}
4815 	if (bootverbose || mwl_rxdesc != MWL_RXDESC)
4816 		device_printf(sc->sc_dev, "using %u rx descriptors\n", mwl_rxdesc);
4817 	if (bootverbose || mwl_rxbuf != MWL_RXBUF)
4818 		device_printf(sc->sc_dev, "using %u rx buffers\n", mwl_rxbuf);
4819 	if (bootverbose || mwl_txbuf != MWL_TXBUF)
4820 		device_printf(sc->sc_dev, "using %u tx buffers\n", mwl_txbuf);
4821 	if (bootverbose && mwl_hal_ismbsscapable(sc->sc_mh))
4822 		device_printf(sc->sc_dev, "multi-bss support\n");
4823 #ifdef MWL_TX_NODROP
4824 	if (bootverbose)
4825 		device_printf(sc->sc_dev, "no tx drop\n");
4826 #endif
4827 }
4828