xref: /freebsd/sys/dev/mwl/if_mwl.c (revision 2589197a)
1 /*-
2  * SPDX-License-Identifier: BSD-2-Clause
3  *
4  * Copyright (c) 2007-2009 Sam Leffler, Errno Consulting
5  * Copyright (c) 2007-2008 Marvell Semiconductor, Inc.
6  * All rights reserved.
7  *
8  * Redistribution and use in source and binary forms, with or without
9  * modification, are permitted provided that the following conditions
10  * are met:
11  * 1. Redistributions of source code must retain the above copyright
12  *    notice, this list of conditions and the following disclaimer,
13  *    without modification.
14  * 2. Redistributions in binary form must reproduce at minimum a disclaimer
15  *    similar to the "NO WARRANTY" disclaimer below ("Disclaimer") and any
16  *    redistribution must be conditioned upon including a substantially
17  *    similar Disclaimer requirement for further binary redistribution.
18  *
19  * NO WARRANTY
20  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
21  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
22  * LIMITED TO, THE IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTIBILITY
23  * AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL
24  * THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY,
25  * OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
28  * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
30  * THE POSSIBILITY OF SUCH DAMAGES.
31  */
32 
33 #include <sys/cdefs.h>
34 /*
35  * Driver for the Marvell 88W8363 Wireless LAN controller.
36  */
37 
38 #include "opt_inet.h"
39 #include "opt_mwl.h"
40 #include "opt_wlan.h"
41 
42 #include <sys/param.h>
43 #include <sys/systm.h>
44 #include <sys/sysctl.h>
45 #include <sys/mbuf.h>
46 #include <sys/malloc.h>
47 #include <sys/lock.h>
48 #include <sys/mutex.h>
49 #include <sys/kernel.h>
50 #include <sys/socket.h>
51 #include <sys/sockio.h>
52 #include <sys/errno.h>
53 #include <sys/callout.h>
54 #include <sys/bus.h>
55 #include <sys/endian.h>
56 #include <sys/kthread.h>
57 #include <sys/taskqueue.h>
58 
59 #include <machine/bus.h>
60 
61 #include <net/if.h>
62 #include <net/if_var.h>
63 #include <net/if_dl.h>
64 #include <net/if_media.h>
65 #include <net/if_types.h>
66 #include <net/if_arp.h>
67 #include <net/ethernet.h>
68 #include <net/if_llc.h>
69 
70 #include <net/bpf.h>
71 
72 #include <net80211/ieee80211_var.h>
73 #include <net80211/ieee80211_input.h>
74 #include <net80211/ieee80211_regdomain.h>
75 
76 #ifdef INET
77 #include <netinet/in.h>
78 #include <netinet/if_ether.h>
79 #endif /* INET */
80 
81 #include <dev/mwl/if_mwlvar.h>
82 #include <dev/mwl/mwldiag.h>
83 
84 static struct ieee80211vap *mwl_vap_create(struct ieee80211com *,
85 		    const char [IFNAMSIZ], int, enum ieee80211_opmode, int,
86 		    const uint8_t [IEEE80211_ADDR_LEN],
87 		    const uint8_t [IEEE80211_ADDR_LEN]);
88 static void	mwl_vap_delete(struct ieee80211vap *);
89 static int	mwl_setupdma(struct mwl_softc *);
90 static int	mwl_hal_reset(struct mwl_softc *sc);
91 static int	mwl_init(struct mwl_softc *);
92 static void	mwl_parent(struct ieee80211com *);
93 static int	mwl_reset(struct ieee80211vap *, u_long);
94 static void	mwl_stop(struct mwl_softc *);
95 static void	mwl_start(struct mwl_softc *);
96 static int	mwl_transmit(struct ieee80211com *, struct mbuf *);
97 static int	mwl_raw_xmit(struct ieee80211_node *, struct mbuf *,
98 			const struct ieee80211_bpf_params *);
99 static int	mwl_media_change(if_t);
100 static void	mwl_watchdog(void *);
101 static int	mwl_ioctl(struct ieee80211com *, u_long, void *);
102 static void	mwl_radar_proc(void *, int);
103 static void	mwl_chanswitch_proc(void *, int);
104 static void	mwl_bawatchdog_proc(void *, int);
105 static int	mwl_key_alloc(struct ieee80211vap *,
106 			struct ieee80211_key *,
107 			ieee80211_keyix *, ieee80211_keyix *);
108 static int	mwl_key_delete(struct ieee80211vap *,
109 			const struct ieee80211_key *);
110 static int	mwl_key_set(struct ieee80211vap *,
111 			const struct ieee80211_key *);
112 static int	_mwl_key_set(struct ieee80211vap *,
113 			const struct ieee80211_key *,
114 			const uint8_t mac[IEEE80211_ADDR_LEN]);
115 static int	mwl_mode_init(struct mwl_softc *);
116 static void	mwl_update_mcast(struct ieee80211com *);
117 static void	mwl_update_promisc(struct ieee80211com *);
118 static void	mwl_updateslot(struct ieee80211com *);
119 static int	mwl_beacon_setup(struct ieee80211vap *);
120 static void	mwl_beacon_update(struct ieee80211vap *, int);
121 #ifdef MWL_HOST_PS_SUPPORT
122 static void	mwl_update_ps(struct ieee80211vap *, int);
123 static int	mwl_set_tim(struct ieee80211_node *, int);
124 #endif
125 static int	mwl_dma_setup(struct mwl_softc *);
126 static void	mwl_dma_cleanup(struct mwl_softc *);
127 static struct ieee80211_node *mwl_node_alloc(struct ieee80211vap *,
128 		    const uint8_t [IEEE80211_ADDR_LEN]);
129 static void	mwl_node_cleanup(struct ieee80211_node *);
130 static void	mwl_node_drain(struct ieee80211_node *);
131 static void	mwl_node_getsignal(const struct ieee80211_node *,
132 			int8_t *, int8_t *);
133 static void	mwl_node_getmimoinfo(const struct ieee80211_node *,
134 			struct ieee80211_mimo_info *);
135 static int	mwl_rxbuf_init(struct mwl_softc *, struct mwl_rxbuf *);
136 static void	mwl_rx_proc(void *, int);
137 static void	mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *, int);
138 static int	mwl_tx_setup(struct mwl_softc *, int, int);
139 static int	mwl_wme_update(struct ieee80211com *);
140 static void	mwl_tx_cleanupq(struct mwl_softc *, struct mwl_txq *);
141 static void	mwl_tx_cleanup(struct mwl_softc *);
142 static uint16_t	mwl_calcformat(uint8_t rate, const struct ieee80211_node *);
143 static int	mwl_tx_start(struct mwl_softc *, struct ieee80211_node *,
144 			     struct mwl_txbuf *, struct mbuf *);
145 static void	mwl_tx_proc(void *, int);
146 static int	mwl_chan_set(struct mwl_softc *, struct ieee80211_channel *);
147 static void	mwl_draintxq(struct mwl_softc *);
148 static void	mwl_cleartxq(struct mwl_softc *, struct ieee80211vap *);
149 static int	mwl_recv_action(struct ieee80211_node *,
150 			const struct ieee80211_frame *,
151 			const uint8_t *, const uint8_t *);
152 static int	mwl_addba_request(struct ieee80211_node *,
153 			struct ieee80211_tx_ampdu *, int dialogtoken,
154 			int baparamset, int batimeout);
155 static int	mwl_addba_response(struct ieee80211_node *,
156 			struct ieee80211_tx_ampdu *, int status,
157 			int baparamset, int batimeout);
158 static void	mwl_addba_stop(struct ieee80211_node *,
159 			struct ieee80211_tx_ampdu *);
160 static int	mwl_startrecv(struct mwl_softc *);
161 static MWL_HAL_APMODE mwl_getapmode(const struct ieee80211vap *,
162 			struct ieee80211_channel *);
163 static int	mwl_setapmode(struct ieee80211vap *, struct ieee80211_channel*);
164 static void	mwl_scan_start(struct ieee80211com *);
165 static void	mwl_scan_end(struct ieee80211com *);
166 static void	mwl_set_channel(struct ieee80211com *);
167 static int	mwl_peerstadb(struct ieee80211_node *,
168 			int aid, int staid, MWL_HAL_PEERINFO *pi);
169 static int	mwl_localstadb(struct ieee80211vap *);
170 static int	mwl_newstate(struct ieee80211vap *, enum ieee80211_state, int);
171 static int	allocstaid(struct mwl_softc *sc, int aid);
172 static void	delstaid(struct mwl_softc *sc, int staid);
173 static void	mwl_newassoc(struct ieee80211_node *, int);
174 static void	mwl_agestations(void *);
175 static int	mwl_setregdomain(struct ieee80211com *,
176 			struct ieee80211_regdomain *, int,
177 			struct ieee80211_channel []);
178 static void	mwl_getradiocaps(struct ieee80211com *, int, int *,
179 			struct ieee80211_channel []);
180 static int	mwl_getchannels(struct mwl_softc *);
181 
182 static void	mwl_sysctlattach(struct mwl_softc *);
183 static void	mwl_announce(struct mwl_softc *);
184 
185 SYSCTL_NODE(_hw, OID_AUTO, mwl, CTLFLAG_RD | CTLFLAG_MPSAFE, 0,
186     "Marvell driver parameters");
187 
188 static	int mwl_rxdesc = MWL_RXDESC;		/* # rx desc's to allocate */
189 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdesc, CTLFLAG_RW, &mwl_rxdesc,
190 	    0, "rx descriptors allocated");
191 static	int mwl_rxbuf = MWL_RXBUF;		/* # rx buffers to allocate */
192 SYSCTL_INT(_hw_mwl, OID_AUTO, rxbuf, CTLFLAG_RWTUN, &mwl_rxbuf,
193 	    0, "rx buffers allocated");
194 static	int mwl_txbuf = MWL_TXBUF;		/* # tx buffers to allocate */
195 SYSCTL_INT(_hw_mwl, OID_AUTO, txbuf, CTLFLAG_RWTUN, &mwl_txbuf,
196 	    0, "tx buffers allocated");
197 static	int mwl_txcoalesce = 8;		/* # tx packets to q before poking f/w*/
198 SYSCTL_INT(_hw_mwl, OID_AUTO, txcoalesce, CTLFLAG_RWTUN, &mwl_txcoalesce,
199 	    0, "tx buffers to send at once");
200 static	int mwl_rxquota = MWL_RXBUF;		/* # max buffers to process */
201 SYSCTL_INT(_hw_mwl, OID_AUTO, rxquota, CTLFLAG_RWTUN, &mwl_rxquota,
202 	    0, "max rx buffers to process per interrupt");
203 static	int mwl_rxdmalow = 3;			/* # min buffers for wakeup */
204 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdmalow, CTLFLAG_RWTUN, &mwl_rxdmalow,
205 	    0, "min free rx buffers before restarting traffic");
206 
207 #ifdef MWL_DEBUG
208 static	int mwl_debug = 0;
209 SYSCTL_INT(_hw_mwl, OID_AUTO, debug, CTLFLAG_RWTUN, &mwl_debug,
210 	    0, "control debugging printfs");
211 enum {
212 	MWL_DEBUG_XMIT		= 0x00000001,	/* basic xmit operation */
213 	MWL_DEBUG_XMIT_DESC	= 0x00000002,	/* xmit descriptors */
214 	MWL_DEBUG_RECV		= 0x00000004,	/* basic recv operation */
215 	MWL_DEBUG_RECV_DESC	= 0x00000008,	/* recv descriptors */
216 	MWL_DEBUG_RESET		= 0x00000010,	/* reset processing */
217 	MWL_DEBUG_BEACON 	= 0x00000020,	/* beacon handling */
218 	MWL_DEBUG_INTR		= 0x00000040,	/* ISR */
219 	MWL_DEBUG_TX_PROC	= 0x00000080,	/* tx ISR proc */
220 	MWL_DEBUG_RX_PROC	= 0x00000100,	/* rx ISR proc */
221 	MWL_DEBUG_KEYCACHE	= 0x00000200,	/* key cache management */
222 	MWL_DEBUG_STATE		= 0x00000400,	/* 802.11 state transitions */
223 	MWL_DEBUG_NODE		= 0x00000800,	/* node management */
224 	MWL_DEBUG_RECV_ALL	= 0x00001000,	/* trace all frames (beacons) */
225 	MWL_DEBUG_TSO		= 0x00002000,	/* TSO processing */
226 	MWL_DEBUG_AMPDU		= 0x00004000,	/* BA stream handling */
227 	MWL_DEBUG_ANY		= 0xffffffff
228 };
229 #define	IS_BEACON(wh) \
230     ((wh->i_fc[0] & (IEEE80211_FC0_TYPE_MASK|IEEE80211_FC0_SUBTYPE_MASK)) == \
231 	 (IEEE80211_FC0_TYPE_MGT|IEEE80211_FC0_SUBTYPE_BEACON))
232 #define	IFF_DUMPPKTS_RECV(sc, wh) \
233     ((sc->sc_debug & MWL_DEBUG_RECV) && \
234       ((sc->sc_debug & MWL_DEBUG_RECV_ALL) || !IS_BEACON(wh)))
235 #define	IFF_DUMPPKTS_XMIT(sc) \
236 	(sc->sc_debug & MWL_DEBUG_XMIT)
237 
238 #define	DPRINTF(sc, m, fmt, ...) do {				\
239 	if (sc->sc_debug & (m))					\
240 		printf(fmt, __VA_ARGS__);			\
241 } while (0)
242 #define	KEYPRINTF(sc, hk, mac) do {				\
243 	if (sc->sc_debug & MWL_DEBUG_KEYCACHE)			\
244 		mwl_keyprint(sc, __func__, hk, mac);		\
245 } while (0)
246 static	void mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix);
247 static	void mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix);
248 #else
249 #define	IFF_DUMPPKTS_RECV(sc, wh)	0
250 #define	IFF_DUMPPKTS_XMIT(sc)		0
251 #define	DPRINTF(sc, m, fmt, ...)	do { (void )sc; } while (0)
252 #define	KEYPRINTF(sc, k, mac)		do { (void )sc; } while (0)
253 #endif
254 
255 static MALLOC_DEFINE(M_MWLDEV, "mwldev", "mwl driver dma buffers");
256 
257 /*
258  * Each packet has fixed front matter: a 2-byte length
259  * of the payload, followed by a 4-address 802.11 header
260  * (regardless of the actual header and always w/o any
261  * QoS header).  The payload then follows.
262  */
263 struct mwltxrec {
264 	uint16_t fwlen;
265 	struct ieee80211_frame_addr4 wh;
266 } __packed;
267 
268 /*
269  * Read/Write shorthands for accesses to BAR 0.  Note
270  * that all BAR 1 operations are done in the "hal" and
271  * there should be no reference to them here.
272  */
273 #ifdef MWL_DEBUG
274 static __inline uint32_t
RD4(struct mwl_softc * sc,bus_size_t off)275 RD4(struct mwl_softc *sc, bus_size_t off)
276 {
277 	return bus_space_read_4(sc->sc_io0t, sc->sc_io0h, off);
278 }
279 #endif
280 
281 static __inline void
WR4(struct mwl_softc * sc,bus_size_t off,uint32_t val)282 WR4(struct mwl_softc *sc, bus_size_t off, uint32_t val)
283 {
284 	bus_space_write_4(sc->sc_io0t, sc->sc_io0h, off, val);
285 }
286 
287 int
mwl_attach(uint16_t devid,struct mwl_softc * sc)288 mwl_attach(uint16_t devid, struct mwl_softc *sc)
289 {
290 	struct ieee80211com *ic = &sc->sc_ic;
291 	struct mwl_hal *mh;
292 	int error = 0;
293 
294 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: devid 0x%x\n", __func__, devid);
295 
296 	/*
297 	 * Setup the RX free list lock early, so it can be consistently
298 	 * removed.
299 	 */
300 	MWL_RXFREE_INIT(sc);
301 
302 	mh = mwl_hal_attach(sc->sc_dev, devid,
303 	    sc->sc_io1h, sc->sc_io1t, sc->sc_dmat);
304 	if (mh == NULL) {
305 		device_printf(sc->sc_dev, "unable to attach HAL\n");
306 		error = EIO;
307 		goto bad;
308 	}
309 	sc->sc_mh = mh;
310 	/*
311 	 * Load firmware so we can get setup.  We arbitrarily
312 	 * pick station firmware; we'll re-load firmware as
313 	 * needed so setting up the wrong mode isn't a big deal.
314 	 */
315 	if (mwl_hal_fwload(mh, NULL) != 0) {
316 		device_printf(sc->sc_dev, "unable to setup builtin firmware\n");
317 		error = EIO;
318 		goto bad1;
319 	}
320 	if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
321 		device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
322 		error = EIO;
323 		goto bad1;
324 	}
325 	error = mwl_getchannels(sc);
326 	if (error != 0)
327 		goto bad1;
328 
329 	sc->sc_txantenna = 0;		/* h/w default */
330 	sc->sc_rxantenna = 0;		/* h/w default */
331 	sc->sc_invalid = 0;		/* ready to go, enable int handling */
332 	sc->sc_ageinterval = MWL_AGEINTERVAL;
333 
334 	/*
335 	 * Allocate tx+rx descriptors and populate the lists.
336 	 * We immediately push the information to the firmware
337 	 * as otherwise it gets upset.
338 	 */
339 	error = mwl_dma_setup(sc);
340 	if (error != 0) {
341 		device_printf(sc->sc_dev, "failed to setup descriptors: %d\n",
342 		    error);
343 		goto bad1;
344 	}
345 	error = mwl_setupdma(sc);	/* push to firmware */
346 	if (error != 0)			/* NB: mwl_setupdma prints msg */
347 		goto bad1;
348 
349 	callout_init(&sc->sc_timer, 1);
350 	callout_init_mtx(&sc->sc_watchdog, &sc->sc_mtx, 0);
351 	mbufq_init(&sc->sc_snd, ifqmaxlen);
352 
353 	sc->sc_tq = taskqueue_create("mwl_taskq", M_NOWAIT,
354 		taskqueue_thread_enqueue, &sc->sc_tq);
355 	taskqueue_start_threads(&sc->sc_tq, 1, PI_NET,
356 		"%s taskq", device_get_nameunit(sc->sc_dev));
357 
358 	NET_TASK_INIT(&sc->sc_rxtask, 0, mwl_rx_proc, sc);
359 	TASK_INIT(&sc->sc_radartask, 0, mwl_radar_proc, sc);
360 	TASK_INIT(&sc->sc_chanswitchtask, 0, mwl_chanswitch_proc, sc);
361 	TASK_INIT(&sc->sc_bawatchdogtask, 0, mwl_bawatchdog_proc, sc);
362 
363 	/* NB: insure BK queue is the lowest priority h/w queue */
364 	if (!mwl_tx_setup(sc, WME_AC_BK, MWL_WME_AC_BK)) {
365 		device_printf(sc->sc_dev,
366 		    "unable to setup xmit queue for %s traffic!\n",
367 		     ieee80211_wme_acnames[WME_AC_BK]);
368 		error = EIO;
369 		goto bad2;
370 	}
371 	if (!mwl_tx_setup(sc, WME_AC_BE, MWL_WME_AC_BE) ||
372 	    !mwl_tx_setup(sc, WME_AC_VI, MWL_WME_AC_VI) ||
373 	    !mwl_tx_setup(sc, WME_AC_VO, MWL_WME_AC_VO)) {
374 		/*
375 		 * Not enough hardware tx queues to properly do WME;
376 		 * just punt and assign them all to the same h/w queue.
377 		 * We could do a better job of this if, for example,
378 		 * we allocate queues when we switch from station to
379 		 * AP mode.
380 		 */
381 		if (sc->sc_ac2q[WME_AC_VI] != NULL)
382 			mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_VI]);
383 		if (sc->sc_ac2q[WME_AC_BE] != NULL)
384 			mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_BE]);
385 		sc->sc_ac2q[WME_AC_BE] = sc->sc_ac2q[WME_AC_BK];
386 		sc->sc_ac2q[WME_AC_VI] = sc->sc_ac2q[WME_AC_BK];
387 		sc->sc_ac2q[WME_AC_VO] = sc->sc_ac2q[WME_AC_BK];
388 	}
389 	TASK_INIT(&sc->sc_txtask, 0, mwl_tx_proc, sc);
390 
391 	ic->ic_softc = sc;
392 	ic->ic_name = device_get_nameunit(sc->sc_dev);
393 	/* XXX not right but it's not used anywhere important */
394 	ic->ic_phytype = IEEE80211_T_OFDM;
395 	ic->ic_opmode = IEEE80211_M_STA;
396 	ic->ic_caps =
397 		  IEEE80211_C_STA		/* station mode supported */
398 		| IEEE80211_C_HOSTAP		/* hostap mode */
399 		| IEEE80211_C_MONITOR		/* monitor mode */
400 #if 0
401 		| IEEE80211_C_IBSS		/* ibss, nee adhoc, mode */
402 		| IEEE80211_C_AHDEMO		/* adhoc demo mode */
403 #endif
404 		| IEEE80211_C_MBSS		/* mesh point link mode */
405 		| IEEE80211_C_WDS		/* WDS supported */
406 		| IEEE80211_C_SHPREAMBLE	/* short preamble supported */
407 		| IEEE80211_C_SHSLOT		/* short slot time supported */
408 		| IEEE80211_C_WME		/* WME/WMM supported */
409 		| IEEE80211_C_BURST		/* xmit bursting supported */
410 		| IEEE80211_C_WPA		/* capable of WPA1+WPA2 */
411 		| IEEE80211_C_BGSCAN		/* capable of bg scanning */
412 		| IEEE80211_C_TXFRAG		/* handle tx frags */
413 		| IEEE80211_C_TXPMGT		/* capable of txpow mgt */
414 		| IEEE80211_C_DFS		/* DFS supported */
415 		;
416 
417 	ic->ic_htcaps =
418 		  IEEE80211_HTCAP_SMPS_ENA	/* SM PS mode enabled */
419 		| IEEE80211_HTCAP_CHWIDTH40	/* 40MHz channel width */
420 		| IEEE80211_HTCAP_SHORTGI20	/* short GI in 20MHz */
421 		| IEEE80211_HTCAP_SHORTGI40	/* short GI in 40MHz */
422 		| IEEE80211_HTCAP_RXSTBC_2STREAM/* 1-2 spatial streams */
423 #if MWL_AGGR_SIZE == 7935
424 		| IEEE80211_HTCAP_MAXAMSDU_7935	/* max A-MSDU length */
425 #else
426 		| IEEE80211_HTCAP_MAXAMSDU_3839	/* max A-MSDU length */
427 #endif
428 #if 0
429 		| IEEE80211_HTCAP_PSMP		/* PSMP supported */
430 		| IEEE80211_HTCAP_40INTOLERANT	/* 40MHz intolerant */
431 #endif
432 		/* s/w capabilities */
433 		| IEEE80211_HTC_HT		/* HT operation */
434 		| IEEE80211_HTC_AMPDU		/* tx A-MPDU */
435 		| IEEE80211_HTC_AMSDU		/* tx A-MSDU */
436 		| IEEE80211_HTC_SMPS		/* SMPS available */
437 		;
438 
439 	/*
440 	 * Mark h/w crypto support.
441 	 * XXX no way to query h/w support.
442 	 */
443 	ic->ic_cryptocaps |= IEEE80211_CRYPTO_WEP
444 			  |  IEEE80211_CRYPTO_AES_CCM
445 			  |  IEEE80211_CRYPTO_TKIP
446 			  |  IEEE80211_CRYPTO_TKIPMIC
447 			  ;
448 	/*
449 	 * Transmit requires space in the packet for a special
450 	 * format transmit record and optional padding between
451 	 * this record and the payload.  Ask the net80211 layer
452 	 * to arrange this when encapsulating packets so we can
453 	 * add it efficiently.
454 	 */
455 	ic->ic_headroom = sizeof(struct mwltxrec) -
456 		sizeof(struct ieee80211_frame);
457 
458 	IEEE80211_ADDR_COPY(ic->ic_macaddr, sc->sc_hwspecs.macAddr);
459 
460 	/* call MI attach routine. */
461 	ieee80211_ifattach(ic);
462 	ic->ic_setregdomain = mwl_setregdomain;
463 	ic->ic_getradiocaps = mwl_getradiocaps;
464 	/* override default methods */
465 	ic->ic_raw_xmit = mwl_raw_xmit;
466 	ic->ic_newassoc = mwl_newassoc;
467 	ic->ic_updateslot = mwl_updateslot;
468 	ic->ic_update_mcast = mwl_update_mcast;
469 	ic->ic_update_promisc = mwl_update_promisc;
470 	ic->ic_wme.wme_update = mwl_wme_update;
471 	ic->ic_transmit = mwl_transmit;
472 	ic->ic_ioctl = mwl_ioctl;
473 	ic->ic_parent = mwl_parent;
474 
475 	ic->ic_node_alloc = mwl_node_alloc;
476 	sc->sc_node_cleanup = ic->ic_node_cleanup;
477 	ic->ic_node_cleanup = mwl_node_cleanup;
478 	sc->sc_node_drain = ic->ic_node_drain;
479 	ic->ic_node_drain = mwl_node_drain;
480 	ic->ic_node_getsignal = mwl_node_getsignal;
481 	ic->ic_node_getmimoinfo = mwl_node_getmimoinfo;
482 
483 	ic->ic_scan_start = mwl_scan_start;
484 	ic->ic_scan_end = mwl_scan_end;
485 	ic->ic_set_channel = mwl_set_channel;
486 
487 	sc->sc_recv_action = ic->ic_recv_action;
488 	ic->ic_recv_action = mwl_recv_action;
489 	sc->sc_addba_request = ic->ic_addba_request;
490 	ic->ic_addba_request = mwl_addba_request;
491 	sc->sc_addba_response = ic->ic_addba_response;
492 	ic->ic_addba_response = mwl_addba_response;
493 	sc->sc_addba_stop = ic->ic_addba_stop;
494 	ic->ic_addba_stop = mwl_addba_stop;
495 
496 	ic->ic_vap_create = mwl_vap_create;
497 	ic->ic_vap_delete = mwl_vap_delete;
498 
499 	ieee80211_radiotap_attach(ic,
500 	    &sc->sc_tx_th.wt_ihdr, sizeof(sc->sc_tx_th),
501 		MWL_TX_RADIOTAP_PRESENT,
502 	    &sc->sc_rx_th.wr_ihdr, sizeof(sc->sc_rx_th),
503 		MWL_RX_RADIOTAP_PRESENT);
504 	/*
505 	 * Setup dynamic sysctl's now that country code and
506 	 * regdomain are available from the hal.
507 	 */
508 	mwl_sysctlattach(sc);
509 
510 	if (bootverbose)
511 		ieee80211_announce(ic);
512 	mwl_announce(sc);
513 	return 0;
514 bad2:
515 	mwl_dma_cleanup(sc);
516 bad1:
517 	mwl_hal_detach(mh);
518 bad:
519 	MWL_RXFREE_DESTROY(sc);
520 	sc->sc_invalid = 1;
521 	return error;
522 }
523 
524 int
mwl_detach(struct mwl_softc * sc)525 mwl_detach(struct mwl_softc *sc)
526 {
527 	struct ieee80211com *ic = &sc->sc_ic;
528 
529 	MWL_LOCK(sc);
530 	mwl_stop(sc);
531 	MWL_UNLOCK(sc);
532 	/*
533 	 * NB: the order of these is important:
534 	 * o call the 802.11 layer before detaching the hal to
535 	 *   insure callbacks into the driver to delete global
536 	 *   key cache entries can be handled
537 	 * o reclaim the tx queue data structures after calling
538 	 *   the 802.11 layer as we'll get called back to reclaim
539 	 *   node state and potentially want to use them
540 	 * o to cleanup the tx queues the hal is called, so detach
541 	 *   it last
542 	 * Other than that, it's straightforward...
543 	 */
544 	ieee80211_ifdetach(ic);
545 	callout_drain(&sc->sc_watchdog);
546 	mwl_dma_cleanup(sc);
547 	MWL_RXFREE_DESTROY(sc);
548 	mwl_tx_cleanup(sc);
549 	mwl_hal_detach(sc->sc_mh);
550 	mbufq_drain(&sc->sc_snd);
551 
552 	return 0;
553 }
554 
555 /*
556  * MAC address handling for multiple BSS on the same radio.
557  * The first vap uses the MAC address from the EEPROM.  For
558  * subsequent vap's we set the U/L bit (bit 1) in the MAC
559  * address and use the next six bits as an index.
560  */
561 static void
assign_address(struct mwl_softc * sc,uint8_t mac[IEEE80211_ADDR_LEN],int clone)562 assign_address(struct mwl_softc *sc, uint8_t mac[IEEE80211_ADDR_LEN], int clone)
563 {
564 	int i;
565 
566 	if (clone && mwl_hal_ismbsscapable(sc->sc_mh)) {
567 		/* NB: we only do this if h/w supports multiple bssid */
568 		for (i = 0; i < 32; i++)
569 			if ((sc->sc_bssidmask & (1<<i)) == 0)
570 				break;
571 		if (i != 0)
572 			mac[0] |= (i << 2)|0x2;
573 	} else
574 		i = 0;
575 	sc->sc_bssidmask |= 1<<i;
576 	if (i == 0)
577 		sc->sc_nbssid0++;
578 }
579 
580 static void
reclaim_address(struct mwl_softc * sc,const uint8_t mac[IEEE80211_ADDR_LEN])581 reclaim_address(struct mwl_softc *sc, const uint8_t mac[IEEE80211_ADDR_LEN])
582 {
583 	int i = mac[0] >> 2;
584 	if (i != 0 || --sc->sc_nbssid0 == 0)
585 		sc->sc_bssidmask &= ~(1<<i);
586 }
587 
588 static struct ieee80211vap *
mwl_vap_create(struct ieee80211com * ic,const char name[IFNAMSIZ],int unit,enum ieee80211_opmode opmode,int flags,const uint8_t bssid[IEEE80211_ADDR_LEN],const uint8_t mac0[IEEE80211_ADDR_LEN])589 mwl_vap_create(struct ieee80211com *ic, const char name[IFNAMSIZ], int unit,
590     enum ieee80211_opmode opmode, int flags,
591     const uint8_t bssid[IEEE80211_ADDR_LEN],
592     const uint8_t mac0[IEEE80211_ADDR_LEN])
593 {
594 	struct mwl_softc *sc = ic->ic_softc;
595 	struct mwl_hal *mh = sc->sc_mh;
596 	struct ieee80211vap *vap, *apvap;
597 	struct mwl_hal_vap *hvap;
598 	struct mwl_vap *mvp;
599 	uint8_t mac[IEEE80211_ADDR_LEN];
600 
601 	IEEE80211_ADDR_COPY(mac, mac0);
602 	switch (opmode) {
603 	case IEEE80211_M_HOSTAP:
604 	case IEEE80211_M_MBSS:
605 		if ((flags & IEEE80211_CLONE_MACADDR) == 0)
606 			assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
607 		hvap = mwl_hal_newvap(mh, MWL_HAL_AP, mac);
608 		if (hvap == NULL) {
609 			if ((flags & IEEE80211_CLONE_MACADDR) == 0)
610 				reclaim_address(sc, mac);
611 			return NULL;
612 		}
613 		break;
614 	case IEEE80211_M_STA:
615 		if ((flags & IEEE80211_CLONE_MACADDR) == 0)
616 			assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
617 		hvap = mwl_hal_newvap(mh, MWL_HAL_STA, mac);
618 		if (hvap == NULL) {
619 			if ((flags & IEEE80211_CLONE_MACADDR) == 0)
620 				reclaim_address(sc, mac);
621 			return NULL;
622 		}
623 		/* no h/w beacon miss support; always use s/w */
624 		flags |= IEEE80211_CLONE_NOBEACONS;
625 		break;
626 	case IEEE80211_M_WDS:
627 		hvap = NULL;		/* NB: we use associated AP vap */
628 		if (sc->sc_napvaps == 0)
629 			return NULL;	/* no existing AP vap */
630 		break;
631 	case IEEE80211_M_MONITOR:
632 		hvap = NULL;
633 		break;
634 	case IEEE80211_M_IBSS:
635 	case IEEE80211_M_AHDEMO:
636 	default:
637 		return NULL;
638 	}
639 
640 	mvp = malloc(sizeof(struct mwl_vap), M_80211_VAP, M_WAITOK | M_ZERO);
641 	mvp->mv_hvap = hvap;
642 	if (opmode == IEEE80211_M_WDS) {
643 		/*
644 		 * WDS vaps must have an associated AP vap; find one.
645 		 * XXX not right.
646 		 */
647 		TAILQ_FOREACH(apvap, &ic->ic_vaps, iv_next)
648 			if (apvap->iv_opmode == IEEE80211_M_HOSTAP) {
649 				mvp->mv_ap_hvap = MWL_VAP(apvap)->mv_hvap;
650 				break;
651 			}
652 		KASSERT(mvp->mv_ap_hvap != NULL, ("no ap vap"));
653 	}
654 	vap = &mvp->mv_vap;
655 	ieee80211_vap_setup(ic, vap, name, unit, opmode, flags, bssid);
656 	/* override with driver methods */
657 	mvp->mv_newstate = vap->iv_newstate;
658 	vap->iv_newstate = mwl_newstate;
659 	vap->iv_max_keyix = 0;	/* XXX */
660 	vap->iv_key_alloc = mwl_key_alloc;
661 	vap->iv_key_delete = mwl_key_delete;
662 	vap->iv_key_set = mwl_key_set;
663 #ifdef MWL_HOST_PS_SUPPORT
664 	if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS) {
665 		vap->iv_update_ps = mwl_update_ps;
666 		mvp->mv_set_tim = vap->iv_set_tim;
667 		vap->iv_set_tim = mwl_set_tim;
668 	}
669 #endif
670 	vap->iv_reset = mwl_reset;
671 	vap->iv_update_beacon = mwl_beacon_update;
672 
673 	/* override max aid so sta's cannot assoc when we're out of sta id's */
674 	vap->iv_max_aid = MWL_MAXSTAID;
675 	/* override default A-MPDU rx parameters */
676 	vap->iv_ampdu_rxmax = IEEE80211_HTCAP_MAXRXAMPDU_64K;
677 	vap->iv_ampdu_density = IEEE80211_HTCAP_MPDUDENSITY_4;
678 
679 	/* complete setup */
680 	ieee80211_vap_attach(vap, mwl_media_change, ieee80211_media_status,
681 	    mac);
682 
683 	switch (vap->iv_opmode) {
684 	case IEEE80211_M_HOSTAP:
685 	case IEEE80211_M_MBSS:
686 	case IEEE80211_M_STA:
687 		/*
688 		 * Setup sta db entry for local address.
689 		 */
690 		mwl_localstadb(vap);
691 		if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
692 		    vap->iv_opmode == IEEE80211_M_MBSS)
693 			sc->sc_napvaps++;
694 		else
695 			sc->sc_nstavaps++;
696 		break;
697 	case IEEE80211_M_WDS:
698 		sc->sc_nwdsvaps++;
699 		break;
700 	default:
701 		break;
702 	}
703 	/*
704 	 * Setup overall operating mode.
705 	 */
706 	if (sc->sc_napvaps)
707 		ic->ic_opmode = IEEE80211_M_HOSTAP;
708 	else if (sc->sc_nstavaps)
709 		ic->ic_opmode = IEEE80211_M_STA;
710 	else
711 		ic->ic_opmode = opmode;
712 
713 	return vap;
714 }
715 
716 static void
mwl_vap_delete(struct ieee80211vap * vap)717 mwl_vap_delete(struct ieee80211vap *vap)
718 {
719 	struct mwl_vap *mvp = MWL_VAP(vap);
720 	struct mwl_softc *sc = vap->iv_ic->ic_softc;
721 	struct mwl_hal *mh = sc->sc_mh;
722 	struct mwl_hal_vap *hvap = mvp->mv_hvap;
723 	enum ieee80211_opmode opmode = vap->iv_opmode;
724 
725 	/* XXX disallow ap vap delete if WDS still present */
726 	if (sc->sc_running) {
727 		/* quiesce h/w while we remove the vap */
728 		mwl_hal_intrset(mh, 0);		/* disable interrupts */
729 	}
730 	ieee80211_vap_detach(vap);
731 	switch (opmode) {
732 	case IEEE80211_M_HOSTAP:
733 	case IEEE80211_M_MBSS:
734 	case IEEE80211_M_STA:
735 		KASSERT(hvap != NULL, ("no hal vap handle"));
736 		(void) mwl_hal_delstation(hvap, vap->iv_myaddr);
737 		mwl_hal_delvap(hvap);
738 		if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS)
739 			sc->sc_napvaps--;
740 		else
741 			sc->sc_nstavaps--;
742 		/* XXX don't do it for IEEE80211_CLONE_MACADDR */
743 		reclaim_address(sc, vap->iv_myaddr);
744 		break;
745 	case IEEE80211_M_WDS:
746 		sc->sc_nwdsvaps--;
747 		break;
748 	default:
749 		break;
750 	}
751 	mwl_cleartxq(sc, vap);
752 	free(mvp, M_80211_VAP);
753 	if (sc->sc_running)
754 		mwl_hal_intrset(mh, sc->sc_imask);
755 }
756 
757 void
mwl_suspend(struct mwl_softc * sc)758 mwl_suspend(struct mwl_softc *sc)
759 {
760 
761 	MWL_LOCK(sc);
762 	mwl_stop(sc);
763 	MWL_UNLOCK(sc);
764 }
765 
766 void
mwl_resume(struct mwl_softc * sc)767 mwl_resume(struct mwl_softc *sc)
768 {
769 	int error = EDOOFUS;
770 
771 	MWL_LOCK(sc);
772 	if (sc->sc_ic.ic_nrunning > 0)
773 		error = mwl_init(sc);
774 	MWL_UNLOCK(sc);
775 
776 	if (error == 0)
777 		ieee80211_start_all(&sc->sc_ic);	/* start all vap's */
778 }
779 
780 void
mwl_shutdown(void * arg)781 mwl_shutdown(void *arg)
782 {
783 	struct mwl_softc *sc = arg;
784 
785 	MWL_LOCK(sc);
786 	mwl_stop(sc);
787 	MWL_UNLOCK(sc);
788 }
789 
790 /*
791  * Interrupt handler.  Most of the actual processing is deferred.
792  */
793 void
mwl_intr(void * arg)794 mwl_intr(void *arg)
795 {
796 	struct mwl_softc *sc = arg;
797 	struct mwl_hal *mh = sc->sc_mh;
798 	uint32_t status;
799 
800 	if (sc->sc_invalid) {
801 		/*
802 		 * The hardware is not ready/present, don't touch anything.
803 		 * Note this can happen early on if the IRQ is shared.
804 		 */
805 		DPRINTF(sc, MWL_DEBUG_ANY, "%s: invalid; ignored\n", __func__);
806 		return;
807 	}
808 	/*
809 	 * Figure out the reason(s) for the interrupt.
810 	 */
811 	mwl_hal_getisr(mh, &status);		/* NB: clears ISR too */
812 	if (status == 0)			/* must be a shared irq */
813 		return;
814 
815 	DPRINTF(sc, MWL_DEBUG_INTR, "%s: status 0x%x imask 0x%x\n",
816 	    __func__, status, sc->sc_imask);
817 	if (status & MACREG_A2HRIC_BIT_RX_RDY)
818 		taskqueue_enqueue(sc->sc_tq, &sc->sc_rxtask);
819 	if (status & MACREG_A2HRIC_BIT_TX_DONE)
820 		taskqueue_enqueue(sc->sc_tq, &sc->sc_txtask);
821 	if (status & MACREG_A2HRIC_BIT_BA_WATCHDOG)
822 		taskqueue_enqueue(sc->sc_tq, &sc->sc_bawatchdogtask);
823 	if (status & MACREG_A2HRIC_BIT_OPC_DONE)
824 		mwl_hal_cmddone(mh);
825 	if (status & MACREG_A2HRIC_BIT_MAC_EVENT) {
826 		;
827 	}
828 	if (status & MACREG_A2HRIC_BIT_ICV_ERROR) {
829 		/* TKIP ICV error */
830 		sc->sc_stats.mst_rx_badtkipicv++;
831 	}
832 	if (status & MACREG_A2HRIC_BIT_QUEUE_EMPTY) {
833 		/* 11n aggregation queue is empty, re-fill */
834 		;
835 	}
836 	if (status & MACREG_A2HRIC_BIT_QUEUE_FULL) {
837 		;
838 	}
839 	if (status & MACREG_A2HRIC_BIT_RADAR_DETECT) {
840 		/* radar detected, process event */
841 		taskqueue_enqueue(sc->sc_tq, &sc->sc_radartask);
842 	}
843 	if (status & MACREG_A2HRIC_BIT_CHAN_SWITCH) {
844 		/* DFS channel switch */
845 		taskqueue_enqueue(sc->sc_tq, &sc->sc_chanswitchtask);
846 	}
847 }
848 
849 static void
mwl_radar_proc(void * arg,int pending)850 mwl_radar_proc(void *arg, int pending)
851 {
852 	struct mwl_softc *sc = arg;
853 	struct ieee80211com *ic = &sc->sc_ic;
854 
855 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: radar detected, pending %u\n",
856 	    __func__, pending);
857 
858 	sc->sc_stats.mst_radardetect++;
859 	/* XXX stop h/w BA streams? */
860 
861 	IEEE80211_LOCK(ic);
862 	ieee80211_dfs_notify_radar(ic, ic->ic_curchan);
863 	IEEE80211_UNLOCK(ic);
864 }
865 
866 static void
mwl_chanswitch_proc(void * arg,int pending)867 mwl_chanswitch_proc(void *arg, int pending)
868 {
869 	struct mwl_softc *sc = arg;
870 	struct ieee80211com *ic = &sc->sc_ic;
871 
872 	DPRINTF(sc, MWL_DEBUG_ANY, "%s: channel switch notice, pending %u\n",
873 	    __func__, pending);
874 
875 	IEEE80211_LOCK(ic);
876 	sc->sc_csapending = 0;
877 	ieee80211_csa_completeswitch(ic);
878 	IEEE80211_UNLOCK(ic);
879 }
880 
881 static void
mwl_bawatchdog(const MWL_HAL_BASTREAM * sp)882 mwl_bawatchdog(const MWL_HAL_BASTREAM *sp)
883 {
884 	struct ieee80211_node *ni = sp->data[0];
885 
886 	/* send DELBA and drop the stream */
887 	ieee80211_ampdu_stop(ni, sp->data[1], IEEE80211_REASON_UNSPECIFIED);
888 }
889 
890 static void
mwl_bawatchdog_proc(void * arg,int pending)891 mwl_bawatchdog_proc(void *arg, int pending)
892 {
893 	struct mwl_softc *sc = arg;
894 	struct mwl_hal *mh = sc->sc_mh;
895 	const MWL_HAL_BASTREAM *sp;
896 	uint8_t bitmap, n;
897 
898 	sc->sc_stats.mst_bawatchdog++;
899 
900 	if (mwl_hal_getwatchdogbitmap(mh, &bitmap) != 0) {
901 		DPRINTF(sc, MWL_DEBUG_AMPDU,
902 		    "%s: could not get bitmap\n", __func__);
903 		sc->sc_stats.mst_bawatchdog_failed++;
904 		return;
905 	}
906 	DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: bitmap 0x%x\n", __func__, bitmap);
907 	if (bitmap == 0xff) {
908 		n = 0;
909 		/* disable all ba streams */
910 		for (bitmap = 0; bitmap < 8; bitmap++) {
911 			sp = mwl_hal_bastream_lookup(mh, bitmap);
912 			if (sp != NULL) {
913 				mwl_bawatchdog(sp);
914 				n++;
915 			}
916 		}
917 		if (n == 0) {
918 			DPRINTF(sc, MWL_DEBUG_AMPDU,
919 			    "%s: no BA streams found\n", __func__);
920 			sc->sc_stats.mst_bawatchdog_empty++;
921 		}
922 	} else if (bitmap != 0xaa) {
923 		/* disable a single ba stream */
924 		sp = mwl_hal_bastream_lookup(mh, bitmap);
925 		if (sp != NULL) {
926 			mwl_bawatchdog(sp);
927 		} else {
928 			DPRINTF(sc, MWL_DEBUG_AMPDU,
929 			    "%s: no BA stream %d\n", __func__, bitmap);
930 			sc->sc_stats.mst_bawatchdog_notfound++;
931 		}
932 	}
933 }
934 
935 /*
936  * Convert net80211 channel to a HAL channel.
937  */
938 static void
mwl_mapchan(MWL_HAL_CHANNEL * hc,const struct ieee80211_channel * chan)939 mwl_mapchan(MWL_HAL_CHANNEL *hc, const struct ieee80211_channel *chan)
940 {
941 	hc->channel = chan->ic_ieee;
942 
943 	*(uint32_t *)&hc->channelFlags = 0;
944 	if (IEEE80211_IS_CHAN_2GHZ(chan))
945 		hc->channelFlags.FreqBand = MWL_FREQ_BAND_2DOT4GHZ;
946 	else if (IEEE80211_IS_CHAN_5GHZ(chan))
947 		hc->channelFlags.FreqBand = MWL_FREQ_BAND_5GHZ;
948 	if (IEEE80211_IS_CHAN_HT40(chan)) {
949 		hc->channelFlags.ChnlWidth = MWL_CH_40_MHz_WIDTH;
950 		if (IEEE80211_IS_CHAN_HT40U(chan))
951 			hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_ABOVE_CTRL_CH;
952 		else
953 			hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_BELOW_CTRL_CH;
954 	} else
955 		hc->channelFlags.ChnlWidth = MWL_CH_20_MHz_WIDTH;
956 	/* XXX 10MHz channels */
957 }
958 
959 /*
960  * Inform firmware of our tx/rx dma setup.  The BAR 0
961  * writes below are for compatibility with older firmware.
962  * For current firmware we send this information with a
963  * cmd block via mwl_hal_sethwdma.
964  */
965 static int
mwl_setupdma(struct mwl_softc * sc)966 mwl_setupdma(struct mwl_softc *sc)
967 {
968 	int error, i;
969 
970 	sc->sc_hwdma.rxDescRead = sc->sc_rxdma.dd_desc_paddr;
971 	WR4(sc, sc->sc_hwspecs.rxDescRead, sc->sc_hwdma.rxDescRead);
972 	WR4(sc, sc->sc_hwspecs.rxDescWrite, sc->sc_hwdma.rxDescRead);
973 
974 	for (i = 0; i < MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES; i++) {
975 		struct mwl_txq *txq = &sc->sc_txq[i];
976 		sc->sc_hwdma.wcbBase[i] = txq->dma.dd_desc_paddr;
977 		WR4(sc, sc->sc_hwspecs.wcbBase[i], sc->sc_hwdma.wcbBase[i]);
978 	}
979 	sc->sc_hwdma.maxNumTxWcb = mwl_txbuf;
980 	sc->sc_hwdma.maxNumWCB = MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES;
981 
982 	error = mwl_hal_sethwdma(sc->sc_mh, &sc->sc_hwdma);
983 	if (error != 0) {
984 		device_printf(sc->sc_dev,
985 		    "unable to setup tx/rx dma; hal status %u\n", error);
986 		/* XXX */
987 	}
988 	return error;
989 }
990 
991 /*
992  * Inform firmware of tx rate parameters.
993  * Called after a channel change.
994  */
995 static int
mwl_setcurchanrates(struct mwl_softc * sc)996 mwl_setcurchanrates(struct mwl_softc *sc)
997 {
998 	struct ieee80211com *ic = &sc->sc_ic;
999 	const struct ieee80211_rateset *rs;
1000 	MWL_HAL_TXRATE rates;
1001 
1002 	memset(&rates, 0, sizeof(rates));
1003 	rs = ieee80211_get_suprates(ic, ic->ic_curchan);
1004 	/* rate used to send management frames */
1005 	rates.MgtRate = rs->rs_rates[0] & IEEE80211_RATE_VAL;
1006 	/* rate used to send multicast frames */
1007 	rates.McastRate = rates.MgtRate;
1008 
1009 	return mwl_hal_settxrate_auto(sc->sc_mh, &rates);
1010 }
1011 
1012 /*
1013  * Inform firmware of tx rate parameters.  Called whenever
1014  * user-settable params change and after a channel change.
1015  */
1016 static int
mwl_setrates(struct ieee80211vap * vap)1017 mwl_setrates(struct ieee80211vap *vap)
1018 {
1019 	struct mwl_vap *mvp = MWL_VAP(vap);
1020 	struct ieee80211_node *ni = vap->iv_bss;
1021 	const struct ieee80211_txparam *tp = ni->ni_txparms;
1022 	MWL_HAL_TXRATE rates;
1023 
1024 	KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1025 
1026 	/*
1027 	 * Update the h/w rate map.
1028 	 * NB: 0x80 for MCS is passed through unchanged
1029 	 */
1030 	memset(&rates, 0, sizeof(rates));
1031 	/* rate used to send management frames */
1032 	rates.MgtRate = tp->mgmtrate;
1033 	/* rate used to send multicast frames */
1034 	rates.McastRate = tp->mcastrate;
1035 
1036 	/* while here calculate EAPOL fixed rate cookie */
1037 	mvp->mv_eapolformat = htole16(mwl_calcformat(rates.MgtRate, ni));
1038 
1039 	return mwl_hal_settxrate(mvp->mv_hvap,
1040 	    tp->ucastrate != IEEE80211_FIXED_RATE_NONE ?
1041 		RATE_FIXED : RATE_AUTO, &rates);
1042 }
1043 
1044 /*
1045  * Setup a fixed xmit rate cookie for EAPOL frames.
1046  */
1047 static void
mwl_seteapolformat(struct ieee80211vap * vap)1048 mwl_seteapolformat(struct ieee80211vap *vap)
1049 {
1050 	struct mwl_vap *mvp = MWL_VAP(vap);
1051 	struct ieee80211_node *ni = vap->iv_bss;
1052 	enum ieee80211_phymode mode;
1053 	uint8_t rate;
1054 
1055 	KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1056 
1057 	mode = ieee80211_chan2mode(ni->ni_chan);
1058 	/*
1059 	 * Use legacy rates when operating a mixed HT+non-HT bss.
1060 	 * NB: this may violate POLA for sta and wds vap's.
1061 	 */
1062 	if (mode == IEEE80211_MODE_11NA &&
1063 	    (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1064 		rate = vap->iv_txparms[IEEE80211_MODE_11A].mgmtrate;
1065 	else if (mode == IEEE80211_MODE_11NG &&
1066 	    (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1067 		rate = vap->iv_txparms[IEEE80211_MODE_11G].mgmtrate;
1068 	else
1069 		rate = vap->iv_txparms[mode].mgmtrate;
1070 
1071 	mvp->mv_eapolformat = htole16(mwl_calcformat(rate, ni));
1072 }
1073 
1074 /*
1075  * Map SKU+country code to region code for radar bin'ing.
1076  */
1077 static int
mwl_map2regioncode(const struct ieee80211_regdomain * rd)1078 mwl_map2regioncode(const struct ieee80211_regdomain *rd)
1079 {
1080 	switch (rd->regdomain) {
1081 	case SKU_FCC:
1082 	case SKU_FCC3:
1083 		return DOMAIN_CODE_FCC;
1084 	case SKU_CA:
1085 		return DOMAIN_CODE_IC;
1086 	case SKU_ETSI:
1087 	case SKU_ETSI2:
1088 	case SKU_ETSI3:
1089 		if (rd->country == CTRY_SPAIN)
1090 			return DOMAIN_CODE_SPAIN;
1091 		if (rd->country == CTRY_FRANCE || rd->country == CTRY_FRANCE2)
1092 			return DOMAIN_CODE_FRANCE;
1093 		/* XXX force 1.3.1 radar type */
1094 		return DOMAIN_CODE_ETSI_131;
1095 	case SKU_JAPAN:
1096 		return DOMAIN_CODE_MKK;
1097 	case SKU_ROW:
1098 		return DOMAIN_CODE_DGT;	/* Taiwan */
1099 	case SKU_APAC:
1100 	case SKU_APAC2:
1101 	case SKU_APAC3:
1102 		return DOMAIN_CODE_AUS;	/* Australia */
1103 	}
1104 	/* XXX KOREA? */
1105 	return DOMAIN_CODE_FCC;			/* XXX? */
1106 }
1107 
1108 static int
mwl_hal_reset(struct mwl_softc * sc)1109 mwl_hal_reset(struct mwl_softc *sc)
1110 {
1111 	struct ieee80211com *ic = &sc->sc_ic;
1112 	struct mwl_hal *mh = sc->sc_mh;
1113 
1114 	mwl_hal_setantenna(mh, WL_ANTENNATYPE_RX, sc->sc_rxantenna);
1115 	mwl_hal_setantenna(mh, WL_ANTENNATYPE_TX, sc->sc_txantenna);
1116 	mwl_hal_setradio(mh, 1, WL_AUTO_PREAMBLE);
1117 	mwl_hal_setwmm(sc->sc_mh, (ic->ic_flags & IEEE80211_F_WME) != 0);
1118 	mwl_chan_set(sc, ic->ic_curchan);
1119 	/* NB: RF/RA performance tuned for indoor mode */
1120 	mwl_hal_setrateadaptmode(mh, 0);
1121 	mwl_hal_setoptimizationlevel(mh,
1122 	    (ic->ic_flags & IEEE80211_F_BURST) != 0);
1123 
1124 	mwl_hal_setregioncode(mh, mwl_map2regioncode(&ic->ic_regdomain));
1125 
1126 	mwl_hal_setaggampduratemode(mh, 1, 80);		/* XXX */
1127 	mwl_hal_setcfend(mh, 0);			/* XXX */
1128 
1129 	return 1;
1130 }
1131 
1132 static int
mwl_init(struct mwl_softc * sc)1133 mwl_init(struct mwl_softc *sc)
1134 {
1135 	struct mwl_hal *mh = sc->sc_mh;
1136 	int error = 0;
1137 
1138 	MWL_LOCK_ASSERT(sc);
1139 
1140 	/*
1141 	 * Stop anything previously setup.  This is safe
1142 	 * whether this is the first time through or not.
1143 	 */
1144 	mwl_stop(sc);
1145 
1146 	/*
1147 	 * Push vap-independent state to the firmware.
1148 	 */
1149 	if (!mwl_hal_reset(sc)) {
1150 		device_printf(sc->sc_dev, "unable to reset hardware\n");
1151 		return EIO;
1152 	}
1153 
1154 	/*
1155 	 * Setup recv (once); transmit is already good to go.
1156 	 */
1157 	error = mwl_startrecv(sc);
1158 	if (error != 0) {
1159 		device_printf(sc->sc_dev, "unable to start recv logic\n");
1160 		return error;
1161 	}
1162 
1163 	/*
1164 	 * Enable interrupts.
1165 	 */
1166 	sc->sc_imask = MACREG_A2HRIC_BIT_RX_RDY
1167 		     | MACREG_A2HRIC_BIT_TX_DONE
1168 		     | MACREG_A2HRIC_BIT_OPC_DONE
1169 #if 0
1170 		     | MACREG_A2HRIC_BIT_MAC_EVENT
1171 #endif
1172 		     | MACREG_A2HRIC_BIT_ICV_ERROR
1173 		     | MACREG_A2HRIC_BIT_RADAR_DETECT
1174 		     | MACREG_A2HRIC_BIT_CHAN_SWITCH
1175 #if 0
1176 		     | MACREG_A2HRIC_BIT_QUEUE_EMPTY
1177 #endif
1178 		     | MACREG_A2HRIC_BIT_BA_WATCHDOG
1179 		     | MACREQ_A2HRIC_BIT_TX_ACK
1180 		     ;
1181 
1182 	sc->sc_running = 1;
1183 	mwl_hal_intrset(mh, sc->sc_imask);
1184 	callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
1185 
1186 	return 0;
1187 }
1188 
1189 static void
mwl_stop(struct mwl_softc * sc)1190 mwl_stop(struct mwl_softc *sc)
1191 {
1192 
1193 	MWL_LOCK_ASSERT(sc);
1194 	if (sc->sc_running) {
1195 		/*
1196 		 * Shutdown the hardware and driver.
1197 		 */
1198 		sc->sc_running = 0;
1199 		callout_stop(&sc->sc_watchdog);
1200 		sc->sc_tx_timer = 0;
1201 		mwl_draintxq(sc);
1202 	}
1203 }
1204 
1205 static int
mwl_reset_vap(struct ieee80211vap * vap,int state)1206 mwl_reset_vap(struct ieee80211vap *vap, int state)
1207 {
1208 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1209 	struct ieee80211com *ic = vap->iv_ic;
1210 
1211 	if (state == IEEE80211_S_RUN)
1212 		mwl_setrates(vap);
1213 	/* XXX off by 1? */
1214 	mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
1215 	/* XXX auto? 20/40 split? */
1216 	mwl_hal_sethtgi(hvap, (vap->iv_flags_ht &
1217 	    (IEEE80211_FHT_SHORTGI20|IEEE80211_FHT_SHORTGI40)) ? 1 : 0);
1218 	mwl_hal_setnprot(hvap, ic->ic_htprotmode == IEEE80211_PROT_NONE ?
1219 	    HTPROTECT_NONE : HTPROTECT_AUTO);
1220 	/* XXX txpower cap */
1221 
1222 	/* re-setup beacons */
1223 	if (state == IEEE80211_S_RUN &&
1224 	    (vap->iv_opmode == IEEE80211_M_HOSTAP ||
1225 	     vap->iv_opmode == IEEE80211_M_MBSS ||
1226 	     vap->iv_opmode == IEEE80211_M_IBSS)) {
1227 		mwl_setapmode(vap, vap->iv_bss->ni_chan);
1228 		mwl_hal_setnprotmode(hvap, _IEEE80211_MASKSHIFT(
1229 		    ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1230 		return mwl_beacon_setup(vap);
1231 	}
1232 	return 0;
1233 }
1234 
1235 /*
1236  * Reset the hardware w/o losing operational state.
1237  * Used to reset or reload hardware state for a vap.
1238  */
1239 static int
mwl_reset(struct ieee80211vap * vap,u_long cmd)1240 mwl_reset(struct ieee80211vap *vap, u_long cmd)
1241 {
1242 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1243 	int error = 0;
1244 
1245 	if (hvap != NULL) {			/* WDS, MONITOR, etc. */
1246 		struct ieee80211com *ic = vap->iv_ic;
1247 		struct mwl_softc *sc = ic->ic_softc;
1248 		struct mwl_hal *mh = sc->sc_mh;
1249 
1250 		/* XXX handle DWDS sta vap change */
1251 		/* XXX do we need to disable interrupts? */
1252 		mwl_hal_intrset(mh, 0);		/* disable interrupts */
1253 		error = mwl_reset_vap(vap, vap->iv_state);
1254 		mwl_hal_intrset(mh, sc->sc_imask);
1255 	}
1256 	return error;
1257 }
1258 
1259 /*
1260  * Allocate a tx buffer for sending a frame.  The
1261  * packet is assumed to have the WME AC stored so
1262  * we can use it to select the appropriate h/w queue.
1263  */
1264 static struct mwl_txbuf *
mwl_gettxbuf(struct mwl_softc * sc,struct mwl_txq * txq)1265 mwl_gettxbuf(struct mwl_softc *sc, struct mwl_txq *txq)
1266 {
1267 	struct mwl_txbuf *bf;
1268 
1269 	/*
1270 	 * Grab a TX buffer and associated resources.
1271 	 */
1272 	MWL_TXQ_LOCK(txq);
1273 	bf = STAILQ_FIRST(&txq->free);
1274 	if (bf != NULL) {
1275 		STAILQ_REMOVE_HEAD(&txq->free, bf_list);
1276 		txq->nfree--;
1277 	}
1278 	MWL_TXQ_UNLOCK(txq);
1279 	if (bf == NULL)
1280 		DPRINTF(sc, MWL_DEBUG_XMIT,
1281 		    "%s: out of xmit buffers on q %d\n", __func__, txq->qnum);
1282 	return bf;
1283 }
1284 
1285 /*
1286  * Return a tx buffer to the queue it came from.  Note there
1287  * are two cases because we must preserve the order of buffers
1288  * as it reflects the fixed order of descriptors in memory
1289  * (the firmware pre-fetches descriptors so we cannot reorder).
1290  */
1291 static void
mwl_puttxbuf_head(struct mwl_txq * txq,struct mwl_txbuf * bf)1292 mwl_puttxbuf_head(struct mwl_txq *txq, struct mwl_txbuf *bf)
1293 {
1294 	bf->bf_m = NULL;
1295 	bf->bf_node = NULL;
1296 	MWL_TXQ_LOCK(txq);
1297 	STAILQ_INSERT_HEAD(&txq->free, bf, bf_list);
1298 	txq->nfree++;
1299 	MWL_TXQ_UNLOCK(txq);
1300 }
1301 
1302 static void
mwl_puttxbuf_tail(struct mwl_txq * txq,struct mwl_txbuf * bf)1303 mwl_puttxbuf_tail(struct mwl_txq *txq, struct mwl_txbuf *bf)
1304 {
1305 	bf->bf_m = NULL;
1306 	bf->bf_node = NULL;
1307 	MWL_TXQ_LOCK(txq);
1308 	STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
1309 	txq->nfree++;
1310 	MWL_TXQ_UNLOCK(txq);
1311 }
1312 
1313 static int
mwl_transmit(struct ieee80211com * ic,struct mbuf * m)1314 mwl_transmit(struct ieee80211com *ic, struct mbuf *m)
1315 {
1316 	struct mwl_softc *sc = ic->ic_softc;
1317 	int error;
1318 
1319 	MWL_LOCK(sc);
1320 	if (!sc->sc_running) {
1321 		MWL_UNLOCK(sc);
1322 		return (ENXIO);
1323 	}
1324 	error = mbufq_enqueue(&sc->sc_snd, m);
1325 	if (error) {
1326 		MWL_UNLOCK(sc);
1327 		return (error);
1328 	}
1329 	mwl_start(sc);
1330 	MWL_UNLOCK(sc);
1331 	return (0);
1332 }
1333 
1334 static void
mwl_start(struct mwl_softc * sc)1335 mwl_start(struct mwl_softc *sc)
1336 {
1337 	struct ieee80211_node *ni;
1338 	struct mwl_txbuf *bf;
1339 	struct mbuf *m;
1340 	struct mwl_txq *txq = NULL;	/* XXX silence gcc */
1341 	int nqueued;
1342 
1343 	MWL_LOCK_ASSERT(sc);
1344 	if (!sc->sc_running || sc->sc_invalid)
1345 		return;
1346 	nqueued = 0;
1347 	while ((m = mbufq_dequeue(&sc->sc_snd)) != NULL) {
1348 		/*
1349 		 * Grab the node for the destination.
1350 		 */
1351 		ni = (struct ieee80211_node *) m->m_pkthdr.rcvif;
1352 		KASSERT(ni != NULL, ("no node"));
1353 		m->m_pkthdr.rcvif = NULL;	/* committed, clear ref */
1354 		/*
1355 		 * Grab a TX buffer and associated resources.
1356 		 * We honor the classification by the 802.11 layer.
1357 		 */
1358 		txq = sc->sc_ac2q[M_WME_GETAC(m)];
1359 		bf = mwl_gettxbuf(sc, txq);
1360 		if (bf == NULL) {
1361 			m_freem(m);
1362 			ieee80211_free_node(ni);
1363 #ifdef MWL_TX_NODROP
1364 			sc->sc_stats.mst_tx_qstop++;
1365 			break;
1366 #else
1367 			DPRINTF(sc, MWL_DEBUG_XMIT,
1368 			    "%s: tail drop on q %d\n", __func__, txq->qnum);
1369 			sc->sc_stats.mst_tx_qdrop++;
1370 			continue;
1371 #endif /* MWL_TX_NODROP */
1372 		}
1373 
1374 		/*
1375 		 * Pass the frame to the h/w for transmission.
1376 		 */
1377 		if (mwl_tx_start(sc, ni, bf, m)) {
1378 			if_inc_counter(ni->ni_vap->iv_ifp,
1379 			    IFCOUNTER_OERRORS, 1);
1380 			mwl_puttxbuf_head(txq, bf);
1381 			ieee80211_free_node(ni);
1382 			continue;
1383 		}
1384 		nqueued++;
1385 		if (nqueued >= mwl_txcoalesce) {
1386 			/*
1387 			 * Poke the firmware to process queued frames;
1388 			 * see below about (lack of) locking.
1389 			 */
1390 			nqueued = 0;
1391 			mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1392 		}
1393 	}
1394 	if (nqueued) {
1395 		/*
1396 		 * NB: We don't need to lock against tx done because
1397 		 * this just prods the firmware to check the transmit
1398 		 * descriptors.  The firmware will also start fetching
1399 		 * descriptors by itself if it notices new ones are
1400 		 * present when it goes to deliver a tx done interrupt
1401 		 * to the host. So if we race with tx done processing
1402 		 * it's ok.  Delivering the kick here rather than in
1403 		 * mwl_tx_start is an optimization to avoid poking the
1404 		 * firmware for each packet.
1405 		 *
1406 		 * NB: the queue id isn't used so 0 is ok.
1407 		 */
1408 		mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1409 	}
1410 }
1411 
1412 static int
mwl_raw_xmit(struct ieee80211_node * ni,struct mbuf * m,const struct ieee80211_bpf_params * params)1413 mwl_raw_xmit(struct ieee80211_node *ni, struct mbuf *m,
1414 	const struct ieee80211_bpf_params *params)
1415 {
1416 	struct ieee80211com *ic = ni->ni_ic;
1417 	struct mwl_softc *sc = ic->ic_softc;
1418 	struct mwl_txbuf *bf;
1419 	struct mwl_txq *txq;
1420 
1421 	if (!sc->sc_running || sc->sc_invalid) {
1422 		m_freem(m);
1423 		return ENETDOWN;
1424 	}
1425 	/*
1426 	 * Grab a TX buffer and associated resources.
1427 	 * Note that we depend on the classification
1428 	 * by the 802.11 layer to get to the right h/w
1429 	 * queue.  Management frames must ALWAYS go on
1430 	 * queue 1 but we cannot just force that here
1431 	 * because we may receive non-mgt frames.
1432 	 */
1433 	txq = sc->sc_ac2q[M_WME_GETAC(m)];
1434 	bf = mwl_gettxbuf(sc, txq);
1435 	if (bf == NULL) {
1436 		sc->sc_stats.mst_tx_qstop++;
1437 		m_freem(m);
1438 		return ENOBUFS;
1439 	}
1440 	/*
1441 	 * Pass the frame to the h/w for transmission.
1442 	 */
1443 	if (mwl_tx_start(sc, ni, bf, m)) {
1444 		mwl_puttxbuf_head(txq, bf);
1445 
1446 		return EIO;		/* XXX */
1447 	}
1448 	/*
1449 	 * NB: We don't need to lock against tx done because
1450 	 * this just prods the firmware to check the transmit
1451 	 * descriptors.  The firmware will also start fetching
1452 	 * descriptors by itself if it notices new ones are
1453 	 * present when it goes to deliver a tx done interrupt
1454 	 * to the host. So if we race with tx done processing
1455 	 * it's ok.  Delivering the kick here rather than in
1456 	 * mwl_tx_start is an optimization to avoid poking the
1457 	 * firmware for each packet.
1458 	 *
1459 	 * NB: the queue id isn't used so 0 is ok.
1460 	 */
1461 	mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1462 	return 0;
1463 }
1464 
1465 static int
mwl_media_change(if_t ifp)1466 mwl_media_change(if_t ifp)
1467 {
1468 	struct ieee80211vap *vap;
1469 	int error;
1470 
1471 	/* NB: only the fixed rate can change and that doesn't need a reset */
1472 	error = ieee80211_media_change(ifp);
1473 	if (error != 0)
1474 		return (error);
1475 
1476 	vap = if_getsoftc(ifp);
1477 	mwl_setrates(vap);
1478 	return (0);
1479 }
1480 
1481 #ifdef MWL_DEBUG
1482 static void
mwl_keyprint(struct mwl_softc * sc,const char * tag,const MWL_HAL_KEYVAL * hk,const uint8_t mac[IEEE80211_ADDR_LEN])1483 mwl_keyprint(struct mwl_softc *sc, const char *tag,
1484 	const MWL_HAL_KEYVAL *hk, const uint8_t mac[IEEE80211_ADDR_LEN])
1485 {
1486 	static const char *ciphers[] = {
1487 		"WEP",
1488 		"TKIP",
1489 		"AES-CCM",
1490 	};
1491 	int i, n;
1492 
1493 	printf("%s: [%u] %-7s", tag, hk->keyIndex, ciphers[hk->keyTypeId]);
1494 	for (i = 0, n = hk->keyLen; i < n; i++)
1495 		printf(" %02x", hk->key.aes[i]);
1496 	printf(" mac %s", ether_sprintf(mac));
1497 	if (hk->keyTypeId == KEY_TYPE_ID_TKIP) {
1498 		printf(" %s", "rxmic");
1499 		for (i = 0; i < sizeof(hk->key.tkip.rxMic); i++)
1500 			printf(" %02x", hk->key.tkip.rxMic[i]);
1501 		printf(" txmic");
1502 		for (i = 0; i < sizeof(hk->key.tkip.txMic); i++)
1503 			printf(" %02x", hk->key.tkip.txMic[i]);
1504 	}
1505 	printf(" flags 0x%x\n", hk->keyFlags);
1506 }
1507 #endif
1508 
1509 /*
1510  * Allocate a key cache slot for a unicast key.  The
1511  * firmware handles key allocation and every station is
1512  * guaranteed key space so we are always successful.
1513  */
1514 static int
mwl_key_alloc(struct ieee80211vap * vap,struct ieee80211_key * k,ieee80211_keyix * keyix,ieee80211_keyix * rxkeyix)1515 mwl_key_alloc(struct ieee80211vap *vap, struct ieee80211_key *k,
1516 	ieee80211_keyix *keyix, ieee80211_keyix *rxkeyix)
1517 {
1518 	struct mwl_softc *sc = vap->iv_ic->ic_softc;
1519 
1520 	if (k->wk_keyix != IEEE80211_KEYIX_NONE ||
1521 	    (k->wk_flags & IEEE80211_KEY_GROUP)) {
1522 		if (!ieee80211_is_key_global(vap, k)) {
1523 			/* should not happen */
1524 			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1525 				"%s: bogus group key\n", __func__);
1526 			return 0;
1527 		}
1528 		/* give the caller what they requested */
1529 		*keyix = *rxkeyix = ieee80211_crypto_get_key_wepidx(vap, k);
1530 	} else {
1531 		/*
1532 		 * Firmware handles key allocation.
1533 		 */
1534 		*keyix = *rxkeyix = 0;
1535 	}
1536 	return 1;
1537 }
1538 
1539 /*
1540  * Delete a key entry allocated by mwl_key_alloc.
1541  */
1542 static int
mwl_key_delete(struct ieee80211vap * vap,const struct ieee80211_key * k)1543 mwl_key_delete(struct ieee80211vap *vap, const struct ieee80211_key *k)
1544 {
1545 	struct mwl_softc *sc = vap->iv_ic->ic_softc;
1546 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1547 	MWL_HAL_KEYVAL hk;
1548 	const uint8_t bcastaddr[IEEE80211_ADDR_LEN] =
1549 	    { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff };
1550 
1551 	if (hvap == NULL) {
1552 		if (vap->iv_opmode != IEEE80211_M_WDS) {
1553 			/* XXX monitor mode? */
1554 			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1555 			    "%s: no hvap for opmode %d\n", __func__,
1556 			    vap->iv_opmode);
1557 			return 0;
1558 		}
1559 		hvap = MWL_VAP(vap)->mv_ap_hvap;
1560 	}
1561 
1562 	DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: delete key %u\n",
1563 	    __func__, k->wk_keyix);
1564 
1565 	memset(&hk, 0, sizeof(hk));
1566 	hk.keyIndex = k->wk_keyix;
1567 	switch (k->wk_cipher->ic_cipher) {
1568 	case IEEE80211_CIPHER_WEP:
1569 		hk.keyTypeId = KEY_TYPE_ID_WEP;
1570 		break;
1571 	case IEEE80211_CIPHER_TKIP:
1572 		hk.keyTypeId = KEY_TYPE_ID_TKIP;
1573 		break;
1574 	case IEEE80211_CIPHER_AES_CCM:
1575 		hk.keyTypeId = KEY_TYPE_ID_AES;
1576 		break;
1577 	default:
1578 		/* XXX should not happen */
1579 		DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1580 		    __func__, k->wk_cipher->ic_cipher);
1581 		return 0;
1582 	}
1583 	return (mwl_hal_keyreset(hvap, &hk, bcastaddr) == 0);	/*XXX*/
1584 }
1585 
1586 static __inline int
addgroupflags(MWL_HAL_KEYVAL * hk,const struct ieee80211_key * k)1587 addgroupflags(MWL_HAL_KEYVAL *hk, const struct ieee80211_key *k)
1588 {
1589 	if (k->wk_flags & IEEE80211_KEY_GROUP) {
1590 		if (k->wk_flags & IEEE80211_KEY_XMIT)
1591 			hk->keyFlags |= KEY_FLAG_TXGROUPKEY;
1592 		if (k->wk_flags & IEEE80211_KEY_RECV)
1593 			hk->keyFlags |= KEY_FLAG_RXGROUPKEY;
1594 		return 1;
1595 	} else
1596 		return 0;
1597 }
1598 
1599 /*
1600  * Set the key cache contents for the specified key.  Key cache
1601  * slot(s) must already have been allocated by mwl_key_alloc.
1602  */
1603 static int
mwl_key_set(struct ieee80211vap * vap,const struct ieee80211_key * k)1604 mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k)
1605 {
1606 	return (_mwl_key_set(vap, k, k->wk_macaddr));
1607 }
1608 
1609 static int
_mwl_key_set(struct ieee80211vap * vap,const struct ieee80211_key * k,const uint8_t mac[IEEE80211_ADDR_LEN])1610 _mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k,
1611 	const uint8_t mac[IEEE80211_ADDR_LEN])
1612 {
1613 #define	GRPXMIT	(IEEE80211_KEY_XMIT | IEEE80211_KEY_GROUP)
1614 /* NB: static wep keys are marked GROUP+tx/rx; GTK will be tx or rx */
1615 #define	IEEE80211_IS_STATICKEY(k) \
1616 	(((k)->wk_flags & (GRPXMIT|IEEE80211_KEY_RECV)) == \
1617 	 (GRPXMIT|IEEE80211_KEY_RECV))
1618 	struct mwl_softc *sc = vap->iv_ic->ic_softc;
1619 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1620 	const struct ieee80211_cipher *cip = k->wk_cipher;
1621 	const uint8_t *macaddr;
1622 	MWL_HAL_KEYVAL hk;
1623 
1624 	KASSERT((k->wk_flags & IEEE80211_KEY_SWCRYPT) == 0,
1625 		("s/w crypto set?"));
1626 
1627 	if (hvap == NULL) {
1628 		if (vap->iv_opmode != IEEE80211_M_WDS) {
1629 			/* XXX monitor mode? */
1630 			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1631 			    "%s: no hvap for opmode %d\n", __func__,
1632 			    vap->iv_opmode);
1633 			return 0;
1634 		}
1635 		hvap = MWL_VAP(vap)->mv_ap_hvap;
1636 	}
1637 	memset(&hk, 0, sizeof(hk));
1638 	hk.keyIndex = k->wk_keyix;
1639 	switch (cip->ic_cipher) {
1640 	case IEEE80211_CIPHER_WEP:
1641 		hk.keyTypeId = KEY_TYPE_ID_WEP;
1642 		hk.keyLen = k->wk_keylen;
1643 		if (k->wk_keyix == vap->iv_def_txkey)
1644 			hk.keyFlags = KEY_FLAG_WEP_TXKEY;
1645 		if (!IEEE80211_IS_STATICKEY(k)) {
1646 			/* NB: WEP is never used for the PTK */
1647 			(void) addgroupflags(&hk, k);
1648 		}
1649 		break;
1650 	case IEEE80211_CIPHER_TKIP:
1651 		hk.keyTypeId = KEY_TYPE_ID_TKIP;
1652 		hk.key.tkip.tsc.high = (uint32_t)(k->wk_keytsc >> 16);
1653 		hk.key.tkip.tsc.low = (uint16_t)k->wk_keytsc;
1654 		hk.keyFlags = KEY_FLAG_TSC_VALID | KEY_FLAG_MICKEY_VALID;
1655 		hk.keyLen = k->wk_keylen + IEEE80211_MICBUF_SIZE;
1656 		if (!addgroupflags(&hk, k))
1657 			hk.keyFlags |= KEY_FLAG_PAIRWISE;
1658 		break;
1659 	case IEEE80211_CIPHER_AES_CCM:
1660 		hk.keyTypeId = KEY_TYPE_ID_AES;
1661 		hk.keyLen = k->wk_keylen;
1662 		if (!addgroupflags(&hk, k))
1663 			hk.keyFlags |= KEY_FLAG_PAIRWISE;
1664 		break;
1665 	default:
1666 		/* XXX should not happen */
1667 		DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1668 		    __func__, k->wk_cipher->ic_cipher);
1669 		return 0;
1670 	}
1671 	/*
1672 	 * NB: tkip mic keys get copied here too; the layout
1673 	 *     just happens to match that in ieee80211_key.
1674 	 */
1675 	memcpy(hk.key.aes, k->wk_key, hk.keyLen);
1676 
1677 	/*
1678 	 * Locate address of sta db entry for writing key;
1679 	 * the convention unfortunately is somewhat different
1680 	 * than how net80211, hostapd, and wpa_supplicant think.
1681 	 */
1682 	if (vap->iv_opmode == IEEE80211_M_STA) {
1683 		/*
1684 		 * NB: keys plumbed before the sta reaches AUTH state
1685 		 * will be discarded or written to the wrong sta db
1686 		 * entry because iv_bss is meaningless.  This is ok
1687 		 * (right now) because we handle deferred plumbing of
1688 		 * WEP keys when the sta reaches AUTH state.
1689 		 */
1690 		macaddr = vap->iv_bss->ni_bssid;
1691 		if ((k->wk_flags & IEEE80211_KEY_GROUP) == 0) {
1692 			/* XXX plumb to local sta db too for static key wep */
1693 			mwl_hal_keyset(hvap, &hk, vap->iv_myaddr);
1694 		}
1695 	} else if (vap->iv_opmode == IEEE80211_M_WDS &&
1696 	    vap->iv_state != IEEE80211_S_RUN) {
1697 		/*
1698 		 * Prior to RUN state a WDS vap will not it's BSS node
1699 		 * setup so we will plumb the key to the wrong mac
1700 		 * address (it'll be our local address).  Workaround
1701 		 * this for the moment by grabbing the correct address.
1702 		 */
1703 		macaddr = vap->iv_des_bssid;
1704 	} else if ((k->wk_flags & GRPXMIT) == GRPXMIT)
1705 		macaddr = vap->iv_myaddr;
1706 	else
1707 		macaddr = mac;
1708 	KEYPRINTF(sc, &hk, macaddr);
1709 	return (mwl_hal_keyset(hvap, &hk, macaddr) == 0);
1710 #undef IEEE80211_IS_STATICKEY
1711 #undef GRPXMIT
1712 }
1713 
1714 /*
1715  * Set the multicast filter contents into the hardware.
1716  * XXX f/w has no support; just defer to the os.
1717  */
1718 static void
mwl_setmcastfilter(struct mwl_softc * sc)1719 mwl_setmcastfilter(struct mwl_softc *sc)
1720 {
1721 #if 0
1722 	struct ether_multi *enm;
1723 	struct ether_multistep estep;
1724 	uint8_t macs[IEEE80211_ADDR_LEN*MWL_HAL_MCAST_MAX];/* XXX stack use */
1725 	uint8_t *mp;
1726 	int nmc;
1727 
1728 	mp = macs;
1729 	nmc = 0;
1730 	ETHER_FIRST_MULTI(estep, &sc->sc_ec, enm);
1731 	while (enm != NULL) {
1732 		/* XXX Punt on ranges. */
1733 		if (nmc == MWL_HAL_MCAST_MAX ||
1734 		    !IEEE80211_ADDR_EQ(enm->enm_addrlo, enm->enm_addrhi)) {
1735 			if_setflagsbit(ifp, IFF_ALLMULTI, 0);
1736 			return;
1737 		}
1738 		IEEE80211_ADDR_COPY(mp, enm->enm_addrlo);
1739 		mp += IEEE80211_ADDR_LEN, nmc++;
1740 		ETHER_NEXT_MULTI(estep, enm);
1741 	}
1742 	if_setflagsbit(ifp, 0, IFF_ALLMULTI);
1743 	mwl_hal_setmcast(sc->sc_mh, nmc, macs);
1744 #endif
1745 }
1746 
1747 static int
mwl_mode_init(struct mwl_softc * sc)1748 mwl_mode_init(struct mwl_softc *sc)
1749 {
1750 	struct ieee80211com *ic = &sc->sc_ic;
1751 	struct mwl_hal *mh = sc->sc_mh;
1752 
1753 	mwl_hal_setpromisc(mh, ic->ic_promisc > 0);
1754 	mwl_setmcastfilter(sc);
1755 
1756 	return 0;
1757 }
1758 
1759 /*
1760  * Callback from the 802.11 layer after a multicast state change.
1761  */
1762 static void
mwl_update_mcast(struct ieee80211com * ic)1763 mwl_update_mcast(struct ieee80211com *ic)
1764 {
1765 	struct mwl_softc *sc = ic->ic_softc;
1766 
1767 	mwl_setmcastfilter(sc);
1768 }
1769 
1770 /*
1771  * Callback from the 802.11 layer after a promiscuous mode change.
1772  * Note this interface does not check the operating mode as this
1773  * is an internal callback and we are expected to honor the current
1774  * state (e.g. this is used for setting the interface in promiscuous
1775  * mode when operating in hostap mode to do ACS).
1776  */
1777 static void
mwl_update_promisc(struct ieee80211com * ic)1778 mwl_update_promisc(struct ieee80211com *ic)
1779 {
1780 	struct mwl_softc *sc = ic->ic_softc;
1781 
1782 	mwl_hal_setpromisc(sc->sc_mh, ic->ic_promisc > 0);
1783 }
1784 
1785 /*
1786  * Callback from the 802.11 layer to update the slot time
1787  * based on the current setting.  We use it to notify the
1788  * firmware of ERP changes and the f/w takes care of things
1789  * like slot time and preamble.
1790  */
1791 static void
mwl_updateslot(struct ieee80211com * ic)1792 mwl_updateslot(struct ieee80211com *ic)
1793 {
1794 	struct mwl_softc *sc = ic->ic_softc;
1795 	struct mwl_hal *mh = sc->sc_mh;
1796 	int prot;
1797 
1798 	/* NB: can be called early; suppress needless cmds */
1799 	if (!sc->sc_running)
1800 		return;
1801 
1802 	/*
1803 	 * Calculate the ERP flags.  The firwmare will use
1804 	 * this to carry out the appropriate measures.
1805 	 */
1806 	prot = 0;
1807 	if (IEEE80211_IS_CHAN_ANYG(ic->ic_curchan)) {
1808 		if ((ic->ic_flags & IEEE80211_F_SHSLOT) == 0)
1809 			prot |= IEEE80211_ERP_NON_ERP_PRESENT;
1810 		if (ic->ic_flags & IEEE80211_F_USEPROT)
1811 			prot |= IEEE80211_ERP_USE_PROTECTION;
1812 		if (ic->ic_flags & IEEE80211_F_USEBARKER)
1813 			prot |= IEEE80211_ERP_LONG_PREAMBLE;
1814 	}
1815 
1816 	DPRINTF(sc, MWL_DEBUG_RESET,
1817 	    "%s: chan %u MHz/flags 0x%x %s slot, (prot 0x%x ic_flags 0x%x)\n",
1818 	    __func__, ic->ic_curchan->ic_freq, ic->ic_curchan->ic_flags,
1819 	    ic->ic_flags & IEEE80211_F_SHSLOT ? "short" : "long", prot,
1820 	    ic->ic_flags);
1821 
1822 	mwl_hal_setgprot(mh, prot);
1823 }
1824 
1825 /*
1826  * Setup the beacon frame.
1827  */
1828 static int
mwl_beacon_setup(struct ieee80211vap * vap)1829 mwl_beacon_setup(struct ieee80211vap *vap)
1830 {
1831 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1832 	struct ieee80211_node *ni = vap->iv_bss;
1833 	struct mbuf *m;
1834 
1835 	m = ieee80211_beacon_alloc(ni);
1836 	if (m == NULL)
1837 		return ENOBUFS;
1838 	mwl_hal_setbeacon(hvap, mtod(m, const void *), m->m_len);
1839 	m_free(m);
1840 
1841 	return 0;
1842 }
1843 
1844 /*
1845  * Update the beacon frame in response to a change.
1846  */
1847 static void
mwl_beacon_update(struct ieee80211vap * vap,int item)1848 mwl_beacon_update(struct ieee80211vap *vap, int item)
1849 {
1850 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1851 	struct ieee80211com *ic = vap->iv_ic;
1852 
1853 	KASSERT(hvap != NULL, ("no beacon"));
1854 	switch (item) {
1855 	case IEEE80211_BEACON_ERP:
1856 		mwl_updateslot(ic);
1857 		break;
1858 	case IEEE80211_BEACON_HTINFO:
1859 		mwl_hal_setnprotmode(hvap, _IEEE80211_MASKSHIFT(
1860 		    ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1861 		break;
1862 	case IEEE80211_BEACON_CAPS:
1863 	case IEEE80211_BEACON_WME:
1864 	case IEEE80211_BEACON_APPIE:
1865 	case IEEE80211_BEACON_CSA:
1866 		break;
1867 	case IEEE80211_BEACON_TIM:
1868 		/* NB: firmware always forms TIM */
1869 		return;
1870 	}
1871 	/* XXX retain beacon frame and update */
1872 	mwl_beacon_setup(vap);
1873 }
1874 
1875 static void
mwl_load_cb(void * arg,bus_dma_segment_t * segs,int nsegs,int error)1876 mwl_load_cb(void *arg, bus_dma_segment_t *segs, int nsegs, int error)
1877 {
1878 	bus_addr_t *paddr = (bus_addr_t*) arg;
1879 	KASSERT(error == 0, ("error %u on bus_dma callback", error));
1880 	*paddr = segs->ds_addr;
1881 }
1882 
1883 #ifdef MWL_HOST_PS_SUPPORT
1884 /*
1885  * Handle power save station occupancy changes.
1886  */
1887 static void
mwl_update_ps(struct ieee80211vap * vap,int nsta)1888 mwl_update_ps(struct ieee80211vap *vap, int nsta)
1889 {
1890 	struct mwl_vap *mvp = MWL_VAP(vap);
1891 
1892 	if (nsta == 0 || mvp->mv_last_ps_sta == 0)
1893 		mwl_hal_setpowersave_bss(mvp->mv_hvap, nsta);
1894 	mvp->mv_last_ps_sta = nsta;
1895 }
1896 
1897 /*
1898  * Handle associated station power save state changes.
1899  */
1900 static int
mwl_set_tim(struct ieee80211_node * ni,int set)1901 mwl_set_tim(struct ieee80211_node *ni, int set)
1902 {
1903 	struct ieee80211vap *vap = ni->ni_vap;
1904 	struct mwl_vap *mvp = MWL_VAP(vap);
1905 
1906 	if (mvp->mv_set_tim(ni, set)) {		/* NB: state change */
1907 		mwl_hal_setpowersave_sta(mvp->mv_hvap,
1908 		    IEEE80211_AID(ni->ni_associd), set);
1909 		return 1;
1910 	} else
1911 		return 0;
1912 }
1913 #endif /* MWL_HOST_PS_SUPPORT */
1914 
1915 static int
mwl_desc_setup(struct mwl_softc * sc,const char * name,struct mwl_descdma * dd,int nbuf,size_t bufsize,int ndesc,size_t descsize)1916 mwl_desc_setup(struct mwl_softc *sc, const char *name,
1917 	struct mwl_descdma *dd,
1918 	int nbuf, size_t bufsize, int ndesc, size_t descsize)
1919 {
1920 	uint8_t *ds;
1921 	int error;
1922 
1923 	DPRINTF(sc, MWL_DEBUG_RESET,
1924 	    "%s: %s DMA: %u bufs (%ju) %u desc/buf (%ju)\n",
1925 	    __func__, name, nbuf, (uintmax_t) bufsize,
1926 	    ndesc, (uintmax_t) descsize);
1927 
1928 	dd->dd_name = name;
1929 	dd->dd_desc_len = nbuf * ndesc * descsize;
1930 
1931 	/*
1932 	 * Setup DMA descriptor area.
1933 	 */
1934 	error = bus_dma_tag_create(bus_get_dma_tag(sc->sc_dev),	/* parent */
1935 		       PAGE_SIZE, 0,		/* alignment, bounds */
1936 		       BUS_SPACE_MAXADDR_32BIT,	/* lowaddr */
1937 		       BUS_SPACE_MAXADDR,	/* highaddr */
1938 		       NULL, NULL,		/* filter, filterarg */
1939 		       dd->dd_desc_len,		/* maxsize */
1940 		       1,			/* nsegments */
1941 		       dd->dd_desc_len,		/* maxsegsize */
1942 		       BUS_DMA_ALLOCNOW,	/* flags */
1943 		       NULL,			/* lockfunc */
1944 		       NULL,			/* lockarg */
1945 		       &dd->dd_dmat);
1946 	if (error != 0) {
1947 		device_printf(sc->sc_dev, "cannot allocate %s DMA tag\n", dd->dd_name);
1948 		return error;
1949 	}
1950 
1951 	/* allocate descriptors */
1952 	error = bus_dmamem_alloc(dd->dd_dmat, (void**) &dd->dd_desc,
1953 				 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
1954 				 &dd->dd_dmamap);
1955 	if (error != 0) {
1956 		device_printf(sc->sc_dev, "unable to alloc memory for %u %s descriptors, "
1957 			"error %u\n", nbuf * ndesc, dd->dd_name, error);
1958 		goto fail1;
1959 	}
1960 
1961 	error = bus_dmamap_load(dd->dd_dmat, dd->dd_dmamap,
1962 				dd->dd_desc, dd->dd_desc_len,
1963 				mwl_load_cb, &dd->dd_desc_paddr,
1964 				BUS_DMA_NOWAIT);
1965 	if (error != 0) {
1966 		device_printf(sc->sc_dev, "unable to map %s descriptors, error %u\n",
1967 			dd->dd_name, error);
1968 		goto fail2;
1969 	}
1970 
1971 	ds = dd->dd_desc;
1972 	memset(ds, 0, dd->dd_desc_len);
1973 	DPRINTF(sc, MWL_DEBUG_RESET,
1974 	    "%s: %s DMA map: %p (%lu) -> 0x%jx (%lu)\n",
1975 	    __func__, dd->dd_name, ds, (u_long) dd->dd_desc_len,
1976 	    (uintmax_t) dd->dd_desc_paddr, /*XXX*/ (u_long) dd->dd_desc_len);
1977 
1978 	return 0;
1979 fail2:
1980 	bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
1981 fail1:
1982 	bus_dma_tag_destroy(dd->dd_dmat);
1983 	memset(dd, 0, sizeof(*dd));
1984 	return error;
1985 #undef DS2PHYS
1986 }
1987 
1988 static void
mwl_desc_cleanup(struct mwl_softc * sc,struct mwl_descdma * dd)1989 mwl_desc_cleanup(struct mwl_softc *sc, struct mwl_descdma *dd)
1990 {
1991 	bus_dmamap_unload(dd->dd_dmat, dd->dd_dmamap);
1992 	bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
1993 	bus_dma_tag_destroy(dd->dd_dmat);
1994 
1995 	memset(dd, 0, sizeof(*dd));
1996 }
1997 
1998 /*
1999  * Construct a tx q's free list.  The order of entries on
2000  * the list must reflect the physical layout of tx descriptors
2001  * because the firmware pre-fetches descriptors.
2002  *
2003  * XXX might be better to use indices into the buffer array.
2004  */
2005 static void
mwl_txq_reset(struct mwl_softc * sc,struct mwl_txq * txq)2006 mwl_txq_reset(struct mwl_softc *sc, struct mwl_txq *txq)
2007 {
2008 	struct mwl_txbuf *bf;
2009 	int i;
2010 
2011 	bf = txq->dma.dd_bufptr;
2012 	STAILQ_INIT(&txq->free);
2013 	for (i = 0; i < mwl_txbuf; i++, bf++)
2014 		STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
2015 	txq->nfree = i;
2016 }
2017 
2018 #define	DS2PHYS(_dd, _ds) \
2019 	((_dd)->dd_desc_paddr + ((caddr_t)(_ds) - (caddr_t)(_dd)->dd_desc))
2020 
2021 static int
mwl_txdma_setup(struct mwl_softc * sc,struct mwl_txq * txq)2022 mwl_txdma_setup(struct mwl_softc *sc, struct mwl_txq *txq)
2023 {
2024 	int error, bsize, i;
2025 	struct mwl_txbuf *bf;
2026 	struct mwl_txdesc *ds;
2027 
2028 	error = mwl_desc_setup(sc, "tx", &txq->dma,
2029 			mwl_txbuf, sizeof(struct mwl_txbuf),
2030 			MWL_TXDESC, sizeof(struct mwl_txdesc));
2031 	if (error != 0)
2032 		return error;
2033 
2034 	/* allocate and setup tx buffers */
2035 	bsize = mwl_txbuf * sizeof(struct mwl_txbuf);
2036 	bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2037 	if (bf == NULL) {
2038 		device_printf(sc->sc_dev, "malloc of %u tx buffers failed\n",
2039 			mwl_txbuf);
2040 		return ENOMEM;
2041 	}
2042 	txq->dma.dd_bufptr = bf;
2043 
2044 	ds = txq->dma.dd_desc;
2045 	for (i = 0; i < mwl_txbuf; i++, bf++, ds += MWL_TXDESC) {
2046 		bf->bf_desc = ds;
2047 		bf->bf_daddr = DS2PHYS(&txq->dma, ds);
2048 		error = bus_dmamap_create(sc->sc_dmat, BUS_DMA_NOWAIT,
2049 				&bf->bf_dmamap);
2050 		if (error != 0) {
2051 			device_printf(sc->sc_dev, "unable to create dmamap for tx "
2052 				"buffer %u, error %u\n", i, error);
2053 			return error;
2054 		}
2055 	}
2056 	mwl_txq_reset(sc, txq);
2057 	return 0;
2058 }
2059 
2060 static void
mwl_txdma_cleanup(struct mwl_softc * sc,struct mwl_txq * txq)2061 mwl_txdma_cleanup(struct mwl_softc *sc, struct mwl_txq *txq)
2062 {
2063 	struct mwl_txbuf *bf;
2064 	int i;
2065 
2066 	bf = txq->dma.dd_bufptr;
2067 	for (i = 0; i < mwl_txbuf; i++, bf++) {
2068 		KASSERT(bf->bf_m == NULL, ("mbuf on free list"));
2069 		KASSERT(bf->bf_node == NULL, ("node on free list"));
2070 		if (bf->bf_dmamap != NULL)
2071 			bus_dmamap_destroy(sc->sc_dmat, bf->bf_dmamap);
2072 	}
2073 	STAILQ_INIT(&txq->free);
2074 	txq->nfree = 0;
2075 	if (txq->dma.dd_bufptr != NULL) {
2076 		free(txq->dma.dd_bufptr, M_MWLDEV);
2077 		txq->dma.dd_bufptr = NULL;
2078 	}
2079 	if (txq->dma.dd_desc_len != 0)
2080 		mwl_desc_cleanup(sc, &txq->dma);
2081 }
2082 
2083 static int
mwl_rxdma_setup(struct mwl_softc * sc)2084 mwl_rxdma_setup(struct mwl_softc *sc)
2085 {
2086 	int error, jumbosize, bsize, i;
2087 	struct mwl_rxbuf *bf;
2088 	struct mwl_jumbo *rbuf;
2089 	struct mwl_rxdesc *ds;
2090 	caddr_t data;
2091 
2092 	error = mwl_desc_setup(sc, "rx", &sc->sc_rxdma,
2093 			mwl_rxdesc, sizeof(struct mwl_rxbuf),
2094 			1, sizeof(struct mwl_rxdesc));
2095 	if (error != 0)
2096 		return error;
2097 
2098 	/*
2099 	 * Receive is done to a private pool of jumbo buffers.
2100 	 * This allows us to attach to mbuf's and avoid re-mapping
2101 	 * memory on each rx we post.  We allocate a large chunk
2102 	 * of memory and manage it in the driver.  The mbuf free
2103 	 * callback method is used to reclaim frames after sending
2104 	 * them up the stack.  By default we allocate 2x the number of
2105 	 * rx descriptors configured so we have some slop to hold
2106 	 * us while frames are processed.
2107 	 */
2108 	if (mwl_rxbuf < 2*mwl_rxdesc) {
2109 		device_printf(sc->sc_dev,
2110 		    "too few rx dma buffers (%d); increasing to %d\n",
2111 		    mwl_rxbuf, 2*mwl_rxdesc);
2112 		mwl_rxbuf = 2*mwl_rxdesc;
2113 	}
2114 	jumbosize = roundup(MWL_AGGR_SIZE, PAGE_SIZE);
2115 	sc->sc_rxmemsize = mwl_rxbuf*jumbosize;
2116 
2117 	error = bus_dma_tag_create(sc->sc_dmat,	/* parent */
2118 		       PAGE_SIZE, 0,		/* alignment, bounds */
2119 		       BUS_SPACE_MAXADDR_32BIT,	/* lowaddr */
2120 		       BUS_SPACE_MAXADDR,	/* highaddr */
2121 		       NULL, NULL,		/* filter, filterarg */
2122 		       sc->sc_rxmemsize,	/* maxsize */
2123 		       1,			/* nsegments */
2124 		       sc->sc_rxmemsize,	/* maxsegsize */
2125 		       BUS_DMA_ALLOCNOW,	/* flags */
2126 		       NULL,			/* lockfunc */
2127 		       NULL,			/* lockarg */
2128 		       &sc->sc_rxdmat);
2129 	if (error != 0) {
2130 		device_printf(sc->sc_dev, "could not create rx DMA tag\n");
2131 		return error;
2132 	}
2133 
2134 	error = bus_dmamem_alloc(sc->sc_rxdmat, (void**) &sc->sc_rxmem,
2135 				 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
2136 				 &sc->sc_rxmap);
2137 	if (error != 0) {
2138 		device_printf(sc->sc_dev, "could not alloc %ju bytes of rx DMA memory\n",
2139 		    (uintmax_t) sc->sc_rxmemsize);
2140 		return error;
2141 	}
2142 
2143 	error = bus_dmamap_load(sc->sc_rxdmat, sc->sc_rxmap,
2144 				sc->sc_rxmem, sc->sc_rxmemsize,
2145 				mwl_load_cb, &sc->sc_rxmem_paddr,
2146 				BUS_DMA_NOWAIT);
2147 	if (error != 0) {
2148 		device_printf(sc->sc_dev, "could not load rx DMA map\n");
2149 		return error;
2150 	}
2151 
2152 	/*
2153 	 * Allocate rx buffers and set them up.
2154 	 */
2155 	bsize = mwl_rxdesc * sizeof(struct mwl_rxbuf);
2156 	bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2157 	if (bf == NULL) {
2158 		device_printf(sc->sc_dev, "malloc of %u rx buffers failed\n", bsize);
2159 		return error;
2160 	}
2161 	sc->sc_rxdma.dd_bufptr = bf;
2162 
2163 	STAILQ_INIT(&sc->sc_rxbuf);
2164 	ds = sc->sc_rxdma.dd_desc;
2165 	for (i = 0; i < mwl_rxdesc; i++, bf++, ds++) {
2166 		bf->bf_desc = ds;
2167 		bf->bf_daddr = DS2PHYS(&sc->sc_rxdma, ds);
2168 		/* pre-assign dma buffer */
2169 		bf->bf_data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2170 		/* NB: tail is intentional to preserve descriptor order */
2171 		STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list);
2172 	}
2173 
2174 	/*
2175 	 * Place remainder of dma memory buffers on the free list.
2176 	 */
2177 	SLIST_INIT(&sc->sc_rxfree);
2178 	for (; i < mwl_rxbuf; i++) {
2179 		data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2180 		rbuf = MWL_JUMBO_DATA2BUF(data);
2181 		SLIST_INSERT_HEAD(&sc->sc_rxfree, rbuf, next);
2182 		sc->sc_nrxfree++;
2183 	}
2184 	return 0;
2185 }
2186 #undef DS2PHYS
2187 
2188 static void
mwl_rxdma_cleanup(struct mwl_softc * sc)2189 mwl_rxdma_cleanup(struct mwl_softc *sc)
2190 {
2191 	if (sc->sc_rxmem_paddr != 0) {
2192 		bus_dmamap_unload(sc->sc_rxdmat, sc->sc_rxmap);
2193 		sc->sc_rxmem_paddr = 0;
2194 	}
2195 	if (sc->sc_rxmem != NULL) {
2196 		bus_dmamem_free(sc->sc_rxdmat, sc->sc_rxmem, sc->sc_rxmap);
2197 		sc->sc_rxmem = NULL;
2198 	}
2199 	if (sc->sc_rxdma.dd_bufptr != NULL) {
2200 		free(sc->sc_rxdma.dd_bufptr, M_MWLDEV);
2201 		sc->sc_rxdma.dd_bufptr = NULL;
2202 	}
2203 	if (sc->sc_rxdma.dd_desc_len != 0)
2204 		mwl_desc_cleanup(sc, &sc->sc_rxdma);
2205 }
2206 
2207 static int
mwl_dma_setup(struct mwl_softc * sc)2208 mwl_dma_setup(struct mwl_softc *sc)
2209 {
2210 	int error, i;
2211 
2212 	error = mwl_rxdma_setup(sc);
2213 	if (error != 0) {
2214 		mwl_rxdma_cleanup(sc);
2215 		return error;
2216 	}
2217 
2218 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
2219 		error = mwl_txdma_setup(sc, &sc->sc_txq[i]);
2220 		if (error != 0) {
2221 			mwl_dma_cleanup(sc);
2222 			return error;
2223 		}
2224 	}
2225 	return 0;
2226 }
2227 
2228 static void
mwl_dma_cleanup(struct mwl_softc * sc)2229 mwl_dma_cleanup(struct mwl_softc *sc)
2230 {
2231 	int i;
2232 
2233 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2234 		mwl_txdma_cleanup(sc, &sc->sc_txq[i]);
2235 	mwl_rxdma_cleanup(sc);
2236 }
2237 
2238 static struct ieee80211_node *
mwl_node_alloc(struct ieee80211vap * vap,const uint8_t mac[IEEE80211_ADDR_LEN])2239 mwl_node_alloc(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
2240 {
2241 	struct ieee80211com *ic = vap->iv_ic;
2242 	struct mwl_softc *sc = ic->ic_softc;
2243 	const size_t space = sizeof(struct mwl_node);
2244 	struct mwl_node *mn;
2245 
2246 	mn = malloc(space, M_80211_NODE, M_NOWAIT|M_ZERO);
2247 	if (mn == NULL) {
2248 		/* XXX stat+msg */
2249 		return NULL;
2250 	}
2251 	DPRINTF(sc, MWL_DEBUG_NODE, "%s: mn %p\n", __func__, mn);
2252 	return &mn->mn_node;
2253 }
2254 
2255 static void
mwl_node_cleanup(struct ieee80211_node * ni)2256 mwl_node_cleanup(struct ieee80211_node *ni)
2257 {
2258 	struct ieee80211com *ic = ni->ni_ic;
2259         struct mwl_softc *sc = ic->ic_softc;
2260 	struct mwl_node *mn = MWL_NODE(ni);
2261 
2262 	DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p ic %p staid %d\n",
2263 	    __func__, ni, ni->ni_ic, mn->mn_staid);
2264 
2265 	if (mn->mn_staid != 0) {
2266 		struct ieee80211vap *vap = ni->ni_vap;
2267 
2268 		if (mn->mn_hvap != NULL) {
2269 			if (vap->iv_opmode == IEEE80211_M_STA)
2270 				mwl_hal_delstation(mn->mn_hvap, vap->iv_myaddr);
2271 			else
2272 				mwl_hal_delstation(mn->mn_hvap, ni->ni_macaddr);
2273 		}
2274 		/*
2275 		 * NB: legacy WDS peer sta db entry is installed using
2276 		 * the associate ap's hvap; use it again to delete it.
2277 		 * XXX can vap be NULL?
2278 		 */
2279 		else if (vap->iv_opmode == IEEE80211_M_WDS &&
2280 		    MWL_VAP(vap)->mv_ap_hvap != NULL)
2281 			mwl_hal_delstation(MWL_VAP(vap)->mv_ap_hvap,
2282 			    ni->ni_macaddr);
2283 		delstaid(sc, mn->mn_staid);
2284 		mn->mn_staid = 0;
2285 	}
2286 	sc->sc_node_cleanup(ni);
2287 }
2288 
2289 /*
2290  * Reclaim rx dma buffers from packets sitting on the ampdu
2291  * reorder queue for a station.  We replace buffers with a
2292  * system cluster (if available).
2293  */
2294 static void
mwl_ampdu_rxdma_reclaim(struct ieee80211_rx_ampdu * rap)2295 mwl_ampdu_rxdma_reclaim(struct ieee80211_rx_ampdu *rap)
2296 {
2297 #if 0
2298 	int i, n, off;
2299 	struct mbuf *m;
2300 	void *cl;
2301 
2302 	n = rap->rxa_qframes;
2303 	for (i = 0; i < rap->rxa_wnd && n > 0; i++) {
2304 		m = rap->rxa_m[i];
2305 		if (m == NULL)
2306 			continue;
2307 		n--;
2308 		/* our dma buffers have a well-known free routine */
2309 		if ((m->m_flags & M_EXT) == 0 ||
2310 		    m->m_ext.ext_free != mwl_ext_free)
2311 			continue;
2312 		/*
2313 		 * Try to allocate a cluster and move the data.
2314 		 */
2315 		off = m->m_data - m->m_ext.ext_buf;
2316 		if (off + m->m_pkthdr.len > MCLBYTES) {
2317 			/* XXX no AMSDU for now */
2318 			continue;
2319 		}
2320 		cl = pool_cache_get_paddr(&mclpool_cache, 0,
2321 		    &m->m_ext.ext_paddr);
2322 		if (cl != NULL) {
2323 			/*
2324 			 * Copy the existing data to the cluster, remove
2325 			 * the rx dma buffer, and attach the cluster in
2326 			 * its place.  Note we preserve the offset to the
2327 			 * data so frames being bridged can still prepend
2328 			 * their headers without adding another mbuf.
2329 			 */
2330 			memcpy((caddr_t) cl + off, m->m_data, m->m_pkthdr.len);
2331 			MEXTREMOVE(m);
2332 			MEXTADD(m, cl, MCLBYTES, 0, NULL, &mclpool_cache);
2333 			/* setup mbuf like _MCLGET does */
2334 			m->m_flags |= M_CLUSTER | M_EXT_RW;
2335 			_MOWNERREF(m, M_EXT | M_CLUSTER);
2336 			/* NB: m_data is clobbered by MEXTADDR, adjust */
2337 			m->m_data += off;
2338 		}
2339 	}
2340 #endif
2341 }
2342 
2343 /*
2344  * Callback to reclaim resources.  We first let the
2345  * net80211 layer do it's thing, then if we are still
2346  * blocked by a lack of rx dma buffers we walk the ampdu
2347  * reorder q's to reclaim buffers by copying to a system
2348  * cluster.
2349  */
2350 static void
mwl_node_drain(struct ieee80211_node * ni)2351 mwl_node_drain(struct ieee80211_node *ni)
2352 {
2353 	struct ieee80211com *ic = ni->ni_ic;
2354         struct mwl_softc *sc = ic->ic_softc;
2355 	struct mwl_node *mn = MWL_NODE(ni);
2356 
2357 	DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p vap %p staid %d\n",
2358 	    __func__, ni, ni->ni_vap, mn->mn_staid);
2359 
2360 	/* NB: call up first to age out ampdu q's */
2361 	sc->sc_node_drain(ni);
2362 
2363 	/* XXX better to not check low water mark? */
2364 	if (sc->sc_rxblocked && mn->mn_staid != 0 &&
2365 	    (ni->ni_flags & IEEE80211_NODE_HT)) {
2366 		uint8_t tid;
2367 		/*
2368 		 * Walk the reorder q and reclaim rx dma buffers by copying
2369 		 * the packet contents into clusters.
2370 		 */
2371 		for (tid = 0; tid < WME_NUM_TID; tid++) {
2372 			struct ieee80211_rx_ampdu *rap;
2373 
2374 			rap = &ni->ni_rx_ampdu[tid];
2375 			if ((rap->rxa_flags & IEEE80211_AGGR_XCHGPEND) == 0)
2376 				continue;
2377 			if (rap->rxa_qframes)
2378 				mwl_ampdu_rxdma_reclaim(rap);
2379 		}
2380 	}
2381 }
2382 
2383 static void
mwl_node_getsignal(const struct ieee80211_node * ni,int8_t * rssi,int8_t * noise)2384 mwl_node_getsignal(const struct ieee80211_node *ni, int8_t *rssi, int8_t *noise)
2385 {
2386 	*rssi = ni->ni_ic->ic_node_getrssi(ni);
2387 #ifdef MWL_ANT_INFO_SUPPORT
2388 #if 0
2389 	/* XXX need to smooth data */
2390 	*noise = -MWL_NODE_CONST(ni)->mn_ai.nf;
2391 #else
2392 	*noise = -95;		/* XXX */
2393 #endif
2394 #else
2395 	*noise = -95;		/* XXX */
2396 #endif
2397 }
2398 
2399 /*
2400  * Convert Hardware per-antenna rssi info to common format:
2401  * Let a1, a2, a3 represent the amplitudes per chain
2402  * Let amax represent max[a1, a2, a3]
2403  * Rssi1_dBm = RSSI_dBm + 20*log10(a1/amax)
2404  * Rssi1_dBm = RSSI_dBm + 20*log10(a1) - 20*log10(amax)
2405  * We store a table that is 4*20*log10(idx) - the extra 4 is to store or
2406  * maintain some extra precision.
2407  *
2408  * Values are stored in .5 db format capped at 127.
2409  */
2410 static void
mwl_node_getmimoinfo(const struct ieee80211_node * ni,struct ieee80211_mimo_info * mi)2411 mwl_node_getmimoinfo(const struct ieee80211_node *ni,
2412 	struct ieee80211_mimo_info *mi)
2413 {
2414 #define	CVT(_dst, _src) do {						\
2415 	(_dst) = rssi + ((logdbtbl[_src] - logdbtbl[rssi_max]) >> 2);	\
2416 	(_dst) = (_dst) > 64 ? 127 : ((_dst) << 1);			\
2417 } while (0)
2418 	static const int8_t logdbtbl[32] = {
2419 	       0,   0,  24,  38,  48,  56,  62,  68,
2420 	      72,  76,  80,  83,  86,  89,  92,  94,
2421 	      96,  98, 100, 102, 104, 106, 107, 109,
2422 	     110, 112, 113, 115, 116, 117, 118, 119
2423 	};
2424 	const struct mwl_node *mn = MWL_NODE_CONST(ni);
2425 	uint8_t rssi = mn->mn_ai.rsvd1/2;		/* XXX */
2426 	uint32_t rssi_max;
2427 
2428 	rssi_max = mn->mn_ai.rssi_a;
2429 	if (mn->mn_ai.rssi_b > rssi_max)
2430 		rssi_max = mn->mn_ai.rssi_b;
2431 	if (mn->mn_ai.rssi_c > rssi_max)
2432 		rssi_max = mn->mn_ai.rssi_c;
2433 
2434 	CVT(mi->ch[0].rssi[0], mn->mn_ai.rssi_a);
2435 	CVT(mi->ch[1].rssi[0], mn->mn_ai.rssi_b);
2436 	CVT(mi->ch[2].rssi[0], mn->mn_ai.rssi_c);
2437 
2438 	mi->ch[0].noise[0] = mn->mn_ai.nf_a;
2439 	mi->ch[1].noise[0] = mn->mn_ai.nf_b;
2440 	mi->ch[2].noise[0] = mn->mn_ai.nf_c;
2441 #undef CVT
2442 }
2443 
2444 static __inline void *
mwl_getrxdma(struct mwl_softc * sc)2445 mwl_getrxdma(struct mwl_softc *sc)
2446 {
2447 	struct mwl_jumbo *buf;
2448 	void *data;
2449 
2450 	/*
2451 	 * Allocate from jumbo pool.
2452 	 */
2453 	MWL_RXFREE_LOCK(sc);
2454 	buf = SLIST_FIRST(&sc->sc_rxfree);
2455 	if (buf == NULL) {
2456 		DPRINTF(sc, MWL_DEBUG_ANY,
2457 		    "%s: out of rx dma buffers\n", __func__);
2458 		sc->sc_stats.mst_rx_nodmabuf++;
2459 		data = NULL;
2460 	} else {
2461 		SLIST_REMOVE_HEAD(&sc->sc_rxfree, next);
2462 		sc->sc_nrxfree--;
2463 		data = MWL_JUMBO_BUF2DATA(buf);
2464 	}
2465 	MWL_RXFREE_UNLOCK(sc);
2466 	return data;
2467 }
2468 
2469 static __inline void
mwl_putrxdma(struct mwl_softc * sc,void * data)2470 mwl_putrxdma(struct mwl_softc *sc, void *data)
2471 {
2472 	struct mwl_jumbo *buf;
2473 
2474 	/* XXX bounds check data */
2475 	MWL_RXFREE_LOCK(sc);
2476 	buf = MWL_JUMBO_DATA2BUF(data);
2477 	SLIST_INSERT_HEAD(&sc->sc_rxfree, buf, next);
2478 	sc->sc_nrxfree++;
2479 	MWL_RXFREE_UNLOCK(sc);
2480 }
2481 
2482 static int
mwl_rxbuf_init(struct mwl_softc * sc,struct mwl_rxbuf * bf)2483 mwl_rxbuf_init(struct mwl_softc *sc, struct mwl_rxbuf *bf)
2484 {
2485 	struct mwl_rxdesc *ds;
2486 
2487 	ds = bf->bf_desc;
2488 	if (bf->bf_data == NULL) {
2489 		bf->bf_data = mwl_getrxdma(sc);
2490 		if (bf->bf_data == NULL) {
2491 			/* mark descriptor to be skipped */
2492 			ds->RxControl = EAGLE_RXD_CTRL_OS_OWN;
2493 			/* NB: don't need PREREAD */
2494 			MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREWRITE);
2495 			sc->sc_stats.mst_rxbuf_failed++;
2496 			return ENOMEM;
2497 		}
2498 	}
2499 	/*
2500 	 * NB: DMA buffer contents is known to be unmodified
2501 	 *     so there's no need to flush the data cache.
2502 	 */
2503 
2504 	/*
2505 	 * Setup descriptor.
2506 	 */
2507 	ds->QosCtrl = 0;
2508 	ds->RSSI = 0;
2509 	ds->Status = EAGLE_RXD_STATUS_IDLE;
2510 	ds->Channel = 0;
2511 	ds->PktLen = htole16(MWL_AGGR_SIZE);
2512 	ds->SQ2 = 0;
2513 	ds->pPhysBuffData = htole32(MWL_JUMBO_DMA_ADDR(sc, bf->bf_data));
2514 	/* NB: don't touch pPhysNext, set once */
2515 	ds->RxControl = EAGLE_RXD_CTRL_DRIVER_OWN;
2516 	MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
2517 
2518 	return 0;
2519 }
2520 
2521 static void
mwl_ext_free(struct mbuf * m)2522 mwl_ext_free(struct mbuf *m)
2523 {
2524 	struct mwl_softc *sc = m->m_ext.ext_arg1;
2525 
2526 	/* XXX bounds check data */
2527 	mwl_putrxdma(sc, m->m_ext.ext_buf);
2528 	/*
2529 	 * If we were previously blocked by a lack of rx dma buffers
2530 	 * check if we now have enough to restart rx interrupt handling.
2531 	 */
2532 	if (sc->sc_rxblocked && sc->sc_nrxfree > mwl_rxdmalow) {
2533 		sc->sc_rxblocked = 0;
2534 		mwl_hal_intrset(sc->sc_mh, sc->sc_imask);
2535 	}
2536 }
2537 
2538 struct mwl_frame_bar {
2539 	u_int8_t	i_fc[2];
2540 	u_int8_t	i_dur[2];
2541 	u_int8_t	i_ra[IEEE80211_ADDR_LEN];
2542 	u_int8_t	i_ta[IEEE80211_ADDR_LEN];
2543 	/* ctl, seq, FCS */
2544 } __packed;
2545 
2546 /*
2547  * Like ieee80211_anyhdrsize, but handles BAR frames
2548  * specially so the logic below to piece the 802.11
2549  * header together works.
2550  */
2551 static __inline int
mwl_anyhdrsize(const void * data)2552 mwl_anyhdrsize(const void *data)
2553 {
2554 	const struct ieee80211_frame *wh = data;
2555 
2556 	if ((wh->i_fc[0]&IEEE80211_FC0_TYPE_MASK) == IEEE80211_FC0_TYPE_CTL) {
2557 		switch (wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK) {
2558 		case IEEE80211_FC0_SUBTYPE_CTS:
2559 		case IEEE80211_FC0_SUBTYPE_ACK:
2560 			return sizeof(struct ieee80211_frame_ack);
2561 		case IEEE80211_FC0_SUBTYPE_BAR:
2562 			return sizeof(struct mwl_frame_bar);
2563 		}
2564 		return sizeof(struct ieee80211_frame_min);
2565 	} else
2566 		return ieee80211_hdrsize(data);
2567 }
2568 
2569 static void
mwl_handlemicerror(struct ieee80211com * ic,const uint8_t * data)2570 mwl_handlemicerror(struct ieee80211com *ic, const uint8_t *data)
2571 {
2572 	const struct ieee80211_frame *wh;
2573 	struct ieee80211_node *ni;
2574 
2575 	wh = (const struct ieee80211_frame *)(data + sizeof(uint16_t));
2576 	ni = ieee80211_find_rxnode(ic, (const struct ieee80211_frame_min *) wh);
2577 	if (ni != NULL) {
2578 		ieee80211_notify_michael_failure(ni->ni_vap, wh, 0);
2579 		ieee80211_free_node(ni);
2580 	}
2581 }
2582 
2583 /*
2584  * Convert hardware signal strength to rssi.  The value
2585  * provided by the device has the noise floor added in;
2586  * we need to compensate for this but we don't have that
2587  * so we use a fixed value.
2588  *
2589  * The offset of 8 is good for both 2.4 and 5GHz.  The LNA
2590  * offset is already set as part of the initial gain.  This
2591  * will give at least +/- 3dB for 2.4GHz and +/- 5dB for 5GHz.
2592  */
2593 static __inline int
cvtrssi(uint8_t ssi)2594 cvtrssi(uint8_t ssi)
2595 {
2596 	int rssi = (int) ssi + 8;
2597 	/* XXX hack guess until we have a real noise floor */
2598 	rssi = 2*(87 - rssi);	/* NB: .5 dBm units */
2599 	return (rssi < 0 ? 0 : rssi > 127 ? 127 : rssi);
2600 }
2601 
2602 static void
mwl_rx_proc(void * arg,int npending)2603 mwl_rx_proc(void *arg, int npending)
2604 {
2605 	struct mwl_softc *sc = arg;
2606 	struct ieee80211com *ic = &sc->sc_ic;
2607 	struct mwl_rxbuf *bf;
2608 	struct mwl_rxdesc *ds;
2609 	struct mbuf *m;
2610 	struct ieee80211_qosframe *wh;
2611 	struct ieee80211_node *ni;
2612 	struct mwl_node *mn;
2613 	int off, len, hdrlen, pktlen, rssi, ntodo;
2614 	uint8_t *data, status;
2615 	void *newdata;
2616 	int16_t nf;
2617 
2618 	DPRINTF(sc, MWL_DEBUG_RX_PROC, "%s: pending %u rdptr 0x%x wrptr 0x%x\n",
2619 	    __func__, npending, RD4(sc, sc->sc_hwspecs.rxDescRead),
2620 	    RD4(sc, sc->sc_hwspecs.rxDescWrite));
2621 	nf = -96;			/* XXX */
2622 	bf = sc->sc_rxnext;
2623 	for (ntodo = mwl_rxquota; ntodo > 0; ntodo--) {
2624 		if (bf == NULL)
2625 			bf = STAILQ_FIRST(&sc->sc_rxbuf);
2626 		ds = bf->bf_desc;
2627 		data = bf->bf_data;
2628 		if (data == NULL) {
2629 			/*
2630 			 * If data allocation failed previously there
2631 			 * will be no buffer; try again to re-populate it.
2632 			 * Note the firmware will not advance to the next
2633 			 * descriptor with a dma buffer so we must mimic
2634 			 * this or we'll get out of sync.
2635 			 */
2636 			DPRINTF(sc, MWL_DEBUG_ANY,
2637 			    "%s: rx buf w/o dma memory\n", __func__);
2638 			(void) mwl_rxbuf_init(sc, bf);
2639 			sc->sc_stats.mst_rx_dmabufmissing++;
2640 			break;
2641 		}
2642 		MWL_RXDESC_SYNC(sc, ds,
2643 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
2644 		if (ds->RxControl != EAGLE_RXD_CTRL_DMA_OWN)
2645 			break;
2646 #ifdef MWL_DEBUG
2647 		if (sc->sc_debug & MWL_DEBUG_RECV_DESC)
2648 			mwl_printrxbuf(bf, 0);
2649 #endif
2650 		status = ds->Status;
2651 		if (status & EAGLE_RXD_STATUS_DECRYPT_ERR_MASK) {
2652 			counter_u64_add(ic->ic_ierrors, 1);
2653 			sc->sc_stats.mst_rx_crypto++;
2654 			/*
2655 			 * NB: Check EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR
2656 			 *     for backwards compatibility.
2657 			 */
2658 			if (status != EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR &&
2659 			    (status & EAGLE_RXD_STATUS_TKIP_MIC_DECRYPT_ERR)) {
2660 				/*
2661 				 * MIC error, notify upper layers.
2662 				 */
2663 				bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap,
2664 				    BUS_DMASYNC_POSTREAD);
2665 				mwl_handlemicerror(ic, data);
2666 				sc->sc_stats.mst_rx_tkipmic++;
2667 			}
2668 			/* XXX too painful to tap packets */
2669 			goto rx_next;
2670 		}
2671 		/*
2672 		 * Sync the data buffer.
2673 		 */
2674 		len = le16toh(ds->PktLen);
2675 		bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap, BUS_DMASYNC_POSTREAD);
2676 		/*
2677 		 * The 802.11 header is provided all or in part at the front;
2678 		 * use it to calculate the true size of the header that we'll
2679 		 * construct below.  We use this to figure out where to copy
2680 		 * payload prior to constructing the header.
2681 		 */
2682 		hdrlen = mwl_anyhdrsize(data + sizeof(uint16_t));
2683 		off = sizeof(uint16_t) + sizeof(struct ieee80211_frame_addr4);
2684 
2685 		/* calculate rssi early so we can re-use for each aggregate */
2686 		rssi = cvtrssi(ds->RSSI);
2687 
2688 		pktlen = hdrlen + (len - off);
2689 		/*
2690 		 * NB: we know our frame is at least as large as
2691 		 * IEEE80211_MIN_LEN because there is a 4-address
2692 		 * frame at the front.  Hence there's no need to
2693 		 * vet the packet length.  If the frame in fact
2694 		 * is too small it should be discarded at the
2695 		 * net80211 layer.
2696 		 */
2697 
2698 		/*
2699 		 * Attach dma buffer to an mbuf.  We tried
2700 		 * doing this based on the packet size (i.e.
2701 		 * copying small packets) but it turns out to
2702 		 * be a net loss.  The tradeoff might be system
2703 		 * dependent (cache architecture is important).
2704 		 */
2705 		MGETHDR(m, M_NOWAIT, MT_DATA);
2706 		if (m == NULL) {
2707 			DPRINTF(sc, MWL_DEBUG_ANY,
2708 			    "%s: no rx mbuf\n", __func__);
2709 			sc->sc_stats.mst_rx_nombuf++;
2710 			goto rx_next;
2711 		}
2712 		/*
2713 		 * Acquire the replacement dma buffer before
2714 		 * processing the frame.  If we're out of dma
2715 		 * buffers we disable rx interrupts and wait
2716 		 * for the free pool to reach mlw_rxdmalow buffers
2717 		 * before starting to do work again.  If the firmware
2718 		 * runs out of descriptors then it will toss frames
2719 		 * which is better than our doing it as that can
2720 		 * starve our processing.  It is also important that
2721 		 * we always process rx'd frames in case they are
2722 		 * A-MPDU as otherwise the host's view of the BA
2723 		 * window may get out of sync with the firmware.
2724 		 */
2725 		newdata = mwl_getrxdma(sc);
2726 		if (newdata == NULL) {
2727 			/* NB: stat+msg in mwl_getrxdma */
2728 			m_free(m);
2729 			/* disable RX interrupt and mark state */
2730 			mwl_hal_intrset(sc->sc_mh,
2731 			    sc->sc_imask &~ MACREG_A2HRIC_BIT_RX_RDY);
2732 			sc->sc_rxblocked = 1;
2733 			ieee80211_drain(ic);
2734 			/* XXX check rxblocked and immediately start again? */
2735 			goto rx_stop;
2736 		}
2737 		bf->bf_data = newdata;
2738 		/*
2739 		 * Attach the dma buffer to the mbuf;
2740 		 * mwl_rxbuf_init will re-setup the rx
2741 		 * descriptor using the replacement dma
2742 		 * buffer we just installed above.
2743 		 */
2744 		m_extadd(m, data, MWL_AGGR_SIZE, mwl_ext_free, sc, NULL, 0,
2745 		    EXT_NET_DRV);
2746 		m->m_data += off - hdrlen;
2747 		m->m_pkthdr.len = m->m_len = pktlen;
2748 		/* NB: dma buffer assumed read-only */
2749 
2750 		/*
2751 		 * Piece 802.11 header together.
2752 		 */
2753 		wh = mtod(m, struct ieee80211_qosframe *);
2754 		/* NB: don't need to do this sometimes but ... */
2755 		/* XXX special case so we can memcpy after m_devget? */
2756 		ovbcopy(data + sizeof(uint16_t), wh, hdrlen);
2757 		if (IEEE80211_QOS_HAS_SEQ(wh))
2758 			*(uint16_t *)ieee80211_getqos(wh) = ds->QosCtrl;
2759 		/*
2760 		 * The f/w strips WEP header but doesn't clear
2761 		 * the WEP bit; mark the packet with M_WEP so
2762 		 * net80211 will treat the data as decrypted.
2763 		 * While here also clear the PWR_MGT bit since
2764 		 * power save is handled by the firmware and
2765 		 * passing this up will potentially cause the
2766 		 * upper layer to put a station in power save
2767 		 * (except when configured with MWL_HOST_PS_SUPPORT).
2768 		 */
2769 		if (wh->i_fc[1] & IEEE80211_FC1_PROTECTED)
2770 			m->m_flags |= M_WEP;
2771 #ifdef MWL_HOST_PS_SUPPORT
2772 		wh->i_fc[1] &= ~IEEE80211_FC1_PROTECTED;
2773 #else
2774 		wh->i_fc[1] &= ~(IEEE80211_FC1_PROTECTED |
2775 		    IEEE80211_FC1_PWR_MGT);
2776 #endif
2777 
2778 		if (ieee80211_radiotap_active(ic)) {
2779 			struct mwl_rx_radiotap_header *tap = &sc->sc_rx_th;
2780 
2781 			tap->wr_flags = 0;
2782 			tap->wr_rate = ds->Rate;
2783 			tap->wr_antsignal = rssi + nf;
2784 			tap->wr_antnoise = nf;
2785 		}
2786 		if (IFF_DUMPPKTS_RECV(sc, wh)) {
2787 			ieee80211_dump_pkt(ic, mtod(m, caddr_t),
2788 			    len, ds->Rate, rssi);
2789 		}
2790 		/* dispatch */
2791 		ni = ieee80211_find_rxnode(ic,
2792 		    (const struct ieee80211_frame_min *) wh);
2793 		if (ni != NULL) {
2794 			mn = MWL_NODE(ni);
2795 #ifdef MWL_ANT_INFO_SUPPORT
2796 			mn->mn_ai.rssi_a = ds->ai.rssi_a;
2797 			mn->mn_ai.rssi_b = ds->ai.rssi_b;
2798 			mn->mn_ai.rssi_c = ds->ai.rssi_c;
2799 			mn->mn_ai.rsvd1 = rssi;
2800 #endif
2801 			/* tag AMPDU aggregates for reorder processing */
2802 			if (ni->ni_flags & IEEE80211_NODE_HT)
2803 				m->m_flags |= M_AMPDU;
2804 			(void) ieee80211_input(ni, m, rssi, nf);
2805 			ieee80211_free_node(ni);
2806 		} else
2807 			(void) ieee80211_input_all(ic, m, rssi, nf);
2808 rx_next:
2809 		/* NB: ignore ENOMEM so we process more descriptors */
2810 		(void) mwl_rxbuf_init(sc, bf);
2811 		bf = STAILQ_NEXT(bf, bf_list);
2812 	}
2813 rx_stop:
2814 	sc->sc_rxnext = bf;
2815 
2816 	if (mbufq_first(&sc->sc_snd) != NULL) {
2817 		/* NB: kick fw; the tx thread may have been preempted */
2818 		mwl_hal_txstart(sc->sc_mh, 0);
2819 		mwl_start(sc);
2820 	}
2821 }
2822 
2823 static void
mwl_txq_init(struct mwl_softc * sc,struct mwl_txq * txq,int qnum)2824 mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *txq, int qnum)
2825 {
2826 	struct mwl_txbuf *bf, *bn;
2827 	struct mwl_txdesc *ds;
2828 
2829 	MWL_TXQ_LOCK_INIT(sc, txq);
2830 	txq->qnum = qnum;
2831 	txq->txpri = 0;	/* XXX */
2832 #if 0
2833 	/* NB: q setup by mwl_txdma_setup XXX */
2834 	STAILQ_INIT(&txq->free);
2835 #endif
2836 	STAILQ_FOREACH(bf, &txq->free, bf_list) {
2837 		bf->bf_txq = txq;
2838 
2839 		ds = bf->bf_desc;
2840 		bn = STAILQ_NEXT(bf, bf_list);
2841 		if (bn == NULL)
2842 			bn = STAILQ_FIRST(&txq->free);
2843 		ds->pPhysNext = htole32(bn->bf_daddr);
2844 	}
2845 	STAILQ_INIT(&txq->active);
2846 }
2847 
2848 /*
2849  * Setup a hardware data transmit queue for the specified
2850  * access control.  We record the mapping from ac's
2851  * to h/w queues for use by mwl_tx_start.
2852  */
2853 static int
mwl_tx_setup(struct mwl_softc * sc,int ac,int mvtype)2854 mwl_tx_setup(struct mwl_softc *sc, int ac, int mvtype)
2855 {
2856 	struct mwl_txq *txq;
2857 
2858 	if (ac >= nitems(sc->sc_ac2q)) {
2859 		device_printf(sc->sc_dev, "AC %u out of range, max %zu!\n",
2860 			ac, nitems(sc->sc_ac2q));
2861 		return 0;
2862 	}
2863 	if (mvtype >= MWL_NUM_TX_QUEUES) {
2864 		device_printf(sc->sc_dev, "mvtype %u out of range, max %u!\n",
2865 			mvtype, MWL_NUM_TX_QUEUES);
2866 		return 0;
2867 	}
2868 	txq = &sc->sc_txq[mvtype];
2869 	mwl_txq_init(sc, txq, mvtype);
2870 	sc->sc_ac2q[ac] = txq;
2871 	return 1;
2872 }
2873 
2874 /*
2875  * Update WME parameters for a transmit queue.
2876  */
2877 static int
mwl_txq_update(struct mwl_softc * sc,int ac)2878 mwl_txq_update(struct mwl_softc *sc, int ac)
2879 {
2880 #define	MWL_EXPONENT_TO_VALUE(v)	((1<<v)-1)
2881 	struct ieee80211com *ic = &sc->sc_ic;
2882 	struct chanAccParams chp;
2883 	struct mwl_txq *txq = sc->sc_ac2q[ac];
2884 	struct wmeParams *wmep;
2885 	struct mwl_hal *mh = sc->sc_mh;
2886 	int aifs, cwmin, cwmax, txoplim;
2887 
2888 	ieee80211_wme_ic_getparams(ic, &chp);
2889 	wmep = &chp.cap_wmeParams[ac];
2890 
2891 	aifs = wmep->wmep_aifsn;
2892 	/* XXX in sta mode need to pass log values for cwmin/max */
2893 	cwmin = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmin);
2894 	cwmax = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmax);
2895 	txoplim = wmep->wmep_txopLimit;		/* NB: units of 32us */
2896 
2897 	if (mwl_hal_setedcaparams(mh, txq->qnum, cwmin, cwmax, aifs, txoplim)) {
2898 		device_printf(sc->sc_dev, "unable to update hardware queue "
2899 			"parameters for %s traffic!\n",
2900 			ieee80211_wme_acnames[ac]);
2901 		return 0;
2902 	}
2903 	return 1;
2904 #undef MWL_EXPONENT_TO_VALUE
2905 }
2906 
2907 /*
2908  * Callback from the 802.11 layer to update WME parameters.
2909  */
2910 static int
mwl_wme_update(struct ieee80211com * ic)2911 mwl_wme_update(struct ieee80211com *ic)
2912 {
2913 	struct mwl_softc *sc = ic->ic_softc;
2914 
2915 	return !mwl_txq_update(sc, WME_AC_BE) ||
2916 	    !mwl_txq_update(sc, WME_AC_BK) ||
2917 	    !mwl_txq_update(sc, WME_AC_VI) ||
2918 	    !mwl_txq_update(sc, WME_AC_VO) ? EIO : 0;
2919 }
2920 
2921 /*
2922  * Reclaim resources for a setup queue.
2923  */
2924 static void
mwl_tx_cleanupq(struct mwl_softc * sc,struct mwl_txq * txq)2925 mwl_tx_cleanupq(struct mwl_softc *sc, struct mwl_txq *txq)
2926 {
2927 	/* XXX hal work? */
2928 	MWL_TXQ_LOCK_DESTROY(txq);
2929 }
2930 
2931 /*
2932  * Reclaim all tx queue resources.
2933  */
2934 static void
mwl_tx_cleanup(struct mwl_softc * sc)2935 mwl_tx_cleanup(struct mwl_softc *sc)
2936 {
2937 	int i;
2938 
2939 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2940 		mwl_tx_cleanupq(sc, &sc->sc_txq[i]);
2941 }
2942 
2943 static int
mwl_tx_dmasetup(struct mwl_softc * sc,struct mwl_txbuf * bf,struct mbuf * m0)2944 mwl_tx_dmasetup(struct mwl_softc *sc, struct mwl_txbuf *bf, struct mbuf *m0)
2945 {
2946 	struct mbuf *m;
2947 	int error;
2948 
2949 	/*
2950 	 * Load the DMA map so any coalescing is done.  This
2951 	 * also calculates the number of descriptors we need.
2952 	 */
2953 	error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
2954 				     bf->bf_segs, &bf->bf_nseg,
2955 				     BUS_DMA_NOWAIT);
2956 	if (error == EFBIG) {
2957 		/* XXX packet requires too many descriptors */
2958 		bf->bf_nseg = MWL_TXDESC+1;
2959 	} else if (error != 0) {
2960 		sc->sc_stats.mst_tx_busdma++;
2961 		m_freem(m0);
2962 		return error;
2963 	}
2964 	/*
2965 	 * Discard null packets and check for packets that
2966 	 * require too many TX descriptors.  We try to convert
2967 	 * the latter to a cluster.
2968 	 */
2969 	if (error == EFBIG) {		/* too many desc's, linearize */
2970 		sc->sc_stats.mst_tx_linear++;
2971 #if MWL_TXDESC > 1
2972 		m = m_collapse(m0, M_NOWAIT, MWL_TXDESC);
2973 #else
2974 		m = m_defrag(m0, M_NOWAIT);
2975 #endif
2976 		if (m == NULL) {
2977 			m_freem(m0);
2978 			sc->sc_stats.mst_tx_nombuf++;
2979 			return ENOMEM;
2980 		}
2981 		m0 = m;
2982 		error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
2983 					     bf->bf_segs, &bf->bf_nseg,
2984 					     BUS_DMA_NOWAIT);
2985 		if (error != 0) {
2986 			sc->sc_stats.mst_tx_busdma++;
2987 			m_freem(m0);
2988 			return error;
2989 		}
2990 		KASSERT(bf->bf_nseg <= MWL_TXDESC,
2991 		    ("too many segments after defrag; nseg %u", bf->bf_nseg));
2992 	} else if (bf->bf_nseg == 0) {		/* null packet, discard */
2993 		sc->sc_stats.mst_tx_nodata++;
2994 		m_freem(m0);
2995 		return EIO;
2996 	}
2997 	DPRINTF(sc, MWL_DEBUG_XMIT, "%s: m %p len %u\n",
2998 		__func__, m0, m0->m_pkthdr.len);
2999 	bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap, BUS_DMASYNC_PREWRITE);
3000 	bf->bf_m = m0;
3001 
3002 	return 0;
3003 }
3004 
3005 static __inline int
mwl_cvtlegacyrate(int rate)3006 mwl_cvtlegacyrate(int rate)
3007 {
3008 	switch (rate) {
3009 	case 2:	 return 0;
3010 	case 4:	 return 1;
3011 	case 11: return 2;
3012 	case 22: return 3;
3013 	case 44: return 4;
3014 	case 12: return 5;
3015 	case 18: return 6;
3016 	case 24: return 7;
3017 	case 36: return 8;
3018 	case 48: return 9;
3019 	case 72: return 10;
3020 	case 96: return 11;
3021 	case 108:return 12;
3022 	}
3023 	return 0;
3024 }
3025 
3026 /*
3027  * Calculate fixed tx rate information per client state;
3028  * this value is suitable for writing to the Format field
3029  * of a tx descriptor.
3030  */
3031 static uint16_t
mwl_calcformat(uint8_t rate,const struct ieee80211_node * ni)3032 mwl_calcformat(uint8_t rate, const struct ieee80211_node *ni)
3033 {
3034 	uint16_t fmt;
3035 
3036 	fmt = _IEEE80211_SHIFTMASK(3, EAGLE_TXD_ANTENNA)
3037 	    | (IEEE80211_IS_CHAN_HT40D(ni->ni_chan) ?
3038 		EAGLE_TXD_EXTCHAN_LO : EAGLE_TXD_EXTCHAN_HI);
3039 	if (rate & IEEE80211_RATE_MCS) {	/* HT MCS */
3040 		fmt |= EAGLE_TXD_FORMAT_HT
3041 		    /* NB: 0x80 implicitly stripped from ucastrate */
3042 		    | _IEEE80211_SHIFTMASK(rate, EAGLE_TXD_RATE);
3043 		/* XXX short/long GI may be wrong; re-check */
3044 		if (IEEE80211_IS_CHAN_HT40(ni->ni_chan)) {
3045 			fmt |= EAGLE_TXD_CHW_40
3046 			    | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI40 ?
3047 			        EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3048 		} else {
3049 			fmt |= EAGLE_TXD_CHW_20
3050 			    | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI20 ?
3051 			        EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3052 		}
3053 	} else {			/* legacy rate */
3054 		fmt |= EAGLE_TXD_FORMAT_LEGACY
3055 		    | _IEEE80211_SHIFTMASK(mwl_cvtlegacyrate(rate),
3056 			EAGLE_TXD_RATE)
3057 		    | EAGLE_TXD_CHW_20
3058 		    /* XXX iv_flags & IEEE80211_F_SHPREAMBLE? */
3059 		    | (ni->ni_capinfo & IEEE80211_CAPINFO_SHORT_PREAMBLE ?
3060 			EAGLE_TXD_PREAMBLE_SHORT : EAGLE_TXD_PREAMBLE_LONG);
3061 	}
3062 	return fmt;
3063 }
3064 
3065 static int
mwl_tx_start(struct mwl_softc * sc,struct ieee80211_node * ni,struct mwl_txbuf * bf,struct mbuf * m0)3066 mwl_tx_start(struct mwl_softc *sc, struct ieee80211_node *ni, struct mwl_txbuf *bf,
3067     struct mbuf *m0)
3068 {
3069 	struct ieee80211com *ic = &sc->sc_ic;
3070 	struct ieee80211vap *vap = ni->ni_vap;
3071 	int error, iswep, ismcast;
3072 	int hdrlen, pktlen;
3073 	struct mwl_txdesc *ds;
3074 	struct mwl_txq *txq;
3075 	struct ieee80211_frame *wh;
3076 	struct mwltxrec *tr;
3077 	struct mwl_node *mn;
3078 	uint16_t qos;
3079 #if MWL_TXDESC > 1
3080 	int i;
3081 #endif
3082 
3083 	wh = mtod(m0, struct ieee80211_frame *);
3084 	iswep = wh->i_fc[1] & IEEE80211_FC1_PROTECTED;
3085 	ismcast = IEEE80211_IS_MULTICAST(wh->i_addr1);
3086 	hdrlen = ieee80211_anyhdrsize(wh);
3087 	pktlen = m0->m_pkthdr.len;
3088 	if (IEEE80211_QOS_HAS_SEQ(wh)) {
3089 		qos = *(uint16_t *)ieee80211_getqos(wh);
3090 	} else
3091 		qos = 0;
3092 
3093 	if (iswep) {
3094 		const struct ieee80211_cipher *cip;
3095 		struct ieee80211_key *k;
3096 
3097 		/*
3098 		 * Construct the 802.11 header+trailer for an encrypted
3099 		 * frame. The only reason this can fail is because of an
3100 		 * unknown or unsupported cipher/key type.
3101 		 *
3102 		 * NB: we do this even though the firmware will ignore
3103 		 *     what we've done for WEP and TKIP as we need the
3104 		 *     ExtIV filled in for CCMP and this also adjusts
3105 		 *     the headers which simplifies our work below.
3106 		 */
3107 		k = ieee80211_crypto_encap(ni, m0);
3108 		if (k == NULL) {
3109 			/*
3110 			 * This can happen when the key is yanked after the
3111 			 * frame was queued.  Just discard the frame; the
3112 			 * 802.11 layer counts failures and provides
3113 			 * debugging/diagnostics.
3114 			 */
3115 			m_freem(m0);
3116 			return EIO;
3117 		}
3118 		/*
3119 		 * Adjust the packet length for the crypto additions
3120 		 * done during encap and any other bits that the f/w
3121 		 * will add later on.
3122 		 */
3123 		cip = k->wk_cipher;
3124 		pktlen += cip->ic_header + cip->ic_miclen + cip->ic_trailer;
3125 
3126 		/* packet header may have moved, reset our local pointer */
3127 		wh = mtod(m0, struct ieee80211_frame *);
3128 	}
3129 
3130 	if (ieee80211_radiotap_active_vap(vap)) {
3131 		sc->sc_tx_th.wt_flags = 0;	/* XXX */
3132 		if (iswep)
3133 			sc->sc_tx_th.wt_flags |= IEEE80211_RADIOTAP_F_WEP;
3134 #if 0
3135 		sc->sc_tx_th.wt_rate = ds->DataRate;
3136 #endif
3137 		sc->sc_tx_th.wt_txpower = ni->ni_txpower;
3138 		sc->sc_tx_th.wt_antenna = sc->sc_txantenna;
3139 
3140 		ieee80211_radiotap_tx(vap, m0);
3141 	}
3142 	/*
3143 	 * Copy up/down the 802.11 header; the firmware requires
3144 	 * we present a 2-byte payload length followed by a
3145 	 * 4-address header (w/o QoS), followed (optionally) by
3146 	 * any WEP/ExtIV header (but only filled in for CCMP).
3147 	 * We are assured the mbuf has sufficient headroom to
3148 	 * prepend in-place by the setup of ic_headroom in
3149 	 * mwl_attach.
3150 	 */
3151 	if (hdrlen < sizeof(struct mwltxrec)) {
3152 		const int space = sizeof(struct mwltxrec) - hdrlen;
3153 		if (M_LEADINGSPACE(m0) < space) {
3154 			/* NB: should never happen */
3155 			device_printf(sc->sc_dev,
3156 			    "not enough headroom, need %d found %zd, "
3157 			    "m_flags 0x%x m_len %d\n",
3158 			    space, M_LEADINGSPACE(m0), m0->m_flags, m0->m_len);
3159 			ieee80211_dump_pkt(ic,
3160 			    mtod(m0, const uint8_t *), m0->m_len, 0, -1);
3161 			m_freem(m0);
3162 			sc->sc_stats.mst_tx_noheadroom++;
3163 			return EIO;
3164 		}
3165 		M_PREPEND(m0, space, M_NOWAIT);
3166 	}
3167 	tr = mtod(m0, struct mwltxrec *);
3168 	if (wh != (struct ieee80211_frame *) &tr->wh)
3169 		ovbcopy(wh, &tr->wh, hdrlen);
3170 	/*
3171 	 * Note: the "firmware length" is actually the length
3172 	 * of the fully formed "802.11 payload".  That is, it's
3173 	 * everything except for the 802.11 header.  In particular
3174 	 * this includes all crypto material including the MIC!
3175 	 */
3176 	tr->fwlen = htole16(pktlen - hdrlen);
3177 
3178 	/*
3179 	 * Load the DMA map so any coalescing is done.  This
3180 	 * also calculates the number of descriptors we need.
3181 	 */
3182 	error = mwl_tx_dmasetup(sc, bf, m0);
3183 	if (error != 0) {
3184 		/* NB: stat collected in mwl_tx_dmasetup */
3185 		DPRINTF(sc, MWL_DEBUG_XMIT,
3186 		    "%s: unable to setup dma\n", __func__);
3187 		return error;
3188 	}
3189 	bf->bf_node = ni;			/* NB: held reference */
3190 	m0 = bf->bf_m;				/* NB: may have changed */
3191 	tr = mtod(m0, struct mwltxrec *);
3192 	wh = (struct ieee80211_frame *)&tr->wh;
3193 
3194 	/*
3195 	 * Formulate tx descriptor.
3196 	 */
3197 	ds = bf->bf_desc;
3198 	txq = bf->bf_txq;
3199 
3200 	ds->QosCtrl = qos;			/* NB: already little-endian */
3201 #if MWL_TXDESC == 1
3202 	/*
3203 	 * NB: multiframes should be zero because the descriptors
3204 	 *     are initialized to zero.  This should handle the case
3205 	 *     where the driver is built with MWL_TXDESC=1 but we are
3206 	 *     using firmware with multi-segment support.
3207 	 */
3208 	ds->PktPtr = htole32(bf->bf_segs[0].ds_addr);
3209 	ds->PktLen = htole16(bf->bf_segs[0].ds_len);
3210 #else
3211 	ds->multiframes = htole32(bf->bf_nseg);
3212 	ds->PktLen = htole16(m0->m_pkthdr.len);
3213 	for (i = 0; i < bf->bf_nseg; i++) {
3214 		ds->PktPtrArray[i] = htole32(bf->bf_segs[i].ds_addr);
3215 		ds->PktLenArray[i] = htole16(bf->bf_segs[i].ds_len);
3216 	}
3217 #endif
3218 	/* NB: pPhysNext, DataRate, and SapPktInfo setup once, don't touch */
3219 	ds->Format = 0;
3220 	ds->pad = 0;
3221 	ds->ack_wcb_addr = 0;
3222 
3223 	mn = MWL_NODE(ni);
3224 	/*
3225 	 * Select transmit rate.
3226 	 */
3227 	switch (wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK) {
3228 	case IEEE80211_FC0_TYPE_MGT:
3229 		sc->sc_stats.mst_tx_mgmt++;
3230 		/* fall thru... */
3231 	case IEEE80211_FC0_TYPE_CTL:
3232 		/* NB: assign to BE q to avoid bursting */
3233 		ds->TxPriority = MWL_WME_AC_BE;
3234 		break;
3235 	case IEEE80211_FC0_TYPE_DATA:
3236 		if (!ismcast) {
3237 			const struct ieee80211_txparam *tp = ni->ni_txparms;
3238 			/*
3239 			 * EAPOL frames get forced to a fixed rate and w/o
3240 			 * aggregation; otherwise check for any fixed rate
3241 			 * for the client (may depend on association state).
3242 			 */
3243 			if (m0->m_flags & M_EAPOL) {
3244 				const struct mwl_vap *mvp = MWL_VAP_CONST(vap);
3245 				ds->Format = mvp->mv_eapolformat;
3246 				ds->pad = htole16(
3247 				    EAGLE_TXD_FIXED_RATE | EAGLE_TXD_DONT_AGGR);
3248 			} else if (tp->ucastrate != IEEE80211_FIXED_RATE_NONE) {
3249 				/* XXX pre-calculate per node */
3250 				ds->Format = htole16(
3251 				    mwl_calcformat(tp->ucastrate, ni));
3252 				ds->pad = htole16(EAGLE_TXD_FIXED_RATE);
3253 			}
3254 			/* NB: EAPOL frames will never have qos set */
3255 			if (qos == 0)
3256 				ds->TxPriority = txq->qnum;
3257 #if MWL_MAXBA > 3
3258 			else if (mwl_bastream_match(&mn->mn_ba[3], qos))
3259 				ds->TxPriority = mn->mn_ba[3].txq;
3260 #endif
3261 #if MWL_MAXBA > 2
3262 			else if (mwl_bastream_match(&mn->mn_ba[2], qos))
3263 				ds->TxPriority = mn->mn_ba[2].txq;
3264 #endif
3265 #if MWL_MAXBA > 1
3266 			else if (mwl_bastream_match(&mn->mn_ba[1], qos))
3267 				ds->TxPriority = mn->mn_ba[1].txq;
3268 #endif
3269 #if MWL_MAXBA > 0
3270 			else if (mwl_bastream_match(&mn->mn_ba[0], qos))
3271 				ds->TxPriority = mn->mn_ba[0].txq;
3272 #endif
3273 			else
3274 				ds->TxPriority = txq->qnum;
3275 		} else
3276 			ds->TxPriority = txq->qnum;
3277 		break;
3278 	default:
3279 		device_printf(sc->sc_dev, "bogus frame type 0x%x (%s)\n",
3280 			wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK, __func__);
3281 		sc->sc_stats.mst_tx_badframetype++;
3282 		m_freem(m0);
3283 		return EIO;
3284 	}
3285 
3286 	if (IFF_DUMPPKTS_XMIT(sc))
3287 		ieee80211_dump_pkt(ic,
3288 		    mtod(m0, const uint8_t *)+sizeof(uint16_t),
3289 		    m0->m_len - sizeof(uint16_t), ds->DataRate, -1);
3290 
3291 	MWL_TXQ_LOCK(txq);
3292 	ds->Status = htole32(EAGLE_TXD_STATUS_FW_OWNED);
3293 	STAILQ_INSERT_TAIL(&txq->active, bf, bf_list);
3294 	MWL_TXDESC_SYNC(txq, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
3295 
3296 	sc->sc_tx_timer = 5;
3297 	MWL_TXQ_UNLOCK(txq);
3298 
3299 	return 0;
3300 }
3301 
3302 static __inline int
mwl_cvtlegacyrix(int rix)3303 mwl_cvtlegacyrix(int rix)
3304 {
3305 	static const int ieeerates[] =
3306 	    { 2, 4, 11, 22, 44, 12, 18, 24, 36, 48, 72, 96, 108 };
3307 	return (rix < nitems(ieeerates) ? ieeerates[rix] : 0);
3308 }
3309 
3310 /*
3311  * Process completed xmit descriptors from the specified queue.
3312  */
3313 static int
mwl_tx_processq(struct mwl_softc * sc,struct mwl_txq * txq)3314 mwl_tx_processq(struct mwl_softc *sc, struct mwl_txq *txq)
3315 {
3316 #define	EAGLE_TXD_STATUS_MCAST \
3317 	(EAGLE_TXD_STATUS_MULTICAST_TX | EAGLE_TXD_STATUS_BROADCAST_TX)
3318 	struct ieee80211com *ic = &sc->sc_ic;
3319 	struct mwl_txbuf *bf;
3320 	struct mwl_txdesc *ds;
3321 	struct ieee80211_node *ni;
3322 	int nreaped;
3323 	uint32_t status;
3324 
3325 	DPRINTF(sc, MWL_DEBUG_TX_PROC, "%s: tx queue %u\n", __func__, txq->qnum);
3326 	for (nreaped = 0;; nreaped++) {
3327 		MWL_TXQ_LOCK(txq);
3328 		bf = STAILQ_FIRST(&txq->active);
3329 		if (bf == NULL) {
3330 			MWL_TXQ_UNLOCK(txq);
3331 			break;
3332 		}
3333 		ds = bf->bf_desc;
3334 		MWL_TXDESC_SYNC(txq, ds,
3335 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
3336 		if (ds->Status & htole32(EAGLE_TXD_STATUS_FW_OWNED)) {
3337 			MWL_TXQ_UNLOCK(txq);
3338 			break;
3339 		}
3340 		STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3341 		MWL_TXQ_UNLOCK(txq);
3342 
3343 #ifdef MWL_DEBUG
3344 		if (sc->sc_debug & MWL_DEBUG_XMIT_DESC)
3345 			mwl_printtxbuf(bf, txq->qnum, nreaped);
3346 #endif
3347 		ni = bf->bf_node;
3348 		if (ni != NULL) {
3349 			status = le32toh(ds->Status);
3350 			if (status & EAGLE_TXD_STATUS_OK) {
3351 				uint16_t Format = le16toh(ds->Format);
3352 				uint8_t txant = _IEEE80211_MASKSHIFT(Format,
3353 				    EAGLE_TXD_ANTENNA);
3354 
3355 				sc->sc_stats.mst_ant_tx[txant]++;
3356 				if (status & EAGLE_TXD_STATUS_OK_RETRY)
3357 					sc->sc_stats.mst_tx_retries++;
3358 				if (status & EAGLE_TXD_STATUS_OK_MORE_RETRY)
3359 					sc->sc_stats.mst_tx_mretries++;
3360 				if (txq->qnum >= MWL_WME_AC_VO)
3361 					ic->ic_wme.wme_hipri_traffic++;
3362 				ni->ni_txrate = _IEEE80211_MASKSHIFT(Format,
3363 				    EAGLE_TXD_RATE);
3364 				if ((Format & EAGLE_TXD_FORMAT_HT) == 0) {
3365 					ni->ni_txrate = mwl_cvtlegacyrix(
3366 					    ni->ni_txrate);
3367 				} else
3368 					ni->ni_txrate |= IEEE80211_RATE_MCS;
3369 				sc->sc_stats.mst_tx_rate = ni->ni_txrate;
3370 			} else {
3371 				if (status & EAGLE_TXD_STATUS_FAILED_LINK_ERROR)
3372 					sc->sc_stats.mst_tx_linkerror++;
3373 				if (status & EAGLE_TXD_STATUS_FAILED_XRETRY)
3374 					sc->sc_stats.mst_tx_xretries++;
3375 				if (status & EAGLE_TXD_STATUS_FAILED_AGING)
3376 					sc->sc_stats.mst_tx_aging++;
3377 				if (bf->bf_m->m_flags & M_FF)
3378 					sc->sc_stats.mst_ff_txerr++;
3379 			}
3380 			if (bf->bf_m->m_flags & M_TXCB)
3381 				/* XXX strip fw len in case header inspected */
3382 				m_adj(bf->bf_m, sizeof(uint16_t));
3383 			ieee80211_tx_complete(ni, bf->bf_m,
3384 			    (status & EAGLE_TXD_STATUS_OK) == 0);
3385 		} else
3386 			m_freem(bf->bf_m);
3387 		ds->Status = htole32(EAGLE_TXD_STATUS_IDLE);
3388 
3389 		bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap,
3390 		    BUS_DMASYNC_POSTWRITE);
3391 		bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3392 
3393 		mwl_puttxbuf_tail(txq, bf);
3394 	}
3395 	return nreaped;
3396 #undef EAGLE_TXD_STATUS_MCAST
3397 }
3398 
3399 /*
3400  * Deferred processing of transmit interrupt; special-cased
3401  * for four hardware queues, 0-3.
3402  */
3403 static void
mwl_tx_proc(void * arg,int npending)3404 mwl_tx_proc(void *arg, int npending)
3405 {
3406 	struct mwl_softc *sc = arg;
3407 	int nreaped;
3408 
3409 	/*
3410 	 * Process each active queue.
3411 	 */
3412 	nreaped = 0;
3413 	if (!STAILQ_EMPTY(&sc->sc_txq[0].active))
3414 		nreaped += mwl_tx_processq(sc, &sc->sc_txq[0]);
3415 	if (!STAILQ_EMPTY(&sc->sc_txq[1].active))
3416 		nreaped += mwl_tx_processq(sc, &sc->sc_txq[1]);
3417 	if (!STAILQ_EMPTY(&sc->sc_txq[2].active))
3418 		nreaped += mwl_tx_processq(sc, &sc->sc_txq[2]);
3419 	if (!STAILQ_EMPTY(&sc->sc_txq[3].active))
3420 		nreaped += mwl_tx_processq(sc, &sc->sc_txq[3]);
3421 
3422 	if (nreaped != 0) {
3423 		sc->sc_tx_timer = 0;
3424 		if (mbufq_first(&sc->sc_snd) != NULL) {
3425 			/* NB: kick fw; the tx thread may have been preempted */
3426 			mwl_hal_txstart(sc->sc_mh, 0);
3427 			mwl_start(sc);
3428 		}
3429 	}
3430 }
3431 
3432 static void
mwl_tx_draintxq(struct mwl_softc * sc,struct mwl_txq * txq)3433 mwl_tx_draintxq(struct mwl_softc *sc, struct mwl_txq *txq)
3434 {
3435 	struct ieee80211_node *ni;
3436 	struct mwl_txbuf *bf;
3437 	u_int ix __unused;
3438 
3439 	/*
3440 	 * NB: this assumes output has been stopped and
3441 	 *     we do not need to block mwl_tx_tasklet
3442 	 */
3443 	for (ix = 0;; ix++) {
3444 		MWL_TXQ_LOCK(txq);
3445 		bf = STAILQ_FIRST(&txq->active);
3446 		if (bf == NULL) {
3447 			MWL_TXQ_UNLOCK(txq);
3448 			break;
3449 		}
3450 		STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3451 		MWL_TXQ_UNLOCK(txq);
3452 #ifdef MWL_DEBUG
3453 		if (sc->sc_debug & MWL_DEBUG_RESET) {
3454 			struct ieee80211com *ic = &sc->sc_ic;
3455 			const struct mwltxrec *tr =
3456 			    mtod(bf->bf_m, const struct mwltxrec *);
3457 			mwl_printtxbuf(bf, txq->qnum, ix);
3458 			ieee80211_dump_pkt(ic, (const uint8_t *)&tr->wh,
3459 				bf->bf_m->m_len - sizeof(tr->fwlen), 0, -1);
3460 		}
3461 #endif /* MWL_DEBUG */
3462 		bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3463 		ni = bf->bf_node;
3464 		if (ni != NULL) {
3465 			/*
3466 			 * Reclaim node reference.
3467 			 */
3468 			ieee80211_free_node(ni);
3469 		}
3470 		m_freem(bf->bf_m);
3471 
3472 		mwl_puttxbuf_tail(txq, bf);
3473 	}
3474 }
3475 
3476 /*
3477  * Drain the transmit queues and reclaim resources.
3478  */
3479 static void
mwl_draintxq(struct mwl_softc * sc)3480 mwl_draintxq(struct mwl_softc *sc)
3481 {
3482 	int i;
3483 
3484 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3485 		mwl_tx_draintxq(sc, &sc->sc_txq[i]);
3486 	sc->sc_tx_timer = 0;
3487 }
3488 
3489 #ifdef MWL_DIAGAPI
3490 /*
3491  * Reset the transmit queues to a pristine state after a fw download.
3492  */
3493 static void
mwl_resettxq(struct mwl_softc * sc)3494 mwl_resettxq(struct mwl_softc *sc)
3495 {
3496 	int i;
3497 
3498 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3499 		mwl_txq_reset(sc, &sc->sc_txq[i]);
3500 }
3501 #endif /* MWL_DIAGAPI */
3502 
3503 /*
3504  * Clear the transmit queues of any frames submitted for the
3505  * specified vap.  This is done when the vap is deleted so we
3506  * don't potentially reference the vap after it is gone.
3507  * Note we cannot remove the frames; we only reclaim the node
3508  * reference.
3509  */
3510 static void
mwl_cleartxq(struct mwl_softc * sc,struct ieee80211vap * vap)3511 mwl_cleartxq(struct mwl_softc *sc, struct ieee80211vap *vap)
3512 {
3513 	struct mwl_txq *txq;
3514 	struct mwl_txbuf *bf;
3515 	int i;
3516 
3517 	for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
3518 		txq = &sc->sc_txq[i];
3519 		MWL_TXQ_LOCK(txq);
3520 		STAILQ_FOREACH(bf, &txq->active, bf_list) {
3521 			struct ieee80211_node *ni = bf->bf_node;
3522 			if (ni != NULL && ni->ni_vap == vap) {
3523 				bf->bf_node = NULL;
3524 				ieee80211_free_node(ni);
3525 			}
3526 		}
3527 		MWL_TXQ_UNLOCK(txq);
3528 	}
3529 }
3530 
3531 static int
mwl_recv_action(struct ieee80211_node * ni,const struct ieee80211_frame * wh,const uint8_t * frm,const uint8_t * efrm)3532 mwl_recv_action(struct ieee80211_node *ni, const struct ieee80211_frame *wh,
3533 	const uint8_t *frm, const uint8_t *efrm)
3534 {
3535 	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3536 	const struct ieee80211_action *ia;
3537 
3538 	ia = (const struct ieee80211_action *) frm;
3539 	if (ia->ia_category == IEEE80211_ACTION_CAT_HT &&
3540 	    ia->ia_action == IEEE80211_ACTION_HT_MIMOPWRSAVE) {
3541 		const struct ieee80211_action_ht_mimopowersave *mps =
3542 		    (const struct ieee80211_action_ht_mimopowersave *) ia;
3543 
3544 		mwl_hal_setmimops(sc->sc_mh, ni->ni_macaddr,
3545 		    mps->am_control & IEEE80211_A_HT_MIMOPWRSAVE_ENA,
3546 		    _IEEE80211_MASKSHIFT(mps->am_control,
3547 			IEEE80211_A_HT_MIMOPWRSAVE_MODE));
3548 		return 0;
3549 	} else
3550 		return sc->sc_recv_action(ni, wh, frm, efrm);
3551 }
3552 
3553 static int
mwl_addba_request(struct ieee80211_node * ni,struct ieee80211_tx_ampdu * tap,int dialogtoken,int baparamset,int batimeout)3554 mwl_addba_request(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3555 	int dialogtoken, int baparamset, int batimeout)
3556 {
3557 	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3558 	struct ieee80211vap *vap = ni->ni_vap;
3559 	struct mwl_node *mn = MWL_NODE(ni);
3560 	struct mwl_bastate *bas;
3561 
3562 	bas = tap->txa_private;
3563 	if (bas == NULL) {
3564 		const MWL_HAL_BASTREAM *sp;
3565 		/*
3566 		 * Check for a free BA stream slot.
3567 		 */
3568 #if MWL_MAXBA > 3
3569 		if (mn->mn_ba[3].bastream == NULL)
3570 			bas = &mn->mn_ba[3];
3571 		else
3572 #endif
3573 #if MWL_MAXBA > 2
3574 		if (mn->mn_ba[2].bastream == NULL)
3575 			bas = &mn->mn_ba[2];
3576 		else
3577 #endif
3578 #if MWL_MAXBA > 1
3579 		if (mn->mn_ba[1].bastream == NULL)
3580 			bas = &mn->mn_ba[1];
3581 		else
3582 #endif
3583 #if MWL_MAXBA > 0
3584 		if (mn->mn_ba[0].bastream == NULL)
3585 			bas = &mn->mn_ba[0];
3586 		else
3587 #endif
3588 		{
3589 			/* sta already has max BA streams */
3590 			/* XXX assign BA stream to highest priority tid */
3591 			DPRINTF(sc, MWL_DEBUG_AMPDU,
3592 			    "%s: already has max bastreams\n", __func__);
3593 			sc->sc_stats.mst_ampdu_reject++;
3594 			return 0;
3595 		}
3596 		/* NB: no held reference to ni */
3597 		sp = mwl_hal_bastream_alloc(MWL_VAP(vap)->mv_hvap,
3598 		    (baparamset & IEEE80211_BAPS_POLICY_IMMEDIATE) != 0,
3599 		    ni->ni_macaddr, tap->txa_tid, ni->ni_htparam,
3600 		    ni, tap);
3601 		if (sp == NULL) {
3602 			/*
3603 			 * No available stream, return 0 so no
3604 			 * a-mpdu aggregation will be done.
3605 			 */
3606 			DPRINTF(sc, MWL_DEBUG_AMPDU,
3607 			    "%s: no bastream available\n", __func__);
3608 			sc->sc_stats.mst_ampdu_nostream++;
3609 			return 0;
3610 		}
3611 		DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: alloc bastream %p\n",
3612 		    __func__, sp);
3613 		/* NB: qos is left zero so we won't match in mwl_tx_start */
3614 		bas->bastream = sp;
3615 		tap->txa_private = bas;
3616 	}
3617 	/* fetch current seq# from the firmware; if available */
3618 	if (mwl_hal_bastream_get_seqno(sc->sc_mh, bas->bastream,
3619 	    vap->iv_opmode == IEEE80211_M_STA ? vap->iv_myaddr : ni->ni_macaddr,
3620 	    &tap->txa_start) != 0)
3621 		tap->txa_start = 0;
3622 	return sc->sc_addba_request(ni, tap, dialogtoken, baparamset, batimeout);
3623 }
3624 
3625 static int
mwl_addba_response(struct ieee80211_node * ni,struct ieee80211_tx_ampdu * tap,int code,int baparamset,int batimeout)3626 mwl_addba_response(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3627 	int code, int baparamset, int batimeout)
3628 {
3629 	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3630 	struct mwl_bastate *bas;
3631 
3632 	bas = tap->txa_private;
3633 	if (bas == NULL) {
3634 		/* XXX should not happen */
3635 		DPRINTF(sc, MWL_DEBUG_AMPDU,
3636 		    "%s: no BA stream allocated, TID %d\n",
3637 		    __func__, tap->txa_tid);
3638 		sc->sc_stats.mst_addba_nostream++;
3639 		return 0;
3640 	}
3641 	if (code == IEEE80211_STATUS_SUCCESS) {
3642 		struct ieee80211vap *vap = ni->ni_vap;
3643 		int bufsiz, error;
3644 
3645 		/*
3646 		 * Tell the firmware to setup the BA stream;
3647 		 * we know resources are available because we
3648 		 * pre-allocated one before forming the request.
3649 		 */
3650 		bufsiz = _IEEE80211_MASKSHIFT(baparamset, IEEE80211_BAPS_BUFSIZ);
3651 		if (bufsiz == 0)
3652 			bufsiz = IEEE80211_AGGR_BAWMAX;
3653 		error = mwl_hal_bastream_create(MWL_VAP(vap)->mv_hvap,
3654 		    bas->bastream, bufsiz, bufsiz, tap->txa_start);
3655 		if (error != 0) {
3656 			/*
3657 			 * Setup failed, return immediately so no a-mpdu
3658 			 * aggregation will be done.
3659 			 */
3660 			mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3661 			mwl_bastream_free(bas);
3662 			tap->txa_private = NULL;
3663 
3664 			DPRINTF(sc, MWL_DEBUG_AMPDU,
3665 			    "%s: create failed, error %d, bufsiz %d TID %d "
3666 			    "htparam 0x%x\n", __func__, error, bufsiz,
3667 			    tap->txa_tid, ni->ni_htparam);
3668 			sc->sc_stats.mst_bacreate_failed++;
3669 			return 0;
3670 		}
3671 		/* NB: cache txq to avoid ptr indirect */
3672 		mwl_bastream_setup(bas, tap->txa_tid, bas->bastream->txq);
3673 		DPRINTF(sc, MWL_DEBUG_AMPDU,
3674 		    "%s: bastream %p assigned to txq %d TID %d bufsiz %d "
3675 		    "htparam 0x%x\n", __func__, bas->bastream,
3676 		    bas->txq, tap->txa_tid, bufsiz, ni->ni_htparam);
3677 	} else {
3678 		/*
3679 		 * Other side NAK'd us; return the resources.
3680 		 */
3681 		DPRINTF(sc, MWL_DEBUG_AMPDU,
3682 		    "%s: request failed with code %d, destroy bastream %p\n",
3683 		    __func__, code, bas->bastream);
3684 		mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3685 		mwl_bastream_free(bas);
3686 		tap->txa_private = NULL;
3687 	}
3688 	/* NB: firmware sends BAR so we don't need to */
3689 	return sc->sc_addba_response(ni, tap, code, baparamset, batimeout);
3690 }
3691 
3692 static void
mwl_addba_stop(struct ieee80211_node * ni,struct ieee80211_tx_ampdu * tap)3693 mwl_addba_stop(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap)
3694 {
3695 	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3696 	struct mwl_bastate *bas;
3697 
3698 	bas = tap->txa_private;
3699 	if (bas != NULL) {
3700 		DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: destroy bastream %p\n",
3701 		    __func__, bas->bastream);
3702 		mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3703 		mwl_bastream_free(bas);
3704 		tap->txa_private = NULL;
3705 	}
3706 	sc->sc_addba_stop(ni, tap);
3707 }
3708 
3709 /*
3710  * Setup the rx data structures.  This should only be
3711  * done once or we may get out of sync with the firmware.
3712  */
3713 static int
mwl_startrecv(struct mwl_softc * sc)3714 mwl_startrecv(struct mwl_softc *sc)
3715 {
3716 	if (!sc->sc_recvsetup) {
3717 		struct mwl_rxbuf *bf, *prev;
3718 		struct mwl_rxdesc *ds;
3719 
3720 		prev = NULL;
3721 		STAILQ_FOREACH(bf, &sc->sc_rxbuf, bf_list) {
3722 			int error = mwl_rxbuf_init(sc, bf);
3723 			if (error != 0) {
3724 				DPRINTF(sc, MWL_DEBUG_RECV,
3725 					"%s: mwl_rxbuf_init failed %d\n",
3726 					__func__, error);
3727 				return error;
3728 			}
3729 			if (prev != NULL) {
3730 				ds = prev->bf_desc;
3731 				ds->pPhysNext = htole32(bf->bf_daddr);
3732 			}
3733 			prev = bf;
3734 		}
3735 		if (prev != NULL) {
3736 			ds = prev->bf_desc;
3737 			ds->pPhysNext =
3738 			    htole32(STAILQ_FIRST(&sc->sc_rxbuf)->bf_daddr);
3739 		}
3740 		sc->sc_recvsetup = 1;
3741 	}
3742 	mwl_mode_init(sc);		/* set filters, etc. */
3743 	return 0;
3744 }
3745 
3746 static MWL_HAL_APMODE
mwl_getapmode(const struct ieee80211vap * vap,struct ieee80211_channel * chan)3747 mwl_getapmode(const struct ieee80211vap *vap, struct ieee80211_channel *chan)
3748 {
3749 	MWL_HAL_APMODE mode;
3750 
3751 	if (IEEE80211_IS_CHAN_HT(chan)) {
3752 		if (vap->iv_flags_ht & IEEE80211_FHT_PUREN)
3753 			mode = AP_MODE_N_ONLY;
3754 		else if (IEEE80211_IS_CHAN_5GHZ(chan))
3755 			mode = AP_MODE_AandN;
3756 		else if (vap->iv_flags & IEEE80211_F_PUREG)
3757 			mode = AP_MODE_GandN;
3758 		else
3759 			mode = AP_MODE_BandGandN;
3760 	} else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3761 		if (vap->iv_flags & IEEE80211_F_PUREG)
3762 			mode = AP_MODE_G_ONLY;
3763 		else
3764 			mode = AP_MODE_MIXED;
3765 	} else if (IEEE80211_IS_CHAN_B(chan))
3766 		mode = AP_MODE_B_ONLY;
3767 	else if (IEEE80211_IS_CHAN_A(chan))
3768 		mode = AP_MODE_A_ONLY;
3769 	else
3770 		mode = AP_MODE_MIXED;		/* XXX should not happen? */
3771 	return mode;
3772 }
3773 
3774 static int
mwl_setapmode(struct ieee80211vap * vap,struct ieee80211_channel * chan)3775 mwl_setapmode(struct ieee80211vap *vap, struct ieee80211_channel *chan)
3776 {
3777 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
3778 	return mwl_hal_setapmode(hvap, mwl_getapmode(vap, chan));
3779 }
3780 
3781 /*
3782  * Set/change channels.
3783  */
3784 static int
mwl_chan_set(struct mwl_softc * sc,struct ieee80211_channel * chan)3785 mwl_chan_set(struct mwl_softc *sc, struct ieee80211_channel *chan)
3786 {
3787 	struct mwl_hal *mh = sc->sc_mh;
3788 	struct ieee80211com *ic = &sc->sc_ic;
3789 	MWL_HAL_CHANNEL hchan;
3790 	int maxtxpow;
3791 
3792 	DPRINTF(sc, MWL_DEBUG_RESET, "%s: chan %u MHz/flags 0x%x\n",
3793 	    __func__, chan->ic_freq, chan->ic_flags);
3794 
3795 	/*
3796 	 * Convert to a HAL channel description with
3797 	 * the flags constrained to reflect the current
3798 	 * operating mode.
3799 	 */
3800 	mwl_mapchan(&hchan, chan);
3801 	mwl_hal_intrset(mh, 0);		/* disable interrupts */
3802 #if 0
3803 	mwl_draintxq(sc);		/* clear pending tx frames */
3804 #endif
3805 	mwl_hal_setchannel(mh, &hchan);
3806 	/*
3807 	 * Tx power is cap'd by the regulatory setting and
3808 	 * possibly a user-set limit.  We pass the min of
3809 	 * these to the hal to apply them to the cal data
3810 	 * for this channel.
3811 	 * XXX min bound?
3812 	 */
3813 	maxtxpow = 2*chan->ic_maxregpower;
3814 	if (maxtxpow > ic->ic_txpowlimit)
3815 		maxtxpow = ic->ic_txpowlimit;
3816 	mwl_hal_settxpower(mh, &hchan, maxtxpow / 2);
3817 	/* NB: potentially change mcast/mgt rates */
3818 	mwl_setcurchanrates(sc);
3819 
3820 	/*
3821 	 * Update internal state.
3822 	 */
3823 	sc->sc_tx_th.wt_chan_freq = htole16(chan->ic_freq);
3824 	sc->sc_rx_th.wr_chan_freq = htole16(chan->ic_freq);
3825 	if (IEEE80211_IS_CHAN_A(chan)) {
3826 		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_A);
3827 		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_A);
3828 	} else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3829 		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_G);
3830 		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_G);
3831 	} else {
3832 		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_B);
3833 		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_B);
3834 	}
3835 	sc->sc_curchan = hchan;
3836 	mwl_hal_intrset(mh, sc->sc_imask);
3837 
3838 	return 0;
3839 }
3840 
3841 static void
mwl_scan_start(struct ieee80211com * ic)3842 mwl_scan_start(struct ieee80211com *ic)
3843 {
3844 	struct mwl_softc *sc = ic->ic_softc;
3845 
3846 	DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3847 }
3848 
3849 static void
mwl_scan_end(struct ieee80211com * ic)3850 mwl_scan_end(struct ieee80211com *ic)
3851 {
3852 	struct mwl_softc *sc = ic->ic_softc;
3853 
3854 	DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3855 }
3856 
3857 static void
mwl_set_channel(struct ieee80211com * ic)3858 mwl_set_channel(struct ieee80211com *ic)
3859 {
3860 	struct mwl_softc *sc = ic->ic_softc;
3861 
3862 	(void) mwl_chan_set(sc, ic->ic_curchan);
3863 }
3864 
3865 /*
3866  * Handle a channel switch request.  We inform the firmware
3867  * and mark the global state to suppress various actions.
3868  * NB: we issue only one request to the fw; we may be called
3869  * multiple times if there are multiple vap's.
3870  */
3871 static void
mwl_startcsa(struct ieee80211vap * vap)3872 mwl_startcsa(struct ieee80211vap *vap)
3873 {
3874 	struct ieee80211com *ic = vap->iv_ic;
3875 	struct mwl_softc *sc = ic->ic_softc;
3876 	MWL_HAL_CHANNEL hchan;
3877 
3878 	if (sc->sc_csapending)
3879 		return;
3880 
3881 	mwl_mapchan(&hchan, ic->ic_csa_newchan);
3882 	/* 1 =>'s quiet channel */
3883 	mwl_hal_setchannelswitchie(sc->sc_mh, &hchan, 1, ic->ic_csa_count);
3884 	sc->sc_csapending = 1;
3885 }
3886 
3887 /*
3888  * Plumb any static WEP key for the station.  This is
3889  * necessary as we must propagate the key from the
3890  * global key table of the vap to each sta db entry.
3891  */
3892 static void
mwl_setanywepkey(struct ieee80211vap * vap,const uint8_t mac[IEEE80211_ADDR_LEN])3893 mwl_setanywepkey(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
3894 {
3895 	if ((vap->iv_flags & (IEEE80211_F_PRIVACY|IEEE80211_F_WPA)) ==
3896 		IEEE80211_F_PRIVACY &&
3897 	    vap->iv_def_txkey != IEEE80211_KEYIX_NONE &&
3898 	    vap->iv_nw_keys[vap->iv_def_txkey].wk_keyix != IEEE80211_KEYIX_NONE)
3899 		(void) _mwl_key_set(vap, &vap->iv_nw_keys[vap->iv_def_txkey],
3900 				    mac);
3901 }
3902 
3903 static int
mwl_peerstadb(struct ieee80211_node * ni,int aid,int staid,MWL_HAL_PEERINFO * pi)3904 mwl_peerstadb(struct ieee80211_node *ni, int aid, int staid, MWL_HAL_PEERINFO *pi)
3905 {
3906 #define	WME(ie) ((const struct ieee80211_wme_info *) ie)
3907 	struct ieee80211vap *vap = ni->ni_vap;
3908 	struct mwl_hal_vap *hvap;
3909 	int error;
3910 
3911 	if (vap->iv_opmode == IEEE80211_M_WDS) {
3912 		/*
3913 		 * WDS vap's do not have a f/w vap; instead they piggyback
3914 		 * on an AP vap and we must install the sta db entry and
3915 		 * crypto state using that AP's handle (the WDS vap has none).
3916 		 */
3917 		hvap = MWL_VAP(vap)->mv_ap_hvap;
3918 	} else
3919 		hvap = MWL_VAP(vap)->mv_hvap;
3920 	error = mwl_hal_newstation(hvap, ni->ni_macaddr,
3921 	    aid, staid, pi,
3922 	    ni->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT),
3923 	    ni->ni_ies.wme_ie != NULL ? WME(ni->ni_ies.wme_ie)->wme_info : 0);
3924 	if (error == 0) {
3925 		/*
3926 		 * Setup security for this station.  For sta mode this is
3927 		 * needed even though do the same thing on transition to
3928 		 * AUTH state because the call to mwl_hal_newstation
3929 		 * clobbers the crypto state we setup.
3930 		 */
3931 		mwl_setanywepkey(vap, ni->ni_macaddr);
3932 	}
3933 	return error;
3934 #undef WME
3935 }
3936 
3937 static void
mwl_setglobalkeys(struct ieee80211vap * vap)3938 mwl_setglobalkeys(struct ieee80211vap *vap)
3939 {
3940 	struct ieee80211_key *wk;
3941 
3942 	wk = &vap->iv_nw_keys[0];
3943 	for (; wk < &vap->iv_nw_keys[IEEE80211_WEP_NKID]; wk++)
3944 		if (wk->wk_keyix != IEEE80211_KEYIX_NONE)
3945 			(void) _mwl_key_set(vap, wk, vap->iv_myaddr);
3946 }
3947 
3948 /*
3949  * Convert a legacy rate set to a firmware bitmask.
3950  */
3951 static uint32_t
get_rate_bitmap(const struct ieee80211_rateset * rs)3952 get_rate_bitmap(const struct ieee80211_rateset *rs)
3953 {
3954 	uint32_t rates;
3955 	int i;
3956 
3957 	rates = 0;
3958 	for (i = 0; i < rs->rs_nrates; i++)
3959 		switch (rs->rs_rates[i] & IEEE80211_RATE_VAL) {
3960 		case 2:	  rates |= 0x001; break;
3961 		case 4:	  rates |= 0x002; break;
3962 		case 11:  rates |= 0x004; break;
3963 		case 22:  rates |= 0x008; break;
3964 		case 44:  rates |= 0x010; break;
3965 		case 12:  rates |= 0x020; break;
3966 		case 18:  rates |= 0x040; break;
3967 		case 24:  rates |= 0x080; break;
3968 		case 36:  rates |= 0x100; break;
3969 		case 48:  rates |= 0x200; break;
3970 		case 72:  rates |= 0x400; break;
3971 		case 96:  rates |= 0x800; break;
3972 		case 108: rates |= 0x1000; break;
3973 		}
3974 	return rates;
3975 }
3976 
3977 /*
3978  * Construct an HT firmware bitmask from an HT rate set.
3979  */
3980 static uint32_t
get_htrate_bitmap(const struct ieee80211_htrateset * rs)3981 get_htrate_bitmap(const struct ieee80211_htrateset *rs)
3982 {
3983 	uint32_t rates;
3984 	int i;
3985 
3986 	rates = 0;
3987 	for (i = 0; i < rs->rs_nrates; i++) {
3988 		if (rs->rs_rates[i] < 16)
3989 			rates |= 1<<rs->rs_rates[i];
3990 	}
3991 	return rates;
3992 }
3993 
3994 /*
3995  * Craft station database entry for station.
3996  * NB: use host byte order here, the hal handles byte swapping.
3997  */
3998 static MWL_HAL_PEERINFO *
mkpeerinfo(MWL_HAL_PEERINFO * pi,const struct ieee80211_node * ni)3999 mkpeerinfo(MWL_HAL_PEERINFO *pi, const struct ieee80211_node *ni)
4000 {
4001 	const struct ieee80211vap *vap = ni->ni_vap;
4002 
4003 	memset(pi, 0, sizeof(*pi));
4004 	pi->LegacyRateBitMap = get_rate_bitmap(&ni->ni_rates);
4005 	pi->CapInfo = ni->ni_capinfo;
4006 	if (ni->ni_flags & IEEE80211_NODE_HT) {
4007 		/* HT capabilities, etc */
4008 		pi->HTCapabilitiesInfo = ni->ni_htcap;
4009 		/* XXX pi.HTCapabilitiesInfo */
4010 	        pi->MacHTParamInfo = ni->ni_htparam;
4011 		pi->HTRateBitMap = get_htrate_bitmap(&ni->ni_htrates);
4012 		pi->AddHtInfo.ControlChan = ni->ni_htctlchan;
4013 		pi->AddHtInfo.AddChan = ni->ni_ht2ndchan;
4014 		pi->AddHtInfo.OpMode = ni->ni_htopmode;
4015 		pi->AddHtInfo.stbc = ni->ni_htstbc;
4016 
4017 		/* constrain according to local configuration */
4018 		if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI40) == 0)
4019 			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI40;
4020 		if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI20) == 0)
4021 			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI20;
4022 		if (ni->ni_chw != 40)
4023 			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_CHWIDTH40;
4024 	}
4025 	return pi;
4026 }
4027 
4028 /*
4029  * Re-create the local sta db entry for a vap to ensure
4030  * up to date WME state is pushed to the firmware.  Because
4031  * this resets crypto state this must be followed by a
4032  * reload of any keys in the global key table.
4033  */
4034 static int
mwl_localstadb(struct ieee80211vap * vap)4035 mwl_localstadb(struct ieee80211vap *vap)
4036 {
4037 #define	WME(ie) ((const struct ieee80211_wme_info *) ie)
4038 	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
4039 	struct ieee80211_node *bss;
4040 	MWL_HAL_PEERINFO pi;
4041 	int error;
4042 
4043 	switch (vap->iv_opmode) {
4044 	case IEEE80211_M_STA:
4045 		bss = vap->iv_bss;
4046 		error = mwl_hal_newstation(hvap, vap->iv_myaddr, 0, 0,
4047 		    vap->iv_state == IEEE80211_S_RUN ?
4048 			mkpeerinfo(&pi, bss) : NULL,
4049 		    (bss->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT)),
4050 		    bss->ni_ies.wme_ie != NULL ?
4051 			WME(bss->ni_ies.wme_ie)->wme_info : 0);
4052 		if (error == 0)
4053 			mwl_setglobalkeys(vap);
4054 		break;
4055 	case IEEE80211_M_HOSTAP:
4056 	case IEEE80211_M_MBSS:
4057 		error = mwl_hal_newstation(hvap, vap->iv_myaddr,
4058 		    0, 0, NULL, vap->iv_flags & IEEE80211_F_WME, 0);
4059 		if (error == 0)
4060 			mwl_setglobalkeys(vap);
4061 		break;
4062 	default:
4063 		error = 0;
4064 		break;
4065 	}
4066 	return error;
4067 #undef WME
4068 }
4069 
4070 static int
mwl_newstate(struct ieee80211vap * vap,enum ieee80211_state nstate,int arg)4071 mwl_newstate(struct ieee80211vap *vap, enum ieee80211_state nstate, int arg)
4072 {
4073 	struct mwl_vap *mvp = MWL_VAP(vap);
4074 	struct mwl_hal_vap *hvap = mvp->mv_hvap;
4075 	struct ieee80211com *ic = vap->iv_ic;
4076 	struct ieee80211_node *ni = NULL;
4077 	struct mwl_softc *sc = ic->ic_softc;
4078 	struct mwl_hal *mh = sc->sc_mh;
4079 	enum ieee80211_state ostate = vap->iv_state;
4080 	int error;
4081 
4082 	DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: %s -> %s\n",
4083 	    if_name(vap->iv_ifp), __func__,
4084 	    ieee80211_state_name[ostate], ieee80211_state_name[nstate]);
4085 
4086 	callout_stop(&sc->sc_timer);
4087 	/*
4088 	 * Clear current radar detection state.
4089 	 */
4090 	if (ostate == IEEE80211_S_CAC) {
4091 		/* stop quiet mode radar detection */
4092 		mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_STOP);
4093 	} else if (sc->sc_radarena) {
4094 		/* stop in-service radar detection */
4095 		mwl_hal_setradardetection(mh, DR_DFS_DISABLE);
4096 		sc->sc_radarena = 0;
4097 	}
4098 	/*
4099 	 * Carry out per-state actions before doing net80211 work.
4100 	 */
4101 	if (nstate == IEEE80211_S_INIT) {
4102 		/* NB: only ap+sta vap's have a fw entity */
4103 		if (hvap != NULL)
4104 			mwl_hal_stop(hvap);
4105 	} else if (nstate == IEEE80211_S_SCAN) {
4106 		mwl_hal_start(hvap);
4107 		/* NB: this disables beacon frames */
4108 		mwl_hal_setinframode(hvap);
4109 	} else if (nstate == IEEE80211_S_AUTH) {
4110 		/*
4111 		 * Must create a sta db entry in case a WEP key needs to
4112 		 * be plumbed.  This entry will be overwritten if we
4113 		 * associate; otherwise it will be reclaimed on node free.
4114 		 */
4115 		ni = vap->iv_bss;
4116 		MWL_NODE(ni)->mn_hvap = hvap;
4117 		(void) mwl_peerstadb(ni, 0, 0, NULL);
4118 	} else if (nstate == IEEE80211_S_CSA) {
4119 		/* XXX move to below? */
4120 		if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
4121 		    vap->iv_opmode == IEEE80211_M_MBSS)
4122 			mwl_startcsa(vap);
4123 	} else if (nstate == IEEE80211_S_CAC) {
4124 		/* XXX move to below? */
4125 		/* stop ap xmit and enable quiet mode radar detection */
4126 		mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_START);
4127 	}
4128 
4129 	/*
4130 	 * Invoke the parent method to do net80211 work.
4131 	 */
4132 	error = mvp->mv_newstate(vap, nstate, arg);
4133 
4134 	/*
4135 	 * Carry out work that must be done after net80211 runs;
4136 	 * this work requires up to date state (e.g. iv_bss).
4137 	 */
4138 	if (error == 0 && nstate == IEEE80211_S_RUN) {
4139 		/* NB: collect bss node again, it may have changed */
4140 		ni = vap->iv_bss;
4141 
4142 		DPRINTF(sc, MWL_DEBUG_STATE,
4143 		    "%s: %s(RUN): iv_flags 0x%08x bintvl %d bssid %s "
4144 		    "capinfo 0x%04x chan %d\n",
4145 		    if_name(vap->iv_ifp), __func__, vap->iv_flags,
4146 		    ni->ni_intval, ether_sprintf(ni->ni_bssid), ni->ni_capinfo,
4147 		    ieee80211_chan2ieee(ic, ic->ic_curchan));
4148 
4149 		/*
4150 		 * Recreate local sta db entry to update WME/HT state.
4151 		 */
4152 		mwl_localstadb(vap);
4153 		switch (vap->iv_opmode) {
4154 		case IEEE80211_M_HOSTAP:
4155 		case IEEE80211_M_MBSS:
4156 			if (ostate == IEEE80211_S_CAC) {
4157 				/* enable in-service radar detection */
4158 				mwl_hal_setradardetection(mh,
4159 				    DR_IN_SERVICE_MONITOR_START);
4160 				sc->sc_radarena = 1;
4161 			}
4162 			/*
4163 			 * Allocate and setup the beacon frame
4164 			 * (and related state).
4165 			 */
4166 			error = mwl_reset_vap(vap, IEEE80211_S_RUN);
4167 			if (error != 0) {
4168 				DPRINTF(sc, MWL_DEBUG_STATE,
4169 				    "%s: beacon setup failed, error %d\n",
4170 				    __func__, error);
4171 				goto bad;
4172 			}
4173 			/* NB: must be after setting up beacon */
4174 			mwl_hal_start(hvap);
4175 			break;
4176 		case IEEE80211_M_STA:
4177 			DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: aid 0x%x\n",
4178 			    if_name(vap->iv_ifp), __func__, ni->ni_associd);
4179 			/*
4180 			 * Set state now that we're associated.
4181 			 */
4182 			mwl_hal_setassocid(hvap, ni->ni_bssid, ni->ni_associd);
4183 			mwl_setrates(vap);
4184 			mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
4185 			if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4186 			    sc->sc_ndwdsvaps++ == 0)
4187 				mwl_hal_setdwds(mh, 1);
4188 			break;
4189 		case IEEE80211_M_WDS:
4190 			DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: bssid %s\n",
4191 			    if_name(vap->iv_ifp), __func__,
4192 			    ether_sprintf(ni->ni_bssid));
4193 			mwl_seteapolformat(vap);
4194 			break;
4195 		default:
4196 			break;
4197 		}
4198 		/*
4199 		 * Set CS mode according to operating channel;
4200 		 * this mostly an optimization for 5GHz.
4201 		 *
4202 		 * NB: must follow mwl_hal_start which resets csmode
4203 		 */
4204 		if (IEEE80211_IS_CHAN_5GHZ(ic->ic_bsschan))
4205 			mwl_hal_setcsmode(mh, CSMODE_AGGRESSIVE);
4206 		else
4207 			mwl_hal_setcsmode(mh, CSMODE_AUTO_ENA);
4208 		/*
4209 		 * Start timer to prod firmware.
4210 		 */
4211 		if (sc->sc_ageinterval != 0)
4212 			callout_reset(&sc->sc_timer, sc->sc_ageinterval*hz,
4213 			    mwl_agestations, sc);
4214 	} else if (nstate == IEEE80211_S_SLEEP) {
4215 		/* XXX set chip in power save */
4216 	} else if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4217 	    --sc->sc_ndwdsvaps == 0)
4218 		mwl_hal_setdwds(mh, 0);
4219 bad:
4220 	return error;
4221 }
4222 
4223 /*
4224  * Manage station id's; these are separate from AID's
4225  * as AID's may have values out of the range of possible
4226  * station id's acceptable to the firmware.
4227  */
4228 static int
allocstaid(struct mwl_softc * sc,int aid)4229 allocstaid(struct mwl_softc *sc, int aid)
4230 {
4231 	int staid;
4232 
4233 	if (!(0 < aid && aid < MWL_MAXSTAID) || isset(sc->sc_staid, aid)) {
4234 		/* NB: don't use 0 */
4235 		for (staid = 1; staid < MWL_MAXSTAID; staid++)
4236 			if (isclr(sc->sc_staid, staid))
4237 				break;
4238 	} else
4239 		staid = aid;
4240 	setbit(sc->sc_staid, staid);
4241 	return staid;
4242 }
4243 
4244 static void
delstaid(struct mwl_softc * sc,int staid)4245 delstaid(struct mwl_softc *sc, int staid)
4246 {
4247 	clrbit(sc->sc_staid, staid);
4248 }
4249 
4250 /*
4251  * Setup driver-specific state for a newly associated node.
4252  * Note that we're called also on a re-associate, the isnew
4253  * param tells us if this is the first time or not.
4254  */
4255 static void
mwl_newassoc(struct ieee80211_node * ni,int isnew)4256 mwl_newassoc(struct ieee80211_node *ni, int isnew)
4257 {
4258 	struct ieee80211vap *vap = ni->ni_vap;
4259         struct mwl_softc *sc = vap->iv_ic->ic_softc;
4260 	struct mwl_node *mn = MWL_NODE(ni);
4261 	MWL_HAL_PEERINFO pi;
4262 	uint16_t aid;
4263 	int error;
4264 
4265 	aid = IEEE80211_AID(ni->ni_associd);
4266 	if (isnew) {
4267 		mn->mn_staid = allocstaid(sc, aid);
4268 		mn->mn_hvap = MWL_VAP(vap)->mv_hvap;
4269 	} else {
4270 		mn = MWL_NODE(ni);
4271 		/* XXX reset BA stream? */
4272 	}
4273 	DPRINTF(sc, MWL_DEBUG_NODE, "%s: mac %s isnew %d aid %d staid %d\n",
4274 	    __func__, ether_sprintf(ni->ni_macaddr), isnew, aid, mn->mn_staid);
4275 	error = mwl_peerstadb(ni, aid, mn->mn_staid, mkpeerinfo(&pi, ni));
4276 	if (error != 0) {
4277 		DPRINTF(sc, MWL_DEBUG_NODE,
4278 		    "%s: error %d creating sta db entry\n",
4279 		    __func__, error);
4280 		/* XXX how to deal with error? */
4281 	}
4282 }
4283 
4284 /*
4285  * Periodically poke the firmware to age out station state
4286  * (power save queues, pending tx aggregates).
4287  */
4288 static void
mwl_agestations(void * arg)4289 mwl_agestations(void *arg)
4290 {
4291 	struct mwl_softc *sc = arg;
4292 
4293 	mwl_hal_setkeepalive(sc->sc_mh);
4294 	if (sc->sc_ageinterval != 0)		/* NB: catch dynamic changes */
4295 		callout_schedule(&sc->sc_timer, sc->sc_ageinterval*hz);
4296 }
4297 
4298 static const struct mwl_hal_channel *
findhalchannel(const MWL_HAL_CHANNELINFO * ci,int ieee)4299 findhalchannel(const MWL_HAL_CHANNELINFO *ci, int ieee)
4300 {
4301 	int i;
4302 
4303 	for (i = 0; i < ci->nchannels; i++) {
4304 		const struct mwl_hal_channel *hc = &ci->channels[i];
4305 		if (hc->ieee == ieee)
4306 			return hc;
4307 	}
4308 	return NULL;
4309 }
4310 
4311 static int
mwl_setregdomain(struct ieee80211com * ic,struct ieee80211_regdomain * rd,int nchan,struct ieee80211_channel chans[])4312 mwl_setregdomain(struct ieee80211com *ic, struct ieee80211_regdomain *rd,
4313 	int nchan, struct ieee80211_channel chans[])
4314 {
4315 	struct mwl_softc *sc = ic->ic_softc;
4316 	struct mwl_hal *mh = sc->sc_mh;
4317 	const MWL_HAL_CHANNELINFO *ci;
4318 	int i;
4319 
4320 	for (i = 0; i < nchan; i++) {
4321 		struct ieee80211_channel *c = &chans[i];
4322 		const struct mwl_hal_channel *hc;
4323 
4324 		if (IEEE80211_IS_CHAN_2GHZ(c)) {
4325 			mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_2DOT4GHZ,
4326 			    IEEE80211_IS_CHAN_HT40(c) ?
4327 				MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4328 		} else if (IEEE80211_IS_CHAN_5GHZ(c)) {
4329 			mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_5GHZ,
4330 			    IEEE80211_IS_CHAN_HT40(c) ?
4331 				MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4332 		} else {
4333 			device_printf(sc->sc_dev,
4334 			    "%s: channel %u freq %u/0x%x not 2.4/5GHz\n",
4335 			    __func__, c->ic_ieee, c->ic_freq, c->ic_flags);
4336 			return EINVAL;
4337 		}
4338 		/*
4339 		 * Verify channel has cal data and cap tx power.
4340 		 */
4341 		hc = findhalchannel(ci, c->ic_ieee);
4342 		if (hc != NULL) {
4343 			if (c->ic_maxpower > 2*hc->maxTxPow)
4344 				c->ic_maxpower = 2*hc->maxTxPow;
4345 			goto next;
4346 		}
4347 		if (IEEE80211_IS_CHAN_HT40(c)) {
4348 			/*
4349 			 * Look for the extension channel since the
4350 			 * hal table only has the primary channel.
4351 			 */
4352 			hc = findhalchannel(ci, c->ic_extieee);
4353 			if (hc != NULL) {
4354 				if (c->ic_maxpower > 2*hc->maxTxPow)
4355 					c->ic_maxpower = 2*hc->maxTxPow;
4356 				goto next;
4357 			}
4358 		}
4359 		device_printf(sc->sc_dev,
4360 		    "%s: no cal data for channel %u ext %u freq %u/0x%x\n",
4361 		    __func__, c->ic_ieee, c->ic_extieee,
4362 		    c->ic_freq, c->ic_flags);
4363 		return EINVAL;
4364 	next:
4365 		;
4366 	}
4367 	return 0;
4368 }
4369 
4370 #define	IEEE80211_CHAN_HTG	(IEEE80211_CHAN_HT|IEEE80211_CHAN_G)
4371 #define	IEEE80211_CHAN_HTA	(IEEE80211_CHAN_HT|IEEE80211_CHAN_A)
4372 
4373 static void
addht40channels(struct ieee80211_channel chans[],int maxchans,int * nchans,const MWL_HAL_CHANNELINFO * ci,int flags)4374 addht40channels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4375 	const MWL_HAL_CHANNELINFO *ci, int flags)
4376 {
4377 	int i, error;
4378 
4379 	for (i = 0; i < ci->nchannels; i++) {
4380 		const struct mwl_hal_channel *hc = &ci->channels[i];
4381 
4382 		error = ieee80211_add_channel_ht40(chans, maxchans, nchans,
4383 		    hc->ieee, hc->maxTxPow, flags);
4384 		if (error != 0 && error != ENOENT)
4385 			break;
4386 	}
4387 }
4388 
4389 static void
addchannels(struct ieee80211_channel chans[],int maxchans,int * nchans,const MWL_HAL_CHANNELINFO * ci,const uint8_t bands[])4390 addchannels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4391 	const MWL_HAL_CHANNELINFO *ci, const uint8_t bands[])
4392 {
4393 	int i, error;
4394 
4395 	error = 0;
4396 	for (i = 0; i < ci->nchannels && error == 0; i++) {
4397 		const struct mwl_hal_channel *hc = &ci->channels[i];
4398 
4399 		error = ieee80211_add_channel(chans, maxchans, nchans,
4400 		    hc->ieee, hc->freq, hc->maxTxPow, 0, bands);
4401 	}
4402 }
4403 
4404 static void
getchannels(struct mwl_softc * sc,int maxchans,int * nchans,struct ieee80211_channel chans[])4405 getchannels(struct mwl_softc *sc, int maxchans, int *nchans,
4406 	struct ieee80211_channel chans[])
4407 {
4408 	const MWL_HAL_CHANNELINFO *ci;
4409 	uint8_t bands[IEEE80211_MODE_BYTES];
4410 
4411 	/*
4412 	 * Use the channel info from the hal to craft the
4413 	 * channel list.  Note that we pass back an unsorted
4414 	 * list; the caller is required to sort it for us
4415 	 * (if desired).
4416 	 */
4417 	*nchans = 0;
4418 	if (mwl_hal_getchannelinfo(sc->sc_mh,
4419 	    MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0) {
4420 		memset(bands, 0, sizeof(bands));
4421 		setbit(bands, IEEE80211_MODE_11B);
4422 		setbit(bands, IEEE80211_MODE_11G);
4423 		setbit(bands, IEEE80211_MODE_11NG);
4424 		addchannels(chans, maxchans, nchans, ci, bands);
4425 	}
4426 	if (mwl_hal_getchannelinfo(sc->sc_mh,
4427 	    MWL_FREQ_BAND_5GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0) {
4428 		memset(bands, 0, sizeof(bands));
4429 		setbit(bands, IEEE80211_MODE_11A);
4430 		setbit(bands, IEEE80211_MODE_11NA);
4431 		addchannels(chans, maxchans, nchans, ci, bands);
4432 	}
4433 	if (mwl_hal_getchannelinfo(sc->sc_mh,
4434 	    MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4435 		addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG);
4436 	if (mwl_hal_getchannelinfo(sc->sc_mh,
4437 	    MWL_FREQ_BAND_5GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4438 		addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA);
4439 }
4440 
4441 static void
mwl_getradiocaps(struct ieee80211com * ic,int maxchans,int * nchans,struct ieee80211_channel chans[])4442 mwl_getradiocaps(struct ieee80211com *ic,
4443 	int maxchans, int *nchans, struct ieee80211_channel chans[])
4444 {
4445 	struct mwl_softc *sc = ic->ic_softc;
4446 
4447 	getchannels(sc, maxchans, nchans, chans);
4448 }
4449 
4450 static int
mwl_getchannels(struct mwl_softc * sc)4451 mwl_getchannels(struct mwl_softc *sc)
4452 {
4453 	struct ieee80211com *ic = &sc->sc_ic;
4454 
4455 	/*
4456 	 * Use the channel info from the hal to craft the
4457 	 * channel list for net80211.  Note that we pass up
4458 	 * an unsorted list; net80211 will sort it for us.
4459 	 */
4460 	memset(ic->ic_channels, 0, sizeof(ic->ic_channels));
4461 	ic->ic_nchans = 0;
4462 	getchannels(sc, IEEE80211_CHAN_MAX, &ic->ic_nchans, ic->ic_channels);
4463 
4464 	ic->ic_regdomain.regdomain = SKU_DEBUG;
4465 	ic->ic_regdomain.country = CTRY_DEFAULT;
4466 	ic->ic_regdomain.location = 'I';
4467 	ic->ic_regdomain.isocc[0] = ' ';	/* XXX? */
4468 	ic->ic_regdomain.isocc[1] = ' ';
4469 	return (ic->ic_nchans == 0 ? EIO : 0);
4470 }
4471 #undef IEEE80211_CHAN_HTA
4472 #undef IEEE80211_CHAN_HTG
4473 
4474 #ifdef MWL_DEBUG
4475 static void
mwl_printrxbuf(const struct mwl_rxbuf * bf,u_int ix)4476 mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix)
4477 {
4478 	const struct mwl_rxdesc *ds = bf->bf_desc;
4479 	uint32_t status = le32toh(ds->Status);
4480 
4481 	printf("R[%2u] (DS.V:%p DS.P:0x%jx) NEXT:%08x DATA:%08x RC:%02x%s\n"
4482 	       "      STAT:%02x LEN:%04x RSSI:%02x CHAN:%02x RATE:%02x QOS:%04x HT:%04x\n",
4483 	    ix, ds, (uintmax_t)bf->bf_daddr, le32toh(ds->pPhysNext),
4484 	    le32toh(ds->pPhysBuffData), ds->RxControl,
4485 	    ds->RxControl != EAGLE_RXD_CTRL_DRIVER_OWN ?
4486 	        "" : (status & EAGLE_RXD_STATUS_OK) ? " *" : " !",
4487 	    ds->Status, le16toh(ds->PktLen), ds->RSSI, ds->Channel,
4488 	    ds->Rate, le16toh(ds->QosCtrl), le16toh(ds->HtSig2));
4489 }
4490 
4491 static void
mwl_printtxbuf(const struct mwl_txbuf * bf,u_int qnum,u_int ix)4492 mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix)
4493 {
4494 	const struct mwl_txdesc *ds = bf->bf_desc;
4495 	uint32_t status = le32toh(ds->Status);
4496 
4497 	printf("Q%u[%3u]", qnum, ix);
4498 	printf(" (DS.V:%p DS.P:0x%jx)\n", ds, (uintmax_t)bf->bf_daddr);
4499 	printf("    NEXT:%08x DATA:%08x LEN:%04x STAT:%08x%s\n",
4500 	    le32toh(ds->pPhysNext),
4501 	    le32toh(ds->PktPtr), le16toh(ds->PktLen), status,
4502 	    status & EAGLE_TXD_STATUS_USED ?
4503 		"" : (status & 3) != 0 ? " *" : " !");
4504 	printf("    RATE:%02x PRI:%x QOS:%04x SAP:%08x FORMAT:%04x\n",
4505 	    ds->DataRate, ds->TxPriority, le16toh(ds->QosCtrl),
4506 	    le32toh(ds->SapPktInfo), le16toh(ds->Format));
4507 #if MWL_TXDESC > 1
4508 	printf("    MULTIFRAMES:%u LEN:%04x %04x %04x %04x %04x %04x\n"
4509 	    , le32toh(ds->multiframes)
4510 	    , le16toh(ds->PktLenArray[0]), le16toh(ds->PktLenArray[1])
4511 	    , le16toh(ds->PktLenArray[2]), le16toh(ds->PktLenArray[3])
4512 	    , le16toh(ds->PktLenArray[4]), le16toh(ds->PktLenArray[5])
4513 	);
4514 	printf("    DATA:%08x %08x %08x %08x %08x %08x\n"
4515 	    , le32toh(ds->PktPtrArray[0]), le32toh(ds->PktPtrArray[1])
4516 	    , le32toh(ds->PktPtrArray[2]), le32toh(ds->PktPtrArray[3])
4517 	    , le32toh(ds->PktPtrArray[4]), le32toh(ds->PktPtrArray[5])
4518 	);
4519 #endif
4520 #if 0
4521 { const uint8_t *cp = (const uint8_t *) ds;
4522   int i;
4523   for (i = 0; i < sizeof(struct mwl_txdesc); i++) {
4524 	printf("%02x ", cp[i]);
4525 	if (((i+1) % 16) == 0)
4526 		printf("\n");
4527   }
4528   printf("\n");
4529 }
4530 #endif
4531 }
4532 #endif /* MWL_DEBUG */
4533 
4534 #if 0
4535 static void
4536 mwl_txq_dump(struct mwl_txq *txq)
4537 {
4538 	struct mwl_txbuf *bf;
4539 	int i = 0;
4540 
4541 	MWL_TXQ_LOCK(txq);
4542 	STAILQ_FOREACH(bf, &txq->active, bf_list) {
4543 		struct mwl_txdesc *ds = bf->bf_desc;
4544 		MWL_TXDESC_SYNC(txq, ds,
4545 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
4546 #ifdef MWL_DEBUG
4547 		mwl_printtxbuf(bf, txq->qnum, i);
4548 #endif
4549 		i++;
4550 	}
4551 	MWL_TXQ_UNLOCK(txq);
4552 }
4553 #endif
4554 
4555 static void
mwl_watchdog(void * arg)4556 mwl_watchdog(void *arg)
4557 {
4558 	struct mwl_softc *sc = arg;
4559 
4560 	callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
4561 	if (sc->sc_tx_timer == 0 || --sc->sc_tx_timer > 0)
4562 		return;
4563 
4564 	if (sc->sc_running && !sc->sc_invalid) {
4565 		if (mwl_hal_setkeepalive(sc->sc_mh))
4566 			device_printf(sc->sc_dev,
4567 			    "transmit timeout (firmware hung?)\n");
4568 		else
4569 			device_printf(sc->sc_dev,
4570 			    "transmit timeout\n");
4571 #if 0
4572 		mwl_reset(sc);
4573 mwl_txq_dump(&sc->sc_txq[0]);/*XXX*/
4574 #endif
4575 		counter_u64_add(sc->sc_ic.ic_oerrors, 1);
4576 		sc->sc_stats.mst_watchdog++;
4577 	}
4578 }
4579 
4580 #ifdef MWL_DIAGAPI
4581 /*
4582  * Diagnostic interface to the HAL.  This is used by various
4583  * tools to do things like retrieve register contents for
4584  * debugging.  The mechanism is intentionally opaque so that
4585  * it can change frequently w/o concern for compatibility.
4586  */
4587 static int
mwl_ioctl_diag(struct mwl_softc * sc,struct mwl_diag * md)4588 mwl_ioctl_diag(struct mwl_softc *sc, struct mwl_diag *md)
4589 {
4590 	struct mwl_hal *mh = sc->sc_mh;
4591 	u_int id = md->md_id & MWL_DIAG_ID;
4592 	void *indata = NULL;
4593 	void *outdata = NULL;
4594 	u_int32_t insize = md->md_in_size;
4595 	u_int32_t outsize = md->md_out_size;
4596 	int error = 0;
4597 
4598 	if (md->md_id & MWL_DIAG_IN) {
4599 		/*
4600 		 * Copy in data.
4601 		 */
4602 		indata = malloc(insize, M_TEMP, M_NOWAIT);
4603 		if (indata == NULL) {
4604 			error = ENOMEM;
4605 			goto bad;
4606 		}
4607 		error = copyin(md->md_in_data, indata, insize);
4608 		if (error)
4609 			goto bad;
4610 	}
4611 	if (md->md_id & MWL_DIAG_DYN) {
4612 		/*
4613 		 * Allocate a buffer for the results (otherwise the HAL
4614 		 * returns a pointer to a buffer where we can read the
4615 		 * results).  Note that we depend on the HAL leaving this
4616 		 * pointer for us to use below in reclaiming the buffer;
4617 		 * may want to be more defensive.
4618 		 */
4619 		outdata = malloc(outsize, M_TEMP, M_NOWAIT);
4620 		if (outdata == NULL) {
4621 			error = ENOMEM;
4622 			goto bad;
4623 		}
4624 	}
4625 	if (mwl_hal_getdiagstate(mh, id, indata, insize, &outdata, &outsize)) {
4626 		if (outsize < md->md_out_size)
4627 			md->md_out_size = outsize;
4628 		if (outdata != NULL)
4629 			error = copyout(outdata, md->md_out_data,
4630 					md->md_out_size);
4631 	} else {
4632 		error = EINVAL;
4633 	}
4634 bad:
4635 	if ((md->md_id & MWL_DIAG_IN) && indata != NULL)
4636 		free(indata, M_TEMP);
4637 	if ((md->md_id & MWL_DIAG_DYN) && outdata != NULL)
4638 		free(outdata, M_TEMP);
4639 	return error;
4640 }
4641 
4642 static int
mwl_ioctl_reset(struct mwl_softc * sc,struct mwl_diag * md)4643 mwl_ioctl_reset(struct mwl_softc *sc, struct mwl_diag *md)
4644 {
4645 	struct mwl_hal *mh = sc->sc_mh;
4646 	int error;
4647 
4648 	MWL_LOCK_ASSERT(sc);
4649 
4650 	if (md->md_id == 0 && mwl_hal_fwload(mh, NULL) != 0) {
4651 		device_printf(sc->sc_dev, "unable to load firmware\n");
4652 		return EIO;
4653 	}
4654 	if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
4655 		device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
4656 		return EIO;
4657 	}
4658 	error = mwl_setupdma(sc);
4659 	if (error != 0) {
4660 		/* NB: mwl_setupdma prints a msg */
4661 		return error;
4662 	}
4663 	/*
4664 	 * Reset tx/rx data structures; after reload we must
4665 	 * re-start the driver's notion of the next xmit/recv.
4666 	 */
4667 	mwl_draintxq(sc);		/* clear pending frames */
4668 	mwl_resettxq(sc);		/* rebuild tx q lists */
4669 	sc->sc_rxnext = NULL;		/* force rx to start at the list head */
4670 	return 0;
4671 }
4672 #endif /* MWL_DIAGAPI */
4673 
4674 static void
mwl_parent(struct ieee80211com * ic)4675 mwl_parent(struct ieee80211com *ic)
4676 {
4677 	struct mwl_softc *sc = ic->ic_softc;
4678 	int startall = 0;
4679 
4680 	MWL_LOCK(sc);
4681 	if (ic->ic_nrunning > 0) {
4682 		if (sc->sc_running) {
4683 			/*
4684 			 * To avoid rescanning another access point,
4685 			 * do not call mwl_init() here.  Instead,
4686 			 * only reflect promisc mode settings.
4687 			 */
4688 			mwl_mode_init(sc);
4689 		} else {
4690 			/*
4691 			 * Beware of being called during attach/detach
4692 			 * to reset promiscuous mode.  In that case we
4693 			 * will still be marked UP but not RUNNING.
4694 			 * However trying to re-init the interface
4695 			 * is the wrong thing to do as we've already
4696 			 * torn down much of our state.  There's
4697 			 * probably a better way to deal with this.
4698 			 */
4699 			if (!sc->sc_invalid) {
4700 				mwl_init(sc);	/* XXX lose error */
4701 				startall = 1;
4702 			}
4703 		}
4704 	} else
4705 		mwl_stop(sc);
4706 	MWL_UNLOCK(sc);
4707 	if (startall)
4708 		ieee80211_start_all(ic);
4709 }
4710 
4711 static int
mwl_ioctl(struct ieee80211com * ic,u_long cmd,void * data)4712 mwl_ioctl(struct ieee80211com *ic, u_long cmd, void *data)
4713 {
4714 	struct mwl_softc *sc = ic->ic_softc;
4715 	struct ifreq *ifr = data;
4716 	int error = 0;
4717 
4718 	switch (cmd) {
4719 	case SIOCGMVSTATS:
4720 		mwl_hal_gethwstats(sc->sc_mh, &sc->sc_stats.hw_stats);
4721 #if 0
4722 		/* NB: embed these numbers to get a consistent view */
4723 		sc->sc_stats.mst_tx_packets =
4724 		    if_get_counter(ifp, IFCOUNTER_OPACKETS);
4725 		sc->sc_stats.mst_rx_packets =
4726 		    if_get_counter(ifp, IFCOUNTER_IPACKETS);
4727 #endif
4728 		/*
4729 		 * NB: Drop the softc lock in case of a page fault;
4730 		 * we'll accept any potential inconsisentcy in the
4731 		 * statistics.  The alternative is to copy the data
4732 		 * to a local structure.
4733 		 */
4734 		return (copyout(&sc->sc_stats, ifr_data_get_ptr(ifr),
4735 		    sizeof (sc->sc_stats)));
4736 #ifdef MWL_DIAGAPI
4737 	case SIOCGMVDIAG:
4738 		/* XXX check privs */
4739 		return mwl_ioctl_diag(sc, (struct mwl_diag *) ifr);
4740 	case SIOCGMVRESET:
4741 		/* XXX check privs */
4742 		MWL_LOCK(sc);
4743 		error = mwl_ioctl_reset(sc,(struct mwl_diag *) ifr);
4744 		MWL_UNLOCK(sc);
4745 		break;
4746 #endif /* MWL_DIAGAPI */
4747 	default:
4748 		error = ENOTTY;
4749 		break;
4750 	}
4751 	return (error);
4752 }
4753 
4754 #ifdef	MWL_DEBUG
4755 static int
mwl_sysctl_debug(SYSCTL_HANDLER_ARGS)4756 mwl_sysctl_debug(SYSCTL_HANDLER_ARGS)
4757 {
4758 	struct mwl_softc *sc = arg1;
4759 	int debug, error;
4760 
4761 	debug = sc->sc_debug | (mwl_hal_getdebug(sc->sc_mh) << 24);
4762 	error = sysctl_handle_int(oidp, &debug, 0, req);
4763 	if (error || !req->newptr)
4764 		return error;
4765 	mwl_hal_setdebug(sc->sc_mh, debug >> 24);
4766 	sc->sc_debug = debug & 0x00ffffff;
4767 	return 0;
4768 }
4769 #endif /* MWL_DEBUG */
4770 
4771 static void
mwl_sysctlattach(struct mwl_softc * sc)4772 mwl_sysctlattach(struct mwl_softc *sc)
4773 {
4774 #ifdef	MWL_DEBUG
4775 	struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(sc->sc_dev);
4776 	struct sysctl_oid *tree = device_get_sysctl_tree(sc->sc_dev);
4777 
4778 	sc->sc_debug = mwl_debug;
4779 	SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(tree), OID_AUTO, "debug",
4780 	    CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_NEEDGIANT, sc, 0,
4781 	    mwl_sysctl_debug, "I", "control debugging printfs");
4782 #endif
4783 }
4784 
4785 /*
4786  * Announce various information on device/driver attach.
4787  */
4788 static void
mwl_announce(struct mwl_softc * sc)4789 mwl_announce(struct mwl_softc *sc)
4790 {
4791 
4792 	device_printf(sc->sc_dev, "Rev A%d hardware, v%d.%d.%d.%d firmware (regioncode %d)\n",
4793 		sc->sc_hwspecs.hwVersion,
4794 		(sc->sc_hwspecs.fwReleaseNumber>>24) & 0xff,
4795 		(sc->sc_hwspecs.fwReleaseNumber>>16) & 0xff,
4796 		(sc->sc_hwspecs.fwReleaseNumber>>8) & 0xff,
4797 		(sc->sc_hwspecs.fwReleaseNumber>>0) & 0xff,
4798 		sc->sc_hwspecs.regionCode);
4799 	sc->sc_fwrelease = sc->sc_hwspecs.fwReleaseNumber;
4800 
4801 	if (bootverbose) {
4802 		int i;
4803 		for (i = 0; i <= WME_AC_VO; i++) {
4804 			struct mwl_txq *txq = sc->sc_ac2q[i];
4805 			device_printf(sc->sc_dev, "Use hw queue %u for %s traffic\n",
4806 				txq->qnum, ieee80211_wme_acnames[i]);
4807 		}
4808 	}
4809 	if (bootverbose || mwl_rxdesc != MWL_RXDESC)
4810 		device_printf(sc->sc_dev, "using %u rx descriptors\n", mwl_rxdesc);
4811 	if (bootverbose || mwl_rxbuf != MWL_RXBUF)
4812 		device_printf(sc->sc_dev, "using %u rx buffers\n", mwl_rxbuf);
4813 	if (bootverbose || mwl_txbuf != MWL_TXBUF)
4814 		device_printf(sc->sc_dev, "using %u tx buffers\n", mwl_txbuf);
4815 	if (bootverbose && mwl_hal_ismbsscapable(sc->sc_mh))
4816 		device_printf(sc->sc_dev, "multi-bss support\n");
4817 #ifdef MWL_TX_NODROP
4818 	if (bootverbose)
4819 		device_printf(sc->sc_dev, "no tx drop\n");
4820 #endif
4821 }
4822