xref: /netbsd/sys/dev/ic/mtd803.c (revision 03da633e)
1 /* $NetBSD: mtd803.c,v 1.41 2020/01/29 15:00:39 thorpej Exp $ */
2 
3 /*-
4  *
5  * Copyright (c) 2002 The NetBSD Foundation, Inc.
6  * All rights reserved.
7  *
8  * This code is derived from software contributed to The NetBSD Foundation
9  * by Peter Bex <Peter.Bex@student.kun.nl>.
10  *
11  * Redistribution and use in source and binary forms, with or without
12  * modification, are permitted provided that the following conditions
13  * are met:
14  * 1. Redistributions of source code must retain the above copyright
15  *    notice, this list of conditions and the following disclaimer.
16  * 2. Redistributions in binary form must reproduce the above copyright
17  *    notice, this list of conditions and the following disclaimer in the
18  *    documentation and/or other materials provided with the distribution.
19  *
20  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
21  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
22  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
23  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
24  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30  * POSSIBILITY OF SUCH DAMAGE.
31  */
32 
33 /*
34  * TODO:
35  * - Most importantly, get some bus_dmamap_syncs in the correct places.
36  *    I don't have access to a computer with PCI other than i386, and i386
37  *    is just such a machine where dmamap_syncs don't do anything.
38  * - Powerhook for when resuming after standby.
39  * - Watchdog stuff doesn't work yet, the system crashes.
40  * - There seems to be a CardBus version of the card. (see datasheet)
41  *    Perhaps a detach function is necessary then? (free buffs, stop rx/tx etc)
42  * - When you enable the TXBUN (Tx buffer unavailable) interrupt, it gets
43  *    raised every time a packet is sent. Strange, since everything works anyway
44  */
45 
46 #include <sys/cdefs.h>
47 __KERNEL_RCSID(0, "$NetBSD: mtd803.c,v 1.41 2020/01/29 15:00:39 thorpej Exp $");
48 
49 
50 #include <sys/param.h>
51 #include <sys/mbuf.h>
52 #include <sys/systm.h>
53 #include <sys/device.h>
54 #include <sys/socket.h>
55 #include <sys/ioctl.h>
56 #include <sys/syslog.h>
57 
58 #include <net/if.h>
59 #include <net/if_ether.h>
60 #include <net/if_media.h>
61 #include <net/bpf.h>
62 
63 #ifdef INET
64 #include <netinet/in.h>
65 #include <netinet/if_inarp.h>
66 #include <netinet/in_systm.h>
67 #include <netinet/in_var.h>
68 #include <netinet/ip.h>
69 #endif
70 
71 #include <sys/bus.h>
72 
73 #include <dev/ic/mtd803reg.h>
74 #include <dev/ic/mtd803var.h>
75 #include <dev/mii/mii.h>
76 #include <dev/mii/miivar.h>
77 
78 /*
79  * Device driver for the MTD803 3-in-1 Fast Ethernet Controller
80  * Written by Peter Bex (peter.bex@student.kun.nl)
81  *
82  * Datasheet at:   http://www.myson.com.tw   or   http://www.century-semi.com
83  */
84 
85 #define MTD_READ_1(sc, reg) \
86 	bus_space_read_1((sc)->bus_tag, (sc)->bus_handle, (reg))
87 #define MTD_WRITE_1(sc, reg, data) \
88 	bus_space_write_1((sc)->bus_tag, (sc)->bus_handle, (reg), (data))
89 
90 #define MTD_READ_2(sc, reg) \
91 	bus_space_read_2((sc)->bus_tag, (sc)->bus_handle, (reg))
92 #define MTD_WRITE_2(sc, reg, data) \
93 	bus_space_write_2((sc)->bus_tag, (sc)->bus_handle, (reg), (data))
94 
95 #define MTD_READ_4(sc, reg) \
96 	bus_space_read_4((sc)->bus_tag, (sc)->bus_handle, (reg))
97 #define MTD_WRITE_4(sc, reg, data) \
98 	bus_space_write_4((sc)->bus_tag, (sc)->bus_handle, (reg), (data))
99 
100 #define MTD_SETBIT(sc, reg, x) \
101 	MTD_WRITE_4((sc), (reg), MTD_READ_4((sc), (reg)) | (x))
102 #define MTD_CLRBIT(sc, reg, x) \
103 	MTD_WRITE_4((sc), (reg), MTD_READ_4((sc), (reg)) & ~(x))
104 
105 #define ETHER_CRC32(buf, len)	(ether_crc32_be((buf), (len)))
106 
107 int mtd_mii_readreg(device_t, int, int, uint16_t *);
108 int mtd_mii_writereg(device_t, int, int, uint16_t);
109 void mtd_mii_statchg(struct ifnet *);
110 
111 void mtd_start(struct ifnet *);
112 void mtd_stop(struct ifnet *, int);
113 int mtd_ioctl(struct ifnet *, u_long, void *);
114 void mtd_setmulti(struct mtd_softc *);
115 void mtd_watchdog(struct ifnet *);
116 
117 int mtd_init(struct ifnet *);
118 void mtd_reset(struct mtd_softc *);
119 void mtd_shutdown(void *);
120 int mtd_init_desc(struct mtd_softc *);
121 int mtd_put(struct mtd_softc *, int, struct mbuf *);
122 struct mbuf *mtd_get(struct mtd_softc *, int, int);
123 
124 int mtd_rxirq(struct mtd_softc *);
125 int mtd_txirq(struct mtd_softc *);
126 int mtd_bufirq(struct mtd_softc *);
127 
128 
129 int
mtd_config(struct mtd_softc * sc)130 mtd_config(struct mtd_softc *sc)
131 {
132 	struct ifnet *ifp = &sc->ethercom.ec_if;
133 	struct mii_data *mii = &sc->mii;
134 	int i;
135 
136 	/* Read station address */
137 	for (i = 0; i < ETHER_ADDR_LEN; ++i)
138 		sc->eaddr[i] = MTD_READ_1(sc, MTD_PAR0 + i);
139 
140 	/* Initialize ifnet structure */
141 	memcpy(ifp->if_xname, device_xname(sc->dev), IFNAMSIZ);
142 	ifp->if_softc = sc;
143 	ifp->if_init = mtd_init;
144 	ifp->if_start = mtd_start;
145 	ifp->if_stop = mtd_stop;
146 	ifp->if_ioctl = mtd_ioctl;
147 	ifp->if_watchdog = mtd_watchdog;
148 	ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
149 	IFQ_SET_READY(&ifp->if_snd);
150 
151 	/* Setup MII interface */
152 	mii->mii_ifp = ifp;
153 	mii->mii_readreg = mtd_mii_readreg;
154 	mii->mii_writereg = mtd_mii_writereg;
155 	mii->mii_statchg = mtd_mii_statchg;
156 
157 	sc->ethercom.ec_mii = mii;
158 	ifmedia_init(&mii->mii_media, 0, ether_mediachange, ether_mediastatus);
159 
160 	mii_attach(sc->dev, mii, 0xffffffff, MII_PHY_ANY, 0, 0);
161 
162 	if (LIST_FIRST(&mii->mii_phys) == NULL) {
163 		aprint_error_dev(sc->dev, "Unable to configure MII\n");
164 		return 1;
165 	} else
166 		ifmedia_set(&mii->mii_media, IFM_ETHER | IFM_AUTO);
167 
168 	if (mtd_init_desc(sc))
169 		return 1;
170 
171 	/* Attach interface */
172 	if_attach(ifp);
173 	ether_ifattach(ifp, sc->eaddr);
174 
175 	/* Initialise random source */
176 	rnd_attach_source(&sc->rnd_src, device_xname(sc->dev),
177 			  RND_TYPE_NET, RND_FLAG_DEFAULT);
178 
179 	/* Add shutdown hook to reset card when we reboot */
180 	sc->sd_hook = shutdownhook_establish(mtd_shutdown, sc);
181 
182 	return 0;
183 }
184 
185 
186 /*
187  * mtd_init
188  * Must be called at splnet()
189  */
190 int
mtd_init(struct ifnet * ifp)191 mtd_init(struct ifnet *ifp)
192 {
193 	struct mtd_softc *sc = ifp->if_softc;
194 
195 	mtd_reset(sc);
196 
197 	/*
198 	 * Set cache alignment and burst length. Don't really know what these
199 	 * mean, so their values are probably suboptimal.
200 	 */
201 	MTD_WRITE_4(sc, MTD_BCR, MTD_BCR_BLEN16);
202 
203 	MTD_WRITE_4(sc, MTD_RXTXR, MTD_TX_STFWD | MTD_TX_FDPLX);
204 
205 	/* Promiscuous mode? */
206 	if (ifp->if_flags & IFF_PROMISC)
207 		MTD_SETBIT(sc, MTD_RXTXR, MTD_RX_PROM);
208 	else
209 		MTD_CLRBIT(sc, MTD_RXTXR, MTD_RX_PROM);
210 
211 	/* Broadcast mode? */
212 	if (ifp->if_flags & IFF_BROADCAST)
213 		MTD_SETBIT(sc, MTD_RXTXR, MTD_RX_ABROAD);
214 	else
215 		MTD_CLRBIT(sc, MTD_RXTXR, MTD_RX_ABROAD);
216 
217 	mtd_setmulti(sc);
218 
219 	/* Enable interrupts */
220 	MTD_WRITE_4(sc, MTD_IMR, MTD_IMR_MASK);
221 	MTD_WRITE_4(sc, MTD_ISR, MTD_ISR_ENABLE);
222 
223 	/* Set descriptor base addresses */
224 	MTD_WRITE_4(sc, MTD_TXLBA, htole32(sc->desc_dma_map->dm_segs[0].ds_addr
225 				+ sizeof(struct mtd_desc) * MTD_NUM_RXD));
226 	MTD_WRITE_4(sc, MTD_RXLBA,
227 		htole32(sc->desc_dma_map->dm_segs[0].ds_addr));
228 
229 	/* Enable receiver and transmitter */
230 	MTD_SETBIT(sc, MTD_RXTXR, MTD_RX_ENABLE);
231 	MTD_SETBIT(sc, MTD_RXTXR, MTD_TX_ENABLE);
232 
233 	/* Interface is running */
234 	ifp->if_flags |= IFF_RUNNING;
235 	ifp->if_flags &= ~IFF_OACTIVE;
236 
237 	return 0;
238 }
239 
240 
241 int
mtd_init_desc(struct mtd_softc * sc)242 mtd_init_desc(struct mtd_softc *sc)
243 {
244 	int rseg, err, i;
245 	bus_dma_segment_t seg;
246 	bus_size_t size;
247 
248 	/* Allocate memory for descriptors */
249 	size = (MTD_NUM_RXD + MTD_NUM_TXD) * sizeof(struct mtd_desc);
250 
251 	/* Allocate DMA-safe memory */
252 	if ((err = bus_dmamem_alloc(sc->dma_tag, size, MTD_DMA_ALIGN,
253 			 0, &seg, 1, &rseg, BUS_DMA_NOWAIT)) != 0) {
254 		aprint_error_dev(sc->dev,
255 		    "unable to allocate DMA buffer, error = %d\n", err);
256 		return 1;
257 	}
258 
259 	/* Map memory to kernel addressable space */
260 	if ((err = bus_dmamem_map(sc->dma_tag, &seg, 1, size,
261 		(void **)&sc->desc, BUS_DMA_NOWAIT | BUS_DMA_COHERENT)) != 0) {
262 		aprint_error_dev(sc->dev,
263 		    "unable to map DMA buffer, error = %d\n", err);
264 		bus_dmamem_free(sc->dma_tag, &seg, rseg);
265 		return 1;
266 	}
267 
268 	/* Create a DMA map */
269 	if ((err = bus_dmamap_create(sc->dma_tag, size, 1,
270 		size, 0, BUS_DMA_NOWAIT, &sc->desc_dma_map)) != 0) {
271 		aprint_error_dev(sc->dev,
272 		    "unable to create DMA map, error = %d\n", err);
273 		bus_dmamem_unmap(sc->dma_tag, (void *)sc->desc, size);
274 		bus_dmamem_free(sc->dma_tag, &seg, rseg);
275 		return 1;
276 	}
277 
278 	/* Load the DMA map */
279 	if ((err = bus_dmamap_load(sc->dma_tag, sc->desc_dma_map, sc->desc,
280 		size, NULL, BUS_DMA_NOWAIT)) != 0) {
281 		aprint_error_dev(sc->dev,
282 		    "unable to load DMA map, error = %d\n", err);
283 		bus_dmamap_destroy(sc->dma_tag, sc->desc_dma_map);
284 		bus_dmamem_unmap(sc->dma_tag, (void *)sc->desc, size);
285 		bus_dmamem_free(sc->dma_tag, &seg, rseg);
286 		return 1;
287 	}
288 
289 	/* Allocate memory for the buffers */
290 	size = MTD_NUM_RXD * MTD_RXBUF_SIZE + MTD_NUM_TXD * MTD_TXBUF_SIZE;
291 
292 	/* Allocate DMA-safe memory */
293 	if ((err = bus_dmamem_alloc(sc->dma_tag, size, MTD_DMA_ALIGN,
294 			 0, &seg, 1, &rseg, BUS_DMA_NOWAIT)) != 0) {
295 		aprint_error_dev(sc->dev,
296 		    "unable to allocate DMA buffer, error = %d\n", err);
297 
298 		/* Undo DMA map for descriptors */
299 		bus_dmamap_unload(sc->dma_tag, sc->desc_dma_map);
300 		bus_dmamap_destroy(sc->dma_tag, sc->desc_dma_map);
301 		bus_dmamem_unmap(sc->dma_tag, (void *)sc->desc, size);
302 		bus_dmamem_free(sc->dma_tag, &seg, rseg);
303 		return 1;
304 	}
305 
306 	/* Map memory to kernel addressable space */
307 	if ((err = bus_dmamem_map(sc->dma_tag, &seg, 1, size,
308 		&sc->buf, BUS_DMA_NOWAIT | BUS_DMA_COHERENT)) != 0) {
309 		aprint_error_dev(sc->dev,
310 		    "unable to map DMA buffer, error = %d\n", err);
311 		bus_dmamem_free(sc->dma_tag, &seg, rseg);
312 
313 		/* Undo DMA map for descriptors */
314 		bus_dmamap_unload(sc->dma_tag, sc->desc_dma_map);
315 		bus_dmamap_destroy(sc->dma_tag, sc->desc_dma_map);
316 		bus_dmamem_unmap(sc->dma_tag, (void *)sc->desc, size);
317 		bus_dmamem_free(sc->dma_tag, &seg, rseg);
318 		return 1;
319 	}
320 
321 	/* Create a DMA map */
322 	if ((err = bus_dmamap_create(sc->dma_tag, size, 1,
323 		size, 0, BUS_DMA_NOWAIT, &sc->buf_dma_map)) != 0) {
324 		aprint_error_dev(sc->dev,
325 		    "unable to create DMA map, error = %d\n", err);
326 		bus_dmamem_unmap(sc->dma_tag, sc->buf, size);
327 		bus_dmamem_free(sc->dma_tag, &seg, rseg);
328 
329 		/* Undo DMA map for descriptors */
330 		bus_dmamap_unload(sc->dma_tag, sc->desc_dma_map);
331 		bus_dmamap_destroy(sc->dma_tag, sc->desc_dma_map);
332 		bus_dmamem_unmap(sc->dma_tag, (void *)sc->desc, size);
333 		bus_dmamem_free(sc->dma_tag, &seg, rseg);
334 		return 1;
335 	}
336 
337 	/* Load the DMA map */
338 	if ((err = bus_dmamap_load(sc->dma_tag, sc->buf_dma_map, sc->buf,
339 		size, NULL, BUS_DMA_NOWAIT)) != 0) {
340 		aprint_error_dev(sc->dev,
341 		    "unable to load DMA map, error = %d\n", err);
342 		bus_dmamap_destroy(sc->dma_tag, sc->buf_dma_map);
343 		bus_dmamem_unmap(sc->dma_tag, sc->buf, size);
344 		bus_dmamem_free(sc->dma_tag, &seg, rseg);
345 
346 		/* Undo DMA map for descriptors */
347 		bus_dmamap_unload(sc->dma_tag, sc->desc_dma_map);
348 		bus_dmamap_destroy(sc->dma_tag, sc->desc_dma_map);
349 		bus_dmamem_unmap(sc->dma_tag, (void *)sc->desc, size);
350 		bus_dmamem_free(sc->dma_tag, &seg, rseg);
351 		return 1;
352 	}
353 
354 	/* Descriptors are stored as a circular linked list */
355 	/* Fill in rx descriptors */
356 	for (i = 0; i < MTD_NUM_RXD; ++i) {
357 		sc->desc[i].stat = MTD_RXD_OWNER;
358 		if (i == MTD_NUM_RXD - 1) {	/* Last descriptor */
359 			/* Link back to first rx descriptor */
360 			sc->desc[i].next =
361 				htole32(sc->desc_dma_map->dm_segs[0].ds_addr);
362 		} else {
363 			/* Link forward to next rx descriptor */
364 			sc->desc[i].next =
365 			htole32(sc->desc_dma_map->dm_segs[0].ds_addr
366 					+ (i + 1) * sizeof(struct mtd_desc));
367 		}
368 		sc->desc[i].conf = MTD_RXBUF_SIZE & MTD_RXD_CONF_BUFS;
369 		/* Set buffer's address */
370 		sc->desc[i].data = htole32(sc->buf_dma_map->dm_segs[0].ds_addr
371 					+ i * MTD_RXBUF_SIZE);
372 	}
373 
374 	/* Fill in tx descriptors */
375 	for (/* i = MTD_NUM_RXD */; i < (MTD_NUM_TXD + MTD_NUM_RXD); ++i) {
376 		sc->desc[i].stat = 0;	/* At least, NOT MTD_TXD_OWNER! */
377 		if (i == (MTD_NUM_RXD + MTD_NUM_TXD - 1)) { /* Last descr */
378 			/* Link back to first tx descriptor */
379 			sc->desc[i].next =
380 				htole32(sc->desc_dma_map->dm_segs[0].ds_addr
381 				    +MTD_NUM_RXD * sizeof(struct mtd_desc));
382 		} else {
383 			/* Link forward to next tx descriptor */
384 			sc->desc[i].next =
385 				htole32(sc->desc_dma_map->dm_segs[0].ds_addr
386 				    + (i + 1) * sizeof(struct mtd_desc));
387 		}
388 		/* sc->desc[i].conf = MTD_TXBUF_SIZE & MTD_TXD_CONF_BUFS; */
389 		/* Set buffer's address */
390 		sc->desc[i].data = htole32(sc->buf_dma_map->dm_segs[0].ds_addr
391 					+ MTD_NUM_RXD * MTD_RXBUF_SIZE
392 					+ (i - MTD_NUM_RXD) * MTD_TXBUF_SIZE);
393 	}
394 
395 	return 0;
396 }
397 
398 
399 void
mtd_mii_statchg(struct ifnet * ifp)400 mtd_mii_statchg(struct ifnet *ifp)
401 {
402 	/* Should we do something here? :) */
403 }
404 
405 
406 int
mtd_mii_readreg(device_t self,int phy,int reg,uint16_t * val)407 mtd_mii_readreg(device_t self, int phy, int reg, uint16_t *val)
408 {
409 	struct mtd_softc *sc = device_private(self);
410 
411 	*val = MTD_READ_2(sc, MTD_PHYBASE + reg * 2);
412 
413 	return 0;
414 }
415 
416 
417 int
mtd_mii_writereg(device_t self,int phy,int reg,uint16_t val)418 mtd_mii_writereg(device_t self, int phy, int reg, uint16_t val)
419 {
420 	struct mtd_softc *sc = device_private(self);
421 
422 	MTD_WRITE_2(sc, MTD_PHYBASE + reg * 2, val);
423 
424 	return 0;
425 }
426 
427 
428 int
mtd_put(struct mtd_softc * sc,int index,struct mbuf * m)429 mtd_put(struct mtd_softc *sc, int index, struct mbuf *m)
430 {
431 	int len, tlen;
432 	char *buf = (char *)sc->buf + MTD_NUM_RXD * MTD_RXBUF_SIZE
433 			+ index * MTD_TXBUF_SIZE;
434 	struct mbuf *n;
435 
436 	for (tlen = 0; m != NULL; m = n) {
437 		len = m->m_len;
438 		if (len == 0) {
439 			n = m_free(m);
440 			continue;
441 		} else if (tlen > MTD_TXBUF_SIZE) {
442 			/* XXX FIXME: No idea what to do here. */
443 			aprint_error_dev(sc->dev,
444 			    "packet too large! Size = %i\n", tlen);
445 			n = m_free(m);
446 			continue;
447 		}
448 		memcpy(buf, mtod(m, void *), len);
449 		buf += len;
450 		tlen += len;
451 		n = m_free(m);
452 	}
453 	sc->desc[MTD_NUM_RXD + index].conf = MTD_TXD_CONF_PAD | MTD_TXD_CONF_CRC
454 		| MTD_TXD_CONF_IRQC
455 		| ((tlen << MTD_TXD_PKTS_SHIFT) & MTD_TXD_CONF_PKTS)
456 		| (tlen & MTD_TXD_CONF_BUFS);
457 
458 	return tlen;
459 }
460 
461 
462 void
mtd_start(struct ifnet * ifp)463 mtd_start(struct ifnet *ifp)
464 {
465 	struct mtd_softc *sc = ifp->if_softc;
466 	struct mbuf *m;
467 	int first_tx = sc->cur_tx;
468 
469 	/* Don't transmit when the interface is busy or inactive */
470 	if ((ifp->if_flags & (IFF_RUNNING | IFF_OACTIVE)) != IFF_RUNNING)
471 		return;
472 
473 	for (;;) {
474 		IF_DEQUEUE(&ifp->if_snd, m);
475 
476 		if (m == NULL)
477 			break;
478 
479 		bpf_mtap(ifp, m, BPF_D_OUT);
480 
481 		/* Copy mbuf chain into tx buffer */
482 		(void)mtd_put(sc, sc->cur_tx, m);
483 
484 		if (sc->cur_tx != first_tx)
485 			sc->desc[MTD_NUM_RXD + sc->cur_tx].stat = MTD_TXD_OWNER;
486 
487 		if (++sc->cur_tx >= MTD_NUM_TXD)
488 			sc->cur_tx = 0;
489 	}
490 	/* Mark first & last descriptor */
491 	sc->desc[MTD_NUM_RXD + first_tx].conf |= MTD_TXD_CONF_FSD;
492 
493 	if (sc->cur_tx == 0)
494 		sc->desc[MTD_NUM_RXD + MTD_NUM_TXD - 1].conf
495 		    |= MTD_TXD_CONF_LSD;
496 	else
497 		sc->desc[MTD_NUM_RXD + sc->cur_tx - 1].conf
498 		    |= MTD_TXD_CONF_LSD;
499 
500 	/* Give first descriptor to chip to complete transaction */
501 	sc->desc[MTD_NUM_RXD + first_tx].stat = MTD_TXD_OWNER;
502 
503 	/* Transmit polling demand */
504 	MTD_WRITE_4(sc, MTD_TXPDR, MTD_TXPDR_DEMAND);
505 
506 	/* XXX FIXME: Set up a watchdog timer */
507 	/* ifp->if_timer = 5; */
508 }
509 
510 
511 void
mtd_stop(struct ifnet * ifp,int disable)512 mtd_stop(struct ifnet *ifp, int disable)
513 {
514 	struct mtd_softc *sc = ifp->if_softc;
515 
516 	/* Disable transmitter and receiver */
517 	MTD_CLRBIT(sc, MTD_RXTXR, MTD_TX_ENABLE);
518 	MTD_CLRBIT(sc, MTD_RXTXR, MTD_RX_ENABLE);
519 
520 	/* Disable interrupts */
521 	MTD_WRITE_4(sc, MTD_IMR, 0x00000000);
522 
523 	/* Must do more at disable??... */
524 	if (disable) {
525 		/* Delete tx and rx descriptor base addresses */
526 		MTD_WRITE_4(sc, MTD_RXLBA, 0x00000000);
527 		MTD_WRITE_4(sc, MTD_TXLBA, 0x00000000);
528 	}
529 
530 	ifp->if_timer = 0;
531 	ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
532 }
533 
534 
535 void
mtd_watchdog(struct ifnet * ifp)536 mtd_watchdog(struct ifnet *ifp)
537 {
538 	struct mtd_softc *sc = ifp->if_softc;
539 	int s;
540 
541 	log(LOG_ERR, "%s: device timeout\n", device_xname(sc->dev));
542 	if_statinc(ifp, if_oerrors);
543 
544 	mtd_stop(ifp, 0);
545 
546 	s = splnet();
547 	mtd_init(ifp);
548 	splx(s);
549 
550 	return;
551 }
552 
553 
554 int
mtd_ioctl(struct ifnet * ifp,u_long cmd,void * data)555 mtd_ioctl(struct ifnet *ifp, u_long cmd, void *data)
556 {
557 	struct mtd_softc *sc = ifp->if_softc;
558 	int s, error = 0;
559 
560 	s = splnet();
561 
562 	if ((error = ether_ioctl(ifp, cmd, data)) == ENETRESET) {
563 		/*
564 		 * Multicast list has changed; set the hardware
565 		 * filter accordingly.
566 		 */
567 		 if (ifp->if_flags & IFF_RUNNING)
568 			 mtd_setmulti(sc);
569 		 error = 0;
570 	}
571 
572 	splx(s);
573 	return error;
574 }
575 
576 
577 struct mbuf *
mtd_get(struct mtd_softc * sc,int index,int totlen)578 mtd_get(struct mtd_softc *sc, int index, int totlen)
579 {
580 	struct ifnet *ifp = &sc->ethercom.ec_if;
581 	struct mbuf *m, *m0, *newm;
582 	int len;
583 	char *buf = (char *)sc->buf + index * MTD_RXBUF_SIZE;
584 
585 	MGETHDR(m0, M_DONTWAIT, MT_DATA);
586 	if (m0 == NULL)
587 		return NULL;
588 
589 	m_set_rcvif(m0, ifp);
590 	m0->m_pkthdr.len = totlen;
591 	m = m0;
592 	len = MHLEN;
593 
594 	while (totlen > 0) {
595 		if (totlen >= MINCLSIZE) {
596 			MCLGET(m, M_DONTWAIT);
597 			if (!(m->m_flags & M_EXT)) {
598 				m_freem(m0);
599 				return NULL;
600 			}
601 			len = MCLBYTES;
602 		}
603 
604 		if (m == m0) {
605 			char *newdata = (char *)
606 				ALIGN(m->m_data + sizeof(struct ether_header)) -
607 				sizeof(struct ether_header);
608 			len -= newdata - m->m_data;
609 			m->m_data = newdata;
610 		}
611 
612 		m->m_len = len = uimin(totlen, len);
613 		memcpy(mtod(m, void *), buf, len);
614 		buf += len;
615 
616 		totlen -= len;
617 		if (totlen > 0) {
618 			MGET(newm, M_DONTWAIT, MT_DATA);
619 			if (newm == NULL) {
620 				m_freem(m0);
621 				return NULL;
622 			}
623 			len = MLEN;
624 			m = m->m_next = newm;
625 		}
626 	}
627 
628 	return m0;
629 }
630 
631 
632 int
mtd_rxirq(struct mtd_softc * sc)633 mtd_rxirq(struct mtd_softc *sc)
634 {
635 	struct ifnet *ifp = &sc->ethercom.ec_if;
636 	int len;
637 	struct mbuf *m;
638 
639 	for (; !(sc->desc[sc->cur_rx].stat & MTD_RXD_OWNER);) {
640 		/* Error summary set? */
641 		if (sc->desc[sc->cur_rx].stat & MTD_RXD_ERRSUM) {
642 			aprint_error_dev(sc->dev,
643 			    "received packet with errors\n");
644 			/* Give up packet, since an error occurred */
645 			sc->desc[sc->cur_rx].stat = MTD_RXD_OWNER;
646 			sc->desc[sc->cur_rx].conf = MTD_RXBUF_SIZE &
647 							MTD_RXD_CONF_BUFS;
648 			if_statinc(ifp, if_ierrors);
649 			if (++sc->cur_rx >= MTD_NUM_RXD)
650 				sc->cur_rx = 0;
651 			continue;
652 		}
653 		/* Get buffer length */
654 		len = (sc->desc[sc->cur_rx].stat & MTD_RXD_FLEN)
655 			>> MTD_RXD_FLEN_SHIFT;
656 		len -= ETHER_CRC_LEN;
657 
658 		/* Check packet size */
659 		if (len <= sizeof(struct ether_header)) {
660 			aprint_error_dev(sc->dev,
661 			    "invalid packet size %d; dropping\n", len);
662 			sc->desc[sc->cur_rx].stat = MTD_RXD_OWNER;
663 			sc->desc[sc->cur_rx].conf = MTD_RXBUF_SIZE &
664 							MTD_RXD_CONF_BUFS;
665 			if_statinc(ifp, if_ierrors);
666 			if (++sc->cur_rx >= MTD_NUM_RXD)
667 				sc->cur_rx = 0;
668 			continue;
669 		}
670 
671 		m = mtd_get(sc, (sc->cur_rx), len);
672 
673 		/* Give descriptor back to card */
674 		sc->desc[sc->cur_rx].conf = MTD_RXBUF_SIZE & MTD_RXD_CONF_BUFS;
675 		sc->desc[sc->cur_rx].stat = MTD_RXD_OWNER;
676 
677 		if (++sc->cur_rx >= MTD_NUM_RXD)
678 			sc->cur_rx = 0;
679 
680 		if (m == NULL) {
681 			aprint_error_dev(sc->dev,
682 			    "error pulling packet off interface\n");
683 			if_statinc(ifp, if_ierrors);
684 			continue;
685 		}
686 
687 		/* Pass the packet up */
688 		if_percpuq_enqueue(ifp->if_percpuq, m);
689 	}
690 
691 	return 1;
692 }
693 
694 
695 int
mtd_txirq(struct mtd_softc * sc)696 mtd_txirq(struct mtd_softc *sc)
697 {
698 	struct ifnet *ifp = &sc->ethercom.ec_if;
699 
700 	/* Clear timeout */
701 	ifp->if_timer = 0;
702 
703 	ifp->if_flags &= ~IFF_OACTIVE;
704 	if_statinc(ifp, if_opackets);
705 
706 	/* XXX FIXME If there is some queued, do an mtd_start? */
707 
708 	return 1;
709 }
710 
711 
712 int
mtd_bufirq(struct mtd_softc * sc)713 mtd_bufirq(struct mtd_softc *sc)
714 {
715 	struct ifnet *ifp = &sc->ethercom.ec_if;
716 
717 	/* Clear timeout */
718 	ifp->if_timer = 0;
719 
720 	/* XXX FIXME: Do something here to make sure we get some buffers! */
721 
722 	return 1;
723 }
724 
725 
726 int
mtd_irq_h(void * args)727 mtd_irq_h(void *args)
728 {
729 	struct mtd_softc *sc = args;
730 	struct ifnet *ifp = &sc->ethercom.ec_if;
731 	uint32_t status;
732 	int r = 0;
733 
734 	if (!(ifp->if_flags & IFF_RUNNING) || !device_is_active(sc->dev))
735 		return 0;
736 
737 	/* Disable interrupts */
738 	MTD_WRITE_4(sc, MTD_IMR, 0x00000000);
739 
740 	for (;;) {
741 		status = MTD_READ_4(sc, MTD_ISR);
742 
743 		/* Add random seed before masking out bits */
744 		if (status)
745 			rnd_add_uint32(&sc->rnd_src, status);
746 
747 		status &= MTD_ISR_MASK;
748 		if (!status)		/* We didn't ask for this */
749 			break;
750 
751 		MTD_WRITE_4(sc, MTD_ISR, status);
752 
753 		/* NOTE: Perhaps we should reset with some of these errors? */
754 
755 		if (status & MTD_ISR_RXBUN) {
756 			aprint_error_dev(sc->dev,
757 			    "receive buffer unavailable\n");
758 			if_statinc(ifp, if_ierrors);
759 		}
760 
761 		if (status & MTD_ISR_RXERR) {
762 			aprint_error_dev(sc->dev, "receive error\n");
763 			if_statinc(ifp, if_ierrors);
764 		}
765 
766 		if (status & MTD_ISR_TXBUN) {
767 			aprint_error_dev(sc->dev,
768 			    "transmit buffer unavailable\n");
769 			if_statinc(ifp, if_ierrors);
770 		}
771 
772 		if ((status & MTD_ISR_PDF)) {
773 			aprint_error_dev(sc->dev,
774 			    "parallel detection fault\n");
775 			if_statinc(ifp, if_ierrors);
776 		}
777 
778 		if (status & MTD_ISR_FBUSERR) {
779 			aprint_error_dev(sc->dev, "fatal bus error\n");
780 			if_statinc(ifp, if_ierrors);
781 		}
782 
783 		if (status & MTD_ISR_TARERR) {
784 			aprint_error_dev(sc->dev, "target error\n");
785 			if_statinc(ifp, if_ierrors);
786 		}
787 
788 		if (status & MTD_ISR_MASTERR) {
789 			aprint_error_dev(sc->dev, "master error\n");
790 			if_statinc(ifp, if_ierrors);
791 		}
792 
793 		if (status & MTD_ISR_PARERR) {
794 			aprint_error_dev(sc->dev, "parity error\n");
795 			if_statinc(ifp, if_ierrors);
796 		}
797 
798 		if (status & MTD_ISR_RXIRQ)	/* Receive interrupt */
799 			r |= mtd_rxirq(sc);
800 
801 		if (status & MTD_ISR_TXIRQ)	/* Transmit interrupt */
802 			r |= mtd_txirq(sc);
803 
804 		if (status & MTD_ISR_TXEARLY)	/* Transmit early */
805 			r |= mtd_txirq(sc);
806 
807 		if (status & MTD_ISR_TXBUN)	/* Transmit buffer n/a */
808 			r |= mtd_bufirq(sc);
809 
810 	}
811 
812 	/* Enable interrupts */
813 	MTD_WRITE_4(sc, MTD_IMR, MTD_IMR_MASK);
814 
815 	return r;
816 }
817 
818 
819 void
mtd_setmulti(struct mtd_softc * sc)820 mtd_setmulti(struct mtd_softc *sc)
821 {
822 	struct ethercom *ec = &sc->ethercom;
823 	struct ifnet *ifp = &ec->ec_if;
824 	uint32_t rxtx_stat;
825 	uint32_t hash[2] = {0, 0};
826 	uint32_t crc;
827 	struct ether_multi *enm;
828 	struct ether_multistep step;
829 	int mcnt = 0;
830 
831 	/* Get old status */
832 	rxtx_stat = MTD_READ_4(sc, MTD_RXTXR);
833 
834 	if ((ifp->if_flags & IFF_ALLMULTI) || (ifp->if_flags & IFF_PROMISC)) {
835 		rxtx_stat |= MTD_RX_AMULTI;
836 		MTD_WRITE_4(sc, MTD_RXTXR, rxtx_stat);
837 		MTD_WRITE_4(sc, MTD_MAR0, MTD_ALL_ADDR);
838 		MTD_WRITE_4(sc, MTD_MAR1, MTD_ALL_ADDR);
839 		return;
840 	}
841 
842 	ETHER_LOCK(ec);
843 	ETHER_FIRST_MULTI(step, ec, enm);
844 	while (enm != NULL) {
845 		/* We need the 6 most significant bits of the CRC */
846 		crc = ETHER_CRC32(enm->enm_addrlo, ETHER_ADDR_LEN) >> 26;
847 
848 		hash[crc >> 5] |= 1 << (crc & 0xf);
849 
850 		++mcnt;
851 		ETHER_NEXT_MULTI(step, enm);
852 	}
853 	ETHER_UNLOCK(ec);
854 
855 	/* Accept multicast bit needs to be on? */
856 	if (mcnt)
857 		rxtx_stat |= MTD_RX_AMULTI;
858 	else
859 		rxtx_stat &= ~MTD_RX_AMULTI;
860 
861 	/* Write out the hash */
862 	MTD_WRITE_4(sc, MTD_MAR0, hash[0]);
863 	MTD_WRITE_4(sc, MTD_MAR1, hash[1]);
864 	MTD_WRITE_4(sc, MTD_RXTXR, rxtx_stat);
865 }
866 
867 
868 void
mtd_reset(struct mtd_softc * sc)869 mtd_reset(struct mtd_softc *sc)
870 {
871 	int i;
872 
873 	MTD_SETBIT(sc, MTD_BCR, MTD_BCR_RESET);
874 
875 	/* Reset descriptor status */
876 	sc->cur_tx = 0;
877 	sc->cur_rx = 0;
878 
879 	/* Wait until done with reset */
880 	for (i = 0; i < MTD_TIMEOUT; ++i) {
881 		DELAY(10);
882 		if (!(MTD_READ_4(sc, MTD_BCR) & MTD_BCR_RESET))
883 			break;
884 	}
885 
886 	if (i == MTD_TIMEOUT)
887 		aprint_error_dev(sc->dev, "reset timed out\n");
888 
889 	/* Wait a little so chip can stabilize */
890 	DELAY(1000);
891 }
892 
893 
894 void
mtd_shutdown(void * arg)895 mtd_shutdown (void *arg)
896 {
897 	struct mtd_softc *sc = arg;
898 	struct ifnet *ifp = &sc->ethercom.ec_if;
899 
900 	rnd_detach_source(&sc->rnd_src);
901 	mtd_stop(ifp, 1);
902 }
903