1 /* 2 * Copyright (c) 1997, 1998 3 * Bill Paul <wpaul@ctr.columbia.edu>. All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 3. All advertising materials mentioning features or use of this software 14 * must display the following acknowledgement: 15 * This product includes software developed by Bill Paul. 16 * 4. Neither the name of the author nor the names of any co-contributors 17 * may be used to endorse or promote products derived from this software 18 * without specific prior written permission. 19 * 20 * THIS SOFTWARE IS PROVIDED BY Bill Paul AND CONTRIBUTORS ``AS IS'' AND 21 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 22 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 23 * ARE DISCLAIMED. IN NO EVENT SHALL Bill Paul OR THE VOICES IN HIS HEAD 24 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 25 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 26 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 27 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 28 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 29 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF 30 * THE POSSIBILITY OF SUCH DAMAGE. 31 * 32 * $FreeBSD: src/sys/pci/if_wb.c,v 1.26.2.6 2003/03/05 18:42:34 njl Exp $ 33 * $DragonFly: src/sys/dev/netif/wb/if_wb.c,v 1.34 2005/12/31 14:08:00 sephe Exp $ 34 */ 35 36 /* 37 * Winbond fast ethernet PCI NIC driver 38 * 39 * Supports various cheap network adapters based on the Winbond W89C840F 40 * fast ethernet controller chip. This includes adapters manufactured by 41 * Winbond itself and some made by Linksys. 42 * 43 * Written by Bill Paul <wpaul@ctr.columbia.edu> 44 * Electrical Engineering Department 45 * Columbia University, New York City 46 */ 47 48 /* 49 * The Winbond W89C840F chip is a bus master; in some ways it resembles 50 * a DEC 'tulip' chip, only not as complicated. Unfortunately, it has 51 * one major difference which is that while the registers do many of 52 * the same things as a tulip adapter, the offsets are different: where 53 * tulip registers are typically spaced 8 bytes apart, the Winbond 54 * registers are spaced 4 bytes apart. The receiver filter is also 55 * programmed differently. 56 * 57 * Like the tulip, the Winbond chip uses small descriptors containing 58 * a status word, a control word and 32-bit areas that can either be used 59 * to point to two external data blocks, or to point to a single block 60 * and another descriptor in a linked list. Descriptors can be grouped 61 * together in blocks to form fixed length rings or can be chained 62 * together in linked lists. A single packet may be spread out over 63 * several descriptors if necessary. 64 * 65 * For the receive ring, this driver uses a linked list of descriptors, 66 * each pointing to a single mbuf cluster buffer, which us large enough 67 * to hold an entire packet. The link list is looped back to created a 68 * closed ring. 69 * 70 * For transmission, the driver creates a linked list of 'super descriptors' 71 * which each contain several individual descriptors linked toghether. 72 * Each 'super descriptor' contains WB_MAXFRAGS descriptors, which we 73 * abuse as fragment pointers. This allows us to use a buffer managment 74 * scheme very similar to that used in the ThunderLAN and Etherlink XL 75 * drivers. 76 * 77 * Autonegotiation is performed using the external PHY via the MII bus. 78 * The sample boards I have all use a Davicom PHY. 79 * 80 * Note: the author of the Linux driver for the Winbond chip alludes 81 * to some sort of flaw in the chip's design that seems to mandate some 82 * drastic workaround which signigicantly impairs transmit performance. 83 * I have no idea what he's on about: transmit performance with all 84 * three of my test boards seems fine. 85 */ 86 87 #include "opt_bdg.h" 88 89 #include <sys/param.h> 90 #include <sys/systm.h> 91 #include <sys/sockio.h> 92 #include <sys/mbuf.h> 93 #include <sys/malloc.h> 94 #include <sys/kernel.h> 95 #include <sys/socket.h> 96 #include <sys/queue.h> 97 #include <sys/serialize.h> 98 #include <sys/thread2.h> 99 100 #include <net/if.h> 101 #include <net/ifq_var.h> 102 #include <net/if_arp.h> 103 #include <net/ethernet.h> 104 #include <net/if_dl.h> 105 #include <net/if_media.h> 106 107 #include <net/bpf.h> 108 109 #include <vm/vm.h> /* for vtophys */ 110 #include <vm/pmap.h> /* for vtophys */ 111 #include <machine/bus.h> 112 #include <machine/resource.h> 113 #include <sys/bus.h> 114 #include <sys/rman.h> 115 116 #include <bus/pci/pcireg.h> 117 #include <bus/pci/pcivar.h> 118 119 #include <dev/netif/mii_layer/mii.h> 120 #include <dev/netif/mii_layer/miivar.h> 121 122 /* "controller miibus0" required. See GENERIC if you get errors here. */ 123 #include "miibus_if.h" 124 125 #define WB_USEIOSPACE 126 127 #include "if_wbreg.h" 128 129 /* 130 * Various supported device vendors/types and their names. 131 */ 132 static struct wb_type wb_devs[] = { 133 { WB_VENDORID, WB_DEVICEID_840F, 134 "Winbond W89C840F 10/100BaseTX" }, 135 { CP_VENDORID, CP_DEVICEID_RL100, 136 "Compex RL100-ATX 10/100baseTX" }, 137 { 0, 0, NULL } 138 }; 139 140 static int wb_probe(device_t); 141 static int wb_attach(device_t); 142 static int wb_detach(device_t); 143 144 static void wb_bfree(void *); 145 static int wb_newbuf(struct wb_softc *, struct wb_chain_onefrag *, 146 struct mbuf *); 147 static int wb_encap(struct wb_softc *, struct wb_chain *, struct mbuf *); 148 149 static void wb_rxeof(struct wb_softc *); 150 static void wb_rxeoc(struct wb_softc *); 151 static void wb_txeof(struct wb_softc *); 152 static void wb_txeoc(struct wb_softc *); 153 static void wb_intr(void *); 154 static void wb_tick(void *); 155 static void wb_start(struct ifnet *); 156 static int wb_ioctl(struct ifnet *, u_long, caddr_t, struct ucred *); 157 static void wb_init(void *); 158 static void wb_stop(struct wb_softc *); 159 static void wb_watchdog(struct ifnet *); 160 static void wb_shutdown(device_t); 161 static int wb_ifmedia_upd(struct ifnet *); 162 static void wb_ifmedia_sts(struct ifnet *, struct ifmediareq *); 163 164 static void wb_eeprom_putbyte(struct wb_softc *, int); 165 static void wb_eeprom_getword(struct wb_softc *, int, uint16_t *); 166 static void wb_read_eeprom(struct wb_softc *, caddr_t, int, int); 167 static void wb_mii_sync(struct wb_softc *); 168 static void wb_mii_send(struct wb_softc *, uint32_t, int); 169 static int wb_mii_readreg(struct wb_softc *, struct wb_mii_frame *); 170 static int wb_mii_writereg(struct wb_softc *, struct wb_mii_frame *); 171 172 static void wb_setcfg(struct wb_softc *, uint32_t); 173 static void wb_setmulti(struct wb_softc *); 174 static void wb_reset(struct wb_softc *); 175 static void wb_fixmedia(struct wb_softc *); 176 static int wb_list_rx_init(struct wb_softc *); 177 static int wb_list_tx_init(struct wb_softc *); 178 179 static int wb_miibus_readreg(device_t, int, int); 180 static int wb_miibus_writereg(device_t, int, int, int); 181 static void wb_miibus_statchg(device_t); 182 183 #ifdef WB_USEIOSPACE 184 #define WB_RES SYS_RES_IOPORT 185 #define WB_RID WB_PCI_LOIO 186 #else 187 #define WB_RES SYS_RES_MEMORY 188 #define WB_RID WB_PCI_LOMEM 189 #endif 190 191 static device_method_t wb_methods[] = { 192 /* Device interface */ 193 DEVMETHOD(device_probe, wb_probe), 194 DEVMETHOD(device_attach, wb_attach), 195 DEVMETHOD(device_detach, wb_detach), 196 DEVMETHOD(device_shutdown, wb_shutdown), 197 198 /* bus interface, for miibus */ 199 DEVMETHOD(bus_print_child, bus_generic_print_child), 200 DEVMETHOD(bus_driver_added, bus_generic_driver_added), 201 202 /* MII interface */ 203 DEVMETHOD(miibus_readreg, wb_miibus_readreg), 204 DEVMETHOD(miibus_writereg, wb_miibus_writereg), 205 DEVMETHOD(miibus_statchg, wb_miibus_statchg), 206 { 0, 0 } 207 }; 208 209 static DEFINE_CLASS_0(wb, wb_driver, wb_methods, sizeof(struct wb_softc)); 210 static devclass_t wb_devclass; 211 212 DECLARE_DUMMY_MODULE(if_wb); 213 DRIVER_MODULE(if_wb, pci, wb_driver, wb_devclass, 0, 0); 214 DRIVER_MODULE(miibus, wb, miibus_driver, miibus_devclass, 0, 0); 215 216 #define WB_SETBIT(sc, reg, x) \ 217 CSR_WRITE_4(sc, reg, CSR_READ_4(sc, reg) | (x)) 218 219 #define WB_CLRBIT(sc, reg, x) \ 220 CSR_WRITE_4(sc, reg, CSR_READ_4(sc, reg) & ~(x)) 221 222 #define SIO_SET(x) \ 223 CSR_WRITE_4(sc, WB_SIO, CSR_READ_4(sc, WB_SIO) | (x)) 224 225 #define SIO_CLR(x) \ 226 CSR_WRITE_4(sc, WB_SIO, CSR_READ_4(sc, WB_SIO) & ~(x)) 227 228 /* 229 * Send a read command and address to the EEPROM, check for ACK. 230 */ 231 static void 232 wb_eeprom_putbyte(struct wb_softc *sc, int addr) 233 { 234 int d, i; 235 236 d = addr | WB_EECMD_READ; 237 238 /* 239 * Feed in each bit and stobe the clock. 240 */ 241 for (i = 0x400; i; i >>= 1) { 242 if (d & i) 243 SIO_SET(WB_SIO_EE_DATAIN); 244 else 245 SIO_CLR(WB_SIO_EE_DATAIN); 246 DELAY(100); 247 SIO_SET(WB_SIO_EE_CLK); 248 DELAY(150); 249 SIO_CLR(WB_SIO_EE_CLK); 250 DELAY(100); 251 } 252 } 253 254 /* 255 * Read a word of data stored in the EEPROM at address 'addr.' 256 */ 257 static void 258 wb_eeprom_getword(struct wb_softc *sc, int addr, uint16_t *dest) 259 { 260 int i; 261 uint16_t word = 0; 262 263 /* Enter EEPROM access mode. */ 264 CSR_WRITE_4(sc, WB_SIO, WB_SIO_EESEL|WB_SIO_EE_CS); 265 266 /* 267 * Send address of word we want to read. 268 */ 269 wb_eeprom_putbyte(sc, addr); 270 271 CSR_WRITE_4(sc, WB_SIO, WB_SIO_EESEL|WB_SIO_EE_CS); 272 273 /* 274 * Start reading bits from EEPROM. 275 */ 276 for (i = 0x8000; i; i >>= 1) { 277 SIO_SET(WB_SIO_EE_CLK); 278 DELAY(100); 279 if (CSR_READ_4(sc, WB_SIO) & WB_SIO_EE_DATAOUT) 280 word |= i; 281 SIO_CLR(WB_SIO_EE_CLK); 282 DELAY(100); 283 } 284 285 /* Turn off EEPROM access mode. */ 286 CSR_WRITE_4(sc, WB_SIO, 0); 287 288 *dest = word; 289 } 290 291 /* 292 * Read a sequence of words from the EEPROM. 293 */ 294 static void 295 wb_read_eeprom(struct wb_softc *sc, caddr_t dest, int off, int cnt) 296 { 297 int i; 298 uint16_t word = 0, *ptr; 299 300 for (i = 0; i < cnt; i++) { 301 wb_eeprom_getword(sc, off + i, &word); 302 ptr = (uint16_t *)(dest + (i * 2)); 303 *ptr = word; 304 } 305 } 306 307 /* 308 * Sync the PHYs by setting data bit and strobing the clock 32 times. 309 */ 310 static void 311 wb_mii_sync(struct wb_softc *sc) 312 { 313 int i; 314 315 SIO_SET(WB_SIO_MII_DIR | WB_SIO_MII_DATAIN); 316 317 for (i = 0; i < 32; i++) { 318 SIO_SET(WB_SIO_MII_CLK); 319 DELAY(1); 320 SIO_CLR(WB_SIO_MII_CLK); 321 DELAY(1); 322 } 323 } 324 325 /* 326 * Clock a series of bits through the MII. 327 */ 328 static void 329 wb_mii_send(struct wb_softc *sc, uint32_t bits, int cnt) 330 { 331 int i; 332 333 SIO_CLR(WB_SIO_MII_CLK); 334 335 for (i = (0x1 << (cnt - 1)); i; i >>= 1) { 336 if (bits & i) 337 SIO_SET(WB_SIO_MII_DATAIN); 338 else 339 SIO_CLR(WB_SIO_MII_DATAIN); 340 DELAY(1); 341 SIO_CLR(WB_SIO_MII_CLK); 342 DELAY(1); 343 SIO_SET(WB_SIO_MII_CLK); 344 } 345 } 346 347 /* 348 * Read an PHY register through the MII. 349 */ 350 static int 351 wb_mii_readreg(struct wb_softc *sc, struct wb_mii_frame *frame) 352 { 353 int ack, i; 354 355 crit_enter(); 356 357 /* 358 * Set up frame for RX. 359 */ 360 frame->mii_stdelim = WB_MII_STARTDELIM; 361 frame->mii_opcode = WB_MII_READOP; 362 frame->mii_turnaround = 0; 363 frame->mii_data = 0; 364 365 CSR_WRITE_4(sc, WB_SIO, 0); 366 367 /* 368 * Turn on data xmit. 369 */ 370 SIO_SET(WB_SIO_MII_DIR); 371 372 wb_mii_sync(sc); 373 374 /* 375 * Send command/address info. 376 */ 377 wb_mii_send(sc, frame->mii_stdelim, 2); 378 wb_mii_send(sc, frame->mii_opcode, 2); 379 wb_mii_send(sc, frame->mii_phyaddr, 5); 380 wb_mii_send(sc, frame->mii_regaddr, 5); 381 382 /* Idle bit */ 383 SIO_CLR((WB_SIO_MII_CLK | WB_SIO_MII_DATAIN)); 384 DELAY(1); 385 SIO_SET(WB_SIO_MII_CLK); 386 DELAY(1); 387 388 /* Turn off xmit. */ 389 SIO_CLR(WB_SIO_MII_DIR); 390 /* Check for ack */ 391 SIO_CLR(WB_SIO_MII_CLK); 392 DELAY(1); 393 ack = CSR_READ_4(sc, WB_SIO) & WB_SIO_MII_DATAOUT; 394 SIO_SET(WB_SIO_MII_CLK); 395 DELAY(1); 396 SIO_CLR(WB_SIO_MII_CLK); 397 DELAY(1); 398 SIO_SET(WB_SIO_MII_CLK); 399 DELAY(1); 400 401 /* 402 * Now try reading data bits. If the ack failed, we still 403 * need to clock through 16 cycles to keep the PHY(s) in sync. 404 */ 405 if (ack) { 406 for(i = 0; i < 16; i++) { 407 SIO_CLR(WB_SIO_MII_CLK); 408 DELAY(1); 409 SIO_SET(WB_SIO_MII_CLK); 410 DELAY(1); 411 } 412 goto fail; 413 } 414 415 for (i = 0x8000; i; i >>= 1) { 416 SIO_CLR(WB_SIO_MII_CLK); 417 DELAY(1); 418 if (!ack) { 419 if (CSR_READ_4(sc, WB_SIO) & WB_SIO_MII_DATAOUT) 420 frame->mii_data |= i; 421 DELAY(1); 422 } 423 SIO_SET(WB_SIO_MII_CLK); 424 DELAY(1); 425 } 426 427 fail: 428 429 SIO_CLR(WB_SIO_MII_CLK); 430 DELAY(1); 431 SIO_SET(WB_SIO_MII_CLK); 432 DELAY(1); 433 434 crit_exit(); 435 436 if (ack) 437 return(1); 438 return(0); 439 } 440 441 /* 442 * Write to a PHY register through the MII. 443 */ 444 static int 445 wb_mii_writereg(struct wb_softc *sc, struct wb_mii_frame *frame) 446 { 447 448 crit_enter(); 449 /* 450 * Set up frame for TX. 451 */ 452 453 frame->mii_stdelim = WB_MII_STARTDELIM; 454 frame->mii_opcode = WB_MII_WRITEOP; 455 frame->mii_turnaround = WB_MII_TURNAROUND; 456 457 /* 458 * Turn on data output. 459 */ 460 SIO_SET(WB_SIO_MII_DIR); 461 462 wb_mii_sync(sc); 463 464 wb_mii_send(sc, frame->mii_stdelim, 2); 465 wb_mii_send(sc, frame->mii_opcode, 2); 466 wb_mii_send(sc, frame->mii_phyaddr, 5); 467 wb_mii_send(sc, frame->mii_regaddr, 5); 468 wb_mii_send(sc, frame->mii_turnaround, 2); 469 wb_mii_send(sc, frame->mii_data, 16); 470 471 /* Idle bit. */ 472 SIO_SET(WB_SIO_MII_CLK); 473 DELAY(1); 474 SIO_CLR(WB_SIO_MII_CLK); 475 DELAY(1); 476 477 /* 478 * Turn off xmit. 479 */ 480 SIO_CLR(WB_SIO_MII_DIR); 481 482 crit_exit(); 483 484 return(0); 485 } 486 487 static int 488 wb_miibus_readreg(device_t dev, int phy, int reg) 489 { 490 struct wb_softc *sc = device_get_softc(dev); 491 struct wb_mii_frame frame; 492 493 bzero(&frame, sizeof(frame)); 494 495 frame.mii_phyaddr = phy; 496 frame.mii_regaddr = reg; 497 wb_mii_readreg(sc, &frame); 498 499 return(frame.mii_data); 500 } 501 502 static int 503 wb_miibus_writereg(device_t dev, int phy, int reg, int data) 504 { 505 struct wb_softc *sc = device_get_softc(dev); 506 struct wb_mii_frame frame; 507 508 bzero(&frame, sizeof(frame)); 509 510 frame.mii_phyaddr = phy; 511 frame.mii_regaddr = reg; 512 frame.mii_data = data; 513 514 wb_mii_writereg(sc, &frame); 515 516 return(0); 517 } 518 519 static void 520 wb_miibus_statchg(device_t dev) 521 { 522 struct wb_softc *sc = device_get_softc(dev); 523 struct mii_data *mii; 524 525 mii = device_get_softc(sc->wb_miibus); 526 wb_setcfg(sc, mii->mii_media_active); 527 } 528 529 /* 530 * Program the 64-bit multicast hash filter. 531 */ 532 static void 533 wb_setmulti(struct wb_softc *sc) 534 { 535 struct ifnet *ifp = &sc->arpcom.ac_if; 536 int h = 0, mcnt = 0; 537 uint32_t hashes[2] = { 0, 0 }; 538 struct ifmultiaddr *ifma; 539 uint32_t rxfilt; 540 541 rxfilt = CSR_READ_4(sc, WB_NETCFG); 542 543 if (ifp->if_flags & IFF_ALLMULTI || ifp->if_flags & IFF_PROMISC) { 544 rxfilt |= WB_NETCFG_RX_MULTI; 545 CSR_WRITE_4(sc, WB_NETCFG, rxfilt); 546 CSR_WRITE_4(sc, WB_MAR0, 0xFFFFFFFF); 547 CSR_WRITE_4(sc, WB_MAR1, 0xFFFFFFFF); 548 return; 549 } 550 551 /* first, zot all the existing hash bits */ 552 CSR_WRITE_4(sc, WB_MAR0, 0); 553 CSR_WRITE_4(sc, WB_MAR1, 0); 554 555 /* now program new ones */ 556 LIST_FOREACH(ifma, &ifp->if_multiaddrs, ifma_link) { 557 if (ifma->ifma_addr->sa_family != AF_LINK) 558 continue; 559 h = ~ether_crc32_be(LLADDR((struct sockaddr_dl *) 560 ifma->ifma_addr), ETHER_ADDR_LEN) >> 26; 561 if (h < 32) 562 hashes[0] |= (1 << h); 563 else 564 hashes[1] |= (1 << (h - 32)); 565 mcnt++; 566 } 567 568 if (mcnt) 569 rxfilt |= WB_NETCFG_RX_MULTI; 570 else 571 rxfilt &= ~WB_NETCFG_RX_MULTI; 572 573 CSR_WRITE_4(sc, WB_MAR0, hashes[0]); 574 CSR_WRITE_4(sc, WB_MAR1, hashes[1]); 575 CSR_WRITE_4(sc, WB_NETCFG, rxfilt); 576 } 577 578 /* 579 * The Winbond manual states that in order to fiddle with the 580 * 'full-duplex' and '100Mbps' bits in the netconfig register, we 581 * first have to put the transmit and/or receive logic in the idle state. 582 */ 583 static void 584 wb_setcfg(struct wb_softc *sc, uint32_t media) 585 { 586 int i, restart = 0; 587 588 if (CSR_READ_4(sc, WB_NETCFG) & (WB_NETCFG_TX_ON | WB_NETCFG_RX_ON)) { 589 restart = 1; 590 WB_CLRBIT(sc, WB_NETCFG, (WB_NETCFG_TX_ON | WB_NETCFG_RX_ON)); 591 592 for (i = 0; i < WB_TIMEOUT; i++) { 593 DELAY(10); 594 if ((CSR_READ_4(sc, WB_ISR) & WB_ISR_TX_IDLE) && 595 (CSR_READ_4(sc, WB_ISR) & WB_ISR_RX_IDLE)) 596 break; 597 } 598 599 if (i == WB_TIMEOUT) { 600 if_printf(&sc->arpcom.ac_if, "failed to force tx and " 601 "rx to idle state\n"); 602 } 603 } 604 605 if (IFM_SUBTYPE(media) == IFM_10_T) 606 WB_CLRBIT(sc, WB_NETCFG, WB_NETCFG_100MBPS); 607 else 608 WB_SETBIT(sc, WB_NETCFG, WB_NETCFG_100MBPS); 609 610 if ((media & IFM_GMASK) == IFM_FDX) 611 WB_SETBIT(sc, WB_NETCFG, WB_NETCFG_FULLDUPLEX); 612 else 613 WB_CLRBIT(sc, WB_NETCFG, WB_NETCFG_FULLDUPLEX); 614 615 if (restart) 616 WB_SETBIT(sc, WB_NETCFG, WB_NETCFG_TX_ON | WB_NETCFG_RX_ON); 617 } 618 619 static void 620 wb_reset(struct wb_softc *sc) 621 { 622 int i; 623 struct mii_data *mii; 624 625 CSR_WRITE_4(sc, WB_NETCFG, 0); 626 CSR_WRITE_4(sc, WB_BUSCTL, 0); 627 CSR_WRITE_4(sc, WB_TXADDR, 0); 628 CSR_WRITE_4(sc, WB_RXADDR, 0); 629 630 WB_SETBIT(sc, WB_BUSCTL, WB_BUSCTL_RESET); 631 WB_SETBIT(sc, WB_BUSCTL, WB_BUSCTL_RESET); 632 633 for (i = 0; i < WB_TIMEOUT; i++) { 634 DELAY(10); 635 if ((CSR_READ_4(sc, WB_BUSCTL) & WB_BUSCTL_RESET) == 0) 636 break; 637 } 638 if (i == WB_TIMEOUT) 639 if_printf(&sc->arpcom.ac_if, "reset never completed!\n"); 640 641 /* Wait a little while for the chip to get its brains in order. */ 642 DELAY(1000); 643 644 if (sc->wb_miibus == NULL) 645 return; 646 647 mii = device_get_softc(sc->wb_miibus); 648 if (mii == NULL) 649 return; 650 651 if (mii->mii_instance) { 652 struct mii_softc *miisc; 653 LIST_FOREACH(miisc, &mii->mii_phys, mii_list) 654 mii_phy_reset(miisc); 655 } 656 } 657 658 static void 659 wb_fixmedia(struct wb_softc *sc) 660 { 661 struct mii_data *mii; 662 uint32_t media; 663 664 if (sc->wb_miibus == NULL) 665 return; 666 667 mii = device_get_softc(sc->wb_miibus); 668 669 mii_pollstat(mii); 670 if (IFM_SUBTYPE(mii->mii_media_active) == IFM_10_T) { 671 media = mii->mii_media_active & ~IFM_10_T; 672 media |= IFM_100_TX; 673 } else if (IFM_SUBTYPE(mii->mii_media_active) == IFM_100_TX) { 674 media = mii->mii_media_active & ~IFM_100_TX; 675 media |= IFM_10_T; 676 } else 677 return; 678 679 ifmedia_set(&mii->mii_media, media); 680 } 681 682 /* 683 * Probe for a Winbond chip. Check the PCI vendor and device 684 * IDs against our list and return a device name if we find a match. 685 */ 686 static int wb_probe(device_t dev) 687 { 688 struct wb_type *t; 689 uint16_t vendor, product; 690 691 vendor = pci_get_vendor(dev); 692 product = pci_get_device(dev); 693 694 for (t = wb_devs; t->wb_name != NULL; t++) { 695 if (vendor == t->wb_vid && product == t->wb_did) { 696 device_set_desc(dev, t->wb_name); 697 return(0); 698 } 699 } 700 701 return(ENXIO); 702 } 703 704 /* 705 * Attach the interface. Allocate softc structures, do ifmedia 706 * setup and ethernet/BPF attach. 707 */ 708 static int 709 wb_attach(device_t dev) 710 { 711 u_char eaddr[ETHER_ADDR_LEN]; 712 struct wb_softc *sc; 713 struct ifnet *ifp; 714 int error = 0, rid; 715 716 sc = device_get_softc(dev); 717 callout_init(&sc->wb_stat_timer); 718 719 /* 720 * Handle power management nonsense. 721 */ 722 if (pci_get_powerstate(dev) != PCI_POWERSTATE_D0) { 723 uint32_t iobase, membase, irq; 724 725 /* Save important PCI config data. */ 726 iobase = pci_read_config(dev, WB_PCI_LOIO, 4); 727 membase = pci_read_config(dev, WB_PCI_LOMEM, 4); 728 irq = pci_read_config(dev, WB_PCI_INTLINE, 4); 729 730 /* Reset the power state. */ 731 device_printf(dev, "chip is in D%d power mode " 732 "-- setting to D0\n", pci_get_powerstate(dev)); 733 pci_set_powerstate(dev, PCI_POWERSTATE_D0); 734 735 /* Restore PCI config data. */ 736 pci_write_config(dev, WB_PCI_LOIO, iobase, 4); 737 pci_write_config(dev, WB_PCI_LOMEM, membase, 4); 738 pci_write_config(dev, WB_PCI_INTLINE, irq, 4); 739 } 740 741 pci_enable_busmaster(dev); 742 743 rid = WB_RID; 744 sc->wb_res = bus_alloc_resource_any(dev, WB_RES, &rid, RF_ACTIVE); 745 746 if (sc->wb_res == NULL) { 747 device_printf(dev, "couldn't map ports/memory\n"); 748 error = ENXIO; 749 goto fail; 750 } 751 752 sc->wb_btag = rman_get_bustag(sc->wb_res); 753 sc->wb_bhandle = rman_get_bushandle(sc->wb_res); 754 755 /* Allocate interrupt */ 756 rid = 0; 757 sc->wb_irq = bus_alloc_resource_any(dev, SYS_RES_IRQ, &rid, 758 RF_SHAREABLE | RF_ACTIVE); 759 760 if (sc->wb_irq == NULL) { 761 device_printf(dev, "couldn't map interrupt\n"); 762 error = ENXIO; 763 goto fail; 764 } 765 766 /* Save the cache line size. */ 767 sc->wb_cachesize = pci_read_config(dev, WB_PCI_CACHELEN, 4) & 0xFF; 768 769 ifp = &sc->arpcom.ac_if; 770 if_initname(ifp, device_get_name(dev), device_get_unit(dev)); 771 772 /* Reset the adapter. */ 773 wb_reset(sc); 774 775 /* 776 * Get station address from the EEPROM. 777 */ 778 wb_read_eeprom(sc, (caddr_t)&eaddr, 0, 3); 779 780 sc->wb_ldata = contigmalloc(sizeof(struct wb_list_data) + 8, M_DEVBUF, 781 M_WAITOK, 0, 0xffffffff, PAGE_SIZE, 0); 782 783 if (sc->wb_ldata == NULL) { 784 device_printf(dev, "no memory for list buffers!\n"); 785 error = ENXIO; 786 goto fail; 787 } 788 789 bzero(sc->wb_ldata, sizeof(struct wb_list_data)); 790 791 ifp->if_softc = sc; 792 ifp->if_mtu = ETHERMTU; 793 ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST; 794 ifp->if_ioctl = wb_ioctl; 795 ifp->if_start = wb_start; 796 ifp->if_watchdog = wb_watchdog; 797 ifp->if_init = wb_init; 798 ifp->if_baudrate = 10000000; 799 ifq_set_maxlen(&ifp->if_snd, WB_TX_LIST_CNT - 1); 800 ifq_set_ready(&ifp->if_snd); 801 802 /* 803 * Do MII setup. 804 */ 805 if (mii_phy_probe(dev, &sc->wb_miibus, 806 wb_ifmedia_upd, wb_ifmedia_sts)) { 807 error = ENXIO; 808 goto fail; 809 } 810 811 /* 812 * Call MI attach routine. 813 */ 814 ether_ifattach(ifp, eaddr, NULL); 815 816 error = bus_setup_intr(dev, sc->wb_irq, INTR_NETSAFE, 817 wb_intr, sc, &sc->wb_intrhand, 818 ifp->if_serializer); 819 820 if (error) { 821 device_printf(dev, "couldn't set up irq\n"); 822 ether_ifdetach(ifp); 823 goto fail; 824 } 825 826 return(0); 827 828 fail: 829 wb_detach(dev); 830 return(error); 831 } 832 833 static int 834 wb_detach(device_t dev) 835 { 836 struct wb_softc *sc = device_get_softc(dev); 837 struct ifnet *ifp = &sc->arpcom.ac_if; 838 839 840 if (device_is_attached(dev)) { 841 lwkt_serialize_enter(ifp->if_serializer); 842 wb_stop(sc); 843 bus_teardown_intr(dev, sc->wb_irq, sc->wb_intrhand); 844 lwkt_serialize_exit(ifp->if_serializer); 845 846 ether_ifdetach(ifp); 847 } 848 849 if (sc->wb_miibus) 850 device_delete_child(dev, sc->wb_miibus); 851 bus_generic_detach(dev); 852 853 if (sc->wb_irq); 854 bus_release_resource(dev, SYS_RES_IRQ, 0, sc->wb_irq); 855 if (sc->wb_res) 856 bus_release_resource(dev, WB_RES, WB_RID, sc->wb_res); 857 if (sc->wb_ldata_ptr) { 858 contigfree(sc->wb_ldata_ptr, sizeof(struct wb_list_data) + 8, 859 M_DEVBUF); 860 } 861 862 return(0); 863 } 864 865 /* 866 * Initialize the transmit descriptors. 867 */ 868 static int 869 wb_list_tx_init(struct wb_softc *sc) 870 { 871 struct wb_chain_data *cd; 872 struct wb_list_data *ld; 873 int i, nexti; 874 875 cd = &sc->wb_cdata; 876 ld = sc->wb_ldata; 877 878 for (i = 0; i < WB_TX_LIST_CNT; i++) { 879 nexti = (i == WB_TX_LIST_CNT - 1) ? 0 : i + 1; 880 cd->wb_tx_chain[i].wb_ptr = &ld->wb_tx_list[i]; 881 cd->wb_tx_chain[i].wb_nextdesc = &cd->wb_tx_chain[nexti]; 882 } 883 884 cd->wb_tx_free = &cd->wb_tx_chain[0]; 885 cd->wb_tx_tail = cd->wb_tx_head = NULL; 886 887 return(0); 888 } 889 890 /* 891 * Initialize the RX descriptors and allocate mbufs for them. Note that 892 * we arrange the descriptors in a closed ring, so that the last descriptor 893 * points back to the first. 894 */ 895 static int 896 wb_list_rx_init(struct wb_softc *sc) 897 { 898 struct wb_chain_data *cd; 899 struct wb_list_data *ld; 900 int i, nexti; 901 902 cd = &sc->wb_cdata; 903 ld = sc->wb_ldata; 904 905 for (i = 0; i < WB_RX_LIST_CNT; i++) { 906 cd->wb_rx_chain[i].wb_ptr = &ld->wb_rx_list[i]; 907 cd->wb_rx_chain[i].wb_buf = &ld->wb_rxbufs[i]; 908 if (wb_newbuf(sc, &cd->wb_rx_chain[i], NULL) == ENOBUFS) 909 return(ENOBUFS); 910 nexti = (WB_RX_LIST_CNT - 1) ? 0 : i + 1; 911 cd->wb_rx_chain[i].wb_nextdesc = &cd->wb_rx_chain[nexti]; 912 ld->wb_rx_list[i].wb_next = vtophys(&ld->wb_rx_list[nexti]); 913 } 914 915 cd->wb_rx_head = &cd->wb_rx_chain[0]; 916 917 return(0); 918 } 919 920 static void 921 wb_bfree(void *arg) 922 { 923 } 924 925 /* 926 * Initialize an RX descriptor and attach an MBUF cluster. 927 */ 928 static int 929 wb_newbuf(struct wb_softc *sc, struct wb_chain_onefrag *c, struct mbuf *m) 930 { 931 struct mbuf *m_new = NULL; 932 933 if (m == NULL) { 934 MGETHDR(m_new, MB_DONTWAIT, MT_DATA); 935 if (m_new == NULL) 936 return(ENOBUFS); 937 938 m_new->m_data = m_new->m_ext.ext_buf = c->wb_buf; 939 m_new->m_flags |= M_EXT; 940 m_new->m_ext.ext_size = m_new->m_pkthdr.len = 941 m_new->m_len = WB_BUFBYTES; 942 m_new->m_ext.ext_free = wb_bfree; 943 m_new->m_ext.ext_ref = wb_bfree; 944 } else { 945 m_new = m; 946 m_new->m_len = m_new->m_pkthdr.len = WB_BUFBYTES; 947 m_new->m_data = m_new->m_ext.ext_buf; 948 } 949 950 m_adj(m_new, sizeof(uint64_t)); 951 952 c->wb_mbuf = m_new; 953 c->wb_ptr->wb_data = vtophys(mtod(m_new, caddr_t)); 954 c->wb_ptr->wb_ctl = WB_RXCTL_RLINK | 1536; 955 c->wb_ptr->wb_status = WB_RXSTAT; 956 957 return(0); 958 } 959 960 /* 961 * A frame has been uploaded: pass the resulting mbuf chain up to 962 * the higher level protocols. 963 */ 964 static void 965 wb_rxeof(struct wb_softc *sc) 966 { 967 struct ifnet *ifp = &sc->arpcom.ac_if; 968 struct mbuf *m, *m0; 969 struct wb_chain_onefrag *cur_rx; 970 int total_len = 0; 971 uint32_t rxstat; 972 973 for (;;) { 974 rxstat = sc->wb_cdata.wb_rx_head->wb_ptr->wb_status; 975 if ((rxstat & WB_RXSTAT_OWN) == 0) 976 break; 977 978 cur_rx = sc->wb_cdata.wb_rx_head; 979 sc->wb_cdata.wb_rx_head = cur_rx->wb_nextdesc; 980 981 m = cur_rx->wb_mbuf; 982 983 if ((rxstat & WB_RXSTAT_MIIERR) || 984 (WB_RXBYTES(cur_rx->wb_ptr->wb_status) < WB_MIN_FRAMELEN) || 985 (WB_RXBYTES(cur_rx->wb_ptr->wb_status) > 1536) || 986 (rxstat & WB_RXSTAT_LASTFRAG) == 0|| 987 (rxstat & WB_RXSTAT_RXCMP) == 0) { 988 ifp->if_ierrors++; 989 wb_newbuf(sc, cur_rx, m); 990 if_printf(ifp, "receiver babbling: possible chip " 991 "bug, forcing reset\n"); 992 wb_fixmedia(sc); 993 wb_reset(sc); 994 wb_init(sc); 995 return; 996 } 997 998 if (rxstat & WB_RXSTAT_RXERR) { 999 ifp->if_ierrors++; 1000 wb_newbuf(sc, cur_rx, m); 1001 break; 1002 } 1003 1004 /* No errors; receive the packet. */ 1005 total_len = WB_RXBYTES(cur_rx->wb_ptr->wb_status); 1006 1007 /* 1008 * XXX The Winbond chip includes the CRC with every 1009 * received frame, and there's no way to turn this 1010 * behavior off (at least, I can't find anything in 1011 * the manual that explains how to do it) so we have 1012 * to trim off the CRC manually. 1013 */ 1014 total_len -= ETHER_CRC_LEN; 1015 1016 m0 = m_devget(mtod(m, char *) - ETHER_ALIGN, 1017 total_len + ETHER_ALIGN, 0, ifp, NULL); 1018 wb_newbuf(sc, cur_rx, m); 1019 if (m0 == NULL) { 1020 ifp->if_ierrors++; 1021 break; 1022 } 1023 m_adj(m0, ETHER_ALIGN); 1024 m = m0; 1025 1026 ifp->if_ipackets++; 1027 ifp->if_input(ifp, m); 1028 } 1029 } 1030 1031 static void 1032 wb_rxeoc(struct wb_softc *sc) 1033 { 1034 wb_rxeof(sc); 1035 1036 WB_CLRBIT(sc, WB_NETCFG, WB_NETCFG_RX_ON); 1037 CSR_WRITE_4(sc, WB_RXADDR, vtophys(&sc->wb_ldata->wb_rx_list[0])); 1038 WB_SETBIT(sc, WB_NETCFG, WB_NETCFG_RX_ON); 1039 if (CSR_READ_4(sc, WB_ISR) & WB_RXSTATE_SUSPEND) 1040 CSR_WRITE_4(sc, WB_RXSTART, 0xFFFFFFFF); 1041 } 1042 1043 /* 1044 * A frame was downloaded to the chip. It's safe for us to clean up 1045 * the list buffers. 1046 */ 1047 static void 1048 wb_txeof(struct wb_softc *sc) 1049 { 1050 struct ifnet *ifp = &sc->arpcom.ac_if; 1051 struct wb_chain *cur_tx; 1052 1053 /* Clear the timeout timer. */ 1054 ifp->if_timer = 0; 1055 1056 if (sc->wb_cdata.wb_tx_head == NULL) 1057 return; 1058 1059 /* 1060 * Go through our tx list and free mbufs for those 1061 * frames that have been transmitted. 1062 */ 1063 while(sc->wb_cdata.wb_tx_head->wb_mbuf != NULL) { 1064 uint32_t txstat; 1065 1066 cur_tx = sc->wb_cdata.wb_tx_head; 1067 txstat = WB_TXSTATUS(cur_tx); 1068 1069 if ((txstat & WB_TXSTAT_OWN) || txstat == WB_UNSENT) 1070 break; 1071 1072 if (txstat & WB_TXSTAT_TXERR) { 1073 ifp->if_oerrors++; 1074 if (txstat & WB_TXSTAT_ABORT) 1075 ifp->if_collisions++; 1076 if (txstat & WB_TXSTAT_LATECOLL) 1077 ifp->if_collisions++; 1078 } 1079 1080 ifp->if_collisions += (txstat & WB_TXSTAT_COLLCNT) >> 3; 1081 1082 ifp->if_opackets++; 1083 m_freem(cur_tx->wb_mbuf); 1084 cur_tx->wb_mbuf = NULL; 1085 1086 if (sc->wb_cdata.wb_tx_head == sc->wb_cdata.wb_tx_tail) { 1087 sc->wb_cdata.wb_tx_head = NULL; 1088 sc->wb_cdata.wb_tx_tail = NULL; 1089 break; 1090 } 1091 1092 sc->wb_cdata.wb_tx_head = cur_tx->wb_nextdesc; 1093 } 1094 } 1095 1096 /* 1097 * TX 'end of channel' interrupt handler. 1098 */ 1099 static void 1100 wb_txeoc(struct wb_softc *sc) 1101 { 1102 struct ifnet *ifp = &sc->arpcom.ac_if; 1103 1104 ifp->if_timer = 0; 1105 1106 if (sc->wb_cdata.wb_tx_head == NULL) { 1107 ifp->if_flags &= ~IFF_OACTIVE; 1108 sc->wb_cdata.wb_tx_tail = NULL; 1109 } else if (WB_TXOWN(sc->wb_cdata.wb_tx_head) == WB_UNSENT) { 1110 WB_TXOWN(sc->wb_cdata.wb_tx_head) = WB_TXSTAT_OWN; 1111 ifp->if_timer = 5; 1112 CSR_WRITE_4(sc, WB_TXSTART, 0xFFFFFFFF); 1113 } 1114 } 1115 1116 static void 1117 wb_intr(void *arg) 1118 { 1119 struct wb_softc *sc = arg; 1120 struct ifnet *ifp = &sc->arpcom.ac_if; 1121 uint32_t status; 1122 1123 if ((ifp->if_flags & IFF_UP) == 0) 1124 return; 1125 1126 /* Disable interrupts. */ 1127 CSR_WRITE_4(sc, WB_IMR, 0x00000000); 1128 1129 for (;;) { 1130 status = CSR_READ_4(sc, WB_ISR); 1131 if (status) 1132 CSR_WRITE_4(sc, WB_ISR, status); 1133 1134 if ((status & WB_INTRS) == 0) 1135 break; 1136 1137 if ((status & WB_ISR_RX_NOBUF) || (status & WB_ISR_RX_ERR)) { 1138 ifp->if_ierrors++; 1139 wb_reset(sc); 1140 if (status & WB_ISR_RX_ERR) 1141 wb_fixmedia(sc); 1142 wb_init(sc); 1143 continue; 1144 } 1145 1146 if (status & WB_ISR_RX_OK) 1147 wb_rxeof(sc); 1148 1149 if (status & WB_ISR_RX_IDLE) 1150 wb_rxeoc(sc); 1151 1152 if (status & WB_ISR_TX_OK) 1153 wb_txeof(sc); 1154 1155 if (status & WB_ISR_TX_NOBUF) 1156 wb_txeoc(sc); 1157 1158 if (status & WB_ISR_TX_IDLE) { 1159 wb_txeof(sc); 1160 if (sc->wb_cdata.wb_tx_head != NULL) { 1161 WB_SETBIT(sc, WB_NETCFG, WB_NETCFG_TX_ON); 1162 CSR_WRITE_4(sc, WB_TXSTART, 0xFFFFFFFF); 1163 } 1164 } 1165 1166 if (status & WB_ISR_TX_UNDERRUN) { 1167 ifp->if_oerrors++; 1168 wb_txeof(sc); 1169 WB_CLRBIT(sc, WB_NETCFG, WB_NETCFG_TX_ON); 1170 /* Jack up TX threshold */ 1171 sc->wb_txthresh += WB_TXTHRESH_CHUNK; 1172 WB_CLRBIT(sc, WB_NETCFG, WB_NETCFG_TX_THRESH); 1173 WB_SETBIT(sc, WB_NETCFG, WB_TXTHRESH(sc->wb_txthresh)); 1174 WB_SETBIT(sc, WB_NETCFG, WB_NETCFG_TX_ON); 1175 } 1176 1177 if (status & WB_ISR_BUS_ERR) { 1178 wb_reset(sc); 1179 wb_init(sc); 1180 } 1181 } 1182 1183 /* Re-enable interrupts. */ 1184 CSR_WRITE_4(sc, WB_IMR, WB_INTRS); 1185 1186 if (!ifq_is_empty(&ifp->if_snd)) 1187 wb_start(ifp); 1188 } 1189 1190 static void 1191 wb_tick(void *xsc) 1192 { 1193 struct wb_softc *sc = xsc; 1194 struct ifnet *ifp = &sc->arpcom.ac_if; 1195 struct mii_data *mii = device_get_softc(sc->wb_miibus); 1196 1197 lwkt_serialize_enter(ifp->if_serializer); 1198 mii_tick(mii); 1199 callout_reset(&sc->wb_stat_timer, hz, wb_tick, sc); 1200 lwkt_serialize_exit(ifp->if_serializer); 1201 } 1202 1203 /* 1204 * Encapsulate an mbuf chain in a descriptor by coupling the mbuf data 1205 * pointers to the fragment pointers. 1206 */ 1207 static int 1208 wb_encap(struct wb_softc *sc, struct wb_chain *c, struct mbuf *m_head) 1209 { 1210 struct wb_desc *f = NULL; 1211 struct mbuf *m; 1212 int frag, total_len; 1213 1214 /* 1215 * Start packing the mbufs in this chain into 1216 * the fragment pointers. Stop when we run out 1217 * of fragments or hit the end of the mbuf chain. 1218 */ 1219 total_len = 0; 1220 1221 for (m = m_head, frag = 0; m != NULL; m = m->m_next) { 1222 if (m->m_len != 0) { 1223 if (frag == WB_MAXFRAGS) 1224 break; 1225 total_len += m->m_len; 1226 f = &c->wb_ptr->wb_frag[frag]; 1227 f->wb_ctl = WB_TXCTL_TLINK | m->m_len; 1228 if (frag == 0) { 1229 f->wb_ctl |= WB_TXCTL_FIRSTFRAG; 1230 f->wb_status = 0; 1231 } else { 1232 f->wb_status = WB_TXSTAT_OWN; 1233 } 1234 f->wb_next = vtophys(&c->wb_ptr->wb_frag[frag + 1]); 1235 f->wb_data = vtophys(mtod(m, vm_offset_t)); 1236 frag++; 1237 } 1238 } 1239 1240 /* 1241 * Handle special case: we used up all 16 fragments, 1242 * but we have more mbufs left in the chain. Copy the 1243 * data into an mbuf cluster. Note that we don't 1244 * bother clearing the values in the other fragment 1245 * pointers/counters; it wouldn't gain us anything, 1246 * and would waste cycles. 1247 */ 1248 if (m != NULL) { 1249 struct mbuf *m_new = NULL; 1250 1251 MGETHDR(m_new, MB_DONTWAIT, MT_DATA); 1252 if (m_new == NULL) 1253 return(1); 1254 if (m_head->m_pkthdr.len > MHLEN) { 1255 MCLGET(m_new, MB_DONTWAIT); 1256 if ((m_new->m_flags & M_EXT) == 0) { 1257 m_freem(m_new); 1258 return(1); 1259 } 1260 } 1261 m_copydata(m_head, 0, m_head->m_pkthdr.len, 1262 mtod(m_new, caddr_t)); 1263 m_new->m_pkthdr.len = m_new->m_len = m_head->m_pkthdr.len; 1264 m_freem(m_head); 1265 m_head = m_new; 1266 f = &c->wb_ptr->wb_frag[0]; 1267 f->wb_status = 0; 1268 f->wb_data = vtophys(mtod(m_new, caddr_t)); 1269 f->wb_ctl = total_len = m_new->m_len; 1270 f->wb_ctl |= WB_TXCTL_TLINK|WB_TXCTL_FIRSTFRAG; 1271 frag = 1; 1272 } 1273 1274 if (total_len < WB_MIN_FRAMELEN) { 1275 f = &c->wb_ptr->wb_frag[frag]; 1276 f->wb_ctl = WB_MIN_FRAMELEN - total_len; 1277 f->wb_data = vtophys(&sc->wb_cdata.wb_pad); 1278 f->wb_ctl |= WB_TXCTL_TLINK; 1279 f->wb_status = WB_TXSTAT_OWN; 1280 frag++; 1281 } 1282 1283 c->wb_mbuf = m_head; 1284 c->wb_lastdesc = frag - 1; 1285 WB_TXCTL(c) |= WB_TXCTL_LASTFRAG; 1286 WB_TXNEXT(c) = vtophys(&c->wb_nextdesc->wb_ptr->wb_frag[0]); 1287 1288 return(0); 1289 } 1290 1291 /* 1292 * Main transmit routine. To avoid having to do mbuf copies, we put pointers 1293 * to the mbuf data regions directly in the transmit lists. We also save a 1294 * copy of the pointers since the transmit list fragment pointers are 1295 * physical addresses. 1296 */ 1297 static void 1298 wb_start(struct ifnet *ifp) 1299 { 1300 struct wb_softc *sc = ifp->if_softc; 1301 struct mbuf *m_head = NULL; 1302 struct wb_chain *cur_tx = NULL, *start_tx; 1303 1304 /* 1305 * Check for an available queue slot. If there are none, 1306 * punt. 1307 */ 1308 if (sc->wb_cdata.wb_tx_free->wb_mbuf != NULL) { 1309 ifp->if_flags |= IFF_OACTIVE; 1310 return; 1311 } 1312 1313 start_tx = sc->wb_cdata.wb_tx_free; 1314 1315 while (sc->wb_cdata.wb_tx_free->wb_mbuf == NULL) { 1316 m_head = ifq_dequeue(&ifp->if_snd, NULL); 1317 if (m_head == NULL) 1318 break; 1319 1320 /* Pick a descriptor off the free list. */ 1321 cur_tx = sc->wb_cdata.wb_tx_free; 1322 sc->wb_cdata.wb_tx_free = cur_tx->wb_nextdesc; 1323 1324 /* Pack the data into the descriptor. */ 1325 wb_encap(sc, cur_tx, m_head); 1326 1327 if (cur_tx != start_tx) 1328 WB_TXOWN(cur_tx) = WB_TXSTAT_OWN; 1329 1330 BPF_MTAP(ifp, cur_tx->wb_mbuf); 1331 } 1332 1333 /* 1334 * If there are no packets queued, bail. 1335 */ 1336 if (cur_tx == NULL) 1337 return; 1338 1339 /* 1340 * Place the request for the upload interrupt 1341 * in the last descriptor in the chain. This way, if 1342 * we're chaining several packets at once, we'll only 1343 * get an interupt once for the whole chain rather than 1344 * once for each packet. 1345 */ 1346 WB_TXCTL(cur_tx) |= WB_TXCTL_FINT; 1347 cur_tx->wb_ptr->wb_frag[0].wb_ctl |= WB_TXCTL_FINT; 1348 sc->wb_cdata.wb_tx_tail = cur_tx; 1349 1350 if (sc->wb_cdata.wb_tx_head == NULL) { 1351 sc->wb_cdata.wb_tx_head = start_tx; 1352 WB_TXOWN(start_tx) = WB_TXSTAT_OWN; 1353 CSR_WRITE_4(sc, WB_TXSTART, 0xFFFFFFFF); 1354 } else { 1355 /* 1356 * We need to distinguish between the case where 1357 * the own bit is clear because the chip cleared it 1358 * and where the own bit is clear because we haven't 1359 * set it yet. The magic value WB_UNSET is just some 1360 * ramdomly chosen number which doesn't have the own 1361 * bit set. When we actually transmit the frame, the 1362 * status word will have _only_ the own bit set, so 1363 * the txeoc handler will be able to tell if it needs 1364 * to initiate another transmission to flush out pending 1365 * frames. 1366 */ 1367 WB_TXOWN(start_tx) = WB_UNSENT; 1368 } 1369 1370 /* 1371 * Set a timeout in case the chip goes out to lunch. 1372 */ 1373 ifp->if_timer = 5; 1374 } 1375 1376 static void 1377 wb_init(void *xsc) 1378 { 1379 struct wb_softc *sc = xsc; 1380 struct ifnet *ifp = &sc->arpcom.ac_if; 1381 int i; 1382 struct mii_data *mii; 1383 1384 crit_enter(); 1385 1386 mii = device_get_softc(sc->wb_miibus); 1387 1388 /* 1389 * Cancel pending I/O and free all RX/TX buffers. 1390 */ 1391 wb_stop(sc); 1392 wb_reset(sc); 1393 1394 sc->wb_txthresh = WB_TXTHRESH_INIT; 1395 1396 /* 1397 * Set cache alignment and burst length. 1398 */ 1399 #ifdef foo 1400 CSR_WRITE_4(sc, WB_BUSCTL, WB_BUSCTL_CONFIG); 1401 WB_CLRBIT(sc, WB_NETCFG, WB_NETCFG_TX_THRESH); 1402 WB_SETBIT(sc, WB_NETCFG, WB_TXTHRESH(sc->wb_txthresh)); 1403 #endif 1404 1405 CSR_WRITE_4(sc, WB_BUSCTL, WB_BUSCTL_MUSTBEONE | WB_BUSCTL_ARBITRATION); 1406 WB_SETBIT(sc, WB_BUSCTL, WB_BURSTLEN_16LONG); 1407 switch(sc->wb_cachesize) { 1408 case 32: 1409 WB_SETBIT(sc, WB_BUSCTL, WB_CACHEALIGN_32LONG); 1410 break; 1411 case 16: 1412 WB_SETBIT(sc, WB_BUSCTL, WB_CACHEALIGN_16LONG); 1413 break; 1414 case 8: 1415 WB_SETBIT(sc, WB_BUSCTL, WB_CACHEALIGN_8LONG); 1416 break; 1417 case 0: 1418 default: 1419 WB_SETBIT(sc, WB_BUSCTL, WB_CACHEALIGN_NONE); 1420 break; 1421 } 1422 1423 /* This doesn't tend to work too well at 100Mbps. */ 1424 WB_CLRBIT(sc, WB_NETCFG, WB_NETCFG_TX_EARLY_ON); 1425 1426 /* Init our MAC address */ 1427 for (i = 0; i < ETHER_ADDR_LEN; i++) 1428 CSR_WRITE_1(sc, WB_NODE0 + i, sc->arpcom.ac_enaddr[i]); 1429 1430 /* Init circular RX list. */ 1431 if (wb_list_rx_init(sc) == ENOBUFS) { 1432 if_printf(ifp, "initialization failed: no " 1433 "memory for rx buffers\n"); 1434 wb_stop(sc); 1435 crit_exit(); 1436 return; 1437 } 1438 1439 /* Init TX descriptors. */ 1440 wb_list_tx_init(sc); 1441 1442 /* If we want promiscuous mode, set the allframes bit. */ 1443 if (ifp->if_flags & IFF_PROMISC) 1444 WB_SETBIT(sc, WB_NETCFG, WB_NETCFG_RX_ALLPHYS); 1445 else 1446 WB_CLRBIT(sc, WB_NETCFG, WB_NETCFG_RX_ALLPHYS); 1447 1448 /* 1449 * Set capture broadcast bit to capture broadcast frames. 1450 */ 1451 if (ifp->if_flags & IFF_BROADCAST) 1452 WB_SETBIT(sc, WB_NETCFG, WB_NETCFG_RX_BROAD); 1453 else 1454 WB_CLRBIT(sc, WB_NETCFG, WB_NETCFG_RX_BROAD); 1455 1456 /* 1457 * Program the multicast filter, if necessary. 1458 */ 1459 wb_setmulti(sc); 1460 1461 /* 1462 * Load the address of the RX list. 1463 */ 1464 WB_CLRBIT(sc, WB_NETCFG, WB_NETCFG_RX_ON); 1465 CSR_WRITE_4(sc, WB_RXADDR, vtophys(&sc->wb_ldata->wb_rx_list[0])); 1466 1467 /* 1468 * Enable interrupts. 1469 */ 1470 CSR_WRITE_4(sc, WB_IMR, WB_INTRS); 1471 CSR_WRITE_4(sc, WB_ISR, 0xFFFFFFFF); 1472 1473 /* Enable receiver and transmitter. */ 1474 WB_SETBIT(sc, WB_NETCFG, WB_NETCFG_RX_ON); 1475 CSR_WRITE_4(sc, WB_RXSTART, 0xFFFFFFFF); 1476 1477 WB_CLRBIT(sc, WB_NETCFG, WB_NETCFG_TX_ON); 1478 CSR_WRITE_4(sc, WB_TXADDR, vtophys(&sc->wb_ldata->wb_tx_list[0])); 1479 WB_SETBIT(sc, WB_NETCFG, WB_NETCFG_TX_ON); 1480 1481 mii_mediachg(mii); 1482 1483 ifp->if_flags |= IFF_RUNNING; 1484 ifp->if_flags &= ~IFF_OACTIVE; 1485 1486 crit_exit(); 1487 1488 callout_reset(&sc->wb_stat_timer, hz, wb_tick, sc); 1489 } 1490 1491 /* 1492 * Set media options. 1493 */ 1494 static int 1495 wb_ifmedia_upd(struct ifnet *ifp) 1496 { 1497 struct wb_softc *sc = ifp->if_softc; 1498 1499 if (ifp->if_flags & IFF_UP) 1500 wb_init(sc); 1501 1502 return(0); 1503 } 1504 1505 /* 1506 * Report current media status. 1507 */ 1508 static void 1509 wb_ifmedia_sts(struct ifnet *ifp, struct ifmediareq *ifmr) 1510 { 1511 struct wb_softc *sc = ifp->if_softc; 1512 struct mii_data *mii = device_get_softc(sc->wb_miibus); 1513 1514 mii_pollstat(mii); 1515 ifmr->ifm_active = mii->mii_media_active; 1516 ifmr->ifm_status = mii->mii_media_status; 1517 } 1518 1519 static int 1520 wb_ioctl(struct ifnet *ifp, u_long command, caddr_t data, struct ucred *cr) 1521 { 1522 struct wb_softc *sc = ifp->if_softc; 1523 struct mii_data *mii; 1524 struct ifreq *ifr = (struct ifreq *) data; 1525 int error = 0; 1526 1527 crit_enter(); 1528 1529 switch(command) { 1530 case SIOCSIFFLAGS: 1531 if (ifp->if_flags & IFF_UP) 1532 wb_init(sc); 1533 else if (ifp->if_flags & IFF_RUNNING) 1534 wb_stop(sc); 1535 error = 0; 1536 break; 1537 case SIOCADDMULTI: 1538 case SIOCDELMULTI: 1539 wb_setmulti(sc); 1540 error = 0; 1541 break; 1542 case SIOCGIFMEDIA: 1543 case SIOCSIFMEDIA: 1544 mii = device_get_softc(sc->wb_miibus); 1545 error = ifmedia_ioctl(ifp, ifr, &mii->mii_media, command); 1546 break; 1547 default: 1548 error = ether_ioctl(ifp, command, data); 1549 break; 1550 } 1551 1552 crit_exit(); 1553 1554 return(error); 1555 } 1556 1557 static void 1558 wb_watchdog(struct ifnet *ifp) 1559 { 1560 struct wb_softc *sc = ifp->if_softc; 1561 1562 ifp->if_oerrors++; 1563 if_printf(ifp, "watchdog timeout\n"); 1564 #ifdef foo 1565 if ((wb_phy_readreg(sc, PHY_BMSR) & PHY_BMSR_LINKSTAT) == 0) 1566 if_printf(ifp, "no carrier - transceiver cable problem?\n"); 1567 #endif 1568 wb_stop(sc); 1569 wb_reset(sc); 1570 wb_init(sc); 1571 1572 if (!ifq_is_empty(&ifp->if_snd)) 1573 wb_start(ifp); 1574 } 1575 1576 /* 1577 * Stop the adapter and free any mbufs allocated to the 1578 * RX and TX lists. 1579 */ 1580 static void 1581 wb_stop(struct wb_softc *sc) 1582 { 1583 struct ifnet *ifp = &sc->arpcom.ac_if; 1584 int i; 1585 1586 ifp->if_timer = 0; 1587 1588 callout_stop(&sc->wb_stat_timer); 1589 1590 WB_CLRBIT(sc, WB_NETCFG, (WB_NETCFG_RX_ON | WB_NETCFG_TX_ON)); 1591 CSR_WRITE_4(sc, WB_IMR, 0x00000000); 1592 CSR_WRITE_4(sc, WB_TXADDR, 0x00000000); 1593 CSR_WRITE_4(sc, WB_RXADDR, 0x00000000); 1594 1595 /* 1596 * Free data in the RX lists. 1597 */ 1598 for (i = 0; i < WB_RX_LIST_CNT; i++) { 1599 if (sc->wb_cdata.wb_rx_chain[i].wb_mbuf != NULL) { 1600 m_freem(sc->wb_cdata.wb_rx_chain[i].wb_mbuf); 1601 sc->wb_cdata.wb_rx_chain[i].wb_mbuf = NULL; 1602 } 1603 } 1604 bzero(&sc->wb_ldata->wb_rx_list, sizeof(sc->wb_ldata->wb_rx_list)); 1605 1606 /* 1607 * Free the TX list buffers. 1608 */ 1609 for (i = 0; i < WB_TX_LIST_CNT; i++) { 1610 if (sc->wb_cdata.wb_tx_chain[i].wb_mbuf != NULL) { 1611 m_freem(sc->wb_cdata.wb_tx_chain[i].wb_mbuf); 1612 sc->wb_cdata.wb_tx_chain[i].wb_mbuf = NULL; 1613 } 1614 } 1615 1616 bzero(&sc->wb_ldata->wb_tx_list, sizeof(sc->wb_ldata->wb_tx_list)); 1617 1618 ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE); 1619 } 1620 1621 /* 1622 * Stop all chip I/O so that the kernel's probe routines don't 1623 * get confused by errant DMAs when rebooting. 1624 */ 1625 static void 1626 wb_shutdown(device_t dev) 1627 { 1628 struct wb_softc *sc = device_get_softc(dev); 1629 struct ifnet *ifp = &sc->arpcom.ac_if; 1630 1631 lwkt_serialize_enter(ifp->if_serializer); 1632 wb_stop(sc); 1633 lwkt_serialize_exit(ifp->if_serializer); 1634 } 1635 1636