1 /*	tulip_core.c: A DEC 21x4x-family ethernet driver for Linux.
2 
3 	Copyright 2000,2001  The Linux Kernel Team
4 	Written/copyright 1994-2001 by Donald Becker.
5 
6 	This software may be used and distributed according to the terms
7 	of the GNU General Public License, incorporated herein by reference.
8 
9 	Please submit bugs to http://bugzilla.kernel.org/ .
10 */
11 
12 #define pr_fmt(fmt) "tulip: " fmt
13 
14 #define DRV_NAME	"tulip"
15 
16 #include <linux/module.h>
17 #include <linux/pci.h>
18 #include <linux/slab.h>
19 #include "tulip.h"
20 #include <linux/init.h>
21 #include <linux/interrupt.h>
22 #include <linux/etherdevice.h>
23 #include <linux/delay.h>
24 #include <linux/mii.h>
25 #include <linux/crc32.h>
26 #include <asm/unaligned.h>
27 #include <linux/uaccess.h>
28 
29 #ifdef CONFIG_SPARC
30 #include <asm/prom.h>
31 #endif
32 
33 /* A few user-configurable values. */
34 
35 /* Maximum events (Rx packets, etc.) to handle at each interrupt. */
36 static unsigned int max_interrupt_work = 25;
37 
38 #define MAX_UNITS 8
39 /* Used to pass the full-duplex flag, etc. */
40 static int full_duplex[MAX_UNITS];
41 static int options[MAX_UNITS];
42 static int mtu[MAX_UNITS];			/* Jumbo MTU for interfaces. */
43 
44 /*  The possible media types that can be set in options[] are: */
45 const char * const medianame[32] = {
46 	"10baseT", "10base2", "AUI", "100baseTx",
47 	"10baseT-FDX", "100baseTx-FDX", "100baseT4", "100baseFx",
48 	"100baseFx-FDX", "MII 10baseT", "MII 10baseT-FDX", "MII",
49 	"10baseT(forced)", "MII 100baseTx", "MII 100baseTx-FDX", "MII 100baseT4",
50 	"MII 100baseFx-HDX", "MII 100baseFx-FDX", "Home-PNA 1Mbps", "Invalid-19",
51 	"","","","", "","","","",  "","","","Transceiver reset",
52 };
53 
54 /* Set the copy breakpoint for the copy-only-tiny-buffer Rx structure. */
55 #if defined(__alpha__) || defined(__arm__) || defined(__hppa__) || \
56 	defined(CONFIG_SPARC) || defined(__ia64__) || \
57 	defined(__sh__) || defined(__mips__)
58 static int rx_copybreak = 1518;
59 #else
60 static int rx_copybreak = 100;
61 #endif
62 
63 /*
64   Set the bus performance register.
65 	Typical: Set 16 longword cache alignment, no burst limit.
66 	Cache alignment bits 15:14	     Burst length 13:8
67 		0000	No alignment  0x00000000 unlimited		0800 8 longwords
68 		4000	8  longwords		0100 1 longword		1000 16 longwords
69 		8000	16 longwords		0200 2 longwords	2000 32 longwords
70 		C000	32  longwords		0400 4 longwords
71 	Warning: many older 486 systems are broken and require setting 0x00A04800
72 	   8 longword cache alignment, 8 longword burst.
73 	ToDo: Non-Intel setting could be better.
74 */
75 
76 #if defined(__alpha__) || defined(__ia64__)
77 static int csr0 = 0x01A00000 | 0xE000;
78 #elif defined(__i386__) || defined(__powerpc__) || defined(__x86_64__)
79 static int csr0 = 0x01A00000 | 0x8000;
80 #elif defined(CONFIG_SPARC) || defined(__hppa__)
81 /* The UltraSparc PCI controllers will disconnect at every 64-byte
82  * crossing anyways so it makes no sense to tell Tulip to burst
83  * any more than that.
84  */
85 static int csr0 = 0x01A00000 | 0x9000;
86 #elif defined(__arm__) || defined(__sh__)
87 static int csr0 = 0x01A00000 | 0x4800;
88 #elif defined(__mips__)
89 static int csr0 = 0x00200000 | 0x4000;
90 #else
91 static int csr0;
92 #endif
93 
94 /* Operational parameters that usually are not changed. */
95 /* Time in jiffies before concluding the transmitter is hung. */
96 #define TX_TIMEOUT  (4*HZ)
97 
98 
99 MODULE_AUTHOR("The Linux Kernel Team");
100 MODULE_DESCRIPTION("Digital 21*4* Tulip ethernet driver");
101 MODULE_LICENSE("GPL");
102 module_param(tulip_debug, int, 0);
103 module_param(max_interrupt_work, int, 0);
104 module_param(rx_copybreak, int, 0);
105 module_param(csr0, int, 0);
106 module_param_array(options, int, NULL, 0);
107 module_param_array(full_duplex, int, NULL, 0);
108 
109 #ifdef TULIP_DEBUG
110 int tulip_debug = TULIP_DEBUG;
111 #else
112 int tulip_debug = 1;
113 #endif
114 
115 static void tulip_timer(struct timer_list *t)
116 {
117 	struct tulip_private *tp = from_timer(tp, t, timer);
118 	struct net_device *dev = tp->dev;
119 
120 	if (netif_running(dev))
121 		schedule_work(&tp->media_work);
122 }
123 
124 /*
125  * This table use during operation for capabilities and media timer.
126  *
127  * It is indexed via the values in 'enum chips'
128  */
129 
130 const struct tulip_chip_table tulip_tbl[] = {
131   { }, /* placeholder for array, slot unused currently */
132   { }, /* placeholder for array, slot unused currently */
133 
134   /* DC21140 */
135   { "Digital DS21140 Tulip", 128, 0x0001ebef,
136 	HAS_MII | HAS_MEDIA_TABLE | CSR12_IN_SROM | HAS_PCI_MWI, tulip_timer,
137 	tulip_media_task },
138 
139   /* DC21142, DC21143 */
140   { "Digital DS21142/43 Tulip", 128, 0x0801fbff,
141 	HAS_MII | HAS_MEDIA_TABLE | ALWAYS_CHECK_MII | HAS_ACPI | HAS_NWAY
142 	| HAS_INTR_MITIGATION | HAS_PCI_MWI, tulip_timer, t21142_media_task },
143 
144   /* LC82C168 */
145   { "Lite-On 82c168 PNIC", 256, 0x0001fbef,
146 	HAS_MII | HAS_PNICNWAY, pnic_timer, },
147 
148   /* MX98713 */
149   { "Macronix 98713 PMAC", 128, 0x0001ebef,
150 	HAS_MII | HAS_MEDIA_TABLE | CSR12_IN_SROM, mxic_timer, },
151 
152   /* MX98715 */
153   { "Macronix 98715 PMAC", 256, 0x0001ebef,
154 	HAS_MEDIA_TABLE, mxic_timer, },
155 
156   /* MX98725 */
157   { "Macronix 98725 PMAC", 256, 0x0001ebef,
158 	HAS_MEDIA_TABLE, mxic_timer, },
159 
160   /* AX88140 */
161   { "ASIX AX88140", 128, 0x0001fbff,
162 	HAS_MII | HAS_MEDIA_TABLE | CSR12_IN_SROM | MC_HASH_ONLY
163 	| IS_ASIX, tulip_timer, tulip_media_task },
164 
165   /* PNIC2 */
166   { "Lite-On PNIC-II", 256, 0x0801fbff,
167 	HAS_MII | HAS_NWAY | HAS_8023X | HAS_PCI_MWI, pnic2_timer, },
168 
169   /* COMET */
170   { "ADMtek Comet", 256, 0x0001abef,
171 	HAS_MII | MC_HASH_ONLY | COMET_MAC_ADDR, comet_timer, },
172 
173   /* COMPEX9881 */
174   { "Compex 9881 PMAC", 128, 0x0001ebef,
175 	HAS_MII | HAS_MEDIA_TABLE | CSR12_IN_SROM, mxic_timer, },
176 
177   /* I21145 */
178   { "Intel DS21145 Tulip", 128, 0x0801fbff,
179 	HAS_MII | HAS_MEDIA_TABLE | ALWAYS_CHECK_MII | HAS_ACPI
180 	| HAS_NWAY | HAS_PCI_MWI, tulip_timer, tulip_media_task },
181 
182   /* DM910X */
183 #ifdef CONFIG_TULIP_DM910X
184   { "Davicom DM9102/DM9102A", 128, 0x0001ebef,
185 	HAS_MII | HAS_MEDIA_TABLE | CSR12_IN_SROM | HAS_ACPI,
186 	tulip_timer, tulip_media_task },
187 #else
188   { NULL },
189 #endif
190 
191   /* RS7112 */
192   { "Conexant LANfinity", 256, 0x0001ebef,
193 	HAS_MII | HAS_ACPI, tulip_timer, tulip_media_task },
194 
195 };
196 
197 
198 static const struct pci_device_id tulip_pci_tbl[] = {
199 	{ 0x1011, 0x0009, PCI_ANY_ID, PCI_ANY_ID, 0, 0, DC21140 },
200 	{ 0x1011, 0x0019, PCI_ANY_ID, PCI_ANY_ID, 0, 0, DC21143 },
201 	{ 0x11AD, 0x0002, PCI_ANY_ID, PCI_ANY_ID, 0, 0, LC82C168 },
202 	{ 0x10d9, 0x0512, PCI_ANY_ID, PCI_ANY_ID, 0, 0, MX98713 },
203 	{ 0x10d9, 0x0531, PCI_ANY_ID, PCI_ANY_ID, 0, 0, MX98715 },
204 /*	{ 0x10d9, 0x0531, PCI_ANY_ID, PCI_ANY_ID, 0, 0, MX98725 },*/
205 	{ 0x125B, 0x1400, PCI_ANY_ID, PCI_ANY_ID, 0, 0, AX88140 },
206 	{ 0x11AD, 0xc115, PCI_ANY_ID, PCI_ANY_ID, 0, 0, PNIC2 },
207 	{ 0x1317, 0x0981, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
208 	{ 0x1317, 0x0985, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
209 	{ 0x1317, 0x1985, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
210 	{ 0x1317, 0x9511, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
211 	{ 0x13D1, 0xAB02, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
212 	{ 0x13D1, 0xAB03, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
213 	{ 0x13D1, 0xAB08, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
214 	{ 0x104A, 0x0981, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
215 	{ 0x104A, 0x2774, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
216 	{ 0x1259, 0xa120, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
217 	{ 0x11F6, 0x9881, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMPEX9881 },
218 	{ 0x8086, 0x0039, PCI_ANY_ID, PCI_ANY_ID, 0, 0, I21145 },
219 #ifdef CONFIG_TULIP_DM910X
220 	{ 0x1282, 0x9100, PCI_ANY_ID, PCI_ANY_ID, 0, 0, DM910X },
221 	{ 0x1282, 0x9102, PCI_ANY_ID, PCI_ANY_ID, 0, 0, DM910X },
222 #endif
223 	{ 0x1113, 0x1216, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
224 	{ 0x1113, 0x1217, PCI_ANY_ID, PCI_ANY_ID, 0, 0, MX98715 },
225 	{ 0x1113, 0x9511, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
226 	{ 0x1186, 0x1541, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
227 	{ 0x1186, 0x1561, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
228 	{ 0x1186, 0x1591, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
229 	{ 0x14f1, 0x1803, PCI_ANY_ID, PCI_ANY_ID, 0, 0, CONEXANT },
230 	{ 0x1626, 0x8410, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
231 	{ 0x1737, 0xAB09, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
232 	{ 0x1737, 0xAB08, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
233 	{ 0x17B3, 0xAB08, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
234 	{ 0x10b7, 0x9300, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET }, /* 3Com 3CSOHO100B-TX */
235 	{ 0x14ea, 0xab08, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET }, /* Planex FNW-3602-TX */
236 	{ 0x1414, 0x0001, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET }, /* Microsoft MN-120 */
237 	{ 0x1414, 0x0002, PCI_ANY_ID, PCI_ANY_ID, 0, 0, COMET },
238 	{ } /* terminate list */
239 };
240 MODULE_DEVICE_TABLE(pci, tulip_pci_tbl);
241 
242 
243 /* A full-duplex map for media types. */
244 const char tulip_media_cap[32] =
245 {0,0,0,16,  3,19,16,24,  27,4,7,5, 0,20,23,20,  28,31,0,0, };
246 
247 static void tulip_tx_timeout(struct net_device *dev, unsigned int txqueue);
248 static void tulip_init_ring(struct net_device *dev);
249 static void tulip_free_ring(struct net_device *dev);
250 static netdev_tx_t tulip_start_xmit(struct sk_buff *skb,
251 					  struct net_device *dev);
252 static int tulip_open(struct net_device *dev);
253 static int tulip_close(struct net_device *dev);
254 static void tulip_up(struct net_device *dev);
255 static void tulip_down(struct net_device *dev);
256 static struct net_device_stats *tulip_get_stats(struct net_device *dev);
257 static int private_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
258 static void set_rx_mode(struct net_device *dev);
259 static void tulip_set_wolopts(struct pci_dev *pdev, u32 wolopts);
260 #ifdef CONFIG_NET_POLL_CONTROLLER
261 static void poll_tulip(struct net_device *dev);
262 #endif
263 
264 static void tulip_set_power_state (struct tulip_private *tp,
265 				   int sleep, int snooze)
266 {
267 	if (tp->flags & HAS_ACPI) {
268 		u32 tmp, newtmp;
269 		pci_read_config_dword (tp->pdev, CFDD, &tmp);
270 		newtmp = tmp & ~(CFDD_Sleep | CFDD_Snooze);
271 		if (sleep)
272 			newtmp |= CFDD_Sleep;
273 		else if (snooze)
274 			newtmp |= CFDD_Snooze;
275 		if (tmp != newtmp)
276 			pci_write_config_dword (tp->pdev, CFDD, newtmp);
277 	}
278 
279 }
280 
281 
282 static void tulip_up(struct net_device *dev)
283 {
284 	struct tulip_private *tp = netdev_priv(dev);
285 	void __iomem *ioaddr = tp->base_addr;
286 	int next_tick = 3*HZ;
287 	u32 reg;
288 	int i;
289 
290 #ifdef CONFIG_TULIP_NAPI
291 	napi_enable(&tp->napi);
292 #endif
293 
294 	/* Wake the chip from sleep/snooze mode. */
295 	tulip_set_power_state (tp, 0, 0);
296 
297 	/* Disable all WOL events */
298 	pci_enable_wake(tp->pdev, PCI_D3hot, 0);
299 	pci_enable_wake(tp->pdev, PCI_D3cold, 0);
300 	tulip_set_wolopts(tp->pdev, 0);
301 
302 	/* On some chip revs we must set the MII/SYM port before the reset!? */
303 	if (tp->mii_cnt  ||  (tp->mtable  &&  tp->mtable->has_mii))
304 		iowrite32(0x00040000, ioaddr + CSR6);
305 
306 	/* Reset the chip, holding bit 0 set at least 50 PCI cycles. */
307 	iowrite32(0x00000001, ioaddr + CSR0);
308 	pci_read_config_dword(tp->pdev, PCI_COMMAND, &reg);  /* flush write */
309 	udelay(100);
310 
311 	/* Deassert reset.
312 	   Wait the specified 50 PCI cycles after a reset by initializing
313 	   Tx and Rx queues and the address filter list. */
314 	iowrite32(tp->csr0, ioaddr + CSR0);
315 	pci_read_config_dword(tp->pdev, PCI_COMMAND, &reg);  /* flush write */
316 	udelay(100);
317 
318 	if (tulip_debug > 1)
319 		netdev_dbg(dev, "tulip_up(), irq==%d\n", tp->pdev->irq);
320 
321 	iowrite32(tp->rx_ring_dma, ioaddr + CSR3);
322 	iowrite32(tp->tx_ring_dma, ioaddr + CSR4);
323 	tp->cur_rx = tp->cur_tx = 0;
324 	tp->dirty_rx = tp->dirty_tx = 0;
325 
326 	if (tp->flags & MC_HASH_ONLY) {
327 		u32 addr_low = get_unaligned_le32(dev->dev_addr);
328 		u32 addr_high = get_unaligned_le16(dev->dev_addr + 4);
329 		if (tp->chip_id == AX88140) {
330 			iowrite32(0, ioaddr + CSR13);
331 			iowrite32(addr_low,  ioaddr + CSR14);
332 			iowrite32(1, ioaddr + CSR13);
333 			iowrite32(addr_high, ioaddr + CSR14);
334 		} else if (tp->flags & COMET_MAC_ADDR) {
335 			iowrite32(addr_low,  ioaddr + 0xA4);
336 			iowrite32(addr_high, ioaddr + 0xA8);
337 			iowrite32(0, ioaddr + CSR27);
338 			iowrite32(0, ioaddr + CSR28);
339 		}
340 	} else {
341 		/* This is set_rx_mode(), but without starting the transmitter. */
342 		const u16 *eaddrs = (const u16 *)dev->dev_addr;
343 		u16 *setup_frm = &tp->setup_frame[15*6];
344 		dma_addr_t mapping;
345 
346 		/* 21140 bug: you must add the broadcast address. */
347 		memset(tp->setup_frame, 0xff, sizeof(tp->setup_frame));
348 		/* Fill the final entry of the table with our physical address. */
349 		*setup_frm++ = eaddrs[0]; *setup_frm++ = eaddrs[0];
350 		*setup_frm++ = eaddrs[1]; *setup_frm++ = eaddrs[1];
351 		*setup_frm++ = eaddrs[2]; *setup_frm++ = eaddrs[2];
352 
353 		mapping = dma_map_single(&tp->pdev->dev, tp->setup_frame,
354 					 sizeof(tp->setup_frame),
355 					 DMA_TO_DEVICE);
356 		tp->tx_buffers[tp->cur_tx].skb = NULL;
357 		tp->tx_buffers[tp->cur_tx].mapping = mapping;
358 
359 		/* Put the setup frame on the Tx list. */
360 		tp->tx_ring[tp->cur_tx].length = cpu_to_le32(0x08000000 | 192);
361 		tp->tx_ring[tp->cur_tx].buffer1 = cpu_to_le32(mapping);
362 		tp->tx_ring[tp->cur_tx].status = cpu_to_le32(DescOwned);
363 
364 		tp->cur_tx++;
365 	}
366 
367 	tp->saved_if_port = dev->if_port;
368 	if (dev->if_port == 0)
369 		dev->if_port = tp->default_port;
370 
371 	/* Allow selecting a default media. */
372 	i = 0;
373 	if (tp->mtable == NULL)
374 		goto media_picked;
375 	if (dev->if_port) {
376 		int looking_for = tulip_media_cap[dev->if_port] & MediaIsMII ? 11 :
377 			(dev->if_port == 12 ? 0 : dev->if_port);
378 		for (i = 0; i < tp->mtable->leafcount; i++)
379 			if (tp->mtable->mleaf[i].media == looking_for) {
380 				dev_info(&dev->dev,
381 					 "Using user-specified media %s\n",
382 					 medianame[dev->if_port]);
383 				goto media_picked;
384 			}
385 	}
386 	if ((tp->mtable->defaultmedia & 0x0800) == 0) {
387 		int looking_for = tp->mtable->defaultmedia & MEDIA_MASK;
388 		for (i = 0; i < tp->mtable->leafcount; i++)
389 			if (tp->mtable->mleaf[i].media == looking_for) {
390 				dev_info(&dev->dev,
391 					 "Using EEPROM-set media %s\n",
392 					 medianame[looking_for]);
393 				goto media_picked;
394 			}
395 	}
396 	/* Start sensing first non-full-duplex media. */
397 	for (i = tp->mtable->leafcount - 1;
398 		 (tulip_media_cap[tp->mtable->mleaf[i].media] & MediaAlwaysFD) && i > 0; i--)
399 		;
400 media_picked:
401 
402 	tp->csr6 = 0;
403 	tp->cur_index = i;
404 	tp->nwayset = 0;
405 
406 	if (dev->if_port) {
407 		if (tp->chip_id == DC21143  &&
408 		    (tulip_media_cap[dev->if_port] & MediaIsMII)) {
409 			/* We must reset the media CSRs when we force-select MII mode. */
410 			iowrite32(0x0000, ioaddr + CSR13);
411 			iowrite32(0x0000, ioaddr + CSR14);
412 			iowrite32(0x0008, ioaddr + CSR15);
413 		}
414 		tulip_select_media(dev, 1);
415 	} else if (tp->chip_id == DC21142) {
416 		if (tp->mii_cnt) {
417 			tulip_select_media(dev, 1);
418 			if (tulip_debug > 1)
419 				dev_info(&dev->dev,
420 					 "Using MII transceiver %d, status %04x\n",
421 					 tp->phys[0],
422 					 tulip_mdio_read(dev, tp->phys[0], 1));
423 			iowrite32(csr6_mask_defstate, ioaddr + CSR6);
424 			tp->csr6 = csr6_mask_hdcap;
425 			dev->if_port = 11;
426 			iowrite32(0x0000, ioaddr + CSR13);
427 			iowrite32(0x0000, ioaddr + CSR14);
428 		} else
429 			t21142_start_nway(dev);
430 	} else if (tp->chip_id == PNIC2) {
431 	        /* for initial startup advertise 10/100 Full and Half */
432 	        tp->sym_advertise = 0x01E0;
433                 /* enable autonegotiate end interrupt */
434 	        iowrite32(ioread32(ioaddr+CSR5)| 0x00008010, ioaddr + CSR5);
435 	        iowrite32(ioread32(ioaddr+CSR7)| 0x00008010, ioaddr + CSR7);
436 		pnic2_start_nway(dev);
437 	} else if (tp->chip_id == LC82C168  &&  ! tp->medialock) {
438 		if (tp->mii_cnt) {
439 			dev->if_port = 11;
440 			tp->csr6 = 0x814C0000 | (tp->full_duplex ? 0x0200 : 0);
441 			iowrite32(0x0001, ioaddr + CSR15);
442 		} else if (ioread32(ioaddr + CSR5) & TPLnkPass)
443 			pnic_do_nway(dev);
444 		else {
445 			/* Start with 10mbps to do autonegotiation. */
446 			iowrite32(0x32, ioaddr + CSR12);
447 			tp->csr6 = 0x00420000;
448 			iowrite32(0x0001B078, ioaddr + 0xB8);
449 			iowrite32(0x0201B078, ioaddr + 0xB8);
450 			next_tick = 1*HZ;
451 		}
452 	} else if ((tp->chip_id == MX98713 || tp->chip_id == COMPEX9881) &&
453 		   ! tp->medialock) {
454 		dev->if_port = 0;
455 		tp->csr6 = 0x01880000 | (tp->full_duplex ? 0x0200 : 0);
456 		iowrite32(0x0f370000 | ioread16(ioaddr + 0x80), ioaddr + 0x80);
457 	} else if (tp->chip_id == MX98715 || tp->chip_id == MX98725) {
458 		/* Provided by BOLO, Macronix - 12/10/1998. */
459 		dev->if_port = 0;
460 		tp->csr6 = 0x01a80200;
461 		iowrite32(0x0f370000 | ioread16(ioaddr + 0x80), ioaddr + 0x80);
462 		iowrite32(0x11000 | ioread16(ioaddr + 0xa0), ioaddr + 0xa0);
463 	} else if (tp->chip_id == COMET || tp->chip_id == CONEXANT) {
464 		/* Enable automatic Tx underrun recovery. */
465 		iowrite32(ioread32(ioaddr + 0x88) | 1, ioaddr + 0x88);
466 		dev->if_port = tp->mii_cnt ? 11 : 0;
467 		tp->csr6 = 0x00040000;
468 	} else if (tp->chip_id == AX88140) {
469 		tp->csr6 = tp->mii_cnt ? 0x00040100 : 0x00000100;
470 	} else
471 		tulip_select_media(dev, 1);
472 
473 	/* Start the chip's Tx to process setup frame. */
474 	tulip_stop_rxtx(tp);
475 	barrier();
476 	udelay(5);
477 	iowrite32(tp->csr6 | TxOn, ioaddr + CSR6);
478 
479 	/* Enable interrupts by setting the interrupt mask. */
480 	iowrite32(tulip_tbl[tp->chip_id].valid_intrs, ioaddr + CSR5);
481 	iowrite32(tulip_tbl[tp->chip_id].valid_intrs, ioaddr + CSR7);
482 	tulip_start_rxtx(tp);
483 	iowrite32(0, ioaddr + CSR2);		/* Rx poll demand */
484 
485 	if (tulip_debug > 2) {
486 		netdev_dbg(dev, "Done tulip_up(), CSR0 %08x, CSR5 %08x CSR6 %08x\n",
487 			   ioread32(ioaddr + CSR0),
488 			   ioread32(ioaddr + CSR5),
489 			   ioread32(ioaddr + CSR6));
490 	}
491 
492 	/* Set the timer to switch to check for link beat and perhaps switch
493 	   to an alternate media type. */
494 	tp->timer.expires = RUN_AT(next_tick);
495 	add_timer(&tp->timer);
496 #ifdef CONFIG_TULIP_NAPI
497 	timer_setup(&tp->oom_timer, oom_timer, 0);
498 #endif
499 }
500 
501 static int
502 tulip_open(struct net_device *dev)
503 {
504 	struct tulip_private *tp = netdev_priv(dev);
505 	int retval;
506 
507 	tulip_init_ring (dev);
508 
509 	retval = request_irq(tp->pdev->irq, tulip_interrupt, IRQF_SHARED,
510 			     dev->name, dev);
511 	if (retval)
512 		goto free_ring;
513 
514 	tulip_up (dev);
515 
516 	netif_start_queue (dev);
517 
518 	return 0;
519 
520 free_ring:
521 	tulip_free_ring (dev);
522 	return retval;
523 }
524 
525 
526 static void tulip_tx_timeout(struct net_device *dev, unsigned int txqueue)
527 {
528 	struct tulip_private *tp = netdev_priv(dev);
529 	void __iomem *ioaddr = tp->base_addr;
530 	unsigned long flags;
531 
532 	spin_lock_irqsave (&tp->lock, flags);
533 
534 	if (tulip_media_cap[dev->if_port] & MediaIsMII) {
535 		/* Do nothing -- the media monitor should handle this. */
536 		if (tulip_debug > 1)
537 			dev_warn(&dev->dev,
538 				 "Transmit timeout using MII device\n");
539 	} else if (tp->chip_id == DC21140 || tp->chip_id == DC21142 ||
540 		   tp->chip_id == MX98713 || tp->chip_id == COMPEX9881 ||
541 		   tp->chip_id == DM910X) {
542 		dev_warn(&dev->dev,
543 			 "21140 transmit timed out, status %08x, SIA %08x %08x %08x %08x, resetting...\n",
544 			 ioread32(ioaddr + CSR5), ioread32(ioaddr + CSR12),
545 			 ioread32(ioaddr + CSR13), ioread32(ioaddr + CSR14),
546 			 ioread32(ioaddr + CSR15));
547 		tp->timeout_recovery = 1;
548 		schedule_work(&tp->media_work);
549 		goto out_unlock;
550 	} else if (tp->chip_id == PNIC2) {
551 		dev_warn(&dev->dev,
552 			 "PNIC2 transmit timed out, status %08x, CSR6/7 %08x / %08x CSR12 %08x, resetting...\n",
553 			 (int)ioread32(ioaddr + CSR5),
554 			 (int)ioread32(ioaddr + CSR6),
555 			 (int)ioread32(ioaddr + CSR7),
556 			 (int)ioread32(ioaddr + CSR12));
557 	} else {
558 		dev_warn(&dev->dev,
559 			 "Transmit timed out, status %08x, CSR12 %08x, resetting...\n",
560 			 ioread32(ioaddr + CSR5), ioread32(ioaddr + CSR12));
561 		dev->if_port = 0;
562 	}
563 
564 #if defined(way_too_many_messages)
565 	if (tulip_debug > 3) {
566 		int i;
567 		for (i = 0; i < RX_RING_SIZE; i++) {
568 			u8 *buf = (u8 *)(tp->rx_ring[i].buffer1);
569 			int j;
570 			printk(KERN_DEBUG
571 			       "%2d: %08x %08x %08x %08x  %02x %02x %02x\n",
572 			       i,
573 			       (unsigned int)tp->rx_ring[i].status,
574 			       (unsigned int)tp->rx_ring[i].length,
575 			       (unsigned int)tp->rx_ring[i].buffer1,
576 			       (unsigned int)tp->rx_ring[i].buffer2,
577 			       buf[0], buf[1], buf[2]);
578 			for (j = 0; ((j < 1600) && buf[j] != 0xee); j++)
579 				if (j < 100)
580 					pr_cont(" %02x", buf[j]);
581 			pr_cont(" j=%d\n", j);
582 		}
583 		printk(KERN_DEBUG "  Rx ring %p: ", tp->rx_ring);
584 		for (i = 0; i < RX_RING_SIZE; i++)
585 			pr_cont(" %08x", (unsigned int)tp->rx_ring[i].status);
586 		printk(KERN_DEBUG "  Tx ring %p: ", tp->tx_ring);
587 		for (i = 0; i < TX_RING_SIZE; i++)
588 			pr_cont(" %08x", (unsigned int)tp->tx_ring[i].status);
589 		pr_cont("\n");
590 	}
591 #endif
592 
593 	tulip_tx_timeout_complete(tp, ioaddr);
594 
595 out_unlock:
596 	spin_unlock_irqrestore (&tp->lock, flags);
597 	netif_trans_update(dev); /* prevent tx timeout */
598 	netif_wake_queue (dev);
599 }
600 
601 
602 /* Initialize the Rx and Tx rings, along with various 'dev' bits. */
603 static void tulip_init_ring(struct net_device *dev)
604 {
605 	struct tulip_private *tp = netdev_priv(dev);
606 	int i;
607 
608 	tp->susp_rx = 0;
609 	tp->ttimer = 0;
610 	tp->nir = 0;
611 
612 	for (i = 0; i < RX_RING_SIZE; i++) {
613 		tp->rx_ring[i].status = 0x00000000;
614 		tp->rx_ring[i].length = cpu_to_le32(PKT_BUF_SZ);
615 		tp->rx_ring[i].buffer2 = cpu_to_le32(tp->rx_ring_dma + sizeof(struct tulip_rx_desc) * (i + 1));
616 		tp->rx_buffers[i].skb = NULL;
617 		tp->rx_buffers[i].mapping = 0;
618 	}
619 	/* Mark the last entry as wrapping the ring. */
620 	tp->rx_ring[i-1].length = cpu_to_le32(PKT_BUF_SZ | DESC_RING_WRAP);
621 	tp->rx_ring[i-1].buffer2 = cpu_to_le32(tp->rx_ring_dma);
622 
623 	for (i = 0; i < RX_RING_SIZE; i++) {
624 		dma_addr_t mapping;
625 
626 		/* Note the receive buffer must be longword aligned.
627 		   netdev_alloc_skb() provides 16 byte alignment.  But do *not*
628 		   use skb_reserve() to align the IP header! */
629 		struct sk_buff *skb = netdev_alloc_skb(dev, PKT_BUF_SZ);
630 		tp->rx_buffers[i].skb = skb;
631 		if (skb == NULL)
632 			break;
633 		mapping = dma_map_single(&tp->pdev->dev, skb->data,
634 					 PKT_BUF_SZ, DMA_FROM_DEVICE);
635 		tp->rx_buffers[i].mapping = mapping;
636 		tp->rx_ring[i].status = cpu_to_le32(DescOwned);	/* Owned by Tulip chip */
637 		tp->rx_ring[i].buffer1 = cpu_to_le32(mapping);
638 	}
639 	tp->dirty_rx = (unsigned int)(i - RX_RING_SIZE);
640 
641 	/* The Tx buffer descriptor is filled in as needed, but we
642 	   do need to clear the ownership bit. */
643 	for (i = 0; i < TX_RING_SIZE; i++) {
644 		tp->tx_buffers[i].skb = NULL;
645 		tp->tx_buffers[i].mapping = 0;
646 		tp->tx_ring[i].status = 0x00000000;
647 		tp->tx_ring[i].buffer2 = cpu_to_le32(tp->tx_ring_dma + sizeof(struct tulip_tx_desc) * (i + 1));
648 	}
649 	tp->tx_ring[i-1].buffer2 = cpu_to_le32(tp->tx_ring_dma);
650 }
651 
652 static netdev_tx_t
653 tulip_start_xmit(struct sk_buff *skb, struct net_device *dev)
654 {
655 	struct tulip_private *tp = netdev_priv(dev);
656 	int entry;
657 	u32 flag;
658 	dma_addr_t mapping;
659 	unsigned long flags;
660 
661 	spin_lock_irqsave(&tp->lock, flags);
662 
663 	/* Calculate the next Tx descriptor entry. */
664 	entry = tp->cur_tx % TX_RING_SIZE;
665 
666 	tp->tx_buffers[entry].skb = skb;
667 	mapping = dma_map_single(&tp->pdev->dev, skb->data, skb->len,
668 				 DMA_TO_DEVICE);
669 	tp->tx_buffers[entry].mapping = mapping;
670 	tp->tx_ring[entry].buffer1 = cpu_to_le32(mapping);
671 
672 	if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE/2) {/* Typical path */
673 		flag = 0x60000000; /* No interrupt */
674 	} else if (tp->cur_tx - tp->dirty_tx == TX_RING_SIZE/2) {
675 		flag = 0xe0000000; /* Tx-done intr. */
676 	} else if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE - 2) {
677 		flag = 0x60000000; /* No Tx-done intr. */
678 	} else {		/* Leave room for set_rx_mode() to fill entries. */
679 		flag = 0xe0000000; /* Tx-done intr. */
680 		netif_stop_queue(dev);
681 	}
682 	if (entry == TX_RING_SIZE-1)
683 		flag = 0xe0000000 | DESC_RING_WRAP;
684 
685 	tp->tx_ring[entry].length = cpu_to_le32(skb->len | flag);
686 	/* if we were using Transmit Automatic Polling, we would need a
687 	 * wmb() here. */
688 	tp->tx_ring[entry].status = cpu_to_le32(DescOwned);
689 	wmb();
690 
691 	tp->cur_tx++;
692 
693 	/* Trigger an immediate transmit demand. */
694 	iowrite32(0, tp->base_addr + CSR1);
695 
696 	spin_unlock_irqrestore(&tp->lock, flags);
697 
698 	return NETDEV_TX_OK;
699 }
700 
701 static void tulip_clean_tx_ring(struct tulip_private *tp)
702 {
703 	unsigned int dirty_tx;
704 
705 	for (dirty_tx = tp->dirty_tx ; tp->cur_tx - dirty_tx > 0;
706 		dirty_tx++) {
707 		int entry = dirty_tx % TX_RING_SIZE;
708 		int status = le32_to_cpu(tp->tx_ring[entry].status);
709 
710 		if (status < 0) {
711 			tp->dev->stats.tx_errors++;	/* It wasn't Txed */
712 			tp->tx_ring[entry].status = 0;
713 		}
714 
715 		/* Check for Tx filter setup frames. */
716 		if (tp->tx_buffers[entry].skb == NULL) {
717 			/* test because dummy frames not mapped */
718 			if (tp->tx_buffers[entry].mapping)
719 				dma_unmap_single(&tp->pdev->dev,
720 						 tp->tx_buffers[entry].mapping,
721 						 sizeof(tp->setup_frame),
722 						 DMA_TO_DEVICE);
723 			continue;
724 		}
725 
726 		dma_unmap_single(&tp->pdev->dev,
727 				 tp->tx_buffers[entry].mapping,
728 				 tp->tx_buffers[entry].skb->len,
729 				 DMA_TO_DEVICE);
730 
731 		/* Free the original skb. */
732 		dev_kfree_skb_irq(tp->tx_buffers[entry].skb);
733 		tp->tx_buffers[entry].skb = NULL;
734 		tp->tx_buffers[entry].mapping = 0;
735 	}
736 }
737 
738 static void tulip_down (struct net_device *dev)
739 {
740 	struct tulip_private *tp = netdev_priv(dev);
741 	void __iomem *ioaddr = tp->base_addr;
742 	unsigned long flags;
743 
744 	cancel_work_sync(&tp->media_work);
745 
746 #ifdef CONFIG_TULIP_NAPI
747 	napi_disable(&tp->napi);
748 #endif
749 
750 	del_timer_sync (&tp->timer);
751 #ifdef CONFIG_TULIP_NAPI
752 	del_timer_sync (&tp->oom_timer);
753 #endif
754 	spin_lock_irqsave (&tp->lock, flags);
755 
756 	/* Disable interrupts by clearing the interrupt mask. */
757 	iowrite32 (0x00000000, ioaddr + CSR7);
758 
759 	/* Stop the Tx and Rx processes. */
760 	tulip_stop_rxtx(tp);
761 
762 	/* prepare receive buffers */
763 	tulip_refill_rx(dev);
764 
765 	/* release any unconsumed transmit buffers */
766 	tulip_clean_tx_ring(tp);
767 
768 	if (ioread32(ioaddr + CSR6) != 0xffffffff)
769 		dev->stats.rx_missed_errors += ioread32(ioaddr + CSR8) & 0xffff;
770 
771 	spin_unlock_irqrestore (&tp->lock, flags);
772 
773 	timer_setup(&tp->timer, tulip_tbl[tp->chip_id].media_timer, 0);
774 
775 	dev->if_port = tp->saved_if_port;
776 
777 	/* Leave the driver in snooze, not sleep, mode. */
778 	tulip_set_power_state (tp, 0, 1);
779 }
780 
781 static void tulip_free_ring (struct net_device *dev)
782 {
783 	struct tulip_private *tp = netdev_priv(dev);
784 	int i;
785 
786 	/* Free all the skbuffs in the Rx queue. */
787 	for (i = 0; i < RX_RING_SIZE; i++) {
788 		struct sk_buff *skb = tp->rx_buffers[i].skb;
789 		dma_addr_t mapping = tp->rx_buffers[i].mapping;
790 
791 		tp->rx_buffers[i].skb = NULL;
792 		tp->rx_buffers[i].mapping = 0;
793 
794 		tp->rx_ring[i].status = 0;	/* Not owned by Tulip chip. */
795 		tp->rx_ring[i].length = 0;
796 		/* An invalid address. */
797 		tp->rx_ring[i].buffer1 = cpu_to_le32(0xBADF00D0);
798 		if (skb) {
799 			dma_unmap_single(&tp->pdev->dev, mapping, PKT_BUF_SZ,
800 					 DMA_FROM_DEVICE);
801 			dev_kfree_skb (skb);
802 		}
803 	}
804 
805 	for (i = 0; i < TX_RING_SIZE; i++) {
806 		struct sk_buff *skb = tp->tx_buffers[i].skb;
807 
808 		if (skb != NULL) {
809 			dma_unmap_single(&tp->pdev->dev,
810 					 tp->tx_buffers[i].mapping, skb->len,
811 					 DMA_TO_DEVICE);
812 			dev_kfree_skb (skb);
813 		}
814 		tp->tx_buffers[i].skb = NULL;
815 		tp->tx_buffers[i].mapping = 0;
816 	}
817 }
818 
819 static int tulip_close (struct net_device *dev)
820 {
821 	struct tulip_private *tp = netdev_priv(dev);
822 	void __iomem *ioaddr = tp->base_addr;
823 
824 	netif_stop_queue (dev);
825 
826 	tulip_down (dev);
827 
828 	if (tulip_debug > 1)
829 		netdev_dbg(dev, "Shutting down ethercard, status was %02x\n",
830 			   ioread32 (ioaddr + CSR5));
831 
832 	free_irq (tp->pdev->irq, dev);
833 
834 	tulip_free_ring (dev);
835 
836 	return 0;
837 }
838 
839 static struct net_device_stats *tulip_get_stats(struct net_device *dev)
840 {
841 	struct tulip_private *tp = netdev_priv(dev);
842 	void __iomem *ioaddr = tp->base_addr;
843 
844 	if (netif_running(dev)) {
845 		unsigned long flags;
846 
847 		spin_lock_irqsave (&tp->lock, flags);
848 
849 		dev->stats.rx_missed_errors += ioread32(ioaddr + CSR8) & 0xffff;
850 
851 		spin_unlock_irqrestore(&tp->lock, flags);
852 	}
853 
854 	return &dev->stats;
855 }
856 
857 
858 static void tulip_get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info)
859 {
860 	struct tulip_private *np = netdev_priv(dev);
861 	strlcpy(info->driver, DRV_NAME, sizeof(info->driver));
862 	strlcpy(info->bus_info, pci_name(np->pdev), sizeof(info->bus_info));
863 }
864 
865 
866 static int tulip_ethtool_set_wol(struct net_device *dev,
867 				 struct ethtool_wolinfo *wolinfo)
868 {
869 	struct tulip_private *tp = netdev_priv(dev);
870 
871 	if (wolinfo->wolopts & (~tp->wolinfo.supported))
872 		   return -EOPNOTSUPP;
873 
874 	tp->wolinfo.wolopts = wolinfo->wolopts;
875 	device_set_wakeup_enable(&tp->pdev->dev, tp->wolinfo.wolopts);
876 	return 0;
877 }
878 
879 static void tulip_ethtool_get_wol(struct net_device *dev,
880 				  struct ethtool_wolinfo *wolinfo)
881 {
882 	struct tulip_private *tp = netdev_priv(dev);
883 
884 	wolinfo->supported = tp->wolinfo.supported;
885 	wolinfo->wolopts = tp->wolinfo.wolopts;
886 	return;
887 }
888 
889 
890 static const struct ethtool_ops ops = {
891 	.get_drvinfo = tulip_get_drvinfo,
892 	.set_wol     = tulip_ethtool_set_wol,
893 	.get_wol     = tulip_ethtool_get_wol,
894 };
895 
896 /* Provide ioctl() calls to examine the MII xcvr state. */
897 static int private_ioctl (struct net_device *dev, struct ifreq *rq, int cmd)
898 {
899 	struct tulip_private *tp = netdev_priv(dev);
900 	void __iomem *ioaddr = tp->base_addr;
901 	struct mii_ioctl_data *data = if_mii(rq);
902 	const unsigned int phy_idx = 0;
903 	int phy = tp->phys[phy_idx] & 0x1f;
904 	unsigned int regnum = data->reg_num;
905 
906 	switch (cmd) {
907 	case SIOCGMIIPHY:		/* Get address of MII PHY in use. */
908 		if (tp->mii_cnt)
909 			data->phy_id = phy;
910 		else if (tp->flags & HAS_NWAY)
911 			data->phy_id = 32;
912 		else if (tp->chip_id == COMET)
913 			data->phy_id = 1;
914 		else
915 			return -ENODEV;
916 		fallthrough;
917 
918 	case SIOCGMIIREG:		/* Read MII PHY register. */
919 		if (data->phy_id == 32 && (tp->flags & HAS_NWAY)) {
920 			int csr12 = ioread32 (ioaddr + CSR12);
921 			int csr14 = ioread32 (ioaddr + CSR14);
922 			switch (regnum) {
923 			case 0:
924                                 if (((csr14<<5) & 0x1000) ||
925                                         (dev->if_port == 5 && tp->nwayset))
926                                         data->val_out = 0x1000;
927                                 else
928                                         data->val_out = (tulip_media_cap[dev->if_port]&MediaIs100 ? 0x2000 : 0)
929                                                 | (tulip_media_cap[dev->if_port]&MediaIsFD ? 0x0100 : 0);
930 				break;
931 			case 1:
932                                 data->val_out =
933 					0x1848 +
934 					((csr12&0x7000) == 0x5000 ? 0x20 : 0) +
935 					((csr12&0x06) == 6 ? 0 : 4);
936                                 data->val_out |= 0x6048;
937 				break;
938 			case 4:
939                                 /* Advertised value, bogus 10baseTx-FD value from CSR6. */
940                                 data->val_out =
941 					((ioread32(ioaddr + CSR6) >> 3) & 0x0040) +
942 					((csr14 >> 1) & 0x20) + 1;
943                                 data->val_out |= ((csr14 >> 9) & 0x03C0);
944 				break;
945 			case 5: data->val_out = tp->lpar; break;
946 			default: data->val_out = 0; break;
947 			}
948 		} else {
949 			data->val_out = tulip_mdio_read (dev, data->phy_id & 0x1f, regnum);
950 		}
951 		return 0;
952 
953 	case SIOCSMIIREG:		/* Write MII PHY register. */
954 		if (regnum & ~0x1f)
955 			return -EINVAL;
956 		if (data->phy_id == phy) {
957 			u16 value = data->val_in;
958 			switch (regnum) {
959 			case 0:	/* Check for autonegotiation on or reset. */
960 				tp->full_duplex_lock = (value & 0x9000) ? 0 : 1;
961 				if (tp->full_duplex_lock)
962 					tp->full_duplex = (value & 0x0100) ? 1 : 0;
963 				break;
964 			case 4:
965 				tp->advertising[phy_idx] =
966 				tp->mii_advertise = data->val_in;
967 				break;
968 			}
969 		}
970 		if (data->phy_id == 32 && (tp->flags & HAS_NWAY)) {
971 			u16 value = data->val_in;
972 			if (regnum == 0) {
973 			  if ((value & 0x1200) == 0x1200) {
974 			    if (tp->chip_id == PNIC2) {
975                                    pnic2_start_nway (dev);
976                             } else {
977 				   t21142_start_nway (dev);
978                             }
979 			  }
980 			} else if (regnum == 4)
981 				tp->sym_advertise = value;
982 		} else {
983 			tulip_mdio_write (dev, data->phy_id & 0x1f, regnum, data->val_in);
984 		}
985 		return 0;
986 	default:
987 		return -EOPNOTSUPP;
988 	}
989 
990 	return -EOPNOTSUPP;
991 }
992 
993 
994 /* Set or clear the multicast filter for this adaptor.
995    Note that we only use exclusion around actually queueing the
996    new frame, not around filling tp->setup_frame.  This is non-deterministic
997    when re-entered but still correct. */
998 
999 static void build_setup_frame_hash(u16 *setup_frm, struct net_device *dev)
1000 {
1001 	struct tulip_private *tp = netdev_priv(dev);
1002 	u16 hash_table[32];
1003 	struct netdev_hw_addr *ha;
1004 	const u16 *eaddrs;
1005 	int i;
1006 
1007 	memset(hash_table, 0, sizeof(hash_table));
1008 	__set_bit_le(255, hash_table);			/* Broadcast entry */
1009 	/* This should work on big-endian machines as well. */
1010 	netdev_for_each_mc_addr(ha, dev) {
1011 		int index = ether_crc_le(ETH_ALEN, ha->addr) & 0x1ff;
1012 
1013 		__set_bit_le(index, hash_table);
1014 	}
1015 	for (i = 0; i < 32; i++) {
1016 		*setup_frm++ = hash_table[i];
1017 		*setup_frm++ = hash_table[i];
1018 	}
1019 	setup_frm = &tp->setup_frame[13*6];
1020 
1021 	/* Fill the final entry with our physical address. */
1022 	eaddrs = (const u16 *)dev->dev_addr;
1023 	*setup_frm++ = eaddrs[0]; *setup_frm++ = eaddrs[0];
1024 	*setup_frm++ = eaddrs[1]; *setup_frm++ = eaddrs[1];
1025 	*setup_frm++ = eaddrs[2]; *setup_frm++ = eaddrs[2];
1026 }
1027 
1028 static void build_setup_frame_perfect(u16 *setup_frm, struct net_device *dev)
1029 {
1030 	struct tulip_private *tp = netdev_priv(dev);
1031 	struct netdev_hw_addr *ha;
1032 	const u16 *eaddrs;
1033 
1034 	/* We have <= 14 addresses so we can use the wonderful
1035 	   16 address perfect filtering of the Tulip. */
1036 	netdev_for_each_mc_addr(ha, dev) {
1037 		eaddrs = (u16 *) ha->addr;
1038 		*setup_frm++ = *eaddrs; *setup_frm++ = *eaddrs++;
1039 		*setup_frm++ = *eaddrs; *setup_frm++ = *eaddrs++;
1040 		*setup_frm++ = *eaddrs; *setup_frm++ = *eaddrs++;
1041 	}
1042 	/* Fill the unused entries with the broadcast address. */
1043 	memset(setup_frm, 0xff, (15 - netdev_mc_count(dev)) * 12);
1044 	setup_frm = &tp->setup_frame[15*6];
1045 
1046 	/* Fill the final entry with our physical address. */
1047 	eaddrs = (const u16 *)dev->dev_addr;
1048 	*setup_frm++ = eaddrs[0]; *setup_frm++ = eaddrs[0];
1049 	*setup_frm++ = eaddrs[1]; *setup_frm++ = eaddrs[1];
1050 	*setup_frm++ = eaddrs[2]; *setup_frm++ = eaddrs[2];
1051 }
1052 
1053 
1054 static void set_rx_mode(struct net_device *dev)
1055 {
1056 	struct tulip_private *tp = netdev_priv(dev);
1057 	void __iomem *ioaddr = tp->base_addr;
1058 	int csr6;
1059 
1060 	csr6 = ioread32(ioaddr + CSR6) & ~0x00D5;
1061 
1062 	tp->csr6 &= ~0x00D5;
1063 	if (dev->flags & IFF_PROMISC) {			/* Set promiscuous. */
1064 		tp->csr6 |= AcceptAllMulticast | AcceptAllPhys;
1065 		csr6 |= AcceptAllMulticast | AcceptAllPhys;
1066 	} else if ((netdev_mc_count(dev) > 1000) ||
1067 		   (dev->flags & IFF_ALLMULTI)) {
1068 		/* Too many to filter well -- accept all multicasts. */
1069 		tp->csr6 |= AcceptAllMulticast;
1070 		csr6 |= AcceptAllMulticast;
1071 	} else	if (tp->flags & MC_HASH_ONLY) {
1072 		/* Some work-alikes have only a 64-entry hash filter table. */
1073 		/* Should verify correctness on big-endian/__powerpc__ */
1074 		struct netdev_hw_addr *ha;
1075 		if (netdev_mc_count(dev) > 64) {
1076 			/* Arbitrary non-effective limit. */
1077 			tp->csr6 |= AcceptAllMulticast;
1078 			csr6 |= AcceptAllMulticast;
1079 		} else {
1080 			u32 mc_filter[2] = {0, 0};		 /* Multicast hash filter */
1081 			int filterbit;
1082 			netdev_for_each_mc_addr(ha, dev) {
1083 				if (tp->flags & COMET_MAC_ADDR)
1084 					filterbit = ether_crc_le(ETH_ALEN,
1085 								 ha->addr);
1086 				else
1087 					filterbit = ether_crc(ETH_ALEN,
1088 							      ha->addr) >> 26;
1089 				filterbit &= 0x3f;
1090 				mc_filter[filterbit >> 5] |= 1 << (filterbit & 31);
1091 				if (tulip_debug > 2)
1092 					dev_info(&dev->dev,
1093 						 "Added filter for %pM  %08x bit %d\n",
1094 						 ha->addr,
1095 						 ether_crc(ETH_ALEN, ha->addr),
1096 						 filterbit);
1097 			}
1098 			if (mc_filter[0] == tp->mc_filter[0]  &&
1099 				mc_filter[1] == tp->mc_filter[1])
1100 				;				/* No change. */
1101 			else if (tp->flags & IS_ASIX) {
1102 				iowrite32(2, ioaddr + CSR13);
1103 				iowrite32(mc_filter[0], ioaddr + CSR14);
1104 				iowrite32(3, ioaddr + CSR13);
1105 				iowrite32(mc_filter[1], ioaddr + CSR14);
1106 			} else if (tp->flags & COMET_MAC_ADDR) {
1107 				iowrite32(mc_filter[0], ioaddr + CSR27);
1108 				iowrite32(mc_filter[1], ioaddr + CSR28);
1109 			}
1110 			tp->mc_filter[0] = mc_filter[0];
1111 			tp->mc_filter[1] = mc_filter[1];
1112 		}
1113 	} else {
1114 		unsigned long flags;
1115 		u32 tx_flags = 0x08000000 | 192;
1116 
1117 		/* Note that only the low-address shortword of setup_frame is valid!
1118 		   The values are doubled for big-endian architectures. */
1119 		if (netdev_mc_count(dev) > 14) {
1120 			/* Must use a multicast hash table. */
1121 			build_setup_frame_hash(tp->setup_frame, dev);
1122 			tx_flags = 0x08400000 | 192;
1123 		} else {
1124 			build_setup_frame_perfect(tp->setup_frame, dev);
1125 		}
1126 
1127 		spin_lock_irqsave(&tp->lock, flags);
1128 
1129 		if (tp->cur_tx - tp->dirty_tx > TX_RING_SIZE - 2) {
1130 			/* Same setup recently queued, we need not add it. */
1131 		} else {
1132 			unsigned int entry;
1133 			int dummy = -1;
1134 
1135 			/* Now add this frame to the Tx list. */
1136 
1137 			entry = tp->cur_tx++ % TX_RING_SIZE;
1138 
1139 			if (entry != 0) {
1140 				/* Avoid a chip errata by prefixing a dummy entry. */
1141 				tp->tx_buffers[entry].skb = NULL;
1142 				tp->tx_buffers[entry].mapping = 0;
1143 				tp->tx_ring[entry].length =
1144 					(entry == TX_RING_SIZE-1) ? cpu_to_le32(DESC_RING_WRAP) : 0;
1145 				tp->tx_ring[entry].buffer1 = 0;
1146 				/* Must set DescOwned later to avoid race with chip */
1147 				dummy = entry;
1148 				entry = tp->cur_tx++ % TX_RING_SIZE;
1149 
1150 			}
1151 
1152 			tp->tx_buffers[entry].skb = NULL;
1153 			tp->tx_buffers[entry].mapping =
1154 				dma_map_single(&tp->pdev->dev,
1155 					       tp->setup_frame,
1156 					       sizeof(tp->setup_frame),
1157 					       DMA_TO_DEVICE);
1158 			/* Put the setup frame on the Tx list. */
1159 			if (entry == TX_RING_SIZE-1)
1160 				tx_flags |= DESC_RING_WRAP;		/* Wrap ring. */
1161 			tp->tx_ring[entry].length = cpu_to_le32(tx_flags);
1162 			tp->tx_ring[entry].buffer1 =
1163 				cpu_to_le32(tp->tx_buffers[entry].mapping);
1164 			tp->tx_ring[entry].status = cpu_to_le32(DescOwned);
1165 			if (dummy >= 0)
1166 				tp->tx_ring[dummy].status = cpu_to_le32(DescOwned);
1167 			if (tp->cur_tx - tp->dirty_tx >= TX_RING_SIZE - 2)
1168 				netif_stop_queue(dev);
1169 
1170 			/* Trigger an immediate transmit demand. */
1171 			iowrite32(0, ioaddr + CSR1);
1172 		}
1173 
1174 		spin_unlock_irqrestore(&tp->lock, flags);
1175 	}
1176 
1177 	iowrite32(csr6, ioaddr + CSR6);
1178 }
1179 
1180 #ifdef CONFIG_TULIP_MWI
1181 static void tulip_mwi_config(struct pci_dev *pdev, struct net_device *dev)
1182 {
1183 	struct tulip_private *tp = netdev_priv(dev);
1184 	u8 cache;
1185 	u16 pci_command;
1186 	u32 csr0;
1187 
1188 	if (tulip_debug > 3)
1189 		netdev_dbg(dev, "tulip_mwi_config()\n");
1190 
1191 	tp->csr0 = csr0 = 0;
1192 
1193 	/* if we have any cache line size at all, we can do MRM and MWI */
1194 	csr0 |= MRM | MWI;
1195 
1196 	/* Enable MWI in the standard PCI command bit.
1197 	 * Check for the case where MWI is desired but not available
1198 	 */
1199 	pci_try_set_mwi(pdev);
1200 
1201 	/* read result from hardware (in case bit refused to enable) */
1202 	pci_read_config_word(pdev, PCI_COMMAND, &pci_command);
1203 	if ((csr0 & MWI) && (!(pci_command & PCI_COMMAND_INVALIDATE)))
1204 		csr0 &= ~MWI;
1205 
1206 	/* if cache line size hardwired to zero, no MWI */
1207 	pci_read_config_byte(pdev, PCI_CACHE_LINE_SIZE, &cache);
1208 	if ((csr0 & MWI) && (cache == 0)) {
1209 		csr0 &= ~MWI;
1210 		pci_clear_mwi(pdev);
1211 	}
1212 
1213 	/* assign per-cacheline-size cache alignment and
1214 	 * burst length values
1215 	 */
1216 	switch (cache) {
1217 	case 8:
1218 		csr0 |= MRL | (1 << CALShift) | (16 << BurstLenShift);
1219 		break;
1220 	case 16:
1221 		csr0 |= MRL | (2 << CALShift) | (16 << BurstLenShift);
1222 		break;
1223 	case 32:
1224 		csr0 |= MRL | (3 << CALShift) | (32 << BurstLenShift);
1225 		break;
1226 	default:
1227 		cache = 0;
1228 		break;
1229 	}
1230 
1231 	/* if we have a good cache line size, we by now have a good
1232 	 * csr0, so save it and exit
1233 	 */
1234 	if (cache)
1235 		goto out;
1236 
1237 	/* we don't have a good csr0 or cache line size, disable MWI */
1238 	if (csr0 & MWI) {
1239 		pci_clear_mwi(pdev);
1240 		csr0 &= ~MWI;
1241 	}
1242 
1243 	/* sane defaults for burst length and cache alignment
1244 	 * originally from de4x5 driver
1245 	 */
1246 	csr0 |= (8 << BurstLenShift) | (1 << CALShift);
1247 
1248 out:
1249 	tp->csr0 = csr0;
1250 	if (tulip_debug > 2)
1251 		netdev_dbg(dev, "MWI config cacheline=%d, csr0=%08x\n",
1252 			   cache, csr0);
1253 }
1254 #endif
1255 
1256 /*
1257  *	Chips that have the MRM/reserved bit quirk and the burst quirk. That
1258  *	is the DM910X and the on chip ULi devices
1259  */
1260 
1261 static int tulip_uli_dm_quirk(struct pci_dev *pdev)
1262 {
1263 	if (pdev->vendor == 0x1282 && pdev->device == 0x9102)
1264 		return 1;
1265 	return 0;
1266 }
1267 
1268 static const struct net_device_ops tulip_netdev_ops = {
1269 	.ndo_open		= tulip_open,
1270 	.ndo_start_xmit		= tulip_start_xmit,
1271 	.ndo_tx_timeout		= tulip_tx_timeout,
1272 	.ndo_stop		= tulip_close,
1273 	.ndo_get_stats		= tulip_get_stats,
1274 	.ndo_eth_ioctl		= private_ioctl,
1275 	.ndo_set_rx_mode	= set_rx_mode,
1276 	.ndo_set_mac_address	= eth_mac_addr,
1277 	.ndo_validate_addr	= eth_validate_addr,
1278 #ifdef CONFIG_NET_POLL_CONTROLLER
1279 	.ndo_poll_controller	 = poll_tulip,
1280 #endif
1281 };
1282 
1283 static const struct pci_device_id early_486_chipsets[] = {
1284 	{ PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82424) },
1285 	{ PCI_DEVICE(PCI_VENDOR_ID_SI, PCI_DEVICE_ID_SI_496) },
1286 	{ },
1287 };
1288 
1289 static int tulip_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
1290 {
1291 	struct tulip_private *tp;
1292 	/* See note below on the multiport cards. */
1293 	static unsigned char last_phys_addr[ETH_ALEN] = {
1294 		0x00, 'L', 'i', 'n', 'u', 'x'
1295 	};
1296 #if defined(__i386__) || defined(__x86_64__)	/* Patch up x86 BIOS bug. */
1297 	static int last_irq;
1298 #endif
1299 	int i, irq;
1300 	unsigned short sum;
1301 	unsigned char *ee_data;
1302 	struct net_device *dev;
1303 	void __iomem *ioaddr;
1304 	static int board_idx = -1;
1305 	int chip_idx = ent->driver_data;
1306 	const char *chip_name = tulip_tbl[chip_idx].chip_name;
1307 	unsigned int eeprom_missing = 0;
1308 	u8 addr[ETH_ALEN] __aligned(2);
1309 	unsigned int force_csr0 = 0;
1310 
1311 	board_idx++;
1312 
1313 	/*
1314 	 *	Lan media wire a tulip chip to a wan interface. Needs a very
1315 	 *	different driver (lmc driver)
1316 	 */
1317 
1318         if (pdev->subsystem_vendor == PCI_VENDOR_ID_LMC) {
1319 		pr_err("skipping LMC card\n");
1320 		return -ENODEV;
1321 	} else if (pdev->subsystem_vendor == PCI_VENDOR_ID_SBE &&
1322 		   (pdev->subsystem_device == PCI_SUBDEVICE_ID_SBE_T3E3 ||
1323 		    pdev->subsystem_device == PCI_SUBDEVICE_ID_SBE_2T3E3_P0 ||
1324 		    pdev->subsystem_device == PCI_SUBDEVICE_ID_SBE_2T3E3_P1)) {
1325 		pr_err("skipping SBE T3E3 port\n");
1326 		return -ENODEV;
1327 	}
1328 
1329 	/*
1330 	 *	DM910x chips should be handled by the dmfe driver, except
1331 	 *	on-board chips on SPARC systems.  Also, early DM9100s need
1332 	 *	software CRC which only the dmfe driver supports.
1333 	 */
1334 
1335 #ifdef CONFIG_TULIP_DM910X
1336 	if (chip_idx == DM910X) {
1337 		struct device_node *dp;
1338 
1339 		if (pdev->vendor == 0x1282 && pdev->device == 0x9100 &&
1340 		    pdev->revision < 0x30) {
1341 			pr_info("skipping early DM9100 with Crc bug (use dmfe)\n");
1342 			return -ENODEV;
1343 		}
1344 
1345 		dp = pci_device_to_OF_node(pdev);
1346 		if (!(dp && of_get_property(dp, "local-mac-address", NULL))) {
1347 			pr_info("skipping DM910x expansion card (use dmfe)\n");
1348 			return -ENODEV;
1349 		}
1350 	}
1351 #endif
1352 
1353 	/*
1354 	 *	Looks for early PCI chipsets where people report hangs
1355 	 *	without the workarounds being on.
1356 	 */
1357 
1358 	/* 1. Intel Saturn. Switch to 8 long words burst, 8 long word cache
1359 	      aligned.  Aries might need this too. The Saturn errata are not
1360 	      pretty reading but thankfully it's an old 486 chipset.
1361 
1362 	   2. The dreaded SiS496 486 chipset. Same workaround as Intel
1363 	      Saturn.
1364 	*/
1365 
1366 	if (pci_dev_present(early_486_chipsets)) {
1367 		csr0 = MRL | MRM | (8 << BurstLenShift) | (1 << CALShift);
1368 		force_csr0 = 1;
1369 	}
1370 
1371 	/* bugfix: the ASIX must have a burst limit or horrible things happen. */
1372 	if (chip_idx == AX88140) {
1373 		if ((csr0 & 0x3f00) == 0)
1374 			csr0 |= 0x2000;
1375 	}
1376 
1377 	/* PNIC doesn't have MWI/MRL/MRM... */
1378 	if (chip_idx == LC82C168)
1379 		csr0 &= ~0xfff10000; /* zero reserved bits 31:20, 16 */
1380 
1381 	/* DM9102A has troubles with MRM & clear reserved bits 24:22, 20, 16, 7:1 */
1382 	if (tulip_uli_dm_quirk(pdev)) {
1383 		csr0 &= ~0x01f100ff;
1384 #if defined(CONFIG_SPARC)
1385                 csr0 = (csr0 & ~0xff00) | 0xe000;
1386 #endif
1387 	}
1388 	/*
1389 	 *	And back to business
1390 	 */
1391 
1392 	i = pci_enable_device(pdev);
1393 	if (i) {
1394 		pr_err("Cannot enable tulip board #%d, aborting\n", board_idx);
1395 		return i;
1396 	}
1397 
1398 	irq = pdev->irq;
1399 
1400 	/* alloc_etherdev ensures aligned and zeroed private structures */
1401 	dev = alloc_etherdev (sizeof (*tp));
1402 	if (!dev) {
1403 		pci_disable_device(pdev);
1404 		return -ENOMEM;
1405 	}
1406 
1407 	SET_NETDEV_DEV(dev, &pdev->dev);
1408 	if (pci_resource_len (pdev, 0) < tulip_tbl[chip_idx].io_size) {
1409 		pr_err("%s: I/O region (0x%llx@0x%llx) too small, aborting\n",
1410 		       pci_name(pdev),
1411 		       (unsigned long long)pci_resource_len (pdev, 0),
1412 		       (unsigned long long)pci_resource_start (pdev, 0));
1413 		goto err_out_free_netdev;
1414 	}
1415 
1416 	/* grab all resources from both PIO and MMIO regions, as we
1417 	 * don't want anyone else messing around with our hardware */
1418 	if (pci_request_regions (pdev, DRV_NAME))
1419 		goto err_out_free_netdev;
1420 
1421 	ioaddr =  pci_iomap(pdev, TULIP_BAR, tulip_tbl[chip_idx].io_size);
1422 
1423 	if (!ioaddr)
1424 		goto err_out_free_res;
1425 
1426 	/*
1427 	 * initialize private data structure 'tp'
1428 	 * it is zeroed and aligned in alloc_etherdev
1429 	 */
1430 	tp = netdev_priv(dev);
1431 	tp->dev = dev;
1432 
1433 	tp->rx_ring = dma_alloc_coherent(&pdev->dev,
1434 					 sizeof(struct tulip_rx_desc) * RX_RING_SIZE +
1435 					 sizeof(struct tulip_tx_desc) * TX_RING_SIZE,
1436 					 &tp->rx_ring_dma, GFP_KERNEL);
1437 	if (!tp->rx_ring)
1438 		goto err_out_mtable;
1439 	tp->tx_ring = (struct tulip_tx_desc *)(tp->rx_ring + RX_RING_SIZE);
1440 	tp->tx_ring_dma = tp->rx_ring_dma + sizeof(struct tulip_rx_desc) * RX_RING_SIZE;
1441 
1442 	tp->chip_id = chip_idx;
1443 	tp->flags = tulip_tbl[chip_idx].flags;
1444 
1445 	tp->wolinfo.supported = 0;
1446 	tp->wolinfo.wolopts = 0;
1447 	/* COMET: Enable power management only for AN983B */
1448 	if (chip_idx == COMET ) {
1449 		u32 sig;
1450 		pci_read_config_dword (pdev, 0x80, &sig);
1451 		if (sig == 0x09811317) {
1452 			tp->flags |= COMET_PM;
1453 			tp->wolinfo.supported = WAKE_PHY | WAKE_MAGIC;
1454 			pr_info("%s: Enabled WOL support for AN983B\n",
1455 				__func__);
1456 		}
1457 	}
1458 	tp->pdev = pdev;
1459 	tp->base_addr = ioaddr;
1460 	tp->revision = pdev->revision;
1461 	tp->csr0 = csr0;
1462 	spin_lock_init(&tp->lock);
1463 	spin_lock_init(&tp->mii_lock);
1464 	timer_setup(&tp->timer, tulip_tbl[tp->chip_id].media_timer, 0);
1465 
1466 	INIT_WORK(&tp->media_work, tulip_tbl[tp->chip_id].media_task);
1467 
1468 #ifdef CONFIG_TULIP_MWI
1469 	if (!force_csr0 && (tp->flags & HAS_PCI_MWI))
1470 		tulip_mwi_config (pdev, dev);
1471 #endif
1472 
1473 	/* Stop the chip's Tx and Rx processes. */
1474 	tulip_stop_rxtx(tp);
1475 
1476 	pci_set_master(pdev);
1477 
1478 #ifdef CONFIG_GSC
1479 	if (pdev->subsystem_vendor == PCI_VENDOR_ID_HP) {
1480 		switch (pdev->subsystem_device) {
1481 		default:
1482 			break;
1483 		case 0x1061:
1484 		case 0x1062:
1485 		case 0x1063:
1486 		case 0x1098:
1487 		case 0x1099:
1488 		case 0x10EE:
1489 			tp->flags |= HAS_SWAPPED_SEEPROM | NEEDS_FAKE_MEDIA_TABLE;
1490 			chip_name = "GSC DS21140 Tulip";
1491 		}
1492 	}
1493 #endif
1494 
1495 	/* Clear the missed-packet counter. */
1496 	ioread32(ioaddr + CSR8);
1497 
1498 	/* The station address ROM is read byte serially.  The register must
1499 	   be polled, waiting for the value to be read bit serially from the
1500 	   EEPROM.
1501 	   */
1502 	ee_data = tp->eeprom;
1503 	memset(ee_data, 0, sizeof(tp->eeprom));
1504 	sum = 0;
1505 	if (chip_idx == LC82C168) {
1506 		for (i = 0; i < 3; i++) {
1507 			int value, boguscnt = 100000;
1508 			iowrite32(0x600 | i, ioaddr + 0x98);
1509 			do {
1510 				value = ioread32(ioaddr + CSR9);
1511 			} while (value < 0  && --boguscnt > 0);
1512 			put_unaligned_le16(value, ((__le16 *)addr) + i);
1513 			sum += value & 0xffff;
1514 		}
1515 		eth_hw_addr_set(dev, addr);
1516 	} else if (chip_idx == COMET) {
1517 		/* No need to read the EEPROM. */
1518 		put_unaligned_le32(ioread32(ioaddr + 0xA4), addr);
1519 		put_unaligned_le16(ioread32(ioaddr + 0xA8), addr + 4);
1520 		eth_hw_addr_set(dev, addr);
1521 		for (i = 0; i < 6; i ++)
1522 			sum += dev->dev_addr[i];
1523 	} else {
1524 		/* A serial EEPROM interface, we read now and sort it out later. */
1525 		int sa_offset = 0;
1526 		int ee_addr_size = tulip_read_eeprom(dev, 0xff, 8) & 0x40000 ? 8 : 6;
1527 		int ee_max_addr = ((1 << ee_addr_size) - 1) * sizeof(u16);
1528 
1529 		if (ee_max_addr > sizeof(tp->eeprom))
1530 			ee_max_addr = sizeof(tp->eeprom);
1531 
1532 		for (i = 0; i < ee_max_addr ; i += sizeof(u16)) {
1533 			u16 data = tulip_read_eeprom(dev, i/2, ee_addr_size);
1534 			ee_data[i] = data & 0xff;
1535 			ee_data[i + 1] = data >> 8;
1536 		}
1537 
1538 		/* DEC now has a specification (see Notes) but early board makers
1539 		   just put the address in the first EEPROM locations. */
1540 		/* This does  memcmp(ee_data, ee_data+16, 8) */
1541 		for (i = 0; i < 8; i ++)
1542 			if (ee_data[i] != ee_data[16+i])
1543 				sa_offset = 20;
1544 		if (chip_idx == CONEXANT) {
1545 			/* Check that the tuple type and length is correct. */
1546 			if (ee_data[0x198] == 0x04  &&  ee_data[0x199] == 6)
1547 				sa_offset = 0x19A;
1548 		} else if (ee_data[0] == 0xff  &&  ee_data[1] == 0xff &&
1549 				   ee_data[2] == 0) {
1550 			sa_offset = 2;		/* Grrr, damn Matrox boards. */
1551 		}
1552 #ifdef CONFIG_MIPS_COBALT
1553                if ((pdev->bus->number == 0) &&
1554                    ((PCI_SLOT(pdev->devfn) == 7) ||
1555                     (PCI_SLOT(pdev->devfn) == 12))) {
1556                        /* Cobalt MAC address in first EEPROM locations. */
1557                        sa_offset = 0;
1558 		       /* Ensure our media table fixup get's applied */
1559 		       memcpy(ee_data + 16, ee_data, 8);
1560                }
1561 #endif
1562 #ifdef CONFIG_GSC
1563 		/* Check to see if we have a broken srom */
1564 		if (ee_data[0] == 0x61 && ee_data[1] == 0x10) {
1565 			/* pci_vendor_id and subsystem_id are swapped */
1566 			ee_data[0] = ee_data[2];
1567 			ee_data[1] = ee_data[3];
1568 			ee_data[2] = 0x61;
1569 			ee_data[3] = 0x10;
1570 
1571 			/* HSC-PCI boards need to be byte-swaped and shifted
1572 			 * up 1 word.  This shift needs to happen at the end
1573 			 * of the MAC first because of the 2 byte overlap.
1574 			 */
1575 			for (i = 4; i >= 0; i -= 2) {
1576 				ee_data[17 + i + 3] = ee_data[17 + i];
1577 				ee_data[16 + i + 5] = ee_data[16 + i];
1578 			}
1579 		}
1580 #endif
1581 
1582 		for (i = 0; i < 6; i ++) {
1583 			addr[i] = ee_data[i + sa_offset];
1584 			sum += ee_data[i + sa_offset];
1585 		}
1586 		eth_hw_addr_set(dev, addr);
1587 	}
1588 	/* Lite-On boards have the address byte-swapped. */
1589 	if ((dev->dev_addr[0] == 0xA0 ||
1590 	     dev->dev_addr[0] == 0xC0 ||
1591 	     dev->dev_addr[0] == 0x02) &&
1592 	    dev->dev_addr[1] == 0x00) {
1593 		for (i = 0; i < 6; i+=2) {
1594 			addr[i] = dev->dev_addr[i+1];
1595 			addr[i+1] = dev->dev_addr[i];
1596 		}
1597 		eth_hw_addr_set(dev, addr);
1598 	}
1599 
1600 	/* On the Zynx 315 Etherarray and other multiport boards only the
1601 	   first Tulip has an EEPROM.
1602 	   On Sparc systems the mac address is held in the OBP property
1603 	   "local-mac-address".
1604 	   The addresses of the subsequent ports are derived from the first.
1605 	   Many PCI BIOSes also incorrectly report the IRQ line, so we correct
1606 	   that here as well. */
1607 	if (sum == 0  || sum == 6*0xff) {
1608 #if defined(CONFIG_SPARC)
1609 		struct device_node *dp = pci_device_to_OF_node(pdev);
1610 		const unsigned char *addr2;
1611 		int len;
1612 #endif
1613 		eeprom_missing = 1;
1614 		for (i = 0; i < 5; i++)
1615 			addr[i] = last_phys_addr[i];
1616 		addr[i] = last_phys_addr[i] + 1;
1617 		eth_hw_addr_set(dev, addr);
1618 #if defined(CONFIG_SPARC)
1619 		addr2 = of_get_property(dp, "local-mac-address", &len);
1620 		if (addr2 && len == ETH_ALEN)
1621 			eth_hw_addr_set(dev, addr2);
1622 #endif
1623 #if defined(__i386__) || defined(__x86_64__)	/* Patch up x86 BIOS bug. */
1624 		if (last_irq)
1625 			irq = last_irq;
1626 #endif
1627 	}
1628 
1629 	for (i = 0; i < 6; i++)
1630 		last_phys_addr[i] = dev->dev_addr[i];
1631 #if defined(__i386__) || defined(__x86_64__)	/* Patch up x86 BIOS bug. */
1632 	last_irq = irq;
1633 #endif
1634 
1635 	/* The lower four bits are the media type. */
1636 	if (board_idx >= 0  &&  board_idx < MAX_UNITS) {
1637 		if (options[board_idx] & MEDIA_MASK)
1638 			tp->default_port = options[board_idx] & MEDIA_MASK;
1639 		if ((options[board_idx] & FullDuplex) || full_duplex[board_idx] > 0)
1640 			tp->full_duplex = 1;
1641 		if (mtu[board_idx] > 0)
1642 			dev->mtu = mtu[board_idx];
1643 	}
1644 	if (dev->mem_start & MEDIA_MASK)
1645 		tp->default_port = dev->mem_start & MEDIA_MASK;
1646 	if (tp->default_port) {
1647 		pr_info(DRV_NAME "%d: Transceiver selection forced to %s\n",
1648 			board_idx, medianame[tp->default_port & MEDIA_MASK]);
1649 		tp->medialock = 1;
1650 		if (tulip_media_cap[tp->default_port] & MediaAlwaysFD)
1651 			tp->full_duplex = 1;
1652 	}
1653 	if (tp->full_duplex)
1654 		tp->full_duplex_lock = 1;
1655 
1656 	if (tulip_media_cap[tp->default_port] & MediaIsMII) {
1657 		static const u16 media2advert[] = {
1658 			0x20, 0x40, 0x03e0, 0x60, 0x80, 0x100, 0x200
1659 		};
1660 		tp->mii_advertise = media2advert[tp->default_port - 9];
1661 		tp->mii_advertise |= (tp->flags & HAS_8023X); /* Matching bits! */
1662 	}
1663 
1664 	if (tp->flags & HAS_MEDIA_TABLE) {
1665 		sprintf(dev->name, DRV_NAME "%d", board_idx);	/* hack */
1666 		tulip_parse_eeprom(dev);
1667 		strcpy(dev->name, "eth%d");			/* un-hack */
1668 	}
1669 
1670 	if ((tp->flags & ALWAYS_CHECK_MII) ||
1671 		(tp->mtable  &&  tp->mtable->has_mii) ||
1672 		( ! tp->mtable  &&  (tp->flags & HAS_MII))) {
1673 		if (tp->mtable  &&  tp->mtable->has_mii) {
1674 			for (i = 0; i < tp->mtable->leafcount; i++)
1675 				if (tp->mtable->mleaf[i].media == 11) {
1676 					tp->cur_index = i;
1677 					tp->saved_if_port = dev->if_port;
1678 					tulip_select_media(dev, 2);
1679 					dev->if_port = tp->saved_if_port;
1680 					break;
1681 				}
1682 		}
1683 
1684 		/* Find the connected MII xcvrs.
1685 		   Doing this in open() would allow detecting external xcvrs
1686 		   later, but takes much time. */
1687 		tulip_find_mii (dev, board_idx);
1688 	}
1689 
1690 	/* The Tulip-specific entries in the device structure. */
1691 	dev->netdev_ops = &tulip_netdev_ops;
1692 	dev->watchdog_timeo = TX_TIMEOUT;
1693 #ifdef CONFIG_TULIP_NAPI
1694 	netif_napi_add(dev, &tp->napi, tulip_poll, 16);
1695 #endif
1696 	dev->ethtool_ops = &ops;
1697 
1698 	if (register_netdev(dev))
1699 		goto err_out_free_ring;
1700 
1701 	pci_set_drvdata(pdev, dev);
1702 
1703 	dev_info(&dev->dev,
1704 #ifdef CONFIG_TULIP_MMIO
1705 		 "%s rev %d at MMIO %#llx,%s %pM, IRQ %d\n",
1706 #else
1707 		 "%s rev %d at Port %#llx,%s %pM, IRQ %d\n",
1708 #endif
1709 		 chip_name, pdev->revision,
1710 		 (unsigned long long)pci_resource_start(pdev, TULIP_BAR),
1711 		 eeprom_missing ? " EEPROM not present," : "",
1712 		 dev->dev_addr, irq);
1713 
1714         if (tp->chip_id == PNIC2)
1715 		tp->link_change = pnic2_lnk_change;
1716 	else if (tp->flags & HAS_NWAY)
1717 		tp->link_change = t21142_lnk_change;
1718 	else if (tp->flags & HAS_PNICNWAY)
1719 		tp->link_change = pnic_lnk_change;
1720 
1721 	/* Reset the xcvr interface and turn on heartbeat. */
1722 	switch (chip_idx) {
1723 	case DC21140:
1724 	case DM910X:
1725 	default:
1726 		if (tp->mtable)
1727 			iowrite32(tp->mtable->csr12dir | 0x100, ioaddr + CSR12);
1728 		break;
1729 	case DC21142:
1730 		if (tp->mii_cnt  ||  tulip_media_cap[dev->if_port] & MediaIsMII) {
1731 			iowrite32(csr6_mask_defstate, ioaddr + CSR6);
1732 			iowrite32(0x0000, ioaddr + CSR13);
1733 			iowrite32(0x0000, ioaddr + CSR14);
1734 			iowrite32(csr6_mask_hdcap, ioaddr + CSR6);
1735 		} else
1736 			t21142_start_nway(dev);
1737 		break;
1738 	case PNIC2:
1739 	        /* just do a reset for sanity sake */
1740 		iowrite32(0x0000, ioaddr + CSR13);
1741 		iowrite32(0x0000, ioaddr + CSR14);
1742 		break;
1743 	case LC82C168:
1744 		if ( ! tp->mii_cnt) {
1745 			tp->nway = 1;
1746 			tp->nwayset = 0;
1747 			iowrite32(csr6_ttm | csr6_ca, ioaddr + CSR6);
1748 			iowrite32(0x30, ioaddr + CSR12);
1749 			iowrite32(0x0001F078, ioaddr + CSR6);
1750 			iowrite32(0x0201F078, ioaddr + CSR6); /* Turn on autonegotiation. */
1751 		}
1752 		break;
1753 	case MX98713:
1754 	case COMPEX9881:
1755 		iowrite32(0x00000000, ioaddr + CSR6);
1756 		iowrite32(0x000711C0, ioaddr + CSR14); /* Turn on NWay. */
1757 		iowrite32(0x00000001, ioaddr + CSR13);
1758 		break;
1759 	case MX98715:
1760 	case MX98725:
1761 		iowrite32(0x01a80000, ioaddr + CSR6);
1762 		iowrite32(0xFFFFFFFF, ioaddr + CSR14);
1763 		iowrite32(0x00001000, ioaddr + CSR12);
1764 		break;
1765 	case COMET:
1766 		/* No initialization necessary. */
1767 		break;
1768 	}
1769 
1770 	/* put the chip in snooze mode until opened */
1771 	tulip_set_power_state (tp, 0, 1);
1772 
1773 	return 0;
1774 
1775 err_out_free_ring:
1776 	dma_free_coherent(&pdev->dev,
1777 			  sizeof(struct tulip_rx_desc) * RX_RING_SIZE +
1778 			  sizeof(struct tulip_tx_desc) * TX_RING_SIZE,
1779 			  tp->rx_ring, tp->rx_ring_dma);
1780 
1781 err_out_mtable:
1782 	kfree (tp->mtable);
1783 	pci_iounmap(pdev, ioaddr);
1784 
1785 err_out_free_res:
1786 	pci_release_regions (pdev);
1787 
1788 err_out_free_netdev:
1789 	free_netdev (dev);
1790 	pci_disable_device(pdev);
1791 	return -ENODEV;
1792 }
1793 
1794 
1795 /* set the registers according to the given wolopts */
1796 static void tulip_set_wolopts (struct pci_dev *pdev, u32 wolopts)
1797 {
1798 	struct net_device *dev = pci_get_drvdata(pdev);
1799 	struct tulip_private *tp = netdev_priv(dev);
1800 	void __iomem *ioaddr = tp->base_addr;
1801 
1802 	if (tp->flags & COMET_PM) {
1803 		unsigned int tmp;
1804 
1805 		tmp = ioread32(ioaddr + CSR18);
1806 		tmp &= ~(comet_csr18_pmes_sticky | comet_csr18_apm_mode | comet_csr18_d3a);
1807 		tmp |= comet_csr18_pm_mode;
1808 		iowrite32(tmp, ioaddr + CSR18);
1809 
1810 		/* Set the Wake-up Control/Status Register to the given WOL options*/
1811 		tmp = ioread32(ioaddr + CSR13);
1812 		tmp &= ~(comet_csr13_linkoffe | comet_csr13_linkone | comet_csr13_wfre | comet_csr13_lsce | comet_csr13_mpre);
1813 		if (wolopts & WAKE_MAGIC)
1814 			tmp |= comet_csr13_mpre;
1815 		if (wolopts & WAKE_PHY)
1816 			tmp |= comet_csr13_linkoffe | comet_csr13_linkone | comet_csr13_lsce;
1817 		/* Clear the event flags */
1818 		tmp |= comet_csr13_wfr | comet_csr13_mpr | comet_csr13_lsc;
1819 		iowrite32(tmp, ioaddr + CSR13);
1820 	}
1821 }
1822 
1823 static int __maybe_unused tulip_suspend(struct device *dev_d)
1824 {
1825 	struct net_device *dev = dev_get_drvdata(dev_d);
1826 	struct tulip_private *tp = netdev_priv(dev);
1827 
1828 	if (!dev)
1829 		return -EINVAL;
1830 
1831 	if (!netif_running(dev))
1832 		goto save_state;
1833 
1834 	tulip_down(dev);
1835 
1836 	netif_device_detach(dev);
1837 	/* FIXME: it needlessly adds an error path. */
1838 	free_irq(tp->pdev->irq, dev);
1839 
1840 save_state:
1841 	tulip_set_wolopts(to_pci_dev(dev_d), tp->wolinfo.wolopts);
1842 	device_set_wakeup_enable(dev_d, !!tp->wolinfo.wolopts);
1843 
1844 	return 0;
1845 }
1846 
1847 static int __maybe_unused tulip_resume(struct device *dev_d)
1848 {
1849 	struct pci_dev *pdev = to_pci_dev(dev_d);
1850 	struct net_device *dev = dev_get_drvdata(dev_d);
1851 	struct tulip_private *tp = netdev_priv(dev);
1852 	void __iomem *ioaddr = tp->base_addr;
1853 	unsigned int tmp;
1854 	int retval = 0;
1855 
1856 	if (!dev)
1857 		return -EINVAL;
1858 
1859 	if (!netif_running(dev))
1860 		return 0;
1861 
1862 	retval = request_irq(pdev->irq, tulip_interrupt, IRQF_SHARED,
1863 			     dev->name, dev);
1864 	if (retval) {
1865 		pr_err("request_irq failed in resume\n");
1866 		return retval;
1867 	}
1868 
1869 	if (tp->flags & COMET_PM) {
1870 		device_set_wakeup_enable(dev_d, 0);
1871 
1872 		/* Clear the PMES flag */
1873 		tmp = ioread32(ioaddr + CSR20);
1874 		tmp |= comet_csr20_pmes;
1875 		iowrite32(tmp, ioaddr + CSR20);
1876 
1877 		/* Disable all wake-up events */
1878 		tulip_set_wolopts(pdev, 0);
1879 	}
1880 	netif_device_attach(dev);
1881 
1882 	if (netif_running(dev))
1883 		tulip_up(dev);
1884 
1885 	return 0;
1886 }
1887 
1888 static void tulip_remove_one(struct pci_dev *pdev)
1889 {
1890 	struct net_device *dev = pci_get_drvdata (pdev);
1891 	struct tulip_private *tp;
1892 
1893 	if (!dev)
1894 		return;
1895 
1896 	tp = netdev_priv(dev);
1897 	unregister_netdev(dev);
1898 	dma_free_coherent(&pdev->dev,
1899 			  sizeof(struct tulip_rx_desc) * RX_RING_SIZE +
1900 			  sizeof(struct tulip_tx_desc) * TX_RING_SIZE,
1901 			  tp->rx_ring, tp->rx_ring_dma);
1902 	kfree (tp->mtable);
1903 	pci_iounmap(pdev, tp->base_addr);
1904 	free_netdev (dev);
1905 	pci_release_regions (pdev);
1906 	pci_disable_device(pdev);
1907 
1908 	/* pci_power_off (pdev, -1); */
1909 }
1910 
1911 #ifdef CONFIG_NET_POLL_CONTROLLER
1912 /*
1913  * Polling 'interrupt' - used by things like netconsole to send skbs
1914  * without having to re-enable interrupts. It's not called while
1915  * the interrupt routine is executing.
1916  */
1917 
1918 static void poll_tulip (struct net_device *dev)
1919 {
1920 	struct tulip_private *tp = netdev_priv(dev);
1921 	const int irq = tp->pdev->irq;
1922 
1923 	/* disable_irq here is not very nice, but with the lockless
1924 	   interrupt handler we have no other choice. */
1925 	disable_irq(irq);
1926 	tulip_interrupt (irq, dev);
1927 	enable_irq(irq);
1928 }
1929 #endif
1930 
1931 static SIMPLE_DEV_PM_OPS(tulip_pm_ops, tulip_suspend, tulip_resume);
1932 
1933 static struct pci_driver tulip_driver = {
1934 	.name		= DRV_NAME,
1935 	.id_table	= tulip_pci_tbl,
1936 	.probe		= tulip_init_one,
1937 	.remove		= tulip_remove_one,
1938 	.driver.pm	= &tulip_pm_ops,
1939 };
1940 
1941 
1942 static int __init tulip_init (void)
1943 {
1944 	if (!csr0) {
1945 		pr_warn("tulip: unknown CPU architecture, using default csr0\n");
1946 		/* default to 8 longword cache line alignment */
1947 		csr0 = 0x00A00000 | 0x4800;
1948 	}
1949 
1950 	/* copy module parms into globals */
1951 	tulip_rx_copybreak = rx_copybreak;
1952 	tulip_max_interrupt_work = max_interrupt_work;
1953 
1954 	/* probe for and init boards */
1955 	return pci_register_driver(&tulip_driver);
1956 }
1957 
1958 
1959 static void __exit tulip_cleanup (void)
1960 {
1961 	pci_unregister_driver (&tulip_driver);
1962 }
1963 
1964 
1965 module_init(tulip_init);
1966 module_exit(tulip_cleanup);
1967