xref: /dragonfly/sys/dev/netif/ix/ixgbe_82598.c (revision 7ff0fc30)
1 /******************************************************************************
2 
3   Copyright (c) 2001-2017, Intel Corporation
4   All rights reserved.
5 
6   Redistribution and use in source and binary forms, with or without
7   modification, are permitted provided that the following conditions are met:
8 
9    1. Redistributions of source code must retain the above copyright notice,
10       this list of conditions and the following disclaimer.
11 
12    2. Redistributions in binary form must reproduce the above copyright
13       notice, this list of conditions and the following disclaimer in the
14       documentation and/or other materials provided with the distribution.
15 
16    3. Neither the name of the Intel Corporation nor the names of its
17       contributors may be used to endorse or promote products derived from
18       this software without specific prior written permission.
19 
20   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21   AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22   IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23   ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24   LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25   CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26   SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27   INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28   CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30   POSSIBILITY OF SUCH DAMAGE.
31 
32 ******************************************************************************/
33 /*$FreeBSD$*/
34 
35 #include "ixgbe_type.h"
36 #include "ixgbe_82598.h"
37 #include "ixgbe_api.h"
38 #include "ixgbe_common.h"
39 #include "ixgbe_phy.h"
40 
41 #define IXGBE_82598_MAX_TX_QUEUES 32
42 #define IXGBE_82598_MAX_RX_QUEUES 64
43 #define IXGBE_82598_RAR_ENTRIES   16
44 #define IXGBE_82598_MC_TBL_SIZE  128
45 #define IXGBE_82598_VFT_TBL_SIZE 128
46 #define IXGBE_82598_RX_PB_SIZE   512
47 
48 static s32 ixgbe_get_link_capabilities_82598(struct ixgbe_hw *hw,
49 					     ixgbe_link_speed *speed,
50 					     bool *autoneg);
51 static enum ixgbe_media_type ixgbe_get_media_type_82598(struct ixgbe_hw *hw);
52 static s32 ixgbe_start_mac_link_82598(struct ixgbe_hw *hw,
53 				      bool autoneg_wait_to_complete);
54 static s32 ixgbe_check_mac_link_82598(struct ixgbe_hw *hw,
55 				      ixgbe_link_speed *speed, bool *link_up,
56 				      bool link_up_wait_to_complete);
57 static s32 ixgbe_setup_mac_link_82598(struct ixgbe_hw *hw,
58 				      ixgbe_link_speed speed,
59 				      bool autoneg_wait_to_complete);
60 static s32 ixgbe_setup_copper_link_82598(struct ixgbe_hw *hw,
61 					 ixgbe_link_speed speed,
62 					 bool autoneg_wait_to_complete);
63 static s32 ixgbe_reset_hw_82598(struct ixgbe_hw *hw);
64 static s32 ixgbe_clear_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq);
65 static s32 ixgbe_clear_vfta_82598(struct ixgbe_hw *hw);
66 static void ixgbe_set_rxpba_82598(struct ixgbe_hw *hw, int num_pb,
67 				  u32 headroom, int strategy);
68 static s32 ixgbe_read_i2c_sff8472_82598(struct ixgbe_hw *hw, u8 byte_offset,
69 					u8 *sff8472_data);
70 /**
71  *  ixgbe_set_pcie_completion_timeout - set pci-e completion timeout
72  *  @hw: pointer to the HW structure
73  *
74  *  The defaults for 82598 should be in the range of 50us to 50ms,
75  *  however the hardware default for these parts is 500us to 1ms which is less
76  *  than the 10ms recommended by the pci-e spec.  To address this we need to
77  *  increase the value to either 10ms to 250ms for capability version 1 config,
78  *  or 16ms to 55ms for version 2.
79  **/
80 void ixgbe_set_pcie_completion_timeout(struct ixgbe_hw *hw)
81 {
82 	u32 gcr = IXGBE_READ_REG(hw, IXGBE_GCR);
83 	u16 pcie_devctl2;
84 
85 	/* only take action if timeout value is defaulted to 0 */
86 	if (gcr & IXGBE_GCR_CMPL_TMOUT_MASK)
87 		goto out;
88 
89 	/*
90 	 * if capababilities version is type 1 we can write the
91 	 * timeout of 10ms to 250ms through the GCR register
92 	 */
93 	if (!(gcr & IXGBE_GCR_CAP_VER2)) {
94 		gcr |= IXGBE_GCR_CMPL_TMOUT_10ms;
95 		goto out;
96 	}
97 
98 	/*
99 	 * for version 2 capabilities we need to write the config space
100 	 * directly in order to set the completion timeout value for
101 	 * 16ms to 55ms
102 	 */
103 	pcie_devctl2 = IXGBE_READ_PCIE_WORD(hw, IXGBE_PCI_DEVICE_CONTROL2);
104 	pcie_devctl2 |= IXGBE_PCI_DEVICE_CONTROL2_16ms;
105 	IXGBE_WRITE_PCIE_WORD(hw, IXGBE_PCI_DEVICE_CONTROL2, pcie_devctl2);
106 out:
107 	/* disable completion timeout resend */
108 	gcr &= ~IXGBE_GCR_CMPL_TMOUT_RESEND;
109 	IXGBE_WRITE_REG(hw, IXGBE_GCR, gcr);
110 }
111 
112 /**
113  *  ixgbe_init_ops_82598 - Inits func ptrs and MAC type
114  *  @hw: pointer to hardware structure
115  *
116  *  Initialize the function pointers and assign the MAC type for 82598.
117  *  Does not touch the hardware.
118  **/
119 s32 ixgbe_init_ops_82598(struct ixgbe_hw *hw)
120 {
121 	struct ixgbe_mac_info *mac = &hw->mac;
122 	struct ixgbe_phy_info *phy = &hw->phy;
123 	s32 ret_val;
124 
125 	DEBUGFUNC("ixgbe_init_ops_82598");
126 
127 	ret_val = ixgbe_init_phy_ops_generic(hw);
128 	ret_val = ixgbe_init_ops_generic(hw);
129 
130 	/* PHY */
131 	phy->ops.init = ixgbe_init_phy_ops_82598;
132 
133 	/* MAC */
134 	mac->ops.start_hw = ixgbe_start_hw_82598;
135 	mac->ops.enable_relaxed_ordering = ixgbe_enable_relaxed_ordering_82598;
136 	mac->ops.reset_hw = ixgbe_reset_hw_82598;
137 	mac->ops.get_media_type = ixgbe_get_media_type_82598;
138 	mac->ops.get_supported_physical_layer =
139 				ixgbe_get_supported_physical_layer_82598;
140 	mac->ops.read_analog_reg8 = ixgbe_read_analog_reg8_82598;
141 	mac->ops.write_analog_reg8 = ixgbe_write_analog_reg8_82598;
142 	mac->ops.set_lan_id = ixgbe_set_lan_id_multi_port_pcie_82598;
143 	mac->ops.enable_rx_dma = ixgbe_enable_rx_dma_82598;
144 
145 	/* RAR, Multicast, VLAN */
146 	mac->ops.set_vmdq = ixgbe_set_vmdq_82598;
147 	mac->ops.clear_vmdq = ixgbe_clear_vmdq_82598;
148 	mac->ops.set_vfta = ixgbe_set_vfta_82598;
149 	mac->ops.set_vlvf = NULL;
150 	mac->ops.clear_vfta = ixgbe_clear_vfta_82598;
151 
152 	/* Flow Control */
153 	mac->ops.fc_enable = ixgbe_fc_enable_82598;
154 
155 	mac->mcft_size		= IXGBE_82598_MC_TBL_SIZE;
156 	mac->vft_size		= IXGBE_82598_VFT_TBL_SIZE;
157 	mac->num_rar_entries	= IXGBE_82598_RAR_ENTRIES;
158 	mac->rx_pb_size		= IXGBE_82598_RX_PB_SIZE;
159 	mac->max_rx_queues	= IXGBE_82598_MAX_RX_QUEUES;
160 	mac->max_tx_queues	= IXGBE_82598_MAX_TX_QUEUES;
161 	mac->max_msix_vectors	= ixgbe_get_pcie_msix_count_generic(hw);
162 
163 	/* SFP+ Module */
164 	phy->ops.read_i2c_eeprom = ixgbe_read_i2c_eeprom_82598;
165 	phy->ops.read_i2c_sff8472 = ixgbe_read_i2c_sff8472_82598;
166 
167 	/* Link */
168 	mac->ops.check_link = ixgbe_check_mac_link_82598;
169 	mac->ops.setup_link = ixgbe_setup_mac_link_82598;
170 	mac->ops.flap_tx_laser = NULL;
171 	mac->ops.get_link_capabilities = ixgbe_get_link_capabilities_82598;
172 	mac->ops.setup_rxpba = ixgbe_set_rxpba_82598;
173 
174 	/* Manageability interface */
175 	mac->ops.set_fw_drv_ver = NULL;
176 
177 	mac->ops.get_rtrup2tc = NULL;
178 
179 	return ret_val;
180 }
181 
182 /**
183  *  ixgbe_init_phy_ops_82598 - PHY/SFP specific init
184  *  @hw: pointer to hardware structure
185  *
186  *  Initialize any function pointers that were not able to be
187  *  set during init_shared_code because the PHY/SFP type was
188  *  not known.  Perform the SFP init if necessary.
189  *
190  **/
191 s32 ixgbe_init_phy_ops_82598(struct ixgbe_hw *hw)
192 {
193 	struct ixgbe_mac_info *mac = &hw->mac;
194 	struct ixgbe_phy_info *phy = &hw->phy;
195 	s32 ret_val = IXGBE_SUCCESS;
196 	u16 list_offset, data_offset;
197 
198 	DEBUGFUNC("ixgbe_init_phy_ops_82598");
199 
200 	/* Identify the PHY */
201 	phy->ops.identify(hw);
202 
203 	/* Overwrite the link function pointers if copper PHY */
204 	if (mac->ops.get_media_type(hw) == ixgbe_media_type_copper) {
205 		mac->ops.setup_link = ixgbe_setup_copper_link_82598;
206 		mac->ops.get_link_capabilities =
207 				ixgbe_get_copper_link_capabilities_generic;
208 	}
209 
210 	switch (hw->phy.type) {
211 	case ixgbe_phy_tn:
212 		phy->ops.setup_link = ixgbe_setup_phy_link_tnx;
213 		phy->ops.check_link = ixgbe_check_phy_link_tnx;
214 		phy->ops.get_firmware_version =
215 					ixgbe_get_phy_firmware_version_tnx;
216 		break;
217 	case ixgbe_phy_nl:
218 		phy->ops.reset = ixgbe_reset_phy_nl;
219 
220 		/* Call SFP+ identify routine to get the SFP+ module type */
221 		ret_val = phy->ops.identify_sfp(hw);
222 		if (ret_val != IXGBE_SUCCESS)
223 			goto out;
224 		else if (hw->phy.sfp_type == ixgbe_sfp_type_unknown) {
225 			ret_val = IXGBE_ERR_SFP_NOT_SUPPORTED;
226 			goto out;
227 		}
228 
229 		/* Check to see if SFP+ module is supported */
230 		ret_val = ixgbe_get_sfp_init_sequence_offsets(hw,
231 							      &list_offset,
232 							      &data_offset);
233 		if (ret_val != IXGBE_SUCCESS) {
234 			ret_val = IXGBE_ERR_SFP_NOT_SUPPORTED;
235 			goto out;
236 		}
237 		break;
238 	default:
239 		break;
240 	}
241 
242 out:
243 	return ret_val;
244 }
245 
246 /**
247  *  ixgbe_start_hw_82598 - Prepare hardware for Tx/Rx
248  *  @hw: pointer to hardware structure
249  *
250  *  Starts the hardware using the generic start_hw function.
251  *  Disables relaxed ordering Then set pcie completion timeout
252  *
253  **/
254 s32 ixgbe_start_hw_82598(struct ixgbe_hw *hw)
255 {
256 	u32 regval;
257 	u32 i;
258 	s32 ret_val = IXGBE_SUCCESS;
259 
260 	DEBUGFUNC("ixgbe_start_hw_82598");
261 
262 	ret_val = ixgbe_start_hw_generic(hw);
263 	if (ret_val)
264 		return ret_val;
265 
266 	/* Disable relaxed ordering */
267 	for (i = 0; ((i < hw->mac.max_tx_queues) &&
268 	     (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
269 		regval = IXGBE_READ_REG(hw, IXGBE_DCA_TXCTRL(i));
270 		regval &= ~IXGBE_DCA_TXCTRL_DESC_WRO_EN;
271 		IXGBE_WRITE_REG(hw, IXGBE_DCA_TXCTRL(i), regval);
272 	}
273 
274 	for (i = 0; ((i < hw->mac.max_rx_queues) &&
275 	     (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
276 		regval = IXGBE_READ_REG(hw, IXGBE_DCA_RXCTRL(i));
277 		regval &= ~(IXGBE_DCA_RXCTRL_DATA_WRO_EN |
278 			    IXGBE_DCA_RXCTRL_HEAD_WRO_EN);
279 		IXGBE_WRITE_REG(hw, IXGBE_DCA_RXCTRL(i), regval);
280 	}
281 
282 	/* set the completion timeout for interface */
283 	ixgbe_set_pcie_completion_timeout(hw);
284 
285 	return ret_val;
286 }
287 
288 /**
289  *  ixgbe_get_link_capabilities_82598 - Determines link capabilities
290  *  @hw: pointer to hardware structure
291  *  @speed: pointer to link speed
292  *  @autoneg: boolean auto-negotiation value
293  *
294  *  Determines the link capabilities by reading the AUTOC register.
295  **/
296 static s32 ixgbe_get_link_capabilities_82598(struct ixgbe_hw *hw,
297 					     ixgbe_link_speed *speed,
298 					     bool *autoneg)
299 {
300 	s32 status = IXGBE_SUCCESS;
301 	u32 autoc = 0;
302 
303 	DEBUGFUNC("ixgbe_get_link_capabilities_82598");
304 
305 	/*
306 	 * Determine link capabilities based on the stored value of AUTOC,
307 	 * which represents EEPROM defaults.  If AUTOC value has not been
308 	 * stored, use the current register value.
309 	 */
310 	if (hw->mac.orig_link_settings_stored)
311 		autoc = hw->mac.orig_autoc;
312 	else
313 		autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
314 
315 	switch (autoc & IXGBE_AUTOC_LMS_MASK) {
316 	case IXGBE_AUTOC_LMS_1G_LINK_NO_AN:
317 		*speed = IXGBE_LINK_SPEED_1GB_FULL;
318 		*autoneg = FALSE;
319 		break;
320 
321 	case IXGBE_AUTOC_LMS_10G_LINK_NO_AN:
322 		*speed = IXGBE_LINK_SPEED_10GB_FULL;
323 		*autoneg = FALSE;
324 		break;
325 
326 	case IXGBE_AUTOC_LMS_1G_AN:
327 		*speed = IXGBE_LINK_SPEED_1GB_FULL;
328 		*autoneg = TRUE;
329 		break;
330 
331 	case IXGBE_AUTOC_LMS_KX4_AN:
332 	case IXGBE_AUTOC_LMS_KX4_AN_1G_AN:
333 		*speed = IXGBE_LINK_SPEED_UNKNOWN;
334 		if (autoc & IXGBE_AUTOC_KX4_SUPP)
335 			*speed |= IXGBE_LINK_SPEED_10GB_FULL;
336 		if (autoc & IXGBE_AUTOC_KX_SUPP)
337 			*speed |= IXGBE_LINK_SPEED_1GB_FULL;
338 		*autoneg = TRUE;
339 		break;
340 
341 	default:
342 		status = IXGBE_ERR_LINK_SETUP;
343 		break;
344 	}
345 
346 	return status;
347 }
348 
349 /**
350  *  ixgbe_get_media_type_82598 - Determines media type
351  *  @hw: pointer to hardware structure
352  *
353  *  Returns the media type (fiber, copper, backplane)
354  **/
355 static enum ixgbe_media_type ixgbe_get_media_type_82598(struct ixgbe_hw *hw)
356 {
357 	enum ixgbe_media_type media_type;
358 
359 	DEBUGFUNC("ixgbe_get_media_type_82598");
360 
361 	/* Detect if there is a copper PHY attached. */
362 	switch (hw->phy.type) {
363 	case ixgbe_phy_cu_unknown:
364 	case ixgbe_phy_tn:
365 		media_type = ixgbe_media_type_copper;
366 		goto out;
367 	default:
368 		break;
369 	}
370 
371 	/* Media type for I82598 is based on device ID */
372 	switch (hw->device_id) {
373 	case IXGBE_DEV_ID_82598:
374 	case IXGBE_DEV_ID_82598_BX:
375 		/* Default device ID is mezzanine card KX/KX4 */
376 		media_type = ixgbe_media_type_backplane;
377 		break;
378 	case IXGBE_DEV_ID_82598AF_DUAL_PORT:
379 	case IXGBE_DEV_ID_82598AF_SINGLE_PORT:
380 	case IXGBE_DEV_ID_82598_DA_DUAL_PORT:
381 	case IXGBE_DEV_ID_82598_SR_DUAL_PORT_EM:
382 	case IXGBE_DEV_ID_82598EB_XF_LR:
383 	case IXGBE_DEV_ID_82598EB_SFP_LOM:
384 		media_type = ixgbe_media_type_fiber;
385 		break;
386 	case IXGBE_DEV_ID_82598EB_CX4:
387 	case IXGBE_DEV_ID_82598_CX4_DUAL_PORT:
388 		media_type = ixgbe_media_type_cx4;
389 		break;
390 	case IXGBE_DEV_ID_82598AT:
391 	case IXGBE_DEV_ID_82598AT2:
392 		media_type = ixgbe_media_type_copper;
393 		break;
394 	default:
395 		media_type = ixgbe_media_type_unknown;
396 		break;
397 	}
398 out:
399 	return media_type;
400 }
401 
402 /**
403  *  ixgbe_fc_enable_82598 - Enable flow control
404  *  @hw: pointer to hardware structure
405  *
406  *  Enable flow control according to the current settings.
407  **/
408 s32 ixgbe_fc_enable_82598(struct ixgbe_hw *hw)
409 {
410 	s32 ret_val = IXGBE_SUCCESS;
411 	u32 fctrl_reg;
412 	u32 rmcs_reg;
413 	u32 reg;
414 	u32 fcrtl, fcrth;
415 	u32 link_speed = 0;
416 	int i;
417 	bool link_up;
418 
419 	DEBUGFUNC("ixgbe_fc_enable_82598");
420 
421 	/* Validate the water mark configuration */
422 	if (!hw->fc.pause_time) {
423 		ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
424 		goto out;
425 	}
426 
427 	/* Low water mark of zero causes XOFF floods */
428 	for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
429 		if ((hw->fc.current_mode & ixgbe_fc_tx_pause) &&
430 		    hw->fc.high_water[i]) {
431 			if (!hw->fc.low_water[i] ||
432 			    hw->fc.low_water[i] >= hw->fc.high_water[i]) {
433 				DEBUGOUT("Invalid water mark configuration\n");
434 				ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
435 				goto out;
436 			}
437 		}
438 	}
439 
440 	/*
441 	 * On 82598 having Rx FC on causes resets while doing 1G
442 	 * so if it's on turn it off once we know link_speed. For
443 	 * more details see 82598 Specification update.
444 	 */
445 	hw->mac.ops.check_link(hw, &link_speed, &link_up, FALSE);
446 	if (link_up && link_speed == IXGBE_LINK_SPEED_1GB_FULL) {
447 		switch (hw->fc.requested_mode) {
448 		case ixgbe_fc_full:
449 			hw->fc.requested_mode = ixgbe_fc_tx_pause;
450 			break;
451 		case ixgbe_fc_rx_pause:
452 			hw->fc.requested_mode = ixgbe_fc_none;
453 			break;
454 		default:
455 			/* no change */
456 			break;
457 		}
458 	}
459 
460 	/* Negotiate the fc mode to use */
461 	ixgbe_fc_autoneg(hw);
462 
463 	/* Disable any previous flow control settings */
464 	fctrl_reg = IXGBE_READ_REG(hw, IXGBE_FCTRL);
465 	fctrl_reg &= ~(IXGBE_FCTRL_RFCE | IXGBE_FCTRL_RPFCE);
466 
467 	rmcs_reg = IXGBE_READ_REG(hw, IXGBE_RMCS);
468 	rmcs_reg &= ~(IXGBE_RMCS_TFCE_PRIORITY | IXGBE_RMCS_TFCE_802_3X);
469 
470 	/*
471 	 * The possible values of fc.current_mode are:
472 	 * 0: Flow control is completely disabled
473 	 * 1: Rx flow control is enabled (we can receive pause frames,
474 	 *    but not send pause frames).
475 	 * 2: Tx flow control is enabled (we can send pause frames but
476 	 *     we do not support receiving pause frames).
477 	 * 3: Both Rx and Tx flow control (symmetric) are enabled.
478 	 * other: Invalid.
479 	 */
480 	switch (hw->fc.current_mode) {
481 	case ixgbe_fc_none:
482 		/*
483 		 * Flow control is disabled by software override or autoneg.
484 		 * The code below will actually disable it in the HW.
485 		 */
486 		break;
487 	case ixgbe_fc_rx_pause:
488 		/*
489 		 * Rx Flow control is enabled and Tx Flow control is
490 		 * disabled by software override. Since there really
491 		 * isn't a way to advertise that we are capable of RX
492 		 * Pause ONLY, we will advertise that we support both
493 		 * symmetric and asymmetric Rx PAUSE.  Later, we will
494 		 * disable the adapter's ability to send PAUSE frames.
495 		 */
496 		fctrl_reg |= IXGBE_FCTRL_RFCE;
497 		break;
498 	case ixgbe_fc_tx_pause:
499 		/*
500 		 * Tx Flow control is enabled, and Rx Flow control is
501 		 * disabled by software override.
502 		 */
503 		rmcs_reg |= IXGBE_RMCS_TFCE_802_3X;
504 		break;
505 	case ixgbe_fc_full:
506 		/* Flow control (both Rx and Tx) is enabled by SW override. */
507 		fctrl_reg |= IXGBE_FCTRL_RFCE;
508 		rmcs_reg |= IXGBE_RMCS_TFCE_802_3X;
509 		break;
510 	default:
511 		DEBUGOUT("Flow control param set incorrectly\n");
512 		ret_val = IXGBE_ERR_CONFIG;
513 		goto out;
514 		break;
515 	}
516 
517 	/* Set 802.3x based flow control settings. */
518 	fctrl_reg |= IXGBE_FCTRL_DPF;
519 	IXGBE_WRITE_REG(hw, IXGBE_FCTRL, fctrl_reg);
520 	IXGBE_WRITE_REG(hw, IXGBE_RMCS, rmcs_reg);
521 
522 	/* Set up and enable Rx high/low water mark thresholds, enable XON. */
523 	for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
524 		if ((hw->fc.current_mode & ixgbe_fc_tx_pause) &&
525 		    hw->fc.high_water[i]) {
526 			fcrtl = (hw->fc.low_water[i] << 10) | IXGBE_FCRTL_XONE;
527 			fcrth = (hw->fc.high_water[i] << 10) | IXGBE_FCRTH_FCEN;
528 			IXGBE_WRITE_REG(hw, IXGBE_FCRTL(i), fcrtl);
529 			IXGBE_WRITE_REG(hw, IXGBE_FCRTH(i), fcrth);
530 		} else {
531 			IXGBE_WRITE_REG(hw, IXGBE_FCRTL(i), 0);
532 			IXGBE_WRITE_REG(hw, IXGBE_FCRTH(i), 0);
533 		}
534 
535 	}
536 
537 	/* Configure pause time (2 TCs per register) */
538 	reg = hw->fc.pause_time * 0x00010001;
539 	for (i = 0; i < (IXGBE_DCB_MAX_TRAFFIC_CLASS / 2); i++)
540 		IXGBE_WRITE_REG(hw, IXGBE_FCTTV(i), reg);
541 
542 	/* Configure flow control refresh threshold value */
543 	IXGBE_WRITE_REG(hw, IXGBE_FCRTV, hw->fc.pause_time / 2);
544 
545 out:
546 	return ret_val;
547 }
548 
549 /**
550  *  ixgbe_start_mac_link_82598 - Configures MAC link settings
551  *  @hw: pointer to hardware structure
552  *  @autoneg_wait_to_complete: TRUE when waiting for completion is needed
553  *
554  *  Configures link settings based on values in the ixgbe_hw struct.
555  *  Restarts the link.  Performs autonegotiation if needed.
556  **/
557 static s32 ixgbe_start_mac_link_82598(struct ixgbe_hw *hw,
558 				      bool autoneg_wait_to_complete)
559 {
560 	u32 autoc_reg;
561 	u32 links_reg;
562 	u32 i;
563 	s32 status = IXGBE_SUCCESS;
564 
565 	DEBUGFUNC("ixgbe_start_mac_link_82598");
566 
567 	/* Restart link */
568 	autoc_reg = IXGBE_READ_REG(hw, IXGBE_AUTOC);
569 	autoc_reg |= IXGBE_AUTOC_AN_RESTART;
570 	IXGBE_WRITE_REG(hw, IXGBE_AUTOC, autoc_reg);
571 
572 	/* Only poll for autoneg to complete if specified to do so */
573 	if (autoneg_wait_to_complete) {
574 		if ((autoc_reg & IXGBE_AUTOC_LMS_MASK) ==
575 		     IXGBE_AUTOC_LMS_KX4_AN ||
576 		    (autoc_reg & IXGBE_AUTOC_LMS_MASK) ==
577 		     IXGBE_AUTOC_LMS_KX4_AN_1G_AN) {
578 			links_reg = 0; /* Just in case Autoneg time = 0 */
579 			for (i = 0; i < IXGBE_AUTO_NEG_TIME; i++) {
580 				links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
581 				if (links_reg & IXGBE_LINKS_KX_AN_COMP)
582 					break;
583 				msec_delay(100);
584 			}
585 			if (!(links_reg & IXGBE_LINKS_KX_AN_COMP)) {
586 				status = IXGBE_ERR_AUTONEG_NOT_COMPLETE;
587 				DEBUGOUT("Autonegotiation did not complete.\n");
588 			}
589 		}
590 	}
591 
592 	/* Add delay to filter out noises during initial link setup */
593 	msec_delay(50);
594 
595 	return status;
596 }
597 
598 /**
599  *  ixgbe_validate_link_ready - Function looks for phy link
600  *  @hw: pointer to hardware structure
601  *
602  *  Function indicates success when phy link is available. If phy is not ready
603  *  within 5 seconds of MAC indicating link, the function returns error.
604  **/
605 static s32 ixgbe_validate_link_ready(struct ixgbe_hw *hw)
606 {
607 	u32 timeout;
608 	u16 an_reg;
609 
610 	if (hw->device_id != IXGBE_DEV_ID_82598AT2)
611 		return IXGBE_SUCCESS;
612 
613 	for (timeout = 0;
614 	     timeout < IXGBE_VALIDATE_LINK_READY_TIMEOUT; timeout++) {
615 		hw->phy.ops.read_reg(hw, IXGBE_MDIO_AUTO_NEG_STATUS,
616 				     IXGBE_MDIO_AUTO_NEG_DEV_TYPE, &an_reg);
617 
618 		if ((an_reg & IXGBE_MII_AUTONEG_COMPLETE) &&
619 		    (an_reg & IXGBE_MII_AUTONEG_LINK_UP))
620 			break;
621 
622 		msec_delay(100);
623 	}
624 
625 	if (timeout == IXGBE_VALIDATE_LINK_READY_TIMEOUT) {
626 		DEBUGOUT("Link was indicated but link is down\n");
627 		return IXGBE_ERR_LINK_SETUP;
628 	}
629 
630 	return IXGBE_SUCCESS;
631 }
632 
633 /**
634  *  ixgbe_check_mac_link_82598 - Get link/speed status
635  *  @hw: pointer to hardware structure
636  *  @speed: pointer to link speed
637  *  @link_up: TRUE is link is up, FALSE otherwise
638  *  @link_up_wait_to_complete: bool used to wait for link up or not
639  *
640  *  Reads the links register to determine if link is up and the current speed
641  **/
642 static s32 ixgbe_check_mac_link_82598(struct ixgbe_hw *hw,
643 				      ixgbe_link_speed *speed, bool *link_up,
644 				      bool link_up_wait_to_complete)
645 {
646 	u32 links_reg;
647 	u32 i;
648 	u16 link_reg, adapt_comp_reg;
649 
650 	DEBUGFUNC("ixgbe_check_mac_link_82598");
651 
652 	/*
653 	 * SERDES PHY requires us to read link status from undocumented
654 	 * register 0xC79F.  Bit 0 set indicates link is up/ready; clear
655 	 * indicates link down.  OxC00C is read to check that the XAUI lanes
656 	 * are active.  Bit 0 clear indicates active; set indicates inactive.
657 	 */
658 	if (hw->phy.type == ixgbe_phy_nl) {
659 		hw->phy.ops.read_reg(hw, 0xC79F, IXGBE_TWINAX_DEV, &link_reg);
660 		hw->phy.ops.read_reg(hw, 0xC79F, IXGBE_TWINAX_DEV, &link_reg);
661 		hw->phy.ops.read_reg(hw, 0xC00C, IXGBE_TWINAX_DEV,
662 				     &adapt_comp_reg);
663 		if (link_up_wait_to_complete) {
664 			for (i = 0; i < hw->mac.max_link_up_time; i++) {
665 				if ((link_reg & 1) &&
666 				    ((adapt_comp_reg & 1) == 0)) {
667 					*link_up = TRUE;
668 					break;
669 				} else {
670 					*link_up = FALSE;
671 				}
672 				msec_delay(100);
673 				hw->phy.ops.read_reg(hw, 0xC79F,
674 						     IXGBE_TWINAX_DEV,
675 						     &link_reg);
676 				hw->phy.ops.read_reg(hw, 0xC00C,
677 						     IXGBE_TWINAX_DEV,
678 						     &adapt_comp_reg);
679 			}
680 		} else {
681 			if ((link_reg & 1) && ((adapt_comp_reg & 1) == 0))
682 				*link_up = TRUE;
683 			else
684 				*link_up = FALSE;
685 		}
686 
687 		if (*link_up == FALSE)
688 			goto out;
689 	}
690 
691 	links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
692 	if (link_up_wait_to_complete) {
693 		for (i = 0; i < hw->mac.max_link_up_time; i++) {
694 			if (links_reg & IXGBE_LINKS_UP) {
695 				*link_up = TRUE;
696 				break;
697 			} else {
698 				*link_up = FALSE;
699 			}
700 			msec_delay(100);
701 			links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
702 		}
703 	} else {
704 		if (links_reg & IXGBE_LINKS_UP)
705 			*link_up = TRUE;
706 		else
707 			*link_up = FALSE;
708 	}
709 
710 	if (links_reg & IXGBE_LINKS_SPEED)
711 		*speed = IXGBE_LINK_SPEED_10GB_FULL;
712 	else
713 		*speed = IXGBE_LINK_SPEED_1GB_FULL;
714 
715 	if ((hw->device_id == IXGBE_DEV_ID_82598AT2) && (*link_up == TRUE) &&
716 	    (ixgbe_validate_link_ready(hw) != IXGBE_SUCCESS))
717 		*link_up = FALSE;
718 
719 out:
720 	return IXGBE_SUCCESS;
721 }
722 
723 /**
724  *  ixgbe_setup_mac_link_82598 - Set MAC link speed
725  *  @hw: pointer to hardware structure
726  *  @speed: new link speed
727  *  @autoneg_wait_to_complete: TRUE when waiting for completion is needed
728  *
729  *  Set the link speed in the AUTOC register and restarts link.
730  **/
731 static s32 ixgbe_setup_mac_link_82598(struct ixgbe_hw *hw,
732 				      ixgbe_link_speed speed,
733 				      bool autoneg_wait_to_complete)
734 {
735 	bool autoneg = FALSE;
736 	s32 status = IXGBE_SUCCESS;
737 	ixgbe_link_speed link_capabilities = IXGBE_LINK_SPEED_UNKNOWN;
738 	u32 curr_autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
739 	u32 autoc = curr_autoc;
740 	u32 link_mode = autoc & IXGBE_AUTOC_LMS_MASK;
741 
742 	DEBUGFUNC("ixgbe_setup_mac_link_82598");
743 
744 	/* Check to see if speed passed in is supported. */
745 	ixgbe_get_link_capabilities(hw, &link_capabilities, &autoneg);
746 	speed &= link_capabilities;
747 
748 	if (speed == IXGBE_LINK_SPEED_UNKNOWN)
749 		status = IXGBE_ERR_LINK_SETUP;
750 
751 	/* Set KX4/KX support according to speed requested */
752 	else if (link_mode == IXGBE_AUTOC_LMS_KX4_AN ||
753 		 link_mode == IXGBE_AUTOC_LMS_KX4_AN_1G_AN) {
754 		autoc &= ~IXGBE_AUTOC_KX4_KX_SUPP_MASK;
755 		if (speed & IXGBE_LINK_SPEED_10GB_FULL)
756 			autoc |= IXGBE_AUTOC_KX4_SUPP;
757 		if (speed & IXGBE_LINK_SPEED_1GB_FULL)
758 			autoc |= IXGBE_AUTOC_KX_SUPP;
759 		if (autoc != curr_autoc)
760 			IXGBE_WRITE_REG(hw, IXGBE_AUTOC, autoc);
761 	}
762 
763 	if (status == IXGBE_SUCCESS) {
764 		/*
765 		 * Setup and restart the link based on the new values in
766 		 * ixgbe_hw This will write the AUTOC register based on the new
767 		 * stored values
768 		 */
769 		status = ixgbe_start_mac_link_82598(hw,
770 						    autoneg_wait_to_complete);
771 	}
772 
773 	return status;
774 }
775 
776 
777 /**
778  *  ixgbe_setup_copper_link_82598 - Set the PHY autoneg advertised field
779  *  @hw: pointer to hardware structure
780  *  @speed: new link speed
781  *  @autoneg_wait_to_complete: TRUE if waiting is needed to complete
782  *
783  *  Sets the link speed in the AUTOC register in the MAC and restarts link.
784  **/
785 static s32 ixgbe_setup_copper_link_82598(struct ixgbe_hw *hw,
786 					 ixgbe_link_speed speed,
787 					 bool autoneg_wait_to_complete)
788 {
789 	s32 status;
790 
791 	DEBUGFUNC("ixgbe_setup_copper_link_82598");
792 
793 	/* Setup the PHY according to input speed */
794 	status = hw->phy.ops.setup_link_speed(hw, speed,
795 					      autoneg_wait_to_complete);
796 	/* Set up MAC */
797 	ixgbe_start_mac_link_82598(hw, autoneg_wait_to_complete);
798 
799 	return status;
800 }
801 
802 /**
803  *  ixgbe_reset_hw_82598 - Performs hardware reset
804  *  @hw: pointer to hardware structure
805  *
806  *  Resets the hardware by resetting the transmit and receive units, masks and
807  *  clears all interrupts, performing a PHY reset, and performing a link (MAC)
808  *  reset.
809  **/
810 static s32 ixgbe_reset_hw_82598(struct ixgbe_hw *hw)
811 {
812 	s32 status = IXGBE_SUCCESS;
813 	s32 phy_status = IXGBE_SUCCESS;
814 	u32 ctrl;
815 	u32 gheccr;
816 	u32 i;
817 	u32 autoc;
818 	u8  analog_val;
819 
820 	DEBUGFUNC("ixgbe_reset_hw_82598");
821 
822 	/* Call adapter stop to disable tx/rx and clear interrupts */
823 	status = hw->mac.ops.stop_adapter(hw);
824 	if (status != IXGBE_SUCCESS)
825 		goto reset_hw_out;
826 
827 	/*
828 	 * Power up the Atlas Tx lanes if they are currently powered down.
829 	 * Atlas Tx lanes are powered down for MAC loopback tests, but
830 	 * they are not automatically restored on reset.
831 	 */
832 	hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK, &analog_val);
833 	if (analog_val & IXGBE_ATLAS_PDN_TX_REG_EN) {
834 		/* Enable Tx Atlas so packets can be transmitted again */
835 		hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK,
836 					     &analog_val);
837 		analog_val &= ~IXGBE_ATLAS_PDN_TX_REG_EN;
838 		hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK,
839 					      analog_val);
840 
841 		hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_10G,
842 					     &analog_val);
843 		analog_val &= ~IXGBE_ATLAS_PDN_TX_10G_QL_ALL;
844 		hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_10G,
845 					      analog_val);
846 
847 		hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_1G,
848 					     &analog_val);
849 		analog_val &= ~IXGBE_ATLAS_PDN_TX_1G_QL_ALL;
850 		hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_1G,
851 					      analog_val);
852 
853 		hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_AN,
854 					     &analog_val);
855 		analog_val &= ~IXGBE_ATLAS_PDN_TX_AN_QL_ALL;
856 		hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_AN,
857 					      analog_val);
858 	}
859 
860 	/* Reset PHY */
861 	if (hw->phy.reset_disable == FALSE) {
862 		/* PHY ops must be identified and initialized prior to reset */
863 
864 		/* Init PHY and function pointers, perform SFP setup */
865 		phy_status = hw->phy.ops.init(hw);
866 		if (phy_status == IXGBE_ERR_SFP_NOT_SUPPORTED)
867 			goto reset_hw_out;
868 		if (phy_status == IXGBE_ERR_SFP_NOT_PRESENT)
869 			goto mac_reset_top;
870 
871 		hw->phy.ops.reset(hw);
872 	}
873 
874 mac_reset_top:
875 	/*
876 	 * Issue global reset to the MAC.  This needs to be a SW reset.
877 	 * If link reset is used, it might reset the MAC when mng is using it
878 	 */
879 	ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL) | IXGBE_CTRL_RST;
880 	IXGBE_WRITE_REG(hw, IXGBE_CTRL, ctrl);
881 	IXGBE_WRITE_FLUSH(hw);
882 
883 	/* Poll for reset bit to self-clear indicating reset is complete */
884 	for (i = 0; i < 10; i++) {
885 		usec_delay(1);
886 		ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL);
887 		if (!(ctrl & IXGBE_CTRL_RST))
888 			break;
889 	}
890 	if (ctrl & IXGBE_CTRL_RST) {
891 		status = IXGBE_ERR_RESET_FAILED;
892 		DEBUGOUT("Reset polling failed to complete.\n");
893 	}
894 
895 	msec_delay(50);
896 
897 	/*
898 	 * Double resets are required for recovery from certain error
899 	 * conditions.  Between resets, it is necessary to stall to allow time
900 	 * for any pending HW events to complete.
901 	 */
902 	if (hw->mac.flags & IXGBE_FLAGS_DOUBLE_RESET_REQUIRED) {
903 		hw->mac.flags &= ~IXGBE_FLAGS_DOUBLE_RESET_REQUIRED;
904 		goto mac_reset_top;
905 	}
906 
907 	gheccr = IXGBE_READ_REG(hw, IXGBE_GHECCR);
908 	gheccr &= ~((1 << 21) | (1 << 18) | (1 << 9) | (1 << 6));
909 	IXGBE_WRITE_REG(hw, IXGBE_GHECCR, gheccr);
910 
911 	/*
912 	 * Store the original AUTOC value if it has not been
913 	 * stored off yet.  Otherwise restore the stored original
914 	 * AUTOC value since the reset operation sets back to deaults.
915 	 */
916 	autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
917 	if (hw->mac.orig_link_settings_stored == FALSE) {
918 		hw->mac.orig_autoc = autoc;
919 		hw->mac.orig_link_settings_stored = TRUE;
920 	} else if (autoc != hw->mac.orig_autoc) {
921 		IXGBE_WRITE_REG(hw, IXGBE_AUTOC, hw->mac.orig_autoc);
922 	}
923 
924 	/* Store the permanent mac address */
925 	hw->mac.ops.get_mac_addr(hw, hw->mac.perm_addr);
926 
927 	/*
928 	 * Store MAC address from RAR0, clear receive address registers, and
929 	 * clear the multicast table
930 	 */
931 	hw->mac.ops.init_rx_addrs(hw);
932 
933 reset_hw_out:
934 	if (phy_status != IXGBE_SUCCESS)
935 		status = phy_status;
936 
937 	return status;
938 }
939 
940 /**
941  *  ixgbe_set_vmdq_82598 - Associate a VMDq set index with a rx address
942  *  @hw: pointer to hardware struct
943  *  @rar: receive address register index to associate with a VMDq index
944  *  @vmdq: VMDq set index
945  **/
946 s32 ixgbe_set_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq)
947 {
948 	u32 rar_high;
949 	u32 rar_entries = hw->mac.num_rar_entries;
950 
951 	DEBUGFUNC("ixgbe_set_vmdq_82598");
952 
953 	/* Make sure we are using a valid rar index range */
954 	if (rar >= rar_entries) {
955 		DEBUGOUT1("RAR index %d is out of range.\n", rar);
956 		return IXGBE_ERR_INVALID_ARGUMENT;
957 	}
958 
959 	rar_high = IXGBE_READ_REG(hw, IXGBE_RAH(rar));
960 	rar_high &= ~IXGBE_RAH_VIND_MASK;
961 	rar_high |= ((vmdq << IXGBE_RAH_VIND_SHIFT) & IXGBE_RAH_VIND_MASK);
962 	IXGBE_WRITE_REG(hw, IXGBE_RAH(rar), rar_high);
963 	return IXGBE_SUCCESS;
964 }
965 
966 /**
967  *  ixgbe_clear_vmdq_82598 - Disassociate a VMDq set index from an rx address
968  *  @hw: pointer to hardware struct
969  *  @rar: receive address register index to associate with a VMDq index
970  *  @vmdq: VMDq clear index (not used in 82598, but elsewhere)
971  **/
972 static s32 ixgbe_clear_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq)
973 {
974 	u32 rar_high;
975 	u32 rar_entries = hw->mac.num_rar_entries;
976 
977 	UNREFERENCED_1PARAMETER(vmdq);
978 
979 	/* Make sure we are using a valid rar index range */
980 	if (rar >= rar_entries) {
981 		DEBUGOUT1("RAR index %d is out of range.\n", rar);
982 		return IXGBE_ERR_INVALID_ARGUMENT;
983 	}
984 
985 	rar_high = IXGBE_READ_REG(hw, IXGBE_RAH(rar));
986 	if (rar_high & IXGBE_RAH_VIND_MASK) {
987 		rar_high &= ~IXGBE_RAH_VIND_MASK;
988 		IXGBE_WRITE_REG(hw, IXGBE_RAH(rar), rar_high);
989 	}
990 
991 	return IXGBE_SUCCESS;
992 }
993 
994 /**
995  *  ixgbe_set_vfta_82598 - Set VLAN filter table
996  *  @hw: pointer to hardware structure
997  *  @vlan: VLAN id to write to VLAN filter
998  *  @vind: VMDq output index that maps queue to VLAN id in VFTA
999  *  @vlan_on: boolean flag to turn on/off VLAN in VFTA
1000  *  @vlvf_bypass: boolean flag - unused
1001  *
1002  *  Turn on/off specified VLAN in the VLAN filter table.
1003  **/
1004 s32 ixgbe_set_vfta_82598(struct ixgbe_hw *hw, u32 vlan, u32 vind,
1005 			 bool vlan_on, bool vlvf_bypass)
1006 {
1007 	u32 regindex;
1008 	u32 bitindex;
1009 	u32 bits;
1010 	u32 vftabyte;
1011 
1012 	UNREFERENCED_1PARAMETER(vlvf_bypass);
1013 
1014 	DEBUGFUNC("ixgbe_set_vfta_82598");
1015 
1016 	if (vlan > 4095)
1017 		return IXGBE_ERR_PARAM;
1018 
1019 	/* Determine 32-bit word position in array */
1020 	regindex = (vlan >> 5) & 0x7F;   /* upper seven bits */
1021 
1022 	/* Determine the location of the (VMD) queue index */
1023 	vftabyte =  ((vlan >> 3) & 0x03); /* bits (4:3) indicating byte array */
1024 	bitindex = (vlan & 0x7) << 2;    /* lower 3 bits indicate nibble */
1025 
1026 	/* Set the nibble for VMD queue index */
1027 	bits = IXGBE_READ_REG(hw, IXGBE_VFTAVIND(vftabyte, regindex));
1028 	bits &= (~(0x0F << bitindex));
1029 	bits |= (vind << bitindex);
1030 	IXGBE_WRITE_REG(hw, IXGBE_VFTAVIND(vftabyte, regindex), bits);
1031 
1032 	/* Determine the location of the bit for this VLAN id */
1033 	bitindex = vlan & 0x1F;   /* lower five bits */
1034 
1035 	bits = IXGBE_READ_REG(hw, IXGBE_VFTA(regindex));
1036 	if (vlan_on)
1037 		/* Turn on this VLAN id */
1038 		bits |= (1 << bitindex);
1039 	else
1040 		/* Turn off this VLAN id */
1041 		bits &= ~(1 << bitindex);
1042 	IXGBE_WRITE_REG(hw, IXGBE_VFTA(regindex), bits);
1043 
1044 	return IXGBE_SUCCESS;
1045 }
1046 
1047 /**
1048  *  ixgbe_clear_vfta_82598 - Clear VLAN filter table
1049  *  @hw: pointer to hardware structure
1050  *
1051  *  Clears the VLAN filer table, and the VMDq index associated with the filter
1052  **/
1053 static s32 ixgbe_clear_vfta_82598(struct ixgbe_hw *hw)
1054 {
1055 	u32 offset;
1056 	u32 vlanbyte;
1057 
1058 	DEBUGFUNC("ixgbe_clear_vfta_82598");
1059 
1060 	for (offset = 0; offset < hw->mac.vft_size; offset++)
1061 		IXGBE_WRITE_REG(hw, IXGBE_VFTA(offset), 0);
1062 
1063 	for (vlanbyte = 0; vlanbyte < 4; vlanbyte++)
1064 		for (offset = 0; offset < hw->mac.vft_size; offset++)
1065 			IXGBE_WRITE_REG(hw, IXGBE_VFTAVIND(vlanbyte, offset),
1066 					0);
1067 
1068 	return IXGBE_SUCCESS;
1069 }
1070 
1071 /**
1072  *  ixgbe_read_analog_reg8_82598 - Reads 8 bit Atlas analog register
1073  *  @hw: pointer to hardware structure
1074  *  @reg: analog register to read
1075  *  @val: read value
1076  *
1077  *  Performs read operation to Atlas analog register specified.
1078  **/
1079 s32 ixgbe_read_analog_reg8_82598(struct ixgbe_hw *hw, u32 reg, u8 *val)
1080 {
1081 	u32  atlas_ctl;
1082 
1083 	DEBUGFUNC("ixgbe_read_analog_reg8_82598");
1084 
1085 	IXGBE_WRITE_REG(hw, IXGBE_ATLASCTL,
1086 			IXGBE_ATLASCTL_WRITE_CMD | (reg << 8));
1087 	IXGBE_WRITE_FLUSH(hw);
1088 	usec_delay(10);
1089 	atlas_ctl = IXGBE_READ_REG(hw, IXGBE_ATLASCTL);
1090 	*val = (u8)atlas_ctl;
1091 
1092 	return IXGBE_SUCCESS;
1093 }
1094 
1095 /**
1096  *  ixgbe_write_analog_reg8_82598 - Writes 8 bit Atlas analog register
1097  *  @hw: pointer to hardware structure
1098  *  @reg: atlas register to write
1099  *  @val: value to write
1100  *
1101  *  Performs write operation to Atlas analog register specified.
1102  **/
1103 s32 ixgbe_write_analog_reg8_82598(struct ixgbe_hw *hw, u32 reg, u8 val)
1104 {
1105 	u32  atlas_ctl;
1106 
1107 	DEBUGFUNC("ixgbe_write_analog_reg8_82598");
1108 
1109 	atlas_ctl = (reg << 8) | val;
1110 	IXGBE_WRITE_REG(hw, IXGBE_ATLASCTL, atlas_ctl);
1111 	IXGBE_WRITE_FLUSH(hw);
1112 	usec_delay(10);
1113 
1114 	return IXGBE_SUCCESS;
1115 }
1116 
1117 /**
1118  *  ixgbe_read_i2c_phy_82598 - Reads 8 bit word over I2C interface.
1119  *  @hw: pointer to hardware structure
1120  *  @dev_addr: address to read from
1121  *  @byte_offset: byte offset to read from dev_addr
1122  *  @eeprom_data: value read
1123  *
1124  *  Performs 8 byte read operation to SFP module's EEPROM over I2C interface.
1125  **/
1126 static s32 ixgbe_read_i2c_phy_82598(struct ixgbe_hw *hw, u8 dev_addr,
1127 				    u8 byte_offset, u8 *eeprom_data)
1128 {
1129 	s32 status = IXGBE_SUCCESS;
1130 	u16 sfp_addr = 0;
1131 	u16 sfp_data = 0;
1132 	u16 sfp_stat = 0;
1133 	u16 gssr;
1134 	u32 i;
1135 
1136 	DEBUGFUNC("ixgbe_read_i2c_phy_82598");
1137 
1138 	if (IXGBE_READ_REG(hw, IXGBE_STATUS) & IXGBE_STATUS_LAN_ID_1)
1139 		gssr = IXGBE_GSSR_PHY1_SM;
1140 	else
1141 		gssr = IXGBE_GSSR_PHY0_SM;
1142 
1143 	if (hw->mac.ops.acquire_swfw_sync(hw, gssr) != IXGBE_SUCCESS)
1144 		return IXGBE_ERR_SWFW_SYNC;
1145 
1146 	if (hw->phy.type == ixgbe_phy_nl) {
1147 		/*
1148 		 * NetLogic phy SDA/SCL registers are at addresses 0xC30A to
1149 		 * 0xC30D. These registers are used to talk to the SFP+
1150 		 * module's EEPROM through the SDA/SCL (I2C) interface.
1151 		 */
1152 		sfp_addr = (dev_addr << 8) + byte_offset;
1153 		sfp_addr = (sfp_addr | IXGBE_I2C_EEPROM_READ_MASK);
1154 		hw->phy.ops.write_reg_mdi(hw,
1155 					  IXGBE_MDIO_PMA_PMD_SDA_SCL_ADDR,
1156 					  IXGBE_MDIO_PMA_PMD_DEV_TYPE,
1157 					  sfp_addr);
1158 
1159 		/* Poll status */
1160 		for (i = 0; i < 100; i++) {
1161 			hw->phy.ops.read_reg_mdi(hw,
1162 						IXGBE_MDIO_PMA_PMD_SDA_SCL_STAT,
1163 						IXGBE_MDIO_PMA_PMD_DEV_TYPE,
1164 						&sfp_stat);
1165 			sfp_stat = sfp_stat & IXGBE_I2C_EEPROM_STATUS_MASK;
1166 			if (sfp_stat != IXGBE_I2C_EEPROM_STATUS_IN_PROGRESS)
1167 				break;
1168 			msec_delay(10);
1169 		}
1170 
1171 		if (sfp_stat != IXGBE_I2C_EEPROM_STATUS_PASS) {
1172 			DEBUGOUT("EEPROM read did not pass.\n");
1173 			status = IXGBE_ERR_SFP_NOT_PRESENT;
1174 			goto out;
1175 		}
1176 
1177 		/* Read data */
1178 		hw->phy.ops.read_reg_mdi(hw, IXGBE_MDIO_PMA_PMD_SDA_SCL_DATA,
1179 					IXGBE_MDIO_PMA_PMD_DEV_TYPE, &sfp_data);
1180 
1181 		*eeprom_data = (u8)(sfp_data >> 8);
1182 	} else {
1183 		status = IXGBE_ERR_PHY;
1184 	}
1185 
1186 out:
1187 	hw->mac.ops.release_swfw_sync(hw, gssr);
1188 	return status;
1189 }
1190 
1191 /**
1192  *  ixgbe_read_i2c_eeprom_82598 - Reads 8 bit word over I2C interface.
1193  *  @hw: pointer to hardware structure
1194  *  @byte_offset: EEPROM byte offset to read
1195  *  @eeprom_data: value read
1196  *
1197  *  Performs 8 byte read operation to SFP module's EEPROM over I2C interface.
1198  **/
1199 s32 ixgbe_read_i2c_eeprom_82598(struct ixgbe_hw *hw, u8 byte_offset,
1200 				u8 *eeprom_data)
1201 {
1202 	return ixgbe_read_i2c_phy_82598(hw, IXGBE_I2C_EEPROM_DEV_ADDR,
1203 					byte_offset, eeprom_data);
1204 }
1205 
1206 /**
1207  *  ixgbe_read_i2c_sff8472_82598 - Reads 8 bit word over I2C interface.
1208  *  @hw: pointer to hardware structure
1209  *  @byte_offset: byte offset at address 0xA2
1210  *  @sff8472_data: value read
1211  *
1212  *  Performs 8 byte read operation to SFP module's SFF-8472 data over I2C
1213  **/
1214 static s32 ixgbe_read_i2c_sff8472_82598(struct ixgbe_hw *hw, u8 byte_offset,
1215 					u8 *sff8472_data)
1216 {
1217 	return ixgbe_read_i2c_phy_82598(hw, IXGBE_I2C_EEPROM_DEV_ADDR2,
1218 					byte_offset, sff8472_data);
1219 }
1220 
1221 /**
1222  *  ixgbe_get_supported_physical_layer_82598 - Returns physical layer type
1223  *  @hw: pointer to hardware structure
1224  *
1225  *  Determines physical layer capabilities of the current configuration.
1226  **/
1227 u64 ixgbe_get_supported_physical_layer_82598(struct ixgbe_hw *hw)
1228 {
1229 	u64 physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1230 	u32 autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
1231 	u32 pma_pmd_10g = autoc & IXGBE_AUTOC_10G_PMA_PMD_MASK;
1232 	u32 pma_pmd_1g = autoc & IXGBE_AUTOC_1G_PMA_PMD_MASK;
1233 	u16 ext_ability = 0;
1234 
1235 	DEBUGFUNC("ixgbe_get_supported_physical_layer_82598");
1236 
1237 	hw->phy.ops.identify(hw);
1238 
1239 	/* Copper PHY must be checked before AUTOC LMS to determine correct
1240 	 * physical layer because 10GBase-T PHYs use LMS = KX4/KX */
1241 	switch (hw->phy.type) {
1242 	case ixgbe_phy_tn:
1243 	case ixgbe_phy_cu_unknown:
1244 		hw->phy.ops.read_reg(hw, IXGBE_MDIO_PHY_EXT_ABILITY,
1245 		IXGBE_MDIO_PMA_PMD_DEV_TYPE, &ext_ability);
1246 		if (ext_ability & IXGBE_MDIO_PHY_10GBASET_ABILITY)
1247 			physical_layer |= IXGBE_PHYSICAL_LAYER_10GBASE_T;
1248 		if (ext_ability & IXGBE_MDIO_PHY_1000BASET_ABILITY)
1249 			physical_layer |= IXGBE_PHYSICAL_LAYER_1000BASE_T;
1250 		if (ext_ability & IXGBE_MDIO_PHY_100BASETX_ABILITY)
1251 			physical_layer |= IXGBE_PHYSICAL_LAYER_100BASE_TX;
1252 		goto out;
1253 	default:
1254 		break;
1255 	}
1256 
1257 	switch (autoc & IXGBE_AUTOC_LMS_MASK) {
1258 	case IXGBE_AUTOC_LMS_1G_AN:
1259 	case IXGBE_AUTOC_LMS_1G_LINK_NO_AN:
1260 		if (pma_pmd_1g == IXGBE_AUTOC_1G_KX)
1261 			physical_layer = IXGBE_PHYSICAL_LAYER_1000BASE_KX;
1262 		else
1263 			physical_layer = IXGBE_PHYSICAL_LAYER_1000BASE_BX;
1264 		break;
1265 	case IXGBE_AUTOC_LMS_10G_LINK_NO_AN:
1266 		if (pma_pmd_10g == IXGBE_AUTOC_10G_CX4)
1267 			physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_CX4;
1268 		else if (pma_pmd_10g == IXGBE_AUTOC_10G_KX4)
1269 			physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_KX4;
1270 		else /* XAUI */
1271 			physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1272 		break;
1273 	case IXGBE_AUTOC_LMS_KX4_AN:
1274 	case IXGBE_AUTOC_LMS_KX4_AN_1G_AN:
1275 		if (autoc & IXGBE_AUTOC_KX_SUPP)
1276 			physical_layer |= IXGBE_PHYSICAL_LAYER_1000BASE_KX;
1277 		if (autoc & IXGBE_AUTOC_KX4_SUPP)
1278 			physical_layer |= IXGBE_PHYSICAL_LAYER_10GBASE_KX4;
1279 		break;
1280 	default:
1281 		break;
1282 	}
1283 
1284 	if (hw->phy.type == ixgbe_phy_nl) {
1285 		hw->phy.ops.identify_sfp(hw);
1286 
1287 		switch (hw->phy.sfp_type) {
1288 		case ixgbe_sfp_type_da_cu:
1289 			physical_layer = IXGBE_PHYSICAL_LAYER_SFP_PLUS_CU;
1290 			break;
1291 		case ixgbe_sfp_type_sr:
1292 			physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_SR;
1293 			break;
1294 		case ixgbe_sfp_type_lr:
1295 			physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_LR;
1296 			break;
1297 		default:
1298 			physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1299 			break;
1300 		}
1301 	}
1302 
1303 	switch (hw->device_id) {
1304 	case IXGBE_DEV_ID_82598_DA_DUAL_PORT:
1305 		physical_layer = IXGBE_PHYSICAL_LAYER_SFP_PLUS_CU;
1306 		break;
1307 	case IXGBE_DEV_ID_82598AF_DUAL_PORT:
1308 	case IXGBE_DEV_ID_82598AF_SINGLE_PORT:
1309 	case IXGBE_DEV_ID_82598_SR_DUAL_PORT_EM:
1310 		physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_SR;
1311 		break;
1312 	case IXGBE_DEV_ID_82598EB_XF_LR:
1313 		physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_LR;
1314 		break;
1315 	default:
1316 		break;
1317 	}
1318 
1319 out:
1320 	return physical_layer;
1321 }
1322 
1323 /**
1324  *  ixgbe_set_lan_id_multi_port_pcie_82598 - Set LAN id for PCIe multiple
1325  *  port devices.
1326  *  @hw: pointer to the HW structure
1327  *
1328  *  Calls common function and corrects issue with some single port devices
1329  *  that enable LAN1 but not LAN0.
1330  **/
1331 void ixgbe_set_lan_id_multi_port_pcie_82598(struct ixgbe_hw *hw)
1332 {
1333 	struct ixgbe_bus_info *bus = &hw->bus;
1334 	u16 pci_gen = 0;
1335 	u16 pci_ctrl2 = 0;
1336 
1337 	DEBUGFUNC("ixgbe_set_lan_id_multi_port_pcie_82598");
1338 
1339 	ixgbe_set_lan_id_multi_port_pcie(hw);
1340 
1341 	/* check if LAN0 is disabled */
1342 	hw->eeprom.ops.read(hw, IXGBE_PCIE_GENERAL_PTR, &pci_gen);
1343 	if ((pci_gen != 0) && (pci_gen != 0xFFFF)) {
1344 
1345 		hw->eeprom.ops.read(hw, pci_gen + IXGBE_PCIE_CTRL2, &pci_ctrl2);
1346 
1347 		/* if LAN0 is completely disabled force function to 0 */
1348 		if ((pci_ctrl2 & IXGBE_PCIE_CTRL2_LAN_DISABLE) &&
1349 		    !(pci_ctrl2 & IXGBE_PCIE_CTRL2_DISABLE_SELECT) &&
1350 		    !(pci_ctrl2 & IXGBE_PCIE_CTRL2_DUMMY_ENABLE)) {
1351 
1352 			bus->func = 0;
1353 		}
1354 	}
1355 }
1356 
1357 /**
1358  *  ixgbe_enable_relaxed_ordering_82598 - enable relaxed ordering
1359  *  @hw: pointer to hardware structure
1360  *
1361  **/
1362 void ixgbe_enable_relaxed_ordering_82598(struct ixgbe_hw *hw)
1363 {
1364 	u32 regval;
1365 	u32 i;
1366 
1367 	DEBUGFUNC("ixgbe_enable_relaxed_ordering_82598");
1368 
1369 	/* Enable relaxed ordering */
1370 	for (i = 0; ((i < hw->mac.max_tx_queues) &&
1371 	     (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
1372 		regval = IXGBE_READ_REG(hw, IXGBE_DCA_TXCTRL(i));
1373 		regval |= IXGBE_DCA_TXCTRL_DESC_WRO_EN;
1374 		IXGBE_WRITE_REG(hw, IXGBE_DCA_TXCTRL(i), regval);
1375 	}
1376 
1377 	for (i = 0; ((i < hw->mac.max_rx_queues) &&
1378 	     (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
1379 		regval = IXGBE_READ_REG(hw, IXGBE_DCA_RXCTRL(i));
1380 		regval |= IXGBE_DCA_RXCTRL_DATA_WRO_EN |
1381 			  IXGBE_DCA_RXCTRL_HEAD_WRO_EN;
1382 		IXGBE_WRITE_REG(hw, IXGBE_DCA_RXCTRL(i), regval);
1383 	}
1384 
1385 }
1386 
1387 /**
1388  * ixgbe_set_rxpba_82598 - Initialize RX packet buffer
1389  * @hw: pointer to hardware structure
1390  * @num_pb: number of packet buffers to allocate
1391  * @headroom: reserve n KB of headroom
1392  * @strategy: packet buffer allocation strategy
1393  **/
1394 static void ixgbe_set_rxpba_82598(struct ixgbe_hw *hw, int num_pb,
1395 				  u32 headroom, int strategy)
1396 {
1397 	u32 rxpktsize = IXGBE_RXPBSIZE_64KB;
1398 	u8 i = 0;
1399 	UNREFERENCED_1PARAMETER(headroom);
1400 
1401 	if (!num_pb)
1402 		return;
1403 
1404 	/* Setup Rx packet buffer sizes */
1405 	switch (strategy) {
1406 	case PBA_STRATEGY_WEIGHTED:
1407 		/* Setup the first four at 80KB */
1408 		rxpktsize = IXGBE_RXPBSIZE_80KB;
1409 		for (; i < 4; i++)
1410 			IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), rxpktsize);
1411 		/* Setup the last four at 48KB...don't re-init i */
1412 		rxpktsize = IXGBE_RXPBSIZE_48KB;
1413 		/* Fall Through */
1414 	case PBA_STRATEGY_EQUAL:
1415 	default:
1416 		/* Divide the remaining Rx packet buffer evenly among the TCs */
1417 		for (; i < IXGBE_MAX_PACKET_BUFFERS; i++)
1418 			IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), rxpktsize);
1419 		break;
1420 	}
1421 
1422 	/* Setup Tx packet buffer sizes */
1423 	for (i = 0; i < IXGBE_MAX_PACKET_BUFFERS; i++)
1424 		IXGBE_WRITE_REG(hw, IXGBE_TXPBSIZE(i), IXGBE_TXPBSIZE_40KB);
1425 }
1426 
1427 /**
1428  *  ixgbe_enable_rx_dma_82598 - Enable the Rx DMA unit
1429  *  @hw: pointer to hardware structure
1430  *  @regval: register value to write to RXCTRL
1431  *
1432  *  Enables the Rx DMA unit
1433  **/
1434 s32 ixgbe_enable_rx_dma_82598(struct ixgbe_hw *hw, u32 regval)
1435 {
1436 	DEBUGFUNC("ixgbe_enable_rx_dma_82598");
1437 
1438 	IXGBE_WRITE_REG(hw, IXGBE_RXCTRL, regval);
1439 
1440 	return IXGBE_SUCCESS;
1441 }
1442