1 /******************************************************************************
2 
3   Copyright (c) 2001-2013, Intel Corporation
4   All rights reserved.
5 
6   Redistribution and use in source and binary forms, with or without
7   modification, are permitted provided that the following conditions are met:
8 
9    1. Redistributions of source code must retain the above copyright notice,
10       this list of conditions and the following disclaimer.
11 
12    2. Redistributions in binary form must reproduce the above copyright
13       notice, this list of conditions and the following disclaimer in the
14       documentation and/or other materials provided with the distribution.
15 
16    3. Neither the name of the Intel Corporation nor the names of its
17       contributors may be used to endorse or promote products derived from
18       this software without specific prior written permission.
19 
20   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21   AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22   IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23   ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24   LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25   CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26   SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27   INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28   CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30   POSSIBILITY OF SUCH DAMAGE.
31 
32 ******************************************************************************/
33 /*$FreeBSD: head/sys/dev/ixgbe/ixgbe_82598.c 251964 2013-06-18 21:28:19Z jfv $*/
34 /*$NetBSD: ixgbe_82598.c,v 1.5 2015/08/05 04:08:44 msaitoh Exp $*/
35 
36 #include "ixgbe_type.h"
37 #include "ixgbe_82598.h"
38 #include "ixgbe_api.h"
39 #include "ixgbe_common.h"
40 #include "ixgbe_phy.h"
41 
42 static s32 ixgbe_get_link_capabilities_82598(struct ixgbe_hw *hw,
43 					     ixgbe_link_speed *speed,
44 					     bool *autoneg);
45 static enum ixgbe_media_type ixgbe_get_media_type_82598(struct ixgbe_hw *hw);
46 static s32 ixgbe_start_mac_link_82598(struct ixgbe_hw *hw,
47 				      bool autoneg_wait_to_complete);
48 static s32 ixgbe_check_mac_link_82598(struct ixgbe_hw *hw,
49 				      ixgbe_link_speed *speed, bool *link_up,
50 				      bool link_up_wait_to_complete);
51 static s32 ixgbe_setup_mac_link_82598(struct ixgbe_hw *hw,
52 				      ixgbe_link_speed speed,
53 				      bool autoneg_wait_to_complete);
54 static s32 ixgbe_setup_copper_link_82598(struct ixgbe_hw *hw,
55 					 ixgbe_link_speed speed,
56 					 bool autoneg_wait_to_complete);
57 static s32 ixgbe_reset_hw_82598(struct ixgbe_hw *hw);
58 static s32 ixgbe_clear_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq);
59 static s32 ixgbe_clear_vfta_82598(struct ixgbe_hw *hw);
60 static void ixgbe_set_rxpba_82598(struct ixgbe_hw *hw, int num_pb,
61 				  u32 headroom, int strategy);
62 static s32 ixgbe_read_i2c_sff8472_82598(struct ixgbe_hw *hw, u8 byte_offset,
63 					u8 *sff8472_data);
64 /**
65  *  ixgbe_set_pcie_completion_timeout - set pci-e completion timeout
66  *  @hw: pointer to the HW structure
67  *
68  *  The defaults for 82598 should be in the range of 50us to 50ms,
69  *  however the hardware default for these parts is 500us to 1ms which is less
70  *  than the 10ms recommended by the pci-e spec.  To address this we need to
71  *  increase the value to either 10ms to 250ms for capability version 1 config,
72  *  or 16ms to 55ms for version 2.
73  **/
ixgbe_set_pcie_completion_timeout(struct ixgbe_hw * hw)74 void ixgbe_set_pcie_completion_timeout(struct ixgbe_hw *hw)
75 {
76 	u32 gcr = IXGBE_READ_REG(hw, IXGBE_GCR);
77 	u16 pcie_devctl2;
78 
79 	/* only take action if timeout value is defaulted to 0 */
80 	if (gcr & IXGBE_GCR_CMPL_TMOUT_MASK)
81 		goto out;
82 
83 	/*
84 	 * if capababilities version is type 1 we can write the
85 	 * timeout of 10ms to 250ms through the GCR register
86 	 */
87 	if (!(gcr & IXGBE_GCR_CAP_VER2)) {
88 		gcr |= IXGBE_GCR_CMPL_TMOUT_10ms;
89 		goto out;
90 	}
91 
92 	/*
93 	 * for version 2 capabilities we need to write the config space
94 	 * directly in order to set the completion timeout value for
95 	 * 16ms to 55ms
96 	 */
97 	pcie_devctl2 = IXGBE_READ_PCIE_WORD(hw, IXGBE_PCI_DEVICE_CONTROL2);
98 	pcie_devctl2 |= IXGBE_PCI_DEVICE_CONTROL2_16ms;
99 	IXGBE_WRITE_PCIE_WORD(hw, IXGBE_PCI_DEVICE_CONTROL2, pcie_devctl2);
100 out:
101 	/* disable completion timeout resend */
102 	gcr &= ~IXGBE_GCR_CMPL_TMOUT_RESEND;
103 	IXGBE_WRITE_REG(hw, IXGBE_GCR, gcr);
104 }
105 
106 /**
107  *  ixgbe_init_ops_82598 - Inits func ptrs and MAC type
108  *  @hw: pointer to hardware structure
109  *
110  *  Initialize the function pointers and assign the MAC type for 82598.
111  *  Does not touch the hardware.
112  **/
ixgbe_init_ops_82598(struct ixgbe_hw * hw)113 s32 ixgbe_init_ops_82598(struct ixgbe_hw *hw)
114 {
115 	struct ixgbe_mac_info *mac = &hw->mac;
116 	struct ixgbe_phy_info *phy = &hw->phy;
117 	s32 ret_val;
118 
119 	DEBUGFUNC("ixgbe_init_ops_82598");
120 
121 	ret_val = ixgbe_init_phy_ops_generic(hw);
122 	ret_val = ixgbe_init_ops_generic(hw);
123 
124 	/* PHY */
125 	phy->ops.init = &ixgbe_init_phy_ops_82598;
126 
127 	/* MAC */
128 	mac->ops.start_hw = &ixgbe_start_hw_82598;
129 	mac->ops.enable_relaxed_ordering = &ixgbe_enable_relaxed_ordering_82598;
130 	mac->ops.reset_hw = &ixgbe_reset_hw_82598;
131 	mac->ops.get_media_type = &ixgbe_get_media_type_82598;
132 	mac->ops.get_supported_physical_layer =
133 				&ixgbe_get_supported_physical_layer_82598;
134 	mac->ops.read_analog_reg8 = &ixgbe_read_analog_reg8_82598;
135 	mac->ops.write_analog_reg8 = &ixgbe_write_analog_reg8_82598;
136 	mac->ops.set_lan_id = &ixgbe_set_lan_id_multi_port_pcie_82598;
137 
138 	/* RAR, Multicast, VLAN */
139 	mac->ops.set_vmdq = &ixgbe_set_vmdq_82598;
140 	mac->ops.clear_vmdq = &ixgbe_clear_vmdq_82598;
141 	mac->ops.set_vfta = &ixgbe_set_vfta_82598;
142 	mac->ops.set_vlvf = NULL;
143 	mac->ops.clear_vfta = &ixgbe_clear_vfta_82598;
144 
145 	/* Flow Control */
146 	mac->ops.fc_enable = &ixgbe_fc_enable_82598;
147 
148 	mac->mcft_size		= 128;
149 	mac->vft_size		= 128;
150 	mac->num_rar_entries	= 16;
151 	mac->rx_pb_size		= 512;
152 	mac->max_tx_queues	= 32;
153 	mac->max_rx_queues	= 64;
154 	mac->max_msix_vectors	= ixgbe_get_pcie_msix_count_generic(hw);
155 
156 	/* SFP+ Module */
157 	phy->ops.read_i2c_eeprom = &ixgbe_read_i2c_eeprom_82598;
158 	phy->ops.read_i2c_sff8472 = &ixgbe_read_i2c_sff8472_82598;
159 
160 	/* Link */
161 	mac->ops.check_link = &ixgbe_check_mac_link_82598;
162 	mac->ops.setup_link = &ixgbe_setup_mac_link_82598;
163 	mac->ops.flap_tx_laser = NULL;
164 	mac->ops.get_link_capabilities = &ixgbe_get_link_capabilities_82598;
165 	mac->ops.setup_rxpba = &ixgbe_set_rxpba_82598;
166 
167 	/* Manageability interface */
168 	mac->ops.set_fw_drv_ver = NULL;
169 
170 	mac->ops.get_rtrup2tc = NULL;
171 
172 	return ret_val;
173 }
174 
175 /**
176  *  ixgbe_init_phy_ops_82598 - PHY/SFP specific init
177  *  @hw: pointer to hardware structure
178  *
179  *  Initialize any function pointers that were not able to be
180  *  set during init_shared_code because the PHY/SFP type was
181  *  not known.  Perform the SFP init if necessary.
182  *
183  **/
ixgbe_init_phy_ops_82598(struct ixgbe_hw * hw)184 s32 ixgbe_init_phy_ops_82598(struct ixgbe_hw *hw)
185 {
186 	struct ixgbe_mac_info *mac = &hw->mac;
187 	struct ixgbe_phy_info *phy = &hw->phy;
188 	s32 ret_val = IXGBE_SUCCESS;
189 	u16 list_offset, data_offset;
190 
191 	DEBUGFUNC("ixgbe_init_phy_ops_82598");
192 
193 	/* Identify the PHY */
194 	phy->ops.identify(hw);
195 
196 	/* Overwrite the link function pointers if copper PHY */
197 	if (mac->ops.get_media_type(hw) == ixgbe_media_type_copper) {
198 		mac->ops.setup_link = &ixgbe_setup_copper_link_82598;
199 		mac->ops.get_link_capabilities =
200 				&ixgbe_get_copper_link_capabilities_generic;
201 	}
202 
203 	switch (hw->phy.type) {
204 	case ixgbe_phy_tn:
205 		phy->ops.setup_link = &ixgbe_setup_phy_link_tnx;
206 		phy->ops.check_link = &ixgbe_check_phy_link_tnx;
207 		phy->ops.get_firmware_version =
208 					&ixgbe_get_phy_firmware_version_tnx;
209 		break;
210 	case ixgbe_phy_nl:
211 		phy->ops.reset = &ixgbe_reset_phy_nl;
212 
213 		/* Call SFP+ identify routine to get the SFP+ module type */
214 		ret_val = phy->ops.identify_sfp(hw);
215 		if (ret_val != IXGBE_SUCCESS)
216 			goto out;
217 		else if (hw->phy.sfp_type == ixgbe_sfp_type_unknown) {
218 			ret_val = IXGBE_ERR_SFP_NOT_SUPPORTED;
219 			goto out;
220 		}
221 
222 		/* Check to see if SFP+ module is supported */
223 		ret_val = ixgbe_get_sfp_init_sequence_offsets(hw,
224 							      &list_offset,
225 							      &data_offset);
226 		if (ret_val != IXGBE_SUCCESS) {
227 			ret_val = IXGBE_ERR_SFP_NOT_SUPPORTED;
228 			goto out;
229 		}
230 		break;
231 	default:
232 		break;
233 	}
234 
235 out:
236 	return ret_val;
237 }
238 
239 /**
240  *  ixgbe_start_hw_82598 - Prepare hardware for Tx/Rx
241  *  @hw: pointer to hardware structure
242  *
243  *  Starts the hardware using the generic start_hw function.
244  *  Disables relaxed ordering Then set pcie completion timeout
245  *
246  **/
ixgbe_start_hw_82598(struct ixgbe_hw * hw)247 s32 ixgbe_start_hw_82598(struct ixgbe_hw *hw)
248 {
249 	u32 regval;
250 	u32 i;
251 	s32 ret_val = IXGBE_SUCCESS;
252 
253 	DEBUGFUNC("ixgbe_start_hw_82598");
254 
255 	ret_val = ixgbe_start_hw_generic(hw);
256 
257 	/* Disable relaxed ordering */
258 	for (i = 0; ((i < hw->mac.max_tx_queues) &&
259 	     (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
260 		regval = IXGBE_READ_REG(hw, IXGBE_DCA_TXCTRL(i));
261 		regval &= ~IXGBE_DCA_TXCTRL_DESC_WRO_EN;
262 		IXGBE_WRITE_REG(hw, IXGBE_DCA_TXCTRL(i), regval);
263 	}
264 
265 	for (i = 0; ((i < hw->mac.max_rx_queues) &&
266 	     (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
267 		regval = IXGBE_READ_REG(hw, IXGBE_DCA_RXCTRL(i));
268 		regval &= ~(IXGBE_DCA_RXCTRL_DATA_WRO_EN |
269 			    IXGBE_DCA_RXCTRL_HEAD_WRO_EN);
270 		IXGBE_WRITE_REG(hw, IXGBE_DCA_RXCTRL(i), regval);
271 	}
272 
273 	/* set the completion timeout for interface */
274 	if (ret_val == IXGBE_SUCCESS)
275 		ixgbe_set_pcie_completion_timeout(hw);
276 
277 	return ret_val;
278 }
279 
280 /**
281  *  ixgbe_get_link_capabilities_82598 - Determines link capabilities
282  *  @hw: pointer to hardware structure
283  *  @speed: pointer to link speed
284  *  @autoneg: boolean auto-negotiation value
285  *
286  *  Determines the link capabilities by reading the AUTOC register.
287  **/
ixgbe_get_link_capabilities_82598(struct ixgbe_hw * hw,ixgbe_link_speed * speed,bool * autoneg)288 static s32 ixgbe_get_link_capabilities_82598(struct ixgbe_hw *hw,
289 					     ixgbe_link_speed *speed,
290 					     bool *autoneg)
291 {
292 	s32 status = IXGBE_SUCCESS;
293 	u32 autoc = 0;
294 
295 	DEBUGFUNC("ixgbe_get_link_capabilities_82598");
296 
297 	/*
298 	 * Determine link capabilities based on the stored value of AUTOC,
299 	 * which represents EEPROM defaults.  If AUTOC value has not been
300 	 * stored, use the current register value.
301 	 */
302 	if (hw->mac.orig_link_settings_stored)
303 		autoc = hw->mac.orig_autoc;
304 	else
305 		autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
306 
307 	switch (autoc & IXGBE_AUTOC_LMS_MASK) {
308 	case IXGBE_AUTOC_LMS_1G_LINK_NO_AN:
309 		*speed = IXGBE_LINK_SPEED_1GB_FULL;
310 		*autoneg = FALSE;
311 		break;
312 
313 	case IXGBE_AUTOC_LMS_10G_LINK_NO_AN:
314 		*speed = IXGBE_LINK_SPEED_10GB_FULL;
315 		*autoneg = FALSE;
316 		break;
317 
318 	case IXGBE_AUTOC_LMS_1G_AN:
319 		*speed = IXGBE_LINK_SPEED_1GB_FULL;
320 		*autoneg = TRUE;
321 		break;
322 
323 	case IXGBE_AUTOC_LMS_KX4_AN:
324 	case IXGBE_AUTOC_LMS_KX4_AN_1G_AN:
325 		*speed = IXGBE_LINK_SPEED_UNKNOWN;
326 		if (autoc & IXGBE_AUTOC_KX4_SUPP)
327 			*speed |= IXGBE_LINK_SPEED_10GB_FULL;
328 		if (autoc & IXGBE_AUTOC_KX_SUPP)
329 			*speed |= IXGBE_LINK_SPEED_1GB_FULL;
330 		*autoneg = TRUE;
331 		break;
332 
333 	default:
334 		status = IXGBE_ERR_LINK_SETUP;
335 		break;
336 	}
337 
338 	return status;
339 }
340 
341 /**
342  *  ixgbe_get_media_type_82598 - Determines media type
343  *  @hw: pointer to hardware structure
344  *
345  *  Returns the media type (fiber, copper, backplane)
346  **/
ixgbe_get_media_type_82598(struct ixgbe_hw * hw)347 static enum ixgbe_media_type ixgbe_get_media_type_82598(struct ixgbe_hw *hw)
348 {
349 	enum ixgbe_media_type media_type;
350 
351 	DEBUGFUNC("ixgbe_get_media_type_82598");
352 
353 	/* Detect if there is a copper PHY attached. */
354 	switch (hw->phy.type) {
355 	case ixgbe_phy_cu_unknown:
356 	case ixgbe_phy_tn:
357 		media_type = ixgbe_media_type_copper;
358 		goto out;
359 	default:
360 		break;
361 	}
362 
363 	/* Media type for I82598 is based on device ID */
364 	switch (hw->device_id) {
365 	case IXGBE_DEV_ID_82598:
366 	case IXGBE_DEV_ID_82598_BX:
367 		/* Default device ID is mezzanine card KX/KX4 */
368 		media_type = ixgbe_media_type_backplane;
369 		break;
370 	case IXGBE_DEV_ID_82598AF_DUAL_PORT:
371 	case IXGBE_DEV_ID_82598AF_SINGLE_PORT:
372 	case IXGBE_DEV_ID_82598_DA_DUAL_PORT:
373 	case IXGBE_DEV_ID_82598_SR_DUAL_PORT_EM:
374 	case IXGBE_DEV_ID_82598EB_XF_LR:
375 	case IXGBE_DEV_ID_82598EB_SFP_LOM:
376 		media_type = ixgbe_media_type_fiber;
377 		break;
378 	case IXGBE_DEV_ID_82598EB_CX4:
379 	case IXGBE_DEV_ID_82598_CX4_DUAL_PORT:
380 		media_type = ixgbe_media_type_cx4;
381 		break;
382 	case IXGBE_DEV_ID_82598AT:
383 	case IXGBE_DEV_ID_82598AT2:
384 		media_type = ixgbe_media_type_copper;
385 		break;
386 	default:
387 		media_type = ixgbe_media_type_unknown;
388 		break;
389 	}
390 out:
391 	return media_type;
392 }
393 
394 /**
395  *  ixgbe_fc_enable_82598 - Enable flow control
396  *  @hw: pointer to hardware structure
397  *
398  *  Enable flow control according to the current settings.
399  **/
ixgbe_fc_enable_82598(struct ixgbe_hw * hw)400 s32 ixgbe_fc_enable_82598(struct ixgbe_hw *hw)
401 {
402 	s32 ret_val = IXGBE_SUCCESS;
403 	u32 fctrl_reg;
404 	u32 rmcs_reg;
405 	u32 reg;
406 	u32 fcrtl, fcrth;
407 	u32 link_speed = 0;
408 	int i;
409 	bool link_up;
410 
411 	DEBUGFUNC("ixgbe_fc_enable_82598");
412 
413 	/* Validate the water mark configuration */
414 	if (!hw->fc.pause_time) {
415 		ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
416 		goto out;
417 	}
418 
419 	/* Low water mark of zero causes XOFF floods */
420 	for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
421 		if ((hw->fc.current_mode & ixgbe_fc_tx_pause) &&
422 		    hw->fc.high_water[i]) {
423 			if (!hw->fc.low_water[i] ||
424 			    hw->fc.low_water[i] >= hw->fc.high_water[i]) {
425 				DEBUGOUT("Invalid water mark configuration\n");
426 				ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
427 				goto out;
428 			}
429 		}
430 	}
431 
432 	/*
433 	 * On 82598 having Rx FC on causes resets while doing 1G
434 	 * so if it's on turn it off once we know link_speed. For
435 	 * more details see 82598 Specification update.
436 	 */
437 	hw->mac.ops.check_link(hw, &link_speed, &link_up, FALSE);
438 	if (link_up && link_speed == IXGBE_LINK_SPEED_1GB_FULL) {
439 		switch (hw->fc.requested_mode) {
440 		case ixgbe_fc_full:
441 			hw->fc.requested_mode = ixgbe_fc_tx_pause;
442 			break;
443 		case ixgbe_fc_rx_pause:
444 			hw->fc.requested_mode = ixgbe_fc_none;
445 			break;
446 		default:
447 			/* no change */
448 			break;
449 		}
450 	}
451 
452 	/* Negotiate the fc mode to use */
453 	ixgbe_fc_autoneg(hw);
454 
455 	/* Disable any previous flow control settings */
456 	fctrl_reg = IXGBE_READ_REG(hw, IXGBE_FCTRL);
457 	fctrl_reg &= ~(IXGBE_FCTRL_RFCE | IXGBE_FCTRL_RPFCE);
458 
459 	rmcs_reg = IXGBE_READ_REG(hw, IXGBE_RMCS);
460 	rmcs_reg &= ~(IXGBE_RMCS_TFCE_PRIORITY | IXGBE_RMCS_TFCE_802_3X);
461 
462 	/*
463 	 * The possible values of fc.current_mode are:
464 	 * 0: Flow control is completely disabled
465 	 * 1: Rx flow control is enabled (we can receive pause frames,
466 	 *    but not send pause frames).
467 	 * 2: Tx flow control is enabled (we can send pause frames but
468 	 *     we do not support receiving pause frames).
469 	 * 3: Both Rx and Tx flow control (symmetric) are enabled.
470 	 * other: Invalid.
471 	 */
472 	switch (hw->fc.current_mode) {
473 	case ixgbe_fc_none:
474 		/*
475 		 * Flow control is disabled by software override or autoneg.
476 		 * The code below will actually disable it in the HW.
477 		 */
478 		break;
479 	case ixgbe_fc_rx_pause:
480 		/*
481 		 * Rx Flow control is enabled and Tx Flow control is
482 		 * disabled by software override. Since there really
483 		 * isn't a way to advertise that we are capable of RX
484 		 * Pause ONLY, we will advertise that we support both
485 		 * symmetric and asymmetric Rx PAUSE.  Later, we will
486 		 * disable the adapter's ability to send PAUSE frames.
487 		 */
488 		fctrl_reg |= IXGBE_FCTRL_RFCE;
489 		break;
490 	case ixgbe_fc_tx_pause:
491 		/*
492 		 * Tx Flow control is enabled, and Rx Flow control is
493 		 * disabled by software override.
494 		 */
495 		rmcs_reg |= IXGBE_RMCS_TFCE_802_3X;
496 		break;
497 	case ixgbe_fc_full:
498 		/* Flow control (both Rx and Tx) is enabled by SW override. */
499 		fctrl_reg |= IXGBE_FCTRL_RFCE;
500 		rmcs_reg |= IXGBE_RMCS_TFCE_802_3X;
501 		break;
502 	default:
503 		DEBUGOUT("Flow control param set incorrectly\n");
504 		ret_val = IXGBE_ERR_CONFIG;
505 		goto out;
506 		break;
507 	}
508 
509 	/* Set 802.3x based flow control settings. */
510 	fctrl_reg |= IXGBE_FCTRL_DPF;
511 	IXGBE_WRITE_REG(hw, IXGBE_FCTRL, fctrl_reg);
512 	IXGBE_WRITE_REG(hw, IXGBE_RMCS, rmcs_reg);
513 
514 	/* Set up and enable Rx high/low water mark thresholds, enable XON. */
515 	for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
516 		if ((hw->fc.current_mode & ixgbe_fc_tx_pause) &&
517 		    hw->fc.high_water[i]) {
518 			fcrtl = (hw->fc.low_water[i] << 10) | IXGBE_FCRTL_XONE;
519 			fcrth = (hw->fc.high_water[i] << 10) | IXGBE_FCRTH_FCEN;
520 			IXGBE_WRITE_REG(hw, IXGBE_FCRTL(i), fcrtl);
521 			IXGBE_WRITE_REG(hw, IXGBE_FCRTH(i), fcrth);
522 		} else {
523 			IXGBE_WRITE_REG(hw, IXGBE_FCRTL(i), 0);
524 			IXGBE_WRITE_REG(hw, IXGBE_FCRTH(i), 0);
525 		}
526 
527 	}
528 
529 	/* Configure pause time (2 TCs per register) */
530 	reg = hw->fc.pause_time * 0x00010001;
531 	for (i = 0; i < (IXGBE_DCB_MAX_TRAFFIC_CLASS / 2); i++)
532 		IXGBE_WRITE_REG(hw, IXGBE_FCTTV(i), reg);
533 
534 	/* Configure flow control refresh threshold value */
535 	IXGBE_WRITE_REG(hw, IXGBE_FCRTV, hw->fc.pause_time / 2);
536 
537 out:
538 	return ret_val;
539 }
540 
541 /**
542  *  ixgbe_start_mac_link_82598 - Configures MAC link settings
543  *  @hw: pointer to hardware structure
544  *
545  *  Configures link settings based on values in the ixgbe_hw struct.
546  *  Restarts the link.  Performs autonegotiation if needed.
547  **/
ixgbe_start_mac_link_82598(struct ixgbe_hw * hw,bool autoneg_wait_to_complete)548 static s32 ixgbe_start_mac_link_82598(struct ixgbe_hw *hw,
549 				      bool autoneg_wait_to_complete)
550 {
551 	u32 autoc_reg;
552 	u32 links_reg;
553 	u32 i;
554 	s32 status = IXGBE_SUCCESS;
555 
556 	DEBUGFUNC("ixgbe_start_mac_link_82598");
557 
558 	/* Restart link */
559 	autoc_reg = IXGBE_READ_REG(hw, IXGBE_AUTOC);
560 	autoc_reg |= IXGBE_AUTOC_AN_RESTART;
561 	IXGBE_WRITE_REG(hw, IXGBE_AUTOC, autoc_reg);
562 
563 	/* Only poll for autoneg to complete if specified to do so */
564 	if (autoneg_wait_to_complete) {
565 		if ((autoc_reg & IXGBE_AUTOC_LMS_MASK) ==
566 		     IXGBE_AUTOC_LMS_KX4_AN ||
567 		    (autoc_reg & IXGBE_AUTOC_LMS_MASK) ==
568 		     IXGBE_AUTOC_LMS_KX4_AN_1G_AN) {
569 			links_reg = 0; /* Just in case Autoneg time = 0 */
570 			for (i = 0; i < IXGBE_AUTO_NEG_TIME; i++) {
571 				links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
572 				if (links_reg & IXGBE_LINKS_KX_AN_COMP)
573 					break;
574 				msec_delay(100);
575 			}
576 			if (!(links_reg & IXGBE_LINKS_KX_AN_COMP)) {
577 				status = IXGBE_ERR_AUTONEG_NOT_COMPLETE;
578 				DEBUGOUT("Autonegotiation did not complete.\n");
579 			}
580 		}
581 	}
582 
583 	/* Add delay to filter out noises during initial link setup */
584 	msec_delay(50);
585 
586 	return status;
587 }
588 
589 /**
590  *  ixgbe_validate_link_ready - Function looks for phy link
591  *  @hw: pointer to hardware structure
592  *
593  *  Function indicates success when phy link is available. If phy is not ready
594  *  within 5 seconds of MAC indicating link, the function returns error.
595  **/
ixgbe_validate_link_ready(struct ixgbe_hw * hw)596 static s32 ixgbe_validate_link_ready(struct ixgbe_hw *hw)
597 {
598 	u32 timeout;
599 	u16 an_reg;
600 
601 	if (hw->device_id != IXGBE_DEV_ID_82598AT2)
602 		return IXGBE_SUCCESS;
603 
604 	for (timeout = 0;
605 	     timeout < IXGBE_VALIDATE_LINK_READY_TIMEOUT; timeout++) {
606 		hw->phy.ops.read_reg(hw, IXGBE_MDIO_AUTO_NEG_STATUS,
607 				     IXGBE_MDIO_AUTO_NEG_DEV_TYPE, &an_reg);
608 
609 		if ((an_reg & IXGBE_MII_AUTONEG_COMPLETE) &&
610 		    (an_reg & IXGBE_MII_AUTONEG_LINK_UP))
611 			break;
612 
613 		msec_delay(100);
614 	}
615 
616 	if (timeout == IXGBE_VALIDATE_LINK_READY_TIMEOUT) {
617 		DEBUGOUT("Link was indicated but link is down\n");
618 		return IXGBE_ERR_LINK_SETUP;
619 	}
620 
621 	return IXGBE_SUCCESS;
622 }
623 
624 /**
625  *  ixgbe_check_mac_link_82598 - Get link/speed status
626  *  @hw: pointer to hardware structure
627  *  @speed: pointer to link speed
628  *  @link_up: TRUE is link is up, FALSE otherwise
629  *  @link_up_wait_to_complete: bool used to wait for link up or not
630  *
631  *  Reads the links register to determine if link is up and the current speed
632  **/
ixgbe_check_mac_link_82598(struct ixgbe_hw * hw,ixgbe_link_speed * speed,bool * link_up,bool link_up_wait_to_complete)633 static s32 ixgbe_check_mac_link_82598(struct ixgbe_hw *hw,
634 				      ixgbe_link_speed *speed, bool *link_up,
635 				      bool link_up_wait_to_complete)
636 {
637 	u32 links_reg;
638 	u32 i;
639 	u16 link_reg, adapt_comp_reg;
640 
641 	DEBUGFUNC("ixgbe_check_mac_link_82598");
642 
643 	/*
644 	 * SERDES PHY requires us to read link status from undocumented
645 	 * register 0xC79F.  Bit 0 set indicates link is up/ready; clear
646 	 * indicates link down.  OxC00C is read to check that the XAUI lanes
647 	 * are active.  Bit 0 clear indicates active; set indicates inactive.
648 	 */
649 	if (hw->phy.type == ixgbe_phy_nl) {
650 		hw->phy.ops.read_reg(hw, 0xC79F, IXGBE_TWINAX_DEV, &link_reg);
651 		hw->phy.ops.read_reg(hw, 0xC79F, IXGBE_TWINAX_DEV, &link_reg);
652 		hw->phy.ops.read_reg(hw, 0xC00C, IXGBE_TWINAX_DEV,
653 				     &adapt_comp_reg);
654 		if (link_up_wait_to_complete) {
655 			for (i = 0; i < IXGBE_LINK_UP_TIME; i++) {
656 				if ((link_reg & 1) &&
657 				    ((adapt_comp_reg & 1) == 0)) {
658 					*link_up = TRUE;
659 					break;
660 				} else {
661 					*link_up = FALSE;
662 				}
663 				msec_delay(100);
664 				hw->phy.ops.read_reg(hw, 0xC79F,
665 						     IXGBE_TWINAX_DEV,
666 						     &link_reg);
667 				hw->phy.ops.read_reg(hw, 0xC00C,
668 						     IXGBE_TWINAX_DEV,
669 						     &adapt_comp_reg);
670 			}
671 		} else {
672 			if ((link_reg & 1) && ((adapt_comp_reg & 1) == 0))
673 				*link_up = TRUE;
674 			else
675 				*link_up = FALSE;
676 		}
677 
678 		if (*link_up == FALSE)
679 			goto out;
680 	}
681 
682 	links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
683 	if (link_up_wait_to_complete) {
684 		for (i = 0; i < IXGBE_LINK_UP_TIME; i++) {
685 			if (links_reg & IXGBE_LINKS_UP) {
686 				*link_up = TRUE;
687 				break;
688 			} else {
689 				*link_up = FALSE;
690 			}
691 			msec_delay(100);
692 			links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
693 		}
694 	} else {
695 		if (links_reg & IXGBE_LINKS_UP)
696 			*link_up = TRUE;
697 		else
698 			*link_up = FALSE;
699 	}
700 
701 	if (links_reg & IXGBE_LINKS_SPEED)
702 		*speed = IXGBE_LINK_SPEED_10GB_FULL;
703 	else
704 		*speed = IXGBE_LINK_SPEED_1GB_FULL;
705 
706 	if ((hw->device_id == IXGBE_DEV_ID_82598AT2) && (*link_up == TRUE) &&
707 	    (ixgbe_validate_link_ready(hw) != IXGBE_SUCCESS))
708 		*link_up = FALSE;
709 
710 out:
711 	return IXGBE_SUCCESS;
712 }
713 
714 /**
715  *  ixgbe_setup_mac_link_82598 - Set MAC link speed
716  *  @hw: pointer to hardware structure
717  *  @speed: new link speed
718  *  @autoneg_wait_to_complete: TRUE when waiting for completion is needed
719  *
720  *  Set the link speed in the AUTOC register and restarts link.
721  **/
ixgbe_setup_mac_link_82598(struct ixgbe_hw * hw,ixgbe_link_speed speed,bool autoneg_wait_to_complete)722 static s32 ixgbe_setup_mac_link_82598(struct ixgbe_hw *hw,
723 				      ixgbe_link_speed speed,
724 				      bool autoneg_wait_to_complete)
725 {
726 	bool autoneg = FALSE;
727 	s32 status = IXGBE_SUCCESS;
728 	ixgbe_link_speed link_capabilities = IXGBE_LINK_SPEED_UNKNOWN;
729 	u32 curr_autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
730 	u32 autoc = curr_autoc;
731 	u32 link_mode = autoc & IXGBE_AUTOC_LMS_MASK;
732 
733 	DEBUGFUNC("ixgbe_setup_mac_link_82598");
734 
735 	/* Check to see if speed passed in is supported. */
736 	ixgbe_get_link_capabilities(hw, &link_capabilities, &autoneg);
737 	speed &= link_capabilities;
738 
739 	if (speed == IXGBE_LINK_SPEED_UNKNOWN)
740 		status = IXGBE_ERR_LINK_SETUP;
741 
742 	/* Set KX4/KX support according to speed requested */
743 	else if (link_mode == IXGBE_AUTOC_LMS_KX4_AN ||
744 		 link_mode == IXGBE_AUTOC_LMS_KX4_AN_1G_AN) {
745 		autoc &= ~IXGBE_AUTOC_KX4_KX_SUPP_MASK;
746 		if (speed & IXGBE_LINK_SPEED_10GB_FULL)
747 			autoc |= IXGBE_AUTOC_KX4_SUPP;
748 		if (speed & IXGBE_LINK_SPEED_1GB_FULL)
749 			autoc |= IXGBE_AUTOC_KX_SUPP;
750 		if (autoc != curr_autoc)
751 			IXGBE_WRITE_REG(hw, IXGBE_AUTOC, autoc);
752 	}
753 
754 	if (status == IXGBE_SUCCESS) {
755 		/*
756 		 * Setup and restart the link based on the new values in
757 		 * ixgbe_hw This will write the AUTOC register based on the new
758 		 * stored values
759 		 */
760 		status = ixgbe_start_mac_link_82598(hw,
761 						    autoneg_wait_to_complete);
762 	}
763 
764 	return status;
765 }
766 
767 
768 /**
769  *  ixgbe_setup_copper_link_82598 - Set the PHY autoneg advertised field
770  *  @hw: pointer to hardware structure
771  *  @speed: new link speed
772  *  @autoneg_wait_to_complete: TRUE if waiting is needed to complete
773  *
774  *  Sets the link speed in the AUTOC register in the MAC and restarts link.
775  **/
ixgbe_setup_copper_link_82598(struct ixgbe_hw * hw,ixgbe_link_speed speed,bool autoneg_wait_to_complete)776 static s32 ixgbe_setup_copper_link_82598(struct ixgbe_hw *hw,
777 					 ixgbe_link_speed speed,
778 					 bool autoneg_wait_to_complete)
779 {
780 	s32 status;
781 
782 	DEBUGFUNC("ixgbe_setup_copper_link_82598");
783 
784 	/* Setup the PHY according to input speed */
785 	status = hw->phy.ops.setup_link_speed(hw, speed,
786 					      autoneg_wait_to_complete);
787 	/* Set up MAC */
788 	ixgbe_start_mac_link_82598(hw, autoneg_wait_to_complete);
789 
790 	return status;
791 }
792 
793 /**
794  *  ixgbe_reset_hw_82598 - Performs hardware reset
795  *  @hw: pointer to hardware structure
796  *
797  *  Resets the hardware by resetting the transmit and receive units, masks and
798  *  clears all interrupts, performing a PHY reset, and performing a link (MAC)
799  *  reset.
800  **/
ixgbe_reset_hw_82598(struct ixgbe_hw * hw)801 static s32 ixgbe_reset_hw_82598(struct ixgbe_hw *hw)
802 {
803 	s32 status = IXGBE_SUCCESS;
804 	s32 phy_status = IXGBE_SUCCESS;
805 	u32 ctrl;
806 	u32 gheccr;
807 	u32 i;
808 	u32 autoc;
809 	u8  analog_val;
810 
811 	DEBUGFUNC("ixgbe_reset_hw_82598");
812 
813 	/* Call adapter stop to disable tx/rx and clear interrupts */
814 	status = hw->mac.ops.stop_adapter(hw);
815 	if (status != IXGBE_SUCCESS)
816 		goto reset_hw_out;
817 
818 	/*
819 	 * Power up the Atlas Tx lanes if they are currently powered down.
820 	 * Atlas Tx lanes are powered down for MAC loopback tests, but
821 	 * they are not automatically restored on reset.
822 	 */
823 	hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK, &analog_val);
824 	if (analog_val & IXGBE_ATLAS_PDN_TX_REG_EN) {
825 		/* Enable Tx Atlas so packets can be transmitted again */
826 		hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK,
827 					     &analog_val);
828 		analog_val &= ~IXGBE_ATLAS_PDN_TX_REG_EN;
829 		hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK,
830 					      analog_val);
831 
832 		hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_10G,
833 					     &analog_val);
834 		analog_val &= ~IXGBE_ATLAS_PDN_TX_10G_QL_ALL;
835 		hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_10G,
836 					      analog_val);
837 
838 		hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_1G,
839 					     &analog_val);
840 		analog_val &= ~IXGBE_ATLAS_PDN_TX_1G_QL_ALL;
841 		hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_1G,
842 					      analog_val);
843 
844 		hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_AN,
845 					     &analog_val);
846 		analog_val &= ~IXGBE_ATLAS_PDN_TX_AN_QL_ALL;
847 		hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_AN,
848 					      analog_val);
849 	}
850 
851 	/* Reset PHY */
852 	if (hw->phy.reset_disable == FALSE) {
853 		/* PHY ops must be identified and initialized prior to reset */
854 
855 		/* Init PHY and function pointers, perform SFP setup */
856 		phy_status = hw->phy.ops.init(hw);
857 		if (phy_status == IXGBE_ERR_SFP_NOT_SUPPORTED)
858 			goto reset_hw_out;
859 		if (phy_status == IXGBE_ERR_SFP_NOT_PRESENT)
860 			goto mac_reset_top;
861 
862 		hw->phy.ops.reset(hw);
863 	}
864 
865 mac_reset_top:
866 	/*
867 	 * Issue global reset to the MAC.  This needs to be a SW reset.
868 	 * If link reset is used, it might reset the MAC when mng is using it
869 	 */
870 	ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL) | IXGBE_CTRL_RST;
871 	IXGBE_WRITE_REG(hw, IXGBE_CTRL, ctrl);
872 	IXGBE_WRITE_FLUSH(hw);
873 
874 	/* Poll for reset bit to self-clear indicating reset is complete */
875 	for (i = 0; i < 10; i++) {
876 		usec_delay(1);
877 		ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL);
878 		if (!(ctrl & IXGBE_CTRL_RST))
879 			break;
880 	}
881 	if (ctrl & IXGBE_CTRL_RST) {
882 		status = IXGBE_ERR_RESET_FAILED;
883 		DEBUGOUT("Reset polling failed to complete.\n");
884 	}
885 
886 	msec_delay(50);
887 
888 	/*
889 	 * Double resets are required for recovery from certain error
890 	 * conditions.  Between resets, it is necessary to stall to allow time
891 	 * for any pending HW events to complete.
892 	 */
893 	if (hw->mac.flags & IXGBE_FLAGS_DOUBLE_RESET_REQUIRED) {
894 		hw->mac.flags &= ~IXGBE_FLAGS_DOUBLE_RESET_REQUIRED;
895 		goto mac_reset_top;
896 	}
897 
898 	gheccr = IXGBE_READ_REG(hw, IXGBE_GHECCR);
899 	gheccr &= ~((1 << 21) | (1 << 18) | (1 << 9) | (1 << 6));
900 	IXGBE_WRITE_REG(hw, IXGBE_GHECCR, gheccr);
901 
902 	/*
903 	 * Store the original AUTOC value if it has not been
904 	 * stored off yet.  Otherwise restore the stored original
905 	 * AUTOC value since the reset operation sets back to deaults.
906 	 */
907 	autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
908 	if (hw->mac.orig_link_settings_stored == FALSE) {
909 		hw->mac.orig_autoc = autoc;
910 		hw->mac.orig_link_settings_stored = TRUE;
911 	} else if (autoc != hw->mac.orig_autoc) {
912 		IXGBE_WRITE_REG(hw, IXGBE_AUTOC, hw->mac.orig_autoc);
913 	}
914 
915 	/* Store the permanent mac address */
916 	hw->mac.ops.get_mac_addr(hw, hw->mac.perm_addr);
917 
918 	/*
919 	 * Store MAC address from RAR0, clear receive address registers, and
920 	 * clear the multicast table
921 	 */
922 	hw->mac.ops.init_rx_addrs(hw);
923 
924 reset_hw_out:
925 	if (phy_status != IXGBE_SUCCESS)
926 		status = phy_status;
927 
928 	return status;
929 }
930 
931 /**
932  *  ixgbe_set_vmdq_82598 - Associate a VMDq set index with a rx address
933  *  @hw: pointer to hardware struct
934  *  @rar: receive address register index to associate with a VMDq index
935  *  @vmdq: VMDq set index
936  **/
ixgbe_set_vmdq_82598(struct ixgbe_hw * hw,u32 rar,u32 vmdq)937 s32 ixgbe_set_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq)
938 {
939 	u32 rar_high;
940 	u32 rar_entries = hw->mac.num_rar_entries;
941 
942 	DEBUGFUNC("ixgbe_set_vmdq_82598");
943 
944 	/* Make sure we are using a valid rar index range */
945 	if (rar >= rar_entries) {
946 		DEBUGOUT1("RAR index %d is out of range.\n", rar);
947 		return IXGBE_ERR_INVALID_ARGUMENT;
948 	}
949 
950 	rar_high = IXGBE_READ_REG(hw, IXGBE_RAH(rar));
951 	rar_high &= ~IXGBE_RAH_VIND_MASK;
952 	rar_high |= ((vmdq << IXGBE_RAH_VIND_SHIFT) & IXGBE_RAH_VIND_MASK);
953 	IXGBE_WRITE_REG(hw, IXGBE_RAH(rar), rar_high);
954 	return IXGBE_SUCCESS;
955 }
956 
957 /**
958  *  ixgbe_clear_vmdq_82598 - Disassociate a VMDq set index from an rx address
959  *  @hw: pointer to hardware struct
960  *  @rar: receive address register index to associate with a VMDq index
961  *  @vmdq: VMDq clear index (not used in 82598, but elsewhere)
962  **/
ixgbe_clear_vmdq_82598(struct ixgbe_hw * hw,u32 rar,u32 vmdq)963 static s32 ixgbe_clear_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq)
964 {
965 	u32 rar_high;
966 	u32 rar_entries = hw->mac.num_rar_entries;
967 
968 	UNREFERENCED_1PARAMETER(vmdq);
969 
970 	/* Make sure we are using a valid rar index range */
971 	if (rar >= rar_entries) {
972 		DEBUGOUT1("RAR index %d is out of range.\n", rar);
973 		return IXGBE_ERR_INVALID_ARGUMENT;
974 	}
975 
976 	rar_high = IXGBE_READ_REG(hw, IXGBE_RAH(rar));
977 	if (rar_high & IXGBE_RAH_VIND_MASK) {
978 		rar_high &= ~IXGBE_RAH_VIND_MASK;
979 		IXGBE_WRITE_REG(hw, IXGBE_RAH(rar), rar_high);
980 	}
981 
982 	return IXGBE_SUCCESS;
983 }
984 
985 /**
986  *  ixgbe_set_vfta_82598 - Set VLAN filter table
987  *  @hw: pointer to hardware structure
988  *  @vlan: VLAN id to write to VLAN filter
989  *  @vind: VMDq output index that maps queue to VLAN id in VFTA
990  *  @vlan_on: boolean flag to turn on/off VLAN in VFTA
991  *
992  *  Turn on/off specified VLAN in the VLAN filter table.
993  **/
ixgbe_set_vfta_82598(struct ixgbe_hw * hw,u32 vlan,u32 vind,bool vlan_on)994 s32 ixgbe_set_vfta_82598(struct ixgbe_hw *hw, u32 vlan, u32 vind,
995 			 bool vlan_on)
996 {
997 	u32 regindex;
998 	u32 bitindex;
999 	u32 bits;
1000 	u32 vftabyte;
1001 
1002 	DEBUGFUNC("ixgbe_set_vfta_82598");
1003 
1004 	if (vlan > 4095)
1005 		return IXGBE_ERR_PARAM;
1006 
1007 	/* Determine 32-bit word position in array */
1008 	regindex = (vlan >> 5) & 0x7F;   /* upper seven bits */
1009 
1010 	/* Determine the location of the (VMD) queue index */
1011 	vftabyte =  ((vlan >> 3) & 0x03); /* bits (4:3) indicating byte array */
1012 	bitindex = (vlan & 0x7) << 2;    /* lower 3 bits indicate nibble */
1013 
1014 	/* Set the nibble for VMD queue index */
1015 	bits = IXGBE_READ_REG(hw, IXGBE_VFTAVIND(vftabyte, regindex));
1016 	bits &= (~(0x0F << bitindex));
1017 	bits |= (vind << bitindex);
1018 	IXGBE_WRITE_REG(hw, IXGBE_VFTAVIND(vftabyte, regindex), bits);
1019 
1020 	/* Determine the location of the bit for this VLAN id */
1021 	bitindex = vlan & 0x1F;   /* lower five bits */
1022 
1023 	bits = IXGBE_READ_REG(hw, IXGBE_VFTA(regindex));
1024 	if (vlan_on)
1025 		/* Turn on this VLAN id */
1026 		bits |= (1 << bitindex);
1027 	else
1028 		/* Turn off this VLAN id */
1029 		bits &= ~(1 << bitindex);
1030 	IXGBE_WRITE_REG(hw, IXGBE_VFTA(regindex), bits);
1031 
1032 	return IXGBE_SUCCESS;
1033 }
1034 
1035 /**
1036  *  ixgbe_clear_vfta_82598 - Clear VLAN filter table
1037  *  @hw: pointer to hardware structure
1038  *
1039  *  Clears the VLAN filer table, and the VMDq index associated with the filter
1040  **/
ixgbe_clear_vfta_82598(struct ixgbe_hw * hw)1041 static s32 ixgbe_clear_vfta_82598(struct ixgbe_hw *hw)
1042 {
1043 	u32 offset;
1044 	u32 vlanbyte;
1045 
1046 	DEBUGFUNC("ixgbe_clear_vfta_82598");
1047 
1048 	for (offset = 0; offset < hw->mac.vft_size; offset++)
1049 		IXGBE_WRITE_REG(hw, IXGBE_VFTA(offset), 0);
1050 
1051 	for (vlanbyte = 0; vlanbyte < 4; vlanbyte++)
1052 		for (offset = 0; offset < hw->mac.vft_size; offset++)
1053 			IXGBE_WRITE_REG(hw, IXGBE_VFTAVIND(vlanbyte, offset),
1054 					0);
1055 
1056 	return IXGBE_SUCCESS;
1057 }
1058 
1059 /**
1060  *  ixgbe_read_analog_reg8_82598 - Reads 8 bit Atlas analog register
1061  *  @hw: pointer to hardware structure
1062  *  @reg: analog register to read
1063  *  @val: read value
1064  *
1065  *  Performs read operation to Atlas analog register specified.
1066  **/
ixgbe_read_analog_reg8_82598(struct ixgbe_hw * hw,u32 reg,u8 * val)1067 s32 ixgbe_read_analog_reg8_82598(struct ixgbe_hw *hw, u32 reg, u8 *val)
1068 {
1069 	u32  atlas_ctl;
1070 
1071 	DEBUGFUNC("ixgbe_read_analog_reg8_82598");
1072 
1073 	IXGBE_WRITE_REG(hw, IXGBE_ATLASCTL,
1074 			IXGBE_ATLASCTL_WRITE_CMD | (reg << 8));
1075 	IXGBE_WRITE_FLUSH(hw);
1076 	usec_delay(10);
1077 	atlas_ctl = IXGBE_READ_REG(hw, IXGBE_ATLASCTL);
1078 	*val = (u8)atlas_ctl;
1079 
1080 	return IXGBE_SUCCESS;
1081 }
1082 
1083 /**
1084  *  ixgbe_write_analog_reg8_82598 - Writes 8 bit Atlas analog register
1085  *  @hw: pointer to hardware structure
1086  *  @reg: atlas register to write
1087  *  @val: value to write
1088  *
1089  *  Performs write operation to Atlas analog register specified.
1090  **/
ixgbe_write_analog_reg8_82598(struct ixgbe_hw * hw,u32 reg,u8 val)1091 s32 ixgbe_write_analog_reg8_82598(struct ixgbe_hw *hw, u32 reg, u8 val)
1092 {
1093 	u32  atlas_ctl;
1094 
1095 	DEBUGFUNC("ixgbe_write_analog_reg8_82598");
1096 
1097 	atlas_ctl = (reg << 8) | val;
1098 	IXGBE_WRITE_REG(hw, IXGBE_ATLASCTL, atlas_ctl);
1099 	IXGBE_WRITE_FLUSH(hw);
1100 	usec_delay(10);
1101 
1102 	return IXGBE_SUCCESS;
1103 }
1104 
1105 /**
1106  *  ixgbe_read_i2c_phy_82598 - Reads 8 bit word over I2C interface.
1107  *  @hw: pointer to hardware structure
1108  *  @dev_addr: address to read from
1109  *  @byte_offset: byte offset to read from dev_addr
1110  *  @eeprom_data: value read
1111  *
1112  *  Performs 8 byte read operation to SFP module's EEPROM over I2C interface.
1113  **/
ixgbe_read_i2c_phy_82598(struct ixgbe_hw * hw,u8 dev_addr,u8 byte_offset,u8 * eeprom_data)1114 static s32 ixgbe_read_i2c_phy_82598(struct ixgbe_hw *hw, u8 dev_addr,
1115 				    u8 byte_offset, u8 *eeprom_data)
1116 {
1117 	s32 status = IXGBE_SUCCESS;
1118 	u16 sfp_addr = 0;
1119 	u16 sfp_data = 0;
1120 	u16 sfp_stat = 0;
1121 	u16 gssr;
1122 	u32 i;
1123 
1124 	DEBUGFUNC("ixgbe_read_i2c_phy_82598");
1125 
1126 	if (IXGBE_READ_REG(hw, IXGBE_STATUS) & IXGBE_STATUS_LAN_ID_1)
1127 		gssr = IXGBE_GSSR_PHY1_SM;
1128 	else
1129 		gssr = IXGBE_GSSR_PHY0_SM;
1130 
1131 	if (hw->mac.ops.acquire_swfw_sync(hw, gssr) != IXGBE_SUCCESS)
1132 		return IXGBE_ERR_SWFW_SYNC;
1133 
1134 	if (hw->phy.type == ixgbe_phy_nl) {
1135 		/*
1136 		 * NetLogic phy SDA/SCL registers are at addresses 0xC30A to
1137 		 * 0xC30D. These registers are used to talk to the SFP+
1138 		 * module's EEPROM through the SDA/SCL (I2C) interface.
1139 		 */
1140 		sfp_addr = (dev_addr << 8) + byte_offset;
1141 		sfp_addr = (sfp_addr | IXGBE_I2C_EEPROM_READ_MASK);
1142 		hw->phy.ops.write_reg_mdi(hw,
1143 					  IXGBE_MDIO_PMA_PMD_SDA_SCL_ADDR,
1144 					  IXGBE_MDIO_PMA_PMD_DEV_TYPE,
1145 					  sfp_addr);
1146 
1147 		/* Poll status */
1148 		for (i = 0; i < 100; i++) {
1149 			hw->phy.ops.read_reg_mdi(hw,
1150 						IXGBE_MDIO_PMA_PMD_SDA_SCL_STAT,
1151 						IXGBE_MDIO_PMA_PMD_DEV_TYPE,
1152 						&sfp_stat);
1153 			sfp_stat = sfp_stat & IXGBE_I2C_EEPROM_STATUS_MASK;
1154 			if (sfp_stat != IXGBE_I2C_EEPROM_STATUS_IN_PROGRESS)
1155 				break;
1156 			msec_delay(10);
1157 		}
1158 
1159 		if (sfp_stat != IXGBE_I2C_EEPROM_STATUS_PASS) {
1160 			DEBUGOUT("EEPROM read did not pass.\n");
1161 			status = IXGBE_ERR_SFP_NOT_PRESENT;
1162 			goto out;
1163 		}
1164 
1165 		/* Read data */
1166 		hw->phy.ops.read_reg_mdi(hw, IXGBE_MDIO_PMA_PMD_SDA_SCL_DATA,
1167 					IXGBE_MDIO_PMA_PMD_DEV_TYPE, &sfp_data);
1168 
1169 		*eeprom_data = (u8)(sfp_data >> 8);
1170 	} else {
1171 		status = IXGBE_ERR_PHY;
1172 	}
1173 
1174 out:
1175 	hw->mac.ops.release_swfw_sync(hw, gssr);
1176 	return status;
1177 }
1178 
1179 /**
1180  *  ixgbe_read_i2c_eeprom_82598 - Reads 8 bit word over I2C interface.
1181  *  @hw: pointer to hardware structure
1182  *  @byte_offset: EEPROM byte offset to read
1183  *  @eeprom_data: value read
1184  *
1185  *  Performs 8 byte read operation to SFP module's EEPROM over I2C interface.
1186  **/
ixgbe_read_i2c_eeprom_82598(struct ixgbe_hw * hw,u8 byte_offset,u8 * eeprom_data)1187 s32 ixgbe_read_i2c_eeprom_82598(struct ixgbe_hw *hw, u8 byte_offset,
1188 				u8 *eeprom_data)
1189 {
1190 	return ixgbe_read_i2c_phy_82598(hw, IXGBE_I2C_EEPROM_DEV_ADDR,
1191 					byte_offset, eeprom_data);
1192 }
1193 
1194 /**
1195  *  ixgbe_read_i2c_sff8472_82598 - Reads 8 bit word over I2C interface.
1196  *  @hw: pointer to hardware structure
1197  *  @byte_offset: byte offset at address 0xA2
1198  *  @eeprom_data: value read
1199  *
1200  *  Performs 8 byte read operation to SFP module's SFF-8472 data over I2C
1201  **/
ixgbe_read_i2c_sff8472_82598(struct ixgbe_hw * hw,u8 byte_offset,u8 * sff8472_data)1202 static s32 ixgbe_read_i2c_sff8472_82598(struct ixgbe_hw *hw, u8 byte_offset,
1203 					u8 *sff8472_data)
1204 {
1205 	return ixgbe_read_i2c_phy_82598(hw, IXGBE_I2C_EEPROM_DEV_ADDR2,
1206 					byte_offset, sff8472_data);
1207 }
1208 
1209 /**
1210  *  ixgbe_get_supported_physical_layer_82598 - Returns physical layer type
1211  *  @hw: pointer to hardware structure
1212  *
1213  *  Determines physical layer capabilities of the current configuration.
1214  **/
ixgbe_get_supported_physical_layer_82598(struct ixgbe_hw * hw)1215 u32 ixgbe_get_supported_physical_layer_82598(struct ixgbe_hw *hw)
1216 {
1217 	u32 physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1218 	u32 autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
1219 	u32 pma_pmd_10g = autoc & IXGBE_AUTOC_10G_PMA_PMD_MASK;
1220 	u32 pma_pmd_1g = autoc & IXGBE_AUTOC_1G_PMA_PMD_MASK;
1221 	u16 ext_ability = 0;
1222 
1223 	DEBUGFUNC("ixgbe_get_supported_physical_layer_82598");
1224 
1225 	hw->phy.ops.identify(hw);
1226 
1227 	/* Copper PHY must be checked before AUTOC LMS to determine correct
1228 	 * physical layer because 10GBase-T PHYs use LMS = KX4/KX */
1229 	switch (hw->phy.type) {
1230 	case ixgbe_phy_tn:
1231 	case ixgbe_phy_cu_unknown:
1232 		hw->phy.ops.read_reg(hw, IXGBE_MDIO_PHY_EXT_ABILITY,
1233 		IXGBE_MDIO_PMA_PMD_DEV_TYPE, &ext_ability);
1234 		if (ext_ability & IXGBE_MDIO_PHY_10GBASET_ABILITY)
1235 			physical_layer |= IXGBE_PHYSICAL_LAYER_10GBASE_T;
1236 		if (ext_ability & IXGBE_MDIO_PHY_1000BASET_ABILITY)
1237 			physical_layer |= IXGBE_PHYSICAL_LAYER_1000BASE_T;
1238 		if (ext_ability & IXGBE_MDIO_PHY_100BASETX_ABILITY)
1239 			physical_layer |= IXGBE_PHYSICAL_LAYER_100BASE_TX;
1240 		goto out;
1241 	default:
1242 		break;
1243 	}
1244 
1245 	switch (autoc & IXGBE_AUTOC_LMS_MASK) {
1246 	case IXGBE_AUTOC_LMS_1G_AN:
1247 	case IXGBE_AUTOC_LMS_1G_LINK_NO_AN:
1248 		if (pma_pmd_1g == IXGBE_AUTOC_1G_KX)
1249 			physical_layer = IXGBE_PHYSICAL_LAYER_1000BASE_KX;
1250 		else
1251 			physical_layer = IXGBE_PHYSICAL_LAYER_1000BASE_BX;
1252 		break;
1253 	case IXGBE_AUTOC_LMS_10G_LINK_NO_AN:
1254 		if (pma_pmd_10g == IXGBE_AUTOC_10G_CX4)
1255 			physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_CX4;
1256 		else if (pma_pmd_10g == IXGBE_AUTOC_10G_KX4)
1257 			physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_KX4;
1258 		else /* XAUI */
1259 			physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1260 		break;
1261 	case IXGBE_AUTOC_LMS_KX4_AN:
1262 	case IXGBE_AUTOC_LMS_KX4_AN_1G_AN:
1263 		if (autoc & IXGBE_AUTOC_KX_SUPP)
1264 			physical_layer |= IXGBE_PHYSICAL_LAYER_1000BASE_KX;
1265 		if (autoc & IXGBE_AUTOC_KX4_SUPP)
1266 			physical_layer |= IXGBE_PHYSICAL_LAYER_10GBASE_KX4;
1267 		break;
1268 	default:
1269 		break;
1270 	}
1271 
1272 	if (hw->phy.type == ixgbe_phy_nl) {
1273 		hw->phy.ops.identify_sfp(hw);
1274 
1275 		switch (hw->phy.sfp_type) {
1276 		case ixgbe_sfp_type_da_cu:
1277 			physical_layer = IXGBE_PHYSICAL_LAYER_SFP_PLUS_CU;
1278 			break;
1279 		case ixgbe_sfp_type_sr:
1280 			physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_SR;
1281 			break;
1282 		case ixgbe_sfp_type_lr:
1283 			physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_LR;
1284 			break;
1285 		default:
1286 			physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1287 			break;
1288 		}
1289 	}
1290 
1291 	switch (hw->device_id) {
1292 	case IXGBE_DEV_ID_82598_DA_DUAL_PORT:
1293 		physical_layer = IXGBE_PHYSICAL_LAYER_SFP_PLUS_CU;
1294 		break;
1295 	case IXGBE_DEV_ID_82598AF_DUAL_PORT:
1296 	case IXGBE_DEV_ID_82598AF_SINGLE_PORT:
1297 	case IXGBE_DEV_ID_82598_SR_DUAL_PORT_EM:
1298 		physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_SR;
1299 		break;
1300 	case IXGBE_DEV_ID_82598EB_XF_LR:
1301 		physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_LR;
1302 		break;
1303 	default:
1304 		break;
1305 	}
1306 
1307 out:
1308 	return physical_layer;
1309 }
1310 
1311 /**
1312  *  ixgbe_set_lan_id_multi_port_pcie_82598 - Set LAN id for PCIe multiple
1313  *  port devices.
1314  *  @hw: pointer to the HW structure
1315  *
1316  *  Calls common function and corrects issue with some single port devices
1317  *  that enable LAN1 but not LAN0.
1318  **/
ixgbe_set_lan_id_multi_port_pcie_82598(struct ixgbe_hw * hw)1319 void ixgbe_set_lan_id_multi_port_pcie_82598(struct ixgbe_hw *hw)
1320 {
1321 	struct ixgbe_bus_info *bus = &hw->bus;
1322 	u16 pci_gen = 0;
1323 	u16 pci_ctrl2 = 0;
1324 
1325 	DEBUGFUNC("ixgbe_set_lan_id_multi_port_pcie_82598");
1326 
1327 	ixgbe_set_lan_id_multi_port_pcie(hw);
1328 
1329 	/* check if LAN0 is disabled */
1330 	hw->eeprom.ops.read(hw, IXGBE_PCIE_GENERAL_PTR, &pci_gen);
1331 	if ((pci_gen != 0) && (pci_gen != 0xFFFF)) {
1332 
1333 		hw->eeprom.ops.read(hw, pci_gen + IXGBE_PCIE_CTRL2, &pci_ctrl2);
1334 
1335 		/* if LAN0 is completely disabled force function to 0 */
1336 		if ((pci_ctrl2 & IXGBE_PCIE_CTRL2_LAN_DISABLE) &&
1337 		    !(pci_ctrl2 & IXGBE_PCIE_CTRL2_DISABLE_SELECT) &&
1338 		    !(pci_ctrl2 & IXGBE_PCIE_CTRL2_DUMMY_ENABLE)) {
1339 
1340 			bus->func = 0;
1341 		}
1342 	}
1343 }
1344 
1345 /**
1346  *  ixgbe_enable_relaxed_ordering_82598 - enable relaxed ordering
1347  *  @hw: pointer to hardware structure
1348  *
1349  **/
ixgbe_enable_relaxed_ordering_82598(struct ixgbe_hw * hw)1350 void ixgbe_enable_relaxed_ordering_82598(struct ixgbe_hw *hw)
1351 {
1352 	u32 regval;
1353 	u32 i;
1354 
1355 	DEBUGFUNC("ixgbe_enable_relaxed_ordering_82598");
1356 
1357 	/* Enable relaxed ordering */
1358 	for (i = 0; ((i < hw->mac.max_tx_queues) &&
1359 	     (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
1360 		regval = IXGBE_READ_REG(hw, IXGBE_DCA_TXCTRL(i));
1361 		regval |= IXGBE_DCA_TXCTRL_DESC_WRO_EN;
1362 		IXGBE_WRITE_REG(hw, IXGBE_DCA_TXCTRL(i), regval);
1363 	}
1364 
1365 	for (i = 0; ((i < hw->mac.max_rx_queues) &&
1366 	     (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
1367 		regval = IXGBE_READ_REG(hw, IXGBE_DCA_RXCTRL(i));
1368 		regval |= IXGBE_DCA_RXCTRL_DATA_WRO_EN |
1369 			  IXGBE_DCA_RXCTRL_HEAD_WRO_EN;
1370 		IXGBE_WRITE_REG(hw, IXGBE_DCA_RXCTRL(i), regval);
1371 	}
1372 
1373 }
1374 
1375 /**
1376  * ixgbe_set_rxpba_82598 - Initialize RX packet buffer
1377  * @hw: pointer to hardware structure
1378  * @num_pb: number of packet buffers to allocate
1379  * @headroom: reserve n KB of headroom
1380  * @strategy: packet buffer allocation strategy
1381  **/
ixgbe_set_rxpba_82598(struct ixgbe_hw * hw,int num_pb,u32 headroom,int strategy)1382 static void ixgbe_set_rxpba_82598(struct ixgbe_hw *hw, int num_pb,
1383 				  u32 headroom, int strategy)
1384 {
1385 	u32 rxpktsize = IXGBE_RXPBSIZE_64KB;
1386 	u8 i = 0;
1387 	UNREFERENCED_1PARAMETER(headroom);
1388 
1389 	if (!num_pb)
1390 		return;
1391 
1392 	/* Setup Rx packet buffer sizes */
1393 	switch (strategy) {
1394 	case PBA_STRATEGY_WEIGHTED:
1395 		/* Setup the first four at 80KB */
1396 		rxpktsize = IXGBE_RXPBSIZE_80KB;
1397 		for (; i < 4; i++)
1398 			IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), rxpktsize);
1399 		/* Setup the last four at 48KB...don't re-init i */
1400 		rxpktsize = IXGBE_RXPBSIZE_48KB;
1401 		/* Fall Through */
1402 	case PBA_STRATEGY_EQUAL:
1403 	default:
1404 		/* Divide the remaining Rx packet buffer evenly among the TCs */
1405 		for (; i < IXGBE_MAX_PACKET_BUFFERS; i++)
1406 			IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), rxpktsize);
1407 		break;
1408 	}
1409 
1410 	/* Setup Tx packet buffer sizes */
1411 	for (i = 0; i < IXGBE_MAX_PACKET_BUFFERS; i++)
1412 		IXGBE_WRITE_REG(hw, IXGBE_TXPBSIZE(i), IXGBE_TXPBSIZE_40KB);
1413 
1414 	return;
1415 }
1416