xref: /freebsd/sys/dev/cxgbe/cudbg/cudbg_wtp.c (revision d93a896e)
1 /*-
2  * Copyright (c) 2017 Chelsio Communications, Inc.
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  *
14  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24  * SUCH DAMAGE.
25  */
26 
27 #include <sys/cdefs.h>
28 __FBSDID("$FreeBSD$");
29 
30 #include <sys/types.h>
31 #include <sys/param.h>
32 
33 #include "common/common.h"
34 #include "common/t4_regs.h"
35 #include "cudbg.h"
36 #include "cudbg_lib_common.h"
37 #include "cudbg_entity.h"
38 
39 int collect_wtp_data(struct cudbg_init *pdbg_init,
40 		     struct cudbg_buffer *dbg_buff,
41 		     struct cudbg_error *cudbg_err);
42 /*SGE_DEBUG Registers.*/
43 #define TP_MIB_SIZE	    0x5e
44 
45 struct sge_debug_reg_data {
46 	/*indx0*/
47 	u32 reserved1:4;
48 	u32 reserved2:4;
49 	u32 debug_uP_SOP_cnt:4;
50 	u32 debug_uP_EOP_cnt:4;
51 	u32 debug_CIM_SOP1_cnt:4;
52 	u32 debug_CIM_EOP1_cnt:4;
53 	u32 debug_CIM_SOP0_cnt:4;
54 	u32 debug_CIM_EOP0_cnt:4;
55 
56 	/*indx1*/
57 	u32 reserved3:32;
58 
59 	/*indx2*/
60 	u32 debug_T_Rx_SOP1_cnt:4;
61 	u32 debug_T_Rx_EOP1_cnt:4;
62 	u32 debug_T_Rx_SOP0_cnt:4;
63 	u32 debug_T_Rx_EOP0_cnt:4;
64 	u32 debug_U_Rx_SOP1_cnt:4;
65 	u32 debug_U_Rx_EOP1_cnt:4;
66 	u32 debug_U_Rx_SOP0_cnt:4;
67 	u32 debug_U_Rx_EOP0_cnt:4;
68 
69 	/*indx3*/
70 	u32 reserved4:32;
71 
72 	/*indx4*/
73 	u32 debug_UD_Rx_SOP3_cnt:4;
74 	u32 debug_UD_Rx_EOP3_cnt:4;
75 	u32 debug_UD_Rx_SOP2_cnt:4;
76 	u32 debug_UD_Rx_EOP2_cnt:4;
77 	u32 debug_UD_Rx_SOP1_cnt:4;
78 	u32 debug_UD_Rx_EOP1_cnt:4;
79 	u32 debug_UD_Rx_SOP0_cnt:4;
80 	u32 debug_UD_Rx_EOP0_cnt:4;
81 
82 	/*indx5*/
83 	u32 reserved5:32;
84 
85 	/*indx6*/
86 	u32 debug_U_Tx_SOP3_cnt:4;
87 	u32 debug_U_Tx_EOP3_cnt:4;
88 	u32 debug_U_Tx_SOP2_cnt:4;
89 	u32 debug_U_Tx_EOP2_cnt:4;
90 	u32 debug_U_Tx_SOP1_cnt:4;
91 	u32 debug_U_Tx_EOP1_cnt:4;
92 	u32 debug_U_Tx_SOP0_cnt:4;
93 	u32 debug_U_Tx_EOP0_cnt:4;
94 
95 	/*indx7*/
96 	u32 reserved6:32;
97 
98 	/*indx8*/
99 	u32  debug_PC_Rsp_SOP1_cnt:4;
100 	u32  debug_PC_Rsp_EOP1_cnt:4;
101 	u32  debug_PC_Rsp_SOP0_cnt:4;
102 	u32  debug_PC_Rsp_EOP0_cnt:4;
103 	u32  debug_PC_Req_SOP1_cnt:4;
104 	u32  debug_PC_Req_EOP1_cnt:4;
105 	u32  debug_PC_Req_SOP0_cnt:4;
106 	u32  debug_PC_Req_EOP0_cnt:4;
107 
108 	/*indx9*/
109 	u32 reserved7:32;
110 
111 	/*indx10*/
112 	u32  debug_PD_Req_SOP3_cnt:4;
113 	u32  debug_PD_Req_EOP3_cnt:4;
114 	u32  debug_PD_Req_SOP2_cnt:4;
115 	u32  debug_PD_Req_EOP2_cnt:4;
116 	u32  debug_PD_Req_SOP1_cnt:4;
117 	u32  debug_PD_Req_EOP1_cnt:4;
118 	u32  debug_PD_Req_SOP0_cnt:4;
119 	u32  debug_PD_Req_EOP0_cnt:4;
120 
121 	/*indx11*/
122 	u32 reserved8:32;
123 
124 	/*indx12*/
125 	u32  debug_PD_Rsp_SOP3_cnt:4;
126 	u32  debug_PD_Rsp_EOP3_cnt:4;
127 	u32  debug_PD_Rsp_SOP2_cnt:4;
128 	u32  debug_PD_Rsp_EOP2_cnt:4;
129 	u32  debug_PD_Rsp_SOP1_cnt:4;
130 	u32  debug_PD_Rsp_EOP1_cnt:4;
131 	u32  debug_PD_Rsp_SOP0_cnt:4;
132 	u32  debug_PD_Rsp_EOP0_cnt:4;
133 
134 	/*indx13*/
135 	u32 reserved9:32;
136 
137 	/*indx14*/
138 	u32  debug_CPLSW_TP_Rx_SOP1_cnt:4;
139 	u32  debug_CPLSW_TP_Rx_EOP1_cnt:4;
140 	u32  debug_CPLSW_TP_Rx_SOP0_cnt:4;
141 	u32  debug_CPLSW_TP_Rx_EOP0_cnt:4;
142 	u32  debug_CPLSW_CIM_SOP1_cnt:4;
143 	u32  debug_CPLSW_CIM_EOP1_cnt:4;
144 	u32  debug_CPLSW_CIM_SOP0_cnt:4;
145 	u32  debug_CPLSW_CIM_EOP0_cnt:4;
146 
147 	/*indx15*/
148 	u32 reserved10:32;
149 
150 	/*indx16*/
151 	u32  debug_PD_Req_Rd3_cnt:4;
152 	u32  debug_PD_Req_Rd2_cnt:4;
153 	u32  debug_PD_Req_Rd1_cnt:4;
154 	u32  debug_PD_Req_Rd0_cnt:4;
155 	u32  debug_PD_Req_Int3_cnt:4;
156 	u32  debug_PD_Req_Int2_cnt:4;
157 	u32  debug_PD_Req_Int1_cnt:4;
158 	u32  debug_PD_Req_Int0_cnt:4;
159 
160 };
161 
162 struct tp_mib_type tp_mib[] = {
163 	{"tp_mib_mac_in_err_0", 0x0},
164 	{"tp_mib_mac_in_err_1", 0x1},
165 	{"tp_mib_mac_in_err_2", 0x2},
166 	{"tp_mib_mac_in_err_3", 0x3},
167 	{"tp_mib_hdr_in_err_0", 0x4},
168 	{"tp_mib_hdr_in_err_1", 0x5},
169 	{"tp_mib_hdr_in_err_2", 0x6},
170 	{"tp_mib_hdr_in_err_3", 0x7},
171 	{"tp_mib_tcp_in_err_0", 0x8},
172 	{"tp_mib_tcp_in_err_1", 0x9},
173 	{"tp_mib_tcp_in_err_2", 0xa},
174 	{"tp_mib_tcp_in_err_3", 0xb},
175 	{"tp_mib_tcp_out_rst", 0xc},
176 	{"tp_mib_tcp_in_seg_hi", 0x10},
177 	{"tp_mib_tcp_in_seg_lo", 0x11},
178 	{"tp_mib_tcp_out_seg_hi", 0x12},
179 	{"tp_mib_tcp_out_seg_lo", 0x13},
180 	{"tp_mib_tcp_rxt_seg_hi", 0x14},
181 	{"tp_mib_tcp_rxt_seg_lo", 0x15},
182 	{"tp_mib_tnl_cng_drop_0", 0x18},
183 	{"tp_mib_tnl_cng_drop_1", 0x19},
184 	{"tp_mib_tnl_cng_drop_2", 0x1a},
185 	{"tp_mib_tnl_cng_drop_3", 0x1b},
186 	{"tp_mib_ofd_chn_drop_0", 0x1c},
187 	{"tp_mib_ofd_chn_drop_1", 0x1d},
188 	{"tp_mib_ofd_chn_drop_2", 0x1e},
189 	{"tp_mib_ofd_chn_drop_3", 0x1f},
190 	{"tp_mib_tnl_out_pkt_0", 0x20},
191 	{"tp_mib_tnl_out_pkt_1", 0x21},
192 	{"tp_mib_tnl_out_pkt_2", 0x22},
193 	{"tp_mib_tnl_out_pkt_3", 0x23},
194 	{"tp_mib_tnl_in_pkt_0", 0x24},
195 	{"tp_mib_tnl_in_pkt_1", 0x25},
196 	{"tp_mib_tnl_in_pkt_2", 0x26},
197 	{"tp_mib_tnl_in_pkt_3", 0x27},
198 	{"tp_mib_tcp_v6in_err_0", 0x28},
199 	{"tp_mib_tcp_v6in_err_1", 0x29},
200 	{"tp_mib_tcp_v6in_err_2", 0x2a},
201 	{"tp_mib_tcp_v6in_err_3", 0x2b},
202 	{"tp_mib_tcp_v6out_rst", 0x2c},
203 	{"tp_mib_tcp_v6in_seg_hi", 0x30},
204 	{"tp_mib_tcp_v6in_seg_lo", 0x31},
205 	{"tp_mib_tcp_v6out_seg_hi", 0x32},
206 	{"tp_mib_tcp_v6out_seg_lo", 0x33},
207 	{"tp_mib_tcp_v6rxt_seg_hi", 0x34},
208 	{"tp_mib_tcp_v6rxt_seg_lo", 0x35},
209 	{"tp_mib_ofd_arp_drop", 0x36},
210 	{"tp_mib_ofd_dfr_drop", 0x37},
211 	{"tp_mib_cpl_in_req_0", 0x38},
212 	{"tp_mib_cpl_in_req_1", 0x39},
213 	{"tp_mib_cpl_in_req_2", 0x3a},
214 	{"tp_mib_cpl_in_req_3", 0x3b},
215 	{"tp_mib_cpl_out_rsp_0", 0x3c},
216 	{"tp_mib_cpl_out_rsp_1", 0x3d},
217 	{"tp_mib_cpl_out_rsp_2", 0x3e},
218 	{"tp_mib_cpl_out_rsp_3", 0x3f},
219 	{"tp_mib_tnl_lpbk_0", 0x40},
220 	{"tp_mib_tnl_lpbk_1", 0x41},
221 	{"tp_mib_tnl_lpbk_2", 0x42},
222 	{"tp_mib_tnl_lpbk_3", 0x43},
223 	{"tp_mib_tnl_drop_0", 0x44},
224 	{"tp_mib_tnl_drop_1", 0x45},
225 	{"tp_mib_tnl_drop_2", 0x46},
226 	{"tp_mib_tnl_drop_3", 0x47},
227 	{"tp_mib_fcoe_ddp_0", 0x48},
228 	{"tp_mib_fcoe_ddp_1", 0x49},
229 	{"tp_mib_fcoe_ddp_2", 0x4a},
230 	{"tp_mib_fcoe_ddp_3", 0x4b},
231 	{"tp_mib_fcoe_drop_0", 0x4c},
232 	{"tp_mib_fcoe_drop_1", 0x4d},
233 	{"tp_mib_fcoe_drop_2", 0x4e},
234 	{"tp_mib_fcoe_drop_3", 0x4f},
235 	{"tp_mib_fcoe_byte_0_hi", 0x50},
236 	{"tp_mib_fcoe_byte_0_lo", 0x51},
237 	{"tp_mib_fcoe_byte_1_hi", 0x52},
238 	{"tp_mib_fcoe_byte_1_lo", 0x53},
239 	{"tp_mib_fcoe_byte_2_hi", 0x54},
240 	{"tp_mib_fcoe_byte_2_lo", 0x55},
241 	{"tp_mib_fcoe_byte_3_hi", 0x56},
242 	{"tp_mib_fcoe_byte_3_lo", 0x57},
243 	{"tp_mib_ofd_vln_drop_0", 0x58},
244 	{"tp_mib_ofd_vln_drop_1", 0x59},
245 	{"tp_mib_ofd_vln_drop_2", 0x5a},
246 	{"tp_mib_ofd_vln_drop_3", 0x5b},
247 	{"tp_mib_usm_pkts", 0x5c},
248 	{"tp_mib_usm_drop", 0x5d},
249 	{"tp_mib_usm_bytes_hi", 0x5e},
250 	{"tp_mib_usm_bytes_lo", 0x5f},
251 	{"tp_mib_tid_del", 0x60},
252 	{"tp_mib_tid_inv", 0x61},
253 	{"tp_mib_tid_act", 0x62},
254 	{"tp_mib_tid_pas", 0x63},
255 	{"tp_mib_rqe_dfr_mod", 0x64},
256 	{"tp_mib_rqe_dfr_pkt", 0x65}
257 };
258 
259 static u32 read_sge_debug_data(struct cudbg_init *pdbg_init, u32 *sge_dbg_reg)
260 {
261 	struct adapter *padap = pdbg_init->adap;
262 	u32 value;
263 	int i = 0;
264 
265 	for (i = 0; i <= 15; i++) {
266 		t4_write_reg(padap, A_SGE_DEBUG_INDEX, (u32)i);
267 		value = t4_read_reg(padap, A_SGE_DEBUG_DATA_LOW);
268 		/*printf("LOW	 0x%08x\n", value);*/
269 		sge_dbg_reg[(i << 1) | 1] = HTONL_NIBBLE(value);
270 		value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH);
271 		/*printf("HIGH	 0x%08x\n", value);*/
272 		sge_dbg_reg[(i << 1)] = HTONL_NIBBLE(value);
273 	}
274 	return 0;
275 }
276 
277 static u32 read_tp_mib_data(struct cudbg_init *pdbg_init,
278 			    struct tp_mib_data **ppTp_Mib)
279 {
280 	struct adapter *padap = pdbg_init->adap;
281 	u32 i = 0;
282 
283 	for (i = 0; i < TP_MIB_SIZE; i++) {
284 		t4_tp_mib_read(padap, &tp_mib[i].value, 1,
285 				  (u32)tp_mib[i].addr, true);
286 	}
287 	*ppTp_Mib = (struct tp_mib_data *)&tp_mib[0];
288 
289 	return 0;
290 }
291 
292 static int t5_wtp_data(struct cudbg_init *pdbg_init,
293 		       struct cudbg_buffer *dbg_buff,
294 		       struct cudbg_error *cudbg_err)
295 {
296 	struct adapter *padap = pdbg_init->adap;
297 	struct sge_debug_reg_data *sge_dbg_reg = NULL;
298 	struct cudbg_buffer scratch_buff;
299 	struct tp_mib_data *ptp_mib = NULL;
300 	struct wtp_data *wtp;
301 	u32 Sge_Dbg[32] = {0};
302 	u32 value = 0;
303 	u32 i = 0;
304 	u32 drop = 0;
305 	u32 err = 0;
306 	u32 offset;
307 	int rc = 0;
308 
309 	rc = get_scratch_buff(dbg_buff, sizeof(struct wtp_data), &scratch_buff);
310 
311 	if (rc)
312 		goto err;
313 
314 	offset = scratch_buff.offset;
315 	wtp = (struct wtp_data *)((char *)scratch_buff.data + offset);
316 
317 	read_sge_debug_data(pdbg_init, Sge_Dbg);
318 	read_tp_mib_data(pdbg_init, &ptp_mib);
319 
320 	sge_dbg_reg = (struct sge_debug_reg_data *) &Sge_Dbg[0];
321 
322 	/*#######################################################################*/
323 	/*# TX PATH, starting from pcie*/
324 	/*#######################################################################*/
325 
326 	/* Get Reqests of commmands from SGE to PCIE*/
327 
328 	wtp->sge_pcie_cmd_req.sop[0] =	sge_dbg_reg->debug_PC_Req_SOP0_cnt;
329 	wtp->sge_pcie_cmd_req.sop[1] =	sge_dbg_reg->debug_PC_Req_SOP1_cnt;
330 
331 	wtp->sge_pcie_cmd_req.eop[0] =	sge_dbg_reg->debug_PC_Req_EOP0_cnt;
332 	wtp->sge_pcie_cmd_req.eop[1] =	sge_dbg_reg->debug_PC_Req_EOP1_cnt;
333 
334 	/* Get Reqests of commmands from PCIE to core*/
335 	value = t4_read_reg(padap, A_PCIE_CMDR_REQ_CNT);
336 
337 	wtp->pcie_core_cmd_req.sop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/
338 	wtp->pcie_core_cmd_req.sop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/
339 	/* there is no EOP for this, so we fake it.*/
340 	wtp->pcie_core_cmd_req.eop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/
341 	wtp->pcie_core_cmd_req.eop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/
342 
343 	/* Get DMA stats*/
344 	for (i = 0; i < 4; i++) {
345 		value = t4_read_reg(padap, A_PCIE_T5_DMA_STAT3 + (i * 0x10));
346 		wtp->pcie_t5_dma_stat3.sop[i] = value & 0xFF;
347 		wtp->pcie_t5_dma_stat3.eop[i] = ((value >> 16) & 0xFF);
348 	}
349 
350 	/* Get SGE debug data high index 6*/
351 	value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_6);
352 	wtp->sge_debug_data_high_index_6.sop[0] = ((value >> 4) & 0x0F);
353 	wtp->sge_debug_data_high_index_6.eop[0] = ((value >> 0) & 0x0F);
354 	wtp->sge_debug_data_high_index_6.sop[1] = ((value >> 12) & 0x0F);
355 	wtp->sge_debug_data_high_index_6.eop[1] = ((value >> 8) & 0x0F);
356 	wtp->sge_debug_data_high_index_6.sop[2] = ((value >> 20) & 0x0F);
357 	wtp->sge_debug_data_high_index_6.eop[2] = ((value >> 16) & 0x0F);
358 	wtp->sge_debug_data_high_index_6.sop[3] = ((value >> 28) & 0x0F);
359 	wtp->sge_debug_data_high_index_6.eop[3] = ((value >> 24) & 0x0F);
360 
361 	/* Get SGE debug data high index 3*/
362 	value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_3);
363 	wtp->sge_debug_data_high_index_3.sop[0] = ((value >> 4) & 0x0F);
364 	wtp->sge_debug_data_high_index_3.eop[0] = ((value >> 0) & 0x0F);
365 	wtp->sge_debug_data_high_index_3.sop[1] = ((value >> 12) & 0x0F);
366 	wtp->sge_debug_data_high_index_3.eop[1] = ((value >> 8) & 0x0F);
367 	wtp->sge_debug_data_high_index_3.sop[2] = ((value >> 20) & 0x0F);
368 	wtp->sge_debug_data_high_index_3.eop[2] = ((value >> 16) & 0x0F);
369 	wtp->sge_debug_data_high_index_3.sop[3] = ((value >> 28) & 0x0F);
370 	wtp->sge_debug_data_high_index_3.eop[3] = ((value >> 24) & 0x0F);
371 
372 	/* Get ULP SE CNT CHx*/
373 	for (i = 0; i < 4; i++) {
374 		value = t4_read_reg(padap, A_ULP_TX_SE_CNT_CH0 + (i * 4));
375 		wtp->ulp_se_cnt_chx.sop[i] = ((value >> 28) & 0x0F);
376 		wtp->ulp_se_cnt_chx.eop[i] = ((value >> 24) & 0x0F);
377 	}
378 
379 	/* Get MAC PORTx PKT COUNT*/
380 	for (i = 0; i < 4; i++) {
381 		value = t4_read_reg(padap, 0x3081c + ((i * 4) << 12));
382 		wtp->mac_portx_pkt_count.sop[i] = ((value >> 24) & 0xFF);
383 		wtp->mac_portx_pkt_count.eop[i] = ((value >> 16) & 0xFF);
384 		wtp->mac_porrx_pkt_count.sop[i] = ((value >> 8) & 0xFF);
385 		wtp->mac_porrx_pkt_count.eop[i] = ((value >> 0) & 0xFF);
386 	}
387 
388 	/* Get mac portx aFramesTransmittedok*/
389 	for (i = 0; i < 4; i++) {
390 		value = t4_read_reg(padap, 0x30a80 + ((i * 4) << 12));
391 		wtp->mac_portx_aframestra_ok.sop[i] = (value & 0xFF);
392 		wtp->mac_portx_aframestra_ok.eop[i] = (value & 0xFF);
393 	}
394 
395 	/* Get command respones from core to PCIE*/
396 	value = t4_read_reg(padap, A_PCIE_CMDR_RSP_CNT);
397 
398 	wtp->core_pcie_cmd_rsp.sop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/
399 	wtp->core_pcie_cmd_rsp.sop[1] = ((value >> 16) & 0xFF); /*bit 16:23*/
400 
401 	wtp->core_pcie_cmd_rsp.eop[0] = ((value >> 8) & 0xFF); /*bit 8:15*/
402 	wtp->core_pcie_cmd_rsp.eop[1] = ((value >> 24) & 0xFF); /*bit 24:31*/
403 
404 	/*Get command Resposes from PCIE to SGE*/
405 	wtp->pcie_sge_cmd_rsp.sop[0] = sge_dbg_reg->debug_PC_Rsp_SOP0_cnt;
406 	wtp->pcie_sge_cmd_rsp.sop[1] = sge_dbg_reg->debug_PC_Rsp_SOP1_cnt;
407 
408 	wtp->pcie_sge_cmd_rsp.eop[0] = sge_dbg_reg->debug_PC_Rsp_EOP0_cnt;
409 	wtp->pcie_sge_cmd_rsp.eop[1] = sge_dbg_reg->debug_PC_Rsp_EOP1_cnt;
410 
411 	/* Get commands sent from SGE to CIM/uP*/
412 	wtp->sge_cim.sop[0] = sge_dbg_reg->debug_CIM_SOP0_cnt;
413 	wtp->sge_cim.sop[1] = sge_dbg_reg->debug_CIM_SOP1_cnt;
414 
415 	wtp->sge_cim.eop[0] = sge_dbg_reg->debug_CIM_EOP0_cnt;
416 	wtp->sge_cim.eop[1] = sge_dbg_reg->debug_CIM_EOP1_cnt;
417 
418 	/* Get Reqests of data from PCIE by SGE*/
419 	wtp->utx_sge_dma_req.sop[0] = sge_dbg_reg->debug_UD_Rx_SOP0_cnt;
420 	wtp->utx_sge_dma_req.sop[1] = sge_dbg_reg->debug_UD_Rx_SOP1_cnt;
421 	wtp->utx_sge_dma_req.sop[2] = sge_dbg_reg->debug_UD_Rx_SOP2_cnt;
422 	wtp->utx_sge_dma_req.sop[3] = sge_dbg_reg->debug_UD_Rx_SOP3_cnt;
423 
424 	wtp->utx_sge_dma_req.eop[0] = sge_dbg_reg->debug_UD_Rx_EOP0_cnt;
425 	wtp->utx_sge_dma_req.eop[1] = sge_dbg_reg->debug_UD_Rx_EOP1_cnt;
426 	wtp->utx_sge_dma_req.eop[2] = sge_dbg_reg->debug_UD_Rx_EOP2_cnt;
427 	wtp->utx_sge_dma_req.eop[3] = sge_dbg_reg->debug_UD_Rx_EOP3_cnt;
428 
429 	/* Get Reqests of data from PCIE by SGE*/
430 	wtp->sge_pcie_dma_req.sop[0] = sge_dbg_reg->debug_PD_Req_Rd0_cnt;
431 	wtp->sge_pcie_dma_req.sop[1] = sge_dbg_reg->debug_PD_Req_Rd1_cnt;
432 	wtp->sge_pcie_dma_req.sop[2] = sge_dbg_reg->debug_PD_Req_Rd2_cnt;
433 	wtp->sge_pcie_dma_req.sop[3] = sge_dbg_reg->debug_PD_Req_Rd3_cnt;
434 	/*no EOP's, so fake it.*/
435 	wtp->sge_pcie_dma_req.eop[0] = sge_dbg_reg->debug_PD_Req_Rd0_cnt;
436 	wtp->sge_pcie_dma_req.eop[1] = sge_dbg_reg->debug_PD_Req_Rd1_cnt;
437 	wtp->sge_pcie_dma_req.eop[2] = sge_dbg_reg->debug_PD_Req_Rd2_cnt;
438 	wtp->sge_pcie_dma_req.eop[3] = sge_dbg_reg->debug_PD_Req_Rd3_cnt;
439 
440 	/* Get Reqests of data from PCIE to core*/
441 	value = t4_read_reg(padap, A_PCIE_DMAR_REQ_CNT);
442 
443 	wtp->pcie_core_dma_req.sop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/
444 	wtp->pcie_core_dma_req.sop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/
445 	wtp->pcie_core_dma_req.sop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/
446 	wtp->pcie_core_dma_req.sop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/
447 	/* There is no eop so fake it.*/
448 	wtp->pcie_core_dma_req.eop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/
449 	wtp->pcie_core_dma_req.eop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/
450 	wtp->pcie_core_dma_req.eop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/
451 	wtp->pcie_core_dma_req.eop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/
452 
453 	/* Get data responses from core to PCIE*/
454 	value = t4_read_reg(padap, A_PCIE_DMAR_RSP_SOP_CNT);
455 
456 	wtp->core_pcie_dma_rsp.sop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/
457 	wtp->core_pcie_dma_rsp.sop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/
458 	wtp->core_pcie_dma_rsp.sop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/
459 	wtp->core_pcie_dma_rsp.sop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/
460 
461 	value = t4_read_reg(padap, A_PCIE_DMAR_RSP_EOP_CNT);
462 
463 	wtp->core_pcie_dma_rsp.eop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/
464 	wtp->core_pcie_dma_rsp.eop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/
465 	wtp->core_pcie_dma_rsp.eop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/
466 	wtp->core_pcie_dma_rsp.eop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/
467 
468 	/* Get PCIE_DATA to SGE*/
469 	wtp->pcie_sge_dma_rsp.sop[0] = sge_dbg_reg->debug_PD_Rsp_SOP0_cnt;
470 	wtp->pcie_sge_dma_rsp.sop[1] = sge_dbg_reg->debug_PD_Rsp_SOP1_cnt;
471 	wtp->pcie_sge_dma_rsp.sop[2] = sge_dbg_reg->debug_PD_Rsp_SOP2_cnt;
472 	wtp->pcie_sge_dma_rsp.sop[3] = sge_dbg_reg->debug_PD_Rsp_SOP3_cnt;
473 
474 	wtp->pcie_sge_dma_rsp.eop[0] = sge_dbg_reg->debug_PD_Rsp_EOP0_cnt;
475 	wtp->pcie_sge_dma_rsp.eop[1] = sge_dbg_reg->debug_PD_Rsp_EOP1_cnt;
476 	wtp->pcie_sge_dma_rsp.eop[2] = sge_dbg_reg->debug_PD_Rsp_EOP2_cnt;
477 	wtp->pcie_sge_dma_rsp.eop[3] = sge_dbg_reg->debug_PD_Rsp_EOP3_cnt;
478 
479 	/*Get SGE to ULP_TX*/
480 	wtp->sge_utx.sop[0] = sge_dbg_reg->debug_U_Tx_SOP0_cnt;
481 	wtp->sge_utx.sop[1] = sge_dbg_reg->debug_U_Tx_SOP1_cnt;
482 	wtp->sge_utx.sop[2] = sge_dbg_reg->debug_U_Tx_SOP2_cnt;
483 	wtp->sge_utx.sop[3] = sge_dbg_reg->debug_U_Tx_SOP3_cnt;
484 
485 	wtp->sge_utx.eop[0] = sge_dbg_reg->debug_U_Tx_EOP0_cnt;
486 	wtp->sge_utx.eop[1] = sge_dbg_reg->debug_U_Tx_EOP1_cnt;
487 	wtp->sge_utx.eop[2] = sge_dbg_reg->debug_U_Tx_EOP2_cnt;
488 	wtp->sge_utx.eop[3] = sge_dbg_reg->debug_U_Tx_EOP3_cnt;
489 
490 	/* Get ULP_TX to TP*/
491 	for (i = 0; i < 4; i++) {
492 		value = t4_read_reg(padap, (A_ULP_TX_SE_CNT_CH0 + (i*4)));
493 
494 		wtp->utx_tp.sop[i] = ((value >> 28) & 0xF); /*bits 28:31*/
495 		wtp->utx_tp.eop[i] = ((value >> 24) & 0xF); /*bits 24:27*/
496 	}
497 
498 	/* Get TP_DBG_CSIDE registers*/
499 	for (i = 0; i < 4; i++) {
500 		t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_CSIDE_RX0 + i),
501 			       true);
502 
503 		wtp->utx_tpcside.sop[i]   = ((value >> 28) & 0xF);/*bits 28:31*/
504 		wtp->utx_tpcside.eop[i]   = ((value >> 24) & 0xF);/*bits 24:27*/
505 		wtp->tpcside_rxpld.sop[i] = ((value >> 20) & 0xF);/*bits 20:23*/
506 		wtp->tpcside_rxpld.eop[i] = ((value >> 16) & 0xF);/*bits 16:19*/
507 		wtp->tpcside_rxarb.sop[i] = ((value >> 12) & 0xF);/*bits 12:15*/
508 		wtp->tpcside_rxarb.eop[i] = ((value >> 8) & 0xF); /*bits 8:11*/
509 		wtp->tpcside_rxcpl.sop[i] = ((value >> 4) & 0xF); /*bits 4:7*/
510 		wtp->tpcside_rxcpl.eop[i] = ((value >> 0) & 0xF); /*bits 0:3*/
511 	}
512 
513 	/* TP_DBG_ESIDE*/
514 	for (i = 0; i < 4; i++) {
515 		t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_ESIDE_PKT0 + i),
516 			       true);
517 
518 		wtp->tpeside_mps.sop[i] = ((value >> 28) & 0xF); /*bits 28:31*/
519 		wtp->tpeside_mps.eop[i] = ((value >> 24) & 0xF); /*bits 24:27*/
520 		wtp->tpeside_pm.sop[i]	= ((value >> 20) & 0xF); /*bits 20:23*/
521 		wtp->tpeside_pm.eop[i]	= ((value >> 16) & 0xF); /*bits 16:19*/
522 		wtp->mps_tpeside.sop[i] = ((value >> 12) & 0xF); /*bits 12:15*/
523 		wtp->mps_tpeside.eop[i] = ((value >> 8) & 0xF); /*bits 8:11*/
524 		wtp->tpeside_pld.sop[i] = ((value >> 4) & 0xF); /*bits 4:7*/
525 		wtp->tpeside_pld.eop[i] = ((value >> 0) & 0xF); /*bits 0:3*/
526 
527 	}
528 
529 	/*PCIE CMD STAT2*/
530 	for (i = 0; i < 3; i++) {
531 		value = t4_read_reg(padap, 0x5988 + (i * 0x10));
532 		wtp->pcie_cmd_stat2.sop[i] = value & 0xFF;
533 		wtp->pcie_cmd_stat2.eop[i] = value & 0xFF;
534 	}
535 
536 	/*PCIE cmd stat3*/
537 	for (i = 0; i < 3; i++) {
538 		value = t4_read_reg(padap, 0x598c + (i * 0x10));
539 		wtp->pcie_cmd_stat3.sop[i] = value & 0xFF;
540 		wtp->pcie_cmd_stat3.eop[i] = value & 0xFF;
541 	}
542 
543 	/* ULP_RX input/output*/
544 	for (i = 0; i < 2; i++) {
545 		value = t4_read_reg(padap, (A_ULP_RX_SE_CNT_CH0 + (i*4)));
546 
547 		wtp->pmrx_ulprx.sop[i]	  = ((value >> 4) & 0xF); /*bits 4:7*/
548 		wtp->pmrx_ulprx.eop[i]	  = ((value >> 0) & 0xF); /*bits 0:3*/
549 		wtp->ulprx_tpcside.sop[i] = ((value >> 28) & 0xF);/*bits 28:31*/
550 		wtp->ulprx_tpcside.eop[i] = ((value >> 24) & 0xF);/*bits 24:27*/
551 	}
552 
553 	/* Get the MPS input from TP*/
554 	drop = 0;
555 	for (i = 0; i < 2; i++) {
556 		value = t4_read_reg(padap, (A_MPS_TX_SE_CNT_TP01 + (i << 2)));
557 		wtp->tp_mps.sop[(i*2)]	   = ((value >> 8) & 0xFF); /*bit 8:15*/
558 		wtp->tp_mps.eop[(i*2)]	   = ((value >> 0) & 0xFF); /*bit 0:7*/
559 		wtp->tp_mps.sop[(i*2) + 1] = ((value >> 24) & 0xFF);/*bit 24:31
560 								    */
561 		wtp->tp_mps.eop[(i*2) + 1] = ((value >> 16) & 0xFF);/*bit 16:23
562 								    */
563 	}
564 	drop  = ptp_mib->TP_MIB_OFD_ARP_DROP.value;
565 	drop += ptp_mib->TP_MIB_OFD_DFR_DROP.value;
566 
567 	drop += ptp_mib->TP_MIB_TNL_DROP_0.value;
568 	drop += ptp_mib->TP_MIB_TNL_DROP_1.value;
569 	drop += ptp_mib->TP_MIB_TNL_DROP_2.value;
570 	drop += ptp_mib->TP_MIB_TNL_DROP_3.value;
571 
572 	wtp->tp_mps.drops = drop;
573 
574 	/* Get the MPS output to the MAC's*/
575 	drop = 0;
576 	for (i = 0; i < 2; i++) {
577 		value = t4_read_reg(padap, (A_MPS_TX_SE_CNT_MAC01 + (i << 2)));
578 		wtp->mps_xgm.sop[(i*2)]     = ((value >> 8) & 0xFF);/*bit 8:15*/
579 		wtp->mps_xgm.eop[(i*2)]     = ((value >> 0) & 0xFF);/*bit 0:7*/
580 		wtp->mps_xgm.sop[(i*2) + 1] = ((value >> 24) & 0xFF);/*bit 24:31
581 								     */
582 		wtp->mps_xgm.eop[(i*2) + 1] = ((value >> 16) & 0xFF);/*bit 16:23
583 								     */
584 	}
585 	for (i = 0; i < 4; i++) {
586 		value = t4_read_reg(padap,
587 				(T5_PORT0_REG(A_MPS_PORT_STAT_TX_PORT_DROP_L) +
588 				(i * T5_PORT_STRIDE)));
589 		drop += value;
590 	}
591 	wtp->mps_xgm.drops = (drop & 0xFF);
592 
593 	/* Get the SOP/EOP counters into and out of MAC. [JHANEL] I think this
594 	 * is*/
595 	/* clear on read, so you have to read both TX and RX path at same
596 	 * time.*/
597 	drop = 0;
598 	for (i = 0; i < 4; i++) {
599 		value = t4_read_reg(padap,
600 				(T5_PORT0_REG(A_MAC_PORT_PKT_COUNT) +
601 				(i * T5_PORT_STRIDE)));
602 
603 		wtp->tx_xgm_xgm.sop[i] = ((value >> 24) & 0xFF); /*bit 24:31*/
604 		wtp->tx_xgm_xgm.eop[i] = ((value >> 16) & 0xFF); /*bit 16:23*/
605 		wtp->rx_xgm_xgm.sop[i] = ((value >> 8) & 0xFF); /*bit 8:15*/
606 		wtp->rx_xgm_xgm.eop[i] = ((value >> 0) & 0xFF); /*bit 0:7*/
607 	}
608 
609 	/* Get the MAC's output to the wire*/
610 	drop = 0;
611 	for (i = 0; i < 4; i++) {
612 		value = t4_read_reg(padap,
613 				(T5_PORT0_REG(A_MAC_PORT_AFRAMESTRANSMITTEDOK) +
614 				(i * T5_PORT_STRIDE)));
615 		wtp->xgm_wire.sop[i] = (value);
616 		wtp->xgm_wire.eop[i] = (value); /* No EOP for XGMAC, so fake
617 						   it.*/
618 	}
619 
620 	/*########################################################################*/
621 	/*# RX PATH, starting from wire*/
622 	/*########################################################################*/
623 
624 	/* Add up the wire input to the MAC*/
625 	drop = 0;
626 	for (i = 0; i < 4; i++) {
627 		value = t4_read_reg(padap,
628 				(T5_PORT0_REG(A_MAC_PORT_AFRAMESRECEIVEDOK) +
629 				(i * T5_PORT_STRIDE)));
630 
631 		wtp->wire_xgm.sop[i] = (value);
632 		wtp->wire_xgm.eop[i] = (value); /* No EOP for XGMAC, so fake
633 						   it.*/
634 	}
635 
636 	/* Already read the rx_xgm_xgm when reading TX path.*/
637 
638 	/* Add up SOP/EOP's on all 8 MPS buffer channels*/
639 	drop = 0;
640 	for (i = 0; i < 8; i++) {
641 		value = t4_read_reg(padap, (A_MPS_RX_SE_CNT_IN0 + (i << 2)));
642 
643 		wtp->xgm_mps.sop[i] = ((value >> 8) & 0xFF); /*bits 8:15*/
644 		wtp->xgm_mps.eop[i] = ((value >> 0) & 0xFF); /*bits 0:7*/
645 	}
646 	for (i = 0; i < 4; i++) {
647 		value = t4_read_reg(padap, (A_MPS_RX_CLS_DROP_CNT0 + (i << 2)));
648 		/* typo in JHANEL's code.*/
649 		drop += (value & 0xFFFF) + ((value >> 16) & 0xFFFF);
650 	}
651 	wtp->xgm_mps.cls_drop = drop & 0xFF;
652 
653 	/* Add up the overflow drops on all 4 ports.*/
654 	drop = 0;
655 	for (i = 0; i < 4; i++) {
656 		value = t4_read_reg(padap,
657 				    (A_MPS_STAT_RX_BG_0_MAC_DROP_FRAME_L +
658 				     (i << 3)));
659 		drop += value;
660 		value = t4_read_reg(padap,
661 				    (A_MPS_STAT_RX_BG_0_MAC_DROP_FRAME_L +
662 				     (i << 2)));
663 		value = t4_read_reg(padap,
664 				    (A_MPS_STAT_RX_BG_0_LB_DROP_FRAME_L +
665 				     (i << 3)));
666 		drop += value;
667 		value = t4_read_reg(padap,
668 				    (A_MPS_STAT_RX_BG_0_MAC_DROP_FRAME_L +
669 				     (i << 2)));
670 
671 		value = t4_read_reg(padap,
672 				    (A_MPS_STAT_RX_BG_0_MAC_TRUNC_FRAME_L +
673 				     (i << 3)));
674 		drop += value;
675 		value = t4_read_reg(padap,
676 				    (A_MPS_STAT_RX_BG_0_MAC_TRUNC_FRAME_L +
677 				     (i << 3)));
678 		value = t4_read_reg(padap,
679 				    (A_MPS_STAT_RX_BG_0_LB_TRUNC_FRAME_L +
680 				     (i << 3)));
681 		drop += value;
682 		value = t4_read_reg(padap,
683 				    (A_MPS_STAT_RX_BG_0_MAC_TRUNC_FRAME_L +
684 				     (i << 3)));
685 
686 		value = t4_read_reg(padap,
687 			T5_PORT0_REG(A_MPS_PORT_STAT_LB_PORT_DROP_FRAMES) +
688 			(i * T5_PORT_STRIDE));
689 		drop += value;
690 	}
691 	wtp->xgm_mps.drop = (drop & 0xFF);
692 
693 	/* Add up the MPS errors that should result in dropped packets*/
694 	err = 0;
695 	for (i = 0; i < 4; i++) {
696 
697 		value = t4_read_reg(padap,
698 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_ERROR_L) +
699 			(i * T5_PORT_STRIDE)));
700 		err += value;
701 		value = t4_read_reg(padap,
702 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_ERROR_L) +
703 			(i * T5_PORT_STRIDE) + 4));
704 
705 		value = t4_read_reg(padap,
706 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_CRC_ERROR_L) +
707 			(i * T5_PORT_STRIDE)));
708 		err += value;
709 		value = t4_read_reg(padap,
710 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_CRC_ERROR_L) +
711 			(i * T5_PORT_STRIDE) + 4));
712 
713 		value = t4_read_reg(padap,
714 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_CRC_ERROR_L) +
715 			(i * T5_PORT_STRIDE)));
716 		err += value;
717 		value = t4_read_reg(padap,
718 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_CRC_ERROR_L) +
719 			(i * T5_PORT_STRIDE) + 4));
720 
721 		value = t4_read_reg(padap,
722 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LEN_ERROR_L) +
723 			(i * T5_PORT_STRIDE)));
724 		err += value;
725 		value = t4_read_reg(padap,
726 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LEN_ERROR_L) +
727 			(i * T5_PORT_STRIDE) + 4));
728 
729 		value = t4_read_reg(padap,
730 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_SYM_ERROR_L) +
731 			(i * T5_PORT_STRIDE)));
732 		err += value;
733 		value = t4_read_reg(padap,
734 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_SYM_ERROR_L) +
735 			(i * T5_PORT_STRIDE) + 4));
736 
737 		value = t4_read_reg(padap,
738 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LESS_64B_L) +
739 			(i * T5_PORT_STRIDE)));
740 		err += value;
741 		value = t4_read_reg(padap,
742 			(T5_PORT0_REG((A_MPS_PORT_STAT_RX_PORT_LESS_64B_L) +
743 			(i * T5_PORT_STRIDE) + 4)));
744 	}
745 	wtp->xgm_mps.err = (err & 0xFF);
746 
747 	drop = 0;
748 	for (i = 0; i < 2; i++) {
749 		value = t4_read_reg(padap, (A_MPS_RX_SE_CNT_OUT01 + (i << 2)));
750 
751 		wtp->mps_tp.sop[(i*2)]	   = ((value >> 8) & 0xFF); /*bit 8:15*/
752 		wtp->mps_tp.eop[(i*2)]	   = ((value >> 0) & 0xFF); /*bit 0:7*/
753 		wtp->mps_tp.sop[(i*2) + 1] = ((value >> 24) & 0xFF);/*bit 24:31
754 								    */
755 		wtp->mps_tp.eop[(i*2) + 1] = ((value >> 16) & 0xFF);/*bit 16:23
756 								    */
757 	}
758 	drop = ptp_mib->TP_MIB_TNL_CNG_DROP_0.value;
759 	drop += ptp_mib->TP_MIB_TNL_CNG_DROP_1.value;
760 	drop += ptp_mib->TP_MIB_TNL_CNG_DROP_2.value;
761 	drop += ptp_mib->TP_MIB_TNL_CNG_DROP_3.value;
762 	drop += ptp_mib->TP_MIB_OFD_CHN_DROP_0.value;
763 	drop += ptp_mib->TP_MIB_OFD_CHN_DROP_1.value;
764 	drop += ptp_mib->TP_MIB_OFD_CHN_DROP_2.value;
765 	drop += ptp_mib->TP_MIB_OFD_CHN_DROP_3.value;
766 	drop += ptp_mib->TP_MIB_FCOE_DROP_0.value;
767 	drop += ptp_mib->TP_MIB_FCOE_DROP_1.value;
768 	drop += ptp_mib->TP_MIB_FCOE_DROP_2.value;
769 	drop += ptp_mib->TP_MIB_FCOE_DROP_3.value;
770 	drop += ptp_mib->TP_MIB_OFD_VLN_DROP_0.value;
771 	drop += ptp_mib->TP_MIB_OFD_VLN_DROP_1.value;
772 	drop += ptp_mib->TP_MIB_OFD_VLN_DROP_2.value;
773 	drop += ptp_mib->TP_MIB_OFD_VLN_DROP_3.value;
774 	drop += ptp_mib->TP_MIB_USM_DROP.value;
775 
776 	wtp->mps_tp.drops = drop;
777 
778 	/* Get TP_DBG_CSIDE_TX registers*/
779 	for (i = 0; i < 4; i++) {
780 		t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_CSIDE_RX0 + i),
781 			       true);
782 
783 		wtp->tpcside_csw.sop[i]   = ((value >> 28) & 0xF);/*bits 28:31*/
784 		wtp->tpcside_csw.eop[i]   = ((value >> 24) & 0xF);/*bits 24:27*/
785 		wtp->tpcside_pm.sop[i]	  = ((value >> 20) & 0xF);/*bits 20:23*/
786 		wtp->tpcside_pm.eop[i]	  = ((value >> 16) & 0xF);/*bits 16:19*/
787 		wtp->tpcside_uturn.sop[i] = ((value >> 12) & 0xF);/*bits 12:15*/
788 		wtp->tpcside_uturn.eop[i] = ((value >> 8) & 0xF); /*bits 8:11*/
789 		wtp->tpcside_txcpl.sop[i] = ((value >> 4) & 0xF); /*bits 4:7*/
790 		wtp->tpcside_txcpl.eop[i] = ((value >> 0) & 0xF); /*bits 0:3*/
791 	}
792 
793 	/* TP to CPL_SWITCH*/
794 	wtp->tp_csw.sop[0] = sge_dbg_reg->debug_CPLSW_TP_Rx_SOP0_cnt;
795 	wtp->tp_csw.sop[1] = sge_dbg_reg->debug_CPLSW_TP_Rx_SOP1_cnt;
796 
797 	wtp->tp_csw.eop[0] = sge_dbg_reg->debug_CPLSW_TP_Rx_EOP0_cnt;
798 	wtp->tp_csw.eop[1] = sge_dbg_reg->debug_CPLSW_TP_Rx_EOP1_cnt;
799 
800 	/* TP/CPL_SWITCH to SGE*/
801 	wtp->csw_sge.sop[0] = sge_dbg_reg->debug_T_Rx_SOP0_cnt;
802 	wtp->csw_sge.sop[1] = sge_dbg_reg->debug_T_Rx_SOP1_cnt;
803 
804 	wtp->csw_sge.eop[0] = sge_dbg_reg->debug_T_Rx_EOP0_cnt;
805 	wtp->csw_sge.eop[1] = sge_dbg_reg->debug_T_Rx_EOP1_cnt;
806 
807 	wtp->sge_pcie.sop[0] = sge_dbg_reg->debug_PD_Req_SOP0_cnt;
808 	wtp->sge_pcie.sop[1] = sge_dbg_reg->debug_PD_Req_SOP1_cnt;
809 	wtp->sge_pcie.sop[2] = sge_dbg_reg->debug_PD_Req_SOP2_cnt;
810 	wtp->sge_pcie.sop[3] = sge_dbg_reg->debug_PD_Req_SOP3_cnt;
811 
812 	wtp->sge_pcie.eop[0] = sge_dbg_reg->debug_PD_Req_EOP0_cnt;
813 	wtp->sge_pcie.eop[1] = sge_dbg_reg->debug_PD_Req_EOP1_cnt;
814 	wtp->sge_pcie.eop[2] = sge_dbg_reg->debug_PD_Req_EOP2_cnt;
815 	wtp->sge_pcie.eop[3] = sge_dbg_reg->debug_PD_Req_EOP3_cnt;
816 
817 	wtp->sge_pcie_ints.sop[0] = sge_dbg_reg->debug_PD_Req_Int0_cnt;
818 	wtp->sge_pcie_ints.sop[1] = sge_dbg_reg->debug_PD_Req_Int1_cnt;
819 	wtp->sge_pcie_ints.sop[2] = sge_dbg_reg->debug_PD_Req_Int2_cnt;
820 	wtp->sge_pcie_ints.sop[3] = sge_dbg_reg->debug_PD_Req_Int3_cnt;
821 	/* NO EOP, so fake it.*/
822 	wtp->sge_pcie_ints.eop[0] = sge_dbg_reg->debug_PD_Req_Int0_cnt;
823 	wtp->sge_pcie_ints.eop[1] = sge_dbg_reg->debug_PD_Req_Int1_cnt;
824 	wtp->sge_pcie_ints.eop[2] = sge_dbg_reg->debug_PD_Req_Int2_cnt;
825 	wtp->sge_pcie_ints.eop[3] = sge_dbg_reg->debug_PD_Req_Int3_cnt;
826 
827 	/*Get PCIE DMA1 STAT2*/
828 	for (i = 0; i < 4; i++) {
829 		value = t4_read_reg(padap, A_PCIE_T5_DMA_STAT2 + (i * 0x10));
830 		wtp->pcie_dma1_stat2.sop[i] = ((value >> 8) & 0x0F);
831 		wtp->pcie_dma1_stat2.eop[i] = ((value >> 8) & 0x0F);
832 		wtp->pcie_dma1_stat2_core.sop[i] += value & 0x0F;
833 		wtp->pcie_dma1_stat2_core.eop[i] += value & 0x0F;
834 	}
835 
836 	/* Get mac porrx aFramesTransmittedok*/
837 	for (i = 0; i < 4; i++) {
838 		value = t4_read_reg(padap, 0x30a88 + ((i * 4) << 12));
839 		wtp->mac_porrx_aframestra_ok.sop[i] = (value & 0xFF);
840 		wtp->mac_porrx_aframestra_ok.eop[i] = (value & 0xFF);
841 	}
842 
843 	/*Get SGE debug data high index 7*/
844 	value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_7);
845 	wtp->sge_debug_data_high_indx7.sop[0] = ((value >> 4) & 0x0F);
846 	wtp->sge_debug_data_high_indx7.eop[0] = ((value >> 0) & 0x0F);
847 	wtp->sge_debug_data_high_indx7.sop[1] = ((value >> 12) & 0x0F);
848 	wtp->sge_debug_data_high_indx7.eop[1] = ((value >> 8) & 0x0F);
849 	wtp->sge_debug_data_high_indx7.sop[2] = ((value >> 20) & 0x0F);
850 	wtp->sge_debug_data_high_indx7.eop[2] = ((value >> 16) & 0x0F);
851 	wtp->sge_debug_data_high_indx7.sop[3] = ((value >> 28) & 0x0F);
852 	wtp->sge_debug_data_high_indx7.eop[3] = ((value >> 24) & 0x0F);
853 
854 	/*Get SGE debug data high index 1*/
855 	value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_1);
856 	wtp->sge_debug_data_high_indx1.sop[0] = ((value >> 20) & 0x0F);
857 	wtp->sge_debug_data_high_indx1.eop[0] = ((value >> 16) & 0x0F);
858 	wtp->sge_debug_data_high_indx1.sop[1] = ((value >> 28) & 0x0F);
859 	wtp->sge_debug_data_high_indx1.eop[1] = ((value >> 24) & 0x0F);
860 
861 	/*Get TP debug CSIDE Tx registers*/
862 	for (i = 0; i < 2; i++) {
863 		t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_CSIDE_TX0 + i),
864 			       true);
865 
866 		wtp->utx_tpcside_tx.sop[i] = ((value >> 28) & 0xF);/*bits 28:31
867 								   */
868 		wtp->utx_tpcside_tx.eop[i] = ((value >> 24) & 0xF);
869 	}
870 
871 	/*Get SGE debug data high index 9*/
872 	value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_9);
873 	wtp->sge_debug_data_high_indx9.sop[0] = ((value >> 20) & 0x0F);
874 	wtp->sge_debug_data_high_indx9.sop[1] = ((value >> 28) & 0x0F);
875 	wtp->sge_debug_data_high_indx9.eop[0] = ((value >> 16) & 0x0F);
876 	wtp->sge_debug_data_high_indx9.eop[1] = ((value >> 24) & 0x0F);
877 	wtp->sge_work_req_pkt.sop[0] = ((value >> 4) & 0x0F);
878 	wtp->sge_work_req_pkt.sop[1] = ((value >> 12) & 0x0F);
879 
880 	/*Get LE DB response count*/
881 	value = t4_read_reg(padap, A_LE_DB_REQ_RSP_CNT);
882 	wtp->le_db_rsp_cnt.sop = value & 0xF;
883 	wtp->le_db_rsp_cnt.eop = (value >> 16) & 0xF;
884 
885 	/*Get TP debug Eside PKTx*/
886 	for (i = 0; i < 4; i++) {
887 		t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_ESIDE_PKT0 + i),
888 			       true);
889 
890 		wtp->tp_dbg_eside_pktx.sop[i] = ((value >> 12) & 0xF);
891 		wtp->tp_dbg_eside_pktx.eop[i] = ((value >> 8) & 0xF);
892 	}
893 
894 	/* Get data responses from core to PCIE*/
895 	value = t4_read_reg(padap, A_PCIE_DMAW_SOP_CNT);
896 
897 	wtp->pcie_core_dmaw.sop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/
898 	wtp->pcie_core_dmaw.sop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/
899 	wtp->pcie_core_dmaw.sop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/
900 	wtp->pcie_core_dmaw.sop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/
901 
902 	value = t4_read_reg(padap, A_PCIE_DMAW_EOP_CNT);
903 
904 	wtp->pcie_core_dmaw.eop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/
905 	wtp->pcie_core_dmaw.eop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/
906 	wtp->pcie_core_dmaw.eop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/
907 	wtp->pcie_core_dmaw.eop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/
908 
909 	value = t4_read_reg(padap, A_PCIE_DMAI_CNT);
910 
911 	wtp->pcie_core_dmai.sop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/
912 	wtp->pcie_core_dmai.sop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/
913 	wtp->pcie_core_dmai.sop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/
914 	wtp->pcie_core_dmai.sop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/
915 	/* no eop for interrups, just fake it.*/
916 	wtp->pcie_core_dmai.eop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/
917 	wtp->pcie_core_dmai.eop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/
918 	wtp->pcie_core_dmai.eop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/
919 	wtp->pcie_core_dmai.eop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/
920 
921 	rc = write_compression_hdr(&scratch_buff, dbg_buff);
922 
923 	if (rc)
924 		goto err1;
925 
926 	rc = compress_buff(&scratch_buff, dbg_buff);
927 
928 err1:
929 	release_scratch_buff(&scratch_buff, dbg_buff);
930 err:
931 	return rc;
932 }
933 
934 static int t6_wtp_data(struct cudbg_init *pdbg_init,
935 		       struct cudbg_buffer *dbg_buff,
936 		       struct cudbg_error *cudbg_err)
937 {
938 	struct adapter *padap = pdbg_init->adap;
939 	struct sge_debug_reg_data *sge_dbg_reg = NULL;
940 	struct cudbg_buffer scratch_buff;
941 	struct tp_mib_data *ptp_mib = NULL;
942 	struct wtp_data *wtp;
943 	u32 Sge_Dbg[32] = {0};
944 	u32 value = 0;
945 	u32 i = 0;
946 	u32 drop = 0;
947 	u32 err = 0;
948 	u32 offset;
949 	int rc = 0;
950 
951 	rc = get_scratch_buff(dbg_buff, sizeof(struct wtp_data), &scratch_buff);
952 
953 	if (rc)
954 		goto err;
955 
956 	offset = scratch_buff.offset;
957 	wtp = (struct wtp_data *)((char *)scratch_buff.data + offset);
958 
959 	read_sge_debug_data(pdbg_init, Sge_Dbg);
960 	read_tp_mib_data(pdbg_init, &ptp_mib);
961 
962 	sge_dbg_reg = (struct sge_debug_reg_data *) &Sge_Dbg[0];
963 
964 	/*# TX PATH*/
965 
966 	/*PCIE CMD STAT2*/
967 	value = t4_read_reg(padap, A_PCIE_T5_CMD_STAT2);
968 	wtp->pcie_cmd_stat2.sop[0] = value & 0xFF;
969 	wtp->pcie_cmd_stat2.eop[0] = value & 0xFF;
970 
971 	value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_7);
972 	wtp->sge_pcie_cmd_req.sop[0] = ((value >> 20) & 0x0F);
973 	wtp->sge_pcie_cmd_req.eop[0] = ((value >> 16) & 0x0F);
974 	wtp->sge_pcie_cmd_req.sop[1] = ((value >> 28) & 0x0F);
975 	wtp->sge_pcie_cmd_req.eop[1] = ((value >> 24) & 0x0F);
976 
977 	value = t4_read_reg(padap, A_PCIE_T5_CMD_STAT3);
978 	wtp->pcie_cmd_stat3.sop[0] = value & 0xFF;
979 	wtp->pcie_cmd_stat3.eop[0] = value & 0xFF;
980 
981 	/*Get command Resposes from PCIE to SGE*/
982 	wtp->pcie_sge_cmd_rsp.sop[0] = sge_dbg_reg->debug_PC_Rsp_SOP0_cnt;
983 	wtp->pcie_sge_cmd_rsp.eop[0] = sge_dbg_reg->debug_PC_Rsp_EOP0_cnt;
984 	wtp->pcie_sge_cmd_rsp.sop[1] = sge_dbg_reg->debug_PC_Rsp_SOP1_cnt;
985 	wtp->pcie_sge_cmd_rsp.eop[1] = sge_dbg_reg->debug_PC_Rsp_EOP0_cnt;
986 
987 	/* Get commands sent from SGE to CIM/uP*/
988 	wtp->sge_cim.sop[0] = sge_dbg_reg->debug_CIM_SOP0_cnt;
989 	wtp->sge_cim.sop[1] = sge_dbg_reg->debug_CIM_SOP1_cnt;
990 
991 	wtp->sge_cim.eop[0] = sge_dbg_reg->debug_CIM_EOP0_cnt;
992 	wtp->sge_cim.eop[1] = sge_dbg_reg->debug_CIM_EOP1_cnt;
993 
994 	/*Get SGE debug data high index 9*/
995 	value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_9);
996 	wtp->sge_work_req_pkt.sop[0] = ((value >> 4) & 0x0F);
997 	wtp->sge_work_req_pkt.eop[0] = ((value >> 0) & 0x0F);
998 
999 	for (i = 0; i < 2; i++) {
1000 		value = t4_read_reg(padap, A_PCIE_T5_DMA_STAT2 + (i * 0x10));
1001 		wtp->pcie_dma1_stat2.sop[i] = ((value >> 8) & 0x0F);
1002 		wtp->pcie_dma1_stat2.eop[i] = ((value >> 8) & 0x0F);
1003 		wtp->pcie_dma1_stat2_core.sop[i] = value & 0x0F;
1004 		wtp->pcie_dma1_stat2_core.eop[i] = value & 0x0F;
1005 	}
1006 
1007 	/* Get DMA0 stats3*/
1008 	for (i = 0; i < 2; i++) {
1009 		value = t4_read_reg(padap, A_PCIE_T5_DMA_STAT3 + (i * 0x10));
1010 		wtp->pcie_t5_dma_stat3.sop[i] = value & 0xFF;
1011 		wtp->pcie_t5_dma_stat3.eop[i] = ((value >> 16) & 0xFF);
1012 	}
1013 
1014 	/* Get ULP SE CNT CHx*/
1015 	for (i = 0; i < 4; i++) {
1016 		value = t4_read_reg(padap, A_ULP_TX_SE_CNT_CH0 + (i * 4));
1017 		wtp->ulp_se_cnt_chx.sop[i] = ((value >> 28) & 0x0F);
1018 		wtp->ulp_se_cnt_chx.eop[i] = ((value >> 24) & 0x0F);
1019 	}
1020 
1021 	/* Get TP_DBG_CSIDE registers*/
1022 	for (i = 0; i < 4; i++) {
1023 		t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_CSIDE_RX0 + i),
1024 			       true);
1025 
1026 		wtp->utx_tpcside.sop[i]   = ((value >> 28) & 0xF);/*bits 28:31*/
1027 		wtp->utx_tpcside.eop[i]   = ((value >> 24) & 0xF);/*bits 24:27*/
1028 		wtp->tpcside_rxarb.sop[i] = ((value >> 12) & 0xF);/*bits 12:15*/
1029 		wtp->tpcside_rxarb.eop[i] = ((value >> 8) & 0xF); /*bits 8:11*/
1030 	}
1031 
1032 	for (i = 0; i < 4; i++) {
1033 		t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_ESIDE_PKT0 + i),
1034 			       true);
1035 
1036 
1037 		wtp->tpeside_mps.sop[i] = ((value >> 28) & 0xF); /*bits 28:31*/
1038 		wtp->tpeside_mps.eop[i] = ((value >> 24) & 0xF); /*bits 24:27*/
1039 	}
1040 
1041 	for (i = 0; i < 2; i++) {
1042 		value = t4_read_reg(padap, (A_MPS_TX_SE_CNT_TP01 + (i << 2)));
1043 		wtp->tp_mps.sop[(i*2)]	   = ((value >> 8) & 0xFF); /*bit 8:15*/
1044 		wtp->tp_mps.eop[(i*2)]	   = ((value >> 0) & 0xFF); /*bit 0:7*/
1045 		wtp->tp_mps.sop[(i*2) + 1] = ((value >> 24) & 0xFF);/*bit 24:31
1046 								    */
1047 		wtp->tp_mps.eop[(i*2) + 1] = ((value >> 16) & 0xFF);/*bit 16:23
1048 								    */
1049 	}
1050 
1051 	for (i = 0; i < 2; i++) {
1052 		value = t4_read_reg(padap, (A_MPS_TX_SE_CNT_MAC01 + (i << 2)));
1053 		wtp->mps_xgm.sop[(i*2)]     = ((value >> 8) & 0xFF);/*bit 8:15*/
1054 		wtp->mps_xgm.eop[(i*2)]     = ((value >> 0) & 0xFF); /*bit 0:7*/
1055 		wtp->mps_xgm.sop[(i*2) + 1] = ((value >> 24) & 0xFF);/*bit 24:31
1056 								     */
1057 		wtp->mps_xgm.eop[(i*2) + 1] = ((value >> 16) & 0xFF);/*bit 16:23
1058 								     */
1059 	}
1060 
1061 	/* Get MAC PORTx PKT COUNT*/
1062 	for (i = 0; i < 2; i++) {
1063 		value = t4_read_reg(padap, 0x3081c + ((i * 4) << 12));
1064 		wtp->mac_portx_pkt_count.sop[i] = ((value >> 24) & 0xFF);
1065 		wtp->mac_portx_pkt_count.eop[i] = ((value >> 16) & 0xFF);
1066 		wtp->mac_porrx_pkt_count.sop[i] = ((value >> 8) & 0xFF);
1067 		wtp->mac_porrx_pkt_count.eop[i] = ((value >> 0) & 0xFF);
1068 	}
1069 
1070 	for (i = 0; i < 2; i++) {
1071 		value = t4_read_reg(padap, 0x30f20 + ((i * 4) << 12));
1072 		wtp->mac_portx_aframestra_ok.sop[i] = value & 0xff;
1073 		wtp->mac_portx_aframestra_ok.eop[i] = value & 0xff;
1074 	}
1075 
1076 	/*MAC_PORT_MTIP_1G10G_TX_etherStatsPkts*/
1077 
1078 	for (i = 0; i < 2; i++) {
1079 		value = t4_read_reg(padap, 0x30f60 + ((i * 4) << 12));
1080 		wtp->mac_portx_etherstatspkts.sop[i] = value & 0xff;
1081 		wtp->mac_portx_etherstatspkts.eop[i] = value & 0xff;
1082 	}
1083 
1084 	/*RX path*/
1085 
1086 	value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_7);
1087 	wtp->sge_debug_data_high_indx7.sop[0] = ((value >> 4) & 0x0F);
1088 	wtp->sge_debug_data_high_indx7.eop[0] = ((value >> 0) & 0x0F);
1089 	wtp->sge_debug_data_high_indx7.sop[1] = ((value >> 12) & 0x0F);
1090 	wtp->sge_debug_data_high_indx7.eop[1] = ((value >> 8) & 0x0F);
1091 
1092 	/*Get SGE debug data high index 1*/
1093 	value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_1);
1094 	wtp->sge_debug_data_high_indx1.sop[0] = ((value >> 20) & 0x0F);
1095 	wtp->sge_debug_data_high_indx1.eop[0] = ((value >> 16) & 0x0F);
1096 	wtp->sge_debug_data_high_indx1.sop[1] = ((value >> 28) & 0x0F);
1097 	wtp->sge_debug_data_high_indx1.eop[1] = ((value >> 24) & 0x0F);
1098 
1099 	value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_9);
1100 	wtp->sge_debug_data_high_indx9.sop[0] = ((value >> 20) & 0x0F);
1101 	wtp->sge_debug_data_high_indx9.sop[1] = ((value >> 28) & 0x0F);
1102 
1103 	wtp->sge_debug_data_high_indx9.eop[0] = ((value >> 16) & 0x0F);
1104 	wtp->sge_debug_data_high_indx9.eop[1] = ((value >> 24) & 0x0F);
1105 
1106 	for (i = 0; i < 2; i++) {
1107 		t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_CSIDE_TX0 + i),
1108 			       true);
1109 
1110 		wtp->utx_tpcside_tx.sop[i] = ((value >> 28) & 0xF);/*bits 28:31
1111 								   */
1112 		wtp->utx_tpcside_tx.eop[i]   = ((value >> 24) & 0xF);
1113 	}
1114 
1115 	/*ULP_RX input/output*/
1116 	for (i = 0; i < 2; i++) {
1117 		value = t4_read_reg(padap, (A_ULP_RX_SE_CNT_CH0 + (i*4)));
1118 
1119 		wtp->pmrx_ulprx.sop[i]	  = ((value >> 4) & 0xF); /*bits 4:7*/
1120 		wtp->pmrx_ulprx.eop[i]	  = ((value >> 0) & 0xF); /*bits 0:3*/
1121 		wtp->ulprx_tpcside.sop[i] = ((value >> 28) & 0xF);/*bits 28:31*/
1122 		wtp->ulprx_tpcside.eop[i] = ((value >> 24) & 0xF);/*bits 24:27*/
1123 	}
1124 
1125 	/*Get LE DB response count*/
1126 	value = t4_read_reg(padap, A_LE_DB_REQ_RSP_CNT);
1127 	wtp->le_db_rsp_cnt.sop = value & 0xF;
1128 	wtp->le_db_rsp_cnt.eop = (value >> 16) & 0xF;
1129 
1130 	/*Get TP debug Eside PKTx*/
1131 	for (i = 0; i < 4; i++) {
1132 		t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_ESIDE_PKT0 + i),
1133 			       true);
1134 
1135 		wtp->tp_dbg_eside_pktx.sop[i] = ((value >> 12) & 0xF);
1136 		wtp->tp_dbg_eside_pktx.eop[i] = ((value >> 8) & 0xF);
1137 	}
1138 
1139 	drop = 0;
1140 	/*MPS_RX_SE_CNT_OUT01*/
1141 	value = t4_read_reg(padap, (A_MPS_RX_SE_CNT_OUT01 + (i << 2)));
1142 	wtp->mps_tp.sop[0] = ((value >> 8) & 0xFF); /*bit 8:15*/
1143 	wtp->mps_tp.eop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/
1144 	wtp->mps_tp.sop[1] = ((value >> 24) & 0xFF); /*bit 24:31*/
1145 	wtp->mps_tp.eop[1] = ((value >> 16) & 0xFF); /*bit 16:23*/
1146 
1147 	drop = ptp_mib->TP_MIB_TNL_CNG_DROP_0.value;
1148 	drop += ptp_mib->TP_MIB_TNL_CNG_DROP_1.value;
1149 	drop += ptp_mib->TP_MIB_OFD_CHN_DROP_0.value;
1150 	drop += ptp_mib->TP_MIB_OFD_CHN_DROP_1.value;
1151 	drop += ptp_mib->TP_MIB_FCOE_DROP_0.value;
1152 	drop += ptp_mib->TP_MIB_FCOE_DROP_1.value;
1153 	drop += ptp_mib->TP_MIB_OFD_VLN_DROP_0.value;
1154 	drop += ptp_mib->TP_MIB_OFD_VLN_DROP_1.value;
1155 	drop += ptp_mib->TP_MIB_USM_DROP.value;
1156 
1157 	wtp->mps_tp.drops = drop;
1158 
1159 	drop = 0;
1160 	for (i = 0; i < 8; i++) {
1161 		value = t4_read_reg(padap, (A_MPS_RX_SE_CNT_IN0 + (i << 2)));
1162 
1163 		wtp->xgm_mps.sop[i] = ((value >> 8) & 0xFF); /*bits 8:15*/
1164 		wtp->xgm_mps.eop[i] = ((value >> 0) & 0xFF); /*bits 0:7*/
1165 	}
1166 	for (i = 0; i < 2; i++) {
1167 		value = t4_read_reg(padap, (A_MPS_RX_CLS_DROP_CNT0 + (i << 2)));
1168 		drop += (value & 0xFFFF) + ((value >> 16) & 0xFFFF);
1169 	}
1170 	wtp->xgm_mps.cls_drop = drop & 0xFF;
1171 
1172 	for (i = 0; i < 2; i++) {
1173 		value = t4_read_reg(padap, 0x30e20 + ((i * 4) << 12));
1174 		wtp->mac_porrx_aframestra_ok.sop[i] = value & 0xff;
1175 		wtp->mac_porrx_aframestra_ok.eop[i] = value & 0xff;
1176 	}
1177 
1178 	/*MAC_PORT_MTIP_1G10G_RX_etherStatsPkts*/
1179 	for (i = 0; i < 2; i++) {
1180 		value = t4_read_reg(padap, 0x30e60 + ((i * 4) << 12));
1181 		wtp->mac_porrx_etherstatspkts.sop[i] = value & 0xff;
1182 		wtp->mac_porrx_etherstatspkts.eop[i] = value & 0xff;
1183 	}
1184 
1185 	wtp->sge_pcie_ints.sop[0] = sge_dbg_reg->debug_PD_Req_Int0_cnt;
1186 	wtp->sge_pcie_ints.sop[1] = sge_dbg_reg->debug_PD_Req_Int1_cnt;
1187 	wtp->sge_pcie_ints.sop[2] = sge_dbg_reg->debug_PD_Req_Int2_cnt;
1188 	wtp->sge_pcie_ints.sop[3] = sge_dbg_reg->debug_PD_Req_Int3_cnt;
1189 
1190 	/* Add up the overflow drops on all 4 ports.*/
1191 	drop = 0;
1192 	for (i = 0; i < 2; i++) {
1193 		value = t4_read_reg(padap,
1194 				    (A_MPS_STAT_RX_BG_0_MAC_DROP_FRAME_L +
1195 				     (i << 3)));
1196 		drop += value;
1197 		value = t4_read_reg(padap,
1198 				    (A_MPS_STAT_RX_BG_0_MAC_DROP_FRAME_L +
1199 				     (i << 2)));
1200 		value = t4_read_reg(padap,
1201 				    (A_MPS_STAT_RX_BG_0_LB_DROP_FRAME_L +
1202 				     (i << 3)));
1203 		drop += value;
1204 		value = t4_read_reg(padap,
1205 				    (A_MPS_STAT_RX_BG_0_MAC_DROP_FRAME_L +
1206 				     (i << 2)));
1207 
1208 		value = t4_read_reg(padap,
1209 				    (A_MPS_STAT_RX_BG_0_MAC_TRUNC_FRAME_L +
1210 				     (i << 3)));
1211 		drop += value;
1212 		value = t4_read_reg(padap,
1213 				    (A_MPS_STAT_RX_BG_0_MAC_TRUNC_FRAME_L +
1214 				     (i << 3)));
1215 		value = t4_read_reg(padap,
1216 				    (A_MPS_STAT_RX_BG_0_LB_TRUNC_FRAME_L +
1217 				     (i << 3)));
1218 		drop += value;
1219 		value = t4_read_reg(padap,
1220 				    (A_MPS_STAT_RX_BG_0_MAC_TRUNC_FRAME_L +
1221 				     (i << 3)));
1222 
1223 		value = t4_read_reg(padap,
1224 			(T5_PORT0_REG(A_MPS_PORT_STAT_LB_PORT_DROP_FRAMES) +
1225 			(i * T5_PORT_STRIDE)));
1226 		drop += value;
1227 	}
1228 	wtp->xgm_mps.drop = (drop & 0xFF);
1229 
1230 	/* Add up the MPS errors that should result in dropped packets*/
1231 	err = 0;
1232 	for (i = 0; i < 2; i++) {
1233 
1234 		value = t4_read_reg(padap,
1235 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_ERROR_L) +
1236 			(i * T5_PORT_STRIDE)));
1237 		err += value;
1238 		value = t4_read_reg(padap,
1239 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_ERROR_L) +
1240 			(i * T5_PORT_STRIDE) + 4));
1241 
1242 		value = t4_read_reg(padap,
1243 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_CRC_ERROR_L) +
1244 			(i * T5_PORT_STRIDE)));
1245 		err += value;
1246 		value = t4_read_reg(padap,
1247 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_CRC_ERROR_L) +
1248 			(i * T5_PORT_STRIDE) + 4));
1249 
1250 		value = t4_read_reg(padap,
1251 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_CRC_ERROR_L) +
1252 				     (i * T5_PORT_STRIDE)));
1253 		err += value;
1254 		value = t4_read_reg(padap,
1255 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_CRC_ERROR_L) +
1256 			(i * T5_PORT_STRIDE) + 4));
1257 
1258 		value = t4_read_reg(padap,
1259 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LEN_ERROR_L) +
1260 			(i * T5_PORT_STRIDE)));
1261 		err += value;
1262 		value = t4_read_reg(padap,
1263 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LEN_ERROR_L) +
1264 			(i * T5_PORT_STRIDE) + 4));
1265 
1266 		value = t4_read_reg(padap,
1267 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_SYM_ERROR_L) +
1268 			(i * T5_PORT_STRIDE)));
1269 		err += value;
1270 		value = t4_read_reg(padap,
1271 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_SYM_ERROR_L) +
1272 			(i * T5_PORT_STRIDE) + 4));
1273 
1274 		value = t4_read_reg(padap,
1275 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LESS_64B_L) +
1276 			(i * T5_PORT_STRIDE)));
1277 		err += value;
1278 		value = t4_read_reg(padap,
1279 			(T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LESS_64B_L) +
1280 			(i * T5_PORT_STRIDE) + 4));
1281 	}
1282 	wtp->xgm_mps.err = (err & 0xFF);
1283 
1284 	rc = write_compression_hdr(&scratch_buff, dbg_buff);
1285 
1286 	if (rc)
1287 		goto err1;
1288 
1289 	rc = compress_buff(&scratch_buff, dbg_buff);
1290 
1291 err1:
1292 	release_scratch_buff(&scratch_buff, dbg_buff);
1293 err:
1294 	return rc;
1295 }
1296 
1297 int collect_wtp_data(struct cudbg_init *pdbg_init,
1298 		     struct cudbg_buffer *dbg_buff,
1299 		     struct cudbg_error *cudbg_err)
1300 {
1301 	struct adapter *padap = pdbg_init->adap;
1302 	int rc = -1;
1303 
1304 	if (is_t5(padap))
1305 		rc = t5_wtp_data(pdbg_init, dbg_buff, cudbg_err);
1306 	else if (is_t6(padap))
1307 		rc = t6_wtp_data(pdbg_init, dbg_buff, cudbg_err);
1308 
1309 	return rc;
1310 }
1311