1 /*
2  * This file is part of the Chelsio T4 Ethernet driver for Linux.
3  *
4  * Copyright (c) 2003-2014 Chelsio Communications, Inc. All rights reserved.
5  *
6  * This software is available to you under a choice of one of two
7  * licenses.  You may choose to be licensed under the terms of the GNU
8  * General Public License (GPL) Version 2, available from the file
9  * COPYING in the main directory of this source tree, or the
10  * OpenIB.org BSD license below:
11  *
12  *     Redistribution and use in source and binary forms, with or
13  *     without modification, are permitted provided that the following
14  *     conditions are met:
15  *
16  *      - Redistributions of source code must retain the above
17  *        copyright notice, this list of conditions and the following
18  *        disclaimer.
19  *
20  *      - Redistributions in binary form must reproduce the above
21  *        copyright notice, this list of conditions and the following
22  *        disclaimer in the documentation and/or other materials
23  *        provided with the distribution.
24  *
25  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
26  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
27  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
28  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
29  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
30  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
31  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
32  * SOFTWARE.
33  */
34 
35 #include <linux/seq_file.h>
36 #include <linux/debugfs.h>
37 #include <linux/string_helpers.h>
38 #include <linux/sort.h>
39 #include <linux/ctype.h>
40 
41 #include "cxgb4.h"
42 #include "t4_regs.h"
43 #include "t4_values.h"
44 #include "t4fw_api.h"
45 #include "cxgb4_debugfs.h"
46 #include "clip_tbl.h"
47 #include "l2t.h"
48 #include "cudbg_if.h"
49 #include "cudbg_lib_common.h"
50 #include "cudbg_entity.h"
51 #include "cudbg_lib.h"
52 #include "cxgb4_tc_mqprio.h"
53 
54 /* generic seq_file support for showing a table of size rows x width. */
55 static void *seq_tab_get_idx(struct seq_tab *tb, loff_t pos)
56 {
57 	pos -= tb->skip_first;
58 	return pos >= tb->rows ? NULL : &tb->data[pos * tb->width];
59 }
60 
61 static void *seq_tab_start(struct seq_file *seq, loff_t *pos)
62 {
63 	struct seq_tab *tb = seq->private;
64 
65 	if (tb->skip_first && *pos == 0)
66 		return SEQ_START_TOKEN;
67 
68 	return seq_tab_get_idx(tb, *pos);
69 }
70 
71 static void *seq_tab_next(struct seq_file *seq, void *v, loff_t *pos)
72 {
73 	v = seq_tab_get_idx(seq->private, *pos + 1);
74 	++(*pos);
75 	return v;
76 }
77 
78 static void seq_tab_stop(struct seq_file *seq, void *v)
79 {
80 }
81 
82 static int seq_tab_show(struct seq_file *seq, void *v)
83 {
84 	const struct seq_tab *tb = seq->private;
85 
86 	return tb->show(seq, v, ((char *)v - tb->data) / tb->width);
87 }
88 
89 static const struct seq_operations seq_tab_ops = {
90 	.start = seq_tab_start,
91 	.next  = seq_tab_next,
92 	.stop  = seq_tab_stop,
93 	.show  = seq_tab_show
94 };
95 
96 struct seq_tab *seq_open_tab(struct file *f, unsigned int rows,
97 			     unsigned int width, unsigned int have_header,
98 			     int (*show)(struct seq_file *seq, void *v, int i))
99 {
100 	struct seq_tab *p;
101 
102 	p = __seq_open_private(f, &seq_tab_ops, sizeof(*p) + rows * width);
103 	if (p) {
104 		p->show = show;
105 		p->rows = rows;
106 		p->width = width;
107 		p->skip_first = have_header != 0;
108 	}
109 	return p;
110 }
111 
112 /* Trim the size of a seq_tab to the supplied number of rows.  The operation is
113  * irreversible.
114  */
115 static int seq_tab_trim(struct seq_tab *p, unsigned int new_rows)
116 {
117 	if (new_rows > p->rows)
118 		return -EINVAL;
119 	p->rows = new_rows;
120 	return 0;
121 }
122 
123 static int cim_la_show(struct seq_file *seq, void *v, int idx)
124 {
125 	if (v == SEQ_START_TOKEN)
126 		seq_puts(seq, "Status   Data      PC     LS0Stat  LS0Addr "
127 			 "            LS0Data\n");
128 	else {
129 		const u32 *p = v;
130 
131 		seq_printf(seq,
132 			   "  %02x  %x%07x %x%07x %08x %08x %08x%08x%08x%08x\n",
133 			   (p[0] >> 4) & 0xff, p[0] & 0xf, p[1] >> 4,
134 			   p[1] & 0xf, p[2] >> 4, p[2] & 0xf, p[3], p[4], p[5],
135 			   p[6], p[7]);
136 	}
137 	return 0;
138 }
139 
140 static int cim_la_show_3in1(struct seq_file *seq, void *v, int idx)
141 {
142 	if (v == SEQ_START_TOKEN) {
143 		seq_puts(seq, "Status   Data      PC\n");
144 	} else {
145 		const u32 *p = v;
146 
147 		seq_printf(seq, "  %02x   %08x %08x\n", p[5] & 0xff, p[6],
148 			   p[7]);
149 		seq_printf(seq, "  %02x   %02x%06x %02x%06x\n",
150 			   (p[3] >> 8) & 0xff, p[3] & 0xff, p[4] >> 8,
151 			   p[4] & 0xff, p[5] >> 8);
152 		seq_printf(seq, "  %02x   %x%07x %x%07x\n", (p[0] >> 4) & 0xff,
153 			   p[0] & 0xf, p[1] >> 4, p[1] & 0xf, p[2] >> 4);
154 	}
155 	return 0;
156 }
157 
158 static int cim_la_show_t6(struct seq_file *seq, void *v, int idx)
159 {
160 	if (v == SEQ_START_TOKEN) {
161 		seq_puts(seq, "Status   Inst    Data      PC     LS0Stat  "
162 			 "LS0Addr  LS0Data  LS1Stat  LS1Addr  LS1Data\n");
163 	} else {
164 		const u32 *p = v;
165 
166 		seq_printf(seq, "  %02x   %04x%04x %04x%04x %04x%04x %08x %08x %08x %08x %08x %08x\n",
167 			   (p[9] >> 16) & 0xff,       /* Status */
168 			   p[9] & 0xffff, p[8] >> 16, /* Inst */
169 			   p[8] & 0xffff, p[7] >> 16, /* Data */
170 			   p[7] & 0xffff, p[6] >> 16, /* PC */
171 			   p[2], p[1], p[0],      /* LS0 Stat, Addr and Data */
172 			   p[5], p[4], p[3]);     /* LS1 Stat, Addr and Data */
173 	}
174 	return 0;
175 }
176 
177 static int cim_la_show_pc_t6(struct seq_file *seq, void *v, int idx)
178 {
179 	if (v == SEQ_START_TOKEN) {
180 		seq_puts(seq, "Status   Inst    Data      PC\n");
181 	} else {
182 		const u32 *p = v;
183 
184 		seq_printf(seq, "  %02x   %08x %08x %08x\n",
185 			   p[3] & 0xff, p[2], p[1], p[0]);
186 		seq_printf(seq, "  %02x   %02x%06x %02x%06x %02x%06x\n",
187 			   (p[6] >> 8) & 0xff, p[6] & 0xff, p[5] >> 8,
188 			   p[5] & 0xff, p[4] >> 8, p[4] & 0xff, p[3] >> 8);
189 		seq_printf(seq, "  %02x   %04x%04x %04x%04x %04x%04x\n",
190 			   (p[9] >> 16) & 0xff, p[9] & 0xffff, p[8] >> 16,
191 			   p[8] & 0xffff, p[7] >> 16, p[7] & 0xffff,
192 			   p[6] >> 16);
193 	}
194 	return 0;
195 }
196 
197 static int cim_la_open(struct inode *inode, struct file *file)
198 {
199 	int ret;
200 	unsigned int cfg;
201 	struct seq_tab *p;
202 	struct adapter *adap = inode->i_private;
203 
204 	ret = t4_cim_read(adap, UP_UP_DBG_LA_CFG_A, 1, &cfg);
205 	if (ret)
206 		return ret;
207 
208 	if (is_t6(adap->params.chip)) {
209 		/* +1 to account for integer division of CIMLA_SIZE/10 */
210 		p = seq_open_tab(file, (adap->params.cim_la_size / 10) + 1,
211 				 10 * sizeof(u32), 1,
212 				 cfg & UPDBGLACAPTPCONLY_F ?
213 					cim_la_show_pc_t6 : cim_la_show_t6);
214 	} else {
215 		p = seq_open_tab(file, adap->params.cim_la_size / 8,
216 				 8 * sizeof(u32), 1,
217 				 cfg & UPDBGLACAPTPCONLY_F ? cim_la_show_3in1 :
218 							     cim_la_show);
219 	}
220 	if (!p)
221 		return -ENOMEM;
222 
223 	ret = t4_cim_read_la(adap, (u32 *)p->data, NULL);
224 	if (ret)
225 		seq_release_private(inode, file);
226 	return ret;
227 }
228 
229 static const struct file_operations cim_la_fops = {
230 	.owner   = THIS_MODULE,
231 	.open    = cim_la_open,
232 	.read    = seq_read,
233 	.llseek  = seq_lseek,
234 	.release = seq_release_private
235 };
236 
237 static int cim_pif_la_show(struct seq_file *seq, void *v, int idx)
238 {
239 	const u32 *p = v;
240 
241 	if (v == SEQ_START_TOKEN) {
242 		seq_puts(seq, "Cntl ID DataBE   Addr                 Data\n");
243 	} else if (idx < CIM_PIFLA_SIZE) {
244 		seq_printf(seq, " %02x  %02x  %04x  %08x %08x%08x%08x%08x\n",
245 			   (p[5] >> 22) & 0xff, (p[5] >> 16) & 0x3f,
246 			   p[5] & 0xffff, p[4], p[3], p[2], p[1], p[0]);
247 	} else {
248 		if (idx == CIM_PIFLA_SIZE)
249 			seq_puts(seq, "\nCntl ID               Data\n");
250 		seq_printf(seq, " %02x  %02x %08x%08x%08x%08x\n",
251 			   (p[4] >> 6) & 0xff, p[4] & 0x3f,
252 			   p[3], p[2], p[1], p[0]);
253 	}
254 	return 0;
255 }
256 
257 static int cim_pif_la_open(struct inode *inode, struct file *file)
258 {
259 	struct seq_tab *p;
260 	struct adapter *adap = inode->i_private;
261 
262 	p = seq_open_tab(file, 2 * CIM_PIFLA_SIZE, 6 * sizeof(u32), 1,
263 			 cim_pif_la_show);
264 	if (!p)
265 		return -ENOMEM;
266 
267 	t4_cim_read_pif_la(adap, (u32 *)p->data,
268 			   (u32 *)p->data + 6 * CIM_PIFLA_SIZE, NULL, NULL);
269 	return 0;
270 }
271 
272 static const struct file_operations cim_pif_la_fops = {
273 	.owner   = THIS_MODULE,
274 	.open    = cim_pif_la_open,
275 	.read    = seq_read,
276 	.llseek  = seq_lseek,
277 	.release = seq_release_private
278 };
279 
280 static int cim_ma_la_show(struct seq_file *seq, void *v, int idx)
281 {
282 	const u32 *p = v;
283 
284 	if (v == SEQ_START_TOKEN) {
285 		seq_puts(seq, "\n");
286 	} else if (idx < CIM_MALA_SIZE) {
287 		seq_printf(seq, "%02x%08x%08x%08x%08x\n",
288 			   p[4], p[3], p[2], p[1], p[0]);
289 	} else {
290 		if (idx == CIM_MALA_SIZE)
291 			seq_puts(seq,
292 				 "\nCnt ID Tag UE       Data       RDY VLD\n");
293 		seq_printf(seq, "%3u %2u  %x   %u %08x%08x  %u   %u\n",
294 			   (p[2] >> 10) & 0xff, (p[2] >> 7) & 7,
295 			   (p[2] >> 3) & 0xf, (p[2] >> 2) & 1,
296 			   (p[1] >> 2) | ((p[2] & 3) << 30),
297 			   (p[0] >> 2) | ((p[1] & 3) << 30), (p[0] >> 1) & 1,
298 			   p[0] & 1);
299 	}
300 	return 0;
301 }
302 
303 static int cim_ma_la_open(struct inode *inode, struct file *file)
304 {
305 	struct seq_tab *p;
306 	struct adapter *adap = inode->i_private;
307 
308 	p = seq_open_tab(file, 2 * CIM_MALA_SIZE, 5 * sizeof(u32), 1,
309 			 cim_ma_la_show);
310 	if (!p)
311 		return -ENOMEM;
312 
313 	t4_cim_read_ma_la(adap, (u32 *)p->data,
314 			  (u32 *)p->data + 5 * CIM_MALA_SIZE);
315 	return 0;
316 }
317 
318 static const struct file_operations cim_ma_la_fops = {
319 	.owner   = THIS_MODULE,
320 	.open    = cim_ma_la_open,
321 	.read    = seq_read,
322 	.llseek  = seq_lseek,
323 	.release = seq_release_private
324 };
325 
326 static int cim_qcfg_show(struct seq_file *seq, void *v)
327 {
328 	static const char * const qname[] = {
329 		"TP0", "TP1", "ULP", "SGE0", "SGE1", "NC-SI",
330 		"ULP0", "ULP1", "ULP2", "ULP3", "SGE", "NC-SI",
331 		"SGE0-RX", "SGE1-RX"
332 	};
333 
334 	int i;
335 	struct adapter *adap = seq->private;
336 	u16 base[CIM_NUM_IBQ + CIM_NUM_OBQ_T5];
337 	u16 size[CIM_NUM_IBQ + CIM_NUM_OBQ_T5];
338 	u32 stat[(4 * (CIM_NUM_IBQ + CIM_NUM_OBQ_T5))];
339 	u16 thres[CIM_NUM_IBQ];
340 	u32 obq_wr_t4[2 * CIM_NUM_OBQ], *wr;
341 	u32 obq_wr_t5[2 * CIM_NUM_OBQ_T5];
342 	u32 *p = stat;
343 	int cim_num_obq = is_t4(adap->params.chip) ?
344 				CIM_NUM_OBQ : CIM_NUM_OBQ_T5;
345 
346 	i = t4_cim_read(adap, is_t4(adap->params.chip) ? UP_IBQ_0_RDADDR_A :
347 			UP_IBQ_0_SHADOW_RDADDR_A,
348 			ARRAY_SIZE(stat), stat);
349 	if (!i) {
350 		if (is_t4(adap->params.chip)) {
351 			i = t4_cim_read(adap, UP_OBQ_0_REALADDR_A,
352 					ARRAY_SIZE(obq_wr_t4), obq_wr_t4);
353 			wr = obq_wr_t4;
354 		} else {
355 			i = t4_cim_read(adap, UP_OBQ_0_SHADOW_REALADDR_A,
356 					ARRAY_SIZE(obq_wr_t5), obq_wr_t5);
357 			wr = obq_wr_t5;
358 		}
359 	}
360 	if (i)
361 		return i;
362 
363 	t4_read_cimq_cfg(adap, base, size, thres);
364 
365 	seq_printf(seq,
366 		   "  Queue  Base  Size Thres  RdPtr WrPtr  SOP  EOP Avail\n");
367 	for (i = 0; i < CIM_NUM_IBQ; i++, p += 4)
368 		seq_printf(seq, "%7s %5x %5u %5u %6x  %4x %4u %4u %5u\n",
369 			   qname[i], base[i], size[i], thres[i],
370 			   IBQRDADDR_G(p[0]), IBQWRADDR_G(p[1]),
371 			   QUESOPCNT_G(p[3]), QUEEOPCNT_G(p[3]),
372 			   QUEREMFLITS_G(p[2]) * 16);
373 	for ( ; i < CIM_NUM_IBQ + cim_num_obq; i++, p += 4, wr += 2)
374 		seq_printf(seq, "%7s %5x %5u %12x  %4x %4u %4u %5u\n",
375 			   qname[i], base[i], size[i],
376 			   QUERDADDR_G(p[0]) & 0x3fff, wr[0] - base[i],
377 			   QUESOPCNT_G(p[3]), QUEEOPCNT_G(p[3]),
378 			   QUEREMFLITS_G(p[2]) * 16);
379 	return 0;
380 }
381 DEFINE_SHOW_ATTRIBUTE(cim_qcfg);
382 
383 static int cimq_show(struct seq_file *seq, void *v, int idx)
384 {
385 	const u32 *p = v;
386 
387 	seq_printf(seq, "%#06x: %08x %08x %08x %08x\n", idx * 16, p[0], p[1],
388 		   p[2], p[3]);
389 	return 0;
390 }
391 
392 static int cim_ibq_open(struct inode *inode, struct file *file)
393 {
394 	int ret;
395 	struct seq_tab *p;
396 	unsigned int qid = (uintptr_t)inode->i_private & 7;
397 	struct adapter *adap = inode->i_private - qid;
398 
399 	p = seq_open_tab(file, CIM_IBQ_SIZE, 4 * sizeof(u32), 0, cimq_show);
400 	if (!p)
401 		return -ENOMEM;
402 
403 	ret = t4_read_cim_ibq(adap, qid, (u32 *)p->data, CIM_IBQ_SIZE * 4);
404 	if (ret < 0)
405 		seq_release_private(inode, file);
406 	else
407 		ret = 0;
408 	return ret;
409 }
410 
411 static const struct file_operations cim_ibq_fops = {
412 	.owner   = THIS_MODULE,
413 	.open    = cim_ibq_open,
414 	.read    = seq_read,
415 	.llseek  = seq_lseek,
416 	.release = seq_release_private
417 };
418 
419 static int cim_obq_open(struct inode *inode, struct file *file)
420 {
421 	int ret;
422 	struct seq_tab *p;
423 	unsigned int qid = (uintptr_t)inode->i_private & 7;
424 	struct adapter *adap = inode->i_private - qid;
425 
426 	p = seq_open_tab(file, 6 * CIM_OBQ_SIZE, 4 * sizeof(u32), 0, cimq_show);
427 	if (!p)
428 		return -ENOMEM;
429 
430 	ret = t4_read_cim_obq(adap, qid, (u32 *)p->data, 6 * CIM_OBQ_SIZE * 4);
431 	if (ret < 0) {
432 		seq_release_private(inode, file);
433 	} else {
434 		seq_tab_trim(p, ret / 4);
435 		ret = 0;
436 	}
437 	return ret;
438 }
439 
440 static const struct file_operations cim_obq_fops = {
441 	.owner   = THIS_MODULE,
442 	.open    = cim_obq_open,
443 	.read    = seq_read,
444 	.llseek  = seq_lseek,
445 	.release = seq_release_private
446 };
447 
448 struct field_desc {
449 	const char *name;
450 	unsigned int start;
451 	unsigned int width;
452 };
453 
454 static void field_desc_show(struct seq_file *seq, u64 v,
455 			    const struct field_desc *p)
456 {
457 	char buf[32];
458 	int line_size = 0;
459 
460 	while (p->name) {
461 		u64 mask = (1ULL << p->width) - 1;
462 		int len = scnprintf(buf, sizeof(buf), "%s: %llu", p->name,
463 				    ((unsigned long long)v >> p->start) & mask);
464 
465 		if (line_size + len >= 79) {
466 			line_size = 8;
467 			seq_puts(seq, "\n        ");
468 		}
469 		seq_printf(seq, "%s ", buf);
470 		line_size += len + 1;
471 		p++;
472 	}
473 	seq_putc(seq, '\n');
474 }
475 
476 static struct field_desc tp_la0[] = {
477 	{ "RcfOpCodeOut", 60, 4 },
478 	{ "State", 56, 4 },
479 	{ "WcfState", 52, 4 },
480 	{ "RcfOpcSrcOut", 50, 2 },
481 	{ "CRxError", 49, 1 },
482 	{ "ERxError", 48, 1 },
483 	{ "SanityFailed", 47, 1 },
484 	{ "SpuriousMsg", 46, 1 },
485 	{ "FlushInputMsg", 45, 1 },
486 	{ "FlushInputCpl", 44, 1 },
487 	{ "RssUpBit", 43, 1 },
488 	{ "RssFilterHit", 42, 1 },
489 	{ "Tid", 32, 10 },
490 	{ "InitTcb", 31, 1 },
491 	{ "LineNumber", 24, 7 },
492 	{ "Emsg", 23, 1 },
493 	{ "EdataOut", 22, 1 },
494 	{ "Cmsg", 21, 1 },
495 	{ "CdataOut", 20, 1 },
496 	{ "EreadPdu", 19, 1 },
497 	{ "CreadPdu", 18, 1 },
498 	{ "TunnelPkt", 17, 1 },
499 	{ "RcfPeerFin", 16, 1 },
500 	{ "RcfReasonOut", 12, 4 },
501 	{ "TxCchannel", 10, 2 },
502 	{ "RcfTxChannel", 8, 2 },
503 	{ "RxEchannel", 6, 2 },
504 	{ "RcfRxChannel", 5, 1 },
505 	{ "RcfDataOutSrdy", 4, 1 },
506 	{ "RxDvld", 3, 1 },
507 	{ "RxOoDvld", 2, 1 },
508 	{ "RxCongestion", 1, 1 },
509 	{ "TxCongestion", 0, 1 },
510 	{ NULL }
511 };
512 
513 static int tp_la_show(struct seq_file *seq, void *v, int idx)
514 {
515 	const u64 *p = v;
516 
517 	field_desc_show(seq, *p, tp_la0);
518 	return 0;
519 }
520 
521 static int tp_la_show2(struct seq_file *seq, void *v, int idx)
522 {
523 	const u64 *p = v;
524 
525 	if (idx)
526 		seq_putc(seq, '\n');
527 	field_desc_show(seq, p[0], tp_la0);
528 	if (idx < (TPLA_SIZE / 2 - 1) || p[1] != ~0ULL)
529 		field_desc_show(seq, p[1], tp_la0);
530 	return 0;
531 }
532 
533 static int tp_la_show3(struct seq_file *seq, void *v, int idx)
534 {
535 	static struct field_desc tp_la1[] = {
536 		{ "CplCmdIn", 56, 8 },
537 		{ "CplCmdOut", 48, 8 },
538 		{ "ESynOut", 47, 1 },
539 		{ "EAckOut", 46, 1 },
540 		{ "EFinOut", 45, 1 },
541 		{ "ERstOut", 44, 1 },
542 		{ "SynIn", 43, 1 },
543 		{ "AckIn", 42, 1 },
544 		{ "FinIn", 41, 1 },
545 		{ "RstIn", 40, 1 },
546 		{ "DataIn", 39, 1 },
547 		{ "DataInVld", 38, 1 },
548 		{ "PadIn", 37, 1 },
549 		{ "RxBufEmpty", 36, 1 },
550 		{ "RxDdp", 35, 1 },
551 		{ "RxFbCongestion", 34, 1 },
552 		{ "TxFbCongestion", 33, 1 },
553 		{ "TxPktSumSrdy", 32, 1 },
554 		{ "RcfUlpType", 28, 4 },
555 		{ "Eread", 27, 1 },
556 		{ "Ebypass", 26, 1 },
557 		{ "Esave", 25, 1 },
558 		{ "Static0", 24, 1 },
559 		{ "Cread", 23, 1 },
560 		{ "Cbypass", 22, 1 },
561 		{ "Csave", 21, 1 },
562 		{ "CPktOut", 20, 1 },
563 		{ "RxPagePoolFull", 18, 2 },
564 		{ "RxLpbkPkt", 17, 1 },
565 		{ "TxLpbkPkt", 16, 1 },
566 		{ "RxVfValid", 15, 1 },
567 		{ "SynLearned", 14, 1 },
568 		{ "SetDelEntry", 13, 1 },
569 		{ "SetInvEntry", 12, 1 },
570 		{ "CpcmdDvld", 11, 1 },
571 		{ "CpcmdSave", 10, 1 },
572 		{ "RxPstructsFull", 8, 2 },
573 		{ "EpcmdDvld", 7, 1 },
574 		{ "EpcmdFlush", 6, 1 },
575 		{ "EpcmdTrimPrefix", 5, 1 },
576 		{ "EpcmdTrimPostfix", 4, 1 },
577 		{ "ERssIp4Pkt", 3, 1 },
578 		{ "ERssIp6Pkt", 2, 1 },
579 		{ "ERssTcpUdpPkt", 1, 1 },
580 		{ "ERssFceFipPkt", 0, 1 },
581 		{ NULL }
582 	};
583 	static struct field_desc tp_la2[] = {
584 		{ "CplCmdIn", 56, 8 },
585 		{ "MpsVfVld", 55, 1 },
586 		{ "MpsPf", 52, 3 },
587 		{ "MpsVf", 44, 8 },
588 		{ "SynIn", 43, 1 },
589 		{ "AckIn", 42, 1 },
590 		{ "FinIn", 41, 1 },
591 		{ "RstIn", 40, 1 },
592 		{ "DataIn", 39, 1 },
593 		{ "DataInVld", 38, 1 },
594 		{ "PadIn", 37, 1 },
595 		{ "RxBufEmpty", 36, 1 },
596 		{ "RxDdp", 35, 1 },
597 		{ "RxFbCongestion", 34, 1 },
598 		{ "TxFbCongestion", 33, 1 },
599 		{ "TxPktSumSrdy", 32, 1 },
600 		{ "RcfUlpType", 28, 4 },
601 		{ "Eread", 27, 1 },
602 		{ "Ebypass", 26, 1 },
603 		{ "Esave", 25, 1 },
604 		{ "Static0", 24, 1 },
605 		{ "Cread", 23, 1 },
606 		{ "Cbypass", 22, 1 },
607 		{ "Csave", 21, 1 },
608 		{ "CPktOut", 20, 1 },
609 		{ "RxPagePoolFull", 18, 2 },
610 		{ "RxLpbkPkt", 17, 1 },
611 		{ "TxLpbkPkt", 16, 1 },
612 		{ "RxVfValid", 15, 1 },
613 		{ "SynLearned", 14, 1 },
614 		{ "SetDelEntry", 13, 1 },
615 		{ "SetInvEntry", 12, 1 },
616 		{ "CpcmdDvld", 11, 1 },
617 		{ "CpcmdSave", 10, 1 },
618 		{ "RxPstructsFull", 8, 2 },
619 		{ "EpcmdDvld", 7, 1 },
620 		{ "EpcmdFlush", 6, 1 },
621 		{ "EpcmdTrimPrefix", 5, 1 },
622 		{ "EpcmdTrimPostfix", 4, 1 },
623 		{ "ERssIp4Pkt", 3, 1 },
624 		{ "ERssIp6Pkt", 2, 1 },
625 		{ "ERssTcpUdpPkt", 1, 1 },
626 		{ "ERssFceFipPkt", 0, 1 },
627 		{ NULL }
628 	};
629 	const u64 *p = v;
630 
631 	if (idx)
632 		seq_putc(seq, '\n');
633 	field_desc_show(seq, p[0], tp_la0);
634 	if (idx < (TPLA_SIZE / 2 - 1) || p[1] != ~0ULL)
635 		field_desc_show(seq, p[1], (p[0] & BIT(17)) ? tp_la2 : tp_la1);
636 	return 0;
637 }
638 
639 static int tp_la_open(struct inode *inode, struct file *file)
640 {
641 	struct seq_tab *p;
642 	struct adapter *adap = inode->i_private;
643 
644 	switch (DBGLAMODE_G(t4_read_reg(adap, TP_DBG_LA_CONFIG_A))) {
645 	case 2:
646 		p = seq_open_tab(file, TPLA_SIZE / 2, 2 * sizeof(u64), 0,
647 				 tp_la_show2);
648 		break;
649 	case 3:
650 		p = seq_open_tab(file, TPLA_SIZE / 2, 2 * sizeof(u64), 0,
651 				 tp_la_show3);
652 		break;
653 	default:
654 		p = seq_open_tab(file, TPLA_SIZE, sizeof(u64), 0, tp_la_show);
655 	}
656 	if (!p)
657 		return -ENOMEM;
658 
659 	t4_tp_read_la(adap, (u64 *)p->data, NULL);
660 	return 0;
661 }
662 
663 static ssize_t tp_la_write(struct file *file, const char __user *buf,
664 			   size_t count, loff_t *pos)
665 {
666 	int err;
667 	char s[32];
668 	unsigned long val;
669 	size_t size = min(sizeof(s) - 1, count);
670 	struct adapter *adap = file_inode(file)->i_private;
671 
672 	if (copy_from_user(s, buf, size))
673 		return -EFAULT;
674 	s[size] = '\0';
675 	err = kstrtoul(s, 0, &val);
676 	if (err)
677 		return err;
678 	if (val > 0xffff)
679 		return -EINVAL;
680 	adap->params.tp.la_mask = val << 16;
681 	t4_set_reg_field(adap, TP_DBG_LA_CONFIG_A, 0xffff0000U,
682 			 adap->params.tp.la_mask);
683 	return count;
684 }
685 
686 static const struct file_operations tp_la_fops = {
687 	.owner   = THIS_MODULE,
688 	.open    = tp_la_open,
689 	.read    = seq_read,
690 	.llseek  = seq_lseek,
691 	.release = seq_release_private,
692 	.write   = tp_la_write
693 };
694 
695 static int ulprx_la_show(struct seq_file *seq, void *v, int idx)
696 {
697 	const u32 *p = v;
698 
699 	if (v == SEQ_START_TOKEN)
700 		seq_puts(seq, "      Pcmd        Type   Message"
701 			 "                Data\n");
702 	else
703 		seq_printf(seq, "%08x%08x  %4x  %08x  %08x%08x%08x%08x\n",
704 			   p[1], p[0], p[2], p[3], p[7], p[6], p[5], p[4]);
705 	return 0;
706 }
707 
708 static int ulprx_la_open(struct inode *inode, struct file *file)
709 {
710 	struct seq_tab *p;
711 	struct adapter *adap = inode->i_private;
712 
713 	p = seq_open_tab(file, ULPRX_LA_SIZE, 8 * sizeof(u32), 1,
714 			 ulprx_la_show);
715 	if (!p)
716 		return -ENOMEM;
717 
718 	t4_ulprx_read_la(adap, (u32 *)p->data);
719 	return 0;
720 }
721 
722 static const struct file_operations ulprx_la_fops = {
723 	.owner   = THIS_MODULE,
724 	.open    = ulprx_la_open,
725 	.read    = seq_read,
726 	.llseek  = seq_lseek,
727 	.release = seq_release_private
728 };
729 
730 /* Show the PM memory stats.  These stats include:
731  *
732  * TX:
733  *   Read: memory read operation
734  *   Write Bypass: cut-through
735  *   Bypass + mem: cut-through and save copy
736  *
737  * RX:
738  *   Read: memory read
739  *   Write Bypass: cut-through
740  *   Flush: payload trim or drop
741  */
742 static int pm_stats_show(struct seq_file *seq, void *v)
743 {
744 	static const char * const tx_pm_stats[] = {
745 		"Read:", "Write bypass:", "Write mem:", "Bypass + mem:"
746 	};
747 	static const char * const rx_pm_stats[] = {
748 		"Read:", "Write bypass:", "Write mem:", "Flush:"
749 	};
750 
751 	int i;
752 	u32 tx_cnt[T6_PM_NSTATS], rx_cnt[T6_PM_NSTATS];
753 	u64 tx_cyc[T6_PM_NSTATS], rx_cyc[T6_PM_NSTATS];
754 	struct adapter *adap = seq->private;
755 
756 	t4_pmtx_get_stats(adap, tx_cnt, tx_cyc);
757 	t4_pmrx_get_stats(adap, rx_cnt, rx_cyc);
758 
759 	seq_printf(seq, "%13s %10s  %20s\n", " ", "Tx pcmds", "Tx bytes");
760 	for (i = 0; i < PM_NSTATS - 1; i++)
761 		seq_printf(seq, "%-13s %10u  %20llu\n",
762 			   tx_pm_stats[i], tx_cnt[i], tx_cyc[i]);
763 
764 	seq_printf(seq, "%13s %10s  %20s\n", " ", "Rx pcmds", "Rx bytes");
765 	for (i = 0; i < PM_NSTATS - 1; i++)
766 		seq_printf(seq, "%-13s %10u  %20llu\n",
767 			   rx_pm_stats[i], rx_cnt[i], rx_cyc[i]);
768 
769 	if (CHELSIO_CHIP_VERSION(adap->params.chip) > CHELSIO_T5) {
770 		/* In T5 the granularity of the total wait is too fine.
771 		 * It is not useful as it reaches the max value too fast.
772 		 * Hence display this Input FIFO wait for T6 onwards.
773 		 */
774 		seq_printf(seq, "%13s %10s  %20s\n",
775 			   " ", "Total wait", "Total Occupancy");
776 		seq_printf(seq, "Tx FIFO wait  %10u  %20llu\n",
777 			   tx_cnt[i], tx_cyc[i]);
778 		seq_printf(seq, "Rx FIFO wait  %10u  %20llu\n",
779 			   rx_cnt[i], rx_cyc[i]);
780 
781 		/* Skip index 6 as there is nothing useful ihere */
782 		i += 2;
783 
784 		/* At index 7, a new stat for read latency (count, total wait)
785 		 * is added.
786 		 */
787 		seq_printf(seq, "%13s %10s  %20s\n",
788 			   " ", "Reads", "Total wait");
789 		seq_printf(seq, "Tx latency    %10u  %20llu\n",
790 			   tx_cnt[i], tx_cyc[i]);
791 		seq_printf(seq, "Rx latency    %10u  %20llu\n",
792 			   rx_cnt[i], rx_cyc[i]);
793 	}
794 	return 0;
795 }
796 
797 static int pm_stats_open(struct inode *inode, struct file *file)
798 {
799 	return single_open(file, pm_stats_show, inode->i_private);
800 }
801 
802 static ssize_t pm_stats_clear(struct file *file, const char __user *buf,
803 			      size_t count, loff_t *pos)
804 {
805 	struct adapter *adap = file_inode(file)->i_private;
806 
807 	t4_write_reg(adap, PM_RX_STAT_CONFIG_A, 0);
808 	t4_write_reg(adap, PM_TX_STAT_CONFIG_A, 0);
809 	return count;
810 }
811 
812 static const struct file_operations pm_stats_debugfs_fops = {
813 	.owner   = THIS_MODULE,
814 	.open    = pm_stats_open,
815 	.read    = seq_read,
816 	.llseek  = seq_lseek,
817 	.release = single_release,
818 	.write   = pm_stats_clear
819 };
820 
821 static int tx_rate_show(struct seq_file *seq, void *v)
822 {
823 	u64 nrate[NCHAN], orate[NCHAN];
824 	struct adapter *adap = seq->private;
825 
826 	t4_get_chan_txrate(adap, nrate, orate);
827 	if (adap->params.arch.nchan == NCHAN) {
828 		seq_puts(seq, "              channel 0   channel 1   "
829 			 "channel 2   channel 3\n");
830 		seq_printf(seq, "NIC B/s:     %10llu  %10llu  %10llu  %10llu\n",
831 			   (unsigned long long)nrate[0],
832 			   (unsigned long long)nrate[1],
833 			   (unsigned long long)nrate[2],
834 			   (unsigned long long)nrate[3]);
835 		seq_printf(seq, "Offload B/s: %10llu  %10llu  %10llu  %10llu\n",
836 			   (unsigned long long)orate[0],
837 			   (unsigned long long)orate[1],
838 			   (unsigned long long)orate[2],
839 			   (unsigned long long)orate[3]);
840 	} else {
841 		seq_puts(seq, "              channel 0   channel 1\n");
842 		seq_printf(seq, "NIC B/s:     %10llu  %10llu\n",
843 			   (unsigned long long)nrate[0],
844 			   (unsigned long long)nrate[1]);
845 		seq_printf(seq, "Offload B/s: %10llu  %10llu\n",
846 			   (unsigned long long)orate[0],
847 			   (unsigned long long)orate[1]);
848 	}
849 	return 0;
850 }
851 DEFINE_SHOW_ATTRIBUTE(tx_rate);
852 
853 static int cctrl_tbl_show(struct seq_file *seq, void *v)
854 {
855 	static const char * const dec_fac[] = {
856 		"0.5", "0.5625", "0.625", "0.6875", "0.75", "0.8125", "0.875",
857 		"0.9375" };
858 
859 	int i;
860 	u16 (*incr)[NCCTRL_WIN];
861 	struct adapter *adap = seq->private;
862 
863 	incr = kmalloc_array(NMTUS, sizeof(*incr), GFP_KERNEL);
864 	if (!incr)
865 		return -ENOMEM;
866 
867 	t4_read_cong_tbl(adap, incr);
868 
869 	for (i = 0; i < NCCTRL_WIN; ++i) {
870 		seq_printf(seq, "%2d: %4u %4u %4u %4u %4u %4u %4u %4u\n", i,
871 			   incr[0][i], incr[1][i], incr[2][i], incr[3][i],
872 			   incr[4][i], incr[5][i], incr[6][i], incr[7][i]);
873 		seq_printf(seq, "%8u %4u %4u %4u %4u %4u %4u %4u %5u %s\n",
874 			   incr[8][i], incr[9][i], incr[10][i], incr[11][i],
875 			   incr[12][i], incr[13][i], incr[14][i], incr[15][i],
876 			   adap->params.a_wnd[i],
877 			   dec_fac[adap->params.b_wnd[i]]);
878 	}
879 
880 	kfree(incr);
881 	return 0;
882 }
883 DEFINE_SHOW_ATTRIBUTE(cctrl_tbl);
884 
885 /* Format a value in a unit that differs from the value's native unit by the
886  * given factor.
887  */
888 static char *unit_conv(char *buf, size_t len, unsigned int val,
889 		       unsigned int factor)
890 {
891 	unsigned int rem = val % factor;
892 
893 	if (rem == 0) {
894 		snprintf(buf, len, "%u", val / factor);
895 	} else {
896 		while (rem % 10 == 0)
897 			rem /= 10;
898 		snprintf(buf, len, "%u.%u", val / factor, rem);
899 	}
900 	return buf;
901 }
902 
903 static int clk_show(struct seq_file *seq, void *v)
904 {
905 	char buf[32];
906 	struct adapter *adap = seq->private;
907 	unsigned int cclk_ps = 1000000000 / adap->params.vpd.cclk;  /* in ps */
908 	u32 res = t4_read_reg(adap, TP_TIMER_RESOLUTION_A);
909 	unsigned int tre = TIMERRESOLUTION_G(res);
910 	unsigned int dack_re = DELAYEDACKRESOLUTION_G(res);
911 	unsigned long long tp_tick_us = (cclk_ps << tre) / 1000000; /* in us */
912 
913 	seq_printf(seq, "Core clock period: %s ns\n",
914 		   unit_conv(buf, sizeof(buf), cclk_ps, 1000));
915 	seq_printf(seq, "TP timer tick: %s us\n",
916 		   unit_conv(buf, sizeof(buf), (cclk_ps << tre), 1000000));
917 	seq_printf(seq, "TCP timestamp tick: %s us\n",
918 		   unit_conv(buf, sizeof(buf),
919 			     (cclk_ps << TIMESTAMPRESOLUTION_G(res)), 1000000));
920 	seq_printf(seq, "DACK tick: %s us\n",
921 		   unit_conv(buf, sizeof(buf), (cclk_ps << dack_re), 1000000));
922 	seq_printf(seq, "DACK timer: %u us\n",
923 		   ((cclk_ps << dack_re) / 1000000) *
924 		   t4_read_reg(adap, TP_DACK_TIMER_A));
925 	seq_printf(seq, "Retransmit min: %llu us\n",
926 		   tp_tick_us * t4_read_reg(adap, TP_RXT_MIN_A));
927 	seq_printf(seq, "Retransmit max: %llu us\n",
928 		   tp_tick_us * t4_read_reg(adap, TP_RXT_MAX_A));
929 	seq_printf(seq, "Persist timer min: %llu us\n",
930 		   tp_tick_us * t4_read_reg(adap, TP_PERS_MIN_A));
931 	seq_printf(seq, "Persist timer max: %llu us\n",
932 		   tp_tick_us * t4_read_reg(adap, TP_PERS_MAX_A));
933 	seq_printf(seq, "Keepalive idle timer: %llu us\n",
934 		   tp_tick_us * t4_read_reg(adap, TP_KEEP_IDLE_A));
935 	seq_printf(seq, "Keepalive interval: %llu us\n",
936 		   tp_tick_us * t4_read_reg(adap, TP_KEEP_INTVL_A));
937 	seq_printf(seq, "Initial SRTT: %llu us\n",
938 		   tp_tick_us * INITSRTT_G(t4_read_reg(adap, TP_INIT_SRTT_A)));
939 	seq_printf(seq, "FINWAIT2 timer: %llu us\n",
940 		   tp_tick_us * t4_read_reg(adap, TP_FINWAIT2_TIMER_A));
941 
942 	return 0;
943 }
944 DEFINE_SHOW_ATTRIBUTE(clk);
945 
946 /* Firmware Device Log dump. */
947 static const char * const devlog_level_strings[] = {
948 	[FW_DEVLOG_LEVEL_EMERG]		= "EMERG",
949 	[FW_DEVLOG_LEVEL_CRIT]		= "CRIT",
950 	[FW_DEVLOG_LEVEL_ERR]		= "ERR",
951 	[FW_DEVLOG_LEVEL_NOTICE]	= "NOTICE",
952 	[FW_DEVLOG_LEVEL_INFO]		= "INFO",
953 	[FW_DEVLOG_LEVEL_DEBUG]		= "DEBUG"
954 };
955 
956 static const char * const devlog_facility_strings[] = {
957 	[FW_DEVLOG_FACILITY_CORE]	= "CORE",
958 	[FW_DEVLOG_FACILITY_CF]         = "CF",
959 	[FW_DEVLOG_FACILITY_SCHED]	= "SCHED",
960 	[FW_DEVLOG_FACILITY_TIMER]	= "TIMER",
961 	[FW_DEVLOG_FACILITY_RES]	= "RES",
962 	[FW_DEVLOG_FACILITY_HW]		= "HW",
963 	[FW_DEVLOG_FACILITY_FLR]	= "FLR",
964 	[FW_DEVLOG_FACILITY_DMAQ]	= "DMAQ",
965 	[FW_DEVLOG_FACILITY_PHY]	= "PHY",
966 	[FW_DEVLOG_FACILITY_MAC]	= "MAC",
967 	[FW_DEVLOG_FACILITY_PORT]	= "PORT",
968 	[FW_DEVLOG_FACILITY_VI]		= "VI",
969 	[FW_DEVLOG_FACILITY_FILTER]	= "FILTER",
970 	[FW_DEVLOG_FACILITY_ACL]	= "ACL",
971 	[FW_DEVLOG_FACILITY_TM]		= "TM",
972 	[FW_DEVLOG_FACILITY_QFC]	= "QFC",
973 	[FW_DEVLOG_FACILITY_DCB]	= "DCB",
974 	[FW_DEVLOG_FACILITY_ETH]	= "ETH",
975 	[FW_DEVLOG_FACILITY_OFLD]	= "OFLD",
976 	[FW_DEVLOG_FACILITY_RI]		= "RI",
977 	[FW_DEVLOG_FACILITY_ISCSI]	= "ISCSI",
978 	[FW_DEVLOG_FACILITY_FCOE]	= "FCOE",
979 	[FW_DEVLOG_FACILITY_FOISCSI]	= "FOISCSI",
980 	[FW_DEVLOG_FACILITY_FOFCOE]	= "FOFCOE"
981 };
982 
983 /* Information gathered by Device Log Open routine for the display routine.
984  */
985 struct devlog_info {
986 	unsigned int nentries;		/* number of entries in log[] */
987 	unsigned int first;		/* first [temporal] entry in log[] */
988 	struct fw_devlog_e log[];	/* Firmware Device Log */
989 };
990 
991 /* Dump a Firmaware Device Log entry.
992  */
993 static int devlog_show(struct seq_file *seq, void *v)
994 {
995 	if (v == SEQ_START_TOKEN)
996 		seq_printf(seq, "%10s  %15s  %8s  %8s  %s\n",
997 			   "Seq#", "Tstamp", "Level", "Facility", "Message");
998 	else {
999 		struct devlog_info *dinfo = seq->private;
1000 		int fidx = (uintptr_t)v - 2;
1001 		unsigned long index;
1002 		struct fw_devlog_e *e;
1003 
1004 		/* Get a pointer to the log entry to display.  Skip unused log
1005 		 * entries.
1006 		 */
1007 		index = dinfo->first + fidx;
1008 		if (index >= dinfo->nentries)
1009 			index -= dinfo->nentries;
1010 		e = &dinfo->log[index];
1011 		if (e->timestamp == 0)
1012 			return 0;
1013 
1014 		/* Print the message.  This depends on the firmware using
1015 		 * exactly the same formating strings as the kernel so we may
1016 		 * eventually have to put a format interpreter in here ...
1017 		 */
1018 		seq_printf(seq, "%10d  %15llu  %8s  %8s  ",
1019 			   be32_to_cpu(e->seqno),
1020 			   be64_to_cpu(e->timestamp),
1021 			   (e->level < ARRAY_SIZE(devlog_level_strings)
1022 			    ? devlog_level_strings[e->level]
1023 			    : "UNKNOWN"),
1024 			   (e->facility < ARRAY_SIZE(devlog_facility_strings)
1025 			    ? devlog_facility_strings[e->facility]
1026 			    : "UNKNOWN"));
1027 		seq_printf(seq, e->fmt,
1028 			   be32_to_cpu(e->params[0]),
1029 			   be32_to_cpu(e->params[1]),
1030 			   be32_to_cpu(e->params[2]),
1031 			   be32_to_cpu(e->params[3]),
1032 			   be32_to_cpu(e->params[4]),
1033 			   be32_to_cpu(e->params[5]),
1034 			   be32_to_cpu(e->params[6]),
1035 			   be32_to_cpu(e->params[7]));
1036 	}
1037 	return 0;
1038 }
1039 
1040 /* Sequential File Operations for Device Log.
1041  */
1042 static inline void *devlog_get_idx(struct devlog_info *dinfo, loff_t pos)
1043 {
1044 	if (pos > dinfo->nentries)
1045 		return NULL;
1046 
1047 	return (void *)(uintptr_t)(pos + 1);
1048 }
1049 
1050 static void *devlog_start(struct seq_file *seq, loff_t *pos)
1051 {
1052 	struct devlog_info *dinfo = seq->private;
1053 
1054 	return (*pos
1055 		? devlog_get_idx(dinfo, *pos)
1056 		: SEQ_START_TOKEN);
1057 }
1058 
1059 static void *devlog_next(struct seq_file *seq, void *v, loff_t *pos)
1060 {
1061 	struct devlog_info *dinfo = seq->private;
1062 
1063 	(*pos)++;
1064 	return devlog_get_idx(dinfo, *pos);
1065 }
1066 
1067 static void devlog_stop(struct seq_file *seq, void *v)
1068 {
1069 }
1070 
1071 static const struct seq_operations devlog_seq_ops = {
1072 	.start = devlog_start,
1073 	.next  = devlog_next,
1074 	.stop  = devlog_stop,
1075 	.show  = devlog_show
1076 };
1077 
1078 /* Set up for reading the firmware's device log.  We read the entire log here
1079  * and then display it incrementally in devlog_show().
1080  */
1081 static int devlog_open(struct inode *inode, struct file *file)
1082 {
1083 	struct adapter *adap = inode->i_private;
1084 	struct devlog_params *dparams = &adap->params.devlog;
1085 	struct devlog_info *dinfo;
1086 	unsigned int index;
1087 	u32 fseqno;
1088 	int ret;
1089 
1090 	/* If we don't know where the log is we can't do anything.
1091 	 */
1092 	if (dparams->start == 0)
1093 		return -ENXIO;
1094 
1095 	/* Allocate the space to read in the firmware's device log and set up
1096 	 * for the iterated call to our display function.
1097 	 */
1098 	dinfo = __seq_open_private(file, &devlog_seq_ops,
1099 				   sizeof(*dinfo) + dparams->size);
1100 	if (!dinfo)
1101 		return -ENOMEM;
1102 
1103 	/* Record the basic log buffer information and read in the raw log.
1104 	 */
1105 	dinfo->nentries = (dparams->size / sizeof(struct fw_devlog_e));
1106 	dinfo->first = 0;
1107 	spin_lock(&adap->win0_lock);
1108 	ret = t4_memory_rw(adap, adap->params.drv_memwin, dparams->memtype,
1109 			   dparams->start, dparams->size, (__be32 *)dinfo->log,
1110 			   T4_MEMORY_READ);
1111 	spin_unlock(&adap->win0_lock);
1112 	if (ret) {
1113 		seq_release_private(inode, file);
1114 		return ret;
1115 	}
1116 
1117 	/* Find the earliest (lowest Sequence Number) log entry in the
1118 	 * circular Device Log.
1119 	 */
1120 	for (fseqno = ~((u32)0), index = 0; index < dinfo->nentries; index++) {
1121 		struct fw_devlog_e *e = &dinfo->log[index];
1122 		__u32 seqno;
1123 
1124 		if (e->timestamp == 0)
1125 			continue;
1126 
1127 		seqno = be32_to_cpu(e->seqno);
1128 		if (seqno < fseqno) {
1129 			fseqno = seqno;
1130 			dinfo->first = index;
1131 		}
1132 	}
1133 	return 0;
1134 }
1135 
1136 static const struct file_operations devlog_fops = {
1137 	.owner   = THIS_MODULE,
1138 	.open    = devlog_open,
1139 	.read    = seq_read,
1140 	.llseek  = seq_lseek,
1141 	.release = seq_release_private
1142 };
1143 
1144 /* Show Firmware Mailbox Command/Reply Log
1145  *
1146  * Note that we don't do any locking when dumping the Firmware Mailbox Log so
1147  * it's possible that we can catch things during a log update and therefore
1148  * see partially corrupted log entries.  But it's probably Good Enough(tm).
1149  * If we ever decide that we want to make sure that we're dumping a coherent
1150  * log, we'd need to perform locking in the mailbox logging and in
1151  * mboxlog_open() where we'd need to grab the entire mailbox log in one go
1152  * like we do for the Firmware Device Log.
1153  */
1154 static int mboxlog_show(struct seq_file *seq, void *v)
1155 {
1156 	struct adapter *adapter = seq->private;
1157 	struct mbox_cmd_log *log = adapter->mbox_log;
1158 	struct mbox_cmd *entry;
1159 	int entry_idx, i;
1160 
1161 	if (v == SEQ_START_TOKEN) {
1162 		seq_printf(seq,
1163 			   "%10s  %15s  %5s  %5s  %s\n",
1164 			   "Seq#", "Tstamp", "Atime", "Etime",
1165 			   "Command/Reply");
1166 		return 0;
1167 	}
1168 
1169 	entry_idx = log->cursor + ((uintptr_t)v - 2);
1170 	if (entry_idx >= log->size)
1171 		entry_idx -= log->size;
1172 	entry = mbox_cmd_log_entry(log, entry_idx);
1173 
1174 	/* skip over unused entries */
1175 	if (entry->timestamp == 0)
1176 		return 0;
1177 
1178 	seq_printf(seq, "%10u  %15llu  %5d  %5d",
1179 		   entry->seqno, entry->timestamp,
1180 		   entry->access, entry->execute);
1181 	for (i = 0; i < MBOX_LEN / 8; i++) {
1182 		u64 flit = entry->cmd[i];
1183 		u32 hi = (u32)(flit >> 32);
1184 		u32 lo = (u32)flit;
1185 
1186 		seq_printf(seq, "  %08x %08x", hi, lo);
1187 	}
1188 	seq_puts(seq, "\n");
1189 	return 0;
1190 }
1191 
1192 static inline void *mboxlog_get_idx(struct seq_file *seq, loff_t pos)
1193 {
1194 	struct adapter *adapter = seq->private;
1195 	struct mbox_cmd_log *log = adapter->mbox_log;
1196 
1197 	return ((pos <= log->size) ? (void *)(uintptr_t)(pos + 1) : NULL);
1198 }
1199 
1200 static void *mboxlog_start(struct seq_file *seq, loff_t *pos)
1201 {
1202 	return *pos ? mboxlog_get_idx(seq, *pos) : SEQ_START_TOKEN;
1203 }
1204 
1205 static void *mboxlog_next(struct seq_file *seq, void *v, loff_t *pos)
1206 {
1207 	++*pos;
1208 	return mboxlog_get_idx(seq, *pos);
1209 }
1210 
1211 static void mboxlog_stop(struct seq_file *seq, void *v)
1212 {
1213 }
1214 
1215 static const struct seq_operations mboxlog_seq_ops = {
1216 	.start = mboxlog_start,
1217 	.next  = mboxlog_next,
1218 	.stop  = mboxlog_stop,
1219 	.show  = mboxlog_show
1220 };
1221 
1222 static int mboxlog_open(struct inode *inode, struct file *file)
1223 {
1224 	int res = seq_open(file, &mboxlog_seq_ops);
1225 
1226 	if (!res) {
1227 		struct seq_file *seq = file->private_data;
1228 
1229 		seq->private = inode->i_private;
1230 	}
1231 	return res;
1232 }
1233 
1234 static const struct file_operations mboxlog_fops = {
1235 	.owner   = THIS_MODULE,
1236 	.open    = mboxlog_open,
1237 	.read    = seq_read,
1238 	.llseek  = seq_lseek,
1239 	.release = seq_release,
1240 };
1241 
1242 static int mbox_show(struct seq_file *seq, void *v)
1243 {
1244 	static const char * const owner[] = { "none", "FW", "driver",
1245 					      "unknown", "<unread>" };
1246 
1247 	int i;
1248 	unsigned int mbox = (uintptr_t)seq->private & 7;
1249 	struct adapter *adap = seq->private - mbox;
1250 	void __iomem *addr = adap->regs + PF_REG(mbox, CIM_PF_MAILBOX_DATA_A);
1251 
1252 	/* For T4 we don't have a shadow copy of the Mailbox Control register.
1253 	 * And since reading that real register causes a side effect of
1254 	 * granting ownership, we're best of simply not reading it at all.
1255 	 */
1256 	if (is_t4(adap->params.chip)) {
1257 		i = 4; /* index of "<unread>" */
1258 	} else {
1259 		unsigned int ctrl_reg = CIM_PF_MAILBOX_CTRL_SHADOW_COPY_A;
1260 		void __iomem *ctrl = adap->regs + PF_REG(mbox, ctrl_reg);
1261 
1262 		i = MBOWNER_G(readl(ctrl));
1263 	}
1264 
1265 	seq_printf(seq, "mailbox owned by %s\n\n", owner[i]);
1266 
1267 	for (i = 0; i < MBOX_LEN; i += 8)
1268 		seq_printf(seq, "%016llx\n",
1269 			   (unsigned long long)readq(addr + i));
1270 	return 0;
1271 }
1272 
1273 static int mbox_open(struct inode *inode, struct file *file)
1274 {
1275 	return single_open(file, mbox_show, inode->i_private);
1276 }
1277 
1278 static ssize_t mbox_write(struct file *file, const char __user *buf,
1279 			  size_t count, loff_t *pos)
1280 {
1281 	int i;
1282 	char c = '\n', s[256];
1283 	unsigned long long data[8];
1284 	const struct inode *ino;
1285 	unsigned int mbox;
1286 	struct adapter *adap;
1287 	void __iomem *addr;
1288 	void __iomem *ctrl;
1289 
1290 	if (count > sizeof(s) - 1 || !count)
1291 		return -EINVAL;
1292 	if (copy_from_user(s, buf, count))
1293 		return -EFAULT;
1294 	s[count] = '\0';
1295 
1296 	if (sscanf(s, "%llx %llx %llx %llx %llx %llx %llx %llx%c", &data[0],
1297 		   &data[1], &data[2], &data[3], &data[4], &data[5], &data[6],
1298 		   &data[7], &c) < 8 || c != '\n')
1299 		return -EINVAL;
1300 
1301 	ino = file_inode(file);
1302 	mbox = (uintptr_t)ino->i_private & 7;
1303 	adap = ino->i_private - mbox;
1304 	addr = adap->regs + PF_REG(mbox, CIM_PF_MAILBOX_DATA_A);
1305 	ctrl = addr + MBOX_LEN;
1306 
1307 	if (MBOWNER_G(readl(ctrl)) != X_MBOWNER_PL)
1308 		return -EBUSY;
1309 
1310 	for (i = 0; i < 8; i++)
1311 		writeq(data[i], addr + 8 * i);
1312 
1313 	writel(MBMSGVALID_F | MBOWNER_V(X_MBOWNER_FW), ctrl);
1314 	return count;
1315 }
1316 
1317 static const struct file_operations mbox_debugfs_fops = {
1318 	.owner   = THIS_MODULE,
1319 	.open    = mbox_open,
1320 	.read    = seq_read,
1321 	.llseek  = seq_lseek,
1322 	.release = single_release,
1323 	.write   = mbox_write
1324 };
1325 
1326 static int mps_trc_show(struct seq_file *seq, void *v)
1327 {
1328 	int enabled, i;
1329 	struct trace_params tp;
1330 	unsigned int trcidx = (uintptr_t)seq->private & 3;
1331 	struct adapter *adap = seq->private - trcidx;
1332 
1333 	t4_get_trace_filter(adap, &tp, trcidx, &enabled);
1334 	if (!enabled) {
1335 		seq_puts(seq, "tracer is disabled\n");
1336 		return 0;
1337 	}
1338 
1339 	if (tp.skip_ofst * 8 >= TRACE_LEN) {
1340 		dev_err(adap->pdev_dev, "illegal trace pattern skip offset\n");
1341 		return -EINVAL;
1342 	}
1343 	if (tp.port < 8) {
1344 		i = adap->chan_map[tp.port & 3];
1345 		if (i >= MAX_NPORTS) {
1346 			dev_err(adap->pdev_dev, "tracer %u is assigned "
1347 				"to non-existing port\n", trcidx);
1348 			return -EINVAL;
1349 		}
1350 		seq_printf(seq, "tracer is capturing %s %s, ",
1351 			   adap->port[i]->name, tp.port < 4 ? "Rx" : "Tx");
1352 	} else
1353 		seq_printf(seq, "tracer is capturing loopback %d, ",
1354 			   tp.port - 8);
1355 	seq_printf(seq, "snap length: %u, min length: %u\n", tp.snap_len,
1356 		   tp.min_len);
1357 	seq_printf(seq, "packets captured %smatch filter\n",
1358 		   tp.invert ? "do not " : "");
1359 
1360 	if (tp.skip_ofst) {
1361 		seq_puts(seq, "filter pattern: ");
1362 		for (i = 0; i < tp.skip_ofst * 2; i += 2)
1363 			seq_printf(seq, "%08x%08x", tp.data[i], tp.data[i + 1]);
1364 		seq_putc(seq, '/');
1365 		for (i = 0; i < tp.skip_ofst * 2; i += 2)
1366 			seq_printf(seq, "%08x%08x", tp.mask[i], tp.mask[i + 1]);
1367 		seq_puts(seq, "@0\n");
1368 	}
1369 
1370 	seq_puts(seq, "filter pattern: ");
1371 	for (i = tp.skip_ofst * 2; i < TRACE_LEN / 4; i += 2)
1372 		seq_printf(seq, "%08x%08x", tp.data[i], tp.data[i + 1]);
1373 	seq_putc(seq, '/');
1374 	for (i = tp.skip_ofst * 2; i < TRACE_LEN / 4; i += 2)
1375 		seq_printf(seq, "%08x%08x", tp.mask[i], tp.mask[i + 1]);
1376 	seq_printf(seq, "@%u\n", (tp.skip_ofst + tp.skip_len) * 8);
1377 	return 0;
1378 }
1379 
1380 static int mps_trc_open(struct inode *inode, struct file *file)
1381 {
1382 	return single_open(file, mps_trc_show, inode->i_private);
1383 }
1384 
1385 static unsigned int xdigit2int(unsigned char c)
1386 {
1387 	return isdigit(c) ? c - '0' : tolower(c) - 'a' + 10;
1388 }
1389 
1390 #define TRC_PORT_NONE 0xff
1391 #define TRC_RSS_ENABLE 0x33
1392 #define TRC_RSS_DISABLE 0x13
1393 
1394 /* Set an MPS trace filter.  Syntax is:
1395  *
1396  * disable
1397  *
1398  * to disable tracing, or
1399  *
1400  * interface qid=<qid no> [snaplen=<val>] [minlen=<val>] [not] [<pattern>]...
1401  *
1402  * where interface is one of rxN, txN, or loopbackN, N = 0..3, qid can be one
1403  * of the NIC's response qid obtained from sge_qinfo and pattern has the form
1404  *
1405  * <pattern data>[/<pattern mask>][@<anchor>]
1406  *
1407  * Up to 2 filter patterns can be specified.  If 2 are supplied the first one
1408  * must be anchored at 0.  An omitted mask is taken as a mask of 1s, an omitted
1409  * anchor is taken as 0.
1410  */
1411 static ssize_t mps_trc_write(struct file *file, const char __user *buf,
1412 			     size_t count, loff_t *pos)
1413 {
1414 	int i, enable, ret;
1415 	u32 *data, *mask;
1416 	struct trace_params tp;
1417 	const struct inode *ino;
1418 	unsigned int trcidx;
1419 	char *s, *p, *word, *end;
1420 	struct adapter *adap;
1421 	u32 j;
1422 
1423 	ino = file_inode(file);
1424 	trcidx = (uintptr_t)ino->i_private & 3;
1425 	adap = ino->i_private - trcidx;
1426 
1427 	/* Don't accept input more than 1K, can't be anything valid except lots
1428 	 * of whitespace.  Well, use less.
1429 	 */
1430 	if (count > 1024)
1431 		return -EFBIG;
1432 	p = s = kzalloc(count + 1, GFP_USER);
1433 	if (!s)
1434 		return -ENOMEM;
1435 	if (copy_from_user(s, buf, count)) {
1436 		count = -EFAULT;
1437 		goto out;
1438 	}
1439 
1440 	if (s[count - 1] == '\n')
1441 		s[count - 1] = '\0';
1442 
1443 	enable = strcmp("disable", s) != 0;
1444 	if (!enable)
1445 		goto apply;
1446 
1447 	/* enable or disable trace multi rss filter */
1448 	if (adap->trace_rss)
1449 		t4_write_reg(adap, MPS_TRC_CFG_A, TRC_RSS_ENABLE);
1450 	else
1451 		t4_write_reg(adap, MPS_TRC_CFG_A, TRC_RSS_DISABLE);
1452 
1453 	memset(&tp, 0, sizeof(tp));
1454 	tp.port = TRC_PORT_NONE;
1455 	i = 0;	/* counts pattern nibbles */
1456 
1457 	while (p) {
1458 		while (isspace(*p))
1459 			p++;
1460 		word = strsep(&p, " ");
1461 		if (!*word)
1462 			break;
1463 
1464 		if (!strncmp(word, "qid=", 4)) {
1465 			end = (char *)word + 4;
1466 			ret = kstrtouint(end, 10, &j);
1467 			if (ret)
1468 				goto out;
1469 			if (!adap->trace_rss) {
1470 				t4_write_reg(adap, MPS_T5_TRC_RSS_CONTROL_A, j);
1471 				continue;
1472 			}
1473 
1474 			switch (trcidx) {
1475 			case 0:
1476 				t4_write_reg(adap, MPS_TRC_RSS_CONTROL_A, j);
1477 				break;
1478 			case 1:
1479 				t4_write_reg(adap,
1480 					     MPS_TRC_FILTER1_RSS_CONTROL_A, j);
1481 				break;
1482 			case 2:
1483 				t4_write_reg(adap,
1484 					     MPS_TRC_FILTER2_RSS_CONTROL_A, j);
1485 				break;
1486 			case 3:
1487 				t4_write_reg(adap,
1488 					     MPS_TRC_FILTER3_RSS_CONTROL_A, j);
1489 				break;
1490 			}
1491 			continue;
1492 		}
1493 		if (!strncmp(word, "snaplen=", 8)) {
1494 			end = (char *)word + 8;
1495 			ret = kstrtouint(end, 10, &j);
1496 			if (ret || j > 9600) {
1497 inval:				count = -EINVAL;
1498 				goto out;
1499 			}
1500 			tp.snap_len = j;
1501 			continue;
1502 		}
1503 		if (!strncmp(word, "minlen=", 7)) {
1504 			end = (char *)word + 7;
1505 			ret = kstrtouint(end, 10, &j);
1506 			if (ret || j > TFMINPKTSIZE_M)
1507 				goto inval;
1508 			tp.min_len = j;
1509 			continue;
1510 		}
1511 		if (!strcmp(word, "not")) {
1512 			tp.invert = !tp.invert;
1513 			continue;
1514 		}
1515 		if (!strncmp(word, "loopback", 8) && tp.port == TRC_PORT_NONE) {
1516 			if (word[8] < '0' || word[8] > '3' || word[9])
1517 				goto inval;
1518 			tp.port = word[8] - '0' + 8;
1519 			continue;
1520 		}
1521 		if (!strncmp(word, "tx", 2) && tp.port == TRC_PORT_NONE) {
1522 			if (word[2] < '0' || word[2] > '3' || word[3])
1523 				goto inval;
1524 			tp.port = word[2] - '0' + 4;
1525 			if (adap->chan_map[tp.port & 3] >= MAX_NPORTS)
1526 				goto inval;
1527 			continue;
1528 		}
1529 		if (!strncmp(word, "rx", 2) && tp.port == TRC_PORT_NONE) {
1530 			if (word[2] < '0' || word[2] > '3' || word[3])
1531 				goto inval;
1532 			tp.port = word[2] - '0';
1533 			if (adap->chan_map[tp.port] >= MAX_NPORTS)
1534 				goto inval;
1535 			continue;
1536 		}
1537 		if (!isxdigit(*word))
1538 			goto inval;
1539 
1540 		/* we have found a trace pattern */
1541 		if (i) {                            /* split pattern */
1542 			if (tp.skip_len)            /* too many splits */
1543 				goto inval;
1544 			tp.skip_ofst = i / 16;
1545 		}
1546 
1547 		data = &tp.data[i / 8];
1548 		mask = &tp.mask[i / 8];
1549 		j = i;
1550 
1551 		while (isxdigit(*word)) {
1552 			if (i >= TRACE_LEN * 2) {
1553 				count = -EFBIG;
1554 				goto out;
1555 			}
1556 			*data = (*data << 4) + xdigit2int(*word++);
1557 			if (++i % 8 == 0)
1558 				data++;
1559 		}
1560 		if (*word == '/') {
1561 			word++;
1562 			while (isxdigit(*word)) {
1563 				if (j >= i)         /* mask longer than data */
1564 					goto inval;
1565 				*mask = (*mask << 4) + xdigit2int(*word++);
1566 				if (++j % 8 == 0)
1567 					mask++;
1568 			}
1569 			if (i != j)                 /* mask shorter than data */
1570 				goto inval;
1571 		} else {                            /* no mask, use all 1s */
1572 			for ( ; i - j >= 8; j += 8)
1573 				*mask++ = 0xffffffff;
1574 			if (i % 8)
1575 				*mask = (1 << (i % 8) * 4) - 1;
1576 		}
1577 		if (*word == '@') {
1578 			end = (char *)word + 1;
1579 			ret = kstrtouint(end, 10, &j);
1580 			if (*end && *end != '\n')
1581 				goto inval;
1582 			if (j & 7)          /* doesn't start at multiple of 8 */
1583 				goto inval;
1584 			j /= 8;
1585 			if (j < tp.skip_ofst)     /* overlaps earlier pattern */
1586 				goto inval;
1587 			if (j - tp.skip_ofst > 31)            /* skip too big */
1588 				goto inval;
1589 			tp.skip_len = j - tp.skip_ofst;
1590 		}
1591 		if (i % 8) {
1592 			*data <<= (8 - i % 8) * 4;
1593 			*mask <<= (8 - i % 8) * 4;
1594 			i = (i + 15) & ~15;         /* 8-byte align */
1595 		}
1596 	}
1597 
1598 	if (tp.port == TRC_PORT_NONE)
1599 		goto inval;
1600 
1601 apply:
1602 	i = t4_set_trace_filter(adap, &tp, trcidx, enable);
1603 	if (i)
1604 		count = i;
1605 out:
1606 	kfree(s);
1607 	return count;
1608 }
1609 
1610 static const struct file_operations mps_trc_debugfs_fops = {
1611 	.owner   = THIS_MODULE,
1612 	.open    = mps_trc_open,
1613 	.read    = seq_read,
1614 	.llseek  = seq_lseek,
1615 	.release = single_release,
1616 	.write   = mps_trc_write
1617 };
1618 
1619 static ssize_t flash_read(struct file *file, char __user *buf, size_t count,
1620 			  loff_t *ppos)
1621 {
1622 	loff_t pos = *ppos;
1623 	loff_t avail = file_inode(file)->i_size;
1624 	struct adapter *adap = file->private_data;
1625 
1626 	if (pos < 0)
1627 		return -EINVAL;
1628 	if (pos >= avail)
1629 		return 0;
1630 	if (count > avail - pos)
1631 		count = avail - pos;
1632 
1633 	while (count) {
1634 		size_t len;
1635 		int ret, ofst;
1636 		u8 data[256];
1637 
1638 		ofst = pos & 3;
1639 		len = min(count + ofst, sizeof(data));
1640 		ret = t4_read_flash(adap, pos - ofst, (len + 3) / 4,
1641 				    (u32 *)data, 1);
1642 		if (ret)
1643 			return ret;
1644 
1645 		len -= ofst;
1646 		if (copy_to_user(buf, data + ofst, len))
1647 			return -EFAULT;
1648 
1649 		buf += len;
1650 		pos += len;
1651 		count -= len;
1652 	}
1653 	count = pos - *ppos;
1654 	*ppos = pos;
1655 	return count;
1656 }
1657 
1658 static const struct file_operations flash_debugfs_fops = {
1659 	.owner   = THIS_MODULE,
1660 	.open    = mem_open,
1661 	.read    = flash_read,
1662 	.llseek  = default_llseek,
1663 };
1664 
1665 static inline void tcamxy2valmask(u64 x, u64 y, u8 *addr, u64 *mask)
1666 {
1667 	*mask = x | y;
1668 	y = (__force u64)cpu_to_be64(y);
1669 	memcpy(addr, (char *)&y + 2, ETH_ALEN);
1670 }
1671 
1672 static int mps_tcam_show(struct seq_file *seq, void *v)
1673 {
1674 	struct adapter *adap = seq->private;
1675 	unsigned int chip_ver = CHELSIO_CHIP_VERSION(adap->params.chip);
1676 	if (v == SEQ_START_TOKEN) {
1677 		if (chip_ver > CHELSIO_T5) {
1678 			seq_puts(seq, "Idx  Ethernet address     Mask     "
1679 				 "  VNI   Mask   IVLAN Vld "
1680 				 "DIP_Hit   Lookup  Port "
1681 				 "Vld Ports PF  VF                           "
1682 				 "Replication                                "
1683 				 "    P0 P1 P2 P3  ML\n");
1684 		} else {
1685 			if (adap->params.arch.mps_rplc_size > 128)
1686 				seq_puts(seq, "Idx  Ethernet address     Mask     "
1687 					 "Vld Ports PF  VF                           "
1688 					 "Replication                                "
1689 					 "    P0 P1 P2 P3  ML\n");
1690 			else
1691 				seq_puts(seq, "Idx  Ethernet address     Mask     "
1692 					 "Vld Ports PF  VF              Replication"
1693 					 "	         P0 P1 P2 P3  ML\n");
1694 		}
1695 	} else {
1696 		u64 mask;
1697 		u8 addr[ETH_ALEN];
1698 		bool replicate, dip_hit = false, vlan_vld = false;
1699 		unsigned int idx = (uintptr_t)v - 2;
1700 		u64 tcamy, tcamx, val;
1701 		u32 cls_lo, cls_hi, ctl, data2, vnix = 0, vniy = 0;
1702 		u32 rplc[8] = {0};
1703 		u8 lookup_type = 0, port_num = 0;
1704 		u16 ivlan = 0;
1705 
1706 		if (chip_ver > CHELSIO_T5) {
1707 			/* CtlCmdType - 0: Read, 1: Write
1708 			 * CtlTcamSel - 0: TCAM0, 1: TCAM1
1709 			 * CtlXYBitSel- 0: Y bit, 1: X bit
1710 			 */
1711 
1712 			/* Read tcamy */
1713 			ctl = CTLCMDTYPE_V(0) | CTLXYBITSEL_V(0);
1714 			if (idx < 256)
1715 				ctl |= CTLTCAMINDEX_V(idx) | CTLTCAMSEL_V(0);
1716 			else
1717 				ctl |= CTLTCAMINDEX_V(idx - 256) |
1718 				       CTLTCAMSEL_V(1);
1719 			t4_write_reg(adap, MPS_CLS_TCAM_DATA2_CTL_A, ctl);
1720 			val = t4_read_reg(adap, MPS_CLS_TCAM_DATA1_A);
1721 			tcamy = DMACH_G(val) << 32;
1722 			tcamy |= t4_read_reg(adap, MPS_CLS_TCAM_DATA0_A);
1723 			data2 = t4_read_reg(adap, MPS_CLS_TCAM_DATA2_CTL_A);
1724 			lookup_type = DATALKPTYPE_G(data2);
1725 			/* 0 - Outer header, 1 - Inner header
1726 			 * [71:48] bit locations are overloaded for
1727 			 * outer vs. inner lookup types.
1728 			 */
1729 			if (lookup_type && (lookup_type != DATALKPTYPE_M)) {
1730 				/* Inner header VNI */
1731 				vniy = (data2 & DATAVIDH2_F) |
1732 				       (DATAVIDH1_G(data2) << 16) | VIDL_G(val);
1733 				dip_hit = data2 & DATADIPHIT_F;
1734 			} else {
1735 				vlan_vld = data2 & DATAVIDH2_F;
1736 				ivlan = VIDL_G(val);
1737 			}
1738 			port_num = DATAPORTNUM_G(data2);
1739 
1740 			/* Read tcamx. Change the control param */
1741 			vnix = 0;
1742 			ctl |= CTLXYBITSEL_V(1);
1743 			t4_write_reg(adap, MPS_CLS_TCAM_DATA2_CTL_A, ctl);
1744 			val = t4_read_reg(adap, MPS_CLS_TCAM_DATA1_A);
1745 			tcamx = DMACH_G(val) << 32;
1746 			tcamx |= t4_read_reg(adap, MPS_CLS_TCAM_DATA0_A);
1747 			data2 = t4_read_reg(adap, MPS_CLS_TCAM_DATA2_CTL_A);
1748 			if (lookup_type && (lookup_type != DATALKPTYPE_M)) {
1749 				/* Inner header VNI mask */
1750 				vnix = (data2 & DATAVIDH2_F) |
1751 				       (DATAVIDH1_G(data2) << 16) | VIDL_G(val);
1752 			}
1753 		} else {
1754 			tcamy = t4_read_reg64(adap, MPS_CLS_TCAM_Y_L(idx));
1755 			tcamx = t4_read_reg64(adap, MPS_CLS_TCAM_X_L(idx));
1756 		}
1757 
1758 		cls_lo = t4_read_reg(adap, MPS_CLS_SRAM_L(idx));
1759 		cls_hi = t4_read_reg(adap, MPS_CLS_SRAM_H(idx));
1760 
1761 		if (tcamx & tcamy) {
1762 			seq_printf(seq, "%3u         -\n", idx);
1763 			goto out;
1764 		}
1765 
1766 		rplc[0] = rplc[1] = rplc[2] = rplc[3] = 0;
1767 		if (chip_ver > CHELSIO_T5)
1768 			replicate = (cls_lo & T6_REPLICATE_F);
1769 		else
1770 			replicate = (cls_lo & REPLICATE_F);
1771 
1772 		if (replicate) {
1773 			struct fw_ldst_cmd ldst_cmd;
1774 			int ret;
1775 			struct fw_ldst_mps_rplc mps_rplc;
1776 			u32 ldst_addrspc;
1777 
1778 			memset(&ldst_cmd, 0, sizeof(ldst_cmd));
1779 			ldst_addrspc =
1780 				FW_LDST_CMD_ADDRSPACE_V(FW_LDST_ADDRSPC_MPS);
1781 			ldst_cmd.op_to_addrspace =
1782 				htonl(FW_CMD_OP_V(FW_LDST_CMD) |
1783 				      FW_CMD_REQUEST_F |
1784 				      FW_CMD_READ_F |
1785 				      ldst_addrspc);
1786 			ldst_cmd.cycles_to_len16 = htonl(FW_LEN16(ldst_cmd));
1787 			ldst_cmd.u.mps.rplc.fid_idx =
1788 				htons(FW_LDST_CMD_FID_V(FW_LDST_MPS_RPLC) |
1789 				      FW_LDST_CMD_IDX_V(idx));
1790 			ret = t4_wr_mbox(adap, adap->mbox, &ldst_cmd,
1791 					 sizeof(ldst_cmd), &ldst_cmd);
1792 			if (ret)
1793 				dev_warn(adap->pdev_dev, "Can't read MPS "
1794 					 "replication map for idx %d: %d\n",
1795 					 idx, -ret);
1796 			else {
1797 				mps_rplc = ldst_cmd.u.mps.rplc;
1798 				rplc[0] = ntohl(mps_rplc.rplc31_0);
1799 				rplc[1] = ntohl(mps_rplc.rplc63_32);
1800 				rplc[2] = ntohl(mps_rplc.rplc95_64);
1801 				rplc[3] = ntohl(mps_rplc.rplc127_96);
1802 				if (adap->params.arch.mps_rplc_size > 128) {
1803 					rplc[4] = ntohl(mps_rplc.rplc159_128);
1804 					rplc[5] = ntohl(mps_rplc.rplc191_160);
1805 					rplc[6] = ntohl(mps_rplc.rplc223_192);
1806 					rplc[7] = ntohl(mps_rplc.rplc255_224);
1807 				}
1808 			}
1809 		}
1810 
1811 		tcamxy2valmask(tcamx, tcamy, addr, &mask);
1812 		if (chip_ver > CHELSIO_T5) {
1813 			/* Inner header lookup */
1814 			if (lookup_type && (lookup_type != DATALKPTYPE_M)) {
1815 				seq_printf(seq,
1816 					   "%3u %pM %012llx %06x %06x    -    -   %3c      'I'  %4x   %3c   %#x%4u%4d",
1817 					   idx, addr,
1818 					   (unsigned long long)mask,
1819 					   vniy, (vnix | vniy),
1820 					   dip_hit ? 'Y' : 'N',
1821 					   port_num,
1822 					   (cls_lo & T6_SRAM_VLD_F) ? 'Y' : 'N',
1823 					   PORTMAP_G(cls_hi),
1824 					   T6_PF_G(cls_lo),
1825 					   (cls_lo & T6_VF_VALID_F) ?
1826 					   T6_VF_G(cls_lo) : -1);
1827 			} else {
1828 				seq_printf(seq,
1829 					   "%3u %pM %012llx    -       -   ",
1830 					   idx, addr,
1831 					   (unsigned long long)mask);
1832 
1833 				if (vlan_vld)
1834 					seq_printf(seq, "%4u   Y     ", ivlan);
1835 				else
1836 					seq_puts(seq, "  -    N     ");
1837 
1838 				seq_printf(seq,
1839 					   "-      %3c  %4x   %3c   %#x%4u%4d",
1840 					   lookup_type ? 'I' : 'O', port_num,
1841 					   (cls_lo & T6_SRAM_VLD_F) ? 'Y' : 'N',
1842 					   PORTMAP_G(cls_hi),
1843 					   T6_PF_G(cls_lo),
1844 					   (cls_lo & T6_VF_VALID_F) ?
1845 					   T6_VF_G(cls_lo) : -1);
1846 			}
1847 		} else
1848 			seq_printf(seq, "%3u %pM %012llx%3c   %#x%4u%4d",
1849 				   idx, addr, (unsigned long long)mask,
1850 				   (cls_lo & SRAM_VLD_F) ? 'Y' : 'N',
1851 				   PORTMAP_G(cls_hi),
1852 				   PF_G(cls_lo),
1853 				   (cls_lo & VF_VALID_F) ? VF_G(cls_lo) : -1);
1854 
1855 		if (replicate) {
1856 			if (adap->params.arch.mps_rplc_size > 128)
1857 				seq_printf(seq, " %08x %08x %08x %08x "
1858 					   "%08x %08x %08x %08x",
1859 					   rplc[7], rplc[6], rplc[5], rplc[4],
1860 					   rplc[3], rplc[2], rplc[1], rplc[0]);
1861 			else
1862 				seq_printf(seq, " %08x %08x %08x %08x",
1863 					   rplc[3], rplc[2], rplc[1], rplc[0]);
1864 		} else {
1865 			if (adap->params.arch.mps_rplc_size > 128)
1866 				seq_printf(seq, "%72c", ' ');
1867 			else
1868 				seq_printf(seq, "%36c", ' ');
1869 		}
1870 
1871 		if (chip_ver > CHELSIO_T5)
1872 			seq_printf(seq, "%4u%3u%3u%3u %#x\n",
1873 				   T6_SRAM_PRIO0_G(cls_lo),
1874 				   T6_SRAM_PRIO1_G(cls_lo),
1875 				   T6_SRAM_PRIO2_G(cls_lo),
1876 				   T6_SRAM_PRIO3_G(cls_lo),
1877 				   (cls_lo >> T6_MULTILISTEN0_S) & 0xf);
1878 		else
1879 			seq_printf(seq, "%4u%3u%3u%3u %#x\n",
1880 				   SRAM_PRIO0_G(cls_lo), SRAM_PRIO1_G(cls_lo),
1881 				   SRAM_PRIO2_G(cls_lo), SRAM_PRIO3_G(cls_lo),
1882 				   (cls_lo >> MULTILISTEN0_S) & 0xf);
1883 	}
1884 out:	return 0;
1885 }
1886 
1887 static inline void *mps_tcam_get_idx(struct seq_file *seq, loff_t pos)
1888 {
1889 	struct adapter *adap = seq->private;
1890 	int max_mac_addr = is_t4(adap->params.chip) ?
1891 				NUM_MPS_CLS_SRAM_L_INSTANCES :
1892 				NUM_MPS_T5_CLS_SRAM_L_INSTANCES;
1893 	return ((pos <= max_mac_addr) ? (void *)(uintptr_t)(pos + 1) : NULL);
1894 }
1895 
1896 static void *mps_tcam_start(struct seq_file *seq, loff_t *pos)
1897 {
1898 	return *pos ? mps_tcam_get_idx(seq, *pos) : SEQ_START_TOKEN;
1899 }
1900 
1901 static void *mps_tcam_next(struct seq_file *seq, void *v, loff_t *pos)
1902 {
1903 	++*pos;
1904 	return mps_tcam_get_idx(seq, *pos);
1905 }
1906 
1907 static void mps_tcam_stop(struct seq_file *seq, void *v)
1908 {
1909 }
1910 
1911 static const struct seq_operations mps_tcam_seq_ops = {
1912 	.start = mps_tcam_start,
1913 	.next  = mps_tcam_next,
1914 	.stop  = mps_tcam_stop,
1915 	.show  = mps_tcam_show
1916 };
1917 
1918 static int mps_tcam_open(struct inode *inode, struct file *file)
1919 {
1920 	int res = seq_open(file, &mps_tcam_seq_ops);
1921 
1922 	if (!res) {
1923 		struct seq_file *seq = file->private_data;
1924 
1925 		seq->private = inode->i_private;
1926 	}
1927 	return res;
1928 }
1929 
1930 static const struct file_operations mps_tcam_debugfs_fops = {
1931 	.owner   = THIS_MODULE,
1932 	.open    = mps_tcam_open,
1933 	.read    = seq_read,
1934 	.llseek  = seq_lseek,
1935 	.release = seq_release,
1936 };
1937 
1938 /* Display various sensor information.
1939  */
1940 static int sensors_show(struct seq_file *seq, void *v)
1941 {
1942 	struct adapter *adap = seq->private;
1943 	u32 param[7], val[7];
1944 	int ret;
1945 
1946 	/* Note that if the sensors haven't been initialized and turned on
1947 	 * we'll get values of 0, so treat those as "<unknown>" ...
1948 	 */
1949 	param[0] = (FW_PARAMS_MNEM_V(FW_PARAMS_MNEM_DEV) |
1950 		    FW_PARAMS_PARAM_X_V(FW_PARAMS_PARAM_DEV_DIAG) |
1951 		    FW_PARAMS_PARAM_Y_V(FW_PARAM_DEV_DIAG_TMP));
1952 	param[1] = (FW_PARAMS_MNEM_V(FW_PARAMS_MNEM_DEV) |
1953 		    FW_PARAMS_PARAM_X_V(FW_PARAMS_PARAM_DEV_DIAG) |
1954 		    FW_PARAMS_PARAM_Y_V(FW_PARAM_DEV_DIAG_VDD));
1955 	ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 2,
1956 			      param, val);
1957 
1958 	if (ret < 0 || val[0] == 0)
1959 		seq_puts(seq, "Temperature: <unknown>\n");
1960 	else
1961 		seq_printf(seq, "Temperature: %dC\n", val[0]);
1962 
1963 	if (ret < 0 || val[1] == 0)
1964 		seq_puts(seq, "Core VDD:    <unknown>\n");
1965 	else
1966 		seq_printf(seq, "Core VDD:    %dmV\n", val[1]);
1967 
1968 	return 0;
1969 }
1970 DEFINE_SHOW_ATTRIBUTE(sensors);
1971 
1972 #if IS_ENABLED(CONFIG_IPV6)
1973 DEFINE_SHOW_ATTRIBUTE(clip_tbl);
1974 #endif
1975 
1976 /*RSS Table.
1977  */
1978 
1979 static int rss_show(struct seq_file *seq, void *v, int idx)
1980 {
1981 	u16 *entry = v;
1982 
1983 	seq_printf(seq, "%4d:  %4u  %4u  %4u  %4u  %4u  %4u  %4u  %4u\n",
1984 		   idx * 8, entry[0], entry[1], entry[2], entry[3], entry[4],
1985 		   entry[5], entry[6], entry[7]);
1986 	return 0;
1987 }
1988 
1989 static int rss_open(struct inode *inode, struct file *file)
1990 {
1991 	struct adapter *adap = inode->i_private;
1992 	int ret, nentries;
1993 	struct seq_tab *p;
1994 
1995 	nentries = t4_chip_rss_size(adap);
1996 	p = seq_open_tab(file, nentries / 8, 8 * sizeof(u16), 0, rss_show);
1997 	if (!p)
1998 		return -ENOMEM;
1999 
2000 	ret = t4_read_rss(adap, (u16 *)p->data);
2001 	if (ret)
2002 		seq_release_private(inode, file);
2003 
2004 	return ret;
2005 }
2006 
2007 static const struct file_operations rss_debugfs_fops = {
2008 	.owner   = THIS_MODULE,
2009 	.open    = rss_open,
2010 	.read    = seq_read,
2011 	.llseek  = seq_lseek,
2012 	.release = seq_release_private
2013 };
2014 
2015 /* RSS Configuration.
2016  */
2017 
2018 /* Small utility function to return the strings "yes" or "no" if the supplied
2019  * argument is non-zero.
2020  */
2021 static const char *yesno(int x)
2022 {
2023 	static const char *yes = "yes";
2024 	static const char *no = "no";
2025 
2026 	return x ? yes : no;
2027 }
2028 
2029 static int rss_config_show(struct seq_file *seq, void *v)
2030 {
2031 	struct adapter *adapter = seq->private;
2032 	static const char * const keymode[] = {
2033 		"global",
2034 		"global and per-VF scramble",
2035 		"per-PF and per-VF scramble",
2036 		"per-VF and per-VF scramble",
2037 	};
2038 	u32 rssconf;
2039 
2040 	rssconf = t4_read_reg(adapter, TP_RSS_CONFIG_A);
2041 	seq_printf(seq, "TP_RSS_CONFIG: %#x\n", rssconf);
2042 	seq_printf(seq, "  Tnl4TupEnIpv6: %3s\n", yesno(rssconf &
2043 							TNL4TUPENIPV6_F));
2044 	seq_printf(seq, "  Tnl2TupEnIpv6: %3s\n", yesno(rssconf &
2045 							TNL2TUPENIPV6_F));
2046 	seq_printf(seq, "  Tnl4TupEnIpv4: %3s\n", yesno(rssconf &
2047 							TNL4TUPENIPV4_F));
2048 	seq_printf(seq, "  Tnl2TupEnIpv4: %3s\n", yesno(rssconf &
2049 							TNL2TUPENIPV4_F));
2050 	seq_printf(seq, "  TnlTcpSel:     %3s\n", yesno(rssconf & TNLTCPSEL_F));
2051 	seq_printf(seq, "  TnlIp6Sel:     %3s\n", yesno(rssconf & TNLIP6SEL_F));
2052 	seq_printf(seq, "  TnlVrtSel:     %3s\n", yesno(rssconf & TNLVRTSEL_F));
2053 	seq_printf(seq, "  TnlMapEn:      %3s\n", yesno(rssconf & TNLMAPEN_F));
2054 	seq_printf(seq, "  OfdHashSave:   %3s\n", yesno(rssconf &
2055 							OFDHASHSAVE_F));
2056 	seq_printf(seq, "  OfdVrtSel:     %3s\n", yesno(rssconf & OFDVRTSEL_F));
2057 	seq_printf(seq, "  OfdMapEn:      %3s\n", yesno(rssconf & OFDMAPEN_F));
2058 	seq_printf(seq, "  OfdLkpEn:      %3s\n", yesno(rssconf & OFDLKPEN_F));
2059 	seq_printf(seq, "  Syn4TupEnIpv6: %3s\n", yesno(rssconf &
2060 							SYN4TUPENIPV6_F));
2061 	seq_printf(seq, "  Syn2TupEnIpv6: %3s\n", yesno(rssconf &
2062 							SYN2TUPENIPV6_F));
2063 	seq_printf(seq, "  Syn4TupEnIpv4: %3s\n", yesno(rssconf &
2064 							SYN4TUPENIPV4_F));
2065 	seq_printf(seq, "  Syn2TupEnIpv4: %3s\n", yesno(rssconf &
2066 							SYN2TUPENIPV4_F));
2067 	seq_printf(seq, "  Syn4TupEnIpv6: %3s\n", yesno(rssconf &
2068 							SYN4TUPENIPV6_F));
2069 	seq_printf(seq, "  SynIp6Sel:     %3s\n", yesno(rssconf & SYNIP6SEL_F));
2070 	seq_printf(seq, "  SynVrt6Sel:    %3s\n", yesno(rssconf & SYNVRTSEL_F));
2071 	seq_printf(seq, "  SynMapEn:      %3s\n", yesno(rssconf & SYNMAPEN_F));
2072 	seq_printf(seq, "  SynLkpEn:      %3s\n", yesno(rssconf & SYNLKPEN_F));
2073 	seq_printf(seq, "  ChnEn:         %3s\n", yesno(rssconf &
2074 							CHANNELENABLE_F));
2075 	seq_printf(seq, "  PrtEn:         %3s\n", yesno(rssconf &
2076 							PORTENABLE_F));
2077 	seq_printf(seq, "  TnlAllLkp:     %3s\n", yesno(rssconf &
2078 							TNLALLLOOKUP_F));
2079 	seq_printf(seq, "  VrtEn:         %3s\n", yesno(rssconf &
2080 							VIRTENABLE_F));
2081 	seq_printf(seq, "  CngEn:         %3s\n", yesno(rssconf &
2082 							CONGESTIONENABLE_F));
2083 	seq_printf(seq, "  HashToeplitz:  %3s\n", yesno(rssconf &
2084 							HASHTOEPLITZ_F));
2085 	seq_printf(seq, "  Udp4En:        %3s\n", yesno(rssconf & UDPENABLE_F));
2086 	seq_printf(seq, "  Disable:       %3s\n", yesno(rssconf & DISABLE_F));
2087 
2088 	seq_puts(seq, "\n");
2089 
2090 	rssconf = t4_read_reg(adapter, TP_RSS_CONFIG_TNL_A);
2091 	seq_printf(seq, "TP_RSS_CONFIG_TNL: %#x\n", rssconf);
2092 	seq_printf(seq, "  MaskSize:      %3d\n", MASKSIZE_G(rssconf));
2093 	seq_printf(seq, "  MaskFilter:    %3d\n", MASKFILTER_G(rssconf));
2094 	if (CHELSIO_CHIP_VERSION(adapter->params.chip) > CHELSIO_T5) {
2095 		seq_printf(seq, "  HashAll:     %3s\n",
2096 			   yesno(rssconf & HASHALL_F));
2097 		seq_printf(seq, "  HashEth:     %3s\n",
2098 			   yesno(rssconf & HASHETH_F));
2099 	}
2100 	seq_printf(seq, "  UseWireCh:     %3s\n", yesno(rssconf & USEWIRECH_F));
2101 
2102 	seq_puts(seq, "\n");
2103 
2104 	rssconf = t4_read_reg(adapter, TP_RSS_CONFIG_OFD_A);
2105 	seq_printf(seq, "TP_RSS_CONFIG_OFD: %#x\n", rssconf);
2106 	seq_printf(seq, "  MaskSize:      %3d\n", MASKSIZE_G(rssconf));
2107 	seq_printf(seq, "  RRCplMapEn:    %3s\n", yesno(rssconf &
2108 							RRCPLMAPEN_F));
2109 	seq_printf(seq, "  RRCplQueWidth: %3d\n", RRCPLQUEWIDTH_G(rssconf));
2110 
2111 	seq_puts(seq, "\n");
2112 
2113 	rssconf = t4_read_reg(adapter, TP_RSS_CONFIG_SYN_A);
2114 	seq_printf(seq, "TP_RSS_CONFIG_SYN: %#x\n", rssconf);
2115 	seq_printf(seq, "  MaskSize:      %3d\n", MASKSIZE_G(rssconf));
2116 	seq_printf(seq, "  UseWireCh:     %3s\n", yesno(rssconf & USEWIRECH_F));
2117 
2118 	seq_puts(seq, "\n");
2119 
2120 	rssconf = t4_read_reg(adapter, TP_RSS_CONFIG_VRT_A);
2121 	seq_printf(seq, "TP_RSS_CONFIG_VRT: %#x\n", rssconf);
2122 	if (CHELSIO_CHIP_VERSION(adapter->params.chip) > CHELSIO_T5) {
2123 		seq_printf(seq, "  KeyWrAddrX:     %3d\n",
2124 			   KEYWRADDRX_G(rssconf));
2125 		seq_printf(seq, "  KeyExtend:      %3s\n",
2126 			   yesno(rssconf & KEYEXTEND_F));
2127 	}
2128 	seq_printf(seq, "  VfRdRg:        %3s\n", yesno(rssconf & VFRDRG_F));
2129 	seq_printf(seq, "  VfRdEn:        %3s\n", yesno(rssconf & VFRDEN_F));
2130 	seq_printf(seq, "  VfPerrEn:      %3s\n", yesno(rssconf & VFPERREN_F));
2131 	seq_printf(seq, "  KeyPerrEn:     %3s\n", yesno(rssconf & KEYPERREN_F));
2132 	seq_printf(seq, "  DisVfVlan:     %3s\n", yesno(rssconf &
2133 							DISABLEVLAN_F));
2134 	seq_printf(seq, "  EnUpSwt:       %3s\n", yesno(rssconf & ENABLEUP0_F));
2135 	seq_printf(seq, "  HashDelay:     %3d\n", HASHDELAY_G(rssconf));
2136 	if (CHELSIO_CHIP_VERSION(adapter->params.chip) <= CHELSIO_T5)
2137 		seq_printf(seq, "  VfWrAddr:      %3d\n", VFWRADDR_G(rssconf));
2138 	else
2139 		seq_printf(seq, "  VfWrAddr:      %3d\n",
2140 			   T6_VFWRADDR_G(rssconf));
2141 	seq_printf(seq, "  KeyMode:       %s\n", keymode[KEYMODE_G(rssconf)]);
2142 	seq_printf(seq, "  VfWrEn:        %3s\n", yesno(rssconf & VFWREN_F));
2143 	seq_printf(seq, "  KeyWrEn:       %3s\n", yesno(rssconf & KEYWREN_F));
2144 	seq_printf(seq, "  KeyWrAddr:     %3d\n", KEYWRADDR_G(rssconf));
2145 
2146 	seq_puts(seq, "\n");
2147 
2148 	rssconf = t4_read_reg(adapter, TP_RSS_CONFIG_CNG_A);
2149 	seq_printf(seq, "TP_RSS_CONFIG_CNG: %#x\n", rssconf);
2150 	seq_printf(seq, "  ChnCount3:     %3s\n", yesno(rssconf & CHNCOUNT3_F));
2151 	seq_printf(seq, "  ChnCount2:     %3s\n", yesno(rssconf & CHNCOUNT2_F));
2152 	seq_printf(seq, "  ChnCount1:     %3s\n", yesno(rssconf & CHNCOUNT1_F));
2153 	seq_printf(seq, "  ChnCount0:     %3s\n", yesno(rssconf & CHNCOUNT0_F));
2154 	seq_printf(seq, "  ChnUndFlow3:   %3s\n", yesno(rssconf &
2155 							CHNUNDFLOW3_F));
2156 	seq_printf(seq, "  ChnUndFlow2:   %3s\n", yesno(rssconf &
2157 							CHNUNDFLOW2_F));
2158 	seq_printf(seq, "  ChnUndFlow1:   %3s\n", yesno(rssconf &
2159 							CHNUNDFLOW1_F));
2160 	seq_printf(seq, "  ChnUndFlow0:   %3s\n", yesno(rssconf &
2161 							CHNUNDFLOW0_F));
2162 	seq_printf(seq, "  RstChn3:       %3s\n", yesno(rssconf & RSTCHN3_F));
2163 	seq_printf(seq, "  RstChn2:       %3s\n", yesno(rssconf & RSTCHN2_F));
2164 	seq_printf(seq, "  RstChn1:       %3s\n", yesno(rssconf & RSTCHN1_F));
2165 	seq_printf(seq, "  RstChn0:       %3s\n", yesno(rssconf & RSTCHN0_F));
2166 	seq_printf(seq, "  UpdVld:        %3s\n", yesno(rssconf & UPDVLD_F));
2167 	seq_printf(seq, "  Xoff:          %3s\n", yesno(rssconf & XOFF_F));
2168 	seq_printf(seq, "  UpdChn3:       %3s\n", yesno(rssconf & UPDCHN3_F));
2169 	seq_printf(seq, "  UpdChn2:       %3s\n", yesno(rssconf & UPDCHN2_F));
2170 	seq_printf(seq, "  UpdChn1:       %3s\n", yesno(rssconf & UPDCHN1_F));
2171 	seq_printf(seq, "  UpdChn0:       %3s\n", yesno(rssconf & UPDCHN0_F));
2172 	seq_printf(seq, "  Queue:         %3d\n", QUEUE_G(rssconf));
2173 
2174 	return 0;
2175 }
2176 DEFINE_SHOW_ATTRIBUTE(rss_config);
2177 
2178 /* RSS Secret Key.
2179  */
2180 
2181 static int rss_key_show(struct seq_file *seq, void *v)
2182 {
2183 	u32 key[10];
2184 
2185 	t4_read_rss_key(seq->private, key, true);
2186 	seq_printf(seq, "%08x%08x%08x%08x%08x%08x%08x%08x%08x%08x\n",
2187 		   key[9], key[8], key[7], key[6], key[5], key[4], key[3],
2188 		   key[2], key[1], key[0]);
2189 	return 0;
2190 }
2191 
2192 static int rss_key_open(struct inode *inode, struct file *file)
2193 {
2194 	return single_open(file, rss_key_show, inode->i_private);
2195 }
2196 
2197 static ssize_t rss_key_write(struct file *file, const char __user *buf,
2198 			     size_t count, loff_t *pos)
2199 {
2200 	int i, j;
2201 	u32 key[10];
2202 	char s[100], *p;
2203 	struct adapter *adap = file_inode(file)->i_private;
2204 
2205 	if (count > sizeof(s) - 1)
2206 		return -EINVAL;
2207 	if (copy_from_user(s, buf, count))
2208 		return -EFAULT;
2209 	for (i = count; i > 0 && isspace(s[i - 1]); i--)
2210 		;
2211 	s[i] = '\0';
2212 
2213 	for (p = s, i = 9; i >= 0; i--) {
2214 		key[i] = 0;
2215 		for (j = 0; j < 8; j++, p++) {
2216 			if (!isxdigit(*p))
2217 				return -EINVAL;
2218 			key[i] = (key[i] << 4) | hex2val(*p);
2219 		}
2220 	}
2221 
2222 	t4_write_rss_key(adap, key, -1, true);
2223 	return count;
2224 }
2225 
2226 static const struct file_operations rss_key_debugfs_fops = {
2227 	.owner   = THIS_MODULE,
2228 	.open    = rss_key_open,
2229 	.read    = seq_read,
2230 	.llseek  = seq_lseek,
2231 	.release = single_release,
2232 	.write   = rss_key_write
2233 };
2234 
2235 /* PF RSS Configuration.
2236  */
2237 
2238 struct rss_pf_conf {
2239 	u32 rss_pf_map;
2240 	u32 rss_pf_mask;
2241 	u32 rss_pf_config;
2242 };
2243 
2244 static int rss_pf_config_show(struct seq_file *seq, void *v, int idx)
2245 {
2246 	struct rss_pf_conf *pfconf;
2247 
2248 	if (v == SEQ_START_TOKEN) {
2249 		/* use the 0th entry to dump the PF Map Index Size */
2250 		pfconf = seq->private + offsetof(struct seq_tab, data);
2251 		seq_printf(seq, "PF Map Index Size = %d\n\n",
2252 			   LKPIDXSIZE_G(pfconf->rss_pf_map));
2253 
2254 		seq_puts(seq, "     RSS              PF   VF    Hash Tuple Enable         Default\n");
2255 		seq_puts(seq, "     Enable       IPF Mask Mask  IPv6      IPv4      UDP   Queue\n");
2256 		seq_puts(seq, " PF  Map Chn Prt  Map Size Size  Four Two  Four Two  Four  Ch1  Ch0\n");
2257 	} else {
2258 		#define G_PFnLKPIDX(map, n) \
2259 			(((map) >> PF1LKPIDX_S*(n)) & PF0LKPIDX_M)
2260 		#define G_PFnMSKSIZE(mask, n) \
2261 			(((mask) >> PF1MSKSIZE_S*(n)) & PF1MSKSIZE_M)
2262 
2263 		pfconf = v;
2264 		seq_printf(seq, "%3d  %3s %3s %3s  %3d  %3d  %3d   %3s %3s   %3s %3s   %3s  %3d  %3d\n",
2265 			   idx,
2266 			   yesno(pfconf->rss_pf_config & MAPENABLE_F),
2267 			   yesno(pfconf->rss_pf_config & CHNENABLE_F),
2268 			   yesno(pfconf->rss_pf_config & PRTENABLE_F),
2269 			   G_PFnLKPIDX(pfconf->rss_pf_map, idx),
2270 			   G_PFnMSKSIZE(pfconf->rss_pf_mask, idx),
2271 			   IVFWIDTH_G(pfconf->rss_pf_config),
2272 			   yesno(pfconf->rss_pf_config & IP6FOURTUPEN_F),
2273 			   yesno(pfconf->rss_pf_config & IP6TWOTUPEN_F),
2274 			   yesno(pfconf->rss_pf_config & IP4FOURTUPEN_F),
2275 			   yesno(pfconf->rss_pf_config & IP4TWOTUPEN_F),
2276 			   yesno(pfconf->rss_pf_config & UDPFOURTUPEN_F),
2277 			   CH1DEFAULTQUEUE_G(pfconf->rss_pf_config),
2278 			   CH0DEFAULTQUEUE_G(pfconf->rss_pf_config));
2279 
2280 		#undef G_PFnLKPIDX
2281 		#undef G_PFnMSKSIZE
2282 	}
2283 	return 0;
2284 }
2285 
2286 static int rss_pf_config_open(struct inode *inode, struct file *file)
2287 {
2288 	struct adapter *adapter = inode->i_private;
2289 	struct seq_tab *p;
2290 	u32 rss_pf_map, rss_pf_mask;
2291 	struct rss_pf_conf *pfconf;
2292 	int pf;
2293 
2294 	p = seq_open_tab(file, 8, sizeof(*pfconf), 1, rss_pf_config_show);
2295 	if (!p)
2296 		return -ENOMEM;
2297 
2298 	pfconf = (struct rss_pf_conf *)p->data;
2299 	rss_pf_map = t4_read_rss_pf_map(adapter, true);
2300 	rss_pf_mask = t4_read_rss_pf_mask(adapter, true);
2301 	for (pf = 0; pf < 8; pf++) {
2302 		pfconf[pf].rss_pf_map = rss_pf_map;
2303 		pfconf[pf].rss_pf_mask = rss_pf_mask;
2304 		t4_read_rss_pf_config(adapter, pf, &pfconf[pf].rss_pf_config,
2305 				      true);
2306 	}
2307 	return 0;
2308 }
2309 
2310 static const struct file_operations rss_pf_config_debugfs_fops = {
2311 	.owner   = THIS_MODULE,
2312 	.open    = rss_pf_config_open,
2313 	.read    = seq_read,
2314 	.llseek  = seq_lseek,
2315 	.release = seq_release_private
2316 };
2317 
2318 /* VF RSS Configuration.
2319  */
2320 
2321 struct rss_vf_conf {
2322 	u32 rss_vf_vfl;
2323 	u32 rss_vf_vfh;
2324 };
2325 
2326 static int rss_vf_config_show(struct seq_file *seq, void *v, int idx)
2327 {
2328 	if (v == SEQ_START_TOKEN) {
2329 		seq_puts(seq, "     RSS                     Hash Tuple Enable\n");
2330 		seq_puts(seq, "     Enable   IVF  Dis  Enb  IPv6      IPv4      UDP    Def  Secret Key\n");
2331 		seq_puts(seq, " VF  Chn Prt  Map  VLAN  uP  Four Two  Four Two  Four   Que  Idx       Hash\n");
2332 	} else {
2333 		struct rss_vf_conf *vfconf = v;
2334 
2335 		seq_printf(seq, "%3d  %3s %3s  %3d   %3s %3s   %3s %3s   %3s  %3s   %3s  %4d  %3d %#10x\n",
2336 			   idx,
2337 			   yesno(vfconf->rss_vf_vfh & VFCHNEN_F),
2338 			   yesno(vfconf->rss_vf_vfh & VFPRTEN_F),
2339 			   VFLKPIDX_G(vfconf->rss_vf_vfh),
2340 			   yesno(vfconf->rss_vf_vfh & VFVLNEX_F),
2341 			   yesno(vfconf->rss_vf_vfh & VFUPEN_F),
2342 			   yesno(vfconf->rss_vf_vfh & VFIP4FOURTUPEN_F),
2343 			   yesno(vfconf->rss_vf_vfh & VFIP6TWOTUPEN_F),
2344 			   yesno(vfconf->rss_vf_vfh & VFIP4FOURTUPEN_F),
2345 			   yesno(vfconf->rss_vf_vfh & VFIP4TWOTUPEN_F),
2346 			   yesno(vfconf->rss_vf_vfh & ENABLEUDPHASH_F),
2347 			   DEFAULTQUEUE_G(vfconf->rss_vf_vfh),
2348 			   KEYINDEX_G(vfconf->rss_vf_vfh),
2349 			   vfconf->rss_vf_vfl);
2350 	}
2351 	return 0;
2352 }
2353 
2354 static int rss_vf_config_open(struct inode *inode, struct file *file)
2355 {
2356 	struct adapter *adapter = inode->i_private;
2357 	struct seq_tab *p;
2358 	struct rss_vf_conf *vfconf;
2359 	int vf, vfcount = adapter->params.arch.vfcount;
2360 
2361 	p = seq_open_tab(file, vfcount, sizeof(*vfconf), 1, rss_vf_config_show);
2362 	if (!p)
2363 		return -ENOMEM;
2364 
2365 	vfconf = (struct rss_vf_conf *)p->data;
2366 	for (vf = 0; vf < vfcount; vf++) {
2367 		t4_read_rss_vf_config(adapter, vf, &vfconf[vf].rss_vf_vfl,
2368 				      &vfconf[vf].rss_vf_vfh, true);
2369 	}
2370 	return 0;
2371 }
2372 
2373 static const struct file_operations rss_vf_config_debugfs_fops = {
2374 	.owner   = THIS_MODULE,
2375 	.open    = rss_vf_config_open,
2376 	.read    = seq_read,
2377 	.llseek  = seq_lseek,
2378 	.release = seq_release_private
2379 };
2380 
2381 #ifdef CONFIG_CHELSIO_T4_DCB
2382 
2383 /* Data Center Briging information for each port.
2384  */
2385 static int dcb_info_show(struct seq_file *seq, void *v)
2386 {
2387 	struct adapter *adap = seq->private;
2388 
2389 	if (v == SEQ_START_TOKEN) {
2390 		seq_puts(seq, "Data Center Bridging Information\n");
2391 	} else {
2392 		int port = (uintptr_t)v - 2;
2393 		struct net_device *dev = adap->port[port];
2394 		struct port_info *pi = netdev2pinfo(dev);
2395 		struct port_dcb_info *dcb = &pi->dcb;
2396 
2397 		seq_puts(seq, "\n");
2398 		seq_printf(seq, "Port: %d (DCB negotiated: %s)\n",
2399 			   port,
2400 			   cxgb4_dcb_enabled(dev) ? "yes" : "no");
2401 
2402 		if (cxgb4_dcb_enabled(dev))
2403 			seq_printf(seq, "[ DCBx Version %s ]\n",
2404 				   dcb_ver_array[dcb->dcb_version]);
2405 
2406 		if (dcb->msgs) {
2407 			int i;
2408 
2409 			seq_puts(seq, "\n  Index\t\t\t  :\t");
2410 			for (i = 0; i < 8; i++)
2411 				seq_printf(seq, " %3d", i);
2412 			seq_puts(seq, "\n\n");
2413 		}
2414 
2415 		if (dcb->msgs & CXGB4_DCB_FW_PGID) {
2416 			int prio, pgid;
2417 
2418 			seq_puts(seq, "  Priority Group IDs\t  :\t");
2419 			for (prio = 0; prio < 8; prio++) {
2420 				pgid = (dcb->pgid >> 4 * (7 - prio)) & 0xf;
2421 				seq_printf(seq, " %3d", pgid);
2422 			}
2423 			seq_puts(seq, "\n");
2424 		}
2425 
2426 		if (dcb->msgs & CXGB4_DCB_FW_PGRATE) {
2427 			int pg;
2428 
2429 			seq_puts(seq, "  Priority Group BW(%)\t  :\t");
2430 			for (pg = 0; pg < 8; pg++)
2431 				seq_printf(seq, " %3d", dcb->pgrate[pg]);
2432 			seq_puts(seq, "\n");
2433 
2434 			if (dcb->dcb_version == FW_PORT_DCB_VER_IEEE) {
2435 				seq_puts(seq, "  TSA Algorithm\t\t  :\t");
2436 				for (pg = 0; pg < 8; pg++)
2437 					seq_printf(seq, " %3d", dcb->tsa[pg]);
2438 				seq_puts(seq, "\n");
2439 			}
2440 
2441 			seq_printf(seq, "  Max PG Traffic Classes  [%3d  ]\n",
2442 				   dcb->pg_num_tcs_supported);
2443 
2444 			seq_puts(seq, "\n");
2445 		}
2446 
2447 		if (dcb->msgs & CXGB4_DCB_FW_PRIORATE) {
2448 			int prio;
2449 
2450 			seq_puts(seq, "  Priority Rate\t:\t");
2451 			for (prio = 0; prio < 8; prio++)
2452 				seq_printf(seq, " %3d", dcb->priorate[prio]);
2453 			seq_puts(seq, "\n");
2454 		}
2455 
2456 		if (dcb->msgs & CXGB4_DCB_FW_PFC) {
2457 			int prio;
2458 
2459 			seq_puts(seq, "  Priority Flow Control   :\t");
2460 			for (prio = 0; prio < 8; prio++) {
2461 				int pfcen = (dcb->pfcen >> 1 * (7 - prio))
2462 					    & 0x1;
2463 				seq_printf(seq, " %3d", pfcen);
2464 			}
2465 			seq_puts(seq, "\n");
2466 
2467 			seq_printf(seq, "  Max PFC Traffic Classes [%3d  ]\n",
2468 				   dcb->pfc_num_tcs_supported);
2469 
2470 			seq_puts(seq, "\n");
2471 		}
2472 
2473 		if (dcb->msgs & CXGB4_DCB_FW_APP_ID) {
2474 			int app, napps;
2475 
2476 			seq_puts(seq, "  Application Information:\n");
2477 			seq_puts(seq, "  App    Priority    Selection         Protocol\n");
2478 			seq_puts(seq, "  Index  Map         Field             ID\n");
2479 			for (app = 0, napps = 0;
2480 			     app < CXGB4_MAX_DCBX_APP_SUPPORTED; app++) {
2481 				struct app_priority *ap;
2482 				static const char * const sel_names[] = {
2483 					"Ethertype",
2484 					"Socket TCP",
2485 					"Socket UDP",
2486 					"Socket All",
2487 				};
2488 				const char *sel_name;
2489 
2490 				ap = &dcb->app_priority[app];
2491 				/* skip empty slots */
2492 				if (ap->protocolid == 0)
2493 					continue;
2494 				napps++;
2495 
2496 				if (ap->sel_field < ARRAY_SIZE(sel_names))
2497 					sel_name = sel_names[ap->sel_field];
2498 				else
2499 					sel_name = "UNKNOWN";
2500 
2501 				seq_printf(seq, "  %3d    %#04x        %-10s (%d)    %#06x (%d)\n",
2502 					   app,
2503 					   ap->user_prio_map,
2504 					   sel_name, ap->sel_field,
2505 					   ap->protocolid, ap->protocolid);
2506 			}
2507 			if (napps == 0)
2508 				seq_puts(seq, "    --- None ---\n");
2509 		}
2510 	}
2511 	return 0;
2512 }
2513 
2514 static inline void *dcb_info_get_idx(struct adapter *adap, loff_t pos)
2515 {
2516 	return (pos <= adap->params.nports
2517 		? (void *)((uintptr_t)pos + 1)
2518 		: NULL);
2519 }
2520 
2521 static void *dcb_info_start(struct seq_file *seq, loff_t *pos)
2522 {
2523 	struct adapter *adap = seq->private;
2524 
2525 	return (*pos
2526 		? dcb_info_get_idx(adap, *pos)
2527 		: SEQ_START_TOKEN);
2528 }
2529 
2530 static void dcb_info_stop(struct seq_file *seq, void *v)
2531 {
2532 }
2533 
2534 static void *dcb_info_next(struct seq_file *seq, void *v, loff_t *pos)
2535 {
2536 	struct adapter *adap = seq->private;
2537 
2538 	(*pos)++;
2539 	return dcb_info_get_idx(adap, *pos);
2540 }
2541 
2542 static const struct seq_operations dcb_info_seq_ops = {
2543 	.start = dcb_info_start,
2544 	.next  = dcb_info_next,
2545 	.stop  = dcb_info_stop,
2546 	.show  = dcb_info_show
2547 };
2548 
2549 static int dcb_info_open(struct inode *inode, struct file *file)
2550 {
2551 	int res = seq_open(file, &dcb_info_seq_ops);
2552 
2553 	if (!res) {
2554 		struct seq_file *seq = file->private_data;
2555 
2556 		seq->private = inode->i_private;
2557 	}
2558 	return res;
2559 }
2560 
2561 static const struct file_operations dcb_info_debugfs_fops = {
2562 	.owner   = THIS_MODULE,
2563 	.open    = dcb_info_open,
2564 	.read    = seq_read,
2565 	.llseek  = seq_lseek,
2566 	.release = seq_release,
2567 };
2568 #endif /* CONFIG_CHELSIO_T4_DCB */
2569 
2570 static int resources_show(struct seq_file *seq, void *v)
2571 {
2572 	struct adapter *adapter = seq->private;
2573 	struct pf_resources *pfres = &adapter->params.pfres;
2574 
2575 	#define S(desc, fmt, var) \
2576 		seq_printf(seq, "%-60s " fmt "\n", \
2577 			   desc " (" #var "):", pfres->var)
2578 
2579 	S("Virtual Interfaces", "%d", nvi);
2580 	S("Egress Queues", "%d", neq);
2581 	S("Ethernet Control", "%d", nethctrl);
2582 	S("Ingress Queues/w Free Lists/Interrupts", "%d", niqflint);
2583 	S("Ingress Queues", "%d", niq);
2584 	S("Traffic Class", "%d", tc);
2585 	S("Port Access Rights Mask", "%#x", pmask);
2586 	S("MAC Address Filters", "%d", nexactf);
2587 	S("Firmware Command Read Capabilities", "%#x", r_caps);
2588 	S("Firmware Command Write/Execute Capabilities", "%#x", wx_caps);
2589 
2590 	#undef S
2591 
2592 	return 0;
2593 }
2594 DEFINE_SHOW_ATTRIBUTE(resources);
2595 
2596 /**
2597  * ethqset2pinfo - return port_info of an Ethernet Queue Set
2598  * @adap: the adapter
2599  * @qset: Ethernet Queue Set
2600  */
2601 static inline struct port_info *ethqset2pinfo(struct adapter *adap, int qset)
2602 {
2603 	int pidx;
2604 
2605 	for_each_port(adap, pidx) {
2606 		struct port_info *pi = adap2pinfo(adap, pidx);
2607 
2608 		if (qset >= pi->first_qset &&
2609 		    qset < pi->first_qset + pi->nqsets)
2610 			return pi;
2611 	}
2612 
2613 	/* should never happen! */
2614 	BUG();
2615 	return NULL;
2616 }
2617 
2618 static int sge_qinfo_uld_txq_entries(const struct adapter *adap, int uld)
2619 {
2620 	const struct sge_uld_txq_info *utxq_info = adap->sge.uld_txq_info[uld];
2621 
2622 	if (!utxq_info)
2623 		return 0;
2624 
2625 	return DIV_ROUND_UP(utxq_info->ntxq, 4);
2626 }
2627 
2628 static int sge_qinfo_uld_rspq_entries(const struct adapter *adap, int uld,
2629 				      bool ciq)
2630 {
2631 	const struct sge_uld_rxq_info *urxq_info = adap->sge.uld_rxq_info[uld];
2632 
2633 	if (!urxq_info)
2634 		return 0;
2635 
2636 	return ciq ? DIV_ROUND_UP(urxq_info->nciq, 4) :
2637 		     DIV_ROUND_UP(urxq_info->nrxq, 4);
2638 }
2639 
2640 static int sge_qinfo_uld_rxq_entries(const struct adapter *adap, int uld)
2641 {
2642 	return sge_qinfo_uld_rspq_entries(adap, uld, false);
2643 }
2644 
2645 static int sge_qinfo_uld_ciq_entries(const struct adapter *adap, int uld)
2646 {
2647 	return sge_qinfo_uld_rspq_entries(adap, uld, true);
2648 }
2649 
2650 static int sge_qinfo_show(struct seq_file *seq, void *v)
2651 {
2652 	int eth_entries, ctrl_entries, eohw_entries = 0, eosw_entries = 0;
2653 	int uld_rxq_entries[CXGB4_ULD_MAX] = { 0 };
2654 	int uld_ciq_entries[CXGB4_ULD_MAX] = { 0 };
2655 	int uld_txq_entries[CXGB4_TX_MAX] = { 0 };
2656 	const struct sge_uld_txq_info *utxq_info;
2657 	const struct sge_uld_rxq_info *urxq_info;
2658 	struct cxgb4_tc_port_mqprio *port_mqprio;
2659 	struct adapter *adap = seq->private;
2660 	int i, j, n, r = (uintptr_t)v - 1;
2661 	struct sge *s = &adap->sge;
2662 
2663 	eth_entries = DIV_ROUND_UP(adap->sge.ethqsets, 4);
2664 	ctrl_entries = DIV_ROUND_UP(MAX_CTRL_QUEUES, 4);
2665 
2666 	if (r)
2667 		seq_putc(seq, '\n');
2668 
2669 #define S3(fmt_spec, s, v) \
2670 do { \
2671 	seq_printf(seq, "%-12s", s); \
2672 	for (i = 0; i < n; ++i) \
2673 		seq_printf(seq, " %16" fmt_spec, v); \
2674 		seq_putc(seq, '\n'); \
2675 } while (0)
2676 #define S(s, v) S3("s", s, v)
2677 #define T3(fmt_spec, s, v) S3(fmt_spec, s, tx[i].v)
2678 #define T(s, v) S3("u", s, tx[i].v)
2679 #define TL(s, v) T3("lu", s, v)
2680 #define R3(fmt_spec, s, v) S3(fmt_spec, s, rx[i].v)
2681 #define R(s, v) S3("u", s, rx[i].v)
2682 #define RL(s, v) R3("lu", s, v)
2683 
2684 	if (r < eth_entries) {
2685 		int base_qset = r * 4;
2686 		const struct sge_eth_rxq *rx = &s->ethrxq[base_qset];
2687 		const struct sge_eth_txq *tx = &s->ethtxq[base_qset];
2688 
2689 		n = min(4, s->ethqsets - 4 * r);
2690 
2691 		S("QType:", "Ethernet");
2692 		S("Interface:",
2693 		  rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A");
2694 		T("TxQ ID:", q.cntxt_id);
2695 		T("TxQ size:", q.size);
2696 		T("TxQ inuse:", q.in_use);
2697 		T("TxQ CIDX:", q.cidx);
2698 		T("TxQ PIDX:", q.pidx);
2699 #ifdef CONFIG_CHELSIO_T4_DCB
2700 		T("DCB Prio:", dcb_prio);
2701 		S3("u", "DCB PGID:",
2702 		   (ethqset2pinfo(adap, base_qset + i)->dcb.pgid >>
2703 		    4*(7-tx[i].dcb_prio)) & 0xf);
2704 		S3("u", "DCB PFC:",
2705 		   (ethqset2pinfo(adap, base_qset + i)->dcb.pfcen >>
2706 		    1*(7-tx[i].dcb_prio)) & 0x1);
2707 #endif
2708 		R("RspQ ID:", rspq.abs_id);
2709 		R("RspQ size:", rspq.size);
2710 		R("RspQE size:", rspq.iqe_len);
2711 		R("RspQ CIDX:", rspq.cidx);
2712 		R("RspQ Gen:", rspq.gen);
2713 		S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
2714 		S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]);
2715 		R("FL ID:", fl.cntxt_id);
2716 		R("FL size:", fl.size - 8);
2717 		R("FL pend:", fl.pend_cred);
2718 		R("FL avail:", fl.avail);
2719 		R("FL PIDX:", fl.pidx);
2720 		R("FL CIDX:", fl.cidx);
2721 		RL("RxPackets:", stats.pkts);
2722 		RL("RxCSO:", stats.rx_cso);
2723 		RL("VLANxtract:", stats.vlan_ex);
2724 		RL("LROmerged:", stats.lro_merged);
2725 		RL("LROpackets:", stats.lro_pkts);
2726 		RL("RxDrops:", stats.rx_drops);
2727 		RL("RxBadPkts:", stats.bad_rx_pkts);
2728 		TL("TSO:", tso);
2729 		TL("USO:", uso);
2730 		TL("TxCSO:", tx_cso);
2731 		TL("VLANins:", vlan_ins);
2732 		TL("TxQFull:", q.stops);
2733 		TL("TxQRestarts:", q.restarts);
2734 		TL("TxMapErr:", mapping_err);
2735 		RL("FLAllocErr:", fl.alloc_failed);
2736 		RL("FLLrgAlcErr:", fl.large_alloc_failed);
2737 		RL("FLMapErr:", fl.mapping_err);
2738 		RL("FLLow:", fl.low);
2739 		RL("FLStarving:", fl.starving);
2740 
2741 		goto out;
2742 	}
2743 
2744 	r -= eth_entries;
2745 	if (!adap->tc_mqprio)
2746 		goto skip_mqprio;
2747 
2748 	mutex_lock(&adap->tc_mqprio->mqprio_mutex);
2749 	if (!refcount_read(&adap->tc_mqprio->refcnt)) {
2750 		mutex_unlock(&adap->tc_mqprio->mqprio_mutex);
2751 		goto skip_mqprio;
2752 	}
2753 
2754 	eohw_entries = DIV_ROUND_UP(adap->sge.eoqsets, 4);
2755 	if (r < eohw_entries) {
2756 		int base_qset = r * 4;
2757 		const struct sge_ofld_rxq *rx = &s->eohw_rxq[base_qset];
2758 		const struct sge_eohw_txq *tx = &s->eohw_txq[base_qset];
2759 
2760 		n = min(4, s->eoqsets - 4 * r);
2761 
2762 		S("QType:", "ETHOFLD");
2763 		S("Interface:",
2764 		  rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A");
2765 		T("TxQ ID:", q.cntxt_id);
2766 		T("TxQ size:", q.size);
2767 		T("TxQ inuse:", q.in_use);
2768 		T("TxQ CIDX:", q.cidx);
2769 		T("TxQ PIDX:", q.pidx);
2770 		R("RspQ ID:", rspq.abs_id);
2771 		R("RspQ size:", rspq.size);
2772 		R("RspQE size:", rspq.iqe_len);
2773 		R("RspQ CIDX:", rspq.cidx);
2774 		R("RspQ Gen:", rspq.gen);
2775 		S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
2776 		S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]);
2777 		R("FL ID:", fl.cntxt_id);
2778 		S3("u", "FL size:", rx->fl.size ? rx->fl.size - 8 : 0);
2779 		R("FL pend:", fl.pend_cred);
2780 		R("FL avail:", fl.avail);
2781 		R("FL PIDX:", fl.pidx);
2782 		R("FL CIDX:", fl.cidx);
2783 		RL("RxPackets:", stats.pkts);
2784 		RL("RxImm:", stats.imm);
2785 		RL("RxAN", stats.an);
2786 		RL("RxNoMem", stats.nomem);
2787 		TL("TSO:", tso);
2788 		TL("USO:", uso);
2789 		TL("TxCSO:", tx_cso);
2790 		TL("VLANins:", vlan_ins);
2791 		TL("TxQFull:", q.stops);
2792 		TL("TxQRestarts:", q.restarts);
2793 		TL("TxMapErr:", mapping_err);
2794 		RL("FLAllocErr:", fl.alloc_failed);
2795 		RL("FLLrgAlcErr:", fl.large_alloc_failed);
2796 		RL("FLMapErr:", fl.mapping_err);
2797 		RL("FLLow:", fl.low);
2798 		RL("FLStarving:", fl.starving);
2799 
2800 		mutex_unlock(&adap->tc_mqprio->mqprio_mutex);
2801 		goto out;
2802 	}
2803 
2804 	r -= eohw_entries;
2805 	for (j = 0; j < adap->params.nports; j++) {
2806 		int entries;
2807 		u8 tc;
2808 
2809 		port_mqprio = &adap->tc_mqprio->port_mqprio[j];
2810 		entries = 0;
2811 		for (tc = 0; tc < port_mqprio->mqprio.qopt.num_tc; tc++)
2812 			entries += port_mqprio->mqprio.qopt.count[tc];
2813 
2814 		if (!entries)
2815 			continue;
2816 
2817 		eosw_entries = DIV_ROUND_UP(entries, 4);
2818 		if (r < eosw_entries) {
2819 			const struct sge_eosw_txq *tx;
2820 
2821 			n = min(4, entries - 4 * r);
2822 			tx = &port_mqprio->eosw_txq[4 * r];
2823 
2824 			S("QType:", "EOSW-TXQ");
2825 			S("Interface:",
2826 			  adap->port[j] ? adap->port[j]->name : "N/A");
2827 			T("EOTID:", hwtid);
2828 			T("HWQID:", hwqid);
2829 			T("State:", state);
2830 			T("Size:", ndesc);
2831 			T("In-Use:", inuse);
2832 			T("Credits:", cred);
2833 			T("Compl:", ncompl);
2834 			T("Last-Compl:", last_compl);
2835 			T("PIDX:", pidx);
2836 			T("Last-PIDX:", last_pidx);
2837 			T("CIDX:", cidx);
2838 			T("Last-CIDX:", last_cidx);
2839 			T("FLOWC-IDX:", flowc_idx);
2840 
2841 			mutex_unlock(&adap->tc_mqprio->mqprio_mutex);
2842 			goto out;
2843 		}
2844 
2845 		r -= eosw_entries;
2846 	}
2847 	mutex_unlock(&adap->tc_mqprio->mqprio_mutex);
2848 
2849 skip_mqprio:
2850 	if (!is_uld(adap))
2851 		goto skip_uld;
2852 
2853 	mutex_lock(&uld_mutex);
2854 	if (s->uld_txq_info)
2855 		for (i = 0; i < ARRAY_SIZE(uld_txq_entries); i++)
2856 			uld_txq_entries[i] = sge_qinfo_uld_txq_entries(adap, i);
2857 
2858 	if (s->uld_rxq_info) {
2859 		for (i = 0; i < ARRAY_SIZE(uld_rxq_entries); i++) {
2860 			uld_rxq_entries[i] = sge_qinfo_uld_rxq_entries(adap, i);
2861 			uld_ciq_entries[i] = sge_qinfo_uld_ciq_entries(adap, i);
2862 		}
2863 	}
2864 
2865 	if (r < uld_txq_entries[CXGB4_TX_OFLD]) {
2866 		const struct sge_uld_txq *tx;
2867 
2868 		utxq_info = s->uld_txq_info[CXGB4_TX_OFLD];
2869 		tx = &utxq_info->uldtxq[r * 4];
2870 		n = min(4, utxq_info->ntxq - 4 * r);
2871 
2872 		S("QType:", "OFLD-TXQ");
2873 		T("TxQ ID:", q.cntxt_id);
2874 		T("TxQ size:", q.size);
2875 		T("TxQ inuse:", q.in_use);
2876 		T("TxQ CIDX:", q.cidx);
2877 		T("TxQ PIDX:", q.pidx);
2878 
2879 		goto unlock;
2880 	}
2881 
2882 	r -= uld_txq_entries[CXGB4_TX_OFLD];
2883 	if (r < uld_rxq_entries[CXGB4_ULD_RDMA]) {
2884 		const struct sge_ofld_rxq *rx;
2885 
2886 		urxq_info = s->uld_rxq_info[CXGB4_ULD_RDMA];
2887 		rx = &urxq_info->uldrxq[r * 4];
2888 		n = min(4, urxq_info->nrxq - 4 * r);
2889 
2890 		S("QType:", "RDMA-CPL");
2891 		S("Interface:",
2892 		  rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A");
2893 		R("RspQ ID:", rspq.abs_id);
2894 		R("RspQ size:", rspq.size);
2895 		R("RspQE size:", rspq.iqe_len);
2896 		R("RspQ CIDX:", rspq.cidx);
2897 		R("RspQ Gen:", rspq.gen);
2898 		S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
2899 		S3("u", "Intr pktcnt:",	s->counter_val[rx[i].rspq.pktcnt_idx]);
2900 		R("FL ID:", fl.cntxt_id);
2901 		R("FL size:", fl.size - 8);
2902 		R("FL pend:", fl.pend_cred);
2903 		R("FL avail:", fl.avail);
2904 		R("FL PIDX:", fl.pidx);
2905 		R("FL CIDX:", fl.cidx);
2906 
2907 		goto unlock;
2908 	}
2909 
2910 	r -= uld_rxq_entries[CXGB4_ULD_RDMA];
2911 	if (r < uld_ciq_entries[CXGB4_ULD_RDMA]) {
2912 		const struct sge_ofld_rxq *rx;
2913 		int ciq_idx = 0;
2914 
2915 		urxq_info = s->uld_rxq_info[CXGB4_ULD_RDMA];
2916 		ciq_idx = urxq_info->nrxq + (r * 4);
2917 		rx = &urxq_info->uldrxq[ciq_idx];
2918 		n = min(4, urxq_info->nciq - 4 * r);
2919 
2920 		S("QType:", "RDMA-CIQ");
2921 		S("Interface:",
2922 		  rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A");
2923 		R("RspQ ID:", rspq.abs_id);
2924 		R("RspQ size:", rspq.size);
2925 		R("RspQE size:", rspq.iqe_len);
2926 		R("RspQ CIDX:", rspq.cidx);
2927 		R("RspQ Gen:", rspq.gen);
2928 		S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
2929 		S3("u", "Intr pktcnt:",	s->counter_val[rx[i].rspq.pktcnt_idx]);
2930 
2931 		goto unlock;
2932 	}
2933 
2934 	r -= uld_ciq_entries[CXGB4_ULD_RDMA];
2935 	if (r < uld_rxq_entries[CXGB4_ULD_ISCSI]) {
2936 		const struct sge_ofld_rxq *rx;
2937 
2938 		urxq_info = s->uld_rxq_info[CXGB4_ULD_ISCSI];
2939 		rx = &urxq_info->uldrxq[r * 4];
2940 		n = min(4, urxq_info->nrxq - 4 * r);
2941 
2942 		S("QType:", "iSCSI");
2943 		R("RspQ ID:", rspq.abs_id);
2944 		R("RspQ size:", rspq.size);
2945 		R("RspQE size:", rspq.iqe_len);
2946 		R("RspQ CIDX:", rspq.cidx);
2947 		R("RspQ Gen:", rspq.gen);
2948 		S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
2949 		S3("u", "Intr pktcnt:",	s->counter_val[rx[i].rspq.pktcnt_idx]);
2950 		R("FL ID:", fl.cntxt_id);
2951 		R("FL size:", fl.size - 8);
2952 		R("FL pend:", fl.pend_cred);
2953 		R("FL avail:", fl.avail);
2954 		R("FL PIDX:", fl.pidx);
2955 		R("FL CIDX:", fl.cidx);
2956 
2957 		goto unlock;
2958 	}
2959 
2960 	r -= uld_rxq_entries[CXGB4_ULD_ISCSI];
2961 	if (r < uld_rxq_entries[CXGB4_ULD_ISCSIT]) {
2962 		const struct sge_ofld_rxq *rx;
2963 
2964 		urxq_info = s->uld_rxq_info[CXGB4_ULD_ISCSIT];
2965 		rx = &urxq_info->uldrxq[r * 4];
2966 		n = min(4, urxq_info->nrxq - 4 * r);
2967 
2968 		S("QType:", "iSCSIT");
2969 		R("RspQ ID:", rspq.abs_id);
2970 		R("RspQ size:", rspq.size);
2971 		R("RspQE size:", rspq.iqe_len);
2972 		R("RspQ CIDX:", rspq.cidx);
2973 		R("RspQ Gen:", rspq.gen);
2974 		S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
2975 		S3("u", "Intr pktcnt:",	s->counter_val[rx[i].rspq.pktcnt_idx]);
2976 		R("FL ID:", fl.cntxt_id);
2977 		R("FL size:", fl.size - 8);
2978 		R("FL pend:", fl.pend_cred);
2979 		R("FL avail:", fl.avail);
2980 		R("FL PIDX:", fl.pidx);
2981 		R("FL CIDX:", fl.cidx);
2982 
2983 		goto unlock;
2984 	}
2985 
2986 	r -= uld_rxq_entries[CXGB4_ULD_ISCSIT];
2987 	if (r < uld_rxq_entries[CXGB4_ULD_TLS]) {
2988 		const struct sge_ofld_rxq *rx;
2989 
2990 		urxq_info = s->uld_rxq_info[CXGB4_ULD_TLS];
2991 		rx = &urxq_info->uldrxq[r * 4];
2992 		n = min(4, urxq_info->nrxq - 4 * r);
2993 
2994 		S("QType:", "TLS");
2995 		R("RspQ ID:", rspq.abs_id);
2996 		R("RspQ size:", rspq.size);
2997 		R("RspQE size:", rspq.iqe_len);
2998 		R("RspQ CIDX:", rspq.cidx);
2999 		R("RspQ Gen:", rspq.gen);
3000 		S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
3001 		S3("u", "Intr pktcnt:",	s->counter_val[rx[i].rspq.pktcnt_idx]);
3002 		R("FL ID:", fl.cntxt_id);
3003 		R("FL size:", fl.size - 8);
3004 		R("FL pend:", fl.pend_cred);
3005 		R("FL avail:", fl.avail);
3006 		R("FL PIDX:", fl.pidx);
3007 		R("FL CIDX:", fl.cidx);
3008 
3009 		goto unlock;
3010 	}
3011 
3012 	r -= uld_rxq_entries[CXGB4_ULD_TLS];
3013 	if (r < uld_txq_entries[CXGB4_TX_CRYPTO]) {
3014 		const struct sge_ofld_rxq *rx;
3015 		const struct sge_uld_txq *tx;
3016 
3017 		utxq_info = s->uld_txq_info[CXGB4_TX_CRYPTO];
3018 		urxq_info = s->uld_rxq_info[CXGB4_ULD_CRYPTO];
3019 		tx = &utxq_info->uldtxq[r * 4];
3020 		rx = &urxq_info->uldrxq[r * 4];
3021 		n = min(4, utxq_info->ntxq - 4 * r);
3022 
3023 		S("QType:", "Crypto");
3024 		T("TxQ ID:", q.cntxt_id);
3025 		T("TxQ size:", q.size);
3026 		T("TxQ inuse:", q.in_use);
3027 		T("TxQ CIDX:", q.cidx);
3028 		T("TxQ PIDX:", q.pidx);
3029 		R("RspQ ID:", rspq.abs_id);
3030 		R("RspQ size:", rspq.size);
3031 		R("RspQE size:", rspq.iqe_len);
3032 		R("RspQ CIDX:", rspq.cidx);
3033 		R("RspQ Gen:", rspq.gen);
3034 		S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
3035 		S3("u", "Intr pktcnt:",	s->counter_val[rx[i].rspq.pktcnt_idx]);
3036 		R("FL ID:", fl.cntxt_id);
3037 		R("FL size:", fl.size - 8);
3038 		R("FL pend:", fl.pend_cred);
3039 		R("FL avail:", fl.avail);
3040 		R("FL PIDX:", fl.pidx);
3041 		R("FL CIDX:", fl.cidx);
3042 
3043 		goto unlock;
3044 	}
3045 
3046 	r -= uld_txq_entries[CXGB4_TX_CRYPTO];
3047 	mutex_unlock(&uld_mutex);
3048 
3049 skip_uld:
3050 	if (r < ctrl_entries) {
3051 		const struct sge_ctrl_txq *tx = &s->ctrlq[r * 4];
3052 
3053 		n = min(4, adap->params.nports - 4 * r);
3054 
3055 		S("QType:", "Control");
3056 		T("TxQ ID:", q.cntxt_id);
3057 		T("TxQ size:", q.size);
3058 		T("TxQ inuse:", q.in_use);
3059 		T("TxQ CIDX:", q.cidx);
3060 		T("TxQ PIDX:", q.pidx);
3061 		TL("TxQFull:", q.stops);
3062 		TL("TxQRestarts:", q.restarts);
3063 
3064 		goto out;
3065 	}
3066 
3067 	r -= ctrl_entries;
3068 	if (r < 1) {
3069 		const struct sge_rspq *evtq = &s->fw_evtq;
3070 
3071 		seq_printf(seq, "%-12s %16s\n", "QType:", "FW event queue");
3072 		seq_printf(seq, "%-12s %16u\n", "RspQ ID:", evtq->abs_id);
3073 		seq_printf(seq, "%-12s %16u\n", "RspQ size:", evtq->size);
3074 		seq_printf(seq, "%-12s %16u\n", "RspQE size:", evtq->iqe_len);
3075 		seq_printf(seq, "%-12s %16u\n", "RspQ CIDX:", evtq->cidx);
3076 		seq_printf(seq, "%-12s %16u\n", "RspQ Gen:", evtq->gen);
3077 		seq_printf(seq, "%-12s %16u\n", "Intr delay:",
3078 			   qtimer_val(adap, evtq));
3079 		seq_printf(seq, "%-12s %16u\n", "Intr pktcnt:",
3080 			   s->counter_val[evtq->pktcnt_idx]);
3081 
3082 		goto out;
3083 	}
3084 
3085 #undef R
3086 #undef RL
3087 #undef T
3088 #undef TL
3089 #undef S
3090 #undef R3
3091 #undef T3
3092 #undef S3
3093 out:
3094 	return 0;
3095 
3096 unlock:
3097 	mutex_unlock(&uld_mutex);
3098 	return 0;
3099 }
3100 
3101 static int sge_queue_entries(const struct adapter *adap)
3102 {
3103 	int i, tot_uld_entries = 0, eohw_entries = 0, eosw_entries = 0;
3104 
3105 	if (adap->tc_mqprio) {
3106 		struct cxgb4_tc_port_mqprio *port_mqprio;
3107 		u8 tc;
3108 
3109 		mutex_lock(&adap->tc_mqprio->mqprio_mutex);
3110 		if (adap->sge.eohw_txq)
3111 			eohw_entries = DIV_ROUND_UP(adap->sge.eoqsets, 4);
3112 
3113 		for (i = 0; i < adap->params.nports; i++) {
3114 			u32 entries = 0;
3115 
3116 			port_mqprio = &adap->tc_mqprio->port_mqprio[i];
3117 			for (tc = 0; tc < port_mqprio->mqprio.qopt.num_tc; tc++)
3118 				entries += port_mqprio->mqprio.qopt.count[tc];
3119 
3120 			if (entries)
3121 				eosw_entries += DIV_ROUND_UP(entries, 4);
3122 		}
3123 		mutex_unlock(&adap->tc_mqprio->mqprio_mutex);
3124 	}
3125 
3126 	if (!is_uld(adap))
3127 		goto lld_only;
3128 
3129 	mutex_lock(&uld_mutex);
3130 	for (i = 0; i < CXGB4_TX_MAX; i++)
3131 		tot_uld_entries += sge_qinfo_uld_txq_entries(adap, i);
3132 
3133 	for (i = 0; i < CXGB4_ULD_MAX; i++) {
3134 		tot_uld_entries += sge_qinfo_uld_rxq_entries(adap, i);
3135 		tot_uld_entries += sge_qinfo_uld_ciq_entries(adap, i);
3136 	}
3137 	mutex_unlock(&uld_mutex);
3138 
3139 lld_only:
3140 	return DIV_ROUND_UP(adap->sge.ethqsets, 4) +
3141 	       eohw_entries + eosw_entries + tot_uld_entries +
3142 	       DIV_ROUND_UP(MAX_CTRL_QUEUES, 4) + 1;
3143 }
3144 
3145 static void *sge_queue_start(struct seq_file *seq, loff_t *pos)
3146 {
3147 	int entries = sge_queue_entries(seq->private);
3148 
3149 	return *pos < entries ? (void *)((uintptr_t)*pos + 1) : NULL;
3150 }
3151 
3152 static void sge_queue_stop(struct seq_file *seq, void *v)
3153 {
3154 }
3155 
3156 static void *sge_queue_next(struct seq_file *seq, void *v, loff_t *pos)
3157 {
3158 	int entries = sge_queue_entries(seq->private);
3159 
3160 	++*pos;
3161 	return *pos < entries ? (void *)((uintptr_t)*pos + 1) : NULL;
3162 }
3163 
3164 static const struct seq_operations sge_qinfo_seq_ops = {
3165 	.start = sge_queue_start,
3166 	.next  = sge_queue_next,
3167 	.stop  = sge_queue_stop,
3168 	.show  = sge_qinfo_show
3169 };
3170 
3171 static int sge_qinfo_open(struct inode *inode, struct file *file)
3172 {
3173 	int res = seq_open(file, &sge_qinfo_seq_ops);
3174 
3175 	if (!res) {
3176 		struct seq_file *seq = file->private_data;
3177 
3178 		seq->private = inode->i_private;
3179 	}
3180 	return res;
3181 }
3182 
3183 static const struct file_operations sge_qinfo_debugfs_fops = {
3184 	.owner   = THIS_MODULE,
3185 	.open    = sge_qinfo_open,
3186 	.read    = seq_read,
3187 	.llseek  = seq_lseek,
3188 	.release = seq_release,
3189 };
3190 
3191 int mem_open(struct inode *inode, struct file *file)
3192 {
3193 	unsigned int mem;
3194 	struct adapter *adap;
3195 
3196 	file->private_data = inode->i_private;
3197 
3198 	mem = (uintptr_t)file->private_data & 0x7;
3199 	adap = file->private_data - mem;
3200 
3201 	(void)t4_fwcache(adap, FW_PARAM_DEV_FWCACHE_FLUSH);
3202 
3203 	return 0;
3204 }
3205 
3206 static ssize_t mem_read(struct file *file, char __user *buf, size_t count,
3207 			loff_t *ppos)
3208 {
3209 	loff_t pos = *ppos;
3210 	loff_t avail = file_inode(file)->i_size;
3211 	unsigned int mem = (uintptr_t)file->private_data & 0x7;
3212 	struct adapter *adap = file->private_data - mem;
3213 	__be32 *data;
3214 	int ret;
3215 
3216 	if (pos < 0)
3217 		return -EINVAL;
3218 	if (pos >= avail)
3219 		return 0;
3220 	if (count > avail - pos)
3221 		count = avail - pos;
3222 
3223 	data = kvzalloc(count, GFP_KERNEL);
3224 	if (!data)
3225 		return -ENOMEM;
3226 
3227 	spin_lock(&adap->win0_lock);
3228 	ret = t4_memory_rw(adap, 0, mem, pos, count, data, T4_MEMORY_READ);
3229 	spin_unlock(&adap->win0_lock);
3230 	if (ret) {
3231 		kvfree(data);
3232 		return ret;
3233 	}
3234 	ret = copy_to_user(buf, data, count);
3235 
3236 	kvfree(data);
3237 	if (ret)
3238 		return -EFAULT;
3239 
3240 	*ppos = pos + count;
3241 	return count;
3242 }
3243 static const struct file_operations mem_debugfs_fops = {
3244 	.owner   = THIS_MODULE,
3245 	.open    = simple_open,
3246 	.read    = mem_read,
3247 	.llseek  = default_llseek,
3248 };
3249 
3250 static int tid_info_show(struct seq_file *seq, void *v)
3251 {
3252 	struct adapter *adap = seq->private;
3253 	const struct tid_info *t;
3254 	enum chip_type chip;
3255 
3256 	t = &adap->tids;
3257 	chip = CHELSIO_CHIP_VERSION(adap->params.chip);
3258 	if (t4_read_reg(adap, LE_DB_CONFIG_A) & HASHEN_F) {
3259 		unsigned int sb;
3260 		seq_printf(seq, "Connections in use: %u\n",
3261 			   atomic_read(&t->conns_in_use));
3262 
3263 		if (chip <= CHELSIO_T5)
3264 			sb = t4_read_reg(adap, LE_DB_SERVER_INDEX_A) / 4;
3265 		else
3266 			sb = t4_read_reg(adap, LE_DB_SRVR_START_INDEX_A);
3267 
3268 		if (sb) {
3269 			seq_printf(seq, "TID range: %u..%u/%u..%u", t->tid_base,
3270 				   sb - 1, adap->tids.hash_base,
3271 				   t->tid_base + t->ntids - 1);
3272 			seq_printf(seq, ", in use: %u/%u\n",
3273 				   atomic_read(&t->tids_in_use),
3274 				   atomic_read(&t->hash_tids_in_use));
3275 		} else if (adap->flags & CXGB4_FW_OFLD_CONN) {
3276 			seq_printf(seq, "TID range: %u..%u/%u..%u",
3277 				   t->aftid_base,
3278 				   t->aftid_end,
3279 				   adap->tids.hash_base,
3280 				   t->tid_base + t->ntids - 1);
3281 			seq_printf(seq, ", in use: %u/%u\n",
3282 				   atomic_read(&t->tids_in_use),
3283 				   atomic_read(&t->hash_tids_in_use));
3284 		} else {
3285 			seq_printf(seq, "TID range: %u..%u",
3286 				   adap->tids.hash_base,
3287 				   t->tid_base + t->ntids - 1);
3288 			seq_printf(seq, ", in use: %u\n",
3289 				   atomic_read(&t->hash_tids_in_use));
3290 		}
3291 	} else if (t->ntids) {
3292 		seq_printf(seq, "Connections in use: %u\n",
3293 			   atomic_read(&t->conns_in_use));
3294 
3295 		seq_printf(seq, "TID range: %u..%u", t->tid_base,
3296 			   t->tid_base + t->ntids - 1);
3297 		seq_printf(seq, ", in use: %u\n",
3298 			   atomic_read(&t->tids_in_use));
3299 	}
3300 
3301 	if (t->nstids)
3302 		seq_printf(seq, "STID range: %u..%u, in use-IPv4/IPv6: %u/%u\n",
3303 			   (!t->stid_base &&
3304 			   (chip <= CHELSIO_T5)) ?
3305 			   t->stid_base + 1 : t->stid_base,
3306 			   t->stid_base + t->nstids - 1,
3307 			   t->stids_in_use - t->v6_stids_in_use,
3308 			   t->v6_stids_in_use);
3309 
3310 	if (t->natids)
3311 		seq_printf(seq, "ATID range: 0..%u, in use: %u\n",
3312 			   t->natids - 1, t->atids_in_use);
3313 	seq_printf(seq, "FTID range: %u..%u\n", t->ftid_base,
3314 		   t->ftid_base + t->nftids - 1);
3315 	if (t->nsftids)
3316 		seq_printf(seq, "SFTID range: %u..%u in use: %u\n",
3317 			   t->sftid_base, t->sftid_base + t->nsftids - 2,
3318 			   t->sftids_in_use);
3319 	if (t->nhpftids)
3320 		seq_printf(seq, "HPFTID range: %u..%u\n", t->hpftid_base,
3321 			   t->hpftid_base + t->nhpftids - 1);
3322 	if (t->neotids)
3323 		seq_printf(seq, "EOTID range: %u..%u, in use: %u\n",
3324 			   t->eotid_base, t->eotid_base + t->neotids - 1,
3325 			   atomic_read(&t->eotids_in_use));
3326 	if (t->ntids)
3327 		seq_printf(seq, "HW TID usage: %u IP users, %u IPv6 users\n",
3328 			   t4_read_reg(adap, LE_DB_ACT_CNT_IPV4_A),
3329 			   t4_read_reg(adap, LE_DB_ACT_CNT_IPV6_A));
3330 	return 0;
3331 }
3332 DEFINE_SHOW_ATTRIBUTE(tid_info);
3333 
3334 static void add_debugfs_mem(struct adapter *adap, const char *name,
3335 			    unsigned int idx, unsigned int size_mb)
3336 {
3337 	debugfs_create_file_size(name, 0400, adap->debugfs_root,
3338 				 (void *)adap + idx, &mem_debugfs_fops,
3339 				 size_mb << 20);
3340 }
3341 
3342 static ssize_t blocked_fl_read(struct file *filp, char __user *ubuf,
3343 			       size_t count, loff_t *ppos)
3344 {
3345 	int len;
3346 	const struct adapter *adap = filp->private_data;
3347 	char *buf;
3348 	ssize_t size = (adap->sge.egr_sz + 3) / 4 +
3349 			adap->sge.egr_sz / 32 + 2; /* includes ,/\n/\0 */
3350 
3351 	buf = kzalloc(size, GFP_KERNEL);
3352 	if (!buf)
3353 		return -ENOMEM;
3354 
3355 	len = snprintf(buf, size - 1, "%*pb\n",
3356 		       adap->sge.egr_sz, adap->sge.blocked_fl);
3357 	len += sprintf(buf + len, "\n");
3358 	size = simple_read_from_buffer(ubuf, count, ppos, buf, len);
3359 	kfree(buf);
3360 	return size;
3361 }
3362 
3363 static ssize_t blocked_fl_write(struct file *filp, const char __user *ubuf,
3364 				size_t count, loff_t *ppos)
3365 {
3366 	int err;
3367 	unsigned long *t;
3368 	struct adapter *adap = filp->private_data;
3369 
3370 	t = kcalloc(BITS_TO_LONGS(adap->sge.egr_sz), sizeof(long), GFP_KERNEL);
3371 	if (!t)
3372 		return -ENOMEM;
3373 
3374 	err = bitmap_parse_user(ubuf, count, t, adap->sge.egr_sz);
3375 	if (err) {
3376 		kfree(t);
3377 		return err;
3378 	}
3379 
3380 	bitmap_copy(adap->sge.blocked_fl, t, adap->sge.egr_sz);
3381 	kfree(t);
3382 	return count;
3383 }
3384 
3385 static const struct file_operations blocked_fl_fops = {
3386 	.owner   = THIS_MODULE,
3387 	.open    = simple_open,
3388 	.read    = blocked_fl_read,
3389 	.write   = blocked_fl_write,
3390 	.llseek  = generic_file_llseek,
3391 };
3392 
3393 static void mem_region_show(struct seq_file *seq, const char *name,
3394 			    unsigned int from, unsigned int to)
3395 {
3396 	char buf[40];
3397 
3398 	string_get_size((u64)to - from + 1, 1, STRING_UNITS_2, buf,
3399 			sizeof(buf));
3400 	seq_printf(seq, "%-15s %#x-%#x [%s]\n", name, from, to, buf);
3401 }
3402 
3403 static int meminfo_show(struct seq_file *seq, void *v)
3404 {
3405 	static const char * const memory[] = { "EDC0:", "EDC1:", "MC:",
3406 					       "MC0:", "MC1:", "HMA:"};
3407 	struct adapter *adap = seq->private;
3408 	struct cudbg_meminfo meminfo;
3409 	int i, rc;
3410 
3411 	memset(&meminfo, 0, sizeof(struct cudbg_meminfo));
3412 	rc = cudbg_fill_meminfo(adap, &meminfo);
3413 	if (rc)
3414 		return -ENXIO;
3415 
3416 	for (i = 0; i < meminfo.avail_c; i++)
3417 		mem_region_show(seq, memory[meminfo.avail[i].idx],
3418 				meminfo.avail[i].base,
3419 				meminfo.avail[i].limit - 1);
3420 
3421 	seq_putc(seq, '\n');
3422 	for (i = 0; i < meminfo.mem_c; i++) {
3423 		if (meminfo.mem[i].idx >= ARRAY_SIZE(cudbg_region))
3424 			continue;                        /* skip holes */
3425 		if (!meminfo.mem[i].limit)
3426 			meminfo.mem[i].limit =
3427 				i < meminfo.mem_c - 1 ?
3428 				meminfo.mem[i + 1].base - 1 : ~0;
3429 		mem_region_show(seq, cudbg_region[meminfo.mem[i].idx],
3430 				meminfo.mem[i].base, meminfo.mem[i].limit);
3431 	}
3432 
3433 	seq_putc(seq, '\n');
3434 	mem_region_show(seq, "uP RAM:", meminfo.up_ram_lo, meminfo.up_ram_hi);
3435 	mem_region_show(seq, "uP Extmem2:", meminfo.up_extmem2_lo,
3436 			meminfo.up_extmem2_hi);
3437 
3438 	seq_printf(seq, "\n%u Rx pages (%u free) of size %uKiB for %u channels\n",
3439 		   meminfo.rx_pages_data[0], meminfo.free_rx_cnt,
3440 		   meminfo.rx_pages_data[1], meminfo.rx_pages_data[2]);
3441 
3442 	seq_printf(seq, "%u Tx pages (%u free) of size %u%ciB for %u channels\n",
3443 		   meminfo.tx_pages_data[0], meminfo.free_tx_cnt,
3444 		   meminfo.tx_pages_data[1], meminfo.tx_pages_data[2],
3445 		   meminfo.tx_pages_data[3]);
3446 
3447 	seq_printf(seq, "%u p-structs (%u free)\n\n",
3448 		   meminfo.p_structs, meminfo.p_structs_free_cnt);
3449 
3450 	for (i = 0; i < 4; i++)
3451 		/* For T6 these are MAC buffer groups */
3452 		seq_printf(seq, "Port %d using %u pages out of %u allocated\n",
3453 			   i, meminfo.port_used[i], meminfo.port_alloc[i]);
3454 
3455 	for (i = 0; i < adap->params.arch.nchan; i++)
3456 		/* For T6 these are MAC buffer groups */
3457 		seq_printf(seq,
3458 			   "Loopback %d using %u pages out of %u allocated\n",
3459 			   i, meminfo.loopback_used[i],
3460 			   meminfo.loopback_alloc[i]);
3461 
3462 	return 0;
3463 }
3464 DEFINE_SHOW_ATTRIBUTE(meminfo);
3465 
3466 static int chcr_stats_show(struct seq_file *seq, void *v)
3467 {
3468 	struct adapter *adap = seq->private;
3469 
3470 	seq_puts(seq, "Chelsio Crypto Accelerator Stats \n");
3471 	seq_printf(seq, "Cipher Ops: %10u \n",
3472 		   atomic_read(&adap->chcr_stats.cipher_rqst));
3473 	seq_printf(seq, "Digest Ops: %10u \n",
3474 		   atomic_read(&adap->chcr_stats.digest_rqst));
3475 	seq_printf(seq, "Aead Ops: %10u \n",
3476 		   atomic_read(&adap->chcr_stats.aead_rqst));
3477 	seq_printf(seq, "Completion: %10u \n",
3478 		   atomic_read(&adap->chcr_stats.complete));
3479 	seq_printf(seq, "Error: %10u \n",
3480 		   atomic_read(&adap->chcr_stats.error));
3481 	seq_printf(seq, "Fallback: %10u \n",
3482 		   atomic_read(&adap->chcr_stats.fallback));
3483 	seq_printf(seq, "IPSec PDU: %10u\n",
3484 		   atomic_read(&adap->chcr_stats.ipsec_cnt));
3485 	seq_printf(seq, "TLS PDU Tx: %10u\n",
3486 		   atomic_read(&adap->chcr_stats.tls_pdu_tx));
3487 	seq_printf(seq, "TLS PDU Rx: %10u\n",
3488 		   atomic_read(&adap->chcr_stats.tls_pdu_rx));
3489 	seq_printf(seq, "TLS Keys (DDR) Count: %10u\n",
3490 		   atomic_read(&adap->chcr_stats.tls_key));
3491 #ifdef CONFIG_CHELSIO_TLS_DEVICE
3492 	seq_puts(seq, "\nChelsio KTLS Crypto Accelerator Stats\n");
3493 	seq_printf(seq, "Tx TLS offload refcount:          %20u\n",
3494 		   refcount_read(&adap->chcr_ktls.ktls_refcount));
3495 	seq_printf(seq, "Tx HW offload contexts added:     %20llu\n",
3496 		   atomic64_read(&adap->chcr_stats.ktls_tx_ctx));
3497 	seq_printf(seq, "Tx connection created:            %20llu\n",
3498 		   atomic64_read(&adap->chcr_stats.ktls_tx_connection_open));
3499 	seq_printf(seq, "Tx connection failed:             %20llu\n",
3500 		   atomic64_read(&adap->chcr_stats.ktls_tx_connection_fail));
3501 	seq_printf(seq, "Tx connection closed:             %20llu\n",
3502 		   atomic64_read(&adap->chcr_stats.ktls_tx_connection_close));
3503 	seq_printf(seq, "Packets passed for encryption :   %20llu\n",
3504 		   atomic64_read(&adap->chcr_stats.ktls_tx_encrypted_packets));
3505 	seq_printf(seq, "Bytes passed for encryption :     %20llu\n",
3506 		   atomic64_read(&adap->chcr_stats.ktls_tx_encrypted_bytes));
3507 	seq_printf(seq, "Tx records send:                  %20llu\n",
3508 		   atomic64_read(&adap->chcr_stats.ktls_tx_send_records));
3509 	seq_printf(seq, "Tx partial start of records:      %20llu\n",
3510 		   atomic64_read(&adap->chcr_stats.ktls_tx_start_pkts));
3511 	seq_printf(seq, "Tx partial middle of records:     %20llu\n",
3512 		   atomic64_read(&adap->chcr_stats.ktls_tx_middle_pkts));
3513 	seq_printf(seq, "Tx partial end of record:         %20llu\n",
3514 		   atomic64_read(&adap->chcr_stats.ktls_tx_end_pkts));
3515 	seq_printf(seq, "Tx complete records:              %20llu\n",
3516 		   atomic64_read(&adap->chcr_stats.ktls_tx_complete_pkts));
3517 	seq_printf(seq, "TX trim pkts :                    %20llu\n",
3518 		   atomic64_read(&adap->chcr_stats.ktls_tx_trimmed_pkts));
3519 	seq_printf(seq, "Tx out of order packets:          %20llu\n",
3520 		   atomic64_read(&adap->chcr_stats.ktls_tx_ooo));
3521 	seq_printf(seq, "Tx drop pkts before HW offload:   %20llu\n",
3522 		   atomic64_read(&adap->chcr_stats.ktls_tx_skip_no_sync_data));
3523 	seq_printf(seq, "Tx drop not synced packets:       %20llu\n",
3524 		   atomic64_read(&adap->chcr_stats.ktls_tx_drop_no_sync_data));
3525 	seq_printf(seq, "Tx drop bypass req:               %20llu\n",
3526 		   atomic64_read(&adap->chcr_stats.ktls_tx_drop_bypass_req));
3527 #endif
3528 
3529 	return 0;
3530 }
3531 DEFINE_SHOW_ATTRIBUTE(chcr_stats);
3532 
3533 #define PRINT_ADAP_STATS(string, value) \
3534 	seq_printf(seq, "%-25s %-20llu\n", (string), \
3535 		   (unsigned long long)(value))
3536 
3537 #define PRINT_CH_STATS(string, value) \
3538 do { \
3539 	seq_printf(seq, "%-25s ", (string)); \
3540 	for (i = 0; i < adap->params.arch.nchan; i++) \
3541 		seq_printf(seq, "%-20llu ", \
3542 			   (unsigned long long)stats.value[i]); \
3543 	seq_printf(seq, "\n"); \
3544 } while (0)
3545 
3546 #define PRINT_CH_STATS2(string, value) \
3547 do { \
3548 	seq_printf(seq, "%-25s ", (string)); \
3549 	for (i = 0; i < adap->params.arch.nchan; i++) \
3550 		seq_printf(seq, "%-20llu ", \
3551 			   (unsigned long long)stats[i].value); \
3552 	seq_printf(seq, "\n"); \
3553 } while (0)
3554 
3555 static void show_tcp_stats(struct seq_file *seq)
3556 {
3557 	struct adapter *adap = seq->private;
3558 	struct tp_tcp_stats v4, v6;
3559 
3560 	spin_lock(&adap->stats_lock);
3561 	t4_tp_get_tcp_stats(adap, &v4, &v6, false);
3562 	spin_unlock(&adap->stats_lock);
3563 
3564 	PRINT_ADAP_STATS("tcp_ipv4_out_rsts:", v4.tcp_out_rsts);
3565 	PRINT_ADAP_STATS("tcp_ipv4_in_segs:", v4.tcp_in_segs);
3566 	PRINT_ADAP_STATS("tcp_ipv4_out_segs:", v4.tcp_out_segs);
3567 	PRINT_ADAP_STATS("tcp_ipv4_retrans_segs:", v4.tcp_retrans_segs);
3568 	PRINT_ADAP_STATS("tcp_ipv6_out_rsts:", v6.tcp_out_rsts);
3569 	PRINT_ADAP_STATS("tcp_ipv6_in_segs:", v6.tcp_in_segs);
3570 	PRINT_ADAP_STATS("tcp_ipv6_out_segs:", v6.tcp_out_segs);
3571 	PRINT_ADAP_STATS("tcp_ipv6_retrans_segs:", v6.tcp_retrans_segs);
3572 }
3573 
3574 static void show_ddp_stats(struct seq_file *seq)
3575 {
3576 	struct adapter *adap = seq->private;
3577 	struct tp_usm_stats stats;
3578 
3579 	spin_lock(&adap->stats_lock);
3580 	t4_get_usm_stats(adap, &stats, false);
3581 	spin_unlock(&adap->stats_lock);
3582 
3583 	PRINT_ADAP_STATS("usm_ddp_frames:", stats.frames);
3584 	PRINT_ADAP_STATS("usm_ddp_octets:", stats.octets);
3585 	PRINT_ADAP_STATS("usm_ddp_drops:", stats.drops);
3586 }
3587 
3588 static void show_rdma_stats(struct seq_file *seq)
3589 {
3590 	struct adapter *adap = seq->private;
3591 	struct tp_rdma_stats stats;
3592 
3593 	spin_lock(&adap->stats_lock);
3594 	t4_tp_get_rdma_stats(adap, &stats, false);
3595 	spin_unlock(&adap->stats_lock);
3596 
3597 	PRINT_ADAP_STATS("rdma_no_rqe_mod_defer:", stats.rqe_dfr_mod);
3598 	PRINT_ADAP_STATS("rdma_no_rqe_pkt_defer:", stats.rqe_dfr_pkt);
3599 }
3600 
3601 static void show_tp_err_adapter_stats(struct seq_file *seq)
3602 {
3603 	struct adapter *adap = seq->private;
3604 	struct tp_err_stats stats;
3605 
3606 	spin_lock(&adap->stats_lock);
3607 	t4_tp_get_err_stats(adap, &stats, false);
3608 	spin_unlock(&adap->stats_lock);
3609 
3610 	PRINT_ADAP_STATS("tp_err_ofld_no_neigh:", stats.ofld_no_neigh);
3611 	PRINT_ADAP_STATS("tp_err_ofld_cong_defer:", stats.ofld_cong_defer);
3612 }
3613 
3614 static void show_cpl_stats(struct seq_file *seq)
3615 {
3616 	struct adapter *adap = seq->private;
3617 	struct tp_cpl_stats stats;
3618 	u8 i;
3619 
3620 	spin_lock(&adap->stats_lock);
3621 	t4_tp_get_cpl_stats(adap, &stats, false);
3622 	spin_unlock(&adap->stats_lock);
3623 
3624 	PRINT_CH_STATS("tp_cpl_requests:", req);
3625 	PRINT_CH_STATS("tp_cpl_responses:", rsp);
3626 }
3627 
3628 static void show_tp_err_channel_stats(struct seq_file *seq)
3629 {
3630 	struct adapter *adap = seq->private;
3631 	struct tp_err_stats stats;
3632 	u8 i;
3633 
3634 	spin_lock(&adap->stats_lock);
3635 	t4_tp_get_err_stats(adap, &stats, false);
3636 	spin_unlock(&adap->stats_lock);
3637 
3638 	PRINT_CH_STATS("tp_mac_in_errs:", mac_in_errs);
3639 	PRINT_CH_STATS("tp_hdr_in_errs:", hdr_in_errs);
3640 	PRINT_CH_STATS("tp_tcp_in_errs:", tcp_in_errs);
3641 	PRINT_CH_STATS("tp_tcp6_in_errs:", tcp6_in_errs);
3642 	PRINT_CH_STATS("tp_tnl_cong_drops:", tnl_cong_drops);
3643 	PRINT_CH_STATS("tp_tnl_tx_drops:", tnl_tx_drops);
3644 	PRINT_CH_STATS("tp_ofld_vlan_drops:", ofld_vlan_drops);
3645 	PRINT_CH_STATS("tp_ofld_chan_drops:", ofld_chan_drops);
3646 }
3647 
3648 static void show_fcoe_stats(struct seq_file *seq)
3649 {
3650 	struct adapter *adap = seq->private;
3651 	struct tp_fcoe_stats stats[NCHAN];
3652 	u8 i;
3653 
3654 	spin_lock(&adap->stats_lock);
3655 	for (i = 0; i < adap->params.arch.nchan; i++)
3656 		t4_get_fcoe_stats(adap, i, &stats[i], false);
3657 	spin_unlock(&adap->stats_lock);
3658 
3659 	PRINT_CH_STATS2("fcoe_octets_ddp", octets_ddp);
3660 	PRINT_CH_STATS2("fcoe_frames_ddp", frames_ddp);
3661 	PRINT_CH_STATS2("fcoe_frames_drop", frames_drop);
3662 }
3663 
3664 #undef PRINT_CH_STATS2
3665 #undef PRINT_CH_STATS
3666 #undef PRINT_ADAP_STATS
3667 
3668 static int tp_stats_show(struct seq_file *seq, void *v)
3669 {
3670 	struct adapter *adap = seq->private;
3671 
3672 	seq_puts(seq, "\n--------Adapter Stats--------\n");
3673 	show_tcp_stats(seq);
3674 	show_ddp_stats(seq);
3675 	show_rdma_stats(seq);
3676 	show_tp_err_adapter_stats(seq);
3677 
3678 	seq_puts(seq, "\n-------- Channel Stats --------\n");
3679 	if (adap->params.arch.nchan == NCHAN)
3680 		seq_printf(seq, "%-25s %-20s %-20s %-20s %-20s\n",
3681 			   " ", "channel 0", "channel 1",
3682 			   "channel 2", "channel 3");
3683 	else
3684 		seq_printf(seq, "%-25s %-20s %-20s\n",
3685 			   " ", "channel 0", "channel 1");
3686 	show_cpl_stats(seq);
3687 	show_tp_err_channel_stats(seq);
3688 	show_fcoe_stats(seq);
3689 
3690 	return 0;
3691 }
3692 DEFINE_SHOW_ATTRIBUTE(tp_stats);
3693 
3694 /* Add an array of Debug FS files.
3695  */
3696 void add_debugfs_files(struct adapter *adap,
3697 		       struct t4_debugfs_entry *files,
3698 		       unsigned int nfiles)
3699 {
3700 	int i;
3701 
3702 	/* debugfs support is best effort */
3703 	for (i = 0; i < nfiles; i++)
3704 		debugfs_create_file(files[i].name, files[i].mode,
3705 				    adap->debugfs_root,
3706 				    (void *)adap + files[i].data,
3707 				    files[i].ops);
3708 }
3709 
3710 int t4_setup_debugfs(struct adapter *adap)
3711 {
3712 	int i;
3713 	u32 size = 0;
3714 
3715 	static struct t4_debugfs_entry t4_debugfs_files[] = {
3716 		{ "cim_la", &cim_la_fops, 0400, 0 },
3717 		{ "cim_pif_la", &cim_pif_la_fops, 0400, 0 },
3718 		{ "cim_ma_la", &cim_ma_la_fops, 0400, 0 },
3719 		{ "cim_qcfg", &cim_qcfg_fops, 0400, 0 },
3720 		{ "clk", &clk_fops, 0400, 0 },
3721 		{ "devlog", &devlog_fops, 0400, 0 },
3722 		{ "mboxlog", &mboxlog_fops, 0400, 0 },
3723 		{ "mbox0", &mbox_debugfs_fops, 0600, 0 },
3724 		{ "mbox1", &mbox_debugfs_fops, 0600, 1 },
3725 		{ "mbox2", &mbox_debugfs_fops, 0600, 2 },
3726 		{ "mbox3", &mbox_debugfs_fops, 0600, 3 },
3727 		{ "mbox4", &mbox_debugfs_fops, 0600, 4 },
3728 		{ "mbox5", &mbox_debugfs_fops, 0600, 5 },
3729 		{ "mbox6", &mbox_debugfs_fops, 0600, 6 },
3730 		{ "mbox7", &mbox_debugfs_fops, 0600, 7 },
3731 		{ "trace0", &mps_trc_debugfs_fops, 0600, 0 },
3732 		{ "trace1", &mps_trc_debugfs_fops, 0600, 1 },
3733 		{ "trace2", &mps_trc_debugfs_fops, 0600, 2 },
3734 		{ "trace3", &mps_trc_debugfs_fops, 0600, 3 },
3735 		{ "l2t", &t4_l2t_fops, 0400, 0},
3736 		{ "mps_tcam", &mps_tcam_debugfs_fops, 0400, 0 },
3737 		{ "rss", &rss_debugfs_fops, 0400, 0 },
3738 		{ "rss_config", &rss_config_fops, 0400, 0 },
3739 		{ "rss_key", &rss_key_debugfs_fops, 0400, 0 },
3740 		{ "rss_pf_config", &rss_pf_config_debugfs_fops, 0400, 0 },
3741 		{ "rss_vf_config", &rss_vf_config_debugfs_fops, 0400, 0 },
3742 		{ "resources", &resources_fops, 0400, 0 },
3743 #ifdef CONFIG_CHELSIO_T4_DCB
3744 		{ "dcb_info", &dcb_info_debugfs_fops, 0400, 0 },
3745 #endif
3746 		{ "sge_qinfo", &sge_qinfo_debugfs_fops, 0400, 0 },
3747 		{ "ibq_tp0",  &cim_ibq_fops, 0400, 0 },
3748 		{ "ibq_tp1",  &cim_ibq_fops, 0400, 1 },
3749 		{ "ibq_ulp",  &cim_ibq_fops, 0400, 2 },
3750 		{ "ibq_sge0", &cim_ibq_fops, 0400, 3 },
3751 		{ "ibq_sge1", &cim_ibq_fops, 0400, 4 },
3752 		{ "ibq_ncsi", &cim_ibq_fops, 0400, 5 },
3753 		{ "obq_ulp0", &cim_obq_fops, 0400, 0 },
3754 		{ "obq_ulp1", &cim_obq_fops, 0400, 1 },
3755 		{ "obq_ulp2", &cim_obq_fops, 0400, 2 },
3756 		{ "obq_ulp3", &cim_obq_fops, 0400, 3 },
3757 		{ "obq_sge",  &cim_obq_fops, 0400, 4 },
3758 		{ "obq_ncsi", &cim_obq_fops, 0400, 5 },
3759 		{ "tp_la", &tp_la_fops, 0400, 0 },
3760 		{ "ulprx_la", &ulprx_la_fops, 0400, 0 },
3761 		{ "sensors", &sensors_fops, 0400, 0 },
3762 		{ "pm_stats", &pm_stats_debugfs_fops, 0400, 0 },
3763 		{ "tx_rate", &tx_rate_fops, 0400, 0 },
3764 		{ "cctrl", &cctrl_tbl_fops, 0400, 0 },
3765 #if IS_ENABLED(CONFIG_IPV6)
3766 		{ "clip_tbl", &clip_tbl_fops, 0400, 0 },
3767 #endif
3768 		{ "tids", &tid_info_fops, 0400, 0},
3769 		{ "blocked_fl", &blocked_fl_fops, 0600, 0 },
3770 		{ "meminfo", &meminfo_fops, 0400, 0 },
3771 		{ "crypto", &chcr_stats_fops, 0400, 0 },
3772 		{ "tp_stats", &tp_stats_fops, 0400, 0 },
3773 	};
3774 
3775 	/* Debug FS nodes common to all T5 and later adapters.
3776 	 */
3777 	static struct t4_debugfs_entry t5_debugfs_files[] = {
3778 		{ "obq_sge_rx_q0", &cim_obq_fops, 0400, 6 },
3779 		{ "obq_sge_rx_q1", &cim_obq_fops, 0400, 7 },
3780 	};
3781 
3782 	add_debugfs_files(adap,
3783 			  t4_debugfs_files,
3784 			  ARRAY_SIZE(t4_debugfs_files));
3785 	if (!is_t4(adap->params.chip))
3786 		add_debugfs_files(adap,
3787 				  t5_debugfs_files,
3788 				  ARRAY_SIZE(t5_debugfs_files));
3789 
3790 	i = t4_read_reg(adap, MA_TARGET_MEM_ENABLE_A);
3791 	if (i & EDRAM0_ENABLE_F) {
3792 		size = t4_read_reg(adap, MA_EDRAM0_BAR_A);
3793 		add_debugfs_mem(adap, "edc0", MEM_EDC0, EDRAM0_SIZE_G(size));
3794 	}
3795 	if (i & EDRAM1_ENABLE_F) {
3796 		size = t4_read_reg(adap, MA_EDRAM1_BAR_A);
3797 		add_debugfs_mem(adap, "edc1", MEM_EDC1, EDRAM1_SIZE_G(size));
3798 	}
3799 	if (is_t5(adap->params.chip)) {
3800 		if (i & EXT_MEM0_ENABLE_F) {
3801 			size = t4_read_reg(adap, MA_EXT_MEMORY0_BAR_A);
3802 			add_debugfs_mem(adap, "mc0", MEM_MC0,
3803 					EXT_MEM0_SIZE_G(size));
3804 		}
3805 		if (i & EXT_MEM1_ENABLE_F) {
3806 			size = t4_read_reg(adap, MA_EXT_MEMORY1_BAR_A);
3807 			add_debugfs_mem(adap, "mc1", MEM_MC1,
3808 					EXT_MEM1_SIZE_G(size));
3809 		}
3810 	} else {
3811 		if (i & EXT_MEM_ENABLE_F) {
3812 			size = t4_read_reg(adap, MA_EXT_MEMORY_BAR_A);
3813 			add_debugfs_mem(adap, "mc", MEM_MC,
3814 					EXT_MEM_SIZE_G(size));
3815 		}
3816 
3817 		if (i & HMA_MUX_F) {
3818 			size = t4_read_reg(adap, MA_EXT_MEMORY1_BAR_A);
3819 			add_debugfs_mem(adap, "hma", MEM_HMA,
3820 					EXT_MEM1_SIZE_G(size));
3821 		}
3822 	}
3823 
3824 	debugfs_create_file_size("flash", 0400, adap->debugfs_root, adap,
3825 				 &flash_debugfs_fops, adap->params.sf_size);
3826 	debugfs_create_bool("use_backdoor", 0600,
3827 			    adap->debugfs_root, &adap->use_bd);
3828 	debugfs_create_bool("trace_rss", 0600,
3829 			    adap->debugfs_root, &adap->trace_rss);
3830 
3831 	return 0;
3832 }
3833