xref: /openbmc/linux/drivers/scsi/csiostor/csio_wr.h (revision 2c684d89)
1 /*
2  * This file is part of the Chelsio FCoE driver for Linux.
3  *
4  * Copyright (c) 2008-2012 Chelsio Communications, Inc. All rights reserved.
5  *
6  * This software is available to you under a choice of one of two
7  * licenses.  You may choose to be licensed under the terms of the GNU
8  * General Public License (GPL) Version 2, available from the file
9  * COPYING in the main directory of this source tree, or the
10  * OpenIB.org BSD license below:
11  *
12  *     Redistribution and use in source and binary forms, with or
13  *     without modification, are permitted provided that the following
14  *     conditions are met:
15  *
16  *      - Redistributions of source code must retain the above
17  *        copyright notice, this list of conditions and the following
18  *        disclaimer.
19  *
20  *      - Redistributions in binary form must reproduce the above
21  *        copyright notice, this list of conditions and the following
22  *        disclaimer in the documentation and/or other materials
23  *        provided with the distribution.
24  *
25  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
26  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
27  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
28  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
29  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
30  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
31  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
32  * SOFTWARE.
33  */
34 
35 #ifndef __CSIO_WR_H__
36 #define __CSIO_WR_H__
37 
38 #include <linux/cache.h>
39 
40 #include "csio_defs.h"
41 #include "t4fw_api.h"
42 #include "t4fw_api_stor.h"
43 
44 /*
45  * SGE register field values.
46  */
47 #define X_INGPCIEBOUNDARY_32B		0
48 #define X_INGPCIEBOUNDARY_64B		1
49 #define X_INGPCIEBOUNDARY_128B		2
50 #define X_INGPCIEBOUNDARY_256B		3
51 #define X_INGPCIEBOUNDARY_512B		4
52 #define X_INGPCIEBOUNDARY_1024B		5
53 #define X_INGPCIEBOUNDARY_2048B		6
54 #define X_INGPCIEBOUNDARY_4096B		7
55 
56 /* GTS register */
57 #define X_TIMERREG_COUNTER0		0
58 #define X_TIMERREG_COUNTER1		1
59 #define X_TIMERREG_COUNTER2		2
60 #define X_TIMERREG_COUNTER3		3
61 #define X_TIMERREG_COUNTER4		4
62 #define X_TIMERREG_COUNTER5		5
63 #define X_TIMERREG_RESTART_COUNTER	6
64 #define X_TIMERREG_UPDATE_CIDX		7
65 
66 /*
67  * Egress Context field values
68  */
69 #define X_FETCHBURSTMIN_16B		0
70 #define X_FETCHBURSTMIN_32B		1
71 #define X_FETCHBURSTMIN_64B		2
72 #define X_FETCHBURSTMIN_128B		3
73 
74 #define X_FETCHBURSTMAX_64B		0
75 #define X_FETCHBURSTMAX_128B		1
76 #define X_FETCHBURSTMAX_256B		2
77 #define X_FETCHBURSTMAX_512B		3
78 
79 #define X_HOSTFCMODE_NONE		0
80 #define X_HOSTFCMODE_INGRESS_QUEUE	1
81 #define X_HOSTFCMODE_STATUS_PAGE	2
82 #define X_HOSTFCMODE_BOTH		3
83 
84 /*
85  * Ingress Context field values
86  */
87 #define X_UPDATESCHEDULING_TIMER	0
88 #define X_UPDATESCHEDULING_COUNTER_OPTTIMER	1
89 
90 #define X_UPDATEDELIVERY_NONE		0
91 #define X_UPDATEDELIVERY_INTERRUPT	1
92 #define X_UPDATEDELIVERY_STATUS_PAGE	2
93 #define X_UPDATEDELIVERY_BOTH		3
94 
95 #define X_INTERRUPTDESTINATION_PCIE	0
96 #define X_INTERRUPTDESTINATION_IQ	1
97 
98 #define X_RSPD_TYPE_FLBUF		0
99 #define X_RSPD_TYPE_CPL			1
100 #define X_RSPD_TYPE_INTR		2
101 
102 /* WR status is at the same position as retval in a CMD header */
103 #define csio_wr_status(_wr)		\
104 		(FW_CMD_RETVAL_G(ntohl(((struct fw_cmd_hdr *)(_wr))->lo)))
105 
106 struct csio_hw;
107 
108 extern int csio_intr_coalesce_cnt;
109 extern int csio_intr_coalesce_time;
110 
111 /* Ingress queue params */
112 struct csio_iq_params {
113 
114 	uint8_t		iq_start:1;
115 	uint8_t		iq_stop:1;
116 	uint8_t		pfn:3;
117 
118 	uint8_t		vfn;
119 
120 	uint16_t	physiqid;
121 	uint16_t	iqid;
122 
123 	uint16_t	fl0id;
124 	uint16_t	fl1id;
125 
126 	uint8_t		viid;
127 
128 	uint8_t		type;
129 	uint8_t		iqasynch;
130 	uint8_t		reserved4;
131 
132 	uint8_t		iqandst;
133 	uint8_t		iqanus;
134 	uint8_t		iqanud;
135 
136 	uint16_t	iqandstindex;
137 
138 	uint8_t		iqdroprss;
139 	uint8_t		iqpciech;
140 	uint8_t		iqdcaen;
141 
142 	uint8_t		iqdcacpu;
143 	uint8_t		iqintcntthresh;
144 	uint8_t		iqo;
145 
146 	uint8_t		iqcprio;
147 	uint8_t		iqesize;
148 
149 	uint16_t	iqsize;
150 
151 	uint64_t	iqaddr;
152 
153 	uint8_t		iqflintiqhsen;
154 	uint8_t		reserved5;
155 	uint8_t		iqflintcongen;
156 	uint8_t		iqflintcngchmap;
157 
158 	uint32_t	reserved6;
159 
160 	uint8_t		fl0hostfcmode;
161 	uint8_t		fl0cprio;
162 	uint8_t		fl0paden;
163 	uint8_t		fl0packen;
164 	uint8_t		fl0congen;
165 	uint8_t		fl0dcaen;
166 
167 	uint8_t		fl0dcacpu;
168 	uint8_t		fl0fbmin;
169 
170 	uint8_t		fl0fbmax;
171 	uint8_t		fl0cidxfthresho;
172 	uint8_t		fl0cidxfthresh;
173 
174 	uint16_t	fl0size;
175 
176 	uint64_t	fl0addr;
177 
178 	uint64_t	reserved7;
179 
180 	uint8_t		fl1hostfcmode;
181 	uint8_t		fl1cprio;
182 	uint8_t		fl1paden;
183 	uint8_t		fl1packen;
184 	uint8_t		fl1congen;
185 	uint8_t		fl1dcaen;
186 
187 	uint8_t		fl1dcacpu;
188 	uint8_t		fl1fbmin;
189 
190 	uint8_t		fl1fbmax;
191 	uint8_t		fl1cidxfthresho;
192 	uint8_t		fl1cidxfthresh;
193 
194 	uint16_t	fl1size;
195 
196 	uint64_t	fl1addr;
197 };
198 
199 /* Egress queue params */
200 struct csio_eq_params {
201 
202 	uint8_t		pfn;
203 	uint8_t		vfn;
204 
205 	uint8_t		eqstart:1;
206 	uint8_t		eqstop:1;
207 
208 	uint16_t        physeqid;
209 	uint32_t	eqid;
210 
211 	uint8_t		hostfcmode:2;
212 	uint8_t		cprio:1;
213 	uint8_t		pciechn:3;
214 
215 	uint16_t	iqid;
216 
217 	uint8_t		dcaen:1;
218 	uint8_t		dcacpu:5;
219 
220 	uint8_t		fbmin:3;
221 	uint8_t		fbmax:3;
222 
223 	uint8_t		cidxfthresho:1;
224 	uint8_t		cidxfthresh:3;
225 
226 	uint16_t	eqsize;
227 
228 	uint64_t	eqaddr;
229 };
230 
231 struct csio_dma_buf {
232 	struct list_head	list;
233 	void			*vaddr;		/* Virtual address */
234 	dma_addr_t		paddr;		/* Physical address */
235 	uint32_t		len;		/* Buffer size */
236 };
237 
238 /* Generic I/O request structure */
239 struct csio_ioreq {
240 	struct csio_sm		sm;		/* SM, List
241 						 * should be the first member
242 						 */
243 	int			iq_idx;		/* Ingress queue index */
244 	int			eq_idx;		/* Egress queue index */
245 	uint32_t		nsge;		/* Number of SG elements */
246 	uint32_t		tmo;		/* Driver timeout */
247 	uint32_t		datadir;	/* Data direction */
248 	struct csio_dma_buf	dma_buf;	/* Req/resp DMA buffers */
249 	uint16_t		wr_status;	/* WR completion status */
250 	int16_t			drv_status;	/* Driver internal status */
251 	struct csio_lnode	*lnode;		/* Owner lnode */
252 	struct csio_rnode	*rnode;		/* Src/destination rnode */
253 	void (*io_cbfn) (struct csio_hw *, struct csio_ioreq *);
254 						/* completion callback */
255 	void			*scratch1;	/* Scratch area 1.
256 						 */
257 	void			*scratch2;	/* Scratch area 2. */
258 	struct list_head	gen_list;	/* Any list associated with
259 						 * this ioreq.
260 						 */
261 	uint64_t		fw_handle;	/* Unique handle passed
262 						 * to FW
263 						 */
264 	uint8_t			dcopy;		/* Data copy required */
265 	uint8_t			reserved1;
266 	uint16_t		reserved2;
267 	struct completion	cmplobj;	/* ioreq completion object */
268 } ____cacheline_aligned_in_smp;
269 
270 /*
271  * Egress status page for egress cidx updates
272  */
273 struct csio_qstatus_page {
274 	__be32 qid;
275 	__be16 cidx;
276 	__be16 pidx;
277 };
278 
279 
280 enum {
281 	CSIO_MAX_FLBUF_PER_IQWR = 4,
282 	CSIO_QCREDIT_SZ  = 64,			/* pidx/cidx increments
283 						 * in bytes
284 						 */
285 	CSIO_MAX_QID = 0xFFFF,
286 	CSIO_MAX_IQ = 128,
287 
288 	CSIO_SGE_NTIMERS = 6,
289 	CSIO_SGE_NCOUNTERS = 4,
290 	CSIO_SGE_FL_SIZE_REGS = 16,
291 };
292 
293 /* Defines for type */
294 enum {
295 	CSIO_EGRESS	= 1,
296 	CSIO_INGRESS	= 2,
297 	CSIO_FREELIST	= 3,
298 };
299 
300 /*
301  * Structure for footer (last 2 flits) of Ingress Queue Entry.
302  */
303 struct csio_iqwr_footer {
304 	__be32			hdrbuflen_pidx;
305 	__be32			pldbuflen_qid;
306 	union {
307 		u8		type_gen;
308 		__be64		last_flit;
309 	} u;
310 };
311 
312 #define IQWRF_NEWBUF		(1 << 31)
313 #define IQWRF_LEN_GET(x)	(((x) >> 0) & 0x7fffffffU)
314 #define IQWRF_GEN_SHIFT		7
315 #define IQWRF_TYPE_GET(x)	(((x) >> 4) & 0x3U)
316 
317 
318 /*
319  * WR pair:
320  * ========
321  * A WR can start towards the end of a queue, and then continue at the
322  * beginning, since the queue is considered to be circular. This will
323  * require a pair of address/len to be passed back to the caller -
324  * hence the Work request pair structure.
325  */
326 struct csio_wr_pair {
327 	void			*addr1;
328 	uint32_t		size1;
329 	void			*addr2;
330 	uint32_t		size2;
331 };
332 
333 /*
334  * The following structure is used by ingress processing to return the
335  * free list buffers to consumers.
336  */
337 struct csio_fl_dma_buf {
338 	struct csio_dma_buf	flbufs[CSIO_MAX_FLBUF_PER_IQWR];
339 						/* Freelist DMA buffers */
340 	int			offset;		/* Offset within the
341 						 * first FL buf.
342 						 */
343 	uint32_t		totlen;		/* Total length */
344 	uint8_t			defer_free;	/* Free of buffer can
345 						 * deferred
346 						 */
347 };
348 
349 /* Data-types */
350 typedef void (*iq_handler_t)(struct csio_hw *, void *, uint32_t,
351 			     struct csio_fl_dma_buf *, void *);
352 
353 struct csio_iq {
354 	uint16_t		iqid;		/* Queue ID */
355 	uint16_t		physiqid;	/* Physical Queue ID */
356 	uint16_t		genbit;		/* Generation bit,
357 						 * initially set to 1
358 						 */
359 	int			flq_idx;	/* Freelist queue index */
360 	iq_handler_t		iq_intx_handler; /* IQ INTx handler routine */
361 };
362 
363 struct csio_eq {
364 	uint16_t		eqid;		/* Qid */
365 	uint16_t		physeqid;	/* Physical Queue ID */
366 	uint8_t			wrap[512];	/* Temp area for q-wrap around*/
367 };
368 
369 struct csio_fl {
370 	uint16_t		flid;		/* Qid */
371 	uint16_t		packen;		/* Packing enabled? */
372 	int			offset;		/* Offset within FL buf */
373 	int			sreg;		/* Size register */
374 	struct csio_dma_buf	*bufs;		/* Free list buffer ptr array
375 						 * indexed using flq->cidx/pidx
376 						 */
377 };
378 
379 struct csio_qstats {
380 	uint32_t	n_tot_reqs;		/* Total no. of Requests */
381 	uint32_t	n_tot_rsps;		/* Total no. of responses */
382 	uint32_t	n_qwrap;		/* Queue wraps */
383 	uint32_t	n_eq_wr_split;		/* Number of split EQ WRs */
384 	uint32_t	n_qentry;		/* Queue entry */
385 	uint32_t	n_qempty;		/* Queue empty */
386 	uint32_t	n_qfull;		/* Queue fulls */
387 	uint32_t	n_rsp_unknown;		/* Unknown response type */
388 	uint32_t	n_stray_comp;		/* Stray completion intr */
389 	uint32_t	n_flq_refill;		/* Number of FL refills */
390 };
391 
392 /* Queue metadata */
393 struct csio_q {
394 	uint16_t		type;		/* Type: Ingress/Egress/FL */
395 	uint16_t		pidx;		/* producer index */
396 	uint16_t		cidx;		/* consumer index */
397 	uint16_t		inc_idx;	/* Incremental index */
398 	uint32_t		wr_sz;		/* Size of all WRs in this q
399 						 * if fixed
400 						 */
401 	void			*vstart;	/* Base virtual address
402 						 * of queue
403 						 */
404 	void			*vwrap;		/* Virtual end address to
405 						 * wrap around at
406 						 */
407 	uint32_t		credits;	/* Size of queue in credits */
408 	void			*owner;		/* Owner */
409 	union {					/* Queue contexts */
410 		struct csio_iq	iq;
411 		struct csio_eq	eq;
412 		struct csio_fl	fl;
413 	} un;
414 
415 	dma_addr_t		pstart;		/* Base physical address of
416 						 * queue
417 						 */
418 	uint32_t		portid;		/* PCIE Channel */
419 	uint32_t		size;		/* Size of queue in bytes */
420 	struct csio_qstats	stats;		/* Statistics */
421 } ____cacheline_aligned_in_smp;
422 
423 struct csio_sge {
424 	uint32_t	csio_fl_align;		/* Calculated and cached
425 						 * for fast path
426 						 */
427 	uint32_t	sge_control;		/* padding, boundaries,
428 						 * lengths, etc.
429 						 */
430 	uint32_t	sge_host_page_size;	/* Host page size */
431 	uint32_t	sge_fl_buf_size[CSIO_SGE_FL_SIZE_REGS];
432 						/* free list buffer sizes */
433 	uint16_t	timer_val[CSIO_SGE_NTIMERS];
434 	uint8_t		counter_val[CSIO_SGE_NCOUNTERS];
435 };
436 
437 /* Work request module */
438 struct csio_wrm {
439 	int			num_q;		/* Number of queues */
440 	struct csio_q		**q_arr;	/* Array of queue pointers
441 						 * allocated dynamically
442 						 * based on configured values
443 						 */
444 	uint32_t		fw_iq_start;	/* Start ID of IQ for this fn*/
445 	uint32_t		fw_eq_start;	/* Start ID of EQ for this fn*/
446 	struct csio_q		*intr_map[CSIO_MAX_IQ];
447 						/* IQ-id to IQ map table. */
448 	int			free_qidx;	/* queue idx of free queue */
449 	struct csio_sge		sge;		/* SGE params */
450 };
451 
452 #define csio_get_q(__hw, __idx)		((__hw)->wrm.q_arr[__idx])
453 #define	csio_q_type(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->type)
454 #define	csio_q_pidx(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->pidx)
455 #define	csio_q_cidx(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->cidx)
456 #define	csio_q_inc_idx(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->inc_idx)
457 #define	csio_q_vstart(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->vstart)
458 #define	csio_q_pstart(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->pstart)
459 #define	csio_q_size(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->size)
460 #define	csio_q_credits(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->credits)
461 #define	csio_q_portid(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->portid)
462 #define	csio_q_wr_sz(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->wr_sz)
463 #define	csio_q_iqid(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->un.iq.iqid)
464 #define csio_q_physiqid(__hw, __idx)					\
465 				((__hw)->wrm.q_arr[(__idx)]->un.iq.physiqid)
466 #define csio_q_iq_flq_idx(__hw, __idx)					\
467 				((__hw)->wrm.q_arr[(__idx)]->un.iq.flq_idx)
468 #define	csio_q_eqid(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->un.eq.eqid)
469 #define	csio_q_flid(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->un.fl.flid)
470 
471 #define csio_q_physeqid(__hw, __idx)					\
472 				((__hw)->wrm.q_arr[(__idx)]->un.eq.physeqid)
473 #define csio_iq_has_fl(__iq)		((__iq)->un.iq.flq_idx != -1)
474 
475 #define csio_q_iq_to_flid(__hw, __iq_idx)				\
476 	csio_q_flid((__hw), (__hw)->wrm.q_arr[(__iq_qidx)]->un.iq.flq_idx)
477 #define csio_q_set_intr_map(__hw, __iq_idx, __rel_iq_id)		\
478 		(__hw)->wrm.intr_map[__rel_iq_id] = csio_get_q(__hw, __iq_idx)
479 #define	csio_q_eq_wrap(__hw, __idx)	((__hw)->wrm.q_arr[(__idx)]->un.eq.wrap)
480 
481 struct csio_mb;
482 
483 int csio_wr_alloc_q(struct csio_hw *, uint32_t, uint32_t,
484 		    uint16_t, void *, uint32_t, int, iq_handler_t);
485 int csio_wr_iq_create(struct csio_hw *, void *, int,
486 				uint32_t, uint8_t, bool,
487 				void (*)(struct csio_hw *, struct csio_mb *));
488 int csio_wr_eq_create(struct csio_hw *, void *, int, int, uint8_t,
489 				void (*)(struct csio_hw *, struct csio_mb *));
490 int csio_wr_destroy_queues(struct csio_hw *, bool cmd);
491 
492 
493 int csio_wr_get(struct csio_hw *, int, uint32_t,
494 			  struct csio_wr_pair *);
495 void csio_wr_copy_to_wrp(void *, struct csio_wr_pair *, uint32_t, uint32_t);
496 int csio_wr_issue(struct csio_hw *, int, bool);
497 int csio_wr_process_iq(struct csio_hw *, struct csio_q *,
498 				 void (*)(struct csio_hw *, void *,
499 					  uint32_t, struct csio_fl_dma_buf *,
500 					  void *),
501 				 void *);
502 int csio_wr_process_iq_idx(struct csio_hw *, int,
503 				 void (*)(struct csio_hw *, void *,
504 					  uint32_t, struct csio_fl_dma_buf *,
505 					  void *),
506 				 void *);
507 
508 void csio_wr_sge_init(struct csio_hw *);
509 int csio_wrm_init(struct csio_wrm *, struct csio_hw *);
510 void csio_wrm_exit(struct csio_wrm *, struct csio_hw *);
511 
512 #endif /* ifndef __CSIO_WR_H__ */
513