xref: /openbmc/linux/drivers/scsi/bfa/bfa_ioc.c (revision d0b73b48)
1 /*
2  * Copyright (c) 2005-2010 Brocade Communications Systems, Inc.
3  * All rights reserved
4  * www.brocade.com
5  *
6  * Linux driver for Brocade Fibre Channel Host Bus Adapter.
7  *
8  * This program is free software; you can redistribute it and/or modify it
9  * under the terms of the GNU General Public License (GPL) Version 2 as
10  * published by the Free Software Foundation
11  *
12  * This program is distributed in the hope that it will be useful, but
13  * WITHOUT ANY WARRANTY; without even the implied warranty of
14  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
15  * General Public License for more details.
16  */
17 
18 #include "bfad_drv.h"
19 #include "bfad_im.h"
20 #include "bfa_ioc.h"
21 #include "bfi_reg.h"
22 #include "bfa_defs.h"
23 #include "bfa_defs_svc.h"
24 
25 BFA_TRC_FILE(CNA, IOC);
26 
27 /*
28  * IOC local definitions
29  */
30 #define BFA_IOC_TOV		3000	/* msecs */
31 #define BFA_IOC_HWSEM_TOV	500	/* msecs */
32 #define BFA_IOC_HB_TOV		500	/* msecs */
33 #define BFA_IOC_TOV_RECOVER	 BFA_IOC_HB_TOV
34 #define BFA_IOC_POLL_TOV	BFA_TIMER_FREQ
35 
36 #define bfa_ioc_timer_start(__ioc)					\
37 	bfa_timer_begin((__ioc)->timer_mod, &(__ioc)->ioc_timer,	\
38 			bfa_ioc_timeout, (__ioc), BFA_IOC_TOV)
39 #define bfa_ioc_timer_stop(__ioc)   bfa_timer_stop(&(__ioc)->ioc_timer)
40 
41 #define bfa_hb_timer_start(__ioc)					\
42 	bfa_timer_begin((__ioc)->timer_mod, &(__ioc)->hb_timer,		\
43 			bfa_ioc_hb_check, (__ioc), BFA_IOC_HB_TOV)
44 #define bfa_hb_timer_stop(__ioc)	bfa_timer_stop(&(__ioc)->hb_timer)
45 
46 #define BFA_DBG_FWTRC_OFF(_fn)	(BFI_IOC_TRC_OFF + BFA_DBG_FWTRC_LEN * (_fn))
47 
48 /*
49  * Asic specific macros : see bfa_hw_cb.c and bfa_hw_ct.c for details.
50  */
51 
52 #define bfa_ioc_firmware_lock(__ioc)			\
53 			((__ioc)->ioc_hwif->ioc_firmware_lock(__ioc))
54 #define bfa_ioc_firmware_unlock(__ioc)			\
55 			((__ioc)->ioc_hwif->ioc_firmware_unlock(__ioc))
56 #define bfa_ioc_reg_init(__ioc) ((__ioc)->ioc_hwif->ioc_reg_init(__ioc))
57 #define bfa_ioc_map_port(__ioc) ((__ioc)->ioc_hwif->ioc_map_port(__ioc))
58 #define bfa_ioc_notify_fail(__ioc)              \
59 			((__ioc)->ioc_hwif->ioc_notify_fail(__ioc))
60 #define bfa_ioc_sync_start(__ioc)               \
61 			((__ioc)->ioc_hwif->ioc_sync_start(__ioc))
62 #define bfa_ioc_sync_join(__ioc)                \
63 			((__ioc)->ioc_hwif->ioc_sync_join(__ioc))
64 #define bfa_ioc_sync_leave(__ioc)               \
65 			((__ioc)->ioc_hwif->ioc_sync_leave(__ioc))
66 #define bfa_ioc_sync_ack(__ioc)                 \
67 			((__ioc)->ioc_hwif->ioc_sync_ack(__ioc))
68 #define bfa_ioc_sync_complete(__ioc)            \
69 			((__ioc)->ioc_hwif->ioc_sync_complete(__ioc))
70 
71 #define bfa_ioc_mbox_cmd_pending(__ioc)		\
72 			(!list_empty(&((__ioc)->mbox_mod.cmd_q)) || \
73 			readl((__ioc)->ioc_regs.hfn_mbox_cmd))
74 
75 bfa_boolean_t bfa_auto_recover = BFA_TRUE;
76 
77 /*
78  * forward declarations
79  */
80 static void bfa_ioc_hw_sem_get(struct bfa_ioc_s *ioc);
81 static void bfa_ioc_hwinit(struct bfa_ioc_s *ioc, bfa_boolean_t force);
82 static void bfa_ioc_timeout(void *ioc);
83 static void bfa_ioc_poll_fwinit(struct bfa_ioc_s *ioc);
84 static void bfa_ioc_send_enable(struct bfa_ioc_s *ioc);
85 static void bfa_ioc_send_disable(struct bfa_ioc_s *ioc);
86 static void bfa_ioc_send_getattr(struct bfa_ioc_s *ioc);
87 static void bfa_ioc_hb_monitor(struct bfa_ioc_s *ioc);
88 static void bfa_ioc_mbox_poll(struct bfa_ioc_s *ioc);
89 static void bfa_ioc_mbox_flush(struct bfa_ioc_s *ioc);
90 static void bfa_ioc_recover(struct bfa_ioc_s *ioc);
91 static void bfa_ioc_event_notify(struct bfa_ioc_s *ioc ,
92 				enum bfa_ioc_event_e event);
93 static void bfa_ioc_disable_comp(struct bfa_ioc_s *ioc);
94 static void bfa_ioc_lpu_stop(struct bfa_ioc_s *ioc);
95 static void bfa_ioc_fail_notify(struct bfa_ioc_s *ioc);
96 static void bfa_ioc_pf_fwmismatch(struct bfa_ioc_s *ioc);
97 
98 /*
99  * IOC state machine definitions/declarations
100  */
101 enum ioc_event {
102 	IOC_E_RESET		= 1,	/*  IOC reset request		*/
103 	IOC_E_ENABLE		= 2,	/*  IOC enable request		*/
104 	IOC_E_DISABLE		= 3,	/*  IOC disable request	*/
105 	IOC_E_DETACH		= 4,	/*  driver detach cleanup	*/
106 	IOC_E_ENABLED		= 5,	/*  f/w enabled		*/
107 	IOC_E_FWRSP_GETATTR	= 6,	/*  IOC get attribute response	*/
108 	IOC_E_DISABLED		= 7,	/*  f/w disabled		*/
109 	IOC_E_PFFAILED		= 8,	/*  failure notice by iocpf sm	*/
110 	IOC_E_HBFAIL		= 9,	/*  heartbeat failure		*/
111 	IOC_E_HWERROR		= 10,	/*  hardware error interrupt	*/
112 	IOC_E_TIMEOUT		= 11,	/*  timeout			*/
113 	IOC_E_HWFAILED		= 12,	/*  PCI mapping failure notice	*/
114 };
115 
116 bfa_fsm_state_decl(bfa_ioc, uninit, struct bfa_ioc_s, enum ioc_event);
117 bfa_fsm_state_decl(bfa_ioc, reset, struct bfa_ioc_s, enum ioc_event);
118 bfa_fsm_state_decl(bfa_ioc, enabling, struct bfa_ioc_s, enum ioc_event);
119 bfa_fsm_state_decl(bfa_ioc, getattr, struct bfa_ioc_s, enum ioc_event);
120 bfa_fsm_state_decl(bfa_ioc, op, struct bfa_ioc_s, enum ioc_event);
121 bfa_fsm_state_decl(bfa_ioc, fail_retry, struct bfa_ioc_s, enum ioc_event);
122 bfa_fsm_state_decl(bfa_ioc, fail, struct bfa_ioc_s, enum ioc_event);
123 bfa_fsm_state_decl(bfa_ioc, disabling, struct bfa_ioc_s, enum ioc_event);
124 bfa_fsm_state_decl(bfa_ioc, disabled, struct bfa_ioc_s, enum ioc_event);
125 bfa_fsm_state_decl(bfa_ioc, hwfail, struct bfa_ioc_s, enum ioc_event);
126 
127 static struct bfa_sm_table_s ioc_sm_table[] = {
128 	{BFA_SM(bfa_ioc_sm_uninit), BFA_IOC_UNINIT},
129 	{BFA_SM(bfa_ioc_sm_reset), BFA_IOC_RESET},
130 	{BFA_SM(bfa_ioc_sm_enabling), BFA_IOC_ENABLING},
131 	{BFA_SM(bfa_ioc_sm_getattr), BFA_IOC_GETATTR},
132 	{BFA_SM(bfa_ioc_sm_op), BFA_IOC_OPERATIONAL},
133 	{BFA_SM(bfa_ioc_sm_fail_retry), BFA_IOC_INITFAIL},
134 	{BFA_SM(bfa_ioc_sm_fail), BFA_IOC_FAIL},
135 	{BFA_SM(bfa_ioc_sm_disabling), BFA_IOC_DISABLING},
136 	{BFA_SM(bfa_ioc_sm_disabled), BFA_IOC_DISABLED},
137 	{BFA_SM(bfa_ioc_sm_hwfail), BFA_IOC_HWFAIL},
138 };
139 
140 /*
141  * IOCPF state machine definitions/declarations
142  */
143 
144 #define bfa_iocpf_timer_start(__ioc)					\
145 	bfa_timer_begin((__ioc)->timer_mod, &(__ioc)->ioc_timer,	\
146 			bfa_iocpf_timeout, (__ioc), BFA_IOC_TOV)
147 #define bfa_iocpf_timer_stop(__ioc)	bfa_timer_stop(&(__ioc)->ioc_timer)
148 
149 #define bfa_iocpf_poll_timer_start(__ioc)				\
150 	bfa_timer_begin((__ioc)->timer_mod, &(__ioc)->ioc_timer,	\
151 			bfa_iocpf_poll_timeout, (__ioc), BFA_IOC_POLL_TOV)
152 
153 #define bfa_sem_timer_start(__ioc)					\
154 	bfa_timer_begin((__ioc)->timer_mod, &(__ioc)->sem_timer,	\
155 			bfa_iocpf_sem_timeout, (__ioc), BFA_IOC_HWSEM_TOV)
156 #define bfa_sem_timer_stop(__ioc)	bfa_timer_stop(&(__ioc)->sem_timer)
157 
158 /*
159  * Forward declareations for iocpf state machine
160  */
161 static void bfa_iocpf_timeout(void *ioc_arg);
162 static void bfa_iocpf_sem_timeout(void *ioc_arg);
163 static void bfa_iocpf_poll_timeout(void *ioc_arg);
164 
165 /*
166  * IOCPF state machine events
167  */
168 enum iocpf_event {
169 	IOCPF_E_ENABLE		= 1,	/*  IOCPF enable request	*/
170 	IOCPF_E_DISABLE		= 2,	/*  IOCPF disable request	*/
171 	IOCPF_E_STOP		= 3,	/*  stop on driver detach	*/
172 	IOCPF_E_FWREADY		= 4,	/*  f/w initialization done	*/
173 	IOCPF_E_FWRSP_ENABLE	= 5,	/*  enable f/w response	*/
174 	IOCPF_E_FWRSP_DISABLE	= 6,	/*  disable f/w response	*/
175 	IOCPF_E_FAIL		= 7,	/*  failure notice by ioc sm	*/
176 	IOCPF_E_INITFAIL	= 8,	/*  init fail notice by ioc sm	*/
177 	IOCPF_E_GETATTRFAIL	= 9,	/*  init fail notice by ioc sm	*/
178 	IOCPF_E_SEMLOCKED	= 10,	/*  h/w semaphore is locked	*/
179 	IOCPF_E_TIMEOUT		= 11,	/*  f/w response timeout	*/
180 	IOCPF_E_SEM_ERROR	= 12,	/*  h/w sem mapping error	*/
181 };
182 
183 /*
184  * IOCPF states
185  */
186 enum bfa_iocpf_state {
187 	BFA_IOCPF_RESET		= 1,	/*  IOC is in reset state */
188 	BFA_IOCPF_SEMWAIT	= 2,	/*  Waiting for IOC h/w semaphore */
189 	BFA_IOCPF_HWINIT	= 3,	/*  IOC h/w is being initialized */
190 	BFA_IOCPF_READY		= 4,	/*  IOCPF is initialized */
191 	BFA_IOCPF_INITFAIL	= 5,	/*  IOCPF failed */
192 	BFA_IOCPF_FAIL		= 6,	/*  IOCPF failed */
193 	BFA_IOCPF_DISABLING	= 7,	/*  IOCPF is being disabled */
194 	BFA_IOCPF_DISABLED	= 8,	/*  IOCPF is disabled */
195 	BFA_IOCPF_FWMISMATCH	= 9,	/*  IOC f/w different from drivers */
196 };
197 
198 bfa_fsm_state_decl(bfa_iocpf, reset, struct bfa_iocpf_s, enum iocpf_event);
199 bfa_fsm_state_decl(bfa_iocpf, fwcheck, struct bfa_iocpf_s, enum iocpf_event);
200 bfa_fsm_state_decl(bfa_iocpf, mismatch, struct bfa_iocpf_s, enum iocpf_event);
201 bfa_fsm_state_decl(bfa_iocpf, semwait, struct bfa_iocpf_s, enum iocpf_event);
202 bfa_fsm_state_decl(bfa_iocpf, hwinit, struct bfa_iocpf_s, enum iocpf_event);
203 bfa_fsm_state_decl(bfa_iocpf, enabling, struct bfa_iocpf_s, enum iocpf_event);
204 bfa_fsm_state_decl(bfa_iocpf, ready, struct bfa_iocpf_s, enum iocpf_event);
205 bfa_fsm_state_decl(bfa_iocpf, initfail_sync, struct bfa_iocpf_s,
206 						enum iocpf_event);
207 bfa_fsm_state_decl(bfa_iocpf, initfail, struct bfa_iocpf_s, enum iocpf_event);
208 bfa_fsm_state_decl(bfa_iocpf, fail_sync, struct bfa_iocpf_s, enum iocpf_event);
209 bfa_fsm_state_decl(bfa_iocpf, fail, struct bfa_iocpf_s, enum iocpf_event);
210 bfa_fsm_state_decl(bfa_iocpf, disabling, struct bfa_iocpf_s, enum iocpf_event);
211 bfa_fsm_state_decl(bfa_iocpf, disabling_sync, struct bfa_iocpf_s,
212 						enum iocpf_event);
213 bfa_fsm_state_decl(bfa_iocpf, disabled, struct bfa_iocpf_s, enum iocpf_event);
214 
215 static struct bfa_sm_table_s iocpf_sm_table[] = {
216 	{BFA_SM(bfa_iocpf_sm_reset), BFA_IOCPF_RESET},
217 	{BFA_SM(bfa_iocpf_sm_fwcheck), BFA_IOCPF_FWMISMATCH},
218 	{BFA_SM(bfa_iocpf_sm_mismatch), BFA_IOCPF_FWMISMATCH},
219 	{BFA_SM(bfa_iocpf_sm_semwait), BFA_IOCPF_SEMWAIT},
220 	{BFA_SM(bfa_iocpf_sm_hwinit), BFA_IOCPF_HWINIT},
221 	{BFA_SM(bfa_iocpf_sm_enabling), BFA_IOCPF_HWINIT},
222 	{BFA_SM(bfa_iocpf_sm_ready), BFA_IOCPF_READY},
223 	{BFA_SM(bfa_iocpf_sm_initfail_sync), BFA_IOCPF_INITFAIL},
224 	{BFA_SM(bfa_iocpf_sm_initfail), BFA_IOCPF_INITFAIL},
225 	{BFA_SM(bfa_iocpf_sm_fail_sync), BFA_IOCPF_FAIL},
226 	{BFA_SM(bfa_iocpf_sm_fail), BFA_IOCPF_FAIL},
227 	{BFA_SM(bfa_iocpf_sm_disabling), BFA_IOCPF_DISABLING},
228 	{BFA_SM(bfa_iocpf_sm_disabling_sync), BFA_IOCPF_DISABLING},
229 	{BFA_SM(bfa_iocpf_sm_disabled), BFA_IOCPF_DISABLED},
230 };
231 
232 /*
233  * IOC State Machine
234  */
235 
236 /*
237  * Beginning state. IOC uninit state.
238  */
239 
240 static void
241 bfa_ioc_sm_uninit_entry(struct bfa_ioc_s *ioc)
242 {
243 }
244 
245 /*
246  * IOC is in uninit state.
247  */
248 static void
249 bfa_ioc_sm_uninit(struct bfa_ioc_s *ioc, enum ioc_event event)
250 {
251 	bfa_trc(ioc, event);
252 
253 	switch (event) {
254 	case IOC_E_RESET:
255 		bfa_fsm_set_state(ioc, bfa_ioc_sm_reset);
256 		break;
257 
258 	default:
259 		bfa_sm_fault(ioc, event);
260 	}
261 }
262 /*
263  * Reset entry actions -- initialize state machine
264  */
265 static void
266 bfa_ioc_sm_reset_entry(struct bfa_ioc_s *ioc)
267 {
268 	bfa_fsm_set_state(&ioc->iocpf, bfa_iocpf_sm_reset);
269 }
270 
271 /*
272  * IOC is in reset state.
273  */
274 static void
275 bfa_ioc_sm_reset(struct bfa_ioc_s *ioc, enum ioc_event event)
276 {
277 	bfa_trc(ioc, event);
278 
279 	switch (event) {
280 	case IOC_E_ENABLE:
281 		bfa_fsm_set_state(ioc, bfa_ioc_sm_enabling);
282 		break;
283 
284 	case IOC_E_DISABLE:
285 		bfa_ioc_disable_comp(ioc);
286 		break;
287 
288 	case IOC_E_DETACH:
289 		bfa_fsm_set_state(ioc, bfa_ioc_sm_uninit);
290 		break;
291 
292 	default:
293 		bfa_sm_fault(ioc, event);
294 	}
295 }
296 
297 
298 static void
299 bfa_ioc_sm_enabling_entry(struct bfa_ioc_s *ioc)
300 {
301 	bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_ENABLE);
302 }
303 
304 /*
305  * Host IOC function is being enabled, awaiting response from firmware.
306  * Semaphore is acquired.
307  */
308 static void
309 bfa_ioc_sm_enabling(struct bfa_ioc_s *ioc, enum ioc_event event)
310 {
311 	bfa_trc(ioc, event);
312 
313 	switch (event) {
314 	case IOC_E_ENABLED:
315 		bfa_fsm_set_state(ioc, bfa_ioc_sm_getattr);
316 		break;
317 
318 	case IOC_E_PFFAILED:
319 		/* !!! fall through !!! */
320 	case IOC_E_HWERROR:
321 		ioc->cbfn->enable_cbfn(ioc->bfa, BFA_STATUS_IOC_FAILURE);
322 		bfa_fsm_set_state(ioc, bfa_ioc_sm_fail);
323 		if (event != IOC_E_PFFAILED)
324 			bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_INITFAIL);
325 		break;
326 
327 	case IOC_E_HWFAILED:
328 		ioc->cbfn->enable_cbfn(ioc->bfa, BFA_STATUS_IOC_FAILURE);
329 		bfa_fsm_set_state(ioc, bfa_ioc_sm_hwfail);
330 		break;
331 
332 	case IOC_E_DISABLE:
333 		bfa_fsm_set_state(ioc, bfa_ioc_sm_disabling);
334 		break;
335 
336 	case IOC_E_DETACH:
337 		bfa_fsm_set_state(ioc, bfa_ioc_sm_uninit);
338 		bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_STOP);
339 		break;
340 
341 	case IOC_E_ENABLE:
342 		break;
343 
344 	default:
345 		bfa_sm_fault(ioc, event);
346 	}
347 }
348 
349 
350 static void
351 bfa_ioc_sm_getattr_entry(struct bfa_ioc_s *ioc)
352 {
353 	bfa_ioc_timer_start(ioc);
354 	bfa_ioc_send_getattr(ioc);
355 }
356 
357 /*
358  * IOC configuration in progress. Timer is active.
359  */
360 static void
361 bfa_ioc_sm_getattr(struct bfa_ioc_s *ioc, enum ioc_event event)
362 {
363 	bfa_trc(ioc, event);
364 
365 	switch (event) {
366 	case IOC_E_FWRSP_GETATTR:
367 		bfa_ioc_timer_stop(ioc);
368 		bfa_fsm_set_state(ioc, bfa_ioc_sm_op);
369 		break;
370 
371 	case IOC_E_PFFAILED:
372 	case IOC_E_HWERROR:
373 		bfa_ioc_timer_stop(ioc);
374 		/* !!! fall through !!! */
375 	case IOC_E_TIMEOUT:
376 		ioc->cbfn->enable_cbfn(ioc->bfa, BFA_STATUS_IOC_FAILURE);
377 		bfa_fsm_set_state(ioc, bfa_ioc_sm_fail);
378 		if (event != IOC_E_PFFAILED)
379 			bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_GETATTRFAIL);
380 		break;
381 
382 	case IOC_E_DISABLE:
383 		bfa_ioc_timer_stop(ioc);
384 		bfa_fsm_set_state(ioc, bfa_ioc_sm_disabling);
385 		break;
386 
387 	case IOC_E_ENABLE:
388 		break;
389 
390 	default:
391 		bfa_sm_fault(ioc, event);
392 	}
393 }
394 
395 static void
396 bfa_ioc_sm_op_entry(struct bfa_ioc_s *ioc)
397 {
398 	struct bfad_s *bfad = (struct bfad_s *)ioc->bfa->bfad;
399 
400 	ioc->cbfn->enable_cbfn(ioc->bfa, BFA_STATUS_OK);
401 	bfa_ioc_event_notify(ioc, BFA_IOC_E_ENABLED);
402 	bfa_ioc_hb_monitor(ioc);
403 	BFA_LOG(KERN_INFO, bfad, bfa_log_level, "IOC enabled\n");
404 	bfa_ioc_aen_post(ioc, BFA_IOC_AEN_ENABLE);
405 }
406 
407 static void
408 bfa_ioc_sm_op(struct bfa_ioc_s *ioc, enum ioc_event event)
409 {
410 	bfa_trc(ioc, event);
411 
412 	switch (event) {
413 	case IOC_E_ENABLE:
414 		break;
415 
416 	case IOC_E_DISABLE:
417 		bfa_hb_timer_stop(ioc);
418 		bfa_fsm_set_state(ioc, bfa_ioc_sm_disabling);
419 		break;
420 
421 	case IOC_E_PFFAILED:
422 	case IOC_E_HWERROR:
423 		bfa_hb_timer_stop(ioc);
424 		/* !!! fall through !!! */
425 	case IOC_E_HBFAIL:
426 		if (ioc->iocpf.auto_recover)
427 			bfa_fsm_set_state(ioc, bfa_ioc_sm_fail_retry);
428 		else
429 			bfa_fsm_set_state(ioc, bfa_ioc_sm_fail);
430 
431 		bfa_ioc_fail_notify(ioc);
432 
433 		if (event != IOC_E_PFFAILED)
434 			bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_FAIL);
435 		break;
436 
437 	default:
438 		bfa_sm_fault(ioc, event);
439 	}
440 }
441 
442 
443 static void
444 bfa_ioc_sm_disabling_entry(struct bfa_ioc_s *ioc)
445 {
446 	struct bfad_s *bfad = (struct bfad_s *)ioc->bfa->bfad;
447 	bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_DISABLE);
448 	BFA_LOG(KERN_INFO, bfad, bfa_log_level, "IOC disabled\n");
449 	bfa_ioc_aen_post(ioc, BFA_IOC_AEN_DISABLE);
450 }
451 
452 /*
453  * IOC is being disabled
454  */
455 static void
456 bfa_ioc_sm_disabling(struct bfa_ioc_s *ioc, enum ioc_event event)
457 {
458 	bfa_trc(ioc, event);
459 
460 	switch (event) {
461 	case IOC_E_DISABLED:
462 		bfa_fsm_set_state(ioc, bfa_ioc_sm_disabled);
463 		break;
464 
465 	case IOC_E_HWERROR:
466 		/*
467 		 * No state change.  Will move to disabled state
468 		 * after iocpf sm completes failure processing and
469 		 * moves to disabled state.
470 		 */
471 		bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_FAIL);
472 		break;
473 
474 	case IOC_E_HWFAILED:
475 		bfa_fsm_set_state(ioc, bfa_ioc_sm_hwfail);
476 		bfa_ioc_disable_comp(ioc);
477 		break;
478 
479 	default:
480 		bfa_sm_fault(ioc, event);
481 	}
482 }
483 
484 /*
485  * IOC disable completion entry.
486  */
487 static void
488 bfa_ioc_sm_disabled_entry(struct bfa_ioc_s *ioc)
489 {
490 	bfa_ioc_disable_comp(ioc);
491 }
492 
493 static void
494 bfa_ioc_sm_disabled(struct bfa_ioc_s *ioc, enum ioc_event event)
495 {
496 	bfa_trc(ioc, event);
497 
498 	switch (event) {
499 	case IOC_E_ENABLE:
500 		bfa_fsm_set_state(ioc, bfa_ioc_sm_enabling);
501 		break;
502 
503 	case IOC_E_DISABLE:
504 		ioc->cbfn->disable_cbfn(ioc->bfa);
505 		break;
506 
507 	case IOC_E_DETACH:
508 		bfa_fsm_set_state(ioc, bfa_ioc_sm_uninit);
509 		bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_STOP);
510 		break;
511 
512 	default:
513 		bfa_sm_fault(ioc, event);
514 	}
515 }
516 
517 
518 static void
519 bfa_ioc_sm_fail_retry_entry(struct bfa_ioc_s *ioc)
520 {
521 	bfa_trc(ioc, 0);
522 }
523 
524 /*
525  * Hardware initialization retry.
526  */
527 static void
528 bfa_ioc_sm_fail_retry(struct bfa_ioc_s *ioc, enum ioc_event event)
529 {
530 	bfa_trc(ioc, event);
531 
532 	switch (event) {
533 	case IOC_E_ENABLED:
534 		bfa_fsm_set_state(ioc, bfa_ioc_sm_getattr);
535 		break;
536 
537 	case IOC_E_PFFAILED:
538 	case IOC_E_HWERROR:
539 		/*
540 		 * Initialization retry failed.
541 		 */
542 		ioc->cbfn->enable_cbfn(ioc->bfa, BFA_STATUS_IOC_FAILURE);
543 		bfa_fsm_set_state(ioc, bfa_ioc_sm_fail);
544 		if (event != IOC_E_PFFAILED)
545 			bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_INITFAIL);
546 		break;
547 
548 	case IOC_E_HWFAILED:
549 		ioc->cbfn->enable_cbfn(ioc->bfa, BFA_STATUS_IOC_FAILURE);
550 		bfa_fsm_set_state(ioc, bfa_ioc_sm_hwfail);
551 		break;
552 
553 	case IOC_E_ENABLE:
554 		break;
555 
556 	case IOC_E_DISABLE:
557 		bfa_fsm_set_state(ioc, bfa_ioc_sm_disabling);
558 		break;
559 
560 	case IOC_E_DETACH:
561 		bfa_fsm_set_state(ioc, bfa_ioc_sm_uninit);
562 		bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_STOP);
563 		break;
564 
565 	default:
566 		bfa_sm_fault(ioc, event);
567 	}
568 }
569 
570 
571 static void
572 bfa_ioc_sm_fail_entry(struct bfa_ioc_s *ioc)
573 {
574 	bfa_trc(ioc, 0);
575 }
576 
577 /*
578  * IOC failure.
579  */
580 static void
581 bfa_ioc_sm_fail(struct bfa_ioc_s *ioc, enum ioc_event event)
582 {
583 	bfa_trc(ioc, event);
584 
585 	switch (event) {
586 
587 	case IOC_E_ENABLE:
588 		ioc->cbfn->enable_cbfn(ioc->bfa, BFA_STATUS_IOC_FAILURE);
589 		break;
590 
591 	case IOC_E_DISABLE:
592 		bfa_fsm_set_state(ioc, bfa_ioc_sm_disabling);
593 		break;
594 
595 	case IOC_E_DETACH:
596 		bfa_fsm_set_state(ioc, bfa_ioc_sm_uninit);
597 		bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_STOP);
598 		break;
599 
600 	case IOC_E_HWERROR:
601 	case IOC_E_HWFAILED:
602 		/*
603 		 * HB failure / HW error notification, ignore.
604 		 */
605 		break;
606 	default:
607 		bfa_sm_fault(ioc, event);
608 	}
609 }
610 
611 static void
612 bfa_ioc_sm_hwfail_entry(struct bfa_ioc_s *ioc)
613 {
614 	bfa_trc(ioc, 0);
615 }
616 
617 static void
618 bfa_ioc_sm_hwfail(struct bfa_ioc_s *ioc, enum ioc_event event)
619 {
620 	bfa_trc(ioc, event);
621 
622 	switch (event) {
623 	case IOC_E_ENABLE:
624 		ioc->cbfn->enable_cbfn(ioc->bfa, BFA_STATUS_IOC_FAILURE);
625 		break;
626 
627 	case IOC_E_DISABLE:
628 		ioc->cbfn->disable_cbfn(ioc->bfa);
629 		break;
630 
631 	case IOC_E_DETACH:
632 		bfa_fsm_set_state(ioc, bfa_ioc_sm_uninit);
633 		break;
634 
635 	case IOC_E_HWERROR:
636 		/* Ignore - already in hwfail state */
637 		break;
638 
639 	default:
640 		bfa_sm_fault(ioc, event);
641 	}
642 }
643 
644 /*
645  * IOCPF State Machine
646  */
647 
648 /*
649  * Reset entry actions -- initialize state machine
650  */
651 static void
652 bfa_iocpf_sm_reset_entry(struct bfa_iocpf_s *iocpf)
653 {
654 	iocpf->fw_mismatch_notified = BFA_FALSE;
655 	iocpf->auto_recover = bfa_auto_recover;
656 }
657 
658 /*
659  * Beginning state. IOC is in reset state.
660  */
661 static void
662 bfa_iocpf_sm_reset(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
663 {
664 	struct bfa_ioc_s *ioc = iocpf->ioc;
665 
666 	bfa_trc(ioc, event);
667 
668 	switch (event) {
669 	case IOCPF_E_ENABLE:
670 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_fwcheck);
671 		break;
672 
673 	case IOCPF_E_STOP:
674 		break;
675 
676 	default:
677 		bfa_sm_fault(ioc, event);
678 	}
679 }
680 
681 /*
682  * Semaphore should be acquired for version check.
683  */
684 static void
685 bfa_iocpf_sm_fwcheck_entry(struct bfa_iocpf_s *iocpf)
686 {
687 	struct bfi_ioc_image_hdr_s	fwhdr;
688 	u32	r32, fwstate, pgnum, pgoff, loff = 0;
689 	int	i;
690 
691 	/*
692 	 * Spin on init semaphore to serialize.
693 	 */
694 	r32 = readl(iocpf->ioc->ioc_regs.ioc_init_sem_reg);
695 	while (r32 & 0x1) {
696 		udelay(20);
697 		r32 = readl(iocpf->ioc->ioc_regs.ioc_init_sem_reg);
698 	}
699 
700 	/* h/w sem init */
701 	fwstate = readl(iocpf->ioc->ioc_regs.ioc_fwstate);
702 	if (fwstate == BFI_IOC_UNINIT) {
703 		writel(1, iocpf->ioc->ioc_regs.ioc_init_sem_reg);
704 		goto sem_get;
705 	}
706 
707 	bfa_ioc_fwver_get(iocpf->ioc, &fwhdr);
708 
709 	if (swab32(fwhdr.exec) == BFI_FWBOOT_TYPE_NORMAL) {
710 		writel(1, iocpf->ioc->ioc_regs.ioc_init_sem_reg);
711 		goto sem_get;
712 	}
713 
714 	/*
715 	 * Clear fwver hdr
716 	 */
717 	pgnum = PSS_SMEM_PGNUM(iocpf->ioc->ioc_regs.smem_pg0, loff);
718 	pgoff = PSS_SMEM_PGOFF(loff);
719 	writel(pgnum, iocpf->ioc->ioc_regs.host_page_num_fn);
720 
721 	for (i = 0; i < sizeof(struct bfi_ioc_image_hdr_s) / sizeof(u32); i++) {
722 		bfa_mem_write(iocpf->ioc->ioc_regs.smem_page_start, loff, 0);
723 		loff += sizeof(u32);
724 	}
725 
726 	bfa_trc(iocpf->ioc, fwstate);
727 	bfa_trc(iocpf->ioc, swab32(fwhdr.exec));
728 	writel(BFI_IOC_UNINIT, iocpf->ioc->ioc_regs.ioc_fwstate);
729 	writel(BFI_IOC_UNINIT, iocpf->ioc->ioc_regs.alt_ioc_fwstate);
730 
731 	/*
732 	 * Unlock the hw semaphore. Should be here only once per boot.
733 	 */
734 	bfa_ioc_ownership_reset(iocpf->ioc);
735 
736 	/*
737 	 * unlock init semaphore.
738 	 */
739 	writel(1, iocpf->ioc->ioc_regs.ioc_init_sem_reg);
740 
741 sem_get:
742 	bfa_ioc_hw_sem_get(iocpf->ioc);
743 }
744 
745 /*
746  * Awaiting h/w semaphore to continue with version check.
747  */
748 static void
749 bfa_iocpf_sm_fwcheck(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
750 {
751 	struct bfa_ioc_s *ioc = iocpf->ioc;
752 
753 	bfa_trc(ioc, event);
754 
755 	switch (event) {
756 	case IOCPF_E_SEMLOCKED:
757 		if (bfa_ioc_firmware_lock(ioc)) {
758 			if (bfa_ioc_sync_start(ioc)) {
759 				bfa_ioc_sync_join(ioc);
760 				bfa_fsm_set_state(iocpf, bfa_iocpf_sm_hwinit);
761 			} else {
762 				bfa_ioc_firmware_unlock(ioc);
763 				writel(1, ioc->ioc_regs.ioc_sem_reg);
764 				bfa_sem_timer_start(ioc);
765 			}
766 		} else {
767 			writel(1, ioc->ioc_regs.ioc_sem_reg);
768 			bfa_fsm_set_state(iocpf, bfa_iocpf_sm_mismatch);
769 		}
770 		break;
771 
772 	case IOCPF_E_SEM_ERROR:
773 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_fail);
774 		bfa_fsm_send_event(ioc, IOC_E_HWFAILED);
775 		break;
776 
777 	case IOCPF_E_DISABLE:
778 		bfa_sem_timer_stop(ioc);
779 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_reset);
780 		bfa_fsm_send_event(ioc, IOC_E_DISABLED);
781 		break;
782 
783 	case IOCPF_E_STOP:
784 		bfa_sem_timer_stop(ioc);
785 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_reset);
786 		break;
787 
788 	default:
789 		bfa_sm_fault(ioc, event);
790 	}
791 }
792 
793 /*
794  * Notify enable completion callback.
795  */
796 static void
797 bfa_iocpf_sm_mismatch_entry(struct bfa_iocpf_s *iocpf)
798 {
799 	/*
800 	 * Call only the first time sm enters fwmismatch state.
801 	 */
802 	if (iocpf->fw_mismatch_notified == BFA_FALSE)
803 		bfa_ioc_pf_fwmismatch(iocpf->ioc);
804 
805 	iocpf->fw_mismatch_notified = BFA_TRUE;
806 	bfa_iocpf_timer_start(iocpf->ioc);
807 }
808 
809 /*
810  * Awaiting firmware version match.
811  */
812 static void
813 bfa_iocpf_sm_mismatch(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
814 {
815 	struct bfa_ioc_s *ioc = iocpf->ioc;
816 
817 	bfa_trc(ioc, event);
818 
819 	switch (event) {
820 	case IOCPF_E_TIMEOUT:
821 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_fwcheck);
822 		break;
823 
824 	case IOCPF_E_DISABLE:
825 		bfa_iocpf_timer_stop(ioc);
826 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_reset);
827 		bfa_fsm_send_event(ioc, IOC_E_DISABLED);
828 		break;
829 
830 	case IOCPF_E_STOP:
831 		bfa_iocpf_timer_stop(ioc);
832 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_reset);
833 		break;
834 
835 	default:
836 		bfa_sm_fault(ioc, event);
837 	}
838 }
839 
840 /*
841  * Request for semaphore.
842  */
843 static void
844 bfa_iocpf_sm_semwait_entry(struct bfa_iocpf_s *iocpf)
845 {
846 	bfa_ioc_hw_sem_get(iocpf->ioc);
847 }
848 
849 /*
850  * Awaiting semaphore for h/w initialzation.
851  */
852 static void
853 bfa_iocpf_sm_semwait(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
854 {
855 	struct bfa_ioc_s *ioc = iocpf->ioc;
856 
857 	bfa_trc(ioc, event);
858 
859 	switch (event) {
860 	case IOCPF_E_SEMLOCKED:
861 		if (bfa_ioc_sync_complete(ioc)) {
862 			bfa_ioc_sync_join(ioc);
863 			bfa_fsm_set_state(iocpf, bfa_iocpf_sm_hwinit);
864 		} else {
865 			writel(1, ioc->ioc_regs.ioc_sem_reg);
866 			bfa_sem_timer_start(ioc);
867 		}
868 		break;
869 
870 	case IOCPF_E_SEM_ERROR:
871 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_fail);
872 		bfa_fsm_send_event(ioc, IOC_E_HWFAILED);
873 		break;
874 
875 	case IOCPF_E_DISABLE:
876 		bfa_sem_timer_stop(ioc);
877 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_disabling_sync);
878 		break;
879 
880 	default:
881 		bfa_sm_fault(ioc, event);
882 	}
883 }
884 
885 static void
886 bfa_iocpf_sm_hwinit_entry(struct bfa_iocpf_s *iocpf)
887 {
888 	iocpf->poll_time = 0;
889 	bfa_ioc_hwinit(iocpf->ioc, BFA_FALSE);
890 }
891 
892 /*
893  * Hardware is being initialized. Interrupts are enabled.
894  * Holding hardware semaphore lock.
895  */
896 static void
897 bfa_iocpf_sm_hwinit(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
898 {
899 	struct bfa_ioc_s *ioc = iocpf->ioc;
900 
901 	bfa_trc(ioc, event);
902 
903 	switch (event) {
904 	case IOCPF_E_FWREADY:
905 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_enabling);
906 		break;
907 
908 	case IOCPF_E_TIMEOUT:
909 		writel(1, ioc->ioc_regs.ioc_sem_reg);
910 		bfa_fsm_send_event(ioc, IOC_E_PFFAILED);
911 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_initfail_sync);
912 		break;
913 
914 	case IOCPF_E_DISABLE:
915 		bfa_iocpf_timer_stop(ioc);
916 		bfa_ioc_sync_leave(ioc);
917 		writel(1, ioc->ioc_regs.ioc_sem_reg);
918 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_disabled);
919 		break;
920 
921 	default:
922 		bfa_sm_fault(ioc, event);
923 	}
924 }
925 
926 static void
927 bfa_iocpf_sm_enabling_entry(struct bfa_iocpf_s *iocpf)
928 {
929 	bfa_iocpf_timer_start(iocpf->ioc);
930 	/*
931 	 * Enable Interrupts before sending fw IOC ENABLE cmd.
932 	 */
933 	iocpf->ioc->cbfn->reset_cbfn(iocpf->ioc->bfa);
934 	bfa_ioc_send_enable(iocpf->ioc);
935 }
936 
937 /*
938  * Host IOC function is being enabled, awaiting response from firmware.
939  * Semaphore is acquired.
940  */
941 static void
942 bfa_iocpf_sm_enabling(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
943 {
944 	struct bfa_ioc_s *ioc = iocpf->ioc;
945 
946 	bfa_trc(ioc, event);
947 
948 	switch (event) {
949 	case IOCPF_E_FWRSP_ENABLE:
950 		bfa_iocpf_timer_stop(ioc);
951 		writel(1, ioc->ioc_regs.ioc_sem_reg);
952 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_ready);
953 		break;
954 
955 	case IOCPF_E_INITFAIL:
956 		bfa_iocpf_timer_stop(ioc);
957 		/*
958 		 * !!! fall through !!!
959 		 */
960 
961 	case IOCPF_E_TIMEOUT:
962 		writel(1, ioc->ioc_regs.ioc_sem_reg);
963 		if (event == IOCPF_E_TIMEOUT)
964 			bfa_fsm_send_event(ioc, IOC_E_PFFAILED);
965 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_initfail_sync);
966 		break;
967 
968 	case IOCPF_E_DISABLE:
969 		bfa_iocpf_timer_stop(ioc);
970 		writel(1, ioc->ioc_regs.ioc_sem_reg);
971 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_disabling);
972 		break;
973 
974 	default:
975 		bfa_sm_fault(ioc, event);
976 	}
977 }
978 
979 static void
980 bfa_iocpf_sm_ready_entry(struct bfa_iocpf_s *iocpf)
981 {
982 	bfa_fsm_send_event(iocpf->ioc, IOC_E_ENABLED);
983 }
984 
985 static void
986 bfa_iocpf_sm_ready(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
987 {
988 	struct bfa_ioc_s *ioc = iocpf->ioc;
989 
990 	bfa_trc(ioc, event);
991 
992 	switch (event) {
993 	case IOCPF_E_DISABLE:
994 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_disabling);
995 		break;
996 
997 	case IOCPF_E_GETATTRFAIL:
998 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_initfail_sync);
999 		break;
1000 
1001 	case IOCPF_E_FAIL:
1002 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_fail_sync);
1003 		break;
1004 
1005 	default:
1006 		bfa_sm_fault(ioc, event);
1007 	}
1008 }
1009 
1010 static void
1011 bfa_iocpf_sm_disabling_entry(struct bfa_iocpf_s *iocpf)
1012 {
1013 	bfa_iocpf_timer_start(iocpf->ioc);
1014 	bfa_ioc_send_disable(iocpf->ioc);
1015 }
1016 
1017 /*
1018  * IOC is being disabled
1019  */
1020 static void
1021 bfa_iocpf_sm_disabling(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
1022 {
1023 	struct bfa_ioc_s *ioc = iocpf->ioc;
1024 
1025 	bfa_trc(ioc, event);
1026 
1027 	switch (event) {
1028 	case IOCPF_E_FWRSP_DISABLE:
1029 		bfa_iocpf_timer_stop(ioc);
1030 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_disabling_sync);
1031 		break;
1032 
1033 	case IOCPF_E_FAIL:
1034 		bfa_iocpf_timer_stop(ioc);
1035 		/*
1036 		 * !!! fall through !!!
1037 		 */
1038 
1039 	case IOCPF_E_TIMEOUT:
1040 		writel(BFI_IOC_FAIL, ioc->ioc_regs.ioc_fwstate);
1041 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_disabling_sync);
1042 		break;
1043 
1044 	case IOCPF_E_FWRSP_ENABLE:
1045 		break;
1046 
1047 	default:
1048 		bfa_sm_fault(ioc, event);
1049 	}
1050 }
1051 
1052 static void
1053 bfa_iocpf_sm_disabling_sync_entry(struct bfa_iocpf_s *iocpf)
1054 {
1055 	bfa_ioc_hw_sem_get(iocpf->ioc);
1056 }
1057 
1058 /*
1059  * IOC hb ack request is being removed.
1060  */
1061 static void
1062 bfa_iocpf_sm_disabling_sync(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
1063 {
1064 	struct bfa_ioc_s *ioc = iocpf->ioc;
1065 
1066 	bfa_trc(ioc, event);
1067 
1068 	switch (event) {
1069 	case IOCPF_E_SEMLOCKED:
1070 		bfa_ioc_sync_leave(ioc);
1071 		writel(1, ioc->ioc_regs.ioc_sem_reg);
1072 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_disabled);
1073 		break;
1074 
1075 	case IOCPF_E_SEM_ERROR:
1076 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_fail);
1077 		bfa_fsm_send_event(ioc, IOC_E_HWFAILED);
1078 		break;
1079 
1080 	case IOCPF_E_FAIL:
1081 		break;
1082 
1083 	default:
1084 		bfa_sm_fault(ioc, event);
1085 	}
1086 }
1087 
1088 /*
1089  * IOC disable completion entry.
1090  */
1091 static void
1092 bfa_iocpf_sm_disabled_entry(struct bfa_iocpf_s *iocpf)
1093 {
1094 	bfa_ioc_mbox_flush(iocpf->ioc);
1095 	bfa_fsm_send_event(iocpf->ioc, IOC_E_DISABLED);
1096 }
1097 
1098 static void
1099 bfa_iocpf_sm_disabled(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
1100 {
1101 	struct bfa_ioc_s *ioc = iocpf->ioc;
1102 
1103 	bfa_trc(ioc, event);
1104 
1105 	switch (event) {
1106 	case IOCPF_E_ENABLE:
1107 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_semwait);
1108 		break;
1109 
1110 	case IOCPF_E_STOP:
1111 		bfa_ioc_firmware_unlock(ioc);
1112 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_reset);
1113 		break;
1114 
1115 	default:
1116 		bfa_sm_fault(ioc, event);
1117 	}
1118 }
1119 
1120 static void
1121 bfa_iocpf_sm_initfail_sync_entry(struct bfa_iocpf_s *iocpf)
1122 {
1123 	bfa_ioc_debug_save_ftrc(iocpf->ioc);
1124 	bfa_ioc_hw_sem_get(iocpf->ioc);
1125 }
1126 
1127 /*
1128  * Hardware initialization failed.
1129  */
1130 static void
1131 bfa_iocpf_sm_initfail_sync(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
1132 {
1133 	struct bfa_ioc_s *ioc = iocpf->ioc;
1134 
1135 	bfa_trc(ioc, event);
1136 
1137 	switch (event) {
1138 	case IOCPF_E_SEMLOCKED:
1139 		bfa_ioc_notify_fail(ioc);
1140 		bfa_ioc_sync_leave(ioc);
1141 		writel(BFI_IOC_FAIL, ioc->ioc_regs.ioc_fwstate);
1142 		writel(1, ioc->ioc_regs.ioc_sem_reg);
1143 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_initfail);
1144 		break;
1145 
1146 	case IOCPF_E_SEM_ERROR:
1147 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_fail);
1148 		bfa_fsm_send_event(ioc, IOC_E_HWFAILED);
1149 		break;
1150 
1151 	case IOCPF_E_DISABLE:
1152 		bfa_sem_timer_stop(ioc);
1153 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_disabling_sync);
1154 		break;
1155 
1156 	case IOCPF_E_STOP:
1157 		bfa_sem_timer_stop(ioc);
1158 		bfa_ioc_firmware_unlock(ioc);
1159 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_reset);
1160 		break;
1161 
1162 	case IOCPF_E_FAIL:
1163 		break;
1164 
1165 	default:
1166 		bfa_sm_fault(ioc, event);
1167 	}
1168 }
1169 
1170 static void
1171 bfa_iocpf_sm_initfail_entry(struct bfa_iocpf_s *iocpf)
1172 {
1173 	bfa_trc(iocpf->ioc, 0);
1174 }
1175 
1176 /*
1177  * Hardware initialization failed.
1178  */
1179 static void
1180 bfa_iocpf_sm_initfail(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
1181 {
1182 	struct bfa_ioc_s *ioc = iocpf->ioc;
1183 
1184 	bfa_trc(ioc, event);
1185 
1186 	switch (event) {
1187 	case IOCPF_E_DISABLE:
1188 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_disabled);
1189 		break;
1190 
1191 	case IOCPF_E_STOP:
1192 		bfa_ioc_firmware_unlock(ioc);
1193 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_reset);
1194 		break;
1195 
1196 	default:
1197 		bfa_sm_fault(ioc, event);
1198 	}
1199 }
1200 
1201 static void
1202 bfa_iocpf_sm_fail_sync_entry(struct bfa_iocpf_s *iocpf)
1203 {
1204 	/*
1205 	 * Mark IOC as failed in hardware and stop firmware.
1206 	 */
1207 	bfa_ioc_lpu_stop(iocpf->ioc);
1208 
1209 	/*
1210 	 * Flush any queued up mailbox requests.
1211 	 */
1212 	bfa_ioc_mbox_flush(iocpf->ioc);
1213 
1214 	bfa_ioc_hw_sem_get(iocpf->ioc);
1215 }
1216 
1217 static void
1218 bfa_iocpf_sm_fail_sync(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
1219 {
1220 	struct bfa_ioc_s *ioc = iocpf->ioc;
1221 
1222 	bfa_trc(ioc, event);
1223 
1224 	switch (event) {
1225 	case IOCPF_E_SEMLOCKED:
1226 		bfa_ioc_sync_ack(ioc);
1227 		bfa_ioc_notify_fail(ioc);
1228 		if (!iocpf->auto_recover) {
1229 			bfa_ioc_sync_leave(ioc);
1230 			writel(BFI_IOC_FAIL, ioc->ioc_regs.ioc_fwstate);
1231 			writel(1, ioc->ioc_regs.ioc_sem_reg);
1232 			bfa_fsm_set_state(iocpf, bfa_iocpf_sm_fail);
1233 		} else {
1234 			if (bfa_ioc_sync_complete(ioc))
1235 				bfa_fsm_set_state(iocpf, bfa_iocpf_sm_hwinit);
1236 			else {
1237 				writel(1, ioc->ioc_regs.ioc_sem_reg);
1238 				bfa_fsm_set_state(iocpf, bfa_iocpf_sm_semwait);
1239 			}
1240 		}
1241 		break;
1242 
1243 	case IOCPF_E_SEM_ERROR:
1244 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_fail);
1245 		bfa_fsm_send_event(ioc, IOC_E_HWFAILED);
1246 		break;
1247 
1248 	case IOCPF_E_DISABLE:
1249 		bfa_sem_timer_stop(ioc);
1250 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_disabling_sync);
1251 		break;
1252 
1253 	case IOCPF_E_FAIL:
1254 		break;
1255 
1256 	default:
1257 		bfa_sm_fault(ioc, event);
1258 	}
1259 }
1260 
1261 static void
1262 bfa_iocpf_sm_fail_entry(struct bfa_iocpf_s *iocpf)
1263 {
1264 	bfa_trc(iocpf->ioc, 0);
1265 }
1266 
1267 /*
1268  * IOC is in failed state.
1269  */
1270 static void
1271 bfa_iocpf_sm_fail(struct bfa_iocpf_s *iocpf, enum iocpf_event event)
1272 {
1273 	struct bfa_ioc_s *ioc = iocpf->ioc;
1274 
1275 	bfa_trc(ioc, event);
1276 
1277 	switch (event) {
1278 	case IOCPF_E_DISABLE:
1279 		bfa_fsm_set_state(iocpf, bfa_iocpf_sm_disabled);
1280 		break;
1281 
1282 	default:
1283 		bfa_sm_fault(ioc, event);
1284 	}
1285 }
1286 
1287 /*
1288  *  BFA IOC private functions
1289  */
1290 
1291 /*
1292  * Notify common modules registered for notification.
1293  */
1294 static void
1295 bfa_ioc_event_notify(struct bfa_ioc_s *ioc, enum bfa_ioc_event_e event)
1296 {
1297 	struct bfa_ioc_notify_s	*notify;
1298 	struct list_head	*qe;
1299 
1300 	list_for_each(qe, &ioc->notify_q) {
1301 		notify = (struct bfa_ioc_notify_s *)qe;
1302 		notify->cbfn(notify->cbarg, event);
1303 	}
1304 }
1305 
1306 static void
1307 bfa_ioc_disable_comp(struct bfa_ioc_s *ioc)
1308 {
1309 	ioc->cbfn->disable_cbfn(ioc->bfa);
1310 	bfa_ioc_event_notify(ioc, BFA_IOC_E_DISABLED);
1311 }
1312 
1313 bfa_boolean_t
1314 bfa_ioc_sem_get(void __iomem *sem_reg)
1315 {
1316 	u32 r32;
1317 	int cnt = 0;
1318 #define BFA_SEM_SPINCNT	3000
1319 
1320 	r32 = readl(sem_reg);
1321 
1322 	while ((r32 & 1) && (cnt < BFA_SEM_SPINCNT)) {
1323 		cnt++;
1324 		udelay(2);
1325 		r32 = readl(sem_reg);
1326 	}
1327 
1328 	if (!(r32 & 1))
1329 		return BFA_TRUE;
1330 
1331 	return BFA_FALSE;
1332 }
1333 
1334 static void
1335 bfa_ioc_hw_sem_get(struct bfa_ioc_s *ioc)
1336 {
1337 	u32	r32;
1338 
1339 	/*
1340 	 * First read to the semaphore register will return 0, subsequent reads
1341 	 * will return 1. Semaphore is released by writing 1 to the register
1342 	 */
1343 	r32 = readl(ioc->ioc_regs.ioc_sem_reg);
1344 	if (r32 == ~0) {
1345 		WARN_ON(r32 == ~0);
1346 		bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_SEM_ERROR);
1347 		return;
1348 	}
1349 	if (!(r32 & 1)) {
1350 		bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_SEMLOCKED);
1351 		return;
1352 	}
1353 
1354 	bfa_sem_timer_start(ioc);
1355 }
1356 
1357 /*
1358  * Initialize LPU local memory (aka secondary memory / SRAM)
1359  */
1360 static void
1361 bfa_ioc_lmem_init(struct bfa_ioc_s *ioc)
1362 {
1363 	u32	pss_ctl;
1364 	int		i;
1365 #define PSS_LMEM_INIT_TIME  10000
1366 
1367 	pss_ctl = readl(ioc->ioc_regs.pss_ctl_reg);
1368 	pss_ctl &= ~__PSS_LMEM_RESET;
1369 	pss_ctl |= __PSS_LMEM_INIT_EN;
1370 
1371 	/*
1372 	 * i2c workaround 12.5khz clock
1373 	 */
1374 	pss_ctl |= __PSS_I2C_CLK_DIV(3UL);
1375 	writel(pss_ctl, ioc->ioc_regs.pss_ctl_reg);
1376 
1377 	/*
1378 	 * wait for memory initialization to be complete
1379 	 */
1380 	i = 0;
1381 	do {
1382 		pss_ctl = readl(ioc->ioc_regs.pss_ctl_reg);
1383 		i++;
1384 	} while (!(pss_ctl & __PSS_LMEM_INIT_DONE) && (i < PSS_LMEM_INIT_TIME));
1385 
1386 	/*
1387 	 * If memory initialization is not successful, IOC timeout will catch
1388 	 * such failures.
1389 	 */
1390 	WARN_ON(!(pss_ctl & __PSS_LMEM_INIT_DONE));
1391 	bfa_trc(ioc, pss_ctl);
1392 
1393 	pss_ctl &= ~(__PSS_LMEM_INIT_DONE | __PSS_LMEM_INIT_EN);
1394 	writel(pss_ctl, ioc->ioc_regs.pss_ctl_reg);
1395 }
1396 
1397 static void
1398 bfa_ioc_lpu_start(struct bfa_ioc_s *ioc)
1399 {
1400 	u32	pss_ctl;
1401 
1402 	/*
1403 	 * Take processor out of reset.
1404 	 */
1405 	pss_ctl = readl(ioc->ioc_regs.pss_ctl_reg);
1406 	pss_ctl &= ~__PSS_LPU0_RESET;
1407 
1408 	writel(pss_ctl, ioc->ioc_regs.pss_ctl_reg);
1409 }
1410 
1411 static void
1412 bfa_ioc_lpu_stop(struct bfa_ioc_s *ioc)
1413 {
1414 	u32	pss_ctl;
1415 
1416 	/*
1417 	 * Put processors in reset.
1418 	 */
1419 	pss_ctl = readl(ioc->ioc_regs.pss_ctl_reg);
1420 	pss_ctl |= (__PSS_LPU0_RESET | __PSS_LPU1_RESET);
1421 
1422 	writel(pss_ctl, ioc->ioc_regs.pss_ctl_reg);
1423 }
1424 
1425 /*
1426  * Get driver and firmware versions.
1427  */
1428 void
1429 bfa_ioc_fwver_get(struct bfa_ioc_s *ioc, struct bfi_ioc_image_hdr_s *fwhdr)
1430 {
1431 	u32	pgnum, pgoff;
1432 	u32	loff = 0;
1433 	int		i;
1434 	u32	*fwsig = (u32 *) fwhdr;
1435 
1436 	pgnum = PSS_SMEM_PGNUM(ioc->ioc_regs.smem_pg0, loff);
1437 	pgoff = PSS_SMEM_PGOFF(loff);
1438 	writel(pgnum, ioc->ioc_regs.host_page_num_fn);
1439 
1440 	for (i = 0; i < (sizeof(struct bfi_ioc_image_hdr_s) / sizeof(u32));
1441 	     i++) {
1442 		fwsig[i] =
1443 			bfa_mem_read(ioc->ioc_regs.smem_page_start, loff);
1444 		loff += sizeof(u32);
1445 	}
1446 }
1447 
1448 /*
1449  * Returns TRUE if same.
1450  */
1451 bfa_boolean_t
1452 bfa_ioc_fwver_cmp(struct bfa_ioc_s *ioc, struct bfi_ioc_image_hdr_s *fwhdr)
1453 {
1454 	struct bfi_ioc_image_hdr_s *drv_fwhdr;
1455 	int i;
1456 
1457 	drv_fwhdr = (struct bfi_ioc_image_hdr_s *)
1458 		bfa_cb_image_get_chunk(bfa_ioc_asic_gen(ioc), 0);
1459 
1460 	for (i = 0; i < BFI_IOC_MD5SUM_SZ; i++) {
1461 		if (fwhdr->md5sum[i] != cpu_to_le32(drv_fwhdr->md5sum[i])) {
1462 			bfa_trc(ioc, i);
1463 			bfa_trc(ioc, fwhdr->md5sum[i]);
1464 			bfa_trc(ioc, drv_fwhdr->md5sum[i]);
1465 			return BFA_FALSE;
1466 		}
1467 	}
1468 
1469 	bfa_trc(ioc, fwhdr->md5sum[0]);
1470 	return BFA_TRUE;
1471 }
1472 
1473 /*
1474  * Return true if current running version is valid. Firmware signature and
1475  * execution context (driver/bios) must match.
1476  */
1477 static bfa_boolean_t
1478 bfa_ioc_fwver_valid(struct bfa_ioc_s *ioc, u32 boot_env)
1479 {
1480 	struct bfi_ioc_image_hdr_s fwhdr, *drv_fwhdr;
1481 
1482 	bfa_ioc_fwver_get(ioc, &fwhdr);
1483 	drv_fwhdr = (struct bfi_ioc_image_hdr_s *)
1484 		bfa_cb_image_get_chunk(bfa_ioc_asic_gen(ioc), 0);
1485 
1486 	if (fwhdr.signature != cpu_to_le32(drv_fwhdr->signature)) {
1487 		bfa_trc(ioc, fwhdr.signature);
1488 		bfa_trc(ioc, drv_fwhdr->signature);
1489 		return BFA_FALSE;
1490 	}
1491 
1492 	if (swab32(fwhdr.bootenv) != boot_env) {
1493 		bfa_trc(ioc, fwhdr.bootenv);
1494 		bfa_trc(ioc, boot_env);
1495 		return BFA_FALSE;
1496 	}
1497 
1498 	return bfa_ioc_fwver_cmp(ioc, &fwhdr);
1499 }
1500 
1501 /*
1502  * Conditionally flush any pending message from firmware at start.
1503  */
1504 static void
1505 bfa_ioc_msgflush(struct bfa_ioc_s *ioc)
1506 {
1507 	u32	r32;
1508 
1509 	r32 = readl(ioc->ioc_regs.lpu_mbox_cmd);
1510 	if (r32)
1511 		writel(1, ioc->ioc_regs.lpu_mbox_cmd);
1512 }
1513 
1514 static void
1515 bfa_ioc_hwinit(struct bfa_ioc_s *ioc, bfa_boolean_t force)
1516 {
1517 	enum bfi_ioc_state ioc_fwstate;
1518 	bfa_boolean_t fwvalid;
1519 	u32 boot_type;
1520 	u32 boot_env;
1521 
1522 	ioc_fwstate = readl(ioc->ioc_regs.ioc_fwstate);
1523 
1524 	if (force)
1525 		ioc_fwstate = BFI_IOC_UNINIT;
1526 
1527 	bfa_trc(ioc, ioc_fwstate);
1528 
1529 	boot_type = BFI_FWBOOT_TYPE_NORMAL;
1530 	boot_env = BFI_FWBOOT_ENV_OS;
1531 
1532 	/*
1533 	 * check if firmware is valid
1534 	 */
1535 	fwvalid = (ioc_fwstate == BFI_IOC_UNINIT) ?
1536 		BFA_FALSE : bfa_ioc_fwver_valid(ioc, boot_env);
1537 
1538 	if (!fwvalid) {
1539 		bfa_ioc_boot(ioc, boot_type, boot_env);
1540 		bfa_ioc_poll_fwinit(ioc);
1541 		return;
1542 	}
1543 
1544 	/*
1545 	 * If hardware initialization is in progress (initialized by other IOC),
1546 	 * just wait for an initialization completion interrupt.
1547 	 */
1548 	if (ioc_fwstate == BFI_IOC_INITING) {
1549 		bfa_ioc_poll_fwinit(ioc);
1550 		return;
1551 	}
1552 
1553 	/*
1554 	 * If IOC function is disabled and firmware version is same,
1555 	 * just re-enable IOC.
1556 	 *
1557 	 * If option rom, IOC must not be in operational state. With
1558 	 * convergence, IOC will be in operational state when 2nd driver
1559 	 * is loaded.
1560 	 */
1561 	if (ioc_fwstate == BFI_IOC_DISABLED || ioc_fwstate == BFI_IOC_OP) {
1562 
1563 		/*
1564 		 * When using MSI-X any pending firmware ready event should
1565 		 * be flushed. Otherwise MSI-X interrupts are not delivered.
1566 		 */
1567 		bfa_ioc_msgflush(ioc);
1568 		bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_FWREADY);
1569 		return;
1570 	}
1571 
1572 	/*
1573 	 * Initialize the h/w for any other states.
1574 	 */
1575 	bfa_ioc_boot(ioc, boot_type, boot_env);
1576 	bfa_ioc_poll_fwinit(ioc);
1577 }
1578 
1579 static void
1580 bfa_ioc_timeout(void *ioc_arg)
1581 {
1582 	struct bfa_ioc_s  *ioc = (struct bfa_ioc_s *) ioc_arg;
1583 
1584 	bfa_trc(ioc, 0);
1585 	bfa_fsm_send_event(ioc, IOC_E_TIMEOUT);
1586 }
1587 
1588 void
1589 bfa_ioc_mbox_send(struct bfa_ioc_s *ioc, void *ioc_msg, int len)
1590 {
1591 	u32 *msgp = (u32 *) ioc_msg;
1592 	u32 i;
1593 
1594 	bfa_trc(ioc, msgp[0]);
1595 	bfa_trc(ioc, len);
1596 
1597 	WARN_ON(len > BFI_IOC_MSGLEN_MAX);
1598 
1599 	/*
1600 	 * first write msg to mailbox registers
1601 	 */
1602 	for (i = 0; i < len / sizeof(u32); i++)
1603 		writel(cpu_to_le32(msgp[i]),
1604 			ioc->ioc_regs.hfn_mbox + i * sizeof(u32));
1605 
1606 	for (; i < BFI_IOC_MSGLEN_MAX / sizeof(u32); i++)
1607 		writel(0, ioc->ioc_regs.hfn_mbox + i * sizeof(u32));
1608 
1609 	/*
1610 	 * write 1 to mailbox CMD to trigger LPU event
1611 	 */
1612 	writel(1, ioc->ioc_regs.hfn_mbox_cmd);
1613 	(void) readl(ioc->ioc_regs.hfn_mbox_cmd);
1614 }
1615 
1616 static void
1617 bfa_ioc_send_enable(struct bfa_ioc_s *ioc)
1618 {
1619 	struct bfi_ioc_ctrl_req_s enable_req;
1620 	struct timeval tv;
1621 
1622 	bfi_h2i_set(enable_req.mh, BFI_MC_IOC, BFI_IOC_H2I_ENABLE_REQ,
1623 		    bfa_ioc_portid(ioc));
1624 	enable_req.clscode = cpu_to_be16(ioc->clscode);
1625 	do_gettimeofday(&tv);
1626 	enable_req.tv_sec = be32_to_cpu(tv.tv_sec);
1627 	bfa_ioc_mbox_send(ioc, &enable_req, sizeof(struct bfi_ioc_ctrl_req_s));
1628 }
1629 
1630 static void
1631 bfa_ioc_send_disable(struct bfa_ioc_s *ioc)
1632 {
1633 	struct bfi_ioc_ctrl_req_s disable_req;
1634 
1635 	bfi_h2i_set(disable_req.mh, BFI_MC_IOC, BFI_IOC_H2I_DISABLE_REQ,
1636 		    bfa_ioc_portid(ioc));
1637 	bfa_ioc_mbox_send(ioc, &disable_req, sizeof(struct bfi_ioc_ctrl_req_s));
1638 }
1639 
1640 static void
1641 bfa_ioc_send_getattr(struct bfa_ioc_s *ioc)
1642 {
1643 	struct bfi_ioc_getattr_req_s	attr_req;
1644 
1645 	bfi_h2i_set(attr_req.mh, BFI_MC_IOC, BFI_IOC_H2I_GETATTR_REQ,
1646 		    bfa_ioc_portid(ioc));
1647 	bfa_dma_be_addr_set(attr_req.attr_addr, ioc->attr_dma.pa);
1648 	bfa_ioc_mbox_send(ioc, &attr_req, sizeof(attr_req));
1649 }
1650 
1651 static void
1652 bfa_ioc_hb_check(void *cbarg)
1653 {
1654 	struct bfa_ioc_s  *ioc = cbarg;
1655 	u32	hb_count;
1656 
1657 	hb_count = readl(ioc->ioc_regs.heartbeat);
1658 	if (ioc->hb_count == hb_count) {
1659 		bfa_ioc_recover(ioc);
1660 		return;
1661 	} else {
1662 		ioc->hb_count = hb_count;
1663 	}
1664 
1665 	bfa_ioc_mbox_poll(ioc);
1666 	bfa_hb_timer_start(ioc);
1667 }
1668 
1669 static void
1670 bfa_ioc_hb_monitor(struct bfa_ioc_s *ioc)
1671 {
1672 	ioc->hb_count = readl(ioc->ioc_regs.heartbeat);
1673 	bfa_hb_timer_start(ioc);
1674 }
1675 
1676 /*
1677  *	Initiate a full firmware download.
1678  */
1679 static void
1680 bfa_ioc_download_fw(struct bfa_ioc_s *ioc, u32 boot_type,
1681 		    u32 boot_env)
1682 {
1683 	u32 *fwimg;
1684 	u32 pgnum, pgoff;
1685 	u32 loff = 0;
1686 	u32 chunkno = 0;
1687 	u32 i;
1688 	u32 asicmode;
1689 
1690 	bfa_trc(ioc, bfa_cb_image_get_size(bfa_ioc_asic_gen(ioc)));
1691 	fwimg = bfa_cb_image_get_chunk(bfa_ioc_asic_gen(ioc), chunkno);
1692 
1693 	pgnum = PSS_SMEM_PGNUM(ioc->ioc_regs.smem_pg0, loff);
1694 	pgoff = PSS_SMEM_PGOFF(loff);
1695 
1696 	writel(pgnum, ioc->ioc_regs.host_page_num_fn);
1697 
1698 	for (i = 0; i < bfa_cb_image_get_size(bfa_ioc_asic_gen(ioc)); i++) {
1699 
1700 		if (BFA_IOC_FLASH_CHUNK_NO(i) != chunkno) {
1701 			chunkno = BFA_IOC_FLASH_CHUNK_NO(i);
1702 			fwimg = bfa_cb_image_get_chunk(bfa_ioc_asic_gen(ioc),
1703 					BFA_IOC_FLASH_CHUNK_ADDR(chunkno));
1704 		}
1705 
1706 		/*
1707 		 * write smem
1708 		 */
1709 		bfa_mem_write(ioc->ioc_regs.smem_page_start, loff,
1710 			cpu_to_le32(fwimg[BFA_IOC_FLASH_OFFSET_IN_CHUNK(i)]));
1711 
1712 		loff += sizeof(u32);
1713 
1714 		/*
1715 		 * handle page offset wrap around
1716 		 */
1717 		loff = PSS_SMEM_PGOFF(loff);
1718 		if (loff == 0) {
1719 			pgnum++;
1720 			writel(pgnum, ioc->ioc_regs.host_page_num_fn);
1721 		}
1722 	}
1723 
1724 	writel(PSS_SMEM_PGNUM(ioc->ioc_regs.smem_pg0, 0),
1725 			ioc->ioc_regs.host_page_num_fn);
1726 
1727 	/*
1728 	 * Set boot type and device mode at the end.
1729 	 */
1730 	asicmode = BFI_FWBOOT_DEVMODE(ioc->asic_gen, ioc->asic_mode,
1731 				ioc->port0_mode, ioc->port1_mode);
1732 	bfa_mem_write(ioc->ioc_regs.smem_page_start, BFI_FWBOOT_DEVMODE_OFF,
1733 			swab32(asicmode));
1734 	bfa_mem_write(ioc->ioc_regs.smem_page_start, BFI_FWBOOT_TYPE_OFF,
1735 			swab32(boot_type));
1736 	bfa_mem_write(ioc->ioc_regs.smem_page_start, BFI_FWBOOT_ENV_OFF,
1737 			swab32(boot_env));
1738 }
1739 
1740 
1741 /*
1742  * Update BFA configuration from firmware configuration.
1743  */
1744 static void
1745 bfa_ioc_getattr_reply(struct bfa_ioc_s *ioc)
1746 {
1747 	struct bfi_ioc_attr_s	*attr = ioc->attr;
1748 
1749 	attr->adapter_prop  = be32_to_cpu(attr->adapter_prop);
1750 	attr->card_type     = be32_to_cpu(attr->card_type);
1751 	attr->maxfrsize	    = be16_to_cpu(attr->maxfrsize);
1752 	ioc->fcmode	= (attr->port_mode == BFI_PORT_MODE_FC);
1753 	attr->mfg_year	= be16_to_cpu(attr->mfg_year);
1754 
1755 	bfa_fsm_send_event(ioc, IOC_E_FWRSP_GETATTR);
1756 }
1757 
1758 /*
1759  * Attach time initialization of mbox logic.
1760  */
1761 static void
1762 bfa_ioc_mbox_attach(struct bfa_ioc_s *ioc)
1763 {
1764 	struct bfa_ioc_mbox_mod_s	*mod = &ioc->mbox_mod;
1765 	int	mc;
1766 
1767 	INIT_LIST_HEAD(&mod->cmd_q);
1768 	for (mc = 0; mc < BFI_MC_MAX; mc++) {
1769 		mod->mbhdlr[mc].cbfn = NULL;
1770 		mod->mbhdlr[mc].cbarg = ioc->bfa;
1771 	}
1772 }
1773 
1774 /*
1775  * Mbox poll timer -- restarts any pending mailbox requests.
1776  */
1777 static void
1778 bfa_ioc_mbox_poll(struct bfa_ioc_s *ioc)
1779 {
1780 	struct bfa_ioc_mbox_mod_s	*mod = &ioc->mbox_mod;
1781 	struct bfa_mbox_cmd_s		*cmd;
1782 	u32			stat;
1783 
1784 	/*
1785 	 * If no command pending, do nothing
1786 	 */
1787 	if (list_empty(&mod->cmd_q))
1788 		return;
1789 
1790 	/*
1791 	 * If previous command is not yet fetched by firmware, do nothing
1792 	 */
1793 	stat = readl(ioc->ioc_regs.hfn_mbox_cmd);
1794 	if (stat)
1795 		return;
1796 
1797 	/*
1798 	 * Enqueue command to firmware.
1799 	 */
1800 	bfa_q_deq(&mod->cmd_q, &cmd);
1801 	bfa_ioc_mbox_send(ioc, cmd->msg, sizeof(cmd->msg));
1802 }
1803 
1804 /*
1805  * Cleanup any pending requests.
1806  */
1807 static void
1808 bfa_ioc_mbox_flush(struct bfa_ioc_s *ioc)
1809 {
1810 	struct bfa_ioc_mbox_mod_s	*mod = &ioc->mbox_mod;
1811 	struct bfa_mbox_cmd_s		*cmd;
1812 
1813 	while (!list_empty(&mod->cmd_q))
1814 		bfa_q_deq(&mod->cmd_q, &cmd);
1815 }
1816 
1817 /*
1818  * Read data from SMEM to host through PCI memmap
1819  *
1820  * @param[in]	ioc	memory for IOC
1821  * @param[in]	tbuf	app memory to store data from smem
1822  * @param[in]	soff	smem offset
1823  * @param[in]	sz	size of smem in bytes
1824  */
1825 static bfa_status_t
1826 bfa_ioc_smem_read(struct bfa_ioc_s *ioc, void *tbuf, u32 soff, u32 sz)
1827 {
1828 	u32 pgnum, loff;
1829 	__be32 r32;
1830 	int i, len;
1831 	u32 *buf = tbuf;
1832 
1833 	pgnum = PSS_SMEM_PGNUM(ioc->ioc_regs.smem_pg0, soff);
1834 	loff = PSS_SMEM_PGOFF(soff);
1835 	bfa_trc(ioc, pgnum);
1836 	bfa_trc(ioc, loff);
1837 	bfa_trc(ioc, sz);
1838 
1839 	/*
1840 	 *  Hold semaphore to serialize pll init and fwtrc.
1841 	 */
1842 	if (BFA_FALSE == bfa_ioc_sem_get(ioc->ioc_regs.ioc_init_sem_reg)) {
1843 		bfa_trc(ioc, 0);
1844 		return BFA_STATUS_FAILED;
1845 	}
1846 
1847 	writel(pgnum, ioc->ioc_regs.host_page_num_fn);
1848 
1849 	len = sz/sizeof(u32);
1850 	bfa_trc(ioc, len);
1851 	for (i = 0; i < len; i++) {
1852 		r32 = bfa_mem_read(ioc->ioc_regs.smem_page_start, loff);
1853 		buf[i] = be32_to_cpu(r32);
1854 		loff += sizeof(u32);
1855 
1856 		/*
1857 		 * handle page offset wrap around
1858 		 */
1859 		loff = PSS_SMEM_PGOFF(loff);
1860 		if (loff == 0) {
1861 			pgnum++;
1862 			writel(pgnum, ioc->ioc_regs.host_page_num_fn);
1863 		}
1864 	}
1865 	writel(PSS_SMEM_PGNUM(ioc->ioc_regs.smem_pg0, 0),
1866 			ioc->ioc_regs.host_page_num_fn);
1867 	/*
1868 	 *  release semaphore.
1869 	 */
1870 	readl(ioc->ioc_regs.ioc_init_sem_reg);
1871 	writel(1, ioc->ioc_regs.ioc_init_sem_reg);
1872 
1873 	bfa_trc(ioc, pgnum);
1874 	return BFA_STATUS_OK;
1875 }
1876 
1877 /*
1878  * Clear SMEM data from host through PCI memmap
1879  *
1880  * @param[in]	ioc	memory for IOC
1881  * @param[in]	soff	smem offset
1882  * @param[in]	sz	size of smem in bytes
1883  */
1884 static bfa_status_t
1885 bfa_ioc_smem_clr(struct bfa_ioc_s *ioc, u32 soff, u32 sz)
1886 {
1887 	int i, len;
1888 	u32 pgnum, loff;
1889 
1890 	pgnum = PSS_SMEM_PGNUM(ioc->ioc_regs.smem_pg0, soff);
1891 	loff = PSS_SMEM_PGOFF(soff);
1892 	bfa_trc(ioc, pgnum);
1893 	bfa_trc(ioc, loff);
1894 	bfa_trc(ioc, sz);
1895 
1896 	/*
1897 	 *  Hold semaphore to serialize pll init and fwtrc.
1898 	 */
1899 	if (BFA_FALSE == bfa_ioc_sem_get(ioc->ioc_regs.ioc_init_sem_reg)) {
1900 		bfa_trc(ioc, 0);
1901 		return BFA_STATUS_FAILED;
1902 	}
1903 
1904 	writel(pgnum, ioc->ioc_regs.host_page_num_fn);
1905 
1906 	len = sz/sizeof(u32); /* len in words */
1907 	bfa_trc(ioc, len);
1908 	for (i = 0; i < len; i++) {
1909 		bfa_mem_write(ioc->ioc_regs.smem_page_start, loff, 0);
1910 		loff += sizeof(u32);
1911 
1912 		/*
1913 		 * handle page offset wrap around
1914 		 */
1915 		loff = PSS_SMEM_PGOFF(loff);
1916 		if (loff == 0) {
1917 			pgnum++;
1918 			writel(pgnum, ioc->ioc_regs.host_page_num_fn);
1919 		}
1920 	}
1921 	writel(PSS_SMEM_PGNUM(ioc->ioc_regs.smem_pg0, 0),
1922 			ioc->ioc_regs.host_page_num_fn);
1923 
1924 	/*
1925 	 *  release semaphore.
1926 	 */
1927 	readl(ioc->ioc_regs.ioc_init_sem_reg);
1928 	writel(1, ioc->ioc_regs.ioc_init_sem_reg);
1929 	bfa_trc(ioc, pgnum);
1930 	return BFA_STATUS_OK;
1931 }
1932 
1933 static void
1934 bfa_ioc_fail_notify(struct bfa_ioc_s *ioc)
1935 {
1936 	struct bfad_s *bfad = (struct bfad_s *)ioc->bfa->bfad;
1937 
1938 	/*
1939 	 * Notify driver and common modules registered for notification.
1940 	 */
1941 	ioc->cbfn->hbfail_cbfn(ioc->bfa);
1942 	bfa_ioc_event_notify(ioc, BFA_IOC_E_FAILED);
1943 
1944 	bfa_ioc_debug_save_ftrc(ioc);
1945 
1946 	BFA_LOG(KERN_CRIT, bfad, bfa_log_level,
1947 		"Heart Beat of IOC has failed\n");
1948 	bfa_ioc_aen_post(ioc, BFA_IOC_AEN_HBFAIL);
1949 
1950 }
1951 
1952 static void
1953 bfa_ioc_pf_fwmismatch(struct bfa_ioc_s *ioc)
1954 {
1955 	struct bfad_s *bfad = (struct bfad_s *)ioc->bfa->bfad;
1956 	/*
1957 	 * Provide enable completion callback.
1958 	 */
1959 	ioc->cbfn->enable_cbfn(ioc->bfa, BFA_STATUS_IOC_FAILURE);
1960 	BFA_LOG(KERN_WARNING, bfad, bfa_log_level,
1961 		"Running firmware version is incompatible "
1962 		"with the driver version\n");
1963 	bfa_ioc_aen_post(ioc, BFA_IOC_AEN_FWMISMATCH);
1964 }
1965 
1966 bfa_status_t
1967 bfa_ioc_pll_init(struct bfa_ioc_s *ioc)
1968 {
1969 
1970 	/*
1971 	 *  Hold semaphore so that nobody can access the chip during init.
1972 	 */
1973 	bfa_ioc_sem_get(ioc->ioc_regs.ioc_init_sem_reg);
1974 
1975 	bfa_ioc_pll_init_asic(ioc);
1976 
1977 	ioc->pllinit = BFA_TRUE;
1978 
1979 	/*
1980 	 * Initialize LMEM
1981 	 */
1982 	bfa_ioc_lmem_init(ioc);
1983 
1984 	/*
1985 	 *  release semaphore.
1986 	 */
1987 	readl(ioc->ioc_regs.ioc_init_sem_reg);
1988 	writel(1, ioc->ioc_regs.ioc_init_sem_reg);
1989 
1990 	return BFA_STATUS_OK;
1991 }
1992 
1993 /*
1994  * Interface used by diag module to do firmware boot with memory test
1995  * as the entry vector.
1996  */
1997 void
1998 bfa_ioc_boot(struct bfa_ioc_s *ioc, u32 boot_type, u32 boot_env)
1999 {
2000 	bfa_ioc_stats(ioc, ioc_boots);
2001 
2002 	if (bfa_ioc_pll_init(ioc) != BFA_STATUS_OK)
2003 		return;
2004 
2005 	/*
2006 	 * Initialize IOC state of all functions on a chip reset.
2007 	 */
2008 	if (boot_type == BFI_FWBOOT_TYPE_MEMTEST) {
2009 		writel(BFI_IOC_MEMTEST, ioc->ioc_regs.ioc_fwstate);
2010 		writel(BFI_IOC_MEMTEST, ioc->ioc_regs.alt_ioc_fwstate);
2011 	} else {
2012 		writel(BFI_IOC_INITING, ioc->ioc_regs.ioc_fwstate);
2013 		writel(BFI_IOC_INITING, ioc->ioc_regs.alt_ioc_fwstate);
2014 	}
2015 
2016 	bfa_ioc_msgflush(ioc);
2017 	bfa_ioc_download_fw(ioc, boot_type, boot_env);
2018 	bfa_ioc_lpu_start(ioc);
2019 }
2020 
2021 /*
2022  * Enable/disable IOC failure auto recovery.
2023  */
2024 void
2025 bfa_ioc_auto_recover(bfa_boolean_t auto_recover)
2026 {
2027 	bfa_auto_recover = auto_recover;
2028 }
2029 
2030 
2031 
2032 bfa_boolean_t
2033 bfa_ioc_is_operational(struct bfa_ioc_s *ioc)
2034 {
2035 	return bfa_fsm_cmp_state(ioc, bfa_ioc_sm_op);
2036 }
2037 
2038 bfa_boolean_t
2039 bfa_ioc_is_initialized(struct bfa_ioc_s *ioc)
2040 {
2041 	u32 r32 = readl(ioc->ioc_regs.ioc_fwstate);
2042 
2043 	return ((r32 != BFI_IOC_UNINIT) &&
2044 		(r32 != BFI_IOC_INITING) &&
2045 		(r32 != BFI_IOC_MEMTEST));
2046 }
2047 
2048 bfa_boolean_t
2049 bfa_ioc_msgget(struct bfa_ioc_s *ioc, void *mbmsg)
2050 {
2051 	__be32	*msgp = mbmsg;
2052 	u32	r32;
2053 	int		i;
2054 
2055 	r32 = readl(ioc->ioc_regs.lpu_mbox_cmd);
2056 	if ((r32 & 1) == 0)
2057 		return BFA_FALSE;
2058 
2059 	/*
2060 	 * read the MBOX msg
2061 	 */
2062 	for (i = 0; i < (sizeof(union bfi_ioc_i2h_msg_u) / sizeof(u32));
2063 	     i++) {
2064 		r32 = readl(ioc->ioc_regs.lpu_mbox +
2065 				   i * sizeof(u32));
2066 		msgp[i] = cpu_to_be32(r32);
2067 	}
2068 
2069 	/*
2070 	 * turn off mailbox interrupt by clearing mailbox status
2071 	 */
2072 	writel(1, ioc->ioc_regs.lpu_mbox_cmd);
2073 	readl(ioc->ioc_regs.lpu_mbox_cmd);
2074 
2075 	return BFA_TRUE;
2076 }
2077 
2078 void
2079 bfa_ioc_isr(struct bfa_ioc_s *ioc, struct bfi_mbmsg_s *m)
2080 {
2081 	union bfi_ioc_i2h_msg_u	*msg;
2082 	struct bfa_iocpf_s *iocpf = &ioc->iocpf;
2083 
2084 	msg = (union bfi_ioc_i2h_msg_u *) m;
2085 
2086 	bfa_ioc_stats(ioc, ioc_isrs);
2087 
2088 	switch (msg->mh.msg_id) {
2089 	case BFI_IOC_I2H_HBEAT:
2090 		break;
2091 
2092 	case BFI_IOC_I2H_ENABLE_REPLY:
2093 		ioc->port_mode = ioc->port_mode_cfg =
2094 				(enum bfa_mode_s)msg->fw_event.port_mode;
2095 		ioc->ad_cap_bm = msg->fw_event.cap_bm;
2096 		bfa_fsm_send_event(iocpf, IOCPF_E_FWRSP_ENABLE);
2097 		break;
2098 
2099 	case BFI_IOC_I2H_DISABLE_REPLY:
2100 		bfa_fsm_send_event(iocpf, IOCPF_E_FWRSP_DISABLE);
2101 		break;
2102 
2103 	case BFI_IOC_I2H_GETATTR_REPLY:
2104 		bfa_ioc_getattr_reply(ioc);
2105 		break;
2106 
2107 	default:
2108 		bfa_trc(ioc, msg->mh.msg_id);
2109 		WARN_ON(1);
2110 	}
2111 }
2112 
2113 /*
2114  * IOC attach time initialization and setup.
2115  *
2116  * @param[in]	ioc	memory for IOC
2117  * @param[in]	bfa	driver instance structure
2118  */
2119 void
2120 bfa_ioc_attach(struct bfa_ioc_s *ioc, void *bfa, struct bfa_ioc_cbfn_s *cbfn,
2121 	       struct bfa_timer_mod_s *timer_mod)
2122 {
2123 	ioc->bfa	= bfa;
2124 	ioc->cbfn	= cbfn;
2125 	ioc->timer_mod	= timer_mod;
2126 	ioc->fcmode	= BFA_FALSE;
2127 	ioc->pllinit	= BFA_FALSE;
2128 	ioc->dbg_fwsave_once = BFA_TRUE;
2129 	ioc->iocpf.ioc	= ioc;
2130 
2131 	bfa_ioc_mbox_attach(ioc);
2132 	INIT_LIST_HEAD(&ioc->notify_q);
2133 
2134 	bfa_fsm_set_state(ioc, bfa_ioc_sm_uninit);
2135 	bfa_fsm_send_event(ioc, IOC_E_RESET);
2136 }
2137 
2138 /*
2139  * Driver detach time IOC cleanup.
2140  */
2141 void
2142 bfa_ioc_detach(struct bfa_ioc_s *ioc)
2143 {
2144 	bfa_fsm_send_event(ioc, IOC_E_DETACH);
2145 	INIT_LIST_HEAD(&ioc->notify_q);
2146 }
2147 
2148 /*
2149  * Setup IOC PCI properties.
2150  *
2151  * @param[in]	pcidev	PCI device information for this IOC
2152  */
2153 void
2154 bfa_ioc_pci_init(struct bfa_ioc_s *ioc, struct bfa_pcidev_s *pcidev,
2155 		enum bfi_pcifn_class clscode)
2156 {
2157 	ioc->clscode	= clscode;
2158 	ioc->pcidev	= *pcidev;
2159 
2160 	/*
2161 	 * Initialize IOC and device personality
2162 	 */
2163 	ioc->port0_mode = ioc->port1_mode = BFI_PORT_MODE_FC;
2164 	ioc->asic_mode  = BFI_ASIC_MODE_FC;
2165 
2166 	switch (pcidev->device_id) {
2167 	case BFA_PCI_DEVICE_ID_FC_8G1P:
2168 	case BFA_PCI_DEVICE_ID_FC_8G2P:
2169 		ioc->asic_gen = BFI_ASIC_GEN_CB;
2170 		ioc->fcmode = BFA_TRUE;
2171 		ioc->port_mode = ioc->port_mode_cfg = BFA_MODE_HBA;
2172 		ioc->ad_cap_bm = BFA_CM_HBA;
2173 		break;
2174 
2175 	case BFA_PCI_DEVICE_ID_CT:
2176 		ioc->asic_gen = BFI_ASIC_GEN_CT;
2177 		ioc->port0_mode = ioc->port1_mode = BFI_PORT_MODE_ETH;
2178 		ioc->asic_mode  = BFI_ASIC_MODE_ETH;
2179 		ioc->port_mode = ioc->port_mode_cfg = BFA_MODE_CNA;
2180 		ioc->ad_cap_bm = BFA_CM_CNA;
2181 		break;
2182 
2183 	case BFA_PCI_DEVICE_ID_CT_FC:
2184 		ioc->asic_gen = BFI_ASIC_GEN_CT;
2185 		ioc->fcmode = BFA_TRUE;
2186 		ioc->port_mode = ioc->port_mode_cfg = BFA_MODE_HBA;
2187 		ioc->ad_cap_bm = BFA_CM_HBA;
2188 		break;
2189 
2190 	case BFA_PCI_DEVICE_ID_CT2:
2191 		ioc->asic_gen = BFI_ASIC_GEN_CT2;
2192 		if (clscode == BFI_PCIFN_CLASS_FC &&
2193 		    pcidev->ssid == BFA_PCI_CT2_SSID_FC) {
2194 			ioc->asic_mode  = BFI_ASIC_MODE_FC16;
2195 			ioc->fcmode = BFA_TRUE;
2196 			ioc->port_mode = ioc->port_mode_cfg = BFA_MODE_HBA;
2197 			ioc->ad_cap_bm = BFA_CM_HBA;
2198 		} else {
2199 			ioc->port0_mode = ioc->port1_mode = BFI_PORT_MODE_ETH;
2200 			ioc->asic_mode  = BFI_ASIC_MODE_ETH;
2201 			if (pcidev->ssid == BFA_PCI_CT2_SSID_FCoE) {
2202 				ioc->port_mode =
2203 				ioc->port_mode_cfg = BFA_MODE_CNA;
2204 				ioc->ad_cap_bm = BFA_CM_CNA;
2205 			} else {
2206 				ioc->port_mode =
2207 				ioc->port_mode_cfg = BFA_MODE_NIC;
2208 				ioc->ad_cap_bm = BFA_CM_NIC;
2209 			}
2210 		}
2211 		break;
2212 
2213 	default:
2214 		WARN_ON(1);
2215 	}
2216 
2217 	/*
2218 	 * Set asic specific interfaces. See bfa_ioc_cb.c and bfa_ioc_ct.c
2219 	 */
2220 	if (ioc->asic_gen == BFI_ASIC_GEN_CB)
2221 		bfa_ioc_set_cb_hwif(ioc);
2222 	else if (ioc->asic_gen == BFI_ASIC_GEN_CT)
2223 		bfa_ioc_set_ct_hwif(ioc);
2224 	else {
2225 		WARN_ON(ioc->asic_gen != BFI_ASIC_GEN_CT2);
2226 		bfa_ioc_set_ct2_hwif(ioc);
2227 		bfa_ioc_ct2_poweron(ioc);
2228 	}
2229 
2230 	bfa_ioc_map_port(ioc);
2231 	bfa_ioc_reg_init(ioc);
2232 }
2233 
2234 /*
2235  * Initialize IOC dma memory
2236  *
2237  * @param[in]	dm_kva	kernel virtual address of IOC dma memory
2238  * @param[in]	dm_pa	physical address of IOC dma memory
2239  */
2240 void
2241 bfa_ioc_mem_claim(struct bfa_ioc_s *ioc,  u8 *dm_kva, u64 dm_pa)
2242 {
2243 	/*
2244 	 * dma memory for firmware attribute
2245 	 */
2246 	ioc->attr_dma.kva = dm_kva;
2247 	ioc->attr_dma.pa = dm_pa;
2248 	ioc->attr = (struct bfi_ioc_attr_s *) dm_kva;
2249 }
2250 
2251 void
2252 bfa_ioc_enable(struct bfa_ioc_s *ioc)
2253 {
2254 	bfa_ioc_stats(ioc, ioc_enables);
2255 	ioc->dbg_fwsave_once = BFA_TRUE;
2256 
2257 	bfa_fsm_send_event(ioc, IOC_E_ENABLE);
2258 }
2259 
2260 void
2261 bfa_ioc_disable(struct bfa_ioc_s *ioc)
2262 {
2263 	bfa_ioc_stats(ioc, ioc_disables);
2264 	bfa_fsm_send_event(ioc, IOC_E_DISABLE);
2265 }
2266 
2267 void
2268 bfa_ioc_suspend(struct bfa_ioc_s *ioc)
2269 {
2270 	ioc->dbg_fwsave_once = BFA_TRUE;
2271 	bfa_fsm_send_event(ioc, IOC_E_HWERROR);
2272 }
2273 
2274 /*
2275  * Initialize memory for saving firmware trace. Driver must initialize
2276  * trace memory before call bfa_ioc_enable().
2277  */
2278 void
2279 bfa_ioc_debug_memclaim(struct bfa_ioc_s *ioc, void *dbg_fwsave)
2280 {
2281 	ioc->dbg_fwsave	    = dbg_fwsave;
2282 	ioc->dbg_fwsave_len = BFA_DBG_FWTRC_LEN;
2283 }
2284 
2285 /*
2286  * Register mailbox message handler functions
2287  *
2288  * @param[in]	ioc		IOC instance
2289  * @param[in]	mcfuncs		message class handler functions
2290  */
2291 void
2292 bfa_ioc_mbox_register(struct bfa_ioc_s *ioc, bfa_ioc_mbox_mcfunc_t *mcfuncs)
2293 {
2294 	struct bfa_ioc_mbox_mod_s	*mod = &ioc->mbox_mod;
2295 	int				mc;
2296 
2297 	for (mc = 0; mc < BFI_MC_MAX; mc++)
2298 		mod->mbhdlr[mc].cbfn = mcfuncs[mc];
2299 }
2300 
2301 /*
2302  * Register mailbox message handler function, to be called by common modules
2303  */
2304 void
2305 bfa_ioc_mbox_regisr(struct bfa_ioc_s *ioc, enum bfi_mclass mc,
2306 		    bfa_ioc_mbox_mcfunc_t cbfn, void *cbarg)
2307 {
2308 	struct bfa_ioc_mbox_mod_s	*mod = &ioc->mbox_mod;
2309 
2310 	mod->mbhdlr[mc].cbfn	= cbfn;
2311 	mod->mbhdlr[mc].cbarg	= cbarg;
2312 }
2313 
2314 /*
2315  * Queue a mailbox command request to firmware. Waits if mailbox is busy.
2316  * Responsibility of caller to serialize
2317  *
2318  * @param[in]	ioc	IOC instance
2319  * @param[i]	cmd	Mailbox command
2320  */
2321 void
2322 bfa_ioc_mbox_queue(struct bfa_ioc_s *ioc, struct bfa_mbox_cmd_s *cmd)
2323 {
2324 	struct bfa_ioc_mbox_mod_s	*mod = &ioc->mbox_mod;
2325 	u32			stat;
2326 
2327 	/*
2328 	 * If a previous command is pending, queue new command
2329 	 */
2330 	if (!list_empty(&mod->cmd_q)) {
2331 		list_add_tail(&cmd->qe, &mod->cmd_q);
2332 		return;
2333 	}
2334 
2335 	/*
2336 	 * If mailbox is busy, queue command for poll timer
2337 	 */
2338 	stat = readl(ioc->ioc_regs.hfn_mbox_cmd);
2339 	if (stat) {
2340 		list_add_tail(&cmd->qe, &mod->cmd_q);
2341 		return;
2342 	}
2343 
2344 	/*
2345 	 * mailbox is free -- queue command to firmware
2346 	 */
2347 	bfa_ioc_mbox_send(ioc, cmd->msg, sizeof(cmd->msg));
2348 }
2349 
2350 /*
2351  * Handle mailbox interrupts
2352  */
2353 void
2354 bfa_ioc_mbox_isr(struct bfa_ioc_s *ioc)
2355 {
2356 	struct bfa_ioc_mbox_mod_s	*mod = &ioc->mbox_mod;
2357 	struct bfi_mbmsg_s		m;
2358 	int				mc;
2359 
2360 	if (bfa_ioc_msgget(ioc, &m)) {
2361 		/*
2362 		 * Treat IOC message class as special.
2363 		 */
2364 		mc = m.mh.msg_class;
2365 		if (mc == BFI_MC_IOC) {
2366 			bfa_ioc_isr(ioc, &m);
2367 			return;
2368 		}
2369 
2370 		if ((mc >= BFI_MC_MAX) || (mod->mbhdlr[mc].cbfn == NULL))
2371 			return;
2372 
2373 		mod->mbhdlr[mc].cbfn(mod->mbhdlr[mc].cbarg, &m);
2374 	}
2375 
2376 	bfa_ioc_lpu_read_stat(ioc);
2377 
2378 	/*
2379 	 * Try to send pending mailbox commands
2380 	 */
2381 	bfa_ioc_mbox_poll(ioc);
2382 }
2383 
2384 void
2385 bfa_ioc_error_isr(struct bfa_ioc_s *ioc)
2386 {
2387 	bfa_ioc_stats(ioc, ioc_hbfails);
2388 	ioc->stats.hb_count = ioc->hb_count;
2389 	bfa_fsm_send_event(ioc, IOC_E_HWERROR);
2390 }
2391 
2392 /*
2393  * return true if IOC is disabled
2394  */
2395 bfa_boolean_t
2396 bfa_ioc_is_disabled(struct bfa_ioc_s *ioc)
2397 {
2398 	return bfa_fsm_cmp_state(ioc, bfa_ioc_sm_disabling) ||
2399 		bfa_fsm_cmp_state(ioc, bfa_ioc_sm_disabled);
2400 }
2401 
2402 /*
2403  * return true if IOC firmware is different.
2404  */
2405 bfa_boolean_t
2406 bfa_ioc_fw_mismatch(struct bfa_ioc_s *ioc)
2407 {
2408 	return bfa_fsm_cmp_state(ioc, bfa_ioc_sm_reset) ||
2409 		bfa_fsm_cmp_state(&ioc->iocpf, bfa_iocpf_sm_fwcheck) ||
2410 		bfa_fsm_cmp_state(&ioc->iocpf, bfa_iocpf_sm_mismatch);
2411 }
2412 
2413 #define bfa_ioc_state_disabled(__sm)		\
2414 	(((__sm) == BFI_IOC_UNINIT) ||		\
2415 	 ((__sm) == BFI_IOC_INITING) ||		\
2416 	 ((__sm) == BFI_IOC_HWINIT) ||		\
2417 	 ((__sm) == BFI_IOC_DISABLED) ||	\
2418 	 ((__sm) == BFI_IOC_FAIL) ||		\
2419 	 ((__sm) == BFI_IOC_CFG_DISABLED))
2420 
2421 /*
2422  * Check if adapter is disabled -- both IOCs should be in a disabled
2423  * state.
2424  */
2425 bfa_boolean_t
2426 bfa_ioc_adapter_is_disabled(struct bfa_ioc_s *ioc)
2427 {
2428 	u32	ioc_state;
2429 
2430 	if (!bfa_fsm_cmp_state(ioc, bfa_ioc_sm_disabled))
2431 		return BFA_FALSE;
2432 
2433 	ioc_state = readl(ioc->ioc_regs.ioc_fwstate);
2434 	if (!bfa_ioc_state_disabled(ioc_state))
2435 		return BFA_FALSE;
2436 
2437 	if (ioc->pcidev.device_id != BFA_PCI_DEVICE_ID_FC_8G1P) {
2438 		ioc_state = readl(ioc->ioc_regs.alt_ioc_fwstate);
2439 		if (!bfa_ioc_state_disabled(ioc_state))
2440 			return BFA_FALSE;
2441 	}
2442 
2443 	return BFA_TRUE;
2444 }
2445 
2446 /*
2447  * Reset IOC fwstate registers.
2448  */
2449 void
2450 bfa_ioc_reset_fwstate(struct bfa_ioc_s *ioc)
2451 {
2452 	writel(BFI_IOC_UNINIT, ioc->ioc_regs.ioc_fwstate);
2453 	writel(BFI_IOC_UNINIT, ioc->ioc_regs.alt_ioc_fwstate);
2454 }
2455 
2456 #define BFA_MFG_NAME "Brocade"
2457 void
2458 bfa_ioc_get_adapter_attr(struct bfa_ioc_s *ioc,
2459 			 struct bfa_adapter_attr_s *ad_attr)
2460 {
2461 	struct bfi_ioc_attr_s	*ioc_attr;
2462 
2463 	ioc_attr = ioc->attr;
2464 
2465 	bfa_ioc_get_adapter_serial_num(ioc, ad_attr->serial_num);
2466 	bfa_ioc_get_adapter_fw_ver(ioc, ad_attr->fw_ver);
2467 	bfa_ioc_get_adapter_optrom_ver(ioc, ad_attr->optrom_ver);
2468 	bfa_ioc_get_adapter_manufacturer(ioc, ad_attr->manufacturer);
2469 	memcpy(&ad_attr->vpd, &ioc_attr->vpd,
2470 		      sizeof(struct bfa_mfg_vpd_s));
2471 
2472 	ad_attr->nports = bfa_ioc_get_nports(ioc);
2473 	ad_attr->max_speed = bfa_ioc_speed_sup(ioc);
2474 
2475 	bfa_ioc_get_adapter_model(ioc, ad_attr->model);
2476 	/* For now, model descr uses same model string */
2477 	bfa_ioc_get_adapter_model(ioc, ad_attr->model_descr);
2478 
2479 	ad_attr->card_type = ioc_attr->card_type;
2480 	ad_attr->is_mezz = bfa_mfg_is_mezz(ioc_attr->card_type);
2481 
2482 	if (BFI_ADAPTER_IS_SPECIAL(ioc_attr->adapter_prop))
2483 		ad_attr->prototype = 1;
2484 	else
2485 		ad_attr->prototype = 0;
2486 
2487 	ad_attr->pwwn = ioc->attr->pwwn;
2488 	ad_attr->mac  = bfa_ioc_get_mac(ioc);
2489 
2490 	ad_attr->pcie_gen = ioc_attr->pcie_gen;
2491 	ad_attr->pcie_lanes = ioc_attr->pcie_lanes;
2492 	ad_attr->pcie_lanes_orig = ioc_attr->pcie_lanes_orig;
2493 	ad_attr->asic_rev = ioc_attr->asic_rev;
2494 
2495 	bfa_ioc_get_pci_chip_rev(ioc, ad_attr->hw_ver);
2496 
2497 	ad_attr->cna_capable = bfa_ioc_is_cna(ioc);
2498 	ad_attr->trunk_capable = (ad_attr->nports > 1) &&
2499 				  !bfa_ioc_is_cna(ioc) && !ad_attr->is_mezz;
2500 	ad_attr->mfg_day = ioc_attr->mfg_day;
2501 	ad_attr->mfg_month = ioc_attr->mfg_month;
2502 	ad_attr->mfg_year = ioc_attr->mfg_year;
2503 }
2504 
2505 enum bfa_ioc_type_e
2506 bfa_ioc_get_type(struct bfa_ioc_s *ioc)
2507 {
2508 	if (ioc->clscode == BFI_PCIFN_CLASS_ETH)
2509 		return BFA_IOC_TYPE_LL;
2510 
2511 	WARN_ON(ioc->clscode != BFI_PCIFN_CLASS_FC);
2512 
2513 	return (ioc->attr->port_mode == BFI_PORT_MODE_FC)
2514 		? BFA_IOC_TYPE_FC : BFA_IOC_TYPE_FCoE;
2515 }
2516 
2517 void
2518 bfa_ioc_get_adapter_serial_num(struct bfa_ioc_s *ioc, char *serial_num)
2519 {
2520 	memset((void *)serial_num, 0, BFA_ADAPTER_SERIAL_NUM_LEN);
2521 	memcpy((void *)serial_num,
2522 			(void *)ioc->attr->brcd_serialnum,
2523 			BFA_ADAPTER_SERIAL_NUM_LEN);
2524 }
2525 
2526 void
2527 bfa_ioc_get_adapter_fw_ver(struct bfa_ioc_s *ioc, char *fw_ver)
2528 {
2529 	memset((void *)fw_ver, 0, BFA_VERSION_LEN);
2530 	memcpy(fw_ver, ioc->attr->fw_version, BFA_VERSION_LEN);
2531 }
2532 
2533 void
2534 bfa_ioc_get_pci_chip_rev(struct bfa_ioc_s *ioc, char *chip_rev)
2535 {
2536 	WARN_ON(!chip_rev);
2537 
2538 	memset((void *)chip_rev, 0, BFA_IOC_CHIP_REV_LEN);
2539 
2540 	chip_rev[0] = 'R';
2541 	chip_rev[1] = 'e';
2542 	chip_rev[2] = 'v';
2543 	chip_rev[3] = '-';
2544 	chip_rev[4] = ioc->attr->asic_rev;
2545 	chip_rev[5] = '\0';
2546 }
2547 
2548 void
2549 bfa_ioc_get_adapter_optrom_ver(struct bfa_ioc_s *ioc, char *optrom_ver)
2550 {
2551 	memset((void *)optrom_ver, 0, BFA_VERSION_LEN);
2552 	memcpy(optrom_ver, ioc->attr->optrom_version,
2553 		      BFA_VERSION_LEN);
2554 }
2555 
2556 void
2557 bfa_ioc_get_adapter_manufacturer(struct bfa_ioc_s *ioc, char *manufacturer)
2558 {
2559 	memset((void *)manufacturer, 0, BFA_ADAPTER_MFG_NAME_LEN);
2560 	memcpy(manufacturer, BFA_MFG_NAME, BFA_ADAPTER_MFG_NAME_LEN);
2561 }
2562 
2563 void
2564 bfa_ioc_get_adapter_model(struct bfa_ioc_s *ioc, char *model)
2565 {
2566 	struct bfi_ioc_attr_s	*ioc_attr;
2567 
2568 	WARN_ON(!model);
2569 	memset((void *)model, 0, BFA_ADAPTER_MODEL_NAME_LEN);
2570 
2571 	ioc_attr = ioc->attr;
2572 
2573 	snprintf(model, BFA_ADAPTER_MODEL_NAME_LEN, "%s-%u",
2574 			BFA_MFG_NAME, ioc_attr->card_type);
2575 }
2576 
2577 enum bfa_ioc_state
2578 bfa_ioc_get_state(struct bfa_ioc_s *ioc)
2579 {
2580 	enum bfa_iocpf_state iocpf_st;
2581 	enum bfa_ioc_state ioc_st = bfa_sm_to_state(ioc_sm_table, ioc->fsm);
2582 
2583 	if (ioc_st == BFA_IOC_ENABLING ||
2584 		ioc_st == BFA_IOC_FAIL || ioc_st == BFA_IOC_INITFAIL) {
2585 
2586 		iocpf_st = bfa_sm_to_state(iocpf_sm_table, ioc->iocpf.fsm);
2587 
2588 		switch (iocpf_st) {
2589 		case BFA_IOCPF_SEMWAIT:
2590 			ioc_st = BFA_IOC_SEMWAIT;
2591 			break;
2592 
2593 		case BFA_IOCPF_HWINIT:
2594 			ioc_st = BFA_IOC_HWINIT;
2595 			break;
2596 
2597 		case BFA_IOCPF_FWMISMATCH:
2598 			ioc_st = BFA_IOC_FWMISMATCH;
2599 			break;
2600 
2601 		case BFA_IOCPF_FAIL:
2602 			ioc_st = BFA_IOC_FAIL;
2603 			break;
2604 
2605 		case BFA_IOCPF_INITFAIL:
2606 			ioc_st = BFA_IOC_INITFAIL;
2607 			break;
2608 
2609 		default:
2610 			break;
2611 		}
2612 	}
2613 
2614 	return ioc_st;
2615 }
2616 
2617 void
2618 bfa_ioc_get_attr(struct bfa_ioc_s *ioc, struct bfa_ioc_attr_s *ioc_attr)
2619 {
2620 	memset((void *)ioc_attr, 0, sizeof(struct bfa_ioc_attr_s));
2621 
2622 	ioc_attr->state = bfa_ioc_get_state(ioc);
2623 	ioc_attr->port_id = ioc->port_id;
2624 	ioc_attr->port_mode = ioc->port_mode;
2625 	ioc_attr->port_mode_cfg = ioc->port_mode_cfg;
2626 	ioc_attr->cap_bm = ioc->ad_cap_bm;
2627 
2628 	ioc_attr->ioc_type = bfa_ioc_get_type(ioc);
2629 
2630 	bfa_ioc_get_adapter_attr(ioc, &ioc_attr->adapter_attr);
2631 
2632 	ioc_attr->pci_attr.device_id = ioc->pcidev.device_id;
2633 	ioc_attr->pci_attr.pcifn = ioc->pcidev.pci_func;
2634 	bfa_ioc_get_pci_chip_rev(ioc, ioc_attr->pci_attr.chip_rev);
2635 }
2636 
2637 mac_t
2638 bfa_ioc_get_mac(struct bfa_ioc_s *ioc)
2639 {
2640 	/*
2641 	 * Check the IOC type and return the appropriate MAC
2642 	 */
2643 	if (bfa_ioc_get_type(ioc) == BFA_IOC_TYPE_FCoE)
2644 		return ioc->attr->fcoe_mac;
2645 	else
2646 		return ioc->attr->mac;
2647 }
2648 
2649 mac_t
2650 bfa_ioc_get_mfg_mac(struct bfa_ioc_s *ioc)
2651 {
2652 	mac_t	m;
2653 
2654 	m = ioc->attr->mfg_mac;
2655 	if (bfa_mfg_is_old_wwn_mac_model(ioc->attr->card_type))
2656 		m.mac[MAC_ADDRLEN - 1] += bfa_ioc_pcifn(ioc);
2657 	else
2658 		bfa_mfg_increment_wwn_mac(&(m.mac[MAC_ADDRLEN-3]),
2659 			bfa_ioc_pcifn(ioc));
2660 
2661 	return m;
2662 }
2663 
2664 /*
2665  * Send AEN notification
2666  */
2667 void
2668 bfa_ioc_aen_post(struct bfa_ioc_s *ioc, enum bfa_ioc_aen_event event)
2669 {
2670 	struct bfad_s *bfad = (struct bfad_s *)ioc->bfa->bfad;
2671 	struct bfa_aen_entry_s	*aen_entry;
2672 	enum bfa_ioc_type_e ioc_type;
2673 
2674 	bfad_get_aen_entry(bfad, aen_entry);
2675 	if (!aen_entry)
2676 		return;
2677 
2678 	ioc_type = bfa_ioc_get_type(ioc);
2679 	switch (ioc_type) {
2680 	case BFA_IOC_TYPE_FC:
2681 		aen_entry->aen_data.ioc.pwwn = ioc->attr->pwwn;
2682 		break;
2683 	case BFA_IOC_TYPE_FCoE:
2684 		aen_entry->aen_data.ioc.pwwn = ioc->attr->pwwn;
2685 		aen_entry->aen_data.ioc.mac = bfa_ioc_get_mac(ioc);
2686 		break;
2687 	case BFA_IOC_TYPE_LL:
2688 		aen_entry->aen_data.ioc.mac = bfa_ioc_get_mac(ioc);
2689 		break;
2690 	default:
2691 		WARN_ON(ioc_type != BFA_IOC_TYPE_FC);
2692 		break;
2693 	}
2694 
2695 	/* Send the AEN notification */
2696 	aen_entry->aen_data.ioc.ioc_type = ioc_type;
2697 	bfad_im_post_vendor_event(aen_entry, bfad, ++ioc->ioc_aen_seq,
2698 				  BFA_AEN_CAT_IOC, event);
2699 }
2700 
2701 /*
2702  * Retrieve saved firmware trace from a prior IOC failure.
2703  */
2704 bfa_status_t
2705 bfa_ioc_debug_fwsave(struct bfa_ioc_s *ioc, void *trcdata, int *trclen)
2706 {
2707 	int	tlen;
2708 
2709 	if (ioc->dbg_fwsave_len == 0)
2710 		return BFA_STATUS_ENOFSAVE;
2711 
2712 	tlen = *trclen;
2713 	if (tlen > ioc->dbg_fwsave_len)
2714 		tlen = ioc->dbg_fwsave_len;
2715 
2716 	memcpy(trcdata, ioc->dbg_fwsave, tlen);
2717 	*trclen = tlen;
2718 	return BFA_STATUS_OK;
2719 }
2720 
2721 
2722 /*
2723  * Retrieve saved firmware trace from a prior IOC failure.
2724  */
2725 bfa_status_t
2726 bfa_ioc_debug_fwtrc(struct bfa_ioc_s *ioc, void *trcdata, int *trclen)
2727 {
2728 	u32 loff = BFA_DBG_FWTRC_OFF(bfa_ioc_portid(ioc));
2729 	int tlen;
2730 	bfa_status_t status;
2731 
2732 	bfa_trc(ioc, *trclen);
2733 
2734 	tlen = *trclen;
2735 	if (tlen > BFA_DBG_FWTRC_LEN)
2736 		tlen = BFA_DBG_FWTRC_LEN;
2737 
2738 	status = bfa_ioc_smem_read(ioc, trcdata, loff, tlen);
2739 	*trclen = tlen;
2740 	return status;
2741 }
2742 
2743 static void
2744 bfa_ioc_send_fwsync(struct bfa_ioc_s *ioc)
2745 {
2746 	struct bfa_mbox_cmd_s cmd;
2747 	struct bfi_ioc_ctrl_req_s *req = (struct bfi_ioc_ctrl_req_s *) cmd.msg;
2748 
2749 	bfi_h2i_set(req->mh, BFI_MC_IOC, BFI_IOC_H2I_DBG_SYNC,
2750 		    bfa_ioc_portid(ioc));
2751 	req->clscode = cpu_to_be16(ioc->clscode);
2752 	bfa_ioc_mbox_queue(ioc, &cmd);
2753 }
2754 
2755 static void
2756 bfa_ioc_fwsync(struct bfa_ioc_s *ioc)
2757 {
2758 	u32 fwsync_iter = 1000;
2759 
2760 	bfa_ioc_send_fwsync(ioc);
2761 
2762 	/*
2763 	 * After sending a fw sync mbox command wait for it to
2764 	 * take effect.  We will not wait for a response because
2765 	 *    1. fw_sync mbox cmd doesn't have a response.
2766 	 *    2. Even if we implement that,  interrupts might not
2767 	 *	 be enabled when we call this function.
2768 	 * So, just keep checking if any mbox cmd is pending, and
2769 	 * after waiting for a reasonable amount of time, go ahead.
2770 	 * It is possible that fw has crashed and the mbox command
2771 	 * is never acknowledged.
2772 	 */
2773 	while (bfa_ioc_mbox_cmd_pending(ioc) && fwsync_iter > 0)
2774 		fwsync_iter--;
2775 }
2776 
2777 /*
2778  * Dump firmware smem
2779  */
2780 bfa_status_t
2781 bfa_ioc_debug_fwcore(struct bfa_ioc_s *ioc, void *buf,
2782 				u32 *offset, int *buflen)
2783 {
2784 	u32 loff;
2785 	int dlen;
2786 	bfa_status_t status;
2787 	u32 smem_len = BFA_IOC_FW_SMEM_SIZE(ioc);
2788 
2789 	if (*offset >= smem_len) {
2790 		*offset = *buflen = 0;
2791 		return BFA_STATUS_EINVAL;
2792 	}
2793 
2794 	loff = *offset;
2795 	dlen = *buflen;
2796 
2797 	/*
2798 	 * First smem read, sync smem before proceeding
2799 	 * No need to sync before reading every chunk.
2800 	 */
2801 	if (loff == 0)
2802 		bfa_ioc_fwsync(ioc);
2803 
2804 	if ((loff + dlen) >= smem_len)
2805 		dlen = smem_len - loff;
2806 
2807 	status = bfa_ioc_smem_read(ioc, buf, loff, dlen);
2808 
2809 	if (status != BFA_STATUS_OK) {
2810 		*offset = *buflen = 0;
2811 		return status;
2812 	}
2813 
2814 	*offset += dlen;
2815 
2816 	if (*offset >= smem_len)
2817 		*offset = 0;
2818 
2819 	*buflen = dlen;
2820 
2821 	return status;
2822 }
2823 
2824 /*
2825  * Firmware statistics
2826  */
2827 bfa_status_t
2828 bfa_ioc_fw_stats_get(struct bfa_ioc_s *ioc, void *stats)
2829 {
2830 	u32 loff = BFI_IOC_FWSTATS_OFF + \
2831 		BFI_IOC_FWSTATS_SZ * (bfa_ioc_portid(ioc));
2832 	int tlen;
2833 	bfa_status_t status;
2834 
2835 	if (ioc->stats_busy) {
2836 		bfa_trc(ioc, ioc->stats_busy);
2837 		return BFA_STATUS_DEVBUSY;
2838 	}
2839 	ioc->stats_busy = BFA_TRUE;
2840 
2841 	tlen = sizeof(struct bfa_fw_stats_s);
2842 	status = bfa_ioc_smem_read(ioc, stats, loff, tlen);
2843 
2844 	ioc->stats_busy = BFA_FALSE;
2845 	return status;
2846 }
2847 
2848 bfa_status_t
2849 bfa_ioc_fw_stats_clear(struct bfa_ioc_s *ioc)
2850 {
2851 	u32 loff = BFI_IOC_FWSTATS_OFF + \
2852 		BFI_IOC_FWSTATS_SZ * (bfa_ioc_portid(ioc));
2853 	int tlen;
2854 	bfa_status_t status;
2855 
2856 	if (ioc->stats_busy) {
2857 		bfa_trc(ioc, ioc->stats_busy);
2858 		return BFA_STATUS_DEVBUSY;
2859 	}
2860 	ioc->stats_busy = BFA_TRUE;
2861 
2862 	tlen = sizeof(struct bfa_fw_stats_s);
2863 	status = bfa_ioc_smem_clr(ioc, loff, tlen);
2864 
2865 	ioc->stats_busy = BFA_FALSE;
2866 	return status;
2867 }
2868 
2869 /*
2870  * Save firmware trace if configured.
2871  */
2872 void
2873 bfa_ioc_debug_save_ftrc(struct bfa_ioc_s *ioc)
2874 {
2875 	int		tlen;
2876 
2877 	if (ioc->dbg_fwsave_once) {
2878 		ioc->dbg_fwsave_once = BFA_FALSE;
2879 		if (ioc->dbg_fwsave_len) {
2880 			tlen = ioc->dbg_fwsave_len;
2881 			bfa_ioc_debug_fwtrc(ioc, ioc->dbg_fwsave, &tlen);
2882 		}
2883 	}
2884 }
2885 
2886 /*
2887  * Firmware failure detected. Start recovery actions.
2888  */
2889 static void
2890 bfa_ioc_recover(struct bfa_ioc_s *ioc)
2891 {
2892 	bfa_ioc_stats(ioc, ioc_hbfails);
2893 	ioc->stats.hb_count = ioc->hb_count;
2894 	bfa_fsm_send_event(ioc, IOC_E_HBFAIL);
2895 }
2896 
2897 /*
2898  *  BFA IOC PF private functions
2899  */
2900 static void
2901 bfa_iocpf_timeout(void *ioc_arg)
2902 {
2903 	struct bfa_ioc_s  *ioc = (struct bfa_ioc_s *) ioc_arg;
2904 
2905 	bfa_trc(ioc, 0);
2906 	bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_TIMEOUT);
2907 }
2908 
2909 static void
2910 bfa_iocpf_sem_timeout(void *ioc_arg)
2911 {
2912 	struct bfa_ioc_s  *ioc = (struct bfa_ioc_s *) ioc_arg;
2913 
2914 	bfa_ioc_hw_sem_get(ioc);
2915 }
2916 
2917 static void
2918 bfa_ioc_poll_fwinit(struct bfa_ioc_s *ioc)
2919 {
2920 	u32 fwstate = readl(ioc->ioc_regs.ioc_fwstate);
2921 
2922 	bfa_trc(ioc, fwstate);
2923 
2924 	if (fwstate == BFI_IOC_DISABLED) {
2925 		bfa_fsm_send_event(&ioc->iocpf, IOCPF_E_FWREADY);
2926 		return;
2927 	}
2928 
2929 	if (ioc->iocpf.poll_time >= (3 * BFA_IOC_TOV))
2930 		bfa_iocpf_timeout(ioc);
2931 	else {
2932 		ioc->iocpf.poll_time += BFA_IOC_POLL_TOV;
2933 		bfa_iocpf_poll_timer_start(ioc);
2934 	}
2935 }
2936 
2937 static void
2938 bfa_iocpf_poll_timeout(void *ioc_arg)
2939 {
2940 	struct bfa_ioc_s *ioc = (struct bfa_ioc_s *) ioc_arg;
2941 
2942 	bfa_ioc_poll_fwinit(ioc);
2943 }
2944 
2945 /*
2946  *  bfa timer function
2947  */
2948 void
2949 bfa_timer_beat(struct bfa_timer_mod_s *mod)
2950 {
2951 	struct list_head *qh = &mod->timer_q;
2952 	struct list_head *qe, *qe_next;
2953 	struct bfa_timer_s *elem;
2954 	struct list_head timedout_q;
2955 
2956 	INIT_LIST_HEAD(&timedout_q);
2957 
2958 	qe = bfa_q_next(qh);
2959 
2960 	while (qe != qh) {
2961 		qe_next = bfa_q_next(qe);
2962 
2963 		elem = (struct bfa_timer_s *) qe;
2964 		if (elem->timeout <= BFA_TIMER_FREQ) {
2965 			elem->timeout = 0;
2966 			list_del(&elem->qe);
2967 			list_add_tail(&elem->qe, &timedout_q);
2968 		} else {
2969 			elem->timeout -= BFA_TIMER_FREQ;
2970 		}
2971 
2972 		qe = qe_next;	/* go to next elem */
2973 	}
2974 
2975 	/*
2976 	 * Pop all the timeout entries
2977 	 */
2978 	while (!list_empty(&timedout_q)) {
2979 		bfa_q_deq(&timedout_q, &elem);
2980 		elem->timercb(elem->arg);
2981 	}
2982 }
2983 
2984 /*
2985  * Should be called with lock protection
2986  */
2987 void
2988 bfa_timer_begin(struct bfa_timer_mod_s *mod, struct bfa_timer_s *timer,
2989 		    void (*timercb) (void *), void *arg, unsigned int timeout)
2990 {
2991 
2992 	WARN_ON(timercb == NULL);
2993 	WARN_ON(bfa_q_is_on_q(&mod->timer_q, timer));
2994 
2995 	timer->timeout = timeout;
2996 	timer->timercb = timercb;
2997 	timer->arg = arg;
2998 
2999 	list_add_tail(&timer->qe, &mod->timer_q);
3000 }
3001 
3002 /*
3003  * Should be called with lock protection
3004  */
3005 void
3006 bfa_timer_stop(struct bfa_timer_s *timer)
3007 {
3008 	WARN_ON(list_empty(&timer->qe));
3009 
3010 	list_del(&timer->qe);
3011 }
3012 
3013 /*
3014  *	ASIC block related
3015  */
3016 static void
3017 bfa_ablk_config_swap(struct bfa_ablk_cfg_s *cfg)
3018 {
3019 	struct bfa_ablk_cfg_inst_s *cfg_inst;
3020 	int i, j;
3021 	u16	be16;
3022 
3023 	for (i = 0; i < BFA_ABLK_MAX; i++) {
3024 		cfg_inst = &cfg->inst[i];
3025 		for (j = 0; j < BFA_ABLK_MAX_PFS; j++) {
3026 			be16 = cfg_inst->pf_cfg[j].pers;
3027 			cfg_inst->pf_cfg[j].pers = be16_to_cpu(be16);
3028 			be16 = cfg_inst->pf_cfg[j].num_qpairs;
3029 			cfg_inst->pf_cfg[j].num_qpairs = be16_to_cpu(be16);
3030 			be16 = cfg_inst->pf_cfg[j].num_vectors;
3031 			cfg_inst->pf_cfg[j].num_vectors = be16_to_cpu(be16);
3032 			be16 = cfg_inst->pf_cfg[j].bw_min;
3033 			cfg_inst->pf_cfg[j].bw_min = be16_to_cpu(be16);
3034 			be16 = cfg_inst->pf_cfg[j].bw_max;
3035 			cfg_inst->pf_cfg[j].bw_max = be16_to_cpu(be16);
3036 		}
3037 	}
3038 }
3039 
3040 static void
3041 bfa_ablk_isr(void *cbarg, struct bfi_mbmsg_s *msg)
3042 {
3043 	struct bfa_ablk_s *ablk = (struct bfa_ablk_s *)cbarg;
3044 	struct bfi_ablk_i2h_rsp_s *rsp = (struct bfi_ablk_i2h_rsp_s *)msg;
3045 	bfa_ablk_cbfn_t cbfn;
3046 
3047 	WARN_ON(msg->mh.msg_class != BFI_MC_ABLK);
3048 	bfa_trc(ablk->ioc, msg->mh.msg_id);
3049 
3050 	switch (msg->mh.msg_id) {
3051 	case BFI_ABLK_I2H_QUERY:
3052 		if (rsp->status == BFA_STATUS_OK) {
3053 			memcpy(ablk->cfg, ablk->dma_addr.kva,
3054 				sizeof(struct bfa_ablk_cfg_s));
3055 			bfa_ablk_config_swap(ablk->cfg);
3056 			ablk->cfg = NULL;
3057 		}
3058 		break;
3059 
3060 	case BFI_ABLK_I2H_ADPT_CONFIG:
3061 	case BFI_ABLK_I2H_PORT_CONFIG:
3062 		/* update config port mode */
3063 		ablk->ioc->port_mode_cfg = rsp->port_mode;
3064 
3065 	case BFI_ABLK_I2H_PF_DELETE:
3066 	case BFI_ABLK_I2H_PF_UPDATE:
3067 	case BFI_ABLK_I2H_OPTROM_ENABLE:
3068 	case BFI_ABLK_I2H_OPTROM_DISABLE:
3069 		/* No-op */
3070 		break;
3071 
3072 	case BFI_ABLK_I2H_PF_CREATE:
3073 		*(ablk->pcifn) = rsp->pcifn;
3074 		ablk->pcifn = NULL;
3075 		break;
3076 
3077 	default:
3078 		WARN_ON(1);
3079 	}
3080 
3081 	ablk->busy = BFA_FALSE;
3082 	if (ablk->cbfn) {
3083 		cbfn = ablk->cbfn;
3084 		ablk->cbfn = NULL;
3085 		cbfn(ablk->cbarg, rsp->status);
3086 	}
3087 }
3088 
3089 static void
3090 bfa_ablk_notify(void *cbarg, enum bfa_ioc_event_e event)
3091 {
3092 	struct bfa_ablk_s *ablk = (struct bfa_ablk_s *)cbarg;
3093 
3094 	bfa_trc(ablk->ioc, event);
3095 
3096 	switch (event) {
3097 	case BFA_IOC_E_ENABLED:
3098 		WARN_ON(ablk->busy != BFA_FALSE);
3099 		break;
3100 
3101 	case BFA_IOC_E_DISABLED:
3102 	case BFA_IOC_E_FAILED:
3103 		/* Fail any pending requests */
3104 		ablk->pcifn = NULL;
3105 		if (ablk->busy) {
3106 			if (ablk->cbfn)
3107 				ablk->cbfn(ablk->cbarg, BFA_STATUS_FAILED);
3108 			ablk->cbfn = NULL;
3109 			ablk->busy = BFA_FALSE;
3110 		}
3111 		break;
3112 
3113 	default:
3114 		WARN_ON(1);
3115 		break;
3116 	}
3117 }
3118 
3119 u32
3120 bfa_ablk_meminfo(void)
3121 {
3122 	return BFA_ROUNDUP(sizeof(struct bfa_ablk_cfg_s), BFA_DMA_ALIGN_SZ);
3123 }
3124 
3125 void
3126 bfa_ablk_memclaim(struct bfa_ablk_s *ablk, u8 *dma_kva, u64 dma_pa)
3127 {
3128 	ablk->dma_addr.kva = dma_kva;
3129 	ablk->dma_addr.pa  = dma_pa;
3130 }
3131 
3132 void
3133 bfa_ablk_attach(struct bfa_ablk_s *ablk, struct bfa_ioc_s *ioc)
3134 {
3135 	ablk->ioc = ioc;
3136 
3137 	bfa_ioc_mbox_regisr(ablk->ioc, BFI_MC_ABLK, bfa_ablk_isr, ablk);
3138 	bfa_q_qe_init(&ablk->ioc_notify);
3139 	bfa_ioc_notify_init(&ablk->ioc_notify, bfa_ablk_notify, ablk);
3140 	list_add_tail(&ablk->ioc_notify.qe, &ablk->ioc->notify_q);
3141 }
3142 
3143 bfa_status_t
3144 bfa_ablk_query(struct bfa_ablk_s *ablk, struct bfa_ablk_cfg_s *ablk_cfg,
3145 		bfa_ablk_cbfn_t cbfn, void *cbarg)
3146 {
3147 	struct bfi_ablk_h2i_query_s *m;
3148 
3149 	WARN_ON(!ablk_cfg);
3150 
3151 	if (!bfa_ioc_is_operational(ablk->ioc)) {
3152 		bfa_trc(ablk->ioc, BFA_STATUS_IOC_FAILURE);
3153 		return BFA_STATUS_IOC_FAILURE;
3154 	}
3155 
3156 	if (ablk->busy) {
3157 		bfa_trc(ablk->ioc, BFA_STATUS_DEVBUSY);
3158 		return  BFA_STATUS_DEVBUSY;
3159 	}
3160 
3161 	ablk->cfg = ablk_cfg;
3162 	ablk->cbfn  = cbfn;
3163 	ablk->cbarg = cbarg;
3164 	ablk->busy  = BFA_TRUE;
3165 
3166 	m = (struct bfi_ablk_h2i_query_s *)ablk->mb.msg;
3167 	bfi_h2i_set(m->mh, BFI_MC_ABLK, BFI_ABLK_H2I_QUERY,
3168 		    bfa_ioc_portid(ablk->ioc));
3169 	bfa_dma_be_addr_set(m->addr, ablk->dma_addr.pa);
3170 	bfa_ioc_mbox_queue(ablk->ioc, &ablk->mb);
3171 
3172 	return BFA_STATUS_OK;
3173 }
3174 
3175 bfa_status_t
3176 bfa_ablk_pf_create(struct bfa_ablk_s *ablk, u16 *pcifn,
3177 		u8 port, enum bfi_pcifn_class personality,
3178 		u16 bw_min, u16 bw_max,
3179 		bfa_ablk_cbfn_t cbfn, void *cbarg)
3180 {
3181 	struct bfi_ablk_h2i_pf_req_s *m;
3182 
3183 	if (!bfa_ioc_is_operational(ablk->ioc)) {
3184 		bfa_trc(ablk->ioc, BFA_STATUS_IOC_FAILURE);
3185 		return BFA_STATUS_IOC_FAILURE;
3186 	}
3187 
3188 	if (ablk->busy) {
3189 		bfa_trc(ablk->ioc, BFA_STATUS_DEVBUSY);
3190 		return  BFA_STATUS_DEVBUSY;
3191 	}
3192 
3193 	ablk->pcifn = pcifn;
3194 	ablk->cbfn = cbfn;
3195 	ablk->cbarg = cbarg;
3196 	ablk->busy  = BFA_TRUE;
3197 
3198 	m = (struct bfi_ablk_h2i_pf_req_s *)ablk->mb.msg;
3199 	bfi_h2i_set(m->mh, BFI_MC_ABLK, BFI_ABLK_H2I_PF_CREATE,
3200 		    bfa_ioc_portid(ablk->ioc));
3201 	m->pers = cpu_to_be16((u16)personality);
3202 	m->bw_min = cpu_to_be16(bw_min);
3203 	m->bw_max = cpu_to_be16(bw_max);
3204 	m->port = port;
3205 	bfa_ioc_mbox_queue(ablk->ioc, &ablk->mb);
3206 
3207 	return BFA_STATUS_OK;
3208 }
3209 
3210 bfa_status_t
3211 bfa_ablk_pf_delete(struct bfa_ablk_s *ablk, int pcifn,
3212 		bfa_ablk_cbfn_t cbfn, void *cbarg)
3213 {
3214 	struct bfi_ablk_h2i_pf_req_s *m;
3215 
3216 	if (!bfa_ioc_is_operational(ablk->ioc)) {
3217 		bfa_trc(ablk->ioc, BFA_STATUS_IOC_FAILURE);
3218 		return BFA_STATUS_IOC_FAILURE;
3219 	}
3220 
3221 	if (ablk->busy) {
3222 		bfa_trc(ablk->ioc, BFA_STATUS_DEVBUSY);
3223 		return  BFA_STATUS_DEVBUSY;
3224 	}
3225 
3226 	ablk->cbfn  = cbfn;
3227 	ablk->cbarg = cbarg;
3228 	ablk->busy  = BFA_TRUE;
3229 
3230 	m = (struct bfi_ablk_h2i_pf_req_s *)ablk->mb.msg;
3231 	bfi_h2i_set(m->mh, BFI_MC_ABLK, BFI_ABLK_H2I_PF_DELETE,
3232 		    bfa_ioc_portid(ablk->ioc));
3233 	m->pcifn = (u8)pcifn;
3234 	bfa_ioc_mbox_queue(ablk->ioc, &ablk->mb);
3235 
3236 	return BFA_STATUS_OK;
3237 }
3238 
3239 bfa_status_t
3240 bfa_ablk_adapter_config(struct bfa_ablk_s *ablk, enum bfa_mode_s mode,
3241 		int max_pf, int max_vf, bfa_ablk_cbfn_t cbfn, void *cbarg)
3242 {
3243 	struct bfi_ablk_h2i_cfg_req_s *m;
3244 
3245 	if (!bfa_ioc_is_operational(ablk->ioc)) {
3246 		bfa_trc(ablk->ioc, BFA_STATUS_IOC_FAILURE);
3247 		return BFA_STATUS_IOC_FAILURE;
3248 	}
3249 
3250 	if (ablk->busy) {
3251 		bfa_trc(ablk->ioc, BFA_STATUS_DEVBUSY);
3252 		return  BFA_STATUS_DEVBUSY;
3253 	}
3254 
3255 	ablk->cbfn  = cbfn;
3256 	ablk->cbarg = cbarg;
3257 	ablk->busy  = BFA_TRUE;
3258 
3259 	m = (struct bfi_ablk_h2i_cfg_req_s *)ablk->mb.msg;
3260 	bfi_h2i_set(m->mh, BFI_MC_ABLK, BFI_ABLK_H2I_ADPT_CONFIG,
3261 		    bfa_ioc_portid(ablk->ioc));
3262 	m->mode = (u8)mode;
3263 	m->max_pf = (u8)max_pf;
3264 	m->max_vf = (u8)max_vf;
3265 	bfa_ioc_mbox_queue(ablk->ioc, &ablk->mb);
3266 
3267 	return BFA_STATUS_OK;
3268 }
3269 
3270 bfa_status_t
3271 bfa_ablk_port_config(struct bfa_ablk_s *ablk, int port, enum bfa_mode_s mode,
3272 		int max_pf, int max_vf, bfa_ablk_cbfn_t cbfn, void *cbarg)
3273 {
3274 	struct bfi_ablk_h2i_cfg_req_s *m;
3275 
3276 	if (!bfa_ioc_is_operational(ablk->ioc)) {
3277 		bfa_trc(ablk->ioc, BFA_STATUS_IOC_FAILURE);
3278 		return BFA_STATUS_IOC_FAILURE;
3279 	}
3280 
3281 	if (ablk->busy) {
3282 		bfa_trc(ablk->ioc, BFA_STATUS_DEVBUSY);
3283 		return  BFA_STATUS_DEVBUSY;
3284 	}
3285 
3286 	ablk->cbfn  = cbfn;
3287 	ablk->cbarg = cbarg;
3288 	ablk->busy  = BFA_TRUE;
3289 
3290 	m = (struct bfi_ablk_h2i_cfg_req_s *)ablk->mb.msg;
3291 	bfi_h2i_set(m->mh, BFI_MC_ABLK, BFI_ABLK_H2I_PORT_CONFIG,
3292 		bfa_ioc_portid(ablk->ioc));
3293 	m->port = (u8)port;
3294 	m->mode = (u8)mode;
3295 	m->max_pf = (u8)max_pf;
3296 	m->max_vf = (u8)max_vf;
3297 	bfa_ioc_mbox_queue(ablk->ioc, &ablk->mb);
3298 
3299 	return BFA_STATUS_OK;
3300 }
3301 
3302 bfa_status_t
3303 bfa_ablk_pf_update(struct bfa_ablk_s *ablk, int pcifn, u16 bw_min,
3304 		   u16 bw_max, bfa_ablk_cbfn_t cbfn, void *cbarg)
3305 {
3306 	struct bfi_ablk_h2i_pf_req_s *m;
3307 
3308 	if (!bfa_ioc_is_operational(ablk->ioc)) {
3309 		bfa_trc(ablk->ioc, BFA_STATUS_IOC_FAILURE);
3310 		return BFA_STATUS_IOC_FAILURE;
3311 	}
3312 
3313 	if (ablk->busy) {
3314 		bfa_trc(ablk->ioc, BFA_STATUS_DEVBUSY);
3315 		return  BFA_STATUS_DEVBUSY;
3316 	}
3317 
3318 	ablk->cbfn  = cbfn;
3319 	ablk->cbarg = cbarg;
3320 	ablk->busy  = BFA_TRUE;
3321 
3322 	m = (struct bfi_ablk_h2i_pf_req_s *)ablk->mb.msg;
3323 	bfi_h2i_set(m->mh, BFI_MC_ABLK, BFI_ABLK_H2I_PF_UPDATE,
3324 		bfa_ioc_portid(ablk->ioc));
3325 	m->pcifn = (u8)pcifn;
3326 	m->bw_min = cpu_to_be16(bw_min);
3327 	m->bw_max = cpu_to_be16(bw_max);
3328 	bfa_ioc_mbox_queue(ablk->ioc, &ablk->mb);
3329 
3330 	return BFA_STATUS_OK;
3331 }
3332 
3333 bfa_status_t
3334 bfa_ablk_optrom_en(struct bfa_ablk_s *ablk, bfa_ablk_cbfn_t cbfn, void *cbarg)
3335 {
3336 	struct bfi_ablk_h2i_optrom_s *m;
3337 
3338 	if (!bfa_ioc_is_operational(ablk->ioc)) {
3339 		bfa_trc(ablk->ioc, BFA_STATUS_IOC_FAILURE);
3340 		return BFA_STATUS_IOC_FAILURE;
3341 	}
3342 
3343 	if (ablk->busy) {
3344 		bfa_trc(ablk->ioc, BFA_STATUS_DEVBUSY);
3345 		return  BFA_STATUS_DEVBUSY;
3346 	}
3347 
3348 	ablk->cbfn  = cbfn;
3349 	ablk->cbarg = cbarg;
3350 	ablk->busy  = BFA_TRUE;
3351 
3352 	m = (struct bfi_ablk_h2i_optrom_s *)ablk->mb.msg;
3353 	bfi_h2i_set(m->mh, BFI_MC_ABLK, BFI_ABLK_H2I_OPTROM_ENABLE,
3354 		bfa_ioc_portid(ablk->ioc));
3355 	bfa_ioc_mbox_queue(ablk->ioc, &ablk->mb);
3356 
3357 	return BFA_STATUS_OK;
3358 }
3359 
3360 bfa_status_t
3361 bfa_ablk_optrom_dis(struct bfa_ablk_s *ablk, bfa_ablk_cbfn_t cbfn, void *cbarg)
3362 {
3363 	struct bfi_ablk_h2i_optrom_s *m;
3364 
3365 	if (!bfa_ioc_is_operational(ablk->ioc)) {
3366 		bfa_trc(ablk->ioc, BFA_STATUS_IOC_FAILURE);
3367 		return BFA_STATUS_IOC_FAILURE;
3368 	}
3369 
3370 	if (ablk->busy) {
3371 		bfa_trc(ablk->ioc, BFA_STATUS_DEVBUSY);
3372 		return  BFA_STATUS_DEVBUSY;
3373 	}
3374 
3375 	ablk->cbfn  = cbfn;
3376 	ablk->cbarg = cbarg;
3377 	ablk->busy  = BFA_TRUE;
3378 
3379 	m = (struct bfi_ablk_h2i_optrom_s *)ablk->mb.msg;
3380 	bfi_h2i_set(m->mh, BFI_MC_ABLK, BFI_ABLK_H2I_OPTROM_DISABLE,
3381 		bfa_ioc_portid(ablk->ioc));
3382 	bfa_ioc_mbox_queue(ablk->ioc, &ablk->mb);
3383 
3384 	return BFA_STATUS_OK;
3385 }
3386 
3387 /*
3388  *	SFP module specific
3389  */
3390 
3391 /* forward declarations */
3392 static void bfa_sfp_getdata_send(struct bfa_sfp_s *sfp);
3393 static void bfa_sfp_media_get(struct bfa_sfp_s *sfp);
3394 static bfa_status_t bfa_sfp_speed_valid(struct bfa_sfp_s *sfp,
3395 				enum bfa_port_speed portspeed);
3396 
3397 static void
3398 bfa_cb_sfp_show(struct bfa_sfp_s *sfp)
3399 {
3400 	bfa_trc(sfp, sfp->lock);
3401 	if (sfp->cbfn)
3402 		sfp->cbfn(sfp->cbarg, sfp->status);
3403 	sfp->lock = 0;
3404 	sfp->cbfn = NULL;
3405 }
3406 
3407 static void
3408 bfa_cb_sfp_state_query(struct bfa_sfp_s *sfp)
3409 {
3410 	bfa_trc(sfp, sfp->portspeed);
3411 	if (sfp->media) {
3412 		bfa_sfp_media_get(sfp);
3413 		if (sfp->state_query_cbfn)
3414 			sfp->state_query_cbfn(sfp->state_query_cbarg,
3415 					sfp->status);
3416 			sfp->media = NULL;
3417 		}
3418 
3419 		if (sfp->portspeed) {
3420 			sfp->status = bfa_sfp_speed_valid(sfp, sfp->portspeed);
3421 			if (sfp->state_query_cbfn)
3422 				sfp->state_query_cbfn(sfp->state_query_cbarg,
3423 						sfp->status);
3424 				sfp->portspeed = BFA_PORT_SPEED_UNKNOWN;
3425 		}
3426 
3427 		sfp->state_query_lock = 0;
3428 		sfp->state_query_cbfn = NULL;
3429 }
3430 
3431 /*
3432  *	IOC event handler.
3433  */
3434 static void
3435 bfa_sfp_notify(void *sfp_arg, enum bfa_ioc_event_e event)
3436 {
3437 	struct bfa_sfp_s *sfp = sfp_arg;
3438 
3439 	bfa_trc(sfp, event);
3440 	bfa_trc(sfp, sfp->lock);
3441 	bfa_trc(sfp, sfp->state_query_lock);
3442 
3443 	switch (event) {
3444 	case BFA_IOC_E_DISABLED:
3445 	case BFA_IOC_E_FAILED:
3446 		if (sfp->lock) {
3447 			sfp->status = BFA_STATUS_IOC_FAILURE;
3448 			bfa_cb_sfp_show(sfp);
3449 		}
3450 
3451 		if (sfp->state_query_lock) {
3452 			sfp->status = BFA_STATUS_IOC_FAILURE;
3453 			bfa_cb_sfp_state_query(sfp);
3454 		}
3455 		break;
3456 
3457 	default:
3458 		break;
3459 	}
3460 }
3461 
3462 /*
3463  * SFP's State Change Notification post to AEN
3464  */
3465 static void
3466 bfa_sfp_scn_aen_post(struct bfa_sfp_s *sfp, struct bfi_sfp_scn_s *rsp)
3467 {
3468 	struct bfad_s *bfad = (struct bfad_s *)sfp->ioc->bfa->bfad;
3469 	struct bfa_aen_entry_s  *aen_entry;
3470 	enum bfa_port_aen_event aen_evt = 0;
3471 
3472 	bfa_trc(sfp, (((u64)rsp->pomlvl) << 16) | (((u64)rsp->sfpid) << 8) |
3473 		      ((u64)rsp->event));
3474 
3475 	bfad_get_aen_entry(bfad, aen_entry);
3476 	if (!aen_entry)
3477 		return;
3478 
3479 	aen_entry->aen_data.port.ioc_type = bfa_ioc_get_type(sfp->ioc);
3480 	aen_entry->aen_data.port.pwwn = sfp->ioc->attr->pwwn;
3481 	aen_entry->aen_data.port.mac = bfa_ioc_get_mac(sfp->ioc);
3482 
3483 	switch (rsp->event) {
3484 	case BFA_SFP_SCN_INSERTED:
3485 		aen_evt = BFA_PORT_AEN_SFP_INSERT;
3486 		break;
3487 	case BFA_SFP_SCN_REMOVED:
3488 		aen_evt = BFA_PORT_AEN_SFP_REMOVE;
3489 		break;
3490 	case BFA_SFP_SCN_FAILED:
3491 		aen_evt = BFA_PORT_AEN_SFP_ACCESS_ERROR;
3492 		break;
3493 	case BFA_SFP_SCN_UNSUPPORT:
3494 		aen_evt = BFA_PORT_AEN_SFP_UNSUPPORT;
3495 		break;
3496 	case BFA_SFP_SCN_POM:
3497 		aen_evt = BFA_PORT_AEN_SFP_POM;
3498 		aen_entry->aen_data.port.level = rsp->pomlvl;
3499 		break;
3500 	default:
3501 		bfa_trc(sfp, rsp->event);
3502 		WARN_ON(1);
3503 	}
3504 
3505 	/* Send the AEN notification */
3506 	bfad_im_post_vendor_event(aen_entry, bfad, ++sfp->ioc->ioc_aen_seq,
3507 				  BFA_AEN_CAT_PORT, aen_evt);
3508 }
3509 
3510 /*
3511  *	SFP get data send
3512  */
3513 static void
3514 bfa_sfp_getdata_send(struct bfa_sfp_s *sfp)
3515 {
3516 	struct bfi_sfp_req_s *req = (struct bfi_sfp_req_s *)sfp->mbcmd.msg;
3517 
3518 	bfa_trc(sfp, req->memtype);
3519 
3520 	/* build host command */
3521 	bfi_h2i_set(req->mh, BFI_MC_SFP, BFI_SFP_H2I_SHOW,
3522 			bfa_ioc_portid(sfp->ioc));
3523 
3524 	/* send mbox cmd */
3525 	bfa_ioc_mbox_queue(sfp->ioc, &sfp->mbcmd);
3526 }
3527 
3528 /*
3529  *	SFP is valid, read sfp data
3530  */
3531 static void
3532 bfa_sfp_getdata(struct bfa_sfp_s *sfp, enum bfi_sfp_mem_e memtype)
3533 {
3534 	struct bfi_sfp_req_s *req = (struct bfi_sfp_req_s *)sfp->mbcmd.msg;
3535 
3536 	WARN_ON(sfp->lock != 0);
3537 	bfa_trc(sfp, sfp->state);
3538 
3539 	sfp->lock = 1;
3540 	sfp->memtype = memtype;
3541 	req->memtype = memtype;
3542 
3543 	/* Setup SG list */
3544 	bfa_alen_set(&req->alen, sizeof(struct sfp_mem_s), sfp->dbuf_pa);
3545 
3546 	bfa_sfp_getdata_send(sfp);
3547 }
3548 
3549 /*
3550  *	SFP scn handler
3551  */
3552 static void
3553 bfa_sfp_scn(struct bfa_sfp_s *sfp, struct bfi_mbmsg_s *msg)
3554 {
3555 	struct bfi_sfp_scn_s *rsp = (struct bfi_sfp_scn_s *) msg;
3556 
3557 	switch (rsp->event) {
3558 	case BFA_SFP_SCN_INSERTED:
3559 		sfp->state = BFA_SFP_STATE_INSERTED;
3560 		sfp->data_valid = 0;
3561 		bfa_sfp_scn_aen_post(sfp, rsp);
3562 		break;
3563 	case BFA_SFP_SCN_REMOVED:
3564 		sfp->state = BFA_SFP_STATE_REMOVED;
3565 		sfp->data_valid = 0;
3566 		bfa_sfp_scn_aen_post(sfp, rsp);
3567 		 break;
3568 	case BFA_SFP_SCN_FAILED:
3569 		sfp->state = BFA_SFP_STATE_FAILED;
3570 		sfp->data_valid = 0;
3571 		bfa_sfp_scn_aen_post(sfp, rsp);
3572 		break;
3573 	case BFA_SFP_SCN_UNSUPPORT:
3574 		sfp->state = BFA_SFP_STATE_UNSUPPORT;
3575 		bfa_sfp_scn_aen_post(sfp, rsp);
3576 		if (!sfp->lock)
3577 			bfa_sfp_getdata(sfp, BFI_SFP_MEM_ALL);
3578 		break;
3579 	case BFA_SFP_SCN_POM:
3580 		bfa_sfp_scn_aen_post(sfp, rsp);
3581 		break;
3582 	case BFA_SFP_SCN_VALID:
3583 		sfp->state = BFA_SFP_STATE_VALID;
3584 		if (!sfp->lock)
3585 			bfa_sfp_getdata(sfp, BFI_SFP_MEM_ALL);
3586 		break;
3587 	default:
3588 		bfa_trc(sfp, rsp->event);
3589 		WARN_ON(1);
3590 	}
3591 }
3592 
3593 /*
3594  * SFP show complete
3595  */
3596 static void
3597 bfa_sfp_show_comp(struct bfa_sfp_s *sfp, struct bfi_mbmsg_s *msg)
3598 {
3599 	struct bfi_sfp_rsp_s *rsp = (struct bfi_sfp_rsp_s *) msg;
3600 
3601 	if (!sfp->lock) {
3602 		/*
3603 		 * receiving response after ioc failure
3604 		 */
3605 		bfa_trc(sfp, sfp->lock);
3606 		return;
3607 	}
3608 
3609 	bfa_trc(sfp, rsp->status);
3610 	if (rsp->status == BFA_STATUS_OK) {
3611 		sfp->data_valid = 1;
3612 		if (sfp->state == BFA_SFP_STATE_VALID)
3613 			sfp->status = BFA_STATUS_OK;
3614 		else if (sfp->state == BFA_SFP_STATE_UNSUPPORT)
3615 			sfp->status = BFA_STATUS_SFP_UNSUPP;
3616 		else
3617 			bfa_trc(sfp, sfp->state);
3618 	} else {
3619 		sfp->data_valid = 0;
3620 		sfp->status = rsp->status;
3621 		/* sfpshow shouldn't change sfp state */
3622 	}
3623 
3624 	bfa_trc(sfp, sfp->memtype);
3625 	if (sfp->memtype == BFI_SFP_MEM_DIAGEXT) {
3626 		bfa_trc(sfp, sfp->data_valid);
3627 		if (sfp->data_valid) {
3628 			u32	size = sizeof(struct sfp_mem_s);
3629 			u8 *des = (u8 *) &(sfp->sfpmem->srlid_base);
3630 			memcpy(des, sfp->dbuf_kva, size);
3631 		}
3632 		/*
3633 		 * Queue completion callback.
3634 		 */
3635 		bfa_cb_sfp_show(sfp);
3636 	} else
3637 		sfp->lock = 0;
3638 
3639 	bfa_trc(sfp, sfp->state_query_lock);
3640 	if (sfp->state_query_lock) {
3641 		sfp->state = rsp->state;
3642 		/* Complete callback */
3643 		bfa_cb_sfp_state_query(sfp);
3644 	}
3645 }
3646 
3647 /*
3648  *	SFP query fw sfp state
3649  */
3650 static void
3651 bfa_sfp_state_query(struct bfa_sfp_s *sfp)
3652 {
3653 	struct bfi_sfp_req_s *req = (struct bfi_sfp_req_s *)sfp->mbcmd.msg;
3654 
3655 	/* Should not be doing query if not in _INIT state */
3656 	WARN_ON(sfp->state != BFA_SFP_STATE_INIT);
3657 	WARN_ON(sfp->state_query_lock != 0);
3658 	bfa_trc(sfp, sfp->state);
3659 
3660 	sfp->state_query_lock = 1;
3661 	req->memtype = 0;
3662 
3663 	if (!sfp->lock)
3664 		bfa_sfp_getdata(sfp, BFI_SFP_MEM_ALL);
3665 }
3666 
3667 static void
3668 bfa_sfp_media_get(struct bfa_sfp_s *sfp)
3669 {
3670 	enum bfa_defs_sfp_media_e *media = sfp->media;
3671 
3672 	*media = BFA_SFP_MEDIA_UNKNOWN;
3673 
3674 	if (sfp->state == BFA_SFP_STATE_UNSUPPORT)
3675 		*media = BFA_SFP_MEDIA_UNSUPPORT;
3676 	else if (sfp->state == BFA_SFP_STATE_VALID) {
3677 		union sfp_xcvr_e10g_code_u e10g;
3678 		struct sfp_mem_s *sfpmem = (struct sfp_mem_s *)sfp->dbuf_kva;
3679 		u16 xmtr_tech = (sfpmem->srlid_base.xcvr[4] & 0x3) << 7 |
3680 				(sfpmem->srlid_base.xcvr[5] >> 1);
3681 
3682 		e10g.b = sfpmem->srlid_base.xcvr[0];
3683 		bfa_trc(sfp, e10g.b);
3684 		bfa_trc(sfp, xmtr_tech);
3685 		/* check fc transmitter tech */
3686 		if ((xmtr_tech & SFP_XMTR_TECH_CU) ||
3687 		    (xmtr_tech & SFP_XMTR_TECH_CP) ||
3688 		    (xmtr_tech & SFP_XMTR_TECH_CA))
3689 			*media = BFA_SFP_MEDIA_CU;
3690 		else if ((xmtr_tech & SFP_XMTR_TECH_EL_INTRA) ||
3691 			 (xmtr_tech & SFP_XMTR_TECH_EL_INTER))
3692 			*media = BFA_SFP_MEDIA_EL;
3693 		else if ((xmtr_tech & SFP_XMTR_TECH_LL) ||
3694 			 (xmtr_tech & SFP_XMTR_TECH_LC))
3695 			*media = BFA_SFP_MEDIA_LW;
3696 		else if ((xmtr_tech & SFP_XMTR_TECH_SL) ||
3697 			 (xmtr_tech & SFP_XMTR_TECH_SN) ||
3698 			 (xmtr_tech & SFP_XMTR_TECH_SA))
3699 			*media = BFA_SFP_MEDIA_SW;
3700 		/* Check 10G Ethernet Compilance code */
3701 		else if (e10g.r.e10g_sr)
3702 			*media = BFA_SFP_MEDIA_SW;
3703 		else if (e10g.r.e10g_lrm && e10g.r.e10g_lr)
3704 			*media = BFA_SFP_MEDIA_LW;
3705 		else if (e10g.r.e10g_unall)
3706 			*media = BFA_SFP_MEDIA_UNKNOWN;
3707 		else
3708 			bfa_trc(sfp, 0);
3709 	} else
3710 		bfa_trc(sfp, sfp->state);
3711 }
3712 
3713 static bfa_status_t
3714 bfa_sfp_speed_valid(struct bfa_sfp_s *sfp, enum bfa_port_speed portspeed)
3715 {
3716 	struct sfp_mem_s *sfpmem = (struct sfp_mem_s *)sfp->dbuf_kva;
3717 	struct sfp_xcvr_s *xcvr = (struct sfp_xcvr_s *) sfpmem->srlid_base.xcvr;
3718 	union sfp_xcvr_fc3_code_u fc3 = xcvr->fc3;
3719 	union sfp_xcvr_e10g_code_u e10g = xcvr->e10g;
3720 
3721 	if (portspeed == BFA_PORT_SPEED_10GBPS) {
3722 		if (e10g.r.e10g_sr || e10g.r.e10g_lr)
3723 			return BFA_STATUS_OK;
3724 		else {
3725 			bfa_trc(sfp, e10g.b);
3726 			return BFA_STATUS_UNSUPP_SPEED;
3727 		}
3728 	}
3729 	if (((portspeed & BFA_PORT_SPEED_16GBPS) && fc3.r.mb1600) ||
3730 	    ((portspeed & BFA_PORT_SPEED_8GBPS) && fc3.r.mb800) ||
3731 	    ((portspeed & BFA_PORT_SPEED_4GBPS) && fc3.r.mb400) ||
3732 	    ((portspeed & BFA_PORT_SPEED_2GBPS) && fc3.r.mb200) ||
3733 	    ((portspeed & BFA_PORT_SPEED_1GBPS) && fc3.r.mb100))
3734 		return BFA_STATUS_OK;
3735 	else {
3736 		bfa_trc(sfp, portspeed);
3737 		bfa_trc(sfp, fc3.b);
3738 		bfa_trc(sfp, e10g.b);
3739 		return BFA_STATUS_UNSUPP_SPEED;
3740 	}
3741 }
3742 
3743 /*
3744  *	SFP hmbox handler
3745  */
3746 void
3747 bfa_sfp_intr(void *sfparg, struct bfi_mbmsg_s *msg)
3748 {
3749 	struct bfa_sfp_s *sfp = sfparg;
3750 
3751 	switch (msg->mh.msg_id) {
3752 	case BFI_SFP_I2H_SHOW:
3753 		bfa_sfp_show_comp(sfp, msg);
3754 		break;
3755 
3756 	case BFI_SFP_I2H_SCN:
3757 		bfa_sfp_scn(sfp, msg);
3758 		break;
3759 
3760 	default:
3761 		bfa_trc(sfp, msg->mh.msg_id);
3762 		WARN_ON(1);
3763 	}
3764 }
3765 
3766 /*
3767  *	Return DMA memory needed by sfp module.
3768  */
3769 u32
3770 bfa_sfp_meminfo(void)
3771 {
3772 	return BFA_ROUNDUP(sizeof(struct sfp_mem_s), BFA_DMA_ALIGN_SZ);
3773 }
3774 
3775 /*
3776  *	Attach virtual and physical memory for SFP.
3777  */
3778 void
3779 bfa_sfp_attach(struct bfa_sfp_s *sfp, struct bfa_ioc_s *ioc, void *dev,
3780 		struct bfa_trc_mod_s *trcmod)
3781 {
3782 	sfp->dev = dev;
3783 	sfp->ioc = ioc;
3784 	sfp->trcmod = trcmod;
3785 
3786 	sfp->cbfn = NULL;
3787 	sfp->cbarg = NULL;
3788 	sfp->sfpmem = NULL;
3789 	sfp->lock = 0;
3790 	sfp->data_valid = 0;
3791 	sfp->state = BFA_SFP_STATE_INIT;
3792 	sfp->state_query_lock = 0;
3793 	sfp->state_query_cbfn = NULL;
3794 	sfp->state_query_cbarg = NULL;
3795 	sfp->media = NULL;
3796 	sfp->portspeed = BFA_PORT_SPEED_UNKNOWN;
3797 	sfp->is_elb = BFA_FALSE;
3798 
3799 	bfa_ioc_mbox_regisr(sfp->ioc, BFI_MC_SFP, bfa_sfp_intr, sfp);
3800 	bfa_q_qe_init(&sfp->ioc_notify);
3801 	bfa_ioc_notify_init(&sfp->ioc_notify, bfa_sfp_notify, sfp);
3802 	list_add_tail(&sfp->ioc_notify.qe, &sfp->ioc->notify_q);
3803 }
3804 
3805 /*
3806  *	Claim Memory for SFP
3807  */
3808 void
3809 bfa_sfp_memclaim(struct bfa_sfp_s *sfp, u8 *dm_kva, u64 dm_pa)
3810 {
3811 	sfp->dbuf_kva   = dm_kva;
3812 	sfp->dbuf_pa    = dm_pa;
3813 	memset(sfp->dbuf_kva, 0, sizeof(struct sfp_mem_s));
3814 
3815 	dm_kva += BFA_ROUNDUP(sizeof(struct sfp_mem_s), BFA_DMA_ALIGN_SZ);
3816 	dm_pa += BFA_ROUNDUP(sizeof(struct sfp_mem_s), BFA_DMA_ALIGN_SZ);
3817 }
3818 
3819 /*
3820  * Show SFP eeprom content
3821  *
3822  * @param[in] sfp   - bfa sfp module
3823  *
3824  * @param[out] sfpmem - sfp eeprom data
3825  *
3826  */
3827 bfa_status_t
3828 bfa_sfp_show(struct bfa_sfp_s *sfp, struct sfp_mem_s *sfpmem,
3829 		bfa_cb_sfp_t cbfn, void *cbarg)
3830 {
3831 
3832 	if (!bfa_ioc_is_operational(sfp->ioc)) {
3833 		bfa_trc(sfp, 0);
3834 		return BFA_STATUS_IOC_NON_OP;
3835 	}
3836 
3837 	if (sfp->lock) {
3838 		bfa_trc(sfp, 0);
3839 		return BFA_STATUS_DEVBUSY;
3840 	}
3841 
3842 	sfp->cbfn = cbfn;
3843 	sfp->cbarg = cbarg;
3844 	sfp->sfpmem = sfpmem;
3845 
3846 	bfa_sfp_getdata(sfp, BFI_SFP_MEM_DIAGEXT);
3847 	return BFA_STATUS_OK;
3848 }
3849 
3850 /*
3851  * Return SFP Media type
3852  *
3853  * @param[in] sfp   - bfa sfp module
3854  *
3855  * @param[out] media - port speed from user
3856  *
3857  */
3858 bfa_status_t
3859 bfa_sfp_media(struct bfa_sfp_s *sfp, enum bfa_defs_sfp_media_e *media,
3860 		bfa_cb_sfp_t cbfn, void *cbarg)
3861 {
3862 	if (!bfa_ioc_is_operational(sfp->ioc)) {
3863 		bfa_trc(sfp, 0);
3864 		return BFA_STATUS_IOC_NON_OP;
3865 	}
3866 
3867 	sfp->media = media;
3868 	if (sfp->state == BFA_SFP_STATE_INIT) {
3869 		if (sfp->state_query_lock) {
3870 			bfa_trc(sfp, 0);
3871 			return BFA_STATUS_DEVBUSY;
3872 		} else {
3873 			sfp->state_query_cbfn = cbfn;
3874 			sfp->state_query_cbarg = cbarg;
3875 			bfa_sfp_state_query(sfp);
3876 			return BFA_STATUS_SFP_NOT_READY;
3877 		}
3878 	}
3879 
3880 	bfa_sfp_media_get(sfp);
3881 	return BFA_STATUS_OK;
3882 }
3883 
3884 /*
3885  * Check if user set port speed is allowed by the SFP
3886  *
3887  * @param[in] sfp   - bfa sfp module
3888  * @param[in] portspeed - port speed from user
3889  *
3890  */
3891 bfa_status_t
3892 bfa_sfp_speed(struct bfa_sfp_s *sfp, enum bfa_port_speed portspeed,
3893 		bfa_cb_sfp_t cbfn, void *cbarg)
3894 {
3895 	WARN_ON(portspeed == BFA_PORT_SPEED_UNKNOWN);
3896 
3897 	if (!bfa_ioc_is_operational(sfp->ioc))
3898 		return BFA_STATUS_IOC_NON_OP;
3899 
3900 	/* For Mezz card, all speed is allowed */
3901 	if (bfa_mfg_is_mezz(sfp->ioc->attr->card_type))
3902 		return BFA_STATUS_OK;
3903 
3904 	/* Check SFP state */
3905 	sfp->portspeed = portspeed;
3906 	if (sfp->state == BFA_SFP_STATE_INIT) {
3907 		if (sfp->state_query_lock) {
3908 			bfa_trc(sfp, 0);
3909 			return BFA_STATUS_DEVBUSY;
3910 		} else {
3911 			sfp->state_query_cbfn = cbfn;
3912 			sfp->state_query_cbarg = cbarg;
3913 			bfa_sfp_state_query(sfp);
3914 			return BFA_STATUS_SFP_NOT_READY;
3915 		}
3916 	}
3917 
3918 	if (sfp->state == BFA_SFP_STATE_REMOVED ||
3919 	    sfp->state == BFA_SFP_STATE_FAILED) {
3920 		bfa_trc(sfp, sfp->state);
3921 		return BFA_STATUS_NO_SFP_DEV;
3922 	}
3923 
3924 	if (sfp->state == BFA_SFP_STATE_INSERTED) {
3925 		bfa_trc(sfp, sfp->state);
3926 		return BFA_STATUS_DEVBUSY;  /* sfp is reading data */
3927 	}
3928 
3929 	/* For eloopback, all speed is allowed */
3930 	if (sfp->is_elb)
3931 		return BFA_STATUS_OK;
3932 
3933 	return bfa_sfp_speed_valid(sfp, portspeed);
3934 }
3935 
3936 /*
3937  *	Flash module specific
3938  */
3939 
3940 /*
3941  * FLASH DMA buffer should be big enough to hold both MFG block and
3942  * asic block(64k) at the same time and also should be 2k aligned to
3943  * avoid write segement to cross sector boundary.
3944  */
3945 #define BFA_FLASH_SEG_SZ	2048
3946 #define BFA_FLASH_DMA_BUF_SZ	\
3947 	BFA_ROUNDUP(0x010000 + sizeof(struct bfa_mfg_block_s), BFA_FLASH_SEG_SZ)
3948 
3949 static void
3950 bfa_flash_aen_audit_post(struct bfa_ioc_s *ioc, enum bfa_audit_aen_event event,
3951 			int inst, int type)
3952 {
3953 	struct bfad_s *bfad = (struct bfad_s *)ioc->bfa->bfad;
3954 	struct bfa_aen_entry_s  *aen_entry;
3955 
3956 	bfad_get_aen_entry(bfad, aen_entry);
3957 	if (!aen_entry)
3958 		return;
3959 
3960 	aen_entry->aen_data.audit.pwwn = ioc->attr->pwwn;
3961 	aen_entry->aen_data.audit.partition_inst = inst;
3962 	aen_entry->aen_data.audit.partition_type = type;
3963 
3964 	/* Send the AEN notification */
3965 	bfad_im_post_vendor_event(aen_entry, bfad, ++ioc->ioc_aen_seq,
3966 				  BFA_AEN_CAT_AUDIT, event);
3967 }
3968 
3969 static void
3970 bfa_flash_cb(struct bfa_flash_s *flash)
3971 {
3972 	flash->op_busy = 0;
3973 	if (flash->cbfn)
3974 		flash->cbfn(flash->cbarg, flash->status);
3975 }
3976 
3977 static void
3978 bfa_flash_notify(void *cbarg, enum bfa_ioc_event_e event)
3979 {
3980 	struct bfa_flash_s	*flash = cbarg;
3981 
3982 	bfa_trc(flash, event);
3983 	switch (event) {
3984 	case BFA_IOC_E_DISABLED:
3985 	case BFA_IOC_E_FAILED:
3986 		if (flash->op_busy) {
3987 			flash->status = BFA_STATUS_IOC_FAILURE;
3988 			flash->cbfn(flash->cbarg, flash->status);
3989 			flash->op_busy = 0;
3990 		}
3991 		break;
3992 
3993 	default:
3994 		break;
3995 	}
3996 }
3997 
3998 /*
3999  * Send flash attribute query request.
4000  *
4001  * @param[in] cbarg - callback argument
4002  */
4003 static void
4004 bfa_flash_query_send(void *cbarg)
4005 {
4006 	struct bfa_flash_s *flash = cbarg;
4007 	struct bfi_flash_query_req_s *msg =
4008 			(struct bfi_flash_query_req_s *) flash->mb.msg;
4009 
4010 	bfi_h2i_set(msg->mh, BFI_MC_FLASH, BFI_FLASH_H2I_QUERY_REQ,
4011 		bfa_ioc_portid(flash->ioc));
4012 	bfa_alen_set(&msg->alen, sizeof(struct bfa_flash_attr_s),
4013 		flash->dbuf_pa);
4014 	bfa_ioc_mbox_queue(flash->ioc, &flash->mb);
4015 }
4016 
4017 /*
4018  * Send flash write request.
4019  *
4020  * @param[in] cbarg - callback argument
4021  */
4022 static void
4023 bfa_flash_write_send(struct bfa_flash_s *flash)
4024 {
4025 	struct bfi_flash_write_req_s *msg =
4026 			(struct bfi_flash_write_req_s *) flash->mb.msg;
4027 	u32	len;
4028 
4029 	msg->type = be32_to_cpu(flash->type);
4030 	msg->instance = flash->instance;
4031 	msg->offset = be32_to_cpu(flash->addr_off + flash->offset);
4032 	len = (flash->residue < BFA_FLASH_DMA_BUF_SZ) ?
4033 		flash->residue : BFA_FLASH_DMA_BUF_SZ;
4034 	msg->length = be32_to_cpu(len);
4035 
4036 	/* indicate if it's the last msg of the whole write operation */
4037 	msg->last = (len == flash->residue) ? 1 : 0;
4038 
4039 	bfi_h2i_set(msg->mh, BFI_MC_FLASH, BFI_FLASH_H2I_WRITE_REQ,
4040 			bfa_ioc_portid(flash->ioc));
4041 	bfa_alen_set(&msg->alen, len, flash->dbuf_pa);
4042 	memcpy(flash->dbuf_kva, flash->ubuf + flash->offset, len);
4043 	bfa_ioc_mbox_queue(flash->ioc, &flash->mb);
4044 
4045 	flash->residue -= len;
4046 	flash->offset += len;
4047 }
4048 
4049 /*
4050  * Send flash read request.
4051  *
4052  * @param[in] cbarg - callback argument
4053  */
4054 static void
4055 bfa_flash_read_send(void *cbarg)
4056 {
4057 	struct bfa_flash_s *flash = cbarg;
4058 	struct bfi_flash_read_req_s *msg =
4059 			(struct bfi_flash_read_req_s *) flash->mb.msg;
4060 	u32	len;
4061 
4062 	msg->type = be32_to_cpu(flash->type);
4063 	msg->instance = flash->instance;
4064 	msg->offset = be32_to_cpu(flash->addr_off + flash->offset);
4065 	len = (flash->residue < BFA_FLASH_DMA_BUF_SZ) ?
4066 			flash->residue : BFA_FLASH_DMA_BUF_SZ;
4067 	msg->length = be32_to_cpu(len);
4068 	bfi_h2i_set(msg->mh, BFI_MC_FLASH, BFI_FLASH_H2I_READ_REQ,
4069 		bfa_ioc_portid(flash->ioc));
4070 	bfa_alen_set(&msg->alen, len, flash->dbuf_pa);
4071 	bfa_ioc_mbox_queue(flash->ioc, &flash->mb);
4072 }
4073 
4074 /*
4075  * Send flash erase request.
4076  *
4077  * @param[in] cbarg - callback argument
4078  */
4079 static void
4080 bfa_flash_erase_send(void *cbarg)
4081 {
4082 	struct bfa_flash_s *flash = cbarg;
4083 	struct bfi_flash_erase_req_s *msg =
4084 			(struct bfi_flash_erase_req_s *) flash->mb.msg;
4085 
4086 	msg->type = be32_to_cpu(flash->type);
4087 	msg->instance = flash->instance;
4088 	bfi_h2i_set(msg->mh, BFI_MC_FLASH, BFI_FLASH_H2I_ERASE_REQ,
4089 			bfa_ioc_portid(flash->ioc));
4090 	bfa_ioc_mbox_queue(flash->ioc, &flash->mb);
4091 }
4092 
4093 /*
4094  * Process flash response messages upon receiving interrupts.
4095  *
4096  * @param[in] flasharg - flash structure
4097  * @param[in] msg - message structure
4098  */
4099 static void
4100 bfa_flash_intr(void *flasharg, struct bfi_mbmsg_s *msg)
4101 {
4102 	struct bfa_flash_s *flash = flasharg;
4103 	u32	status;
4104 
4105 	union {
4106 		struct bfi_flash_query_rsp_s *query;
4107 		struct bfi_flash_erase_rsp_s *erase;
4108 		struct bfi_flash_write_rsp_s *write;
4109 		struct bfi_flash_read_rsp_s *read;
4110 		struct bfi_flash_event_s *event;
4111 		struct bfi_mbmsg_s   *msg;
4112 	} m;
4113 
4114 	m.msg = msg;
4115 	bfa_trc(flash, msg->mh.msg_id);
4116 
4117 	if (!flash->op_busy && msg->mh.msg_id != BFI_FLASH_I2H_EVENT) {
4118 		/* receiving response after ioc failure */
4119 		bfa_trc(flash, 0x9999);
4120 		return;
4121 	}
4122 
4123 	switch (msg->mh.msg_id) {
4124 	case BFI_FLASH_I2H_QUERY_RSP:
4125 		status = be32_to_cpu(m.query->status);
4126 		bfa_trc(flash, status);
4127 		if (status == BFA_STATUS_OK) {
4128 			u32	i;
4129 			struct bfa_flash_attr_s *attr, *f;
4130 
4131 			attr = (struct bfa_flash_attr_s *) flash->ubuf;
4132 			f = (struct bfa_flash_attr_s *) flash->dbuf_kva;
4133 			attr->status = be32_to_cpu(f->status);
4134 			attr->npart = be32_to_cpu(f->npart);
4135 			bfa_trc(flash, attr->status);
4136 			bfa_trc(flash, attr->npart);
4137 			for (i = 0; i < attr->npart; i++) {
4138 				attr->part[i].part_type =
4139 					be32_to_cpu(f->part[i].part_type);
4140 				attr->part[i].part_instance =
4141 					be32_to_cpu(f->part[i].part_instance);
4142 				attr->part[i].part_off =
4143 					be32_to_cpu(f->part[i].part_off);
4144 				attr->part[i].part_size =
4145 					be32_to_cpu(f->part[i].part_size);
4146 				attr->part[i].part_len =
4147 					be32_to_cpu(f->part[i].part_len);
4148 				attr->part[i].part_status =
4149 					be32_to_cpu(f->part[i].part_status);
4150 			}
4151 		}
4152 		flash->status = status;
4153 		bfa_flash_cb(flash);
4154 		break;
4155 	case BFI_FLASH_I2H_ERASE_RSP:
4156 		status = be32_to_cpu(m.erase->status);
4157 		bfa_trc(flash, status);
4158 		flash->status = status;
4159 		bfa_flash_cb(flash);
4160 		break;
4161 	case BFI_FLASH_I2H_WRITE_RSP:
4162 		status = be32_to_cpu(m.write->status);
4163 		bfa_trc(flash, status);
4164 		if (status != BFA_STATUS_OK || flash->residue == 0) {
4165 			flash->status = status;
4166 			bfa_flash_cb(flash);
4167 		} else {
4168 			bfa_trc(flash, flash->offset);
4169 			bfa_flash_write_send(flash);
4170 		}
4171 		break;
4172 	case BFI_FLASH_I2H_READ_RSP:
4173 		status = be32_to_cpu(m.read->status);
4174 		bfa_trc(flash, status);
4175 		if (status != BFA_STATUS_OK) {
4176 			flash->status = status;
4177 			bfa_flash_cb(flash);
4178 		} else {
4179 			u32 len = be32_to_cpu(m.read->length);
4180 			bfa_trc(flash, flash->offset);
4181 			bfa_trc(flash, len);
4182 			memcpy(flash->ubuf + flash->offset,
4183 				flash->dbuf_kva, len);
4184 			flash->residue -= len;
4185 			flash->offset += len;
4186 			if (flash->residue == 0) {
4187 				flash->status = status;
4188 				bfa_flash_cb(flash);
4189 			} else
4190 				bfa_flash_read_send(flash);
4191 		}
4192 		break;
4193 	case BFI_FLASH_I2H_BOOT_VER_RSP:
4194 		break;
4195 	case BFI_FLASH_I2H_EVENT:
4196 		status = be32_to_cpu(m.event->status);
4197 		bfa_trc(flash, status);
4198 		if (status == BFA_STATUS_BAD_FWCFG)
4199 			bfa_ioc_aen_post(flash->ioc, BFA_IOC_AEN_FWCFG_ERROR);
4200 		else if (status == BFA_STATUS_INVALID_VENDOR) {
4201 			u32 param;
4202 			param = be32_to_cpu(m.event->param);
4203 			bfa_trc(flash, param);
4204 			bfa_ioc_aen_post(flash->ioc,
4205 				BFA_IOC_AEN_INVALID_VENDOR);
4206 		}
4207 		break;
4208 
4209 	default:
4210 		WARN_ON(1);
4211 	}
4212 }
4213 
4214 /*
4215  * Flash memory info API.
4216  *
4217  * @param[in] mincfg - minimal cfg variable
4218  */
4219 u32
4220 bfa_flash_meminfo(bfa_boolean_t mincfg)
4221 {
4222 	/* min driver doesn't need flash */
4223 	if (mincfg)
4224 		return 0;
4225 	return BFA_ROUNDUP(BFA_FLASH_DMA_BUF_SZ, BFA_DMA_ALIGN_SZ);
4226 }
4227 
4228 /*
4229  * Flash attach API.
4230  *
4231  * @param[in] flash - flash structure
4232  * @param[in] ioc  - ioc structure
4233  * @param[in] dev  - device structure
4234  * @param[in] trcmod - trace module
4235  * @param[in] logmod - log module
4236  */
4237 void
4238 bfa_flash_attach(struct bfa_flash_s *flash, struct bfa_ioc_s *ioc, void *dev,
4239 		struct bfa_trc_mod_s *trcmod, bfa_boolean_t mincfg)
4240 {
4241 	flash->ioc = ioc;
4242 	flash->trcmod = trcmod;
4243 	flash->cbfn = NULL;
4244 	flash->cbarg = NULL;
4245 	flash->op_busy = 0;
4246 
4247 	bfa_ioc_mbox_regisr(flash->ioc, BFI_MC_FLASH, bfa_flash_intr, flash);
4248 	bfa_q_qe_init(&flash->ioc_notify);
4249 	bfa_ioc_notify_init(&flash->ioc_notify, bfa_flash_notify, flash);
4250 	list_add_tail(&flash->ioc_notify.qe, &flash->ioc->notify_q);
4251 
4252 	/* min driver doesn't need flash */
4253 	if (mincfg) {
4254 		flash->dbuf_kva = NULL;
4255 		flash->dbuf_pa = 0;
4256 	}
4257 }
4258 
4259 /*
4260  * Claim memory for flash
4261  *
4262  * @param[in] flash - flash structure
4263  * @param[in] dm_kva - pointer to virtual memory address
4264  * @param[in] dm_pa - physical memory address
4265  * @param[in] mincfg - minimal cfg variable
4266  */
4267 void
4268 bfa_flash_memclaim(struct bfa_flash_s *flash, u8 *dm_kva, u64 dm_pa,
4269 		bfa_boolean_t mincfg)
4270 {
4271 	if (mincfg)
4272 		return;
4273 
4274 	flash->dbuf_kva = dm_kva;
4275 	flash->dbuf_pa = dm_pa;
4276 	memset(flash->dbuf_kva, 0, BFA_FLASH_DMA_BUF_SZ);
4277 	dm_kva += BFA_ROUNDUP(BFA_FLASH_DMA_BUF_SZ, BFA_DMA_ALIGN_SZ);
4278 	dm_pa += BFA_ROUNDUP(BFA_FLASH_DMA_BUF_SZ, BFA_DMA_ALIGN_SZ);
4279 }
4280 
4281 /*
4282  * Get flash attribute.
4283  *
4284  * @param[in] flash - flash structure
4285  * @param[in] attr - flash attribute structure
4286  * @param[in] cbfn - callback function
4287  * @param[in] cbarg - callback argument
4288  *
4289  * Return status.
4290  */
4291 bfa_status_t
4292 bfa_flash_get_attr(struct bfa_flash_s *flash, struct bfa_flash_attr_s *attr,
4293 		bfa_cb_flash_t cbfn, void *cbarg)
4294 {
4295 	bfa_trc(flash, BFI_FLASH_H2I_QUERY_REQ);
4296 
4297 	if (!bfa_ioc_is_operational(flash->ioc))
4298 		return BFA_STATUS_IOC_NON_OP;
4299 
4300 	if (flash->op_busy) {
4301 		bfa_trc(flash, flash->op_busy);
4302 		return BFA_STATUS_DEVBUSY;
4303 	}
4304 
4305 	flash->op_busy = 1;
4306 	flash->cbfn = cbfn;
4307 	flash->cbarg = cbarg;
4308 	flash->ubuf = (u8 *) attr;
4309 	bfa_flash_query_send(flash);
4310 
4311 	return BFA_STATUS_OK;
4312 }
4313 
4314 /*
4315  * Erase flash partition.
4316  *
4317  * @param[in] flash - flash structure
4318  * @param[in] type - flash partition type
4319  * @param[in] instance - flash partition instance
4320  * @param[in] cbfn - callback function
4321  * @param[in] cbarg - callback argument
4322  *
4323  * Return status.
4324  */
4325 bfa_status_t
4326 bfa_flash_erase_part(struct bfa_flash_s *flash, enum bfa_flash_part_type type,
4327 		u8 instance, bfa_cb_flash_t cbfn, void *cbarg)
4328 {
4329 	bfa_trc(flash, BFI_FLASH_H2I_ERASE_REQ);
4330 	bfa_trc(flash, type);
4331 	bfa_trc(flash, instance);
4332 
4333 	if (!bfa_ioc_is_operational(flash->ioc))
4334 		return BFA_STATUS_IOC_NON_OP;
4335 
4336 	if (flash->op_busy) {
4337 		bfa_trc(flash, flash->op_busy);
4338 		return BFA_STATUS_DEVBUSY;
4339 	}
4340 
4341 	flash->op_busy = 1;
4342 	flash->cbfn = cbfn;
4343 	flash->cbarg = cbarg;
4344 	flash->type = type;
4345 	flash->instance = instance;
4346 
4347 	bfa_flash_erase_send(flash);
4348 	bfa_flash_aen_audit_post(flash->ioc, BFA_AUDIT_AEN_FLASH_ERASE,
4349 				instance, type);
4350 	return BFA_STATUS_OK;
4351 }
4352 
4353 /*
4354  * Update flash partition.
4355  *
4356  * @param[in] flash - flash structure
4357  * @param[in] type - flash partition type
4358  * @param[in] instance - flash partition instance
4359  * @param[in] buf - update data buffer
4360  * @param[in] len - data buffer length
4361  * @param[in] offset - offset relative to the partition starting address
4362  * @param[in] cbfn - callback function
4363  * @param[in] cbarg - callback argument
4364  *
4365  * Return status.
4366  */
4367 bfa_status_t
4368 bfa_flash_update_part(struct bfa_flash_s *flash, enum bfa_flash_part_type type,
4369 		u8 instance, void *buf, u32 len, u32 offset,
4370 		bfa_cb_flash_t cbfn, void *cbarg)
4371 {
4372 	bfa_trc(flash, BFI_FLASH_H2I_WRITE_REQ);
4373 	bfa_trc(flash, type);
4374 	bfa_trc(flash, instance);
4375 	bfa_trc(flash, len);
4376 	bfa_trc(flash, offset);
4377 
4378 	if (!bfa_ioc_is_operational(flash->ioc))
4379 		return BFA_STATUS_IOC_NON_OP;
4380 
4381 	/*
4382 	 * 'len' must be in word (4-byte) boundary
4383 	 * 'offset' must be in sector (16kb) boundary
4384 	 */
4385 	if (!len || (len & 0x03) || (offset & 0x00003FFF))
4386 		return BFA_STATUS_FLASH_BAD_LEN;
4387 
4388 	if (type == BFA_FLASH_PART_MFG)
4389 		return BFA_STATUS_EINVAL;
4390 
4391 	if (flash->op_busy) {
4392 		bfa_trc(flash, flash->op_busy);
4393 		return BFA_STATUS_DEVBUSY;
4394 	}
4395 
4396 	flash->op_busy = 1;
4397 	flash->cbfn = cbfn;
4398 	flash->cbarg = cbarg;
4399 	flash->type = type;
4400 	flash->instance = instance;
4401 	flash->residue = len;
4402 	flash->offset = 0;
4403 	flash->addr_off = offset;
4404 	flash->ubuf = buf;
4405 
4406 	bfa_flash_write_send(flash);
4407 	return BFA_STATUS_OK;
4408 }
4409 
4410 /*
4411  * Read flash partition.
4412  *
4413  * @param[in] flash - flash structure
4414  * @param[in] type - flash partition type
4415  * @param[in] instance - flash partition instance
4416  * @param[in] buf - read data buffer
4417  * @param[in] len - data buffer length
4418  * @param[in] offset - offset relative to the partition starting address
4419  * @param[in] cbfn - callback function
4420  * @param[in] cbarg - callback argument
4421  *
4422  * Return status.
4423  */
4424 bfa_status_t
4425 bfa_flash_read_part(struct bfa_flash_s *flash, enum bfa_flash_part_type type,
4426 		u8 instance, void *buf, u32 len, u32 offset,
4427 		bfa_cb_flash_t cbfn, void *cbarg)
4428 {
4429 	bfa_trc(flash, BFI_FLASH_H2I_READ_REQ);
4430 	bfa_trc(flash, type);
4431 	bfa_trc(flash, instance);
4432 	bfa_trc(flash, len);
4433 	bfa_trc(flash, offset);
4434 
4435 	if (!bfa_ioc_is_operational(flash->ioc))
4436 		return BFA_STATUS_IOC_NON_OP;
4437 
4438 	/*
4439 	 * 'len' must be in word (4-byte) boundary
4440 	 * 'offset' must be in sector (16kb) boundary
4441 	 */
4442 	if (!len || (len & 0x03) || (offset & 0x00003FFF))
4443 		return BFA_STATUS_FLASH_BAD_LEN;
4444 
4445 	if (flash->op_busy) {
4446 		bfa_trc(flash, flash->op_busy);
4447 		return BFA_STATUS_DEVBUSY;
4448 	}
4449 
4450 	flash->op_busy = 1;
4451 	flash->cbfn = cbfn;
4452 	flash->cbarg = cbarg;
4453 	flash->type = type;
4454 	flash->instance = instance;
4455 	flash->residue = len;
4456 	flash->offset = 0;
4457 	flash->addr_off = offset;
4458 	flash->ubuf = buf;
4459 	bfa_flash_read_send(flash);
4460 
4461 	return BFA_STATUS_OK;
4462 }
4463 
4464 /*
4465  *	DIAG module specific
4466  */
4467 
4468 #define BFA_DIAG_MEMTEST_TOV	50000	/* memtest timeout in msec */
4469 #define CT2_BFA_DIAG_MEMTEST_TOV	(9*30*1000)  /* 4.5 min */
4470 
4471 /* IOC event handler */
4472 static void
4473 bfa_diag_notify(void *diag_arg, enum bfa_ioc_event_e event)
4474 {
4475 	struct bfa_diag_s *diag = diag_arg;
4476 
4477 	bfa_trc(diag, event);
4478 	bfa_trc(diag, diag->block);
4479 	bfa_trc(diag, diag->fwping.lock);
4480 	bfa_trc(diag, diag->tsensor.lock);
4481 
4482 	switch (event) {
4483 	case BFA_IOC_E_DISABLED:
4484 	case BFA_IOC_E_FAILED:
4485 		if (diag->fwping.lock) {
4486 			diag->fwping.status = BFA_STATUS_IOC_FAILURE;
4487 			diag->fwping.cbfn(diag->fwping.cbarg,
4488 					diag->fwping.status);
4489 			diag->fwping.lock = 0;
4490 		}
4491 
4492 		if (diag->tsensor.lock) {
4493 			diag->tsensor.status = BFA_STATUS_IOC_FAILURE;
4494 			diag->tsensor.cbfn(diag->tsensor.cbarg,
4495 					   diag->tsensor.status);
4496 			diag->tsensor.lock = 0;
4497 		}
4498 
4499 		if (diag->block) {
4500 			if (diag->timer_active) {
4501 				bfa_timer_stop(&diag->timer);
4502 				diag->timer_active = 0;
4503 			}
4504 
4505 			diag->status = BFA_STATUS_IOC_FAILURE;
4506 			diag->cbfn(diag->cbarg, diag->status);
4507 			diag->block = 0;
4508 		}
4509 		break;
4510 
4511 	default:
4512 		break;
4513 	}
4514 }
4515 
4516 static void
4517 bfa_diag_memtest_done(void *cbarg)
4518 {
4519 	struct bfa_diag_s *diag = cbarg;
4520 	struct bfa_ioc_s  *ioc = diag->ioc;
4521 	struct bfa_diag_memtest_result *res = diag->result;
4522 	u32	loff = BFI_BOOT_MEMTEST_RES_ADDR;
4523 	u32	pgnum, pgoff, i;
4524 
4525 	pgnum = PSS_SMEM_PGNUM(ioc->ioc_regs.smem_pg0, loff);
4526 	pgoff = PSS_SMEM_PGOFF(loff);
4527 
4528 	writel(pgnum, ioc->ioc_regs.host_page_num_fn);
4529 
4530 	for (i = 0; i < (sizeof(struct bfa_diag_memtest_result) /
4531 			 sizeof(u32)); i++) {
4532 		/* read test result from smem */
4533 		*((u32 *) res + i) =
4534 			bfa_mem_read(ioc->ioc_regs.smem_page_start, loff);
4535 		loff += sizeof(u32);
4536 	}
4537 
4538 	/* Reset IOC fwstates to BFI_IOC_UNINIT */
4539 	bfa_ioc_reset_fwstate(ioc);
4540 
4541 	res->status = swab32(res->status);
4542 	bfa_trc(diag, res->status);
4543 
4544 	if (res->status == BFI_BOOT_MEMTEST_RES_SIG)
4545 		diag->status = BFA_STATUS_OK;
4546 	else {
4547 		diag->status = BFA_STATUS_MEMTEST_FAILED;
4548 		res->addr = swab32(res->addr);
4549 		res->exp = swab32(res->exp);
4550 		res->act = swab32(res->act);
4551 		res->err_status = swab32(res->err_status);
4552 		res->err_status1 = swab32(res->err_status1);
4553 		res->err_addr = swab32(res->err_addr);
4554 		bfa_trc(diag, res->addr);
4555 		bfa_trc(diag, res->exp);
4556 		bfa_trc(diag, res->act);
4557 		bfa_trc(diag, res->err_status);
4558 		bfa_trc(diag, res->err_status1);
4559 		bfa_trc(diag, res->err_addr);
4560 	}
4561 	diag->timer_active = 0;
4562 	diag->cbfn(diag->cbarg, diag->status);
4563 	diag->block = 0;
4564 }
4565 
4566 /*
4567  * Firmware ping
4568  */
4569 
4570 /*
4571  * Perform DMA test directly
4572  */
4573 static void
4574 diag_fwping_send(struct bfa_diag_s *diag)
4575 {
4576 	struct bfi_diag_fwping_req_s *fwping_req;
4577 	u32	i;
4578 
4579 	bfa_trc(diag, diag->fwping.dbuf_pa);
4580 
4581 	/* fill DMA area with pattern */
4582 	for (i = 0; i < (BFI_DIAG_DMA_BUF_SZ >> 2); i++)
4583 		*((u32 *)diag->fwping.dbuf_kva + i) = diag->fwping.data;
4584 
4585 	/* Fill mbox msg */
4586 	fwping_req = (struct bfi_diag_fwping_req_s *)diag->fwping.mbcmd.msg;
4587 
4588 	/* Setup SG list */
4589 	bfa_alen_set(&fwping_req->alen, BFI_DIAG_DMA_BUF_SZ,
4590 			diag->fwping.dbuf_pa);
4591 	/* Set up dma count */
4592 	fwping_req->count = cpu_to_be32(diag->fwping.count);
4593 	/* Set up data pattern */
4594 	fwping_req->data = diag->fwping.data;
4595 
4596 	/* build host command */
4597 	bfi_h2i_set(fwping_req->mh, BFI_MC_DIAG, BFI_DIAG_H2I_FWPING,
4598 		bfa_ioc_portid(diag->ioc));
4599 
4600 	/* send mbox cmd */
4601 	bfa_ioc_mbox_queue(diag->ioc, &diag->fwping.mbcmd);
4602 }
4603 
4604 static void
4605 diag_fwping_comp(struct bfa_diag_s *diag,
4606 		 struct bfi_diag_fwping_rsp_s *diag_rsp)
4607 {
4608 	u32	rsp_data = diag_rsp->data;
4609 	u8	rsp_dma_status = diag_rsp->dma_status;
4610 
4611 	bfa_trc(diag, rsp_data);
4612 	bfa_trc(diag, rsp_dma_status);
4613 
4614 	if (rsp_dma_status == BFA_STATUS_OK) {
4615 		u32	i, pat;
4616 		pat = (diag->fwping.count & 0x1) ? ~(diag->fwping.data) :
4617 			diag->fwping.data;
4618 		/* Check mbox data */
4619 		if (diag->fwping.data != rsp_data) {
4620 			bfa_trc(diag, rsp_data);
4621 			diag->fwping.result->dmastatus =
4622 					BFA_STATUS_DATACORRUPTED;
4623 			diag->fwping.status = BFA_STATUS_DATACORRUPTED;
4624 			diag->fwping.cbfn(diag->fwping.cbarg,
4625 					diag->fwping.status);
4626 			diag->fwping.lock = 0;
4627 			return;
4628 		}
4629 		/* Check dma pattern */
4630 		for (i = 0; i < (BFI_DIAG_DMA_BUF_SZ >> 2); i++) {
4631 			if (*((u32 *)diag->fwping.dbuf_kva + i) != pat) {
4632 				bfa_trc(diag, i);
4633 				bfa_trc(diag, pat);
4634 				bfa_trc(diag,
4635 					*((u32 *)diag->fwping.dbuf_kva + i));
4636 				diag->fwping.result->dmastatus =
4637 						BFA_STATUS_DATACORRUPTED;
4638 				diag->fwping.status = BFA_STATUS_DATACORRUPTED;
4639 				diag->fwping.cbfn(diag->fwping.cbarg,
4640 						diag->fwping.status);
4641 				diag->fwping.lock = 0;
4642 				return;
4643 			}
4644 		}
4645 		diag->fwping.result->dmastatus = BFA_STATUS_OK;
4646 		diag->fwping.status = BFA_STATUS_OK;
4647 		diag->fwping.cbfn(diag->fwping.cbarg, diag->fwping.status);
4648 		diag->fwping.lock = 0;
4649 	} else {
4650 		diag->fwping.status = BFA_STATUS_HDMA_FAILED;
4651 		diag->fwping.cbfn(diag->fwping.cbarg, diag->fwping.status);
4652 		diag->fwping.lock = 0;
4653 	}
4654 }
4655 
4656 /*
4657  * Temperature Sensor
4658  */
4659 
4660 static void
4661 diag_tempsensor_send(struct bfa_diag_s *diag)
4662 {
4663 	struct bfi_diag_ts_req_s *msg;
4664 
4665 	msg = (struct bfi_diag_ts_req_s *)diag->tsensor.mbcmd.msg;
4666 	bfa_trc(diag, msg->temp);
4667 	/* build host command */
4668 	bfi_h2i_set(msg->mh, BFI_MC_DIAG, BFI_DIAG_H2I_TEMPSENSOR,
4669 		bfa_ioc_portid(diag->ioc));
4670 	/* send mbox cmd */
4671 	bfa_ioc_mbox_queue(diag->ioc, &diag->tsensor.mbcmd);
4672 }
4673 
4674 static void
4675 diag_tempsensor_comp(struct bfa_diag_s *diag, bfi_diag_ts_rsp_t *rsp)
4676 {
4677 	if (!diag->tsensor.lock) {
4678 		/* receiving response after ioc failure */
4679 		bfa_trc(diag, diag->tsensor.lock);
4680 		return;
4681 	}
4682 
4683 	/*
4684 	 * ASIC junction tempsensor is a reg read operation
4685 	 * it will always return OK
4686 	 */
4687 	diag->tsensor.temp->temp = be16_to_cpu(rsp->temp);
4688 	diag->tsensor.temp->ts_junc = rsp->ts_junc;
4689 	diag->tsensor.temp->ts_brd = rsp->ts_brd;
4690 
4691 	if (rsp->ts_brd) {
4692 		/* tsensor.temp->status is brd_temp status */
4693 		diag->tsensor.temp->status = rsp->status;
4694 		if (rsp->status == BFA_STATUS_OK) {
4695 			diag->tsensor.temp->brd_temp =
4696 				be16_to_cpu(rsp->brd_temp);
4697 		} else
4698 			diag->tsensor.temp->brd_temp = 0;
4699 	}
4700 
4701 	bfa_trc(diag, rsp->status);
4702 	bfa_trc(diag, rsp->ts_junc);
4703 	bfa_trc(diag, rsp->temp);
4704 	bfa_trc(diag, rsp->ts_brd);
4705 	bfa_trc(diag, rsp->brd_temp);
4706 
4707 	/* tsensor status is always good bcos we always have junction temp */
4708 	diag->tsensor.status = BFA_STATUS_OK;
4709 	diag->tsensor.cbfn(diag->tsensor.cbarg, diag->tsensor.status);
4710 	diag->tsensor.lock = 0;
4711 }
4712 
4713 /*
4714  *	LED Test command
4715  */
4716 static void
4717 diag_ledtest_send(struct bfa_diag_s *diag, struct bfa_diag_ledtest_s *ledtest)
4718 {
4719 	struct bfi_diag_ledtest_req_s  *msg;
4720 
4721 	msg = (struct bfi_diag_ledtest_req_s *)diag->ledtest.mbcmd.msg;
4722 	/* build host command */
4723 	bfi_h2i_set(msg->mh, BFI_MC_DIAG, BFI_DIAG_H2I_LEDTEST,
4724 			bfa_ioc_portid(diag->ioc));
4725 
4726 	/*
4727 	 * convert the freq from N blinks per 10 sec to
4728 	 * crossbow ontime value. We do it here because division is need
4729 	 */
4730 	if (ledtest->freq)
4731 		ledtest->freq = 500 / ledtest->freq;
4732 
4733 	if (ledtest->freq == 0)
4734 		ledtest->freq = 1;
4735 
4736 	bfa_trc(diag, ledtest->freq);
4737 	/* mcpy(&ledtest_req->req, ledtest, sizeof(bfa_diag_ledtest_t)); */
4738 	msg->cmd = (u8) ledtest->cmd;
4739 	msg->color = (u8) ledtest->color;
4740 	msg->portid = bfa_ioc_portid(diag->ioc);
4741 	msg->led = ledtest->led;
4742 	msg->freq = cpu_to_be16(ledtest->freq);
4743 
4744 	/* send mbox cmd */
4745 	bfa_ioc_mbox_queue(diag->ioc, &diag->ledtest.mbcmd);
4746 }
4747 
4748 static void
4749 diag_ledtest_comp(struct bfa_diag_s *diag, struct bfi_diag_ledtest_rsp_s *msg)
4750 {
4751 	bfa_trc(diag, diag->ledtest.lock);
4752 	diag->ledtest.lock = BFA_FALSE;
4753 	/* no bfa_cb_queue is needed because driver is not waiting */
4754 }
4755 
4756 /*
4757  * Port beaconing
4758  */
4759 static void
4760 diag_portbeacon_send(struct bfa_diag_s *diag, bfa_boolean_t beacon, u32 sec)
4761 {
4762 	struct bfi_diag_portbeacon_req_s *msg;
4763 
4764 	msg = (struct bfi_diag_portbeacon_req_s *)diag->beacon.mbcmd.msg;
4765 	/* build host command */
4766 	bfi_h2i_set(msg->mh, BFI_MC_DIAG, BFI_DIAG_H2I_PORTBEACON,
4767 		bfa_ioc_portid(diag->ioc));
4768 	msg->beacon = beacon;
4769 	msg->period = cpu_to_be32(sec);
4770 	/* send mbox cmd */
4771 	bfa_ioc_mbox_queue(diag->ioc, &diag->beacon.mbcmd);
4772 }
4773 
4774 static void
4775 diag_portbeacon_comp(struct bfa_diag_s *diag)
4776 {
4777 	bfa_trc(diag, diag->beacon.state);
4778 	diag->beacon.state = BFA_FALSE;
4779 	if (diag->cbfn_beacon)
4780 		diag->cbfn_beacon(diag->dev, BFA_FALSE, diag->beacon.link_e2e);
4781 }
4782 
4783 /*
4784  *	Diag hmbox handler
4785  */
4786 void
4787 bfa_diag_intr(void *diagarg, struct bfi_mbmsg_s *msg)
4788 {
4789 	struct bfa_diag_s *diag = diagarg;
4790 
4791 	switch (msg->mh.msg_id) {
4792 	case BFI_DIAG_I2H_PORTBEACON:
4793 		diag_portbeacon_comp(diag);
4794 		break;
4795 	case BFI_DIAG_I2H_FWPING:
4796 		diag_fwping_comp(diag, (struct bfi_diag_fwping_rsp_s *) msg);
4797 		break;
4798 	case BFI_DIAG_I2H_TEMPSENSOR:
4799 		diag_tempsensor_comp(diag, (bfi_diag_ts_rsp_t *) msg);
4800 		break;
4801 	case BFI_DIAG_I2H_LEDTEST:
4802 		diag_ledtest_comp(diag, (struct bfi_diag_ledtest_rsp_s *) msg);
4803 		break;
4804 	default:
4805 		bfa_trc(diag, msg->mh.msg_id);
4806 		WARN_ON(1);
4807 	}
4808 }
4809 
4810 /*
4811  * Gen RAM Test
4812  *
4813  *   @param[in] *diag           - diag data struct
4814  *   @param[in] *memtest        - mem test params input from upper layer,
4815  *   @param[in] pattern         - mem test pattern
4816  *   @param[in] *result         - mem test result
4817  *   @param[in] cbfn            - mem test callback functioin
4818  *   @param[in] cbarg           - callback functioin arg
4819  *
4820  *   @param[out]
4821  */
4822 bfa_status_t
4823 bfa_diag_memtest(struct bfa_diag_s *diag, struct bfa_diag_memtest_s *memtest,
4824 		u32 pattern, struct bfa_diag_memtest_result *result,
4825 		bfa_cb_diag_t cbfn, void *cbarg)
4826 {
4827 	u32	memtest_tov;
4828 
4829 	bfa_trc(diag, pattern);
4830 
4831 	if (!bfa_ioc_adapter_is_disabled(diag->ioc))
4832 		return BFA_STATUS_ADAPTER_ENABLED;
4833 
4834 	/* check to see if there is another destructive diag cmd running */
4835 	if (diag->block) {
4836 		bfa_trc(diag, diag->block);
4837 		return BFA_STATUS_DEVBUSY;
4838 	} else
4839 		diag->block = 1;
4840 
4841 	diag->result = result;
4842 	diag->cbfn = cbfn;
4843 	diag->cbarg = cbarg;
4844 
4845 	/* download memtest code and take LPU0 out of reset */
4846 	bfa_ioc_boot(diag->ioc, BFI_FWBOOT_TYPE_MEMTEST, BFI_FWBOOT_ENV_OS);
4847 
4848 	memtest_tov = (bfa_ioc_asic_gen(diag->ioc) == BFI_ASIC_GEN_CT2) ?
4849 		       CT2_BFA_DIAG_MEMTEST_TOV : BFA_DIAG_MEMTEST_TOV;
4850 	bfa_timer_begin(diag->ioc->timer_mod, &diag->timer,
4851 			bfa_diag_memtest_done, diag, memtest_tov);
4852 	diag->timer_active = 1;
4853 	return BFA_STATUS_OK;
4854 }
4855 
4856 /*
4857  * DIAG firmware ping command
4858  *
4859  *   @param[in] *diag           - diag data struct
4860  *   @param[in] cnt             - dma loop count for testing PCIE
4861  *   @param[in] data            - data pattern to pass in fw
4862  *   @param[in] *result         - pt to bfa_diag_fwping_result_t data struct
4863  *   @param[in] cbfn            - callback function
4864  *   @param[in] *cbarg          - callback functioin arg
4865  *
4866  *   @param[out]
4867  */
4868 bfa_status_t
4869 bfa_diag_fwping(struct bfa_diag_s *diag, u32 cnt, u32 data,
4870 		struct bfa_diag_results_fwping *result, bfa_cb_diag_t cbfn,
4871 		void *cbarg)
4872 {
4873 	bfa_trc(diag, cnt);
4874 	bfa_trc(diag, data);
4875 
4876 	if (!bfa_ioc_is_operational(diag->ioc))
4877 		return BFA_STATUS_IOC_NON_OP;
4878 
4879 	if (bfa_asic_id_ct2(bfa_ioc_devid((diag->ioc))) &&
4880 	    ((diag->ioc)->clscode == BFI_PCIFN_CLASS_ETH))
4881 		return BFA_STATUS_CMD_NOTSUPP;
4882 
4883 	/* check to see if there is another destructive diag cmd running */
4884 	if (diag->block || diag->fwping.lock) {
4885 		bfa_trc(diag, diag->block);
4886 		bfa_trc(diag, diag->fwping.lock);
4887 		return BFA_STATUS_DEVBUSY;
4888 	}
4889 
4890 	/* Initialization */
4891 	diag->fwping.lock = 1;
4892 	diag->fwping.cbfn = cbfn;
4893 	diag->fwping.cbarg = cbarg;
4894 	diag->fwping.result = result;
4895 	diag->fwping.data = data;
4896 	diag->fwping.count = cnt;
4897 
4898 	/* Init test results */
4899 	diag->fwping.result->data = 0;
4900 	diag->fwping.result->status = BFA_STATUS_OK;
4901 
4902 	/* kick off the first ping */
4903 	diag_fwping_send(diag);
4904 	return BFA_STATUS_OK;
4905 }
4906 
4907 /*
4908  * Read Temperature Sensor
4909  *
4910  *   @param[in] *diag           - diag data struct
4911  *   @param[in] *result         - pt to bfa_diag_temp_t data struct
4912  *   @param[in] cbfn            - callback function
4913  *   @param[in] *cbarg          - callback functioin arg
4914  *
4915  *   @param[out]
4916  */
4917 bfa_status_t
4918 bfa_diag_tsensor_query(struct bfa_diag_s *diag,
4919 		struct bfa_diag_results_tempsensor_s *result,
4920 		bfa_cb_diag_t cbfn, void *cbarg)
4921 {
4922 	/* check to see if there is a destructive diag cmd running */
4923 	if (diag->block || diag->tsensor.lock) {
4924 		bfa_trc(diag, diag->block);
4925 		bfa_trc(diag, diag->tsensor.lock);
4926 		return BFA_STATUS_DEVBUSY;
4927 	}
4928 
4929 	if (!bfa_ioc_is_operational(diag->ioc))
4930 		return BFA_STATUS_IOC_NON_OP;
4931 
4932 	/* Init diag mod params */
4933 	diag->tsensor.lock = 1;
4934 	diag->tsensor.temp = result;
4935 	diag->tsensor.cbfn = cbfn;
4936 	diag->tsensor.cbarg = cbarg;
4937 	diag->tsensor.status = BFA_STATUS_OK;
4938 
4939 	/* Send msg to fw */
4940 	diag_tempsensor_send(diag);
4941 
4942 	return BFA_STATUS_OK;
4943 }
4944 
4945 /*
4946  * LED Test command
4947  *
4948  *   @param[in] *diag           - diag data struct
4949  *   @param[in] *ledtest        - pt to ledtest data structure
4950  *
4951  *   @param[out]
4952  */
4953 bfa_status_t
4954 bfa_diag_ledtest(struct bfa_diag_s *diag, struct bfa_diag_ledtest_s *ledtest)
4955 {
4956 	bfa_trc(diag, ledtest->cmd);
4957 
4958 	if (!bfa_ioc_is_operational(diag->ioc))
4959 		return BFA_STATUS_IOC_NON_OP;
4960 
4961 	if (diag->beacon.state)
4962 		return BFA_STATUS_BEACON_ON;
4963 
4964 	if (diag->ledtest.lock)
4965 		return BFA_STATUS_LEDTEST_OP;
4966 
4967 	/* Send msg to fw */
4968 	diag->ledtest.lock = BFA_TRUE;
4969 	diag_ledtest_send(diag, ledtest);
4970 
4971 	return BFA_STATUS_OK;
4972 }
4973 
4974 /*
4975  * Port beaconing command
4976  *
4977  *   @param[in] *diag           - diag data struct
4978  *   @param[in] beacon          - port beaconing 1:ON   0:OFF
4979  *   @param[in] link_e2e_beacon - link beaconing 1:ON   0:OFF
4980  *   @param[in] sec             - beaconing duration in seconds
4981  *
4982  *   @param[out]
4983  */
4984 bfa_status_t
4985 bfa_diag_beacon_port(struct bfa_diag_s *diag, bfa_boolean_t beacon,
4986 		bfa_boolean_t link_e2e_beacon, uint32_t sec)
4987 {
4988 	bfa_trc(diag, beacon);
4989 	bfa_trc(diag, link_e2e_beacon);
4990 	bfa_trc(diag, sec);
4991 
4992 	if (!bfa_ioc_is_operational(diag->ioc))
4993 		return BFA_STATUS_IOC_NON_OP;
4994 
4995 	if (diag->ledtest.lock)
4996 		return BFA_STATUS_LEDTEST_OP;
4997 
4998 	if (diag->beacon.state && beacon)       /* beacon alread on */
4999 		return BFA_STATUS_BEACON_ON;
5000 
5001 	diag->beacon.state	= beacon;
5002 	diag->beacon.link_e2e	= link_e2e_beacon;
5003 	if (diag->cbfn_beacon)
5004 		diag->cbfn_beacon(diag->dev, beacon, link_e2e_beacon);
5005 
5006 	/* Send msg to fw */
5007 	diag_portbeacon_send(diag, beacon, sec);
5008 
5009 	return BFA_STATUS_OK;
5010 }
5011 
5012 /*
5013  * Return DMA memory needed by diag module.
5014  */
5015 u32
5016 bfa_diag_meminfo(void)
5017 {
5018 	return BFA_ROUNDUP(BFI_DIAG_DMA_BUF_SZ, BFA_DMA_ALIGN_SZ);
5019 }
5020 
5021 /*
5022  *	Attach virtual and physical memory for Diag.
5023  */
5024 void
5025 bfa_diag_attach(struct bfa_diag_s *diag, struct bfa_ioc_s *ioc, void *dev,
5026 	bfa_cb_diag_beacon_t cbfn_beacon, struct bfa_trc_mod_s *trcmod)
5027 {
5028 	diag->dev = dev;
5029 	diag->ioc = ioc;
5030 	diag->trcmod = trcmod;
5031 
5032 	diag->block = 0;
5033 	diag->cbfn = NULL;
5034 	diag->cbarg = NULL;
5035 	diag->result = NULL;
5036 	diag->cbfn_beacon = cbfn_beacon;
5037 
5038 	bfa_ioc_mbox_regisr(diag->ioc, BFI_MC_DIAG, bfa_diag_intr, diag);
5039 	bfa_q_qe_init(&diag->ioc_notify);
5040 	bfa_ioc_notify_init(&diag->ioc_notify, bfa_diag_notify, diag);
5041 	list_add_tail(&diag->ioc_notify.qe, &diag->ioc->notify_q);
5042 }
5043 
5044 void
5045 bfa_diag_memclaim(struct bfa_diag_s *diag, u8 *dm_kva, u64 dm_pa)
5046 {
5047 	diag->fwping.dbuf_kva = dm_kva;
5048 	diag->fwping.dbuf_pa = dm_pa;
5049 	memset(diag->fwping.dbuf_kva, 0, BFI_DIAG_DMA_BUF_SZ);
5050 }
5051 
5052 /*
5053  *	PHY module specific
5054  */
5055 #define BFA_PHY_DMA_BUF_SZ	0x02000         /* 8k dma buffer */
5056 #define BFA_PHY_LOCK_STATUS	0x018878        /* phy semaphore status reg */
5057 
5058 static void
5059 bfa_phy_ntoh32(u32 *obuf, u32 *ibuf, int sz)
5060 {
5061 	int i, m = sz >> 2;
5062 
5063 	for (i = 0; i < m; i++)
5064 		obuf[i] = be32_to_cpu(ibuf[i]);
5065 }
5066 
5067 static bfa_boolean_t
5068 bfa_phy_present(struct bfa_phy_s *phy)
5069 {
5070 	return (phy->ioc->attr->card_type == BFA_MFG_TYPE_LIGHTNING);
5071 }
5072 
5073 static void
5074 bfa_phy_notify(void *cbarg, enum bfa_ioc_event_e event)
5075 {
5076 	struct bfa_phy_s *phy = cbarg;
5077 
5078 	bfa_trc(phy, event);
5079 
5080 	switch (event) {
5081 	case BFA_IOC_E_DISABLED:
5082 	case BFA_IOC_E_FAILED:
5083 		if (phy->op_busy) {
5084 			phy->status = BFA_STATUS_IOC_FAILURE;
5085 			phy->cbfn(phy->cbarg, phy->status);
5086 			phy->op_busy = 0;
5087 		}
5088 		break;
5089 
5090 	default:
5091 		break;
5092 	}
5093 }
5094 
5095 /*
5096  * Send phy attribute query request.
5097  *
5098  * @param[in] cbarg - callback argument
5099  */
5100 static void
5101 bfa_phy_query_send(void *cbarg)
5102 {
5103 	struct bfa_phy_s *phy = cbarg;
5104 	struct bfi_phy_query_req_s *msg =
5105 			(struct bfi_phy_query_req_s *) phy->mb.msg;
5106 
5107 	msg->instance = phy->instance;
5108 	bfi_h2i_set(msg->mh, BFI_MC_PHY, BFI_PHY_H2I_QUERY_REQ,
5109 		bfa_ioc_portid(phy->ioc));
5110 	bfa_alen_set(&msg->alen, sizeof(struct bfa_phy_attr_s), phy->dbuf_pa);
5111 	bfa_ioc_mbox_queue(phy->ioc, &phy->mb);
5112 }
5113 
5114 /*
5115  * Send phy write request.
5116  *
5117  * @param[in] cbarg - callback argument
5118  */
5119 static void
5120 bfa_phy_write_send(void *cbarg)
5121 {
5122 	struct bfa_phy_s *phy = cbarg;
5123 	struct bfi_phy_write_req_s *msg =
5124 			(struct bfi_phy_write_req_s *) phy->mb.msg;
5125 	u32	len;
5126 	u16	*buf, *dbuf;
5127 	int	i, sz;
5128 
5129 	msg->instance = phy->instance;
5130 	msg->offset = cpu_to_be32(phy->addr_off + phy->offset);
5131 	len = (phy->residue < BFA_PHY_DMA_BUF_SZ) ?
5132 			phy->residue : BFA_PHY_DMA_BUF_SZ;
5133 	msg->length = cpu_to_be32(len);
5134 
5135 	/* indicate if it's the last msg of the whole write operation */
5136 	msg->last = (len == phy->residue) ? 1 : 0;
5137 
5138 	bfi_h2i_set(msg->mh, BFI_MC_PHY, BFI_PHY_H2I_WRITE_REQ,
5139 		bfa_ioc_portid(phy->ioc));
5140 	bfa_alen_set(&msg->alen, len, phy->dbuf_pa);
5141 
5142 	buf = (u16 *) (phy->ubuf + phy->offset);
5143 	dbuf = (u16 *)phy->dbuf_kva;
5144 	sz = len >> 1;
5145 	for (i = 0; i < sz; i++)
5146 		buf[i] = cpu_to_be16(dbuf[i]);
5147 
5148 	bfa_ioc_mbox_queue(phy->ioc, &phy->mb);
5149 
5150 	phy->residue -= len;
5151 	phy->offset += len;
5152 }
5153 
5154 /*
5155  * Send phy read request.
5156  *
5157  * @param[in] cbarg - callback argument
5158  */
5159 static void
5160 bfa_phy_read_send(void *cbarg)
5161 {
5162 	struct bfa_phy_s *phy = cbarg;
5163 	struct bfi_phy_read_req_s *msg =
5164 			(struct bfi_phy_read_req_s *) phy->mb.msg;
5165 	u32	len;
5166 
5167 	msg->instance = phy->instance;
5168 	msg->offset = cpu_to_be32(phy->addr_off + phy->offset);
5169 	len = (phy->residue < BFA_PHY_DMA_BUF_SZ) ?
5170 			phy->residue : BFA_PHY_DMA_BUF_SZ;
5171 	msg->length = cpu_to_be32(len);
5172 	bfi_h2i_set(msg->mh, BFI_MC_PHY, BFI_PHY_H2I_READ_REQ,
5173 		bfa_ioc_portid(phy->ioc));
5174 	bfa_alen_set(&msg->alen, len, phy->dbuf_pa);
5175 	bfa_ioc_mbox_queue(phy->ioc, &phy->mb);
5176 }
5177 
5178 /*
5179  * Send phy stats request.
5180  *
5181  * @param[in] cbarg - callback argument
5182  */
5183 static void
5184 bfa_phy_stats_send(void *cbarg)
5185 {
5186 	struct bfa_phy_s *phy = cbarg;
5187 	struct bfi_phy_stats_req_s *msg =
5188 			(struct bfi_phy_stats_req_s *) phy->mb.msg;
5189 
5190 	msg->instance = phy->instance;
5191 	bfi_h2i_set(msg->mh, BFI_MC_PHY, BFI_PHY_H2I_STATS_REQ,
5192 		bfa_ioc_portid(phy->ioc));
5193 	bfa_alen_set(&msg->alen, sizeof(struct bfa_phy_stats_s), phy->dbuf_pa);
5194 	bfa_ioc_mbox_queue(phy->ioc, &phy->mb);
5195 }
5196 
5197 /*
5198  * Flash memory info API.
5199  *
5200  * @param[in] mincfg - minimal cfg variable
5201  */
5202 u32
5203 bfa_phy_meminfo(bfa_boolean_t mincfg)
5204 {
5205 	/* min driver doesn't need phy */
5206 	if (mincfg)
5207 		return 0;
5208 
5209 	return BFA_ROUNDUP(BFA_PHY_DMA_BUF_SZ, BFA_DMA_ALIGN_SZ);
5210 }
5211 
5212 /*
5213  * Flash attach API.
5214  *
5215  * @param[in] phy - phy structure
5216  * @param[in] ioc  - ioc structure
5217  * @param[in] dev  - device structure
5218  * @param[in] trcmod - trace module
5219  * @param[in] logmod - log module
5220  */
5221 void
5222 bfa_phy_attach(struct bfa_phy_s *phy, struct bfa_ioc_s *ioc, void *dev,
5223 		struct bfa_trc_mod_s *trcmod, bfa_boolean_t mincfg)
5224 {
5225 	phy->ioc = ioc;
5226 	phy->trcmod = trcmod;
5227 	phy->cbfn = NULL;
5228 	phy->cbarg = NULL;
5229 	phy->op_busy = 0;
5230 
5231 	bfa_ioc_mbox_regisr(phy->ioc, BFI_MC_PHY, bfa_phy_intr, phy);
5232 	bfa_q_qe_init(&phy->ioc_notify);
5233 	bfa_ioc_notify_init(&phy->ioc_notify, bfa_phy_notify, phy);
5234 	list_add_tail(&phy->ioc_notify.qe, &phy->ioc->notify_q);
5235 
5236 	/* min driver doesn't need phy */
5237 	if (mincfg) {
5238 		phy->dbuf_kva = NULL;
5239 		phy->dbuf_pa = 0;
5240 	}
5241 }
5242 
5243 /*
5244  * Claim memory for phy
5245  *
5246  * @param[in] phy - phy structure
5247  * @param[in] dm_kva - pointer to virtual memory address
5248  * @param[in] dm_pa - physical memory address
5249  * @param[in] mincfg - minimal cfg variable
5250  */
5251 void
5252 bfa_phy_memclaim(struct bfa_phy_s *phy, u8 *dm_kva, u64 dm_pa,
5253 		bfa_boolean_t mincfg)
5254 {
5255 	if (mincfg)
5256 		return;
5257 
5258 	phy->dbuf_kva = dm_kva;
5259 	phy->dbuf_pa = dm_pa;
5260 	memset(phy->dbuf_kva, 0, BFA_PHY_DMA_BUF_SZ);
5261 	dm_kva += BFA_ROUNDUP(BFA_PHY_DMA_BUF_SZ, BFA_DMA_ALIGN_SZ);
5262 	dm_pa += BFA_ROUNDUP(BFA_PHY_DMA_BUF_SZ, BFA_DMA_ALIGN_SZ);
5263 }
5264 
5265 bfa_boolean_t
5266 bfa_phy_busy(struct bfa_ioc_s *ioc)
5267 {
5268 	void __iomem	*rb;
5269 
5270 	rb = bfa_ioc_bar0(ioc);
5271 	return readl(rb + BFA_PHY_LOCK_STATUS);
5272 }
5273 
5274 /*
5275  * Get phy attribute.
5276  *
5277  * @param[in] phy - phy structure
5278  * @param[in] attr - phy attribute structure
5279  * @param[in] cbfn - callback function
5280  * @param[in] cbarg - callback argument
5281  *
5282  * Return status.
5283  */
5284 bfa_status_t
5285 bfa_phy_get_attr(struct bfa_phy_s *phy, u8 instance,
5286 		struct bfa_phy_attr_s *attr, bfa_cb_phy_t cbfn, void *cbarg)
5287 {
5288 	bfa_trc(phy, BFI_PHY_H2I_QUERY_REQ);
5289 	bfa_trc(phy, instance);
5290 
5291 	if (!bfa_phy_present(phy))
5292 		return BFA_STATUS_PHY_NOT_PRESENT;
5293 
5294 	if (!bfa_ioc_is_operational(phy->ioc))
5295 		return BFA_STATUS_IOC_NON_OP;
5296 
5297 	if (phy->op_busy || bfa_phy_busy(phy->ioc)) {
5298 		bfa_trc(phy, phy->op_busy);
5299 		return BFA_STATUS_DEVBUSY;
5300 	}
5301 
5302 	phy->op_busy = 1;
5303 	phy->cbfn = cbfn;
5304 	phy->cbarg = cbarg;
5305 	phy->instance = instance;
5306 	phy->ubuf = (uint8_t *) attr;
5307 	bfa_phy_query_send(phy);
5308 
5309 	return BFA_STATUS_OK;
5310 }
5311 
5312 /*
5313  * Get phy stats.
5314  *
5315  * @param[in] phy - phy structure
5316  * @param[in] instance - phy image instance
5317  * @param[in] stats - pointer to phy stats
5318  * @param[in] cbfn - callback function
5319  * @param[in] cbarg - callback argument
5320  *
5321  * Return status.
5322  */
5323 bfa_status_t
5324 bfa_phy_get_stats(struct bfa_phy_s *phy, u8 instance,
5325 		struct bfa_phy_stats_s *stats,
5326 		bfa_cb_phy_t cbfn, void *cbarg)
5327 {
5328 	bfa_trc(phy, BFI_PHY_H2I_STATS_REQ);
5329 	bfa_trc(phy, instance);
5330 
5331 	if (!bfa_phy_present(phy))
5332 		return BFA_STATUS_PHY_NOT_PRESENT;
5333 
5334 	if (!bfa_ioc_is_operational(phy->ioc))
5335 		return BFA_STATUS_IOC_NON_OP;
5336 
5337 	if (phy->op_busy || bfa_phy_busy(phy->ioc)) {
5338 		bfa_trc(phy, phy->op_busy);
5339 		return BFA_STATUS_DEVBUSY;
5340 	}
5341 
5342 	phy->op_busy = 1;
5343 	phy->cbfn = cbfn;
5344 	phy->cbarg = cbarg;
5345 	phy->instance = instance;
5346 	phy->ubuf = (u8 *) stats;
5347 	bfa_phy_stats_send(phy);
5348 
5349 	return BFA_STATUS_OK;
5350 }
5351 
5352 /*
5353  * Update phy image.
5354  *
5355  * @param[in] phy - phy structure
5356  * @param[in] instance - phy image instance
5357  * @param[in] buf - update data buffer
5358  * @param[in] len - data buffer length
5359  * @param[in] offset - offset relative to starting address
5360  * @param[in] cbfn - callback function
5361  * @param[in] cbarg - callback argument
5362  *
5363  * Return status.
5364  */
5365 bfa_status_t
5366 bfa_phy_update(struct bfa_phy_s *phy, u8 instance,
5367 		void *buf, u32 len, u32 offset,
5368 		bfa_cb_phy_t cbfn, void *cbarg)
5369 {
5370 	bfa_trc(phy, BFI_PHY_H2I_WRITE_REQ);
5371 	bfa_trc(phy, instance);
5372 	bfa_trc(phy, len);
5373 	bfa_trc(phy, offset);
5374 
5375 	if (!bfa_phy_present(phy))
5376 		return BFA_STATUS_PHY_NOT_PRESENT;
5377 
5378 	if (!bfa_ioc_is_operational(phy->ioc))
5379 		return BFA_STATUS_IOC_NON_OP;
5380 
5381 	/* 'len' must be in word (4-byte) boundary */
5382 	if (!len || (len & 0x03))
5383 		return BFA_STATUS_FAILED;
5384 
5385 	if (phy->op_busy || bfa_phy_busy(phy->ioc)) {
5386 		bfa_trc(phy, phy->op_busy);
5387 		return BFA_STATUS_DEVBUSY;
5388 	}
5389 
5390 	phy->op_busy = 1;
5391 	phy->cbfn = cbfn;
5392 	phy->cbarg = cbarg;
5393 	phy->instance = instance;
5394 	phy->residue = len;
5395 	phy->offset = 0;
5396 	phy->addr_off = offset;
5397 	phy->ubuf = buf;
5398 
5399 	bfa_phy_write_send(phy);
5400 	return BFA_STATUS_OK;
5401 }
5402 
5403 /*
5404  * Read phy image.
5405  *
5406  * @param[in] phy - phy structure
5407  * @param[in] instance - phy image instance
5408  * @param[in] buf - read data buffer
5409  * @param[in] len - data buffer length
5410  * @param[in] offset - offset relative to starting address
5411  * @param[in] cbfn - callback function
5412  * @param[in] cbarg - callback argument
5413  *
5414  * Return status.
5415  */
5416 bfa_status_t
5417 bfa_phy_read(struct bfa_phy_s *phy, u8 instance,
5418 		void *buf, u32 len, u32 offset,
5419 		bfa_cb_phy_t cbfn, void *cbarg)
5420 {
5421 	bfa_trc(phy, BFI_PHY_H2I_READ_REQ);
5422 	bfa_trc(phy, instance);
5423 	bfa_trc(phy, len);
5424 	bfa_trc(phy, offset);
5425 
5426 	if (!bfa_phy_present(phy))
5427 		return BFA_STATUS_PHY_NOT_PRESENT;
5428 
5429 	if (!bfa_ioc_is_operational(phy->ioc))
5430 		return BFA_STATUS_IOC_NON_OP;
5431 
5432 	/* 'len' must be in word (4-byte) boundary */
5433 	if (!len || (len & 0x03))
5434 		return BFA_STATUS_FAILED;
5435 
5436 	if (phy->op_busy || bfa_phy_busy(phy->ioc)) {
5437 		bfa_trc(phy, phy->op_busy);
5438 		return BFA_STATUS_DEVBUSY;
5439 	}
5440 
5441 	phy->op_busy = 1;
5442 	phy->cbfn = cbfn;
5443 	phy->cbarg = cbarg;
5444 	phy->instance = instance;
5445 	phy->residue = len;
5446 	phy->offset = 0;
5447 	phy->addr_off = offset;
5448 	phy->ubuf = buf;
5449 	bfa_phy_read_send(phy);
5450 
5451 	return BFA_STATUS_OK;
5452 }
5453 
5454 /*
5455  * Process phy response messages upon receiving interrupts.
5456  *
5457  * @param[in] phyarg - phy structure
5458  * @param[in] msg - message structure
5459  */
5460 void
5461 bfa_phy_intr(void *phyarg, struct bfi_mbmsg_s *msg)
5462 {
5463 	struct bfa_phy_s *phy = phyarg;
5464 	u32	status;
5465 
5466 	union {
5467 		struct bfi_phy_query_rsp_s *query;
5468 		struct bfi_phy_stats_rsp_s *stats;
5469 		struct bfi_phy_write_rsp_s *write;
5470 		struct bfi_phy_read_rsp_s *read;
5471 		struct bfi_mbmsg_s   *msg;
5472 	} m;
5473 
5474 	m.msg = msg;
5475 	bfa_trc(phy, msg->mh.msg_id);
5476 
5477 	if (!phy->op_busy) {
5478 		/* receiving response after ioc failure */
5479 		bfa_trc(phy, 0x9999);
5480 		return;
5481 	}
5482 
5483 	switch (msg->mh.msg_id) {
5484 	case BFI_PHY_I2H_QUERY_RSP:
5485 		status = be32_to_cpu(m.query->status);
5486 		bfa_trc(phy, status);
5487 
5488 		if (status == BFA_STATUS_OK) {
5489 			struct bfa_phy_attr_s *attr =
5490 				(struct bfa_phy_attr_s *) phy->ubuf;
5491 			bfa_phy_ntoh32((u32 *)attr, (u32 *)phy->dbuf_kva,
5492 					sizeof(struct bfa_phy_attr_s));
5493 			bfa_trc(phy, attr->status);
5494 			bfa_trc(phy, attr->length);
5495 		}
5496 
5497 		phy->status = status;
5498 		phy->op_busy = 0;
5499 		if (phy->cbfn)
5500 			phy->cbfn(phy->cbarg, phy->status);
5501 		break;
5502 	case BFI_PHY_I2H_STATS_RSP:
5503 		status = be32_to_cpu(m.stats->status);
5504 		bfa_trc(phy, status);
5505 
5506 		if (status == BFA_STATUS_OK) {
5507 			struct bfa_phy_stats_s *stats =
5508 				(struct bfa_phy_stats_s *) phy->ubuf;
5509 			bfa_phy_ntoh32((u32 *)stats, (u32 *)phy->dbuf_kva,
5510 				sizeof(struct bfa_phy_stats_s));
5511 				bfa_trc(phy, stats->status);
5512 		}
5513 
5514 		phy->status = status;
5515 		phy->op_busy = 0;
5516 		if (phy->cbfn)
5517 			phy->cbfn(phy->cbarg, phy->status);
5518 		break;
5519 	case BFI_PHY_I2H_WRITE_RSP:
5520 		status = be32_to_cpu(m.write->status);
5521 		bfa_trc(phy, status);
5522 
5523 		if (status != BFA_STATUS_OK || phy->residue == 0) {
5524 			phy->status = status;
5525 			phy->op_busy = 0;
5526 			if (phy->cbfn)
5527 				phy->cbfn(phy->cbarg, phy->status);
5528 		} else {
5529 			bfa_trc(phy, phy->offset);
5530 			bfa_phy_write_send(phy);
5531 		}
5532 		break;
5533 	case BFI_PHY_I2H_READ_RSP:
5534 		status = be32_to_cpu(m.read->status);
5535 		bfa_trc(phy, status);
5536 
5537 		if (status != BFA_STATUS_OK) {
5538 			phy->status = status;
5539 			phy->op_busy = 0;
5540 			if (phy->cbfn)
5541 				phy->cbfn(phy->cbarg, phy->status);
5542 		} else {
5543 			u32 len = be32_to_cpu(m.read->length);
5544 			u16 *buf = (u16 *)(phy->ubuf + phy->offset);
5545 			u16 *dbuf = (u16 *)phy->dbuf_kva;
5546 			int i, sz = len >> 1;
5547 
5548 			bfa_trc(phy, phy->offset);
5549 			bfa_trc(phy, len);
5550 
5551 			for (i = 0; i < sz; i++)
5552 				buf[i] = be16_to_cpu(dbuf[i]);
5553 
5554 			phy->residue -= len;
5555 			phy->offset += len;
5556 
5557 			if (phy->residue == 0) {
5558 				phy->status = status;
5559 				phy->op_busy = 0;
5560 				if (phy->cbfn)
5561 					phy->cbfn(phy->cbarg, phy->status);
5562 			} else
5563 				bfa_phy_read_send(phy);
5564 		}
5565 		break;
5566 	default:
5567 		WARN_ON(1);
5568 	}
5569 }
5570 
5571 /*
5572  *	DCONF module specific
5573  */
5574 
5575 BFA_MODULE(dconf);
5576 
5577 /*
5578  * DCONF state machine events
5579  */
5580 enum bfa_dconf_event {
5581 	BFA_DCONF_SM_INIT		= 1,	/* dconf Init */
5582 	BFA_DCONF_SM_FLASH_COMP		= 2,	/* read/write to flash */
5583 	BFA_DCONF_SM_WR			= 3,	/* binding change, map */
5584 	BFA_DCONF_SM_TIMEOUT		= 4,	/* Start timer */
5585 	BFA_DCONF_SM_EXIT		= 5,	/* exit dconf module */
5586 	BFA_DCONF_SM_IOCDISABLE		= 6,	/* IOC disable event */
5587 };
5588 
5589 /* forward declaration of DCONF state machine */
5590 static void bfa_dconf_sm_uninit(struct bfa_dconf_mod_s *dconf,
5591 				enum bfa_dconf_event event);
5592 static void bfa_dconf_sm_flash_read(struct bfa_dconf_mod_s *dconf,
5593 				enum bfa_dconf_event event);
5594 static void bfa_dconf_sm_ready(struct bfa_dconf_mod_s *dconf,
5595 				enum bfa_dconf_event event);
5596 static void bfa_dconf_sm_dirty(struct bfa_dconf_mod_s *dconf,
5597 				enum bfa_dconf_event event);
5598 static void bfa_dconf_sm_sync(struct bfa_dconf_mod_s *dconf,
5599 				enum bfa_dconf_event event);
5600 static void bfa_dconf_sm_final_sync(struct bfa_dconf_mod_s *dconf,
5601 				enum bfa_dconf_event event);
5602 static void bfa_dconf_sm_iocdown_dirty(struct bfa_dconf_mod_s *dconf,
5603 				enum bfa_dconf_event event);
5604 
5605 static void bfa_dconf_cbfn(void *dconf, bfa_status_t status);
5606 static void bfa_dconf_timer(void *cbarg);
5607 static bfa_status_t bfa_dconf_flash_write(struct bfa_dconf_mod_s *dconf);
5608 static void bfa_dconf_init_cb(void *arg, bfa_status_t status);
5609 
5610 /*
5611  * Beginning state of dconf module. Waiting for an event to start.
5612  */
5613 static void
5614 bfa_dconf_sm_uninit(struct bfa_dconf_mod_s *dconf, enum bfa_dconf_event event)
5615 {
5616 	bfa_status_t bfa_status;
5617 	bfa_trc(dconf->bfa, event);
5618 
5619 	switch (event) {
5620 	case BFA_DCONF_SM_INIT:
5621 		if (dconf->min_cfg) {
5622 			bfa_trc(dconf->bfa, dconf->min_cfg);
5623 			bfa_fsm_send_event(&dconf->bfa->iocfc,
5624 					IOCFC_E_DCONF_DONE);
5625 			return;
5626 		}
5627 		bfa_sm_set_state(dconf, bfa_dconf_sm_flash_read);
5628 		bfa_timer_start(dconf->bfa, &dconf->timer,
5629 			bfa_dconf_timer, dconf, 2 * BFA_DCONF_UPDATE_TOV);
5630 		bfa_status = bfa_flash_read_part(BFA_FLASH(dconf->bfa),
5631 					BFA_FLASH_PART_DRV, dconf->instance,
5632 					dconf->dconf,
5633 					sizeof(struct bfa_dconf_s), 0,
5634 					bfa_dconf_init_cb, dconf->bfa);
5635 		if (bfa_status != BFA_STATUS_OK) {
5636 			bfa_timer_stop(&dconf->timer);
5637 			bfa_dconf_init_cb(dconf->bfa, BFA_STATUS_FAILED);
5638 			bfa_sm_set_state(dconf, bfa_dconf_sm_uninit);
5639 			return;
5640 		}
5641 		break;
5642 	case BFA_DCONF_SM_EXIT:
5643 		bfa_fsm_send_event(&dconf->bfa->iocfc, IOCFC_E_DCONF_DONE);
5644 	case BFA_DCONF_SM_IOCDISABLE:
5645 	case BFA_DCONF_SM_WR:
5646 	case BFA_DCONF_SM_FLASH_COMP:
5647 		break;
5648 	default:
5649 		bfa_sm_fault(dconf->bfa, event);
5650 	}
5651 }
5652 
5653 /*
5654  * Read flash for dconf entries and make a call back to the driver once done.
5655  */
5656 static void
5657 bfa_dconf_sm_flash_read(struct bfa_dconf_mod_s *dconf,
5658 			enum bfa_dconf_event event)
5659 {
5660 	bfa_trc(dconf->bfa, event);
5661 
5662 	switch (event) {
5663 	case BFA_DCONF_SM_FLASH_COMP:
5664 		bfa_timer_stop(&dconf->timer);
5665 		bfa_sm_set_state(dconf, bfa_dconf_sm_ready);
5666 		break;
5667 	case BFA_DCONF_SM_TIMEOUT:
5668 		bfa_sm_set_state(dconf, bfa_dconf_sm_ready);
5669 		bfa_ioc_suspend(&dconf->bfa->ioc);
5670 		break;
5671 	case BFA_DCONF_SM_EXIT:
5672 		bfa_timer_stop(&dconf->timer);
5673 		bfa_sm_set_state(dconf, bfa_dconf_sm_uninit);
5674 		bfa_fsm_send_event(&dconf->bfa->iocfc, IOCFC_E_DCONF_DONE);
5675 		break;
5676 	case BFA_DCONF_SM_IOCDISABLE:
5677 		bfa_timer_stop(&dconf->timer);
5678 		bfa_sm_set_state(dconf, bfa_dconf_sm_uninit);
5679 		break;
5680 	default:
5681 		bfa_sm_fault(dconf->bfa, event);
5682 	}
5683 }
5684 
5685 /*
5686  * DCONF Module is in ready state. Has completed the initialization.
5687  */
5688 static void
5689 bfa_dconf_sm_ready(struct bfa_dconf_mod_s *dconf, enum bfa_dconf_event event)
5690 {
5691 	bfa_trc(dconf->bfa, event);
5692 
5693 	switch (event) {
5694 	case BFA_DCONF_SM_WR:
5695 		bfa_timer_start(dconf->bfa, &dconf->timer,
5696 			bfa_dconf_timer, dconf, BFA_DCONF_UPDATE_TOV);
5697 		bfa_sm_set_state(dconf, bfa_dconf_sm_dirty);
5698 		break;
5699 	case BFA_DCONF_SM_EXIT:
5700 		bfa_sm_set_state(dconf, bfa_dconf_sm_uninit);
5701 		bfa_fsm_send_event(&dconf->bfa->iocfc, IOCFC_E_DCONF_DONE);
5702 		break;
5703 	case BFA_DCONF_SM_INIT:
5704 	case BFA_DCONF_SM_IOCDISABLE:
5705 		break;
5706 	default:
5707 		bfa_sm_fault(dconf->bfa, event);
5708 	}
5709 }
5710 
5711 /*
5712  * entries are dirty, write back to the flash.
5713  */
5714 
5715 static void
5716 bfa_dconf_sm_dirty(struct bfa_dconf_mod_s *dconf, enum bfa_dconf_event event)
5717 {
5718 	bfa_trc(dconf->bfa, event);
5719 
5720 	switch (event) {
5721 	case BFA_DCONF_SM_TIMEOUT:
5722 		bfa_sm_set_state(dconf, bfa_dconf_sm_sync);
5723 		bfa_dconf_flash_write(dconf);
5724 		break;
5725 	case BFA_DCONF_SM_WR:
5726 		bfa_timer_stop(&dconf->timer);
5727 		bfa_timer_start(dconf->bfa, &dconf->timer,
5728 			bfa_dconf_timer, dconf, BFA_DCONF_UPDATE_TOV);
5729 		break;
5730 	case BFA_DCONF_SM_EXIT:
5731 		bfa_timer_stop(&dconf->timer);
5732 		bfa_timer_start(dconf->bfa, &dconf->timer,
5733 			bfa_dconf_timer, dconf, BFA_DCONF_UPDATE_TOV);
5734 		bfa_sm_set_state(dconf, bfa_dconf_sm_final_sync);
5735 		bfa_dconf_flash_write(dconf);
5736 		break;
5737 	case BFA_DCONF_SM_FLASH_COMP:
5738 		break;
5739 	case BFA_DCONF_SM_IOCDISABLE:
5740 		bfa_timer_stop(&dconf->timer);
5741 		bfa_sm_set_state(dconf, bfa_dconf_sm_iocdown_dirty);
5742 		break;
5743 	default:
5744 		bfa_sm_fault(dconf->bfa, event);
5745 	}
5746 }
5747 
5748 /*
5749  * Sync the dconf entries to the flash.
5750  */
5751 static void
5752 bfa_dconf_sm_final_sync(struct bfa_dconf_mod_s *dconf,
5753 			enum bfa_dconf_event event)
5754 {
5755 	bfa_trc(dconf->bfa, event);
5756 
5757 	switch (event) {
5758 	case BFA_DCONF_SM_IOCDISABLE:
5759 	case BFA_DCONF_SM_FLASH_COMP:
5760 		bfa_timer_stop(&dconf->timer);
5761 	case BFA_DCONF_SM_TIMEOUT:
5762 		bfa_sm_set_state(dconf, bfa_dconf_sm_uninit);
5763 		bfa_fsm_send_event(&dconf->bfa->iocfc, IOCFC_E_DCONF_DONE);
5764 		break;
5765 	default:
5766 		bfa_sm_fault(dconf->bfa, event);
5767 	}
5768 }
5769 
5770 static void
5771 bfa_dconf_sm_sync(struct bfa_dconf_mod_s *dconf, enum bfa_dconf_event event)
5772 {
5773 	bfa_trc(dconf->bfa, event);
5774 
5775 	switch (event) {
5776 	case BFA_DCONF_SM_FLASH_COMP:
5777 		bfa_sm_set_state(dconf, bfa_dconf_sm_ready);
5778 		break;
5779 	case BFA_DCONF_SM_WR:
5780 		bfa_timer_start(dconf->bfa, &dconf->timer,
5781 			bfa_dconf_timer, dconf, BFA_DCONF_UPDATE_TOV);
5782 		bfa_sm_set_state(dconf, bfa_dconf_sm_dirty);
5783 		break;
5784 	case BFA_DCONF_SM_EXIT:
5785 		bfa_timer_start(dconf->bfa, &dconf->timer,
5786 			bfa_dconf_timer, dconf, BFA_DCONF_UPDATE_TOV);
5787 		bfa_sm_set_state(dconf, bfa_dconf_sm_final_sync);
5788 		break;
5789 	case BFA_DCONF_SM_IOCDISABLE:
5790 		bfa_sm_set_state(dconf, bfa_dconf_sm_iocdown_dirty);
5791 		break;
5792 	default:
5793 		bfa_sm_fault(dconf->bfa, event);
5794 	}
5795 }
5796 
5797 static void
5798 bfa_dconf_sm_iocdown_dirty(struct bfa_dconf_mod_s *dconf,
5799 			enum bfa_dconf_event event)
5800 {
5801 	bfa_trc(dconf->bfa, event);
5802 
5803 	switch (event) {
5804 	case BFA_DCONF_SM_INIT:
5805 		bfa_timer_start(dconf->bfa, &dconf->timer,
5806 			bfa_dconf_timer, dconf, BFA_DCONF_UPDATE_TOV);
5807 		bfa_sm_set_state(dconf, bfa_dconf_sm_dirty);
5808 		break;
5809 	case BFA_DCONF_SM_EXIT:
5810 		bfa_sm_set_state(dconf, bfa_dconf_sm_uninit);
5811 		bfa_fsm_send_event(&dconf->bfa->iocfc, IOCFC_E_DCONF_DONE);
5812 		break;
5813 	case BFA_DCONF_SM_IOCDISABLE:
5814 		break;
5815 	default:
5816 		bfa_sm_fault(dconf->bfa, event);
5817 	}
5818 }
5819 
5820 /*
5821  * Compute and return memory needed by DRV_CFG module.
5822  */
5823 static void
5824 bfa_dconf_meminfo(struct bfa_iocfc_cfg_s *cfg, struct bfa_meminfo_s *meminfo,
5825 		  struct bfa_s *bfa)
5826 {
5827 	struct bfa_mem_kva_s *dconf_kva = BFA_MEM_DCONF_KVA(bfa);
5828 
5829 	if (cfg->drvcfg.min_cfg)
5830 		bfa_mem_kva_setup(meminfo, dconf_kva,
5831 				sizeof(struct bfa_dconf_hdr_s));
5832 	else
5833 		bfa_mem_kva_setup(meminfo, dconf_kva,
5834 				sizeof(struct bfa_dconf_s));
5835 }
5836 
5837 static void
5838 bfa_dconf_attach(struct bfa_s *bfa, void *bfad, struct bfa_iocfc_cfg_s *cfg,
5839 		struct bfa_pcidev_s *pcidev)
5840 {
5841 	struct bfa_dconf_mod_s *dconf = BFA_DCONF_MOD(bfa);
5842 
5843 	dconf->bfad = bfad;
5844 	dconf->bfa = bfa;
5845 	dconf->instance = bfa->ioc.port_id;
5846 	bfa_trc(bfa, dconf->instance);
5847 
5848 	dconf->dconf = (struct bfa_dconf_s *) bfa_mem_kva_curp(dconf);
5849 	if (cfg->drvcfg.min_cfg) {
5850 		bfa_mem_kva_curp(dconf) += sizeof(struct bfa_dconf_hdr_s);
5851 		dconf->min_cfg = BFA_TRUE;
5852 	} else {
5853 		dconf->min_cfg = BFA_FALSE;
5854 		bfa_mem_kva_curp(dconf) += sizeof(struct bfa_dconf_s);
5855 	}
5856 
5857 	bfa_dconf_read_data_valid(bfa) = BFA_FALSE;
5858 	bfa_sm_set_state(dconf, bfa_dconf_sm_uninit);
5859 }
5860 
5861 static void
5862 bfa_dconf_init_cb(void *arg, bfa_status_t status)
5863 {
5864 	struct bfa_s *bfa = arg;
5865 	struct bfa_dconf_mod_s *dconf = BFA_DCONF_MOD(bfa);
5866 
5867 	if (status == BFA_STATUS_OK) {
5868 		bfa_dconf_read_data_valid(bfa) = BFA_TRUE;
5869 		if (dconf->dconf->hdr.signature != BFI_DCONF_SIGNATURE)
5870 			dconf->dconf->hdr.signature = BFI_DCONF_SIGNATURE;
5871 		if (dconf->dconf->hdr.version != BFI_DCONF_VERSION)
5872 			dconf->dconf->hdr.version = BFI_DCONF_VERSION;
5873 	}
5874 	bfa_sm_send_event(dconf, BFA_DCONF_SM_FLASH_COMP);
5875 	bfa_fsm_send_event(&bfa->iocfc, IOCFC_E_DCONF_DONE);
5876 }
5877 
5878 void
5879 bfa_dconf_modinit(struct bfa_s *bfa)
5880 {
5881 	struct bfa_dconf_mod_s *dconf = BFA_DCONF_MOD(bfa);
5882 	bfa_sm_send_event(dconf, BFA_DCONF_SM_INIT);
5883 }
5884 static void
5885 bfa_dconf_start(struct bfa_s *bfa)
5886 {
5887 }
5888 
5889 static void
5890 bfa_dconf_stop(struct bfa_s *bfa)
5891 {
5892 }
5893 
5894 static void bfa_dconf_timer(void *cbarg)
5895 {
5896 	struct bfa_dconf_mod_s *dconf = cbarg;
5897 	bfa_sm_send_event(dconf, BFA_DCONF_SM_TIMEOUT);
5898 }
5899 static void
5900 bfa_dconf_iocdisable(struct bfa_s *bfa)
5901 {
5902 	struct bfa_dconf_mod_s *dconf = BFA_DCONF_MOD(bfa);
5903 	bfa_sm_send_event(dconf, BFA_DCONF_SM_IOCDISABLE);
5904 }
5905 
5906 static void
5907 bfa_dconf_detach(struct bfa_s *bfa)
5908 {
5909 }
5910 
5911 static bfa_status_t
5912 bfa_dconf_flash_write(struct bfa_dconf_mod_s *dconf)
5913 {
5914 	bfa_status_t bfa_status;
5915 	bfa_trc(dconf->bfa, 0);
5916 
5917 	bfa_status = bfa_flash_update_part(BFA_FLASH(dconf->bfa),
5918 				BFA_FLASH_PART_DRV, dconf->instance,
5919 				dconf->dconf,  sizeof(struct bfa_dconf_s), 0,
5920 				bfa_dconf_cbfn, dconf);
5921 	if (bfa_status != BFA_STATUS_OK)
5922 		WARN_ON(bfa_status);
5923 	bfa_trc(dconf->bfa, bfa_status);
5924 
5925 	return bfa_status;
5926 }
5927 
5928 bfa_status_t
5929 bfa_dconf_update(struct bfa_s *bfa)
5930 {
5931 	struct bfa_dconf_mod_s *dconf = BFA_DCONF_MOD(bfa);
5932 	bfa_trc(dconf->bfa, 0);
5933 	if (bfa_sm_cmp_state(dconf, bfa_dconf_sm_iocdown_dirty))
5934 		return BFA_STATUS_FAILED;
5935 
5936 	if (dconf->min_cfg) {
5937 		bfa_trc(dconf->bfa, dconf->min_cfg);
5938 		return BFA_STATUS_FAILED;
5939 	}
5940 
5941 	bfa_sm_send_event(dconf, BFA_DCONF_SM_WR);
5942 	return BFA_STATUS_OK;
5943 }
5944 
5945 static void
5946 bfa_dconf_cbfn(void *arg, bfa_status_t status)
5947 {
5948 	struct bfa_dconf_mod_s *dconf = arg;
5949 	WARN_ON(status);
5950 	bfa_sm_send_event(dconf, BFA_DCONF_SM_FLASH_COMP);
5951 }
5952 
5953 void
5954 bfa_dconf_modexit(struct bfa_s *bfa)
5955 {
5956 	struct bfa_dconf_mod_s *dconf = BFA_DCONF_MOD(bfa);
5957 	bfa_sm_send_event(dconf, BFA_DCONF_SM_EXIT);
5958 }
5959 
5960 /*
5961  * FRU specific functions
5962  */
5963 
5964 #define BFA_FRU_DMA_BUF_SZ	0x02000		/* 8k dma buffer */
5965 #define BFA_FRU_CHINOOK_MAX_SIZE 0x10000
5966 #define BFA_FRU_LIGHTNING_MAX_SIZE 0x200
5967 
5968 static void
5969 bfa_fru_notify(void *cbarg, enum bfa_ioc_event_e event)
5970 {
5971 	struct bfa_fru_s *fru = cbarg;
5972 
5973 	bfa_trc(fru, event);
5974 
5975 	switch (event) {
5976 	case BFA_IOC_E_DISABLED:
5977 	case BFA_IOC_E_FAILED:
5978 		if (fru->op_busy) {
5979 			fru->status = BFA_STATUS_IOC_FAILURE;
5980 			fru->cbfn(fru->cbarg, fru->status);
5981 			fru->op_busy = 0;
5982 		}
5983 		break;
5984 
5985 	default:
5986 		break;
5987 	}
5988 }
5989 
5990 /*
5991  * Send fru write request.
5992  *
5993  * @param[in] cbarg - callback argument
5994  */
5995 static void
5996 bfa_fru_write_send(void *cbarg, enum bfi_fru_h2i_msgs msg_type)
5997 {
5998 	struct bfa_fru_s *fru = cbarg;
5999 	struct bfi_fru_write_req_s *msg =
6000 			(struct bfi_fru_write_req_s *) fru->mb.msg;
6001 	u32 len;
6002 
6003 	msg->offset = cpu_to_be32(fru->addr_off + fru->offset);
6004 	len = (fru->residue < BFA_FRU_DMA_BUF_SZ) ?
6005 				fru->residue : BFA_FRU_DMA_BUF_SZ;
6006 	msg->length = cpu_to_be32(len);
6007 
6008 	/*
6009 	 * indicate if it's the last msg of the whole write operation
6010 	 */
6011 	msg->last = (len == fru->residue) ? 1 : 0;
6012 
6013 	bfi_h2i_set(msg->mh, BFI_MC_FRU, msg_type, bfa_ioc_portid(fru->ioc));
6014 	bfa_alen_set(&msg->alen, len, fru->dbuf_pa);
6015 
6016 	memcpy(fru->dbuf_kva, fru->ubuf + fru->offset, len);
6017 	bfa_ioc_mbox_queue(fru->ioc, &fru->mb);
6018 
6019 	fru->residue -= len;
6020 	fru->offset += len;
6021 }
6022 
6023 /*
6024  * Send fru read request.
6025  *
6026  * @param[in] cbarg - callback argument
6027  */
6028 static void
6029 bfa_fru_read_send(void *cbarg, enum bfi_fru_h2i_msgs msg_type)
6030 {
6031 	struct bfa_fru_s *fru = cbarg;
6032 	struct bfi_fru_read_req_s *msg =
6033 			(struct bfi_fru_read_req_s *) fru->mb.msg;
6034 	u32 len;
6035 
6036 	msg->offset = cpu_to_be32(fru->addr_off + fru->offset);
6037 	len = (fru->residue < BFA_FRU_DMA_BUF_SZ) ?
6038 				fru->residue : BFA_FRU_DMA_BUF_SZ;
6039 	msg->length = cpu_to_be32(len);
6040 	bfi_h2i_set(msg->mh, BFI_MC_FRU, msg_type, bfa_ioc_portid(fru->ioc));
6041 	bfa_alen_set(&msg->alen, len, fru->dbuf_pa);
6042 	bfa_ioc_mbox_queue(fru->ioc, &fru->mb);
6043 }
6044 
6045 /*
6046  * Flash memory info API.
6047  *
6048  * @param[in] mincfg - minimal cfg variable
6049  */
6050 u32
6051 bfa_fru_meminfo(bfa_boolean_t mincfg)
6052 {
6053 	/* min driver doesn't need fru */
6054 	if (mincfg)
6055 		return 0;
6056 
6057 	return BFA_ROUNDUP(BFA_FRU_DMA_BUF_SZ, BFA_DMA_ALIGN_SZ);
6058 }
6059 
6060 /*
6061  * Flash attach API.
6062  *
6063  * @param[in] fru - fru structure
6064  * @param[in] ioc  - ioc structure
6065  * @param[in] dev  - device structure
6066  * @param[in] trcmod - trace module
6067  * @param[in] logmod - log module
6068  */
6069 void
6070 bfa_fru_attach(struct bfa_fru_s *fru, struct bfa_ioc_s *ioc, void *dev,
6071 	struct bfa_trc_mod_s *trcmod, bfa_boolean_t mincfg)
6072 {
6073 	fru->ioc = ioc;
6074 	fru->trcmod = trcmod;
6075 	fru->cbfn = NULL;
6076 	fru->cbarg = NULL;
6077 	fru->op_busy = 0;
6078 
6079 	bfa_ioc_mbox_regisr(fru->ioc, BFI_MC_FRU, bfa_fru_intr, fru);
6080 	bfa_q_qe_init(&fru->ioc_notify);
6081 	bfa_ioc_notify_init(&fru->ioc_notify, bfa_fru_notify, fru);
6082 	list_add_tail(&fru->ioc_notify.qe, &fru->ioc->notify_q);
6083 
6084 	/* min driver doesn't need fru */
6085 	if (mincfg) {
6086 		fru->dbuf_kva = NULL;
6087 		fru->dbuf_pa = 0;
6088 	}
6089 }
6090 
6091 /*
6092  * Claim memory for fru
6093  *
6094  * @param[in] fru - fru structure
6095  * @param[in] dm_kva - pointer to virtual memory address
6096  * @param[in] dm_pa - frusical memory address
6097  * @param[in] mincfg - minimal cfg variable
6098  */
6099 void
6100 bfa_fru_memclaim(struct bfa_fru_s *fru, u8 *dm_kva, u64 dm_pa,
6101 	bfa_boolean_t mincfg)
6102 {
6103 	if (mincfg)
6104 		return;
6105 
6106 	fru->dbuf_kva = dm_kva;
6107 	fru->dbuf_pa = dm_pa;
6108 	memset(fru->dbuf_kva, 0, BFA_FRU_DMA_BUF_SZ);
6109 	dm_kva += BFA_ROUNDUP(BFA_FRU_DMA_BUF_SZ, BFA_DMA_ALIGN_SZ);
6110 	dm_pa += BFA_ROUNDUP(BFA_FRU_DMA_BUF_SZ, BFA_DMA_ALIGN_SZ);
6111 }
6112 
6113 /*
6114  * Update fru vpd image.
6115  *
6116  * @param[in] fru - fru structure
6117  * @param[in] buf - update data buffer
6118  * @param[in] len - data buffer length
6119  * @param[in] offset - offset relative to starting address
6120  * @param[in] cbfn - callback function
6121  * @param[in] cbarg - callback argument
6122  *
6123  * Return status.
6124  */
6125 bfa_status_t
6126 bfa_fruvpd_update(struct bfa_fru_s *fru, void *buf, u32 len, u32 offset,
6127 		  bfa_cb_fru_t cbfn, void *cbarg)
6128 {
6129 	bfa_trc(fru, BFI_FRUVPD_H2I_WRITE_REQ);
6130 	bfa_trc(fru, len);
6131 	bfa_trc(fru, offset);
6132 
6133 	if (fru->ioc->asic_gen != BFI_ASIC_GEN_CT2)
6134 		return BFA_STATUS_FRU_NOT_PRESENT;
6135 
6136 	if (fru->ioc->attr->card_type != BFA_MFG_TYPE_CHINOOK)
6137 		return BFA_STATUS_CMD_NOTSUPP;
6138 
6139 	if (!bfa_ioc_is_operational(fru->ioc))
6140 		return BFA_STATUS_IOC_NON_OP;
6141 
6142 	if (fru->op_busy) {
6143 		bfa_trc(fru, fru->op_busy);
6144 		return BFA_STATUS_DEVBUSY;
6145 	}
6146 
6147 	fru->op_busy = 1;
6148 
6149 	fru->cbfn = cbfn;
6150 	fru->cbarg = cbarg;
6151 	fru->residue = len;
6152 	fru->offset = 0;
6153 	fru->addr_off = offset;
6154 	fru->ubuf = buf;
6155 
6156 	bfa_fru_write_send(fru, BFI_FRUVPD_H2I_WRITE_REQ);
6157 
6158 	return BFA_STATUS_OK;
6159 }
6160 
6161 /*
6162  * Read fru vpd image.
6163  *
6164  * @param[in] fru - fru structure
6165  * @param[in] buf - read data buffer
6166  * @param[in] len - data buffer length
6167  * @param[in] offset - offset relative to starting address
6168  * @param[in] cbfn - callback function
6169  * @param[in] cbarg - callback argument
6170  *
6171  * Return status.
6172  */
6173 bfa_status_t
6174 bfa_fruvpd_read(struct bfa_fru_s *fru, void *buf, u32 len, u32 offset,
6175 		bfa_cb_fru_t cbfn, void *cbarg)
6176 {
6177 	bfa_trc(fru, BFI_FRUVPD_H2I_READ_REQ);
6178 	bfa_trc(fru, len);
6179 	bfa_trc(fru, offset);
6180 
6181 	if (fru->ioc->asic_gen != BFI_ASIC_GEN_CT2)
6182 		return BFA_STATUS_FRU_NOT_PRESENT;
6183 
6184 	if (fru->ioc->attr->card_type != BFA_MFG_TYPE_CHINOOK)
6185 		return BFA_STATUS_CMD_NOTSUPP;
6186 
6187 	if (!bfa_ioc_is_operational(fru->ioc))
6188 		return BFA_STATUS_IOC_NON_OP;
6189 
6190 	if (fru->op_busy) {
6191 		bfa_trc(fru, fru->op_busy);
6192 		return BFA_STATUS_DEVBUSY;
6193 	}
6194 
6195 	fru->op_busy = 1;
6196 
6197 	fru->cbfn = cbfn;
6198 	fru->cbarg = cbarg;
6199 	fru->residue = len;
6200 	fru->offset = 0;
6201 	fru->addr_off = offset;
6202 	fru->ubuf = buf;
6203 	bfa_fru_read_send(fru, BFI_FRUVPD_H2I_READ_REQ);
6204 
6205 	return BFA_STATUS_OK;
6206 }
6207 
6208 /*
6209  * Get maximum size fru vpd image.
6210  *
6211  * @param[in] fru - fru structure
6212  * @param[out] size - maximum size of fru vpd data
6213  *
6214  * Return status.
6215  */
6216 bfa_status_t
6217 bfa_fruvpd_get_max_size(struct bfa_fru_s *fru, u32 *max_size)
6218 {
6219 	if (fru->ioc->asic_gen != BFI_ASIC_GEN_CT2)
6220 		return BFA_STATUS_FRU_NOT_PRESENT;
6221 
6222 	if (!bfa_ioc_is_operational(fru->ioc))
6223 		return BFA_STATUS_IOC_NON_OP;
6224 
6225 	if (fru->ioc->attr->card_type == BFA_MFG_TYPE_CHINOOK)
6226 		*max_size = BFA_FRU_CHINOOK_MAX_SIZE;
6227 	else
6228 		return BFA_STATUS_CMD_NOTSUPP;
6229 	return BFA_STATUS_OK;
6230 }
6231 /*
6232  * tfru write.
6233  *
6234  * @param[in] fru - fru structure
6235  * @param[in] buf - update data buffer
6236  * @param[in] len - data buffer length
6237  * @param[in] offset - offset relative to starting address
6238  * @param[in] cbfn - callback function
6239  * @param[in] cbarg - callback argument
6240  *
6241  * Return status.
6242  */
6243 bfa_status_t
6244 bfa_tfru_write(struct bfa_fru_s *fru, void *buf, u32 len, u32 offset,
6245 	       bfa_cb_fru_t cbfn, void *cbarg)
6246 {
6247 	bfa_trc(fru, BFI_TFRU_H2I_WRITE_REQ);
6248 	bfa_trc(fru, len);
6249 	bfa_trc(fru, offset);
6250 	bfa_trc(fru, *((u8 *) buf));
6251 
6252 	if (fru->ioc->asic_gen != BFI_ASIC_GEN_CT2)
6253 		return BFA_STATUS_FRU_NOT_PRESENT;
6254 
6255 	if (!bfa_ioc_is_operational(fru->ioc))
6256 		return BFA_STATUS_IOC_NON_OP;
6257 
6258 	if (fru->op_busy) {
6259 		bfa_trc(fru, fru->op_busy);
6260 		return BFA_STATUS_DEVBUSY;
6261 	}
6262 
6263 	fru->op_busy = 1;
6264 
6265 	fru->cbfn = cbfn;
6266 	fru->cbarg = cbarg;
6267 	fru->residue = len;
6268 	fru->offset = 0;
6269 	fru->addr_off = offset;
6270 	fru->ubuf = buf;
6271 
6272 	bfa_fru_write_send(fru, BFI_TFRU_H2I_WRITE_REQ);
6273 
6274 	return BFA_STATUS_OK;
6275 }
6276 
6277 /*
6278  * tfru read.
6279  *
6280  * @param[in] fru - fru structure
6281  * @param[in] buf - read data buffer
6282  * @param[in] len - data buffer length
6283  * @param[in] offset - offset relative to starting address
6284  * @param[in] cbfn - callback function
6285  * @param[in] cbarg - callback argument
6286  *
6287  * Return status.
6288  */
6289 bfa_status_t
6290 bfa_tfru_read(struct bfa_fru_s *fru, void *buf, u32 len, u32 offset,
6291 	      bfa_cb_fru_t cbfn, void *cbarg)
6292 {
6293 	bfa_trc(fru, BFI_TFRU_H2I_READ_REQ);
6294 	bfa_trc(fru, len);
6295 	bfa_trc(fru, offset);
6296 
6297 	if (fru->ioc->asic_gen != BFI_ASIC_GEN_CT2)
6298 		return BFA_STATUS_FRU_NOT_PRESENT;
6299 
6300 	if (!bfa_ioc_is_operational(fru->ioc))
6301 		return BFA_STATUS_IOC_NON_OP;
6302 
6303 	if (fru->op_busy) {
6304 		bfa_trc(fru, fru->op_busy);
6305 		return BFA_STATUS_DEVBUSY;
6306 	}
6307 
6308 	fru->op_busy = 1;
6309 
6310 	fru->cbfn = cbfn;
6311 	fru->cbarg = cbarg;
6312 	fru->residue = len;
6313 	fru->offset = 0;
6314 	fru->addr_off = offset;
6315 	fru->ubuf = buf;
6316 	bfa_fru_read_send(fru, BFI_TFRU_H2I_READ_REQ);
6317 
6318 	return BFA_STATUS_OK;
6319 }
6320 
6321 /*
6322  * Process fru response messages upon receiving interrupts.
6323  *
6324  * @param[in] fruarg - fru structure
6325  * @param[in] msg - message structure
6326  */
6327 void
6328 bfa_fru_intr(void *fruarg, struct bfi_mbmsg_s *msg)
6329 {
6330 	struct bfa_fru_s *fru = fruarg;
6331 	struct bfi_fru_rsp_s *rsp = (struct bfi_fru_rsp_s *)msg;
6332 	u32 status;
6333 
6334 	bfa_trc(fru, msg->mh.msg_id);
6335 
6336 	if (!fru->op_busy) {
6337 		/*
6338 		 * receiving response after ioc failure
6339 		 */
6340 		bfa_trc(fru, 0x9999);
6341 		return;
6342 	}
6343 
6344 	switch (msg->mh.msg_id) {
6345 	case BFI_FRUVPD_I2H_WRITE_RSP:
6346 	case BFI_TFRU_I2H_WRITE_RSP:
6347 		status = be32_to_cpu(rsp->status);
6348 		bfa_trc(fru, status);
6349 
6350 		if (status != BFA_STATUS_OK || fru->residue == 0) {
6351 			fru->status = status;
6352 			fru->op_busy = 0;
6353 			if (fru->cbfn)
6354 				fru->cbfn(fru->cbarg, fru->status);
6355 		} else {
6356 			bfa_trc(fru, fru->offset);
6357 			if (msg->mh.msg_id == BFI_FRUVPD_I2H_WRITE_RSP)
6358 				bfa_fru_write_send(fru,
6359 					BFI_FRUVPD_H2I_WRITE_REQ);
6360 			else
6361 				bfa_fru_write_send(fru,
6362 					BFI_TFRU_H2I_WRITE_REQ);
6363 		}
6364 		break;
6365 	case BFI_FRUVPD_I2H_READ_RSP:
6366 	case BFI_TFRU_I2H_READ_RSP:
6367 		status = be32_to_cpu(rsp->status);
6368 		bfa_trc(fru, status);
6369 
6370 		if (status != BFA_STATUS_OK) {
6371 			fru->status = status;
6372 			fru->op_busy = 0;
6373 			if (fru->cbfn)
6374 				fru->cbfn(fru->cbarg, fru->status);
6375 		} else {
6376 			u32 len = be32_to_cpu(rsp->length);
6377 
6378 			bfa_trc(fru, fru->offset);
6379 			bfa_trc(fru, len);
6380 
6381 			memcpy(fru->ubuf + fru->offset, fru->dbuf_kva, len);
6382 			fru->residue -= len;
6383 			fru->offset += len;
6384 
6385 			if (fru->residue == 0) {
6386 				fru->status = status;
6387 				fru->op_busy = 0;
6388 				if (fru->cbfn)
6389 					fru->cbfn(fru->cbarg, fru->status);
6390 			} else {
6391 				if (msg->mh.msg_id == BFI_FRUVPD_I2H_READ_RSP)
6392 					bfa_fru_read_send(fru,
6393 						BFI_FRUVPD_H2I_READ_REQ);
6394 				else
6395 					bfa_fru_read_send(fru,
6396 						BFI_TFRU_H2I_READ_REQ);
6397 			}
6398 		}
6399 		break;
6400 	default:
6401 		WARN_ON(1);
6402 	}
6403 }
6404