xref: /openbmc/linux/drivers/scsi/aacraid/src.c (revision 5927145e)
1 /*
2  *	Adaptec AAC series RAID controller driver
3  *	(c) Copyright 2001 Red Hat Inc.
4  *
5  * based on the old aacraid driver that is..
6  * Adaptec aacraid device driver for Linux.
7  *
8  * Copyright (c) 2000-2010 Adaptec, Inc.
9  *               2010-2015 PMC-Sierra, Inc. (aacraid@pmc-sierra.com)
10  *		 2016-2017 Microsemi Corp. (aacraid@microsemi.com)
11  *
12  * This program is free software; you can redistribute it and/or modify
13  * it under the terms of the GNU General Public License as published by
14  * the Free Software Foundation; either version 2, or (at your option)
15  * any later version.
16  *
17  * This program is distributed in the hope that it will be useful,
18  * but WITHOUT ANY WARRANTY; without even the implied warranty of
19  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
20  * GNU General Public License for more details.
21  *
22  * You should have received a copy of the GNU General Public License
23  * along with this program; see the file COPYING.  If not, write to
24  * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA.
25  *
26  * Module Name:
27  *  src.c
28  *
29  * Abstract: Hardware Device Interface for PMC SRC based controllers
30  *
31  */
32 
33 #include <linux/kernel.h>
34 #include <linux/init.h>
35 #include <linux/types.h>
36 #include <linux/pci.h>
37 #include <linux/spinlock.h>
38 #include <linux/slab.h>
39 #include <linux/blkdev.h>
40 #include <linux/delay.h>
41 #include <linux/completion.h>
42 #include <linux/time.h>
43 #include <linux/interrupt.h>
44 #include <scsi/scsi_host.h>
45 
46 #include "aacraid.h"
47 
48 static int aac_src_get_sync_status(struct aac_dev *dev);
49 
50 static irqreturn_t aac_src_intr_message(int irq, void *dev_id)
51 {
52 	struct aac_msix_ctx *ctx;
53 	struct aac_dev *dev;
54 	unsigned long bellbits, bellbits_shifted;
55 	int vector_no;
56 	int isFastResponse, mode;
57 	u32 index, handle;
58 
59 	ctx = (struct aac_msix_ctx *)dev_id;
60 	dev = ctx->dev;
61 	vector_no = ctx->vector_no;
62 
63 	if (dev->msi_enabled) {
64 		mode = AAC_INT_MODE_MSI;
65 		if (vector_no == 0) {
66 			bellbits = src_readl(dev, MUnit.ODR_MSI);
67 			if (bellbits & 0x40000)
68 				mode |= AAC_INT_MODE_AIF;
69 			if (bellbits & 0x1000)
70 				mode |= AAC_INT_MODE_SYNC;
71 		}
72 	} else {
73 		mode = AAC_INT_MODE_INTX;
74 		bellbits = src_readl(dev, MUnit.ODR_R);
75 		if (bellbits & PmDoorBellResponseSent) {
76 			bellbits = PmDoorBellResponseSent;
77 			src_writel(dev, MUnit.ODR_C, bellbits);
78 			src_readl(dev, MUnit.ODR_C);
79 		} else {
80 			bellbits_shifted = (bellbits >> SRC_ODR_SHIFT);
81 			src_writel(dev, MUnit.ODR_C, bellbits);
82 			src_readl(dev, MUnit.ODR_C);
83 
84 			if (bellbits_shifted & DoorBellAifPending)
85 				mode |= AAC_INT_MODE_AIF;
86 			else if (bellbits_shifted & OUTBOUNDDOORBELL_0)
87 				mode |= AAC_INT_MODE_SYNC;
88 		}
89 	}
90 
91 	if (mode & AAC_INT_MODE_SYNC) {
92 		unsigned long sflags;
93 		struct list_head *entry;
94 		int send_it = 0;
95 		extern int aac_sync_mode;
96 
97 		if (!aac_sync_mode && !dev->msi_enabled) {
98 			src_writel(dev, MUnit.ODR_C, bellbits);
99 			src_readl(dev, MUnit.ODR_C);
100 		}
101 
102 		if (dev->sync_fib) {
103 			if (dev->sync_fib->callback)
104 				dev->sync_fib->callback(dev->sync_fib->callback_data,
105 					dev->sync_fib);
106 			spin_lock_irqsave(&dev->sync_fib->event_lock, sflags);
107 			if (dev->sync_fib->flags & FIB_CONTEXT_FLAG_WAIT) {
108 				dev->management_fib_count--;
109 				up(&dev->sync_fib->event_wait);
110 			}
111 			spin_unlock_irqrestore(&dev->sync_fib->event_lock,
112 						sflags);
113 			spin_lock_irqsave(&dev->sync_lock, sflags);
114 			if (!list_empty(&dev->sync_fib_list)) {
115 				entry = dev->sync_fib_list.next;
116 				dev->sync_fib = list_entry(entry,
117 							   struct fib,
118 							   fiblink);
119 				list_del(entry);
120 				send_it = 1;
121 			} else {
122 				dev->sync_fib = NULL;
123 			}
124 			spin_unlock_irqrestore(&dev->sync_lock, sflags);
125 			if (send_it) {
126 				aac_adapter_sync_cmd(dev, SEND_SYNCHRONOUS_FIB,
127 					(u32)dev->sync_fib->hw_fib_pa,
128 					0, 0, 0, 0, 0,
129 					NULL, NULL, NULL, NULL, NULL);
130 			}
131 		}
132 		if (!dev->msi_enabled)
133 			mode = 0;
134 
135 	}
136 
137 	if (mode & AAC_INT_MODE_AIF) {
138 		/* handle AIF */
139 		if (dev->sa_firmware) {
140 			u32 events = src_readl(dev, MUnit.SCR0);
141 
142 			aac_intr_normal(dev, events, 1, 0, NULL);
143 			writel(events, &dev->IndexRegs->Mailbox[0]);
144 			src_writel(dev, MUnit.IDR, 1 << 23);
145 		} else {
146 			if (dev->aif_thread && dev->fsa_dev)
147 				aac_intr_normal(dev, 0, 2, 0, NULL);
148 		}
149 		if (dev->msi_enabled)
150 			aac_src_access_devreg(dev, AAC_CLEAR_AIF_BIT);
151 		mode = 0;
152 	}
153 
154 	if (mode) {
155 		index = dev->host_rrq_idx[vector_no];
156 
157 		for (;;) {
158 			isFastResponse = 0;
159 			/* remove toggle bit (31) */
160 			handle = le32_to_cpu((dev->host_rrq[index])
161 				& 0x7fffffff);
162 			/* check fast response bits (30, 1) */
163 			if (handle & 0x40000000)
164 				isFastResponse = 1;
165 			handle &= 0x0000ffff;
166 			if (handle == 0)
167 				break;
168 			handle >>= 2;
169 			if (dev->msi_enabled && dev->max_msix > 1)
170 				atomic_dec(&dev->rrq_outstanding[vector_no]);
171 			aac_intr_normal(dev, handle, 0, isFastResponse, NULL);
172 			dev->host_rrq[index++] = 0;
173 			if (index == (vector_no + 1) * dev->vector_cap)
174 				index = vector_no * dev->vector_cap;
175 			dev->host_rrq_idx[vector_no] = index;
176 		}
177 		mode = 0;
178 	}
179 
180 	return IRQ_HANDLED;
181 }
182 
183 /**
184  *	aac_src_disable_interrupt	-	Disable interrupts
185  *	@dev: Adapter
186  */
187 
188 static void aac_src_disable_interrupt(struct aac_dev *dev)
189 {
190 	src_writel(dev, MUnit.OIMR, dev->OIMR = 0xffffffff);
191 }
192 
193 /**
194  *	aac_src_enable_interrupt_message	-	Enable interrupts
195  *	@dev: Adapter
196  */
197 
198 static void aac_src_enable_interrupt_message(struct aac_dev *dev)
199 {
200 	aac_src_access_devreg(dev, AAC_ENABLE_INTERRUPT);
201 }
202 
203 /**
204  *	src_sync_cmd	-	send a command and wait
205  *	@dev: Adapter
206  *	@command: Command to execute
207  *	@p1: first parameter
208  *	@ret: adapter status
209  *
210  *	This routine will send a synchronous command to the adapter and wait
211  *	for its	completion.
212  */
213 
214 static int src_sync_cmd(struct aac_dev *dev, u32 command,
215 	u32 p1, u32 p2, u32 p3, u32 p4, u32 p5, u32 p6,
216 	u32 *status, u32 * r1, u32 * r2, u32 * r3, u32 * r4)
217 {
218 	unsigned long start;
219 	unsigned long delay;
220 	int ok;
221 
222 	/*
223 	 *	Write the command into Mailbox 0
224 	 */
225 	writel(command, &dev->IndexRegs->Mailbox[0]);
226 	/*
227 	 *	Write the parameters into Mailboxes 1 - 6
228 	 */
229 	writel(p1, &dev->IndexRegs->Mailbox[1]);
230 	writel(p2, &dev->IndexRegs->Mailbox[2]);
231 	writel(p3, &dev->IndexRegs->Mailbox[3]);
232 	writel(p4, &dev->IndexRegs->Mailbox[4]);
233 
234 	/*
235 	 *	Clear the synch command doorbell to start on a clean slate.
236 	 */
237 	if (!dev->msi_enabled)
238 		src_writel(dev,
239 			   MUnit.ODR_C,
240 			   OUTBOUNDDOORBELL_0 << SRC_ODR_SHIFT);
241 
242 	/*
243 	 *	Disable doorbell interrupts
244 	 */
245 	src_writel(dev, MUnit.OIMR, dev->OIMR = 0xffffffff);
246 
247 	/*
248 	 *	Force the completion of the mask register write before issuing
249 	 *	the interrupt.
250 	 */
251 	src_readl(dev, MUnit.OIMR);
252 
253 	/*
254 	 *	Signal that there is a new synch command
255 	 */
256 	src_writel(dev, MUnit.IDR, INBOUNDDOORBELL_0 << SRC_IDR_SHIFT);
257 
258 	if (!dev->sync_mode || command != SEND_SYNCHRONOUS_FIB) {
259 		ok = 0;
260 		start = jiffies;
261 
262 		if (command == IOP_RESET_ALWAYS) {
263 			/* Wait up to 10 sec */
264 			delay = 10*HZ;
265 		} else {
266 			/* Wait up to 5 minutes */
267 			delay = 300*HZ;
268 		}
269 		while (time_before(jiffies, start+delay)) {
270 			udelay(5);	/* Delay 5 microseconds to let Mon960 get info. */
271 			/*
272 			 *	Mon960 will set doorbell0 bit when it has completed the command.
273 			 */
274 			if (aac_src_get_sync_status(dev) & OUTBOUNDDOORBELL_0) {
275 				/*
276 				 *	Clear the doorbell.
277 				 */
278 				if (dev->msi_enabled)
279 					aac_src_access_devreg(dev,
280 						AAC_CLEAR_SYNC_BIT);
281 				else
282 					src_writel(dev,
283 						MUnit.ODR_C,
284 						OUTBOUNDDOORBELL_0 << SRC_ODR_SHIFT);
285 				ok = 1;
286 				break;
287 			}
288 			/*
289 			 *	Yield the processor in case we are slow
290 			 */
291 			msleep(1);
292 		}
293 		if (unlikely(ok != 1)) {
294 			/*
295 			 *	Restore interrupt mask even though we timed out
296 			 */
297 			aac_adapter_enable_int(dev);
298 			return -ETIMEDOUT;
299 		}
300 		/*
301 		 *	Pull the synch status from Mailbox 0.
302 		 */
303 		if (status)
304 			*status = readl(&dev->IndexRegs->Mailbox[0]);
305 		if (r1)
306 			*r1 = readl(&dev->IndexRegs->Mailbox[1]);
307 		if (r2)
308 			*r2 = readl(&dev->IndexRegs->Mailbox[2]);
309 		if (r3)
310 			*r3 = readl(&dev->IndexRegs->Mailbox[3]);
311 		if (r4)
312 			*r4 = readl(&dev->IndexRegs->Mailbox[4]);
313 		if (command == GET_COMM_PREFERRED_SETTINGS)
314 			dev->max_msix =
315 				readl(&dev->IndexRegs->Mailbox[5]) & 0xFFFF;
316 		/*
317 		 *	Clear the synch command doorbell.
318 		 */
319 		if (!dev->msi_enabled)
320 			src_writel(dev,
321 				MUnit.ODR_C,
322 				OUTBOUNDDOORBELL_0 << SRC_ODR_SHIFT);
323 	}
324 
325 	/*
326 	 *	Restore interrupt mask
327 	 */
328 	aac_adapter_enable_int(dev);
329 	return 0;
330 }
331 
332 /**
333  *	aac_src_interrupt_adapter	-	interrupt adapter
334  *	@dev: Adapter
335  *
336  *	Send an interrupt to the i960 and breakpoint it.
337  */
338 
339 static void aac_src_interrupt_adapter(struct aac_dev *dev)
340 {
341 	src_sync_cmd(dev, BREAKPOINT_REQUEST,
342 		0, 0, 0, 0, 0, 0,
343 		NULL, NULL, NULL, NULL, NULL);
344 }
345 
346 /**
347  *	aac_src_notify_adapter		-	send an event to the adapter
348  *	@dev: Adapter
349  *	@event: Event to send
350  *
351  *	Notify the i960 that something it probably cares about has
352  *	happened.
353  */
354 
355 static void aac_src_notify_adapter(struct aac_dev *dev, u32 event)
356 {
357 	switch (event) {
358 
359 	case AdapNormCmdQue:
360 		src_writel(dev, MUnit.ODR_C,
361 			INBOUNDDOORBELL_1 << SRC_ODR_SHIFT);
362 		break;
363 	case HostNormRespNotFull:
364 		src_writel(dev, MUnit.ODR_C,
365 			INBOUNDDOORBELL_4 << SRC_ODR_SHIFT);
366 		break;
367 	case AdapNormRespQue:
368 		src_writel(dev, MUnit.ODR_C,
369 			INBOUNDDOORBELL_2 << SRC_ODR_SHIFT);
370 		break;
371 	case HostNormCmdNotFull:
372 		src_writel(dev, MUnit.ODR_C,
373 			INBOUNDDOORBELL_3 << SRC_ODR_SHIFT);
374 		break;
375 	case FastIo:
376 		src_writel(dev, MUnit.ODR_C,
377 			INBOUNDDOORBELL_6 << SRC_ODR_SHIFT);
378 		break;
379 	case AdapPrintfDone:
380 		src_writel(dev, MUnit.ODR_C,
381 			INBOUNDDOORBELL_5 << SRC_ODR_SHIFT);
382 		break;
383 	default:
384 		BUG();
385 		break;
386 	}
387 }
388 
389 /**
390  *	aac_src_start_adapter		-	activate adapter
391  *	@dev:	Adapter
392  *
393  *	Start up processing on an i960 based AAC adapter
394  */
395 
396 static void aac_src_start_adapter(struct aac_dev *dev)
397 {
398 	union aac_init *init;
399 	int i;
400 
401 	 /* reset host_rrq_idx first */
402 	for (i = 0; i < dev->max_msix; i++) {
403 		dev->host_rrq_idx[i] = i * dev->vector_cap;
404 		atomic_set(&dev->rrq_outstanding[i], 0);
405 	}
406 	atomic_set(&dev->msix_counter, 0);
407 	dev->fibs_pushed_no = 0;
408 
409 	init = dev->init;
410 	if (dev->comm_interface == AAC_COMM_MESSAGE_TYPE3) {
411 		init->r8.host_elapsed_seconds = cpu_to_le32(get_seconds());
412 		src_sync_cmd(dev, INIT_STRUCT_BASE_ADDRESS,
413 			lower_32_bits(dev->init_pa),
414 			upper_32_bits(dev->init_pa),
415 			sizeof(struct _r8) +
416 			(AAC_MAX_HRRQ - 1) * sizeof(struct _rrq),
417 			0, 0, 0, NULL, NULL, NULL, NULL, NULL);
418 	} else {
419 		init->r7.host_elapsed_seconds = cpu_to_le32(get_seconds());
420 		// We can only use a 32 bit address here
421 		src_sync_cmd(dev, INIT_STRUCT_BASE_ADDRESS,
422 			(u32)(ulong)dev->init_pa, 0, 0, 0, 0, 0,
423 			NULL, NULL, NULL, NULL, NULL);
424 	}
425 
426 }
427 
428 /**
429  *	aac_src_check_health
430  *	@dev: device to check if healthy
431  *
432  *	Will attempt to determine if the specified adapter is alive and
433  *	capable of handling requests, returning 0 if alive.
434  */
435 static int aac_src_check_health(struct aac_dev *dev)
436 {
437 	u32 status = src_readl(dev, MUnit.OMR);
438 
439 	/*
440 	 *	Check to see if the board panic'd.
441 	 */
442 	if (unlikely(status & KERNEL_PANIC))
443 		goto err_blink;
444 
445 	/*
446 	 *	Check to see if the board failed any self tests.
447 	 */
448 	if (unlikely(status & SELF_TEST_FAILED))
449 		goto err_out;
450 
451 	/*
452 	 *	Check to see if the board failed any self tests.
453 	 */
454 	if (unlikely(status & MONITOR_PANIC))
455 		goto err_out;
456 
457 	/*
458 	 *	Wait for the adapter to be up and running.
459 	 */
460 	if (unlikely(!(status & KERNEL_UP_AND_RUNNING)))
461 		return -3;
462 	/*
463 	 *	Everything is OK
464 	 */
465 	return 0;
466 
467 err_out:
468 	return -1;
469 
470 err_blink:
471 	return (status >> 16) & 0xFF;
472 }
473 
474 static inline u32 aac_get_vector(struct aac_dev *dev)
475 {
476 	return atomic_inc_return(&dev->msix_counter)%dev->max_msix;
477 }
478 
479 /**
480  *	aac_src_deliver_message
481  *	@fib: fib to issue
482  *
483  *	Will send a fib, returning 0 if successful.
484  */
485 static int aac_src_deliver_message(struct fib *fib)
486 {
487 	struct aac_dev *dev = fib->dev;
488 	struct aac_queue *q = &dev->queues->queue[AdapNormCmdQueue];
489 	u32 fibsize;
490 	dma_addr_t address;
491 	struct aac_fib_xporthdr *pFibX;
492 	int native_hba;
493 #if !defined(writeq)
494 	unsigned long flags;
495 #endif
496 
497 	u16 vector_no;
498 
499 	atomic_inc(&q->numpending);
500 
501 	native_hba = (fib->flags & FIB_CONTEXT_FLAG_NATIVE_HBA) ? 1 : 0;
502 
503 
504 	if (dev->msi_enabled && dev->max_msix > 1 &&
505 		(native_hba || fib->hw_fib_va->header.Command != AifRequest)) {
506 
507 		if ((dev->comm_interface == AAC_COMM_MESSAGE_TYPE3)
508 			&& dev->sa_firmware)
509 			vector_no = aac_get_vector(dev);
510 		else
511 			vector_no = fib->vector_no;
512 
513 		if (native_hba) {
514 			if (fib->flags & FIB_CONTEXT_FLAG_NATIVE_HBA_TMF) {
515 				struct aac_hba_tm_req *tm_req;
516 
517 				tm_req = (struct aac_hba_tm_req *)
518 						fib->hw_fib_va;
519 				if (tm_req->iu_type ==
520 					HBA_IU_TYPE_SCSI_TM_REQ) {
521 					((struct aac_hba_tm_req *)
522 						fib->hw_fib_va)->reply_qid
523 							= vector_no;
524 					((struct aac_hba_tm_req *)
525 						fib->hw_fib_va)->request_id
526 							+= (vector_no << 16);
527 				} else {
528 					((struct aac_hba_reset_req *)
529 						fib->hw_fib_va)->reply_qid
530 							= vector_no;
531 					((struct aac_hba_reset_req *)
532 						fib->hw_fib_va)->request_id
533 							+= (vector_no << 16);
534 				}
535 			} else {
536 				((struct aac_hba_cmd_req *)
537 					fib->hw_fib_va)->reply_qid
538 						= vector_no;
539 				((struct aac_hba_cmd_req *)
540 					fib->hw_fib_va)->request_id
541 						+= (vector_no << 16);
542 			}
543 		} else {
544 			fib->hw_fib_va->header.Handle += (vector_no << 16);
545 		}
546 	} else {
547 		vector_no = 0;
548 	}
549 
550 	atomic_inc(&dev->rrq_outstanding[vector_no]);
551 
552 	if (native_hba) {
553 		address = fib->hw_fib_pa;
554 		fibsize = (fib->hbacmd_size + 127) / 128 - 1;
555 		if (fibsize > 31)
556 			fibsize = 31;
557 		address |= fibsize;
558 #if defined(writeq)
559 		src_writeq(dev, MUnit.IQN_L, (u64)address);
560 #else
561 		spin_lock_irqsave(&fib->dev->iq_lock, flags);
562 		src_writel(dev, MUnit.IQN_H,
563 			upper_32_bits(address) & 0xffffffff);
564 		src_writel(dev, MUnit.IQN_L, address & 0xffffffff);
565 		spin_unlock_irqrestore(&fib->dev->iq_lock, flags);
566 #endif
567 	} else {
568 		if (dev->comm_interface == AAC_COMM_MESSAGE_TYPE2 ||
569 			dev->comm_interface == AAC_COMM_MESSAGE_TYPE3) {
570 			/* Calculate the amount to the fibsize bits */
571 			fibsize = (le16_to_cpu(fib->hw_fib_va->header.Size)
572 				+ 127) / 128 - 1;
573 			/* New FIB header, 32-bit */
574 			address = fib->hw_fib_pa;
575 			fib->hw_fib_va->header.StructType = FIB_MAGIC2;
576 			fib->hw_fib_va->header.SenderFibAddress =
577 				cpu_to_le32((u32)address);
578 			fib->hw_fib_va->header.u.TimeStamp = 0;
579 			WARN_ON(upper_32_bits(address) != 0L);
580 		} else {
581 			/* Calculate the amount to the fibsize bits */
582 			fibsize = (sizeof(struct aac_fib_xporthdr) +
583 				le16_to_cpu(fib->hw_fib_va->header.Size)
584 				+ 127) / 128 - 1;
585 			/* Fill XPORT header */
586 			pFibX = (struct aac_fib_xporthdr *)
587 				((unsigned char *)fib->hw_fib_va -
588 				sizeof(struct aac_fib_xporthdr));
589 			pFibX->Handle = fib->hw_fib_va->header.Handle;
590 			pFibX->HostAddress =
591 				cpu_to_le64((u64)fib->hw_fib_pa);
592 			pFibX->Size = cpu_to_le32(
593 				le16_to_cpu(fib->hw_fib_va->header.Size));
594 			address = fib->hw_fib_pa -
595 				(u64)sizeof(struct aac_fib_xporthdr);
596 		}
597 		if (fibsize > 31)
598 			fibsize = 31;
599 		address |= fibsize;
600 
601 #if defined(writeq)
602 		src_writeq(dev, MUnit.IQ_L, (u64)address);
603 #else
604 		spin_lock_irqsave(&fib->dev->iq_lock, flags);
605 		src_writel(dev, MUnit.IQ_H,
606 			upper_32_bits(address) & 0xffffffff);
607 		src_writel(dev, MUnit.IQ_L, address & 0xffffffff);
608 		spin_unlock_irqrestore(&fib->dev->iq_lock, flags);
609 #endif
610 	}
611 	return 0;
612 }
613 
614 /**
615  *	aac_src_ioremap
616  *	@size: mapping resize request
617  *
618  */
619 static int aac_src_ioremap(struct aac_dev *dev, u32 size)
620 {
621 	if (!size) {
622 		iounmap(dev->regs.src.bar1);
623 		dev->regs.src.bar1 = NULL;
624 		iounmap(dev->regs.src.bar0);
625 		dev->base = dev->regs.src.bar0 = NULL;
626 		return 0;
627 	}
628 	dev->regs.src.bar1 = ioremap(pci_resource_start(dev->pdev, 2),
629 		AAC_MIN_SRC_BAR1_SIZE);
630 	dev->base = NULL;
631 	if (dev->regs.src.bar1 == NULL)
632 		return -1;
633 	dev->base = dev->regs.src.bar0 = ioremap(dev->base_start, size);
634 	if (dev->base == NULL) {
635 		iounmap(dev->regs.src.bar1);
636 		dev->regs.src.bar1 = NULL;
637 		return -1;
638 	}
639 	dev->IndexRegs = &((struct src_registers __iomem *)
640 		dev->base)->u.tupelo.IndexRegs;
641 	return 0;
642 }
643 
644 /**
645  *  aac_srcv_ioremap
646  *	@size: mapping resize request
647  *
648  */
649 static int aac_srcv_ioremap(struct aac_dev *dev, u32 size)
650 {
651 	if (!size) {
652 		iounmap(dev->regs.src.bar0);
653 		dev->base = dev->regs.src.bar0 = NULL;
654 		return 0;
655 	}
656 
657 	dev->regs.src.bar1 =
658 	ioremap(pci_resource_start(dev->pdev, 2), AAC_MIN_SRCV_BAR1_SIZE);
659 	dev->base = NULL;
660 	if (dev->regs.src.bar1 == NULL)
661 		return -1;
662 	dev->base = dev->regs.src.bar0 = ioremap(dev->base_start, size);
663 	if (dev->base == NULL) {
664 		iounmap(dev->regs.src.bar1);
665 		dev->regs.src.bar1 = NULL;
666 		return -1;
667 	}
668 	dev->IndexRegs = &((struct src_registers __iomem *)
669 		dev->base)->u.denali.IndexRegs;
670 	return 0;
671 }
672 
673 void aac_set_intx_mode(struct aac_dev *dev)
674 {
675 	if (dev->msi_enabled) {
676 		aac_src_access_devreg(dev, AAC_ENABLE_INTX);
677 		dev->msi_enabled = 0;
678 		msleep(5000); /* Delay 5 seconds */
679 	}
680 }
681 
682 static void aac_dump_fw_fib_iop_reset(struct aac_dev *dev)
683 {
684 	__le32 supported_options3;
685 
686 	if (!aac_fib_dump)
687 		return;
688 
689 	supported_options3  = dev->supplement_adapter_info.supported_options3;
690 	if (!(supported_options3 & AAC_OPTION_SUPPORTED3_IOP_RESET_FIB_DUMP))
691 		return;
692 
693 	aac_adapter_sync_cmd(dev, IOP_RESET_FW_FIB_DUMP,
694 			0, 0, 0,  0, 0, 0, NULL, NULL, NULL, NULL, NULL);
695 }
696 
697 static bool aac_is_ctrl_up_and_running(struct aac_dev *dev)
698 {
699 	bool ctrl_up = true;
700 	unsigned long status, start;
701 	bool is_up = false;
702 
703 	start = jiffies;
704 	do {
705 		schedule();
706 		status = src_readl(dev, MUnit.OMR);
707 
708 		if (status == 0xffffffff)
709 			status = 0;
710 
711 		if (status & KERNEL_BOOTING) {
712 			start = jiffies;
713 			continue;
714 		}
715 
716 		if (time_after(jiffies, start+HZ*SOFT_RESET_TIME)) {
717 			ctrl_up = false;
718 			break;
719 		}
720 
721 		is_up = status & KERNEL_UP_AND_RUNNING;
722 
723 	} while (!is_up);
724 
725 	return ctrl_up;
726 }
727 
728 static void aac_notify_fw_of_iop_reset(struct aac_dev *dev)
729 {
730 	aac_adapter_sync_cmd(dev, IOP_RESET_ALWAYS, 0, 0, 0, 0, 0, 0, NULL,
731 						NULL, NULL, NULL, NULL);
732 }
733 
734 static void aac_send_iop_reset(struct aac_dev *dev)
735 {
736 	aac_dump_fw_fib_iop_reset(dev);
737 
738 	aac_notify_fw_of_iop_reset(dev);
739 
740 	aac_set_intx_mode(dev);
741 
742 	src_writel(dev, MUnit.IDR, IOP_SRC_RESET_MASK);
743 
744 	msleep(5000);
745 }
746 
747 static void aac_send_hardware_soft_reset(struct aac_dev *dev)
748 {
749 	u_int32_t val;
750 
751 	val = readl(((char *)(dev->base) + IBW_SWR_OFFSET));
752 	val |= 0x01;
753 	writel(val, ((char *)(dev->base) + IBW_SWR_OFFSET));
754 	msleep_interruptible(20000);
755 }
756 
757 static int aac_src_restart_adapter(struct aac_dev *dev, int bled, u8 reset_type)
758 {
759 	bool is_ctrl_up;
760 	int ret = 0;
761 
762 	if (bled < 0)
763 		goto invalid_out;
764 
765 	if (bled)
766 		dev_err(&dev->pdev->dev, "adapter kernel panic'd %x.\n", bled);
767 
768 	/*
769 	 * When there is a BlinkLED, IOP_RESET has not effect
770 	 */
771 	if (bled >= 2 && dev->sa_firmware && reset_type & HW_IOP_RESET)
772 		reset_type &= ~HW_IOP_RESET;
773 
774 	dev->a_ops.adapter_enable_int = aac_src_disable_interrupt;
775 
776 	dev_err(&dev->pdev->dev, "Controller reset type is %d\n", reset_type);
777 
778 	if (reset_type & HW_IOP_RESET) {
779 		dev_info(&dev->pdev->dev, "Issuing IOP reset\n");
780 		aac_send_iop_reset(dev);
781 
782 		/*
783 		 * Creates a delay or wait till up and running comes thru
784 		 */
785 		is_ctrl_up = aac_is_ctrl_up_and_running(dev);
786 		if (!is_ctrl_up)
787 			dev_err(&dev->pdev->dev, "IOP reset failed\n");
788 		else {
789 			dev_info(&dev->pdev->dev, "IOP reset succeded\n");
790 			goto set_startup;
791 		}
792 	}
793 
794 	if (!dev->sa_firmware) {
795 		dev_err(&dev->pdev->dev, "ARC Reset attempt failed\n");
796 		ret = -ENODEV;
797 		goto out;
798 	}
799 
800 	if (reset_type & HW_SOFT_RESET) {
801 		dev_info(&dev->pdev->dev, "Issuing SOFT reset\n");
802 		aac_send_hardware_soft_reset(dev);
803 		dev->msi_enabled = 0;
804 
805 		is_ctrl_up = aac_is_ctrl_up_and_running(dev);
806 		if (!is_ctrl_up) {
807 			dev_err(&dev->pdev->dev, "SOFT reset failed\n");
808 			ret = -ENODEV;
809 			goto out;
810 		} else
811 			dev_info(&dev->pdev->dev, "SOFT reset succeded\n");
812 	}
813 
814 set_startup:
815 	if (startup_timeout < 300)
816 		startup_timeout = 300;
817 
818 out:
819 	return ret;
820 
821 invalid_out:
822 	if (src_readl(dev, MUnit.OMR) & KERNEL_PANIC)
823 		ret = -ENODEV;
824 goto out;
825 }
826 
827 /**
828  *	aac_src_select_comm	-	Select communications method
829  *	@dev: Adapter
830  *	@comm: communications method
831  */
832 static int aac_src_select_comm(struct aac_dev *dev, int comm)
833 {
834 	switch (comm) {
835 	case AAC_COMM_MESSAGE:
836 		dev->a_ops.adapter_intr = aac_src_intr_message;
837 		dev->a_ops.adapter_deliver = aac_src_deliver_message;
838 		break;
839 	default:
840 		return 1;
841 	}
842 	return 0;
843 }
844 
845 /**
846  *  aac_src_init	-	initialize an Cardinal Frey Bar card
847  *  @dev: device to configure
848  *
849  */
850 
851 int aac_src_init(struct aac_dev *dev)
852 {
853 	unsigned long start;
854 	unsigned long status;
855 	int restart = 0;
856 	int instance = dev->id;
857 	const char *name = dev->name;
858 
859 	dev->a_ops.adapter_ioremap = aac_src_ioremap;
860 	dev->a_ops.adapter_comm = aac_src_select_comm;
861 
862 	dev->base_size = AAC_MIN_SRC_BAR0_SIZE;
863 	if (aac_adapter_ioremap(dev, dev->base_size)) {
864 		printk(KERN_WARNING "%s: unable to map adapter.\n", name);
865 		goto error_iounmap;
866 	}
867 
868 	/* Failure to reset here is an option ... */
869 	dev->a_ops.adapter_sync_cmd = src_sync_cmd;
870 	dev->a_ops.adapter_enable_int = aac_src_disable_interrupt;
871 
872 	if (dev->init_reset) {
873 		dev->init_reset = false;
874 		if (!aac_src_restart_adapter(dev, 0, IOP_HWSOFT_RESET))
875 			++restart;
876 	}
877 
878 	/*
879 	 *	Check to see if the board panic'd while booting.
880 	 */
881 	status = src_readl(dev, MUnit.OMR);
882 	if (status & KERNEL_PANIC) {
883 		if (aac_src_restart_adapter(dev,
884 			aac_src_check_health(dev), IOP_HWSOFT_RESET))
885 			goto error_iounmap;
886 		++restart;
887 	}
888 	/*
889 	 *	Check to see if the board failed any self tests.
890 	 */
891 	status = src_readl(dev, MUnit.OMR);
892 	if (status & SELF_TEST_FAILED) {
893 		printk(KERN_ERR "%s%d: adapter self-test failed.\n",
894 			dev->name, instance);
895 		goto error_iounmap;
896 	}
897 	/*
898 	 *	Check to see if the monitor panic'd while booting.
899 	 */
900 	if (status & MONITOR_PANIC) {
901 		printk(KERN_ERR "%s%d: adapter monitor panic.\n",
902 			dev->name, instance);
903 		goto error_iounmap;
904 	}
905 	start = jiffies;
906 	/*
907 	 *	Wait for the adapter to be up and running. Wait up to 3 minutes
908 	 */
909 	while (!((status = src_readl(dev, MUnit.OMR)) &
910 		KERNEL_UP_AND_RUNNING)) {
911 		if ((restart &&
912 		  (status & (KERNEL_PANIC|SELF_TEST_FAILED|MONITOR_PANIC))) ||
913 		  time_after(jiffies, start+HZ*startup_timeout)) {
914 			printk(KERN_ERR "%s%d: adapter kernel failed to start, init status = %lx.\n",
915 					dev->name, instance, status);
916 			goto error_iounmap;
917 		}
918 		if (!restart &&
919 		  ((status & (KERNEL_PANIC|SELF_TEST_FAILED|MONITOR_PANIC)) ||
920 		  time_after(jiffies, start + HZ *
921 		  ((startup_timeout > 60)
922 		    ? (startup_timeout - 60)
923 		    : (startup_timeout / 2))))) {
924 			if (likely(!aac_src_restart_adapter(dev,
925 				aac_src_check_health(dev), IOP_HWSOFT_RESET)))
926 				start = jiffies;
927 			++restart;
928 		}
929 		msleep(1);
930 	}
931 	if (restart && aac_commit)
932 		aac_commit = 1;
933 	/*
934 	 *	Fill in the common function dispatch table.
935 	 */
936 	dev->a_ops.adapter_interrupt = aac_src_interrupt_adapter;
937 	dev->a_ops.adapter_disable_int = aac_src_disable_interrupt;
938 	dev->a_ops.adapter_enable_int = aac_src_disable_interrupt;
939 	dev->a_ops.adapter_notify = aac_src_notify_adapter;
940 	dev->a_ops.adapter_sync_cmd = src_sync_cmd;
941 	dev->a_ops.adapter_check_health = aac_src_check_health;
942 	dev->a_ops.adapter_restart = aac_src_restart_adapter;
943 	dev->a_ops.adapter_start = aac_src_start_adapter;
944 
945 	/*
946 	 *	First clear out all interrupts.  Then enable the one's that we
947 	 *	can handle.
948 	 */
949 	aac_adapter_comm(dev, AAC_COMM_MESSAGE);
950 	aac_adapter_disable_int(dev);
951 	src_writel(dev, MUnit.ODR_C, 0xffffffff);
952 	aac_adapter_enable_int(dev);
953 
954 	if (aac_init_adapter(dev) == NULL)
955 		goto error_iounmap;
956 	if (dev->comm_interface != AAC_COMM_MESSAGE_TYPE1)
957 		goto error_iounmap;
958 
959 	dev->msi = !pci_enable_msi(dev->pdev);
960 
961 	dev->aac_msix[0].vector_no = 0;
962 	dev->aac_msix[0].dev = dev;
963 
964 	if (request_irq(dev->pdev->irq, dev->a_ops.adapter_intr,
965 			IRQF_SHARED, "aacraid", &(dev->aac_msix[0]))  < 0) {
966 
967 		if (dev->msi)
968 			pci_disable_msi(dev->pdev);
969 
970 		printk(KERN_ERR "%s%d: Interrupt unavailable.\n",
971 			name, instance);
972 		goto error_iounmap;
973 	}
974 	dev->dbg_base = pci_resource_start(dev->pdev, 2);
975 	dev->dbg_base_mapped = dev->regs.src.bar1;
976 	dev->dbg_size = AAC_MIN_SRC_BAR1_SIZE;
977 	dev->a_ops.adapter_enable_int = aac_src_enable_interrupt_message;
978 
979 	aac_adapter_enable_int(dev);
980 
981 	if (!dev->sync_mode) {
982 		/*
983 		 * Tell the adapter that all is configured, and it can
984 		 * start accepting requests
985 		 */
986 		aac_src_start_adapter(dev);
987 	}
988 	return 0;
989 
990 error_iounmap:
991 
992 	return -1;
993 }
994 
995 /**
996  *  aac_srcv_init	-	initialize an SRCv card
997  *  @dev: device to configure
998  *
999  */
1000 
1001 int aac_srcv_init(struct aac_dev *dev)
1002 {
1003 	unsigned long start;
1004 	unsigned long status;
1005 	int restart = 0;
1006 	int instance = dev->id;
1007 	const char *name = dev->name;
1008 
1009 	dev->a_ops.adapter_ioremap = aac_srcv_ioremap;
1010 	dev->a_ops.adapter_comm = aac_src_select_comm;
1011 
1012 	dev->base_size = AAC_MIN_SRCV_BAR0_SIZE;
1013 	if (aac_adapter_ioremap(dev, dev->base_size)) {
1014 		printk(KERN_WARNING "%s: unable to map adapter.\n", name);
1015 		goto error_iounmap;
1016 	}
1017 
1018 	/* Failure to reset here is an option ... */
1019 	dev->a_ops.adapter_sync_cmd = src_sync_cmd;
1020 	dev->a_ops.adapter_enable_int = aac_src_disable_interrupt;
1021 
1022 	if (dev->init_reset) {
1023 		dev->init_reset = false;
1024 		if (!aac_src_restart_adapter(dev, 0, IOP_HWSOFT_RESET))
1025 			++restart;
1026 	}
1027 
1028 	/*
1029 	 *	Check to see if flash update is running.
1030 	 *	Wait for the adapter to be up and running. Wait up to 5 minutes
1031 	 */
1032 	status = src_readl(dev, MUnit.OMR);
1033 	if (status & FLASH_UPD_PENDING) {
1034 		start = jiffies;
1035 		do {
1036 			status = src_readl(dev, MUnit.OMR);
1037 			if (time_after(jiffies, start+HZ*FWUPD_TIMEOUT)) {
1038 				printk(KERN_ERR "%s%d: adapter flash update failed.\n",
1039 					dev->name, instance);
1040 				goto error_iounmap;
1041 			}
1042 		} while (!(status & FLASH_UPD_SUCCESS) &&
1043 			 !(status & FLASH_UPD_FAILED));
1044 		/* Delay 10 seconds.
1045 		 * Because right now FW is doing a soft reset,
1046 		 * do not read scratch pad register at this time
1047 		 */
1048 		ssleep(10);
1049 	}
1050 	/*
1051 	 *	Check to see if the board panic'd while booting.
1052 	 */
1053 	status = src_readl(dev, MUnit.OMR);
1054 	if (status & KERNEL_PANIC) {
1055 		if (aac_src_restart_adapter(dev,
1056 			aac_src_check_health(dev), IOP_HWSOFT_RESET))
1057 			goto error_iounmap;
1058 		++restart;
1059 	}
1060 	/*
1061 	 *	Check to see if the board failed any self tests.
1062 	 */
1063 	status = src_readl(dev, MUnit.OMR);
1064 	if (status & SELF_TEST_FAILED) {
1065 		printk(KERN_ERR "%s%d: adapter self-test failed.\n", dev->name, instance);
1066 		goto error_iounmap;
1067 	}
1068 	/*
1069 	 *	Check to see if the monitor panic'd while booting.
1070 	 */
1071 	if (status & MONITOR_PANIC) {
1072 		printk(KERN_ERR "%s%d: adapter monitor panic.\n", dev->name, instance);
1073 		goto error_iounmap;
1074 	}
1075 	start = jiffies;
1076 	/*
1077 	 *	Wait for the adapter to be up and running. Wait up to 3 minutes
1078 	 */
1079 	while (!((status = src_readl(dev, MUnit.OMR)) &
1080 		KERNEL_UP_AND_RUNNING) ||
1081 		status == 0xffffffff) {
1082 		if ((restart &&
1083 		  (status & (KERNEL_PANIC|SELF_TEST_FAILED|MONITOR_PANIC))) ||
1084 		  time_after(jiffies, start+HZ*startup_timeout)) {
1085 			printk(KERN_ERR "%s%d: adapter kernel failed to start, init status = %lx.\n",
1086 					dev->name, instance, status);
1087 			goto error_iounmap;
1088 		}
1089 		if (!restart &&
1090 		  ((status & (KERNEL_PANIC|SELF_TEST_FAILED|MONITOR_PANIC)) ||
1091 		  time_after(jiffies, start + HZ *
1092 		  ((startup_timeout > 60)
1093 		    ? (startup_timeout - 60)
1094 		    : (startup_timeout / 2))))) {
1095 			if (likely(!aac_src_restart_adapter(dev,
1096 				aac_src_check_health(dev), IOP_HWSOFT_RESET)))
1097 				start = jiffies;
1098 			++restart;
1099 		}
1100 		msleep(1);
1101 	}
1102 	if (restart && aac_commit)
1103 		aac_commit = 1;
1104 	/*
1105 	 *	Fill in the common function dispatch table.
1106 	 */
1107 	dev->a_ops.adapter_interrupt = aac_src_interrupt_adapter;
1108 	dev->a_ops.adapter_disable_int = aac_src_disable_interrupt;
1109 	dev->a_ops.adapter_enable_int = aac_src_disable_interrupt;
1110 	dev->a_ops.adapter_notify = aac_src_notify_adapter;
1111 	dev->a_ops.adapter_sync_cmd = src_sync_cmd;
1112 	dev->a_ops.adapter_check_health = aac_src_check_health;
1113 	dev->a_ops.adapter_restart = aac_src_restart_adapter;
1114 	dev->a_ops.adapter_start = aac_src_start_adapter;
1115 
1116 	/*
1117 	 *	First clear out all interrupts.  Then enable the one's that we
1118 	 *	can handle.
1119 	 */
1120 	aac_adapter_comm(dev, AAC_COMM_MESSAGE);
1121 	aac_adapter_disable_int(dev);
1122 	src_writel(dev, MUnit.ODR_C, 0xffffffff);
1123 	aac_adapter_enable_int(dev);
1124 
1125 	if (aac_init_adapter(dev) == NULL)
1126 		goto error_iounmap;
1127 	if ((dev->comm_interface != AAC_COMM_MESSAGE_TYPE2) &&
1128 		(dev->comm_interface != AAC_COMM_MESSAGE_TYPE3))
1129 		goto error_iounmap;
1130 	if (dev->msi_enabled)
1131 		aac_src_access_devreg(dev, AAC_ENABLE_MSIX);
1132 
1133 	if (aac_acquire_irq(dev))
1134 		goto error_iounmap;
1135 
1136 	dev->dbg_base = pci_resource_start(dev->pdev, 2);
1137 	dev->dbg_base_mapped = dev->regs.src.bar1;
1138 	dev->dbg_size = AAC_MIN_SRCV_BAR1_SIZE;
1139 	dev->a_ops.adapter_enable_int = aac_src_enable_interrupt_message;
1140 
1141 	aac_adapter_enable_int(dev);
1142 
1143 	if (!dev->sync_mode) {
1144 		/*
1145 		 * Tell the adapter that all is configured, and it can
1146 		 * start accepting requests
1147 		 */
1148 		aac_src_start_adapter(dev);
1149 	}
1150 	return 0;
1151 
1152 error_iounmap:
1153 
1154 	return -1;
1155 }
1156 
1157 void aac_src_access_devreg(struct aac_dev *dev, int mode)
1158 {
1159 	u_int32_t val;
1160 
1161 	switch (mode) {
1162 	case AAC_ENABLE_INTERRUPT:
1163 		src_writel(dev,
1164 			   MUnit.OIMR,
1165 			   dev->OIMR = (dev->msi_enabled ?
1166 					AAC_INT_ENABLE_TYPE1_MSIX :
1167 					AAC_INT_ENABLE_TYPE1_INTX));
1168 		break;
1169 
1170 	case AAC_DISABLE_INTERRUPT:
1171 		src_writel(dev,
1172 			   MUnit.OIMR,
1173 			   dev->OIMR = AAC_INT_DISABLE_ALL);
1174 		break;
1175 
1176 	case AAC_ENABLE_MSIX:
1177 		/* set bit 6 */
1178 		val = src_readl(dev, MUnit.IDR);
1179 		val |= 0x40;
1180 		src_writel(dev,  MUnit.IDR, val);
1181 		src_readl(dev, MUnit.IDR);
1182 		/* unmask int. */
1183 		val = PMC_ALL_INTERRUPT_BITS;
1184 		src_writel(dev, MUnit.IOAR, val);
1185 		val = src_readl(dev, MUnit.OIMR);
1186 		src_writel(dev,
1187 			   MUnit.OIMR,
1188 			   val & (~(PMC_GLOBAL_INT_BIT2 | PMC_GLOBAL_INT_BIT0)));
1189 		break;
1190 
1191 	case AAC_DISABLE_MSIX:
1192 		/* reset bit 6 */
1193 		val = src_readl(dev, MUnit.IDR);
1194 		val &= ~0x40;
1195 		src_writel(dev, MUnit.IDR, val);
1196 		src_readl(dev, MUnit.IDR);
1197 		break;
1198 
1199 	case AAC_CLEAR_AIF_BIT:
1200 		/* set bit 5 */
1201 		val = src_readl(dev, MUnit.IDR);
1202 		val |= 0x20;
1203 		src_writel(dev, MUnit.IDR, val);
1204 		src_readl(dev, MUnit.IDR);
1205 		break;
1206 
1207 	case AAC_CLEAR_SYNC_BIT:
1208 		/* set bit 4 */
1209 		val = src_readl(dev, MUnit.IDR);
1210 		val |= 0x10;
1211 		src_writel(dev, MUnit.IDR, val);
1212 		src_readl(dev, MUnit.IDR);
1213 		break;
1214 
1215 	case AAC_ENABLE_INTX:
1216 		/* set bit 7 */
1217 		val = src_readl(dev, MUnit.IDR);
1218 		val |= 0x80;
1219 		src_writel(dev, MUnit.IDR, val);
1220 		src_readl(dev, MUnit.IDR);
1221 		/* unmask int. */
1222 		val = PMC_ALL_INTERRUPT_BITS;
1223 		src_writel(dev, MUnit.IOAR, val);
1224 		src_readl(dev, MUnit.IOAR);
1225 		val = src_readl(dev, MUnit.OIMR);
1226 		src_writel(dev, MUnit.OIMR,
1227 				val & (~(PMC_GLOBAL_INT_BIT2)));
1228 		break;
1229 
1230 	default:
1231 		break;
1232 	}
1233 }
1234 
1235 static int aac_src_get_sync_status(struct aac_dev *dev)
1236 {
1237 
1238 	int val;
1239 
1240 	if (dev->msi_enabled)
1241 		val = src_readl(dev, MUnit.ODR_MSI) & 0x1000 ? 1 : 0;
1242 	else
1243 		val = src_readl(dev, MUnit.ODR_R) >> SRC_ODR_SHIFT;
1244 
1245 	return val;
1246 }
1247