1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  *
4  *			Linux MegaRAID device driver
5  *
6  * Copyright (c) 2003-2004  LSI Logic Corporation.
7  *
8  * FILE		: megaraid_mm.c
9  * Version	: v2.20.2.7 (Jul 16 2006)
10  *
11  * Common management module
12  */
13 #include <linux/sched.h>
14 #include <linux/slab.h>
15 #include <linux/mutex.h>
16 #include "megaraid_mm.h"
17 
18 
19 // Entry points for char node driver
20 static DEFINE_MUTEX(mraid_mm_mutex);
21 static int mraid_mm_open(struct inode *, struct file *);
22 static long mraid_mm_unlocked_ioctl(struct file *, uint, unsigned long);
23 
24 
25 // routines to convert to and from the old the format
26 static int mimd_to_kioc(mimd_t __user *, mraid_mmadp_t *, uioc_t *);
27 static int kioc_to_mimd(uioc_t *, mimd_t __user *);
28 
29 
30 // Helper functions
31 static int handle_drvrcmd(void __user *, uint8_t, int *);
32 static int lld_ioctl(mraid_mmadp_t *, uioc_t *);
33 static void ioctl_done(uioc_t *);
34 static void lld_timedout(struct timer_list *);
35 static void hinfo_to_cinfo(mraid_hba_info_t *, mcontroller_t *);
36 static mraid_mmadp_t *mraid_mm_get_adapter(mimd_t __user *, int *);
37 static uioc_t *mraid_mm_alloc_kioc(mraid_mmadp_t *);
38 static void mraid_mm_dealloc_kioc(mraid_mmadp_t *, uioc_t *);
39 static int mraid_mm_attach_buf(mraid_mmadp_t *, uioc_t *, int);
40 static int mraid_mm_setup_dma_pools(mraid_mmadp_t *);
41 static void mraid_mm_free_adp_resources(mraid_mmadp_t *);
42 static void mraid_mm_teardown_dma_pools(mraid_mmadp_t *);
43 
44 MODULE_AUTHOR("LSI Logic Corporation");
45 MODULE_DESCRIPTION("LSI Logic Management Module");
46 MODULE_LICENSE("GPL");
47 MODULE_VERSION(LSI_COMMON_MOD_VERSION);
48 
49 static int dbglevel = CL_ANN;
50 module_param_named(dlevel, dbglevel, int, 0);
51 MODULE_PARM_DESC(dlevel, "Debug level (default=0)");
52 
53 EXPORT_SYMBOL(mraid_mm_register_adp);
54 EXPORT_SYMBOL(mraid_mm_unregister_adp);
55 EXPORT_SYMBOL(mraid_mm_adapter_app_handle);
56 
57 static uint32_t drvr_ver	= 0x02200207;
58 
59 static int adapters_count_g;
60 static struct list_head adapters_list_g;
61 
62 static wait_queue_head_t wait_q;
63 
64 static const struct file_operations lsi_fops = {
65 	.open	= mraid_mm_open,
66 	.unlocked_ioctl = mraid_mm_unlocked_ioctl,
67 	.compat_ioctl = compat_ptr_ioctl,
68 	.owner	= THIS_MODULE,
69 	.llseek = noop_llseek,
70 };
71 
72 static struct miscdevice megaraid_mm_dev = {
73 	.minor	= MISC_DYNAMIC_MINOR,
74 	.name   = "megadev0",
75 	.fops   = &lsi_fops,
76 };
77 
78 /**
79  * mraid_mm_open - open routine for char node interface
80  * @inode	: unused
81  * @filep	: unused
82  *
83  * Allow ioctl operations by apps only if they have superuser privilege.
84  */
85 static int
86 mraid_mm_open(struct inode *inode, struct file *filep)
87 {
88 	/*
89 	 * Only allow superuser to access private ioctl interface
90 	 */
91 	if (!capable(CAP_SYS_ADMIN)) return (-EACCES);
92 
93 	return 0;
94 }
95 
96 /**
97  * mraid_mm_ioctl - module entry-point for ioctls
98  * @filep	: file operations pointer (ignored)
99  * @cmd		: ioctl command
100  * @arg		: user ioctl packet
101  */
102 static int
103 mraid_mm_ioctl(struct file *filep, unsigned int cmd, unsigned long arg)
104 {
105 	uioc_t		*kioc;
106 	char		signature[EXT_IOCTL_SIGN_SZ]	= {0};
107 	int		rval;
108 	mraid_mmadp_t	*adp;
109 	uint8_t		old_ioctl;
110 	int		drvrcmd_rval;
111 	void __user *argp = (void __user *)arg;
112 
113 	/*
114 	 * Make sure only USCSICMD are issued through this interface.
115 	 * MIMD application would still fire different command.
116 	 */
117 
118 	if ((_IOC_TYPE(cmd) != MEGAIOC_MAGIC) && (cmd != USCSICMD)) {
119 		return (-EINVAL);
120 	}
121 
122 	/*
123 	 * Look for signature to see if this is the new or old ioctl format.
124 	 */
125 	if (copy_from_user(signature, argp, EXT_IOCTL_SIGN_SZ)) {
126 		con_log(CL_ANN, (KERN_WARNING
127 			"megaraid cmm: copy from usr addr failed\n"));
128 		return (-EFAULT);
129 	}
130 
131 	if (memcmp(signature, EXT_IOCTL_SIGN, EXT_IOCTL_SIGN_SZ) == 0)
132 		old_ioctl = 0;
133 	else
134 		old_ioctl = 1;
135 
136 	/*
137 	 * At present, we don't support the new ioctl packet
138 	 */
139 	if (!old_ioctl )
140 		return (-EINVAL);
141 
142 	/*
143 	 * If it is a driver ioctl (as opposed to fw ioctls), then we can
144 	 * handle the command locally. rval > 0 means it is not a drvr cmd
145 	 */
146 	rval = handle_drvrcmd(argp, old_ioctl, &drvrcmd_rval);
147 
148 	if (rval < 0)
149 		return rval;
150 	else if (rval == 0)
151 		return drvrcmd_rval;
152 
153 	rval = 0;
154 	if ((adp = mraid_mm_get_adapter(argp, &rval)) == NULL) {
155 		return rval;
156 	}
157 
158 	/*
159 	 * Check if adapter can accept ioctl. We may have marked it offline
160 	 * if any previous kioc had timedout on this controller.
161 	 */
162 	if (!adp->quiescent) {
163 		con_log(CL_ANN, (KERN_WARNING
164 			"megaraid cmm: controller cannot accept cmds due to "
165 			"earlier errors\n" ));
166 		return -EFAULT;
167 	}
168 
169 	/*
170 	 * The following call will block till a kioc is available
171 	 * or return NULL if the list head is empty for the pointer
172 	 * of type mraid_mmapt passed to mraid_mm_alloc_kioc
173 	 */
174 	kioc = mraid_mm_alloc_kioc(adp);
175 	if (!kioc)
176 		return -ENXIO;
177 
178 	/*
179 	 * User sent the old mimd_t ioctl packet. Convert it to uioc_t.
180 	 */
181 	if ((rval = mimd_to_kioc(argp, adp, kioc))) {
182 		mraid_mm_dealloc_kioc(adp, kioc);
183 		return rval;
184 	}
185 
186 	kioc->done = ioctl_done;
187 
188 	/*
189 	 * Issue the IOCTL to the low level driver. After the IOCTL completes
190 	 * release the kioc if and only if it was _not_ timedout. If it was
191 	 * timedout, that means that resources are still with low level driver.
192 	 */
193 	if ((rval = lld_ioctl(adp, kioc))) {
194 
195 		if (!kioc->timedout)
196 			mraid_mm_dealloc_kioc(adp, kioc);
197 
198 		return rval;
199 	}
200 
201 	/*
202 	 * Convert the kioc back to user space
203 	 */
204 	rval = kioc_to_mimd(kioc, argp);
205 
206 	/*
207 	 * Return the kioc to free pool
208 	 */
209 	mraid_mm_dealloc_kioc(adp, kioc);
210 
211 	return rval;
212 }
213 
214 static long
215 mraid_mm_unlocked_ioctl(struct file *filep, unsigned int cmd,
216 		        unsigned long arg)
217 {
218 	int err;
219 
220 	mutex_lock(&mraid_mm_mutex);
221 	err = mraid_mm_ioctl(filep, cmd, arg);
222 	mutex_unlock(&mraid_mm_mutex);
223 
224 	return err;
225 }
226 
227 /**
228  * mraid_mm_get_adapter - Returns corresponding adapters for the mimd packet
229  * @umimd	: User space mimd_t ioctl packet
230  * @rval	: returned success/error status
231  *
232  * The function return value is a pointer to the located @adapter.
233  */
234 static mraid_mmadp_t *
235 mraid_mm_get_adapter(mimd_t __user *umimd, int *rval)
236 {
237 	mraid_mmadp_t	*adapter;
238 	mimd_t		mimd;
239 	uint32_t	adapno;
240 	int		iterator;
241 	bool		is_found;
242 
243 	if (copy_from_user(&mimd, umimd, sizeof(mimd_t))) {
244 		*rval = -EFAULT;
245 		return NULL;
246 	}
247 
248 	adapno = GETADAP(mimd.ui.fcs.adapno);
249 
250 	if (adapno >= adapters_count_g) {
251 		*rval = -ENODEV;
252 		return NULL;
253 	}
254 
255 	adapter = NULL;
256 	iterator = 0;
257 	is_found = false;
258 
259 	list_for_each_entry(adapter, &adapters_list_g, list) {
260 		if (iterator++ == adapno) {
261 			is_found = true;
262 			break;
263 		}
264 	}
265 
266 	if (!is_found) {
267 		*rval = -ENODEV;
268 		return NULL;
269 	}
270 
271 	return adapter;
272 }
273 
274 /**
275  * handle_drvrcmd - Checks if the opcode is a driver cmd and if it is, handles it.
276  * @arg		: packet sent by the user app
277  * @old_ioctl	: mimd if 1; uioc otherwise
278  * @rval	: pointer for command's returned value (not function status)
279  */
280 static int
281 handle_drvrcmd(void __user *arg, uint8_t old_ioctl, int *rval)
282 {
283 	mimd_t		__user *umimd;
284 	mimd_t		kmimd;
285 	uint8_t		opcode;
286 	uint8_t		subopcode;
287 
288 	if (old_ioctl)
289 		goto old_packet;
290 	else
291 		goto new_packet;
292 
293 new_packet:
294 	return (-ENOTSUPP);
295 
296 old_packet:
297 	*rval = 0;
298 	umimd = arg;
299 
300 	if (copy_from_user(&kmimd, umimd, sizeof(mimd_t)))
301 		return (-EFAULT);
302 
303 	opcode		= kmimd.ui.fcs.opcode;
304 	subopcode	= kmimd.ui.fcs.subopcode;
305 
306 	/*
307 	 * If the opcode is 0x82 and the subopcode is either GET_DRVRVER or
308 	 * GET_NUMADP, then we can handle. Otherwise we should return 1 to
309 	 * indicate that we cannot handle this.
310 	 */
311 	if (opcode != 0x82)
312 		return 1;
313 
314 	switch (subopcode) {
315 
316 	case MEGAIOC_QDRVRVER:
317 
318 		if (copy_to_user(kmimd.data, &drvr_ver, sizeof(uint32_t)))
319 			return (-EFAULT);
320 
321 		return 0;
322 
323 	case MEGAIOC_QNADAP:
324 
325 		*rval = adapters_count_g;
326 
327 		if (copy_to_user(kmimd.data, &adapters_count_g,
328 				sizeof(uint32_t)))
329 			return (-EFAULT);
330 
331 		return 0;
332 
333 	default:
334 		/* cannot handle */
335 		return 1;
336 	}
337 
338 	return 0;
339 }
340 
341 
342 /**
343  * mimd_to_kioc	- Converter from old to new ioctl format
344  * @umimd	: user space old MIMD IOCTL
345  * @adp		: adapter softstate
346  * @kioc	: kernel space new format IOCTL
347  *
348  * Routine to convert MIMD interface IOCTL to new interface IOCTL packet. The
349  * new packet is in kernel space so that driver can perform operations on it
350  * freely.
351  */
352 
353 static int
354 mimd_to_kioc(mimd_t __user *umimd, mraid_mmadp_t *adp, uioc_t *kioc)
355 {
356 	mbox64_t		*mbox64;
357 	mbox_t			*mbox;
358 	mraid_passthru_t	*pthru32;
359 	uint32_t		adapno;
360 	uint8_t			opcode;
361 	uint8_t			subopcode;
362 	mimd_t			mimd;
363 
364 	if (copy_from_user(&mimd, umimd, sizeof(mimd_t)))
365 		return (-EFAULT);
366 
367 	/*
368 	 * Applications are not allowed to send extd pthru
369 	 */
370 	if ((mimd.mbox[0] == MBOXCMD_PASSTHRU64) ||
371 			(mimd.mbox[0] == MBOXCMD_EXTPTHRU))
372 		return (-EINVAL);
373 
374 	opcode		= mimd.ui.fcs.opcode;
375 	subopcode	= mimd.ui.fcs.subopcode;
376 	adapno		= GETADAP(mimd.ui.fcs.adapno);
377 
378 	if (adapno >= adapters_count_g)
379 		return (-ENODEV);
380 
381 	kioc->adapno	= adapno;
382 	kioc->mb_type	= MBOX_LEGACY;
383 	kioc->app_type	= APPTYPE_MIMD;
384 
385 	switch (opcode) {
386 
387 	case 0x82:
388 
389 		if (subopcode == MEGAIOC_QADAPINFO) {
390 
391 			kioc->opcode	= GET_ADAP_INFO;
392 			kioc->data_dir	= UIOC_RD;
393 			kioc->xferlen	= sizeof(mraid_hba_info_t);
394 
395 			if (mraid_mm_attach_buf(adp, kioc, kioc->xferlen))
396 				return (-ENOMEM);
397 		}
398 		else {
399 			con_log(CL_ANN, (KERN_WARNING
400 					"megaraid cmm: Invalid subop\n"));
401 			return (-EINVAL);
402 		}
403 
404 		break;
405 
406 	case 0x81:
407 
408 		kioc->opcode		= MBOX_CMD;
409 		kioc->xferlen		= mimd.ui.fcs.length;
410 		kioc->user_data_len	= kioc->xferlen;
411 		kioc->user_data		= mimd.ui.fcs.buffer;
412 
413 		if (mraid_mm_attach_buf(adp, kioc, kioc->xferlen))
414 			return (-ENOMEM);
415 
416 		if (mimd.outlen) kioc->data_dir  = UIOC_RD;
417 		if (mimd.inlen) kioc->data_dir |= UIOC_WR;
418 
419 		break;
420 
421 	case 0x80:
422 
423 		kioc->opcode		= MBOX_CMD;
424 		kioc->xferlen		= (mimd.outlen > mimd.inlen) ?
425 						mimd.outlen : mimd.inlen;
426 		kioc->user_data_len	= kioc->xferlen;
427 		kioc->user_data		= mimd.data;
428 
429 		if (mraid_mm_attach_buf(adp, kioc, kioc->xferlen))
430 			return (-ENOMEM);
431 
432 		if (mimd.outlen) kioc->data_dir  = UIOC_RD;
433 		if (mimd.inlen) kioc->data_dir |= UIOC_WR;
434 
435 		break;
436 
437 	default:
438 		return (-EINVAL);
439 	}
440 
441 	/*
442 	 * If driver command, nothing else to do
443 	 */
444 	if (opcode == 0x82)
445 		return 0;
446 
447 	/*
448 	 * This is a mailbox cmd; copy the mailbox from mimd
449 	 */
450 	mbox64	= (mbox64_t *)((unsigned long)kioc->cmdbuf);
451 	mbox	= &mbox64->mbox32;
452 	memcpy(mbox, mimd.mbox, 14);
453 
454 	if (mbox->cmd != MBOXCMD_PASSTHRU) {	// regular DCMD
455 
456 		mbox->xferaddr	= (uint32_t)kioc->buf_paddr;
457 
458 		if (kioc->data_dir & UIOC_WR) {
459 			if (copy_from_user(kioc->buf_vaddr, kioc->user_data,
460 							kioc->xferlen)) {
461 				return (-EFAULT);
462 			}
463 		}
464 
465 		return 0;
466 	}
467 
468 	/*
469 	 * This is a regular 32-bit pthru cmd; mbox points to pthru struct.
470 	 * Just like in above case, the beginning for memblk is treated as
471 	 * a mailbox. The passthru will begin at next 1K boundary. And the
472 	 * data will start 1K after that.
473 	 */
474 	pthru32			= kioc->pthru32;
475 	kioc->user_pthru	= &umimd->pthru;
476 	mbox->xferaddr		= (uint32_t)kioc->pthru32_h;
477 
478 	if (copy_from_user(pthru32, kioc->user_pthru,
479 			sizeof(mraid_passthru_t))) {
480 		return (-EFAULT);
481 	}
482 
483 	pthru32->dataxferaddr	= kioc->buf_paddr;
484 	if (kioc->data_dir & UIOC_WR) {
485 		if (pthru32->dataxferlen > kioc->xferlen)
486 			return -EINVAL;
487 		if (copy_from_user(kioc->buf_vaddr, kioc->user_data,
488 						pthru32->dataxferlen)) {
489 			return (-EFAULT);
490 		}
491 	}
492 
493 	return 0;
494 }
495 
496 /**
497  * mraid_mm_attach_buf - Attach a free dma buffer for required size
498  * @adp		: Adapter softstate
499  * @kioc	: kioc that the buffer needs to be attached to
500  * @xferlen	: required length for buffer
501  *
502  * First we search for a pool with smallest buffer that is >= @xferlen. If
503  * that pool has no free buffer, we will try for the next bigger size. If none
504  * is available, we will try to allocate the smallest buffer that is >=
505  * @xferlen and attach it the pool.
506  */
507 static int
508 mraid_mm_attach_buf(mraid_mmadp_t *adp, uioc_t *kioc, int xferlen)
509 {
510 	mm_dmapool_t	*pool;
511 	int		right_pool = -1;
512 	unsigned long	flags;
513 	int		i;
514 
515 	kioc->pool_index	= -1;
516 	kioc->buf_vaddr		= NULL;
517 	kioc->buf_paddr		= 0;
518 	kioc->free_buf		= 0;
519 
520 	/*
521 	 * We need xferlen amount of memory. See if we can get it from our
522 	 * dma pools. If we don't get exact size, we will try bigger buffer
523 	 */
524 
525 	for (i = 0; i < MAX_DMA_POOLS; i++) {
526 
527 		pool = &adp->dma_pool_list[i];
528 
529 		if (xferlen > pool->buf_size)
530 			continue;
531 
532 		if (right_pool == -1)
533 			right_pool = i;
534 
535 		spin_lock_irqsave(&pool->lock, flags);
536 
537 		if (!pool->in_use) {
538 
539 			pool->in_use		= 1;
540 			kioc->pool_index	= i;
541 			kioc->buf_vaddr		= pool->vaddr;
542 			kioc->buf_paddr		= pool->paddr;
543 
544 			spin_unlock_irqrestore(&pool->lock, flags);
545 			return 0;
546 		}
547 		else {
548 			spin_unlock_irqrestore(&pool->lock, flags);
549 			continue;
550 		}
551 	}
552 
553 	/*
554 	 * If xferlen doesn't match any of our pools, return error
555 	 */
556 	if (right_pool == -1)
557 		return -EINVAL;
558 
559 	/*
560 	 * We did not get any buffer from the preallocated pool. Let us try
561 	 * to allocate one new buffer. NOTE: This is a blocking call.
562 	 */
563 	pool = &adp->dma_pool_list[right_pool];
564 
565 	spin_lock_irqsave(&pool->lock, flags);
566 
567 	kioc->pool_index	= right_pool;
568 	kioc->free_buf		= 1;
569 	kioc->buf_vaddr		= dma_pool_alloc(pool->handle, GFP_ATOMIC,
570 							&kioc->buf_paddr);
571 	spin_unlock_irqrestore(&pool->lock, flags);
572 
573 	if (!kioc->buf_vaddr)
574 		return -ENOMEM;
575 
576 	return 0;
577 }
578 
579 /**
580  * mraid_mm_alloc_kioc - Returns a uioc_t from free list
581  * @adp	: Adapter softstate for this module
582  *
583  * The kioc_semaphore is initialized with number of kioc nodes in the
584  * free kioc pool. If the kioc pool is empty, this function blocks till
585  * a kioc becomes free.
586  */
587 static uioc_t *
588 mraid_mm_alloc_kioc(mraid_mmadp_t *adp)
589 {
590 	uioc_t			*kioc;
591 	struct list_head*	head;
592 	unsigned long		flags;
593 
594 	down(&adp->kioc_semaphore);
595 
596 	spin_lock_irqsave(&adp->kioc_pool_lock, flags);
597 
598 	head = &adp->kioc_pool;
599 
600 	if (list_empty(head)) {
601 		up(&adp->kioc_semaphore);
602 		spin_unlock_irqrestore(&adp->kioc_pool_lock, flags);
603 
604 		con_log(CL_ANN, ("megaraid cmm: kioc list empty!\n"));
605 		return NULL;
606 	}
607 
608 	kioc = list_entry(head->next, uioc_t, list);
609 	list_del_init(&kioc->list);
610 
611 	spin_unlock_irqrestore(&adp->kioc_pool_lock, flags);
612 
613 	memset((caddr_t)(unsigned long)kioc->cmdbuf, 0, sizeof(mbox64_t));
614 	memset((caddr_t) kioc->pthru32, 0, sizeof(mraid_passthru_t));
615 
616 	kioc->buf_vaddr		= NULL;
617 	kioc->buf_paddr		= 0;
618 	kioc->pool_index	=-1;
619 	kioc->free_buf		= 0;
620 	kioc->user_data		= NULL;
621 	kioc->user_data_len	= 0;
622 	kioc->user_pthru	= NULL;
623 	kioc->timedout		= 0;
624 
625 	return kioc;
626 }
627 
628 /**
629  * mraid_mm_dealloc_kioc - Return kioc to free pool
630  * @adp		: Adapter softstate
631  * @kioc	: uioc_t node to be returned to free pool
632  */
633 static void
634 mraid_mm_dealloc_kioc(mraid_mmadp_t *adp, uioc_t *kioc)
635 {
636 	mm_dmapool_t	*pool;
637 	unsigned long	flags;
638 
639 	if (kioc->pool_index != -1) {
640 		pool = &adp->dma_pool_list[kioc->pool_index];
641 
642 		/* This routine may be called in non-isr context also */
643 		spin_lock_irqsave(&pool->lock, flags);
644 
645 		/*
646 		 * While attaching the dma buffer, if we didn't get the
647 		 * required buffer from the pool, we would have allocated
648 		 * it at the run time and set the free_buf flag. We must
649 		 * free that buffer. Otherwise, just mark that the buffer is
650 		 * not in use
651 		 */
652 		if (kioc->free_buf == 1)
653 			dma_pool_free(pool->handle, kioc->buf_vaddr,
654 							kioc->buf_paddr);
655 		else
656 			pool->in_use = 0;
657 
658 		spin_unlock_irqrestore(&pool->lock, flags);
659 	}
660 
661 	/* Return the kioc to the free pool */
662 	spin_lock_irqsave(&adp->kioc_pool_lock, flags);
663 	list_add(&kioc->list, &adp->kioc_pool);
664 	spin_unlock_irqrestore(&adp->kioc_pool_lock, flags);
665 
666 	/* increment the free kioc count */
667 	up(&adp->kioc_semaphore);
668 
669 	return;
670 }
671 
672 /**
673  * lld_ioctl - Routine to issue ioctl to low level drvr
674  * @adp		: The adapter handle
675  * @kioc	: The ioctl packet with kernel addresses
676  */
677 static int
678 lld_ioctl(mraid_mmadp_t *adp, uioc_t *kioc)
679 {
680 	int			rval;
681 	struct uioc_timeout	timeout = { };
682 
683 	kioc->status	= -ENODATA;
684 	rval		= adp->issue_uioc(adp->drvr_data, kioc, IOCTL_ISSUE);
685 
686 	if (rval) return rval;
687 
688 	/*
689 	 * Start the timer
690 	 */
691 	if (adp->timeout > 0) {
692 		timeout.uioc = kioc;
693 		timer_setup_on_stack(&timeout.timer, lld_timedout, 0);
694 
695 		timeout.timer.expires	= jiffies + adp->timeout * HZ;
696 
697 		add_timer(&timeout.timer);
698 	}
699 
700 	/*
701 	 * Wait till the low level driver completes the ioctl. After this
702 	 * call, the ioctl either completed successfully or timedout.
703 	 */
704 	wait_event(wait_q, (kioc->status != -ENODATA));
705 	if (timeout.timer.function) {
706 		del_timer_sync(&timeout.timer);
707 		destroy_timer_on_stack(&timeout.timer);
708 	}
709 
710 	/*
711 	 * If the command had timedout, we mark the controller offline
712 	 * before returning
713 	 */
714 	if (kioc->timedout) {
715 		adp->quiescent = 0;
716 	}
717 
718 	return kioc->status;
719 }
720 
721 
722 /**
723  * ioctl_done - callback from the low level driver
724  * @kioc	: completed ioctl packet
725  */
726 static void
727 ioctl_done(uioc_t *kioc)
728 {
729 	uint32_t	adapno;
730 	int		iterator;
731 	mraid_mmadp_t*	adapter;
732 	bool		is_found;
733 
734 	/*
735 	 * When the kioc returns from driver, make sure it still doesn't
736 	 * have ENODATA in status. Otherwise, driver will hang on wait_event
737 	 * forever
738 	 */
739 	if (kioc->status == -ENODATA) {
740 		con_log(CL_ANN, (KERN_WARNING
741 			"megaraid cmm: lld didn't change status!\n"));
742 
743 		kioc->status = -EINVAL;
744 	}
745 
746 	/*
747 	 * Check if this kioc was timedout before. If so, nobody is waiting
748 	 * on this kioc. We don't have to wake up anybody. Instead, we just
749 	 * have to free the kioc
750 	 */
751 	if (kioc->timedout) {
752 		iterator	= 0;
753 		adapter		= NULL;
754 		adapno		= kioc->adapno;
755 		is_found	= false;
756 
757 		con_log(CL_ANN, ( KERN_WARNING "megaraid cmm: completed "
758 					"ioctl that was timedout before\n"));
759 
760 		list_for_each_entry(adapter, &adapters_list_g, list) {
761 			if (iterator++ == adapno) {
762 				is_found = true;
763 				break;
764 			}
765 		}
766 
767 		kioc->timedout = 0;
768 
769 		if (is_found)
770 			mraid_mm_dealloc_kioc( adapter, kioc );
771 
772 	}
773 	else {
774 		wake_up(&wait_q);
775 	}
776 }
777 
778 
779 /**
780  * lld_timedout	- callback from the expired timer
781  * @t		: timer that timed out
782  */
783 static void
784 lld_timedout(struct timer_list *t)
785 {
786 	struct uioc_timeout *timeout = from_timer(timeout, t, timer);
787 	uioc_t *kioc	= timeout->uioc;
788 
789 	kioc->status 	= -ETIME;
790 	kioc->timedout	= 1;
791 
792 	con_log(CL_ANN, (KERN_WARNING "megaraid cmm: ioctl timed out\n"));
793 
794 	wake_up(&wait_q);
795 }
796 
797 
798 /**
799  * kioc_to_mimd	- Converter from new back to old format
800  * @kioc	: Kernel space IOCTL packet (successfully issued)
801  * @mimd	: User space MIMD packet
802  */
803 static int
804 kioc_to_mimd(uioc_t *kioc, mimd_t __user *mimd)
805 {
806 	mimd_t			kmimd;
807 	uint8_t			opcode;
808 	uint8_t			subopcode;
809 
810 	mbox64_t		*mbox64;
811 	mraid_passthru_t	__user *upthru32;
812 	mraid_passthru_t	*kpthru32;
813 	mcontroller_t		cinfo;
814 	mraid_hba_info_t	*hinfo;
815 
816 
817 	if (copy_from_user(&kmimd, mimd, sizeof(mimd_t)))
818 		return (-EFAULT);
819 
820 	opcode		= kmimd.ui.fcs.opcode;
821 	subopcode	= kmimd.ui.fcs.subopcode;
822 
823 	if (opcode == 0x82) {
824 		switch (subopcode) {
825 
826 		case MEGAIOC_QADAPINFO:
827 
828 			hinfo = (mraid_hba_info_t *)(unsigned long)
829 					kioc->buf_vaddr;
830 
831 			hinfo_to_cinfo(hinfo, &cinfo);
832 
833 			if (copy_to_user(kmimd.data, &cinfo, sizeof(cinfo)))
834 				return (-EFAULT);
835 
836 			return 0;
837 
838 		default:
839 			return (-EINVAL);
840 		}
841 
842 		return 0;
843 	}
844 
845 	mbox64 = (mbox64_t *)(unsigned long)kioc->cmdbuf;
846 
847 	if (kioc->user_pthru) {
848 
849 		upthru32 = kioc->user_pthru;
850 		kpthru32 = kioc->pthru32;
851 
852 		if (copy_to_user(&upthru32->scsistatus,
853 					&kpthru32->scsistatus,
854 					sizeof(uint8_t))) {
855 			return (-EFAULT);
856 		}
857 	}
858 
859 	if (kioc->user_data) {
860 		if (copy_to_user(kioc->user_data, kioc->buf_vaddr,
861 					kioc->user_data_len)) {
862 			return (-EFAULT);
863 		}
864 	}
865 
866 	if (copy_to_user(&mimd->mbox[17],
867 			&mbox64->mbox32.status, sizeof(uint8_t))) {
868 		return (-EFAULT);
869 	}
870 
871 	return 0;
872 }
873 
874 
875 /**
876  * hinfo_to_cinfo - Convert new format hba info into old format
877  * @hinfo	: New format, more comprehensive adapter info
878  * @cinfo	: Old format adapter info to support mimd_t apps
879  */
880 static void
881 hinfo_to_cinfo(mraid_hba_info_t *hinfo, mcontroller_t *cinfo)
882 {
883 	if (!hinfo || !cinfo)
884 		return;
885 
886 	cinfo->base		= hinfo->baseport;
887 	cinfo->irq		= hinfo->irq;
888 	cinfo->numldrv		= hinfo->num_ldrv;
889 	cinfo->pcibus		= hinfo->pci_bus;
890 	cinfo->pcidev		= hinfo->pci_slot;
891 	cinfo->pcifun		= PCI_FUNC(hinfo->pci_dev_fn);
892 	cinfo->pciid		= hinfo->pci_device_id;
893 	cinfo->pcivendor	= hinfo->pci_vendor_id;
894 	cinfo->pcislot		= hinfo->pci_slot;
895 	cinfo->uid		= hinfo->unique_id;
896 }
897 
898 
899 /**
900  * mraid_mm_register_adp - Registration routine for low level drivers
901  * @lld_adp	: Adapter object
902  */
903 int
904 mraid_mm_register_adp(mraid_mmadp_t *lld_adp)
905 {
906 	mraid_mmadp_t	*adapter;
907 	mbox64_t	*mbox_list;
908 	uioc_t		*kioc;
909 	uint32_t	rval;
910 	int		i;
911 
912 
913 	if (lld_adp->drvr_type != DRVRTYPE_MBOX)
914 		return (-EINVAL);
915 
916 	adapter = kzalloc(sizeof(mraid_mmadp_t), GFP_KERNEL);
917 
918 	if (!adapter)
919 		return -ENOMEM;
920 
921 
922 	adapter->unique_id	= lld_adp->unique_id;
923 	adapter->drvr_type	= lld_adp->drvr_type;
924 	adapter->drvr_data	= lld_adp->drvr_data;
925 	adapter->pdev		= lld_adp->pdev;
926 	adapter->issue_uioc	= lld_adp->issue_uioc;
927 	adapter->timeout	= lld_adp->timeout;
928 	adapter->max_kioc	= lld_adp->max_kioc;
929 	adapter->quiescent	= 1;
930 
931 	/*
932 	 * Allocate single blocks of memory for all required kiocs,
933 	 * mailboxes and passthru structures.
934 	 */
935 	adapter->kioc_list	= kmalloc_array(lld_adp->max_kioc,
936 						  sizeof(uioc_t),
937 						  GFP_KERNEL);
938 	adapter->mbox_list	= kmalloc_array(lld_adp->max_kioc,
939 						  sizeof(mbox64_t),
940 						  GFP_KERNEL);
941 	adapter->pthru_dma_pool = dma_pool_create("megaraid mm pthru pool",
942 						&adapter->pdev->dev,
943 						sizeof(mraid_passthru_t),
944 						16, 0);
945 
946 	if (!adapter->kioc_list || !adapter->mbox_list ||
947 			!adapter->pthru_dma_pool) {
948 
949 		con_log(CL_ANN, (KERN_WARNING
950 			"megaraid cmm: out of memory, %s %d\n", __func__,
951 			__LINE__));
952 
953 		rval = (-ENOMEM);
954 
955 		goto memalloc_error;
956 	}
957 
958 	/*
959 	 * Slice kioc_list and make a kioc_pool with the individiual kiocs
960 	 */
961 	INIT_LIST_HEAD(&adapter->kioc_pool);
962 	spin_lock_init(&adapter->kioc_pool_lock);
963 	sema_init(&adapter->kioc_semaphore, lld_adp->max_kioc);
964 
965 	mbox_list	= (mbox64_t *)adapter->mbox_list;
966 
967 	for (i = 0; i < lld_adp->max_kioc; i++) {
968 
969 		kioc		= adapter->kioc_list + i;
970 		kioc->cmdbuf	= (uint64_t)(unsigned long)(mbox_list + i);
971 		kioc->pthru32	= dma_pool_alloc(adapter->pthru_dma_pool,
972 						GFP_KERNEL, &kioc->pthru32_h);
973 
974 		if (!kioc->pthru32) {
975 
976 			con_log(CL_ANN, (KERN_WARNING
977 				"megaraid cmm: out of memory, %s %d\n",
978 					__func__, __LINE__));
979 
980 			rval = (-ENOMEM);
981 
982 			goto pthru_dma_pool_error;
983 		}
984 
985 		list_add_tail(&kioc->list, &adapter->kioc_pool);
986 	}
987 
988 	// Setup the dma pools for data buffers
989 	if ((rval = mraid_mm_setup_dma_pools(adapter)) != 0) {
990 		goto dma_pool_error;
991 	}
992 
993 	list_add_tail(&adapter->list, &adapters_list_g);
994 
995 	adapters_count_g++;
996 
997 	return 0;
998 
999 dma_pool_error:
1000 	/* Do nothing */
1001 
1002 pthru_dma_pool_error:
1003 
1004 	for (i = 0; i < lld_adp->max_kioc; i++) {
1005 		kioc = adapter->kioc_list + i;
1006 		if (kioc->pthru32) {
1007 			dma_pool_free(adapter->pthru_dma_pool, kioc->pthru32,
1008 				kioc->pthru32_h);
1009 		}
1010 	}
1011 
1012 memalloc_error:
1013 
1014 	kfree(adapter->kioc_list);
1015 	kfree(adapter->mbox_list);
1016 
1017 	dma_pool_destroy(adapter->pthru_dma_pool);
1018 
1019 	kfree(adapter);
1020 
1021 	return rval;
1022 }
1023 
1024 
1025 /**
1026  * mraid_mm_adapter_app_handle - return the application handle for this adapter
1027  * @unique_id	: adapter unique identifier
1028  *
1029  * For the given driver data, locate the adapter in our global list and
1030  * return the corresponding handle, which is also used by applications to
1031  * uniquely identify an adapter.
1032  *
1033  * Return adapter handle if found in the list.
1034  * Return 0 if adapter could not be located, should never happen though.
1035  */
1036 uint32_t
1037 mraid_mm_adapter_app_handle(uint32_t unique_id)
1038 {
1039 	mraid_mmadp_t	*adapter;
1040 	mraid_mmadp_t	*tmp;
1041 	int		index = 0;
1042 
1043 	list_for_each_entry_safe(adapter, tmp, &adapters_list_g, list) {
1044 
1045 		if (adapter->unique_id == unique_id) {
1046 
1047 			return MKADAP(index);
1048 		}
1049 
1050 		index++;
1051 	}
1052 
1053 	return 0;
1054 }
1055 
1056 
1057 /**
1058  * mraid_mm_setup_dma_pools - Set up dma buffer pools per adapter
1059  * @adp	: Adapter softstate
1060  *
1061  * We maintain a pool of dma buffers per each adapter. Each pool has one
1062  * buffer. E.g, we may have 5 dma pools - one each for 4k, 8k ... 64k buffers.
1063  * We have just one 4k buffer in 4k pool, one 8k buffer in 8k pool etc. We
1064  * dont' want to waste too much memory by allocating more buffers per each
1065  * pool.
1066  */
1067 static int
1068 mraid_mm_setup_dma_pools(mraid_mmadp_t *adp)
1069 {
1070 	mm_dmapool_t	*pool;
1071 	int		bufsize;
1072 	int		i;
1073 
1074 	/*
1075 	 * Create MAX_DMA_POOLS number of pools
1076 	 */
1077 	bufsize = MRAID_MM_INIT_BUFF_SIZE;
1078 
1079 	for (i = 0; i < MAX_DMA_POOLS; i++){
1080 
1081 		pool = &adp->dma_pool_list[i];
1082 
1083 		pool->buf_size = bufsize;
1084 		spin_lock_init(&pool->lock);
1085 
1086 		pool->handle = dma_pool_create("megaraid mm data buffer",
1087 						&adp->pdev->dev, bufsize,
1088 						16, 0);
1089 
1090 		if (!pool->handle) {
1091 			goto dma_pool_setup_error;
1092 		}
1093 
1094 		pool->vaddr = dma_pool_alloc(pool->handle, GFP_KERNEL,
1095 							&pool->paddr);
1096 
1097 		if (!pool->vaddr)
1098 			goto dma_pool_setup_error;
1099 
1100 		bufsize = bufsize * 2;
1101 	}
1102 
1103 	return 0;
1104 
1105 dma_pool_setup_error:
1106 
1107 	mraid_mm_teardown_dma_pools(adp);
1108 	return (-ENOMEM);
1109 }
1110 
1111 
1112 /**
1113  * mraid_mm_unregister_adp - Unregister routine for low level drivers
1114  * @unique_id	: UID of the adpater
1115  *
1116  * Assumes no outstanding ioctls to llds.
1117  */
1118 int
1119 mraid_mm_unregister_adp(uint32_t unique_id)
1120 {
1121 	mraid_mmadp_t	*adapter;
1122 	mraid_mmadp_t	*tmp;
1123 
1124 	list_for_each_entry_safe(adapter, tmp, &adapters_list_g, list) {
1125 
1126 
1127 		if (adapter->unique_id == unique_id) {
1128 
1129 			adapters_count_g--;
1130 
1131 			list_del_init(&adapter->list);
1132 
1133 			mraid_mm_free_adp_resources(adapter);
1134 
1135 			kfree(adapter);
1136 
1137 			con_log(CL_ANN, (
1138 				"megaraid cmm: Unregistered one adapter:%#x\n",
1139 				unique_id));
1140 
1141 			return 0;
1142 		}
1143 	}
1144 
1145 	return (-ENODEV);
1146 }
1147 
1148 /**
1149  * mraid_mm_free_adp_resources - Free adapter softstate
1150  * @adp	: Adapter softstate
1151  */
1152 static void
1153 mraid_mm_free_adp_resources(mraid_mmadp_t *adp)
1154 {
1155 	uioc_t	*kioc;
1156 	int	i;
1157 
1158 	mraid_mm_teardown_dma_pools(adp);
1159 
1160 	for (i = 0; i < adp->max_kioc; i++) {
1161 
1162 		kioc = adp->kioc_list + i;
1163 
1164 		dma_pool_free(adp->pthru_dma_pool, kioc->pthru32,
1165 				kioc->pthru32_h);
1166 	}
1167 
1168 	kfree(adp->kioc_list);
1169 	kfree(adp->mbox_list);
1170 
1171 	dma_pool_destroy(adp->pthru_dma_pool);
1172 
1173 
1174 	return;
1175 }
1176 
1177 
1178 /**
1179  * mraid_mm_teardown_dma_pools - Free all per adapter dma buffers
1180  * @adp	: Adapter softstate
1181  */
1182 static void
1183 mraid_mm_teardown_dma_pools(mraid_mmadp_t *adp)
1184 {
1185 	int		i;
1186 	mm_dmapool_t	*pool;
1187 
1188 	for (i = 0; i < MAX_DMA_POOLS; i++) {
1189 
1190 		pool = &adp->dma_pool_list[i];
1191 
1192 		if (pool->handle) {
1193 
1194 			if (pool->vaddr)
1195 				dma_pool_free(pool->handle, pool->vaddr,
1196 							pool->paddr);
1197 
1198 			dma_pool_destroy(pool->handle);
1199 			pool->handle = NULL;
1200 		}
1201 	}
1202 
1203 	return;
1204 }
1205 
1206 /**
1207  * mraid_mm_init	- Module entry point
1208  */
1209 static int __init
1210 mraid_mm_init(void)
1211 {
1212 	int err;
1213 
1214 	// Announce the driver version
1215 	con_log(CL_ANN, (KERN_INFO "megaraid cmm: %s %s\n",
1216 		LSI_COMMON_MOD_VERSION, LSI_COMMON_MOD_EXT_VERSION));
1217 
1218 	err = misc_register(&megaraid_mm_dev);
1219 	if (err < 0) {
1220 		con_log(CL_ANN, ("megaraid cmm: cannot register misc device\n"));
1221 		return err;
1222 	}
1223 
1224 	init_waitqueue_head(&wait_q);
1225 
1226 	INIT_LIST_HEAD(&adapters_list_g);
1227 
1228 	return 0;
1229 }
1230 
1231 
1232 /**
1233  * mraid_mm_exit	- Module exit point
1234  */
1235 static void __exit
1236 mraid_mm_exit(void)
1237 {
1238 	con_log(CL_DLEVEL1 , ("exiting common mod\n"));
1239 
1240 	misc_deregister(&megaraid_mm_dev);
1241 }
1242 
1243 module_init(mraid_mm_init);
1244 module_exit(mraid_mm_exit);
1245 
1246 /* vi: set ts=8 sw=8 tw=78: */
1247