xref: /openbmc/linux/drivers/hsi/clients/cmt_speech.c (revision 0c6dfa75)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * cmt_speech.c - HSI CMT speech driver
4  *
5  * Copyright (C) 2008,2009,2010 Nokia Corporation. All rights reserved.
6  *
7  * Contact: Kai Vehmanen <kai.vehmanen@nokia.com>
8  * Original author: Peter Ujfalusi <peter.ujfalusi@nokia.com>
9  */
10 
11 #include <linux/errno.h>
12 #include <linux/module.h>
13 #include <linux/types.h>
14 #include <linux/init.h>
15 #include <linux/device.h>
16 #include <linux/miscdevice.h>
17 #include <linux/mm.h>
18 #include <linux/slab.h>
19 #include <linux/fs.h>
20 #include <linux/poll.h>
21 #include <linux/sched/signal.h>
22 #include <linux/ioctl.h>
23 #include <linux/uaccess.h>
24 #include <linux/pm_qos.h>
25 #include <linux/hsi/hsi.h>
26 #include <linux/hsi/ssi_protocol.h>
27 #include <linux/hsi/cs-protocol.h>
28 
29 #define CS_MMAP_SIZE	PAGE_SIZE
30 
31 struct char_queue {
32 	struct list_head	list;
33 	u32			msg;
34 };
35 
36 struct cs_char {
37 	unsigned int		opened;
38 	struct hsi_client	*cl;
39 	struct cs_hsi_iface	*hi;
40 	struct list_head	chardev_queue;
41 	struct list_head	dataind_queue;
42 	int			dataind_pending;
43 	/* mmap things */
44 	unsigned long		mmap_base;
45 	unsigned long		mmap_size;
46 	spinlock_t		lock;
47 	struct fasync_struct	*async_queue;
48 	wait_queue_head_t	wait;
49 	/* hsi channel ids */
50 	int                     channel_id_cmd;
51 	int                     channel_id_data;
52 };
53 
54 #define SSI_CHANNEL_STATE_READING	1
55 #define SSI_CHANNEL_STATE_WRITING	(1 << 1)
56 #define SSI_CHANNEL_STATE_POLL		(1 << 2)
57 #define SSI_CHANNEL_STATE_ERROR		(1 << 3)
58 
59 #define TARGET_MASK			0xf000000
60 #define TARGET_REMOTE			(1 << CS_DOMAIN_SHIFT)
61 #define TARGET_LOCAL			0
62 
63 /* Number of pre-allocated commands buffers */
64 #define CS_MAX_CMDS		        4
65 
66 /*
67  * During data transfers, transactions must be handled
68  * within 20ms (fixed value in cmtspeech HSI protocol)
69  */
70 #define CS_QOS_LATENCY_FOR_DATA_USEC	20000
71 
72 /* Timeout to wait for pending HSI transfers to complete */
73 #define CS_HSI_TRANSFER_TIMEOUT_MS      500
74 
75 
76 #define RX_PTR_BOUNDARY_SHIFT		8
77 #define RX_PTR_MAX_SHIFT		(RX_PTR_BOUNDARY_SHIFT + \
78 						CS_MAX_BUFFERS_SHIFT)
79 struct cs_hsi_iface {
80 	struct hsi_client		*cl;
81 	struct hsi_client		*master;
82 
83 	unsigned int			iface_state;
84 	unsigned int			wakeline_state;
85 	unsigned int			control_state;
86 	unsigned int			data_state;
87 
88 	/* state exposed to application */
89 	struct cs_mmap_config_block	*mmap_cfg;
90 
91 	unsigned long			mmap_base;
92 	unsigned long			mmap_size;
93 
94 	unsigned int			rx_slot;
95 	unsigned int			tx_slot;
96 
97 	/* note: for security reasons, we do not trust the contents of
98 	 * mmap_cfg, but instead duplicate the variables here */
99 	unsigned int			buf_size;
100 	unsigned int			rx_bufs;
101 	unsigned int			tx_bufs;
102 	unsigned int			rx_ptr_boundary;
103 	unsigned int			rx_offsets[CS_MAX_BUFFERS];
104 	unsigned int			tx_offsets[CS_MAX_BUFFERS];
105 
106 	/* size of aligned memory blocks */
107 	unsigned int			slot_size;
108 	unsigned int			flags;
109 
110 	struct list_head		cmdqueue;
111 
112 	struct hsi_msg			*data_rx_msg;
113 	struct hsi_msg			*data_tx_msg;
114 	wait_queue_head_t		datawait;
115 
116 	struct pm_qos_request           pm_qos_req;
117 
118 	spinlock_t			lock;
119 };
120 
121 static struct cs_char cs_char_data;
122 
123 static void cs_hsi_read_on_control(struct cs_hsi_iface *hi);
124 static void cs_hsi_read_on_data(struct cs_hsi_iface *hi);
125 
126 static inline void rx_ptr_shift_too_big(void)
127 {
128 	BUILD_BUG_ON((1LLU << RX_PTR_MAX_SHIFT) > UINT_MAX);
129 }
130 
131 static void cs_notify(u32 message, struct list_head *head)
132 {
133 	struct char_queue *entry;
134 
135 	spin_lock(&cs_char_data.lock);
136 
137 	if (!cs_char_data.opened) {
138 		spin_unlock(&cs_char_data.lock);
139 		goto out;
140 	}
141 
142 	entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
143 	if (!entry) {
144 		dev_err(&cs_char_data.cl->device,
145 			"Can't allocate new entry for the queue.\n");
146 		spin_unlock(&cs_char_data.lock);
147 		goto out;
148 	}
149 
150 	entry->msg = message;
151 	list_add_tail(&entry->list, head);
152 
153 	spin_unlock(&cs_char_data.lock);
154 
155 	wake_up_interruptible(&cs_char_data.wait);
156 	kill_fasync(&cs_char_data.async_queue, SIGIO, POLL_IN);
157 
158 out:
159 	return;
160 }
161 
162 static u32 cs_pop_entry(struct list_head *head)
163 {
164 	struct char_queue *entry;
165 	u32 data;
166 
167 	entry = list_entry(head->next, struct char_queue, list);
168 	data = entry->msg;
169 	list_del(&entry->list);
170 	kfree(entry);
171 
172 	return data;
173 }
174 
175 static void cs_notify_control(u32 message)
176 {
177 	cs_notify(message, &cs_char_data.chardev_queue);
178 }
179 
180 static void cs_notify_data(u32 message, int maxlength)
181 {
182 	cs_notify(message, &cs_char_data.dataind_queue);
183 
184 	spin_lock(&cs_char_data.lock);
185 	cs_char_data.dataind_pending++;
186 	while (cs_char_data.dataind_pending > maxlength &&
187 				!list_empty(&cs_char_data.dataind_queue)) {
188 		dev_dbg(&cs_char_data.cl->device, "data notification "
189 		"queue overrun (%u entries)\n", cs_char_data.dataind_pending);
190 
191 		cs_pop_entry(&cs_char_data.dataind_queue);
192 		cs_char_data.dataind_pending--;
193 	}
194 	spin_unlock(&cs_char_data.lock);
195 }
196 
197 static inline void cs_set_cmd(struct hsi_msg *msg, u32 cmd)
198 {
199 	u32 *data = sg_virt(msg->sgt.sgl);
200 	*data = cmd;
201 }
202 
203 static inline u32 cs_get_cmd(struct hsi_msg *msg)
204 {
205 	u32 *data = sg_virt(msg->sgt.sgl);
206 	return *data;
207 }
208 
209 static void cs_release_cmd(struct hsi_msg *msg)
210 {
211 	struct cs_hsi_iface *hi = msg->context;
212 
213 	list_add_tail(&msg->link, &hi->cmdqueue);
214 }
215 
216 static void cs_cmd_destructor(struct hsi_msg *msg)
217 {
218 	struct cs_hsi_iface *hi = msg->context;
219 
220 	spin_lock(&hi->lock);
221 
222 	dev_dbg(&cs_char_data.cl->device, "control cmd destructor\n");
223 
224 	if (hi->iface_state != CS_STATE_CLOSED)
225 		dev_err(&hi->cl->device, "Cmd flushed while driver active\n");
226 
227 	if (msg->ttype == HSI_MSG_READ)
228 		hi->control_state &=
229 			~(SSI_CHANNEL_STATE_POLL | SSI_CHANNEL_STATE_READING);
230 	else if (msg->ttype == HSI_MSG_WRITE &&
231 			hi->control_state & SSI_CHANNEL_STATE_WRITING)
232 		hi->control_state &= ~SSI_CHANNEL_STATE_WRITING;
233 
234 	cs_release_cmd(msg);
235 
236 	spin_unlock(&hi->lock);
237 }
238 
239 static struct hsi_msg *cs_claim_cmd(struct cs_hsi_iface* ssi)
240 {
241 	struct hsi_msg *msg;
242 
243 	BUG_ON(list_empty(&ssi->cmdqueue));
244 
245 	msg = list_first_entry(&ssi->cmdqueue, struct hsi_msg, link);
246 	list_del(&msg->link);
247 	msg->destructor = cs_cmd_destructor;
248 
249 	return msg;
250 }
251 
252 static void cs_free_cmds(struct cs_hsi_iface *ssi)
253 {
254 	struct hsi_msg *msg, *tmp;
255 
256 	list_for_each_entry_safe(msg, tmp, &ssi->cmdqueue, link) {
257 		list_del(&msg->link);
258 		msg->destructor = NULL;
259 		kfree(sg_virt(msg->sgt.sgl));
260 		hsi_free_msg(msg);
261 	}
262 }
263 
264 static int cs_alloc_cmds(struct cs_hsi_iface *hi)
265 {
266 	struct hsi_msg *msg;
267 	u32 *buf;
268 	unsigned int i;
269 
270 	INIT_LIST_HEAD(&hi->cmdqueue);
271 
272 	for (i = 0; i < CS_MAX_CMDS; i++) {
273 		msg = hsi_alloc_msg(1, GFP_KERNEL);
274 		if (!msg)
275 			goto out;
276 		buf = kmalloc(sizeof(*buf), GFP_KERNEL);
277 		if (!buf) {
278 			hsi_free_msg(msg);
279 			goto out;
280 		}
281 		sg_init_one(msg->sgt.sgl, buf, sizeof(*buf));
282 		msg->channel = cs_char_data.channel_id_cmd;
283 		msg->context = hi;
284 		list_add_tail(&msg->link, &hi->cmdqueue);
285 	}
286 
287 	return 0;
288 
289 out:
290 	cs_free_cmds(hi);
291 	return -ENOMEM;
292 }
293 
294 static void cs_hsi_data_destructor(struct hsi_msg *msg)
295 {
296 	struct cs_hsi_iface *hi = msg->context;
297 	const char *dir = (msg->ttype == HSI_MSG_READ) ? "TX" : "RX";
298 
299 	dev_dbg(&cs_char_data.cl->device, "Freeing data %s message\n", dir);
300 
301 	spin_lock(&hi->lock);
302 	if (hi->iface_state != CS_STATE_CLOSED)
303 		dev_err(&cs_char_data.cl->device,
304 				"Data %s flush while device active\n", dir);
305 	if (msg->ttype == HSI_MSG_READ)
306 		hi->data_state &=
307 			~(SSI_CHANNEL_STATE_POLL | SSI_CHANNEL_STATE_READING);
308 	else
309 		hi->data_state &= ~SSI_CHANNEL_STATE_WRITING;
310 
311 	msg->status = HSI_STATUS_COMPLETED;
312 	if (unlikely(waitqueue_active(&hi->datawait)))
313 		wake_up_interruptible(&hi->datawait);
314 
315 	spin_unlock(&hi->lock);
316 }
317 
318 static int cs_hsi_alloc_data(struct cs_hsi_iface *hi)
319 {
320 	struct hsi_msg *txmsg, *rxmsg;
321 	int res = 0;
322 
323 	rxmsg = hsi_alloc_msg(1, GFP_KERNEL);
324 	if (!rxmsg) {
325 		res = -ENOMEM;
326 		goto out1;
327 	}
328 	rxmsg->channel = cs_char_data.channel_id_data;
329 	rxmsg->destructor = cs_hsi_data_destructor;
330 	rxmsg->context = hi;
331 
332 	txmsg = hsi_alloc_msg(1, GFP_KERNEL);
333 	if (!txmsg) {
334 		res = -ENOMEM;
335 		goto out2;
336 	}
337 	txmsg->channel = cs_char_data.channel_id_data;
338 	txmsg->destructor = cs_hsi_data_destructor;
339 	txmsg->context = hi;
340 
341 	hi->data_rx_msg = rxmsg;
342 	hi->data_tx_msg = txmsg;
343 
344 	return 0;
345 
346 out2:
347 	hsi_free_msg(rxmsg);
348 out1:
349 	return res;
350 }
351 
352 static void cs_hsi_free_data_msg(struct hsi_msg *msg)
353 {
354 	WARN_ON(msg->status != HSI_STATUS_COMPLETED &&
355 					msg->status != HSI_STATUS_ERROR);
356 	hsi_free_msg(msg);
357 }
358 
359 static void cs_hsi_free_data(struct cs_hsi_iface *hi)
360 {
361 	cs_hsi_free_data_msg(hi->data_rx_msg);
362 	cs_hsi_free_data_msg(hi->data_tx_msg);
363 }
364 
365 static inline void __cs_hsi_error_pre(struct cs_hsi_iface *hi,
366 					struct hsi_msg *msg, const char *info,
367 					unsigned int *state)
368 {
369 	spin_lock(&hi->lock);
370 	dev_err(&hi->cl->device, "HSI %s error, msg %d, state %u\n",
371 		info, msg->status, *state);
372 }
373 
374 static inline void __cs_hsi_error_post(struct cs_hsi_iface *hi)
375 {
376 	spin_unlock(&hi->lock);
377 }
378 
379 static inline void __cs_hsi_error_read_bits(unsigned int *state)
380 {
381 	*state |= SSI_CHANNEL_STATE_ERROR;
382 	*state &= ~(SSI_CHANNEL_STATE_READING | SSI_CHANNEL_STATE_POLL);
383 }
384 
385 static inline void __cs_hsi_error_write_bits(unsigned int *state)
386 {
387 	*state |= SSI_CHANNEL_STATE_ERROR;
388 	*state &= ~SSI_CHANNEL_STATE_WRITING;
389 }
390 
391 static void cs_hsi_control_read_error(struct cs_hsi_iface *hi,
392 							struct hsi_msg *msg)
393 {
394 	__cs_hsi_error_pre(hi, msg, "control read", &hi->control_state);
395 	cs_release_cmd(msg);
396 	__cs_hsi_error_read_bits(&hi->control_state);
397 	__cs_hsi_error_post(hi);
398 }
399 
400 static void cs_hsi_control_write_error(struct cs_hsi_iface *hi,
401 							struct hsi_msg *msg)
402 {
403 	__cs_hsi_error_pre(hi, msg, "control write", &hi->control_state);
404 	cs_release_cmd(msg);
405 	__cs_hsi_error_write_bits(&hi->control_state);
406 	__cs_hsi_error_post(hi);
407 
408 }
409 
410 static void cs_hsi_data_read_error(struct cs_hsi_iface *hi, struct hsi_msg *msg)
411 {
412 	__cs_hsi_error_pre(hi, msg, "data read", &hi->data_state);
413 	__cs_hsi_error_read_bits(&hi->data_state);
414 	__cs_hsi_error_post(hi);
415 }
416 
417 static void cs_hsi_data_write_error(struct cs_hsi_iface *hi,
418 							struct hsi_msg *msg)
419 {
420 	__cs_hsi_error_pre(hi, msg, "data write", &hi->data_state);
421 	__cs_hsi_error_write_bits(&hi->data_state);
422 	__cs_hsi_error_post(hi);
423 }
424 
425 static void cs_hsi_read_on_control_complete(struct hsi_msg *msg)
426 {
427 	u32 cmd = cs_get_cmd(msg);
428 	struct cs_hsi_iface *hi = msg->context;
429 
430 	spin_lock(&hi->lock);
431 	hi->control_state &= ~SSI_CHANNEL_STATE_READING;
432 	if (msg->status == HSI_STATUS_ERROR) {
433 		dev_err(&hi->cl->device, "Control RX error detected\n");
434 		spin_unlock(&hi->lock);
435 		cs_hsi_control_read_error(hi, msg);
436 		goto out;
437 	}
438 	dev_dbg(&hi->cl->device, "Read on control: %08X\n", cmd);
439 	cs_release_cmd(msg);
440 	if (hi->flags & CS_FEAT_TSTAMP_RX_CTRL) {
441 		struct timespec64 tspec;
442 		struct cs_timestamp *tstamp =
443 			&hi->mmap_cfg->tstamp_rx_ctrl;
444 
445 		ktime_get_ts64(&tspec);
446 
447 		tstamp->tv_sec = (__u32) tspec.tv_sec;
448 		tstamp->tv_nsec = (__u32) tspec.tv_nsec;
449 	}
450 	spin_unlock(&hi->lock);
451 
452 	cs_notify_control(cmd);
453 
454 out:
455 	cs_hsi_read_on_control(hi);
456 }
457 
458 static void cs_hsi_peek_on_control_complete(struct hsi_msg *msg)
459 {
460 	struct cs_hsi_iface *hi = msg->context;
461 	int ret;
462 
463 	if (msg->status == HSI_STATUS_ERROR) {
464 		dev_err(&hi->cl->device, "Control peek RX error detected\n");
465 		cs_hsi_control_read_error(hi, msg);
466 		return;
467 	}
468 
469 	WARN_ON(!(hi->control_state & SSI_CHANNEL_STATE_READING));
470 
471 	dev_dbg(&hi->cl->device, "Peek on control complete, reading\n");
472 	msg->sgt.nents = 1;
473 	msg->complete = cs_hsi_read_on_control_complete;
474 	ret = hsi_async_read(hi->cl, msg);
475 	if (ret)
476 		cs_hsi_control_read_error(hi, msg);
477 }
478 
479 static void cs_hsi_read_on_control(struct cs_hsi_iface *hi)
480 {
481 	struct hsi_msg *msg;
482 	int ret;
483 
484 	spin_lock(&hi->lock);
485 	if (hi->control_state & SSI_CHANNEL_STATE_READING) {
486 		dev_err(&hi->cl->device, "Control read already pending (%d)\n",
487 			hi->control_state);
488 		spin_unlock(&hi->lock);
489 		return;
490 	}
491 	if (hi->control_state & SSI_CHANNEL_STATE_ERROR) {
492 		dev_err(&hi->cl->device, "Control read error (%d)\n",
493 			hi->control_state);
494 		spin_unlock(&hi->lock);
495 		return;
496 	}
497 	hi->control_state |= SSI_CHANNEL_STATE_READING;
498 	dev_dbg(&hi->cl->device, "Issuing RX on control\n");
499 	msg = cs_claim_cmd(hi);
500 	spin_unlock(&hi->lock);
501 
502 	msg->sgt.nents = 0;
503 	msg->complete = cs_hsi_peek_on_control_complete;
504 	ret = hsi_async_read(hi->cl, msg);
505 	if (ret)
506 		cs_hsi_control_read_error(hi, msg);
507 }
508 
509 static void cs_hsi_write_on_control_complete(struct hsi_msg *msg)
510 {
511 	struct cs_hsi_iface *hi = msg->context;
512 	if (msg->status == HSI_STATUS_COMPLETED) {
513 		spin_lock(&hi->lock);
514 		hi->control_state &= ~SSI_CHANNEL_STATE_WRITING;
515 		cs_release_cmd(msg);
516 		spin_unlock(&hi->lock);
517 	} else if (msg->status == HSI_STATUS_ERROR) {
518 		cs_hsi_control_write_error(hi, msg);
519 	} else {
520 		dev_err(&hi->cl->device,
521 			"unexpected status in control write callback %d\n",
522 			msg->status);
523 	}
524 }
525 
526 static int cs_hsi_write_on_control(struct cs_hsi_iface *hi, u32 message)
527 {
528 	struct hsi_msg *msg;
529 	int ret;
530 
531 	spin_lock(&hi->lock);
532 	if (hi->control_state & SSI_CHANNEL_STATE_ERROR) {
533 		spin_unlock(&hi->lock);
534 		return -EIO;
535 	}
536 	if (hi->control_state & SSI_CHANNEL_STATE_WRITING) {
537 		dev_err(&hi->cl->device,
538 			"Write still pending on control channel.\n");
539 		spin_unlock(&hi->lock);
540 		return -EBUSY;
541 	}
542 	hi->control_state |= SSI_CHANNEL_STATE_WRITING;
543 	msg = cs_claim_cmd(hi);
544 	spin_unlock(&hi->lock);
545 
546 	cs_set_cmd(msg, message);
547 	msg->sgt.nents = 1;
548 	msg->complete = cs_hsi_write_on_control_complete;
549 	dev_dbg(&hi->cl->device,
550 		"Sending control message %08X\n", message);
551 	ret = hsi_async_write(hi->cl, msg);
552 	if (ret) {
553 		dev_err(&hi->cl->device,
554 			"async_write failed with %d\n", ret);
555 		cs_hsi_control_write_error(hi, msg);
556 	}
557 
558 	/*
559 	 * Make sure control read is always pending when issuing
560 	 * new control writes. This is needed as the controller
561 	 * may flush our messages if e.g. the peer device reboots
562 	 * unexpectedly (and we cannot directly resubmit a new read from
563 	 * the message destructor; see cs_cmd_destructor()).
564 	 */
565 	if (!(hi->control_state & SSI_CHANNEL_STATE_READING)) {
566 		dev_err(&hi->cl->device, "Restarting control reads\n");
567 		cs_hsi_read_on_control(hi);
568 	}
569 
570 	return 0;
571 }
572 
573 static void cs_hsi_read_on_data_complete(struct hsi_msg *msg)
574 {
575 	struct cs_hsi_iface *hi = msg->context;
576 	u32 payload;
577 
578 	if (unlikely(msg->status == HSI_STATUS_ERROR)) {
579 		cs_hsi_data_read_error(hi, msg);
580 		return;
581 	}
582 
583 	spin_lock(&hi->lock);
584 	WARN_ON(!(hi->data_state & SSI_CHANNEL_STATE_READING));
585 	hi->data_state &= ~SSI_CHANNEL_STATE_READING;
586 	payload = CS_RX_DATA_RECEIVED;
587 	payload |= hi->rx_slot;
588 	hi->rx_slot++;
589 	hi->rx_slot %= hi->rx_ptr_boundary;
590 	/* expose current rx ptr in mmap area */
591 	hi->mmap_cfg->rx_ptr = hi->rx_slot;
592 	if (unlikely(waitqueue_active(&hi->datawait)))
593 		wake_up_interruptible(&hi->datawait);
594 	spin_unlock(&hi->lock);
595 
596 	cs_notify_data(payload, hi->rx_bufs);
597 	cs_hsi_read_on_data(hi);
598 }
599 
600 static void cs_hsi_peek_on_data_complete(struct hsi_msg *msg)
601 {
602 	struct cs_hsi_iface *hi = msg->context;
603 	u32 *address;
604 	int ret;
605 
606 	if (unlikely(msg->status == HSI_STATUS_ERROR)) {
607 		cs_hsi_data_read_error(hi, msg);
608 		return;
609 	}
610 	if (unlikely(hi->iface_state != CS_STATE_CONFIGURED)) {
611 		dev_err(&hi->cl->device, "Data received in invalid state\n");
612 		cs_hsi_data_read_error(hi, msg);
613 		return;
614 	}
615 
616 	spin_lock(&hi->lock);
617 	WARN_ON(!(hi->data_state & SSI_CHANNEL_STATE_POLL));
618 	hi->data_state &= ~SSI_CHANNEL_STATE_POLL;
619 	hi->data_state |= SSI_CHANNEL_STATE_READING;
620 	spin_unlock(&hi->lock);
621 
622 	address = (u32 *)(hi->mmap_base +
623 				hi->rx_offsets[hi->rx_slot % hi->rx_bufs]);
624 	sg_init_one(msg->sgt.sgl, address, hi->buf_size);
625 	msg->sgt.nents = 1;
626 	msg->complete = cs_hsi_read_on_data_complete;
627 	ret = hsi_async_read(hi->cl, msg);
628 	if (ret)
629 		cs_hsi_data_read_error(hi, msg);
630 }
631 
632 /*
633  * Read/write transaction is ongoing. Returns false if in
634  * SSI_CHANNEL_STATE_POLL state.
635  */
636 static inline int cs_state_xfer_active(unsigned int state)
637 {
638 	return (state & SSI_CHANNEL_STATE_WRITING) ||
639 		(state & SSI_CHANNEL_STATE_READING);
640 }
641 
642 /*
643  * No pending read/writes
644  */
645 static inline int cs_state_idle(unsigned int state)
646 {
647 	return !(state & ~SSI_CHANNEL_STATE_ERROR);
648 }
649 
650 static void cs_hsi_read_on_data(struct cs_hsi_iface *hi)
651 {
652 	struct hsi_msg *rxmsg;
653 	int ret;
654 
655 	spin_lock(&hi->lock);
656 	if (hi->data_state &
657 		(SSI_CHANNEL_STATE_READING | SSI_CHANNEL_STATE_POLL)) {
658 		dev_dbg(&hi->cl->device, "Data read already pending (%u)\n",
659 			hi->data_state);
660 		spin_unlock(&hi->lock);
661 		return;
662 	}
663 	hi->data_state |= SSI_CHANNEL_STATE_POLL;
664 	spin_unlock(&hi->lock);
665 
666 	rxmsg = hi->data_rx_msg;
667 	sg_init_one(rxmsg->sgt.sgl, (void *)hi->mmap_base, 0);
668 	rxmsg->sgt.nents = 0;
669 	rxmsg->complete = cs_hsi_peek_on_data_complete;
670 
671 	ret = hsi_async_read(hi->cl, rxmsg);
672 	if (ret)
673 		cs_hsi_data_read_error(hi, rxmsg);
674 }
675 
676 static void cs_hsi_write_on_data_complete(struct hsi_msg *msg)
677 {
678 	struct cs_hsi_iface *hi = msg->context;
679 
680 	if (msg->status == HSI_STATUS_COMPLETED) {
681 		spin_lock(&hi->lock);
682 		hi->data_state &= ~SSI_CHANNEL_STATE_WRITING;
683 		if (unlikely(waitqueue_active(&hi->datawait)))
684 			wake_up_interruptible(&hi->datawait);
685 		spin_unlock(&hi->lock);
686 	} else {
687 		cs_hsi_data_write_error(hi, msg);
688 	}
689 }
690 
691 static int cs_hsi_write_on_data(struct cs_hsi_iface *hi, unsigned int slot)
692 {
693 	u32 *address;
694 	struct hsi_msg *txmsg;
695 	int ret;
696 
697 	spin_lock(&hi->lock);
698 	if (hi->iface_state != CS_STATE_CONFIGURED) {
699 		dev_err(&hi->cl->device, "Not configured, aborting\n");
700 		ret = -EINVAL;
701 		goto error;
702 	}
703 	if (hi->data_state & SSI_CHANNEL_STATE_ERROR) {
704 		dev_err(&hi->cl->device, "HSI error, aborting\n");
705 		ret = -EIO;
706 		goto error;
707 	}
708 	if (hi->data_state & SSI_CHANNEL_STATE_WRITING) {
709 		dev_err(&hi->cl->device, "Write pending on data channel.\n");
710 		ret = -EBUSY;
711 		goto error;
712 	}
713 	hi->data_state |= SSI_CHANNEL_STATE_WRITING;
714 	spin_unlock(&hi->lock);
715 
716 	hi->tx_slot = slot;
717 	address = (u32 *)(hi->mmap_base + hi->tx_offsets[hi->tx_slot]);
718 	txmsg = hi->data_tx_msg;
719 	sg_init_one(txmsg->sgt.sgl, address, hi->buf_size);
720 	txmsg->complete = cs_hsi_write_on_data_complete;
721 	ret = hsi_async_write(hi->cl, txmsg);
722 	if (ret)
723 		cs_hsi_data_write_error(hi, txmsg);
724 
725 	return ret;
726 
727 error:
728 	spin_unlock(&hi->lock);
729 	if (ret == -EIO)
730 		cs_hsi_data_write_error(hi, hi->data_tx_msg);
731 
732 	return ret;
733 }
734 
735 static unsigned int cs_hsi_get_state(struct cs_hsi_iface *hi)
736 {
737 	return hi->iface_state;
738 }
739 
740 static int cs_hsi_command(struct cs_hsi_iface *hi, u32 cmd)
741 {
742 	int ret = 0;
743 
744 	local_bh_disable();
745 	switch (cmd & TARGET_MASK) {
746 	case TARGET_REMOTE:
747 		ret = cs_hsi_write_on_control(hi, cmd);
748 		break;
749 	case TARGET_LOCAL:
750 		if ((cmd & CS_CMD_MASK) == CS_TX_DATA_READY)
751 			ret = cs_hsi_write_on_data(hi, cmd & CS_PARAM_MASK);
752 		else
753 			ret = -EINVAL;
754 		break;
755 	default:
756 		ret = -EINVAL;
757 		break;
758 	}
759 	local_bh_enable();
760 
761 	return ret;
762 }
763 
764 static void cs_hsi_set_wakeline(struct cs_hsi_iface *hi, bool new_state)
765 {
766 	int change = 0;
767 
768 	spin_lock_bh(&hi->lock);
769 	if (hi->wakeline_state != new_state) {
770 		hi->wakeline_state = new_state;
771 		change = 1;
772 		dev_dbg(&hi->cl->device, "setting wake line to %d (%p)\n",
773 			new_state, hi->cl);
774 	}
775 	spin_unlock_bh(&hi->lock);
776 
777 	if (change) {
778 		if (new_state)
779 			ssip_slave_start_tx(hi->master);
780 		else
781 			ssip_slave_stop_tx(hi->master);
782 	}
783 
784 	dev_dbg(&hi->cl->device, "wake line set to %d (%p)\n",
785 		new_state, hi->cl);
786 }
787 
788 static void set_buffer_sizes(struct cs_hsi_iface *hi, int rx_bufs, int tx_bufs)
789 {
790 	hi->rx_bufs = rx_bufs;
791 	hi->tx_bufs = tx_bufs;
792 	hi->mmap_cfg->rx_bufs = rx_bufs;
793 	hi->mmap_cfg->tx_bufs = tx_bufs;
794 
795 	if (hi->flags & CS_FEAT_ROLLING_RX_COUNTER) {
796 		/*
797 		 * For more robust overrun detection, let the rx
798 		 * pointer run in range 0..'boundary-1'. Boundary
799 		 * is a multiple of rx_bufs, and limited in max size
800 		 * by RX_PTR_MAX_SHIFT to allow for fast ptr-diff
801 		 * calculation.
802 		 */
803 		hi->rx_ptr_boundary = (rx_bufs << RX_PTR_BOUNDARY_SHIFT);
804 		hi->mmap_cfg->rx_ptr_boundary = hi->rx_ptr_boundary;
805 	} else {
806 		hi->rx_ptr_boundary = hi->rx_bufs;
807 	}
808 }
809 
810 static int check_buf_params(struct cs_hsi_iface *hi,
811 					const struct cs_buffer_config *buf_cfg)
812 {
813 	size_t buf_size_aligned = L1_CACHE_ALIGN(buf_cfg->buf_size) *
814 					(buf_cfg->rx_bufs + buf_cfg->tx_bufs);
815 	size_t ctrl_size_aligned = L1_CACHE_ALIGN(sizeof(*hi->mmap_cfg));
816 	int r = 0;
817 
818 	if (buf_cfg->rx_bufs > CS_MAX_BUFFERS ||
819 					buf_cfg->tx_bufs > CS_MAX_BUFFERS) {
820 		r = -EINVAL;
821 	} else if ((buf_size_aligned + ctrl_size_aligned) >= hi->mmap_size) {
822 		dev_err(&hi->cl->device, "No space for the requested buffer "
823 			"configuration\n");
824 		r = -ENOBUFS;
825 	}
826 
827 	return r;
828 }
829 
830 /*
831  * Block until pending data transfers have completed.
832  */
833 static int cs_hsi_data_sync(struct cs_hsi_iface *hi)
834 {
835 	int r = 0;
836 
837 	spin_lock_bh(&hi->lock);
838 
839 	if (!cs_state_xfer_active(hi->data_state)) {
840 		dev_dbg(&hi->cl->device, "hsi_data_sync break, idle\n");
841 		goto out;
842 	}
843 
844 	for (;;) {
845 		int s;
846 		DEFINE_WAIT(wait);
847 		if (!cs_state_xfer_active(hi->data_state))
848 			goto out;
849 		if (signal_pending(current)) {
850 			r = -ERESTARTSYS;
851 			goto out;
852 		}
853 		/*
854 		 * prepare_to_wait must be called with hi->lock held
855 		 * so that callbacks can check for waitqueue_active()
856 		 */
857 		prepare_to_wait(&hi->datawait, &wait, TASK_INTERRUPTIBLE);
858 		spin_unlock_bh(&hi->lock);
859 		s = schedule_timeout(
860 			msecs_to_jiffies(CS_HSI_TRANSFER_TIMEOUT_MS));
861 		spin_lock_bh(&hi->lock);
862 		finish_wait(&hi->datawait, &wait);
863 		if (!s) {
864 			dev_dbg(&hi->cl->device,
865 				"hsi_data_sync timeout after %d ms\n",
866 				CS_HSI_TRANSFER_TIMEOUT_MS);
867 			r = -EIO;
868 			goto out;
869 		}
870 	}
871 
872 out:
873 	spin_unlock_bh(&hi->lock);
874 	dev_dbg(&hi->cl->device, "hsi_data_sync done with res %d\n", r);
875 
876 	return r;
877 }
878 
879 static void cs_hsi_data_enable(struct cs_hsi_iface *hi,
880 					struct cs_buffer_config *buf_cfg)
881 {
882 	unsigned int data_start, i;
883 
884 	BUG_ON(hi->buf_size == 0);
885 
886 	set_buffer_sizes(hi, buf_cfg->rx_bufs, buf_cfg->tx_bufs);
887 
888 	hi->slot_size = L1_CACHE_ALIGN(hi->buf_size);
889 	dev_dbg(&hi->cl->device,
890 			"setting slot size to %u, buf size %u, align %u\n",
891 			hi->slot_size, hi->buf_size, L1_CACHE_BYTES);
892 
893 	data_start = L1_CACHE_ALIGN(sizeof(*hi->mmap_cfg));
894 	dev_dbg(&hi->cl->device,
895 			"setting data start at %u, cfg block %u, align %u\n",
896 			data_start, sizeof(*hi->mmap_cfg), L1_CACHE_BYTES);
897 
898 	for (i = 0; i < hi->mmap_cfg->rx_bufs; i++) {
899 		hi->rx_offsets[i] = data_start + i * hi->slot_size;
900 		hi->mmap_cfg->rx_offsets[i] = hi->rx_offsets[i];
901 		dev_dbg(&hi->cl->device, "DL buf #%u at %u\n",
902 					i, hi->rx_offsets[i]);
903 	}
904 	for (i = 0; i < hi->mmap_cfg->tx_bufs; i++) {
905 		hi->tx_offsets[i] = data_start +
906 			(i + hi->mmap_cfg->rx_bufs) * hi->slot_size;
907 		hi->mmap_cfg->tx_offsets[i] = hi->tx_offsets[i];
908 		dev_dbg(&hi->cl->device, "UL buf #%u at %u\n",
909 					i, hi->rx_offsets[i]);
910 	}
911 
912 	hi->iface_state = CS_STATE_CONFIGURED;
913 }
914 
915 static void cs_hsi_data_disable(struct cs_hsi_iface *hi, int old_state)
916 {
917 	if (old_state == CS_STATE_CONFIGURED) {
918 		dev_dbg(&hi->cl->device,
919 			"closing data channel with slot size 0\n");
920 		hi->iface_state = CS_STATE_OPENED;
921 	}
922 }
923 
924 static int cs_hsi_buf_config(struct cs_hsi_iface *hi,
925 					struct cs_buffer_config *buf_cfg)
926 {
927 	int r = 0;
928 	unsigned int old_state = hi->iface_state;
929 
930 	spin_lock_bh(&hi->lock);
931 	/* Prevent new transactions during buffer reconfig */
932 	if (old_state == CS_STATE_CONFIGURED)
933 		hi->iface_state = CS_STATE_OPENED;
934 	spin_unlock_bh(&hi->lock);
935 
936 	/*
937 	 * make sure that no non-zero data reads are ongoing before
938 	 * proceeding to change the buffer layout
939 	 */
940 	r = cs_hsi_data_sync(hi);
941 	if (r < 0)
942 		return r;
943 
944 	WARN_ON(cs_state_xfer_active(hi->data_state));
945 
946 	spin_lock_bh(&hi->lock);
947 	r = check_buf_params(hi, buf_cfg);
948 	if (r < 0)
949 		goto error;
950 
951 	hi->buf_size = buf_cfg->buf_size;
952 	hi->mmap_cfg->buf_size = hi->buf_size;
953 	hi->flags = buf_cfg->flags;
954 
955 	hi->rx_slot = 0;
956 	hi->tx_slot = 0;
957 	hi->slot_size = 0;
958 
959 	if (hi->buf_size)
960 		cs_hsi_data_enable(hi, buf_cfg);
961 	else
962 		cs_hsi_data_disable(hi, old_state);
963 
964 	spin_unlock_bh(&hi->lock);
965 
966 	if (old_state != hi->iface_state) {
967 		if (hi->iface_state == CS_STATE_CONFIGURED) {
968 			cpu_latency_qos_add_request(&hi->pm_qos_req,
969 				CS_QOS_LATENCY_FOR_DATA_USEC);
970 			local_bh_disable();
971 			cs_hsi_read_on_data(hi);
972 			local_bh_enable();
973 		} else if (old_state == CS_STATE_CONFIGURED) {
974 			cpu_latency_qos_remove_request(&hi->pm_qos_req);
975 		}
976 	}
977 	return r;
978 
979 error:
980 	spin_unlock_bh(&hi->lock);
981 	return r;
982 }
983 
984 static int cs_hsi_start(struct cs_hsi_iface **hi, struct hsi_client *cl,
985 			unsigned long mmap_base, unsigned long mmap_size)
986 {
987 	int err = 0;
988 	struct cs_hsi_iface *hsi_if = kzalloc(sizeof(*hsi_if), GFP_KERNEL);
989 
990 	dev_dbg(&cl->device, "cs_hsi_start\n");
991 
992 	if (!hsi_if) {
993 		err = -ENOMEM;
994 		goto leave0;
995 	}
996 	spin_lock_init(&hsi_if->lock);
997 	hsi_if->cl = cl;
998 	hsi_if->iface_state = CS_STATE_CLOSED;
999 	hsi_if->mmap_cfg = (struct cs_mmap_config_block *)mmap_base;
1000 	hsi_if->mmap_base = mmap_base;
1001 	hsi_if->mmap_size = mmap_size;
1002 	memset(hsi_if->mmap_cfg, 0, sizeof(*hsi_if->mmap_cfg));
1003 	init_waitqueue_head(&hsi_if->datawait);
1004 	err = cs_alloc_cmds(hsi_if);
1005 	if (err < 0) {
1006 		dev_err(&cl->device, "Unable to alloc HSI messages\n");
1007 		goto leave1;
1008 	}
1009 	err = cs_hsi_alloc_data(hsi_if);
1010 	if (err < 0) {
1011 		dev_err(&cl->device, "Unable to alloc HSI messages for data\n");
1012 		goto leave2;
1013 	}
1014 	err = hsi_claim_port(cl, 1);
1015 	if (err < 0) {
1016 		dev_err(&cl->device,
1017 				"Could not open, HSI port already claimed\n");
1018 		goto leave3;
1019 	}
1020 	hsi_if->master = ssip_slave_get_master(cl);
1021 	if (IS_ERR(hsi_if->master)) {
1022 		err = PTR_ERR(hsi_if->master);
1023 		dev_err(&cl->device, "Could not get HSI master client\n");
1024 		goto leave4;
1025 	}
1026 	if (!ssip_slave_running(hsi_if->master)) {
1027 		err = -ENODEV;
1028 		dev_err(&cl->device,
1029 				"HSI port not initialized\n");
1030 		goto leave4;
1031 	}
1032 
1033 	hsi_if->iface_state = CS_STATE_OPENED;
1034 	local_bh_disable();
1035 	cs_hsi_read_on_control(hsi_if);
1036 	local_bh_enable();
1037 
1038 	dev_dbg(&cl->device, "cs_hsi_start...done\n");
1039 
1040 	BUG_ON(!hi);
1041 	*hi = hsi_if;
1042 
1043 	return 0;
1044 
1045 leave4:
1046 	hsi_release_port(cl);
1047 leave3:
1048 	cs_hsi_free_data(hsi_if);
1049 leave2:
1050 	cs_free_cmds(hsi_if);
1051 leave1:
1052 	kfree(hsi_if);
1053 leave0:
1054 	dev_dbg(&cl->device, "cs_hsi_start...done/error\n\n");
1055 
1056 	return err;
1057 }
1058 
1059 static void cs_hsi_stop(struct cs_hsi_iface *hi)
1060 {
1061 	dev_dbg(&hi->cl->device, "cs_hsi_stop\n");
1062 	cs_hsi_set_wakeline(hi, 0);
1063 	ssip_slave_put_master(hi->master);
1064 
1065 	/* hsi_release_port() needs to be called with CS_STATE_CLOSED */
1066 	hi->iface_state = CS_STATE_CLOSED;
1067 	hsi_release_port(hi->cl);
1068 
1069 	/*
1070 	 * hsi_release_port() should flush out all the pending
1071 	 * messages, so cs_state_idle() should be true for both
1072 	 * control and data channels.
1073 	 */
1074 	WARN_ON(!cs_state_idle(hi->control_state));
1075 	WARN_ON(!cs_state_idle(hi->data_state));
1076 
1077 	if (cpu_latency_qos_request_active(&hi->pm_qos_req))
1078 		cpu_latency_qos_remove_request(&hi->pm_qos_req);
1079 
1080 	spin_lock_bh(&hi->lock);
1081 	cs_hsi_free_data(hi);
1082 	cs_free_cmds(hi);
1083 	spin_unlock_bh(&hi->lock);
1084 	kfree(hi);
1085 }
1086 
1087 static vm_fault_t cs_char_vma_fault(struct vm_fault *vmf)
1088 {
1089 	struct cs_char *csdata = vmf->vma->vm_private_data;
1090 	struct page *page;
1091 
1092 	page = virt_to_page((void *)csdata->mmap_base);
1093 	get_page(page);
1094 	vmf->page = page;
1095 
1096 	return 0;
1097 }
1098 
1099 static const struct vm_operations_struct cs_char_vm_ops = {
1100 	.fault	= cs_char_vma_fault,
1101 };
1102 
1103 static int cs_char_fasync(int fd, struct file *file, int on)
1104 {
1105 	struct cs_char *csdata = file->private_data;
1106 
1107 	if (fasync_helper(fd, file, on, &csdata->async_queue) < 0)
1108 		return -EIO;
1109 
1110 	return 0;
1111 }
1112 
1113 static __poll_t cs_char_poll(struct file *file, poll_table *wait)
1114 {
1115 	struct cs_char *csdata = file->private_data;
1116 	__poll_t ret = 0;
1117 
1118 	poll_wait(file, &cs_char_data.wait, wait);
1119 	spin_lock_bh(&csdata->lock);
1120 	if (!list_empty(&csdata->chardev_queue))
1121 		ret = EPOLLIN | EPOLLRDNORM;
1122 	else if (!list_empty(&csdata->dataind_queue))
1123 		ret = EPOLLIN | EPOLLRDNORM;
1124 	spin_unlock_bh(&csdata->lock);
1125 
1126 	return ret;
1127 }
1128 
1129 static ssize_t cs_char_read(struct file *file, char __user *buf, size_t count,
1130 								loff_t *unused)
1131 {
1132 	struct cs_char *csdata = file->private_data;
1133 	u32 data;
1134 	ssize_t retval;
1135 
1136 	if (count < sizeof(data))
1137 		return -EINVAL;
1138 
1139 	for (;;) {
1140 		DEFINE_WAIT(wait);
1141 
1142 		spin_lock_bh(&csdata->lock);
1143 		if (!list_empty(&csdata->chardev_queue)) {
1144 			data = cs_pop_entry(&csdata->chardev_queue);
1145 		} else if (!list_empty(&csdata->dataind_queue)) {
1146 			data = cs_pop_entry(&csdata->dataind_queue);
1147 			csdata->dataind_pending--;
1148 		} else {
1149 			data = 0;
1150 		}
1151 		spin_unlock_bh(&csdata->lock);
1152 
1153 		if (data)
1154 			break;
1155 		if (file->f_flags & O_NONBLOCK) {
1156 			retval = -EAGAIN;
1157 			goto out;
1158 		} else if (signal_pending(current)) {
1159 			retval = -ERESTARTSYS;
1160 			goto out;
1161 		}
1162 		prepare_to_wait_exclusive(&csdata->wait, &wait,
1163 						TASK_INTERRUPTIBLE);
1164 		schedule();
1165 		finish_wait(&csdata->wait, &wait);
1166 	}
1167 
1168 	retval = put_user(data, (u32 __user *)buf);
1169 	if (!retval)
1170 		retval = sizeof(data);
1171 
1172 out:
1173 	return retval;
1174 }
1175 
1176 static ssize_t cs_char_write(struct file *file, const char __user *buf,
1177 						size_t count, loff_t *unused)
1178 {
1179 	struct cs_char *csdata = file->private_data;
1180 	u32 data;
1181 	int err;
1182 	ssize_t	retval;
1183 
1184 	if (count < sizeof(data))
1185 		return -EINVAL;
1186 
1187 	if (get_user(data, (u32 __user *)buf))
1188 		retval = -EFAULT;
1189 	else
1190 		retval = count;
1191 
1192 	err = cs_hsi_command(csdata->hi, data);
1193 	if (err < 0)
1194 		retval = err;
1195 
1196 	return retval;
1197 }
1198 
1199 static long cs_char_ioctl(struct file *file, unsigned int cmd,
1200 				unsigned long arg)
1201 {
1202 	struct cs_char *csdata = file->private_data;
1203 	int r = 0;
1204 
1205 	switch (cmd) {
1206 	case CS_GET_STATE: {
1207 		unsigned int state;
1208 
1209 		state = cs_hsi_get_state(csdata->hi);
1210 		if (copy_to_user((void __user *)arg, &state, sizeof(state)))
1211 			r = -EFAULT;
1212 
1213 		break;
1214 	}
1215 	case CS_SET_WAKELINE: {
1216 		unsigned int state;
1217 
1218 		if (copy_from_user(&state, (void __user *)arg, sizeof(state))) {
1219 			r = -EFAULT;
1220 			break;
1221 		}
1222 
1223 		if (state > 1) {
1224 			r = -EINVAL;
1225 			break;
1226 		}
1227 
1228 		cs_hsi_set_wakeline(csdata->hi, !!state);
1229 
1230 		break;
1231 	}
1232 	case CS_GET_IF_VERSION: {
1233 		unsigned int ifver = CS_IF_VERSION;
1234 
1235 		if (copy_to_user((void __user *)arg, &ifver, sizeof(ifver)))
1236 			r = -EFAULT;
1237 
1238 		break;
1239 	}
1240 	case CS_CONFIG_BUFS: {
1241 		struct cs_buffer_config buf_cfg;
1242 
1243 		if (copy_from_user(&buf_cfg, (void __user *)arg,
1244 							sizeof(buf_cfg)))
1245 			r = -EFAULT;
1246 		else
1247 			r = cs_hsi_buf_config(csdata->hi, &buf_cfg);
1248 
1249 		break;
1250 	}
1251 	default:
1252 		r = -ENOTTY;
1253 		break;
1254 	}
1255 
1256 	return r;
1257 }
1258 
1259 static int cs_char_mmap(struct file *file, struct vm_area_struct *vma)
1260 {
1261 	if (vma->vm_end < vma->vm_start)
1262 		return -EINVAL;
1263 
1264 	if (vma_pages(vma) != 1)
1265 		return -EINVAL;
1266 
1267 	vm_flags_set(vma, VM_IO | VM_DONTDUMP | VM_DONTEXPAND);
1268 	vma->vm_ops = &cs_char_vm_ops;
1269 	vma->vm_private_data = file->private_data;
1270 
1271 	return 0;
1272 }
1273 
1274 static int cs_char_open(struct inode *unused, struct file *file)
1275 {
1276 	int ret = 0;
1277 	unsigned long p;
1278 
1279 	spin_lock_bh(&cs_char_data.lock);
1280 	if (cs_char_data.opened) {
1281 		ret = -EBUSY;
1282 		spin_unlock_bh(&cs_char_data.lock);
1283 		goto out1;
1284 	}
1285 	cs_char_data.opened = 1;
1286 	cs_char_data.dataind_pending = 0;
1287 	spin_unlock_bh(&cs_char_data.lock);
1288 
1289 	p = get_zeroed_page(GFP_KERNEL);
1290 	if (!p) {
1291 		ret = -ENOMEM;
1292 		goto out2;
1293 	}
1294 
1295 	ret = cs_hsi_start(&cs_char_data.hi, cs_char_data.cl, p, CS_MMAP_SIZE);
1296 	if (ret) {
1297 		dev_err(&cs_char_data.cl->device, "Unable to initialize HSI\n");
1298 		goto out3;
1299 	}
1300 
1301 	/* these are only used in release so lock not needed */
1302 	cs_char_data.mmap_base = p;
1303 	cs_char_data.mmap_size = CS_MMAP_SIZE;
1304 
1305 	file->private_data = &cs_char_data;
1306 
1307 	return 0;
1308 
1309 out3:
1310 	free_page(p);
1311 out2:
1312 	spin_lock_bh(&cs_char_data.lock);
1313 	cs_char_data.opened = 0;
1314 	spin_unlock_bh(&cs_char_data.lock);
1315 out1:
1316 	return ret;
1317 }
1318 
1319 static void cs_free_char_queue(struct list_head *head)
1320 {
1321 	struct char_queue *entry;
1322 	struct list_head *cursor, *next;
1323 
1324 	if (!list_empty(head)) {
1325 		list_for_each_safe(cursor, next, head) {
1326 			entry = list_entry(cursor, struct char_queue, list);
1327 			list_del(&entry->list);
1328 			kfree(entry);
1329 		}
1330 	}
1331 
1332 }
1333 
1334 static int cs_char_release(struct inode *unused, struct file *file)
1335 {
1336 	struct cs_char *csdata = file->private_data;
1337 
1338 	cs_hsi_stop(csdata->hi);
1339 	spin_lock_bh(&csdata->lock);
1340 	csdata->hi = NULL;
1341 	free_page(csdata->mmap_base);
1342 	cs_free_char_queue(&csdata->chardev_queue);
1343 	cs_free_char_queue(&csdata->dataind_queue);
1344 	csdata->opened = 0;
1345 	spin_unlock_bh(&csdata->lock);
1346 
1347 	return 0;
1348 }
1349 
1350 static const struct file_operations cs_char_fops = {
1351 	.owner		= THIS_MODULE,
1352 	.read		= cs_char_read,
1353 	.write		= cs_char_write,
1354 	.poll		= cs_char_poll,
1355 	.unlocked_ioctl	= cs_char_ioctl,
1356 	.mmap		= cs_char_mmap,
1357 	.open		= cs_char_open,
1358 	.release	= cs_char_release,
1359 	.fasync		= cs_char_fasync,
1360 };
1361 
1362 static struct miscdevice cs_char_miscdev = {
1363 	.minor	= MISC_DYNAMIC_MINOR,
1364 	.name	= "cmt_speech",
1365 	.fops	= &cs_char_fops
1366 };
1367 
1368 static int cs_hsi_client_probe(struct device *dev)
1369 {
1370 	int err = 0;
1371 	struct hsi_client *cl = to_hsi_client(dev);
1372 
1373 	dev_dbg(dev, "hsi_client_probe\n");
1374 	init_waitqueue_head(&cs_char_data.wait);
1375 	spin_lock_init(&cs_char_data.lock);
1376 	cs_char_data.opened = 0;
1377 	cs_char_data.cl = cl;
1378 	cs_char_data.hi = NULL;
1379 	INIT_LIST_HEAD(&cs_char_data.chardev_queue);
1380 	INIT_LIST_HEAD(&cs_char_data.dataind_queue);
1381 
1382 	cs_char_data.channel_id_cmd = hsi_get_channel_id_by_name(cl,
1383 		"speech-control");
1384 	if (cs_char_data.channel_id_cmd < 0) {
1385 		err = cs_char_data.channel_id_cmd;
1386 		dev_err(dev, "Could not get cmd channel (%d)\n", err);
1387 		return err;
1388 	}
1389 
1390 	cs_char_data.channel_id_data = hsi_get_channel_id_by_name(cl,
1391 		"speech-data");
1392 	if (cs_char_data.channel_id_data < 0) {
1393 		err = cs_char_data.channel_id_data;
1394 		dev_err(dev, "Could not get data channel (%d)\n", err);
1395 		return err;
1396 	}
1397 
1398 	err = misc_register(&cs_char_miscdev);
1399 	if (err)
1400 		dev_err(dev, "Failed to register: %d\n", err);
1401 
1402 	return err;
1403 }
1404 
1405 static int cs_hsi_client_remove(struct device *dev)
1406 {
1407 	struct cs_hsi_iface *hi;
1408 
1409 	dev_dbg(dev, "hsi_client_remove\n");
1410 	misc_deregister(&cs_char_miscdev);
1411 	spin_lock_bh(&cs_char_data.lock);
1412 	hi = cs_char_data.hi;
1413 	cs_char_data.hi = NULL;
1414 	spin_unlock_bh(&cs_char_data.lock);
1415 	if (hi)
1416 		cs_hsi_stop(hi);
1417 
1418 	return 0;
1419 }
1420 
1421 static struct hsi_client_driver cs_hsi_driver = {
1422 	.driver = {
1423 		.name	= "cmt-speech",
1424 		.owner	= THIS_MODULE,
1425 		.probe	= cs_hsi_client_probe,
1426 		.remove	= cs_hsi_client_remove,
1427 	},
1428 };
1429 
1430 static int __init cs_char_init(void)
1431 {
1432 	pr_info("CMT speech driver added\n");
1433 	return hsi_register_client_driver(&cs_hsi_driver);
1434 }
1435 module_init(cs_char_init);
1436 
1437 static void __exit cs_char_exit(void)
1438 {
1439 	hsi_unregister_client_driver(&cs_hsi_driver);
1440 	pr_info("CMT speech driver removed\n");
1441 }
1442 module_exit(cs_char_exit);
1443 
1444 MODULE_ALIAS("hsi:cmt-speech");
1445 MODULE_AUTHOR("Kai Vehmanen <kai.vehmanen@nokia.com>");
1446 MODULE_AUTHOR("Peter Ujfalusi <peter.ujfalusi@nokia.com>");
1447 MODULE_DESCRIPTION("CMT speech driver");
1448 MODULE_LICENSE("GPL v2");
1449