1 /*
2  * ssi_protocol.c
3  *
4  * Implementation of the SSI McSAAB improved protocol.
5  *
6  * Copyright (C) 2010 Nokia Corporation. All rights reserved.
7  * Copyright (C) 2013 Sebastian Reichel <sre@kernel.org>
8  *
9  * Contact: Carlos Chinea <carlos.chinea@nokia.com>
10  *
11  * This program is free software; you can redistribute it and/or
12  * modify it under the terms of the GNU General Public License
13  * version 2 as published by the Free Software Foundation.
14  *
15  * This program is distributed in the hope that it will be useful, but
16  * WITHOUT ANY WARRANTY; without even the implied warranty of
17  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
18  * General Public License for more details.
19  *
20  * You should have received a copy of the GNU General Public License
21  * along with this program; if not, write to the Free Software
22  * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
23  * 02110-1301 USA
24  */
25 
26 #include <linux/atomic.h>
27 #include <linux/clk.h>
28 #include <linux/device.h>
29 #include <linux/err.h>
30 #include <linux/gpio.h>
31 #include <linux/if_ether.h>
32 #include <linux/if_arp.h>
33 #include <linux/if_phonet.h>
34 #include <linux/init.h>
35 #include <linux/irq.h>
36 #include <linux/list.h>
37 #include <linux/module.h>
38 #include <linux/netdevice.h>
39 #include <linux/notifier.h>
40 #include <linux/scatterlist.h>
41 #include <linux/skbuff.h>
42 #include <linux/slab.h>
43 #include <linux/spinlock.h>
44 #include <linux/timer.h>
45 #include <linux/hsi/hsi.h>
46 #include <linux/hsi/ssi_protocol.h>
47 
48 void ssi_waketest(struct hsi_client *cl, unsigned int enable);
49 
50 #define SSIP_TXQUEUE_LEN	100
51 #define SSIP_MAX_MTU		65535
52 #define SSIP_DEFAULT_MTU	4000
53 #define PN_MEDIA_SOS		21
54 #define SSIP_MIN_PN_HDR		6	/* FIXME: Revisit */
55 #define SSIP_WDTOUT		2000	/* FIXME: has to be 500 msecs */
56 #define SSIP_KATOUT		15	/* 15 msecs */
57 #define SSIP_MAX_CMDS		5 /* Number of pre-allocated commands buffers */
58 #define SSIP_BYTES_TO_FRAMES(x) ((((x) - 1) >> 2) + 1)
59 #define SSIP_CMT_LOADER_SYNC	0x11223344
60 /*
61  * SSI protocol command definitions
62  */
63 #define SSIP_COMMAND(data)	((data) >> 28)
64 #define SSIP_PAYLOAD(data)	((data) & 0xfffffff)
65 /* Commands */
66 #define SSIP_SW_BREAK		0
67 #define SSIP_BOOTINFO_REQ	1
68 #define SSIP_BOOTINFO_RESP	2
69 #define SSIP_WAKETEST_RESULT	3
70 #define SSIP_START_TRANS	4
71 #define SSIP_READY		5
72 /* Payloads */
73 #define SSIP_DATA_VERSION(data)	((data) & 0xff)
74 #define SSIP_LOCAL_VERID	1
75 #define SSIP_WAKETEST_OK	0
76 #define SSIP_WAKETEST_FAILED	1
77 #define SSIP_PDU_LENGTH(data)	(((data) >> 8) & 0xffff)
78 #define SSIP_MSG_ID(data)	((data) & 0xff)
79 /* Generic Command */
80 #define SSIP_CMD(cmd, payload)	(((cmd) << 28) | ((payload) & 0xfffffff))
81 /* Commands for the control channel */
82 #define SSIP_BOOTINFO_REQ_CMD(ver) \
83 		SSIP_CMD(SSIP_BOOTINFO_REQ, SSIP_DATA_VERSION(ver))
84 #define SSIP_BOOTINFO_RESP_CMD(ver) \
85 		SSIP_CMD(SSIP_BOOTINFO_RESP, SSIP_DATA_VERSION(ver))
86 #define SSIP_START_TRANS_CMD(pdulen, id) \
87 		SSIP_CMD(SSIP_START_TRANS, (((pdulen) << 8) | SSIP_MSG_ID(id)))
88 #define SSIP_READY_CMD		SSIP_CMD(SSIP_READY, 0)
89 #define SSIP_SWBREAK_CMD	SSIP_CMD(SSIP_SW_BREAK, 0)
90 
91 #define SSIP_WAKETEST_FLAG 0
92 
93 /* Main state machine states */
94 enum {
95 	INIT,
96 	HANDSHAKE,
97 	ACTIVE,
98 };
99 
100 /* Send state machine states */
101 enum {
102 	SEND_IDLE,
103 	WAIT4READY,
104 	SEND_READY,
105 	SENDING,
106 	SENDING_SWBREAK,
107 };
108 
109 /* Receive state machine states */
110 enum {
111 	RECV_IDLE,
112 	RECV_READY,
113 	RECEIVING,
114 };
115 
116 /**
117  * struct ssi_protocol - SSI protocol (McSAAB) data
118  * @main_state: Main state machine
119  * @send_state: TX state machine
120  * @recv_state: RX state machine
121  * @flags: Flags, currently only used to follow wake line test
122  * @rxid: RX data id
123  * @txid: TX data id
124  * @txqueue_len: TX queue length
125  * @tx_wd: TX watchdog
126  * @rx_wd: RX watchdog
127  * @keep_alive: Workaround for SSI HW bug
128  * @lock: To serialize access to this struct
129  * @netdev: Phonet network device
130  * @txqueue: TX data queue
131  * @cmdqueue: Queue of free commands
132  * @cl: HSI client own reference
133  * @link: Link for ssip_list
134  * @tx_usecount: Refcount to keep track the slaves that use the wake line
135  * @channel_id_cmd: HSI channel id for command stream
136  * @channel_id_data: HSI channel id for data stream
137  */
138 struct ssi_protocol {
139 	unsigned int		main_state;
140 	unsigned int		send_state;
141 	unsigned int		recv_state;
142 	unsigned long		flags;
143 	u8			rxid;
144 	u8			txid;
145 	unsigned int		txqueue_len;
146 	struct timer_list	tx_wd;
147 	struct timer_list	rx_wd;
148 	struct timer_list	keep_alive; /* wake-up workaround */
149 	spinlock_t		lock;
150 	struct net_device	*netdev;
151 	struct list_head	txqueue;
152 	struct list_head	cmdqueue;
153 	struct work_struct	work;
154 	struct hsi_client	*cl;
155 	struct list_head	link;
156 	atomic_t		tx_usecnt;
157 	int			channel_id_cmd;
158 	int			channel_id_data;
159 };
160 
161 /* List of ssi protocol instances */
162 static LIST_HEAD(ssip_list);
163 
164 static void ssip_rxcmd_complete(struct hsi_msg *msg);
165 
166 static inline void ssip_set_cmd(struct hsi_msg *msg, u32 cmd)
167 {
168 	u32 *data;
169 
170 	data = sg_virt(msg->sgt.sgl);
171 	*data = cmd;
172 }
173 
174 static inline u32 ssip_get_cmd(struct hsi_msg *msg)
175 {
176 	u32 *data;
177 
178 	data = sg_virt(msg->sgt.sgl);
179 
180 	return *data;
181 }
182 
183 static void ssip_skb_to_msg(struct sk_buff *skb, struct hsi_msg *msg)
184 {
185 	skb_frag_t *frag;
186 	struct scatterlist *sg;
187 	int i;
188 
189 	BUG_ON(msg->sgt.nents != (unsigned int)(skb_shinfo(skb)->nr_frags + 1));
190 
191 	sg = msg->sgt.sgl;
192 	sg_set_buf(sg, skb->data, skb_headlen(skb));
193 	for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
194 		sg = sg_next(sg);
195 		BUG_ON(!sg);
196 		frag = &skb_shinfo(skb)->frags[i];
197 		sg_set_page(sg, frag->page.p, frag->size, frag->page_offset);
198 	}
199 }
200 
201 static void ssip_free_data(struct hsi_msg *msg)
202 {
203 	struct sk_buff *skb;
204 
205 	skb = msg->context;
206 	pr_debug("free data: msg %p context %p skb %p\n", msg, msg->context,
207 								skb);
208 	msg->destructor = NULL;
209 	dev_kfree_skb(skb);
210 	hsi_free_msg(msg);
211 }
212 
213 static struct hsi_msg *ssip_alloc_data(struct ssi_protocol *ssi,
214 					struct sk_buff *skb, gfp_t flags)
215 {
216 	struct hsi_msg *msg;
217 
218 	msg = hsi_alloc_msg(skb_shinfo(skb)->nr_frags + 1, flags);
219 	if (!msg)
220 		return NULL;
221 	ssip_skb_to_msg(skb, msg);
222 	msg->destructor = ssip_free_data;
223 	msg->channel = ssi->channel_id_data;
224 	msg->context = skb;
225 
226 	return msg;
227 }
228 
229 static inline void ssip_release_cmd(struct hsi_msg *msg)
230 {
231 	struct ssi_protocol *ssi = hsi_client_drvdata(msg->cl);
232 
233 	dev_dbg(&msg->cl->device, "Release cmd 0x%08x\n", ssip_get_cmd(msg));
234 	spin_lock_bh(&ssi->lock);
235 	list_add_tail(&msg->link, &ssi->cmdqueue);
236 	spin_unlock_bh(&ssi->lock);
237 }
238 
239 static struct hsi_msg *ssip_claim_cmd(struct ssi_protocol *ssi)
240 {
241 	struct hsi_msg *msg;
242 
243 	BUG_ON(list_empty(&ssi->cmdqueue));
244 
245 	spin_lock_bh(&ssi->lock);
246 	msg = list_first_entry(&ssi->cmdqueue, struct hsi_msg, link);
247 	list_del(&msg->link);
248 	spin_unlock_bh(&ssi->lock);
249 	msg->destructor = ssip_release_cmd;
250 
251 	return msg;
252 }
253 
254 static void ssip_free_cmds(struct ssi_protocol *ssi)
255 {
256 	struct hsi_msg *msg, *tmp;
257 
258 	list_for_each_entry_safe(msg, tmp, &ssi->cmdqueue, link) {
259 		list_del(&msg->link);
260 		msg->destructor = NULL;
261 		kfree(sg_virt(msg->sgt.sgl));
262 		hsi_free_msg(msg);
263 	}
264 }
265 
266 static int ssip_alloc_cmds(struct ssi_protocol *ssi)
267 {
268 	struct hsi_msg *msg;
269 	u32 *buf;
270 	unsigned int i;
271 
272 	for (i = 0; i < SSIP_MAX_CMDS; i++) {
273 		msg = hsi_alloc_msg(1, GFP_KERNEL);
274 		if (!msg)
275 			goto out;
276 		buf = kmalloc(sizeof(*buf), GFP_KERNEL);
277 		if (!buf) {
278 			hsi_free_msg(msg);
279 			goto out;
280 		}
281 		sg_init_one(msg->sgt.sgl, buf, sizeof(*buf));
282 		msg->channel = ssi->channel_id_cmd;
283 		list_add_tail(&msg->link, &ssi->cmdqueue);
284 	}
285 
286 	return 0;
287 out:
288 	ssip_free_cmds(ssi);
289 
290 	return -ENOMEM;
291 }
292 
293 static void ssip_set_rxstate(struct ssi_protocol *ssi, unsigned int state)
294 {
295 	ssi->recv_state = state;
296 	switch (state) {
297 	case RECV_IDLE:
298 		del_timer(&ssi->rx_wd);
299 		if (ssi->send_state == SEND_IDLE)
300 			del_timer(&ssi->keep_alive);
301 		break;
302 	case RECV_READY:
303 		/* CMT speech workaround */
304 		if (atomic_read(&ssi->tx_usecnt))
305 			break;
306 		/* Otherwise fall through */
307 	case RECEIVING:
308 		mod_timer(&ssi->keep_alive, jiffies +
309 						msecs_to_jiffies(SSIP_KATOUT));
310 		mod_timer(&ssi->rx_wd, jiffies + msecs_to_jiffies(SSIP_WDTOUT));
311 		break;
312 	default:
313 		break;
314 	}
315 }
316 
317 static void ssip_set_txstate(struct ssi_protocol *ssi, unsigned int state)
318 {
319 	ssi->send_state = state;
320 	switch (state) {
321 	case SEND_IDLE:
322 	case SEND_READY:
323 		del_timer(&ssi->tx_wd);
324 		if (ssi->recv_state == RECV_IDLE)
325 			del_timer(&ssi->keep_alive);
326 		break;
327 	case WAIT4READY:
328 	case SENDING:
329 	case SENDING_SWBREAK:
330 		mod_timer(&ssi->keep_alive,
331 				jiffies + msecs_to_jiffies(SSIP_KATOUT));
332 		mod_timer(&ssi->tx_wd, jiffies + msecs_to_jiffies(SSIP_WDTOUT));
333 		break;
334 	default:
335 		break;
336 	}
337 }
338 
339 struct hsi_client *ssip_slave_get_master(struct hsi_client *slave)
340 {
341 	struct hsi_client *master = ERR_PTR(-ENODEV);
342 	struct ssi_protocol *ssi;
343 
344 	list_for_each_entry(ssi, &ssip_list, link)
345 		if (slave->device.parent == ssi->cl->device.parent) {
346 			master = ssi->cl;
347 			break;
348 		}
349 
350 	return master;
351 }
352 EXPORT_SYMBOL_GPL(ssip_slave_get_master);
353 
354 int ssip_slave_start_tx(struct hsi_client *master)
355 {
356 	struct ssi_protocol *ssi = hsi_client_drvdata(master);
357 
358 	dev_dbg(&master->device, "start TX %d\n", atomic_read(&ssi->tx_usecnt));
359 	spin_lock_bh(&ssi->lock);
360 	if (ssi->send_state == SEND_IDLE) {
361 		ssip_set_txstate(ssi, WAIT4READY);
362 		hsi_start_tx(master);
363 	}
364 	spin_unlock_bh(&ssi->lock);
365 	atomic_inc(&ssi->tx_usecnt);
366 
367 	return 0;
368 }
369 EXPORT_SYMBOL_GPL(ssip_slave_start_tx);
370 
371 int ssip_slave_stop_tx(struct hsi_client *master)
372 {
373 	struct ssi_protocol *ssi = hsi_client_drvdata(master);
374 
375 	WARN_ON_ONCE(atomic_read(&ssi->tx_usecnt) == 0);
376 
377 	if (atomic_dec_and_test(&ssi->tx_usecnt)) {
378 		spin_lock_bh(&ssi->lock);
379 		if ((ssi->send_state == SEND_READY) ||
380 			(ssi->send_state == WAIT4READY)) {
381 			ssip_set_txstate(ssi, SEND_IDLE);
382 			hsi_stop_tx(master);
383 		}
384 		spin_unlock_bh(&ssi->lock);
385 	}
386 	dev_dbg(&master->device, "stop TX %d\n", atomic_read(&ssi->tx_usecnt));
387 
388 	return 0;
389 }
390 EXPORT_SYMBOL_GPL(ssip_slave_stop_tx);
391 
392 int ssip_slave_running(struct hsi_client *master)
393 {
394 	struct ssi_protocol *ssi = hsi_client_drvdata(master);
395 	return netif_running(ssi->netdev);
396 }
397 EXPORT_SYMBOL_GPL(ssip_slave_running);
398 
399 static void ssip_reset(struct hsi_client *cl)
400 {
401 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
402 	struct list_head *head, *tmp;
403 	struct hsi_msg *msg;
404 
405 	if (netif_running(ssi->netdev))
406 		netif_carrier_off(ssi->netdev);
407 	hsi_flush(cl);
408 	spin_lock_bh(&ssi->lock);
409 	if (ssi->send_state != SEND_IDLE)
410 		hsi_stop_tx(cl);
411 	spin_unlock_bh(&ssi->lock);
412 	if (test_and_clear_bit(SSIP_WAKETEST_FLAG, &ssi->flags))
413 		ssi_waketest(cl, 0); /* FIXME: To be removed */
414 	spin_lock_bh(&ssi->lock);
415 	del_timer(&ssi->rx_wd);
416 	del_timer(&ssi->tx_wd);
417 	del_timer(&ssi->keep_alive);
418 	ssi->main_state = 0;
419 	ssi->send_state = 0;
420 	ssi->recv_state = 0;
421 	ssi->flags = 0;
422 	ssi->rxid = 0;
423 	ssi->txid = 0;
424 	list_for_each_safe(head, tmp, &ssi->txqueue) {
425 		msg = list_entry(head, struct hsi_msg, link);
426 		dev_dbg(&cl->device, "Pending TX data\n");
427 		list_del(head);
428 		ssip_free_data(msg);
429 	}
430 	ssi->txqueue_len = 0;
431 	spin_unlock_bh(&ssi->lock);
432 }
433 
434 static void ssip_dump_state(struct hsi_client *cl)
435 {
436 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
437 	struct hsi_msg *msg;
438 
439 	spin_lock_bh(&ssi->lock);
440 	dev_err(&cl->device, "Main state: %d\n", ssi->main_state);
441 	dev_err(&cl->device, "Recv state: %d\n", ssi->recv_state);
442 	dev_err(&cl->device, "Send state: %d\n", ssi->send_state);
443 	dev_err(&cl->device, "CMT %s\n", (ssi->main_state == ACTIVE) ?
444 							"Online" : "Offline");
445 	dev_err(&cl->device, "Wake test %d\n",
446 				test_bit(SSIP_WAKETEST_FLAG, &ssi->flags));
447 	dev_err(&cl->device, "Data RX id: %d\n", ssi->rxid);
448 	dev_err(&cl->device, "Data TX id: %d\n", ssi->txid);
449 
450 	list_for_each_entry(msg, &ssi->txqueue, link)
451 		dev_err(&cl->device, "pending TX data (%p)\n", msg);
452 	spin_unlock_bh(&ssi->lock);
453 }
454 
455 static void ssip_error(struct hsi_client *cl)
456 {
457 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
458 	struct hsi_msg *msg;
459 
460 	ssip_dump_state(cl);
461 	ssip_reset(cl);
462 	msg = ssip_claim_cmd(ssi);
463 	msg->complete = ssip_rxcmd_complete;
464 	hsi_async_read(cl, msg);
465 }
466 
467 static void ssip_keep_alive(unsigned long data)
468 {
469 	struct hsi_client *cl = (struct hsi_client *)data;
470 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
471 
472 	dev_dbg(&cl->device, "Keep alive kick in: m(%d) r(%d) s(%d)\n",
473 		ssi->main_state, ssi->recv_state, ssi->send_state);
474 
475 	spin_lock(&ssi->lock);
476 	if (ssi->recv_state == RECV_IDLE)
477 		switch (ssi->send_state) {
478 		case SEND_READY:
479 			if (atomic_read(&ssi->tx_usecnt) == 0)
480 				break;
481 			/*
482 			 * Fall through. Workaround for cmt-speech
483 			 * in that case we relay on audio timers.
484 			 */
485 		case SEND_IDLE:
486 			spin_unlock(&ssi->lock);
487 			return;
488 		}
489 	mod_timer(&ssi->keep_alive, jiffies + msecs_to_jiffies(SSIP_KATOUT));
490 	spin_unlock(&ssi->lock);
491 }
492 
493 static void ssip_wd(unsigned long data)
494 {
495 	struct hsi_client *cl = (struct hsi_client *)data;
496 
497 	dev_err(&cl->device, "Watchdog trigerred\n");
498 	ssip_error(cl);
499 }
500 
501 static void ssip_send_bootinfo_req_cmd(struct hsi_client *cl)
502 {
503 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
504 	struct hsi_msg *msg;
505 
506 	dev_dbg(&cl->device, "Issuing BOOT INFO REQ command\n");
507 	msg = ssip_claim_cmd(ssi);
508 	ssip_set_cmd(msg, SSIP_BOOTINFO_REQ_CMD(SSIP_LOCAL_VERID));
509 	msg->complete = ssip_release_cmd;
510 	hsi_async_write(cl, msg);
511 	dev_dbg(&cl->device, "Issuing RX command\n");
512 	msg = ssip_claim_cmd(ssi);
513 	msg->complete = ssip_rxcmd_complete;
514 	hsi_async_read(cl, msg);
515 }
516 
517 static void ssip_start_rx(struct hsi_client *cl)
518 {
519 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
520 	struct hsi_msg *msg;
521 
522 	dev_dbg(&cl->device, "RX start M(%d) R(%d)\n", ssi->main_state,
523 						ssi->recv_state);
524 	spin_lock_bh(&ssi->lock);
525 	/*
526 	 * We can have two UP events in a row due to a short low
527 	 * high transition. Therefore we need to ignore the sencond UP event.
528 	 */
529 	if ((ssi->main_state != ACTIVE) || (ssi->recv_state == RECV_READY)) {
530 		spin_unlock_bh(&ssi->lock);
531 		return;
532 	}
533 	ssip_set_rxstate(ssi, RECV_READY);
534 	spin_unlock_bh(&ssi->lock);
535 
536 	msg = ssip_claim_cmd(ssi);
537 	ssip_set_cmd(msg, SSIP_READY_CMD);
538 	msg->complete = ssip_release_cmd;
539 	dev_dbg(&cl->device, "Send READY\n");
540 	hsi_async_write(cl, msg);
541 }
542 
543 static void ssip_stop_rx(struct hsi_client *cl)
544 {
545 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
546 
547 	dev_dbg(&cl->device, "RX stop M(%d)\n", ssi->main_state);
548 	spin_lock_bh(&ssi->lock);
549 	if (likely(ssi->main_state == ACTIVE))
550 		ssip_set_rxstate(ssi, RECV_IDLE);
551 	spin_unlock_bh(&ssi->lock);
552 }
553 
554 static void ssip_free_strans(struct hsi_msg *msg)
555 {
556 	ssip_free_data(msg->context);
557 	ssip_release_cmd(msg);
558 }
559 
560 static void ssip_strans_complete(struct hsi_msg *msg)
561 {
562 	struct hsi_client *cl = msg->cl;
563 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
564 	struct hsi_msg *data;
565 
566 	data = msg->context;
567 	ssip_release_cmd(msg);
568 	spin_lock_bh(&ssi->lock);
569 	ssip_set_txstate(ssi, SENDING);
570 	spin_unlock_bh(&ssi->lock);
571 	hsi_async_write(cl, data);
572 }
573 
574 static int ssip_xmit(struct hsi_client *cl)
575 {
576 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
577 	struct hsi_msg *msg, *dmsg;
578 	struct sk_buff *skb;
579 
580 	spin_lock_bh(&ssi->lock);
581 	if (list_empty(&ssi->txqueue)) {
582 		spin_unlock_bh(&ssi->lock);
583 		return 0;
584 	}
585 	dmsg = list_first_entry(&ssi->txqueue, struct hsi_msg, link);
586 	list_del(&dmsg->link);
587 	ssi->txqueue_len--;
588 	spin_unlock_bh(&ssi->lock);
589 
590 	msg = ssip_claim_cmd(ssi);
591 	skb = dmsg->context;
592 	msg->context = dmsg;
593 	msg->complete = ssip_strans_complete;
594 	msg->destructor = ssip_free_strans;
595 
596 	spin_lock_bh(&ssi->lock);
597 	ssip_set_cmd(msg, SSIP_START_TRANS_CMD(SSIP_BYTES_TO_FRAMES(skb->len),
598 								ssi->txid));
599 	ssi->txid++;
600 	ssip_set_txstate(ssi, SENDING);
601 	spin_unlock_bh(&ssi->lock);
602 
603 	dev_dbg(&cl->device, "Send STRANS (%d frames)\n",
604 						SSIP_BYTES_TO_FRAMES(skb->len));
605 
606 	return hsi_async_write(cl, msg);
607 }
608 
609 /* In soft IRQ context */
610 static void ssip_pn_rx(struct sk_buff *skb)
611 {
612 	struct net_device *dev = skb->dev;
613 
614 	if (unlikely(!netif_running(dev))) {
615 		dev_dbg(&dev->dev, "Drop RX packet\n");
616 		dev->stats.rx_dropped++;
617 		dev_kfree_skb(skb);
618 		return;
619 	}
620 	if (unlikely(!pskb_may_pull(skb, SSIP_MIN_PN_HDR))) {
621 		dev_dbg(&dev->dev, "Error drop RX packet\n");
622 		dev->stats.rx_errors++;
623 		dev->stats.rx_length_errors++;
624 		dev_kfree_skb(skb);
625 		return;
626 	}
627 	dev->stats.rx_packets++;
628 	dev->stats.rx_bytes += skb->len;
629 
630 	/* length field is exchanged in network byte order */
631 	((u16 *)skb->data)[2] = ntohs(((u16 *)skb->data)[2]);
632 	dev_dbg(&dev->dev, "RX length fixed (%04x -> %u)\n",
633 			((u16 *)skb->data)[2], ntohs(((u16 *)skb->data)[2]));
634 
635 	skb->protocol = htons(ETH_P_PHONET);
636 	skb_reset_mac_header(skb);
637 	__skb_pull(skb, 1);
638 	netif_rx(skb);
639 }
640 
641 static void ssip_rx_data_complete(struct hsi_msg *msg)
642 {
643 	struct hsi_client *cl = msg->cl;
644 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
645 	struct sk_buff *skb;
646 
647 	if (msg->status == HSI_STATUS_ERROR) {
648 		dev_err(&cl->device, "RX data error\n");
649 		ssip_free_data(msg);
650 		ssip_error(cl);
651 		return;
652 	}
653 	del_timer(&ssi->rx_wd); /* FIXME: Revisit */
654 	skb = msg->context;
655 	ssip_pn_rx(skb);
656 	hsi_free_msg(msg);
657 }
658 
659 static void ssip_rx_bootinforeq(struct hsi_client *cl, u32 cmd)
660 {
661 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
662 	struct hsi_msg *msg;
663 
664 	/* Workaroud: Ignore CMT Loader message leftover */
665 	if (cmd == SSIP_CMT_LOADER_SYNC)
666 		return;
667 
668 	switch (ssi->main_state) {
669 	case ACTIVE:
670 		dev_err(&cl->device, "Boot info req on active state\n");
671 		ssip_error(cl);
672 		/* Fall through */
673 	case INIT:
674 	case HANDSHAKE:
675 		spin_lock_bh(&ssi->lock);
676 		ssi->main_state = HANDSHAKE;
677 		spin_unlock_bh(&ssi->lock);
678 
679 		if (!test_and_set_bit(SSIP_WAKETEST_FLAG, &ssi->flags))
680 			ssi_waketest(cl, 1); /* FIXME: To be removed */
681 
682 		spin_lock_bh(&ssi->lock);
683 		/* Start boot handshake watchdog */
684 		mod_timer(&ssi->tx_wd, jiffies + msecs_to_jiffies(SSIP_WDTOUT));
685 		spin_unlock_bh(&ssi->lock);
686 		dev_dbg(&cl->device, "Send BOOTINFO_RESP\n");
687 		if (SSIP_DATA_VERSION(cmd) != SSIP_LOCAL_VERID)
688 			dev_warn(&cl->device, "boot info req verid mismatch\n");
689 		msg = ssip_claim_cmd(ssi);
690 		ssip_set_cmd(msg, SSIP_BOOTINFO_RESP_CMD(SSIP_LOCAL_VERID));
691 		msg->complete = ssip_release_cmd;
692 		hsi_async_write(cl, msg);
693 		break;
694 	default:
695 		dev_dbg(&cl->device, "Wrong state M(%d)\n", ssi->main_state);
696 		break;
697 	}
698 }
699 
700 static void ssip_rx_bootinforesp(struct hsi_client *cl, u32 cmd)
701 {
702 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
703 
704 	if (SSIP_DATA_VERSION(cmd) != SSIP_LOCAL_VERID)
705 		dev_warn(&cl->device, "boot info resp verid mismatch\n");
706 
707 	spin_lock_bh(&ssi->lock);
708 	if (ssi->main_state != ACTIVE)
709 		/* Use tx_wd as a boot watchdog in non ACTIVE state */
710 		mod_timer(&ssi->tx_wd, jiffies + msecs_to_jiffies(SSIP_WDTOUT));
711 	else
712 		dev_dbg(&cl->device, "boot info resp ignored M(%d)\n",
713 							ssi->main_state);
714 	spin_unlock_bh(&ssi->lock);
715 }
716 
717 static void ssip_rx_waketest(struct hsi_client *cl, u32 cmd)
718 {
719 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
720 	unsigned int wkres = SSIP_PAYLOAD(cmd);
721 
722 	spin_lock_bh(&ssi->lock);
723 	if (ssi->main_state != HANDSHAKE) {
724 		dev_dbg(&cl->device, "wake lines test ignored M(%d)\n",
725 							ssi->main_state);
726 		spin_unlock_bh(&ssi->lock);
727 		return;
728 	}
729 	spin_unlock_bh(&ssi->lock);
730 
731 	if (test_and_clear_bit(SSIP_WAKETEST_FLAG, &ssi->flags))
732 		ssi_waketest(cl, 0); /* FIXME: To be removed */
733 
734 	spin_lock_bh(&ssi->lock);
735 	ssi->main_state = ACTIVE;
736 	del_timer(&ssi->tx_wd); /* Stop boot handshake timer */
737 	spin_unlock_bh(&ssi->lock);
738 
739 	dev_notice(&cl->device, "WAKELINES TEST %s\n",
740 				wkres & SSIP_WAKETEST_FAILED ? "FAILED" : "OK");
741 	if (wkres & SSIP_WAKETEST_FAILED) {
742 		ssip_error(cl);
743 		return;
744 	}
745 	dev_dbg(&cl->device, "CMT is ONLINE\n");
746 	netif_wake_queue(ssi->netdev);
747 	netif_carrier_on(ssi->netdev);
748 }
749 
750 static void ssip_rx_ready(struct hsi_client *cl)
751 {
752 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
753 
754 	spin_lock_bh(&ssi->lock);
755 	if (unlikely(ssi->main_state != ACTIVE)) {
756 		dev_dbg(&cl->device, "READY on wrong state: S(%d) M(%d)\n",
757 					ssi->send_state, ssi->main_state);
758 		spin_unlock_bh(&ssi->lock);
759 		return;
760 	}
761 	if (ssi->send_state != WAIT4READY) {
762 		dev_dbg(&cl->device, "Ignore spurious READY command\n");
763 		spin_unlock_bh(&ssi->lock);
764 		return;
765 	}
766 	ssip_set_txstate(ssi, SEND_READY);
767 	spin_unlock_bh(&ssi->lock);
768 	ssip_xmit(cl);
769 }
770 
771 static void ssip_rx_strans(struct hsi_client *cl, u32 cmd)
772 {
773 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
774 	struct sk_buff *skb;
775 	struct hsi_msg *msg;
776 	int len = SSIP_PDU_LENGTH(cmd);
777 
778 	dev_dbg(&cl->device, "RX strans: %d frames\n", len);
779 	spin_lock_bh(&ssi->lock);
780 	if (unlikely(ssi->main_state != ACTIVE)) {
781 		dev_err(&cl->device, "START TRANS wrong state: S(%d) M(%d)\n",
782 					ssi->send_state, ssi->main_state);
783 		spin_unlock_bh(&ssi->lock);
784 		return;
785 	}
786 	ssip_set_rxstate(ssi, RECEIVING);
787 	if (unlikely(SSIP_MSG_ID(cmd) != ssi->rxid)) {
788 		dev_err(&cl->device, "START TRANS id %d expected %d\n",
789 					SSIP_MSG_ID(cmd), ssi->rxid);
790 		spin_unlock_bh(&ssi->lock);
791 		goto out1;
792 	}
793 	ssi->rxid++;
794 	spin_unlock_bh(&ssi->lock);
795 	skb = netdev_alloc_skb(ssi->netdev, len * 4);
796 	if (unlikely(!skb)) {
797 		dev_err(&cl->device, "No memory for rx skb\n");
798 		goto out1;
799 	}
800 	skb->dev = ssi->netdev;
801 	skb_put(skb, len * 4);
802 	msg = ssip_alloc_data(ssi, skb, GFP_ATOMIC);
803 	if (unlikely(!msg)) {
804 		dev_err(&cl->device, "No memory for RX data msg\n");
805 		goto out2;
806 	}
807 	msg->complete = ssip_rx_data_complete;
808 	hsi_async_read(cl, msg);
809 
810 	return;
811 out2:
812 	dev_kfree_skb(skb);
813 out1:
814 	ssip_error(cl);
815 }
816 
817 static void ssip_rxcmd_complete(struct hsi_msg *msg)
818 {
819 	struct hsi_client *cl = msg->cl;
820 	u32 cmd = ssip_get_cmd(msg);
821 	unsigned int cmdid = SSIP_COMMAND(cmd);
822 
823 	if (msg->status == HSI_STATUS_ERROR) {
824 		dev_err(&cl->device, "RX error detected\n");
825 		ssip_release_cmd(msg);
826 		ssip_error(cl);
827 		return;
828 	}
829 	hsi_async_read(cl, msg);
830 	dev_dbg(&cl->device, "RX cmd: 0x%08x\n", cmd);
831 	switch (cmdid) {
832 	case SSIP_SW_BREAK:
833 		/* Ignored */
834 		break;
835 	case SSIP_BOOTINFO_REQ:
836 		ssip_rx_bootinforeq(cl, cmd);
837 		break;
838 	case SSIP_BOOTINFO_RESP:
839 		ssip_rx_bootinforesp(cl, cmd);
840 		break;
841 	case SSIP_WAKETEST_RESULT:
842 		ssip_rx_waketest(cl, cmd);
843 		break;
844 	case SSIP_START_TRANS:
845 		ssip_rx_strans(cl, cmd);
846 		break;
847 	case SSIP_READY:
848 		ssip_rx_ready(cl);
849 		break;
850 	default:
851 		dev_warn(&cl->device, "command 0x%08x not supported\n", cmd);
852 		break;
853 	}
854 }
855 
856 static void ssip_swbreak_complete(struct hsi_msg *msg)
857 {
858 	struct hsi_client *cl = msg->cl;
859 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
860 
861 	ssip_release_cmd(msg);
862 	spin_lock_bh(&ssi->lock);
863 	if (list_empty(&ssi->txqueue)) {
864 		if (atomic_read(&ssi->tx_usecnt)) {
865 			ssip_set_txstate(ssi, SEND_READY);
866 		} else {
867 			ssip_set_txstate(ssi, SEND_IDLE);
868 			hsi_stop_tx(cl);
869 		}
870 		spin_unlock_bh(&ssi->lock);
871 	} else {
872 		spin_unlock_bh(&ssi->lock);
873 		ssip_xmit(cl);
874 	}
875 	netif_wake_queue(ssi->netdev);
876 }
877 
878 static void ssip_tx_data_complete(struct hsi_msg *msg)
879 {
880 	struct hsi_client *cl = msg->cl;
881 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
882 	struct hsi_msg *cmsg;
883 
884 	if (msg->status == HSI_STATUS_ERROR) {
885 		dev_err(&cl->device, "TX data error\n");
886 		ssip_error(cl);
887 		goto out;
888 	}
889 	spin_lock_bh(&ssi->lock);
890 	if (list_empty(&ssi->txqueue)) {
891 		ssip_set_txstate(ssi, SENDING_SWBREAK);
892 		spin_unlock_bh(&ssi->lock);
893 		cmsg = ssip_claim_cmd(ssi);
894 		ssip_set_cmd(cmsg, SSIP_SWBREAK_CMD);
895 		cmsg->complete = ssip_swbreak_complete;
896 		dev_dbg(&cl->device, "Send SWBREAK\n");
897 		hsi_async_write(cl, cmsg);
898 	} else {
899 		spin_unlock_bh(&ssi->lock);
900 		ssip_xmit(cl);
901 	}
902 out:
903 	ssip_free_data(msg);
904 }
905 
906 static void ssip_port_event(struct hsi_client *cl, unsigned long event)
907 {
908 	switch (event) {
909 	case HSI_EVENT_START_RX:
910 		ssip_start_rx(cl);
911 		break;
912 	case HSI_EVENT_STOP_RX:
913 		ssip_stop_rx(cl);
914 		break;
915 	default:
916 		return;
917 	}
918 }
919 
920 static int ssip_pn_open(struct net_device *dev)
921 {
922 	struct hsi_client *cl = to_hsi_client(dev->dev.parent);
923 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
924 	int err;
925 
926 	err = hsi_claim_port(cl, 1);
927 	if (err < 0) {
928 		dev_err(&cl->device, "SSI port already claimed\n");
929 		return err;
930 	}
931 	err = hsi_register_port_event(cl, ssip_port_event);
932 	if (err < 0) {
933 		dev_err(&cl->device, "Register HSI port event failed (%d)\n",
934 			err);
935 		return err;
936 	}
937 	dev_dbg(&cl->device, "Configuring SSI port\n");
938 	hsi_setup(cl);
939 
940 	if (!test_and_set_bit(SSIP_WAKETEST_FLAG, &ssi->flags))
941 		ssi_waketest(cl, 1); /* FIXME: To be removed */
942 
943 	spin_lock_bh(&ssi->lock);
944 	ssi->main_state = HANDSHAKE;
945 	spin_unlock_bh(&ssi->lock);
946 
947 	ssip_send_bootinfo_req_cmd(cl);
948 
949 	return 0;
950 }
951 
952 static int ssip_pn_stop(struct net_device *dev)
953 {
954 	struct hsi_client *cl = to_hsi_client(dev->dev.parent);
955 
956 	ssip_reset(cl);
957 	hsi_unregister_port_event(cl);
958 	hsi_release_port(cl);
959 
960 	return 0;
961 }
962 
963 static int ssip_pn_set_mtu(struct net_device *dev, int new_mtu)
964 {
965 	if (new_mtu > SSIP_MAX_MTU || new_mtu < PHONET_MIN_MTU)
966 		return -EINVAL;
967 	dev->mtu = new_mtu;
968 
969 	return 0;
970 }
971 
972 static void ssip_xmit_work(struct work_struct *work)
973 {
974 	struct ssi_protocol *ssi =
975 				container_of(work, struct ssi_protocol, work);
976 	struct hsi_client *cl = ssi->cl;
977 
978 	ssip_xmit(cl);
979 }
980 
981 static int ssip_pn_xmit(struct sk_buff *skb, struct net_device *dev)
982 {
983 	struct hsi_client *cl = to_hsi_client(dev->dev.parent);
984 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
985 	struct hsi_msg *msg;
986 
987 	if ((skb->protocol != htons(ETH_P_PHONET)) ||
988 					(skb->len < SSIP_MIN_PN_HDR))
989 		goto drop;
990 	/* Pad to 32-bits - FIXME: Revisit*/
991 	if ((skb->len & 3) && skb_pad(skb, 4 - (skb->len & 3)))
992 		goto drop;
993 
994 	/*
995 	 * Modem sends Phonet messages over SSI with its own endianess...
996 	 * Assume that modem has the same endianess as we do.
997 	 */
998 	if (skb_cow_head(skb, 0))
999 		goto drop;
1000 
1001 	/* length field is exchanged in network byte order */
1002 	((u16 *)skb->data)[2] = htons(((u16 *)skb->data)[2]);
1003 
1004 	msg = ssip_alloc_data(ssi, skb, GFP_ATOMIC);
1005 	if (!msg) {
1006 		dev_dbg(&cl->device, "Dropping tx data: No memory\n");
1007 		goto drop;
1008 	}
1009 	msg->complete = ssip_tx_data_complete;
1010 
1011 	spin_lock_bh(&ssi->lock);
1012 	if (unlikely(ssi->main_state != ACTIVE)) {
1013 		spin_unlock_bh(&ssi->lock);
1014 		dev_dbg(&cl->device, "Dropping tx data: CMT is OFFLINE\n");
1015 		goto drop2;
1016 	}
1017 	list_add_tail(&msg->link, &ssi->txqueue);
1018 	ssi->txqueue_len++;
1019 	if (dev->tx_queue_len < ssi->txqueue_len) {
1020 		dev_info(&cl->device, "TX queue full %d\n", ssi->txqueue_len);
1021 		netif_stop_queue(dev);
1022 	}
1023 	if (ssi->send_state == SEND_IDLE) {
1024 		ssip_set_txstate(ssi, WAIT4READY);
1025 		spin_unlock_bh(&ssi->lock);
1026 		dev_dbg(&cl->device, "Start TX qlen %d\n", ssi->txqueue_len);
1027 		hsi_start_tx(cl);
1028 	} else if (ssi->send_state == SEND_READY) {
1029 		/* Needed for cmt-speech workaround */
1030 		dev_dbg(&cl->device, "Start TX on SEND READY qlen %d\n",
1031 							ssi->txqueue_len);
1032 		spin_unlock_bh(&ssi->lock);
1033 		schedule_work(&ssi->work);
1034 	} else {
1035 		spin_unlock_bh(&ssi->lock);
1036 	}
1037 	dev->stats.tx_packets++;
1038 	dev->stats.tx_bytes += skb->len;
1039 
1040 	return 0;
1041 drop2:
1042 	hsi_free_msg(msg);
1043 drop:
1044 	dev->stats.tx_dropped++;
1045 	dev_kfree_skb(skb);
1046 
1047 	return 0;
1048 }
1049 
1050 /* CMT reset event handler */
1051 void ssip_reset_event(struct hsi_client *master)
1052 {
1053 	struct ssi_protocol *ssi = hsi_client_drvdata(master);
1054 	dev_err(&ssi->cl->device, "CMT reset detected!\n");
1055 	ssip_error(ssi->cl);
1056 }
1057 EXPORT_SYMBOL_GPL(ssip_reset_event);
1058 
1059 static const struct net_device_ops ssip_pn_ops = {
1060 	.ndo_open	= ssip_pn_open,
1061 	.ndo_stop	= ssip_pn_stop,
1062 	.ndo_start_xmit	= ssip_pn_xmit,
1063 	.ndo_change_mtu	= ssip_pn_set_mtu,
1064 };
1065 
1066 static void ssip_pn_setup(struct net_device *dev)
1067 {
1068 	dev->features		= 0;
1069 	dev->netdev_ops		= &ssip_pn_ops;
1070 	dev->type		= ARPHRD_PHONET;
1071 	dev->flags		= IFF_POINTOPOINT | IFF_NOARP;
1072 	dev->mtu		= SSIP_DEFAULT_MTU;
1073 	dev->hard_header_len	= 1;
1074 	dev->dev_addr[0]	= PN_MEDIA_SOS;
1075 	dev->addr_len		= 1;
1076 	dev->tx_queue_len	= SSIP_TXQUEUE_LEN;
1077 
1078 	dev->destructor		= free_netdev;
1079 	dev->header_ops		= &phonet_header_ops;
1080 }
1081 
1082 static int ssi_protocol_probe(struct device *dev)
1083 {
1084 	static const char ifname[] = "phonet%d";
1085 	struct hsi_client *cl = to_hsi_client(dev);
1086 	struct ssi_protocol *ssi;
1087 	int err;
1088 
1089 	ssi = kzalloc(sizeof(*ssi), GFP_KERNEL);
1090 	if (!ssi) {
1091 		dev_err(dev, "No memory for ssi protocol\n");
1092 		return -ENOMEM;
1093 	}
1094 
1095 	spin_lock_init(&ssi->lock);
1096 	init_timer_deferrable(&ssi->rx_wd);
1097 	init_timer_deferrable(&ssi->tx_wd);
1098 	init_timer(&ssi->keep_alive);
1099 	ssi->rx_wd.data = (unsigned long)cl;
1100 	ssi->rx_wd.function = ssip_wd;
1101 	ssi->tx_wd.data = (unsigned long)cl;
1102 	ssi->tx_wd.function = ssip_wd;
1103 	ssi->keep_alive.data = (unsigned long)cl;
1104 	ssi->keep_alive.function = ssip_keep_alive;
1105 	INIT_LIST_HEAD(&ssi->txqueue);
1106 	INIT_LIST_HEAD(&ssi->cmdqueue);
1107 	atomic_set(&ssi->tx_usecnt, 0);
1108 	hsi_client_set_drvdata(cl, ssi);
1109 	ssi->cl = cl;
1110 	INIT_WORK(&ssi->work, ssip_xmit_work);
1111 
1112 	ssi->channel_id_cmd = hsi_get_channel_id_by_name(cl, "mcsaab-control");
1113 	if (ssi->channel_id_cmd < 0) {
1114 		err = ssi->channel_id_cmd;
1115 		dev_err(dev, "Could not get cmd channel (%d)\n", err);
1116 		goto out;
1117 	}
1118 
1119 	ssi->channel_id_data = hsi_get_channel_id_by_name(cl, "mcsaab-data");
1120 	if (ssi->channel_id_data < 0) {
1121 		err = ssi->channel_id_data;
1122 		dev_err(dev, "Could not get data channel (%d)\n", err);
1123 		goto out;
1124 	}
1125 
1126 	err = ssip_alloc_cmds(ssi);
1127 	if (err < 0) {
1128 		dev_err(dev, "No memory for commands\n");
1129 		goto out;
1130 	}
1131 
1132 	ssi->netdev = alloc_netdev(0, ifname, NET_NAME_UNKNOWN, ssip_pn_setup);
1133 	if (!ssi->netdev) {
1134 		dev_err(dev, "No memory for netdev\n");
1135 		err = -ENOMEM;
1136 		goto out1;
1137 	}
1138 
1139 	SET_NETDEV_DEV(ssi->netdev, dev);
1140 	netif_carrier_off(ssi->netdev);
1141 	err = register_netdev(ssi->netdev);
1142 	if (err < 0) {
1143 		dev_err(dev, "Register netdev failed (%d)\n", err);
1144 		goto out2;
1145 	}
1146 
1147 	list_add(&ssi->link, &ssip_list);
1148 
1149 	dev_dbg(dev, "channel configuration: cmd=%d, data=%d\n",
1150 		ssi->channel_id_cmd, ssi->channel_id_data);
1151 
1152 	return 0;
1153 out2:
1154 	free_netdev(ssi->netdev);
1155 out1:
1156 	ssip_free_cmds(ssi);
1157 out:
1158 	kfree(ssi);
1159 
1160 	return err;
1161 }
1162 
1163 static int ssi_protocol_remove(struct device *dev)
1164 {
1165 	struct hsi_client *cl = to_hsi_client(dev);
1166 	struct ssi_protocol *ssi = hsi_client_drvdata(cl);
1167 
1168 	list_del(&ssi->link);
1169 	unregister_netdev(ssi->netdev);
1170 	ssip_free_cmds(ssi);
1171 	hsi_client_set_drvdata(cl, NULL);
1172 	kfree(ssi);
1173 
1174 	return 0;
1175 }
1176 
1177 static struct hsi_client_driver ssip_driver = {
1178 	.driver = {
1179 		.name	= "ssi-protocol",
1180 		.owner	= THIS_MODULE,
1181 		.probe	= ssi_protocol_probe,
1182 		.remove	= ssi_protocol_remove,
1183 	},
1184 };
1185 
1186 static int __init ssip_init(void)
1187 {
1188 	pr_info("SSI protocol aka McSAAB added\n");
1189 
1190 	return hsi_register_client_driver(&ssip_driver);
1191 }
1192 module_init(ssip_init);
1193 
1194 static void __exit ssip_exit(void)
1195 {
1196 	hsi_unregister_client_driver(&ssip_driver);
1197 	pr_info("SSI protocol driver removed\n");
1198 }
1199 module_exit(ssip_exit);
1200 
1201 MODULE_ALIAS("hsi:ssi-protocol");
1202 MODULE_AUTHOR("Carlos Chinea <carlos.chinea@nokia.com>");
1203 MODULE_AUTHOR("Remi Denis-Courmont <remi.denis-courmont@nokia.com>");
1204 MODULE_DESCRIPTION("SSI protocol improved aka McSAAB");
1205 MODULE_LICENSE("GPL");
1206