1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * ssi_protocol.c 4 * 5 * Implementation of the SSI McSAAB improved protocol. 6 * 7 * Copyright (C) 2010 Nokia Corporation. All rights reserved. 8 * Copyright (C) 2013 Sebastian Reichel <sre@kernel.org> 9 * 10 * Contact: Carlos Chinea <carlos.chinea@nokia.com> 11 */ 12 13 #include <linux/atomic.h> 14 #include <linux/clk.h> 15 #include <linux/device.h> 16 #include <linux/err.h> 17 #include <linux/gpio.h> 18 #include <linux/if_ether.h> 19 #include <linux/if_arp.h> 20 #include <linux/if_phonet.h> 21 #include <linux/init.h> 22 #include <linux/irq.h> 23 #include <linux/list.h> 24 #include <linux/module.h> 25 #include <linux/netdevice.h> 26 #include <linux/notifier.h> 27 #include <linux/scatterlist.h> 28 #include <linux/skbuff.h> 29 #include <linux/slab.h> 30 #include <linux/spinlock.h> 31 #include <linux/timer.h> 32 #include <linux/hsi/hsi.h> 33 #include <linux/hsi/ssi_protocol.h> 34 35 void ssi_waketest(struct hsi_client *cl, unsigned int enable); 36 37 #define SSIP_TXQUEUE_LEN 100 38 #define SSIP_MAX_MTU 65535 39 #define SSIP_DEFAULT_MTU 4000 40 #define PN_MEDIA_SOS 21 41 #define SSIP_MIN_PN_HDR 6 /* FIXME: Revisit */ 42 #define SSIP_WDTOUT 2000 /* FIXME: has to be 500 msecs */ 43 #define SSIP_KATOUT 15 /* 15 msecs */ 44 #define SSIP_MAX_CMDS 5 /* Number of pre-allocated commands buffers */ 45 #define SSIP_BYTES_TO_FRAMES(x) ((((x) - 1) >> 2) + 1) 46 #define SSIP_CMT_LOADER_SYNC 0x11223344 47 /* 48 * SSI protocol command definitions 49 */ 50 #define SSIP_COMMAND(data) ((data) >> 28) 51 #define SSIP_PAYLOAD(data) ((data) & 0xfffffff) 52 /* Commands */ 53 #define SSIP_SW_BREAK 0 54 #define SSIP_BOOTINFO_REQ 1 55 #define SSIP_BOOTINFO_RESP 2 56 #define SSIP_WAKETEST_RESULT 3 57 #define SSIP_START_TRANS 4 58 #define SSIP_READY 5 59 /* Payloads */ 60 #define SSIP_DATA_VERSION(data) ((data) & 0xff) 61 #define SSIP_LOCAL_VERID 1 62 #define SSIP_WAKETEST_OK 0 63 #define SSIP_WAKETEST_FAILED 1 64 #define SSIP_PDU_LENGTH(data) (((data) >> 8) & 0xffff) 65 #define SSIP_MSG_ID(data) ((data) & 0xff) 66 /* Generic Command */ 67 #define SSIP_CMD(cmd, payload) (((cmd) << 28) | ((payload) & 0xfffffff)) 68 /* Commands for the control channel */ 69 #define SSIP_BOOTINFO_REQ_CMD(ver) \ 70 SSIP_CMD(SSIP_BOOTINFO_REQ, SSIP_DATA_VERSION(ver)) 71 #define SSIP_BOOTINFO_RESP_CMD(ver) \ 72 SSIP_CMD(SSIP_BOOTINFO_RESP, SSIP_DATA_VERSION(ver)) 73 #define SSIP_START_TRANS_CMD(pdulen, id) \ 74 SSIP_CMD(SSIP_START_TRANS, (((pdulen) << 8) | SSIP_MSG_ID(id))) 75 #define SSIP_READY_CMD SSIP_CMD(SSIP_READY, 0) 76 #define SSIP_SWBREAK_CMD SSIP_CMD(SSIP_SW_BREAK, 0) 77 78 #define SSIP_WAKETEST_FLAG 0 79 80 /* Main state machine states */ 81 enum { 82 INIT, 83 HANDSHAKE, 84 ACTIVE, 85 }; 86 87 /* Send state machine states */ 88 enum { 89 SEND_IDLE, 90 WAIT4READY, 91 SEND_READY, 92 SENDING, 93 SENDING_SWBREAK, 94 }; 95 96 /* Receive state machine states */ 97 enum { 98 RECV_IDLE, 99 RECV_READY, 100 RECEIVING, 101 }; 102 103 /** 104 * struct ssi_protocol - SSI protocol (McSAAB) data 105 * @main_state: Main state machine 106 * @send_state: TX state machine 107 * @recv_state: RX state machine 108 * @flags: Flags, currently only used to follow wake line test 109 * @rxid: RX data id 110 * @txid: TX data id 111 * @txqueue_len: TX queue length 112 * @tx_wd: TX watchdog 113 * @rx_wd: RX watchdog 114 * @keep_alive: Workaround for SSI HW bug 115 * @lock: To serialize access to this struct 116 * @netdev: Phonet network device 117 * @txqueue: TX data queue 118 * @cmdqueue: Queue of free commands 119 * @cl: HSI client own reference 120 * @link: Link for ssip_list 121 * @tx_usecount: Refcount to keep track the slaves that use the wake line 122 * @channel_id_cmd: HSI channel id for command stream 123 * @channel_id_data: HSI channel id for data stream 124 */ 125 struct ssi_protocol { 126 unsigned int main_state; 127 unsigned int send_state; 128 unsigned int recv_state; 129 unsigned long flags; 130 u8 rxid; 131 u8 txid; 132 unsigned int txqueue_len; 133 struct timer_list tx_wd; 134 struct timer_list rx_wd; 135 struct timer_list keep_alive; /* wake-up workaround */ 136 spinlock_t lock; 137 struct net_device *netdev; 138 struct list_head txqueue; 139 struct list_head cmdqueue; 140 struct work_struct work; 141 struct hsi_client *cl; 142 struct list_head link; 143 atomic_t tx_usecnt; 144 int channel_id_cmd; 145 int channel_id_data; 146 }; 147 148 /* List of ssi protocol instances */ 149 static LIST_HEAD(ssip_list); 150 151 static void ssip_rxcmd_complete(struct hsi_msg *msg); 152 153 static inline void ssip_set_cmd(struct hsi_msg *msg, u32 cmd) 154 { 155 u32 *data; 156 157 data = sg_virt(msg->sgt.sgl); 158 *data = cmd; 159 } 160 161 static inline u32 ssip_get_cmd(struct hsi_msg *msg) 162 { 163 u32 *data; 164 165 data = sg_virt(msg->sgt.sgl); 166 167 return *data; 168 } 169 170 static void ssip_skb_to_msg(struct sk_buff *skb, struct hsi_msg *msg) 171 { 172 skb_frag_t *frag; 173 struct scatterlist *sg; 174 int i; 175 176 BUG_ON(msg->sgt.nents != (unsigned int)(skb_shinfo(skb)->nr_frags + 1)); 177 178 sg = msg->sgt.sgl; 179 sg_set_buf(sg, skb->data, skb_headlen(skb)); 180 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { 181 sg = sg_next(sg); 182 BUG_ON(!sg); 183 frag = &skb_shinfo(skb)->frags[i]; 184 sg_set_page(sg, skb_frag_page(frag), skb_frag_size(frag), 185 skb_frag_off(frag)); 186 } 187 } 188 189 static void ssip_free_data(struct hsi_msg *msg) 190 { 191 struct sk_buff *skb; 192 193 skb = msg->context; 194 pr_debug("free data: msg %p context %p skb %p\n", msg, msg->context, 195 skb); 196 msg->destructor = NULL; 197 dev_kfree_skb(skb); 198 hsi_free_msg(msg); 199 } 200 201 static struct hsi_msg *ssip_alloc_data(struct ssi_protocol *ssi, 202 struct sk_buff *skb, gfp_t flags) 203 { 204 struct hsi_msg *msg; 205 206 msg = hsi_alloc_msg(skb_shinfo(skb)->nr_frags + 1, flags); 207 if (!msg) 208 return NULL; 209 ssip_skb_to_msg(skb, msg); 210 msg->destructor = ssip_free_data; 211 msg->channel = ssi->channel_id_data; 212 msg->context = skb; 213 214 return msg; 215 } 216 217 static inline void ssip_release_cmd(struct hsi_msg *msg) 218 { 219 struct ssi_protocol *ssi = hsi_client_drvdata(msg->cl); 220 221 dev_dbg(&msg->cl->device, "Release cmd 0x%08x\n", ssip_get_cmd(msg)); 222 spin_lock_bh(&ssi->lock); 223 list_add_tail(&msg->link, &ssi->cmdqueue); 224 spin_unlock_bh(&ssi->lock); 225 } 226 227 static struct hsi_msg *ssip_claim_cmd(struct ssi_protocol *ssi) 228 { 229 struct hsi_msg *msg; 230 231 BUG_ON(list_empty(&ssi->cmdqueue)); 232 233 spin_lock_bh(&ssi->lock); 234 msg = list_first_entry(&ssi->cmdqueue, struct hsi_msg, link); 235 list_del(&msg->link); 236 spin_unlock_bh(&ssi->lock); 237 msg->destructor = ssip_release_cmd; 238 239 return msg; 240 } 241 242 static void ssip_free_cmds(struct ssi_protocol *ssi) 243 { 244 struct hsi_msg *msg, *tmp; 245 246 list_for_each_entry_safe(msg, tmp, &ssi->cmdqueue, link) { 247 list_del(&msg->link); 248 msg->destructor = NULL; 249 kfree(sg_virt(msg->sgt.sgl)); 250 hsi_free_msg(msg); 251 } 252 } 253 254 static int ssip_alloc_cmds(struct ssi_protocol *ssi) 255 { 256 struct hsi_msg *msg; 257 u32 *buf; 258 unsigned int i; 259 260 for (i = 0; i < SSIP_MAX_CMDS; i++) { 261 msg = hsi_alloc_msg(1, GFP_KERNEL); 262 if (!msg) 263 goto out; 264 buf = kmalloc(sizeof(*buf), GFP_KERNEL); 265 if (!buf) { 266 hsi_free_msg(msg); 267 goto out; 268 } 269 sg_init_one(msg->sgt.sgl, buf, sizeof(*buf)); 270 msg->channel = ssi->channel_id_cmd; 271 list_add_tail(&msg->link, &ssi->cmdqueue); 272 } 273 274 return 0; 275 out: 276 ssip_free_cmds(ssi); 277 278 return -ENOMEM; 279 } 280 281 static void ssip_set_rxstate(struct ssi_protocol *ssi, unsigned int state) 282 { 283 ssi->recv_state = state; 284 switch (state) { 285 case RECV_IDLE: 286 del_timer(&ssi->rx_wd); 287 if (ssi->send_state == SEND_IDLE) 288 del_timer(&ssi->keep_alive); 289 break; 290 case RECV_READY: 291 /* CMT speech workaround */ 292 if (atomic_read(&ssi->tx_usecnt)) 293 break; 294 /* Else, fall through */ 295 case RECEIVING: 296 mod_timer(&ssi->keep_alive, jiffies + 297 msecs_to_jiffies(SSIP_KATOUT)); 298 mod_timer(&ssi->rx_wd, jiffies + msecs_to_jiffies(SSIP_WDTOUT)); 299 break; 300 default: 301 break; 302 } 303 } 304 305 static void ssip_set_txstate(struct ssi_protocol *ssi, unsigned int state) 306 { 307 ssi->send_state = state; 308 switch (state) { 309 case SEND_IDLE: 310 case SEND_READY: 311 del_timer(&ssi->tx_wd); 312 if (ssi->recv_state == RECV_IDLE) 313 del_timer(&ssi->keep_alive); 314 break; 315 case WAIT4READY: 316 case SENDING: 317 case SENDING_SWBREAK: 318 mod_timer(&ssi->keep_alive, 319 jiffies + msecs_to_jiffies(SSIP_KATOUT)); 320 mod_timer(&ssi->tx_wd, jiffies + msecs_to_jiffies(SSIP_WDTOUT)); 321 break; 322 default: 323 break; 324 } 325 } 326 327 struct hsi_client *ssip_slave_get_master(struct hsi_client *slave) 328 { 329 struct hsi_client *master = ERR_PTR(-ENODEV); 330 struct ssi_protocol *ssi; 331 332 list_for_each_entry(ssi, &ssip_list, link) 333 if (slave->device.parent == ssi->cl->device.parent) { 334 master = ssi->cl; 335 break; 336 } 337 338 return master; 339 } 340 EXPORT_SYMBOL_GPL(ssip_slave_get_master); 341 342 int ssip_slave_start_tx(struct hsi_client *master) 343 { 344 struct ssi_protocol *ssi = hsi_client_drvdata(master); 345 346 dev_dbg(&master->device, "start TX %d\n", atomic_read(&ssi->tx_usecnt)); 347 spin_lock_bh(&ssi->lock); 348 if (ssi->send_state == SEND_IDLE) { 349 ssip_set_txstate(ssi, WAIT4READY); 350 hsi_start_tx(master); 351 } 352 spin_unlock_bh(&ssi->lock); 353 atomic_inc(&ssi->tx_usecnt); 354 355 return 0; 356 } 357 EXPORT_SYMBOL_GPL(ssip_slave_start_tx); 358 359 int ssip_slave_stop_tx(struct hsi_client *master) 360 { 361 struct ssi_protocol *ssi = hsi_client_drvdata(master); 362 363 WARN_ON_ONCE(atomic_read(&ssi->tx_usecnt) == 0); 364 365 if (atomic_dec_and_test(&ssi->tx_usecnt)) { 366 spin_lock_bh(&ssi->lock); 367 if ((ssi->send_state == SEND_READY) || 368 (ssi->send_state == WAIT4READY)) { 369 ssip_set_txstate(ssi, SEND_IDLE); 370 hsi_stop_tx(master); 371 } 372 spin_unlock_bh(&ssi->lock); 373 } 374 dev_dbg(&master->device, "stop TX %d\n", atomic_read(&ssi->tx_usecnt)); 375 376 return 0; 377 } 378 EXPORT_SYMBOL_GPL(ssip_slave_stop_tx); 379 380 int ssip_slave_running(struct hsi_client *master) 381 { 382 struct ssi_protocol *ssi = hsi_client_drvdata(master); 383 return netif_running(ssi->netdev); 384 } 385 EXPORT_SYMBOL_GPL(ssip_slave_running); 386 387 static void ssip_reset(struct hsi_client *cl) 388 { 389 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 390 struct list_head *head, *tmp; 391 struct hsi_msg *msg; 392 393 if (netif_running(ssi->netdev)) 394 netif_carrier_off(ssi->netdev); 395 hsi_flush(cl); 396 spin_lock_bh(&ssi->lock); 397 if (ssi->send_state != SEND_IDLE) 398 hsi_stop_tx(cl); 399 spin_unlock_bh(&ssi->lock); 400 if (test_and_clear_bit(SSIP_WAKETEST_FLAG, &ssi->flags)) 401 ssi_waketest(cl, 0); /* FIXME: To be removed */ 402 spin_lock_bh(&ssi->lock); 403 del_timer(&ssi->rx_wd); 404 del_timer(&ssi->tx_wd); 405 del_timer(&ssi->keep_alive); 406 ssi->main_state = 0; 407 ssi->send_state = 0; 408 ssi->recv_state = 0; 409 ssi->flags = 0; 410 ssi->rxid = 0; 411 ssi->txid = 0; 412 list_for_each_safe(head, tmp, &ssi->txqueue) { 413 msg = list_entry(head, struct hsi_msg, link); 414 dev_dbg(&cl->device, "Pending TX data\n"); 415 list_del(head); 416 ssip_free_data(msg); 417 } 418 ssi->txqueue_len = 0; 419 spin_unlock_bh(&ssi->lock); 420 } 421 422 static void ssip_dump_state(struct hsi_client *cl) 423 { 424 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 425 struct hsi_msg *msg; 426 427 spin_lock_bh(&ssi->lock); 428 dev_err(&cl->device, "Main state: %d\n", ssi->main_state); 429 dev_err(&cl->device, "Recv state: %d\n", ssi->recv_state); 430 dev_err(&cl->device, "Send state: %d\n", ssi->send_state); 431 dev_err(&cl->device, "CMT %s\n", (ssi->main_state == ACTIVE) ? 432 "Online" : "Offline"); 433 dev_err(&cl->device, "Wake test %d\n", 434 test_bit(SSIP_WAKETEST_FLAG, &ssi->flags)); 435 dev_err(&cl->device, "Data RX id: %d\n", ssi->rxid); 436 dev_err(&cl->device, "Data TX id: %d\n", ssi->txid); 437 438 list_for_each_entry(msg, &ssi->txqueue, link) 439 dev_err(&cl->device, "pending TX data (%p)\n", msg); 440 spin_unlock_bh(&ssi->lock); 441 } 442 443 static void ssip_error(struct hsi_client *cl) 444 { 445 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 446 struct hsi_msg *msg; 447 448 ssip_dump_state(cl); 449 ssip_reset(cl); 450 msg = ssip_claim_cmd(ssi); 451 msg->complete = ssip_rxcmd_complete; 452 hsi_async_read(cl, msg); 453 } 454 455 static void ssip_keep_alive(struct timer_list *t) 456 { 457 struct ssi_protocol *ssi = from_timer(ssi, t, keep_alive); 458 struct hsi_client *cl = ssi->cl; 459 460 dev_dbg(&cl->device, "Keep alive kick in: m(%d) r(%d) s(%d)\n", 461 ssi->main_state, ssi->recv_state, ssi->send_state); 462 463 spin_lock(&ssi->lock); 464 if (ssi->recv_state == RECV_IDLE) 465 switch (ssi->send_state) { 466 case SEND_READY: 467 if (atomic_read(&ssi->tx_usecnt) == 0) 468 break; 469 /* Fall through */ 470 /* 471 * Workaround for cmt-speech in that case 472 * we relay on audio timers. 473 */ 474 case SEND_IDLE: 475 spin_unlock(&ssi->lock); 476 return; 477 } 478 mod_timer(&ssi->keep_alive, jiffies + msecs_to_jiffies(SSIP_KATOUT)); 479 spin_unlock(&ssi->lock); 480 } 481 482 static void ssip_rx_wd(struct timer_list *t) 483 { 484 struct ssi_protocol *ssi = from_timer(ssi, t, rx_wd); 485 struct hsi_client *cl = ssi->cl; 486 487 dev_err(&cl->device, "Watchdog triggered\n"); 488 ssip_error(cl); 489 } 490 491 static void ssip_tx_wd(struct timer_list *t) 492 { 493 struct ssi_protocol *ssi = from_timer(ssi, t, tx_wd); 494 struct hsi_client *cl = ssi->cl; 495 496 dev_err(&cl->device, "Watchdog triggered\n"); 497 ssip_error(cl); 498 } 499 500 static void ssip_send_bootinfo_req_cmd(struct hsi_client *cl) 501 { 502 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 503 struct hsi_msg *msg; 504 505 dev_dbg(&cl->device, "Issuing BOOT INFO REQ command\n"); 506 msg = ssip_claim_cmd(ssi); 507 ssip_set_cmd(msg, SSIP_BOOTINFO_REQ_CMD(SSIP_LOCAL_VERID)); 508 msg->complete = ssip_release_cmd; 509 hsi_async_write(cl, msg); 510 dev_dbg(&cl->device, "Issuing RX command\n"); 511 msg = ssip_claim_cmd(ssi); 512 msg->complete = ssip_rxcmd_complete; 513 hsi_async_read(cl, msg); 514 } 515 516 static void ssip_start_rx(struct hsi_client *cl) 517 { 518 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 519 struct hsi_msg *msg; 520 521 dev_dbg(&cl->device, "RX start M(%d) R(%d)\n", ssi->main_state, 522 ssi->recv_state); 523 spin_lock_bh(&ssi->lock); 524 /* 525 * We can have two UP events in a row due to a short low 526 * high transition. Therefore we need to ignore the sencond UP event. 527 */ 528 if ((ssi->main_state != ACTIVE) || (ssi->recv_state == RECV_READY)) { 529 spin_unlock_bh(&ssi->lock); 530 return; 531 } 532 ssip_set_rxstate(ssi, RECV_READY); 533 spin_unlock_bh(&ssi->lock); 534 535 msg = ssip_claim_cmd(ssi); 536 ssip_set_cmd(msg, SSIP_READY_CMD); 537 msg->complete = ssip_release_cmd; 538 dev_dbg(&cl->device, "Send READY\n"); 539 hsi_async_write(cl, msg); 540 } 541 542 static void ssip_stop_rx(struct hsi_client *cl) 543 { 544 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 545 546 dev_dbg(&cl->device, "RX stop M(%d)\n", ssi->main_state); 547 spin_lock_bh(&ssi->lock); 548 if (likely(ssi->main_state == ACTIVE)) 549 ssip_set_rxstate(ssi, RECV_IDLE); 550 spin_unlock_bh(&ssi->lock); 551 } 552 553 static void ssip_free_strans(struct hsi_msg *msg) 554 { 555 ssip_free_data(msg->context); 556 ssip_release_cmd(msg); 557 } 558 559 static void ssip_strans_complete(struct hsi_msg *msg) 560 { 561 struct hsi_client *cl = msg->cl; 562 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 563 struct hsi_msg *data; 564 565 data = msg->context; 566 ssip_release_cmd(msg); 567 spin_lock_bh(&ssi->lock); 568 ssip_set_txstate(ssi, SENDING); 569 spin_unlock_bh(&ssi->lock); 570 hsi_async_write(cl, data); 571 } 572 573 static int ssip_xmit(struct hsi_client *cl) 574 { 575 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 576 struct hsi_msg *msg, *dmsg; 577 struct sk_buff *skb; 578 579 spin_lock_bh(&ssi->lock); 580 if (list_empty(&ssi->txqueue)) { 581 spin_unlock_bh(&ssi->lock); 582 return 0; 583 } 584 dmsg = list_first_entry(&ssi->txqueue, struct hsi_msg, link); 585 list_del(&dmsg->link); 586 ssi->txqueue_len--; 587 spin_unlock_bh(&ssi->lock); 588 589 msg = ssip_claim_cmd(ssi); 590 skb = dmsg->context; 591 msg->context = dmsg; 592 msg->complete = ssip_strans_complete; 593 msg->destructor = ssip_free_strans; 594 595 spin_lock_bh(&ssi->lock); 596 ssip_set_cmd(msg, SSIP_START_TRANS_CMD(SSIP_BYTES_TO_FRAMES(skb->len), 597 ssi->txid)); 598 ssi->txid++; 599 ssip_set_txstate(ssi, SENDING); 600 spin_unlock_bh(&ssi->lock); 601 602 dev_dbg(&cl->device, "Send STRANS (%d frames)\n", 603 SSIP_BYTES_TO_FRAMES(skb->len)); 604 605 return hsi_async_write(cl, msg); 606 } 607 608 /* In soft IRQ context */ 609 static void ssip_pn_rx(struct sk_buff *skb) 610 { 611 struct net_device *dev = skb->dev; 612 613 if (unlikely(!netif_running(dev))) { 614 dev_dbg(&dev->dev, "Drop RX packet\n"); 615 dev->stats.rx_dropped++; 616 dev_kfree_skb(skb); 617 return; 618 } 619 if (unlikely(!pskb_may_pull(skb, SSIP_MIN_PN_HDR))) { 620 dev_dbg(&dev->dev, "Error drop RX packet\n"); 621 dev->stats.rx_errors++; 622 dev->stats.rx_length_errors++; 623 dev_kfree_skb(skb); 624 return; 625 } 626 dev->stats.rx_packets++; 627 dev->stats.rx_bytes += skb->len; 628 629 /* length field is exchanged in network byte order */ 630 ((u16 *)skb->data)[2] = ntohs(((u16 *)skb->data)[2]); 631 dev_dbg(&dev->dev, "RX length fixed (%04x -> %u)\n", 632 ((u16 *)skb->data)[2], ntohs(((u16 *)skb->data)[2])); 633 634 skb->protocol = htons(ETH_P_PHONET); 635 skb_reset_mac_header(skb); 636 __skb_pull(skb, 1); 637 netif_rx(skb); 638 } 639 640 static void ssip_rx_data_complete(struct hsi_msg *msg) 641 { 642 struct hsi_client *cl = msg->cl; 643 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 644 struct sk_buff *skb; 645 646 if (msg->status == HSI_STATUS_ERROR) { 647 dev_err(&cl->device, "RX data error\n"); 648 ssip_free_data(msg); 649 ssip_error(cl); 650 return; 651 } 652 del_timer(&ssi->rx_wd); /* FIXME: Revisit */ 653 skb = msg->context; 654 ssip_pn_rx(skb); 655 hsi_free_msg(msg); 656 } 657 658 static void ssip_rx_bootinforeq(struct hsi_client *cl, u32 cmd) 659 { 660 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 661 struct hsi_msg *msg; 662 663 /* Workaroud: Ignore CMT Loader message leftover */ 664 if (cmd == SSIP_CMT_LOADER_SYNC) 665 return; 666 667 switch (ssi->main_state) { 668 case ACTIVE: 669 dev_err(&cl->device, "Boot info req on active state\n"); 670 ssip_error(cl); 671 /* Fall through */ 672 case INIT: 673 case HANDSHAKE: 674 spin_lock_bh(&ssi->lock); 675 ssi->main_state = HANDSHAKE; 676 spin_unlock_bh(&ssi->lock); 677 678 if (!test_and_set_bit(SSIP_WAKETEST_FLAG, &ssi->flags)) 679 ssi_waketest(cl, 1); /* FIXME: To be removed */ 680 681 spin_lock_bh(&ssi->lock); 682 /* Start boot handshake watchdog */ 683 mod_timer(&ssi->tx_wd, jiffies + msecs_to_jiffies(SSIP_WDTOUT)); 684 spin_unlock_bh(&ssi->lock); 685 dev_dbg(&cl->device, "Send BOOTINFO_RESP\n"); 686 if (SSIP_DATA_VERSION(cmd) != SSIP_LOCAL_VERID) 687 dev_warn(&cl->device, "boot info req verid mismatch\n"); 688 msg = ssip_claim_cmd(ssi); 689 ssip_set_cmd(msg, SSIP_BOOTINFO_RESP_CMD(SSIP_LOCAL_VERID)); 690 msg->complete = ssip_release_cmd; 691 hsi_async_write(cl, msg); 692 break; 693 default: 694 dev_dbg(&cl->device, "Wrong state M(%d)\n", ssi->main_state); 695 break; 696 } 697 } 698 699 static void ssip_rx_bootinforesp(struct hsi_client *cl, u32 cmd) 700 { 701 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 702 703 if (SSIP_DATA_VERSION(cmd) != SSIP_LOCAL_VERID) 704 dev_warn(&cl->device, "boot info resp verid mismatch\n"); 705 706 spin_lock_bh(&ssi->lock); 707 if (ssi->main_state != ACTIVE) 708 /* Use tx_wd as a boot watchdog in non ACTIVE state */ 709 mod_timer(&ssi->tx_wd, jiffies + msecs_to_jiffies(SSIP_WDTOUT)); 710 else 711 dev_dbg(&cl->device, "boot info resp ignored M(%d)\n", 712 ssi->main_state); 713 spin_unlock_bh(&ssi->lock); 714 } 715 716 static void ssip_rx_waketest(struct hsi_client *cl, u32 cmd) 717 { 718 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 719 unsigned int wkres = SSIP_PAYLOAD(cmd); 720 721 spin_lock_bh(&ssi->lock); 722 if (ssi->main_state != HANDSHAKE) { 723 dev_dbg(&cl->device, "wake lines test ignored M(%d)\n", 724 ssi->main_state); 725 spin_unlock_bh(&ssi->lock); 726 return; 727 } 728 spin_unlock_bh(&ssi->lock); 729 730 if (test_and_clear_bit(SSIP_WAKETEST_FLAG, &ssi->flags)) 731 ssi_waketest(cl, 0); /* FIXME: To be removed */ 732 733 spin_lock_bh(&ssi->lock); 734 ssi->main_state = ACTIVE; 735 del_timer(&ssi->tx_wd); /* Stop boot handshake timer */ 736 spin_unlock_bh(&ssi->lock); 737 738 dev_notice(&cl->device, "WAKELINES TEST %s\n", 739 wkres & SSIP_WAKETEST_FAILED ? "FAILED" : "OK"); 740 if (wkres & SSIP_WAKETEST_FAILED) { 741 ssip_error(cl); 742 return; 743 } 744 dev_dbg(&cl->device, "CMT is ONLINE\n"); 745 netif_wake_queue(ssi->netdev); 746 netif_carrier_on(ssi->netdev); 747 } 748 749 static void ssip_rx_ready(struct hsi_client *cl) 750 { 751 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 752 753 spin_lock_bh(&ssi->lock); 754 if (unlikely(ssi->main_state != ACTIVE)) { 755 dev_dbg(&cl->device, "READY on wrong state: S(%d) M(%d)\n", 756 ssi->send_state, ssi->main_state); 757 spin_unlock_bh(&ssi->lock); 758 return; 759 } 760 if (ssi->send_state != WAIT4READY) { 761 dev_dbg(&cl->device, "Ignore spurious READY command\n"); 762 spin_unlock_bh(&ssi->lock); 763 return; 764 } 765 ssip_set_txstate(ssi, SEND_READY); 766 spin_unlock_bh(&ssi->lock); 767 ssip_xmit(cl); 768 } 769 770 static void ssip_rx_strans(struct hsi_client *cl, u32 cmd) 771 { 772 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 773 struct sk_buff *skb; 774 struct hsi_msg *msg; 775 int len = SSIP_PDU_LENGTH(cmd); 776 777 dev_dbg(&cl->device, "RX strans: %d frames\n", len); 778 spin_lock_bh(&ssi->lock); 779 if (unlikely(ssi->main_state != ACTIVE)) { 780 dev_err(&cl->device, "START TRANS wrong state: S(%d) M(%d)\n", 781 ssi->send_state, ssi->main_state); 782 spin_unlock_bh(&ssi->lock); 783 return; 784 } 785 ssip_set_rxstate(ssi, RECEIVING); 786 if (unlikely(SSIP_MSG_ID(cmd) != ssi->rxid)) { 787 dev_err(&cl->device, "START TRANS id %d expected %d\n", 788 SSIP_MSG_ID(cmd), ssi->rxid); 789 spin_unlock_bh(&ssi->lock); 790 goto out1; 791 } 792 ssi->rxid++; 793 spin_unlock_bh(&ssi->lock); 794 skb = netdev_alloc_skb(ssi->netdev, len * 4); 795 if (unlikely(!skb)) { 796 dev_err(&cl->device, "No memory for rx skb\n"); 797 goto out1; 798 } 799 skb->dev = ssi->netdev; 800 skb_put(skb, len * 4); 801 msg = ssip_alloc_data(ssi, skb, GFP_ATOMIC); 802 if (unlikely(!msg)) { 803 dev_err(&cl->device, "No memory for RX data msg\n"); 804 goto out2; 805 } 806 msg->complete = ssip_rx_data_complete; 807 hsi_async_read(cl, msg); 808 809 return; 810 out2: 811 dev_kfree_skb(skb); 812 out1: 813 ssip_error(cl); 814 } 815 816 static void ssip_rxcmd_complete(struct hsi_msg *msg) 817 { 818 struct hsi_client *cl = msg->cl; 819 u32 cmd = ssip_get_cmd(msg); 820 unsigned int cmdid = SSIP_COMMAND(cmd); 821 822 if (msg->status == HSI_STATUS_ERROR) { 823 dev_err(&cl->device, "RX error detected\n"); 824 ssip_release_cmd(msg); 825 ssip_error(cl); 826 return; 827 } 828 hsi_async_read(cl, msg); 829 dev_dbg(&cl->device, "RX cmd: 0x%08x\n", cmd); 830 switch (cmdid) { 831 case SSIP_SW_BREAK: 832 /* Ignored */ 833 break; 834 case SSIP_BOOTINFO_REQ: 835 ssip_rx_bootinforeq(cl, cmd); 836 break; 837 case SSIP_BOOTINFO_RESP: 838 ssip_rx_bootinforesp(cl, cmd); 839 break; 840 case SSIP_WAKETEST_RESULT: 841 ssip_rx_waketest(cl, cmd); 842 break; 843 case SSIP_START_TRANS: 844 ssip_rx_strans(cl, cmd); 845 break; 846 case SSIP_READY: 847 ssip_rx_ready(cl); 848 break; 849 default: 850 dev_warn(&cl->device, "command 0x%08x not supported\n", cmd); 851 break; 852 } 853 } 854 855 static void ssip_swbreak_complete(struct hsi_msg *msg) 856 { 857 struct hsi_client *cl = msg->cl; 858 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 859 860 ssip_release_cmd(msg); 861 spin_lock_bh(&ssi->lock); 862 if (list_empty(&ssi->txqueue)) { 863 if (atomic_read(&ssi->tx_usecnt)) { 864 ssip_set_txstate(ssi, SEND_READY); 865 } else { 866 ssip_set_txstate(ssi, SEND_IDLE); 867 hsi_stop_tx(cl); 868 } 869 spin_unlock_bh(&ssi->lock); 870 } else { 871 spin_unlock_bh(&ssi->lock); 872 ssip_xmit(cl); 873 } 874 netif_wake_queue(ssi->netdev); 875 } 876 877 static void ssip_tx_data_complete(struct hsi_msg *msg) 878 { 879 struct hsi_client *cl = msg->cl; 880 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 881 struct hsi_msg *cmsg; 882 883 if (msg->status == HSI_STATUS_ERROR) { 884 dev_err(&cl->device, "TX data error\n"); 885 ssip_error(cl); 886 goto out; 887 } 888 spin_lock_bh(&ssi->lock); 889 if (list_empty(&ssi->txqueue)) { 890 ssip_set_txstate(ssi, SENDING_SWBREAK); 891 spin_unlock_bh(&ssi->lock); 892 cmsg = ssip_claim_cmd(ssi); 893 ssip_set_cmd(cmsg, SSIP_SWBREAK_CMD); 894 cmsg->complete = ssip_swbreak_complete; 895 dev_dbg(&cl->device, "Send SWBREAK\n"); 896 hsi_async_write(cl, cmsg); 897 } else { 898 spin_unlock_bh(&ssi->lock); 899 ssip_xmit(cl); 900 } 901 out: 902 ssip_free_data(msg); 903 } 904 905 static void ssip_port_event(struct hsi_client *cl, unsigned long event) 906 { 907 switch (event) { 908 case HSI_EVENT_START_RX: 909 ssip_start_rx(cl); 910 break; 911 case HSI_EVENT_STOP_RX: 912 ssip_stop_rx(cl); 913 break; 914 default: 915 return; 916 } 917 } 918 919 static int ssip_pn_open(struct net_device *dev) 920 { 921 struct hsi_client *cl = to_hsi_client(dev->dev.parent); 922 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 923 int err; 924 925 err = hsi_claim_port(cl, 1); 926 if (err < 0) { 927 dev_err(&cl->device, "SSI port already claimed\n"); 928 return err; 929 } 930 err = hsi_register_port_event(cl, ssip_port_event); 931 if (err < 0) { 932 dev_err(&cl->device, "Register HSI port event failed (%d)\n", 933 err); 934 return err; 935 } 936 dev_dbg(&cl->device, "Configuring SSI port\n"); 937 hsi_setup(cl); 938 939 if (!test_and_set_bit(SSIP_WAKETEST_FLAG, &ssi->flags)) 940 ssi_waketest(cl, 1); /* FIXME: To be removed */ 941 942 spin_lock_bh(&ssi->lock); 943 ssi->main_state = HANDSHAKE; 944 spin_unlock_bh(&ssi->lock); 945 946 ssip_send_bootinfo_req_cmd(cl); 947 948 return 0; 949 } 950 951 static int ssip_pn_stop(struct net_device *dev) 952 { 953 struct hsi_client *cl = to_hsi_client(dev->dev.parent); 954 955 ssip_reset(cl); 956 hsi_unregister_port_event(cl); 957 hsi_release_port(cl); 958 959 return 0; 960 } 961 962 static void ssip_xmit_work(struct work_struct *work) 963 { 964 struct ssi_protocol *ssi = 965 container_of(work, struct ssi_protocol, work); 966 struct hsi_client *cl = ssi->cl; 967 968 ssip_xmit(cl); 969 } 970 971 static int ssip_pn_xmit(struct sk_buff *skb, struct net_device *dev) 972 { 973 struct hsi_client *cl = to_hsi_client(dev->dev.parent); 974 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 975 struct hsi_msg *msg; 976 977 if ((skb->protocol != htons(ETH_P_PHONET)) || 978 (skb->len < SSIP_MIN_PN_HDR)) 979 goto drop; 980 /* Pad to 32-bits - FIXME: Revisit*/ 981 if ((skb->len & 3) && skb_pad(skb, 4 - (skb->len & 3))) 982 goto inc_dropped; 983 984 /* 985 * Modem sends Phonet messages over SSI with its own endianness. 986 * Assume that modem has the same endianness as we do. 987 */ 988 if (skb_cow_head(skb, 0)) 989 goto drop; 990 991 /* length field is exchanged in network byte order */ 992 ((u16 *)skb->data)[2] = htons(((u16 *)skb->data)[2]); 993 994 msg = ssip_alloc_data(ssi, skb, GFP_ATOMIC); 995 if (!msg) { 996 dev_dbg(&cl->device, "Dropping tx data: No memory\n"); 997 goto drop; 998 } 999 msg->complete = ssip_tx_data_complete; 1000 1001 spin_lock_bh(&ssi->lock); 1002 if (unlikely(ssi->main_state != ACTIVE)) { 1003 spin_unlock_bh(&ssi->lock); 1004 dev_dbg(&cl->device, "Dropping tx data: CMT is OFFLINE\n"); 1005 goto drop2; 1006 } 1007 list_add_tail(&msg->link, &ssi->txqueue); 1008 ssi->txqueue_len++; 1009 if (dev->tx_queue_len < ssi->txqueue_len) { 1010 dev_info(&cl->device, "TX queue full %d\n", ssi->txqueue_len); 1011 netif_stop_queue(dev); 1012 } 1013 if (ssi->send_state == SEND_IDLE) { 1014 ssip_set_txstate(ssi, WAIT4READY); 1015 spin_unlock_bh(&ssi->lock); 1016 dev_dbg(&cl->device, "Start TX qlen %d\n", ssi->txqueue_len); 1017 hsi_start_tx(cl); 1018 } else if (ssi->send_state == SEND_READY) { 1019 /* Needed for cmt-speech workaround */ 1020 dev_dbg(&cl->device, "Start TX on SEND READY qlen %d\n", 1021 ssi->txqueue_len); 1022 spin_unlock_bh(&ssi->lock); 1023 schedule_work(&ssi->work); 1024 } else { 1025 spin_unlock_bh(&ssi->lock); 1026 } 1027 dev->stats.tx_packets++; 1028 dev->stats.tx_bytes += skb->len; 1029 1030 return 0; 1031 drop2: 1032 hsi_free_msg(msg); 1033 drop: 1034 dev_kfree_skb(skb); 1035 inc_dropped: 1036 dev->stats.tx_dropped++; 1037 1038 return 0; 1039 } 1040 1041 /* CMT reset event handler */ 1042 void ssip_reset_event(struct hsi_client *master) 1043 { 1044 struct ssi_protocol *ssi = hsi_client_drvdata(master); 1045 dev_err(&ssi->cl->device, "CMT reset detected!\n"); 1046 ssip_error(ssi->cl); 1047 } 1048 EXPORT_SYMBOL_GPL(ssip_reset_event); 1049 1050 static const struct net_device_ops ssip_pn_ops = { 1051 .ndo_open = ssip_pn_open, 1052 .ndo_stop = ssip_pn_stop, 1053 .ndo_start_xmit = ssip_pn_xmit, 1054 }; 1055 1056 static void ssip_pn_setup(struct net_device *dev) 1057 { 1058 dev->features = 0; 1059 dev->netdev_ops = &ssip_pn_ops; 1060 dev->type = ARPHRD_PHONET; 1061 dev->flags = IFF_POINTOPOINT | IFF_NOARP; 1062 dev->mtu = SSIP_DEFAULT_MTU; 1063 dev->hard_header_len = 1; 1064 dev->dev_addr[0] = PN_MEDIA_SOS; 1065 dev->addr_len = 1; 1066 dev->tx_queue_len = SSIP_TXQUEUE_LEN; 1067 1068 dev->needs_free_netdev = true; 1069 dev->header_ops = &phonet_header_ops; 1070 } 1071 1072 static int ssi_protocol_probe(struct device *dev) 1073 { 1074 static const char ifname[] = "phonet%d"; 1075 struct hsi_client *cl = to_hsi_client(dev); 1076 struct ssi_protocol *ssi; 1077 int err; 1078 1079 ssi = kzalloc(sizeof(*ssi), GFP_KERNEL); 1080 if (!ssi) 1081 return -ENOMEM; 1082 1083 spin_lock_init(&ssi->lock); 1084 timer_setup(&ssi->rx_wd, ssip_rx_wd, TIMER_DEFERRABLE); 1085 timer_setup(&ssi->tx_wd, ssip_tx_wd, TIMER_DEFERRABLE); 1086 timer_setup(&ssi->keep_alive, ssip_keep_alive, 0); 1087 INIT_LIST_HEAD(&ssi->txqueue); 1088 INIT_LIST_HEAD(&ssi->cmdqueue); 1089 atomic_set(&ssi->tx_usecnt, 0); 1090 hsi_client_set_drvdata(cl, ssi); 1091 ssi->cl = cl; 1092 INIT_WORK(&ssi->work, ssip_xmit_work); 1093 1094 ssi->channel_id_cmd = hsi_get_channel_id_by_name(cl, "mcsaab-control"); 1095 if (ssi->channel_id_cmd < 0) { 1096 err = ssi->channel_id_cmd; 1097 dev_err(dev, "Could not get cmd channel (%d)\n", err); 1098 goto out; 1099 } 1100 1101 ssi->channel_id_data = hsi_get_channel_id_by_name(cl, "mcsaab-data"); 1102 if (ssi->channel_id_data < 0) { 1103 err = ssi->channel_id_data; 1104 dev_err(dev, "Could not get data channel (%d)\n", err); 1105 goto out; 1106 } 1107 1108 err = ssip_alloc_cmds(ssi); 1109 if (err < 0) { 1110 dev_err(dev, "No memory for commands\n"); 1111 goto out; 1112 } 1113 1114 ssi->netdev = alloc_netdev(0, ifname, NET_NAME_UNKNOWN, ssip_pn_setup); 1115 if (!ssi->netdev) { 1116 dev_err(dev, "No memory for netdev\n"); 1117 err = -ENOMEM; 1118 goto out1; 1119 } 1120 1121 /* MTU range: 6 - 65535 */ 1122 ssi->netdev->min_mtu = PHONET_MIN_MTU; 1123 ssi->netdev->max_mtu = SSIP_MAX_MTU; 1124 1125 SET_NETDEV_DEV(ssi->netdev, dev); 1126 netif_carrier_off(ssi->netdev); 1127 err = register_netdev(ssi->netdev); 1128 if (err < 0) { 1129 dev_err(dev, "Register netdev failed (%d)\n", err); 1130 goto out2; 1131 } 1132 1133 list_add(&ssi->link, &ssip_list); 1134 1135 dev_dbg(dev, "channel configuration: cmd=%d, data=%d\n", 1136 ssi->channel_id_cmd, ssi->channel_id_data); 1137 1138 return 0; 1139 out2: 1140 free_netdev(ssi->netdev); 1141 out1: 1142 ssip_free_cmds(ssi); 1143 out: 1144 kfree(ssi); 1145 1146 return err; 1147 } 1148 1149 static int ssi_protocol_remove(struct device *dev) 1150 { 1151 struct hsi_client *cl = to_hsi_client(dev); 1152 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 1153 1154 list_del(&ssi->link); 1155 unregister_netdev(ssi->netdev); 1156 ssip_free_cmds(ssi); 1157 hsi_client_set_drvdata(cl, NULL); 1158 kfree(ssi); 1159 1160 return 0; 1161 } 1162 1163 static struct hsi_client_driver ssip_driver = { 1164 .driver = { 1165 .name = "ssi-protocol", 1166 .owner = THIS_MODULE, 1167 .probe = ssi_protocol_probe, 1168 .remove = ssi_protocol_remove, 1169 }, 1170 }; 1171 1172 static int __init ssip_init(void) 1173 { 1174 pr_info("SSI protocol aka McSAAB added\n"); 1175 1176 return hsi_register_client_driver(&ssip_driver); 1177 } 1178 module_init(ssip_init); 1179 1180 static void __exit ssip_exit(void) 1181 { 1182 hsi_unregister_client_driver(&ssip_driver); 1183 pr_info("SSI protocol driver removed\n"); 1184 } 1185 module_exit(ssip_exit); 1186 1187 MODULE_ALIAS("hsi:ssi-protocol"); 1188 MODULE_AUTHOR("Carlos Chinea <carlos.chinea@nokia.com>"); 1189 MODULE_AUTHOR("Remi Denis-Courmont <remi.denis-courmont@nokia.com>"); 1190 MODULE_DESCRIPTION("SSI protocol improved aka McSAAB"); 1191 MODULE_LICENSE("GPL"); 1192