1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * ssi_protocol.c 4 * 5 * Implementation of the SSI McSAAB improved protocol. 6 * 7 * Copyright (C) 2010 Nokia Corporation. All rights reserved. 8 * Copyright (C) 2013 Sebastian Reichel <sre@kernel.org> 9 * 10 * Contact: Carlos Chinea <carlos.chinea@nokia.com> 11 */ 12 13 #include <linux/atomic.h> 14 #include <linux/clk.h> 15 #include <linux/device.h> 16 #include <linux/err.h> 17 #include <linux/gpio.h> 18 #include <linux/if_ether.h> 19 #include <linux/if_arp.h> 20 #include <linux/if_phonet.h> 21 #include <linux/init.h> 22 #include <linux/irq.h> 23 #include <linux/list.h> 24 #include <linux/module.h> 25 #include <linux/netdevice.h> 26 #include <linux/notifier.h> 27 #include <linux/scatterlist.h> 28 #include <linux/skbuff.h> 29 #include <linux/slab.h> 30 #include <linux/spinlock.h> 31 #include <linux/timer.h> 32 #include <linux/hsi/hsi.h> 33 #include <linux/hsi/ssi_protocol.h> 34 35 void ssi_waketest(struct hsi_client *cl, unsigned int enable); 36 37 #define SSIP_TXQUEUE_LEN 100 38 #define SSIP_MAX_MTU 65535 39 #define SSIP_DEFAULT_MTU 4000 40 #define PN_MEDIA_SOS 21 41 #define SSIP_MIN_PN_HDR 6 /* FIXME: Revisit */ 42 #define SSIP_WDTOUT 2000 /* FIXME: has to be 500 msecs */ 43 #define SSIP_KATOUT 15 /* 15 msecs */ 44 #define SSIP_MAX_CMDS 5 /* Number of pre-allocated commands buffers */ 45 #define SSIP_BYTES_TO_FRAMES(x) ((((x) - 1) >> 2) + 1) 46 #define SSIP_CMT_LOADER_SYNC 0x11223344 47 /* 48 * SSI protocol command definitions 49 */ 50 #define SSIP_COMMAND(data) ((data) >> 28) 51 #define SSIP_PAYLOAD(data) ((data) & 0xfffffff) 52 /* Commands */ 53 #define SSIP_SW_BREAK 0 54 #define SSIP_BOOTINFO_REQ 1 55 #define SSIP_BOOTINFO_RESP 2 56 #define SSIP_WAKETEST_RESULT 3 57 #define SSIP_START_TRANS 4 58 #define SSIP_READY 5 59 /* Payloads */ 60 #define SSIP_DATA_VERSION(data) ((data) & 0xff) 61 #define SSIP_LOCAL_VERID 1 62 #define SSIP_WAKETEST_OK 0 63 #define SSIP_WAKETEST_FAILED 1 64 #define SSIP_PDU_LENGTH(data) (((data) >> 8) & 0xffff) 65 #define SSIP_MSG_ID(data) ((data) & 0xff) 66 /* Generic Command */ 67 #define SSIP_CMD(cmd, payload) (((cmd) << 28) | ((payload) & 0xfffffff)) 68 /* Commands for the control channel */ 69 #define SSIP_BOOTINFO_REQ_CMD(ver) \ 70 SSIP_CMD(SSIP_BOOTINFO_REQ, SSIP_DATA_VERSION(ver)) 71 #define SSIP_BOOTINFO_RESP_CMD(ver) \ 72 SSIP_CMD(SSIP_BOOTINFO_RESP, SSIP_DATA_VERSION(ver)) 73 #define SSIP_START_TRANS_CMD(pdulen, id) \ 74 SSIP_CMD(SSIP_START_TRANS, (((pdulen) << 8) | SSIP_MSG_ID(id))) 75 #define SSIP_READY_CMD SSIP_CMD(SSIP_READY, 0) 76 #define SSIP_SWBREAK_CMD SSIP_CMD(SSIP_SW_BREAK, 0) 77 78 #define SSIP_WAKETEST_FLAG 0 79 80 /* Main state machine states */ 81 enum { 82 INIT, 83 HANDSHAKE, 84 ACTIVE, 85 }; 86 87 /* Send state machine states */ 88 enum { 89 SEND_IDLE, 90 WAIT4READY, 91 SEND_READY, 92 SENDING, 93 SENDING_SWBREAK, 94 }; 95 96 /* Receive state machine states */ 97 enum { 98 RECV_IDLE, 99 RECV_READY, 100 RECEIVING, 101 }; 102 103 /** 104 * struct ssi_protocol - SSI protocol (McSAAB) data 105 * @main_state: Main state machine 106 * @send_state: TX state machine 107 * @recv_state: RX state machine 108 * @flags: Flags, currently only used to follow wake line test 109 * @rxid: RX data id 110 * @txid: TX data id 111 * @txqueue_len: TX queue length 112 * @tx_wd: TX watchdog 113 * @rx_wd: RX watchdog 114 * @keep_alive: Workaround for SSI HW bug 115 * @lock: To serialize access to this struct 116 * @netdev: Phonet network device 117 * @txqueue: TX data queue 118 * @cmdqueue: Queue of free commands 119 * @cl: HSI client own reference 120 * @link: Link for ssip_list 121 * @tx_usecount: Refcount to keep track the slaves that use the wake line 122 * @channel_id_cmd: HSI channel id for command stream 123 * @channel_id_data: HSI channel id for data stream 124 */ 125 struct ssi_protocol { 126 unsigned int main_state; 127 unsigned int send_state; 128 unsigned int recv_state; 129 unsigned long flags; 130 u8 rxid; 131 u8 txid; 132 unsigned int txqueue_len; 133 struct timer_list tx_wd; 134 struct timer_list rx_wd; 135 struct timer_list keep_alive; /* wake-up workaround */ 136 spinlock_t lock; 137 struct net_device *netdev; 138 struct list_head txqueue; 139 struct list_head cmdqueue; 140 struct work_struct work; 141 struct hsi_client *cl; 142 struct list_head link; 143 atomic_t tx_usecnt; 144 int channel_id_cmd; 145 int channel_id_data; 146 }; 147 148 /* List of ssi protocol instances */ 149 static LIST_HEAD(ssip_list); 150 151 static void ssip_rxcmd_complete(struct hsi_msg *msg); 152 153 static inline void ssip_set_cmd(struct hsi_msg *msg, u32 cmd) 154 { 155 u32 *data; 156 157 data = sg_virt(msg->sgt.sgl); 158 *data = cmd; 159 } 160 161 static inline u32 ssip_get_cmd(struct hsi_msg *msg) 162 { 163 u32 *data; 164 165 data = sg_virt(msg->sgt.sgl); 166 167 return *data; 168 } 169 170 static void ssip_skb_to_msg(struct sk_buff *skb, struct hsi_msg *msg) 171 { 172 skb_frag_t *frag; 173 struct scatterlist *sg; 174 int i; 175 176 BUG_ON(msg->sgt.nents != (unsigned int)(skb_shinfo(skb)->nr_frags + 1)); 177 178 sg = msg->sgt.sgl; 179 sg_set_buf(sg, skb->data, skb_headlen(skb)); 180 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { 181 sg = sg_next(sg); 182 BUG_ON(!sg); 183 frag = &skb_shinfo(skb)->frags[i]; 184 sg_set_page(sg, skb_frag_page(frag), skb_frag_size(frag), 185 skb_frag_off(frag)); 186 } 187 } 188 189 static void ssip_free_data(struct hsi_msg *msg) 190 { 191 struct sk_buff *skb; 192 193 skb = msg->context; 194 pr_debug("free data: msg %p context %p skb %p\n", msg, msg->context, 195 skb); 196 msg->destructor = NULL; 197 dev_kfree_skb(skb); 198 hsi_free_msg(msg); 199 } 200 201 static struct hsi_msg *ssip_alloc_data(struct ssi_protocol *ssi, 202 struct sk_buff *skb, gfp_t flags) 203 { 204 struct hsi_msg *msg; 205 206 msg = hsi_alloc_msg(skb_shinfo(skb)->nr_frags + 1, flags); 207 if (!msg) 208 return NULL; 209 ssip_skb_to_msg(skb, msg); 210 msg->destructor = ssip_free_data; 211 msg->channel = ssi->channel_id_data; 212 msg->context = skb; 213 214 return msg; 215 } 216 217 static inline void ssip_release_cmd(struct hsi_msg *msg) 218 { 219 struct ssi_protocol *ssi = hsi_client_drvdata(msg->cl); 220 221 dev_dbg(&msg->cl->device, "Release cmd 0x%08x\n", ssip_get_cmd(msg)); 222 spin_lock_bh(&ssi->lock); 223 list_add_tail(&msg->link, &ssi->cmdqueue); 224 spin_unlock_bh(&ssi->lock); 225 } 226 227 static struct hsi_msg *ssip_claim_cmd(struct ssi_protocol *ssi) 228 { 229 struct hsi_msg *msg; 230 231 BUG_ON(list_empty(&ssi->cmdqueue)); 232 233 spin_lock_bh(&ssi->lock); 234 msg = list_first_entry(&ssi->cmdqueue, struct hsi_msg, link); 235 list_del(&msg->link); 236 spin_unlock_bh(&ssi->lock); 237 msg->destructor = ssip_release_cmd; 238 239 return msg; 240 } 241 242 static void ssip_free_cmds(struct ssi_protocol *ssi) 243 { 244 struct hsi_msg *msg, *tmp; 245 246 list_for_each_entry_safe(msg, tmp, &ssi->cmdqueue, link) { 247 list_del(&msg->link); 248 msg->destructor = NULL; 249 kfree(sg_virt(msg->sgt.sgl)); 250 hsi_free_msg(msg); 251 } 252 } 253 254 static int ssip_alloc_cmds(struct ssi_protocol *ssi) 255 { 256 struct hsi_msg *msg; 257 u32 *buf; 258 unsigned int i; 259 260 for (i = 0; i < SSIP_MAX_CMDS; i++) { 261 msg = hsi_alloc_msg(1, GFP_KERNEL); 262 if (!msg) 263 goto out; 264 buf = kmalloc(sizeof(*buf), GFP_KERNEL); 265 if (!buf) { 266 hsi_free_msg(msg); 267 goto out; 268 } 269 sg_init_one(msg->sgt.sgl, buf, sizeof(*buf)); 270 msg->channel = ssi->channel_id_cmd; 271 list_add_tail(&msg->link, &ssi->cmdqueue); 272 } 273 274 return 0; 275 out: 276 ssip_free_cmds(ssi); 277 278 return -ENOMEM; 279 } 280 281 static void ssip_set_rxstate(struct ssi_protocol *ssi, unsigned int state) 282 { 283 ssi->recv_state = state; 284 switch (state) { 285 case RECV_IDLE: 286 del_timer(&ssi->rx_wd); 287 if (ssi->send_state == SEND_IDLE) 288 del_timer(&ssi->keep_alive); 289 break; 290 case RECV_READY: 291 /* CMT speech workaround */ 292 if (atomic_read(&ssi->tx_usecnt)) 293 break; 294 /* Otherwise fall through */ 295 case RECEIVING: 296 mod_timer(&ssi->keep_alive, jiffies + 297 msecs_to_jiffies(SSIP_KATOUT)); 298 mod_timer(&ssi->rx_wd, jiffies + msecs_to_jiffies(SSIP_WDTOUT)); 299 break; 300 default: 301 break; 302 } 303 } 304 305 static void ssip_set_txstate(struct ssi_protocol *ssi, unsigned int state) 306 { 307 ssi->send_state = state; 308 switch (state) { 309 case SEND_IDLE: 310 case SEND_READY: 311 del_timer(&ssi->tx_wd); 312 if (ssi->recv_state == RECV_IDLE) 313 del_timer(&ssi->keep_alive); 314 break; 315 case WAIT4READY: 316 case SENDING: 317 case SENDING_SWBREAK: 318 mod_timer(&ssi->keep_alive, 319 jiffies + msecs_to_jiffies(SSIP_KATOUT)); 320 mod_timer(&ssi->tx_wd, jiffies + msecs_to_jiffies(SSIP_WDTOUT)); 321 break; 322 default: 323 break; 324 } 325 } 326 327 struct hsi_client *ssip_slave_get_master(struct hsi_client *slave) 328 { 329 struct hsi_client *master = ERR_PTR(-ENODEV); 330 struct ssi_protocol *ssi; 331 332 list_for_each_entry(ssi, &ssip_list, link) 333 if (slave->device.parent == ssi->cl->device.parent) { 334 master = ssi->cl; 335 break; 336 } 337 338 return master; 339 } 340 EXPORT_SYMBOL_GPL(ssip_slave_get_master); 341 342 int ssip_slave_start_tx(struct hsi_client *master) 343 { 344 struct ssi_protocol *ssi = hsi_client_drvdata(master); 345 346 dev_dbg(&master->device, "start TX %d\n", atomic_read(&ssi->tx_usecnt)); 347 spin_lock_bh(&ssi->lock); 348 if (ssi->send_state == SEND_IDLE) { 349 ssip_set_txstate(ssi, WAIT4READY); 350 hsi_start_tx(master); 351 } 352 spin_unlock_bh(&ssi->lock); 353 atomic_inc(&ssi->tx_usecnt); 354 355 return 0; 356 } 357 EXPORT_SYMBOL_GPL(ssip_slave_start_tx); 358 359 int ssip_slave_stop_tx(struct hsi_client *master) 360 { 361 struct ssi_protocol *ssi = hsi_client_drvdata(master); 362 363 WARN_ON_ONCE(atomic_read(&ssi->tx_usecnt) == 0); 364 365 if (atomic_dec_and_test(&ssi->tx_usecnt)) { 366 spin_lock_bh(&ssi->lock); 367 if ((ssi->send_state == SEND_READY) || 368 (ssi->send_state == WAIT4READY)) { 369 ssip_set_txstate(ssi, SEND_IDLE); 370 hsi_stop_tx(master); 371 } 372 spin_unlock_bh(&ssi->lock); 373 } 374 dev_dbg(&master->device, "stop TX %d\n", atomic_read(&ssi->tx_usecnt)); 375 376 return 0; 377 } 378 EXPORT_SYMBOL_GPL(ssip_slave_stop_tx); 379 380 int ssip_slave_running(struct hsi_client *master) 381 { 382 struct ssi_protocol *ssi = hsi_client_drvdata(master); 383 return netif_running(ssi->netdev); 384 } 385 EXPORT_SYMBOL_GPL(ssip_slave_running); 386 387 static void ssip_reset(struct hsi_client *cl) 388 { 389 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 390 struct list_head *head, *tmp; 391 struct hsi_msg *msg; 392 393 if (netif_running(ssi->netdev)) 394 netif_carrier_off(ssi->netdev); 395 hsi_flush(cl); 396 spin_lock_bh(&ssi->lock); 397 if (ssi->send_state != SEND_IDLE) 398 hsi_stop_tx(cl); 399 spin_unlock_bh(&ssi->lock); 400 if (test_and_clear_bit(SSIP_WAKETEST_FLAG, &ssi->flags)) 401 ssi_waketest(cl, 0); /* FIXME: To be removed */ 402 spin_lock_bh(&ssi->lock); 403 del_timer(&ssi->rx_wd); 404 del_timer(&ssi->tx_wd); 405 del_timer(&ssi->keep_alive); 406 ssi->main_state = 0; 407 ssi->send_state = 0; 408 ssi->recv_state = 0; 409 ssi->flags = 0; 410 ssi->rxid = 0; 411 ssi->txid = 0; 412 list_for_each_safe(head, tmp, &ssi->txqueue) { 413 msg = list_entry(head, struct hsi_msg, link); 414 dev_dbg(&cl->device, "Pending TX data\n"); 415 list_del(head); 416 ssip_free_data(msg); 417 } 418 ssi->txqueue_len = 0; 419 spin_unlock_bh(&ssi->lock); 420 } 421 422 static void ssip_dump_state(struct hsi_client *cl) 423 { 424 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 425 struct hsi_msg *msg; 426 427 spin_lock_bh(&ssi->lock); 428 dev_err(&cl->device, "Main state: %d\n", ssi->main_state); 429 dev_err(&cl->device, "Recv state: %d\n", ssi->recv_state); 430 dev_err(&cl->device, "Send state: %d\n", ssi->send_state); 431 dev_err(&cl->device, "CMT %s\n", (ssi->main_state == ACTIVE) ? 432 "Online" : "Offline"); 433 dev_err(&cl->device, "Wake test %d\n", 434 test_bit(SSIP_WAKETEST_FLAG, &ssi->flags)); 435 dev_err(&cl->device, "Data RX id: %d\n", ssi->rxid); 436 dev_err(&cl->device, "Data TX id: %d\n", ssi->txid); 437 438 list_for_each_entry(msg, &ssi->txqueue, link) 439 dev_err(&cl->device, "pending TX data (%p)\n", msg); 440 spin_unlock_bh(&ssi->lock); 441 } 442 443 static void ssip_error(struct hsi_client *cl) 444 { 445 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 446 struct hsi_msg *msg; 447 448 ssip_dump_state(cl); 449 ssip_reset(cl); 450 msg = ssip_claim_cmd(ssi); 451 msg->complete = ssip_rxcmd_complete; 452 hsi_async_read(cl, msg); 453 } 454 455 static void ssip_keep_alive(struct timer_list *t) 456 { 457 struct ssi_protocol *ssi = from_timer(ssi, t, keep_alive); 458 struct hsi_client *cl = ssi->cl; 459 460 dev_dbg(&cl->device, "Keep alive kick in: m(%d) r(%d) s(%d)\n", 461 ssi->main_state, ssi->recv_state, ssi->send_state); 462 463 spin_lock(&ssi->lock); 464 if (ssi->recv_state == RECV_IDLE) 465 switch (ssi->send_state) { 466 case SEND_READY: 467 if (atomic_read(&ssi->tx_usecnt) == 0) 468 break; 469 /* 470 * Fall through. Workaround for cmt-speech 471 * in that case we relay on audio timers. 472 */ 473 case SEND_IDLE: 474 spin_unlock(&ssi->lock); 475 return; 476 } 477 mod_timer(&ssi->keep_alive, jiffies + msecs_to_jiffies(SSIP_KATOUT)); 478 spin_unlock(&ssi->lock); 479 } 480 481 static void ssip_rx_wd(struct timer_list *t) 482 { 483 struct ssi_protocol *ssi = from_timer(ssi, t, rx_wd); 484 struct hsi_client *cl = ssi->cl; 485 486 dev_err(&cl->device, "Watchdog triggered\n"); 487 ssip_error(cl); 488 } 489 490 static void ssip_tx_wd(struct timer_list *t) 491 { 492 struct ssi_protocol *ssi = from_timer(ssi, t, tx_wd); 493 struct hsi_client *cl = ssi->cl; 494 495 dev_err(&cl->device, "Watchdog triggered\n"); 496 ssip_error(cl); 497 } 498 499 static void ssip_send_bootinfo_req_cmd(struct hsi_client *cl) 500 { 501 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 502 struct hsi_msg *msg; 503 504 dev_dbg(&cl->device, "Issuing BOOT INFO REQ command\n"); 505 msg = ssip_claim_cmd(ssi); 506 ssip_set_cmd(msg, SSIP_BOOTINFO_REQ_CMD(SSIP_LOCAL_VERID)); 507 msg->complete = ssip_release_cmd; 508 hsi_async_write(cl, msg); 509 dev_dbg(&cl->device, "Issuing RX command\n"); 510 msg = ssip_claim_cmd(ssi); 511 msg->complete = ssip_rxcmd_complete; 512 hsi_async_read(cl, msg); 513 } 514 515 static void ssip_start_rx(struct hsi_client *cl) 516 { 517 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 518 struct hsi_msg *msg; 519 520 dev_dbg(&cl->device, "RX start M(%d) R(%d)\n", ssi->main_state, 521 ssi->recv_state); 522 spin_lock_bh(&ssi->lock); 523 /* 524 * We can have two UP events in a row due to a short low 525 * high transition. Therefore we need to ignore the sencond UP event. 526 */ 527 if ((ssi->main_state != ACTIVE) || (ssi->recv_state == RECV_READY)) { 528 spin_unlock_bh(&ssi->lock); 529 return; 530 } 531 ssip_set_rxstate(ssi, RECV_READY); 532 spin_unlock_bh(&ssi->lock); 533 534 msg = ssip_claim_cmd(ssi); 535 ssip_set_cmd(msg, SSIP_READY_CMD); 536 msg->complete = ssip_release_cmd; 537 dev_dbg(&cl->device, "Send READY\n"); 538 hsi_async_write(cl, msg); 539 } 540 541 static void ssip_stop_rx(struct hsi_client *cl) 542 { 543 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 544 545 dev_dbg(&cl->device, "RX stop M(%d)\n", ssi->main_state); 546 spin_lock_bh(&ssi->lock); 547 if (likely(ssi->main_state == ACTIVE)) 548 ssip_set_rxstate(ssi, RECV_IDLE); 549 spin_unlock_bh(&ssi->lock); 550 } 551 552 static void ssip_free_strans(struct hsi_msg *msg) 553 { 554 ssip_free_data(msg->context); 555 ssip_release_cmd(msg); 556 } 557 558 static void ssip_strans_complete(struct hsi_msg *msg) 559 { 560 struct hsi_client *cl = msg->cl; 561 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 562 struct hsi_msg *data; 563 564 data = msg->context; 565 ssip_release_cmd(msg); 566 spin_lock_bh(&ssi->lock); 567 ssip_set_txstate(ssi, SENDING); 568 spin_unlock_bh(&ssi->lock); 569 hsi_async_write(cl, data); 570 } 571 572 static int ssip_xmit(struct hsi_client *cl) 573 { 574 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 575 struct hsi_msg *msg, *dmsg; 576 struct sk_buff *skb; 577 578 spin_lock_bh(&ssi->lock); 579 if (list_empty(&ssi->txqueue)) { 580 spin_unlock_bh(&ssi->lock); 581 return 0; 582 } 583 dmsg = list_first_entry(&ssi->txqueue, struct hsi_msg, link); 584 list_del(&dmsg->link); 585 ssi->txqueue_len--; 586 spin_unlock_bh(&ssi->lock); 587 588 msg = ssip_claim_cmd(ssi); 589 skb = dmsg->context; 590 msg->context = dmsg; 591 msg->complete = ssip_strans_complete; 592 msg->destructor = ssip_free_strans; 593 594 spin_lock_bh(&ssi->lock); 595 ssip_set_cmd(msg, SSIP_START_TRANS_CMD(SSIP_BYTES_TO_FRAMES(skb->len), 596 ssi->txid)); 597 ssi->txid++; 598 ssip_set_txstate(ssi, SENDING); 599 spin_unlock_bh(&ssi->lock); 600 601 dev_dbg(&cl->device, "Send STRANS (%d frames)\n", 602 SSIP_BYTES_TO_FRAMES(skb->len)); 603 604 return hsi_async_write(cl, msg); 605 } 606 607 /* In soft IRQ context */ 608 static void ssip_pn_rx(struct sk_buff *skb) 609 { 610 struct net_device *dev = skb->dev; 611 612 if (unlikely(!netif_running(dev))) { 613 dev_dbg(&dev->dev, "Drop RX packet\n"); 614 dev->stats.rx_dropped++; 615 dev_kfree_skb(skb); 616 return; 617 } 618 if (unlikely(!pskb_may_pull(skb, SSIP_MIN_PN_HDR))) { 619 dev_dbg(&dev->dev, "Error drop RX packet\n"); 620 dev->stats.rx_errors++; 621 dev->stats.rx_length_errors++; 622 dev_kfree_skb(skb); 623 return; 624 } 625 dev->stats.rx_packets++; 626 dev->stats.rx_bytes += skb->len; 627 628 /* length field is exchanged in network byte order */ 629 ((u16 *)skb->data)[2] = ntohs(((u16 *)skb->data)[2]); 630 dev_dbg(&dev->dev, "RX length fixed (%04x -> %u)\n", 631 ((u16 *)skb->data)[2], ntohs(((u16 *)skb->data)[2])); 632 633 skb->protocol = htons(ETH_P_PHONET); 634 skb_reset_mac_header(skb); 635 __skb_pull(skb, 1); 636 netif_rx(skb); 637 } 638 639 static void ssip_rx_data_complete(struct hsi_msg *msg) 640 { 641 struct hsi_client *cl = msg->cl; 642 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 643 struct sk_buff *skb; 644 645 if (msg->status == HSI_STATUS_ERROR) { 646 dev_err(&cl->device, "RX data error\n"); 647 ssip_free_data(msg); 648 ssip_error(cl); 649 return; 650 } 651 del_timer(&ssi->rx_wd); /* FIXME: Revisit */ 652 skb = msg->context; 653 ssip_pn_rx(skb); 654 hsi_free_msg(msg); 655 } 656 657 static void ssip_rx_bootinforeq(struct hsi_client *cl, u32 cmd) 658 { 659 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 660 struct hsi_msg *msg; 661 662 /* Workaroud: Ignore CMT Loader message leftover */ 663 if (cmd == SSIP_CMT_LOADER_SYNC) 664 return; 665 666 switch (ssi->main_state) { 667 case ACTIVE: 668 dev_err(&cl->device, "Boot info req on active state\n"); 669 ssip_error(cl); 670 /* Fall through */ 671 case INIT: 672 case HANDSHAKE: 673 spin_lock_bh(&ssi->lock); 674 ssi->main_state = HANDSHAKE; 675 spin_unlock_bh(&ssi->lock); 676 677 if (!test_and_set_bit(SSIP_WAKETEST_FLAG, &ssi->flags)) 678 ssi_waketest(cl, 1); /* FIXME: To be removed */ 679 680 spin_lock_bh(&ssi->lock); 681 /* Start boot handshake watchdog */ 682 mod_timer(&ssi->tx_wd, jiffies + msecs_to_jiffies(SSIP_WDTOUT)); 683 spin_unlock_bh(&ssi->lock); 684 dev_dbg(&cl->device, "Send BOOTINFO_RESP\n"); 685 if (SSIP_DATA_VERSION(cmd) != SSIP_LOCAL_VERID) 686 dev_warn(&cl->device, "boot info req verid mismatch\n"); 687 msg = ssip_claim_cmd(ssi); 688 ssip_set_cmd(msg, SSIP_BOOTINFO_RESP_CMD(SSIP_LOCAL_VERID)); 689 msg->complete = ssip_release_cmd; 690 hsi_async_write(cl, msg); 691 break; 692 default: 693 dev_dbg(&cl->device, "Wrong state M(%d)\n", ssi->main_state); 694 break; 695 } 696 } 697 698 static void ssip_rx_bootinforesp(struct hsi_client *cl, u32 cmd) 699 { 700 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 701 702 if (SSIP_DATA_VERSION(cmd) != SSIP_LOCAL_VERID) 703 dev_warn(&cl->device, "boot info resp verid mismatch\n"); 704 705 spin_lock_bh(&ssi->lock); 706 if (ssi->main_state != ACTIVE) 707 /* Use tx_wd as a boot watchdog in non ACTIVE state */ 708 mod_timer(&ssi->tx_wd, jiffies + msecs_to_jiffies(SSIP_WDTOUT)); 709 else 710 dev_dbg(&cl->device, "boot info resp ignored M(%d)\n", 711 ssi->main_state); 712 spin_unlock_bh(&ssi->lock); 713 } 714 715 static void ssip_rx_waketest(struct hsi_client *cl, u32 cmd) 716 { 717 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 718 unsigned int wkres = SSIP_PAYLOAD(cmd); 719 720 spin_lock_bh(&ssi->lock); 721 if (ssi->main_state != HANDSHAKE) { 722 dev_dbg(&cl->device, "wake lines test ignored M(%d)\n", 723 ssi->main_state); 724 spin_unlock_bh(&ssi->lock); 725 return; 726 } 727 spin_unlock_bh(&ssi->lock); 728 729 if (test_and_clear_bit(SSIP_WAKETEST_FLAG, &ssi->flags)) 730 ssi_waketest(cl, 0); /* FIXME: To be removed */ 731 732 spin_lock_bh(&ssi->lock); 733 ssi->main_state = ACTIVE; 734 del_timer(&ssi->tx_wd); /* Stop boot handshake timer */ 735 spin_unlock_bh(&ssi->lock); 736 737 dev_notice(&cl->device, "WAKELINES TEST %s\n", 738 wkres & SSIP_WAKETEST_FAILED ? "FAILED" : "OK"); 739 if (wkres & SSIP_WAKETEST_FAILED) { 740 ssip_error(cl); 741 return; 742 } 743 dev_dbg(&cl->device, "CMT is ONLINE\n"); 744 netif_wake_queue(ssi->netdev); 745 netif_carrier_on(ssi->netdev); 746 } 747 748 static void ssip_rx_ready(struct hsi_client *cl) 749 { 750 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 751 752 spin_lock_bh(&ssi->lock); 753 if (unlikely(ssi->main_state != ACTIVE)) { 754 dev_dbg(&cl->device, "READY on wrong state: S(%d) M(%d)\n", 755 ssi->send_state, ssi->main_state); 756 spin_unlock_bh(&ssi->lock); 757 return; 758 } 759 if (ssi->send_state != WAIT4READY) { 760 dev_dbg(&cl->device, "Ignore spurious READY command\n"); 761 spin_unlock_bh(&ssi->lock); 762 return; 763 } 764 ssip_set_txstate(ssi, SEND_READY); 765 spin_unlock_bh(&ssi->lock); 766 ssip_xmit(cl); 767 } 768 769 static void ssip_rx_strans(struct hsi_client *cl, u32 cmd) 770 { 771 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 772 struct sk_buff *skb; 773 struct hsi_msg *msg; 774 int len = SSIP_PDU_LENGTH(cmd); 775 776 dev_dbg(&cl->device, "RX strans: %d frames\n", len); 777 spin_lock_bh(&ssi->lock); 778 if (unlikely(ssi->main_state != ACTIVE)) { 779 dev_err(&cl->device, "START TRANS wrong state: S(%d) M(%d)\n", 780 ssi->send_state, ssi->main_state); 781 spin_unlock_bh(&ssi->lock); 782 return; 783 } 784 ssip_set_rxstate(ssi, RECEIVING); 785 if (unlikely(SSIP_MSG_ID(cmd) != ssi->rxid)) { 786 dev_err(&cl->device, "START TRANS id %d expected %d\n", 787 SSIP_MSG_ID(cmd), ssi->rxid); 788 spin_unlock_bh(&ssi->lock); 789 goto out1; 790 } 791 ssi->rxid++; 792 spin_unlock_bh(&ssi->lock); 793 skb = netdev_alloc_skb(ssi->netdev, len * 4); 794 if (unlikely(!skb)) { 795 dev_err(&cl->device, "No memory for rx skb\n"); 796 goto out1; 797 } 798 skb->dev = ssi->netdev; 799 skb_put(skb, len * 4); 800 msg = ssip_alloc_data(ssi, skb, GFP_ATOMIC); 801 if (unlikely(!msg)) { 802 dev_err(&cl->device, "No memory for RX data msg\n"); 803 goto out2; 804 } 805 msg->complete = ssip_rx_data_complete; 806 hsi_async_read(cl, msg); 807 808 return; 809 out2: 810 dev_kfree_skb(skb); 811 out1: 812 ssip_error(cl); 813 } 814 815 static void ssip_rxcmd_complete(struct hsi_msg *msg) 816 { 817 struct hsi_client *cl = msg->cl; 818 u32 cmd = ssip_get_cmd(msg); 819 unsigned int cmdid = SSIP_COMMAND(cmd); 820 821 if (msg->status == HSI_STATUS_ERROR) { 822 dev_err(&cl->device, "RX error detected\n"); 823 ssip_release_cmd(msg); 824 ssip_error(cl); 825 return; 826 } 827 hsi_async_read(cl, msg); 828 dev_dbg(&cl->device, "RX cmd: 0x%08x\n", cmd); 829 switch (cmdid) { 830 case SSIP_SW_BREAK: 831 /* Ignored */ 832 break; 833 case SSIP_BOOTINFO_REQ: 834 ssip_rx_bootinforeq(cl, cmd); 835 break; 836 case SSIP_BOOTINFO_RESP: 837 ssip_rx_bootinforesp(cl, cmd); 838 break; 839 case SSIP_WAKETEST_RESULT: 840 ssip_rx_waketest(cl, cmd); 841 break; 842 case SSIP_START_TRANS: 843 ssip_rx_strans(cl, cmd); 844 break; 845 case SSIP_READY: 846 ssip_rx_ready(cl); 847 break; 848 default: 849 dev_warn(&cl->device, "command 0x%08x not supported\n", cmd); 850 break; 851 } 852 } 853 854 static void ssip_swbreak_complete(struct hsi_msg *msg) 855 { 856 struct hsi_client *cl = msg->cl; 857 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 858 859 ssip_release_cmd(msg); 860 spin_lock_bh(&ssi->lock); 861 if (list_empty(&ssi->txqueue)) { 862 if (atomic_read(&ssi->tx_usecnt)) { 863 ssip_set_txstate(ssi, SEND_READY); 864 } else { 865 ssip_set_txstate(ssi, SEND_IDLE); 866 hsi_stop_tx(cl); 867 } 868 spin_unlock_bh(&ssi->lock); 869 } else { 870 spin_unlock_bh(&ssi->lock); 871 ssip_xmit(cl); 872 } 873 netif_wake_queue(ssi->netdev); 874 } 875 876 static void ssip_tx_data_complete(struct hsi_msg *msg) 877 { 878 struct hsi_client *cl = msg->cl; 879 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 880 struct hsi_msg *cmsg; 881 882 if (msg->status == HSI_STATUS_ERROR) { 883 dev_err(&cl->device, "TX data error\n"); 884 ssip_error(cl); 885 goto out; 886 } 887 spin_lock_bh(&ssi->lock); 888 if (list_empty(&ssi->txqueue)) { 889 ssip_set_txstate(ssi, SENDING_SWBREAK); 890 spin_unlock_bh(&ssi->lock); 891 cmsg = ssip_claim_cmd(ssi); 892 ssip_set_cmd(cmsg, SSIP_SWBREAK_CMD); 893 cmsg->complete = ssip_swbreak_complete; 894 dev_dbg(&cl->device, "Send SWBREAK\n"); 895 hsi_async_write(cl, cmsg); 896 } else { 897 spin_unlock_bh(&ssi->lock); 898 ssip_xmit(cl); 899 } 900 out: 901 ssip_free_data(msg); 902 } 903 904 static void ssip_port_event(struct hsi_client *cl, unsigned long event) 905 { 906 switch (event) { 907 case HSI_EVENT_START_RX: 908 ssip_start_rx(cl); 909 break; 910 case HSI_EVENT_STOP_RX: 911 ssip_stop_rx(cl); 912 break; 913 default: 914 return; 915 } 916 } 917 918 static int ssip_pn_open(struct net_device *dev) 919 { 920 struct hsi_client *cl = to_hsi_client(dev->dev.parent); 921 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 922 int err; 923 924 err = hsi_claim_port(cl, 1); 925 if (err < 0) { 926 dev_err(&cl->device, "SSI port already claimed\n"); 927 return err; 928 } 929 err = hsi_register_port_event(cl, ssip_port_event); 930 if (err < 0) { 931 dev_err(&cl->device, "Register HSI port event failed (%d)\n", 932 err); 933 return err; 934 } 935 dev_dbg(&cl->device, "Configuring SSI port\n"); 936 hsi_setup(cl); 937 938 if (!test_and_set_bit(SSIP_WAKETEST_FLAG, &ssi->flags)) 939 ssi_waketest(cl, 1); /* FIXME: To be removed */ 940 941 spin_lock_bh(&ssi->lock); 942 ssi->main_state = HANDSHAKE; 943 spin_unlock_bh(&ssi->lock); 944 945 ssip_send_bootinfo_req_cmd(cl); 946 947 return 0; 948 } 949 950 static int ssip_pn_stop(struct net_device *dev) 951 { 952 struct hsi_client *cl = to_hsi_client(dev->dev.parent); 953 954 ssip_reset(cl); 955 hsi_unregister_port_event(cl); 956 hsi_release_port(cl); 957 958 return 0; 959 } 960 961 static void ssip_xmit_work(struct work_struct *work) 962 { 963 struct ssi_protocol *ssi = 964 container_of(work, struct ssi_protocol, work); 965 struct hsi_client *cl = ssi->cl; 966 967 ssip_xmit(cl); 968 } 969 970 static int ssip_pn_xmit(struct sk_buff *skb, struct net_device *dev) 971 { 972 struct hsi_client *cl = to_hsi_client(dev->dev.parent); 973 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 974 struct hsi_msg *msg; 975 976 if ((skb->protocol != htons(ETH_P_PHONET)) || 977 (skb->len < SSIP_MIN_PN_HDR)) 978 goto drop; 979 /* Pad to 32-bits - FIXME: Revisit*/ 980 if ((skb->len & 3) && skb_pad(skb, 4 - (skb->len & 3))) 981 goto inc_dropped; 982 983 /* 984 * Modem sends Phonet messages over SSI with its own endianness. 985 * Assume that modem has the same endianness as we do. 986 */ 987 if (skb_cow_head(skb, 0)) 988 goto drop; 989 990 /* length field is exchanged in network byte order */ 991 ((u16 *)skb->data)[2] = htons(((u16 *)skb->data)[2]); 992 993 msg = ssip_alloc_data(ssi, skb, GFP_ATOMIC); 994 if (!msg) { 995 dev_dbg(&cl->device, "Dropping tx data: No memory\n"); 996 goto drop; 997 } 998 msg->complete = ssip_tx_data_complete; 999 1000 spin_lock_bh(&ssi->lock); 1001 if (unlikely(ssi->main_state != ACTIVE)) { 1002 spin_unlock_bh(&ssi->lock); 1003 dev_dbg(&cl->device, "Dropping tx data: CMT is OFFLINE\n"); 1004 goto drop2; 1005 } 1006 list_add_tail(&msg->link, &ssi->txqueue); 1007 ssi->txqueue_len++; 1008 if (dev->tx_queue_len < ssi->txqueue_len) { 1009 dev_info(&cl->device, "TX queue full %d\n", ssi->txqueue_len); 1010 netif_stop_queue(dev); 1011 } 1012 if (ssi->send_state == SEND_IDLE) { 1013 ssip_set_txstate(ssi, WAIT4READY); 1014 spin_unlock_bh(&ssi->lock); 1015 dev_dbg(&cl->device, "Start TX qlen %d\n", ssi->txqueue_len); 1016 hsi_start_tx(cl); 1017 } else if (ssi->send_state == SEND_READY) { 1018 /* Needed for cmt-speech workaround */ 1019 dev_dbg(&cl->device, "Start TX on SEND READY qlen %d\n", 1020 ssi->txqueue_len); 1021 spin_unlock_bh(&ssi->lock); 1022 schedule_work(&ssi->work); 1023 } else { 1024 spin_unlock_bh(&ssi->lock); 1025 } 1026 dev->stats.tx_packets++; 1027 dev->stats.tx_bytes += skb->len; 1028 1029 return 0; 1030 drop2: 1031 hsi_free_msg(msg); 1032 drop: 1033 dev_kfree_skb(skb); 1034 inc_dropped: 1035 dev->stats.tx_dropped++; 1036 1037 return 0; 1038 } 1039 1040 /* CMT reset event handler */ 1041 void ssip_reset_event(struct hsi_client *master) 1042 { 1043 struct ssi_protocol *ssi = hsi_client_drvdata(master); 1044 dev_err(&ssi->cl->device, "CMT reset detected!\n"); 1045 ssip_error(ssi->cl); 1046 } 1047 EXPORT_SYMBOL_GPL(ssip_reset_event); 1048 1049 static const struct net_device_ops ssip_pn_ops = { 1050 .ndo_open = ssip_pn_open, 1051 .ndo_stop = ssip_pn_stop, 1052 .ndo_start_xmit = ssip_pn_xmit, 1053 }; 1054 1055 static void ssip_pn_setup(struct net_device *dev) 1056 { 1057 dev->features = 0; 1058 dev->netdev_ops = &ssip_pn_ops; 1059 dev->type = ARPHRD_PHONET; 1060 dev->flags = IFF_POINTOPOINT | IFF_NOARP; 1061 dev->mtu = SSIP_DEFAULT_MTU; 1062 dev->hard_header_len = 1; 1063 dev->dev_addr[0] = PN_MEDIA_SOS; 1064 dev->addr_len = 1; 1065 dev->tx_queue_len = SSIP_TXQUEUE_LEN; 1066 1067 dev->needs_free_netdev = true; 1068 dev->header_ops = &phonet_header_ops; 1069 } 1070 1071 static int ssi_protocol_probe(struct device *dev) 1072 { 1073 static const char ifname[] = "phonet%d"; 1074 struct hsi_client *cl = to_hsi_client(dev); 1075 struct ssi_protocol *ssi; 1076 int err; 1077 1078 ssi = kzalloc(sizeof(*ssi), GFP_KERNEL); 1079 if (!ssi) 1080 return -ENOMEM; 1081 1082 spin_lock_init(&ssi->lock); 1083 timer_setup(&ssi->rx_wd, ssip_rx_wd, TIMER_DEFERRABLE); 1084 timer_setup(&ssi->tx_wd, ssip_tx_wd, TIMER_DEFERRABLE); 1085 timer_setup(&ssi->keep_alive, ssip_keep_alive, 0); 1086 INIT_LIST_HEAD(&ssi->txqueue); 1087 INIT_LIST_HEAD(&ssi->cmdqueue); 1088 atomic_set(&ssi->tx_usecnt, 0); 1089 hsi_client_set_drvdata(cl, ssi); 1090 ssi->cl = cl; 1091 INIT_WORK(&ssi->work, ssip_xmit_work); 1092 1093 ssi->channel_id_cmd = hsi_get_channel_id_by_name(cl, "mcsaab-control"); 1094 if (ssi->channel_id_cmd < 0) { 1095 err = ssi->channel_id_cmd; 1096 dev_err(dev, "Could not get cmd channel (%d)\n", err); 1097 goto out; 1098 } 1099 1100 ssi->channel_id_data = hsi_get_channel_id_by_name(cl, "mcsaab-data"); 1101 if (ssi->channel_id_data < 0) { 1102 err = ssi->channel_id_data; 1103 dev_err(dev, "Could not get data channel (%d)\n", err); 1104 goto out; 1105 } 1106 1107 err = ssip_alloc_cmds(ssi); 1108 if (err < 0) { 1109 dev_err(dev, "No memory for commands\n"); 1110 goto out; 1111 } 1112 1113 ssi->netdev = alloc_netdev(0, ifname, NET_NAME_UNKNOWN, ssip_pn_setup); 1114 if (!ssi->netdev) { 1115 dev_err(dev, "No memory for netdev\n"); 1116 err = -ENOMEM; 1117 goto out1; 1118 } 1119 1120 /* MTU range: 6 - 65535 */ 1121 ssi->netdev->min_mtu = PHONET_MIN_MTU; 1122 ssi->netdev->max_mtu = SSIP_MAX_MTU; 1123 1124 SET_NETDEV_DEV(ssi->netdev, dev); 1125 netif_carrier_off(ssi->netdev); 1126 err = register_netdev(ssi->netdev); 1127 if (err < 0) { 1128 dev_err(dev, "Register netdev failed (%d)\n", err); 1129 goto out2; 1130 } 1131 1132 list_add(&ssi->link, &ssip_list); 1133 1134 dev_dbg(dev, "channel configuration: cmd=%d, data=%d\n", 1135 ssi->channel_id_cmd, ssi->channel_id_data); 1136 1137 return 0; 1138 out2: 1139 free_netdev(ssi->netdev); 1140 out1: 1141 ssip_free_cmds(ssi); 1142 out: 1143 kfree(ssi); 1144 1145 return err; 1146 } 1147 1148 static int ssi_protocol_remove(struct device *dev) 1149 { 1150 struct hsi_client *cl = to_hsi_client(dev); 1151 struct ssi_protocol *ssi = hsi_client_drvdata(cl); 1152 1153 list_del(&ssi->link); 1154 unregister_netdev(ssi->netdev); 1155 ssip_free_cmds(ssi); 1156 hsi_client_set_drvdata(cl, NULL); 1157 kfree(ssi); 1158 1159 return 0; 1160 } 1161 1162 static struct hsi_client_driver ssip_driver = { 1163 .driver = { 1164 .name = "ssi-protocol", 1165 .owner = THIS_MODULE, 1166 .probe = ssi_protocol_probe, 1167 .remove = ssi_protocol_remove, 1168 }, 1169 }; 1170 1171 static int __init ssip_init(void) 1172 { 1173 pr_info("SSI protocol aka McSAAB added\n"); 1174 1175 return hsi_register_client_driver(&ssip_driver); 1176 } 1177 module_init(ssip_init); 1178 1179 static void __exit ssip_exit(void) 1180 { 1181 hsi_unregister_client_driver(&ssip_driver); 1182 pr_info("SSI protocol driver removed\n"); 1183 } 1184 module_exit(ssip_exit); 1185 1186 MODULE_ALIAS("hsi:ssi-protocol"); 1187 MODULE_AUTHOR("Carlos Chinea <carlos.chinea@nokia.com>"); 1188 MODULE_AUTHOR("Remi Denis-Courmont <remi.denis-courmont@nokia.com>"); 1189 MODULE_DESCRIPTION("SSI protocol improved aka McSAAB"); 1190 MODULE_LICENSE("GPL"); 1191