1 // SPDX-License-Identifier: GPL-2.0-only 2 /* Copyright (c) 2021, Linaro Ltd <loic.poulain@linaro.org> */ 3 4 #include <linux/err.h> 5 #include <linux/errno.h> 6 #include <linux/debugfs.h> 7 #include <linux/fs.h> 8 #include <linux/init.h> 9 #include <linux/idr.h> 10 #include <linux/kernel.h> 11 #include <linux/module.h> 12 #include <linux/poll.h> 13 #include <linux/skbuff.h> 14 #include <linux/slab.h> 15 #include <linux/types.h> 16 #include <linux/uaccess.h> 17 #include <linux/termios.h> 18 #include <linux/wwan.h> 19 #include <net/rtnetlink.h> 20 #include <uapi/linux/wwan.h> 21 22 /* Maximum number of minors in use */ 23 #define WWAN_MAX_MINORS (1 << MINORBITS) 24 25 static DEFINE_MUTEX(wwan_register_lock); /* WWAN device create|remove lock */ 26 static DEFINE_IDA(minors); /* minors for WWAN port chardevs */ 27 static DEFINE_IDA(wwan_dev_ids); /* for unique WWAN device IDs */ 28 static struct class *wwan_class; 29 static int wwan_major; 30 static struct dentry *wwan_debugfs_dir; 31 32 #define to_wwan_dev(d) container_of(d, struct wwan_device, dev) 33 #define to_wwan_port(d) container_of(d, struct wwan_port, dev) 34 35 /* WWAN port flags */ 36 #define WWAN_PORT_TX_OFF 0 37 38 /** 39 * struct wwan_device - The structure that defines a WWAN device 40 * 41 * @id: WWAN device unique ID. 42 * @dev: Underlying device. 43 * @port_id: Current available port ID to pick. 44 * @ops: wwan device ops 45 * @ops_ctxt: context to pass to ops 46 * @debugfs_dir: WWAN device debugfs dir 47 */ 48 struct wwan_device { 49 unsigned int id; 50 struct device dev; 51 atomic_t port_id; 52 const struct wwan_ops *ops; 53 void *ops_ctxt; 54 #ifdef CONFIG_WWAN_DEBUGFS 55 struct dentry *debugfs_dir; 56 #endif 57 }; 58 59 /** 60 * struct wwan_port - The structure that defines a WWAN port 61 * @type: Port type 62 * @start_count: Port start counter 63 * @flags: Store port state and capabilities 64 * @ops: Pointer to WWAN port operations 65 * @ops_lock: Protect port ops 66 * @dev: Underlying device 67 * @rxq: Buffer inbound queue 68 * @waitqueue: The waitqueue for port fops (read/write/poll) 69 * @data_lock: Port specific data access serialization 70 * @at_data: AT port specific data 71 */ 72 struct wwan_port { 73 enum wwan_port_type type; 74 unsigned int start_count; 75 unsigned long flags; 76 const struct wwan_port_ops *ops; 77 struct mutex ops_lock; /* Serialize ops + protect against removal */ 78 struct device dev; 79 struct sk_buff_head rxq; 80 wait_queue_head_t waitqueue; 81 struct mutex data_lock; /* Port specific data access serialization */ 82 union { 83 struct { 84 struct ktermios termios; 85 int mdmbits; 86 } at_data; 87 }; 88 }; 89 90 static ssize_t index_show(struct device *dev, struct device_attribute *attr, char *buf) 91 { 92 struct wwan_device *wwan = to_wwan_dev(dev); 93 94 return sprintf(buf, "%d\n", wwan->id); 95 } 96 static DEVICE_ATTR_RO(index); 97 98 static struct attribute *wwan_dev_attrs[] = { 99 &dev_attr_index.attr, 100 NULL, 101 }; 102 ATTRIBUTE_GROUPS(wwan_dev); 103 104 static void wwan_dev_destroy(struct device *dev) 105 { 106 struct wwan_device *wwandev = to_wwan_dev(dev); 107 108 ida_free(&wwan_dev_ids, wwandev->id); 109 kfree(wwandev); 110 } 111 112 static const struct device_type wwan_dev_type = { 113 .name = "wwan_dev", 114 .release = wwan_dev_destroy, 115 .groups = wwan_dev_groups, 116 }; 117 118 static int wwan_dev_parent_match(struct device *dev, const void *parent) 119 { 120 return (dev->type == &wwan_dev_type && 121 (dev->parent == parent || dev == parent)); 122 } 123 124 static struct wwan_device *wwan_dev_get_by_parent(struct device *parent) 125 { 126 struct device *dev; 127 128 dev = class_find_device(wwan_class, NULL, parent, wwan_dev_parent_match); 129 if (!dev) 130 return ERR_PTR(-ENODEV); 131 132 return to_wwan_dev(dev); 133 } 134 135 static int wwan_dev_name_match(struct device *dev, const void *name) 136 { 137 return dev->type == &wwan_dev_type && 138 strcmp(dev_name(dev), name) == 0; 139 } 140 141 static struct wwan_device *wwan_dev_get_by_name(const char *name) 142 { 143 struct device *dev; 144 145 dev = class_find_device(wwan_class, NULL, name, wwan_dev_name_match); 146 if (!dev) 147 return ERR_PTR(-ENODEV); 148 149 return to_wwan_dev(dev); 150 } 151 152 #ifdef CONFIG_WWAN_DEBUGFS 153 struct dentry *wwan_get_debugfs_dir(struct device *parent) 154 { 155 struct wwan_device *wwandev; 156 157 wwandev = wwan_dev_get_by_parent(parent); 158 if (IS_ERR(wwandev)) 159 return ERR_CAST(wwandev); 160 161 return wwandev->debugfs_dir; 162 } 163 EXPORT_SYMBOL_GPL(wwan_get_debugfs_dir); 164 165 static int wwan_dev_debugfs_match(struct device *dev, const void *dir) 166 { 167 struct wwan_device *wwandev; 168 169 if (dev->type != &wwan_dev_type) 170 return 0; 171 172 wwandev = to_wwan_dev(dev); 173 174 return wwandev->debugfs_dir == dir; 175 } 176 177 static struct wwan_device *wwan_dev_get_by_debugfs(struct dentry *dir) 178 { 179 struct device *dev; 180 181 dev = class_find_device(wwan_class, NULL, dir, wwan_dev_debugfs_match); 182 if (!dev) 183 return ERR_PTR(-ENODEV); 184 185 return to_wwan_dev(dev); 186 } 187 188 void wwan_put_debugfs_dir(struct dentry *dir) 189 { 190 struct wwan_device *wwandev = wwan_dev_get_by_debugfs(dir); 191 192 if (WARN_ON(IS_ERR(wwandev))) 193 return; 194 195 /* wwan_dev_get_by_debugfs() also got a reference */ 196 put_device(&wwandev->dev); 197 put_device(&wwandev->dev); 198 } 199 EXPORT_SYMBOL_GPL(wwan_put_debugfs_dir); 200 #endif 201 202 /* This function allocates and registers a new WWAN device OR if a WWAN device 203 * already exist for the given parent, it gets a reference and return it. 204 * This function is not exported (for now), it is called indirectly via 205 * wwan_create_port(). 206 */ 207 static struct wwan_device *wwan_create_dev(struct device *parent) 208 { 209 struct wwan_device *wwandev; 210 int err, id; 211 212 /* The 'find-alloc-register' operation must be protected against 213 * concurrent execution, a WWAN device is possibly shared between 214 * multiple callers or concurrently unregistered from wwan_remove_dev(). 215 */ 216 mutex_lock(&wwan_register_lock); 217 218 /* If wwandev already exists, return it */ 219 wwandev = wwan_dev_get_by_parent(parent); 220 if (!IS_ERR(wwandev)) 221 goto done_unlock; 222 223 id = ida_alloc(&wwan_dev_ids, GFP_KERNEL); 224 if (id < 0) { 225 wwandev = ERR_PTR(id); 226 goto done_unlock; 227 } 228 229 wwandev = kzalloc(sizeof(*wwandev), GFP_KERNEL); 230 if (!wwandev) { 231 wwandev = ERR_PTR(-ENOMEM); 232 ida_free(&wwan_dev_ids, id); 233 goto done_unlock; 234 } 235 236 wwandev->dev.parent = parent; 237 wwandev->dev.class = wwan_class; 238 wwandev->dev.type = &wwan_dev_type; 239 wwandev->id = id; 240 dev_set_name(&wwandev->dev, "wwan%d", wwandev->id); 241 242 err = device_register(&wwandev->dev); 243 if (err) { 244 put_device(&wwandev->dev); 245 wwandev = ERR_PTR(err); 246 goto done_unlock; 247 } 248 249 #ifdef CONFIG_WWAN_DEBUGFS 250 wwandev->debugfs_dir = 251 debugfs_create_dir(kobject_name(&wwandev->dev.kobj), 252 wwan_debugfs_dir); 253 #endif 254 255 done_unlock: 256 mutex_unlock(&wwan_register_lock); 257 258 return wwandev; 259 } 260 261 static int is_wwan_child(struct device *dev, void *data) 262 { 263 return dev->class == wwan_class; 264 } 265 266 static void wwan_remove_dev(struct wwan_device *wwandev) 267 { 268 int ret; 269 270 /* Prevent concurrent picking from wwan_create_dev */ 271 mutex_lock(&wwan_register_lock); 272 273 /* WWAN device is created and registered (get+add) along with its first 274 * child port, and subsequent port registrations only grab a reference 275 * (get). The WWAN device must then be unregistered (del+put) along with 276 * its last port, and reference simply dropped (put) otherwise. In the 277 * same fashion, we must not unregister it when the ops are still there. 278 */ 279 if (wwandev->ops) 280 ret = 1; 281 else 282 ret = device_for_each_child(&wwandev->dev, NULL, is_wwan_child); 283 284 if (!ret) { 285 #ifdef CONFIG_WWAN_DEBUGFS 286 debugfs_remove_recursive(wwandev->debugfs_dir); 287 #endif 288 device_unregister(&wwandev->dev); 289 } else { 290 put_device(&wwandev->dev); 291 } 292 293 mutex_unlock(&wwan_register_lock); 294 } 295 296 /* ------- WWAN port management ------- */ 297 298 static const struct { 299 const char * const name; /* Port type name */ 300 const char * const devsuf; /* Port devce name suffix */ 301 } wwan_port_types[WWAN_PORT_MAX + 1] = { 302 [WWAN_PORT_AT] = { 303 .name = "AT", 304 .devsuf = "at", 305 }, 306 [WWAN_PORT_MBIM] = { 307 .name = "MBIM", 308 .devsuf = "mbim", 309 }, 310 [WWAN_PORT_QMI] = { 311 .name = "QMI", 312 .devsuf = "qmi", 313 }, 314 [WWAN_PORT_QCDM] = { 315 .name = "QCDM", 316 .devsuf = "qcdm", 317 }, 318 [WWAN_PORT_FIREHOSE] = { 319 .name = "FIREHOSE", 320 .devsuf = "firehose", 321 }, 322 [WWAN_PORT_XMMRPC] = { 323 .name = "XMMRPC", 324 .devsuf = "xmmrpc", 325 }, 326 }; 327 328 static ssize_t type_show(struct device *dev, struct device_attribute *attr, 329 char *buf) 330 { 331 struct wwan_port *port = to_wwan_port(dev); 332 333 return sprintf(buf, "%s\n", wwan_port_types[port->type].name); 334 } 335 static DEVICE_ATTR_RO(type); 336 337 static struct attribute *wwan_port_attrs[] = { 338 &dev_attr_type.attr, 339 NULL, 340 }; 341 ATTRIBUTE_GROUPS(wwan_port); 342 343 static void wwan_port_destroy(struct device *dev) 344 { 345 struct wwan_port *port = to_wwan_port(dev); 346 347 ida_free(&minors, MINOR(port->dev.devt)); 348 mutex_destroy(&port->data_lock); 349 mutex_destroy(&port->ops_lock); 350 kfree(port); 351 } 352 353 static const struct device_type wwan_port_dev_type = { 354 .name = "wwan_port", 355 .release = wwan_port_destroy, 356 .groups = wwan_port_groups, 357 }; 358 359 static int wwan_port_minor_match(struct device *dev, const void *minor) 360 { 361 return (dev->type == &wwan_port_dev_type && 362 MINOR(dev->devt) == *(unsigned int *)minor); 363 } 364 365 static struct wwan_port *wwan_port_get_by_minor(unsigned int minor) 366 { 367 struct device *dev; 368 369 dev = class_find_device(wwan_class, NULL, &minor, wwan_port_minor_match); 370 if (!dev) 371 return ERR_PTR(-ENODEV); 372 373 return to_wwan_port(dev); 374 } 375 376 /* Allocate and set unique name based on passed format 377 * 378 * Name allocation approach is highly inspired by the __dev_alloc_name() 379 * function. 380 * 381 * To avoid names collision, the caller must prevent the new port device 382 * registration as well as concurrent invocation of this function. 383 */ 384 static int __wwan_port_dev_assign_name(struct wwan_port *port, const char *fmt) 385 { 386 struct wwan_device *wwandev = to_wwan_dev(port->dev.parent); 387 const unsigned int max_ports = PAGE_SIZE * 8; 388 struct class_dev_iter iter; 389 unsigned long *idmap; 390 struct device *dev; 391 char buf[0x20]; 392 int id; 393 394 idmap = (unsigned long *)get_zeroed_page(GFP_KERNEL); 395 if (!idmap) 396 return -ENOMEM; 397 398 /* Collect ids of same name format ports */ 399 class_dev_iter_init(&iter, wwan_class, NULL, &wwan_port_dev_type); 400 while ((dev = class_dev_iter_next(&iter))) { 401 if (dev->parent != &wwandev->dev) 402 continue; 403 if (sscanf(dev_name(dev), fmt, &id) != 1) 404 continue; 405 if (id < 0 || id >= max_ports) 406 continue; 407 set_bit(id, idmap); 408 } 409 class_dev_iter_exit(&iter); 410 411 /* Allocate unique id */ 412 id = find_first_zero_bit(idmap, max_ports); 413 free_page((unsigned long)idmap); 414 415 snprintf(buf, sizeof(buf), fmt, id); /* Name generation */ 416 417 dev = device_find_child_by_name(&wwandev->dev, buf); 418 if (dev) { 419 put_device(dev); 420 return -ENFILE; 421 } 422 423 return dev_set_name(&port->dev, buf); 424 } 425 426 struct wwan_port *wwan_create_port(struct device *parent, 427 enum wwan_port_type type, 428 const struct wwan_port_ops *ops, 429 void *drvdata) 430 { 431 struct wwan_device *wwandev; 432 struct wwan_port *port; 433 char namefmt[0x20]; 434 int minor, err; 435 436 if (type > WWAN_PORT_MAX || !ops) 437 return ERR_PTR(-EINVAL); 438 439 /* A port is always a child of a WWAN device, retrieve (allocate or 440 * pick) the WWAN device based on the provided parent device. 441 */ 442 wwandev = wwan_create_dev(parent); 443 if (IS_ERR(wwandev)) 444 return ERR_CAST(wwandev); 445 446 /* A port is exposed as character device, get a minor */ 447 minor = ida_alloc_range(&minors, 0, WWAN_MAX_MINORS - 1, GFP_KERNEL); 448 if (minor < 0) { 449 err = minor; 450 goto error_wwandev_remove; 451 } 452 453 port = kzalloc(sizeof(*port), GFP_KERNEL); 454 if (!port) { 455 err = -ENOMEM; 456 ida_free(&minors, minor); 457 goto error_wwandev_remove; 458 } 459 460 port->type = type; 461 port->ops = ops; 462 mutex_init(&port->ops_lock); 463 skb_queue_head_init(&port->rxq); 464 init_waitqueue_head(&port->waitqueue); 465 mutex_init(&port->data_lock); 466 467 port->dev.parent = &wwandev->dev; 468 port->dev.class = wwan_class; 469 port->dev.type = &wwan_port_dev_type; 470 port->dev.devt = MKDEV(wwan_major, minor); 471 dev_set_drvdata(&port->dev, drvdata); 472 473 /* allocate unique name based on wwan device id, port type and number */ 474 snprintf(namefmt, sizeof(namefmt), "wwan%u%s%%d", wwandev->id, 475 wwan_port_types[port->type].devsuf); 476 477 /* Serialize ports registration */ 478 mutex_lock(&wwan_register_lock); 479 480 __wwan_port_dev_assign_name(port, namefmt); 481 err = device_register(&port->dev); 482 483 mutex_unlock(&wwan_register_lock); 484 485 if (err) 486 goto error_put_device; 487 488 return port; 489 490 error_put_device: 491 put_device(&port->dev); 492 error_wwandev_remove: 493 wwan_remove_dev(wwandev); 494 495 return ERR_PTR(err); 496 } 497 EXPORT_SYMBOL_GPL(wwan_create_port); 498 499 void wwan_remove_port(struct wwan_port *port) 500 { 501 struct wwan_device *wwandev = to_wwan_dev(port->dev.parent); 502 503 mutex_lock(&port->ops_lock); 504 if (port->start_count) 505 port->ops->stop(port); 506 port->ops = NULL; /* Prevent any new port operations (e.g. from fops) */ 507 mutex_unlock(&port->ops_lock); 508 509 wake_up_interruptible(&port->waitqueue); 510 511 skb_queue_purge(&port->rxq); 512 dev_set_drvdata(&port->dev, NULL); 513 device_unregister(&port->dev); 514 515 /* Release related wwan device */ 516 wwan_remove_dev(wwandev); 517 } 518 EXPORT_SYMBOL_GPL(wwan_remove_port); 519 520 void wwan_port_rx(struct wwan_port *port, struct sk_buff *skb) 521 { 522 skb_queue_tail(&port->rxq, skb); 523 wake_up_interruptible(&port->waitqueue); 524 } 525 EXPORT_SYMBOL_GPL(wwan_port_rx); 526 527 void wwan_port_txon(struct wwan_port *port) 528 { 529 clear_bit(WWAN_PORT_TX_OFF, &port->flags); 530 wake_up_interruptible(&port->waitqueue); 531 } 532 EXPORT_SYMBOL_GPL(wwan_port_txon); 533 534 void wwan_port_txoff(struct wwan_port *port) 535 { 536 set_bit(WWAN_PORT_TX_OFF, &port->flags); 537 } 538 EXPORT_SYMBOL_GPL(wwan_port_txoff); 539 540 void *wwan_port_get_drvdata(struct wwan_port *port) 541 { 542 return dev_get_drvdata(&port->dev); 543 } 544 EXPORT_SYMBOL_GPL(wwan_port_get_drvdata); 545 546 static int wwan_port_op_start(struct wwan_port *port) 547 { 548 int ret = 0; 549 550 mutex_lock(&port->ops_lock); 551 if (!port->ops) { /* Port got unplugged */ 552 ret = -ENODEV; 553 goto out_unlock; 554 } 555 556 /* If port is already started, don't start again */ 557 if (!port->start_count) 558 ret = port->ops->start(port); 559 560 if (!ret) 561 port->start_count++; 562 563 out_unlock: 564 mutex_unlock(&port->ops_lock); 565 566 return ret; 567 } 568 569 static void wwan_port_op_stop(struct wwan_port *port) 570 { 571 mutex_lock(&port->ops_lock); 572 port->start_count--; 573 if (!port->start_count) { 574 if (port->ops) 575 port->ops->stop(port); 576 skb_queue_purge(&port->rxq); 577 } 578 mutex_unlock(&port->ops_lock); 579 } 580 581 static int wwan_port_op_tx(struct wwan_port *port, struct sk_buff *skb, 582 bool nonblock) 583 { 584 int ret; 585 586 mutex_lock(&port->ops_lock); 587 if (!port->ops) { /* Port got unplugged */ 588 ret = -ENODEV; 589 goto out_unlock; 590 } 591 592 if (nonblock || !port->ops->tx_blocking) 593 ret = port->ops->tx(port, skb); 594 else 595 ret = port->ops->tx_blocking(port, skb); 596 597 out_unlock: 598 mutex_unlock(&port->ops_lock); 599 600 return ret; 601 } 602 603 static bool is_read_blocked(struct wwan_port *port) 604 { 605 return skb_queue_empty(&port->rxq) && port->ops; 606 } 607 608 static bool is_write_blocked(struct wwan_port *port) 609 { 610 return test_bit(WWAN_PORT_TX_OFF, &port->flags) && port->ops; 611 } 612 613 static int wwan_wait_rx(struct wwan_port *port, bool nonblock) 614 { 615 if (!is_read_blocked(port)) 616 return 0; 617 618 if (nonblock) 619 return -EAGAIN; 620 621 if (wait_event_interruptible(port->waitqueue, !is_read_blocked(port))) 622 return -ERESTARTSYS; 623 624 return 0; 625 } 626 627 static int wwan_wait_tx(struct wwan_port *port, bool nonblock) 628 { 629 if (!is_write_blocked(port)) 630 return 0; 631 632 if (nonblock) 633 return -EAGAIN; 634 635 if (wait_event_interruptible(port->waitqueue, !is_write_blocked(port))) 636 return -ERESTARTSYS; 637 638 return 0; 639 } 640 641 static int wwan_port_fops_open(struct inode *inode, struct file *file) 642 { 643 struct wwan_port *port; 644 int err = 0; 645 646 port = wwan_port_get_by_minor(iminor(inode)); 647 if (IS_ERR(port)) 648 return PTR_ERR(port); 649 650 file->private_data = port; 651 stream_open(inode, file); 652 653 err = wwan_port_op_start(port); 654 if (err) 655 put_device(&port->dev); 656 657 return err; 658 } 659 660 static int wwan_port_fops_release(struct inode *inode, struct file *filp) 661 { 662 struct wwan_port *port = filp->private_data; 663 664 wwan_port_op_stop(port); 665 put_device(&port->dev); 666 667 return 0; 668 } 669 670 static ssize_t wwan_port_fops_read(struct file *filp, char __user *buf, 671 size_t count, loff_t *ppos) 672 { 673 struct wwan_port *port = filp->private_data; 674 struct sk_buff *skb; 675 size_t copied; 676 int ret; 677 678 ret = wwan_wait_rx(port, !!(filp->f_flags & O_NONBLOCK)); 679 if (ret) 680 return ret; 681 682 skb = skb_dequeue(&port->rxq); 683 if (!skb) 684 return -EIO; 685 686 copied = min_t(size_t, count, skb->len); 687 if (copy_to_user(buf, skb->data, copied)) { 688 kfree_skb(skb); 689 return -EFAULT; 690 } 691 skb_pull(skb, copied); 692 693 /* skb is not fully consumed, keep it in the queue */ 694 if (skb->len) 695 skb_queue_head(&port->rxq, skb); 696 else 697 consume_skb(skb); 698 699 return copied; 700 } 701 702 static ssize_t wwan_port_fops_write(struct file *filp, const char __user *buf, 703 size_t count, loff_t *offp) 704 { 705 struct wwan_port *port = filp->private_data; 706 struct sk_buff *skb; 707 int ret; 708 709 ret = wwan_wait_tx(port, !!(filp->f_flags & O_NONBLOCK)); 710 if (ret) 711 return ret; 712 713 skb = alloc_skb(count, GFP_KERNEL); 714 if (!skb) 715 return -ENOMEM; 716 717 if (copy_from_user(skb_put(skb, count), buf, count)) { 718 kfree_skb(skb); 719 return -EFAULT; 720 } 721 722 ret = wwan_port_op_tx(port, skb, !!(filp->f_flags & O_NONBLOCK)); 723 if (ret) { 724 kfree_skb(skb); 725 return ret; 726 } 727 728 return count; 729 } 730 731 static __poll_t wwan_port_fops_poll(struct file *filp, poll_table *wait) 732 { 733 struct wwan_port *port = filp->private_data; 734 __poll_t mask = 0; 735 736 poll_wait(filp, &port->waitqueue, wait); 737 738 mutex_lock(&port->ops_lock); 739 if (port->ops && port->ops->tx_poll) 740 mask |= port->ops->tx_poll(port, filp, wait); 741 else if (!is_write_blocked(port)) 742 mask |= EPOLLOUT | EPOLLWRNORM; 743 if (!is_read_blocked(port)) 744 mask |= EPOLLIN | EPOLLRDNORM; 745 if (!port->ops) 746 mask |= EPOLLHUP | EPOLLERR; 747 mutex_unlock(&port->ops_lock); 748 749 return mask; 750 } 751 752 /* Implements minimalistic stub terminal IOCTLs support */ 753 static long wwan_port_fops_at_ioctl(struct wwan_port *port, unsigned int cmd, 754 unsigned long arg) 755 { 756 int ret = 0; 757 758 mutex_lock(&port->data_lock); 759 760 switch (cmd) { 761 case TCFLSH: 762 break; 763 764 case TCGETS: 765 if (copy_to_user((void __user *)arg, &port->at_data.termios, 766 sizeof(struct termios))) 767 ret = -EFAULT; 768 break; 769 770 case TCSETS: 771 case TCSETSW: 772 case TCSETSF: 773 if (copy_from_user(&port->at_data.termios, (void __user *)arg, 774 sizeof(struct termios))) 775 ret = -EFAULT; 776 break; 777 778 #ifdef TCGETS2 779 case TCGETS2: 780 if (copy_to_user((void __user *)arg, &port->at_data.termios, 781 sizeof(struct termios2))) 782 ret = -EFAULT; 783 break; 784 785 case TCSETS2: 786 case TCSETSW2: 787 case TCSETSF2: 788 if (copy_from_user(&port->at_data.termios, (void __user *)arg, 789 sizeof(struct termios2))) 790 ret = -EFAULT; 791 break; 792 #endif 793 794 case TIOCMGET: 795 ret = put_user(port->at_data.mdmbits, (int __user *)arg); 796 break; 797 798 case TIOCMSET: 799 case TIOCMBIC: 800 case TIOCMBIS: { 801 int mdmbits; 802 803 if (copy_from_user(&mdmbits, (int __user *)arg, sizeof(int))) { 804 ret = -EFAULT; 805 break; 806 } 807 if (cmd == TIOCMBIC) 808 port->at_data.mdmbits &= ~mdmbits; 809 else if (cmd == TIOCMBIS) 810 port->at_data.mdmbits |= mdmbits; 811 else 812 port->at_data.mdmbits = mdmbits; 813 break; 814 } 815 816 default: 817 ret = -ENOIOCTLCMD; 818 } 819 820 mutex_unlock(&port->data_lock); 821 822 return ret; 823 } 824 825 static long wwan_port_fops_ioctl(struct file *filp, unsigned int cmd, 826 unsigned long arg) 827 { 828 struct wwan_port *port = filp->private_data; 829 int res; 830 831 if (port->type == WWAN_PORT_AT) { /* AT port specific IOCTLs */ 832 res = wwan_port_fops_at_ioctl(port, cmd, arg); 833 if (res != -ENOIOCTLCMD) 834 return res; 835 } 836 837 switch (cmd) { 838 case TIOCINQ: { /* aka SIOCINQ aka FIONREAD */ 839 unsigned long flags; 840 struct sk_buff *skb; 841 int amount = 0; 842 843 spin_lock_irqsave(&port->rxq.lock, flags); 844 skb_queue_walk(&port->rxq, skb) 845 amount += skb->len; 846 spin_unlock_irqrestore(&port->rxq.lock, flags); 847 848 return put_user(amount, (int __user *)arg); 849 } 850 851 default: 852 return -ENOIOCTLCMD; 853 } 854 } 855 856 static const struct file_operations wwan_port_fops = { 857 .owner = THIS_MODULE, 858 .open = wwan_port_fops_open, 859 .release = wwan_port_fops_release, 860 .read = wwan_port_fops_read, 861 .write = wwan_port_fops_write, 862 .poll = wwan_port_fops_poll, 863 .unlocked_ioctl = wwan_port_fops_ioctl, 864 #ifdef CONFIG_COMPAT 865 .compat_ioctl = compat_ptr_ioctl, 866 #endif 867 .llseek = noop_llseek, 868 }; 869 870 static int wwan_rtnl_validate(struct nlattr *tb[], struct nlattr *data[], 871 struct netlink_ext_ack *extack) 872 { 873 if (!data) 874 return -EINVAL; 875 876 if (!tb[IFLA_PARENT_DEV_NAME]) 877 return -EINVAL; 878 879 if (!data[IFLA_WWAN_LINK_ID]) 880 return -EINVAL; 881 882 return 0; 883 } 884 885 static struct device_type wwan_type = { .name = "wwan" }; 886 887 static struct net_device *wwan_rtnl_alloc(struct nlattr *tb[], 888 const char *ifname, 889 unsigned char name_assign_type, 890 unsigned int num_tx_queues, 891 unsigned int num_rx_queues) 892 { 893 const char *devname = nla_data(tb[IFLA_PARENT_DEV_NAME]); 894 struct wwan_device *wwandev = wwan_dev_get_by_name(devname); 895 struct net_device *dev; 896 unsigned int priv_size; 897 898 if (IS_ERR(wwandev)) 899 return ERR_CAST(wwandev); 900 901 /* only supported if ops were registered (not just ports) */ 902 if (!wwandev->ops) { 903 dev = ERR_PTR(-EOPNOTSUPP); 904 goto out; 905 } 906 907 priv_size = sizeof(struct wwan_netdev_priv) + wwandev->ops->priv_size; 908 dev = alloc_netdev_mqs(priv_size, ifname, name_assign_type, 909 wwandev->ops->setup, num_tx_queues, num_rx_queues); 910 911 if (dev) { 912 SET_NETDEV_DEV(dev, &wwandev->dev); 913 SET_NETDEV_DEVTYPE(dev, &wwan_type); 914 } 915 916 out: 917 /* release the reference */ 918 put_device(&wwandev->dev); 919 return dev; 920 } 921 922 static int wwan_rtnl_newlink(struct net *src_net, struct net_device *dev, 923 struct nlattr *tb[], struct nlattr *data[], 924 struct netlink_ext_ack *extack) 925 { 926 struct wwan_device *wwandev = wwan_dev_get_by_parent(dev->dev.parent); 927 u32 link_id = nla_get_u32(data[IFLA_WWAN_LINK_ID]); 928 struct wwan_netdev_priv *priv = netdev_priv(dev); 929 int ret; 930 931 if (IS_ERR(wwandev)) 932 return PTR_ERR(wwandev); 933 934 /* shouldn't have a netdev (left) with us as parent so WARN */ 935 if (WARN_ON(!wwandev->ops)) { 936 ret = -EOPNOTSUPP; 937 goto out; 938 } 939 940 priv->link_id = link_id; 941 if (wwandev->ops->newlink) 942 ret = wwandev->ops->newlink(wwandev->ops_ctxt, dev, 943 link_id, extack); 944 else 945 ret = register_netdevice(dev); 946 947 out: 948 /* release the reference */ 949 put_device(&wwandev->dev); 950 return ret; 951 } 952 953 static void wwan_rtnl_dellink(struct net_device *dev, struct list_head *head) 954 { 955 struct wwan_device *wwandev = wwan_dev_get_by_parent(dev->dev.parent); 956 957 if (IS_ERR(wwandev)) 958 return; 959 960 /* shouldn't have a netdev (left) with us as parent so WARN */ 961 if (WARN_ON(!wwandev->ops)) 962 goto out; 963 964 if (wwandev->ops->dellink) 965 wwandev->ops->dellink(wwandev->ops_ctxt, dev, head); 966 else 967 unregister_netdevice_queue(dev, head); 968 969 out: 970 /* release the reference */ 971 put_device(&wwandev->dev); 972 } 973 974 static size_t wwan_rtnl_get_size(const struct net_device *dev) 975 { 976 return 977 nla_total_size(4) + /* IFLA_WWAN_LINK_ID */ 978 0; 979 } 980 981 static int wwan_rtnl_fill_info(struct sk_buff *skb, 982 const struct net_device *dev) 983 { 984 struct wwan_netdev_priv *priv = netdev_priv(dev); 985 986 if (nla_put_u32(skb, IFLA_WWAN_LINK_ID, priv->link_id)) 987 goto nla_put_failure; 988 989 return 0; 990 991 nla_put_failure: 992 return -EMSGSIZE; 993 } 994 995 static const struct nla_policy wwan_rtnl_policy[IFLA_WWAN_MAX + 1] = { 996 [IFLA_WWAN_LINK_ID] = { .type = NLA_U32 }, 997 }; 998 999 static struct rtnl_link_ops wwan_rtnl_link_ops __read_mostly = { 1000 .kind = "wwan", 1001 .maxtype = __IFLA_WWAN_MAX, 1002 .alloc = wwan_rtnl_alloc, 1003 .validate = wwan_rtnl_validate, 1004 .newlink = wwan_rtnl_newlink, 1005 .dellink = wwan_rtnl_dellink, 1006 .get_size = wwan_rtnl_get_size, 1007 .fill_info = wwan_rtnl_fill_info, 1008 .policy = wwan_rtnl_policy, 1009 }; 1010 1011 static void wwan_create_default_link(struct wwan_device *wwandev, 1012 u32 def_link_id) 1013 { 1014 struct nlattr *tb[IFLA_MAX + 1], *linkinfo[IFLA_INFO_MAX + 1]; 1015 struct nlattr *data[IFLA_WWAN_MAX + 1]; 1016 struct net_device *dev; 1017 struct nlmsghdr *nlh; 1018 struct sk_buff *msg; 1019 1020 /* Forge attributes required to create a WWAN netdev. We first 1021 * build a netlink message and then parse it. This looks 1022 * odd, but such approach is less error prone. 1023 */ 1024 msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); 1025 if (WARN_ON(!msg)) 1026 return; 1027 nlh = nlmsg_put(msg, 0, 0, RTM_NEWLINK, 0, 0); 1028 if (WARN_ON(!nlh)) 1029 goto free_attrs; 1030 1031 if (nla_put_string(msg, IFLA_PARENT_DEV_NAME, dev_name(&wwandev->dev))) 1032 goto free_attrs; 1033 tb[IFLA_LINKINFO] = nla_nest_start(msg, IFLA_LINKINFO); 1034 if (!tb[IFLA_LINKINFO]) 1035 goto free_attrs; 1036 linkinfo[IFLA_INFO_DATA] = nla_nest_start(msg, IFLA_INFO_DATA); 1037 if (!linkinfo[IFLA_INFO_DATA]) 1038 goto free_attrs; 1039 if (nla_put_u32(msg, IFLA_WWAN_LINK_ID, def_link_id)) 1040 goto free_attrs; 1041 nla_nest_end(msg, linkinfo[IFLA_INFO_DATA]); 1042 nla_nest_end(msg, tb[IFLA_LINKINFO]); 1043 1044 nlmsg_end(msg, nlh); 1045 1046 /* The next three parsing calls can not fail */ 1047 nlmsg_parse_deprecated(nlh, 0, tb, IFLA_MAX, NULL, NULL); 1048 nla_parse_nested_deprecated(linkinfo, IFLA_INFO_MAX, tb[IFLA_LINKINFO], 1049 NULL, NULL); 1050 nla_parse_nested_deprecated(data, IFLA_WWAN_MAX, 1051 linkinfo[IFLA_INFO_DATA], NULL, NULL); 1052 1053 rtnl_lock(); 1054 1055 dev = rtnl_create_link(&init_net, "wwan%d", NET_NAME_ENUM, 1056 &wwan_rtnl_link_ops, tb, NULL); 1057 if (WARN_ON(IS_ERR(dev))) 1058 goto unlock; 1059 1060 if (WARN_ON(wwan_rtnl_newlink(&init_net, dev, tb, data, NULL))) { 1061 free_netdev(dev); 1062 goto unlock; 1063 } 1064 1065 rtnl_configure_link(dev, NULL, 0, NULL); /* Link initialized, notify new link */ 1066 1067 unlock: 1068 rtnl_unlock(); 1069 1070 free_attrs: 1071 nlmsg_free(msg); 1072 } 1073 1074 /** 1075 * wwan_register_ops - register WWAN device ops 1076 * @parent: Device to use as parent and shared by all WWAN ports and 1077 * created netdevs 1078 * @ops: operations to register 1079 * @ctxt: context to pass to operations 1080 * @def_link_id: id of the default link that will be automatically created by 1081 * the WWAN core for the WWAN device. The default link will not be created 1082 * if the passed value is WWAN_NO_DEFAULT_LINK. 1083 * 1084 * Returns: 0 on success, a negative error code on failure 1085 */ 1086 int wwan_register_ops(struct device *parent, const struct wwan_ops *ops, 1087 void *ctxt, u32 def_link_id) 1088 { 1089 struct wwan_device *wwandev; 1090 1091 if (WARN_ON(!parent || !ops || !ops->setup)) 1092 return -EINVAL; 1093 1094 wwandev = wwan_create_dev(parent); 1095 if (IS_ERR(wwandev)) 1096 return PTR_ERR(wwandev); 1097 1098 if (WARN_ON(wwandev->ops)) { 1099 wwan_remove_dev(wwandev); 1100 return -EBUSY; 1101 } 1102 1103 wwandev->ops = ops; 1104 wwandev->ops_ctxt = ctxt; 1105 1106 /* NB: we do not abort ops registration in case of default link 1107 * creation failure. Link ops is the management interface, while the 1108 * default link creation is a service option. And we should not prevent 1109 * a user from manually creating a link latter if service option failed 1110 * now. 1111 */ 1112 if (def_link_id != WWAN_NO_DEFAULT_LINK) 1113 wwan_create_default_link(wwandev, def_link_id); 1114 1115 return 0; 1116 } 1117 EXPORT_SYMBOL_GPL(wwan_register_ops); 1118 1119 /* Enqueue child netdev deletion */ 1120 static int wwan_child_dellink(struct device *dev, void *data) 1121 { 1122 struct list_head *kill_list = data; 1123 1124 if (dev->type == &wwan_type) 1125 wwan_rtnl_dellink(to_net_dev(dev), kill_list); 1126 1127 return 0; 1128 } 1129 1130 /** 1131 * wwan_unregister_ops - remove WWAN device ops 1132 * @parent: Device to use as parent and shared by all WWAN ports and 1133 * created netdevs 1134 */ 1135 void wwan_unregister_ops(struct device *parent) 1136 { 1137 struct wwan_device *wwandev = wwan_dev_get_by_parent(parent); 1138 LIST_HEAD(kill_list); 1139 1140 if (WARN_ON(IS_ERR(wwandev))) 1141 return; 1142 if (WARN_ON(!wwandev->ops)) { 1143 put_device(&wwandev->dev); 1144 return; 1145 } 1146 1147 /* put the reference obtained by wwan_dev_get_by_parent(), 1148 * we should still have one (that the owner is giving back 1149 * now) due to the ops being assigned. 1150 */ 1151 put_device(&wwandev->dev); 1152 1153 rtnl_lock(); /* Prevent concurent netdev(s) creation/destroying */ 1154 1155 /* Remove all child netdev(s), using batch removing */ 1156 device_for_each_child(&wwandev->dev, &kill_list, 1157 wwan_child_dellink); 1158 unregister_netdevice_many(&kill_list); 1159 1160 wwandev->ops = NULL; /* Finally remove ops */ 1161 1162 rtnl_unlock(); 1163 1164 wwandev->ops_ctxt = NULL; 1165 wwan_remove_dev(wwandev); 1166 } 1167 EXPORT_SYMBOL_GPL(wwan_unregister_ops); 1168 1169 static int __init wwan_init(void) 1170 { 1171 int err; 1172 1173 err = rtnl_link_register(&wwan_rtnl_link_ops); 1174 if (err) 1175 return err; 1176 1177 wwan_class = class_create("wwan"); 1178 if (IS_ERR(wwan_class)) { 1179 err = PTR_ERR(wwan_class); 1180 goto unregister; 1181 } 1182 1183 /* chrdev used for wwan ports */ 1184 wwan_major = __register_chrdev(0, 0, WWAN_MAX_MINORS, "wwan_port", 1185 &wwan_port_fops); 1186 if (wwan_major < 0) { 1187 err = wwan_major; 1188 goto destroy; 1189 } 1190 1191 #ifdef CONFIG_WWAN_DEBUGFS 1192 wwan_debugfs_dir = debugfs_create_dir("wwan", NULL); 1193 #endif 1194 1195 return 0; 1196 1197 destroy: 1198 class_destroy(wwan_class); 1199 unregister: 1200 rtnl_link_unregister(&wwan_rtnl_link_ops); 1201 return err; 1202 } 1203 1204 static void __exit wwan_exit(void) 1205 { 1206 debugfs_remove_recursive(wwan_debugfs_dir); 1207 __unregister_chrdev(wwan_major, 0, WWAN_MAX_MINORS, "wwan_port"); 1208 rtnl_link_unregister(&wwan_rtnl_link_ops); 1209 class_destroy(wwan_class); 1210 } 1211 1212 module_init(wwan_init); 1213 module_exit(wwan_exit); 1214 1215 MODULE_AUTHOR("Loic Poulain <loic.poulain@linaro.org>"); 1216 MODULE_DESCRIPTION("WWAN core"); 1217 MODULE_LICENSE("GPL v2"); 1218