1 /* 2 * 3 * Intel Management Engine Interface (Intel MEI) Linux driver 4 * Copyright (c) 2003-2012, Intel Corporation. 5 * 6 * This program is free software; you can redistribute it and/or modify it 7 * under the terms and conditions of the GNU General Public License, 8 * version 2, as published by the Free Software Foundation. 9 * 10 * This program is distributed in the hope it will be useful, but WITHOUT 11 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 12 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for 13 * more details. 14 * 15 */ 16 #include <linux/module.h> 17 #include <linux/moduleparam.h> 18 #include <linux/kernel.h> 19 #include <linux/device.h> 20 #include <linux/slab.h> 21 #include <linux/fs.h> 22 #include <linux/errno.h> 23 #include <linux/types.h> 24 #include <linux/fcntl.h> 25 #include <linux/aio.h> 26 #include <linux/poll.h> 27 #include <linux/init.h> 28 #include <linux/ioctl.h> 29 #include <linux/cdev.h> 30 #include <linux/sched.h> 31 #include <linux/uuid.h> 32 #include <linux/compat.h> 33 #include <linux/jiffies.h> 34 #include <linux/interrupt.h> 35 36 #include <linux/mei.h> 37 38 #include "mei_dev.h" 39 #include "client.h" 40 41 /** 42 * mei_open - the open function 43 * 44 * @inode: pointer to inode structure 45 * @file: pointer to file structure 46 * 47 * Return: 0 on success, <0 on error 48 */ 49 static int mei_open(struct inode *inode, struct file *file) 50 { 51 struct mei_device *dev; 52 struct mei_cl *cl; 53 54 int err; 55 56 dev = container_of(inode->i_cdev, struct mei_device, cdev); 57 if (!dev) 58 return -ENODEV; 59 60 mutex_lock(&dev->device_lock); 61 62 cl = NULL; 63 64 err = -ENODEV; 65 if (dev->dev_state != MEI_DEV_ENABLED) { 66 dev_dbg(dev->dev, "dev_state != MEI_ENABLED dev_state = %s\n", 67 mei_dev_state_str(dev->dev_state)); 68 goto err_unlock; 69 } 70 71 err = -ENOMEM; 72 cl = mei_cl_allocate(dev); 73 if (!cl) 74 goto err_unlock; 75 76 /* open_handle_count check is handled in the mei_cl_link */ 77 err = mei_cl_link(cl, MEI_HOST_CLIENT_ID_ANY); 78 if (err) 79 goto err_unlock; 80 81 file->private_data = cl; 82 83 mutex_unlock(&dev->device_lock); 84 85 return nonseekable_open(inode, file); 86 87 err_unlock: 88 mutex_unlock(&dev->device_lock); 89 kfree(cl); 90 return err; 91 } 92 93 /** 94 * mei_release - the release function 95 * 96 * @inode: pointer to inode structure 97 * @file: pointer to file structure 98 * 99 * Return: 0 on success, <0 on error 100 */ 101 static int mei_release(struct inode *inode, struct file *file) 102 { 103 struct mei_cl *cl = file->private_data; 104 struct mei_cl_cb *cb; 105 struct mei_device *dev; 106 int rets = 0; 107 108 if (WARN_ON(!cl || !cl->dev)) 109 return -ENODEV; 110 111 dev = cl->dev; 112 113 mutex_lock(&dev->device_lock); 114 if (cl == &dev->iamthif_cl) { 115 rets = mei_amthif_release(dev, file); 116 goto out; 117 } 118 if (cl->state == MEI_FILE_CONNECTED) { 119 cl->state = MEI_FILE_DISCONNECTING; 120 cl_dbg(dev, cl, "disconnecting\n"); 121 rets = mei_cl_disconnect(cl); 122 } 123 mei_cl_flush_queues(cl); 124 cl_dbg(dev, cl, "removing\n"); 125 126 mei_cl_unlink(cl); 127 128 129 /* free read cb */ 130 cb = NULL; 131 if (cl->read_cb) { 132 cb = mei_cl_find_read_cb(cl); 133 /* Remove entry from read list */ 134 if (cb) 135 list_del(&cb->list); 136 137 cb = cl->read_cb; 138 cl->read_cb = NULL; 139 } 140 141 file->private_data = NULL; 142 143 mei_io_cb_free(cb); 144 145 kfree(cl); 146 out: 147 mutex_unlock(&dev->device_lock); 148 return rets; 149 } 150 151 152 /** 153 * mei_read - the read function. 154 * 155 * @file: pointer to file structure 156 * @ubuf: pointer to user buffer 157 * @length: buffer length 158 * @offset: data offset in buffer 159 * 160 * Return: >=0 data length on success , <0 on error 161 */ 162 static ssize_t mei_read(struct file *file, char __user *ubuf, 163 size_t length, loff_t *offset) 164 { 165 struct mei_cl *cl = file->private_data; 166 struct mei_cl_cb *cb_pos = NULL; 167 struct mei_cl_cb *cb = NULL; 168 struct mei_device *dev; 169 int rets; 170 int err; 171 172 173 if (WARN_ON(!cl || !cl->dev)) 174 return -ENODEV; 175 176 dev = cl->dev; 177 178 179 mutex_lock(&dev->device_lock); 180 if (dev->dev_state != MEI_DEV_ENABLED) { 181 rets = -ENODEV; 182 goto out; 183 } 184 185 if (length == 0) { 186 rets = 0; 187 goto out; 188 } 189 190 if (cl == &dev->iamthif_cl) { 191 rets = mei_amthif_read(dev, file, ubuf, length, offset); 192 goto out; 193 } 194 195 if (cl->read_cb) { 196 cb = cl->read_cb; 197 /* read what left */ 198 if (cb->buf_idx > *offset) 199 goto copy_buffer; 200 /* offset is beyond buf_idx we have no more data return 0 */ 201 if (cb->buf_idx > 0 && cb->buf_idx <= *offset) { 202 rets = 0; 203 goto free; 204 } 205 /* Offset needs to be cleaned for contiguous reads*/ 206 if (cb->buf_idx == 0 && *offset > 0) 207 *offset = 0; 208 } else if (*offset > 0) { 209 *offset = 0; 210 } 211 212 err = mei_cl_read_start(cl, length); 213 if (err && err != -EBUSY) { 214 dev_dbg(dev->dev, 215 "mei start read failure with status = %d\n", err); 216 rets = err; 217 goto out; 218 } 219 220 if (MEI_READ_COMPLETE != cl->reading_state && 221 !waitqueue_active(&cl->rx_wait)) { 222 if (file->f_flags & O_NONBLOCK) { 223 rets = -EAGAIN; 224 goto out; 225 } 226 227 mutex_unlock(&dev->device_lock); 228 229 if (wait_event_interruptible(cl->rx_wait, 230 MEI_READ_COMPLETE == cl->reading_state || 231 mei_cl_is_transitioning(cl))) { 232 233 if (signal_pending(current)) 234 return -EINTR; 235 return -ERESTARTSYS; 236 } 237 238 mutex_lock(&dev->device_lock); 239 if (mei_cl_is_transitioning(cl)) { 240 rets = -EBUSY; 241 goto out; 242 } 243 } 244 245 cb = cl->read_cb; 246 247 if (!cb) { 248 rets = -ENODEV; 249 goto out; 250 } 251 if (cl->reading_state != MEI_READ_COMPLETE) { 252 rets = 0; 253 goto out; 254 } 255 /* now copy the data to user space */ 256 copy_buffer: 257 dev_dbg(dev->dev, "buf.size = %d buf.idx= %ld\n", 258 cb->response_buffer.size, cb->buf_idx); 259 if (length == 0 || ubuf == NULL || *offset > cb->buf_idx) { 260 rets = -EMSGSIZE; 261 goto free; 262 } 263 264 /* length is being truncated to PAGE_SIZE, 265 * however buf_idx may point beyond that */ 266 length = min_t(size_t, length, cb->buf_idx - *offset); 267 268 if (copy_to_user(ubuf, cb->response_buffer.data + *offset, length)) { 269 dev_dbg(dev->dev, "failed to copy data to userland\n"); 270 rets = -EFAULT; 271 goto free; 272 } 273 274 rets = length; 275 *offset += length; 276 if ((unsigned long)*offset < cb->buf_idx) 277 goto out; 278 279 free: 280 cb_pos = mei_cl_find_read_cb(cl); 281 /* Remove entry from read list */ 282 if (cb_pos) 283 list_del(&cb_pos->list); 284 mei_io_cb_free(cb); 285 cl->reading_state = MEI_IDLE; 286 cl->read_cb = NULL; 287 out: 288 dev_dbg(dev->dev, "end mei read rets= %d\n", rets); 289 mutex_unlock(&dev->device_lock); 290 return rets; 291 } 292 /** 293 * mei_write - the write function. 294 * 295 * @file: pointer to file structure 296 * @ubuf: pointer to user buffer 297 * @length: buffer length 298 * @offset: data offset in buffer 299 * 300 * Return: >=0 data length on success , <0 on error 301 */ 302 static ssize_t mei_write(struct file *file, const char __user *ubuf, 303 size_t length, loff_t *offset) 304 { 305 struct mei_cl *cl = file->private_data; 306 struct mei_me_client *me_cl = NULL; 307 struct mei_cl_cb *write_cb = NULL; 308 struct mei_device *dev; 309 unsigned long timeout = 0; 310 int rets; 311 312 if (WARN_ON(!cl || !cl->dev)) 313 return -ENODEV; 314 315 dev = cl->dev; 316 317 mutex_lock(&dev->device_lock); 318 319 if (dev->dev_state != MEI_DEV_ENABLED) { 320 rets = -ENODEV; 321 goto out; 322 } 323 324 me_cl = mei_me_cl_by_uuid_id(dev, &cl->cl_uuid, cl->me_client_id); 325 if (!me_cl) { 326 rets = -ENOTTY; 327 goto out; 328 } 329 330 if (length == 0) { 331 rets = 0; 332 goto out; 333 } 334 335 if (length > me_cl->props.max_msg_length) { 336 rets = -EFBIG; 337 goto out; 338 } 339 340 if (cl->state != MEI_FILE_CONNECTED) { 341 dev_err(dev->dev, "host client = %d, is not connected to ME client = %d", 342 cl->host_client_id, cl->me_client_id); 343 rets = -ENODEV; 344 goto out; 345 } 346 if (cl == &dev->iamthif_cl) { 347 write_cb = mei_amthif_find_read_list_entry(dev, file); 348 349 if (write_cb) { 350 timeout = write_cb->read_time + 351 mei_secs_to_jiffies(MEI_IAMTHIF_READ_TIMER); 352 353 if (time_after(jiffies, timeout) || 354 cl->reading_state == MEI_READ_COMPLETE) { 355 *offset = 0; 356 list_del(&write_cb->list); 357 mei_io_cb_free(write_cb); 358 write_cb = NULL; 359 } 360 } 361 } 362 363 /* free entry used in read */ 364 if (cl->reading_state == MEI_READ_COMPLETE) { 365 *offset = 0; 366 write_cb = mei_cl_find_read_cb(cl); 367 if (write_cb) { 368 list_del(&write_cb->list); 369 mei_io_cb_free(write_cb); 370 write_cb = NULL; 371 cl->reading_state = MEI_IDLE; 372 cl->read_cb = NULL; 373 } 374 } else if (cl->reading_state == MEI_IDLE) 375 *offset = 0; 376 377 378 write_cb = mei_io_cb_init(cl, file); 379 if (!write_cb) { 380 rets = -ENOMEM; 381 goto out; 382 } 383 rets = mei_io_cb_alloc_req_buf(write_cb, length); 384 if (rets) 385 goto out; 386 387 rets = copy_from_user(write_cb->request_buffer.data, ubuf, length); 388 if (rets) { 389 dev_dbg(dev->dev, "failed to copy data from userland\n"); 390 rets = -EFAULT; 391 goto out; 392 } 393 394 if (cl == &dev->iamthif_cl) { 395 rets = mei_amthif_write(dev, write_cb); 396 397 if (rets) { 398 dev_err(dev->dev, 399 "amthif write failed with status = %d\n", rets); 400 goto out; 401 } 402 mei_me_cl_put(me_cl); 403 mutex_unlock(&dev->device_lock); 404 return length; 405 } 406 407 rets = mei_cl_write(cl, write_cb, false); 408 out: 409 mei_me_cl_put(me_cl); 410 mutex_unlock(&dev->device_lock); 411 if (rets < 0) 412 mei_io_cb_free(write_cb); 413 return rets; 414 } 415 416 /** 417 * mei_ioctl_connect_client - the connect to fw client IOCTL function 418 * 419 * @file: private data of the file object 420 * @data: IOCTL connect data, input and output parameters 421 * 422 * Locking: called under "dev->device_lock" lock 423 * 424 * Return: 0 on success, <0 on failure. 425 */ 426 static int mei_ioctl_connect_client(struct file *file, 427 struct mei_connect_client_data *data) 428 { 429 struct mei_device *dev; 430 struct mei_client *client; 431 struct mei_me_client *me_cl; 432 struct mei_cl *cl; 433 int rets; 434 435 cl = file->private_data; 436 dev = cl->dev; 437 438 if (dev->dev_state != MEI_DEV_ENABLED) 439 return -ENODEV; 440 441 if (cl->state != MEI_FILE_INITIALIZING && 442 cl->state != MEI_FILE_DISCONNECTED) 443 return -EBUSY; 444 445 /* find ME client we're trying to connect to */ 446 me_cl = mei_me_cl_by_uuid(dev, &data->in_client_uuid); 447 if (!me_cl || me_cl->props.fixed_address) { 448 dev_dbg(dev->dev, "Cannot connect to FW Client UUID = %pUl\n", 449 &data->in_client_uuid); 450 return -ENOTTY; 451 } 452 453 cl->me_client_id = me_cl->client_id; 454 cl->cl_uuid = me_cl->props.protocol_name; 455 456 dev_dbg(dev->dev, "Connect to FW Client ID = %d\n", 457 cl->me_client_id); 458 dev_dbg(dev->dev, "FW Client - Protocol Version = %d\n", 459 me_cl->props.protocol_version); 460 dev_dbg(dev->dev, "FW Client - Max Msg Len = %d\n", 461 me_cl->props.max_msg_length); 462 463 /* if we're connecting to amthif client then we will use the 464 * existing connection 465 */ 466 if (uuid_le_cmp(data->in_client_uuid, mei_amthif_guid) == 0) { 467 dev_dbg(dev->dev, "FW Client is amthi\n"); 468 if (dev->iamthif_cl.state != MEI_FILE_CONNECTED) { 469 rets = -ENODEV; 470 goto end; 471 } 472 mei_cl_unlink(cl); 473 474 kfree(cl); 475 cl = NULL; 476 dev->iamthif_open_count++; 477 file->private_data = &dev->iamthif_cl; 478 479 client = &data->out_client_properties; 480 client->max_msg_length = me_cl->props.max_msg_length; 481 client->protocol_version = me_cl->props.protocol_version; 482 rets = dev->iamthif_cl.status; 483 484 goto end; 485 } 486 487 /* prepare the output buffer */ 488 client = &data->out_client_properties; 489 client->max_msg_length = me_cl->props.max_msg_length; 490 client->protocol_version = me_cl->props.protocol_version; 491 dev_dbg(dev->dev, "Can connect?\n"); 492 493 rets = mei_cl_connect(cl, file); 494 495 end: 496 mei_me_cl_put(me_cl); 497 return rets; 498 } 499 500 /** 501 * mei_ioctl - the IOCTL function 502 * 503 * @file: pointer to file structure 504 * @cmd: ioctl command 505 * @data: pointer to mei message structure 506 * 507 * Return: 0 on success , <0 on error 508 */ 509 static long mei_ioctl(struct file *file, unsigned int cmd, unsigned long data) 510 { 511 struct mei_device *dev; 512 struct mei_cl *cl = file->private_data; 513 struct mei_connect_client_data connect_data; 514 int rets; 515 516 517 if (WARN_ON(!cl || !cl->dev)) 518 return -ENODEV; 519 520 dev = cl->dev; 521 522 dev_dbg(dev->dev, "IOCTL cmd = 0x%x", cmd); 523 524 mutex_lock(&dev->device_lock); 525 if (dev->dev_state != MEI_DEV_ENABLED) { 526 rets = -ENODEV; 527 goto out; 528 } 529 530 switch (cmd) { 531 case IOCTL_MEI_CONNECT_CLIENT: 532 dev_dbg(dev->dev, ": IOCTL_MEI_CONNECT_CLIENT.\n"); 533 if (copy_from_user(&connect_data, (char __user *)data, 534 sizeof(struct mei_connect_client_data))) { 535 dev_dbg(dev->dev, "failed to copy data from userland\n"); 536 rets = -EFAULT; 537 goto out; 538 } 539 540 rets = mei_ioctl_connect_client(file, &connect_data); 541 if (rets) 542 goto out; 543 544 /* if all is ok, copying the data back to user. */ 545 if (copy_to_user((char __user *)data, &connect_data, 546 sizeof(struct mei_connect_client_data))) { 547 dev_dbg(dev->dev, "failed to copy data to userland\n"); 548 rets = -EFAULT; 549 goto out; 550 } 551 552 break; 553 554 default: 555 dev_err(dev->dev, ": unsupported ioctl %d.\n", cmd); 556 rets = -ENOIOCTLCMD; 557 } 558 559 out: 560 mutex_unlock(&dev->device_lock); 561 return rets; 562 } 563 564 /** 565 * mei_compat_ioctl - the compat IOCTL function 566 * 567 * @file: pointer to file structure 568 * @cmd: ioctl command 569 * @data: pointer to mei message structure 570 * 571 * Return: 0 on success , <0 on error 572 */ 573 #ifdef CONFIG_COMPAT 574 static long mei_compat_ioctl(struct file *file, 575 unsigned int cmd, unsigned long data) 576 { 577 return mei_ioctl(file, cmd, (unsigned long)compat_ptr(data)); 578 } 579 #endif 580 581 582 /** 583 * mei_poll - the poll function 584 * 585 * @file: pointer to file structure 586 * @wait: pointer to poll_table structure 587 * 588 * Return: poll mask 589 */ 590 static unsigned int mei_poll(struct file *file, poll_table *wait) 591 { 592 struct mei_cl *cl = file->private_data; 593 struct mei_device *dev; 594 unsigned int mask = 0; 595 596 if (WARN_ON(!cl || !cl->dev)) 597 return POLLERR; 598 599 dev = cl->dev; 600 601 mutex_lock(&dev->device_lock); 602 603 if (!mei_cl_is_connected(cl)) { 604 mask = POLLERR; 605 goto out; 606 } 607 608 mutex_unlock(&dev->device_lock); 609 610 611 if (cl == &dev->iamthif_cl) 612 return mei_amthif_poll(dev, file, wait); 613 614 poll_wait(file, &cl->tx_wait, wait); 615 616 mutex_lock(&dev->device_lock); 617 618 if (!mei_cl_is_connected(cl)) { 619 mask = POLLERR; 620 goto out; 621 } 622 623 mask |= (POLLIN | POLLRDNORM); 624 625 out: 626 mutex_unlock(&dev->device_lock); 627 return mask; 628 } 629 630 /** 631 * fw_status_show - mei device attribute show method 632 * 633 * @device: device pointer 634 * @attr: attribute pointer 635 * @buf: char out buffer 636 * 637 * Return: number of the bytes printed into buf or error 638 */ 639 static ssize_t fw_status_show(struct device *device, 640 struct device_attribute *attr, char *buf) 641 { 642 struct mei_device *dev = dev_get_drvdata(device); 643 struct mei_fw_status fw_status; 644 int err, i; 645 ssize_t cnt = 0; 646 647 mutex_lock(&dev->device_lock); 648 err = mei_fw_status(dev, &fw_status); 649 mutex_unlock(&dev->device_lock); 650 if (err) { 651 dev_err(device, "read fw_status error = %d\n", err); 652 return err; 653 } 654 655 for (i = 0; i < fw_status.count; i++) 656 cnt += scnprintf(buf + cnt, PAGE_SIZE - cnt, "%08X\n", 657 fw_status.status[i]); 658 return cnt; 659 } 660 static DEVICE_ATTR_RO(fw_status); 661 662 static struct attribute *mei_attrs[] = { 663 &dev_attr_fw_status.attr, 664 NULL 665 }; 666 ATTRIBUTE_GROUPS(mei); 667 668 /* 669 * file operations structure will be used for mei char device. 670 */ 671 static const struct file_operations mei_fops = { 672 .owner = THIS_MODULE, 673 .read = mei_read, 674 .unlocked_ioctl = mei_ioctl, 675 #ifdef CONFIG_COMPAT 676 .compat_ioctl = mei_compat_ioctl, 677 #endif 678 .open = mei_open, 679 .release = mei_release, 680 .write = mei_write, 681 .poll = mei_poll, 682 .llseek = no_llseek 683 }; 684 685 static struct class *mei_class; 686 static dev_t mei_devt; 687 #define MEI_MAX_DEVS MINORMASK 688 static DEFINE_MUTEX(mei_minor_lock); 689 static DEFINE_IDR(mei_idr); 690 691 /** 692 * mei_minor_get - obtain next free device minor number 693 * 694 * @dev: device pointer 695 * 696 * Return: allocated minor, or -ENOSPC if no free minor left 697 */ 698 static int mei_minor_get(struct mei_device *dev) 699 { 700 int ret; 701 702 mutex_lock(&mei_minor_lock); 703 ret = idr_alloc(&mei_idr, dev, 0, MEI_MAX_DEVS, GFP_KERNEL); 704 if (ret >= 0) 705 dev->minor = ret; 706 else if (ret == -ENOSPC) 707 dev_err(dev->dev, "too many mei devices\n"); 708 709 mutex_unlock(&mei_minor_lock); 710 return ret; 711 } 712 713 /** 714 * mei_minor_free - mark device minor number as free 715 * 716 * @dev: device pointer 717 */ 718 static void mei_minor_free(struct mei_device *dev) 719 { 720 mutex_lock(&mei_minor_lock); 721 idr_remove(&mei_idr, dev->minor); 722 mutex_unlock(&mei_minor_lock); 723 } 724 725 int mei_register(struct mei_device *dev, struct device *parent) 726 { 727 struct device *clsdev; /* class device */ 728 int ret, devno; 729 730 ret = mei_minor_get(dev); 731 if (ret < 0) 732 return ret; 733 734 /* Fill in the data structures */ 735 devno = MKDEV(MAJOR(mei_devt), dev->minor); 736 cdev_init(&dev->cdev, &mei_fops); 737 dev->cdev.owner = mei_fops.owner; 738 739 /* Add the device */ 740 ret = cdev_add(&dev->cdev, devno, 1); 741 if (ret) { 742 dev_err(parent, "unable to add device %d:%d\n", 743 MAJOR(mei_devt), dev->minor); 744 goto err_dev_add; 745 } 746 747 clsdev = device_create_with_groups(mei_class, parent, devno, 748 dev, mei_groups, 749 "mei%d", dev->minor); 750 751 if (IS_ERR(clsdev)) { 752 dev_err(parent, "unable to create device %d:%d\n", 753 MAJOR(mei_devt), dev->minor); 754 ret = PTR_ERR(clsdev); 755 goto err_dev_create; 756 } 757 758 ret = mei_dbgfs_register(dev, dev_name(clsdev)); 759 if (ret) { 760 dev_err(clsdev, "cannot register debugfs ret = %d\n", ret); 761 goto err_dev_dbgfs; 762 } 763 764 return 0; 765 766 err_dev_dbgfs: 767 device_destroy(mei_class, devno); 768 err_dev_create: 769 cdev_del(&dev->cdev); 770 err_dev_add: 771 mei_minor_free(dev); 772 return ret; 773 } 774 EXPORT_SYMBOL_GPL(mei_register); 775 776 void mei_deregister(struct mei_device *dev) 777 { 778 int devno; 779 780 devno = dev->cdev.dev; 781 cdev_del(&dev->cdev); 782 783 mei_dbgfs_deregister(dev); 784 785 device_destroy(mei_class, devno); 786 787 mei_minor_free(dev); 788 } 789 EXPORT_SYMBOL_GPL(mei_deregister); 790 791 static int __init mei_init(void) 792 { 793 int ret; 794 795 mei_class = class_create(THIS_MODULE, "mei"); 796 if (IS_ERR(mei_class)) { 797 pr_err("couldn't create class\n"); 798 ret = PTR_ERR(mei_class); 799 goto err; 800 } 801 802 ret = alloc_chrdev_region(&mei_devt, 0, MEI_MAX_DEVS, "mei"); 803 if (ret < 0) { 804 pr_err("unable to allocate char dev region\n"); 805 goto err_class; 806 } 807 808 ret = mei_cl_bus_init(); 809 if (ret < 0) { 810 pr_err("unable to initialize bus\n"); 811 goto err_chrdev; 812 } 813 814 return 0; 815 816 err_chrdev: 817 unregister_chrdev_region(mei_devt, MEI_MAX_DEVS); 818 err_class: 819 class_destroy(mei_class); 820 err: 821 return ret; 822 } 823 824 static void __exit mei_exit(void) 825 { 826 unregister_chrdev_region(mei_devt, MEI_MAX_DEVS); 827 class_destroy(mei_class); 828 mei_cl_bus_exit(); 829 } 830 831 module_init(mei_init); 832 module_exit(mei_exit); 833 834 MODULE_AUTHOR("Intel Corporation"); 835 MODULE_DESCRIPTION("Intel(R) Management Engine Interface"); 836 MODULE_LICENSE("GPL v2"); 837 838