1 /* 2 * Copyright (C) 2012 IBM Corporation 3 * 4 * Author: Ashley Lai <ashleydlai@gmail.com> 5 * 6 * Maintained by: <tpmdd-devel@lists.sourceforge.net> 7 * 8 * Device driver for TCG/TCPA TPM (trusted platform module). 9 * Specifications at www.trustedcomputinggroup.org 10 * 11 * This program is free software; you can redistribute it and/or 12 * modify it under the terms of the GNU General Public License as 13 * published by the Free Software Foundation, version 2 of the 14 * License. 15 * 16 */ 17 18 #include <linux/dma-mapping.h> 19 #include <linux/dmapool.h> 20 #include <linux/slab.h> 21 #include <asm/vio.h> 22 #include <asm/irq.h> 23 #include <linux/types.h> 24 #include <linux/list.h> 25 #include <linux/spinlock.h> 26 #include <linux/interrupt.h> 27 #include <linux/wait.h> 28 #include <asm/prom.h> 29 30 #include "tpm.h" 31 #include "tpm_ibmvtpm.h" 32 33 static const char tpm_ibmvtpm_driver_name[] = "tpm_ibmvtpm"; 34 35 static struct vio_device_id tpm_ibmvtpm_device_table[] = { 36 { "IBM,vtpm", "IBM,vtpm"}, 37 { "", "" } 38 }; 39 MODULE_DEVICE_TABLE(vio, tpm_ibmvtpm_device_table); 40 41 /** 42 * ibmvtpm_send_crq - Send a CRQ request 43 * @vdev: vio device struct 44 * @w1: first word 45 * @w2: second word 46 * 47 * Return value: 48 * 0 -Sucess 49 * Non-zero - Failure 50 */ 51 static int ibmvtpm_send_crq(struct vio_dev *vdev, u64 w1, u64 w2) 52 { 53 return plpar_hcall_norets(H_SEND_CRQ, vdev->unit_address, w1, w2); 54 } 55 56 /** 57 * tpm_ibmvtpm_recv - Receive data after send 58 * @chip: tpm chip struct 59 * @buf: buffer to read 60 * count: size of buffer 61 * 62 * Return value: 63 * Number of bytes read 64 */ 65 static int tpm_ibmvtpm_recv(struct tpm_chip *chip, u8 *buf, size_t count) 66 { 67 struct ibmvtpm_dev *ibmvtpm = dev_get_drvdata(&chip->dev); 68 u16 len; 69 int sig; 70 71 if (!ibmvtpm->rtce_buf) { 72 dev_err(ibmvtpm->dev, "ibmvtpm device is not ready\n"); 73 return 0; 74 } 75 76 sig = wait_event_interruptible(ibmvtpm->wq, !ibmvtpm->tpm_processing_cmd); 77 if (sig) 78 return -EINTR; 79 80 len = ibmvtpm->res_len; 81 82 if (count < len) { 83 dev_err(ibmvtpm->dev, 84 "Invalid size in recv: count=%zd, crq_size=%d\n", 85 count, len); 86 return -EIO; 87 } 88 89 spin_lock(&ibmvtpm->rtce_lock); 90 memcpy((void *)buf, (void *)ibmvtpm->rtce_buf, len); 91 memset(ibmvtpm->rtce_buf, 0, len); 92 ibmvtpm->res_len = 0; 93 spin_unlock(&ibmvtpm->rtce_lock); 94 return len; 95 } 96 97 /** 98 * tpm_ibmvtpm_send - Send tpm request 99 * @chip: tpm chip struct 100 * @buf: buffer contains data to send 101 * count: size of buffer 102 * 103 * Return value: 104 * Number of bytes sent 105 */ 106 static int tpm_ibmvtpm_send(struct tpm_chip *chip, u8 *buf, size_t count) 107 { 108 struct ibmvtpm_dev *ibmvtpm = dev_get_drvdata(&chip->dev); 109 struct ibmvtpm_crq crq; 110 __be64 *word = (__be64 *)&crq; 111 int rc, sig; 112 113 if (!ibmvtpm->rtce_buf) { 114 dev_err(ibmvtpm->dev, "ibmvtpm device is not ready\n"); 115 return 0; 116 } 117 118 if (count > ibmvtpm->rtce_size) { 119 dev_err(ibmvtpm->dev, 120 "Invalid size in send: count=%zd, rtce_size=%d\n", 121 count, ibmvtpm->rtce_size); 122 return -EIO; 123 } 124 125 if (ibmvtpm->tpm_processing_cmd) { 126 dev_info(ibmvtpm->dev, 127 "Need to wait for TPM to finish\n"); 128 /* wait for previous command to finish */ 129 sig = wait_event_interruptible(ibmvtpm->wq, !ibmvtpm->tpm_processing_cmd); 130 if (sig) 131 return -EINTR; 132 } 133 134 spin_lock(&ibmvtpm->rtce_lock); 135 ibmvtpm->res_len = 0; 136 memcpy((void *)ibmvtpm->rtce_buf, (void *)buf, count); 137 crq.valid = (u8)IBMVTPM_VALID_CMD; 138 crq.msg = (u8)VTPM_TPM_COMMAND; 139 crq.len = cpu_to_be16(count); 140 crq.data = cpu_to_be32(ibmvtpm->rtce_dma_handle); 141 142 /* 143 * set the processing flag before the Hcall, since we may get the 144 * result (interrupt) before even being able to check rc. 145 */ 146 ibmvtpm->tpm_processing_cmd = true; 147 148 rc = ibmvtpm_send_crq(ibmvtpm->vdev, be64_to_cpu(word[0]), 149 be64_to_cpu(word[1])); 150 if (rc != H_SUCCESS) { 151 dev_err(ibmvtpm->dev, "tpm_ibmvtpm_send failed rc=%d\n", rc); 152 rc = 0; 153 ibmvtpm->tpm_processing_cmd = false; 154 } else 155 rc = count; 156 157 spin_unlock(&ibmvtpm->rtce_lock); 158 return rc; 159 } 160 161 static void tpm_ibmvtpm_cancel(struct tpm_chip *chip) 162 { 163 return; 164 } 165 166 static u8 tpm_ibmvtpm_status(struct tpm_chip *chip) 167 { 168 return 0; 169 } 170 171 /** 172 * ibmvtpm_crq_get_rtce_size - Send a CRQ request to get rtce size 173 * @ibmvtpm: vtpm device struct 174 * 175 * Return value: 176 * 0 - Success 177 * Non-zero - Failure 178 */ 179 static int ibmvtpm_crq_get_rtce_size(struct ibmvtpm_dev *ibmvtpm) 180 { 181 struct ibmvtpm_crq crq; 182 u64 *buf = (u64 *) &crq; 183 int rc; 184 185 crq.valid = (u8)IBMVTPM_VALID_CMD; 186 crq.msg = (u8)VTPM_GET_RTCE_BUFFER_SIZE; 187 188 rc = ibmvtpm_send_crq(ibmvtpm->vdev, cpu_to_be64(buf[0]), 189 cpu_to_be64(buf[1])); 190 if (rc != H_SUCCESS) 191 dev_err(ibmvtpm->dev, 192 "ibmvtpm_crq_get_rtce_size failed rc=%d\n", rc); 193 194 return rc; 195 } 196 197 /** 198 * ibmvtpm_crq_get_version - Send a CRQ request to get vtpm version 199 * - Note that this is vtpm version and not tpm version 200 * @ibmvtpm: vtpm device struct 201 * 202 * Return value: 203 * 0 - Success 204 * Non-zero - Failure 205 */ 206 static int ibmvtpm_crq_get_version(struct ibmvtpm_dev *ibmvtpm) 207 { 208 struct ibmvtpm_crq crq; 209 u64 *buf = (u64 *) &crq; 210 int rc; 211 212 crq.valid = (u8)IBMVTPM_VALID_CMD; 213 crq.msg = (u8)VTPM_GET_VERSION; 214 215 rc = ibmvtpm_send_crq(ibmvtpm->vdev, cpu_to_be64(buf[0]), 216 cpu_to_be64(buf[1])); 217 if (rc != H_SUCCESS) 218 dev_err(ibmvtpm->dev, 219 "ibmvtpm_crq_get_version failed rc=%d\n", rc); 220 221 return rc; 222 } 223 224 /** 225 * ibmvtpm_crq_send_init_complete - Send a CRQ initialize complete message 226 * @ibmvtpm: vtpm device struct 227 * 228 * Return value: 229 * 0 - Success 230 * Non-zero - Failure 231 */ 232 static int ibmvtpm_crq_send_init_complete(struct ibmvtpm_dev *ibmvtpm) 233 { 234 int rc; 235 236 rc = ibmvtpm_send_crq(ibmvtpm->vdev, INIT_CRQ_COMP_CMD, 0); 237 if (rc != H_SUCCESS) 238 dev_err(ibmvtpm->dev, 239 "ibmvtpm_crq_send_init_complete failed rc=%d\n", rc); 240 241 return rc; 242 } 243 244 /** 245 * ibmvtpm_crq_send_init - Send a CRQ initialize message 246 * @ibmvtpm: vtpm device struct 247 * 248 * Return value: 249 * 0 - Success 250 * Non-zero - Failure 251 */ 252 static int ibmvtpm_crq_send_init(struct ibmvtpm_dev *ibmvtpm) 253 { 254 int rc; 255 256 rc = ibmvtpm_send_crq(ibmvtpm->vdev, INIT_CRQ_CMD, 0); 257 if (rc != H_SUCCESS) 258 dev_err(ibmvtpm->dev, 259 "ibmvtpm_crq_send_init failed rc=%d\n", rc); 260 261 return rc; 262 } 263 264 /** 265 * tpm_ibmvtpm_remove - ibm vtpm remove entry point 266 * @vdev: vio device struct 267 * 268 * Return value: 269 * 0 270 */ 271 static int tpm_ibmvtpm_remove(struct vio_dev *vdev) 272 { 273 struct tpm_chip *chip = dev_get_drvdata(&vdev->dev); 274 struct ibmvtpm_dev *ibmvtpm = dev_get_drvdata(&chip->dev); 275 int rc = 0; 276 277 tpm_chip_unregister(chip); 278 279 free_irq(vdev->irq, ibmvtpm); 280 281 do { 282 if (rc) 283 msleep(100); 284 rc = plpar_hcall_norets(H_FREE_CRQ, vdev->unit_address); 285 } while (rc == H_BUSY || H_IS_LONG_BUSY(rc)); 286 287 dma_unmap_single(ibmvtpm->dev, ibmvtpm->crq_dma_handle, 288 CRQ_RES_BUF_SIZE, DMA_BIDIRECTIONAL); 289 free_page((unsigned long)ibmvtpm->crq_queue.crq_addr); 290 291 if (ibmvtpm->rtce_buf) { 292 dma_unmap_single(ibmvtpm->dev, ibmvtpm->rtce_dma_handle, 293 ibmvtpm->rtce_size, DMA_BIDIRECTIONAL); 294 kfree(ibmvtpm->rtce_buf); 295 } 296 297 kfree(ibmvtpm); 298 299 return 0; 300 } 301 302 /** 303 * tpm_ibmvtpm_get_desired_dma - Get DMA size needed by this driver 304 * @vdev: vio device struct 305 * 306 * Return value: 307 * Number of bytes the driver needs to DMA map 308 */ 309 static unsigned long tpm_ibmvtpm_get_desired_dma(struct vio_dev *vdev) 310 { 311 struct tpm_chip *chip = dev_get_drvdata(&vdev->dev); 312 struct ibmvtpm_dev *ibmvtpm = dev_get_drvdata(&chip->dev); 313 314 /* ibmvtpm initializes at probe time, so the data we are 315 * asking for may not be set yet. Estimate that 4K required 316 * for TCE-mapped buffer in addition to CRQ. 317 */ 318 if (!ibmvtpm) 319 return CRQ_RES_BUF_SIZE + PAGE_SIZE; 320 321 return CRQ_RES_BUF_SIZE + ibmvtpm->rtce_size; 322 } 323 324 /** 325 * tpm_ibmvtpm_suspend - Suspend 326 * @dev: device struct 327 * 328 * Return value: 329 * 0 330 */ 331 static int tpm_ibmvtpm_suspend(struct device *dev) 332 { 333 struct tpm_chip *chip = dev_get_drvdata(dev); 334 struct ibmvtpm_dev *ibmvtpm = dev_get_drvdata(&chip->dev); 335 struct ibmvtpm_crq crq; 336 u64 *buf = (u64 *) &crq; 337 int rc = 0; 338 339 crq.valid = (u8)IBMVTPM_VALID_CMD; 340 crq.msg = (u8)VTPM_PREPARE_TO_SUSPEND; 341 342 rc = ibmvtpm_send_crq(ibmvtpm->vdev, cpu_to_be64(buf[0]), 343 cpu_to_be64(buf[1])); 344 if (rc != H_SUCCESS) 345 dev_err(ibmvtpm->dev, 346 "tpm_ibmvtpm_suspend failed rc=%d\n", rc); 347 348 return rc; 349 } 350 351 /** 352 * ibmvtpm_reset_crq - Reset CRQ 353 * @ibmvtpm: ibm vtpm struct 354 * 355 * Return value: 356 * 0 - Success 357 * Non-zero - Failure 358 */ 359 static int ibmvtpm_reset_crq(struct ibmvtpm_dev *ibmvtpm) 360 { 361 int rc = 0; 362 363 do { 364 if (rc) 365 msleep(100); 366 rc = plpar_hcall_norets(H_FREE_CRQ, 367 ibmvtpm->vdev->unit_address); 368 } while (rc == H_BUSY || H_IS_LONG_BUSY(rc)); 369 370 memset(ibmvtpm->crq_queue.crq_addr, 0, CRQ_RES_BUF_SIZE); 371 ibmvtpm->crq_queue.index = 0; 372 373 return plpar_hcall_norets(H_REG_CRQ, ibmvtpm->vdev->unit_address, 374 ibmvtpm->crq_dma_handle, CRQ_RES_BUF_SIZE); 375 } 376 377 /** 378 * tpm_ibmvtpm_resume - Resume from suspend 379 * @dev: device struct 380 * 381 * Return value: 382 * 0 383 */ 384 static int tpm_ibmvtpm_resume(struct device *dev) 385 { 386 struct tpm_chip *chip = dev_get_drvdata(dev); 387 struct ibmvtpm_dev *ibmvtpm = dev_get_drvdata(&chip->dev); 388 int rc = 0; 389 390 do { 391 if (rc) 392 msleep(100); 393 rc = plpar_hcall_norets(H_ENABLE_CRQ, 394 ibmvtpm->vdev->unit_address); 395 } while (rc == H_IN_PROGRESS || rc == H_BUSY || H_IS_LONG_BUSY(rc)); 396 397 if (rc) { 398 dev_err(dev, "Error enabling ibmvtpm rc=%d\n", rc); 399 return rc; 400 } 401 402 rc = vio_enable_interrupts(ibmvtpm->vdev); 403 if (rc) { 404 dev_err(dev, "Error vio_enable_interrupts rc=%d\n", rc); 405 return rc; 406 } 407 408 rc = ibmvtpm_crq_send_init(ibmvtpm); 409 if (rc) 410 dev_err(dev, "Error send_init rc=%d\n", rc); 411 412 return rc; 413 } 414 415 static bool tpm_ibmvtpm_req_canceled(struct tpm_chip *chip, u8 status) 416 { 417 return (status == 0); 418 } 419 420 static const struct tpm_class_ops tpm_ibmvtpm = { 421 .recv = tpm_ibmvtpm_recv, 422 .send = tpm_ibmvtpm_send, 423 .cancel = tpm_ibmvtpm_cancel, 424 .status = tpm_ibmvtpm_status, 425 .req_complete_mask = 0, 426 .req_complete_val = 0, 427 .req_canceled = tpm_ibmvtpm_req_canceled, 428 }; 429 430 static const struct dev_pm_ops tpm_ibmvtpm_pm_ops = { 431 .suspend = tpm_ibmvtpm_suspend, 432 .resume = tpm_ibmvtpm_resume, 433 }; 434 435 /** 436 * ibmvtpm_crq_get_next - Get next responded crq 437 * @ibmvtpm vtpm device struct 438 * 439 * Return value: 440 * vtpm crq pointer 441 */ 442 static struct ibmvtpm_crq *ibmvtpm_crq_get_next(struct ibmvtpm_dev *ibmvtpm) 443 { 444 struct ibmvtpm_crq_queue *crq_q = &ibmvtpm->crq_queue; 445 struct ibmvtpm_crq *crq = &crq_q->crq_addr[crq_q->index]; 446 447 if (crq->valid & VTPM_MSG_RES) { 448 if (++crq_q->index == crq_q->num_entry) 449 crq_q->index = 0; 450 smp_rmb(); 451 } else 452 crq = NULL; 453 return crq; 454 } 455 456 /** 457 * ibmvtpm_crq_process - Process responded crq 458 * @crq crq to be processed 459 * @ibmvtpm vtpm device struct 460 * 461 * Return value: 462 * Nothing 463 */ 464 static void ibmvtpm_crq_process(struct ibmvtpm_crq *crq, 465 struct ibmvtpm_dev *ibmvtpm) 466 { 467 int rc = 0; 468 469 switch (crq->valid) { 470 case VALID_INIT_CRQ: 471 switch (crq->msg) { 472 case INIT_CRQ_RES: 473 dev_info(ibmvtpm->dev, "CRQ initialized\n"); 474 rc = ibmvtpm_crq_send_init_complete(ibmvtpm); 475 if (rc) 476 dev_err(ibmvtpm->dev, "Unable to send CRQ init complete rc=%d\n", rc); 477 return; 478 case INIT_CRQ_COMP_RES: 479 dev_info(ibmvtpm->dev, 480 "CRQ initialization completed\n"); 481 return; 482 default: 483 dev_err(ibmvtpm->dev, "Unknown crq message type: %d\n", crq->msg); 484 return; 485 } 486 case IBMVTPM_VALID_CMD: 487 switch (crq->msg) { 488 case VTPM_GET_RTCE_BUFFER_SIZE_RES: 489 if (be16_to_cpu(crq->len) <= 0) { 490 dev_err(ibmvtpm->dev, "Invalid rtce size\n"); 491 return; 492 } 493 ibmvtpm->rtce_size = be16_to_cpu(crq->len); 494 ibmvtpm->rtce_buf = kmalloc(ibmvtpm->rtce_size, 495 GFP_ATOMIC); 496 if (!ibmvtpm->rtce_buf) { 497 dev_err(ibmvtpm->dev, "Failed to allocate memory for rtce buffer\n"); 498 return; 499 } 500 501 ibmvtpm->rtce_dma_handle = dma_map_single(ibmvtpm->dev, 502 ibmvtpm->rtce_buf, ibmvtpm->rtce_size, 503 DMA_BIDIRECTIONAL); 504 505 if (dma_mapping_error(ibmvtpm->dev, 506 ibmvtpm->rtce_dma_handle)) { 507 kfree(ibmvtpm->rtce_buf); 508 ibmvtpm->rtce_buf = NULL; 509 dev_err(ibmvtpm->dev, "Failed to dma map rtce buffer\n"); 510 } 511 512 return; 513 case VTPM_GET_VERSION_RES: 514 ibmvtpm->vtpm_version = be32_to_cpu(crq->data); 515 return; 516 case VTPM_TPM_COMMAND_RES: 517 /* len of the data in rtce buffer */ 518 ibmvtpm->res_len = be16_to_cpu(crq->len); 519 ibmvtpm->tpm_processing_cmd = false; 520 wake_up_interruptible(&ibmvtpm->wq); 521 return; 522 default: 523 return; 524 } 525 } 526 return; 527 } 528 529 /** 530 * ibmvtpm_interrupt - Interrupt handler 531 * @irq: irq number to handle 532 * @vtpm_instance: vtpm that received interrupt 533 * 534 * Returns: 535 * IRQ_HANDLED 536 **/ 537 static irqreturn_t ibmvtpm_interrupt(int irq, void *vtpm_instance) 538 { 539 struct ibmvtpm_dev *ibmvtpm = (struct ibmvtpm_dev *) vtpm_instance; 540 struct ibmvtpm_crq *crq; 541 542 /* while loop is needed for initial setup (get version and 543 * get rtce_size). There should be only one tpm request at any 544 * given time. 545 */ 546 while ((crq = ibmvtpm_crq_get_next(ibmvtpm)) != NULL) { 547 ibmvtpm_crq_process(crq, ibmvtpm); 548 crq->valid = 0; 549 smp_wmb(); 550 } 551 552 return IRQ_HANDLED; 553 } 554 555 /** 556 * tpm_ibmvtpm_probe - ibm vtpm initialize entry point 557 * @vio_dev: vio device struct 558 * @id: vio device id struct 559 * 560 * Return value: 561 * 0 - Success 562 * Non-zero - Failure 563 */ 564 static int tpm_ibmvtpm_probe(struct vio_dev *vio_dev, 565 const struct vio_device_id *id) 566 { 567 struct ibmvtpm_dev *ibmvtpm; 568 struct device *dev = &vio_dev->dev; 569 struct ibmvtpm_crq_queue *crq_q; 570 struct tpm_chip *chip; 571 int rc = -ENOMEM, rc1; 572 573 chip = tpmm_chip_alloc(dev, &tpm_ibmvtpm); 574 if (IS_ERR(chip)) 575 return PTR_ERR(chip); 576 577 ibmvtpm = kzalloc(sizeof(struct ibmvtpm_dev), GFP_KERNEL); 578 if (!ibmvtpm) { 579 dev_err(dev, "kzalloc for ibmvtpm failed\n"); 580 goto cleanup; 581 } 582 583 ibmvtpm->dev = dev; 584 ibmvtpm->vdev = vio_dev; 585 586 crq_q = &ibmvtpm->crq_queue; 587 crq_q->crq_addr = (struct ibmvtpm_crq *)get_zeroed_page(GFP_KERNEL); 588 if (!crq_q->crq_addr) { 589 dev_err(dev, "Unable to allocate memory for crq_addr\n"); 590 goto cleanup; 591 } 592 593 crq_q->num_entry = CRQ_RES_BUF_SIZE / sizeof(*crq_q->crq_addr); 594 ibmvtpm->crq_dma_handle = dma_map_single(dev, crq_q->crq_addr, 595 CRQ_RES_BUF_SIZE, 596 DMA_BIDIRECTIONAL); 597 598 if (dma_mapping_error(dev, ibmvtpm->crq_dma_handle)) { 599 dev_err(dev, "dma mapping failed\n"); 600 goto cleanup; 601 } 602 603 rc = plpar_hcall_norets(H_REG_CRQ, vio_dev->unit_address, 604 ibmvtpm->crq_dma_handle, CRQ_RES_BUF_SIZE); 605 if (rc == H_RESOURCE) 606 rc = ibmvtpm_reset_crq(ibmvtpm); 607 608 if (rc) { 609 dev_err(dev, "Unable to register CRQ rc=%d\n", rc); 610 goto reg_crq_cleanup; 611 } 612 613 rc = request_irq(vio_dev->irq, ibmvtpm_interrupt, 0, 614 tpm_ibmvtpm_driver_name, ibmvtpm); 615 if (rc) { 616 dev_err(dev, "Error %d register irq 0x%x\n", rc, vio_dev->irq); 617 goto init_irq_cleanup; 618 } 619 620 rc = vio_enable_interrupts(vio_dev); 621 if (rc) { 622 dev_err(dev, "Error %d enabling interrupts\n", rc); 623 goto init_irq_cleanup; 624 } 625 626 init_waitqueue_head(&ibmvtpm->wq); 627 628 crq_q->index = 0; 629 630 dev_set_drvdata(&chip->dev, ibmvtpm); 631 632 spin_lock_init(&ibmvtpm->rtce_lock); 633 634 rc = ibmvtpm_crq_send_init(ibmvtpm); 635 if (rc) 636 goto init_irq_cleanup; 637 638 rc = ibmvtpm_crq_get_version(ibmvtpm); 639 if (rc) 640 goto init_irq_cleanup; 641 642 rc = ibmvtpm_crq_get_rtce_size(ibmvtpm); 643 if (rc) 644 goto init_irq_cleanup; 645 646 return tpm_chip_register(chip); 647 init_irq_cleanup: 648 do { 649 rc1 = plpar_hcall_norets(H_FREE_CRQ, vio_dev->unit_address); 650 } while (rc1 == H_BUSY || H_IS_LONG_BUSY(rc1)); 651 reg_crq_cleanup: 652 dma_unmap_single(dev, ibmvtpm->crq_dma_handle, CRQ_RES_BUF_SIZE, 653 DMA_BIDIRECTIONAL); 654 cleanup: 655 if (ibmvtpm) { 656 if (crq_q->crq_addr) 657 free_page((unsigned long)crq_q->crq_addr); 658 kfree(ibmvtpm); 659 } 660 661 return rc; 662 } 663 664 static struct vio_driver ibmvtpm_driver = { 665 .id_table = tpm_ibmvtpm_device_table, 666 .probe = tpm_ibmvtpm_probe, 667 .remove = tpm_ibmvtpm_remove, 668 .get_desired_dma = tpm_ibmvtpm_get_desired_dma, 669 .name = tpm_ibmvtpm_driver_name, 670 .pm = &tpm_ibmvtpm_pm_ops, 671 }; 672 673 /** 674 * ibmvtpm_module_init - Initialize ibm vtpm module 675 * 676 * Return value: 677 * 0 -Success 678 * Non-zero - Failure 679 */ 680 static int __init ibmvtpm_module_init(void) 681 { 682 return vio_register_driver(&ibmvtpm_driver); 683 } 684 685 /** 686 * ibmvtpm_module_exit - Teardown ibm vtpm module 687 * 688 * Return value: 689 * Nothing 690 */ 691 static void __exit ibmvtpm_module_exit(void) 692 { 693 vio_unregister_driver(&ibmvtpm_driver); 694 } 695 696 module_init(ibmvtpm_module_init); 697 module_exit(ibmvtpm_module_exit); 698 699 MODULE_AUTHOR("adlai@us.ibm.com"); 700 MODULE_DESCRIPTION("IBM vTPM Driver"); 701 MODULE_VERSION("1.0"); 702 MODULE_LICENSE("GPL"); 703