1 /* 2 * character device driver for reading z/VM system service records 3 * 4 * 5 * Copyright IBM Corp. 2004, 2009 6 * character device driver for reading z/VM system service records, 7 * Version 1.0 8 * Author(s): Xenia Tkatschow <xenia@us.ibm.com> 9 * Stefan Weinhuber <wein@de.ibm.com> 10 * 11 */ 12 13 #define KMSG_COMPONENT "vmlogrdr" 14 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt 15 16 #include <linux/module.h> 17 #include <linux/init.h> 18 #include <linux/slab.h> 19 #include <linux/errno.h> 20 #include <linux/types.h> 21 #include <linux/interrupt.h> 22 #include <linux/spinlock.h> 23 #include <linux/atomic.h> 24 #include <asm/uaccess.h> 25 #include <asm/cpcmd.h> 26 #include <asm/debug.h> 27 #include <asm/ebcdic.h> 28 #include <net/iucv/iucv.h> 29 #include <linux/kmod.h> 30 #include <linux/cdev.h> 31 #include <linux/device.h> 32 #include <linux/string.h> 33 34 MODULE_AUTHOR 35 ("(C) 2004 IBM Corporation by Xenia Tkatschow (xenia@us.ibm.com)\n" 36 " Stefan Weinhuber (wein@de.ibm.com)"); 37 MODULE_DESCRIPTION ("Character device driver for reading z/VM " 38 "system service records."); 39 MODULE_LICENSE("GPL"); 40 41 42 /* 43 * The size of the buffer for iucv data transfer is one page, 44 * but in addition to the data we read from iucv we also 45 * place an integer and some characters into that buffer, 46 * so the maximum size for record data is a little less then 47 * one page. 48 */ 49 #define NET_BUFFER_SIZE (PAGE_SIZE - sizeof(int) - sizeof(FENCE)) 50 51 /* 52 * The elements that are concurrently accessed by bottom halves are 53 * connection_established, iucv_path_severed, local_interrupt_buffer 54 * and receive_ready. The first three can be protected by 55 * priv_lock. receive_ready is atomic, so it can be incremented and 56 * decremented without holding a lock. 57 * The variable dev_in_use needs to be protected by the lock, since 58 * it's a flag used by open to make sure that the device is opened only 59 * by one user at the same time. 60 */ 61 struct vmlogrdr_priv_t { 62 char system_service[8]; 63 char internal_name[8]; 64 char recording_name[8]; 65 struct iucv_path *path; 66 int connection_established; 67 int iucv_path_severed; 68 struct iucv_message local_interrupt_buffer; 69 atomic_t receive_ready; 70 int minor_num; 71 char * buffer; 72 char * current_position; 73 int remaining; 74 ulong residual_length; 75 int buffer_free; 76 int dev_in_use; /* 1: already opened, 0: not opened*/ 77 spinlock_t priv_lock; 78 struct device *device; 79 struct device *class_device; 80 int autorecording; 81 int autopurge; 82 }; 83 84 85 /* 86 * File operation structure for vmlogrdr devices 87 */ 88 static int vmlogrdr_open(struct inode *, struct file *); 89 static int vmlogrdr_release(struct inode *, struct file *); 90 static ssize_t vmlogrdr_read (struct file *filp, char __user *data, 91 size_t count, loff_t * ppos); 92 93 static const struct file_operations vmlogrdr_fops = { 94 .owner = THIS_MODULE, 95 .open = vmlogrdr_open, 96 .release = vmlogrdr_release, 97 .read = vmlogrdr_read, 98 .llseek = no_llseek, 99 }; 100 101 102 static void vmlogrdr_iucv_path_complete(struct iucv_path *, u8 ipuser[16]); 103 static void vmlogrdr_iucv_path_severed(struct iucv_path *, u8 ipuser[16]); 104 static void vmlogrdr_iucv_message_pending(struct iucv_path *, 105 struct iucv_message *); 106 107 108 static struct iucv_handler vmlogrdr_iucv_handler = { 109 .path_complete = vmlogrdr_iucv_path_complete, 110 .path_severed = vmlogrdr_iucv_path_severed, 111 .message_pending = vmlogrdr_iucv_message_pending, 112 }; 113 114 115 static DECLARE_WAIT_QUEUE_HEAD(conn_wait_queue); 116 static DECLARE_WAIT_QUEUE_HEAD(read_wait_queue); 117 118 /* 119 * pointer to system service private structure 120 * minor number 0 --> logrec 121 * minor number 1 --> account 122 * minor number 2 --> symptom 123 */ 124 125 static struct vmlogrdr_priv_t sys_ser[] = { 126 { .system_service = "*LOGREC ", 127 .internal_name = "logrec", 128 .recording_name = "EREP", 129 .minor_num = 0, 130 .buffer_free = 1, 131 .priv_lock = __SPIN_LOCK_UNLOCKED(sys_ser[0].priv_lock), 132 .autorecording = 1, 133 .autopurge = 1, 134 }, 135 { .system_service = "*ACCOUNT", 136 .internal_name = "account", 137 .recording_name = "ACCOUNT", 138 .minor_num = 1, 139 .buffer_free = 1, 140 .priv_lock = __SPIN_LOCK_UNLOCKED(sys_ser[1].priv_lock), 141 .autorecording = 1, 142 .autopurge = 1, 143 }, 144 { .system_service = "*SYMPTOM", 145 .internal_name = "symptom", 146 .recording_name = "SYMPTOM", 147 .minor_num = 2, 148 .buffer_free = 1, 149 .priv_lock = __SPIN_LOCK_UNLOCKED(sys_ser[2].priv_lock), 150 .autorecording = 1, 151 .autopurge = 1, 152 } 153 }; 154 155 #define MAXMINOR (sizeof(sys_ser)/sizeof(struct vmlogrdr_priv_t)) 156 157 static char FENCE[] = {"EOR"}; 158 static int vmlogrdr_major = 0; 159 static struct cdev *vmlogrdr_cdev = NULL; 160 static int recording_class_AB; 161 162 163 static void vmlogrdr_iucv_path_complete(struct iucv_path *path, u8 ipuser[16]) 164 { 165 struct vmlogrdr_priv_t * logptr = path->private; 166 167 spin_lock(&logptr->priv_lock); 168 logptr->connection_established = 1; 169 spin_unlock(&logptr->priv_lock); 170 wake_up(&conn_wait_queue); 171 } 172 173 174 static void vmlogrdr_iucv_path_severed(struct iucv_path *path, u8 ipuser[16]) 175 { 176 struct vmlogrdr_priv_t * logptr = path->private; 177 u8 reason = (u8) ipuser[8]; 178 179 pr_err("vmlogrdr: connection severed with reason %i\n", reason); 180 181 iucv_path_sever(path, NULL); 182 kfree(path); 183 logptr->path = NULL; 184 185 spin_lock(&logptr->priv_lock); 186 logptr->connection_established = 0; 187 logptr->iucv_path_severed = 1; 188 spin_unlock(&logptr->priv_lock); 189 190 wake_up(&conn_wait_queue); 191 /* just in case we're sleeping waiting for a record */ 192 wake_up_interruptible(&read_wait_queue); 193 } 194 195 196 static void vmlogrdr_iucv_message_pending(struct iucv_path *path, 197 struct iucv_message *msg) 198 { 199 struct vmlogrdr_priv_t * logptr = path->private; 200 201 /* 202 * This function is the bottom half so it should be quick. 203 * Copy the external interrupt data into our local eib and increment 204 * the usage count 205 */ 206 spin_lock(&logptr->priv_lock); 207 memcpy(&logptr->local_interrupt_buffer, msg, sizeof(*msg)); 208 atomic_inc(&logptr->receive_ready); 209 spin_unlock(&logptr->priv_lock); 210 wake_up_interruptible(&read_wait_queue); 211 } 212 213 214 static int vmlogrdr_get_recording_class_AB(void) 215 { 216 static const char cp_command[] = "QUERY COMMAND RECORDING "; 217 char cp_response[80]; 218 char *tail; 219 int len,i; 220 221 cpcmd(cp_command, cp_response, sizeof(cp_response), NULL); 222 len = strnlen(cp_response,sizeof(cp_response)); 223 // now the parsing 224 tail=strnchr(cp_response,len,'='); 225 if (!tail) 226 return 0; 227 tail++; 228 if (!strncmp("ANY",tail,3)) 229 return 1; 230 if (!strncmp("NONE",tail,4)) 231 return 0; 232 /* 233 * expect comma separated list of classes here, if one of them 234 * is A or B return 1 otherwise 0 235 */ 236 for (i=tail-cp_response; i<len; i++) 237 if ( cp_response[i]=='A' || cp_response[i]=='B' ) 238 return 1; 239 return 0; 240 } 241 242 243 static int vmlogrdr_recording(struct vmlogrdr_priv_t * logptr, 244 int action, int purge) 245 { 246 247 char cp_command[80]; 248 char cp_response[160]; 249 char *onoff, *qid_string; 250 int rc; 251 252 onoff = ((action == 1) ? "ON" : "OFF"); 253 qid_string = ((recording_class_AB == 1) ? " QID * " : ""); 254 255 /* 256 * The recording commands needs to be called with option QID 257 * for guests that have previlege classes A or B. 258 * Purging has to be done as separate step, because recording 259 * can't be switched on as long as records are on the queue. 260 * Doing both at the same time doesn't work. 261 */ 262 if (purge && (action == 1)) { 263 memset(cp_command, 0x00, sizeof(cp_command)); 264 memset(cp_response, 0x00, sizeof(cp_response)); 265 snprintf(cp_command, sizeof(cp_command), 266 "RECORDING %s PURGE %s", 267 logptr->recording_name, 268 qid_string); 269 cpcmd(cp_command, cp_response, sizeof(cp_response), NULL); 270 } 271 272 memset(cp_command, 0x00, sizeof(cp_command)); 273 memset(cp_response, 0x00, sizeof(cp_response)); 274 snprintf(cp_command, sizeof(cp_command), "RECORDING %s %s %s", 275 logptr->recording_name, 276 onoff, 277 qid_string); 278 cpcmd(cp_command, cp_response, sizeof(cp_response), NULL); 279 /* The recording command will usually answer with 'Command complete' 280 * on success, but when the specific service was never connected 281 * before then there might be an additional informational message 282 * 'HCPCRC8072I Recording entry not found' before the 283 * 'Command complete'. So I use strstr rather then the strncmp. 284 */ 285 if (strstr(cp_response,"Command complete")) 286 rc = 0; 287 else 288 rc = -EIO; 289 /* 290 * If we turn recording off, we have to purge any remaining records 291 * afterwards, as a large number of queued records may impact z/VM 292 * performance. 293 */ 294 if (purge && (action == 0)) { 295 memset(cp_command, 0x00, sizeof(cp_command)); 296 memset(cp_response, 0x00, sizeof(cp_response)); 297 snprintf(cp_command, sizeof(cp_command), 298 "RECORDING %s PURGE %s", 299 logptr->recording_name, 300 qid_string); 301 cpcmd(cp_command, cp_response, sizeof(cp_response), NULL); 302 } 303 304 return rc; 305 } 306 307 308 static int vmlogrdr_open (struct inode *inode, struct file *filp) 309 { 310 int dev_num = 0; 311 struct vmlogrdr_priv_t * logptr = NULL; 312 int connect_rc = 0; 313 int ret; 314 315 dev_num = iminor(inode); 316 if (dev_num >= MAXMINOR) 317 return -ENODEV; 318 logptr = &sys_ser[dev_num]; 319 320 /* 321 * only allow for blocking reads to be open 322 */ 323 if (filp->f_flags & O_NONBLOCK) 324 return -EOPNOTSUPP; 325 326 /* Besure this device hasn't already been opened */ 327 spin_lock_bh(&logptr->priv_lock); 328 if (logptr->dev_in_use) { 329 spin_unlock_bh(&logptr->priv_lock); 330 return -EBUSY; 331 } 332 logptr->dev_in_use = 1; 333 logptr->connection_established = 0; 334 logptr->iucv_path_severed = 0; 335 atomic_set(&logptr->receive_ready, 0); 336 logptr->buffer_free = 1; 337 spin_unlock_bh(&logptr->priv_lock); 338 339 /* set the file options */ 340 filp->private_data = logptr; 341 filp->f_op = &vmlogrdr_fops; 342 343 /* start recording for this service*/ 344 if (logptr->autorecording) { 345 ret = vmlogrdr_recording(logptr,1,logptr->autopurge); 346 if (ret) 347 pr_warning("vmlogrdr: failed to start " 348 "recording automatically\n"); 349 } 350 351 /* create connection to the system service */ 352 logptr->path = iucv_path_alloc(10, 0, GFP_KERNEL); 353 if (!logptr->path) 354 goto out_dev; 355 connect_rc = iucv_path_connect(logptr->path, &vmlogrdr_iucv_handler, 356 logptr->system_service, NULL, NULL, 357 logptr); 358 if (connect_rc) { 359 pr_err("vmlogrdr: iucv connection to %s " 360 "failed with rc %i \n", 361 logptr->system_service, connect_rc); 362 goto out_path; 363 } 364 365 /* We've issued the connect and now we must wait for a 366 * ConnectionComplete or ConnectinSevered Interrupt 367 * before we can continue to process. 368 */ 369 wait_event(conn_wait_queue, (logptr->connection_established) 370 || (logptr->iucv_path_severed)); 371 if (logptr->iucv_path_severed) 372 goto out_record; 373 nonseekable_open(inode, filp); 374 return 0; 375 376 out_record: 377 if (logptr->autorecording) 378 vmlogrdr_recording(logptr,0,logptr->autopurge); 379 out_path: 380 kfree(logptr->path); /* kfree(NULL) is ok. */ 381 logptr->path = NULL; 382 out_dev: 383 logptr->dev_in_use = 0; 384 return -EIO; 385 } 386 387 388 static int vmlogrdr_release (struct inode *inode, struct file *filp) 389 { 390 int ret; 391 392 struct vmlogrdr_priv_t * logptr = filp->private_data; 393 394 iucv_path_sever(logptr->path, NULL); 395 kfree(logptr->path); 396 logptr->path = NULL; 397 if (logptr->autorecording) { 398 ret = vmlogrdr_recording(logptr,0,logptr->autopurge); 399 if (ret) 400 pr_warning("vmlogrdr: failed to stop " 401 "recording automatically\n"); 402 } 403 logptr->dev_in_use = 0; 404 405 return 0; 406 } 407 408 409 static int vmlogrdr_receive_data(struct vmlogrdr_priv_t *priv) 410 { 411 int rc, *temp; 412 /* we need to keep track of two data sizes here: 413 * The number of bytes we need to receive from iucv and 414 * the total number of bytes we actually write into the buffer. 415 */ 416 int user_data_count, iucv_data_count; 417 char * buffer; 418 419 if (atomic_read(&priv->receive_ready)) { 420 spin_lock_bh(&priv->priv_lock); 421 if (priv->residual_length){ 422 /* receive second half of a record */ 423 iucv_data_count = priv->residual_length; 424 user_data_count = 0; 425 buffer = priv->buffer; 426 } else { 427 /* receive a new record: 428 * We need to return the total length of the record 429 * + size of FENCE in the first 4 bytes of the buffer. 430 */ 431 iucv_data_count = priv->local_interrupt_buffer.length; 432 user_data_count = sizeof(int); 433 temp = (int*)priv->buffer; 434 *temp= iucv_data_count + sizeof(FENCE); 435 buffer = priv->buffer + sizeof(int); 436 } 437 /* 438 * If the record is bigger than our buffer, we receive only 439 * a part of it. We can get the rest later. 440 */ 441 if (iucv_data_count > NET_BUFFER_SIZE) 442 iucv_data_count = NET_BUFFER_SIZE; 443 rc = iucv_message_receive(priv->path, 444 &priv->local_interrupt_buffer, 445 0, buffer, iucv_data_count, 446 &priv->residual_length); 447 spin_unlock_bh(&priv->priv_lock); 448 /* An rc of 5 indicates that the record was bigger than 449 * the buffer, which is OK for us. A 9 indicates that the 450 * record was purged befor we could receive it. 451 */ 452 if (rc == 5) 453 rc = 0; 454 if (rc == 9) 455 atomic_set(&priv->receive_ready, 0); 456 } else { 457 rc = 1; 458 } 459 if (!rc) { 460 priv->buffer_free = 0; 461 user_data_count += iucv_data_count; 462 priv->current_position = priv->buffer; 463 if (priv->residual_length == 0){ 464 /* the whole record has been captured, 465 * now add the fence */ 466 atomic_dec(&priv->receive_ready); 467 buffer = priv->buffer + user_data_count; 468 memcpy(buffer, FENCE, sizeof(FENCE)); 469 user_data_count += sizeof(FENCE); 470 } 471 priv->remaining = user_data_count; 472 } 473 474 return rc; 475 } 476 477 478 static ssize_t vmlogrdr_read(struct file *filp, char __user *data, 479 size_t count, loff_t * ppos) 480 { 481 int rc; 482 struct vmlogrdr_priv_t * priv = filp->private_data; 483 484 while (priv->buffer_free) { 485 rc = vmlogrdr_receive_data(priv); 486 if (rc) { 487 rc = wait_event_interruptible(read_wait_queue, 488 atomic_read(&priv->receive_ready)); 489 if (rc) 490 return rc; 491 } 492 } 493 /* copy only up to end of record */ 494 if (count > priv->remaining) 495 count = priv->remaining; 496 497 if (copy_to_user(data, priv->current_position, count)) 498 return -EFAULT; 499 500 *ppos += count; 501 priv->current_position += count; 502 priv->remaining -= count; 503 504 /* if all data has been transferred, set buffer free */ 505 if (priv->remaining == 0) 506 priv->buffer_free = 1; 507 508 return count; 509 } 510 511 static ssize_t vmlogrdr_autopurge_store(struct device * dev, 512 struct device_attribute *attr, 513 const char * buf, size_t count) 514 { 515 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev); 516 ssize_t ret = count; 517 518 switch (buf[0]) { 519 case '0': 520 priv->autopurge=0; 521 break; 522 case '1': 523 priv->autopurge=1; 524 break; 525 default: 526 ret = -EINVAL; 527 } 528 return ret; 529 } 530 531 532 static ssize_t vmlogrdr_autopurge_show(struct device *dev, 533 struct device_attribute *attr, 534 char *buf) 535 { 536 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev); 537 return sprintf(buf, "%u\n", priv->autopurge); 538 } 539 540 541 static DEVICE_ATTR(autopurge, 0644, vmlogrdr_autopurge_show, 542 vmlogrdr_autopurge_store); 543 544 545 static ssize_t vmlogrdr_purge_store(struct device * dev, 546 struct device_attribute *attr, 547 const char * buf, size_t count) 548 { 549 550 char cp_command[80]; 551 char cp_response[80]; 552 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev); 553 554 if (buf[0] != '1') 555 return -EINVAL; 556 557 memset(cp_command, 0x00, sizeof(cp_command)); 558 memset(cp_response, 0x00, sizeof(cp_response)); 559 560 /* 561 * The recording command needs to be called with option QID 562 * for guests that have previlege classes A or B. 563 * Other guests will not recognize the command and we have to 564 * issue the same command without the QID parameter. 565 */ 566 567 if (recording_class_AB) 568 snprintf(cp_command, sizeof(cp_command), 569 "RECORDING %s PURGE QID * ", 570 priv->recording_name); 571 else 572 snprintf(cp_command, sizeof(cp_command), 573 "RECORDING %s PURGE ", 574 priv->recording_name); 575 576 cpcmd(cp_command, cp_response, sizeof(cp_response), NULL); 577 578 return count; 579 } 580 581 582 static DEVICE_ATTR(purge, 0200, NULL, vmlogrdr_purge_store); 583 584 585 static ssize_t vmlogrdr_autorecording_store(struct device *dev, 586 struct device_attribute *attr, 587 const char *buf, size_t count) 588 { 589 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev); 590 ssize_t ret = count; 591 592 switch (buf[0]) { 593 case '0': 594 priv->autorecording=0; 595 break; 596 case '1': 597 priv->autorecording=1; 598 break; 599 default: 600 ret = -EINVAL; 601 } 602 return ret; 603 } 604 605 606 static ssize_t vmlogrdr_autorecording_show(struct device *dev, 607 struct device_attribute *attr, 608 char *buf) 609 { 610 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev); 611 return sprintf(buf, "%u\n", priv->autorecording); 612 } 613 614 615 static DEVICE_ATTR(autorecording, 0644, vmlogrdr_autorecording_show, 616 vmlogrdr_autorecording_store); 617 618 619 static ssize_t vmlogrdr_recording_store(struct device * dev, 620 struct device_attribute *attr, 621 const char * buf, size_t count) 622 { 623 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev); 624 ssize_t ret; 625 626 switch (buf[0]) { 627 case '0': 628 ret = vmlogrdr_recording(priv,0,0); 629 break; 630 case '1': 631 ret = vmlogrdr_recording(priv,1,0); 632 break; 633 default: 634 ret = -EINVAL; 635 } 636 if (ret) 637 return ret; 638 else 639 return count; 640 641 } 642 643 644 static DEVICE_ATTR(recording, 0200, NULL, vmlogrdr_recording_store); 645 646 647 static ssize_t vmlogrdr_recording_status_show(struct device_driver *driver, 648 char *buf) 649 { 650 651 static const char cp_command[] = "QUERY RECORDING "; 652 int len; 653 654 cpcmd(cp_command, buf, 4096, NULL); 655 len = strlen(buf); 656 return len; 657 } 658 static DRIVER_ATTR(recording_status, 0444, vmlogrdr_recording_status_show, 659 NULL); 660 static struct attribute *vmlogrdr_drv_attrs[] = { 661 &driver_attr_recording_status.attr, 662 NULL, 663 }; 664 static struct attribute_group vmlogrdr_drv_attr_group = { 665 .attrs = vmlogrdr_drv_attrs, 666 }; 667 static const struct attribute_group *vmlogrdr_drv_attr_groups[] = { 668 &vmlogrdr_drv_attr_group, 669 NULL, 670 }; 671 672 static struct attribute *vmlogrdr_attrs[] = { 673 &dev_attr_autopurge.attr, 674 &dev_attr_purge.attr, 675 &dev_attr_autorecording.attr, 676 &dev_attr_recording.attr, 677 NULL, 678 }; 679 static struct attribute_group vmlogrdr_attr_group = { 680 .attrs = vmlogrdr_attrs, 681 }; 682 static const struct attribute_group *vmlogrdr_attr_groups[] = { 683 &vmlogrdr_attr_group, 684 NULL, 685 }; 686 687 static int vmlogrdr_pm_prepare(struct device *dev) 688 { 689 int rc; 690 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev); 691 692 rc = 0; 693 if (priv) { 694 spin_lock_bh(&priv->priv_lock); 695 if (priv->dev_in_use) 696 rc = -EBUSY; 697 spin_unlock_bh(&priv->priv_lock); 698 } 699 if (rc) 700 pr_err("vmlogrdr: device %s is busy. Refuse to suspend.\n", 701 dev_name(dev)); 702 return rc; 703 } 704 705 706 static const struct dev_pm_ops vmlogrdr_pm_ops = { 707 .prepare = vmlogrdr_pm_prepare, 708 }; 709 710 static struct class *vmlogrdr_class; 711 static struct device_driver vmlogrdr_driver = { 712 .name = "vmlogrdr", 713 .bus = &iucv_bus, 714 .pm = &vmlogrdr_pm_ops, 715 .groups = vmlogrdr_drv_attr_groups, 716 }; 717 718 static int vmlogrdr_register_driver(void) 719 { 720 int ret; 721 722 /* Register with iucv driver */ 723 ret = iucv_register(&vmlogrdr_iucv_handler, 1); 724 if (ret) 725 goto out; 726 727 ret = driver_register(&vmlogrdr_driver); 728 if (ret) 729 goto out_iucv; 730 731 vmlogrdr_class = class_create(THIS_MODULE, "vmlogrdr"); 732 if (IS_ERR(vmlogrdr_class)) { 733 ret = PTR_ERR(vmlogrdr_class); 734 vmlogrdr_class = NULL; 735 goto out_driver; 736 } 737 return 0; 738 739 out_driver: 740 driver_unregister(&vmlogrdr_driver); 741 out_iucv: 742 iucv_unregister(&vmlogrdr_iucv_handler, 1); 743 out: 744 return ret; 745 } 746 747 748 static void vmlogrdr_unregister_driver(void) 749 { 750 class_destroy(vmlogrdr_class); 751 vmlogrdr_class = NULL; 752 driver_unregister(&vmlogrdr_driver); 753 iucv_unregister(&vmlogrdr_iucv_handler, 1); 754 } 755 756 757 static int vmlogrdr_register_device(struct vmlogrdr_priv_t *priv) 758 { 759 struct device *dev; 760 int ret; 761 762 dev = kzalloc(sizeof(struct device), GFP_KERNEL); 763 if (dev) { 764 dev_set_name(dev, "%s", priv->internal_name); 765 dev->bus = &iucv_bus; 766 dev->parent = iucv_root; 767 dev->driver = &vmlogrdr_driver; 768 dev->groups = vmlogrdr_attr_groups; 769 dev_set_drvdata(dev, priv); 770 /* 771 * The release function could be called after the 772 * module has been unloaded. It's _only_ task is to 773 * free the struct. Therefore, we specify kfree() 774 * directly here. (Probably a little bit obfuscating 775 * but legitime ...). 776 */ 777 dev->release = (void (*)(struct device *))kfree; 778 } else 779 return -ENOMEM; 780 ret = device_register(dev); 781 if (ret) { 782 put_device(dev); 783 return ret; 784 } 785 786 priv->class_device = device_create(vmlogrdr_class, dev, 787 MKDEV(vmlogrdr_major, 788 priv->minor_num), 789 priv, "%s", dev_name(dev)); 790 if (IS_ERR(priv->class_device)) { 791 ret = PTR_ERR(priv->class_device); 792 priv->class_device=NULL; 793 device_unregister(dev); 794 return ret; 795 } 796 priv->device = dev; 797 return 0; 798 } 799 800 801 static int vmlogrdr_unregister_device(struct vmlogrdr_priv_t *priv) 802 { 803 device_destroy(vmlogrdr_class, MKDEV(vmlogrdr_major, priv->minor_num)); 804 if (priv->device != NULL) { 805 device_unregister(priv->device); 806 priv->device=NULL; 807 } 808 return 0; 809 } 810 811 812 static int vmlogrdr_register_cdev(dev_t dev) 813 { 814 int rc = 0; 815 vmlogrdr_cdev = cdev_alloc(); 816 if (!vmlogrdr_cdev) { 817 return -ENOMEM; 818 } 819 vmlogrdr_cdev->owner = THIS_MODULE; 820 vmlogrdr_cdev->ops = &vmlogrdr_fops; 821 vmlogrdr_cdev->dev = dev; 822 rc = cdev_add(vmlogrdr_cdev, vmlogrdr_cdev->dev, MAXMINOR); 823 if (!rc) 824 return 0; 825 826 // cleanup: cdev is not fully registered, no cdev_del here! 827 kobject_put(&vmlogrdr_cdev->kobj); 828 vmlogrdr_cdev=NULL; 829 return rc; 830 } 831 832 833 static void vmlogrdr_cleanup(void) 834 { 835 int i; 836 837 if (vmlogrdr_cdev) { 838 cdev_del(vmlogrdr_cdev); 839 vmlogrdr_cdev=NULL; 840 } 841 for (i=0; i < MAXMINOR; ++i ) { 842 vmlogrdr_unregister_device(&sys_ser[i]); 843 free_page((unsigned long)sys_ser[i].buffer); 844 } 845 vmlogrdr_unregister_driver(); 846 if (vmlogrdr_major) { 847 unregister_chrdev_region(MKDEV(vmlogrdr_major, 0), MAXMINOR); 848 vmlogrdr_major=0; 849 } 850 } 851 852 853 static int __init vmlogrdr_init(void) 854 { 855 int rc; 856 int i; 857 dev_t dev; 858 859 if (! MACHINE_IS_VM) { 860 pr_err("not running under VM, driver not loaded.\n"); 861 return -ENODEV; 862 } 863 864 recording_class_AB = vmlogrdr_get_recording_class_AB(); 865 866 rc = alloc_chrdev_region(&dev, 0, MAXMINOR, "vmlogrdr"); 867 if (rc) 868 return rc; 869 vmlogrdr_major = MAJOR(dev); 870 871 rc=vmlogrdr_register_driver(); 872 if (rc) 873 goto cleanup; 874 875 for (i=0; i < MAXMINOR; ++i ) { 876 sys_ser[i].buffer = (char *) get_zeroed_page(GFP_KERNEL); 877 if (!sys_ser[i].buffer) { 878 rc = -ENOMEM; 879 break; 880 } 881 sys_ser[i].current_position = sys_ser[i].buffer; 882 rc=vmlogrdr_register_device(&sys_ser[i]); 883 if (rc) 884 break; 885 } 886 if (rc) 887 goto cleanup; 888 889 rc = vmlogrdr_register_cdev(dev); 890 if (rc) 891 goto cleanup; 892 return 0; 893 894 cleanup: 895 vmlogrdr_cleanup(); 896 return rc; 897 } 898 899 900 static void __exit vmlogrdr_exit(void) 901 { 902 vmlogrdr_cleanup(); 903 return; 904 } 905 906 907 module_init(vmlogrdr_init); 908 module_exit(vmlogrdr_exit); 909