1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Character device driver for extended error reporting. 4 * 5 * Copyright IBM Corp. 2005 6 * extended error reporting for DASD ECKD devices 7 * Author(s): Stefan Weinhuber <wein@de.ibm.com> 8 */ 9 10 #define KMSG_COMPONENT "dasd-eckd" 11 12 #include <linux/init.h> 13 #include <linux/fs.h> 14 #include <linux/kernel.h> 15 #include <linux/miscdevice.h> 16 #include <linux/module.h> 17 #include <linux/moduleparam.h> 18 #include <linux/device.h> 19 #include <linux/poll.h> 20 #include <linux/mutex.h> 21 #include <linux/err.h> 22 #include <linux/slab.h> 23 24 #include <linux/uaccess.h> 25 #include <linux/atomic.h> 26 #include <asm/ebcdic.h> 27 28 #include "dasd_int.h" 29 #include "dasd_eckd.h" 30 31 #ifdef PRINTK_HEADER 32 #undef PRINTK_HEADER 33 #endif /* PRINTK_HEADER */ 34 #define PRINTK_HEADER "dasd(eer):" 35 36 /* 37 * SECTION: the internal buffer 38 */ 39 40 /* 41 * The internal buffer is meant to store obaque blobs of data, so it does 42 * not know of higher level concepts like triggers. 43 * It consists of a number of pages that are used as a ringbuffer. Each data 44 * blob is stored in a simple record that consists of an integer, which 45 * contains the size of the following data, and the data bytes themselfes. 46 * 47 * To allow for multiple independent readers we create one internal buffer 48 * each time the device is opened and destroy the buffer when the file is 49 * closed again. The number of pages used for this buffer is determined by 50 * the module parmeter eer_pages. 51 * 52 * One record can be written to a buffer by using the functions 53 * - dasd_eer_start_record (one time per record to write the size to the 54 * buffer and reserve the space for the data) 55 * - dasd_eer_write_buffer (one or more times per record to write the data) 56 * The data can be written in several steps but you will have to compute 57 * the total size up front for the invocation of dasd_eer_start_record. 58 * If the ringbuffer is full, dasd_eer_start_record will remove the required 59 * number of old records. 60 * 61 * A record is typically read in two steps, first read the integer that 62 * specifies the size of the following data, then read the data. 63 * Both can be done by 64 * - dasd_eer_read_buffer 65 * 66 * For all mentioned functions you need to get the bufferlock first and keep 67 * it until a complete record is written or read. 68 * 69 * All information necessary to keep track of an internal buffer is kept in 70 * a struct eerbuffer. The buffer specific to a file pointer is strored in 71 * the private_data field of that file. To be able to write data to all 72 * existing buffers, each buffer is also added to the bufferlist. 73 * If the user does not want to read a complete record in one go, we have to 74 * keep track of the rest of the record. residual stores the number of bytes 75 * that are still to deliver. If the rest of the record is invalidated between 76 * two reads then residual will be set to -1 so that the next read will fail. 77 * All entries in the eerbuffer structure are protected with the bufferlock. 78 * To avoid races between writing to a buffer on the one side and creating 79 * and destroying buffers on the other side, the bufferlock must also be used 80 * to protect the bufferlist. 81 */ 82 83 static int eer_pages = 5; 84 module_param(eer_pages, int, S_IRUGO|S_IWUSR); 85 86 struct eerbuffer { 87 struct list_head list; 88 char **buffer; 89 int buffersize; 90 int buffer_page_count; 91 int head; 92 int tail; 93 int residual; 94 }; 95 96 static LIST_HEAD(bufferlist); 97 static DEFINE_SPINLOCK(bufferlock); 98 static DECLARE_WAIT_QUEUE_HEAD(dasd_eer_read_wait_queue); 99 100 /* 101 * How many free bytes are available on the buffer. 102 * Needs to be called with bufferlock held. 103 */ 104 static int dasd_eer_get_free_bytes(struct eerbuffer *eerb) 105 { 106 if (eerb->head < eerb->tail) 107 return eerb->tail - eerb->head - 1; 108 return eerb->buffersize - eerb->head + eerb->tail -1; 109 } 110 111 /* 112 * How many bytes of buffer space are used. 113 * Needs to be called with bufferlock held. 114 */ 115 static int dasd_eer_get_filled_bytes(struct eerbuffer *eerb) 116 { 117 118 if (eerb->head >= eerb->tail) 119 return eerb->head - eerb->tail; 120 return eerb->buffersize - eerb->tail + eerb->head; 121 } 122 123 /* 124 * The dasd_eer_write_buffer function just copies count bytes of data 125 * to the buffer. Make sure to call dasd_eer_start_record first, to 126 * make sure that enough free space is available. 127 * Needs to be called with bufferlock held. 128 */ 129 static void dasd_eer_write_buffer(struct eerbuffer *eerb, 130 char *data, int count) 131 { 132 133 unsigned long headindex,localhead; 134 unsigned long rest, len; 135 char *nextdata; 136 137 nextdata = data; 138 rest = count; 139 while (rest > 0) { 140 headindex = eerb->head / PAGE_SIZE; 141 localhead = eerb->head % PAGE_SIZE; 142 len = min(rest, PAGE_SIZE - localhead); 143 memcpy(eerb->buffer[headindex]+localhead, nextdata, len); 144 nextdata += len; 145 rest -= len; 146 eerb->head += len; 147 if (eerb->head == eerb->buffersize) 148 eerb->head = 0; /* wrap around */ 149 BUG_ON(eerb->head > eerb->buffersize); 150 } 151 } 152 153 /* 154 * Needs to be called with bufferlock held. 155 */ 156 static int dasd_eer_read_buffer(struct eerbuffer *eerb, char *data, int count) 157 { 158 159 unsigned long tailindex,localtail; 160 unsigned long rest, len, finalcount; 161 char *nextdata; 162 163 finalcount = min(count, dasd_eer_get_filled_bytes(eerb)); 164 nextdata = data; 165 rest = finalcount; 166 while (rest > 0) { 167 tailindex = eerb->tail / PAGE_SIZE; 168 localtail = eerb->tail % PAGE_SIZE; 169 len = min(rest, PAGE_SIZE - localtail); 170 memcpy(nextdata, eerb->buffer[tailindex] + localtail, len); 171 nextdata += len; 172 rest -= len; 173 eerb->tail += len; 174 if (eerb->tail == eerb->buffersize) 175 eerb->tail = 0; /* wrap around */ 176 BUG_ON(eerb->tail > eerb->buffersize); 177 } 178 return finalcount; 179 } 180 181 /* 182 * Whenever you want to write a blob of data to the internal buffer you 183 * have to start by using this function first. It will write the number 184 * of bytes that will be written to the buffer. If necessary it will remove 185 * old records to make room for the new one. 186 * Needs to be called with bufferlock held. 187 */ 188 static int dasd_eer_start_record(struct eerbuffer *eerb, int count) 189 { 190 int tailcount; 191 192 if (count + sizeof(count) > eerb->buffersize) 193 return -ENOMEM; 194 while (dasd_eer_get_free_bytes(eerb) < count + sizeof(count)) { 195 if (eerb->residual > 0) { 196 eerb->tail += eerb->residual; 197 if (eerb->tail >= eerb->buffersize) 198 eerb->tail -= eerb->buffersize; 199 eerb->residual = -1; 200 } 201 dasd_eer_read_buffer(eerb, (char *) &tailcount, 202 sizeof(tailcount)); 203 eerb->tail += tailcount; 204 if (eerb->tail >= eerb->buffersize) 205 eerb->tail -= eerb->buffersize; 206 } 207 dasd_eer_write_buffer(eerb, (char*) &count, sizeof(count)); 208 209 return 0; 210 }; 211 212 /* 213 * Release pages that are not used anymore. 214 */ 215 static void dasd_eer_free_buffer_pages(char **buf, int no_pages) 216 { 217 int i; 218 219 for (i = 0; i < no_pages; i++) 220 free_page((unsigned long) buf[i]); 221 } 222 223 /* 224 * Allocate a new set of memory pages. 225 */ 226 static int dasd_eer_allocate_buffer_pages(char **buf, int no_pages) 227 { 228 int i; 229 230 for (i = 0; i < no_pages; i++) { 231 buf[i] = (char *) get_zeroed_page(GFP_KERNEL); 232 if (!buf[i]) { 233 dasd_eer_free_buffer_pages(buf, i); 234 return -ENOMEM; 235 } 236 } 237 return 0; 238 } 239 240 /* 241 * SECTION: The extended error reporting functionality 242 */ 243 244 /* 245 * When a DASD device driver wants to report an error, it calls the 246 * function dasd_eer_write and gives the respective trigger ID as 247 * parameter. Currently there are four kinds of triggers: 248 * 249 * DASD_EER_FATALERROR: all kinds of unrecoverable I/O problems 250 * DASD_EER_PPRCSUSPEND: PPRC was suspended 251 * DASD_EER_NOPATH: There is no path to the device left. 252 * DASD_EER_STATECHANGE: The state of the device has changed. 253 * 254 * For the first three triggers all required information can be supplied by 255 * the caller. For these triggers a record is written by the function 256 * dasd_eer_write_standard_trigger. 257 * 258 * The DASD_EER_STATECHANGE trigger is special since a sense subsystem 259 * status ccw need to be executed to gather the necessary sense data first. 260 * The dasd_eer_snss function will queue the SNSS request and the request 261 * callback will then call dasd_eer_write with the DASD_EER_STATCHANGE 262 * trigger. 263 * 264 * To avoid memory allocations at runtime, the necessary memory is allocated 265 * when the extended error reporting is enabled for a device (by 266 * dasd_eer_probe). There is one sense subsystem status request for each 267 * eer enabled DASD device. The presence of the cqr in device->eer_cqr 268 * indicates that eer is enable for the device. The use of the snss request 269 * is protected by the DASD_FLAG_EER_IN_USE bit. When this flag indicates 270 * that the cqr is currently in use, dasd_eer_snss cannot start a second 271 * request but sets the DASD_FLAG_EER_SNSS flag instead. The callback of 272 * the SNSS request will check the bit and call dasd_eer_snss again. 273 */ 274 275 #define SNSS_DATA_SIZE 44 276 277 #define DASD_EER_BUSID_SIZE 10 278 struct dasd_eer_header { 279 __u32 total_size; 280 __u32 trigger; 281 __u64 tv_sec; 282 __u64 tv_usec; 283 char busid[DASD_EER_BUSID_SIZE]; 284 } __attribute__ ((packed)); 285 286 /* 287 * The following function can be used for those triggers that have 288 * all necessary data available when the function is called. 289 * If the parameter cqr is not NULL, the chain of requests will be searched 290 * for valid sense data, and all valid sense data sets will be added to 291 * the triggers data. 292 */ 293 static void dasd_eer_write_standard_trigger(struct dasd_device *device, 294 struct dasd_ccw_req *cqr, 295 int trigger) 296 { 297 struct dasd_ccw_req *temp_cqr; 298 int data_size; 299 struct timespec64 ts; 300 struct dasd_eer_header header; 301 unsigned long flags; 302 struct eerbuffer *eerb; 303 char *sense; 304 305 /* go through cqr chain and count the valid sense data sets */ 306 data_size = 0; 307 for (temp_cqr = cqr; temp_cqr; temp_cqr = temp_cqr->refers) 308 if (dasd_get_sense(&temp_cqr->irb)) 309 data_size += 32; 310 311 header.total_size = sizeof(header) + data_size + 4; /* "EOR" */ 312 header.trigger = trigger; 313 ktime_get_real_ts64(&ts); 314 header.tv_sec = ts.tv_sec; 315 header.tv_usec = ts.tv_nsec / NSEC_PER_USEC; 316 strscpy(header.busid, dev_name(&device->cdev->dev), 317 DASD_EER_BUSID_SIZE); 318 319 spin_lock_irqsave(&bufferlock, flags); 320 list_for_each_entry(eerb, &bufferlist, list) { 321 dasd_eer_start_record(eerb, header.total_size); 322 dasd_eer_write_buffer(eerb, (char *) &header, sizeof(header)); 323 for (temp_cqr = cqr; temp_cqr; temp_cqr = temp_cqr->refers) { 324 sense = dasd_get_sense(&temp_cqr->irb); 325 if (sense) 326 dasd_eer_write_buffer(eerb, sense, 32); 327 } 328 dasd_eer_write_buffer(eerb, "EOR", 4); 329 } 330 spin_unlock_irqrestore(&bufferlock, flags); 331 wake_up_interruptible(&dasd_eer_read_wait_queue); 332 } 333 334 /* 335 * This function writes a DASD_EER_STATECHANGE trigger. 336 */ 337 static void dasd_eer_write_snss_trigger(struct dasd_device *device, 338 struct dasd_ccw_req *cqr, 339 int trigger) 340 { 341 int data_size; 342 int snss_rc; 343 struct timespec64 ts; 344 struct dasd_eer_header header; 345 unsigned long flags; 346 struct eerbuffer *eerb; 347 348 snss_rc = (cqr->status == DASD_CQR_DONE) ? 0 : -EIO; 349 if (snss_rc) 350 data_size = 0; 351 else 352 data_size = SNSS_DATA_SIZE; 353 354 header.total_size = sizeof(header) + data_size + 4; /* "EOR" */ 355 header.trigger = DASD_EER_STATECHANGE; 356 ktime_get_real_ts64(&ts); 357 header.tv_sec = ts.tv_sec; 358 header.tv_usec = ts.tv_nsec / NSEC_PER_USEC; 359 strscpy(header.busid, dev_name(&device->cdev->dev), 360 DASD_EER_BUSID_SIZE); 361 362 spin_lock_irqsave(&bufferlock, flags); 363 list_for_each_entry(eerb, &bufferlist, list) { 364 dasd_eer_start_record(eerb, header.total_size); 365 dasd_eer_write_buffer(eerb, (char *) &header , sizeof(header)); 366 if (!snss_rc) 367 dasd_eer_write_buffer(eerb, cqr->data, SNSS_DATA_SIZE); 368 dasd_eer_write_buffer(eerb, "EOR", 4); 369 } 370 spin_unlock_irqrestore(&bufferlock, flags); 371 wake_up_interruptible(&dasd_eer_read_wait_queue); 372 } 373 374 /* 375 * This function is called for all triggers. It calls the appropriate 376 * function that writes the actual trigger records. 377 */ 378 void dasd_eer_write(struct dasd_device *device, struct dasd_ccw_req *cqr, 379 unsigned int id) 380 { 381 if (!device->eer_cqr) 382 return; 383 switch (id) { 384 case DASD_EER_FATALERROR: 385 case DASD_EER_PPRCSUSPEND: 386 dasd_eer_write_standard_trigger(device, cqr, id); 387 break; 388 case DASD_EER_NOPATH: 389 case DASD_EER_NOSPC: 390 dasd_eer_write_standard_trigger(device, NULL, id); 391 break; 392 case DASD_EER_STATECHANGE: 393 dasd_eer_write_snss_trigger(device, cqr, id); 394 break; 395 default: /* unknown trigger, so we write it without any sense data */ 396 dasd_eer_write_standard_trigger(device, NULL, id); 397 break; 398 } 399 } 400 EXPORT_SYMBOL(dasd_eer_write); 401 402 /* 403 * Start a sense subsystem status request. 404 * Needs to be called with the device held. 405 */ 406 void dasd_eer_snss(struct dasd_device *device) 407 { 408 struct dasd_ccw_req *cqr; 409 410 cqr = device->eer_cqr; 411 if (!cqr) /* Device not eer enabled. */ 412 return; 413 if (test_and_set_bit(DASD_FLAG_EER_IN_USE, &device->flags)) { 414 /* Sense subsystem status request in use. */ 415 set_bit(DASD_FLAG_EER_SNSS, &device->flags); 416 return; 417 } 418 /* cdev is already locked, can't use dasd_add_request_head */ 419 clear_bit(DASD_FLAG_EER_SNSS, &device->flags); 420 cqr->status = DASD_CQR_QUEUED; 421 list_add(&cqr->devlist, &device->ccw_queue); 422 dasd_schedule_device_bh(device); 423 } 424 425 /* 426 * Callback function for use with sense subsystem status request. 427 */ 428 static void dasd_eer_snss_cb(struct dasd_ccw_req *cqr, void *data) 429 { 430 struct dasd_device *device = cqr->startdev; 431 unsigned long flags; 432 433 dasd_eer_write(device, cqr, DASD_EER_STATECHANGE); 434 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags); 435 if (device->eer_cqr == cqr) { 436 clear_bit(DASD_FLAG_EER_IN_USE, &device->flags); 437 if (test_bit(DASD_FLAG_EER_SNSS, &device->flags)) 438 /* Another SNSS has been requested in the meantime. */ 439 dasd_eer_snss(device); 440 cqr = NULL; 441 } 442 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags); 443 if (cqr) 444 /* 445 * Extended error recovery has been switched off while 446 * the SNSS request was running. It could even have 447 * been switched off and on again in which case there 448 * is a new ccw in device->eer_cqr. Free the "old" 449 * snss request now. 450 */ 451 dasd_sfree_request(cqr, device); 452 } 453 454 /* 455 * Enable error reporting on a given device. 456 */ 457 int dasd_eer_enable(struct dasd_device *device) 458 { 459 struct dasd_ccw_req *cqr = NULL; 460 unsigned long flags; 461 struct ccw1 *ccw; 462 int rc = 0; 463 464 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags); 465 if (device->eer_cqr) 466 goto out; 467 else if (!device->discipline || 468 strcmp(device->discipline->name, "ECKD")) 469 rc = -EMEDIUMTYPE; 470 else if (test_bit(DASD_FLAG_OFFLINE, &device->flags)) 471 rc = -EBUSY; 472 473 if (rc) 474 goto out; 475 476 cqr = dasd_smalloc_request(DASD_ECKD_MAGIC, 1 /* SNSS */, 477 SNSS_DATA_SIZE, device, NULL); 478 if (IS_ERR(cqr)) { 479 rc = -ENOMEM; 480 cqr = NULL; 481 goto out; 482 } 483 484 cqr->startdev = device; 485 cqr->retries = 255; 486 cqr->expires = 10 * HZ; 487 clear_bit(DASD_CQR_FLAGS_USE_ERP, &cqr->flags); 488 set_bit(DASD_CQR_ALLOW_SLOCK, &cqr->flags); 489 490 ccw = cqr->cpaddr; 491 ccw->cmd_code = DASD_ECKD_CCW_SNSS; 492 ccw->count = SNSS_DATA_SIZE; 493 ccw->flags = 0; 494 ccw->cda = (__u32)(addr_t) cqr->data; 495 496 cqr->buildclk = get_tod_clock(); 497 cqr->status = DASD_CQR_FILLED; 498 cqr->callback = dasd_eer_snss_cb; 499 500 if (!device->eer_cqr) { 501 device->eer_cqr = cqr; 502 cqr = NULL; 503 } 504 505 out: 506 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags); 507 508 if (cqr) 509 dasd_sfree_request(cqr, device); 510 511 return rc; 512 } 513 514 /* 515 * Disable error reporting on a given device. 516 */ 517 void dasd_eer_disable(struct dasd_device *device) 518 { 519 struct dasd_ccw_req *cqr; 520 unsigned long flags; 521 int in_use; 522 523 if (!device->eer_cqr) 524 return; 525 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags); 526 cqr = device->eer_cqr; 527 device->eer_cqr = NULL; 528 clear_bit(DASD_FLAG_EER_SNSS, &device->flags); 529 in_use = test_and_clear_bit(DASD_FLAG_EER_IN_USE, &device->flags); 530 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags); 531 if (cqr && !in_use) 532 dasd_sfree_request(cqr, device); 533 } 534 535 /* 536 * SECTION: the device operations 537 */ 538 539 /* 540 * On the one side we need a lock to access our internal buffer, on the 541 * other side a copy_to_user can sleep. So we need to copy the data we have 542 * to transfer in a readbuffer, which is protected by the readbuffer_mutex. 543 */ 544 static char readbuffer[PAGE_SIZE]; 545 static DEFINE_MUTEX(readbuffer_mutex); 546 547 static int dasd_eer_open(struct inode *inp, struct file *filp) 548 { 549 struct eerbuffer *eerb; 550 unsigned long flags; 551 552 eerb = kzalloc(sizeof(struct eerbuffer), GFP_KERNEL); 553 if (!eerb) 554 return -ENOMEM; 555 eerb->buffer_page_count = eer_pages; 556 if (eerb->buffer_page_count < 1 || 557 eerb->buffer_page_count > INT_MAX / PAGE_SIZE) { 558 kfree(eerb); 559 DBF_EVENT(DBF_WARNING, "can't open device since module " 560 "parameter eer_pages is smaller than 1 or" 561 " bigger than %d", (int)(INT_MAX / PAGE_SIZE)); 562 return -EINVAL; 563 } 564 eerb->buffersize = eerb->buffer_page_count * PAGE_SIZE; 565 eerb->buffer = kmalloc_array(eerb->buffer_page_count, sizeof(char *), 566 GFP_KERNEL); 567 if (!eerb->buffer) { 568 kfree(eerb); 569 return -ENOMEM; 570 } 571 if (dasd_eer_allocate_buffer_pages(eerb->buffer, 572 eerb->buffer_page_count)) { 573 kfree(eerb->buffer); 574 kfree(eerb); 575 return -ENOMEM; 576 } 577 filp->private_data = eerb; 578 spin_lock_irqsave(&bufferlock, flags); 579 list_add(&eerb->list, &bufferlist); 580 spin_unlock_irqrestore(&bufferlock, flags); 581 582 return nonseekable_open(inp,filp); 583 } 584 585 static int dasd_eer_close(struct inode *inp, struct file *filp) 586 { 587 struct eerbuffer *eerb; 588 unsigned long flags; 589 590 eerb = (struct eerbuffer *) filp->private_data; 591 spin_lock_irqsave(&bufferlock, flags); 592 list_del(&eerb->list); 593 spin_unlock_irqrestore(&bufferlock, flags); 594 dasd_eer_free_buffer_pages(eerb->buffer, eerb->buffer_page_count); 595 kfree(eerb->buffer); 596 kfree(eerb); 597 598 return 0; 599 } 600 601 static ssize_t dasd_eer_read(struct file *filp, char __user *buf, 602 size_t count, loff_t *ppos) 603 { 604 int tc,rc; 605 int tailcount,effective_count; 606 unsigned long flags; 607 struct eerbuffer *eerb; 608 609 eerb = (struct eerbuffer *) filp->private_data; 610 if (mutex_lock_interruptible(&readbuffer_mutex)) 611 return -ERESTARTSYS; 612 613 spin_lock_irqsave(&bufferlock, flags); 614 615 if (eerb->residual < 0) { /* the remainder of this record */ 616 /* has been deleted */ 617 eerb->residual = 0; 618 spin_unlock_irqrestore(&bufferlock, flags); 619 mutex_unlock(&readbuffer_mutex); 620 return -EIO; 621 } else if (eerb->residual > 0) { 622 /* OK we still have a second half of a record to deliver */ 623 effective_count = min(eerb->residual, (int) count); 624 eerb->residual -= effective_count; 625 } else { 626 tc = 0; 627 while (!tc) { 628 tc = dasd_eer_read_buffer(eerb, (char *) &tailcount, 629 sizeof(tailcount)); 630 if (!tc) { 631 /* no data available */ 632 spin_unlock_irqrestore(&bufferlock, flags); 633 mutex_unlock(&readbuffer_mutex); 634 if (filp->f_flags & O_NONBLOCK) 635 return -EAGAIN; 636 rc = wait_event_interruptible( 637 dasd_eer_read_wait_queue, 638 eerb->head != eerb->tail); 639 if (rc) 640 return rc; 641 if (mutex_lock_interruptible(&readbuffer_mutex)) 642 return -ERESTARTSYS; 643 spin_lock_irqsave(&bufferlock, flags); 644 } 645 } 646 WARN_ON(tc != sizeof(tailcount)); 647 effective_count = min(tailcount,(int)count); 648 eerb->residual = tailcount - effective_count; 649 } 650 651 tc = dasd_eer_read_buffer(eerb, readbuffer, effective_count); 652 WARN_ON(tc != effective_count); 653 654 spin_unlock_irqrestore(&bufferlock, flags); 655 656 if (copy_to_user(buf, readbuffer, effective_count)) { 657 mutex_unlock(&readbuffer_mutex); 658 return -EFAULT; 659 } 660 661 mutex_unlock(&readbuffer_mutex); 662 return effective_count; 663 } 664 665 static __poll_t dasd_eer_poll(struct file *filp, poll_table *ptable) 666 { 667 __poll_t mask; 668 unsigned long flags; 669 struct eerbuffer *eerb; 670 671 eerb = (struct eerbuffer *) filp->private_data; 672 poll_wait(filp, &dasd_eer_read_wait_queue, ptable); 673 spin_lock_irqsave(&bufferlock, flags); 674 if (eerb->head != eerb->tail) 675 mask = EPOLLIN | EPOLLRDNORM ; 676 else 677 mask = 0; 678 spin_unlock_irqrestore(&bufferlock, flags); 679 return mask; 680 } 681 682 static const struct file_operations dasd_eer_fops = { 683 .open = &dasd_eer_open, 684 .release = &dasd_eer_close, 685 .read = &dasd_eer_read, 686 .poll = &dasd_eer_poll, 687 .owner = THIS_MODULE, 688 .llseek = noop_llseek, 689 }; 690 691 static struct miscdevice *dasd_eer_dev = NULL; 692 693 int __init dasd_eer_init(void) 694 { 695 int rc; 696 697 dasd_eer_dev = kzalloc(sizeof(*dasd_eer_dev), GFP_KERNEL); 698 if (!dasd_eer_dev) 699 return -ENOMEM; 700 701 dasd_eer_dev->minor = MISC_DYNAMIC_MINOR; 702 dasd_eer_dev->name = "dasd_eer"; 703 dasd_eer_dev->fops = &dasd_eer_fops; 704 705 rc = misc_register(dasd_eer_dev); 706 if (rc) { 707 kfree(dasd_eer_dev); 708 dasd_eer_dev = NULL; 709 DBF_EVENT(DBF_ERR, "%s", "dasd_eer_init could not " 710 "register misc device"); 711 return rc; 712 } 713 714 return 0; 715 } 716 717 void dasd_eer_exit(void) 718 { 719 if (dasd_eer_dev) { 720 misc_deregister(dasd_eer_dev); 721 kfree(dasd_eer_dev); 722 dasd_eer_dev = NULL; 723 } 724 } 725