1 /* 2 * watchdog_dev.c 3 * 4 * (c) Copyright 2008-2011 Alan Cox <alan@lxorguk.ukuu.org.uk>, 5 * All Rights Reserved. 6 * 7 * (c) Copyright 2008-2011 Wim Van Sebroeck <wim@iguana.be>. 8 * 9 * 10 * This source code is part of the generic code that can be used 11 * by all the watchdog timer drivers. 12 * 13 * This part of the generic code takes care of the following 14 * misc device: /dev/watchdog. 15 * 16 * Based on source code of the following authors: 17 * Matt Domsch <Matt_Domsch@dell.com>, 18 * Rob Radez <rob@osinvestor.com>, 19 * Rusty Lynch <rusty@linux.co.intel.com> 20 * Satyam Sharma <satyam@infradead.org> 21 * Randy Dunlap <randy.dunlap@oracle.com> 22 * 23 * This program is free software; you can redistribute it and/or 24 * modify it under the terms of the GNU General Public License 25 * as published by the Free Software Foundation; either version 26 * 2 of the License, or (at your option) any later version. 27 * 28 * Neither Alan Cox, CymruNet Ltd., Wim Van Sebroeck nor Iguana vzw. 29 * admit liability nor provide warranty for any of this software. 30 * This material is provided "AS-IS" and at no charge. 31 */ 32 33 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 34 35 #include <linux/cdev.h> /* For character device */ 36 #include <linux/errno.h> /* For the -ENODEV/... values */ 37 #include <linux/fs.h> /* For file operations */ 38 #include <linux/init.h> /* For __init/__exit/... */ 39 #include <linux/jiffies.h> /* For timeout functions */ 40 #include <linux/kernel.h> /* For printk/panic/... */ 41 #include <linux/kref.h> /* For data references */ 42 #include <linux/miscdevice.h> /* For handling misc devices */ 43 #include <linux/module.h> /* For module stuff/... */ 44 #include <linux/mutex.h> /* For mutexes */ 45 #include <linux/slab.h> /* For memory functions */ 46 #include <linux/types.h> /* For standard types (like size_t) */ 47 #include <linux/watchdog.h> /* For watchdog specific items */ 48 #include <linux/workqueue.h> /* For workqueue */ 49 #include <linux/uaccess.h> /* For copy_to_user/put_user/... */ 50 51 #include "watchdog_core.h" 52 53 /* 54 * struct watchdog_core_data - watchdog core internal data 55 * @kref: Reference count. 56 * @cdev: The watchdog's Character device. 57 * @wdd: Pointer to watchdog device. 58 * @lock: Lock for watchdog core. 59 * @status: Watchdog core internal status bits. 60 */ 61 struct watchdog_core_data { 62 struct kref kref; 63 struct cdev cdev; 64 struct watchdog_device *wdd; 65 struct mutex lock; 66 unsigned long last_keepalive; 67 unsigned long last_hw_keepalive; 68 struct delayed_work work; 69 unsigned long status; /* Internal status bits */ 70 #define _WDOG_DEV_OPEN 0 /* Opened ? */ 71 #define _WDOG_ALLOW_RELEASE 1 /* Did we receive the magic char ? */ 72 #define _WDOG_KEEPALIVE 2 /* Did we receive a keepalive ? */ 73 }; 74 75 /* the dev_t structure to store the dynamically allocated watchdog devices */ 76 static dev_t watchdog_devt; 77 /* Reference to watchdog device behind /dev/watchdog */ 78 static struct watchdog_core_data *old_wd_data; 79 80 static struct workqueue_struct *watchdog_wq; 81 82 static inline bool watchdog_need_worker(struct watchdog_device *wdd) 83 { 84 /* All variables in milli-seconds */ 85 unsigned int hm = wdd->max_hw_heartbeat_ms; 86 unsigned int t = wdd->timeout * 1000; 87 88 /* 89 * A worker to generate heartbeat requests is needed if all of the 90 * following conditions are true. 91 * - Userspace activated the watchdog. 92 * - The driver provided a value for the maximum hardware timeout, and 93 * thus is aware that the framework supports generating heartbeat 94 * requests. 95 * - Userspace requests a longer timeout than the hardware can handle. 96 * 97 * Alternatively, if userspace has not opened the watchdog 98 * device, we take care of feeding the watchdog if it is 99 * running. 100 */ 101 return (hm && watchdog_active(wdd) && t > hm) || 102 (t && !watchdog_active(wdd) && watchdog_hw_running(wdd)); 103 } 104 105 static long watchdog_next_keepalive(struct watchdog_device *wdd) 106 { 107 struct watchdog_core_data *wd_data = wdd->wd_data; 108 unsigned int timeout_ms = wdd->timeout * 1000; 109 unsigned long keepalive_interval; 110 unsigned long last_heartbeat; 111 unsigned long virt_timeout; 112 unsigned int hw_heartbeat_ms; 113 114 virt_timeout = wd_data->last_keepalive + msecs_to_jiffies(timeout_ms); 115 hw_heartbeat_ms = min_not_zero(timeout_ms, wdd->max_hw_heartbeat_ms); 116 keepalive_interval = msecs_to_jiffies(hw_heartbeat_ms / 2); 117 118 if (!watchdog_active(wdd)) 119 return keepalive_interval; 120 121 /* 122 * To ensure that the watchdog times out wdd->timeout seconds 123 * after the most recent ping from userspace, the last 124 * worker ping has to come in hw_heartbeat_ms before this timeout. 125 */ 126 last_heartbeat = virt_timeout - msecs_to_jiffies(hw_heartbeat_ms); 127 return min_t(long, last_heartbeat - jiffies, keepalive_interval); 128 } 129 130 static inline void watchdog_update_worker(struct watchdog_device *wdd) 131 { 132 struct watchdog_core_data *wd_data = wdd->wd_data; 133 134 if (watchdog_need_worker(wdd)) { 135 long t = watchdog_next_keepalive(wdd); 136 137 if (t > 0) 138 mod_delayed_work(watchdog_wq, &wd_data->work, t); 139 } else { 140 cancel_delayed_work(&wd_data->work); 141 } 142 } 143 144 static int __watchdog_ping(struct watchdog_device *wdd) 145 { 146 struct watchdog_core_data *wd_data = wdd->wd_data; 147 unsigned long earliest_keepalive = wd_data->last_hw_keepalive + 148 msecs_to_jiffies(wdd->min_hw_heartbeat_ms); 149 int err; 150 151 if (time_is_after_jiffies(earliest_keepalive)) { 152 mod_delayed_work(watchdog_wq, &wd_data->work, 153 earliest_keepalive - jiffies); 154 return 0; 155 } 156 157 wd_data->last_hw_keepalive = jiffies; 158 159 if (wdd->ops->ping) 160 err = wdd->ops->ping(wdd); /* ping the watchdog */ 161 else 162 err = wdd->ops->start(wdd); /* restart watchdog */ 163 164 watchdog_update_worker(wdd); 165 166 return err; 167 } 168 169 /* 170 * watchdog_ping: ping the watchdog. 171 * @wdd: the watchdog device to ping 172 * 173 * The caller must hold wd_data->lock. 174 * 175 * If the watchdog has no own ping operation then it needs to be 176 * restarted via the start operation. This wrapper function does 177 * exactly that. 178 * We only ping when the watchdog device is running. 179 */ 180 181 static int watchdog_ping(struct watchdog_device *wdd) 182 { 183 struct watchdog_core_data *wd_data = wdd->wd_data; 184 185 if (!watchdog_active(wdd) && !watchdog_hw_running(wdd)) 186 return 0; 187 188 set_bit(_WDOG_KEEPALIVE, &wd_data->status); 189 190 wd_data->last_keepalive = jiffies; 191 return __watchdog_ping(wdd); 192 } 193 194 static void watchdog_ping_work(struct work_struct *work) 195 { 196 struct watchdog_core_data *wd_data; 197 struct watchdog_device *wdd; 198 199 wd_data = container_of(to_delayed_work(work), struct watchdog_core_data, 200 work); 201 202 mutex_lock(&wd_data->lock); 203 wdd = wd_data->wdd; 204 if (wdd && (watchdog_active(wdd) || watchdog_hw_running(wdd))) 205 __watchdog_ping(wdd); 206 mutex_unlock(&wd_data->lock); 207 } 208 209 /* 210 * watchdog_start: wrapper to start the watchdog. 211 * @wdd: the watchdog device to start 212 * 213 * The caller must hold wd_data->lock. 214 * 215 * Start the watchdog if it is not active and mark it active. 216 * This function returns zero on success or a negative errno code for 217 * failure. 218 */ 219 220 static int watchdog_start(struct watchdog_device *wdd) 221 { 222 struct watchdog_core_data *wd_data = wdd->wd_data; 223 unsigned long started_at; 224 int err; 225 226 if (watchdog_active(wdd)) 227 return 0; 228 229 set_bit(_WDOG_KEEPALIVE, &wd_data->status); 230 231 started_at = jiffies; 232 if (watchdog_hw_running(wdd) && wdd->ops->ping) 233 err = wdd->ops->ping(wdd); 234 else 235 err = wdd->ops->start(wdd); 236 if (err == 0) { 237 set_bit(WDOG_ACTIVE, &wdd->status); 238 wd_data->last_keepalive = started_at; 239 watchdog_update_worker(wdd); 240 } 241 242 return err; 243 } 244 245 /* 246 * watchdog_stop: wrapper to stop the watchdog. 247 * @wdd: the watchdog device to stop 248 * 249 * The caller must hold wd_data->lock. 250 * 251 * Stop the watchdog if it is still active and unmark it active. 252 * This function returns zero on success or a negative errno code for 253 * failure. 254 * If the 'nowayout' feature was set, the watchdog cannot be stopped. 255 */ 256 257 static int watchdog_stop(struct watchdog_device *wdd) 258 { 259 int err = 0; 260 261 if (!watchdog_active(wdd)) 262 return 0; 263 264 if (test_bit(WDOG_NO_WAY_OUT, &wdd->status)) { 265 pr_info("watchdog%d: nowayout prevents watchdog being stopped!\n", 266 wdd->id); 267 return -EBUSY; 268 } 269 270 if (wdd->ops->stop) { 271 clear_bit(WDOG_HW_RUNNING, &wdd->status); 272 err = wdd->ops->stop(wdd); 273 } else { 274 set_bit(WDOG_HW_RUNNING, &wdd->status); 275 } 276 277 if (err == 0) { 278 clear_bit(WDOG_ACTIVE, &wdd->status); 279 watchdog_update_worker(wdd); 280 } 281 282 return err; 283 } 284 285 /* 286 * watchdog_get_status: wrapper to get the watchdog status 287 * @wdd: the watchdog device to get the status from 288 * 289 * The caller must hold wd_data->lock. 290 * 291 * Get the watchdog's status flags. 292 */ 293 294 static unsigned int watchdog_get_status(struct watchdog_device *wdd) 295 { 296 struct watchdog_core_data *wd_data = wdd->wd_data; 297 unsigned int status; 298 299 if (wdd->ops->status) 300 status = wdd->ops->status(wdd); 301 else 302 status = wdd->bootstatus & (WDIOF_CARDRESET | 303 WDIOF_OVERHEAT | 304 WDIOF_FANFAULT | 305 WDIOF_EXTERN1 | 306 WDIOF_EXTERN2 | 307 WDIOF_POWERUNDER | 308 WDIOF_POWEROVER); 309 310 if (test_bit(_WDOG_ALLOW_RELEASE, &wd_data->status)) 311 status |= WDIOF_MAGICCLOSE; 312 313 if (test_and_clear_bit(_WDOG_KEEPALIVE, &wd_data->status)) 314 status |= WDIOF_KEEPALIVEPING; 315 316 return status; 317 } 318 319 /* 320 * watchdog_set_timeout: set the watchdog timer timeout 321 * @wdd: the watchdog device to set the timeout for 322 * @timeout: timeout to set in seconds 323 * 324 * The caller must hold wd_data->lock. 325 */ 326 327 static int watchdog_set_timeout(struct watchdog_device *wdd, 328 unsigned int timeout) 329 { 330 int err = 0; 331 332 if (!(wdd->info->options & WDIOF_SETTIMEOUT)) 333 return -EOPNOTSUPP; 334 335 if (watchdog_timeout_invalid(wdd, timeout)) 336 return -EINVAL; 337 338 if (wdd->ops->set_timeout) 339 err = wdd->ops->set_timeout(wdd, timeout); 340 else 341 wdd->timeout = timeout; 342 343 watchdog_update_worker(wdd); 344 345 return err; 346 } 347 348 /* 349 * watchdog_get_timeleft: wrapper to get the time left before a reboot 350 * @wdd: the watchdog device to get the remaining time from 351 * @timeleft: the time that's left 352 * 353 * The caller must hold wd_data->lock. 354 * 355 * Get the time before a watchdog will reboot (if not pinged). 356 */ 357 358 static int watchdog_get_timeleft(struct watchdog_device *wdd, 359 unsigned int *timeleft) 360 { 361 *timeleft = 0; 362 363 if (!wdd->ops->get_timeleft) 364 return -EOPNOTSUPP; 365 366 *timeleft = wdd->ops->get_timeleft(wdd); 367 368 return 0; 369 } 370 371 #ifdef CONFIG_WATCHDOG_SYSFS 372 static ssize_t nowayout_show(struct device *dev, struct device_attribute *attr, 373 char *buf) 374 { 375 struct watchdog_device *wdd = dev_get_drvdata(dev); 376 377 return sprintf(buf, "%d\n", !!test_bit(WDOG_NO_WAY_OUT, &wdd->status)); 378 } 379 static DEVICE_ATTR_RO(nowayout); 380 381 static ssize_t status_show(struct device *dev, struct device_attribute *attr, 382 char *buf) 383 { 384 struct watchdog_device *wdd = dev_get_drvdata(dev); 385 struct watchdog_core_data *wd_data = wdd->wd_data; 386 unsigned int status; 387 388 mutex_lock(&wd_data->lock); 389 status = watchdog_get_status(wdd); 390 mutex_unlock(&wd_data->lock); 391 392 return sprintf(buf, "0x%x\n", status); 393 } 394 static DEVICE_ATTR_RO(status); 395 396 static ssize_t bootstatus_show(struct device *dev, 397 struct device_attribute *attr, char *buf) 398 { 399 struct watchdog_device *wdd = dev_get_drvdata(dev); 400 401 return sprintf(buf, "%u\n", wdd->bootstatus); 402 } 403 static DEVICE_ATTR_RO(bootstatus); 404 405 static ssize_t timeleft_show(struct device *dev, struct device_attribute *attr, 406 char *buf) 407 { 408 struct watchdog_device *wdd = dev_get_drvdata(dev); 409 struct watchdog_core_data *wd_data = wdd->wd_data; 410 ssize_t status; 411 unsigned int val; 412 413 mutex_lock(&wd_data->lock); 414 status = watchdog_get_timeleft(wdd, &val); 415 mutex_unlock(&wd_data->lock); 416 if (!status) 417 status = sprintf(buf, "%u\n", val); 418 419 return status; 420 } 421 static DEVICE_ATTR_RO(timeleft); 422 423 static ssize_t timeout_show(struct device *dev, struct device_attribute *attr, 424 char *buf) 425 { 426 struct watchdog_device *wdd = dev_get_drvdata(dev); 427 428 return sprintf(buf, "%u\n", wdd->timeout); 429 } 430 static DEVICE_ATTR_RO(timeout); 431 432 static ssize_t identity_show(struct device *dev, struct device_attribute *attr, 433 char *buf) 434 { 435 struct watchdog_device *wdd = dev_get_drvdata(dev); 436 437 return sprintf(buf, "%s\n", wdd->info->identity); 438 } 439 static DEVICE_ATTR_RO(identity); 440 441 static ssize_t state_show(struct device *dev, struct device_attribute *attr, 442 char *buf) 443 { 444 struct watchdog_device *wdd = dev_get_drvdata(dev); 445 446 if (watchdog_active(wdd)) 447 return sprintf(buf, "active\n"); 448 449 return sprintf(buf, "inactive\n"); 450 } 451 static DEVICE_ATTR_RO(state); 452 453 static umode_t wdt_is_visible(struct kobject *kobj, struct attribute *attr, 454 int n) 455 { 456 struct device *dev = container_of(kobj, struct device, kobj); 457 struct watchdog_device *wdd = dev_get_drvdata(dev); 458 umode_t mode = attr->mode; 459 460 if (attr == &dev_attr_timeleft.attr && !wdd->ops->get_timeleft) 461 mode = 0; 462 463 return mode; 464 } 465 static struct attribute *wdt_attrs[] = { 466 &dev_attr_state.attr, 467 &dev_attr_identity.attr, 468 &dev_attr_timeout.attr, 469 &dev_attr_timeleft.attr, 470 &dev_attr_bootstatus.attr, 471 &dev_attr_status.attr, 472 &dev_attr_nowayout.attr, 473 NULL, 474 }; 475 476 static const struct attribute_group wdt_group = { 477 .attrs = wdt_attrs, 478 .is_visible = wdt_is_visible, 479 }; 480 __ATTRIBUTE_GROUPS(wdt); 481 #else 482 #define wdt_groups NULL 483 #endif 484 485 /* 486 * watchdog_ioctl_op: call the watchdog drivers ioctl op if defined 487 * @wdd: the watchdog device to do the ioctl on 488 * @cmd: watchdog command 489 * @arg: argument pointer 490 * 491 * The caller must hold wd_data->lock. 492 */ 493 494 static int watchdog_ioctl_op(struct watchdog_device *wdd, unsigned int cmd, 495 unsigned long arg) 496 { 497 if (!wdd->ops->ioctl) 498 return -ENOIOCTLCMD; 499 500 return wdd->ops->ioctl(wdd, cmd, arg); 501 } 502 503 /* 504 * watchdog_write: writes to the watchdog. 505 * @file: file from VFS 506 * @data: user address of data 507 * @len: length of data 508 * @ppos: pointer to the file offset 509 * 510 * A write to a watchdog device is defined as a keepalive ping. 511 * Writing the magic 'V' sequence allows the next close to turn 512 * off the watchdog (if 'nowayout' is not set). 513 */ 514 515 static ssize_t watchdog_write(struct file *file, const char __user *data, 516 size_t len, loff_t *ppos) 517 { 518 struct watchdog_core_data *wd_data = file->private_data; 519 struct watchdog_device *wdd; 520 int err; 521 size_t i; 522 char c; 523 524 if (len == 0) 525 return 0; 526 527 /* 528 * Note: just in case someone wrote the magic character 529 * five months ago... 530 */ 531 clear_bit(_WDOG_ALLOW_RELEASE, &wd_data->status); 532 533 /* scan to see whether or not we got the magic character */ 534 for (i = 0; i != len; i++) { 535 if (get_user(c, data + i)) 536 return -EFAULT; 537 if (c == 'V') 538 set_bit(_WDOG_ALLOW_RELEASE, &wd_data->status); 539 } 540 541 /* someone wrote to us, so we send the watchdog a keepalive ping */ 542 543 err = -ENODEV; 544 mutex_lock(&wd_data->lock); 545 wdd = wd_data->wdd; 546 if (wdd) 547 err = watchdog_ping(wdd); 548 mutex_unlock(&wd_data->lock); 549 550 if (err < 0) 551 return err; 552 553 return len; 554 } 555 556 /* 557 * watchdog_ioctl: handle the different ioctl's for the watchdog device. 558 * @file: file handle to the device 559 * @cmd: watchdog command 560 * @arg: argument pointer 561 * 562 * The watchdog API defines a common set of functions for all watchdogs 563 * according to their available features. 564 */ 565 566 static long watchdog_ioctl(struct file *file, unsigned int cmd, 567 unsigned long arg) 568 { 569 struct watchdog_core_data *wd_data = file->private_data; 570 void __user *argp = (void __user *)arg; 571 struct watchdog_device *wdd; 572 int __user *p = argp; 573 unsigned int val; 574 int err; 575 576 mutex_lock(&wd_data->lock); 577 578 wdd = wd_data->wdd; 579 if (!wdd) { 580 err = -ENODEV; 581 goto out_ioctl; 582 } 583 584 err = watchdog_ioctl_op(wdd, cmd, arg); 585 if (err != -ENOIOCTLCMD) 586 goto out_ioctl; 587 588 switch (cmd) { 589 case WDIOC_GETSUPPORT: 590 err = copy_to_user(argp, wdd->info, 591 sizeof(struct watchdog_info)) ? -EFAULT : 0; 592 break; 593 case WDIOC_GETSTATUS: 594 val = watchdog_get_status(wdd); 595 err = put_user(val, p); 596 break; 597 case WDIOC_GETBOOTSTATUS: 598 err = put_user(wdd->bootstatus, p); 599 break; 600 case WDIOC_SETOPTIONS: 601 if (get_user(val, p)) { 602 err = -EFAULT; 603 break; 604 } 605 if (val & WDIOS_DISABLECARD) { 606 err = watchdog_stop(wdd); 607 if (err < 0) 608 break; 609 } 610 if (val & WDIOS_ENABLECARD) 611 err = watchdog_start(wdd); 612 break; 613 case WDIOC_KEEPALIVE: 614 if (!(wdd->info->options & WDIOF_KEEPALIVEPING)) { 615 err = -EOPNOTSUPP; 616 break; 617 } 618 err = watchdog_ping(wdd); 619 break; 620 case WDIOC_SETTIMEOUT: 621 if (get_user(val, p)) { 622 err = -EFAULT; 623 break; 624 } 625 err = watchdog_set_timeout(wdd, val); 626 if (err < 0) 627 break; 628 /* If the watchdog is active then we send a keepalive ping 629 * to make sure that the watchdog keep's running (and if 630 * possible that it takes the new timeout) */ 631 err = watchdog_ping(wdd); 632 if (err < 0) 633 break; 634 /* Fall */ 635 case WDIOC_GETTIMEOUT: 636 /* timeout == 0 means that we don't know the timeout */ 637 if (wdd->timeout == 0) { 638 err = -EOPNOTSUPP; 639 break; 640 } 641 err = put_user(wdd->timeout, p); 642 break; 643 case WDIOC_GETTIMELEFT: 644 err = watchdog_get_timeleft(wdd, &val); 645 if (err < 0) 646 break; 647 err = put_user(val, p); 648 break; 649 default: 650 err = -ENOTTY; 651 break; 652 } 653 654 out_ioctl: 655 mutex_unlock(&wd_data->lock); 656 return err; 657 } 658 659 /* 660 * watchdog_open: open the /dev/watchdog* devices. 661 * @inode: inode of device 662 * @file: file handle to device 663 * 664 * When the /dev/watchdog* device gets opened, we start the watchdog. 665 * Watch out: the /dev/watchdog device is single open, so we make sure 666 * it can only be opened once. 667 */ 668 669 static int watchdog_open(struct inode *inode, struct file *file) 670 { 671 struct watchdog_core_data *wd_data; 672 struct watchdog_device *wdd; 673 int err; 674 675 /* Get the corresponding watchdog device */ 676 if (imajor(inode) == MISC_MAJOR) 677 wd_data = old_wd_data; 678 else 679 wd_data = container_of(inode->i_cdev, struct watchdog_core_data, 680 cdev); 681 682 /* the watchdog is single open! */ 683 if (test_and_set_bit(_WDOG_DEV_OPEN, &wd_data->status)) 684 return -EBUSY; 685 686 wdd = wd_data->wdd; 687 688 /* 689 * If the /dev/watchdog device is open, we don't want the module 690 * to be unloaded. 691 */ 692 if (!watchdog_hw_running(wdd) && !try_module_get(wdd->ops->owner)) { 693 err = -EBUSY; 694 goto out_clear; 695 } 696 697 err = watchdog_start(wdd); 698 if (err < 0) 699 goto out_mod; 700 701 file->private_data = wd_data; 702 703 if (!watchdog_hw_running(wdd)) 704 kref_get(&wd_data->kref); 705 706 /* dev/watchdog is a virtual (and thus non-seekable) filesystem */ 707 return nonseekable_open(inode, file); 708 709 out_mod: 710 module_put(wd_data->wdd->ops->owner); 711 out_clear: 712 clear_bit(_WDOG_DEV_OPEN, &wd_data->status); 713 return err; 714 } 715 716 static void watchdog_core_data_release(struct kref *kref) 717 { 718 struct watchdog_core_data *wd_data; 719 720 wd_data = container_of(kref, struct watchdog_core_data, kref); 721 722 kfree(wd_data); 723 } 724 725 /* 726 * watchdog_release: release the watchdog device. 727 * @inode: inode of device 728 * @file: file handle to device 729 * 730 * This is the code for when /dev/watchdog gets closed. We will only 731 * stop the watchdog when we have received the magic char (and nowayout 732 * was not set), else the watchdog will keep running. 733 */ 734 735 static int watchdog_release(struct inode *inode, struct file *file) 736 { 737 struct watchdog_core_data *wd_data = file->private_data; 738 struct watchdog_device *wdd; 739 int err = -EBUSY; 740 bool running; 741 742 mutex_lock(&wd_data->lock); 743 744 wdd = wd_data->wdd; 745 if (!wdd) 746 goto done; 747 748 /* 749 * We only stop the watchdog if we received the magic character 750 * or if WDIOF_MAGICCLOSE is not set. If nowayout was set then 751 * watchdog_stop will fail. 752 */ 753 if (!test_bit(WDOG_ACTIVE, &wdd->status)) 754 err = 0; 755 else if (test_and_clear_bit(_WDOG_ALLOW_RELEASE, &wd_data->status) || 756 !(wdd->info->options & WDIOF_MAGICCLOSE)) 757 err = watchdog_stop(wdd); 758 759 /* If the watchdog was not stopped, send a keepalive ping */ 760 if (err < 0) { 761 pr_crit("watchdog%d: watchdog did not stop!\n", wdd->id); 762 watchdog_ping(wdd); 763 } 764 765 watchdog_update_worker(wdd); 766 767 /* make sure that /dev/watchdog can be re-opened */ 768 clear_bit(_WDOG_DEV_OPEN, &wd_data->status); 769 770 done: 771 running = wdd && watchdog_hw_running(wdd); 772 mutex_unlock(&wd_data->lock); 773 /* 774 * Allow the owner module to be unloaded again unless the watchdog 775 * is still running. If the watchdog is still running, it can not 776 * be stopped, and its driver must not be unloaded. 777 */ 778 if (!running) { 779 module_put(wd_data->cdev.owner); 780 kref_put(&wd_data->kref, watchdog_core_data_release); 781 } 782 return 0; 783 } 784 785 static const struct file_operations watchdog_fops = { 786 .owner = THIS_MODULE, 787 .write = watchdog_write, 788 .unlocked_ioctl = watchdog_ioctl, 789 .open = watchdog_open, 790 .release = watchdog_release, 791 }; 792 793 static struct miscdevice watchdog_miscdev = { 794 .minor = WATCHDOG_MINOR, 795 .name = "watchdog", 796 .fops = &watchdog_fops, 797 }; 798 799 /* 800 * watchdog_cdev_register: register watchdog character device 801 * @wdd: watchdog device 802 * @devno: character device number 803 * 804 * Register a watchdog character device including handling the legacy 805 * /dev/watchdog node. /dev/watchdog is actually a miscdevice and 806 * thus we set it up like that. 807 */ 808 809 static int watchdog_cdev_register(struct watchdog_device *wdd, dev_t devno) 810 { 811 struct watchdog_core_data *wd_data; 812 int err; 813 814 wd_data = kzalloc(sizeof(struct watchdog_core_data), GFP_KERNEL); 815 if (!wd_data) 816 return -ENOMEM; 817 kref_init(&wd_data->kref); 818 mutex_init(&wd_data->lock); 819 820 wd_data->wdd = wdd; 821 wdd->wd_data = wd_data; 822 823 if (!watchdog_wq) 824 return -ENODEV; 825 826 INIT_DELAYED_WORK(&wd_data->work, watchdog_ping_work); 827 828 if (wdd->id == 0) { 829 old_wd_data = wd_data; 830 watchdog_miscdev.parent = wdd->parent; 831 err = misc_register(&watchdog_miscdev); 832 if (err != 0) { 833 pr_err("%s: cannot register miscdev on minor=%d (err=%d).\n", 834 wdd->info->identity, WATCHDOG_MINOR, err); 835 if (err == -EBUSY) 836 pr_err("%s: a legacy watchdog module is probably present.\n", 837 wdd->info->identity); 838 old_wd_data = NULL; 839 kfree(wd_data); 840 return err; 841 } 842 } 843 844 /* Fill in the data structures */ 845 cdev_init(&wd_data->cdev, &watchdog_fops); 846 wd_data->cdev.owner = wdd->ops->owner; 847 848 /* Add the device */ 849 err = cdev_add(&wd_data->cdev, devno, 1); 850 if (err) { 851 pr_err("watchdog%d unable to add device %d:%d\n", 852 wdd->id, MAJOR(watchdog_devt), wdd->id); 853 if (wdd->id == 0) { 854 misc_deregister(&watchdog_miscdev); 855 old_wd_data = NULL; 856 kref_put(&wd_data->kref, watchdog_core_data_release); 857 } 858 return err; 859 } 860 861 /* Record time of most recent heartbeat as 'just before now'. */ 862 wd_data->last_hw_keepalive = jiffies - 1; 863 864 /* 865 * If the watchdog is running, prevent its driver from being unloaded, 866 * and schedule an immediate ping. 867 */ 868 if (watchdog_hw_running(wdd)) { 869 __module_get(wdd->ops->owner); 870 kref_get(&wd_data->kref); 871 queue_delayed_work(watchdog_wq, &wd_data->work, 0); 872 } 873 874 return 0; 875 } 876 877 /* 878 * watchdog_cdev_unregister: unregister watchdog character device 879 * @watchdog: watchdog device 880 * 881 * Unregister watchdog character device and if needed the legacy 882 * /dev/watchdog device. 883 */ 884 885 static void watchdog_cdev_unregister(struct watchdog_device *wdd) 886 { 887 struct watchdog_core_data *wd_data = wdd->wd_data; 888 889 cdev_del(&wd_data->cdev); 890 if (wdd->id == 0) { 891 misc_deregister(&watchdog_miscdev); 892 old_wd_data = NULL; 893 } 894 895 mutex_lock(&wd_data->lock); 896 wd_data->wdd = NULL; 897 wdd->wd_data = NULL; 898 mutex_unlock(&wd_data->lock); 899 900 cancel_delayed_work_sync(&wd_data->work); 901 902 kref_put(&wd_data->kref, watchdog_core_data_release); 903 } 904 905 static struct class watchdog_class = { 906 .name = "watchdog", 907 .owner = THIS_MODULE, 908 .dev_groups = wdt_groups, 909 }; 910 911 /* 912 * watchdog_dev_register: register a watchdog device 913 * @wdd: watchdog device 914 * 915 * Register a watchdog device including handling the legacy 916 * /dev/watchdog node. /dev/watchdog is actually a miscdevice and 917 * thus we set it up like that. 918 */ 919 920 int watchdog_dev_register(struct watchdog_device *wdd) 921 { 922 struct device *dev; 923 dev_t devno; 924 int ret; 925 926 devno = MKDEV(MAJOR(watchdog_devt), wdd->id); 927 928 ret = watchdog_cdev_register(wdd, devno); 929 if (ret) 930 return ret; 931 932 dev = device_create_with_groups(&watchdog_class, wdd->parent, 933 devno, wdd, wdd->groups, 934 "watchdog%d", wdd->id); 935 if (IS_ERR(dev)) { 936 watchdog_cdev_unregister(wdd); 937 return PTR_ERR(dev); 938 } 939 940 return ret; 941 } 942 943 /* 944 * watchdog_dev_unregister: unregister a watchdog device 945 * @watchdog: watchdog device 946 * 947 * Unregister watchdog device and if needed the legacy 948 * /dev/watchdog device. 949 */ 950 951 void watchdog_dev_unregister(struct watchdog_device *wdd) 952 { 953 device_destroy(&watchdog_class, wdd->wd_data->cdev.dev); 954 watchdog_cdev_unregister(wdd); 955 } 956 957 /* 958 * watchdog_dev_init: init dev part of watchdog core 959 * 960 * Allocate a range of chardev nodes to use for watchdog devices 961 */ 962 963 int __init watchdog_dev_init(void) 964 { 965 int err; 966 967 watchdog_wq = alloc_workqueue("watchdogd", 968 WQ_HIGHPRI | WQ_MEM_RECLAIM, 0); 969 if (!watchdog_wq) { 970 pr_err("Failed to create watchdog workqueue\n"); 971 return -ENOMEM; 972 } 973 974 err = class_register(&watchdog_class); 975 if (err < 0) { 976 pr_err("couldn't register class\n"); 977 goto err_register; 978 } 979 980 err = alloc_chrdev_region(&watchdog_devt, 0, MAX_DOGS, "watchdog"); 981 if (err < 0) { 982 pr_err("watchdog: unable to allocate char dev region\n"); 983 goto err_alloc; 984 } 985 986 return 0; 987 988 err_alloc: 989 class_unregister(&watchdog_class); 990 err_register: 991 destroy_workqueue(watchdog_wq); 992 return err; 993 } 994 995 /* 996 * watchdog_dev_exit: exit dev part of watchdog core 997 * 998 * Release the range of chardev nodes used for watchdog devices 999 */ 1000 1001 void __exit watchdog_dev_exit(void) 1002 { 1003 unregister_chrdev_region(watchdog_devt, MAX_DOGS); 1004 class_unregister(&watchdog_class); 1005 destroy_workqueue(watchdog_wq); 1006 } 1007