1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * (C) Copyright 2002-2004, 2007 Greg Kroah-Hartman <greg@kroah.com> 4 * (C) Copyright 2007 Novell Inc. 5 */ 6 7 #include <linux/pci.h> 8 #include <linux/module.h> 9 #include <linux/init.h> 10 #include <linux/device.h> 11 #include <linux/mempolicy.h> 12 #include <linux/string.h> 13 #include <linux/slab.h> 14 #include <linux/sched.h> 15 #include <linux/sched/isolation.h> 16 #include <linux/cpu.h> 17 #include <linux/pm_runtime.h> 18 #include <linux/suspend.h> 19 #include <linux/kexec.h> 20 #include <linux/of_device.h> 21 #include <linux/acpi.h> 22 #include <linux/dma-map-ops.h> 23 #include "pci.h" 24 #include "pcie/portdrv.h" 25 26 struct pci_dynid { 27 struct list_head node; 28 struct pci_device_id id; 29 }; 30 31 /** 32 * pci_add_dynid - add a new PCI device ID to this driver and re-probe devices 33 * @drv: target pci driver 34 * @vendor: PCI vendor ID 35 * @device: PCI device ID 36 * @subvendor: PCI subvendor ID 37 * @subdevice: PCI subdevice ID 38 * @class: PCI class 39 * @class_mask: PCI class mask 40 * @driver_data: private driver data 41 * 42 * Adds a new dynamic pci device ID to this driver and causes the 43 * driver to probe for all devices again. @drv must have been 44 * registered prior to calling this function. 45 * 46 * CONTEXT: 47 * Does GFP_KERNEL allocation. 48 * 49 * RETURNS: 50 * 0 on success, -errno on failure. 51 */ 52 int pci_add_dynid(struct pci_driver *drv, 53 unsigned int vendor, unsigned int device, 54 unsigned int subvendor, unsigned int subdevice, 55 unsigned int class, unsigned int class_mask, 56 unsigned long driver_data) 57 { 58 struct pci_dynid *dynid; 59 60 dynid = kzalloc(sizeof(*dynid), GFP_KERNEL); 61 if (!dynid) 62 return -ENOMEM; 63 64 dynid->id.vendor = vendor; 65 dynid->id.device = device; 66 dynid->id.subvendor = subvendor; 67 dynid->id.subdevice = subdevice; 68 dynid->id.class = class; 69 dynid->id.class_mask = class_mask; 70 dynid->id.driver_data = driver_data; 71 72 spin_lock(&drv->dynids.lock); 73 list_add_tail(&dynid->node, &drv->dynids.list); 74 spin_unlock(&drv->dynids.lock); 75 76 return driver_attach(&drv->driver); 77 } 78 EXPORT_SYMBOL_GPL(pci_add_dynid); 79 80 static void pci_free_dynids(struct pci_driver *drv) 81 { 82 struct pci_dynid *dynid, *n; 83 84 spin_lock(&drv->dynids.lock); 85 list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { 86 list_del(&dynid->node); 87 kfree(dynid); 88 } 89 spin_unlock(&drv->dynids.lock); 90 } 91 92 /** 93 * pci_match_id - See if a PCI device matches a given pci_id table 94 * @ids: array of PCI device ID structures to search in 95 * @dev: the PCI device structure to match against. 96 * 97 * Used by a driver to check whether a PCI device is in its list of 98 * supported devices. Returns the matching pci_device_id structure or 99 * %NULL if there is no match. 100 * 101 * Deprecated; don't use this as it will not catch any dynamic IDs 102 * that a driver might want to check for. 103 */ 104 const struct pci_device_id *pci_match_id(const struct pci_device_id *ids, 105 struct pci_dev *dev) 106 { 107 if (ids) { 108 while (ids->vendor || ids->subvendor || ids->class_mask) { 109 if (pci_match_one_device(ids, dev)) 110 return ids; 111 ids++; 112 } 113 } 114 return NULL; 115 } 116 EXPORT_SYMBOL(pci_match_id); 117 118 static const struct pci_device_id pci_device_id_any = { 119 .vendor = PCI_ANY_ID, 120 .device = PCI_ANY_ID, 121 .subvendor = PCI_ANY_ID, 122 .subdevice = PCI_ANY_ID, 123 }; 124 125 /** 126 * pci_match_device - See if a device matches a driver's list of IDs 127 * @drv: the PCI driver to match against 128 * @dev: the PCI device structure to match against 129 * 130 * Used by a driver to check whether a PCI device is in its list of 131 * supported devices or in the dynids list, which may have been augmented 132 * via the sysfs "new_id" file. Returns the matching pci_device_id 133 * structure or %NULL if there is no match. 134 */ 135 static const struct pci_device_id *pci_match_device(struct pci_driver *drv, 136 struct pci_dev *dev) 137 { 138 struct pci_dynid *dynid; 139 const struct pci_device_id *found_id = NULL, *ids; 140 141 /* When driver_override is set, only bind to the matching driver */ 142 if (dev->driver_override && strcmp(dev->driver_override, drv->name)) 143 return NULL; 144 145 /* Look at the dynamic ids first, before the static ones */ 146 spin_lock(&drv->dynids.lock); 147 list_for_each_entry(dynid, &drv->dynids.list, node) { 148 if (pci_match_one_device(&dynid->id, dev)) { 149 found_id = &dynid->id; 150 break; 151 } 152 } 153 spin_unlock(&drv->dynids.lock); 154 155 if (found_id) 156 return found_id; 157 158 for (ids = drv->id_table; (found_id = pci_match_id(ids, dev)); 159 ids = found_id + 1) { 160 /* 161 * The match table is split based on driver_override. 162 * In case override_only was set, enforce driver_override 163 * matching. 164 */ 165 if (found_id->override_only) { 166 if (dev->driver_override) 167 return found_id; 168 } else { 169 return found_id; 170 } 171 } 172 173 /* driver_override will always match, send a dummy id */ 174 if (dev->driver_override) 175 return &pci_device_id_any; 176 return NULL; 177 } 178 179 /** 180 * new_id_store - sysfs frontend to pci_add_dynid() 181 * @driver: target device driver 182 * @buf: buffer for scanning device ID data 183 * @count: input size 184 * 185 * Allow PCI IDs to be added to an existing driver via sysfs. 186 */ 187 static ssize_t new_id_store(struct device_driver *driver, const char *buf, 188 size_t count) 189 { 190 struct pci_driver *pdrv = to_pci_driver(driver); 191 const struct pci_device_id *ids = pdrv->id_table; 192 u32 vendor, device, subvendor = PCI_ANY_ID, 193 subdevice = PCI_ANY_ID, class = 0, class_mask = 0; 194 unsigned long driver_data = 0; 195 int fields = 0; 196 int retval = 0; 197 198 fields = sscanf(buf, "%x %x %x %x %x %x %lx", 199 &vendor, &device, &subvendor, &subdevice, 200 &class, &class_mask, &driver_data); 201 if (fields < 2) 202 return -EINVAL; 203 204 if (fields != 7) { 205 struct pci_dev *pdev = kzalloc(sizeof(*pdev), GFP_KERNEL); 206 if (!pdev) 207 return -ENOMEM; 208 209 pdev->vendor = vendor; 210 pdev->device = device; 211 pdev->subsystem_vendor = subvendor; 212 pdev->subsystem_device = subdevice; 213 pdev->class = class; 214 215 if (pci_match_device(pdrv, pdev)) 216 retval = -EEXIST; 217 218 kfree(pdev); 219 220 if (retval) 221 return retval; 222 } 223 224 /* Only accept driver_data values that match an existing id_table 225 entry */ 226 if (ids) { 227 retval = -EINVAL; 228 while (ids->vendor || ids->subvendor || ids->class_mask) { 229 if (driver_data == ids->driver_data) { 230 retval = 0; 231 break; 232 } 233 ids++; 234 } 235 if (retval) /* No match */ 236 return retval; 237 } 238 239 retval = pci_add_dynid(pdrv, vendor, device, subvendor, subdevice, 240 class, class_mask, driver_data); 241 if (retval) 242 return retval; 243 return count; 244 } 245 static DRIVER_ATTR_WO(new_id); 246 247 /** 248 * remove_id_store - remove a PCI device ID from this driver 249 * @driver: target device driver 250 * @buf: buffer for scanning device ID data 251 * @count: input size 252 * 253 * Removes a dynamic pci device ID to this driver. 254 */ 255 static ssize_t remove_id_store(struct device_driver *driver, const char *buf, 256 size_t count) 257 { 258 struct pci_dynid *dynid, *n; 259 struct pci_driver *pdrv = to_pci_driver(driver); 260 u32 vendor, device, subvendor = PCI_ANY_ID, 261 subdevice = PCI_ANY_ID, class = 0, class_mask = 0; 262 int fields = 0; 263 size_t retval = -ENODEV; 264 265 fields = sscanf(buf, "%x %x %x %x %x %x", 266 &vendor, &device, &subvendor, &subdevice, 267 &class, &class_mask); 268 if (fields < 2) 269 return -EINVAL; 270 271 spin_lock(&pdrv->dynids.lock); 272 list_for_each_entry_safe(dynid, n, &pdrv->dynids.list, node) { 273 struct pci_device_id *id = &dynid->id; 274 if ((id->vendor == vendor) && 275 (id->device == device) && 276 (subvendor == PCI_ANY_ID || id->subvendor == subvendor) && 277 (subdevice == PCI_ANY_ID || id->subdevice == subdevice) && 278 !((id->class ^ class) & class_mask)) { 279 list_del(&dynid->node); 280 kfree(dynid); 281 retval = count; 282 break; 283 } 284 } 285 spin_unlock(&pdrv->dynids.lock); 286 287 return retval; 288 } 289 static DRIVER_ATTR_WO(remove_id); 290 291 static struct attribute *pci_drv_attrs[] = { 292 &driver_attr_new_id.attr, 293 &driver_attr_remove_id.attr, 294 NULL, 295 }; 296 ATTRIBUTE_GROUPS(pci_drv); 297 298 struct drv_dev_and_id { 299 struct pci_driver *drv; 300 struct pci_dev *dev; 301 const struct pci_device_id *id; 302 }; 303 304 static long local_pci_probe(void *_ddi) 305 { 306 struct drv_dev_and_id *ddi = _ddi; 307 struct pci_dev *pci_dev = ddi->dev; 308 struct pci_driver *pci_drv = ddi->drv; 309 struct device *dev = &pci_dev->dev; 310 int rc; 311 312 /* 313 * Unbound PCI devices are always put in D0, regardless of 314 * runtime PM status. During probe, the device is set to 315 * active and the usage count is incremented. If the driver 316 * supports runtime PM, it should call pm_runtime_put_noidle(), 317 * or any other runtime PM helper function decrementing the usage 318 * count, in its probe routine and pm_runtime_get_noresume() in 319 * its remove routine. 320 */ 321 pm_runtime_get_sync(dev); 322 pci_dev->driver = pci_drv; 323 rc = pci_drv->probe(pci_dev, ddi->id); 324 if (!rc) 325 return rc; 326 if (rc < 0) { 327 pci_dev->driver = NULL; 328 pm_runtime_put_sync(dev); 329 return rc; 330 } 331 /* 332 * Probe function should return < 0 for failure, 0 for success 333 * Treat values > 0 as success, but warn. 334 */ 335 pci_warn(pci_dev, "Driver probe function unexpectedly returned %d\n", 336 rc); 337 return 0; 338 } 339 340 static bool pci_physfn_is_probed(struct pci_dev *dev) 341 { 342 #ifdef CONFIG_PCI_IOV 343 return dev->is_virtfn && dev->physfn->is_probed; 344 #else 345 return false; 346 #endif 347 } 348 349 static int pci_call_probe(struct pci_driver *drv, struct pci_dev *dev, 350 const struct pci_device_id *id) 351 { 352 int error, node, cpu; 353 struct drv_dev_and_id ddi = { drv, dev, id }; 354 355 /* 356 * Execute driver initialization on node where the device is 357 * attached. This way the driver likely allocates its local memory 358 * on the right node. 359 */ 360 node = dev_to_node(&dev->dev); 361 dev->is_probed = 1; 362 363 cpu_hotplug_disable(); 364 365 /* 366 * Prevent nesting work_on_cpu() for the case where a Virtual Function 367 * device is probed from work_on_cpu() of the Physical device. 368 */ 369 if (node < 0 || node >= MAX_NUMNODES || !node_online(node) || 370 pci_physfn_is_probed(dev)) { 371 cpu = nr_cpu_ids; 372 } else { 373 cpumask_var_t wq_domain_mask; 374 375 if (!zalloc_cpumask_var(&wq_domain_mask, GFP_KERNEL)) { 376 error = -ENOMEM; 377 goto out; 378 } 379 cpumask_and(wq_domain_mask, 380 housekeeping_cpumask(HK_TYPE_WQ), 381 housekeeping_cpumask(HK_TYPE_DOMAIN)); 382 383 cpu = cpumask_any_and(cpumask_of_node(node), 384 wq_domain_mask); 385 free_cpumask_var(wq_domain_mask); 386 } 387 388 if (cpu < nr_cpu_ids) 389 error = work_on_cpu(cpu, local_pci_probe, &ddi); 390 else 391 error = local_pci_probe(&ddi); 392 out: 393 dev->is_probed = 0; 394 cpu_hotplug_enable(); 395 return error; 396 } 397 398 /** 399 * __pci_device_probe - check if a driver wants to claim a specific PCI device 400 * @drv: driver to call to check if it wants the PCI device 401 * @pci_dev: PCI device being probed 402 * 403 * returns 0 on success, else error. 404 * side-effect: pci_dev->driver is set to drv when drv claims pci_dev. 405 */ 406 static int __pci_device_probe(struct pci_driver *drv, struct pci_dev *pci_dev) 407 { 408 const struct pci_device_id *id; 409 int error = 0; 410 411 if (drv->probe) { 412 error = -ENODEV; 413 414 id = pci_match_device(drv, pci_dev); 415 if (id) 416 error = pci_call_probe(drv, pci_dev, id); 417 } 418 return error; 419 } 420 421 int __weak pcibios_alloc_irq(struct pci_dev *dev) 422 { 423 return 0; 424 } 425 426 void __weak pcibios_free_irq(struct pci_dev *dev) 427 { 428 } 429 430 #ifdef CONFIG_PCI_IOV 431 static inline bool pci_device_can_probe(struct pci_dev *pdev) 432 { 433 return (!pdev->is_virtfn || pdev->physfn->sriov->drivers_autoprobe || 434 pdev->driver_override); 435 } 436 #else 437 static inline bool pci_device_can_probe(struct pci_dev *pdev) 438 { 439 return true; 440 } 441 #endif 442 443 static int pci_device_probe(struct device *dev) 444 { 445 int error; 446 struct pci_dev *pci_dev = to_pci_dev(dev); 447 struct pci_driver *drv = to_pci_driver(dev->driver); 448 449 if (!pci_device_can_probe(pci_dev)) 450 return -ENODEV; 451 452 pci_assign_irq(pci_dev); 453 454 error = pcibios_alloc_irq(pci_dev); 455 if (error < 0) 456 return error; 457 458 pci_dev_get(pci_dev); 459 error = __pci_device_probe(drv, pci_dev); 460 if (error) { 461 pcibios_free_irq(pci_dev); 462 pci_dev_put(pci_dev); 463 } 464 465 return error; 466 } 467 468 static void pci_device_remove(struct device *dev) 469 { 470 struct pci_dev *pci_dev = to_pci_dev(dev); 471 struct pci_driver *drv = pci_dev->driver; 472 473 if (drv->remove) { 474 pm_runtime_get_sync(dev); 475 drv->remove(pci_dev); 476 pm_runtime_put_noidle(dev); 477 } 478 pcibios_free_irq(pci_dev); 479 pci_dev->driver = NULL; 480 pci_iov_remove(pci_dev); 481 482 /* Undo the runtime PM settings in local_pci_probe() */ 483 pm_runtime_put_sync(dev); 484 485 /* 486 * If the device is still on, set the power state as "unknown", 487 * since it might change by the next time we load the driver. 488 */ 489 if (pci_dev->current_state == PCI_D0) 490 pci_dev->current_state = PCI_UNKNOWN; 491 492 /* 493 * We would love to complain here if pci_dev->is_enabled is set, that 494 * the driver should have called pci_disable_device(), but the 495 * unfortunate fact is there are too many odd BIOS and bridge setups 496 * that don't like drivers doing that all of the time. 497 * Oh well, we can dream of sane hardware when we sleep, no matter how 498 * horrible the crap we have to deal with is when we are awake... 499 */ 500 501 pci_dev_put(pci_dev); 502 } 503 504 static void pci_device_shutdown(struct device *dev) 505 { 506 struct pci_dev *pci_dev = to_pci_dev(dev); 507 struct pci_driver *drv = pci_dev->driver; 508 509 pm_runtime_resume(dev); 510 511 if (drv && drv->shutdown) 512 drv->shutdown(pci_dev); 513 514 /* 515 * If this is a kexec reboot, turn off Bus Master bit on the 516 * device to tell it to not continue to do DMA. Don't touch 517 * devices in D3cold or unknown states. 518 * If it is not a kexec reboot, firmware will hit the PCI 519 * devices with big hammer and stop their DMA any way. 520 */ 521 if (kexec_in_progress && (pci_dev->current_state <= PCI_D3hot)) 522 pci_clear_master(pci_dev); 523 } 524 525 #ifdef CONFIG_PM 526 527 /* Auxiliary functions used for system resume and run-time resume. */ 528 529 /** 530 * pci_restore_standard_config - restore standard config registers of PCI device 531 * @pci_dev: PCI device to handle 532 */ 533 static int pci_restore_standard_config(struct pci_dev *pci_dev) 534 { 535 pci_update_current_state(pci_dev, PCI_UNKNOWN); 536 537 if (pci_dev->current_state != PCI_D0) { 538 int error = pci_set_power_state(pci_dev, PCI_D0); 539 if (error) 540 return error; 541 } 542 543 pci_restore_state(pci_dev); 544 pci_pme_restore(pci_dev); 545 return 0; 546 } 547 548 static void pci_pm_default_resume(struct pci_dev *pci_dev) 549 { 550 pci_fixup_device(pci_fixup_resume, pci_dev); 551 pci_enable_wake(pci_dev, PCI_D0, false); 552 } 553 554 #endif 555 556 #ifdef CONFIG_PM_SLEEP 557 558 static void pci_pm_default_resume_early(struct pci_dev *pci_dev) 559 { 560 pci_power_up(pci_dev); 561 pci_update_current_state(pci_dev, PCI_D0); 562 pci_restore_state(pci_dev); 563 pci_pme_restore(pci_dev); 564 } 565 566 /* 567 * Default "suspend" method for devices that have no driver provided suspend, 568 * or not even a driver at all (second part). 569 */ 570 static void pci_pm_set_unknown_state(struct pci_dev *pci_dev) 571 { 572 /* 573 * mark its power state as "unknown", since we don't know if 574 * e.g. the BIOS will change its device state when we suspend. 575 */ 576 if (pci_dev->current_state == PCI_D0) 577 pci_dev->current_state = PCI_UNKNOWN; 578 } 579 580 /* 581 * Default "resume" method for devices that have no driver provided resume, 582 * or not even a driver at all (second part). 583 */ 584 static int pci_pm_reenable_device(struct pci_dev *pci_dev) 585 { 586 int retval; 587 588 /* if the device was enabled before suspend, re-enable */ 589 retval = pci_reenable_device(pci_dev); 590 /* 591 * if the device was busmaster before the suspend, make it busmaster 592 * again 593 */ 594 if (pci_dev->is_busmaster) 595 pci_set_master(pci_dev); 596 597 return retval; 598 } 599 600 static int pci_legacy_suspend(struct device *dev, pm_message_t state) 601 { 602 struct pci_dev *pci_dev = to_pci_dev(dev); 603 struct pci_driver *drv = pci_dev->driver; 604 605 if (drv && drv->suspend) { 606 pci_power_t prev = pci_dev->current_state; 607 int error; 608 609 error = drv->suspend(pci_dev, state); 610 suspend_report_result(dev, drv->suspend, error); 611 if (error) 612 return error; 613 614 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 615 && pci_dev->current_state != PCI_UNKNOWN) { 616 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 617 "PCI PM: Device state not saved by %pS\n", 618 drv->suspend); 619 } 620 } 621 622 pci_fixup_device(pci_fixup_suspend, pci_dev); 623 624 return 0; 625 } 626 627 static int pci_legacy_suspend_late(struct device *dev, pm_message_t state) 628 { 629 struct pci_dev *pci_dev = to_pci_dev(dev); 630 631 if (!pci_dev->state_saved) 632 pci_save_state(pci_dev); 633 634 pci_pm_set_unknown_state(pci_dev); 635 636 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 637 638 return 0; 639 } 640 641 static int pci_legacy_resume(struct device *dev) 642 { 643 struct pci_dev *pci_dev = to_pci_dev(dev); 644 struct pci_driver *drv = pci_dev->driver; 645 646 pci_fixup_device(pci_fixup_resume, pci_dev); 647 648 return drv && drv->resume ? 649 drv->resume(pci_dev) : pci_pm_reenable_device(pci_dev); 650 } 651 652 /* Auxiliary functions used by the new power management framework */ 653 654 static void pci_pm_default_suspend(struct pci_dev *pci_dev) 655 { 656 /* Disable non-bridge devices without PM support */ 657 if (!pci_has_subordinate(pci_dev)) 658 pci_disable_enabled_device(pci_dev); 659 } 660 661 static bool pci_has_legacy_pm_support(struct pci_dev *pci_dev) 662 { 663 struct pci_driver *drv = pci_dev->driver; 664 bool ret = drv && (drv->suspend || drv->resume); 665 666 /* 667 * Legacy PM support is used by default, so warn if the new framework is 668 * supported as well. Drivers are supposed to support either the 669 * former, or the latter, but not both at the same time. 670 */ 671 pci_WARN(pci_dev, ret && drv->driver.pm, "device %04x:%04x\n", 672 pci_dev->vendor, pci_dev->device); 673 674 return ret; 675 } 676 677 /* New power management framework */ 678 679 static int pci_pm_prepare(struct device *dev) 680 { 681 struct pci_dev *pci_dev = to_pci_dev(dev); 682 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 683 684 if (pm && pm->prepare) { 685 int error = pm->prepare(dev); 686 if (error < 0) 687 return error; 688 689 if (!error && dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_PREPARE)) 690 return 0; 691 } 692 if (pci_dev_need_resume(pci_dev)) 693 return 0; 694 695 /* 696 * The PME setting needs to be adjusted here in case the direct-complete 697 * optimization is used with respect to this device. 698 */ 699 pci_dev_adjust_pme(pci_dev); 700 return 1; 701 } 702 703 static void pci_pm_complete(struct device *dev) 704 { 705 struct pci_dev *pci_dev = to_pci_dev(dev); 706 707 pci_dev_complete_resume(pci_dev); 708 pm_generic_complete(dev); 709 710 /* Resume device if platform firmware has put it in reset-power-on */ 711 if (pm_runtime_suspended(dev) && pm_resume_via_firmware()) { 712 pci_power_t pre_sleep_state = pci_dev->current_state; 713 714 pci_refresh_power_state(pci_dev); 715 /* 716 * On platforms with ACPI this check may also trigger for 717 * devices sharing power resources if one of those power 718 * resources has been activated as a result of a change of the 719 * power state of another device sharing it. However, in that 720 * case it is also better to resume the device, in general. 721 */ 722 if (pci_dev->current_state < pre_sleep_state) 723 pm_request_resume(dev); 724 } 725 } 726 727 #else /* !CONFIG_PM_SLEEP */ 728 729 #define pci_pm_prepare NULL 730 #define pci_pm_complete NULL 731 732 #endif /* !CONFIG_PM_SLEEP */ 733 734 #ifdef CONFIG_SUSPEND 735 static void pcie_pme_root_status_cleanup(struct pci_dev *pci_dev) 736 { 737 /* 738 * Some BIOSes forget to clear Root PME Status bits after system 739 * wakeup, which breaks ACPI-based runtime wakeup on PCI Express. 740 * Clear those bits now just in case (shouldn't hurt). 741 */ 742 if (pci_is_pcie(pci_dev) && 743 (pci_pcie_type(pci_dev) == PCI_EXP_TYPE_ROOT_PORT || 744 pci_pcie_type(pci_dev) == PCI_EXP_TYPE_RC_EC)) 745 pcie_clear_root_pme_status(pci_dev); 746 } 747 748 static int pci_pm_suspend(struct device *dev) 749 { 750 struct pci_dev *pci_dev = to_pci_dev(dev); 751 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 752 753 pci_dev->skip_bus_pm = false; 754 755 if (pci_has_legacy_pm_support(pci_dev)) 756 return pci_legacy_suspend(dev, PMSG_SUSPEND); 757 758 if (!pm) { 759 pci_pm_default_suspend(pci_dev); 760 return 0; 761 } 762 763 /* 764 * PCI devices suspended at run time may need to be resumed at this 765 * point, because in general it may be necessary to reconfigure them for 766 * system suspend. Namely, if the device is expected to wake up the 767 * system from the sleep state, it may have to be reconfigured for this 768 * purpose, or if the device is not expected to wake up the system from 769 * the sleep state, it should be prevented from signaling wakeup events 770 * going forward. 771 * 772 * Also if the driver of the device does not indicate that its system 773 * suspend callbacks can cope with runtime-suspended devices, it is 774 * better to resume the device from runtime suspend here. 775 */ 776 if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || 777 pci_dev_need_resume(pci_dev)) { 778 pm_runtime_resume(dev); 779 pci_dev->state_saved = false; 780 } else { 781 pci_dev_adjust_pme(pci_dev); 782 } 783 784 if (pm->suspend) { 785 pci_power_t prev = pci_dev->current_state; 786 int error; 787 788 error = pm->suspend(dev); 789 suspend_report_result(dev, pm->suspend, error); 790 if (error) 791 return error; 792 793 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 794 && pci_dev->current_state != PCI_UNKNOWN) { 795 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 796 "PCI PM: State of device not saved by %pS\n", 797 pm->suspend); 798 } 799 } 800 801 return 0; 802 } 803 804 static int pci_pm_suspend_late(struct device *dev) 805 { 806 if (dev_pm_skip_suspend(dev)) 807 return 0; 808 809 pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); 810 811 return pm_generic_suspend_late(dev); 812 } 813 814 static int pci_pm_suspend_noirq(struct device *dev) 815 { 816 struct pci_dev *pci_dev = to_pci_dev(dev); 817 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 818 819 if (dev_pm_skip_suspend(dev)) 820 return 0; 821 822 if (pci_has_legacy_pm_support(pci_dev)) 823 return pci_legacy_suspend_late(dev, PMSG_SUSPEND); 824 825 if (!pm) { 826 pci_save_state(pci_dev); 827 goto Fixup; 828 } 829 830 if (pm->suspend_noirq) { 831 pci_power_t prev = pci_dev->current_state; 832 int error; 833 834 error = pm->suspend_noirq(dev); 835 suspend_report_result(dev, pm->suspend_noirq, error); 836 if (error) 837 return error; 838 839 if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 840 && pci_dev->current_state != PCI_UNKNOWN) { 841 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 842 "PCI PM: State of device not saved by %pS\n", 843 pm->suspend_noirq); 844 goto Fixup; 845 } 846 } 847 848 if (pci_dev->skip_bus_pm) { 849 /* 850 * Either the device is a bridge with a child in D0 below it, or 851 * the function is running for the second time in a row without 852 * going through full resume, which is possible only during 853 * suspend-to-idle in a spurious wakeup case. The device should 854 * be in D0 at this point, but if it is a bridge, it may be 855 * necessary to save its state. 856 */ 857 if (!pci_dev->state_saved) 858 pci_save_state(pci_dev); 859 } else if (!pci_dev->state_saved) { 860 pci_save_state(pci_dev); 861 if (pci_power_manageable(pci_dev)) 862 pci_prepare_to_sleep(pci_dev); 863 } 864 865 pci_dbg(pci_dev, "PCI PM: Suspend power state: %s\n", 866 pci_power_name(pci_dev->current_state)); 867 868 if (pci_dev->current_state == PCI_D0) { 869 pci_dev->skip_bus_pm = true; 870 /* 871 * Per PCI PM r1.2, table 6-1, a bridge must be in D0 if any 872 * downstream device is in D0, so avoid changing the power state 873 * of the parent bridge by setting the skip_bus_pm flag for it. 874 */ 875 if (pci_dev->bus->self) 876 pci_dev->bus->self->skip_bus_pm = true; 877 } 878 879 if (pci_dev->skip_bus_pm && pm_suspend_no_platform()) { 880 pci_dbg(pci_dev, "PCI PM: Skipped\n"); 881 goto Fixup; 882 } 883 884 pci_pm_set_unknown_state(pci_dev); 885 886 /* 887 * Some BIOSes from ASUS have a bug: If a USB EHCI host controller's 888 * PCI COMMAND register isn't 0, the BIOS assumes that the controller 889 * hasn't been quiesced and tries to turn it off. If the controller 890 * is already in D3, this can hang or cause memory corruption. 891 * 892 * Since the value of the COMMAND register doesn't matter once the 893 * device has been suspended, we can safely set it to 0 here. 894 */ 895 if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) 896 pci_write_config_word(pci_dev, PCI_COMMAND, 0); 897 898 Fixup: 899 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 900 901 /* 902 * If the target system sleep state is suspend-to-idle, it is sufficient 903 * to check whether or not the device's wakeup settings are good for 904 * runtime PM. Otherwise, the pm_resume_via_firmware() check will cause 905 * pci_pm_complete() to take care of fixing up the device's state 906 * anyway, if need be. 907 */ 908 if (device_can_wakeup(dev) && !device_may_wakeup(dev)) 909 dev->power.may_skip_resume = false; 910 911 return 0; 912 } 913 914 static int pci_pm_resume_noirq(struct device *dev) 915 { 916 struct pci_dev *pci_dev = to_pci_dev(dev); 917 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 918 pci_power_t prev_state = pci_dev->current_state; 919 bool skip_bus_pm = pci_dev->skip_bus_pm; 920 921 if (dev_pm_skip_resume(dev)) 922 return 0; 923 924 /* 925 * In the suspend-to-idle case, devices left in D0 during suspend will 926 * stay in D0, so it is not necessary to restore or update their 927 * configuration here and attempting to put them into D0 again is 928 * pointless, so avoid doing that. 929 */ 930 if (!(skip_bus_pm && pm_suspend_no_platform())) 931 pci_pm_default_resume_early(pci_dev); 932 933 pci_fixup_device(pci_fixup_resume_early, pci_dev); 934 pcie_pme_root_status_cleanup(pci_dev); 935 936 if (!skip_bus_pm && prev_state == PCI_D3cold) 937 pci_bridge_wait_for_secondary_bus(pci_dev); 938 939 if (pci_has_legacy_pm_support(pci_dev)) 940 return 0; 941 942 if (pm && pm->resume_noirq) 943 return pm->resume_noirq(dev); 944 945 return 0; 946 } 947 948 static int pci_pm_resume_early(struct device *dev) 949 { 950 if (dev_pm_skip_resume(dev)) 951 return 0; 952 953 return pm_generic_resume_early(dev); 954 } 955 956 static int pci_pm_resume(struct device *dev) 957 { 958 struct pci_dev *pci_dev = to_pci_dev(dev); 959 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 960 961 /* 962 * This is necessary for the suspend error path in which resume is 963 * called without restoring the standard config registers of the device. 964 */ 965 if (pci_dev->state_saved) 966 pci_restore_standard_config(pci_dev); 967 968 if (pci_has_legacy_pm_support(pci_dev)) 969 return pci_legacy_resume(dev); 970 971 pci_pm_default_resume(pci_dev); 972 973 if (pm) { 974 if (pm->resume) 975 return pm->resume(dev); 976 } else { 977 pci_pm_reenable_device(pci_dev); 978 } 979 980 return 0; 981 } 982 983 #else /* !CONFIG_SUSPEND */ 984 985 #define pci_pm_suspend NULL 986 #define pci_pm_suspend_late NULL 987 #define pci_pm_suspend_noirq NULL 988 #define pci_pm_resume NULL 989 #define pci_pm_resume_early NULL 990 #define pci_pm_resume_noirq NULL 991 992 #endif /* !CONFIG_SUSPEND */ 993 994 #ifdef CONFIG_HIBERNATE_CALLBACKS 995 996 static int pci_pm_freeze(struct device *dev) 997 { 998 struct pci_dev *pci_dev = to_pci_dev(dev); 999 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1000 1001 if (pci_has_legacy_pm_support(pci_dev)) 1002 return pci_legacy_suspend(dev, PMSG_FREEZE); 1003 1004 if (!pm) { 1005 pci_pm_default_suspend(pci_dev); 1006 return 0; 1007 } 1008 1009 /* 1010 * Resume all runtime-suspended devices before creating a snapshot 1011 * image of system memory, because the restore kernel generally cannot 1012 * be expected to always handle them consistently and they need to be 1013 * put into the runtime-active metastate during system resume anyway, 1014 * so it is better to ensure that the state saved in the image will be 1015 * always consistent with that. 1016 */ 1017 pm_runtime_resume(dev); 1018 pci_dev->state_saved = false; 1019 1020 if (pm->freeze) { 1021 int error; 1022 1023 error = pm->freeze(dev); 1024 suspend_report_result(dev, pm->freeze, error); 1025 if (error) 1026 return error; 1027 } 1028 1029 return 0; 1030 } 1031 1032 static int pci_pm_freeze_noirq(struct device *dev) 1033 { 1034 struct pci_dev *pci_dev = to_pci_dev(dev); 1035 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1036 1037 if (pci_has_legacy_pm_support(pci_dev)) 1038 return pci_legacy_suspend_late(dev, PMSG_FREEZE); 1039 1040 if (pm && pm->freeze_noirq) { 1041 int error; 1042 1043 error = pm->freeze_noirq(dev); 1044 suspend_report_result(dev, pm->freeze_noirq, error); 1045 if (error) 1046 return error; 1047 } 1048 1049 if (!pci_dev->state_saved) 1050 pci_save_state(pci_dev); 1051 1052 pci_pm_set_unknown_state(pci_dev); 1053 1054 return 0; 1055 } 1056 1057 static int pci_pm_thaw_noirq(struct device *dev) 1058 { 1059 struct pci_dev *pci_dev = to_pci_dev(dev); 1060 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1061 1062 /* 1063 * The pm->thaw_noirq() callback assumes the device has been 1064 * returned to D0 and its config state has been restored. 1065 * 1066 * In addition, pci_restore_state() restores MSI-X state in MMIO 1067 * space, which requires the device to be in D0, so return it to D0 1068 * in case the driver's "freeze" callbacks put it into a low-power 1069 * state. 1070 */ 1071 pci_set_power_state(pci_dev, PCI_D0); 1072 pci_restore_state(pci_dev); 1073 1074 if (pci_has_legacy_pm_support(pci_dev)) 1075 return 0; 1076 1077 if (pm && pm->thaw_noirq) 1078 return pm->thaw_noirq(dev); 1079 1080 return 0; 1081 } 1082 1083 static int pci_pm_thaw(struct device *dev) 1084 { 1085 struct pci_dev *pci_dev = to_pci_dev(dev); 1086 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1087 int error = 0; 1088 1089 if (pci_has_legacy_pm_support(pci_dev)) 1090 return pci_legacy_resume(dev); 1091 1092 if (pm) { 1093 if (pm->thaw) 1094 error = pm->thaw(dev); 1095 } else { 1096 pci_pm_reenable_device(pci_dev); 1097 } 1098 1099 pci_dev->state_saved = false; 1100 1101 return error; 1102 } 1103 1104 static int pci_pm_poweroff(struct device *dev) 1105 { 1106 struct pci_dev *pci_dev = to_pci_dev(dev); 1107 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1108 1109 if (pci_has_legacy_pm_support(pci_dev)) 1110 return pci_legacy_suspend(dev, PMSG_HIBERNATE); 1111 1112 if (!pm) { 1113 pci_pm_default_suspend(pci_dev); 1114 return 0; 1115 } 1116 1117 /* The reason to do that is the same as in pci_pm_suspend(). */ 1118 if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || 1119 pci_dev_need_resume(pci_dev)) { 1120 pm_runtime_resume(dev); 1121 pci_dev->state_saved = false; 1122 } else { 1123 pci_dev_adjust_pme(pci_dev); 1124 } 1125 1126 if (pm->poweroff) { 1127 int error; 1128 1129 error = pm->poweroff(dev); 1130 suspend_report_result(dev, pm->poweroff, error); 1131 if (error) 1132 return error; 1133 } 1134 1135 return 0; 1136 } 1137 1138 static int pci_pm_poweroff_late(struct device *dev) 1139 { 1140 if (dev_pm_skip_suspend(dev)) 1141 return 0; 1142 1143 pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); 1144 1145 return pm_generic_poweroff_late(dev); 1146 } 1147 1148 static int pci_pm_poweroff_noirq(struct device *dev) 1149 { 1150 struct pci_dev *pci_dev = to_pci_dev(dev); 1151 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1152 1153 if (dev_pm_skip_suspend(dev)) 1154 return 0; 1155 1156 if (pci_has_legacy_pm_support(pci_dev)) 1157 return pci_legacy_suspend_late(dev, PMSG_HIBERNATE); 1158 1159 if (!pm) { 1160 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 1161 return 0; 1162 } 1163 1164 if (pm->poweroff_noirq) { 1165 int error; 1166 1167 error = pm->poweroff_noirq(dev); 1168 suspend_report_result(dev, pm->poweroff_noirq, error); 1169 if (error) 1170 return error; 1171 } 1172 1173 if (!pci_dev->state_saved && !pci_has_subordinate(pci_dev)) 1174 pci_prepare_to_sleep(pci_dev); 1175 1176 /* 1177 * The reason for doing this here is the same as for the analogous code 1178 * in pci_pm_suspend_noirq(). 1179 */ 1180 if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) 1181 pci_write_config_word(pci_dev, PCI_COMMAND, 0); 1182 1183 pci_fixup_device(pci_fixup_suspend_late, pci_dev); 1184 1185 return 0; 1186 } 1187 1188 static int pci_pm_restore_noirq(struct device *dev) 1189 { 1190 struct pci_dev *pci_dev = to_pci_dev(dev); 1191 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1192 1193 pci_pm_default_resume_early(pci_dev); 1194 pci_fixup_device(pci_fixup_resume_early, pci_dev); 1195 1196 if (pci_has_legacy_pm_support(pci_dev)) 1197 return 0; 1198 1199 if (pm && pm->restore_noirq) 1200 return pm->restore_noirq(dev); 1201 1202 return 0; 1203 } 1204 1205 static int pci_pm_restore(struct device *dev) 1206 { 1207 struct pci_dev *pci_dev = to_pci_dev(dev); 1208 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1209 1210 /* 1211 * This is necessary for the hibernation error path in which restore is 1212 * called without restoring the standard config registers of the device. 1213 */ 1214 if (pci_dev->state_saved) 1215 pci_restore_standard_config(pci_dev); 1216 1217 if (pci_has_legacy_pm_support(pci_dev)) 1218 return pci_legacy_resume(dev); 1219 1220 pci_pm_default_resume(pci_dev); 1221 1222 if (pm) { 1223 if (pm->restore) 1224 return pm->restore(dev); 1225 } else { 1226 pci_pm_reenable_device(pci_dev); 1227 } 1228 1229 return 0; 1230 } 1231 1232 #else /* !CONFIG_HIBERNATE_CALLBACKS */ 1233 1234 #define pci_pm_freeze NULL 1235 #define pci_pm_freeze_noirq NULL 1236 #define pci_pm_thaw NULL 1237 #define pci_pm_thaw_noirq NULL 1238 #define pci_pm_poweroff NULL 1239 #define pci_pm_poweroff_late NULL 1240 #define pci_pm_poweroff_noirq NULL 1241 #define pci_pm_restore NULL 1242 #define pci_pm_restore_noirq NULL 1243 1244 #endif /* !CONFIG_HIBERNATE_CALLBACKS */ 1245 1246 #ifdef CONFIG_PM 1247 1248 static int pci_pm_runtime_suspend(struct device *dev) 1249 { 1250 struct pci_dev *pci_dev = to_pci_dev(dev); 1251 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1252 pci_power_t prev = pci_dev->current_state; 1253 int error; 1254 1255 /* 1256 * If pci_dev->driver is not set (unbound), we leave the device in D0, 1257 * but it may go to D3cold when the bridge above it runtime suspends. 1258 * Save its config space in case that happens. 1259 */ 1260 if (!pci_dev->driver) { 1261 pci_save_state(pci_dev); 1262 return 0; 1263 } 1264 1265 pci_dev->state_saved = false; 1266 if (pm && pm->runtime_suspend) { 1267 error = pm->runtime_suspend(dev); 1268 /* 1269 * -EBUSY and -EAGAIN is used to request the runtime PM core 1270 * to schedule a new suspend, so log the event only with debug 1271 * log level. 1272 */ 1273 if (error == -EBUSY || error == -EAGAIN) { 1274 pci_dbg(pci_dev, "can't suspend now (%ps returned %d)\n", 1275 pm->runtime_suspend, error); 1276 return error; 1277 } else if (error) { 1278 pci_err(pci_dev, "can't suspend (%ps returned %d)\n", 1279 pm->runtime_suspend, error); 1280 return error; 1281 } 1282 } 1283 1284 pci_fixup_device(pci_fixup_suspend, pci_dev); 1285 1286 if (pm && pm->runtime_suspend 1287 && !pci_dev->state_saved && pci_dev->current_state != PCI_D0 1288 && pci_dev->current_state != PCI_UNKNOWN) { 1289 pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, 1290 "PCI PM: State of device not saved by %pS\n", 1291 pm->runtime_suspend); 1292 return 0; 1293 } 1294 1295 if (!pci_dev->state_saved) { 1296 pci_save_state(pci_dev); 1297 pci_finish_runtime_suspend(pci_dev); 1298 } 1299 1300 return 0; 1301 } 1302 1303 static int pci_pm_runtime_resume(struct device *dev) 1304 { 1305 struct pci_dev *pci_dev = to_pci_dev(dev); 1306 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1307 pci_power_t prev_state = pci_dev->current_state; 1308 int error = 0; 1309 1310 /* 1311 * Restoring config space is necessary even if the device is not bound 1312 * to a driver because although we left it in D0, it may have gone to 1313 * D3cold when the bridge above it runtime suspended. 1314 */ 1315 pci_restore_standard_config(pci_dev); 1316 1317 if (!pci_dev->driver) 1318 return 0; 1319 1320 pci_fixup_device(pci_fixup_resume_early, pci_dev); 1321 pci_pm_default_resume(pci_dev); 1322 1323 if (prev_state == PCI_D3cold) 1324 pci_bridge_wait_for_secondary_bus(pci_dev); 1325 1326 if (pm && pm->runtime_resume) 1327 error = pm->runtime_resume(dev); 1328 1329 pci_dev->runtime_d3cold = false; 1330 1331 return error; 1332 } 1333 1334 static int pci_pm_runtime_idle(struct device *dev) 1335 { 1336 struct pci_dev *pci_dev = to_pci_dev(dev); 1337 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 1338 1339 /* 1340 * If pci_dev->driver is not set (unbound), the device should 1341 * always remain in D0 regardless of the runtime PM status 1342 */ 1343 if (!pci_dev->driver) 1344 return 0; 1345 1346 if (!pm) 1347 return -ENOSYS; 1348 1349 if (pm->runtime_idle) 1350 return pm->runtime_idle(dev); 1351 1352 return 0; 1353 } 1354 1355 static const struct dev_pm_ops pci_dev_pm_ops = { 1356 .prepare = pci_pm_prepare, 1357 .complete = pci_pm_complete, 1358 .suspend = pci_pm_suspend, 1359 .suspend_late = pci_pm_suspend_late, 1360 .resume = pci_pm_resume, 1361 .resume_early = pci_pm_resume_early, 1362 .freeze = pci_pm_freeze, 1363 .thaw = pci_pm_thaw, 1364 .poweroff = pci_pm_poweroff, 1365 .poweroff_late = pci_pm_poweroff_late, 1366 .restore = pci_pm_restore, 1367 .suspend_noirq = pci_pm_suspend_noirq, 1368 .resume_noirq = pci_pm_resume_noirq, 1369 .freeze_noirq = pci_pm_freeze_noirq, 1370 .thaw_noirq = pci_pm_thaw_noirq, 1371 .poweroff_noirq = pci_pm_poweroff_noirq, 1372 .restore_noirq = pci_pm_restore_noirq, 1373 .runtime_suspend = pci_pm_runtime_suspend, 1374 .runtime_resume = pci_pm_runtime_resume, 1375 .runtime_idle = pci_pm_runtime_idle, 1376 }; 1377 1378 #define PCI_PM_OPS_PTR (&pci_dev_pm_ops) 1379 1380 #else /* !CONFIG_PM */ 1381 1382 #define pci_pm_runtime_suspend NULL 1383 #define pci_pm_runtime_resume NULL 1384 #define pci_pm_runtime_idle NULL 1385 1386 #define PCI_PM_OPS_PTR NULL 1387 1388 #endif /* !CONFIG_PM */ 1389 1390 /** 1391 * __pci_register_driver - register a new pci driver 1392 * @drv: the driver structure to register 1393 * @owner: owner module of drv 1394 * @mod_name: module name string 1395 * 1396 * Adds the driver structure to the list of registered drivers. 1397 * Returns a negative value on error, otherwise 0. 1398 * If no error occurred, the driver remains registered even if 1399 * no device was claimed during registration. 1400 */ 1401 int __pci_register_driver(struct pci_driver *drv, struct module *owner, 1402 const char *mod_name) 1403 { 1404 /* initialize common driver fields */ 1405 drv->driver.name = drv->name; 1406 drv->driver.bus = &pci_bus_type; 1407 drv->driver.owner = owner; 1408 drv->driver.mod_name = mod_name; 1409 drv->driver.groups = drv->groups; 1410 drv->driver.dev_groups = drv->dev_groups; 1411 1412 spin_lock_init(&drv->dynids.lock); 1413 INIT_LIST_HEAD(&drv->dynids.list); 1414 1415 /* register with core */ 1416 return driver_register(&drv->driver); 1417 } 1418 EXPORT_SYMBOL(__pci_register_driver); 1419 1420 /** 1421 * pci_unregister_driver - unregister a pci driver 1422 * @drv: the driver structure to unregister 1423 * 1424 * Deletes the driver structure from the list of registered PCI drivers, 1425 * gives it a chance to clean up by calling its remove() function for 1426 * each device it was responsible for, and marks those devices as 1427 * driverless. 1428 */ 1429 1430 void pci_unregister_driver(struct pci_driver *drv) 1431 { 1432 driver_unregister(&drv->driver); 1433 pci_free_dynids(drv); 1434 } 1435 EXPORT_SYMBOL(pci_unregister_driver); 1436 1437 static struct pci_driver pci_compat_driver = { 1438 .name = "compat" 1439 }; 1440 1441 /** 1442 * pci_dev_driver - get the pci_driver of a device 1443 * @dev: the device to query 1444 * 1445 * Returns the appropriate pci_driver structure or %NULL if there is no 1446 * registered driver for the device. 1447 */ 1448 struct pci_driver *pci_dev_driver(const struct pci_dev *dev) 1449 { 1450 if (dev->driver) 1451 return dev->driver; 1452 else { 1453 int i; 1454 for (i = 0; i <= PCI_ROM_RESOURCE; i++) 1455 if (dev->resource[i].flags & IORESOURCE_BUSY) 1456 return &pci_compat_driver; 1457 } 1458 return NULL; 1459 } 1460 EXPORT_SYMBOL(pci_dev_driver); 1461 1462 /** 1463 * pci_bus_match - Tell if a PCI device structure has a matching PCI device id structure 1464 * @dev: the PCI device structure to match against 1465 * @drv: the device driver to search for matching PCI device id structures 1466 * 1467 * Used by a driver to check whether a PCI device present in the 1468 * system is in its list of supported devices. Returns the matching 1469 * pci_device_id structure or %NULL if there is no match. 1470 */ 1471 static int pci_bus_match(struct device *dev, struct device_driver *drv) 1472 { 1473 struct pci_dev *pci_dev = to_pci_dev(dev); 1474 struct pci_driver *pci_drv; 1475 const struct pci_device_id *found_id; 1476 1477 if (!pci_dev->match_driver) 1478 return 0; 1479 1480 pci_drv = to_pci_driver(drv); 1481 found_id = pci_match_device(pci_drv, pci_dev); 1482 if (found_id) 1483 return 1; 1484 1485 return 0; 1486 } 1487 1488 /** 1489 * pci_dev_get - increments the reference count of the pci device structure 1490 * @dev: the device being referenced 1491 * 1492 * Each live reference to a device should be refcounted. 1493 * 1494 * Drivers for PCI devices should normally record such references in 1495 * their probe() methods, when they bind to a device, and release 1496 * them by calling pci_dev_put(), in their disconnect() methods. 1497 * 1498 * A pointer to the device with the incremented reference counter is returned. 1499 */ 1500 struct pci_dev *pci_dev_get(struct pci_dev *dev) 1501 { 1502 if (dev) 1503 get_device(&dev->dev); 1504 return dev; 1505 } 1506 EXPORT_SYMBOL(pci_dev_get); 1507 1508 /** 1509 * pci_dev_put - release a use of the pci device structure 1510 * @dev: device that's been disconnected 1511 * 1512 * Must be called when a user of a device is finished with it. When the last 1513 * user of the device calls this function, the memory of the device is freed. 1514 */ 1515 void pci_dev_put(struct pci_dev *dev) 1516 { 1517 if (dev) 1518 put_device(&dev->dev); 1519 } 1520 EXPORT_SYMBOL(pci_dev_put); 1521 1522 static int pci_uevent(struct device *dev, struct kobj_uevent_env *env) 1523 { 1524 struct pci_dev *pdev; 1525 1526 if (!dev) 1527 return -ENODEV; 1528 1529 pdev = to_pci_dev(dev); 1530 1531 if (add_uevent_var(env, "PCI_CLASS=%04X", pdev->class)) 1532 return -ENOMEM; 1533 1534 if (add_uevent_var(env, "PCI_ID=%04X:%04X", pdev->vendor, pdev->device)) 1535 return -ENOMEM; 1536 1537 if (add_uevent_var(env, "PCI_SUBSYS_ID=%04X:%04X", pdev->subsystem_vendor, 1538 pdev->subsystem_device)) 1539 return -ENOMEM; 1540 1541 if (add_uevent_var(env, "PCI_SLOT_NAME=%s", pci_name(pdev))) 1542 return -ENOMEM; 1543 1544 if (add_uevent_var(env, "MODALIAS=pci:v%08Xd%08Xsv%08Xsd%08Xbc%02Xsc%02Xi%02X", 1545 pdev->vendor, pdev->device, 1546 pdev->subsystem_vendor, pdev->subsystem_device, 1547 (u8)(pdev->class >> 16), (u8)(pdev->class >> 8), 1548 (u8)(pdev->class))) 1549 return -ENOMEM; 1550 1551 return 0; 1552 } 1553 1554 #if defined(CONFIG_PCIEAER) || defined(CONFIG_EEH) 1555 /** 1556 * pci_uevent_ers - emit a uevent during recovery path of PCI device 1557 * @pdev: PCI device undergoing error recovery 1558 * @err_type: type of error event 1559 */ 1560 void pci_uevent_ers(struct pci_dev *pdev, enum pci_ers_result err_type) 1561 { 1562 int idx = 0; 1563 char *envp[3]; 1564 1565 switch (err_type) { 1566 case PCI_ERS_RESULT_NONE: 1567 case PCI_ERS_RESULT_CAN_RECOVER: 1568 envp[idx++] = "ERROR_EVENT=BEGIN_RECOVERY"; 1569 envp[idx++] = "DEVICE_ONLINE=0"; 1570 break; 1571 case PCI_ERS_RESULT_RECOVERED: 1572 envp[idx++] = "ERROR_EVENT=SUCCESSFUL_RECOVERY"; 1573 envp[idx++] = "DEVICE_ONLINE=1"; 1574 break; 1575 case PCI_ERS_RESULT_DISCONNECT: 1576 envp[idx++] = "ERROR_EVENT=FAILED_RECOVERY"; 1577 envp[idx++] = "DEVICE_ONLINE=0"; 1578 break; 1579 default: 1580 break; 1581 } 1582 1583 if (idx > 0) { 1584 envp[idx++] = NULL; 1585 kobject_uevent_env(&pdev->dev.kobj, KOBJ_CHANGE, envp); 1586 } 1587 } 1588 #endif 1589 1590 static int pci_bus_num_vf(struct device *dev) 1591 { 1592 return pci_num_vf(to_pci_dev(dev)); 1593 } 1594 1595 /** 1596 * pci_dma_configure - Setup DMA configuration 1597 * @dev: ptr to dev structure 1598 * 1599 * Function to update PCI devices's DMA configuration using the same 1600 * info from the OF node or ACPI node of host bridge's parent (if any). 1601 */ 1602 static int pci_dma_configure(struct device *dev) 1603 { 1604 struct device *bridge; 1605 int ret = 0; 1606 1607 bridge = pci_get_host_bridge_device(to_pci_dev(dev)); 1608 1609 if (IS_ENABLED(CONFIG_OF) && bridge->parent && 1610 bridge->parent->of_node) { 1611 ret = of_dma_configure(dev, bridge->parent->of_node, true); 1612 } else if (has_acpi_companion(bridge)) { 1613 struct acpi_device *adev = to_acpi_device_node(bridge->fwnode); 1614 1615 ret = acpi_dma_configure(dev, acpi_get_dma_attr(adev)); 1616 } 1617 1618 pci_put_host_bridge_device(bridge); 1619 return ret; 1620 } 1621 1622 struct bus_type pci_bus_type = { 1623 .name = "pci", 1624 .match = pci_bus_match, 1625 .uevent = pci_uevent, 1626 .probe = pci_device_probe, 1627 .remove = pci_device_remove, 1628 .shutdown = pci_device_shutdown, 1629 .dev_groups = pci_dev_groups, 1630 .bus_groups = pci_bus_groups, 1631 .drv_groups = pci_drv_groups, 1632 .pm = PCI_PM_OPS_PTR, 1633 .num_vf = pci_bus_num_vf, 1634 .dma_configure = pci_dma_configure, 1635 }; 1636 EXPORT_SYMBOL(pci_bus_type); 1637 1638 #ifdef CONFIG_PCIEPORTBUS 1639 static int pcie_port_bus_match(struct device *dev, struct device_driver *drv) 1640 { 1641 struct pcie_device *pciedev; 1642 struct pcie_port_service_driver *driver; 1643 1644 if (drv->bus != &pcie_port_bus_type || dev->bus != &pcie_port_bus_type) 1645 return 0; 1646 1647 pciedev = to_pcie_device(dev); 1648 driver = to_service_driver(drv); 1649 1650 if (driver->service != pciedev->service) 1651 return 0; 1652 1653 if (driver->port_type != PCIE_ANY_PORT && 1654 driver->port_type != pci_pcie_type(pciedev->port)) 1655 return 0; 1656 1657 return 1; 1658 } 1659 1660 struct bus_type pcie_port_bus_type = { 1661 .name = "pci_express", 1662 .match = pcie_port_bus_match, 1663 }; 1664 EXPORT_SYMBOL_GPL(pcie_port_bus_type); 1665 #endif 1666 1667 static int __init pci_driver_init(void) 1668 { 1669 int ret; 1670 1671 ret = bus_register(&pci_bus_type); 1672 if (ret) 1673 return ret; 1674 1675 #ifdef CONFIG_PCIEPORTBUS 1676 ret = bus_register(&pcie_port_bus_type); 1677 if (ret) 1678 return ret; 1679 #endif 1680 dma_debug_add_bus(&pci_bus_type); 1681 return 0; 1682 } 1683 postcore_initcall(pci_driver_init); 1684