1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Copyright IBM Corp. 2012 4 * 5 * Author(s): 6 * Jan Glauber <jang@linux.vnet.ibm.com> 7 * 8 * The System z PCI code is a rewrite from a prototype by 9 * the following people (Kudoz!): 10 * Alexander Schmidt 11 * Christoph Raisch 12 * Hannes Hering 13 * Hoang-Nam Nguyen 14 * Jan-Bernd Themann 15 * Stefan Roscher 16 * Thomas Klein 17 */ 18 19 #define KMSG_COMPONENT "zpci" 20 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt 21 22 #include <linux/kernel.h> 23 #include <linux/slab.h> 24 #include <linux/err.h> 25 #include <linux/export.h> 26 #include <linux/delay.h> 27 #include <linux/seq_file.h> 28 #include <linux/jump_label.h> 29 #include <linux/pci.h> 30 #include <linux/printk.h> 31 32 #include <asm/isc.h> 33 #include <asm/airq.h> 34 #include <asm/facility.h> 35 #include <asm/pci_insn.h> 36 #include <asm/pci_clp.h> 37 #include <asm/pci_dma.h> 38 39 /* list of all detected zpci devices */ 40 static LIST_HEAD(zpci_list); 41 static DEFINE_SPINLOCK(zpci_list_lock); 42 43 static DECLARE_BITMAP(zpci_domain, ZPCI_NR_DEVICES); 44 static DEFINE_SPINLOCK(zpci_domain_lock); 45 46 #define ZPCI_IOMAP_ENTRIES \ 47 min(((unsigned long) ZPCI_NR_DEVICES * PCI_STD_NUM_BARS / 2), \ 48 ZPCI_IOMAP_MAX_ENTRIES) 49 50 static DEFINE_SPINLOCK(zpci_iomap_lock); 51 static unsigned long *zpci_iomap_bitmap; 52 struct zpci_iomap_entry *zpci_iomap_start; 53 EXPORT_SYMBOL_GPL(zpci_iomap_start); 54 55 DEFINE_STATIC_KEY_FALSE(have_mio); 56 57 static struct kmem_cache *zdev_fmb_cache; 58 59 struct zpci_dev *get_zdev_by_fid(u32 fid) 60 { 61 struct zpci_dev *tmp, *zdev = NULL; 62 63 spin_lock(&zpci_list_lock); 64 list_for_each_entry(tmp, &zpci_list, entry) { 65 if (tmp->fid == fid) { 66 zdev = tmp; 67 break; 68 } 69 } 70 spin_unlock(&zpci_list_lock); 71 return zdev; 72 } 73 74 void zpci_remove_reserved_devices(void) 75 { 76 struct zpci_dev *tmp, *zdev; 77 enum zpci_state state; 78 LIST_HEAD(remove); 79 80 spin_lock(&zpci_list_lock); 81 list_for_each_entry_safe(zdev, tmp, &zpci_list, entry) { 82 if (zdev->state == ZPCI_FN_STATE_STANDBY && 83 !clp_get_state(zdev->fid, &state) && 84 state == ZPCI_FN_STATE_RESERVED) 85 list_move_tail(&zdev->entry, &remove); 86 } 87 spin_unlock(&zpci_list_lock); 88 89 list_for_each_entry_safe(zdev, tmp, &remove, entry) 90 zpci_remove_device(zdev); 91 } 92 93 static struct zpci_dev *get_zdev_by_bus(struct pci_bus *bus) 94 { 95 return (bus && bus->sysdata) ? (struct zpci_dev *) bus->sysdata : NULL; 96 } 97 98 int pci_domain_nr(struct pci_bus *bus) 99 { 100 return ((struct zpci_dev *) bus->sysdata)->domain; 101 } 102 EXPORT_SYMBOL_GPL(pci_domain_nr); 103 104 int pci_proc_domain(struct pci_bus *bus) 105 { 106 return pci_domain_nr(bus); 107 } 108 EXPORT_SYMBOL_GPL(pci_proc_domain); 109 110 /* Modify PCI: Register I/O address translation parameters */ 111 int zpci_register_ioat(struct zpci_dev *zdev, u8 dmaas, 112 u64 base, u64 limit, u64 iota) 113 { 114 u64 req = ZPCI_CREATE_REQ(zdev->fh, dmaas, ZPCI_MOD_FC_REG_IOAT); 115 struct zpci_fib fib = {0}; 116 u8 status; 117 118 WARN_ON_ONCE(iota & 0x3fff); 119 fib.pba = base; 120 fib.pal = limit; 121 fib.iota = iota | ZPCI_IOTA_RTTO_FLAG; 122 return zpci_mod_fc(req, &fib, &status) ? -EIO : 0; 123 } 124 125 /* Modify PCI: Unregister I/O address translation parameters */ 126 int zpci_unregister_ioat(struct zpci_dev *zdev, u8 dmaas) 127 { 128 u64 req = ZPCI_CREATE_REQ(zdev->fh, dmaas, ZPCI_MOD_FC_DEREG_IOAT); 129 struct zpci_fib fib = {0}; 130 u8 cc, status; 131 132 cc = zpci_mod_fc(req, &fib, &status); 133 if (cc == 3) /* Function already gone. */ 134 cc = 0; 135 return cc ? -EIO : 0; 136 } 137 138 /* Modify PCI: Set PCI function measurement parameters */ 139 int zpci_fmb_enable_device(struct zpci_dev *zdev) 140 { 141 u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_SET_MEASURE); 142 struct zpci_fib fib = {0}; 143 u8 cc, status; 144 145 if (zdev->fmb || sizeof(*zdev->fmb) < zdev->fmb_length) 146 return -EINVAL; 147 148 zdev->fmb = kmem_cache_zalloc(zdev_fmb_cache, GFP_KERNEL); 149 if (!zdev->fmb) 150 return -ENOMEM; 151 WARN_ON((u64) zdev->fmb & 0xf); 152 153 /* reset software counters */ 154 atomic64_set(&zdev->allocated_pages, 0); 155 atomic64_set(&zdev->mapped_pages, 0); 156 atomic64_set(&zdev->unmapped_pages, 0); 157 158 fib.fmb_addr = virt_to_phys(zdev->fmb); 159 cc = zpci_mod_fc(req, &fib, &status); 160 if (cc) { 161 kmem_cache_free(zdev_fmb_cache, zdev->fmb); 162 zdev->fmb = NULL; 163 } 164 return cc ? -EIO : 0; 165 } 166 167 /* Modify PCI: Disable PCI function measurement */ 168 int zpci_fmb_disable_device(struct zpci_dev *zdev) 169 { 170 u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_SET_MEASURE); 171 struct zpci_fib fib = {0}; 172 u8 cc, status; 173 174 if (!zdev->fmb) 175 return -EINVAL; 176 177 /* Function measurement is disabled if fmb address is zero */ 178 cc = zpci_mod_fc(req, &fib, &status); 179 if (cc == 3) /* Function already gone. */ 180 cc = 0; 181 182 if (!cc) { 183 kmem_cache_free(zdev_fmb_cache, zdev->fmb); 184 zdev->fmb = NULL; 185 } 186 return cc ? -EIO : 0; 187 } 188 189 static int zpci_cfg_load(struct zpci_dev *zdev, int offset, u32 *val, u8 len) 190 { 191 u64 req = ZPCI_CREATE_REQ(zdev->fh, ZPCI_PCIAS_CFGSPC, len); 192 u64 data; 193 int rc; 194 195 rc = __zpci_load(&data, req, offset); 196 if (!rc) { 197 data = le64_to_cpu((__force __le64) data); 198 data >>= (8 - len) * 8; 199 *val = (u32) data; 200 } else 201 *val = 0xffffffff; 202 return rc; 203 } 204 205 static int zpci_cfg_store(struct zpci_dev *zdev, int offset, u32 val, u8 len) 206 { 207 u64 req = ZPCI_CREATE_REQ(zdev->fh, ZPCI_PCIAS_CFGSPC, len); 208 u64 data = val; 209 int rc; 210 211 data <<= (8 - len) * 8; 212 data = (__force u64) cpu_to_le64(data); 213 rc = __zpci_store(data, req, offset); 214 return rc; 215 } 216 217 resource_size_t pcibios_align_resource(void *data, const struct resource *res, 218 resource_size_t size, 219 resource_size_t align) 220 { 221 return 0; 222 } 223 224 /* combine single writes by using store-block insn */ 225 void __iowrite64_copy(void __iomem *to, const void *from, size_t count) 226 { 227 zpci_memcpy_toio(to, from, count); 228 } 229 230 void __iomem *ioremap(unsigned long ioaddr, unsigned long size) 231 { 232 struct vm_struct *area; 233 unsigned long offset; 234 235 if (!size) 236 return NULL; 237 238 if (!static_branch_unlikely(&have_mio)) 239 return (void __iomem *) ioaddr; 240 241 offset = ioaddr & ~PAGE_MASK; 242 ioaddr &= PAGE_MASK; 243 size = PAGE_ALIGN(size + offset); 244 area = get_vm_area(size, VM_IOREMAP); 245 if (!area) 246 return NULL; 247 248 if (ioremap_page_range((unsigned long) area->addr, 249 (unsigned long) area->addr + size, 250 ioaddr, PAGE_KERNEL)) { 251 vunmap(area->addr); 252 return NULL; 253 } 254 return (void __iomem *) ((unsigned long) area->addr + offset); 255 } 256 EXPORT_SYMBOL(ioremap); 257 258 void iounmap(volatile void __iomem *addr) 259 { 260 if (static_branch_likely(&have_mio)) 261 vunmap((__force void *) ((unsigned long) addr & PAGE_MASK)); 262 } 263 EXPORT_SYMBOL(iounmap); 264 265 /* Create a virtual mapping cookie for a PCI BAR */ 266 static void __iomem *pci_iomap_range_fh(struct pci_dev *pdev, int bar, 267 unsigned long offset, unsigned long max) 268 { 269 struct zpci_dev *zdev = to_zpci(pdev); 270 int idx; 271 272 idx = zdev->bars[bar].map_idx; 273 spin_lock(&zpci_iomap_lock); 274 /* Detect overrun */ 275 WARN_ON(!++zpci_iomap_start[idx].count); 276 zpci_iomap_start[idx].fh = zdev->fh; 277 zpci_iomap_start[idx].bar = bar; 278 spin_unlock(&zpci_iomap_lock); 279 280 return (void __iomem *) ZPCI_ADDR(idx) + offset; 281 } 282 283 static void __iomem *pci_iomap_range_mio(struct pci_dev *pdev, int bar, 284 unsigned long offset, 285 unsigned long max) 286 { 287 unsigned long barsize = pci_resource_len(pdev, bar); 288 struct zpci_dev *zdev = to_zpci(pdev); 289 void __iomem *iova; 290 291 iova = ioremap((unsigned long) zdev->bars[bar].mio_wt, barsize); 292 return iova ? iova + offset : iova; 293 } 294 295 void __iomem *pci_iomap_range(struct pci_dev *pdev, int bar, 296 unsigned long offset, unsigned long max) 297 { 298 if (bar >= PCI_STD_NUM_BARS || !pci_resource_len(pdev, bar)) 299 return NULL; 300 301 if (static_branch_likely(&have_mio)) 302 return pci_iomap_range_mio(pdev, bar, offset, max); 303 else 304 return pci_iomap_range_fh(pdev, bar, offset, max); 305 } 306 EXPORT_SYMBOL(pci_iomap_range); 307 308 void __iomem *pci_iomap(struct pci_dev *dev, int bar, unsigned long maxlen) 309 { 310 return pci_iomap_range(dev, bar, 0, maxlen); 311 } 312 EXPORT_SYMBOL(pci_iomap); 313 314 static void __iomem *pci_iomap_wc_range_mio(struct pci_dev *pdev, int bar, 315 unsigned long offset, unsigned long max) 316 { 317 unsigned long barsize = pci_resource_len(pdev, bar); 318 struct zpci_dev *zdev = to_zpci(pdev); 319 void __iomem *iova; 320 321 iova = ioremap((unsigned long) zdev->bars[bar].mio_wb, barsize); 322 return iova ? iova + offset : iova; 323 } 324 325 void __iomem *pci_iomap_wc_range(struct pci_dev *pdev, int bar, 326 unsigned long offset, unsigned long max) 327 { 328 if (bar >= PCI_STD_NUM_BARS || !pci_resource_len(pdev, bar)) 329 return NULL; 330 331 if (static_branch_likely(&have_mio)) 332 return pci_iomap_wc_range_mio(pdev, bar, offset, max); 333 else 334 return pci_iomap_range_fh(pdev, bar, offset, max); 335 } 336 EXPORT_SYMBOL(pci_iomap_wc_range); 337 338 void __iomem *pci_iomap_wc(struct pci_dev *dev, int bar, unsigned long maxlen) 339 { 340 return pci_iomap_wc_range(dev, bar, 0, maxlen); 341 } 342 EXPORT_SYMBOL(pci_iomap_wc); 343 344 static void pci_iounmap_fh(struct pci_dev *pdev, void __iomem *addr) 345 { 346 unsigned int idx = ZPCI_IDX(addr); 347 348 spin_lock(&zpci_iomap_lock); 349 /* Detect underrun */ 350 WARN_ON(!zpci_iomap_start[idx].count); 351 if (!--zpci_iomap_start[idx].count) { 352 zpci_iomap_start[idx].fh = 0; 353 zpci_iomap_start[idx].bar = 0; 354 } 355 spin_unlock(&zpci_iomap_lock); 356 } 357 358 static void pci_iounmap_mio(struct pci_dev *pdev, void __iomem *addr) 359 { 360 iounmap(addr); 361 } 362 363 void pci_iounmap(struct pci_dev *pdev, void __iomem *addr) 364 { 365 if (static_branch_likely(&have_mio)) 366 pci_iounmap_mio(pdev, addr); 367 else 368 pci_iounmap_fh(pdev, addr); 369 } 370 EXPORT_SYMBOL(pci_iounmap); 371 372 static int pci_read(struct pci_bus *bus, unsigned int devfn, int where, 373 int size, u32 *val) 374 { 375 struct zpci_dev *zdev = get_zdev_by_bus(bus); 376 int ret; 377 378 if (!zdev || devfn != ZPCI_DEVFN) 379 ret = -ENODEV; 380 else 381 ret = zpci_cfg_load(zdev, where, val, size); 382 383 return ret; 384 } 385 386 static int pci_write(struct pci_bus *bus, unsigned int devfn, int where, 387 int size, u32 val) 388 { 389 struct zpci_dev *zdev = get_zdev_by_bus(bus); 390 int ret; 391 392 if (!zdev || devfn != ZPCI_DEVFN) 393 ret = -ENODEV; 394 else 395 ret = zpci_cfg_store(zdev, where, val, size); 396 397 return ret; 398 } 399 400 static struct pci_ops pci_root_ops = { 401 .read = pci_read, 402 .write = pci_write, 403 }; 404 405 #ifdef CONFIG_PCI_IOV 406 static struct resource iov_res = { 407 .name = "PCI IOV res", 408 .start = 0, 409 .end = -1, 410 .flags = IORESOURCE_MEM, 411 }; 412 #endif 413 414 static void zpci_map_resources(struct pci_dev *pdev) 415 { 416 struct zpci_dev *zdev = to_zpci(pdev); 417 resource_size_t len; 418 int i; 419 420 for (i = 0; i < PCI_STD_NUM_BARS; i++) { 421 len = pci_resource_len(pdev, i); 422 if (!len) 423 continue; 424 425 if (zpci_use_mio(zdev)) 426 pdev->resource[i].start = 427 (resource_size_t __force) zdev->bars[i].mio_wt; 428 else 429 pdev->resource[i].start = (resource_size_t __force) 430 pci_iomap_range_fh(pdev, i, 0, 0); 431 pdev->resource[i].end = pdev->resource[i].start + len - 1; 432 } 433 434 #ifdef CONFIG_PCI_IOV 435 for (i = 0; i < PCI_SRIOV_NUM_BARS; i++) { 436 int bar = i + PCI_IOV_RESOURCES; 437 438 len = pci_resource_len(pdev, bar); 439 if (!len) 440 continue; 441 pdev->resource[bar].parent = &iov_res; 442 } 443 #endif 444 } 445 446 static void zpci_unmap_resources(struct pci_dev *pdev) 447 { 448 struct zpci_dev *zdev = to_zpci(pdev); 449 resource_size_t len; 450 int i; 451 452 if (zpci_use_mio(zdev)) 453 return; 454 455 for (i = 0; i < PCI_STD_NUM_BARS; i++) { 456 len = pci_resource_len(pdev, i); 457 if (!len) 458 continue; 459 pci_iounmap_fh(pdev, (void __iomem __force *) 460 pdev->resource[i].start); 461 } 462 } 463 464 static int zpci_alloc_iomap(struct zpci_dev *zdev) 465 { 466 unsigned long entry; 467 468 spin_lock(&zpci_iomap_lock); 469 entry = find_first_zero_bit(zpci_iomap_bitmap, ZPCI_IOMAP_ENTRIES); 470 if (entry == ZPCI_IOMAP_ENTRIES) { 471 spin_unlock(&zpci_iomap_lock); 472 return -ENOSPC; 473 } 474 set_bit(entry, zpci_iomap_bitmap); 475 spin_unlock(&zpci_iomap_lock); 476 return entry; 477 } 478 479 static void zpci_free_iomap(struct zpci_dev *zdev, int entry) 480 { 481 spin_lock(&zpci_iomap_lock); 482 memset(&zpci_iomap_start[entry], 0, sizeof(struct zpci_iomap_entry)); 483 clear_bit(entry, zpci_iomap_bitmap); 484 spin_unlock(&zpci_iomap_lock); 485 } 486 487 static struct resource *__alloc_res(struct zpci_dev *zdev, unsigned long start, 488 unsigned long size, unsigned long flags) 489 { 490 struct resource *r; 491 492 r = kzalloc(sizeof(*r), GFP_KERNEL); 493 if (!r) 494 return NULL; 495 496 r->start = start; 497 r->end = r->start + size - 1; 498 r->flags = flags; 499 r->name = zdev->res_name; 500 501 if (request_resource(&iomem_resource, r)) { 502 kfree(r); 503 return NULL; 504 } 505 return r; 506 } 507 508 static int zpci_setup_bus_resources(struct zpci_dev *zdev, 509 struct list_head *resources) 510 { 511 unsigned long addr, size, flags; 512 struct resource *res; 513 int i, entry; 514 515 snprintf(zdev->res_name, sizeof(zdev->res_name), 516 "PCI Bus %04x:%02x", zdev->domain, ZPCI_BUS_NR); 517 518 for (i = 0; i < PCI_STD_NUM_BARS; i++) { 519 if (!zdev->bars[i].size) 520 continue; 521 entry = zpci_alloc_iomap(zdev); 522 if (entry < 0) 523 return entry; 524 zdev->bars[i].map_idx = entry; 525 526 /* only MMIO is supported */ 527 flags = IORESOURCE_MEM; 528 if (zdev->bars[i].val & 8) 529 flags |= IORESOURCE_PREFETCH; 530 if (zdev->bars[i].val & 4) 531 flags |= IORESOURCE_MEM_64; 532 533 if (zpci_use_mio(zdev)) 534 addr = (unsigned long) zdev->bars[i].mio_wt; 535 else 536 addr = ZPCI_ADDR(entry); 537 size = 1UL << zdev->bars[i].size; 538 539 res = __alloc_res(zdev, addr, size, flags); 540 if (!res) { 541 zpci_free_iomap(zdev, entry); 542 return -ENOMEM; 543 } 544 zdev->bars[i].res = res; 545 pci_add_resource(resources, res); 546 } 547 548 return 0; 549 } 550 551 static void zpci_cleanup_bus_resources(struct zpci_dev *zdev) 552 { 553 int i; 554 555 for (i = 0; i < PCI_STD_NUM_BARS; i++) { 556 if (!zdev->bars[i].size || !zdev->bars[i].res) 557 continue; 558 559 zpci_free_iomap(zdev, zdev->bars[i].map_idx); 560 release_resource(zdev->bars[i].res); 561 kfree(zdev->bars[i].res); 562 } 563 } 564 565 int pcibios_add_device(struct pci_dev *pdev) 566 { 567 struct resource *res; 568 int i; 569 570 if (pdev->is_physfn) 571 pdev->no_vf_scan = 1; 572 573 pdev->dev.groups = zpci_attr_groups; 574 pdev->dev.dma_ops = &s390_pci_dma_ops; 575 zpci_map_resources(pdev); 576 577 for (i = 0; i < PCI_STD_NUM_BARS; i++) { 578 res = &pdev->resource[i]; 579 if (res->parent || !res->flags) 580 continue; 581 pci_claim_resource(pdev, i); 582 } 583 584 return 0; 585 } 586 587 void pcibios_release_device(struct pci_dev *pdev) 588 { 589 zpci_unmap_resources(pdev); 590 } 591 592 int pcibios_enable_device(struct pci_dev *pdev, int mask) 593 { 594 struct zpci_dev *zdev = to_zpci(pdev); 595 596 zpci_debug_init_device(zdev, dev_name(&pdev->dev)); 597 zpci_fmb_enable_device(zdev); 598 599 return pci_enable_resources(pdev, mask); 600 } 601 602 void pcibios_disable_device(struct pci_dev *pdev) 603 { 604 struct zpci_dev *zdev = to_zpci(pdev); 605 606 zpci_fmb_disable_device(zdev); 607 zpci_debug_exit_device(zdev); 608 } 609 610 #ifdef CONFIG_HIBERNATE_CALLBACKS 611 static int zpci_restore(struct device *dev) 612 { 613 struct pci_dev *pdev = to_pci_dev(dev); 614 struct zpci_dev *zdev = to_zpci(pdev); 615 int ret = 0; 616 617 if (zdev->state != ZPCI_FN_STATE_ONLINE) 618 goto out; 619 620 ret = clp_enable_fh(zdev, ZPCI_NR_DMA_SPACES); 621 if (ret) 622 goto out; 623 624 zpci_map_resources(pdev); 625 zpci_register_ioat(zdev, 0, zdev->start_dma, zdev->end_dma, 626 (u64) zdev->dma_table); 627 628 out: 629 return ret; 630 } 631 632 static int zpci_freeze(struct device *dev) 633 { 634 struct pci_dev *pdev = to_pci_dev(dev); 635 struct zpci_dev *zdev = to_zpci(pdev); 636 637 if (zdev->state != ZPCI_FN_STATE_ONLINE) 638 return 0; 639 640 zpci_unregister_ioat(zdev, 0); 641 zpci_unmap_resources(pdev); 642 return clp_disable_fh(zdev); 643 } 644 645 struct dev_pm_ops pcibios_pm_ops = { 646 .thaw_noirq = zpci_restore, 647 .freeze_noirq = zpci_freeze, 648 .restore_noirq = zpci_restore, 649 .poweroff_noirq = zpci_freeze, 650 }; 651 #endif /* CONFIG_HIBERNATE_CALLBACKS */ 652 653 static int zpci_alloc_domain(struct zpci_dev *zdev) 654 { 655 if (zpci_unique_uid) { 656 zdev->domain = (u16) zdev->uid; 657 if (zdev->domain >= ZPCI_NR_DEVICES) 658 return 0; 659 660 spin_lock(&zpci_domain_lock); 661 if (test_bit(zdev->domain, zpci_domain)) { 662 spin_unlock(&zpci_domain_lock); 663 pr_err("Adding PCI function %08x failed because domain %04x is already assigned\n", 664 zdev->fid, zdev->domain); 665 return -EEXIST; 666 } 667 set_bit(zdev->domain, zpci_domain); 668 spin_unlock(&zpci_domain_lock); 669 return 0; 670 } 671 672 spin_lock(&zpci_domain_lock); 673 zdev->domain = find_first_zero_bit(zpci_domain, ZPCI_NR_DEVICES); 674 if (zdev->domain == ZPCI_NR_DEVICES) { 675 spin_unlock(&zpci_domain_lock); 676 pr_err("Adding PCI function %08x failed because the configured limit of %d is reached\n", 677 zdev->fid, ZPCI_NR_DEVICES); 678 return -ENOSPC; 679 } 680 set_bit(zdev->domain, zpci_domain); 681 spin_unlock(&zpci_domain_lock); 682 return 0; 683 } 684 685 static void zpci_free_domain(struct zpci_dev *zdev) 686 { 687 if (zdev->domain >= ZPCI_NR_DEVICES) 688 return; 689 690 spin_lock(&zpci_domain_lock); 691 clear_bit(zdev->domain, zpci_domain); 692 spin_unlock(&zpci_domain_lock); 693 } 694 695 void pcibios_remove_bus(struct pci_bus *bus) 696 { 697 struct zpci_dev *zdev = get_zdev_by_bus(bus); 698 699 zpci_exit_slot(zdev); 700 zpci_cleanup_bus_resources(zdev); 701 zpci_destroy_iommu(zdev); 702 zpci_free_domain(zdev); 703 704 spin_lock(&zpci_list_lock); 705 list_del(&zdev->entry); 706 spin_unlock(&zpci_list_lock); 707 708 zpci_dbg(3, "rem fid:%x\n", zdev->fid); 709 kfree(zdev); 710 } 711 712 static int zpci_scan_bus(struct zpci_dev *zdev) 713 { 714 LIST_HEAD(resources); 715 int ret; 716 717 ret = zpci_setup_bus_resources(zdev, &resources); 718 if (ret) 719 goto error; 720 721 zdev->bus = pci_scan_root_bus(NULL, ZPCI_BUS_NR, &pci_root_ops, 722 zdev, &resources); 723 if (!zdev->bus) { 724 ret = -EIO; 725 goto error; 726 } 727 zdev->bus->max_bus_speed = zdev->max_bus_speed; 728 pci_bus_add_devices(zdev->bus); 729 return 0; 730 731 error: 732 zpci_cleanup_bus_resources(zdev); 733 pci_free_resource_list(&resources); 734 return ret; 735 } 736 737 int zpci_enable_device(struct zpci_dev *zdev) 738 { 739 int rc; 740 741 rc = clp_enable_fh(zdev, ZPCI_NR_DMA_SPACES); 742 if (rc) 743 goto out; 744 745 rc = zpci_dma_init_device(zdev); 746 if (rc) 747 goto out_dma; 748 749 zdev->state = ZPCI_FN_STATE_ONLINE; 750 return 0; 751 752 out_dma: 753 clp_disable_fh(zdev); 754 out: 755 return rc; 756 } 757 EXPORT_SYMBOL_GPL(zpci_enable_device); 758 759 int zpci_disable_device(struct zpci_dev *zdev) 760 { 761 zpci_dma_exit_device(zdev); 762 return clp_disable_fh(zdev); 763 } 764 EXPORT_SYMBOL_GPL(zpci_disable_device); 765 766 int zpci_create_device(struct zpci_dev *zdev) 767 { 768 int rc; 769 770 rc = zpci_alloc_domain(zdev); 771 if (rc) 772 goto out; 773 774 rc = zpci_init_iommu(zdev); 775 if (rc) 776 goto out_free; 777 778 mutex_init(&zdev->lock); 779 if (zdev->state == ZPCI_FN_STATE_CONFIGURED) { 780 rc = zpci_enable_device(zdev); 781 if (rc) 782 goto out_destroy_iommu; 783 } 784 rc = zpci_scan_bus(zdev); 785 if (rc) 786 goto out_disable; 787 788 spin_lock(&zpci_list_lock); 789 list_add_tail(&zdev->entry, &zpci_list); 790 spin_unlock(&zpci_list_lock); 791 792 zpci_init_slot(zdev); 793 794 return 0; 795 796 out_disable: 797 if (zdev->state == ZPCI_FN_STATE_ONLINE) 798 zpci_disable_device(zdev); 799 out_destroy_iommu: 800 zpci_destroy_iommu(zdev); 801 out_free: 802 zpci_free_domain(zdev); 803 out: 804 return rc; 805 } 806 807 void zpci_remove_device(struct zpci_dev *zdev) 808 { 809 if (!zdev->bus) 810 return; 811 812 pci_stop_root_bus(zdev->bus); 813 pci_remove_root_bus(zdev->bus); 814 } 815 816 int zpci_report_error(struct pci_dev *pdev, 817 struct zpci_report_error_header *report) 818 { 819 struct zpci_dev *zdev = to_zpci(pdev); 820 821 return sclp_pci_report(report, zdev->fh, zdev->fid); 822 } 823 EXPORT_SYMBOL(zpci_report_error); 824 825 static int zpci_mem_init(void) 826 { 827 BUILD_BUG_ON(!is_power_of_2(__alignof__(struct zpci_fmb)) || 828 __alignof__(struct zpci_fmb) < sizeof(struct zpci_fmb)); 829 830 zdev_fmb_cache = kmem_cache_create("PCI_FMB_cache", sizeof(struct zpci_fmb), 831 __alignof__(struct zpci_fmb), 0, NULL); 832 if (!zdev_fmb_cache) 833 goto error_fmb; 834 835 zpci_iomap_start = kcalloc(ZPCI_IOMAP_ENTRIES, 836 sizeof(*zpci_iomap_start), GFP_KERNEL); 837 if (!zpci_iomap_start) 838 goto error_iomap; 839 840 zpci_iomap_bitmap = kcalloc(BITS_TO_LONGS(ZPCI_IOMAP_ENTRIES), 841 sizeof(*zpci_iomap_bitmap), GFP_KERNEL); 842 if (!zpci_iomap_bitmap) 843 goto error_iomap_bitmap; 844 845 return 0; 846 error_iomap_bitmap: 847 kfree(zpci_iomap_start); 848 error_iomap: 849 kmem_cache_destroy(zdev_fmb_cache); 850 error_fmb: 851 return -ENOMEM; 852 } 853 854 static void zpci_mem_exit(void) 855 { 856 kfree(zpci_iomap_bitmap); 857 kfree(zpci_iomap_start); 858 kmem_cache_destroy(zdev_fmb_cache); 859 } 860 861 static unsigned int s390_pci_probe __initdata = 1; 862 static unsigned int s390_pci_no_mio __initdata; 863 unsigned int s390_pci_force_floating __initdata; 864 static unsigned int s390_pci_initialized; 865 866 char * __init pcibios_setup(char *str) 867 { 868 if (!strcmp(str, "off")) { 869 s390_pci_probe = 0; 870 return NULL; 871 } 872 if (!strcmp(str, "nomio")) { 873 s390_pci_no_mio = 1; 874 return NULL; 875 } 876 if (!strcmp(str, "force_floating")) { 877 s390_pci_force_floating = 1; 878 return NULL; 879 } 880 return str; 881 } 882 883 bool zpci_is_enabled(void) 884 { 885 return s390_pci_initialized; 886 } 887 888 static int __init pci_base_init(void) 889 { 890 int rc; 891 892 if (!s390_pci_probe) 893 return 0; 894 895 if (!test_facility(69) || !test_facility(71)) 896 return 0; 897 898 if (test_facility(153) && !s390_pci_no_mio) { 899 static_branch_enable(&have_mio); 900 ctl_set_bit(2, 5); 901 } 902 903 rc = zpci_debug_init(); 904 if (rc) 905 goto out; 906 907 rc = zpci_mem_init(); 908 if (rc) 909 goto out_mem; 910 911 rc = zpci_irq_init(); 912 if (rc) 913 goto out_irq; 914 915 rc = zpci_dma_init(); 916 if (rc) 917 goto out_dma; 918 919 rc = clp_scan_pci_devices(); 920 if (rc) 921 goto out_find; 922 923 s390_pci_initialized = 1; 924 return 0; 925 926 out_find: 927 zpci_dma_exit(); 928 out_dma: 929 zpci_irq_exit(); 930 out_irq: 931 zpci_mem_exit(); 932 out_mem: 933 zpci_debug_exit(); 934 out: 935 return rc; 936 } 937 subsys_initcall_sync(pci_base_init); 938 939 void zpci_rescan(void) 940 { 941 if (zpci_is_enabled()) 942 clp_rescan_pci_devices_simple(NULL); 943 } 944