1 /* 2 * Low-Level PCI Access for i386 machines 3 * 4 * Copyright 1993, 1994 Drew Eckhardt 5 * Visionary Computing 6 * (Unix and Linux consulting and custom programming) 7 * Drew@Colorado.EDU 8 * +1 (303) 786-7975 9 * 10 * Drew's work was sponsored by: 11 * iX Multiuser Multitasking Magazine 12 * Hannover, Germany 13 * hm@ix.de 14 * 15 * Copyright 1997--2000 Martin Mares <mj@ucw.cz> 16 * 17 * For more information, please consult the following manuals (look at 18 * http://www.pcisig.com/ for how to get them): 19 * 20 * PCI BIOS Specification 21 * PCI Local Bus Specification 22 * PCI to PCI Bridge Specification 23 * PCI System Design Guide 24 * 25 */ 26 27 #include <linux/types.h> 28 #include <linux/kernel.h> 29 #include <linux/export.h> 30 #include <linux/pci.h> 31 #include <linux/init.h> 32 #include <linux/ioport.h> 33 #include <linux/errno.h> 34 #include <linux/bootmem.h> 35 36 #include <asm/pat.h> 37 #include <asm/e820.h> 38 #include <asm/pci_x86.h> 39 #include <asm/io_apic.h> 40 41 42 /* 43 * This list of dynamic mappings is for temporarily maintaining 44 * original BIOS BAR addresses for possible reinstatement. 45 */ 46 struct pcibios_fwaddrmap { 47 struct list_head list; 48 struct pci_dev *dev; 49 resource_size_t fw_addr[DEVICE_COUNT_RESOURCE]; 50 }; 51 52 static LIST_HEAD(pcibios_fwaddrmappings); 53 static DEFINE_SPINLOCK(pcibios_fwaddrmap_lock); 54 static bool pcibios_fw_addr_done; 55 56 /* Must be called with 'pcibios_fwaddrmap_lock' lock held. */ 57 static struct pcibios_fwaddrmap *pcibios_fwaddrmap_lookup(struct pci_dev *dev) 58 { 59 struct pcibios_fwaddrmap *map; 60 61 WARN_ON_SMP(!spin_is_locked(&pcibios_fwaddrmap_lock)); 62 63 list_for_each_entry(map, &pcibios_fwaddrmappings, list) 64 if (map->dev == dev) 65 return map; 66 67 return NULL; 68 } 69 70 static void 71 pcibios_save_fw_addr(struct pci_dev *dev, int idx, resource_size_t fw_addr) 72 { 73 unsigned long flags; 74 struct pcibios_fwaddrmap *map; 75 76 if (pcibios_fw_addr_done) 77 return; 78 79 spin_lock_irqsave(&pcibios_fwaddrmap_lock, flags); 80 map = pcibios_fwaddrmap_lookup(dev); 81 if (!map) { 82 spin_unlock_irqrestore(&pcibios_fwaddrmap_lock, flags); 83 map = kzalloc(sizeof(*map), GFP_KERNEL); 84 if (!map) 85 return; 86 87 map->dev = pci_dev_get(dev); 88 map->fw_addr[idx] = fw_addr; 89 INIT_LIST_HEAD(&map->list); 90 91 spin_lock_irqsave(&pcibios_fwaddrmap_lock, flags); 92 list_add_tail(&map->list, &pcibios_fwaddrmappings); 93 } else 94 map->fw_addr[idx] = fw_addr; 95 spin_unlock_irqrestore(&pcibios_fwaddrmap_lock, flags); 96 } 97 98 resource_size_t pcibios_retrieve_fw_addr(struct pci_dev *dev, int idx) 99 { 100 unsigned long flags; 101 struct pcibios_fwaddrmap *map; 102 resource_size_t fw_addr = 0; 103 104 if (pcibios_fw_addr_done) 105 return 0; 106 107 spin_lock_irqsave(&pcibios_fwaddrmap_lock, flags); 108 map = pcibios_fwaddrmap_lookup(dev); 109 if (map) 110 fw_addr = map->fw_addr[idx]; 111 spin_unlock_irqrestore(&pcibios_fwaddrmap_lock, flags); 112 113 return fw_addr; 114 } 115 116 static void __init pcibios_fw_addr_list_del(void) 117 { 118 unsigned long flags; 119 struct pcibios_fwaddrmap *entry, *next; 120 121 spin_lock_irqsave(&pcibios_fwaddrmap_lock, flags); 122 list_for_each_entry_safe(entry, next, &pcibios_fwaddrmappings, list) { 123 list_del(&entry->list); 124 pci_dev_put(entry->dev); 125 kfree(entry); 126 } 127 spin_unlock_irqrestore(&pcibios_fwaddrmap_lock, flags); 128 pcibios_fw_addr_done = true; 129 } 130 131 static int 132 skip_isa_ioresource_align(struct pci_dev *dev) { 133 134 if ((pci_probe & PCI_CAN_SKIP_ISA_ALIGN) && 135 !(dev->bus->bridge_ctl & PCI_BRIDGE_CTL_ISA)) 136 return 1; 137 return 0; 138 } 139 140 /* 141 * We need to avoid collisions with `mirrored' VGA ports 142 * and other strange ISA hardware, so we always want the 143 * addresses to be allocated in the 0x000-0x0ff region 144 * modulo 0x400. 145 * 146 * Why? Because some silly external IO cards only decode 147 * the low 10 bits of the IO address. The 0x00-0xff region 148 * is reserved for motherboard devices that decode all 16 149 * bits, so it's ok to allocate at, say, 0x2800-0x28ff, 150 * but we want to try to avoid allocating at 0x2900-0x2bff 151 * which might have be mirrored at 0x0100-0x03ff.. 152 */ 153 resource_size_t 154 pcibios_align_resource(void *data, const struct resource *res, 155 resource_size_t size, resource_size_t align) 156 { 157 struct pci_dev *dev = data; 158 resource_size_t start = res->start; 159 160 if (res->flags & IORESOURCE_IO) { 161 if (skip_isa_ioresource_align(dev)) 162 return start; 163 if (start & 0x300) 164 start = (start + 0x3ff) & ~0x3ff; 165 } 166 return start; 167 } 168 EXPORT_SYMBOL(pcibios_align_resource); 169 170 /* 171 * Handle resources of PCI devices. If the world were perfect, we could 172 * just allocate all the resource regions and do nothing more. It isn't. 173 * On the other hand, we cannot just re-allocate all devices, as it would 174 * require us to know lots of host bridge internals. So we attempt to 175 * keep as much of the original configuration as possible, but tweak it 176 * when it's found to be wrong. 177 * 178 * Known BIOS problems we have to work around: 179 * - I/O or memory regions not configured 180 * - regions configured, but not enabled in the command register 181 * - bogus I/O addresses above 64K used 182 * - expansion ROMs left enabled (this may sound harmless, but given 183 * the fact the PCI specs explicitly allow address decoders to be 184 * shared between expansion ROMs and other resource regions, it's 185 * at least dangerous) 186 * - bad resource sizes or overlaps with other regions 187 * 188 * Our solution: 189 * (1) Allocate resources for all buses behind PCI-to-PCI bridges. 190 * This gives us fixed barriers on where we can allocate. 191 * (2) Allocate resources for all enabled devices. If there is 192 * a collision, just mark the resource as unallocated. Also 193 * disable expansion ROMs during this step. 194 * (3) Try to allocate resources for disabled devices. If the 195 * resources were assigned correctly, everything goes well, 196 * if they weren't, they won't disturb allocation of other 197 * resources. 198 * (4) Assign new addresses to resources which were either 199 * not configured at all or misconfigured. If explicitly 200 * requested by the user, configure expansion ROM address 201 * as well. 202 */ 203 204 static void pcibios_allocate_bridge_resources(struct pci_dev *dev) 205 { 206 int idx; 207 struct resource *r; 208 209 for (idx = PCI_BRIDGE_RESOURCES; idx < PCI_NUM_RESOURCES; idx++) { 210 r = &dev->resource[idx]; 211 if (!r->flags) 212 continue; 213 if (!r->start || pci_claim_resource(dev, idx) < 0) { 214 /* 215 * Something is wrong with the region. 216 * Invalidate the resource to prevent 217 * child resource allocations in this 218 * range. 219 */ 220 r->start = r->end = 0; 221 r->flags = 0; 222 } 223 } 224 } 225 226 static void pcibios_allocate_bus_resources(struct pci_bus *bus) 227 { 228 struct pci_bus *child; 229 230 /* Depth-First Search on bus tree */ 231 if (bus->self) 232 pcibios_allocate_bridge_resources(bus->self); 233 list_for_each_entry(child, &bus->children, node) 234 pcibios_allocate_bus_resources(child); 235 } 236 237 struct pci_check_idx_range { 238 int start; 239 int end; 240 }; 241 242 static void pcibios_allocate_dev_resources(struct pci_dev *dev, int pass) 243 { 244 int idx, disabled, i; 245 u16 command; 246 struct resource *r; 247 248 struct pci_check_idx_range idx_range[] = { 249 { PCI_STD_RESOURCES, PCI_STD_RESOURCE_END }, 250 #ifdef CONFIG_PCI_IOV 251 { PCI_IOV_RESOURCES, PCI_IOV_RESOURCE_END }, 252 #endif 253 }; 254 255 pci_read_config_word(dev, PCI_COMMAND, &command); 256 for (i = 0; i < ARRAY_SIZE(idx_range); i++) 257 for (idx = idx_range[i].start; idx <= idx_range[i].end; idx++) { 258 r = &dev->resource[idx]; 259 if (r->parent) /* Already allocated */ 260 continue; 261 if (!r->start) /* Address not assigned at all */ 262 continue; 263 if (r->flags & IORESOURCE_IO) 264 disabled = !(command & PCI_COMMAND_IO); 265 else 266 disabled = !(command & PCI_COMMAND_MEMORY); 267 if (pass == disabled) { 268 dev_dbg(&dev->dev, 269 "BAR %d: reserving %pr (d=%d, p=%d)\n", 270 idx, r, disabled, pass); 271 if (pci_claim_resource(dev, idx) < 0) { 272 /* We'll assign a new address later */ 273 pcibios_save_fw_addr(dev, 274 idx, r->start); 275 r->end -= r->start; 276 r->start = 0; 277 } 278 } 279 } 280 if (!pass) { 281 r = &dev->resource[PCI_ROM_RESOURCE]; 282 if (r->flags & IORESOURCE_ROM_ENABLE) { 283 /* Turn the ROM off, leave the resource region, 284 * but keep it unregistered. */ 285 u32 reg; 286 dev_dbg(&dev->dev, "disabling ROM %pR\n", r); 287 r->flags &= ~IORESOURCE_ROM_ENABLE; 288 pci_read_config_dword(dev, dev->rom_base_reg, ®); 289 pci_write_config_dword(dev, dev->rom_base_reg, 290 reg & ~PCI_ROM_ADDRESS_ENABLE); 291 } 292 } 293 } 294 295 static void pcibios_allocate_resources(struct pci_bus *bus, int pass) 296 { 297 struct pci_dev *dev; 298 struct pci_bus *child; 299 300 list_for_each_entry(dev, &bus->devices, bus_list) { 301 pcibios_allocate_dev_resources(dev, pass); 302 303 child = dev->subordinate; 304 if (child) 305 pcibios_allocate_resources(child, pass); 306 } 307 } 308 309 static void pcibios_allocate_dev_rom_resource(struct pci_dev *dev) 310 { 311 struct resource *r; 312 313 /* 314 * Try to use BIOS settings for ROMs, otherwise let 315 * pci_assign_unassigned_resources() allocate the new 316 * addresses. 317 */ 318 r = &dev->resource[PCI_ROM_RESOURCE]; 319 if (!r->flags || !r->start) 320 return; 321 322 if (pci_claim_resource(dev, PCI_ROM_RESOURCE) < 0) { 323 r->end -= r->start; 324 r->start = 0; 325 } 326 } 327 static void pcibios_allocate_rom_resources(struct pci_bus *bus) 328 { 329 struct pci_dev *dev; 330 struct pci_bus *child; 331 332 list_for_each_entry(dev, &bus->devices, bus_list) { 333 pcibios_allocate_dev_rom_resource(dev); 334 335 child = dev->subordinate; 336 if (child) 337 pcibios_allocate_rom_resources(child); 338 } 339 } 340 341 static int __init pcibios_assign_resources(void) 342 { 343 struct pci_bus *bus; 344 345 if (!(pci_probe & PCI_ASSIGN_ROMS)) 346 list_for_each_entry(bus, &pci_root_buses, node) 347 pcibios_allocate_rom_resources(bus); 348 349 pci_assign_unassigned_resources(); 350 pcibios_fw_addr_list_del(); 351 352 return 0; 353 } 354 355 void pcibios_resource_survey_bus(struct pci_bus *bus) 356 { 357 dev_printk(KERN_DEBUG, &bus->dev, "Allocating resources\n"); 358 359 pcibios_allocate_bus_resources(bus); 360 361 pcibios_allocate_resources(bus, 0); 362 pcibios_allocate_resources(bus, 1); 363 364 if (!(pci_probe & PCI_ASSIGN_ROMS)) 365 pcibios_allocate_rom_resources(bus); 366 } 367 368 void __init pcibios_resource_survey(void) 369 { 370 struct pci_bus *bus; 371 372 DBG("PCI: Allocating resources\n"); 373 374 list_for_each_entry(bus, &pci_root_buses, node) 375 pcibios_allocate_bus_resources(bus); 376 377 list_for_each_entry(bus, &pci_root_buses, node) 378 pcibios_allocate_resources(bus, 0); 379 list_for_each_entry(bus, &pci_root_buses, node) 380 pcibios_allocate_resources(bus, 1); 381 382 e820_reserve_resources_late(); 383 /* 384 * Insert the IO APIC resources after PCI initialization has 385 * occurred to handle IO APICS that are mapped in on a BAR in 386 * PCI space, but before trying to assign unassigned pci res. 387 */ 388 ioapic_insert_resources(); 389 } 390 391 /** 392 * called in fs_initcall (one below subsys_initcall), 393 * give a chance for motherboard reserve resources 394 */ 395 fs_initcall(pcibios_assign_resources); 396 397 static const struct vm_operations_struct pci_mmap_ops = { 398 .access = generic_access_phys, 399 }; 400 401 int pci_mmap_page_range(struct pci_dev *dev, struct vm_area_struct *vma, 402 enum pci_mmap_state mmap_state, int write_combine) 403 { 404 unsigned long prot; 405 406 /* I/O space cannot be accessed via normal processor loads and 407 * stores on this platform. 408 */ 409 if (mmap_state == pci_mmap_io) 410 return -EINVAL; 411 412 prot = pgprot_val(vma->vm_page_prot); 413 414 /* 415 * Return error if pat is not enabled and write_combine is requested. 416 * Caller can followup with UC MINUS request and add a WC mtrr if there 417 * is a free mtrr slot. 418 */ 419 if (!pat_enabled && write_combine) 420 return -EINVAL; 421 422 if (pat_enabled && write_combine) 423 prot |= _PAGE_CACHE_WC; 424 else if (pat_enabled || boot_cpu_data.x86 > 3) 425 /* 426 * ioremap() and ioremap_nocache() defaults to UC MINUS for now. 427 * To avoid attribute conflicts, request UC MINUS here 428 * as well. 429 */ 430 prot |= _PAGE_CACHE_UC_MINUS; 431 432 prot |= _PAGE_IOMAP; /* creating a mapping for IO */ 433 434 vma->vm_page_prot = __pgprot(prot); 435 436 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, 437 vma->vm_end - vma->vm_start, 438 vma->vm_page_prot)) 439 return -EAGAIN; 440 441 vma->vm_ops = &pci_mmap_ops; 442 443 return 0; 444 } 445