1 /* 2 * Copyright(c) 2007 - 2009 Intel Corporation. All rights reserved. 3 * 4 * This program is free software; you can redistribute it and/or modify it 5 * under the terms of the GNU General Public License as published by the Free 6 * Software Foundation; either version 2 of the License, or (at your option) 7 * any later version. 8 * 9 * This program is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for 12 * more details. 13 * 14 * You should have received a copy of the GNU General Public License along with 15 * this program; if not, write to the Free Software Foundation, Inc., 59 16 * Temple Place - Suite 330, Boston, MA 02111-1307, USA. 17 * 18 * The full GNU General Public License is included in this distribution in the 19 * file called COPYING. 20 */ 21 22 /* 23 * This driver supports an interface for DCA clients and providers to meet. 24 */ 25 26 #include <linux/kernel.h> 27 #include <linux/notifier.h> 28 #include <linux/device.h> 29 #include <linux/dca.h> 30 #include <linux/slab.h> 31 32 #define DCA_VERSION "1.12.1" 33 34 MODULE_VERSION(DCA_VERSION); 35 MODULE_LICENSE("GPL"); 36 MODULE_AUTHOR("Intel Corporation"); 37 38 static DEFINE_SPINLOCK(dca_lock); 39 40 static LIST_HEAD(dca_domains); 41 42 static BLOCKING_NOTIFIER_HEAD(dca_provider_chain); 43 44 static int dca_providers_blocked; 45 46 static struct pci_bus *dca_pci_rc_from_dev(struct device *dev) 47 { 48 struct pci_dev *pdev = to_pci_dev(dev); 49 struct pci_bus *bus = pdev->bus; 50 51 while (bus->parent) 52 bus = bus->parent; 53 54 return bus; 55 } 56 57 static struct dca_domain *dca_allocate_domain(struct pci_bus *rc) 58 { 59 struct dca_domain *domain; 60 61 domain = kzalloc(sizeof(*domain), GFP_NOWAIT); 62 if (!domain) 63 return NULL; 64 65 INIT_LIST_HEAD(&domain->dca_providers); 66 domain->pci_rc = rc; 67 68 return domain; 69 } 70 71 static void dca_free_domain(struct dca_domain *domain) 72 { 73 list_del(&domain->node); 74 kfree(domain); 75 } 76 77 static int dca_provider_ioat_ver_3_0(struct device *dev) 78 { 79 struct pci_dev *pdev = to_pci_dev(dev); 80 81 return ((pdev->vendor == PCI_VENDOR_ID_INTEL) && 82 ((pdev->device == PCI_DEVICE_ID_INTEL_IOAT_TBG0) || 83 (pdev->device == PCI_DEVICE_ID_INTEL_IOAT_TBG1) || 84 (pdev->device == PCI_DEVICE_ID_INTEL_IOAT_TBG2) || 85 (pdev->device == PCI_DEVICE_ID_INTEL_IOAT_TBG3) || 86 (pdev->device == PCI_DEVICE_ID_INTEL_IOAT_TBG4) || 87 (pdev->device == PCI_DEVICE_ID_INTEL_IOAT_TBG5) || 88 (pdev->device == PCI_DEVICE_ID_INTEL_IOAT_TBG6) || 89 (pdev->device == PCI_DEVICE_ID_INTEL_IOAT_TBG7))); 90 } 91 92 static void unregister_dca_providers(void) 93 { 94 struct dca_provider *dca, *_dca; 95 struct list_head unregistered_providers; 96 struct dca_domain *domain; 97 unsigned long flags; 98 99 blocking_notifier_call_chain(&dca_provider_chain, 100 DCA_PROVIDER_REMOVE, NULL); 101 102 INIT_LIST_HEAD(&unregistered_providers); 103 104 spin_lock_irqsave(&dca_lock, flags); 105 106 if (list_empty(&dca_domains)) { 107 spin_unlock_irqrestore(&dca_lock, flags); 108 return; 109 } 110 111 /* at this point only one domain in the list is expected */ 112 domain = list_first_entry(&dca_domains, struct dca_domain, node); 113 if (!domain) 114 return; 115 116 list_for_each_entry_safe(dca, _dca, &domain->dca_providers, node) { 117 list_del(&dca->node); 118 list_add(&dca->node, &unregistered_providers); 119 } 120 121 dca_free_domain(domain); 122 123 spin_unlock_irqrestore(&dca_lock, flags); 124 125 list_for_each_entry_safe(dca, _dca, &unregistered_providers, node) { 126 dca_sysfs_remove_provider(dca); 127 list_del(&dca->node); 128 } 129 } 130 131 static struct dca_domain *dca_find_domain(struct pci_bus *rc) 132 { 133 struct dca_domain *domain; 134 135 list_for_each_entry(domain, &dca_domains, node) 136 if (domain->pci_rc == rc) 137 return domain; 138 139 return NULL; 140 } 141 142 static struct dca_domain *dca_get_domain(struct device *dev) 143 { 144 struct pci_bus *rc; 145 struct dca_domain *domain; 146 147 rc = dca_pci_rc_from_dev(dev); 148 domain = dca_find_domain(rc); 149 150 if (!domain) { 151 if (dca_provider_ioat_ver_3_0(dev) && !list_empty(&dca_domains)) { 152 dca_providers_blocked = 1; 153 } else { 154 domain = dca_allocate_domain(rc); 155 if (domain) 156 list_add(&domain->node, &dca_domains); 157 } 158 } 159 160 return domain; 161 } 162 163 static struct dca_provider *dca_find_provider_by_dev(struct device *dev) 164 { 165 struct dca_provider *dca; 166 struct pci_bus *rc; 167 struct dca_domain *domain; 168 169 if (dev) { 170 rc = dca_pci_rc_from_dev(dev); 171 domain = dca_find_domain(rc); 172 if (!domain) 173 return NULL; 174 } else { 175 if (!list_empty(&dca_domains)) 176 domain = list_first_entry(&dca_domains, 177 struct dca_domain, 178 node); 179 else 180 return NULL; 181 } 182 183 list_for_each_entry(dca, &domain->dca_providers, node) 184 if ((!dev) || (dca->ops->dev_managed(dca, dev))) 185 return dca; 186 187 return NULL; 188 } 189 190 /** 191 * dca_add_requester - add a dca client to the list 192 * @dev - the device that wants dca service 193 */ 194 int dca_add_requester(struct device *dev) 195 { 196 struct dca_provider *dca; 197 int err, slot = -ENODEV; 198 unsigned long flags; 199 struct pci_bus *pci_rc; 200 struct dca_domain *domain; 201 202 if (!dev) 203 return -EFAULT; 204 205 spin_lock_irqsave(&dca_lock, flags); 206 207 /* check if the requester has not been added already */ 208 dca = dca_find_provider_by_dev(dev); 209 if (dca) { 210 spin_unlock_irqrestore(&dca_lock, flags); 211 return -EEXIST; 212 } 213 214 pci_rc = dca_pci_rc_from_dev(dev); 215 domain = dca_find_domain(pci_rc); 216 if (!domain) { 217 spin_unlock_irqrestore(&dca_lock, flags); 218 return -ENODEV; 219 } 220 221 list_for_each_entry(dca, &domain->dca_providers, node) { 222 slot = dca->ops->add_requester(dca, dev); 223 if (slot >= 0) 224 break; 225 } 226 227 spin_unlock_irqrestore(&dca_lock, flags); 228 229 if (slot < 0) 230 return slot; 231 232 err = dca_sysfs_add_req(dca, dev, slot); 233 if (err) { 234 spin_lock_irqsave(&dca_lock, flags); 235 if (dca == dca_find_provider_by_dev(dev)) 236 dca->ops->remove_requester(dca, dev); 237 spin_unlock_irqrestore(&dca_lock, flags); 238 return err; 239 } 240 241 return 0; 242 } 243 EXPORT_SYMBOL_GPL(dca_add_requester); 244 245 /** 246 * dca_remove_requester - remove a dca client from the list 247 * @dev - the device that wants dca service 248 */ 249 int dca_remove_requester(struct device *dev) 250 { 251 struct dca_provider *dca; 252 int slot; 253 unsigned long flags; 254 255 if (!dev) 256 return -EFAULT; 257 258 spin_lock_irqsave(&dca_lock, flags); 259 dca = dca_find_provider_by_dev(dev); 260 if (!dca) { 261 spin_unlock_irqrestore(&dca_lock, flags); 262 return -ENODEV; 263 } 264 slot = dca->ops->remove_requester(dca, dev); 265 spin_unlock_irqrestore(&dca_lock, flags); 266 267 if (slot < 0) 268 return slot; 269 270 dca_sysfs_remove_req(dca, slot); 271 272 return 0; 273 } 274 EXPORT_SYMBOL_GPL(dca_remove_requester); 275 276 /** 277 * dca_common_get_tag - return the dca tag (serves both new and old api) 278 * @dev - the device that wants dca service 279 * @cpu - the cpuid as returned by get_cpu() 280 */ 281 u8 dca_common_get_tag(struct device *dev, int cpu) 282 { 283 struct dca_provider *dca; 284 u8 tag; 285 unsigned long flags; 286 287 spin_lock_irqsave(&dca_lock, flags); 288 289 dca = dca_find_provider_by_dev(dev); 290 if (!dca) { 291 spin_unlock_irqrestore(&dca_lock, flags); 292 return -ENODEV; 293 } 294 tag = dca->ops->get_tag(dca, dev, cpu); 295 296 spin_unlock_irqrestore(&dca_lock, flags); 297 return tag; 298 } 299 300 /** 301 * dca3_get_tag - return the dca tag to the requester device 302 * for the given cpu (new api) 303 * @dev - the device that wants dca service 304 * @cpu - the cpuid as returned by get_cpu() 305 */ 306 u8 dca3_get_tag(struct device *dev, int cpu) 307 { 308 if (!dev) 309 return -EFAULT; 310 311 return dca_common_get_tag(dev, cpu); 312 } 313 EXPORT_SYMBOL_GPL(dca3_get_tag); 314 315 /** 316 * dca_get_tag - return the dca tag for the given cpu (old api) 317 * @cpu - the cpuid as returned by get_cpu() 318 */ 319 u8 dca_get_tag(int cpu) 320 { 321 struct device *dev = NULL; 322 323 return dca_common_get_tag(dev, cpu); 324 } 325 EXPORT_SYMBOL_GPL(dca_get_tag); 326 327 /** 328 * alloc_dca_provider - get data struct for describing a dca provider 329 * @ops - pointer to struct of dca operation function pointers 330 * @priv_size - size of extra mem to be added for provider's needs 331 */ 332 struct dca_provider *alloc_dca_provider(struct dca_ops *ops, int priv_size) 333 { 334 struct dca_provider *dca; 335 int alloc_size; 336 337 alloc_size = (sizeof(*dca) + priv_size); 338 dca = kzalloc(alloc_size, GFP_KERNEL); 339 if (!dca) 340 return NULL; 341 dca->ops = ops; 342 343 return dca; 344 } 345 EXPORT_SYMBOL_GPL(alloc_dca_provider); 346 347 /** 348 * free_dca_provider - release the dca provider data struct 349 * @ops - pointer to struct of dca operation function pointers 350 * @priv_size - size of extra mem to be added for provider's needs 351 */ 352 void free_dca_provider(struct dca_provider *dca) 353 { 354 kfree(dca); 355 } 356 EXPORT_SYMBOL_GPL(free_dca_provider); 357 358 /** 359 * register_dca_provider - register a dca provider 360 * @dca - struct created by alloc_dca_provider() 361 * @dev - device providing dca services 362 */ 363 int register_dca_provider(struct dca_provider *dca, struct device *dev) 364 { 365 int err; 366 unsigned long flags; 367 struct dca_domain *domain; 368 369 spin_lock_irqsave(&dca_lock, flags); 370 if (dca_providers_blocked) { 371 spin_unlock_irqrestore(&dca_lock, flags); 372 return -ENODEV; 373 } 374 spin_unlock_irqrestore(&dca_lock, flags); 375 376 err = dca_sysfs_add_provider(dca, dev); 377 if (err) 378 return err; 379 380 spin_lock_irqsave(&dca_lock, flags); 381 domain = dca_get_domain(dev); 382 if (!domain) { 383 if (dca_providers_blocked) { 384 spin_unlock_irqrestore(&dca_lock, flags); 385 dca_sysfs_remove_provider(dca); 386 unregister_dca_providers(); 387 } else { 388 spin_unlock_irqrestore(&dca_lock, flags); 389 } 390 return -ENODEV; 391 } 392 list_add(&dca->node, &domain->dca_providers); 393 spin_unlock_irqrestore(&dca_lock, flags); 394 395 blocking_notifier_call_chain(&dca_provider_chain, 396 DCA_PROVIDER_ADD, NULL); 397 return 0; 398 } 399 EXPORT_SYMBOL_GPL(register_dca_provider); 400 401 /** 402 * unregister_dca_provider - remove a dca provider 403 * @dca - struct created by alloc_dca_provider() 404 */ 405 void unregister_dca_provider(struct dca_provider *dca, struct device *dev) 406 { 407 unsigned long flags; 408 struct pci_bus *pci_rc; 409 struct dca_domain *domain; 410 411 blocking_notifier_call_chain(&dca_provider_chain, 412 DCA_PROVIDER_REMOVE, NULL); 413 414 spin_lock_irqsave(&dca_lock, flags); 415 416 list_del(&dca->node); 417 418 pci_rc = dca_pci_rc_from_dev(dev); 419 domain = dca_find_domain(pci_rc); 420 if (list_empty(&domain->dca_providers)) 421 dca_free_domain(domain); 422 423 spin_unlock_irqrestore(&dca_lock, flags); 424 425 dca_sysfs_remove_provider(dca); 426 } 427 EXPORT_SYMBOL_GPL(unregister_dca_provider); 428 429 /** 430 * dca_register_notify - register a client's notifier callback 431 */ 432 void dca_register_notify(struct notifier_block *nb) 433 { 434 blocking_notifier_chain_register(&dca_provider_chain, nb); 435 } 436 EXPORT_SYMBOL_GPL(dca_register_notify); 437 438 /** 439 * dca_unregister_notify - remove a client's notifier callback 440 */ 441 void dca_unregister_notify(struct notifier_block *nb) 442 { 443 blocking_notifier_chain_unregister(&dca_provider_chain, nb); 444 } 445 EXPORT_SYMBOL_GPL(dca_unregister_notify); 446 447 static int __init dca_init(void) 448 { 449 pr_info("dca service started, version %s\n", DCA_VERSION); 450 return dca_sysfs_init(); 451 } 452 453 static void __exit dca_exit(void) 454 { 455 dca_sysfs_exit(); 456 } 457 458 arch_initcall(dca_init); 459 module_exit(dca_exit); 460 461