1 /* 2 * Copyright (c) 2004 Topspin Communications. All rights reserved. 3 * Copyright (c) 2005 Sun Microsystems, Inc. All rights reserved. 4 * 5 * This software is available to you under a choice of one of two 6 * licenses. You may choose to be licensed under the terms of the GNU 7 * General Public License (GPL) Version 2, available from the file 8 * COPYING in the main directory of this source tree, or the 9 * OpenIB.org BSD license below: 10 * 11 * Redistribution and use in source and binary forms, with or 12 * without modification, are permitted provided that the following 13 * conditions are met: 14 * 15 * - Redistributions of source code must retain the above 16 * copyright notice, this list of conditions and the following 17 * disclaimer. 18 * 19 * - Redistributions in binary form must reproduce the above 20 * copyright notice, this list of conditions and the following 21 * disclaimer in the documentation and/or other materials 22 * provided with the distribution. 23 * 24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 25 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 26 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 27 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 28 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 29 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 30 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 31 * SOFTWARE. 32 */ 33 34 #include <linux/module.h> 35 #include <linux/string.h> 36 #include <linux/errno.h> 37 #include <linux/kernel.h> 38 #include <linux/slab.h> 39 #include <linux/init.h> 40 #include <linux/mutex.h> 41 #include <linux/netdevice.h> 42 #include <rdma/rdma_netlink.h> 43 #include <rdma/ib_addr.h> 44 #include <rdma/ib_cache.h> 45 46 #include "core_priv.h" 47 48 MODULE_AUTHOR("Roland Dreier"); 49 MODULE_DESCRIPTION("core kernel InfiniBand API"); 50 MODULE_LICENSE("Dual BSD/GPL"); 51 52 struct ib_client_data { 53 struct list_head list; 54 struct ib_client *client; 55 void * data; 56 /* The device or client is going down. Do not call client or device 57 * callbacks other than remove(). */ 58 bool going_down; 59 }; 60 61 struct workqueue_struct *ib_wq; 62 EXPORT_SYMBOL_GPL(ib_wq); 63 64 /* The device_list and client_list contain devices and clients after their 65 * registration has completed, and the devices and clients are removed 66 * during unregistration. */ 67 static LIST_HEAD(device_list); 68 static LIST_HEAD(client_list); 69 70 /* 71 * device_mutex and lists_rwsem protect access to both device_list and 72 * client_list. device_mutex protects writer access by device and client 73 * registration / de-registration. lists_rwsem protects reader access to 74 * these lists. Iterators of these lists must lock it for read, while updates 75 * to the lists must be done with a write lock. A special case is when the 76 * device_mutex is locked. In this case locking the lists for read access is 77 * not necessary as the device_mutex implies it. 78 * 79 * lists_rwsem also protects access to the client data list. 80 */ 81 static DEFINE_MUTEX(device_mutex); 82 static DECLARE_RWSEM(lists_rwsem); 83 84 85 static int ib_device_check_mandatory(struct ib_device *device) 86 { 87 #define IB_MANDATORY_FUNC(x) { offsetof(struct ib_device, x), #x } 88 static const struct { 89 size_t offset; 90 char *name; 91 } mandatory_table[] = { 92 IB_MANDATORY_FUNC(query_device), 93 IB_MANDATORY_FUNC(query_port), 94 IB_MANDATORY_FUNC(query_pkey), 95 IB_MANDATORY_FUNC(query_gid), 96 IB_MANDATORY_FUNC(alloc_pd), 97 IB_MANDATORY_FUNC(dealloc_pd), 98 IB_MANDATORY_FUNC(create_ah), 99 IB_MANDATORY_FUNC(destroy_ah), 100 IB_MANDATORY_FUNC(create_qp), 101 IB_MANDATORY_FUNC(modify_qp), 102 IB_MANDATORY_FUNC(destroy_qp), 103 IB_MANDATORY_FUNC(post_send), 104 IB_MANDATORY_FUNC(post_recv), 105 IB_MANDATORY_FUNC(create_cq), 106 IB_MANDATORY_FUNC(destroy_cq), 107 IB_MANDATORY_FUNC(poll_cq), 108 IB_MANDATORY_FUNC(req_notify_cq), 109 IB_MANDATORY_FUNC(get_dma_mr), 110 IB_MANDATORY_FUNC(dereg_mr), 111 IB_MANDATORY_FUNC(get_port_immutable) 112 }; 113 int i; 114 115 for (i = 0; i < ARRAY_SIZE(mandatory_table); ++i) { 116 if (!*(void **) ((void *) device + mandatory_table[i].offset)) { 117 printk(KERN_WARNING "Device %s is missing mandatory function %s\n", 118 device->name, mandatory_table[i].name); 119 return -EINVAL; 120 } 121 } 122 123 return 0; 124 } 125 126 static struct ib_device *__ib_device_get_by_name(const char *name) 127 { 128 struct ib_device *device; 129 130 list_for_each_entry(device, &device_list, core_list) 131 if (!strncmp(name, device->name, IB_DEVICE_NAME_MAX)) 132 return device; 133 134 return NULL; 135 } 136 137 138 static int alloc_name(char *name) 139 { 140 unsigned long *inuse; 141 char buf[IB_DEVICE_NAME_MAX]; 142 struct ib_device *device; 143 int i; 144 145 inuse = (unsigned long *) get_zeroed_page(GFP_KERNEL); 146 if (!inuse) 147 return -ENOMEM; 148 149 list_for_each_entry(device, &device_list, core_list) { 150 if (!sscanf(device->name, name, &i)) 151 continue; 152 if (i < 0 || i >= PAGE_SIZE * 8) 153 continue; 154 snprintf(buf, sizeof buf, name, i); 155 if (!strncmp(buf, device->name, IB_DEVICE_NAME_MAX)) 156 set_bit(i, inuse); 157 } 158 159 i = find_first_zero_bit(inuse, PAGE_SIZE * 8); 160 free_page((unsigned long) inuse); 161 snprintf(buf, sizeof buf, name, i); 162 163 if (__ib_device_get_by_name(buf)) 164 return -ENFILE; 165 166 strlcpy(name, buf, IB_DEVICE_NAME_MAX); 167 return 0; 168 } 169 170 static void ib_device_release(struct device *device) 171 { 172 struct ib_device *dev = container_of(device, struct ib_device, dev); 173 174 ib_cache_release_one(dev); 175 kfree(dev->port_immutable); 176 kfree(dev); 177 } 178 179 static int ib_device_uevent(struct device *device, 180 struct kobj_uevent_env *env) 181 { 182 struct ib_device *dev = container_of(device, struct ib_device, dev); 183 184 if (add_uevent_var(env, "NAME=%s", dev->name)) 185 return -ENOMEM; 186 187 /* 188 * It would be nice to pass the node GUID with the event... 189 */ 190 191 return 0; 192 } 193 194 static struct class ib_class = { 195 .name = "infiniband", 196 .dev_release = ib_device_release, 197 .dev_uevent = ib_device_uevent, 198 }; 199 200 /** 201 * ib_alloc_device - allocate an IB device struct 202 * @size:size of structure to allocate 203 * 204 * Low-level drivers should use ib_alloc_device() to allocate &struct 205 * ib_device. @size is the size of the structure to be allocated, 206 * including any private data used by the low-level driver. 207 * ib_dealloc_device() must be used to free structures allocated with 208 * ib_alloc_device(). 209 */ 210 struct ib_device *ib_alloc_device(size_t size) 211 { 212 struct ib_device *device; 213 214 if (WARN_ON(size < sizeof(struct ib_device))) 215 return NULL; 216 217 device = kzalloc(size, GFP_KERNEL); 218 if (!device) 219 return NULL; 220 221 device->dev.class = &ib_class; 222 device_initialize(&device->dev); 223 224 dev_set_drvdata(&device->dev, device); 225 226 INIT_LIST_HEAD(&device->event_handler_list); 227 spin_lock_init(&device->event_handler_lock); 228 spin_lock_init(&device->client_data_lock); 229 INIT_LIST_HEAD(&device->client_data_list); 230 INIT_LIST_HEAD(&device->port_list); 231 232 return device; 233 } 234 EXPORT_SYMBOL(ib_alloc_device); 235 236 /** 237 * ib_dealloc_device - free an IB device struct 238 * @device:structure to free 239 * 240 * Free a structure allocated with ib_alloc_device(). 241 */ 242 void ib_dealloc_device(struct ib_device *device) 243 { 244 WARN_ON(device->reg_state != IB_DEV_UNREGISTERED && 245 device->reg_state != IB_DEV_UNINITIALIZED); 246 kobject_put(&device->dev.kobj); 247 } 248 EXPORT_SYMBOL(ib_dealloc_device); 249 250 static int add_client_context(struct ib_device *device, struct ib_client *client) 251 { 252 struct ib_client_data *context; 253 unsigned long flags; 254 255 context = kmalloc(sizeof *context, GFP_KERNEL); 256 if (!context) { 257 printk(KERN_WARNING "Couldn't allocate client context for %s/%s\n", 258 device->name, client->name); 259 return -ENOMEM; 260 } 261 262 context->client = client; 263 context->data = NULL; 264 context->going_down = false; 265 266 down_write(&lists_rwsem); 267 spin_lock_irqsave(&device->client_data_lock, flags); 268 list_add(&context->list, &device->client_data_list); 269 spin_unlock_irqrestore(&device->client_data_lock, flags); 270 up_write(&lists_rwsem); 271 272 return 0; 273 } 274 275 static int verify_immutable(const struct ib_device *dev, u8 port) 276 { 277 return WARN_ON(!rdma_cap_ib_mad(dev, port) && 278 rdma_max_mad_size(dev, port) != 0); 279 } 280 281 static int read_port_immutable(struct ib_device *device) 282 { 283 int ret; 284 u8 start_port = rdma_start_port(device); 285 u8 end_port = rdma_end_port(device); 286 u8 port; 287 288 /** 289 * device->port_immutable is indexed directly by the port number to make 290 * access to this data as efficient as possible. 291 * 292 * Therefore port_immutable is declared as a 1 based array with 293 * potential empty slots at the beginning. 294 */ 295 device->port_immutable = kzalloc(sizeof(*device->port_immutable) 296 * (end_port + 1), 297 GFP_KERNEL); 298 if (!device->port_immutable) 299 return -ENOMEM; 300 301 for (port = start_port; port <= end_port; ++port) { 302 ret = device->get_port_immutable(device, port, 303 &device->port_immutable[port]); 304 if (ret) 305 return ret; 306 307 if (verify_immutable(device, port)) 308 return -EINVAL; 309 } 310 return 0; 311 } 312 313 /** 314 * ib_register_device - Register an IB device with IB core 315 * @device:Device to register 316 * 317 * Low-level drivers use ib_register_device() to register their 318 * devices with the IB core. All registered clients will receive a 319 * callback for each device that is added. @device must be allocated 320 * with ib_alloc_device(). 321 */ 322 int ib_register_device(struct ib_device *device, 323 int (*port_callback)(struct ib_device *, 324 u8, struct kobject *)) 325 { 326 int ret; 327 struct ib_client *client; 328 329 mutex_lock(&device_mutex); 330 331 if (strchr(device->name, '%')) { 332 ret = alloc_name(device->name); 333 if (ret) 334 goto out; 335 } 336 337 if (ib_device_check_mandatory(device)) { 338 ret = -EINVAL; 339 goto out; 340 } 341 342 ret = read_port_immutable(device); 343 if (ret) { 344 printk(KERN_WARNING "Couldn't create per port immutable data %s\n", 345 device->name); 346 goto out; 347 } 348 349 ret = ib_cache_setup_one(device); 350 if (ret) { 351 printk(KERN_WARNING "Couldn't set up InfiniBand P_Key/GID cache\n"); 352 goto out; 353 } 354 355 ret = ib_device_register_sysfs(device, port_callback); 356 if (ret) { 357 printk(KERN_WARNING "Couldn't register device %s with driver model\n", 358 device->name); 359 ib_cache_cleanup_one(device); 360 goto out; 361 } 362 363 device->reg_state = IB_DEV_REGISTERED; 364 365 list_for_each_entry(client, &client_list, list) 366 if (client->add && !add_client_context(device, client)) 367 client->add(device); 368 369 down_write(&lists_rwsem); 370 list_add_tail(&device->core_list, &device_list); 371 up_write(&lists_rwsem); 372 out: 373 mutex_unlock(&device_mutex); 374 return ret; 375 } 376 EXPORT_SYMBOL(ib_register_device); 377 378 /** 379 * ib_unregister_device - Unregister an IB device 380 * @device:Device to unregister 381 * 382 * Unregister an IB device. All clients will receive a remove callback. 383 */ 384 void ib_unregister_device(struct ib_device *device) 385 { 386 struct ib_client_data *context, *tmp; 387 unsigned long flags; 388 389 mutex_lock(&device_mutex); 390 391 down_write(&lists_rwsem); 392 list_del(&device->core_list); 393 spin_lock_irqsave(&device->client_data_lock, flags); 394 list_for_each_entry_safe(context, tmp, &device->client_data_list, list) 395 context->going_down = true; 396 spin_unlock_irqrestore(&device->client_data_lock, flags); 397 downgrade_write(&lists_rwsem); 398 399 list_for_each_entry_safe(context, tmp, &device->client_data_list, 400 list) { 401 if (context->client->remove) 402 context->client->remove(device, context->data); 403 } 404 up_read(&lists_rwsem); 405 406 mutex_unlock(&device_mutex); 407 408 ib_device_unregister_sysfs(device); 409 ib_cache_cleanup_one(device); 410 411 down_write(&lists_rwsem); 412 spin_lock_irqsave(&device->client_data_lock, flags); 413 list_for_each_entry_safe(context, tmp, &device->client_data_list, list) 414 kfree(context); 415 spin_unlock_irqrestore(&device->client_data_lock, flags); 416 up_write(&lists_rwsem); 417 418 device->reg_state = IB_DEV_UNREGISTERED; 419 } 420 EXPORT_SYMBOL(ib_unregister_device); 421 422 /** 423 * ib_register_client - Register an IB client 424 * @client:Client to register 425 * 426 * Upper level users of the IB drivers can use ib_register_client() to 427 * register callbacks for IB device addition and removal. When an IB 428 * device is added, each registered client's add method will be called 429 * (in the order the clients were registered), and when a device is 430 * removed, each client's remove method will be called (in the reverse 431 * order that clients were registered). In addition, when 432 * ib_register_client() is called, the client will receive an add 433 * callback for all devices already registered. 434 */ 435 int ib_register_client(struct ib_client *client) 436 { 437 struct ib_device *device; 438 439 mutex_lock(&device_mutex); 440 441 list_for_each_entry(device, &device_list, core_list) 442 if (client->add && !add_client_context(device, client)) 443 client->add(device); 444 445 down_write(&lists_rwsem); 446 list_add_tail(&client->list, &client_list); 447 up_write(&lists_rwsem); 448 449 mutex_unlock(&device_mutex); 450 451 return 0; 452 } 453 EXPORT_SYMBOL(ib_register_client); 454 455 /** 456 * ib_unregister_client - Unregister an IB client 457 * @client:Client to unregister 458 * 459 * Upper level users use ib_unregister_client() to remove their client 460 * registration. When ib_unregister_client() is called, the client 461 * will receive a remove callback for each IB device still registered. 462 */ 463 void ib_unregister_client(struct ib_client *client) 464 { 465 struct ib_client_data *context, *tmp; 466 struct ib_device *device; 467 unsigned long flags; 468 469 mutex_lock(&device_mutex); 470 471 down_write(&lists_rwsem); 472 list_del(&client->list); 473 up_write(&lists_rwsem); 474 475 list_for_each_entry(device, &device_list, core_list) { 476 struct ib_client_data *found_context = NULL; 477 478 down_write(&lists_rwsem); 479 spin_lock_irqsave(&device->client_data_lock, flags); 480 list_for_each_entry_safe(context, tmp, &device->client_data_list, list) 481 if (context->client == client) { 482 context->going_down = true; 483 found_context = context; 484 break; 485 } 486 spin_unlock_irqrestore(&device->client_data_lock, flags); 487 up_write(&lists_rwsem); 488 489 if (client->remove) 490 client->remove(device, found_context ? 491 found_context->data : NULL); 492 493 if (!found_context) { 494 pr_warn("No client context found for %s/%s\n", 495 device->name, client->name); 496 continue; 497 } 498 499 down_write(&lists_rwsem); 500 spin_lock_irqsave(&device->client_data_lock, flags); 501 list_del(&found_context->list); 502 kfree(found_context); 503 spin_unlock_irqrestore(&device->client_data_lock, flags); 504 up_write(&lists_rwsem); 505 } 506 507 mutex_unlock(&device_mutex); 508 } 509 EXPORT_SYMBOL(ib_unregister_client); 510 511 /** 512 * ib_get_client_data - Get IB client context 513 * @device:Device to get context for 514 * @client:Client to get context for 515 * 516 * ib_get_client_data() returns client context set with 517 * ib_set_client_data(). 518 */ 519 void *ib_get_client_data(struct ib_device *device, struct ib_client *client) 520 { 521 struct ib_client_data *context; 522 void *ret = NULL; 523 unsigned long flags; 524 525 spin_lock_irqsave(&device->client_data_lock, flags); 526 list_for_each_entry(context, &device->client_data_list, list) 527 if (context->client == client) { 528 ret = context->data; 529 break; 530 } 531 spin_unlock_irqrestore(&device->client_data_lock, flags); 532 533 return ret; 534 } 535 EXPORT_SYMBOL(ib_get_client_data); 536 537 /** 538 * ib_set_client_data - Set IB client context 539 * @device:Device to set context for 540 * @client:Client to set context for 541 * @data:Context to set 542 * 543 * ib_set_client_data() sets client context that can be retrieved with 544 * ib_get_client_data(). 545 */ 546 void ib_set_client_data(struct ib_device *device, struct ib_client *client, 547 void *data) 548 { 549 struct ib_client_data *context; 550 unsigned long flags; 551 552 spin_lock_irqsave(&device->client_data_lock, flags); 553 list_for_each_entry(context, &device->client_data_list, list) 554 if (context->client == client) { 555 context->data = data; 556 goto out; 557 } 558 559 printk(KERN_WARNING "No client context found for %s/%s\n", 560 device->name, client->name); 561 562 out: 563 spin_unlock_irqrestore(&device->client_data_lock, flags); 564 } 565 EXPORT_SYMBOL(ib_set_client_data); 566 567 /** 568 * ib_register_event_handler - Register an IB event handler 569 * @event_handler:Handler to register 570 * 571 * ib_register_event_handler() registers an event handler that will be 572 * called back when asynchronous IB events occur (as defined in 573 * chapter 11 of the InfiniBand Architecture Specification). This 574 * callback may occur in interrupt context. 575 */ 576 int ib_register_event_handler (struct ib_event_handler *event_handler) 577 { 578 unsigned long flags; 579 580 spin_lock_irqsave(&event_handler->device->event_handler_lock, flags); 581 list_add_tail(&event_handler->list, 582 &event_handler->device->event_handler_list); 583 spin_unlock_irqrestore(&event_handler->device->event_handler_lock, flags); 584 585 return 0; 586 } 587 EXPORT_SYMBOL(ib_register_event_handler); 588 589 /** 590 * ib_unregister_event_handler - Unregister an event handler 591 * @event_handler:Handler to unregister 592 * 593 * Unregister an event handler registered with 594 * ib_register_event_handler(). 595 */ 596 int ib_unregister_event_handler(struct ib_event_handler *event_handler) 597 { 598 unsigned long flags; 599 600 spin_lock_irqsave(&event_handler->device->event_handler_lock, flags); 601 list_del(&event_handler->list); 602 spin_unlock_irqrestore(&event_handler->device->event_handler_lock, flags); 603 604 return 0; 605 } 606 EXPORT_SYMBOL(ib_unregister_event_handler); 607 608 /** 609 * ib_dispatch_event - Dispatch an asynchronous event 610 * @event:Event to dispatch 611 * 612 * Low-level drivers must call ib_dispatch_event() to dispatch the 613 * event to all registered event handlers when an asynchronous event 614 * occurs. 615 */ 616 void ib_dispatch_event(struct ib_event *event) 617 { 618 unsigned long flags; 619 struct ib_event_handler *handler; 620 621 spin_lock_irqsave(&event->device->event_handler_lock, flags); 622 623 list_for_each_entry(handler, &event->device->event_handler_list, list) 624 handler->handler(handler, event); 625 626 spin_unlock_irqrestore(&event->device->event_handler_lock, flags); 627 } 628 EXPORT_SYMBOL(ib_dispatch_event); 629 630 /** 631 * ib_query_device - Query IB device attributes 632 * @device:Device to query 633 * @device_attr:Device attributes 634 * 635 * ib_query_device() returns the attributes of a device through the 636 * @device_attr pointer. 637 */ 638 int ib_query_device(struct ib_device *device, 639 struct ib_device_attr *device_attr) 640 { 641 struct ib_udata uhw = {.outlen = 0, .inlen = 0}; 642 643 memset(device_attr, 0, sizeof(*device_attr)); 644 645 return device->query_device(device, device_attr, &uhw); 646 } 647 EXPORT_SYMBOL(ib_query_device); 648 649 /** 650 * ib_query_port - Query IB port attributes 651 * @device:Device to query 652 * @port_num:Port number to query 653 * @port_attr:Port attributes 654 * 655 * ib_query_port() returns the attributes of a port through the 656 * @port_attr pointer. 657 */ 658 int ib_query_port(struct ib_device *device, 659 u8 port_num, 660 struct ib_port_attr *port_attr) 661 { 662 if (port_num < rdma_start_port(device) || port_num > rdma_end_port(device)) 663 return -EINVAL; 664 665 return device->query_port(device, port_num, port_attr); 666 } 667 EXPORT_SYMBOL(ib_query_port); 668 669 /** 670 * ib_query_gid - Get GID table entry 671 * @device:Device to query 672 * @port_num:Port number to query 673 * @index:GID table index to query 674 * @gid:Returned GID 675 * 676 * ib_query_gid() fetches the specified GID table entry. 677 */ 678 int ib_query_gid(struct ib_device *device, 679 u8 port_num, int index, union ib_gid *gid) 680 { 681 if (rdma_cap_roce_gid_table(device, port_num)) 682 return ib_get_cached_gid(device, port_num, index, gid); 683 684 return device->query_gid(device, port_num, index, gid); 685 } 686 EXPORT_SYMBOL(ib_query_gid); 687 688 /** 689 * ib_enum_roce_netdev - enumerate all RoCE ports 690 * @ib_dev : IB device we want to query 691 * @filter: Should we call the callback? 692 * @filter_cookie: Cookie passed to filter 693 * @cb: Callback to call for each found RoCE ports 694 * @cookie: Cookie passed back to the callback 695 * 696 * Enumerates all of the physical RoCE ports of ib_dev 697 * which are related to netdevice and calls callback() on each 698 * device for which filter() function returns non zero. 699 */ 700 void ib_enum_roce_netdev(struct ib_device *ib_dev, 701 roce_netdev_filter filter, 702 void *filter_cookie, 703 roce_netdev_callback cb, 704 void *cookie) 705 { 706 u8 port; 707 708 for (port = rdma_start_port(ib_dev); port <= rdma_end_port(ib_dev); 709 port++) 710 if (rdma_protocol_roce(ib_dev, port)) { 711 struct net_device *idev = NULL; 712 713 if (ib_dev->get_netdev) 714 idev = ib_dev->get_netdev(ib_dev, port); 715 716 if (idev && 717 idev->reg_state >= NETREG_UNREGISTERED) { 718 dev_put(idev); 719 idev = NULL; 720 } 721 722 if (filter(ib_dev, port, idev, filter_cookie)) 723 cb(ib_dev, port, idev, cookie); 724 725 if (idev) 726 dev_put(idev); 727 } 728 } 729 730 /** 731 * ib_enum_all_roce_netdevs - enumerate all RoCE devices 732 * @filter: Should we call the callback? 733 * @filter_cookie: Cookie passed to filter 734 * @cb: Callback to call for each found RoCE ports 735 * @cookie: Cookie passed back to the callback 736 * 737 * Enumerates all RoCE devices' physical ports which are related 738 * to netdevices and calls callback() on each device for which 739 * filter() function returns non zero. 740 */ 741 void ib_enum_all_roce_netdevs(roce_netdev_filter filter, 742 void *filter_cookie, 743 roce_netdev_callback cb, 744 void *cookie) 745 { 746 struct ib_device *dev; 747 748 down_read(&lists_rwsem); 749 list_for_each_entry(dev, &device_list, core_list) 750 ib_enum_roce_netdev(dev, filter, filter_cookie, cb, cookie); 751 up_read(&lists_rwsem); 752 } 753 754 /** 755 * ib_query_pkey - Get P_Key table entry 756 * @device:Device to query 757 * @port_num:Port number to query 758 * @index:P_Key table index to query 759 * @pkey:Returned P_Key 760 * 761 * ib_query_pkey() fetches the specified P_Key table entry. 762 */ 763 int ib_query_pkey(struct ib_device *device, 764 u8 port_num, u16 index, u16 *pkey) 765 { 766 return device->query_pkey(device, port_num, index, pkey); 767 } 768 EXPORT_SYMBOL(ib_query_pkey); 769 770 /** 771 * ib_modify_device - Change IB device attributes 772 * @device:Device to modify 773 * @device_modify_mask:Mask of attributes to change 774 * @device_modify:New attribute values 775 * 776 * ib_modify_device() changes a device's attributes as specified by 777 * the @device_modify_mask and @device_modify structure. 778 */ 779 int ib_modify_device(struct ib_device *device, 780 int device_modify_mask, 781 struct ib_device_modify *device_modify) 782 { 783 if (!device->modify_device) 784 return -ENOSYS; 785 786 return device->modify_device(device, device_modify_mask, 787 device_modify); 788 } 789 EXPORT_SYMBOL(ib_modify_device); 790 791 /** 792 * ib_modify_port - Modifies the attributes for the specified port. 793 * @device: The device to modify. 794 * @port_num: The number of the port to modify. 795 * @port_modify_mask: Mask used to specify which attributes of the port 796 * to change. 797 * @port_modify: New attribute values for the port. 798 * 799 * ib_modify_port() changes a port's attributes as specified by the 800 * @port_modify_mask and @port_modify structure. 801 */ 802 int ib_modify_port(struct ib_device *device, 803 u8 port_num, int port_modify_mask, 804 struct ib_port_modify *port_modify) 805 { 806 if (!device->modify_port) 807 return -ENOSYS; 808 809 if (port_num < rdma_start_port(device) || port_num > rdma_end_port(device)) 810 return -EINVAL; 811 812 return device->modify_port(device, port_num, port_modify_mask, 813 port_modify); 814 } 815 EXPORT_SYMBOL(ib_modify_port); 816 817 /** 818 * ib_find_gid - Returns the port number and GID table index where 819 * a specified GID value occurs. 820 * @device: The device to query. 821 * @gid: The GID value to search for. 822 * @port_num: The port number of the device where the GID value was found. 823 * @index: The index into the GID table where the GID was found. This 824 * parameter may be NULL. 825 */ 826 int ib_find_gid(struct ib_device *device, union ib_gid *gid, 827 u8 *port_num, u16 *index) 828 { 829 union ib_gid tmp_gid; 830 int ret, port, i; 831 832 for (port = rdma_start_port(device); port <= rdma_end_port(device); ++port) { 833 if (rdma_cap_roce_gid_table(device, port)) { 834 if (!ib_cache_gid_find_by_port(device, gid, port, 835 NULL, index)) { 836 *port_num = port; 837 return 0; 838 } 839 } 840 841 for (i = 0; i < device->port_immutable[port].gid_tbl_len; ++i) { 842 ret = ib_query_gid(device, port, i, &tmp_gid); 843 if (ret) 844 return ret; 845 if (!memcmp(&tmp_gid, gid, sizeof *gid)) { 846 *port_num = port; 847 if (index) 848 *index = i; 849 return 0; 850 } 851 } 852 } 853 854 return -ENOENT; 855 } 856 EXPORT_SYMBOL(ib_find_gid); 857 858 /** 859 * ib_find_pkey - Returns the PKey table index where a specified 860 * PKey value occurs. 861 * @device: The device to query. 862 * @port_num: The port number of the device to search for the PKey. 863 * @pkey: The PKey value to search for. 864 * @index: The index into the PKey table where the PKey was found. 865 */ 866 int ib_find_pkey(struct ib_device *device, 867 u8 port_num, u16 pkey, u16 *index) 868 { 869 int ret, i; 870 u16 tmp_pkey; 871 int partial_ix = -1; 872 873 for (i = 0; i < device->port_immutable[port_num].pkey_tbl_len; ++i) { 874 ret = ib_query_pkey(device, port_num, i, &tmp_pkey); 875 if (ret) 876 return ret; 877 if ((pkey & 0x7fff) == (tmp_pkey & 0x7fff)) { 878 /* if there is full-member pkey take it.*/ 879 if (tmp_pkey & 0x8000) { 880 *index = i; 881 return 0; 882 } 883 if (partial_ix < 0) 884 partial_ix = i; 885 } 886 } 887 888 /*no full-member, if exists take the limited*/ 889 if (partial_ix >= 0) { 890 *index = partial_ix; 891 return 0; 892 } 893 return -ENOENT; 894 } 895 EXPORT_SYMBOL(ib_find_pkey); 896 897 /** 898 * ib_get_net_dev_by_params() - Return the appropriate net_dev 899 * for a received CM request 900 * @dev: An RDMA device on which the request has been received. 901 * @port: Port number on the RDMA device. 902 * @pkey: The Pkey the request came on. 903 * @gid: A GID that the net_dev uses to communicate. 904 * @addr: Contains the IP address that the request specified as its 905 * destination. 906 */ 907 struct net_device *ib_get_net_dev_by_params(struct ib_device *dev, 908 u8 port, 909 u16 pkey, 910 const union ib_gid *gid, 911 const struct sockaddr *addr) 912 { 913 struct net_device *net_dev = NULL; 914 struct ib_client_data *context; 915 916 if (!rdma_protocol_ib(dev, port)) 917 return NULL; 918 919 down_read(&lists_rwsem); 920 921 list_for_each_entry(context, &dev->client_data_list, list) { 922 struct ib_client *client = context->client; 923 924 if (context->going_down) 925 continue; 926 927 if (client->get_net_dev_by_params) { 928 net_dev = client->get_net_dev_by_params(dev, port, pkey, 929 gid, addr, 930 context->data); 931 if (net_dev) 932 break; 933 } 934 } 935 936 up_read(&lists_rwsem); 937 938 return net_dev; 939 } 940 EXPORT_SYMBOL(ib_get_net_dev_by_params); 941 942 static int __init ib_core_init(void) 943 { 944 int ret; 945 946 ib_wq = alloc_workqueue("infiniband", 0, 0); 947 if (!ib_wq) 948 return -ENOMEM; 949 950 ret = class_register(&ib_class); 951 if (ret) { 952 printk(KERN_WARNING "Couldn't create InfiniBand device class\n"); 953 goto err; 954 } 955 956 ret = ibnl_init(); 957 if (ret) { 958 printk(KERN_WARNING "Couldn't init IB netlink interface\n"); 959 goto err_sysfs; 960 } 961 962 ib_cache_setup(); 963 964 return 0; 965 966 err_sysfs: 967 class_unregister(&ib_class); 968 969 err: 970 destroy_workqueue(ib_wq); 971 return ret; 972 } 973 974 static void __exit ib_core_cleanup(void) 975 { 976 ib_cache_cleanup(); 977 ibnl_cleanup(); 978 class_unregister(&ib_class); 979 /* Make sure that any pending umem accounting work is done. */ 980 destroy_workqueue(ib_wq); 981 } 982 983 module_init(ib_core_init); 984 module_exit(ib_core_cleanup); 985