1 /* 2 * Copyright (c) 2016 Hisilicon Limited. 3 * 4 * This software is available to you under a choice of one of two 5 * licenses. You may choose to be licensed under the terms of the GNU 6 * General Public License (GPL) Version 2, available from the file 7 * COPYING in the main directory of this source tree, or the 8 * OpenIB.org BSD license below: 9 * 10 * Redistribution and use in source and binary forms, with or 11 * without modification, are permitted provided that the following 12 * conditions are met: 13 * 14 * - Redistributions of source code must retain the above 15 * copyright notice, this list of conditions and the following 16 * disclaimer. 17 * 18 * - Redistributions in binary form must reproduce the above 19 * copyright notice, this list of conditions and the following 20 * disclaimer in the documentation and/or other materials 21 * provided with the distribution. 22 * 23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 30 * SOFTWARE. 31 */ 32 33 #include <linux/platform_device.h> 34 #include <linux/pci.h> 35 #include "hns_roce_device.h" 36 37 void hns_roce_init_pd_table(struct hns_roce_dev *hr_dev) 38 { 39 struct hns_roce_ida *pd_ida = &hr_dev->pd_ida; 40 41 ida_init(&pd_ida->ida); 42 pd_ida->max = hr_dev->caps.num_pds - 1; 43 pd_ida->min = hr_dev->caps.reserved_pds; 44 } 45 46 int hns_roce_alloc_pd(struct ib_pd *ibpd, struct ib_udata *udata) 47 { 48 struct ib_device *ib_dev = ibpd->device; 49 struct hns_roce_dev *hr_dev = to_hr_dev(ib_dev); 50 struct hns_roce_ida *pd_ida = &hr_dev->pd_ida; 51 struct hns_roce_pd *pd = to_hr_pd(ibpd); 52 int ret = 0; 53 int id; 54 55 id = ida_alloc_range(&pd_ida->ida, pd_ida->min, pd_ida->max, 56 GFP_KERNEL); 57 if (id < 0) { 58 ibdev_err(ib_dev, "failed to alloc pd, id = %d.\n", id); 59 return -ENOMEM; 60 } 61 pd->pdn = (unsigned long)id; 62 63 if (udata) { 64 struct hns_roce_ib_alloc_pd_resp resp = {.pdn = pd->pdn}; 65 66 ret = ib_copy_to_udata(udata, &resp, 67 min(udata->outlen, sizeof(resp))); 68 if (ret) { 69 ida_free(&pd_ida->ida, id); 70 ibdev_err(ib_dev, "failed to copy to udata, ret = %d\n", ret); 71 } 72 } 73 74 return ret; 75 } 76 77 int hns_roce_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata) 78 { 79 struct hns_roce_dev *hr_dev = to_hr_dev(pd->device); 80 81 ida_free(&hr_dev->pd_ida.ida, (int)to_hr_pd(pd)->pdn); 82 83 return 0; 84 } 85 86 int hns_roce_uar_alloc(struct hns_roce_dev *hr_dev, struct hns_roce_uar *uar) 87 { 88 struct resource *res; 89 int ret; 90 91 /* Using bitmap to manager UAR index */ 92 ret = hns_roce_bitmap_alloc(&hr_dev->uar_table.bitmap, &uar->logic_idx); 93 if (ret) 94 return -ENOMEM; 95 96 if (uar->logic_idx > 0 && hr_dev->caps.phy_num_uars > 1) 97 uar->index = (uar->logic_idx - 1) % 98 (hr_dev->caps.phy_num_uars - 1) + 1; 99 else 100 uar->index = 0; 101 102 if (!dev_is_pci(hr_dev->dev)) { 103 res = platform_get_resource(hr_dev->pdev, IORESOURCE_MEM, 0); 104 if (!res) { 105 dev_err(&hr_dev->pdev->dev, "memory resource not found!\n"); 106 return -EINVAL; 107 } 108 uar->pfn = ((res->start) >> PAGE_SHIFT) + uar->index; 109 } else { 110 uar->pfn = ((pci_resource_start(hr_dev->pci_dev, 2)) 111 >> PAGE_SHIFT); 112 } 113 114 return 0; 115 } 116 117 void hns_roce_uar_free(struct hns_roce_dev *hr_dev, struct hns_roce_uar *uar) 118 { 119 hns_roce_bitmap_free(&hr_dev->uar_table.bitmap, uar->logic_idx); 120 } 121 122 int hns_roce_init_uar_table(struct hns_roce_dev *hr_dev) 123 { 124 return hns_roce_bitmap_init(&hr_dev->uar_table.bitmap, 125 hr_dev->caps.num_uars, 126 hr_dev->caps.num_uars - 1, 127 hr_dev->caps.reserved_uars, 0); 128 } 129 130 void hns_roce_cleanup_uar_table(struct hns_roce_dev *hr_dev) 131 { 132 hns_roce_bitmap_cleanup(&hr_dev->uar_table.bitmap); 133 } 134 135 static int hns_roce_xrcd_alloc(struct hns_roce_dev *hr_dev, u32 *xrcdn) 136 { 137 struct hns_roce_ida *xrcd_ida = &hr_dev->xrcd_ida; 138 int id; 139 140 id = ida_alloc_range(&xrcd_ida->ida, xrcd_ida->min, xrcd_ida->max, 141 GFP_KERNEL); 142 if (id < 0) { 143 ibdev_err(&hr_dev->ib_dev, "failed to alloc xrcdn(%d).\n", id); 144 return -ENOMEM; 145 } 146 *xrcdn = (u32)id; 147 148 return 0; 149 } 150 151 void hns_roce_init_xrcd_table(struct hns_roce_dev *hr_dev) 152 { 153 struct hns_roce_ida *xrcd_ida = &hr_dev->xrcd_ida; 154 155 ida_init(&xrcd_ida->ida); 156 xrcd_ida->max = hr_dev->caps.num_xrcds - 1; 157 xrcd_ida->min = hr_dev->caps.reserved_xrcds; 158 } 159 160 int hns_roce_alloc_xrcd(struct ib_xrcd *ib_xrcd, struct ib_udata *udata) 161 { 162 struct hns_roce_dev *hr_dev = to_hr_dev(ib_xrcd->device); 163 struct hns_roce_xrcd *xrcd = to_hr_xrcd(ib_xrcd); 164 int ret; 165 166 if (!(hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_XRC)) 167 return -EINVAL; 168 169 ret = hns_roce_xrcd_alloc(hr_dev, &xrcd->xrcdn); 170 if (ret) 171 return ret; 172 173 return 0; 174 } 175 176 int hns_roce_dealloc_xrcd(struct ib_xrcd *ib_xrcd, struct ib_udata *udata) 177 { 178 struct hns_roce_dev *hr_dev = to_hr_dev(ib_xrcd->device); 179 u32 xrcdn = to_hr_xrcd(ib_xrcd)->xrcdn; 180 181 ida_free(&hr_dev->xrcd_ida.ida, (int)xrcdn); 182 183 return 0; 184 } 185