1 /* 2 * Copyright (c) 2016, Mellanox Technologies. All rights reserved. 3 * 4 * This software is available to you under a choice of one of two 5 * licenses. You may choose to be licensed under the terms of the GNU 6 * General Public License (GPL) Version 2, available from the file 7 * COPYING in the main directory of this source tree, or the 8 * OpenIB.org BSD license below: 9 * 10 * Redistribution and use in source and binary forms, with or 11 * without modification, are permitted provided that the following 12 * conditions are met: 13 * 14 * - Redistributions of source code must retain the above 15 * copyright notice, this list of conditions and the following 16 * disclaimer. 17 * 18 * - Redistributions in binary form must reproduce the above 19 * copyright notice, this list of conditions and the following 20 * disclaimer in the documentation and/or other materials 21 * provided with the distribution. 22 * 23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 30 * SOFTWARE. 31 */ 32 33 #include "mlx5_ib.h" 34 35 struct mlx5_ib_gsi_wr { 36 struct ib_cqe cqe; 37 struct ib_wc wc; 38 bool completed:1; 39 }; 40 41 static bool mlx5_ib_deth_sqpn_cap(struct mlx5_ib_dev *dev) 42 { 43 return MLX5_CAP_GEN(dev->mdev, set_deth_sqpn); 44 } 45 46 /* Call with gsi->lock locked */ 47 static void generate_completions(struct mlx5_ib_qp *mqp) 48 { 49 struct mlx5_ib_gsi_qp *gsi = &mqp->gsi; 50 struct ib_cq *gsi_cq = mqp->ibqp.send_cq; 51 struct mlx5_ib_gsi_wr *wr; 52 u32 index; 53 54 for (index = gsi->outstanding_ci; index != gsi->outstanding_pi; 55 index++) { 56 wr = &gsi->outstanding_wrs[index % gsi->cap.max_send_wr]; 57 58 if (!wr->completed) 59 break; 60 61 WARN_ON_ONCE(mlx5_ib_generate_wc(gsi_cq, &wr->wc)); 62 wr->completed = false; 63 } 64 65 gsi->outstanding_ci = index; 66 } 67 68 static void handle_single_completion(struct ib_cq *cq, struct ib_wc *wc) 69 { 70 struct mlx5_ib_gsi_qp *gsi = cq->cq_context; 71 struct mlx5_ib_gsi_wr *wr = 72 container_of(wc->wr_cqe, struct mlx5_ib_gsi_wr, cqe); 73 struct mlx5_ib_qp *mqp = container_of(gsi, struct mlx5_ib_qp, gsi); 74 u64 wr_id; 75 unsigned long flags; 76 77 spin_lock_irqsave(&gsi->lock, flags); 78 wr->completed = true; 79 wr_id = wr->wc.wr_id; 80 wr->wc = *wc; 81 wr->wc.wr_id = wr_id; 82 wr->wc.qp = &mqp->ibqp; 83 84 generate_completions(mqp); 85 spin_unlock_irqrestore(&gsi->lock, flags); 86 } 87 88 int mlx5_ib_create_gsi(struct ib_pd *pd, struct mlx5_ib_qp *mqp, 89 struct ib_qp_init_attr *attr) 90 { 91 struct mlx5_ib_dev *dev = to_mdev(pd->device); 92 struct mlx5_ib_gsi_qp *gsi; 93 struct ib_qp_init_attr hw_init_attr = *attr; 94 const u8 port_num = attr->port_num; 95 int num_qps = 0; 96 int ret; 97 98 if (mlx5_ib_deth_sqpn_cap(dev)) { 99 if (MLX5_CAP_GEN(dev->mdev, 100 port_type) == MLX5_CAP_PORT_TYPE_IB) 101 num_qps = pd->device->attrs.max_pkeys; 102 else if (dev->lag_active) 103 num_qps = dev->lag_ports; 104 } 105 106 gsi = &mqp->gsi; 107 gsi->tx_qps = kcalloc(num_qps, sizeof(*gsi->tx_qps), GFP_KERNEL); 108 if (!gsi->tx_qps) 109 return -ENOMEM; 110 111 gsi->outstanding_wrs = 112 kcalloc(attr->cap.max_send_wr, sizeof(*gsi->outstanding_wrs), 113 GFP_KERNEL); 114 if (!gsi->outstanding_wrs) { 115 ret = -ENOMEM; 116 goto err_free_tx; 117 } 118 119 if (dev->devr.ports[port_num - 1].gsi) { 120 mlx5_ib_warn(dev, "GSI QP already exists on port %d\n", 121 port_num); 122 ret = -EBUSY; 123 goto err_free_wrs; 124 } 125 gsi->num_qps = num_qps; 126 spin_lock_init(&gsi->lock); 127 128 gsi->cap = attr->cap; 129 gsi->port_num = port_num; 130 131 gsi->cq = ib_alloc_cq(pd->device, gsi, attr->cap.max_send_wr, 0, 132 IB_POLL_SOFTIRQ); 133 if (IS_ERR(gsi->cq)) { 134 mlx5_ib_warn(dev, "unable to create send CQ for GSI QP. error %ld\n", 135 PTR_ERR(gsi->cq)); 136 ret = PTR_ERR(gsi->cq); 137 goto err_free_wrs; 138 } 139 140 hw_init_attr.qp_type = MLX5_IB_QPT_HW_GSI; 141 hw_init_attr.send_cq = gsi->cq; 142 if (num_qps) { 143 hw_init_attr.cap.max_send_wr = 0; 144 hw_init_attr.cap.max_send_sge = 0; 145 hw_init_attr.cap.max_inline_data = 0; 146 } 147 148 gsi->rx_qp = ib_create_qp(pd, &hw_init_attr); 149 if (IS_ERR(gsi->rx_qp)) { 150 mlx5_ib_warn(dev, "unable to create hardware GSI QP. error %ld\n", 151 PTR_ERR(gsi->rx_qp)); 152 ret = PTR_ERR(gsi->rx_qp); 153 goto err_destroy_cq; 154 } 155 156 dev->devr.ports[attr->port_num - 1].gsi = gsi; 157 return 0; 158 159 err_destroy_cq: 160 ib_free_cq(gsi->cq); 161 err_free_wrs: 162 kfree(gsi->outstanding_wrs); 163 err_free_tx: 164 kfree(gsi->tx_qps); 165 return ret; 166 } 167 168 int mlx5_ib_destroy_gsi(struct mlx5_ib_qp *mqp) 169 { 170 struct mlx5_ib_dev *dev = to_mdev(mqp->ibqp.device); 171 struct mlx5_ib_gsi_qp *gsi = &mqp->gsi; 172 const int port_num = gsi->port_num; 173 int qp_index; 174 int ret; 175 176 ret = ib_destroy_qp(gsi->rx_qp); 177 if (ret) { 178 mlx5_ib_warn(dev, "unable to destroy hardware GSI QP. error %d\n", 179 ret); 180 return ret; 181 } 182 dev->devr.ports[port_num - 1].gsi = NULL; 183 gsi->rx_qp = NULL; 184 185 for (qp_index = 0; qp_index < gsi->num_qps; ++qp_index) { 186 if (!gsi->tx_qps[qp_index]) 187 continue; 188 WARN_ON_ONCE(ib_destroy_qp(gsi->tx_qps[qp_index])); 189 gsi->tx_qps[qp_index] = NULL; 190 } 191 192 ib_free_cq(gsi->cq); 193 194 kfree(gsi->outstanding_wrs); 195 kfree(gsi->tx_qps); 196 return 0; 197 } 198 199 static struct ib_qp *create_gsi_ud_qp(struct mlx5_ib_gsi_qp *gsi) 200 { 201 struct ib_pd *pd = gsi->rx_qp->pd; 202 struct ib_qp_init_attr init_attr = { 203 .event_handler = gsi->rx_qp->event_handler, 204 .qp_context = gsi->rx_qp->qp_context, 205 .send_cq = gsi->cq, 206 .recv_cq = gsi->rx_qp->recv_cq, 207 .cap = { 208 .max_send_wr = gsi->cap.max_send_wr, 209 .max_send_sge = gsi->cap.max_send_sge, 210 .max_inline_data = gsi->cap.max_inline_data, 211 }, 212 .qp_type = IB_QPT_UD, 213 .create_flags = MLX5_IB_QP_CREATE_SQPN_QP1, 214 }; 215 216 return ib_create_qp(pd, &init_attr); 217 } 218 219 static int modify_to_rts(struct mlx5_ib_gsi_qp *gsi, struct ib_qp *qp, 220 u16 pkey_index) 221 { 222 struct mlx5_ib_dev *dev = to_mdev(qp->device); 223 struct ib_qp_attr attr; 224 int mask; 225 int ret; 226 227 mask = IB_QP_STATE | IB_QP_PKEY_INDEX | IB_QP_QKEY | IB_QP_PORT; 228 attr.qp_state = IB_QPS_INIT; 229 attr.pkey_index = pkey_index; 230 attr.qkey = IB_QP1_QKEY; 231 attr.port_num = gsi->port_num; 232 ret = ib_modify_qp(qp, &attr, mask); 233 if (ret) { 234 mlx5_ib_err(dev, "could not change QP%d state to INIT: %d\n", 235 qp->qp_num, ret); 236 return ret; 237 } 238 239 attr.qp_state = IB_QPS_RTR; 240 ret = ib_modify_qp(qp, &attr, IB_QP_STATE); 241 if (ret) { 242 mlx5_ib_err(dev, "could not change QP%d state to RTR: %d\n", 243 qp->qp_num, ret); 244 return ret; 245 } 246 247 attr.qp_state = IB_QPS_RTS; 248 attr.sq_psn = 0; 249 ret = ib_modify_qp(qp, &attr, IB_QP_STATE | IB_QP_SQ_PSN); 250 if (ret) { 251 mlx5_ib_err(dev, "could not change QP%d state to RTS: %d\n", 252 qp->qp_num, ret); 253 return ret; 254 } 255 256 return 0; 257 } 258 259 static void setup_qp(struct mlx5_ib_gsi_qp *gsi, u16 qp_index) 260 { 261 struct ib_device *device = gsi->rx_qp->device; 262 struct mlx5_ib_dev *dev = to_mdev(device); 263 int pkey_index = qp_index; 264 struct mlx5_ib_qp *mqp; 265 struct ib_qp *qp; 266 unsigned long flags; 267 u16 pkey; 268 int ret; 269 270 if (MLX5_CAP_GEN(dev->mdev, port_type) != MLX5_CAP_PORT_TYPE_IB) 271 pkey_index = 0; 272 273 ret = ib_query_pkey(device, gsi->port_num, pkey_index, &pkey); 274 if (ret) { 275 mlx5_ib_warn(dev, "unable to read P_Key at port %d, index %d\n", 276 gsi->port_num, qp_index); 277 return; 278 } 279 280 if (!pkey) { 281 mlx5_ib_dbg(dev, "invalid P_Key at port %d, index %d. Skipping.\n", 282 gsi->port_num, qp_index); 283 return; 284 } 285 286 spin_lock_irqsave(&gsi->lock, flags); 287 qp = gsi->tx_qps[qp_index]; 288 spin_unlock_irqrestore(&gsi->lock, flags); 289 if (qp) { 290 mlx5_ib_dbg(dev, "already existing GSI TX QP at port %d, index %d. Skipping\n", 291 gsi->port_num, qp_index); 292 return; 293 } 294 295 qp = create_gsi_ud_qp(gsi); 296 if (IS_ERR(qp)) { 297 mlx5_ib_warn(dev, "unable to create hardware UD QP for GSI: %ld\n", 298 PTR_ERR(qp)); 299 return; 300 } 301 302 mqp = to_mqp(qp); 303 if (dev->lag_active) 304 mqp->gsi_lag_port = qp_index + 1; 305 ret = modify_to_rts(gsi, qp, pkey_index); 306 if (ret) 307 goto err_destroy_qp; 308 309 spin_lock_irqsave(&gsi->lock, flags); 310 WARN_ON_ONCE(gsi->tx_qps[qp_index]); 311 gsi->tx_qps[qp_index] = qp; 312 spin_unlock_irqrestore(&gsi->lock, flags); 313 314 return; 315 316 err_destroy_qp: 317 WARN_ON_ONCE(qp); 318 } 319 320 int mlx5_ib_gsi_modify_qp(struct ib_qp *qp, struct ib_qp_attr *attr, 321 int attr_mask) 322 { 323 struct mlx5_ib_dev *dev = to_mdev(qp->device); 324 struct mlx5_ib_qp *mqp = to_mqp(qp); 325 struct mlx5_ib_gsi_qp *gsi = &mqp->gsi; 326 u16 qp_index; 327 int ret; 328 329 mlx5_ib_dbg(dev, "modifying GSI QP to state %d\n", attr->qp_state); 330 331 ret = ib_modify_qp(gsi->rx_qp, attr, attr_mask); 332 if (ret) { 333 mlx5_ib_warn(dev, "unable to modify GSI rx QP: %d\n", ret); 334 return ret; 335 } 336 337 if (to_mqp(gsi->rx_qp)->state != IB_QPS_RTS) 338 return 0; 339 340 for (qp_index = 0; qp_index < gsi->num_qps; ++qp_index) 341 setup_qp(gsi, qp_index); 342 return 0; 343 } 344 345 int mlx5_ib_gsi_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, 346 int qp_attr_mask, 347 struct ib_qp_init_attr *qp_init_attr) 348 { 349 struct mlx5_ib_qp *mqp = to_mqp(qp); 350 struct mlx5_ib_gsi_qp *gsi = &mqp->gsi; 351 int ret; 352 353 ret = ib_query_qp(gsi->rx_qp, qp_attr, qp_attr_mask, qp_init_attr); 354 qp_init_attr->cap = gsi->cap; 355 return ret; 356 } 357 358 /* Call with gsi->lock locked */ 359 static int mlx5_ib_add_outstanding_wr(struct mlx5_ib_qp *mqp, 360 struct ib_ud_wr *wr, struct ib_wc *wc) 361 { 362 struct mlx5_ib_gsi_qp *gsi = &mqp->gsi; 363 struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device); 364 struct mlx5_ib_gsi_wr *gsi_wr; 365 366 if (gsi->outstanding_pi == gsi->outstanding_ci + gsi->cap.max_send_wr) { 367 mlx5_ib_warn(dev, "no available GSI work request.\n"); 368 return -ENOMEM; 369 } 370 371 gsi_wr = &gsi->outstanding_wrs[gsi->outstanding_pi % 372 gsi->cap.max_send_wr]; 373 gsi->outstanding_pi++; 374 375 if (!wc) { 376 memset(&gsi_wr->wc, 0, sizeof(gsi_wr->wc)); 377 gsi_wr->wc.pkey_index = wr->pkey_index; 378 gsi_wr->wc.wr_id = wr->wr.wr_id; 379 } else { 380 gsi_wr->wc = *wc; 381 gsi_wr->completed = true; 382 } 383 384 gsi_wr->cqe.done = &handle_single_completion; 385 wr->wr.wr_cqe = &gsi_wr->cqe; 386 387 return 0; 388 } 389 390 /* Call with gsi->lock locked */ 391 static int mlx5_ib_gsi_silent_drop(struct mlx5_ib_qp *mqp, struct ib_ud_wr *wr) 392 { 393 struct ib_wc wc = { 394 { .wr_id = wr->wr.wr_id }, 395 .status = IB_WC_SUCCESS, 396 .opcode = IB_WC_SEND, 397 .qp = &mqp->ibqp, 398 }; 399 int ret; 400 401 ret = mlx5_ib_add_outstanding_wr(mqp, wr, &wc); 402 if (ret) 403 return ret; 404 405 generate_completions(mqp); 406 407 return 0; 408 } 409 410 /* Call with gsi->lock locked */ 411 static struct ib_qp *get_tx_qp(struct mlx5_ib_gsi_qp *gsi, struct ib_ud_wr *wr) 412 { 413 struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device); 414 struct mlx5_ib_ah *ah = to_mah(wr->ah); 415 int qp_index = wr->pkey_index; 416 417 if (!gsi->num_qps) 418 return gsi->rx_qp; 419 420 if (dev->lag_active && ah->xmit_port) 421 qp_index = ah->xmit_port - 1; 422 423 if (qp_index >= gsi->num_qps) 424 return NULL; 425 426 return gsi->tx_qps[qp_index]; 427 } 428 429 int mlx5_ib_gsi_post_send(struct ib_qp *qp, const struct ib_send_wr *wr, 430 const struct ib_send_wr **bad_wr) 431 { 432 struct mlx5_ib_qp *mqp = to_mqp(qp); 433 struct mlx5_ib_gsi_qp *gsi = &mqp->gsi; 434 struct ib_qp *tx_qp; 435 unsigned long flags; 436 int ret; 437 438 for (; wr; wr = wr->next) { 439 struct ib_ud_wr cur_wr = *ud_wr(wr); 440 441 cur_wr.wr.next = NULL; 442 443 spin_lock_irqsave(&gsi->lock, flags); 444 tx_qp = get_tx_qp(gsi, &cur_wr); 445 if (!tx_qp) { 446 ret = mlx5_ib_gsi_silent_drop(mqp, &cur_wr); 447 if (ret) 448 goto err; 449 spin_unlock_irqrestore(&gsi->lock, flags); 450 continue; 451 } 452 453 ret = mlx5_ib_add_outstanding_wr(mqp, &cur_wr, NULL); 454 if (ret) 455 goto err; 456 457 ret = ib_post_send(tx_qp, &cur_wr.wr, bad_wr); 458 if (ret) { 459 /* Undo the effect of adding the outstanding wr */ 460 gsi->outstanding_pi--; 461 goto err; 462 } 463 spin_unlock_irqrestore(&gsi->lock, flags); 464 } 465 466 return 0; 467 468 err: 469 spin_unlock_irqrestore(&gsi->lock, flags); 470 *bad_wr = wr; 471 return ret; 472 } 473 474 int mlx5_ib_gsi_post_recv(struct ib_qp *qp, const struct ib_recv_wr *wr, 475 const struct ib_recv_wr **bad_wr) 476 { 477 struct mlx5_ib_qp *mqp = to_mqp(qp); 478 struct mlx5_ib_gsi_qp *gsi = &mqp->gsi; 479 480 return ib_post_recv(gsi->rx_qp, wr, bad_wr); 481 } 482 483 void mlx5_ib_gsi_pkey_change(struct mlx5_ib_gsi_qp *gsi) 484 { 485 u16 qp_index; 486 487 for (qp_index = 0; qp_index < gsi->num_qps; ++qp_index) 488 setup_qp(gsi, qp_index); 489 } 490