1 /* 2 * Copyright (c) 2016 Mellanox Technologies Ltd. All rights reserved. 3 * Copyright (c) 2015 System Fabric Works, Inc. All rights reserved. 4 * 5 * This software is available to you under a choice of one of two 6 * licenses. You may choose to be licensed under the terms of the GNU 7 * General Public License (GPL) Version 2, available from the file 8 * COPYING in the main directory of this source tree, or the 9 * OpenIB.org BSD license below: 10 * 11 * Redistribution and use in source and binary forms, with or 12 * without modification, are permitted provided that the following 13 * conditions are met: 14 * 15 * - Redistributions of source code must retain the above 16 * copyright notice, this list of conditions and the following 17 * disclaimer. 18 * 19 * - Redistributions in binary form must reproduce the above 20 * copyright notice, this list of conditions and the following 21 * disclaimer in the documentation and/or other materials 22 * provided with the distribution. 23 * 24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 25 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 26 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 27 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 28 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 29 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 30 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 31 * SOFTWARE. 32 */ 33 34 #include <linux/skbuff.h> 35 36 #include "rxe.h" 37 #include "rxe_loc.h" 38 #include "rxe_queue.h" 39 40 static int next_opcode(struct rxe_qp *qp, struct rxe_send_wqe *wqe, 41 unsigned opcode); 42 43 static inline void retry_first_write_send(struct rxe_qp *qp, 44 struct rxe_send_wqe *wqe, 45 unsigned mask, int npsn) 46 { 47 int i; 48 49 for (i = 0; i < npsn; i++) { 50 int to_send = (wqe->dma.resid > qp->mtu) ? 51 qp->mtu : wqe->dma.resid; 52 53 qp->req.opcode = next_opcode(qp, wqe, 54 wqe->wr.opcode); 55 56 if (wqe->wr.send_flags & IB_SEND_INLINE) { 57 wqe->dma.resid -= to_send; 58 wqe->dma.sge_offset += to_send; 59 } else { 60 advance_dma_data(&wqe->dma, to_send); 61 } 62 if (mask & WR_WRITE_MASK) 63 wqe->iova += qp->mtu; 64 } 65 } 66 67 static void req_retry(struct rxe_qp *qp) 68 { 69 struct rxe_send_wqe *wqe; 70 unsigned int wqe_index; 71 unsigned int mask; 72 int npsn; 73 int first = 1; 74 75 wqe = queue_head(qp->sq.queue); 76 npsn = (qp->comp.psn - wqe->first_psn) & BTH_PSN_MASK; 77 78 qp->req.wqe_index = consumer_index(qp->sq.queue); 79 qp->req.psn = qp->comp.psn; 80 qp->req.opcode = -1; 81 82 for (wqe_index = consumer_index(qp->sq.queue); 83 wqe_index != producer_index(qp->sq.queue); 84 wqe_index = next_index(qp->sq.queue, wqe_index)) { 85 wqe = addr_from_index(qp->sq.queue, wqe_index); 86 mask = wr_opcode_mask(wqe->wr.opcode, qp); 87 88 if (wqe->state == wqe_state_posted) 89 break; 90 91 if (wqe->state == wqe_state_done) 92 continue; 93 94 wqe->iova = (mask & WR_ATOMIC_MASK) ? 95 wqe->wr.wr.atomic.remote_addr : 96 (mask & WR_READ_OR_WRITE_MASK) ? 97 wqe->wr.wr.rdma.remote_addr : 98 0; 99 100 if (!first || (mask & WR_READ_MASK) == 0) { 101 wqe->dma.resid = wqe->dma.length; 102 wqe->dma.cur_sge = 0; 103 wqe->dma.sge_offset = 0; 104 } 105 106 if (first) { 107 first = 0; 108 109 if (mask & WR_WRITE_OR_SEND_MASK) 110 retry_first_write_send(qp, wqe, mask, npsn); 111 112 if (mask & WR_READ_MASK) 113 wqe->iova += npsn * qp->mtu; 114 } 115 116 wqe->state = wqe_state_posted; 117 } 118 } 119 120 void rnr_nak_timer(unsigned long data) 121 { 122 struct rxe_qp *qp = (struct rxe_qp *)data; 123 124 pr_debug("rnr nak timer fired\n"); 125 rxe_run_task(&qp->req.task, 1); 126 } 127 128 static struct rxe_send_wqe *req_next_wqe(struct rxe_qp *qp) 129 { 130 struct rxe_send_wqe *wqe = queue_head(qp->sq.queue); 131 unsigned long flags; 132 133 if (unlikely(qp->req.state == QP_STATE_DRAIN)) { 134 /* check to see if we are drained; 135 * state_lock used by requester and completer 136 */ 137 spin_lock_irqsave(&qp->state_lock, flags); 138 do { 139 if (qp->req.state != QP_STATE_DRAIN) { 140 /* comp just finished */ 141 spin_unlock_irqrestore(&qp->state_lock, 142 flags); 143 break; 144 } 145 146 if (wqe && ((qp->req.wqe_index != 147 consumer_index(qp->sq.queue)) || 148 (wqe->state != wqe_state_posted))) { 149 /* comp not done yet */ 150 spin_unlock_irqrestore(&qp->state_lock, 151 flags); 152 break; 153 } 154 155 qp->req.state = QP_STATE_DRAINED; 156 spin_unlock_irqrestore(&qp->state_lock, flags); 157 158 if (qp->ibqp.event_handler) { 159 struct ib_event ev; 160 161 ev.device = qp->ibqp.device; 162 ev.element.qp = &qp->ibqp; 163 ev.event = IB_EVENT_SQ_DRAINED; 164 qp->ibqp.event_handler(&ev, 165 qp->ibqp.qp_context); 166 } 167 } while (0); 168 } 169 170 if (qp->req.wqe_index == producer_index(qp->sq.queue)) 171 return NULL; 172 173 wqe = addr_from_index(qp->sq.queue, qp->req.wqe_index); 174 175 if (unlikely((qp->req.state == QP_STATE_DRAIN || 176 qp->req.state == QP_STATE_DRAINED) && 177 (wqe->state != wqe_state_processing))) 178 return NULL; 179 180 if (unlikely((wqe->wr.send_flags & IB_SEND_FENCE) && 181 (qp->req.wqe_index != consumer_index(qp->sq.queue)))) { 182 qp->req.wait_fence = 1; 183 return NULL; 184 } 185 186 wqe->mask = wr_opcode_mask(wqe->wr.opcode, qp); 187 return wqe; 188 } 189 190 static int next_opcode_rc(struct rxe_qp *qp, unsigned opcode, int fits) 191 { 192 switch (opcode) { 193 case IB_WR_RDMA_WRITE: 194 if (qp->req.opcode == IB_OPCODE_RC_RDMA_WRITE_FIRST || 195 qp->req.opcode == IB_OPCODE_RC_RDMA_WRITE_MIDDLE) 196 return fits ? 197 IB_OPCODE_RC_RDMA_WRITE_LAST : 198 IB_OPCODE_RC_RDMA_WRITE_MIDDLE; 199 else 200 return fits ? 201 IB_OPCODE_RC_RDMA_WRITE_ONLY : 202 IB_OPCODE_RC_RDMA_WRITE_FIRST; 203 204 case IB_WR_RDMA_WRITE_WITH_IMM: 205 if (qp->req.opcode == IB_OPCODE_RC_RDMA_WRITE_FIRST || 206 qp->req.opcode == IB_OPCODE_RC_RDMA_WRITE_MIDDLE) 207 return fits ? 208 IB_OPCODE_RC_RDMA_WRITE_LAST_WITH_IMMEDIATE : 209 IB_OPCODE_RC_RDMA_WRITE_MIDDLE; 210 else 211 return fits ? 212 IB_OPCODE_RC_RDMA_WRITE_ONLY_WITH_IMMEDIATE : 213 IB_OPCODE_RC_RDMA_WRITE_FIRST; 214 215 case IB_WR_SEND: 216 if (qp->req.opcode == IB_OPCODE_RC_SEND_FIRST || 217 qp->req.opcode == IB_OPCODE_RC_SEND_MIDDLE) 218 return fits ? 219 IB_OPCODE_RC_SEND_LAST : 220 IB_OPCODE_RC_SEND_MIDDLE; 221 else 222 return fits ? 223 IB_OPCODE_RC_SEND_ONLY : 224 IB_OPCODE_RC_SEND_FIRST; 225 226 case IB_WR_SEND_WITH_IMM: 227 if (qp->req.opcode == IB_OPCODE_RC_SEND_FIRST || 228 qp->req.opcode == IB_OPCODE_RC_SEND_MIDDLE) 229 return fits ? 230 IB_OPCODE_RC_SEND_LAST_WITH_IMMEDIATE : 231 IB_OPCODE_RC_SEND_MIDDLE; 232 else 233 return fits ? 234 IB_OPCODE_RC_SEND_ONLY_WITH_IMMEDIATE : 235 IB_OPCODE_RC_SEND_FIRST; 236 237 case IB_WR_RDMA_READ: 238 return IB_OPCODE_RC_RDMA_READ_REQUEST; 239 240 case IB_WR_ATOMIC_CMP_AND_SWP: 241 return IB_OPCODE_RC_COMPARE_SWAP; 242 243 case IB_WR_ATOMIC_FETCH_AND_ADD: 244 return IB_OPCODE_RC_FETCH_ADD; 245 246 case IB_WR_SEND_WITH_INV: 247 if (qp->req.opcode == IB_OPCODE_RC_SEND_FIRST || 248 qp->req.opcode == IB_OPCODE_RC_SEND_MIDDLE) 249 return fits ? IB_OPCODE_RC_SEND_LAST_WITH_INVALIDATE : 250 IB_OPCODE_RC_SEND_MIDDLE; 251 else 252 return fits ? IB_OPCODE_RC_SEND_ONLY_WITH_INVALIDATE : 253 IB_OPCODE_RC_SEND_FIRST; 254 case IB_WR_REG_MR: 255 case IB_WR_LOCAL_INV: 256 return opcode; 257 } 258 259 return -EINVAL; 260 } 261 262 static int next_opcode_uc(struct rxe_qp *qp, unsigned opcode, int fits) 263 { 264 switch (opcode) { 265 case IB_WR_RDMA_WRITE: 266 if (qp->req.opcode == IB_OPCODE_UC_RDMA_WRITE_FIRST || 267 qp->req.opcode == IB_OPCODE_UC_RDMA_WRITE_MIDDLE) 268 return fits ? 269 IB_OPCODE_UC_RDMA_WRITE_LAST : 270 IB_OPCODE_UC_RDMA_WRITE_MIDDLE; 271 else 272 return fits ? 273 IB_OPCODE_UC_RDMA_WRITE_ONLY : 274 IB_OPCODE_UC_RDMA_WRITE_FIRST; 275 276 case IB_WR_RDMA_WRITE_WITH_IMM: 277 if (qp->req.opcode == IB_OPCODE_UC_RDMA_WRITE_FIRST || 278 qp->req.opcode == IB_OPCODE_UC_RDMA_WRITE_MIDDLE) 279 return fits ? 280 IB_OPCODE_UC_RDMA_WRITE_LAST_WITH_IMMEDIATE : 281 IB_OPCODE_UC_RDMA_WRITE_MIDDLE; 282 else 283 return fits ? 284 IB_OPCODE_UC_RDMA_WRITE_ONLY_WITH_IMMEDIATE : 285 IB_OPCODE_UC_RDMA_WRITE_FIRST; 286 287 case IB_WR_SEND: 288 if (qp->req.opcode == IB_OPCODE_UC_SEND_FIRST || 289 qp->req.opcode == IB_OPCODE_UC_SEND_MIDDLE) 290 return fits ? 291 IB_OPCODE_UC_SEND_LAST : 292 IB_OPCODE_UC_SEND_MIDDLE; 293 else 294 return fits ? 295 IB_OPCODE_UC_SEND_ONLY : 296 IB_OPCODE_UC_SEND_FIRST; 297 298 case IB_WR_SEND_WITH_IMM: 299 if (qp->req.opcode == IB_OPCODE_UC_SEND_FIRST || 300 qp->req.opcode == IB_OPCODE_UC_SEND_MIDDLE) 301 return fits ? 302 IB_OPCODE_UC_SEND_LAST_WITH_IMMEDIATE : 303 IB_OPCODE_UC_SEND_MIDDLE; 304 else 305 return fits ? 306 IB_OPCODE_UC_SEND_ONLY_WITH_IMMEDIATE : 307 IB_OPCODE_UC_SEND_FIRST; 308 } 309 310 return -EINVAL; 311 } 312 313 static int next_opcode(struct rxe_qp *qp, struct rxe_send_wqe *wqe, 314 unsigned opcode) 315 { 316 int fits = (wqe->dma.resid <= qp->mtu); 317 318 switch (qp_type(qp)) { 319 case IB_QPT_RC: 320 return next_opcode_rc(qp, opcode, fits); 321 322 case IB_QPT_UC: 323 return next_opcode_uc(qp, opcode, fits); 324 325 case IB_QPT_SMI: 326 case IB_QPT_UD: 327 case IB_QPT_GSI: 328 switch (opcode) { 329 case IB_WR_SEND: 330 return IB_OPCODE_UD_SEND_ONLY; 331 332 case IB_WR_SEND_WITH_IMM: 333 return IB_OPCODE_UD_SEND_ONLY_WITH_IMMEDIATE; 334 } 335 break; 336 337 default: 338 break; 339 } 340 341 return -EINVAL; 342 } 343 344 static inline int check_init_depth(struct rxe_qp *qp, struct rxe_send_wqe *wqe) 345 { 346 int depth; 347 348 if (wqe->has_rd_atomic) 349 return 0; 350 351 qp->req.need_rd_atomic = 1; 352 depth = atomic_dec_return(&qp->req.rd_atomic); 353 354 if (depth >= 0) { 355 qp->req.need_rd_atomic = 0; 356 wqe->has_rd_atomic = 1; 357 return 0; 358 } 359 360 atomic_inc(&qp->req.rd_atomic); 361 return -EAGAIN; 362 } 363 364 static inline int get_mtu(struct rxe_qp *qp, struct rxe_send_wqe *wqe) 365 { 366 struct rxe_dev *rxe = to_rdev(qp->ibqp.device); 367 struct rxe_port *port; 368 struct rxe_av *av; 369 370 if ((qp_type(qp) == IB_QPT_RC) || (qp_type(qp) == IB_QPT_UC)) 371 return qp->mtu; 372 373 av = &wqe->av; 374 port = &rxe->port; 375 376 return port->mtu_cap; 377 } 378 379 static struct sk_buff *init_req_packet(struct rxe_qp *qp, 380 struct rxe_send_wqe *wqe, 381 int opcode, int payload, 382 struct rxe_pkt_info *pkt) 383 { 384 struct rxe_dev *rxe = to_rdev(qp->ibqp.device); 385 struct rxe_port *port = &rxe->port; 386 struct sk_buff *skb; 387 struct rxe_send_wr *ibwr = &wqe->wr; 388 struct rxe_av *av; 389 int pad = (-payload) & 0x3; 390 int paylen; 391 int solicited; 392 u16 pkey; 393 u32 qp_num; 394 int ack_req; 395 396 /* length from start of bth to end of icrc */ 397 paylen = rxe_opcode[opcode].length + payload + pad + RXE_ICRC_SIZE; 398 399 /* pkt->hdr, rxe, port_num and mask are initialized in ifc 400 * layer 401 */ 402 pkt->opcode = opcode; 403 pkt->qp = qp; 404 pkt->psn = qp->req.psn; 405 pkt->mask = rxe_opcode[opcode].mask; 406 pkt->paylen = paylen; 407 pkt->offset = 0; 408 pkt->wqe = wqe; 409 410 /* init skb */ 411 av = rxe_get_av(pkt); 412 skb = rxe->ifc_ops->init_packet(rxe, av, paylen, pkt); 413 if (unlikely(!skb)) 414 return NULL; 415 416 /* init bth */ 417 solicited = (ibwr->send_flags & IB_SEND_SOLICITED) && 418 (pkt->mask & RXE_END_MASK) && 419 ((pkt->mask & (RXE_SEND_MASK)) || 420 (pkt->mask & (RXE_WRITE_MASK | RXE_IMMDT_MASK)) == 421 (RXE_WRITE_MASK | RXE_IMMDT_MASK)); 422 423 pkey = (qp_type(qp) == IB_QPT_GSI) ? 424 port->pkey_tbl[ibwr->wr.ud.pkey_index] : 425 port->pkey_tbl[qp->attr.pkey_index]; 426 427 qp_num = (pkt->mask & RXE_DETH_MASK) ? ibwr->wr.ud.remote_qpn : 428 qp->attr.dest_qp_num; 429 430 ack_req = ((pkt->mask & RXE_END_MASK) || 431 (qp->req.noack_pkts++ > RXE_MAX_PKT_PER_ACK)); 432 if (ack_req) 433 qp->req.noack_pkts = 0; 434 435 bth_init(pkt, pkt->opcode, solicited, 0, pad, pkey, qp_num, 436 ack_req, pkt->psn); 437 438 /* init optional headers */ 439 if (pkt->mask & RXE_RETH_MASK) { 440 reth_set_rkey(pkt, ibwr->wr.rdma.rkey); 441 reth_set_va(pkt, wqe->iova); 442 reth_set_len(pkt, wqe->dma.length); 443 } 444 445 if (pkt->mask & RXE_IMMDT_MASK) 446 immdt_set_imm(pkt, ibwr->ex.imm_data); 447 448 if (pkt->mask & RXE_IETH_MASK) 449 ieth_set_rkey(pkt, ibwr->ex.invalidate_rkey); 450 451 if (pkt->mask & RXE_ATMETH_MASK) { 452 atmeth_set_va(pkt, wqe->iova); 453 if (opcode == IB_OPCODE_RC_COMPARE_SWAP || 454 opcode == IB_OPCODE_RD_COMPARE_SWAP) { 455 atmeth_set_swap_add(pkt, ibwr->wr.atomic.swap); 456 atmeth_set_comp(pkt, ibwr->wr.atomic.compare_add); 457 } else { 458 atmeth_set_swap_add(pkt, ibwr->wr.atomic.compare_add); 459 } 460 atmeth_set_rkey(pkt, ibwr->wr.atomic.rkey); 461 } 462 463 if (pkt->mask & RXE_DETH_MASK) { 464 if (qp->ibqp.qp_num == 1) 465 deth_set_qkey(pkt, GSI_QKEY); 466 else 467 deth_set_qkey(pkt, ibwr->wr.ud.remote_qkey); 468 deth_set_sqp(pkt, qp->ibqp.qp_num); 469 } 470 471 return skb; 472 } 473 474 static int fill_packet(struct rxe_qp *qp, struct rxe_send_wqe *wqe, 475 struct rxe_pkt_info *pkt, struct sk_buff *skb, 476 int paylen) 477 { 478 struct rxe_dev *rxe = to_rdev(qp->ibqp.device); 479 u32 crc = 0; 480 u32 *p; 481 int err; 482 483 err = rxe->ifc_ops->prepare(rxe, pkt, skb, &crc); 484 if (err) 485 return err; 486 487 if (pkt->mask & RXE_WRITE_OR_SEND) { 488 if (wqe->wr.send_flags & IB_SEND_INLINE) { 489 u8 *tmp = &wqe->dma.inline_data[wqe->dma.sge_offset]; 490 491 crc = crc32_le(crc, tmp, paylen); 492 493 memcpy(payload_addr(pkt), tmp, paylen); 494 495 wqe->dma.resid -= paylen; 496 wqe->dma.sge_offset += paylen; 497 } else { 498 err = copy_data(rxe, qp->pd, 0, &wqe->dma, 499 payload_addr(pkt), paylen, 500 from_mem_obj, 501 &crc); 502 if (err) 503 return err; 504 } 505 } 506 p = payload_addr(pkt) + paylen + bth_pad(pkt); 507 508 *p = ~crc; 509 510 return 0; 511 } 512 513 static void update_wqe_state(struct rxe_qp *qp, 514 struct rxe_send_wqe *wqe, 515 struct rxe_pkt_info *pkt, 516 enum wqe_state *prev_state) 517 { 518 enum wqe_state prev_state_ = wqe->state; 519 520 if (pkt->mask & RXE_END_MASK) { 521 if (qp_type(qp) == IB_QPT_RC) 522 wqe->state = wqe_state_pending; 523 } else { 524 wqe->state = wqe_state_processing; 525 } 526 527 *prev_state = prev_state_; 528 } 529 530 static void update_state(struct rxe_qp *qp, struct rxe_send_wqe *wqe, 531 struct rxe_pkt_info *pkt, int payload) 532 { 533 /* number of packets left to send including current one */ 534 int num_pkt = (wqe->dma.resid + payload + qp->mtu - 1) / qp->mtu; 535 536 /* handle zero length packet case */ 537 if (num_pkt == 0) 538 num_pkt = 1; 539 540 if (pkt->mask & RXE_START_MASK) { 541 wqe->first_psn = qp->req.psn; 542 wqe->last_psn = (qp->req.psn + num_pkt - 1) & BTH_PSN_MASK; 543 } 544 545 if (pkt->mask & RXE_READ_MASK) 546 qp->req.psn = (wqe->first_psn + num_pkt) & BTH_PSN_MASK; 547 else 548 qp->req.psn = (qp->req.psn + 1) & BTH_PSN_MASK; 549 550 qp->req.opcode = pkt->opcode; 551 552 553 if (pkt->mask & RXE_END_MASK) 554 qp->req.wqe_index = next_index(qp->sq.queue, qp->req.wqe_index); 555 556 qp->need_req_skb = 0; 557 558 if (qp->qp_timeout_jiffies && !timer_pending(&qp->retrans_timer)) 559 mod_timer(&qp->retrans_timer, 560 jiffies + qp->qp_timeout_jiffies); 561 } 562 563 int rxe_requester(void *arg) 564 { 565 struct rxe_qp *qp = (struct rxe_qp *)arg; 566 struct rxe_pkt_info pkt; 567 struct sk_buff *skb; 568 struct rxe_send_wqe *wqe; 569 unsigned mask; 570 int payload; 571 int mtu; 572 int opcode; 573 int ret; 574 enum wqe_state prev_state; 575 576 next_wqe: 577 if (unlikely(!qp->valid || qp->req.state == QP_STATE_ERROR)) 578 goto exit; 579 580 if (unlikely(qp->req.state == QP_STATE_RESET)) { 581 qp->req.wqe_index = consumer_index(qp->sq.queue); 582 qp->req.opcode = -1; 583 qp->req.need_rd_atomic = 0; 584 qp->req.wait_psn = 0; 585 qp->req.need_retry = 0; 586 goto exit; 587 } 588 589 if (unlikely(qp->req.need_retry)) { 590 req_retry(qp); 591 qp->req.need_retry = 0; 592 } 593 594 wqe = req_next_wqe(qp); 595 if (unlikely(!wqe)) 596 goto exit; 597 598 if (wqe->mask & WR_REG_MASK) { 599 if (wqe->wr.opcode == IB_WR_LOCAL_INV) { 600 struct rxe_dev *rxe = to_rdev(qp->ibqp.device); 601 struct rxe_mem *rmr; 602 603 rmr = rxe_pool_get_index(&rxe->mr_pool, 604 wqe->wr.ex.invalidate_rkey >> 8); 605 if (!rmr) { 606 pr_err("No mr for key %#x\n", wqe->wr.ex.invalidate_rkey); 607 wqe->state = wqe_state_error; 608 wqe->status = IB_WC_MW_BIND_ERR; 609 goto exit; 610 } 611 rmr->state = RXE_MEM_STATE_FREE; 612 wqe->state = wqe_state_done; 613 wqe->status = IB_WC_SUCCESS; 614 } else if (wqe->wr.opcode == IB_WR_REG_MR) { 615 struct rxe_mem *rmr = to_rmr(wqe->wr.wr.reg.mr); 616 617 rmr->state = RXE_MEM_STATE_VALID; 618 rmr->access = wqe->wr.wr.reg.access; 619 rmr->lkey = wqe->wr.wr.reg.key; 620 rmr->rkey = wqe->wr.wr.reg.key; 621 wqe->state = wqe_state_done; 622 wqe->status = IB_WC_SUCCESS; 623 } else { 624 goto exit; 625 } 626 qp->req.wqe_index = next_index(qp->sq.queue, 627 qp->req.wqe_index); 628 goto next_wqe; 629 } 630 631 if (unlikely(qp_type(qp) == IB_QPT_RC && 632 qp->req.psn > (qp->comp.psn + RXE_MAX_UNACKED_PSNS))) { 633 qp->req.wait_psn = 1; 634 goto exit; 635 } 636 637 /* Limit the number of inflight SKBs per QP */ 638 if (unlikely(atomic_read(&qp->skb_out) > 639 RXE_INFLIGHT_SKBS_PER_QP_HIGH)) { 640 qp->need_req_skb = 1; 641 goto exit; 642 } 643 644 opcode = next_opcode(qp, wqe, wqe->wr.opcode); 645 if (unlikely(opcode < 0)) { 646 wqe->status = IB_WC_LOC_QP_OP_ERR; 647 goto exit; 648 } 649 650 mask = rxe_opcode[opcode].mask; 651 if (unlikely(mask & RXE_READ_OR_ATOMIC)) { 652 if (check_init_depth(qp, wqe)) 653 goto exit; 654 } 655 656 mtu = get_mtu(qp, wqe); 657 payload = (mask & RXE_WRITE_OR_SEND) ? wqe->dma.resid : 0; 658 if (payload > mtu) { 659 if (qp_type(qp) == IB_QPT_UD) { 660 /* C10-93.1.1: If the total sum of all the buffer lengths specified for a 661 * UD message exceeds the MTU of the port as returned by QueryHCA, the CI 662 * shall not emit any packets for this message. Further, the CI shall not 663 * generate an error due to this condition. 664 */ 665 666 /* fake a successful UD send */ 667 wqe->first_psn = qp->req.psn; 668 wqe->last_psn = qp->req.psn; 669 qp->req.psn = (qp->req.psn + 1) & BTH_PSN_MASK; 670 qp->req.opcode = IB_OPCODE_UD_SEND_ONLY; 671 qp->req.wqe_index = next_index(qp->sq.queue, 672 qp->req.wqe_index); 673 wqe->state = wqe_state_done; 674 wqe->status = IB_WC_SUCCESS; 675 goto complete; 676 } 677 payload = mtu; 678 } 679 680 skb = init_req_packet(qp, wqe, opcode, payload, &pkt); 681 if (unlikely(!skb)) { 682 pr_err("Failed allocating skb\n"); 683 goto err; 684 } 685 686 if (fill_packet(qp, wqe, &pkt, skb, payload)) { 687 pr_debug("Error during fill packet\n"); 688 goto err; 689 } 690 691 update_wqe_state(qp, wqe, &pkt, &prev_state); 692 ret = rxe_xmit_packet(to_rdev(qp->ibqp.device), qp, &pkt, skb); 693 if (ret) { 694 qp->need_req_skb = 1; 695 kfree_skb(skb); 696 697 wqe->state = prev_state; 698 699 if (ret == -EAGAIN) { 700 rxe_run_task(&qp->req.task, 1); 701 goto exit; 702 } 703 704 goto err; 705 } 706 707 update_state(qp, wqe, &pkt, payload); 708 709 goto next_wqe; 710 711 err: 712 kfree_skb(skb); 713 wqe->status = IB_WC_LOC_PROT_ERR; 714 wqe->state = wqe_state_error; 715 716 complete: 717 if (qp_type(qp) != IB_QPT_RC) { 718 while (rxe_completer(qp) == 0) 719 ; 720 } 721 722 return 0; 723 724 exit: 725 return -EAGAIN; 726 } 727