1 // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
2 /*
3 * Copyright (c) 2020, Mellanox Technologies inc. All rights reserved.
4 */
5
6 #include <linux/gfp.h>
7 #include <linux/mlx5/qp.h>
8 #include <linux/mlx5/driver.h>
9 #include "wr.h"
10 #include "umr.h"
11
12 static const u32 mlx5_ib_opcode[] = {
13 [IB_WR_SEND] = MLX5_OPCODE_SEND,
14 [IB_WR_LSO] = MLX5_OPCODE_LSO,
15 [IB_WR_SEND_WITH_IMM] = MLX5_OPCODE_SEND_IMM,
16 [IB_WR_RDMA_WRITE] = MLX5_OPCODE_RDMA_WRITE,
17 [IB_WR_RDMA_WRITE_WITH_IMM] = MLX5_OPCODE_RDMA_WRITE_IMM,
18 [IB_WR_RDMA_READ] = MLX5_OPCODE_RDMA_READ,
19 [IB_WR_ATOMIC_CMP_AND_SWP] = MLX5_OPCODE_ATOMIC_CS,
20 [IB_WR_ATOMIC_FETCH_AND_ADD] = MLX5_OPCODE_ATOMIC_FA,
21 [IB_WR_SEND_WITH_INV] = MLX5_OPCODE_SEND_INVAL,
22 [IB_WR_LOCAL_INV] = MLX5_OPCODE_UMR,
23 [IB_WR_REG_MR] = MLX5_OPCODE_UMR,
24 [IB_WR_MASKED_ATOMIC_CMP_AND_SWP] = MLX5_OPCODE_ATOMIC_MASKED_CS,
25 [IB_WR_MASKED_ATOMIC_FETCH_AND_ADD] = MLX5_OPCODE_ATOMIC_MASKED_FA,
26 [MLX5_IB_WR_UMR] = MLX5_OPCODE_UMR,
27 };
28
mlx5r_wq_overflow(struct mlx5_ib_wq * wq,int nreq,struct ib_cq * ib_cq)29 int mlx5r_wq_overflow(struct mlx5_ib_wq *wq, int nreq, struct ib_cq *ib_cq)
30 {
31 struct mlx5_ib_cq *cq;
32 unsigned int cur;
33
34 cur = wq->head - wq->tail;
35 if (likely(cur + nreq < wq->max_post))
36 return 0;
37
38 cq = to_mcq(ib_cq);
39 spin_lock(&cq->lock);
40 cur = wq->head - wq->tail;
41 spin_unlock(&cq->lock);
42
43 return cur + nreq >= wq->max_post;
44 }
45
set_raddr_seg(struct mlx5_wqe_raddr_seg * rseg,u64 remote_addr,u32 rkey)46 static __always_inline void set_raddr_seg(struct mlx5_wqe_raddr_seg *rseg,
47 u64 remote_addr, u32 rkey)
48 {
49 rseg->raddr = cpu_to_be64(remote_addr);
50 rseg->rkey = cpu_to_be32(rkey);
51 rseg->reserved = 0;
52 }
53
set_eth_seg(const struct ib_send_wr * wr,struct mlx5_ib_qp * qp,void ** seg,int * size,void ** cur_edge)54 static void set_eth_seg(const struct ib_send_wr *wr, struct mlx5_ib_qp *qp,
55 void **seg, int *size, void **cur_edge)
56 {
57 struct mlx5_wqe_eth_seg *eseg = *seg;
58
59 memset(eseg, 0, sizeof(struct mlx5_wqe_eth_seg));
60
61 if (wr->send_flags & IB_SEND_IP_CSUM)
62 eseg->cs_flags = MLX5_ETH_WQE_L3_CSUM |
63 MLX5_ETH_WQE_L4_CSUM;
64
65 if (wr->opcode == IB_WR_LSO) {
66 struct ib_ud_wr *ud_wr = container_of(wr, struct ib_ud_wr, wr);
67 size_t left, copysz;
68 void *pdata = ud_wr->header;
69 size_t stride;
70
71 left = ud_wr->hlen;
72 eseg->mss = cpu_to_be16(ud_wr->mss);
73 eseg->inline_hdr.sz = cpu_to_be16(left);
74
75 /* mlx5r_memcpy_send_wqe should get a 16B align address. Hence,
76 * we first copy up to the current edge and then, if needed,
77 * continue to mlx5r_memcpy_send_wqe.
78 */
79 copysz = min_t(u64, *cur_edge - (void *)eseg->inline_hdr.start,
80 left);
81 memcpy(eseg->inline_hdr.data, pdata, copysz);
82 stride = ALIGN(sizeof(struct mlx5_wqe_eth_seg) -
83 sizeof(eseg->inline_hdr.start) + copysz, 16);
84 *size += stride / 16;
85 *seg += stride;
86
87 if (copysz < left) {
88 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
89 left -= copysz;
90 pdata += copysz;
91 mlx5r_memcpy_send_wqe(&qp->sq, cur_edge, seg, size,
92 pdata, left);
93 }
94
95 return;
96 }
97
98 *seg += sizeof(struct mlx5_wqe_eth_seg);
99 *size += sizeof(struct mlx5_wqe_eth_seg) / 16;
100 }
101
set_datagram_seg(struct mlx5_wqe_datagram_seg * dseg,const struct ib_send_wr * wr)102 static void set_datagram_seg(struct mlx5_wqe_datagram_seg *dseg,
103 const struct ib_send_wr *wr)
104 {
105 memcpy(&dseg->av, &to_mah(ud_wr(wr)->ah)->av, sizeof(struct mlx5_av));
106 dseg->av.dqp_dct =
107 cpu_to_be32(ud_wr(wr)->remote_qpn | MLX5_EXTENDED_UD_AV);
108 dseg->av.key.qkey.qkey = cpu_to_be32(ud_wr(wr)->remote_qkey);
109 }
110
set_data_ptr_seg(struct mlx5_wqe_data_seg * dseg,struct ib_sge * sg)111 static void set_data_ptr_seg(struct mlx5_wqe_data_seg *dseg, struct ib_sge *sg)
112 {
113 dseg->byte_count = cpu_to_be32(sg->length);
114 dseg->lkey = cpu_to_be32(sg->lkey);
115 dseg->addr = cpu_to_be64(sg->addr);
116 }
117
frwr_mkey_mask(bool atomic)118 static __be64 frwr_mkey_mask(bool atomic)
119 {
120 u64 result;
121
122 result = MLX5_MKEY_MASK_LEN |
123 MLX5_MKEY_MASK_PAGE_SIZE |
124 MLX5_MKEY_MASK_START_ADDR |
125 MLX5_MKEY_MASK_EN_RINVAL |
126 MLX5_MKEY_MASK_KEY |
127 MLX5_MKEY_MASK_LR |
128 MLX5_MKEY_MASK_LW |
129 MLX5_MKEY_MASK_RR |
130 MLX5_MKEY_MASK_RW |
131 MLX5_MKEY_MASK_SMALL_FENCE |
132 MLX5_MKEY_MASK_FREE;
133
134 if (atomic)
135 result |= MLX5_MKEY_MASK_A;
136
137 return cpu_to_be64(result);
138 }
139
sig_mkey_mask(void)140 static __be64 sig_mkey_mask(void)
141 {
142 u64 result;
143
144 result = MLX5_MKEY_MASK_LEN |
145 MLX5_MKEY_MASK_PAGE_SIZE |
146 MLX5_MKEY_MASK_START_ADDR |
147 MLX5_MKEY_MASK_EN_SIGERR |
148 MLX5_MKEY_MASK_EN_RINVAL |
149 MLX5_MKEY_MASK_KEY |
150 MLX5_MKEY_MASK_LR |
151 MLX5_MKEY_MASK_LW |
152 MLX5_MKEY_MASK_RR |
153 MLX5_MKEY_MASK_RW |
154 MLX5_MKEY_MASK_SMALL_FENCE |
155 MLX5_MKEY_MASK_FREE |
156 MLX5_MKEY_MASK_BSF_EN;
157
158 return cpu_to_be64(result);
159 }
160
set_reg_umr_seg(struct mlx5_wqe_umr_ctrl_seg * umr,struct mlx5_ib_mr * mr,u8 flags,bool atomic)161 static void set_reg_umr_seg(struct mlx5_wqe_umr_ctrl_seg *umr,
162 struct mlx5_ib_mr *mr, u8 flags, bool atomic)
163 {
164 int size = (mr->mmkey.ndescs + mr->meta_ndescs) * mr->desc_size;
165
166 memset(umr, 0, sizeof(*umr));
167
168 umr->flags = flags;
169 umr->xlt_octowords = cpu_to_be16(mlx5r_umr_get_xlt_octo(size));
170 umr->mkey_mask = frwr_mkey_mask(atomic);
171 }
172
set_linv_umr_seg(struct mlx5_wqe_umr_ctrl_seg * umr)173 static void set_linv_umr_seg(struct mlx5_wqe_umr_ctrl_seg *umr)
174 {
175 memset(umr, 0, sizeof(*umr));
176 umr->mkey_mask = cpu_to_be64(MLX5_MKEY_MASK_FREE);
177 umr->flags = MLX5_UMR_INLINE;
178 }
179
get_umr_flags(int acc)180 static u8 get_umr_flags(int acc)
181 {
182 return (acc & IB_ACCESS_REMOTE_ATOMIC ? MLX5_PERM_ATOMIC : 0) |
183 (acc & IB_ACCESS_REMOTE_WRITE ? MLX5_PERM_REMOTE_WRITE : 0) |
184 (acc & IB_ACCESS_REMOTE_READ ? MLX5_PERM_REMOTE_READ : 0) |
185 (acc & IB_ACCESS_LOCAL_WRITE ? MLX5_PERM_LOCAL_WRITE : 0) |
186 MLX5_PERM_LOCAL_READ | MLX5_PERM_UMR_EN;
187 }
188
set_reg_mkey_seg(struct mlx5_mkey_seg * seg,struct mlx5_ib_mr * mr,u32 key,int access)189 static void set_reg_mkey_seg(struct mlx5_mkey_seg *seg,
190 struct mlx5_ib_mr *mr,
191 u32 key, int access)
192 {
193 int ndescs = ALIGN(mr->mmkey.ndescs + mr->meta_ndescs, 8) >> 1;
194
195 memset(seg, 0, sizeof(*seg));
196
197 if (mr->access_mode == MLX5_MKC_ACCESS_MODE_MTT)
198 seg->log2_page_size = ilog2(mr->ibmr.page_size);
199 else if (mr->access_mode == MLX5_MKC_ACCESS_MODE_KLMS)
200 /* KLMs take twice the size of MTTs */
201 ndescs *= 2;
202
203 seg->flags = get_umr_flags(access) | mr->access_mode;
204 seg->qpn_mkey7_0 = cpu_to_be32((key & 0xff) | 0xffffff00);
205 seg->flags_pd = cpu_to_be32(MLX5_MKEY_REMOTE_INVAL);
206 seg->start_addr = cpu_to_be64(mr->ibmr.iova);
207 seg->len = cpu_to_be64(mr->ibmr.length);
208 seg->xlt_oct_size = cpu_to_be32(ndescs);
209 }
210
set_linv_mkey_seg(struct mlx5_mkey_seg * seg)211 static void set_linv_mkey_seg(struct mlx5_mkey_seg *seg)
212 {
213 memset(seg, 0, sizeof(*seg));
214 seg->status = MLX5_MKEY_STATUS_FREE;
215 }
216
set_reg_data_seg(struct mlx5_wqe_data_seg * dseg,struct mlx5_ib_mr * mr,struct mlx5_ib_pd * pd)217 static void set_reg_data_seg(struct mlx5_wqe_data_seg *dseg,
218 struct mlx5_ib_mr *mr,
219 struct mlx5_ib_pd *pd)
220 {
221 int bcount = mr->desc_size * (mr->mmkey.ndescs + mr->meta_ndescs);
222
223 dseg->addr = cpu_to_be64(mr->desc_map);
224 dseg->byte_count = cpu_to_be32(ALIGN(bcount, 64));
225 dseg->lkey = cpu_to_be32(pd->ibpd.local_dma_lkey);
226 }
227
send_ieth(const struct ib_send_wr * wr)228 static __be32 send_ieth(const struct ib_send_wr *wr)
229 {
230 switch (wr->opcode) {
231 case IB_WR_SEND_WITH_IMM:
232 case IB_WR_RDMA_WRITE_WITH_IMM:
233 return wr->ex.imm_data;
234
235 case IB_WR_SEND_WITH_INV:
236 return cpu_to_be32(wr->ex.invalidate_rkey);
237
238 default:
239 return 0;
240 }
241 }
242
calc_sig(void * wqe,int size)243 static u8 calc_sig(void *wqe, int size)
244 {
245 u8 *p = wqe;
246 u8 res = 0;
247 int i;
248
249 for (i = 0; i < size; i++)
250 res ^= p[i];
251
252 return ~res;
253 }
254
wq_sig(void * wqe)255 static u8 wq_sig(void *wqe)
256 {
257 return calc_sig(wqe, (*((u8 *)wqe + 8) & 0x3f) << 4);
258 }
259
set_data_inl_seg(struct mlx5_ib_qp * qp,const struct ib_send_wr * wr,void ** wqe,int * wqe_sz,void ** cur_edge)260 static int set_data_inl_seg(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr,
261 void **wqe, int *wqe_sz, void **cur_edge)
262 {
263 struct mlx5_wqe_inline_seg *seg;
264 size_t offset;
265 int inl = 0;
266 int i;
267
268 seg = *wqe;
269 *wqe += sizeof(*seg);
270 offset = sizeof(*seg);
271
272 for (i = 0; i < wr->num_sge; i++) {
273 size_t len = wr->sg_list[i].length;
274 void *addr = (void *)(unsigned long)(wr->sg_list[i].addr);
275
276 inl += len;
277
278 if (unlikely(inl > qp->max_inline_data))
279 return -ENOMEM;
280
281 while (likely(len)) {
282 size_t leftlen;
283 size_t copysz;
284
285 handle_post_send_edge(&qp->sq, wqe,
286 *wqe_sz + (offset >> 4),
287 cur_edge);
288
289 leftlen = *cur_edge - *wqe;
290 copysz = min_t(size_t, leftlen, len);
291
292 memcpy(*wqe, addr, copysz);
293 len -= copysz;
294 addr += copysz;
295 *wqe += copysz;
296 offset += copysz;
297 }
298 }
299
300 seg->byte_count = cpu_to_be32(inl | MLX5_INLINE_SEG);
301
302 *wqe_sz += ALIGN(inl + sizeof(seg->byte_count), 16) / 16;
303
304 return 0;
305 }
306
prot_field_size(enum ib_signature_type type)307 static u16 prot_field_size(enum ib_signature_type type)
308 {
309 switch (type) {
310 case IB_SIG_TYPE_T10_DIF:
311 return MLX5_DIF_SIZE;
312 default:
313 return 0;
314 }
315 }
316
bs_selector(int block_size)317 static u8 bs_selector(int block_size)
318 {
319 switch (block_size) {
320 case 512: return 0x1;
321 case 520: return 0x2;
322 case 4096: return 0x3;
323 case 4160: return 0x4;
324 case 1073741824: return 0x5;
325 default: return 0;
326 }
327 }
328
mlx5_fill_inl_bsf(struct ib_sig_domain * domain,struct mlx5_bsf_inl * inl)329 static void mlx5_fill_inl_bsf(struct ib_sig_domain *domain,
330 struct mlx5_bsf_inl *inl)
331 {
332 /* Valid inline section and allow BSF refresh */
333 inl->vld_refresh = cpu_to_be16(MLX5_BSF_INL_VALID |
334 MLX5_BSF_REFRESH_DIF);
335 inl->dif_apptag = cpu_to_be16(domain->sig.dif.app_tag);
336 inl->dif_reftag = cpu_to_be32(domain->sig.dif.ref_tag);
337 /* repeating block */
338 inl->rp_inv_seed = MLX5_BSF_REPEAT_BLOCK;
339 inl->sig_type = domain->sig.dif.bg_type == IB_T10DIF_CRC ?
340 MLX5_DIF_CRC : MLX5_DIF_IPCS;
341
342 if (domain->sig.dif.ref_remap)
343 inl->dif_inc_ref_guard_check |= MLX5_BSF_INC_REFTAG;
344
345 if (domain->sig.dif.app_escape) {
346 if (domain->sig.dif.ref_escape)
347 inl->dif_inc_ref_guard_check |= MLX5_BSF_APPREF_ESCAPE;
348 else
349 inl->dif_inc_ref_guard_check |= MLX5_BSF_APPTAG_ESCAPE;
350 }
351
352 inl->dif_app_bitmask_check =
353 cpu_to_be16(domain->sig.dif.apptag_check_mask);
354 }
355
mlx5_set_bsf(struct ib_mr * sig_mr,struct ib_sig_attrs * sig_attrs,struct mlx5_bsf * bsf,u32 data_size)356 static int mlx5_set_bsf(struct ib_mr *sig_mr,
357 struct ib_sig_attrs *sig_attrs,
358 struct mlx5_bsf *bsf, u32 data_size)
359 {
360 struct mlx5_core_sig_ctx *msig = to_mmr(sig_mr)->sig;
361 struct mlx5_bsf_basic *basic = &bsf->basic;
362 struct ib_sig_domain *mem = &sig_attrs->mem;
363 struct ib_sig_domain *wire = &sig_attrs->wire;
364
365 memset(bsf, 0, sizeof(*bsf));
366
367 /* Basic + Extended + Inline */
368 basic->bsf_size_sbs = 1 << 7;
369 /* Input domain check byte mask */
370 basic->check_byte_mask = sig_attrs->check_mask;
371 basic->raw_data_size = cpu_to_be32(data_size);
372
373 /* Memory domain */
374 switch (sig_attrs->mem.sig_type) {
375 case IB_SIG_TYPE_NONE:
376 break;
377 case IB_SIG_TYPE_T10_DIF:
378 basic->mem.bs_selector = bs_selector(mem->sig.dif.pi_interval);
379 basic->m_bfs_psv = cpu_to_be32(msig->psv_memory.psv_idx);
380 mlx5_fill_inl_bsf(mem, &bsf->m_inl);
381 break;
382 default:
383 return -EINVAL;
384 }
385
386 /* Wire domain */
387 switch (sig_attrs->wire.sig_type) {
388 case IB_SIG_TYPE_NONE:
389 break;
390 case IB_SIG_TYPE_T10_DIF:
391 if (mem->sig.dif.pi_interval == wire->sig.dif.pi_interval &&
392 mem->sig_type == wire->sig_type) {
393 /* Same block structure */
394 basic->bsf_size_sbs |= 1 << 4;
395 if (mem->sig.dif.bg_type == wire->sig.dif.bg_type)
396 basic->wire.copy_byte_mask |= MLX5_CPY_GRD_MASK;
397 if (mem->sig.dif.app_tag == wire->sig.dif.app_tag)
398 basic->wire.copy_byte_mask |= MLX5_CPY_APP_MASK;
399 if (mem->sig.dif.ref_tag == wire->sig.dif.ref_tag)
400 basic->wire.copy_byte_mask |= MLX5_CPY_REF_MASK;
401 } else
402 basic->wire.bs_selector =
403 bs_selector(wire->sig.dif.pi_interval);
404
405 basic->w_bfs_psv = cpu_to_be32(msig->psv_wire.psv_idx);
406 mlx5_fill_inl_bsf(wire, &bsf->w_inl);
407 break;
408 default:
409 return -EINVAL;
410 }
411
412 return 0;
413 }
414
415
set_sig_data_segment(const struct ib_send_wr * send_wr,struct ib_mr * sig_mr,struct ib_sig_attrs * sig_attrs,struct mlx5_ib_qp * qp,void ** seg,int * size,void ** cur_edge)416 static int set_sig_data_segment(const struct ib_send_wr *send_wr,
417 struct ib_mr *sig_mr,
418 struct ib_sig_attrs *sig_attrs,
419 struct mlx5_ib_qp *qp, void **seg, int *size,
420 void **cur_edge)
421 {
422 struct mlx5_bsf *bsf;
423 u32 data_len;
424 u32 data_key;
425 u64 data_va;
426 u32 prot_len = 0;
427 u32 prot_key = 0;
428 u64 prot_va = 0;
429 bool prot = false;
430 int ret;
431 int wqe_size;
432 struct mlx5_ib_mr *mr = to_mmr(sig_mr);
433 struct mlx5_ib_mr *pi_mr = mr->pi_mr;
434
435 data_len = pi_mr->data_length;
436 data_key = pi_mr->ibmr.lkey;
437 data_va = pi_mr->data_iova;
438 if (pi_mr->meta_ndescs) {
439 prot_len = pi_mr->meta_length;
440 prot_key = pi_mr->ibmr.lkey;
441 prot_va = pi_mr->pi_iova;
442 prot = true;
443 }
444
445 if (!prot || (data_key == prot_key && data_va == prot_va &&
446 data_len == prot_len)) {
447 /**
448 * Source domain doesn't contain signature information
449 * or data and protection are interleaved in memory.
450 * So need construct:
451 * ------------------
452 * | data_klm |
453 * ------------------
454 * | BSF |
455 * ------------------
456 **/
457 struct mlx5_klm *data_klm = *seg;
458
459 data_klm->bcount = cpu_to_be32(data_len);
460 data_klm->key = cpu_to_be32(data_key);
461 data_klm->va = cpu_to_be64(data_va);
462 wqe_size = ALIGN(sizeof(*data_klm), 64);
463 } else {
464 /**
465 * Source domain contains signature information
466 * So need construct a strided block format:
467 * ---------------------------
468 * | stride_block_ctrl |
469 * ---------------------------
470 * | data_klm |
471 * ---------------------------
472 * | prot_klm |
473 * ---------------------------
474 * | BSF |
475 * ---------------------------
476 **/
477 struct mlx5_stride_block_ctrl_seg *sblock_ctrl;
478 struct mlx5_stride_block_entry *data_sentry;
479 struct mlx5_stride_block_entry *prot_sentry;
480 u16 block_size = sig_attrs->mem.sig.dif.pi_interval;
481 int prot_size;
482
483 sblock_ctrl = *seg;
484 data_sentry = (void *)sblock_ctrl + sizeof(*sblock_ctrl);
485 prot_sentry = (void *)data_sentry + sizeof(*data_sentry);
486
487 prot_size = prot_field_size(sig_attrs->mem.sig_type);
488 if (!prot_size) {
489 pr_err("Bad block size given: %u\n", block_size);
490 return -EINVAL;
491 }
492 sblock_ctrl->bcount_per_cycle = cpu_to_be32(block_size +
493 prot_size);
494 sblock_ctrl->op = cpu_to_be32(MLX5_STRIDE_BLOCK_OP);
495 sblock_ctrl->repeat_count = cpu_to_be32(data_len / block_size);
496 sblock_ctrl->num_entries = cpu_to_be16(2);
497
498 data_sentry->bcount = cpu_to_be16(block_size);
499 data_sentry->key = cpu_to_be32(data_key);
500 data_sentry->va = cpu_to_be64(data_va);
501 data_sentry->stride = cpu_to_be16(block_size);
502
503 prot_sentry->bcount = cpu_to_be16(prot_size);
504 prot_sentry->key = cpu_to_be32(prot_key);
505 prot_sentry->va = cpu_to_be64(prot_va);
506 prot_sentry->stride = cpu_to_be16(prot_size);
507
508 wqe_size = ALIGN(sizeof(*sblock_ctrl) + sizeof(*data_sentry) +
509 sizeof(*prot_sentry), 64);
510 }
511
512 *seg += wqe_size;
513 *size += wqe_size / 16;
514 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
515
516 bsf = *seg;
517 ret = mlx5_set_bsf(sig_mr, sig_attrs, bsf, data_len);
518 if (ret)
519 return -EINVAL;
520
521 *seg += sizeof(*bsf);
522 *size += sizeof(*bsf) / 16;
523 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
524
525 return 0;
526 }
527
set_sig_mkey_segment(struct mlx5_mkey_seg * seg,struct ib_mr * sig_mr,int access_flags,u32 size,u32 length,u32 pdn)528 static void set_sig_mkey_segment(struct mlx5_mkey_seg *seg,
529 struct ib_mr *sig_mr, int access_flags,
530 u32 size, u32 length, u32 pdn)
531 {
532 u32 sig_key = sig_mr->rkey;
533 u8 sigerr = to_mmr(sig_mr)->sig->sigerr_count & 1;
534
535 memset(seg, 0, sizeof(*seg));
536
537 seg->flags = get_umr_flags(access_flags) | MLX5_MKC_ACCESS_MODE_KLMS;
538 seg->qpn_mkey7_0 = cpu_to_be32((sig_key & 0xff) | 0xffffff00);
539 seg->flags_pd = cpu_to_be32(MLX5_MKEY_REMOTE_INVAL | sigerr << 26 |
540 MLX5_MKEY_BSF_EN | pdn);
541 seg->len = cpu_to_be64(length);
542 seg->xlt_oct_size = cpu_to_be32(mlx5r_umr_get_xlt_octo(size));
543 seg->bsfs_octo_size = cpu_to_be32(MLX5_MKEY_BSF_OCTO_SIZE);
544 }
545
set_sig_umr_segment(struct mlx5_wqe_umr_ctrl_seg * umr,u32 size)546 static void set_sig_umr_segment(struct mlx5_wqe_umr_ctrl_seg *umr,
547 u32 size)
548 {
549 memset(umr, 0, sizeof(*umr));
550
551 umr->flags = MLX5_FLAGS_INLINE | MLX5_FLAGS_CHECK_FREE;
552 umr->xlt_octowords = cpu_to_be16(mlx5r_umr_get_xlt_octo(size));
553 umr->bsf_octowords = cpu_to_be16(MLX5_MKEY_BSF_OCTO_SIZE);
554 umr->mkey_mask = sig_mkey_mask();
555 }
556
set_pi_umr_wr(const struct ib_send_wr * send_wr,struct mlx5_ib_qp * qp,void ** seg,int * size,void ** cur_edge)557 static int set_pi_umr_wr(const struct ib_send_wr *send_wr,
558 struct mlx5_ib_qp *qp, void **seg, int *size,
559 void **cur_edge)
560 {
561 const struct ib_reg_wr *wr = reg_wr(send_wr);
562 struct mlx5_ib_mr *sig_mr = to_mmr(wr->mr);
563 struct mlx5_ib_mr *pi_mr = sig_mr->pi_mr;
564 struct ib_sig_attrs *sig_attrs = sig_mr->ibmr.sig_attrs;
565 u32 pdn = to_mpd(qp->ibqp.pd)->pdn;
566 u32 xlt_size;
567 int region_len, ret;
568
569 if (unlikely(send_wr->num_sge != 0) ||
570 unlikely(wr->access & IB_ACCESS_REMOTE_ATOMIC) ||
571 unlikely(!sig_mr->sig) || unlikely(!qp->ibqp.integrity_en) ||
572 unlikely(!sig_mr->sig->sig_status_checked))
573 return -EINVAL;
574
575 /* length of the protected region, data + protection */
576 region_len = pi_mr->ibmr.length;
577
578 /**
579 * KLM octoword size - if protection was provided
580 * then we use strided block format (3 octowords),
581 * else we use single KLM (1 octoword)
582 **/
583 if (sig_attrs->mem.sig_type != IB_SIG_TYPE_NONE)
584 xlt_size = 0x30;
585 else
586 xlt_size = sizeof(struct mlx5_klm);
587
588 set_sig_umr_segment(*seg, xlt_size);
589 *seg += sizeof(struct mlx5_wqe_umr_ctrl_seg);
590 *size += sizeof(struct mlx5_wqe_umr_ctrl_seg) / 16;
591 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
592
593 set_sig_mkey_segment(*seg, wr->mr, wr->access, xlt_size, region_len,
594 pdn);
595 *seg += sizeof(struct mlx5_mkey_seg);
596 *size += sizeof(struct mlx5_mkey_seg) / 16;
597 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
598
599 ret = set_sig_data_segment(send_wr, wr->mr, sig_attrs, qp, seg, size,
600 cur_edge);
601 if (ret)
602 return ret;
603
604 sig_mr->sig->sig_status_checked = false;
605 return 0;
606 }
607
set_psv_wr(struct ib_sig_domain * domain,u32 psv_idx,void ** seg,int * size)608 static int set_psv_wr(struct ib_sig_domain *domain,
609 u32 psv_idx, void **seg, int *size)
610 {
611 struct mlx5_seg_set_psv *psv_seg = *seg;
612
613 memset(psv_seg, 0, sizeof(*psv_seg));
614 psv_seg->psv_num = cpu_to_be32(psv_idx);
615 switch (domain->sig_type) {
616 case IB_SIG_TYPE_NONE:
617 break;
618 case IB_SIG_TYPE_T10_DIF:
619 psv_seg->transient_sig = cpu_to_be32(domain->sig.dif.bg << 16 |
620 domain->sig.dif.app_tag);
621 psv_seg->ref_tag = cpu_to_be32(domain->sig.dif.ref_tag);
622 break;
623 default:
624 pr_err("Bad signature type (%d) is given.\n",
625 domain->sig_type);
626 return -EINVAL;
627 }
628
629 *seg += sizeof(*psv_seg);
630 *size += sizeof(*psv_seg) / 16;
631
632 return 0;
633 }
634
set_reg_wr(struct mlx5_ib_qp * qp,const struct ib_reg_wr * wr,void ** seg,int * size,void ** cur_edge,bool check_not_free)635 static int set_reg_wr(struct mlx5_ib_qp *qp,
636 const struct ib_reg_wr *wr,
637 void **seg, int *size, void **cur_edge,
638 bool check_not_free)
639 {
640 struct mlx5_ib_mr *mr = to_mmr(wr->mr);
641 struct mlx5_ib_pd *pd = to_mpd(qp->ibqp.pd);
642 struct mlx5_ib_dev *dev = to_mdev(pd->ibpd.device);
643 int mr_list_size = (mr->mmkey.ndescs + mr->meta_ndescs) * mr->desc_size;
644 bool umr_inline = mr_list_size <= MLX5_IB_SQ_UMR_INLINE_THRESHOLD;
645 bool atomic = wr->access & IB_ACCESS_REMOTE_ATOMIC;
646 u8 flags = 0;
647
648 /* Matches access in mlx5_set_umr_free_mkey().
649 * Relaxed Ordering is set implicitly in mlx5_set_umr_free_mkey() and
650 * kernel ULPs are not aware of it, so we don't set it here.
651 */
652 if (!mlx5r_umr_can_reconfig(dev, 0, wr->access)) {
653 mlx5_ib_warn(
654 to_mdev(qp->ibqp.device),
655 "Fast update for MR access flags is not possible\n");
656 return -EINVAL;
657 }
658
659 if (unlikely(wr->wr.send_flags & IB_SEND_INLINE)) {
660 mlx5_ib_warn(to_mdev(qp->ibqp.device),
661 "Invalid IB_SEND_INLINE send flag\n");
662 return -EINVAL;
663 }
664
665 if (check_not_free)
666 flags |= MLX5_UMR_CHECK_NOT_FREE;
667 if (umr_inline)
668 flags |= MLX5_UMR_INLINE;
669
670 set_reg_umr_seg(*seg, mr, flags, atomic);
671 *seg += sizeof(struct mlx5_wqe_umr_ctrl_seg);
672 *size += sizeof(struct mlx5_wqe_umr_ctrl_seg) / 16;
673 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
674
675 set_reg_mkey_seg(*seg, mr, wr->key, wr->access);
676 *seg += sizeof(struct mlx5_mkey_seg);
677 *size += sizeof(struct mlx5_mkey_seg) / 16;
678 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
679
680 if (umr_inline) {
681 mlx5r_memcpy_send_wqe(&qp->sq, cur_edge, seg, size, mr->descs,
682 mr_list_size);
683 *size = ALIGN(*size, MLX5_SEND_WQE_BB >> 4);
684 } else {
685 set_reg_data_seg(*seg, mr, pd);
686 *seg += sizeof(struct mlx5_wqe_data_seg);
687 *size += (sizeof(struct mlx5_wqe_data_seg) / 16);
688 }
689 return 0;
690 }
691
set_linv_wr(struct mlx5_ib_qp * qp,void ** seg,int * size,void ** cur_edge)692 static void set_linv_wr(struct mlx5_ib_qp *qp, void **seg, int *size,
693 void **cur_edge)
694 {
695 set_linv_umr_seg(*seg);
696 *seg += sizeof(struct mlx5_wqe_umr_ctrl_seg);
697 *size += sizeof(struct mlx5_wqe_umr_ctrl_seg) / 16;
698 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
699 set_linv_mkey_seg(*seg);
700 *seg += sizeof(struct mlx5_mkey_seg);
701 *size += sizeof(struct mlx5_mkey_seg) / 16;
702 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
703 }
704
dump_wqe(struct mlx5_ib_qp * qp,u32 idx,int size_16)705 static void dump_wqe(struct mlx5_ib_qp *qp, u32 idx, int size_16)
706 {
707 __be32 *p = NULL;
708 int i, j;
709
710 pr_debug("dump WQE index %u:\n", idx);
711 for (i = 0, j = 0; i < size_16 * 4; i += 4, j += 4) {
712 if ((i & 0xf) == 0) {
713 p = mlx5_frag_buf_get_wqe(&qp->sq.fbc, idx);
714 pr_debug("WQBB at %p:\n", (void *)p);
715 j = 0;
716 idx = (idx + 1) & (qp->sq.wqe_cnt - 1);
717 }
718 pr_debug("%08x %08x %08x %08x\n", be32_to_cpu(p[j]),
719 be32_to_cpu(p[j + 1]), be32_to_cpu(p[j + 2]),
720 be32_to_cpu(p[j + 3]));
721 }
722 }
723
mlx5r_begin_wqe(struct mlx5_ib_qp * qp,void ** seg,struct mlx5_wqe_ctrl_seg ** ctrl,unsigned int * idx,int * size,void ** cur_edge,int nreq,__be32 general_id,bool send_signaled,bool solicited)724 int mlx5r_begin_wqe(struct mlx5_ib_qp *qp, void **seg,
725 struct mlx5_wqe_ctrl_seg **ctrl, unsigned int *idx,
726 int *size, void **cur_edge, int nreq, __be32 general_id,
727 bool send_signaled, bool solicited)
728 {
729 if (unlikely(mlx5r_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)))
730 return -ENOMEM;
731
732 *idx = qp->sq.cur_post & (qp->sq.wqe_cnt - 1);
733 *seg = mlx5_frag_buf_get_wqe(&qp->sq.fbc, *idx);
734 *ctrl = *seg;
735 *(uint32_t *)(*seg + 8) = 0;
736 (*ctrl)->general_id = general_id;
737 (*ctrl)->fm_ce_se = qp->sq_signal_bits |
738 (send_signaled ? MLX5_WQE_CTRL_CQ_UPDATE : 0) |
739 (solicited ? MLX5_WQE_CTRL_SOLICITED : 0);
740
741 *seg += sizeof(**ctrl);
742 *size = sizeof(**ctrl) / 16;
743 *cur_edge = qp->sq.cur_edge;
744
745 return 0;
746 }
747
begin_wqe(struct mlx5_ib_qp * qp,void ** seg,struct mlx5_wqe_ctrl_seg ** ctrl,const struct ib_send_wr * wr,unsigned int * idx,int * size,void ** cur_edge,int nreq)748 static int begin_wqe(struct mlx5_ib_qp *qp, void **seg,
749 struct mlx5_wqe_ctrl_seg **ctrl,
750 const struct ib_send_wr *wr, unsigned int *idx, int *size,
751 void **cur_edge, int nreq)
752 {
753 return mlx5r_begin_wqe(qp, seg, ctrl, idx, size, cur_edge, nreq,
754 send_ieth(wr), wr->send_flags & IB_SEND_SIGNALED,
755 wr->send_flags & IB_SEND_SOLICITED);
756 }
757
mlx5r_finish_wqe(struct mlx5_ib_qp * qp,struct mlx5_wqe_ctrl_seg * ctrl,void * seg,u8 size,void * cur_edge,unsigned int idx,u64 wr_id,int nreq,u8 fence,u32 mlx5_opcode)758 void mlx5r_finish_wqe(struct mlx5_ib_qp *qp, struct mlx5_wqe_ctrl_seg *ctrl,
759 void *seg, u8 size, void *cur_edge, unsigned int idx,
760 u64 wr_id, int nreq, u8 fence, u32 mlx5_opcode)
761 {
762 u8 opmod = 0;
763
764 ctrl->opmod_idx_opcode = cpu_to_be32(((u32)(qp->sq.cur_post) << 8) |
765 mlx5_opcode | ((u32)opmod << 24));
766 ctrl->qpn_ds = cpu_to_be32(size | (qp->trans_qp.base.mqp.qpn << 8));
767 ctrl->fm_ce_se |= fence;
768 if (unlikely(qp->flags_en & MLX5_QP_FLAG_SIGNATURE))
769 ctrl->signature = wq_sig(ctrl);
770
771 qp->sq.wrid[idx] = wr_id;
772 qp->sq.w_list[idx].opcode = mlx5_opcode;
773 qp->sq.wqe_head[idx] = qp->sq.head + nreq;
774 qp->sq.cur_post += DIV_ROUND_UP(size * 16, MLX5_SEND_WQE_BB);
775 qp->sq.w_list[idx].next = qp->sq.cur_post;
776
777 /* We save the edge which was possibly updated during the WQE
778 * construction, into SQ's cache.
779 */
780 seg = PTR_ALIGN(seg, MLX5_SEND_WQE_BB);
781 qp->sq.cur_edge = (unlikely(seg == cur_edge)) ?
782 get_sq_edge(&qp->sq, qp->sq.cur_post &
783 (qp->sq.wqe_cnt - 1)) :
784 cur_edge;
785 }
786
handle_rdma_op(const struct ib_send_wr * wr,void ** seg,int * size)787 static void handle_rdma_op(const struct ib_send_wr *wr, void **seg, int *size)
788 {
789 set_raddr_seg(*seg, rdma_wr(wr)->remote_addr, rdma_wr(wr)->rkey);
790 *seg += sizeof(struct mlx5_wqe_raddr_seg);
791 *size += sizeof(struct mlx5_wqe_raddr_seg) / 16;
792 }
793
handle_local_inv(struct mlx5_ib_qp * qp,const struct ib_send_wr * wr,struct mlx5_wqe_ctrl_seg ** ctrl,void ** seg,int * size,void ** cur_edge,unsigned int idx)794 static void handle_local_inv(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr,
795 struct mlx5_wqe_ctrl_seg **ctrl, void **seg,
796 int *size, void **cur_edge, unsigned int idx)
797 {
798 qp->sq.wr_data[idx] = IB_WR_LOCAL_INV;
799 (*ctrl)->imm = cpu_to_be32(wr->ex.invalidate_rkey);
800 set_linv_wr(qp, seg, size, cur_edge);
801 }
802
handle_reg_mr(struct mlx5_ib_qp * qp,const struct ib_send_wr * wr,struct mlx5_wqe_ctrl_seg ** ctrl,void ** seg,int * size,void ** cur_edge,unsigned int idx)803 static int handle_reg_mr(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr,
804 struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size,
805 void **cur_edge, unsigned int idx)
806 {
807 qp->sq.wr_data[idx] = IB_WR_REG_MR;
808 (*ctrl)->imm = cpu_to_be32(reg_wr(wr)->key);
809 return set_reg_wr(qp, reg_wr(wr), seg, size, cur_edge, true);
810 }
811
handle_psv(struct mlx5_ib_dev * dev,struct mlx5_ib_qp * qp,const struct ib_send_wr * wr,struct mlx5_wqe_ctrl_seg ** ctrl,void ** seg,int * size,void ** cur_edge,unsigned int * idx,int nreq,struct ib_sig_domain * domain,u32 psv_index,u8 next_fence)812 static int handle_psv(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp,
813 const struct ib_send_wr *wr,
814 struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size,
815 void **cur_edge, unsigned int *idx, int nreq,
816 struct ib_sig_domain *domain, u32 psv_index,
817 u8 next_fence)
818 {
819 int err;
820
821 /*
822 * SET_PSV WQEs are not signaled and solicited on error.
823 */
824 err = mlx5r_begin_wqe(qp, seg, ctrl, idx, size, cur_edge, nreq,
825 send_ieth(wr), false, true);
826 if (unlikely(err)) {
827 mlx5_ib_warn(dev, "\n");
828 err = -ENOMEM;
829 goto out;
830 }
831 err = set_psv_wr(domain, psv_index, seg, size);
832 if (unlikely(err)) {
833 mlx5_ib_warn(dev, "\n");
834 goto out;
835 }
836 mlx5r_finish_wqe(qp, *ctrl, *seg, *size, *cur_edge, *idx, wr->wr_id,
837 nreq, next_fence, MLX5_OPCODE_SET_PSV);
838
839 out:
840 return err;
841 }
842
handle_reg_mr_integrity(struct mlx5_ib_dev * dev,struct mlx5_ib_qp * qp,const struct ib_send_wr * wr,struct mlx5_wqe_ctrl_seg ** ctrl,void ** seg,int * size,void ** cur_edge,unsigned int * idx,int nreq,u8 fence,u8 next_fence)843 static int handle_reg_mr_integrity(struct mlx5_ib_dev *dev,
844 struct mlx5_ib_qp *qp,
845 const struct ib_send_wr *wr,
846 struct mlx5_wqe_ctrl_seg **ctrl, void **seg,
847 int *size, void **cur_edge,
848 unsigned int *idx, int nreq, u8 fence,
849 u8 next_fence)
850 {
851 struct mlx5_ib_mr *mr;
852 struct mlx5_ib_mr *pi_mr;
853 struct mlx5_ib_mr pa_pi_mr;
854 struct ib_sig_attrs *sig_attrs;
855 struct ib_reg_wr reg_pi_wr;
856 int err;
857
858 qp->sq.wr_data[*idx] = IB_WR_REG_MR_INTEGRITY;
859
860 mr = to_mmr(reg_wr(wr)->mr);
861 pi_mr = mr->pi_mr;
862
863 if (pi_mr) {
864 memset(®_pi_wr, 0,
865 sizeof(struct ib_reg_wr));
866
867 reg_pi_wr.mr = &pi_mr->ibmr;
868 reg_pi_wr.access = reg_wr(wr)->access;
869 reg_pi_wr.key = pi_mr->ibmr.rkey;
870
871 (*ctrl)->imm = cpu_to_be32(reg_pi_wr.key);
872 /* UMR for data + prot registration */
873 err = set_reg_wr(qp, ®_pi_wr, seg, size, cur_edge, false);
874 if (unlikely(err))
875 goto out;
876
877 mlx5r_finish_wqe(qp, *ctrl, *seg, *size, *cur_edge, *idx,
878 wr->wr_id, nreq, fence, MLX5_OPCODE_UMR);
879
880 err = begin_wqe(qp, seg, ctrl, wr, idx, size, cur_edge, nreq);
881 if (unlikely(err)) {
882 mlx5_ib_warn(dev, "\n");
883 err = -ENOMEM;
884 goto out;
885 }
886 } else {
887 memset(&pa_pi_mr, 0, sizeof(struct mlx5_ib_mr));
888 /* No UMR, use local_dma_lkey */
889 pa_pi_mr.ibmr.lkey = mr->ibmr.pd->local_dma_lkey;
890 pa_pi_mr.mmkey.ndescs = mr->mmkey.ndescs;
891 pa_pi_mr.data_length = mr->data_length;
892 pa_pi_mr.data_iova = mr->data_iova;
893 if (mr->meta_ndescs) {
894 pa_pi_mr.meta_ndescs = mr->meta_ndescs;
895 pa_pi_mr.meta_length = mr->meta_length;
896 pa_pi_mr.pi_iova = mr->pi_iova;
897 }
898
899 pa_pi_mr.ibmr.length = mr->ibmr.length;
900 mr->pi_mr = &pa_pi_mr;
901 }
902 (*ctrl)->imm = cpu_to_be32(mr->ibmr.rkey);
903 /* UMR for sig MR */
904 err = set_pi_umr_wr(wr, qp, seg, size, cur_edge);
905 if (unlikely(err)) {
906 mlx5_ib_warn(dev, "\n");
907 goto out;
908 }
909 mlx5r_finish_wqe(qp, *ctrl, *seg, *size, *cur_edge, *idx, wr->wr_id,
910 nreq, fence, MLX5_OPCODE_UMR);
911
912 sig_attrs = mr->ibmr.sig_attrs;
913 err = handle_psv(dev, qp, wr, ctrl, seg, size, cur_edge, idx, nreq,
914 &sig_attrs->mem, mr->sig->psv_memory.psv_idx,
915 next_fence);
916 if (unlikely(err))
917 goto out;
918
919 err = handle_psv(dev, qp, wr, ctrl, seg, size, cur_edge, idx, nreq,
920 &sig_attrs->wire, mr->sig->psv_wire.psv_idx,
921 next_fence);
922 if (unlikely(err))
923 goto out;
924
925 qp->next_fence = MLX5_FENCE_MODE_INITIATOR_SMALL;
926
927 out:
928 return err;
929 }
930
handle_qpt_rc(struct mlx5_ib_dev * dev,struct mlx5_ib_qp * qp,const struct ib_send_wr * wr,struct mlx5_wqe_ctrl_seg ** ctrl,void ** seg,int * size,void ** cur_edge,unsigned int * idx,int nreq,u8 fence,u8 next_fence,int * num_sge)931 static int handle_qpt_rc(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp,
932 const struct ib_send_wr *wr,
933 struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size,
934 void **cur_edge, unsigned int *idx, int nreq, u8 fence,
935 u8 next_fence, int *num_sge)
936 {
937 int err = 0;
938
939 switch (wr->opcode) {
940 case IB_WR_RDMA_READ:
941 case IB_WR_RDMA_WRITE:
942 case IB_WR_RDMA_WRITE_WITH_IMM:
943 handle_rdma_op(wr, seg, size);
944 break;
945
946 case IB_WR_ATOMIC_CMP_AND_SWP:
947 case IB_WR_ATOMIC_FETCH_AND_ADD:
948 case IB_WR_MASKED_ATOMIC_CMP_AND_SWP:
949 mlx5_ib_warn(dev, "Atomic operations are not supported yet\n");
950 err = -EOPNOTSUPP;
951 goto out;
952
953 case IB_WR_LOCAL_INV:
954 handle_local_inv(qp, wr, ctrl, seg, size, cur_edge, *idx);
955 *num_sge = 0;
956 break;
957
958 case IB_WR_REG_MR:
959 err = handle_reg_mr(qp, wr, ctrl, seg, size, cur_edge, *idx);
960 if (unlikely(err))
961 goto out;
962 *num_sge = 0;
963 break;
964
965 case IB_WR_REG_MR_INTEGRITY:
966 err = handle_reg_mr_integrity(dev, qp, wr, ctrl, seg, size,
967 cur_edge, idx, nreq, fence,
968 next_fence);
969 if (unlikely(err))
970 goto out;
971 *num_sge = 0;
972 break;
973
974 default:
975 break;
976 }
977
978 out:
979 return err;
980 }
981
handle_qpt_uc(const struct ib_send_wr * wr,void ** seg,int * size)982 static void handle_qpt_uc(const struct ib_send_wr *wr, void **seg, int *size)
983 {
984 switch (wr->opcode) {
985 case IB_WR_RDMA_WRITE:
986 case IB_WR_RDMA_WRITE_WITH_IMM:
987 handle_rdma_op(wr, seg, size);
988 break;
989 default:
990 break;
991 }
992 }
993
handle_qpt_hw_gsi(struct mlx5_ib_qp * qp,const struct ib_send_wr * wr,void ** seg,int * size,void ** cur_edge)994 static void handle_qpt_hw_gsi(struct mlx5_ib_qp *qp,
995 const struct ib_send_wr *wr, void **seg,
996 int *size, void **cur_edge)
997 {
998 set_datagram_seg(*seg, wr);
999 *seg += sizeof(struct mlx5_wqe_datagram_seg);
1000 *size += sizeof(struct mlx5_wqe_datagram_seg) / 16;
1001 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
1002 }
1003
handle_qpt_ud(struct mlx5_ib_qp * qp,const struct ib_send_wr * wr,void ** seg,int * size,void ** cur_edge)1004 static void handle_qpt_ud(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr,
1005 void **seg, int *size, void **cur_edge)
1006 {
1007 set_datagram_seg(*seg, wr);
1008 *seg += sizeof(struct mlx5_wqe_datagram_seg);
1009 *size += sizeof(struct mlx5_wqe_datagram_seg) / 16;
1010 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
1011
1012 /* handle qp that supports ud offload */
1013 if (qp->flags & IB_QP_CREATE_IPOIB_UD_LSO) {
1014 struct mlx5_wqe_eth_pad *pad;
1015
1016 pad = *seg;
1017 memset(pad, 0, sizeof(struct mlx5_wqe_eth_pad));
1018 *seg += sizeof(struct mlx5_wqe_eth_pad);
1019 *size += sizeof(struct mlx5_wqe_eth_pad) / 16;
1020 set_eth_seg(wr, qp, seg, size, cur_edge);
1021 handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
1022 }
1023 }
1024
mlx5r_ring_db(struct mlx5_ib_qp * qp,unsigned int nreq,struct mlx5_wqe_ctrl_seg * ctrl)1025 void mlx5r_ring_db(struct mlx5_ib_qp *qp, unsigned int nreq,
1026 struct mlx5_wqe_ctrl_seg *ctrl)
1027 {
1028 struct mlx5_bf *bf = &qp->bf;
1029
1030 qp->sq.head += nreq;
1031
1032 /* Make sure that descriptors are written before
1033 * updating doorbell record and ringing the doorbell
1034 */
1035 wmb();
1036
1037 qp->db.db[MLX5_SND_DBR] = cpu_to_be32(qp->sq.cur_post);
1038
1039 /* Make sure doorbell record is visible to the HCA before
1040 * we hit doorbell.
1041 */
1042 wmb();
1043
1044 mlx5_write64((__be32 *)ctrl, bf->bfreg->map + bf->offset);
1045 /* Make sure doorbells don't leak out of SQ spinlock
1046 * and reach the HCA out of order.
1047 */
1048 bf->offset ^= bf->buf_size;
1049 }
1050
mlx5_ib_post_send(struct ib_qp * ibqp,const struct ib_send_wr * wr,const struct ib_send_wr ** bad_wr,bool drain)1051 int mlx5_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
1052 const struct ib_send_wr **bad_wr, bool drain)
1053 {
1054 struct mlx5_wqe_ctrl_seg *ctrl = NULL; /* compiler warning */
1055 struct mlx5_ib_dev *dev = to_mdev(ibqp->device);
1056 struct mlx5_core_dev *mdev = dev->mdev;
1057 struct mlx5_ib_qp *qp = to_mqp(ibqp);
1058 struct mlx5_wqe_xrc_seg *xrc;
1059 void *cur_edge;
1060 int size;
1061 unsigned long flags;
1062 unsigned int idx;
1063 int err = 0;
1064 int num_sge;
1065 void *seg;
1066 int nreq;
1067 int i;
1068 u8 next_fence = 0;
1069 u8 fence;
1070
1071 if (unlikely(mdev->state == MLX5_DEVICE_STATE_INTERNAL_ERROR &&
1072 !drain)) {
1073 *bad_wr = wr;
1074 return -EIO;
1075 }
1076
1077 if (qp->type == IB_QPT_GSI)
1078 return mlx5_ib_gsi_post_send(ibqp, wr, bad_wr);
1079
1080 spin_lock_irqsave(&qp->sq.lock, flags);
1081
1082 for (nreq = 0; wr; nreq++, wr = wr->next) {
1083 if (unlikely(wr->opcode >= ARRAY_SIZE(mlx5_ib_opcode))) {
1084 mlx5_ib_warn(dev, "\n");
1085 err = -EINVAL;
1086 *bad_wr = wr;
1087 goto out;
1088 }
1089
1090 num_sge = wr->num_sge;
1091 if (unlikely(num_sge > qp->sq.max_gs)) {
1092 mlx5_ib_warn(dev, "\n");
1093 err = -EINVAL;
1094 *bad_wr = wr;
1095 goto out;
1096 }
1097
1098 err = begin_wqe(qp, &seg, &ctrl, wr, &idx, &size, &cur_edge,
1099 nreq);
1100 if (err) {
1101 mlx5_ib_warn(dev, "\n");
1102 err = -ENOMEM;
1103 *bad_wr = wr;
1104 goto out;
1105 }
1106
1107 if (wr->opcode == IB_WR_REG_MR ||
1108 wr->opcode == IB_WR_REG_MR_INTEGRITY) {
1109 fence = dev->umr_fence;
1110 next_fence = MLX5_FENCE_MODE_INITIATOR_SMALL;
1111 } else {
1112 if (wr->send_flags & IB_SEND_FENCE) {
1113 if (qp->next_fence)
1114 fence = MLX5_FENCE_MODE_SMALL_AND_FENCE;
1115 else
1116 fence = MLX5_FENCE_MODE_FENCE;
1117 } else {
1118 fence = qp->next_fence;
1119 }
1120 }
1121
1122 switch (qp->type) {
1123 case IB_QPT_XRC_INI:
1124 xrc = seg;
1125 seg += sizeof(*xrc);
1126 size += sizeof(*xrc) / 16;
1127 fallthrough;
1128 case IB_QPT_RC:
1129 err = handle_qpt_rc(dev, qp, wr, &ctrl, &seg, &size,
1130 &cur_edge, &idx, nreq, fence,
1131 next_fence, &num_sge);
1132 if (unlikely(err)) {
1133 *bad_wr = wr;
1134 goto out;
1135 } else if (wr->opcode == IB_WR_REG_MR_INTEGRITY) {
1136 goto skip_psv;
1137 }
1138 break;
1139
1140 case IB_QPT_UC:
1141 handle_qpt_uc(wr, &seg, &size);
1142 break;
1143 case IB_QPT_SMI:
1144 if (unlikely(!dev->port_caps[qp->port - 1].has_smi)) {
1145 mlx5_ib_warn(dev, "Send SMP MADs is not allowed\n");
1146 err = -EPERM;
1147 *bad_wr = wr;
1148 goto out;
1149 }
1150 fallthrough;
1151 case MLX5_IB_QPT_HW_GSI:
1152 handle_qpt_hw_gsi(qp, wr, &seg, &size, &cur_edge);
1153 break;
1154 case IB_QPT_UD:
1155 handle_qpt_ud(qp, wr, &seg, &size, &cur_edge);
1156 break;
1157
1158 default:
1159 break;
1160 }
1161
1162 if (wr->send_flags & IB_SEND_INLINE && num_sge) {
1163 err = set_data_inl_seg(qp, wr, &seg, &size, &cur_edge);
1164 if (unlikely(err)) {
1165 mlx5_ib_warn(dev, "\n");
1166 *bad_wr = wr;
1167 goto out;
1168 }
1169 } else {
1170 for (i = 0; i < num_sge; i++) {
1171 handle_post_send_edge(&qp->sq, &seg, size,
1172 &cur_edge);
1173 if (unlikely(!wr->sg_list[i].length))
1174 continue;
1175
1176 set_data_ptr_seg(
1177 (struct mlx5_wqe_data_seg *)seg,
1178 wr->sg_list + i);
1179 size += sizeof(struct mlx5_wqe_data_seg) / 16;
1180 seg += sizeof(struct mlx5_wqe_data_seg);
1181 }
1182 }
1183
1184 qp->next_fence = next_fence;
1185 mlx5r_finish_wqe(qp, ctrl, seg, size, cur_edge, idx, wr->wr_id,
1186 nreq, fence, mlx5_ib_opcode[wr->opcode]);
1187 skip_psv:
1188 if (0)
1189 dump_wqe(qp, idx, size);
1190 }
1191
1192 out:
1193 if (likely(nreq))
1194 mlx5r_ring_db(qp, nreq, ctrl);
1195
1196 spin_unlock_irqrestore(&qp->sq.lock, flags);
1197
1198 return err;
1199 }
1200
set_sig_seg(struct mlx5_rwqe_sig * sig,int max_gs)1201 static void set_sig_seg(struct mlx5_rwqe_sig *sig, int max_gs)
1202 {
1203 sig->signature = calc_sig(sig, (max_gs + 1) << 2);
1204 }
1205
mlx5_ib_post_recv(struct ib_qp * ibqp,const struct ib_recv_wr * wr,const struct ib_recv_wr ** bad_wr,bool drain)1206 int mlx5_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
1207 const struct ib_recv_wr **bad_wr, bool drain)
1208 {
1209 struct mlx5_ib_qp *qp = to_mqp(ibqp);
1210 struct mlx5_wqe_data_seg *scat;
1211 struct mlx5_rwqe_sig *sig;
1212 struct mlx5_ib_dev *dev = to_mdev(ibqp->device);
1213 struct mlx5_core_dev *mdev = dev->mdev;
1214 unsigned long flags;
1215 int err = 0;
1216 int nreq;
1217 int ind;
1218 int i;
1219
1220 if (unlikely(mdev->state == MLX5_DEVICE_STATE_INTERNAL_ERROR &&
1221 !drain)) {
1222 *bad_wr = wr;
1223 return -EIO;
1224 }
1225
1226 if (qp->type == IB_QPT_GSI)
1227 return mlx5_ib_gsi_post_recv(ibqp, wr, bad_wr);
1228
1229 spin_lock_irqsave(&qp->rq.lock, flags);
1230
1231 ind = qp->rq.head & (qp->rq.wqe_cnt - 1);
1232
1233 for (nreq = 0; wr; nreq++, wr = wr->next) {
1234 if (mlx5r_wq_overflow(&qp->rq, nreq, qp->ibqp.recv_cq)) {
1235 err = -ENOMEM;
1236 *bad_wr = wr;
1237 goto out;
1238 }
1239
1240 if (unlikely(wr->num_sge > qp->rq.max_gs)) {
1241 err = -EINVAL;
1242 *bad_wr = wr;
1243 goto out;
1244 }
1245
1246 scat = mlx5_frag_buf_get_wqe(&qp->rq.fbc, ind);
1247 if (qp->flags_en & MLX5_QP_FLAG_SIGNATURE)
1248 scat++;
1249
1250 for (i = 0; i < wr->num_sge; i++)
1251 set_data_ptr_seg(scat + i, wr->sg_list + i);
1252
1253 if (i < qp->rq.max_gs) {
1254 scat[i].byte_count = 0;
1255 scat[i].lkey = dev->mkeys.terminate_scatter_list_mkey;
1256 scat[i].addr = 0;
1257 }
1258
1259 if (qp->flags_en & MLX5_QP_FLAG_SIGNATURE) {
1260 sig = (struct mlx5_rwqe_sig *)scat;
1261 set_sig_seg(sig, qp->rq.max_gs);
1262 }
1263
1264 qp->rq.wrid[ind] = wr->wr_id;
1265
1266 ind = (ind + 1) & (qp->rq.wqe_cnt - 1);
1267 }
1268
1269 out:
1270 if (likely(nreq)) {
1271 qp->rq.head += nreq;
1272
1273 /* Make sure that descriptors are written before
1274 * doorbell record.
1275 */
1276 wmb();
1277
1278 *qp->db.db = cpu_to_be32(qp->rq.head & 0xffff);
1279 }
1280
1281 spin_unlock_irqrestore(&qp->rq.lock, flags);
1282
1283 return err;
1284 }
1285