1 /* 2 * Broadcom NetXtreme-E RoCE driver. 3 * 4 * Copyright (c) 2016 - 2017, Broadcom. All rights reserved. The term 5 * Broadcom refers to Broadcom Limited and/or its subsidiaries. 6 * 7 * This software is available to you under a choice of one of two 8 * licenses. You may choose to be licensed under the terms of the GNU 9 * General Public License (GPL) Version 2, available from the file 10 * COPYING in the main directory of this source tree, or the 11 * BSD license below: 12 * 13 * Redistribution and use in source and binary forms, with or without 14 * modification, are permitted provided that the following conditions 15 * are met: 16 * 17 * 1. Redistributions of source code must retain the above copyright 18 * notice, this list of conditions and the following disclaimer. 19 * 2. Redistributions in binary form must reproduce the above copyright 20 * notice, this list of conditions and the following disclaimer in 21 * the documentation and/or other materials provided with the 22 * distribution. 23 * 24 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' 25 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, 26 * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 27 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS 28 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 29 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 30 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR 31 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, 32 * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE 33 * OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN 34 * IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 35 * 36 * Description: IB Verbs interpreter (header) 37 */ 38 39 #ifndef __BNXT_RE_IB_VERBS_H__ 40 #define __BNXT_RE_IB_VERBS_H__ 41 42 struct bnxt_re_gid_ctx { 43 u32 idx; 44 u32 refcnt; 45 }; 46 47 #define BNXT_RE_FENCE_BYTES 64 48 struct bnxt_re_fence_data { 49 u32 size; 50 u8 va[BNXT_RE_FENCE_BYTES]; 51 dma_addr_t dma_addr; 52 struct bnxt_re_mr *mr; 53 struct ib_mw *mw; 54 struct bnxt_qplib_swqe bind_wqe; 55 u32 bind_rkey; 56 }; 57 58 struct bnxt_re_pd { 59 struct bnxt_re_dev *rdev; 60 struct ib_pd ib_pd; 61 struct bnxt_qplib_pd qplib_pd; 62 struct bnxt_re_fence_data fence; 63 }; 64 65 struct bnxt_re_ah { 66 struct bnxt_re_dev *rdev; 67 struct ib_ah ib_ah; 68 struct bnxt_qplib_ah qplib_ah; 69 }; 70 71 struct bnxt_re_qp { 72 struct list_head list; 73 struct bnxt_re_dev *rdev; 74 struct ib_qp ib_qp; 75 spinlock_t sq_lock; /* protect sq */ 76 spinlock_t rq_lock; /* protect rq */ 77 struct bnxt_qplib_qp qplib_qp; 78 struct ib_umem *sumem; 79 struct ib_umem *rumem; 80 /* QP1 */ 81 u32 send_psn; 82 struct ib_ud_header qp1_hdr; 83 }; 84 85 struct bnxt_re_cq { 86 struct bnxt_re_dev *rdev; 87 spinlock_t cq_lock; /* protect cq */ 88 u16 cq_count; 89 u16 cq_period; 90 struct ib_cq ib_cq; 91 struct bnxt_qplib_cq qplib_cq; 92 struct bnxt_qplib_cqe *cql; 93 #define MAX_CQL_PER_POLL 1024 94 u32 max_cql; 95 struct ib_umem *umem; 96 }; 97 98 struct bnxt_re_mr { 99 struct bnxt_re_dev *rdev; 100 struct ib_mr ib_mr; 101 struct ib_umem *ib_umem; 102 struct bnxt_qplib_mrw qplib_mr; 103 u32 npages; 104 u64 *pages; 105 struct bnxt_qplib_frpl qplib_frpl; 106 }; 107 108 struct bnxt_re_frpl { 109 struct bnxt_re_dev *rdev; 110 struct bnxt_qplib_frpl qplib_frpl; 111 u64 *page_list; 112 }; 113 114 struct bnxt_re_fmr { 115 struct bnxt_re_dev *rdev; 116 struct ib_fmr ib_fmr; 117 struct bnxt_qplib_mrw qplib_fmr; 118 }; 119 120 struct bnxt_re_mw { 121 struct bnxt_re_dev *rdev; 122 struct ib_mw ib_mw; 123 struct bnxt_qplib_mrw qplib_mw; 124 }; 125 126 struct bnxt_re_ucontext { 127 struct bnxt_re_dev *rdev; 128 struct ib_ucontext ib_uctx; 129 struct bnxt_qplib_dpi dpi; 130 void *shpg; 131 spinlock_t sh_lock; /* protect shpg */ 132 }; 133 134 struct net_device *bnxt_re_get_netdev(struct ib_device *ibdev, u8 port_num); 135 136 int bnxt_re_query_device(struct ib_device *ibdev, 137 struct ib_device_attr *ib_attr, 138 struct ib_udata *udata); 139 int bnxt_re_modify_device(struct ib_device *ibdev, 140 int device_modify_mask, 141 struct ib_device_modify *device_modify); 142 int bnxt_re_query_port(struct ib_device *ibdev, u8 port_num, 143 struct ib_port_attr *port_attr); 144 int bnxt_re_modify_port(struct ib_device *ibdev, u8 port_num, 145 int port_modify_mask, 146 struct ib_port_modify *port_modify); 147 int bnxt_re_get_port_immutable(struct ib_device *ibdev, u8 port_num, 148 struct ib_port_immutable *immutable); 149 int bnxt_re_query_pkey(struct ib_device *ibdev, u8 port_num, 150 u16 index, u16 *pkey); 151 int bnxt_re_del_gid(struct ib_device *ibdev, u8 port_num, 152 unsigned int index, void **context); 153 int bnxt_re_add_gid(struct ib_device *ibdev, u8 port_num, 154 unsigned int index, const union ib_gid *gid, 155 const struct ib_gid_attr *attr, void **context); 156 int bnxt_re_query_gid(struct ib_device *ibdev, u8 port_num, 157 int index, union ib_gid *gid); 158 enum rdma_link_layer bnxt_re_get_link_layer(struct ib_device *ibdev, 159 u8 port_num); 160 struct ib_pd *bnxt_re_alloc_pd(struct ib_device *ibdev, 161 struct ib_ucontext *context, 162 struct ib_udata *udata); 163 int bnxt_re_dealloc_pd(struct ib_pd *pd); 164 struct ib_ah *bnxt_re_create_ah(struct ib_pd *pd, 165 struct rdma_ah_attr *ah_attr, 166 struct ib_udata *udata); 167 int bnxt_re_modify_ah(struct ib_ah *ah, struct rdma_ah_attr *ah_attr); 168 int bnxt_re_query_ah(struct ib_ah *ah, struct rdma_ah_attr *ah_attr); 169 int bnxt_re_destroy_ah(struct ib_ah *ah); 170 struct ib_qp *bnxt_re_create_qp(struct ib_pd *pd, 171 struct ib_qp_init_attr *qp_init_attr, 172 struct ib_udata *udata); 173 int bnxt_re_modify_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, 174 int qp_attr_mask, struct ib_udata *udata); 175 int bnxt_re_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr, 176 int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr); 177 int bnxt_re_destroy_qp(struct ib_qp *qp); 178 int bnxt_re_post_send(struct ib_qp *qp, struct ib_send_wr *send_wr, 179 struct ib_send_wr **bad_send_wr); 180 int bnxt_re_post_recv(struct ib_qp *qp, struct ib_recv_wr *recv_wr, 181 struct ib_recv_wr **bad_recv_wr); 182 struct ib_cq *bnxt_re_create_cq(struct ib_device *ibdev, 183 const struct ib_cq_init_attr *attr, 184 struct ib_ucontext *context, 185 struct ib_udata *udata); 186 int bnxt_re_destroy_cq(struct ib_cq *cq); 187 int bnxt_re_poll_cq(struct ib_cq *cq, int num_entries, struct ib_wc *wc); 188 int bnxt_re_req_notify_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags); 189 struct ib_mr *bnxt_re_get_dma_mr(struct ib_pd *pd, int mr_access_flags); 190 191 int bnxt_re_map_mr_sg(struct ib_mr *ib_mr, struct scatterlist *sg, int sg_nents, 192 unsigned int *sg_offset); 193 struct ib_mr *bnxt_re_alloc_mr(struct ib_pd *ib_pd, enum ib_mr_type mr_type, 194 u32 max_num_sg); 195 int bnxt_re_dereg_mr(struct ib_mr *mr); 196 struct ib_mw *bnxt_re_alloc_mw(struct ib_pd *ib_pd, enum ib_mw_type type, 197 struct ib_udata *udata); 198 int bnxt_re_dealloc_mw(struct ib_mw *mw); 199 struct ib_mr *bnxt_re_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, 200 u64 virt_addr, int mr_access_flags, 201 struct ib_udata *udata); 202 struct ib_ucontext *bnxt_re_alloc_ucontext(struct ib_device *ibdev, 203 struct ib_udata *udata); 204 int bnxt_re_dealloc_ucontext(struct ib_ucontext *context); 205 int bnxt_re_mmap(struct ib_ucontext *context, struct vm_area_struct *vma); 206 #endif /* __BNXT_RE_IB_VERBS_H__ */ 207