1 /* 2 * Copyright (c) 2013-2015, Mellanox Technologies, Ltd. All rights reserved. 3 * 4 * This software is available to you under a choice of one of two 5 * licenses. You may choose to be licensed under the terms of the GNU 6 * General Public License (GPL) Version 2, available from the file 7 * COPYING in the main directory of this source tree, or the 8 * OpenIB.org BSD license below: 9 * 10 * Redistribution and use in source and binary forms, with or 11 * without modification, are permitted provided that the following 12 * conditions are met: 13 * 14 * - Redistributions of source code must retain the above 15 * copyright notice, this list of conditions and the following 16 * disclaimer. 17 * 18 * - Redistributions in binary form must reproduce the above 19 * copyright notice, this list of conditions and the following 20 * disclaimer in the documentation and/or other materials 21 * provided with the distribution. 22 * 23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 30 * SOFTWARE. 31 */ 32 33 #include <linux/mlx5/driver.h> 34 #include "wq.h" 35 #include "mlx5_core.h" 36 37 u32 mlx5_wq_cyc_get_size(struct mlx5_wq_cyc *wq) 38 { 39 return (u32)wq->fbc.sz_m1 + 1; 40 } 41 42 u32 mlx5_cqwq_get_size(struct mlx5_cqwq *wq) 43 { 44 return wq->fbc.sz_m1 + 1; 45 } 46 47 u32 mlx5_wq_ll_get_size(struct mlx5_wq_ll *wq) 48 { 49 return (u32)wq->fbc.sz_m1 + 1; 50 } 51 52 static u32 wq_get_byte_sz(u8 log_sz, u8 log_stride) 53 { 54 return ((u32)1 << log_sz) << log_stride; 55 } 56 57 int mlx5_wq_cyc_create(struct mlx5_core_dev *mdev, struct mlx5_wq_param *param, 58 void *wqc, struct mlx5_wq_cyc *wq, 59 struct mlx5_wq_ctrl *wq_ctrl) 60 { 61 u8 log_wq_stride = MLX5_GET(wq, wqc, log_wq_stride); 62 u8 log_wq_sz = MLX5_GET(wq, wqc, log_wq_sz); 63 struct mlx5_frag_buf_ctrl *fbc = &wq->fbc; 64 int err; 65 66 err = mlx5_db_alloc_node(mdev, &wq_ctrl->db, param->db_numa_node); 67 if (err) { 68 mlx5_core_warn(mdev, "mlx5_db_alloc_node() failed, %d\n", err); 69 return err; 70 } 71 72 wq->db = wq_ctrl->db.db; 73 74 err = mlx5_frag_buf_alloc_node(mdev, wq_get_byte_sz(log_wq_sz, log_wq_stride), 75 &wq_ctrl->buf, param->buf_numa_node); 76 if (err) { 77 mlx5_core_warn(mdev, "mlx5_frag_buf_alloc_node() failed, %d\n", err); 78 goto err_db_free; 79 } 80 81 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc); 82 wq->sz = mlx5_wq_cyc_get_size(wq); 83 84 wq_ctrl->mdev = mdev; 85 86 return 0; 87 88 err_db_free: 89 mlx5_db_free(mdev, &wq_ctrl->db); 90 91 return err; 92 } 93 94 int mlx5_wq_qp_create(struct mlx5_core_dev *mdev, struct mlx5_wq_param *param, 95 void *qpc, struct mlx5_wq_qp *wq, 96 struct mlx5_wq_ctrl *wq_ctrl) 97 { 98 u8 log_rq_stride = MLX5_GET(qpc, qpc, log_rq_stride) + 4; 99 u8 log_rq_sz = MLX5_GET(qpc, qpc, log_rq_size); 100 u8 log_sq_stride = ilog2(MLX5_SEND_WQE_BB); 101 u8 log_sq_sz = MLX5_GET(qpc, qpc, log_sq_size); 102 103 u32 rq_byte_size; 104 int err; 105 106 107 108 err = mlx5_db_alloc_node(mdev, &wq_ctrl->db, param->db_numa_node); 109 if (err) { 110 mlx5_core_warn(mdev, "mlx5_db_alloc_node() failed, %d\n", err); 111 return err; 112 } 113 114 err = mlx5_frag_buf_alloc_node(mdev, 115 wq_get_byte_sz(log_rq_sz, log_rq_stride) + 116 wq_get_byte_sz(log_sq_sz, log_sq_stride), 117 &wq_ctrl->buf, param->buf_numa_node); 118 if (err) { 119 mlx5_core_warn(mdev, "mlx5_frag_buf_alloc_node() failed, %d\n", err); 120 goto err_db_free; 121 } 122 123 mlx5_init_fbc(wq_ctrl->buf.frags, log_rq_stride, log_rq_sz, &wq->rq.fbc); 124 125 rq_byte_size = wq_get_byte_sz(log_rq_sz, log_rq_stride); 126 127 if (rq_byte_size < PAGE_SIZE) { 128 /* SQ starts within the same page of the RQ */ 129 u16 sq_strides_offset = rq_byte_size / MLX5_SEND_WQE_BB; 130 131 mlx5_init_fbc_offset(wq_ctrl->buf.frags, 132 log_sq_stride, log_sq_sz, sq_strides_offset, 133 &wq->sq.fbc); 134 } else { 135 u16 rq_npages = rq_byte_size >> PAGE_SHIFT; 136 137 mlx5_init_fbc(wq_ctrl->buf.frags + rq_npages, 138 log_sq_stride, log_sq_sz, &wq->sq.fbc); 139 } 140 141 wq->rq.db = &wq_ctrl->db.db[MLX5_RCV_DBR]; 142 wq->sq.db = &wq_ctrl->db.db[MLX5_SND_DBR]; 143 144 wq_ctrl->mdev = mdev; 145 146 return 0; 147 148 err_db_free: 149 mlx5_db_free(mdev, &wq_ctrl->db); 150 151 return err; 152 } 153 154 int mlx5_cqwq_create(struct mlx5_core_dev *mdev, struct mlx5_wq_param *param, 155 void *cqc, struct mlx5_cqwq *wq, 156 struct mlx5_wq_ctrl *wq_ctrl) 157 { 158 u8 log_wq_stride = MLX5_GET(cqc, cqc, cqe_sz) + 6; 159 u8 log_wq_sz = MLX5_GET(cqc, cqc, log_cq_size); 160 int err; 161 162 err = mlx5_db_alloc_node(mdev, &wq_ctrl->db, param->db_numa_node); 163 if (err) { 164 mlx5_core_warn(mdev, "mlx5_db_alloc_node() failed, %d\n", err); 165 return err; 166 } 167 168 wq->db = wq_ctrl->db.db; 169 170 err = mlx5_frag_buf_alloc_node(mdev, wq_get_byte_sz(log_wq_sz, log_wq_stride), 171 &wq_ctrl->buf, 172 param->buf_numa_node); 173 if (err) { 174 mlx5_core_warn(mdev, "mlx5_frag_buf_alloc_node() failed, %d\n", 175 err); 176 goto err_db_free; 177 } 178 179 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, &wq->fbc); 180 181 wq_ctrl->mdev = mdev; 182 183 return 0; 184 185 err_db_free: 186 mlx5_db_free(mdev, &wq_ctrl->db); 187 188 return err; 189 } 190 191 int mlx5_wq_ll_create(struct mlx5_core_dev *mdev, struct mlx5_wq_param *param, 192 void *wqc, struct mlx5_wq_ll *wq, 193 struct mlx5_wq_ctrl *wq_ctrl) 194 { 195 u8 log_wq_stride = MLX5_GET(wq, wqc, log_wq_stride); 196 u8 log_wq_sz = MLX5_GET(wq, wqc, log_wq_sz); 197 struct mlx5_frag_buf_ctrl *fbc = &wq->fbc; 198 struct mlx5_wqe_srq_next_seg *next_seg; 199 int err; 200 int i; 201 202 err = mlx5_db_alloc_node(mdev, &wq_ctrl->db, param->db_numa_node); 203 if (err) { 204 mlx5_core_warn(mdev, "mlx5_db_alloc_node() failed, %d\n", err); 205 return err; 206 } 207 208 wq->db = wq_ctrl->db.db; 209 210 err = mlx5_frag_buf_alloc_node(mdev, wq_get_byte_sz(log_wq_sz, log_wq_stride), 211 &wq_ctrl->buf, param->buf_numa_node); 212 if (err) { 213 mlx5_core_warn(mdev, "mlx5_frag_buf_alloc_node() failed, %d\n", err); 214 goto err_db_free; 215 } 216 217 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc); 218 219 for (i = 0; i < fbc->sz_m1; i++) { 220 next_seg = mlx5_wq_ll_get_wqe(wq, i); 221 next_seg->next_wqe_index = cpu_to_be16(i + 1); 222 } 223 next_seg = mlx5_wq_ll_get_wqe(wq, i); 224 wq->tail_next = &next_seg->next_wqe_index; 225 226 wq_ctrl->mdev = mdev; 227 228 return 0; 229 230 err_db_free: 231 mlx5_db_free(mdev, &wq_ctrl->db); 232 233 return err; 234 } 235 236 void mlx5_wq_destroy(struct mlx5_wq_ctrl *wq_ctrl) 237 { 238 mlx5_frag_buf_free(wq_ctrl->mdev, &wq_ctrl->buf); 239 mlx5_db_free(wq_ctrl->mdev, &wq_ctrl->db); 240 } 241 242