1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Huawei HiNIC PCI Express Linux driver 4 * Copyright(c) 2017 Huawei Technologies Co., Ltd 5 */ 6 7 #include <linux/kernel.h> 8 #include <linux/types.h> 9 #include <linux/errno.h> 10 #include <linux/pci.h> 11 #include <linux/device.h> 12 #include <linux/slab.h> 13 #include <linux/vmalloc.h> 14 #include <linux/spinlock.h> 15 #include <linux/sizes.h> 16 #include <linux/atomic.h> 17 #include <linux/log2.h> 18 #include <linux/io.h> 19 #include <linux/completion.h> 20 #include <linux/err.h> 21 #include <asm/byteorder.h> 22 #include <asm/barrier.h> 23 24 #include "hinic_common.h" 25 #include "hinic_hw_if.h" 26 #include "hinic_hw_eqs.h" 27 #include "hinic_hw_mgmt.h" 28 #include "hinic_hw_wqe.h" 29 #include "hinic_hw_wq.h" 30 #include "hinic_hw_cmdq.h" 31 #include "hinic_hw_io.h" 32 #include "hinic_hw_dev.h" 33 34 #define CMDQ_CEQE_TYPE_SHIFT 0 35 36 #define CMDQ_CEQE_TYPE_MASK 0x7 37 38 #define CMDQ_CEQE_GET(val, member) \ 39 (((val) >> CMDQ_CEQE_##member##_SHIFT) \ 40 & CMDQ_CEQE_##member##_MASK) 41 42 #define CMDQ_WQE_ERRCODE_VAL_SHIFT 20 43 44 #define CMDQ_WQE_ERRCODE_VAL_MASK 0xF 45 46 #define CMDQ_WQE_ERRCODE_GET(val, member) \ 47 (((val) >> CMDQ_WQE_ERRCODE_##member##_SHIFT) \ 48 & CMDQ_WQE_ERRCODE_##member##_MASK) 49 50 #define CMDQ_DB_PI_OFF(pi) (((u16)LOWER_8_BITS(pi)) << 3) 51 52 #define CMDQ_DB_ADDR(db_base, pi) ((db_base) + CMDQ_DB_PI_OFF(pi)) 53 54 #define CMDQ_WQE_HEADER(wqe) ((struct hinic_cmdq_header *)(wqe)) 55 56 #define CMDQ_WQE_COMPLETED(ctrl_info) \ 57 HINIC_CMDQ_CTRL_GET(ctrl_info, HW_BUSY_BIT) 58 59 #define FIRST_DATA_TO_WRITE_LAST sizeof(u64) 60 61 #define CMDQ_DB_OFF SZ_2K 62 63 #define CMDQ_WQEBB_SIZE 64 64 #define CMDQ_WQE_SIZE 64 65 #define CMDQ_DEPTH SZ_4K 66 67 #define CMDQ_WQ_PAGE_SIZE SZ_256K 68 69 #define WQE_LCMD_SIZE 64 70 #define WQE_SCMD_SIZE 64 71 72 #define COMPLETE_LEN 3 73 74 #define CMDQ_TIMEOUT 1000 75 76 #define CMDQ_PFN(addr, page_size) ((addr) >> (ilog2(page_size))) 77 78 #define cmdq_to_cmdqs(cmdq) container_of((cmdq) - (cmdq)->cmdq_type, \ 79 struct hinic_cmdqs, cmdq[0]) 80 81 #define cmdqs_to_func_to_io(cmdqs) container_of(cmdqs, \ 82 struct hinic_func_to_io, \ 83 cmdqs) 84 85 enum completion_format { 86 COMPLETE_DIRECT = 0, 87 COMPLETE_SGE = 1, 88 }; 89 90 enum data_format { 91 DATA_SGE = 0, 92 DATA_DIRECT = 1, 93 }; 94 95 enum bufdesc_len { 96 BUFDESC_LCMD_LEN = 2, /* 16 bytes - 2(8 byte unit) */ 97 BUFDESC_SCMD_LEN = 3, /* 24 bytes - 3(8 byte unit) */ 98 }; 99 100 enum ctrl_sect_len { 101 CTRL_SECT_LEN = 1, /* 4 bytes (ctrl) - 1(8 byte unit) */ 102 CTRL_DIRECT_SECT_LEN = 2, /* 12 bytes (ctrl + rsvd) - 2(8 byte unit) */ 103 }; 104 105 enum cmdq_scmd_type { 106 CMDQ_SET_ARM_CMD = 2, 107 }; 108 109 enum cmdq_cmd_type { 110 CMDQ_CMD_SYNC_DIRECT_RESP = 0, 111 CMDQ_CMD_SYNC_SGE_RESP = 1, 112 }; 113 114 enum completion_request { 115 NO_CEQ = 0, 116 CEQ_SET = 1, 117 }; 118 119 /** 120 * hinic_alloc_cmdq_buf - alloc buffer for sending command 121 * @cmdqs: the cmdqs 122 * @cmdq_buf: the buffer returned in this struct 123 * 124 * Return 0 - Success, negative - Failure 125 **/ 126 int hinic_alloc_cmdq_buf(struct hinic_cmdqs *cmdqs, 127 struct hinic_cmdq_buf *cmdq_buf) 128 { 129 struct hinic_hwif *hwif = cmdqs->hwif; 130 struct pci_dev *pdev = hwif->pdev; 131 132 cmdq_buf->buf = dma_pool_alloc(cmdqs->cmdq_buf_pool, GFP_KERNEL, 133 &cmdq_buf->dma_addr); 134 if (!cmdq_buf->buf) { 135 dev_err(&pdev->dev, "Failed to allocate cmd from the pool\n"); 136 return -ENOMEM; 137 } 138 139 return 0; 140 } 141 142 /** 143 * hinic_free_cmdq_buf - free buffer 144 * @cmdqs: the cmdqs 145 * @cmdq_buf: the buffer to free that is in this struct 146 **/ 147 void hinic_free_cmdq_buf(struct hinic_cmdqs *cmdqs, 148 struct hinic_cmdq_buf *cmdq_buf) 149 { 150 dma_pool_free(cmdqs->cmdq_buf_pool, cmdq_buf->buf, cmdq_buf->dma_addr); 151 } 152 153 static unsigned int cmdq_wqe_size_from_bdlen(enum bufdesc_len len) 154 { 155 unsigned int wqe_size = 0; 156 157 switch (len) { 158 case BUFDESC_LCMD_LEN: 159 wqe_size = WQE_LCMD_SIZE; 160 break; 161 case BUFDESC_SCMD_LEN: 162 wqe_size = WQE_SCMD_SIZE; 163 break; 164 } 165 166 return wqe_size; 167 } 168 169 static void cmdq_set_sge_completion(struct hinic_cmdq_completion *completion, 170 struct hinic_cmdq_buf *buf_out) 171 { 172 struct hinic_sge_resp *sge_resp = &completion->sge_resp; 173 174 hinic_set_sge(&sge_resp->sge, buf_out->dma_addr, buf_out->size); 175 } 176 177 static void cmdq_prepare_wqe_ctrl(struct hinic_cmdq_wqe *wqe, int wrapped, 178 enum hinic_cmd_ack_type ack_type, 179 enum hinic_mod_type mod, u8 cmd, u16 prod_idx, 180 enum completion_format complete_format, 181 enum data_format data_format, 182 enum bufdesc_len buf_len) 183 { 184 struct hinic_cmdq_wqe_lcmd *wqe_lcmd; 185 struct hinic_cmdq_wqe_scmd *wqe_scmd; 186 enum ctrl_sect_len ctrl_len; 187 struct hinic_ctrl *ctrl; 188 u32 saved_data; 189 190 if (data_format == DATA_SGE) { 191 wqe_lcmd = &wqe->wqe_lcmd; 192 193 wqe_lcmd->status.status_info = 0; 194 ctrl = &wqe_lcmd->ctrl; 195 ctrl_len = CTRL_SECT_LEN; 196 } else { 197 wqe_scmd = &wqe->direct_wqe.wqe_scmd; 198 199 wqe_scmd->status.status_info = 0; 200 ctrl = &wqe_scmd->ctrl; 201 ctrl_len = CTRL_DIRECT_SECT_LEN; 202 } 203 204 ctrl->ctrl_info = HINIC_CMDQ_CTRL_SET(prod_idx, PI) | 205 HINIC_CMDQ_CTRL_SET(cmd, CMD) | 206 HINIC_CMDQ_CTRL_SET(mod, MOD) | 207 HINIC_CMDQ_CTRL_SET(ack_type, ACK_TYPE); 208 209 CMDQ_WQE_HEADER(wqe)->header_info = 210 HINIC_CMDQ_WQE_HEADER_SET(buf_len, BUFDESC_LEN) | 211 HINIC_CMDQ_WQE_HEADER_SET(complete_format, COMPLETE_FMT) | 212 HINIC_CMDQ_WQE_HEADER_SET(data_format, DATA_FMT) | 213 HINIC_CMDQ_WQE_HEADER_SET(CEQ_SET, COMPLETE_REQ) | 214 HINIC_CMDQ_WQE_HEADER_SET(COMPLETE_LEN, COMPLETE_SECT_LEN) | 215 HINIC_CMDQ_WQE_HEADER_SET(ctrl_len, CTRL_LEN) | 216 HINIC_CMDQ_WQE_HEADER_SET(wrapped, TOGGLED_WRAPPED); 217 218 saved_data = CMDQ_WQE_HEADER(wqe)->saved_data; 219 saved_data = HINIC_SAVED_DATA_CLEAR(saved_data, ARM); 220 221 if (cmd == CMDQ_SET_ARM_CMD && mod == HINIC_MOD_COMM) 222 CMDQ_WQE_HEADER(wqe)->saved_data |= 223 HINIC_SAVED_DATA_SET(1, ARM); 224 else 225 CMDQ_WQE_HEADER(wqe)->saved_data = saved_data; 226 } 227 228 static void cmdq_set_lcmd_bufdesc(struct hinic_cmdq_wqe_lcmd *wqe_lcmd, 229 struct hinic_cmdq_buf *buf_in) 230 { 231 hinic_set_sge(&wqe_lcmd->buf_desc.sge, buf_in->dma_addr, buf_in->size); 232 } 233 234 static void cmdq_set_direct_wqe_data(struct hinic_cmdq_direct_wqe *wqe, 235 void *buf_in, u32 in_size) 236 { 237 struct hinic_cmdq_wqe_scmd *wqe_scmd = &wqe->wqe_scmd; 238 239 wqe_scmd->buf_desc.buf_len = in_size; 240 memcpy(wqe_scmd->buf_desc.data, buf_in, in_size); 241 } 242 243 static void cmdq_set_lcmd_wqe(struct hinic_cmdq_wqe *wqe, 244 enum cmdq_cmd_type cmd_type, 245 struct hinic_cmdq_buf *buf_in, 246 struct hinic_cmdq_buf *buf_out, int wrapped, 247 enum hinic_cmd_ack_type ack_type, 248 enum hinic_mod_type mod, u8 cmd, u16 prod_idx) 249 { 250 struct hinic_cmdq_wqe_lcmd *wqe_lcmd = &wqe->wqe_lcmd; 251 enum completion_format complete_format; 252 253 switch (cmd_type) { 254 case CMDQ_CMD_SYNC_SGE_RESP: 255 complete_format = COMPLETE_SGE; 256 cmdq_set_sge_completion(&wqe_lcmd->completion, buf_out); 257 break; 258 case CMDQ_CMD_SYNC_DIRECT_RESP: 259 complete_format = COMPLETE_DIRECT; 260 wqe_lcmd->completion.direct_resp = 0; 261 break; 262 } 263 264 cmdq_prepare_wqe_ctrl(wqe, wrapped, ack_type, mod, cmd, 265 prod_idx, complete_format, DATA_SGE, 266 BUFDESC_LCMD_LEN); 267 268 cmdq_set_lcmd_bufdesc(wqe_lcmd, buf_in); 269 } 270 271 static void cmdq_set_direct_wqe(struct hinic_cmdq_wqe *wqe, 272 enum cmdq_cmd_type cmd_type, 273 void *buf_in, u16 in_size, 274 struct hinic_cmdq_buf *buf_out, int wrapped, 275 enum hinic_cmd_ack_type ack_type, 276 enum hinic_mod_type mod, u8 cmd, u16 prod_idx) 277 { 278 struct hinic_cmdq_direct_wqe *direct_wqe = &wqe->direct_wqe; 279 enum completion_format complete_format; 280 struct hinic_cmdq_wqe_scmd *wqe_scmd; 281 282 wqe_scmd = &direct_wqe->wqe_scmd; 283 284 switch (cmd_type) { 285 case CMDQ_CMD_SYNC_SGE_RESP: 286 complete_format = COMPLETE_SGE; 287 cmdq_set_sge_completion(&wqe_scmd->completion, buf_out); 288 break; 289 case CMDQ_CMD_SYNC_DIRECT_RESP: 290 complete_format = COMPLETE_DIRECT; 291 wqe_scmd->completion.direct_resp = 0; 292 break; 293 } 294 295 cmdq_prepare_wqe_ctrl(wqe, wrapped, ack_type, mod, cmd, prod_idx, 296 complete_format, DATA_DIRECT, BUFDESC_SCMD_LEN); 297 298 cmdq_set_direct_wqe_data(direct_wqe, buf_in, in_size); 299 } 300 301 static void cmdq_wqe_fill(void *dst, void *src) 302 { 303 memcpy(dst + FIRST_DATA_TO_WRITE_LAST, src + FIRST_DATA_TO_WRITE_LAST, 304 CMDQ_WQE_SIZE - FIRST_DATA_TO_WRITE_LAST); 305 306 wmb(); /* The first 8 bytes should be written last */ 307 308 *(u64 *)dst = *(u64 *)src; 309 } 310 311 static void cmdq_fill_db(u32 *db_info, 312 enum hinic_cmdq_type cmdq_type, u16 prod_idx) 313 { 314 *db_info = HINIC_CMDQ_DB_INFO_SET(UPPER_8_BITS(prod_idx), HI_PROD_IDX) | 315 HINIC_CMDQ_DB_INFO_SET(HINIC_CTRL_PATH, PATH) | 316 HINIC_CMDQ_DB_INFO_SET(cmdq_type, CMDQ_TYPE) | 317 HINIC_CMDQ_DB_INFO_SET(HINIC_DB_CMDQ_TYPE, DB_TYPE); 318 } 319 320 static void cmdq_set_db(struct hinic_cmdq *cmdq, 321 enum hinic_cmdq_type cmdq_type, u16 prod_idx) 322 { 323 u32 db_info; 324 325 cmdq_fill_db(&db_info, cmdq_type, prod_idx); 326 327 /* The data that is written to HW should be in Big Endian Format */ 328 db_info = cpu_to_be32(db_info); 329 330 wmb(); /* write all before the doorbell */ 331 332 writel(db_info, CMDQ_DB_ADDR(cmdq->db_base, prod_idx)); 333 } 334 335 static int cmdq_sync_cmd_direct_resp(struct hinic_cmdq *cmdq, 336 enum hinic_mod_type mod, u8 cmd, 337 struct hinic_cmdq_buf *buf_in, 338 u64 *resp) 339 { 340 struct hinic_cmdq_wqe *curr_cmdq_wqe, cmdq_wqe; 341 u16 curr_prod_idx, next_prod_idx; 342 int errcode, wrapped, num_wqebbs; 343 struct hinic_wq *wq = cmdq->wq; 344 struct hinic_hw_wqe *hw_wqe; 345 struct completion done; 346 347 /* Keep doorbell index correct. bh - for tasklet(ceq). */ 348 spin_lock_bh(&cmdq->cmdq_lock); 349 350 /* WQE_SIZE = WQEBB_SIZE, we will get the wq element and not shadow*/ 351 hw_wqe = hinic_get_wqe(wq, WQE_LCMD_SIZE, &curr_prod_idx); 352 if (IS_ERR(hw_wqe)) { 353 spin_unlock_bh(&cmdq->cmdq_lock); 354 return -EBUSY; 355 } 356 357 curr_cmdq_wqe = &hw_wqe->cmdq_wqe; 358 359 wrapped = cmdq->wrapped; 360 361 num_wqebbs = ALIGN(WQE_LCMD_SIZE, wq->wqebb_size) / wq->wqebb_size; 362 next_prod_idx = curr_prod_idx + num_wqebbs; 363 if (next_prod_idx >= wq->q_depth) { 364 cmdq->wrapped = !cmdq->wrapped; 365 next_prod_idx -= wq->q_depth; 366 } 367 368 cmdq->errcode[curr_prod_idx] = &errcode; 369 370 init_completion(&done); 371 cmdq->done[curr_prod_idx] = &done; 372 373 cmdq_set_lcmd_wqe(&cmdq_wqe, CMDQ_CMD_SYNC_DIRECT_RESP, buf_in, NULL, 374 wrapped, HINIC_CMD_ACK_TYPE_CMDQ, mod, cmd, 375 curr_prod_idx); 376 377 /* The data that is written to HW should be in Big Endian Format */ 378 hinic_cpu_to_be32(&cmdq_wqe, WQE_LCMD_SIZE); 379 380 /* CMDQ WQE is not shadow, therefore wqe will be written to wq */ 381 cmdq_wqe_fill(curr_cmdq_wqe, &cmdq_wqe); 382 383 cmdq_set_db(cmdq, HINIC_CMDQ_SYNC, next_prod_idx); 384 385 spin_unlock_bh(&cmdq->cmdq_lock); 386 387 if (!wait_for_completion_timeout(&done, 388 msecs_to_jiffies(CMDQ_TIMEOUT))) { 389 spin_lock_bh(&cmdq->cmdq_lock); 390 391 if (cmdq->errcode[curr_prod_idx] == &errcode) 392 cmdq->errcode[curr_prod_idx] = NULL; 393 394 if (cmdq->done[curr_prod_idx] == &done) 395 cmdq->done[curr_prod_idx] = NULL; 396 397 spin_unlock_bh(&cmdq->cmdq_lock); 398 399 hinic_dump_ceq_info(cmdq->hwdev); 400 return -ETIMEDOUT; 401 } 402 403 smp_rmb(); /* read error code after completion */ 404 405 if (resp) { 406 struct hinic_cmdq_wqe_lcmd *wqe_lcmd = &curr_cmdq_wqe->wqe_lcmd; 407 408 *resp = cpu_to_be64(wqe_lcmd->completion.direct_resp); 409 } 410 411 if (errcode != 0) 412 return -EFAULT; 413 414 return 0; 415 } 416 417 static int cmdq_set_arm_bit(struct hinic_cmdq *cmdq, void *buf_in, 418 u16 in_size) 419 { 420 struct hinic_cmdq_wqe *curr_cmdq_wqe, cmdq_wqe; 421 u16 curr_prod_idx, next_prod_idx; 422 struct hinic_wq *wq = cmdq->wq; 423 struct hinic_hw_wqe *hw_wqe; 424 int wrapped, num_wqebbs; 425 426 /* Keep doorbell index correct */ 427 spin_lock(&cmdq->cmdq_lock); 428 429 /* WQE_SIZE = WQEBB_SIZE, we will get the wq element and not shadow*/ 430 hw_wqe = hinic_get_wqe(wq, WQE_SCMD_SIZE, &curr_prod_idx); 431 if (IS_ERR(hw_wqe)) { 432 spin_unlock(&cmdq->cmdq_lock); 433 return -EBUSY; 434 } 435 436 curr_cmdq_wqe = &hw_wqe->cmdq_wqe; 437 438 wrapped = cmdq->wrapped; 439 440 num_wqebbs = ALIGN(WQE_SCMD_SIZE, wq->wqebb_size) / wq->wqebb_size; 441 next_prod_idx = curr_prod_idx + num_wqebbs; 442 if (next_prod_idx >= wq->q_depth) { 443 cmdq->wrapped = !cmdq->wrapped; 444 next_prod_idx -= wq->q_depth; 445 } 446 447 cmdq_set_direct_wqe(&cmdq_wqe, CMDQ_CMD_SYNC_DIRECT_RESP, buf_in, 448 in_size, NULL, wrapped, HINIC_CMD_ACK_TYPE_CMDQ, 449 HINIC_MOD_COMM, CMDQ_SET_ARM_CMD, curr_prod_idx); 450 451 /* The data that is written to HW should be in Big Endian Format */ 452 hinic_cpu_to_be32(&cmdq_wqe, WQE_SCMD_SIZE); 453 454 /* cmdq wqe is not shadow, therefore wqe will be written to wq */ 455 cmdq_wqe_fill(curr_cmdq_wqe, &cmdq_wqe); 456 457 cmdq_set_db(cmdq, HINIC_CMDQ_SYNC, next_prod_idx); 458 459 spin_unlock(&cmdq->cmdq_lock); 460 return 0; 461 } 462 463 static int cmdq_params_valid(struct hinic_cmdq_buf *buf_in) 464 { 465 if (buf_in->size > HINIC_CMDQ_MAX_DATA_SIZE) 466 return -EINVAL; 467 468 return 0; 469 } 470 471 /** 472 * hinic_cmdq_direct_resp - send command with direct data as resp 473 * @cmdqs: the cmdqs 474 * @mod: module on the card that will handle the command 475 * @cmd: the command 476 * @buf_in: the buffer for the command 477 * @resp: the response to return 478 * 479 * Return 0 - Success, negative - Failure 480 **/ 481 int hinic_cmdq_direct_resp(struct hinic_cmdqs *cmdqs, 482 enum hinic_mod_type mod, u8 cmd, 483 struct hinic_cmdq_buf *buf_in, u64 *resp) 484 { 485 struct hinic_hwif *hwif = cmdqs->hwif; 486 struct pci_dev *pdev = hwif->pdev; 487 int err; 488 489 err = cmdq_params_valid(buf_in); 490 if (err) { 491 dev_err(&pdev->dev, "Invalid CMDQ parameters\n"); 492 return err; 493 } 494 495 return cmdq_sync_cmd_direct_resp(&cmdqs->cmdq[HINIC_CMDQ_SYNC], 496 mod, cmd, buf_in, resp); 497 } 498 499 /** 500 * hinic_set_arm_bit - set arm bit for enable interrupt again 501 * @cmdqs: the cmdqs 502 * @q_type: type of queue to set the arm bit for 503 * @q_id: the queue number 504 * 505 * Return 0 - Success, negative - Failure 506 **/ 507 static int hinic_set_arm_bit(struct hinic_cmdqs *cmdqs, 508 enum hinic_set_arm_qtype q_type, u32 q_id) 509 { 510 struct hinic_cmdq *cmdq = &cmdqs->cmdq[HINIC_CMDQ_SYNC]; 511 struct hinic_hwif *hwif = cmdqs->hwif; 512 struct pci_dev *pdev = hwif->pdev; 513 struct hinic_cmdq_arm_bit arm_bit; 514 int err; 515 516 arm_bit.q_type = q_type; 517 arm_bit.q_id = q_id; 518 519 err = cmdq_set_arm_bit(cmdq, &arm_bit, sizeof(arm_bit)); 520 if (err) { 521 dev_err(&pdev->dev, "Failed to set arm for qid %d\n", q_id); 522 return err; 523 } 524 525 return 0; 526 } 527 528 static void clear_wqe_complete_bit(struct hinic_cmdq *cmdq, 529 struct hinic_cmdq_wqe *wqe) 530 { 531 u32 header_info = be32_to_cpu(CMDQ_WQE_HEADER(wqe)->header_info); 532 unsigned int bufdesc_len, wqe_size; 533 struct hinic_ctrl *ctrl; 534 535 bufdesc_len = HINIC_CMDQ_WQE_HEADER_GET(header_info, BUFDESC_LEN); 536 wqe_size = cmdq_wqe_size_from_bdlen(bufdesc_len); 537 if (wqe_size == WQE_LCMD_SIZE) { 538 struct hinic_cmdq_wqe_lcmd *wqe_lcmd = &wqe->wqe_lcmd; 539 540 ctrl = &wqe_lcmd->ctrl; 541 } else { 542 struct hinic_cmdq_direct_wqe *direct_wqe = &wqe->direct_wqe; 543 struct hinic_cmdq_wqe_scmd *wqe_scmd; 544 545 wqe_scmd = &direct_wqe->wqe_scmd; 546 ctrl = &wqe_scmd->ctrl; 547 } 548 549 /* clear HW busy bit */ 550 ctrl->ctrl_info = 0; 551 552 wmb(); /* verify wqe is clear */ 553 } 554 555 /** 556 * cmdq_arm_ceq_handler - cmdq completion event handler for arm command 557 * @cmdq: the cmdq of the arm command 558 * @wqe: the wqe of the arm command 559 * 560 * Return 0 - Success, negative - Failure 561 **/ 562 static int cmdq_arm_ceq_handler(struct hinic_cmdq *cmdq, 563 struct hinic_cmdq_wqe *wqe) 564 { 565 struct hinic_cmdq_direct_wqe *direct_wqe = &wqe->direct_wqe; 566 struct hinic_cmdq_wqe_scmd *wqe_scmd; 567 struct hinic_ctrl *ctrl; 568 u32 ctrl_info; 569 570 wqe_scmd = &direct_wqe->wqe_scmd; 571 ctrl = &wqe_scmd->ctrl; 572 ctrl_info = be32_to_cpu(ctrl->ctrl_info); 573 574 /* HW should toggle the HW BUSY BIT */ 575 if (!CMDQ_WQE_COMPLETED(ctrl_info)) 576 return -EBUSY; 577 578 clear_wqe_complete_bit(cmdq, wqe); 579 580 hinic_put_wqe(cmdq->wq, WQE_SCMD_SIZE); 581 return 0; 582 } 583 584 static void cmdq_update_errcode(struct hinic_cmdq *cmdq, u16 prod_idx, 585 int errcode) 586 { 587 if (cmdq->errcode[prod_idx]) 588 *cmdq->errcode[prod_idx] = errcode; 589 } 590 591 /** 592 * cmdq_sync_cmd_handler - cmdq completion event handler for sync command 593 * @cmdq: the cmdq of the command 594 * @cons_idx: the consumer index to update the error code for 595 * @errcode: the error code 596 **/ 597 static void cmdq_sync_cmd_handler(struct hinic_cmdq *cmdq, u16 cons_idx, 598 int errcode) 599 { 600 u16 prod_idx = cons_idx; 601 602 spin_lock(&cmdq->cmdq_lock); 603 cmdq_update_errcode(cmdq, prod_idx, errcode); 604 605 wmb(); /* write all before update for the command request */ 606 607 if (cmdq->done[prod_idx]) 608 complete(cmdq->done[prod_idx]); 609 spin_unlock(&cmdq->cmdq_lock); 610 } 611 612 static int cmdq_cmd_ceq_handler(struct hinic_cmdq *cmdq, u16 ci, 613 struct hinic_cmdq_wqe *cmdq_wqe) 614 { 615 struct hinic_cmdq_wqe_lcmd *wqe_lcmd = &cmdq_wqe->wqe_lcmd; 616 struct hinic_status *status = &wqe_lcmd->status; 617 struct hinic_ctrl *ctrl = &wqe_lcmd->ctrl; 618 int errcode; 619 620 if (!CMDQ_WQE_COMPLETED(be32_to_cpu(ctrl->ctrl_info))) 621 return -EBUSY; 622 623 dma_rmb(); 624 625 errcode = CMDQ_WQE_ERRCODE_GET(be32_to_cpu(status->status_info), VAL); 626 627 cmdq_sync_cmd_handler(cmdq, ci, errcode); 628 629 clear_wqe_complete_bit(cmdq, cmdq_wqe); 630 hinic_put_wqe(cmdq->wq, WQE_LCMD_SIZE); 631 return 0; 632 } 633 634 /** 635 * cmdq_ceq_handler - cmdq completion event handler 636 * @handle: private data for the handler(cmdqs) 637 * @ceqe_data: ceq element data 638 **/ 639 static void cmdq_ceq_handler(void *handle, u32 ceqe_data) 640 { 641 enum hinic_cmdq_type cmdq_type = CMDQ_CEQE_GET(ceqe_data, TYPE); 642 struct hinic_cmdqs *cmdqs = (struct hinic_cmdqs *)handle; 643 struct hinic_cmdq *cmdq = &cmdqs->cmdq[cmdq_type]; 644 struct hinic_cmdq_header *header; 645 struct hinic_hw_wqe *hw_wqe; 646 int err, set_arm = 0; 647 u32 saved_data; 648 u16 ci; 649 650 /* Read the smallest wqe size for getting wqe size */ 651 while ((hw_wqe = hinic_read_wqe(cmdq->wq, WQE_SCMD_SIZE, &ci))) { 652 if (IS_ERR(hw_wqe)) 653 break; 654 655 header = CMDQ_WQE_HEADER(&hw_wqe->cmdq_wqe); 656 saved_data = be32_to_cpu(header->saved_data); 657 658 if (HINIC_SAVED_DATA_GET(saved_data, ARM)) { 659 /* arm_bit was set until here */ 660 set_arm = 0; 661 662 if (cmdq_arm_ceq_handler(cmdq, &hw_wqe->cmdq_wqe)) 663 break; 664 } else { 665 set_arm = 1; 666 667 hw_wqe = hinic_read_wqe(cmdq->wq, WQE_LCMD_SIZE, &ci); 668 if (IS_ERR(hw_wqe)) 669 break; 670 671 if (cmdq_cmd_ceq_handler(cmdq, ci, &hw_wqe->cmdq_wqe)) 672 break; 673 } 674 } 675 676 if (set_arm) { 677 struct hinic_hwif *hwif = cmdqs->hwif; 678 struct pci_dev *pdev = hwif->pdev; 679 680 err = hinic_set_arm_bit(cmdqs, HINIC_SET_ARM_CMDQ, cmdq_type); 681 if (err) 682 dev_err(&pdev->dev, "Failed to set arm for CMDQ\n"); 683 } 684 } 685 686 /** 687 * cmdq_init_queue_ctxt - init the queue ctxt of a cmdq 688 * @cmdq_ctxt: cmdq ctxt to initialize 689 * @cmdq: the cmdq 690 * @cmdq_pages: the memory of the queue 691 **/ 692 static void cmdq_init_queue_ctxt(struct hinic_cmdq_ctxt *cmdq_ctxt, 693 struct hinic_cmdq *cmdq, 694 struct hinic_cmdq_pages *cmdq_pages) 695 { 696 struct hinic_cmdq_ctxt_info *ctxt_info = &cmdq_ctxt->ctxt_info; 697 u64 wq_first_page_paddr, cmdq_first_block_paddr, pfn; 698 struct hinic_cmdqs *cmdqs = cmdq_to_cmdqs(cmdq); 699 struct hinic_wq *wq = cmdq->wq; 700 701 /* The data in the HW is in Big Endian Format */ 702 wq_first_page_paddr = be64_to_cpu(*wq->block_vaddr); 703 704 pfn = CMDQ_PFN(wq_first_page_paddr, SZ_4K); 705 706 ctxt_info->curr_wqe_page_pfn = 707 HINIC_CMDQ_CTXT_PAGE_INFO_SET(pfn, CURR_WQE_PAGE_PFN) | 708 HINIC_CMDQ_CTXT_PAGE_INFO_SET(HINIC_CEQ_ID_CMDQ, EQ_ID) | 709 HINIC_CMDQ_CTXT_PAGE_INFO_SET(1, CEQ_ARM) | 710 HINIC_CMDQ_CTXT_PAGE_INFO_SET(1, CEQ_EN) | 711 HINIC_CMDQ_CTXT_PAGE_INFO_SET(cmdq->wrapped, WRAPPED); 712 713 if (wq->num_q_pages != 1) { 714 /* block PFN - Read Modify Write */ 715 cmdq_first_block_paddr = cmdq_pages->page_paddr; 716 717 pfn = CMDQ_PFN(cmdq_first_block_paddr, wq->wq_page_size); 718 } 719 720 ctxt_info->wq_block_pfn = 721 HINIC_CMDQ_CTXT_BLOCK_INFO_SET(pfn, WQ_BLOCK_PFN) | 722 HINIC_CMDQ_CTXT_BLOCK_INFO_SET(atomic_read(&wq->cons_idx), CI); 723 724 cmdq_ctxt->func_idx = HINIC_HWIF_FUNC_IDX(cmdqs->hwif); 725 cmdq_ctxt->ppf_idx = HINIC_HWIF_PPF_IDX(cmdqs->hwif); 726 cmdq_ctxt->cmdq_type = cmdq->cmdq_type; 727 } 728 729 /** 730 * init_cmdq - initialize cmdq 731 * @cmdq: the cmdq 732 * @wq: the wq attaced to the cmdq 733 * @q_type: the cmdq type of the cmdq 734 * @db_area: doorbell area for the cmdq 735 * 736 * Return 0 - Success, negative - Failure 737 **/ 738 static int init_cmdq(struct hinic_cmdq *cmdq, struct hinic_wq *wq, 739 enum hinic_cmdq_type q_type, void __iomem *db_area) 740 { 741 int err; 742 743 cmdq->wq = wq; 744 cmdq->cmdq_type = q_type; 745 cmdq->wrapped = 1; 746 747 spin_lock_init(&cmdq->cmdq_lock); 748 749 cmdq->done = vzalloc(array_size(sizeof(*cmdq->done), wq->q_depth)); 750 if (!cmdq->done) 751 return -ENOMEM; 752 753 cmdq->errcode = vzalloc(array_size(sizeof(*cmdq->errcode), 754 wq->q_depth)); 755 if (!cmdq->errcode) { 756 err = -ENOMEM; 757 goto err_errcode; 758 } 759 760 cmdq->db_base = db_area + CMDQ_DB_OFF; 761 return 0; 762 763 err_errcode: 764 vfree(cmdq->done); 765 return err; 766 } 767 768 /** 769 * free_cmdq - Free cmdq 770 * @cmdq: the cmdq to free 771 **/ 772 static void free_cmdq(struct hinic_cmdq *cmdq) 773 { 774 vfree(cmdq->errcode); 775 vfree(cmdq->done); 776 } 777 778 /** 779 * init_cmdqs_ctxt - write the cmdq ctxt to HW after init all cmdq 780 * @hwdev: the NIC HW device 781 * @cmdqs: cmdqs to write the ctxts for 782 * @db_area: db_area for all the cmdqs 783 * 784 * Return 0 - Success, negative - Failure 785 **/ 786 static int init_cmdqs_ctxt(struct hinic_hwdev *hwdev, 787 struct hinic_cmdqs *cmdqs, void __iomem **db_area) 788 { 789 struct hinic_hwif *hwif = hwdev->hwif; 790 enum hinic_cmdq_type type, cmdq_type; 791 struct hinic_cmdq_ctxt *cmdq_ctxts; 792 struct pci_dev *pdev = hwif->pdev; 793 struct hinic_pfhwdev *pfhwdev; 794 int err; 795 796 cmdq_ctxts = devm_kcalloc(&pdev->dev, HINIC_MAX_CMDQ_TYPES, 797 sizeof(*cmdq_ctxts), GFP_KERNEL); 798 if (!cmdq_ctxts) 799 return -ENOMEM; 800 801 pfhwdev = container_of(hwdev, struct hinic_pfhwdev, hwdev); 802 803 cmdq_type = HINIC_CMDQ_SYNC; 804 for (; cmdq_type < HINIC_MAX_CMDQ_TYPES; cmdq_type++) { 805 cmdqs->cmdq[cmdq_type].hwdev = hwdev; 806 err = init_cmdq(&cmdqs->cmdq[cmdq_type], 807 &cmdqs->saved_wqs[cmdq_type], cmdq_type, 808 db_area[cmdq_type]); 809 if (err) { 810 dev_err(&pdev->dev, "Failed to initialize cmdq\n"); 811 goto err_init_cmdq; 812 } 813 814 cmdq_init_queue_ctxt(&cmdq_ctxts[cmdq_type], 815 &cmdqs->cmdq[cmdq_type], 816 &cmdqs->cmdq_pages); 817 } 818 819 /* Write the CMDQ ctxts */ 820 cmdq_type = HINIC_CMDQ_SYNC; 821 for (; cmdq_type < HINIC_MAX_CMDQ_TYPES; cmdq_type++) { 822 err = hinic_msg_to_mgmt(&pfhwdev->pf_to_mgmt, HINIC_MOD_COMM, 823 HINIC_COMM_CMD_CMDQ_CTXT_SET, 824 &cmdq_ctxts[cmdq_type], 825 sizeof(cmdq_ctxts[cmdq_type]), 826 NULL, NULL, HINIC_MGMT_MSG_SYNC); 827 if (err) { 828 dev_err(&pdev->dev, "Failed to set CMDQ CTXT type = %d\n", 829 cmdq_type); 830 goto err_write_cmdq_ctxt; 831 } 832 } 833 834 devm_kfree(&pdev->dev, cmdq_ctxts); 835 return 0; 836 837 err_write_cmdq_ctxt: 838 cmdq_type = HINIC_MAX_CMDQ_TYPES; 839 840 err_init_cmdq: 841 for (type = HINIC_CMDQ_SYNC; type < cmdq_type; type++) 842 free_cmdq(&cmdqs->cmdq[type]); 843 844 devm_kfree(&pdev->dev, cmdq_ctxts); 845 return err; 846 } 847 848 static int hinic_set_cmdq_depth(struct hinic_hwdev *hwdev, u16 cmdq_depth) 849 { 850 struct hinic_cmd_hw_ioctxt hw_ioctxt = { 0 }; 851 struct hinic_pfhwdev *pfhwdev; 852 853 pfhwdev = container_of(hwdev, struct hinic_pfhwdev, hwdev); 854 855 hw_ioctxt.func_idx = HINIC_HWIF_FUNC_IDX(hwdev->hwif); 856 hw_ioctxt.ppf_idx = HINIC_HWIF_PPF_IDX(hwdev->hwif); 857 858 hw_ioctxt.set_cmdq_depth = HW_IOCTXT_SET_CMDQ_DEPTH_ENABLE; 859 hw_ioctxt.cmdq_depth = (u8)ilog2(cmdq_depth); 860 861 return hinic_msg_to_mgmt(&pfhwdev->pf_to_mgmt, HINIC_MOD_COMM, 862 HINIC_COMM_CMD_HWCTXT_SET, 863 &hw_ioctxt, sizeof(hw_ioctxt), NULL, 864 NULL, HINIC_MGMT_MSG_SYNC); 865 } 866 867 /** 868 * hinic_init_cmdqs - init all cmdqs 869 * @cmdqs: cmdqs to init 870 * @hwif: HW interface for accessing cmdqs 871 * @db_area: doorbell areas for all the cmdqs 872 * 873 * Return 0 - Success, negative - Failure 874 **/ 875 int hinic_init_cmdqs(struct hinic_cmdqs *cmdqs, struct hinic_hwif *hwif, 876 void __iomem **db_area) 877 { 878 struct hinic_func_to_io *func_to_io = cmdqs_to_func_to_io(cmdqs); 879 struct pci_dev *pdev = hwif->pdev; 880 struct hinic_hwdev *hwdev; 881 u16 max_wqe_size; 882 int err; 883 884 cmdqs->hwif = hwif; 885 cmdqs->cmdq_buf_pool = dma_pool_create("hinic_cmdq", &pdev->dev, 886 HINIC_CMDQ_BUF_SIZE, 887 HINIC_CMDQ_BUF_SIZE, 0); 888 if (!cmdqs->cmdq_buf_pool) 889 return -ENOMEM; 890 891 cmdqs->saved_wqs = devm_kcalloc(&pdev->dev, HINIC_MAX_CMDQ_TYPES, 892 sizeof(*cmdqs->saved_wqs), GFP_KERNEL); 893 if (!cmdqs->saved_wqs) { 894 err = -ENOMEM; 895 goto err_saved_wqs; 896 } 897 898 max_wqe_size = WQE_LCMD_SIZE; 899 err = hinic_wqs_cmdq_alloc(&cmdqs->cmdq_pages, cmdqs->saved_wqs, hwif, 900 HINIC_MAX_CMDQ_TYPES, CMDQ_WQEBB_SIZE, 901 CMDQ_WQ_PAGE_SIZE, CMDQ_DEPTH, max_wqe_size); 902 if (err) { 903 dev_err(&pdev->dev, "Failed to allocate CMDQ wqs\n"); 904 goto err_cmdq_wqs; 905 } 906 907 hwdev = container_of(func_to_io, struct hinic_hwdev, func_to_io); 908 err = init_cmdqs_ctxt(hwdev, cmdqs, db_area); 909 if (err) { 910 dev_err(&pdev->dev, "Failed to write cmdq ctxt\n"); 911 goto err_cmdq_ctxt; 912 } 913 914 hinic_ceq_register_cb(&func_to_io->ceqs, HINIC_CEQ_CMDQ, cmdqs, 915 cmdq_ceq_handler); 916 917 err = hinic_set_cmdq_depth(hwdev, CMDQ_DEPTH); 918 if (err) { 919 dev_err(&hwif->pdev->dev, "Failed to set cmdq depth\n"); 920 goto err_set_cmdq_depth; 921 } 922 923 return 0; 924 925 err_set_cmdq_depth: 926 hinic_ceq_unregister_cb(&func_to_io->ceqs, HINIC_CEQ_CMDQ); 927 free_cmdq(&cmdqs->cmdq[HINIC_CMDQ_SYNC]); 928 err_cmdq_ctxt: 929 hinic_wqs_cmdq_free(&cmdqs->cmdq_pages, cmdqs->saved_wqs, 930 HINIC_MAX_CMDQ_TYPES); 931 932 err_cmdq_wqs: 933 devm_kfree(&pdev->dev, cmdqs->saved_wqs); 934 935 err_saved_wqs: 936 dma_pool_destroy(cmdqs->cmdq_buf_pool); 937 return err; 938 } 939 940 /** 941 * hinic_free_cmdqs - free all cmdqs 942 * @cmdqs: cmdqs to free 943 **/ 944 void hinic_free_cmdqs(struct hinic_cmdqs *cmdqs) 945 { 946 struct hinic_func_to_io *func_to_io = cmdqs_to_func_to_io(cmdqs); 947 struct hinic_hwif *hwif = cmdqs->hwif; 948 struct pci_dev *pdev = hwif->pdev; 949 enum hinic_cmdq_type cmdq_type; 950 951 hinic_ceq_unregister_cb(&func_to_io->ceqs, HINIC_CEQ_CMDQ); 952 953 cmdq_type = HINIC_CMDQ_SYNC; 954 for (; cmdq_type < HINIC_MAX_CMDQ_TYPES; cmdq_type++) 955 free_cmdq(&cmdqs->cmdq[cmdq_type]); 956 957 hinic_wqs_cmdq_free(&cmdqs->cmdq_pages, cmdqs->saved_wqs, 958 HINIC_MAX_CMDQ_TYPES); 959 960 devm_kfree(&pdev->dev, cmdqs->saved_wqs); 961 962 dma_pool_destroy(cmdqs->cmdq_buf_pool); 963 } 964