1 /* 2 * QLogic qlcnic NIC Driver 3 * Copyright (c) 2009-2013 QLogic Corporation 4 * 5 * See LICENSE.qlcnic for copyright and licensing details. 6 */ 7 8 #include <net/ip.h> 9 10 #include "qlcnic.h" 11 #include "qlcnic_hdr.h" 12 #include "qlcnic_83xx_hw.h" 13 #include "qlcnic_hw.h" 14 15 #define QLC_83XX_MINIDUMP_FLASH 0x520000 16 #define QLC_83XX_OCM_INDEX 3 17 #define QLC_83XX_PCI_INDEX 0 18 #define QLC_83XX_DMA_ENGINE_INDEX 8 19 20 static const u32 qlcnic_ms_read_data[] = { 21 0x410000A8, 0x410000AC, 0x410000B8, 0x410000BC 22 }; 23 24 #define QLCNIC_DUMP_WCRB BIT_0 25 #define QLCNIC_DUMP_RWCRB BIT_1 26 #define QLCNIC_DUMP_ANDCRB BIT_2 27 #define QLCNIC_DUMP_ORCRB BIT_3 28 #define QLCNIC_DUMP_POLLCRB BIT_4 29 #define QLCNIC_DUMP_RD_SAVE BIT_5 30 #define QLCNIC_DUMP_WRT_SAVED BIT_6 31 #define QLCNIC_DUMP_MOD_SAVE_ST BIT_7 32 #define QLCNIC_DUMP_SKIP BIT_7 33 34 #define QLCNIC_DUMP_MASK_MAX 0xff 35 36 struct qlcnic_pex_dma_descriptor { 37 u32 read_data_size; 38 u32 dma_desc_cmd; 39 u32 src_addr_low; 40 u32 src_addr_high; 41 u32 dma_bus_addr_low; 42 u32 dma_bus_addr_high; 43 u32 rsvd[6]; 44 } __packed; 45 46 struct qlcnic_common_entry_hdr { 47 u32 type; 48 u32 offset; 49 u32 cap_size; 50 #if defined(__LITTLE_ENDIAN) 51 u8 mask; 52 u8 rsvd[2]; 53 u8 flags; 54 #else 55 u8 flags; 56 u8 rsvd[2]; 57 u8 mask; 58 #endif 59 } __packed; 60 61 struct __crb { 62 u32 addr; 63 #if defined(__LITTLE_ENDIAN) 64 u8 stride; 65 u8 rsvd1[3]; 66 #else 67 u8 rsvd1[3]; 68 u8 stride; 69 #endif 70 u32 data_size; 71 u32 no_ops; 72 u32 rsvd2[4]; 73 } __packed; 74 75 struct __ctrl { 76 u32 addr; 77 #if defined(__LITTLE_ENDIAN) 78 u8 stride; 79 u8 index_a; 80 u16 timeout; 81 #else 82 u16 timeout; 83 u8 index_a; 84 u8 stride; 85 #endif 86 u32 data_size; 87 u32 no_ops; 88 #if defined(__LITTLE_ENDIAN) 89 u8 opcode; 90 u8 index_v; 91 u8 shl_val; 92 u8 shr_val; 93 #else 94 u8 shr_val; 95 u8 shl_val; 96 u8 index_v; 97 u8 opcode; 98 #endif 99 u32 val1; 100 u32 val2; 101 u32 val3; 102 } __packed; 103 104 struct __cache { 105 u32 addr; 106 #if defined(__LITTLE_ENDIAN) 107 u16 stride; 108 u16 init_tag_val; 109 #else 110 u16 init_tag_val; 111 u16 stride; 112 #endif 113 u32 size; 114 u32 no_ops; 115 u32 ctrl_addr; 116 u32 ctrl_val; 117 u32 read_addr; 118 #if defined(__LITTLE_ENDIAN) 119 u8 read_addr_stride; 120 u8 read_addr_num; 121 u8 rsvd1[2]; 122 #else 123 u8 rsvd1[2]; 124 u8 read_addr_num; 125 u8 read_addr_stride; 126 #endif 127 } __packed; 128 129 struct __ocm { 130 u8 rsvd[8]; 131 u32 size; 132 u32 no_ops; 133 u8 rsvd1[8]; 134 u32 read_addr; 135 u32 read_addr_stride; 136 } __packed; 137 138 struct __mem { 139 u32 desc_card_addr; 140 u32 dma_desc_cmd; 141 u32 start_dma_cmd; 142 u32 rsvd[3]; 143 u32 addr; 144 u32 size; 145 } __packed; 146 147 struct __mux { 148 u32 addr; 149 u8 rsvd[4]; 150 u32 size; 151 u32 no_ops; 152 u32 val; 153 u32 val_stride; 154 u32 read_addr; 155 u8 rsvd2[4]; 156 } __packed; 157 158 struct __queue { 159 u32 sel_addr; 160 #if defined(__LITTLE_ENDIAN) 161 u16 stride; 162 u8 rsvd[2]; 163 #else 164 u8 rsvd[2]; 165 u16 stride; 166 #endif 167 u32 size; 168 u32 no_ops; 169 u8 rsvd2[8]; 170 u32 read_addr; 171 #if defined(__LITTLE_ENDIAN) 172 u8 read_addr_stride; 173 u8 read_addr_cnt; 174 u8 rsvd3[2]; 175 #else 176 u8 rsvd3[2]; 177 u8 read_addr_cnt; 178 u8 read_addr_stride; 179 #endif 180 } __packed; 181 182 struct __pollrd { 183 u32 sel_addr; 184 u32 read_addr; 185 u32 sel_val; 186 #if defined(__LITTLE_ENDIAN) 187 u16 sel_val_stride; 188 u16 no_ops; 189 #else 190 u16 no_ops; 191 u16 sel_val_stride; 192 #endif 193 u32 poll_wait; 194 u32 poll_mask; 195 u32 data_size; 196 u8 rsvd[4]; 197 } __packed; 198 199 struct __mux2 { 200 u32 sel_addr1; 201 u32 sel_addr2; 202 u32 sel_val1; 203 u32 sel_val2; 204 u32 no_ops; 205 u32 sel_val_mask; 206 u32 read_addr; 207 #if defined(__LITTLE_ENDIAN) 208 u8 sel_val_stride; 209 u8 data_size; 210 u8 rsvd[2]; 211 #else 212 u8 rsvd[2]; 213 u8 data_size; 214 u8 sel_val_stride; 215 #endif 216 } __packed; 217 218 struct __pollrdmwr { 219 u32 addr1; 220 u32 addr2; 221 u32 val1; 222 u32 val2; 223 u32 poll_wait; 224 u32 poll_mask; 225 u32 mod_mask; 226 u32 data_size; 227 } __packed; 228 229 struct qlcnic_dump_entry { 230 struct qlcnic_common_entry_hdr hdr; 231 union { 232 struct __crb crb; 233 struct __cache cache; 234 struct __ocm ocm; 235 struct __mem mem; 236 struct __mux mux; 237 struct __queue que; 238 struct __ctrl ctrl; 239 struct __pollrdmwr pollrdmwr; 240 struct __mux2 mux2; 241 struct __pollrd pollrd; 242 } region; 243 } __packed; 244 245 enum qlcnic_minidump_opcode { 246 QLCNIC_DUMP_NOP = 0, 247 QLCNIC_DUMP_READ_CRB = 1, 248 QLCNIC_DUMP_READ_MUX = 2, 249 QLCNIC_DUMP_QUEUE = 3, 250 QLCNIC_DUMP_BRD_CONFIG = 4, 251 QLCNIC_DUMP_READ_OCM = 6, 252 QLCNIC_DUMP_PEG_REG = 7, 253 QLCNIC_DUMP_L1_DTAG = 8, 254 QLCNIC_DUMP_L1_ITAG = 9, 255 QLCNIC_DUMP_L1_DATA = 11, 256 QLCNIC_DUMP_L1_INST = 12, 257 QLCNIC_DUMP_L2_DTAG = 21, 258 QLCNIC_DUMP_L2_ITAG = 22, 259 QLCNIC_DUMP_L2_DATA = 23, 260 QLCNIC_DUMP_L2_INST = 24, 261 QLCNIC_DUMP_POLL_RD = 35, 262 QLCNIC_READ_MUX2 = 36, 263 QLCNIC_READ_POLLRDMWR = 37, 264 QLCNIC_DUMP_READ_ROM = 71, 265 QLCNIC_DUMP_READ_MEM = 72, 266 QLCNIC_DUMP_READ_CTRL = 98, 267 QLCNIC_DUMP_TLHDR = 99, 268 QLCNIC_DUMP_RDEND = 255 269 }; 270 271 inline u32 qlcnic_82xx_get_saved_state(void *t_hdr, u32 index) 272 { 273 struct qlcnic_82xx_dump_template_hdr *hdr = t_hdr; 274 275 return hdr->saved_state[index]; 276 } 277 278 inline void qlcnic_82xx_set_saved_state(void *t_hdr, u32 index, 279 u32 value) 280 { 281 struct qlcnic_82xx_dump_template_hdr *hdr = t_hdr; 282 283 hdr->saved_state[index] = value; 284 } 285 286 void qlcnic_82xx_cache_tmpl_hdr_values(struct qlcnic_fw_dump *fw_dump) 287 { 288 struct qlcnic_82xx_dump_template_hdr *hdr; 289 290 hdr = fw_dump->tmpl_hdr; 291 fw_dump->tmpl_hdr_size = hdr->size; 292 fw_dump->version = hdr->version; 293 fw_dump->num_entries = hdr->num_entries; 294 fw_dump->offset = hdr->offset; 295 296 hdr->drv_cap_mask = hdr->cap_mask; 297 fw_dump->cap_mask = hdr->cap_mask; 298 299 fw_dump->use_pex_dma = (hdr->capabilities & BIT_0) ? true : false; 300 } 301 302 inline u32 qlcnic_82xx_get_cap_size(void *t_hdr, int index) 303 { 304 struct qlcnic_82xx_dump_template_hdr *hdr = t_hdr; 305 306 return hdr->cap_sizes[index]; 307 } 308 309 void qlcnic_82xx_set_sys_info(void *t_hdr, int idx, u32 value) 310 { 311 struct qlcnic_82xx_dump_template_hdr *hdr = t_hdr; 312 313 hdr->sys_info[idx] = value; 314 } 315 316 void qlcnic_82xx_store_cap_mask(void *tmpl_hdr, u32 mask) 317 { 318 struct qlcnic_82xx_dump_template_hdr *hdr = tmpl_hdr; 319 320 hdr->drv_cap_mask = mask; 321 } 322 323 inline u32 qlcnic_83xx_get_saved_state(void *t_hdr, u32 index) 324 { 325 struct qlcnic_83xx_dump_template_hdr *hdr = t_hdr; 326 327 return hdr->saved_state[index]; 328 } 329 330 inline void qlcnic_83xx_set_saved_state(void *t_hdr, u32 index, 331 u32 value) 332 { 333 struct qlcnic_83xx_dump_template_hdr *hdr = t_hdr; 334 335 hdr->saved_state[index] = value; 336 } 337 338 #define QLCNIC_TEMPLATE_VERSION (0x20001) 339 340 void qlcnic_83xx_cache_tmpl_hdr_values(struct qlcnic_fw_dump *fw_dump) 341 { 342 struct qlcnic_83xx_dump_template_hdr *hdr; 343 344 hdr = fw_dump->tmpl_hdr; 345 fw_dump->tmpl_hdr_size = hdr->size; 346 fw_dump->version = hdr->version; 347 fw_dump->num_entries = hdr->num_entries; 348 fw_dump->offset = hdr->offset; 349 350 hdr->drv_cap_mask = hdr->cap_mask; 351 fw_dump->cap_mask = hdr->cap_mask; 352 353 fw_dump->use_pex_dma = (fw_dump->version & 0xfffff) >= 354 QLCNIC_TEMPLATE_VERSION; 355 } 356 357 inline u32 qlcnic_83xx_get_cap_size(void *t_hdr, int index) 358 { 359 struct qlcnic_83xx_dump_template_hdr *hdr = t_hdr; 360 361 return hdr->cap_sizes[index]; 362 } 363 364 void qlcnic_83xx_set_sys_info(void *t_hdr, int idx, u32 value) 365 { 366 struct qlcnic_83xx_dump_template_hdr *hdr = t_hdr; 367 368 hdr->sys_info[idx] = value; 369 } 370 371 void qlcnic_83xx_store_cap_mask(void *tmpl_hdr, u32 mask) 372 { 373 struct qlcnic_83xx_dump_template_hdr *hdr; 374 375 hdr = tmpl_hdr; 376 hdr->drv_cap_mask = mask; 377 } 378 379 struct qlcnic_dump_operations { 380 enum qlcnic_minidump_opcode opcode; 381 u32 (*handler)(struct qlcnic_adapter *, struct qlcnic_dump_entry *, 382 __le32 *); 383 }; 384 385 static u32 qlcnic_dump_crb(struct qlcnic_adapter *adapter, 386 struct qlcnic_dump_entry *entry, __le32 *buffer) 387 { 388 int i; 389 u32 addr, data; 390 struct __crb *crb = &entry->region.crb; 391 392 addr = crb->addr; 393 394 for (i = 0; i < crb->no_ops; i++) { 395 data = qlcnic_ind_rd(adapter, addr); 396 *buffer++ = cpu_to_le32(addr); 397 *buffer++ = cpu_to_le32(data); 398 addr += crb->stride; 399 } 400 return crb->no_ops * 2 * sizeof(u32); 401 } 402 403 static u32 qlcnic_dump_ctrl(struct qlcnic_adapter *adapter, 404 struct qlcnic_dump_entry *entry, __le32 *buffer) 405 { 406 void *hdr = adapter->ahw->fw_dump.tmpl_hdr; 407 struct __ctrl *ctr = &entry->region.ctrl; 408 int i, k, timeout = 0; 409 u32 addr, data, temp; 410 u8 no_ops; 411 412 addr = ctr->addr; 413 no_ops = ctr->no_ops; 414 415 for (i = 0; i < no_ops; i++) { 416 k = 0; 417 for (k = 0; k < 8; k++) { 418 if (!(ctr->opcode & (1 << k))) 419 continue; 420 switch (1 << k) { 421 case QLCNIC_DUMP_WCRB: 422 qlcnic_ind_wr(adapter, addr, ctr->val1); 423 break; 424 case QLCNIC_DUMP_RWCRB: 425 data = qlcnic_ind_rd(adapter, addr); 426 qlcnic_ind_wr(adapter, addr, data); 427 break; 428 case QLCNIC_DUMP_ANDCRB: 429 data = qlcnic_ind_rd(adapter, addr); 430 qlcnic_ind_wr(adapter, addr, 431 (data & ctr->val2)); 432 break; 433 case QLCNIC_DUMP_ORCRB: 434 data = qlcnic_ind_rd(adapter, addr); 435 qlcnic_ind_wr(adapter, addr, 436 (data | ctr->val3)); 437 break; 438 case QLCNIC_DUMP_POLLCRB: 439 while (timeout <= ctr->timeout) { 440 data = qlcnic_ind_rd(adapter, addr); 441 if ((data & ctr->val2) == ctr->val1) 442 break; 443 usleep_range(1000, 2000); 444 timeout++; 445 } 446 if (timeout > ctr->timeout) { 447 dev_info(&adapter->pdev->dev, 448 "Timed out, aborting poll CRB\n"); 449 return -EINVAL; 450 } 451 break; 452 case QLCNIC_DUMP_RD_SAVE: 453 temp = ctr->index_a; 454 if (temp) 455 addr = qlcnic_get_saved_state(adapter, 456 hdr, 457 temp); 458 data = qlcnic_ind_rd(adapter, addr); 459 qlcnic_set_saved_state(adapter, hdr, 460 ctr->index_v, data); 461 break; 462 case QLCNIC_DUMP_WRT_SAVED: 463 temp = ctr->index_v; 464 if (temp) 465 data = qlcnic_get_saved_state(adapter, 466 hdr, 467 temp); 468 else 469 data = ctr->val1; 470 471 temp = ctr->index_a; 472 if (temp) 473 addr = qlcnic_get_saved_state(adapter, 474 hdr, 475 temp); 476 qlcnic_ind_wr(adapter, addr, data); 477 break; 478 case QLCNIC_DUMP_MOD_SAVE_ST: 479 data = qlcnic_get_saved_state(adapter, hdr, 480 ctr->index_v); 481 data <<= ctr->shl_val; 482 data >>= ctr->shr_val; 483 if (ctr->val2) 484 data &= ctr->val2; 485 data |= ctr->val3; 486 data += ctr->val1; 487 qlcnic_set_saved_state(adapter, hdr, 488 ctr->index_v, data); 489 break; 490 default: 491 dev_info(&adapter->pdev->dev, 492 "Unknown opcode\n"); 493 break; 494 } 495 } 496 addr += ctr->stride; 497 } 498 return 0; 499 } 500 501 static u32 qlcnic_dump_mux(struct qlcnic_adapter *adapter, 502 struct qlcnic_dump_entry *entry, __le32 *buffer) 503 { 504 int loop; 505 u32 val, data = 0; 506 struct __mux *mux = &entry->region.mux; 507 508 val = mux->val; 509 for (loop = 0; loop < mux->no_ops; loop++) { 510 qlcnic_ind_wr(adapter, mux->addr, val); 511 data = qlcnic_ind_rd(adapter, mux->read_addr); 512 *buffer++ = cpu_to_le32(val); 513 *buffer++ = cpu_to_le32(data); 514 val += mux->val_stride; 515 } 516 return 2 * mux->no_ops * sizeof(u32); 517 } 518 519 static u32 qlcnic_dump_que(struct qlcnic_adapter *adapter, 520 struct qlcnic_dump_entry *entry, __le32 *buffer) 521 { 522 int i, loop; 523 u32 cnt, addr, data, que_id = 0; 524 struct __queue *que = &entry->region.que; 525 526 addr = que->read_addr; 527 cnt = que->read_addr_cnt; 528 529 for (loop = 0; loop < que->no_ops; loop++) { 530 qlcnic_ind_wr(adapter, que->sel_addr, que_id); 531 addr = que->read_addr; 532 for (i = 0; i < cnt; i++) { 533 data = qlcnic_ind_rd(adapter, addr); 534 *buffer++ = cpu_to_le32(data); 535 addr += que->read_addr_stride; 536 } 537 que_id += que->stride; 538 } 539 return que->no_ops * cnt * sizeof(u32); 540 } 541 542 static u32 qlcnic_dump_ocm(struct qlcnic_adapter *adapter, 543 struct qlcnic_dump_entry *entry, __le32 *buffer) 544 { 545 int i; 546 u32 data; 547 void __iomem *addr; 548 struct __ocm *ocm = &entry->region.ocm; 549 550 addr = adapter->ahw->pci_base0 + ocm->read_addr; 551 for (i = 0; i < ocm->no_ops; i++) { 552 data = readl(addr); 553 *buffer++ = cpu_to_le32(data); 554 addr += ocm->read_addr_stride; 555 } 556 return ocm->no_ops * sizeof(u32); 557 } 558 559 static u32 qlcnic_read_rom(struct qlcnic_adapter *adapter, 560 struct qlcnic_dump_entry *entry, __le32 *buffer) 561 { 562 int i, count = 0; 563 u32 fl_addr, size, val, lck_val, addr; 564 struct __mem *rom = &entry->region.mem; 565 566 fl_addr = rom->addr; 567 size = rom->size / 4; 568 lock_try: 569 lck_val = QLC_SHARED_REG_RD32(adapter, QLCNIC_FLASH_LOCK); 570 if (!lck_val && count < MAX_CTL_CHECK) { 571 usleep_range(10000, 11000); 572 count++; 573 goto lock_try; 574 } 575 QLC_SHARED_REG_WR32(adapter, QLCNIC_FLASH_LOCK_OWNER, 576 adapter->ahw->pci_func); 577 for (i = 0; i < size; i++) { 578 addr = fl_addr & 0xFFFF0000; 579 qlcnic_ind_wr(adapter, FLASH_ROM_WINDOW, addr); 580 addr = LSW(fl_addr) + FLASH_ROM_DATA; 581 val = qlcnic_ind_rd(adapter, addr); 582 fl_addr += 4; 583 *buffer++ = cpu_to_le32(val); 584 } 585 QLC_SHARED_REG_RD32(adapter, QLCNIC_FLASH_UNLOCK); 586 return rom->size; 587 } 588 589 static u32 qlcnic_dump_l1_cache(struct qlcnic_adapter *adapter, 590 struct qlcnic_dump_entry *entry, __le32 *buffer) 591 { 592 int i; 593 u32 cnt, val, data, addr; 594 struct __cache *l1 = &entry->region.cache; 595 596 val = l1->init_tag_val; 597 598 for (i = 0; i < l1->no_ops; i++) { 599 qlcnic_ind_wr(adapter, l1->addr, val); 600 qlcnic_ind_wr(adapter, l1->ctrl_addr, LSW(l1->ctrl_val)); 601 addr = l1->read_addr; 602 cnt = l1->read_addr_num; 603 while (cnt) { 604 data = qlcnic_ind_rd(adapter, addr); 605 *buffer++ = cpu_to_le32(data); 606 addr += l1->read_addr_stride; 607 cnt--; 608 } 609 val += l1->stride; 610 } 611 return l1->no_ops * l1->read_addr_num * sizeof(u32); 612 } 613 614 static u32 qlcnic_dump_l2_cache(struct qlcnic_adapter *adapter, 615 struct qlcnic_dump_entry *entry, __le32 *buffer) 616 { 617 int i; 618 u32 cnt, val, data, addr; 619 u8 poll_mask, poll_to, time_out = 0; 620 struct __cache *l2 = &entry->region.cache; 621 622 val = l2->init_tag_val; 623 poll_mask = LSB(MSW(l2->ctrl_val)); 624 poll_to = MSB(MSW(l2->ctrl_val)); 625 626 for (i = 0; i < l2->no_ops; i++) { 627 qlcnic_ind_wr(adapter, l2->addr, val); 628 if (LSW(l2->ctrl_val)) 629 qlcnic_ind_wr(adapter, l2->ctrl_addr, 630 LSW(l2->ctrl_val)); 631 if (!poll_mask) 632 goto skip_poll; 633 do { 634 data = qlcnic_ind_rd(adapter, l2->ctrl_addr); 635 if (!(data & poll_mask)) 636 break; 637 usleep_range(1000, 2000); 638 time_out++; 639 } while (time_out <= poll_to); 640 641 if (time_out > poll_to) { 642 dev_err(&adapter->pdev->dev, 643 "Timeout exceeded in %s, aborting dump\n", 644 __func__); 645 return -EINVAL; 646 } 647 skip_poll: 648 addr = l2->read_addr; 649 cnt = l2->read_addr_num; 650 while (cnt) { 651 data = qlcnic_ind_rd(adapter, addr); 652 *buffer++ = cpu_to_le32(data); 653 addr += l2->read_addr_stride; 654 cnt--; 655 } 656 val += l2->stride; 657 } 658 return l2->no_ops * l2->read_addr_num * sizeof(u32); 659 } 660 661 static u32 qlcnic_read_memory_test_agent(struct qlcnic_adapter *adapter, 662 struct __mem *mem, __le32 *buffer, 663 int *ret) 664 { 665 u32 addr, data, test; 666 int i, reg_read; 667 668 reg_read = mem->size; 669 addr = mem->addr; 670 /* check for data size of multiple of 16 and 16 byte alignment */ 671 if ((addr & 0xf) || (reg_read%16)) { 672 dev_info(&adapter->pdev->dev, 673 "Unaligned memory addr:0x%x size:0x%x\n", 674 addr, reg_read); 675 *ret = -EINVAL; 676 return 0; 677 } 678 679 mutex_lock(&adapter->ahw->mem_lock); 680 681 while (reg_read != 0) { 682 qlcnic_ind_wr(adapter, QLCNIC_MS_ADDR_LO, addr); 683 qlcnic_ind_wr(adapter, QLCNIC_MS_ADDR_HI, 0); 684 qlcnic_ind_wr(adapter, QLCNIC_MS_CTRL, QLCNIC_TA_START_ENABLE); 685 686 for (i = 0; i < MAX_CTL_CHECK; i++) { 687 test = qlcnic_ind_rd(adapter, QLCNIC_MS_CTRL); 688 if (!(test & TA_CTL_BUSY)) 689 break; 690 } 691 if (i == MAX_CTL_CHECK) { 692 if (printk_ratelimit()) { 693 dev_err(&adapter->pdev->dev, 694 "failed to read through agent\n"); 695 *ret = -EIO; 696 goto out; 697 } 698 } 699 for (i = 0; i < 4; i++) { 700 data = qlcnic_ind_rd(adapter, qlcnic_ms_read_data[i]); 701 *buffer++ = cpu_to_le32(data); 702 } 703 addr += 16; 704 reg_read -= 16; 705 ret += 16; 706 } 707 out: 708 mutex_unlock(&adapter->ahw->mem_lock); 709 return mem->size; 710 } 711 712 /* DMA register base address */ 713 #define QLC_DMA_REG_BASE_ADDR(dma_no) (0x77320000 + (dma_no * 0x10000)) 714 715 /* DMA register offsets w.r.t base address */ 716 #define QLC_DMA_CMD_BUFF_ADDR_LOW 0 717 #define QLC_DMA_CMD_BUFF_ADDR_HI 4 718 #define QLC_DMA_CMD_STATUS_CTRL 8 719 720 static int qlcnic_start_pex_dma(struct qlcnic_adapter *adapter, 721 struct __mem *mem) 722 { 723 struct device *dev = &adapter->pdev->dev; 724 u32 dma_no, dma_base_addr, temp_addr; 725 int i, ret, dma_sts; 726 void *tmpl_hdr; 727 728 tmpl_hdr = adapter->ahw->fw_dump.tmpl_hdr; 729 dma_no = qlcnic_get_saved_state(adapter, tmpl_hdr, 730 QLC_83XX_DMA_ENGINE_INDEX); 731 dma_base_addr = QLC_DMA_REG_BASE_ADDR(dma_no); 732 733 temp_addr = dma_base_addr + QLC_DMA_CMD_BUFF_ADDR_LOW; 734 ret = qlcnic_ind_wr(adapter, temp_addr, mem->desc_card_addr); 735 if (ret) 736 return ret; 737 738 temp_addr = dma_base_addr + QLC_DMA_CMD_BUFF_ADDR_HI; 739 ret = qlcnic_ind_wr(adapter, temp_addr, 0); 740 if (ret) 741 return ret; 742 743 temp_addr = dma_base_addr + QLC_DMA_CMD_STATUS_CTRL; 744 ret = qlcnic_ind_wr(adapter, temp_addr, mem->start_dma_cmd); 745 if (ret) 746 return ret; 747 748 /* Wait for DMA to complete */ 749 temp_addr = dma_base_addr + QLC_DMA_CMD_STATUS_CTRL; 750 for (i = 0; i < 400; i++) { 751 dma_sts = qlcnic_ind_rd(adapter, temp_addr); 752 753 if (dma_sts & BIT_1) 754 usleep_range(250, 500); 755 else 756 break; 757 } 758 759 if (i >= 400) { 760 dev_info(dev, "PEX DMA operation timed out"); 761 ret = -EIO; 762 } 763 764 return ret; 765 } 766 767 static u32 qlcnic_read_memory_pexdma(struct qlcnic_adapter *adapter, 768 struct __mem *mem, 769 __le32 *buffer, int *ret) 770 { 771 struct qlcnic_fw_dump *fw_dump = &adapter->ahw->fw_dump; 772 u32 temp, dma_base_addr, size = 0, read_size = 0; 773 struct qlcnic_pex_dma_descriptor *dma_descr; 774 struct device *dev = &adapter->pdev->dev; 775 dma_addr_t dma_phys_addr; 776 void *dma_buffer; 777 void *tmpl_hdr; 778 779 tmpl_hdr = fw_dump->tmpl_hdr; 780 781 /* Check if DMA engine is available */ 782 temp = qlcnic_get_saved_state(adapter, tmpl_hdr, 783 QLC_83XX_DMA_ENGINE_INDEX); 784 dma_base_addr = QLC_DMA_REG_BASE_ADDR(temp); 785 temp = qlcnic_ind_rd(adapter, 786 dma_base_addr + QLC_DMA_CMD_STATUS_CTRL); 787 788 if (!(temp & BIT_31)) { 789 dev_info(dev, "%s: DMA engine is not available\n", __func__); 790 *ret = -EIO; 791 return 0; 792 } 793 794 /* Create DMA descriptor */ 795 dma_descr = kzalloc(sizeof(struct qlcnic_pex_dma_descriptor), 796 GFP_KERNEL); 797 if (!dma_descr) { 798 *ret = -ENOMEM; 799 return 0; 800 } 801 802 /* dma_desc_cmd 0:15 = 0 803 * dma_desc_cmd 16:19 = mem->dma_desc_cmd 0:3 804 * dma_desc_cmd 20:23 = pci function number 805 * dma_desc_cmd 24:31 = mem->dma_desc_cmd 8:15 806 */ 807 dma_phys_addr = fw_dump->phys_addr; 808 dma_buffer = fw_dump->dma_buffer; 809 temp = 0; 810 temp = mem->dma_desc_cmd & 0xff0f; 811 temp |= (adapter->ahw->pci_func & 0xf) << 4; 812 dma_descr->dma_desc_cmd = (temp << 16) & 0xffff0000; 813 dma_descr->dma_bus_addr_low = LSD(dma_phys_addr); 814 dma_descr->dma_bus_addr_high = MSD(dma_phys_addr); 815 dma_descr->src_addr_high = 0; 816 817 /* Collect memory dump using multiple DMA operations if required */ 818 while (read_size < mem->size) { 819 if (mem->size - read_size >= QLC_PEX_DMA_READ_SIZE) 820 size = QLC_PEX_DMA_READ_SIZE; 821 else 822 size = mem->size - read_size; 823 824 dma_descr->src_addr_low = mem->addr + read_size; 825 dma_descr->read_data_size = size; 826 827 /* Write DMA descriptor to MS memory*/ 828 temp = sizeof(struct qlcnic_pex_dma_descriptor) / 16; 829 *ret = qlcnic_ms_mem_write128(adapter, mem->desc_card_addr, 830 (u32 *)dma_descr, temp); 831 if (*ret) { 832 dev_info(dev, "Failed to write DMA descriptor to MS memory at address 0x%x\n", 833 mem->desc_card_addr); 834 goto free_dma_descr; 835 } 836 837 *ret = qlcnic_start_pex_dma(adapter, mem); 838 if (*ret) { 839 dev_info(dev, "Failed to start PEX DMA operation\n"); 840 goto free_dma_descr; 841 } 842 843 memcpy(buffer, dma_buffer, size); 844 buffer += size / 4; 845 read_size += size; 846 } 847 848 free_dma_descr: 849 kfree(dma_descr); 850 851 return read_size; 852 } 853 854 static u32 qlcnic_read_memory(struct qlcnic_adapter *adapter, 855 struct qlcnic_dump_entry *entry, __le32 *buffer) 856 { 857 struct qlcnic_fw_dump *fw_dump = &adapter->ahw->fw_dump; 858 struct device *dev = &adapter->pdev->dev; 859 struct __mem *mem = &entry->region.mem; 860 u32 data_size; 861 int ret = 0; 862 863 if (fw_dump->use_pex_dma) { 864 data_size = qlcnic_read_memory_pexdma(adapter, mem, buffer, 865 &ret); 866 if (ret) 867 dev_info(dev, 868 "Failed to read memory dump using PEX DMA: mask[0x%x]\n", 869 entry->hdr.mask); 870 else 871 return data_size; 872 } 873 874 data_size = qlcnic_read_memory_test_agent(adapter, mem, buffer, &ret); 875 if (ret) { 876 dev_info(dev, 877 "Failed to read memory dump using test agent method: mask[0x%x]\n", 878 entry->hdr.mask); 879 return 0; 880 } else { 881 return data_size; 882 } 883 } 884 885 static u32 qlcnic_dump_nop(struct qlcnic_adapter *adapter, 886 struct qlcnic_dump_entry *entry, __le32 *buffer) 887 { 888 entry->hdr.flags |= QLCNIC_DUMP_SKIP; 889 return 0; 890 } 891 892 static int qlcnic_valid_dump_entry(struct device *dev, 893 struct qlcnic_dump_entry *entry, u32 size) 894 { 895 int ret = 1; 896 if (size != entry->hdr.cap_size) { 897 dev_err(dev, 898 "Invalid entry, Type:%d\tMask:%d\tSize:%dCap_size:%d\n", 899 entry->hdr.type, entry->hdr.mask, size, 900 entry->hdr.cap_size); 901 ret = 0; 902 } 903 return ret; 904 } 905 906 static u32 qlcnic_read_pollrdmwr(struct qlcnic_adapter *adapter, 907 struct qlcnic_dump_entry *entry, 908 __le32 *buffer) 909 { 910 struct __pollrdmwr *poll = &entry->region.pollrdmwr; 911 u32 data, wait_count, poll_wait, temp; 912 913 poll_wait = poll->poll_wait; 914 915 qlcnic_ind_wr(adapter, poll->addr1, poll->val1); 916 wait_count = 0; 917 918 while (wait_count < poll_wait) { 919 data = qlcnic_ind_rd(adapter, poll->addr1); 920 if ((data & poll->poll_mask) != 0) 921 break; 922 wait_count++; 923 } 924 925 if (wait_count == poll_wait) { 926 dev_err(&adapter->pdev->dev, 927 "Timeout exceeded in %s, aborting dump\n", 928 __func__); 929 return 0; 930 } 931 932 data = qlcnic_ind_rd(adapter, poll->addr2) & poll->mod_mask; 933 qlcnic_ind_wr(adapter, poll->addr2, data); 934 qlcnic_ind_wr(adapter, poll->addr1, poll->val2); 935 wait_count = 0; 936 937 while (wait_count < poll_wait) { 938 temp = qlcnic_ind_rd(adapter, poll->addr1); 939 if ((temp & poll->poll_mask) != 0) 940 break; 941 wait_count++; 942 } 943 944 *buffer++ = cpu_to_le32(poll->addr2); 945 *buffer++ = cpu_to_le32(data); 946 947 return 2 * sizeof(u32); 948 949 } 950 951 static u32 qlcnic_read_pollrd(struct qlcnic_adapter *adapter, 952 struct qlcnic_dump_entry *entry, __le32 *buffer) 953 { 954 struct __pollrd *pollrd = &entry->region.pollrd; 955 u32 data, wait_count, poll_wait, sel_val; 956 int i; 957 958 poll_wait = pollrd->poll_wait; 959 sel_val = pollrd->sel_val; 960 961 for (i = 0; i < pollrd->no_ops; i++) { 962 qlcnic_ind_wr(adapter, pollrd->sel_addr, sel_val); 963 wait_count = 0; 964 while (wait_count < poll_wait) { 965 data = qlcnic_ind_rd(adapter, pollrd->sel_addr); 966 if ((data & pollrd->poll_mask) != 0) 967 break; 968 wait_count++; 969 } 970 971 if (wait_count == poll_wait) { 972 dev_err(&adapter->pdev->dev, 973 "Timeout exceeded in %s, aborting dump\n", 974 __func__); 975 return 0; 976 } 977 978 data = qlcnic_ind_rd(adapter, pollrd->read_addr); 979 *buffer++ = cpu_to_le32(sel_val); 980 *buffer++ = cpu_to_le32(data); 981 sel_val += pollrd->sel_val_stride; 982 } 983 return pollrd->no_ops * (2 * sizeof(u32)); 984 } 985 986 static u32 qlcnic_read_mux2(struct qlcnic_adapter *adapter, 987 struct qlcnic_dump_entry *entry, __le32 *buffer) 988 { 989 struct __mux2 *mux2 = &entry->region.mux2; 990 u32 data; 991 u32 t_sel_val, sel_val1, sel_val2; 992 int i; 993 994 sel_val1 = mux2->sel_val1; 995 sel_val2 = mux2->sel_val2; 996 997 for (i = 0; i < mux2->no_ops; i++) { 998 qlcnic_ind_wr(adapter, mux2->sel_addr1, sel_val1); 999 t_sel_val = sel_val1 & mux2->sel_val_mask; 1000 qlcnic_ind_wr(adapter, mux2->sel_addr2, t_sel_val); 1001 data = qlcnic_ind_rd(adapter, mux2->read_addr); 1002 *buffer++ = cpu_to_le32(t_sel_val); 1003 *buffer++ = cpu_to_le32(data); 1004 qlcnic_ind_wr(adapter, mux2->sel_addr1, sel_val2); 1005 t_sel_val = sel_val2 & mux2->sel_val_mask; 1006 qlcnic_ind_wr(adapter, mux2->sel_addr2, t_sel_val); 1007 data = qlcnic_ind_rd(adapter, mux2->read_addr); 1008 *buffer++ = cpu_to_le32(t_sel_val); 1009 *buffer++ = cpu_to_le32(data); 1010 sel_val1 += mux2->sel_val_stride; 1011 sel_val2 += mux2->sel_val_stride; 1012 } 1013 1014 return mux2->no_ops * (4 * sizeof(u32)); 1015 } 1016 1017 static u32 qlcnic_83xx_dump_rom(struct qlcnic_adapter *adapter, 1018 struct qlcnic_dump_entry *entry, __le32 *buffer) 1019 { 1020 u32 fl_addr, size; 1021 struct __mem *rom = &entry->region.mem; 1022 1023 fl_addr = rom->addr; 1024 size = rom->size / 4; 1025 1026 if (!qlcnic_83xx_lockless_flash_read32(adapter, fl_addr, 1027 (u8 *)buffer, size)) 1028 return rom->size; 1029 1030 return 0; 1031 } 1032 1033 static const struct qlcnic_dump_operations qlcnic_fw_dump_ops[] = { 1034 {QLCNIC_DUMP_NOP, qlcnic_dump_nop}, 1035 {QLCNIC_DUMP_READ_CRB, qlcnic_dump_crb}, 1036 {QLCNIC_DUMP_READ_MUX, qlcnic_dump_mux}, 1037 {QLCNIC_DUMP_QUEUE, qlcnic_dump_que}, 1038 {QLCNIC_DUMP_BRD_CONFIG, qlcnic_read_rom}, 1039 {QLCNIC_DUMP_READ_OCM, qlcnic_dump_ocm}, 1040 {QLCNIC_DUMP_PEG_REG, qlcnic_dump_ctrl}, 1041 {QLCNIC_DUMP_L1_DTAG, qlcnic_dump_l1_cache}, 1042 {QLCNIC_DUMP_L1_ITAG, qlcnic_dump_l1_cache}, 1043 {QLCNIC_DUMP_L1_DATA, qlcnic_dump_l1_cache}, 1044 {QLCNIC_DUMP_L1_INST, qlcnic_dump_l1_cache}, 1045 {QLCNIC_DUMP_L2_DTAG, qlcnic_dump_l2_cache}, 1046 {QLCNIC_DUMP_L2_ITAG, qlcnic_dump_l2_cache}, 1047 {QLCNIC_DUMP_L2_DATA, qlcnic_dump_l2_cache}, 1048 {QLCNIC_DUMP_L2_INST, qlcnic_dump_l2_cache}, 1049 {QLCNIC_DUMP_READ_ROM, qlcnic_read_rom}, 1050 {QLCNIC_DUMP_READ_MEM, qlcnic_read_memory}, 1051 {QLCNIC_DUMP_READ_CTRL, qlcnic_dump_ctrl}, 1052 {QLCNIC_DUMP_TLHDR, qlcnic_dump_nop}, 1053 {QLCNIC_DUMP_RDEND, qlcnic_dump_nop}, 1054 }; 1055 1056 static const struct qlcnic_dump_operations qlcnic_83xx_fw_dump_ops[] = { 1057 {QLCNIC_DUMP_NOP, qlcnic_dump_nop}, 1058 {QLCNIC_DUMP_READ_CRB, qlcnic_dump_crb}, 1059 {QLCNIC_DUMP_READ_MUX, qlcnic_dump_mux}, 1060 {QLCNIC_DUMP_QUEUE, qlcnic_dump_que}, 1061 {QLCNIC_DUMP_BRD_CONFIG, qlcnic_83xx_dump_rom}, 1062 {QLCNIC_DUMP_READ_OCM, qlcnic_dump_ocm}, 1063 {QLCNIC_DUMP_PEG_REG, qlcnic_dump_ctrl}, 1064 {QLCNIC_DUMP_L1_DTAG, qlcnic_dump_l1_cache}, 1065 {QLCNIC_DUMP_L1_ITAG, qlcnic_dump_l1_cache}, 1066 {QLCNIC_DUMP_L1_DATA, qlcnic_dump_l1_cache}, 1067 {QLCNIC_DUMP_L1_INST, qlcnic_dump_l1_cache}, 1068 {QLCNIC_DUMP_L2_DTAG, qlcnic_dump_l2_cache}, 1069 {QLCNIC_DUMP_L2_ITAG, qlcnic_dump_l2_cache}, 1070 {QLCNIC_DUMP_L2_DATA, qlcnic_dump_l2_cache}, 1071 {QLCNIC_DUMP_L2_INST, qlcnic_dump_l2_cache}, 1072 {QLCNIC_DUMP_POLL_RD, qlcnic_read_pollrd}, 1073 {QLCNIC_READ_MUX2, qlcnic_read_mux2}, 1074 {QLCNIC_READ_POLLRDMWR, qlcnic_read_pollrdmwr}, 1075 {QLCNIC_DUMP_READ_ROM, qlcnic_83xx_dump_rom}, 1076 {QLCNIC_DUMP_READ_MEM, qlcnic_read_memory}, 1077 {QLCNIC_DUMP_READ_CTRL, qlcnic_dump_ctrl}, 1078 {QLCNIC_DUMP_TLHDR, qlcnic_dump_nop}, 1079 {QLCNIC_DUMP_RDEND, qlcnic_dump_nop}, 1080 }; 1081 1082 static uint32_t qlcnic_temp_checksum(uint32_t *temp_buffer, u32 temp_size) 1083 { 1084 uint64_t sum = 0; 1085 int count = temp_size / sizeof(uint32_t); 1086 while (count-- > 0) 1087 sum += *temp_buffer++; 1088 while (sum >> 32) 1089 sum = (sum & 0xFFFFFFFF) + (sum >> 32); 1090 return ~sum; 1091 } 1092 1093 static int qlcnic_fw_flash_get_minidump_temp(struct qlcnic_adapter *adapter, 1094 u8 *buffer, u32 size) 1095 { 1096 int ret = 0; 1097 1098 if (qlcnic_82xx_check(adapter)) 1099 return -EIO; 1100 1101 if (qlcnic_83xx_lock_flash(adapter)) 1102 return -EIO; 1103 1104 ret = qlcnic_83xx_lockless_flash_read32(adapter, 1105 QLC_83XX_MINIDUMP_FLASH, 1106 buffer, size / sizeof(u32)); 1107 1108 qlcnic_83xx_unlock_flash(adapter); 1109 1110 return ret; 1111 } 1112 1113 static int 1114 qlcnic_fw_flash_get_minidump_temp_size(struct qlcnic_adapter *adapter, 1115 struct qlcnic_cmd_args *cmd) 1116 { 1117 struct qlcnic_83xx_dump_template_hdr tmp_hdr; 1118 u32 size = sizeof(tmp_hdr) / sizeof(u32); 1119 int ret = 0; 1120 1121 if (qlcnic_82xx_check(adapter)) 1122 return -EIO; 1123 1124 if (qlcnic_83xx_lock_flash(adapter)) 1125 return -EIO; 1126 1127 ret = qlcnic_83xx_lockless_flash_read32(adapter, 1128 QLC_83XX_MINIDUMP_FLASH, 1129 (u8 *)&tmp_hdr, size); 1130 1131 qlcnic_83xx_unlock_flash(adapter); 1132 1133 cmd->rsp.arg[2] = tmp_hdr.size; 1134 cmd->rsp.arg[3] = tmp_hdr.version; 1135 1136 return ret; 1137 } 1138 1139 static int qlcnic_fw_get_minidump_temp_size(struct qlcnic_adapter *adapter, 1140 u32 *version, u32 *temp_size, 1141 u8 *use_flash_temp) 1142 { 1143 int err = 0; 1144 struct qlcnic_cmd_args cmd; 1145 1146 if (qlcnic_alloc_mbx_args(&cmd, adapter, QLCNIC_CMD_TEMP_SIZE)) 1147 return -ENOMEM; 1148 1149 err = qlcnic_issue_cmd(adapter, &cmd); 1150 if (err != QLCNIC_RCODE_SUCCESS) { 1151 if (qlcnic_fw_flash_get_minidump_temp_size(adapter, &cmd)) { 1152 qlcnic_free_mbx_args(&cmd); 1153 return -EIO; 1154 } 1155 *use_flash_temp = 1; 1156 } 1157 1158 *temp_size = cmd.rsp.arg[2]; 1159 *version = cmd.rsp.arg[3]; 1160 qlcnic_free_mbx_args(&cmd); 1161 1162 if (!(*temp_size)) 1163 return -EIO; 1164 1165 return 0; 1166 } 1167 1168 static int __qlcnic_fw_cmd_get_minidump_temp(struct qlcnic_adapter *adapter, 1169 u32 *buffer, u32 temp_size) 1170 { 1171 int err = 0, i; 1172 void *tmp_addr; 1173 __le32 *tmp_buf; 1174 struct qlcnic_cmd_args cmd; 1175 dma_addr_t tmp_addr_t = 0; 1176 1177 tmp_addr = dma_alloc_coherent(&adapter->pdev->dev, temp_size, 1178 &tmp_addr_t, GFP_KERNEL); 1179 if (!tmp_addr) 1180 return -ENOMEM; 1181 1182 if (qlcnic_alloc_mbx_args(&cmd, adapter, QLCNIC_CMD_GET_TEMP_HDR)) { 1183 err = -ENOMEM; 1184 goto free_mem; 1185 } 1186 1187 cmd.req.arg[1] = LSD(tmp_addr_t); 1188 cmd.req.arg[2] = MSD(tmp_addr_t); 1189 cmd.req.arg[3] = temp_size; 1190 err = qlcnic_issue_cmd(adapter, &cmd); 1191 1192 tmp_buf = tmp_addr; 1193 if (err == QLCNIC_RCODE_SUCCESS) { 1194 for (i = 0; i < temp_size / sizeof(u32); i++) 1195 *buffer++ = __le32_to_cpu(*tmp_buf++); 1196 } 1197 1198 qlcnic_free_mbx_args(&cmd); 1199 1200 free_mem: 1201 dma_free_coherent(&adapter->pdev->dev, temp_size, tmp_addr, tmp_addr_t); 1202 1203 return err; 1204 } 1205 1206 int qlcnic_fw_cmd_get_minidump_temp(struct qlcnic_adapter *adapter) 1207 { 1208 struct qlcnic_hardware_context *ahw; 1209 struct qlcnic_fw_dump *fw_dump; 1210 u32 version, csum, *tmp_buf; 1211 u8 use_flash_temp = 0; 1212 u32 temp_size = 0; 1213 void *temp_buffer; 1214 int err; 1215 1216 ahw = adapter->ahw; 1217 fw_dump = &ahw->fw_dump; 1218 err = qlcnic_fw_get_minidump_temp_size(adapter, &version, &temp_size, 1219 &use_flash_temp); 1220 if (err) { 1221 dev_err(&adapter->pdev->dev, 1222 "Can't get template size %d\n", err); 1223 return -EIO; 1224 } 1225 1226 fw_dump->tmpl_hdr = vzalloc(temp_size); 1227 if (!fw_dump->tmpl_hdr) 1228 return -ENOMEM; 1229 1230 tmp_buf = (u32 *)fw_dump->tmpl_hdr; 1231 if (use_flash_temp) 1232 goto flash_temp; 1233 1234 err = __qlcnic_fw_cmd_get_minidump_temp(adapter, tmp_buf, temp_size); 1235 1236 if (err) { 1237 flash_temp: 1238 err = qlcnic_fw_flash_get_minidump_temp(adapter, (u8 *)tmp_buf, 1239 temp_size); 1240 1241 if (err) { 1242 dev_err(&adapter->pdev->dev, 1243 "Failed to get minidump template header %d\n", 1244 err); 1245 vfree(fw_dump->tmpl_hdr); 1246 fw_dump->tmpl_hdr = NULL; 1247 return -EIO; 1248 } 1249 } 1250 1251 csum = qlcnic_temp_checksum((uint32_t *)tmp_buf, temp_size); 1252 1253 if (csum) { 1254 dev_err(&adapter->pdev->dev, 1255 "Template header checksum validation failed\n"); 1256 vfree(fw_dump->tmpl_hdr); 1257 fw_dump->tmpl_hdr = NULL; 1258 return -EIO; 1259 } 1260 1261 qlcnic_cache_tmpl_hdr_values(adapter, fw_dump); 1262 1263 if (fw_dump->use_pex_dma) { 1264 fw_dump->dma_buffer = NULL; 1265 temp_buffer = dma_alloc_coherent(&adapter->pdev->dev, 1266 QLC_PEX_DMA_READ_SIZE, 1267 &fw_dump->phys_addr, 1268 GFP_KERNEL); 1269 if (!temp_buffer) 1270 fw_dump->use_pex_dma = false; 1271 else 1272 fw_dump->dma_buffer = temp_buffer; 1273 } 1274 1275 1276 dev_info(&adapter->pdev->dev, 1277 "Default minidump capture mask 0x%x\n", 1278 fw_dump->cap_mask); 1279 1280 qlcnic_enable_fw_dump_state(adapter); 1281 1282 return 0; 1283 } 1284 1285 int qlcnic_dump_fw(struct qlcnic_adapter *adapter) 1286 { 1287 struct qlcnic_fw_dump *fw_dump = &adapter->ahw->fw_dump; 1288 static const struct qlcnic_dump_operations *fw_dump_ops; 1289 struct qlcnic_83xx_dump_template_hdr *hdr_83xx; 1290 u32 entry_offset, dump, no_entries, buf_offset = 0; 1291 int i, k, ops_cnt, ops_index, dump_size = 0; 1292 struct device *dev = &adapter->pdev->dev; 1293 struct qlcnic_hardware_context *ahw; 1294 struct qlcnic_dump_entry *entry; 1295 void *tmpl_hdr; 1296 u32 ocm_window; 1297 __le32 *buffer; 1298 char mesg[64]; 1299 char *msg[] = {mesg, NULL}; 1300 1301 ahw = adapter->ahw; 1302 tmpl_hdr = fw_dump->tmpl_hdr; 1303 1304 /* Return if we don't have firmware dump template header */ 1305 if (!tmpl_hdr) 1306 return -EIO; 1307 1308 if (!qlcnic_check_fw_dump_state(adapter)) { 1309 dev_info(&adapter->pdev->dev, "Dump not enabled\n"); 1310 return -EIO; 1311 } 1312 1313 if (fw_dump->clr) { 1314 dev_info(&adapter->pdev->dev, 1315 "Previous dump not cleared, not capturing dump\n"); 1316 return -EIO; 1317 } 1318 1319 netif_info(adapter->ahw, drv, adapter->netdev, "Take FW dump\n"); 1320 /* Calculate the size for dump data area only */ 1321 for (i = 2, k = 1; (i & QLCNIC_DUMP_MASK_MAX); i <<= 1, k++) 1322 if (i & fw_dump->cap_mask) 1323 dump_size += qlcnic_get_cap_size(adapter, tmpl_hdr, k); 1324 1325 if (!dump_size) 1326 return -EIO; 1327 1328 fw_dump->data = vzalloc(dump_size); 1329 if (!fw_dump->data) 1330 return -ENOMEM; 1331 1332 buffer = fw_dump->data; 1333 fw_dump->size = dump_size; 1334 no_entries = fw_dump->num_entries; 1335 entry_offset = fw_dump->offset; 1336 qlcnic_set_sys_info(adapter, tmpl_hdr, 0, QLCNIC_DRIVER_VERSION); 1337 qlcnic_set_sys_info(adapter, tmpl_hdr, 1, adapter->fw_version); 1338 1339 if (qlcnic_82xx_check(adapter)) { 1340 ops_cnt = ARRAY_SIZE(qlcnic_fw_dump_ops); 1341 fw_dump_ops = qlcnic_fw_dump_ops; 1342 } else { 1343 hdr_83xx = tmpl_hdr; 1344 ops_cnt = ARRAY_SIZE(qlcnic_83xx_fw_dump_ops); 1345 fw_dump_ops = qlcnic_83xx_fw_dump_ops; 1346 ocm_window = hdr_83xx->ocm_wnd_reg[ahw->pci_func]; 1347 hdr_83xx->saved_state[QLC_83XX_OCM_INDEX] = ocm_window; 1348 hdr_83xx->saved_state[QLC_83XX_PCI_INDEX] = ahw->pci_func; 1349 } 1350 1351 for (i = 0; i < no_entries; i++) { 1352 entry = tmpl_hdr + entry_offset; 1353 if (!(entry->hdr.mask & fw_dump->cap_mask)) { 1354 entry->hdr.flags |= QLCNIC_DUMP_SKIP; 1355 entry_offset += entry->hdr.offset; 1356 continue; 1357 } 1358 1359 /* Find the handler for this entry */ 1360 ops_index = 0; 1361 while (ops_index < ops_cnt) { 1362 if (entry->hdr.type == fw_dump_ops[ops_index].opcode) 1363 break; 1364 ops_index++; 1365 } 1366 1367 if (ops_index == ops_cnt) { 1368 dev_info(dev, "Skipping unknown entry opcode %d\n", 1369 entry->hdr.type); 1370 entry->hdr.flags |= QLCNIC_DUMP_SKIP; 1371 entry_offset += entry->hdr.offset; 1372 continue; 1373 } 1374 1375 /* Collect dump for this entry */ 1376 dump = fw_dump_ops[ops_index].handler(adapter, entry, buffer); 1377 if (!qlcnic_valid_dump_entry(dev, entry, dump)) { 1378 entry->hdr.flags |= QLCNIC_DUMP_SKIP; 1379 entry_offset += entry->hdr.offset; 1380 continue; 1381 } 1382 1383 buf_offset += entry->hdr.cap_size; 1384 entry_offset += entry->hdr.offset; 1385 buffer = fw_dump->data + buf_offset; 1386 } 1387 1388 fw_dump->clr = 1; 1389 snprintf(mesg, sizeof(mesg), "FW_DUMP=%s", adapter->netdev->name); 1390 netdev_info(adapter->netdev, 1391 "Dump data %d bytes captured, dump data address = %p, template header size %d bytes, template address = %p\n", 1392 fw_dump->size, fw_dump->data, fw_dump->tmpl_hdr_size, 1393 fw_dump->tmpl_hdr); 1394 /* Send a udev event to notify availability of FW dump */ 1395 kobject_uevent_env(&dev->kobj, KOBJ_CHANGE, msg); 1396 1397 return 0; 1398 } 1399 1400 static inline bool 1401 qlcnic_83xx_md_check_extended_dump_capability(struct qlcnic_adapter *adapter) 1402 { 1403 /* For special adapters (with 0x8830 device ID), where iSCSI firmware 1404 * dump needs to be captured as part of regular firmware dump 1405 * collection process, firmware exports it's capability through 1406 * capability registers 1407 */ 1408 return ((adapter->pdev->device == PCI_DEVICE_ID_QLOGIC_QLE8830) && 1409 (adapter->ahw->extra_capability[0] & 1410 QLCNIC_FW_CAPABILITY_2_EXT_ISCSI_DUMP)); 1411 } 1412 1413 void qlcnic_83xx_get_minidump_template(struct qlcnic_adapter *adapter) 1414 { 1415 u32 prev_version, current_version; 1416 struct qlcnic_hardware_context *ahw = adapter->ahw; 1417 struct qlcnic_fw_dump *fw_dump = &ahw->fw_dump; 1418 struct pci_dev *pdev = adapter->pdev; 1419 bool extended = false; 1420 1421 prev_version = adapter->fw_version; 1422 current_version = qlcnic_83xx_get_fw_version(adapter); 1423 1424 if (fw_dump->tmpl_hdr == NULL || current_version > prev_version) { 1425 vfree(fw_dump->tmpl_hdr); 1426 1427 if (qlcnic_83xx_md_check_extended_dump_capability(adapter)) 1428 extended = !qlcnic_83xx_extend_md_capab(adapter); 1429 1430 if (!qlcnic_fw_cmd_get_minidump_temp(adapter)) 1431 dev_info(&pdev->dev, "Supports FW dump capability\n"); 1432 1433 /* Once we have minidump template with extended iSCSI dump 1434 * capability, update the minidump capture mask to 0x1f as 1435 * per FW requirement 1436 */ 1437 if (extended) { 1438 struct qlcnic_83xx_dump_template_hdr *hdr; 1439 1440 hdr = fw_dump->tmpl_hdr; 1441 hdr->drv_cap_mask = 0x1f; 1442 fw_dump->cap_mask = 0x1f; 1443 dev_info(&pdev->dev, 1444 "Extended iSCSI dump capability and updated capture mask to 0x1f\n"); 1445 } 1446 } 1447 } 1448