1 /******************************************************************* 2 * This file is part of the Emulex Linux Device Driver for * 3 * Fibre Channel Host Bus Adapters. * 4 * Copyright (C) 2017-2023 Broadcom. All Rights Reserved. The term * 5 * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries. * 6 * Copyright (C) 2004-2016 Emulex. All rights reserved. * 7 * EMULEX and SLI are trademarks of Emulex. * 8 * www.broadcom.com * 9 * Portions Copyright (C) 2004-2005 Christoph Hellwig * 10 * * 11 * This program is free software; you can redistribute it and/or * 12 * modify it under the terms of version 2 of the GNU General * 13 * Public License as published by the Free Software Foundation. * 14 * This program is distributed in the hope that it will be useful. * 15 * ALL EXPRESS OR IMPLIED CONDITIONS, REPRESENTATIONS AND * 16 * WARRANTIES, INCLUDING ANY IMPLIED WARRANTY OF MERCHANTABILITY, * 17 * FITNESS FOR A PARTICULAR PURPOSE, OR NON-INFRINGEMENT, ARE * 18 * DISCLAIMED, EXCEPT TO THE EXTENT THAT SUCH DISCLAIMERS ARE HELD * 19 * TO BE LEGALLY INVALID. See the GNU General Public License for * 20 * more details, a copy of which can be found in the file COPYING * 21 * included with this package. * 22 *******************************************************************/ 23 24 #include <scsi/scsi_host.h> 25 #include <linux/hashtable.h> 26 #include <linux/ktime.h> 27 #include <linux/workqueue.h> 28 29 #if defined(CONFIG_DEBUG_FS) && !defined(CONFIG_SCSI_LPFC_DEBUG_FS) 30 #define CONFIG_SCSI_LPFC_DEBUG_FS 31 #endif 32 33 struct lpfc_sli2_slim; 34 35 #define ELX_MODEL_NAME_SIZE 80 36 37 #define LPFC_PCI_DEV_LP 0x1 38 #define LPFC_PCI_DEV_OC 0x2 39 40 #define LPFC_SLI_REV2 2 41 #define LPFC_SLI_REV3 3 42 #define LPFC_SLI_REV4 4 43 44 #define LPFC_MAX_TARGET 4096 /* max number of targets supported */ 45 #define LPFC_MAX_DISC_THREADS 64 /* max outstanding discovery els 46 requests */ 47 #define LPFC_MAX_NS_RETRY 3 /* Number of retry attempts to contact 48 the NameServer before giving up. */ 49 #define LPFC_CMD_PER_LUN 3 /* max outstanding cmds per lun */ 50 #define LPFC_DEFAULT_SG_SEG_CNT 64 /* sg element count per scsi cmnd */ 51 52 #define LPFC_DEFAULT_XPSGL_SIZE 256 53 #define LPFC_MAX_SG_TABLESIZE 0xffff 54 #define LPFC_MIN_SG_SLI4_BUF_SZ 0x800 /* based on LPFC_DEFAULT_SG_SEG_CNT */ 55 #define LPFC_MAX_BG_SLI4_SEG_CNT_DIF 128 /* sg element count for BlockGuard */ 56 #define LPFC_MAX_SG_SEG_CNT_DIF 512 /* sg element count per scsi cmnd */ 57 #define LPFC_MAX_SG_SEG_CNT 4096 /* sg element count per scsi cmnd */ 58 #define LPFC_MIN_SG_SEG_CNT 32 /* sg element count per scsi cmnd */ 59 #define LPFC_MAX_SGL_SEG_CNT 512 /* SGL element count per scsi cmnd */ 60 #define LPFC_MAX_BPL_SEG_CNT 4096 /* BPL element count per scsi cmnd */ 61 #define LPFC_MAX_NVME_SEG_CNT 256 /* max SGL element cnt per NVME cmnd */ 62 63 #define LPFC_MAX_SGE_SIZE 0x80000000 /* Maximum data allowed in a SGE */ 64 #define LPFC_IOCB_LIST_CNT 2250 /* list of IOCBs for fast-path usage. */ 65 #define LPFC_Q_RAMP_UP_INTERVAL 120 /* lun q_depth ramp up interval */ 66 #define LPFC_VNAME_LEN 100 /* vport symbolic name length */ 67 #define LPFC_TGTQ_RAMPUP_PCENT 5 /* Target queue rampup in percentage */ 68 #define LPFC_MIN_TGT_QDEPTH 10 69 #define LPFC_MAX_TGT_QDEPTH 0xFFFF 70 71 /* 72 * Following time intervals are used of adjusting SCSI device 73 * queue depths when there are driver resource error or Firmware 74 * resource error. 75 */ 76 /* 1 Second */ 77 #define QUEUE_RAMP_DOWN_INTERVAL (msecs_to_jiffies(1000 * 1)) 78 79 /* Number of exchanges reserved for discovery to complete */ 80 #define LPFC_DISC_IOCB_BUFF_COUNT 20 81 82 #define LPFC_HB_MBOX_INTERVAL 5 /* Heart beat interval in seconds. */ 83 #define LPFC_HB_MBOX_TIMEOUT 30 /* Heart beat timeout in seconds. */ 84 85 /* Error Attention event polling interval */ 86 #define LPFC_ERATT_POLL_INTERVAL 5 /* EATT poll interval in seconds */ 87 88 /* Define macros for 64 bit support */ 89 #define putPaddrLow(addr) ((uint32_t) (0xffffffff & (u64)(addr))) 90 #define putPaddrHigh(addr) ((uint32_t) (0xffffffff & (((u64)(addr))>>32))) 91 #define getPaddr(high, low) ((dma_addr_t)( \ 92 (( (u64)(high)<<16 ) << 16)|( (u64)(low)))) 93 /* Provide maximum configuration definitions. */ 94 #define LPFC_DRVR_TIMEOUT 16 /* driver iocb timeout value in sec */ 95 #define FC_MAX_ADPTMSG 64 96 97 #define MAX_HBAEVT 32 98 #define MAX_HBAS_NO_RESET 16 99 100 /* Number of MSI-X vectors the driver uses */ 101 #define LPFC_MSIX_VECTORS 2 102 103 /* lpfc wait event data ready flag */ 104 #define LPFC_DATA_READY 0 /* bit 0 */ 105 106 /* queue dump line buffer size */ 107 #define LPFC_LBUF_SZ 128 108 109 /* mailbox system shutdown options */ 110 #define LPFC_MBX_NO_WAIT 0 111 #define LPFC_MBX_WAIT 1 112 113 #define LPFC_CFG_PARAM_MAGIC_NUM 0xFEAA0005 114 #define LPFC_PORT_CFG_NAME "/cfg/port.cfg" 115 116 #define lpfc_rangecheck(val, min, max) \ 117 ((uint)(val) >= (uint)(min) && (val) <= (max)) 118 119 enum lpfc_polling_flags { 120 ENABLE_FCP_RING_POLLING = 0x1, 121 DISABLE_FCP_RING_INT = 0x2 122 }; 123 124 struct perf_prof { 125 uint16_t cmd_cpu[40]; 126 uint16_t rsp_cpu[40]; 127 uint16_t qh_cpu[40]; 128 uint16_t wqidx[40]; 129 }; 130 131 /* 132 * Provide for FC4 TYPE x28 - NVME. The 133 * bit mask for FCP and NVME is 0x8 identically 134 * because they are 32 bit positions distance. 135 */ 136 #define LPFC_FC4_TYPE_BITMASK 0x00000100 137 138 /* Provide DMA memory definitions the driver uses per port instance. */ 139 struct lpfc_dmabuf { 140 struct list_head list; 141 void *virt; /* virtual address ptr */ 142 dma_addr_t phys; /* mapped address */ 143 uint32_t buffer_tag; /* used for tagged queue ring */ 144 }; 145 146 struct lpfc_nvmet_ctxbuf { 147 struct list_head list; 148 struct lpfc_async_xchg_ctx *context; 149 struct lpfc_iocbq *iocbq; 150 struct lpfc_sglq *sglq; 151 struct work_struct defer_work; 152 }; 153 154 struct lpfc_dma_pool { 155 struct lpfc_dmabuf *elements; 156 uint32_t max_count; 157 uint32_t current_count; 158 }; 159 160 struct hbq_dmabuf { 161 struct lpfc_dmabuf hbuf; 162 struct lpfc_dmabuf dbuf; 163 uint16_t total_size; 164 uint16_t bytes_recv; 165 uint32_t tag; 166 struct lpfc_cq_event cq_event; 167 unsigned long time_stamp; 168 void *context; 169 }; 170 171 struct rqb_dmabuf { 172 struct lpfc_dmabuf hbuf; 173 struct lpfc_dmabuf dbuf; 174 uint16_t total_size; 175 uint16_t bytes_recv; 176 uint16_t idx; 177 struct lpfc_queue *hrq; /* ptr to associated Header RQ */ 178 struct lpfc_queue *drq; /* ptr to associated Data RQ */ 179 }; 180 181 /* Priority bit. Set value to exceed low water mark in lpfc_mem. */ 182 #define MEM_PRI 0x100 183 184 185 /****************************************************************************/ 186 /* Device VPD save area */ 187 /****************************************************************************/ 188 typedef struct lpfc_vpd { 189 uint32_t status; /* vpd status value */ 190 uint32_t length; /* number of bytes actually returned */ 191 struct { 192 uint32_t rsvd1; /* Revision numbers */ 193 uint32_t biuRev; 194 uint32_t smRev; 195 uint32_t smFwRev; 196 uint32_t endecRev; 197 uint16_t rBit; 198 uint8_t fcphHigh; 199 uint8_t fcphLow; 200 uint8_t feaLevelHigh; 201 uint8_t feaLevelLow; 202 uint32_t postKernRev; 203 uint32_t opFwRev; 204 uint8_t opFwName[16]; 205 uint32_t sli1FwRev; 206 uint8_t sli1FwName[16]; 207 uint32_t sli2FwRev; 208 uint8_t sli2FwName[16]; 209 } rev; 210 struct { 211 #ifdef __BIG_ENDIAN_BITFIELD 212 uint32_t rsvd3 :20; /* Reserved */ 213 uint32_t rsvd2 : 3; /* Reserved */ 214 uint32_t cbg : 1; /* Configure BlockGuard */ 215 uint32_t cmv : 1; /* Configure Max VPIs */ 216 uint32_t ccrp : 1; /* Config Command Ring Polling */ 217 uint32_t csah : 1; /* Configure Synchronous Abort Handling */ 218 uint32_t chbs : 1; /* Cofigure Host Backing store */ 219 uint32_t cinb : 1; /* Enable Interrupt Notification Block */ 220 uint32_t cerbm : 1; /* Configure Enhanced Receive Buf Mgmt */ 221 uint32_t cmx : 1; /* Configure Max XRIs */ 222 uint32_t cmr : 1; /* Configure Max RPIs */ 223 #else /* __LITTLE_ENDIAN */ 224 uint32_t cmr : 1; /* Configure Max RPIs */ 225 uint32_t cmx : 1; /* Configure Max XRIs */ 226 uint32_t cerbm : 1; /* Configure Enhanced Receive Buf Mgmt */ 227 uint32_t cinb : 1; /* Enable Interrupt Notification Block */ 228 uint32_t chbs : 1; /* Cofigure Host Backing store */ 229 uint32_t csah : 1; /* Configure Synchronous Abort Handling */ 230 uint32_t ccrp : 1; /* Config Command Ring Polling */ 231 uint32_t cmv : 1; /* Configure Max VPIs */ 232 uint32_t cbg : 1; /* Configure BlockGuard */ 233 uint32_t rsvd2 : 3; /* Reserved */ 234 uint32_t rsvd3 :20; /* Reserved */ 235 #endif 236 } sli3Feat; 237 } lpfc_vpd_t; 238 239 240 /* 241 * lpfc stat counters 242 */ 243 struct lpfc_stats { 244 /* Statistics for ELS commands */ 245 uint32_t elsLogiCol; 246 uint32_t elsRetryExceeded; 247 uint32_t elsXmitRetry; 248 uint32_t elsDelayRetry; 249 uint32_t elsRcvDrop; 250 uint32_t elsRcvFrame; 251 uint32_t elsRcvRSCN; 252 uint32_t elsRcvRNID; 253 uint32_t elsRcvFARP; 254 uint32_t elsRcvFARPR; 255 uint32_t elsRcvFLOGI; 256 uint32_t elsRcvPLOGI; 257 uint32_t elsRcvADISC; 258 uint32_t elsRcvPDISC; 259 uint32_t elsRcvFAN; 260 uint32_t elsRcvLOGO; 261 uint32_t elsRcvPRLO; 262 uint32_t elsRcvPRLI; 263 uint32_t elsRcvLIRR; 264 uint32_t elsRcvRLS; 265 uint32_t elsRcvRPL; 266 uint32_t elsRcvRRQ; 267 uint32_t elsRcvRTV; 268 uint32_t elsRcvECHO; 269 uint32_t elsRcvLCB; 270 uint32_t elsRcvRDP; 271 uint32_t elsRcvRDF; 272 uint32_t elsXmitFLOGI; 273 uint32_t elsXmitFDISC; 274 uint32_t elsXmitPLOGI; 275 uint32_t elsXmitPRLI; 276 uint32_t elsXmitADISC; 277 uint32_t elsXmitLOGO; 278 uint32_t elsXmitSCR; 279 uint32_t elsXmitRSCN; 280 uint32_t elsXmitRNID; 281 uint32_t elsXmitFARP; 282 uint32_t elsXmitFARPR; 283 uint32_t elsXmitACC; 284 uint32_t elsXmitLSRJT; 285 286 uint32_t frameRcvBcast; 287 uint32_t frameRcvMulti; 288 uint32_t strayXmitCmpl; 289 uint32_t frameXmitDelay; 290 uint32_t xriCmdCmpl; 291 uint32_t xriStatErr; 292 uint32_t LinkUp; 293 uint32_t LinkDown; 294 uint32_t LinkMultiEvent; 295 uint32_t NoRcvBuf; 296 uint32_t fcpCmd; 297 uint32_t fcpCmpl; 298 uint32_t fcpRspErr; 299 uint32_t fcpRemoteStop; 300 uint32_t fcpPortRjt; 301 uint32_t fcpPortBusy; 302 uint32_t fcpError; 303 uint32_t fcpLocalErr; 304 }; 305 306 struct lpfc_hba; 307 308 309 #define LPFC_VMID_TIMER 300 /* timer interval in seconds */ 310 311 #define LPFC_MAX_VMID_SIZE 256 312 #define LPFC_COMPRESS_VMID_SIZE 16 313 314 union lpfc_vmid_io_tag { 315 u32 app_id; /* App Id vmid */ 316 u8 cs_ctl_vmid; /* Priority tag vmid */ 317 }; 318 319 #define JIFFIES_PER_HR (HZ * 60 * 60) 320 321 struct lpfc_vmid { 322 u8 flag; 323 #define LPFC_VMID_SLOT_FREE 0x0 324 #define LPFC_VMID_SLOT_USED 0x1 325 #define LPFC_VMID_REQ_REGISTER 0x2 326 #define LPFC_VMID_REGISTERED 0x4 327 #define LPFC_VMID_DE_REGISTER 0x8 328 char host_vmid[LPFC_MAX_VMID_SIZE]; 329 union lpfc_vmid_io_tag un; 330 struct hlist_node hnode; 331 u64 io_rd_cnt; 332 u64 io_wr_cnt; 333 u8 vmid_len; 334 u8 delete_inactive; /* Delete if inactive flag 0 = no, 1 = yes */ 335 u32 hash_index; 336 u64 __percpu *last_io_time; 337 }; 338 339 #define lpfc_vmid_is_type_priority_tag(vport)\ 340 (vport->vmid_priority_tagging ? 1 : 0) 341 342 #define LPFC_VMID_HASH_SIZE 256 343 #define LPFC_VMID_HASH_MASK 255 344 #define LPFC_VMID_HASH_SHIFT 6 345 346 struct lpfc_vmid_context { 347 struct lpfc_vmid *vmp; 348 struct lpfc_nodelist *nlp; 349 bool instantiated; 350 }; 351 352 struct lpfc_vmid_priority_range { 353 u8 low; 354 u8 high; 355 u8 qos; 356 }; 357 358 struct lpfc_vmid_priority_info { 359 u32 num_descriptors; 360 struct lpfc_vmid_priority_range *vmid_range; 361 }; 362 363 #define QFPA_EVEN_ONLY 0x01 364 #define QFPA_ODD_ONLY 0x02 365 #define QFPA_EVEN_ODD 0x03 366 367 enum discovery_state { 368 LPFC_VPORT_UNKNOWN = 0, /* vport state is unknown */ 369 LPFC_VPORT_FAILED = 1, /* vport has failed */ 370 LPFC_LOCAL_CFG_LINK = 6, /* local NPORT Id configured */ 371 LPFC_FLOGI = 7, /* FLOGI sent to Fabric */ 372 LPFC_FDISC = 8, /* FDISC sent for vport */ 373 LPFC_FABRIC_CFG_LINK = 9, /* Fabric assigned NPORT Id 374 * configured */ 375 LPFC_NS_REG = 10, /* Register with NameServer */ 376 LPFC_NS_QRY = 11, /* Query NameServer for NPort ID list */ 377 LPFC_BUILD_DISC_LIST = 12, /* Build ADISC and PLOGI lists for 378 * device authentication / discovery */ 379 LPFC_DISC_AUTH = 13, /* Processing ADISC list */ 380 LPFC_VPORT_READY = 32, 381 }; 382 383 enum hba_state { 384 LPFC_LINK_UNKNOWN = 0, /* HBA state is unknown */ 385 LPFC_WARM_START = 1, /* HBA state after selective reset */ 386 LPFC_INIT_START = 2, /* Initial state after board reset */ 387 LPFC_INIT_MBX_CMDS = 3, /* Initialize HBA with mbox commands */ 388 LPFC_LINK_DOWN = 4, /* HBA initialized, link is down */ 389 LPFC_LINK_UP = 5, /* Link is up - issue READ_LA */ 390 LPFC_CLEAR_LA = 6, /* authentication cmplt - issue 391 * CLEAR_LA */ 392 LPFC_HBA_READY = 32, 393 LPFC_HBA_ERROR = -1 394 }; 395 396 struct lpfc_trunk_link_state { 397 enum hba_state state; 398 uint8_t fault; 399 }; 400 401 struct lpfc_trunk_link { 402 struct lpfc_trunk_link_state link0, 403 link1, 404 link2, 405 link3; 406 u32 phy_lnk_speed; 407 }; 408 409 /* Format of congestion module parameters */ 410 struct lpfc_cgn_param { 411 uint32_t cgn_param_magic; 412 uint8_t cgn_param_version; /* version 1 */ 413 uint8_t cgn_param_mode; /* 0=off 1=managed 2=monitor only */ 414 #define LPFC_CFG_OFF 0 415 #define LPFC_CFG_MANAGED 1 416 #define LPFC_CFG_MONITOR 2 417 uint8_t cgn_rsvd1; 418 uint8_t cgn_rsvd2; 419 uint8_t cgn_param_level0; 420 uint8_t cgn_param_level1; 421 uint8_t cgn_param_level2; 422 uint8_t byte11; 423 uint8_t byte12; 424 uint8_t byte13; 425 uint8_t byte14; 426 uint8_t byte15; 427 }; 428 429 /* Max number of days of congestion data */ 430 #define LPFC_MAX_CGN_DAYS 10 431 432 /* Format of congestion buffer info 433 * This structure defines memory thats allocated and registered with 434 * the HBA firmware. When adding or removing fields from this structure 435 * the alignment must match the HBA firmware. 436 */ 437 438 struct lpfc_cgn_info { 439 /* Header */ 440 __le16 cgn_info_size; /* is sizeof(struct lpfc_cgn_info) */ 441 uint8_t cgn_info_version; /* represents format of structure */ 442 #define LPFC_CGN_INFO_V1 1 443 #define LPFC_CGN_INFO_V2 2 444 #define LPFC_CGN_INFO_V3 3 445 uint8_t cgn_info_mode; /* 0=off 1=managed 2=monitor only */ 446 uint8_t cgn_info_detect; 447 uint8_t cgn_info_action; 448 uint8_t cgn_info_level0; 449 uint8_t cgn_info_level1; 450 uint8_t cgn_info_level2; 451 452 /* Start Time */ 453 uint8_t cgn_info_month; 454 uint8_t cgn_info_day; 455 uint8_t cgn_info_year; 456 uint8_t cgn_info_hour; 457 uint8_t cgn_info_minute; 458 uint8_t cgn_info_second; 459 460 /* minute / hours / daily indices */ 461 uint8_t cgn_index_minute; 462 uint8_t cgn_index_hour; 463 uint8_t cgn_index_day; 464 465 __le16 cgn_warn_freq; 466 __le16 cgn_alarm_freq; 467 __le16 cgn_lunq; 468 uint8_t cgn_pad1[8]; 469 470 /* Driver Information */ 471 __le16 cgn_drvr_min[60]; 472 __le32 cgn_drvr_hr[24]; 473 __le32 cgn_drvr_day[LPFC_MAX_CGN_DAYS]; 474 475 /* Congestion Warnings */ 476 __le16 cgn_warn_min[60]; 477 __le32 cgn_warn_hr[24]; 478 __le32 cgn_warn_day[LPFC_MAX_CGN_DAYS]; 479 480 /* Latency Information */ 481 __le32 cgn_latency_min[60]; 482 __le32 cgn_latency_hr[24]; 483 __le32 cgn_latency_day[LPFC_MAX_CGN_DAYS]; 484 485 /* Bandwidth Information */ 486 __le16 cgn_bw_min[60]; 487 __le16 cgn_bw_hr[24]; 488 __le16 cgn_bw_day[LPFC_MAX_CGN_DAYS]; 489 490 /* Congestion Alarms */ 491 __le16 cgn_alarm_min[60]; 492 __le32 cgn_alarm_hr[24]; 493 __le32 cgn_alarm_day[LPFC_MAX_CGN_DAYS]; 494 495 struct_group(cgn_stat, 496 uint8_t cgn_stat_npm; /* Notifications per minute */ 497 498 /* Start Time */ 499 uint8_t cgn_stat_month; 500 uint8_t cgn_stat_day; 501 uint8_t cgn_stat_year; 502 uint8_t cgn_stat_hour; 503 uint8_t cgn_stat_minute; 504 uint8_t cgn_pad2[2]; 505 506 __le32 cgn_notification; 507 __le32 cgn_peer_notification; 508 __le32 link_integ_notification; 509 __le32 delivery_notification; 510 511 uint8_t cgn_stat_cgn_month; /* Last congestion notification FPIN */ 512 uint8_t cgn_stat_cgn_day; 513 uint8_t cgn_stat_cgn_year; 514 uint8_t cgn_stat_cgn_hour; 515 uint8_t cgn_stat_cgn_min; 516 uint8_t cgn_stat_cgn_sec; 517 518 uint8_t cgn_stat_peer_month; /* Last peer congestion FPIN */ 519 uint8_t cgn_stat_peer_day; 520 uint8_t cgn_stat_peer_year; 521 uint8_t cgn_stat_peer_hour; 522 uint8_t cgn_stat_peer_min; 523 uint8_t cgn_stat_peer_sec; 524 525 uint8_t cgn_stat_lnk_month; /* Last link integrity FPIN */ 526 uint8_t cgn_stat_lnk_day; 527 uint8_t cgn_stat_lnk_year; 528 uint8_t cgn_stat_lnk_hour; 529 uint8_t cgn_stat_lnk_min; 530 uint8_t cgn_stat_lnk_sec; 531 532 uint8_t cgn_stat_del_month; /* Last delivery notification FPIN */ 533 uint8_t cgn_stat_del_day; 534 uint8_t cgn_stat_del_year; 535 uint8_t cgn_stat_del_hour; 536 uint8_t cgn_stat_del_min; 537 uint8_t cgn_stat_del_sec; 538 ); 539 540 __le32 cgn_info_crc; 541 #define LPFC_CGN_CRC32_MAGIC_NUMBER 0x1EDC6F41 542 #define LPFC_CGN_CRC32_SEED 0xFFFFFFFF 543 }; 544 545 #define LPFC_CGN_INFO_SZ (sizeof(struct lpfc_cgn_info) - \ 546 sizeof(uint32_t)) 547 548 struct lpfc_cgn_stat { 549 atomic64_t total_bytes; 550 atomic64_t rcv_bytes; 551 atomic64_t rx_latency; 552 #define LPFC_CGN_NOT_SENT 0xFFFFFFFFFFFFFFFFLL 553 atomic_t rx_io_cnt; 554 }; 555 556 struct lpfc_cgn_acqe_stat { 557 atomic64_t alarm; 558 atomic64_t warn; 559 }; 560 561 struct lpfc_vport { 562 struct lpfc_hba *phba; 563 struct list_head listentry; 564 uint8_t port_type; 565 #define LPFC_PHYSICAL_PORT 1 566 #define LPFC_NPIV_PORT 2 567 #define LPFC_FABRIC_PORT 3 568 enum discovery_state port_state; 569 570 uint16_t vpi; 571 uint16_t vfi; 572 uint8_t vpi_state; 573 #define LPFC_VPI_REGISTERED 0x1 574 575 uint32_t fc_flag; /* FC flags */ 576 /* Several of these flags are HBA centric and should be moved to 577 * phba->link_flag (e.g. FC_PTP, FC_PUBLIC_LOOP) 578 */ 579 #define FC_PT2PT 0x1 /* pt2pt with no fabric */ 580 #define FC_PT2PT_PLOGI 0x2 /* pt2pt initiate PLOGI */ 581 #define FC_DISC_TMO 0x4 /* Discovery timer running */ 582 #define FC_PUBLIC_LOOP 0x8 /* Public loop */ 583 #define FC_LBIT 0x10 /* LOGIN bit in loopinit set */ 584 #define FC_RSCN_MODE 0x20 /* RSCN cmd rcv'ed */ 585 #define FC_NLP_MORE 0x40 /* More node to process in node tbl */ 586 #define FC_OFFLINE_MODE 0x80 /* Interface is offline for diag */ 587 #define FC_FABRIC 0x100 /* We are fabric attached */ 588 #define FC_VPORT_LOGO_RCVD 0x200 /* LOGO received on vport */ 589 #define FC_RSCN_DISCOVERY 0x400 /* Auth all devices after RSCN */ 590 #define FC_LOGO_RCVD_DID_CHNG 0x800 /* FDISC on phys port detect DID chng*/ 591 #define FC_PT2PT_NO_NVME 0x1000 /* Don't send NVME PRLI */ 592 #define FC_SCSI_SCAN_TMO 0x4000 /* scsi scan timer running */ 593 #define FC_ABORT_DISCOVERY 0x8000 /* we want to abort discovery */ 594 #define FC_NDISC_ACTIVE 0x10000 /* NPort discovery active */ 595 #define FC_BYPASSED_MODE 0x20000 /* NPort is in bypassed mode */ 596 #define FC_VPORT_NEEDS_REG_VPI 0x80000 /* Needs to have its vpi registered */ 597 #define FC_RSCN_DEFERRED 0x100000 /* A deferred RSCN being processed */ 598 #define FC_VPORT_NEEDS_INIT_VPI 0x200000 /* Need to INIT_VPI before FDISC */ 599 #define FC_VPORT_CVL_RCVD 0x400000 /* VLink failed due to CVL */ 600 #define FC_VFI_REGISTERED 0x800000 /* VFI is registered */ 601 #define FC_FDISC_COMPLETED 0x1000000/* FDISC completed */ 602 #define FC_DISC_DELAYED 0x2000000/* Delay NPort discovery */ 603 604 uint32_t ct_flags; 605 #define FC_CT_RFF_ID 0x1 /* RFF_ID accepted by switch */ 606 #define FC_CT_RNN_ID 0x2 /* RNN_ID accepted by switch */ 607 #define FC_CT_RSNN_NN 0x4 /* RSNN_NN accepted by switch */ 608 #define FC_CT_RSPN_ID 0x8 /* RSPN_ID accepted by switch */ 609 #define FC_CT_RFT_ID 0x10 /* RFT_ID accepted by switch */ 610 #define FC_CT_RPRT_DEFER 0x20 /* Defer issuing FDMI RPRT */ 611 612 struct list_head fc_nodes; 613 614 /* Keep counters for the number of entries in each list. */ 615 uint16_t fc_plogi_cnt; 616 uint16_t fc_adisc_cnt; 617 uint16_t fc_reglogin_cnt; 618 uint16_t fc_prli_cnt; 619 uint16_t fc_unmap_cnt; 620 uint16_t fc_map_cnt; 621 uint16_t fc_npr_cnt; 622 uint16_t fc_unused_cnt; 623 struct serv_parm fc_sparam; /* buffer for our service parameters */ 624 625 uint32_t fc_myDID; /* fibre channel S_ID */ 626 uint32_t fc_prevDID; /* previous fibre channel S_ID */ 627 struct lpfc_name fabric_portname; 628 struct lpfc_name fabric_nodename; 629 630 int32_t stopped; /* HBA has not been restarted since last ERATT */ 631 uint8_t fc_linkspeed; /* Link speed after last READ_LA */ 632 633 uint32_t num_disc_nodes; /* in addition to hba_state */ 634 uint32_t gidft_inp; /* cnt of outstanding GID_FTs */ 635 636 uint32_t fc_nlp_cnt; /* outstanding NODELIST requests */ 637 uint32_t fc_rscn_id_cnt; /* count of RSCNs payloads in list */ 638 uint32_t fc_rscn_flush; /* flag use of fc_rscn_id_list */ 639 struct lpfc_dmabuf *fc_rscn_id_list[FC_MAX_HOLD_RSCN]; 640 struct lpfc_name fc_nodename; /* fc nodename */ 641 struct lpfc_name fc_portname; /* fc portname */ 642 643 struct lpfc_work_evt disc_timeout_evt; 644 645 struct timer_list fc_disctmo; /* Discovery rescue timer */ 646 uint8_t fc_ns_retry; /* retries for fabric nameserver */ 647 uint32_t fc_prli_sent; /* cntr for outstanding PRLIs */ 648 649 spinlock_t work_port_lock; 650 uint32_t work_port_events; /* Timeout to be handled */ 651 #define WORKER_DISC_TMO 0x1 /* vport: Discovery timeout */ 652 #define WORKER_ELS_TMO 0x2 /* vport: ELS timeout */ 653 #define WORKER_DELAYED_DISC_TMO 0x8 /* vport: delayed discovery */ 654 655 #define WORKER_MBOX_TMO 0x100 /* hba: MBOX timeout */ 656 #define WORKER_HB_TMO 0x200 /* hba: Heart beat timeout */ 657 #define WORKER_FABRIC_BLOCK_TMO 0x400 /* hba: fabric block timeout */ 658 #define WORKER_RAMP_DOWN_QUEUE 0x800 /* hba: Decrease Q depth */ 659 #define WORKER_RAMP_UP_QUEUE 0x1000 /* hba: Increase Q depth */ 660 #define WORKER_SERVICE_TXQ 0x2000 /* hba: IOCBs on the txq */ 661 #define WORKER_CHECK_INACTIVE_VMID 0x4000 /* hba: check inactive vmids */ 662 #define WORKER_CHECK_VMID_ISSUE_QFPA 0x8000 /* vport: Check if qfpa needs 663 * to be issued */ 664 665 struct timer_list els_tmofunc; 666 struct timer_list delayed_disc_tmo; 667 668 uint8_t load_flag; 669 #define FC_LOADING 0x1 /* HBA in process of loading drvr */ 670 #define FC_UNLOADING 0x2 /* HBA in process of unloading drvr */ 671 #define FC_ALLOW_FDMI 0x4 /* port is ready for FDMI requests */ 672 #define FC_ALLOW_VMID 0x8 /* Allow VMID I/Os */ 673 #define FC_DEREGISTER_ALL_APP_ID 0x10 /* Deregister all VMIDs */ 674 /* Vport Config Parameters */ 675 uint32_t cfg_scan_down; 676 uint32_t cfg_lun_queue_depth; 677 uint32_t cfg_nodev_tmo; 678 uint32_t cfg_devloss_tmo; 679 uint32_t cfg_restrict_login; 680 uint32_t cfg_peer_port_login; 681 uint32_t cfg_fcp_class; 682 uint32_t cfg_use_adisc; 683 uint32_t cfg_discovery_threads; 684 uint32_t cfg_log_verbose; 685 uint32_t cfg_enable_fc4_type; 686 uint32_t cfg_max_luns; 687 uint32_t cfg_enable_da_id; 688 uint32_t cfg_max_scsicmpl_time; 689 uint32_t cfg_tgt_queue_depth; 690 uint32_t cfg_first_burst_size; 691 uint32_t dev_loss_tmo_changed; 692 /* VMID parameters */ 693 u8 lpfc_vmid_host_uuid[LPFC_COMPRESS_VMID_SIZE]; 694 u32 max_vmid; /* maximum VMIDs allowed per port */ 695 u32 cur_vmid_cnt; /* Current VMID count */ 696 #define LPFC_MIN_VMID 4 697 #define LPFC_MAX_VMID 255 698 u32 vmid_inactivity_timeout; /* Time after which the VMID */ 699 /* deregisters from switch */ 700 u32 vmid_priority_tagging; 701 #define LPFC_VMID_PRIO_TAG_DISABLE 0 /* Disable */ 702 #define LPFC_VMID_PRIO_TAG_SUP_TARGETS 1 /* Allow supported targets only */ 703 #define LPFC_VMID_PRIO_TAG_ALL_TARGETS 2 /* Allow all targets */ 704 unsigned long *vmid_priority_range; 705 #define LPFC_VMID_MAX_PRIORITY_RANGE 256 706 #define LPFC_VMID_PRIORITY_BITMAP_SIZE 32 707 u8 vmid_flag; 708 #define LPFC_VMID_IN_USE 0x1 709 #define LPFC_VMID_ISSUE_QFPA 0x2 710 #define LPFC_VMID_QFPA_CMPL 0x4 711 #define LPFC_VMID_QOS_ENABLED 0x8 712 #define LPFC_VMID_TIMER_ENBLD 0x10 713 #define LPFC_VMID_TYPE_PRIO 0x20 714 struct fc_qfpa_res *qfpa_res; 715 716 struct fc_vport *fc_vport; 717 718 struct lpfc_vmid *vmid; 719 DECLARE_HASHTABLE(hash_table, 8); 720 rwlock_t vmid_lock; 721 struct lpfc_vmid_priority_info vmid_priority; 722 723 #ifdef CONFIG_SCSI_LPFC_DEBUG_FS 724 struct dentry *debug_disc_trc; 725 struct dentry *debug_nodelist; 726 struct dentry *debug_nvmestat; 727 struct dentry *debug_scsistat; 728 struct dentry *debug_ioktime; 729 struct dentry *debug_hdwqstat; 730 struct dentry *vport_debugfs_root; 731 struct lpfc_debugfs_trc *disc_trc; 732 atomic_t disc_trc_cnt; 733 #endif 734 struct list_head rcv_buffer_list; 735 unsigned long rcv_buffer_time_stamp; 736 uint32_t vport_flag; 737 #define STATIC_VPORT 0x1 738 #define FAWWPN_PARAM_CHG 0x2 739 740 uint16_t fdmi_num_disc; 741 uint32_t fdmi_hba_mask; 742 uint32_t fdmi_port_mask; 743 744 /* There is a single nvme instance per vport. */ 745 struct nvme_fc_local_port *localport; 746 uint8_t nvmei_support; /* driver supports NVME Initiator */ 747 uint32_t last_fcp_wqidx; 748 uint32_t rcv_flogi_cnt; /* How many unsol FLOGIs ACK'd. */ 749 }; 750 751 struct hbq_s { 752 uint16_t entry_count; /* Current number of HBQ slots */ 753 uint16_t buffer_count; /* Current number of buffers posted */ 754 uint32_t next_hbqPutIdx; /* Index to next HBQ slot to use */ 755 uint32_t hbqPutIdx; /* HBQ slot to use */ 756 uint32_t local_hbqGetIdx; /* Local copy of Get index from Port */ 757 void *hbq_virt; /* Virtual ptr to this hbq */ 758 struct list_head hbq_buffer_list; /* buffers assigned to this HBQ */ 759 /* Callback for HBQ buffer allocation */ 760 struct hbq_dmabuf *(*hbq_alloc_buffer) (struct lpfc_hba *); 761 /* Callback for HBQ buffer free */ 762 void (*hbq_free_buffer) (struct lpfc_hba *, 763 struct hbq_dmabuf *); 764 }; 765 766 /* this matches the position in the lpfc_hbq_defs array */ 767 #define LPFC_ELS_HBQ 0 768 #define LPFC_MAX_HBQS 1 769 770 enum hba_temp_state { 771 HBA_NORMAL_TEMP, 772 HBA_OVER_TEMP 773 }; 774 775 enum intr_type_t { 776 NONE = 0, 777 INTx, 778 MSI, 779 MSIX, 780 }; 781 782 #define LPFC_CT_CTX_MAX 64 783 struct unsol_rcv_ct_ctx { 784 uint32_t ctxt_id; 785 uint32_t SID; 786 uint32_t valid; 787 #define UNSOL_INVALID 0 788 #define UNSOL_VALID 1 789 uint16_t oxid; 790 uint16_t rxid; 791 }; 792 793 #define LPFC_USER_LINK_SPEED_AUTO 0 /* auto select (default)*/ 794 #define LPFC_USER_LINK_SPEED_1G 1 /* 1 Gigabaud */ 795 #define LPFC_USER_LINK_SPEED_2G 2 /* 2 Gigabaud */ 796 #define LPFC_USER_LINK_SPEED_4G 4 /* 4 Gigabaud */ 797 #define LPFC_USER_LINK_SPEED_8G 8 /* 8 Gigabaud */ 798 #define LPFC_USER_LINK_SPEED_10G 10 /* 10 Gigabaud */ 799 #define LPFC_USER_LINK_SPEED_16G 16 /* 16 Gigabaud */ 800 #define LPFC_USER_LINK_SPEED_32G 32 /* 32 Gigabaud */ 801 #define LPFC_USER_LINK_SPEED_64G 64 /* 64 Gigabaud */ 802 #define LPFC_USER_LINK_SPEED_MAX LPFC_USER_LINK_SPEED_64G 803 804 #define LPFC_LINK_SPEED_STRING "0, 1, 2, 4, 8, 10, 16, 32, 64" 805 806 enum nemb_type { 807 nemb_mse = 1, 808 nemb_hbd 809 }; 810 811 enum mbox_type { 812 mbox_rd = 1, 813 mbox_wr 814 }; 815 816 enum dma_type { 817 dma_mbox = 1, 818 dma_ebuf 819 }; 820 821 enum sta_type { 822 sta_pre_addr = 1, 823 sta_pos_addr 824 }; 825 826 struct lpfc_mbox_ext_buf_ctx { 827 uint32_t state; 828 #define LPFC_BSG_MBOX_IDLE 0 829 #define LPFC_BSG_MBOX_HOST 1 830 #define LPFC_BSG_MBOX_PORT 2 831 #define LPFC_BSG_MBOX_DONE 3 832 #define LPFC_BSG_MBOX_ABTS 4 833 enum nemb_type nembType; 834 enum mbox_type mboxType; 835 uint32_t numBuf; 836 uint32_t mbxTag; 837 uint32_t seqNum; 838 struct lpfc_dmabuf *mbx_dmabuf; 839 struct list_head ext_dmabuf_list; 840 }; 841 842 struct lpfc_epd_pool { 843 /* Expedite pool */ 844 struct list_head list; 845 u32 count; 846 spinlock_t lock; /* lock for expedite pool */ 847 }; 848 849 enum ras_state { 850 INACTIVE, 851 REG_INPROGRESS, 852 ACTIVE 853 }; 854 855 struct lpfc_ras_fwlog { 856 uint8_t *fwlog_buff; 857 uint32_t fw_buffcount; /* Buffer size posted to FW */ 858 #define LPFC_RAS_BUFF_ENTERIES 16 /* Each entry can hold max of 64k */ 859 #define LPFC_RAS_MAX_ENTRY_SIZE (64 * 1024) 860 #define LPFC_RAS_MIN_BUFF_POST_SIZE (256 * 1024) 861 #define LPFC_RAS_MAX_BUFF_POST_SIZE (1024 * 1024) 862 uint32_t fw_loglevel; /* Log level set */ 863 struct lpfc_dmabuf lwpd; 864 struct list_head fwlog_buff_list; 865 866 /* RAS support status on adapter */ 867 bool ras_hwsupport; /* RAS Support available on HW or not */ 868 bool ras_enabled; /* Ras Enabled for the function */ 869 #define LPFC_RAS_DISABLE_LOGGING 0x00 870 #define LPFC_RAS_ENABLE_LOGGING 0x01 871 enum ras_state state; /* RAS logging running state */ 872 }; 873 874 #define DBG_LOG_STR_SZ 256 875 #define DBG_LOG_SZ 256 876 877 struct dbg_log_ent { 878 char log[DBG_LOG_STR_SZ]; 879 u64 t_ns; 880 }; 881 882 enum lpfc_irq_chann_mode { 883 /* Assign IRQs to all possible cpus that have hardware queues */ 884 NORMAL_MODE, 885 886 /* Assign IRQs only to cpus on the same numa node as HBA */ 887 NUMA_MODE, 888 889 /* Assign IRQs only on non-hyperthreaded CPUs. This is the 890 * same as normal_mode, but assign IRQS only on physical CPUs. 891 */ 892 NHT_MODE, 893 }; 894 895 enum lpfc_hba_bit_flags { 896 FABRIC_COMANDS_BLOCKED, 897 HBA_PCI_ERR, 898 }; 899 900 struct lpfc_hba { 901 /* SCSI interface function jump table entries */ 902 struct lpfc_io_buf * (*lpfc_get_scsi_buf) 903 (struct lpfc_hba *phba, struct lpfc_nodelist *ndlp, 904 struct scsi_cmnd *cmnd); 905 int (*lpfc_scsi_prep_dma_buf) 906 (struct lpfc_hba *, struct lpfc_io_buf *); 907 void (*lpfc_scsi_unprep_dma_buf) 908 (struct lpfc_hba *, struct lpfc_io_buf *); 909 void (*lpfc_release_scsi_buf) 910 (struct lpfc_hba *, struct lpfc_io_buf *); 911 void (*lpfc_rampdown_queue_depth) 912 (struct lpfc_hba *); 913 void (*lpfc_scsi_prep_cmnd) 914 (struct lpfc_vport *, struct lpfc_io_buf *, 915 struct lpfc_nodelist *); 916 int (*lpfc_scsi_prep_cmnd_buf) 917 (struct lpfc_vport *vport, 918 struct lpfc_io_buf *lpfc_cmd, 919 uint8_t tmo); 920 int (*lpfc_scsi_prep_task_mgmt_cmd) 921 (struct lpfc_vport *vport, 922 struct lpfc_io_buf *lpfc_cmd, 923 u64 lun, u8 task_mgmt_cmd); 924 925 /* IOCB interface function jump table entries */ 926 int (*__lpfc_sli_issue_iocb) 927 (struct lpfc_hba *, uint32_t, 928 struct lpfc_iocbq *, uint32_t); 929 int (*__lpfc_sli_issue_fcp_io) 930 (struct lpfc_hba *phba, uint32_t ring_number, 931 struct lpfc_iocbq *piocb, uint32_t flag); 932 void (*__lpfc_sli_release_iocbq)(struct lpfc_hba *, 933 struct lpfc_iocbq *); 934 int (*lpfc_hba_down_post)(struct lpfc_hba *phba); 935 void (*lpfc_scsi_cmd_iocb_cmpl) 936 (struct lpfc_hba *, struct lpfc_iocbq *, struct lpfc_iocbq *); 937 938 /* MBOX interface function jump table entries */ 939 int (*lpfc_sli_issue_mbox) 940 (struct lpfc_hba *, LPFC_MBOXQ_t *, uint32_t); 941 942 /* Slow-path IOCB process function jump table entries */ 943 void (*lpfc_sli_handle_slow_ring_event) 944 (struct lpfc_hba *phba, struct lpfc_sli_ring *pring, 945 uint32_t mask); 946 947 /* INIT device interface function jump table entries */ 948 int (*lpfc_sli_hbq_to_firmware) 949 (struct lpfc_hba *, uint32_t, struct hbq_dmabuf *); 950 int (*lpfc_sli_brdrestart) 951 (struct lpfc_hba *); 952 int (*lpfc_sli_brdready) 953 (struct lpfc_hba *, uint32_t); 954 void (*lpfc_handle_eratt) 955 (struct lpfc_hba *); 956 void (*lpfc_stop_port) 957 (struct lpfc_hba *); 958 int (*lpfc_hba_init_link) 959 (struct lpfc_hba *, uint32_t); 960 int (*lpfc_hba_down_link) 961 (struct lpfc_hba *, uint32_t); 962 int (*lpfc_selective_reset) 963 (struct lpfc_hba *); 964 965 int (*lpfc_bg_scsi_prep_dma_buf) 966 (struct lpfc_hba *, struct lpfc_io_buf *); 967 968 /* Prep SLI WQE/IOCB jump table entries */ 969 void (*__lpfc_sli_prep_els_req_rsp)(struct lpfc_iocbq *cmdiocbq, 970 struct lpfc_vport *vport, 971 struct lpfc_dmabuf *bmp, 972 u16 cmd_size, u32 did, u32 elscmd, 973 u8 tmo, u8 expect_rsp); 974 void (*__lpfc_sli_prep_gen_req)(struct lpfc_iocbq *cmdiocbq, 975 struct lpfc_dmabuf *bmp, u16 rpi, 976 u32 num_entry, u8 tmo); 977 void (*__lpfc_sli_prep_xmit_seq64)(struct lpfc_iocbq *cmdiocbq, 978 struct lpfc_dmabuf *bmp, u16 rpi, 979 u16 ox_id, u32 num_entry, u8 rctl, 980 u8 last_seq, u8 cr_cx_cmd); 981 void (*__lpfc_sli_prep_abort_xri)(struct lpfc_iocbq *cmdiocbq, 982 u16 ulp_context, u16 iotag, 983 u8 ulp_class, u16 cqid, bool ia, 984 bool wqec); 985 986 /* expedite pool */ 987 struct lpfc_epd_pool epd_pool; 988 989 /* SLI4 specific HBA data structure */ 990 struct lpfc_sli4_hba sli4_hba; 991 992 struct workqueue_struct *wq; 993 struct delayed_work eq_delay_work; 994 995 #define LPFC_IDLE_STAT_DELAY 1000 996 struct delayed_work idle_stat_delay_work; 997 998 struct lpfc_sli sli; 999 uint8_t pci_dev_grp; /* lpfc PCI dev group: 0x0, 0x1, 0x2,... */ 1000 uint32_t sli_rev; /* SLI2, SLI3, or SLI4 */ 1001 uint32_t sli3_options; /* Mask of enabled SLI3 options */ 1002 #define LPFC_SLI3_HBQ_ENABLED 0x01 1003 #define LPFC_SLI3_NPIV_ENABLED 0x02 1004 #define LPFC_SLI3_VPORT_TEARDOWN 0x04 1005 #define LPFC_SLI3_CRP_ENABLED 0x08 1006 #define LPFC_SLI3_BG_ENABLED 0x20 1007 #define LPFC_SLI3_DSS_ENABLED 0x40 1008 #define LPFC_SLI4_PERFH_ENABLED 0x80 1009 #define LPFC_SLI4_PHWQ_ENABLED 0x100 1010 uint32_t iocb_cmd_size; 1011 uint32_t iocb_rsp_size; 1012 1013 struct lpfc_trunk_link trunk_link; 1014 enum hba_state link_state; 1015 uint32_t link_flag; /* link state flags */ 1016 #define LS_LOOPBACK_MODE 0x1 /* NPort is in Loopback mode */ 1017 /* This flag is set while issuing */ 1018 /* INIT_LINK mailbox command */ 1019 #define LS_NPIV_FAB_SUPPORTED 0x2 /* Fabric supports NPIV */ 1020 #define LS_IGNORE_ERATT 0x4 /* intr handler should ignore ERATT */ 1021 #define LS_MDS_LINK_DOWN 0x8 /* MDS Diagnostics Link Down */ 1022 #define LS_MDS_LOOPBACK 0x10 /* MDS Diagnostics Link Up (Loopback) */ 1023 #define LS_CT_VEN_RPA 0x20 /* Vendor RPA sent to switch */ 1024 #define LS_EXTERNAL_LOOPBACK 0x40 /* External loopback plug inserted */ 1025 1026 uint32_t hba_flag; /* hba generic flags */ 1027 #define HBA_ERATT_HANDLED 0x1 /* This flag is set when eratt handled */ 1028 #define DEFER_ERATT 0x2 /* Deferred error attention in progress */ 1029 #define HBA_FCOE_MODE 0x4 /* HBA function in FCoE Mode */ 1030 #define HBA_SP_QUEUE_EVT 0x8 /* Slow-path qevt posted to worker thread*/ 1031 #define HBA_POST_RECEIVE_BUFFER 0x10 /* Rcv buffers need to be posted */ 1032 #define HBA_PERSISTENT_TOPO 0x20 /* Persistent topology support in hba */ 1033 #define ELS_XRI_ABORT_EVENT 0x40 /* ELS_XRI abort event was queued */ 1034 #define ASYNC_EVENT 0x80 1035 #define LINK_DISABLED 0x100 /* Link disabled by user */ 1036 #define FCF_TS_INPROG 0x200 /* FCF table scan in progress */ 1037 #define FCF_RR_INPROG 0x400 /* FCF roundrobin flogi in progress */ 1038 #define HBA_FIP_SUPPORT 0x800 /* FIP support in HBA */ 1039 #define HBA_DEVLOSS_TMO 0x2000 /* HBA in devloss timeout */ 1040 #define HBA_RRQ_ACTIVE 0x4000 /* process the rrq active list */ 1041 #define HBA_IOQ_FLUSH 0x8000 /* FCP/NVME I/O queues being flushed */ 1042 #define HBA_RECOVERABLE_UE 0x20000 /* Firmware supports recoverable UE */ 1043 #define HBA_FORCED_LINK_SPEED 0x40000 /* 1044 * Firmware supports Forced Link Speed 1045 * capability 1046 */ 1047 #define HBA_FLOGI_ISSUED 0x100000 /* FLOGI was issued */ 1048 #define HBA_SHORT_CMF 0x200000 /* shorter CMF timer routine */ 1049 #define HBA_CGN_DAY_WRAP 0x400000 /* HBA Congestion info day wraps */ 1050 #define HBA_DEFER_FLOGI 0x800000 /* Defer FLOGI till read_sparm cmpl */ 1051 #define HBA_SETUP 0x1000000 /* Signifies HBA setup is completed */ 1052 #define HBA_NEEDS_CFG_PORT 0x2000000 /* SLI3 - needs a CONFIG_PORT mbox */ 1053 #define HBA_HBEAT_INP 0x4000000 /* mbox HBEAT is in progress */ 1054 #define HBA_HBEAT_TMO 0x8000000 /* HBEAT initiated after timeout */ 1055 #define HBA_FLOGI_OUTSTANDING 0x10000000 /* FLOGI is outstanding */ 1056 #define HBA_RHBA_CMPL 0x20000000 /* RHBA FDMI command is successful */ 1057 1058 struct completion *fw_dump_cmpl; /* cmpl event tracker for fw_dump */ 1059 uint32_t fcp_ring_in_use; /* When polling test if intr-hndlr active*/ 1060 struct lpfc_dmabuf slim2p; 1061 1062 MAILBOX_t *mbox; 1063 uint32_t *mbox_ext; 1064 struct lpfc_mbox_ext_buf_ctx mbox_ext_buf_ctx; 1065 uint32_t ha_copy; 1066 struct _PCB *pcb; 1067 struct _IOCB *IOCBs; 1068 1069 struct lpfc_dmabuf hbqslimp; 1070 1071 uint16_t pci_cfg_value; 1072 1073 uint8_t fc_linkspeed; /* Link speed after last READ_LA */ 1074 1075 uint32_t fc_eventTag; /* event tag for link attention */ 1076 uint32_t link_events; 1077 1078 /* These fields used to be binfo */ 1079 uint32_t fc_pref_DID; /* preferred D_ID */ 1080 uint8_t fc_pref_ALPA; /* preferred AL_PA */ 1081 uint32_t fc_edtovResol; /* E_D_TOV timer resolution */ 1082 uint32_t fc_edtov; /* E_D_TOV timer value */ 1083 uint32_t fc_arbtov; /* ARB_TOV timer value */ 1084 uint32_t fc_ratov; /* R_A_TOV timer value */ 1085 uint32_t fc_rttov; /* R_T_TOV timer value */ 1086 uint32_t fc_altov; /* AL_TOV timer value */ 1087 uint32_t fc_crtov; /* C_R_TOV timer value */ 1088 1089 struct serv_parm fc_fabparam; /* fabric service parameters buffer */ 1090 uint8_t alpa_map[128]; /* AL_PA map from READ_LA */ 1091 1092 uint32_t lmt; 1093 1094 uint32_t fc_topology; /* link topology, from LINK INIT */ 1095 uint32_t fc_topology_changed; /* link topology, from LINK INIT */ 1096 1097 struct lpfc_stats fc_stat; 1098 1099 struct lpfc_nodelist fc_fcpnodev; /* nodelist entry for no device */ 1100 uint32_t nport_event_cnt; /* timestamp for nlplist entry */ 1101 1102 uint8_t wwnn[8]; 1103 uint8_t wwpn[8]; 1104 uint32_t RandomData[7]; 1105 uint8_t fcp_embed_io; 1106 uint8_t nvmet_support; /* driver supports NVMET */ 1107 #define LPFC_NVMET_MAX_PORTS 32 1108 uint8_t mds_diags_support; 1109 uint8_t bbcredit_support; 1110 uint8_t enab_exp_wqcq_pages; 1111 u8 nsler; /* Firmware supports FC-NVMe-2 SLER */ 1112 1113 /* HBA Config Parameters */ 1114 uint32_t cfg_ack0; 1115 uint32_t cfg_xri_rebalancing; 1116 uint32_t cfg_xpsgl; 1117 uint32_t cfg_enable_npiv; 1118 uint32_t cfg_enable_rrq; 1119 uint32_t cfg_topology; 1120 uint32_t cfg_link_speed; 1121 #define LPFC_FCF_FOV 1 /* Fast fcf failover */ 1122 #define LPFC_FCF_PRIORITY 2 /* Priority fcf failover */ 1123 uint32_t cfg_fcf_failover_policy; 1124 uint32_t cfg_fcp_io_sched; 1125 uint32_t cfg_ns_query; 1126 uint32_t cfg_fcp2_no_tgt_reset; 1127 uint32_t cfg_cr_delay; 1128 uint32_t cfg_cr_count; 1129 uint32_t cfg_multi_ring_support; 1130 uint32_t cfg_multi_ring_rctl; 1131 uint32_t cfg_multi_ring_type; 1132 uint32_t cfg_poll; 1133 uint32_t cfg_poll_tmo; 1134 uint32_t cfg_task_mgmt_tmo; 1135 uint32_t cfg_use_msi; 1136 uint32_t cfg_auto_imax; 1137 uint32_t cfg_fcp_imax; 1138 uint32_t cfg_force_rscn; 1139 uint32_t cfg_cq_poll_threshold; 1140 uint32_t cfg_cq_max_proc_limit; 1141 uint32_t cfg_fcp_cpu_map; 1142 uint32_t cfg_fcp_mq_threshold; 1143 uint32_t cfg_hdw_queue; 1144 uint32_t cfg_irq_chann; 1145 uint32_t cfg_suppress_rsp; 1146 uint32_t cfg_nvme_oas; 1147 uint32_t cfg_nvme_embed_cmd; 1148 uint32_t cfg_nvmet_mrq_post; 1149 uint32_t cfg_nvmet_mrq; 1150 uint32_t cfg_enable_nvmet; 1151 uint32_t cfg_nvme_enable_fb; 1152 uint32_t cfg_nvmet_fb_size; 1153 uint32_t cfg_total_seg_cnt; 1154 uint32_t cfg_sg_seg_cnt; 1155 uint32_t cfg_nvme_seg_cnt; 1156 uint32_t cfg_scsi_seg_cnt; 1157 uint32_t cfg_sg_dma_buf_size; 1158 uint32_t cfg_hba_queue_depth; 1159 uint32_t cfg_enable_hba_reset; 1160 uint32_t cfg_enable_hba_heartbeat; 1161 uint32_t cfg_fof; 1162 uint32_t cfg_EnableXLane; 1163 uint8_t cfg_oas_tgt_wwpn[8]; 1164 uint8_t cfg_oas_vpt_wwpn[8]; 1165 uint32_t cfg_oas_lun_state; 1166 #define OAS_LUN_ENABLE 1 1167 #define OAS_LUN_DISABLE 0 1168 uint32_t cfg_oas_lun_status; 1169 #define OAS_LUN_STATUS_EXISTS 0x01 1170 uint32_t cfg_oas_flags; 1171 #define OAS_FIND_ANY_VPORT 0x01 1172 #define OAS_FIND_ANY_TARGET 0x02 1173 #define OAS_LUN_VALID 0x04 1174 uint32_t cfg_oas_priority; 1175 uint32_t cfg_XLanePriority; 1176 uint32_t cfg_enable_bg; 1177 uint32_t cfg_prot_mask; 1178 uint32_t cfg_prot_guard; 1179 uint32_t cfg_hostmem_hgp; 1180 uint32_t cfg_log_verbose; 1181 uint32_t cfg_enable_fc4_type; 1182 #define LPFC_ENABLE_FCP 1 1183 #define LPFC_ENABLE_NVME 2 1184 #define LPFC_ENABLE_BOTH 3 1185 #if (IS_ENABLED(CONFIG_NVME_FC)) 1186 #define LPFC_MAX_ENBL_FC4_TYPE LPFC_ENABLE_BOTH 1187 #define LPFC_DEF_ENBL_FC4_TYPE LPFC_ENABLE_BOTH 1188 #else 1189 #define LPFC_MAX_ENBL_FC4_TYPE LPFC_ENABLE_FCP 1190 #define LPFC_DEF_ENBL_FC4_TYPE LPFC_ENABLE_FCP 1191 #endif 1192 uint32_t cfg_sriov_nr_virtfn; 1193 uint32_t cfg_request_firmware_upgrade; 1194 uint32_t cfg_suppress_link_up; 1195 uint32_t cfg_rrq_xri_bitmap_sz; 1196 u32 cfg_fcp_wait_abts_rsp; 1197 uint32_t cfg_delay_discovery; 1198 uint32_t cfg_sli_mode; 1199 #define LPFC_INITIALIZE_LINK 0 /* do normal init_link mbox */ 1200 #define LPFC_DELAY_INIT_LINK 1 /* layered driver hold off */ 1201 #define LPFC_DELAY_INIT_LINK_INDEFINITELY 2 /* wait, manual intervention */ 1202 uint32_t cfg_fdmi_on; 1203 #define LPFC_FDMI_NO_SUPPORT 0 /* FDMI not supported */ 1204 #define LPFC_FDMI_SUPPORT 1 /* FDMI supported? */ 1205 uint32_t cfg_enable_SmartSAN; 1206 uint32_t cfg_enable_mds_diags; 1207 uint32_t cfg_ras_fwlog_level; 1208 uint32_t cfg_ras_fwlog_buffsize; 1209 uint32_t cfg_ras_fwlog_func; 1210 uint32_t cfg_enable_bbcr; /* Enable BB Credit Recovery */ 1211 uint32_t cfg_enable_dpp; /* Enable Direct Packet Push */ 1212 uint32_t cfg_enable_pbde; 1213 uint32_t cfg_enable_mi; 1214 struct nvmet_fc_target_port *targetport; 1215 lpfc_vpd_t vpd; /* vital product data */ 1216 1217 u32 cfg_max_vmid; /* maximum VMIDs allowed per port */ 1218 u32 cfg_vmid_app_header; 1219 #define LPFC_VMID_APP_HEADER_DISABLE 0 1220 #define LPFC_VMID_APP_HEADER_ENABLE 1 1221 u32 cfg_vmid_priority_tagging; 1222 u32 cfg_vmid_inactivity_timeout; /* Time after which the VMID */ 1223 /* deregisters from switch */ 1224 struct pci_dev *pcidev; 1225 struct list_head work_list; 1226 uint32_t work_ha; /* Host Attention Bits for WT */ 1227 uint32_t work_ha_mask; /* HA Bits owned by WT */ 1228 uint32_t work_hs; /* HS stored in case of ERRAT */ 1229 uint32_t work_status[2]; /* Extra status from SLIM */ 1230 1231 wait_queue_head_t work_waitq; 1232 struct task_struct *worker_thread; 1233 unsigned long data_flags; 1234 uint32_t border_sge_num; 1235 1236 uint32_t hbq_in_use; /* HBQs in use flag */ 1237 uint32_t hbq_count; /* Count of configured HBQs */ 1238 struct hbq_s hbqs[LPFC_MAX_HBQS]; /* local copy of hbq indicies */ 1239 1240 atomic_t fcp_qidx; /* next FCP WQ (RR Policy) */ 1241 atomic_t nvme_qidx; /* next NVME WQ (RR Policy) */ 1242 1243 phys_addr_t pci_bar0_map; /* Physical address for PCI BAR0 */ 1244 phys_addr_t pci_bar1_map; /* Physical address for PCI BAR1 */ 1245 phys_addr_t pci_bar2_map; /* Physical address for PCI BAR2 */ 1246 void __iomem *slim_memmap_p; /* Kernel memory mapped address for 1247 PCI BAR0 */ 1248 void __iomem *ctrl_regs_memmap_p;/* Kernel memory mapped address for 1249 PCI BAR2 */ 1250 1251 void __iomem *pci_bar0_memmap_p; /* Kernel memory mapped address for 1252 PCI BAR0 with dual-ULP support */ 1253 void __iomem *pci_bar2_memmap_p; /* Kernel memory mapped address for 1254 PCI BAR2 with dual-ULP support */ 1255 void __iomem *pci_bar4_memmap_p; /* Kernel memory mapped address for 1256 PCI BAR4 with dual-ULP support */ 1257 #define PCI_64BIT_BAR0 0 1258 #define PCI_64BIT_BAR2 2 1259 #define PCI_64BIT_BAR4 4 1260 void __iomem *MBslimaddr; /* virtual address for mbox cmds */ 1261 void __iomem *HAregaddr; /* virtual address for host attn reg */ 1262 void __iomem *CAregaddr; /* virtual address for chip attn reg */ 1263 void __iomem *HSregaddr; /* virtual address for host status 1264 reg */ 1265 void __iomem *HCregaddr; /* virtual address for host ctl reg */ 1266 1267 struct lpfc_hgp __iomem *host_gp; /* Host side get/put pointers */ 1268 struct lpfc_pgp *port_gp; 1269 uint32_t __iomem *hbq_put; /* Address in SLIM to HBQ put ptrs */ 1270 uint32_t *hbq_get; /* Host mem address of HBQ get ptrs */ 1271 1272 int brd_no; /* FC board number */ 1273 char SerialNumber[32]; /* adapter Serial Number */ 1274 char OptionROMVersion[32]; /* adapter BIOS / Fcode version */ 1275 char BIOSVersion[16]; /* Boot BIOS version */ 1276 char ModelDesc[256]; /* Model Description */ 1277 char ModelName[80]; /* Model Name */ 1278 char ProgramType[256]; /* Program Type */ 1279 char Port[20]; /* Port No */ 1280 uint8_t vpd_flag; /* VPD data flag */ 1281 1282 #define VPD_MODEL_DESC 0x1 /* valid vpd model description */ 1283 #define VPD_MODEL_NAME 0x2 /* valid vpd model name */ 1284 #define VPD_PROGRAM_TYPE 0x4 /* valid vpd program type */ 1285 #define VPD_PORT 0x8 /* valid vpd port data */ 1286 #define VPD_MASK 0xf /* mask for any vpd data */ 1287 1288 1289 struct timer_list fcp_poll_timer; 1290 struct timer_list eratt_poll; 1291 uint32_t eratt_poll_interval; 1292 1293 uint64_t bg_guard_err_cnt; 1294 uint64_t bg_apptag_err_cnt; 1295 uint64_t bg_reftag_err_cnt; 1296 1297 /* fastpath list. */ 1298 spinlock_t scsi_buf_list_get_lock; /* SCSI buf alloc list lock */ 1299 spinlock_t scsi_buf_list_put_lock; /* SCSI buf free list lock */ 1300 struct list_head lpfc_scsi_buf_list_get; 1301 struct list_head lpfc_scsi_buf_list_put; 1302 uint32_t total_scsi_bufs; 1303 struct list_head lpfc_iocb_list; 1304 uint32_t total_iocbq_bufs; 1305 struct list_head active_rrq_list; 1306 spinlock_t hbalock; 1307 struct work_struct unblock_request_work; /* SCSI layer unblock IOs */ 1308 1309 /* dma_mem_pools */ 1310 struct dma_pool *lpfc_sg_dma_buf_pool; 1311 struct dma_pool *lpfc_mbuf_pool; 1312 struct dma_pool *lpfc_hrb_pool; /* header receive buffer pool */ 1313 struct dma_pool *lpfc_drb_pool; /* data receive buffer pool */ 1314 struct dma_pool *lpfc_nvmet_drb_pool; /* data receive buffer pool */ 1315 struct dma_pool *lpfc_hbq_pool; /* SLI3 hbq buffer pool */ 1316 struct dma_pool *lpfc_cmd_rsp_buf_pool; 1317 struct lpfc_dma_pool lpfc_mbuf_safety_pool; 1318 1319 mempool_t *mbox_mem_pool; 1320 mempool_t *nlp_mem_pool; 1321 mempool_t *rrq_pool; 1322 mempool_t *active_rrq_pool; 1323 1324 struct fc_host_statistics link_stats; 1325 enum lpfc_irq_chann_mode irq_chann_mode; 1326 enum intr_type_t intr_type; 1327 uint32_t intr_mode; 1328 #define LPFC_INTR_ERROR 0xFFFFFFFF 1329 struct list_head port_list; 1330 spinlock_t port_list_lock; /* lock for port_list mutations */ 1331 struct lpfc_vport *pport; /* physical lpfc_vport pointer */ 1332 uint16_t max_vpi; /* Maximum virtual nports */ 1333 #define LPFC_MAX_VPI 0xFF /* Max number VPI supported 0 - 0xff */ 1334 #define LPFC_MAX_VPORTS 0x100 /* Max vports per port, with pport */ 1335 uint16_t max_vports; /* 1336 * For IOV HBAs max_vpi can change 1337 * after a reset. max_vports is max 1338 * number of vports present. This can 1339 * be greater than max_vpi. 1340 */ 1341 uint16_t vpi_base; 1342 uint16_t vfi_base; 1343 unsigned long *vpi_bmask; /* vpi allocation table */ 1344 uint16_t *vpi_ids; 1345 uint16_t vpi_count; 1346 struct list_head lpfc_vpi_blk_list; 1347 1348 /* Data structure used by fabric iocb scheduler */ 1349 struct list_head fabric_iocb_list; 1350 atomic_t fabric_iocb_count; 1351 struct timer_list fabric_block_timer; 1352 unsigned long bit_flags; 1353 atomic_t num_rsrc_err; 1354 atomic_t num_cmd_success; 1355 unsigned long last_rsrc_error_time; 1356 unsigned long last_ramp_down_time; 1357 #ifdef CONFIG_SCSI_LPFC_DEBUG_FS 1358 struct dentry *hba_debugfs_root; 1359 atomic_t debugfs_vport_count; 1360 struct dentry *debug_multixri_pools; 1361 struct dentry *debug_hbqinfo; 1362 struct dentry *debug_dumpHostSlim; 1363 struct dentry *debug_dumpHBASlim; 1364 struct dentry *debug_InjErrLBA; /* LBA to inject errors at */ 1365 struct dentry *debug_InjErrNPortID; /* NPortID to inject errors at */ 1366 struct dentry *debug_InjErrWWPN; /* WWPN to inject errors at */ 1367 struct dentry *debug_writeGuard; /* inject write guard_tag errors */ 1368 struct dentry *debug_writeApp; /* inject write app_tag errors */ 1369 struct dentry *debug_writeRef; /* inject write ref_tag errors */ 1370 struct dentry *debug_readGuard; /* inject read guard_tag errors */ 1371 struct dentry *debug_readApp; /* inject read app_tag errors */ 1372 struct dentry *debug_readRef; /* inject read ref_tag errors */ 1373 1374 struct dentry *debug_nvmeio_trc; 1375 struct lpfc_debugfs_nvmeio_trc *nvmeio_trc; 1376 struct dentry *debug_hdwqinfo; 1377 #ifdef LPFC_HDWQ_LOCK_STAT 1378 struct dentry *debug_lockstat; 1379 #endif 1380 struct dentry *debug_cgn_buffer; 1381 struct dentry *debug_rx_monitor; 1382 struct dentry *debug_ras_log; 1383 atomic_t nvmeio_trc_cnt; 1384 uint32_t nvmeio_trc_size; 1385 uint32_t nvmeio_trc_output_idx; 1386 1387 /* T10 DIF error injection */ 1388 uint32_t lpfc_injerr_wgrd_cnt; 1389 uint32_t lpfc_injerr_wapp_cnt; 1390 uint32_t lpfc_injerr_wref_cnt; 1391 uint32_t lpfc_injerr_rgrd_cnt; 1392 uint32_t lpfc_injerr_rapp_cnt; 1393 uint32_t lpfc_injerr_rref_cnt; 1394 uint32_t lpfc_injerr_nportid; 1395 struct lpfc_name lpfc_injerr_wwpn; 1396 sector_t lpfc_injerr_lba; 1397 #define LPFC_INJERR_LBA_OFF (sector_t)(-1) 1398 1399 struct dentry *debug_slow_ring_trc; 1400 struct lpfc_debugfs_trc *slow_ring_trc; 1401 atomic_t slow_ring_trc_cnt; 1402 /* iDiag debugfs sub-directory */ 1403 struct dentry *idiag_root; 1404 struct dentry *idiag_pci_cfg; 1405 struct dentry *idiag_bar_acc; 1406 struct dentry *idiag_que_info; 1407 struct dentry *idiag_que_acc; 1408 struct dentry *idiag_drb_acc; 1409 struct dentry *idiag_ctl_acc; 1410 struct dentry *idiag_mbx_acc; 1411 struct dentry *idiag_ext_acc; 1412 uint8_t lpfc_idiag_last_eq; 1413 #endif 1414 uint16_t nvmeio_trc_on; 1415 1416 /* Used for deferred freeing of ELS data buffers */ 1417 struct list_head elsbuf; 1418 int elsbuf_cnt; 1419 int elsbuf_prev_cnt; 1420 1421 uint8_t temp_sensor_support; 1422 /* Fields used for heart beat. */ 1423 unsigned long last_completion_time; 1424 unsigned long skipped_hb; 1425 struct timer_list hb_tmofunc; 1426 struct timer_list rrq_tmr; 1427 enum hba_temp_state over_temp_state; 1428 /* 1429 * Following bit will be set for all buffer tags which are not 1430 * associated with any HBQ. 1431 */ 1432 #define QUE_BUFTAG_BIT (1<<31) 1433 uint32_t buffer_tag_count; 1434 1435 /* Maximum number of events that can be outstanding at any time*/ 1436 #define LPFC_MAX_EVT_COUNT 512 1437 atomic_t fast_event_count; 1438 uint32_t fcoe_eventtag; 1439 uint32_t fcoe_eventtag_at_fcf_scan; 1440 uint32_t fcoe_cvl_eventtag; 1441 uint32_t fcoe_cvl_eventtag_attn; 1442 struct lpfc_fcf fcf; 1443 uint8_t fc_map[3]; 1444 uint8_t valid_vlan; 1445 uint16_t vlan_id; 1446 struct list_head fcf_conn_rec_list; 1447 1448 bool defer_flogi_acc_flag; 1449 uint16_t defer_flogi_acc_rx_id; 1450 uint16_t defer_flogi_acc_ox_id; 1451 1452 spinlock_t ct_ev_lock; /* synchronize access to ct_ev_waiters */ 1453 struct list_head ct_ev_waiters; 1454 struct unsol_rcv_ct_ctx ct_ctx[LPFC_CT_CTX_MAX]; 1455 uint32_t ctx_idx; 1456 struct timer_list inactive_vmid_poll; 1457 1458 /* RAS Support */ 1459 struct lpfc_ras_fwlog ras_fwlog; 1460 1461 uint32_t iocb_cnt; 1462 uint32_t iocb_max; 1463 atomic_t sdev_cnt; 1464 spinlock_t devicelock; /* lock for luns list */ 1465 mempool_t *device_data_mem_pool; 1466 struct list_head luns; 1467 #define LPFC_TRANSGRESSION_HIGH_TEMPERATURE 0x0080 1468 #define LPFC_TRANSGRESSION_LOW_TEMPERATURE 0x0040 1469 #define LPFC_TRANSGRESSION_HIGH_VOLTAGE 0x0020 1470 #define LPFC_TRANSGRESSION_LOW_VOLTAGE 0x0010 1471 #define LPFC_TRANSGRESSION_HIGH_TXBIAS 0x0008 1472 #define LPFC_TRANSGRESSION_LOW_TXBIAS 0x0004 1473 #define LPFC_TRANSGRESSION_HIGH_TXPOWER 0x0002 1474 #define LPFC_TRANSGRESSION_LOW_TXPOWER 0x0001 1475 #define LPFC_TRANSGRESSION_HIGH_RXPOWER 0x8000 1476 #define LPFC_TRANSGRESSION_LOW_RXPOWER 0x4000 1477 uint16_t sfp_alarm; 1478 uint16_t sfp_warning; 1479 1480 #ifdef CONFIG_SCSI_LPFC_DEBUG_FS 1481 uint16_t hdwqstat_on; 1482 #define LPFC_CHECK_OFF 0 1483 #define LPFC_CHECK_NVME_IO 1 1484 #define LPFC_CHECK_NVMET_IO 2 1485 #define LPFC_CHECK_SCSI_IO 4 1486 uint16_t ktime_on; 1487 uint64_t ktime_data_samples; 1488 uint64_t ktime_status_samples; 1489 uint64_t ktime_last_cmd; 1490 uint64_t ktime_seg1_total; 1491 uint64_t ktime_seg1_min; 1492 uint64_t ktime_seg1_max; 1493 uint64_t ktime_seg2_total; 1494 uint64_t ktime_seg2_min; 1495 uint64_t ktime_seg2_max; 1496 uint64_t ktime_seg3_total; 1497 uint64_t ktime_seg3_min; 1498 uint64_t ktime_seg3_max; 1499 uint64_t ktime_seg4_total; 1500 uint64_t ktime_seg4_min; 1501 uint64_t ktime_seg4_max; 1502 uint64_t ktime_seg5_total; 1503 uint64_t ktime_seg5_min; 1504 uint64_t ktime_seg5_max; 1505 uint64_t ktime_seg6_total; 1506 uint64_t ktime_seg6_min; 1507 uint64_t ktime_seg6_max; 1508 uint64_t ktime_seg7_total; 1509 uint64_t ktime_seg7_min; 1510 uint64_t ktime_seg7_max; 1511 uint64_t ktime_seg8_total; 1512 uint64_t ktime_seg8_min; 1513 uint64_t ktime_seg8_max; 1514 uint64_t ktime_seg9_total; 1515 uint64_t ktime_seg9_min; 1516 uint64_t ktime_seg9_max; 1517 uint64_t ktime_seg10_total; 1518 uint64_t ktime_seg10_min; 1519 uint64_t ktime_seg10_max; 1520 #endif 1521 /* CMF objects */ 1522 struct lpfc_cgn_stat __percpu *cmf_stat; 1523 uint32_t cmf_interval_rate; /* timer interval limit in ms */ 1524 uint32_t cmf_timer_cnt; 1525 #define LPFC_CMF_INTERVAL 90 1526 uint64_t cmf_link_byte_count; 1527 uint64_t cmf_max_line_rate; 1528 uint64_t cmf_max_bytes_per_interval; 1529 uint64_t cmf_last_sync_bw; 1530 #define LPFC_CMF_BLK_SIZE 512 1531 struct hrtimer cmf_timer; 1532 atomic_t cmf_bw_wait; 1533 atomic_t cmf_busy; 1534 atomic_t cmf_stop_io; /* To block request and stop IO's */ 1535 uint32_t cmf_active_mode; 1536 uint32_t cmf_info_per_interval; 1537 #define LPFC_MAX_CMF_INFO 32 1538 struct timespec64 cmf_latency; /* Interval congestion timestamp */ 1539 uint32_t cmf_last_ts; /* Interval congestion time (ms) */ 1540 uint32_t cmf_active_info; 1541 1542 /* Signal / FPIN handling for Congestion Mgmt */ 1543 u8 cgn_reg_fpin; /* Negotiated value from RDF */ 1544 u8 cgn_init_reg_fpin; /* Initial value from READ_CONFIG */ 1545 #define LPFC_CGN_FPIN_NONE 0x0 1546 #define LPFC_CGN_FPIN_WARN 0x1 1547 #define LPFC_CGN_FPIN_ALARM 0x2 1548 #define LPFC_CGN_FPIN_BOTH (LPFC_CGN_FPIN_WARN | LPFC_CGN_FPIN_ALARM) 1549 1550 u8 cgn_reg_signal; /* Negotiated value from EDC */ 1551 u8 cgn_init_reg_signal; /* Initial value from READ_CONFIG */ 1552 /* cgn_reg_signal and cgn_init_reg_signal use 1553 * enum fc_edc_cg_signal_cap_types 1554 */ 1555 u16 cgn_fpin_frequency; /* In units of msecs */ 1556 #define LPFC_FPIN_INIT_FREQ 0xffff 1557 u32 cgn_sig_freq; 1558 u32 cgn_acqe_cnt; 1559 1560 /* RX monitor handling for CMF */ 1561 struct lpfc_rx_info_monitor *rx_monitor; 1562 atomic_t rx_max_read_cnt; /* Maximum read bytes */ 1563 uint64_t rx_block_cnt; 1564 1565 /* Congestion parameters from flash */ 1566 struct lpfc_cgn_param cgn_p; 1567 1568 /* Statistics counter for ACQE cgn alarms and warnings */ 1569 struct lpfc_cgn_acqe_stat cgn_acqe_stat; 1570 1571 /* Congestion buffer information */ 1572 struct lpfc_dmabuf *cgn_i; /* Congestion Info buffer */ 1573 atomic_t cgn_fabric_warn_cnt; /* Total warning cgn events for info */ 1574 atomic_t cgn_fabric_alarm_cnt; /* Total alarm cgn events for info */ 1575 atomic_t cgn_sync_warn_cnt; /* Total warning events for SYNC wqe */ 1576 atomic_t cgn_sync_alarm_cnt; /* Total alarm events for SYNC wqe */ 1577 atomic_t cgn_driver_evt_cnt; /* Total driver cgn events for fmw */ 1578 atomic_t cgn_latency_evt_cnt; 1579 struct timespec64 cgn_daily_ts; 1580 atomic64_t cgn_latency_evt; /* Avg latency per minute */ 1581 unsigned long cgn_evt_timestamp; 1582 #define LPFC_CGN_TIMER_TO_MIN 60000 /* ms in a minute */ 1583 uint32_t cgn_evt_minute; 1584 #define LPFC_SEC_MIN 60 1585 #define LPFC_MIN_HOUR 60 1586 #define LPFC_HOUR_DAY 24 1587 #define LPFC_MIN_DAY (LPFC_MIN_HOUR * LPFC_HOUR_DAY) 1588 1589 struct hlist_node cpuhp; /* used for cpuhp per hba callback */ 1590 struct timer_list cpuhp_poll_timer; 1591 struct list_head poll_list; /* slowpath eq polling list */ 1592 #define LPFC_POLL_HB 1 /* slowpath heartbeat */ 1593 1594 char os_host_name[MAXHOSTNAMELEN]; 1595 1596 /* LD Signaling */ 1597 u32 degrade_activate_threshold; 1598 u32 degrade_deactivate_threshold; 1599 u32 fec_degrade_interval; 1600 1601 atomic_t dbg_log_idx; 1602 atomic_t dbg_log_cnt; 1603 atomic_t dbg_log_dmping; 1604 struct dbg_log_ent dbg_log[DBG_LOG_SZ]; 1605 }; 1606 1607 #define LPFC_MAX_RXMONITOR_ENTRY 800 1608 #define LPFC_MAX_RXMONITOR_DUMP 32 1609 struct rx_info_entry { 1610 uint64_t cmf_bytes; /* Total no of read bytes for CMF_SYNC_WQE */ 1611 uint64_t total_bytes; /* Total no of read bytes requested */ 1612 uint64_t rcv_bytes; /* Total no of read bytes completed */ 1613 uint64_t avg_io_size; 1614 uint64_t avg_io_latency;/* Average io latency in microseconds */ 1615 uint64_t max_read_cnt; /* Maximum read bytes */ 1616 uint64_t max_bytes_per_interval; 1617 uint32_t cmf_busy; 1618 uint32_t cmf_info; /* CMF_SYNC_WQE info */ 1619 uint32_t io_cnt; 1620 uint32_t timer_utilization; 1621 uint32_t timer_interval; 1622 }; 1623 1624 struct lpfc_rx_info_monitor { 1625 struct rx_info_entry *ring; /* info organized in a circular buffer */ 1626 u32 head_idx, tail_idx; /* index to head/tail of ring */ 1627 spinlock_t lock; /* spinlock for ring */ 1628 u32 entries; /* storing number entries/size of ring */ 1629 }; 1630 1631 static inline struct Scsi_Host * 1632 lpfc_shost_from_vport(struct lpfc_vport *vport) 1633 { 1634 return container_of((void *) vport, struct Scsi_Host, hostdata[0]); 1635 } 1636 1637 static inline void 1638 lpfc_set_loopback_flag(struct lpfc_hba *phba) 1639 { 1640 if (phba->cfg_topology == FLAGS_LOCAL_LB) 1641 phba->link_flag |= LS_LOOPBACK_MODE; 1642 else 1643 phba->link_flag &= ~LS_LOOPBACK_MODE; 1644 } 1645 1646 static inline int 1647 lpfc_is_link_up(struct lpfc_hba *phba) 1648 { 1649 return phba->link_state == LPFC_LINK_UP || 1650 phba->link_state == LPFC_CLEAR_LA || 1651 phba->link_state == LPFC_HBA_READY; 1652 } 1653 1654 static inline void 1655 lpfc_worker_wake_up(struct lpfc_hba *phba) 1656 { 1657 /* Set the lpfc data pending flag */ 1658 set_bit(LPFC_DATA_READY, &phba->data_flags); 1659 1660 /* Wake up worker thread */ 1661 wake_up(&phba->work_waitq); 1662 return; 1663 } 1664 1665 static inline int 1666 lpfc_readl(void __iomem *addr, uint32_t *data) 1667 { 1668 uint32_t temp; 1669 temp = readl(addr); 1670 if (temp == 0xffffffff) 1671 return -EIO; 1672 *data = temp; 1673 return 0; 1674 } 1675 1676 static inline int 1677 lpfc_sli_read_hs(struct lpfc_hba *phba) 1678 { 1679 /* 1680 * There was a link/board error. Read the status register to retrieve 1681 * the error event and process it. 1682 */ 1683 phba->sli.slistat.err_attn_event++; 1684 1685 /* Save status info and check for unplug error */ 1686 if (lpfc_readl(phba->HSregaddr, &phba->work_hs) || 1687 lpfc_readl(phba->MBslimaddr + 0xa8, &phba->work_status[0]) || 1688 lpfc_readl(phba->MBslimaddr + 0xac, &phba->work_status[1])) { 1689 return -EIO; 1690 } 1691 1692 /* Clear chip Host Attention error bit */ 1693 writel(HA_ERATT, phba->HAregaddr); 1694 readl(phba->HAregaddr); /* flush */ 1695 phba->pport->stopped = 1; 1696 1697 return 0; 1698 } 1699 1700 static inline struct lpfc_sli_ring * 1701 lpfc_phba_elsring(struct lpfc_hba *phba) 1702 { 1703 /* Return NULL if sli_rev has become invalid due to bad fw */ 1704 if (phba->sli_rev != LPFC_SLI_REV4 && 1705 phba->sli_rev != LPFC_SLI_REV3 && 1706 phba->sli_rev != LPFC_SLI_REV2) 1707 return NULL; 1708 1709 if (phba->sli_rev == LPFC_SLI_REV4) { 1710 if (phba->sli4_hba.els_wq) 1711 return phba->sli4_hba.els_wq->pring; 1712 else 1713 return NULL; 1714 } 1715 return &phba->sli.sli3_ring[LPFC_ELS_RING]; 1716 } 1717 1718 /** 1719 * lpfc_next_online_cpu - Finds next online CPU on cpumask 1720 * @mask: Pointer to phba's cpumask member. 1721 * @start: starting cpu index 1722 * 1723 * Note: If no valid cpu found, then nr_cpu_ids is returned. 1724 * 1725 **/ 1726 static inline unsigned int 1727 lpfc_next_online_cpu(const struct cpumask *mask, unsigned int start) 1728 { 1729 unsigned int cpu_it; 1730 1731 for_each_cpu_wrap(cpu_it, mask, start) { 1732 if (cpu_online(cpu_it)) 1733 break; 1734 } 1735 1736 return cpu_it; 1737 } 1738 /** 1739 * lpfc_sli4_mod_hba_eq_delay - update EQ delay 1740 * @phba: Pointer to HBA context object. 1741 * @q: The Event Queue to update. 1742 * @delay: The delay value (in us) to be written. 1743 * 1744 **/ 1745 static inline void 1746 lpfc_sli4_mod_hba_eq_delay(struct lpfc_hba *phba, struct lpfc_queue *eq, 1747 u32 delay) 1748 { 1749 struct lpfc_register reg_data; 1750 1751 reg_data.word0 = 0; 1752 bf_set(lpfc_sliport_eqdelay_id, ®_data, eq->queue_id); 1753 bf_set(lpfc_sliport_eqdelay_delay, ®_data, delay); 1754 writel(reg_data.word0, phba->sli4_hba.u.if_type2.EQDregaddr); 1755 eq->q_mode = delay; 1756 } 1757 1758 1759 /* 1760 * Macro that declares tables and a routine to perform enum type to 1761 * ascii string lookup. 1762 * 1763 * Defines a <key,value> table for an enum. Uses xxx_INIT defines for 1764 * the enum to populate the table. Macro defines a routine (named 1765 * by caller) that will search all elements of the table for the key 1766 * and return the name string if found or "Unrecognized" if not found. 1767 */ 1768 #define DECLARE_ENUM2STR_LOOKUP(routine, enum_name, enum_init) \ 1769 static struct { \ 1770 enum enum_name value; \ 1771 char *name; \ 1772 } fc_##enum_name##_e2str_names[] = enum_init; \ 1773 static const char *routine(enum enum_name table_key) \ 1774 { \ 1775 int i; \ 1776 char *name = "Unrecognized"; \ 1777 \ 1778 for (i = 0; i < ARRAY_SIZE(fc_##enum_name##_e2str_names); i++) {\ 1779 if (fc_##enum_name##_e2str_names[i].value == table_key) {\ 1780 name = fc_##enum_name##_e2str_names[i].name; \ 1781 break; \ 1782 } \ 1783 } \ 1784 return name; \ 1785 } 1786 1787 /** 1788 * lpfc_is_vmid_enabled - returns if VMID is enabled for either switch types 1789 * @phba: Pointer to HBA context object. 1790 * 1791 * Relationship between the enable, target support and if vmid tag is required 1792 * for the particular combination 1793 * --------------------------------------------------- 1794 * Switch Enable Flag Target Support VMID Needed 1795 * --------------------------------------------------- 1796 * App Id 0 NA N 1797 * App Id 1 0 N 1798 * App Id 1 1 Y 1799 * Pr Tag 0 NA N 1800 * Pr Tag 1 0 N 1801 * Pr Tag 1 1 Y 1802 * Pr Tag 2 * Y 1803 --------------------------------------------------- 1804 * 1805 **/ 1806 static inline int lpfc_is_vmid_enabled(struct lpfc_hba *phba) 1807 { 1808 return phba->cfg_vmid_app_header || phba->cfg_vmid_priority_tagging; 1809 } 1810 1811 static inline 1812 u8 get_job_ulpstatus(struct lpfc_hba *phba, struct lpfc_iocbq *iocbq) 1813 { 1814 if (phba->sli_rev == LPFC_SLI_REV4) 1815 return bf_get(lpfc_wcqe_c_status, &iocbq->wcqe_cmpl); 1816 else 1817 return iocbq->iocb.ulpStatus; 1818 } 1819 1820 static inline 1821 u32 get_job_word4(struct lpfc_hba *phba, struct lpfc_iocbq *iocbq) 1822 { 1823 if (phba->sli_rev == LPFC_SLI_REV4) 1824 return iocbq->wcqe_cmpl.parameter; 1825 else 1826 return iocbq->iocb.un.ulpWord[4]; 1827 } 1828 1829 static inline 1830 u8 get_job_cmnd(struct lpfc_hba *phba, struct lpfc_iocbq *iocbq) 1831 { 1832 if (phba->sli_rev == LPFC_SLI_REV4) 1833 return bf_get(wqe_cmnd, &iocbq->wqe.generic.wqe_com); 1834 else 1835 return iocbq->iocb.ulpCommand; 1836 } 1837 1838 static inline 1839 u16 get_job_ulpcontext(struct lpfc_hba *phba, struct lpfc_iocbq *iocbq) 1840 { 1841 if (phba->sli_rev == LPFC_SLI_REV4) 1842 return bf_get(wqe_ctxt_tag, &iocbq->wqe.generic.wqe_com); 1843 else 1844 return iocbq->iocb.ulpContext; 1845 } 1846 1847 static inline 1848 u16 get_job_rcvoxid(struct lpfc_hba *phba, struct lpfc_iocbq *iocbq) 1849 { 1850 if (phba->sli_rev == LPFC_SLI_REV4) 1851 return bf_get(wqe_rcvoxid, &iocbq->wqe.generic.wqe_com); 1852 else 1853 return iocbq->iocb.unsli3.rcvsli3.ox_id; 1854 } 1855 1856 static inline 1857 u32 get_job_data_placed(struct lpfc_hba *phba, struct lpfc_iocbq *iocbq) 1858 { 1859 if (phba->sli_rev == LPFC_SLI_REV4) 1860 return iocbq->wcqe_cmpl.total_data_placed; 1861 else 1862 return iocbq->iocb.un.genreq64.bdl.bdeSize; 1863 } 1864 1865 static inline 1866 u32 get_job_abtsiotag(struct lpfc_hba *phba, struct lpfc_iocbq *iocbq) 1867 { 1868 if (phba->sli_rev == LPFC_SLI_REV4) 1869 return iocbq->wqe.abort_cmd.wqe_com.abort_tag; 1870 else 1871 return iocbq->iocb.un.acxri.abortIoTag; 1872 } 1873 1874 static inline 1875 u32 get_job_els_rsp64_did(struct lpfc_hba *phba, struct lpfc_iocbq *iocbq) 1876 { 1877 if (phba->sli_rev == LPFC_SLI_REV4) 1878 return bf_get(wqe_els_did, &iocbq->wqe.els_req.wqe_dest); 1879 else 1880 return iocbq->iocb.un.elsreq64.remoteID; 1881 } 1882