/* * Scsi Host Layer for MPT (Message Passing Technology) based controllers * * This code is based on drivers/scsi/mpt3sas/mpt3sas_scsih.c * Copyright (C) 2012-2014 LSI Corporation * Copyright (C) 2013-2014 Avago Technologies * (mailto: MPT-FusionLinux.pdl@avagotech.com) * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License * as published by the Free Software Foundation; either version 2 * of the License, or (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * NO WARRANTY * THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR * CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT * LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT, * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is * solely responsible for determining the appropriateness of using and * distributing the Program and assumes all risks associated with its * exercise of rights under this Agreement, including but not limited to * the risks and costs of program errors, damage to or loss of data, * programs or equipment, and unavailability or interruption of operations. * DISCLAIMER OF LIABILITY * NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR * TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE * USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED * HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, * USA. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "mpt3sas_base.h" #define RAID_CHANNEL 1 #define PCIE_CHANNEL 2 /* forward proto's */ static void _scsih_expander_node_remove(struct MPT3SAS_ADAPTER *ioc, struct _sas_node *sas_expander); static void _firmware_event_work(struct work_struct *work); static void _scsih_remove_device(struct MPT3SAS_ADAPTER *ioc, struct _sas_device *sas_device); static int _scsih_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle, u8 retry_count, u8 is_pd); static int _scsih_pcie_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle); static void _scsih_pcie_device_remove_from_sml(struct MPT3SAS_ADAPTER *ioc, struct _pcie_device *pcie_device); static void _scsih_pcie_check_device(struct MPT3SAS_ADAPTER *ioc, u16 handle); static u8 _scsih_check_for_pending_tm(struct MPT3SAS_ADAPTER *ioc, u16 smid); static void _scsih_complete_devices_scanning(struct MPT3SAS_ADAPTER *ioc); /* global parameters */ LIST_HEAD(mpt3sas_ioc_list); /* global ioc lock for list operations */ DEFINE_SPINLOCK(gioc_lock); MODULE_AUTHOR(MPT3SAS_AUTHOR); MODULE_DESCRIPTION(MPT3SAS_DESCRIPTION); MODULE_LICENSE("GPL"); MODULE_VERSION(MPT3SAS_DRIVER_VERSION); MODULE_ALIAS("mpt2sas"); /* local parameters */ static u8 scsi_io_cb_idx = -1; static u8 tm_cb_idx = -1; static u8 ctl_cb_idx = -1; static u8 base_cb_idx = -1; static u8 port_enable_cb_idx = -1; static u8 transport_cb_idx = -1; static u8 scsih_cb_idx = -1; static u8 config_cb_idx = -1; static int mpt2_ids; static int mpt3_ids; static u8 tm_tr_cb_idx = -1 ; static u8 tm_tr_volume_cb_idx = -1 ; static u8 tm_sas_control_cb_idx = -1; /* command line options */ static u32 logging_level; MODULE_PARM_DESC(logging_level, " bits for enabling additional logging info (default=0)"); static ushort max_sectors = 0xFFFF; module_param(max_sectors, ushort, 0444); MODULE_PARM_DESC(max_sectors, "max sectors, range 64 to 32767 default=32767"); static int missing_delay[2] = {-1, -1}; module_param_array(missing_delay, int, NULL, 0444); MODULE_PARM_DESC(missing_delay, " device missing delay , io missing delay"); /* scsi-mid layer global parmeter is max_report_luns, which is 511 */ #define MPT3SAS_MAX_LUN (16895) static u64 max_lun = MPT3SAS_MAX_LUN; module_param(max_lun, ullong, 0444); MODULE_PARM_DESC(max_lun, " max lun, default=16895 "); static ushort hbas_to_enumerate; module_param(hbas_to_enumerate, ushort, 0444); MODULE_PARM_DESC(hbas_to_enumerate, " 0 - enumerates both SAS 2.0 & SAS 3.0 generation HBAs\n \ 1 - enumerates only SAS 2.0 generation HBAs\n \ 2 - enumerates only SAS 3.0 generation HBAs (default=0)"); /* diag_buffer_enable is bitwise * bit 0 set = TRACE * bit 1 set = SNAPSHOT * bit 2 set = EXTENDED * * Either bit can be set, or both */ static int diag_buffer_enable = -1; module_param(diag_buffer_enable, int, 0444); MODULE_PARM_DESC(diag_buffer_enable, " post diag buffers (TRACE=1/SNAPSHOT=2/EXTENDED=4/default=0)"); static int disable_discovery = -1; module_param(disable_discovery, int, 0444); MODULE_PARM_DESC(disable_discovery, " disable discovery "); /* permit overriding the host protection capabilities mask (EEDP/T10 PI) */ static int prot_mask = -1; module_param(prot_mask, int, 0444); MODULE_PARM_DESC(prot_mask, " host protection capabilities mask, def=7 "); static bool enable_sdev_max_qd; module_param(enable_sdev_max_qd, bool, 0444); MODULE_PARM_DESC(enable_sdev_max_qd, "Enable sdev max qd as can_queue, def=disabled(0)"); static int multipath_on_hba = -1; module_param(multipath_on_hba, int, 0); MODULE_PARM_DESC(multipath_on_hba, "Multipath support to add same target device\n\t\t" "as many times as it is visible to HBA from various paths\n\t\t" "(by default:\n\t\t" "\t SAS 2.0 & SAS 3.0 HBA - This will be disabled,\n\t\t" "\t SAS 3.5 HBA - This will be enabled)"); static int host_tagset_enable = 1; module_param(host_tagset_enable, int, 0444); MODULE_PARM_DESC(host_tagset_enable, "Shared host tagset enable/disable Default: enable(1)"); /* raid transport support */ static struct raid_template *mpt3sas_raid_template; static struct raid_template *mpt2sas_raid_template; /** * struct sense_info - common structure for obtaining sense keys * @skey: sense key * @asc: additional sense code * @ascq: additional sense code qualifier */ struct sense_info { u8 skey; u8 asc; u8 ascq; }; #define MPT3SAS_PROCESS_TRIGGER_DIAG (0xFFFB) #define MPT3SAS_TURN_ON_PFA_LED (0xFFFC) #define MPT3SAS_PORT_ENABLE_COMPLETE (0xFFFD) #define MPT3SAS_ABRT_TASK_SET (0xFFFE) #define MPT3SAS_REMOVE_UNRESPONDING_DEVICES (0xFFFF) /** * struct fw_event_work - firmware event struct * @list: link list framework * @work: work object (ioc->fault_reset_work_q) * @ioc: per adapter object * @device_handle: device handle * @VF_ID: virtual function id * @VP_ID: virtual port id * @ignore: flag meaning this event has been marked to ignore * @event: firmware event MPI2_EVENT_XXX defined in mpi2_ioc.h * @refcount: kref for this event * @event_data: reply event data payload follows * * This object stored on ioc->fw_event_list. */ struct fw_event_work { struct list_head list; struct work_struct work; struct MPT3SAS_ADAPTER *ioc; u16 device_handle; u8 VF_ID; u8 VP_ID; u8 ignore; u16 event; struct kref refcount; char event_data[] __aligned(4); }; static void fw_event_work_free(struct kref *r) { kfree(container_of(r, struct fw_event_work, refcount)); } static void fw_event_work_get(struct fw_event_work *fw_work) { kref_get(&fw_work->refcount); } static void fw_event_work_put(struct fw_event_work *fw_work) { kref_put(&fw_work->refcount, fw_event_work_free); } static struct fw_event_work *alloc_fw_event_work(int len) { struct fw_event_work *fw_event; fw_event = kzalloc(sizeof(*fw_event) + len, GFP_ATOMIC); if (!fw_event) return NULL; kref_init(&fw_event->refcount); return fw_event; } /** * struct _scsi_io_transfer - scsi io transfer * @handle: sas device handle (assigned by firmware) * @is_raid: flag set for hidden raid components * @dir: DMA_TO_DEVICE, DMA_FROM_DEVICE, * @data_length: data transfer length * @data_dma: dma pointer to data * @sense: sense data * @lun: lun number * @cdb_length: cdb length * @cdb: cdb contents * @timeout: timeout for this command * @VF_ID: virtual function id * @VP_ID: virtual port id * @valid_reply: flag set for reply message * @sense_length: sense length * @ioc_status: ioc status * @scsi_state: scsi state * @scsi_status: scsi staus * @log_info: log information * @transfer_length: data length transfer when there is a reply message * * Used for sending internal scsi commands to devices within this module. * Refer to _scsi_send_scsi_io(). */ struct _scsi_io_transfer { u16 handle; u8 is_raid; enum dma_data_direction dir; u32 data_length; dma_addr_t data_dma; u8 sense[SCSI_SENSE_BUFFERSIZE]; u32 lun; u8 cdb_length; u8 cdb[32]; u8 timeout; u8 VF_ID; u8 VP_ID; u8 valid_reply; /* the following bits are only valid when 'valid_reply = 1' */ u32 sense_length; u16 ioc_status; u8 scsi_state; u8 scsi_status; u32 log_info; u32 transfer_length; }; /** * _scsih_set_debug_level - global setting of ioc->logging_level. * @val: ? * @kp: ? * * Note: The logging levels are defined in mpt3sas_debug.h. */ static int _scsih_set_debug_level(const char *val, const struct kernel_param *kp) { int ret = param_set_int(val, kp); struct MPT3SAS_ADAPTER *ioc; if (ret) return ret; pr_info("setting logging_level(0x%08x)\n", logging_level); spin_lock(&gioc_lock); list_for_each_entry(ioc, &mpt3sas_ioc_list, list) ioc->logging_level = logging_level; spin_unlock(&gioc_lock); return 0; } module_param_call(logging_level, _scsih_set_debug_level, param_get_int, &logging_level, 0644); /** * _scsih_srch_boot_sas_address - search based on sas_address * @sas_address: sas address * @boot_device: boot device object from bios page 2 * * Return: 1 when there's a match, 0 means no match. */ static inline int _scsih_srch_boot_sas_address(u64 sas_address, Mpi2BootDeviceSasWwid_t *boot_device) { return (sas_address == le64_to_cpu(boot_device->SASAddress)) ? 1 : 0; } /** * _scsih_srch_boot_device_name - search based on device name * @device_name: device name specified in INDENTIFY fram * @boot_device: boot device object from bios page 2 * * Return: 1 when there's a match, 0 means no match. */ static inline int _scsih_srch_boot_device_name(u64 device_name, Mpi2BootDeviceDeviceName_t *boot_device) { return (device_name == le64_to_cpu(boot_device->DeviceName)) ? 1 : 0; } /** * _scsih_srch_boot_encl_slot - search based on enclosure_logical_id/slot * @enclosure_logical_id: enclosure logical id * @slot_number: slot number * @boot_device: boot device object from bios page 2 * * Return: 1 when there's a match, 0 means no match. */ static inline int _scsih_srch_boot_encl_slot(u64 enclosure_logical_id, u16 slot_number, Mpi2BootDeviceEnclosureSlot_t *boot_device) { return (enclosure_logical_id == le64_to_cpu(boot_device-> EnclosureLogicalID) && slot_number == le16_to_cpu(boot_device-> SlotNumber)) ? 1 : 0; } /** * mpt3sas_get_port_by_id - get hba port entry corresponding to provided * port number from port list * @ioc: per adapter object * @port_id: port number * @bypass_dirty_port_flag: when set look the matching hba port entry even * if hba port entry is marked as dirty. * * Search for hba port entry corresponding to provided port number, * if available return port object otherwise return NULL. */ struct hba_port * mpt3sas_get_port_by_id(struct MPT3SAS_ADAPTER *ioc, u8 port_id, u8 bypass_dirty_port_flag) { struct hba_port *port, *port_next; /* * When multipath_on_hba is disabled then * search the hba_port entry using default * port id i.e. 255 */ if (!ioc->multipath_on_hba) port_id = MULTIPATH_DISABLED_PORT_ID; list_for_each_entry_safe(port, port_next, &ioc->port_table_list, list) { if (port->port_id != port_id) continue; if (bypass_dirty_port_flag) return port; if (port->flags & HBA_PORT_FLAG_DIRTY_PORT) continue; return port; } /* * Allocate hba_port object for default port id (i.e. 255) * when multipath_on_hba is disabled for the HBA. * And add this object to port_table_list. */ if (!ioc->multipath_on_hba) { port = kzalloc(sizeof(struct hba_port), GFP_ATOMIC); if (!port) return NULL; port->port_id = port_id; ioc_info(ioc, "hba_port entry: %p, port: %d is added to hba_port list\n", port, port->port_id); list_add_tail(&port->list, &ioc->port_table_list); return port; } return NULL; } /** * mpt3sas_get_vphy_by_phy - get virtual_phy object corresponding to phy number * @ioc: per adapter object * @port: hba_port object * @phy: phy number * * Return virtual_phy object corresponding to phy number. */ struct virtual_phy * mpt3sas_get_vphy_by_phy(struct MPT3SAS_ADAPTER *ioc, struct hba_port *port, u32 phy) { struct virtual_phy *vphy, *vphy_next; if (!port->vphys_mask) return NULL; list_for_each_entry_safe(vphy, vphy_next, &port->vphys_list, list) { if (vphy->phy_mask & (1 << phy)) return vphy; } return NULL; } /** * _scsih_is_boot_device - search for matching boot device. * @sas_address: sas address * @device_name: device name specified in INDENTIFY fram * @enclosure_logical_id: enclosure logical id * @slot: slot number * @form: specifies boot device form * @boot_device: boot device object from bios page 2 * * Return: 1 when there's a match, 0 means no match. */ static int _scsih_is_boot_device(u64 sas_address, u64 device_name, u64 enclosure_logical_id, u16 slot, u8 form, Mpi2BiosPage2BootDevice_t *boot_device) { int rc = 0; switch (form) { case MPI2_BIOSPAGE2_FORM_SAS_WWID: if (!sas_address) break; rc = _scsih_srch_boot_sas_address( sas_address, &boot_device->SasWwid); break; case MPI2_BIOSPAGE2_FORM_ENCLOSURE_SLOT: if (!enclosure_logical_id) break; rc = _scsih_srch_boot_encl_slot( enclosure_logical_id, slot, &boot_device->EnclosureSlot); break; case MPI2_BIOSPAGE2_FORM_DEVICE_NAME: if (!device_name) break; rc = _scsih_srch_boot_device_name( device_name, &boot_device->DeviceName); break; case MPI2_BIOSPAGE2_FORM_NO_DEVICE_SPECIFIED: break; } return rc; } /** * _scsih_get_sas_address - set the sas_address for given device handle * @ioc: ? * @handle: device handle * @sas_address: sas address * * Return: 0 success, non-zero when failure */ static int _scsih_get_sas_address(struct MPT3SAS_ADAPTER *ioc, u16 handle, u64 *sas_address) { Mpi2SasDevicePage0_t sas_device_pg0; Mpi2ConfigReply_t mpi_reply; u32 ioc_status; *sas_address = 0; if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return -ENXIO; } ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status == MPI2_IOCSTATUS_SUCCESS) { /* For HBA, vSES doesn't return HBA SAS address. Instead return * vSES's sas address. */ if ((handle <= ioc->sas_hba.num_phys) && (!(le32_to_cpu(sas_device_pg0.DeviceInfo) & MPI2_SAS_DEVICE_INFO_SEP))) *sas_address = ioc->sas_hba.sas_address; else *sas_address = le64_to_cpu(sas_device_pg0.SASAddress); return 0; } /* we hit this because the given parent handle doesn't exist */ if (ioc_status == MPI2_IOCSTATUS_CONFIG_INVALID_PAGE) return -ENXIO; /* else error case */ ioc_err(ioc, "handle(0x%04x), ioc_status(0x%04x), failure at %s:%d/%s()!\n", handle, ioc_status, __FILE__, __LINE__, __func__); return -EIO; } /** * _scsih_determine_boot_device - determine boot device. * @ioc: per adapter object * @device: sas_device or pcie_device object * @channel: SAS or PCIe channel * * Determines whether this device should be first reported device to * to scsi-ml or sas transport, this purpose is for persistent boot device. * There are primary, alternate, and current entries in bios page 2. The order * priority is primary, alternate, then current. This routine saves * the corresponding device object. * The saved data to be used later in _scsih_probe_boot_devices(). */ static void _scsih_determine_boot_device(struct MPT3SAS_ADAPTER *ioc, void *device, u32 channel) { struct _sas_device *sas_device; struct _pcie_device *pcie_device; struct _raid_device *raid_device; u64 sas_address; u64 device_name; u64 enclosure_logical_id; u16 slot; /* only process this function when driver loads */ if (!ioc->is_driver_loading) return; /* no Bios, return immediately */ if (!ioc->bios_pg3.BiosVersion) return; if (channel == RAID_CHANNEL) { raid_device = device; sas_address = raid_device->wwid; device_name = 0; enclosure_logical_id = 0; slot = 0; } else if (channel == PCIE_CHANNEL) { pcie_device = device; sas_address = pcie_device->wwid; device_name = 0; enclosure_logical_id = 0; slot = 0; } else { sas_device = device; sas_address = sas_device->sas_address; device_name = sas_device->device_name; enclosure_logical_id = sas_device->enclosure_logical_id; slot = sas_device->slot; } if (!ioc->req_boot_device.device) { if (_scsih_is_boot_device(sas_address, device_name, enclosure_logical_id, slot, (ioc->bios_pg2.ReqBootDeviceForm & MPI2_BIOSPAGE2_FORM_MASK), &ioc->bios_pg2.RequestedBootDevice)) { dinitprintk(ioc, ioc_info(ioc, "%s: req_boot_device(0x%016llx)\n", __func__, (u64)sas_address)); ioc->req_boot_device.device = device; ioc->req_boot_device.channel = channel; } } if (!ioc->req_alt_boot_device.device) { if (_scsih_is_boot_device(sas_address, device_name, enclosure_logical_id, slot, (ioc->bios_pg2.ReqAltBootDeviceForm & MPI2_BIOSPAGE2_FORM_MASK), &ioc->bios_pg2.RequestedAltBootDevice)) { dinitprintk(ioc, ioc_info(ioc, "%s: req_alt_boot_device(0x%016llx)\n", __func__, (u64)sas_address)); ioc->req_alt_boot_device.device = device; ioc->req_alt_boot_device.channel = channel; } } if (!ioc->current_boot_device.device) { if (_scsih_is_boot_device(sas_address, device_name, enclosure_logical_id, slot, (ioc->bios_pg2.CurrentBootDeviceForm & MPI2_BIOSPAGE2_FORM_MASK), &ioc->bios_pg2.CurrentBootDevice)) { dinitprintk(ioc, ioc_info(ioc, "%s: current_boot_device(0x%016llx)\n", __func__, (u64)sas_address)); ioc->current_boot_device.device = device; ioc->current_boot_device.channel = channel; } } } static struct _sas_device * __mpt3sas_get_sdev_from_target(struct MPT3SAS_ADAPTER *ioc, struct MPT3SAS_TARGET *tgt_priv) { struct _sas_device *ret; assert_spin_locked(&ioc->sas_device_lock); ret = tgt_priv->sas_dev; if (ret) sas_device_get(ret); return ret; } static struct _sas_device * mpt3sas_get_sdev_from_target(struct MPT3SAS_ADAPTER *ioc, struct MPT3SAS_TARGET *tgt_priv) { struct _sas_device *ret; unsigned long flags; spin_lock_irqsave(&ioc->sas_device_lock, flags); ret = __mpt3sas_get_sdev_from_target(ioc, tgt_priv); spin_unlock_irqrestore(&ioc->sas_device_lock, flags); return ret; } static struct _pcie_device * __mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER *ioc, struct MPT3SAS_TARGET *tgt_priv) { struct _pcie_device *ret; assert_spin_locked(&ioc->pcie_device_lock); ret = tgt_priv->pcie_dev; if (ret) pcie_device_get(ret); return ret; } /** * mpt3sas_get_pdev_from_target - pcie device search * @ioc: per adapter object * @tgt_priv: starget private object * * Context: This function will acquire ioc->pcie_device_lock and will release * before returning the pcie_device object. * * This searches for pcie_device from target, then return pcie_device object. */ static struct _pcie_device * mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER *ioc, struct MPT3SAS_TARGET *tgt_priv) { struct _pcie_device *ret; unsigned long flags; spin_lock_irqsave(&ioc->pcie_device_lock, flags); ret = __mpt3sas_get_pdev_from_target(ioc, tgt_priv); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); return ret; } /** * __mpt3sas_get_sdev_by_rphy - sas device search * @ioc: per adapter object * @rphy: sas_rphy pointer * * Context: This function will acquire ioc->sas_device_lock and will release * before returning the sas_device object. * * This searches for sas_device from rphy object * then return sas_device object. */ struct _sas_device * __mpt3sas_get_sdev_by_rphy(struct MPT3SAS_ADAPTER *ioc, struct sas_rphy *rphy) { struct _sas_device *sas_device; assert_spin_locked(&ioc->sas_device_lock); list_for_each_entry(sas_device, &ioc->sas_device_list, list) { if (sas_device->rphy != rphy) continue; sas_device_get(sas_device); return sas_device; } sas_device = NULL; list_for_each_entry(sas_device, &ioc->sas_device_init_list, list) { if (sas_device->rphy != rphy) continue; sas_device_get(sas_device); return sas_device; } return NULL; } /** * __mpt3sas_get_sdev_by_addr - get _sas_device object corresponding to provided * sas address from sas_device_list list * @ioc: per adapter object * @sas_address: device sas address * @port: port number * * Search for _sas_device object corresponding to provided sas address, * if available return _sas_device object address otherwise return NULL. */ struct _sas_device * __mpt3sas_get_sdev_by_addr(struct MPT3SAS_ADAPTER *ioc, u64 sas_address, struct hba_port *port) { struct _sas_device *sas_device; if (!port) return NULL; assert_spin_locked(&ioc->sas_device_lock); list_for_each_entry(sas_device, &ioc->sas_device_list, list) { if (sas_device->sas_address != sas_address) continue; if (sas_device->port != port) continue; sas_device_get(sas_device); return sas_device; } list_for_each_entry(sas_device, &ioc->sas_device_init_list, list) { if (sas_device->sas_address != sas_address) continue; if (sas_device->port != port) continue; sas_device_get(sas_device); return sas_device; } return NULL; } /** * mpt3sas_get_sdev_by_addr - sas device search * @ioc: per adapter object * @sas_address: sas address * @port: hba port entry * Context: Calling function should acquire ioc->sas_device_lock * * This searches for sas_device based on sas_address & port number, * then return sas_device object. */ struct _sas_device * mpt3sas_get_sdev_by_addr(struct MPT3SAS_ADAPTER *ioc, u64 sas_address, struct hba_port *port) { struct _sas_device *sas_device; unsigned long flags; spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_by_addr(ioc, sas_address, port); spin_unlock_irqrestore(&ioc->sas_device_lock, flags); return sas_device; } static struct _sas_device * __mpt3sas_get_sdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct _sas_device *sas_device; assert_spin_locked(&ioc->sas_device_lock); list_for_each_entry(sas_device, &ioc->sas_device_list, list) if (sas_device->handle == handle) goto found_device; list_for_each_entry(sas_device, &ioc->sas_device_init_list, list) if (sas_device->handle == handle) goto found_device; return NULL; found_device: sas_device_get(sas_device); return sas_device; } /** * mpt3sas_get_sdev_by_handle - sas device search * @ioc: per adapter object * @handle: sas device handle (assigned by firmware) * Context: Calling function should acquire ioc->sas_device_lock * * This searches for sas_device based on sas_address, then return sas_device * object. */ struct _sas_device * mpt3sas_get_sdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct _sas_device *sas_device; unsigned long flags; spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); spin_unlock_irqrestore(&ioc->sas_device_lock, flags); return sas_device; } /** * _scsih_display_enclosure_chassis_info - display device location info * @ioc: per adapter object * @sas_device: per sas device object * @sdev: scsi device struct * @starget: scsi target struct */ static void _scsih_display_enclosure_chassis_info(struct MPT3SAS_ADAPTER *ioc, struct _sas_device *sas_device, struct scsi_device *sdev, struct scsi_target *starget) { if (sdev) { if (sas_device->enclosure_handle != 0) sdev_printk(KERN_INFO, sdev, "enclosure logical id (0x%016llx), slot(%d) \n", (unsigned long long) sas_device->enclosure_logical_id, sas_device->slot); if (sas_device->connector_name[0] != '\0') sdev_printk(KERN_INFO, sdev, "enclosure level(0x%04x), connector name( %s)\n", sas_device->enclosure_level, sas_device->connector_name); if (sas_device->is_chassis_slot_valid) sdev_printk(KERN_INFO, sdev, "chassis slot(0x%04x)\n", sas_device->chassis_slot); } else if (starget) { if (sas_device->enclosure_handle != 0) starget_printk(KERN_INFO, starget, "enclosure logical id(0x%016llx), slot(%d) \n", (unsigned long long) sas_device->enclosure_logical_id, sas_device->slot); if (sas_device->connector_name[0] != '\0') starget_printk(KERN_INFO, starget, "enclosure level(0x%04x), connector name( %s)\n", sas_device->enclosure_level, sas_device->connector_name); if (sas_device->is_chassis_slot_valid) starget_printk(KERN_INFO, starget, "chassis slot(0x%04x)\n", sas_device->chassis_slot); } else { if (sas_device->enclosure_handle != 0) ioc_info(ioc, "enclosure logical id(0x%016llx), slot(%d)\n", (u64)sas_device->enclosure_logical_id, sas_device->slot); if (sas_device->connector_name[0] != '\0') ioc_info(ioc, "enclosure level(0x%04x), connector name( %s)\n", sas_device->enclosure_level, sas_device->connector_name); if (sas_device->is_chassis_slot_valid) ioc_info(ioc, "chassis slot(0x%04x)\n", sas_device->chassis_slot); } } /** * _scsih_sas_device_remove - remove sas_device from list. * @ioc: per adapter object * @sas_device: the sas_device object * Context: This function will acquire ioc->sas_device_lock. * * If sas_device is on the list, remove it and decrement its reference count. */ static void _scsih_sas_device_remove(struct MPT3SAS_ADAPTER *ioc, struct _sas_device *sas_device) { unsigned long flags; if (!sas_device) return; ioc_info(ioc, "removing handle(0x%04x), sas_addr(0x%016llx)\n", sas_device->handle, (u64)sas_device->sas_address); _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL); /* * The lock serializes access to the list, but we still need to verify * that nobody removed the entry while we were waiting on the lock. */ spin_lock_irqsave(&ioc->sas_device_lock, flags); if (!list_empty(&sas_device->list)) { list_del_init(&sas_device->list); sas_device_put(sas_device); } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); } /** * _scsih_device_remove_by_handle - removing device object by handle * @ioc: per adapter object * @handle: device handle */ static void _scsih_device_remove_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct _sas_device *sas_device; unsigned long flags; if (ioc->shost_recovery) return; spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); if (sas_device) { list_del_init(&sas_device->list); sas_device_put(sas_device); } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); if (sas_device) { _scsih_remove_device(ioc, sas_device); sas_device_put(sas_device); } } /** * mpt3sas_device_remove_by_sas_address - removing device object by * sas address & port number * @ioc: per adapter object * @sas_address: device sas_address * @port: hba port entry * * Return nothing. */ void mpt3sas_device_remove_by_sas_address(struct MPT3SAS_ADAPTER *ioc, u64 sas_address, struct hba_port *port) { struct _sas_device *sas_device; unsigned long flags; if (ioc->shost_recovery) return; spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_by_addr(ioc, sas_address, port); if (sas_device) { list_del_init(&sas_device->list); sas_device_put(sas_device); } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); if (sas_device) { _scsih_remove_device(ioc, sas_device); sas_device_put(sas_device); } } /** * _scsih_sas_device_add - insert sas_device to the list. * @ioc: per adapter object * @sas_device: the sas_device object * Context: This function will acquire ioc->sas_device_lock. * * Adding new object to the ioc->sas_device_list. */ static void _scsih_sas_device_add(struct MPT3SAS_ADAPTER *ioc, struct _sas_device *sas_device) { unsigned long flags; dewtprintk(ioc, ioc_info(ioc, "%s: handle(0x%04x), sas_addr(0x%016llx)\n", __func__, sas_device->handle, (u64)sas_device->sas_address)); dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL)); spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device_get(sas_device); list_add_tail(&sas_device->list, &ioc->sas_device_list); spin_unlock_irqrestore(&ioc->sas_device_lock, flags); if (ioc->hide_drives) { clear_bit(sas_device->handle, ioc->pend_os_device_add); return; } if (!mpt3sas_transport_port_add(ioc, sas_device->handle, sas_device->sas_address_parent, sas_device->port)) { _scsih_sas_device_remove(ioc, sas_device); } else if (!sas_device->starget) { /* * When asyn scanning is enabled, its not possible to remove * devices while scanning is turned on due to an oops in * scsi_sysfs_add_sdev()->add_device()->sysfs_addrm_start() */ if (!ioc->is_driver_loading) { mpt3sas_transport_port_remove(ioc, sas_device->sas_address, sas_device->sas_address_parent, sas_device->port); _scsih_sas_device_remove(ioc, sas_device); } } else clear_bit(sas_device->handle, ioc->pend_os_device_add); } /** * _scsih_sas_device_init_add - insert sas_device to the list. * @ioc: per adapter object * @sas_device: the sas_device object * Context: This function will acquire ioc->sas_device_lock. * * Adding new object at driver load time to the ioc->sas_device_init_list. */ static void _scsih_sas_device_init_add(struct MPT3SAS_ADAPTER *ioc, struct _sas_device *sas_device) { unsigned long flags; dewtprintk(ioc, ioc_info(ioc, "%s: handle(0x%04x), sas_addr(0x%016llx)\n", __func__, sas_device->handle, (u64)sas_device->sas_address)); dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL)); spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device_get(sas_device); list_add_tail(&sas_device->list, &ioc->sas_device_init_list); _scsih_determine_boot_device(ioc, sas_device, 0); spin_unlock_irqrestore(&ioc->sas_device_lock, flags); } static struct _pcie_device * __mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid) { struct _pcie_device *pcie_device; assert_spin_locked(&ioc->pcie_device_lock); list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) if (pcie_device->wwid == wwid) goto found_device; list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list) if (pcie_device->wwid == wwid) goto found_device; return NULL; found_device: pcie_device_get(pcie_device); return pcie_device; } /** * mpt3sas_get_pdev_by_wwid - pcie device search * @ioc: per adapter object * @wwid: wwid * * Context: This function will acquire ioc->pcie_device_lock and will release * before returning the pcie_device object. * * This searches for pcie_device based on wwid, then return pcie_device object. */ static struct _pcie_device * mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid) { struct _pcie_device *pcie_device; unsigned long flags; spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); return pcie_device; } static struct _pcie_device * __mpt3sas_get_pdev_by_idchannel(struct MPT3SAS_ADAPTER *ioc, int id, int channel) { struct _pcie_device *pcie_device; assert_spin_locked(&ioc->pcie_device_lock); list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) if (pcie_device->id == id && pcie_device->channel == channel) goto found_device; list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list) if (pcie_device->id == id && pcie_device->channel == channel) goto found_device; return NULL; found_device: pcie_device_get(pcie_device); return pcie_device; } static struct _pcie_device * __mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct _pcie_device *pcie_device; assert_spin_locked(&ioc->pcie_device_lock); list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) if (pcie_device->handle == handle) goto found_device; list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list) if (pcie_device->handle == handle) goto found_device; return NULL; found_device: pcie_device_get(pcie_device); return pcie_device; } /** * mpt3sas_get_pdev_by_handle - pcie device search * @ioc: per adapter object * @handle: Firmware device handle * * Context: This function will acquire ioc->pcie_device_lock and will release * before returning the pcie_device object. * * This searches for pcie_device based on handle, then return pcie_device * object. */ struct _pcie_device * mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct _pcie_device *pcie_device; unsigned long flags; spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); return pcie_device; } /** * _scsih_set_nvme_max_shutdown_latency - Update max_shutdown_latency. * @ioc: per adapter object * Context: This function will acquire ioc->pcie_device_lock * * Update ioc->max_shutdown_latency to that NVMe drives RTD3 Entry Latency * which has reported maximum among all available NVMe drives. * Minimum max_shutdown_latency will be six seconds. */ static void _scsih_set_nvme_max_shutdown_latency(struct MPT3SAS_ADAPTER *ioc) { struct _pcie_device *pcie_device; unsigned long flags; u16 shutdown_latency = IO_UNIT_CONTROL_SHUTDOWN_TIMEOUT; spin_lock_irqsave(&ioc->pcie_device_lock, flags); list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) { if (pcie_device->shutdown_latency) { if (shutdown_latency < pcie_device->shutdown_latency) shutdown_latency = pcie_device->shutdown_latency; } } ioc->max_shutdown_latency = shutdown_latency; spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); } /** * _scsih_pcie_device_remove - remove pcie_device from list. * @ioc: per adapter object * @pcie_device: the pcie_device object * Context: This function will acquire ioc->pcie_device_lock. * * If pcie_device is on the list, remove it and decrement its reference count. */ static void _scsih_pcie_device_remove(struct MPT3SAS_ADAPTER *ioc, struct _pcie_device *pcie_device) { unsigned long flags; int was_on_pcie_device_list = 0; u8 update_latency = 0; if (!pcie_device) return; ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n", pcie_device->handle, (u64)pcie_device->wwid); if (pcie_device->enclosure_handle != 0) ioc_info(ioc, "removing enclosure logical id(0x%016llx), slot(%d)\n", (u64)pcie_device->enclosure_logical_id, pcie_device->slot); if (pcie_device->connector_name[0] != '\0') ioc_info(ioc, "removing enclosure level(0x%04x), connector name( %s)\n", pcie_device->enclosure_level, pcie_device->connector_name); spin_lock_irqsave(&ioc->pcie_device_lock, flags); if (!list_empty(&pcie_device->list)) { list_del_init(&pcie_device->list); was_on_pcie_device_list = 1; } if (pcie_device->shutdown_latency == ioc->max_shutdown_latency) update_latency = 1; spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); if (was_on_pcie_device_list) { kfree(pcie_device->serial_number); pcie_device_put(pcie_device); } /* * This device's RTD3 Entry Latency matches IOC's * max_shutdown_latency. Recalculate IOC's max_shutdown_latency * from the available drives as current drive is getting removed. */ if (update_latency) _scsih_set_nvme_max_shutdown_latency(ioc); } /** * _scsih_pcie_device_remove_by_handle - removing pcie device object by handle * @ioc: per adapter object * @handle: device handle */ static void _scsih_pcie_device_remove_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct _pcie_device *pcie_device; unsigned long flags; int was_on_pcie_device_list = 0; u8 update_latency = 0; if (ioc->shost_recovery) return; spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle); if (pcie_device) { if (!list_empty(&pcie_device->list)) { list_del_init(&pcie_device->list); was_on_pcie_device_list = 1; pcie_device_put(pcie_device); } if (pcie_device->shutdown_latency == ioc->max_shutdown_latency) update_latency = 1; } spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); if (was_on_pcie_device_list) { _scsih_pcie_device_remove_from_sml(ioc, pcie_device); pcie_device_put(pcie_device); } /* * This device's RTD3 Entry Latency matches IOC's * max_shutdown_latency. Recalculate IOC's max_shutdown_latency * from the available drives as current drive is getting removed. */ if (update_latency) _scsih_set_nvme_max_shutdown_latency(ioc); } /** * _scsih_pcie_device_add - add pcie_device object * @ioc: per adapter object * @pcie_device: pcie_device object * * This is added to the pcie_device_list link list. */ static void _scsih_pcie_device_add(struct MPT3SAS_ADAPTER *ioc, struct _pcie_device *pcie_device) { unsigned long flags; dewtprintk(ioc, ioc_info(ioc, "%s: handle (0x%04x), wwid(0x%016llx)\n", __func__, pcie_device->handle, (u64)pcie_device->wwid)); if (pcie_device->enclosure_handle != 0) dewtprintk(ioc, ioc_info(ioc, "%s: enclosure logical id(0x%016llx), slot( %d)\n", __func__, (u64)pcie_device->enclosure_logical_id, pcie_device->slot)); if (pcie_device->connector_name[0] != '\0') dewtprintk(ioc, ioc_info(ioc, "%s: enclosure level(0x%04x), connector name( %s)\n", __func__, pcie_device->enclosure_level, pcie_device->connector_name)); spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device_get(pcie_device); list_add_tail(&pcie_device->list, &ioc->pcie_device_list); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); if (pcie_device->access_status == MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) { clear_bit(pcie_device->handle, ioc->pend_os_device_add); return; } if (scsi_add_device(ioc->shost, PCIE_CHANNEL, pcie_device->id, 0)) { _scsih_pcie_device_remove(ioc, pcie_device); } else if (!pcie_device->starget) { if (!ioc->is_driver_loading) { /*TODO-- Need to find out whether this condition will occur or not*/ clear_bit(pcie_device->handle, ioc->pend_os_device_add); } } else clear_bit(pcie_device->handle, ioc->pend_os_device_add); } /* * _scsih_pcie_device_init_add - insert pcie_device to the init list. * @ioc: per adapter object * @pcie_device: the pcie_device object * Context: This function will acquire ioc->pcie_device_lock. * * Adding new object at driver load time to the ioc->pcie_device_init_list. */ static void _scsih_pcie_device_init_add(struct MPT3SAS_ADAPTER *ioc, struct _pcie_device *pcie_device) { unsigned long flags; dewtprintk(ioc, ioc_info(ioc, "%s: handle (0x%04x), wwid(0x%016llx)\n", __func__, pcie_device->handle, (u64)pcie_device->wwid)); if (pcie_device->enclosure_handle != 0) dewtprintk(ioc, ioc_info(ioc, "%s: enclosure logical id(0x%016llx), slot( %d)\n", __func__, (u64)pcie_device->enclosure_logical_id, pcie_device->slot)); if (pcie_device->connector_name[0] != '\0') dewtprintk(ioc, ioc_info(ioc, "%s: enclosure level(0x%04x), connector name( %s)\n", __func__, pcie_device->enclosure_level, pcie_device->connector_name)); spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device_get(pcie_device); list_add_tail(&pcie_device->list, &ioc->pcie_device_init_list); if (pcie_device->access_status != MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) _scsih_determine_boot_device(ioc, pcie_device, PCIE_CHANNEL); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); } /** * _scsih_raid_device_find_by_id - raid device search * @ioc: per adapter object * @id: sas device target id * @channel: sas device channel * Context: Calling function should acquire ioc->raid_device_lock * * This searches for raid_device based on target id, then return raid_device * object. */ static struct _raid_device * _scsih_raid_device_find_by_id(struct MPT3SAS_ADAPTER *ioc, int id, int channel) { struct _raid_device *raid_device, *r; r = NULL; list_for_each_entry(raid_device, &ioc->raid_device_list, list) { if (raid_device->id == id && raid_device->channel == channel) { r = raid_device; goto out; } } out: return r; } /** * mpt3sas_raid_device_find_by_handle - raid device search * @ioc: per adapter object * @handle: sas device handle (assigned by firmware) * Context: Calling function should acquire ioc->raid_device_lock * * This searches for raid_device based on handle, then return raid_device * object. */ struct _raid_device * mpt3sas_raid_device_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct _raid_device *raid_device, *r; r = NULL; list_for_each_entry(raid_device, &ioc->raid_device_list, list) { if (raid_device->handle != handle) continue; r = raid_device; goto out; } out: return r; } /** * _scsih_raid_device_find_by_wwid - raid device search * @ioc: per adapter object * @wwid: ? * Context: Calling function should acquire ioc->raid_device_lock * * This searches for raid_device based on wwid, then return raid_device * object. */ static struct _raid_device * _scsih_raid_device_find_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid) { struct _raid_device *raid_device, *r; r = NULL; list_for_each_entry(raid_device, &ioc->raid_device_list, list) { if (raid_device->wwid != wwid) continue; r = raid_device; goto out; } out: return r; } /** * _scsih_raid_device_add - add raid_device object * @ioc: per adapter object * @raid_device: raid_device object * * This is added to the raid_device_list link list. */ static void _scsih_raid_device_add(struct MPT3SAS_ADAPTER *ioc, struct _raid_device *raid_device) { unsigned long flags; dewtprintk(ioc, ioc_info(ioc, "%s: handle(0x%04x), wwid(0x%016llx)\n", __func__, raid_device->handle, (u64)raid_device->wwid)); spin_lock_irqsave(&ioc->raid_device_lock, flags); list_add_tail(&raid_device->list, &ioc->raid_device_list); spin_unlock_irqrestore(&ioc->raid_device_lock, flags); } /** * _scsih_raid_device_remove - delete raid_device object * @ioc: per adapter object * @raid_device: raid_device object * */ static void _scsih_raid_device_remove(struct MPT3SAS_ADAPTER *ioc, struct _raid_device *raid_device) { unsigned long flags; spin_lock_irqsave(&ioc->raid_device_lock, flags); list_del(&raid_device->list); kfree(raid_device); spin_unlock_irqrestore(&ioc->raid_device_lock, flags); } /** * mpt3sas_scsih_expander_find_by_handle - expander device search * @ioc: per adapter object * @handle: expander handle (assigned by firmware) * Context: Calling function should acquire ioc->sas_device_lock * * This searches for expander device based on handle, then returns the * sas_node object. */ struct _sas_node * mpt3sas_scsih_expander_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct _sas_node *sas_expander, *r; r = NULL; list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) { if (sas_expander->handle != handle) continue; r = sas_expander; goto out; } out: return r; } /** * mpt3sas_scsih_enclosure_find_by_handle - exclosure device search * @ioc: per adapter object * @handle: enclosure handle (assigned by firmware) * Context: Calling function should acquire ioc->sas_device_lock * * This searches for enclosure device based on handle, then returns the * enclosure object. */ static struct _enclosure_node * mpt3sas_scsih_enclosure_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct _enclosure_node *enclosure_dev, *r; r = NULL; list_for_each_entry(enclosure_dev, &ioc->enclosure_list, list) { if (le16_to_cpu(enclosure_dev->pg0.EnclosureHandle) != handle) continue; r = enclosure_dev; goto out; } out: return r; } /** * mpt3sas_scsih_expander_find_by_sas_address - expander device search * @ioc: per adapter object * @sas_address: sas address * @port: hba port entry * Context: Calling function should acquire ioc->sas_node_lock. * * This searches for expander device based on sas_address & port number, * then returns the sas_node object. */ struct _sas_node * mpt3sas_scsih_expander_find_by_sas_address(struct MPT3SAS_ADAPTER *ioc, u64 sas_address, struct hba_port *port) { struct _sas_node *sas_expander, *r = NULL; if (!port) return r; list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) { if (sas_expander->sas_address != sas_address) continue; if (sas_expander->port != port) continue; r = sas_expander; goto out; } out: return r; } /** * _scsih_expander_node_add - insert expander device to the list. * @ioc: per adapter object * @sas_expander: the sas_device object * Context: This function will acquire ioc->sas_node_lock. * * Adding new object to the ioc->sas_expander_list. */ static void _scsih_expander_node_add(struct MPT3SAS_ADAPTER *ioc, struct _sas_node *sas_expander) { unsigned long flags; spin_lock_irqsave(&ioc->sas_node_lock, flags); list_add_tail(&sas_expander->list, &ioc->sas_expander_list); spin_unlock_irqrestore(&ioc->sas_node_lock, flags); } /** * _scsih_is_end_device - determines if device is an end device * @device_info: bitfield providing information about the device. * Context: none * * Return: 1 if end device. */ static int _scsih_is_end_device(u32 device_info) { if (device_info & MPI2_SAS_DEVICE_INFO_END_DEVICE && ((device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) | (device_info & MPI2_SAS_DEVICE_INFO_STP_TARGET) | (device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE))) return 1; else return 0; } /** * _scsih_is_nvme_pciescsi_device - determines if * device is an pcie nvme/scsi device * @device_info: bitfield providing information about the device. * Context: none * * Returns 1 if device is pcie device type nvme/scsi. */ static int _scsih_is_nvme_pciescsi_device(u32 device_info) { if (((device_info & MPI26_PCIE_DEVINFO_MASK_DEVICE_TYPE) == MPI26_PCIE_DEVINFO_NVME) || ((device_info & MPI26_PCIE_DEVINFO_MASK_DEVICE_TYPE) == MPI26_PCIE_DEVINFO_SCSI)) return 1; else return 0; } /** * _scsih_scsi_lookup_find_by_target - search for matching channel:id * @ioc: per adapter object * @id: target id * @channel: channel * Context: This function will acquire ioc->scsi_lookup_lock. * * This will search for a matching channel:id in the scsi_lookup array, * returning 1 if found. */ static u8 _scsih_scsi_lookup_find_by_target(struct MPT3SAS_ADAPTER *ioc, int id, int channel) { int smid; struct scsi_cmnd *scmd; for (smid = 1; smid <= ioc->shost->can_queue; smid++) { scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); if (!scmd) continue; if (scmd->device->id == id && scmd->device->channel == channel) return 1; } return 0; } /** * _scsih_scsi_lookup_find_by_lun - search for matching channel:id:lun * @ioc: per adapter object * @id: target id * @lun: lun number * @channel: channel * Context: This function will acquire ioc->scsi_lookup_lock. * * This will search for a matching channel:id:lun in the scsi_lookup array, * returning 1 if found. */ static u8 _scsih_scsi_lookup_find_by_lun(struct MPT3SAS_ADAPTER *ioc, int id, unsigned int lun, int channel) { int smid; struct scsi_cmnd *scmd; for (smid = 1; smid <= ioc->shost->can_queue; smid++) { scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); if (!scmd) continue; if (scmd->device->id == id && scmd->device->channel == channel && scmd->device->lun == lun) return 1; } return 0; } /** * mpt3sas_scsih_scsi_lookup_get - returns scmd entry * @ioc: per adapter object * @smid: system request message index * * Return: the smid stored scmd pointer. * Then will dereference the stored scmd pointer. */ struct scsi_cmnd * mpt3sas_scsih_scsi_lookup_get(struct MPT3SAS_ADAPTER *ioc, u16 smid) { struct scsi_cmnd *scmd = NULL; struct scsiio_tracker *st; Mpi25SCSIIORequest_t *mpi_request; u16 tag = smid - 1; if (smid > 0 && smid <= ioc->scsiio_depth - INTERNAL_SCSIIO_CMDS_COUNT) { u32 unique_tag = ioc->io_queue_num[tag] << BLK_MQ_UNIQUE_TAG_BITS | tag; mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); /* * If SCSI IO request is outstanding at driver level then * DevHandle filed must be non-zero. If DevHandle is zero * then it means that this smid is free at driver level, * so return NULL. */ if (!mpi_request->DevHandle) return scmd; scmd = scsi_host_find_tag(ioc->shost, unique_tag); if (scmd) { st = scsi_cmd_priv(scmd); if (st->cb_idx == 0xFF || st->smid == 0) scmd = NULL; } } return scmd; } /** * scsih_change_queue_depth - setting device queue depth * @sdev: scsi device struct * @qdepth: requested queue depth * * Return: queue depth. */ static int scsih_change_queue_depth(struct scsi_device *sdev, int qdepth) { struct Scsi_Host *shost = sdev->host; int max_depth; struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); struct MPT3SAS_DEVICE *sas_device_priv_data; struct MPT3SAS_TARGET *sas_target_priv_data; struct _sas_device *sas_device; unsigned long flags; max_depth = shost->can_queue; /* * limit max device queue for SATA to 32 if enable_sdev_max_qd * is disabled. */ if (ioc->enable_sdev_max_qd) goto not_sata; sas_device_priv_data = sdev->hostdata; if (!sas_device_priv_data) goto not_sata; sas_target_priv_data = sas_device_priv_data->sas_target; if (!sas_target_priv_data) goto not_sata; if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) goto not_sata; spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_from_target(ioc, sas_target_priv_data); if (sas_device) { if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE) max_depth = MPT3SAS_SATA_QUEUE_DEPTH; sas_device_put(sas_device); } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); not_sata: if (!sdev->tagged_supported) max_depth = 1; if (qdepth > max_depth) qdepth = max_depth; scsi_change_queue_depth(sdev, qdepth); sdev_printk(KERN_INFO, sdev, "qdepth(%d), tagged(%d), scsi_level(%d), cmd_que(%d)\n", sdev->queue_depth, sdev->tagged_supported, sdev->scsi_level, ((sdev->inquiry[7] & 2) >> 1)); return sdev->queue_depth; } /** * mpt3sas_scsih_change_queue_depth - setting device queue depth * @sdev: scsi device struct * @qdepth: requested queue depth * * Returns nothing. */ void mpt3sas_scsih_change_queue_depth(struct scsi_device *sdev, int qdepth) { struct Scsi_Host *shost = sdev->host; struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); if (ioc->enable_sdev_max_qd) qdepth = shost->can_queue; scsih_change_queue_depth(sdev, qdepth); } /** * scsih_target_alloc - target add routine * @starget: scsi target struct * * Return: 0 if ok. Any other return is assumed to be an error and * the device is ignored. */ static int scsih_target_alloc(struct scsi_target *starget) { struct Scsi_Host *shost = dev_to_shost(&starget->dev); struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); struct MPT3SAS_TARGET *sas_target_priv_data; struct _sas_device *sas_device; struct _raid_device *raid_device; struct _pcie_device *pcie_device; unsigned long flags; struct sas_rphy *rphy; sas_target_priv_data = kzalloc(sizeof(*sas_target_priv_data), GFP_KERNEL); if (!sas_target_priv_data) return -ENOMEM; starget->hostdata = sas_target_priv_data; sas_target_priv_data->starget = starget; sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE; /* RAID volumes */ if (starget->channel == RAID_CHANNEL) { spin_lock_irqsave(&ioc->raid_device_lock, flags); raid_device = _scsih_raid_device_find_by_id(ioc, starget->id, starget->channel); if (raid_device) { sas_target_priv_data->handle = raid_device->handle; sas_target_priv_data->sas_address = raid_device->wwid; sas_target_priv_data->flags |= MPT_TARGET_FLAGS_VOLUME; if (ioc->is_warpdrive) sas_target_priv_data->raid_device = raid_device; raid_device->starget = starget; } spin_unlock_irqrestore(&ioc->raid_device_lock, flags); return 0; } /* PCIe devices */ if (starget->channel == PCIE_CHANNEL) { spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device = __mpt3sas_get_pdev_by_idchannel(ioc, starget->id, starget->channel); if (pcie_device) { sas_target_priv_data->handle = pcie_device->handle; sas_target_priv_data->sas_address = pcie_device->wwid; sas_target_priv_data->port = NULL; sas_target_priv_data->pcie_dev = pcie_device; pcie_device->starget = starget; pcie_device->id = starget->id; pcie_device->channel = starget->channel; sas_target_priv_data->flags |= MPT_TARGET_FLAGS_PCIE_DEVICE; if (pcie_device->fast_path) sas_target_priv_data->flags |= MPT_TARGET_FASTPATH_IO; } spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); return 0; } /* sas/sata devices */ spin_lock_irqsave(&ioc->sas_device_lock, flags); rphy = dev_to_rphy(starget->dev.parent); sas_device = __mpt3sas_get_sdev_by_rphy(ioc, rphy); if (sas_device) { sas_target_priv_data->handle = sas_device->handle; sas_target_priv_data->sas_address = sas_device->sas_address; sas_target_priv_data->port = sas_device->port; sas_target_priv_data->sas_dev = sas_device; sas_device->starget = starget; sas_device->id = starget->id; sas_device->channel = starget->channel; if (test_bit(sas_device->handle, ioc->pd_handles)) sas_target_priv_data->flags |= MPT_TARGET_FLAGS_RAID_COMPONENT; if (sas_device->fast_path) sas_target_priv_data->flags |= MPT_TARGET_FASTPATH_IO; } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); return 0; } /** * scsih_target_destroy - target destroy routine * @starget: scsi target struct */ static void scsih_target_destroy(struct scsi_target *starget) { struct Scsi_Host *shost = dev_to_shost(&starget->dev); struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); struct MPT3SAS_TARGET *sas_target_priv_data; struct _sas_device *sas_device; struct _raid_device *raid_device; struct _pcie_device *pcie_device; unsigned long flags; sas_target_priv_data = starget->hostdata; if (!sas_target_priv_data) return; if (starget->channel == RAID_CHANNEL) { spin_lock_irqsave(&ioc->raid_device_lock, flags); raid_device = _scsih_raid_device_find_by_id(ioc, starget->id, starget->channel); if (raid_device) { raid_device->starget = NULL; raid_device->sdev = NULL; } spin_unlock_irqrestore(&ioc->raid_device_lock, flags); goto out; } if (starget->channel == PCIE_CHANNEL) { spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device = __mpt3sas_get_pdev_from_target(ioc, sas_target_priv_data); if (pcie_device && (pcie_device->starget == starget) && (pcie_device->id == starget->id) && (pcie_device->channel == starget->channel)) pcie_device->starget = NULL; if (pcie_device) { /* * Corresponding get() is in _scsih_target_alloc() */ sas_target_priv_data->pcie_dev = NULL; pcie_device_put(pcie_device); pcie_device_put(pcie_device); } spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); goto out; } spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_from_target(ioc, sas_target_priv_data); if (sas_device && (sas_device->starget == starget) && (sas_device->id == starget->id) && (sas_device->channel == starget->channel)) sas_device->starget = NULL; if (sas_device) { /* * Corresponding get() is in _scsih_target_alloc() */ sas_target_priv_data->sas_dev = NULL; sas_device_put(sas_device); sas_device_put(sas_device); } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); out: kfree(sas_target_priv_data); starget->hostdata = NULL; } /** * scsih_slave_alloc - device add routine * @sdev: scsi device struct * * Return: 0 if ok. Any other return is assumed to be an error and * the device is ignored. */ static int scsih_slave_alloc(struct scsi_device *sdev) { struct Scsi_Host *shost; struct MPT3SAS_ADAPTER *ioc; struct MPT3SAS_TARGET *sas_target_priv_data; struct MPT3SAS_DEVICE *sas_device_priv_data; struct scsi_target *starget; struct _raid_device *raid_device; struct _sas_device *sas_device; struct _pcie_device *pcie_device; unsigned long flags; sas_device_priv_data = kzalloc(sizeof(*sas_device_priv_data), GFP_KERNEL); if (!sas_device_priv_data) return -ENOMEM; sas_device_priv_data->lun = sdev->lun; sas_device_priv_data->flags = MPT_DEVICE_FLAGS_INIT; starget = scsi_target(sdev); sas_target_priv_data = starget->hostdata; sas_target_priv_data->num_luns++; sas_device_priv_data->sas_target = sas_target_priv_data; sdev->hostdata = sas_device_priv_data; if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT)) sdev->no_uld_attach = 1; shost = dev_to_shost(&starget->dev); ioc = shost_priv(shost); if (starget->channel == RAID_CHANNEL) { spin_lock_irqsave(&ioc->raid_device_lock, flags); raid_device = _scsih_raid_device_find_by_id(ioc, starget->id, starget->channel); if (raid_device) raid_device->sdev = sdev; /* raid is single lun */ spin_unlock_irqrestore(&ioc->raid_device_lock, flags); } if (starget->channel == PCIE_CHANNEL) { spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, sas_target_priv_data->sas_address); if (pcie_device && (pcie_device->starget == NULL)) { sdev_printk(KERN_INFO, sdev, "%s : pcie_device->starget set to starget @ %d\n", __func__, __LINE__); pcie_device->starget = starget; } if (pcie_device) pcie_device_put(pcie_device); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); } else if (!(sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) { spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_by_addr(ioc, sas_target_priv_data->sas_address, sas_target_priv_data->port); if (sas_device && (sas_device->starget == NULL)) { sdev_printk(KERN_INFO, sdev, "%s : sas_device->starget set to starget @ %d\n", __func__, __LINE__); sas_device->starget = starget; } if (sas_device) sas_device_put(sas_device); spin_unlock_irqrestore(&ioc->sas_device_lock, flags); } return 0; } /** * scsih_slave_destroy - device destroy routine * @sdev: scsi device struct */ static void scsih_slave_destroy(struct scsi_device *sdev) { struct MPT3SAS_TARGET *sas_target_priv_data; struct scsi_target *starget; struct Scsi_Host *shost; struct MPT3SAS_ADAPTER *ioc; struct _sas_device *sas_device; struct _pcie_device *pcie_device; unsigned long flags; if (!sdev->hostdata) return; starget = scsi_target(sdev); sas_target_priv_data = starget->hostdata; sas_target_priv_data->num_luns--; shost = dev_to_shost(&starget->dev); ioc = shost_priv(shost); if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) { spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device = __mpt3sas_get_pdev_from_target(ioc, sas_target_priv_data); if (pcie_device && !sas_target_priv_data->num_luns) pcie_device->starget = NULL; if (pcie_device) pcie_device_put(pcie_device); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); } else if (!(sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) { spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_from_target(ioc, sas_target_priv_data); if (sas_device && !sas_target_priv_data->num_luns) sas_device->starget = NULL; if (sas_device) sas_device_put(sas_device); spin_unlock_irqrestore(&ioc->sas_device_lock, flags); } kfree(sdev->hostdata); sdev->hostdata = NULL; } /** * _scsih_display_sata_capabilities - sata capabilities * @ioc: per adapter object * @handle: device handle * @sdev: scsi device struct */ static void _scsih_display_sata_capabilities(struct MPT3SAS_ADAPTER *ioc, u16 handle, struct scsi_device *sdev) { Mpi2ConfigReply_t mpi_reply; Mpi2SasDevicePage0_t sas_device_pg0; u32 ioc_status; u16 flags; u32 device_info; if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } flags = le16_to_cpu(sas_device_pg0.Flags); device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); sdev_printk(KERN_INFO, sdev, "atapi(%s), ncq(%s), asyn_notify(%s), smart(%s), fua(%s), " "sw_preserve(%s)\n", (device_info & MPI2_SAS_DEVICE_INFO_ATAPI_DEVICE) ? "y" : "n", (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_NCQ_SUPPORTED) ? "y" : "n", (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_ASYNCHRONOUS_NOTIFY) ? "y" : "n", (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_SMART_SUPPORTED) ? "y" : "n", (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_FUA_SUPPORTED) ? "y" : "n", (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_SW_PRESERVE) ? "y" : "n"); } /* * raid transport support - * Enabled for SLES11 and newer, in older kernels the driver will panic when * unloading the driver followed by a load - I believe that the subroutine * raid_class_release() is not cleaning up properly. */ /** * scsih_is_raid - return boolean indicating device is raid volume * @dev: the device struct object */ static int scsih_is_raid(struct device *dev) { struct scsi_device *sdev = to_scsi_device(dev); struct MPT3SAS_ADAPTER *ioc = shost_priv(sdev->host); if (ioc->is_warpdrive) return 0; return (sdev->channel == RAID_CHANNEL) ? 1 : 0; } static int scsih_is_nvme(struct device *dev) { struct scsi_device *sdev = to_scsi_device(dev); return (sdev->channel == PCIE_CHANNEL) ? 1 : 0; } /** * scsih_get_resync - get raid volume resync percent complete * @dev: the device struct object */ static void scsih_get_resync(struct device *dev) { struct scsi_device *sdev = to_scsi_device(dev); struct MPT3SAS_ADAPTER *ioc = shost_priv(sdev->host); static struct _raid_device *raid_device; unsigned long flags; Mpi2RaidVolPage0_t vol_pg0; Mpi2ConfigReply_t mpi_reply; u32 volume_status_flags; u8 percent_complete; u16 handle; percent_complete = 0; handle = 0; if (ioc->is_warpdrive) goto out; spin_lock_irqsave(&ioc->raid_device_lock, flags); raid_device = _scsih_raid_device_find_by_id(ioc, sdev->id, sdev->channel); if (raid_device) { handle = raid_device->handle; percent_complete = raid_device->percent_complete; } spin_unlock_irqrestore(&ioc->raid_device_lock, flags); if (!handle) goto out; if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, &vol_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle, sizeof(Mpi2RaidVolPage0_t))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); percent_complete = 0; goto out; } volume_status_flags = le32_to_cpu(vol_pg0.VolumeStatusFlags); if (!(volume_status_flags & MPI2_RAIDVOL0_STATUS_FLAG_RESYNC_IN_PROGRESS)) percent_complete = 0; out: switch (ioc->hba_mpi_version_belonged) { case MPI2_VERSION: raid_set_resync(mpt2sas_raid_template, dev, percent_complete); break; case MPI25_VERSION: case MPI26_VERSION: raid_set_resync(mpt3sas_raid_template, dev, percent_complete); break; } } /** * scsih_get_state - get raid volume level * @dev: the device struct object */ static void scsih_get_state(struct device *dev) { struct scsi_device *sdev = to_scsi_device(dev); struct MPT3SAS_ADAPTER *ioc = shost_priv(sdev->host); static struct _raid_device *raid_device; unsigned long flags; Mpi2RaidVolPage0_t vol_pg0; Mpi2ConfigReply_t mpi_reply; u32 volstate; enum raid_state state = RAID_STATE_UNKNOWN; u16 handle = 0; spin_lock_irqsave(&ioc->raid_device_lock, flags); raid_device = _scsih_raid_device_find_by_id(ioc, sdev->id, sdev->channel); if (raid_device) handle = raid_device->handle; spin_unlock_irqrestore(&ioc->raid_device_lock, flags); if (!raid_device) goto out; if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, &vol_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle, sizeof(Mpi2RaidVolPage0_t))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } volstate = le32_to_cpu(vol_pg0.VolumeStatusFlags); if (volstate & MPI2_RAIDVOL0_STATUS_FLAG_RESYNC_IN_PROGRESS) { state = RAID_STATE_RESYNCING; goto out; } switch (vol_pg0.VolumeState) { case MPI2_RAID_VOL_STATE_OPTIMAL: case MPI2_RAID_VOL_STATE_ONLINE: state = RAID_STATE_ACTIVE; break; case MPI2_RAID_VOL_STATE_DEGRADED: state = RAID_STATE_DEGRADED; break; case MPI2_RAID_VOL_STATE_FAILED: case MPI2_RAID_VOL_STATE_MISSING: state = RAID_STATE_OFFLINE; break; } out: switch (ioc->hba_mpi_version_belonged) { case MPI2_VERSION: raid_set_state(mpt2sas_raid_template, dev, state); break; case MPI25_VERSION: case MPI26_VERSION: raid_set_state(mpt3sas_raid_template, dev, state); break; } } /** * _scsih_set_level - set raid level * @ioc: ? * @sdev: scsi device struct * @volume_type: volume type */ static void _scsih_set_level(struct MPT3SAS_ADAPTER *ioc, struct scsi_device *sdev, u8 volume_type) { enum raid_level level = RAID_LEVEL_UNKNOWN; switch (volume_type) { case MPI2_RAID_VOL_TYPE_RAID0: level = RAID_LEVEL_0; break; case MPI2_RAID_VOL_TYPE_RAID10: level = RAID_LEVEL_10; break; case MPI2_RAID_VOL_TYPE_RAID1E: level = RAID_LEVEL_1E; break; case MPI2_RAID_VOL_TYPE_RAID1: level = RAID_LEVEL_1; break; } switch (ioc->hba_mpi_version_belonged) { case MPI2_VERSION: raid_set_level(mpt2sas_raid_template, &sdev->sdev_gendev, level); break; case MPI25_VERSION: case MPI26_VERSION: raid_set_level(mpt3sas_raid_template, &sdev->sdev_gendev, level); break; } } /** * _scsih_get_volume_capabilities - volume capabilities * @ioc: per adapter object * @raid_device: the raid_device object * * Return: 0 for success, else 1 */ static int _scsih_get_volume_capabilities(struct MPT3SAS_ADAPTER *ioc, struct _raid_device *raid_device) { Mpi2RaidVolPage0_t *vol_pg0; Mpi2RaidPhysDiskPage0_t pd_pg0; Mpi2SasDevicePage0_t sas_device_pg0; Mpi2ConfigReply_t mpi_reply; u16 sz; u8 num_pds; if ((mpt3sas_config_get_number_pds(ioc, raid_device->handle, &num_pds)) || !num_pds) { dfailprintk(ioc, ioc_warn(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__)); return 1; } raid_device->num_pds = num_pds; sz = offsetof(Mpi2RaidVolPage0_t, PhysDisk) + (num_pds * sizeof(Mpi2RaidVol0PhysDisk_t)); vol_pg0 = kzalloc(sz, GFP_KERNEL); if (!vol_pg0) { dfailprintk(ioc, ioc_warn(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__)); return 1; } if ((mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, vol_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, raid_device->handle, sz))) { dfailprintk(ioc, ioc_warn(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__)); kfree(vol_pg0); return 1; } raid_device->volume_type = vol_pg0->VolumeType; /* figure out what the underlying devices are by * obtaining the device_info bits for the 1st device */ if (!(mpt3sas_config_get_phys_disk_pg0(ioc, &mpi_reply, &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_PHYSDISKNUM, vol_pg0->PhysDisk[0].PhysDiskNum))) { if (!(mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, le16_to_cpu(pd_pg0.DevHandle)))) { raid_device->device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); } } kfree(vol_pg0); return 0; } /** * _scsih_enable_tlr - setting TLR flags * @ioc: per adapter object * @sdev: scsi device struct * * Enabling Transaction Layer Retries for tape devices when * vpd page 0x90 is present * */ static void _scsih_enable_tlr(struct MPT3SAS_ADAPTER *ioc, struct scsi_device *sdev) { /* only for TAPE */ if (sdev->type != TYPE_TAPE) return; if (!(ioc->facts.IOCCapabilities & MPI2_IOCFACTS_CAPABILITY_TLR)) return; sas_enable_tlr(sdev); sdev_printk(KERN_INFO, sdev, "TLR %s\n", sas_is_tlr_enabled(sdev) ? "Enabled" : "Disabled"); return; } /** * scsih_slave_configure - device configure routine. * @sdev: scsi device struct * * Return: 0 if ok. Any other return is assumed to be an error and * the device is ignored. */ static int scsih_slave_configure(struct scsi_device *sdev) { struct Scsi_Host *shost = sdev->host; struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); struct MPT3SAS_DEVICE *sas_device_priv_data; struct MPT3SAS_TARGET *sas_target_priv_data; struct _sas_device *sas_device; struct _pcie_device *pcie_device; struct _raid_device *raid_device; unsigned long flags; int qdepth; u8 ssp_target = 0; char *ds = ""; char *r_level = ""; u16 handle, volume_handle = 0; u64 volume_wwid = 0; qdepth = 1; sas_device_priv_data = sdev->hostdata; sas_device_priv_data->configured_lun = 1; sas_device_priv_data->flags &= ~MPT_DEVICE_FLAGS_INIT; sas_target_priv_data = sas_device_priv_data->sas_target; handle = sas_target_priv_data->handle; /* raid volume handling */ if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME) { spin_lock_irqsave(&ioc->raid_device_lock, flags); raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); spin_unlock_irqrestore(&ioc->raid_device_lock, flags); if (!raid_device) { dfailprintk(ioc, ioc_warn(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__)); return 1; } if (_scsih_get_volume_capabilities(ioc, raid_device)) { dfailprintk(ioc, ioc_warn(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__)); return 1; } /* * WARPDRIVE: Initialize the required data for Direct IO */ mpt3sas_init_warpdrive_properties(ioc, raid_device); /* RAID Queue Depth Support * IS volume = underlying qdepth of drive type, either * MPT3SAS_SAS_QUEUE_DEPTH or MPT3SAS_SATA_QUEUE_DEPTH * IM/IME/R10 = 128 (MPT3SAS_RAID_QUEUE_DEPTH) */ if (raid_device->device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) { qdepth = MPT3SAS_SAS_QUEUE_DEPTH; ds = "SSP"; } else { qdepth = MPT3SAS_SATA_QUEUE_DEPTH; if (raid_device->device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE) ds = "SATA"; else ds = "STP"; } switch (raid_device->volume_type) { case MPI2_RAID_VOL_TYPE_RAID0: r_level = "RAID0"; break; case MPI2_RAID_VOL_TYPE_RAID1E: qdepth = MPT3SAS_RAID_QUEUE_DEPTH; if (ioc->manu_pg10.OEMIdentifier && (le32_to_cpu(ioc->manu_pg10.GenericFlags0) & MFG10_GF0_R10_DISPLAY) && !(raid_device->num_pds % 2)) r_level = "RAID10"; else r_level = "RAID1E"; break; case MPI2_RAID_VOL_TYPE_RAID1: qdepth = MPT3SAS_RAID_QUEUE_DEPTH; r_level = "RAID1"; break; case MPI2_RAID_VOL_TYPE_RAID10: qdepth = MPT3SAS_RAID_QUEUE_DEPTH; r_level = "RAID10"; break; case MPI2_RAID_VOL_TYPE_UNKNOWN: default: qdepth = MPT3SAS_RAID_QUEUE_DEPTH; r_level = "RAIDX"; break; } if (!ioc->hide_ir_msg) sdev_printk(KERN_INFO, sdev, "%s: handle(0x%04x), wwid(0x%016llx)," " pd_count(%d), type(%s)\n", r_level, raid_device->handle, (unsigned long long)raid_device->wwid, raid_device->num_pds, ds); if (shost->max_sectors > MPT3SAS_RAID_MAX_SECTORS) { blk_queue_max_hw_sectors(sdev->request_queue, MPT3SAS_RAID_MAX_SECTORS); sdev_printk(KERN_INFO, sdev, "Set queue's max_sector to: %u\n", MPT3SAS_RAID_MAX_SECTORS); } mpt3sas_scsih_change_queue_depth(sdev, qdepth); /* raid transport support */ if (!ioc->is_warpdrive) _scsih_set_level(ioc, sdev, raid_device->volume_type); return 0; } /* non-raid handling */ if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) { if (mpt3sas_config_get_volume_handle(ioc, handle, &volume_handle)) { dfailprintk(ioc, ioc_warn(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__)); return 1; } if (volume_handle && mpt3sas_config_get_volume_wwid(ioc, volume_handle, &volume_wwid)) { dfailprintk(ioc, ioc_warn(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__)); return 1; } } /* PCIe handling */ if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) { spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, sas_device_priv_data->sas_target->sas_address); if (!pcie_device) { spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); dfailprintk(ioc, ioc_warn(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__)); return 1; } qdepth = MPT3SAS_NVME_QUEUE_DEPTH; ds = "NVMe"; sdev_printk(KERN_INFO, sdev, "%s: handle(0x%04x), wwid(0x%016llx), port(%d)\n", ds, handle, (unsigned long long)pcie_device->wwid, pcie_device->port_num); if (pcie_device->enclosure_handle != 0) sdev_printk(KERN_INFO, sdev, "%s: enclosure logical id(0x%016llx), slot(%d)\n", ds, (unsigned long long)pcie_device->enclosure_logical_id, pcie_device->slot); if (pcie_device->connector_name[0] != '\0') sdev_printk(KERN_INFO, sdev, "%s: enclosure level(0x%04x)," "connector name( %s)\n", ds, pcie_device->enclosure_level, pcie_device->connector_name); if (pcie_device->nvme_mdts) blk_queue_max_hw_sectors(sdev->request_queue, pcie_device->nvme_mdts/512); pcie_device_put(pcie_device); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); mpt3sas_scsih_change_queue_depth(sdev, qdepth); /* Enable QUEUE_FLAG_NOMERGES flag, so that IOs won't be ** merged and can eliminate holes created during merging ** operation. **/ blk_queue_flag_set(QUEUE_FLAG_NOMERGES, sdev->request_queue); blk_queue_virt_boundary(sdev->request_queue, ioc->page_size - 1); return 0; } spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_by_addr(ioc, sas_device_priv_data->sas_target->sas_address, sas_device_priv_data->sas_target->port); if (!sas_device) { spin_unlock_irqrestore(&ioc->sas_device_lock, flags); dfailprintk(ioc, ioc_warn(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__)); return 1; } sas_device->volume_handle = volume_handle; sas_device->volume_wwid = volume_wwid; if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) { qdepth = MPT3SAS_SAS_QUEUE_DEPTH; ssp_target = 1; if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_SEP) { sdev_printk(KERN_WARNING, sdev, "set ignore_delay_remove for handle(0x%04x)\n", sas_device_priv_data->sas_target->handle); sas_device_priv_data->ignore_delay_remove = 1; ds = "SES"; } else ds = "SSP"; } else { qdepth = MPT3SAS_SATA_QUEUE_DEPTH; if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_STP_TARGET) ds = "STP"; else if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE) ds = "SATA"; } sdev_printk(KERN_INFO, sdev, "%s: handle(0x%04x), " \ "sas_addr(0x%016llx), phy(%d), device_name(0x%016llx)\n", ds, handle, (unsigned long long)sas_device->sas_address, sas_device->phy, (unsigned long long)sas_device->device_name); _scsih_display_enclosure_chassis_info(NULL, sas_device, sdev, NULL); sas_device_put(sas_device); spin_unlock_irqrestore(&ioc->sas_device_lock, flags); if (!ssp_target) _scsih_display_sata_capabilities(ioc, handle, sdev); mpt3sas_scsih_change_queue_depth(sdev, qdepth); if (ssp_target) { sas_read_port_mode_page(sdev); _scsih_enable_tlr(ioc, sdev); } return 0; } /** * scsih_bios_param - fetch head, sector, cylinder info for a disk * @sdev: scsi device struct * @bdev: pointer to block device context * @capacity: device size (in 512 byte sectors) * @params: three element array to place output: * params[0] number of heads (max 255) * params[1] number of sectors (max 63) * params[2] number of cylinders */ static int scsih_bios_param(struct scsi_device *sdev, struct block_device *bdev, sector_t capacity, int params[]) { int heads; int sectors; sector_t cylinders; ulong dummy; heads = 64; sectors = 32; dummy = heads * sectors; cylinders = capacity; sector_div(cylinders, dummy); /* * Handle extended translation size for logical drives * > 1Gb */ if ((ulong)capacity >= 0x200000) { heads = 255; sectors = 63; dummy = heads * sectors; cylinders = capacity; sector_div(cylinders, dummy); } /* return result */ params[0] = heads; params[1] = sectors; params[2] = cylinders; return 0; } /** * _scsih_response_code - translation of device response code * @ioc: per adapter object * @response_code: response code returned by the device */ static void _scsih_response_code(struct MPT3SAS_ADAPTER *ioc, u8 response_code) { char *desc; switch (response_code) { case MPI2_SCSITASKMGMT_RSP_TM_COMPLETE: desc = "task management request completed"; break; case MPI2_SCSITASKMGMT_RSP_INVALID_FRAME: desc = "invalid frame"; break; case MPI2_SCSITASKMGMT_RSP_TM_NOT_SUPPORTED: desc = "task management request not supported"; break; case MPI2_SCSITASKMGMT_RSP_TM_FAILED: desc = "task management request failed"; break; case MPI2_SCSITASKMGMT_RSP_TM_SUCCEEDED: desc = "task management request succeeded"; break; case MPI2_SCSITASKMGMT_RSP_TM_INVALID_LUN: desc = "invalid lun"; break; case 0xA: desc = "overlapped tag attempted"; break; case MPI2_SCSITASKMGMT_RSP_IO_QUEUED_ON_IOC: desc = "task queued, however not sent to target"; break; default: desc = "unknown"; break; } ioc_warn(ioc, "response_code(0x%01x): %s\n", response_code, desc); } /** * _scsih_tm_done - tm completion routine * @ioc: per adapter object * @smid: system request message index * @msix_index: MSIX table index supplied by the OS * @reply: reply message frame(lower 32bit addr) * Context: none. * * The callback handler when using scsih_issue_tm. * * Return: 1 meaning mf should be freed from _base_interrupt * 0 means the mf is freed from this function. */ static u8 _scsih_tm_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply) { MPI2DefaultReply_t *mpi_reply; if (ioc->tm_cmds.status == MPT3_CMD_NOT_USED) return 1; if (ioc->tm_cmds.smid != smid) return 1; ioc->tm_cmds.status |= MPT3_CMD_COMPLETE; mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply); if (mpi_reply) { memcpy(ioc->tm_cmds.reply, mpi_reply, mpi_reply->MsgLength*4); ioc->tm_cmds.status |= MPT3_CMD_REPLY_VALID; } ioc->tm_cmds.status &= ~MPT3_CMD_PENDING; complete(&ioc->tm_cmds.done); return 1; } /** * mpt3sas_scsih_set_tm_flag - set per target tm_busy * @ioc: per adapter object * @handle: device handle * * During taskmangement request, we need to freeze the device queue. */ void mpt3sas_scsih_set_tm_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct MPT3SAS_DEVICE *sas_device_priv_data; struct scsi_device *sdev; u8 skip = 0; shost_for_each_device(sdev, ioc->shost) { if (skip) continue; sas_device_priv_data = sdev->hostdata; if (!sas_device_priv_data) continue; if (sas_device_priv_data->sas_target->handle == handle) { sas_device_priv_data->sas_target->tm_busy = 1; skip = 1; ioc->ignore_loginfos = 1; } } } /** * mpt3sas_scsih_clear_tm_flag - clear per target tm_busy * @ioc: per adapter object * @handle: device handle * * During taskmangement request, we need to freeze the device queue. */ void mpt3sas_scsih_clear_tm_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct MPT3SAS_DEVICE *sas_device_priv_data; struct scsi_device *sdev; u8 skip = 0; shost_for_each_device(sdev, ioc->shost) { if (skip) continue; sas_device_priv_data = sdev->hostdata; if (!sas_device_priv_data) continue; if (sas_device_priv_data->sas_target->handle == handle) { sas_device_priv_data->sas_target->tm_busy = 0; skip = 1; ioc->ignore_loginfos = 0; } } } /** * scsih_tm_cmd_map_status - map the target reset & LUN reset TM status * @ioc: per adapter object * @channel: the channel assigned by the OS * @id: the id assigned by the OS * @lun: lun number * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h) * @smid_task: smid assigned to the task * * Look whether TM has aborted the timed out SCSI command, if * TM has aborted the IO then return SUCCESS else return FAILED. */ static int scsih_tm_cmd_map_status(struct MPT3SAS_ADAPTER *ioc, uint channel, uint id, uint lun, u8 type, u16 smid_task) { if (smid_task <= ioc->shost->can_queue) { switch (type) { case MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET: if (!(_scsih_scsi_lookup_find_by_target(ioc, id, channel))) return SUCCESS; break; case MPI2_SCSITASKMGMT_TASKTYPE_ABRT_TASK_SET: case MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET: if (!(_scsih_scsi_lookup_find_by_lun(ioc, id, lun, channel))) return SUCCESS; break; default: return SUCCESS; } } else if (smid_task == ioc->scsih_cmds.smid) { if ((ioc->scsih_cmds.status & MPT3_CMD_COMPLETE) || (ioc->scsih_cmds.status & MPT3_CMD_NOT_USED)) return SUCCESS; } else if (smid_task == ioc->ctl_cmds.smid) { if ((ioc->ctl_cmds.status & MPT3_CMD_COMPLETE) || (ioc->ctl_cmds.status & MPT3_CMD_NOT_USED)) return SUCCESS; } return FAILED; } /** * scsih_tm_post_processing - post processing of target & LUN reset * @ioc: per adapter object * @handle: device handle * @channel: the channel assigned by the OS * @id: the id assigned by the OS * @lun: lun number * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h) * @smid_task: smid assigned to the task * * Post processing of target & LUN reset. Due to interrupt latency * issue it possible that interrupt for aborted IO might not be * received yet. So before returning failure status, poll the * reply descriptor pools for the reply of timed out SCSI command. * Return FAILED status if reply for timed out is not received * otherwise return SUCCESS. */ static int scsih_tm_post_processing(struct MPT3SAS_ADAPTER *ioc, u16 handle, uint channel, uint id, uint lun, u8 type, u16 smid_task) { int rc; rc = scsih_tm_cmd_map_status(ioc, channel, id, lun, type, smid_task); if (rc == SUCCESS) return rc; ioc_info(ioc, "Poll ReplyDescriptor queues for completion of" " smid(%d), task_type(0x%02x), handle(0x%04x)\n", smid_task, type, handle); /* * Due to interrupt latency issues, driver may receive interrupt for * TM first and then for aborted SCSI IO command. So, poll all the * ReplyDescriptor pools before returning the FAILED status to SML. */ mpt3sas_base_mask_interrupts(ioc); mpt3sas_base_sync_reply_irqs(ioc, 1); mpt3sas_base_unmask_interrupts(ioc); return scsih_tm_cmd_map_status(ioc, channel, id, lun, type, smid_task); } /** * mpt3sas_scsih_issue_tm - main routine for sending tm requests * @ioc: per adapter struct * @handle: device handle * @channel: the channel assigned by the OS * @id: the id assigned by the OS * @lun: lun number * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h) * @smid_task: smid assigned to the task * @msix_task: MSIX table index supplied by the OS * @timeout: timeout in seconds * @tr_method: Target Reset Method * Context: user * * A generic API for sending task management requests to firmware. * * The callback index is set inside `ioc->tm_cb_idx`. * The caller is responsible to check for outstanding commands. * * Return: SUCCESS or FAILED. */ int mpt3sas_scsih_issue_tm(struct MPT3SAS_ADAPTER *ioc, u16 handle, uint channel, uint id, u64 lun, u8 type, u16 smid_task, u16 msix_task, u8 timeout, u8 tr_method) { Mpi2SCSITaskManagementRequest_t *mpi_request; Mpi2SCSITaskManagementReply_t *mpi_reply; Mpi25SCSIIORequest_t *request; u16 smid = 0; u32 ioc_state; int rc; u8 issue_reset = 0; lockdep_assert_held(&ioc->tm_cmds.mutex); if (ioc->tm_cmds.status != MPT3_CMD_NOT_USED) { ioc_info(ioc, "%s: tm_cmd busy!!!\n", __func__); return FAILED; } if (ioc->shost_recovery || ioc->remove_host || ioc->pci_error_recovery) { ioc_info(ioc, "%s: host reset in progress!\n", __func__); return FAILED; } ioc_state = mpt3sas_base_get_iocstate(ioc, 0); if (ioc_state & MPI2_DOORBELL_USED) { dhsprintk(ioc, ioc_info(ioc, "unexpected doorbell active!\n")); rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); return (!rc) ? SUCCESS : FAILED; } if ((ioc_state & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_FAULT) { mpt3sas_print_fault_code(ioc, ioc_state & MPI2_DOORBELL_DATA_MASK); rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); return (!rc) ? SUCCESS : FAILED; } else if ((ioc_state & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_COREDUMP) { mpt3sas_print_coredump_info(ioc, ioc_state & MPI2_DOORBELL_DATA_MASK); rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); return (!rc) ? SUCCESS : FAILED; } smid = mpt3sas_base_get_smid_hpr(ioc, ioc->tm_cb_idx); if (!smid) { ioc_err(ioc, "%s: failed obtaining a smid\n", __func__); return FAILED; } dtmprintk(ioc, ioc_info(ioc, "sending tm: handle(0x%04x), task_type(0x%02x), smid(%d), timeout(%d), tr_method(0x%x)\n", handle, type, smid_task, timeout, tr_method)); ioc->tm_cmds.status = MPT3_CMD_PENDING; mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); ioc->tm_cmds.smid = smid; memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t)); memset(ioc->tm_cmds.reply, 0, sizeof(Mpi2SCSITaskManagementReply_t)); mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT; mpi_request->DevHandle = cpu_to_le16(handle); mpi_request->TaskType = type; if (type == MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK || type == MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK) mpi_request->MsgFlags = tr_method; mpi_request->TaskMID = cpu_to_le16(smid_task); int_to_scsilun(lun, (struct scsi_lun *)mpi_request->LUN); mpt3sas_scsih_set_tm_flag(ioc, handle); init_completion(&ioc->tm_cmds.done); ioc->put_smid_hi_priority(ioc, smid, msix_task); wait_for_completion_timeout(&ioc->tm_cmds.done, timeout*HZ); if (!(ioc->tm_cmds.status & MPT3_CMD_COMPLETE)) { mpt3sas_check_cmd_timeout(ioc, ioc->tm_cmds.status, mpi_request, sizeof(Mpi2SCSITaskManagementRequest_t)/4, issue_reset); if (issue_reset) { rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); rc = (!rc) ? SUCCESS : FAILED; goto out; } } /* sync IRQs in case those were busy during flush. */ mpt3sas_base_sync_reply_irqs(ioc, 0); if (ioc->tm_cmds.status & MPT3_CMD_REPLY_VALID) { mpt3sas_trigger_master(ioc, MASTER_TRIGGER_TASK_MANAGMENT); mpi_reply = ioc->tm_cmds.reply; dtmprintk(ioc, ioc_info(ioc, "complete tm: ioc_status(0x%04x), loginfo(0x%08x), term_count(0x%08x)\n", le16_to_cpu(mpi_reply->IOCStatus), le32_to_cpu(mpi_reply->IOCLogInfo), le32_to_cpu(mpi_reply->TerminationCount))); if (ioc->logging_level & MPT_DEBUG_TM) { _scsih_response_code(ioc, mpi_reply->ResponseCode); if (mpi_reply->IOCStatus) _debug_dump_mf(mpi_request, sizeof(Mpi2SCSITaskManagementRequest_t)/4); } } switch (type) { case MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK: rc = SUCCESS; /* * If DevHandle filed in smid_task's entry of request pool * doesn't match with device handle on which this task abort * TM is received then it means that TM has successfully * aborted the timed out command. Since smid_task's entry in * request pool will be memset to zero once the timed out * command is returned to the SML. If the command is not * aborted then smid_task’s entry won’t be cleared and it * will have same DevHandle value on which this task abort TM * is received and driver will return the TM status as FAILED. */ request = mpt3sas_base_get_msg_frame(ioc, smid_task); if (le16_to_cpu(request->DevHandle) != handle) break; ioc_info(ioc, "Task abort tm failed: handle(0x%04x)," "timeout(%d) tr_method(0x%x) smid(%d) msix_index(%d)\n", handle, timeout, tr_method, smid_task, msix_task); rc = FAILED; break; case MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET: case MPI2_SCSITASKMGMT_TASKTYPE_ABRT_TASK_SET: case MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET: rc = scsih_tm_post_processing(ioc, handle, channel, id, lun, type, smid_task); break; case MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK: rc = SUCCESS; break; default: rc = FAILED; break; } out: mpt3sas_scsih_clear_tm_flag(ioc, handle); ioc->tm_cmds.status = MPT3_CMD_NOT_USED; return rc; } int mpt3sas_scsih_issue_locked_tm(struct MPT3SAS_ADAPTER *ioc, u16 handle, uint channel, uint id, u64 lun, u8 type, u16 smid_task, u16 msix_task, u8 timeout, u8 tr_method) { int ret; mutex_lock(&ioc->tm_cmds.mutex); ret = mpt3sas_scsih_issue_tm(ioc, handle, channel, id, lun, type, smid_task, msix_task, timeout, tr_method); mutex_unlock(&ioc->tm_cmds.mutex); return ret; } /** * _scsih_tm_display_info - displays info about the device * @ioc: per adapter struct * @scmd: pointer to scsi command object * * Called by task management callback handlers. */ static void _scsih_tm_display_info(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd) { struct scsi_target *starget = scmd->device->sdev_target; struct MPT3SAS_TARGET *priv_target = starget->hostdata; struct _sas_device *sas_device = NULL; struct _pcie_device *pcie_device = NULL; unsigned long flags; char *device_str = NULL; if (!priv_target) return; if (ioc->hide_ir_msg) device_str = "WarpDrive"; else device_str = "volume"; scsi_print_command(scmd); if (priv_target->flags & MPT_TARGET_FLAGS_VOLUME) { starget_printk(KERN_INFO, starget, "%s handle(0x%04x), %s wwid(0x%016llx)\n", device_str, priv_target->handle, device_str, (unsigned long long)priv_target->sas_address); } else if (priv_target->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) { spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device = __mpt3sas_get_pdev_from_target(ioc, priv_target); if (pcie_device) { starget_printk(KERN_INFO, starget, "handle(0x%04x), wwid(0x%016llx), port(%d)\n", pcie_device->handle, (unsigned long long)pcie_device->wwid, pcie_device->port_num); if (pcie_device->enclosure_handle != 0) starget_printk(KERN_INFO, starget, "enclosure logical id(0x%016llx), slot(%d)\n", (unsigned long long) pcie_device->enclosure_logical_id, pcie_device->slot); if (pcie_device->connector_name[0] != '\0') starget_printk(KERN_INFO, starget, "enclosure level(0x%04x), connector name( %s)\n", pcie_device->enclosure_level, pcie_device->connector_name); pcie_device_put(pcie_device); } spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); } else { spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_from_target(ioc, priv_target); if (sas_device) { if (priv_target->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) { starget_printk(KERN_INFO, starget, "volume handle(0x%04x), " "volume wwid(0x%016llx)\n", sas_device->volume_handle, (unsigned long long)sas_device->volume_wwid); } starget_printk(KERN_INFO, starget, "handle(0x%04x), sas_address(0x%016llx), phy(%d)\n", sas_device->handle, (unsigned long long)sas_device->sas_address, sas_device->phy); _scsih_display_enclosure_chassis_info(NULL, sas_device, NULL, starget); sas_device_put(sas_device); } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); } } /** * scsih_abort - eh threads main abort routine * @scmd: pointer to scsi command object * * Return: SUCCESS if command aborted else FAILED */ static int scsih_abort(struct scsi_cmnd *scmd) { struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host); struct MPT3SAS_DEVICE *sas_device_priv_data; struct scsiio_tracker *st = scsi_cmd_priv(scmd); u16 handle; int r; u8 timeout = 30; struct _pcie_device *pcie_device = NULL; sdev_printk(KERN_INFO, scmd->device, "attempting task abort!" "scmd(0x%p), outstanding for %u ms & timeout %u ms\n", scmd, jiffies_to_msecs(jiffies - scmd->jiffies_at_alloc), (scmd->request->timeout / HZ) * 1000); _scsih_tm_display_info(ioc, scmd); sas_device_priv_data = scmd->device->hostdata; if (!sas_device_priv_data || !sas_device_priv_data->sas_target || ioc->remove_host) { sdev_printk(KERN_INFO, scmd->device, "device been deleted! scmd(0x%p)\n", scmd); scmd->result = DID_NO_CONNECT << 16; scmd->scsi_done(scmd); r = SUCCESS; goto out; } /* check for completed command */ if (st == NULL || st->cb_idx == 0xFF) { sdev_printk(KERN_INFO, scmd->device, "No reference found at " "driver, assuming scmd(0x%p) might have completed\n", scmd); scmd->result = DID_RESET << 16; r = SUCCESS; goto out; } /* for hidden raid components and volumes this is not supported */ if (sas_device_priv_data->sas_target->flags & MPT_TARGET_FLAGS_RAID_COMPONENT || sas_device_priv_data->sas_target->flags & MPT_TARGET_FLAGS_VOLUME) { scmd->result = DID_RESET << 16; r = FAILED; goto out; } mpt3sas_halt_firmware(ioc); handle = sas_device_priv_data->sas_target->handle; pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle); if (pcie_device && (!ioc->tm_custom_handling) && (!(mpt3sas_scsih_is_pcie_scsi_device(pcie_device->device_info)))) timeout = ioc->nvme_abort_timeout; r = mpt3sas_scsih_issue_locked_tm(ioc, handle, scmd->device->channel, scmd->device->id, scmd->device->lun, MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK, st->smid, st->msix_io, timeout, 0); /* Command must be cleared after abort */ if (r == SUCCESS && st->cb_idx != 0xFF) r = FAILED; out: sdev_printk(KERN_INFO, scmd->device, "task abort: %s scmd(0x%p)\n", ((r == SUCCESS) ? "SUCCESS" : "FAILED"), scmd); if (pcie_device) pcie_device_put(pcie_device); return r; } /** * scsih_dev_reset - eh threads main device reset routine * @scmd: pointer to scsi command object * * Return: SUCCESS if command aborted else FAILED */ static int scsih_dev_reset(struct scsi_cmnd *scmd) { struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host); struct MPT3SAS_DEVICE *sas_device_priv_data; struct _sas_device *sas_device = NULL; struct _pcie_device *pcie_device = NULL; u16 handle; u8 tr_method = 0; u8 tr_timeout = 30; int r; struct scsi_target *starget = scmd->device->sdev_target; struct MPT3SAS_TARGET *target_priv_data = starget->hostdata; sdev_printk(KERN_INFO, scmd->device, "attempting device reset! scmd(0x%p)\n", scmd); _scsih_tm_display_info(ioc, scmd); sas_device_priv_data = scmd->device->hostdata; if (!sas_device_priv_data || !sas_device_priv_data->sas_target || ioc->remove_host) { sdev_printk(KERN_INFO, scmd->device, "device been deleted! scmd(0x%p)\n", scmd); scmd->result = DID_NO_CONNECT << 16; scmd->scsi_done(scmd); r = SUCCESS; goto out; } /* for hidden raid components obtain the volume_handle */ handle = 0; if (sas_device_priv_data->sas_target->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) { sas_device = mpt3sas_get_sdev_from_target(ioc, target_priv_data); if (sas_device) handle = sas_device->volume_handle; } else handle = sas_device_priv_data->sas_target->handle; if (!handle) { scmd->result = DID_RESET << 16; r = FAILED; goto out; } pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle); if (pcie_device && (!ioc->tm_custom_handling) && (!(mpt3sas_scsih_is_pcie_scsi_device(pcie_device->device_info)))) { tr_timeout = pcie_device->reset_timeout; tr_method = MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE; } else tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET; r = mpt3sas_scsih_issue_locked_tm(ioc, handle, scmd->device->channel, scmd->device->id, scmd->device->lun, MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET, 0, 0, tr_timeout, tr_method); /* Check for busy commands after reset */ if (r == SUCCESS && scsi_device_busy(scmd->device)) r = FAILED; out: sdev_printk(KERN_INFO, scmd->device, "device reset: %s scmd(0x%p)\n", ((r == SUCCESS) ? "SUCCESS" : "FAILED"), scmd); if (sas_device) sas_device_put(sas_device); if (pcie_device) pcie_device_put(pcie_device); return r; } /** * scsih_target_reset - eh threads main target reset routine * @scmd: pointer to scsi command object * * Return: SUCCESS if command aborted else FAILED */ static int scsih_target_reset(struct scsi_cmnd *scmd) { struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host); struct MPT3SAS_DEVICE *sas_device_priv_data; struct _sas_device *sas_device = NULL; struct _pcie_device *pcie_device = NULL; u16 handle; u8 tr_method = 0; u8 tr_timeout = 30; int r; struct scsi_target *starget = scmd->device->sdev_target; struct MPT3SAS_TARGET *target_priv_data = starget->hostdata; starget_printk(KERN_INFO, starget, "attempting target reset! scmd(0x%p)\n", scmd); _scsih_tm_display_info(ioc, scmd); sas_device_priv_data = scmd->device->hostdata; if (!sas_device_priv_data || !sas_device_priv_data->sas_target || ioc->remove_host) { starget_printk(KERN_INFO, starget, "target been deleted! scmd(0x%p)\n", scmd); scmd->result = DID_NO_CONNECT << 16; scmd->scsi_done(scmd); r = SUCCESS; goto out; } /* for hidden raid components obtain the volume_handle */ handle = 0; if (sas_device_priv_data->sas_target->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) { sas_device = mpt3sas_get_sdev_from_target(ioc, target_priv_data); if (sas_device) handle = sas_device->volume_handle; } else handle = sas_device_priv_data->sas_target->handle; if (!handle) { scmd->result = DID_RESET << 16; r = FAILED; goto out; } pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle); if (pcie_device && (!ioc->tm_custom_handling) && (!(mpt3sas_scsih_is_pcie_scsi_device(pcie_device->device_info)))) { tr_timeout = pcie_device->reset_timeout; tr_method = MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE; } else tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET; r = mpt3sas_scsih_issue_locked_tm(ioc, handle, scmd->device->channel, scmd->device->id, 0, MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET, 0, 0, tr_timeout, tr_method); /* Check for busy commands after reset */ if (r == SUCCESS && atomic_read(&starget->target_busy)) r = FAILED; out: starget_printk(KERN_INFO, starget, "target reset: %s scmd(0x%p)\n", ((r == SUCCESS) ? "SUCCESS" : "FAILED"), scmd); if (sas_device) sas_device_put(sas_device); if (pcie_device) pcie_device_put(pcie_device); return r; } /** * scsih_host_reset - eh threads main host reset routine * @scmd: pointer to scsi command object * * Return: SUCCESS if command aborted else FAILED */ static int scsih_host_reset(struct scsi_cmnd *scmd) { struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host); int r, retval; ioc_info(ioc, "attempting host reset! scmd(0x%p)\n", scmd); scsi_print_command(scmd); if (ioc->is_driver_loading || ioc->remove_host) { ioc_info(ioc, "Blocking the host reset\n"); r = FAILED; goto out; } retval = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); r = (retval < 0) ? FAILED : SUCCESS; out: ioc_info(ioc, "host reset: %s scmd(0x%p)\n", r == SUCCESS ? "SUCCESS" : "FAILED", scmd); return r; } /** * _scsih_fw_event_add - insert and queue up fw_event * @ioc: per adapter object * @fw_event: object describing the event * Context: This function will acquire ioc->fw_event_lock. * * This adds the firmware event object into link list, then queues it up to * be processed from user context. */ static void _scsih_fw_event_add(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { unsigned long flags; if (ioc->firmware_event_thread == NULL) return; spin_lock_irqsave(&ioc->fw_event_lock, flags); fw_event_work_get(fw_event); INIT_LIST_HEAD(&fw_event->list); list_add_tail(&fw_event->list, &ioc->fw_event_list); INIT_WORK(&fw_event->work, _firmware_event_work); fw_event_work_get(fw_event); queue_work(ioc->firmware_event_thread, &fw_event->work); spin_unlock_irqrestore(&ioc->fw_event_lock, flags); } /** * _scsih_fw_event_del_from_list - delete fw_event from the list * @ioc: per adapter object * @fw_event: object describing the event * Context: This function will acquire ioc->fw_event_lock. * * If the fw_event is on the fw_event_list, remove it and do a put. */ static void _scsih_fw_event_del_from_list(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { unsigned long flags; spin_lock_irqsave(&ioc->fw_event_lock, flags); if (!list_empty(&fw_event->list)) { list_del_init(&fw_event->list); fw_event_work_put(fw_event); } spin_unlock_irqrestore(&ioc->fw_event_lock, flags); } /** * mpt3sas_send_trigger_data_event - send event for processing trigger data * @ioc: per adapter object * @event_data: trigger event data */ void mpt3sas_send_trigger_data_event(struct MPT3SAS_ADAPTER *ioc, struct SL_WH_TRIGGERS_EVENT_DATA_T *event_data) { struct fw_event_work *fw_event; u16 sz; if (ioc->is_driver_loading) return; sz = sizeof(*event_data); fw_event = alloc_fw_event_work(sz); if (!fw_event) return; fw_event->event = MPT3SAS_PROCESS_TRIGGER_DIAG; fw_event->ioc = ioc; memcpy(fw_event->event_data, event_data, sizeof(*event_data)); _scsih_fw_event_add(ioc, fw_event); fw_event_work_put(fw_event); } /** * _scsih_error_recovery_delete_devices - remove devices not responding * @ioc: per adapter object */ static void _scsih_error_recovery_delete_devices(struct MPT3SAS_ADAPTER *ioc) { struct fw_event_work *fw_event; fw_event = alloc_fw_event_work(0); if (!fw_event) return; fw_event->event = MPT3SAS_REMOVE_UNRESPONDING_DEVICES; fw_event->ioc = ioc; _scsih_fw_event_add(ioc, fw_event); fw_event_work_put(fw_event); } /** * mpt3sas_port_enable_complete - port enable completed (fake event) * @ioc: per adapter object */ void mpt3sas_port_enable_complete(struct MPT3SAS_ADAPTER *ioc) { struct fw_event_work *fw_event; fw_event = alloc_fw_event_work(0); if (!fw_event) return; fw_event->event = MPT3SAS_PORT_ENABLE_COMPLETE; fw_event->ioc = ioc; _scsih_fw_event_add(ioc, fw_event); fw_event_work_put(fw_event); } static struct fw_event_work *dequeue_next_fw_event(struct MPT3SAS_ADAPTER *ioc) { unsigned long flags; struct fw_event_work *fw_event = NULL; spin_lock_irqsave(&ioc->fw_event_lock, flags); if (!list_empty(&ioc->fw_event_list)) { fw_event = list_first_entry(&ioc->fw_event_list, struct fw_event_work, list); list_del_init(&fw_event->list); } spin_unlock_irqrestore(&ioc->fw_event_lock, flags); return fw_event; } /** * _scsih_fw_event_cleanup_queue - cleanup event queue * @ioc: per adapter object * * Walk the firmware event queue, either killing timers, or waiting * for outstanding events to complete * * Context: task, can sleep */ static void _scsih_fw_event_cleanup_queue(struct MPT3SAS_ADAPTER *ioc) { struct fw_event_work *fw_event; if ((list_empty(&ioc->fw_event_list) && !ioc->current_event) || !ioc->firmware_event_thread) return; /* * Set current running event as ignore, so that * current running event will exit quickly. * As diag reset has occurred it is of no use * to process remaining stale event data entries. */ if (ioc->shost_recovery && ioc->current_event) ioc->current_event->ignore = 1; ioc->fw_events_cleanup = 1; while ((fw_event = dequeue_next_fw_event(ioc)) || (fw_event = ioc->current_event)) { /* * Don't call cancel_work_sync() for current_event * other than MPT3SAS_REMOVE_UNRESPONDING_DEVICES; * otherwise we may observe deadlock if current * hard reset issued as part of processing the current_event. * * Orginal logic of cleaning the current_event is added * for handling the back to back host reset issued by the user. * i.e. during back to back host reset, driver use to process * the two instances of MPT3SAS_REMOVE_UNRESPONDING_DEVICES * event back to back and this made the drives to unregister * the devices from SML. */ if (fw_event == ioc->current_event && ioc->current_event->event != MPT3SAS_REMOVE_UNRESPONDING_DEVICES) { ioc->current_event = NULL; continue; } /* * Driver has to clear ioc->start_scan flag when * it is cleaning up MPT3SAS_PORT_ENABLE_COMPLETE, * otherwise scsi_scan_host() API waits for the * 5 minute timer to expire. If we exit from * scsi_scan_host() early then we can issue the * new port enable request as part of current diag reset. */ if (fw_event->event == MPT3SAS_PORT_ENABLE_COMPLETE) { ioc->port_enable_cmds.status |= MPT3_CMD_RESET; ioc->start_scan = 0; } /* * Wait on the fw_event to complete. If this returns 1, then * the event was never executed, and we need a put for the * reference the work had on the fw_event. * * If it did execute, we wait for it to finish, and the put will * happen from _firmware_event_work() */ if (cancel_work_sync(&fw_event->work)) fw_event_work_put(fw_event); fw_event_work_put(fw_event); } ioc->fw_events_cleanup = 0; } /** * _scsih_internal_device_block - block the sdev device * @sdev: per device object * @sas_device_priv_data : per device driver private data * * make sure device is blocked without error, if not * print an error */ static void _scsih_internal_device_block(struct scsi_device *sdev, struct MPT3SAS_DEVICE *sas_device_priv_data) { int r = 0; sdev_printk(KERN_INFO, sdev, "device_block, handle(0x%04x)\n", sas_device_priv_data->sas_target->handle); sas_device_priv_data->block = 1; r = scsi_internal_device_block_nowait(sdev); if (r == -EINVAL) sdev_printk(KERN_WARNING, sdev, "device_block failed with return(%d) for handle(0x%04x)\n", r, sas_device_priv_data->sas_target->handle); } /** * _scsih_internal_device_unblock - unblock the sdev device * @sdev: per device object * @sas_device_priv_data : per device driver private data * make sure device is unblocked without error, if not retry * by blocking and then unblocking */ static void _scsih_internal_device_unblock(struct scsi_device *sdev, struct MPT3SAS_DEVICE *sas_device_priv_data) { int r = 0; sdev_printk(KERN_WARNING, sdev, "device_unblock and setting to running, " "handle(0x%04x)\n", sas_device_priv_data->sas_target->handle); sas_device_priv_data->block = 0; r = scsi_internal_device_unblock_nowait(sdev, SDEV_RUNNING); if (r == -EINVAL) { /* The device has been set to SDEV_RUNNING by SD layer during * device addition but the request queue is still stopped by * our earlier block call. We need to perform a block again * to get the device to SDEV_BLOCK and then to SDEV_RUNNING */ sdev_printk(KERN_WARNING, sdev, "device_unblock failed with return(%d) for handle(0x%04x) " "performing a block followed by an unblock\n", r, sas_device_priv_data->sas_target->handle); sas_device_priv_data->block = 1; r = scsi_internal_device_block_nowait(sdev); if (r) sdev_printk(KERN_WARNING, sdev, "retried device_block " "failed with return(%d) for handle(0x%04x)\n", r, sas_device_priv_data->sas_target->handle); sas_device_priv_data->block = 0; r = scsi_internal_device_unblock_nowait(sdev, SDEV_RUNNING); if (r) sdev_printk(KERN_WARNING, sdev, "retried device_unblock" " failed with return(%d) for handle(0x%04x)\n", r, sas_device_priv_data->sas_target->handle); } } /** * _scsih_ublock_io_all_device - unblock every device * @ioc: per adapter object * * change the device state from block to running */ static void _scsih_ublock_io_all_device(struct MPT3SAS_ADAPTER *ioc) { struct MPT3SAS_DEVICE *sas_device_priv_data; struct scsi_device *sdev; shost_for_each_device(sdev, ioc->shost) { sas_device_priv_data = sdev->hostdata; if (!sas_device_priv_data) continue; if (!sas_device_priv_data->block) continue; dewtprintk(ioc, sdev_printk(KERN_INFO, sdev, "device_running, handle(0x%04x)\n", sas_device_priv_data->sas_target->handle)); _scsih_internal_device_unblock(sdev, sas_device_priv_data); } } /** * _scsih_ublock_io_device - prepare device to be deleted * @ioc: per adapter object * @sas_address: sas address * @port: hba port entry * * unblock then put device in offline state */ static void _scsih_ublock_io_device(struct MPT3SAS_ADAPTER *ioc, u64 sas_address, struct hba_port *port) { struct MPT3SAS_DEVICE *sas_device_priv_data; struct scsi_device *sdev; shost_for_each_device(sdev, ioc->shost) { sas_device_priv_data = sdev->hostdata; if (!sas_device_priv_data) continue; if (sas_device_priv_data->sas_target->sas_address != sas_address) continue; if (sas_device_priv_data->sas_target->port != port) continue; if (sas_device_priv_data->block) _scsih_internal_device_unblock(sdev, sas_device_priv_data); } } /** * _scsih_block_io_all_device - set the device state to SDEV_BLOCK * @ioc: per adapter object * * During device pull we need to appropriately set the sdev state. */ static void _scsih_block_io_all_device(struct MPT3SAS_ADAPTER *ioc) { struct MPT3SAS_DEVICE *sas_device_priv_data; struct scsi_device *sdev; shost_for_each_device(sdev, ioc->shost) { sas_device_priv_data = sdev->hostdata; if (!sas_device_priv_data) continue; if (sas_device_priv_data->block) continue; if (sas_device_priv_data->ignore_delay_remove) { sdev_printk(KERN_INFO, sdev, "%s skip device_block for SES handle(0x%04x)\n", __func__, sas_device_priv_data->sas_target->handle); continue; } _scsih_internal_device_block(sdev, sas_device_priv_data); } } /** * _scsih_block_io_device - set the device state to SDEV_BLOCK * @ioc: per adapter object * @handle: device handle * * During device pull we need to appropriately set the sdev state. */ static void _scsih_block_io_device(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct MPT3SAS_DEVICE *sas_device_priv_data; struct scsi_device *sdev; struct _sas_device *sas_device; sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); shost_for_each_device(sdev, ioc->shost) { sas_device_priv_data = sdev->hostdata; if (!sas_device_priv_data) continue; if (sas_device_priv_data->sas_target->handle != handle) continue; if (sas_device_priv_data->block) continue; if (sas_device && sas_device->pend_sas_rphy_add) continue; if (sas_device_priv_data->ignore_delay_remove) { sdev_printk(KERN_INFO, sdev, "%s skip device_block for SES handle(0x%04x)\n", __func__, sas_device_priv_data->sas_target->handle); continue; } _scsih_internal_device_block(sdev, sas_device_priv_data); } if (sas_device) sas_device_put(sas_device); } /** * _scsih_block_io_to_children_attached_to_ex * @ioc: per adapter object * @sas_expander: the sas_device object * * This routine set sdev state to SDEV_BLOCK for all devices * attached to this expander. This function called when expander is * pulled. */ static void _scsih_block_io_to_children_attached_to_ex(struct MPT3SAS_ADAPTER *ioc, struct _sas_node *sas_expander) { struct _sas_port *mpt3sas_port; struct _sas_device *sas_device; struct _sas_node *expander_sibling; unsigned long flags; if (!sas_expander) return; list_for_each_entry(mpt3sas_port, &sas_expander->sas_port_list, port_list) { if (mpt3sas_port->remote_identify.device_type == SAS_END_DEVICE) { spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_by_addr(ioc, mpt3sas_port->remote_identify.sas_address, mpt3sas_port->hba_port); if (sas_device) { set_bit(sas_device->handle, ioc->blocking_handles); sas_device_put(sas_device); } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); } } list_for_each_entry(mpt3sas_port, &sas_expander->sas_port_list, port_list) { if (mpt3sas_port->remote_identify.device_type == SAS_EDGE_EXPANDER_DEVICE || mpt3sas_port->remote_identify.device_type == SAS_FANOUT_EXPANDER_DEVICE) { expander_sibling = mpt3sas_scsih_expander_find_by_sas_address( ioc, mpt3sas_port->remote_identify.sas_address, mpt3sas_port->hba_port); _scsih_block_io_to_children_attached_to_ex(ioc, expander_sibling); } } } /** * _scsih_block_io_to_children_attached_directly * @ioc: per adapter object * @event_data: topology change event data * * This routine set sdev state to SDEV_BLOCK for all devices * direct attached during device pull. */ static void _scsih_block_io_to_children_attached_directly(struct MPT3SAS_ADAPTER *ioc, Mpi2EventDataSasTopologyChangeList_t *event_data) { int i; u16 handle; u16 reason_code; for (i = 0; i < event_data->NumEntries; i++) { handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); if (!handle) continue; reason_code = event_data->PHY[i].PhyStatus & MPI2_EVENT_SAS_TOPO_RC_MASK; if (reason_code == MPI2_EVENT_SAS_TOPO_RC_DELAY_NOT_RESPONDING) _scsih_block_io_device(ioc, handle); } } /** * _scsih_block_io_to_pcie_children_attached_directly * @ioc: per adapter object * @event_data: topology change event data * * This routine set sdev state to SDEV_BLOCK for all devices * direct attached during device pull/reconnect. */ static void _scsih_block_io_to_pcie_children_attached_directly(struct MPT3SAS_ADAPTER *ioc, Mpi26EventDataPCIeTopologyChangeList_t *event_data) { int i; u16 handle; u16 reason_code; for (i = 0; i < event_data->NumEntries; i++) { handle = le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle); if (!handle) continue; reason_code = event_data->PortEntry[i].PortStatus; if (reason_code == MPI26_EVENT_PCIE_TOPO_PS_DELAY_NOT_RESPONDING) _scsih_block_io_device(ioc, handle); } } /** * _scsih_tm_tr_send - send task management request * @ioc: per adapter object * @handle: device handle * Context: interrupt time. * * This code is to initiate the device removal handshake protocol * with controller firmware. This function will issue target reset * using high priority request queue. It will send a sas iounit * control request (MPI2_SAS_OP_REMOVE_DEVICE) from this completion. * * This is designed to send muliple task management request at the same * time to the fifo. If the fifo is full, we will append the request, * and process it in a future completion. */ static void _scsih_tm_tr_send(struct MPT3SAS_ADAPTER *ioc, u16 handle) { Mpi2SCSITaskManagementRequest_t *mpi_request; u16 smid; struct _sas_device *sas_device = NULL; struct _pcie_device *pcie_device = NULL; struct MPT3SAS_TARGET *sas_target_priv_data = NULL; u64 sas_address = 0; unsigned long flags; struct _tr_list *delayed_tr; u32 ioc_state; u8 tr_method = 0; struct hba_port *port = NULL; if (ioc->pci_error_recovery) { dewtprintk(ioc, ioc_info(ioc, "%s: host in pci error recovery: handle(0x%04x)\n", __func__, handle)); return; } ioc_state = mpt3sas_base_get_iocstate(ioc, 1); if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) { dewtprintk(ioc, ioc_info(ioc, "%s: host is not operational: handle(0x%04x)\n", __func__, handle)); return; } /* if PD, then return */ if (test_bit(handle, ioc->pd_handles)) return; clear_bit(handle, ioc->pend_os_device_add); spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); if (sas_device && sas_device->starget && sas_device->starget->hostdata) { sas_target_priv_data = sas_device->starget->hostdata; sas_target_priv_data->deleted = 1; sas_address = sas_device->sas_address; port = sas_device->port; } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); if (!sas_device) { spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle); if (pcie_device && pcie_device->starget && pcie_device->starget->hostdata) { sas_target_priv_data = pcie_device->starget->hostdata; sas_target_priv_data->deleted = 1; sas_address = pcie_device->wwid; } spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); if (pcie_device && (!ioc->tm_custom_handling) && (!(mpt3sas_scsih_is_pcie_scsi_device( pcie_device->device_info)))) tr_method = MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE; else tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET; } if (sas_target_priv_data) { dewtprintk(ioc, ioc_info(ioc, "setting delete flag: handle(0x%04x), sas_addr(0x%016llx)\n", handle, (u64)sas_address)); if (sas_device) { if (sas_device->enclosure_handle != 0) dewtprintk(ioc, ioc_info(ioc, "setting delete flag:enclosure logical id(0x%016llx), slot(%d)\n", (u64)sas_device->enclosure_logical_id, sas_device->slot)); if (sas_device->connector_name[0] != '\0') dewtprintk(ioc, ioc_info(ioc, "setting delete flag: enclosure level(0x%04x), connector name( %s)\n", sas_device->enclosure_level, sas_device->connector_name)); } else if (pcie_device) { if (pcie_device->enclosure_handle != 0) dewtprintk(ioc, ioc_info(ioc, "setting delete flag: logical id(0x%016llx), slot(%d)\n", (u64)pcie_device->enclosure_logical_id, pcie_device->slot)); if (pcie_device->connector_name[0] != '\0') dewtprintk(ioc, ioc_info(ioc, "setting delete flag:, enclosure level(0x%04x), connector name( %s)\n", pcie_device->enclosure_level, pcie_device->connector_name)); } _scsih_ublock_io_device(ioc, sas_address, port); sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE; } smid = mpt3sas_base_get_smid_hpr(ioc, ioc->tm_tr_cb_idx); if (!smid) { delayed_tr = kzalloc(sizeof(*delayed_tr), GFP_ATOMIC); if (!delayed_tr) goto out; INIT_LIST_HEAD(&delayed_tr->list); delayed_tr->handle = handle; list_add_tail(&delayed_tr->list, &ioc->delayed_tr_list); dewtprintk(ioc, ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n", handle)); goto out; } dewtprintk(ioc, ioc_info(ioc, "tr_send:handle(0x%04x), (open), smid(%d), cb(%d)\n", handle, smid, ioc->tm_tr_cb_idx)); mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t)); mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT; mpi_request->DevHandle = cpu_to_le16(handle); mpi_request->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET; mpi_request->MsgFlags = tr_method; set_bit(handle, ioc->device_remove_in_progress); ioc->put_smid_hi_priority(ioc, smid, 0); mpt3sas_trigger_master(ioc, MASTER_TRIGGER_DEVICE_REMOVAL); out: if (sas_device) sas_device_put(sas_device); if (pcie_device) pcie_device_put(pcie_device); } /** * _scsih_tm_tr_complete - * @ioc: per adapter object * @smid: system request message index * @msix_index: MSIX table index supplied by the OS * @reply: reply message frame(lower 32bit addr) * Context: interrupt time. * * This is the target reset completion routine. * This code is part of the code to initiate the device removal * handshake protocol with controller firmware. * It will send a sas iounit control request (MPI2_SAS_OP_REMOVE_DEVICE) * * Return: 1 meaning mf should be freed from _base_interrupt * 0 means the mf is freed from this function. */ static u8 _scsih_tm_tr_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply) { u16 handle; Mpi2SCSITaskManagementRequest_t *mpi_request_tm; Mpi2SCSITaskManagementReply_t *mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply); Mpi2SasIoUnitControlRequest_t *mpi_request; u16 smid_sas_ctrl; u32 ioc_state; struct _sc_list *delayed_sc; if (ioc->pci_error_recovery) { dewtprintk(ioc, ioc_info(ioc, "%s: host in pci error recovery\n", __func__)); return 1; } ioc_state = mpt3sas_base_get_iocstate(ioc, 1); if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) { dewtprintk(ioc, ioc_info(ioc, "%s: host is not operational\n", __func__)); return 1; } if (unlikely(!mpi_reply)) { ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return 1; } mpi_request_tm = mpt3sas_base_get_msg_frame(ioc, smid); handle = le16_to_cpu(mpi_request_tm->DevHandle); if (handle != le16_to_cpu(mpi_reply->DevHandle)) { dewtprintk(ioc, ioc_err(ioc, "spurious interrupt: handle(0x%04x:0x%04x), smid(%d)!!!\n", handle, le16_to_cpu(mpi_reply->DevHandle), smid)); return 0; } mpt3sas_trigger_master(ioc, MASTER_TRIGGER_TASK_MANAGMENT); dewtprintk(ioc, ioc_info(ioc, "tr_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x), completed(%d)\n", handle, smid, le16_to_cpu(mpi_reply->IOCStatus), le32_to_cpu(mpi_reply->IOCLogInfo), le32_to_cpu(mpi_reply->TerminationCount))); smid_sas_ctrl = mpt3sas_base_get_smid(ioc, ioc->tm_sas_control_cb_idx); if (!smid_sas_ctrl) { delayed_sc = kzalloc(sizeof(*delayed_sc), GFP_ATOMIC); if (!delayed_sc) return _scsih_check_for_pending_tm(ioc, smid); INIT_LIST_HEAD(&delayed_sc->list); delayed_sc->handle = le16_to_cpu(mpi_request_tm->DevHandle); list_add_tail(&delayed_sc->list, &ioc->delayed_sc_list); dewtprintk(ioc, ioc_info(ioc, "DELAYED:sc:handle(0x%04x), (open)\n", handle)); return _scsih_check_for_pending_tm(ioc, smid); } dewtprintk(ioc, ioc_info(ioc, "sc_send:handle(0x%04x), (open), smid(%d), cb(%d)\n", handle, smid_sas_ctrl, ioc->tm_sas_control_cb_idx)); mpi_request = mpt3sas_base_get_msg_frame(ioc, smid_sas_ctrl); memset(mpi_request, 0, sizeof(Mpi2SasIoUnitControlRequest_t)); mpi_request->Function = MPI2_FUNCTION_SAS_IO_UNIT_CONTROL; mpi_request->Operation = MPI2_SAS_OP_REMOVE_DEVICE; mpi_request->DevHandle = mpi_request_tm->DevHandle; ioc->put_smid_default(ioc, smid_sas_ctrl); return _scsih_check_for_pending_tm(ioc, smid); } /** _scsih_allow_scmd_to_device - check whether scmd needs to * issue to IOC or not. * @ioc: per adapter object * @scmd: pointer to scsi command object * * Returns true if scmd can be issued to IOC otherwise returns false. */ inline bool _scsih_allow_scmd_to_device(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd) { if (ioc->pci_error_recovery) return false; if (ioc->hba_mpi_version_belonged == MPI2_VERSION) { if (ioc->remove_host) return false; return true; } if (ioc->remove_host) { switch (scmd->cmnd[0]) { case SYNCHRONIZE_CACHE: case START_STOP: return true; default: return false; } } return true; } /** * _scsih_sas_control_complete - completion routine * @ioc: per adapter object * @smid: system request message index * @msix_index: MSIX table index supplied by the OS * @reply: reply message frame(lower 32bit addr) * Context: interrupt time. * * This is the sas iounit control completion routine. * This code is part of the code to initiate the device removal * handshake protocol with controller firmware. * * Return: 1 meaning mf should be freed from _base_interrupt * 0 means the mf is freed from this function. */ static u8 _scsih_sas_control_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply) { Mpi2SasIoUnitControlReply_t *mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply); if (likely(mpi_reply)) { dewtprintk(ioc, ioc_info(ioc, "sc_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x)\n", le16_to_cpu(mpi_reply->DevHandle), smid, le16_to_cpu(mpi_reply->IOCStatus), le32_to_cpu(mpi_reply->IOCLogInfo))); if (le16_to_cpu(mpi_reply->IOCStatus) == MPI2_IOCSTATUS_SUCCESS) { clear_bit(le16_to_cpu(mpi_reply->DevHandle), ioc->device_remove_in_progress); } } else { ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); } return mpt3sas_check_for_pending_internal_cmds(ioc, smid); } /** * _scsih_tm_tr_volume_send - send target reset request for volumes * @ioc: per adapter object * @handle: device handle * Context: interrupt time. * * This is designed to send muliple task management request at the same * time to the fifo. If the fifo is full, we will append the request, * and process it in a future completion. */ static void _scsih_tm_tr_volume_send(struct MPT3SAS_ADAPTER *ioc, u16 handle) { Mpi2SCSITaskManagementRequest_t *mpi_request; u16 smid; struct _tr_list *delayed_tr; if (ioc->pci_error_recovery) { dewtprintk(ioc, ioc_info(ioc, "%s: host reset in progress!\n", __func__)); return; } smid = mpt3sas_base_get_smid_hpr(ioc, ioc->tm_tr_volume_cb_idx); if (!smid) { delayed_tr = kzalloc(sizeof(*delayed_tr), GFP_ATOMIC); if (!delayed_tr) return; INIT_LIST_HEAD(&delayed_tr->list); delayed_tr->handle = handle; list_add_tail(&delayed_tr->list, &ioc->delayed_tr_volume_list); dewtprintk(ioc, ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n", handle)); return; } dewtprintk(ioc, ioc_info(ioc, "tr_send:handle(0x%04x), (open), smid(%d), cb(%d)\n", handle, smid, ioc->tm_tr_volume_cb_idx)); mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t)); mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT; mpi_request->DevHandle = cpu_to_le16(handle); mpi_request->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET; ioc->put_smid_hi_priority(ioc, smid, 0); } /** * _scsih_tm_volume_tr_complete - target reset completion * @ioc: per adapter object * @smid: system request message index * @msix_index: MSIX table index supplied by the OS * @reply: reply message frame(lower 32bit addr) * Context: interrupt time. * * Return: 1 meaning mf should be freed from _base_interrupt * 0 means the mf is freed from this function. */ static u8 _scsih_tm_volume_tr_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply) { u16 handle; Mpi2SCSITaskManagementRequest_t *mpi_request_tm; Mpi2SCSITaskManagementReply_t *mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply); if (ioc->shost_recovery || ioc->pci_error_recovery) { dewtprintk(ioc, ioc_info(ioc, "%s: host reset in progress!\n", __func__)); return 1; } if (unlikely(!mpi_reply)) { ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return 1; } mpi_request_tm = mpt3sas_base_get_msg_frame(ioc, smid); handle = le16_to_cpu(mpi_request_tm->DevHandle); if (handle != le16_to_cpu(mpi_reply->DevHandle)) { dewtprintk(ioc, ioc_err(ioc, "spurious interrupt: handle(0x%04x:0x%04x), smid(%d)!!!\n", handle, le16_to_cpu(mpi_reply->DevHandle), smid)); return 0; } dewtprintk(ioc, ioc_info(ioc, "tr_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x), completed(%d)\n", handle, smid, le16_to_cpu(mpi_reply->IOCStatus), le32_to_cpu(mpi_reply->IOCLogInfo), le32_to_cpu(mpi_reply->TerminationCount))); return _scsih_check_for_pending_tm(ioc, smid); } /** * _scsih_issue_delayed_event_ack - issue delayed Event ACK messages * @ioc: per adapter object * @smid: system request message index * @event: Event ID * @event_context: used to track events uniquely * * Context - processed in interrupt context. */ static void _scsih_issue_delayed_event_ack(struct MPT3SAS_ADAPTER *ioc, u16 smid, U16 event, U32 event_context) { Mpi2EventAckRequest_t *ack_request; int i = smid - ioc->internal_smid; unsigned long flags; /* Without releasing the smid just update the * call back index and reuse the same smid for * processing this delayed request */ spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); ioc->internal_lookup[i].cb_idx = ioc->base_cb_idx; spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags); dewtprintk(ioc, ioc_info(ioc, "EVENT ACK: event(0x%04x), smid(%d), cb(%d)\n", le16_to_cpu(event), smid, ioc->base_cb_idx)); ack_request = mpt3sas_base_get_msg_frame(ioc, smid); memset(ack_request, 0, sizeof(Mpi2EventAckRequest_t)); ack_request->Function = MPI2_FUNCTION_EVENT_ACK; ack_request->Event = event; ack_request->EventContext = event_context; ack_request->VF_ID = 0; /* TODO */ ack_request->VP_ID = 0; ioc->put_smid_default(ioc, smid); } /** * _scsih_issue_delayed_sas_io_unit_ctrl - issue delayed * sas_io_unit_ctrl messages * @ioc: per adapter object * @smid: system request message index * @handle: device handle * * Context - processed in interrupt context. */ static void _scsih_issue_delayed_sas_io_unit_ctrl(struct MPT3SAS_ADAPTER *ioc, u16 smid, u16 handle) { Mpi2SasIoUnitControlRequest_t *mpi_request; u32 ioc_state; int i = smid - ioc->internal_smid; unsigned long flags; if (ioc->remove_host) { dewtprintk(ioc, ioc_info(ioc, "%s: host has been removed\n", __func__)); return; } else if (ioc->pci_error_recovery) { dewtprintk(ioc, ioc_info(ioc, "%s: host in pci error recovery\n", __func__)); return; } ioc_state = mpt3sas_base_get_iocstate(ioc, 1); if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) { dewtprintk(ioc, ioc_info(ioc, "%s: host is not operational\n", __func__)); return; } /* Without releasing the smid just update the * call back index and reuse the same smid for * processing this delayed request */ spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); ioc->internal_lookup[i].cb_idx = ioc->tm_sas_control_cb_idx; spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags); dewtprintk(ioc, ioc_info(ioc, "sc_send:handle(0x%04x), (open), smid(%d), cb(%d)\n", handle, smid, ioc->tm_sas_control_cb_idx)); mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); memset(mpi_request, 0, sizeof(Mpi2SasIoUnitControlRequest_t)); mpi_request->Function = MPI2_FUNCTION_SAS_IO_UNIT_CONTROL; mpi_request->Operation = MPI2_SAS_OP_REMOVE_DEVICE; mpi_request->DevHandle = cpu_to_le16(handle); ioc->put_smid_default(ioc, smid); } /** * mpt3sas_check_for_pending_internal_cmds - check for pending internal messages * @ioc: per adapter object * @smid: system request message index * * Context: Executed in interrupt context * * This will check delayed internal messages list, and process the * next request. * * Return: 1 meaning mf should be freed from _base_interrupt * 0 means the mf is freed from this function. */ u8 mpt3sas_check_for_pending_internal_cmds(struct MPT3SAS_ADAPTER *ioc, u16 smid) { struct _sc_list *delayed_sc; struct _event_ack_list *delayed_event_ack; if (!list_empty(&ioc->delayed_event_ack_list)) { delayed_event_ack = list_entry(ioc->delayed_event_ack_list.next, struct _event_ack_list, list); _scsih_issue_delayed_event_ack(ioc, smid, delayed_event_ack->Event, delayed_event_ack->EventContext); list_del(&delayed_event_ack->list); kfree(delayed_event_ack); return 0; } if (!list_empty(&ioc->delayed_sc_list)) { delayed_sc = list_entry(ioc->delayed_sc_list.next, struct _sc_list, list); _scsih_issue_delayed_sas_io_unit_ctrl(ioc, smid, delayed_sc->handle); list_del(&delayed_sc->list); kfree(delayed_sc); return 0; } return 1; } /** * _scsih_check_for_pending_tm - check for pending task management * @ioc: per adapter object * @smid: system request message index * * This will check delayed target reset list, and feed the * next reqeust. * * Return: 1 meaning mf should be freed from _base_interrupt * 0 means the mf is freed from this function. */ static u8 _scsih_check_for_pending_tm(struct MPT3SAS_ADAPTER *ioc, u16 smid) { struct _tr_list *delayed_tr; if (!list_empty(&ioc->delayed_tr_volume_list)) { delayed_tr = list_entry(ioc->delayed_tr_volume_list.next, struct _tr_list, list); mpt3sas_base_free_smid(ioc, smid); _scsih_tm_tr_volume_send(ioc, delayed_tr->handle); list_del(&delayed_tr->list); kfree(delayed_tr); return 0; } if (!list_empty(&ioc->delayed_tr_list)) { delayed_tr = list_entry(ioc->delayed_tr_list.next, struct _tr_list, list); mpt3sas_base_free_smid(ioc, smid); _scsih_tm_tr_send(ioc, delayed_tr->handle); list_del(&delayed_tr->list); kfree(delayed_tr); return 0; } return 1; } /** * _scsih_check_topo_delete_events - sanity check on topo events * @ioc: per adapter object * @event_data: the event data payload * * This routine added to better handle cable breaker. * * This handles the case where driver receives multiple expander * add and delete events in a single shot. When there is a delete event * the routine will void any pending add events waiting in the event queue. */ static void _scsih_check_topo_delete_events(struct MPT3SAS_ADAPTER *ioc, Mpi2EventDataSasTopologyChangeList_t *event_data) { struct fw_event_work *fw_event; Mpi2EventDataSasTopologyChangeList_t *local_event_data; u16 expander_handle; struct _sas_node *sas_expander; unsigned long flags; int i, reason_code; u16 handle; for (i = 0 ; i < event_data->NumEntries; i++) { handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); if (!handle) continue; reason_code = event_data->PHY[i].PhyStatus & MPI2_EVENT_SAS_TOPO_RC_MASK; if (reason_code == MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING) _scsih_tm_tr_send(ioc, handle); } expander_handle = le16_to_cpu(event_data->ExpanderDevHandle); if (expander_handle < ioc->sas_hba.num_phys) { _scsih_block_io_to_children_attached_directly(ioc, event_data); return; } if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_DELAY_NOT_RESPONDING) { /* put expander attached devices into blocking state */ spin_lock_irqsave(&ioc->sas_node_lock, flags); sas_expander = mpt3sas_scsih_expander_find_by_handle(ioc, expander_handle); _scsih_block_io_to_children_attached_to_ex(ioc, sas_expander); spin_unlock_irqrestore(&ioc->sas_node_lock, flags); do { handle = find_first_bit(ioc->blocking_handles, ioc->facts.MaxDevHandle); if (handle < ioc->facts.MaxDevHandle) _scsih_block_io_device(ioc, handle); } while (test_and_clear_bit(handle, ioc->blocking_handles)); } else if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_RESPONDING) _scsih_block_io_to_children_attached_directly(ioc, event_data); if (event_data->ExpStatus != MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING) return; /* mark ignore flag for pending events */ spin_lock_irqsave(&ioc->fw_event_lock, flags); list_for_each_entry(fw_event, &ioc->fw_event_list, list) { if (fw_event->event != MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST || fw_event->ignore) continue; local_event_data = (Mpi2EventDataSasTopologyChangeList_t *) fw_event->event_data; if (local_event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_ADDED || local_event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_RESPONDING) { if (le16_to_cpu(local_event_data->ExpanderDevHandle) == expander_handle) { dewtprintk(ioc, ioc_info(ioc, "setting ignoring flag\n")); fw_event->ignore = 1; } } } spin_unlock_irqrestore(&ioc->fw_event_lock, flags); } /** * _scsih_check_pcie_topo_remove_events - sanity check on topo * events * @ioc: per adapter object * @event_data: the event data payload * * This handles the case where driver receives multiple switch * or device add and delete events in a single shot. When there * is a delete event the routine will void any pending add * events waiting in the event queue. */ static void _scsih_check_pcie_topo_remove_events(struct MPT3SAS_ADAPTER *ioc, Mpi26EventDataPCIeTopologyChangeList_t *event_data) { struct fw_event_work *fw_event; Mpi26EventDataPCIeTopologyChangeList_t *local_event_data; unsigned long flags; int i, reason_code; u16 handle, switch_handle; for (i = 0; i < event_data->NumEntries; i++) { handle = le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle); if (!handle) continue; reason_code = event_data->PortEntry[i].PortStatus; if (reason_code == MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING) _scsih_tm_tr_send(ioc, handle); } switch_handle = le16_to_cpu(event_data->SwitchDevHandle); if (!switch_handle) { _scsih_block_io_to_pcie_children_attached_directly( ioc, event_data); return; } /* TODO We are not supporting cascaded PCIe Switch removal yet*/ if ((event_data->SwitchStatus == MPI26_EVENT_PCIE_TOPO_SS_DELAY_NOT_RESPONDING) || (event_data->SwitchStatus == MPI26_EVENT_PCIE_TOPO_SS_RESPONDING)) _scsih_block_io_to_pcie_children_attached_directly( ioc, event_data); if (event_data->SwitchStatus != MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING) return; /* mark ignore flag for pending events */ spin_lock_irqsave(&ioc->fw_event_lock, flags); list_for_each_entry(fw_event, &ioc->fw_event_list, list) { if (fw_event->event != MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST || fw_event->ignore) continue; local_event_data = (Mpi26EventDataPCIeTopologyChangeList_t *) fw_event->event_data; if (local_event_data->SwitchStatus == MPI2_EVENT_SAS_TOPO_ES_ADDED || local_event_data->SwitchStatus == MPI2_EVENT_SAS_TOPO_ES_RESPONDING) { if (le16_to_cpu(local_event_data->SwitchDevHandle) == switch_handle) { dewtprintk(ioc, ioc_info(ioc, "setting ignoring flag for switch event\n")); fw_event->ignore = 1; } } } spin_unlock_irqrestore(&ioc->fw_event_lock, flags); } /** * _scsih_set_volume_delete_flag - setting volume delete flag * @ioc: per adapter object * @handle: device handle * * This returns nothing. */ static void _scsih_set_volume_delete_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct _raid_device *raid_device; struct MPT3SAS_TARGET *sas_target_priv_data; unsigned long flags; spin_lock_irqsave(&ioc->raid_device_lock, flags); raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); if (raid_device && raid_device->starget && raid_device->starget->hostdata) { sas_target_priv_data = raid_device->starget->hostdata; sas_target_priv_data->deleted = 1; dewtprintk(ioc, ioc_info(ioc, "setting delete flag: handle(0x%04x), wwid(0x%016llx)\n", handle, (u64)raid_device->wwid)); } spin_unlock_irqrestore(&ioc->raid_device_lock, flags); } /** * _scsih_set_volume_handle_for_tr - set handle for target reset to volume * @handle: input handle * @a: handle for volume a * @b: handle for volume b * * IR firmware only supports two raid volumes. The purpose of this * routine is to set the volume handle in either a or b. When the given * input handle is non-zero, or when a and b have not been set before. */ static void _scsih_set_volume_handle_for_tr(u16 handle, u16 *a, u16 *b) { if (!handle || handle == *a || handle == *b) return; if (!*a) *a = handle; else if (!*b) *b = handle; } /** * _scsih_check_ir_config_unhide_events - check for UNHIDE events * @ioc: per adapter object * @event_data: the event data payload * Context: interrupt time. * * This routine will send target reset to volume, followed by target * resets to the PDs. This is called when a PD has been removed, or * volume has been deleted or removed. When the target reset is sent * to volume, the PD target resets need to be queued to start upon * completion of the volume target reset. */ static void _scsih_check_ir_config_unhide_events(struct MPT3SAS_ADAPTER *ioc, Mpi2EventDataIrConfigChangeList_t *event_data) { Mpi2EventIrConfigElement_t *element; int i; u16 handle, volume_handle, a, b; struct _tr_list *delayed_tr; a = 0; b = 0; if (ioc->is_warpdrive) return; /* Volume Resets for Deleted or Removed */ element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; for (i = 0; i < event_data->NumElements; i++, element++) { if (le32_to_cpu(event_data->Flags) & MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG) continue; if (element->ReasonCode == MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED || element->ReasonCode == MPI2_EVENT_IR_CHANGE_RC_REMOVED) { volume_handle = le16_to_cpu(element->VolDevHandle); _scsih_set_volume_delete_flag(ioc, volume_handle); _scsih_set_volume_handle_for_tr(volume_handle, &a, &b); } } /* Volume Resets for UNHIDE events */ element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; for (i = 0; i < event_data->NumElements; i++, element++) { if (le32_to_cpu(event_data->Flags) & MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG) continue; if (element->ReasonCode == MPI2_EVENT_IR_CHANGE_RC_UNHIDE) { volume_handle = le16_to_cpu(element->VolDevHandle); _scsih_set_volume_handle_for_tr(volume_handle, &a, &b); } } if (a) _scsih_tm_tr_volume_send(ioc, a); if (b) _scsih_tm_tr_volume_send(ioc, b); /* PD target resets */ element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; for (i = 0; i < event_data->NumElements; i++, element++) { if (element->ReasonCode != MPI2_EVENT_IR_CHANGE_RC_UNHIDE) continue; handle = le16_to_cpu(element->PhysDiskDevHandle); volume_handle = le16_to_cpu(element->VolDevHandle); clear_bit(handle, ioc->pd_handles); if (!volume_handle) _scsih_tm_tr_send(ioc, handle); else if (volume_handle == a || volume_handle == b) { delayed_tr = kzalloc(sizeof(*delayed_tr), GFP_ATOMIC); BUG_ON(!delayed_tr); INIT_LIST_HEAD(&delayed_tr->list); delayed_tr->handle = handle; list_add_tail(&delayed_tr->list, &ioc->delayed_tr_list); dewtprintk(ioc, ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n", handle)); } else _scsih_tm_tr_send(ioc, handle); } } /** * _scsih_check_volume_delete_events - set delete flag for volumes * @ioc: per adapter object * @event_data: the event data payload * Context: interrupt time. * * This will handle the case when the cable connected to entire volume is * pulled. We will take care of setting the deleted flag so normal IO will * not be sent. */ static void _scsih_check_volume_delete_events(struct MPT3SAS_ADAPTER *ioc, Mpi2EventDataIrVolume_t *event_data) { u32 state; if (event_data->ReasonCode != MPI2_EVENT_IR_VOLUME_RC_STATE_CHANGED) return; state = le32_to_cpu(event_data->NewValue); if (state == MPI2_RAID_VOL_STATE_MISSING || state == MPI2_RAID_VOL_STATE_FAILED) _scsih_set_volume_delete_flag(ioc, le16_to_cpu(event_data->VolDevHandle)); } /** * _scsih_temp_threshold_events - display temperature threshold exceeded events * @ioc: per adapter object * @event_data: the temp threshold event data * Context: interrupt time. */ static void _scsih_temp_threshold_events(struct MPT3SAS_ADAPTER *ioc, Mpi2EventDataTemperature_t *event_data) { u32 doorbell; if (ioc->temp_sensors_count >= event_data->SensorNum) { ioc_err(ioc, "Temperature Threshold flags %s%s%s%s exceeded for Sensor: %d !!!\n", le16_to_cpu(event_data->Status) & 0x1 ? "0 " : " ", le16_to_cpu(event_data->Status) & 0x2 ? "1 " : " ", le16_to_cpu(event_data->Status) & 0x4 ? "2 " : " ", le16_to_cpu(event_data->Status) & 0x8 ? "3 " : " ", event_data->SensorNum); ioc_err(ioc, "Current Temp In Celsius: %d\n", event_data->CurrentTemperature); if (ioc->hba_mpi_version_belonged != MPI2_VERSION) { doorbell = mpt3sas_base_get_iocstate(ioc, 0); if ((doorbell & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_FAULT) { mpt3sas_print_fault_code(ioc, doorbell & MPI2_DOORBELL_DATA_MASK); } else if ((doorbell & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_COREDUMP) { mpt3sas_print_coredump_info(ioc, doorbell & MPI2_DOORBELL_DATA_MASK); } } } } static int _scsih_set_satl_pending(struct scsi_cmnd *scmd, bool pending) { struct MPT3SAS_DEVICE *priv = scmd->device->hostdata; if (scmd->cmnd[0] != ATA_12 && scmd->cmnd[0] != ATA_16) return 0; if (pending) return test_and_set_bit(0, &priv->ata_command_pending); clear_bit(0, &priv->ata_command_pending); return 0; } /** * _scsih_flush_running_cmds - completing outstanding commands. * @ioc: per adapter object * * The flushing out of all pending scmd commands following host reset, * where all IO is dropped to the floor. */ static void _scsih_flush_running_cmds(struct MPT3SAS_ADAPTER *ioc) { struct scsi_cmnd *scmd; struct scsiio_tracker *st; u16 smid; int count = 0; for (smid = 1; smid <= ioc->scsiio_depth; smid++) { scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); if (!scmd) continue; count++; _scsih_set_satl_pending(scmd, false); st = scsi_cmd_priv(scmd); mpt3sas_base_clear_st(ioc, st); scsi_dma_unmap(scmd); if (ioc->pci_error_recovery || ioc->remove_host) scmd->result = DID_NO_CONNECT << 16; else scmd->result = DID_RESET << 16; scmd->scsi_done(scmd); } dtmprintk(ioc, ioc_info(ioc, "completing %d cmds\n", count)); } /** * _scsih_setup_eedp - setup MPI request for EEDP transfer * @ioc: per adapter object * @scmd: pointer to scsi command object * @mpi_request: pointer to the SCSI_IO request message frame * * Supporting protection 1 and 3. */ static void _scsih_setup_eedp(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd, Mpi25SCSIIORequest_t *mpi_request) { u16 eedp_flags; unsigned char prot_op = scsi_get_prot_op(scmd); unsigned char prot_type = scsi_get_prot_type(scmd); Mpi25SCSIIORequest_t *mpi_request_3v = (Mpi25SCSIIORequest_t *)mpi_request; if (prot_type == SCSI_PROT_DIF_TYPE0 || prot_op == SCSI_PROT_NORMAL) return; if (prot_op == SCSI_PROT_READ_STRIP) eedp_flags = MPI2_SCSIIO_EEDPFLAGS_CHECK_REMOVE_OP; else if (prot_op == SCSI_PROT_WRITE_INSERT) eedp_flags = MPI2_SCSIIO_EEDPFLAGS_INSERT_OP; else return; switch (prot_type) { case SCSI_PROT_DIF_TYPE1: case SCSI_PROT_DIF_TYPE2: /* * enable ref/guard checking * auto increment ref tag */ eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_INC_PRI_REFTAG | MPI2_SCSIIO_EEDPFLAGS_CHECK_REFTAG | MPI2_SCSIIO_EEDPFLAGS_CHECK_GUARD; mpi_request->CDB.EEDP32.PrimaryReferenceTag = cpu_to_be32(t10_pi_ref_tag(scmd->request)); break; case SCSI_PROT_DIF_TYPE3: /* * enable guard checking */ eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_CHECK_GUARD; break; } mpi_request_3v->EEDPBlockSize = cpu_to_le16(scmd->device->sector_size); if (ioc->is_gen35_ioc) eedp_flags |= MPI25_SCSIIO_EEDPFLAGS_APPTAG_DISABLE_MODE; mpi_request->EEDPFlags = cpu_to_le16(eedp_flags); } /** * _scsih_eedp_error_handling - return sense code for EEDP errors * @scmd: pointer to scsi command object * @ioc_status: ioc status */ static void _scsih_eedp_error_handling(struct scsi_cmnd *scmd, u16 ioc_status) { u8 ascq; switch (ioc_status) { case MPI2_IOCSTATUS_EEDP_GUARD_ERROR: ascq = 0x01; break; case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR: ascq = 0x02; break; case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR: ascq = 0x03; break; default: ascq = 0x00; break; } scsi_build_sense_buffer(0, scmd->sense_buffer, ILLEGAL_REQUEST, 0x10, ascq); scmd->result = DRIVER_SENSE << 24 | (DID_ABORT << 16) | SAM_STAT_CHECK_CONDITION; } /** * scsih_qcmd - main scsi request entry point * @shost: SCSI host pointer * @scmd: pointer to scsi command object * * The callback index is set inside `ioc->scsi_io_cb_idx`. * * Return: 0 on success. If there's a failure, return either: * SCSI_MLQUEUE_DEVICE_BUSY if the device queue is full, or * SCSI_MLQUEUE_HOST_BUSY if the entire host queue is full */ static int scsih_qcmd(struct Scsi_Host *shost, struct scsi_cmnd *scmd) { struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); struct MPT3SAS_DEVICE *sas_device_priv_data; struct MPT3SAS_TARGET *sas_target_priv_data; struct _raid_device *raid_device; struct request *rq = scmd->request; int class; Mpi25SCSIIORequest_t *mpi_request; struct _pcie_device *pcie_device = NULL; u32 mpi_control; u16 smid; u16 handle; if (ioc->logging_level & MPT_DEBUG_SCSI) scsi_print_command(scmd); sas_device_priv_data = scmd->device->hostdata; if (!sas_device_priv_data || !sas_device_priv_data->sas_target) { scmd->result = DID_NO_CONNECT << 16; scmd->scsi_done(scmd); return 0; } if (!(_scsih_allow_scmd_to_device(ioc, scmd))) { scmd->result = DID_NO_CONNECT << 16; scmd->scsi_done(scmd); return 0; } sas_target_priv_data = sas_device_priv_data->sas_target; /* invalid device handle */ handle = sas_target_priv_data->handle; if (handle == MPT3SAS_INVALID_DEVICE_HANDLE) { scmd->result = DID_NO_CONNECT << 16; scmd->scsi_done(scmd); return 0; } if (ioc->shost_recovery || ioc->ioc_link_reset_in_progress) { /* host recovery or link resets sent via IOCTLs */ return SCSI_MLQUEUE_HOST_BUSY; } else if (sas_target_priv_data->deleted) { /* device has been deleted */ scmd->result = DID_NO_CONNECT << 16; scmd->scsi_done(scmd); return 0; } else if (sas_target_priv_data->tm_busy || sas_device_priv_data->block) { /* device busy with task management */ return SCSI_MLQUEUE_DEVICE_BUSY; } /* * Bug work around for firmware SATL handling. The loop * is based on atomic operations and ensures consistency * since we're lockless at this point */ do { if (test_bit(0, &sas_device_priv_data->ata_command_pending)) return SCSI_MLQUEUE_DEVICE_BUSY; } while (_scsih_set_satl_pending(scmd, true)); if (scmd->sc_data_direction == DMA_FROM_DEVICE) mpi_control = MPI2_SCSIIO_CONTROL_READ; else if (scmd->sc_data_direction == DMA_TO_DEVICE) mpi_control = MPI2_SCSIIO_CONTROL_WRITE; else mpi_control = MPI2_SCSIIO_CONTROL_NODATATRANSFER; /* set tags */ mpi_control |= MPI2_SCSIIO_CONTROL_SIMPLEQ; /* NCQ Prio supported, make sure control indicated high priority */ if (sas_device_priv_data->ncq_prio_enable) { class = IOPRIO_PRIO_CLASS(req_get_ioprio(rq)); if (class == IOPRIO_CLASS_RT) mpi_control |= 1 << MPI2_SCSIIO_CONTROL_CMDPRI_SHIFT; } /* Make sure Device is not raid volume. * We do not expose raid functionality to upper layer for warpdrive. */ if (((!ioc->is_warpdrive && !scsih_is_raid(&scmd->device->sdev_gendev)) && !scsih_is_nvme(&scmd->device->sdev_gendev)) && sas_is_tlr_enabled(scmd->device) && scmd->cmd_len != 32) mpi_control |= MPI2_SCSIIO_CONTROL_TLR_ON; smid = mpt3sas_base_get_smid_scsiio(ioc, ioc->scsi_io_cb_idx, scmd); if (!smid) { ioc_err(ioc, "%s: failed obtaining a smid\n", __func__); _scsih_set_satl_pending(scmd, false); goto out; } mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); memset(mpi_request, 0, ioc->request_sz); _scsih_setup_eedp(ioc, scmd, mpi_request); if (scmd->cmd_len == 32) mpi_control |= 4 << MPI2_SCSIIO_CONTROL_ADDCDBLEN_SHIFT; mpi_request->Function = MPI2_FUNCTION_SCSI_IO_REQUEST; if (sas_device_priv_data->sas_target->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) mpi_request->Function = MPI2_FUNCTION_RAID_SCSI_IO_PASSTHROUGH; else mpi_request->Function = MPI2_FUNCTION_SCSI_IO_REQUEST; mpi_request->DevHandle = cpu_to_le16(handle); mpi_request->DataLength = cpu_to_le32(scsi_bufflen(scmd)); mpi_request->Control = cpu_to_le32(mpi_control); mpi_request->IoFlags = cpu_to_le16(scmd->cmd_len); mpi_request->MsgFlags = MPI2_SCSIIO_MSGFLAGS_SYSTEM_SENSE_ADDR; mpi_request->SenseBufferLength = SCSI_SENSE_BUFFERSIZE; mpi_request->SenseBufferLowAddress = mpt3sas_base_get_sense_buffer_dma(ioc, smid); mpi_request->SGLOffset0 = offsetof(Mpi25SCSIIORequest_t, SGL) / 4; int_to_scsilun(sas_device_priv_data->lun, (struct scsi_lun *) mpi_request->LUN); memcpy(mpi_request->CDB.CDB32, scmd->cmnd, scmd->cmd_len); if (mpi_request->DataLength) { pcie_device = sas_target_priv_data->pcie_dev; if (ioc->build_sg_scmd(ioc, scmd, smid, pcie_device)) { mpt3sas_base_free_smid(ioc, smid); _scsih_set_satl_pending(scmd, false); goto out; } } else ioc->build_zero_len_sge(ioc, &mpi_request->SGL); raid_device = sas_target_priv_data->raid_device; if (raid_device && raid_device->direct_io_enabled) mpt3sas_setup_direct_io(ioc, scmd, raid_device, mpi_request); if (likely(mpi_request->Function == MPI2_FUNCTION_SCSI_IO_REQUEST)) { if (sas_target_priv_data->flags & MPT_TARGET_FASTPATH_IO) { mpi_request->IoFlags = cpu_to_le16(scmd->cmd_len | MPI25_SCSIIO_IOFLAGS_FAST_PATH); ioc->put_smid_fast_path(ioc, smid, handle); } else ioc->put_smid_scsi_io(ioc, smid, le16_to_cpu(mpi_request->DevHandle)); } else ioc->put_smid_default(ioc, smid); return 0; out: return SCSI_MLQUEUE_HOST_BUSY; } /** * _scsih_normalize_sense - normalize descriptor and fixed format sense data * @sense_buffer: sense data returned by target * @data: normalized skey/asc/ascq */ static void _scsih_normalize_sense(char *sense_buffer, struct sense_info *data) { if ((sense_buffer[0] & 0x7F) >= 0x72) { /* descriptor format */ data->skey = sense_buffer[1] & 0x0F; data->asc = sense_buffer[2]; data->ascq = sense_buffer[3]; } else { /* fixed format */ data->skey = sense_buffer[2] & 0x0F; data->asc = sense_buffer[12]; data->ascq = sense_buffer[13]; } } /** * _scsih_scsi_ioc_info - translated non-succesfull SCSI_IO request * @ioc: per adapter object * @scmd: pointer to scsi command object * @mpi_reply: reply mf payload returned from firmware * @smid: ? * * scsi_status - SCSI Status code returned from target device * scsi_state - state info associated with SCSI_IO determined by ioc * ioc_status - ioc supplied status info */ static void _scsih_scsi_ioc_info(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd, Mpi2SCSIIOReply_t *mpi_reply, u16 smid) { u32 response_info; u8 *response_bytes; u16 ioc_status = le16_to_cpu(mpi_reply->IOCStatus) & MPI2_IOCSTATUS_MASK; u8 scsi_state = mpi_reply->SCSIState; u8 scsi_status = mpi_reply->SCSIStatus; char *desc_ioc_state = NULL; char *desc_scsi_status = NULL; char *desc_scsi_state = ioc->tmp_string; u32 log_info = le32_to_cpu(mpi_reply->IOCLogInfo); struct _sas_device *sas_device = NULL; struct _pcie_device *pcie_device = NULL; struct scsi_target *starget = scmd->device->sdev_target; struct MPT3SAS_TARGET *priv_target = starget->hostdata; char *device_str = NULL; if (!priv_target) return; if (ioc->hide_ir_msg) device_str = "WarpDrive"; else device_str = "volume"; if (log_info == 0x31170000) return; switch (ioc_status) { case MPI2_IOCSTATUS_SUCCESS: desc_ioc_state = "success"; break; case MPI2_IOCSTATUS_INVALID_FUNCTION: desc_ioc_state = "invalid function"; break; case MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR: desc_ioc_state = "scsi recovered error"; break; case MPI2_IOCSTATUS_SCSI_INVALID_DEVHANDLE: desc_ioc_state = "scsi invalid dev handle"; break; case MPI2_IOCSTATUS_SCSI_DEVICE_NOT_THERE: desc_ioc_state = "scsi device not there"; break; case MPI2_IOCSTATUS_SCSI_DATA_OVERRUN: desc_ioc_state = "scsi data overrun"; break; case MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN: desc_ioc_state = "scsi data underrun"; break; case MPI2_IOCSTATUS_SCSI_IO_DATA_ERROR: desc_ioc_state = "scsi io data error"; break; case MPI2_IOCSTATUS_SCSI_PROTOCOL_ERROR: desc_ioc_state = "scsi protocol error"; break; case MPI2_IOCSTATUS_SCSI_TASK_TERMINATED: desc_ioc_state = "scsi task terminated"; break; case MPI2_IOCSTATUS_SCSI_RESIDUAL_MISMATCH: desc_ioc_state = "scsi residual mismatch"; break; case MPI2_IOCSTATUS_SCSI_TASK_MGMT_FAILED: desc_ioc_state = "scsi task mgmt failed"; break; case MPI2_IOCSTATUS_SCSI_IOC_TERMINATED: desc_ioc_state = "scsi ioc terminated"; break; case MPI2_IOCSTATUS_SCSI_EXT_TERMINATED: desc_ioc_state = "scsi ext terminated"; break; case MPI2_IOCSTATUS_EEDP_GUARD_ERROR: desc_ioc_state = "eedp guard error"; break; case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR: desc_ioc_state = "eedp ref tag error"; break; case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR: desc_ioc_state = "eedp app tag error"; break; case MPI2_IOCSTATUS_INSUFFICIENT_POWER: desc_ioc_state = "insufficient power"; break; default: desc_ioc_state = "unknown"; break; } switch (scsi_status) { case MPI2_SCSI_STATUS_GOOD: desc_scsi_status = "good"; break; case MPI2_SCSI_STATUS_CHECK_CONDITION: desc_scsi_status = "check condition"; break; case MPI2_SCSI_STATUS_CONDITION_MET: desc_scsi_status = "condition met"; break; case MPI2_SCSI_STATUS_BUSY: desc_scsi_status = "busy"; break; case MPI2_SCSI_STATUS_INTERMEDIATE: desc_scsi_status = "intermediate"; break; case MPI2_SCSI_STATUS_INTERMEDIATE_CONDMET: desc_scsi_status = "intermediate condmet"; break; case MPI2_SCSI_STATUS_RESERVATION_CONFLICT: desc_scsi_status = "reservation conflict"; break; case MPI2_SCSI_STATUS_COMMAND_TERMINATED: desc_scsi_status = "command terminated"; break; case MPI2_SCSI_STATUS_TASK_SET_FULL: desc_scsi_status = "task set full"; break; case MPI2_SCSI_STATUS_ACA_ACTIVE: desc_scsi_status = "aca active"; break; case MPI2_SCSI_STATUS_TASK_ABORTED: desc_scsi_status = "task aborted"; break; default: desc_scsi_status = "unknown"; break; } desc_scsi_state[0] = '\0'; if (!scsi_state) desc_scsi_state = " "; if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) strcat(desc_scsi_state, "response info "); if (scsi_state & MPI2_SCSI_STATE_TERMINATED) strcat(desc_scsi_state, "state terminated "); if (scsi_state & MPI2_SCSI_STATE_NO_SCSI_STATUS) strcat(desc_scsi_state, "no status "); if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_FAILED) strcat(desc_scsi_state, "autosense failed "); if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) strcat(desc_scsi_state, "autosense valid "); scsi_print_command(scmd); if (priv_target->flags & MPT_TARGET_FLAGS_VOLUME) { ioc_warn(ioc, "\t%s wwid(0x%016llx)\n", device_str, (u64)priv_target->sas_address); } else if (priv_target->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) { pcie_device = mpt3sas_get_pdev_from_target(ioc, priv_target); if (pcie_device) { ioc_info(ioc, "\twwid(0x%016llx), port(%d)\n", (u64)pcie_device->wwid, pcie_device->port_num); if (pcie_device->enclosure_handle != 0) ioc_info(ioc, "\tenclosure logical id(0x%016llx), slot(%d)\n", (u64)pcie_device->enclosure_logical_id, pcie_device->slot); if (pcie_device->connector_name[0]) ioc_info(ioc, "\tenclosure level(0x%04x), connector name( %s)\n", pcie_device->enclosure_level, pcie_device->connector_name); pcie_device_put(pcie_device); } } else { sas_device = mpt3sas_get_sdev_from_target(ioc, priv_target); if (sas_device) { ioc_warn(ioc, "\tsas_address(0x%016llx), phy(%d)\n", (u64)sas_device->sas_address, sas_device->phy); _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL); sas_device_put(sas_device); } } ioc_warn(ioc, "\thandle(0x%04x), ioc_status(%s)(0x%04x), smid(%d)\n", le16_to_cpu(mpi_reply->DevHandle), desc_ioc_state, ioc_status, smid); ioc_warn(ioc, "\trequest_len(%d), underflow(%d), resid(%d)\n", scsi_bufflen(scmd), scmd->underflow, scsi_get_resid(scmd)); ioc_warn(ioc, "\ttag(%d), transfer_count(%d), sc->result(0x%08x)\n", le16_to_cpu(mpi_reply->TaskTag), le32_to_cpu(mpi_reply->TransferCount), scmd->result); ioc_warn(ioc, "\tscsi_status(%s)(0x%02x), scsi_state(%s)(0x%02x)\n", desc_scsi_status, scsi_status, desc_scsi_state, scsi_state); if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) { struct sense_info data; _scsih_normalize_sense(scmd->sense_buffer, &data); ioc_warn(ioc, "\t[sense_key,asc,ascq]: [0x%02x,0x%02x,0x%02x], count(%d)\n", data.skey, data.asc, data.ascq, le32_to_cpu(mpi_reply->SenseCount)); } if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) { response_info = le32_to_cpu(mpi_reply->ResponseInfo); response_bytes = (u8 *)&response_info; _scsih_response_code(ioc, response_bytes[0]); } } /** * _scsih_turn_on_pfa_led - illuminate PFA LED * @ioc: per adapter object * @handle: device handle * Context: process */ static void _scsih_turn_on_pfa_led(struct MPT3SAS_ADAPTER *ioc, u16 handle) { Mpi2SepReply_t mpi_reply; Mpi2SepRequest_t mpi_request; struct _sas_device *sas_device; sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); if (!sas_device) return; memset(&mpi_request, 0, sizeof(Mpi2SepRequest_t)); mpi_request.Function = MPI2_FUNCTION_SCSI_ENCLOSURE_PROCESSOR; mpi_request.Action = MPI2_SEP_REQ_ACTION_WRITE_STATUS; mpi_request.SlotStatus = cpu_to_le32(MPI2_SEP_REQ_SLOTSTATUS_PREDICTED_FAULT); mpi_request.DevHandle = cpu_to_le16(handle); mpi_request.Flags = MPI2_SEP_REQ_FLAGS_DEVHANDLE_ADDRESS; if ((mpt3sas_base_scsi_enclosure_processor(ioc, &mpi_reply, &mpi_request)) != 0) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } sas_device->pfa_led_on = 1; if (mpi_reply.IOCStatus || mpi_reply.IOCLogInfo) { dewtprintk(ioc, ioc_info(ioc, "enclosure_processor: ioc_status (0x%04x), loginfo(0x%08x)\n", le16_to_cpu(mpi_reply.IOCStatus), le32_to_cpu(mpi_reply.IOCLogInfo))); goto out; } out: sas_device_put(sas_device); } /** * _scsih_turn_off_pfa_led - turn off Fault LED * @ioc: per adapter object * @sas_device: sas device whose PFA LED has to turned off * Context: process */ static void _scsih_turn_off_pfa_led(struct MPT3SAS_ADAPTER *ioc, struct _sas_device *sas_device) { Mpi2SepReply_t mpi_reply; Mpi2SepRequest_t mpi_request; memset(&mpi_request, 0, sizeof(Mpi2SepRequest_t)); mpi_request.Function = MPI2_FUNCTION_SCSI_ENCLOSURE_PROCESSOR; mpi_request.Action = MPI2_SEP_REQ_ACTION_WRITE_STATUS; mpi_request.SlotStatus = 0; mpi_request.Slot = cpu_to_le16(sas_device->slot); mpi_request.DevHandle = 0; mpi_request.EnclosureHandle = cpu_to_le16(sas_device->enclosure_handle); mpi_request.Flags = MPI2_SEP_REQ_FLAGS_ENCLOSURE_SLOT_ADDRESS; if ((mpt3sas_base_scsi_enclosure_processor(ioc, &mpi_reply, &mpi_request)) != 0) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } if (mpi_reply.IOCStatus || mpi_reply.IOCLogInfo) { dewtprintk(ioc, ioc_info(ioc, "enclosure_processor: ioc_status (0x%04x), loginfo(0x%08x)\n", le16_to_cpu(mpi_reply.IOCStatus), le32_to_cpu(mpi_reply.IOCLogInfo))); return; } } /** * _scsih_send_event_to_turn_on_pfa_led - fire delayed event * @ioc: per adapter object * @handle: device handle * Context: interrupt. */ static void _scsih_send_event_to_turn_on_pfa_led(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct fw_event_work *fw_event; fw_event = alloc_fw_event_work(0); if (!fw_event) return; fw_event->event = MPT3SAS_TURN_ON_PFA_LED; fw_event->device_handle = handle; fw_event->ioc = ioc; _scsih_fw_event_add(ioc, fw_event); fw_event_work_put(fw_event); } /** * _scsih_smart_predicted_fault - process smart errors * @ioc: per adapter object * @handle: device handle * Context: interrupt. */ static void _scsih_smart_predicted_fault(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct scsi_target *starget; struct MPT3SAS_TARGET *sas_target_priv_data; Mpi2EventNotificationReply_t *event_reply; Mpi2EventDataSasDeviceStatusChange_t *event_data; struct _sas_device *sas_device; ssize_t sz; unsigned long flags; /* only handle non-raid devices */ spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); if (!sas_device) goto out_unlock; starget = sas_device->starget; sas_target_priv_data = starget->hostdata; if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) || ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME))) goto out_unlock; _scsih_display_enclosure_chassis_info(NULL, sas_device, NULL, starget); spin_unlock_irqrestore(&ioc->sas_device_lock, flags); if (ioc->pdev->subsystem_vendor == PCI_VENDOR_ID_IBM) _scsih_send_event_to_turn_on_pfa_led(ioc, handle); /* insert into event log */ sz = offsetof(Mpi2EventNotificationReply_t, EventData) + sizeof(Mpi2EventDataSasDeviceStatusChange_t); event_reply = kzalloc(sz, GFP_ATOMIC); if (!event_reply) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } event_reply->Function = MPI2_FUNCTION_EVENT_NOTIFICATION; event_reply->Event = cpu_to_le16(MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE); event_reply->MsgLength = sz/4; event_reply->EventDataLength = cpu_to_le16(sizeof(Mpi2EventDataSasDeviceStatusChange_t)/4); event_data = (Mpi2EventDataSasDeviceStatusChange_t *) event_reply->EventData; event_data->ReasonCode = MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA; event_data->ASC = 0x5D; event_data->DevHandle = cpu_to_le16(handle); event_data->SASAddress = cpu_to_le64(sas_target_priv_data->sas_address); mpt3sas_ctl_add_to_event_log(ioc, event_reply); kfree(event_reply); out: if (sas_device) sas_device_put(sas_device); return; out_unlock: spin_unlock_irqrestore(&ioc->sas_device_lock, flags); goto out; } /** * _scsih_io_done - scsi request callback * @ioc: per adapter object * @smid: system request message index * @msix_index: MSIX table index supplied by the OS * @reply: reply message frame(lower 32bit addr) * * Callback handler when using _scsih_qcmd. * * Return: 1 meaning mf should be freed from _base_interrupt * 0 means the mf is freed from this function. */ static u8 _scsih_io_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply) { Mpi25SCSIIORequest_t *mpi_request; Mpi2SCSIIOReply_t *mpi_reply; struct scsi_cmnd *scmd; struct scsiio_tracker *st; u16 ioc_status; u32 xfer_cnt; u8 scsi_state; u8 scsi_status; u32 log_info; struct MPT3SAS_DEVICE *sas_device_priv_data; u32 response_code = 0; mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply); scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); if (scmd == NULL) return 1; _scsih_set_satl_pending(scmd, false); mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); if (mpi_reply == NULL) { scmd->result = DID_OK << 16; goto out; } sas_device_priv_data = scmd->device->hostdata; if (!sas_device_priv_data || !sas_device_priv_data->sas_target || sas_device_priv_data->sas_target->deleted) { scmd->result = DID_NO_CONNECT << 16; goto out; } ioc_status = le16_to_cpu(mpi_reply->IOCStatus); /* * WARPDRIVE: If direct_io is set then it is directIO, * the failed direct I/O should be redirected to volume */ st = scsi_cmd_priv(scmd); if (st->direct_io && ((ioc_status & MPI2_IOCSTATUS_MASK) != MPI2_IOCSTATUS_SCSI_TASK_TERMINATED)) { st->direct_io = 0; st->scmd = scmd; memcpy(mpi_request->CDB.CDB32, scmd->cmnd, scmd->cmd_len); mpi_request->DevHandle = cpu_to_le16(sas_device_priv_data->sas_target->handle); ioc->put_smid_scsi_io(ioc, smid, sas_device_priv_data->sas_target->handle); return 0; } /* turning off TLR */ scsi_state = mpi_reply->SCSIState; if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) response_code = le32_to_cpu(mpi_reply->ResponseInfo) & 0xFF; if (!sas_device_priv_data->tlr_snoop_check) { sas_device_priv_data->tlr_snoop_check++; if ((!ioc->is_warpdrive && !scsih_is_raid(&scmd->device->sdev_gendev) && !scsih_is_nvme(&scmd->device->sdev_gendev)) && sas_is_tlr_enabled(scmd->device) && response_code == MPI2_SCSITASKMGMT_RSP_INVALID_FRAME) { sas_disable_tlr(scmd->device); sdev_printk(KERN_INFO, scmd->device, "TLR disabled\n"); } } xfer_cnt = le32_to_cpu(mpi_reply->TransferCount); scsi_set_resid(scmd, scsi_bufflen(scmd) - xfer_cnt); if (ioc_status & MPI2_IOCSTATUS_FLAG_LOG_INFO_AVAILABLE) log_info = le32_to_cpu(mpi_reply->IOCLogInfo); else log_info = 0; ioc_status &= MPI2_IOCSTATUS_MASK; scsi_status = mpi_reply->SCSIStatus; if (ioc_status == MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN && xfer_cnt == 0 && (scsi_status == MPI2_SCSI_STATUS_BUSY || scsi_status == MPI2_SCSI_STATUS_RESERVATION_CONFLICT || scsi_status == MPI2_SCSI_STATUS_TASK_SET_FULL)) { ioc_status = MPI2_IOCSTATUS_SUCCESS; } if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) { struct sense_info data; const void *sense_data = mpt3sas_base_get_sense_buffer(ioc, smid); u32 sz = min_t(u32, SCSI_SENSE_BUFFERSIZE, le32_to_cpu(mpi_reply->SenseCount)); memcpy(scmd->sense_buffer, sense_data, sz); _scsih_normalize_sense(scmd->sense_buffer, &data); /* failure prediction threshold exceeded */ if (data.asc == 0x5D) _scsih_smart_predicted_fault(ioc, le16_to_cpu(mpi_reply->DevHandle)); mpt3sas_trigger_scsi(ioc, data.skey, data.asc, data.ascq); if ((ioc->logging_level & MPT_DEBUG_REPLY) && ((scmd->sense_buffer[2] == UNIT_ATTENTION) || (scmd->sense_buffer[2] == MEDIUM_ERROR) || (scmd->sense_buffer[2] == HARDWARE_ERROR))) _scsih_scsi_ioc_info(ioc, scmd, mpi_reply, smid); } switch (ioc_status) { case MPI2_IOCSTATUS_BUSY: case MPI2_IOCSTATUS_INSUFFICIENT_RESOURCES: scmd->result = SAM_STAT_BUSY; break; case MPI2_IOCSTATUS_SCSI_DEVICE_NOT_THERE: scmd->result = DID_NO_CONNECT << 16; break; case MPI2_IOCSTATUS_SCSI_IOC_TERMINATED: if (sas_device_priv_data->block) { scmd->result = DID_TRANSPORT_DISRUPTED << 16; goto out; } if (log_info == 0x31110630) { if (scmd->retries > 2) { scmd->result = DID_NO_CONNECT << 16; scsi_device_set_state(scmd->device, SDEV_OFFLINE); } else { scmd->result = DID_SOFT_ERROR << 16; scmd->device->expecting_cc_ua = 1; } break; } else if (log_info == VIRTUAL_IO_FAILED_RETRY) { scmd->result = DID_RESET << 16; break; } else if ((scmd->device->channel == RAID_CHANNEL) && (scsi_state == (MPI2_SCSI_STATE_TERMINATED | MPI2_SCSI_STATE_NO_SCSI_STATUS))) { scmd->result = DID_RESET << 16; break; } scmd->result = DID_SOFT_ERROR << 16; break; case MPI2_IOCSTATUS_SCSI_TASK_TERMINATED: case MPI2_IOCSTATUS_SCSI_EXT_TERMINATED: scmd->result = DID_RESET << 16; break; case MPI2_IOCSTATUS_SCSI_RESIDUAL_MISMATCH: if ((xfer_cnt == 0) || (scmd->underflow > xfer_cnt)) scmd->result = DID_SOFT_ERROR << 16; else scmd->result = (DID_OK << 16) | scsi_status; break; case MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN: scmd->result = (DID_OK << 16) | scsi_status; if ((scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID)) break; if (xfer_cnt < scmd->underflow) { if (scsi_status == SAM_STAT_BUSY) scmd->result = SAM_STAT_BUSY; else scmd->result = DID_SOFT_ERROR << 16; } else if (scsi_state & (MPI2_SCSI_STATE_AUTOSENSE_FAILED | MPI2_SCSI_STATE_NO_SCSI_STATUS)) scmd->result = DID_SOFT_ERROR << 16; else if (scsi_state & MPI2_SCSI_STATE_TERMINATED) scmd->result = DID_RESET << 16; else if (!xfer_cnt && scmd->cmnd[0] == REPORT_LUNS) { mpi_reply->SCSIState = MPI2_SCSI_STATE_AUTOSENSE_VALID; mpi_reply->SCSIStatus = SAM_STAT_CHECK_CONDITION; scmd->result = (DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION; scmd->sense_buffer[0] = 0x70; scmd->sense_buffer[2] = ILLEGAL_REQUEST; scmd->sense_buffer[12] = 0x20; scmd->sense_buffer[13] = 0; } break; case MPI2_IOCSTATUS_SCSI_DATA_OVERRUN: scsi_set_resid(scmd, 0); fallthrough; case MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR: case MPI2_IOCSTATUS_SUCCESS: scmd->result = (DID_OK << 16) | scsi_status; if (response_code == MPI2_SCSITASKMGMT_RSP_INVALID_FRAME || (scsi_state & (MPI2_SCSI_STATE_AUTOSENSE_FAILED | MPI2_SCSI_STATE_NO_SCSI_STATUS))) scmd->result = DID_SOFT_ERROR << 16; else if (scsi_state & MPI2_SCSI_STATE_TERMINATED) scmd->result = DID_RESET << 16; break; case MPI2_IOCSTATUS_EEDP_GUARD_ERROR: case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR: case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR: _scsih_eedp_error_handling(scmd, ioc_status); break; case MPI2_IOCSTATUS_SCSI_PROTOCOL_ERROR: case MPI2_IOCSTATUS_INVALID_FUNCTION: case MPI2_IOCSTATUS_INVALID_SGL: case MPI2_IOCSTATUS_INTERNAL_ERROR: case MPI2_IOCSTATUS_INVALID_FIELD: case MPI2_IOCSTATUS_INVALID_STATE: case MPI2_IOCSTATUS_SCSI_IO_DATA_ERROR: case MPI2_IOCSTATUS_SCSI_TASK_MGMT_FAILED: case MPI2_IOCSTATUS_INSUFFICIENT_POWER: default: scmd->result = DID_SOFT_ERROR << 16; break; } if (scmd->result && (ioc->logging_level & MPT_DEBUG_REPLY)) _scsih_scsi_ioc_info(ioc , scmd, mpi_reply, smid); out: scsi_dma_unmap(scmd); mpt3sas_base_free_smid(ioc, smid); scmd->scsi_done(scmd); return 0; } /** * _scsih_update_vphys_after_reset - update the Port's * vphys_list after reset * @ioc: per adapter object * * Returns nothing. */ static void _scsih_update_vphys_after_reset(struct MPT3SAS_ADAPTER *ioc) { u16 sz, ioc_status; int i; Mpi2ConfigReply_t mpi_reply; Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL; u16 attached_handle; u64 attached_sas_addr; u8 found = 0, port_id; Mpi2SasPhyPage0_t phy_pg0; struct hba_port *port, *port_next, *mport; struct virtual_phy *vphy, *vphy_next; struct _sas_device *sas_device; /* * Mark all the vphys objects as dirty. */ list_for_each_entry_safe(port, port_next, &ioc->port_table_list, list) { if (!port->vphys_mask) continue; list_for_each_entry_safe(vphy, vphy_next, &port->vphys_list, list) { vphy->flags |= MPT_VPHY_FLAG_DIRTY_PHY; } } /* * Read SASIOUnitPage0 to get each HBA Phy's data. */ sz = offsetof(Mpi2SasIOUnitPage0_t, PhyData) + (ioc->sas_hba.num_phys * sizeof(Mpi2SasIOUnit0PhyData_t)); sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL); if (!sas_iounit_pg0) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply, sas_iounit_pg0, sz)) != 0) goto out; ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) goto out; /* * Loop over each HBA Phy. */ for (i = 0; i < ioc->sas_hba.num_phys; i++) { /* * Check whether Phy's Negotiation Link Rate is > 1.5G or not. */ if ((sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4) < MPI2_SAS_NEG_LINK_RATE_1_5) continue; /* * Check whether Phy is connected to SEP device or not, * if it is SEP device then read the Phy's SASPHYPage0 data to * determine whether Phy is a virtual Phy or not. if it is * virtual phy then it is conformed that the attached remote * device is a HBA's vSES device. */ if (!(le32_to_cpu( sas_iounit_pg0->PhyData[i].ControllerPhyDeviceInfo) & MPI2_SAS_DEVICE_INFO_SEP)) continue; if ((mpt3sas_config_get_phy_pg0(ioc, &mpi_reply, &phy_pg0, i))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); continue; } if (!(le32_to_cpu(phy_pg0.PhyInfo) & MPI2_SAS_PHYINFO_VIRTUAL_PHY)) continue; /* * Get the vSES device's SAS Address. */ attached_handle = le16_to_cpu( sas_iounit_pg0->PhyData[i].AttachedDevHandle); if (_scsih_get_sas_address(ioc, attached_handle, &attached_sas_addr) != 0) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); continue; } found = 0; port = port_next = NULL; /* * Loop over each virtual_phy object from * each port's vphys_list. */ list_for_each_entry_safe(port, port_next, &ioc->port_table_list, list) { if (!port->vphys_mask) continue; list_for_each_entry_safe(vphy, vphy_next, &port->vphys_list, list) { /* * Continue with next virtual_phy object * if the object is not marked as dirty. */ if (!(vphy->flags & MPT_VPHY_FLAG_DIRTY_PHY)) continue; /* * Continue with next virtual_phy object * if the object's SAS Address is not equals * to current Phy's vSES device SAS Address. */ if (vphy->sas_address != attached_sas_addr) continue; /* * Enable current Phy number bit in object's * phy_mask field. */ if (!(vphy->phy_mask & (1 << i))) vphy->phy_mask = (1 << i); /* * Get hba_port object from hba_port table * corresponding to current phy's Port ID. * if there is no hba_port object corresponding * to Phy's Port ID then create a new hba_port * object & add to hba_port table. */ port_id = sas_iounit_pg0->PhyData[i].Port; mport = mpt3sas_get_port_by_id(ioc, port_id, 1); if (!mport) { mport = kzalloc( sizeof(struct hba_port), GFP_KERNEL); if (!mport) break; mport->port_id = port_id; ioc_info(ioc, "%s: hba_port entry: %p, port: %d is added to hba_port list\n", __func__, mport, mport->port_id); list_add_tail(&mport->list, &ioc->port_table_list); } /* * If mport & port pointers are not pointing to * same hba_port object then it means that vSES * device's Port ID got changed after reset and * hence move current virtual_phy object from * port's vphys_list to mport's vphys_list. */ if (port != mport) { if (!mport->vphys_mask) INIT_LIST_HEAD( &mport->vphys_list); mport->vphys_mask |= (1 << i); port->vphys_mask &= ~(1 << i); list_move(&vphy->list, &mport->vphys_list); sas_device = mpt3sas_get_sdev_by_addr( ioc, attached_sas_addr, port); if (sas_device) sas_device->port = mport; } /* * Earlier while updating the hba_port table, * it is determined that there is no other * direct attached device with mport's Port ID, * Hence mport was marked as dirty. Only vSES * device has this Port ID, so unmark the mport * as dirt. */ if (mport->flags & HBA_PORT_FLAG_DIRTY_PORT) { mport->sas_address = 0; mport->phy_mask = 0; mport->flags &= ~HBA_PORT_FLAG_DIRTY_PORT; } /* * Unmark current virtual_phy object as dirty. */ vphy->flags &= ~MPT_VPHY_FLAG_DIRTY_PHY; found = 1; break; } if (found) break; } } out: kfree(sas_iounit_pg0); } /** * _scsih_get_port_table_after_reset - Construct temporary port table * @ioc: per adapter object * @port_table: address where port table needs to be constructed * * return number of HBA port entries available after reset. */ static int _scsih_get_port_table_after_reset(struct MPT3SAS_ADAPTER *ioc, struct hba_port *port_table) { u16 sz, ioc_status; int i, j; Mpi2ConfigReply_t mpi_reply; Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL; u16 attached_handle; u64 attached_sas_addr; u8 found = 0, port_count = 0, port_id; sz = offsetof(Mpi2SasIOUnitPage0_t, PhyData) + (ioc->sas_hba.num_phys * sizeof(Mpi2SasIOUnit0PhyData_t)); sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL); if (!sas_iounit_pg0) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return port_count; } if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply, sas_iounit_pg0, sz)) != 0) goto out; ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) goto out; for (i = 0; i < ioc->sas_hba.num_phys; i++) { found = 0; if ((sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4) < MPI2_SAS_NEG_LINK_RATE_1_5) continue; attached_handle = le16_to_cpu(sas_iounit_pg0->PhyData[i].AttachedDevHandle); if (_scsih_get_sas_address( ioc, attached_handle, &attached_sas_addr) != 0) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); continue; } for (j = 0; j < port_count; j++) { port_id = sas_iounit_pg0->PhyData[i].Port; if (port_table[j].port_id == port_id && port_table[j].sas_address == attached_sas_addr) { port_table[j].phy_mask |= (1 << i); found = 1; break; } } if (found) continue; port_id = sas_iounit_pg0->PhyData[i].Port; port_table[port_count].port_id = port_id; port_table[port_count].phy_mask = (1 << i); port_table[port_count].sas_address = attached_sas_addr; port_count++; } out: kfree(sas_iounit_pg0); return port_count; } enum hba_port_matched_codes { NOT_MATCHED = 0, MATCHED_WITH_ADDR_AND_PHYMASK, MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT, MATCHED_WITH_ADDR_AND_SUBPHYMASK, MATCHED_WITH_ADDR, }; /** * _scsih_look_and_get_matched_port_entry - Get matched hba port entry * from HBA port table * @ioc: per adapter object * @port_entry: hba port entry from temporary port table which needs to be * searched for matched entry in the HBA port table * @matched_port_entry: save matched hba port entry here * @count: count of matched entries * * return type of matched entry found. */ static enum hba_port_matched_codes _scsih_look_and_get_matched_port_entry(struct MPT3SAS_ADAPTER *ioc, struct hba_port *port_entry, struct hba_port **matched_port_entry, int *count) { struct hba_port *port_table_entry, *matched_port = NULL; enum hba_port_matched_codes matched_code = NOT_MATCHED; int lcount = 0; *matched_port_entry = NULL; list_for_each_entry(port_table_entry, &ioc->port_table_list, list) { if (!(port_table_entry->flags & HBA_PORT_FLAG_DIRTY_PORT)) continue; if ((port_table_entry->sas_address == port_entry->sas_address) && (port_table_entry->phy_mask == port_entry->phy_mask)) { matched_code = MATCHED_WITH_ADDR_AND_PHYMASK; matched_port = port_table_entry; break; } if ((port_table_entry->sas_address == port_entry->sas_address) && (port_table_entry->phy_mask & port_entry->phy_mask) && (port_table_entry->port_id == port_entry->port_id)) { matched_code = MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT; matched_port = port_table_entry; continue; } if ((port_table_entry->sas_address == port_entry->sas_address) && (port_table_entry->phy_mask & port_entry->phy_mask)) { if (matched_code == MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT) continue; matched_code = MATCHED_WITH_ADDR_AND_SUBPHYMASK; matched_port = port_table_entry; continue; } if (port_table_entry->sas_address == port_entry->sas_address) { if (matched_code == MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT) continue; if (matched_code == MATCHED_WITH_ADDR_AND_SUBPHYMASK) continue; matched_code = MATCHED_WITH_ADDR; matched_port = port_table_entry; lcount++; } } *matched_port_entry = matched_port; if (matched_code == MATCHED_WITH_ADDR) *count = lcount; return matched_code; } /** * _scsih_del_phy_part_of_anther_port - remove phy if it * is a part of anther port *@ioc: per adapter object *@port_table: port table after reset *@index: hba port entry index *@port_count: number of ports available after host reset *@offset: HBA phy bit offset * */ static void _scsih_del_phy_part_of_anther_port(struct MPT3SAS_ADAPTER *ioc, struct hba_port *port_table, int index, u8 port_count, int offset) { struct _sas_node *sas_node = &ioc->sas_hba; u32 i, found = 0; for (i = 0; i < port_count; i++) { if (i == index) continue; if (port_table[i].phy_mask & (1 << offset)) { mpt3sas_transport_del_phy_from_an_existing_port( ioc, sas_node, &sas_node->phy[offset]); found = 1; break; } } if (!found) port_table[index].phy_mask |= (1 << offset); } /** * _scsih_add_or_del_phys_from_existing_port - add/remove phy to/from * right port *@ioc: per adapter object *@hba_port_entry: hba port table entry *@port_table: temporary port table *@index: hba port entry index *@port_count: number of ports available after host reset * */ static void _scsih_add_or_del_phys_from_existing_port(struct MPT3SAS_ADAPTER *ioc, struct hba_port *hba_port_entry, struct hba_port *port_table, int index, int port_count) { u32 phy_mask, offset = 0; struct _sas_node *sas_node = &ioc->sas_hba; phy_mask = hba_port_entry->phy_mask ^ port_table[index].phy_mask; for (offset = 0; offset < ioc->sas_hba.num_phys; offset++) { if (phy_mask & (1 << offset)) { if (!(port_table[index].phy_mask & (1 << offset))) { _scsih_del_phy_part_of_anther_port( ioc, port_table, index, port_count, offset); continue; } if (sas_node->phy[offset].phy_belongs_to_port) mpt3sas_transport_del_phy_from_an_existing_port( ioc, sas_node, &sas_node->phy[offset]); mpt3sas_transport_add_phy_to_an_existing_port( ioc, sas_node, &sas_node->phy[offset], hba_port_entry->sas_address, hba_port_entry); } } } /** * _scsih_del_dirty_vphy - delete virtual_phy objects marked as dirty. * @ioc: per adapter object * * Returns nothing. */ static void _scsih_del_dirty_vphy(struct MPT3SAS_ADAPTER *ioc) { struct hba_port *port, *port_next; struct virtual_phy *vphy, *vphy_next; list_for_each_entry_safe(port, port_next, &ioc->port_table_list, list) { if (!port->vphys_mask) continue; list_for_each_entry_safe(vphy, vphy_next, &port->vphys_list, list) { if (vphy->flags & MPT_VPHY_FLAG_DIRTY_PHY) { drsprintk(ioc, ioc_info(ioc, "Deleting vphy %p entry from port id: %d\t, Phy_mask 0x%08x\n", vphy, port->port_id, vphy->phy_mask)); port->vphys_mask &= ~vphy->phy_mask; list_del(&vphy->list); kfree(vphy); } } if (!port->vphys_mask && !port->sas_address) port->flags |= HBA_PORT_FLAG_DIRTY_PORT; } } /** * _scsih_del_dirty_port_entries - delete dirty port entries from port list * after host reset *@ioc: per adapter object * */ static void _scsih_del_dirty_port_entries(struct MPT3SAS_ADAPTER *ioc) { struct hba_port *port, *port_next; list_for_each_entry_safe(port, port_next, &ioc->port_table_list, list) { if (!(port->flags & HBA_PORT_FLAG_DIRTY_PORT) || port->flags & HBA_PORT_FLAG_NEW_PORT) continue; drsprintk(ioc, ioc_info(ioc, "Deleting port table entry %p having Port: %d\t Phy_mask 0x%08x\n", port, port->port_id, port->phy_mask)); list_del(&port->list); kfree(port); } } /** * _scsih_sas_port_refresh - Update HBA port table after host reset * @ioc: per adapter object */ static void _scsih_sas_port_refresh(struct MPT3SAS_ADAPTER *ioc) { u32 port_count = 0; struct hba_port *port_table; struct hba_port *port_table_entry; struct hba_port *port_entry = NULL; int i, j, count = 0, lcount = 0; int ret; u64 sas_addr; drsprintk(ioc, ioc_info(ioc, "updating ports for sas_host(0x%016llx)\n", (unsigned long long)ioc->sas_hba.sas_address)); port_table = kcalloc(ioc->sas_hba.num_phys, sizeof(struct hba_port), GFP_KERNEL); if (!port_table) return; port_count = _scsih_get_port_table_after_reset(ioc, port_table); if (!port_count) return; drsprintk(ioc, ioc_info(ioc, "New Port table\n")); for (j = 0; j < port_count; j++) drsprintk(ioc, ioc_info(ioc, "Port: %d\t Phy_mask 0x%08x\t sas_addr(0x%016llx)\n", port_table[j].port_id, port_table[j].phy_mask, port_table[j].sas_address)); list_for_each_entry(port_table_entry, &ioc->port_table_list, list) port_table_entry->flags |= HBA_PORT_FLAG_DIRTY_PORT; drsprintk(ioc, ioc_info(ioc, "Old Port table\n")); port_table_entry = NULL; list_for_each_entry(port_table_entry, &ioc->port_table_list, list) { drsprintk(ioc, ioc_info(ioc, "Port: %d\t Phy_mask 0x%08x\t sas_addr(0x%016llx)\n", port_table_entry->port_id, port_table_entry->phy_mask, port_table_entry->sas_address)); } for (j = 0; j < port_count; j++) { ret = _scsih_look_and_get_matched_port_entry(ioc, &port_table[j], &port_entry, &count); if (!port_entry) { drsprintk(ioc, ioc_info(ioc, "No Matched entry for sas_addr(0x%16llx), Port:%d\n", port_table[j].sas_address, port_table[j].port_id)); continue; } switch (ret) { case MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT: case MATCHED_WITH_ADDR_AND_SUBPHYMASK: _scsih_add_or_del_phys_from_existing_port(ioc, port_entry, port_table, j, port_count); break; case MATCHED_WITH_ADDR: sas_addr = port_table[j].sas_address; for (i = 0; i < port_count; i++) { if (port_table[i].sas_address == sas_addr) lcount++; } if (count > 1 || lcount > 1) port_entry = NULL; else _scsih_add_or_del_phys_from_existing_port(ioc, port_entry, port_table, j, port_count); } if (!port_entry) continue; if (port_entry->port_id != port_table[j].port_id) port_entry->port_id = port_table[j].port_id; port_entry->flags &= ~HBA_PORT_FLAG_DIRTY_PORT; port_entry->phy_mask = port_table[j].phy_mask; } port_table_entry = NULL; } /** * _scsih_alloc_vphy - allocate virtual_phy object * @ioc: per adapter object * @port_id: Port ID number * @phy_num: HBA Phy number * * Returns allocated virtual_phy object. */ static struct virtual_phy * _scsih_alloc_vphy(struct MPT3SAS_ADAPTER *ioc, u8 port_id, u8 phy_num) { struct virtual_phy *vphy; struct hba_port *port; port = mpt3sas_get_port_by_id(ioc, port_id, 0); if (!port) return NULL; vphy = mpt3sas_get_vphy_by_phy(ioc, port, phy_num); if (!vphy) { vphy = kzalloc(sizeof(struct virtual_phy), GFP_KERNEL); if (!vphy) return NULL; if (!port->vphys_mask) INIT_LIST_HEAD(&port->vphys_list); /* * Enable bit corresponding to HBA phy number on its * parent hba_port object's vphys_mask field. */ port->vphys_mask |= (1 << phy_num); vphy->phy_mask |= (1 << phy_num); list_add_tail(&vphy->list, &port->vphys_list); ioc_info(ioc, "vphy entry: %p, port id: %d, phy:%d is added to port's vphys_list\n", vphy, port->port_id, phy_num); } return vphy; } /** * _scsih_sas_host_refresh - refreshing sas host object contents * @ioc: per adapter object * Context: user * * During port enable, fw will send topology events for every device. Its * possible that the handles may change from the previous setting, so this * code keeping handles updating if changed. */ static void _scsih_sas_host_refresh(struct MPT3SAS_ADAPTER *ioc) { u16 sz; u16 ioc_status; int i; Mpi2ConfigReply_t mpi_reply; Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL; u16 attached_handle; u8 link_rate, port_id; struct hba_port *port; Mpi2SasPhyPage0_t phy_pg0; dtmprintk(ioc, ioc_info(ioc, "updating handles for sas_host(0x%016llx)\n", (u64)ioc->sas_hba.sas_address)); sz = offsetof(Mpi2SasIOUnitPage0_t, PhyData) + (ioc->sas_hba.num_phys * sizeof(Mpi2SasIOUnit0PhyData_t)); sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL); if (!sas_iounit_pg0) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply, sas_iounit_pg0, sz)) != 0) goto out; ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) goto out; for (i = 0; i < ioc->sas_hba.num_phys ; i++) { link_rate = sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4; if (i == 0) ioc->sas_hba.handle = le16_to_cpu( sas_iounit_pg0->PhyData[0].ControllerDevHandle); port_id = sas_iounit_pg0->PhyData[i].Port; if (!(mpt3sas_get_port_by_id(ioc, port_id, 0))) { port = kzalloc(sizeof(struct hba_port), GFP_KERNEL); if (!port) goto out; port->port_id = port_id; ioc_info(ioc, "hba_port entry: %p, port: %d is added to hba_port list\n", port, port->port_id); if (ioc->shost_recovery) port->flags = HBA_PORT_FLAG_NEW_PORT; list_add_tail(&port->list, &ioc->port_table_list); } /* * Check whether current Phy belongs to HBA vSES device or not. */ if (le32_to_cpu(sas_iounit_pg0->PhyData[i].ControllerPhyDeviceInfo) & MPI2_SAS_DEVICE_INFO_SEP && (link_rate >= MPI2_SAS_NEG_LINK_RATE_1_5)) { if ((mpt3sas_config_get_phy_pg0(ioc, &mpi_reply, &phy_pg0, i))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } if (!(le32_to_cpu(phy_pg0.PhyInfo) & MPI2_SAS_PHYINFO_VIRTUAL_PHY)) continue; /* * Allocate a virtual_phy object for vSES device, if * this vSES device is hot added. */ if (!_scsih_alloc_vphy(ioc, port_id, i)) goto out; ioc->sas_hba.phy[i].hba_vphy = 1; } ioc->sas_hba.phy[i].handle = ioc->sas_hba.handle; attached_handle = le16_to_cpu(sas_iounit_pg0->PhyData[i]. AttachedDevHandle); if (attached_handle && link_rate < MPI2_SAS_NEG_LINK_RATE_1_5) link_rate = MPI2_SAS_NEG_LINK_RATE_1_5; ioc->sas_hba.phy[i].port = mpt3sas_get_port_by_id(ioc, port_id, 0); mpt3sas_transport_update_links(ioc, ioc->sas_hba.sas_address, attached_handle, i, link_rate, ioc->sas_hba.phy[i].port); } out: kfree(sas_iounit_pg0); } /** * _scsih_sas_host_add - create sas host object * @ioc: per adapter object * * Creating host side data object, stored in ioc->sas_hba */ static void _scsih_sas_host_add(struct MPT3SAS_ADAPTER *ioc) { int i; Mpi2ConfigReply_t mpi_reply; Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL; Mpi2SasIOUnitPage1_t *sas_iounit_pg1 = NULL; Mpi2SasPhyPage0_t phy_pg0; Mpi2SasDevicePage0_t sas_device_pg0; Mpi2SasEnclosurePage0_t enclosure_pg0; u16 ioc_status; u16 sz; u8 device_missing_delay; u8 num_phys, port_id; struct hba_port *port; mpt3sas_config_get_number_hba_phys(ioc, &num_phys); if (!num_phys) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } ioc->sas_hba.phy = kcalloc(num_phys, sizeof(struct _sas_phy), GFP_KERNEL); if (!ioc->sas_hba.phy) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } ioc->sas_hba.num_phys = num_phys; /* sas_iounit page 0 */ sz = offsetof(Mpi2SasIOUnitPage0_t, PhyData) + (ioc->sas_hba.num_phys * sizeof(Mpi2SasIOUnit0PhyData_t)); sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL); if (!sas_iounit_pg0) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply, sas_iounit_pg0, sz))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } /* sas_iounit page 1 */ sz = offsetof(Mpi2SasIOUnitPage1_t, PhyData) + (ioc->sas_hba.num_phys * sizeof(Mpi2SasIOUnit1PhyData_t)); sas_iounit_pg1 = kzalloc(sz, GFP_KERNEL); if (!sas_iounit_pg1) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } if ((mpt3sas_config_get_sas_iounit_pg1(ioc, &mpi_reply, sas_iounit_pg1, sz))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } ioc->io_missing_delay = sas_iounit_pg1->IODeviceMissingDelay; device_missing_delay = sas_iounit_pg1->ReportDeviceMissingDelay; if (device_missing_delay & MPI2_SASIOUNIT1_REPORT_MISSING_UNIT_16) ioc->device_missing_delay = (device_missing_delay & MPI2_SASIOUNIT1_REPORT_MISSING_TIMEOUT_MASK) * 16; else ioc->device_missing_delay = device_missing_delay & MPI2_SASIOUNIT1_REPORT_MISSING_TIMEOUT_MASK; ioc->sas_hba.parent_dev = &ioc->shost->shost_gendev; for (i = 0; i < ioc->sas_hba.num_phys ; i++) { if ((mpt3sas_config_get_phy_pg0(ioc, &mpi_reply, &phy_pg0, i))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } if (i == 0) ioc->sas_hba.handle = le16_to_cpu(sas_iounit_pg0-> PhyData[0].ControllerDevHandle); port_id = sas_iounit_pg0->PhyData[i].Port; if (!(mpt3sas_get_port_by_id(ioc, port_id, 0))) { port = kzalloc(sizeof(struct hba_port), GFP_KERNEL); if (!port) goto out; port->port_id = port_id; ioc_info(ioc, "hba_port entry: %p, port: %d is added to hba_port list\n", port, port->port_id); list_add_tail(&port->list, &ioc->port_table_list); } /* * Check whether current Phy belongs to HBA vSES device or not. */ if ((le32_to_cpu(phy_pg0.PhyInfo) & MPI2_SAS_PHYINFO_VIRTUAL_PHY) && (phy_pg0.NegotiatedLinkRate >> 4) >= MPI2_SAS_NEG_LINK_RATE_1_5) { /* * Allocate a virtual_phy object for vSES device. */ if (!_scsih_alloc_vphy(ioc, port_id, i)) goto out; ioc->sas_hba.phy[i].hba_vphy = 1; } ioc->sas_hba.phy[i].handle = ioc->sas_hba.handle; ioc->sas_hba.phy[i].phy_id = i; ioc->sas_hba.phy[i].port = mpt3sas_get_port_by_id(ioc, port_id, 0); mpt3sas_transport_add_host_phy(ioc, &ioc->sas_hba.phy[i], phy_pg0, ioc->sas_hba.parent_dev); } if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, ioc->sas_hba.handle))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } ioc->sas_hba.enclosure_handle = le16_to_cpu(sas_device_pg0.EnclosureHandle); ioc->sas_hba.sas_address = le64_to_cpu(sas_device_pg0.SASAddress); ioc_info(ioc, "host_add: handle(0x%04x), sas_addr(0x%016llx), phys(%d)\n", ioc->sas_hba.handle, (u64)ioc->sas_hba.sas_address, ioc->sas_hba.num_phys); if (ioc->sas_hba.enclosure_handle) { if (!(mpt3sas_config_get_enclosure_pg0(ioc, &mpi_reply, &enclosure_pg0, MPI2_SAS_ENCLOS_PGAD_FORM_HANDLE, ioc->sas_hba.enclosure_handle))) ioc->sas_hba.enclosure_logical_id = le64_to_cpu(enclosure_pg0.EnclosureLogicalID); } out: kfree(sas_iounit_pg1); kfree(sas_iounit_pg0); } /** * _scsih_expander_add - creating expander object * @ioc: per adapter object * @handle: expander handle * * Creating expander object, stored in ioc->sas_expander_list. * * Return: 0 for success, else error. */ static int _scsih_expander_add(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct _sas_node *sas_expander; struct _enclosure_node *enclosure_dev; Mpi2ConfigReply_t mpi_reply; Mpi2ExpanderPage0_t expander_pg0; Mpi2ExpanderPage1_t expander_pg1; u32 ioc_status; u16 parent_handle; u64 sas_address, sas_address_parent = 0; int i; unsigned long flags; struct _sas_port *mpt3sas_port = NULL; u8 port_id; int rc = 0; if (!handle) return -1; if (ioc->shost_recovery || ioc->pci_error_recovery) return -1; if ((mpt3sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0, MPI2_SAS_EXPAND_PGAD_FORM_HNDL, handle))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return -1; } ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return -1; } /* handle out of order topology events */ parent_handle = le16_to_cpu(expander_pg0.ParentDevHandle); if (_scsih_get_sas_address(ioc, parent_handle, &sas_address_parent) != 0) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return -1; } port_id = expander_pg0.PhysicalPort; if (sas_address_parent != ioc->sas_hba.sas_address) { spin_lock_irqsave(&ioc->sas_node_lock, flags); sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc, sas_address_parent, mpt3sas_get_port_by_id(ioc, port_id, 0)); spin_unlock_irqrestore(&ioc->sas_node_lock, flags); if (!sas_expander) { rc = _scsih_expander_add(ioc, parent_handle); if (rc != 0) return rc; } } spin_lock_irqsave(&ioc->sas_node_lock, flags); sas_address = le64_to_cpu(expander_pg0.SASAddress); sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc, sas_address, mpt3sas_get_port_by_id(ioc, port_id, 0)); spin_unlock_irqrestore(&ioc->sas_node_lock, flags); if (sas_expander) return 0; sas_expander = kzalloc(sizeof(struct _sas_node), GFP_KERNEL); if (!sas_expander) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return -1; } sas_expander->handle = handle; sas_expander->num_phys = expander_pg0.NumPhys; sas_expander->sas_address_parent = sas_address_parent; sas_expander->sas_address = sas_address; sas_expander->port = mpt3sas_get_port_by_id(ioc, port_id, 0); if (!sas_expander->port) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); rc = -1; goto out_fail; } ioc_info(ioc, "expander_add: handle(0x%04x), parent(0x%04x), sas_addr(0x%016llx), phys(%d)\n", handle, parent_handle, (u64)sas_expander->sas_address, sas_expander->num_phys); if (!sas_expander->num_phys) goto out_fail; sas_expander->phy = kcalloc(sas_expander->num_phys, sizeof(struct _sas_phy), GFP_KERNEL); if (!sas_expander->phy) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); rc = -1; goto out_fail; } INIT_LIST_HEAD(&sas_expander->sas_port_list); mpt3sas_port = mpt3sas_transport_port_add(ioc, handle, sas_address_parent, sas_expander->port); if (!mpt3sas_port) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); rc = -1; goto out_fail; } sas_expander->parent_dev = &mpt3sas_port->rphy->dev; sas_expander->rphy = mpt3sas_port->rphy; for (i = 0 ; i < sas_expander->num_phys ; i++) { if ((mpt3sas_config_get_expander_pg1(ioc, &mpi_reply, &expander_pg1, i, handle))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); rc = -1; goto out_fail; } sas_expander->phy[i].handle = handle; sas_expander->phy[i].phy_id = i; sas_expander->phy[i].port = mpt3sas_get_port_by_id(ioc, port_id, 0); if ((mpt3sas_transport_add_expander_phy(ioc, &sas_expander->phy[i], expander_pg1, sas_expander->parent_dev))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); rc = -1; goto out_fail; } } if (sas_expander->enclosure_handle) { enclosure_dev = mpt3sas_scsih_enclosure_find_by_handle(ioc, sas_expander->enclosure_handle); if (enclosure_dev) sas_expander->enclosure_logical_id = le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); } _scsih_expander_node_add(ioc, sas_expander); return 0; out_fail: if (mpt3sas_port) mpt3sas_transport_port_remove(ioc, sas_expander->sas_address, sas_address_parent, sas_expander->port); kfree(sas_expander); return rc; } /** * mpt3sas_expander_remove - removing expander object * @ioc: per adapter object * @sas_address: expander sas_address * @port: hba port entry */ void mpt3sas_expander_remove(struct MPT3SAS_ADAPTER *ioc, u64 sas_address, struct hba_port *port) { struct _sas_node *sas_expander; unsigned long flags; if (ioc->shost_recovery) return; if (!port) return; spin_lock_irqsave(&ioc->sas_node_lock, flags); sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc, sas_address, port); spin_unlock_irqrestore(&ioc->sas_node_lock, flags); if (sas_expander) _scsih_expander_node_remove(ioc, sas_expander); } /** * _scsih_done - internal SCSI_IO callback handler. * @ioc: per adapter object * @smid: system request message index * @msix_index: MSIX table index supplied by the OS * @reply: reply message frame(lower 32bit addr) * * Callback handler when sending internal generated SCSI_IO. * The callback index passed is `ioc->scsih_cb_idx` * * Return: 1 meaning mf should be freed from _base_interrupt * 0 means the mf is freed from this function. */ static u8 _scsih_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply) { MPI2DefaultReply_t *mpi_reply; mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply); if (ioc->scsih_cmds.status == MPT3_CMD_NOT_USED) return 1; if (ioc->scsih_cmds.smid != smid) return 1; ioc->scsih_cmds.status |= MPT3_CMD_COMPLETE; if (mpi_reply) { memcpy(ioc->scsih_cmds.reply, mpi_reply, mpi_reply->MsgLength*4); ioc->scsih_cmds.status |= MPT3_CMD_REPLY_VALID; } ioc->scsih_cmds.status &= ~MPT3_CMD_PENDING; complete(&ioc->scsih_cmds.done); return 1; } #define MPT3_MAX_LUNS (255) /** * _scsih_check_access_status - check access flags * @ioc: per adapter object * @sas_address: sas address * @handle: sas device handle * @access_status: errors returned during discovery of the device * * Return: 0 for success, else failure */ static u8 _scsih_check_access_status(struct MPT3SAS_ADAPTER *ioc, u64 sas_address, u16 handle, u8 access_status) { u8 rc = 1; char *desc = NULL; switch (access_status) { case MPI2_SAS_DEVICE0_ASTATUS_NO_ERRORS: case MPI2_SAS_DEVICE0_ASTATUS_SATA_NEEDS_INITIALIZATION: rc = 0; break; case MPI2_SAS_DEVICE0_ASTATUS_SATA_CAPABILITY_FAILED: desc = "sata capability failed"; break; case MPI2_SAS_DEVICE0_ASTATUS_SATA_AFFILIATION_CONFLICT: desc = "sata affiliation conflict"; break; case MPI2_SAS_DEVICE0_ASTATUS_ROUTE_NOT_ADDRESSABLE: desc = "route not addressable"; break; case MPI2_SAS_DEVICE0_ASTATUS_SMP_ERROR_NOT_ADDRESSABLE: desc = "smp error not addressable"; break; case MPI2_SAS_DEVICE0_ASTATUS_DEVICE_BLOCKED: desc = "device blocked"; break; case MPI2_SAS_DEVICE0_ASTATUS_SATA_INIT_FAILED: case MPI2_SAS_DEVICE0_ASTATUS_SIF_UNKNOWN: case MPI2_SAS_DEVICE0_ASTATUS_SIF_AFFILIATION_CONFLICT: case MPI2_SAS_DEVICE0_ASTATUS_SIF_DIAG: case MPI2_SAS_DEVICE0_ASTATUS_SIF_IDENTIFICATION: case MPI2_SAS_DEVICE0_ASTATUS_SIF_CHECK_POWER: case MPI2_SAS_DEVICE0_ASTATUS_SIF_PIO_SN: case MPI2_SAS_DEVICE0_ASTATUS_SIF_MDMA_SN: case MPI2_SAS_DEVICE0_ASTATUS_SIF_UDMA_SN: case MPI2_SAS_DEVICE0_ASTATUS_SIF_ZONING_VIOLATION: case MPI2_SAS_DEVICE0_ASTATUS_SIF_NOT_ADDRESSABLE: case MPI2_SAS_DEVICE0_ASTATUS_SIF_MAX: desc = "sata initialization failed"; break; default: desc = "unknown"; break; } if (!rc) return 0; ioc_err(ioc, "discovery errors(%s): sas_address(0x%016llx), handle(0x%04x)\n", desc, (u64)sas_address, handle); return rc; } /** * _scsih_check_device - checking device responsiveness * @ioc: per adapter object * @parent_sas_address: sas address of parent expander or sas host * @handle: attached device handle * @phy_number: phy number * @link_rate: new link rate */ static void _scsih_check_device(struct MPT3SAS_ADAPTER *ioc, u64 parent_sas_address, u16 handle, u8 phy_number, u8 link_rate) { Mpi2ConfigReply_t mpi_reply; Mpi2SasDevicePage0_t sas_device_pg0; struct _sas_device *sas_device = NULL; struct _enclosure_node *enclosure_dev = NULL; u32 ioc_status; unsigned long flags; u64 sas_address; struct scsi_target *starget; struct MPT3SAS_TARGET *sas_target_priv_data; u32 device_info; struct hba_port *port; if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) return; ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) return; /* wide port handling ~ we need only handle device once for the phy that * is matched in sas device page zero */ if (phy_number != sas_device_pg0.PhyNum) return; /* check if this is end device */ device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); if (!(_scsih_is_end_device(device_info))) return; spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_address = le64_to_cpu(sas_device_pg0.SASAddress); port = mpt3sas_get_port_by_id(ioc, sas_device_pg0.PhysicalPort, 0); if (!port) goto out_unlock; sas_device = __mpt3sas_get_sdev_by_addr(ioc, sas_address, port); if (!sas_device) goto out_unlock; if (unlikely(sas_device->handle != handle)) { starget = sas_device->starget; sas_target_priv_data = starget->hostdata; starget_printk(KERN_INFO, starget, "handle changed from(0x%04x) to (0x%04x)!!!\n", sas_device->handle, handle); sas_target_priv_data->handle = handle; sas_device->handle = handle; if (le16_to_cpu(sas_device_pg0.Flags) & MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) { sas_device->enclosure_level = sas_device_pg0.EnclosureLevel; memcpy(sas_device->connector_name, sas_device_pg0.ConnectorName, 4); sas_device->connector_name[4] = '\0'; } else { sas_device->enclosure_level = 0; sas_device->connector_name[0] = '\0'; } sas_device->enclosure_handle = le16_to_cpu(sas_device_pg0.EnclosureHandle); sas_device->is_chassis_slot_valid = 0; enclosure_dev = mpt3sas_scsih_enclosure_find_by_handle(ioc, sas_device->enclosure_handle); if (enclosure_dev) { sas_device->enclosure_logical_id = le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); if (le16_to_cpu(enclosure_dev->pg0.Flags) & MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) { sas_device->is_chassis_slot_valid = 1; sas_device->chassis_slot = enclosure_dev->pg0.ChassisSlot; } } } /* check if device is present */ if (!(le16_to_cpu(sas_device_pg0.Flags) & MPI2_SAS_DEVICE0_FLAGS_DEVICE_PRESENT)) { ioc_err(ioc, "device is not present handle(0x%04x), flags!!!\n", handle); goto out_unlock; } /* check if there were any issues with discovery */ if (_scsih_check_access_status(ioc, sas_address, handle, sas_device_pg0.AccessStatus)) goto out_unlock; spin_unlock_irqrestore(&ioc->sas_device_lock, flags); _scsih_ublock_io_device(ioc, sas_address, port); if (sas_device) sas_device_put(sas_device); return; out_unlock: spin_unlock_irqrestore(&ioc->sas_device_lock, flags); if (sas_device) sas_device_put(sas_device); } /** * _scsih_add_device - creating sas device object * @ioc: per adapter object * @handle: sas device handle * @phy_num: phy number end device attached to * @is_pd: is this hidden raid component * * Creating end device object, stored in ioc->sas_device_list. * * Return: 0 for success, non-zero for failure. */ static int _scsih_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle, u8 phy_num, u8 is_pd) { Mpi2ConfigReply_t mpi_reply; Mpi2SasDevicePage0_t sas_device_pg0; struct _sas_device *sas_device; struct _enclosure_node *enclosure_dev = NULL; u32 ioc_status; u64 sas_address; u32 device_info; u8 port_id; if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return -1; } ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return -1; } /* check if this is end device */ device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); if (!(_scsih_is_end_device(device_info))) return -1; set_bit(handle, ioc->pend_os_device_add); sas_address = le64_to_cpu(sas_device_pg0.SASAddress); /* check if device is present */ if (!(le16_to_cpu(sas_device_pg0.Flags) & MPI2_SAS_DEVICE0_FLAGS_DEVICE_PRESENT)) { ioc_err(ioc, "device is not present handle(0x04%x)!!!\n", handle); return -1; } /* check if there were any issues with discovery */ if (_scsih_check_access_status(ioc, sas_address, handle, sas_device_pg0.AccessStatus)) return -1; port_id = sas_device_pg0.PhysicalPort; sas_device = mpt3sas_get_sdev_by_addr(ioc, sas_address, mpt3sas_get_port_by_id(ioc, port_id, 0)); if (sas_device) { clear_bit(handle, ioc->pend_os_device_add); sas_device_put(sas_device); return -1; } if (sas_device_pg0.EnclosureHandle) { enclosure_dev = mpt3sas_scsih_enclosure_find_by_handle(ioc, le16_to_cpu(sas_device_pg0.EnclosureHandle)); if (enclosure_dev == NULL) ioc_info(ioc, "Enclosure handle(0x%04x) doesn't match with enclosure device!\n", sas_device_pg0.EnclosureHandle); } sas_device = kzalloc(sizeof(struct _sas_device), GFP_KERNEL); if (!sas_device) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return 0; } kref_init(&sas_device->refcount); sas_device->handle = handle; if (_scsih_get_sas_address(ioc, le16_to_cpu(sas_device_pg0.ParentDevHandle), &sas_device->sas_address_parent) != 0) ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); sas_device->enclosure_handle = le16_to_cpu(sas_device_pg0.EnclosureHandle); if (sas_device->enclosure_handle != 0) sas_device->slot = le16_to_cpu(sas_device_pg0.Slot); sas_device->device_info = device_info; sas_device->sas_address = sas_address; sas_device->phy = sas_device_pg0.PhyNum; sas_device->fast_path = (le16_to_cpu(sas_device_pg0.Flags) & MPI25_SAS_DEVICE0_FLAGS_FAST_PATH_CAPABLE) ? 1 : 0; sas_device->port = mpt3sas_get_port_by_id(ioc, port_id, 0); if (!sas_device->port) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out; } if (le16_to_cpu(sas_device_pg0.Flags) & MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) { sas_device->enclosure_level = sas_device_pg0.EnclosureLevel; memcpy(sas_device->connector_name, sas_device_pg0.ConnectorName, 4); sas_device->connector_name[4] = '\0'; } else { sas_device->enclosure_level = 0; sas_device->connector_name[0] = '\0'; } /* get enclosure_logical_id & chassis_slot*/ sas_device->is_chassis_slot_valid = 0; if (enclosure_dev) { sas_device->enclosure_logical_id = le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); if (le16_to_cpu(enclosure_dev->pg0.Flags) & MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) { sas_device->is_chassis_slot_valid = 1; sas_device->chassis_slot = enclosure_dev->pg0.ChassisSlot; } } /* get device name */ sas_device->device_name = le64_to_cpu(sas_device_pg0.DeviceName); if (ioc->wait_for_discovery_to_complete) _scsih_sas_device_init_add(ioc, sas_device); else _scsih_sas_device_add(ioc, sas_device); out: sas_device_put(sas_device); return 0; } /** * _scsih_remove_device - removing sas device object * @ioc: per adapter object * @sas_device: the sas_device object */ static void _scsih_remove_device(struct MPT3SAS_ADAPTER *ioc, struct _sas_device *sas_device) { struct MPT3SAS_TARGET *sas_target_priv_data; if ((ioc->pdev->subsystem_vendor == PCI_VENDOR_ID_IBM) && (sas_device->pfa_led_on)) { _scsih_turn_off_pfa_led(ioc, sas_device); sas_device->pfa_led_on = 0; } dewtprintk(ioc, ioc_info(ioc, "%s: enter: handle(0x%04x), sas_addr(0x%016llx)\n", __func__, sas_device->handle, (u64)sas_device->sas_address)); dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL)); if (sas_device->starget && sas_device->starget->hostdata) { sas_target_priv_data = sas_device->starget->hostdata; sas_target_priv_data->deleted = 1; _scsih_ublock_io_device(ioc, sas_device->sas_address, sas_device->port); sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE; } if (!ioc->hide_drives) mpt3sas_transport_port_remove(ioc, sas_device->sas_address, sas_device->sas_address_parent, sas_device->port); ioc_info(ioc, "removing handle(0x%04x), sas_addr(0x%016llx)\n", sas_device->handle, (u64)sas_device->sas_address); _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL); dewtprintk(ioc, ioc_info(ioc, "%s: exit: handle(0x%04x), sas_addr(0x%016llx)\n", __func__, sas_device->handle, (u64)sas_device->sas_address)); dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL)); } /** * _scsih_sas_topology_change_event_debug - debug for topology event * @ioc: per adapter object * @event_data: event data payload * Context: user. */ static void _scsih_sas_topology_change_event_debug(struct MPT3SAS_ADAPTER *ioc, Mpi2EventDataSasTopologyChangeList_t *event_data) { int i; u16 handle; u16 reason_code; u8 phy_number; char *status_str = NULL; u8 link_rate, prev_link_rate; switch (event_data->ExpStatus) { case MPI2_EVENT_SAS_TOPO_ES_ADDED: status_str = "add"; break; case MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING: status_str = "remove"; break; case MPI2_EVENT_SAS_TOPO_ES_RESPONDING: case 0: status_str = "responding"; break; case MPI2_EVENT_SAS_TOPO_ES_DELAY_NOT_RESPONDING: status_str = "remove delay"; break; default: status_str = "unknown status"; break; } ioc_info(ioc, "sas topology change: (%s)\n", status_str); pr_info("\thandle(0x%04x), enclosure_handle(0x%04x) " \ "start_phy(%02d), count(%d)\n", le16_to_cpu(event_data->ExpanderDevHandle), le16_to_cpu(event_data->EnclosureHandle), event_data->StartPhyNum, event_data->NumEntries); for (i = 0; i < event_data->NumEntries; i++) { handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); if (!handle) continue; phy_number = event_data->StartPhyNum + i; reason_code = event_data->PHY[i].PhyStatus & MPI2_EVENT_SAS_TOPO_RC_MASK; switch (reason_code) { case MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED: status_str = "target add"; break; case MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING: status_str = "target remove"; break; case MPI2_EVENT_SAS_TOPO_RC_DELAY_NOT_RESPONDING: status_str = "delay target remove"; break; case MPI2_EVENT_SAS_TOPO_RC_PHY_CHANGED: status_str = "link rate change"; break; case MPI2_EVENT_SAS_TOPO_RC_NO_CHANGE: status_str = "target responding"; break; default: status_str = "unknown"; break; } link_rate = event_data->PHY[i].LinkRate >> 4; prev_link_rate = event_data->PHY[i].LinkRate & 0xF; pr_info("\tphy(%02d), attached_handle(0x%04x): %s:" \ " link rate: new(0x%02x), old(0x%02x)\n", phy_number, handle, status_str, link_rate, prev_link_rate); } } /** * _scsih_sas_topology_change_event - handle topology changes * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. * */ static int _scsih_sas_topology_change_event(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { int i; u16 parent_handle, handle; u16 reason_code; u8 phy_number, max_phys; struct _sas_node *sas_expander; u64 sas_address; unsigned long flags; u8 link_rate, prev_link_rate; struct hba_port *port; Mpi2EventDataSasTopologyChangeList_t *event_data = (Mpi2EventDataSasTopologyChangeList_t *) fw_event->event_data; if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) _scsih_sas_topology_change_event_debug(ioc, event_data); if (ioc->shost_recovery || ioc->remove_host || ioc->pci_error_recovery) return 0; if (!ioc->sas_hba.num_phys) _scsih_sas_host_add(ioc); else _scsih_sas_host_refresh(ioc); if (fw_event->ignore) { dewtprintk(ioc, ioc_info(ioc, "ignoring expander event\n")); return 0; } parent_handle = le16_to_cpu(event_data->ExpanderDevHandle); port = mpt3sas_get_port_by_id(ioc, event_data->PhysicalPort, 0); /* handle expander add */ if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_ADDED) if (_scsih_expander_add(ioc, parent_handle) != 0) return 0; spin_lock_irqsave(&ioc->sas_node_lock, flags); sas_expander = mpt3sas_scsih_expander_find_by_handle(ioc, parent_handle); if (sas_expander) { sas_address = sas_expander->sas_address; max_phys = sas_expander->num_phys; port = sas_expander->port; } else if (parent_handle < ioc->sas_hba.num_phys) { sas_address = ioc->sas_hba.sas_address; max_phys = ioc->sas_hba.num_phys; } else { spin_unlock_irqrestore(&ioc->sas_node_lock, flags); return 0; } spin_unlock_irqrestore(&ioc->sas_node_lock, flags); /* handle siblings events */ for (i = 0; i < event_data->NumEntries; i++) { if (fw_event->ignore) { dewtprintk(ioc, ioc_info(ioc, "ignoring expander event\n")); return 0; } if (ioc->remove_host || ioc->pci_error_recovery) return 0; phy_number = event_data->StartPhyNum + i; if (phy_number >= max_phys) continue; reason_code = event_data->PHY[i].PhyStatus & MPI2_EVENT_SAS_TOPO_RC_MASK; if ((event_data->PHY[i].PhyStatus & MPI2_EVENT_SAS_TOPO_PHYSTATUS_VACANT) && (reason_code != MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING)) continue; handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); if (!handle) continue; link_rate = event_data->PHY[i].LinkRate >> 4; prev_link_rate = event_data->PHY[i].LinkRate & 0xF; switch (reason_code) { case MPI2_EVENT_SAS_TOPO_RC_PHY_CHANGED: if (ioc->shost_recovery) break; if (link_rate == prev_link_rate) break; mpt3sas_transport_update_links(ioc, sas_address, handle, phy_number, link_rate, port); if (link_rate < MPI2_SAS_NEG_LINK_RATE_1_5) break; _scsih_check_device(ioc, sas_address, handle, phy_number, link_rate); if (!test_bit(handle, ioc->pend_os_device_add)) break; fallthrough; case MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED: if (ioc->shost_recovery) break; mpt3sas_transport_update_links(ioc, sas_address, handle, phy_number, link_rate, port); _scsih_add_device(ioc, handle, phy_number, 0); break; case MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING: _scsih_device_remove_by_handle(ioc, handle); break; } } /* handle expander removal */ if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING && sas_expander) mpt3sas_expander_remove(ioc, sas_address, port); return 0; } /** * _scsih_sas_device_status_change_event_debug - debug for device event * @ioc: ? * @event_data: event data payload * Context: user. */ static void _scsih_sas_device_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc, Mpi2EventDataSasDeviceStatusChange_t *event_data) { char *reason_str = NULL; switch (event_data->ReasonCode) { case MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA: reason_str = "smart data"; break; case MPI2_EVENT_SAS_DEV_STAT_RC_UNSUPPORTED: reason_str = "unsupported device discovered"; break; case MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET: reason_str = "internal device reset"; break; case MPI2_EVENT_SAS_DEV_STAT_RC_TASK_ABORT_INTERNAL: reason_str = "internal task abort"; break; case MPI2_EVENT_SAS_DEV_STAT_RC_ABORT_TASK_SET_INTERNAL: reason_str = "internal task abort set"; break; case MPI2_EVENT_SAS_DEV_STAT_RC_CLEAR_TASK_SET_INTERNAL: reason_str = "internal clear task set"; break; case MPI2_EVENT_SAS_DEV_STAT_RC_QUERY_TASK_INTERNAL: reason_str = "internal query task"; break; case MPI2_EVENT_SAS_DEV_STAT_RC_SATA_INIT_FAILURE: reason_str = "sata init failure"; break; case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_INTERNAL_DEV_RESET: reason_str = "internal device reset complete"; break; case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_TASK_ABORT_INTERNAL: reason_str = "internal task abort complete"; break; case MPI2_EVENT_SAS_DEV_STAT_RC_ASYNC_NOTIFICATION: reason_str = "internal async notification"; break; case MPI2_EVENT_SAS_DEV_STAT_RC_EXPANDER_REDUCED_FUNCTIONALITY: reason_str = "expander reduced functionality"; break; case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_EXPANDER_REDUCED_FUNCTIONALITY: reason_str = "expander reduced functionality complete"; break; default: reason_str = "unknown reason"; break; } ioc_info(ioc, "device status change: (%s)\thandle(0x%04x), sas address(0x%016llx), tag(%d)", reason_str, le16_to_cpu(event_data->DevHandle), (u64)le64_to_cpu(event_data->SASAddress), le16_to_cpu(event_data->TaskTag)); if (event_data->ReasonCode == MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA) pr_cont(", ASC(0x%x), ASCQ(0x%x)\n", event_data->ASC, event_data->ASCQ); pr_cont("\n"); } /** * _scsih_sas_device_status_change_event - handle device status change * @ioc: per adapter object * @event_data: The fw event * Context: user. */ static void _scsih_sas_device_status_change_event(struct MPT3SAS_ADAPTER *ioc, Mpi2EventDataSasDeviceStatusChange_t *event_data) { struct MPT3SAS_TARGET *target_priv_data; struct _sas_device *sas_device; u64 sas_address; unsigned long flags; /* In MPI Revision K (0xC), the internal device reset complete was * implemented, so avoid setting tm_busy flag for older firmware. */ if ((ioc->facts.HeaderVersion >> 8) < 0xC) return; if (event_data->ReasonCode != MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET && event_data->ReasonCode != MPI2_EVENT_SAS_DEV_STAT_RC_CMP_INTERNAL_DEV_RESET) return; spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_address = le64_to_cpu(event_data->SASAddress); sas_device = __mpt3sas_get_sdev_by_addr(ioc, sas_address, mpt3sas_get_port_by_id(ioc, event_data->PhysicalPort, 0)); if (!sas_device || !sas_device->starget) goto out; target_priv_data = sas_device->starget->hostdata; if (!target_priv_data) goto out; if (event_data->ReasonCode == MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET) target_priv_data->tm_busy = 1; else target_priv_data->tm_busy = 0; if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) ioc_info(ioc, "%s tm_busy flag for handle(0x%04x)\n", (target_priv_data->tm_busy == 1) ? "Enable" : "Disable", target_priv_data->handle); out: if (sas_device) sas_device_put(sas_device); spin_unlock_irqrestore(&ioc->sas_device_lock, flags); } /** * _scsih_check_pcie_access_status - check access flags * @ioc: per adapter object * @wwid: wwid * @handle: sas device handle * @access_status: errors returned during discovery of the device * * Return: 0 for success, else failure */ static u8 _scsih_check_pcie_access_status(struct MPT3SAS_ADAPTER *ioc, u64 wwid, u16 handle, u8 access_status) { u8 rc = 1; char *desc = NULL; switch (access_status) { case MPI26_PCIEDEV0_ASTATUS_NO_ERRORS: case MPI26_PCIEDEV0_ASTATUS_NEEDS_INITIALIZATION: rc = 0; break; case MPI26_PCIEDEV0_ASTATUS_CAPABILITY_FAILED: desc = "PCIe device capability failed"; break; case MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED: desc = "PCIe device blocked"; ioc_info(ioc, "Device with Access Status (%s): wwid(0x%016llx), " "handle(0x%04x)\n ll only be added to the internal list", desc, (u64)wwid, handle); rc = 0; break; case MPI26_PCIEDEV0_ASTATUS_MEMORY_SPACE_ACCESS_FAILED: desc = "PCIe device mem space access failed"; break; case MPI26_PCIEDEV0_ASTATUS_UNSUPPORTED_DEVICE: desc = "PCIe device unsupported"; break; case MPI26_PCIEDEV0_ASTATUS_MSIX_REQUIRED: desc = "PCIe device MSIx Required"; break; case MPI26_PCIEDEV0_ASTATUS_INIT_FAIL_MAX: desc = "PCIe device init fail max"; break; case MPI26_PCIEDEV0_ASTATUS_UNKNOWN: desc = "PCIe device status unknown"; break; case MPI26_PCIEDEV0_ASTATUS_NVME_READY_TIMEOUT: desc = "nvme ready timeout"; break; case MPI26_PCIEDEV0_ASTATUS_NVME_DEVCFG_UNSUPPORTED: desc = "nvme device configuration unsupported"; break; case MPI26_PCIEDEV0_ASTATUS_NVME_IDENTIFY_FAILED: desc = "nvme identify failed"; break; case MPI26_PCIEDEV0_ASTATUS_NVME_QCONFIG_FAILED: desc = "nvme qconfig failed"; break; case MPI26_PCIEDEV0_ASTATUS_NVME_QCREATION_FAILED: desc = "nvme qcreation failed"; break; case MPI26_PCIEDEV0_ASTATUS_NVME_EVENTCFG_FAILED: desc = "nvme eventcfg failed"; break; case MPI26_PCIEDEV0_ASTATUS_NVME_GET_FEATURE_STAT_FAILED: desc = "nvme get feature stat failed"; break; case MPI26_PCIEDEV0_ASTATUS_NVME_IDLE_TIMEOUT: desc = "nvme idle timeout"; break; case MPI26_PCIEDEV0_ASTATUS_NVME_FAILURE_STATUS: desc = "nvme failure status"; break; default: ioc_err(ioc, "NVMe discovery error(0x%02x): wwid(0x%016llx), handle(0x%04x)\n", access_status, (u64)wwid, handle); return rc; } if (!rc) return rc; ioc_info(ioc, "NVMe discovery error(%s): wwid(0x%016llx), handle(0x%04x)\n", desc, (u64)wwid, handle); return rc; } /** * _scsih_pcie_device_remove_from_sml - removing pcie device * from SML and free up associated memory * @ioc: per adapter object * @pcie_device: the pcie_device object */ static void _scsih_pcie_device_remove_from_sml(struct MPT3SAS_ADAPTER *ioc, struct _pcie_device *pcie_device) { struct MPT3SAS_TARGET *sas_target_priv_data; dewtprintk(ioc, ioc_info(ioc, "%s: enter: handle(0x%04x), wwid(0x%016llx)\n", __func__, pcie_device->handle, (u64)pcie_device->wwid)); if (pcie_device->enclosure_handle != 0) dewtprintk(ioc, ioc_info(ioc, "%s: enter: enclosure logical id(0x%016llx), slot(%d)\n", __func__, (u64)pcie_device->enclosure_logical_id, pcie_device->slot)); if (pcie_device->connector_name[0] != '\0') dewtprintk(ioc, ioc_info(ioc, "%s: enter: enclosure level(0x%04x), connector name(%s)\n", __func__, pcie_device->enclosure_level, pcie_device->connector_name)); if (pcie_device->starget && pcie_device->starget->hostdata) { sas_target_priv_data = pcie_device->starget->hostdata; sas_target_priv_data->deleted = 1; _scsih_ublock_io_device(ioc, pcie_device->wwid, NULL); sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE; } ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n", pcie_device->handle, (u64)pcie_device->wwid); if (pcie_device->enclosure_handle != 0) ioc_info(ioc, "removing : enclosure logical id(0x%016llx), slot(%d)\n", (u64)pcie_device->enclosure_logical_id, pcie_device->slot); if (pcie_device->connector_name[0] != '\0') ioc_info(ioc, "removing: enclosure level(0x%04x), connector name( %s)\n", pcie_device->enclosure_level, pcie_device->connector_name); if (pcie_device->starget && (pcie_device->access_status != MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED)) scsi_remove_target(&pcie_device->starget->dev); dewtprintk(ioc, ioc_info(ioc, "%s: exit: handle(0x%04x), wwid(0x%016llx)\n", __func__, pcie_device->handle, (u64)pcie_device->wwid)); if (pcie_device->enclosure_handle != 0) dewtprintk(ioc, ioc_info(ioc, "%s: exit: enclosure logical id(0x%016llx), slot(%d)\n", __func__, (u64)pcie_device->enclosure_logical_id, pcie_device->slot)); if (pcie_device->connector_name[0] != '\0') dewtprintk(ioc, ioc_info(ioc, "%s: exit: enclosure level(0x%04x), connector name( %s)\n", __func__, pcie_device->enclosure_level, pcie_device->connector_name)); kfree(pcie_device->serial_number); } /** * _scsih_pcie_check_device - checking device responsiveness * @ioc: per adapter object * @handle: attached device handle */ static void _scsih_pcie_check_device(struct MPT3SAS_ADAPTER *ioc, u16 handle) { Mpi2ConfigReply_t mpi_reply; Mpi26PCIeDevicePage0_t pcie_device_pg0; u32 ioc_status; struct _pcie_device *pcie_device; u64 wwid; unsigned long flags; struct scsi_target *starget; struct MPT3SAS_TARGET *sas_target_priv_data; u32 device_info; if ((mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply, &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_HANDLE, handle))) return; ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) return; /* check if this is end device */ device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo); if (!(_scsih_is_nvme_pciescsi_device(device_info))) return; wwid = le64_to_cpu(pcie_device_pg0.WWID); spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid); if (!pcie_device) { spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); return; } if (unlikely(pcie_device->handle != handle)) { starget = pcie_device->starget; sas_target_priv_data = starget->hostdata; pcie_device->access_status = pcie_device_pg0.AccessStatus; starget_printk(KERN_INFO, starget, "handle changed from(0x%04x) to (0x%04x)!!!\n", pcie_device->handle, handle); sas_target_priv_data->handle = handle; pcie_device->handle = handle; if (le32_to_cpu(pcie_device_pg0.Flags) & MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) { pcie_device->enclosure_level = pcie_device_pg0.EnclosureLevel; memcpy(&pcie_device->connector_name[0], &pcie_device_pg0.ConnectorName[0], 4); } else { pcie_device->enclosure_level = 0; pcie_device->connector_name[0] = '\0'; } } /* check if device is present */ if (!(le32_to_cpu(pcie_device_pg0.Flags) & MPI26_PCIEDEV0_FLAGS_DEVICE_PRESENT)) { ioc_info(ioc, "device is not present handle(0x%04x), flags!!!\n", handle); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); pcie_device_put(pcie_device); return; } /* check if there were any issues with discovery */ if (_scsih_check_pcie_access_status(ioc, wwid, handle, pcie_device_pg0.AccessStatus)) { spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); pcie_device_put(pcie_device); return; } spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); pcie_device_put(pcie_device); _scsih_ublock_io_device(ioc, wwid, NULL); return; } /** * _scsih_pcie_add_device - creating pcie device object * @ioc: per adapter object * @handle: pcie device handle * * Creating end device object, stored in ioc->pcie_device_list. * * Return: 1 means queue the event later, 0 means complete the event */ static int _scsih_pcie_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle) { Mpi26PCIeDevicePage0_t pcie_device_pg0; Mpi26PCIeDevicePage2_t pcie_device_pg2; Mpi2ConfigReply_t mpi_reply; struct _pcie_device *pcie_device; struct _enclosure_node *enclosure_dev; u32 ioc_status; u64 wwid; if ((mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply, &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_HANDLE, handle))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return 0; } ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return 0; } set_bit(handle, ioc->pend_os_device_add); wwid = le64_to_cpu(pcie_device_pg0.WWID); /* check if device is present */ if (!(le32_to_cpu(pcie_device_pg0.Flags) & MPI26_PCIEDEV0_FLAGS_DEVICE_PRESENT)) { ioc_err(ioc, "device is not present handle(0x04%x)!!!\n", handle); return 0; } /* check if there were any issues with discovery */ if (_scsih_check_pcie_access_status(ioc, wwid, handle, pcie_device_pg0.AccessStatus)) return 0; if (!(_scsih_is_nvme_pciescsi_device(le32_to_cpu (pcie_device_pg0.DeviceInfo)))) return 0; pcie_device = mpt3sas_get_pdev_by_wwid(ioc, wwid); if (pcie_device) { clear_bit(handle, ioc->pend_os_device_add); pcie_device_put(pcie_device); return 0; } /* PCIe Device Page 2 contains read-only information about a * specific NVMe device; therefore, this page is only * valid for NVMe devices and skip for pcie devices of type scsi. */ if (!(mpt3sas_scsih_is_pcie_scsi_device( le32_to_cpu(pcie_device_pg0.DeviceInfo)))) { if (mpt3sas_config_get_pcie_device_pg2(ioc, &mpi_reply, &pcie_device_pg2, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle)) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return 0; } ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return 0; } } pcie_device = kzalloc(sizeof(struct _pcie_device), GFP_KERNEL); if (!pcie_device) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return 0; } kref_init(&pcie_device->refcount); pcie_device->id = ioc->pcie_target_id++; pcie_device->channel = PCIE_CHANNEL; pcie_device->handle = handle; pcie_device->access_status = pcie_device_pg0.AccessStatus; pcie_device->device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo); pcie_device->wwid = wwid; pcie_device->port_num = pcie_device_pg0.PortNum; pcie_device->fast_path = (le32_to_cpu(pcie_device_pg0.Flags) & MPI26_PCIEDEV0_FLAGS_FAST_PATH_CAPABLE) ? 1 : 0; pcie_device->enclosure_handle = le16_to_cpu(pcie_device_pg0.EnclosureHandle); if (pcie_device->enclosure_handle != 0) pcie_device->slot = le16_to_cpu(pcie_device_pg0.Slot); if (le32_to_cpu(pcie_device_pg0.Flags) & MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) { pcie_device->enclosure_level = pcie_device_pg0.EnclosureLevel; memcpy(&pcie_device->connector_name[0], &pcie_device_pg0.ConnectorName[0], 4); } else { pcie_device->enclosure_level = 0; pcie_device->connector_name[0] = '\0'; } /* get enclosure_logical_id */ if (pcie_device->enclosure_handle) { enclosure_dev = mpt3sas_scsih_enclosure_find_by_handle(ioc, pcie_device->enclosure_handle); if (enclosure_dev) pcie_device->enclosure_logical_id = le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); } /* TODO -- Add device name once FW supports it */ if (!(mpt3sas_scsih_is_pcie_scsi_device( le32_to_cpu(pcie_device_pg0.DeviceInfo)))) { pcie_device->nvme_mdts = le32_to_cpu(pcie_device_pg2.MaximumDataTransferSize); pcie_device->shutdown_latency = le16_to_cpu(pcie_device_pg2.ShutdownLatency); /* * Set IOC's max_shutdown_latency to drive's RTD3 Entry Latency * if drive's RTD3 Entry Latency is greater then IOC's * max_shutdown_latency. */ if (pcie_device->shutdown_latency > ioc->max_shutdown_latency) ioc->max_shutdown_latency = pcie_device->shutdown_latency; if (pcie_device_pg2.ControllerResetTO) pcie_device->reset_timeout = pcie_device_pg2.ControllerResetTO; else pcie_device->reset_timeout = 30; } else pcie_device->reset_timeout = 30; if (ioc->wait_for_discovery_to_complete) _scsih_pcie_device_init_add(ioc, pcie_device); else _scsih_pcie_device_add(ioc, pcie_device); pcie_device_put(pcie_device); return 0; } /** * _scsih_pcie_topology_change_event_debug - debug for topology * event * @ioc: per adapter object * @event_data: event data payload * Context: user. */ static void _scsih_pcie_topology_change_event_debug(struct MPT3SAS_ADAPTER *ioc, Mpi26EventDataPCIeTopologyChangeList_t *event_data) { int i; u16 handle; u16 reason_code; u8 port_number; char *status_str = NULL; u8 link_rate, prev_link_rate; switch (event_data->SwitchStatus) { case MPI26_EVENT_PCIE_TOPO_SS_ADDED: status_str = "add"; break; case MPI26_EVENT_PCIE_TOPO_SS_NOT_RESPONDING: status_str = "remove"; break; case MPI26_EVENT_PCIE_TOPO_SS_RESPONDING: case 0: status_str = "responding"; break; case MPI26_EVENT_PCIE_TOPO_SS_DELAY_NOT_RESPONDING: status_str = "remove delay"; break; default: status_str = "unknown status"; break; } ioc_info(ioc, "pcie topology change: (%s)\n", status_str); pr_info("\tswitch_handle(0x%04x), enclosure_handle(0x%04x)" "start_port(%02d), count(%d)\n", le16_to_cpu(event_data->SwitchDevHandle), le16_to_cpu(event_data->EnclosureHandle), event_data->StartPortNum, event_data->NumEntries); for (i = 0; i < event_data->NumEntries; i++) { handle = le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle); if (!handle) continue; port_number = event_data->StartPortNum + i; reason_code = event_data->PortEntry[i].PortStatus; switch (reason_code) { case MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED: status_str = "target add"; break; case MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING: status_str = "target remove"; break; case MPI26_EVENT_PCIE_TOPO_PS_DELAY_NOT_RESPONDING: status_str = "delay target remove"; break; case MPI26_EVENT_PCIE_TOPO_PS_PORT_CHANGED: status_str = "link rate change"; break; case MPI26_EVENT_PCIE_TOPO_PS_NO_CHANGE: status_str = "target responding"; break; default: status_str = "unknown"; break; } link_rate = event_data->PortEntry[i].CurrentPortInfo & MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK; prev_link_rate = event_data->PortEntry[i].PreviousPortInfo & MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK; pr_info("\tport(%02d), attached_handle(0x%04x): %s:" " link rate: new(0x%02x), old(0x%02x)\n", port_number, handle, status_str, link_rate, prev_link_rate); } } /** * _scsih_pcie_topology_change_event - handle PCIe topology * changes * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. * */ static void _scsih_pcie_topology_change_event(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { int i; u16 handle; u16 reason_code; u8 link_rate, prev_link_rate; unsigned long flags; int rc; Mpi26EventDataPCIeTopologyChangeList_t *event_data = (Mpi26EventDataPCIeTopologyChangeList_t *) fw_event->event_data; struct _pcie_device *pcie_device; if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) _scsih_pcie_topology_change_event_debug(ioc, event_data); if (ioc->shost_recovery || ioc->remove_host || ioc->pci_error_recovery) return; if (fw_event->ignore) { dewtprintk(ioc, ioc_info(ioc, "ignoring switch event\n")); return; } /* handle siblings events */ for (i = 0; i < event_data->NumEntries; i++) { if (fw_event->ignore) { dewtprintk(ioc, ioc_info(ioc, "ignoring switch event\n")); return; } if (ioc->remove_host || ioc->pci_error_recovery) return; reason_code = event_data->PortEntry[i].PortStatus; handle = le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle); if (!handle) continue; link_rate = event_data->PortEntry[i].CurrentPortInfo & MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK; prev_link_rate = event_data->PortEntry[i].PreviousPortInfo & MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK; switch (reason_code) { case MPI26_EVENT_PCIE_TOPO_PS_PORT_CHANGED: if (ioc->shost_recovery) break; if (link_rate == prev_link_rate) break; if (link_rate < MPI26_EVENT_PCIE_TOPO_PI_RATE_2_5) break; _scsih_pcie_check_device(ioc, handle); /* This code after this point handles the test case * where a device has been added, however its returning * BUSY for sometime. Then before the Device Missing * Delay expires and the device becomes READY, the * device is removed and added back. */ spin_lock_irqsave(&ioc->pcie_device_lock, flags); pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); if (pcie_device) { pcie_device_put(pcie_device); break; } if (!test_bit(handle, ioc->pend_os_device_add)) break; dewtprintk(ioc, ioc_info(ioc, "handle(0x%04x) device not found: convert event to a device add\n", handle)); event_data->PortEntry[i].PortStatus &= 0xF0; event_data->PortEntry[i].PortStatus |= MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED; fallthrough; case MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED: if (ioc->shost_recovery) break; if (link_rate < MPI26_EVENT_PCIE_TOPO_PI_RATE_2_5) break; rc = _scsih_pcie_add_device(ioc, handle); if (!rc) { /* mark entry vacant */ /* TODO This needs to be reviewed and fixed, * we dont have an entry * to make an event void like vacant */ event_data->PortEntry[i].PortStatus |= MPI26_EVENT_PCIE_TOPO_PS_NO_CHANGE; } break; case MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING: _scsih_pcie_device_remove_by_handle(ioc, handle); break; } } } /** * _scsih_pcie_device_status_change_event_debug - debug for device event * @ioc: ? * @event_data: event data payload * Context: user. */ static void _scsih_pcie_device_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc, Mpi26EventDataPCIeDeviceStatusChange_t *event_data) { char *reason_str = NULL; switch (event_data->ReasonCode) { case MPI26_EVENT_PCIDEV_STAT_RC_SMART_DATA: reason_str = "smart data"; break; case MPI26_EVENT_PCIDEV_STAT_RC_UNSUPPORTED: reason_str = "unsupported device discovered"; break; case MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET: reason_str = "internal device reset"; break; case MPI26_EVENT_PCIDEV_STAT_RC_TASK_ABORT_INTERNAL: reason_str = "internal task abort"; break; case MPI26_EVENT_PCIDEV_STAT_RC_ABORT_TASK_SET_INTERNAL: reason_str = "internal task abort set"; break; case MPI26_EVENT_PCIDEV_STAT_RC_CLEAR_TASK_SET_INTERNAL: reason_str = "internal clear task set"; break; case MPI26_EVENT_PCIDEV_STAT_RC_QUERY_TASK_INTERNAL: reason_str = "internal query task"; break; case MPI26_EVENT_PCIDEV_STAT_RC_DEV_INIT_FAILURE: reason_str = "device init failure"; break; case MPI26_EVENT_PCIDEV_STAT_RC_CMP_INTERNAL_DEV_RESET: reason_str = "internal device reset complete"; break; case MPI26_EVENT_PCIDEV_STAT_RC_CMP_TASK_ABORT_INTERNAL: reason_str = "internal task abort complete"; break; case MPI26_EVENT_PCIDEV_STAT_RC_ASYNC_NOTIFICATION: reason_str = "internal async notification"; break; case MPI26_EVENT_PCIDEV_STAT_RC_PCIE_HOT_RESET_FAILED: reason_str = "pcie hot reset failed"; break; default: reason_str = "unknown reason"; break; } ioc_info(ioc, "PCIE device status change: (%s)\n" "\thandle(0x%04x), WWID(0x%016llx), tag(%d)", reason_str, le16_to_cpu(event_data->DevHandle), (u64)le64_to_cpu(event_data->WWID), le16_to_cpu(event_data->TaskTag)); if (event_data->ReasonCode == MPI26_EVENT_PCIDEV_STAT_RC_SMART_DATA) pr_cont(", ASC(0x%x), ASCQ(0x%x)\n", event_data->ASC, event_data->ASCQ); pr_cont("\n"); } /** * _scsih_pcie_device_status_change_event - handle device status * change * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. */ static void _scsih_pcie_device_status_change_event(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { struct MPT3SAS_TARGET *target_priv_data; struct _pcie_device *pcie_device; u64 wwid; unsigned long flags; Mpi26EventDataPCIeDeviceStatusChange_t *event_data = (Mpi26EventDataPCIeDeviceStatusChange_t *)fw_event->event_data; if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) _scsih_pcie_device_status_change_event_debug(ioc, event_data); if (event_data->ReasonCode != MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET && event_data->ReasonCode != MPI26_EVENT_PCIDEV_STAT_RC_CMP_INTERNAL_DEV_RESET) return; spin_lock_irqsave(&ioc->pcie_device_lock, flags); wwid = le64_to_cpu(event_data->WWID); pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid); if (!pcie_device || !pcie_device->starget) goto out; target_priv_data = pcie_device->starget->hostdata; if (!target_priv_data) goto out; if (event_data->ReasonCode == MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET) target_priv_data->tm_busy = 1; else target_priv_data->tm_busy = 0; out: if (pcie_device) pcie_device_put(pcie_device); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); } /** * _scsih_sas_enclosure_dev_status_change_event_debug - debug for enclosure * event * @ioc: per adapter object * @event_data: event data payload * Context: user. */ static void _scsih_sas_enclosure_dev_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc, Mpi2EventDataSasEnclDevStatusChange_t *event_data) { char *reason_str = NULL; switch (event_data->ReasonCode) { case MPI2_EVENT_SAS_ENCL_RC_ADDED: reason_str = "enclosure add"; break; case MPI2_EVENT_SAS_ENCL_RC_NOT_RESPONDING: reason_str = "enclosure remove"; break; default: reason_str = "unknown reason"; break; } ioc_info(ioc, "enclosure status change: (%s)\n" "\thandle(0x%04x), enclosure logical id(0x%016llx) number slots(%d)\n", reason_str, le16_to_cpu(event_data->EnclosureHandle), (u64)le64_to_cpu(event_data->EnclosureLogicalID), le16_to_cpu(event_data->StartSlot)); } /** * _scsih_sas_enclosure_dev_status_change_event - handle enclosure events * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. */ static void _scsih_sas_enclosure_dev_status_change_event(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { Mpi2ConfigReply_t mpi_reply; struct _enclosure_node *enclosure_dev = NULL; Mpi2EventDataSasEnclDevStatusChange_t *event_data = (Mpi2EventDataSasEnclDevStatusChange_t *)fw_event->event_data; int rc; u16 enclosure_handle = le16_to_cpu(event_data->EnclosureHandle); if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) _scsih_sas_enclosure_dev_status_change_event_debug(ioc, (Mpi2EventDataSasEnclDevStatusChange_t *) fw_event->event_data); if (ioc->shost_recovery) return; if (enclosure_handle) enclosure_dev = mpt3sas_scsih_enclosure_find_by_handle(ioc, enclosure_handle); switch (event_data->ReasonCode) { case MPI2_EVENT_SAS_ENCL_RC_ADDED: if (!enclosure_dev) { enclosure_dev = kzalloc(sizeof(struct _enclosure_node), GFP_KERNEL); if (!enclosure_dev) { ioc_info(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } rc = mpt3sas_config_get_enclosure_pg0(ioc, &mpi_reply, &enclosure_dev->pg0, MPI2_SAS_ENCLOS_PGAD_FORM_HANDLE, enclosure_handle); if (rc || (le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK)) { kfree(enclosure_dev); return; } list_add_tail(&enclosure_dev->list, &ioc->enclosure_list); } break; case MPI2_EVENT_SAS_ENCL_RC_NOT_RESPONDING: if (enclosure_dev) { list_del(&enclosure_dev->list); kfree(enclosure_dev); } break; default: break; } } /** * _scsih_sas_broadcast_primitive_event - handle broadcast events * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. */ static void _scsih_sas_broadcast_primitive_event(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { struct scsi_cmnd *scmd; struct scsi_device *sdev; struct scsiio_tracker *st; u16 smid, handle; u32 lun; struct MPT3SAS_DEVICE *sas_device_priv_data; u32 termination_count; u32 query_count; Mpi2SCSITaskManagementReply_t *mpi_reply; Mpi2EventDataSasBroadcastPrimitive_t *event_data = (Mpi2EventDataSasBroadcastPrimitive_t *) fw_event->event_data; u16 ioc_status; unsigned long flags; int r; u8 max_retries = 0; u8 task_abort_retries; mutex_lock(&ioc->tm_cmds.mutex); ioc_info(ioc, "%s: enter: phy number(%d), width(%d)\n", __func__, event_data->PhyNum, event_data->PortWidth); _scsih_block_io_all_device(ioc); spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); mpi_reply = ioc->tm_cmds.reply; broadcast_aen_retry: /* sanity checks for retrying this loop */ if (max_retries++ == 5) { dewtprintk(ioc, ioc_info(ioc, "%s: giving up\n", __func__)); goto out; } else if (max_retries > 1) dewtprintk(ioc, ioc_info(ioc, "%s: %d retry\n", __func__, max_retries - 1)); termination_count = 0; query_count = 0; for (smid = 1; smid <= ioc->scsiio_depth; smid++) { if (ioc->shost_recovery) goto out; scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); if (!scmd) continue; st = scsi_cmd_priv(scmd); sdev = scmd->device; sas_device_priv_data = sdev->hostdata; if (!sas_device_priv_data || !sas_device_priv_data->sas_target) continue; /* skip hidden raid components */ if (sas_device_priv_data->sas_target->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) continue; /* skip volumes */ if (sas_device_priv_data->sas_target->flags & MPT_TARGET_FLAGS_VOLUME) continue; /* skip PCIe devices */ if (sas_device_priv_data->sas_target->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) continue; handle = sas_device_priv_data->sas_target->handle; lun = sas_device_priv_data->lun; query_count++; if (ioc->shost_recovery) goto out; spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags); r = mpt3sas_scsih_issue_tm(ioc, handle, 0, 0, lun, MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK, st->smid, st->msix_io, 30, 0); if (r == FAILED) { sdev_printk(KERN_WARNING, sdev, "mpt3sas_scsih_issue_tm: FAILED when sending " "QUERY_TASK: scmd(%p)\n", scmd); spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); goto broadcast_aen_retry; } ioc_status = le16_to_cpu(mpi_reply->IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { sdev_printk(KERN_WARNING, sdev, "query task: FAILED with IOCSTATUS(0x%04x), scmd(%p)\n", ioc_status, scmd); spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); goto broadcast_aen_retry; } /* see if IO is still owned by IOC and target */ if (mpi_reply->ResponseCode == MPI2_SCSITASKMGMT_RSP_TM_SUCCEEDED || mpi_reply->ResponseCode == MPI2_SCSITASKMGMT_RSP_IO_QUEUED_ON_IOC) { spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); continue; } task_abort_retries = 0; tm_retry: if (task_abort_retries++ == 60) { dewtprintk(ioc, ioc_info(ioc, "%s: ABORT_TASK: giving up\n", __func__)); spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); goto broadcast_aen_retry; } if (ioc->shost_recovery) goto out_no_lock; r = mpt3sas_scsih_issue_tm(ioc, handle, sdev->channel, sdev->id, sdev->lun, MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK, st->smid, st->msix_io, 30, 0); if (r == FAILED || st->cb_idx != 0xFF) { sdev_printk(KERN_WARNING, sdev, "mpt3sas_scsih_issue_tm: ABORT_TASK: FAILED : " "scmd(%p)\n", scmd); goto tm_retry; } if (task_abort_retries > 1) sdev_printk(KERN_WARNING, sdev, "mpt3sas_scsih_issue_tm: ABORT_TASK: RETRIES (%d):" " scmd(%p)\n", task_abort_retries - 1, scmd); termination_count += le32_to_cpu(mpi_reply->TerminationCount); spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); } if (ioc->broadcast_aen_pending) { dewtprintk(ioc, ioc_info(ioc, "%s: loop back due to pending AEN\n", __func__)); ioc->broadcast_aen_pending = 0; goto broadcast_aen_retry; } out: spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags); out_no_lock: dewtprintk(ioc, ioc_info(ioc, "%s - exit, query_count = %d termination_count = %d\n", __func__, query_count, termination_count)); ioc->broadcast_aen_busy = 0; if (!ioc->shost_recovery) _scsih_ublock_io_all_device(ioc); mutex_unlock(&ioc->tm_cmds.mutex); } /** * _scsih_sas_discovery_event - handle discovery events * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. */ static void _scsih_sas_discovery_event(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { Mpi2EventDataSasDiscovery_t *event_data = (Mpi2EventDataSasDiscovery_t *) fw_event->event_data; if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) { ioc_info(ioc, "discovery event: (%s)", event_data->ReasonCode == MPI2_EVENT_SAS_DISC_RC_STARTED ? "start" : "stop"); if (event_data->DiscoveryStatus) pr_cont("discovery_status(0x%08x)", le32_to_cpu(event_data->DiscoveryStatus)); pr_cont("\n"); } if (event_data->ReasonCode == MPI2_EVENT_SAS_DISC_RC_STARTED && !ioc->sas_hba.num_phys) { if (disable_discovery > 0 && ioc->shost_recovery) { /* Wait for the reset to complete */ while (ioc->shost_recovery) ssleep(1); } _scsih_sas_host_add(ioc); } } /** * _scsih_sas_device_discovery_error_event - display SAS device discovery error * events * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. */ static void _scsih_sas_device_discovery_error_event(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { Mpi25EventDataSasDeviceDiscoveryError_t *event_data = (Mpi25EventDataSasDeviceDiscoveryError_t *)fw_event->event_data; switch (event_data->ReasonCode) { case MPI25_EVENT_SAS_DISC_ERR_SMP_FAILED: ioc_warn(ioc, "SMP command sent to the expander (handle:0x%04x, sas_address:0x%016llx, physical_port:0x%02x) has failed\n", le16_to_cpu(event_data->DevHandle), (u64)le64_to_cpu(event_data->SASAddress), event_data->PhysicalPort); break; case MPI25_EVENT_SAS_DISC_ERR_SMP_TIMEOUT: ioc_warn(ioc, "SMP command sent to the expander (handle:0x%04x, sas_address:0x%016llx, physical_port:0x%02x) has timed out\n", le16_to_cpu(event_data->DevHandle), (u64)le64_to_cpu(event_data->SASAddress), event_data->PhysicalPort); break; default: break; } } /** * _scsih_pcie_enumeration_event - handle enumeration events * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. */ static void _scsih_pcie_enumeration_event(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { Mpi26EventDataPCIeEnumeration_t *event_data = (Mpi26EventDataPCIeEnumeration_t *)fw_event->event_data; if (!(ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)) return; ioc_info(ioc, "pcie enumeration event: (%s) Flag 0x%02x", (event_data->ReasonCode == MPI26_EVENT_PCIE_ENUM_RC_STARTED) ? "started" : "completed", event_data->Flags); if (event_data->EnumerationStatus) pr_cont("enumeration_status(0x%08x)", le32_to_cpu(event_data->EnumerationStatus)); pr_cont("\n"); } /** * _scsih_ir_fastpath - turn on fastpath for IR physdisk * @ioc: per adapter object * @handle: device handle for physical disk * @phys_disk_num: physical disk number * * Return: 0 for success, else failure. */ static int _scsih_ir_fastpath(struct MPT3SAS_ADAPTER *ioc, u16 handle, u8 phys_disk_num) { Mpi2RaidActionRequest_t *mpi_request; Mpi2RaidActionReply_t *mpi_reply; u16 smid; u8 issue_reset = 0; int rc = 0; u16 ioc_status; u32 log_info; if (ioc->hba_mpi_version_belonged == MPI2_VERSION) return rc; mutex_lock(&ioc->scsih_cmds.mutex); if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) { ioc_err(ioc, "%s: scsih_cmd in use\n", __func__); rc = -EAGAIN; goto out; } ioc->scsih_cmds.status = MPT3_CMD_PENDING; smid = mpt3sas_base_get_smid(ioc, ioc->scsih_cb_idx); if (!smid) { ioc_err(ioc, "%s: failed obtaining a smid\n", __func__); ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; rc = -EAGAIN; goto out; } mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); ioc->scsih_cmds.smid = smid; memset(mpi_request, 0, sizeof(Mpi2RaidActionRequest_t)); mpi_request->Function = MPI2_FUNCTION_RAID_ACTION; mpi_request->Action = MPI2_RAID_ACTION_PHYSDISK_HIDDEN; mpi_request->PhysDiskNum = phys_disk_num; dewtprintk(ioc, ioc_info(ioc, "IR RAID_ACTION: turning fast path on for handle(0x%04x), phys_disk_num (0x%02x)\n", handle, phys_disk_num)); init_completion(&ioc->scsih_cmds.done); ioc->put_smid_default(ioc, smid); wait_for_completion_timeout(&ioc->scsih_cmds.done, 10*HZ); if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) { mpt3sas_check_cmd_timeout(ioc, ioc->scsih_cmds.status, mpi_request, sizeof(Mpi2RaidActionRequest_t)/4, issue_reset); rc = -EFAULT; goto out; } if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) { mpi_reply = ioc->scsih_cmds.reply; ioc_status = le16_to_cpu(mpi_reply->IOCStatus); if (ioc_status & MPI2_IOCSTATUS_FLAG_LOG_INFO_AVAILABLE) log_info = le32_to_cpu(mpi_reply->IOCLogInfo); else log_info = 0; ioc_status &= MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { dewtprintk(ioc, ioc_info(ioc, "IR RAID_ACTION: failed: ioc_status(0x%04x), loginfo(0x%08x)!!!\n", ioc_status, log_info)); rc = -EFAULT; } else dewtprintk(ioc, ioc_info(ioc, "IR RAID_ACTION: completed successfully\n")); } out: ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; mutex_unlock(&ioc->scsih_cmds.mutex); if (issue_reset) mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); return rc; } /** * _scsih_reprobe_lun - reprobing lun * @sdev: scsi device struct * @no_uld_attach: sdev->no_uld_attach flag setting * **/ static void _scsih_reprobe_lun(struct scsi_device *sdev, void *no_uld_attach) { sdev->no_uld_attach = no_uld_attach ? 1 : 0; sdev_printk(KERN_INFO, sdev, "%s raid component\n", sdev->no_uld_attach ? "hiding" : "exposing"); WARN_ON(scsi_device_reprobe(sdev)); } /** * _scsih_sas_volume_add - add new volume * @ioc: per adapter object * @element: IR config element data * Context: user. */ static void _scsih_sas_volume_add(struct MPT3SAS_ADAPTER *ioc, Mpi2EventIrConfigElement_t *element) { struct _raid_device *raid_device; unsigned long flags; u64 wwid; u16 handle = le16_to_cpu(element->VolDevHandle); int rc; mpt3sas_config_get_volume_wwid(ioc, handle, &wwid); if (!wwid) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } spin_lock_irqsave(&ioc->raid_device_lock, flags); raid_device = _scsih_raid_device_find_by_wwid(ioc, wwid); spin_unlock_irqrestore(&ioc->raid_device_lock, flags); if (raid_device) return; raid_device = kzalloc(sizeof(struct _raid_device), GFP_KERNEL); if (!raid_device) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } raid_device->id = ioc->sas_id++; raid_device->channel = RAID_CHANNEL; raid_device->handle = handle; raid_device->wwid = wwid; _scsih_raid_device_add(ioc, raid_device); if (!ioc->wait_for_discovery_to_complete) { rc = scsi_add_device(ioc->shost, RAID_CHANNEL, raid_device->id, 0); if (rc) _scsih_raid_device_remove(ioc, raid_device); } else { spin_lock_irqsave(&ioc->raid_device_lock, flags); _scsih_determine_boot_device(ioc, raid_device, 1); spin_unlock_irqrestore(&ioc->raid_device_lock, flags); } } /** * _scsih_sas_volume_delete - delete volume * @ioc: per adapter object * @handle: volume device handle * Context: user. */ static void _scsih_sas_volume_delete(struct MPT3SAS_ADAPTER *ioc, u16 handle) { struct _raid_device *raid_device; unsigned long flags; struct MPT3SAS_TARGET *sas_target_priv_data; struct scsi_target *starget = NULL; spin_lock_irqsave(&ioc->raid_device_lock, flags); raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); if (raid_device) { if (raid_device->starget) { starget = raid_device->starget; sas_target_priv_data = starget->hostdata; sas_target_priv_data->deleted = 1; } ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n", raid_device->handle, (u64)raid_device->wwid); list_del(&raid_device->list); kfree(raid_device); } spin_unlock_irqrestore(&ioc->raid_device_lock, flags); if (starget) scsi_remove_target(&starget->dev); } /** * _scsih_sas_pd_expose - expose pd component to /dev/sdX * @ioc: per adapter object * @element: IR config element data * Context: user. */ static void _scsih_sas_pd_expose(struct MPT3SAS_ADAPTER *ioc, Mpi2EventIrConfigElement_t *element) { struct _sas_device *sas_device; struct scsi_target *starget = NULL; struct MPT3SAS_TARGET *sas_target_priv_data; unsigned long flags; u16 handle = le16_to_cpu(element->PhysDiskDevHandle); spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); if (sas_device) { sas_device->volume_handle = 0; sas_device->volume_wwid = 0; clear_bit(handle, ioc->pd_handles); if (sas_device->starget && sas_device->starget->hostdata) { starget = sas_device->starget; sas_target_priv_data = starget->hostdata; sas_target_priv_data->flags &= ~MPT_TARGET_FLAGS_RAID_COMPONENT; } } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); if (!sas_device) return; /* exposing raid component */ if (starget) starget_for_each_device(starget, NULL, _scsih_reprobe_lun); sas_device_put(sas_device); } /** * _scsih_sas_pd_hide - hide pd component from /dev/sdX * @ioc: per adapter object * @element: IR config element data * Context: user. */ static void _scsih_sas_pd_hide(struct MPT3SAS_ADAPTER *ioc, Mpi2EventIrConfigElement_t *element) { struct _sas_device *sas_device; struct scsi_target *starget = NULL; struct MPT3SAS_TARGET *sas_target_priv_data; unsigned long flags; u16 handle = le16_to_cpu(element->PhysDiskDevHandle); u16 volume_handle = 0; u64 volume_wwid = 0; mpt3sas_config_get_volume_handle(ioc, handle, &volume_handle); if (volume_handle) mpt3sas_config_get_volume_wwid(ioc, volume_handle, &volume_wwid); spin_lock_irqsave(&ioc->sas_device_lock, flags); sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); if (sas_device) { set_bit(handle, ioc->pd_handles); if (sas_device->starget && sas_device->starget->hostdata) { starget = sas_device->starget; sas_target_priv_data = starget->hostdata; sas_target_priv_data->flags |= MPT_TARGET_FLAGS_RAID_COMPONENT; sas_device->volume_handle = volume_handle; sas_device->volume_wwid = volume_wwid; } } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); if (!sas_device) return; /* hiding raid component */ _scsih_ir_fastpath(ioc, handle, element->PhysDiskNum); if (starget) starget_for_each_device(starget, (void *)1, _scsih_reprobe_lun); sas_device_put(sas_device); } /** * _scsih_sas_pd_delete - delete pd component * @ioc: per adapter object * @element: IR config element data * Context: user. */ static void _scsih_sas_pd_delete(struct MPT3SAS_ADAPTER *ioc, Mpi2EventIrConfigElement_t *element) { u16 handle = le16_to_cpu(element->PhysDiskDevHandle); _scsih_device_remove_by_handle(ioc, handle); } /** * _scsih_sas_pd_add - remove pd component * @ioc: per adapter object * @element: IR config element data * Context: user. */ static void _scsih_sas_pd_add(struct MPT3SAS_ADAPTER *ioc, Mpi2EventIrConfigElement_t *element) { struct _sas_device *sas_device; u16 handle = le16_to_cpu(element->PhysDiskDevHandle); Mpi2ConfigReply_t mpi_reply; Mpi2SasDevicePage0_t sas_device_pg0; u32 ioc_status; u64 sas_address; u16 parent_handle; set_bit(handle, ioc->pd_handles); sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); if (sas_device) { _scsih_ir_fastpath(ioc, handle, element->PhysDiskNum); sas_device_put(sas_device); return; } if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle); if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address)) mpt3sas_transport_update_links(ioc, sas_address, handle, sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5, mpt3sas_get_port_by_id(ioc, sas_device_pg0.PhysicalPort, 0)); _scsih_ir_fastpath(ioc, handle, element->PhysDiskNum); _scsih_add_device(ioc, handle, 0, 1); } /** * _scsih_sas_ir_config_change_event_debug - debug for IR Config Change events * @ioc: per adapter object * @event_data: event data payload * Context: user. */ static void _scsih_sas_ir_config_change_event_debug(struct MPT3SAS_ADAPTER *ioc, Mpi2EventDataIrConfigChangeList_t *event_data) { Mpi2EventIrConfigElement_t *element; u8 element_type; int i; char *reason_str = NULL, *element_str = NULL; element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; ioc_info(ioc, "raid config change: (%s), elements(%d)\n", le32_to_cpu(event_data->Flags) & MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG ? "foreign" : "native", event_data->NumElements); for (i = 0; i < event_data->NumElements; i++, element++) { switch (element->ReasonCode) { case MPI2_EVENT_IR_CHANGE_RC_ADDED: reason_str = "add"; break; case MPI2_EVENT_IR_CHANGE_RC_REMOVED: reason_str = "remove"; break; case MPI2_EVENT_IR_CHANGE_RC_NO_CHANGE: reason_str = "no change"; break; case MPI2_EVENT_IR_CHANGE_RC_HIDE: reason_str = "hide"; break; case MPI2_EVENT_IR_CHANGE_RC_UNHIDE: reason_str = "unhide"; break; case MPI2_EVENT_IR_CHANGE_RC_VOLUME_CREATED: reason_str = "volume_created"; break; case MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED: reason_str = "volume_deleted"; break; case MPI2_EVENT_IR_CHANGE_RC_PD_CREATED: reason_str = "pd_created"; break; case MPI2_EVENT_IR_CHANGE_RC_PD_DELETED: reason_str = "pd_deleted"; break; default: reason_str = "unknown reason"; break; } element_type = le16_to_cpu(element->ElementFlags) & MPI2_EVENT_IR_CHANGE_EFLAGS_ELEMENT_TYPE_MASK; switch (element_type) { case MPI2_EVENT_IR_CHANGE_EFLAGS_VOLUME_ELEMENT: element_str = "volume"; break; case MPI2_EVENT_IR_CHANGE_EFLAGS_VOLPHYSDISK_ELEMENT: element_str = "phys disk"; break; case MPI2_EVENT_IR_CHANGE_EFLAGS_HOTSPARE_ELEMENT: element_str = "hot spare"; break; default: element_str = "unknown element"; break; } pr_info("\t(%s:%s), vol handle(0x%04x), " \ "pd handle(0x%04x), pd num(0x%02x)\n", element_str, reason_str, le16_to_cpu(element->VolDevHandle), le16_to_cpu(element->PhysDiskDevHandle), element->PhysDiskNum); } } /** * _scsih_sas_ir_config_change_event - handle ir configuration change events * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. */ static void _scsih_sas_ir_config_change_event(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { Mpi2EventIrConfigElement_t *element; int i; u8 foreign_config; Mpi2EventDataIrConfigChangeList_t *event_data = (Mpi2EventDataIrConfigChangeList_t *) fw_event->event_data; if ((ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) && (!ioc->hide_ir_msg)) _scsih_sas_ir_config_change_event_debug(ioc, event_data); foreign_config = (le32_to_cpu(event_data->Flags) & MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG) ? 1 : 0; element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; if (ioc->shost_recovery && ioc->hba_mpi_version_belonged != MPI2_VERSION) { for (i = 0; i < event_data->NumElements; i++, element++) { if (element->ReasonCode == MPI2_EVENT_IR_CHANGE_RC_HIDE) _scsih_ir_fastpath(ioc, le16_to_cpu(element->PhysDiskDevHandle), element->PhysDiskNum); } return; } for (i = 0; i < event_data->NumElements; i++, element++) { switch (element->ReasonCode) { case MPI2_EVENT_IR_CHANGE_RC_VOLUME_CREATED: case MPI2_EVENT_IR_CHANGE_RC_ADDED: if (!foreign_config) _scsih_sas_volume_add(ioc, element); break; case MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED: case MPI2_EVENT_IR_CHANGE_RC_REMOVED: if (!foreign_config) _scsih_sas_volume_delete(ioc, le16_to_cpu(element->VolDevHandle)); break; case MPI2_EVENT_IR_CHANGE_RC_PD_CREATED: if (!ioc->is_warpdrive) _scsih_sas_pd_hide(ioc, element); break; case MPI2_EVENT_IR_CHANGE_RC_PD_DELETED: if (!ioc->is_warpdrive) _scsih_sas_pd_expose(ioc, element); break; case MPI2_EVENT_IR_CHANGE_RC_HIDE: if (!ioc->is_warpdrive) _scsih_sas_pd_add(ioc, element); break; case MPI2_EVENT_IR_CHANGE_RC_UNHIDE: if (!ioc->is_warpdrive) _scsih_sas_pd_delete(ioc, element); break; } } } /** * _scsih_sas_ir_volume_event - IR volume event * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. */ static void _scsih_sas_ir_volume_event(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { u64 wwid; unsigned long flags; struct _raid_device *raid_device; u16 handle; u32 state; int rc; Mpi2EventDataIrVolume_t *event_data = (Mpi2EventDataIrVolume_t *) fw_event->event_data; if (ioc->shost_recovery) return; if (event_data->ReasonCode != MPI2_EVENT_IR_VOLUME_RC_STATE_CHANGED) return; handle = le16_to_cpu(event_data->VolDevHandle); state = le32_to_cpu(event_data->NewValue); if (!ioc->hide_ir_msg) dewtprintk(ioc, ioc_info(ioc, "%s: handle(0x%04x), old(0x%08x), new(0x%08x)\n", __func__, handle, le32_to_cpu(event_data->PreviousValue), state)); switch (state) { case MPI2_RAID_VOL_STATE_MISSING: case MPI2_RAID_VOL_STATE_FAILED: _scsih_sas_volume_delete(ioc, handle); break; case MPI2_RAID_VOL_STATE_ONLINE: case MPI2_RAID_VOL_STATE_DEGRADED: case MPI2_RAID_VOL_STATE_OPTIMAL: spin_lock_irqsave(&ioc->raid_device_lock, flags); raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); spin_unlock_irqrestore(&ioc->raid_device_lock, flags); if (raid_device) break; mpt3sas_config_get_volume_wwid(ioc, handle, &wwid); if (!wwid) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); break; } raid_device = kzalloc(sizeof(struct _raid_device), GFP_KERNEL); if (!raid_device) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); break; } raid_device->id = ioc->sas_id++; raid_device->channel = RAID_CHANNEL; raid_device->handle = handle; raid_device->wwid = wwid; _scsih_raid_device_add(ioc, raid_device); rc = scsi_add_device(ioc->shost, RAID_CHANNEL, raid_device->id, 0); if (rc) _scsih_raid_device_remove(ioc, raid_device); break; case MPI2_RAID_VOL_STATE_INITIALIZING: default: break; } } /** * _scsih_sas_ir_physical_disk_event - PD event * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. */ static void _scsih_sas_ir_physical_disk_event(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { u16 handle, parent_handle; u32 state; struct _sas_device *sas_device; Mpi2ConfigReply_t mpi_reply; Mpi2SasDevicePage0_t sas_device_pg0; u32 ioc_status; Mpi2EventDataIrPhysicalDisk_t *event_data = (Mpi2EventDataIrPhysicalDisk_t *) fw_event->event_data; u64 sas_address; if (ioc->shost_recovery) return; if (event_data->ReasonCode != MPI2_EVENT_IR_PHYSDISK_RC_STATE_CHANGED) return; handle = le16_to_cpu(event_data->PhysDiskDevHandle); state = le32_to_cpu(event_data->NewValue); if (!ioc->hide_ir_msg) dewtprintk(ioc, ioc_info(ioc, "%s: handle(0x%04x), old(0x%08x), new(0x%08x)\n", __func__, handle, le32_to_cpu(event_data->PreviousValue), state)); switch (state) { case MPI2_RAID_PD_STATE_ONLINE: case MPI2_RAID_PD_STATE_DEGRADED: case MPI2_RAID_PD_STATE_REBUILDING: case MPI2_RAID_PD_STATE_OPTIMAL: case MPI2_RAID_PD_STATE_HOT_SPARE: if (!ioc->is_warpdrive) set_bit(handle, ioc->pd_handles); sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); if (sas_device) { sas_device_put(sas_device); return; } if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle); if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address)) mpt3sas_transport_update_links(ioc, sas_address, handle, sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5, mpt3sas_get_port_by_id(ioc, sas_device_pg0.PhysicalPort, 0)); _scsih_add_device(ioc, handle, 0, 1); break; case MPI2_RAID_PD_STATE_OFFLINE: case MPI2_RAID_PD_STATE_NOT_CONFIGURED: case MPI2_RAID_PD_STATE_NOT_COMPATIBLE: default: break; } } /** * _scsih_sas_ir_operation_status_event_debug - debug for IR op event * @ioc: per adapter object * @event_data: event data payload * Context: user. */ static void _scsih_sas_ir_operation_status_event_debug(struct MPT3SAS_ADAPTER *ioc, Mpi2EventDataIrOperationStatus_t *event_data) { char *reason_str = NULL; switch (event_data->RAIDOperation) { case MPI2_EVENT_IR_RAIDOP_RESYNC: reason_str = "resync"; break; case MPI2_EVENT_IR_RAIDOP_ONLINE_CAP_EXPANSION: reason_str = "online capacity expansion"; break; case MPI2_EVENT_IR_RAIDOP_CONSISTENCY_CHECK: reason_str = "consistency check"; break; case MPI2_EVENT_IR_RAIDOP_BACKGROUND_INIT: reason_str = "background init"; break; case MPI2_EVENT_IR_RAIDOP_MAKE_DATA_CONSISTENT: reason_str = "make data consistent"; break; } if (!reason_str) return; ioc_info(ioc, "raid operational status: (%s)\thandle(0x%04x), percent complete(%d)\n", reason_str, le16_to_cpu(event_data->VolDevHandle), event_data->PercentComplete); } /** * _scsih_sas_ir_operation_status_event - handle RAID operation events * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. */ static void _scsih_sas_ir_operation_status_event(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { Mpi2EventDataIrOperationStatus_t *event_data = (Mpi2EventDataIrOperationStatus_t *) fw_event->event_data; static struct _raid_device *raid_device; unsigned long flags; u16 handle; if ((ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) && (!ioc->hide_ir_msg)) _scsih_sas_ir_operation_status_event_debug(ioc, event_data); /* code added for raid transport support */ if (event_data->RAIDOperation == MPI2_EVENT_IR_RAIDOP_RESYNC) { spin_lock_irqsave(&ioc->raid_device_lock, flags); handle = le16_to_cpu(event_data->VolDevHandle); raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); if (raid_device) raid_device->percent_complete = event_data->PercentComplete; spin_unlock_irqrestore(&ioc->raid_device_lock, flags); } } /** * _scsih_prep_device_scan - initialize parameters prior to device scan * @ioc: per adapter object * * Set the deleted flag prior to device scan. If the device is found during * the scan, then we clear the deleted flag. */ static void _scsih_prep_device_scan(struct MPT3SAS_ADAPTER *ioc) { struct MPT3SAS_DEVICE *sas_device_priv_data; struct scsi_device *sdev; shost_for_each_device(sdev, ioc->shost) { sas_device_priv_data = sdev->hostdata; if (sas_device_priv_data && sas_device_priv_data->sas_target) sas_device_priv_data->sas_target->deleted = 1; } } /** * _scsih_mark_responding_sas_device - mark a sas_devices as responding * @ioc: per adapter object * @sas_device_pg0: SAS Device page 0 * * After host reset, find out whether devices are still responding. * Used in _scsih_remove_unresponsive_sas_devices. */ static void _scsih_mark_responding_sas_device(struct MPT3SAS_ADAPTER *ioc, Mpi2SasDevicePage0_t *sas_device_pg0) { struct MPT3SAS_TARGET *sas_target_priv_data = NULL; struct scsi_target *starget; struct _sas_device *sas_device = NULL; struct _enclosure_node *enclosure_dev = NULL; unsigned long flags; struct hba_port *port = mpt3sas_get_port_by_id( ioc, sas_device_pg0->PhysicalPort, 0); if (sas_device_pg0->EnclosureHandle) { enclosure_dev = mpt3sas_scsih_enclosure_find_by_handle(ioc, le16_to_cpu(sas_device_pg0->EnclosureHandle)); if (enclosure_dev == NULL) ioc_info(ioc, "Enclosure handle(0x%04x) doesn't match with enclosure device!\n", sas_device_pg0->EnclosureHandle); } spin_lock_irqsave(&ioc->sas_device_lock, flags); list_for_each_entry(sas_device, &ioc->sas_device_list, list) { if (sas_device->sas_address != le64_to_cpu( sas_device_pg0->SASAddress)) continue; if (sas_device->slot != le16_to_cpu(sas_device_pg0->Slot)) continue; if (sas_device->port != port) continue; sas_device->responding = 1; starget = sas_device->starget; if (starget && starget->hostdata) { sas_target_priv_data = starget->hostdata; sas_target_priv_data->tm_busy = 0; sas_target_priv_data->deleted = 0; } else sas_target_priv_data = NULL; if (starget) { starget_printk(KERN_INFO, starget, "handle(0x%04x), sas_addr(0x%016llx)\n", le16_to_cpu(sas_device_pg0->DevHandle), (unsigned long long) sas_device->sas_address); if (sas_device->enclosure_handle != 0) starget_printk(KERN_INFO, starget, "enclosure logical id(0x%016llx), slot(%d)\n", (unsigned long long) sas_device->enclosure_logical_id, sas_device->slot); } if (le16_to_cpu(sas_device_pg0->Flags) & MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) { sas_device->enclosure_level = sas_device_pg0->EnclosureLevel; memcpy(&sas_device->connector_name[0], &sas_device_pg0->ConnectorName[0], 4); } else { sas_device->enclosure_level = 0; sas_device->connector_name[0] = '\0'; } sas_device->enclosure_handle = le16_to_cpu(sas_device_pg0->EnclosureHandle); sas_device->is_chassis_slot_valid = 0; if (enclosure_dev) { sas_device->enclosure_logical_id = le64_to_cpu( enclosure_dev->pg0.EnclosureLogicalID); if (le16_to_cpu(enclosure_dev->pg0.Flags) & MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) { sas_device->is_chassis_slot_valid = 1; sas_device->chassis_slot = enclosure_dev->pg0.ChassisSlot; } } if (sas_device->handle == le16_to_cpu( sas_device_pg0->DevHandle)) goto out; pr_info("\thandle changed from(0x%04x)!!!\n", sas_device->handle); sas_device->handle = le16_to_cpu( sas_device_pg0->DevHandle); if (sas_target_priv_data) sas_target_priv_data->handle = le16_to_cpu(sas_device_pg0->DevHandle); goto out; } out: spin_unlock_irqrestore(&ioc->sas_device_lock, flags); } /** * _scsih_create_enclosure_list_after_reset - Free Existing list, * And create enclosure list by scanning all Enclosure Page(0)s * @ioc: per adapter object */ static void _scsih_create_enclosure_list_after_reset(struct MPT3SAS_ADAPTER *ioc) { struct _enclosure_node *enclosure_dev; Mpi2ConfigReply_t mpi_reply; u16 enclosure_handle; int rc; /* Free existing enclosure list */ mpt3sas_free_enclosure_list(ioc); /* Re constructing enclosure list after reset*/ enclosure_handle = 0xFFFF; do { enclosure_dev = kzalloc(sizeof(struct _enclosure_node), GFP_KERNEL); if (!enclosure_dev) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } rc = mpt3sas_config_get_enclosure_pg0(ioc, &mpi_reply, &enclosure_dev->pg0, MPI2_SAS_ENCLOS_PGAD_FORM_GET_NEXT_HANDLE, enclosure_handle); if (rc || (le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK)) { kfree(enclosure_dev); return; } list_add_tail(&enclosure_dev->list, &ioc->enclosure_list); enclosure_handle = le16_to_cpu(enclosure_dev->pg0.EnclosureHandle); } while (1); } /** * _scsih_search_responding_sas_devices - * @ioc: per adapter object * * After host reset, find out whether devices are still responding. * If not remove. */ static void _scsih_search_responding_sas_devices(struct MPT3SAS_ADAPTER *ioc) { Mpi2SasDevicePage0_t sas_device_pg0; Mpi2ConfigReply_t mpi_reply; u16 ioc_status; u16 handle; u32 device_info; ioc_info(ioc, "search for end-devices: start\n"); if (list_empty(&ioc->sas_device_list)) goto out; handle = 0xFFFF; while (!(mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_GET_NEXT_HANDLE, handle))) { ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) break; handle = le16_to_cpu(sas_device_pg0.DevHandle); device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); if (!(_scsih_is_end_device(device_info))) continue; _scsih_mark_responding_sas_device(ioc, &sas_device_pg0); } out: ioc_info(ioc, "search for end-devices: complete\n"); } /** * _scsih_mark_responding_pcie_device - mark a pcie_device as responding * @ioc: per adapter object * @pcie_device_pg0: PCIe Device page 0 * * After host reset, find out whether devices are still responding. * Used in _scsih_remove_unresponding_devices. */ static void _scsih_mark_responding_pcie_device(struct MPT3SAS_ADAPTER *ioc, Mpi26PCIeDevicePage0_t *pcie_device_pg0) { struct MPT3SAS_TARGET *sas_target_priv_data = NULL; struct scsi_target *starget; struct _pcie_device *pcie_device; unsigned long flags; spin_lock_irqsave(&ioc->pcie_device_lock, flags); list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) { if ((pcie_device->wwid == le64_to_cpu(pcie_device_pg0->WWID)) && (pcie_device->slot == le16_to_cpu( pcie_device_pg0->Slot))) { pcie_device->access_status = pcie_device_pg0->AccessStatus; pcie_device->responding = 1; starget = pcie_device->starget; if (starget && starget->hostdata) { sas_target_priv_data = starget->hostdata; sas_target_priv_data->tm_busy = 0; sas_target_priv_data->deleted = 0; } else sas_target_priv_data = NULL; if (starget) { starget_printk(KERN_INFO, starget, "handle(0x%04x), wwid(0x%016llx) ", pcie_device->handle, (unsigned long long)pcie_device->wwid); if (pcie_device->enclosure_handle != 0) starget_printk(KERN_INFO, starget, "enclosure logical id(0x%016llx), " "slot(%d)\n", (unsigned long long) pcie_device->enclosure_logical_id, pcie_device->slot); } if (((le32_to_cpu(pcie_device_pg0->Flags)) & MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) && (ioc->hba_mpi_version_belonged != MPI2_VERSION)) { pcie_device->enclosure_level = pcie_device_pg0->EnclosureLevel; memcpy(&pcie_device->connector_name[0], &pcie_device_pg0->ConnectorName[0], 4); } else { pcie_device->enclosure_level = 0; pcie_device->connector_name[0] = '\0'; } if (pcie_device->handle == le16_to_cpu( pcie_device_pg0->DevHandle)) goto out; pr_info("\thandle changed from(0x%04x)!!!\n", pcie_device->handle); pcie_device->handle = le16_to_cpu( pcie_device_pg0->DevHandle); if (sas_target_priv_data) sas_target_priv_data->handle = le16_to_cpu(pcie_device_pg0->DevHandle); goto out; } } out: spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); } /** * _scsih_search_responding_pcie_devices - * @ioc: per adapter object * * After host reset, find out whether devices are still responding. * If not remove. */ static void _scsih_search_responding_pcie_devices(struct MPT3SAS_ADAPTER *ioc) { Mpi26PCIeDevicePage0_t pcie_device_pg0; Mpi2ConfigReply_t mpi_reply; u16 ioc_status; u16 handle; u32 device_info; ioc_info(ioc, "search for end-devices: start\n"); if (list_empty(&ioc->pcie_device_list)) goto out; handle = 0xFFFF; while (!(mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply, &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_GET_NEXT_HANDLE, handle))) { ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_info(ioc, "\tbreak from %s: ioc_status(0x%04x), loginfo(0x%08x)\n", __func__, ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); break; } handle = le16_to_cpu(pcie_device_pg0.DevHandle); device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo); if (!(_scsih_is_nvme_pciescsi_device(device_info))) continue; _scsih_mark_responding_pcie_device(ioc, &pcie_device_pg0); } out: ioc_info(ioc, "search for PCIe end-devices: complete\n"); } /** * _scsih_mark_responding_raid_device - mark a raid_device as responding * @ioc: per adapter object * @wwid: world wide identifier for raid volume * @handle: device handle * * After host reset, find out whether devices are still responding. * Used in _scsih_remove_unresponsive_raid_devices. */ static void _scsih_mark_responding_raid_device(struct MPT3SAS_ADAPTER *ioc, u64 wwid, u16 handle) { struct MPT3SAS_TARGET *sas_target_priv_data = NULL; struct scsi_target *starget; struct _raid_device *raid_device; unsigned long flags; spin_lock_irqsave(&ioc->raid_device_lock, flags); list_for_each_entry(raid_device, &ioc->raid_device_list, list) { if (raid_device->wwid == wwid && raid_device->starget) { starget = raid_device->starget; if (starget && starget->hostdata) { sas_target_priv_data = starget->hostdata; sas_target_priv_data->deleted = 0; } else sas_target_priv_data = NULL; raid_device->responding = 1; spin_unlock_irqrestore(&ioc->raid_device_lock, flags); starget_printk(KERN_INFO, raid_device->starget, "handle(0x%04x), wwid(0x%016llx)\n", handle, (unsigned long long)raid_device->wwid); /* * WARPDRIVE: The handles of the PDs might have changed * across the host reset so re-initialize the * required data for Direct IO */ mpt3sas_init_warpdrive_properties(ioc, raid_device); spin_lock_irqsave(&ioc->raid_device_lock, flags); if (raid_device->handle == handle) { spin_unlock_irqrestore(&ioc->raid_device_lock, flags); return; } pr_info("\thandle changed from(0x%04x)!!!\n", raid_device->handle); raid_device->handle = handle; if (sas_target_priv_data) sas_target_priv_data->handle = handle; spin_unlock_irqrestore(&ioc->raid_device_lock, flags); return; } } spin_unlock_irqrestore(&ioc->raid_device_lock, flags); } /** * _scsih_search_responding_raid_devices - * @ioc: per adapter object * * After host reset, find out whether devices are still responding. * If not remove. */ static void _scsih_search_responding_raid_devices(struct MPT3SAS_ADAPTER *ioc) { Mpi2RaidVolPage1_t volume_pg1; Mpi2RaidVolPage0_t volume_pg0; Mpi2RaidPhysDiskPage0_t pd_pg0; Mpi2ConfigReply_t mpi_reply; u16 ioc_status; u16 handle; u8 phys_disk_num; if (!ioc->ir_firmware) return; ioc_info(ioc, "search for raid volumes: start\n"); if (list_empty(&ioc->raid_device_list)) goto out; handle = 0xFFFF; while (!(mpt3sas_config_get_raid_volume_pg1(ioc, &mpi_reply, &volume_pg1, MPI2_RAID_VOLUME_PGAD_FORM_GET_NEXT_HANDLE, handle))) { ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) break; handle = le16_to_cpu(volume_pg1.DevHandle); if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, &volume_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle, sizeof(Mpi2RaidVolPage0_t))) continue; if (volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_OPTIMAL || volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_ONLINE || volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_DEGRADED) _scsih_mark_responding_raid_device(ioc, le64_to_cpu(volume_pg1.WWID), handle); } /* refresh the pd_handles */ if (!ioc->is_warpdrive) { phys_disk_num = 0xFF; memset(ioc->pd_handles, 0, ioc->pd_handles_sz); while (!(mpt3sas_config_get_phys_disk_pg0(ioc, &mpi_reply, &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_GET_NEXT_PHYSDISKNUM, phys_disk_num))) { ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) break; phys_disk_num = pd_pg0.PhysDiskNum; handle = le16_to_cpu(pd_pg0.DevHandle); set_bit(handle, ioc->pd_handles); } } out: ioc_info(ioc, "search for responding raid volumes: complete\n"); } /** * _scsih_mark_responding_expander - mark a expander as responding * @ioc: per adapter object * @expander_pg0:SAS Expander Config Page0 * * After host reset, find out whether devices are still responding. * Used in _scsih_remove_unresponsive_expanders. */ static void _scsih_mark_responding_expander(struct MPT3SAS_ADAPTER *ioc, Mpi2ExpanderPage0_t *expander_pg0) { struct _sas_node *sas_expander = NULL; unsigned long flags; int i; struct _enclosure_node *enclosure_dev = NULL; u16 handle = le16_to_cpu(expander_pg0->DevHandle); u16 enclosure_handle = le16_to_cpu(expander_pg0->EnclosureHandle); u64 sas_address = le64_to_cpu(expander_pg0->SASAddress); struct hba_port *port = mpt3sas_get_port_by_id( ioc, expander_pg0->PhysicalPort, 0); if (enclosure_handle) enclosure_dev = mpt3sas_scsih_enclosure_find_by_handle(ioc, enclosure_handle); spin_lock_irqsave(&ioc->sas_node_lock, flags); list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) { if (sas_expander->sas_address != sas_address) continue; if (sas_expander->port != port) continue; sas_expander->responding = 1; if (enclosure_dev) { sas_expander->enclosure_logical_id = le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); sas_expander->enclosure_handle = le16_to_cpu(expander_pg0->EnclosureHandle); } if (sas_expander->handle == handle) goto out; pr_info("\texpander(0x%016llx): handle changed" \ " from(0x%04x) to (0x%04x)!!!\n", (unsigned long long)sas_expander->sas_address, sas_expander->handle, handle); sas_expander->handle = handle; for (i = 0 ; i < sas_expander->num_phys ; i++) sas_expander->phy[i].handle = handle; goto out; } out: spin_unlock_irqrestore(&ioc->sas_node_lock, flags); } /** * _scsih_search_responding_expanders - * @ioc: per adapter object * * After host reset, find out whether devices are still responding. * If not remove. */ static void _scsih_search_responding_expanders(struct MPT3SAS_ADAPTER *ioc) { Mpi2ExpanderPage0_t expander_pg0; Mpi2ConfigReply_t mpi_reply; u16 ioc_status; u64 sas_address; u16 handle; u8 port; ioc_info(ioc, "search for expanders: start\n"); if (list_empty(&ioc->sas_expander_list)) goto out; handle = 0xFFFF; while (!(mpt3sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0, MPI2_SAS_EXPAND_PGAD_FORM_GET_NEXT_HNDL, handle))) { ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) break; handle = le16_to_cpu(expander_pg0.DevHandle); sas_address = le64_to_cpu(expander_pg0.SASAddress); port = expander_pg0.PhysicalPort; pr_info( "\texpander present: handle(0x%04x), sas_addr(0x%016llx), port:%d\n", handle, (unsigned long long)sas_address, (ioc->multipath_on_hba ? port : MULTIPATH_DISABLED_PORT_ID)); _scsih_mark_responding_expander(ioc, &expander_pg0); } out: ioc_info(ioc, "search for expanders: complete\n"); } /** * _scsih_remove_unresponding_devices - removing unresponding devices * @ioc: per adapter object */ static void _scsih_remove_unresponding_devices(struct MPT3SAS_ADAPTER *ioc) { struct _sas_device *sas_device, *sas_device_next; struct _sas_node *sas_expander, *sas_expander_next; struct _raid_device *raid_device, *raid_device_next; struct _pcie_device *pcie_device, *pcie_device_next; struct list_head tmp_list; unsigned long flags; LIST_HEAD(head); ioc_info(ioc, "removing unresponding devices: start\n"); /* removing unresponding end devices */ ioc_info(ioc, "removing unresponding devices: end-devices\n"); /* * Iterate, pulling off devices marked as non-responding. We become the * owner for the reference the list had on any object we prune. */ spin_lock_irqsave(&ioc->sas_device_lock, flags); /* * Clean up the sas_device_init_list list as * driver goes for fresh scan as part of diag reset. */ list_for_each_entry_safe(sas_device, sas_device_next, &ioc->sas_device_init_list, list) { list_del_init(&sas_device->list); sas_device_put(sas_device); } list_for_each_entry_safe(sas_device, sas_device_next, &ioc->sas_device_list, list) { if (!sas_device->responding) list_move_tail(&sas_device->list, &head); else sas_device->responding = 0; } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); /* * Now, uninitialize and remove the unresponding devices we pruned. */ list_for_each_entry_safe(sas_device, sas_device_next, &head, list) { _scsih_remove_device(ioc, sas_device); list_del_init(&sas_device->list); sas_device_put(sas_device); } ioc_info(ioc, "Removing unresponding devices: pcie end-devices\n"); INIT_LIST_HEAD(&head); spin_lock_irqsave(&ioc->pcie_device_lock, flags); /* * Clean up the pcie_device_init_list list as * driver goes for fresh scan as part of diag reset. */ list_for_each_entry_safe(pcie_device, pcie_device_next, &ioc->pcie_device_init_list, list) { list_del_init(&pcie_device->list); pcie_device_put(pcie_device); } list_for_each_entry_safe(pcie_device, pcie_device_next, &ioc->pcie_device_list, list) { if (!pcie_device->responding) list_move_tail(&pcie_device->list, &head); else pcie_device->responding = 0; } spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); list_for_each_entry_safe(pcie_device, pcie_device_next, &head, list) { _scsih_pcie_device_remove_from_sml(ioc, pcie_device); list_del_init(&pcie_device->list); pcie_device_put(pcie_device); } /* removing unresponding volumes */ if (ioc->ir_firmware) { ioc_info(ioc, "removing unresponding devices: volumes\n"); list_for_each_entry_safe(raid_device, raid_device_next, &ioc->raid_device_list, list) { if (!raid_device->responding) _scsih_sas_volume_delete(ioc, raid_device->handle); else raid_device->responding = 0; } } /* removing unresponding expanders */ ioc_info(ioc, "removing unresponding devices: expanders\n"); spin_lock_irqsave(&ioc->sas_node_lock, flags); INIT_LIST_HEAD(&tmp_list); list_for_each_entry_safe(sas_expander, sas_expander_next, &ioc->sas_expander_list, list) { if (!sas_expander->responding) list_move_tail(&sas_expander->list, &tmp_list); else sas_expander->responding = 0; } spin_unlock_irqrestore(&ioc->sas_node_lock, flags); list_for_each_entry_safe(sas_expander, sas_expander_next, &tmp_list, list) { _scsih_expander_node_remove(ioc, sas_expander); } ioc_info(ioc, "removing unresponding devices: complete\n"); /* unblock devices */ _scsih_ublock_io_all_device(ioc); } static void _scsih_refresh_expander_links(struct MPT3SAS_ADAPTER *ioc, struct _sas_node *sas_expander, u16 handle) { Mpi2ExpanderPage1_t expander_pg1; Mpi2ConfigReply_t mpi_reply; int i; for (i = 0 ; i < sas_expander->num_phys ; i++) { if ((mpt3sas_config_get_expander_pg1(ioc, &mpi_reply, &expander_pg1, i, handle))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return; } mpt3sas_transport_update_links(ioc, sas_expander->sas_address, le16_to_cpu(expander_pg1.AttachedDevHandle), i, expander_pg1.NegotiatedLinkRate >> 4, sas_expander->port); } } /** * _scsih_scan_for_devices_after_reset - scan for devices after host reset * @ioc: per adapter object */ static void _scsih_scan_for_devices_after_reset(struct MPT3SAS_ADAPTER *ioc) { Mpi2ExpanderPage0_t expander_pg0; Mpi2SasDevicePage0_t sas_device_pg0; Mpi26PCIeDevicePage0_t pcie_device_pg0; Mpi2RaidVolPage1_t *volume_pg1; Mpi2RaidVolPage0_t *volume_pg0; Mpi2RaidPhysDiskPage0_t pd_pg0; Mpi2EventIrConfigElement_t element; Mpi2ConfigReply_t mpi_reply; u8 phys_disk_num, port_id; u16 ioc_status; u16 handle, parent_handle; u64 sas_address; struct _sas_device *sas_device; struct _pcie_device *pcie_device; struct _sas_node *expander_device; static struct _raid_device *raid_device; u8 retry_count; unsigned long flags; volume_pg0 = kzalloc(sizeof(*volume_pg0), GFP_KERNEL); if (!volume_pg0) return; volume_pg1 = kzalloc(sizeof(*volume_pg1), GFP_KERNEL); if (!volume_pg1) { kfree(volume_pg0); return; } ioc_info(ioc, "scan devices: start\n"); _scsih_sas_host_refresh(ioc); ioc_info(ioc, "\tscan devices: expanders start\n"); /* expanders */ handle = 0xFFFF; while (!(mpt3sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0, MPI2_SAS_EXPAND_PGAD_FORM_GET_NEXT_HNDL, handle))) { ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_info(ioc, "\tbreak from expander scan: ioc_status(0x%04x), loginfo(0x%08x)\n", ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); break; } handle = le16_to_cpu(expander_pg0.DevHandle); spin_lock_irqsave(&ioc->sas_node_lock, flags); port_id = expander_pg0.PhysicalPort; expander_device = mpt3sas_scsih_expander_find_by_sas_address( ioc, le64_to_cpu(expander_pg0.SASAddress), mpt3sas_get_port_by_id(ioc, port_id, 0)); spin_unlock_irqrestore(&ioc->sas_node_lock, flags); if (expander_device) _scsih_refresh_expander_links(ioc, expander_device, handle); else { ioc_info(ioc, "\tBEFORE adding expander: handle (0x%04x), sas_addr(0x%016llx)\n", handle, (u64)le64_to_cpu(expander_pg0.SASAddress)); _scsih_expander_add(ioc, handle); ioc_info(ioc, "\tAFTER adding expander: handle (0x%04x), sas_addr(0x%016llx)\n", handle, (u64)le64_to_cpu(expander_pg0.SASAddress)); } } ioc_info(ioc, "\tscan devices: expanders complete\n"); if (!ioc->ir_firmware) goto skip_to_sas; ioc_info(ioc, "\tscan devices: phys disk start\n"); /* phys disk */ phys_disk_num = 0xFF; while (!(mpt3sas_config_get_phys_disk_pg0(ioc, &mpi_reply, &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_GET_NEXT_PHYSDISKNUM, phys_disk_num))) { ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_info(ioc, "\tbreak from phys disk scan: ioc_status(0x%04x), loginfo(0x%08x)\n", ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); break; } phys_disk_num = pd_pg0.PhysDiskNum; handle = le16_to_cpu(pd_pg0.DevHandle); sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); if (sas_device) { sas_device_put(sas_device); continue; } if (mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle) != 0) continue; ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_info(ioc, "\tbreak from phys disk scan ioc_status(0x%04x), loginfo(0x%08x)\n", ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); break; } parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle); if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address)) { ioc_info(ioc, "\tBEFORE adding phys disk: handle (0x%04x), sas_addr(0x%016llx)\n", handle, (u64)le64_to_cpu(sas_device_pg0.SASAddress)); port_id = sas_device_pg0.PhysicalPort; mpt3sas_transport_update_links(ioc, sas_address, handle, sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5, mpt3sas_get_port_by_id(ioc, port_id, 0)); set_bit(handle, ioc->pd_handles); retry_count = 0; /* This will retry adding the end device. * _scsih_add_device() will decide on retries and * return "1" when it should be retried */ while (_scsih_add_device(ioc, handle, retry_count++, 1)) { ssleep(1); } ioc_info(ioc, "\tAFTER adding phys disk: handle (0x%04x), sas_addr(0x%016llx)\n", handle, (u64)le64_to_cpu(sas_device_pg0.SASAddress)); } } ioc_info(ioc, "\tscan devices: phys disk complete\n"); ioc_info(ioc, "\tscan devices: volumes start\n"); /* volumes */ handle = 0xFFFF; while (!(mpt3sas_config_get_raid_volume_pg1(ioc, &mpi_reply, volume_pg1, MPI2_RAID_VOLUME_PGAD_FORM_GET_NEXT_HANDLE, handle))) { ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_info(ioc, "\tbreak from volume scan: ioc_status(0x%04x), loginfo(0x%08x)\n", ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); break; } handle = le16_to_cpu(volume_pg1->DevHandle); spin_lock_irqsave(&ioc->raid_device_lock, flags); raid_device = _scsih_raid_device_find_by_wwid(ioc, le64_to_cpu(volume_pg1->WWID)); spin_unlock_irqrestore(&ioc->raid_device_lock, flags); if (raid_device) continue; if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, volume_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle, sizeof(Mpi2RaidVolPage0_t))) continue; ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_info(ioc, "\tbreak from volume scan: ioc_status(0x%04x), loginfo(0x%08x)\n", ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); break; } if (volume_pg0->VolumeState == MPI2_RAID_VOL_STATE_OPTIMAL || volume_pg0->VolumeState == MPI2_RAID_VOL_STATE_ONLINE || volume_pg0->VolumeState == MPI2_RAID_VOL_STATE_DEGRADED) { memset(&element, 0, sizeof(Mpi2EventIrConfigElement_t)); element.ReasonCode = MPI2_EVENT_IR_CHANGE_RC_ADDED; element.VolDevHandle = volume_pg1->DevHandle; ioc_info(ioc, "\tBEFORE adding volume: handle (0x%04x)\n", volume_pg1->DevHandle); _scsih_sas_volume_add(ioc, &element); ioc_info(ioc, "\tAFTER adding volume: handle (0x%04x)\n", volume_pg1->DevHandle); } } ioc_info(ioc, "\tscan devices: volumes complete\n"); skip_to_sas: ioc_info(ioc, "\tscan devices: end devices start\n"); /* sas devices */ handle = 0xFFFF; while (!(mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_GET_NEXT_HANDLE, handle))) { ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_info(ioc, "\tbreak from end device scan: ioc_status(0x%04x), loginfo(0x%08x)\n", ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); break; } handle = le16_to_cpu(sas_device_pg0.DevHandle); if (!(_scsih_is_end_device( le32_to_cpu(sas_device_pg0.DeviceInfo)))) continue; port_id = sas_device_pg0.PhysicalPort; sas_device = mpt3sas_get_sdev_by_addr(ioc, le64_to_cpu(sas_device_pg0.SASAddress), mpt3sas_get_port_by_id(ioc, port_id, 0)); if (sas_device) { sas_device_put(sas_device); continue; } parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle); if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address)) { ioc_info(ioc, "\tBEFORE adding end device: handle (0x%04x), sas_addr(0x%016llx)\n", handle, (u64)le64_to_cpu(sas_device_pg0.SASAddress)); mpt3sas_transport_update_links(ioc, sas_address, handle, sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5, mpt3sas_get_port_by_id(ioc, port_id, 0)); retry_count = 0; /* This will retry adding the end device. * _scsih_add_device() will decide on retries and * return "1" when it should be retried */ while (_scsih_add_device(ioc, handle, retry_count++, 0)) { ssleep(1); } ioc_info(ioc, "\tAFTER adding end device: handle (0x%04x), sas_addr(0x%016llx)\n", handle, (u64)le64_to_cpu(sas_device_pg0.SASAddress)); } } ioc_info(ioc, "\tscan devices: end devices complete\n"); ioc_info(ioc, "\tscan devices: pcie end devices start\n"); /* pcie devices */ handle = 0xFFFF; while (!(mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply, &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_GET_NEXT_HANDLE, handle))) { ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { ioc_info(ioc, "\tbreak from pcie end device scan: ioc_status(0x%04x), loginfo(0x%08x)\n", ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); break; } handle = le16_to_cpu(pcie_device_pg0.DevHandle); if (!(_scsih_is_nvme_pciescsi_device( le32_to_cpu(pcie_device_pg0.DeviceInfo)))) continue; pcie_device = mpt3sas_get_pdev_by_wwid(ioc, le64_to_cpu(pcie_device_pg0.WWID)); if (pcie_device) { pcie_device_put(pcie_device); continue; } retry_count = 0; parent_handle = le16_to_cpu(pcie_device_pg0.ParentDevHandle); _scsih_pcie_add_device(ioc, handle); ioc_info(ioc, "\tAFTER adding pcie end device: handle (0x%04x), wwid(0x%016llx)\n", handle, (u64)le64_to_cpu(pcie_device_pg0.WWID)); } kfree(volume_pg0); kfree(volume_pg1); ioc_info(ioc, "\tpcie devices: pcie end devices complete\n"); ioc_info(ioc, "scan devices: complete\n"); } /** * mpt3sas_scsih_pre_reset_handler - reset callback handler (for scsih) * @ioc: per adapter object * * The handler for doing any required cleanup or initialization. */ void mpt3sas_scsih_pre_reset_handler(struct MPT3SAS_ADAPTER *ioc) { dtmprintk(ioc, ioc_info(ioc, "%s: MPT3_IOC_PRE_RESET\n", __func__)); } /** * mpt3sas_scsih_clear_outstanding_scsi_tm_commands - clears outstanding * scsi & tm cmds. * @ioc: per adapter object * * The handler for doing any required cleanup or initialization. */ void mpt3sas_scsih_clear_outstanding_scsi_tm_commands(struct MPT3SAS_ADAPTER *ioc) { dtmprintk(ioc, ioc_info(ioc, "%s: clear outstanding scsi & tm cmds\n", __func__)); if (ioc->scsih_cmds.status & MPT3_CMD_PENDING) { ioc->scsih_cmds.status |= MPT3_CMD_RESET; mpt3sas_base_free_smid(ioc, ioc->scsih_cmds.smid); complete(&ioc->scsih_cmds.done); } if (ioc->tm_cmds.status & MPT3_CMD_PENDING) { ioc->tm_cmds.status |= MPT3_CMD_RESET; mpt3sas_base_free_smid(ioc, ioc->tm_cmds.smid); complete(&ioc->tm_cmds.done); } memset(ioc->pend_os_device_add, 0, ioc->pend_os_device_add_sz); memset(ioc->device_remove_in_progress, 0, ioc->device_remove_in_progress_sz); _scsih_fw_event_cleanup_queue(ioc); _scsih_flush_running_cmds(ioc); } /** * mpt3sas_scsih_reset_done_handler - reset callback handler (for scsih) * @ioc: per adapter object * * The handler for doing any required cleanup or initialization. */ void mpt3sas_scsih_reset_done_handler(struct MPT3SAS_ADAPTER *ioc) { dtmprintk(ioc, ioc_info(ioc, "%s: MPT3_IOC_DONE_RESET\n", __func__)); if (!(disable_discovery > 0 && !ioc->sas_hba.num_phys)) { if (ioc->multipath_on_hba) { _scsih_sas_port_refresh(ioc); _scsih_update_vphys_after_reset(ioc); } _scsih_prep_device_scan(ioc); _scsih_create_enclosure_list_after_reset(ioc); _scsih_search_responding_sas_devices(ioc); _scsih_search_responding_pcie_devices(ioc); _scsih_search_responding_raid_devices(ioc); _scsih_search_responding_expanders(ioc); _scsih_error_recovery_delete_devices(ioc); } } /** * _mpt3sas_fw_work - delayed task for processing firmware events * @ioc: per adapter object * @fw_event: The fw_event_work object * Context: user. */ static void _mpt3sas_fw_work(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) { ioc->current_event = fw_event; _scsih_fw_event_del_from_list(ioc, fw_event); /* the queue is being flushed so ignore this event */ if (ioc->remove_host || ioc->pci_error_recovery) { fw_event_work_put(fw_event); ioc->current_event = NULL; return; } switch (fw_event->event) { case MPT3SAS_PROCESS_TRIGGER_DIAG: mpt3sas_process_trigger_data(ioc, (struct SL_WH_TRIGGERS_EVENT_DATA_T *) fw_event->event_data); break; case MPT3SAS_REMOVE_UNRESPONDING_DEVICES: while (scsi_host_in_recovery(ioc->shost) || ioc->shost_recovery) { /* * If we're unloading or cancelling the work, bail. * Otherwise, this can become an infinite loop. */ if (ioc->remove_host || ioc->fw_events_cleanup) goto out; ssleep(1); } _scsih_remove_unresponding_devices(ioc); _scsih_del_dirty_vphy(ioc); _scsih_del_dirty_port_entries(ioc); _scsih_scan_for_devices_after_reset(ioc); /* * If diag reset has occurred during the driver load * then driver has to complete the driver load operation * by executing the following items: *- Register the devices from sas_device_init_list to SML *- clear is_driver_loading flag, *- start the watchdog thread. * In happy driver load path, above things are taken care of when * driver executes scsih_scan_finished(). */ if (ioc->is_driver_loading) _scsih_complete_devices_scanning(ioc); _scsih_set_nvme_max_shutdown_latency(ioc); break; case MPT3SAS_PORT_ENABLE_COMPLETE: ioc->start_scan = 0; if (missing_delay[0] != -1 && missing_delay[1] != -1) mpt3sas_base_update_missing_delay(ioc, missing_delay[0], missing_delay[1]); dewtprintk(ioc, ioc_info(ioc, "port enable: complete from worker thread\n")); break; case MPT3SAS_TURN_ON_PFA_LED: _scsih_turn_on_pfa_led(ioc, fw_event->device_handle); break; case MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST: _scsih_sas_topology_change_event(ioc, fw_event); break; case MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE: if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) _scsih_sas_device_status_change_event_debug(ioc, (Mpi2EventDataSasDeviceStatusChange_t *) fw_event->event_data); break; case MPI2_EVENT_SAS_DISCOVERY: _scsih_sas_discovery_event(ioc, fw_event); break; case MPI2_EVENT_SAS_DEVICE_DISCOVERY_ERROR: _scsih_sas_device_discovery_error_event(ioc, fw_event); break; case MPI2_EVENT_SAS_BROADCAST_PRIMITIVE: _scsih_sas_broadcast_primitive_event(ioc, fw_event); break; case MPI2_EVENT_SAS_ENCL_DEVICE_STATUS_CHANGE: _scsih_sas_enclosure_dev_status_change_event(ioc, fw_event); break; case MPI2_EVENT_IR_CONFIGURATION_CHANGE_LIST: _scsih_sas_ir_config_change_event(ioc, fw_event); break; case MPI2_EVENT_IR_VOLUME: _scsih_sas_ir_volume_event(ioc, fw_event); break; case MPI2_EVENT_IR_PHYSICAL_DISK: _scsih_sas_ir_physical_disk_event(ioc, fw_event); break; case MPI2_EVENT_IR_OPERATION_STATUS: _scsih_sas_ir_operation_status_event(ioc, fw_event); break; case MPI2_EVENT_PCIE_DEVICE_STATUS_CHANGE: _scsih_pcie_device_status_change_event(ioc, fw_event); break; case MPI2_EVENT_PCIE_ENUMERATION: _scsih_pcie_enumeration_event(ioc, fw_event); break; case MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST: _scsih_pcie_topology_change_event(ioc, fw_event); ioc->current_event = NULL; return; break; } out: fw_event_work_put(fw_event); ioc->current_event = NULL; } /** * _firmware_event_work * @work: The fw_event_work object * Context: user. * * wrappers for the work thread handling firmware events */ static void _firmware_event_work(struct work_struct *work) { struct fw_event_work *fw_event = container_of(work, struct fw_event_work, work); _mpt3sas_fw_work(fw_event->ioc, fw_event); } /** * mpt3sas_scsih_event_callback - firmware event handler (called at ISR time) * @ioc: per adapter object * @msix_index: MSIX table index supplied by the OS * @reply: reply message frame(lower 32bit addr) * Context: interrupt. * * This function merely adds a new work task into ioc->firmware_event_thread. * The tasks are worked from _firmware_event_work in user context. * * Return: 1 meaning mf should be freed from _base_interrupt * 0 means the mf is freed from this function. */ u8 mpt3sas_scsih_event_callback(struct MPT3SAS_ADAPTER *ioc, u8 msix_index, u32 reply) { struct fw_event_work *fw_event; Mpi2EventNotificationReply_t *mpi_reply; u16 event; u16 sz; Mpi26EventDataActiveCableExcept_t *ActiveCableEventData; /* events turned off due to host reset */ if (ioc->pci_error_recovery) return 1; mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply); if (unlikely(!mpi_reply)) { ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return 1; } event = le16_to_cpu(mpi_reply->Event); if (event != MPI2_EVENT_LOG_ENTRY_ADDED) mpt3sas_trigger_event(ioc, event, 0); switch (event) { /* handle these */ case MPI2_EVENT_SAS_BROADCAST_PRIMITIVE: { Mpi2EventDataSasBroadcastPrimitive_t *baen_data = (Mpi2EventDataSasBroadcastPrimitive_t *) mpi_reply->EventData; if (baen_data->Primitive != MPI2_EVENT_PRIMITIVE_ASYNCHRONOUS_EVENT) return 1; if (ioc->broadcast_aen_busy) { ioc->broadcast_aen_pending++; return 1; } else ioc->broadcast_aen_busy = 1; break; } case MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST: _scsih_check_topo_delete_events(ioc, (Mpi2EventDataSasTopologyChangeList_t *) mpi_reply->EventData); /* * No need to add the topology change list * event to fw event work queue when * diag reset is going on. Since during diag * reset driver scan the devices by reading * sas device page0's not by processing the * events. */ if (ioc->shost_recovery) return 1; break; case MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST: _scsih_check_pcie_topo_remove_events(ioc, (Mpi26EventDataPCIeTopologyChangeList_t *) mpi_reply->EventData); if (ioc->shost_recovery) return 1; break; case MPI2_EVENT_IR_CONFIGURATION_CHANGE_LIST: _scsih_check_ir_config_unhide_events(ioc, (Mpi2EventDataIrConfigChangeList_t *) mpi_reply->EventData); break; case MPI2_EVENT_IR_VOLUME: _scsih_check_volume_delete_events(ioc, (Mpi2EventDataIrVolume_t *) mpi_reply->EventData); break; case MPI2_EVENT_LOG_ENTRY_ADDED: { Mpi2EventDataLogEntryAdded_t *log_entry; u32 *log_code; if (!ioc->is_warpdrive) break; log_entry = (Mpi2EventDataLogEntryAdded_t *) mpi_reply->EventData; log_code = (u32 *)log_entry->LogData; if (le16_to_cpu(log_entry->LogEntryQualifier) != MPT2_WARPDRIVE_LOGENTRY) break; switch (le32_to_cpu(*log_code)) { case MPT2_WARPDRIVE_LC_SSDT: ioc_warn(ioc, "WarpDrive Warning: IO Throttling has occurred in the WarpDrive subsystem. Check WarpDrive documentation for additional details.\n"); break; case MPT2_WARPDRIVE_LC_SSDLW: ioc_warn(ioc, "WarpDrive Warning: Program/Erase Cycles for the WarpDrive subsystem in degraded range. Check WarpDrive documentation for additional details.\n"); break; case MPT2_WARPDRIVE_LC_SSDLF: ioc_err(ioc, "WarpDrive Fatal Error: There are no Program/Erase Cycles for the WarpDrive subsystem. The storage device will be in read-only mode. Check WarpDrive documentation for additional details.\n"); break; case MPT2_WARPDRIVE_LC_BRMF: ioc_err(ioc, "WarpDrive Fatal Error: The Backup Rail Monitor has failed on the WarpDrive subsystem. Check WarpDrive documentation for additional details.\n"); break; } break; } case MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE: _scsih_sas_device_status_change_event(ioc, (Mpi2EventDataSasDeviceStatusChange_t *) mpi_reply->EventData); break; case MPI2_EVENT_IR_OPERATION_STATUS: case MPI2_EVENT_SAS_DISCOVERY: case MPI2_EVENT_SAS_DEVICE_DISCOVERY_ERROR: case MPI2_EVENT_SAS_ENCL_DEVICE_STATUS_CHANGE: case MPI2_EVENT_IR_PHYSICAL_DISK: case MPI2_EVENT_PCIE_ENUMERATION: case MPI2_EVENT_PCIE_DEVICE_STATUS_CHANGE: break; case MPI2_EVENT_TEMP_THRESHOLD: _scsih_temp_threshold_events(ioc, (Mpi2EventDataTemperature_t *) mpi_reply->EventData); break; case MPI2_EVENT_ACTIVE_CABLE_EXCEPTION: ActiveCableEventData = (Mpi26EventDataActiveCableExcept_t *) mpi_reply->EventData; switch (ActiveCableEventData->ReasonCode) { case MPI26_EVENT_ACTIVE_CABLE_INSUFFICIENT_POWER: ioc_notice(ioc, "Currently an active cable with ReceptacleID %d\n", ActiveCableEventData->ReceptacleID); pr_notice("cannot be powered and devices connected\n"); pr_notice("to this active cable will not be seen\n"); pr_notice("This active cable requires %d mW of power\n", le32_to_cpu( ActiveCableEventData->ActiveCablePowerRequirement)); break; case MPI26_EVENT_ACTIVE_CABLE_DEGRADED: ioc_notice(ioc, "Currently a cable with ReceptacleID %d\n", ActiveCableEventData->ReceptacleID); pr_notice( "is not running at optimal speed(12 Gb/s rate)\n"); break; } break; default: /* ignore the rest */ return 1; } sz = le16_to_cpu(mpi_reply->EventDataLength) * 4; fw_event = alloc_fw_event_work(sz); if (!fw_event) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); return 1; } memcpy(fw_event->event_data, mpi_reply->EventData, sz); fw_event->ioc = ioc; fw_event->VF_ID = mpi_reply->VF_ID; fw_event->VP_ID = mpi_reply->VP_ID; fw_event->event = event; _scsih_fw_event_add(ioc, fw_event); fw_event_work_put(fw_event); return 1; } /** * _scsih_expander_node_remove - removing expander device from list. * @ioc: per adapter object * @sas_expander: the sas_device object * * Removing object and freeing associated memory from the * ioc->sas_expander_list. */ static void _scsih_expander_node_remove(struct MPT3SAS_ADAPTER *ioc, struct _sas_node *sas_expander) { struct _sas_port *mpt3sas_port, *next; unsigned long flags; /* remove sibling ports attached to this expander */ list_for_each_entry_safe(mpt3sas_port, next, &sas_expander->sas_port_list, port_list) { if (ioc->shost_recovery) return; if (mpt3sas_port->remote_identify.device_type == SAS_END_DEVICE) mpt3sas_device_remove_by_sas_address(ioc, mpt3sas_port->remote_identify.sas_address, mpt3sas_port->hba_port); else if (mpt3sas_port->remote_identify.device_type == SAS_EDGE_EXPANDER_DEVICE || mpt3sas_port->remote_identify.device_type == SAS_FANOUT_EXPANDER_DEVICE) mpt3sas_expander_remove(ioc, mpt3sas_port->remote_identify.sas_address, mpt3sas_port->hba_port); } mpt3sas_transport_port_remove(ioc, sas_expander->sas_address, sas_expander->sas_address_parent, sas_expander->port); ioc_info(ioc, "expander_remove: handle(0x%04x), sas_addr(0x%016llx), port:%d\n", sas_expander->handle, (unsigned long long) sas_expander->sas_address, sas_expander->port->port_id); spin_lock_irqsave(&ioc->sas_node_lock, flags); list_del(&sas_expander->list); spin_unlock_irqrestore(&ioc->sas_node_lock, flags); kfree(sas_expander->phy); kfree(sas_expander); } /** * _scsih_nvme_shutdown - NVMe shutdown notification * @ioc: per adapter object * * Sending IoUnitControl request with shutdown operation code to alert IOC that * the host system is shutting down so that IOC can issue NVMe shutdown to * NVMe drives attached to it. */ static void _scsih_nvme_shutdown(struct MPT3SAS_ADAPTER *ioc) { Mpi26IoUnitControlRequest_t *mpi_request; Mpi26IoUnitControlReply_t *mpi_reply; u16 smid; /* are there any NVMe devices ? */ if (list_empty(&ioc->pcie_device_list)) return; mutex_lock(&ioc->scsih_cmds.mutex); if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) { ioc_err(ioc, "%s: scsih_cmd in use\n", __func__); goto out; } ioc->scsih_cmds.status = MPT3_CMD_PENDING; smid = mpt3sas_base_get_smid(ioc, ioc->scsih_cb_idx); if (!smid) { ioc_err(ioc, "%s: failed obtaining a smid\n", __func__); ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; goto out; } mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); ioc->scsih_cmds.smid = smid; memset(mpi_request, 0, sizeof(Mpi26IoUnitControlRequest_t)); mpi_request->Function = MPI2_FUNCTION_IO_UNIT_CONTROL; mpi_request->Operation = MPI26_CTRL_OP_SHUTDOWN; init_completion(&ioc->scsih_cmds.done); ioc->put_smid_default(ioc, smid); /* Wait for max_shutdown_latency seconds */ ioc_info(ioc, "Io Unit Control shutdown (sending), Shutdown latency %d sec\n", ioc->max_shutdown_latency); wait_for_completion_timeout(&ioc->scsih_cmds.done, ioc->max_shutdown_latency*HZ); if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) { ioc_err(ioc, "%s: timeout\n", __func__); goto out; } if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) { mpi_reply = ioc->scsih_cmds.reply; ioc_info(ioc, "Io Unit Control shutdown (complete):" "ioc_status(0x%04x), loginfo(0x%08x)\n", le16_to_cpu(mpi_reply->IOCStatus), le32_to_cpu(mpi_reply->IOCLogInfo)); } out: ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; mutex_unlock(&ioc->scsih_cmds.mutex); } /** * _scsih_ir_shutdown - IR shutdown notification * @ioc: per adapter object * * Sending RAID Action to alert the Integrated RAID subsystem of the IOC that * the host system is shutting down. */ static void _scsih_ir_shutdown(struct MPT3SAS_ADAPTER *ioc) { Mpi2RaidActionRequest_t *mpi_request; Mpi2RaidActionReply_t *mpi_reply; u16 smid; /* is IR firmware build loaded ? */ if (!ioc->ir_firmware) return; /* are there any volumes ? */ if (list_empty(&ioc->raid_device_list)) return; mutex_lock(&ioc->scsih_cmds.mutex); if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) { ioc_err(ioc, "%s: scsih_cmd in use\n", __func__); goto out; } ioc->scsih_cmds.status = MPT3_CMD_PENDING; smid = mpt3sas_base_get_smid(ioc, ioc->scsih_cb_idx); if (!smid) { ioc_err(ioc, "%s: failed obtaining a smid\n", __func__); ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; goto out; } mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); ioc->scsih_cmds.smid = smid; memset(mpi_request, 0, sizeof(Mpi2RaidActionRequest_t)); mpi_request->Function = MPI2_FUNCTION_RAID_ACTION; mpi_request->Action = MPI2_RAID_ACTION_SYSTEM_SHUTDOWN_INITIATED; if (!ioc->hide_ir_msg) ioc_info(ioc, "IR shutdown (sending)\n"); init_completion(&ioc->scsih_cmds.done); ioc->put_smid_default(ioc, smid); wait_for_completion_timeout(&ioc->scsih_cmds.done, 10*HZ); if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) { ioc_err(ioc, "%s: timeout\n", __func__); goto out; } if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) { mpi_reply = ioc->scsih_cmds.reply; if (!ioc->hide_ir_msg) ioc_info(ioc, "IR shutdown (complete): ioc_status(0x%04x), loginfo(0x%08x)\n", le16_to_cpu(mpi_reply->IOCStatus), le32_to_cpu(mpi_reply->IOCLogInfo)); } out: ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; mutex_unlock(&ioc->scsih_cmds.mutex); } /** * _scsih_get_shost_and_ioc - get shost and ioc * and verify whether they are NULL or not * @pdev: PCI device struct * @shost: address of scsi host pointer * @ioc: address of HBA adapter pointer * * Return zero if *shost and *ioc are not NULL otherwise return error number. */ static int _scsih_get_shost_and_ioc(struct pci_dev *pdev, struct Scsi_Host **shost, struct MPT3SAS_ADAPTER **ioc) { *shost = pci_get_drvdata(pdev); if (*shost == NULL) { dev_err(&pdev->dev, "pdev's driver data is null\n"); return -ENXIO; } *ioc = shost_priv(*shost); if (*ioc == NULL) { dev_err(&pdev->dev, "shost's private data is null\n"); return -ENXIO; } return 0; } /** * scsih_remove - detach and remove add host * @pdev: PCI device struct * * Routine called when unloading the driver. */ static void scsih_remove(struct pci_dev *pdev) { struct Scsi_Host *shost; struct MPT3SAS_ADAPTER *ioc; struct _sas_port *mpt3sas_port, *next_port; struct _raid_device *raid_device, *next; struct MPT3SAS_TARGET *sas_target_priv_data; struct _pcie_device *pcie_device, *pcienext; struct workqueue_struct *wq; unsigned long flags; Mpi2ConfigReply_t mpi_reply; struct hba_port *port, *port_next; if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc)) return; ioc->remove_host = 1; if (!pci_device_is_present(pdev)) _scsih_flush_running_cmds(ioc); _scsih_fw_event_cleanup_queue(ioc); spin_lock_irqsave(&ioc->fw_event_lock, flags); wq = ioc->firmware_event_thread; ioc->firmware_event_thread = NULL; spin_unlock_irqrestore(&ioc->fw_event_lock, flags); if (wq) destroy_workqueue(wq); /* * Copy back the unmodified ioc page1. so that on next driver load, * current modified changes on ioc page1 won't take effect. */ if (ioc->is_aero_ioc) mpt3sas_config_set_ioc_pg1(ioc, &mpi_reply, &ioc->ioc_pg1_copy); /* release all the volumes */ _scsih_ir_shutdown(ioc); mpt3sas_destroy_debugfs(ioc); sas_remove_host(shost); list_for_each_entry_safe(raid_device, next, &ioc->raid_device_list, list) { if (raid_device->starget) { sas_target_priv_data = raid_device->starget->hostdata; sas_target_priv_data->deleted = 1; scsi_remove_target(&raid_device->starget->dev); } ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n", raid_device->handle, (u64)raid_device->wwid); _scsih_raid_device_remove(ioc, raid_device); } list_for_each_entry_safe(pcie_device, pcienext, &ioc->pcie_device_list, list) { _scsih_pcie_device_remove_from_sml(ioc, pcie_device); list_del_init(&pcie_device->list); pcie_device_put(pcie_device); } /* free ports attached to the sas_host */ list_for_each_entry_safe(mpt3sas_port, next_port, &ioc->sas_hba.sas_port_list, port_list) { if (mpt3sas_port->remote_identify.device_type == SAS_END_DEVICE) mpt3sas_device_remove_by_sas_address(ioc, mpt3sas_port->remote_identify.sas_address, mpt3sas_port->hba_port); else if (mpt3sas_port->remote_identify.device_type == SAS_EDGE_EXPANDER_DEVICE || mpt3sas_port->remote_identify.device_type == SAS_FANOUT_EXPANDER_DEVICE) mpt3sas_expander_remove(ioc, mpt3sas_port->remote_identify.sas_address, mpt3sas_port->hba_port); } list_for_each_entry_safe(port, port_next, &ioc->port_table_list, list) { list_del(&port->list); kfree(port); } /* free phys attached to the sas_host */ if (ioc->sas_hba.num_phys) { kfree(ioc->sas_hba.phy); ioc->sas_hba.phy = NULL; ioc->sas_hba.num_phys = 0; } mpt3sas_base_detach(ioc); spin_lock(&gioc_lock); list_del(&ioc->list); spin_unlock(&gioc_lock); scsi_host_put(shost); } /** * scsih_shutdown - routine call during system shutdown * @pdev: PCI device struct */ static void scsih_shutdown(struct pci_dev *pdev) { struct Scsi_Host *shost; struct MPT3SAS_ADAPTER *ioc; struct workqueue_struct *wq; unsigned long flags; Mpi2ConfigReply_t mpi_reply; if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc)) return; ioc->remove_host = 1; if (!pci_device_is_present(pdev)) _scsih_flush_running_cmds(ioc); _scsih_fw_event_cleanup_queue(ioc); spin_lock_irqsave(&ioc->fw_event_lock, flags); wq = ioc->firmware_event_thread; ioc->firmware_event_thread = NULL; spin_unlock_irqrestore(&ioc->fw_event_lock, flags); if (wq) destroy_workqueue(wq); /* * Copy back the unmodified ioc page1 so that on next driver load, * current modified changes on ioc page1 won't take effect. */ if (ioc->is_aero_ioc) mpt3sas_config_set_ioc_pg1(ioc, &mpi_reply, &ioc->ioc_pg1_copy); _scsih_ir_shutdown(ioc); _scsih_nvme_shutdown(ioc); mpt3sas_base_detach(ioc); } /** * _scsih_probe_boot_devices - reports 1st device * @ioc: per adapter object * * If specified in bios page 2, this routine reports the 1st * device scsi-ml or sas transport for persistent boot device * purposes. Please refer to function _scsih_determine_boot_device() */ static void _scsih_probe_boot_devices(struct MPT3SAS_ADAPTER *ioc) { u32 channel; void *device; struct _sas_device *sas_device; struct _raid_device *raid_device; struct _pcie_device *pcie_device; u16 handle; u64 sas_address_parent; u64 sas_address; unsigned long flags; int rc; int tid; struct hba_port *port; /* no Bios, return immediately */ if (!ioc->bios_pg3.BiosVersion) return; device = NULL; if (ioc->req_boot_device.device) { device = ioc->req_boot_device.device; channel = ioc->req_boot_device.channel; } else if (ioc->req_alt_boot_device.device) { device = ioc->req_alt_boot_device.device; channel = ioc->req_alt_boot_device.channel; } else if (ioc->current_boot_device.device) { device = ioc->current_boot_device.device; channel = ioc->current_boot_device.channel; } if (!device) return; if (channel == RAID_CHANNEL) { raid_device = device; /* * If this boot vd is already registered with SML then * no need to register it again as part of device scanning * after diag reset during driver load operation. */ if (raid_device->starget) return; rc = scsi_add_device(ioc->shost, RAID_CHANNEL, raid_device->id, 0); if (rc) _scsih_raid_device_remove(ioc, raid_device); } else if (channel == PCIE_CHANNEL) { pcie_device = device; /* * If this boot NVMe device is already registered with SML then * no need to register it again as part of device scanning * after diag reset during driver load operation. */ if (pcie_device->starget) return; spin_lock_irqsave(&ioc->pcie_device_lock, flags); tid = pcie_device->id; list_move_tail(&pcie_device->list, &ioc->pcie_device_list); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); rc = scsi_add_device(ioc->shost, PCIE_CHANNEL, tid, 0); if (rc) _scsih_pcie_device_remove(ioc, pcie_device); } else { sas_device = device; /* * If this boot sas/sata device is already registered with SML * then no need to register it again as part of device scanning * after diag reset during driver load operation. */ if (sas_device->starget) return; spin_lock_irqsave(&ioc->sas_device_lock, flags); handle = sas_device->handle; sas_address_parent = sas_device->sas_address_parent; sas_address = sas_device->sas_address; port = sas_device->port; list_move_tail(&sas_device->list, &ioc->sas_device_list); spin_unlock_irqrestore(&ioc->sas_device_lock, flags); if (ioc->hide_drives) return; if (!port) return; if (!mpt3sas_transport_port_add(ioc, handle, sas_address_parent, port)) { _scsih_sas_device_remove(ioc, sas_device); } else if (!sas_device->starget) { if (!ioc->is_driver_loading) { mpt3sas_transport_port_remove(ioc, sas_address, sas_address_parent, port); _scsih_sas_device_remove(ioc, sas_device); } } } } /** * _scsih_probe_raid - reporting raid volumes to scsi-ml * @ioc: per adapter object * * Called during initial loading of the driver. */ static void _scsih_probe_raid(struct MPT3SAS_ADAPTER *ioc) { struct _raid_device *raid_device, *raid_next; int rc; list_for_each_entry_safe(raid_device, raid_next, &ioc->raid_device_list, list) { if (raid_device->starget) continue; rc = scsi_add_device(ioc->shost, RAID_CHANNEL, raid_device->id, 0); if (rc) _scsih_raid_device_remove(ioc, raid_device); } } static struct _sas_device *get_next_sas_device(struct MPT3SAS_ADAPTER *ioc) { struct _sas_device *sas_device = NULL; unsigned long flags; spin_lock_irqsave(&ioc->sas_device_lock, flags); if (!list_empty(&ioc->sas_device_init_list)) { sas_device = list_first_entry(&ioc->sas_device_init_list, struct _sas_device, list); sas_device_get(sas_device); } spin_unlock_irqrestore(&ioc->sas_device_lock, flags); return sas_device; } static void sas_device_make_active(struct MPT3SAS_ADAPTER *ioc, struct _sas_device *sas_device) { unsigned long flags; spin_lock_irqsave(&ioc->sas_device_lock, flags); /* * Since we dropped the lock during the call to port_add(), we need to * be careful here that somebody else didn't move or delete this item * while we were busy with other things. * * If it was on the list, we need a put() for the reference the list * had. Either way, we need a get() for the destination list. */ if (!list_empty(&sas_device->list)) { list_del_init(&sas_device->list); sas_device_put(sas_device); } sas_device_get(sas_device); list_add_tail(&sas_device->list, &ioc->sas_device_list); spin_unlock_irqrestore(&ioc->sas_device_lock, flags); } /** * _scsih_probe_sas - reporting sas devices to sas transport * @ioc: per adapter object * * Called during initial loading of the driver. */ static void _scsih_probe_sas(struct MPT3SAS_ADAPTER *ioc) { struct _sas_device *sas_device; if (ioc->hide_drives) return; while ((sas_device = get_next_sas_device(ioc))) { if (!mpt3sas_transport_port_add(ioc, sas_device->handle, sas_device->sas_address_parent, sas_device->port)) { _scsih_sas_device_remove(ioc, sas_device); sas_device_put(sas_device); continue; } else if (!sas_device->starget) { /* * When asyn scanning is enabled, its not possible to * remove devices while scanning is turned on due to an * oops in scsi_sysfs_add_sdev()->add_device()-> * sysfs_addrm_start() */ if (!ioc->is_driver_loading) { mpt3sas_transport_port_remove(ioc, sas_device->sas_address, sas_device->sas_address_parent, sas_device->port); _scsih_sas_device_remove(ioc, sas_device); sas_device_put(sas_device); continue; } } sas_device_make_active(ioc, sas_device); sas_device_put(sas_device); } } /** * get_next_pcie_device - Get the next pcie device * @ioc: per adapter object * * Get the next pcie device from pcie_device_init_list list. * * Return: pcie device structure if pcie_device_init_list list is not empty * otherwise returns NULL */ static struct _pcie_device *get_next_pcie_device(struct MPT3SAS_ADAPTER *ioc) { struct _pcie_device *pcie_device = NULL; unsigned long flags; spin_lock_irqsave(&ioc->pcie_device_lock, flags); if (!list_empty(&ioc->pcie_device_init_list)) { pcie_device = list_first_entry(&ioc->pcie_device_init_list, struct _pcie_device, list); pcie_device_get(pcie_device); } spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); return pcie_device; } /** * pcie_device_make_active - Add pcie device to pcie_device_list list * @ioc: per adapter object * @pcie_device: pcie device object * * Add the pcie device which has registered with SCSI Transport Later to * pcie_device_list list */ static void pcie_device_make_active(struct MPT3SAS_ADAPTER *ioc, struct _pcie_device *pcie_device) { unsigned long flags; spin_lock_irqsave(&ioc->pcie_device_lock, flags); if (!list_empty(&pcie_device->list)) { list_del_init(&pcie_device->list); pcie_device_put(pcie_device); } pcie_device_get(pcie_device); list_add_tail(&pcie_device->list, &ioc->pcie_device_list); spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); } /** * _scsih_probe_pcie - reporting PCIe devices to scsi-ml * @ioc: per adapter object * * Called during initial loading of the driver. */ static void _scsih_probe_pcie(struct MPT3SAS_ADAPTER *ioc) { struct _pcie_device *pcie_device; int rc; /* PCIe Device List */ while ((pcie_device = get_next_pcie_device(ioc))) { if (pcie_device->starget) { pcie_device_put(pcie_device); continue; } if (pcie_device->access_status == MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) { pcie_device_make_active(ioc, pcie_device); pcie_device_put(pcie_device); continue; } rc = scsi_add_device(ioc->shost, PCIE_CHANNEL, pcie_device->id, 0); if (rc) { _scsih_pcie_device_remove(ioc, pcie_device); pcie_device_put(pcie_device); continue; } else if (!pcie_device->starget) { /* * When async scanning is enabled, its not possible to * remove devices while scanning is turned on due to an * oops in scsi_sysfs_add_sdev()->add_device()-> * sysfs_addrm_start() */ if (!ioc->is_driver_loading) { /* TODO-- Need to find out whether this condition will * occur or not */ _scsih_pcie_device_remove(ioc, pcie_device); pcie_device_put(pcie_device); continue; } } pcie_device_make_active(ioc, pcie_device); pcie_device_put(pcie_device); } } /** * _scsih_probe_devices - probing for devices * @ioc: per adapter object * * Called during initial loading of the driver. */ static void _scsih_probe_devices(struct MPT3SAS_ADAPTER *ioc) { u16 volume_mapping_flags; if (!(ioc->facts.ProtocolFlags & MPI2_IOCFACTS_PROTOCOL_SCSI_INITIATOR)) return; /* return when IOC doesn't support initiator mode */ _scsih_probe_boot_devices(ioc); if (ioc->ir_firmware) { volume_mapping_flags = le16_to_cpu(ioc->ioc_pg8.IRVolumeMappingFlags) & MPI2_IOCPAGE8_IRFLAGS_MASK_VOLUME_MAPPING_MODE; if (volume_mapping_flags == MPI2_IOCPAGE8_IRFLAGS_LOW_VOLUME_MAPPING) { _scsih_probe_raid(ioc); _scsih_probe_sas(ioc); } else { _scsih_probe_sas(ioc); _scsih_probe_raid(ioc); } } else { _scsih_probe_sas(ioc); _scsih_probe_pcie(ioc); } } /** * scsih_scan_start - scsi lld callback for .scan_start * @shost: SCSI host pointer * * The shost has the ability to discover targets on its own instead * of scanning the entire bus. In our implemention, we will kick off * firmware discovery. */ static void scsih_scan_start(struct Scsi_Host *shost) { struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); int rc; if (diag_buffer_enable != -1 && diag_buffer_enable != 0) mpt3sas_enable_diag_buffer(ioc, diag_buffer_enable); else if (ioc->manu_pg11.HostTraceBufferMaxSizeKB != 0) mpt3sas_enable_diag_buffer(ioc, 1); if (disable_discovery > 0) return; ioc->start_scan = 1; rc = mpt3sas_port_enable(ioc); if (rc != 0) ioc_info(ioc, "port enable: FAILED\n"); } /** * _scsih_complete_devices_scanning - add the devices to sml and * complete ioc initialization. * @ioc: per adapter object * * Return nothing. */ static void _scsih_complete_devices_scanning(struct MPT3SAS_ADAPTER *ioc) { if (ioc->wait_for_discovery_to_complete) { ioc->wait_for_discovery_to_complete = 0; _scsih_probe_devices(ioc); } mpt3sas_base_start_watchdog(ioc); ioc->is_driver_loading = 0; } /** * scsih_scan_finished - scsi lld callback for .scan_finished * @shost: SCSI host pointer * @time: elapsed time of the scan in jiffies * * This function will be called periodicallyn until it returns 1 with the * scsi_host and the elapsed time of the scan in jiffies. In our implemention, * we wait for firmware discovery to complete, then return 1. */ static int scsih_scan_finished(struct Scsi_Host *shost, unsigned long time) { struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); u32 ioc_state; int issue_hard_reset = 0; if (disable_discovery > 0) { ioc->is_driver_loading = 0; ioc->wait_for_discovery_to_complete = 0; return 1; } if (time >= (300 * HZ)) { ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED; ioc_info(ioc, "port enable: FAILED with timeout (timeout=300s)\n"); ioc->is_driver_loading = 0; return 1; } if (ioc->start_scan) { ioc_state = mpt3sas_base_get_iocstate(ioc, 0); if ((ioc_state & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_FAULT) { mpt3sas_print_fault_code(ioc, ioc_state & MPI2_DOORBELL_DATA_MASK); issue_hard_reset = 1; goto out; } else if ((ioc_state & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_COREDUMP) { mpt3sas_base_coredump_info(ioc, ioc_state & MPI2_DOORBELL_DATA_MASK); mpt3sas_base_wait_for_coredump_completion(ioc, __func__); issue_hard_reset = 1; goto out; } return 0; } if (ioc->port_enable_cmds.status & MPT3_CMD_RESET) { ioc_info(ioc, "port enable: aborted due to diag reset\n"); ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED; goto out; } if (ioc->start_scan_failed) { ioc_info(ioc, "port enable: FAILED with (ioc_status=0x%08x)\n", ioc->start_scan_failed); ioc->is_driver_loading = 0; ioc->wait_for_discovery_to_complete = 0; ioc->remove_host = 1; return 1; } ioc_info(ioc, "port enable: SUCCESS\n"); ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED; _scsih_complete_devices_scanning(ioc); out: if (issue_hard_reset) { ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED; if (mpt3sas_base_hard_reset_handler(ioc, SOFT_RESET)) ioc->is_driver_loading = 0; } return 1; } /** * scsih_map_queues - map reply queues with request queues * @shost: SCSI host pointer */ static int scsih_map_queues(struct Scsi_Host *shost) { struct MPT3SAS_ADAPTER *ioc = (struct MPT3SAS_ADAPTER *)shost->hostdata; if (ioc->shost->nr_hw_queues == 1) return 0; return blk_mq_pci_map_queues(&shost->tag_set.map[HCTX_TYPE_DEFAULT], ioc->pdev, ioc->high_iops_queues); } /* shost template for SAS 2.0 HBA devices */ static struct scsi_host_template mpt2sas_driver_template = { .module = THIS_MODULE, .name = "Fusion MPT SAS Host", .proc_name = MPT2SAS_DRIVER_NAME, .queuecommand = scsih_qcmd, .target_alloc = scsih_target_alloc, .slave_alloc = scsih_slave_alloc, .slave_configure = scsih_slave_configure, .target_destroy = scsih_target_destroy, .slave_destroy = scsih_slave_destroy, .scan_finished = scsih_scan_finished, .scan_start = scsih_scan_start, .change_queue_depth = scsih_change_queue_depth, .eh_abort_handler = scsih_abort, .eh_device_reset_handler = scsih_dev_reset, .eh_target_reset_handler = scsih_target_reset, .eh_host_reset_handler = scsih_host_reset, .bios_param = scsih_bios_param, .can_queue = 1, .this_id = -1, .sg_tablesize = MPT2SAS_SG_DEPTH, .max_sectors = 32767, .cmd_per_lun = 7, .shost_attrs = mpt3sas_host_attrs, .sdev_attrs = mpt3sas_dev_attrs, .track_queue_depth = 1, .cmd_size = sizeof(struct scsiio_tracker), }; /* raid transport support for SAS 2.0 HBA devices */ static struct raid_function_template mpt2sas_raid_functions = { .cookie = &mpt2sas_driver_template, .is_raid = scsih_is_raid, .get_resync = scsih_get_resync, .get_state = scsih_get_state, }; /* shost template for SAS 3.0 HBA devices */ static struct scsi_host_template mpt3sas_driver_template = { .module = THIS_MODULE, .name = "Fusion MPT SAS Host", .proc_name = MPT3SAS_DRIVER_NAME, .queuecommand = scsih_qcmd, .target_alloc = scsih_target_alloc, .slave_alloc = scsih_slave_alloc, .slave_configure = scsih_slave_configure, .target_destroy = scsih_target_destroy, .slave_destroy = scsih_slave_destroy, .scan_finished = scsih_scan_finished, .scan_start = scsih_scan_start, .change_queue_depth = scsih_change_queue_depth, .eh_abort_handler = scsih_abort, .eh_device_reset_handler = scsih_dev_reset, .eh_target_reset_handler = scsih_target_reset, .eh_host_reset_handler = scsih_host_reset, .bios_param = scsih_bios_param, .can_queue = 1, .this_id = -1, .sg_tablesize = MPT3SAS_SG_DEPTH, .max_sectors = 32767, .max_segment_size = 0xffffffff, .cmd_per_lun = 7, .shost_attrs = mpt3sas_host_attrs, .sdev_attrs = mpt3sas_dev_attrs, .track_queue_depth = 1, .cmd_size = sizeof(struct scsiio_tracker), .map_queues = scsih_map_queues, }; /* raid transport support for SAS 3.0 HBA devices */ static struct raid_function_template mpt3sas_raid_functions = { .cookie = &mpt3sas_driver_template, .is_raid = scsih_is_raid, .get_resync = scsih_get_resync, .get_state = scsih_get_state, }; /** * _scsih_determine_hba_mpi_version - determine in which MPI version class * this device belongs to. * @pdev: PCI device struct * * return MPI2_VERSION for SAS 2.0 HBA devices, * MPI25_VERSION for SAS 3.0 HBA devices, and * MPI26 VERSION for Cutlass & Invader SAS 3.0 HBA devices */ static u16 _scsih_determine_hba_mpi_version(struct pci_dev *pdev) { switch (pdev->device) { case MPI2_MFGPAGE_DEVID_SSS6200: case MPI2_MFGPAGE_DEVID_SAS2004: case MPI2_MFGPAGE_DEVID_SAS2008: case MPI2_MFGPAGE_DEVID_SAS2108_1: case MPI2_MFGPAGE_DEVID_SAS2108_2: case MPI2_MFGPAGE_DEVID_SAS2108_3: case MPI2_MFGPAGE_DEVID_SAS2116_1: case MPI2_MFGPAGE_DEVID_SAS2116_2: case MPI2_MFGPAGE_DEVID_SAS2208_1: case MPI2_MFGPAGE_DEVID_SAS2208_2: case MPI2_MFGPAGE_DEVID_SAS2208_3: case MPI2_MFGPAGE_DEVID_SAS2208_4: case MPI2_MFGPAGE_DEVID_SAS2208_5: case MPI2_MFGPAGE_DEVID_SAS2208_6: case MPI2_MFGPAGE_DEVID_SAS2308_1: case MPI2_MFGPAGE_DEVID_SAS2308_2: case MPI2_MFGPAGE_DEVID_SAS2308_3: case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP: case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1: return MPI2_VERSION; case MPI25_MFGPAGE_DEVID_SAS3004: case MPI25_MFGPAGE_DEVID_SAS3008: case MPI25_MFGPAGE_DEVID_SAS3108_1: case MPI25_MFGPAGE_DEVID_SAS3108_2: case MPI25_MFGPAGE_DEVID_SAS3108_5: case MPI25_MFGPAGE_DEVID_SAS3108_6: return MPI25_VERSION; case MPI26_MFGPAGE_DEVID_SAS3216: case MPI26_MFGPAGE_DEVID_SAS3224: case MPI26_MFGPAGE_DEVID_SAS3316_1: case MPI26_MFGPAGE_DEVID_SAS3316_2: case MPI26_MFGPAGE_DEVID_SAS3316_3: case MPI26_MFGPAGE_DEVID_SAS3316_4: case MPI26_MFGPAGE_DEVID_SAS3324_1: case MPI26_MFGPAGE_DEVID_SAS3324_2: case MPI26_MFGPAGE_DEVID_SAS3324_3: case MPI26_MFGPAGE_DEVID_SAS3324_4: case MPI26_MFGPAGE_DEVID_SAS3508: case MPI26_MFGPAGE_DEVID_SAS3508_1: case MPI26_MFGPAGE_DEVID_SAS3408: case MPI26_MFGPAGE_DEVID_SAS3516: case MPI26_MFGPAGE_DEVID_SAS3516_1: case MPI26_MFGPAGE_DEVID_SAS3416: case MPI26_MFGPAGE_DEVID_SAS3616: case MPI26_ATLAS_PCIe_SWITCH_DEVID: case MPI26_MFGPAGE_DEVID_CFG_SEC_3916: case MPI26_MFGPAGE_DEVID_HARD_SEC_3916: case MPI26_MFGPAGE_DEVID_CFG_SEC_3816: case MPI26_MFGPAGE_DEVID_HARD_SEC_3816: case MPI26_MFGPAGE_DEVID_INVALID0_3916: case MPI26_MFGPAGE_DEVID_INVALID1_3916: case MPI26_MFGPAGE_DEVID_INVALID0_3816: case MPI26_MFGPAGE_DEVID_INVALID1_3816: return MPI26_VERSION; } return 0; } /** * _scsih_probe - attach and add scsi host * @pdev: PCI device struct * @id: pci device id * * Return: 0 success, anything else error. */ static int _scsih_probe(struct pci_dev *pdev, const struct pci_device_id *id) { struct MPT3SAS_ADAPTER *ioc; struct Scsi_Host *shost = NULL; int rv; u16 hba_mpi_version; /* Determine in which MPI version class this pci device belongs */ hba_mpi_version = _scsih_determine_hba_mpi_version(pdev); if (hba_mpi_version == 0) return -ENODEV; /* Enumerate only SAS 2.0 HBA's if hbas_to_enumerate is one, * for other generation HBA's return with -ENODEV */ if ((hbas_to_enumerate == 1) && (hba_mpi_version != MPI2_VERSION)) return -ENODEV; /* Enumerate only SAS 3.0 HBA's if hbas_to_enumerate is two, * for other generation HBA's return with -ENODEV */ if ((hbas_to_enumerate == 2) && (!(hba_mpi_version == MPI25_VERSION || hba_mpi_version == MPI26_VERSION))) return -ENODEV; switch (hba_mpi_version) { case MPI2_VERSION: pci_disable_link_state(pdev, PCIE_LINK_STATE_L0S | PCIE_LINK_STATE_L1 | PCIE_LINK_STATE_CLKPM); /* Use mpt2sas driver host template for SAS 2.0 HBA's */ shost = scsi_host_alloc(&mpt2sas_driver_template, sizeof(struct MPT3SAS_ADAPTER)); if (!shost) return -ENODEV; ioc = shost_priv(shost); memset(ioc, 0, sizeof(struct MPT3SAS_ADAPTER)); ioc->hba_mpi_version_belonged = hba_mpi_version; ioc->id = mpt2_ids++; sprintf(ioc->driver_name, "%s", MPT2SAS_DRIVER_NAME); switch (pdev->device) { case MPI2_MFGPAGE_DEVID_SSS6200: ioc->is_warpdrive = 1; ioc->hide_ir_msg = 1; break; case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP: case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1: ioc->is_mcpu_endpoint = 1; break; default: ioc->mfg_pg10_hide_flag = MFG_PAGE10_EXPOSE_ALL_DISKS; break; } if (multipath_on_hba == -1 || multipath_on_hba == 0) ioc->multipath_on_hba = 0; else ioc->multipath_on_hba = 1; break; case MPI25_VERSION: case MPI26_VERSION: /* Use mpt3sas driver host template for SAS 3.0 HBA's */ shost = scsi_host_alloc(&mpt3sas_driver_template, sizeof(struct MPT3SAS_ADAPTER)); if (!shost) return -ENODEV; ioc = shost_priv(shost); memset(ioc, 0, sizeof(struct MPT3SAS_ADAPTER)); ioc->hba_mpi_version_belonged = hba_mpi_version; ioc->id = mpt3_ids++; sprintf(ioc->driver_name, "%s", MPT3SAS_DRIVER_NAME); switch (pdev->device) { case MPI26_MFGPAGE_DEVID_SAS3508: case MPI26_MFGPAGE_DEVID_SAS3508_1: case MPI26_MFGPAGE_DEVID_SAS3408: case MPI26_MFGPAGE_DEVID_SAS3516: case MPI26_MFGPAGE_DEVID_SAS3516_1: case MPI26_MFGPAGE_DEVID_SAS3416: case MPI26_MFGPAGE_DEVID_SAS3616: case MPI26_ATLAS_PCIe_SWITCH_DEVID: ioc->is_gen35_ioc = 1; break; case MPI26_MFGPAGE_DEVID_INVALID0_3816: case MPI26_MFGPAGE_DEVID_INVALID0_3916: dev_err(&pdev->dev, "HBA with DeviceId 0x%04x, sub VendorId 0x%04x, sub DeviceId 0x%04x is Invalid", pdev->device, pdev->subsystem_vendor, pdev->subsystem_device); return 1; case MPI26_MFGPAGE_DEVID_INVALID1_3816: case MPI26_MFGPAGE_DEVID_INVALID1_3916: dev_err(&pdev->dev, "HBA with DeviceId 0x%04x, sub VendorId 0x%04x, sub DeviceId 0x%04x is Tampered", pdev->device, pdev->subsystem_vendor, pdev->subsystem_device); return 1; case MPI26_MFGPAGE_DEVID_CFG_SEC_3816: case MPI26_MFGPAGE_DEVID_CFG_SEC_3916: dev_info(&pdev->dev, "HBA is in Configurable Secure mode\n"); fallthrough; case MPI26_MFGPAGE_DEVID_HARD_SEC_3816: case MPI26_MFGPAGE_DEVID_HARD_SEC_3916: ioc->is_aero_ioc = ioc->is_gen35_ioc = 1; break; default: ioc->is_gen35_ioc = ioc->is_aero_ioc = 0; } if ((ioc->hba_mpi_version_belonged == MPI25_VERSION && pdev->revision >= SAS3_PCI_DEVICE_C0_REVISION) || (ioc->hba_mpi_version_belonged == MPI26_VERSION)) { ioc->combined_reply_queue = 1; if (ioc->is_gen35_ioc) ioc->combined_reply_index_count = MPT3_SUP_REPLY_POST_HOST_INDEX_REG_COUNT_G35; else ioc->combined_reply_index_count = MPT3_SUP_REPLY_POST_HOST_INDEX_REG_COUNT_G3; } switch (ioc->is_gen35_ioc) { case 0: if (multipath_on_hba == -1 || multipath_on_hba == 0) ioc->multipath_on_hba = 0; else ioc->multipath_on_hba = 1; break; case 1: if (multipath_on_hba == -1 || multipath_on_hba > 0) ioc->multipath_on_hba = 1; else ioc->multipath_on_hba = 0; break; default: break; } break; default: return -ENODEV; } INIT_LIST_HEAD(&ioc->list); spin_lock(&gioc_lock); list_add_tail(&ioc->list, &mpt3sas_ioc_list); spin_unlock(&gioc_lock); ioc->shost = shost; ioc->pdev = pdev; ioc->scsi_io_cb_idx = scsi_io_cb_idx; ioc->tm_cb_idx = tm_cb_idx; ioc->ctl_cb_idx = ctl_cb_idx; ioc->base_cb_idx = base_cb_idx; ioc->port_enable_cb_idx = port_enable_cb_idx; ioc->transport_cb_idx = transport_cb_idx; ioc->scsih_cb_idx = scsih_cb_idx; ioc->config_cb_idx = config_cb_idx; ioc->tm_tr_cb_idx = tm_tr_cb_idx; ioc->tm_tr_volume_cb_idx = tm_tr_volume_cb_idx; ioc->tm_sas_control_cb_idx = tm_sas_control_cb_idx; ioc->logging_level = logging_level; ioc->schedule_dead_ioc_flush_running_cmds = &_scsih_flush_running_cmds; /* Host waits for minimum of six seconds */ ioc->max_shutdown_latency = IO_UNIT_CONTROL_SHUTDOWN_TIMEOUT; /* * Enable MEMORY MOVE support flag. */ ioc->drv_support_bitmap |= MPT_DRV_SUPPORT_BITMAP_MEMMOVE; /* Enable ADDITIONAL QUERY support flag. */ ioc->drv_support_bitmap |= MPT_DRV_SUPPORT_BITMAP_ADDNLQUERY; ioc->enable_sdev_max_qd = enable_sdev_max_qd; /* misc semaphores and spin locks */ mutex_init(&ioc->reset_in_progress_mutex); /* initializing pci_access_mutex lock */ mutex_init(&ioc->pci_access_mutex); spin_lock_init(&ioc->ioc_reset_in_progress_lock); spin_lock_init(&ioc->scsi_lookup_lock); spin_lock_init(&ioc->sas_device_lock); spin_lock_init(&ioc->sas_node_lock); spin_lock_init(&ioc->fw_event_lock); spin_lock_init(&ioc->raid_device_lock); spin_lock_init(&ioc->pcie_device_lock); spin_lock_init(&ioc->diag_trigger_lock); INIT_LIST_HEAD(&ioc->sas_device_list); INIT_LIST_HEAD(&ioc->sas_device_init_list); INIT_LIST_HEAD(&ioc->sas_expander_list); INIT_LIST_HEAD(&ioc->enclosure_list); INIT_LIST_HEAD(&ioc->pcie_device_list); INIT_LIST_HEAD(&ioc->pcie_device_init_list); INIT_LIST_HEAD(&ioc->fw_event_list); INIT_LIST_HEAD(&ioc->raid_device_list); INIT_LIST_HEAD(&ioc->sas_hba.sas_port_list); INIT_LIST_HEAD(&ioc->delayed_tr_list); INIT_LIST_HEAD(&ioc->delayed_sc_list); INIT_LIST_HEAD(&ioc->delayed_event_ack_list); INIT_LIST_HEAD(&ioc->delayed_tr_volume_list); INIT_LIST_HEAD(&ioc->reply_queue_list); INIT_LIST_HEAD(&ioc->port_table_list); sprintf(ioc->name, "%s_cm%d", ioc->driver_name, ioc->id); /* init shost parameters */ shost->max_cmd_len = 32; shost->max_lun = max_lun; shost->transportt = mpt3sas_transport_template; shost->unique_id = ioc->id; if (ioc->is_mcpu_endpoint) { /* mCPU MPI support 64K max IO */ shost->max_sectors = 128; ioc_info(ioc, "The max_sectors value is set to %d\n", shost->max_sectors); } else { if (max_sectors != 0xFFFF) { if (max_sectors < 64) { shost->max_sectors = 64; ioc_warn(ioc, "Invalid value %d passed for max_sectors, range is 64 to 32767. Assigning value of 64.\n", max_sectors); } else if (max_sectors > 32767) { shost->max_sectors = 32767; ioc_warn(ioc, "Invalid value %d passed for max_sectors, range is 64 to 32767.Assigning default value of 32767.\n", max_sectors); } else { shost->max_sectors = max_sectors & 0xFFFE; ioc_info(ioc, "The max_sectors value is set to %d\n", shost->max_sectors); } } } /* register EEDP capabilities with SCSI layer */ if (prot_mask >= 0) scsi_host_set_prot(shost, (prot_mask & 0x07)); else scsi_host_set_prot(shost, SHOST_DIF_TYPE1_PROTECTION | SHOST_DIF_TYPE2_PROTECTION | SHOST_DIF_TYPE3_PROTECTION); scsi_host_set_guard(shost, SHOST_DIX_GUARD_CRC); /* event thread */ snprintf(ioc->firmware_event_name, sizeof(ioc->firmware_event_name), "fw_event_%s%d", ioc->driver_name, ioc->id); ioc->firmware_event_thread = alloc_ordered_workqueue( ioc->firmware_event_name, 0); if (!ioc->firmware_event_thread) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); rv = -ENODEV; goto out_thread_fail; } ioc->is_driver_loading = 1; if ((mpt3sas_base_attach(ioc))) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); rv = -ENODEV; goto out_attach_fail; } if (ioc->is_warpdrive) { if (ioc->mfg_pg10_hide_flag == MFG_PAGE10_EXPOSE_ALL_DISKS) ioc->hide_drives = 0; else if (ioc->mfg_pg10_hide_flag == MFG_PAGE10_HIDE_ALL_DISKS) ioc->hide_drives = 1; else { if (mpt3sas_get_num_volumes(ioc)) ioc->hide_drives = 1; else ioc->hide_drives = 0; } } else ioc->hide_drives = 0; shost->host_tagset = 0; shost->nr_hw_queues = 1; if (ioc->is_gen35_ioc && ioc->reply_queue_count > 1 && host_tagset_enable && ioc->smp_affinity_enable) { shost->host_tagset = 1; shost->nr_hw_queues = ioc->reply_queue_count - ioc->high_iops_queues; dev_info(&ioc->pdev->dev, "Max SCSIIO MPT commands: %d shared with nr_hw_queues = %d\n", shost->can_queue, shost->nr_hw_queues); } rv = scsi_add_host(shost, &pdev->dev); if (rv) { ioc_err(ioc, "failure at %s:%d/%s()!\n", __FILE__, __LINE__, __func__); goto out_add_shost_fail; } scsi_scan_host(shost); mpt3sas_setup_debugfs(ioc); return 0; out_add_shost_fail: mpt3sas_base_detach(ioc); out_attach_fail: destroy_workqueue(ioc->firmware_event_thread); out_thread_fail: spin_lock(&gioc_lock); list_del(&ioc->list); spin_unlock(&gioc_lock); scsi_host_put(shost); return rv; } /** * scsih_suspend - power management suspend main entry point * @dev: Device struct * * Return: 0 success, anything else error. */ static int __maybe_unused scsih_suspend(struct device *dev) { struct pci_dev *pdev = to_pci_dev(dev); struct Scsi_Host *shost; struct MPT3SAS_ADAPTER *ioc; int rc; rc = _scsih_get_shost_and_ioc(pdev, &shost, &ioc); if (rc) return rc; mpt3sas_base_stop_watchdog(ioc); flush_scheduled_work(); scsi_block_requests(shost); _scsih_nvme_shutdown(ioc); ioc_info(ioc, "pdev=0x%p, slot=%s, entering operating state\n", pdev, pci_name(pdev)); mpt3sas_base_free_resources(ioc); return 0; } /** * scsih_resume - power management resume main entry point * @dev: Device struct * * Return: 0 success, anything else error. */ static int __maybe_unused scsih_resume(struct device *dev) { struct pci_dev *pdev = to_pci_dev(dev); struct Scsi_Host *shost; struct MPT3SAS_ADAPTER *ioc; pci_power_t device_state = pdev->current_state; int r; r = _scsih_get_shost_and_ioc(pdev, &shost, &ioc); if (r) return r; ioc_info(ioc, "pdev=0x%p, slot=%s, previous operating state [D%d]\n", pdev, pci_name(pdev), device_state); ioc->pdev = pdev; r = mpt3sas_base_map_resources(ioc); if (r) return r; ioc_info(ioc, "Issuing Hard Reset as part of OS Resume\n"); mpt3sas_base_hard_reset_handler(ioc, SOFT_RESET); scsi_unblock_requests(shost); mpt3sas_base_start_watchdog(ioc); return 0; } /** * scsih_pci_error_detected - Called when a PCI error is detected. * @pdev: PCI device struct * @state: PCI channel state * * Description: Called when a PCI error is detected. * * Return: PCI_ERS_RESULT_NEED_RESET or PCI_ERS_RESULT_DISCONNECT. */ static pci_ers_result_t scsih_pci_error_detected(struct pci_dev *pdev, pci_channel_state_t state) { struct Scsi_Host *shost; struct MPT3SAS_ADAPTER *ioc; if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc)) return PCI_ERS_RESULT_DISCONNECT; ioc_info(ioc, "PCI error: detected callback, state(%d)!!\n", state); switch (state) { case pci_channel_io_normal: return PCI_ERS_RESULT_CAN_RECOVER; case pci_channel_io_frozen: /* Fatal error, prepare for slot reset */ ioc->pci_error_recovery = 1; scsi_block_requests(ioc->shost); mpt3sas_base_stop_watchdog(ioc); mpt3sas_base_free_resources(ioc); return PCI_ERS_RESULT_NEED_RESET; case pci_channel_io_perm_failure: /* Permanent error, prepare for device removal */ ioc->pci_error_recovery = 1; mpt3sas_base_stop_watchdog(ioc); _scsih_flush_running_cmds(ioc); return PCI_ERS_RESULT_DISCONNECT; } return PCI_ERS_RESULT_NEED_RESET; } /** * scsih_pci_slot_reset - Called when PCI slot has been reset. * @pdev: PCI device struct * * Description: This routine is called by the pci error recovery * code after the PCI slot has been reset, just before we * should resume normal operations. */ static pci_ers_result_t scsih_pci_slot_reset(struct pci_dev *pdev) { struct Scsi_Host *shost; struct MPT3SAS_ADAPTER *ioc; int rc; if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc)) return PCI_ERS_RESULT_DISCONNECT; ioc_info(ioc, "PCI error: slot reset callback!!\n"); ioc->pci_error_recovery = 0; ioc->pdev = pdev; pci_restore_state(pdev); rc = mpt3sas_base_map_resources(ioc); if (rc) return PCI_ERS_RESULT_DISCONNECT; ioc_info(ioc, "Issuing Hard Reset as part of PCI Slot Reset\n"); rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); ioc_warn(ioc, "hard reset: %s\n", (rc == 0) ? "success" : "failed"); if (!rc) return PCI_ERS_RESULT_RECOVERED; else return PCI_ERS_RESULT_DISCONNECT; } /** * scsih_pci_resume() - resume normal ops after PCI reset * @pdev: pointer to PCI device * * Called when the error recovery driver tells us that its * OK to resume normal operation. Use completion to allow * halted scsi ops to resume. */ static void scsih_pci_resume(struct pci_dev *pdev) { struct Scsi_Host *shost; struct MPT3SAS_ADAPTER *ioc; if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc)) return; ioc_info(ioc, "PCI error: resume callback!!\n"); mpt3sas_base_start_watchdog(ioc); scsi_unblock_requests(ioc->shost); } /** * scsih_pci_mmio_enabled - Enable MMIO and dump debug registers * @pdev: pointer to PCI device */ static pci_ers_result_t scsih_pci_mmio_enabled(struct pci_dev *pdev) { struct Scsi_Host *shost; struct MPT3SAS_ADAPTER *ioc; if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc)) return PCI_ERS_RESULT_DISCONNECT; ioc_info(ioc, "PCI error: mmio enabled callback!!\n"); /* TODO - dump whatever for debugging purposes */ /* This called only if scsih_pci_error_detected returns * PCI_ERS_RESULT_CAN_RECOVER. Read/write to the device still * works, no need to reset slot. */ return PCI_ERS_RESULT_RECOVERED; } /** * scsih_ncq_prio_supp - Check for NCQ command priority support * @sdev: scsi device struct * * This is called when a user indicates they would like to enable * ncq command priorities. This works only on SATA devices. */ bool scsih_ncq_prio_supp(struct scsi_device *sdev) { unsigned char *buf; bool ncq_prio_supp = false; if (!scsi_device_supports_vpd(sdev)) return ncq_prio_supp; buf = kmalloc(SCSI_VPD_PG_LEN, GFP_KERNEL); if (!buf) return ncq_prio_supp; if (!scsi_get_vpd_page(sdev, 0x89, buf, SCSI_VPD_PG_LEN)) ncq_prio_supp = (buf[213] >> 4) & 1; kfree(buf); return ncq_prio_supp; } /* * The pci device ids are defined in mpi/mpi2_cnfg.h. */ static const struct pci_device_id mpt3sas_pci_table[] = { /* Spitfire ~ 2004 */ { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2004, PCI_ANY_ID, PCI_ANY_ID }, /* Falcon ~ 2008 */ { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2008, PCI_ANY_ID, PCI_ANY_ID }, /* Liberator ~ 2108 */ { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_1, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_2, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_3, PCI_ANY_ID, PCI_ANY_ID }, /* Meteor ~ 2116 */ { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2116_1, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2116_2, PCI_ANY_ID, PCI_ANY_ID }, /* Thunderbolt ~ 2208 */ { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_1, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_2, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_3, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_4, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_5, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_6, PCI_ANY_ID, PCI_ANY_ID }, /* Mustang ~ 2308 */ { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_1, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_2, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_3, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1, PCI_ANY_ID, PCI_ANY_ID }, /* SSS6200 */ { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SSS6200, PCI_ANY_ID, PCI_ANY_ID }, /* Fury ~ 3004 and 3008 */ { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3004, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3008, PCI_ANY_ID, PCI_ANY_ID }, /* Invader ~ 3108 */ { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_1, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_2, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_5, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_6, PCI_ANY_ID, PCI_ANY_ID }, /* Cutlass ~ 3216 and 3224 */ { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3216, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3224, PCI_ANY_ID, PCI_ANY_ID }, /* Intruder ~ 3316 and 3324 */ { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_1, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_2, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_3, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_4, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_1, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_2, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_3, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_4, PCI_ANY_ID, PCI_ANY_ID }, /* Ventura, Crusader, Harpoon & Tomcat ~ 3516, 3416, 3508 & 3408*/ { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3508, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3508_1, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3408, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3516, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3516_1, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3416, PCI_ANY_ID, PCI_ANY_ID }, /* Mercator ~ 3616*/ { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3616, PCI_ANY_ID, PCI_ANY_ID }, /* Aero SI 0x00E1 Configurable Secure * 0x00E2 Hard Secure */ { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_CFG_SEC_3916, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_HARD_SEC_3916, PCI_ANY_ID, PCI_ANY_ID }, /* * Aero SI –> 0x00E0 Invalid, 0x00E3 Tampered */ { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID0_3916, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID1_3916, PCI_ANY_ID, PCI_ANY_ID }, /* Atlas PCIe Switch Management Port */ { MPI2_MFGPAGE_VENDORID_LSI, MPI26_ATLAS_PCIe_SWITCH_DEVID, PCI_ANY_ID, PCI_ANY_ID }, /* Sea SI 0x00E5 Configurable Secure * 0x00E6 Hard Secure */ { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_CFG_SEC_3816, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_HARD_SEC_3816, PCI_ANY_ID, PCI_ANY_ID }, /* * Sea SI –> 0x00E4 Invalid, 0x00E7 Tampered */ { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID0_3816, PCI_ANY_ID, PCI_ANY_ID }, { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID1_3816, PCI_ANY_ID, PCI_ANY_ID }, {0} /* Terminating entry */ }; MODULE_DEVICE_TABLE(pci, mpt3sas_pci_table); static struct pci_error_handlers _mpt3sas_err_handler = { .error_detected = scsih_pci_error_detected, .mmio_enabled = scsih_pci_mmio_enabled, .slot_reset = scsih_pci_slot_reset, .resume = scsih_pci_resume, }; static SIMPLE_DEV_PM_OPS(scsih_pm_ops, scsih_suspend, scsih_resume); static struct pci_driver mpt3sas_driver = { .name = MPT3SAS_DRIVER_NAME, .id_table = mpt3sas_pci_table, .probe = _scsih_probe, .remove = scsih_remove, .shutdown = scsih_shutdown, .err_handler = &_mpt3sas_err_handler, .driver.pm = &scsih_pm_ops, }; /** * scsih_init - main entry point for this driver. * * Return: 0 success, anything else error. */ static int scsih_init(void) { mpt2_ids = 0; mpt3_ids = 0; mpt3sas_base_initialize_callback_handler(); /* queuecommand callback hander */ scsi_io_cb_idx = mpt3sas_base_register_callback_handler(_scsih_io_done); /* task management callback handler */ tm_cb_idx = mpt3sas_base_register_callback_handler(_scsih_tm_done); /* base internal commands callback handler */ base_cb_idx = mpt3sas_base_register_callback_handler(mpt3sas_base_done); port_enable_cb_idx = mpt3sas_base_register_callback_handler( mpt3sas_port_enable_done); /* transport internal commands callback handler */ transport_cb_idx = mpt3sas_base_register_callback_handler( mpt3sas_transport_done); /* scsih internal commands callback handler */ scsih_cb_idx = mpt3sas_base_register_callback_handler(_scsih_done); /* configuration page API internal commands callback handler */ config_cb_idx = mpt3sas_base_register_callback_handler( mpt3sas_config_done); /* ctl module callback handler */ ctl_cb_idx = mpt3sas_base_register_callback_handler(mpt3sas_ctl_done); tm_tr_cb_idx = mpt3sas_base_register_callback_handler( _scsih_tm_tr_complete); tm_tr_volume_cb_idx = mpt3sas_base_register_callback_handler( _scsih_tm_volume_tr_complete); tm_sas_control_cb_idx = mpt3sas_base_register_callback_handler( _scsih_sas_control_complete); mpt3sas_init_debugfs(); return 0; } /** * scsih_exit - exit point for this driver (when it is a module). * * Return: 0 success, anything else error. */ static void scsih_exit(void) { mpt3sas_base_release_callback_handler(scsi_io_cb_idx); mpt3sas_base_release_callback_handler(tm_cb_idx); mpt3sas_base_release_callback_handler(base_cb_idx); mpt3sas_base_release_callback_handler(port_enable_cb_idx); mpt3sas_base_release_callback_handler(transport_cb_idx); mpt3sas_base_release_callback_handler(scsih_cb_idx); mpt3sas_base_release_callback_handler(config_cb_idx); mpt3sas_base_release_callback_handler(ctl_cb_idx); mpt3sas_base_release_callback_handler(tm_tr_cb_idx); mpt3sas_base_release_callback_handler(tm_tr_volume_cb_idx); mpt3sas_base_release_callback_handler(tm_sas_control_cb_idx); /* raid transport support */ if (hbas_to_enumerate != 1) raid_class_release(mpt3sas_raid_template); if (hbas_to_enumerate != 2) raid_class_release(mpt2sas_raid_template); sas_release_transport(mpt3sas_transport_template); mpt3sas_exit_debugfs(); } /** * _mpt3sas_init - main entry point for this driver. * * Return: 0 success, anything else error. */ static int __init _mpt3sas_init(void) { int error; pr_info("%s version %s loaded\n", MPT3SAS_DRIVER_NAME, MPT3SAS_DRIVER_VERSION); mpt3sas_transport_template = sas_attach_transport(&mpt3sas_transport_functions); if (!mpt3sas_transport_template) return -ENODEV; /* No need attach mpt3sas raid functions template * if hbas_to_enumarate value is one. */ if (hbas_to_enumerate != 1) { mpt3sas_raid_template = raid_class_attach(&mpt3sas_raid_functions); if (!mpt3sas_raid_template) { sas_release_transport(mpt3sas_transport_template); return -ENODEV; } } /* No need to attach mpt2sas raid functions template * if hbas_to_enumarate value is two */ if (hbas_to_enumerate != 2) { mpt2sas_raid_template = raid_class_attach(&mpt2sas_raid_functions); if (!mpt2sas_raid_template) { sas_release_transport(mpt3sas_transport_template); return -ENODEV; } } error = scsih_init(); if (error) { scsih_exit(); return error; } mpt3sas_ctl_init(hbas_to_enumerate); error = pci_register_driver(&mpt3sas_driver); if (error) scsih_exit(); return error; } /** * _mpt3sas_exit - exit point for this driver (when it is a module). * */ static void __exit _mpt3sas_exit(void) { pr_info("mpt3sas version %s unloading\n", MPT3SAS_DRIVER_VERSION); mpt3sas_ctl_exit(hbas_to_enumerate); pci_unregister_driver(&mpt3sas_driver); scsih_exit(); } module_init(_mpt3sas_init); module_exit(_mpt3sas_exit);