Lines Matching +full:hba +full:- +full:cap
1 // SPDX-License-Identifier: GPL-2.0-only
3 * Copyright (c) 2013-2016, Linux Foundation. All rights reserved.
16 #include <linux/reset-controller.h>
22 #include "ufshcd-pltfrm.h"
24 #include "ufs-qcom.h"
102 static int ufs_qcom_set_dme_vs_core_clk_ctrl_clear_div(struct ufs_hba *hba,
114 if (host->hba->caps & UFSHCD_CAP_CRYPTO) in ufs_qcom_ice_enable()
115 qcom_ice_enable(host->ice); in ufs_qcom_ice_enable()
120 struct ufs_hba *hba = host->hba; in ufs_qcom_ice_init() local
121 struct device *dev = hba->dev; in ufs_qcom_ice_init()
125 if (ice == ERR_PTR(-EOPNOTSUPP)) { in ufs_qcom_ice_init()
133 host->ice = ice; in ufs_qcom_ice_init()
134 hba->caps |= UFSHCD_CAP_CRYPTO; in ufs_qcom_ice_init()
141 if (host->hba->caps & UFSHCD_CAP_CRYPTO) in ufs_qcom_ice_resume()
142 return qcom_ice_resume(host->ice); in ufs_qcom_ice_resume()
149 if (host->hba->caps & UFSHCD_CAP_CRYPTO) in ufs_qcom_ice_suspend()
150 return qcom_ice_suspend(host->ice); in ufs_qcom_ice_suspend()
155 static int ufs_qcom_ice_program_key(struct ufs_hba *hba, in ufs_qcom_ice_program_key() argument
159 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_ice_program_key()
160 union ufs_crypto_cap_entry cap; in ufs_qcom_ice_program_key() local
162 if (!(cfg->config_enable & UFS_CRYPTO_CONFIGURATION_ENABLE)) in ufs_qcom_ice_program_key()
163 return qcom_ice_evict_key(host->ice, slot); in ufs_qcom_ice_program_key()
165 /* Only AES-256-XTS has been tested so far. */ in ufs_qcom_ice_program_key()
166 cap = hba->crypto_cap_array[cfg->crypto_cap_idx]; in ufs_qcom_ice_program_key()
167 if (cap.algorithm_id != UFS_CRYPTO_ALG_AES_XTS || in ufs_qcom_ice_program_key()
168 cap.key_size != UFS_CRYPTO_KEY_SIZE_256) in ufs_qcom_ice_program_key()
169 return -EOPNOTSUPP; in ufs_qcom_ice_program_key()
171 return qcom_ice_program_key(host->ice, in ufs_qcom_ice_program_key()
174 cfg->crypto_key, in ufs_qcom_ice_program_key()
175 cfg->data_unit_size, slot); in ufs_qcom_ice_program_key()
216 if (optional && err == -ENOENT) { in ufs_qcom_host_clk_get()
221 if (err != -EPROBE_DEFER) in ufs_qcom_host_clk_get()
241 if (!host->is_lane_clks_enabled) in ufs_qcom_disable_lane_clks()
244 clk_disable_unprepare(host->tx_l1_sync_clk); in ufs_qcom_disable_lane_clks()
245 clk_disable_unprepare(host->tx_l0_sync_clk); in ufs_qcom_disable_lane_clks()
246 clk_disable_unprepare(host->rx_l1_sync_clk); in ufs_qcom_disable_lane_clks()
247 clk_disable_unprepare(host->rx_l0_sync_clk); in ufs_qcom_disable_lane_clks()
249 host->is_lane_clks_enabled = false; in ufs_qcom_disable_lane_clks()
255 struct device *dev = host->hba->dev; in ufs_qcom_enable_lane_clks()
257 if (host->is_lane_clks_enabled) in ufs_qcom_enable_lane_clks()
261 host->rx_l0_sync_clk); in ufs_qcom_enable_lane_clks()
266 host->tx_l0_sync_clk); in ufs_qcom_enable_lane_clks()
271 host->rx_l1_sync_clk); in ufs_qcom_enable_lane_clks()
276 host->tx_l1_sync_clk); in ufs_qcom_enable_lane_clks()
280 host->is_lane_clks_enabled = true; in ufs_qcom_enable_lane_clks()
285 clk_disable_unprepare(host->rx_l1_sync_clk); in ufs_qcom_enable_lane_clks()
287 clk_disable_unprepare(host->tx_l0_sync_clk); in ufs_qcom_enable_lane_clks()
289 clk_disable_unprepare(host->rx_l0_sync_clk); in ufs_qcom_enable_lane_clks()
297 struct device *dev = host->hba->dev; in ufs_qcom_init_lane_clks()
303 &host->rx_l0_sync_clk, false); in ufs_qcom_init_lane_clks()
308 &host->tx_l0_sync_clk, false); in ufs_qcom_init_lane_clks()
313 if (host->hba->lanes_per_direction > 1) { in ufs_qcom_init_lane_clks()
315 &host->rx_l1_sync_clk, false); in ufs_qcom_init_lane_clks()
320 &host->tx_l1_sync_clk, true); in ufs_qcom_init_lane_clks()
326 static int ufs_qcom_check_hibern8(struct ufs_hba *hba) in ufs_qcom_check_hibern8() argument
333 err = ufshcd_dme_get(hba, in ufs_qcom_check_hibern8()
349 err = ufshcd_dme_get(hba, in ufs_qcom_check_hibern8()
355 dev_err(hba->dev, "%s: unable to get TX_FSM_STATE, err %d\n", in ufs_qcom_check_hibern8()
359 dev_err(hba->dev, "%s: invalid TX_FSM_STATE = %d\n", in ufs_qcom_check_hibern8()
368 ufshcd_rmwl(host->hba, QUNIPRO_SEL, in ufs_qcom_select_unipro_mode()
372 if (host->hw_ver.major >= 0x05) in ufs_qcom_select_unipro_mode()
373 ufshcd_rmwl(host->hba, QUNIPRO_G4_SEL, 0, REG_UFS_CFG0); in ufs_qcom_select_unipro_mode()
377 * ufs_qcom_host_reset - reset host controller and PHY
379 static int ufs_qcom_host_reset(struct ufs_hba *hba) in ufs_qcom_host_reset() argument
382 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_host_reset()
385 if (!host->core_reset) { in ufs_qcom_host_reset()
386 dev_warn(hba->dev, "%s: reset control not set\n", __func__); in ufs_qcom_host_reset()
390 reenable_intr = hba->is_irq_enabled; in ufs_qcom_host_reset()
391 disable_irq(hba->irq); in ufs_qcom_host_reset()
392 hba->is_irq_enabled = false; in ufs_qcom_host_reset()
394 ret = reset_control_assert(host->core_reset); in ufs_qcom_host_reset()
396 dev_err(hba->dev, "%s: core_reset assert failed, err = %d\n", in ufs_qcom_host_reset()
403 * is at least 3-4 sleep clock (32.7KHz) cycles, which comes to in ufs_qcom_host_reset()
408 ret = reset_control_deassert(host->core_reset); in ufs_qcom_host_reset()
410 dev_err(hba->dev, "%s: core_reset deassert failed, err = %d\n", in ufs_qcom_host_reset()
416 enable_irq(hba->irq); in ufs_qcom_host_reset()
417 hba->is_irq_enabled = true; in ufs_qcom_host_reset()
423 static u32 ufs_qcom_get_hs_gear(struct ufs_hba *hba) in ufs_qcom_get_hs_gear() argument
425 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_get_hs_gear()
427 if (host->hw_ver.major == 0x1) { in ufs_qcom_get_hs_gear()
429 * HS-G3 operations may not reliably work on legacy QCOM in ufs_qcom_get_hs_gear()
433 * Hence downgrade the maximum supported gear to HS-G2. in ufs_qcom_get_hs_gear()
436 } else if (host->hw_ver.major >= 0x4) { in ufs_qcom_get_hs_gear()
437 return UFS_QCOM_MAX_GEAR(ufshcd_readl(hba, REG_UFS_PARAM0)); in ufs_qcom_get_hs_gear()
440 /* Default is HS-G3 */ in ufs_qcom_get_hs_gear()
444 static int ufs_qcom_power_up_sequence(struct ufs_hba *hba) in ufs_qcom_power_up_sequence() argument
446 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_power_up_sequence()
447 struct phy *phy = host->generic_phy; in ufs_qcom_power_up_sequence()
451 ret = ufs_qcom_host_reset(hba); in ufs_qcom_power_up_sequence()
453 dev_warn(hba->dev, "%s: host reset returned %d\n", in ufs_qcom_power_up_sequence()
456 if (phy->power_count) { in ufs_qcom_power_up_sequence()
461 /* phy initialization - calibrate the phy */ in ufs_qcom_power_up_sequence()
464 dev_err(hba->dev, "%s: phy init failed, ret = %d\n", in ufs_qcom_power_up_sequence()
469 phy_set_mode_ext(phy, PHY_MODE_UFS_HS_B, host->hs_gear); in ufs_qcom_power_up_sequence()
471 /* power on phy - start serdes and phy's power and clocks */ in ufs_qcom_power_up_sequence()
474 dev_err(hba->dev, "%s: phy power on failed, ret = %d\n", in ufs_qcom_power_up_sequence()
491 * Internal hardware sub-modules within the UTP controller control the CGCs.
492 * Hardware CGCs disable the clock to inactivate UTP sub-modules not involved
497 static void ufs_qcom_enable_hw_clk_gating(struct ufs_hba *hba) in ufs_qcom_enable_hw_clk_gating() argument
499 ufshcd_writel(hba, in ufs_qcom_enable_hw_clk_gating()
500 ufshcd_readl(hba, REG_UFS_CFG2) | REG_UFS_CFG2_CGC_EN_ALL, in ufs_qcom_enable_hw_clk_gating()
504 ufshcd_readl(hba, REG_UFS_CFG2); in ufs_qcom_enable_hw_clk_gating()
507 static int ufs_qcom_hce_enable_notify(struct ufs_hba *hba, in ufs_qcom_hce_enable_notify() argument
510 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_hce_enable_notify()
515 ufs_qcom_power_up_sequence(hba); in ufs_qcom_hce_enable_notify()
525 err = ufs_qcom_check_hibern8(hba); in ufs_qcom_hce_enable_notify()
526 ufs_qcom_enable_hw_clk_gating(hba); in ufs_qcom_hce_enable_notify()
530 dev_err(hba->dev, "%s: invalid status %d\n", __func__, status); in ufs_qcom_hce_enable_notify()
531 err = -EINVAL; in ufs_qcom_hce_enable_notify()
538 * Return: zero for success and non-zero in case of a failure.
540 static int ufs_qcom_cfg_timers(struct ufs_hba *hba, u32 gear, in ufs_qcom_cfg_timers() argument
543 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_cfg_timers()
576 if (ufs_qcom_cap_qunipro(host) && !ufshcd_is_intr_aggr_allowed(hba)) in ufs_qcom_cfg_timers()
580 dev_err(hba->dev, "%s: invalid gear = %d\n", __func__, gear); in ufs_qcom_cfg_timers()
581 return -EINVAL; in ufs_qcom_cfg_timers()
584 list_for_each_entry(clki, &hba->clk_list_head, list) { in ufs_qcom_cfg_timers()
585 if (!strcmp(clki->name, "core_clk")) in ufs_qcom_cfg_timers()
586 core_clk_rate = clk_get_rate(clki->clk); in ufs_qcom_cfg_timers()
594 if (ufshcd_readl(hba, REG_UFS_SYS1CLK_1US) != core_clk_cycles_per_us) { in ufs_qcom_cfg_timers()
595 ufshcd_writel(hba, core_clk_cycles_per_us, REG_UFS_SYS1CLK_1US); in ufs_qcom_cfg_timers()
600 ufshcd_readl(hba, REG_UFS_SYS1CLK_1US); in ufs_qcom_cfg_timers()
615 dev_err(hba->dev, in ufs_qcom_cfg_timers()
619 return -EINVAL; in ufs_qcom_cfg_timers()
621 tx_clk_cycles_per_us = hs_fr_table_rA[gear-1][1]; in ufs_qcom_cfg_timers()
624 dev_err(hba->dev, in ufs_qcom_cfg_timers()
628 return -EINVAL; in ufs_qcom_cfg_timers()
630 tx_clk_cycles_per_us = hs_fr_table_rB[gear-1][1]; in ufs_qcom_cfg_timers()
632 dev_err(hba->dev, "%s: invalid rate = %d\n", in ufs_qcom_cfg_timers()
634 return -EINVAL; in ufs_qcom_cfg_timers()
640 dev_err(hba->dev, in ufs_qcom_cfg_timers()
644 return -EINVAL; in ufs_qcom_cfg_timers()
646 tx_clk_cycles_per_us = pwm_fr_table[gear-1][1]; in ufs_qcom_cfg_timers()
650 dev_err(hba->dev, "%s: invalid mode = %d\n", __func__, hs); in ufs_qcom_cfg_timers()
651 return -EINVAL; in ufs_qcom_cfg_timers()
654 if (ufshcd_readl(hba, REG_UFS_TX_SYMBOL_CLK_NS_US) != in ufs_qcom_cfg_timers()
657 ufshcd_writel(hba, core_clk_period_in_ns | tx_clk_cycles_per_us, in ufs_qcom_cfg_timers()
666 if (update_link_startup_timer && host->hw_ver.major != 0x5) { in ufs_qcom_cfg_timers()
667 ufshcd_writel(hba, ((core_clk_rate / MSEC_PER_SEC) * 100), in ufs_qcom_cfg_timers()
679 static int ufs_qcom_link_startup_notify(struct ufs_hba *hba, in ufs_qcom_link_startup_notify() argument
683 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_link_startup_notify()
687 if (ufs_qcom_cfg_timers(hba, UFS_PWM_G1, SLOWAUTO_MODE, in ufs_qcom_link_startup_notify()
689 dev_err(hba->dev, "%s: ufs_qcom_cfg_timers() failed\n", in ufs_qcom_link_startup_notify()
691 return -EINVAL; in ufs_qcom_link_startup_notify()
699 err = ufs_qcom_set_dme_vs_core_clk_ctrl_clear_div(hba, in ufs_qcom_link_startup_notify()
709 if (ufshcd_get_local_unipro_ver(hba) != UFS_UNIPRO_VER_1_41) in ufs_qcom_link_startup_notify()
710 err = ufshcd_disable_host_tx_lcc(hba); in ufs_qcom_link_startup_notify()
720 static void ufs_qcom_device_reset_ctrl(struct ufs_hba *hba, bool asserted) in ufs_qcom_device_reset_ctrl() argument
722 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_device_reset_ctrl()
725 if (!host->device_reset) in ufs_qcom_device_reset_ctrl()
728 gpiod_set_value_cansleep(host->device_reset, asserted); in ufs_qcom_device_reset_ctrl()
731 static int ufs_qcom_suspend(struct ufs_hba *hba, enum ufs_pm_op pm_op, in ufs_qcom_suspend() argument
734 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_suspend()
735 struct phy *phy = host->generic_phy; in ufs_qcom_suspend()
740 if (ufs_qcom_is_link_off(hba)) { in ufs_qcom_suspend()
750 ufs_qcom_device_reset_ctrl(hba, true); in ufs_qcom_suspend()
752 } else if (!ufs_qcom_is_link_active(hba)) { in ufs_qcom_suspend()
759 static int ufs_qcom_resume(struct ufs_hba *hba, enum ufs_pm_op pm_op) in ufs_qcom_resume() argument
761 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_resume()
762 struct phy *phy = host->generic_phy; in ufs_qcom_resume()
765 if (ufs_qcom_is_link_off(hba)) { in ufs_qcom_resume()
768 dev_err(hba->dev, "%s: failed PHY power on: %d\n", in ufs_qcom_resume()
777 } else if (!ufs_qcom_is_link_active(hba)) { in ufs_qcom_resume()
788 if (host->dev_ref_clk_ctrl_mmio && in ufs_qcom_dev_ref_clk_ctrl()
789 (enable ^ host->is_dev_ref_clk_enabled)) { in ufs_qcom_dev_ref_clk_ctrl()
790 u32 temp = readl_relaxed(host->dev_ref_clk_ctrl_mmio); in ufs_qcom_dev_ref_clk_ctrl()
793 temp |= host->dev_ref_clk_en_mask; in ufs_qcom_dev_ref_clk_ctrl()
795 temp &= ~host->dev_ref_clk_en_mask; in ufs_qcom_dev_ref_clk_ctrl()
806 gating_wait = host->hba->dev_info.clk_gating_wait_us; in ufs_qcom_dev_ref_clk_ctrl()
814 * HS-MODE to LS-MODE or HIBERN8 state. Give it in ufs_qcom_dev_ref_clk_ctrl()
822 writel_relaxed(temp, host->dev_ref_clk_ctrl_mmio); in ufs_qcom_dev_ref_clk_ctrl()
828 readl(host->dev_ref_clk_ctrl_mmio); in ufs_qcom_dev_ref_clk_ctrl()
838 host->is_dev_ref_clk_enabled = enable; in ufs_qcom_dev_ref_clk_ctrl()
844 struct device *dev = host->hba->dev; in ufs_qcom_icc_set_bw()
847 ret = icc_set_bw(host->icc_ddr, 0, mem_bw); in ufs_qcom_icc_set_bw()
853 ret = icc_set_bw(host->icc_cpu, 0, cfg_bw); in ufs_qcom_icc_set_bw()
864 struct ufs_pa_layer_attr *p = &host->dev_req_params; in ufs_qcom_get_bw_table()
865 int gear = max_t(u32, p->gear_rx, p->gear_tx); in ufs_qcom_get_bw_table()
866 int lane = max_t(u32, p->lane_rx, p->lane_tx); in ufs_qcom_get_bw_table()
869 if (p->hs_rate == PA_HS_MODE_B) in ufs_qcom_get_bw_table()
887 static int ufs_qcom_pwr_change_notify(struct ufs_hba *hba, in ufs_qcom_pwr_change_notify() argument
892 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_pwr_change_notify()
898 return -EINVAL; in ufs_qcom_pwr_change_notify()
907 ufs_qcom_cap.hs_tx_gear = ufs_qcom_cap.hs_rx_gear = ufs_qcom_get_hs_gear(hba); in ufs_qcom_pwr_change_notify()
913 dev_err(hba->dev, "%s: failed to determine capabilities\n", in ufs_qcom_pwr_change_notify()
923 if (dev_req_params->gear_tx > host->hs_gear) in ufs_qcom_pwr_change_notify()
924 host->hs_gear = dev_req_params->gear_tx; in ufs_qcom_pwr_change_notify()
927 if (!ufshcd_is_hs_mode(&hba->pwr_info) && in ufs_qcom_pwr_change_notify()
931 if (host->hw_ver.major >= 0x4) { in ufs_qcom_pwr_change_notify()
932 ufshcd_dme_configure_adapt(hba, in ufs_qcom_pwr_change_notify()
933 dev_req_params->gear_tx, in ufs_qcom_pwr_change_notify()
938 if (ufs_qcom_cfg_timers(hba, dev_req_params->gear_rx, in ufs_qcom_pwr_change_notify()
939 dev_req_params->pwr_rx, in ufs_qcom_pwr_change_notify()
940 dev_req_params->hs_rate, false)) { in ufs_qcom_pwr_change_notify()
941 dev_err(hba->dev, "%s: ufs_qcom_cfg_timers() failed\n", in ufs_qcom_pwr_change_notify()
948 ret = -EINVAL; in ufs_qcom_pwr_change_notify()
952 memcpy(&host->dev_req_params, in ufs_qcom_pwr_change_notify()
958 if (ufshcd_is_hs_mode(&hba->pwr_info) && in ufs_qcom_pwr_change_notify()
963 ret = -EINVAL; in ufs_qcom_pwr_change_notify()
970 static int ufs_qcom_quirk_host_pa_saveconfigtime(struct ufs_hba *hba) in ufs_qcom_quirk_host_pa_saveconfigtime() argument
975 err = ufshcd_dme_get(hba, UIC_ARG_MIB(PA_VS_CONFIG_REG1), in ufs_qcom_quirk_host_pa_saveconfigtime()
981 return ufshcd_dme_set(hba, UIC_ARG_MIB(PA_VS_CONFIG_REG1), in ufs_qcom_quirk_host_pa_saveconfigtime()
985 static int ufs_qcom_apply_dev_quirks(struct ufs_hba *hba) in ufs_qcom_apply_dev_quirks() argument
989 if (hba->dev_quirks & UFS_DEVICE_QUIRK_HOST_PA_SAVECONFIGTIME) in ufs_qcom_apply_dev_quirks()
990 err = ufs_qcom_quirk_host_pa_saveconfigtime(hba); in ufs_qcom_apply_dev_quirks()
992 if (hba->dev_info.wmanufacturerid == UFS_VENDOR_WDC) in ufs_qcom_apply_dev_quirks()
993 hba->dev_quirks |= UFS_DEVICE_QUIRK_HOST_PA_TACTIVATE; in ufs_qcom_apply_dev_quirks()
998 static u32 ufs_qcom_get_ufs_hci_version(struct ufs_hba *hba) in ufs_qcom_get_ufs_hci_version() argument
1000 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_get_ufs_hci_version()
1002 if (host->hw_ver.major == 0x1) in ufs_qcom_get_ufs_hci_version()
1009 * ufs_qcom_advertise_quirks - advertise the known QCOM UFS controller quirks
1010 * @hba: host controller instance
1017 static void ufs_qcom_advertise_quirks(struct ufs_hba *hba) in ufs_qcom_advertise_quirks() argument
1019 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_advertise_quirks()
1021 if (host->hw_ver.major == 0x01) { in ufs_qcom_advertise_quirks()
1022 hba->quirks |= UFSHCD_QUIRK_DELAY_BEFORE_DME_CMDS in ufs_qcom_advertise_quirks()
1026 if (host->hw_ver.minor == 0x0001 && host->hw_ver.step == 0x0001) in ufs_qcom_advertise_quirks()
1027 hba->quirks |= UFSHCD_QUIRK_BROKEN_INTR_AGGR; in ufs_qcom_advertise_quirks()
1029 hba->quirks |= UFSHCD_QUIRK_BROKEN_LCC; in ufs_qcom_advertise_quirks()
1032 if (host->hw_ver.major == 0x2) { in ufs_qcom_advertise_quirks()
1033 hba->quirks |= UFSHCD_QUIRK_BROKEN_UFS_HCI_VERSION; in ufs_qcom_advertise_quirks()
1037 hba->quirks |= (UFSHCD_QUIRK_DELAY_BEFORE_DME_CMDS in ufs_qcom_advertise_quirks()
1042 if (host->hw_ver.major > 0x3) in ufs_qcom_advertise_quirks()
1043 hba->quirks |= UFSHCD_QUIRK_REINIT_AFTER_MAX_GEAR_SWITCH; in ufs_qcom_advertise_quirks()
1046 static void ufs_qcom_set_caps(struct ufs_hba *hba) in ufs_qcom_set_caps() argument
1048 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_set_caps()
1050 hba->caps |= UFSHCD_CAP_CLK_GATING | UFSHCD_CAP_HIBERN8_WITH_CLK_GATING; in ufs_qcom_set_caps()
1051 hba->caps |= UFSHCD_CAP_CLK_SCALING | UFSHCD_CAP_WB_WITH_CLK_SCALING; in ufs_qcom_set_caps()
1052 hba->caps |= UFSHCD_CAP_AUTO_BKOPS_SUSPEND; in ufs_qcom_set_caps()
1053 hba->caps |= UFSHCD_CAP_WB_EN; in ufs_qcom_set_caps()
1054 hba->caps |= UFSHCD_CAP_AGGR_POWER_COLLAPSE; in ufs_qcom_set_caps()
1055 hba->caps |= UFSHCD_CAP_RPM_AUTOSUSPEND; in ufs_qcom_set_caps()
1057 if (host->hw_ver.major >= 0x2) { in ufs_qcom_set_caps()
1058 host->caps = UFS_QCOM_CAP_QUNIPRO | in ufs_qcom_set_caps()
1064 * ufs_qcom_setup_clocks - enables/disable clocks
1065 * @hba: host controller instance
1069 * Return: 0 on success, non-zero on failure.
1071 static int ufs_qcom_setup_clocks(struct ufs_hba *hba, bool on, in ufs_qcom_setup_clocks() argument
1074 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_setup_clocks()
1089 if (!ufs_qcom_is_link_active(hba)) { in ufs_qcom_setup_clocks()
1098 if (ufshcd_is_hs_mode(&hba->pwr_info)) in ufs_qcom_setup_clocks()
1115 ufs_qcom_assert_reset(host->hba); in ufs_qcom_reset_assert()
1126 ufs_qcom_deassert_reset(host->hba); in ufs_qcom_reset_deassert()
1143 struct device *dev = host->hba->dev; in ufs_qcom_icc_init()
1146 host->icc_ddr = devm_of_icc_get(dev, "ufs-ddr"); in ufs_qcom_icc_init()
1147 if (IS_ERR(host->icc_ddr)) in ufs_qcom_icc_init()
1148 return dev_err_probe(dev, PTR_ERR(host->icc_ddr), in ufs_qcom_icc_init()
1151 host->icc_cpu = devm_of_icc_get(dev, "cpu-ufs"); in ufs_qcom_icc_init()
1152 if (IS_ERR(host->icc_cpu)) in ufs_qcom_icc_init()
1153 return dev_err_probe(dev, PTR_ERR(host->icc_cpu), in ufs_qcom_icc_init()
1170 * ufs_qcom_init - bind phy with controller
1171 * @hba: host controller instance
1176 * Return: -EPROBE_DEFER if binding fails, returns negative error
1179 static int ufs_qcom_init(struct ufs_hba *hba) in ufs_qcom_init() argument
1182 struct device *dev = hba->dev; in ufs_qcom_init()
1191 return -ENOMEM; in ufs_qcom_init()
1194 /* Make a two way bind between the qcom host and the hba */ in ufs_qcom_init()
1195 host->hba = hba; in ufs_qcom_init()
1196 ufshcd_set_variant(hba, host); in ufs_qcom_init()
1199 host->core_reset = devm_reset_control_get_optional(hba->dev, "rst"); in ufs_qcom_init()
1200 if (IS_ERR(host->core_reset)) { in ufs_qcom_init()
1201 err = dev_err_probe(dev, PTR_ERR(host->core_reset), in ufs_qcom_init()
1206 /* Fire up the reset controller. Failure here is non-fatal. */ in ufs_qcom_init()
1207 host->rcdev.of_node = dev->of_node; in ufs_qcom_init()
1208 host->rcdev.ops = &ufs_qcom_reset_ops; in ufs_qcom_init()
1209 host->rcdev.owner = dev->driver->owner; in ufs_qcom_init()
1210 host->rcdev.nr_resets = 1; in ufs_qcom_init()
1211 err = devm_reset_controller_register(dev, &host->rcdev); in ufs_qcom_init()
1216 host->generic_phy = devm_phy_get(dev, "ufsphy"); in ufs_qcom_init()
1217 if (IS_ERR(host->generic_phy)) { in ufs_qcom_init()
1218 err = dev_err_probe(dev, PTR_ERR(host->generic_phy), "Failed to get PHY\n"); in ufs_qcom_init()
1227 host->device_reset = devm_gpiod_get_optional(dev, "reset", in ufs_qcom_init()
1229 if (IS_ERR(host->device_reset)) { in ufs_qcom_init()
1230 err = PTR_ERR(host->device_reset); in ufs_qcom_init()
1231 if (err != -EPROBE_DEFER) in ufs_qcom_init()
1236 ufs_qcom_get_controller_revision(hba, &host->hw_ver.major, in ufs_qcom_init()
1237 &host->hw_ver.minor, &host->hw_ver.step); in ufs_qcom_init()
1243 if (host->hw_ver.major >= 0x02) { in ufs_qcom_init()
1244 host->dev_ref_clk_ctrl_mmio = hba->mmio_base + REG_UFS_CFG1; in ufs_qcom_init()
1245 host->dev_ref_clk_en_mask = BIT(26); in ufs_qcom_init()
1251 host->dev_ref_clk_ctrl_mmio = in ufs_qcom_init()
1253 if (IS_ERR(host->dev_ref_clk_ctrl_mmio)) in ufs_qcom_init()
1254 host->dev_ref_clk_ctrl_mmio = NULL; in ufs_qcom_init()
1255 host->dev_ref_clk_en_mask = BIT(5); in ufs_qcom_init()
1259 list_for_each_entry(clki, &hba->clk_list_head, list) { in ufs_qcom_init()
1260 if (!strcmp(clki->name, "core_clk_unipro")) in ufs_qcom_init()
1261 clki->keep_link_active = true; in ufs_qcom_init()
1268 ufs_qcom_set_caps(hba); in ufs_qcom_init()
1269 ufs_qcom_advertise_quirks(hba); in ufs_qcom_init()
1275 ufs_qcom_setup_clocks(hba, true, POST_CHANGE); in ufs_qcom_init()
1277 if (hba->dev->id < MAX_UFS_QCOM_HOSTS) in ufs_qcom_init()
1278 ufs_qcom_hosts[hba->dev->id] = host; in ufs_qcom_init()
1283 /* Failure is non-fatal */ in ufs_qcom_init()
1291 host->hs_gear = UFS_HS_G2; in ufs_qcom_init()
1296 ufshcd_set_variant(hba, NULL); in ufs_qcom_init()
1301 static void ufs_qcom_exit(struct ufs_hba *hba) in ufs_qcom_exit() argument
1303 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_exit()
1306 phy_power_off(host->generic_phy); in ufs_qcom_exit()
1307 phy_exit(host->generic_phy); in ufs_qcom_exit()
1310 static int ufs_qcom_set_dme_vs_core_clk_ctrl_clear_div(struct ufs_hba *hba, in ufs_qcom_set_dme_vs_core_clk_ctrl_clear_div() argument
1317 return -EINVAL; in ufs_qcom_set_dme_vs_core_clk_ctrl_clear_div()
1319 err = ufshcd_dme_get(hba, in ufs_qcom_set_dme_vs_core_clk_ctrl_clear_div()
1331 return ufshcd_dme_set(hba, in ufs_qcom_set_dme_vs_core_clk_ctrl_clear_div()
1336 static int ufs_qcom_clk_scale_up_pre_change(struct ufs_hba *hba) in ufs_qcom_clk_scale_up_pre_change() argument
1342 static int ufs_qcom_clk_scale_up_post_change(struct ufs_hba *hba) in ufs_qcom_clk_scale_up_post_change() argument
1344 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_clk_scale_up_post_change()
1350 return ufs_qcom_set_dme_vs_core_clk_ctrl_clear_div(hba, 150); in ufs_qcom_clk_scale_up_post_change()
1353 static int ufs_qcom_clk_scale_down_pre_change(struct ufs_hba *hba) in ufs_qcom_clk_scale_down_pre_change() argument
1355 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_clk_scale_down_pre_change()
1362 err = ufshcd_dme_get(hba, in ufs_qcom_clk_scale_down_pre_change()
1370 err = ufshcd_dme_set(hba, in ufs_qcom_clk_scale_down_pre_change()
1378 static int ufs_qcom_clk_scale_down_post_change(struct ufs_hba *hba) in ufs_qcom_clk_scale_down_post_change() argument
1380 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_clk_scale_down_post_change()
1386 return ufs_qcom_set_dme_vs_core_clk_ctrl_clear_div(hba, 75); in ufs_qcom_clk_scale_down_post_change()
1389 static int ufs_qcom_clk_scale_notify(struct ufs_hba *hba, in ufs_qcom_clk_scale_notify() argument
1392 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_clk_scale_notify()
1393 struct ufs_pa_layer_attr *dev_req_params = &host->dev_req_params; in ufs_qcom_clk_scale_notify()
1397 if (!ufshcd_is_hba_active(hba)) in ufs_qcom_clk_scale_notify()
1401 err = ufshcd_uic_hibern8_enter(hba); in ufs_qcom_clk_scale_notify()
1405 err = ufs_qcom_clk_scale_up_pre_change(hba); in ufs_qcom_clk_scale_notify()
1407 err = ufs_qcom_clk_scale_down_pre_change(hba); in ufs_qcom_clk_scale_notify()
1410 ufshcd_uic_hibern8_exit(hba); in ufs_qcom_clk_scale_notify()
1415 err = ufs_qcom_clk_scale_up_post_change(hba); in ufs_qcom_clk_scale_notify()
1417 err = ufs_qcom_clk_scale_down_post_change(hba); in ufs_qcom_clk_scale_notify()
1421 ufshcd_uic_hibern8_exit(hba); in ufs_qcom_clk_scale_notify()
1425 ufs_qcom_cfg_timers(hba, in ufs_qcom_clk_scale_notify()
1426 dev_req_params->gear_rx, in ufs_qcom_clk_scale_notify()
1427 dev_req_params->pwr_rx, in ufs_qcom_clk_scale_notify()
1428 dev_req_params->hs_rate, in ufs_qcom_clk_scale_notify()
1431 ufshcd_uic_hibern8_exit(hba); in ufs_qcom_clk_scale_notify()
1439 ufshcd_rmwl(host->hba, UFS_REG_TEST_BUS_EN, in ufs_qcom_enable_test_bus()
1441 ufshcd_rmwl(host->hba, TEST_BUS_EN, TEST_BUS_EN, REG_UFS_CFG1); in ufs_qcom_enable_test_bus()
1447 host->testbus.select_major = TSTBUS_UNIPRO; in ufs_qcom_get_default_testbus_cfg()
1448 host->testbus.select_minor = 37; in ufs_qcom_get_default_testbus_cfg()
1453 if (host->testbus.select_major >= TSTBUS_MAX) { in ufs_qcom_testbus_cfg_is_ok()
1454 dev_err(host->hba->dev, in ufs_qcom_testbus_cfg_is_ok()
1456 __func__, host->testbus.select_major); in ufs_qcom_testbus_cfg_is_ok()
1470 return -EINVAL; in ufs_qcom_testbus_config()
1473 return -EPERM; in ufs_qcom_testbus_config()
1475 switch (host->testbus.select_major) { in ufs_qcom_testbus_config()
1532 ufshcd_rmwl(host->hba, TEST_BUS_SEL, in ufs_qcom_testbus_config()
1533 (u32)host->testbus.select_major << 19, in ufs_qcom_testbus_config()
1535 ufshcd_rmwl(host->hba, mask, in ufs_qcom_testbus_config()
1536 (u32)host->testbus.select_minor << offset, in ufs_qcom_testbus_config()
1548 static void ufs_qcom_dump_dbg_regs(struct ufs_hba *hba) in ufs_qcom_dump_dbg_regs() argument
1553 host = ufshcd_get_variant(hba); in ufs_qcom_dump_dbg_regs()
1555 ufshcd_dump_regs(hba, REG_UFS_SYS1CLK_1US, 16 * 4, in ufs_qcom_dump_dbg_regs()
1559 ufshcd_dump_regs(hba, reg, 44 * 4, "UFS_UFS_DBG_RD_REG_OCSC "); in ufs_qcom_dump_dbg_regs()
1561 reg = ufshcd_readl(hba, REG_UFS_CFG1); in ufs_qcom_dump_dbg_regs()
1563 ufshcd_writel(hba, reg, REG_UFS_CFG1); in ufs_qcom_dump_dbg_regs()
1566 ufshcd_dump_regs(hba, reg, 32 * 4, "UFS_UFS_DBG_RD_EDTL_RAM "); in ufs_qcom_dump_dbg_regs()
1569 ufshcd_dump_regs(hba, reg, 128 * 4, "UFS_UFS_DBG_RD_DESC_RAM "); in ufs_qcom_dump_dbg_regs()
1572 ufshcd_dump_regs(hba, reg, 64 * 4, "UFS_UFS_DBG_RD_PRDT_RAM "); in ufs_qcom_dump_dbg_regs()
1574 /* clear bit 17 - UTP_DBG_RAMS_EN */ in ufs_qcom_dump_dbg_regs()
1575 ufshcd_rmwl(hba, UTP_DBG_RAMS_EN, 0, REG_UFS_CFG1); in ufs_qcom_dump_dbg_regs()
1578 ufshcd_dump_regs(hba, reg, 4 * 4, "UFS_DBG_RD_REG_UAWM "); in ufs_qcom_dump_dbg_regs()
1581 ufshcd_dump_regs(hba, reg, 4 * 4, "UFS_DBG_RD_REG_UARM "); in ufs_qcom_dump_dbg_regs()
1584 ufshcd_dump_regs(hba, reg, 48 * 4, "UFS_DBG_RD_REG_TXUC "); in ufs_qcom_dump_dbg_regs()
1587 ufshcd_dump_regs(hba, reg, 27 * 4, "UFS_DBG_RD_REG_RXUC "); in ufs_qcom_dump_dbg_regs()
1590 ufshcd_dump_regs(hba, reg, 19 * 4, "UFS_DBG_RD_REG_DFC "); in ufs_qcom_dump_dbg_regs()
1593 ufshcd_dump_regs(hba, reg, 34 * 4, "UFS_DBG_RD_REG_TRLUT "); in ufs_qcom_dump_dbg_regs()
1596 ufshcd_dump_regs(hba, reg, 9 * 4, "UFS_DBG_RD_REG_TMRLUT "); in ufs_qcom_dump_dbg_regs()
1600 * ufs_qcom_device_reset() - toggle the (optional) device reset line
1601 * @hba: per-adapter instance
1605 static int ufs_qcom_device_reset(struct ufs_hba *hba) in ufs_qcom_device_reset() argument
1607 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_device_reset()
1610 if (!host->device_reset) in ufs_qcom_device_reset()
1611 return -EOPNOTSUPP; in ufs_qcom_device_reset()
1617 ufs_qcom_device_reset_ctrl(hba, true); in ufs_qcom_device_reset()
1620 ufs_qcom_device_reset_ctrl(hba, false); in ufs_qcom_device_reset()
1627 static void ufs_qcom_config_scaling_param(struct ufs_hba *hba, in ufs_qcom_config_scaling_param() argument
1631 p->polling_ms = 60; in ufs_qcom_config_scaling_param()
1632 p->timer = DEVFREQ_TIMER_DELAYED; in ufs_qcom_config_scaling_param()
1633 d->upthreshold = 70; in ufs_qcom_config_scaling_param()
1634 d->downdifferential = 5; in ufs_qcom_config_scaling_param()
1637 static void ufs_qcom_config_scaling_param(struct ufs_hba *hba, in ufs_qcom_config_scaling_param() argument
1660 static int ufs_qcom_mcq_config_resource(struct ufs_hba *hba) in ufs_qcom_mcq_config_resource() argument
1662 struct platform_device *pdev = to_platform_device(hba->dev); in ufs_qcom_mcq_config_resource()
1667 memcpy(hba->res, ufs_res_info, sizeof(ufs_res_info)); in ufs_qcom_mcq_config_resource()
1670 res = &hba->res[i]; in ufs_qcom_mcq_config_resource()
1671 res->resource = platform_get_resource_byname(pdev, in ufs_qcom_mcq_config_resource()
1673 res->name); in ufs_qcom_mcq_config_resource()
1674 if (!res->resource) { in ufs_qcom_mcq_config_resource()
1675 dev_info(hba->dev, "Resource %s not provided\n", res->name); in ufs_qcom_mcq_config_resource()
1677 return -ENODEV; in ufs_qcom_mcq_config_resource()
1680 res_mem = res->resource; in ufs_qcom_mcq_config_resource()
1681 res->base = hba->mmio_base; in ufs_qcom_mcq_config_resource()
1685 res->base = devm_ioremap_resource(hba->dev, res->resource); in ufs_qcom_mcq_config_resource()
1686 if (IS_ERR(res->base)) { in ufs_qcom_mcq_config_resource()
1687 dev_err(hba->dev, "Failed to map res %s, err=%d\n", in ufs_qcom_mcq_config_resource()
1688 res->name, (int)PTR_ERR(res->base)); in ufs_qcom_mcq_config_resource()
1689 ret = PTR_ERR(res->base); in ufs_qcom_mcq_config_resource()
1690 res->base = NULL; in ufs_qcom_mcq_config_resource()
1696 res = &hba->res[RES_MCQ]; in ufs_qcom_mcq_config_resource()
1698 if (res->base) in ufs_qcom_mcq_config_resource()
1702 res_mcq = devm_kzalloc(hba->dev, sizeof(*res_mcq), GFP_KERNEL); in ufs_qcom_mcq_config_resource()
1704 return -ENOMEM; in ufs_qcom_mcq_config_resource()
1706 res_mcq->start = res_mem->start + in ufs_qcom_mcq_config_resource()
1707 MCQ_SQATTR_OFFSET(hba->mcq_capabilities); in ufs_qcom_mcq_config_resource()
1708 res_mcq->end = res_mcq->start + hba->nr_hw_queues * MCQ_QCFG_SIZE - 1; in ufs_qcom_mcq_config_resource()
1709 res_mcq->flags = res_mem->flags; in ufs_qcom_mcq_config_resource()
1710 res_mcq->name = "mcq"; in ufs_qcom_mcq_config_resource()
1714 dev_err(hba->dev, "Failed to insert MCQ resource, err=%d\n", in ufs_qcom_mcq_config_resource()
1719 res->base = devm_ioremap_resource(hba->dev, res_mcq); in ufs_qcom_mcq_config_resource()
1720 if (IS_ERR(res->base)) { in ufs_qcom_mcq_config_resource()
1721 dev_err(hba->dev, "MCQ registers mapping failed, err=%d\n", in ufs_qcom_mcq_config_resource()
1722 (int)PTR_ERR(res->base)); in ufs_qcom_mcq_config_resource()
1723 ret = PTR_ERR(res->base); in ufs_qcom_mcq_config_resource()
1728 hba->mcq_base = res->base; in ufs_qcom_mcq_config_resource()
1731 res->base = NULL; in ufs_qcom_mcq_config_resource()
1736 static int ufs_qcom_op_runtime_config(struct ufs_hba *hba) in ufs_qcom_op_runtime_config() argument
1742 mem_res = &hba->res[RES_UFS]; in ufs_qcom_op_runtime_config()
1743 sqdao_res = &hba->res[RES_MCQ_SQD]; in ufs_qcom_op_runtime_config()
1745 if (!mem_res->base || !sqdao_res->base) in ufs_qcom_op_runtime_config()
1746 return -EINVAL; in ufs_qcom_op_runtime_config()
1749 opr = &hba->mcq_opr[i]; in ufs_qcom_op_runtime_config()
1750 opr->offset = sqdao_res->resource->start - in ufs_qcom_op_runtime_config()
1751 mem_res->resource->start + 0x40 * i; in ufs_qcom_op_runtime_config()
1752 opr->stride = 0x100; in ufs_qcom_op_runtime_config()
1753 opr->base = sqdao_res->base + 0x40 * i; in ufs_qcom_op_runtime_config()
1759 static int ufs_qcom_get_hba_mac(struct ufs_hba *hba) in ufs_qcom_get_hba_mac() argument
1765 static int ufs_qcom_get_outstanding_cqs(struct ufs_hba *hba, in ufs_qcom_get_outstanding_cqs() argument
1768 struct ufshcd_res_info *mcq_vs_res = &hba->res[RES_MCQ_VS]; in ufs_qcom_get_outstanding_cqs()
1770 if (!mcq_vs_res->base) in ufs_qcom_get_outstanding_cqs()
1771 return -EINVAL; in ufs_qcom_get_outstanding_cqs()
1773 *ocqs = readl(mcq_vs_res->base + UFS_MEM_CQIS_VS); in ufs_qcom_get_outstanding_cqs()
1781 struct ufs_hba *hba = dev_get_drvdata(dev); in ufs_qcom_write_msi_msg() local
1783 ufshcd_mcq_config_esi(hba, msg); in ufs_qcom_write_msi_msg()
1790 struct ufs_hba *hba = dev_get_drvdata(dev); in ufs_qcom_mcq_esi_handler() local
1791 u32 id = desc->msi_index; in ufs_qcom_mcq_esi_handler()
1792 struct ufs_hw_queue *hwq = &hba->uhq[id]; in ufs_qcom_mcq_esi_handler()
1794 ufshcd_mcq_write_cqis(hba, 0x1, id); in ufs_qcom_mcq_esi_handler()
1795 ufshcd_mcq_poll_cqe_lock(hba, hwq); in ufs_qcom_mcq_esi_handler()
1800 static int ufs_qcom_config_esi(struct ufs_hba *hba) in ufs_qcom_config_esi() argument
1802 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_config_esi()
1807 if (host->esi_enabled) in ufs_qcom_config_esi()
1814 nr_irqs = hba->nr_hw_queues - hba->nr_queues[HCTX_TYPE_POLL]; in ufs_qcom_config_esi()
1815 ret = platform_msi_domain_alloc_irqs(hba->dev, nr_irqs, in ufs_qcom_config_esi()
1818 dev_err(hba->dev, "Failed to request Platform MSI %d\n", ret); in ufs_qcom_config_esi()
1822 msi_lock_descs(hba->dev); in ufs_qcom_config_esi()
1823 msi_for_each_desc(desc, hba->dev, MSI_DESC_ALL) { in ufs_qcom_config_esi()
1824 ret = devm_request_irq(hba->dev, desc->irq, in ufs_qcom_config_esi()
1826 IRQF_SHARED, "qcom-mcq-esi", desc); in ufs_qcom_config_esi()
1828 dev_err(hba->dev, "%s: Fail to request IRQ for %d, err = %d\n", in ufs_qcom_config_esi()
1829 __func__, desc->irq, ret); in ufs_qcom_config_esi()
1834 msi_unlock_descs(hba->dev); in ufs_qcom_config_esi()
1838 msi_lock_descs(hba->dev); in ufs_qcom_config_esi()
1839 msi_for_each_desc(desc, hba->dev, MSI_DESC_ALL) { in ufs_qcom_config_esi()
1842 devm_free_irq(hba->dev, desc->irq, hba); in ufs_qcom_config_esi()
1844 msi_unlock_descs(hba->dev); in ufs_qcom_config_esi()
1845 platform_msi_domain_free_irqs(hba->dev); in ufs_qcom_config_esi()
1847 if (host->hw_ver.major == 6 && host->hw_ver.minor == 0 && in ufs_qcom_config_esi()
1848 host->hw_ver.step == 0) { in ufs_qcom_config_esi()
1849 ufshcd_writel(hba, in ufs_qcom_config_esi()
1850 ufshcd_readl(hba, REG_UFS_CFG3) | 0x1F000, in ufs_qcom_config_esi()
1853 ufshcd_mcq_enable_esi(hba); in ufs_qcom_config_esi()
1858 host->esi_enabled = true; in ufs_qcom_config_esi()
1864 * struct ufs_hba_qcom_vops - UFS QCOM specific variant operations
1894 * ufs_qcom_probe - probe routine of the driver
1897 * Return: zero for success and non-zero for failure.
1902 struct device *dev = &pdev->dev; in ufs_qcom_probe()
1913 * ufs_qcom_remove - set driver_data of the device to NULL
1920 struct ufs_hba *hba = platform_get_drvdata(pdev); in ufs_qcom_remove() local
1922 pm_runtime_get_sync(&(pdev)->dev); in ufs_qcom_remove()
1923 ufshcd_remove(hba); in ufs_qcom_remove()
1924 platform_msi_domain_free_irqs(hba->dev); in ufs_qcom_remove()
1959 .name = "ufshcd-qcom",