xref: /openbmc/linux/drivers/ufs/core/ufshcd-priv.h (revision f762326b)
1 /* SPDX-License-Identifier: GPL-2.0-or-later */
2 
3 #ifndef _UFSHCD_PRIV_H_
4 #define _UFSHCD_PRIV_H_
5 
6 #include <linux/pm_runtime.h>
7 #include <ufs/ufshcd.h>
8 
9 static inline bool ufshcd_is_user_access_allowed(struct ufs_hba *hba)
10 {
11 	return !hba->shutting_down;
12 }
13 
14 void ufshcd_schedule_eh_work(struct ufs_hba *hba);
15 
16 static inline bool ufshcd_keep_autobkops_enabled_except_suspend(
17 							struct ufs_hba *hba)
18 {
19 	return hba->caps & UFSHCD_CAP_KEEP_AUTO_BKOPS_ENABLED_EXCEPT_SUSPEND;
20 }
21 
22 static inline u8 ufshcd_wb_get_query_index(struct ufs_hba *hba)
23 {
24 	if (hba->dev_info.wb_buffer_type == WB_BUF_MODE_LU_DEDICATED)
25 		return hba->dev_info.wb_dedicated_lu;
26 	return 0;
27 }
28 
29 static inline bool ufshcd_is_wb_buf_flush_allowed(struct ufs_hba *hba)
30 {
31 	return ufshcd_is_wb_allowed(hba) &&
32 		!(hba->quirks & UFSHCI_QUIRK_SKIP_MANUAL_WB_FLUSH_CTRL);
33 }
34 
35 #ifdef CONFIG_SCSI_UFS_HWMON
36 void ufs_hwmon_probe(struct ufs_hba *hba, u8 mask);
37 void ufs_hwmon_remove(struct ufs_hba *hba);
38 void ufs_hwmon_notify_event(struct ufs_hba *hba, u8 ee_mask);
39 #else
40 static inline void ufs_hwmon_probe(struct ufs_hba *hba, u8 mask) {}
41 static inline void ufs_hwmon_remove(struct ufs_hba *hba) {}
42 static inline void ufs_hwmon_notify_event(struct ufs_hba *hba, u8 ee_mask) {}
43 #endif
44 
45 int ufshcd_query_descriptor_retry(struct ufs_hba *hba,
46 				  enum query_opcode opcode,
47 				  enum desc_idn idn, u8 index,
48 				  u8 selector,
49 				  u8 *desc_buf, int *buf_len);
50 int ufshcd_read_desc_param(struct ufs_hba *hba,
51 			   enum desc_idn desc_id,
52 			   int desc_index,
53 			   u8 param_offset,
54 			   u8 *param_read_buf,
55 			   u8 param_size);
56 int ufshcd_query_attr_retry(struct ufs_hba *hba, enum query_opcode opcode,
57 			    enum attr_idn idn, u8 index, u8 selector,
58 			    u32 *attr_val);
59 int ufshcd_query_attr(struct ufs_hba *hba, enum query_opcode opcode,
60 		      enum attr_idn idn, u8 index, u8 selector, u32 *attr_val);
61 int ufshcd_query_flag(struct ufs_hba *hba, enum query_opcode opcode,
62 	enum flag_idn idn, u8 index, bool *flag_res);
63 void ufshcd_auto_hibern8_update(struct ufs_hba *hba, u32 ahit);
64 void ufshcd_compl_one_cqe(struct ufs_hba *hba, int task_tag,
65 			  struct cq_entry *cqe);
66 int ufshcd_mcq_init(struct ufs_hba *hba);
67 int ufshcd_mcq_decide_queue_depth(struct ufs_hba *hba);
68 int ufshcd_mcq_memory_alloc(struct ufs_hba *hba);
69 void ufshcd_mcq_make_queues_operational(struct ufs_hba *hba);
70 void ufshcd_mcq_config_mac(struct ufs_hba *hba, u32 max_active_cmds);
71 void ufshcd_mcq_select_mcq_mode(struct ufs_hba *hba);
72 u32 ufshcd_mcq_read_cqis(struct ufs_hba *hba, int i);
73 void ufshcd_mcq_write_cqis(struct ufs_hba *hba, u32 val, int i);
74 struct ufs_hw_queue *ufshcd_mcq_req_to_hwq(struct ufs_hba *hba,
75 					   struct request *req);
76 unsigned long ufshcd_mcq_poll_cqe_lock(struct ufs_hba *hba,
77 				       struct ufs_hw_queue *hwq);
78 void ufshcd_mcq_compl_all_cqes_lock(struct ufs_hba *hba,
79 				    struct ufs_hw_queue *hwq);
80 bool ufshcd_cmd_inflight(struct scsi_cmnd *cmd);
81 int ufshcd_mcq_sq_cleanup(struct ufs_hba *hba, int task_tag);
82 int ufshcd_mcq_abort(struct scsi_cmnd *cmd);
83 int ufshcd_try_to_abort_task(struct ufs_hba *hba, int tag);
84 void ufshcd_release_scsi_cmd(struct ufs_hba *hba,
85 			     struct ufshcd_lrb *lrbp);
86 
87 #define UFSHCD_MCQ_IO_QUEUE_OFFSET	1
88 #define SD_ASCII_STD true
89 #define SD_RAW false
90 int ufshcd_read_string_desc(struct ufs_hba *hba, u8 desc_index,
91 			    u8 **buf, bool ascii);
92 
93 int ufshcd_send_uic_cmd(struct ufs_hba *hba, struct uic_command *uic_cmd);
94 
95 int ufshcd_exec_raw_upiu_cmd(struct ufs_hba *hba,
96 			     struct utp_upiu_req *req_upiu,
97 			     struct utp_upiu_req *rsp_upiu,
98 			     int msgcode,
99 			     u8 *desc_buff, int *buff_len,
100 			     enum query_opcode desc_op);
101 
102 int ufshcd_wb_toggle(struct ufs_hba *hba, bool enable);
103 
104 /* Wrapper functions for safely calling variant operations */
105 static inline const char *ufshcd_get_var_name(struct ufs_hba *hba)
106 {
107 	if (hba->vops)
108 		return hba->vops->name;
109 	return "";
110 }
111 
112 static inline void ufshcd_vops_exit(struct ufs_hba *hba)
113 {
114 	if (hba->vops && hba->vops->exit)
115 		return hba->vops->exit(hba);
116 }
117 
118 static inline u32 ufshcd_vops_get_ufs_hci_version(struct ufs_hba *hba)
119 {
120 	if (hba->vops && hba->vops->get_ufs_hci_version)
121 		return hba->vops->get_ufs_hci_version(hba);
122 
123 	return ufshcd_readl(hba, REG_UFS_VERSION);
124 }
125 
126 static inline int ufshcd_vops_clk_scale_notify(struct ufs_hba *hba,
127 			bool up, enum ufs_notify_change_status status)
128 {
129 	if (hba->vops && hba->vops->clk_scale_notify)
130 		return hba->vops->clk_scale_notify(hba, up, status);
131 	return 0;
132 }
133 
134 static inline void ufshcd_vops_event_notify(struct ufs_hba *hba,
135 					    enum ufs_event_type evt,
136 					    void *data)
137 {
138 	if (hba->vops && hba->vops->event_notify)
139 		hba->vops->event_notify(hba, evt, data);
140 }
141 
142 static inline int ufshcd_vops_setup_clocks(struct ufs_hba *hba, bool on,
143 					enum ufs_notify_change_status status)
144 {
145 	if (hba->vops && hba->vops->setup_clocks)
146 		return hba->vops->setup_clocks(hba, on, status);
147 	return 0;
148 }
149 
150 static inline int ufshcd_vops_hce_enable_notify(struct ufs_hba *hba,
151 						bool status)
152 {
153 	if (hba->vops && hba->vops->hce_enable_notify)
154 		return hba->vops->hce_enable_notify(hba, status);
155 
156 	return 0;
157 }
158 static inline int ufshcd_vops_link_startup_notify(struct ufs_hba *hba,
159 						bool status)
160 {
161 	if (hba->vops && hba->vops->link_startup_notify)
162 		return hba->vops->link_startup_notify(hba, status);
163 
164 	return 0;
165 }
166 
167 static inline int ufshcd_vops_pwr_change_notify(struct ufs_hba *hba,
168 				  enum ufs_notify_change_status status,
169 				  struct ufs_pa_layer_attr *dev_max_params,
170 				  struct ufs_pa_layer_attr *dev_req_params)
171 {
172 	if (hba->vops && hba->vops->pwr_change_notify)
173 		return hba->vops->pwr_change_notify(hba, status,
174 					dev_max_params, dev_req_params);
175 
176 	return -ENOTSUPP;
177 }
178 
179 static inline void ufshcd_vops_setup_task_mgmt(struct ufs_hba *hba,
180 					int tag, u8 tm_function)
181 {
182 	if (hba->vops && hba->vops->setup_task_mgmt)
183 		return hba->vops->setup_task_mgmt(hba, tag, tm_function);
184 }
185 
186 static inline void ufshcd_vops_hibern8_notify(struct ufs_hba *hba,
187 					enum uic_cmd_dme cmd,
188 					enum ufs_notify_change_status status)
189 {
190 	if (hba->vops && hba->vops->hibern8_notify)
191 		return hba->vops->hibern8_notify(hba, cmd, status);
192 }
193 
194 static inline int ufshcd_vops_apply_dev_quirks(struct ufs_hba *hba)
195 {
196 	if (hba->vops && hba->vops->apply_dev_quirks)
197 		return hba->vops->apply_dev_quirks(hba);
198 	return 0;
199 }
200 
201 static inline void ufshcd_vops_fixup_dev_quirks(struct ufs_hba *hba)
202 {
203 	if (hba->vops && hba->vops->fixup_dev_quirks)
204 		hba->vops->fixup_dev_quirks(hba);
205 }
206 
207 static inline int ufshcd_vops_suspend(struct ufs_hba *hba, enum ufs_pm_op op,
208 				enum ufs_notify_change_status status)
209 {
210 	if (hba->vops && hba->vops->suspend)
211 		return hba->vops->suspend(hba, op, status);
212 
213 	return 0;
214 }
215 
216 static inline int ufshcd_vops_resume(struct ufs_hba *hba, enum ufs_pm_op op)
217 {
218 	if (hba->vops && hba->vops->resume)
219 		return hba->vops->resume(hba, op);
220 
221 	return 0;
222 }
223 
224 static inline void ufshcd_vops_dbg_register_dump(struct ufs_hba *hba)
225 {
226 	if (hba->vops && hba->vops->dbg_register_dump)
227 		hba->vops->dbg_register_dump(hba);
228 }
229 
230 static inline int ufshcd_vops_device_reset(struct ufs_hba *hba)
231 {
232 	if (hba->vops && hba->vops->device_reset)
233 		return hba->vops->device_reset(hba);
234 
235 	return -EOPNOTSUPP;
236 }
237 
238 static inline void ufshcd_vops_config_scaling_param(struct ufs_hba *hba,
239 		struct devfreq_dev_profile *p,
240 		struct devfreq_simple_ondemand_data *data)
241 {
242 	if (hba->vops && hba->vops->config_scaling_param)
243 		hba->vops->config_scaling_param(hba, p, data);
244 }
245 
246 static inline void ufshcd_vops_reinit_notify(struct ufs_hba *hba)
247 {
248 	if (hba->vops && hba->vops->reinit_notify)
249 		hba->vops->reinit_notify(hba);
250 }
251 
252 static inline int ufshcd_vops_mcq_config_resource(struct ufs_hba *hba)
253 {
254 	if (hba->vops && hba->vops->mcq_config_resource)
255 		return hba->vops->mcq_config_resource(hba);
256 
257 	return -EOPNOTSUPP;
258 }
259 
260 static inline int ufshcd_mcq_vops_get_hba_mac(struct ufs_hba *hba)
261 {
262 	if (hba->vops && hba->vops->get_hba_mac)
263 		return hba->vops->get_hba_mac(hba);
264 
265 	return -EOPNOTSUPP;
266 }
267 
268 static inline int ufshcd_mcq_vops_op_runtime_config(struct ufs_hba *hba)
269 {
270 	if (hba->vops && hba->vops->op_runtime_config)
271 		return hba->vops->op_runtime_config(hba);
272 
273 	return -EOPNOTSUPP;
274 }
275 
276 static inline int ufshcd_vops_get_outstanding_cqs(struct ufs_hba *hba,
277 						  unsigned long *ocqs)
278 {
279 	if (hba->vops && hba->vops->get_outstanding_cqs)
280 		return hba->vops->get_outstanding_cqs(hba, ocqs);
281 
282 	return -EOPNOTSUPP;
283 }
284 
285 static inline int ufshcd_mcq_vops_config_esi(struct ufs_hba *hba)
286 {
287 	if (hba->vops && hba->vops->config_esi)
288 		return hba->vops->config_esi(hba);
289 
290 	return -EOPNOTSUPP;
291 }
292 
293 extern const struct ufs_pm_lvl_states ufs_pm_lvl_states[];
294 
295 /**
296  * ufshcd_scsi_to_upiu_lun - maps scsi LUN to UPIU LUN
297  * @scsi_lun: scsi LUN id
298  *
299  * Returns UPIU LUN id
300  */
301 static inline u8 ufshcd_scsi_to_upiu_lun(unsigned int scsi_lun)
302 {
303 	if (scsi_is_wlun(scsi_lun))
304 		return (scsi_lun & UFS_UPIU_MAX_UNIT_NUM_ID)
305 			| UFS_UPIU_WLUN_ID;
306 	else
307 		return scsi_lun & UFS_UPIU_MAX_UNIT_NUM_ID;
308 }
309 
310 int __ufshcd_write_ee_control(struct ufs_hba *hba, u32 ee_ctrl_mask);
311 int ufshcd_write_ee_control(struct ufs_hba *hba);
312 int ufshcd_update_ee_control(struct ufs_hba *hba, u16 *mask,
313 			     const u16 *other_mask, u16 set, u16 clr);
314 
315 static inline int ufshcd_update_ee_drv_mask(struct ufs_hba *hba,
316 					    u16 set, u16 clr)
317 {
318 	return ufshcd_update_ee_control(hba, &hba->ee_drv_mask,
319 					&hba->ee_usr_mask, set, clr);
320 }
321 
322 static inline int ufshcd_update_ee_usr_mask(struct ufs_hba *hba,
323 					    u16 set, u16 clr)
324 {
325 	return ufshcd_update_ee_control(hba, &hba->ee_usr_mask,
326 					&hba->ee_drv_mask, set, clr);
327 }
328 
329 static inline int ufshcd_rpm_get_sync(struct ufs_hba *hba)
330 {
331 	return pm_runtime_get_sync(&hba->ufs_device_wlun->sdev_gendev);
332 }
333 
334 static inline int ufshcd_rpm_put_sync(struct ufs_hba *hba)
335 {
336 	return pm_runtime_put_sync(&hba->ufs_device_wlun->sdev_gendev);
337 }
338 
339 static inline void ufshcd_rpm_get_noresume(struct ufs_hba *hba)
340 {
341 	pm_runtime_get_noresume(&hba->ufs_device_wlun->sdev_gendev);
342 }
343 
344 static inline int ufshcd_rpm_resume(struct ufs_hba *hba)
345 {
346 	return pm_runtime_resume(&hba->ufs_device_wlun->sdev_gendev);
347 }
348 
349 static inline int ufshcd_rpm_put(struct ufs_hba *hba)
350 {
351 	return pm_runtime_put(&hba->ufs_device_wlun->sdev_gendev);
352 }
353 
354 /**
355  * ufs_is_valid_unit_desc_lun - checks if the given LUN has a unit descriptor
356  * @dev_info: pointer of instance of struct ufs_dev_info
357  * @lun: LU number to check
358  * @return: true if the lun has a matching unit descriptor, false otherwise
359  */
360 static inline bool ufs_is_valid_unit_desc_lun(struct ufs_dev_info *dev_info, u8 lun)
361 {
362 	if (!dev_info || !dev_info->max_lu_supported) {
363 		pr_err("Max General LU supported by UFS isn't initialized\n");
364 		return false;
365 	}
366 	return lun == UFS_UPIU_RPMB_WLUN || (lun < dev_info->max_lu_supported);
367 }
368 
369 static inline void ufshcd_inc_sq_tail(struct ufs_hw_queue *q)
370 	__must_hold(&q->sq_lock)
371 {
372 	u32 mask = q->max_entries - 1;
373 	u32 val;
374 
375 	q->sq_tail_slot = (q->sq_tail_slot + 1) & mask;
376 	val = q->sq_tail_slot * sizeof(struct utp_transfer_req_desc);
377 	writel(val, q->mcq_sq_tail);
378 }
379 
380 static inline void ufshcd_mcq_update_cq_tail_slot(struct ufs_hw_queue *q)
381 {
382 	u32 val = readl(q->mcq_cq_tail);
383 
384 	q->cq_tail_slot = val / sizeof(struct cq_entry);
385 }
386 
387 static inline bool ufshcd_mcq_is_cq_empty(struct ufs_hw_queue *q)
388 {
389 	return q->cq_head_slot == q->cq_tail_slot;
390 }
391 
392 static inline void ufshcd_mcq_inc_cq_head_slot(struct ufs_hw_queue *q)
393 {
394 	q->cq_head_slot++;
395 	if (q->cq_head_slot == q->max_entries)
396 		q->cq_head_slot = 0;
397 }
398 
399 static inline void ufshcd_mcq_update_cq_head(struct ufs_hw_queue *q)
400 {
401 	writel(q->cq_head_slot * sizeof(struct cq_entry), q->mcq_cq_head);
402 }
403 
404 static inline struct cq_entry *ufshcd_mcq_cur_cqe(struct ufs_hw_queue *q)
405 {
406 	struct cq_entry *cqe = q->cqe_base_addr;
407 
408 	return cqe + q->cq_head_slot;
409 }
410 
411 static inline u32 ufshcd_mcq_get_sq_head_slot(struct ufs_hw_queue *q)
412 {
413 	u32 val = readl(q->mcq_sq_head);
414 
415 	return val / sizeof(struct utp_transfer_req_desc);
416 }
417 
418 #endif /* _UFSHCD_PRIV_H_ */
419