1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Serial Attached SCSI (SAS) class SCSI Host glue.
4  *
5  * Copyright (C) 2005 Adaptec, Inc.  All rights reserved.
6  * Copyright (C) 2005 Luben Tuikov <luben_tuikov@adaptec.com>
7  */
8 
9 #include <linux/kthread.h>
10 #include <linux/firmware.h>
11 #include <linux/export.h>
12 #include <linux/ctype.h>
13 #include <linux/kernel.h>
14 
15 #include "sas_internal.h"
16 
17 #include <scsi/scsi_host.h>
18 #include <scsi/scsi_device.h>
19 #include <scsi/scsi_tcq.h>
20 #include <scsi/scsi.h>
21 #include <scsi/scsi_eh.h>
22 #include <scsi/scsi_transport.h>
23 #include <scsi/scsi_transport_sas.h>
24 #include <scsi/sas_ata.h>
25 #include "scsi_sas_internal.h"
26 #include "scsi_transport_api.h"
27 #include "scsi_priv.h"
28 
29 #include <linux/err.h>
30 #include <linux/blkdev.h>
31 #include <linux/freezer.h>
32 #include <linux/gfp.h>
33 #include <linux/scatterlist.h>
34 #include <linux/libata.h>
35 
36 /* record final status and free the task */
sas_end_task(struct scsi_cmnd * sc,struct sas_task * task)37 static void sas_end_task(struct scsi_cmnd *sc, struct sas_task *task)
38 {
39 	struct task_status_struct *ts = &task->task_status;
40 	enum scsi_host_status hs = DID_OK;
41 	enum exec_status stat = SAS_SAM_STAT_GOOD;
42 
43 	if (ts->resp == SAS_TASK_UNDELIVERED) {
44 		/* transport error */
45 		hs = DID_NO_CONNECT;
46 	} else { /* ts->resp == SAS_TASK_COMPLETE */
47 		/* task delivered, what happened afterwards? */
48 		switch (ts->stat) {
49 		case SAS_DEV_NO_RESPONSE:
50 		case SAS_INTERRUPTED:
51 		case SAS_PHY_DOWN:
52 		case SAS_NAK_R_ERR:
53 		case SAS_OPEN_TO:
54 			hs = DID_NO_CONNECT;
55 			break;
56 		case SAS_DATA_UNDERRUN:
57 			scsi_set_resid(sc, ts->residual);
58 			if (scsi_bufflen(sc) - scsi_get_resid(sc) < sc->underflow)
59 				hs = DID_ERROR;
60 			break;
61 		case SAS_DATA_OVERRUN:
62 			hs = DID_ERROR;
63 			break;
64 		case SAS_QUEUE_FULL:
65 			hs = DID_SOFT_ERROR; /* retry */
66 			break;
67 		case SAS_DEVICE_UNKNOWN:
68 			hs = DID_BAD_TARGET;
69 			break;
70 		case SAS_OPEN_REJECT:
71 			if (ts->open_rej_reason == SAS_OREJ_RSVD_RETRY)
72 				hs = DID_SOFT_ERROR; /* retry */
73 			else
74 				hs = DID_ERROR;
75 			break;
76 		case SAS_PROTO_RESPONSE:
77 			pr_notice("LLDD:%s sent SAS_PROTO_RESP for an SSP task; please report this\n",
78 				  task->dev->port->ha->sas_ha_name);
79 			break;
80 		case SAS_ABORTED_TASK:
81 			hs = DID_ABORT;
82 			break;
83 		case SAS_SAM_STAT_CHECK_CONDITION:
84 			memcpy(sc->sense_buffer, ts->buf,
85 			       min(SCSI_SENSE_BUFFERSIZE, ts->buf_valid_size));
86 			stat = SAS_SAM_STAT_CHECK_CONDITION;
87 			break;
88 		default:
89 			stat = ts->stat;
90 			break;
91 		}
92 	}
93 
94 	sc->result = (hs << 16) | stat;
95 	ASSIGN_SAS_TASK(sc, NULL);
96 	sas_free_task(task);
97 }
98 
sas_scsi_task_done(struct sas_task * task)99 static void sas_scsi_task_done(struct sas_task *task)
100 {
101 	struct scsi_cmnd *sc = task->uldd_task;
102 	struct domain_device *dev = task->dev;
103 	struct sas_ha_struct *ha = dev->port->ha;
104 	unsigned long flags;
105 
106 	spin_lock_irqsave(&dev->done_lock, flags);
107 	if (test_bit(SAS_HA_FROZEN, &ha->state))
108 		task = NULL;
109 	else
110 		ASSIGN_SAS_TASK(sc, NULL);
111 	spin_unlock_irqrestore(&dev->done_lock, flags);
112 
113 	if (unlikely(!task)) {
114 		/* task will be completed by the error handler */
115 		pr_debug("task done but aborted\n");
116 		return;
117 	}
118 
119 	if (unlikely(!sc)) {
120 		pr_debug("task_done called with non existing SCSI cmnd!\n");
121 		sas_free_task(task);
122 		return;
123 	}
124 
125 	sas_end_task(sc, task);
126 	scsi_done(sc);
127 }
128 
sas_create_task(struct scsi_cmnd * cmd,struct domain_device * dev,gfp_t gfp_flags)129 static struct sas_task *sas_create_task(struct scsi_cmnd *cmd,
130 					       struct domain_device *dev,
131 					       gfp_t gfp_flags)
132 {
133 	struct sas_task *task = sas_alloc_task(gfp_flags);
134 	struct scsi_lun lun;
135 
136 	if (!task)
137 		return NULL;
138 
139 	task->uldd_task = cmd;
140 	ASSIGN_SAS_TASK(cmd, task);
141 
142 	task->dev = dev;
143 	task->task_proto = task->dev->tproto; /* BUG_ON(!SSP) */
144 
145 	int_to_scsilun(cmd->device->lun, &lun);
146 	memcpy(task->ssp_task.LUN, &lun.scsi_lun, 8);
147 	task->ssp_task.task_attr = TASK_ATTR_SIMPLE;
148 	task->ssp_task.cmd = cmd;
149 
150 	task->scatter = scsi_sglist(cmd);
151 	task->num_scatter = scsi_sg_count(cmd);
152 	task->total_xfer_len = scsi_bufflen(cmd);
153 	task->data_dir = cmd->sc_data_direction;
154 
155 	task->task_done = sas_scsi_task_done;
156 
157 	return task;
158 }
159 
sas_queuecommand(struct Scsi_Host * host,struct scsi_cmnd * cmd)160 int sas_queuecommand(struct Scsi_Host *host, struct scsi_cmnd *cmd)
161 {
162 	struct sas_internal *i = to_sas_internal(host->transportt);
163 	struct domain_device *dev = cmd_to_domain_dev(cmd);
164 	struct sas_task *task;
165 	int res = 0;
166 
167 	/* If the device fell off, no sense in issuing commands */
168 	if (test_bit(SAS_DEV_GONE, &dev->state)) {
169 		cmd->result = DID_BAD_TARGET << 16;
170 		goto out_done;
171 	}
172 
173 	if (dev_is_sata(dev)) {
174 		spin_lock_irq(dev->sata_dev.ap->lock);
175 		res = ata_sas_queuecmd(cmd, dev->sata_dev.ap);
176 		spin_unlock_irq(dev->sata_dev.ap->lock);
177 		return res;
178 	}
179 
180 	task = sas_create_task(cmd, dev, GFP_ATOMIC);
181 	if (!task)
182 		return SCSI_MLQUEUE_HOST_BUSY;
183 
184 	res = i->dft->lldd_execute_task(task, GFP_ATOMIC);
185 	if (res)
186 		goto out_free_task;
187 	return 0;
188 
189 out_free_task:
190 	pr_debug("lldd_execute_task returned: %d\n", res);
191 	ASSIGN_SAS_TASK(cmd, NULL);
192 	sas_free_task(task);
193 	if (res == -SAS_QUEUE_FULL)
194 		cmd->result = DID_SOFT_ERROR << 16; /* retry */
195 	else
196 		cmd->result = DID_ERROR << 16;
197 out_done:
198 	scsi_done(cmd);
199 	return 0;
200 }
201 EXPORT_SYMBOL_GPL(sas_queuecommand);
202 
sas_eh_finish_cmd(struct scsi_cmnd * cmd)203 static void sas_eh_finish_cmd(struct scsi_cmnd *cmd)
204 {
205 	struct sas_ha_struct *sas_ha = SHOST_TO_SAS_HA(cmd->device->host);
206 	struct domain_device *dev = cmd_to_domain_dev(cmd);
207 	struct sas_task *task = TO_SAS_TASK(cmd);
208 
209 	/* At this point, we only get called following an actual abort
210 	 * of the task, so we should be guaranteed not to be racing with
211 	 * any completions from the LLD.  Task is freed after this.
212 	 */
213 	sas_end_task(cmd, task);
214 
215 	if (dev_is_sata(dev)) {
216 		/* defer commands to libata so that libata EH can
217 		 * handle ata qcs correctly
218 		 */
219 		list_move_tail(&cmd->eh_entry, &sas_ha->eh_ata_q);
220 		return;
221 	}
222 
223 	/* now finish the command and move it on to the error
224 	 * handler done list, this also takes it off the
225 	 * error handler pending list.
226 	 */
227 	scsi_eh_finish_cmd(cmd, &sas_ha->eh_done_q);
228 }
229 
sas_scsi_clear_queue_lu(struct list_head * error_q,struct scsi_cmnd * my_cmd)230 static void sas_scsi_clear_queue_lu(struct list_head *error_q, struct scsi_cmnd *my_cmd)
231 {
232 	struct scsi_cmnd *cmd, *n;
233 
234 	list_for_each_entry_safe(cmd, n, error_q, eh_entry) {
235 		if (cmd->device->sdev_target == my_cmd->device->sdev_target &&
236 		    cmd->device->lun == my_cmd->device->lun)
237 			sas_eh_finish_cmd(cmd);
238 	}
239 }
240 
sas_scsi_clear_queue_I_T(struct list_head * error_q,struct domain_device * dev)241 static void sas_scsi_clear_queue_I_T(struct list_head *error_q,
242 				     struct domain_device *dev)
243 {
244 	struct scsi_cmnd *cmd, *n;
245 
246 	list_for_each_entry_safe(cmd, n, error_q, eh_entry) {
247 		struct domain_device *x = cmd_to_domain_dev(cmd);
248 
249 		if (x == dev)
250 			sas_eh_finish_cmd(cmd);
251 	}
252 }
253 
sas_scsi_clear_queue_port(struct list_head * error_q,struct asd_sas_port * port)254 static void sas_scsi_clear_queue_port(struct list_head *error_q,
255 				      struct asd_sas_port *port)
256 {
257 	struct scsi_cmnd *cmd, *n;
258 
259 	list_for_each_entry_safe(cmd, n, error_q, eh_entry) {
260 		struct domain_device *dev = cmd_to_domain_dev(cmd);
261 		struct asd_sas_port *x = dev->port;
262 
263 		if (x == port)
264 			sas_eh_finish_cmd(cmd);
265 	}
266 }
267 
268 enum task_disposition {
269 	TASK_IS_DONE,
270 	TASK_IS_ABORTED,
271 	TASK_IS_AT_LU,
272 	TASK_IS_NOT_AT_LU,
273 	TASK_ABORT_FAILED,
274 };
275 
sas_scsi_find_task(struct sas_task * task)276 static enum task_disposition sas_scsi_find_task(struct sas_task *task)
277 {
278 	unsigned long flags;
279 	int i, res;
280 	struct sas_internal *si =
281 		to_sas_internal(task->dev->port->ha->shost->transportt);
282 
283 	for (i = 0; i < 5; i++) {
284 		pr_notice("%s: aborting task 0x%p\n", __func__, task);
285 		res = si->dft->lldd_abort_task(task);
286 
287 		spin_lock_irqsave(&task->task_state_lock, flags);
288 		if (task->task_state_flags & SAS_TASK_STATE_DONE) {
289 			spin_unlock_irqrestore(&task->task_state_lock, flags);
290 			pr_debug("%s: task 0x%p is done\n", __func__, task);
291 			return TASK_IS_DONE;
292 		}
293 		spin_unlock_irqrestore(&task->task_state_lock, flags);
294 
295 		if (res == TMF_RESP_FUNC_COMPLETE) {
296 			pr_notice("%s: task 0x%p is aborted\n",
297 				  __func__, task);
298 			return TASK_IS_ABORTED;
299 		} else if (si->dft->lldd_query_task) {
300 			pr_notice("%s: querying task 0x%p\n", __func__, task);
301 			res = si->dft->lldd_query_task(task);
302 			switch (res) {
303 			case TMF_RESP_FUNC_SUCC:
304 				pr_notice("%s: task 0x%p at LU\n", __func__,
305 					  task);
306 				return TASK_IS_AT_LU;
307 			case TMF_RESP_FUNC_COMPLETE:
308 				pr_notice("%s: task 0x%p not at LU\n",
309 					  __func__, task);
310 				return TASK_IS_NOT_AT_LU;
311 			case TMF_RESP_FUNC_FAILED:
312 				pr_notice("%s: task 0x%p failed to abort\n",
313 					  __func__, task);
314 				return TASK_ABORT_FAILED;
315 			default:
316 				pr_notice("%s: task 0x%p result code %d not handled\n",
317 					  __func__, task, res);
318 			}
319 		}
320 	}
321 	return TASK_ABORT_FAILED;
322 }
323 
sas_recover_lu(struct domain_device * dev,struct scsi_cmnd * cmd)324 static int sas_recover_lu(struct domain_device *dev, struct scsi_cmnd *cmd)
325 {
326 	int res = TMF_RESP_FUNC_FAILED;
327 	struct scsi_lun lun;
328 	struct sas_internal *i =
329 		to_sas_internal(dev->port->ha->shost->transportt);
330 
331 	int_to_scsilun(cmd->device->lun, &lun);
332 
333 	pr_notice("eh: device %016llx LUN 0x%llx has the task\n",
334 		  SAS_ADDR(dev->sas_addr),
335 		  cmd->device->lun);
336 
337 	if (i->dft->lldd_abort_task_set)
338 		res = i->dft->lldd_abort_task_set(dev, lun.scsi_lun);
339 
340 	if (res == TMF_RESP_FUNC_FAILED) {
341 		if (i->dft->lldd_clear_task_set)
342 			res = i->dft->lldd_clear_task_set(dev, lun.scsi_lun);
343 	}
344 
345 	if (res == TMF_RESP_FUNC_FAILED) {
346 		if (i->dft->lldd_lu_reset)
347 			res = i->dft->lldd_lu_reset(dev, lun.scsi_lun);
348 	}
349 
350 	return res;
351 }
352 
sas_recover_I_T(struct domain_device * dev)353 static int sas_recover_I_T(struct domain_device *dev)
354 {
355 	int res = TMF_RESP_FUNC_FAILED;
356 	struct sas_internal *i =
357 		to_sas_internal(dev->port->ha->shost->transportt);
358 
359 	pr_notice("I_T nexus reset for dev %016llx\n",
360 		  SAS_ADDR(dev->sas_addr));
361 
362 	if (i->dft->lldd_I_T_nexus_reset)
363 		res = i->dft->lldd_I_T_nexus_reset(dev);
364 
365 	return res;
366 }
367 
368 /* take a reference on the last known good phy for this device */
sas_get_local_phy(struct domain_device * dev)369 struct sas_phy *sas_get_local_phy(struct domain_device *dev)
370 {
371 	struct sas_ha_struct *ha = dev->port->ha;
372 	struct sas_phy *phy;
373 	unsigned long flags;
374 
375 	/* a published domain device always has a valid phy, it may be
376 	 * stale, but it is never NULL
377 	 */
378 	BUG_ON(!dev->phy);
379 
380 	spin_lock_irqsave(&ha->phy_port_lock, flags);
381 	phy = dev->phy;
382 	get_device(&phy->dev);
383 	spin_unlock_irqrestore(&ha->phy_port_lock, flags);
384 
385 	return phy;
386 }
387 EXPORT_SYMBOL_GPL(sas_get_local_phy);
388 
sas_queue_reset(struct domain_device * dev,int reset_type,u64 lun)389 static int sas_queue_reset(struct domain_device *dev, int reset_type, u64 lun)
390 {
391 	struct sas_ha_struct *ha = dev->port->ha;
392 	int scheduled = 0, tries = 100;
393 
394 	/* ata: promote lun reset to bus reset */
395 	if (dev_is_sata(dev)) {
396 		sas_ata_schedule_reset(dev);
397 		return SUCCESS;
398 	}
399 
400 	while (!scheduled && tries--) {
401 		spin_lock_irq(&ha->lock);
402 		if (!test_bit(SAS_DEV_EH_PENDING, &dev->state) &&
403 		    !test_bit(reset_type, &dev->state)) {
404 			scheduled = 1;
405 			ha->eh_active++;
406 			list_add_tail(&dev->ssp_dev.eh_list_node, &ha->eh_dev_q);
407 			set_bit(SAS_DEV_EH_PENDING, &dev->state);
408 			set_bit(reset_type, &dev->state);
409 			int_to_scsilun(lun, &dev->ssp_dev.reset_lun);
410 			scsi_schedule_eh(ha->shost);
411 		}
412 		spin_unlock_irq(&ha->lock);
413 
414 		if (scheduled)
415 			return SUCCESS;
416 	}
417 
418 	pr_warn("%s reset of %s failed\n",
419 		reset_type == SAS_DEV_LU_RESET ? "LUN" : "Bus",
420 		dev_name(&dev->rphy->dev));
421 
422 	return FAILED;
423 }
424 
sas_eh_abort_handler(struct scsi_cmnd * cmd)425 int sas_eh_abort_handler(struct scsi_cmnd *cmd)
426 {
427 	int res = TMF_RESP_FUNC_FAILED;
428 	struct sas_task *task = TO_SAS_TASK(cmd);
429 	struct Scsi_Host *host = cmd->device->host;
430 	struct domain_device *dev = cmd_to_domain_dev(cmd);
431 	struct sas_internal *i = to_sas_internal(host->transportt);
432 	unsigned long flags;
433 
434 	if (!i->dft->lldd_abort_task)
435 		return FAILED;
436 
437 	spin_lock_irqsave(host->host_lock, flags);
438 	/* We cannot do async aborts for SATA devices */
439 	if (dev_is_sata(dev) && !host->host_eh_scheduled) {
440 		spin_unlock_irqrestore(host->host_lock, flags);
441 		return FAILED;
442 	}
443 	spin_unlock_irqrestore(host->host_lock, flags);
444 
445 	if (task)
446 		res = i->dft->lldd_abort_task(task);
447 	else
448 		pr_notice("no task to abort\n");
449 	if (res == TMF_RESP_FUNC_SUCC || res == TMF_RESP_FUNC_COMPLETE)
450 		return SUCCESS;
451 
452 	return FAILED;
453 }
454 EXPORT_SYMBOL_GPL(sas_eh_abort_handler);
455 
456 /* Attempt to send a LUN reset message to a device */
sas_eh_device_reset_handler(struct scsi_cmnd * cmd)457 int sas_eh_device_reset_handler(struct scsi_cmnd *cmd)
458 {
459 	int res;
460 	struct scsi_lun lun;
461 	struct Scsi_Host *host = cmd->device->host;
462 	struct domain_device *dev = cmd_to_domain_dev(cmd);
463 	struct sas_internal *i = to_sas_internal(host->transportt);
464 
465 	if (current != host->ehandler)
466 		return sas_queue_reset(dev, SAS_DEV_LU_RESET, cmd->device->lun);
467 
468 	int_to_scsilun(cmd->device->lun, &lun);
469 
470 	if (!i->dft->lldd_lu_reset)
471 		return FAILED;
472 
473 	res = i->dft->lldd_lu_reset(dev, lun.scsi_lun);
474 	if (res == TMF_RESP_FUNC_SUCC || res == TMF_RESP_FUNC_COMPLETE)
475 		return SUCCESS;
476 
477 	return FAILED;
478 }
479 EXPORT_SYMBOL_GPL(sas_eh_device_reset_handler);
480 
sas_eh_target_reset_handler(struct scsi_cmnd * cmd)481 int sas_eh_target_reset_handler(struct scsi_cmnd *cmd)
482 {
483 	int res;
484 	struct Scsi_Host *host = cmd->device->host;
485 	struct domain_device *dev = cmd_to_domain_dev(cmd);
486 	struct sas_internal *i = to_sas_internal(host->transportt);
487 
488 	if (current != host->ehandler)
489 		return sas_queue_reset(dev, SAS_DEV_RESET, 0);
490 
491 	if (!i->dft->lldd_I_T_nexus_reset)
492 		return FAILED;
493 
494 	res = i->dft->lldd_I_T_nexus_reset(dev);
495 	if (res == TMF_RESP_FUNC_SUCC || res == TMF_RESP_FUNC_COMPLETE ||
496 	    res == -ENODEV)
497 		return SUCCESS;
498 
499 	return FAILED;
500 }
501 EXPORT_SYMBOL_GPL(sas_eh_target_reset_handler);
502 
503 /* Try to reset a device */
try_to_reset_cmd_device(struct scsi_cmnd * cmd)504 static int try_to_reset_cmd_device(struct scsi_cmnd *cmd)
505 {
506 	int res;
507 	struct Scsi_Host *shost = cmd->device->host;
508 
509 	if (!shost->hostt->eh_device_reset_handler)
510 		goto try_target_reset;
511 
512 	res = shost->hostt->eh_device_reset_handler(cmd);
513 	if (res == SUCCESS)
514 		return res;
515 
516 try_target_reset:
517 	if (shost->hostt->eh_target_reset_handler)
518 		return shost->hostt->eh_target_reset_handler(cmd);
519 
520 	return FAILED;
521 }
522 
sas_eh_handle_sas_errors(struct Scsi_Host * shost,struct list_head * work_q)523 static void sas_eh_handle_sas_errors(struct Scsi_Host *shost, struct list_head *work_q)
524 {
525 	struct scsi_cmnd *cmd, *n;
526 	enum task_disposition res = TASK_IS_DONE;
527 	int tmf_resp, need_reset;
528 	struct sas_internal *i = to_sas_internal(shost->transportt);
529 	unsigned long flags;
530 	struct sas_ha_struct *ha = SHOST_TO_SAS_HA(shost);
531 	LIST_HEAD(done);
532 
533 	/* clean out any commands that won the completion vs eh race */
534 	list_for_each_entry_safe(cmd, n, work_q, eh_entry) {
535 		struct domain_device *dev = cmd_to_domain_dev(cmd);
536 		struct sas_task *task;
537 
538 		spin_lock_irqsave(&dev->done_lock, flags);
539 		/* by this point the lldd has either observed
540 		 * SAS_HA_FROZEN and is leaving the task alone, or has
541 		 * won the race with eh and decided to complete it
542 		 */
543 		task = TO_SAS_TASK(cmd);
544 		spin_unlock_irqrestore(&dev->done_lock, flags);
545 
546 		if (!task)
547 			list_move_tail(&cmd->eh_entry, &done);
548 	}
549 
550  Again:
551 	list_for_each_entry_safe(cmd, n, work_q, eh_entry) {
552 		struct sas_task *task = TO_SAS_TASK(cmd);
553 
554 		list_del_init(&cmd->eh_entry);
555 
556 		spin_lock_irqsave(&task->task_state_lock, flags);
557 		need_reset = task->task_state_flags & SAS_TASK_NEED_DEV_RESET;
558 		spin_unlock_irqrestore(&task->task_state_lock, flags);
559 
560 		if (need_reset) {
561 			pr_notice("%s: task 0x%p requests reset\n",
562 				  __func__, task);
563 			goto reset;
564 		}
565 
566 		pr_debug("trying to find task 0x%p\n", task);
567 		res = sas_scsi_find_task(task);
568 
569 		switch (res) {
570 		case TASK_IS_DONE:
571 			pr_notice("%s: task 0x%p is done\n", __func__,
572 				    task);
573 			sas_eh_finish_cmd(cmd);
574 			continue;
575 		case TASK_IS_ABORTED:
576 			pr_notice("%s: task 0x%p is aborted\n",
577 				  __func__, task);
578 			sas_eh_finish_cmd(cmd);
579 			continue;
580 		case TASK_IS_AT_LU:
581 			pr_info("task 0x%p is at LU: lu recover\n", task);
582  reset:
583 			tmf_resp = sas_recover_lu(task->dev, cmd);
584 			if (tmf_resp == TMF_RESP_FUNC_COMPLETE) {
585 				pr_notice("dev %016llx LU 0x%llx is recovered\n",
586 					  SAS_ADDR(task->dev),
587 					  cmd->device->lun);
588 				sas_eh_finish_cmd(cmd);
589 				sas_scsi_clear_queue_lu(work_q, cmd);
590 				goto Again;
591 			}
592 			fallthrough;
593 		case TASK_IS_NOT_AT_LU:
594 		case TASK_ABORT_FAILED:
595 			pr_notice("task 0x%p is not at LU: I_T recover\n",
596 				  task);
597 			tmf_resp = sas_recover_I_T(task->dev);
598 			if (tmf_resp == TMF_RESP_FUNC_COMPLETE ||
599 			    tmf_resp == -ENODEV) {
600 				struct domain_device *dev = task->dev;
601 				pr_notice("I_T %016llx recovered\n",
602 					  SAS_ADDR(task->dev->sas_addr));
603 				sas_eh_finish_cmd(cmd);
604 				sas_scsi_clear_queue_I_T(work_q, dev);
605 				goto Again;
606 			}
607 			/* Hammer time :-) */
608 			try_to_reset_cmd_device(cmd);
609 			if (i->dft->lldd_clear_nexus_port) {
610 				struct asd_sas_port *port = task->dev->port;
611 				pr_debug("clearing nexus for port:%d\n",
612 					  port->id);
613 				res = i->dft->lldd_clear_nexus_port(port);
614 				if (res == TMF_RESP_FUNC_COMPLETE) {
615 					pr_notice("clear nexus port:%d succeeded\n",
616 						  port->id);
617 					sas_eh_finish_cmd(cmd);
618 					sas_scsi_clear_queue_port(work_q,
619 								  port);
620 					goto Again;
621 				}
622 			}
623 			if (i->dft->lldd_clear_nexus_ha) {
624 				pr_debug("clear nexus ha\n");
625 				res = i->dft->lldd_clear_nexus_ha(ha);
626 				if (res == TMF_RESP_FUNC_COMPLETE) {
627 					pr_notice("clear nexus ha succeeded\n");
628 					sas_eh_finish_cmd(cmd);
629 					goto clear_q;
630 				}
631 			}
632 			/* If we are here -- this means that no amount
633 			 * of effort could recover from errors.  Quite
634 			 * possibly the HA just disappeared.
635 			 */
636 			pr_err("error from device %016llx, LUN 0x%llx couldn't be recovered in any way\n",
637 			       SAS_ADDR(task->dev->sas_addr),
638 			       cmd->device->lun);
639 
640 			sas_eh_finish_cmd(cmd);
641 			goto clear_q;
642 		}
643 	}
644  out:
645 	list_splice_tail(&done, work_q);
646 	list_splice_tail_init(&ha->eh_ata_q, work_q);
647 	return;
648 
649  clear_q:
650 	pr_debug("--- Exit %s -- clear_q\n", __func__);
651 	list_for_each_entry_safe(cmd, n, work_q, eh_entry)
652 		sas_eh_finish_cmd(cmd);
653 	goto out;
654 }
655 
sas_eh_handle_resets(struct Scsi_Host * shost)656 static void sas_eh_handle_resets(struct Scsi_Host *shost)
657 {
658 	struct sas_ha_struct *ha = SHOST_TO_SAS_HA(shost);
659 	struct sas_internal *i = to_sas_internal(shost->transportt);
660 
661 	/* handle directed resets to sas devices */
662 	spin_lock_irq(&ha->lock);
663 	while (!list_empty(&ha->eh_dev_q)) {
664 		struct domain_device *dev;
665 		struct ssp_device *ssp;
666 
667 		ssp = list_entry(ha->eh_dev_q.next, typeof(*ssp), eh_list_node);
668 		list_del_init(&ssp->eh_list_node);
669 		dev = container_of(ssp, typeof(*dev), ssp_dev);
670 		kref_get(&dev->kref);
671 		WARN_ONCE(dev_is_sata(dev), "ssp reset to ata device?\n");
672 
673 		spin_unlock_irq(&ha->lock);
674 
675 		if (test_and_clear_bit(SAS_DEV_LU_RESET, &dev->state))
676 			i->dft->lldd_lu_reset(dev, ssp->reset_lun.scsi_lun);
677 
678 		if (test_and_clear_bit(SAS_DEV_RESET, &dev->state))
679 			i->dft->lldd_I_T_nexus_reset(dev);
680 
681 		sas_put_device(dev);
682 		spin_lock_irq(&ha->lock);
683 		clear_bit(SAS_DEV_EH_PENDING, &dev->state);
684 		ha->eh_active--;
685 	}
686 	spin_unlock_irq(&ha->lock);
687 }
688 
689 
sas_scsi_recover_host(struct Scsi_Host * shost)690 void sas_scsi_recover_host(struct Scsi_Host *shost)
691 {
692 	struct sas_ha_struct *ha = SHOST_TO_SAS_HA(shost);
693 	LIST_HEAD(eh_work_q);
694 	int tries = 0;
695 	bool retry;
696 
697 retry:
698 	tries++;
699 	retry = true;
700 	spin_lock_irq(shost->host_lock);
701 	list_splice_init(&shost->eh_cmd_q, &eh_work_q);
702 	spin_unlock_irq(shost->host_lock);
703 
704 	pr_notice("Enter %s busy: %d failed: %d\n",
705 		  __func__, scsi_host_busy(shost), shost->host_failed);
706 	/*
707 	 * Deal with commands that still have SAS tasks (i.e. they didn't
708 	 * complete via the normal sas_task completion mechanism),
709 	 * SAS_HA_FROZEN gives eh dominion over all sas_task completion.
710 	 */
711 	set_bit(SAS_HA_FROZEN, &ha->state);
712 	sas_eh_handle_sas_errors(shost, &eh_work_q);
713 	clear_bit(SAS_HA_FROZEN, &ha->state);
714 	if (list_empty(&eh_work_q))
715 		goto out;
716 
717 	/*
718 	 * Now deal with SCSI commands that completed ok but have a an error
719 	 * code (and hopefully sense data) attached.  This is roughly what
720 	 * scsi_unjam_host does, but we skip scsi_eh_abort_cmds because any
721 	 * command we see here has no sas_task and is thus unknown to the HA.
722 	 */
723 	sas_ata_eh(shost, &eh_work_q);
724 	if (!scsi_eh_get_sense(&eh_work_q, &ha->eh_done_q))
725 		scsi_eh_ready_devs(shost, &eh_work_q, &ha->eh_done_q);
726 
727 out:
728 	sas_eh_handle_resets(shost);
729 
730 	/* now link into libata eh --- if we have any ata devices */
731 	sas_ata_strategy_handler(shost);
732 
733 	scsi_eh_flush_done_q(&ha->eh_done_q);
734 
735 	/* check if any new eh work was scheduled during the last run */
736 	spin_lock_irq(&ha->lock);
737 	if (ha->eh_active == 0) {
738 		shost->host_eh_scheduled = 0;
739 		retry = false;
740 	}
741 	spin_unlock_irq(&ha->lock);
742 
743 	if (retry)
744 		goto retry;
745 
746 	pr_notice("--- Exit %s: busy: %d failed: %d tries: %d\n",
747 		  __func__, scsi_host_busy(shost),
748 		  shost->host_failed, tries);
749 }
750 
sas_ioctl(struct scsi_device * sdev,unsigned int cmd,void __user * arg)751 int sas_ioctl(struct scsi_device *sdev, unsigned int cmd, void __user *arg)
752 {
753 	struct domain_device *dev = sdev_to_domain_dev(sdev);
754 
755 	if (dev_is_sata(dev))
756 		return ata_sas_scsi_ioctl(dev->sata_dev.ap, sdev, cmd, arg);
757 
758 	return -EINVAL;
759 }
760 EXPORT_SYMBOL_GPL(sas_ioctl);
761 
sas_find_dev_by_rphy(struct sas_rphy * rphy)762 struct domain_device *sas_find_dev_by_rphy(struct sas_rphy *rphy)
763 {
764 	struct Scsi_Host *shost = dev_to_shost(rphy->dev.parent);
765 	struct sas_ha_struct *ha = SHOST_TO_SAS_HA(shost);
766 	struct domain_device *found_dev = NULL;
767 	int i;
768 	unsigned long flags;
769 
770 	spin_lock_irqsave(&ha->phy_port_lock, flags);
771 	for (i = 0; i < ha->num_phys; i++) {
772 		struct asd_sas_port *port = ha->sas_port[i];
773 		struct domain_device *dev;
774 
775 		spin_lock(&port->dev_list_lock);
776 		list_for_each_entry(dev, &port->dev_list, dev_list_node) {
777 			if (rphy == dev->rphy) {
778 				found_dev = dev;
779 				spin_unlock(&port->dev_list_lock);
780 				goto found;
781 			}
782 		}
783 		spin_unlock(&port->dev_list_lock);
784 	}
785  found:
786 	spin_unlock_irqrestore(&ha->phy_port_lock, flags);
787 
788 	return found_dev;
789 }
790 
sas_target_alloc(struct scsi_target * starget)791 int sas_target_alloc(struct scsi_target *starget)
792 {
793 	struct sas_rphy *rphy = dev_to_rphy(starget->dev.parent);
794 	struct domain_device *found_dev = sas_find_dev_by_rphy(rphy);
795 
796 	if (!found_dev)
797 		return -ENODEV;
798 
799 	kref_get(&found_dev->kref);
800 	starget->hostdata = found_dev;
801 	return 0;
802 }
803 EXPORT_SYMBOL_GPL(sas_target_alloc);
804 
805 #define SAS_DEF_QD 256
806 
sas_slave_configure(struct scsi_device * scsi_dev)807 int sas_slave_configure(struct scsi_device *scsi_dev)
808 {
809 	struct domain_device *dev = sdev_to_domain_dev(scsi_dev);
810 
811 	BUG_ON(dev->rphy->identify.device_type != SAS_END_DEVICE);
812 
813 	if (dev_is_sata(dev)) {
814 		ata_sas_slave_configure(scsi_dev, dev->sata_dev.ap);
815 		return 0;
816 	}
817 
818 	sas_read_port_mode_page(scsi_dev);
819 
820 	if (scsi_dev->tagged_supported) {
821 		scsi_change_queue_depth(scsi_dev, SAS_DEF_QD);
822 	} else {
823 		pr_notice("device %016llx, LUN 0x%llx doesn't support TCQ\n",
824 			  SAS_ADDR(dev->sas_addr), scsi_dev->lun);
825 		scsi_change_queue_depth(scsi_dev, 1);
826 	}
827 
828 	scsi_dev->allow_restart = 1;
829 
830 	return 0;
831 }
832 EXPORT_SYMBOL_GPL(sas_slave_configure);
833 
sas_change_queue_depth(struct scsi_device * sdev,int depth)834 int sas_change_queue_depth(struct scsi_device *sdev, int depth)
835 {
836 	struct domain_device *dev = sdev_to_domain_dev(sdev);
837 
838 	if (dev_is_sata(dev))
839 		return ata_change_queue_depth(dev->sata_dev.ap, sdev, depth);
840 
841 	if (!sdev->tagged_supported)
842 		depth = 1;
843 	return scsi_change_queue_depth(sdev, depth);
844 }
845 EXPORT_SYMBOL_GPL(sas_change_queue_depth);
846 
sas_bios_param(struct scsi_device * scsi_dev,struct block_device * bdev,sector_t capacity,int * hsc)847 int sas_bios_param(struct scsi_device *scsi_dev,
848 			  struct block_device *bdev,
849 			  sector_t capacity, int *hsc)
850 {
851 	hsc[0] = 255;
852 	hsc[1] = 63;
853 	sector_div(capacity, 255*63);
854 	hsc[2] = capacity;
855 
856 	return 0;
857 }
858 EXPORT_SYMBOL_GPL(sas_bios_param);
859 
sas_task_internal_done(struct sas_task * task)860 void sas_task_internal_done(struct sas_task *task)
861 {
862 	del_timer(&task->slow_task->timer);
863 	complete(&task->slow_task->completion);
864 }
865 
sas_task_internal_timedout(struct timer_list * t)866 void sas_task_internal_timedout(struct timer_list *t)
867 {
868 	struct sas_task_slow *slow = from_timer(slow, t, timer);
869 	struct sas_task *task = slow->task;
870 	bool is_completed = true;
871 	unsigned long flags;
872 
873 	spin_lock_irqsave(&task->task_state_lock, flags);
874 	if (!(task->task_state_flags & SAS_TASK_STATE_DONE)) {
875 		task->task_state_flags |= SAS_TASK_STATE_ABORTED;
876 		is_completed = false;
877 	}
878 	spin_unlock_irqrestore(&task->task_state_lock, flags);
879 
880 	if (!is_completed)
881 		complete(&task->slow_task->completion);
882 }
883 
884 #define TASK_TIMEOUT			(20 * HZ)
885 #define TASK_RETRY			3
886 
sas_execute_internal_abort(struct domain_device * device,enum sas_internal_abort type,u16 tag,unsigned int qid,void * data)887 static int sas_execute_internal_abort(struct domain_device *device,
888 				      enum sas_internal_abort type, u16 tag,
889 				      unsigned int qid, void *data)
890 {
891 	struct sas_ha_struct *ha = device->port->ha;
892 	struct sas_internal *i = to_sas_internal(ha->shost->transportt);
893 	struct sas_task *task = NULL;
894 	int res, retry;
895 
896 	for (retry = 0; retry < TASK_RETRY; retry++) {
897 		task = sas_alloc_slow_task(GFP_KERNEL);
898 		if (!task)
899 			return -ENOMEM;
900 
901 		task->dev = device;
902 		task->task_proto = SAS_PROTOCOL_INTERNAL_ABORT;
903 		task->task_done = sas_task_internal_done;
904 		task->slow_task->timer.function = sas_task_internal_timedout;
905 		task->slow_task->timer.expires = jiffies + TASK_TIMEOUT;
906 		add_timer(&task->slow_task->timer);
907 
908 		task->abort_task.tag = tag;
909 		task->abort_task.type = type;
910 		task->abort_task.qid = qid;
911 
912 		res = i->dft->lldd_execute_task(task, GFP_KERNEL);
913 		if (res) {
914 			del_timer_sync(&task->slow_task->timer);
915 			pr_err("Executing internal abort failed %016llx (%d)\n",
916 			       SAS_ADDR(device->sas_addr), res);
917 			break;
918 		}
919 
920 		wait_for_completion(&task->slow_task->completion);
921 		res = TMF_RESP_FUNC_FAILED;
922 
923 		/* Even if the internal abort timed out, return direct. */
924 		if (task->task_state_flags & SAS_TASK_STATE_ABORTED) {
925 			bool quit = true;
926 
927 			if (i->dft->lldd_abort_timeout)
928 				quit = i->dft->lldd_abort_timeout(task, data);
929 			else
930 				pr_err("Internal abort: timeout %016llx\n",
931 				       SAS_ADDR(device->sas_addr));
932 			res = -EIO;
933 			if (quit)
934 				break;
935 		}
936 
937 		if (task->task_status.resp == SAS_TASK_COMPLETE &&
938 			task->task_status.stat == SAS_SAM_STAT_GOOD) {
939 			res = TMF_RESP_FUNC_COMPLETE;
940 			break;
941 		}
942 
943 		if (task->task_status.resp == SAS_TASK_COMPLETE &&
944 			task->task_status.stat == TMF_RESP_FUNC_SUCC) {
945 			res = TMF_RESP_FUNC_SUCC;
946 			break;
947 		}
948 
949 		pr_err("Internal abort: task to dev %016llx response: 0x%x status 0x%x\n",
950 		       SAS_ADDR(device->sas_addr), task->task_status.resp,
951 		       task->task_status.stat);
952 		sas_free_task(task);
953 		task = NULL;
954 	}
955 	BUG_ON(retry == TASK_RETRY && task != NULL);
956 	sas_free_task(task);
957 	return res;
958 }
959 
sas_execute_internal_abort_single(struct domain_device * device,u16 tag,unsigned int qid,void * data)960 int sas_execute_internal_abort_single(struct domain_device *device, u16 tag,
961 				      unsigned int qid, void *data)
962 {
963 	return sas_execute_internal_abort(device, SAS_INTERNAL_ABORT_SINGLE,
964 					  tag, qid, data);
965 }
966 EXPORT_SYMBOL_GPL(sas_execute_internal_abort_single);
967 
sas_execute_internal_abort_dev(struct domain_device * device,unsigned int qid,void * data)968 int sas_execute_internal_abort_dev(struct domain_device *device,
969 				   unsigned int qid, void *data)
970 {
971 	return sas_execute_internal_abort(device, SAS_INTERNAL_ABORT_DEV,
972 					  SCSI_NO_TAG, qid, data);
973 }
974 EXPORT_SYMBOL_GPL(sas_execute_internal_abort_dev);
975 
sas_execute_tmf(struct domain_device * device,void * parameter,int para_len,int force_phy_id,struct sas_tmf_task * tmf)976 int sas_execute_tmf(struct domain_device *device, void *parameter,
977 		    int para_len, int force_phy_id,
978 		    struct sas_tmf_task *tmf)
979 {
980 	struct sas_task *task;
981 	struct sas_internal *i =
982 		to_sas_internal(device->port->ha->shost->transportt);
983 	int res, retry;
984 
985 	for (retry = 0; retry < TASK_RETRY; retry++) {
986 		task = sas_alloc_slow_task(GFP_KERNEL);
987 		if (!task)
988 			return -ENOMEM;
989 
990 		task->dev = device;
991 		task->task_proto = device->tproto;
992 
993 		if (dev_is_sata(device)) {
994 			task->ata_task.device_control_reg_update = 1;
995 			if (force_phy_id >= 0) {
996 				task->ata_task.force_phy = true;
997 				task->ata_task.force_phy_id = force_phy_id;
998 			}
999 			memcpy(&task->ata_task.fis, parameter, para_len);
1000 		} else {
1001 			memcpy(&task->ssp_task, parameter, para_len);
1002 		}
1003 
1004 		task->task_done = sas_task_internal_done;
1005 		task->tmf = tmf;
1006 
1007 		task->slow_task->timer.function = sas_task_internal_timedout;
1008 		task->slow_task->timer.expires = jiffies + TASK_TIMEOUT;
1009 		add_timer(&task->slow_task->timer);
1010 
1011 		res = i->dft->lldd_execute_task(task, GFP_KERNEL);
1012 		if (res) {
1013 			del_timer_sync(&task->slow_task->timer);
1014 			pr_err("executing TMF task failed %016llx (%d)\n",
1015 			       SAS_ADDR(device->sas_addr), res);
1016 			break;
1017 		}
1018 
1019 		wait_for_completion(&task->slow_task->completion);
1020 
1021 		if (i->dft->lldd_tmf_exec_complete)
1022 			i->dft->lldd_tmf_exec_complete(device);
1023 
1024 		res = TMF_RESP_FUNC_FAILED;
1025 
1026 		if ((task->task_state_flags & SAS_TASK_STATE_ABORTED)) {
1027 			if (!(task->task_state_flags & SAS_TASK_STATE_DONE)) {
1028 				pr_err("TMF task timeout for %016llx and not done\n",
1029 				       SAS_ADDR(device->sas_addr));
1030 				if (i->dft->lldd_tmf_aborted)
1031 					i->dft->lldd_tmf_aborted(task);
1032 				break;
1033 			}
1034 			pr_warn("TMF task timeout for %016llx and done\n",
1035 				SAS_ADDR(device->sas_addr));
1036 		}
1037 
1038 		if (task->task_status.resp == SAS_TASK_COMPLETE &&
1039 		    task->task_status.stat == TMF_RESP_FUNC_COMPLETE) {
1040 			res = TMF_RESP_FUNC_COMPLETE;
1041 			break;
1042 		}
1043 
1044 		if (task->task_status.resp == SAS_TASK_COMPLETE &&
1045 		    task->task_status.stat == TMF_RESP_FUNC_SUCC) {
1046 			res = TMF_RESP_FUNC_SUCC;
1047 			break;
1048 		}
1049 
1050 		if (task->task_status.resp == SAS_TASK_COMPLETE &&
1051 		    task->task_status.stat == SAS_DATA_UNDERRUN) {
1052 			/* no error, but return the number of bytes of
1053 			 * underrun
1054 			 */
1055 			pr_warn("TMF task to dev %016llx resp: 0x%x sts 0x%x underrun\n",
1056 				SAS_ADDR(device->sas_addr),
1057 				task->task_status.resp,
1058 				task->task_status.stat);
1059 			res = task->task_status.residual;
1060 			break;
1061 		}
1062 
1063 		if (task->task_status.resp == SAS_TASK_COMPLETE &&
1064 		    task->task_status.stat == SAS_DATA_OVERRUN) {
1065 			pr_warn("TMF task blocked task error %016llx\n",
1066 				SAS_ADDR(device->sas_addr));
1067 			res = -EMSGSIZE;
1068 			break;
1069 		}
1070 
1071 		if (task->task_status.resp == SAS_TASK_COMPLETE &&
1072 		    task->task_status.stat == SAS_OPEN_REJECT) {
1073 			pr_warn("TMF task open reject failed  %016llx\n",
1074 				SAS_ADDR(device->sas_addr));
1075 			res = -EIO;
1076 		} else {
1077 			pr_warn("TMF task to dev %016llx resp: 0x%x status 0x%x\n",
1078 				SAS_ADDR(device->sas_addr),
1079 				task->task_status.resp,
1080 				task->task_status.stat);
1081 		}
1082 		sas_free_task(task);
1083 		task = NULL;
1084 	}
1085 
1086 	if (retry == TASK_RETRY)
1087 		pr_warn("executing TMF for %016llx failed after %d attempts!\n",
1088 			SAS_ADDR(device->sas_addr), TASK_RETRY);
1089 	sas_free_task(task);
1090 
1091 	return res;
1092 }
1093 
sas_execute_ssp_tmf(struct domain_device * device,u8 * lun,struct sas_tmf_task * tmf)1094 static int sas_execute_ssp_tmf(struct domain_device *device, u8 *lun,
1095 			       struct sas_tmf_task *tmf)
1096 {
1097 	struct sas_ssp_task ssp_task;
1098 
1099 	if (!(device->tproto & SAS_PROTOCOL_SSP))
1100 		return TMF_RESP_FUNC_ESUPP;
1101 
1102 	memcpy(ssp_task.LUN, lun, 8);
1103 
1104 	return sas_execute_tmf(device, &ssp_task, sizeof(ssp_task), -1, tmf);
1105 }
1106 
sas_abort_task_set(struct domain_device * dev,u8 * lun)1107 int sas_abort_task_set(struct domain_device *dev, u8 *lun)
1108 {
1109 	struct sas_tmf_task tmf_task = {
1110 		.tmf = TMF_ABORT_TASK_SET,
1111 	};
1112 
1113 	return sas_execute_ssp_tmf(dev, lun, &tmf_task);
1114 }
1115 EXPORT_SYMBOL_GPL(sas_abort_task_set);
1116 
sas_clear_task_set(struct domain_device * dev,u8 * lun)1117 int sas_clear_task_set(struct domain_device *dev, u8 *lun)
1118 {
1119 	struct sas_tmf_task tmf_task = {
1120 		.tmf = TMF_CLEAR_TASK_SET,
1121 	};
1122 
1123 	return sas_execute_ssp_tmf(dev, lun, &tmf_task);
1124 }
1125 EXPORT_SYMBOL_GPL(sas_clear_task_set);
1126 
sas_lu_reset(struct domain_device * dev,u8 * lun)1127 int sas_lu_reset(struct domain_device *dev, u8 *lun)
1128 {
1129 	struct sas_tmf_task tmf_task = {
1130 		.tmf = TMF_LU_RESET,
1131 	};
1132 
1133 	return sas_execute_ssp_tmf(dev, lun, &tmf_task);
1134 }
1135 EXPORT_SYMBOL_GPL(sas_lu_reset);
1136 
sas_query_task(struct sas_task * task,u16 tag)1137 int sas_query_task(struct sas_task *task, u16 tag)
1138 {
1139 	struct sas_tmf_task tmf_task = {
1140 		.tmf = TMF_QUERY_TASK,
1141 		.tag_of_task_to_be_managed = tag,
1142 	};
1143 	struct scsi_cmnd *cmnd = task->uldd_task;
1144 	struct domain_device *dev = task->dev;
1145 	struct scsi_lun lun;
1146 
1147 	int_to_scsilun(cmnd->device->lun, &lun);
1148 
1149 	return sas_execute_ssp_tmf(dev, lun.scsi_lun, &tmf_task);
1150 }
1151 EXPORT_SYMBOL_GPL(sas_query_task);
1152 
sas_abort_task(struct sas_task * task,u16 tag)1153 int sas_abort_task(struct sas_task *task, u16 tag)
1154 {
1155 	struct sas_tmf_task tmf_task = {
1156 		.tmf = TMF_ABORT_TASK,
1157 		.tag_of_task_to_be_managed = tag,
1158 	};
1159 	struct scsi_cmnd *cmnd = task->uldd_task;
1160 	struct domain_device *dev = task->dev;
1161 	struct scsi_lun lun;
1162 
1163 	int_to_scsilun(cmnd->device->lun, &lun);
1164 
1165 	return sas_execute_ssp_tmf(dev, lun.scsi_lun, &tmf_task);
1166 }
1167 EXPORT_SYMBOL_GPL(sas_abort_task);
1168 
1169 /*
1170  * Tell an upper layer that it needs to initiate an abort for a given task.
1171  * This should only ever be called by an LLDD.
1172  */
sas_task_abort(struct sas_task * task)1173 void sas_task_abort(struct sas_task *task)
1174 {
1175 	struct scsi_cmnd *sc = task->uldd_task;
1176 
1177 	/* Escape for libsas internal commands */
1178 	if (!sc) {
1179 		struct sas_task_slow *slow = task->slow_task;
1180 
1181 		if (!slow)
1182 			return;
1183 		if (!del_timer(&slow->timer))
1184 			return;
1185 		slow->timer.function(&slow->timer);
1186 		return;
1187 	}
1188 
1189 	if (dev_is_sata(task->dev))
1190 		sas_ata_task_abort(task);
1191 	else
1192 		blk_abort_request(scsi_cmd_to_rq(sc));
1193 }
1194 EXPORT_SYMBOL_GPL(sas_task_abort);
1195 
sas_slave_alloc(struct scsi_device * sdev)1196 int sas_slave_alloc(struct scsi_device *sdev)
1197 {
1198 	if (dev_is_sata(sdev_to_domain_dev(sdev)) && sdev->lun)
1199 		return -ENXIO;
1200 
1201 	return 0;
1202 }
1203 EXPORT_SYMBOL_GPL(sas_slave_alloc);
1204 
sas_target_destroy(struct scsi_target * starget)1205 void sas_target_destroy(struct scsi_target *starget)
1206 {
1207 	struct domain_device *found_dev = starget->hostdata;
1208 
1209 	if (!found_dev)
1210 		return;
1211 
1212 	starget->hostdata = NULL;
1213 	sas_put_device(found_dev);
1214 }
1215 EXPORT_SYMBOL_GPL(sas_target_destroy);
1216 
1217 #define SAS_STRING_ADDR_SIZE	16
1218 
sas_request_addr(struct Scsi_Host * shost,u8 * addr)1219 int sas_request_addr(struct Scsi_Host *shost, u8 *addr)
1220 {
1221 	int res;
1222 	const struct firmware *fw;
1223 
1224 	res = request_firmware(&fw, "sas_addr", &shost->shost_gendev);
1225 	if (res)
1226 		return res;
1227 
1228 	if (fw->size < SAS_STRING_ADDR_SIZE) {
1229 		res = -ENODEV;
1230 		goto out;
1231 	}
1232 
1233 	res = hex2bin(addr, fw->data, strnlen(fw->data, SAS_ADDR_SIZE * 2) / 2);
1234 	if (res)
1235 		goto out;
1236 
1237 out:
1238 	release_firmware(fw);
1239 	return res;
1240 }
1241 EXPORT_SYMBOL_GPL(sas_request_addr);
1242 
1243