xref: /openbmc/linux/drivers/s390/scsi/zfcp_fsf.c (revision 11a163f2)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * zfcp device driver
4  *
5  * Implementation of FSF commands.
6  *
7  * Copyright IBM Corp. 2002, 2020
8  */
9 
10 #define KMSG_COMPONENT "zfcp"
11 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
12 
13 #include <linux/blktrace_api.h>
14 #include <linux/jiffies.h>
15 #include <linux/types.h>
16 #include <linux/slab.h>
17 #include <scsi/fc/fc_els.h>
18 #include "zfcp_ext.h"
19 #include "zfcp_fc.h"
20 #include "zfcp_dbf.h"
21 #include "zfcp_qdio.h"
22 #include "zfcp_reqlist.h"
23 #include "zfcp_diag.h"
24 
25 /* timeout for FSF requests sent during scsi_eh: abort or FCP TMF */
26 #define ZFCP_FSF_SCSI_ER_TIMEOUT (10*HZ)
27 /* timeout for: exchange config/port data outside ERP, or open/close WKA port */
28 #define ZFCP_FSF_REQUEST_TIMEOUT (60*HZ)
29 
30 struct kmem_cache *zfcp_fsf_qtcb_cache;
31 
32 static bool ber_stop = true;
33 module_param(ber_stop, bool, 0600);
34 MODULE_PARM_DESC(ber_stop,
35 		 "Shuts down FCP devices for FCP channels that report a bit-error count in excess of its threshold (default on)");
36 
37 static void zfcp_fsf_request_timeout_handler(struct timer_list *t)
38 {
39 	struct zfcp_fsf_req *fsf_req = from_timer(fsf_req, t, timer);
40 	struct zfcp_adapter *adapter = fsf_req->adapter;
41 
42 	zfcp_qdio_siosl(adapter);
43 	zfcp_erp_adapter_reopen(adapter, ZFCP_STATUS_COMMON_ERP_FAILED,
44 				"fsrth_1");
45 }
46 
47 static void zfcp_fsf_start_timer(struct zfcp_fsf_req *fsf_req,
48 				 unsigned long timeout)
49 {
50 	fsf_req->timer.function = zfcp_fsf_request_timeout_handler;
51 	fsf_req->timer.expires = jiffies + timeout;
52 	add_timer(&fsf_req->timer);
53 }
54 
55 static void zfcp_fsf_start_erp_timer(struct zfcp_fsf_req *fsf_req)
56 {
57 	BUG_ON(!fsf_req->erp_action);
58 	fsf_req->timer.function = zfcp_erp_timeout_handler;
59 	fsf_req->timer.expires = jiffies + 30 * HZ;
60 	add_timer(&fsf_req->timer);
61 }
62 
63 /* association between FSF command and FSF QTCB type */
64 static u32 fsf_qtcb_type[] = {
65 	[FSF_QTCB_FCP_CMND] =             FSF_IO_COMMAND,
66 	[FSF_QTCB_ABORT_FCP_CMND] =       FSF_SUPPORT_COMMAND,
67 	[FSF_QTCB_OPEN_PORT_WITH_DID] =   FSF_SUPPORT_COMMAND,
68 	[FSF_QTCB_OPEN_LUN] =             FSF_SUPPORT_COMMAND,
69 	[FSF_QTCB_CLOSE_LUN] =            FSF_SUPPORT_COMMAND,
70 	[FSF_QTCB_CLOSE_PORT] =           FSF_SUPPORT_COMMAND,
71 	[FSF_QTCB_CLOSE_PHYSICAL_PORT] =  FSF_SUPPORT_COMMAND,
72 	[FSF_QTCB_SEND_ELS] =             FSF_SUPPORT_COMMAND,
73 	[FSF_QTCB_SEND_GENERIC] =         FSF_SUPPORT_COMMAND,
74 	[FSF_QTCB_EXCHANGE_CONFIG_DATA] = FSF_CONFIG_COMMAND,
75 	[FSF_QTCB_EXCHANGE_PORT_DATA] =   FSF_PORT_COMMAND,
76 	[FSF_QTCB_DOWNLOAD_CONTROL_FILE] = FSF_SUPPORT_COMMAND,
77 	[FSF_QTCB_UPLOAD_CONTROL_FILE] =  FSF_SUPPORT_COMMAND
78 };
79 
80 static void zfcp_fsf_class_not_supp(struct zfcp_fsf_req *req)
81 {
82 	dev_err(&req->adapter->ccw_device->dev, "FCP device not "
83 		"operational because of an unsupported FC class\n");
84 	zfcp_erp_adapter_shutdown(req->adapter, 0, "fscns_1");
85 	req->status |= ZFCP_STATUS_FSFREQ_ERROR;
86 }
87 
88 /**
89  * zfcp_fsf_req_free - free memory used by fsf request
90  * @req: pointer to struct zfcp_fsf_req
91  */
92 void zfcp_fsf_req_free(struct zfcp_fsf_req *req)
93 {
94 	if (likely(req->pool)) {
95 		if (likely(!zfcp_fsf_req_is_status_read_buffer(req)))
96 			mempool_free(req->qtcb, req->adapter->pool.qtcb_pool);
97 		mempool_free(req, req->pool);
98 		return;
99 	}
100 
101 	if (likely(!zfcp_fsf_req_is_status_read_buffer(req)))
102 		kmem_cache_free(zfcp_fsf_qtcb_cache, req->qtcb);
103 	kfree(req);
104 }
105 
106 static void zfcp_fsf_status_read_port_closed(struct zfcp_fsf_req *req)
107 {
108 	unsigned long flags;
109 	struct fsf_status_read_buffer *sr_buf = req->data;
110 	struct zfcp_adapter *adapter = req->adapter;
111 	struct zfcp_port *port;
112 	int d_id = ntoh24(sr_buf->d_id);
113 
114 	read_lock_irqsave(&adapter->port_list_lock, flags);
115 	list_for_each_entry(port, &adapter->port_list, list)
116 		if (port->d_id == d_id) {
117 			zfcp_erp_port_reopen(port, 0, "fssrpc1");
118 			break;
119 		}
120 	read_unlock_irqrestore(&adapter->port_list_lock, flags);
121 }
122 
123 void zfcp_fsf_fc_host_link_down(struct zfcp_adapter *adapter)
124 {
125 	struct Scsi_Host *shost = adapter->scsi_host;
126 
127 	adapter->hydra_version = 0;
128 	adapter->peer_wwpn = 0;
129 	adapter->peer_wwnn = 0;
130 	adapter->peer_d_id = 0;
131 
132 	/* if there is no shost yet, we have nothing to zero-out */
133 	if (shost == NULL)
134 		return;
135 
136 	fc_host_port_id(shost) = 0;
137 	fc_host_fabric_name(shost) = 0;
138 	fc_host_speed(shost) = FC_PORTSPEED_UNKNOWN;
139 	fc_host_port_type(shost) = FC_PORTTYPE_UNKNOWN;
140 	snprintf(fc_host_model(shost), FC_SYMBOLIC_NAME_SIZE, "0x%04x", 0);
141 	memset(fc_host_active_fc4s(shost), 0, FC_FC4_LIST_SIZE);
142 }
143 
144 static void zfcp_fsf_link_down_info_eval(struct zfcp_fsf_req *req,
145 					 struct fsf_link_down_info *link_down)
146 {
147 	struct zfcp_adapter *adapter = req->adapter;
148 
149 	if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED)
150 		return;
151 
152 	atomic_or(ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED, &adapter->status);
153 
154 	zfcp_scsi_schedule_rports_block(adapter);
155 
156 	zfcp_fsf_fc_host_link_down(adapter);
157 
158 	if (!link_down)
159 		goto out;
160 
161 	switch (link_down->error_code) {
162 	case FSF_PSQ_LINK_NO_LIGHT:
163 		dev_warn(&req->adapter->ccw_device->dev,
164 			 "There is no light signal from the local "
165 			 "fibre channel cable\n");
166 		break;
167 	case FSF_PSQ_LINK_WRAP_PLUG:
168 		dev_warn(&req->adapter->ccw_device->dev,
169 			 "There is a wrap plug instead of a fibre "
170 			 "channel cable\n");
171 		break;
172 	case FSF_PSQ_LINK_NO_FCP:
173 		dev_warn(&req->adapter->ccw_device->dev,
174 			 "The adjacent fibre channel node does not "
175 			 "support FCP\n");
176 		break;
177 	case FSF_PSQ_LINK_FIRMWARE_UPDATE:
178 		dev_warn(&req->adapter->ccw_device->dev,
179 			 "The FCP device is suspended because of a "
180 			 "firmware update\n");
181 		break;
182 	case FSF_PSQ_LINK_INVALID_WWPN:
183 		dev_warn(&req->adapter->ccw_device->dev,
184 			 "The FCP device detected a WWPN that is "
185 			 "duplicate or not valid\n");
186 		break;
187 	case FSF_PSQ_LINK_NO_NPIV_SUPPORT:
188 		dev_warn(&req->adapter->ccw_device->dev,
189 			 "The fibre channel fabric does not support NPIV\n");
190 		break;
191 	case FSF_PSQ_LINK_NO_FCP_RESOURCES:
192 		dev_warn(&req->adapter->ccw_device->dev,
193 			 "The FCP adapter cannot support more NPIV ports\n");
194 		break;
195 	case FSF_PSQ_LINK_NO_FABRIC_RESOURCES:
196 		dev_warn(&req->adapter->ccw_device->dev,
197 			 "The adjacent switch cannot support "
198 			 "more NPIV ports\n");
199 		break;
200 	case FSF_PSQ_LINK_FABRIC_LOGIN_UNABLE:
201 		dev_warn(&req->adapter->ccw_device->dev,
202 			 "The FCP adapter could not log in to the "
203 			 "fibre channel fabric\n");
204 		break;
205 	case FSF_PSQ_LINK_WWPN_ASSIGNMENT_CORRUPTED:
206 		dev_warn(&req->adapter->ccw_device->dev,
207 			 "The WWPN assignment file on the FCP adapter "
208 			 "has been damaged\n");
209 		break;
210 	case FSF_PSQ_LINK_MODE_TABLE_CURRUPTED:
211 		dev_warn(&req->adapter->ccw_device->dev,
212 			 "The mode table on the FCP adapter "
213 			 "has been damaged\n");
214 		break;
215 	case FSF_PSQ_LINK_NO_WWPN_ASSIGNMENT:
216 		dev_warn(&req->adapter->ccw_device->dev,
217 			 "All NPIV ports on the FCP adapter have "
218 			 "been assigned\n");
219 		break;
220 	default:
221 		dev_warn(&req->adapter->ccw_device->dev,
222 			 "The link between the FCP adapter and "
223 			 "the FC fabric is down\n");
224 	}
225 out:
226 	zfcp_erp_set_adapter_status(adapter, ZFCP_STATUS_COMMON_ERP_FAILED);
227 }
228 
229 static void zfcp_fsf_status_read_link_down(struct zfcp_fsf_req *req)
230 {
231 	struct fsf_status_read_buffer *sr_buf = req->data;
232 	struct fsf_link_down_info *ldi =
233 		(struct fsf_link_down_info *) &sr_buf->payload;
234 
235 	switch (sr_buf->status_subtype) {
236 	case FSF_STATUS_READ_SUB_NO_PHYSICAL_LINK:
237 	case FSF_STATUS_READ_SUB_FDISC_FAILED:
238 		zfcp_fsf_link_down_info_eval(req, ldi);
239 		break;
240 	case FSF_STATUS_READ_SUB_FIRMWARE_UPDATE:
241 		zfcp_fsf_link_down_info_eval(req, NULL);
242 	}
243 }
244 
245 static void zfcp_fsf_status_read_handler(struct zfcp_fsf_req *req)
246 {
247 	struct zfcp_adapter *adapter = req->adapter;
248 	struct fsf_status_read_buffer *sr_buf = req->data;
249 
250 	if (req->status & ZFCP_STATUS_FSFREQ_DISMISSED) {
251 		zfcp_dbf_hba_fsf_uss("fssrh_1", req);
252 		mempool_free(virt_to_page(sr_buf), adapter->pool.sr_data);
253 		zfcp_fsf_req_free(req);
254 		return;
255 	}
256 
257 	zfcp_dbf_hba_fsf_uss("fssrh_4", req);
258 
259 	switch (sr_buf->status_type) {
260 	case FSF_STATUS_READ_PORT_CLOSED:
261 		zfcp_fsf_status_read_port_closed(req);
262 		break;
263 	case FSF_STATUS_READ_INCOMING_ELS:
264 		zfcp_fc_incoming_els(req);
265 		break;
266 	case FSF_STATUS_READ_SENSE_DATA_AVAIL:
267 		break;
268 	case FSF_STATUS_READ_BIT_ERROR_THRESHOLD:
269 		zfcp_dbf_hba_bit_err("fssrh_3", req);
270 		if (ber_stop) {
271 			dev_warn(&adapter->ccw_device->dev,
272 				 "All paths over this FCP device are disused because of excessive bit errors\n");
273 			zfcp_erp_adapter_shutdown(adapter, 0, "fssrh_b");
274 		} else {
275 			dev_warn(&adapter->ccw_device->dev,
276 				 "The error threshold for checksum statistics has been exceeded\n");
277 		}
278 		break;
279 	case FSF_STATUS_READ_LINK_DOWN:
280 		zfcp_fsf_status_read_link_down(req);
281 		zfcp_fc_enqueue_event(adapter, FCH_EVT_LINKDOWN, 0);
282 		break;
283 	case FSF_STATUS_READ_LINK_UP:
284 		dev_info(&adapter->ccw_device->dev,
285 			 "The local link has been restored\n");
286 		/* All ports should be marked as ready to run again */
287 		zfcp_erp_set_adapter_status(adapter,
288 					    ZFCP_STATUS_COMMON_RUNNING);
289 		zfcp_erp_adapter_reopen(adapter,
290 					ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED |
291 					ZFCP_STATUS_COMMON_ERP_FAILED,
292 					"fssrh_2");
293 		zfcp_fc_enqueue_event(adapter, FCH_EVT_LINKUP, 0);
294 
295 		break;
296 	case FSF_STATUS_READ_NOTIFICATION_LOST:
297 		if (sr_buf->status_subtype & FSF_STATUS_READ_SUB_INCOMING_ELS)
298 			zfcp_fc_conditional_port_scan(adapter);
299 		break;
300 	case FSF_STATUS_READ_FEATURE_UPDATE_ALERT:
301 		adapter->adapter_features = sr_buf->payload.word[0];
302 		break;
303 	}
304 
305 	mempool_free(virt_to_page(sr_buf), adapter->pool.sr_data);
306 	zfcp_fsf_req_free(req);
307 
308 	atomic_inc(&adapter->stat_miss);
309 	queue_work(adapter->work_queue, &adapter->stat_work);
310 }
311 
312 static void zfcp_fsf_fsfstatus_qual_eval(struct zfcp_fsf_req *req)
313 {
314 	switch (req->qtcb->header.fsf_status_qual.word[0]) {
315 	case FSF_SQ_FCP_RSP_AVAILABLE:
316 	case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
317 	case FSF_SQ_NO_RETRY_POSSIBLE:
318 	case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
319 		return;
320 	case FSF_SQ_COMMAND_ABORTED:
321 		break;
322 	case FSF_SQ_NO_RECOM:
323 		dev_err(&req->adapter->ccw_device->dev,
324 			"The FCP adapter reported a problem "
325 			"that cannot be recovered\n");
326 		zfcp_qdio_siosl(req->adapter);
327 		zfcp_erp_adapter_shutdown(req->adapter, 0, "fsfsqe1");
328 		break;
329 	}
330 	/* all non-return stats set FSFREQ_ERROR*/
331 	req->status |= ZFCP_STATUS_FSFREQ_ERROR;
332 }
333 
334 static void zfcp_fsf_fsfstatus_eval(struct zfcp_fsf_req *req)
335 {
336 	if (unlikely(req->status & ZFCP_STATUS_FSFREQ_ERROR))
337 		return;
338 
339 	switch (req->qtcb->header.fsf_status) {
340 	case FSF_UNKNOWN_COMMAND:
341 		dev_err(&req->adapter->ccw_device->dev,
342 			"The FCP adapter does not recognize the command 0x%x\n",
343 			req->qtcb->header.fsf_command);
344 		zfcp_erp_adapter_shutdown(req->adapter, 0, "fsfse_1");
345 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
346 		break;
347 	case FSF_ADAPTER_STATUS_AVAILABLE:
348 		zfcp_fsf_fsfstatus_qual_eval(req);
349 		break;
350 	}
351 }
352 
353 static void zfcp_fsf_protstatus_eval(struct zfcp_fsf_req *req)
354 {
355 	struct zfcp_adapter *adapter = req->adapter;
356 	struct fsf_qtcb *qtcb = req->qtcb;
357 	union fsf_prot_status_qual *psq = &qtcb->prefix.prot_status_qual;
358 
359 	zfcp_dbf_hba_fsf_response(req);
360 
361 	if (req->status & ZFCP_STATUS_FSFREQ_DISMISSED) {
362 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
363 		return;
364 	}
365 
366 	switch (qtcb->prefix.prot_status) {
367 	case FSF_PROT_GOOD:
368 	case FSF_PROT_FSF_STATUS_PRESENTED:
369 		return;
370 	case FSF_PROT_QTCB_VERSION_ERROR:
371 		dev_err(&adapter->ccw_device->dev,
372 			"QTCB version 0x%x not supported by FCP adapter "
373 			"(0x%x to 0x%x)\n", FSF_QTCB_CURRENT_VERSION,
374 			psq->word[0], psq->word[1]);
375 		zfcp_erp_adapter_shutdown(adapter, 0, "fspse_1");
376 		break;
377 	case FSF_PROT_ERROR_STATE:
378 	case FSF_PROT_SEQ_NUMB_ERROR:
379 		zfcp_erp_adapter_reopen(adapter, 0, "fspse_2");
380 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
381 		break;
382 	case FSF_PROT_UNSUPP_QTCB_TYPE:
383 		dev_err(&adapter->ccw_device->dev,
384 			"The QTCB type is not supported by the FCP adapter\n");
385 		zfcp_erp_adapter_shutdown(adapter, 0, "fspse_3");
386 		break;
387 	case FSF_PROT_HOST_CONNECTION_INITIALIZING:
388 		atomic_or(ZFCP_STATUS_ADAPTER_HOST_CON_INIT,
389 				&adapter->status);
390 		break;
391 	case FSF_PROT_DUPLICATE_REQUEST_ID:
392 		dev_err(&adapter->ccw_device->dev,
393 			"0x%Lx is an ambiguous request identifier\n",
394 			(unsigned long long)qtcb->bottom.support.req_handle);
395 		zfcp_erp_adapter_shutdown(adapter, 0, "fspse_4");
396 		break;
397 	case FSF_PROT_LINK_DOWN:
398 		zfcp_fsf_link_down_info_eval(req, &psq->link_down_info);
399 		/* go through reopen to flush pending requests */
400 		zfcp_erp_adapter_reopen(adapter, 0, "fspse_6");
401 		break;
402 	case FSF_PROT_REEST_QUEUE:
403 		/* All ports should be marked as ready to run again */
404 		zfcp_erp_set_adapter_status(adapter,
405 					    ZFCP_STATUS_COMMON_RUNNING);
406 		zfcp_erp_adapter_reopen(adapter,
407 					ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED |
408 					ZFCP_STATUS_COMMON_ERP_FAILED,
409 					"fspse_8");
410 		break;
411 	default:
412 		dev_err(&adapter->ccw_device->dev,
413 			"0x%x is not a valid transfer protocol status\n",
414 			qtcb->prefix.prot_status);
415 		zfcp_qdio_siosl(adapter);
416 		zfcp_erp_adapter_shutdown(adapter, 0, "fspse_9");
417 	}
418 	req->status |= ZFCP_STATUS_FSFREQ_ERROR;
419 }
420 
421 /**
422  * zfcp_fsf_req_complete - process completion of a FSF request
423  * @req: The FSF request that has been completed.
424  *
425  * When a request has been completed either from the FCP adapter,
426  * or it has been dismissed due to a queue shutdown, this function
427  * is called to process the completion status and trigger further
428  * events related to the FSF request.
429  * Caller must ensure that the request has been removed from
430  * adapter->req_list, to protect against concurrent modification
431  * by zfcp_erp_strategy_check_fsfreq().
432  */
433 static void zfcp_fsf_req_complete(struct zfcp_fsf_req *req)
434 {
435 	struct zfcp_erp_action *erp_action;
436 
437 	if (unlikely(zfcp_fsf_req_is_status_read_buffer(req))) {
438 		zfcp_fsf_status_read_handler(req);
439 		return;
440 	}
441 
442 	del_timer_sync(&req->timer);
443 	zfcp_fsf_protstatus_eval(req);
444 	zfcp_fsf_fsfstatus_eval(req);
445 	req->handler(req);
446 
447 	erp_action = req->erp_action;
448 	if (erp_action)
449 		zfcp_erp_notify(erp_action, 0);
450 
451 	if (likely(req->status & ZFCP_STATUS_FSFREQ_CLEANUP))
452 		zfcp_fsf_req_free(req);
453 	else
454 		complete(&req->completion);
455 }
456 
457 /**
458  * zfcp_fsf_req_dismiss_all - dismiss all fsf requests
459  * @adapter: pointer to struct zfcp_adapter
460  *
461  * Never ever call this without shutting down the adapter first.
462  * Otherwise the adapter would continue using and corrupting s390 storage.
463  * Included BUG_ON() call to ensure this is done.
464  * ERP is supposed to be the only user of this function.
465  */
466 void zfcp_fsf_req_dismiss_all(struct zfcp_adapter *adapter)
467 {
468 	struct zfcp_fsf_req *req, *tmp;
469 	LIST_HEAD(remove_queue);
470 
471 	BUG_ON(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP);
472 	zfcp_reqlist_move(adapter->req_list, &remove_queue);
473 
474 	list_for_each_entry_safe(req, tmp, &remove_queue, list) {
475 		list_del(&req->list);
476 		req->status |= ZFCP_STATUS_FSFREQ_DISMISSED;
477 		zfcp_fsf_req_complete(req);
478 	}
479 }
480 
481 #define ZFCP_FSF_PORTSPEED_1GBIT	(1 <<  0)
482 #define ZFCP_FSF_PORTSPEED_2GBIT	(1 <<  1)
483 #define ZFCP_FSF_PORTSPEED_4GBIT	(1 <<  2)
484 #define ZFCP_FSF_PORTSPEED_10GBIT	(1 <<  3)
485 #define ZFCP_FSF_PORTSPEED_8GBIT	(1 <<  4)
486 #define ZFCP_FSF_PORTSPEED_16GBIT	(1 <<  5)
487 #define ZFCP_FSF_PORTSPEED_32GBIT	(1 <<  6)
488 #define ZFCP_FSF_PORTSPEED_64GBIT	(1 <<  7)
489 #define ZFCP_FSF_PORTSPEED_128GBIT	(1 <<  8)
490 #define ZFCP_FSF_PORTSPEED_NOT_NEGOTIATED (1 << 15)
491 
492 u32 zfcp_fsf_convert_portspeed(u32 fsf_speed)
493 {
494 	u32 fdmi_speed = 0;
495 	if (fsf_speed & ZFCP_FSF_PORTSPEED_1GBIT)
496 		fdmi_speed |= FC_PORTSPEED_1GBIT;
497 	if (fsf_speed & ZFCP_FSF_PORTSPEED_2GBIT)
498 		fdmi_speed |= FC_PORTSPEED_2GBIT;
499 	if (fsf_speed & ZFCP_FSF_PORTSPEED_4GBIT)
500 		fdmi_speed |= FC_PORTSPEED_4GBIT;
501 	if (fsf_speed & ZFCP_FSF_PORTSPEED_10GBIT)
502 		fdmi_speed |= FC_PORTSPEED_10GBIT;
503 	if (fsf_speed & ZFCP_FSF_PORTSPEED_8GBIT)
504 		fdmi_speed |= FC_PORTSPEED_8GBIT;
505 	if (fsf_speed & ZFCP_FSF_PORTSPEED_16GBIT)
506 		fdmi_speed |= FC_PORTSPEED_16GBIT;
507 	if (fsf_speed & ZFCP_FSF_PORTSPEED_32GBIT)
508 		fdmi_speed |= FC_PORTSPEED_32GBIT;
509 	if (fsf_speed & ZFCP_FSF_PORTSPEED_64GBIT)
510 		fdmi_speed |= FC_PORTSPEED_64GBIT;
511 	if (fsf_speed & ZFCP_FSF_PORTSPEED_128GBIT)
512 		fdmi_speed |= FC_PORTSPEED_128GBIT;
513 	if (fsf_speed & ZFCP_FSF_PORTSPEED_NOT_NEGOTIATED)
514 		fdmi_speed |= FC_PORTSPEED_NOT_NEGOTIATED;
515 	return fdmi_speed;
516 }
517 
518 static int zfcp_fsf_exchange_config_evaluate(struct zfcp_fsf_req *req)
519 {
520 	struct fsf_qtcb_bottom_config *bottom = &req->qtcb->bottom.config;
521 	struct zfcp_adapter *adapter = req->adapter;
522 	struct fc_els_flogi *plogi;
523 
524 	/* adjust pointers for missing command code */
525 	plogi = (struct fc_els_flogi *) ((u8 *)&bottom->plogi_payload
526 					- sizeof(u32));
527 
528 	if (req->data)
529 		memcpy(req->data, bottom, sizeof(*bottom));
530 
531 	adapter->timer_ticks = bottom->timer_interval & ZFCP_FSF_TIMER_INT_MASK;
532 	adapter->stat_read_buf_num = max(bottom->status_read_buf_num,
533 					 (u16)FSF_STATUS_READS_RECOM);
534 
535 	/* no error return above here, otherwise must fix call chains */
536 	/* do not evaluate invalid fields */
537 	if (req->qtcb->header.fsf_status == FSF_EXCHANGE_CONFIG_DATA_INCOMPLETE)
538 		return 0;
539 
540 	adapter->hydra_version = bottom->adapter_type;
541 
542 	switch (bottom->fc_topology) {
543 	case FSF_TOPO_P2P:
544 		adapter->peer_d_id = ntoh24(bottom->peer_d_id);
545 		adapter->peer_wwpn = be64_to_cpu(plogi->fl_wwpn);
546 		adapter->peer_wwnn = be64_to_cpu(plogi->fl_wwnn);
547 		break;
548 	case FSF_TOPO_FABRIC:
549 		break;
550 	case FSF_TOPO_AL:
551 	default:
552 		dev_err(&adapter->ccw_device->dev,
553 			"Unknown or unsupported arbitrated loop "
554 			"fibre channel topology detected\n");
555 		zfcp_erp_adapter_shutdown(adapter, 0, "fsece_1");
556 		return -EIO;
557 	}
558 
559 	return 0;
560 }
561 
562 static void zfcp_fsf_exchange_config_data_handler(struct zfcp_fsf_req *req)
563 {
564 	struct zfcp_adapter *adapter = req->adapter;
565 	struct zfcp_diag_header *const diag_hdr =
566 		&adapter->diagnostics->config_data.header;
567 	struct fsf_qtcb *qtcb = req->qtcb;
568 	struct fsf_qtcb_bottom_config *bottom = &qtcb->bottom.config;
569 
570 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
571 		return;
572 
573 	adapter->fsf_lic_version = bottom->lic_version;
574 	adapter->adapter_features = bottom->adapter_features;
575 	adapter->connection_features = bottom->connection_features;
576 	adapter->peer_wwpn = 0;
577 	adapter->peer_wwnn = 0;
578 	adapter->peer_d_id = 0;
579 
580 	switch (qtcb->header.fsf_status) {
581 	case FSF_GOOD:
582 		/*
583 		 * usually we wait with an update till the cache is too old,
584 		 * but because we have the data available, update it anyway
585 		 */
586 		zfcp_diag_update_xdata(diag_hdr, bottom, false);
587 
588 		zfcp_scsi_shost_update_config_data(adapter, bottom, false);
589 		if (zfcp_fsf_exchange_config_evaluate(req))
590 			return;
591 
592 		if (bottom->max_qtcb_size < sizeof(struct fsf_qtcb)) {
593 			dev_err(&adapter->ccw_device->dev,
594 				"FCP adapter maximum QTCB size (%d bytes) "
595 				"is too small\n",
596 				bottom->max_qtcb_size);
597 			zfcp_erp_adapter_shutdown(adapter, 0, "fsecdh1");
598 			return;
599 		}
600 		atomic_or(ZFCP_STATUS_ADAPTER_XCONFIG_OK,
601 				&adapter->status);
602 		break;
603 	case FSF_EXCHANGE_CONFIG_DATA_INCOMPLETE:
604 		zfcp_diag_update_xdata(diag_hdr, bottom, true);
605 		req->status |= ZFCP_STATUS_FSFREQ_XDATAINCOMPLETE;
606 
607 		/* avoids adapter shutdown to be able to recognize
608 		 * events such as LINK UP */
609 		atomic_or(ZFCP_STATUS_ADAPTER_XCONFIG_OK,
610 				&adapter->status);
611 		zfcp_fsf_link_down_info_eval(req,
612 			&qtcb->header.fsf_status_qual.link_down_info);
613 
614 		zfcp_scsi_shost_update_config_data(adapter, bottom, true);
615 		if (zfcp_fsf_exchange_config_evaluate(req))
616 			return;
617 		break;
618 	default:
619 		zfcp_erp_adapter_shutdown(adapter, 0, "fsecdh3");
620 		return;
621 	}
622 
623 	if (adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT)
624 		adapter->hardware_version = bottom->hardware_version;
625 
626 	if (FSF_QTCB_CURRENT_VERSION < bottom->low_qtcb_version) {
627 		dev_err(&adapter->ccw_device->dev,
628 			"The FCP adapter only supports newer "
629 			"control block versions\n");
630 		zfcp_erp_adapter_shutdown(adapter, 0, "fsecdh4");
631 		return;
632 	}
633 	if (FSF_QTCB_CURRENT_VERSION > bottom->high_qtcb_version) {
634 		dev_err(&adapter->ccw_device->dev,
635 			"The FCP adapter only supports older "
636 			"control block versions\n");
637 		zfcp_erp_adapter_shutdown(adapter, 0, "fsecdh5");
638 	}
639 }
640 
641 /*
642  * Mapping of FC Endpoint Security flag masks to mnemonics
643  *
644  * NOTE: Update macro ZFCP_FSF_MAX_FC_SECURITY_MNEMONIC_LENGTH when making any
645  *       changes.
646  */
647 static const struct {
648 	u32	mask;
649 	char	*name;
650 } zfcp_fsf_fc_security_mnemonics[] = {
651 	{ FSF_FC_SECURITY_AUTH,		"Authentication" },
652 	{ FSF_FC_SECURITY_ENC_FCSP2 |
653 	  FSF_FC_SECURITY_ENC_ERAS,	"Encryption" },
654 };
655 
656 /* maximum strlen(zfcp_fsf_fc_security_mnemonics[...].name) + 1 */
657 #define ZFCP_FSF_MAX_FC_SECURITY_MNEMONIC_LENGTH 15
658 
659 /**
660  * zfcp_fsf_scnprint_fc_security() - translate FC Endpoint Security flags into
661  *                                   mnemonics and place in a buffer
662  * @buf        : the buffer to place the translated FC Endpoint Security flag(s)
663  *               into
664  * @size       : the size of the buffer, including the trailing null space
665  * @fc_security: one or more FC Endpoint Security flags, or zero
666  * @fmt        : specifies whether a list or a single item is to be put into the
667  *               buffer
668  *
669  * The Fibre Channel (FC) Endpoint Security flags are translated into mnemonics.
670  * If the FC Endpoint Security flags are zero "none" is placed into the buffer.
671  *
672  * With ZFCP_FSF_PRINT_FMT_LIST the mnemonics are placed as a list separated by
673  * a comma followed by a space into the buffer. If one or more FC Endpoint
674  * Security flags cannot be translated into a mnemonic, as they are undefined
675  * in zfcp_fsf_fc_security_mnemonics, their bitwise ORed value in hexadecimal
676  * representation is placed into the buffer.
677  *
678  * With ZFCP_FSF_PRINT_FMT_SINGLEITEM only one single mnemonic is placed into
679  * the buffer. If the FC Endpoint Security flag cannot be translated, as it is
680  * undefined in zfcp_fsf_fc_security_mnemonics, its value in hexadecimal
681  * representation is placed into the buffer. If more than one FC Endpoint
682  * Security flag was specified, their value in hexadecimal representation is
683  * placed into the buffer. The macro ZFCP_FSF_MAX_FC_SECURITY_MNEMONIC_LENGTH
684  * can be used to define a buffer that is large enough to hold one mnemonic.
685  *
686  * Return: The number of characters written into buf not including the trailing
687  *         '\0'. If size is == 0 the function returns 0.
688  */
689 ssize_t zfcp_fsf_scnprint_fc_security(char *buf, size_t size, u32 fc_security,
690 				      enum zfcp_fsf_print_fmt fmt)
691 {
692 	const char *prefix = "";
693 	ssize_t len = 0;
694 	int i;
695 
696 	if (fc_security == 0)
697 		return scnprintf(buf, size, "none");
698 	if (fmt == ZFCP_FSF_PRINT_FMT_SINGLEITEM && hweight32(fc_security) != 1)
699 		return scnprintf(buf, size, "0x%08x", fc_security);
700 
701 	for (i = 0; i < ARRAY_SIZE(zfcp_fsf_fc_security_mnemonics); i++) {
702 		if (!(fc_security & zfcp_fsf_fc_security_mnemonics[i].mask))
703 			continue;
704 
705 		len += scnprintf(buf + len, size - len, "%s%s", prefix,
706 				 zfcp_fsf_fc_security_mnemonics[i].name);
707 		prefix = ", ";
708 		fc_security &= ~zfcp_fsf_fc_security_mnemonics[i].mask;
709 	}
710 
711 	if (fc_security != 0)
712 		len += scnprintf(buf + len, size - len, "%s0x%08x",
713 				 prefix, fc_security);
714 
715 	return len;
716 }
717 
718 static void zfcp_fsf_dbf_adapter_fc_security(struct zfcp_adapter *adapter,
719 					     struct zfcp_fsf_req *req)
720 {
721 	if (adapter->fc_security_algorithms ==
722 	    adapter->fc_security_algorithms_old) {
723 		/* no change, no trace */
724 		return;
725 	}
726 
727 	zfcp_dbf_hba_fsf_fces("fsfcesa", req, ZFCP_DBF_INVALID_WWPN,
728 			      adapter->fc_security_algorithms_old,
729 			      adapter->fc_security_algorithms);
730 
731 	adapter->fc_security_algorithms_old = adapter->fc_security_algorithms;
732 }
733 
734 static void zfcp_fsf_exchange_port_evaluate(struct zfcp_fsf_req *req)
735 {
736 	struct zfcp_adapter *adapter = req->adapter;
737 	struct fsf_qtcb_bottom_port *bottom = &req->qtcb->bottom.port;
738 
739 	if (req->data)
740 		memcpy(req->data, bottom, sizeof(*bottom));
741 
742 	if (adapter->adapter_features & FSF_FEATURE_FC_SECURITY)
743 		adapter->fc_security_algorithms =
744 			bottom->fc_security_algorithms;
745 	else
746 		adapter->fc_security_algorithms = 0;
747 	zfcp_fsf_dbf_adapter_fc_security(adapter, req);
748 }
749 
750 static void zfcp_fsf_exchange_port_data_handler(struct zfcp_fsf_req *req)
751 {
752 	struct zfcp_diag_header *const diag_hdr =
753 		&req->adapter->diagnostics->port_data.header;
754 	struct fsf_qtcb *qtcb = req->qtcb;
755 	struct fsf_qtcb_bottom_port *bottom = &qtcb->bottom.port;
756 
757 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
758 		return;
759 
760 	switch (qtcb->header.fsf_status) {
761 	case FSF_GOOD:
762 		/*
763 		 * usually we wait with an update till the cache is too old,
764 		 * but because we have the data available, update it anyway
765 		 */
766 		zfcp_diag_update_xdata(diag_hdr, bottom, false);
767 
768 		zfcp_scsi_shost_update_port_data(req->adapter, bottom);
769 		zfcp_fsf_exchange_port_evaluate(req);
770 		break;
771 	case FSF_EXCHANGE_CONFIG_DATA_INCOMPLETE:
772 		zfcp_diag_update_xdata(diag_hdr, bottom, true);
773 		req->status |= ZFCP_STATUS_FSFREQ_XDATAINCOMPLETE;
774 
775 		zfcp_fsf_link_down_info_eval(req,
776 			&qtcb->header.fsf_status_qual.link_down_info);
777 
778 		zfcp_scsi_shost_update_port_data(req->adapter, bottom);
779 		zfcp_fsf_exchange_port_evaluate(req);
780 		break;
781 	}
782 }
783 
784 static struct zfcp_fsf_req *zfcp_fsf_alloc(mempool_t *pool)
785 {
786 	struct zfcp_fsf_req *req;
787 
788 	if (likely(pool))
789 		req = mempool_alloc(pool, GFP_ATOMIC);
790 	else
791 		req = kmalloc(sizeof(*req), GFP_ATOMIC);
792 
793 	if (unlikely(!req))
794 		return NULL;
795 
796 	memset(req, 0, sizeof(*req));
797 	req->pool = pool;
798 	return req;
799 }
800 
801 static struct fsf_qtcb *zfcp_fsf_qtcb_alloc(mempool_t *pool)
802 {
803 	struct fsf_qtcb *qtcb;
804 
805 	if (likely(pool))
806 		qtcb = mempool_alloc(pool, GFP_ATOMIC);
807 	else
808 		qtcb = kmem_cache_alloc(zfcp_fsf_qtcb_cache, GFP_ATOMIC);
809 
810 	if (unlikely(!qtcb))
811 		return NULL;
812 
813 	memset(qtcb, 0, sizeof(*qtcb));
814 	return qtcb;
815 }
816 
817 static struct zfcp_fsf_req *zfcp_fsf_req_create(struct zfcp_qdio *qdio,
818 						u32 fsf_cmd, u8 sbtype,
819 						mempool_t *pool)
820 {
821 	struct zfcp_adapter *adapter = qdio->adapter;
822 	struct zfcp_fsf_req *req = zfcp_fsf_alloc(pool);
823 
824 	if (unlikely(!req))
825 		return ERR_PTR(-ENOMEM);
826 
827 	if (adapter->req_no == 0)
828 		adapter->req_no++;
829 
830 	INIT_LIST_HEAD(&req->list);
831 	timer_setup(&req->timer, NULL, 0);
832 	init_completion(&req->completion);
833 
834 	req->adapter = adapter;
835 	req->req_id = adapter->req_no;
836 
837 	if (likely(fsf_cmd != FSF_QTCB_UNSOLICITED_STATUS)) {
838 		if (likely(pool))
839 			req->qtcb = zfcp_fsf_qtcb_alloc(
840 				adapter->pool.qtcb_pool);
841 		else
842 			req->qtcb = zfcp_fsf_qtcb_alloc(NULL);
843 
844 		if (unlikely(!req->qtcb)) {
845 			zfcp_fsf_req_free(req);
846 			return ERR_PTR(-ENOMEM);
847 		}
848 
849 		req->qtcb->prefix.req_seq_no = adapter->fsf_req_seq_no;
850 		req->qtcb->prefix.req_id = req->req_id;
851 		req->qtcb->prefix.ulp_info = 26;
852 		req->qtcb->prefix.qtcb_type = fsf_qtcb_type[fsf_cmd];
853 		req->qtcb->prefix.qtcb_version = FSF_QTCB_CURRENT_VERSION;
854 		req->qtcb->header.req_handle = req->req_id;
855 		req->qtcb->header.fsf_command = fsf_cmd;
856 	}
857 
858 	zfcp_qdio_req_init(adapter->qdio, &req->qdio_req, req->req_id, sbtype,
859 			   req->qtcb, sizeof(struct fsf_qtcb));
860 
861 	return req;
862 }
863 
864 static int zfcp_fsf_req_send(struct zfcp_fsf_req *req)
865 {
866 	const bool is_srb = zfcp_fsf_req_is_status_read_buffer(req);
867 	struct zfcp_adapter *adapter = req->adapter;
868 	struct zfcp_qdio *qdio = adapter->qdio;
869 	int req_id = req->req_id;
870 
871 	zfcp_reqlist_add(adapter->req_list, req);
872 
873 	req->qdio_req.qdio_outb_usage = atomic_read(&qdio->req_q_free);
874 	req->issued = get_tod_clock();
875 	if (zfcp_qdio_send(qdio, &req->qdio_req)) {
876 		del_timer_sync(&req->timer);
877 		/* lookup request again, list might have changed */
878 		zfcp_reqlist_find_rm(adapter->req_list, req_id);
879 		zfcp_erp_adapter_reopen(adapter, 0, "fsrs__1");
880 		return -EIO;
881 	}
882 
883 	/*
884 	 * NOTE: DO NOT TOUCH ASYNC req PAST THIS POINT.
885 	 *	 ONLY TOUCH SYNC req AGAIN ON req->completion.
886 	 *
887 	 * The request might complete and be freed concurrently at any point
888 	 * now. This is not protected by the QDIO-lock (req_q_lock). So any
889 	 * uncontrolled access after this might result in an use-after-free bug.
890 	 * Only if the request doesn't have ZFCP_STATUS_FSFREQ_CLEANUP set, and
891 	 * when it is completed via req->completion, is it safe to use req
892 	 * again.
893 	 */
894 
895 	/* Don't increase for unsolicited status */
896 	if (!is_srb)
897 		adapter->fsf_req_seq_no++;
898 	adapter->req_no++;
899 
900 	return 0;
901 }
902 
903 /**
904  * zfcp_fsf_status_read - send status read request
905  * @qdio: pointer to struct zfcp_qdio
906  * Returns: 0 on success, ERROR otherwise
907  */
908 int zfcp_fsf_status_read(struct zfcp_qdio *qdio)
909 {
910 	struct zfcp_adapter *adapter = qdio->adapter;
911 	struct zfcp_fsf_req *req;
912 	struct fsf_status_read_buffer *sr_buf;
913 	struct page *page;
914 	int retval = -EIO;
915 
916 	spin_lock_irq(&qdio->req_q_lock);
917 	if (zfcp_qdio_sbal_get(qdio))
918 		goto out;
919 
920 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_UNSOLICITED_STATUS,
921 				  SBAL_SFLAGS0_TYPE_STATUS,
922 				  adapter->pool.status_read_req);
923 	if (IS_ERR(req)) {
924 		retval = PTR_ERR(req);
925 		goto out;
926 	}
927 
928 	page = mempool_alloc(adapter->pool.sr_data, GFP_ATOMIC);
929 	if (!page) {
930 		retval = -ENOMEM;
931 		goto failed_buf;
932 	}
933 	sr_buf = page_address(page);
934 	memset(sr_buf, 0, sizeof(*sr_buf));
935 	req->data = sr_buf;
936 
937 	zfcp_qdio_fill_next(qdio, &req->qdio_req, sr_buf, sizeof(*sr_buf));
938 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
939 
940 	retval = zfcp_fsf_req_send(req);
941 	if (retval)
942 		goto failed_req_send;
943 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
944 
945 	goto out;
946 
947 failed_req_send:
948 	req->data = NULL;
949 	mempool_free(virt_to_page(sr_buf), adapter->pool.sr_data);
950 failed_buf:
951 	zfcp_dbf_hba_fsf_uss("fssr__1", req);
952 	zfcp_fsf_req_free(req);
953 out:
954 	spin_unlock_irq(&qdio->req_q_lock);
955 	return retval;
956 }
957 
958 static void zfcp_fsf_abort_fcp_command_handler(struct zfcp_fsf_req *req)
959 {
960 	struct scsi_device *sdev = req->data;
961 	struct zfcp_scsi_dev *zfcp_sdev;
962 	union fsf_status_qual *fsq = &req->qtcb->header.fsf_status_qual;
963 
964 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
965 		return;
966 
967 	zfcp_sdev = sdev_to_zfcp(sdev);
968 
969 	switch (req->qtcb->header.fsf_status) {
970 	case FSF_PORT_HANDLE_NOT_VALID:
971 		if (fsq->word[0] == fsq->word[1]) {
972 			zfcp_erp_adapter_reopen(zfcp_sdev->port->adapter, 0,
973 						"fsafch1");
974 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
975 		}
976 		break;
977 	case FSF_LUN_HANDLE_NOT_VALID:
978 		if (fsq->word[0] == fsq->word[1]) {
979 			zfcp_erp_port_reopen(zfcp_sdev->port, 0, "fsafch2");
980 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
981 		}
982 		break;
983 	case FSF_FCP_COMMAND_DOES_NOT_EXIST:
984 		req->status |= ZFCP_STATUS_FSFREQ_ABORTNOTNEEDED;
985 		break;
986 	case FSF_PORT_BOXED:
987 		zfcp_erp_set_port_status(zfcp_sdev->port,
988 					 ZFCP_STATUS_COMMON_ACCESS_BOXED);
989 		zfcp_erp_port_reopen(zfcp_sdev->port,
990 				     ZFCP_STATUS_COMMON_ERP_FAILED, "fsafch3");
991 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
992 		break;
993 	case FSF_LUN_BOXED:
994 		zfcp_erp_set_lun_status(sdev, ZFCP_STATUS_COMMON_ACCESS_BOXED);
995 		zfcp_erp_lun_reopen(sdev, ZFCP_STATUS_COMMON_ERP_FAILED,
996 				    "fsafch4");
997 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
998                 break;
999 	case FSF_ADAPTER_STATUS_AVAILABLE:
1000 		switch (fsq->word[0]) {
1001 		case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
1002 			zfcp_fc_test_link(zfcp_sdev->port);
1003 			fallthrough;
1004 		case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
1005 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1006 			break;
1007 		}
1008 		break;
1009 	case FSF_GOOD:
1010 		req->status |= ZFCP_STATUS_FSFREQ_ABORTSUCCEEDED;
1011 		break;
1012 	}
1013 }
1014 
1015 /**
1016  * zfcp_fsf_abort_fcp_cmnd - abort running SCSI command
1017  * @scmnd: The SCSI command to abort
1018  * Returns: pointer to struct zfcp_fsf_req
1019  */
1020 
1021 struct zfcp_fsf_req *zfcp_fsf_abort_fcp_cmnd(struct scsi_cmnd *scmnd)
1022 {
1023 	struct zfcp_fsf_req *req = NULL;
1024 	struct scsi_device *sdev = scmnd->device;
1025 	struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(sdev);
1026 	struct zfcp_qdio *qdio = zfcp_sdev->port->adapter->qdio;
1027 	unsigned long old_req_id = (unsigned long) scmnd->host_scribble;
1028 
1029 	spin_lock_irq(&qdio->req_q_lock);
1030 	if (zfcp_qdio_sbal_get(qdio))
1031 		goto out;
1032 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_ABORT_FCP_CMND,
1033 				  SBAL_SFLAGS0_TYPE_READ,
1034 				  qdio->adapter->pool.scsi_abort);
1035 	if (IS_ERR(req)) {
1036 		req = NULL;
1037 		goto out;
1038 	}
1039 
1040 	if (unlikely(!(atomic_read(&zfcp_sdev->status) &
1041 		       ZFCP_STATUS_COMMON_UNBLOCKED)))
1042 		goto out_error_free;
1043 
1044 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1045 
1046 	req->data = sdev;
1047 	req->handler = zfcp_fsf_abort_fcp_command_handler;
1048 	req->qtcb->header.lun_handle = zfcp_sdev->lun_handle;
1049 	req->qtcb->header.port_handle = zfcp_sdev->port->handle;
1050 	req->qtcb->bottom.support.req_handle = (u64) old_req_id;
1051 
1052 	zfcp_fsf_start_timer(req, ZFCP_FSF_SCSI_ER_TIMEOUT);
1053 	if (!zfcp_fsf_req_send(req)) {
1054 		/* NOTE: DO NOT TOUCH req, UNTIL IT COMPLETES! */
1055 		goto out;
1056 	}
1057 
1058 out_error_free:
1059 	zfcp_fsf_req_free(req);
1060 	req = NULL;
1061 out:
1062 	spin_unlock_irq(&qdio->req_q_lock);
1063 	return req;
1064 }
1065 
1066 static void zfcp_fsf_send_ct_handler(struct zfcp_fsf_req *req)
1067 {
1068 	struct zfcp_adapter *adapter = req->adapter;
1069 	struct zfcp_fsf_ct_els *ct = req->data;
1070 	struct fsf_qtcb_header *header = &req->qtcb->header;
1071 
1072 	ct->status = -EINVAL;
1073 
1074 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
1075 		goto skip_fsfstatus;
1076 
1077 	switch (header->fsf_status) {
1078         case FSF_GOOD:
1079 		ct->status = 0;
1080 		zfcp_dbf_san_res("fsscth2", req);
1081 		break;
1082         case FSF_SERVICE_CLASS_NOT_SUPPORTED:
1083 		zfcp_fsf_class_not_supp(req);
1084 		break;
1085         case FSF_ADAPTER_STATUS_AVAILABLE:
1086                 switch (header->fsf_status_qual.word[0]){
1087                 case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
1088                 case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
1089 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1090 			break;
1091                 }
1092                 break;
1093         case FSF_PORT_BOXED:
1094 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1095 		break;
1096 	case FSF_PORT_HANDLE_NOT_VALID:
1097 		zfcp_erp_adapter_reopen(adapter, 0, "fsscth1");
1098 		fallthrough;
1099 	case FSF_GENERIC_COMMAND_REJECTED:
1100 	case FSF_PAYLOAD_SIZE_MISMATCH:
1101 	case FSF_REQUEST_SIZE_TOO_LARGE:
1102 	case FSF_RESPONSE_SIZE_TOO_LARGE:
1103 	case FSF_SBAL_MISMATCH:
1104 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1105 		break;
1106 	}
1107 
1108 skip_fsfstatus:
1109 	if (ct->handler)
1110 		ct->handler(ct->handler_data);
1111 }
1112 
1113 static void zfcp_fsf_setup_ct_els_unchained(struct zfcp_qdio *qdio,
1114 					    struct zfcp_qdio_req *q_req,
1115 					    struct scatterlist *sg_req,
1116 					    struct scatterlist *sg_resp)
1117 {
1118 	zfcp_qdio_fill_next(qdio, q_req, sg_virt(sg_req), sg_req->length);
1119 	zfcp_qdio_fill_next(qdio, q_req, sg_virt(sg_resp), sg_resp->length);
1120 	zfcp_qdio_set_sbale_last(qdio, q_req);
1121 }
1122 
1123 static int zfcp_fsf_setup_ct_els_sbals(struct zfcp_fsf_req *req,
1124 				       struct scatterlist *sg_req,
1125 				       struct scatterlist *sg_resp)
1126 {
1127 	struct zfcp_adapter *adapter = req->adapter;
1128 	struct zfcp_qdio *qdio = adapter->qdio;
1129 	struct fsf_qtcb *qtcb = req->qtcb;
1130 	u32 feat = adapter->adapter_features;
1131 
1132 	if (zfcp_adapter_multi_buffer_active(adapter)) {
1133 		if (zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, sg_req))
1134 			return -EIO;
1135 		qtcb->bottom.support.req_buf_length =
1136 			zfcp_qdio_real_bytes(sg_req);
1137 		if (zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, sg_resp))
1138 			return -EIO;
1139 		qtcb->bottom.support.resp_buf_length =
1140 			zfcp_qdio_real_bytes(sg_resp);
1141 
1142 		zfcp_qdio_set_data_div(qdio, &req->qdio_req, sg_nents(sg_req));
1143 		zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1144 		zfcp_qdio_set_scount(qdio, &req->qdio_req);
1145 		return 0;
1146 	}
1147 
1148 	/* use single, unchained SBAL if it can hold the request */
1149 	if (zfcp_qdio_sg_one_sbale(sg_req) && zfcp_qdio_sg_one_sbale(sg_resp)) {
1150 		zfcp_fsf_setup_ct_els_unchained(qdio, &req->qdio_req,
1151 						sg_req, sg_resp);
1152 		return 0;
1153 	}
1154 
1155 	if (!(feat & FSF_FEATURE_ELS_CT_CHAINED_SBALS))
1156 		return -EOPNOTSUPP;
1157 
1158 	if (zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, sg_req))
1159 		return -EIO;
1160 
1161 	qtcb->bottom.support.req_buf_length = zfcp_qdio_real_bytes(sg_req);
1162 
1163 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1164 	zfcp_qdio_skip_to_last_sbale(qdio, &req->qdio_req);
1165 
1166 	if (zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, sg_resp))
1167 		return -EIO;
1168 
1169 	qtcb->bottom.support.resp_buf_length = zfcp_qdio_real_bytes(sg_resp);
1170 
1171 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1172 
1173 	return 0;
1174 }
1175 
1176 static int zfcp_fsf_setup_ct_els(struct zfcp_fsf_req *req,
1177 				 struct scatterlist *sg_req,
1178 				 struct scatterlist *sg_resp,
1179 				 unsigned int timeout)
1180 {
1181 	int ret;
1182 
1183 	ret = zfcp_fsf_setup_ct_els_sbals(req, sg_req, sg_resp);
1184 	if (ret)
1185 		return ret;
1186 
1187 	/* common settings for ct/gs and els requests */
1188 	if (timeout > 255)
1189 		timeout = 255; /* max value accepted by hardware */
1190 	req->qtcb->bottom.support.service_class = FSF_CLASS_3;
1191 	req->qtcb->bottom.support.timeout = timeout;
1192 	zfcp_fsf_start_timer(req, (timeout + 10) * HZ);
1193 
1194 	return 0;
1195 }
1196 
1197 /**
1198  * zfcp_fsf_send_ct - initiate a Generic Service request (FC-GS)
1199  * @wka_port: pointer to zfcp WKA port to send CT/GS to
1200  * @ct: pointer to struct zfcp_send_ct with data for request
1201  * @pool: if non-null this mempool is used to allocate struct zfcp_fsf_req
1202  * @timeout: timeout that hardware should use, and a later software timeout
1203  */
1204 int zfcp_fsf_send_ct(struct zfcp_fc_wka_port *wka_port,
1205 		     struct zfcp_fsf_ct_els *ct, mempool_t *pool,
1206 		     unsigned int timeout)
1207 {
1208 	struct zfcp_qdio *qdio = wka_port->adapter->qdio;
1209 	struct zfcp_fsf_req *req;
1210 	int ret = -EIO;
1211 
1212 	spin_lock_irq(&qdio->req_q_lock);
1213 	if (zfcp_qdio_sbal_get(qdio))
1214 		goto out;
1215 
1216 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_SEND_GENERIC,
1217 				  SBAL_SFLAGS0_TYPE_WRITE_READ, pool);
1218 
1219 	if (IS_ERR(req)) {
1220 		ret = PTR_ERR(req);
1221 		goto out;
1222 	}
1223 
1224 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1225 	ret = zfcp_fsf_setup_ct_els(req, ct->req, ct->resp, timeout);
1226 	if (ret)
1227 		goto failed_send;
1228 
1229 	req->handler = zfcp_fsf_send_ct_handler;
1230 	req->qtcb->header.port_handle = wka_port->handle;
1231 	ct->d_id = wka_port->d_id;
1232 	req->data = ct;
1233 
1234 	zfcp_dbf_san_req("fssct_1", req, wka_port->d_id);
1235 
1236 	ret = zfcp_fsf_req_send(req);
1237 	if (ret)
1238 		goto failed_send;
1239 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1240 
1241 	goto out;
1242 
1243 failed_send:
1244 	zfcp_fsf_req_free(req);
1245 out:
1246 	spin_unlock_irq(&qdio->req_q_lock);
1247 	return ret;
1248 }
1249 
1250 static void zfcp_fsf_send_els_handler(struct zfcp_fsf_req *req)
1251 {
1252 	struct zfcp_fsf_ct_els *send_els = req->data;
1253 	struct fsf_qtcb_header *header = &req->qtcb->header;
1254 
1255 	send_els->status = -EINVAL;
1256 
1257 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
1258 		goto skip_fsfstatus;
1259 
1260 	switch (header->fsf_status) {
1261 	case FSF_GOOD:
1262 		send_els->status = 0;
1263 		zfcp_dbf_san_res("fsselh1", req);
1264 		break;
1265 	case FSF_SERVICE_CLASS_NOT_SUPPORTED:
1266 		zfcp_fsf_class_not_supp(req);
1267 		break;
1268 	case FSF_ADAPTER_STATUS_AVAILABLE:
1269 		switch (header->fsf_status_qual.word[0]){
1270 		case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
1271 		case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
1272 		case FSF_SQ_RETRY_IF_POSSIBLE:
1273 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1274 			break;
1275 		}
1276 		break;
1277 	case FSF_ELS_COMMAND_REJECTED:
1278 	case FSF_PAYLOAD_SIZE_MISMATCH:
1279 	case FSF_REQUEST_SIZE_TOO_LARGE:
1280 	case FSF_RESPONSE_SIZE_TOO_LARGE:
1281 		break;
1282 	case FSF_SBAL_MISMATCH:
1283 		/* should never occur, avoided in zfcp_fsf_send_els */
1284 		fallthrough;
1285 	default:
1286 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1287 		break;
1288 	}
1289 skip_fsfstatus:
1290 	if (send_els->handler)
1291 		send_els->handler(send_els->handler_data);
1292 }
1293 
1294 /**
1295  * zfcp_fsf_send_els - initiate an ELS command (FC-FS)
1296  * @adapter: pointer to zfcp adapter
1297  * @d_id: N_Port_ID to send ELS to
1298  * @els: pointer to struct zfcp_send_els with data for the command
1299  * @timeout: timeout that hardware should use, and a later software timeout
1300  */
1301 int zfcp_fsf_send_els(struct zfcp_adapter *adapter, u32 d_id,
1302 		      struct zfcp_fsf_ct_els *els, unsigned int timeout)
1303 {
1304 	struct zfcp_fsf_req *req;
1305 	struct zfcp_qdio *qdio = adapter->qdio;
1306 	int ret = -EIO;
1307 
1308 	spin_lock_irq(&qdio->req_q_lock);
1309 	if (zfcp_qdio_sbal_get(qdio))
1310 		goto out;
1311 
1312 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_SEND_ELS,
1313 				  SBAL_SFLAGS0_TYPE_WRITE_READ, NULL);
1314 
1315 	if (IS_ERR(req)) {
1316 		ret = PTR_ERR(req);
1317 		goto out;
1318 	}
1319 
1320 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1321 
1322 	if (!zfcp_adapter_multi_buffer_active(adapter))
1323 		zfcp_qdio_sbal_limit(qdio, &req->qdio_req, 2);
1324 
1325 	ret = zfcp_fsf_setup_ct_els(req, els->req, els->resp, timeout);
1326 
1327 	if (ret)
1328 		goto failed_send;
1329 
1330 	hton24(req->qtcb->bottom.support.d_id, d_id);
1331 	req->handler = zfcp_fsf_send_els_handler;
1332 	els->d_id = d_id;
1333 	req->data = els;
1334 
1335 	zfcp_dbf_san_req("fssels1", req, d_id);
1336 
1337 	ret = zfcp_fsf_req_send(req);
1338 	if (ret)
1339 		goto failed_send;
1340 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1341 
1342 	goto out;
1343 
1344 failed_send:
1345 	zfcp_fsf_req_free(req);
1346 out:
1347 	spin_unlock_irq(&qdio->req_q_lock);
1348 	return ret;
1349 }
1350 
1351 int zfcp_fsf_exchange_config_data(struct zfcp_erp_action *erp_action)
1352 {
1353 	struct zfcp_fsf_req *req;
1354 	struct zfcp_qdio *qdio = erp_action->adapter->qdio;
1355 	int retval = -EIO;
1356 
1357 	spin_lock_irq(&qdio->req_q_lock);
1358 	if (zfcp_qdio_sbal_get(qdio))
1359 		goto out;
1360 
1361 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_CONFIG_DATA,
1362 				  SBAL_SFLAGS0_TYPE_READ,
1363 				  qdio->adapter->pool.erp_req);
1364 
1365 	if (IS_ERR(req)) {
1366 		retval = PTR_ERR(req);
1367 		goto out;
1368 	}
1369 
1370 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1371 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1372 
1373 	req->qtcb->bottom.config.feature_selection =
1374 			FSF_FEATURE_NOTIFICATION_LOST |
1375 			FSF_FEATURE_UPDATE_ALERT |
1376 			FSF_FEATURE_REQUEST_SFP_DATA |
1377 			FSF_FEATURE_FC_SECURITY;
1378 	req->erp_action = erp_action;
1379 	req->handler = zfcp_fsf_exchange_config_data_handler;
1380 	erp_action->fsf_req_id = req->req_id;
1381 
1382 	zfcp_fsf_start_erp_timer(req);
1383 	retval = zfcp_fsf_req_send(req);
1384 	if (retval) {
1385 		zfcp_fsf_req_free(req);
1386 		erp_action->fsf_req_id = 0;
1387 	}
1388 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1389 out:
1390 	spin_unlock_irq(&qdio->req_q_lock);
1391 	return retval;
1392 }
1393 
1394 
1395 /**
1396  * zfcp_fsf_exchange_config_data_sync() - Request information about FCP channel.
1397  * @qdio: pointer to the QDIO-Queue to use for sending the command.
1398  * @data: pointer to the QTCB-Bottom for storing the result of the command,
1399  *	  might be %NULL.
1400  *
1401  * Returns:
1402  * * 0		- Exchange Config Data was successful, @data is complete
1403  * * -EIO	- Exchange Config Data was not successful, @data is invalid
1404  * * -EAGAIN	- @data contains incomplete data
1405  * * -ENOMEM	- Some memory allocation failed along the way
1406  */
1407 int zfcp_fsf_exchange_config_data_sync(struct zfcp_qdio *qdio,
1408 				       struct fsf_qtcb_bottom_config *data)
1409 {
1410 	struct zfcp_fsf_req *req = NULL;
1411 	int retval = -EIO;
1412 
1413 	spin_lock_irq(&qdio->req_q_lock);
1414 	if (zfcp_qdio_sbal_get(qdio))
1415 		goto out_unlock;
1416 
1417 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_CONFIG_DATA,
1418 				  SBAL_SFLAGS0_TYPE_READ, NULL);
1419 
1420 	if (IS_ERR(req)) {
1421 		retval = PTR_ERR(req);
1422 		goto out_unlock;
1423 	}
1424 
1425 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1426 	req->handler = zfcp_fsf_exchange_config_data_handler;
1427 
1428 	req->qtcb->bottom.config.feature_selection =
1429 			FSF_FEATURE_NOTIFICATION_LOST |
1430 			FSF_FEATURE_UPDATE_ALERT |
1431 			FSF_FEATURE_REQUEST_SFP_DATA |
1432 			FSF_FEATURE_FC_SECURITY;
1433 
1434 	if (data)
1435 		req->data = data;
1436 
1437 	zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT);
1438 	retval = zfcp_fsf_req_send(req);
1439 	spin_unlock_irq(&qdio->req_q_lock);
1440 
1441 	if (!retval) {
1442 		/* NOTE: ONLY TOUCH SYNC req AGAIN ON req->completion. */
1443 		wait_for_completion(&req->completion);
1444 
1445 		if (req->status &
1446 		    (ZFCP_STATUS_FSFREQ_ERROR | ZFCP_STATUS_FSFREQ_DISMISSED))
1447 			retval = -EIO;
1448 		else if (req->status & ZFCP_STATUS_FSFREQ_XDATAINCOMPLETE)
1449 			retval = -EAGAIN;
1450 	}
1451 
1452 	zfcp_fsf_req_free(req);
1453 	return retval;
1454 
1455 out_unlock:
1456 	spin_unlock_irq(&qdio->req_q_lock);
1457 	return retval;
1458 }
1459 
1460 /**
1461  * zfcp_fsf_exchange_port_data - request information about local port
1462  * @erp_action: ERP action for the adapter for which port data is requested
1463  * Returns: 0 on success, error otherwise
1464  */
1465 int zfcp_fsf_exchange_port_data(struct zfcp_erp_action *erp_action)
1466 {
1467 	struct zfcp_qdio *qdio = erp_action->adapter->qdio;
1468 	struct zfcp_fsf_req *req;
1469 	int retval = -EIO;
1470 
1471 	if (!(qdio->adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT))
1472 		return -EOPNOTSUPP;
1473 
1474 	spin_lock_irq(&qdio->req_q_lock);
1475 	if (zfcp_qdio_sbal_get(qdio))
1476 		goto out;
1477 
1478 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_PORT_DATA,
1479 				  SBAL_SFLAGS0_TYPE_READ,
1480 				  qdio->adapter->pool.erp_req);
1481 
1482 	if (IS_ERR(req)) {
1483 		retval = PTR_ERR(req);
1484 		goto out;
1485 	}
1486 
1487 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1488 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1489 
1490 	req->handler = zfcp_fsf_exchange_port_data_handler;
1491 	req->erp_action = erp_action;
1492 	erp_action->fsf_req_id = req->req_id;
1493 
1494 	zfcp_fsf_start_erp_timer(req);
1495 	retval = zfcp_fsf_req_send(req);
1496 	if (retval) {
1497 		zfcp_fsf_req_free(req);
1498 		erp_action->fsf_req_id = 0;
1499 	}
1500 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1501 out:
1502 	spin_unlock_irq(&qdio->req_q_lock);
1503 	return retval;
1504 }
1505 
1506 /**
1507  * zfcp_fsf_exchange_port_data_sync() - Request information about local port.
1508  * @qdio: pointer to the QDIO-Queue to use for sending the command.
1509  * @data: pointer to the QTCB-Bottom for storing the result of the command,
1510  *	  might be %NULL.
1511  *
1512  * Returns:
1513  * * 0		- Exchange Port Data was successful, @data is complete
1514  * * -EIO	- Exchange Port Data was not successful, @data is invalid
1515  * * -EAGAIN	- @data contains incomplete data
1516  * * -ENOMEM	- Some memory allocation failed along the way
1517  * * -EOPNOTSUPP	- This operation is not supported
1518  */
1519 int zfcp_fsf_exchange_port_data_sync(struct zfcp_qdio *qdio,
1520 				     struct fsf_qtcb_bottom_port *data)
1521 {
1522 	struct zfcp_fsf_req *req = NULL;
1523 	int retval = -EIO;
1524 
1525 	if (!(qdio->adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT))
1526 		return -EOPNOTSUPP;
1527 
1528 	spin_lock_irq(&qdio->req_q_lock);
1529 	if (zfcp_qdio_sbal_get(qdio))
1530 		goto out_unlock;
1531 
1532 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_PORT_DATA,
1533 				  SBAL_SFLAGS0_TYPE_READ, NULL);
1534 
1535 	if (IS_ERR(req)) {
1536 		retval = PTR_ERR(req);
1537 		goto out_unlock;
1538 	}
1539 
1540 	if (data)
1541 		req->data = data;
1542 
1543 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1544 
1545 	req->handler = zfcp_fsf_exchange_port_data_handler;
1546 	zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT);
1547 	retval = zfcp_fsf_req_send(req);
1548 	spin_unlock_irq(&qdio->req_q_lock);
1549 
1550 	if (!retval) {
1551 		/* NOTE: ONLY TOUCH SYNC req AGAIN ON req->completion. */
1552 		wait_for_completion(&req->completion);
1553 
1554 		if (req->status &
1555 		    (ZFCP_STATUS_FSFREQ_ERROR | ZFCP_STATUS_FSFREQ_DISMISSED))
1556 			retval = -EIO;
1557 		else if (req->status & ZFCP_STATUS_FSFREQ_XDATAINCOMPLETE)
1558 			retval = -EAGAIN;
1559 	}
1560 
1561 	zfcp_fsf_req_free(req);
1562 	return retval;
1563 
1564 out_unlock:
1565 	spin_unlock_irq(&qdio->req_q_lock);
1566 	return retval;
1567 }
1568 
1569 static void zfcp_fsf_log_port_fc_security(struct zfcp_port *port,
1570 					  struct zfcp_fsf_req *req)
1571 {
1572 	char mnemonic_old[ZFCP_FSF_MAX_FC_SECURITY_MNEMONIC_LENGTH];
1573 	char mnemonic_new[ZFCP_FSF_MAX_FC_SECURITY_MNEMONIC_LENGTH];
1574 
1575 	if (port->connection_info == port->connection_info_old) {
1576 		/* no change, no log nor trace */
1577 		return;
1578 	}
1579 
1580 	zfcp_dbf_hba_fsf_fces("fsfcesp", req, port->wwpn,
1581 			      port->connection_info_old,
1582 			      port->connection_info);
1583 
1584 	zfcp_fsf_scnprint_fc_security(mnemonic_old, sizeof(mnemonic_old),
1585 				      port->connection_info_old,
1586 				      ZFCP_FSF_PRINT_FMT_SINGLEITEM);
1587 	zfcp_fsf_scnprint_fc_security(mnemonic_new, sizeof(mnemonic_new),
1588 				      port->connection_info,
1589 				      ZFCP_FSF_PRINT_FMT_SINGLEITEM);
1590 
1591 	if (strncmp(mnemonic_old, mnemonic_new,
1592 		    ZFCP_FSF_MAX_FC_SECURITY_MNEMONIC_LENGTH) == 0) {
1593 		/* no change in string representation, no log */
1594 		goto out;
1595 	}
1596 
1597 	if (port->connection_info_old == 0) {
1598 		/* activation */
1599 		dev_info(&port->adapter->ccw_device->dev,
1600 			 "FC Endpoint Security of connection to remote port 0x%16llx enabled: %s\n",
1601 			 port->wwpn, mnemonic_new);
1602 	} else if (port->connection_info == 0) {
1603 		/* deactivation */
1604 		dev_warn(&port->adapter->ccw_device->dev,
1605 			 "FC Endpoint Security of connection to remote port 0x%16llx disabled: was %s\n",
1606 			 port->wwpn, mnemonic_old);
1607 	} else {
1608 		/* change */
1609 		dev_warn(&port->adapter->ccw_device->dev,
1610 			 "FC Endpoint Security of connection to remote port 0x%16llx changed: from %s to %s\n",
1611 			 port->wwpn, mnemonic_old, mnemonic_new);
1612 	}
1613 
1614 out:
1615 	port->connection_info_old = port->connection_info;
1616 }
1617 
1618 static void zfcp_fsf_log_security_error(const struct device *dev, u32 fsf_sqw0,
1619 					u64 wwpn)
1620 {
1621 	switch (fsf_sqw0) {
1622 
1623 	/*
1624 	 * Open Port command error codes
1625 	 */
1626 
1627 	case FSF_SQ_SECURITY_REQUIRED:
1628 		dev_warn_ratelimited(dev,
1629 				     "FC Endpoint Security error: FC security is required but not supported or configured on remote port 0x%016llx\n",
1630 				     wwpn);
1631 		break;
1632 	case FSF_SQ_SECURITY_TIMEOUT:
1633 		dev_warn_ratelimited(dev,
1634 				     "FC Endpoint Security error: a timeout prevented opening remote port 0x%016llx\n",
1635 				     wwpn);
1636 		break;
1637 	case FSF_SQ_SECURITY_KM_UNAVAILABLE:
1638 		dev_warn_ratelimited(dev,
1639 				     "FC Endpoint Security error: opening remote port 0x%016llx failed because local and external key manager cannot communicate\n",
1640 				     wwpn);
1641 		break;
1642 	case FSF_SQ_SECURITY_RKM_UNAVAILABLE:
1643 		dev_warn_ratelimited(dev,
1644 				     "FC Endpoint Security error: opening remote port 0x%016llx failed because it cannot communicate with the external key manager\n",
1645 				     wwpn);
1646 		break;
1647 	case FSF_SQ_SECURITY_AUTH_FAILURE:
1648 		dev_warn_ratelimited(dev,
1649 				     "FC Endpoint Security error: the device could not verify the identity of remote port 0x%016llx\n",
1650 				     wwpn);
1651 		break;
1652 
1653 	/*
1654 	 * Send FCP command error codes
1655 	 */
1656 
1657 	case FSF_SQ_SECURITY_ENC_FAILURE:
1658 		dev_warn_ratelimited(dev,
1659 				     "FC Endpoint Security error: FC connection to remote port 0x%016llx closed because encryption broke down\n",
1660 				     wwpn);
1661 		break;
1662 
1663 	/*
1664 	 * Unknown error codes
1665 	 */
1666 
1667 	default:
1668 		dev_warn_ratelimited(dev,
1669 				     "FC Endpoint Security error: the device issued an unknown error code 0x%08x related to the FC connection to remote port 0x%016llx\n",
1670 				     fsf_sqw0, wwpn);
1671 	}
1672 }
1673 
1674 static void zfcp_fsf_open_port_handler(struct zfcp_fsf_req *req)
1675 {
1676 	struct zfcp_adapter *adapter = req->adapter;
1677 	struct zfcp_port *port = req->data;
1678 	struct fsf_qtcb_header *header = &req->qtcb->header;
1679 	struct fsf_qtcb_bottom_support *bottom = &req->qtcb->bottom.support;
1680 	struct fc_els_flogi *plogi;
1681 
1682 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
1683 		goto out;
1684 
1685 	switch (header->fsf_status) {
1686 	case FSF_PORT_ALREADY_OPEN:
1687 		break;
1688 	case FSF_MAXIMUM_NUMBER_OF_PORTS_EXCEEDED:
1689 		dev_warn(&adapter->ccw_device->dev,
1690 			 "Not enough FCP adapter resources to open "
1691 			 "remote port 0x%016Lx\n",
1692 			 (unsigned long long)port->wwpn);
1693 		zfcp_erp_set_port_status(port,
1694 					 ZFCP_STATUS_COMMON_ERP_FAILED);
1695 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1696 		break;
1697 	case FSF_SECURITY_ERROR:
1698 		zfcp_fsf_log_security_error(&req->adapter->ccw_device->dev,
1699 					    header->fsf_status_qual.word[0],
1700 					    port->wwpn);
1701 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1702 		break;
1703 	case FSF_ADAPTER_STATUS_AVAILABLE:
1704 		switch (header->fsf_status_qual.word[0]) {
1705 		case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
1706 			/* no zfcp_fc_test_link() with failed open port */
1707 			fallthrough;
1708 		case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
1709 		case FSF_SQ_NO_RETRY_POSSIBLE:
1710 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1711 			break;
1712 		}
1713 		break;
1714 	case FSF_GOOD:
1715 		port->handle = header->port_handle;
1716 		if (adapter->adapter_features & FSF_FEATURE_FC_SECURITY)
1717 			port->connection_info = bottom->connection_info;
1718 		else
1719 			port->connection_info = 0;
1720 		zfcp_fsf_log_port_fc_security(port, req);
1721 		atomic_or(ZFCP_STATUS_COMMON_OPEN |
1722 				ZFCP_STATUS_PORT_PHYS_OPEN, &port->status);
1723 		atomic_andnot(ZFCP_STATUS_COMMON_ACCESS_BOXED,
1724 		                  &port->status);
1725 		/* check whether D_ID has changed during open */
1726 		/*
1727 		 * FIXME: This check is not airtight, as the FCP channel does
1728 		 * not monitor closures of target port connections caused on
1729 		 * the remote side. Thus, they might miss out on invalidating
1730 		 * locally cached WWPNs (and other N_Port parameters) of gone
1731 		 * target ports. So, our heroic attempt to make things safe
1732 		 * could be undermined by 'open port' response data tagged with
1733 		 * obsolete WWPNs. Another reason to monitor potential
1734 		 * connection closures ourself at least (by interpreting
1735 		 * incoming ELS' and unsolicited status). It just crosses my
1736 		 * mind that one should be able to cross-check by means of
1737 		 * another GID_PN straight after a port has been opened.
1738 		 * Alternately, an ADISC/PDISC ELS should suffice, as well.
1739 		 */
1740 		plogi = (struct fc_els_flogi *) bottom->els;
1741 		if (bottom->els1_length >= FSF_PLOGI_MIN_LEN)
1742 			zfcp_fc_plogi_evaluate(port, plogi);
1743 		break;
1744 	case FSF_UNKNOWN_OP_SUBTYPE:
1745 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1746 		break;
1747 	}
1748 
1749 out:
1750 	put_device(&port->dev);
1751 }
1752 
1753 /**
1754  * zfcp_fsf_open_port - create and send open port request
1755  * @erp_action: pointer to struct zfcp_erp_action
1756  * Returns: 0 on success, error otherwise
1757  */
1758 int zfcp_fsf_open_port(struct zfcp_erp_action *erp_action)
1759 {
1760 	struct zfcp_qdio *qdio = erp_action->adapter->qdio;
1761 	struct zfcp_port *port = erp_action->port;
1762 	struct zfcp_fsf_req *req;
1763 	int retval = -EIO;
1764 
1765 	spin_lock_irq(&qdio->req_q_lock);
1766 	if (zfcp_qdio_sbal_get(qdio))
1767 		goto out;
1768 
1769 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_OPEN_PORT_WITH_DID,
1770 				  SBAL_SFLAGS0_TYPE_READ,
1771 				  qdio->adapter->pool.erp_req);
1772 
1773 	if (IS_ERR(req)) {
1774 		retval = PTR_ERR(req);
1775 		goto out;
1776 	}
1777 
1778 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1779 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1780 
1781 	req->handler = zfcp_fsf_open_port_handler;
1782 	hton24(req->qtcb->bottom.support.d_id, port->d_id);
1783 	req->data = port;
1784 	req->erp_action = erp_action;
1785 	erp_action->fsf_req_id = req->req_id;
1786 	get_device(&port->dev);
1787 
1788 	zfcp_fsf_start_erp_timer(req);
1789 	retval = zfcp_fsf_req_send(req);
1790 	if (retval) {
1791 		zfcp_fsf_req_free(req);
1792 		erp_action->fsf_req_id = 0;
1793 		put_device(&port->dev);
1794 	}
1795 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1796 out:
1797 	spin_unlock_irq(&qdio->req_q_lock);
1798 	return retval;
1799 }
1800 
1801 static void zfcp_fsf_close_port_handler(struct zfcp_fsf_req *req)
1802 {
1803 	struct zfcp_port *port = req->data;
1804 
1805 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
1806 		return;
1807 
1808 	switch (req->qtcb->header.fsf_status) {
1809 	case FSF_PORT_HANDLE_NOT_VALID:
1810 		zfcp_erp_adapter_reopen(port->adapter, 0, "fscph_1");
1811 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1812 		break;
1813 	case FSF_ADAPTER_STATUS_AVAILABLE:
1814 		break;
1815 	case FSF_GOOD:
1816 		zfcp_erp_clear_port_status(port, ZFCP_STATUS_COMMON_OPEN);
1817 		break;
1818 	}
1819 }
1820 
1821 /**
1822  * zfcp_fsf_close_port - create and send close port request
1823  * @erp_action: pointer to struct zfcp_erp_action
1824  * Returns: 0 on success, error otherwise
1825  */
1826 int zfcp_fsf_close_port(struct zfcp_erp_action *erp_action)
1827 {
1828 	struct zfcp_qdio *qdio = erp_action->adapter->qdio;
1829 	struct zfcp_fsf_req *req;
1830 	int retval = -EIO;
1831 
1832 	spin_lock_irq(&qdio->req_q_lock);
1833 	if (zfcp_qdio_sbal_get(qdio))
1834 		goto out;
1835 
1836 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_PORT,
1837 				  SBAL_SFLAGS0_TYPE_READ,
1838 				  qdio->adapter->pool.erp_req);
1839 
1840 	if (IS_ERR(req)) {
1841 		retval = PTR_ERR(req);
1842 		goto out;
1843 	}
1844 
1845 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1846 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1847 
1848 	req->handler = zfcp_fsf_close_port_handler;
1849 	req->data = erp_action->port;
1850 	req->erp_action = erp_action;
1851 	req->qtcb->header.port_handle = erp_action->port->handle;
1852 	erp_action->fsf_req_id = req->req_id;
1853 
1854 	zfcp_fsf_start_erp_timer(req);
1855 	retval = zfcp_fsf_req_send(req);
1856 	if (retval) {
1857 		zfcp_fsf_req_free(req);
1858 		erp_action->fsf_req_id = 0;
1859 	}
1860 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1861 out:
1862 	spin_unlock_irq(&qdio->req_q_lock);
1863 	return retval;
1864 }
1865 
1866 static void zfcp_fsf_open_wka_port_handler(struct zfcp_fsf_req *req)
1867 {
1868 	struct zfcp_fc_wka_port *wka_port = req->data;
1869 	struct fsf_qtcb_header *header = &req->qtcb->header;
1870 
1871 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR) {
1872 		wka_port->status = ZFCP_FC_WKA_PORT_OFFLINE;
1873 		goto out;
1874 	}
1875 
1876 	switch (header->fsf_status) {
1877 	case FSF_MAXIMUM_NUMBER_OF_PORTS_EXCEEDED:
1878 		dev_warn(&req->adapter->ccw_device->dev,
1879 			 "Opening WKA port 0x%x failed\n", wka_port->d_id);
1880 		fallthrough;
1881 	case FSF_ADAPTER_STATUS_AVAILABLE:
1882 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1883 		wka_port->status = ZFCP_FC_WKA_PORT_OFFLINE;
1884 		break;
1885 	case FSF_GOOD:
1886 		wka_port->handle = header->port_handle;
1887 		fallthrough;
1888 	case FSF_PORT_ALREADY_OPEN:
1889 		wka_port->status = ZFCP_FC_WKA_PORT_ONLINE;
1890 	}
1891 out:
1892 	wake_up(&wka_port->completion_wq);
1893 }
1894 
1895 /**
1896  * zfcp_fsf_open_wka_port - create and send open wka-port request
1897  * @wka_port: pointer to struct zfcp_fc_wka_port
1898  * Returns: 0 on success, error otherwise
1899  */
1900 int zfcp_fsf_open_wka_port(struct zfcp_fc_wka_port *wka_port)
1901 {
1902 	struct zfcp_qdio *qdio = wka_port->adapter->qdio;
1903 	struct zfcp_fsf_req *req;
1904 	unsigned long req_id = 0;
1905 	int retval = -EIO;
1906 
1907 	spin_lock_irq(&qdio->req_q_lock);
1908 	if (zfcp_qdio_sbal_get(qdio))
1909 		goto out;
1910 
1911 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_OPEN_PORT_WITH_DID,
1912 				  SBAL_SFLAGS0_TYPE_READ,
1913 				  qdio->adapter->pool.erp_req);
1914 
1915 	if (IS_ERR(req)) {
1916 		retval = PTR_ERR(req);
1917 		goto out;
1918 	}
1919 
1920 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1921 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1922 
1923 	req->handler = zfcp_fsf_open_wka_port_handler;
1924 	hton24(req->qtcb->bottom.support.d_id, wka_port->d_id);
1925 	req->data = wka_port;
1926 
1927 	req_id = req->req_id;
1928 
1929 	zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT);
1930 	retval = zfcp_fsf_req_send(req);
1931 	if (retval)
1932 		zfcp_fsf_req_free(req);
1933 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1934 out:
1935 	spin_unlock_irq(&qdio->req_q_lock);
1936 	if (!retval)
1937 		zfcp_dbf_rec_run_wka("fsowp_1", wka_port, req_id);
1938 	return retval;
1939 }
1940 
1941 static void zfcp_fsf_close_wka_port_handler(struct zfcp_fsf_req *req)
1942 {
1943 	struct zfcp_fc_wka_port *wka_port = req->data;
1944 
1945 	if (req->qtcb->header.fsf_status == FSF_PORT_HANDLE_NOT_VALID) {
1946 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
1947 		zfcp_erp_adapter_reopen(wka_port->adapter, 0, "fscwph1");
1948 	}
1949 
1950 	wka_port->status = ZFCP_FC_WKA_PORT_OFFLINE;
1951 	wake_up(&wka_port->completion_wq);
1952 }
1953 
1954 /**
1955  * zfcp_fsf_close_wka_port - create and send close wka port request
1956  * @wka_port: WKA port to open
1957  * Returns: 0 on success, error otherwise
1958  */
1959 int zfcp_fsf_close_wka_port(struct zfcp_fc_wka_port *wka_port)
1960 {
1961 	struct zfcp_qdio *qdio = wka_port->adapter->qdio;
1962 	struct zfcp_fsf_req *req;
1963 	unsigned long req_id = 0;
1964 	int retval = -EIO;
1965 
1966 	spin_lock_irq(&qdio->req_q_lock);
1967 	if (zfcp_qdio_sbal_get(qdio))
1968 		goto out;
1969 
1970 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_PORT,
1971 				  SBAL_SFLAGS0_TYPE_READ,
1972 				  qdio->adapter->pool.erp_req);
1973 
1974 	if (IS_ERR(req)) {
1975 		retval = PTR_ERR(req);
1976 		goto out;
1977 	}
1978 
1979 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
1980 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
1981 
1982 	req->handler = zfcp_fsf_close_wka_port_handler;
1983 	req->data = wka_port;
1984 	req->qtcb->header.port_handle = wka_port->handle;
1985 
1986 	req_id = req->req_id;
1987 
1988 	zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT);
1989 	retval = zfcp_fsf_req_send(req);
1990 	if (retval)
1991 		zfcp_fsf_req_free(req);
1992 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
1993 out:
1994 	spin_unlock_irq(&qdio->req_q_lock);
1995 	if (!retval)
1996 		zfcp_dbf_rec_run_wka("fscwp_1", wka_port, req_id);
1997 	return retval;
1998 }
1999 
2000 static void zfcp_fsf_close_physical_port_handler(struct zfcp_fsf_req *req)
2001 {
2002 	struct zfcp_port *port = req->data;
2003 	struct fsf_qtcb_header *header = &req->qtcb->header;
2004 	struct scsi_device *sdev;
2005 
2006 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
2007 		return;
2008 
2009 	switch (header->fsf_status) {
2010 	case FSF_PORT_HANDLE_NOT_VALID:
2011 		zfcp_erp_adapter_reopen(port->adapter, 0, "fscpph1");
2012 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2013 		break;
2014 	case FSF_PORT_BOXED:
2015 		/* can't use generic zfcp_erp_modify_port_status because
2016 		 * ZFCP_STATUS_COMMON_OPEN must not be reset for the port */
2017 		atomic_andnot(ZFCP_STATUS_PORT_PHYS_OPEN, &port->status);
2018 		shost_for_each_device(sdev, port->adapter->scsi_host)
2019 			if (sdev_to_zfcp(sdev)->port == port)
2020 				atomic_andnot(ZFCP_STATUS_COMMON_OPEN,
2021 						  &sdev_to_zfcp(sdev)->status);
2022 		zfcp_erp_set_port_status(port, ZFCP_STATUS_COMMON_ACCESS_BOXED);
2023 		zfcp_erp_port_reopen(port, ZFCP_STATUS_COMMON_ERP_FAILED,
2024 				     "fscpph2");
2025 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2026 		break;
2027 	case FSF_ADAPTER_STATUS_AVAILABLE:
2028 		switch (header->fsf_status_qual.word[0]) {
2029 		case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
2030 		case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
2031 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2032 			break;
2033 		}
2034 		break;
2035 	case FSF_GOOD:
2036 		/* can't use generic zfcp_erp_modify_port_status because
2037 		 * ZFCP_STATUS_COMMON_OPEN must not be reset for the port
2038 		 */
2039 		atomic_andnot(ZFCP_STATUS_PORT_PHYS_OPEN, &port->status);
2040 		shost_for_each_device(sdev, port->adapter->scsi_host)
2041 			if (sdev_to_zfcp(sdev)->port == port)
2042 				atomic_andnot(ZFCP_STATUS_COMMON_OPEN,
2043 						  &sdev_to_zfcp(sdev)->status);
2044 		break;
2045 	}
2046 }
2047 
2048 /**
2049  * zfcp_fsf_close_physical_port - close physical port
2050  * @erp_action: pointer to struct zfcp_erp_action
2051  * Returns: 0 on success
2052  */
2053 int zfcp_fsf_close_physical_port(struct zfcp_erp_action *erp_action)
2054 {
2055 	struct zfcp_qdio *qdio = erp_action->adapter->qdio;
2056 	struct zfcp_fsf_req *req;
2057 	int retval = -EIO;
2058 
2059 	spin_lock_irq(&qdio->req_q_lock);
2060 	if (zfcp_qdio_sbal_get(qdio))
2061 		goto out;
2062 
2063 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_PHYSICAL_PORT,
2064 				  SBAL_SFLAGS0_TYPE_READ,
2065 				  qdio->adapter->pool.erp_req);
2066 
2067 	if (IS_ERR(req)) {
2068 		retval = PTR_ERR(req);
2069 		goto out;
2070 	}
2071 
2072 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
2073 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
2074 
2075 	req->data = erp_action->port;
2076 	req->qtcb->header.port_handle = erp_action->port->handle;
2077 	req->erp_action = erp_action;
2078 	req->handler = zfcp_fsf_close_physical_port_handler;
2079 	erp_action->fsf_req_id = req->req_id;
2080 
2081 	zfcp_fsf_start_erp_timer(req);
2082 	retval = zfcp_fsf_req_send(req);
2083 	if (retval) {
2084 		zfcp_fsf_req_free(req);
2085 		erp_action->fsf_req_id = 0;
2086 	}
2087 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
2088 out:
2089 	spin_unlock_irq(&qdio->req_q_lock);
2090 	return retval;
2091 }
2092 
2093 static void zfcp_fsf_open_lun_handler(struct zfcp_fsf_req *req)
2094 {
2095 	struct zfcp_adapter *adapter = req->adapter;
2096 	struct scsi_device *sdev = req->data;
2097 	struct zfcp_scsi_dev *zfcp_sdev;
2098 	struct fsf_qtcb_header *header = &req->qtcb->header;
2099 	union fsf_status_qual *qual = &header->fsf_status_qual;
2100 
2101 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
2102 		return;
2103 
2104 	zfcp_sdev = sdev_to_zfcp(sdev);
2105 
2106 	atomic_andnot(ZFCP_STATUS_COMMON_ACCESS_DENIED |
2107 			  ZFCP_STATUS_COMMON_ACCESS_BOXED,
2108 			  &zfcp_sdev->status);
2109 
2110 	switch (header->fsf_status) {
2111 
2112 	case FSF_PORT_HANDLE_NOT_VALID:
2113 		zfcp_erp_adapter_reopen(adapter, 0, "fsouh_1");
2114 		fallthrough;
2115 	case FSF_LUN_ALREADY_OPEN:
2116 		break;
2117 	case FSF_PORT_BOXED:
2118 		zfcp_erp_set_port_status(zfcp_sdev->port,
2119 					 ZFCP_STATUS_COMMON_ACCESS_BOXED);
2120 		zfcp_erp_port_reopen(zfcp_sdev->port,
2121 				     ZFCP_STATUS_COMMON_ERP_FAILED, "fsouh_2");
2122 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2123 		break;
2124 	case FSF_LUN_SHARING_VIOLATION:
2125 		if (qual->word[0])
2126 			dev_warn(&zfcp_sdev->port->adapter->ccw_device->dev,
2127 				 "LUN 0x%016Lx on port 0x%016Lx is already in "
2128 				 "use by CSS%d, MIF Image ID %x\n",
2129 				 zfcp_scsi_dev_lun(sdev),
2130 				 (unsigned long long)zfcp_sdev->port->wwpn,
2131 				 qual->fsf_queue_designator.cssid,
2132 				 qual->fsf_queue_designator.hla);
2133 		zfcp_erp_set_lun_status(sdev,
2134 					ZFCP_STATUS_COMMON_ERP_FAILED |
2135 					ZFCP_STATUS_COMMON_ACCESS_DENIED);
2136 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2137 		break;
2138 	case FSF_MAXIMUM_NUMBER_OF_LUNS_EXCEEDED:
2139 		dev_warn(&adapter->ccw_device->dev,
2140 			 "No handle is available for LUN "
2141 			 "0x%016Lx on port 0x%016Lx\n",
2142 			 (unsigned long long)zfcp_scsi_dev_lun(sdev),
2143 			 (unsigned long long)zfcp_sdev->port->wwpn);
2144 		zfcp_erp_set_lun_status(sdev, ZFCP_STATUS_COMMON_ERP_FAILED);
2145 		fallthrough;
2146 	case FSF_INVALID_COMMAND_OPTION:
2147 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2148 		break;
2149 	case FSF_ADAPTER_STATUS_AVAILABLE:
2150 		switch (header->fsf_status_qual.word[0]) {
2151 		case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
2152 			zfcp_fc_test_link(zfcp_sdev->port);
2153 			fallthrough;
2154 		case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
2155 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2156 			break;
2157 		}
2158 		break;
2159 
2160 	case FSF_GOOD:
2161 		zfcp_sdev->lun_handle = header->lun_handle;
2162 		atomic_or(ZFCP_STATUS_COMMON_OPEN, &zfcp_sdev->status);
2163 		break;
2164 	}
2165 }
2166 
2167 /**
2168  * zfcp_fsf_open_lun - open LUN
2169  * @erp_action: pointer to struct zfcp_erp_action
2170  * Returns: 0 on success, error otherwise
2171  */
2172 int zfcp_fsf_open_lun(struct zfcp_erp_action *erp_action)
2173 {
2174 	struct zfcp_adapter *adapter = erp_action->adapter;
2175 	struct zfcp_qdio *qdio = adapter->qdio;
2176 	struct zfcp_fsf_req *req;
2177 	int retval = -EIO;
2178 
2179 	spin_lock_irq(&qdio->req_q_lock);
2180 	if (zfcp_qdio_sbal_get(qdio))
2181 		goto out;
2182 
2183 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_OPEN_LUN,
2184 				  SBAL_SFLAGS0_TYPE_READ,
2185 				  adapter->pool.erp_req);
2186 
2187 	if (IS_ERR(req)) {
2188 		retval = PTR_ERR(req);
2189 		goto out;
2190 	}
2191 
2192 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
2193 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
2194 
2195 	req->qtcb->header.port_handle = erp_action->port->handle;
2196 	req->qtcb->bottom.support.fcp_lun = zfcp_scsi_dev_lun(erp_action->sdev);
2197 	req->handler = zfcp_fsf_open_lun_handler;
2198 	req->data = erp_action->sdev;
2199 	req->erp_action = erp_action;
2200 	erp_action->fsf_req_id = req->req_id;
2201 
2202 	if (!(adapter->connection_features & FSF_FEATURE_NPIV_MODE))
2203 		req->qtcb->bottom.support.option = FSF_OPEN_LUN_SUPPRESS_BOXING;
2204 
2205 	zfcp_fsf_start_erp_timer(req);
2206 	retval = zfcp_fsf_req_send(req);
2207 	if (retval) {
2208 		zfcp_fsf_req_free(req);
2209 		erp_action->fsf_req_id = 0;
2210 	}
2211 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
2212 out:
2213 	spin_unlock_irq(&qdio->req_q_lock);
2214 	return retval;
2215 }
2216 
2217 static void zfcp_fsf_close_lun_handler(struct zfcp_fsf_req *req)
2218 {
2219 	struct scsi_device *sdev = req->data;
2220 	struct zfcp_scsi_dev *zfcp_sdev;
2221 
2222 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
2223 		return;
2224 
2225 	zfcp_sdev = sdev_to_zfcp(sdev);
2226 
2227 	switch (req->qtcb->header.fsf_status) {
2228 	case FSF_PORT_HANDLE_NOT_VALID:
2229 		zfcp_erp_adapter_reopen(zfcp_sdev->port->adapter, 0, "fscuh_1");
2230 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2231 		break;
2232 	case FSF_LUN_HANDLE_NOT_VALID:
2233 		zfcp_erp_port_reopen(zfcp_sdev->port, 0, "fscuh_2");
2234 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2235 		break;
2236 	case FSF_PORT_BOXED:
2237 		zfcp_erp_set_port_status(zfcp_sdev->port,
2238 					 ZFCP_STATUS_COMMON_ACCESS_BOXED);
2239 		zfcp_erp_port_reopen(zfcp_sdev->port,
2240 				     ZFCP_STATUS_COMMON_ERP_FAILED, "fscuh_3");
2241 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2242 		break;
2243 	case FSF_ADAPTER_STATUS_AVAILABLE:
2244 		switch (req->qtcb->header.fsf_status_qual.word[0]) {
2245 		case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE:
2246 			zfcp_fc_test_link(zfcp_sdev->port);
2247 			fallthrough;
2248 		case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED:
2249 			req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2250 			break;
2251 		}
2252 		break;
2253 	case FSF_GOOD:
2254 		atomic_andnot(ZFCP_STATUS_COMMON_OPEN, &zfcp_sdev->status);
2255 		break;
2256 	}
2257 }
2258 
2259 /**
2260  * zfcp_fsf_close_LUN - close LUN
2261  * @erp_action: pointer to erp_action triggering the "close LUN"
2262  * Returns: 0 on success, error otherwise
2263  */
2264 int zfcp_fsf_close_lun(struct zfcp_erp_action *erp_action)
2265 {
2266 	struct zfcp_qdio *qdio = erp_action->adapter->qdio;
2267 	struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(erp_action->sdev);
2268 	struct zfcp_fsf_req *req;
2269 	int retval = -EIO;
2270 
2271 	spin_lock_irq(&qdio->req_q_lock);
2272 	if (zfcp_qdio_sbal_get(qdio))
2273 		goto out;
2274 
2275 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_LUN,
2276 				  SBAL_SFLAGS0_TYPE_READ,
2277 				  qdio->adapter->pool.erp_req);
2278 
2279 	if (IS_ERR(req)) {
2280 		retval = PTR_ERR(req);
2281 		goto out;
2282 	}
2283 
2284 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
2285 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
2286 
2287 	req->qtcb->header.port_handle = erp_action->port->handle;
2288 	req->qtcb->header.lun_handle = zfcp_sdev->lun_handle;
2289 	req->handler = zfcp_fsf_close_lun_handler;
2290 	req->data = erp_action->sdev;
2291 	req->erp_action = erp_action;
2292 	erp_action->fsf_req_id = req->req_id;
2293 
2294 	zfcp_fsf_start_erp_timer(req);
2295 	retval = zfcp_fsf_req_send(req);
2296 	if (retval) {
2297 		zfcp_fsf_req_free(req);
2298 		erp_action->fsf_req_id = 0;
2299 	}
2300 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
2301 out:
2302 	spin_unlock_irq(&qdio->req_q_lock);
2303 	return retval;
2304 }
2305 
2306 static void zfcp_fsf_update_lat(struct zfcp_latency_record *lat_rec, u32 lat)
2307 {
2308 	lat_rec->sum += lat;
2309 	lat_rec->min = min(lat_rec->min, lat);
2310 	lat_rec->max = max(lat_rec->max, lat);
2311 }
2312 
2313 static void zfcp_fsf_req_trace(struct zfcp_fsf_req *req, struct scsi_cmnd *scsi)
2314 {
2315 	struct fsf_qual_latency_info *lat_in;
2316 	struct zfcp_latency_cont *lat = NULL;
2317 	struct zfcp_scsi_dev *zfcp_sdev;
2318 	struct zfcp_blk_drv_data blktrc;
2319 	int ticks = req->adapter->timer_ticks;
2320 
2321 	lat_in = &req->qtcb->prefix.prot_status_qual.latency_info;
2322 
2323 	blktrc.flags = 0;
2324 	blktrc.magic = ZFCP_BLK_DRV_DATA_MAGIC;
2325 	if (req->status & ZFCP_STATUS_FSFREQ_ERROR)
2326 		blktrc.flags |= ZFCP_BLK_REQ_ERROR;
2327 	blktrc.inb_usage = 0;
2328 	blktrc.outb_usage = req->qdio_req.qdio_outb_usage;
2329 
2330 	if (req->adapter->adapter_features & FSF_FEATURE_MEASUREMENT_DATA &&
2331 	    !(req->status & ZFCP_STATUS_FSFREQ_ERROR)) {
2332 		zfcp_sdev = sdev_to_zfcp(scsi->device);
2333 		blktrc.flags |= ZFCP_BLK_LAT_VALID;
2334 		blktrc.channel_lat = lat_in->channel_lat * ticks;
2335 		blktrc.fabric_lat = lat_in->fabric_lat * ticks;
2336 
2337 		switch (req->qtcb->bottom.io.data_direction) {
2338 		case FSF_DATADIR_DIF_READ_STRIP:
2339 		case FSF_DATADIR_DIF_READ_CONVERT:
2340 		case FSF_DATADIR_READ:
2341 			lat = &zfcp_sdev->latencies.read;
2342 			break;
2343 		case FSF_DATADIR_DIF_WRITE_INSERT:
2344 		case FSF_DATADIR_DIF_WRITE_CONVERT:
2345 		case FSF_DATADIR_WRITE:
2346 			lat = &zfcp_sdev->latencies.write;
2347 			break;
2348 		case FSF_DATADIR_CMND:
2349 			lat = &zfcp_sdev->latencies.cmd;
2350 			break;
2351 		}
2352 
2353 		if (lat) {
2354 			spin_lock(&zfcp_sdev->latencies.lock);
2355 			zfcp_fsf_update_lat(&lat->channel, lat_in->channel_lat);
2356 			zfcp_fsf_update_lat(&lat->fabric, lat_in->fabric_lat);
2357 			lat->counter++;
2358 			spin_unlock(&zfcp_sdev->latencies.lock);
2359 		}
2360 	}
2361 
2362 	blk_add_driver_data(scsi->request->q, scsi->request, &blktrc,
2363 			    sizeof(blktrc));
2364 }
2365 
2366 /**
2367  * zfcp_fsf_fcp_handler_common() - FCP response handler common to I/O and TMF.
2368  * @req: Pointer to FSF request.
2369  * @sdev: Pointer to SCSI device as request context.
2370  */
2371 static void zfcp_fsf_fcp_handler_common(struct zfcp_fsf_req *req,
2372 					struct scsi_device *sdev)
2373 {
2374 	struct zfcp_scsi_dev *zfcp_sdev;
2375 	struct fsf_qtcb_header *header = &req->qtcb->header;
2376 
2377 	if (unlikely(req->status & ZFCP_STATUS_FSFREQ_ERROR))
2378 		return;
2379 
2380 	zfcp_sdev = sdev_to_zfcp(sdev);
2381 
2382 	switch (header->fsf_status) {
2383 	case FSF_HANDLE_MISMATCH:
2384 	case FSF_PORT_HANDLE_NOT_VALID:
2385 		zfcp_erp_adapter_reopen(req->adapter, 0, "fssfch1");
2386 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2387 		break;
2388 	case FSF_FCPLUN_NOT_VALID:
2389 	case FSF_LUN_HANDLE_NOT_VALID:
2390 		zfcp_erp_port_reopen(zfcp_sdev->port, 0, "fssfch2");
2391 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2392 		break;
2393 	case FSF_SERVICE_CLASS_NOT_SUPPORTED:
2394 		zfcp_fsf_class_not_supp(req);
2395 		break;
2396 	case FSF_DIRECTION_INDICATOR_NOT_VALID:
2397 		dev_err(&req->adapter->ccw_device->dev,
2398 			"Incorrect direction %d, LUN 0x%016Lx on port "
2399 			"0x%016Lx closed\n",
2400 			req->qtcb->bottom.io.data_direction,
2401 			(unsigned long long)zfcp_scsi_dev_lun(sdev),
2402 			(unsigned long long)zfcp_sdev->port->wwpn);
2403 		zfcp_erp_adapter_shutdown(req->adapter, 0, "fssfch3");
2404 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2405 		break;
2406 	case FSF_CMND_LENGTH_NOT_VALID:
2407 		dev_err(&req->adapter->ccw_device->dev,
2408 			"Incorrect FCP_CMND length %d, FCP device closed\n",
2409 			req->qtcb->bottom.io.fcp_cmnd_length);
2410 		zfcp_erp_adapter_shutdown(req->adapter, 0, "fssfch4");
2411 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2412 		break;
2413 	case FSF_PORT_BOXED:
2414 		zfcp_erp_set_port_status(zfcp_sdev->port,
2415 					 ZFCP_STATUS_COMMON_ACCESS_BOXED);
2416 		zfcp_erp_port_reopen(zfcp_sdev->port,
2417 				     ZFCP_STATUS_COMMON_ERP_FAILED, "fssfch5");
2418 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2419 		break;
2420 	case FSF_LUN_BOXED:
2421 		zfcp_erp_set_lun_status(sdev, ZFCP_STATUS_COMMON_ACCESS_BOXED);
2422 		zfcp_erp_lun_reopen(sdev, ZFCP_STATUS_COMMON_ERP_FAILED,
2423 				    "fssfch6");
2424 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2425 		break;
2426 	case FSF_ADAPTER_STATUS_AVAILABLE:
2427 		if (header->fsf_status_qual.word[0] ==
2428 		    FSF_SQ_INVOKE_LINK_TEST_PROCEDURE)
2429 			zfcp_fc_test_link(zfcp_sdev->port);
2430 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2431 		break;
2432 	case FSF_SECURITY_ERROR:
2433 		zfcp_fsf_log_security_error(&req->adapter->ccw_device->dev,
2434 					    header->fsf_status_qual.word[0],
2435 					    zfcp_sdev->port->wwpn);
2436 		zfcp_erp_port_forced_reopen(zfcp_sdev->port, 0, "fssfch7");
2437 		req->status |= ZFCP_STATUS_FSFREQ_ERROR;
2438 		break;
2439 	}
2440 }
2441 
2442 static void zfcp_fsf_fcp_cmnd_handler(struct zfcp_fsf_req *req)
2443 {
2444 	struct scsi_cmnd *scpnt;
2445 	struct fcp_resp_with_ext *fcp_rsp;
2446 	unsigned long flags;
2447 
2448 	read_lock_irqsave(&req->adapter->abort_lock, flags);
2449 
2450 	scpnt = req->data;
2451 	if (unlikely(!scpnt)) {
2452 		read_unlock_irqrestore(&req->adapter->abort_lock, flags);
2453 		return;
2454 	}
2455 
2456 	zfcp_fsf_fcp_handler_common(req, scpnt->device);
2457 
2458 	if (unlikely(req->status & ZFCP_STATUS_FSFREQ_ERROR)) {
2459 		set_host_byte(scpnt, DID_TRANSPORT_DISRUPTED);
2460 		goto skip_fsfstatus;
2461 	}
2462 
2463 	switch (req->qtcb->header.fsf_status) {
2464 	case FSF_INCONSISTENT_PROT_DATA:
2465 	case FSF_INVALID_PROT_PARM:
2466 		set_host_byte(scpnt, DID_ERROR);
2467 		goto skip_fsfstatus;
2468 	case FSF_BLOCK_GUARD_CHECK_FAILURE:
2469 		zfcp_scsi_dif_sense_error(scpnt, 0x1);
2470 		goto skip_fsfstatus;
2471 	case FSF_APP_TAG_CHECK_FAILURE:
2472 		zfcp_scsi_dif_sense_error(scpnt, 0x2);
2473 		goto skip_fsfstatus;
2474 	case FSF_REF_TAG_CHECK_FAILURE:
2475 		zfcp_scsi_dif_sense_error(scpnt, 0x3);
2476 		goto skip_fsfstatus;
2477 	}
2478 	BUILD_BUG_ON(sizeof(struct fcp_resp_with_ext) > FSF_FCP_RSP_SIZE);
2479 	fcp_rsp = &req->qtcb->bottom.io.fcp_rsp.iu;
2480 	zfcp_fc_eval_fcp_rsp(fcp_rsp, scpnt);
2481 
2482 skip_fsfstatus:
2483 	zfcp_fsf_req_trace(req, scpnt);
2484 	zfcp_dbf_scsi_result(scpnt, req);
2485 
2486 	scpnt->host_scribble = NULL;
2487 	(scpnt->scsi_done) (scpnt);
2488 	/*
2489 	 * We must hold this lock until scsi_done has been called.
2490 	 * Otherwise we may call scsi_done after abort regarding this
2491 	 * command has completed.
2492 	 * Note: scsi_done must not block!
2493 	 */
2494 	read_unlock_irqrestore(&req->adapter->abort_lock, flags);
2495 }
2496 
2497 static int zfcp_fsf_set_data_dir(struct scsi_cmnd *scsi_cmnd, u32 *data_dir)
2498 {
2499 	switch (scsi_get_prot_op(scsi_cmnd)) {
2500 	case SCSI_PROT_NORMAL:
2501 		switch (scsi_cmnd->sc_data_direction) {
2502 		case DMA_NONE:
2503 			*data_dir = FSF_DATADIR_CMND;
2504 			break;
2505 		case DMA_FROM_DEVICE:
2506 			*data_dir = FSF_DATADIR_READ;
2507 			break;
2508 		case DMA_TO_DEVICE:
2509 			*data_dir = FSF_DATADIR_WRITE;
2510 			break;
2511 		case DMA_BIDIRECTIONAL:
2512 			return -EINVAL;
2513 		}
2514 		break;
2515 
2516 	case SCSI_PROT_READ_STRIP:
2517 		*data_dir = FSF_DATADIR_DIF_READ_STRIP;
2518 		break;
2519 	case SCSI_PROT_WRITE_INSERT:
2520 		*data_dir = FSF_DATADIR_DIF_WRITE_INSERT;
2521 		break;
2522 	case SCSI_PROT_READ_PASS:
2523 		*data_dir = FSF_DATADIR_DIF_READ_CONVERT;
2524 		break;
2525 	case SCSI_PROT_WRITE_PASS:
2526 		*data_dir = FSF_DATADIR_DIF_WRITE_CONVERT;
2527 		break;
2528 	default:
2529 		return -EINVAL;
2530 	}
2531 
2532 	return 0;
2533 }
2534 
2535 /**
2536  * zfcp_fsf_fcp_cmnd - initiate an FCP command (for a SCSI command)
2537  * @scsi_cmnd: scsi command to be sent
2538  */
2539 int zfcp_fsf_fcp_cmnd(struct scsi_cmnd *scsi_cmnd)
2540 {
2541 	struct zfcp_fsf_req *req;
2542 	struct fcp_cmnd *fcp_cmnd;
2543 	u8 sbtype = SBAL_SFLAGS0_TYPE_READ;
2544 	int retval = -EIO;
2545 	struct scsi_device *sdev = scsi_cmnd->device;
2546 	struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(sdev);
2547 	struct zfcp_adapter *adapter = zfcp_sdev->port->adapter;
2548 	struct zfcp_qdio *qdio = adapter->qdio;
2549 	struct fsf_qtcb_bottom_io *io;
2550 	unsigned long flags;
2551 
2552 	if (unlikely(!(atomic_read(&zfcp_sdev->status) &
2553 		       ZFCP_STATUS_COMMON_UNBLOCKED)))
2554 		return -EBUSY;
2555 
2556 	spin_lock_irqsave(&qdio->req_q_lock, flags);
2557 	if (atomic_read(&qdio->req_q_free) <= 0) {
2558 		atomic_inc(&qdio->req_q_full);
2559 		goto out;
2560 	}
2561 
2562 	if (scsi_cmnd->sc_data_direction == DMA_TO_DEVICE)
2563 		sbtype = SBAL_SFLAGS0_TYPE_WRITE;
2564 
2565 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_FCP_CMND,
2566 				  sbtype, adapter->pool.scsi_req);
2567 
2568 	if (IS_ERR(req)) {
2569 		retval = PTR_ERR(req);
2570 		goto out;
2571 	}
2572 
2573 	scsi_cmnd->host_scribble = (unsigned char *) req->req_id;
2574 
2575 	io = &req->qtcb->bottom.io;
2576 	req->status |= ZFCP_STATUS_FSFREQ_CLEANUP;
2577 	req->data = scsi_cmnd;
2578 	req->handler = zfcp_fsf_fcp_cmnd_handler;
2579 	req->qtcb->header.lun_handle = zfcp_sdev->lun_handle;
2580 	req->qtcb->header.port_handle = zfcp_sdev->port->handle;
2581 	io->service_class = FSF_CLASS_3;
2582 	io->fcp_cmnd_length = FCP_CMND_LEN;
2583 
2584 	if (scsi_get_prot_op(scsi_cmnd) != SCSI_PROT_NORMAL) {
2585 		io->data_block_length = scsi_cmnd->device->sector_size;
2586 		io->ref_tag_value = scsi_get_lba(scsi_cmnd) & 0xFFFFFFFF;
2587 	}
2588 
2589 	if (zfcp_fsf_set_data_dir(scsi_cmnd, &io->data_direction))
2590 		goto failed_scsi_cmnd;
2591 
2592 	BUILD_BUG_ON(sizeof(struct fcp_cmnd) > FSF_FCP_CMND_SIZE);
2593 	fcp_cmnd = &req->qtcb->bottom.io.fcp_cmnd.iu;
2594 	zfcp_fc_scsi_to_fcp(fcp_cmnd, scsi_cmnd);
2595 
2596 	if ((scsi_get_prot_op(scsi_cmnd) != SCSI_PROT_NORMAL) &&
2597 	    scsi_prot_sg_count(scsi_cmnd)) {
2598 		zfcp_qdio_set_data_div(qdio, &req->qdio_req,
2599 				       scsi_prot_sg_count(scsi_cmnd));
2600 		retval = zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req,
2601 						 scsi_prot_sglist(scsi_cmnd));
2602 		if (retval)
2603 			goto failed_scsi_cmnd;
2604 		io->prot_data_length = zfcp_qdio_real_bytes(
2605 						scsi_prot_sglist(scsi_cmnd));
2606 	}
2607 
2608 	retval = zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req,
2609 					 scsi_sglist(scsi_cmnd));
2610 	if (unlikely(retval))
2611 		goto failed_scsi_cmnd;
2612 
2613 	zfcp_qdio_set_sbale_last(adapter->qdio, &req->qdio_req);
2614 	if (zfcp_adapter_multi_buffer_active(adapter))
2615 		zfcp_qdio_set_scount(qdio, &req->qdio_req);
2616 
2617 	retval = zfcp_fsf_req_send(req);
2618 	if (unlikely(retval))
2619 		goto failed_scsi_cmnd;
2620 	/* NOTE: DO NOT TOUCH req PAST THIS POINT! */
2621 
2622 	goto out;
2623 
2624 failed_scsi_cmnd:
2625 	zfcp_fsf_req_free(req);
2626 	scsi_cmnd->host_scribble = NULL;
2627 out:
2628 	spin_unlock_irqrestore(&qdio->req_q_lock, flags);
2629 	return retval;
2630 }
2631 
2632 static void zfcp_fsf_fcp_task_mgmt_handler(struct zfcp_fsf_req *req)
2633 {
2634 	struct scsi_device *sdev = req->data;
2635 	struct fcp_resp_with_ext *fcp_rsp;
2636 	struct fcp_resp_rsp_info *rsp_info;
2637 
2638 	zfcp_fsf_fcp_handler_common(req, sdev);
2639 
2640 	fcp_rsp = &req->qtcb->bottom.io.fcp_rsp.iu;
2641 	rsp_info = (struct fcp_resp_rsp_info *) &fcp_rsp[1];
2642 
2643 	if ((rsp_info->rsp_code != FCP_TMF_CMPL) ||
2644 	     (req->status & ZFCP_STATUS_FSFREQ_ERROR))
2645 		req->status |= ZFCP_STATUS_FSFREQ_TMFUNCFAILED;
2646 }
2647 
2648 /**
2649  * zfcp_fsf_fcp_task_mgmt() - Send SCSI task management command (TMF).
2650  * @sdev: Pointer to SCSI device to send the task management command to.
2651  * @tm_flags: Unsigned byte for task management flags.
2652  *
2653  * Return: On success pointer to struct zfcp_fsf_req, %NULL otherwise.
2654  */
2655 struct zfcp_fsf_req *zfcp_fsf_fcp_task_mgmt(struct scsi_device *sdev,
2656 					    u8 tm_flags)
2657 {
2658 	struct zfcp_fsf_req *req = NULL;
2659 	struct fcp_cmnd *fcp_cmnd;
2660 	struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(sdev);
2661 	struct zfcp_qdio *qdio = zfcp_sdev->port->adapter->qdio;
2662 
2663 	if (unlikely(!(atomic_read(&zfcp_sdev->status) &
2664 		       ZFCP_STATUS_COMMON_UNBLOCKED)))
2665 		return NULL;
2666 
2667 	spin_lock_irq(&qdio->req_q_lock);
2668 	if (zfcp_qdio_sbal_get(qdio))
2669 		goto out;
2670 
2671 	req = zfcp_fsf_req_create(qdio, FSF_QTCB_FCP_CMND,
2672 				  SBAL_SFLAGS0_TYPE_WRITE,
2673 				  qdio->adapter->pool.scsi_req);
2674 
2675 	if (IS_ERR(req)) {
2676 		req = NULL;
2677 		goto out;
2678 	}
2679 
2680 	req->data = sdev;
2681 
2682 	req->handler = zfcp_fsf_fcp_task_mgmt_handler;
2683 	req->qtcb->header.lun_handle = zfcp_sdev->lun_handle;
2684 	req->qtcb->header.port_handle = zfcp_sdev->port->handle;
2685 	req->qtcb->bottom.io.data_direction = FSF_DATADIR_CMND;
2686 	req->qtcb->bottom.io.service_class = FSF_CLASS_3;
2687 	req->qtcb->bottom.io.fcp_cmnd_length = FCP_CMND_LEN;
2688 
2689 	zfcp_qdio_set_sbale_last(qdio, &req->qdio_req);
2690 
2691 	fcp_cmnd = &req->qtcb->bottom.io.fcp_cmnd.iu;
2692 	zfcp_fc_fcp_tm(fcp_cmnd, sdev, tm_flags);
2693 
2694 	zfcp_fsf_start_timer(req, ZFCP_FSF_SCSI_ER_TIMEOUT);
2695 	if (!zfcp_fsf_req_send(req)) {
2696 		/* NOTE: DO NOT TOUCH req, UNTIL IT COMPLETES! */
2697 		goto out;
2698 	}
2699 
2700 	zfcp_fsf_req_free(req);
2701 	req = NULL;
2702 out:
2703 	spin_unlock_irq(&qdio->req_q_lock);
2704 	return req;
2705 }
2706 
2707 /**
2708  * zfcp_fsf_reqid_check - validate req_id contained in SBAL returned by QDIO
2709  * @qdio: pointer to struct zfcp_qdio
2710  * @sbal_idx: response queue index of SBAL to be processed
2711  */
2712 void zfcp_fsf_reqid_check(struct zfcp_qdio *qdio, int sbal_idx)
2713 {
2714 	struct zfcp_adapter *adapter = qdio->adapter;
2715 	struct qdio_buffer *sbal = qdio->res_q[sbal_idx];
2716 	struct qdio_buffer_element *sbale;
2717 	struct zfcp_fsf_req *fsf_req;
2718 	unsigned long req_id;
2719 	int idx;
2720 
2721 	for (idx = 0; idx < QDIO_MAX_ELEMENTS_PER_BUFFER; idx++) {
2722 
2723 		sbale = &sbal->element[idx];
2724 		req_id = sbale->addr;
2725 		fsf_req = zfcp_reqlist_find_rm(adapter->req_list, req_id);
2726 
2727 		if (!fsf_req) {
2728 			/*
2729 			 * Unknown request means that we have potentially memory
2730 			 * corruption and must stop the machine immediately.
2731 			 */
2732 			zfcp_qdio_siosl(adapter);
2733 			panic("error: unknown req_id (%lx) on adapter %s.\n",
2734 			      req_id, dev_name(&adapter->ccw_device->dev));
2735 		}
2736 
2737 		zfcp_fsf_req_complete(fsf_req);
2738 
2739 		if (likely(sbale->eflags & SBAL_EFLAGS_LAST_ENTRY))
2740 			break;
2741 	}
2742 }
2743