1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * Generic SCSI-3 ALUA SCSI Device Handler
4 *
5 * Copyright (C) 2007-2010 Hannes Reinecke, SUSE Linux Products GmbH.
6 * All rights reserved.
7 */
8 #include <linux/slab.h>
9 #include <linux/delay.h>
10 #include <linux/module.h>
11 #include <asm/unaligned.h>
12 #include <scsi/scsi.h>
13 #include <scsi/scsi_proto.h>
14 #include <scsi/scsi_dbg.h>
15 #include <scsi/scsi_eh.h>
16 #include <scsi/scsi_dh.h>
17
18 #define ALUA_DH_NAME "alua"
19 #define ALUA_DH_VER "2.0"
20
21 #define TPGS_SUPPORT_NONE 0x00
22 #define TPGS_SUPPORT_OPTIMIZED 0x01
23 #define TPGS_SUPPORT_NONOPTIMIZED 0x02
24 #define TPGS_SUPPORT_STANDBY 0x04
25 #define TPGS_SUPPORT_UNAVAILABLE 0x08
26 #define TPGS_SUPPORT_LBA_DEPENDENT 0x10
27 #define TPGS_SUPPORT_OFFLINE 0x40
28 #define TPGS_SUPPORT_TRANSITION 0x80
29 #define TPGS_SUPPORT_ALL 0xdf
30
31 #define RTPG_FMT_MASK 0x70
32 #define RTPG_FMT_EXT_HDR 0x10
33
34 #define TPGS_MODE_UNINITIALIZED -1
35 #define TPGS_MODE_NONE 0x0
36 #define TPGS_MODE_IMPLICIT 0x1
37 #define TPGS_MODE_EXPLICIT 0x2
38
39 #define ALUA_RTPG_SIZE 128
40 #define ALUA_FAILOVER_TIMEOUT 60
41 #define ALUA_FAILOVER_RETRIES 5
42 #define ALUA_RTPG_DELAY_MSECS 5
43 #define ALUA_RTPG_RETRY_DELAY 2
44
45 /* device handler flags */
46 #define ALUA_OPTIMIZE_STPG 0x01
47 #define ALUA_RTPG_EXT_HDR_UNSUPP 0x02
48 /* State machine flags */
49 #define ALUA_PG_RUN_RTPG 0x10
50 #define ALUA_PG_RUN_STPG 0x20
51 #define ALUA_PG_RUNNING 0x40
52
53 static uint optimize_stpg;
54 module_param(optimize_stpg, uint, S_IRUGO|S_IWUSR);
55 MODULE_PARM_DESC(optimize_stpg, "Allow use of a non-optimized path, rather than sending a STPG, when implicit TPGS is supported (0=No,1=Yes). Default is 0.");
56
57 static LIST_HEAD(port_group_list);
58 static DEFINE_SPINLOCK(port_group_lock);
59 static struct workqueue_struct *kaluad_wq;
60
61 struct alua_port_group {
62 struct kref kref;
63 struct rcu_head rcu;
64 struct list_head node;
65 struct list_head dh_list;
66 unsigned char device_id_str[256];
67 int device_id_len;
68 int group_id;
69 int tpgs;
70 int state;
71 int pref;
72 int valid_states;
73 unsigned flags; /* used for optimizing STPG */
74 unsigned char transition_tmo;
75 unsigned long expiry;
76 unsigned long interval;
77 struct delayed_work rtpg_work;
78 spinlock_t lock;
79 struct list_head rtpg_list;
80 struct scsi_device *rtpg_sdev;
81 };
82
83 struct alua_dh_data {
84 struct list_head node;
85 struct alua_port_group __rcu *pg;
86 int group_id;
87 spinlock_t pg_lock;
88 struct scsi_device *sdev;
89 int init_error;
90 struct mutex init_mutex;
91 bool disabled;
92 };
93
94 struct alua_queue_data {
95 struct list_head entry;
96 activate_complete callback_fn;
97 void *callback_data;
98 };
99
100 #define ALUA_POLICY_SWITCH_CURRENT 0
101 #define ALUA_POLICY_SWITCH_ALL 1
102
103 static void alua_rtpg_work(struct work_struct *work);
104 static bool alua_rtpg_queue(struct alua_port_group *pg,
105 struct scsi_device *sdev,
106 struct alua_queue_data *qdata, bool force);
107 static void alua_check(struct scsi_device *sdev, bool force);
108
release_port_group(struct kref * kref)109 static void release_port_group(struct kref *kref)
110 {
111 struct alua_port_group *pg;
112
113 pg = container_of(kref, struct alua_port_group, kref);
114 if (pg->rtpg_sdev)
115 flush_delayed_work(&pg->rtpg_work);
116 spin_lock(&port_group_lock);
117 list_del(&pg->node);
118 spin_unlock(&port_group_lock);
119 kfree_rcu(pg, rcu);
120 }
121
122 /*
123 * submit_rtpg - Issue a REPORT TARGET GROUP STATES command
124 * @sdev: sdev the command should be sent to
125 */
submit_rtpg(struct scsi_device * sdev,unsigned char * buff,int bufflen,struct scsi_sense_hdr * sshdr,int flags)126 static int submit_rtpg(struct scsi_device *sdev, unsigned char *buff,
127 int bufflen, struct scsi_sense_hdr *sshdr, int flags)
128 {
129 u8 cdb[MAX_COMMAND_SIZE];
130 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV |
131 REQ_FAILFAST_TRANSPORT | REQ_FAILFAST_DRIVER;
132 const struct scsi_exec_args exec_args = {
133 .sshdr = sshdr,
134 };
135
136 /* Prepare the command. */
137 memset(cdb, 0x0, MAX_COMMAND_SIZE);
138 cdb[0] = MAINTENANCE_IN;
139 if (!(flags & ALUA_RTPG_EXT_HDR_UNSUPP))
140 cdb[1] = MI_REPORT_TARGET_PGS | MI_EXT_HDR_PARAM_FMT;
141 else
142 cdb[1] = MI_REPORT_TARGET_PGS;
143 put_unaligned_be32(bufflen, &cdb[6]);
144
145 return scsi_execute_cmd(sdev, cdb, opf, buff, bufflen,
146 ALUA_FAILOVER_TIMEOUT * HZ,
147 ALUA_FAILOVER_RETRIES, &exec_args);
148 }
149
150 /*
151 * submit_stpg - Issue a SET TARGET PORT GROUP command
152 *
153 * Currently we're only setting the current target port group state
154 * to 'active/optimized' and let the array firmware figure out
155 * the states of the remaining groups.
156 */
submit_stpg(struct scsi_device * sdev,int group_id,struct scsi_sense_hdr * sshdr)157 static int submit_stpg(struct scsi_device *sdev, int group_id,
158 struct scsi_sense_hdr *sshdr)
159 {
160 u8 cdb[MAX_COMMAND_SIZE];
161 unsigned char stpg_data[8];
162 int stpg_len = 8;
163 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV |
164 REQ_FAILFAST_TRANSPORT | REQ_FAILFAST_DRIVER;
165 const struct scsi_exec_args exec_args = {
166 .sshdr = sshdr,
167 };
168
169 /* Prepare the data buffer */
170 memset(stpg_data, 0, stpg_len);
171 stpg_data[4] = SCSI_ACCESS_STATE_OPTIMAL;
172 put_unaligned_be16(group_id, &stpg_data[6]);
173
174 /* Prepare the command. */
175 memset(cdb, 0x0, MAX_COMMAND_SIZE);
176 cdb[0] = MAINTENANCE_OUT;
177 cdb[1] = MO_SET_TARGET_PGS;
178 put_unaligned_be32(stpg_len, &cdb[6]);
179
180 return scsi_execute_cmd(sdev, cdb, opf, stpg_data,
181 stpg_len, ALUA_FAILOVER_TIMEOUT * HZ,
182 ALUA_FAILOVER_RETRIES, &exec_args);
183 }
184
alua_find_get_pg(char * id_str,size_t id_size,int group_id)185 static struct alua_port_group *alua_find_get_pg(char *id_str, size_t id_size,
186 int group_id)
187 {
188 struct alua_port_group *pg;
189
190 if (!id_str || !id_size || !strlen(id_str))
191 return NULL;
192
193 list_for_each_entry(pg, &port_group_list, node) {
194 if (pg->group_id != group_id)
195 continue;
196 if (!pg->device_id_len || pg->device_id_len != id_size)
197 continue;
198 if (strncmp(pg->device_id_str, id_str, id_size))
199 continue;
200 if (!kref_get_unless_zero(&pg->kref))
201 continue;
202 return pg;
203 }
204
205 return NULL;
206 }
207
208 /*
209 * alua_alloc_pg - Allocate a new port_group structure
210 * @sdev: scsi device
211 * @group_id: port group id
212 * @tpgs: target port group settings
213 *
214 * Allocate a new port_group structure for a given
215 * device.
216 */
alua_alloc_pg(struct scsi_device * sdev,int group_id,int tpgs)217 static struct alua_port_group *alua_alloc_pg(struct scsi_device *sdev,
218 int group_id, int tpgs)
219 {
220 struct alua_port_group *pg, *tmp_pg;
221
222 pg = kzalloc(sizeof(struct alua_port_group), GFP_KERNEL);
223 if (!pg)
224 return ERR_PTR(-ENOMEM);
225
226 pg->device_id_len = scsi_vpd_lun_id(sdev, pg->device_id_str,
227 sizeof(pg->device_id_str));
228 if (pg->device_id_len <= 0) {
229 /*
230 * TPGS supported but no device identification found.
231 * Generate private device identification.
232 */
233 sdev_printk(KERN_INFO, sdev,
234 "%s: No device descriptors found\n",
235 ALUA_DH_NAME);
236 pg->device_id_str[0] = '\0';
237 pg->device_id_len = 0;
238 }
239 pg->group_id = group_id;
240 pg->tpgs = tpgs;
241 pg->state = SCSI_ACCESS_STATE_OPTIMAL;
242 pg->valid_states = TPGS_SUPPORT_ALL;
243 if (optimize_stpg)
244 pg->flags |= ALUA_OPTIMIZE_STPG;
245 kref_init(&pg->kref);
246 INIT_DELAYED_WORK(&pg->rtpg_work, alua_rtpg_work);
247 INIT_LIST_HEAD(&pg->rtpg_list);
248 INIT_LIST_HEAD(&pg->node);
249 INIT_LIST_HEAD(&pg->dh_list);
250 spin_lock_init(&pg->lock);
251
252 spin_lock(&port_group_lock);
253 tmp_pg = alua_find_get_pg(pg->device_id_str, pg->device_id_len,
254 group_id);
255 if (tmp_pg) {
256 spin_unlock(&port_group_lock);
257 kfree(pg);
258 return tmp_pg;
259 }
260
261 list_add(&pg->node, &port_group_list);
262 spin_unlock(&port_group_lock);
263
264 return pg;
265 }
266
267 /*
268 * alua_check_tpgs - Evaluate TPGS setting
269 * @sdev: device to be checked
270 *
271 * Examine the TPGS setting of the sdev to find out if ALUA
272 * is supported.
273 */
alua_check_tpgs(struct scsi_device * sdev)274 static int alua_check_tpgs(struct scsi_device *sdev)
275 {
276 int tpgs = TPGS_MODE_NONE;
277
278 /*
279 * ALUA support for non-disk devices is fraught with
280 * difficulties, so disable it for now.
281 */
282 if (sdev->type != TYPE_DISK) {
283 sdev_printk(KERN_INFO, sdev,
284 "%s: disable for non-disk devices\n",
285 ALUA_DH_NAME);
286 return tpgs;
287 }
288
289 tpgs = scsi_device_tpgs(sdev);
290 switch (tpgs) {
291 case TPGS_MODE_EXPLICIT|TPGS_MODE_IMPLICIT:
292 sdev_printk(KERN_INFO, sdev,
293 "%s: supports implicit and explicit TPGS\n",
294 ALUA_DH_NAME);
295 break;
296 case TPGS_MODE_EXPLICIT:
297 sdev_printk(KERN_INFO, sdev, "%s: supports explicit TPGS\n",
298 ALUA_DH_NAME);
299 break;
300 case TPGS_MODE_IMPLICIT:
301 sdev_printk(KERN_INFO, sdev, "%s: supports implicit TPGS\n",
302 ALUA_DH_NAME);
303 break;
304 case TPGS_MODE_NONE:
305 sdev_printk(KERN_INFO, sdev, "%s: not supported\n",
306 ALUA_DH_NAME);
307 break;
308 default:
309 sdev_printk(KERN_INFO, sdev,
310 "%s: unsupported TPGS setting %d\n",
311 ALUA_DH_NAME, tpgs);
312 tpgs = TPGS_MODE_NONE;
313 break;
314 }
315
316 return tpgs;
317 }
318
319 /*
320 * alua_check_vpd - Evaluate INQUIRY vpd page 0x83
321 * @sdev: device to be checked
322 *
323 * Extract the relative target port and the target port group
324 * descriptor from the list of identificators.
325 */
alua_check_vpd(struct scsi_device * sdev,struct alua_dh_data * h,int tpgs)326 static int alua_check_vpd(struct scsi_device *sdev, struct alua_dh_data *h,
327 int tpgs)
328 {
329 int rel_port = -1, group_id;
330 struct alua_port_group *pg, *old_pg = NULL;
331 bool pg_updated = false;
332 unsigned long flags;
333
334 group_id = scsi_vpd_tpg_id(sdev, &rel_port);
335 if (group_id < 0) {
336 /*
337 * Internal error; TPGS supported but required
338 * VPD identification descriptors not present.
339 * Disable ALUA support
340 */
341 sdev_printk(KERN_INFO, sdev,
342 "%s: No target port descriptors found\n",
343 ALUA_DH_NAME);
344 return SCSI_DH_DEV_UNSUPP;
345 }
346
347 pg = alua_alloc_pg(sdev, group_id, tpgs);
348 if (IS_ERR(pg)) {
349 if (PTR_ERR(pg) == -ENOMEM)
350 return SCSI_DH_NOMEM;
351 return SCSI_DH_DEV_UNSUPP;
352 }
353 if (pg->device_id_len)
354 sdev_printk(KERN_INFO, sdev,
355 "%s: device %s port group %x rel port %x\n",
356 ALUA_DH_NAME, pg->device_id_str,
357 group_id, rel_port);
358 else
359 sdev_printk(KERN_INFO, sdev,
360 "%s: port group %x rel port %x\n",
361 ALUA_DH_NAME, group_id, rel_port);
362
363 kref_get(&pg->kref);
364
365 /* Check for existing port group references */
366 spin_lock(&h->pg_lock);
367 old_pg = rcu_dereference_protected(h->pg, lockdep_is_held(&h->pg_lock));
368 if (old_pg != pg) {
369 /* port group has changed. Update to new port group */
370 if (h->pg) {
371 spin_lock_irqsave(&old_pg->lock, flags);
372 list_del_rcu(&h->node);
373 spin_unlock_irqrestore(&old_pg->lock, flags);
374 }
375 rcu_assign_pointer(h->pg, pg);
376 pg_updated = true;
377 }
378
379 spin_lock_irqsave(&pg->lock, flags);
380 if (pg_updated)
381 list_add_rcu(&h->node, &pg->dh_list);
382 spin_unlock_irqrestore(&pg->lock, flags);
383
384 spin_unlock(&h->pg_lock);
385
386 alua_rtpg_queue(pg, sdev, NULL, true);
387 kref_put(&pg->kref, release_port_group);
388
389 if (old_pg)
390 kref_put(&old_pg->kref, release_port_group);
391
392 return SCSI_DH_OK;
393 }
394
print_alua_state(unsigned char state)395 static char print_alua_state(unsigned char state)
396 {
397 switch (state) {
398 case SCSI_ACCESS_STATE_OPTIMAL:
399 return 'A';
400 case SCSI_ACCESS_STATE_ACTIVE:
401 return 'N';
402 case SCSI_ACCESS_STATE_STANDBY:
403 return 'S';
404 case SCSI_ACCESS_STATE_UNAVAILABLE:
405 return 'U';
406 case SCSI_ACCESS_STATE_LBA:
407 return 'L';
408 case SCSI_ACCESS_STATE_OFFLINE:
409 return 'O';
410 case SCSI_ACCESS_STATE_TRANSITIONING:
411 return 'T';
412 default:
413 return 'X';
414 }
415 }
416
alua_handle_state_transition(struct scsi_device * sdev)417 static void alua_handle_state_transition(struct scsi_device *sdev)
418 {
419 struct alua_dh_data *h = sdev->handler_data;
420 struct alua_port_group *pg;
421
422 rcu_read_lock();
423 pg = rcu_dereference(h->pg);
424 if (pg)
425 pg->state = SCSI_ACCESS_STATE_TRANSITIONING;
426 rcu_read_unlock();
427 alua_check(sdev, false);
428 }
429
alua_check_sense(struct scsi_device * sdev,struct scsi_sense_hdr * sense_hdr)430 static enum scsi_disposition alua_check_sense(struct scsi_device *sdev,
431 struct scsi_sense_hdr *sense_hdr)
432 {
433 switch (sense_hdr->sense_key) {
434 case NOT_READY:
435 if (sense_hdr->asc == 0x04 && sense_hdr->ascq == 0x0a) {
436 /*
437 * LUN Not Accessible - ALUA state transition
438 */
439 alua_handle_state_transition(sdev);
440 return NEEDS_RETRY;
441 }
442 break;
443 case UNIT_ATTENTION:
444 if (sense_hdr->asc == 0x04 && sense_hdr->ascq == 0x0a) {
445 /*
446 * LUN Not Accessible - ALUA state transition
447 */
448 alua_handle_state_transition(sdev);
449 return NEEDS_RETRY;
450 }
451 if (sense_hdr->asc == 0x29 && sense_hdr->ascq == 0x00) {
452 /*
453 * Power On, Reset, or Bus Device Reset.
454 * Might have obscured a state transition,
455 * so schedule a recheck.
456 */
457 alua_check(sdev, true);
458 return ADD_TO_MLQUEUE;
459 }
460 if (sense_hdr->asc == 0x29 && sense_hdr->ascq == 0x04)
461 /*
462 * Device internal reset
463 */
464 return ADD_TO_MLQUEUE;
465 if (sense_hdr->asc == 0x2a && sense_hdr->ascq == 0x01)
466 /*
467 * Mode Parameters Changed
468 */
469 return ADD_TO_MLQUEUE;
470 if (sense_hdr->asc == 0x2a && sense_hdr->ascq == 0x06) {
471 /*
472 * ALUA state changed
473 */
474 alua_check(sdev, true);
475 return ADD_TO_MLQUEUE;
476 }
477 if (sense_hdr->asc == 0x2a && sense_hdr->ascq == 0x07) {
478 /*
479 * Implicit ALUA state transition failed
480 */
481 alua_check(sdev, true);
482 return ADD_TO_MLQUEUE;
483 }
484 if (sense_hdr->asc == 0x3f && sense_hdr->ascq == 0x03)
485 /*
486 * Inquiry data has changed
487 */
488 return ADD_TO_MLQUEUE;
489 if (sense_hdr->asc == 0x3f && sense_hdr->ascq == 0x0e)
490 /*
491 * REPORTED_LUNS_DATA_HAS_CHANGED is reported
492 * when switching controllers on targets like
493 * Intel Multi-Flex. We can just retry.
494 */
495 return ADD_TO_MLQUEUE;
496 break;
497 }
498
499 return SCSI_RETURN_NOT_HANDLED;
500 }
501
502 /*
503 * alua_tur - Send a TEST UNIT READY
504 * @sdev: device to which the TEST UNIT READY command should be send
505 *
506 * Send a TEST UNIT READY to @sdev to figure out the device state
507 * Returns SCSI_DH_RETRY if the sense code is NOT READY/ALUA TRANSITIONING,
508 * SCSI_DH_OK if no error occurred, and SCSI_DH_IO otherwise.
509 */
alua_tur(struct scsi_device * sdev)510 static int alua_tur(struct scsi_device *sdev)
511 {
512 struct scsi_sense_hdr sense_hdr;
513 int retval;
514
515 retval = scsi_test_unit_ready(sdev, ALUA_FAILOVER_TIMEOUT * HZ,
516 ALUA_FAILOVER_RETRIES, &sense_hdr);
517 if ((sense_hdr.sense_key == NOT_READY ||
518 sense_hdr.sense_key == UNIT_ATTENTION) &&
519 sense_hdr.asc == 0x04 && sense_hdr.ascq == 0x0a)
520 return SCSI_DH_RETRY;
521 else if (retval)
522 return SCSI_DH_IO;
523 else
524 return SCSI_DH_OK;
525 }
526
527 /*
528 * alua_rtpg - Evaluate REPORT TARGET GROUP STATES
529 * @sdev: the device to be evaluated.
530 *
531 * Evaluate the Target Port Group State.
532 * Returns SCSI_DH_DEV_OFFLINED if the path is
533 * found to be unusable.
534 */
alua_rtpg(struct scsi_device * sdev,struct alua_port_group * pg)535 static int alua_rtpg(struct scsi_device *sdev, struct alua_port_group *pg)
536 {
537 struct scsi_sense_hdr sense_hdr;
538 struct alua_port_group *tmp_pg;
539 int len, k, off, bufflen = ALUA_RTPG_SIZE;
540 int group_id_old, state_old, pref_old, valid_states_old;
541 unsigned char *desc, *buff;
542 unsigned err;
543 int retval;
544 unsigned int tpg_desc_tbl_off;
545 unsigned char orig_transition_tmo;
546 unsigned long flags;
547 bool transitioning_sense = false;
548
549 group_id_old = pg->group_id;
550 state_old = pg->state;
551 pref_old = pg->pref;
552 valid_states_old = pg->valid_states;
553
554 if (!pg->expiry) {
555 unsigned long transition_tmo = ALUA_FAILOVER_TIMEOUT * HZ;
556
557 if (pg->transition_tmo)
558 transition_tmo = pg->transition_tmo * HZ;
559
560 pg->expiry = round_jiffies_up(jiffies + transition_tmo);
561 }
562
563 buff = kzalloc(bufflen, GFP_KERNEL);
564 if (!buff)
565 return SCSI_DH_DEV_TEMP_BUSY;
566
567 retry:
568 err = 0;
569 retval = submit_rtpg(sdev, buff, bufflen, &sense_hdr, pg->flags);
570
571 if (retval) {
572 /*
573 * Some (broken) implementations have a habit of returning
574 * an error during things like firmware update etc.
575 * But if the target only supports active/optimized there's
576 * not much we can do; it's not that we can switch paths
577 * or anything.
578 * So ignore any errors to avoid spurious failures during
579 * path failover.
580 */
581 if ((pg->valid_states & ~TPGS_SUPPORT_OPTIMIZED) == 0) {
582 sdev_printk(KERN_INFO, sdev,
583 "%s: ignoring rtpg result %d\n",
584 ALUA_DH_NAME, retval);
585 kfree(buff);
586 return SCSI_DH_OK;
587 }
588 if (retval < 0 || !scsi_sense_valid(&sense_hdr)) {
589 sdev_printk(KERN_INFO, sdev,
590 "%s: rtpg failed, result %d\n",
591 ALUA_DH_NAME, retval);
592 kfree(buff);
593 if (retval < 0)
594 return SCSI_DH_DEV_TEMP_BUSY;
595 if (host_byte(retval) == DID_NO_CONNECT)
596 return SCSI_DH_RES_TEMP_UNAVAIL;
597 return SCSI_DH_IO;
598 }
599
600 /*
601 * submit_rtpg() has failed on existing arrays
602 * when requesting extended header info, and
603 * the array doesn't support extended headers,
604 * even though it shouldn't according to T10.
605 * The retry without rtpg_ext_hdr_req set
606 * handles this.
607 * Note: some arrays return a sense key of ILLEGAL_REQUEST
608 * with ASC 00h if they don't support the extended header.
609 */
610 if (!(pg->flags & ALUA_RTPG_EXT_HDR_UNSUPP) &&
611 sense_hdr.sense_key == ILLEGAL_REQUEST) {
612 pg->flags |= ALUA_RTPG_EXT_HDR_UNSUPP;
613 goto retry;
614 }
615 /*
616 * If the array returns with 'ALUA state transition'
617 * sense code here it cannot return RTPG data during
618 * transition. So set the state to 'transitioning' directly.
619 */
620 if (sense_hdr.sense_key == NOT_READY &&
621 sense_hdr.asc == 0x04 && sense_hdr.ascq == 0x0a) {
622 transitioning_sense = true;
623 goto skip_rtpg;
624 }
625 /*
626 * Retry on any other UNIT ATTENTION occurred.
627 */
628 if (sense_hdr.sense_key == UNIT_ATTENTION)
629 err = SCSI_DH_RETRY;
630 if (err == SCSI_DH_RETRY &&
631 pg->expiry != 0 && time_before(jiffies, pg->expiry)) {
632 sdev_printk(KERN_ERR, sdev, "%s: rtpg retry\n",
633 ALUA_DH_NAME);
634 scsi_print_sense_hdr(sdev, ALUA_DH_NAME, &sense_hdr);
635 kfree(buff);
636 return err;
637 }
638 sdev_printk(KERN_ERR, sdev, "%s: rtpg failed\n",
639 ALUA_DH_NAME);
640 scsi_print_sense_hdr(sdev, ALUA_DH_NAME, &sense_hdr);
641 kfree(buff);
642 pg->expiry = 0;
643 return SCSI_DH_IO;
644 }
645
646 len = get_unaligned_be32(&buff[0]) + 4;
647
648 if (len > bufflen) {
649 /* Resubmit with the correct length */
650 kfree(buff);
651 bufflen = len;
652 buff = kmalloc(bufflen, GFP_KERNEL);
653 if (!buff) {
654 sdev_printk(KERN_WARNING, sdev,
655 "%s: kmalloc buffer failed\n",__func__);
656 /* Temporary failure, bypass */
657 pg->expiry = 0;
658 return SCSI_DH_DEV_TEMP_BUSY;
659 }
660 goto retry;
661 }
662
663 orig_transition_tmo = pg->transition_tmo;
664 if ((buff[4] & RTPG_FMT_MASK) == RTPG_FMT_EXT_HDR && buff[5] != 0)
665 pg->transition_tmo = buff[5];
666 else
667 pg->transition_tmo = ALUA_FAILOVER_TIMEOUT;
668
669 if (orig_transition_tmo != pg->transition_tmo) {
670 sdev_printk(KERN_INFO, sdev,
671 "%s: transition timeout set to %d seconds\n",
672 ALUA_DH_NAME, pg->transition_tmo);
673 pg->expiry = jiffies + pg->transition_tmo * HZ;
674 }
675
676 if ((buff[4] & RTPG_FMT_MASK) == RTPG_FMT_EXT_HDR)
677 tpg_desc_tbl_off = 8;
678 else
679 tpg_desc_tbl_off = 4;
680
681 for (k = tpg_desc_tbl_off, desc = buff + tpg_desc_tbl_off;
682 k < len;
683 k += off, desc += off) {
684 u16 group_id = get_unaligned_be16(&desc[2]);
685
686 spin_lock_irqsave(&port_group_lock, flags);
687 tmp_pg = alua_find_get_pg(pg->device_id_str, pg->device_id_len,
688 group_id);
689 spin_unlock_irqrestore(&port_group_lock, flags);
690 if (tmp_pg) {
691 if (spin_trylock_irqsave(&tmp_pg->lock, flags)) {
692 if ((tmp_pg == pg) ||
693 !(tmp_pg->flags & ALUA_PG_RUNNING)) {
694 struct alua_dh_data *h;
695
696 tmp_pg->state = desc[0] & 0x0f;
697 tmp_pg->pref = desc[0] >> 7;
698 rcu_read_lock();
699 list_for_each_entry_rcu(h,
700 &tmp_pg->dh_list, node) {
701 if (!h->sdev)
702 continue;
703 h->sdev->access_state = desc[0];
704 }
705 rcu_read_unlock();
706 }
707 if (tmp_pg == pg)
708 tmp_pg->valid_states = desc[1];
709 spin_unlock_irqrestore(&tmp_pg->lock, flags);
710 }
711 kref_put(&tmp_pg->kref, release_port_group);
712 }
713 off = 8 + (desc[7] * 4);
714 }
715
716 skip_rtpg:
717 spin_lock_irqsave(&pg->lock, flags);
718 if (transitioning_sense)
719 pg->state = SCSI_ACCESS_STATE_TRANSITIONING;
720
721 if (group_id_old != pg->group_id || state_old != pg->state ||
722 pref_old != pg->pref || valid_states_old != pg->valid_states)
723 sdev_printk(KERN_INFO, sdev,
724 "%s: port group %02x state %c %s supports %c%c%c%c%c%c%c\n",
725 ALUA_DH_NAME, pg->group_id, print_alua_state(pg->state),
726 pg->pref ? "preferred" : "non-preferred",
727 pg->valid_states&TPGS_SUPPORT_TRANSITION?'T':'t',
728 pg->valid_states&TPGS_SUPPORT_OFFLINE?'O':'o',
729 pg->valid_states&TPGS_SUPPORT_LBA_DEPENDENT?'L':'l',
730 pg->valid_states&TPGS_SUPPORT_UNAVAILABLE?'U':'u',
731 pg->valid_states&TPGS_SUPPORT_STANDBY?'S':'s',
732 pg->valid_states&TPGS_SUPPORT_NONOPTIMIZED?'N':'n',
733 pg->valid_states&TPGS_SUPPORT_OPTIMIZED?'A':'a');
734
735 switch (pg->state) {
736 case SCSI_ACCESS_STATE_TRANSITIONING:
737 if (time_before(jiffies, pg->expiry)) {
738 /* State transition, retry */
739 pg->interval = ALUA_RTPG_RETRY_DELAY;
740 err = SCSI_DH_RETRY;
741 } else {
742 struct alua_dh_data *h;
743
744 /* Transitioning time exceeded, set port to standby */
745 err = SCSI_DH_IO;
746 pg->state = SCSI_ACCESS_STATE_STANDBY;
747 pg->expiry = 0;
748 rcu_read_lock();
749 list_for_each_entry_rcu(h, &pg->dh_list, node) {
750 if (!h->sdev)
751 continue;
752 h->sdev->access_state =
753 (pg->state & SCSI_ACCESS_STATE_MASK);
754 if (pg->pref)
755 h->sdev->access_state |=
756 SCSI_ACCESS_STATE_PREFERRED;
757 }
758 rcu_read_unlock();
759 }
760 break;
761 case SCSI_ACCESS_STATE_OFFLINE:
762 /* Path unusable */
763 err = SCSI_DH_DEV_OFFLINED;
764 pg->expiry = 0;
765 break;
766 default:
767 /* Useable path if active */
768 err = SCSI_DH_OK;
769 pg->expiry = 0;
770 break;
771 }
772 spin_unlock_irqrestore(&pg->lock, flags);
773 kfree(buff);
774 return err;
775 }
776
777 /*
778 * alua_stpg - Issue a SET TARGET PORT GROUP command
779 *
780 * Issue a SET TARGET PORT GROUP command and evaluate the
781 * response. Returns SCSI_DH_RETRY per default to trigger
782 * a re-evaluation of the target group state or SCSI_DH_OK
783 * if no further action needs to be taken.
784 */
alua_stpg(struct scsi_device * sdev,struct alua_port_group * pg)785 static unsigned alua_stpg(struct scsi_device *sdev, struct alua_port_group *pg)
786 {
787 int retval;
788 struct scsi_sense_hdr sense_hdr;
789
790 if (!(pg->tpgs & TPGS_MODE_EXPLICIT)) {
791 /* Only implicit ALUA supported, retry */
792 return SCSI_DH_RETRY;
793 }
794 switch (pg->state) {
795 case SCSI_ACCESS_STATE_OPTIMAL:
796 return SCSI_DH_OK;
797 case SCSI_ACCESS_STATE_ACTIVE:
798 if ((pg->flags & ALUA_OPTIMIZE_STPG) &&
799 !pg->pref &&
800 (pg->tpgs & TPGS_MODE_IMPLICIT))
801 return SCSI_DH_OK;
802 break;
803 case SCSI_ACCESS_STATE_STANDBY:
804 case SCSI_ACCESS_STATE_UNAVAILABLE:
805 break;
806 case SCSI_ACCESS_STATE_OFFLINE:
807 return SCSI_DH_IO;
808 case SCSI_ACCESS_STATE_TRANSITIONING:
809 break;
810 default:
811 sdev_printk(KERN_INFO, sdev,
812 "%s: stpg failed, unhandled TPGS state %d",
813 ALUA_DH_NAME, pg->state);
814 return SCSI_DH_NOSYS;
815 }
816 retval = submit_stpg(sdev, pg->group_id, &sense_hdr);
817
818 if (retval) {
819 if (retval < 0 || !scsi_sense_valid(&sense_hdr)) {
820 sdev_printk(KERN_INFO, sdev,
821 "%s: stpg failed, result %d",
822 ALUA_DH_NAME, retval);
823 if (retval < 0)
824 return SCSI_DH_DEV_TEMP_BUSY;
825 } else {
826 sdev_printk(KERN_INFO, sdev, "%s: stpg failed\n",
827 ALUA_DH_NAME);
828 scsi_print_sense_hdr(sdev, ALUA_DH_NAME, &sense_hdr);
829 }
830 }
831 /* Retry RTPG */
832 return SCSI_DH_RETRY;
833 }
834
835 /*
836 * The caller must call scsi_device_put() on the returned pointer if it is not
837 * NULL.
838 */
839 static struct scsi_device * __must_check
alua_rtpg_select_sdev(struct alua_port_group * pg)840 alua_rtpg_select_sdev(struct alua_port_group *pg)
841 {
842 struct alua_dh_data *h;
843 struct scsi_device *sdev = NULL, *prev_sdev;
844
845 lockdep_assert_held(&pg->lock);
846 if (WARN_ON(!pg->rtpg_sdev))
847 return NULL;
848
849 /*
850 * RCU protection isn't necessary for dh_list here
851 * as we hold pg->lock, but for access to h->pg.
852 */
853 rcu_read_lock();
854 list_for_each_entry_rcu(h, &pg->dh_list, node) {
855 if (!h->sdev)
856 continue;
857 if (h->sdev == pg->rtpg_sdev) {
858 h->disabled = true;
859 continue;
860 }
861 if (rcu_dereference(h->pg) == pg &&
862 !h->disabled &&
863 !scsi_device_get(h->sdev)) {
864 sdev = h->sdev;
865 break;
866 }
867 }
868 rcu_read_unlock();
869
870 if (!sdev) {
871 pr_warn("%s: no device found for rtpg\n",
872 (pg->device_id_len ?
873 (char *)pg->device_id_str : "(nameless PG)"));
874 return NULL;
875 }
876
877 sdev_printk(KERN_INFO, sdev, "rtpg retry on different device\n");
878
879 prev_sdev = pg->rtpg_sdev;
880 pg->rtpg_sdev = sdev;
881
882 return prev_sdev;
883 }
884
alua_rtpg_work(struct work_struct * work)885 static void alua_rtpg_work(struct work_struct *work)
886 {
887 struct alua_port_group *pg =
888 container_of(work, struct alua_port_group, rtpg_work.work);
889 struct scsi_device *sdev, *prev_sdev = NULL;
890 LIST_HEAD(qdata_list);
891 int err = SCSI_DH_OK;
892 struct alua_queue_data *qdata, *tmp;
893 struct alua_dh_data *h;
894 unsigned long flags;
895
896 spin_lock_irqsave(&pg->lock, flags);
897 sdev = pg->rtpg_sdev;
898 if (!sdev) {
899 WARN_ON(pg->flags & ALUA_PG_RUN_RTPG);
900 WARN_ON(pg->flags & ALUA_PG_RUN_STPG);
901 spin_unlock_irqrestore(&pg->lock, flags);
902 kref_put(&pg->kref, release_port_group);
903 return;
904 }
905 pg->flags |= ALUA_PG_RUNNING;
906 if (pg->flags & ALUA_PG_RUN_RTPG) {
907 int state = pg->state;
908
909 pg->flags &= ~ALUA_PG_RUN_RTPG;
910 spin_unlock_irqrestore(&pg->lock, flags);
911 if (state == SCSI_ACCESS_STATE_TRANSITIONING) {
912 if (alua_tur(sdev) == SCSI_DH_RETRY) {
913 spin_lock_irqsave(&pg->lock, flags);
914 pg->flags &= ~ALUA_PG_RUNNING;
915 pg->flags |= ALUA_PG_RUN_RTPG;
916 if (!pg->interval)
917 pg->interval = ALUA_RTPG_RETRY_DELAY;
918 spin_unlock_irqrestore(&pg->lock, flags);
919 queue_delayed_work(kaluad_wq, &pg->rtpg_work,
920 pg->interval * HZ);
921 return;
922 }
923 /* Send RTPG on failure or if TUR indicates SUCCESS */
924 }
925 err = alua_rtpg(sdev, pg);
926 spin_lock_irqsave(&pg->lock, flags);
927
928 /* If RTPG failed on the current device, try using another */
929 if (err == SCSI_DH_RES_TEMP_UNAVAIL &&
930 (prev_sdev = alua_rtpg_select_sdev(pg)))
931 err = SCSI_DH_IMM_RETRY;
932
933 if (err == SCSI_DH_RETRY || err == SCSI_DH_IMM_RETRY ||
934 pg->flags & ALUA_PG_RUN_RTPG) {
935 pg->flags &= ~ALUA_PG_RUNNING;
936 if (err == SCSI_DH_IMM_RETRY)
937 pg->interval = 0;
938 else if (!pg->interval && !(pg->flags & ALUA_PG_RUN_RTPG))
939 pg->interval = ALUA_RTPG_RETRY_DELAY;
940 pg->flags |= ALUA_PG_RUN_RTPG;
941 spin_unlock_irqrestore(&pg->lock, flags);
942 goto queue_rtpg;
943 }
944 if (err != SCSI_DH_OK)
945 pg->flags &= ~ALUA_PG_RUN_STPG;
946 }
947 if (pg->flags & ALUA_PG_RUN_STPG) {
948 pg->flags &= ~ALUA_PG_RUN_STPG;
949 spin_unlock_irqrestore(&pg->lock, flags);
950 err = alua_stpg(sdev, pg);
951 spin_lock_irqsave(&pg->lock, flags);
952 if (err == SCSI_DH_RETRY || pg->flags & ALUA_PG_RUN_RTPG) {
953 pg->flags |= ALUA_PG_RUN_RTPG;
954 pg->interval = 0;
955 pg->flags &= ~ALUA_PG_RUNNING;
956 spin_unlock_irqrestore(&pg->lock, flags);
957 goto queue_rtpg;
958 }
959 }
960
961 list_splice_init(&pg->rtpg_list, &qdata_list);
962 /*
963 * We went through an RTPG, for good or bad.
964 * Re-enable all devices for the next attempt.
965 */
966 list_for_each_entry(h, &pg->dh_list, node)
967 h->disabled = false;
968 pg->rtpg_sdev = NULL;
969 spin_unlock_irqrestore(&pg->lock, flags);
970
971 if (prev_sdev)
972 scsi_device_put(prev_sdev);
973
974 list_for_each_entry_safe(qdata, tmp, &qdata_list, entry) {
975 list_del(&qdata->entry);
976 if (qdata->callback_fn)
977 qdata->callback_fn(qdata->callback_data, err);
978 kfree(qdata);
979 }
980 spin_lock_irqsave(&pg->lock, flags);
981 pg->flags &= ~ALUA_PG_RUNNING;
982 spin_unlock_irqrestore(&pg->lock, flags);
983 scsi_device_put(sdev);
984 kref_put(&pg->kref, release_port_group);
985 return;
986
987 queue_rtpg:
988 if (prev_sdev)
989 scsi_device_put(prev_sdev);
990 queue_delayed_work(kaluad_wq, &pg->rtpg_work, pg->interval * HZ);
991 }
992
993 /**
994 * alua_rtpg_queue() - cause RTPG to be submitted asynchronously
995 * @pg: ALUA port group associated with @sdev.
996 * @sdev: SCSI device for which to submit an RTPG.
997 * @qdata: Information about the callback to invoke after the RTPG.
998 * @force: Whether or not to submit an RTPG if a work item that will submit an
999 * RTPG already has been scheduled.
1000 *
1001 * Returns true if and only if alua_rtpg_work() will be called asynchronously.
1002 * That function is responsible for calling @qdata->fn().
1003 *
1004 * Context: may be called from atomic context (alua_check()) only if the caller
1005 * holds an sdev reference.
1006 */
alua_rtpg_queue(struct alua_port_group * pg,struct scsi_device * sdev,struct alua_queue_data * qdata,bool force)1007 static bool alua_rtpg_queue(struct alua_port_group *pg,
1008 struct scsi_device *sdev,
1009 struct alua_queue_data *qdata, bool force)
1010 {
1011 int start_queue = 0;
1012 unsigned long flags;
1013
1014 if (WARN_ON_ONCE(!pg) || scsi_device_get(sdev))
1015 return false;
1016
1017 spin_lock_irqsave(&pg->lock, flags);
1018 if (qdata) {
1019 list_add_tail(&qdata->entry, &pg->rtpg_list);
1020 pg->flags |= ALUA_PG_RUN_STPG;
1021 force = true;
1022 }
1023 if (pg->rtpg_sdev == NULL) {
1024 struct alua_dh_data *h = sdev->handler_data;
1025
1026 rcu_read_lock();
1027 if (h && rcu_dereference(h->pg) == pg) {
1028 pg->interval = 0;
1029 pg->flags |= ALUA_PG_RUN_RTPG;
1030 kref_get(&pg->kref);
1031 pg->rtpg_sdev = sdev;
1032 start_queue = 1;
1033 }
1034 rcu_read_unlock();
1035 } else if (!(pg->flags & ALUA_PG_RUN_RTPG) && force) {
1036 pg->flags |= ALUA_PG_RUN_RTPG;
1037 /* Do not queue if the worker is already running */
1038 if (!(pg->flags & ALUA_PG_RUNNING)) {
1039 kref_get(&pg->kref);
1040 start_queue = 1;
1041 }
1042 }
1043
1044 spin_unlock_irqrestore(&pg->lock, flags);
1045
1046 if (start_queue) {
1047 if (queue_delayed_work(kaluad_wq, &pg->rtpg_work,
1048 msecs_to_jiffies(ALUA_RTPG_DELAY_MSECS)))
1049 sdev = NULL;
1050 else
1051 kref_put(&pg->kref, release_port_group);
1052 }
1053 if (sdev)
1054 scsi_device_put(sdev);
1055
1056 return true;
1057 }
1058
1059 /*
1060 * alua_initialize - Initialize ALUA state
1061 * @sdev: the device to be initialized
1062 *
1063 * For the prep_fn to work correctly we have
1064 * to initialize the ALUA state for the device.
1065 */
alua_initialize(struct scsi_device * sdev,struct alua_dh_data * h)1066 static int alua_initialize(struct scsi_device *sdev, struct alua_dh_data *h)
1067 {
1068 int err = SCSI_DH_DEV_UNSUPP, tpgs;
1069
1070 mutex_lock(&h->init_mutex);
1071 h->disabled = false;
1072 tpgs = alua_check_tpgs(sdev);
1073 if (tpgs != TPGS_MODE_NONE)
1074 err = alua_check_vpd(sdev, h, tpgs);
1075 h->init_error = err;
1076 mutex_unlock(&h->init_mutex);
1077 return err;
1078 }
1079 /*
1080 * alua_set_params - set/unset the optimize flag
1081 * @sdev: device on the path to be activated
1082 * params - parameters in the following format
1083 * "no_of_params\0param1\0param2\0param3\0...\0"
1084 * For example, to set the flag pass the following parameters
1085 * from multipath.conf
1086 * hardware_handler "2 alua 1"
1087 */
alua_set_params(struct scsi_device * sdev,const char * params)1088 static int alua_set_params(struct scsi_device *sdev, const char *params)
1089 {
1090 struct alua_dh_data *h = sdev->handler_data;
1091 struct alua_port_group *pg = NULL;
1092 unsigned int optimize = 0, argc;
1093 const char *p = params;
1094 int result = SCSI_DH_OK;
1095 unsigned long flags;
1096
1097 if ((sscanf(params, "%u", &argc) != 1) || (argc != 1))
1098 return -EINVAL;
1099
1100 while (*p++)
1101 ;
1102 if ((sscanf(p, "%u", &optimize) != 1) || (optimize > 1))
1103 return -EINVAL;
1104
1105 rcu_read_lock();
1106 pg = rcu_dereference(h->pg);
1107 if (!pg) {
1108 rcu_read_unlock();
1109 return -ENXIO;
1110 }
1111 spin_lock_irqsave(&pg->lock, flags);
1112 if (optimize)
1113 pg->flags |= ALUA_OPTIMIZE_STPG;
1114 else
1115 pg->flags &= ~ALUA_OPTIMIZE_STPG;
1116 spin_unlock_irqrestore(&pg->lock, flags);
1117 rcu_read_unlock();
1118
1119 return result;
1120 }
1121
1122 /*
1123 * alua_activate - activate a path
1124 * @sdev: device on the path to be activated
1125 *
1126 * We're currently switching the port group to be activated only and
1127 * let the array figure out the rest.
1128 * There may be other arrays which require us to switch all port groups
1129 * based on a certain policy. But until we actually encounter them it
1130 * should be okay.
1131 */
alua_activate(struct scsi_device * sdev,activate_complete fn,void * data)1132 static int alua_activate(struct scsi_device *sdev,
1133 activate_complete fn, void *data)
1134 {
1135 struct alua_dh_data *h = sdev->handler_data;
1136 int err = SCSI_DH_OK;
1137 struct alua_queue_data *qdata;
1138 struct alua_port_group *pg;
1139
1140 qdata = kzalloc(sizeof(*qdata), GFP_KERNEL);
1141 if (!qdata) {
1142 err = SCSI_DH_RES_TEMP_UNAVAIL;
1143 goto out;
1144 }
1145 qdata->callback_fn = fn;
1146 qdata->callback_data = data;
1147
1148 mutex_lock(&h->init_mutex);
1149 rcu_read_lock();
1150 pg = rcu_dereference(h->pg);
1151 if (!pg || !kref_get_unless_zero(&pg->kref)) {
1152 rcu_read_unlock();
1153 kfree(qdata);
1154 err = h->init_error;
1155 mutex_unlock(&h->init_mutex);
1156 goto out;
1157 }
1158 rcu_read_unlock();
1159 mutex_unlock(&h->init_mutex);
1160
1161 if (alua_rtpg_queue(pg, sdev, qdata, true)) {
1162 fn = NULL;
1163 } else {
1164 kfree(qdata);
1165 err = SCSI_DH_DEV_OFFLINED;
1166 }
1167 kref_put(&pg->kref, release_port_group);
1168 out:
1169 if (fn)
1170 fn(data, err);
1171 return 0;
1172 }
1173
1174 /*
1175 * alua_check - check path status
1176 * @sdev: device on the path to be checked
1177 *
1178 * Check the device status
1179 */
alua_check(struct scsi_device * sdev,bool force)1180 static void alua_check(struct scsi_device *sdev, bool force)
1181 {
1182 struct alua_dh_data *h = sdev->handler_data;
1183 struct alua_port_group *pg;
1184
1185 rcu_read_lock();
1186 pg = rcu_dereference(h->pg);
1187 if (!pg || !kref_get_unless_zero(&pg->kref)) {
1188 rcu_read_unlock();
1189 return;
1190 }
1191 rcu_read_unlock();
1192 alua_rtpg_queue(pg, sdev, NULL, force);
1193 kref_put(&pg->kref, release_port_group);
1194 }
1195
1196 /*
1197 * alua_prep_fn - request callback
1198 *
1199 * Fail I/O to all paths not in state
1200 * active/optimized or active/non-optimized.
1201 */
alua_prep_fn(struct scsi_device * sdev,struct request * req)1202 static blk_status_t alua_prep_fn(struct scsi_device *sdev, struct request *req)
1203 {
1204 struct alua_dh_data *h = sdev->handler_data;
1205 struct alua_port_group *pg;
1206 unsigned char state = SCSI_ACCESS_STATE_OPTIMAL;
1207
1208 rcu_read_lock();
1209 pg = rcu_dereference(h->pg);
1210 if (pg)
1211 state = pg->state;
1212 rcu_read_unlock();
1213
1214 switch (state) {
1215 case SCSI_ACCESS_STATE_OPTIMAL:
1216 case SCSI_ACCESS_STATE_ACTIVE:
1217 case SCSI_ACCESS_STATE_LBA:
1218 case SCSI_ACCESS_STATE_TRANSITIONING:
1219 return BLK_STS_OK;
1220 default:
1221 req->rq_flags |= RQF_QUIET;
1222 return BLK_STS_IOERR;
1223 }
1224 }
1225
alua_rescan(struct scsi_device * sdev)1226 static void alua_rescan(struct scsi_device *sdev)
1227 {
1228 struct alua_dh_data *h = sdev->handler_data;
1229
1230 alua_initialize(sdev, h);
1231 }
1232
1233 /*
1234 * alua_bus_attach - Attach device handler
1235 * @sdev: device to be attached to
1236 */
alua_bus_attach(struct scsi_device * sdev)1237 static int alua_bus_attach(struct scsi_device *sdev)
1238 {
1239 struct alua_dh_data *h;
1240 int err;
1241
1242 h = kzalloc(sizeof(*h) , GFP_KERNEL);
1243 if (!h)
1244 return SCSI_DH_NOMEM;
1245 spin_lock_init(&h->pg_lock);
1246 rcu_assign_pointer(h->pg, NULL);
1247 h->init_error = SCSI_DH_OK;
1248 h->sdev = sdev;
1249 INIT_LIST_HEAD(&h->node);
1250
1251 mutex_init(&h->init_mutex);
1252 err = alua_initialize(sdev, h);
1253 if (err != SCSI_DH_OK && err != SCSI_DH_DEV_OFFLINED)
1254 goto failed;
1255
1256 sdev->handler_data = h;
1257 return SCSI_DH_OK;
1258 failed:
1259 kfree(h);
1260 return err;
1261 }
1262
1263 /*
1264 * alua_bus_detach - Detach device handler
1265 * @sdev: device to be detached from
1266 */
alua_bus_detach(struct scsi_device * sdev)1267 static void alua_bus_detach(struct scsi_device *sdev)
1268 {
1269 struct alua_dh_data *h = sdev->handler_data;
1270 struct alua_port_group *pg;
1271
1272 spin_lock(&h->pg_lock);
1273 pg = rcu_dereference_protected(h->pg, lockdep_is_held(&h->pg_lock));
1274 rcu_assign_pointer(h->pg, NULL);
1275 spin_unlock(&h->pg_lock);
1276 if (pg) {
1277 spin_lock_irq(&pg->lock);
1278 list_del_rcu(&h->node);
1279 spin_unlock_irq(&pg->lock);
1280 kref_put(&pg->kref, release_port_group);
1281 }
1282 sdev->handler_data = NULL;
1283 synchronize_rcu();
1284 kfree(h);
1285 }
1286
1287 static struct scsi_device_handler alua_dh = {
1288 .name = ALUA_DH_NAME,
1289 .module = THIS_MODULE,
1290 .attach = alua_bus_attach,
1291 .detach = alua_bus_detach,
1292 .prep_fn = alua_prep_fn,
1293 .check_sense = alua_check_sense,
1294 .activate = alua_activate,
1295 .rescan = alua_rescan,
1296 .set_params = alua_set_params,
1297 };
1298
alua_init(void)1299 static int __init alua_init(void)
1300 {
1301 int r;
1302
1303 kaluad_wq = alloc_workqueue("kaluad", WQ_MEM_RECLAIM, 0);
1304 if (!kaluad_wq)
1305 return -ENOMEM;
1306
1307 r = scsi_register_device_handler(&alua_dh);
1308 if (r != 0) {
1309 printk(KERN_ERR "%s: Failed to register scsi device handler",
1310 ALUA_DH_NAME);
1311 destroy_workqueue(kaluad_wq);
1312 }
1313 return r;
1314 }
1315
alua_exit(void)1316 static void __exit alua_exit(void)
1317 {
1318 scsi_unregister_device_handler(&alua_dh);
1319 destroy_workqueue(kaluad_wq);
1320 }
1321
1322 module_init(alua_init);
1323 module_exit(alua_exit);
1324
1325 MODULE_DESCRIPTION("DM Multipath ALUA support");
1326 MODULE_AUTHOR("Hannes Reinecke <hare@suse.de>");
1327 MODULE_LICENSE("GPL");
1328 MODULE_VERSION(ALUA_DH_VER);
1329