xref: /openbmc/linux/drivers/memstick/core/ms_block.c (revision cd238eff)
1 /*
2  *  ms_block.c - Sony MemoryStick (legacy) storage support
3 
4  *  Copyright (C) 2013 Maxim Levitsky <maximlevitsky@gmail.com>
5  *
6  * This program is free software; you can redistribute it and/or modify
7  * it under the terms of the GNU General Public License version 2 as
8  * published by the Free Software Foundation.
9  *
10  * Minor portions of the driver were copied from mspro_block.c which is
11  * Copyright (C) 2007 Alex Dubov <oakad@yahoo.com>
12  *
13  */
14 #define DRIVER_NAME "ms_block"
15 #define pr_fmt(fmt) DRIVER_NAME ": " fmt
16 
17 #include <linux/module.h>
18 #include <linux/blk-mq.h>
19 #include <linux/memstick.h>
20 #include <linux/idr.h>
21 #include <linux/hdreg.h>
22 #include <linux/delay.h>
23 #include <linux/slab.h>
24 #include <linux/random.h>
25 #include <linux/bitmap.h>
26 #include <linux/scatterlist.h>
27 #include <linux/jiffies.h>
28 #include <linux/workqueue.h>
29 #include <linux/mutex.h>
30 #include "ms_block.h"
31 
32 static int debug;
33 static int cache_flush_timeout = 1000;
34 static bool verify_writes;
35 
36 /*
37  * Copies section of 'sg_from' starting from offset 'offset' and with length
38  * 'len' To another scatterlist of to_nents enties
39  */
40 static size_t msb_sg_copy(struct scatterlist *sg_from,
41 	struct scatterlist *sg_to, int to_nents, size_t offset, size_t len)
42 {
43 	size_t copied = 0;
44 
45 	while (offset > 0) {
46 		if (offset >= sg_from->length) {
47 			if (sg_is_last(sg_from))
48 				return 0;
49 
50 			offset -= sg_from->length;
51 			sg_from = sg_next(sg_from);
52 			continue;
53 		}
54 
55 		copied = min(len, sg_from->length - offset);
56 		sg_set_page(sg_to, sg_page(sg_from),
57 			copied, sg_from->offset + offset);
58 
59 		len -= copied;
60 		offset = 0;
61 
62 		if (sg_is_last(sg_from) || !len)
63 			goto out;
64 
65 		sg_to = sg_next(sg_to);
66 		to_nents--;
67 		sg_from = sg_next(sg_from);
68 	}
69 
70 	while (len > sg_from->length && to_nents--) {
71 		len -= sg_from->length;
72 		copied += sg_from->length;
73 
74 		sg_set_page(sg_to, sg_page(sg_from),
75 				sg_from->length, sg_from->offset);
76 
77 		if (sg_is_last(sg_from) || !len)
78 			goto out;
79 
80 		sg_from = sg_next(sg_from);
81 		sg_to = sg_next(sg_to);
82 	}
83 
84 	if (len && to_nents) {
85 		sg_set_page(sg_to, sg_page(sg_from), len, sg_from->offset);
86 		copied += len;
87 	}
88 out:
89 	sg_mark_end(sg_to);
90 	return copied;
91 }
92 
93 /*
94  * Compares section of 'sg' starting from offset 'offset' and with length 'len'
95  * to linear buffer of length 'len' at address 'buffer'
96  * Returns 0 if equal and  -1 otherwice
97  */
98 static int msb_sg_compare_to_buffer(struct scatterlist *sg,
99 					size_t offset, u8 *buffer, size_t len)
100 {
101 	int retval = 0, cmplen;
102 	struct sg_mapping_iter miter;
103 
104 	sg_miter_start(&miter, sg, sg_nents(sg),
105 					SG_MITER_ATOMIC | SG_MITER_FROM_SG);
106 
107 	while (sg_miter_next(&miter) && len > 0) {
108 		if (offset >= miter.length) {
109 			offset -= miter.length;
110 			continue;
111 		}
112 
113 		cmplen = min(miter.length - offset, len);
114 		retval = memcmp(miter.addr + offset, buffer, cmplen) ? -1 : 0;
115 		if (retval)
116 			break;
117 
118 		buffer += cmplen;
119 		len -= cmplen;
120 		offset = 0;
121 	}
122 
123 	if (!retval && len)
124 		retval = -1;
125 
126 	sg_miter_stop(&miter);
127 	return retval;
128 }
129 
130 
131 /* Get zone at which block with logical address 'lba' lives
132  * Flash is broken into zones.
133  * Each zone consists of 512 eraseblocks, out of which in first
134  * zone 494 are used and 496 are for all following zones.
135  * Therefore zone #0 hosts blocks 0-493, zone #1 blocks 494-988, etc...
136 */
137 static int msb_get_zone_from_lba(int lba)
138 {
139 	if (lba < 494)
140 		return 0;
141 	return ((lba - 494) / 496) + 1;
142 }
143 
144 /* Get zone of physical block. Trivial */
145 static int msb_get_zone_from_pba(int pba)
146 {
147 	return pba / MS_BLOCKS_IN_ZONE;
148 }
149 
150 /* Debug test to validate free block counts */
151 static int msb_validate_used_block_bitmap(struct msb_data *msb)
152 {
153 	int total_free_blocks = 0;
154 	int i;
155 
156 	if (!debug)
157 		return 0;
158 
159 	for (i = 0; i < msb->zone_count; i++)
160 		total_free_blocks += msb->free_block_count[i];
161 
162 	if (msb->block_count - bitmap_weight(msb->used_blocks_bitmap,
163 					msb->block_count) == total_free_blocks)
164 		return 0;
165 
166 	pr_err("BUG: free block counts don't match the bitmap");
167 	msb->read_only = true;
168 	return -EINVAL;
169 }
170 
171 /* Mark physical block as used */
172 static void msb_mark_block_used(struct msb_data *msb, int pba)
173 {
174 	int zone = msb_get_zone_from_pba(pba);
175 
176 	if (test_bit(pba, msb->used_blocks_bitmap)) {
177 		pr_err(
178 		"BUG: attempt to mark already used pba %d as used", pba);
179 		msb->read_only = true;
180 		return;
181 	}
182 
183 	if (msb_validate_used_block_bitmap(msb))
184 		return;
185 
186 	/* No races because all IO is single threaded */
187 	__set_bit(pba, msb->used_blocks_bitmap);
188 	msb->free_block_count[zone]--;
189 }
190 
191 /* Mark physical block as free */
192 static void msb_mark_block_unused(struct msb_data *msb, int pba)
193 {
194 	int zone = msb_get_zone_from_pba(pba);
195 
196 	if (!test_bit(pba, msb->used_blocks_bitmap)) {
197 		pr_err("BUG: attempt to mark already unused pba %d as unused" , pba);
198 		msb->read_only = true;
199 		return;
200 	}
201 
202 	if (msb_validate_used_block_bitmap(msb))
203 		return;
204 
205 	/* No races because all IO is single threaded */
206 	__clear_bit(pba, msb->used_blocks_bitmap);
207 	msb->free_block_count[zone]++;
208 }
209 
210 /* Invalidate current register window */
211 static void msb_invalidate_reg_window(struct msb_data *msb)
212 {
213 	msb->reg_addr.w_offset = offsetof(struct ms_register, id);
214 	msb->reg_addr.w_length = sizeof(struct ms_id_register);
215 	msb->reg_addr.r_offset = offsetof(struct ms_register, id);
216 	msb->reg_addr.r_length = sizeof(struct ms_id_register);
217 	msb->addr_valid = false;
218 }
219 
220 /* Start a state machine */
221 static int msb_run_state_machine(struct msb_data *msb, int   (*state_func)
222 		(struct memstick_dev *card, struct memstick_request **req))
223 {
224 	struct memstick_dev *card = msb->card;
225 
226 	WARN_ON(msb->state != -1);
227 	msb->int_polling = false;
228 	msb->state = 0;
229 	msb->exit_error = 0;
230 
231 	memset(&card->current_mrq, 0, sizeof(card->current_mrq));
232 
233 	card->next_request = state_func;
234 	memstick_new_req(card->host);
235 	wait_for_completion(&card->mrq_complete);
236 
237 	WARN_ON(msb->state != -1);
238 	return msb->exit_error;
239 }
240 
241 /* State machines call that to exit */
242 static int msb_exit_state_machine(struct msb_data *msb, int error)
243 {
244 	WARN_ON(msb->state == -1);
245 
246 	msb->state = -1;
247 	msb->exit_error = error;
248 	msb->card->next_request = h_msb_default_bad;
249 
250 	/* Invalidate reg window on errors */
251 	if (error)
252 		msb_invalidate_reg_window(msb);
253 
254 	complete(&msb->card->mrq_complete);
255 	return -ENXIO;
256 }
257 
258 /* read INT register */
259 static int msb_read_int_reg(struct msb_data *msb, long timeout)
260 {
261 	struct memstick_request *mrq = &msb->card->current_mrq;
262 
263 	WARN_ON(msb->state == -1);
264 
265 	if (!msb->int_polling) {
266 		msb->int_timeout = jiffies +
267 			msecs_to_jiffies(timeout == -1 ? 500 : timeout);
268 		msb->int_polling = true;
269 	} else if (time_after(jiffies, msb->int_timeout)) {
270 		mrq->data[0] = MEMSTICK_INT_CMDNAK;
271 		return 0;
272 	}
273 
274 	if ((msb->caps & MEMSTICK_CAP_AUTO_GET_INT) &&
275 				mrq->need_card_int && !mrq->error) {
276 		mrq->data[0] = mrq->int_reg;
277 		mrq->need_card_int = false;
278 		return 0;
279 	} else {
280 		memstick_init_req(mrq, MS_TPC_GET_INT, NULL, 1);
281 		return 1;
282 	}
283 }
284 
285 /* Read a register */
286 static int msb_read_regs(struct msb_data *msb, int offset, int len)
287 {
288 	struct memstick_request *req = &msb->card->current_mrq;
289 
290 	if (msb->reg_addr.r_offset != offset ||
291 	    msb->reg_addr.r_length != len || !msb->addr_valid) {
292 
293 		msb->reg_addr.r_offset = offset;
294 		msb->reg_addr.r_length = len;
295 		msb->addr_valid = true;
296 
297 		memstick_init_req(req, MS_TPC_SET_RW_REG_ADRS,
298 			&msb->reg_addr, sizeof(msb->reg_addr));
299 		return 0;
300 	}
301 
302 	memstick_init_req(req, MS_TPC_READ_REG, NULL, len);
303 	return 1;
304 }
305 
306 /* Write a card register */
307 static int msb_write_regs(struct msb_data *msb, int offset, int len, void *buf)
308 {
309 	struct memstick_request *req = &msb->card->current_mrq;
310 
311 	if (msb->reg_addr.w_offset != offset ||
312 		msb->reg_addr.w_length != len  || !msb->addr_valid) {
313 
314 		msb->reg_addr.w_offset = offset;
315 		msb->reg_addr.w_length = len;
316 		msb->addr_valid = true;
317 
318 		memstick_init_req(req, MS_TPC_SET_RW_REG_ADRS,
319 			&msb->reg_addr, sizeof(msb->reg_addr));
320 		return 0;
321 	}
322 
323 	memstick_init_req(req, MS_TPC_WRITE_REG, buf, len);
324 	return 1;
325 }
326 
327 /* Handler for absence of IO */
328 static int h_msb_default_bad(struct memstick_dev *card,
329 						struct memstick_request **mrq)
330 {
331 	return -ENXIO;
332 }
333 
334 /*
335  * This function is a handler for reads of one page from device.
336  * Writes output to msb->current_sg, takes sector address from msb->reg.param
337  * Can also be used to read extra data only. Set params accordintly.
338  */
339 static int h_msb_read_page(struct memstick_dev *card,
340 					struct memstick_request **out_mrq)
341 {
342 	struct msb_data *msb = memstick_get_drvdata(card);
343 	struct memstick_request *mrq = *out_mrq = &card->current_mrq;
344 	struct scatterlist sg[2];
345 	u8 command, intreg;
346 
347 	if (mrq->error) {
348 		dbg("read_page, unknown error");
349 		return msb_exit_state_machine(msb, mrq->error);
350 	}
351 again:
352 	switch (msb->state) {
353 	case MSB_RP_SEND_BLOCK_ADDRESS:
354 		/* msb_write_regs sometimes "fails" because it needs to update
355 			the reg window, and thus it returns request for that.
356 			Then we stay in this state and retry */
357 		if (!msb_write_regs(msb,
358 			offsetof(struct ms_register, param),
359 			sizeof(struct ms_param_register),
360 			(unsigned char *)&msb->regs.param))
361 			return 0;
362 
363 		msb->state = MSB_RP_SEND_READ_COMMAND;
364 		return 0;
365 
366 	case MSB_RP_SEND_READ_COMMAND:
367 		command = MS_CMD_BLOCK_READ;
368 		memstick_init_req(mrq, MS_TPC_SET_CMD, &command, 1);
369 		msb->state = MSB_RP_SEND_INT_REQ;
370 		return 0;
371 
372 	case MSB_RP_SEND_INT_REQ:
373 		msb->state = MSB_RP_RECEIVE_INT_REQ_RESULT;
374 		/* If dont actually need to send the int read request (only in
375 			serial mode), then just fall through */
376 		if (msb_read_int_reg(msb, -1))
377 			return 0;
378 		/* fallthrough */
379 
380 	case MSB_RP_RECEIVE_INT_REQ_RESULT:
381 		intreg = mrq->data[0];
382 		msb->regs.status.interrupt = intreg;
383 
384 		if (intreg & MEMSTICK_INT_CMDNAK)
385 			return msb_exit_state_machine(msb, -EIO);
386 
387 		if (!(intreg & MEMSTICK_INT_CED)) {
388 			msb->state = MSB_RP_SEND_INT_REQ;
389 			goto again;
390 		}
391 
392 		msb->int_polling = false;
393 		msb->state = (intreg & MEMSTICK_INT_ERR) ?
394 			MSB_RP_SEND_READ_STATUS_REG : MSB_RP_SEND_OOB_READ;
395 		goto again;
396 
397 	case MSB_RP_SEND_READ_STATUS_REG:
398 		 /* read the status register to understand source of the INT_ERR */
399 		if (!msb_read_regs(msb,
400 			offsetof(struct ms_register, status),
401 			sizeof(struct ms_status_register)))
402 			return 0;
403 
404 		msb->state = MSB_RP_RECEIVE_STATUS_REG;
405 		return 0;
406 
407 	case MSB_RP_RECEIVE_STATUS_REG:
408 		msb->regs.status = *(struct ms_status_register *)mrq->data;
409 		msb->state = MSB_RP_SEND_OOB_READ;
410 		/* fallthrough */
411 
412 	case MSB_RP_SEND_OOB_READ:
413 		if (!msb_read_regs(msb,
414 			offsetof(struct ms_register, extra_data),
415 			sizeof(struct ms_extra_data_register)))
416 			return 0;
417 
418 		msb->state = MSB_RP_RECEIVE_OOB_READ;
419 		return 0;
420 
421 	case MSB_RP_RECEIVE_OOB_READ:
422 		msb->regs.extra_data =
423 			*(struct ms_extra_data_register *) mrq->data;
424 		msb->state = MSB_RP_SEND_READ_DATA;
425 		/* fallthrough */
426 
427 	case MSB_RP_SEND_READ_DATA:
428 		/* Skip that state if we only read the oob */
429 		if (msb->regs.param.cp == MEMSTICK_CP_EXTRA) {
430 			msb->state = MSB_RP_RECEIVE_READ_DATA;
431 			goto again;
432 		}
433 
434 		sg_init_table(sg, ARRAY_SIZE(sg));
435 		msb_sg_copy(msb->current_sg, sg, ARRAY_SIZE(sg),
436 			msb->current_sg_offset,
437 			msb->page_size);
438 
439 		memstick_init_req_sg(mrq, MS_TPC_READ_LONG_DATA, sg);
440 		msb->state = MSB_RP_RECEIVE_READ_DATA;
441 		return 0;
442 
443 	case MSB_RP_RECEIVE_READ_DATA:
444 		if (!(msb->regs.status.interrupt & MEMSTICK_INT_ERR)) {
445 			msb->current_sg_offset += msb->page_size;
446 			return msb_exit_state_machine(msb, 0);
447 		}
448 
449 		if (msb->regs.status.status1 & MEMSTICK_UNCORR_ERROR) {
450 			dbg("read_page: uncorrectable error");
451 			return msb_exit_state_machine(msb, -EBADMSG);
452 		}
453 
454 		if (msb->regs.status.status1 & MEMSTICK_CORR_ERROR) {
455 			dbg("read_page: correctable error");
456 			msb->current_sg_offset += msb->page_size;
457 			return msb_exit_state_machine(msb, -EUCLEAN);
458 		} else {
459 			dbg("read_page: INT error, but no status error bits");
460 			return msb_exit_state_machine(msb, -EIO);
461 		}
462 	}
463 
464 	BUG();
465 }
466 
467 /*
468  * Handler of writes of exactly one block.
469  * Takes address from msb->regs.param.
470  * Writes same extra data to blocks, also taken
471  * from msb->regs.extra
472  * Returns -EBADMSG if write fails due to uncorrectable error, or -EIO if
473  * device refuses to take the command or something else
474  */
475 static int h_msb_write_block(struct memstick_dev *card,
476 					struct memstick_request **out_mrq)
477 {
478 	struct msb_data *msb = memstick_get_drvdata(card);
479 	struct memstick_request *mrq = *out_mrq = &card->current_mrq;
480 	struct scatterlist sg[2];
481 	u8 intreg, command;
482 
483 	if (mrq->error)
484 		return msb_exit_state_machine(msb, mrq->error);
485 
486 again:
487 	switch (msb->state) {
488 
489 	/* HACK: Jmicon handling of TPCs between 8 and
490 	 *	sizeof(memstick_request.data) is broken due to hardware
491 	 *	bug in PIO mode that is used for these TPCs
492 	 *	Therefore split the write
493 	 */
494 
495 	case MSB_WB_SEND_WRITE_PARAMS:
496 		if (!msb_write_regs(msb,
497 			offsetof(struct ms_register, param),
498 			sizeof(struct ms_param_register),
499 			&msb->regs.param))
500 			return 0;
501 
502 		msb->state = MSB_WB_SEND_WRITE_OOB;
503 		return 0;
504 
505 	case MSB_WB_SEND_WRITE_OOB:
506 		if (!msb_write_regs(msb,
507 			offsetof(struct ms_register, extra_data),
508 			sizeof(struct ms_extra_data_register),
509 			&msb->regs.extra_data))
510 			return 0;
511 		msb->state = MSB_WB_SEND_WRITE_COMMAND;
512 		return 0;
513 
514 
515 	case MSB_WB_SEND_WRITE_COMMAND:
516 		command = MS_CMD_BLOCK_WRITE;
517 		memstick_init_req(mrq, MS_TPC_SET_CMD, &command, 1);
518 		msb->state = MSB_WB_SEND_INT_REQ;
519 		return 0;
520 
521 	case MSB_WB_SEND_INT_REQ:
522 		msb->state = MSB_WB_RECEIVE_INT_REQ;
523 		if (msb_read_int_reg(msb, -1))
524 			return 0;
525 		/* fallthrough */
526 
527 	case MSB_WB_RECEIVE_INT_REQ:
528 		intreg = mrq->data[0];
529 		msb->regs.status.interrupt = intreg;
530 
531 		/* errors mean out of here, and fast... */
532 		if (intreg & (MEMSTICK_INT_CMDNAK))
533 			return msb_exit_state_machine(msb, -EIO);
534 
535 		if (intreg & MEMSTICK_INT_ERR)
536 			return msb_exit_state_machine(msb, -EBADMSG);
537 
538 
539 		/* for last page we need to poll CED */
540 		if (msb->current_page == msb->pages_in_block) {
541 			if (intreg & MEMSTICK_INT_CED)
542 				return msb_exit_state_machine(msb, 0);
543 			msb->state = MSB_WB_SEND_INT_REQ;
544 			goto again;
545 
546 		}
547 
548 		/* for non-last page we need BREQ before writing next chunk */
549 		if (!(intreg & MEMSTICK_INT_BREQ)) {
550 			msb->state = MSB_WB_SEND_INT_REQ;
551 			goto again;
552 		}
553 
554 		msb->int_polling = false;
555 		msb->state = MSB_WB_SEND_WRITE_DATA;
556 		/* fallthrough */
557 
558 	case MSB_WB_SEND_WRITE_DATA:
559 		sg_init_table(sg, ARRAY_SIZE(sg));
560 
561 		if (msb_sg_copy(msb->current_sg, sg, ARRAY_SIZE(sg),
562 			msb->current_sg_offset,
563 			msb->page_size) < msb->page_size)
564 			return msb_exit_state_machine(msb, -EIO);
565 
566 		memstick_init_req_sg(mrq, MS_TPC_WRITE_LONG_DATA, sg);
567 		mrq->need_card_int = 1;
568 		msb->state = MSB_WB_RECEIVE_WRITE_CONFIRMATION;
569 		return 0;
570 
571 	case MSB_WB_RECEIVE_WRITE_CONFIRMATION:
572 		msb->current_page++;
573 		msb->current_sg_offset += msb->page_size;
574 		msb->state = MSB_WB_SEND_INT_REQ;
575 		goto again;
576 	default:
577 		BUG();
578 	}
579 
580 	return 0;
581 }
582 
583 /*
584  * This function is used to send simple IO requests to device that consist
585  * of register write + command
586  */
587 static int h_msb_send_command(struct memstick_dev *card,
588 					struct memstick_request **out_mrq)
589 {
590 	struct msb_data *msb = memstick_get_drvdata(card);
591 	struct memstick_request *mrq = *out_mrq = &card->current_mrq;
592 	u8 intreg;
593 
594 	if (mrq->error) {
595 		dbg("send_command: unknown error");
596 		return msb_exit_state_machine(msb, mrq->error);
597 	}
598 again:
599 	switch (msb->state) {
600 
601 	/* HACK: see h_msb_write_block */
602 	case MSB_SC_SEND_WRITE_PARAMS: /* write param register*/
603 		if (!msb_write_regs(msb,
604 			offsetof(struct ms_register, param),
605 			sizeof(struct ms_param_register),
606 			&msb->regs.param))
607 			return 0;
608 		msb->state = MSB_SC_SEND_WRITE_OOB;
609 		return 0;
610 
611 	case MSB_SC_SEND_WRITE_OOB:
612 		if (!msb->command_need_oob) {
613 			msb->state = MSB_SC_SEND_COMMAND;
614 			goto again;
615 		}
616 
617 		if (!msb_write_regs(msb,
618 			offsetof(struct ms_register, extra_data),
619 			sizeof(struct ms_extra_data_register),
620 			&msb->regs.extra_data))
621 			return 0;
622 
623 		msb->state = MSB_SC_SEND_COMMAND;
624 		return 0;
625 
626 	case MSB_SC_SEND_COMMAND:
627 		memstick_init_req(mrq, MS_TPC_SET_CMD, &msb->command_value, 1);
628 		msb->state = MSB_SC_SEND_INT_REQ;
629 		return 0;
630 
631 	case MSB_SC_SEND_INT_REQ:
632 		msb->state = MSB_SC_RECEIVE_INT_REQ;
633 		if (msb_read_int_reg(msb, -1))
634 			return 0;
635 		/* fallthrough */
636 
637 	case MSB_SC_RECEIVE_INT_REQ:
638 		intreg = mrq->data[0];
639 
640 		if (intreg & MEMSTICK_INT_CMDNAK)
641 			return msb_exit_state_machine(msb, -EIO);
642 		if (intreg & MEMSTICK_INT_ERR)
643 			return msb_exit_state_machine(msb, -EBADMSG);
644 
645 		if (!(intreg & MEMSTICK_INT_CED)) {
646 			msb->state = MSB_SC_SEND_INT_REQ;
647 			goto again;
648 		}
649 
650 		return msb_exit_state_machine(msb, 0);
651 	}
652 
653 	BUG();
654 }
655 
656 /* Small handler for card reset */
657 static int h_msb_reset(struct memstick_dev *card,
658 					struct memstick_request **out_mrq)
659 {
660 	u8 command = MS_CMD_RESET;
661 	struct msb_data *msb = memstick_get_drvdata(card);
662 	struct memstick_request *mrq = *out_mrq = &card->current_mrq;
663 
664 	if (mrq->error)
665 		return msb_exit_state_machine(msb, mrq->error);
666 
667 	switch (msb->state) {
668 	case MSB_RS_SEND:
669 		memstick_init_req(mrq, MS_TPC_SET_CMD, &command, 1);
670 		mrq->need_card_int = 0;
671 		msb->state = MSB_RS_CONFIRM;
672 		return 0;
673 	case MSB_RS_CONFIRM:
674 		return msb_exit_state_machine(msb, 0);
675 	}
676 	BUG();
677 }
678 
679 /* This handler is used to do serial->parallel switch */
680 static int h_msb_parallel_switch(struct memstick_dev *card,
681 					struct memstick_request **out_mrq)
682 {
683 	struct msb_data *msb = memstick_get_drvdata(card);
684 	struct memstick_request *mrq = *out_mrq = &card->current_mrq;
685 	struct memstick_host *host = card->host;
686 
687 	if (mrq->error) {
688 		dbg("parallel_switch: error");
689 		msb->regs.param.system &= ~MEMSTICK_SYS_PAM;
690 		return msb_exit_state_machine(msb, mrq->error);
691 	}
692 
693 	switch (msb->state) {
694 	case MSB_PS_SEND_SWITCH_COMMAND:
695 		/* Set the parallel interface on memstick side */
696 		msb->regs.param.system |= MEMSTICK_SYS_PAM;
697 
698 		if (!msb_write_regs(msb,
699 			offsetof(struct ms_register, param),
700 			1,
701 			(unsigned char *)&msb->regs.param))
702 			return 0;
703 
704 		msb->state = MSB_PS_SWICH_HOST;
705 		return 0;
706 
707 	case MSB_PS_SWICH_HOST:
708 		 /* Set parallel interface on our side + send a dummy request
709 			to see if card responds */
710 		host->set_param(host, MEMSTICK_INTERFACE, MEMSTICK_PAR4);
711 		memstick_init_req(mrq, MS_TPC_GET_INT, NULL, 1);
712 		msb->state = MSB_PS_CONFIRM;
713 		return 0;
714 
715 	case MSB_PS_CONFIRM:
716 		return msb_exit_state_machine(msb, 0);
717 	}
718 
719 	BUG();
720 }
721 
722 static int msb_switch_to_parallel(struct msb_data *msb);
723 
724 /* Reset the card, to guard against hw errors beeing treated as bad blocks */
725 static int msb_reset(struct msb_data *msb, bool full)
726 {
727 
728 	bool was_parallel = msb->regs.param.system & MEMSTICK_SYS_PAM;
729 	struct memstick_dev *card = msb->card;
730 	struct memstick_host *host = card->host;
731 	int error;
732 
733 	/* Reset the card */
734 	msb->regs.param.system = MEMSTICK_SYS_BAMD;
735 
736 	if (full) {
737 		error =  host->set_param(host,
738 					MEMSTICK_POWER, MEMSTICK_POWER_OFF);
739 		if (error)
740 			goto out_error;
741 
742 		msb_invalidate_reg_window(msb);
743 
744 		error = host->set_param(host,
745 					MEMSTICK_POWER, MEMSTICK_POWER_ON);
746 		if (error)
747 			goto out_error;
748 
749 		error = host->set_param(host,
750 					MEMSTICK_INTERFACE, MEMSTICK_SERIAL);
751 		if (error) {
752 out_error:
753 			dbg("Failed to reset the host controller");
754 			msb->read_only = true;
755 			return -EFAULT;
756 		}
757 	}
758 
759 	error = msb_run_state_machine(msb, h_msb_reset);
760 	if (error) {
761 		dbg("Failed to reset the card");
762 		msb->read_only = true;
763 		return -ENODEV;
764 	}
765 
766 	/* Set parallel mode */
767 	if (was_parallel)
768 		msb_switch_to_parallel(msb);
769 	return 0;
770 }
771 
772 /* Attempts to switch interface to parallel mode */
773 static int msb_switch_to_parallel(struct msb_data *msb)
774 {
775 	int error;
776 
777 	error = msb_run_state_machine(msb, h_msb_parallel_switch);
778 	if (error) {
779 		pr_err("Switch to parallel failed");
780 		msb->regs.param.system &= ~MEMSTICK_SYS_PAM;
781 		msb_reset(msb, true);
782 		return -EFAULT;
783 	}
784 
785 	msb->caps |= MEMSTICK_CAP_AUTO_GET_INT;
786 	return 0;
787 }
788 
789 /* Changes overwrite flag on a page */
790 static int msb_set_overwrite_flag(struct msb_data *msb,
791 						u16 pba, u8 page, u8 flag)
792 {
793 	if (msb->read_only)
794 		return -EROFS;
795 
796 	msb->regs.param.block_address = cpu_to_be16(pba);
797 	msb->regs.param.page_address = page;
798 	msb->regs.param.cp = MEMSTICK_CP_OVERWRITE;
799 	msb->regs.extra_data.overwrite_flag = flag;
800 	msb->command_value = MS_CMD_BLOCK_WRITE;
801 	msb->command_need_oob = true;
802 
803 	dbg_verbose("changing overwrite flag to %02x for sector %d, page %d",
804 							flag, pba, page);
805 	return msb_run_state_machine(msb, h_msb_send_command);
806 }
807 
808 static int msb_mark_bad(struct msb_data *msb, int pba)
809 {
810 	pr_notice("marking pba %d as bad", pba);
811 	msb_reset(msb, true);
812 	return msb_set_overwrite_flag(
813 			msb, pba, 0, 0xFF & ~MEMSTICK_OVERWRITE_BKST);
814 }
815 
816 static int msb_mark_page_bad(struct msb_data *msb, int pba, int page)
817 {
818 	dbg("marking page %d of pba %d as bad", page, pba);
819 	msb_reset(msb, true);
820 	return msb_set_overwrite_flag(msb,
821 		pba, page, ~MEMSTICK_OVERWRITE_PGST0);
822 }
823 
824 /* Erases one physical block */
825 static int msb_erase_block(struct msb_data *msb, u16 pba)
826 {
827 	int error, try;
828 	if (msb->read_only)
829 		return -EROFS;
830 
831 	dbg_verbose("erasing pba %d", pba);
832 
833 	for (try = 1; try < 3; try++) {
834 		msb->regs.param.block_address = cpu_to_be16(pba);
835 		msb->regs.param.page_address = 0;
836 		msb->regs.param.cp = MEMSTICK_CP_BLOCK;
837 		msb->command_value = MS_CMD_BLOCK_ERASE;
838 		msb->command_need_oob = false;
839 
840 
841 		error = msb_run_state_machine(msb, h_msb_send_command);
842 		if (!error || msb_reset(msb, true))
843 			break;
844 	}
845 
846 	if (error) {
847 		pr_err("erase failed, marking pba %d as bad", pba);
848 		msb_mark_bad(msb, pba);
849 	}
850 
851 	dbg_verbose("erase success, marking pba %d as unused", pba);
852 	msb_mark_block_unused(msb, pba);
853 	__set_bit(pba, msb->erased_blocks_bitmap);
854 	return error;
855 }
856 
857 /* Reads one page from device */
858 static int msb_read_page(struct msb_data *msb,
859 	u16 pba, u8 page, struct ms_extra_data_register *extra,
860 					struct scatterlist *sg,  int offset)
861 {
862 	int try, error;
863 
864 	if (pba == MS_BLOCK_INVALID) {
865 		unsigned long flags;
866 		struct sg_mapping_iter miter;
867 		size_t len = msb->page_size;
868 
869 		dbg_verbose("read unmapped sector. returning 0xFF");
870 
871 		local_irq_save(flags);
872 		sg_miter_start(&miter, sg, sg_nents(sg),
873 				SG_MITER_ATOMIC | SG_MITER_TO_SG);
874 
875 		while (sg_miter_next(&miter) && len > 0) {
876 
877 			int chunklen;
878 
879 			if (offset && offset >= miter.length) {
880 				offset -= miter.length;
881 				continue;
882 			}
883 
884 			chunklen = min(miter.length - offset, len);
885 			memset(miter.addr + offset, 0xFF, chunklen);
886 			len -= chunklen;
887 			offset = 0;
888 		}
889 
890 		sg_miter_stop(&miter);
891 		local_irq_restore(flags);
892 
893 		if (offset)
894 			return -EFAULT;
895 
896 		if (extra)
897 			memset(extra, 0xFF, sizeof(*extra));
898 		return 0;
899 	}
900 
901 	if (pba >= msb->block_count) {
902 		pr_err("BUG: attempt to read beyond the end of the card at pba %d", pba);
903 		return -EINVAL;
904 	}
905 
906 	for (try = 1; try < 3; try++) {
907 		msb->regs.param.block_address = cpu_to_be16(pba);
908 		msb->regs.param.page_address = page;
909 		msb->regs.param.cp = MEMSTICK_CP_PAGE;
910 
911 		msb->current_sg = sg;
912 		msb->current_sg_offset = offset;
913 		error = msb_run_state_machine(msb, h_msb_read_page);
914 
915 
916 		if (error == -EUCLEAN) {
917 			pr_notice("correctable error on pba %d, page %d",
918 				pba, page);
919 			error = 0;
920 		}
921 
922 		if (!error && extra)
923 			*extra = msb->regs.extra_data;
924 
925 		if (!error || msb_reset(msb, true))
926 			break;
927 
928 	}
929 
930 	/* Mark bad pages */
931 	if (error == -EBADMSG) {
932 		pr_err("uncorrectable error on read of pba %d, page %d",
933 			pba, page);
934 
935 		if (msb->regs.extra_data.overwrite_flag &
936 					MEMSTICK_OVERWRITE_PGST0)
937 			msb_mark_page_bad(msb, pba, page);
938 		return -EBADMSG;
939 	}
940 
941 	if (error)
942 		pr_err("read of pba %d, page %d failed with error %d",
943 			pba, page, error);
944 	return error;
945 }
946 
947 /* Reads oob of page only */
948 static int msb_read_oob(struct msb_data *msb, u16 pba, u16 page,
949 	struct ms_extra_data_register *extra)
950 {
951 	int error;
952 
953 	BUG_ON(!extra);
954 	msb->regs.param.block_address = cpu_to_be16(pba);
955 	msb->regs.param.page_address = page;
956 	msb->regs.param.cp = MEMSTICK_CP_EXTRA;
957 
958 	if (pba > msb->block_count) {
959 		pr_err("BUG: attempt to read beyond the end of card at pba %d", pba);
960 		return -EINVAL;
961 	}
962 
963 	error = msb_run_state_machine(msb, h_msb_read_page);
964 	*extra = msb->regs.extra_data;
965 
966 	if (error == -EUCLEAN) {
967 		pr_notice("correctable error on pba %d, page %d",
968 			pba, page);
969 		return 0;
970 	}
971 
972 	return error;
973 }
974 
975 /* Reads a block and compares it with data contained in scatterlist orig_sg */
976 static int msb_verify_block(struct msb_data *msb, u16 pba,
977 				struct scatterlist *orig_sg,  int offset)
978 {
979 	struct scatterlist sg;
980 	int page = 0, error;
981 
982 	sg_init_one(&sg, msb->block_buffer, msb->block_size);
983 
984 	while (page < msb->pages_in_block) {
985 
986 		error = msb_read_page(msb, pba, page,
987 				NULL, &sg, page * msb->page_size);
988 		if (error)
989 			return error;
990 		page++;
991 	}
992 
993 	if (msb_sg_compare_to_buffer(orig_sg, offset,
994 				msb->block_buffer, msb->block_size))
995 		return -EIO;
996 	return 0;
997 }
998 
999 /* Writes exectly one block + oob */
1000 static int msb_write_block(struct msb_data *msb,
1001 			u16 pba, u32 lba, struct scatterlist *sg, int offset)
1002 {
1003 	int error, current_try = 1;
1004 	BUG_ON(sg->length < msb->page_size);
1005 
1006 	if (msb->read_only)
1007 		return -EROFS;
1008 
1009 	if (pba == MS_BLOCK_INVALID) {
1010 		pr_err(
1011 			"BUG: write: attempt to write MS_BLOCK_INVALID block");
1012 		return -EINVAL;
1013 	}
1014 
1015 	if (pba >= msb->block_count || lba >= msb->logical_block_count) {
1016 		pr_err(
1017 		"BUG: write: attempt to write beyond the end of device");
1018 		return -EINVAL;
1019 	}
1020 
1021 	if (msb_get_zone_from_lba(lba) != msb_get_zone_from_pba(pba)) {
1022 		pr_err("BUG: write: lba zone mismatch");
1023 		return -EINVAL;
1024 	}
1025 
1026 	if (pba == msb->boot_block_locations[0] ||
1027 		pba == msb->boot_block_locations[1]) {
1028 		pr_err("BUG: write: attempt to write to boot blocks!");
1029 		return -EINVAL;
1030 	}
1031 
1032 	while (1) {
1033 
1034 		if (msb->read_only)
1035 			return -EROFS;
1036 
1037 		msb->regs.param.cp = MEMSTICK_CP_BLOCK;
1038 		msb->regs.param.page_address = 0;
1039 		msb->regs.param.block_address = cpu_to_be16(pba);
1040 
1041 		msb->regs.extra_data.management_flag = 0xFF;
1042 		msb->regs.extra_data.overwrite_flag = 0xF8;
1043 		msb->regs.extra_data.logical_address = cpu_to_be16(lba);
1044 
1045 		msb->current_sg = sg;
1046 		msb->current_sg_offset = offset;
1047 		msb->current_page = 0;
1048 
1049 		error = msb_run_state_machine(msb, h_msb_write_block);
1050 
1051 		/* Sector we just wrote to is assumed erased since its pba
1052 			was erased. If it wasn't erased, write will succeed
1053 			and will just clear the bits that were set in the block
1054 			thus test that what we have written,
1055 			matches what we expect.
1056 			We do trust the blocks that we erased */
1057 		if (!error && (verify_writes ||
1058 				!test_bit(pba, msb->erased_blocks_bitmap)))
1059 			error = msb_verify_block(msb, pba, sg, offset);
1060 
1061 		if (!error)
1062 			break;
1063 
1064 		if (current_try > 1 || msb_reset(msb, true))
1065 			break;
1066 
1067 		pr_err("write failed, trying to erase the pba %d", pba);
1068 		error = msb_erase_block(msb, pba);
1069 		if (error)
1070 			break;
1071 
1072 		current_try++;
1073 	}
1074 	return error;
1075 }
1076 
1077 /* Finds a free block for write replacement */
1078 static u16 msb_get_free_block(struct msb_data *msb, int zone)
1079 {
1080 	u16 pos;
1081 	int pba = zone * MS_BLOCKS_IN_ZONE;
1082 	int i;
1083 
1084 	get_random_bytes(&pos, sizeof(pos));
1085 
1086 	if (!msb->free_block_count[zone]) {
1087 		pr_err("NO free blocks in the zone %d, to use for a write, (media is WORN out) switching to RO mode", zone);
1088 		msb->read_only = true;
1089 		return MS_BLOCK_INVALID;
1090 	}
1091 
1092 	pos %= msb->free_block_count[zone];
1093 
1094 	dbg_verbose("have %d choices for a free block, selected randomally: %d",
1095 		msb->free_block_count[zone], pos);
1096 
1097 	pba = find_next_zero_bit(msb->used_blocks_bitmap,
1098 							msb->block_count, pba);
1099 	for (i = 0; i < pos; ++i)
1100 		pba = find_next_zero_bit(msb->used_blocks_bitmap,
1101 						msb->block_count, pba + 1);
1102 
1103 	dbg_verbose("result of the free blocks scan: pba %d", pba);
1104 
1105 	if (pba == msb->block_count || (msb_get_zone_from_pba(pba)) != zone) {
1106 		pr_err("BUG: cant get a free block");
1107 		msb->read_only = true;
1108 		return MS_BLOCK_INVALID;
1109 	}
1110 
1111 	msb_mark_block_used(msb, pba);
1112 	return pba;
1113 }
1114 
1115 static int msb_update_block(struct msb_data *msb, u16 lba,
1116 	struct scatterlist *sg, int offset)
1117 {
1118 	u16 pba, new_pba;
1119 	int error, try;
1120 
1121 	pba = msb->lba_to_pba_table[lba];
1122 	dbg_verbose("start of a block update at lba  %d, pba %d", lba, pba);
1123 
1124 	if (pba != MS_BLOCK_INVALID) {
1125 		dbg_verbose("setting the update flag on the block");
1126 		msb_set_overwrite_flag(msb, pba, 0,
1127 				0xFF & ~MEMSTICK_OVERWRITE_UDST);
1128 	}
1129 
1130 	for (try = 0; try < 3; try++) {
1131 		new_pba = msb_get_free_block(msb,
1132 			msb_get_zone_from_lba(lba));
1133 
1134 		if (new_pba == MS_BLOCK_INVALID) {
1135 			error = -EIO;
1136 			goto out;
1137 		}
1138 
1139 		dbg_verbose("block update: writing updated block to the pba %d",
1140 								new_pba);
1141 		error = msb_write_block(msb, new_pba, lba, sg, offset);
1142 		if (error == -EBADMSG) {
1143 			msb_mark_bad(msb, new_pba);
1144 			continue;
1145 		}
1146 
1147 		if (error)
1148 			goto out;
1149 
1150 		dbg_verbose("block update: erasing the old block");
1151 		msb_erase_block(msb, pba);
1152 		msb->lba_to_pba_table[lba] = new_pba;
1153 		return 0;
1154 	}
1155 out:
1156 	if (error) {
1157 		pr_err("block update error after %d tries,  switching to r/o mode", try);
1158 		msb->read_only = true;
1159 	}
1160 	return error;
1161 }
1162 
1163 /* Converts endiannes in the boot block for easy use */
1164 static void msb_fix_boot_page_endianness(struct ms_boot_page *p)
1165 {
1166 	p->header.block_id = be16_to_cpu(p->header.block_id);
1167 	p->header.format_reserved = be16_to_cpu(p->header.format_reserved);
1168 	p->entry.disabled_block.start_addr
1169 		= be32_to_cpu(p->entry.disabled_block.start_addr);
1170 	p->entry.disabled_block.data_size
1171 		= be32_to_cpu(p->entry.disabled_block.data_size);
1172 	p->entry.cis_idi.start_addr
1173 		= be32_to_cpu(p->entry.cis_idi.start_addr);
1174 	p->entry.cis_idi.data_size
1175 		= be32_to_cpu(p->entry.cis_idi.data_size);
1176 	p->attr.block_size = be16_to_cpu(p->attr.block_size);
1177 	p->attr.number_of_blocks = be16_to_cpu(p->attr.number_of_blocks);
1178 	p->attr.number_of_effective_blocks
1179 		= be16_to_cpu(p->attr.number_of_effective_blocks);
1180 	p->attr.page_size = be16_to_cpu(p->attr.page_size);
1181 	p->attr.memory_manufacturer_code
1182 		= be16_to_cpu(p->attr.memory_manufacturer_code);
1183 	p->attr.memory_device_code = be16_to_cpu(p->attr.memory_device_code);
1184 	p->attr.implemented_capacity
1185 		= be16_to_cpu(p->attr.implemented_capacity);
1186 	p->attr.controller_number = be16_to_cpu(p->attr.controller_number);
1187 	p->attr.controller_function = be16_to_cpu(p->attr.controller_function);
1188 }
1189 
1190 static int msb_read_boot_blocks(struct msb_data *msb)
1191 {
1192 	int pba = 0;
1193 	struct scatterlist sg;
1194 	struct ms_extra_data_register extra;
1195 	struct ms_boot_page *page;
1196 
1197 	msb->boot_block_locations[0] = MS_BLOCK_INVALID;
1198 	msb->boot_block_locations[1] = MS_BLOCK_INVALID;
1199 	msb->boot_block_count = 0;
1200 
1201 	dbg_verbose("Start of a scan for the boot blocks");
1202 
1203 	if (!msb->boot_page) {
1204 		page = kmalloc_array(2, sizeof(struct ms_boot_page),
1205 				     GFP_KERNEL);
1206 		if (!page)
1207 			return -ENOMEM;
1208 
1209 		msb->boot_page = page;
1210 	} else
1211 		page = msb->boot_page;
1212 
1213 	msb->block_count = MS_BLOCK_MAX_BOOT_ADDR;
1214 
1215 	for (pba = 0; pba < MS_BLOCK_MAX_BOOT_ADDR; pba++) {
1216 
1217 		sg_init_one(&sg, page, sizeof(*page));
1218 		if (msb_read_page(msb, pba, 0, &extra, &sg, 0)) {
1219 			dbg("boot scan: can't read pba %d", pba);
1220 			continue;
1221 		}
1222 
1223 		if (extra.management_flag & MEMSTICK_MANAGEMENT_SYSFLG) {
1224 			dbg("management flag doesn't indicate boot block %d",
1225 									pba);
1226 			continue;
1227 		}
1228 
1229 		if (be16_to_cpu(page->header.block_id) != MS_BLOCK_BOOT_ID) {
1230 			dbg("the pba at %d doesn' contain boot block ID", pba);
1231 			continue;
1232 		}
1233 
1234 		msb_fix_boot_page_endianness(page);
1235 		msb->boot_block_locations[msb->boot_block_count] = pba;
1236 
1237 		page++;
1238 		msb->boot_block_count++;
1239 
1240 		if (msb->boot_block_count == 2)
1241 			break;
1242 	}
1243 
1244 	if (!msb->boot_block_count) {
1245 		pr_err("media doesn't contain master page, aborting");
1246 		return -EIO;
1247 	}
1248 
1249 	dbg_verbose("End of scan for boot blocks");
1250 	return 0;
1251 }
1252 
1253 static int msb_read_bad_block_table(struct msb_data *msb, int block_nr)
1254 {
1255 	struct ms_boot_page *boot_block;
1256 	struct scatterlist sg;
1257 	u16 *buffer = NULL;
1258 	int offset = 0;
1259 	int i, error = 0;
1260 	int data_size, data_offset, page, page_offset, size_to_read;
1261 	u16 pba;
1262 
1263 	BUG_ON(block_nr > 1);
1264 	boot_block = &msb->boot_page[block_nr];
1265 	pba = msb->boot_block_locations[block_nr];
1266 
1267 	if (msb->boot_block_locations[block_nr] == MS_BLOCK_INVALID)
1268 		return -EINVAL;
1269 
1270 	data_size = boot_block->entry.disabled_block.data_size;
1271 	data_offset = sizeof(struct ms_boot_page) +
1272 			boot_block->entry.disabled_block.start_addr;
1273 	if (!data_size)
1274 		return 0;
1275 
1276 	page = data_offset / msb->page_size;
1277 	page_offset = data_offset % msb->page_size;
1278 	size_to_read =
1279 		DIV_ROUND_UP(data_size + page_offset, msb->page_size) *
1280 			msb->page_size;
1281 
1282 	dbg("reading bad block of boot block at pba %d, offset %d len %d",
1283 		pba, data_offset, data_size);
1284 
1285 	buffer = kzalloc(size_to_read, GFP_KERNEL);
1286 	if (!buffer)
1287 		return -ENOMEM;
1288 
1289 	/* Read the buffer */
1290 	sg_init_one(&sg, buffer, size_to_read);
1291 
1292 	while (offset < size_to_read) {
1293 		error = msb_read_page(msb, pba, page, NULL, &sg, offset);
1294 		if (error)
1295 			goto out;
1296 
1297 		page++;
1298 		offset += msb->page_size;
1299 
1300 		if (page == msb->pages_in_block) {
1301 			pr_err(
1302 			"bad block table extends beyond the boot block");
1303 			break;
1304 		}
1305 	}
1306 
1307 	/* Process the bad block table */
1308 	for (i = page_offset; i < data_size / sizeof(u16); i++) {
1309 
1310 		u16 bad_block = be16_to_cpu(buffer[i]);
1311 
1312 		if (bad_block >= msb->block_count) {
1313 			dbg("bad block table contains invalid block %d",
1314 								bad_block);
1315 			continue;
1316 		}
1317 
1318 		if (test_bit(bad_block, msb->used_blocks_bitmap))  {
1319 			dbg("duplicate bad block %d in the table",
1320 				bad_block);
1321 			continue;
1322 		}
1323 
1324 		dbg("block %d is marked as factory bad", bad_block);
1325 		msb_mark_block_used(msb, bad_block);
1326 	}
1327 out:
1328 	kfree(buffer);
1329 	return error;
1330 }
1331 
1332 static int msb_ftl_initialize(struct msb_data *msb)
1333 {
1334 	int i;
1335 
1336 	if (msb->ftl_initialized)
1337 		return 0;
1338 
1339 	msb->zone_count = msb->block_count / MS_BLOCKS_IN_ZONE;
1340 	msb->logical_block_count = msb->zone_count * 496 - 2;
1341 
1342 	msb->used_blocks_bitmap = kzalloc(msb->block_count / 8, GFP_KERNEL);
1343 	msb->erased_blocks_bitmap = kzalloc(msb->block_count / 8, GFP_KERNEL);
1344 	msb->lba_to_pba_table =
1345 		kmalloc_array(msb->logical_block_count, sizeof(u16),
1346 			      GFP_KERNEL);
1347 
1348 	if (!msb->used_blocks_bitmap || !msb->lba_to_pba_table ||
1349 						!msb->erased_blocks_bitmap) {
1350 		kfree(msb->used_blocks_bitmap);
1351 		kfree(msb->lba_to_pba_table);
1352 		kfree(msb->erased_blocks_bitmap);
1353 		return -ENOMEM;
1354 	}
1355 
1356 	for (i = 0; i < msb->zone_count; i++)
1357 		msb->free_block_count[i] = MS_BLOCKS_IN_ZONE;
1358 
1359 	memset(msb->lba_to_pba_table, MS_BLOCK_INVALID,
1360 			msb->logical_block_count * sizeof(u16));
1361 
1362 	dbg("initial FTL tables created. Zone count = %d, Logical block count = %d",
1363 		msb->zone_count, msb->logical_block_count);
1364 
1365 	msb->ftl_initialized = true;
1366 	return 0;
1367 }
1368 
1369 static int msb_ftl_scan(struct msb_data *msb)
1370 {
1371 	u16 pba, lba, other_block;
1372 	u8 overwrite_flag, management_flag, other_overwrite_flag;
1373 	int error;
1374 	struct ms_extra_data_register extra;
1375 	u8 *overwrite_flags = kzalloc(msb->block_count, GFP_KERNEL);
1376 
1377 	if (!overwrite_flags)
1378 		return -ENOMEM;
1379 
1380 	dbg("Start of media scanning");
1381 	for (pba = 0; pba < msb->block_count; pba++) {
1382 
1383 		if (pba == msb->boot_block_locations[0] ||
1384 			pba == msb->boot_block_locations[1]) {
1385 			dbg_verbose("pba %05d -> [boot block]", pba);
1386 			msb_mark_block_used(msb, pba);
1387 			continue;
1388 		}
1389 
1390 		if (test_bit(pba, msb->used_blocks_bitmap)) {
1391 			dbg_verbose("pba %05d -> [factory bad]", pba);
1392 			continue;
1393 		}
1394 
1395 		memset(&extra, 0, sizeof(extra));
1396 		error = msb_read_oob(msb, pba, 0, &extra);
1397 
1398 		/* can't trust the page if we can't read the oob */
1399 		if (error == -EBADMSG) {
1400 			pr_notice(
1401 			"oob of pba %d damaged, will try to erase it", pba);
1402 			msb_mark_block_used(msb, pba);
1403 			msb_erase_block(msb, pba);
1404 			continue;
1405 		} else if (error) {
1406 			pr_err("unknown error %d on read of oob of pba %d - aborting",
1407 				error, pba);
1408 
1409 			kfree(overwrite_flags);
1410 			return error;
1411 		}
1412 
1413 		lba = be16_to_cpu(extra.logical_address);
1414 		management_flag = extra.management_flag;
1415 		overwrite_flag = extra.overwrite_flag;
1416 		overwrite_flags[pba] = overwrite_flag;
1417 
1418 		/* Skip bad blocks */
1419 		if (!(overwrite_flag & MEMSTICK_OVERWRITE_BKST)) {
1420 			dbg("pba %05d -> [BAD]", pba);
1421 			msb_mark_block_used(msb, pba);
1422 			continue;
1423 		}
1424 
1425 		/* Skip system/drm blocks */
1426 		if ((management_flag & MEMSTICK_MANAGEMENT_FLAG_NORMAL) !=
1427 			MEMSTICK_MANAGEMENT_FLAG_NORMAL) {
1428 			dbg("pba %05d -> [reserved management flag %02x]",
1429 							pba, management_flag);
1430 			msb_mark_block_used(msb, pba);
1431 			continue;
1432 		}
1433 
1434 		/* Erase temporary tables */
1435 		if (!(management_flag & MEMSTICK_MANAGEMENT_ATFLG)) {
1436 			dbg("pba %05d -> [temp table] - will erase", pba);
1437 
1438 			msb_mark_block_used(msb, pba);
1439 			msb_erase_block(msb, pba);
1440 			continue;
1441 		}
1442 
1443 		if (lba == MS_BLOCK_INVALID) {
1444 			dbg_verbose("pba %05d -> [free]", pba);
1445 			continue;
1446 		}
1447 
1448 		msb_mark_block_used(msb, pba);
1449 
1450 		/* Block has LBA not according to zoning*/
1451 		if (msb_get_zone_from_lba(lba) != msb_get_zone_from_pba(pba)) {
1452 			pr_notice("pba %05d -> [bad lba %05d] - will erase",
1453 								pba, lba);
1454 			msb_erase_block(msb, pba);
1455 			continue;
1456 		}
1457 
1458 		/* No collisions - great */
1459 		if (msb->lba_to_pba_table[lba] == MS_BLOCK_INVALID) {
1460 			dbg_verbose("pba %05d -> [lba %05d]", pba, lba);
1461 			msb->lba_to_pba_table[lba] = pba;
1462 			continue;
1463 		}
1464 
1465 		other_block = msb->lba_to_pba_table[lba];
1466 		other_overwrite_flag = overwrite_flags[other_block];
1467 
1468 		pr_notice("Collision between pba %d and pba %d",
1469 			pba, other_block);
1470 
1471 		if (!(overwrite_flag & MEMSTICK_OVERWRITE_UDST)) {
1472 			pr_notice("pba %d is marked as stable, use it", pba);
1473 			msb_erase_block(msb, other_block);
1474 			msb->lba_to_pba_table[lba] = pba;
1475 			continue;
1476 		}
1477 
1478 		if (!(other_overwrite_flag & MEMSTICK_OVERWRITE_UDST)) {
1479 			pr_notice("pba %d is marked as stable, use it",
1480 								other_block);
1481 			msb_erase_block(msb, pba);
1482 			continue;
1483 		}
1484 
1485 		pr_notice("collision between blocks %d and %d, without stable flag set on both, erasing pba %d",
1486 				pba, other_block, other_block);
1487 
1488 		msb_erase_block(msb, other_block);
1489 		msb->lba_to_pba_table[lba] = pba;
1490 	}
1491 
1492 	dbg("End of media scanning");
1493 	kfree(overwrite_flags);
1494 	return 0;
1495 }
1496 
1497 static void msb_cache_flush_timer(struct timer_list *t)
1498 {
1499 	struct msb_data *msb = from_timer(msb, t, cache_flush_timer);
1500 	msb->need_flush_cache = true;
1501 	queue_work(msb->io_queue, &msb->io_work);
1502 }
1503 
1504 
1505 static void msb_cache_discard(struct msb_data *msb)
1506 {
1507 	if (msb->cache_block_lba == MS_BLOCK_INVALID)
1508 		return;
1509 
1510 	del_timer_sync(&msb->cache_flush_timer);
1511 
1512 	dbg_verbose("Discarding the write cache");
1513 	msb->cache_block_lba = MS_BLOCK_INVALID;
1514 	bitmap_zero(&msb->valid_cache_bitmap, msb->pages_in_block);
1515 }
1516 
1517 static int msb_cache_init(struct msb_data *msb)
1518 {
1519 	timer_setup(&msb->cache_flush_timer, msb_cache_flush_timer, 0);
1520 
1521 	if (!msb->cache)
1522 		msb->cache = kzalloc(msb->block_size, GFP_KERNEL);
1523 	if (!msb->cache)
1524 		return -ENOMEM;
1525 
1526 	msb_cache_discard(msb);
1527 	return 0;
1528 }
1529 
1530 static int msb_cache_flush(struct msb_data *msb)
1531 {
1532 	struct scatterlist sg;
1533 	struct ms_extra_data_register extra;
1534 	int page, offset, error;
1535 	u16 pba, lba;
1536 
1537 	if (msb->read_only)
1538 		return -EROFS;
1539 
1540 	if (msb->cache_block_lba == MS_BLOCK_INVALID)
1541 		return 0;
1542 
1543 	lba = msb->cache_block_lba;
1544 	pba = msb->lba_to_pba_table[lba];
1545 
1546 	dbg_verbose("Flushing the write cache of pba %d (LBA %d)",
1547 						pba, msb->cache_block_lba);
1548 
1549 	sg_init_one(&sg, msb->cache , msb->block_size);
1550 
1551 	/* Read all missing pages in cache */
1552 	for (page = 0; page < msb->pages_in_block; page++) {
1553 
1554 		if (test_bit(page, &msb->valid_cache_bitmap))
1555 			continue;
1556 
1557 		offset = page * msb->page_size;
1558 
1559 		dbg_verbose("reading non-present sector %d of cache block %d",
1560 			page, lba);
1561 		error = msb_read_page(msb, pba, page, &extra, &sg, offset);
1562 
1563 		/* Bad pages are copied with 00 page status */
1564 		if (error == -EBADMSG) {
1565 			pr_err("read error on sector %d, contents probably damaged", page);
1566 			continue;
1567 		}
1568 
1569 		if (error)
1570 			return error;
1571 
1572 		if ((extra.overwrite_flag & MEMSTICK_OV_PG_NORMAL) !=
1573 							MEMSTICK_OV_PG_NORMAL) {
1574 			dbg("page %d is marked as bad", page);
1575 			continue;
1576 		}
1577 
1578 		set_bit(page, &msb->valid_cache_bitmap);
1579 	}
1580 
1581 	/* Write the cache now */
1582 	error = msb_update_block(msb, msb->cache_block_lba, &sg, 0);
1583 	pba = msb->lba_to_pba_table[msb->cache_block_lba];
1584 
1585 	/* Mark invalid pages */
1586 	if (!error) {
1587 		for (page = 0; page < msb->pages_in_block; page++) {
1588 
1589 			if (test_bit(page, &msb->valid_cache_bitmap))
1590 				continue;
1591 
1592 			dbg("marking page %d as containing damaged data",
1593 				page);
1594 			msb_set_overwrite_flag(msb,
1595 				pba , page, 0xFF & ~MEMSTICK_OV_PG_NORMAL);
1596 		}
1597 	}
1598 
1599 	msb_cache_discard(msb);
1600 	return error;
1601 }
1602 
1603 static int msb_cache_write(struct msb_data *msb, int lba,
1604 	int page, bool add_to_cache_only, struct scatterlist *sg, int offset)
1605 {
1606 	int error;
1607 	struct scatterlist sg_tmp[10];
1608 
1609 	if (msb->read_only)
1610 		return -EROFS;
1611 
1612 	if (msb->cache_block_lba == MS_BLOCK_INVALID ||
1613 						lba != msb->cache_block_lba)
1614 		if (add_to_cache_only)
1615 			return 0;
1616 
1617 	/* If we need to write different block */
1618 	if (msb->cache_block_lba != MS_BLOCK_INVALID &&
1619 						lba != msb->cache_block_lba) {
1620 		dbg_verbose("first flush the cache");
1621 		error = msb_cache_flush(msb);
1622 		if (error)
1623 			return error;
1624 	}
1625 
1626 	if (msb->cache_block_lba  == MS_BLOCK_INVALID) {
1627 		msb->cache_block_lba  = lba;
1628 		mod_timer(&msb->cache_flush_timer,
1629 			jiffies + msecs_to_jiffies(cache_flush_timeout));
1630 	}
1631 
1632 	dbg_verbose("Write of LBA %d page %d to cache ", lba, page);
1633 
1634 	sg_init_table(sg_tmp, ARRAY_SIZE(sg_tmp));
1635 	msb_sg_copy(sg, sg_tmp, ARRAY_SIZE(sg_tmp), offset, msb->page_size);
1636 
1637 	sg_copy_to_buffer(sg_tmp, sg_nents(sg_tmp),
1638 		msb->cache + page * msb->page_size, msb->page_size);
1639 
1640 	set_bit(page, &msb->valid_cache_bitmap);
1641 	return 0;
1642 }
1643 
1644 static int msb_cache_read(struct msb_data *msb, int lba,
1645 				int page, struct scatterlist *sg, int offset)
1646 {
1647 	int pba = msb->lba_to_pba_table[lba];
1648 	struct scatterlist sg_tmp[10];
1649 	int error = 0;
1650 
1651 	if (lba == msb->cache_block_lba &&
1652 			test_bit(page, &msb->valid_cache_bitmap)) {
1653 
1654 		dbg_verbose("Read of LBA %d (pba %d) sector %d from cache",
1655 							lba, pba, page);
1656 
1657 		sg_init_table(sg_tmp, ARRAY_SIZE(sg_tmp));
1658 		msb_sg_copy(sg, sg_tmp, ARRAY_SIZE(sg_tmp),
1659 			offset, msb->page_size);
1660 		sg_copy_from_buffer(sg_tmp, sg_nents(sg_tmp),
1661 			msb->cache + msb->page_size * page,
1662 							msb->page_size);
1663 	} else {
1664 		dbg_verbose("Read of LBA %d (pba %d) sector %d from device",
1665 							lba, pba, page);
1666 
1667 		error = msb_read_page(msb, pba, page, NULL, sg, offset);
1668 		if (error)
1669 			return error;
1670 
1671 		msb_cache_write(msb, lba, page, true, sg, offset);
1672 	}
1673 	return error;
1674 }
1675 
1676 /* Emulated geometry table
1677  * This table content isn't that importaint,
1678  * One could put here different values, providing that they still
1679  * cover whole disk.
1680  * 64 MB entry is what windows reports for my 64M memstick */
1681 
1682 static const struct chs_entry chs_table[] = {
1683 /*        size sectors cylynders  heads */
1684 	{ 4,    16,    247,       2  },
1685 	{ 8,    16,    495,       2  },
1686 	{ 16,   16,    495,       4  },
1687 	{ 32,   16,    991,       4  },
1688 	{ 64,   16,    991,       8  },
1689 	{128,   16,    991,       16 },
1690 	{ 0 }
1691 };
1692 
1693 /* Load information about the card */
1694 static int msb_init_card(struct memstick_dev *card)
1695 {
1696 	struct msb_data *msb = memstick_get_drvdata(card);
1697 	struct memstick_host *host = card->host;
1698 	struct ms_boot_page *boot_block;
1699 	int error = 0, i, raw_size_in_megs;
1700 
1701 	msb->caps = 0;
1702 
1703 	if (card->id.class >= MEMSTICK_CLASS_ROM &&
1704 				card->id.class <= MEMSTICK_CLASS_ROM)
1705 		msb->read_only = true;
1706 
1707 	msb->state = -1;
1708 	error = msb_reset(msb, false);
1709 	if (error)
1710 		return error;
1711 
1712 	/* Due to a bug in Jmicron driver written by Alex Dubov,
1713 	 its serial mode barely works,
1714 	 so we switch to parallel mode right away */
1715 	if (host->caps & MEMSTICK_CAP_PAR4)
1716 		msb_switch_to_parallel(msb);
1717 
1718 	msb->page_size = sizeof(struct ms_boot_page);
1719 
1720 	/* Read the boot page */
1721 	error = msb_read_boot_blocks(msb);
1722 	if (error)
1723 		return -EIO;
1724 
1725 	boot_block = &msb->boot_page[0];
1726 
1727 	/* Save intersting attributes from boot page */
1728 	msb->block_count = boot_block->attr.number_of_blocks;
1729 	msb->page_size = boot_block->attr.page_size;
1730 
1731 	msb->pages_in_block = boot_block->attr.block_size * 2;
1732 	msb->block_size = msb->page_size * msb->pages_in_block;
1733 
1734 	if (msb->page_size > PAGE_SIZE) {
1735 		/* this isn't supported by linux at all, anyway*/
1736 		dbg("device page %d size isn't supported", msb->page_size);
1737 		return -EINVAL;
1738 	}
1739 
1740 	msb->block_buffer = kzalloc(msb->block_size, GFP_KERNEL);
1741 	if (!msb->block_buffer)
1742 		return -ENOMEM;
1743 
1744 	raw_size_in_megs = (msb->block_size * msb->block_count) >> 20;
1745 
1746 	for (i = 0; chs_table[i].size; i++) {
1747 
1748 		if (chs_table[i].size != raw_size_in_megs)
1749 			continue;
1750 
1751 		msb->geometry.cylinders = chs_table[i].cyl;
1752 		msb->geometry.heads = chs_table[i].head;
1753 		msb->geometry.sectors = chs_table[i].sec;
1754 		break;
1755 	}
1756 
1757 	if (boot_block->attr.transfer_supporting == 1)
1758 		msb->caps |= MEMSTICK_CAP_PAR4;
1759 
1760 	if (boot_block->attr.device_type & 0x03)
1761 		msb->read_only = true;
1762 
1763 	dbg("Total block count = %d", msb->block_count);
1764 	dbg("Each block consists of %d pages", msb->pages_in_block);
1765 	dbg("Page size = %d bytes", msb->page_size);
1766 	dbg("Parallel mode supported: %d", !!(msb->caps & MEMSTICK_CAP_PAR4));
1767 	dbg("Read only: %d", msb->read_only);
1768 
1769 #if 0
1770 	/* Now we can switch the interface */
1771 	if (host->caps & msb->caps & MEMSTICK_CAP_PAR4)
1772 		msb_switch_to_parallel(msb);
1773 #endif
1774 
1775 	error = msb_cache_init(msb);
1776 	if (error)
1777 		return error;
1778 
1779 	error = msb_ftl_initialize(msb);
1780 	if (error)
1781 		return error;
1782 
1783 
1784 	/* Read the bad block table */
1785 	error = msb_read_bad_block_table(msb, 0);
1786 
1787 	if (error && error != -ENOMEM) {
1788 		dbg("failed to read bad block table from primary boot block, trying from backup");
1789 		error = msb_read_bad_block_table(msb, 1);
1790 	}
1791 
1792 	if (error)
1793 		return error;
1794 
1795 	/* *drum roll* Scan the media */
1796 	error = msb_ftl_scan(msb);
1797 	if (error) {
1798 		pr_err("Scan of media failed");
1799 		return error;
1800 	}
1801 
1802 	return 0;
1803 
1804 }
1805 
1806 static int msb_do_write_request(struct msb_data *msb, int lba,
1807 	int page, struct scatterlist *sg, size_t len, int *sucessfuly_written)
1808 {
1809 	int error = 0;
1810 	off_t offset = 0;
1811 	*sucessfuly_written = 0;
1812 
1813 	while (offset < len) {
1814 		if (page == 0 && len - offset >= msb->block_size) {
1815 
1816 			if (msb->cache_block_lba == lba)
1817 				msb_cache_discard(msb);
1818 
1819 			dbg_verbose("Writing whole lba %d", lba);
1820 			error = msb_update_block(msb, lba, sg, offset);
1821 			if (error)
1822 				return error;
1823 
1824 			offset += msb->block_size;
1825 			*sucessfuly_written += msb->block_size;
1826 			lba++;
1827 			continue;
1828 		}
1829 
1830 		error = msb_cache_write(msb, lba, page, false, sg, offset);
1831 		if (error)
1832 			return error;
1833 
1834 		offset += msb->page_size;
1835 		*sucessfuly_written += msb->page_size;
1836 
1837 		page++;
1838 		if (page == msb->pages_in_block) {
1839 			page = 0;
1840 			lba++;
1841 		}
1842 	}
1843 	return 0;
1844 }
1845 
1846 static int msb_do_read_request(struct msb_data *msb, int lba,
1847 		int page, struct scatterlist *sg, int len, int *sucessfuly_read)
1848 {
1849 	int error = 0;
1850 	int offset = 0;
1851 	*sucessfuly_read = 0;
1852 
1853 	while (offset < len) {
1854 
1855 		error = msb_cache_read(msb, lba, page, sg, offset);
1856 		if (error)
1857 			return error;
1858 
1859 		offset += msb->page_size;
1860 		*sucessfuly_read += msb->page_size;
1861 
1862 		page++;
1863 		if (page == msb->pages_in_block) {
1864 			page = 0;
1865 			lba++;
1866 		}
1867 	}
1868 	return 0;
1869 }
1870 
1871 static void msb_io_work(struct work_struct *work)
1872 {
1873 	struct msb_data *msb = container_of(work, struct msb_data, io_work);
1874 	int page, error, len;
1875 	sector_t lba;
1876 	struct scatterlist *sg = msb->prealloc_sg;
1877 	struct request *req;
1878 
1879 	dbg_verbose("IO: work started");
1880 
1881 	while (1) {
1882 		spin_lock_irq(&msb->q_lock);
1883 
1884 		if (msb->need_flush_cache) {
1885 			msb->need_flush_cache = false;
1886 			spin_unlock_irq(&msb->q_lock);
1887 			msb_cache_flush(msb);
1888 			continue;
1889 		}
1890 
1891 		req = msb->req;
1892 		if (!req) {
1893 			dbg_verbose("IO: no more requests exiting");
1894 			spin_unlock_irq(&msb->q_lock);
1895 			return;
1896 		}
1897 
1898 		spin_unlock_irq(&msb->q_lock);
1899 
1900 		/* process the request */
1901 		dbg_verbose("IO: processing new request");
1902 		blk_rq_map_sg(msb->queue, req, sg);
1903 
1904 		lba = blk_rq_pos(req);
1905 
1906 		sector_div(lba, msb->page_size / 512);
1907 		page = sector_div(lba, msb->pages_in_block);
1908 
1909 		if (rq_data_dir(msb->req) == READ)
1910 			error = msb_do_read_request(msb, lba, page, sg,
1911 				blk_rq_bytes(req), &len);
1912 		else
1913 			error = msb_do_write_request(msb, lba, page, sg,
1914 				blk_rq_bytes(req), &len);
1915 
1916 		if (len && !blk_update_request(req, BLK_STS_OK, len)) {
1917 			__blk_mq_end_request(req, BLK_STS_OK);
1918 			spin_lock_irq(&msb->q_lock);
1919 			msb->req = NULL;
1920 			spin_unlock_irq(&msb->q_lock);
1921 		}
1922 
1923 		if (error && msb->req) {
1924 			blk_status_t ret = errno_to_blk_status(error);
1925 
1926 			dbg_verbose("IO: ending one sector of the request with error");
1927 			blk_mq_end_request(req, ret);
1928 			spin_lock_irq(&msb->q_lock);
1929 			msb->req = NULL;
1930 			spin_unlock_irq(&msb->q_lock);
1931 		}
1932 
1933 		if (msb->req)
1934 			dbg_verbose("IO: request still pending");
1935 	}
1936 }
1937 
1938 static DEFINE_IDR(msb_disk_idr); /*set of used disk numbers */
1939 static DEFINE_MUTEX(msb_disk_lock); /* protects against races in open/release */
1940 
1941 static int msb_bd_open(struct block_device *bdev, fmode_t mode)
1942 {
1943 	struct gendisk *disk = bdev->bd_disk;
1944 	struct msb_data *msb = disk->private_data;
1945 
1946 	dbg_verbose("block device open");
1947 
1948 	mutex_lock(&msb_disk_lock);
1949 
1950 	if (msb && msb->card)
1951 		msb->usage_count++;
1952 
1953 	mutex_unlock(&msb_disk_lock);
1954 	return 0;
1955 }
1956 
1957 static void msb_data_clear(struct msb_data *msb)
1958 {
1959 	kfree(msb->boot_page);
1960 	kfree(msb->used_blocks_bitmap);
1961 	kfree(msb->lba_to_pba_table);
1962 	kfree(msb->cache);
1963 	msb->card = NULL;
1964 }
1965 
1966 static int msb_disk_release(struct gendisk *disk)
1967 {
1968 	struct msb_data *msb = disk->private_data;
1969 
1970 	dbg_verbose("block device release");
1971 	mutex_lock(&msb_disk_lock);
1972 
1973 	if (msb) {
1974 		if (msb->usage_count)
1975 			msb->usage_count--;
1976 
1977 		if (!msb->usage_count) {
1978 			disk->private_data = NULL;
1979 			idr_remove(&msb_disk_idr, msb->disk_id);
1980 			put_disk(disk);
1981 			kfree(msb);
1982 		}
1983 	}
1984 	mutex_unlock(&msb_disk_lock);
1985 	return 0;
1986 }
1987 
1988 static void msb_bd_release(struct gendisk *disk, fmode_t mode)
1989 {
1990 	msb_disk_release(disk);
1991 }
1992 
1993 static int msb_bd_getgeo(struct block_device *bdev,
1994 				 struct hd_geometry *geo)
1995 {
1996 	struct msb_data *msb = bdev->bd_disk->private_data;
1997 	*geo = msb->geometry;
1998 	return 0;
1999 }
2000 
2001 static blk_status_t msb_queue_rq(struct blk_mq_hw_ctx *hctx,
2002 				 const struct blk_mq_queue_data *bd)
2003 {
2004 	struct memstick_dev *card = hctx->queue->queuedata;
2005 	struct msb_data *msb = memstick_get_drvdata(card);
2006 	struct request *req = bd->rq;
2007 
2008 	dbg_verbose("Submit request");
2009 
2010 	spin_lock_irq(&msb->q_lock);
2011 
2012 	if (msb->card_dead) {
2013 		dbg("Refusing requests on removed card");
2014 
2015 		WARN_ON(!msb->io_queue_stopped);
2016 
2017 		spin_unlock_irq(&msb->q_lock);
2018 		blk_mq_start_request(req);
2019 		return BLK_STS_IOERR;
2020 	}
2021 
2022 	if (msb->req) {
2023 		spin_unlock_irq(&msb->q_lock);
2024 		return BLK_STS_DEV_RESOURCE;
2025 	}
2026 
2027 	blk_mq_start_request(req);
2028 	msb->req = req;
2029 
2030 	if (!msb->io_queue_stopped)
2031 		queue_work(msb->io_queue, &msb->io_work);
2032 
2033 	spin_unlock_irq(&msb->q_lock);
2034 	return BLK_STS_OK;
2035 }
2036 
2037 static int msb_check_card(struct memstick_dev *card)
2038 {
2039 	struct msb_data *msb = memstick_get_drvdata(card);
2040 	return (msb->card_dead == 0);
2041 }
2042 
2043 static void msb_stop(struct memstick_dev *card)
2044 {
2045 	struct msb_data *msb = memstick_get_drvdata(card);
2046 	unsigned long flags;
2047 
2048 	dbg("Stopping all msblock IO");
2049 
2050 	blk_mq_stop_hw_queues(msb->queue);
2051 	spin_lock_irqsave(&msb->q_lock, flags);
2052 	msb->io_queue_stopped = true;
2053 	spin_unlock_irqrestore(&msb->q_lock, flags);
2054 
2055 	del_timer_sync(&msb->cache_flush_timer);
2056 	flush_workqueue(msb->io_queue);
2057 
2058 	spin_lock_irqsave(&msb->q_lock, flags);
2059 	if (msb->req) {
2060 		blk_mq_requeue_request(msb->req, false);
2061 		msb->req = NULL;
2062 	}
2063 	spin_unlock_irqrestore(&msb->q_lock, flags);
2064 }
2065 
2066 static void msb_start(struct memstick_dev *card)
2067 {
2068 	struct msb_data *msb = memstick_get_drvdata(card);
2069 	unsigned long flags;
2070 
2071 	dbg("Resuming IO from msblock");
2072 
2073 	msb_invalidate_reg_window(msb);
2074 
2075 	spin_lock_irqsave(&msb->q_lock, flags);
2076 	if (!msb->io_queue_stopped || msb->card_dead) {
2077 		spin_unlock_irqrestore(&msb->q_lock, flags);
2078 		return;
2079 	}
2080 	spin_unlock_irqrestore(&msb->q_lock, flags);
2081 
2082 	/* Kick cache flush anyway, its harmless */
2083 	msb->need_flush_cache = true;
2084 	msb->io_queue_stopped = false;
2085 
2086 	blk_mq_start_hw_queues(msb->queue);
2087 
2088 	queue_work(msb->io_queue, &msb->io_work);
2089 
2090 }
2091 
2092 static const struct block_device_operations msb_bdops = {
2093 	.open    = msb_bd_open,
2094 	.release = msb_bd_release,
2095 	.getgeo  = msb_bd_getgeo,
2096 	.owner   = THIS_MODULE
2097 };
2098 
2099 static const struct blk_mq_ops msb_mq_ops = {
2100 	.queue_rq	= msb_queue_rq,
2101 };
2102 
2103 /* Registers the block device */
2104 static int msb_init_disk(struct memstick_dev *card)
2105 {
2106 	struct msb_data *msb = memstick_get_drvdata(card);
2107 	int rc;
2108 	unsigned long capacity;
2109 
2110 	mutex_lock(&msb_disk_lock);
2111 	msb->disk_id = idr_alloc(&msb_disk_idr, card, 0, 256, GFP_KERNEL);
2112 	mutex_unlock(&msb_disk_lock);
2113 
2114 	if (msb->disk_id  < 0)
2115 		return msb->disk_id;
2116 
2117 	msb->disk = alloc_disk(0);
2118 	if (!msb->disk) {
2119 		rc = -ENOMEM;
2120 		goto out_release_id;
2121 	}
2122 
2123 	msb->queue = blk_mq_init_sq_queue(&msb->tag_set, &msb_mq_ops, 2,
2124 						BLK_MQ_F_SHOULD_MERGE);
2125 	if (IS_ERR(msb->queue)) {
2126 		rc = PTR_ERR(msb->queue);
2127 		msb->queue = NULL;
2128 		goto out_put_disk;
2129 	}
2130 
2131 	msb->queue->queuedata = card;
2132 
2133 	blk_queue_max_hw_sectors(msb->queue, MS_BLOCK_MAX_PAGES);
2134 	blk_queue_max_segments(msb->queue, MS_BLOCK_MAX_SEGS);
2135 	blk_queue_max_segment_size(msb->queue,
2136 				   MS_BLOCK_MAX_PAGES * msb->page_size);
2137 	blk_queue_logical_block_size(msb->queue, msb->page_size);
2138 
2139 	sprintf(msb->disk->disk_name, "msblk%d", msb->disk_id);
2140 	msb->disk->fops = &msb_bdops;
2141 	msb->disk->private_data = msb;
2142 	msb->disk->queue = msb->queue;
2143 	msb->disk->flags |= GENHD_FL_EXT_DEVT;
2144 
2145 	capacity = msb->pages_in_block * msb->logical_block_count;
2146 	capacity *= (msb->page_size / 512);
2147 	set_capacity(msb->disk, capacity);
2148 	dbg("Set total disk size to %lu sectors", capacity);
2149 
2150 	msb->usage_count = 1;
2151 	msb->io_queue = alloc_ordered_workqueue("ms_block", WQ_MEM_RECLAIM);
2152 	INIT_WORK(&msb->io_work, msb_io_work);
2153 	sg_init_table(msb->prealloc_sg, MS_BLOCK_MAX_SEGS+1);
2154 
2155 	if (msb->read_only)
2156 		set_disk_ro(msb->disk, 1);
2157 
2158 	msb_start(card);
2159 	device_add_disk(&card->dev, msb->disk, NULL);
2160 	dbg("Disk added");
2161 	return 0;
2162 
2163 out_put_disk:
2164 	put_disk(msb->disk);
2165 out_release_id:
2166 	mutex_lock(&msb_disk_lock);
2167 	idr_remove(&msb_disk_idr, msb->disk_id);
2168 	mutex_unlock(&msb_disk_lock);
2169 	return rc;
2170 }
2171 
2172 static int msb_probe(struct memstick_dev *card)
2173 {
2174 	struct msb_data *msb;
2175 	int rc = 0;
2176 
2177 	msb = kzalloc(sizeof(struct msb_data), GFP_KERNEL);
2178 	if (!msb)
2179 		return -ENOMEM;
2180 	memstick_set_drvdata(card, msb);
2181 	msb->card = card;
2182 	spin_lock_init(&msb->q_lock);
2183 
2184 	rc = msb_init_card(card);
2185 	if (rc)
2186 		goto out_free;
2187 
2188 	rc = msb_init_disk(card);
2189 	if (!rc) {
2190 		card->check = msb_check_card;
2191 		card->stop = msb_stop;
2192 		card->start = msb_start;
2193 		return 0;
2194 	}
2195 out_free:
2196 	memstick_set_drvdata(card, NULL);
2197 	msb_data_clear(msb);
2198 	kfree(msb);
2199 	return rc;
2200 }
2201 
2202 static void msb_remove(struct memstick_dev *card)
2203 {
2204 	struct msb_data *msb = memstick_get_drvdata(card);
2205 	unsigned long flags;
2206 
2207 	if (!msb->io_queue_stopped)
2208 		msb_stop(card);
2209 
2210 	dbg("Removing the disk device");
2211 
2212 	/* Take care of unhandled + new requests from now on */
2213 	spin_lock_irqsave(&msb->q_lock, flags);
2214 	msb->card_dead = true;
2215 	spin_unlock_irqrestore(&msb->q_lock, flags);
2216 	blk_mq_start_hw_queues(msb->queue);
2217 
2218 	/* Remove the disk */
2219 	del_gendisk(msb->disk);
2220 	blk_cleanup_queue(msb->queue);
2221 	blk_mq_free_tag_set(&msb->tag_set);
2222 	msb->queue = NULL;
2223 
2224 	mutex_lock(&msb_disk_lock);
2225 	msb_data_clear(msb);
2226 	mutex_unlock(&msb_disk_lock);
2227 
2228 	msb_disk_release(msb->disk);
2229 	memstick_set_drvdata(card, NULL);
2230 }
2231 
2232 #ifdef CONFIG_PM
2233 
2234 static int msb_suspend(struct memstick_dev *card, pm_message_t state)
2235 {
2236 	msb_stop(card);
2237 	return 0;
2238 }
2239 
2240 static int msb_resume(struct memstick_dev *card)
2241 {
2242 	struct msb_data *msb = memstick_get_drvdata(card);
2243 	struct msb_data *new_msb = NULL;
2244 	bool card_dead = true;
2245 
2246 #ifndef CONFIG_MEMSTICK_UNSAFE_RESUME
2247 	msb->card_dead = true;
2248 	return 0;
2249 #endif
2250 	mutex_lock(&card->host->lock);
2251 
2252 	new_msb = kzalloc(sizeof(struct msb_data), GFP_KERNEL);
2253 	if (!new_msb)
2254 		goto out;
2255 
2256 	new_msb->card = card;
2257 	memstick_set_drvdata(card, new_msb);
2258 	spin_lock_init(&new_msb->q_lock);
2259 	sg_init_table(msb->prealloc_sg, MS_BLOCK_MAX_SEGS+1);
2260 
2261 	if (msb_init_card(card))
2262 		goto out;
2263 
2264 	if (msb->block_size != new_msb->block_size)
2265 		goto out;
2266 
2267 	if (memcmp(msb->boot_page, new_msb->boot_page,
2268 					sizeof(struct ms_boot_page)))
2269 		goto out;
2270 
2271 	if (msb->logical_block_count != new_msb->logical_block_count ||
2272 		memcmp(msb->lba_to_pba_table, new_msb->lba_to_pba_table,
2273 						msb->logical_block_count))
2274 		goto out;
2275 
2276 	if (msb->block_count != new_msb->block_count ||
2277 		memcmp(msb->used_blocks_bitmap, new_msb->used_blocks_bitmap,
2278 							msb->block_count / 8))
2279 		goto out;
2280 
2281 	card_dead = false;
2282 out:
2283 	if (card_dead)
2284 		dbg("Card was removed/replaced during suspend");
2285 
2286 	msb->card_dead = card_dead;
2287 	memstick_set_drvdata(card, msb);
2288 
2289 	if (new_msb) {
2290 		msb_data_clear(new_msb);
2291 		kfree(new_msb);
2292 	}
2293 
2294 	msb_start(card);
2295 	mutex_unlock(&card->host->lock);
2296 	return 0;
2297 }
2298 #else
2299 
2300 #define msb_suspend NULL
2301 #define msb_resume NULL
2302 
2303 #endif /* CONFIG_PM */
2304 
2305 static struct memstick_device_id msb_id_tbl[] = {
2306 	{MEMSTICK_MATCH_ALL, MEMSTICK_TYPE_LEGACY, MEMSTICK_CATEGORY_STORAGE,
2307 	 MEMSTICK_CLASS_FLASH},
2308 
2309 	{MEMSTICK_MATCH_ALL, MEMSTICK_TYPE_LEGACY, MEMSTICK_CATEGORY_STORAGE,
2310 	 MEMSTICK_CLASS_ROM},
2311 
2312 	{MEMSTICK_MATCH_ALL, MEMSTICK_TYPE_LEGACY, MEMSTICK_CATEGORY_STORAGE,
2313 	 MEMSTICK_CLASS_RO},
2314 
2315 	{MEMSTICK_MATCH_ALL, MEMSTICK_TYPE_LEGACY, MEMSTICK_CATEGORY_STORAGE,
2316 	 MEMSTICK_CLASS_WP},
2317 
2318 	{MEMSTICK_MATCH_ALL, MEMSTICK_TYPE_DUO, MEMSTICK_CATEGORY_STORAGE_DUO,
2319 	 MEMSTICK_CLASS_DUO},
2320 	{}
2321 };
2322 MODULE_DEVICE_TABLE(memstick, msb_id_tbl);
2323 
2324 
2325 static struct memstick_driver msb_driver = {
2326 	.driver = {
2327 		.name  = DRIVER_NAME,
2328 		.owner = THIS_MODULE
2329 	},
2330 	.id_table = msb_id_tbl,
2331 	.probe    = msb_probe,
2332 	.remove   = msb_remove,
2333 	.suspend  = msb_suspend,
2334 	.resume   = msb_resume
2335 };
2336 
2337 static int __init msb_init(void)
2338 {
2339 	int rc = memstick_register_driver(&msb_driver);
2340 	if (rc)
2341 		pr_err("failed to register memstick driver (error %d)\n", rc);
2342 
2343 	return rc;
2344 }
2345 
2346 static void __exit msb_exit(void)
2347 {
2348 	memstick_unregister_driver(&msb_driver);
2349 	idr_destroy(&msb_disk_idr);
2350 }
2351 
2352 module_init(msb_init);
2353 module_exit(msb_exit);
2354 
2355 module_param(cache_flush_timeout, int, S_IRUGO);
2356 MODULE_PARM_DESC(cache_flush_timeout,
2357 				"Cache flush timeout in msec (1000 default)");
2358 module_param(debug, int, S_IRUGO | S_IWUSR);
2359 MODULE_PARM_DESC(debug, "Debug level (0-2)");
2360 
2361 module_param(verify_writes, bool, S_IRUGO);
2362 MODULE_PARM_DESC(verify_writes, "Read back and check all data that is written");
2363 
2364 MODULE_LICENSE("GPL");
2365 MODULE_AUTHOR("Maxim Levitsky");
2366 MODULE_DESCRIPTION("Sony MemoryStick block device driver");
2367