xref: /openbmc/u-boot/drivers/mmc/mmc.c (revision 57efeb04)
1 // SPDX-License-Identifier: GPL-2.0+
2 /*
3  * Copyright 2008, Freescale Semiconductor, Inc
4  * Andy Fleming
5  *
6  * Based vaguely on the Linux code
7  */
8 
9 #include <config.h>
10 #include <common.h>
11 #include <command.h>
12 #include <dm.h>
13 #include <dm/device-internal.h>
14 #include <errno.h>
15 #include <mmc.h>
16 #include <part.h>
17 #include <power/regulator.h>
18 #include <malloc.h>
19 #include <memalign.h>
20 #include <linux/list.h>
21 #include <div64.h>
22 #include "mmc_private.h"
23 
24 static int mmc_set_signal_voltage(struct mmc *mmc, uint signal_voltage);
25 static int mmc_power_cycle(struct mmc *mmc);
26 #if !CONFIG_IS_ENABLED(MMC_TINY)
27 static int mmc_select_mode_and_width(struct mmc *mmc, uint card_caps);
28 #endif
29 
30 #if !CONFIG_IS_ENABLED(DM_MMC)
31 
32 #if CONFIG_IS_ENABLED(MMC_UHS_SUPPORT)
33 static int mmc_wait_dat0(struct mmc *mmc, int state, int timeout)
34 {
35 	return -ENOSYS;
36 }
37 #endif
38 
39 __weak int board_mmc_getwp(struct mmc *mmc)
40 {
41 	return -1;
42 }
43 
44 int mmc_getwp(struct mmc *mmc)
45 {
46 	int wp;
47 
48 	wp = board_mmc_getwp(mmc);
49 
50 	if (wp < 0) {
51 		if (mmc->cfg->ops->getwp)
52 			wp = mmc->cfg->ops->getwp(mmc);
53 		else
54 			wp = 0;
55 	}
56 
57 	return wp;
58 }
59 
60 __weak int board_mmc_getcd(struct mmc *mmc)
61 {
62 	return -1;
63 }
64 #endif
65 
66 #ifdef CONFIG_MMC_TRACE
67 void mmmc_trace_before_send(struct mmc *mmc, struct mmc_cmd *cmd)
68 {
69 	printf("CMD_SEND:%d\n", cmd->cmdidx);
70 	printf("\t\tARG\t\t\t 0x%08x\n", cmd->cmdarg);
71 }
72 
73 void mmmc_trace_after_send(struct mmc *mmc, struct mmc_cmd *cmd, int ret)
74 {
75 	int i;
76 	u8 *ptr;
77 
78 	if (ret) {
79 		printf("\t\tRET\t\t\t %d\n", ret);
80 	} else {
81 		switch (cmd->resp_type) {
82 		case MMC_RSP_NONE:
83 			printf("\t\tMMC_RSP_NONE\n");
84 			break;
85 		case MMC_RSP_R1:
86 			printf("\t\tMMC_RSP_R1,5,6,7 \t 0x%08x \n",
87 				cmd->response[0]);
88 			break;
89 		case MMC_RSP_R1b:
90 			printf("\t\tMMC_RSP_R1b\t\t 0x%08x \n",
91 				cmd->response[0]);
92 			break;
93 		case MMC_RSP_R2:
94 			printf("\t\tMMC_RSP_R2\t\t 0x%08x \n",
95 				cmd->response[0]);
96 			printf("\t\t          \t\t 0x%08x \n",
97 				cmd->response[1]);
98 			printf("\t\t          \t\t 0x%08x \n",
99 				cmd->response[2]);
100 			printf("\t\t          \t\t 0x%08x \n",
101 				cmd->response[3]);
102 			printf("\n");
103 			printf("\t\t\t\t\tDUMPING DATA\n");
104 			for (i = 0; i < 4; i++) {
105 				int j;
106 				printf("\t\t\t\t\t%03d - ", i*4);
107 				ptr = (u8 *)&cmd->response[i];
108 				ptr += 3;
109 				for (j = 0; j < 4; j++)
110 					printf("%02x ", *ptr--);
111 				printf("\n");
112 			}
113 			break;
114 		case MMC_RSP_R3:
115 			printf("\t\tMMC_RSP_R3,4\t\t 0x%08x \n",
116 				cmd->response[0]);
117 			break;
118 		default:
119 			printf("\t\tERROR MMC rsp not supported\n");
120 			break;
121 		}
122 	}
123 }
124 
125 void mmc_trace_state(struct mmc *mmc, struct mmc_cmd *cmd)
126 {
127 	int status;
128 
129 	status = (cmd->response[0] & MMC_STATUS_CURR_STATE) >> 9;
130 	printf("CURR STATE:%d\n", status);
131 }
132 #endif
133 
134 #if CONFIG_IS_ENABLED(MMC_VERBOSE) || defined(DEBUG)
135 const char *mmc_mode_name(enum bus_mode mode)
136 {
137 	static const char *const names[] = {
138 	      [MMC_LEGACY]	= "MMC legacy",
139 	      [SD_LEGACY]	= "SD Legacy",
140 	      [MMC_HS]		= "MMC High Speed (26MHz)",
141 	      [SD_HS]		= "SD High Speed (50MHz)",
142 	      [UHS_SDR12]	= "UHS SDR12 (25MHz)",
143 	      [UHS_SDR25]	= "UHS SDR25 (50MHz)",
144 	      [UHS_SDR50]	= "UHS SDR50 (100MHz)",
145 	      [UHS_SDR104]	= "UHS SDR104 (208MHz)",
146 	      [UHS_DDR50]	= "UHS DDR50 (50MHz)",
147 	      [MMC_HS_52]	= "MMC High Speed (52MHz)",
148 	      [MMC_DDR_52]	= "MMC DDR52 (52MHz)",
149 	      [MMC_HS_200]	= "HS200 (200MHz)",
150 	      [MMC_HS_400]	= "HS400 (200MHz)",
151 	};
152 
153 	if (mode >= MMC_MODES_END)
154 		return "Unknown mode";
155 	else
156 		return names[mode];
157 }
158 #endif
159 
160 static uint mmc_mode2freq(struct mmc *mmc, enum bus_mode mode)
161 {
162 	static const int freqs[] = {
163 	      [MMC_LEGACY]	= 25000000,
164 	      [SD_LEGACY]	= 25000000,
165 	      [MMC_HS]		= 26000000,
166 	      [SD_HS]		= 50000000,
167 	      [MMC_HS_52]	= 52000000,
168 	      [MMC_DDR_52]	= 52000000,
169 	      [UHS_SDR12]	= 25000000,
170 	      [UHS_SDR25]	= 50000000,
171 	      [UHS_SDR50]	= 100000000,
172 	      [UHS_DDR50]	= 50000000,
173 	      [UHS_SDR104]	= 208000000,
174 	      [MMC_HS_200]	= 200000000,
175 	      [MMC_HS_400]	= 200000000,
176 	};
177 
178 	if (mode == MMC_LEGACY)
179 		return mmc->legacy_speed;
180 	else if (mode >= MMC_MODES_END)
181 		return 0;
182 	else
183 		return freqs[mode];
184 }
185 
186 static int mmc_select_mode(struct mmc *mmc, enum bus_mode mode)
187 {
188 	mmc->selected_mode = mode;
189 	mmc->tran_speed = mmc_mode2freq(mmc, mode);
190 	mmc->ddr_mode = mmc_is_mode_ddr(mode);
191 	pr_debug("selecting mode %s (freq : %d MHz)\n", mmc_mode_name(mode),
192 		 mmc->tran_speed / 1000000);
193 	return 0;
194 }
195 
196 #if !CONFIG_IS_ENABLED(DM_MMC)
197 int mmc_send_cmd(struct mmc *mmc, struct mmc_cmd *cmd, struct mmc_data *data)
198 {
199 	int ret;
200 
201 	mmmc_trace_before_send(mmc, cmd);
202 	ret = mmc->cfg->ops->send_cmd(mmc, cmd, data);
203 	mmmc_trace_after_send(mmc, cmd, ret);
204 
205 	return ret;
206 }
207 #endif
208 
209 int mmc_send_status(struct mmc *mmc, int timeout)
210 {
211 	struct mmc_cmd cmd;
212 	int err, retries = 5;
213 
214 	cmd.cmdidx = MMC_CMD_SEND_STATUS;
215 	cmd.resp_type = MMC_RSP_R1;
216 	if (!mmc_host_is_spi(mmc))
217 		cmd.cmdarg = mmc->rca << 16;
218 
219 	while (1) {
220 		err = mmc_send_cmd(mmc, &cmd, NULL);
221 		if (!err) {
222 			if ((cmd.response[0] & MMC_STATUS_RDY_FOR_DATA) &&
223 			    (cmd.response[0] & MMC_STATUS_CURR_STATE) !=
224 			     MMC_STATE_PRG)
225 				break;
226 
227 			if (cmd.response[0] & MMC_STATUS_MASK) {
228 #if !defined(CONFIG_SPL_BUILD) || defined(CONFIG_SPL_LIBCOMMON_SUPPORT)
229 				pr_err("Status Error: 0x%08x\n",
230 				       cmd.response[0]);
231 #endif
232 				return -ECOMM;
233 			}
234 		} else if (--retries < 0)
235 			return err;
236 
237 		if (timeout-- <= 0)
238 			break;
239 
240 		udelay(1000);
241 	}
242 
243 	mmc_trace_state(mmc, &cmd);
244 	if (timeout <= 0) {
245 #if !defined(CONFIG_SPL_BUILD) || defined(CONFIG_SPL_LIBCOMMON_SUPPORT)
246 		pr_err("Timeout waiting card ready\n");
247 #endif
248 		return -ETIMEDOUT;
249 	}
250 
251 	return 0;
252 }
253 
254 int mmc_set_blocklen(struct mmc *mmc, int len)
255 {
256 	struct mmc_cmd cmd;
257 	int err;
258 
259 	if (mmc->ddr_mode)
260 		return 0;
261 
262 	cmd.cmdidx = MMC_CMD_SET_BLOCKLEN;
263 	cmd.resp_type = MMC_RSP_R1;
264 	cmd.cmdarg = len;
265 
266 	err = mmc_send_cmd(mmc, &cmd, NULL);
267 
268 #ifdef CONFIG_MMC_QUIRKS
269 	if (err && (mmc->quirks & MMC_QUIRK_RETRY_SET_BLOCKLEN)) {
270 		int retries = 4;
271 		/*
272 		 * It has been seen that SET_BLOCKLEN may fail on the first
273 		 * attempt, let's try a few more time
274 		 */
275 		do {
276 			err = mmc_send_cmd(mmc, &cmd, NULL);
277 			if (!err)
278 				break;
279 		} while (retries--);
280 	}
281 #endif
282 
283 	return err;
284 }
285 
286 #ifdef MMC_SUPPORTS_TUNING
287 static const u8 tuning_blk_pattern_4bit[] = {
288 	0xff, 0x0f, 0xff, 0x00, 0xff, 0xcc, 0xc3, 0xcc,
289 	0xc3, 0x3c, 0xcc, 0xff, 0xfe, 0xff, 0xfe, 0xef,
290 	0xff, 0xdf, 0xff, 0xdd, 0xff, 0xfb, 0xff, 0xfb,
291 	0xbf, 0xff, 0x7f, 0xff, 0x77, 0xf7, 0xbd, 0xef,
292 	0xff, 0xf0, 0xff, 0xf0, 0x0f, 0xfc, 0xcc, 0x3c,
293 	0xcc, 0x33, 0xcc, 0xcf, 0xff, 0xef, 0xff, 0xee,
294 	0xff, 0xfd, 0xff, 0xfd, 0xdf, 0xff, 0xbf, 0xff,
295 	0xbb, 0xff, 0xf7, 0xff, 0xf7, 0x7f, 0x7b, 0xde,
296 };
297 
298 static const u8 tuning_blk_pattern_8bit[] = {
299 	0xff, 0xff, 0x00, 0xff, 0xff, 0xff, 0x00, 0x00,
300 	0xff, 0xff, 0xcc, 0xcc, 0xcc, 0x33, 0xcc, 0xcc,
301 	0xcc, 0x33, 0x33, 0xcc, 0xcc, 0xcc, 0xff, 0xff,
302 	0xff, 0xee, 0xff, 0xff, 0xff, 0xee, 0xee, 0xff,
303 	0xff, 0xff, 0xdd, 0xff, 0xff, 0xff, 0xdd, 0xdd,
304 	0xff, 0xff, 0xff, 0xbb, 0xff, 0xff, 0xff, 0xbb,
305 	0xbb, 0xff, 0xff, 0xff, 0x77, 0xff, 0xff, 0xff,
306 	0x77, 0x77, 0xff, 0x77, 0xbb, 0xdd, 0xee, 0xff,
307 	0xff, 0xff, 0xff, 0x00, 0xff, 0xff, 0xff, 0x00,
308 	0x00, 0xff, 0xff, 0xcc, 0xcc, 0xcc, 0x33, 0xcc,
309 	0xcc, 0xcc, 0x33, 0x33, 0xcc, 0xcc, 0xcc, 0xff,
310 	0xff, 0xff, 0xee, 0xff, 0xff, 0xff, 0xee, 0xee,
311 	0xff, 0xff, 0xff, 0xdd, 0xff, 0xff, 0xff, 0xdd,
312 	0xdd, 0xff, 0xff, 0xff, 0xbb, 0xff, 0xff, 0xff,
313 	0xbb, 0xbb, 0xff, 0xff, 0xff, 0x77, 0xff, 0xff,
314 	0xff, 0x77, 0x77, 0xff, 0x77, 0xbb, 0xdd, 0xee,
315 };
316 
317 int mmc_send_tuning(struct mmc *mmc, u32 opcode, int *cmd_error)
318 {
319 	struct mmc_cmd cmd;
320 	struct mmc_data data;
321 	const u8 *tuning_block_pattern;
322 	int size, err;
323 
324 	if (mmc->bus_width == 8) {
325 		tuning_block_pattern = tuning_blk_pattern_8bit;
326 		size = sizeof(tuning_blk_pattern_8bit);
327 	} else if (mmc->bus_width == 4) {
328 		tuning_block_pattern = tuning_blk_pattern_4bit;
329 		size = sizeof(tuning_blk_pattern_4bit);
330 	} else {
331 		return -EINVAL;
332 	}
333 
334 	ALLOC_CACHE_ALIGN_BUFFER(u8, data_buf, size);
335 
336 	cmd.cmdidx = opcode;
337 	cmd.cmdarg = 0;
338 	cmd.resp_type = MMC_RSP_R1;
339 
340 	data.dest = (void *)data_buf;
341 	data.blocks = 1;
342 	data.blocksize = size;
343 	data.flags = MMC_DATA_READ;
344 
345 	err = mmc_send_cmd(mmc, &cmd, &data);
346 	if (err)
347 		return err;
348 
349 	if (memcmp(data_buf, tuning_block_pattern, size))
350 		return -EIO;
351 
352 	return 0;
353 }
354 #endif
355 
356 static int mmc_read_blocks(struct mmc *mmc, void *dst, lbaint_t start,
357 			   lbaint_t blkcnt)
358 {
359 	struct mmc_cmd cmd;
360 	struct mmc_data data;
361 
362 	if (blkcnt > 1)
363 		cmd.cmdidx = MMC_CMD_READ_MULTIPLE_BLOCK;
364 	else
365 		cmd.cmdidx = MMC_CMD_READ_SINGLE_BLOCK;
366 
367 	if (mmc->high_capacity)
368 		cmd.cmdarg = start;
369 	else
370 		cmd.cmdarg = start * mmc->read_bl_len;
371 
372 	cmd.resp_type = MMC_RSP_R1;
373 
374 	data.dest = dst;
375 	data.blocks = blkcnt;
376 	data.blocksize = mmc->read_bl_len;
377 	data.flags = MMC_DATA_READ;
378 
379 	if (mmc_send_cmd(mmc, &cmd, &data))
380 		return 0;
381 
382 	if (blkcnt > 1) {
383 		cmd.cmdidx = MMC_CMD_STOP_TRANSMISSION;
384 		cmd.cmdarg = 0;
385 		cmd.resp_type = MMC_RSP_R1b;
386 		if (mmc_send_cmd(mmc, &cmd, NULL)) {
387 #if !defined(CONFIG_SPL_BUILD) || defined(CONFIG_SPL_LIBCOMMON_SUPPORT)
388 			pr_err("mmc fail to send stop cmd\n");
389 #endif
390 			return 0;
391 		}
392 	}
393 
394 	return blkcnt;
395 }
396 
397 #if CONFIG_IS_ENABLED(BLK)
398 ulong mmc_bread(struct udevice *dev, lbaint_t start, lbaint_t blkcnt, void *dst)
399 #else
400 ulong mmc_bread(struct blk_desc *block_dev, lbaint_t start, lbaint_t blkcnt,
401 		void *dst)
402 #endif
403 {
404 #if CONFIG_IS_ENABLED(BLK)
405 	struct blk_desc *block_dev = dev_get_uclass_platdata(dev);
406 #endif
407 	int dev_num = block_dev->devnum;
408 	int err;
409 	lbaint_t cur, blocks_todo = blkcnt;
410 
411 	if (blkcnt == 0)
412 		return 0;
413 
414 	struct mmc *mmc = find_mmc_device(dev_num);
415 	if (!mmc)
416 		return 0;
417 
418 	if (CONFIG_IS_ENABLED(MMC_TINY))
419 		err = mmc_switch_part(mmc, block_dev->hwpart);
420 	else
421 		err = blk_dselect_hwpart(block_dev, block_dev->hwpart);
422 
423 	if (err < 0)
424 		return 0;
425 
426 	if ((start + blkcnt) > block_dev->lba) {
427 #if !defined(CONFIG_SPL_BUILD) || defined(CONFIG_SPL_LIBCOMMON_SUPPORT)
428 		pr_err("MMC: block number 0x" LBAF " exceeds max(0x" LBAF ")\n",
429 		       start + blkcnt, block_dev->lba);
430 #endif
431 		return 0;
432 	}
433 
434 	if (mmc_set_blocklen(mmc, mmc->read_bl_len)) {
435 		pr_debug("%s: Failed to set blocklen\n", __func__);
436 		return 0;
437 	}
438 
439 	do {
440 		cur = (blocks_todo > mmc->cfg->b_max) ?
441 			mmc->cfg->b_max : blocks_todo;
442 		if (mmc_read_blocks(mmc, dst, start, cur) != cur) {
443 			pr_debug("%s: Failed to read blocks\n", __func__);
444 			return 0;
445 		}
446 		blocks_todo -= cur;
447 		start += cur;
448 		dst += cur * mmc->read_bl_len;
449 	} while (blocks_todo > 0);
450 
451 	return blkcnt;
452 }
453 
454 static int mmc_go_idle(struct mmc *mmc)
455 {
456 	struct mmc_cmd cmd;
457 	int err;
458 
459 	udelay(1000);
460 
461 	cmd.cmdidx = MMC_CMD_GO_IDLE_STATE;
462 	cmd.cmdarg = 0;
463 	cmd.resp_type = MMC_RSP_NONE;
464 
465 	err = mmc_send_cmd(mmc, &cmd, NULL);
466 
467 	if (err)
468 		return err;
469 
470 	udelay(2000);
471 
472 	return 0;
473 }
474 
475 #if CONFIG_IS_ENABLED(MMC_UHS_SUPPORT)
476 static int mmc_switch_voltage(struct mmc *mmc, int signal_voltage)
477 {
478 	struct mmc_cmd cmd;
479 	int err = 0;
480 
481 	/*
482 	 * Send CMD11 only if the request is to switch the card to
483 	 * 1.8V signalling.
484 	 */
485 	if (signal_voltage == MMC_SIGNAL_VOLTAGE_330)
486 		return mmc_set_signal_voltage(mmc, signal_voltage);
487 
488 	cmd.cmdidx = SD_CMD_SWITCH_UHS18V;
489 	cmd.cmdarg = 0;
490 	cmd.resp_type = MMC_RSP_R1;
491 
492 	err = mmc_send_cmd(mmc, &cmd, NULL);
493 	if (err)
494 		return err;
495 
496 	if (!mmc_host_is_spi(mmc) && (cmd.response[0] & MMC_STATUS_ERROR))
497 		return -EIO;
498 
499 	/*
500 	 * The card should drive cmd and dat[0:3] low immediately
501 	 * after the response of cmd11, but wait 100 us to be sure
502 	 */
503 	err = mmc_wait_dat0(mmc, 0, 100);
504 	if (err == -ENOSYS)
505 		udelay(100);
506 	else if (err)
507 		return -ETIMEDOUT;
508 
509 	/*
510 	 * During a signal voltage level switch, the clock must be gated
511 	 * for 5 ms according to the SD spec
512 	 */
513 	mmc_set_clock(mmc, mmc->clock, MMC_CLK_DISABLE);
514 
515 	err = mmc_set_signal_voltage(mmc, signal_voltage);
516 	if (err)
517 		return err;
518 
519 	/* Keep clock gated for at least 10 ms, though spec only says 5 ms */
520 	mdelay(10);
521 	mmc_set_clock(mmc, mmc->clock, MMC_CLK_ENABLE);
522 
523 	/*
524 	 * Failure to switch is indicated by the card holding
525 	 * dat[0:3] low. Wait for at least 1 ms according to spec
526 	 */
527 	err = mmc_wait_dat0(mmc, 1, 1000);
528 	if (err == -ENOSYS)
529 		udelay(1000);
530 	else if (err)
531 		return -ETIMEDOUT;
532 
533 	return 0;
534 }
535 #endif
536 
537 static int sd_send_op_cond(struct mmc *mmc, bool uhs_en)
538 {
539 	int timeout = 1000;
540 	int err;
541 	struct mmc_cmd cmd;
542 
543 	while (1) {
544 		cmd.cmdidx = MMC_CMD_APP_CMD;
545 		cmd.resp_type = MMC_RSP_R1;
546 		cmd.cmdarg = 0;
547 
548 		err = mmc_send_cmd(mmc, &cmd, NULL);
549 
550 		if (err)
551 			return err;
552 
553 		cmd.cmdidx = SD_CMD_APP_SEND_OP_COND;
554 		cmd.resp_type = MMC_RSP_R3;
555 
556 		/*
557 		 * Most cards do not answer if some reserved bits
558 		 * in the ocr are set. However, Some controller
559 		 * can set bit 7 (reserved for low voltages), but
560 		 * how to manage low voltages SD card is not yet
561 		 * specified.
562 		 */
563 		cmd.cmdarg = mmc_host_is_spi(mmc) ? 0 :
564 			(mmc->cfg->voltages & 0xff8000);
565 
566 		if (mmc->version == SD_VERSION_2)
567 			cmd.cmdarg |= OCR_HCS;
568 
569 		if (uhs_en)
570 			cmd.cmdarg |= OCR_S18R;
571 
572 		err = mmc_send_cmd(mmc, &cmd, NULL);
573 
574 		if (err)
575 			return err;
576 
577 		if (cmd.response[0] & OCR_BUSY)
578 			break;
579 
580 		if (timeout-- <= 0)
581 			return -EOPNOTSUPP;
582 
583 		udelay(1000);
584 	}
585 
586 	if (mmc->version != SD_VERSION_2)
587 		mmc->version = SD_VERSION_1_0;
588 
589 	if (mmc_host_is_spi(mmc)) { /* read OCR for spi */
590 		cmd.cmdidx = MMC_CMD_SPI_READ_OCR;
591 		cmd.resp_type = MMC_RSP_R3;
592 		cmd.cmdarg = 0;
593 
594 		err = mmc_send_cmd(mmc, &cmd, NULL);
595 
596 		if (err)
597 			return err;
598 	}
599 
600 	mmc->ocr = cmd.response[0];
601 
602 #if CONFIG_IS_ENABLED(MMC_UHS_SUPPORT)
603 	if (uhs_en && !(mmc_host_is_spi(mmc)) && (cmd.response[0] & 0x41000000)
604 	    == 0x41000000) {
605 		err = mmc_switch_voltage(mmc, MMC_SIGNAL_VOLTAGE_180);
606 		if (err)
607 			return err;
608 	}
609 #endif
610 
611 	mmc->high_capacity = ((mmc->ocr & OCR_HCS) == OCR_HCS);
612 	mmc->rca = 0;
613 
614 	return 0;
615 }
616 
617 static int mmc_send_op_cond_iter(struct mmc *mmc, int use_arg)
618 {
619 	struct mmc_cmd cmd;
620 	int err;
621 
622 	cmd.cmdidx = MMC_CMD_SEND_OP_COND;
623 	cmd.resp_type = MMC_RSP_R3;
624 	cmd.cmdarg = 0;
625 	if (use_arg && !mmc_host_is_spi(mmc))
626 		cmd.cmdarg = OCR_HCS |
627 			(mmc->cfg->voltages &
628 			(mmc->ocr & OCR_VOLTAGE_MASK)) |
629 			(mmc->ocr & OCR_ACCESS_MODE);
630 
631 	err = mmc_send_cmd(mmc, &cmd, NULL);
632 	if (err)
633 		return err;
634 	mmc->ocr = cmd.response[0];
635 	return 0;
636 }
637 
638 static int mmc_send_op_cond(struct mmc *mmc)
639 {
640 	int err, i;
641 
642 	/* Some cards seem to need this */
643 	mmc_go_idle(mmc);
644 
645  	/* Asking to the card its capabilities */
646 	for (i = 0; i < 2; i++) {
647 		err = mmc_send_op_cond_iter(mmc, i != 0);
648 		if (err)
649 			return err;
650 
651 		/* exit if not busy (flag seems to be inverted) */
652 		if (mmc->ocr & OCR_BUSY)
653 			break;
654 	}
655 	mmc->op_cond_pending = 1;
656 	return 0;
657 }
658 
659 static int mmc_complete_op_cond(struct mmc *mmc)
660 {
661 	struct mmc_cmd cmd;
662 	int timeout = 1000;
663 	ulong start;
664 	int err;
665 
666 	mmc->op_cond_pending = 0;
667 	if (!(mmc->ocr & OCR_BUSY)) {
668 		/* Some cards seem to need this */
669 		mmc_go_idle(mmc);
670 
671 		start = get_timer(0);
672 		while (1) {
673 			err = mmc_send_op_cond_iter(mmc, 1);
674 			if (err)
675 				return err;
676 			if (mmc->ocr & OCR_BUSY)
677 				break;
678 			if (get_timer(start) > timeout)
679 				return -EOPNOTSUPP;
680 			udelay(100);
681 		}
682 	}
683 
684 	if (mmc_host_is_spi(mmc)) { /* read OCR for spi */
685 		cmd.cmdidx = MMC_CMD_SPI_READ_OCR;
686 		cmd.resp_type = MMC_RSP_R3;
687 		cmd.cmdarg = 0;
688 
689 		err = mmc_send_cmd(mmc, &cmd, NULL);
690 
691 		if (err)
692 			return err;
693 
694 		mmc->ocr = cmd.response[0];
695 	}
696 
697 	mmc->version = MMC_VERSION_UNKNOWN;
698 
699 	mmc->high_capacity = ((mmc->ocr & OCR_HCS) == OCR_HCS);
700 	mmc->rca = 1;
701 
702 	return 0;
703 }
704 
705 
706 static int mmc_send_ext_csd(struct mmc *mmc, u8 *ext_csd)
707 {
708 	struct mmc_cmd cmd;
709 	struct mmc_data data;
710 	int err;
711 
712 	/* Get the Card Status Register */
713 	cmd.cmdidx = MMC_CMD_SEND_EXT_CSD;
714 	cmd.resp_type = MMC_RSP_R1;
715 	cmd.cmdarg = 0;
716 
717 	data.dest = (char *)ext_csd;
718 	data.blocks = 1;
719 	data.blocksize = MMC_MAX_BLOCK_LEN;
720 	data.flags = MMC_DATA_READ;
721 
722 	err = mmc_send_cmd(mmc, &cmd, &data);
723 
724 	return err;
725 }
726 
727 static int __mmc_switch(struct mmc *mmc, u8 set, u8 index, u8 value,
728 			bool send_status)
729 {
730 	struct mmc_cmd cmd;
731 	int timeout = 1000;
732 	int retries = 3;
733 	int ret;
734 
735 	cmd.cmdidx = MMC_CMD_SWITCH;
736 	cmd.resp_type = MMC_RSP_R1b;
737 	cmd.cmdarg = (MMC_SWITCH_MODE_WRITE_BYTE << 24) |
738 				 (index << 16) |
739 				 (value << 8);
740 
741 	while (retries > 0) {
742 		ret = mmc_send_cmd(mmc, &cmd, NULL);
743 
744 		if (ret) {
745 			retries--;
746 			continue;
747 		}
748 
749 		if (!send_status) {
750 			mdelay(50);
751 			return 0;
752 		}
753 
754 		/* Waiting for the ready status */
755 		return mmc_send_status(mmc, timeout);
756 	}
757 
758 	return ret;
759 
760 }
761 
762 int mmc_switch(struct mmc *mmc, u8 set, u8 index, u8 value)
763 {
764 	return __mmc_switch(mmc, set, index, value, true);
765 }
766 
767 #if !CONFIG_IS_ENABLED(MMC_TINY)
768 static int mmc_set_card_speed(struct mmc *mmc, enum bus_mode mode,
769 			      bool hsdowngrade)
770 {
771 	int err;
772 	int speed_bits;
773 
774 	ALLOC_CACHE_ALIGN_BUFFER(u8, test_csd, MMC_MAX_BLOCK_LEN);
775 
776 	switch (mode) {
777 	case MMC_HS:
778 	case MMC_HS_52:
779 	case MMC_DDR_52:
780 		speed_bits = EXT_CSD_TIMING_HS;
781 		break;
782 #if CONFIG_IS_ENABLED(MMC_HS200_SUPPORT)
783 	case MMC_HS_200:
784 		speed_bits = EXT_CSD_TIMING_HS200;
785 		break;
786 #endif
787 #if CONFIG_IS_ENABLED(MMC_HS400_SUPPORT)
788 	case MMC_HS_400:
789 		speed_bits = EXT_CSD_TIMING_HS400;
790 		break;
791 #endif
792 	case MMC_LEGACY:
793 		speed_bits = EXT_CSD_TIMING_LEGACY;
794 		break;
795 	default:
796 		return -EINVAL;
797 	}
798 
799 	err = __mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_HS_TIMING,
800 			   speed_bits, !hsdowngrade);
801 	if (err)
802 		return err;
803 
804 #if CONFIG_IS_ENABLED(MMC_HS200_SUPPORT) || \
805     CONFIG_IS_ENABLED(MMC_HS400_SUPPORT)
806 	/*
807 	 * In case the eMMC is in HS200/HS400 mode and we are downgrading
808 	 * to HS mode, the card clock are still running much faster than
809 	 * the supported HS mode clock, so we can not reliably read out
810 	 * Extended CSD. Reconfigure the controller to run at HS mode.
811 	 */
812 	if (hsdowngrade) {
813 		mmc_select_mode(mmc, MMC_HS);
814 		mmc_set_clock(mmc, mmc_mode2freq(mmc, MMC_HS), false);
815 	}
816 #endif
817 
818 	if ((mode == MMC_HS) || (mode == MMC_HS_52)) {
819 		/* Now check to see that it worked */
820 		err = mmc_send_ext_csd(mmc, test_csd);
821 		if (err)
822 			return err;
823 
824 		/* No high-speed support */
825 		if (!test_csd[EXT_CSD_HS_TIMING])
826 			return -ENOTSUPP;
827 	}
828 
829 	return 0;
830 }
831 
832 static int mmc_get_capabilities(struct mmc *mmc)
833 {
834 	u8 *ext_csd = mmc->ext_csd;
835 	char cardtype;
836 
837 	mmc->card_caps = MMC_MODE_1BIT | MMC_CAP(MMC_LEGACY);
838 
839 	if (mmc_host_is_spi(mmc))
840 		return 0;
841 
842 	/* Only version 4 supports high-speed */
843 	if (mmc->version < MMC_VERSION_4)
844 		return 0;
845 
846 	if (!ext_csd) {
847 		pr_err("No ext_csd found!\n"); /* this should enver happen */
848 		return -ENOTSUPP;
849 	}
850 
851 	mmc->card_caps |= MMC_MODE_4BIT | MMC_MODE_8BIT;
852 
853 	cardtype = ext_csd[EXT_CSD_CARD_TYPE];
854 	mmc->cardtype = cardtype;
855 
856 #if CONFIG_IS_ENABLED(MMC_HS200_SUPPORT)
857 	if (cardtype & (EXT_CSD_CARD_TYPE_HS200_1_2V |
858 			EXT_CSD_CARD_TYPE_HS200_1_8V)) {
859 		mmc->card_caps |= MMC_MODE_HS200;
860 	}
861 #endif
862 #if CONFIG_IS_ENABLED(MMC_HS400_SUPPORT)
863 	if (cardtype & (EXT_CSD_CARD_TYPE_HS400_1_2V |
864 			EXT_CSD_CARD_TYPE_HS400_1_8V)) {
865 		mmc->card_caps |= MMC_MODE_HS400;
866 	}
867 #endif
868 	if (cardtype & EXT_CSD_CARD_TYPE_52) {
869 		if (cardtype & EXT_CSD_CARD_TYPE_DDR_52)
870 			mmc->card_caps |= MMC_MODE_DDR_52MHz;
871 		mmc->card_caps |= MMC_MODE_HS_52MHz;
872 	}
873 	if (cardtype & EXT_CSD_CARD_TYPE_26)
874 		mmc->card_caps |= MMC_MODE_HS;
875 
876 	return 0;
877 }
878 #endif
879 
880 static int mmc_set_capacity(struct mmc *mmc, int part_num)
881 {
882 	switch (part_num) {
883 	case 0:
884 		mmc->capacity = mmc->capacity_user;
885 		break;
886 	case 1:
887 	case 2:
888 		mmc->capacity = mmc->capacity_boot;
889 		break;
890 	case 3:
891 		mmc->capacity = mmc->capacity_rpmb;
892 		break;
893 	case 4:
894 	case 5:
895 	case 6:
896 	case 7:
897 		mmc->capacity = mmc->capacity_gp[part_num - 4];
898 		break;
899 	default:
900 		return -1;
901 	}
902 
903 	mmc_get_blk_desc(mmc)->lba = lldiv(mmc->capacity, mmc->read_bl_len);
904 
905 	return 0;
906 }
907 
908 #if CONFIG_IS_ENABLED(MMC_HS200_SUPPORT)
909 static int mmc_boot_part_access_chk(struct mmc *mmc, unsigned int part_num)
910 {
911 	int forbidden = 0;
912 	bool change = false;
913 
914 	if (part_num & PART_ACCESS_MASK)
915 		forbidden = MMC_CAP(MMC_HS_200);
916 
917 	if (MMC_CAP(mmc->selected_mode) & forbidden) {
918 		pr_debug("selected mode (%s) is forbidden for part %d\n",
919 			 mmc_mode_name(mmc->selected_mode), part_num);
920 		change = true;
921 	} else if (mmc->selected_mode != mmc->best_mode) {
922 		pr_debug("selected mode is not optimal\n");
923 		change = true;
924 	}
925 
926 	if (change)
927 		return mmc_select_mode_and_width(mmc,
928 						 mmc->card_caps & ~forbidden);
929 
930 	return 0;
931 }
932 #else
933 static inline int mmc_boot_part_access_chk(struct mmc *mmc,
934 					   unsigned int part_num)
935 {
936 	return 0;
937 }
938 #endif
939 
940 int mmc_switch_part(struct mmc *mmc, unsigned int part_num)
941 {
942 	int ret;
943 
944 	ret = mmc_boot_part_access_chk(mmc, part_num);
945 	if (ret)
946 		return ret;
947 
948 	ret = mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_PART_CONF,
949 			 (mmc->part_config & ~PART_ACCESS_MASK)
950 			 | (part_num & PART_ACCESS_MASK));
951 
952 	/*
953 	 * Set the capacity if the switch succeeded or was intended
954 	 * to return to representing the raw device.
955 	 */
956 	if ((ret == 0) || ((ret == -ENODEV) && (part_num == 0))) {
957 		ret = mmc_set_capacity(mmc, part_num);
958 		mmc_get_blk_desc(mmc)->hwpart = part_num;
959 	}
960 
961 	return ret;
962 }
963 
964 #if CONFIG_IS_ENABLED(MMC_HW_PARTITIONING)
965 int mmc_hwpart_config(struct mmc *mmc,
966 		      const struct mmc_hwpart_conf *conf,
967 		      enum mmc_hwpart_conf_mode mode)
968 {
969 	u8 part_attrs = 0;
970 	u32 enh_size_mult;
971 	u32 enh_start_addr;
972 	u32 gp_size_mult[4];
973 	u32 max_enh_size_mult;
974 	u32 tot_enh_size_mult = 0;
975 	u8 wr_rel_set;
976 	int i, pidx, err;
977 	ALLOC_CACHE_ALIGN_BUFFER(u8, ext_csd, MMC_MAX_BLOCK_LEN);
978 
979 	if (mode < MMC_HWPART_CONF_CHECK || mode > MMC_HWPART_CONF_COMPLETE)
980 		return -EINVAL;
981 
982 	if (IS_SD(mmc) || (mmc->version < MMC_VERSION_4_41)) {
983 		pr_err("eMMC >= 4.4 required for enhanced user data area\n");
984 		return -EMEDIUMTYPE;
985 	}
986 
987 	if (!(mmc->part_support & PART_SUPPORT)) {
988 		pr_err("Card does not support partitioning\n");
989 		return -EMEDIUMTYPE;
990 	}
991 
992 	if (!mmc->hc_wp_grp_size) {
993 		pr_err("Card does not define HC WP group size\n");
994 		return -EMEDIUMTYPE;
995 	}
996 
997 	/* check partition alignment and total enhanced size */
998 	if (conf->user.enh_size) {
999 		if (conf->user.enh_size % mmc->hc_wp_grp_size ||
1000 		    conf->user.enh_start % mmc->hc_wp_grp_size) {
1001 			pr_err("User data enhanced area not HC WP group "
1002 			       "size aligned\n");
1003 			return -EINVAL;
1004 		}
1005 		part_attrs |= EXT_CSD_ENH_USR;
1006 		enh_size_mult = conf->user.enh_size / mmc->hc_wp_grp_size;
1007 		if (mmc->high_capacity) {
1008 			enh_start_addr = conf->user.enh_start;
1009 		} else {
1010 			enh_start_addr = (conf->user.enh_start << 9);
1011 		}
1012 	} else {
1013 		enh_size_mult = 0;
1014 		enh_start_addr = 0;
1015 	}
1016 	tot_enh_size_mult += enh_size_mult;
1017 
1018 	for (pidx = 0; pidx < 4; pidx++) {
1019 		if (conf->gp_part[pidx].size % mmc->hc_wp_grp_size) {
1020 			pr_err("GP%i partition not HC WP group size "
1021 			       "aligned\n", pidx+1);
1022 			return -EINVAL;
1023 		}
1024 		gp_size_mult[pidx] = conf->gp_part[pidx].size / mmc->hc_wp_grp_size;
1025 		if (conf->gp_part[pidx].size && conf->gp_part[pidx].enhanced) {
1026 			part_attrs |= EXT_CSD_ENH_GP(pidx);
1027 			tot_enh_size_mult += gp_size_mult[pidx];
1028 		}
1029 	}
1030 
1031 	if (part_attrs && ! (mmc->part_support & ENHNCD_SUPPORT)) {
1032 		pr_err("Card does not support enhanced attribute\n");
1033 		return -EMEDIUMTYPE;
1034 	}
1035 
1036 	err = mmc_send_ext_csd(mmc, ext_csd);
1037 	if (err)
1038 		return err;
1039 
1040 	max_enh_size_mult =
1041 		(ext_csd[EXT_CSD_MAX_ENH_SIZE_MULT+2] << 16) +
1042 		(ext_csd[EXT_CSD_MAX_ENH_SIZE_MULT+1] << 8) +
1043 		ext_csd[EXT_CSD_MAX_ENH_SIZE_MULT];
1044 	if (tot_enh_size_mult > max_enh_size_mult) {
1045 		pr_err("Total enhanced size exceeds maximum (%u > %u)\n",
1046 		       tot_enh_size_mult, max_enh_size_mult);
1047 		return -EMEDIUMTYPE;
1048 	}
1049 
1050 	/* The default value of EXT_CSD_WR_REL_SET is device
1051 	 * dependent, the values can only be changed if the
1052 	 * EXT_CSD_HS_CTRL_REL bit is set. The values can be
1053 	 * changed only once and before partitioning is completed. */
1054 	wr_rel_set = ext_csd[EXT_CSD_WR_REL_SET];
1055 	if (conf->user.wr_rel_change) {
1056 		if (conf->user.wr_rel_set)
1057 			wr_rel_set |= EXT_CSD_WR_DATA_REL_USR;
1058 		else
1059 			wr_rel_set &= ~EXT_CSD_WR_DATA_REL_USR;
1060 	}
1061 	for (pidx = 0; pidx < 4; pidx++) {
1062 		if (conf->gp_part[pidx].wr_rel_change) {
1063 			if (conf->gp_part[pidx].wr_rel_set)
1064 				wr_rel_set |= EXT_CSD_WR_DATA_REL_GP(pidx);
1065 			else
1066 				wr_rel_set &= ~EXT_CSD_WR_DATA_REL_GP(pidx);
1067 		}
1068 	}
1069 
1070 	if (wr_rel_set != ext_csd[EXT_CSD_WR_REL_SET] &&
1071 	    !(ext_csd[EXT_CSD_WR_REL_PARAM] & EXT_CSD_HS_CTRL_REL)) {
1072 		puts("Card does not support host controlled partition write "
1073 		     "reliability settings\n");
1074 		return -EMEDIUMTYPE;
1075 	}
1076 
1077 	if (ext_csd[EXT_CSD_PARTITION_SETTING] &
1078 	    EXT_CSD_PARTITION_SETTING_COMPLETED) {
1079 		pr_err("Card already partitioned\n");
1080 		return -EPERM;
1081 	}
1082 
1083 	if (mode == MMC_HWPART_CONF_CHECK)
1084 		return 0;
1085 
1086 	/* Partitioning requires high-capacity size definitions */
1087 	if (!(ext_csd[EXT_CSD_ERASE_GROUP_DEF] & 0x01)) {
1088 		err = mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL,
1089 				 EXT_CSD_ERASE_GROUP_DEF, 1);
1090 
1091 		if (err)
1092 			return err;
1093 
1094 		ext_csd[EXT_CSD_ERASE_GROUP_DEF] = 1;
1095 
1096 		/* update erase group size to be high-capacity */
1097 		mmc->erase_grp_size =
1098 			ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE] * 1024;
1099 
1100 	}
1101 
1102 	/* all OK, write the configuration */
1103 	for (i = 0; i < 4; i++) {
1104 		err = mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL,
1105 				 EXT_CSD_ENH_START_ADDR+i,
1106 				 (enh_start_addr >> (i*8)) & 0xFF);
1107 		if (err)
1108 			return err;
1109 	}
1110 	for (i = 0; i < 3; i++) {
1111 		err = mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL,
1112 				 EXT_CSD_ENH_SIZE_MULT+i,
1113 				 (enh_size_mult >> (i*8)) & 0xFF);
1114 		if (err)
1115 			return err;
1116 	}
1117 	for (pidx = 0; pidx < 4; pidx++) {
1118 		for (i = 0; i < 3; i++) {
1119 			err = mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL,
1120 					 EXT_CSD_GP_SIZE_MULT+pidx*3+i,
1121 					 (gp_size_mult[pidx] >> (i*8)) & 0xFF);
1122 			if (err)
1123 				return err;
1124 		}
1125 	}
1126 	err = mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL,
1127 			 EXT_CSD_PARTITIONS_ATTRIBUTE, part_attrs);
1128 	if (err)
1129 		return err;
1130 
1131 	if (mode == MMC_HWPART_CONF_SET)
1132 		return 0;
1133 
1134 	/* The WR_REL_SET is a write-once register but shall be
1135 	 * written before setting PART_SETTING_COMPLETED. As it is
1136 	 * write-once we can only write it when completing the
1137 	 * partitioning. */
1138 	if (wr_rel_set != ext_csd[EXT_CSD_WR_REL_SET]) {
1139 		err = mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL,
1140 				 EXT_CSD_WR_REL_SET, wr_rel_set);
1141 		if (err)
1142 			return err;
1143 	}
1144 
1145 	/* Setting PART_SETTING_COMPLETED confirms the partition
1146 	 * configuration but it only becomes effective after power
1147 	 * cycle, so we do not adjust the partition related settings
1148 	 * in the mmc struct. */
1149 
1150 	err = mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL,
1151 			 EXT_CSD_PARTITION_SETTING,
1152 			 EXT_CSD_PARTITION_SETTING_COMPLETED);
1153 	if (err)
1154 		return err;
1155 
1156 	return 0;
1157 }
1158 #endif
1159 
1160 #if !CONFIG_IS_ENABLED(DM_MMC)
1161 int mmc_getcd(struct mmc *mmc)
1162 {
1163 	int cd;
1164 
1165 	cd = board_mmc_getcd(mmc);
1166 
1167 	if (cd < 0) {
1168 		if (mmc->cfg->ops->getcd)
1169 			cd = mmc->cfg->ops->getcd(mmc);
1170 		else
1171 			cd = 1;
1172 	}
1173 
1174 	return cd;
1175 }
1176 #endif
1177 
1178 #if !CONFIG_IS_ENABLED(MMC_TINY)
1179 static int sd_switch(struct mmc *mmc, int mode, int group, u8 value, u8 *resp)
1180 {
1181 	struct mmc_cmd cmd;
1182 	struct mmc_data data;
1183 
1184 	/* Switch the frequency */
1185 	cmd.cmdidx = SD_CMD_SWITCH_FUNC;
1186 	cmd.resp_type = MMC_RSP_R1;
1187 	cmd.cmdarg = (mode << 31) | 0xffffff;
1188 	cmd.cmdarg &= ~(0xf << (group * 4));
1189 	cmd.cmdarg |= value << (group * 4);
1190 
1191 	data.dest = (char *)resp;
1192 	data.blocksize = 64;
1193 	data.blocks = 1;
1194 	data.flags = MMC_DATA_READ;
1195 
1196 	return mmc_send_cmd(mmc, &cmd, &data);
1197 }
1198 
1199 static int sd_get_capabilities(struct mmc *mmc)
1200 {
1201 	int err;
1202 	struct mmc_cmd cmd;
1203 	ALLOC_CACHE_ALIGN_BUFFER(__be32, scr, 2);
1204 	ALLOC_CACHE_ALIGN_BUFFER(__be32, switch_status, 16);
1205 	struct mmc_data data;
1206 	int timeout;
1207 #if CONFIG_IS_ENABLED(MMC_UHS_SUPPORT)
1208 	u32 sd3_bus_mode;
1209 #endif
1210 
1211 	mmc->card_caps = MMC_MODE_1BIT | MMC_CAP(SD_LEGACY);
1212 
1213 	if (mmc_host_is_spi(mmc))
1214 		return 0;
1215 
1216 	/* Read the SCR to find out if this card supports higher speeds */
1217 	cmd.cmdidx = MMC_CMD_APP_CMD;
1218 	cmd.resp_type = MMC_RSP_R1;
1219 	cmd.cmdarg = mmc->rca << 16;
1220 
1221 	err = mmc_send_cmd(mmc, &cmd, NULL);
1222 
1223 	if (err)
1224 		return err;
1225 
1226 	cmd.cmdidx = SD_CMD_APP_SEND_SCR;
1227 	cmd.resp_type = MMC_RSP_R1;
1228 	cmd.cmdarg = 0;
1229 
1230 	timeout = 3;
1231 
1232 retry_scr:
1233 	data.dest = (char *)scr;
1234 	data.blocksize = 8;
1235 	data.blocks = 1;
1236 	data.flags = MMC_DATA_READ;
1237 
1238 	err = mmc_send_cmd(mmc, &cmd, &data);
1239 
1240 	if (err) {
1241 		if (timeout--)
1242 			goto retry_scr;
1243 
1244 		return err;
1245 	}
1246 
1247 	mmc->scr[0] = __be32_to_cpu(scr[0]);
1248 	mmc->scr[1] = __be32_to_cpu(scr[1]);
1249 
1250 	switch ((mmc->scr[0] >> 24) & 0xf) {
1251 	case 0:
1252 		mmc->version = SD_VERSION_1_0;
1253 		break;
1254 	case 1:
1255 		mmc->version = SD_VERSION_1_10;
1256 		break;
1257 	case 2:
1258 		mmc->version = SD_VERSION_2;
1259 		if ((mmc->scr[0] >> 15) & 0x1)
1260 			mmc->version = SD_VERSION_3;
1261 		break;
1262 	default:
1263 		mmc->version = SD_VERSION_1_0;
1264 		break;
1265 	}
1266 
1267 	if (mmc->scr[0] & SD_DATA_4BIT)
1268 		mmc->card_caps |= MMC_MODE_4BIT;
1269 
1270 	/* Version 1.0 doesn't support switching */
1271 	if (mmc->version == SD_VERSION_1_0)
1272 		return 0;
1273 
1274 	timeout = 4;
1275 	while (timeout--) {
1276 		err = sd_switch(mmc, SD_SWITCH_CHECK, 0, 1,
1277 				(u8 *)switch_status);
1278 
1279 		if (err)
1280 			return err;
1281 
1282 		/* The high-speed function is busy.  Try again */
1283 		if (!(__be32_to_cpu(switch_status[7]) & SD_HIGHSPEED_BUSY))
1284 			break;
1285 	}
1286 
1287 	/* If high-speed isn't supported, we return */
1288 	if (__be32_to_cpu(switch_status[3]) & SD_HIGHSPEED_SUPPORTED)
1289 		mmc->card_caps |= MMC_CAP(SD_HS);
1290 
1291 #if CONFIG_IS_ENABLED(MMC_UHS_SUPPORT)
1292 	/* Version before 3.0 don't support UHS modes */
1293 	if (mmc->version < SD_VERSION_3)
1294 		return 0;
1295 
1296 	sd3_bus_mode = __be32_to_cpu(switch_status[3]) >> 16 & 0x1f;
1297 	if (sd3_bus_mode & SD_MODE_UHS_SDR104)
1298 		mmc->card_caps |= MMC_CAP(UHS_SDR104);
1299 	if (sd3_bus_mode & SD_MODE_UHS_SDR50)
1300 		mmc->card_caps |= MMC_CAP(UHS_SDR50);
1301 	if (sd3_bus_mode & SD_MODE_UHS_SDR25)
1302 		mmc->card_caps |= MMC_CAP(UHS_SDR25);
1303 	if (sd3_bus_mode & SD_MODE_UHS_SDR12)
1304 		mmc->card_caps |= MMC_CAP(UHS_SDR12);
1305 	if (sd3_bus_mode & SD_MODE_UHS_DDR50)
1306 		mmc->card_caps |= MMC_CAP(UHS_DDR50);
1307 #endif
1308 
1309 	return 0;
1310 }
1311 
1312 static int sd_set_card_speed(struct mmc *mmc, enum bus_mode mode)
1313 {
1314 	int err;
1315 
1316 	ALLOC_CACHE_ALIGN_BUFFER(uint, switch_status, 16);
1317 	int speed;
1318 
1319 	/* SD version 1.00 and 1.01 does not support CMD 6 */
1320 	if (mmc->version == SD_VERSION_1_0)
1321 		return 0;
1322 
1323 	switch (mode) {
1324 	case SD_LEGACY:
1325 		speed = UHS_SDR12_BUS_SPEED;
1326 		break;
1327 	case SD_HS:
1328 		speed = HIGH_SPEED_BUS_SPEED;
1329 		break;
1330 #if CONFIG_IS_ENABLED(MMC_UHS_SUPPORT)
1331 	case UHS_SDR12:
1332 		speed = UHS_SDR12_BUS_SPEED;
1333 		break;
1334 	case UHS_SDR25:
1335 		speed = UHS_SDR25_BUS_SPEED;
1336 		break;
1337 	case UHS_SDR50:
1338 		speed = UHS_SDR50_BUS_SPEED;
1339 		break;
1340 	case UHS_DDR50:
1341 		speed = UHS_DDR50_BUS_SPEED;
1342 		break;
1343 	case UHS_SDR104:
1344 		speed = UHS_SDR104_BUS_SPEED;
1345 		break;
1346 #endif
1347 	default:
1348 		return -EINVAL;
1349 	}
1350 
1351 	err = sd_switch(mmc, SD_SWITCH_SWITCH, 0, speed, (u8 *)switch_status);
1352 	if (err)
1353 		return err;
1354 
1355 	if (((__be32_to_cpu(switch_status[4]) >> 24) & 0xF) != speed)
1356 		return -ENOTSUPP;
1357 
1358 	return 0;
1359 }
1360 
1361 static int sd_select_bus_width(struct mmc *mmc, int w)
1362 {
1363 	int err;
1364 	struct mmc_cmd cmd;
1365 
1366 	if ((w != 4) && (w != 1))
1367 		return -EINVAL;
1368 
1369 	cmd.cmdidx = MMC_CMD_APP_CMD;
1370 	cmd.resp_type = MMC_RSP_R1;
1371 	cmd.cmdarg = mmc->rca << 16;
1372 
1373 	err = mmc_send_cmd(mmc, &cmd, NULL);
1374 	if (err)
1375 		return err;
1376 
1377 	cmd.cmdidx = SD_CMD_APP_SET_BUS_WIDTH;
1378 	cmd.resp_type = MMC_RSP_R1;
1379 	if (w == 4)
1380 		cmd.cmdarg = 2;
1381 	else if (w == 1)
1382 		cmd.cmdarg = 0;
1383 	err = mmc_send_cmd(mmc, &cmd, NULL);
1384 	if (err)
1385 		return err;
1386 
1387 	return 0;
1388 }
1389 #endif
1390 
1391 #if CONFIG_IS_ENABLED(MMC_WRITE)
1392 static int sd_read_ssr(struct mmc *mmc)
1393 {
1394 	static const unsigned int sd_au_size[] = {
1395 		0,		SZ_16K / 512,		SZ_32K / 512,
1396 		SZ_64K / 512,	SZ_128K / 512,		SZ_256K / 512,
1397 		SZ_512K / 512,	SZ_1M / 512,		SZ_2M / 512,
1398 		SZ_4M / 512,	SZ_8M / 512,		(SZ_8M + SZ_4M) / 512,
1399 		SZ_16M / 512,	(SZ_16M + SZ_8M) / 512,	SZ_32M / 512,
1400 		SZ_64M / 512,
1401 	};
1402 	int err, i;
1403 	struct mmc_cmd cmd;
1404 	ALLOC_CACHE_ALIGN_BUFFER(uint, ssr, 16);
1405 	struct mmc_data data;
1406 	int timeout = 3;
1407 	unsigned int au, eo, et, es;
1408 
1409 	cmd.cmdidx = MMC_CMD_APP_CMD;
1410 	cmd.resp_type = MMC_RSP_R1;
1411 	cmd.cmdarg = mmc->rca << 16;
1412 
1413 	err = mmc_send_cmd(mmc, &cmd, NULL);
1414 	if (err)
1415 		return err;
1416 
1417 	cmd.cmdidx = SD_CMD_APP_SD_STATUS;
1418 	cmd.resp_type = MMC_RSP_R1;
1419 	cmd.cmdarg = 0;
1420 
1421 retry_ssr:
1422 	data.dest = (char *)ssr;
1423 	data.blocksize = 64;
1424 	data.blocks = 1;
1425 	data.flags = MMC_DATA_READ;
1426 
1427 	err = mmc_send_cmd(mmc, &cmd, &data);
1428 	if (err) {
1429 		if (timeout--)
1430 			goto retry_ssr;
1431 
1432 		return err;
1433 	}
1434 
1435 	for (i = 0; i < 16; i++)
1436 		ssr[i] = be32_to_cpu(ssr[i]);
1437 
1438 	au = (ssr[2] >> 12) & 0xF;
1439 	if ((au <= 9) || (mmc->version == SD_VERSION_3)) {
1440 		mmc->ssr.au = sd_au_size[au];
1441 		es = (ssr[3] >> 24) & 0xFF;
1442 		es |= (ssr[2] & 0xFF) << 8;
1443 		et = (ssr[3] >> 18) & 0x3F;
1444 		if (es && et) {
1445 			eo = (ssr[3] >> 16) & 0x3;
1446 			mmc->ssr.erase_timeout = (et * 1000) / es;
1447 			mmc->ssr.erase_offset = eo * 1000;
1448 		}
1449 	} else {
1450 		pr_debug("Invalid Allocation Unit Size.\n");
1451 	}
1452 
1453 	return 0;
1454 }
1455 #endif
1456 /* frequency bases */
1457 /* divided by 10 to be nice to platforms without floating point */
1458 static const int fbase[] = {
1459 	10000,
1460 	100000,
1461 	1000000,
1462 	10000000,
1463 };
1464 
1465 /* Multiplier values for TRAN_SPEED.  Multiplied by 10 to be nice
1466  * to platforms without floating point.
1467  */
1468 static const u8 multipliers[] = {
1469 	0,	/* reserved */
1470 	10,
1471 	12,
1472 	13,
1473 	15,
1474 	20,
1475 	25,
1476 	30,
1477 	35,
1478 	40,
1479 	45,
1480 	50,
1481 	55,
1482 	60,
1483 	70,
1484 	80,
1485 };
1486 
1487 static inline int bus_width(uint cap)
1488 {
1489 	if (cap == MMC_MODE_8BIT)
1490 		return 8;
1491 	if (cap == MMC_MODE_4BIT)
1492 		return 4;
1493 	if (cap == MMC_MODE_1BIT)
1494 		return 1;
1495 	pr_warn("invalid bus witdh capability 0x%x\n", cap);
1496 	return 0;
1497 }
1498 
1499 #if !CONFIG_IS_ENABLED(DM_MMC)
1500 #ifdef MMC_SUPPORTS_TUNING
1501 static int mmc_execute_tuning(struct mmc *mmc, uint opcode)
1502 {
1503 	return -ENOTSUPP;
1504 }
1505 #endif
1506 
1507 static void mmc_send_init_stream(struct mmc *mmc)
1508 {
1509 }
1510 
1511 static int mmc_set_ios(struct mmc *mmc)
1512 {
1513 	int ret = 0;
1514 
1515 	if (mmc->cfg->ops->set_ios)
1516 		ret = mmc->cfg->ops->set_ios(mmc);
1517 
1518 	return ret;
1519 }
1520 #endif
1521 
1522 int mmc_set_clock(struct mmc *mmc, uint clock, bool disable)
1523 {
1524 	if (!disable) {
1525 		if (clock > mmc->cfg->f_max)
1526 			clock = mmc->cfg->f_max;
1527 
1528 		if (clock < mmc->cfg->f_min)
1529 			clock = mmc->cfg->f_min;
1530 	}
1531 
1532 	mmc->clock = clock;
1533 	mmc->clk_disable = disable;
1534 
1535 	debug("clock is %s (%dHz)\n", disable ? "disabled" : "enabled", clock);
1536 
1537 	return mmc_set_ios(mmc);
1538 }
1539 
1540 static int mmc_set_bus_width(struct mmc *mmc, uint width)
1541 {
1542 	mmc->bus_width = width;
1543 
1544 	return mmc_set_ios(mmc);
1545 }
1546 
1547 #if CONFIG_IS_ENABLED(MMC_VERBOSE) || defined(DEBUG)
1548 /*
1549  * helper function to display the capabilities in a human
1550  * friendly manner. The capabilities include bus width and
1551  * supported modes.
1552  */
1553 void mmc_dump_capabilities(const char *text, uint caps)
1554 {
1555 	enum bus_mode mode;
1556 
1557 	pr_debug("%s: widths [", text);
1558 	if (caps & MMC_MODE_8BIT)
1559 		pr_debug("8, ");
1560 	if (caps & MMC_MODE_4BIT)
1561 		pr_debug("4, ");
1562 	if (caps & MMC_MODE_1BIT)
1563 		pr_debug("1, ");
1564 	pr_debug("\b\b] modes [");
1565 	for (mode = MMC_LEGACY; mode < MMC_MODES_END; mode++)
1566 		if (MMC_CAP(mode) & caps)
1567 			pr_debug("%s, ", mmc_mode_name(mode));
1568 	pr_debug("\b\b]\n");
1569 }
1570 #endif
1571 
1572 struct mode_width_tuning {
1573 	enum bus_mode mode;
1574 	uint widths;
1575 #ifdef MMC_SUPPORTS_TUNING
1576 	uint tuning;
1577 #endif
1578 };
1579 
1580 #if CONFIG_IS_ENABLED(MMC_IO_VOLTAGE)
1581 int mmc_voltage_to_mv(enum mmc_voltage voltage)
1582 {
1583 	switch (voltage) {
1584 	case MMC_SIGNAL_VOLTAGE_000: return 0;
1585 	case MMC_SIGNAL_VOLTAGE_330: return 3300;
1586 	case MMC_SIGNAL_VOLTAGE_180: return 1800;
1587 	case MMC_SIGNAL_VOLTAGE_120: return 1200;
1588 	}
1589 	return -EINVAL;
1590 }
1591 
1592 static int mmc_set_signal_voltage(struct mmc *mmc, uint signal_voltage)
1593 {
1594 	int err;
1595 
1596 	if (mmc->signal_voltage == signal_voltage)
1597 		return 0;
1598 
1599 	mmc->signal_voltage = signal_voltage;
1600 	err = mmc_set_ios(mmc);
1601 	if (err)
1602 		pr_debug("unable to set voltage (err %d)\n", err);
1603 
1604 	return err;
1605 }
1606 #else
1607 static inline int mmc_set_signal_voltage(struct mmc *mmc, uint signal_voltage)
1608 {
1609 	return 0;
1610 }
1611 #endif
1612 
1613 #if !CONFIG_IS_ENABLED(MMC_TINY)
1614 static const struct mode_width_tuning sd_modes_by_pref[] = {
1615 #if CONFIG_IS_ENABLED(MMC_UHS_SUPPORT)
1616 #ifdef MMC_SUPPORTS_TUNING
1617 	{
1618 		.mode = UHS_SDR104,
1619 		.widths = MMC_MODE_4BIT | MMC_MODE_1BIT,
1620 		.tuning = MMC_CMD_SEND_TUNING_BLOCK
1621 	},
1622 #endif
1623 	{
1624 		.mode = UHS_SDR50,
1625 		.widths = MMC_MODE_4BIT | MMC_MODE_1BIT,
1626 	},
1627 	{
1628 		.mode = UHS_DDR50,
1629 		.widths = MMC_MODE_4BIT | MMC_MODE_1BIT,
1630 	},
1631 	{
1632 		.mode = UHS_SDR25,
1633 		.widths = MMC_MODE_4BIT | MMC_MODE_1BIT,
1634 	},
1635 #endif
1636 	{
1637 		.mode = SD_HS,
1638 		.widths = MMC_MODE_4BIT | MMC_MODE_1BIT,
1639 	},
1640 #if CONFIG_IS_ENABLED(MMC_UHS_SUPPORT)
1641 	{
1642 		.mode = UHS_SDR12,
1643 		.widths = MMC_MODE_4BIT | MMC_MODE_1BIT,
1644 	},
1645 #endif
1646 	{
1647 		.mode = SD_LEGACY,
1648 		.widths = MMC_MODE_4BIT | MMC_MODE_1BIT,
1649 	}
1650 };
1651 
1652 #define for_each_sd_mode_by_pref(caps, mwt) \
1653 	for (mwt = sd_modes_by_pref;\
1654 	     mwt < sd_modes_by_pref + ARRAY_SIZE(sd_modes_by_pref);\
1655 	     mwt++) \
1656 		if (caps & MMC_CAP(mwt->mode))
1657 
1658 static int sd_select_mode_and_width(struct mmc *mmc, uint card_caps)
1659 {
1660 	int err;
1661 	uint widths[] = {MMC_MODE_4BIT, MMC_MODE_1BIT};
1662 	const struct mode_width_tuning *mwt;
1663 #if CONFIG_IS_ENABLED(MMC_UHS_SUPPORT)
1664 	bool uhs_en = (mmc->ocr & OCR_S18R) ? true : false;
1665 #else
1666 	bool uhs_en = false;
1667 #endif
1668 	uint caps;
1669 
1670 #ifdef DEBUG
1671 	mmc_dump_capabilities("sd card", card_caps);
1672 	mmc_dump_capabilities("host", mmc->host_caps);
1673 #endif
1674 
1675 	/* Restrict card's capabilities by what the host can do */
1676 	caps = card_caps & mmc->host_caps;
1677 
1678 	if (!uhs_en)
1679 		caps &= ~UHS_CAPS;
1680 
1681 	for_each_sd_mode_by_pref(caps, mwt) {
1682 		uint *w;
1683 
1684 		for (w = widths; w < widths + ARRAY_SIZE(widths); w++) {
1685 			if (*w & caps & mwt->widths) {
1686 				pr_debug("trying mode %s width %d (at %d MHz)\n",
1687 					 mmc_mode_name(mwt->mode),
1688 					 bus_width(*w),
1689 					 mmc_mode2freq(mmc, mwt->mode) / 1000000);
1690 
1691 				/* configure the bus width (card + host) */
1692 				err = sd_select_bus_width(mmc, bus_width(*w));
1693 				if (err)
1694 					goto error;
1695 				mmc_set_bus_width(mmc, bus_width(*w));
1696 
1697 				/* configure the bus mode (card) */
1698 				err = sd_set_card_speed(mmc, mwt->mode);
1699 				if (err)
1700 					goto error;
1701 
1702 				/* configure the bus mode (host) */
1703 				mmc_select_mode(mmc, mwt->mode);
1704 				mmc_set_clock(mmc, mmc->tran_speed,
1705 						MMC_CLK_ENABLE);
1706 
1707 #ifdef MMC_SUPPORTS_TUNING
1708 				/* execute tuning if needed */
1709 				if (mwt->tuning && !mmc_host_is_spi(mmc)) {
1710 					err = mmc_execute_tuning(mmc,
1711 								 mwt->tuning);
1712 					if (err) {
1713 						pr_debug("tuning failed\n");
1714 						goto error;
1715 					}
1716 				}
1717 #endif
1718 
1719 #if CONFIG_IS_ENABLED(MMC_WRITE)
1720 				err = sd_read_ssr(mmc);
1721 				if (err)
1722 					pr_warn("unable to read ssr\n");
1723 #endif
1724 				if (!err)
1725 					return 0;
1726 
1727 error:
1728 				/* revert to a safer bus speed */
1729 				mmc_select_mode(mmc, SD_LEGACY);
1730 				mmc_set_clock(mmc, mmc->tran_speed,
1731 						MMC_CLK_ENABLE);
1732 			}
1733 		}
1734 	}
1735 
1736 	pr_err("unable to select a mode\n");
1737 	return -ENOTSUPP;
1738 }
1739 
1740 /*
1741  * read the compare the part of ext csd that is constant.
1742  * This can be used to check that the transfer is working
1743  * as expected.
1744  */
1745 static int mmc_read_and_compare_ext_csd(struct mmc *mmc)
1746 {
1747 	int err;
1748 	const u8 *ext_csd = mmc->ext_csd;
1749 	ALLOC_CACHE_ALIGN_BUFFER(u8, test_csd, MMC_MAX_BLOCK_LEN);
1750 
1751 	if (mmc->version < MMC_VERSION_4)
1752 		return 0;
1753 
1754 	err = mmc_send_ext_csd(mmc, test_csd);
1755 	if (err)
1756 		return err;
1757 
1758 	/* Only compare read only fields */
1759 	if (ext_csd[EXT_CSD_PARTITIONING_SUPPORT]
1760 		== test_csd[EXT_CSD_PARTITIONING_SUPPORT] &&
1761 	    ext_csd[EXT_CSD_HC_WP_GRP_SIZE]
1762 		== test_csd[EXT_CSD_HC_WP_GRP_SIZE] &&
1763 	    ext_csd[EXT_CSD_REV]
1764 		== test_csd[EXT_CSD_REV] &&
1765 	    ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE]
1766 		== test_csd[EXT_CSD_HC_ERASE_GRP_SIZE] &&
1767 	    memcmp(&ext_csd[EXT_CSD_SEC_CNT],
1768 		   &test_csd[EXT_CSD_SEC_CNT], 4) == 0)
1769 		return 0;
1770 
1771 	return -EBADMSG;
1772 }
1773 
1774 #if CONFIG_IS_ENABLED(MMC_IO_VOLTAGE)
1775 static int mmc_set_lowest_voltage(struct mmc *mmc, enum bus_mode mode,
1776 				  uint32_t allowed_mask)
1777 {
1778 	u32 card_mask = 0;
1779 
1780 	switch (mode) {
1781 	case MMC_HS_400:
1782 	case MMC_HS_200:
1783 		if (mmc->cardtype & (EXT_CSD_CARD_TYPE_HS200_1_8V |
1784 		    EXT_CSD_CARD_TYPE_HS400_1_8V))
1785 			card_mask |= MMC_SIGNAL_VOLTAGE_180;
1786 		if (mmc->cardtype & (EXT_CSD_CARD_TYPE_HS200_1_2V |
1787 		    EXT_CSD_CARD_TYPE_HS400_1_2V))
1788 			card_mask |= MMC_SIGNAL_VOLTAGE_120;
1789 		break;
1790 	case MMC_DDR_52:
1791 		if (mmc->cardtype & EXT_CSD_CARD_TYPE_DDR_1_8V)
1792 			card_mask |= MMC_SIGNAL_VOLTAGE_330 |
1793 				     MMC_SIGNAL_VOLTAGE_180;
1794 		if (mmc->cardtype & EXT_CSD_CARD_TYPE_DDR_1_2V)
1795 			card_mask |= MMC_SIGNAL_VOLTAGE_120;
1796 		break;
1797 	default:
1798 		card_mask |= MMC_SIGNAL_VOLTAGE_330;
1799 		break;
1800 	}
1801 
1802 	while (card_mask & allowed_mask) {
1803 		enum mmc_voltage best_match;
1804 
1805 		best_match = 1 << (ffs(card_mask & allowed_mask) - 1);
1806 		if (!mmc_set_signal_voltage(mmc,  best_match))
1807 			return 0;
1808 
1809 		allowed_mask &= ~best_match;
1810 	}
1811 
1812 	return -ENOTSUPP;
1813 }
1814 #else
1815 static inline int mmc_set_lowest_voltage(struct mmc *mmc, enum bus_mode mode,
1816 					 uint32_t allowed_mask)
1817 {
1818 	return 0;
1819 }
1820 #endif
1821 
1822 static const struct mode_width_tuning mmc_modes_by_pref[] = {
1823 #if CONFIG_IS_ENABLED(MMC_HS400_SUPPORT)
1824 	{
1825 		.mode = MMC_HS_400,
1826 		.widths = MMC_MODE_8BIT,
1827 		.tuning = MMC_CMD_SEND_TUNING_BLOCK_HS200
1828 	},
1829 #endif
1830 #if CONFIG_IS_ENABLED(MMC_HS200_SUPPORT)
1831 	{
1832 		.mode = MMC_HS_200,
1833 		.widths = MMC_MODE_8BIT | MMC_MODE_4BIT,
1834 		.tuning = MMC_CMD_SEND_TUNING_BLOCK_HS200
1835 	},
1836 #endif
1837 	{
1838 		.mode = MMC_DDR_52,
1839 		.widths = MMC_MODE_8BIT | MMC_MODE_4BIT,
1840 	},
1841 	{
1842 		.mode = MMC_HS_52,
1843 		.widths = MMC_MODE_8BIT | MMC_MODE_4BIT | MMC_MODE_1BIT,
1844 	},
1845 	{
1846 		.mode = MMC_HS,
1847 		.widths = MMC_MODE_8BIT | MMC_MODE_4BIT | MMC_MODE_1BIT,
1848 	},
1849 	{
1850 		.mode = MMC_LEGACY,
1851 		.widths = MMC_MODE_8BIT | MMC_MODE_4BIT | MMC_MODE_1BIT,
1852 	}
1853 };
1854 
1855 #define for_each_mmc_mode_by_pref(caps, mwt) \
1856 	for (mwt = mmc_modes_by_pref;\
1857 	    mwt < mmc_modes_by_pref + ARRAY_SIZE(mmc_modes_by_pref);\
1858 	    mwt++) \
1859 		if (caps & MMC_CAP(mwt->mode))
1860 
1861 static const struct ext_csd_bus_width {
1862 	uint cap;
1863 	bool is_ddr;
1864 	uint ext_csd_bits;
1865 } ext_csd_bus_width[] = {
1866 	{MMC_MODE_8BIT, true, EXT_CSD_DDR_BUS_WIDTH_8},
1867 	{MMC_MODE_4BIT, true, EXT_CSD_DDR_BUS_WIDTH_4},
1868 	{MMC_MODE_8BIT, false, EXT_CSD_BUS_WIDTH_8},
1869 	{MMC_MODE_4BIT, false, EXT_CSD_BUS_WIDTH_4},
1870 	{MMC_MODE_1BIT, false, EXT_CSD_BUS_WIDTH_1},
1871 };
1872 
1873 #if CONFIG_IS_ENABLED(MMC_HS400_SUPPORT)
1874 static int mmc_select_hs400(struct mmc *mmc)
1875 {
1876 	int err;
1877 
1878 	/* Set timing to HS200 for tuning */
1879 	err = mmc_set_card_speed(mmc, MMC_HS_200, false);
1880 	if (err)
1881 		return err;
1882 
1883 	/* configure the bus mode (host) */
1884 	mmc_select_mode(mmc, MMC_HS_200);
1885 	mmc_set_clock(mmc, mmc->tran_speed, false);
1886 
1887 	/* execute tuning if needed */
1888 	err = mmc_execute_tuning(mmc, MMC_CMD_SEND_TUNING_BLOCK_HS200);
1889 	if (err) {
1890 		debug("tuning failed\n");
1891 		return err;
1892 	}
1893 
1894 	/* Set back to HS */
1895 	mmc_set_card_speed(mmc, MMC_HS, true);
1896 
1897 	err = mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_BUS_WIDTH,
1898 			 EXT_CSD_BUS_WIDTH_8 | EXT_CSD_DDR_FLAG);
1899 	if (err)
1900 		return err;
1901 
1902 	err = mmc_set_card_speed(mmc, MMC_HS_400, false);
1903 	if (err)
1904 		return err;
1905 
1906 	mmc_select_mode(mmc, MMC_HS_400);
1907 	err = mmc_set_clock(mmc, mmc->tran_speed, false);
1908 	if (err)
1909 		return err;
1910 
1911 	return 0;
1912 }
1913 #else
1914 static int mmc_select_hs400(struct mmc *mmc)
1915 {
1916 	return -ENOTSUPP;
1917 }
1918 #endif
1919 
1920 #define for_each_supported_width(caps, ddr, ecbv) \
1921 	for (ecbv = ext_csd_bus_width;\
1922 	    ecbv < ext_csd_bus_width + ARRAY_SIZE(ext_csd_bus_width);\
1923 	    ecbv++) \
1924 		if ((ddr == ecbv->is_ddr) && (caps & ecbv->cap))
1925 
1926 static int mmc_select_mode_and_width(struct mmc *mmc, uint card_caps)
1927 {
1928 	int err;
1929 	const struct mode_width_tuning *mwt;
1930 	const struct ext_csd_bus_width *ecbw;
1931 
1932 #ifdef DEBUG
1933 	mmc_dump_capabilities("mmc", card_caps);
1934 	mmc_dump_capabilities("host", mmc->host_caps);
1935 #endif
1936 
1937 	/* Restrict card's capabilities by what the host can do */
1938 	card_caps &= mmc->host_caps;
1939 
1940 	/* Only version 4 of MMC supports wider bus widths */
1941 	if (mmc->version < MMC_VERSION_4)
1942 		return 0;
1943 
1944 	if (!mmc->ext_csd) {
1945 		pr_debug("No ext_csd found!\n"); /* this should enver happen */
1946 		return -ENOTSUPP;
1947 	}
1948 
1949 #if CONFIG_IS_ENABLED(MMC_HS200_SUPPORT) || \
1950     CONFIG_IS_ENABLED(MMC_HS400_SUPPORT)
1951 	/*
1952 	 * In case the eMMC is in HS200/HS400 mode, downgrade to HS mode
1953 	 * before doing anything else, since a transition from either of
1954 	 * the HS200/HS400 mode directly to legacy mode is not supported.
1955 	 */
1956 	if (mmc->selected_mode == MMC_HS_200 ||
1957 	    mmc->selected_mode == MMC_HS_400)
1958 		mmc_set_card_speed(mmc, MMC_HS, true);
1959 	else
1960 #endif
1961 		mmc_set_clock(mmc, mmc->legacy_speed, MMC_CLK_ENABLE);
1962 
1963 	for_each_mmc_mode_by_pref(card_caps, mwt) {
1964 		for_each_supported_width(card_caps & mwt->widths,
1965 					 mmc_is_mode_ddr(mwt->mode), ecbw) {
1966 			enum mmc_voltage old_voltage;
1967 			pr_debug("trying mode %s width %d (at %d MHz)\n",
1968 				 mmc_mode_name(mwt->mode),
1969 				 bus_width(ecbw->cap),
1970 				 mmc_mode2freq(mmc, mwt->mode) / 1000000);
1971 			old_voltage = mmc->signal_voltage;
1972 			err = mmc_set_lowest_voltage(mmc, mwt->mode,
1973 						     MMC_ALL_SIGNAL_VOLTAGE);
1974 			if (err)
1975 				continue;
1976 
1977 			/* configure the bus width (card + host) */
1978 			err = mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL,
1979 				    EXT_CSD_BUS_WIDTH,
1980 				    ecbw->ext_csd_bits & ~EXT_CSD_DDR_FLAG);
1981 			if (err)
1982 				goto error;
1983 			mmc_set_bus_width(mmc, bus_width(ecbw->cap));
1984 
1985 			if (mwt->mode == MMC_HS_400) {
1986 				err = mmc_select_hs400(mmc);
1987 				if (err) {
1988 					printf("Select HS400 failed %d\n", err);
1989 					goto error;
1990 				}
1991 			} else {
1992 				/* configure the bus speed (card) */
1993 				err = mmc_set_card_speed(mmc, mwt->mode, false);
1994 				if (err)
1995 					goto error;
1996 
1997 				/*
1998 				 * configure the bus width AND the ddr mode
1999 				 * (card). The host side will be taken care
2000 				 * of in the next step
2001 				 */
2002 				if (ecbw->ext_csd_bits & EXT_CSD_DDR_FLAG) {
2003 					err = mmc_switch(mmc,
2004 							 EXT_CSD_CMD_SET_NORMAL,
2005 							 EXT_CSD_BUS_WIDTH,
2006 							 ecbw->ext_csd_bits);
2007 					if (err)
2008 						goto error;
2009 				}
2010 
2011 				/* configure the bus mode (host) */
2012 				mmc_select_mode(mmc, mwt->mode);
2013 				mmc_set_clock(mmc, mmc->tran_speed,
2014 					      MMC_CLK_ENABLE);
2015 #ifdef MMC_SUPPORTS_TUNING
2016 
2017 				/* execute tuning if needed */
2018 				if (mwt->tuning) {
2019 					err = mmc_execute_tuning(mmc,
2020 								 mwt->tuning);
2021 					if (err) {
2022 						pr_debug("tuning failed\n");
2023 						goto error;
2024 					}
2025 				}
2026 #endif
2027 			}
2028 
2029 			/* do a transfer to check the configuration */
2030 			err = mmc_read_and_compare_ext_csd(mmc);
2031 			if (!err)
2032 				return 0;
2033 error:
2034 			mmc_set_signal_voltage(mmc, old_voltage);
2035 			/* if an error occured, revert to a safer bus mode */
2036 			mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL,
2037 				   EXT_CSD_BUS_WIDTH, EXT_CSD_BUS_WIDTH_1);
2038 			mmc_select_mode(mmc, MMC_LEGACY);
2039 			mmc_set_bus_width(mmc, 1);
2040 		}
2041 	}
2042 
2043 	pr_err("unable to select a mode\n");
2044 
2045 	return -ENOTSUPP;
2046 }
2047 #endif
2048 
2049 #if CONFIG_IS_ENABLED(MMC_TINY)
2050 DEFINE_CACHE_ALIGN_BUFFER(u8, ext_csd_bkup, MMC_MAX_BLOCK_LEN);
2051 #endif
2052 
2053 static int mmc_startup_v4(struct mmc *mmc)
2054 {
2055 	int err, i;
2056 	u64 capacity;
2057 	bool has_parts = false;
2058 	bool part_completed;
2059 	static const u32 mmc_versions[] = {
2060 		MMC_VERSION_4,
2061 		MMC_VERSION_4_1,
2062 		MMC_VERSION_4_2,
2063 		MMC_VERSION_4_3,
2064 		MMC_VERSION_4_4,
2065 		MMC_VERSION_4_41,
2066 		MMC_VERSION_4_5,
2067 		MMC_VERSION_5_0,
2068 		MMC_VERSION_5_1
2069 	};
2070 
2071 #if CONFIG_IS_ENABLED(MMC_TINY)
2072 	u8 *ext_csd = ext_csd_bkup;
2073 
2074 	if (IS_SD(mmc) || mmc->version < MMC_VERSION_4)
2075 		return 0;
2076 
2077 	if (!mmc->ext_csd)
2078 		memset(ext_csd_bkup, 0, sizeof(ext_csd_bkup));
2079 
2080 	err = mmc_send_ext_csd(mmc, ext_csd);
2081 	if (err)
2082 		goto error;
2083 
2084 	/* store the ext csd for future reference */
2085 	if (!mmc->ext_csd)
2086 		mmc->ext_csd = ext_csd;
2087 #else
2088 	ALLOC_CACHE_ALIGN_BUFFER(u8, ext_csd, MMC_MAX_BLOCK_LEN);
2089 
2090 	if (IS_SD(mmc) || (mmc->version < MMC_VERSION_4))
2091 		return 0;
2092 
2093 	/* check  ext_csd version and capacity */
2094 	err = mmc_send_ext_csd(mmc, ext_csd);
2095 	if (err)
2096 		goto error;
2097 
2098 	/* store the ext csd for future reference */
2099 	if (!mmc->ext_csd)
2100 		mmc->ext_csd = malloc(MMC_MAX_BLOCK_LEN);
2101 	if (!mmc->ext_csd)
2102 		return -ENOMEM;
2103 	memcpy(mmc->ext_csd, ext_csd, MMC_MAX_BLOCK_LEN);
2104 #endif
2105 	if (ext_csd[EXT_CSD_REV] >= ARRAY_SIZE(mmc_versions))
2106 		return -EINVAL;
2107 
2108 	mmc->version = mmc_versions[ext_csd[EXT_CSD_REV]];
2109 
2110 	if (mmc->version >= MMC_VERSION_4_2) {
2111 		/*
2112 		 * According to the JEDEC Standard, the value of
2113 		 * ext_csd's capacity is valid if the value is more
2114 		 * than 2GB
2115 		 */
2116 		capacity = ext_csd[EXT_CSD_SEC_CNT] << 0
2117 				| ext_csd[EXT_CSD_SEC_CNT + 1] << 8
2118 				| ext_csd[EXT_CSD_SEC_CNT + 2] << 16
2119 				| ext_csd[EXT_CSD_SEC_CNT + 3] << 24;
2120 		capacity *= MMC_MAX_BLOCK_LEN;
2121 		if ((capacity >> 20) > 2 * 1024)
2122 			mmc->capacity_user = capacity;
2123 	}
2124 
2125 	/* The partition data may be non-zero but it is only
2126 	 * effective if PARTITION_SETTING_COMPLETED is set in
2127 	 * EXT_CSD, so ignore any data if this bit is not set,
2128 	 * except for enabling the high-capacity group size
2129 	 * definition (see below).
2130 	 */
2131 	part_completed = !!(ext_csd[EXT_CSD_PARTITION_SETTING] &
2132 			    EXT_CSD_PARTITION_SETTING_COMPLETED);
2133 
2134 	/* store the partition info of emmc */
2135 	mmc->part_support = ext_csd[EXT_CSD_PARTITIONING_SUPPORT];
2136 	if ((ext_csd[EXT_CSD_PARTITIONING_SUPPORT] & PART_SUPPORT) ||
2137 	    ext_csd[EXT_CSD_BOOT_MULT])
2138 		mmc->part_config = ext_csd[EXT_CSD_PART_CONF];
2139 	if (part_completed &&
2140 	    (ext_csd[EXT_CSD_PARTITIONING_SUPPORT] & ENHNCD_SUPPORT))
2141 		mmc->part_attr = ext_csd[EXT_CSD_PARTITIONS_ATTRIBUTE];
2142 
2143 	mmc->capacity_boot = ext_csd[EXT_CSD_BOOT_MULT] << 17;
2144 
2145 	mmc->capacity_rpmb = ext_csd[EXT_CSD_RPMB_MULT] << 17;
2146 
2147 	for (i = 0; i < 4; i++) {
2148 		int idx = EXT_CSD_GP_SIZE_MULT + i * 3;
2149 		uint mult = (ext_csd[idx + 2] << 16) +
2150 			(ext_csd[idx + 1] << 8) + ext_csd[idx];
2151 		if (mult)
2152 			has_parts = true;
2153 		if (!part_completed)
2154 			continue;
2155 		mmc->capacity_gp[i] = mult;
2156 		mmc->capacity_gp[i] *=
2157 			ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE];
2158 		mmc->capacity_gp[i] *= ext_csd[EXT_CSD_HC_WP_GRP_SIZE];
2159 		mmc->capacity_gp[i] <<= 19;
2160 	}
2161 
2162 #ifndef CONFIG_SPL_BUILD
2163 	if (part_completed) {
2164 		mmc->enh_user_size =
2165 			(ext_csd[EXT_CSD_ENH_SIZE_MULT + 2] << 16) +
2166 			(ext_csd[EXT_CSD_ENH_SIZE_MULT + 1] << 8) +
2167 			ext_csd[EXT_CSD_ENH_SIZE_MULT];
2168 		mmc->enh_user_size *= ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE];
2169 		mmc->enh_user_size *= ext_csd[EXT_CSD_HC_WP_GRP_SIZE];
2170 		mmc->enh_user_size <<= 19;
2171 		mmc->enh_user_start =
2172 			(ext_csd[EXT_CSD_ENH_START_ADDR + 3] << 24) +
2173 			(ext_csd[EXT_CSD_ENH_START_ADDR + 2] << 16) +
2174 			(ext_csd[EXT_CSD_ENH_START_ADDR + 1] << 8) +
2175 			ext_csd[EXT_CSD_ENH_START_ADDR];
2176 		if (mmc->high_capacity)
2177 			mmc->enh_user_start <<= 9;
2178 	}
2179 #endif
2180 
2181 	/*
2182 	 * Host needs to enable ERASE_GRP_DEF bit if device is
2183 	 * partitioned. This bit will be lost every time after a reset
2184 	 * or power off. This will affect erase size.
2185 	 */
2186 	if (part_completed)
2187 		has_parts = true;
2188 	if ((ext_csd[EXT_CSD_PARTITIONING_SUPPORT] & PART_SUPPORT) &&
2189 	    (ext_csd[EXT_CSD_PARTITIONS_ATTRIBUTE] & PART_ENH_ATTRIB))
2190 		has_parts = true;
2191 	if (has_parts) {
2192 		err = mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL,
2193 				 EXT_CSD_ERASE_GROUP_DEF, 1);
2194 
2195 		if (err)
2196 			goto error;
2197 
2198 		ext_csd[EXT_CSD_ERASE_GROUP_DEF] = 1;
2199 	}
2200 
2201 	if (ext_csd[EXT_CSD_ERASE_GROUP_DEF] & 0x01) {
2202 #if CONFIG_IS_ENABLED(MMC_WRITE)
2203 		/* Read out group size from ext_csd */
2204 		mmc->erase_grp_size =
2205 			ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE] * 1024;
2206 #endif
2207 		/*
2208 		 * if high capacity and partition setting completed
2209 		 * SEC_COUNT is valid even if it is smaller than 2 GiB
2210 		 * JEDEC Standard JESD84-B45, 6.2.4
2211 		 */
2212 		if (mmc->high_capacity && part_completed) {
2213 			capacity = (ext_csd[EXT_CSD_SEC_CNT]) |
2214 				(ext_csd[EXT_CSD_SEC_CNT + 1] << 8) |
2215 				(ext_csd[EXT_CSD_SEC_CNT + 2] << 16) |
2216 				(ext_csd[EXT_CSD_SEC_CNT + 3] << 24);
2217 			capacity *= MMC_MAX_BLOCK_LEN;
2218 			mmc->capacity_user = capacity;
2219 		}
2220 	}
2221 #if CONFIG_IS_ENABLED(MMC_WRITE)
2222 	else {
2223 		/* Calculate the group size from the csd value. */
2224 		int erase_gsz, erase_gmul;
2225 
2226 		erase_gsz = (mmc->csd[2] & 0x00007c00) >> 10;
2227 		erase_gmul = (mmc->csd[2] & 0x000003e0) >> 5;
2228 		mmc->erase_grp_size = (erase_gsz + 1)
2229 			* (erase_gmul + 1);
2230 	}
2231 #endif
2232 #if CONFIG_IS_ENABLED(MMC_HW_PARTITIONING)
2233 	mmc->hc_wp_grp_size = 1024
2234 		* ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE]
2235 		* ext_csd[EXT_CSD_HC_WP_GRP_SIZE];
2236 #endif
2237 
2238 	mmc->wr_rel_set = ext_csd[EXT_CSD_WR_REL_SET];
2239 
2240 	return 0;
2241 error:
2242 	if (mmc->ext_csd) {
2243 #if !CONFIG_IS_ENABLED(MMC_TINY)
2244 		free(mmc->ext_csd);
2245 #endif
2246 		mmc->ext_csd = NULL;
2247 	}
2248 	return err;
2249 }
2250 
2251 static int mmc_startup(struct mmc *mmc)
2252 {
2253 	int err, i;
2254 	uint mult, freq;
2255 	u64 cmult, csize;
2256 	struct mmc_cmd cmd;
2257 	struct blk_desc *bdesc;
2258 
2259 #ifdef CONFIG_MMC_SPI_CRC_ON
2260 	if (mmc_host_is_spi(mmc)) { /* enable CRC check for spi */
2261 		cmd.cmdidx = MMC_CMD_SPI_CRC_ON_OFF;
2262 		cmd.resp_type = MMC_RSP_R1;
2263 		cmd.cmdarg = 1;
2264 		err = mmc_send_cmd(mmc, &cmd, NULL);
2265 		if (err)
2266 			return err;
2267 	}
2268 #endif
2269 
2270 	/* Put the Card in Identify Mode */
2271 	cmd.cmdidx = mmc_host_is_spi(mmc) ? MMC_CMD_SEND_CID :
2272 		MMC_CMD_ALL_SEND_CID; /* cmd not supported in spi */
2273 	cmd.resp_type = MMC_RSP_R2;
2274 	cmd.cmdarg = 0;
2275 
2276 	err = mmc_send_cmd(mmc, &cmd, NULL);
2277 
2278 #ifdef CONFIG_MMC_QUIRKS
2279 	if (err && (mmc->quirks & MMC_QUIRK_RETRY_SEND_CID)) {
2280 		int retries = 4;
2281 		/*
2282 		 * It has been seen that SEND_CID may fail on the first
2283 		 * attempt, let's try a few more time
2284 		 */
2285 		do {
2286 			err = mmc_send_cmd(mmc, &cmd, NULL);
2287 			if (!err)
2288 				break;
2289 		} while (retries--);
2290 	}
2291 #endif
2292 
2293 	if (err)
2294 		return err;
2295 
2296 	memcpy(mmc->cid, cmd.response, 16);
2297 
2298 	/*
2299 	 * For MMC cards, set the Relative Address.
2300 	 * For SD cards, get the Relatvie Address.
2301 	 * This also puts the cards into Standby State
2302 	 */
2303 	if (!mmc_host_is_spi(mmc)) { /* cmd not supported in spi */
2304 		cmd.cmdidx = SD_CMD_SEND_RELATIVE_ADDR;
2305 		cmd.cmdarg = mmc->rca << 16;
2306 		cmd.resp_type = MMC_RSP_R6;
2307 
2308 		err = mmc_send_cmd(mmc, &cmd, NULL);
2309 
2310 		if (err)
2311 			return err;
2312 
2313 		if (IS_SD(mmc))
2314 			mmc->rca = (cmd.response[0] >> 16) & 0xffff;
2315 	}
2316 
2317 	/* Get the Card-Specific Data */
2318 	cmd.cmdidx = MMC_CMD_SEND_CSD;
2319 	cmd.resp_type = MMC_RSP_R2;
2320 	cmd.cmdarg = mmc->rca << 16;
2321 
2322 	err = mmc_send_cmd(mmc, &cmd, NULL);
2323 
2324 	if (err)
2325 		return err;
2326 
2327 	mmc->csd[0] = cmd.response[0];
2328 	mmc->csd[1] = cmd.response[1];
2329 	mmc->csd[2] = cmd.response[2];
2330 	mmc->csd[3] = cmd.response[3];
2331 
2332 	if (mmc->version == MMC_VERSION_UNKNOWN) {
2333 		int version = (cmd.response[0] >> 26) & 0xf;
2334 
2335 		switch (version) {
2336 		case 0:
2337 			mmc->version = MMC_VERSION_1_2;
2338 			break;
2339 		case 1:
2340 			mmc->version = MMC_VERSION_1_4;
2341 			break;
2342 		case 2:
2343 			mmc->version = MMC_VERSION_2_2;
2344 			break;
2345 		case 3:
2346 			mmc->version = MMC_VERSION_3;
2347 			break;
2348 		case 4:
2349 			mmc->version = MMC_VERSION_4;
2350 			break;
2351 		default:
2352 			mmc->version = MMC_VERSION_1_2;
2353 			break;
2354 		}
2355 	}
2356 
2357 	/* divide frequency by 10, since the mults are 10x bigger */
2358 	freq = fbase[(cmd.response[0] & 0x7)];
2359 	mult = multipliers[((cmd.response[0] >> 3) & 0xf)];
2360 
2361 	mmc->legacy_speed = freq * mult;
2362 	mmc_select_mode(mmc, MMC_LEGACY);
2363 
2364 	mmc->dsr_imp = ((cmd.response[1] >> 12) & 0x1);
2365 	mmc->read_bl_len = 1 << ((cmd.response[1] >> 16) & 0xf);
2366 #if CONFIG_IS_ENABLED(MMC_WRITE)
2367 
2368 	if (IS_SD(mmc))
2369 		mmc->write_bl_len = mmc->read_bl_len;
2370 	else
2371 		mmc->write_bl_len = 1 << ((cmd.response[3] >> 22) & 0xf);
2372 #endif
2373 
2374 	if (mmc->high_capacity) {
2375 		csize = (mmc->csd[1] & 0x3f) << 16
2376 			| (mmc->csd[2] & 0xffff0000) >> 16;
2377 		cmult = 8;
2378 	} else {
2379 		csize = (mmc->csd[1] & 0x3ff) << 2
2380 			| (mmc->csd[2] & 0xc0000000) >> 30;
2381 		cmult = (mmc->csd[2] & 0x00038000) >> 15;
2382 	}
2383 
2384 	mmc->capacity_user = (csize + 1) << (cmult + 2);
2385 	mmc->capacity_user *= mmc->read_bl_len;
2386 	mmc->capacity_boot = 0;
2387 	mmc->capacity_rpmb = 0;
2388 	for (i = 0; i < 4; i++)
2389 		mmc->capacity_gp[i] = 0;
2390 
2391 	if (mmc->read_bl_len > MMC_MAX_BLOCK_LEN)
2392 		mmc->read_bl_len = MMC_MAX_BLOCK_LEN;
2393 
2394 #if CONFIG_IS_ENABLED(MMC_WRITE)
2395 	if (mmc->write_bl_len > MMC_MAX_BLOCK_LEN)
2396 		mmc->write_bl_len = MMC_MAX_BLOCK_LEN;
2397 #endif
2398 
2399 	if ((mmc->dsr_imp) && (0xffffffff != mmc->dsr)) {
2400 		cmd.cmdidx = MMC_CMD_SET_DSR;
2401 		cmd.cmdarg = (mmc->dsr & 0xffff) << 16;
2402 		cmd.resp_type = MMC_RSP_NONE;
2403 		if (mmc_send_cmd(mmc, &cmd, NULL))
2404 			pr_warn("MMC: SET_DSR failed\n");
2405 	}
2406 
2407 	/* Select the card, and put it into Transfer Mode */
2408 	if (!mmc_host_is_spi(mmc)) { /* cmd not supported in spi */
2409 		cmd.cmdidx = MMC_CMD_SELECT_CARD;
2410 		cmd.resp_type = MMC_RSP_R1;
2411 		cmd.cmdarg = mmc->rca << 16;
2412 		err = mmc_send_cmd(mmc, &cmd, NULL);
2413 
2414 		if (err)
2415 			return err;
2416 	}
2417 
2418 	/*
2419 	 * For SD, its erase group is always one sector
2420 	 */
2421 #if CONFIG_IS_ENABLED(MMC_WRITE)
2422 	mmc->erase_grp_size = 1;
2423 #endif
2424 	mmc->part_config = MMCPART_NOAVAILABLE;
2425 
2426 	err = mmc_startup_v4(mmc);
2427 	if (err)
2428 		return err;
2429 
2430 	err = mmc_set_capacity(mmc, mmc_get_blk_desc(mmc)->hwpart);
2431 	if (err)
2432 		return err;
2433 
2434 #if CONFIG_IS_ENABLED(MMC_TINY)
2435 	mmc_set_clock(mmc, mmc->legacy_speed, false);
2436 	mmc_select_mode(mmc, IS_SD(mmc) ? SD_LEGACY : MMC_LEGACY);
2437 	mmc_set_bus_width(mmc, 1);
2438 #else
2439 	if (IS_SD(mmc)) {
2440 		err = sd_get_capabilities(mmc);
2441 		if (err)
2442 			return err;
2443 		err = sd_select_mode_and_width(mmc, mmc->card_caps);
2444 	} else {
2445 		err = mmc_get_capabilities(mmc);
2446 		if (err)
2447 			return err;
2448 		mmc_select_mode_and_width(mmc, mmc->card_caps);
2449 	}
2450 #endif
2451 	if (err)
2452 		return err;
2453 
2454 	mmc->best_mode = mmc->selected_mode;
2455 
2456 	/* Fix the block length for DDR mode */
2457 	if (mmc->ddr_mode) {
2458 		mmc->read_bl_len = MMC_MAX_BLOCK_LEN;
2459 #if CONFIG_IS_ENABLED(MMC_WRITE)
2460 		mmc->write_bl_len = MMC_MAX_BLOCK_LEN;
2461 #endif
2462 	}
2463 
2464 	/* fill in device description */
2465 	bdesc = mmc_get_blk_desc(mmc);
2466 	bdesc->lun = 0;
2467 	bdesc->hwpart = 0;
2468 	bdesc->type = 0;
2469 	bdesc->blksz = mmc->read_bl_len;
2470 	bdesc->log2blksz = LOG2(bdesc->blksz);
2471 	bdesc->lba = lldiv(mmc->capacity, mmc->read_bl_len);
2472 #if !defined(CONFIG_SPL_BUILD) || \
2473 		(defined(CONFIG_SPL_LIBCOMMON_SUPPORT) && \
2474 		!defined(CONFIG_USE_TINY_PRINTF))
2475 	sprintf(bdesc->vendor, "Man %06x Snr %04x%04x",
2476 		mmc->cid[0] >> 24, (mmc->cid[2] & 0xffff),
2477 		(mmc->cid[3] >> 16) & 0xffff);
2478 	sprintf(bdesc->product, "%c%c%c%c%c%c", mmc->cid[0] & 0xff,
2479 		(mmc->cid[1] >> 24), (mmc->cid[1] >> 16) & 0xff,
2480 		(mmc->cid[1] >> 8) & 0xff, mmc->cid[1] & 0xff,
2481 		(mmc->cid[2] >> 24) & 0xff);
2482 	sprintf(bdesc->revision, "%d.%d", (mmc->cid[2] >> 20) & 0xf,
2483 		(mmc->cid[2] >> 16) & 0xf);
2484 #else
2485 	bdesc->vendor[0] = 0;
2486 	bdesc->product[0] = 0;
2487 	bdesc->revision[0] = 0;
2488 #endif
2489 
2490 #if !defined(CONFIG_DM_MMC) && (!defined(CONFIG_SPL_BUILD) || defined(CONFIG_SPL_LIBDISK_SUPPORT))
2491 	part_init(bdesc);
2492 #endif
2493 
2494 	return 0;
2495 }
2496 
2497 static int mmc_send_if_cond(struct mmc *mmc)
2498 {
2499 	struct mmc_cmd cmd;
2500 	int err;
2501 
2502 	cmd.cmdidx = SD_CMD_SEND_IF_COND;
2503 	/* We set the bit if the host supports voltages between 2.7 and 3.6 V */
2504 	cmd.cmdarg = ((mmc->cfg->voltages & 0xff8000) != 0) << 8 | 0xaa;
2505 	cmd.resp_type = MMC_RSP_R7;
2506 
2507 	err = mmc_send_cmd(mmc, &cmd, NULL);
2508 
2509 	if (err)
2510 		return err;
2511 
2512 	if ((cmd.response[0] & 0xff) != 0xaa)
2513 		return -EOPNOTSUPP;
2514 	else
2515 		mmc->version = SD_VERSION_2;
2516 
2517 	return 0;
2518 }
2519 
2520 #if !CONFIG_IS_ENABLED(DM_MMC)
2521 /* board-specific MMC power initializations. */
2522 __weak void board_mmc_power_init(void)
2523 {
2524 }
2525 #endif
2526 
2527 static int mmc_power_init(struct mmc *mmc)
2528 {
2529 #if CONFIG_IS_ENABLED(DM_MMC)
2530 #if CONFIG_IS_ENABLED(DM_REGULATOR)
2531 	int ret;
2532 
2533 	ret = device_get_supply_regulator(mmc->dev, "vmmc-supply",
2534 					  &mmc->vmmc_supply);
2535 	if (ret)
2536 		pr_debug("%s: No vmmc supply\n", mmc->dev->name);
2537 
2538 	ret = device_get_supply_regulator(mmc->dev, "vqmmc-supply",
2539 					  &mmc->vqmmc_supply);
2540 	if (ret)
2541 		pr_debug("%s: No vqmmc supply\n", mmc->dev->name);
2542 #endif
2543 #else /* !CONFIG_DM_MMC */
2544 	/*
2545 	 * Driver model should use a regulator, as above, rather than calling
2546 	 * out to board code.
2547 	 */
2548 	board_mmc_power_init();
2549 #endif
2550 	return 0;
2551 }
2552 
2553 /*
2554  * put the host in the initial state:
2555  * - turn on Vdd (card power supply)
2556  * - configure the bus width and clock to minimal values
2557  */
2558 static void mmc_set_initial_state(struct mmc *mmc)
2559 {
2560 	int err;
2561 
2562 	/* First try to set 3.3V. If it fails set to 1.8V */
2563 	err = mmc_set_signal_voltage(mmc, MMC_SIGNAL_VOLTAGE_330);
2564 	if (err != 0)
2565 		err = mmc_set_signal_voltage(mmc, MMC_SIGNAL_VOLTAGE_180);
2566 	if (err != 0)
2567 		pr_warn("mmc: failed to set signal voltage\n");
2568 
2569 	mmc_select_mode(mmc, MMC_LEGACY);
2570 	mmc_set_bus_width(mmc, 1);
2571 	mmc_set_clock(mmc, 0, MMC_CLK_ENABLE);
2572 }
2573 
2574 static int mmc_power_on(struct mmc *mmc)
2575 {
2576 #if CONFIG_IS_ENABLED(DM_MMC) && CONFIG_IS_ENABLED(DM_REGULATOR)
2577 	if (mmc->vmmc_supply) {
2578 		int ret = regulator_set_enable(mmc->vmmc_supply, true);
2579 
2580 		if (ret) {
2581 			puts("Error enabling VMMC supply\n");
2582 			return ret;
2583 		}
2584 	}
2585 #endif
2586 	return 0;
2587 }
2588 
2589 static int mmc_power_off(struct mmc *mmc)
2590 {
2591 	mmc_set_clock(mmc, 0, MMC_CLK_DISABLE);
2592 #if CONFIG_IS_ENABLED(DM_MMC) && CONFIG_IS_ENABLED(DM_REGULATOR)
2593 	if (mmc->vmmc_supply) {
2594 		int ret = regulator_set_enable(mmc->vmmc_supply, false);
2595 
2596 		if (ret) {
2597 			pr_debug("Error disabling VMMC supply\n");
2598 			return ret;
2599 		}
2600 	}
2601 #endif
2602 	return 0;
2603 }
2604 
2605 static int mmc_power_cycle(struct mmc *mmc)
2606 {
2607 	int ret;
2608 
2609 	ret = mmc_power_off(mmc);
2610 	if (ret)
2611 		return ret;
2612 	/*
2613 	 * SD spec recommends at least 1ms of delay. Let's wait for 2ms
2614 	 * to be on the safer side.
2615 	 */
2616 	udelay(2000);
2617 	return mmc_power_on(mmc);
2618 }
2619 
2620 int mmc_get_op_cond(struct mmc *mmc)
2621 {
2622 	bool uhs_en = supports_uhs(mmc->cfg->host_caps);
2623 	int err;
2624 
2625 	if (mmc->has_init)
2626 		return 0;
2627 
2628 #ifdef CONFIG_FSL_ESDHC_ADAPTER_IDENT
2629 	mmc_adapter_card_type_ident();
2630 #endif
2631 	err = mmc_power_init(mmc);
2632 	if (err)
2633 		return err;
2634 
2635 #ifdef CONFIG_MMC_QUIRKS
2636 	mmc->quirks = MMC_QUIRK_RETRY_SET_BLOCKLEN |
2637 		      MMC_QUIRK_RETRY_SEND_CID;
2638 #endif
2639 
2640 	err = mmc_power_cycle(mmc);
2641 	if (err) {
2642 		/*
2643 		 * if power cycling is not supported, we should not try
2644 		 * to use the UHS modes, because we wouldn't be able to
2645 		 * recover from an error during the UHS initialization.
2646 		 */
2647 		pr_debug("Unable to do a full power cycle. Disabling the UHS modes for safety\n");
2648 		uhs_en = false;
2649 		mmc->host_caps &= ~UHS_CAPS;
2650 		err = mmc_power_on(mmc);
2651 	}
2652 	if (err)
2653 		return err;
2654 
2655 #if CONFIG_IS_ENABLED(DM_MMC)
2656 	/* The device has already been probed ready for use */
2657 #else
2658 	/* made sure it's not NULL earlier */
2659 	err = mmc->cfg->ops->init(mmc);
2660 	if (err)
2661 		return err;
2662 #endif
2663 	mmc->ddr_mode = 0;
2664 
2665 retry:
2666 	mmc_set_initial_state(mmc);
2667 	mmc_send_init_stream(mmc);
2668 
2669 	/* Reset the Card */
2670 	err = mmc_go_idle(mmc);
2671 
2672 	if (err)
2673 		return err;
2674 
2675 	/* The internal partition reset to user partition(0) at every CMD0*/
2676 	mmc_get_blk_desc(mmc)->hwpart = 0;
2677 
2678 	/* Test for SD version 2 */
2679 	err = mmc_send_if_cond(mmc);
2680 
2681 	/* Now try to get the SD card's operating condition */
2682 	err = sd_send_op_cond(mmc, uhs_en);
2683 	if (err && uhs_en) {
2684 		uhs_en = false;
2685 		mmc_power_cycle(mmc);
2686 		goto retry;
2687 	}
2688 
2689 	/* If the command timed out, we check for an MMC card */
2690 	if (err == -ETIMEDOUT) {
2691 		err = mmc_send_op_cond(mmc);
2692 
2693 		if (err) {
2694 #if !defined(CONFIG_SPL_BUILD) || defined(CONFIG_SPL_LIBCOMMON_SUPPORT)
2695 			pr_err("Card did not respond to voltage select!\n");
2696 #endif
2697 			return -EOPNOTSUPP;
2698 		}
2699 	}
2700 
2701 	return err;
2702 }
2703 
2704 int mmc_start_init(struct mmc *mmc)
2705 {
2706 	bool no_card;
2707 	int err = 0;
2708 
2709 	/*
2710 	 * all hosts are capable of 1 bit bus-width and able to use the legacy
2711 	 * timings.
2712 	 */
2713 	mmc->host_caps = mmc->cfg->host_caps | MMC_CAP(SD_LEGACY) |
2714 			 MMC_CAP(MMC_LEGACY) | MMC_MODE_1BIT;
2715 
2716 #if !defined(CONFIG_MMC_BROKEN_CD)
2717 	/* we pretend there's no card when init is NULL */
2718 	no_card = mmc_getcd(mmc) == 0;
2719 #else
2720 	no_card = 0;
2721 #endif
2722 #if !CONFIG_IS_ENABLED(DM_MMC)
2723 	no_card = no_card || (mmc->cfg->ops->init == NULL);
2724 #endif
2725 	if (no_card) {
2726 		mmc->has_init = 0;
2727 #if !defined(CONFIG_SPL_BUILD) || defined(CONFIG_SPL_LIBCOMMON_SUPPORT)
2728 		pr_err("MMC: no card present\n");
2729 #endif
2730 		return -ENOMEDIUM;
2731 	}
2732 
2733 	err = mmc_get_op_cond(mmc);
2734 
2735 	if (!err)
2736 		mmc->init_in_progress = 1;
2737 
2738 	return err;
2739 }
2740 
2741 static int mmc_complete_init(struct mmc *mmc)
2742 {
2743 	int err = 0;
2744 
2745 	mmc->init_in_progress = 0;
2746 	if (mmc->op_cond_pending)
2747 		err = mmc_complete_op_cond(mmc);
2748 
2749 	if (!err)
2750 		err = mmc_startup(mmc);
2751 	if (err)
2752 		mmc->has_init = 0;
2753 	else
2754 		mmc->has_init = 1;
2755 	return err;
2756 }
2757 
2758 int mmc_init(struct mmc *mmc)
2759 {
2760 	int err = 0;
2761 	__maybe_unused ulong start;
2762 #if CONFIG_IS_ENABLED(DM_MMC)
2763 	struct mmc_uclass_priv *upriv = dev_get_uclass_priv(mmc->dev);
2764 
2765 	upriv->mmc = mmc;
2766 #endif
2767 	if (mmc->has_init)
2768 		return 0;
2769 
2770 	start = get_timer(0);
2771 
2772 	if (!mmc->init_in_progress)
2773 		err = mmc_start_init(mmc);
2774 
2775 	if (!err)
2776 		err = mmc_complete_init(mmc);
2777 	if (err)
2778 		pr_info("%s: %d, time %lu\n", __func__, err, get_timer(start));
2779 
2780 	return err;
2781 }
2782 
2783 #if CONFIG_IS_ENABLED(MMC_UHS_SUPPORT) || \
2784     CONFIG_IS_ENABLED(MMC_HS200_SUPPORT) || \
2785     CONFIG_IS_ENABLED(MMC_HS400_SUPPORT)
2786 int mmc_deinit(struct mmc *mmc)
2787 {
2788 	u32 caps_filtered;
2789 
2790 	if (!mmc->has_init)
2791 		return 0;
2792 
2793 	if (IS_SD(mmc)) {
2794 		caps_filtered = mmc->card_caps &
2795 			~(MMC_CAP(UHS_SDR12) | MMC_CAP(UHS_SDR25) |
2796 			  MMC_CAP(UHS_SDR50) | MMC_CAP(UHS_DDR50) |
2797 			  MMC_CAP(UHS_SDR104));
2798 
2799 		return sd_select_mode_and_width(mmc, caps_filtered);
2800 	} else {
2801 		caps_filtered = mmc->card_caps &
2802 			~(MMC_CAP(MMC_HS_200) | MMC_CAP(MMC_HS_400));
2803 
2804 		return mmc_select_mode_and_width(mmc, caps_filtered);
2805 	}
2806 }
2807 #endif
2808 
2809 int mmc_set_dsr(struct mmc *mmc, u16 val)
2810 {
2811 	mmc->dsr = val;
2812 	return 0;
2813 }
2814 
2815 /* CPU-specific MMC initializations */
2816 __weak int cpu_mmc_init(bd_t *bis)
2817 {
2818 	return -1;
2819 }
2820 
2821 /* board-specific MMC initializations. */
2822 __weak int board_mmc_init(bd_t *bis)
2823 {
2824 	return -1;
2825 }
2826 
2827 void mmc_set_preinit(struct mmc *mmc, int preinit)
2828 {
2829 	mmc->preinit = preinit;
2830 }
2831 
2832 #if CONFIG_IS_ENABLED(DM_MMC)
2833 static int mmc_probe(bd_t *bis)
2834 {
2835 	int ret, i;
2836 	struct uclass *uc;
2837 	struct udevice *dev;
2838 
2839 	ret = uclass_get(UCLASS_MMC, &uc);
2840 	if (ret)
2841 		return ret;
2842 
2843 	/*
2844 	 * Try to add them in sequence order. Really with driver model we
2845 	 * should allow holes, but the current MMC list does not allow that.
2846 	 * So if we request 0, 1, 3 we will get 0, 1, 2.
2847 	 */
2848 	for (i = 0; ; i++) {
2849 		ret = uclass_get_device_by_seq(UCLASS_MMC, i, &dev);
2850 		if (ret == -ENODEV)
2851 			break;
2852 	}
2853 	uclass_foreach_dev(dev, uc) {
2854 		ret = device_probe(dev);
2855 		if (ret)
2856 			pr_err("%s - probe failed: %d\n", dev->name, ret);
2857 	}
2858 
2859 	return 0;
2860 }
2861 #else
2862 static int mmc_probe(bd_t *bis)
2863 {
2864 	if (board_mmc_init(bis) < 0)
2865 		cpu_mmc_init(bis);
2866 
2867 	return 0;
2868 }
2869 #endif
2870 
2871 int mmc_initialize(bd_t *bis)
2872 {
2873 	static int initialized = 0;
2874 	int ret;
2875 	if (initialized)	/* Avoid initializing mmc multiple times */
2876 		return 0;
2877 	initialized = 1;
2878 
2879 #if !CONFIG_IS_ENABLED(BLK)
2880 #if !CONFIG_IS_ENABLED(MMC_TINY)
2881 	mmc_list_init();
2882 #endif
2883 #endif
2884 	ret = mmc_probe(bis);
2885 	if (ret)
2886 		return ret;
2887 
2888 #ifndef CONFIG_SPL_BUILD
2889 	print_mmc_devices(',');
2890 #endif
2891 
2892 	mmc_do_preinit();
2893 	return 0;
2894 }
2895 
2896 #ifdef CONFIG_CMD_BKOPS_ENABLE
2897 int mmc_set_bkops_enable(struct mmc *mmc)
2898 {
2899 	int err;
2900 	ALLOC_CACHE_ALIGN_BUFFER(u8, ext_csd, MMC_MAX_BLOCK_LEN);
2901 
2902 	err = mmc_send_ext_csd(mmc, ext_csd);
2903 	if (err) {
2904 		puts("Could not get ext_csd register values\n");
2905 		return err;
2906 	}
2907 
2908 	if (!(ext_csd[EXT_CSD_BKOPS_SUPPORT] & 0x1)) {
2909 		puts("Background operations not supported on device\n");
2910 		return -EMEDIUMTYPE;
2911 	}
2912 
2913 	if (ext_csd[EXT_CSD_BKOPS_EN] & 0x1) {
2914 		puts("Background operations already enabled\n");
2915 		return 0;
2916 	}
2917 
2918 	err = mmc_switch(mmc, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_BKOPS_EN, 1);
2919 	if (err) {
2920 		puts("Failed to enable manual background operations\n");
2921 		return err;
2922 	}
2923 
2924 	puts("Enabled manual background operations\n");
2925 
2926 	return 0;
2927 }
2928 #endif
2929