xref: /openbmc/linux/drivers/mmc/core/mmc.c (revision 1f9f6a78)
1 /*
2  *  linux/drivers/mmc/core/mmc.c
3  *
4  *  Copyright (C) 2003-2004 Russell King, All Rights Reserved.
5  *  Copyright (C) 2005-2007 Pierre Ossman, All Rights Reserved.
6  *  MMCv4 support Copyright (C) 2006 Philip Langdale, All Rights Reserved.
7  *
8  * This program is free software; you can redistribute it and/or modify
9  * it under the terms of the GNU General Public License version 2 as
10  * published by the Free Software Foundation.
11  */
12 
13 #include <linux/err.h>
14 #include <linux/slab.h>
15 #include <linux/stat.h>
16 #include <linux/pm_runtime.h>
17 
18 #include <linux/mmc/host.h>
19 #include <linux/mmc/card.h>
20 #include <linux/mmc/mmc.h>
21 
22 #include "core.h"
23 #include "bus.h"
24 #include "mmc_ops.h"
25 #include "sd_ops.h"
26 
27 static const unsigned int tran_exp[] = {
28 	10000,		100000,		1000000,	10000000,
29 	0,		0,		0,		0
30 };
31 
32 static const unsigned char tran_mant[] = {
33 	0,	10,	12,	13,	15,	20,	25,	30,
34 	35,	40,	45,	50,	55,	60,	70,	80,
35 };
36 
37 static const unsigned int tacc_exp[] = {
38 	1,	10,	100,	1000,	10000,	100000,	1000000, 10000000,
39 };
40 
41 static const unsigned int tacc_mant[] = {
42 	0,	10,	12,	13,	15,	20,	25,	30,
43 	35,	40,	45,	50,	55,	60,	70,	80,
44 };
45 
46 #define UNSTUFF_BITS(resp,start,size)					\
47 	({								\
48 		const int __size = size;				\
49 		const u32 __mask = (__size < 32 ? 1 << __size : 0) - 1;	\
50 		const int __off = 3 - ((start) / 32);			\
51 		const int __shft = (start) & 31;			\
52 		u32 __res;						\
53 									\
54 		__res = resp[__off] >> __shft;				\
55 		if (__size + __shft > 32)				\
56 			__res |= resp[__off-1] << ((32 - __shft) % 32);	\
57 		__res & __mask;						\
58 	})
59 
60 /*
61  * Given the decoded CSD structure, decode the raw CID to our CID structure.
62  */
63 static int mmc_decode_cid(struct mmc_card *card)
64 {
65 	u32 *resp = card->raw_cid;
66 
67 	/*
68 	 * The selection of the format here is based upon published
69 	 * specs from sandisk and from what people have reported.
70 	 */
71 	switch (card->csd.mmca_vsn) {
72 	case 0: /* MMC v1.0 - v1.2 */
73 	case 1: /* MMC v1.4 */
74 		card->cid.manfid	= UNSTUFF_BITS(resp, 104, 24);
75 		card->cid.prod_name[0]	= UNSTUFF_BITS(resp, 96, 8);
76 		card->cid.prod_name[1]	= UNSTUFF_BITS(resp, 88, 8);
77 		card->cid.prod_name[2]	= UNSTUFF_BITS(resp, 80, 8);
78 		card->cid.prod_name[3]	= UNSTUFF_BITS(resp, 72, 8);
79 		card->cid.prod_name[4]	= UNSTUFF_BITS(resp, 64, 8);
80 		card->cid.prod_name[5]	= UNSTUFF_BITS(resp, 56, 8);
81 		card->cid.prod_name[6]	= UNSTUFF_BITS(resp, 48, 8);
82 		card->cid.hwrev		= UNSTUFF_BITS(resp, 44, 4);
83 		card->cid.fwrev		= UNSTUFF_BITS(resp, 40, 4);
84 		card->cid.serial	= UNSTUFF_BITS(resp, 16, 24);
85 		card->cid.month		= UNSTUFF_BITS(resp, 12, 4);
86 		card->cid.year		= UNSTUFF_BITS(resp, 8, 4) + 1997;
87 		break;
88 
89 	case 2: /* MMC v2.0 - v2.2 */
90 	case 3: /* MMC v3.1 - v3.3 */
91 	case 4: /* MMC v4 */
92 		card->cid.manfid	= UNSTUFF_BITS(resp, 120, 8);
93 		card->cid.oemid		= UNSTUFF_BITS(resp, 104, 16);
94 		card->cid.prod_name[0]	= UNSTUFF_BITS(resp, 96, 8);
95 		card->cid.prod_name[1]	= UNSTUFF_BITS(resp, 88, 8);
96 		card->cid.prod_name[2]	= UNSTUFF_BITS(resp, 80, 8);
97 		card->cid.prod_name[3]	= UNSTUFF_BITS(resp, 72, 8);
98 		card->cid.prod_name[4]	= UNSTUFF_BITS(resp, 64, 8);
99 		card->cid.prod_name[5]	= UNSTUFF_BITS(resp, 56, 8);
100 		card->cid.prv		= UNSTUFF_BITS(resp, 48, 8);
101 		card->cid.serial	= UNSTUFF_BITS(resp, 16, 32);
102 		card->cid.month		= UNSTUFF_BITS(resp, 12, 4);
103 		card->cid.year		= UNSTUFF_BITS(resp, 8, 4) + 1997;
104 		break;
105 
106 	default:
107 		pr_err("%s: card has unknown MMCA version %d\n",
108 			mmc_hostname(card->host), card->csd.mmca_vsn);
109 		return -EINVAL;
110 	}
111 
112 	return 0;
113 }
114 
115 static void mmc_set_erase_size(struct mmc_card *card)
116 {
117 	if (card->ext_csd.erase_group_def & 1)
118 		card->erase_size = card->ext_csd.hc_erase_size;
119 	else
120 		card->erase_size = card->csd.erase_size;
121 
122 	mmc_init_erase(card);
123 }
124 
125 /*
126  * Given a 128-bit response, decode to our card CSD structure.
127  */
128 static int mmc_decode_csd(struct mmc_card *card)
129 {
130 	struct mmc_csd *csd = &card->csd;
131 	unsigned int e, m, a, b;
132 	u32 *resp = card->raw_csd;
133 
134 	/*
135 	 * We only understand CSD structure v1.1 and v1.2.
136 	 * v1.2 has extra information in bits 15, 11 and 10.
137 	 * We also support eMMC v4.4 & v4.41.
138 	 */
139 	csd->structure = UNSTUFF_BITS(resp, 126, 2);
140 	if (csd->structure == 0) {
141 		pr_err("%s: unrecognised CSD structure version %d\n",
142 			mmc_hostname(card->host), csd->structure);
143 		return -EINVAL;
144 	}
145 
146 	csd->mmca_vsn	 = UNSTUFF_BITS(resp, 122, 4);
147 	m = UNSTUFF_BITS(resp, 115, 4);
148 	e = UNSTUFF_BITS(resp, 112, 3);
149 	csd->tacc_ns	 = (tacc_exp[e] * tacc_mant[m] + 9) / 10;
150 	csd->tacc_clks	 = UNSTUFF_BITS(resp, 104, 8) * 100;
151 
152 	m = UNSTUFF_BITS(resp, 99, 4);
153 	e = UNSTUFF_BITS(resp, 96, 3);
154 	csd->max_dtr	  = tran_exp[e] * tran_mant[m];
155 	csd->cmdclass	  = UNSTUFF_BITS(resp, 84, 12);
156 
157 	e = UNSTUFF_BITS(resp, 47, 3);
158 	m = UNSTUFF_BITS(resp, 62, 12);
159 	csd->capacity	  = (1 + m) << (e + 2);
160 
161 	csd->read_blkbits = UNSTUFF_BITS(resp, 80, 4);
162 	csd->read_partial = UNSTUFF_BITS(resp, 79, 1);
163 	csd->write_misalign = UNSTUFF_BITS(resp, 78, 1);
164 	csd->read_misalign = UNSTUFF_BITS(resp, 77, 1);
165 	csd->dsr_imp = UNSTUFF_BITS(resp, 76, 1);
166 	csd->r2w_factor = UNSTUFF_BITS(resp, 26, 3);
167 	csd->write_blkbits = UNSTUFF_BITS(resp, 22, 4);
168 	csd->write_partial = UNSTUFF_BITS(resp, 21, 1);
169 
170 	if (csd->write_blkbits >= 9) {
171 		a = UNSTUFF_BITS(resp, 42, 5);
172 		b = UNSTUFF_BITS(resp, 37, 5);
173 		csd->erase_size = (a + 1) * (b + 1);
174 		csd->erase_size <<= csd->write_blkbits - 9;
175 	}
176 
177 	return 0;
178 }
179 
180 static void mmc_select_card_type(struct mmc_card *card)
181 {
182 	struct mmc_host *host = card->host;
183 	u8 card_type = card->ext_csd.raw_card_type;
184 	u32 caps = host->caps, caps2 = host->caps2;
185 	unsigned int hs_max_dtr = 0, hs200_max_dtr = 0;
186 	unsigned int avail_type = 0;
187 
188 	if (caps & MMC_CAP_MMC_HIGHSPEED &&
189 	    card_type & EXT_CSD_CARD_TYPE_HS_26) {
190 		hs_max_dtr = MMC_HIGH_26_MAX_DTR;
191 		avail_type |= EXT_CSD_CARD_TYPE_HS_26;
192 	}
193 
194 	if (caps & MMC_CAP_MMC_HIGHSPEED &&
195 	    card_type & EXT_CSD_CARD_TYPE_HS_52) {
196 		hs_max_dtr = MMC_HIGH_52_MAX_DTR;
197 		avail_type |= EXT_CSD_CARD_TYPE_HS_52;
198 	}
199 
200 	if (caps & MMC_CAP_1_8V_DDR &&
201 	    card_type & EXT_CSD_CARD_TYPE_DDR_1_8V) {
202 		hs_max_dtr = MMC_HIGH_DDR_MAX_DTR;
203 		avail_type |= EXT_CSD_CARD_TYPE_DDR_1_8V;
204 	}
205 
206 	if (caps & MMC_CAP_1_2V_DDR &&
207 	    card_type & EXT_CSD_CARD_TYPE_DDR_1_2V) {
208 		hs_max_dtr = MMC_HIGH_DDR_MAX_DTR;
209 		avail_type |= EXT_CSD_CARD_TYPE_DDR_1_2V;
210 	}
211 
212 	if (caps2 & MMC_CAP2_HS200_1_8V_SDR &&
213 	    card_type & EXT_CSD_CARD_TYPE_HS200_1_8V) {
214 		hs200_max_dtr = MMC_HS200_MAX_DTR;
215 		avail_type |= EXT_CSD_CARD_TYPE_HS200_1_8V;
216 	}
217 
218 	if (caps2 & MMC_CAP2_HS200_1_2V_SDR &&
219 	    card_type & EXT_CSD_CARD_TYPE_HS200_1_2V) {
220 		hs200_max_dtr = MMC_HS200_MAX_DTR;
221 		avail_type |= EXT_CSD_CARD_TYPE_HS200_1_2V;
222 	}
223 
224 	if (caps2 & MMC_CAP2_HS400_1_8V &&
225 	    card_type & EXT_CSD_CARD_TYPE_HS400_1_8V) {
226 		hs200_max_dtr = MMC_HS200_MAX_DTR;
227 		avail_type |= EXT_CSD_CARD_TYPE_HS400_1_8V;
228 	}
229 
230 	if (caps2 & MMC_CAP2_HS400_1_2V &&
231 	    card_type & EXT_CSD_CARD_TYPE_HS400_1_2V) {
232 		hs200_max_dtr = MMC_HS200_MAX_DTR;
233 		avail_type |= EXT_CSD_CARD_TYPE_HS400_1_2V;
234 	}
235 
236 	card->ext_csd.hs_max_dtr = hs_max_dtr;
237 	card->ext_csd.hs200_max_dtr = hs200_max_dtr;
238 	card->mmc_avail_type = avail_type;
239 }
240 
241 static void mmc_manage_enhanced_area(struct mmc_card *card, u8 *ext_csd)
242 {
243 	u8 hc_erase_grp_sz, hc_wp_grp_sz;
244 
245 	/*
246 	 * Disable these attributes by default
247 	 */
248 	card->ext_csd.enhanced_area_offset = -EINVAL;
249 	card->ext_csd.enhanced_area_size = -EINVAL;
250 
251 	/*
252 	 * Enhanced area feature support -- check whether the eMMC
253 	 * card has the Enhanced area enabled.  If so, export enhanced
254 	 * area offset and size to user by adding sysfs interface.
255 	 */
256 	if ((ext_csd[EXT_CSD_PARTITION_SUPPORT] & 0x2) &&
257 	    (ext_csd[EXT_CSD_PARTITION_ATTRIBUTE] & 0x1)) {
258 		if (card->ext_csd.partition_setting_completed) {
259 			hc_erase_grp_sz =
260 				ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE];
261 			hc_wp_grp_sz =
262 				ext_csd[EXT_CSD_HC_WP_GRP_SIZE];
263 
264 			/*
265 			 * calculate the enhanced data area offset, in bytes
266 			 */
267 			card->ext_csd.enhanced_area_offset =
268 				(ext_csd[139] << 24) + (ext_csd[138] << 16) +
269 				(ext_csd[137] << 8) + ext_csd[136];
270 			if (mmc_card_blockaddr(card))
271 				card->ext_csd.enhanced_area_offset <<= 9;
272 			/*
273 			 * calculate the enhanced data area size, in kilobytes
274 			 */
275 			card->ext_csd.enhanced_area_size =
276 				(ext_csd[142] << 16) + (ext_csd[141] << 8) +
277 				ext_csd[140];
278 			card->ext_csd.enhanced_area_size *=
279 				(size_t)(hc_erase_grp_sz * hc_wp_grp_sz);
280 			card->ext_csd.enhanced_area_size <<= 9;
281 		} else {
282 			pr_warn("%s: defines enhanced area without partition setting complete\n",
283 				mmc_hostname(card->host));
284 		}
285 	}
286 }
287 
288 static void mmc_manage_gp_partitions(struct mmc_card *card, u8 *ext_csd)
289 {
290 	int idx;
291 	u8 hc_erase_grp_sz, hc_wp_grp_sz;
292 	unsigned int part_size;
293 
294 	/*
295 	 * General purpose partition feature support --
296 	 * If ext_csd has the size of general purpose partitions,
297 	 * set size, part_cfg, partition name in mmc_part.
298 	 */
299 	if (ext_csd[EXT_CSD_PARTITION_SUPPORT] &
300 	    EXT_CSD_PART_SUPPORT_PART_EN) {
301 		hc_erase_grp_sz =
302 			ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE];
303 		hc_wp_grp_sz =
304 			ext_csd[EXT_CSD_HC_WP_GRP_SIZE];
305 
306 		for (idx = 0; idx < MMC_NUM_GP_PARTITION; idx++) {
307 			if (!ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3] &&
308 			    !ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 1] &&
309 			    !ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 2])
310 				continue;
311 			if (card->ext_csd.partition_setting_completed == 0) {
312 				pr_warn("%s: has partition size defined without partition complete\n",
313 					mmc_hostname(card->host));
314 				break;
315 			}
316 			part_size =
317 				(ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 2]
318 				<< 16) +
319 				(ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 1]
320 				<< 8) +
321 				ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3];
322 			part_size *= (size_t)(hc_erase_grp_sz *
323 				hc_wp_grp_sz);
324 			mmc_part_add(card, part_size << 19,
325 				EXT_CSD_PART_CONFIG_ACC_GP0 + idx,
326 				"gp%d", idx, false,
327 				MMC_BLK_DATA_AREA_GP);
328 		}
329 	}
330 }
331 
332 /*
333  * Decode extended CSD.
334  */
335 static int mmc_decode_ext_csd(struct mmc_card *card, u8 *ext_csd)
336 {
337 	int err = 0, idx;
338 	unsigned int part_size;
339 
340 	/* Version is coded in the CSD_STRUCTURE byte in the EXT_CSD register */
341 	card->ext_csd.raw_ext_csd_structure = ext_csd[EXT_CSD_STRUCTURE];
342 	if (card->csd.structure == 3) {
343 		if (card->ext_csd.raw_ext_csd_structure > 2) {
344 			pr_err("%s: unrecognised EXT_CSD structure "
345 				"version %d\n", mmc_hostname(card->host),
346 					card->ext_csd.raw_ext_csd_structure);
347 			err = -EINVAL;
348 			goto out;
349 		}
350 	}
351 
352 	/*
353 	 * The EXT_CSD format is meant to be forward compatible. As long
354 	 * as CSD_STRUCTURE does not change, all values for EXT_CSD_REV
355 	 * are authorized, see JEDEC JESD84-B50 section B.8.
356 	 */
357 	card->ext_csd.rev = ext_csd[EXT_CSD_REV];
358 
359 	card->ext_csd.raw_sectors[0] = ext_csd[EXT_CSD_SEC_CNT + 0];
360 	card->ext_csd.raw_sectors[1] = ext_csd[EXT_CSD_SEC_CNT + 1];
361 	card->ext_csd.raw_sectors[2] = ext_csd[EXT_CSD_SEC_CNT + 2];
362 	card->ext_csd.raw_sectors[3] = ext_csd[EXT_CSD_SEC_CNT + 3];
363 	if (card->ext_csd.rev >= 2) {
364 		card->ext_csd.sectors =
365 			ext_csd[EXT_CSD_SEC_CNT + 0] << 0 |
366 			ext_csd[EXT_CSD_SEC_CNT + 1] << 8 |
367 			ext_csd[EXT_CSD_SEC_CNT + 2] << 16 |
368 			ext_csd[EXT_CSD_SEC_CNT + 3] << 24;
369 
370 		/* Cards with density > 2GiB are sector addressed */
371 		if (card->ext_csd.sectors > (2u * 1024 * 1024 * 1024) / 512)
372 			mmc_card_set_blockaddr(card);
373 	}
374 
375 	card->ext_csd.raw_card_type = ext_csd[EXT_CSD_CARD_TYPE];
376 	mmc_select_card_type(card);
377 
378 	card->ext_csd.raw_s_a_timeout = ext_csd[EXT_CSD_S_A_TIMEOUT];
379 	card->ext_csd.raw_erase_timeout_mult =
380 		ext_csd[EXT_CSD_ERASE_TIMEOUT_MULT];
381 	card->ext_csd.raw_hc_erase_grp_size =
382 		ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE];
383 	if (card->ext_csd.rev >= 3) {
384 		u8 sa_shift = ext_csd[EXT_CSD_S_A_TIMEOUT];
385 		card->ext_csd.part_config = ext_csd[EXT_CSD_PART_CONFIG];
386 
387 		/* EXT_CSD value is in units of 10ms, but we store in ms */
388 		card->ext_csd.part_time = 10 * ext_csd[EXT_CSD_PART_SWITCH_TIME];
389 
390 		/* Sleep / awake timeout in 100ns units */
391 		if (sa_shift > 0 && sa_shift <= 0x17)
392 			card->ext_csd.sa_timeout =
393 					1 << ext_csd[EXT_CSD_S_A_TIMEOUT];
394 		card->ext_csd.erase_group_def =
395 			ext_csd[EXT_CSD_ERASE_GROUP_DEF];
396 		card->ext_csd.hc_erase_timeout = 300 *
397 			ext_csd[EXT_CSD_ERASE_TIMEOUT_MULT];
398 		card->ext_csd.hc_erase_size =
399 			ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE] << 10;
400 
401 		card->ext_csd.rel_sectors = ext_csd[EXT_CSD_REL_WR_SEC_C];
402 
403 		/*
404 		 * There are two boot regions of equal size, defined in
405 		 * multiples of 128K.
406 		 */
407 		if (ext_csd[EXT_CSD_BOOT_MULT] && mmc_boot_partition_access(card->host)) {
408 			for (idx = 0; idx < MMC_NUM_BOOT_PARTITION; idx++) {
409 				part_size = ext_csd[EXT_CSD_BOOT_MULT] << 17;
410 				mmc_part_add(card, part_size,
411 					EXT_CSD_PART_CONFIG_ACC_BOOT0 + idx,
412 					"boot%d", idx, true,
413 					MMC_BLK_DATA_AREA_BOOT);
414 			}
415 		}
416 	}
417 
418 	card->ext_csd.raw_hc_erase_gap_size =
419 		ext_csd[EXT_CSD_HC_WP_GRP_SIZE];
420 	card->ext_csd.raw_sec_trim_mult =
421 		ext_csd[EXT_CSD_SEC_TRIM_MULT];
422 	card->ext_csd.raw_sec_erase_mult =
423 		ext_csd[EXT_CSD_SEC_ERASE_MULT];
424 	card->ext_csd.raw_sec_feature_support =
425 		ext_csd[EXT_CSD_SEC_FEATURE_SUPPORT];
426 	card->ext_csd.raw_trim_mult =
427 		ext_csd[EXT_CSD_TRIM_MULT];
428 	card->ext_csd.raw_partition_support = ext_csd[EXT_CSD_PARTITION_SUPPORT];
429 	if (card->ext_csd.rev >= 4) {
430 		if (ext_csd[EXT_CSD_PARTITION_SETTING_COMPLETED] &
431 		    EXT_CSD_PART_SETTING_COMPLETED)
432 			card->ext_csd.partition_setting_completed = 1;
433 		else
434 			card->ext_csd.partition_setting_completed = 0;
435 
436 		mmc_manage_enhanced_area(card, ext_csd);
437 
438 		mmc_manage_gp_partitions(card, ext_csd);
439 
440 		card->ext_csd.sec_trim_mult =
441 			ext_csd[EXT_CSD_SEC_TRIM_MULT];
442 		card->ext_csd.sec_erase_mult =
443 			ext_csd[EXT_CSD_SEC_ERASE_MULT];
444 		card->ext_csd.sec_feature_support =
445 			ext_csd[EXT_CSD_SEC_FEATURE_SUPPORT];
446 		card->ext_csd.trim_timeout = 300 *
447 			ext_csd[EXT_CSD_TRIM_MULT];
448 
449 		/*
450 		 * Note that the call to mmc_part_add above defaults to read
451 		 * only. If this default assumption is changed, the call must
452 		 * take into account the value of boot_locked below.
453 		 */
454 		card->ext_csd.boot_ro_lock = ext_csd[EXT_CSD_BOOT_WP];
455 		card->ext_csd.boot_ro_lockable = true;
456 
457 		/* Save power class values */
458 		card->ext_csd.raw_pwr_cl_52_195 =
459 			ext_csd[EXT_CSD_PWR_CL_52_195];
460 		card->ext_csd.raw_pwr_cl_26_195 =
461 			ext_csd[EXT_CSD_PWR_CL_26_195];
462 		card->ext_csd.raw_pwr_cl_52_360 =
463 			ext_csd[EXT_CSD_PWR_CL_52_360];
464 		card->ext_csd.raw_pwr_cl_26_360 =
465 			ext_csd[EXT_CSD_PWR_CL_26_360];
466 		card->ext_csd.raw_pwr_cl_200_195 =
467 			ext_csd[EXT_CSD_PWR_CL_200_195];
468 		card->ext_csd.raw_pwr_cl_200_360 =
469 			ext_csd[EXT_CSD_PWR_CL_200_360];
470 		card->ext_csd.raw_pwr_cl_ddr_52_195 =
471 			ext_csd[EXT_CSD_PWR_CL_DDR_52_195];
472 		card->ext_csd.raw_pwr_cl_ddr_52_360 =
473 			ext_csd[EXT_CSD_PWR_CL_DDR_52_360];
474 		card->ext_csd.raw_pwr_cl_ddr_200_360 =
475 			ext_csd[EXT_CSD_PWR_CL_DDR_200_360];
476 	}
477 
478 	if (card->ext_csd.rev >= 5) {
479 		/* Adjust production date as per JEDEC JESD84-B451 */
480 		if (card->cid.year < 2010)
481 			card->cid.year += 16;
482 
483 		/* check whether the eMMC card supports BKOPS */
484 		if (ext_csd[EXT_CSD_BKOPS_SUPPORT] & 0x1) {
485 			card->ext_csd.bkops = 1;
486 			card->ext_csd.bkops_en = ext_csd[EXT_CSD_BKOPS_EN];
487 			card->ext_csd.raw_bkops_status =
488 				ext_csd[EXT_CSD_BKOPS_STATUS];
489 			if (!card->ext_csd.bkops_en)
490 				pr_info("%s: BKOPS_EN bit is not set\n",
491 					mmc_hostname(card->host));
492 		}
493 
494 		/* check whether the eMMC card supports HPI */
495 		if (ext_csd[EXT_CSD_HPI_FEATURES] & 0x1) {
496 			card->ext_csd.hpi = 1;
497 			if (ext_csd[EXT_CSD_HPI_FEATURES] & 0x2)
498 				card->ext_csd.hpi_cmd =	MMC_STOP_TRANSMISSION;
499 			else
500 				card->ext_csd.hpi_cmd = MMC_SEND_STATUS;
501 			/*
502 			 * Indicate the maximum timeout to close
503 			 * a command interrupted by HPI
504 			 */
505 			card->ext_csd.out_of_int_time =
506 				ext_csd[EXT_CSD_OUT_OF_INTERRUPT_TIME] * 10;
507 		}
508 
509 		card->ext_csd.rel_param = ext_csd[EXT_CSD_WR_REL_PARAM];
510 		card->ext_csd.rst_n_function = ext_csd[EXT_CSD_RST_N_FUNCTION];
511 
512 		/*
513 		 * RPMB regions are defined in multiples of 128K.
514 		 */
515 		card->ext_csd.raw_rpmb_size_mult = ext_csd[EXT_CSD_RPMB_MULT];
516 		if (ext_csd[EXT_CSD_RPMB_MULT] && mmc_host_cmd23(card->host)) {
517 			mmc_part_add(card, ext_csd[EXT_CSD_RPMB_MULT] << 17,
518 				EXT_CSD_PART_CONFIG_ACC_RPMB,
519 				"rpmb", 0, false,
520 				MMC_BLK_DATA_AREA_RPMB);
521 		}
522 	}
523 
524 	card->ext_csd.raw_erased_mem_count = ext_csd[EXT_CSD_ERASED_MEM_CONT];
525 	if (ext_csd[EXT_CSD_ERASED_MEM_CONT])
526 		card->erased_byte = 0xFF;
527 	else
528 		card->erased_byte = 0x0;
529 
530 	/* eMMC v4.5 or later */
531 	if (card->ext_csd.rev >= 6) {
532 		card->ext_csd.feature_support |= MMC_DISCARD_FEATURE;
533 
534 		card->ext_csd.generic_cmd6_time = 10 *
535 			ext_csd[EXT_CSD_GENERIC_CMD6_TIME];
536 		card->ext_csd.power_off_longtime = 10 *
537 			ext_csd[EXT_CSD_POWER_OFF_LONG_TIME];
538 
539 		card->ext_csd.cache_size =
540 			ext_csd[EXT_CSD_CACHE_SIZE + 0] << 0 |
541 			ext_csd[EXT_CSD_CACHE_SIZE + 1] << 8 |
542 			ext_csd[EXT_CSD_CACHE_SIZE + 2] << 16 |
543 			ext_csd[EXT_CSD_CACHE_SIZE + 3] << 24;
544 
545 		if (ext_csd[EXT_CSD_DATA_SECTOR_SIZE] == 1)
546 			card->ext_csd.data_sector_size = 4096;
547 		else
548 			card->ext_csd.data_sector_size = 512;
549 
550 		if ((ext_csd[EXT_CSD_DATA_TAG_SUPPORT] & 1) &&
551 		    (ext_csd[EXT_CSD_TAG_UNIT_SIZE] <= 8)) {
552 			card->ext_csd.data_tag_unit_size =
553 			((unsigned int) 1 << ext_csd[EXT_CSD_TAG_UNIT_SIZE]) *
554 			(card->ext_csd.data_sector_size);
555 		} else {
556 			card->ext_csd.data_tag_unit_size = 0;
557 		}
558 
559 		card->ext_csd.max_packed_writes =
560 			ext_csd[EXT_CSD_MAX_PACKED_WRITES];
561 		card->ext_csd.max_packed_reads =
562 			ext_csd[EXT_CSD_MAX_PACKED_READS];
563 	} else {
564 		card->ext_csd.data_sector_size = 512;
565 	}
566 
567 	/* eMMC v5 or later */
568 	if (card->ext_csd.rev >= 7) {
569 		memcpy(card->ext_csd.fwrev, &ext_csd[EXT_CSD_FIRMWARE_VERSION],
570 		       MMC_FIRMWARE_LEN);
571 		card->ext_csd.ffu_capable =
572 			(ext_csd[EXT_CSD_SUPPORTED_MODE] & 0x1) &&
573 			!(ext_csd[EXT_CSD_FW_CONFIG] & 0x1);
574 	}
575 out:
576 	return err;
577 }
578 
579 static int mmc_read_ext_csd(struct mmc_card *card)
580 {
581 	u8 *ext_csd;
582 	int err;
583 
584 	if (!mmc_can_ext_csd(card))
585 		return 0;
586 
587 	err = mmc_get_ext_csd(card, &ext_csd);
588 	if (err) {
589 		/* If the host or the card can't do the switch,
590 		 * fail more gracefully. */
591 		if ((err != -EINVAL)
592 		 && (err != -ENOSYS)
593 		 && (err != -EFAULT))
594 			return err;
595 
596 		/*
597 		 * High capacity cards should have this "magic" size
598 		 * stored in their CSD.
599 		 */
600 		if (card->csd.capacity == (4096 * 512)) {
601 			pr_err("%s: unable to read EXT_CSD on a possible high capacity card. Card will be ignored.\n",
602 				mmc_hostname(card->host));
603 		} else {
604 			pr_warn("%s: unable to read EXT_CSD, performance might suffer\n",
605 				mmc_hostname(card->host));
606 			err = 0;
607 		}
608 
609 		return err;
610 	}
611 
612 	err = mmc_decode_ext_csd(card, ext_csd);
613 	kfree(ext_csd);
614 	return err;
615 }
616 
617 static int mmc_compare_ext_csds(struct mmc_card *card, unsigned bus_width)
618 {
619 	u8 *bw_ext_csd;
620 	int err;
621 
622 	if (bus_width == MMC_BUS_WIDTH_1)
623 		return 0;
624 
625 	err = mmc_get_ext_csd(card, &bw_ext_csd);
626 	if (err)
627 		return err;
628 
629 	/* only compare read only fields */
630 	err = !((card->ext_csd.raw_partition_support ==
631 			bw_ext_csd[EXT_CSD_PARTITION_SUPPORT]) &&
632 		(card->ext_csd.raw_erased_mem_count ==
633 			bw_ext_csd[EXT_CSD_ERASED_MEM_CONT]) &&
634 		(card->ext_csd.rev ==
635 			bw_ext_csd[EXT_CSD_REV]) &&
636 		(card->ext_csd.raw_ext_csd_structure ==
637 			bw_ext_csd[EXT_CSD_STRUCTURE]) &&
638 		(card->ext_csd.raw_card_type ==
639 			bw_ext_csd[EXT_CSD_CARD_TYPE]) &&
640 		(card->ext_csd.raw_s_a_timeout ==
641 			bw_ext_csd[EXT_CSD_S_A_TIMEOUT]) &&
642 		(card->ext_csd.raw_hc_erase_gap_size ==
643 			bw_ext_csd[EXT_CSD_HC_WP_GRP_SIZE]) &&
644 		(card->ext_csd.raw_erase_timeout_mult ==
645 			bw_ext_csd[EXT_CSD_ERASE_TIMEOUT_MULT]) &&
646 		(card->ext_csd.raw_hc_erase_grp_size ==
647 			bw_ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE]) &&
648 		(card->ext_csd.raw_sec_trim_mult ==
649 			bw_ext_csd[EXT_CSD_SEC_TRIM_MULT]) &&
650 		(card->ext_csd.raw_sec_erase_mult ==
651 			bw_ext_csd[EXT_CSD_SEC_ERASE_MULT]) &&
652 		(card->ext_csd.raw_sec_feature_support ==
653 			bw_ext_csd[EXT_CSD_SEC_FEATURE_SUPPORT]) &&
654 		(card->ext_csd.raw_trim_mult ==
655 			bw_ext_csd[EXT_CSD_TRIM_MULT]) &&
656 		(card->ext_csd.raw_sectors[0] ==
657 			bw_ext_csd[EXT_CSD_SEC_CNT + 0]) &&
658 		(card->ext_csd.raw_sectors[1] ==
659 			bw_ext_csd[EXT_CSD_SEC_CNT + 1]) &&
660 		(card->ext_csd.raw_sectors[2] ==
661 			bw_ext_csd[EXT_CSD_SEC_CNT + 2]) &&
662 		(card->ext_csd.raw_sectors[3] ==
663 			bw_ext_csd[EXT_CSD_SEC_CNT + 3]) &&
664 		(card->ext_csd.raw_pwr_cl_52_195 ==
665 			bw_ext_csd[EXT_CSD_PWR_CL_52_195]) &&
666 		(card->ext_csd.raw_pwr_cl_26_195 ==
667 			bw_ext_csd[EXT_CSD_PWR_CL_26_195]) &&
668 		(card->ext_csd.raw_pwr_cl_52_360 ==
669 			bw_ext_csd[EXT_CSD_PWR_CL_52_360]) &&
670 		(card->ext_csd.raw_pwr_cl_26_360 ==
671 			bw_ext_csd[EXT_CSD_PWR_CL_26_360]) &&
672 		(card->ext_csd.raw_pwr_cl_200_195 ==
673 			bw_ext_csd[EXT_CSD_PWR_CL_200_195]) &&
674 		(card->ext_csd.raw_pwr_cl_200_360 ==
675 			bw_ext_csd[EXT_CSD_PWR_CL_200_360]) &&
676 		(card->ext_csd.raw_pwr_cl_ddr_52_195 ==
677 			bw_ext_csd[EXT_CSD_PWR_CL_DDR_52_195]) &&
678 		(card->ext_csd.raw_pwr_cl_ddr_52_360 ==
679 			bw_ext_csd[EXT_CSD_PWR_CL_DDR_52_360]) &&
680 		(card->ext_csd.raw_pwr_cl_ddr_200_360 ==
681 			bw_ext_csd[EXT_CSD_PWR_CL_DDR_200_360]));
682 
683 	if (err)
684 		err = -EINVAL;
685 
686 	kfree(bw_ext_csd);
687 	return err;
688 }
689 
690 MMC_DEV_ATTR(cid, "%08x%08x%08x%08x\n", card->raw_cid[0], card->raw_cid[1],
691 	card->raw_cid[2], card->raw_cid[3]);
692 MMC_DEV_ATTR(csd, "%08x%08x%08x%08x\n", card->raw_csd[0], card->raw_csd[1],
693 	card->raw_csd[2], card->raw_csd[3]);
694 MMC_DEV_ATTR(date, "%02d/%04d\n", card->cid.month, card->cid.year);
695 MMC_DEV_ATTR(erase_size, "%u\n", card->erase_size << 9);
696 MMC_DEV_ATTR(preferred_erase_size, "%u\n", card->pref_erase << 9);
697 MMC_DEV_ATTR(ffu_capable, "%d\n", card->ext_csd.ffu_capable);
698 MMC_DEV_ATTR(hwrev, "0x%x\n", card->cid.hwrev);
699 MMC_DEV_ATTR(manfid, "0x%06x\n", card->cid.manfid);
700 MMC_DEV_ATTR(name, "%s\n", card->cid.prod_name);
701 MMC_DEV_ATTR(oemid, "0x%04x\n", card->cid.oemid);
702 MMC_DEV_ATTR(prv, "0x%x\n", card->cid.prv);
703 MMC_DEV_ATTR(serial, "0x%08x\n", card->cid.serial);
704 MMC_DEV_ATTR(enhanced_area_offset, "%llu\n",
705 		card->ext_csd.enhanced_area_offset);
706 MMC_DEV_ATTR(enhanced_area_size, "%u\n", card->ext_csd.enhanced_area_size);
707 MMC_DEV_ATTR(raw_rpmb_size_mult, "%#x\n", card->ext_csd.raw_rpmb_size_mult);
708 MMC_DEV_ATTR(rel_sectors, "%#x\n", card->ext_csd.rel_sectors);
709 
710 static ssize_t mmc_fwrev_show(struct device *dev,
711 			      struct device_attribute *attr,
712 			      char *buf)
713 {
714 	struct mmc_card *card = mmc_dev_to_card(dev);
715 
716 	if (card->ext_csd.rev < 7) {
717 		return sprintf(buf, "0x%x\n", card->cid.fwrev);
718 	} else {
719 		return sprintf(buf, "0x%*phN\n", MMC_FIRMWARE_LEN,
720 			       card->ext_csd.fwrev);
721 	}
722 }
723 
724 static DEVICE_ATTR(fwrev, S_IRUGO, mmc_fwrev_show, NULL);
725 
726 static struct attribute *mmc_std_attrs[] = {
727 	&dev_attr_cid.attr,
728 	&dev_attr_csd.attr,
729 	&dev_attr_date.attr,
730 	&dev_attr_erase_size.attr,
731 	&dev_attr_preferred_erase_size.attr,
732 	&dev_attr_fwrev.attr,
733 	&dev_attr_ffu_capable.attr,
734 	&dev_attr_hwrev.attr,
735 	&dev_attr_manfid.attr,
736 	&dev_attr_name.attr,
737 	&dev_attr_oemid.attr,
738 	&dev_attr_prv.attr,
739 	&dev_attr_serial.attr,
740 	&dev_attr_enhanced_area_offset.attr,
741 	&dev_attr_enhanced_area_size.attr,
742 	&dev_attr_raw_rpmb_size_mult.attr,
743 	&dev_attr_rel_sectors.attr,
744 	NULL,
745 };
746 ATTRIBUTE_GROUPS(mmc_std);
747 
748 static struct device_type mmc_type = {
749 	.groups = mmc_std_groups,
750 };
751 
752 /*
753  * Select the PowerClass for the current bus width
754  * If power class is defined for 4/8 bit bus in the
755  * extended CSD register, select it by executing the
756  * mmc_switch command.
757  */
758 static int __mmc_select_powerclass(struct mmc_card *card,
759 				   unsigned int bus_width)
760 {
761 	struct mmc_host *host = card->host;
762 	struct mmc_ext_csd *ext_csd = &card->ext_csd;
763 	unsigned int pwrclass_val = 0;
764 	int err = 0;
765 
766 	switch (1 << host->ios.vdd) {
767 	case MMC_VDD_165_195:
768 		if (host->ios.clock <= MMC_HIGH_26_MAX_DTR)
769 			pwrclass_val = ext_csd->raw_pwr_cl_26_195;
770 		else if (host->ios.clock <= MMC_HIGH_52_MAX_DTR)
771 			pwrclass_val = (bus_width <= EXT_CSD_BUS_WIDTH_8) ?
772 				ext_csd->raw_pwr_cl_52_195 :
773 				ext_csd->raw_pwr_cl_ddr_52_195;
774 		else if (host->ios.clock <= MMC_HS200_MAX_DTR)
775 			pwrclass_val = ext_csd->raw_pwr_cl_200_195;
776 		break;
777 	case MMC_VDD_27_28:
778 	case MMC_VDD_28_29:
779 	case MMC_VDD_29_30:
780 	case MMC_VDD_30_31:
781 	case MMC_VDD_31_32:
782 	case MMC_VDD_32_33:
783 	case MMC_VDD_33_34:
784 	case MMC_VDD_34_35:
785 	case MMC_VDD_35_36:
786 		if (host->ios.clock <= MMC_HIGH_26_MAX_DTR)
787 			pwrclass_val = ext_csd->raw_pwr_cl_26_360;
788 		else if (host->ios.clock <= MMC_HIGH_52_MAX_DTR)
789 			pwrclass_val = (bus_width <= EXT_CSD_BUS_WIDTH_8) ?
790 				ext_csd->raw_pwr_cl_52_360 :
791 				ext_csd->raw_pwr_cl_ddr_52_360;
792 		else if (host->ios.clock <= MMC_HS200_MAX_DTR)
793 			pwrclass_val = (bus_width == EXT_CSD_DDR_BUS_WIDTH_8) ?
794 				ext_csd->raw_pwr_cl_ddr_200_360 :
795 				ext_csd->raw_pwr_cl_200_360;
796 		break;
797 	default:
798 		pr_warn("%s: Voltage range not supported for power class\n",
799 			mmc_hostname(host));
800 		return -EINVAL;
801 	}
802 
803 	if (bus_width & (EXT_CSD_BUS_WIDTH_8 | EXT_CSD_DDR_BUS_WIDTH_8))
804 		pwrclass_val = (pwrclass_val & EXT_CSD_PWR_CL_8BIT_MASK) >>
805 				EXT_CSD_PWR_CL_8BIT_SHIFT;
806 	else
807 		pwrclass_val = (pwrclass_val & EXT_CSD_PWR_CL_4BIT_MASK) >>
808 				EXT_CSD_PWR_CL_4BIT_SHIFT;
809 
810 	/* If the power class is different from the default value */
811 	if (pwrclass_val > 0) {
812 		err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
813 				 EXT_CSD_POWER_CLASS,
814 				 pwrclass_val,
815 				 card->ext_csd.generic_cmd6_time);
816 	}
817 
818 	return err;
819 }
820 
821 static int mmc_select_powerclass(struct mmc_card *card)
822 {
823 	struct mmc_host *host = card->host;
824 	u32 bus_width, ext_csd_bits;
825 	int err, ddr;
826 
827 	/* Power class selection is supported for versions >= 4.0 */
828 	if (!mmc_can_ext_csd(card))
829 		return 0;
830 
831 	bus_width = host->ios.bus_width;
832 	/* Power class values are defined only for 4/8 bit bus */
833 	if (bus_width == MMC_BUS_WIDTH_1)
834 		return 0;
835 
836 	ddr = card->mmc_avail_type & EXT_CSD_CARD_TYPE_DDR_52;
837 	if (ddr)
838 		ext_csd_bits = (bus_width == MMC_BUS_WIDTH_8) ?
839 			EXT_CSD_DDR_BUS_WIDTH_8 : EXT_CSD_DDR_BUS_WIDTH_4;
840 	else
841 		ext_csd_bits = (bus_width == MMC_BUS_WIDTH_8) ?
842 			EXT_CSD_BUS_WIDTH_8 :  EXT_CSD_BUS_WIDTH_4;
843 
844 	err = __mmc_select_powerclass(card, ext_csd_bits);
845 	if (err)
846 		pr_warn("%s: power class selection to bus width %d ddr %d failed\n",
847 			mmc_hostname(host), 1 << bus_width, ddr);
848 
849 	return err;
850 }
851 
852 /*
853  * Set the bus speed for the selected speed mode.
854  */
855 static void mmc_set_bus_speed(struct mmc_card *card)
856 {
857 	unsigned int max_dtr = (unsigned int)-1;
858 
859 	if ((mmc_card_hs200(card) || mmc_card_hs400(card)) &&
860 	     max_dtr > card->ext_csd.hs200_max_dtr)
861 		max_dtr = card->ext_csd.hs200_max_dtr;
862 	else if (mmc_card_hs(card) && max_dtr > card->ext_csd.hs_max_dtr)
863 		max_dtr = card->ext_csd.hs_max_dtr;
864 	else if (max_dtr > card->csd.max_dtr)
865 		max_dtr = card->csd.max_dtr;
866 
867 	mmc_set_clock(card->host, max_dtr);
868 }
869 
870 /*
871  * Select the bus width amoung 4-bit and 8-bit(SDR).
872  * If the bus width is changed successfully, return the selected width value.
873  * Zero is returned instead of error value if the wide width is not supported.
874  */
875 static int mmc_select_bus_width(struct mmc_card *card)
876 {
877 	static unsigned ext_csd_bits[] = {
878 		EXT_CSD_BUS_WIDTH_8,
879 		EXT_CSD_BUS_WIDTH_4,
880 	};
881 	static unsigned bus_widths[] = {
882 		MMC_BUS_WIDTH_8,
883 		MMC_BUS_WIDTH_4,
884 	};
885 	struct mmc_host *host = card->host;
886 	unsigned idx, bus_width = 0;
887 	int err = 0;
888 
889 	if (!mmc_can_ext_csd(card) &&
890 	    !(host->caps & (MMC_CAP_4_BIT_DATA | MMC_CAP_8_BIT_DATA)))
891 		return 0;
892 
893 	idx = (host->caps & MMC_CAP_8_BIT_DATA) ? 0 : 1;
894 
895 	/*
896 	 * Unlike SD, MMC cards dont have a configuration register to notify
897 	 * supported bus width. So bus test command should be run to identify
898 	 * the supported bus width or compare the ext csd values of current
899 	 * bus width and ext csd values of 1 bit mode read earlier.
900 	 */
901 	for (; idx < ARRAY_SIZE(bus_widths); idx++) {
902 		/*
903 		 * Host is capable of 8bit transfer, then switch
904 		 * the device to work in 8bit transfer mode. If the
905 		 * mmc switch command returns error then switch to
906 		 * 4bit transfer mode. On success set the corresponding
907 		 * bus width on the host.
908 		 */
909 		err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
910 				 EXT_CSD_BUS_WIDTH,
911 				 ext_csd_bits[idx],
912 				 card->ext_csd.generic_cmd6_time);
913 		if (err)
914 			continue;
915 
916 		bus_width = bus_widths[idx];
917 		mmc_set_bus_width(host, bus_width);
918 
919 		/*
920 		 * If controller can't handle bus width test,
921 		 * compare ext_csd previously read in 1 bit mode
922 		 * against ext_csd at new bus width
923 		 */
924 		if (!(host->caps & MMC_CAP_BUS_WIDTH_TEST))
925 			err = mmc_compare_ext_csds(card, bus_width);
926 		else
927 			err = mmc_bus_test(card, bus_width);
928 
929 		if (!err) {
930 			err = bus_width;
931 			break;
932 		} else {
933 			pr_warn("%s: switch to bus width %d failed\n",
934 				mmc_hostname(host), ext_csd_bits[idx]);
935 		}
936 	}
937 
938 	return err;
939 }
940 
941 /*
942  * Switch to the high-speed mode
943  */
944 static int mmc_select_hs(struct mmc_card *card)
945 {
946 	int err;
947 
948 	err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
949 			   EXT_CSD_HS_TIMING, EXT_CSD_TIMING_HS,
950 			   card->ext_csd.generic_cmd6_time,
951 			   true, true, true);
952 	if (!err)
953 		mmc_set_timing(card->host, MMC_TIMING_MMC_HS);
954 
955 	return err;
956 }
957 
958 /*
959  * Activate wide bus and DDR if supported.
960  */
961 static int mmc_select_hs_ddr(struct mmc_card *card)
962 {
963 	struct mmc_host *host = card->host;
964 	u32 bus_width, ext_csd_bits;
965 	int err = 0;
966 
967 	if (!(card->mmc_avail_type & EXT_CSD_CARD_TYPE_DDR_52))
968 		return 0;
969 
970 	bus_width = host->ios.bus_width;
971 	if (bus_width == MMC_BUS_WIDTH_1)
972 		return 0;
973 
974 	ext_csd_bits = (bus_width == MMC_BUS_WIDTH_8) ?
975 		EXT_CSD_DDR_BUS_WIDTH_8 : EXT_CSD_DDR_BUS_WIDTH_4;
976 
977 	err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
978 			EXT_CSD_BUS_WIDTH,
979 			ext_csd_bits,
980 			card->ext_csd.generic_cmd6_time);
981 	if (err) {
982 		pr_err("%s: switch to bus width %d ddr failed\n",
983 			mmc_hostname(host), 1 << bus_width);
984 		return err;
985 	}
986 
987 	/*
988 	 * eMMC cards can support 3.3V to 1.2V i/o (vccq)
989 	 * signaling.
990 	 *
991 	 * EXT_CSD_CARD_TYPE_DDR_1_8V means 3.3V or 1.8V vccq.
992 	 *
993 	 * 1.8V vccq at 3.3V core voltage (vcc) is not required
994 	 * in the JEDEC spec for DDR.
995 	 *
996 	 * Even (e)MMC card can support 3.3v to 1.2v vccq, but not all
997 	 * host controller can support this, like some of the SDHCI
998 	 * controller which connect to an eMMC device. Some of these
999 	 * host controller still needs to use 1.8v vccq for supporting
1000 	 * DDR mode.
1001 	 *
1002 	 * So the sequence will be:
1003 	 * if (host and device can both support 1.2v IO)
1004 	 *	use 1.2v IO;
1005 	 * else if (host and device can both support 1.8v IO)
1006 	 *	use 1.8v IO;
1007 	 * so if host and device can only support 3.3v IO, this is the
1008 	 * last choice.
1009 	 *
1010 	 * WARNING: eMMC rules are NOT the same as SD DDR
1011 	 */
1012 	err = -EINVAL;
1013 	if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_DDR_1_2V)
1014 		err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_120);
1015 
1016 	if (err && (card->mmc_avail_type & EXT_CSD_CARD_TYPE_DDR_1_8V))
1017 		err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_180);
1018 
1019 	/* make sure vccq is 3.3v after switching disaster */
1020 	if (err)
1021 		err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_330);
1022 
1023 	if (!err)
1024 		mmc_set_timing(host, MMC_TIMING_MMC_DDR52);
1025 
1026 	return err;
1027 }
1028 
1029 static int mmc_select_hs400(struct mmc_card *card)
1030 {
1031 	struct mmc_host *host = card->host;
1032 	int err = 0;
1033 
1034 	/*
1035 	 * HS400 mode requires 8-bit bus width
1036 	 */
1037 	if (!(card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS400 &&
1038 	      host->ios.bus_width == MMC_BUS_WIDTH_8))
1039 		return 0;
1040 
1041 	/*
1042 	 * Before switching to dual data rate operation for HS400,
1043 	 * it is required to convert from HS200 mode to HS mode.
1044 	 */
1045 	mmc_set_timing(card->host, MMC_TIMING_MMC_HS);
1046 	mmc_set_bus_speed(card);
1047 
1048 	err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
1049 			   EXT_CSD_HS_TIMING, EXT_CSD_TIMING_HS,
1050 			   card->ext_csd.generic_cmd6_time,
1051 			   true, true, true);
1052 	if (err) {
1053 		pr_err("%s: switch to high-speed from hs200 failed, err:%d\n",
1054 			mmc_hostname(host), err);
1055 		return err;
1056 	}
1057 
1058 	err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
1059 			 EXT_CSD_BUS_WIDTH,
1060 			 EXT_CSD_DDR_BUS_WIDTH_8,
1061 			 card->ext_csd.generic_cmd6_time);
1062 	if (err) {
1063 		pr_err("%s: switch to bus width for hs400 failed, err:%d\n",
1064 			mmc_hostname(host), err);
1065 		return err;
1066 	}
1067 
1068 	err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
1069 			   EXT_CSD_HS_TIMING, EXT_CSD_TIMING_HS400,
1070 			   card->ext_csd.generic_cmd6_time,
1071 			   true, true, true);
1072 	if (err) {
1073 		pr_err("%s: switch to hs400 failed, err:%d\n",
1074 			 mmc_hostname(host), err);
1075 		return err;
1076 	}
1077 
1078 	mmc_set_timing(host, MMC_TIMING_MMC_HS400);
1079 	mmc_set_bus_speed(card);
1080 
1081 	return 0;
1082 }
1083 
1084 /*
1085  * For device supporting HS200 mode, the following sequence
1086  * should be done before executing the tuning process.
1087  * 1. set the desired bus width(4-bit or 8-bit, 1-bit is not supported)
1088  * 2. switch to HS200 mode
1089  * 3. set the clock to > 52Mhz and <=200MHz
1090  */
1091 static int mmc_select_hs200(struct mmc_card *card)
1092 {
1093 	struct mmc_host *host = card->host;
1094 	int err = -EINVAL;
1095 
1096 	if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS200_1_2V)
1097 		err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_120);
1098 
1099 	if (err && card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS200_1_8V)
1100 		err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_180);
1101 
1102 	/* If fails try again during next card power cycle */
1103 	if (err)
1104 		goto err;
1105 
1106 	/*
1107 	 * Set the bus width(4 or 8) with host's support and
1108 	 * switch to HS200 mode if bus width is set successfully.
1109 	 */
1110 	err = mmc_select_bus_width(card);
1111 	if (!IS_ERR_VALUE(err)) {
1112 		err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
1113 				   EXT_CSD_HS_TIMING, EXT_CSD_TIMING_HS200,
1114 				   card->ext_csd.generic_cmd6_time,
1115 				   true, true, true);
1116 		if (!err)
1117 			mmc_set_timing(host, MMC_TIMING_MMC_HS200);
1118 	}
1119 err:
1120 	return err;
1121 }
1122 
1123 /*
1124  * Activate High Speed or HS200 mode if supported.
1125  */
1126 static int mmc_select_timing(struct mmc_card *card)
1127 {
1128 	int err = 0;
1129 
1130 	if (!mmc_can_ext_csd(card))
1131 		goto bus_speed;
1132 
1133 	if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS200)
1134 		err = mmc_select_hs200(card);
1135 	else if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS)
1136 		err = mmc_select_hs(card);
1137 
1138 	if (err && err != -EBADMSG)
1139 		return err;
1140 
1141 	if (err) {
1142 		pr_warn("%s: switch to %s failed\n",
1143 			mmc_card_hs(card) ? "high-speed" :
1144 			(mmc_card_hs200(card) ? "hs200" : ""),
1145 			mmc_hostname(card->host));
1146 		err = 0;
1147 	}
1148 
1149 bus_speed:
1150 	/*
1151 	 * Set the bus speed to the selected bus timing.
1152 	 * If timing is not selected, backward compatible is the default.
1153 	 */
1154 	mmc_set_bus_speed(card);
1155 	return err;
1156 }
1157 
1158 const u8 tuning_blk_pattern_4bit[MMC_TUNING_BLK_PATTERN_4BIT_SIZE] = {
1159 	0xff, 0x0f, 0xff, 0x00, 0xff, 0xcc, 0xc3, 0xcc,
1160 	0xc3, 0x3c, 0xcc, 0xff, 0xfe, 0xff, 0xfe, 0xef,
1161 	0xff, 0xdf, 0xff, 0xdd, 0xff, 0xfb, 0xff, 0xfb,
1162 	0xbf, 0xff, 0x7f, 0xff, 0x77, 0xf7, 0xbd, 0xef,
1163 	0xff, 0xf0, 0xff, 0xf0, 0x0f, 0xfc, 0xcc, 0x3c,
1164 	0xcc, 0x33, 0xcc, 0xcf, 0xff, 0xef, 0xff, 0xee,
1165 	0xff, 0xfd, 0xff, 0xfd, 0xdf, 0xff, 0xbf, 0xff,
1166 	0xbb, 0xff, 0xf7, 0xff, 0xf7, 0x7f, 0x7b, 0xde,
1167 };
1168 EXPORT_SYMBOL(tuning_blk_pattern_4bit);
1169 
1170 const u8 tuning_blk_pattern_8bit[MMC_TUNING_BLK_PATTERN_8BIT_SIZE] = {
1171 	0xff, 0xff, 0x00, 0xff, 0xff, 0xff, 0x00, 0x00,
1172 	0xff, 0xff, 0xcc, 0xcc, 0xcc, 0x33, 0xcc, 0xcc,
1173 	0xcc, 0x33, 0x33, 0xcc, 0xcc, 0xcc, 0xff, 0xff,
1174 	0xff, 0xee, 0xff, 0xff, 0xff, 0xee, 0xee, 0xff,
1175 	0xff, 0xff, 0xdd, 0xff, 0xff, 0xff, 0xdd, 0xdd,
1176 	0xff, 0xff, 0xff, 0xbb, 0xff, 0xff, 0xff, 0xbb,
1177 	0xbb, 0xff, 0xff, 0xff, 0x77, 0xff, 0xff, 0xff,
1178 	0x77, 0x77, 0xff, 0x77, 0xbb, 0xdd, 0xee, 0xff,
1179 	0xff, 0xff, 0xff, 0x00, 0xff, 0xff, 0xff, 0x00,
1180 	0x00, 0xff, 0xff, 0xcc, 0xcc, 0xcc, 0x33, 0xcc,
1181 	0xcc, 0xcc, 0x33, 0x33, 0xcc, 0xcc, 0xcc, 0xff,
1182 	0xff, 0xff, 0xee, 0xff, 0xff, 0xff, 0xee, 0xee,
1183 	0xff, 0xff, 0xff, 0xdd, 0xff, 0xff, 0xff, 0xdd,
1184 	0xdd, 0xff, 0xff, 0xff, 0xbb, 0xff, 0xff, 0xff,
1185 	0xbb, 0xbb, 0xff, 0xff, 0xff, 0x77, 0xff, 0xff,
1186 	0xff, 0x77, 0x77, 0xff, 0x77, 0xbb, 0xdd, 0xee,
1187 };
1188 EXPORT_SYMBOL(tuning_blk_pattern_8bit);
1189 
1190 /*
1191  * Execute tuning sequence to seek the proper bus operating
1192  * conditions for HS200 and HS400, which sends CMD21 to the device.
1193  */
1194 static int mmc_hs200_tuning(struct mmc_card *card)
1195 {
1196 	struct mmc_host *host = card->host;
1197 	int err = 0;
1198 
1199 	/*
1200 	 * Timing should be adjusted to the HS400 target
1201 	 * operation frequency for tuning process
1202 	 */
1203 	if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS400 &&
1204 	    host->ios.bus_width == MMC_BUS_WIDTH_8)
1205 		if (host->ops->prepare_hs400_tuning)
1206 			host->ops->prepare_hs400_tuning(host, &host->ios);
1207 
1208 	if (host->ops->execute_tuning) {
1209 		mmc_host_clk_hold(host);
1210 		err = host->ops->execute_tuning(host,
1211 				MMC_SEND_TUNING_BLOCK_HS200);
1212 		mmc_host_clk_release(host);
1213 
1214 		if (err)
1215 			pr_err("%s: tuning execution failed\n",
1216 				mmc_hostname(host));
1217 	}
1218 
1219 	return err;
1220 }
1221 
1222 /*
1223  * Handle the detection and initialisation of a card.
1224  *
1225  * In the case of a resume, "oldcard" will contain the card
1226  * we're trying to reinitialise.
1227  */
1228 static int mmc_init_card(struct mmc_host *host, u32 ocr,
1229 	struct mmc_card *oldcard)
1230 {
1231 	struct mmc_card *card;
1232 	int err;
1233 	u32 cid[4];
1234 	u32 rocr;
1235 
1236 	BUG_ON(!host);
1237 	WARN_ON(!host->claimed);
1238 
1239 	/* Set correct bus mode for MMC before attempting init */
1240 	if (!mmc_host_is_spi(host))
1241 		mmc_set_bus_mode(host, MMC_BUSMODE_OPENDRAIN);
1242 
1243 	/*
1244 	 * Since we're changing the OCR value, we seem to
1245 	 * need to tell some cards to go back to the idle
1246 	 * state.  We wait 1ms to give cards time to
1247 	 * respond.
1248 	 * mmc_go_idle is needed for eMMC that are asleep
1249 	 */
1250 	mmc_go_idle(host);
1251 
1252 	/* The extra bit indicates that we support high capacity */
1253 	err = mmc_send_op_cond(host, ocr | (1 << 30), &rocr);
1254 	if (err)
1255 		goto err;
1256 
1257 	/*
1258 	 * For SPI, enable CRC as appropriate.
1259 	 */
1260 	if (mmc_host_is_spi(host)) {
1261 		err = mmc_spi_set_crc(host, use_spi_crc);
1262 		if (err)
1263 			goto err;
1264 	}
1265 
1266 	/*
1267 	 * Fetch CID from card.
1268 	 */
1269 	if (mmc_host_is_spi(host))
1270 		err = mmc_send_cid(host, cid);
1271 	else
1272 		err = mmc_all_send_cid(host, cid);
1273 	if (err)
1274 		goto err;
1275 
1276 	if (oldcard) {
1277 		if (memcmp(cid, oldcard->raw_cid, sizeof(cid)) != 0) {
1278 			err = -ENOENT;
1279 			goto err;
1280 		}
1281 
1282 		card = oldcard;
1283 	} else {
1284 		/*
1285 		 * Allocate card structure.
1286 		 */
1287 		card = mmc_alloc_card(host, &mmc_type);
1288 		if (IS_ERR(card)) {
1289 			err = PTR_ERR(card);
1290 			goto err;
1291 		}
1292 
1293 		card->ocr = ocr;
1294 		card->type = MMC_TYPE_MMC;
1295 		card->rca = 1;
1296 		memcpy(card->raw_cid, cid, sizeof(card->raw_cid));
1297 	}
1298 
1299 	/*
1300 	 * For native busses:  set card RCA and quit open drain mode.
1301 	 */
1302 	if (!mmc_host_is_spi(host)) {
1303 		err = mmc_set_relative_addr(card);
1304 		if (err)
1305 			goto free_card;
1306 
1307 		mmc_set_bus_mode(host, MMC_BUSMODE_PUSHPULL);
1308 	}
1309 
1310 	if (!oldcard) {
1311 		/*
1312 		 * Fetch CSD from card.
1313 		 */
1314 		err = mmc_send_csd(card, card->raw_csd);
1315 		if (err)
1316 			goto free_card;
1317 
1318 		err = mmc_decode_csd(card);
1319 		if (err)
1320 			goto free_card;
1321 		err = mmc_decode_cid(card);
1322 		if (err)
1323 			goto free_card;
1324 	}
1325 
1326 	/*
1327 	 * handling only for cards supporting DSR and hosts requesting
1328 	 * DSR configuration
1329 	 */
1330 	if (card->csd.dsr_imp && host->dsr_req)
1331 		mmc_set_dsr(host);
1332 
1333 	/*
1334 	 * Select card, as all following commands rely on that.
1335 	 */
1336 	if (!mmc_host_is_spi(host)) {
1337 		err = mmc_select_card(card);
1338 		if (err)
1339 			goto free_card;
1340 	}
1341 
1342 	if (!oldcard) {
1343 		/* Read extended CSD. */
1344 		err = mmc_read_ext_csd(card);
1345 		if (err)
1346 			goto free_card;
1347 
1348 		/* If doing byte addressing, check if required to do sector
1349 		 * addressing.  Handle the case of <2GB cards needing sector
1350 		 * addressing.  See section 8.1 JEDEC Standard JED84-A441;
1351 		 * ocr register has bit 30 set for sector addressing.
1352 		 */
1353 		if (!(mmc_card_blockaddr(card)) && (rocr & (1<<30)))
1354 			mmc_card_set_blockaddr(card);
1355 
1356 		/* Erase size depends on CSD and Extended CSD */
1357 		mmc_set_erase_size(card);
1358 	}
1359 
1360 	/*
1361 	 * If enhanced_area_en is TRUE, host needs to enable ERASE_GRP_DEF
1362 	 * bit.  This bit will be lost every time after a reset or power off.
1363 	 */
1364 	if (card->ext_csd.partition_setting_completed ||
1365 	    (card->ext_csd.rev >= 3 && (host->caps2 & MMC_CAP2_HC_ERASE_SZ))) {
1366 		err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
1367 				 EXT_CSD_ERASE_GROUP_DEF, 1,
1368 				 card->ext_csd.generic_cmd6_time);
1369 
1370 		if (err && err != -EBADMSG)
1371 			goto free_card;
1372 
1373 		if (err) {
1374 			err = 0;
1375 			/*
1376 			 * Just disable enhanced area off & sz
1377 			 * will try to enable ERASE_GROUP_DEF
1378 			 * during next time reinit
1379 			 */
1380 			card->ext_csd.enhanced_area_offset = -EINVAL;
1381 			card->ext_csd.enhanced_area_size = -EINVAL;
1382 		} else {
1383 			card->ext_csd.erase_group_def = 1;
1384 			/*
1385 			 * enable ERASE_GRP_DEF successfully.
1386 			 * This will affect the erase size, so
1387 			 * here need to reset erase size
1388 			 */
1389 			mmc_set_erase_size(card);
1390 		}
1391 	}
1392 
1393 	/*
1394 	 * Ensure eMMC user default partition is enabled
1395 	 */
1396 	if (card->ext_csd.part_config & EXT_CSD_PART_CONFIG_ACC_MASK) {
1397 		card->ext_csd.part_config &= ~EXT_CSD_PART_CONFIG_ACC_MASK;
1398 		err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_PART_CONFIG,
1399 				 card->ext_csd.part_config,
1400 				 card->ext_csd.part_time);
1401 		if (err && err != -EBADMSG)
1402 			goto free_card;
1403 	}
1404 
1405 	/*
1406 	 * Enable power_off_notification byte in the ext_csd register
1407 	 */
1408 	if (card->ext_csd.rev >= 6) {
1409 		err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
1410 				 EXT_CSD_POWER_OFF_NOTIFICATION,
1411 				 EXT_CSD_POWER_ON,
1412 				 card->ext_csd.generic_cmd6_time);
1413 		if (err && err != -EBADMSG)
1414 			goto free_card;
1415 
1416 		/*
1417 		 * The err can be -EBADMSG or 0,
1418 		 * so check for success and update the flag
1419 		 */
1420 		if (!err)
1421 			card->ext_csd.power_off_notification = EXT_CSD_POWER_ON;
1422 	}
1423 
1424 	/*
1425 	 * Select timing interface
1426 	 */
1427 	err = mmc_select_timing(card);
1428 	if (err)
1429 		goto free_card;
1430 
1431 	if (mmc_card_hs200(card)) {
1432 		err = mmc_hs200_tuning(card);
1433 		if (err)
1434 			goto free_card;
1435 
1436 		err = mmc_select_hs400(card);
1437 		if (err)
1438 			goto free_card;
1439 	} else if (mmc_card_hs(card)) {
1440 		/* Select the desired bus width optionally */
1441 		err = mmc_select_bus_width(card);
1442 		if (!IS_ERR_VALUE(err)) {
1443 			err = mmc_select_hs_ddr(card);
1444 			if (err)
1445 				goto free_card;
1446 		}
1447 	}
1448 
1449 	/*
1450 	 * Choose the power class with selected bus interface
1451 	 */
1452 	mmc_select_powerclass(card);
1453 
1454 	/*
1455 	 * Enable HPI feature (if supported)
1456 	 */
1457 	if (card->ext_csd.hpi) {
1458 		err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
1459 				EXT_CSD_HPI_MGMT, 1,
1460 				card->ext_csd.generic_cmd6_time);
1461 		if (err && err != -EBADMSG)
1462 			goto free_card;
1463 		if (err) {
1464 			pr_warn("%s: Enabling HPI failed\n",
1465 				mmc_hostname(card->host));
1466 			err = 0;
1467 		} else
1468 			card->ext_csd.hpi_en = 1;
1469 	}
1470 
1471 	/*
1472 	 * If cache size is higher than 0, this indicates
1473 	 * the existence of cache and it can be turned on.
1474 	 */
1475 	if (card->ext_csd.cache_size > 0) {
1476 		err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
1477 				EXT_CSD_CACHE_CTRL, 1,
1478 				card->ext_csd.generic_cmd6_time);
1479 		if (err && err != -EBADMSG)
1480 			goto free_card;
1481 
1482 		/*
1483 		 * Only if no error, cache is turned on successfully.
1484 		 */
1485 		if (err) {
1486 			pr_warn("%s: Cache is supported, but failed to turn on (%d)\n",
1487 				mmc_hostname(card->host), err);
1488 			card->ext_csd.cache_ctrl = 0;
1489 			err = 0;
1490 		} else {
1491 			card->ext_csd.cache_ctrl = 1;
1492 		}
1493 	}
1494 
1495 	/*
1496 	 * The mandatory minimum values are defined for packed command.
1497 	 * read: 5, write: 3
1498 	 */
1499 	if (card->ext_csd.max_packed_writes >= 3 &&
1500 	    card->ext_csd.max_packed_reads >= 5 &&
1501 	    host->caps2 & MMC_CAP2_PACKED_CMD) {
1502 		err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
1503 				EXT_CSD_EXP_EVENTS_CTRL,
1504 				EXT_CSD_PACKED_EVENT_EN,
1505 				card->ext_csd.generic_cmd6_time);
1506 		if (err && err != -EBADMSG)
1507 			goto free_card;
1508 		if (err) {
1509 			pr_warn("%s: Enabling packed event failed\n",
1510 				mmc_hostname(card->host));
1511 			card->ext_csd.packed_event_en = 0;
1512 			err = 0;
1513 		} else {
1514 			card->ext_csd.packed_event_en = 1;
1515 		}
1516 	}
1517 
1518 	if (!oldcard)
1519 		host->card = card;
1520 
1521 	return 0;
1522 
1523 free_card:
1524 	if (!oldcard)
1525 		mmc_remove_card(card);
1526 err:
1527 	return err;
1528 }
1529 
1530 static int mmc_can_sleep(struct mmc_card *card)
1531 {
1532 	return (card && card->ext_csd.rev >= 3);
1533 }
1534 
1535 static int mmc_sleep(struct mmc_host *host)
1536 {
1537 	struct mmc_command cmd = {0};
1538 	struct mmc_card *card = host->card;
1539 	unsigned int timeout_ms = DIV_ROUND_UP(card->ext_csd.sa_timeout, 10000);
1540 	int err;
1541 
1542 	err = mmc_deselect_cards(host);
1543 	if (err)
1544 		return err;
1545 
1546 	cmd.opcode = MMC_SLEEP_AWAKE;
1547 	cmd.arg = card->rca << 16;
1548 	cmd.arg |= 1 << 15;
1549 
1550 	/*
1551 	 * If the max_busy_timeout of the host is specified, validate it against
1552 	 * the sleep cmd timeout. A failure means we need to prevent the host
1553 	 * from doing hw busy detection, which is done by converting to a R1
1554 	 * response instead of a R1B.
1555 	 */
1556 	if (host->max_busy_timeout && (timeout_ms > host->max_busy_timeout)) {
1557 		cmd.flags = MMC_RSP_R1 | MMC_CMD_AC;
1558 	} else {
1559 		cmd.flags = MMC_RSP_R1B | MMC_CMD_AC;
1560 		cmd.busy_timeout = timeout_ms;
1561 	}
1562 
1563 	err = mmc_wait_for_cmd(host, &cmd, 0);
1564 	if (err)
1565 		return err;
1566 
1567 	/*
1568 	 * If the host does not wait while the card signals busy, then we will
1569 	 * will have to wait the sleep/awake timeout.  Note, we cannot use the
1570 	 * SEND_STATUS command to poll the status because that command (and most
1571 	 * others) is invalid while the card sleeps.
1572 	 */
1573 	if (!cmd.busy_timeout || !(host->caps & MMC_CAP_WAIT_WHILE_BUSY))
1574 		mmc_delay(timeout_ms);
1575 
1576 	return err;
1577 }
1578 
1579 static int mmc_can_poweroff_notify(const struct mmc_card *card)
1580 {
1581 	return card &&
1582 		mmc_card_mmc(card) &&
1583 		(card->ext_csd.power_off_notification == EXT_CSD_POWER_ON);
1584 }
1585 
1586 static int mmc_poweroff_notify(struct mmc_card *card, unsigned int notify_type)
1587 {
1588 	unsigned int timeout = card->ext_csd.generic_cmd6_time;
1589 	int err;
1590 
1591 	/* Use EXT_CSD_POWER_OFF_SHORT as default notification type. */
1592 	if (notify_type == EXT_CSD_POWER_OFF_LONG)
1593 		timeout = card->ext_csd.power_off_longtime;
1594 
1595 	err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
1596 			EXT_CSD_POWER_OFF_NOTIFICATION,
1597 			notify_type, timeout, true, false, false);
1598 	if (err)
1599 		pr_err("%s: Power Off Notification timed out, %u\n",
1600 		       mmc_hostname(card->host), timeout);
1601 
1602 	/* Disable the power off notification after the switch operation. */
1603 	card->ext_csd.power_off_notification = EXT_CSD_NO_POWER_NOTIFICATION;
1604 
1605 	return err;
1606 }
1607 
1608 /*
1609  * Host is being removed. Free up the current card.
1610  */
1611 static void mmc_remove(struct mmc_host *host)
1612 {
1613 	BUG_ON(!host);
1614 	BUG_ON(!host->card);
1615 
1616 	mmc_remove_card(host->card);
1617 	host->card = NULL;
1618 }
1619 
1620 /*
1621  * Card detection - card is alive.
1622  */
1623 static int mmc_alive(struct mmc_host *host)
1624 {
1625 	return mmc_send_status(host->card, NULL);
1626 }
1627 
1628 /*
1629  * Card detection callback from host.
1630  */
1631 static void mmc_detect(struct mmc_host *host)
1632 {
1633 	int err;
1634 
1635 	BUG_ON(!host);
1636 	BUG_ON(!host->card);
1637 
1638 	mmc_get_card(host->card);
1639 
1640 	/*
1641 	 * Just check if our card has been removed.
1642 	 */
1643 	err = _mmc_detect_card_removed(host);
1644 
1645 	mmc_put_card(host->card);
1646 
1647 	if (err) {
1648 		mmc_remove(host);
1649 
1650 		mmc_claim_host(host);
1651 		mmc_detach_bus(host);
1652 		mmc_power_off(host);
1653 		mmc_release_host(host);
1654 	}
1655 }
1656 
1657 static int _mmc_suspend(struct mmc_host *host, bool is_suspend)
1658 {
1659 	int err = 0;
1660 	unsigned int notify_type = is_suspend ? EXT_CSD_POWER_OFF_SHORT :
1661 					EXT_CSD_POWER_OFF_LONG;
1662 
1663 	BUG_ON(!host);
1664 	BUG_ON(!host->card);
1665 
1666 	mmc_claim_host(host);
1667 
1668 	if (mmc_card_suspended(host->card))
1669 		goto out;
1670 
1671 	if (mmc_card_doing_bkops(host->card)) {
1672 		err = mmc_stop_bkops(host->card);
1673 		if (err)
1674 			goto out;
1675 	}
1676 
1677 	err = mmc_flush_cache(host->card);
1678 	if (err)
1679 		goto out;
1680 
1681 	if (mmc_can_poweroff_notify(host->card) &&
1682 		((host->caps2 & MMC_CAP2_FULL_PWR_CYCLE) || !is_suspend))
1683 		err = mmc_poweroff_notify(host->card, notify_type);
1684 	else if (mmc_can_sleep(host->card))
1685 		err = mmc_sleep(host);
1686 	else if (!mmc_host_is_spi(host))
1687 		err = mmc_deselect_cards(host);
1688 
1689 	if (!err) {
1690 		mmc_power_off(host);
1691 		mmc_card_set_suspended(host->card);
1692 	}
1693 out:
1694 	mmc_release_host(host);
1695 	return err;
1696 }
1697 
1698 /*
1699  * Suspend callback
1700  */
1701 static int mmc_suspend(struct mmc_host *host)
1702 {
1703 	int err;
1704 
1705 	err = _mmc_suspend(host, true);
1706 	if (!err) {
1707 		pm_runtime_disable(&host->card->dev);
1708 		pm_runtime_set_suspended(&host->card->dev);
1709 	}
1710 
1711 	return err;
1712 }
1713 
1714 /*
1715  * This function tries to determine if the same card is still present
1716  * and, if so, restore all state to it.
1717  */
1718 static int _mmc_resume(struct mmc_host *host)
1719 {
1720 	int err = 0;
1721 
1722 	BUG_ON(!host);
1723 	BUG_ON(!host->card);
1724 
1725 	mmc_claim_host(host);
1726 
1727 	if (!mmc_card_suspended(host->card))
1728 		goto out;
1729 
1730 	mmc_power_up(host, host->card->ocr);
1731 	err = mmc_init_card(host, host->card->ocr, host->card);
1732 	mmc_card_clr_suspended(host->card);
1733 
1734 out:
1735 	mmc_release_host(host);
1736 	return err;
1737 }
1738 
1739 /*
1740  * Shutdown callback
1741  */
1742 static int mmc_shutdown(struct mmc_host *host)
1743 {
1744 	int err = 0;
1745 
1746 	/*
1747 	 * In a specific case for poweroff notify, we need to resume the card
1748 	 * before we can shutdown it properly.
1749 	 */
1750 	if (mmc_can_poweroff_notify(host->card) &&
1751 		!(host->caps2 & MMC_CAP2_FULL_PWR_CYCLE))
1752 		err = _mmc_resume(host);
1753 
1754 	if (!err)
1755 		err = _mmc_suspend(host, false);
1756 
1757 	return err;
1758 }
1759 
1760 /*
1761  * Callback for resume.
1762  */
1763 static int mmc_resume(struct mmc_host *host)
1764 {
1765 	int err = 0;
1766 
1767 	if (!(host->caps & MMC_CAP_RUNTIME_RESUME)) {
1768 		err = _mmc_resume(host);
1769 		pm_runtime_set_active(&host->card->dev);
1770 		pm_runtime_mark_last_busy(&host->card->dev);
1771 	}
1772 	pm_runtime_enable(&host->card->dev);
1773 
1774 	return err;
1775 }
1776 
1777 /*
1778  * Callback for runtime_suspend.
1779  */
1780 static int mmc_runtime_suspend(struct mmc_host *host)
1781 {
1782 	int err;
1783 
1784 	if (!(host->caps & MMC_CAP_AGGRESSIVE_PM))
1785 		return 0;
1786 
1787 	err = _mmc_suspend(host, true);
1788 	if (err)
1789 		pr_err("%s: error %d doing aggessive suspend\n",
1790 			mmc_hostname(host), err);
1791 
1792 	return err;
1793 }
1794 
1795 /*
1796  * Callback for runtime_resume.
1797  */
1798 static int mmc_runtime_resume(struct mmc_host *host)
1799 {
1800 	int err;
1801 
1802 	if (!(host->caps & (MMC_CAP_AGGRESSIVE_PM | MMC_CAP_RUNTIME_RESUME)))
1803 		return 0;
1804 
1805 	err = _mmc_resume(host);
1806 	if (err)
1807 		pr_err("%s: error %d doing aggessive resume\n",
1808 			mmc_hostname(host), err);
1809 
1810 	return 0;
1811 }
1812 
1813 static int mmc_power_restore(struct mmc_host *host)
1814 {
1815 	int ret;
1816 
1817 	mmc_claim_host(host);
1818 	ret = mmc_init_card(host, host->card->ocr, host->card);
1819 	mmc_release_host(host);
1820 
1821 	return ret;
1822 }
1823 
1824 static const struct mmc_bus_ops mmc_ops = {
1825 	.remove = mmc_remove,
1826 	.detect = mmc_detect,
1827 	.suspend = mmc_suspend,
1828 	.resume = mmc_resume,
1829 	.runtime_suspend = mmc_runtime_suspend,
1830 	.runtime_resume = mmc_runtime_resume,
1831 	.power_restore = mmc_power_restore,
1832 	.alive = mmc_alive,
1833 	.shutdown = mmc_shutdown,
1834 };
1835 
1836 /*
1837  * Starting point for MMC card init.
1838  */
1839 int mmc_attach_mmc(struct mmc_host *host)
1840 {
1841 	int err;
1842 	u32 ocr, rocr;
1843 
1844 	BUG_ON(!host);
1845 	WARN_ON(!host->claimed);
1846 
1847 	/* Set correct bus mode for MMC before attempting attach */
1848 	if (!mmc_host_is_spi(host))
1849 		mmc_set_bus_mode(host, MMC_BUSMODE_OPENDRAIN);
1850 
1851 	err = mmc_send_op_cond(host, 0, &ocr);
1852 	if (err)
1853 		return err;
1854 
1855 	mmc_attach_bus(host, &mmc_ops);
1856 	if (host->ocr_avail_mmc)
1857 		host->ocr_avail = host->ocr_avail_mmc;
1858 
1859 	/*
1860 	 * We need to get OCR a different way for SPI.
1861 	 */
1862 	if (mmc_host_is_spi(host)) {
1863 		err = mmc_spi_read_ocr(host, 1, &ocr);
1864 		if (err)
1865 			goto err;
1866 	}
1867 
1868 	rocr = mmc_select_voltage(host, ocr);
1869 
1870 	/*
1871 	 * Can we support the voltage of the card?
1872 	 */
1873 	if (!rocr) {
1874 		err = -EINVAL;
1875 		goto err;
1876 	}
1877 
1878 	/*
1879 	 * Detect and init the card.
1880 	 */
1881 	err = mmc_init_card(host, rocr, NULL);
1882 	if (err)
1883 		goto err;
1884 
1885 	mmc_release_host(host);
1886 	err = mmc_add_card(host->card);
1887 	mmc_claim_host(host);
1888 	if (err)
1889 		goto remove_card;
1890 
1891 	return 0;
1892 
1893 remove_card:
1894 	mmc_release_host(host);
1895 	mmc_remove_card(host->card);
1896 	mmc_claim_host(host);
1897 	host->card = NULL;
1898 err:
1899 	mmc_detach_bus(host);
1900 
1901 	pr_err("%s: error %d whilst initialising MMC card\n",
1902 		mmc_hostname(host), err);
1903 
1904 	return err;
1905 }
1906