xref: /openbmc/linux/drivers/bus/ti-sysc.c (revision 060f35a317ef09101b128f399dce7ed13d019461)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * ti-sysc.c - Texas Instruments sysc interconnect target driver
4  */
5 
6 #include <linux/io.h>
7 #include <linux/clk.h>
8 #include <linux/clkdev.h>
9 #include <linux/cpu_pm.h>
10 #include <linux/delay.h>
11 #include <linux/list.h>
12 #include <linux/module.h>
13 #include <linux/platform_device.h>
14 #include <linux/pm_domain.h>
15 #include <linux/pm_runtime.h>
16 #include <linux/reset.h>
17 #include <linux/of_address.h>
18 #include <linux/of_platform.h>
19 #include <linux/slab.h>
20 #include <linux/sys_soc.h>
21 #include <linux/timekeeping.h>
22 #include <linux/iopoll.h>
23 
24 #include <linux/platform_data/ti-sysc.h>
25 
26 #include <dt-bindings/bus/ti-sysc.h>
27 
28 #define DIS_ISP		BIT(2)
29 #define DIS_IVA		BIT(1)
30 #define DIS_SGX		BIT(0)
31 
32 #define SOC_FLAG(match, flag)	{ .machine = match, .data = (void *)(flag), }
33 
34 #define MAX_MODULE_SOFTRESET_WAIT		10000
35 
36 enum sysc_soc {
37 	SOC_UNKNOWN,
38 	SOC_2420,
39 	SOC_2430,
40 	SOC_3430,
41 	SOC_AM35,
42 	SOC_3630,
43 	SOC_4430,
44 	SOC_4460,
45 	SOC_4470,
46 	SOC_5430,
47 	SOC_AM3,
48 	SOC_AM4,
49 	SOC_DRA7,
50 };
51 
52 struct sysc_address {
53 	unsigned long base;
54 	struct list_head node;
55 };
56 
57 struct sysc_module {
58 	struct sysc *ddata;
59 	struct list_head node;
60 };
61 
62 struct sysc_soc_info {
63 	unsigned long general_purpose:1;
64 	enum sysc_soc soc;
65 	struct mutex list_lock;	/* disabled and restored modules list lock */
66 	struct list_head disabled_modules;
67 	struct list_head restored_modules;
68 	struct notifier_block nb;
69 };
70 
71 enum sysc_clocks {
72 	SYSC_FCK,
73 	SYSC_ICK,
74 	SYSC_OPTFCK0,
75 	SYSC_OPTFCK1,
76 	SYSC_OPTFCK2,
77 	SYSC_OPTFCK3,
78 	SYSC_OPTFCK4,
79 	SYSC_OPTFCK5,
80 	SYSC_OPTFCK6,
81 	SYSC_OPTFCK7,
82 	SYSC_MAX_CLOCKS,
83 };
84 
85 static struct sysc_soc_info *sysc_soc;
86 static const char * const reg_names[] = { "rev", "sysc", "syss", };
87 static const char * const clock_names[SYSC_MAX_CLOCKS] = {
88 	"fck", "ick", "opt0", "opt1", "opt2", "opt3", "opt4",
89 	"opt5", "opt6", "opt7",
90 };
91 
92 #define SYSC_IDLEMODE_MASK		3
93 #define SYSC_CLOCKACTIVITY_MASK		3
94 
95 /**
96  * struct sysc - TI sysc interconnect target module registers and capabilities
97  * @dev: struct device pointer
98  * @module_pa: physical address of the interconnect target module
99  * @module_size: size of the interconnect target module
100  * @module_va: virtual address of the interconnect target module
101  * @offsets: register offsets from module base
102  * @mdata: ti-sysc to hwmod translation data for a module
103  * @clocks: clocks used by the interconnect target module
104  * @clock_roles: clock role names for the found clocks
105  * @nr_clocks: number of clocks used by the interconnect target module
106  * @rsts: resets used by the interconnect target module
107  * @legacy_mode: configured for legacy mode if set
108  * @cap: interconnect target module capabilities
109  * @cfg: interconnect target module configuration
110  * @cookie: data used by legacy platform callbacks
111  * @name: name if available
112  * @revision: interconnect target module revision
113  * @sysconfig: saved sysconfig register value
114  * @reserved: target module is reserved and already in use
115  * @enabled: sysc runtime enabled status
116  * @needs_resume: runtime resume needed on resume from suspend
117  * @child_needs_resume: runtime resume needed for child on resume from suspend
118  * @disable_on_idle: status flag used for disabling modules with resets
119  * @idle_work: work structure used to perform delayed idle on a module
120  * @pre_reset_quirk: module specific pre-reset quirk
121  * @post_reset_quirk: module specific post-reset quirk
122  * @reset_done_quirk: module specific reset done quirk
123  * @module_enable_quirk: module specific enable quirk
124  * @module_disable_quirk: module specific disable quirk
125  * @module_unlock_quirk: module specific sysconfig unlock quirk
126  * @module_lock_quirk: module specific sysconfig lock quirk
127  */
128 struct sysc {
129 	struct device *dev;
130 	u64 module_pa;
131 	u32 module_size;
132 	void __iomem *module_va;
133 	int offsets[SYSC_MAX_REGS];
134 	struct ti_sysc_module_data *mdata;
135 	struct clk **clocks;
136 	const char **clock_roles;
137 	int nr_clocks;
138 	struct reset_control *rsts;
139 	const char *legacy_mode;
140 	const struct sysc_capabilities *cap;
141 	struct sysc_config cfg;
142 	struct ti_sysc_cookie cookie;
143 	const char *name;
144 	u32 revision;
145 	u32 sysconfig;
146 	unsigned int reserved:1;
147 	unsigned int enabled:1;
148 	unsigned int needs_resume:1;
149 	unsigned int child_needs_resume:1;
150 	struct delayed_work idle_work;
151 	void (*pre_reset_quirk)(struct sysc *sysc);
152 	void (*post_reset_quirk)(struct sysc *sysc);
153 	void (*reset_done_quirk)(struct sysc *sysc);
154 	void (*module_enable_quirk)(struct sysc *sysc);
155 	void (*module_disable_quirk)(struct sysc *sysc);
156 	void (*module_unlock_quirk)(struct sysc *sysc);
157 	void (*module_lock_quirk)(struct sysc *sysc);
158 };
159 
160 static void sysc_parse_dts_quirks(struct sysc *ddata, struct device_node *np,
161 				  bool is_child);
162 static int sysc_reset(struct sysc *ddata);
163 
sysc_write(struct sysc * ddata,int offset,u32 value)164 static void sysc_write(struct sysc *ddata, int offset, u32 value)
165 {
166 	if (ddata->cfg.quirks & SYSC_QUIRK_16BIT) {
167 		writew_relaxed(value & 0xffff, ddata->module_va + offset);
168 
169 		/* Only i2c revision has LO and HI register with stride of 4 */
170 		if (ddata->offsets[SYSC_REVISION] >= 0 &&
171 		    offset == ddata->offsets[SYSC_REVISION]) {
172 			u16 hi = value >> 16;
173 
174 			writew_relaxed(hi, ddata->module_va + offset + 4);
175 		}
176 
177 		return;
178 	}
179 
180 	writel_relaxed(value, ddata->module_va + offset);
181 }
182 
sysc_read(struct sysc * ddata,int offset)183 static u32 sysc_read(struct sysc *ddata, int offset)
184 {
185 	if (ddata->cfg.quirks & SYSC_QUIRK_16BIT) {
186 		u32 val;
187 
188 		val = readw_relaxed(ddata->module_va + offset);
189 
190 		/* Only i2c revision has LO and HI register with stride of 4 */
191 		if (ddata->offsets[SYSC_REVISION] >= 0 &&
192 		    offset == ddata->offsets[SYSC_REVISION]) {
193 			u16 tmp = readw_relaxed(ddata->module_va + offset + 4);
194 
195 			val |= tmp << 16;
196 		}
197 
198 		return val;
199 	}
200 
201 	return readl_relaxed(ddata->module_va + offset);
202 }
203 
sysc_opt_clks_needed(struct sysc * ddata)204 static bool sysc_opt_clks_needed(struct sysc *ddata)
205 {
206 	return !!(ddata->cfg.quirks & SYSC_QUIRK_OPT_CLKS_NEEDED);
207 }
208 
sysc_read_revision(struct sysc * ddata)209 static u32 sysc_read_revision(struct sysc *ddata)
210 {
211 	int offset = ddata->offsets[SYSC_REVISION];
212 
213 	if (offset < 0)
214 		return 0;
215 
216 	return sysc_read(ddata, offset);
217 }
218 
sysc_read_sysconfig(struct sysc * ddata)219 static u32 sysc_read_sysconfig(struct sysc *ddata)
220 {
221 	int offset = ddata->offsets[SYSC_SYSCONFIG];
222 
223 	if (offset < 0)
224 		return 0;
225 
226 	return sysc_read(ddata, offset);
227 }
228 
sysc_read_sysstatus(struct sysc * ddata)229 static u32 sysc_read_sysstatus(struct sysc *ddata)
230 {
231 	int offset = ddata->offsets[SYSC_SYSSTATUS];
232 
233 	if (offset < 0)
234 		return 0;
235 
236 	return sysc_read(ddata, offset);
237 }
238 
sysc_poll_reset_sysstatus(struct sysc * ddata)239 static int sysc_poll_reset_sysstatus(struct sysc *ddata)
240 {
241 	int error, retries;
242 	u32 syss_done, rstval;
243 
244 	if (ddata->cfg.quirks & SYSS_QUIRK_RESETDONE_INVERTED)
245 		syss_done = 0;
246 	else
247 		syss_done = ddata->cfg.syss_mask;
248 
249 	if (likely(!timekeeping_suspended)) {
250 		error = readx_poll_timeout_atomic(sysc_read_sysstatus, ddata,
251 				rstval, (rstval & ddata->cfg.syss_mask) ==
252 				syss_done, 100, MAX_MODULE_SOFTRESET_WAIT);
253 	} else {
254 		retries = MAX_MODULE_SOFTRESET_WAIT;
255 		while (retries--) {
256 			rstval = sysc_read_sysstatus(ddata);
257 			if ((rstval & ddata->cfg.syss_mask) == syss_done)
258 				return 0;
259 			udelay(2); /* Account for udelay flakeyness */
260 		}
261 		error = -ETIMEDOUT;
262 	}
263 
264 	return error;
265 }
266 
sysc_poll_reset_sysconfig(struct sysc * ddata)267 static int sysc_poll_reset_sysconfig(struct sysc *ddata)
268 {
269 	int error, retries;
270 	u32 sysc_mask, rstval;
271 
272 	sysc_mask = BIT(ddata->cap->regbits->srst_shift);
273 
274 	if (likely(!timekeeping_suspended)) {
275 		error = readx_poll_timeout_atomic(sysc_read_sysconfig, ddata,
276 				rstval, !(rstval & sysc_mask),
277 				100, MAX_MODULE_SOFTRESET_WAIT);
278 	} else {
279 		retries = MAX_MODULE_SOFTRESET_WAIT;
280 		while (retries--) {
281 			rstval = sysc_read_sysconfig(ddata);
282 			if (!(rstval & sysc_mask))
283 				return 0;
284 			udelay(2); /* Account for udelay flakeyness */
285 		}
286 		error = -ETIMEDOUT;
287 	}
288 
289 	return error;
290 }
291 
292 /* Poll on reset status */
sysc_wait_softreset(struct sysc * ddata)293 static int sysc_wait_softreset(struct sysc *ddata)
294 {
295 	int syss_offset, error = 0;
296 
297 	if (ddata->cap->regbits->srst_shift < 0)
298 		return 0;
299 
300 	syss_offset = ddata->offsets[SYSC_SYSSTATUS];
301 
302 	if (syss_offset >= 0)
303 		error = sysc_poll_reset_sysstatus(ddata);
304 	else if (ddata->cfg.quirks & SYSC_QUIRK_RESET_STATUS)
305 		error = sysc_poll_reset_sysconfig(ddata);
306 
307 	return error;
308 }
309 
sysc_add_named_clock_from_child(struct sysc * ddata,const char * name,const char * optfck_name)310 static int sysc_add_named_clock_from_child(struct sysc *ddata,
311 					   const char *name,
312 					   const char *optfck_name)
313 {
314 	struct device_node *np = ddata->dev->of_node;
315 	struct device_node *child;
316 	struct clk_lookup *cl;
317 	struct clk *clock;
318 	const char *n;
319 
320 	if (name)
321 		n = name;
322 	else
323 		n = optfck_name;
324 
325 	/* Does the clock alias already exist? */
326 	clock = of_clk_get_by_name(np, n);
327 	if (!IS_ERR(clock)) {
328 		clk_put(clock);
329 
330 		return 0;
331 	}
332 
333 	child = of_get_next_available_child(np, NULL);
334 	if (!child)
335 		return -ENODEV;
336 
337 	clock = devm_get_clk_from_child(ddata->dev, child, name);
338 	if (IS_ERR(clock))
339 		return PTR_ERR(clock);
340 
341 	/*
342 	 * Use clkdev_add() instead of clkdev_alloc() to avoid the MAX_DEV_ID
343 	 * limit for clk_get(). If cl ever needs to be freed, it should be done
344 	 * with clkdev_drop().
345 	 */
346 	cl = kzalloc(sizeof(*cl), GFP_KERNEL);
347 	if (!cl)
348 		return -ENOMEM;
349 
350 	cl->con_id = n;
351 	cl->dev_id = dev_name(ddata->dev);
352 	cl->clk = clock;
353 	clkdev_add(cl);
354 
355 	clk_put(clock);
356 
357 	return 0;
358 }
359 
sysc_init_ext_opt_clock(struct sysc * ddata,const char * name)360 static int sysc_init_ext_opt_clock(struct sysc *ddata, const char *name)
361 {
362 	const char *optfck_name;
363 	int error, index;
364 
365 	if (ddata->nr_clocks < SYSC_OPTFCK0)
366 		index = SYSC_OPTFCK0;
367 	else
368 		index = ddata->nr_clocks;
369 
370 	if (name)
371 		optfck_name = name;
372 	else
373 		optfck_name = clock_names[index];
374 
375 	error = sysc_add_named_clock_from_child(ddata, name, optfck_name);
376 	if (error)
377 		return error;
378 
379 	ddata->clock_roles[index] = optfck_name;
380 	ddata->nr_clocks++;
381 
382 	return 0;
383 }
384 
sysc_get_one_clock(struct sysc * ddata,const char * name)385 static int sysc_get_one_clock(struct sysc *ddata, const char *name)
386 {
387 	int error, i, index = -ENODEV;
388 
389 	if (!strncmp(clock_names[SYSC_FCK], name, 3))
390 		index = SYSC_FCK;
391 	else if (!strncmp(clock_names[SYSC_ICK], name, 3))
392 		index = SYSC_ICK;
393 
394 	if (index < 0) {
395 		for (i = SYSC_OPTFCK0; i < SYSC_MAX_CLOCKS; i++) {
396 			if (!ddata->clocks[i]) {
397 				index = i;
398 				break;
399 			}
400 		}
401 	}
402 
403 	if (index < 0) {
404 		dev_err(ddata->dev, "clock %s not added\n", name);
405 		return index;
406 	}
407 
408 	ddata->clocks[index] = devm_clk_get(ddata->dev, name);
409 	if (IS_ERR(ddata->clocks[index])) {
410 		dev_err(ddata->dev, "clock get error for %s: %li\n",
411 			name, PTR_ERR(ddata->clocks[index]));
412 
413 		return PTR_ERR(ddata->clocks[index]);
414 	}
415 
416 	error = clk_prepare(ddata->clocks[index]);
417 	if (error) {
418 		dev_err(ddata->dev, "clock prepare error for %s: %i\n",
419 			name, error);
420 
421 		return error;
422 	}
423 
424 	return 0;
425 }
426 
sysc_get_clocks(struct sysc * ddata)427 static int sysc_get_clocks(struct sysc *ddata)
428 {
429 	struct device_node *np = ddata->dev->of_node;
430 	struct property *prop;
431 	const char *name;
432 	int nr_fck = 0, nr_ick = 0, i, error = 0;
433 
434 	ddata->clock_roles = devm_kcalloc(ddata->dev,
435 					  SYSC_MAX_CLOCKS,
436 					  sizeof(*ddata->clock_roles),
437 					  GFP_KERNEL);
438 	if (!ddata->clock_roles)
439 		return -ENOMEM;
440 
441 	of_property_for_each_string(np, "clock-names", prop, name) {
442 		if (!strncmp(clock_names[SYSC_FCK], name, 3))
443 			nr_fck++;
444 		if (!strncmp(clock_names[SYSC_ICK], name, 3))
445 			nr_ick++;
446 		ddata->clock_roles[ddata->nr_clocks] = name;
447 		ddata->nr_clocks++;
448 	}
449 
450 	if (ddata->nr_clocks < 1)
451 		return 0;
452 
453 	if ((ddata->cfg.quirks & SYSC_QUIRK_EXT_OPT_CLOCK)) {
454 		error = sysc_init_ext_opt_clock(ddata, NULL);
455 		if (error)
456 			return error;
457 	}
458 
459 	if (ddata->nr_clocks > SYSC_MAX_CLOCKS) {
460 		dev_err(ddata->dev, "too many clocks for %pOF\n", np);
461 
462 		return -EINVAL;
463 	}
464 
465 	if (nr_fck > 1 || nr_ick > 1) {
466 		dev_err(ddata->dev, "max one fck and ick for %pOF\n", np);
467 
468 		return -EINVAL;
469 	}
470 
471 	/* Always add a slot for main clocks fck and ick even if unused */
472 	if (!nr_fck)
473 		ddata->nr_clocks++;
474 	if (!nr_ick)
475 		ddata->nr_clocks++;
476 
477 	ddata->clocks = devm_kcalloc(ddata->dev,
478 				     ddata->nr_clocks, sizeof(*ddata->clocks),
479 				     GFP_KERNEL);
480 	if (!ddata->clocks)
481 		return -ENOMEM;
482 
483 	for (i = 0; i < SYSC_MAX_CLOCKS; i++) {
484 		const char *name = ddata->clock_roles[i];
485 
486 		if (!name)
487 			continue;
488 
489 		error = sysc_get_one_clock(ddata, name);
490 		if (error)
491 			return error;
492 	}
493 
494 	return 0;
495 }
496 
sysc_enable_main_clocks(struct sysc * ddata)497 static int sysc_enable_main_clocks(struct sysc *ddata)
498 {
499 	struct clk *clock;
500 	int i, error;
501 
502 	if (!ddata->clocks)
503 		return 0;
504 
505 	for (i = 0; i < SYSC_OPTFCK0; i++) {
506 		clock = ddata->clocks[i];
507 
508 		/* Main clocks may not have ick */
509 		if (IS_ERR_OR_NULL(clock))
510 			continue;
511 
512 		error = clk_enable(clock);
513 		if (error)
514 			goto err_disable;
515 	}
516 
517 	return 0;
518 
519 err_disable:
520 	for (i--; i >= 0; i--) {
521 		clock = ddata->clocks[i];
522 
523 		/* Main clocks may not have ick */
524 		if (IS_ERR_OR_NULL(clock))
525 			continue;
526 
527 		clk_disable(clock);
528 	}
529 
530 	return error;
531 }
532 
sysc_disable_main_clocks(struct sysc * ddata)533 static void sysc_disable_main_clocks(struct sysc *ddata)
534 {
535 	struct clk *clock;
536 	int i;
537 
538 	if (!ddata->clocks)
539 		return;
540 
541 	for (i = 0; i < SYSC_OPTFCK0; i++) {
542 		clock = ddata->clocks[i];
543 		if (IS_ERR_OR_NULL(clock))
544 			continue;
545 
546 		clk_disable(clock);
547 	}
548 }
549 
sysc_enable_opt_clocks(struct sysc * ddata)550 static int sysc_enable_opt_clocks(struct sysc *ddata)
551 {
552 	struct clk *clock;
553 	int i, error;
554 
555 	if (!ddata->clocks || ddata->nr_clocks < SYSC_OPTFCK0 + 1)
556 		return 0;
557 
558 	for (i = SYSC_OPTFCK0; i < SYSC_MAX_CLOCKS; i++) {
559 		clock = ddata->clocks[i];
560 
561 		/* Assume no holes for opt clocks */
562 		if (IS_ERR_OR_NULL(clock))
563 			return 0;
564 
565 		error = clk_enable(clock);
566 		if (error)
567 			goto err_disable;
568 	}
569 
570 	return 0;
571 
572 err_disable:
573 	for (i--; i >= 0; i--) {
574 		clock = ddata->clocks[i];
575 		if (IS_ERR_OR_NULL(clock))
576 			continue;
577 
578 		clk_disable(clock);
579 	}
580 
581 	return error;
582 }
583 
sysc_disable_opt_clocks(struct sysc * ddata)584 static void sysc_disable_opt_clocks(struct sysc *ddata)
585 {
586 	struct clk *clock;
587 	int i;
588 
589 	if (!ddata->clocks || ddata->nr_clocks < SYSC_OPTFCK0 + 1)
590 		return;
591 
592 	for (i = SYSC_OPTFCK0; i < SYSC_MAX_CLOCKS; i++) {
593 		clock = ddata->clocks[i];
594 
595 		/* Assume no holes for opt clocks */
596 		if (IS_ERR_OR_NULL(clock))
597 			return;
598 
599 		clk_disable(clock);
600 	}
601 }
602 
sysc_clkdm_deny_idle(struct sysc * ddata)603 static void sysc_clkdm_deny_idle(struct sysc *ddata)
604 {
605 	struct ti_sysc_platform_data *pdata;
606 
607 	if (ddata->legacy_mode || (ddata->cfg.quirks & SYSC_QUIRK_CLKDM_NOAUTO))
608 		return;
609 
610 	pdata = dev_get_platdata(ddata->dev);
611 	if (pdata && pdata->clkdm_deny_idle)
612 		pdata->clkdm_deny_idle(ddata->dev, &ddata->cookie);
613 }
614 
sysc_clkdm_allow_idle(struct sysc * ddata)615 static void sysc_clkdm_allow_idle(struct sysc *ddata)
616 {
617 	struct ti_sysc_platform_data *pdata;
618 
619 	if (ddata->legacy_mode || (ddata->cfg.quirks & SYSC_QUIRK_CLKDM_NOAUTO))
620 		return;
621 
622 	pdata = dev_get_platdata(ddata->dev);
623 	if (pdata && pdata->clkdm_allow_idle)
624 		pdata->clkdm_allow_idle(ddata->dev, &ddata->cookie);
625 }
626 
627 /**
628  * sysc_init_resets - init rstctrl reset line if configured
629  * @ddata: device driver data
630  *
631  * See sysc_rstctrl_reset_deassert().
632  */
sysc_init_resets(struct sysc * ddata)633 static int sysc_init_resets(struct sysc *ddata)
634 {
635 	ddata->rsts =
636 		devm_reset_control_get_optional_shared(ddata->dev, "rstctrl");
637 
638 	return PTR_ERR_OR_ZERO(ddata->rsts);
639 }
640 
641 /**
642  * sysc_parse_and_check_child_range - parses module IO region from ranges
643  * @ddata: device driver data
644  *
645  * In general we only need rev, syss, and sysc registers and not the whole
646  * module range. But we do want the offsets for these registers from the
647  * module base. This allows us to check them against the legacy hwmod
648  * platform data. Let's also check the ranges are configured properly.
649  */
sysc_parse_and_check_child_range(struct sysc * ddata)650 static int sysc_parse_and_check_child_range(struct sysc *ddata)
651 {
652 	struct device_node *np = ddata->dev->of_node;
653 	struct of_range_parser parser;
654 	struct of_range range;
655 	int error;
656 
657 	error = of_range_parser_init(&parser, np);
658 	if (error)
659 		return error;
660 
661 	for_each_of_range(&parser, &range) {
662 		ddata->module_pa = range.cpu_addr;
663 		ddata->module_size = range.size;
664 		break;
665 	}
666 
667 	return 0;
668 }
669 
670 /* Interconnect instances to probe before l4_per instances */
671 static struct resource early_bus_ranges[] = {
672 	/* am3/4 l4_wkup */
673 	{ .start = 0x44c00000, .end = 0x44c00000 + 0x300000, },
674 	/* omap4/5 and dra7 l4_cfg */
675 	{ .start = 0x4a000000, .end = 0x4a000000 + 0x300000, },
676 	/* omap4 l4_wkup */
677 	{ .start = 0x4a300000, .end = 0x4a300000 + 0x30000,  },
678 	/* omap5 and dra7 l4_wkup without dra7 dcan segment */
679 	{ .start = 0x4ae00000, .end = 0x4ae00000 + 0x30000,  },
680 };
681 
682 static atomic_t sysc_defer = ATOMIC_INIT(10);
683 
684 /**
685  * sysc_defer_non_critical - defer non_critical interconnect probing
686  * @ddata: device driver data
687  *
688  * We want to probe l4_cfg and l4_wkup interconnect instances before any
689  * l4_per instances as l4_per instances depend on resources on l4_cfg and
690  * l4_wkup interconnects.
691  */
sysc_defer_non_critical(struct sysc * ddata)692 static int sysc_defer_non_critical(struct sysc *ddata)
693 {
694 	struct resource *res;
695 	int i;
696 
697 	if (!atomic_read(&sysc_defer))
698 		return 0;
699 
700 	for (i = 0; i < ARRAY_SIZE(early_bus_ranges); i++) {
701 		res = &early_bus_ranges[i];
702 		if (ddata->module_pa >= res->start &&
703 		    ddata->module_pa <= res->end) {
704 			atomic_set(&sysc_defer, 0);
705 
706 			return 0;
707 		}
708 	}
709 
710 	atomic_dec_if_positive(&sysc_defer);
711 
712 	return -EPROBE_DEFER;
713 }
714 
715 static struct device_node *stdout_path;
716 
sysc_init_stdout_path(struct sysc * ddata)717 static void sysc_init_stdout_path(struct sysc *ddata)
718 {
719 	struct device_node *np = NULL;
720 	const char *uart;
721 
722 	if (IS_ERR(stdout_path))
723 		return;
724 
725 	if (stdout_path)
726 		return;
727 
728 	np = of_find_node_by_path("/chosen");
729 	if (!np)
730 		goto err;
731 
732 	uart = of_get_property(np, "stdout-path", NULL);
733 	if (!uart)
734 		goto err;
735 
736 	np = of_find_node_by_path(uart);
737 	if (!np)
738 		goto err;
739 
740 	stdout_path = np;
741 
742 	return;
743 
744 err:
745 	stdout_path = ERR_PTR(-ENODEV);
746 }
747 
sysc_check_quirk_stdout(struct sysc * ddata,struct device_node * np)748 static void sysc_check_quirk_stdout(struct sysc *ddata,
749 				    struct device_node *np)
750 {
751 	sysc_init_stdout_path(ddata);
752 	if (np != stdout_path)
753 		return;
754 
755 	ddata->cfg.quirks |= SYSC_QUIRK_NO_IDLE_ON_INIT |
756 				SYSC_QUIRK_NO_RESET_ON_INIT;
757 }
758 
759 /**
760  * sysc_check_one_child - check child configuration
761  * @ddata: device driver data
762  * @np: child device node
763  *
764  * Let's avoid messy situations where we have new interconnect target
765  * node but children have "ti,hwmods". These belong to the interconnect
766  * target node and are managed by this driver.
767  */
sysc_check_one_child(struct sysc * ddata,struct device_node * np)768 static void sysc_check_one_child(struct sysc *ddata,
769 				 struct device_node *np)
770 {
771 	const char *name;
772 
773 	name = of_get_property(np, "ti,hwmods", NULL);
774 	if (name && !of_device_is_compatible(np, "ti,sysc"))
775 		dev_warn(ddata->dev, "really a child ti,hwmods property?");
776 
777 	sysc_check_quirk_stdout(ddata, np);
778 	sysc_parse_dts_quirks(ddata, np, true);
779 }
780 
sysc_check_children(struct sysc * ddata)781 static void sysc_check_children(struct sysc *ddata)
782 {
783 	struct device_node *child;
784 
785 	for_each_child_of_node(ddata->dev->of_node, child)
786 		sysc_check_one_child(ddata, child);
787 }
788 
789 /*
790  * So far only I2C uses 16-bit read access with clockactivity with revision
791  * in two registers with stride of 4. We can detect this based on the rev
792  * register size to configure things far enough to be able to properly read
793  * the revision register.
794  */
sysc_check_quirk_16bit(struct sysc * ddata,struct resource * res)795 static void sysc_check_quirk_16bit(struct sysc *ddata, struct resource *res)
796 {
797 	if (resource_size(res) == 8)
798 		ddata->cfg.quirks |= SYSC_QUIRK_16BIT | SYSC_QUIRK_USE_CLOCKACT;
799 }
800 
801 /**
802  * sysc_parse_one - parses the interconnect target module registers
803  * @ddata: device driver data
804  * @reg: register to parse
805  */
sysc_parse_one(struct sysc * ddata,enum sysc_registers reg)806 static int sysc_parse_one(struct sysc *ddata, enum sysc_registers reg)
807 {
808 	struct resource *res;
809 	const char *name;
810 
811 	switch (reg) {
812 	case SYSC_REVISION:
813 	case SYSC_SYSCONFIG:
814 	case SYSC_SYSSTATUS:
815 		name = reg_names[reg];
816 		break;
817 	default:
818 		return -EINVAL;
819 	}
820 
821 	res = platform_get_resource_byname(to_platform_device(ddata->dev),
822 					   IORESOURCE_MEM, name);
823 	if (!res) {
824 		ddata->offsets[reg] = -ENODEV;
825 
826 		return 0;
827 	}
828 
829 	ddata->offsets[reg] = res->start - ddata->module_pa;
830 	if (reg == SYSC_REVISION)
831 		sysc_check_quirk_16bit(ddata, res);
832 
833 	return 0;
834 }
835 
sysc_parse_registers(struct sysc * ddata)836 static int sysc_parse_registers(struct sysc *ddata)
837 {
838 	int i, error;
839 
840 	for (i = 0; i < SYSC_MAX_REGS; i++) {
841 		error = sysc_parse_one(ddata, i);
842 		if (error)
843 			return error;
844 	}
845 
846 	return 0;
847 }
848 
849 /**
850  * sysc_check_registers - check for misconfigured register overlaps
851  * @ddata: device driver data
852  */
sysc_check_registers(struct sysc * ddata)853 static int sysc_check_registers(struct sysc *ddata)
854 {
855 	int i, j, nr_regs = 0, nr_matches = 0;
856 
857 	for (i = 0; i < SYSC_MAX_REGS; i++) {
858 		if (ddata->offsets[i] < 0)
859 			continue;
860 
861 		if (ddata->offsets[i] > (ddata->module_size - 4)) {
862 			dev_err(ddata->dev, "register outside module range");
863 
864 				return -EINVAL;
865 		}
866 
867 		for (j = 0; j < SYSC_MAX_REGS; j++) {
868 			if (ddata->offsets[j] < 0)
869 				continue;
870 
871 			if (ddata->offsets[i] == ddata->offsets[j])
872 				nr_matches++;
873 		}
874 		nr_regs++;
875 	}
876 
877 	if (nr_matches > nr_regs) {
878 		dev_err(ddata->dev, "overlapping registers: (%i/%i)",
879 			nr_regs, nr_matches);
880 
881 		return -EINVAL;
882 	}
883 
884 	return 0;
885 }
886 
887 /**
888  * sysc_ioremap - ioremap register space for the interconnect target module
889  * @ddata: device driver data
890  *
891  * Note that the interconnect target module registers can be anywhere
892  * within the interconnect target module range. For example, SGX has
893  * them at offset 0x1fc00 in the 32MB module address space. And cpsw
894  * has them at offset 0x1200 in the CPSW_WR child. Usually the
895  * interconnect target module registers are at the beginning of
896  * the module range though.
897  */
sysc_ioremap(struct sysc * ddata)898 static int sysc_ioremap(struct sysc *ddata)
899 {
900 	int size;
901 
902 	if (ddata->offsets[SYSC_REVISION] < 0 &&
903 	    ddata->offsets[SYSC_SYSCONFIG] < 0 &&
904 	    ddata->offsets[SYSC_SYSSTATUS] < 0) {
905 		size = ddata->module_size;
906 	} else {
907 		size = max3(ddata->offsets[SYSC_REVISION],
908 			    ddata->offsets[SYSC_SYSCONFIG],
909 			    ddata->offsets[SYSC_SYSSTATUS]);
910 
911 		if (size < SZ_1K)
912 			size = SZ_1K;
913 
914 		if ((size + sizeof(u32)) > ddata->module_size)
915 			size = ddata->module_size;
916 	}
917 
918 	ddata->module_va = devm_ioremap(ddata->dev,
919 					ddata->module_pa,
920 					size + sizeof(u32));
921 	if (!ddata->module_va)
922 		return -EIO;
923 
924 	return 0;
925 }
926 
927 /**
928  * sysc_map_and_check_registers - ioremap and check device registers
929  * @ddata: device driver data
930  */
sysc_map_and_check_registers(struct sysc * ddata)931 static int sysc_map_and_check_registers(struct sysc *ddata)
932 {
933 	struct device_node *np = ddata->dev->of_node;
934 	int error;
935 
936 	error = sysc_parse_and_check_child_range(ddata);
937 	if (error)
938 		return error;
939 
940 	error = sysc_defer_non_critical(ddata);
941 	if (error)
942 		return error;
943 
944 	sysc_check_children(ddata);
945 
946 	if (!of_property_present(np, "reg"))
947 		return 0;
948 
949 	error = sysc_parse_registers(ddata);
950 	if (error)
951 		return error;
952 
953 	error = sysc_ioremap(ddata);
954 	if (error)
955 		return error;
956 
957 	error = sysc_check_registers(ddata);
958 	if (error)
959 		return error;
960 
961 	return 0;
962 }
963 
964 /**
965  * sysc_show_rev - read and show interconnect target module revision
966  * @bufp: buffer to print the information to
967  * @ddata: device driver data
968  */
sysc_show_rev(char * bufp,struct sysc * ddata)969 static int sysc_show_rev(char *bufp, struct sysc *ddata)
970 {
971 	int len;
972 
973 	if (ddata->offsets[SYSC_REVISION] < 0)
974 		return sprintf(bufp, ":NA");
975 
976 	len = sprintf(bufp, ":%08x", ddata->revision);
977 
978 	return len;
979 }
980 
sysc_show_reg(struct sysc * ddata,char * bufp,enum sysc_registers reg)981 static int sysc_show_reg(struct sysc *ddata,
982 			 char *bufp, enum sysc_registers reg)
983 {
984 	if (ddata->offsets[reg] < 0)
985 		return sprintf(bufp, ":NA");
986 
987 	return sprintf(bufp, ":%x", ddata->offsets[reg]);
988 }
989 
sysc_show_name(char * bufp,struct sysc * ddata)990 static int sysc_show_name(char *bufp, struct sysc *ddata)
991 {
992 	if (!ddata->name)
993 		return 0;
994 
995 	return sprintf(bufp, ":%s", ddata->name);
996 }
997 
998 /**
999  * sysc_show_registers - show information about interconnect target module
1000  * @ddata: device driver data
1001  */
sysc_show_registers(struct sysc * ddata)1002 static void sysc_show_registers(struct sysc *ddata)
1003 {
1004 	char buf[128];
1005 	char *bufp = buf;
1006 	int i;
1007 
1008 	for (i = 0; i < SYSC_MAX_REGS; i++)
1009 		bufp += sysc_show_reg(ddata, bufp, i);
1010 
1011 	bufp += sysc_show_rev(bufp, ddata);
1012 	bufp += sysc_show_name(bufp, ddata);
1013 
1014 	dev_dbg(ddata->dev, "%llx:%x%s\n",
1015 		ddata->module_pa, ddata->module_size,
1016 		buf);
1017 }
1018 
1019 /**
1020  * sysc_write_sysconfig - handle sysconfig quirks for register write
1021  * @ddata: device driver data
1022  * @value: register value
1023  */
sysc_write_sysconfig(struct sysc * ddata,u32 value)1024 static void sysc_write_sysconfig(struct sysc *ddata, u32 value)
1025 {
1026 	if (ddata->module_unlock_quirk)
1027 		ddata->module_unlock_quirk(ddata);
1028 
1029 	sysc_write(ddata, ddata->offsets[SYSC_SYSCONFIG], value);
1030 
1031 	if (ddata->module_lock_quirk)
1032 		ddata->module_lock_quirk(ddata);
1033 }
1034 
1035 #define SYSC_IDLE_MASK	(SYSC_NR_IDLEMODES - 1)
1036 #define SYSC_CLOCACT_ICK	2
1037 
1038 /* Caller needs to manage sysc_clkdm_deny_idle() and sysc_clkdm_allow_idle() */
sysc_enable_module(struct device * dev)1039 static int sysc_enable_module(struct device *dev)
1040 {
1041 	struct sysc *ddata;
1042 	const struct sysc_regbits *regbits;
1043 	u32 reg, idlemodes, best_mode;
1044 	int error;
1045 
1046 	ddata = dev_get_drvdata(dev);
1047 
1048 	/*
1049 	 * Some modules like DSS reset automatically on idle. Enable optional
1050 	 * reset clocks and wait for OCP softreset to complete.
1051 	 */
1052 	if (ddata->cfg.quirks & SYSC_QUIRK_OPT_CLKS_IN_RESET) {
1053 		error = sysc_enable_opt_clocks(ddata);
1054 		if (error) {
1055 			dev_err(ddata->dev,
1056 				"Optional clocks failed for enable: %i\n",
1057 				error);
1058 			return error;
1059 		}
1060 	}
1061 	/*
1062 	 * Some modules like i2c and hdq1w have unusable reset status unless
1063 	 * the module reset quirk is enabled. Skip status check on enable.
1064 	 */
1065 	if (!(ddata->cfg.quirks & SYSC_MODULE_QUIRK_ENA_RESETDONE)) {
1066 		error = sysc_wait_softreset(ddata);
1067 		if (error)
1068 			dev_warn(ddata->dev, "OCP softreset timed out\n");
1069 	}
1070 	if (ddata->cfg.quirks & SYSC_QUIRK_OPT_CLKS_IN_RESET)
1071 		sysc_disable_opt_clocks(ddata);
1072 
1073 	/*
1074 	 * Some subsystem private interconnects, like DSS top level module,
1075 	 * need only the automatic OCP softreset handling with no sysconfig
1076 	 * register bits to configure.
1077 	 */
1078 	if (ddata->offsets[SYSC_SYSCONFIG] == -ENODEV)
1079 		return 0;
1080 
1081 	regbits = ddata->cap->regbits;
1082 	reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1083 
1084 	/*
1085 	 * Set CLOCKACTIVITY, we only use it for ick. And we only configure it
1086 	 * based on the SYSC_QUIRK_USE_CLOCKACT flag, not based on the hardware
1087 	 * capabilities. See the old HWMOD_SET_DEFAULT_CLOCKACT flag.
1088 	 */
1089 	if (regbits->clkact_shift >= 0 &&
1090 	    (ddata->cfg.quirks & SYSC_QUIRK_USE_CLOCKACT))
1091 		reg |= SYSC_CLOCACT_ICK << regbits->clkact_shift;
1092 
1093 	/* Set SIDLE mode */
1094 	idlemodes = ddata->cfg.sidlemodes;
1095 	if (!idlemodes || regbits->sidle_shift < 0)
1096 		goto set_midle;
1097 
1098 	if (ddata->cfg.quirks & (SYSC_QUIRK_SWSUP_SIDLE |
1099 				 SYSC_QUIRK_SWSUP_SIDLE_ACT)) {
1100 		best_mode = SYSC_IDLE_NO;
1101 
1102 		/* Clear WAKEUP */
1103 		if (regbits->enwkup_shift >= 0 &&
1104 		    ddata->cfg.sysc_val & BIT(regbits->enwkup_shift))
1105 			reg &= ~BIT(regbits->enwkup_shift);
1106 	} else {
1107 		best_mode = fls(ddata->cfg.sidlemodes) - 1;
1108 		if (best_mode > SYSC_IDLE_MASK) {
1109 			dev_err(dev, "%s: invalid sidlemode\n", __func__);
1110 			return -EINVAL;
1111 		}
1112 
1113 		/* Set WAKEUP */
1114 		if (regbits->enwkup_shift >= 0 &&
1115 		    ddata->cfg.sysc_val & BIT(regbits->enwkup_shift))
1116 			reg |= BIT(regbits->enwkup_shift);
1117 	}
1118 
1119 	reg &= ~(SYSC_IDLE_MASK << regbits->sidle_shift);
1120 	reg |= best_mode << regbits->sidle_shift;
1121 	sysc_write_sysconfig(ddata, reg);
1122 
1123 set_midle:
1124 	/* Set MIDLE mode */
1125 	idlemodes = ddata->cfg.midlemodes;
1126 	if (!idlemodes || regbits->midle_shift < 0)
1127 		goto set_autoidle;
1128 
1129 	best_mode = fls(ddata->cfg.midlemodes) - 1;
1130 	if (best_mode > SYSC_IDLE_MASK) {
1131 		dev_err(dev, "%s: invalid midlemode\n", __func__);
1132 		error = -EINVAL;
1133 		goto save_context;
1134 	}
1135 
1136 	if (ddata->cfg.quirks & SYSC_QUIRK_SWSUP_MSTANDBY)
1137 		best_mode = SYSC_IDLE_NO;
1138 
1139 	reg &= ~(SYSC_IDLE_MASK << regbits->midle_shift);
1140 	reg |= best_mode << regbits->midle_shift;
1141 	sysc_write_sysconfig(ddata, reg);
1142 
1143 set_autoidle:
1144 	/* Autoidle bit must enabled separately if available */
1145 	if (regbits->autoidle_shift >= 0 &&
1146 	    ddata->cfg.sysc_val & BIT(regbits->autoidle_shift)) {
1147 		reg |= 1 << regbits->autoidle_shift;
1148 		sysc_write_sysconfig(ddata, reg);
1149 	}
1150 
1151 	error = 0;
1152 
1153 save_context:
1154 	/* Save context and flush posted write */
1155 	ddata->sysconfig = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1156 
1157 	if (ddata->module_enable_quirk)
1158 		ddata->module_enable_quirk(ddata);
1159 
1160 	return error;
1161 }
1162 
sysc_best_idle_mode(u32 idlemodes,u32 * best_mode)1163 static int sysc_best_idle_mode(u32 idlemodes, u32 *best_mode)
1164 {
1165 	if (idlemodes & BIT(SYSC_IDLE_SMART_WKUP))
1166 		*best_mode = SYSC_IDLE_SMART_WKUP;
1167 	else if (idlemodes & BIT(SYSC_IDLE_SMART))
1168 		*best_mode = SYSC_IDLE_SMART;
1169 	else if (idlemodes & BIT(SYSC_IDLE_FORCE))
1170 		*best_mode = SYSC_IDLE_FORCE;
1171 	else
1172 		return -EINVAL;
1173 
1174 	return 0;
1175 }
1176 
1177 /* Caller needs to manage sysc_clkdm_deny_idle() and sysc_clkdm_allow_idle() */
sysc_disable_module(struct device * dev)1178 static int sysc_disable_module(struct device *dev)
1179 {
1180 	struct sysc *ddata;
1181 	const struct sysc_regbits *regbits;
1182 	u32 reg, idlemodes, best_mode;
1183 	int ret;
1184 
1185 	ddata = dev_get_drvdata(dev);
1186 	if (ddata->offsets[SYSC_SYSCONFIG] == -ENODEV)
1187 		return 0;
1188 
1189 	if (ddata->module_disable_quirk)
1190 		ddata->module_disable_quirk(ddata);
1191 
1192 	regbits = ddata->cap->regbits;
1193 	reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1194 
1195 	/* Set MIDLE mode */
1196 	idlemodes = ddata->cfg.midlemodes;
1197 	if (!idlemodes || regbits->midle_shift < 0)
1198 		goto set_sidle;
1199 
1200 	ret = sysc_best_idle_mode(idlemodes, &best_mode);
1201 	if (ret) {
1202 		dev_err(dev, "%s: invalid midlemode\n", __func__);
1203 		return ret;
1204 	}
1205 
1206 	if (ddata->cfg.quirks & (SYSC_QUIRK_SWSUP_MSTANDBY) ||
1207 	    ddata->cfg.quirks & (SYSC_QUIRK_FORCE_MSTANDBY))
1208 		best_mode = SYSC_IDLE_FORCE;
1209 
1210 	reg &= ~(SYSC_IDLE_MASK << regbits->midle_shift);
1211 	reg |= best_mode << regbits->midle_shift;
1212 	sysc_write_sysconfig(ddata, reg);
1213 
1214 set_sidle:
1215 	/* Set SIDLE mode */
1216 	idlemodes = ddata->cfg.sidlemodes;
1217 	if (!idlemodes || regbits->sidle_shift < 0) {
1218 		ret = 0;
1219 		goto save_context;
1220 	}
1221 
1222 	if (ddata->cfg.quirks & SYSC_QUIRK_SWSUP_SIDLE) {
1223 		best_mode = SYSC_IDLE_FORCE;
1224 	} else {
1225 		ret = sysc_best_idle_mode(idlemodes, &best_mode);
1226 		if (ret) {
1227 			dev_err(dev, "%s: invalid sidlemode\n", __func__);
1228 			ret = -EINVAL;
1229 			goto save_context;
1230 		}
1231 	}
1232 
1233 	if (ddata->cfg.quirks & SYSC_QUIRK_SWSUP_SIDLE_ACT) {
1234 		/* Set WAKEUP */
1235 		if (regbits->enwkup_shift >= 0 &&
1236 		    ddata->cfg.sysc_val & BIT(regbits->enwkup_shift))
1237 			reg |= BIT(regbits->enwkup_shift);
1238 	}
1239 
1240 	reg &= ~(SYSC_IDLE_MASK << regbits->sidle_shift);
1241 	reg |= best_mode << regbits->sidle_shift;
1242 	if (regbits->autoidle_shift >= 0 &&
1243 	    ddata->cfg.sysc_val & BIT(regbits->autoidle_shift))
1244 		reg |= 1 << regbits->autoidle_shift;
1245 	sysc_write_sysconfig(ddata, reg);
1246 
1247 	ret = 0;
1248 
1249 save_context:
1250 	/* Save context and flush posted write */
1251 	ddata->sysconfig = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1252 
1253 	return ret;
1254 }
1255 
sysc_runtime_suspend_legacy(struct device * dev,struct sysc * ddata)1256 static int __maybe_unused sysc_runtime_suspend_legacy(struct device *dev,
1257 						      struct sysc *ddata)
1258 {
1259 	struct ti_sysc_platform_data *pdata;
1260 	int error;
1261 
1262 	pdata = dev_get_platdata(ddata->dev);
1263 	if (!pdata)
1264 		return 0;
1265 
1266 	if (!pdata->idle_module)
1267 		return -ENODEV;
1268 
1269 	error = pdata->idle_module(dev, &ddata->cookie);
1270 	if (error)
1271 		dev_err(dev, "%s: could not idle: %i\n",
1272 			__func__, error);
1273 
1274 	reset_control_assert(ddata->rsts);
1275 
1276 	return 0;
1277 }
1278 
sysc_runtime_resume_legacy(struct device * dev,struct sysc * ddata)1279 static int __maybe_unused sysc_runtime_resume_legacy(struct device *dev,
1280 						     struct sysc *ddata)
1281 {
1282 	struct ti_sysc_platform_data *pdata;
1283 	int error;
1284 
1285 	pdata = dev_get_platdata(ddata->dev);
1286 	if (!pdata)
1287 		return 0;
1288 
1289 	if (!pdata->enable_module)
1290 		return -ENODEV;
1291 
1292 	error = pdata->enable_module(dev, &ddata->cookie);
1293 	if (error)
1294 		dev_err(dev, "%s: could not enable: %i\n",
1295 			__func__, error);
1296 
1297 	reset_control_deassert(ddata->rsts);
1298 
1299 	return 0;
1300 }
1301 
sysc_runtime_suspend(struct device * dev)1302 static int __maybe_unused sysc_runtime_suspend(struct device *dev)
1303 {
1304 	struct sysc *ddata;
1305 	int error = 0;
1306 
1307 	ddata = dev_get_drvdata(dev);
1308 
1309 	if (!ddata->enabled)
1310 		return 0;
1311 
1312 	sysc_clkdm_deny_idle(ddata);
1313 
1314 	if (ddata->legacy_mode) {
1315 		error = sysc_runtime_suspend_legacy(dev, ddata);
1316 		if (error)
1317 			goto err_allow_idle;
1318 	} else {
1319 		error = sysc_disable_module(dev);
1320 		if (error)
1321 			goto err_allow_idle;
1322 	}
1323 
1324 	sysc_disable_main_clocks(ddata);
1325 
1326 	if (sysc_opt_clks_needed(ddata))
1327 		sysc_disable_opt_clocks(ddata);
1328 
1329 	ddata->enabled = false;
1330 
1331 err_allow_idle:
1332 	sysc_clkdm_allow_idle(ddata);
1333 
1334 	reset_control_assert(ddata->rsts);
1335 
1336 	return error;
1337 }
1338 
sysc_runtime_resume(struct device * dev)1339 static int __maybe_unused sysc_runtime_resume(struct device *dev)
1340 {
1341 	struct sysc *ddata;
1342 	int error = 0;
1343 
1344 	ddata = dev_get_drvdata(dev);
1345 
1346 	if (ddata->enabled)
1347 		return 0;
1348 
1349 
1350 	sysc_clkdm_deny_idle(ddata);
1351 
1352 	if (sysc_opt_clks_needed(ddata)) {
1353 		error = sysc_enable_opt_clocks(ddata);
1354 		if (error)
1355 			goto err_allow_idle;
1356 	}
1357 
1358 	error = sysc_enable_main_clocks(ddata);
1359 	if (error)
1360 		goto err_opt_clocks;
1361 
1362 	reset_control_deassert(ddata->rsts);
1363 
1364 	if (ddata->legacy_mode) {
1365 		error = sysc_runtime_resume_legacy(dev, ddata);
1366 		if (error)
1367 			goto err_main_clocks;
1368 	} else {
1369 		error = sysc_enable_module(dev);
1370 		if (error)
1371 			goto err_main_clocks;
1372 	}
1373 
1374 	ddata->enabled = true;
1375 
1376 	sysc_clkdm_allow_idle(ddata);
1377 
1378 	return 0;
1379 
1380 err_main_clocks:
1381 	sysc_disable_main_clocks(ddata);
1382 err_opt_clocks:
1383 	if (sysc_opt_clks_needed(ddata))
1384 		sysc_disable_opt_clocks(ddata);
1385 err_allow_idle:
1386 	sysc_clkdm_allow_idle(ddata);
1387 
1388 	return error;
1389 }
1390 
1391 /*
1392  * Checks if device context was lost. Assumes the sysconfig register value
1393  * after lost context is different from the configured value. Only works for
1394  * enabled devices.
1395  *
1396  * Eventually we may want to also add support to using the context lost
1397  * registers that some SoCs have.
1398  */
sysc_check_context(struct sysc * ddata)1399 static int sysc_check_context(struct sysc *ddata)
1400 {
1401 	u32 reg;
1402 
1403 	if (!ddata->enabled)
1404 		return -ENODATA;
1405 
1406 	reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1407 	if (reg == ddata->sysconfig)
1408 		return 0;
1409 
1410 	return -EACCES;
1411 }
1412 
sysc_reinit_module(struct sysc * ddata,bool leave_enabled)1413 static int sysc_reinit_module(struct sysc *ddata, bool leave_enabled)
1414 {
1415 	struct device *dev = ddata->dev;
1416 	int error;
1417 
1418 	if (ddata->enabled) {
1419 		/* Nothing to do if enabled and context not lost */
1420 		error = sysc_check_context(ddata);
1421 		if (!error)
1422 			return 0;
1423 
1424 		/* Disable target module if it is enabled */
1425 		error = sysc_runtime_suspend(dev);
1426 		if (error)
1427 			dev_warn(dev, "reinit suspend failed: %i\n", error);
1428 	}
1429 
1430 	/* Enable target module */
1431 	error = sysc_runtime_resume(dev);
1432 	if (error)
1433 		dev_warn(dev, "reinit resume failed: %i\n", error);
1434 
1435 	/* Some modules like am335x gpmc need reset and restore of sysconfig */
1436 	if (ddata->cfg.quirks & SYSC_QUIRK_RESET_ON_CTX_LOST) {
1437 		error = sysc_reset(ddata);
1438 		if (error)
1439 			dev_warn(dev, "reinit reset failed: %i\n", error);
1440 
1441 		sysc_write_sysconfig(ddata, ddata->sysconfig);
1442 	}
1443 
1444 	if (leave_enabled)
1445 		return error;
1446 
1447 	/* Disable target module if no leave_enabled was set */
1448 	error = sysc_runtime_suspend(dev);
1449 	if (error)
1450 		dev_warn(dev, "reinit suspend failed: %i\n", error);
1451 
1452 	return error;
1453 }
1454 
sysc_noirq_suspend(struct device * dev)1455 static int __maybe_unused sysc_noirq_suspend(struct device *dev)
1456 {
1457 	struct sysc *ddata;
1458 
1459 	ddata = dev_get_drvdata(dev);
1460 
1461 	if (ddata->cfg.quirks &
1462 	    (SYSC_QUIRK_LEGACY_IDLE | SYSC_QUIRK_NO_IDLE))
1463 		return 0;
1464 
1465 	if (!ddata->enabled)
1466 		return 0;
1467 
1468 	ddata->needs_resume = 1;
1469 
1470 	return sysc_runtime_suspend(dev);
1471 }
1472 
sysc_noirq_resume(struct device * dev)1473 static int __maybe_unused sysc_noirq_resume(struct device *dev)
1474 {
1475 	struct sysc *ddata;
1476 	int error = 0;
1477 
1478 	ddata = dev_get_drvdata(dev);
1479 
1480 	if (ddata->cfg.quirks &
1481 	    (SYSC_QUIRK_LEGACY_IDLE | SYSC_QUIRK_NO_IDLE))
1482 		return 0;
1483 
1484 	if (ddata->cfg.quirks & SYSC_QUIRK_REINIT_ON_RESUME) {
1485 		error = sysc_reinit_module(ddata, ddata->needs_resume);
1486 		if (error)
1487 			dev_warn(dev, "noirq_resume failed: %i\n", error);
1488 	} else if (ddata->needs_resume) {
1489 		error = sysc_runtime_resume(dev);
1490 		if (error)
1491 			dev_warn(dev, "noirq_resume failed: %i\n", error);
1492 	}
1493 
1494 	ddata->needs_resume = 0;
1495 
1496 	return error;
1497 }
1498 
1499 static const struct dev_pm_ops sysc_pm_ops = {
1500 	SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(sysc_noirq_suspend, sysc_noirq_resume)
1501 	SET_RUNTIME_PM_OPS(sysc_runtime_suspend,
1502 			   sysc_runtime_resume,
1503 			   NULL)
1504 };
1505 
1506 /* Module revision register based quirks */
1507 struct sysc_revision_quirk {
1508 	const char *name;
1509 	u32 base;
1510 	int rev_offset;
1511 	int sysc_offset;
1512 	int syss_offset;
1513 	u32 revision;
1514 	u32 revision_mask;
1515 	u32 quirks;
1516 };
1517 
1518 #define SYSC_QUIRK(optname, optbase, optrev, optsysc, optsyss,		\
1519 		   optrev_val, optrevmask, optquirkmask)		\
1520 	{								\
1521 		.name = (optname),					\
1522 		.base = (optbase),					\
1523 		.rev_offset = (optrev),					\
1524 		.sysc_offset = (optsysc),				\
1525 		.syss_offset = (optsyss),				\
1526 		.revision = (optrev_val),				\
1527 		.revision_mask = (optrevmask),				\
1528 		.quirks = (optquirkmask),				\
1529 	}
1530 
1531 static const struct sysc_revision_quirk sysc_revision_quirks[] = {
1532 	/* These drivers need to be fixed to not use pm_runtime_irq_safe() */
1533 	SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x00000046, 0xffffffff,
1534 		   SYSC_QUIRK_SWSUP_SIDLE_ACT | SYSC_QUIRK_LEGACY_IDLE),
1535 	SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x00000052, 0xffffffff,
1536 		   SYSC_QUIRK_SWSUP_SIDLE_ACT | SYSC_QUIRK_LEGACY_IDLE),
1537 	/* Uarts on omap4 and later */
1538 	SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x50411e03, 0xffff00ff,
1539 		   SYSC_QUIRK_SWSUP_SIDLE_ACT | SYSC_QUIRK_LEGACY_IDLE),
1540 	SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x47422e03, 0xffffffff,
1541 		   SYSC_QUIRK_SWSUP_SIDLE_ACT | SYSC_QUIRK_LEGACY_IDLE),
1542 	SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x47424e03, 0xffffffff,
1543 		   SYSC_QUIRK_SWSUP_SIDLE_ACT | SYSC_QUIRK_LEGACY_IDLE),
1544 
1545 	/* Quirks that need to be set based on the module address */
1546 	SYSC_QUIRK("mcpdm", 0x40132000, 0, 0x10, -ENODEV, 0x50000800, 0xffffffff,
1547 		   SYSC_QUIRK_EXT_OPT_CLOCK | SYSC_QUIRK_NO_RESET_ON_INIT |
1548 		   SYSC_QUIRK_SWSUP_SIDLE),
1549 
1550 	/* Quirks that need to be set based on detected module */
1551 	SYSC_QUIRK("aess", 0, 0, 0x10, -ENODEV, 0x40000000, 0xffffffff,
1552 		   SYSC_MODULE_QUIRK_AESS),
1553 	/* Errata i893 handling for dra7 dcan1 and 2 */
1554 	SYSC_QUIRK("dcan", 0x4ae3c000, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff,
1555 		   SYSC_QUIRK_CLKDM_NOAUTO),
1556 	SYSC_QUIRK("dcan", 0x48480000, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff,
1557 		   SYSC_QUIRK_CLKDM_NOAUTO),
1558 	SYSC_QUIRK("dss", 0x4832a000, 0, 0x10, 0x14, 0x00000020, 0xffffffff,
1559 		   SYSC_QUIRK_OPT_CLKS_IN_RESET | SYSC_MODULE_QUIRK_DSS_RESET),
1560 	SYSC_QUIRK("dss", 0x58000000, 0, -ENODEV, 0x14, 0x00000040, 0xffffffff,
1561 		   SYSC_QUIRK_OPT_CLKS_IN_RESET | SYSC_MODULE_QUIRK_DSS_RESET),
1562 	SYSC_QUIRK("dss", 0x58000000, 0, -ENODEV, 0x14, 0x00000061, 0xffffffff,
1563 		   SYSC_QUIRK_OPT_CLKS_IN_RESET | SYSC_MODULE_QUIRK_DSS_RESET),
1564 	SYSC_QUIRK("dwc3", 0x48880000, 0, 0x10, -ENODEV, 0x500a0200, 0xffffffff,
1565 		   SYSC_QUIRK_CLKDM_NOAUTO),
1566 	SYSC_QUIRK("dwc3", 0x488c0000, 0, 0x10, -ENODEV, 0x500a0200, 0xffffffff,
1567 		   SYSC_QUIRK_CLKDM_NOAUTO),
1568 	SYSC_QUIRK("gpio", 0, 0, 0x10, 0x114, 0x50600801, 0xffff00ff,
1569 		   SYSC_QUIRK_OPT_CLKS_IN_RESET),
1570 	SYSC_QUIRK("gpmc", 0, 0, 0x10, 0x14, 0x00000060, 0xffffffff,
1571 		   SYSC_QUIRK_REINIT_ON_CTX_LOST | SYSC_QUIRK_RESET_ON_CTX_LOST |
1572 		   SYSC_QUIRK_GPMC_DEBUG),
1573 	SYSC_QUIRK("hdmi", 0, 0, 0x10, -ENODEV, 0x50030200, 0xffffffff,
1574 		   SYSC_QUIRK_OPT_CLKS_NEEDED),
1575 	SYSC_QUIRK("hdq1w", 0, 0, 0x14, 0x18, 0x00000006, 0xffffffff,
1576 		   SYSC_MODULE_QUIRK_HDQ1W | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1577 	SYSC_QUIRK("hdq1w", 0, 0, 0x14, 0x18, 0x0000000a, 0xffffffff,
1578 		   SYSC_MODULE_QUIRK_HDQ1W | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1579 	SYSC_QUIRK("i2c", 0, 0, 0x20, 0x10, 0x00000036, 0x000000ff,
1580 		   SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1581 	SYSC_QUIRK("i2c", 0, 0, 0x20, 0x10, 0x0000003c, 0x000000ff,
1582 		   SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1583 	SYSC_QUIRK("i2c", 0, 0, 0x20, 0x10, 0x00000040, 0x000000ff,
1584 		   SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1585 	SYSC_QUIRK("i2c", 0, 0, 0x10, 0x90, 0x5040000a, 0xfffff0f0,
1586 		   SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1587 	SYSC_QUIRK("gpu", 0x50000000, 0x14, -ENODEV, -ENODEV, 0x00010201, 0xffffffff, 0),
1588 	SYSC_QUIRK("gpu", 0x50000000, 0xfe00, 0xfe10, -ENODEV, 0x40000000 , 0xffffffff,
1589 		   SYSC_MODULE_QUIRK_SGX),
1590 	SYSC_QUIRK("lcdc", 0, 0, 0x54, -ENODEV, 0x4f201000, 0xffffffff,
1591 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1592 	SYSC_QUIRK("mcasp", 0, 0, 0x4, -ENODEV, 0x44306302, 0xffffffff,
1593 		   SYSC_QUIRK_SWSUP_SIDLE),
1594 	SYSC_QUIRK("rtc", 0, 0x74, 0x78, -ENODEV, 0x4eb01908, 0xffff00f0,
1595 		   SYSC_MODULE_QUIRK_RTC_UNLOCK),
1596 	SYSC_QUIRK("tptc", 0, 0, 0x10, -ENODEV, 0x40006c00, 0xffffefff,
1597 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1598 	SYSC_QUIRK("tptc", 0, 0, -ENODEV, -ENODEV, 0x40007c00, 0xffffffff,
1599 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1600 	SYSC_QUIRK("sata", 0, 0xfc, 0x1100, -ENODEV, 0x5e412000, 0xffffffff,
1601 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1602 	SYSC_QUIRK("usb_host_hs", 0, 0, 0x10, 0x14, 0x50700100, 0xffffffff,
1603 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1604 	SYSC_QUIRK("usb_host_hs", 0, 0, 0x10, -ENODEV, 0x50700101, 0xffffffff,
1605 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1606 	SYSC_QUIRK("usb_otg_hs", 0, 0x400, 0x404, 0x408, 0x00000033,
1607 		   0xffffffff, SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1608 		   SYSC_MODULE_QUIRK_OTG),
1609 	SYSC_QUIRK("usb_otg_hs", 0, 0x400, 0x404, 0x408, 0x00000040,
1610 		   0xffffffff, SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1611 		   SYSC_MODULE_QUIRK_OTG),
1612 	SYSC_QUIRK("usb_otg_hs", 0, 0x400, 0x404, 0x408, 0x00000050,
1613 		   0xffffffff, SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1614 		   SYSC_MODULE_QUIRK_OTG),
1615 	SYSC_QUIRK("usb_otg_hs", 0, 0, 0x10, -ENODEV, 0x4ea2080d, 0xffffffff,
1616 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1617 		   SYSC_QUIRK_REINIT_ON_CTX_LOST),
1618 	SYSC_QUIRK("wdt", 0, 0, 0x10, 0x14, 0x502a0500, 0xfffff0f0,
1619 		   SYSC_MODULE_QUIRK_WDT),
1620 	/* PRUSS on am3, am4 and am5 */
1621 	SYSC_QUIRK("pruss", 0, 0x26000, 0x26004, -ENODEV, 0x47000000, 0xff000000,
1622 		   SYSC_MODULE_QUIRK_PRUSS),
1623 	/* Watchdog on am3 and am4 */
1624 	SYSC_QUIRK("wdt", 0x44e35000, 0, 0x10, 0x14, 0x502a0500, 0xfffff0f0,
1625 		   SYSC_MODULE_QUIRK_WDT | SYSC_QUIRK_SWSUP_SIDLE),
1626 
1627 #ifdef DEBUG
1628 	SYSC_QUIRK("adc", 0, 0, 0x10, -ENODEV, 0x47300001, 0xffffffff, 0),
1629 	SYSC_QUIRK("atl", 0, 0, -ENODEV, -ENODEV, 0x0a070100, 0xffffffff, 0),
1630 	SYSC_QUIRK("cm", 0, 0, -ENODEV, -ENODEV, 0x40000301, 0xffffffff, 0),
1631 	SYSC_QUIRK("control", 0, 0, 0x10, -ENODEV, 0x40000900, 0xffffffff, 0),
1632 	SYSC_QUIRK("cpgmac", 0, 0x1200, 0x1208, 0x1204, 0x4edb1902,
1633 		   0xffff00f0, 0),
1634 	SYSC_QUIRK("dcan", 0, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff, 0),
1635 	SYSC_QUIRK("dcan", 0, 0x20, -ENODEV, -ENODEV, 0x4edb1902, 0xffffffff, 0),
1636 	SYSC_QUIRK("dispc", 0x4832a400, 0, 0x10, 0x14, 0x00000030, 0xffffffff, 0),
1637 	SYSC_QUIRK("dispc", 0x58001000, 0, 0x10, 0x14, 0x00000040, 0xffffffff, 0),
1638 	SYSC_QUIRK("dispc", 0x58001000, 0, 0x10, 0x14, 0x00000051, 0xffffffff, 0),
1639 	SYSC_QUIRK("dmic", 0, 0, 0x10, -ENODEV, 0x50010000, 0xffffffff, 0),
1640 	SYSC_QUIRK("dsi", 0x58004000, 0, 0x10, 0x14, 0x00000030, 0xffffffff, 0),
1641 	SYSC_QUIRK("dsi", 0x58005000, 0, 0x10, 0x14, 0x00000030, 0xffffffff, 0),
1642 	SYSC_QUIRK("dsi", 0x58005000, 0, 0x10, 0x14, 0x00000040, 0xffffffff, 0),
1643 	SYSC_QUIRK("dsi", 0x58009000, 0, 0x10, 0x14, 0x00000040, 0xffffffff, 0),
1644 	SYSC_QUIRK("dwc3", 0, 0, 0x10, -ENODEV, 0x500a0200, 0xffffffff, 0),
1645 	SYSC_QUIRK("d2d", 0x4a0b6000, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1646 	SYSC_QUIRK("d2d", 0x4a0cd000, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1647 	SYSC_QUIRK("elm", 0x48080000, 0, 0x10, 0x14, 0x00000020, 0xffffffff, 0),
1648 	SYSC_QUIRK("emif", 0, 0, -ENODEV, -ENODEV, 0x40441403, 0xffff0fff, 0),
1649 	SYSC_QUIRK("emif", 0, 0, -ENODEV, -ENODEV, 0x50440500, 0xffffffff, 0),
1650 	SYSC_QUIRK("epwmss", 0, 0, 0x4, -ENODEV, 0x47400001, 0xffffffff, 0),
1651 	SYSC_QUIRK("gpu", 0, 0x1fc00, 0x1fc10, -ENODEV, 0, 0, 0),
1652 	SYSC_QUIRK("gpu", 0, 0xfe00, 0xfe10, -ENODEV, 0x40000000 , 0xffffffff, 0),
1653 	SYSC_QUIRK("hdmi", 0, 0, 0x10, -ENODEV, 0x50031d00, 0xffffffff, 0),
1654 	SYSC_QUIRK("hsi", 0, 0, 0x10, 0x14, 0x50043101, 0xffffffff, 0),
1655 	SYSC_QUIRK("iss", 0, 0, 0x10, -ENODEV, 0x40000101, 0xffffffff, 0),
1656 	SYSC_QUIRK("keypad", 0x4a31c000, 0, 0x10, 0x14, 0x00000020, 0xffffffff, 0),
1657 	SYSC_QUIRK("mcasp", 0, 0, 0x4, -ENODEV, 0x44307b02, 0xffffffff, 0),
1658 	SYSC_QUIRK("mcbsp", 0, -ENODEV, 0x8c, -ENODEV, 0, 0, 0),
1659 	SYSC_QUIRK("mcspi", 0, 0, 0x10, -ENODEV, 0x40300a0b, 0xffff00ff, 0),
1660 	SYSC_QUIRK("mcspi", 0, 0, 0x110, 0x114, 0x40300a0b, 0xffffffff, 0),
1661 	SYSC_QUIRK("mailbox", 0, 0, 0x10, -ENODEV, 0x00000400, 0xffffffff, 0),
1662 	SYSC_QUIRK("m3", 0, 0, -ENODEV, -ENODEV, 0x5f580105, 0x0fff0f00, 0),
1663 	SYSC_QUIRK("ocp2scp", 0, 0, 0x10, 0x14, 0x50060005, 0xfffffff0, 0),
1664 	SYSC_QUIRK("ocp2scp", 0, 0, -ENODEV, -ENODEV, 0x50060007, 0xffffffff, 0),
1665 	SYSC_QUIRK("padconf", 0, 0, 0x10, -ENODEV, 0x4fff0800, 0xffffffff, 0),
1666 	SYSC_QUIRK("padconf", 0, 0, -ENODEV, -ENODEV, 0x40001100, 0xffffffff, 0),
1667 	SYSC_QUIRK("pcie", 0x51000000, -ENODEV, -ENODEV, -ENODEV, 0, 0, 0),
1668 	SYSC_QUIRK("pcie", 0x51800000, -ENODEV, -ENODEV, -ENODEV, 0, 0, 0),
1669 	SYSC_QUIRK("prcm", 0, 0, -ENODEV, -ENODEV, 0x40000100, 0xffffffff, 0),
1670 	SYSC_QUIRK("prcm", 0, 0, -ENODEV, -ENODEV, 0x00004102, 0xffffffff, 0),
1671 	SYSC_QUIRK("prcm", 0, 0, -ENODEV, -ENODEV, 0x40000400, 0xffffffff, 0),
1672 	SYSC_QUIRK("rfbi", 0x4832a800, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1673 	SYSC_QUIRK("rfbi", 0x58002000, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1674 	SYSC_QUIRK("scm", 0, 0, 0x10, -ENODEV, 0x40000900, 0xffffffff, 0),
1675 	SYSC_QUIRK("scm", 0, 0, -ENODEV, -ENODEV, 0x4e8b0100, 0xffffffff, 0),
1676 	SYSC_QUIRK("scm", 0, 0, -ENODEV, -ENODEV, 0x4f000100, 0xffffffff, 0),
1677 	SYSC_QUIRK("scm", 0, 0, -ENODEV, -ENODEV, 0x40000900, 0xffffffff, 0),
1678 	SYSC_QUIRK("scrm", 0, 0, -ENODEV, -ENODEV, 0x00000010, 0xffffffff, 0),
1679 	SYSC_QUIRK("sdio", 0, 0, 0x10, -ENODEV, 0x40202301, 0xffff0ff0, 0),
1680 	SYSC_QUIRK("sdio", 0, 0x2fc, 0x110, 0x114, 0x31010000, 0xffffffff, 0),
1681 	SYSC_QUIRK("sdma", 0, 0, 0x2c, 0x28, 0x00010900, 0xffffffff, 0),
1682 	SYSC_QUIRK("sham", 0, 0x100, 0x110, 0x114, 0x40000c03, 0xffffffff, 0),
1683 	SYSC_QUIRK("slimbus", 0, 0, 0x10, -ENODEV, 0x40000902, 0xffffffff, 0),
1684 	SYSC_QUIRK("slimbus", 0, 0, 0x10, -ENODEV, 0x40002903, 0xffffffff, 0),
1685 	SYSC_QUIRK("smartreflex", 0, -ENODEV, 0x24, -ENODEV, 0x00000000, 0xffffffff, 0),
1686 	SYSC_QUIRK("smartreflex", 0, -ENODEV, 0x38, -ENODEV, 0x00000000, 0xffffffff, 0),
1687 	SYSC_QUIRK("spinlock", 0, 0, 0x10, -ENODEV, 0x50020000, 0xffffffff, 0),
1688 	SYSC_QUIRK("rng", 0, 0x1fe0, 0x1fe4, -ENODEV, 0x00000020, 0xffffffff, 0),
1689 	SYSC_QUIRK("timer", 0, 0, 0x10, 0x14, 0x00000013, 0xffffffff, 0),
1690 	SYSC_QUIRK("timer", 0, 0, 0x10, 0x14, 0x00000015, 0xffffffff, 0),
1691 	/* Some timers on omap4 and later */
1692 	SYSC_QUIRK("timer", 0, 0, 0x10, -ENODEV, 0x50002100, 0xffffffff, 0),
1693 	SYSC_QUIRK("timer", 0, 0, 0x10, -ENODEV, 0x4fff1301, 0xffff00ff, 0),
1694 	SYSC_QUIRK("timer32k", 0, 0, 0x4, -ENODEV, 0x00000040, 0xffffffff, 0),
1695 	SYSC_QUIRK("timer32k", 0, 0, 0x4, -ENODEV, 0x00000011, 0xffffffff, 0),
1696 	SYSC_QUIRK("timer32k", 0, 0, 0x4, -ENODEV, 0x00000060, 0xffffffff, 0),
1697 	SYSC_QUIRK("tpcc", 0, 0, -ENODEV, -ENODEV, 0x40014c00, 0xffffffff, 0),
1698 	SYSC_QUIRK("usbhstll", 0, 0, 0x10, 0x14, 0x00000004, 0xffffffff, 0),
1699 	SYSC_QUIRK("usbhstll", 0, 0, 0x10, 0x14, 0x00000008, 0xffffffff, 0),
1700 	SYSC_QUIRK("venc", 0x58003000, 0, -ENODEV, -ENODEV, 0x00000002, 0xffffffff, 0),
1701 	SYSC_QUIRK("vfpe", 0, 0, 0x104, -ENODEV, 0x4d001200, 0xffffffff, 0),
1702 #endif
1703 };
1704 
1705 /*
1706  * Early quirks based on module base and register offsets only that are
1707  * needed before the module revision can be read
1708  */
sysc_init_early_quirks(struct sysc * ddata)1709 static void sysc_init_early_quirks(struct sysc *ddata)
1710 {
1711 	const struct sysc_revision_quirk *q;
1712 	int i;
1713 
1714 	for (i = 0; i < ARRAY_SIZE(sysc_revision_quirks); i++) {
1715 		q = &sysc_revision_quirks[i];
1716 
1717 		if (!q->base)
1718 			continue;
1719 
1720 		if (q->base != ddata->module_pa)
1721 			continue;
1722 
1723 		if (q->rev_offset != ddata->offsets[SYSC_REVISION])
1724 			continue;
1725 
1726 		if (q->sysc_offset != ddata->offsets[SYSC_SYSCONFIG])
1727 			continue;
1728 
1729 		if (q->syss_offset != ddata->offsets[SYSC_SYSSTATUS])
1730 			continue;
1731 
1732 		ddata->name = q->name;
1733 		ddata->cfg.quirks |= q->quirks;
1734 	}
1735 }
1736 
1737 /* Quirks that also consider the revision register value */
sysc_init_revision_quirks(struct sysc * ddata)1738 static void sysc_init_revision_quirks(struct sysc *ddata)
1739 {
1740 	const struct sysc_revision_quirk *q;
1741 	int i;
1742 
1743 	for (i = 0; i < ARRAY_SIZE(sysc_revision_quirks); i++) {
1744 		q = &sysc_revision_quirks[i];
1745 
1746 		if (q->base && q->base != ddata->module_pa)
1747 			continue;
1748 
1749 		if (q->rev_offset != ddata->offsets[SYSC_REVISION])
1750 			continue;
1751 
1752 		if (q->sysc_offset != ddata->offsets[SYSC_SYSCONFIG])
1753 			continue;
1754 
1755 		if (q->syss_offset != ddata->offsets[SYSC_SYSSTATUS])
1756 			continue;
1757 
1758 		if (q->revision == ddata->revision ||
1759 		    (q->revision & q->revision_mask) ==
1760 		    (ddata->revision & q->revision_mask)) {
1761 			ddata->name = q->name;
1762 			ddata->cfg.quirks |= q->quirks;
1763 		}
1764 	}
1765 }
1766 
1767 /*
1768  * DSS needs dispc outputs disabled to reset modules. Returns mask of
1769  * enabled DSS interrupts. Eventually we may be able to do this on
1770  * dispc init rather than top-level DSS init.
1771  */
sysc_quirk_dispc(struct sysc * ddata,int dispc_offset,bool disable)1772 static u32 sysc_quirk_dispc(struct sysc *ddata, int dispc_offset,
1773 			    bool disable)
1774 {
1775 	bool lcd_en, digit_en, lcd2_en = false, lcd3_en = false;
1776 	const int lcd_en_mask = BIT(0), digit_en_mask = BIT(1);
1777 	int manager_count;
1778 	bool framedonetv_irq = true;
1779 	u32 val, irq_mask = 0;
1780 
1781 	switch (sysc_soc->soc) {
1782 	case SOC_2420 ... SOC_3630:
1783 		manager_count = 2;
1784 		framedonetv_irq = false;
1785 		break;
1786 	case SOC_4430 ... SOC_4470:
1787 		manager_count = 3;
1788 		break;
1789 	case SOC_5430:
1790 	case SOC_DRA7:
1791 		manager_count = 4;
1792 		break;
1793 	case SOC_AM4:
1794 		manager_count = 1;
1795 		framedonetv_irq = false;
1796 		break;
1797 	case SOC_UNKNOWN:
1798 	default:
1799 		return 0;
1800 	}
1801 
1802 	/* Remap the whole module range to be able to reset dispc outputs */
1803 	devm_iounmap(ddata->dev, ddata->module_va);
1804 	ddata->module_va = devm_ioremap(ddata->dev,
1805 					ddata->module_pa,
1806 					ddata->module_size);
1807 	if (!ddata->module_va)
1808 		return -EIO;
1809 
1810 	/* DISP_CONTROL, shut down lcd and digit on disable if enabled */
1811 	val = sysc_read(ddata, dispc_offset + 0x40);
1812 	lcd_en = val & lcd_en_mask;
1813 	digit_en = val & digit_en_mask;
1814 	if (lcd_en)
1815 		irq_mask |= BIT(0);			/* FRAMEDONE */
1816 	if (digit_en) {
1817 		if (framedonetv_irq)
1818 			irq_mask |= BIT(24);		/* FRAMEDONETV */
1819 		else
1820 			irq_mask |= BIT(2) | BIT(3);	/* EVSYNC bits */
1821 	}
1822 	if (disable && (lcd_en || digit_en))
1823 		sysc_write(ddata, dispc_offset + 0x40,
1824 			   val & ~(lcd_en_mask | digit_en_mask));
1825 
1826 	if (manager_count <= 2)
1827 		return irq_mask;
1828 
1829 	/* DISPC_CONTROL2 */
1830 	val = sysc_read(ddata, dispc_offset + 0x238);
1831 	lcd2_en = val & lcd_en_mask;
1832 	if (lcd2_en)
1833 		irq_mask |= BIT(22);			/* FRAMEDONE2 */
1834 	if (disable && lcd2_en)
1835 		sysc_write(ddata, dispc_offset + 0x238,
1836 			   val & ~lcd_en_mask);
1837 
1838 	if (manager_count <= 3)
1839 		return irq_mask;
1840 
1841 	/* DISPC_CONTROL3 */
1842 	val = sysc_read(ddata, dispc_offset + 0x848);
1843 	lcd3_en = val & lcd_en_mask;
1844 	if (lcd3_en)
1845 		irq_mask |= BIT(30);			/* FRAMEDONE3 */
1846 	if (disable && lcd3_en)
1847 		sysc_write(ddata, dispc_offset + 0x848,
1848 			   val & ~lcd_en_mask);
1849 
1850 	return irq_mask;
1851 }
1852 
1853 /* DSS needs child outputs disabled and SDI registers cleared for reset */
sysc_pre_reset_quirk_dss(struct sysc * ddata)1854 static void sysc_pre_reset_quirk_dss(struct sysc *ddata)
1855 {
1856 	const int dispc_offset = 0x1000;
1857 	int error;
1858 	u32 irq_mask, val;
1859 
1860 	/* Get enabled outputs */
1861 	irq_mask = sysc_quirk_dispc(ddata, dispc_offset, false);
1862 	if (!irq_mask)
1863 		return;
1864 
1865 	/* Clear IRQSTATUS */
1866 	sysc_write(ddata, dispc_offset + 0x18, irq_mask);
1867 
1868 	/* Disable outputs */
1869 	val = sysc_quirk_dispc(ddata, dispc_offset, true);
1870 
1871 	/* Poll IRQSTATUS */
1872 	error = readl_poll_timeout(ddata->module_va + dispc_offset + 0x18,
1873 				   val, val != irq_mask, 100, 50);
1874 	if (error)
1875 		dev_warn(ddata->dev, "%s: timed out %08x !+ %08x\n",
1876 			 __func__, val, irq_mask);
1877 
1878 	if (sysc_soc->soc == SOC_3430 || sysc_soc->soc == SOC_AM35) {
1879 		/* Clear DSS_SDI_CONTROL */
1880 		sysc_write(ddata, 0x44, 0);
1881 
1882 		/* Clear DSS_PLL_CONTROL */
1883 		sysc_write(ddata, 0x48, 0);
1884 	}
1885 
1886 	/* Clear DSS_CONTROL to switch DSS clock sources to PRCM if not */
1887 	sysc_write(ddata, 0x40, 0);
1888 }
1889 
1890 /* 1-wire needs module's internal clocks enabled for reset */
sysc_pre_reset_quirk_hdq1w(struct sysc * ddata)1891 static void sysc_pre_reset_quirk_hdq1w(struct sysc *ddata)
1892 {
1893 	int offset = 0x0c;	/* HDQ_CTRL_STATUS */
1894 	u16 val;
1895 
1896 	val = sysc_read(ddata, offset);
1897 	val |= BIT(5);
1898 	sysc_write(ddata, offset, val);
1899 }
1900 
1901 /* AESS (Audio Engine SubSystem) needs autogating set after enable */
sysc_module_enable_quirk_aess(struct sysc * ddata)1902 static void sysc_module_enable_quirk_aess(struct sysc *ddata)
1903 {
1904 	int offset = 0x7c;	/* AESS_AUTO_GATING_ENABLE */
1905 
1906 	sysc_write(ddata, offset, 1);
1907 }
1908 
1909 /* I2C needs to be disabled for reset */
sysc_clk_quirk_i2c(struct sysc * ddata,bool enable)1910 static void sysc_clk_quirk_i2c(struct sysc *ddata, bool enable)
1911 {
1912 	int offset;
1913 	u16 val;
1914 
1915 	/* I2C_CON, omap2/3 is different from omap4 and later */
1916 	if ((ddata->revision & 0xffffff00) == 0x001f0000)
1917 		offset = 0x24;
1918 	else
1919 		offset = 0xa4;
1920 
1921 	/* I2C_EN */
1922 	val = sysc_read(ddata, offset);
1923 	if (enable)
1924 		val |= BIT(15);
1925 	else
1926 		val &= ~BIT(15);
1927 	sysc_write(ddata, offset, val);
1928 }
1929 
sysc_pre_reset_quirk_i2c(struct sysc * ddata)1930 static void sysc_pre_reset_quirk_i2c(struct sysc *ddata)
1931 {
1932 	sysc_clk_quirk_i2c(ddata, false);
1933 }
1934 
sysc_post_reset_quirk_i2c(struct sysc * ddata)1935 static void sysc_post_reset_quirk_i2c(struct sysc *ddata)
1936 {
1937 	sysc_clk_quirk_i2c(ddata, true);
1938 }
1939 
1940 /* RTC on am3 and 4 needs to be unlocked and locked for sysconfig */
sysc_quirk_rtc(struct sysc * ddata,bool lock)1941 static void sysc_quirk_rtc(struct sysc *ddata, bool lock)
1942 {
1943 	u32 val, kick0_val = 0, kick1_val = 0;
1944 	unsigned long flags;
1945 	int error;
1946 
1947 	if (!lock) {
1948 		kick0_val = 0x83e70b13;
1949 		kick1_val = 0x95a4f1e0;
1950 	}
1951 
1952 	local_irq_save(flags);
1953 	/* RTC_STATUS BUSY bit may stay active for 1/32768 seconds (~30 usec) */
1954 	error = readl_poll_timeout_atomic(ddata->module_va + 0x44, val,
1955 					  !(val & BIT(0)), 100, 50);
1956 	if (error)
1957 		dev_warn(ddata->dev, "rtc busy timeout\n");
1958 	/* Now we have ~15 microseconds to read/write various registers */
1959 	sysc_write(ddata, 0x6c, kick0_val);
1960 	sysc_write(ddata, 0x70, kick1_val);
1961 	local_irq_restore(flags);
1962 }
1963 
sysc_module_unlock_quirk_rtc(struct sysc * ddata)1964 static void sysc_module_unlock_quirk_rtc(struct sysc *ddata)
1965 {
1966 	sysc_quirk_rtc(ddata, false);
1967 }
1968 
sysc_module_lock_quirk_rtc(struct sysc * ddata)1969 static void sysc_module_lock_quirk_rtc(struct sysc *ddata)
1970 {
1971 	sysc_quirk_rtc(ddata, true);
1972 }
1973 
1974 /* OTG omap2430 glue layer up to omap4 needs OTG_FORCESTDBY configured */
sysc_module_enable_quirk_otg(struct sysc * ddata)1975 static void sysc_module_enable_quirk_otg(struct sysc *ddata)
1976 {
1977 	int offset = 0x414;	/* OTG_FORCESTDBY */
1978 
1979 	sysc_write(ddata, offset, 0);
1980 }
1981 
sysc_module_disable_quirk_otg(struct sysc * ddata)1982 static void sysc_module_disable_quirk_otg(struct sysc *ddata)
1983 {
1984 	int offset = 0x414;	/* OTG_FORCESTDBY */
1985 	u32 val = BIT(0);	/* ENABLEFORCE */
1986 
1987 	sysc_write(ddata, offset, val);
1988 }
1989 
1990 /* 36xx SGX needs a quirk for to bypass OCP IPG interrupt logic */
sysc_module_enable_quirk_sgx(struct sysc * ddata)1991 static void sysc_module_enable_quirk_sgx(struct sysc *ddata)
1992 {
1993 	int offset = 0xff08;	/* OCP_DEBUG_CONFIG */
1994 	u32 val = BIT(31);	/* THALIA_INT_BYPASS */
1995 
1996 	sysc_write(ddata, offset, val);
1997 }
1998 
1999 /* Watchdog timer needs a disable sequence after reset */
sysc_reset_done_quirk_wdt(struct sysc * ddata)2000 static void sysc_reset_done_quirk_wdt(struct sysc *ddata)
2001 {
2002 	int wps, spr, error;
2003 	u32 val;
2004 
2005 	wps = 0x34;
2006 	spr = 0x48;
2007 
2008 	sysc_write(ddata, spr, 0xaaaa);
2009 	error = readl_poll_timeout(ddata->module_va + wps, val,
2010 				   !(val & 0x10), 100,
2011 				   MAX_MODULE_SOFTRESET_WAIT);
2012 	if (error)
2013 		dev_warn(ddata->dev, "wdt disable step1 failed\n");
2014 
2015 	sysc_write(ddata, spr, 0x5555);
2016 	error = readl_poll_timeout(ddata->module_va + wps, val,
2017 				   !(val & 0x10), 100,
2018 				   MAX_MODULE_SOFTRESET_WAIT);
2019 	if (error)
2020 		dev_warn(ddata->dev, "wdt disable step2 failed\n");
2021 }
2022 
2023 /* PRUSS needs to set MSTANDBY_INIT inorder to idle properly */
sysc_module_disable_quirk_pruss(struct sysc * ddata)2024 static void sysc_module_disable_quirk_pruss(struct sysc *ddata)
2025 {
2026 	u32 reg;
2027 
2028 	reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
2029 	reg |= SYSC_PRUSS_STANDBY_INIT;
2030 	sysc_write(ddata, ddata->offsets[SYSC_SYSCONFIG], reg);
2031 }
2032 
sysc_init_module_quirks(struct sysc * ddata)2033 static void sysc_init_module_quirks(struct sysc *ddata)
2034 {
2035 	if (ddata->legacy_mode || !ddata->name)
2036 		return;
2037 
2038 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_HDQ1W) {
2039 		ddata->pre_reset_quirk = sysc_pre_reset_quirk_hdq1w;
2040 
2041 		return;
2042 	}
2043 
2044 #ifdef CONFIG_OMAP_GPMC_DEBUG
2045 	if (ddata->cfg.quirks & SYSC_QUIRK_GPMC_DEBUG) {
2046 		ddata->cfg.quirks |= SYSC_QUIRK_NO_RESET_ON_INIT;
2047 
2048 		return;
2049 	}
2050 #endif
2051 
2052 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_I2C) {
2053 		ddata->pre_reset_quirk = sysc_pre_reset_quirk_i2c;
2054 		ddata->post_reset_quirk = sysc_post_reset_quirk_i2c;
2055 
2056 		return;
2057 	}
2058 
2059 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_AESS)
2060 		ddata->module_enable_quirk = sysc_module_enable_quirk_aess;
2061 
2062 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_DSS_RESET)
2063 		ddata->pre_reset_quirk = sysc_pre_reset_quirk_dss;
2064 
2065 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_RTC_UNLOCK) {
2066 		ddata->module_unlock_quirk = sysc_module_unlock_quirk_rtc;
2067 		ddata->module_lock_quirk = sysc_module_lock_quirk_rtc;
2068 
2069 		return;
2070 	}
2071 
2072 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_OTG) {
2073 		ddata->module_enable_quirk = sysc_module_enable_quirk_otg;
2074 		ddata->module_disable_quirk = sysc_module_disable_quirk_otg;
2075 	}
2076 
2077 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_SGX)
2078 		ddata->module_enable_quirk = sysc_module_enable_quirk_sgx;
2079 
2080 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_WDT) {
2081 		ddata->reset_done_quirk = sysc_reset_done_quirk_wdt;
2082 		ddata->module_disable_quirk = sysc_reset_done_quirk_wdt;
2083 	}
2084 
2085 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_PRUSS)
2086 		ddata->module_disable_quirk = sysc_module_disable_quirk_pruss;
2087 }
2088 
sysc_clockdomain_init(struct sysc * ddata)2089 static int sysc_clockdomain_init(struct sysc *ddata)
2090 {
2091 	struct ti_sysc_platform_data *pdata = dev_get_platdata(ddata->dev);
2092 	struct clk *fck = NULL, *ick = NULL;
2093 	int error;
2094 
2095 	if (!pdata || !pdata->init_clockdomain)
2096 		return 0;
2097 
2098 	switch (ddata->nr_clocks) {
2099 	case 2:
2100 		ick = ddata->clocks[SYSC_ICK];
2101 		fallthrough;
2102 	case 1:
2103 		fck = ddata->clocks[SYSC_FCK];
2104 		break;
2105 	case 0:
2106 		return 0;
2107 	}
2108 
2109 	error = pdata->init_clockdomain(ddata->dev, fck, ick, &ddata->cookie);
2110 	if (!error || error == -ENODEV)
2111 		return 0;
2112 
2113 	return error;
2114 }
2115 
2116 /*
2117  * Note that pdata->init_module() typically does a reset first. After
2118  * pdata->init_module() is done, PM runtime can be used for the interconnect
2119  * target module.
2120  */
sysc_legacy_init(struct sysc * ddata)2121 static int sysc_legacy_init(struct sysc *ddata)
2122 {
2123 	struct ti_sysc_platform_data *pdata = dev_get_platdata(ddata->dev);
2124 	int error;
2125 
2126 	if (!pdata || !pdata->init_module)
2127 		return 0;
2128 
2129 	error = pdata->init_module(ddata->dev, ddata->mdata, &ddata->cookie);
2130 	if (error == -EEXIST)
2131 		error = 0;
2132 
2133 	return error;
2134 }
2135 
2136 /*
2137  * Note that the caller must ensure the interconnect target module is enabled
2138  * before calling reset. Otherwise reset will not complete.
2139  */
sysc_reset(struct sysc * ddata)2140 static int sysc_reset(struct sysc *ddata)
2141 {
2142 	int sysc_offset, sysc_val, error;
2143 	u32 sysc_mask;
2144 
2145 	sysc_offset = ddata->offsets[SYSC_SYSCONFIG];
2146 
2147 	if (ddata->legacy_mode ||
2148 	    ddata->cap->regbits->srst_shift < 0 ||
2149 	    ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT)
2150 		return 0;
2151 
2152 	sysc_mask = BIT(ddata->cap->regbits->srst_shift);
2153 
2154 	if (ddata->pre_reset_quirk)
2155 		ddata->pre_reset_quirk(ddata);
2156 
2157 	if (sysc_offset >= 0) {
2158 		sysc_val = sysc_read_sysconfig(ddata);
2159 		sysc_val |= sysc_mask;
2160 		sysc_write(ddata, sysc_offset, sysc_val);
2161 
2162 		/*
2163 		 * Some devices need a delay before reading registers
2164 		 * after reset. Presumably a srst_udelay is not needed
2165 		 * for devices that use a rstctrl register reset.
2166 		 */
2167 		if (ddata->cfg.srst_udelay)
2168 			fsleep(ddata->cfg.srst_udelay);
2169 
2170 		/*
2171 		 * Flush posted write. For devices needing srst_udelay
2172 		 * this should trigger an interconnect error if the
2173 		 * srst_udelay value is needed but not configured.
2174 		 */
2175 		sysc_val = sysc_read_sysconfig(ddata);
2176 	}
2177 
2178 	if (ddata->post_reset_quirk)
2179 		ddata->post_reset_quirk(ddata);
2180 
2181 	error = sysc_wait_softreset(ddata);
2182 	if (error)
2183 		dev_warn(ddata->dev, "OCP softreset timed out\n");
2184 
2185 	if (ddata->reset_done_quirk)
2186 		ddata->reset_done_quirk(ddata);
2187 
2188 	return error;
2189 }
2190 
2191 /*
2192  * At this point the module is configured enough to read the revision but
2193  * module may not be completely configured yet to use PM runtime. Enable
2194  * all clocks directly during init to configure the quirks needed for PM
2195  * runtime based on the revision register.
2196  */
sysc_init_module(struct sysc * ddata)2197 static int sysc_init_module(struct sysc *ddata)
2198 {
2199 	bool rstctrl_deasserted = false;
2200 	int error = 0;
2201 
2202 	error = sysc_clockdomain_init(ddata);
2203 	if (error)
2204 		return error;
2205 
2206 	sysc_clkdm_deny_idle(ddata);
2207 
2208 	/*
2209 	 * Always enable clocks. The bootloader may or may not have enabled
2210 	 * the related clocks.
2211 	 */
2212 	error = sysc_enable_opt_clocks(ddata);
2213 	if (error)
2214 		return error;
2215 
2216 	error = sysc_enable_main_clocks(ddata);
2217 	if (error)
2218 		goto err_opt_clocks;
2219 
2220 	if (!(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT)) {
2221 		error = reset_control_deassert(ddata->rsts);
2222 		if (error)
2223 			goto err_main_clocks;
2224 		rstctrl_deasserted = true;
2225 	}
2226 
2227 	ddata->revision = sysc_read_revision(ddata);
2228 	sysc_init_revision_quirks(ddata);
2229 	sysc_init_module_quirks(ddata);
2230 
2231 	if (ddata->legacy_mode) {
2232 		error = sysc_legacy_init(ddata);
2233 		if (error)
2234 			goto err_main_clocks;
2235 	}
2236 
2237 	if (!ddata->legacy_mode) {
2238 		error = sysc_enable_module(ddata->dev);
2239 		if (error)
2240 			goto err_main_clocks;
2241 	}
2242 
2243 	error = sysc_reset(ddata);
2244 	if (error)
2245 		dev_err(ddata->dev, "Reset failed with %d\n", error);
2246 
2247 	if (error && !ddata->legacy_mode)
2248 		sysc_disable_module(ddata->dev);
2249 
2250 err_main_clocks:
2251 	if (error)
2252 		sysc_disable_main_clocks(ddata);
2253 err_opt_clocks:
2254 	/* No re-enable of clockdomain autoidle to prevent module autoidle */
2255 	if (error) {
2256 		sysc_disable_opt_clocks(ddata);
2257 		sysc_clkdm_allow_idle(ddata);
2258 	}
2259 
2260 	if (error && rstctrl_deasserted &&
2261 	    !(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT))
2262 		reset_control_assert(ddata->rsts);
2263 
2264 	return error;
2265 }
2266 
sysc_init_sysc_mask(struct sysc * ddata)2267 static int sysc_init_sysc_mask(struct sysc *ddata)
2268 {
2269 	struct device_node *np = ddata->dev->of_node;
2270 	int error;
2271 	u32 val;
2272 
2273 	error = of_property_read_u32(np, "ti,sysc-mask", &val);
2274 	if (error)
2275 		return 0;
2276 
2277 	ddata->cfg.sysc_val = val & ddata->cap->sysc_mask;
2278 
2279 	return 0;
2280 }
2281 
sysc_init_idlemode(struct sysc * ddata,u8 * idlemodes,const char * name)2282 static int sysc_init_idlemode(struct sysc *ddata, u8 *idlemodes,
2283 			      const char *name)
2284 {
2285 	struct device_node *np = ddata->dev->of_node;
2286 	u32 val;
2287 
2288 	of_property_for_each_u32(np, name, val) {
2289 		if (val >= SYSC_NR_IDLEMODES) {
2290 			dev_err(ddata->dev, "invalid idlemode: %i\n", val);
2291 			return -EINVAL;
2292 		}
2293 		*idlemodes |=  (1 << val);
2294 	}
2295 
2296 	return 0;
2297 }
2298 
sysc_init_idlemodes(struct sysc * ddata)2299 static int sysc_init_idlemodes(struct sysc *ddata)
2300 {
2301 	int error;
2302 
2303 	error = sysc_init_idlemode(ddata, &ddata->cfg.midlemodes,
2304 				   "ti,sysc-midle");
2305 	if (error)
2306 		return error;
2307 
2308 	error = sysc_init_idlemode(ddata, &ddata->cfg.sidlemodes,
2309 				   "ti,sysc-sidle");
2310 	if (error)
2311 		return error;
2312 
2313 	return 0;
2314 }
2315 
2316 /*
2317  * Only some devices on omap4 and later have SYSCONFIG reset done
2318  * bit. We can detect this if there is no SYSSTATUS at all, or the
2319  * SYSTATUS bit 0 is not used. Note that some SYSSTATUS registers
2320  * have multiple bits for the child devices like OHCI and EHCI.
2321  * Depends on SYSC being parsed first.
2322  */
sysc_init_syss_mask(struct sysc * ddata)2323 static int sysc_init_syss_mask(struct sysc *ddata)
2324 {
2325 	struct device_node *np = ddata->dev->of_node;
2326 	int error;
2327 	u32 val;
2328 
2329 	error = of_property_read_u32(np, "ti,syss-mask", &val);
2330 	if (error) {
2331 		if ((ddata->cap->type == TI_SYSC_OMAP4 ||
2332 		     ddata->cap->type == TI_SYSC_OMAP4_TIMER) &&
2333 		    (ddata->cfg.sysc_val & SYSC_OMAP4_SOFTRESET))
2334 			ddata->cfg.quirks |= SYSC_QUIRK_RESET_STATUS;
2335 
2336 		return 0;
2337 	}
2338 
2339 	if (!(val & 1) && (ddata->cfg.sysc_val & SYSC_OMAP4_SOFTRESET))
2340 		ddata->cfg.quirks |= SYSC_QUIRK_RESET_STATUS;
2341 
2342 	ddata->cfg.syss_mask = val;
2343 
2344 	return 0;
2345 }
2346 
2347 /*
2348  * Many child device drivers need to have fck and opt clocks available
2349  * to get the clock rate for device internal configuration etc.
2350  */
sysc_child_add_named_clock(struct sysc * ddata,struct device * child,const char * name)2351 static int sysc_child_add_named_clock(struct sysc *ddata,
2352 				      struct device *child,
2353 				      const char *name)
2354 {
2355 	struct clk *clk;
2356 	struct clk_lookup *l;
2357 	int error = 0;
2358 
2359 	if (!name)
2360 		return 0;
2361 
2362 	clk = clk_get(child, name);
2363 	if (!IS_ERR(clk)) {
2364 		error = -EEXIST;
2365 		goto put_clk;
2366 	}
2367 
2368 	clk = clk_get(ddata->dev, name);
2369 	if (IS_ERR(clk))
2370 		return -ENODEV;
2371 
2372 	l = clkdev_create(clk, name, dev_name(child));
2373 	if (!l)
2374 		error = -ENOMEM;
2375 put_clk:
2376 	clk_put(clk);
2377 
2378 	return error;
2379 }
2380 
sysc_child_add_clocks(struct sysc * ddata,struct device * child)2381 static int sysc_child_add_clocks(struct sysc *ddata,
2382 				 struct device *child)
2383 {
2384 	int i, error;
2385 
2386 	for (i = 0; i < ddata->nr_clocks; i++) {
2387 		error = sysc_child_add_named_clock(ddata,
2388 						   child,
2389 						   ddata->clock_roles[i]);
2390 		if (error && error != -EEXIST) {
2391 			dev_err(ddata->dev, "could not add child clock %s: %i\n",
2392 				ddata->clock_roles[i], error);
2393 
2394 			return error;
2395 		}
2396 	}
2397 
2398 	return 0;
2399 }
2400 
2401 static struct device_type sysc_device_type = {
2402 };
2403 
sysc_child_to_parent(struct device * dev)2404 static struct sysc *sysc_child_to_parent(struct device *dev)
2405 {
2406 	struct device *parent = dev->parent;
2407 
2408 	if (!parent || parent->type != &sysc_device_type)
2409 		return NULL;
2410 
2411 	return dev_get_drvdata(parent);
2412 }
2413 
sysc_child_runtime_suspend(struct device * dev)2414 static int __maybe_unused sysc_child_runtime_suspend(struct device *dev)
2415 {
2416 	struct sysc *ddata;
2417 	int error;
2418 
2419 	ddata = sysc_child_to_parent(dev);
2420 
2421 	error = pm_generic_runtime_suspend(dev);
2422 	if (error)
2423 		return error;
2424 
2425 	if (!ddata->enabled)
2426 		return 0;
2427 
2428 	return sysc_runtime_suspend(ddata->dev);
2429 }
2430 
sysc_child_runtime_resume(struct device * dev)2431 static int __maybe_unused sysc_child_runtime_resume(struct device *dev)
2432 {
2433 	struct sysc *ddata;
2434 	int error;
2435 
2436 	ddata = sysc_child_to_parent(dev);
2437 
2438 	if (!ddata->enabled) {
2439 		error = sysc_runtime_resume(ddata->dev);
2440 		if (error < 0)
2441 			dev_err(ddata->dev,
2442 				"%s error: %i\n", __func__, error);
2443 	}
2444 
2445 	return pm_generic_runtime_resume(dev);
2446 }
2447 
2448 #ifdef CONFIG_PM_SLEEP
sysc_child_suspend_noirq(struct device * dev)2449 static int sysc_child_suspend_noirq(struct device *dev)
2450 {
2451 	struct sysc *ddata;
2452 	int error;
2453 
2454 	ddata = sysc_child_to_parent(dev);
2455 
2456 	dev_dbg(ddata->dev, "%s %s\n", __func__,
2457 		ddata->name ? ddata->name : "");
2458 
2459 	error = pm_generic_suspend_noirq(dev);
2460 	if (error) {
2461 		dev_err(dev, "%s error at %i: %i\n",
2462 			__func__, __LINE__, error);
2463 
2464 		return error;
2465 	}
2466 
2467 	if (!pm_runtime_status_suspended(dev)) {
2468 		error = pm_generic_runtime_suspend(dev);
2469 		if (error) {
2470 			dev_dbg(dev, "%s busy at %i: %i\n",
2471 				__func__, __LINE__, error);
2472 
2473 			return 0;
2474 		}
2475 
2476 		error = sysc_runtime_suspend(ddata->dev);
2477 		if (error) {
2478 			dev_err(dev, "%s error at %i: %i\n",
2479 				__func__, __LINE__, error);
2480 
2481 			return error;
2482 		}
2483 
2484 		ddata->child_needs_resume = true;
2485 	}
2486 
2487 	return 0;
2488 }
2489 
sysc_child_resume_noirq(struct device * dev)2490 static int sysc_child_resume_noirq(struct device *dev)
2491 {
2492 	struct sysc *ddata;
2493 	int error;
2494 
2495 	ddata = sysc_child_to_parent(dev);
2496 
2497 	dev_dbg(ddata->dev, "%s %s\n", __func__,
2498 		ddata->name ? ddata->name : "");
2499 
2500 	if (ddata->child_needs_resume) {
2501 		ddata->child_needs_resume = false;
2502 
2503 		error = sysc_runtime_resume(ddata->dev);
2504 		if (error)
2505 			dev_err(ddata->dev,
2506 				"%s runtime resume error: %i\n",
2507 				__func__, error);
2508 
2509 		error = pm_generic_runtime_resume(dev);
2510 		if (error)
2511 			dev_err(ddata->dev,
2512 				"%s generic runtime resume: %i\n",
2513 				__func__, error);
2514 	}
2515 
2516 	return pm_generic_resume_noirq(dev);
2517 }
2518 #endif
2519 
2520 static struct dev_pm_domain sysc_child_pm_domain = {
2521 	.ops = {
2522 		SET_RUNTIME_PM_OPS(sysc_child_runtime_suspend,
2523 				   sysc_child_runtime_resume,
2524 				   NULL)
2525 		USE_PLATFORM_PM_SLEEP_OPS
2526 		SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(sysc_child_suspend_noirq,
2527 					      sysc_child_resume_noirq)
2528 	}
2529 };
2530 
2531 /* Caller needs to take list_lock if ever used outside of cpu_pm */
sysc_reinit_modules(struct sysc_soc_info * soc)2532 static void sysc_reinit_modules(struct sysc_soc_info *soc)
2533 {
2534 	struct sysc_module *module;
2535 	struct sysc *ddata;
2536 
2537 	list_for_each_entry(module, &sysc_soc->restored_modules, node) {
2538 		ddata = module->ddata;
2539 		sysc_reinit_module(ddata, ddata->enabled);
2540 	}
2541 }
2542 
2543 /**
2544  * sysc_context_notifier - optionally reset and restore module after idle
2545  * @nb: notifier block
2546  * @cmd: unused
2547  * @v: unused
2548  *
2549  * Some interconnect target modules need to be restored, or reset and restored
2550  * on CPU_PM CPU_PM_CLUSTER_EXIT notifier. This is needed at least for am335x
2551  * OTG and GPMC target modules even if the modules are unused.
2552  */
sysc_context_notifier(struct notifier_block * nb,unsigned long cmd,void * v)2553 static int sysc_context_notifier(struct notifier_block *nb, unsigned long cmd,
2554 				 void *v)
2555 {
2556 	struct sysc_soc_info *soc;
2557 
2558 	soc = container_of(nb, struct sysc_soc_info, nb);
2559 
2560 	switch (cmd) {
2561 	case CPU_CLUSTER_PM_ENTER:
2562 		break;
2563 	case CPU_CLUSTER_PM_ENTER_FAILED:	/* No need to restore context */
2564 		break;
2565 	case CPU_CLUSTER_PM_EXIT:
2566 		sysc_reinit_modules(soc);
2567 		break;
2568 	}
2569 
2570 	return NOTIFY_OK;
2571 }
2572 
2573 /**
2574  * sysc_add_restored - optionally add reset and restore quirk hanlling
2575  * @ddata: device data
2576  */
sysc_add_restored(struct sysc * ddata)2577 static void sysc_add_restored(struct sysc *ddata)
2578 {
2579 	struct sysc_module *restored_module;
2580 
2581 	restored_module = kzalloc(sizeof(*restored_module), GFP_KERNEL);
2582 	if (!restored_module)
2583 		return;
2584 
2585 	restored_module->ddata = ddata;
2586 
2587 	mutex_lock(&sysc_soc->list_lock);
2588 
2589 	list_add(&restored_module->node, &sysc_soc->restored_modules);
2590 
2591 	if (sysc_soc->nb.notifier_call)
2592 		goto out_unlock;
2593 
2594 	sysc_soc->nb.notifier_call = sysc_context_notifier;
2595 	cpu_pm_register_notifier(&sysc_soc->nb);
2596 
2597 out_unlock:
2598 	mutex_unlock(&sysc_soc->list_lock);
2599 }
2600 
2601 /**
2602  * sysc_legacy_idle_quirk - handle children in omap_device compatible way
2603  * @ddata: device driver data
2604  * @child: child device driver
2605  *
2606  * Allow idle for child devices as done with _od_runtime_suspend().
2607  * Otherwise many child devices will not idle because of the permanent
2608  * parent usecount set in pm_runtime_irq_safe().
2609  *
2610  * Note that the long term solution is to just modify the child device
2611  * drivers to not set pm_runtime_irq_safe() and then this can be just
2612  * dropped.
2613  */
sysc_legacy_idle_quirk(struct sysc * ddata,struct device * child)2614 static void sysc_legacy_idle_quirk(struct sysc *ddata, struct device *child)
2615 {
2616 	if (ddata->cfg.quirks & SYSC_QUIRK_LEGACY_IDLE)
2617 		dev_pm_domain_set(child, &sysc_child_pm_domain);
2618 }
2619 
sysc_notifier_call(struct notifier_block * nb,unsigned long event,void * device)2620 static int sysc_notifier_call(struct notifier_block *nb,
2621 			      unsigned long event, void *device)
2622 {
2623 	struct device *dev = device;
2624 	struct sysc *ddata;
2625 	int error;
2626 
2627 	ddata = sysc_child_to_parent(dev);
2628 	if (!ddata)
2629 		return NOTIFY_DONE;
2630 
2631 	switch (event) {
2632 	case BUS_NOTIFY_ADD_DEVICE:
2633 		error = sysc_child_add_clocks(ddata, dev);
2634 		if (error)
2635 			return error;
2636 		sysc_legacy_idle_quirk(ddata, dev);
2637 		break;
2638 	default:
2639 		break;
2640 	}
2641 
2642 	return NOTIFY_DONE;
2643 }
2644 
2645 static struct notifier_block sysc_nb = {
2646 	.notifier_call = sysc_notifier_call,
2647 };
2648 
2649 /* Device tree configured quirks */
2650 struct sysc_dts_quirk {
2651 	const char *name;
2652 	u32 mask;
2653 };
2654 
2655 static const struct sysc_dts_quirk sysc_dts_quirks[] = {
2656 	{ .name = "ti,no-idle-on-init",
2657 	  .mask = SYSC_QUIRK_NO_IDLE_ON_INIT, },
2658 	{ .name = "ti,no-reset-on-init",
2659 	  .mask = SYSC_QUIRK_NO_RESET_ON_INIT, },
2660 	{ .name = "ti,no-idle",
2661 	  .mask = SYSC_QUIRK_NO_IDLE, },
2662 };
2663 
sysc_parse_dts_quirks(struct sysc * ddata,struct device_node * np,bool is_child)2664 static void sysc_parse_dts_quirks(struct sysc *ddata, struct device_node *np,
2665 				  bool is_child)
2666 {
2667 	const struct property *prop;
2668 	int i, len;
2669 
2670 	for (i = 0; i < ARRAY_SIZE(sysc_dts_quirks); i++) {
2671 		const char *name = sysc_dts_quirks[i].name;
2672 
2673 		prop = of_get_property(np, name, &len);
2674 		if (!prop)
2675 			continue;
2676 
2677 		ddata->cfg.quirks |= sysc_dts_quirks[i].mask;
2678 		if (is_child) {
2679 			dev_warn(ddata->dev,
2680 				 "dts flag should be at module level for %s\n",
2681 				 name);
2682 		}
2683 	}
2684 }
2685 
sysc_init_dts_quirks(struct sysc * ddata)2686 static int sysc_init_dts_quirks(struct sysc *ddata)
2687 {
2688 	struct device_node *np = ddata->dev->of_node;
2689 	int error;
2690 	u32 val;
2691 
2692 	ddata->legacy_mode = of_get_property(np, "ti,hwmods", NULL);
2693 
2694 	sysc_parse_dts_quirks(ddata, np, false);
2695 	error = of_property_read_u32(np, "ti,sysc-delay-us", &val);
2696 	if (!error) {
2697 		if (val > 255) {
2698 			dev_warn(ddata->dev, "bad ti,sysc-delay-us: %i\n",
2699 				 val);
2700 		}
2701 
2702 		ddata->cfg.srst_udelay = (u8)val;
2703 	}
2704 
2705 	return 0;
2706 }
2707 
sysc_unprepare(struct sysc * ddata)2708 static void sysc_unprepare(struct sysc *ddata)
2709 {
2710 	int i;
2711 
2712 	if (!ddata->clocks)
2713 		return;
2714 
2715 	for (i = 0; i < SYSC_MAX_CLOCKS; i++) {
2716 		if (!IS_ERR_OR_NULL(ddata->clocks[i]))
2717 			clk_unprepare(ddata->clocks[i]);
2718 	}
2719 }
2720 
2721 /*
2722  * Common sysc register bits found on omap2, also known as type1
2723  */
2724 static const struct sysc_regbits sysc_regbits_omap2 = {
2725 	.dmadisable_shift = -ENODEV,
2726 	.midle_shift = 12,
2727 	.sidle_shift = 3,
2728 	.clkact_shift = 8,
2729 	.emufree_shift = 5,
2730 	.enwkup_shift = 2,
2731 	.srst_shift = 1,
2732 	.autoidle_shift = 0,
2733 };
2734 
2735 static const struct sysc_capabilities sysc_omap2 = {
2736 	.type = TI_SYSC_OMAP2,
2737 	.sysc_mask = SYSC_OMAP2_CLOCKACTIVITY | SYSC_OMAP2_EMUFREE |
2738 		     SYSC_OMAP2_ENAWAKEUP | SYSC_OMAP2_SOFTRESET |
2739 		     SYSC_OMAP2_AUTOIDLE,
2740 	.regbits = &sysc_regbits_omap2,
2741 };
2742 
2743 /* All omap2 and 3 timers, and timers 1, 2 & 10 on omap 4 and 5 */
2744 static const struct sysc_capabilities sysc_omap2_timer = {
2745 	.type = TI_SYSC_OMAP2_TIMER,
2746 	.sysc_mask = SYSC_OMAP2_CLOCKACTIVITY | SYSC_OMAP2_EMUFREE |
2747 		     SYSC_OMAP2_ENAWAKEUP | SYSC_OMAP2_SOFTRESET |
2748 		     SYSC_OMAP2_AUTOIDLE,
2749 	.regbits = &sysc_regbits_omap2,
2750 	.mod_quirks = SYSC_QUIRK_USE_CLOCKACT,
2751 };
2752 
2753 /*
2754  * SHAM2 (SHA1/MD5) sysc found on omap3, a variant of sysc_regbits_omap2
2755  * with different sidle position
2756  */
2757 static const struct sysc_regbits sysc_regbits_omap3_sham = {
2758 	.dmadisable_shift = -ENODEV,
2759 	.midle_shift = -ENODEV,
2760 	.sidle_shift = 4,
2761 	.clkact_shift = -ENODEV,
2762 	.enwkup_shift = -ENODEV,
2763 	.srst_shift = 1,
2764 	.autoidle_shift = 0,
2765 	.emufree_shift = -ENODEV,
2766 };
2767 
2768 static const struct sysc_capabilities sysc_omap3_sham = {
2769 	.type = TI_SYSC_OMAP3_SHAM,
2770 	.sysc_mask = SYSC_OMAP2_SOFTRESET | SYSC_OMAP2_AUTOIDLE,
2771 	.regbits = &sysc_regbits_omap3_sham,
2772 };
2773 
2774 /*
2775  * AES register bits found on omap3 and later, a variant of
2776  * sysc_regbits_omap2 with different sidle position
2777  */
2778 static const struct sysc_regbits sysc_regbits_omap3_aes = {
2779 	.dmadisable_shift = -ENODEV,
2780 	.midle_shift = -ENODEV,
2781 	.sidle_shift = 6,
2782 	.clkact_shift = -ENODEV,
2783 	.enwkup_shift = -ENODEV,
2784 	.srst_shift = 1,
2785 	.autoidle_shift = 0,
2786 	.emufree_shift = -ENODEV,
2787 };
2788 
2789 static const struct sysc_capabilities sysc_omap3_aes = {
2790 	.type = TI_SYSC_OMAP3_AES,
2791 	.sysc_mask = SYSC_OMAP2_SOFTRESET | SYSC_OMAP2_AUTOIDLE,
2792 	.regbits = &sysc_regbits_omap3_aes,
2793 };
2794 
2795 /*
2796  * Common sysc register bits found on omap4, also known as type2
2797  */
2798 static const struct sysc_regbits sysc_regbits_omap4 = {
2799 	.dmadisable_shift = 16,
2800 	.midle_shift = 4,
2801 	.sidle_shift = 2,
2802 	.clkact_shift = -ENODEV,
2803 	.enwkup_shift = -ENODEV,
2804 	.emufree_shift = 1,
2805 	.srst_shift = 0,
2806 	.autoidle_shift = -ENODEV,
2807 };
2808 
2809 static const struct sysc_capabilities sysc_omap4 = {
2810 	.type = TI_SYSC_OMAP4,
2811 	.sysc_mask = SYSC_OMAP4_DMADISABLE | SYSC_OMAP4_FREEEMU |
2812 		     SYSC_OMAP4_SOFTRESET,
2813 	.regbits = &sysc_regbits_omap4,
2814 };
2815 
2816 static const struct sysc_capabilities sysc_omap4_timer = {
2817 	.type = TI_SYSC_OMAP4_TIMER,
2818 	.sysc_mask = SYSC_OMAP4_DMADISABLE | SYSC_OMAP4_FREEEMU |
2819 		     SYSC_OMAP4_SOFTRESET,
2820 	.regbits = &sysc_regbits_omap4,
2821 };
2822 
2823 /*
2824  * Common sysc register bits found on omap4, also known as type3
2825  */
2826 static const struct sysc_regbits sysc_regbits_omap4_simple = {
2827 	.dmadisable_shift = -ENODEV,
2828 	.midle_shift = 2,
2829 	.sidle_shift = 0,
2830 	.clkact_shift = -ENODEV,
2831 	.enwkup_shift = -ENODEV,
2832 	.srst_shift = -ENODEV,
2833 	.emufree_shift = -ENODEV,
2834 	.autoidle_shift = -ENODEV,
2835 };
2836 
2837 static const struct sysc_capabilities sysc_omap4_simple = {
2838 	.type = TI_SYSC_OMAP4_SIMPLE,
2839 	.regbits = &sysc_regbits_omap4_simple,
2840 };
2841 
2842 /*
2843  * SmartReflex sysc found on omap34xx
2844  */
2845 static const struct sysc_regbits sysc_regbits_omap34xx_sr = {
2846 	.dmadisable_shift = -ENODEV,
2847 	.midle_shift = -ENODEV,
2848 	.sidle_shift = -ENODEV,
2849 	.clkact_shift = 20,
2850 	.enwkup_shift = -ENODEV,
2851 	.srst_shift = -ENODEV,
2852 	.emufree_shift = -ENODEV,
2853 	.autoidle_shift = -ENODEV,
2854 };
2855 
2856 static const struct sysc_capabilities sysc_34xx_sr = {
2857 	.type = TI_SYSC_OMAP34XX_SR,
2858 	.sysc_mask = SYSC_OMAP2_CLOCKACTIVITY,
2859 	.regbits = &sysc_regbits_omap34xx_sr,
2860 	.mod_quirks = SYSC_QUIRK_USE_CLOCKACT | SYSC_QUIRK_UNCACHED |
2861 		      SYSC_QUIRK_LEGACY_IDLE,
2862 };
2863 
2864 /*
2865  * SmartReflex sysc found on omap36xx and later
2866  */
2867 static const struct sysc_regbits sysc_regbits_omap36xx_sr = {
2868 	.dmadisable_shift = -ENODEV,
2869 	.midle_shift = -ENODEV,
2870 	.sidle_shift = 24,
2871 	.clkact_shift = -ENODEV,
2872 	.enwkup_shift = 26,
2873 	.srst_shift = -ENODEV,
2874 	.emufree_shift = -ENODEV,
2875 	.autoidle_shift = -ENODEV,
2876 };
2877 
2878 static const struct sysc_capabilities sysc_36xx_sr = {
2879 	.type = TI_SYSC_OMAP36XX_SR,
2880 	.sysc_mask = SYSC_OMAP3_SR_ENAWAKEUP,
2881 	.regbits = &sysc_regbits_omap36xx_sr,
2882 	.mod_quirks = SYSC_QUIRK_UNCACHED | SYSC_QUIRK_LEGACY_IDLE,
2883 };
2884 
2885 static const struct sysc_capabilities sysc_omap4_sr = {
2886 	.type = TI_SYSC_OMAP4_SR,
2887 	.regbits = &sysc_regbits_omap36xx_sr,
2888 	.mod_quirks = SYSC_QUIRK_LEGACY_IDLE,
2889 };
2890 
2891 /*
2892  * McASP register bits found on omap4 and later
2893  */
2894 static const struct sysc_regbits sysc_regbits_omap4_mcasp = {
2895 	.dmadisable_shift = -ENODEV,
2896 	.midle_shift = -ENODEV,
2897 	.sidle_shift = 0,
2898 	.clkact_shift = -ENODEV,
2899 	.enwkup_shift = -ENODEV,
2900 	.srst_shift = -ENODEV,
2901 	.emufree_shift = -ENODEV,
2902 	.autoidle_shift = -ENODEV,
2903 };
2904 
2905 static const struct sysc_capabilities sysc_omap4_mcasp = {
2906 	.type = TI_SYSC_OMAP4_MCASP,
2907 	.regbits = &sysc_regbits_omap4_mcasp,
2908 	.mod_quirks = SYSC_QUIRK_OPT_CLKS_NEEDED,
2909 };
2910 
2911 /*
2912  * McASP found on dra7 and later
2913  */
2914 static const struct sysc_capabilities sysc_dra7_mcasp = {
2915 	.type = TI_SYSC_OMAP4_SIMPLE,
2916 	.regbits = &sysc_regbits_omap4_simple,
2917 	.mod_quirks = SYSC_QUIRK_OPT_CLKS_NEEDED,
2918 };
2919 
2920 /*
2921  * FS USB host found on omap4 and later
2922  */
2923 static const struct sysc_regbits sysc_regbits_omap4_usb_host_fs = {
2924 	.dmadisable_shift = -ENODEV,
2925 	.midle_shift = -ENODEV,
2926 	.sidle_shift = 24,
2927 	.clkact_shift = -ENODEV,
2928 	.enwkup_shift = 26,
2929 	.srst_shift = -ENODEV,
2930 	.emufree_shift = -ENODEV,
2931 	.autoidle_shift = -ENODEV,
2932 };
2933 
2934 static const struct sysc_capabilities sysc_omap4_usb_host_fs = {
2935 	.type = TI_SYSC_OMAP4_USB_HOST_FS,
2936 	.sysc_mask = SYSC_OMAP2_ENAWAKEUP,
2937 	.regbits = &sysc_regbits_omap4_usb_host_fs,
2938 };
2939 
2940 static const struct sysc_regbits sysc_regbits_dra7_mcan = {
2941 	.dmadisable_shift = -ENODEV,
2942 	.midle_shift = -ENODEV,
2943 	.sidle_shift = -ENODEV,
2944 	.clkact_shift = -ENODEV,
2945 	.enwkup_shift = 4,
2946 	.srst_shift = 0,
2947 	.emufree_shift = -ENODEV,
2948 	.autoidle_shift = -ENODEV,
2949 };
2950 
2951 static const struct sysc_capabilities sysc_dra7_mcan = {
2952 	.type = TI_SYSC_DRA7_MCAN,
2953 	.sysc_mask = SYSC_DRA7_MCAN_ENAWAKEUP | SYSC_OMAP4_SOFTRESET,
2954 	.regbits = &sysc_regbits_dra7_mcan,
2955 	.mod_quirks = SYSS_QUIRK_RESETDONE_INVERTED,
2956 };
2957 
2958 /*
2959  * PRUSS found on some AM33xx, AM437x and AM57xx SoCs
2960  */
2961 static const struct sysc_capabilities sysc_pruss = {
2962 	.type = TI_SYSC_PRUSS,
2963 	.sysc_mask = SYSC_PRUSS_STANDBY_INIT | SYSC_PRUSS_SUB_MWAIT,
2964 	.regbits = &sysc_regbits_omap4_simple,
2965 	.mod_quirks = SYSC_MODULE_QUIRK_PRUSS,
2966 };
2967 
sysc_init_pdata(struct sysc * ddata)2968 static int sysc_init_pdata(struct sysc *ddata)
2969 {
2970 	struct ti_sysc_platform_data *pdata = dev_get_platdata(ddata->dev);
2971 	struct ti_sysc_module_data *mdata;
2972 
2973 	if (!pdata)
2974 		return 0;
2975 
2976 	mdata = devm_kzalloc(ddata->dev, sizeof(*mdata), GFP_KERNEL);
2977 	if (!mdata)
2978 		return -ENOMEM;
2979 
2980 	if (ddata->legacy_mode) {
2981 		mdata->name = ddata->legacy_mode;
2982 		mdata->module_pa = ddata->module_pa;
2983 		mdata->module_size = ddata->module_size;
2984 		mdata->offsets = ddata->offsets;
2985 		mdata->nr_offsets = SYSC_MAX_REGS;
2986 		mdata->cap = ddata->cap;
2987 		mdata->cfg = &ddata->cfg;
2988 	}
2989 
2990 	ddata->mdata = mdata;
2991 
2992 	return 0;
2993 }
2994 
sysc_init_match(struct sysc * ddata)2995 static int sysc_init_match(struct sysc *ddata)
2996 {
2997 	const struct sysc_capabilities *cap;
2998 
2999 	cap = of_device_get_match_data(ddata->dev);
3000 	if (!cap)
3001 		return -EINVAL;
3002 
3003 	ddata->cap = cap;
3004 	if (ddata->cap)
3005 		ddata->cfg.quirks |= ddata->cap->mod_quirks;
3006 
3007 	return 0;
3008 }
3009 
ti_sysc_idle(struct work_struct * work)3010 static void ti_sysc_idle(struct work_struct *work)
3011 {
3012 	struct sysc *ddata;
3013 
3014 	ddata = container_of(work, struct sysc, idle_work.work);
3015 
3016 	/*
3017 	 * One time decrement of clock usage counts if left on from init.
3018 	 * Note that we disable opt clocks unconditionally in this case
3019 	 * as they are enabled unconditionally during init without
3020 	 * considering sysc_opt_clks_needed() at that point.
3021 	 */
3022 	if (ddata->cfg.quirks & (SYSC_QUIRK_NO_IDLE |
3023 				 SYSC_QUIRK_NO_IDLE_ON_INIT)) {
3024 		sysc_disable_main_clocks(ddata);
3025 		sysc_disable_opt_clocks(ddata);
3026 		sysc_clkdm_allow_idle(ddata);
3027 	}
3028 
3029 	/* Keep permanent PM runtime usage count for SYSC_QUIRK_NO_IDLE */
3030 	if (ddata->cfg.quirks & SYSC_QUIRK_NO_IDLE)
3031 		return;
3032 
3033 	/*
3034 	 * Decrement PM runtime usage count for SYSC_QUIRK_NO_IDLE_ON_INIT
3035 	 * and SYSC_QUIRK_NO_RESET_ON_INIT
3036 	 */
3037 	if (pm_runtime_active(ddata->dev))
3038 		pm_runtime_put_sync(ddata->dev);
3039 }
3040 
3041 /*
3042  * SoC model and features detection. Only needed for SoCs that need
3043  * special handling for quirks, no need to list others.
3044  */
3045 static const struct soc_device_attribute sysc_soc_match[] = {
3046 	SOC_FLAG("OMAP242*", SOC_2420),
3047 	SOC_FLAG("OMAP243*", SOC_2430),
3048 	SOC_FLAG("AM35*", SOC_AM35),
3049 	SOC_FLAG("OMAP3[45]*", SOC_3430),
3050 	SOC_FLAG("OMAP3[67]*", SOC_3630),
3051 	SOC_FLAG("OMAP443*", SOC_4430),
3052 	SOC_FLAG("OMAP446*", SOC_4460),
3053 	SOC_FLAG("OMAP447*", SOC_4470),
3054 	SOC_FLAG("OMAP54*", SOC_5430),
3055 	SOC_FLAG("AM433", SOC_AM3),
3056 	SOC_FLAG("AM43*", SOC_AM4),
3057 	SOC_FLAG("DRA7*", SOC_DRA7),
3058 
3059 	{ /* sentinel */ }
3060 };
3061 
3062 /*
3063  * List of SoCs variants with disabled features. By default we assume all
3064  * devices in the device tree are available so no need to list those SoCs.
3065  */
3066 static const struct soc_device_attribute sysc_soc_feat_match[] = {
3067 	/* OMAP3430/3530 and AM3517 variants with some accelerators disabled */
3068 	SOC_FLAG("AM3505", DIS_SGX),
3069 	SOC_FLAG("OMAP3525", DIS_SGX),
3070 	SOC_FLAG("OMAP3515", DIS_IVA | DIS_SGX),
3071 	SOC_FLAG("OMAP3503", DIS_ISP | DIS_IVA | DIS_SGX),
3072 
3073 	/* OMAP3630/DM3730 variants with some accelerators disabled */
3074 	SOC_FLAG("AM3703", DIS_IVA | DIS_SGX),
3075 	SOC_FLAG("DM3725", DIS_SGX),
3076 	SOC_FLAG("OMAP3611", DIS_ISP | DIS_IVA | DIS_SGX),
3077 	SOC_FLAG("OMAP3615/AM3715", DIS_IVA),
3078 	SOC_FLAG("OMAP3621", DIS_ISP),
3079 
3080 	{ /* sentinel */ }
3081 };
3082 
sysc_add_disabled(unsigned long base)3083 static int sysc_add_disabled(unsigned long base)
3084 {
3085 	struct sysc_address *disabled_module;
3086 
3087 	disabled_module = kzalloc(sizeof(*disabled_module), GFP_KERNEL);
3088 	if (!disabled_module)
3089 		return -ENOMEM;
3090 
3091 	disabled_module->base = base;
3092 
3093 	mutex_lock(&sysc_soc->list_lock);
3094 	list_add(&disabled_module->node, &sysc_soc->disabled_modules);
3095 	mutex_unlock(&sysc_soc->list_lock);
3096 
3097 	return 0;
3098 }
3099 
3100 /*
3101  * One time init to detect the booted SoC, disable unavailable features
3102  * and initialize list for optional cpu_pm notifier.
3103  *
3104  * Note that we initialize static data shared across all ti-sysc instances
3105  * so ddata is only used for SoC type. This can be called from module_init
3106  * once we no longer need to rely on platform data.
3107  */
sysc_init_static_data(struct sysc * ddata)3108 static int sysc_init_static_data(struct sysc *ddata)
3109 {
3110 	const struct soc_device_attribute *match;
3111 	struct ti_sysc_platform_data *pdata;
3112 	unsigned long features = 0;
3113 	struct device_node *np;
3114 
3115 	if (sysc_soc)
3116 		return 0;
3117 
3118 	sysc_soc = kzalloc(sizeof(*sysc_soc), GFP_KERNEL);
3119 	if (!sysc_soc)
3120 		return -ENOMEM;
3121 
3122 	mutex_init(&sysc_soc->list_lock);
3123 	INIT_LIST_HEAD(&sysc_soc->disabled_modules);
3124 	INIT_LIST_HEAD(&sysc_soc->restored_modules);
3125 	sysc_soc->general_purpose = true;
3126 
3127 	pdata = dev_get_platdata(ddata->dev);
3128 	if (pdata && pdata->soc_type_gp)
3129 		sysc_soc->general_purpose = pdata->soc_type_gp();
3130 
3131 	match = soc_device_match(sysc_soc_match);
3132 	if (match && match->data)
3133 		sysc_soc->soc = (enum sysc_soc)(uintptr_t)match->data;
3134 
3135 	/*
3136 	 * Check and warn about possible old incomplete dtb. We now want to see
3137 	 * simple-pm-bus instead of simple-bus in the dtb for genpd using SoCs.
3138 	 */
3139 	switch (sysc_soc->soc) {
3140 	case SOC_AM3:
3141 	case SOC_AM4:
3142 	case SOC_4430 ... SOC_4470:
3143 	case SOC_5430:
3144 	case SOC_DRA7:
3145 		np = of_find_node_by_path("/ocp");
3146 		WARN_ONCE(np && of_device_is_compatible(np, "simple-bus"),
3147 			  "ti-sysc: Incomplete old dtb, please update\n");
3148 		break;
3149 	default:
3150 		break;
3151 	}
3152 
3153 	/* Ignore devices that are not available on HS and EMU SoCs */
3154 	if (!sysc_soc->general_purpose) {
3155 		switch (sysc_soc->soc) {
3156 		case SOC_3430 ... SOC_3630:
3157 			sysc_add_disabled(0x48304000);	/* timer12 */
3158 			break;
3159 		case SOC_AM3:
3160 			sysc_add_disabled(0x48310000);  /* rng */
3161 			break;
3162 		default:
3163 			break;
3164 		}
3165 	}
3166 
3167 	match = soc_device_match(sysc_soc_feat_match);
3168 	if (!match)
3169 		return 0;
3170 
3171 	if (match->data)
3172 		features = (unsigned long)match->data;
3173 
3174 	/*
3175 	 * Add disabled devices to the list based on the module base.
3176 	 * Note that this must be done before we attempt to access the
3177 	 * device and have module revision checks working.
3178 	 */
3179 	if (features & DIS_ISP)
3180 		sysc_add_disabled(0x480bd400);
3181 	if (features & DIS_IVA)
3182 		sysc_add_disabled(0x5d000000);
3183 	if (features & DIS_SGX)
3184 		sysc_add_disabled(0x50000000);
3185 
3186 	return 0;
3187 }
3188 
sysc_cleanup_static_data(void)3189 static void sysc_cleanup_static_data(void)
3190 {
3191 	struct sysc_module *restored_module;
3192 	struct sysc_address *disabled_module;
3193 	struct list_head *pos, *tmp;
3194 
3195 	if (!sysc_soc)
3196 		return;
3197 
3198 	if (sysc_soc->nb.notifier_call)
3199 		cpu_pm_unregister_notifier(&sysc_soc->nb);
3200 
3201 	mutex_lock(&sysc_soc->list_lock);
3202 	list_for_each_safe(pos, tmp, &sysc_soc->restored_modules) {
3203 		restored_module = list_entry(pos, struct sysc_module, node);
3204 		list_del(pos);
3205 		kfree(restored_module);
3206 	}
3207 	list_for_each_safe(pos, tmp, &sysc_soc->disabled_modules) {
3208 		disabled_module = list_entry(pos, struct sysc_address, node);
3209 		list_del(pos);
3210 		kfree(disabled_module);
3211 	}
3212 	mutex_unlock(&sysc_soc->list_lock);
3213 }
3214 
sysc_check_disabled_devices(struct sysc * ddata)3215 static int sysc_check_disabled_devices(struct sysc *ddata)
3216 {
3217 	struct sysc_address *disabled_module;
3218 	int error = 0;
3219 
3220 	mutex_lock(&sysc_soc->list_lock);
3221 	list_for_each_entry(disabled_module, &sysc_soc->disabled_modules, node) {
3222 		if (ddata->module_pa == disabled_module->base) {
3223 			dev_dbg(ddata->dev, "module disabled for this SoC\n");
3224 			error = -ENODEV;
3225 			break;
3226 		}
3227 	}
3228 	mutex_unlock(&sysc_soc->list_lock);
3229 
3230 	return error;
3231 }
3232 
3233 /*
3234  * Ignore timers tagged with no-reset and no-idle. These are likely in use,
3235  * for example by drivers/clocksource/timer-ti-dm-systimer.c. If more checks
3236  * are needed, we could also look at the timer register configuration.
3237  */
sysc_check_active_timer(struct sysc * ddata)3238 static int sysc_check_active_timer(struct sysc *ddata)
3239 {
3240 	int error;
3241 
3242 	if (ddata->cap->type != TI_SYSC_OMAP2_TIMER &&
3243 	    ddata->cap->type != TI_SYSC_OMAP4_TIMER)
3244 		return 0;
3245 
3246 	/*
3247 	 * Quirk for omap3 beagleboard revision A to B4 to use gpt12.
3248 	 * Revision C and later are fixed with commit 23885389dbbb ("ARM:
3249 	 * dts: Fix timer regression for beagleboard revision c"). This all
3250 	 * can be dropped if we stop supporting old beagleboard revisions
3251 	 * A to B4 at some point.
3252 	 */
3253 	if (sysc_soc->soc == SOC_3430 || sysc_soc->soc == SOC_AM35)
3254 		error = -ENXIO;
3255 	else
3256 		error = -EBUSY;
3257 
3258 	if ((ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT) &&
3259 	    (ddata->cfg.quirks & SYSC_QUIRK_NO_IDLE))
3260 		return error;
3261 
3262 	return 0;
3263 }
3264 
3265 static const struct of_device_id sysc_match_table[] = {
3266 	{ .compatible = "simple-bus", },
3267 	{ /* sentinel */ },
3268 };
3269 
sysc_probe(struct platform_device * pdev)3270 static int sysc_probe(struct platform_device *pdev)
3271 {
3272 	struct ti_sysc_platform_data *pdata = dev_get_platdata(&pdev->dev);
3273 	struct sysc *ddata;
3274 	int error;
3275 
3276 	ddata = devm_kzalloc(&pdev->dev, sizeof(*ddata), GFP_KERNEL);
3277 	if (!ddata)
3278 		return -ENOMEM;
3279 
3280 	ddata->offsets[SYSC_REVISION] = -ENODEV;
3281 	ddata->offsets[SYSC_SYSCONFIG] = -ENODEV;
3282 	ddata->offsets[SYSC_SYSSTATUS] = -ENODEV;
3283 	ddata->dev = &pdev->dev;
3284 	platform_set_drvdata(pdev, ddata);
3285 
3286 	error = sysc_init_static_data(ddata);
3287 	if (error)
3288 		return error;
3289 
3290 	error = sysc_init_match(ddata);
3291 	if (error)
3292 		return error;
3293 
3294 	error = sysc_init_dts_quirks(ddata);
3295 	if (error)
3296 		return error;
3297 
3298 	error = sysc_map_and_check_registers(ddata);
3299 	if (error)
3300 		return error;
3301 
3302 	error = sysc_init_sysc_mask(ddata);
3303 	if (error)
3304 		return error;
3305 
3306 	error = sysc_init_idlemodes(ddata);
3307 	if (error)
3308 		return error;
3309 
3310 	error = sysc_init_syss_mask(ddata);
3311 	if (error)
3312 		return error;
3313 
3314 	error = sysc_init_pdata(ddata);
3315 	if (error)
3316 		return error;
3317 
3318 	sysc_init_early_quirks(ddata);
3319 
3320 	error = sysc_check_disabled_devices(ddata);
3321 	if (error)
3322 		return error;
3323 
3324 	error = sysc_check_active_timer(ddata);
3325 	if (error == -ENXIO)
3326 		ddata->reserved = true;
3327 	else if (error)
3328 		return error;
3329 
3330 	error = sysc_get_clocks(ddata);
3331 	if (error)
3332 		return error;
3333 
3334 	error = sysc_init_resets(ddata);
3335 	if (error)
3336 		goto unprepare;
3337 
3338 	error = sysc_init_module(ddata);
3339 	if (error)
3340 		goto unprepare;
3341 
3342 	pm_runtime_enable(ddata->dev);
3343 	error = pm_runtime_resume_and_get(ddata->dev);
3344 	if (error < 0) {
3345 		pm_runtime_disable(ddata->dev);
3346 		goto unprepare;
3347 	}
3348 
3349 	/* Balance use counts as PM runtime should have enabled these all */
3350 	if (!(ddata->cfg.quirks &
3351 	      (SYSC_QUIRK_NO_IDLE | SYSC_QUIRK_NO_IDLE_ON_INIT))) {
3352 		sysc_disable_main_clocks(ddata);
3353 		sysc_disable_opt_clocks(ddata);
3354 		sysc_clkdm_allow_idle(ddata);
3355 	}
3356 
3357 	if (!(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT))
3358 		reset_control_assert(ddata->rsts);
3359 
3360 	sysc_show_registers(ddata);
3361 
3362 	ddata->dev->type = &sysc_device_type;
3363 
3364 	if (!ddata->reserved) {
3365 		error = of_platform_populate(ddata->dev->of_node,
3366 					     sysc_match_table,
3367 					     pdata ? pdata->auxdata : NULL,
3368 					     ddata->dev);
3369 		if (error)
3370 			goto err;
3371 	}
3372 
3373 	INIT_DELAYED_WORK(&ddata->idle_work, ti_sysc_idle);
3374 
3375 	/* At least earlycon won't survive without deferred idle */
3376 	if (ddata->cfg.quirks & (SYSC_QUIRK_NO_IDLE |
3377 				 SYSC_QUIRK_NO_IDLE_ON_INIT |
3378 				 SYSC_QUIRK_NO_RESET_ON_INIT)) {
3379 		schedule_delayed_work(&ddata->idle_work, 3000);
3380 	} else {
3381 		pm_runtime_put(&pdev->dev);
3382 	}
3383 
3384 	if (ddata->cfg.quirks & SYSC_QUIRK_REINIT_ON_CTX_LOST)
3385 		sysc_add_restored(ddata);
3386 
3387 	return 0;
3388 
3389 err:
3390 	pm_runtime_put_sync(&pdev->dev);
3391 	pm_runtime_disable(&pdev->dev);
3392 unprepare:
3393 	sysc_unprepare(ddata);
3394 
3395 	return error;
3396 }
3397 
sysc_remove(struct platform_device * pdev)3398 static int sysc_remove(struct platform_device *pdev)
3399 {
3400 	struct sysc *ddata = platform_get_drvdata(pdev);
3401 	int error;
3402 
3403 	/* Device can still be enabled, see deferred idle quirk in probe */
3404 	if (cancel_delayed_work_sync(&ddata->idle_work))
3405 		ti_sysc_idle(&ddata->idle_work.work);
3406 
3407 	error = pm_runtime_resume_and_get(ddata->dev);
3408 	if (error < 0) {
3409 		pm_runtime_disable(ddata->dev);
3410 		goto unprepare;
3411 	}
3412 
3413 	of_platform_depopulate(&pdev->dev);
3414 
3415 	pm_runtime_put_sync(&pdev->dev);
3416 	pm_runtime_disable(&pdev->dev);
3417 
3418 	if (!reset_control_status(ddata->rsts))
3419 		reset_control_assert(ddata->rsts);
3420 
3421 unprepare:
3422 	sysc_unprepare(ddata);
3423 
3424 	return 0;
3425 }
3426 
3427 static const struct of_device_id sysc_match[] = {
3428 	{ .compatible = "ti,sysc-omap2", .data = &sysc_omap2, },
3429 	{ .compatible = "ti,sysc-omap2-timer", .data = &sysc_omap2_timer, },
3430 	{ .compatible = "ti,sysc-omap4", .data = &sysc_omap4, },
3431 	{ .compatible = "ti,sysc-omap4-timer", .data = &sysc_omap4_timer, },
3432 	{ .compatible = "ti,sysc-omap4-simple", .data = &sysc_omap4_simple, },
3433 	{ .compatible = "ti,sysc-omap3430-sr", .data = &sysc_34xx_sr, },
3434 	{ .compatible = "ti,sysc-omap3630-sr", .data = &sysc_36xx_sr, },
3435 	{ .compatible = "ti,sysc-omap4-sr", .data = &sysc_omap4_sr, },
3436 	{ .compatible = "ti,sysc-omap3-sham", .data = &sysc_omap3_sham, },
3437 	{ .compatible = "ti,sysc-omap-aes", .data = &sysc_omap3_aes, },
3438 	{ .compatible = "ti,sysc-mcasp", .data = &sysc_omap4_mcasp, },
3439 	{ .compatible = "ti,sysc-dra7-mcasp", .data = &sysc_dra7_mcasp, },
3440 	{ .compatible = "ti,sysc-usb-host-fs",
3441 	  .data = &sysc_omap4_usb_host_fs, },
3442 	{ .compatible = "ti,sysc-dra7-mcan", .data = &sysc_dra7_mcan, },
3443 	{ .compatible = "ti,sysc-pruss", .data = &sysc_pruss, },
3444 	{  },
3445 };
3446 MODULE_DEVICE_TABLE(of, sysc_match);
3447 
3448 static struct platform_driver sysc_driver = {
3449 	.probe		= sysc_probe,
3450 	.remove		= sysc_remove,
3451 	.driver         = {
3452 		.name   = "ti-sysc",
3453 		.of_match_table	= sysc_match,
3454 		.pm = &sysc_pm_ops,
3455 	},
3456 };
3457 
sysc_init(void)3458 static int __init sysc_init(void)
3459 {
3460 	bus_register_notifier(&platform_bus_type, &sysc_nb);
3461 
3462 	return platform_driver_register(&sysc_driver);
3463 }
3464 module_init(sysc_init);
3465 
sysc_exit(void)3466 static void __exit sysc_exit(void)
3467 {
3468 	bus_unregister_notifier(&platform_bus_type, &sysc_nb);
3469 	platform_driver_unregister(&sysc_driver);
3470 	sysc_cleanup_static_data();
3471 }
3472 module_exit(sysc_exit);
3473 
3474 MODULE_DESCRIPTION("TI sysc interconnect target driver");
3475 MODULE_LICENSE("GPL v2");
3476