1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * ti-sysc.c - Texas Instruments sysc interconnect target driver
4 */
5
6 #include <linux/io.h>
7 #include <linux/clk.h>
8 #include <linux/clkdev.h>
9 #include <linux/cpu_pm.h>
10 #include <linux/delay.h>
11 #include <linux/list.h>
12 #include <linux/module.h>
13 #include <linux/platform_device.h>
14 #include <linux/pm_domain.h>
15 #include <linux/pm_runtime.h>
16 #include <linux/reset.h>
17 #include <linux/of_address.h>
18 #include <linux/of_platform.h>
19 #include <linux/slab.h>
20 #include <linux/sys_soc.h>
21 #include <linux/timekeeping.h>
22 #include <linux/iopoll.h>
23
24 #include <linux/platform_data/ti-sysc.h>
25
26 #include <dt-bindings/bus/ti-sysc.h>
27
28 #define DIS_ISP BIT(2)
29 #define DIS_IVA BIT(1)
30 #define DIS_SGX BIT(0)
31
32 #define SOC_FLAG(match, flag) { .machine = match, .data = (void *)(flag), }
33
34 #define MAX_MODULE_SOFTRESET_WAIT 10000
35
36 enum sysc_soc {
37 SOC_UNKNOWN,
38 SOC_2420,
39 SOC_2430,
40 SOC_3430,
41 SOC_AM35,
42 SOC_3630,
43 SOC_4430,
44 SOC_4460,
45 SOC_4470,
46 SOC_5430,
47 SOC_AM3,
48 SOC_AM4,
49 SOC_DRA7,
50 };
51
52 struct sysc_address {
53 unsigned long base;
54 struct list_head node;
55 };
56
57 struct sysc_module {
58 struct sysc *ddata;
59 struct list_head node;
60 };
61
62 struct sysc_soc_info {
63 unsigned long general_purpose:1;
64 enum sysc_soc soc;
65 struct mutex list_lock; /* disabled and restored modules list lock */
66 struct list_head disabled_modules;
67 struct list_head restored_modules;
68 struct notifier_block nb;
69 };
70
71 enum sysc_clocks {
72 SYSC_FCK,
73 SYSC_ICK,
74 SYSC_OPTFCK0,
75 SYSC_OPTFCK1,
76 SYSC_OPTFCK2,
77 SYSC_OPTFCK3,
78 SYSC_OPTFCK4,
79 SYSC_OPTFCK5,
80 SYSC_OPTFCK6,
81 SYSC_OPTFCK7,
82 SYSC_MAX_CLOCKS,
83 };
84
85 static struct sysc_soc_info *sysc_soc;
86 static const char * const reg_names[] = { "rev", "sysc", "syss", };
87 static const char * const clock_names[SYSC_MAX_CLOCKS] = {
88 "fck", "ick", "opt0", "opt1", "opt2", "opt3", "opt4",
89 "opt5", "opt6", "opt7",
90 };
91
92 #define SYSC_IDLEMODE_MASK 3
93 #define SYSC_CLOCKACTIVITY_MASK 3
94
95 /**
96 * struct sysc - TI sysc interconnect target module registers and capabilities
97 * @dev: struct device pointer
98 * @module_pa: physical address of the interconnect target module
99 * @module_size: size of the interconnect target module
100 * @module_va: virtual address of the interconnect target module
101 * @offsets: register offsets from module base
102 * @mdata: ti-sysc to hwmod translation data for a module
103 * @clocks: clocks used by the interconnect target module
104 * @clock_roles: clock role names for the found clocks
105 * @nr_clocks: number of clocks used by the interconnect target module
106 * @rsts: resets used by the interconnect target module
107 * @legacy_mode: configured for legacy mode if set
108 * @cap: interconnect target module capabilities
109 * @cfg: interconnect target module configuration
110 * @cookie: data used by legacy platform callbacks
111 * @name: name if available
112 * @revision: interconnect target module revision
113 * @sysconfig: saved sysconfig register value
114 * @reserved: target module is reserved and already in use
115 * @enabled: sysc runtime enabled status
116 * @needs_resume: runtime resume needed on resume from suspend
117 * @child_needs_resume: runtime resume needed for child on resume from suspend
118 * @disable_on_idle: status flag used for disabling modules with resets
119 * @idle_work: work structure used to perform delayed idle on a module
120 * @pre_reset_quirk: module specific pre-reset quirk
121 * @post_reset_quirk: module specific post-reset quirk
122 * @reset_done_quirk: module specific reset done quirk
123 * @module_enable_quirk: module specific enable quirk
124 * @module_disable_quirk: module specific disable quirk
125 * @module_unlock_quirk: module specific sysconfig unlock quirk
126 * @module_lock_quirk: module specific sysconfig lock quirk
127 */
128 struct sysc {
129 struct device *dev;
130 u64 module_pa;
131 u32 module_size;
132 void __iomem *module_va;
133 int offsets[SYSC_MAX_REGS];
134 struct ti_sysc_module_data *mdata;
135 struct clk **clocks;
136 const char **clock_roles;
137 int nr_clocks;
138 struct reset_control *rsts;
139 const char *legacy_mode;
140 const struct sysc_capabilities *cap;
141 struct sysc_config cfg;
142 struct ti_sysc_cookie cookie;
143 const char *name;
144 u32 revision;
145 u32 sysconfig;
146 unsigned int reserved:1;
147 unsigned int enabled:1;
148 unsigned int needs_resume:1;
149 unsigned int child_needs_resume:1;
150 struct delayed_work idle_work;
151 void (*pre_reset_quirk)(struct sysc *sysc);
152 void (*post_reset_quirk)(struct sysc *sysc);
153 void (*reset_done_quirk)(struct sysc *sysc);
154 void (*module_enable_quirk)(struct sysc *sysc);
155 void (*module_disable_quirk)(struct sysc *sysc);
156 void (*module_unlock_quirk)(struct sysc *sysc);
157 void (*module_lock_quirk)(struct sysc *sysc);
158 };
159
160 static void sysc_parse_dts_quirks(struct sysc *ddata, struct device_node *np,
161 bool is_child);
162 static int sysc_reset(struct sysc *ddata);
163
sysc_write(struct sysc * ddata,int offset,u32 value)164 static void sysc_write(struct sysc *ddata, int offset, u32 value)
165 {
166 if (ddata->cfg.quirks & SYSC_QUIRK_16BIT) {
167 writew_relaxed(value & 0xffff, ddata->module_va + offset);
168
169 /* Only i2c revision has LO and HI register with stride of 4 */
170 if (ddata->offsets[SYSC_REVISION] >= 0 &&
171 offset == ddata->offsets[SYSC_REVISION]) {
172 u16 hi = value >> 16;
173
174 writew_relaxed(hi, ddata->module_va + offset + 4);
175 }
176
177 return;
178 }
179
180 writel_relaxed(value, ddata->module_va + offset);
181 }
182
sysc_read(struct sysc * ddata,int offset)183 static u32 sysc_read(struct sysc *ddata, int offset)
184 {
185 if (ddata->cfg.quirks & SYSC_QUIRK_16BIT) {
186 u32 val;
187
188 val = readw_relaxed(ddata->module_va + offset);
189
190 /* Only i2c revision has LO and HI register with stride of 4 */
191 if (ddata->offsets[SYSC_REVISION] >= 0 &&
192 offset == ddata->offsets[SYSC_REVISION]) {
193 u16 tmp = readw_relaxed(ddata->module_va + offset + 4);
194
195 val |= tmp << 16;
196 }
197
198 return val;
199 }
200
201 return readl_relaxed(ddata->module_va + offset);
202 }
203
sysc_opt_clks_needed(struct sysc * ddata)204 static bool sysc_opt_clks_needed(struct sysc *ddata)
205 {
206 return !!(ddata->cfg.quirks & SYSC_QUIRK_OPT_CLKS_NEEDED);
207 }
208
sysc_read_revision(struct sysc * ddata)209 static u32 sysc_read_revision(struct sysc *ddata)
210 {
211 int offset = ddata->offsets[SYSC_REVISION];
212
213 if (offset < 0)
214 return 0;
215
216 return sysc_read(ddata, offset);
217 }
218
sysc_read_sysconfig(struct sysc * ddata)219 static u32 sysc_read_sysconfig(struct sysc *ddata)
220 {
221 int offset = ddata->offsets[SYSC_SYSCONFIG];
222
223 if (offset < 0)
224 return 0;
225
226 return sysc_read(ddata, offset);
227 }
228
sysc_read_sysstatus(struct sysc * ddata)229 static u32 sysc_read_sysstatus(struct sysc *ddata)
230 {
231 int offset = ddata->offsets[SYSC_SYSSTATUS];
232
233 if (offset < 0)
234 return 0;
235
236 return sysc_read(ddata, offset);
237 }
238
sysc_poll_reset_sysstatus(struct sysc * ddata)239 static int sysc_poll_reset_sysstatus(struct sysc *ddata)
240 {
241 int error, retries;
242 u32 syss_done, rstval;
243
244 if (ddata->cfg.quirks & SYSS_QUIRK_RESETDONE_INVERTED)
245 syss_done = 0;
246 else
247 syss_done = ddata->cfg.syss_mask;
248
249 if (likely(!timekeeping_suspended)) {
250 error = readx_poll_timeout_atomic(sysc_read_sysstatus, ddata,
251 rstval, (rstval & ddata->cfg.syss_mask) ==
252 syss_done, 100, MAX_MODULE_SOFTRESET_WAIT);
253 } else {
254 retries = MAX_MODULE_SOFTRESET_WAIT;
255 while (retries--) {
256 rstval = sysc_read_sysstatus(ddata);
257 if ((rstval & ddata->cfg.syss_mask) == syss_done)
258 return 0;
259 udelay(2); /* Account for udelay flakeyness */
260 }
261 error = -ETIMEDOUT;
262 }
263
264 return error;
265 }
266
sysc_poll_reset_sysconfig(struct sysc * ddata)267 static int sysc_poll_reset_sysconfig(struct sysc *ddata)
268 {
269 int error, retries;
270 u32 sysc_mask, rstval;
271
272 sysc_mask = BIT(ddata->cap->regbits->srst_shift);
273
274 if (likely(!timekeeping_suspended)) {
275 error = readx_poll_timeout_atomic(sysc_read_sysconfig, ddata,
276 rstval, !(rstval & sysc_mask),
277 100, MAX_MODULE_SOFTRESET_WAIT);
278 } else {
279 retries = MAX_MODULE_SOFTRESET_WAIT;
280 while (retries--) {
281 rstval = sysc_read_sysconfig(ddata);
282 if (!(rstval & sysc_mask))
283 return 0;
284 udelay(2); /* Account for udelay flakeyness */
285 }
286 error = -ETIMEDOUT;
287 }
288
289 return error;
290 }
291
292 /* Poll on reset status */
sysc_wait_softreset(struct sysc * ddata)293 static int sysc_wait_softreset(struct sysc *ddata)
294 {
295 int syss_offset, error = 0;
296
297 if (ddata->cap->regbits->srst_shift < 0)
298 return 0;
299
300 syss_offset = ddata->offsets[SYSC_SYSSTATUS];
301
302 if (syss_offset >= 0)
303 error = sysc_poll_reset_sysstatus(ddata);
304 else if (ddata->cfg.quirks & SYSC_QUIRK_RESET_STATUS)
305 error = sysc_poll_reset_sysconfig(ddata);
306
307 return error;
308 }
309
sysc_add_named_clock_from_child(struct sysc * ddata,const char * name,const char * optfck_name)310 static int sysc_add_named_clock_from_child(struct sysc *ddata,
311 const char *name,
312 const char *optfck_name)
313 {
314 struct device_node *np = ddata->dev->of_node;
315 struct device_node *child;
316 struct clk_lookup *cl;
317 struct clk *clock;
318 const char *n;
319
320 if (name)
321 n = name;
322 else
323 n = optfck_name;
324
325 /* Does the clock alias already exist? */
326 clock = of_clk_get_by_name(np, n);
327 if (!IS_ERR(clock)) {
328 clk_put(clock);
329
330 return 0;
331 }
332
333 child = of_get_next_available_child(np, NULL);
334 if (!child)
335 return -ENODEV;
336
337 clock = devm_get_clk_from_child(ddata->dev, child, name);
338 if (IS_ERR(clock))
339 return PTR_ERR(clock);
340
341 /*
342 * Use clkdev_add() instead of clkdev_alloc() to avoid the MAX_DEV_ID
343 * limit for clk_get(). If cl ever needs to be freed, it should be done
344 * with clkdev_drop().
345 */
346 cl = kzalloc(sizeof(*cl), GFP_KERNEL);
347 if (!cl)
348 return -ENOMEM;
349
350 cl->con_id = n;
351 cl->dev_id = dev_name(ddata->dev);
352 cl->clk = clock;
353 clkdev_add(cl);
354
355 clk_put(clock);
356
357 return 0;
358 }
359
sysc_init_ext_opt_clock(struct sysc * ddata,const char * name)360 static int sysc_init_ext_opt_clock(struct sysc *ddata, const char *name)
361 {
362 const char *optfck_name;
363 int error, index;
364
365 if (ddata->nr_clocks < SYSC_OPTFCK0)
366 index = SYSC_OPTFCK0;
367 else
368 index = ddata->nr_clocks;
369
370 if (name)
371 optfck_name = name;
372 else
373 optfck_name = clock_names[index];
374
375 error = sysc_add_named_clock_from_child(ddata, name, optfck_name);
376 if (error)
377 return error;
378
379 ddata->clock_roles[index] = optfck_name;
380 ddata->nr_clocks++;
381
382 return 0;
383 }
384
sysc_get_one_clock(struct sysc * ddata,const char * name)385 static int sysc_get_one_clock(struct sysc *ddata, const char *name)
386 {
387 int error, i, index = -ENODEV;
388
389 if (!strncmp(clock_names[SYSC_FCK], name, 3))
390 index = SYSC_FCK;
391 else if (!strncmp(clock_names[SYSC_ICK], name, 3))
392 index = SYSC_ICK;
393
394 if (index < 0) {
395 for (i = SYSC_OPTFCK0; i < SYSC_MAX_CLOCKS; i++) {
396 if (!ddata->clocks[i]) {
397 index = i;
398 break;
399 }
400 }
401 }
402
403 if (index < 0) {
404 dev_err(ddata->dev, "clock %s not added\n", name);
405 return index;
406 }
407
408 ddata->clocks[index] = devm_clk_get(ddata->dev, name);
409 if (IS_ERR(ddata->clocks[index])) {
410 dev_err(ddata->dev, "clock get error for %s: %li\n",
411 name, PTR_ERR(ddata->clocks[index]));
412
413 return PTR_ERR(ddata->clocks[index]);
414 }
415
416 error = clk_prepare(ddata->clocks[index]);
417 if (error) {
418 dev_err(ddata->dev, "clock prepare error for %s: %i\n",
419 name, error);
420
421 return error;
422 }
423
424 return 0;
425 }
426
sysc_get_clocks(struct sysc * ddata)427 static int sysc_get_clocks(struct sysc *ddata)
428 {
429 struct device_node *np = ddata->dev->of_node;
430 struct property *prop;
431 const char *name;
432 int nr_fck = 0, nr_ick = 0, i, error = 0;
433
434 ddata->clock_roles = devm_kcalloc(ddata->dev,
435 SYSC_MAX_CLOCKS,
436 sizeof(*ddata->clock_roles),
437 GFP_KERNEL);
438 if (!ddata->clock_roles)
439 return -ENOMEM;
440
441 of_property_for_each_string(np, "clock-names", prop, name) {
442 if (!strncmp(clock_names[SYSC_FCK], name, 3))
443 nr_fck++;
444 if (!strncmp(clock_names[SYSC_ICK], name, 3))
445 nr_ick++;
446 ddata->clock_roles[ddata->nr_clocks] = name;
447 ddata->nr_clocks++;
448 }
449
450 if (ddata->nr_clocks < 1)
451 return 0;
452
453 if ((ddata->cfg.quirks & SYSC_QUIRK_EXT_OPT_CLOCK)) {
454 error = sysc_init_ext_opt_clock(ddata, NULL);
455 if (error)
456 return error;
457 }
458
459 if (ddata->nr_clocks > SYSC_MAX_CLOCKS) {
460 dev_err(ddata->dev, "too many clocks for %pOF\n", np);
461
462 return -EINVAL;
463 }
464
465 if (nr_fck > 1 || nr_ick > 1) {
466 dev_err(ddata->dev, "max one fck and ick for %pOF\n", np);
467
468 return -EINVAL;
469 }
470
471 /* Always add a slot for main clocks fck and ick even if unused */
472 if (!nr_fck)
473 ddata->nr_clocks++;
474 if (!nr_ick)
475 ddata->nr_clocks++;
476
477 ddata->clocks = devm_kcalloc(ddata->dev,
478 ddata->nr_clocks, sizeof(*ddata->clocks),
479 GFP_KERNEL);
480 if (!ddata->clocks)
481 return -ENOMEM;
482
483 for (i = 0; i < SYSC_MAX_CLOCKS; i++) {
484 const char *name = ddata->clock_roles[i];
485
486 if (!name)
487 continue;
488
489 error = sysc_get_one_clock(ddata, name);
490 if (error)
491 return error;
492 }
493
494 return 0;
495 }
496
sysc_enable_main_clocks(struct sysc * ddata)497 static int sysc_enable_main_clocks(struct sysc *ddata)
498 {
499 struct clk *clock;
500 int i, error;
501
502 if (!ddata->clocks)
503 return 0;
504
505 for (i = 0; i < SYSC_OPTFCK0; i++) {
506 clock = ddata->clocks[i];
507
508 /* Main clocks may not have ick */
509 if (IS_ERR_OR_NULL(clock))
510 continue;
511
512 error = clk_enable(clock);
513 if (error)
514 goto err_disable;
515 }
516
517 return 0;
518
519 err_disable:
520 for (i--; i >= 0; i--) {
521 clock = ddata->clocks[i];
522
523 /* Main clocks may not have ick */
524 if (IS_ERR_OR_NULL(clock))
525 continue;
526
527 clk_disable(clock);
528 }
529
530 return error;
531 }
532
sysc_disable_main_clocks(struct sysc * ddata)533 static void sysc_disable_main_clocks(struct sysc *ddata)
534 {
535 struct clk *clock;
536 int i;
537
538 if (!ddata->clocks)
539 return;
540
541 for (i = 0; i < SYSC_OPTFCK0; i++) {
542 clock = ddata->clocks[i];
543 if (IS_ERR_OR_NULL(clock))
544 continue;
545
546 clk_disable(clock);
547 }
548 }
549
sysc_enable_opt_clocks(struct sysc * ddata)550 static int sysc_enable_opt_clocks(struct sysc *ddata)
551 {
552 struct clk *clock;
553 int i, error;
554
555 if (!ddata->clocks || ddata->nr_clocks < SYSC_OPTFCK0 + 1)
556 return 0;
557
558 for (i = SYSC_OPTFCK0; i < SYSC_MAX_CLOCKS; i++) {
559 clock = ddata->clocks[i];
560
561 /* Assume no holes for opt clocks */
562 if (IS_ERR_OR_NULL(clock))
563 return 0;
564
565 error = clk_enable(clock);
566 if (error)
567 goto err_disable;
568 }
569
570 return 0;
571
572 err_disable:
573 for (i--; i >= 0; i--) {
574 clock = ddata->clocks[i];
575 if (IS_ERR_OR_NULL(clock))
576 continue;
577
578 clk_disable(clock);
579 }
580
581 return error;
582 }
583
sysc_disable_opt_clocks(struct sysc * ddata)584 static void sysc_disable_opt_clocks(struct sysc *ddata)
585 {
586 struct clk *clock;
587 int i;
588
589 if (!ddata->clocks || ddata->nr_clocks < SYSC_OPTFCK0 + 1)
590 return;
591
592 for (i = SYSC_OPTFCK0; i < SYSC_MAX_CLOCKS; i++) {
593 clock = ddata->clocks[i];
594
595 /* Assume no holes for opt clocks */
596 if (IS_ERR_OR_NULL(clock))
597 return;
598
599 clk_disable(clock);
600 }
601 }
602
sysc_clkdm_deny_idle(struct sysc * ddata)603 static void sysc_clkdm_deny_idle(struct sysc *ddata)
604 {
605 struct ti_sysc_platform_data *pdata;
606
607 if (ddata->legacy_mode || (ddata->cfg.quirks & SYSC_QUIRK_CLKDM_NOAUTO))
608 return;
609
610 pdata = dev_get_platdata(ddata->dev);
611 if (pdata && pdata->clkdm_deny_idle)
612 pdata->clkdm_deny_idle(ddata->dev, &ddata->cookie);
613 }
614
sysc_clkdm_allow_idle(struct sysc * ddata)615 static void sysc_clkdm_allow_idle(struct sysc *ddata)
616 {
617 struct ti_sysc_platform_data *pdata;
618
619 if (ddata->legacy_mode || (ddata->cfg.quirks & SYSC_QUIRK_CLKDM_NOAUTO))
620 return;
621
622 pdata = dev_get_platdata(ddata->dev);
623 if (pdata && pdata->clkdm_allow_idle)
624 pdata->clkdm_allow_idle(ddata->dev, &ddata->cookie);
625 }
626
627 /**
628 * sysc_init_resets - init rstctrl reset line if configured
629 * @ddata: device driver data
630 *
631 * See sysc_rstctrl_reset_deassert().
632 */
sysc_init_resets(struct sysc * ddata)633 static int sysc_init_resets(struct sysc *ddata)
634 {
635 ddata->rsts =
636 devm_reset_control_get_optional_shared(ddata->dev, "rstctrl");
637
638 return PTR_ERR_OR_ZERO(ddata->rsts);
639 }
640
641 /**
642 * sysc_parse_and_check_child_range - parses module IO region from ranges
643 * @ddata: device driver data
644 *
645 * In general we only need rev, syss, and sysc registers and not the whole
646 * module range. But we do want the offsets for these registers from the
647 * module base. This allows us to check them against the legacy hwmod
648 * platform data. Let's also check the ranges are configured properly.
649 */
sysc_parse_and_check_child_range(struct sysc * ddata)650 static int sysc_parse_and_check_child_range(struct sysc *ddata)
651 {
652 struct device_node *np = ddata->dev->of_node;
653 struct of_range_parser parser;
654 struct of_range range;
655 int error;
656
657 error = of_range_parser_init(&parser, np);
658 if (error)
659 return error;
660
661 for_each_of_range(&parser, &range) {
662 ddata->module_pa = range.cpu_addr;
663 ddata->module_size = range.size;
664 break;
665 }
666
667 return 0;
668 }
669
670 static struct device_node *stdout_path;
671
sysc_init_stdout_path(struct sysc * ddata)672 static void sysc_init_stdout_path(struct sysc *ddata)
673 {
674 struct device_node *np = NULL;
675 const char *uart;
676
677 if (IS_ERR(stdout_path))
678 return;
679
680 if (stdout_path)
681 return;
682
683 np = of_find_node_by_path("/chosen");
684 if (!np)
685 goto err;
686
687 uart = of_get_property(np, "stdout-path", NULL);
688 if (!uart)
689 goto err;
690
691 np = of_find_node_by_path(uart);
692 if (!np)
693 goto err;
694
695 stdout_path = np;
696
697 return;
698
699 err:
700 stdout_path = ERR_PTR(-ENODEV);
701 }
702
sysc_check_quirk_stdout(struct sysc * ddata,struct device_node * np)703 static void sysc_check_quirk_stdout(struct sysc *ddata,
704 struct device_node *np)
705 {
706 sysc_init_stdout_path(ddata);
707 if (np != stdout_path)
708 return;
709
710 ddata->cfg.quirks |= SYSC_QUIRK_NO_IDLE_ON_INIT |
711 SYSC_QUIRK_NO_RESET_ON_INIT;
712 }
713
714 /**
715 * sysc_check_one_child - check child configuration
716 * @ddata: device driver data
717 * @np: child device node
718 *
719 * Let's avoid messy situations where we have new interconnect target
720 * node but children have "ti,hwmods". These belong to the interconnect
721 * target node and are managed by this driver.
722 */
sysc_check_one_child(struct sysc * ddata,struct device_node * np)723 static void sysc_check_one_child(struct sysc *ddata,
724 struct device_node *np)
725 {
726 const char *name;
727
728 name = of_get_property(np, "ti,hwmods", NULL);
729 if (name && !of_device_is_compatible(np, "ti,sysc"))
730 dev_warn(ddata->dev, "really a child ti,hwmods property?");
731
732 sysc_check_quirk_stdout(ddata, np);
733 sysc_parse_dts_quirks(ddata, np, true);
734 }
735
sysc_check_children(struct sysc * ddata)736 static void sysc_check_children(struct sysc *ddata)
737 {
738 struct device_node *child;
739
740 for_each_child_of_node(ddata->dev->of_node, child)
741 sysc_check_one_child(ddata, child);
742 }
743
744 /*
745 * So far only I2C uses 16-bit read access with clockactivity with revision
746 * in two registers with stride of 4. We can detect this based on the rev
747 * register size to configure things far enough to be able to properly read
748 * the revision register.
749 */
sysc_check_quirk_16bit(struct sysc * ddata,struct resource * res)750 static void sysc_check_quirk_16bit(struct sysc *ddata, struct resource *res)
751 {
752 if (resource_size(res) == 8)
753 ddata->cfg.quirks |= SYSC_QUIRK_16BIT | SYSC_QUIRK_USE_CLOCKACT;
754 }
755
756 /**
757 * sysc_parse_one - parses the interconnect target module registers
758 * @ddata: device driver data
759 * @reg: register to parse
760 */
sysc_parse_one(struct sysc * ddata,enum sysc_registers reg)761 static int sysc_parse_one(struct sysc *ddata, enum sysc_registers reg)
762 {
763 struct resource *res;
764 const char *name;
765
766 switch (reg) {
767 case SYSC_REVISION:
768 case SYSC_SYSCONFIG:
769 case SYSC_SYSSTATUS:
770 name = reg_names[reg];
771 break;
772 default:
773 return -EINVAL;
774 }
775
776 res = platform_get_resource_byname(to_platform_device(ddata->dev),
777 IORESOURCE_MEM, name);
778 if (!res) {
779 ddata->offsets[reg] = -ENODEV;
780
781 return 0;
782 }
783
784 ddata->offsets[reg] = res->start - ddata->module_pa;
785 if (reg == SYSC_REVISION)
786 sysc_check_quirk_16bit(ddata, res);
787
788 return 0;
789 }
790
sysc_parse_registers(struct sysc * ddata)791 static int sysc_parse_registers(struct sysc *ddata)
792 {
793 int i, error;
794
795 for (i = 0; i < SYSC_MAX_REGS; i++) {
796 error = sysc_parse_one(ddata, i);
797 if (error)
798 return error;
799 }
800
801 return 0;
802 }
803
804 /**
805 * sysc_check_registers - check for misconfigured register overlaps
806 * @ddata: device driver data
807 */
sysc_check_registers(struct sysc * ddata)808 static int sysc_check_registers(struct sysc *ddata)
809 {
810 int i, j, nr_regs = 0, nr_matches = 0;
811
812 for (i = 0; i < SYSC_MAX_REGS; i++) {
813 if (ddata->offsets[i] < 0)
814 continue;
815
816 if (ddata->offsets[i] > (ddata->module_size - 4)) {
817 dev_err(ddata->dev, "register outside module range");
818
819 return -EINVAL;
820 }
821
822 for (j = 0; j < SYSC_MAX_REGS; j++) {
823 if (ddata->offsets[j] < 0)
824 continue;
825
826 if (ddata->offsets[i] == ddata->offsets[j])
827 nr_matches++;
828 }
829 nr_regs++;
830 }
831
832 if (nr_matches > nr_regs) {
833 dev_err(ddata->dev, "overlapping registers: (%i/%i)",
834 nr_regs, nr_matches);
835
836 return -EINVAL;
837 }
838
839 return 0;
840 }
841
842 /**
843 * sysc_ioremap - ioremap register space for the interconnect target module
844 * @ddata: device driver data
845 *
846 * Note that the interconnect target module registers can be anywhere
847 * within the interconnect target module range. For example, SGX has
848 * them at offset 0x1fc00 in the 32MB module address space. And cpsw
849 * has them at offset 0x1200 in the CPSW_WR child. Usually the
850 * interconnect target module registers are at the beginning of
851 * the module range though.
852 */
sysc_ioremap(struct sysc * ddata)853 static int sysc_ioremap(struct sysc *ddata)
854 {
855 int size;
856
857 if (ddata->offsets[SYSC_REVISION] < 0 &&
858 ddata->offsets[SYSC_SYSCONFIG] < 0 &&
859 ddata->offsets[SYSC_SYSSTATUS] < 0) {
860 size = ddata->module_size;
861 } else {
862 size = max3(ddata->offsets[SYSC_REVISION],
863 ddata->offsets[SYSC_SYSCONFIG],
864 ddata->offsets[SYSC_SYSSTATUS]);
865
866 if (size < SZ_1K)
867 size = SZ_1K;
868
869 if ((size + sizeof(u32)) > ddata->module_size)
870 size = ddata->module_size;
871 }
872
873 ddata->module_va = devm_ioremap(ddata->dev,
874 ddata->module_pa,
875 size + sizeof(u32));
876 if (!ddata->module_va)
877 return -EIO;
878
879 return 0;
880 }
881
882 /**
883 * sysc_map_and_check_registers - ioremap and check device registers
884 * @ddata: device driver data
885 */
sysc_map_and_check_registers(struct sysc * ddata)886 static int sysc_map_and_check_registers(struct sysc *ddata)
887 {
888 struct device_node *np = ddata->dev->of_node;
889 int error;
890
891 error = sysc_parse_and_check_child_range(ddata);
892 if (error)
893 return error;
894
895 sysc_check_children(ddata);
896
897 if (!of_property_present(np, "reg"))
898 return 0;
899
900 error = sysc_parse_registers(ddata);
901 if (error)
902 return error;
903
904 error = sysc_ioremap(ddata);
905 if (error)
906 return error;
907
908 error = sysc_check_registers(ddata);
909 if (error)
910 return error;
911
912 return 0;
913 }
914
915 /**
916 * sysc_show_rev - read and show interconnect target module revision
917 * @bufp: buffer to print the information to
918 * @ddata: device driver data
919 */
sysc_show_rev(char * bufp,struct sysc * ddata)920 static int sysc_show_rev(char *bufp, struct sysc *ddata)
921 {
922 int len;
923
924 if (ddata->offsets[SYSC_REVISION] < 0)
925 return sprintf(bufp, ":NA");
926
927 len = sprintf(bufp, ":%08x", ddata->revision);
928
929 return len;
930 }
931
sysc_show_reg(struct sysc * ddata,char * bufp,enum sysc_registers reg)932 static int sysc_show_reg(struct sysc *ddata,
933 char *bufp, enum sysc_registers reg)
934 {
935 if (ddata->offsets[reg] < 0)
936 return sprintf(bufp, ":NA");
937
938 return sprintf(bufp, ":%x", ddata->offsets[reg]);
939 }
940
sysc_show_name(char * bufp,struct sysc * ddata)941 static int sysc_show_name(char *bufp, struct sysc *ddata)
942 {
943 if (!ddata->name)
944 return 0;
945
946 return sprintf(bufp, ":%s", ddata->name);
947 }
948
949 /**
950 * sysc_show_registers - show information about interconnect target module
951 * @ddata: device driver data
952 */
sysc_show_registers(struct sysc * ddata)953 static void sysc_show_registers(struct sysc *ddata)
954 {
955 char buf[128];
956 char *bufp = buf;
957 int i;
958
959 for (i = 0; i < SYSC_MAX_REGS; i++)
960 bufp += sysc_show_reg(ddata, bufp, i);
961
962 bufp += sysc_show_rev(bufp, ddata);
963 bufp += sysc_show_name(bufp, ddata);
964
965 dev_dbg(ddata->dev, "%llx:%x%s\n",
966 ddata->module_pa, ddata->module_size,
967 buf);
968 }
969
970 /**
971 * sysc_write_sysconfig - handle sysconfig quirks for register write
972 * @ddata: device driver data
973 * @value: register value
974 */
sysc_write_sysconfig(struct sysc * ddata,u32 value)975 static void sysc_write_sysconfig(struct sysc *ddata, u32 value)
976 {
977 if (ddata->module_unlock_quirk)
978 ddata->module_unlock_quirk(ddata);
979
980 sysc_write(ddata, ddata->offsets[SYSC_SYSCONFIG], value);
981
982 if (ddata->module_lock_quirk)
983 ddata->module_lock_quirk(ddata);
984 }
985
986 #define SYSC_IDLE_MASK (SYSC_NR_IDLEMODES - 1)
987 #define SYSC_CLOCACT_ICK 2
988
989 /* Caller needs to manage sysc_clkdm_deny_idle() and sysc_clkdm_allow_idle() */
sysc_enable_module(struct device * dev)990 static int sysc_enable_module(struct device *dev)
991 {
992 struct sysc *ddata;
993 const struct sysc_regbits *regbits;
994 u32 reg, idlemodes, best_mode;
995 int error;
996
997 ddata = dev_get_drvdata(dev);
998
999 /*
1000 * Some modules like DSS reset automatically on idle. Enable optional
1001 * reset clocks and wait for OCP softreset to complete.
1002 */
1003 if (ddata->cfg.quirks & SYSC_QUIRK_OPT_CLKS_IN_RESET) {
1004 error = sysc_enable_opt_clocks(ddata);
1005 if (error) {
1006 dev_err(ddata->dev,
1007 "Optional clocks failed for enable: %i\n",
1008 error);
1009 return error;
1010 }
1011 }
1012 /*
1013 * Some modules like i2c and hdq1w have unusable reset status unless
1014 * the module reset quirk is enabled. Skip status check on enable.
1015 */
1016 if (!(ddata->cfg.quirks & SYSC_MODULE_QUIRK_ENA_RESETDONE)) {
1017 error = sysc_wait_softreset(ddata);
1018 if (error)
1019 dev_warn(ddata->dev, "OCP softreset timed out\n");
1020 }
1021 if (ddata->cfg.quirks & SYSC_QUIRK_OPT_CLKS_IN_RESET)
1022 sysc_disable_opt_clocks(ddata);
1023
1024 /*
1025 * Some subsystem private interconnects, like DSS top level module,
1026 * need only the automatic OCP softreset handling with no sysconfig
1027 * register bits to configure.
1028 */
1029 if (ddata->offsets[SYSC_SYSCONFIG] == -ENODEV)
1030 return 0;
1031
1032 regbits = ddata->cap->regbits;
1033 reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1034
1035 /*
1036 * Set CLOCKACTIVITY, we only use it for ick. And we only configure it
1037 * based on the SYSC_QUIRK_USE_CLOCKACT flag, not based on the hardware
1038 * capabilities. See the old HWMOD_SET_DEFAULT_CLOCKACT flag.
1039 */
1040 if (regbits->clkact_shift >= 0 &&
1041 (ddata->cfg.quirks & SYSC_QUIRK_USE_CLOCKACT))
1042 reg |= SYSC_CLOCACT_ICK << regbits->clkact_shift;
1043
1044 /* Set SIDLE mode */
1045 idlemodes = ddata->cfg.sidlemodes;
1046 if (!idlemodes || regbits->sidle_shift < 0)
1047 goto set_midle;
1048
1049 if (ddata->cfg.quirks & (SYSC_QUIRK_SWSUP_SIDLE |
1050 SYSC_QUIRK_SWSUP_SIDLE_ACT)) {
1051 best_mode = SYSC_IDLE_NO;
1052
1053 /* Clear WAKEUP */
1054 if (regbits->enwkup_shift >= 0 &&
1055 ddata->cfg.sysc_val & BIT(regbits->enwkup_shift))
1056 reg &= ~BIT(regbits->enwkup_shift);
1057 } else {
1058 best_mode = fls(ddata->cfg.sidlemodes) - 1;
1059 if (best_mode > SYSC_IDLE_MASK) {
1060 dev_err(dev, "%s: invalid sidlemode\n", __func__);
1061 return -EINVAL;
1062 }
1063
1064 /* Set WAKEUP */
1065 if (regbits->enwkup_shift >= 0 &&
1066 ddata->cfg.sysc_val & BIT(regbits->enwkup_shift))
1067 reg |= BIT(regbits->enwkup_shift);
1068 }
1069
1070 reg &= ~(SYSC_IDLE_MASK << regbits->sidle_shift);
1071 reg |= best_mode << regbits->sidle_shift;
1072 sysc_write_sysconfig(ddata, reg);
1073
1074 set_midle:
1075 /* Set MIDLE mode */
1076 idlemodes = ddata->cfg.midlemodes;
1077 if (!idlemodes || regbits->midle_shift < 0)
1078 goto set_autoidle;
1079
1080 best_mode = fls(ddata->cfg.midlemodes) - 1;
1081 if (best_mode > SYSC_IDLE_MASK) {
1082 dev_err(dev, "%s: invalid midlemode\n", __func__);
1083 error = -EINVAL;
1084 goto save_context;
1085 }
1086
1087 if (ddata->cfg.quirks & SYSC_QUIRK_SWSUP_MSTANDBY)
1088 best_mode = SYSC_IDLE_NO;
1089
1090 reg &= ~(SYSC_IDLE_MASK << regbits->midle_shift);
1091 reg |= best_mode << regbits->midle_shift;
1092 sysc_write_sysconfig(ddata, reg);
1093
1094 set_autoidle:
1095 /* Autoidle bit must enabled separately if available */
1096 if (regbits->autoidle_shift >= 0 &&
1097 ddata->cfg.sysc_val & BIT(regbits->autoidle_shift)) {
1098 reg |= 1 << regbits->autoidle_shift;
1099 sysc_write_sysconfig(ddata, reg);
1100 }
1101
1102 error = 0;
1103
1104 save_context:
1105 /* Save context and flush posted write */
1106 ddata->sysconfig = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1107
1108 if (ddata->module_enable_quirk)
1109 ddata->module_enable_quirk(ddata);
1110
1111 return error;
1112 }
1113
sysc_best_idle_mode(u32 idlemodes,u32 * best_mode)1114 static int sysc_best_idle_mode(u32 idlemodes, u32 *best_mode)
1115 {
1116 if (idlemodes & BIT(SYSC_IDLE_SMART_WKUP))
1117 *best_mode = SYSC_IDLE_SMART_WKUP;
1118 else if (idlemodes & BIT(SYSC_IDLE_SMART))
1119 *best_mode = SYSC_IDLE_SMART;
1120 else if (idlemodes & BIT(SYSC_IDLE_FORCE))
1121 *best_mode = SYSC_IDLE_FORCE;
1122 else
1123 return -EINVAL;
1124
1125 return 0;
1126 }
1127
1128 /* Caller needs to manage sysc_clkdm_deny_idle() and sysc_clkdm_allow_idle() */
sysc_disable_module(struct device * dev)1129 static int sysc_disable_module(struct device *dev)
1130 {
1131 struct sysc *ddata;
1132 const struct sysc_regbits *regbits;
1133 u32 reg, idlemodes, best_mode;
1134 int ret;
1135
1136 ddata = dev_get_drvdata(dev);
1137 if (ddata->offsets[SYSC_SYSCONFIG] == -ENODEV)
1138 return 0;
1139
1140 if (ddata->module_disable_quirk)
1141 ddata->module_disable_quirk(ddata);
1142
1143 regbits = ddata->cap->regbits;
1144 reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1145
1146 /* Set MIDLE mode */
1147 idlemodes = ddata->cfg.midlemodes;
1148 if (!idlemodes || regbits->midle_shift < 0)
1149 goto set_sidle;
1150
1151 ret = sysc_best_idle_mode(idlemodes, &best_mode);
1152 if (ret) {
1153 dev_err(dev, "%s: invalid midlemode\n", __func__);
1154 return ret;
1155 }
1156
1157 if (ddata->cfg.quirks & (SYSC_QUIRK_SWSUP_MSTANDBY) ||
1158 ddata->cfg.quirks & (SYSC_QUIRK_FORCE_MSTANDBY))
1159 best_mode = SYSC_IDLE_FORCE;
1160
1161 reg &= ~(SYSC_IDLE_MASK << regbits->midle_shift);
1162 reg |= best_mode << regbits->midle_shift;
1163 sysc_write_sysconfig(ddata, reg);
1164
1165 set_sidle:
1166 /* Set SIDLE mode */
1167 idlemodes = ddata->cfg.sidlemodes;
1168 if (!idlemodes || regbits->sidle_shift < 0) {
1169 ret = 0;
1170 goto save_context;
1171 }
1172
1173 if (ddata->cfg.quirks & SYSC_QUIRK_SWSUP_SIDLE) {
1174 best_mode = SYSC_IDLE_FORCE;
1175 } else {
1176 ret = sysc_best_idle_mode(idlemodes, &best_mode);
1177 if (ret) {
1178 dev_err(dev, "%s: invalid sidlemode\n", __func__);
1179 ret = -EINVAL;
1180 goto save_context;
1181 }
1182 }
1183
1184 if (ddata->cfg.quirks & SYSC_QUIRK_SWSUP_SIDLE_ACT) {
1185 /* Set WAKEUP */
1186 if (regbits->enwkup_shift >= 0 &&
1187 ddata->cfg.sysc_val & BIT(regbits->enwkup_shift))
1188 reg |= BIT(regbits->enwkup_shift);
1189 }
1190
1191 reg &= ~(SYSC_IDLE_MASK << regbits->sidle_shift);
1192 reg |= best_mode << regbits->sidle_shift;
1193 if (regbits->autoidle_shift >= 0 &&
1194 ddata->cfg.sysc_val & BIT(regbits->autoidle_shift))
1195 reg |= 1 << regbits->autoidle_shift;
1196 sysc_write_sysconfig(ddata, reg);
1197
1198 ret = 0;
1199
1200 save_context:
1201 /* Save context and flush posted write */
1202 ddata->sysconfig = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1203
1204 return ret;
1205 }
1206
sysc_runtime_suspend_legacy(struct device * dev,struct sysc * ddata)1207 static int __maybe_unused sysc_runtime_suspend_legacy(struct device *dev,
1208 struct sysc *ddata)
1209 {
1210 struct ti_sysc_platform_data *pdata;
1211 int error;
1212
1213 pdata = dev_get_platdata(ddata->dev);
1214 if (!pdata)
1215 return 0;
1216
1217 if (!pdata->idle_module)
1218 return -ENODEV;
1219
1220 error = pdata->idle_module(dev, &ddata->cookie);
1221 if (error)
1222 dev_err(dev, "%s: could not idle: %i\n",
1223 __func__, error);
1224
1225 reset_control_assert(ddata->rsts);
1226
1227 return 0;
1228 }
1229
sysc_runtime_resume_legacy(struct device * dev,struct sysc * ddata)1230 static int __maybe_unused sysc_runtime_resume_legacy(struct device *dev,
1231 struct sysc *ddata)
1232 {
1233 struct ti_sysc_platform_data *pdata;
1234 int error;
1235
1236 pdata = dev_get_platdata(ddata->dev);
1237 if (!pdata)
1238 return 0;
1239
1240 if (!pdata->enable_module)
1241 return -ENODEV;
1242
1243 error = pdata->enable_module(dev, &ddata->cookie);
1244 if (error)
1245 dev_err(dev, "%s: could not enable: %i\n",
1246 __func__, error);
1247
1248 reset_control_deassert(ddata->rsts);
1249
1250 return 0;
1251 }
1252
sysc_runtime_suspend(struct device * dev)1253 static int __maybe_unused sysc_runtime_suspend(struct device *dev)
1254 {
1255 struct sysc *ddata;
1256 int error = 0;
1257
1258 ddata = dev_get_drvdata(dev);
1259
1260 if (!ddata->enabled)
1261 return 0;
1262
1263 sysc_clkdm_deny_idle(ddata);
1264
1265 if (ddata->legacy_mode) {
1266 error = sysc_runtime_suspend_legacy(dev, ddata);
1267 if (error)
1268 goto err_allow_idle;
1269 } else {
1270 error = sysc_disable_module(dev);
1271 if (error)
1272 goto err_allow_idle;
1273 }
1274
1275 sysc_disable_main_clocks(ddata);
1276
1277 if (sysc_opt_clks_needed(ddata))
1278 sysc_disable_opt_clocks(ddata);
1279
1280 ddata->enabled = false;
1281
1282 err_allow_idle:
1283 sysc_clkdm_allow_idle(ddata);
1284
1285 reset_control_assert(ddata->rsts);
1286
1287 return error;
1288 }
1289
sysc_runtime_resume(struct device * dev)1290 static int __maybe_unused sysc_runtime_resume(struct device *dev)
1291 {
1292 struct sysc *ddata;
1293 int error = 0;
1294
1295 ddata = dev_get_drvdata(dev);
1296
1297 if (ddata->enabled)
1298 return 0;
1299
1300
1301 sysc_clkdm_deny_idle(ddata);
1302
1303 if (sysc_opt_clks_needed(ddata)) {
1304 error = sysc_enable_opt_clocks(ddata);
1305 if (error)
1306 goto err_allow_idle;
1307 }
1308
1309 error = sysc_enable_main_clocks(ddata);
1310 if (error)
1311 goto err_opt_clocks;
1312
1313 reset_control_deassert(ddata->rsts);
1314
1315 if (ddata->legacy_mode) {
1316 error = sysc_runtime_resume_legacy(dev, ddata);
1317 if (error)
1318 goto err_main_clocks;
1319 } else {
1320 error = sysc_enable_module(dev);
1321 if (error)
1322 goto err_main_clocks;
1323 }
1324
1325 ddata->enabled = true;
1326
1327 sysc_clkdm_allow_idle(ddata);
1328
1329 return 0;
1330
1331 err_main_clocks:
1332 sysc_disable_main_clocks(ddata);
1333 err_opt_clocks:
1334 if (sysc_opt_clks_needed(ddata))
1335 sysc_disable_opt_clocks(ddata);
1336 err_allow_idle:
1337 sysc_clkdm_allow_idle(ddata);
1338
1339 return error;
1340 }
1341
1342 /*
1343 * Checks if device context was lost. Assumes the sysconfig register value
1344 * after lost context is different from the configured value. Only works for
1345 * enabled devices.
1346 *
1347 * Eventually we may want to also add support to using the context lost
1348 * registers that some SoCs have.
1349 */
sysc_check_context(struct sysc * ddata)1350 static int sysc_check_context(struct sysc *ddata)
1351 {
1352 u32 reg;
1353
1354 if (!ddata->enabled)
1355 return -ENODATA;
1356
1357 reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1358 if (reg == ddata->sysconfig)
1359 return 0;
1360
1361 return -EACCES;
1362 }
1363
sysc_reinit_module(struct sysc * ddata,bool leave_enabled)1364 static int sysc_reinit_module(struct sysc *ddata, bool leave_enabled)
1365 {
1366 struct device *dev = ddata->dev;
1367 int error;
1368
1369 if (ddata->enabled) {
1370 /* Nothing to do if enabled and context not lost */
1371 error = sysc_check_context(ddata);
1372 if (!error)
1373 return 0;
1374
1375 /* Disable target module if it is enabled */
1376 error = sysc_runtime_suspend(dev);
1377 if (error)
1378 dev_warn(dev, "reinit suspend failed: %i\n", error);
1379 }
1380
1381 /* Enable target module */
1382 error = sysc_runtime_resume(dev);
1383 if (error)
1384 dev_warn(dev, "reinit resume failed: %i\n", error);
1385
1386 /* Some modules like am335x gpmc need reset and restore of sysconfig */
1387 if (ddata->cfg.quirks & SYSC_QUIRK_RESET_ON_CTX_LOST) {
1388 error = sysc_reset(ddata);
1389 if (error)
1390 dev_warn(dev, "reinit reset failed: %i\n", error);
1391
1392 sysc_write_sysconfig(ddata, ddata->sysconfig);
1393 }
1394
1395 if (leave_enabled)
1396 return error;
1397
1398 /* Disable target module if no leave_enabled was set */
1399 error = sysc_runtime_suspend(dev);
1400 if (error)
1401 dev_warn(dev, "reinit suspend failed: %i\n", error);
1402
1403 return error;
1404 }
1405
sysc_noirq_suspend(struct device * dev)1406 static int __maybe_unused sysc_noirq_suspend(struct device *dev)
1407 {
1408 struct sysc *ddata;
1409
1410 ddata = dev_get_drvdata(dev);
1411
1412 if (ddata->cfg.quirks &
1413 (SYSC_QUIRK_LEGACY_IDLE | SYSC_QUIRK_NO_IDLE))
1414 return 0;
1415
1416 if (!ddata->enabled)
1417 return 0;
1418
1419 ddata->needs_resume = 1;
1420
1421 return sysc_runtime_suspend(dev);
1422 }
1423
sysc_noirq_resume(struct device * dev)1424 static int __maybe_unused sysc_noirq_resume(struct device *dev)
1425 {
1426 struct sysc *ddata;
1427 int error = 0;
1428
1429 ddata = dev_get_drvdata(dev);
1430
1431 if (ddata->cfg.quirks &
1432 (SYSC_QUIRK_LEGACY_IDLE | SYSC_QUIRK_NO_IDLE))
1433 return 0;
1434
1435 if (ddata->cfg.quirks & SYSC_QUIRK_REINIT_ON_RESUME) {
1436 error = sysc_reinit_module(ddata, ddata->needs_resume);
1437 if (error)
1438 dev_warn(dev, "noirq_resume failed: %i\n", error);
1439 } else if (ddata->needs_resume) {
1440 error = sysc_runtime_resume(dev);
1441 if (error)
1442 dev_warn(dev, "noirq_resume failed: %i\n", error);
1443 }
1444
1445 ddata->needs_resume = 0;
1446
1447 return error;
1448 }
1449
1450 static const struct dev_pm_ops sysc_pm_ops = {
1451 SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(sysc_noirq_suspend, sysc_noirq_resume)
1452 SET_RUNTIME_PM_OPS(sysc_runtime_suspend,
1453 sysc_runtime_resume,
1454 NULL)
1455 };
1456
1457 /* Module revision register based quirks */
1458 struct sysc_revision_quirk {
1459 const char *name;
1460 u32 base;
1461 int rev_offset;
1462 int sysc_offset;
1463 int syss_offset;
1464 u32 revision;
1465 u32 revision_mask;
1466 u32 quirks;
1467 };
1468
1469 #define SYSC_QUIRK(optname, optbase, optrev, optsysc, optsyss, \
1470 optrev_val, optrevmask, optquirkmask) \
1471 { \
1472 .name = (optname), \
1473 .base = (optbase), \
1474 .rev_offset = (optrev), \
1475 .sysc_offset = (optsysc), \
1476 .syss_offset = (optsyss), \
1477 .revision = (optrev_val), \
1478 .revision_mask = (optrevmask), \
1479 .quirks = (optquirkmask), \
1480 }
1481
1482 static const struct sysc_revision_quirk sysc_revision_quirks[] = {
1483 /* These drivers need to be fixed to not use pm_runtime_irq_safe() */
1484 SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x00000046, 0xffffffff,
1485 SYSC_QUIRK_SWSUP_SIDLE_ACT | SYSC_QUIRK_LEGACY_IDLE),
1486 SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x00000052, 0xffffffff,
1487 SYSC_QUIRK_SWSUP_SIDLE_ACT | SYSC_QUIRK_LEGACY_IDLE),
1488 /* Uarts on omap4 and later */
1489 SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x50411e03, 0xffff00ff,
1490 SYSC_QUIRK_SWSUP_SIDLE_ACT | SYSC_QUIRK_LEGACY_IDLE),
1491 SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x47422e03, 0xffffffff,
1492 SYSC_QUIRK_SWSUP_SIDLE_ACT | SYSC_QUIRK_LEGACY_IDLE),
1493 SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x47424e03, 0xffffffff,
1494 SYSC_QUIRK_SWSUP_SIDLE_ACT | SYSC_QUIRK_LEGACY_IDLE),
1495
1496 /* Quirks that need to be set based on the module address */
1497 SYSC_QUIRK("mcpdm", 0x40132000, 0, 0x10, -ENODEV, 0x50000800, 0xffffffff,
1498 SYSC_QUIRK_EXT_OPT_CLOCK | SYSC_QUIRK_NO_RESET_ON_INIT |
1499 SYSC_QUIRK_SWSUP_SIDLE),
1500
1501 /* Quirks that need to be set based on detected module */
1502 SYSC_QUIRK("aess", 0, 0, 0x10, -ENODEV, 0x40000000, 0xffffffff,
1503 SYSC_MODULE_QUIRK_AESS),
1504 /* Errata i893 handling for dra7 dcan1 and 2 */
1505 SYSC_QUIRK("dcan", 0x4ae3c000, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff,
1506 SYSC_QUIRK_CLKDM_NOAUTO),
1507 SYSC_QUIRK("dcan", 0x48480000, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff,
1508 SYSC_QUIRK_CLKDM_NOAUTO),
1509 SYSC_QUIRK("dss", 0x4832a000, 0, 0x10, 0x14, 0x00000020, 0xffffffff,
1510 SYSC_QUIRK_OPT_CLKS_IN_RESET | SYSC_MODULE_QUIRK_DSS_RESET),
1511 SYSC_QUIRK("dss", 0x58000000, 0, -ENODEV, 0x14, 0x00000040, 0xffffffff,
1512 SYSC_QUIRK_OPT_CLKS_IN_RESET | SYSC_MODULE_QUIRK_DSS_RESET),
1513 SYSC_QUIRK("dss", 0x58000000, 0, -ENODEV, 0x14, 0x00000061, 0xffffffff,
1514 SYSC_QUIRK_OPT_CLKS_IN_RESET | SYSC_MODULE_QUIRK_DSS_RESET),
1515 SYSC_QUIRK("dwc3", 0x48880000, 0, 0x10, -ENODEV, 0x500a0200, 0xffffffff,
1516 SYSC_QUIRK_CLKDM_NOAUTO),
1517 SYSC_QUIRK("dwc3", 0x488c0000, 0, 0x10, -ENODEV, 0x500a0200, 0xffffffff,
1518 SYSC_QUIRK_CLKDM_NOAUTO),
1519 SYSC_QUIRK("gpio", 0, 0, 0x10, 0x114, 0x50600801, 0xffff00ff,
1520 SYSC_QUIRK_OPT_CLKS_IN_RESET),
1521 SYSC_QUIRK("gpmc", 0, 0, 0x10, 0x14, 0x00000060, 0xffffffff,
1522 SYSC_QUIRK_REINIT_ON_CTX_LOST | SYSC_QUIRK_RESET_ON_CTX_LOST |
1523 SYSC_QUIRK_GPMC_DEBUG),
1524 SYSC_QUIRK("hdmi", 0, 0, 0x10, -ENODEV, 0x50030200, 0xffffffff,
1525 SYSC_QUIRK_OPT_CLKS_NEEDED),
1526 SYSC_QUIRK("hdq1w", 0, 0, 0x14, 0x18, 0x00000006, 0xffffffff,
1527 SYSC_MODULE_QUIRK_HDQ1W | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1528 SYSC_QUIRK("hdq1w", 0, 0, 0x14, 0x18, 0x0000000a, 0xffffffff,
1529 SYSC_MODULE_QUIRK_HDQ1W | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1530 SYSC_QUIRK("i2c", 0, 0, 0x20, 0x10, 0x00000036, 0x000000ff,
1531 SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1532 SYSC_QUIRK("i2c", 0, 0, 0x20, 0x10, 0x0000003c, 0x000000ff,
1533 SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1534 SYSC_QUIRK("i2c", 0, 0, 0x20, 0x10, 0x00000040, 0x000000ff,
1535 SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1536 SYSC_QUIRK("i2c", 0, 0, 0x10, 0x90, 0x5040000a, 0xfffff0f0,
1537 SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1538 SYSC_QUIRK("gpu", 0x50000000, 0x14, -ENODEV, -ENODEV, 0x00010201, 0xffffffff, 0),
1539 SYSC_QUIRK("gpu", 0x50000000, 0xfe00, 0xfe10, -ENODEV, 0x40000000 , 0xffffffff,
1540 SYSC_MODULE_QUIRK_SGX),
1541 SYSC_QUIRK("lcdc", 0, 0, 0x54, -ENODEV, 0x4f201000, 0xffffffff,
1542 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1543 SYSC_QUIRK("mcasp", 0, 0, 0x4, -ENODEV, 0x44306302, 0xffffffff,
1544 SYSC_QUIRK_SWSUP_SIDLE),
1545 SYSC_QUIRK("rtc", 0, 0x74, 0x78, -ENODEV, 0x4eb01908, 0xffff00f0,
1546 SYSC_MODULE_QUIRK_RTC_UNLOCK),
1547 SYSC_QUIRK("tptc", 0, 0, 0x10, -ENODEV, 0x40006c00, 0xffffefff,
1548 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1549 SYSC_QUIRK("tptc", 0, 0, -ENODEV, -ENODEV, 0x40007c00, 0xffffffff,
1550 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1551 SYSC_QUIRK("sata", 0, 0xfc, 0x1100, -ENODEV, 0x5e412000, 0xffffffff,
1552 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1553 SYSC_QUIRK("usb_host_hs", 0, 0, 0x10, 0x14, 0x50700100, 0xffffffff,
1554 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1555 SYSC_QUIRK("usb_host_hs", 0, 0, 0x10, -ENODEV, 0x50700101, 0xffffffff,
1556 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1557 SYSC_QUIRK("usb_otg_hs", 0, 0x400, 0x404, 0x408, 0x00000033,
1558 0xffffffff, SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1559 SYSC_MODULE_QUIRK_OTG),
1560 SYSC_QUIRK("usb_otg_hs", 0, 0x400, 0x404, 0x408, 0x00000040,
1561 0xffffffff, SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1562 SYSC_MODULE_QUIRK_OTG),
1563 SYSC_QUIRK("usb_otg_hs", 0, 0x400, 0x404, 0x408, 0x00000050,
1564 0xffffffff, SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1565 SYSC_MODULE_QUIRK_OTG),
1566 SYSC_QUIRK("usb_otg_hs", 0, 0, 0x10, -ENODEV, 0x4ea2080d, 0xffffffff,
1567 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1568 SYSC_QUIRK_REINIT_ON_CTX_LOST),
1569 SYSC_QUIRK("wdt", 0, 0, 0x10, 0x14, 0x502a0500, 0xfffff0f0,
1570 SYSC_MODULE_QUIRK_WDT),
1571 /* PRUSS on am3, am4 and am5 */
1572 SYSC_QUIRK("pruss", 0, 0x26000, 0x26004, -ENODEV, 0x47000000, 0xff000000,
1573 SYSC_MODULE_QUIRK_PRUSS),
1574 /* Watchdog on am3 and am4 */
1575 SYSC_QUIRK("wdt", 0x44e35000, 0, 0x10, 0x14, 0x502a0500, 0xfffff0f0,
1576 SYSC_MODULE_QUIRK_WDT | SYSC_QUIRK_SWSUP_SIDLE),
1577
1578 #ifdef DEBUG
1579 SYSC_QUIRK("adc", 0, 0, 0x10, -ENODEV, 0x47300001, 0xffffffff, 0),
1580 SYSC_QUIRK("atl", 0, 0, -ENODEV, -ENODEV, 0x0a070100, 0xffffffff, 0),
1581 SYSC_QUIRK("cm", 0, 0, -ENODEV, -ENODEV, 0x40000301, 0xffffffff, 0),
1582 SYSC_QUIRK("control", 0, 0, 0x10, -ENODEV, 0x40000900, 0xffffffff, 0),
1583 SYSC_QUIRK("cpgmac", 0, 0x1200, 0x1208, 0x1204, 0x4edb1902,
1584 0xffff00f0, 0),
1585 SYSC_QUIRK("dcan", 0, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff, 0),
1586 SYSC_QUIRK("dcan", 0, 0x20, -ENODEV, -ENODEV, 0x4edb1902, 0xffffffff, 0),
1587 SYSC_QUIRK("dispc", 0x4832a400, 0, 0x10, 0x14, 0x00000030, 0xffffffff, 0),
1588 SYSC_QUIRK("dispc", 0x58001000, 0, 0x10, 0x14, 0x00000040, 0xffffffff, 0),
1589 SYSC_QUIRK("dispc", 0x58001000, 0, 0x10, 0x14, 0x00000051, 0xffffffff, 0),
1590 SYSC_QUIRK("dmic", 0, 0, 0x10, -ENODEV, 0x50010000, 0xffffffff, 0),
1591 SYSC_QUIRK("dsi", 0x58004000, 0, 0x10, 0x14, 0x00000030, 0xffffffff, 0),
1592 SYSC_QUIRK("dsi", 0x58005000, 0, 0x10, 0x14, 0x00000030, 0xffffffff, 0),
1593 SYSC_QUIRK("dsi", 0x58005000, 0, 0x10, 0x14, 0x00000040, 0xffffffff, 0),
1594 SYSC_QUIRK("dsi", 0x58009000, 0, 0x10, 0x14, 0x00000040, 0xffffffff, 0),
1595 SYSC_QUIRK("dwc3", 0, 0, 0x10, -ENODEV, 0x500a0200, 0xffffffff, 0),
1596 SYSC_QUIRK("d2d", 0x4a0b6000, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1597 SYSC_QUIRK("d2d", 0x4a0cd000, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1598 SYSC_QUIRK("elm", 0x48080000, 0, 0x10, 0x14, 0x00000020, 0xffffffff, 0),
1599 SYSC_QUIRK("emif", 0, 0, -ENODEV, -ENODEV, 0x40441403, 0xffff0fff, 0),
1600 SYSC_QUIRK("emif", 0, 0, -ENODEV, -ENODEV, 0x50440500, 0xffffffff, 0),
1601 SYSC_QUIRK("epwmss", 0, 0, 0x4, -ENODEV, 0x47400001, 0xffffffff, 0),
1602 SYSC_QUIRK("gpu", 0, 0x1fc00, 0x1fc10, -ENODEV, 0, 0, 0),
1603 SYSC_QUIRK("gpu", 0, 0xfe00, 0xfe10, -ENODEV, 0x40000000 , 0xffffffff, 0),
1604 SYSC_QUIRK("hdmi", 0, 0, 0x10, -ENODEV, 0x50031d00, 0xffffffff, 0),
1605 SYSC_QUIRK("hsi", 0, 0, 0x10, 0x14, 0x50043101, 0xffffffff, 0),
1606 SYSC_QUIRK("iss", 0, 0, 0x10, -ENODEV, 0x40000101, 0xffffffff, 0),
1607 SYSC_QUIRK("keypad", 0x4a31c000, 0, 0x10, 0x14, 0x00000020, 0xffffffff, 0),
1608 SYSC_QUIRK("mcasp", 0, 0, 0x4, -ENODEV, 0x44307b02, 0xffffffff, 0),
1609 SYSC_QUIRK("mcbsp", 0, -ENODEV, 0x8c, -ENODEV, 0, 0, 0),
1610 SYSC_QUIRK("mcspi", 0, 0, 0x10, -ENODEV, 0x40300a0b, 0xffff00ff, 0),
1611 SYSC_QUIRK("mcspi", 0, 0, 0x110, 0x114, 0x40300a0b, 0xffffffff, 0),
1612 SYSC_QUIRK("mailbox", 0, 0, 0x10, -ENODEV, 0x00000400, 0xffffffff, 0),
1613 SYSC_QUIRK("m3", 0, 0, -ENODEV, -ENODEV, 0x5f580105, 0x0fff0f00, 0),
1614 SYSC_QUIRK("ocp2scp", 0, 0, 0x10, 0x14, 0x50060005, 0xfffffff0, 0),
1615 SYSC_QUIRK("ocp2scp", 0, 0, -ENODEV, -ENODEV, 0x50060007, 0xffffffff, 0),
1616 SYSC_QUIRK("padconf", 0, 0, 0x10, -ENODEV, 0x4fff0800, 0xffffffff, 0),
1617 SYSC_QUIRK("padconf", 0, 0, -ENODEV, -ENODEV, 0x40001100, 0xffffffff, 0),
1618 SYSC_QUIRK("pcie", 0x51000000, -ENODEV, -ENODEV, -ENODEV, 0, 0, 0),
1619 SYSC_QUIRK("pcie", 0x51800000, -ENODEV, -ENODEV, -ENODEV, 0, 0, 0),
1620 SYSC_QUIRK("prcm", 0, 0, -ENODEV, -ENODEV, 0x40000100, 0xffffffff, 0),
1621 SYSC_QUIRK("prcm", 0, 0, -ENODEV, -ENODEV, 0x00004102, 0xffffffff, 0),
1622 SYSC_QUIRK("prcm", 0, 0, -ENODEV, -ENODEV, 0x40000400, 0xffffffff, 0),
1623 SYSC_QUIRK("rfbi", 0x4832a800, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1624 SYSC_QUIRK("rfbi", 0x58002000, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1625 SYSC_QUIRK("scm", 0, 0, 0x10, -ENODEV, 0x40000900, 0xffffffff, 0),
1626 SYSC_QUIRK("scm", 0, 0, -ENODEV, -ENODEV, 0x4e8b0100, 0xffffffff, 0),
1627 SYSC_QUIRK("scm", 0, 0, -ENODEV, -ENODEV, 0x4f000100, 0xffffffff, 0),
1628 SYSC_QUIRK("scm", 0, 0, -ENODEV, -ENODEV, 0x40000900, 0xffffffff, 0),
1629 SYSC_QUIRK("scrm", 0, 0, -ENODEV, -ENODEV, 0x00000010, 0xffffffff, 0),
1630 SYSC_QUIRK("sdio", 0, 0, 0x10, -ENODEV, 0x40202301, 0xffff0ff0, 0),
1631 SYSC_QUIRK("sdio", 0, 0x2fc, 0x110, 0x114, 0x31010000, 0xffffffff, 0),
1632 SYSC_QUIRK("sdma", 0, 0, 0x2c, 0x28, 0x00010900, 0xffffffff, 0),
1633 SYSC_QUIRK("sham", 0, 0x100, 0x110, 0x114, 0x40000c03, 0xffffffff, 0),
1634 SYSC_QUIRK("slimbus", 0, 0, 0x10, -ENODEV, 0x40000902, 0xffffffff, 0),
1635 SYSC_QUIRK("slimbus", 0, 0, 0x10, -ENODEV, 0x40002903, 0xffffffff, 0),
1636 SYSC_QUIRK("smartreflex", 0, -ENODEV, 0x24, -ENODEV, 0x00000000, 0xffffffff, 0),
1637 SYSC_QUIRK("smartreflex", 0, -ENODEV, 0x38, -ENODEV, 0x00000000, 0xffffffff, 0),
1638 SYSC_QUIRK("spinlock", 0, 0, 0x10, -ENODEV, 0x50020000, 0xffffffff, 0),
1639 SYSC_QUIRK("rng", 0, 0x1fe0, 0x1fe4, -ENODEV, 0x00000020, 0xffffffff, 0),
1640 SYSC_QUIRK("timer", 0, 0, 0x10, 0x14, 0x00000013, 0xffffffff, 0),
1641 SYSC_QUIRK("timer", 0, 0, 0x10, 0x14, 0x00000015, 0xffffffff, 0),
1642 /* Some timers on omap4 and later */
1643 SYSC_QUIRK("timer", 0, 0, 0x10, -ENODEV, 0x50002100, 0xffffffff, 0),
1644 SYSC_QUIRK("timer", 0, 0, 0x10, -ENODEV, 0x4fff1301, 0xffff00ff, 0),
1645 SYSC_QUIRK("timer32k", 0, 0, 0x4, -ENODEV, 0x00000040, 0xffffffff, 0),
1646 SYSC_QUIRK("timer32k", 0, 0, 0x4, -ENODEV, 0x00000011, 0xffffffff, 0),
1647 SYSC_QUIRK("timer32k", 0, 0, 0x4, -ENODEV, 0x00000060, 0xffffffff, 0),
1648 SYSC_QUIRK("tpcc", 0, 0, -ENODEV, -ENODEV, 0x40014c00, 0xffffffff, 0),
1649 SYSC_QUIRK("usbhstll", 0, 0, 0x10, 0x14, 0x00000004, 0xffffffff, 0),
1650 SYSC_QUIRK("usbhstll", 0, 0, 0x10, 0x14, 0x00000008, 0xffffffff, 0),
1651 SYSC_QUIRK("venc", 0x58003000, 0, -ENODEV, -ENODEV, 0x00000002, 0xffffffff, 0),
1652 SYSC_QUIRK("vfpe", 0, 0, 0x104, -ENODEV, 0x4d001200, 0xffffffff, 0),
1653 #endif
1654 };
1655
1656 /*
1657 * Early quirks based on module base and register offsets only that are
1658 * needed before the module revision can be read
1659 */
sysc_init_early_quirks(struct sysc * ddata)1660 static void sysc_init_early_quirks(struct sysc *ddata)
1661 {
1662 const struct sysc_revision_quirk *q;
1663 int i;
1664
1665 for (i = 0; i < ARRAY_SIZE(sysc_revision_quirks); i++) {
1666 q = &sysc_revision_quirks[i];
1667
1668 if (!q->base)
1669 continue;
1670
1671 if (q->base != ddata->module_pa)
1672 continue;
1673
1674 if (q->rev_offset != ddata->offsets[SYSC_REVISION])
1675 continue;
1676
1677 if (q->sysc_offset != ddata->offsets[SYSC_SYSCONFIG])
1678 continue;
1679
1680 if (q->syss_offset != ddata->offsets[SYSC_SYSSTATUS])
1681 continue;
1682
1683 ddata->name = q->name;
1684 ddata->cfg.quirks |= q->quirks;
1685 }
1686 }
1687
1688 /* Quirks that also consider the revision register value */
sysc_init_revision_quirks(struct sysc * ddata)1689 static void sysc_init_revision_quirks(struct sysc *ddata)
1690 {
1691 const struct sysc_revision_quirk *q;
1692 int i;
1693
1694 for (i = 0; i < ARRAY_SIZE(sysc_revision_quirks); i++) {
1695 q = &sysc_revision_quirks[i];
1696
1697 if (q->base && q->base != ddata->module_pa)
1698 continue;
1699
1700 if (q->rev_offset != ddata->offsets[SYSC_REVISION])
1701 continue;
1702
1703 if (q->sysc_offset != ddata->offsets[SYSC_SYSCONFIG])
1704 continue;
1705
1706 if (q->syss_offset != ddata->offsets[SYSC_SYSSTATUS])
1707 continue;
1708
1709 if (q->revision == ddata->revision ||
1710 (q->revision & q->revision_mask) ==
1711 (ddata->revision & q->revision_mask)) {
1712 ddata->name = q->name;
1713 ddata->cfg.quirks |= q->quirks;
1714 }
1715 }
1716 }
1717
1718 /*
1719 * DSS needs dispc outputs disabled to reset modules. Returns mask of
1720 * enabled DSS interrupts. Eventually we may be able to do this on
1721 * dispc init rather than top-level DSS init.
1722 */
sysc_quirk_dispc(struct sysc * ddata,int dispc_offset,bool disable)1723 static u32 sysc_quirk_dispc(struct sysc *ddata, int dispc_offset,
1724 bool disable)
1725 {
1726 bool lcd_en, digit_en, lcd2_en = false, lcd3_en = false;
1727 const int lcd_en_mask = BIT(0), digit_en_mask = BIT(1);
1728 int manager_count;
1729 bool framedonetv_irq = true;
1730 u32 val, irq_mask = 0;
1731
1732 switch (sysc_soc->soc) {
1733 case SOC_2420 ... SOC_3630:
1734 manager_count = 2;
1735 framedonetv_irq = false;
1736 break;
1737 case SOC_4430 ... SOC_4470:
1738 manager_count = 3;
1739 break;
1740 case SOC_5430:
1741 case SOC_DRA7:
1742 manager_count = 4;
1743 break;
1744 case SOC_AM4:
1745 manager_count = 1;
1746 framedonetv_irq = false;
1747 break;
1748 case SOC_UNKNOWN:
1749 default:
1750 return 0;
1751 }
1752
1753 /* Remap the whole module range to be able to reset dispc outputs */
1754 devm_iounmap(ddata->dev, ddata->module_va);
1755 ddata->module_va = devm_ioremap(ddata->dev,
1756 ddata->module_pa,
1757 ddata->module_size);
1758 if (!ddata->module_va)
1759 return -EIO;
1760
1761 /* DISP_CONTROL, shut down lcd and digit on disable if enabled */
1762 val = sysc_read(ddata, dispc_offset + 0x40);
1763 lcd_en = val & lcd_en_mask;
1764 digit_en = val & digit_en_mask;
1765 if (lcd_en)
1766 irq_mask |= BIT(0); /* FRAMEDONE */
1767 if (digit_en) {
1768 if (framedonetv_irq)
1769 irq_mask |= BIT(24); /* FRAMEDONETV */
1770 else
1771 irq_mask |= BIT(2) | BIT(3); /* EVSYNC bits */
1772 }
1773 if (disable && (lcd_en || digit_en))
1774 sysc_write(ddata, dispc_offset + 0x40,
1775 val & ~(lcd_en_mask | digit_en_mask));
1776
1777 if (manager_count <= 2)
1778 return irq_mask;
1779
1780 /* DISPC_CONTROL2 */
1781 val = sysc_read(ddata, dispc_offset + 0x238);
1782 lcd2_en = val & lcd_en_mask;
1783 if (lcd2_en)
1784 irq_mask |= BIT(22); /* FRAMEDONE2 */
1785 if (disable && lcd2_en)
1786 sysc_write(ddata, dispc_offset + 0x238,
1787 val & ~lcd_en_mask);
1788
1789 if (manager_count <= 3)
1790 return irq_mask;
1791
1792 /* DISPC_CONTROL3 */
1793 val = sysc_read(ddata, dispc_offset + 0x848);
1794 lcd3_en = val & lcd_en_mask;
1795 if (lcd3_en)
1796 irq_mask |= BIT(30); /* FRAMEDONE3 */
1797 if (disable && lcd3_en)
1798 sysc_write(ddata, dispc_offset + 0x848,
1799 val & ~lcd_en_mask);
1800
1801 return irq_mask;
1802 }
1803
1804 /* DSS needs child outputs disabled and SDI registers cleared for reset */
sysc_pre_reset_quirk_dss(struct sysc * ddata)1805 static void sysc_pre_reset_quirk_dss(struct sysc *ddata)
1806 {
1807 const int dispc_offset = 0x1000;
1808 int error;
1809 u32 irq_mask, val;
1810
1811 /* Get enabled outputs */
1812 irq_mask = sysc_quirk_dispc(ddata, dispc_offset, false);
1813 if (!irq_mask)
1814 return;
1815
1816 /* Clear IRQSTATUS */
1817 sysc_write(ddata, dispc_offset + 0x18, irq_mask);
1818
1819 /* Disable outputs */
1820 val = sysc_quirk_dispc(ddata, dispc_offset, true);
1821
1822 /* Poll IRQSTATUS */
1823 error = readl_poll_timeout(ddata->module_va + dispc_offset + 0x18,
1824 val, val != irq_mask, 100, 50);
1825 if (error)
1826 dev_warn(ddata->dev, "%s: timed out %08x !+ %08x\n",
1827 __func__, val, irq_mask);
1828
1829 if (sysc_soc->soc == SOC_3430 || sysc_soc->soc == SOC_AM35) {
1830 /* Clear DSS_SDI_CONTROL */
1831 sysc_write(ddata, 0x44, 0);
1832
1833 /* Clear DSS_PLL_CONTROL */
1834 sysc_write(ddata, 0x48, 0);
1835 }
1836
1837 /* Clear DSS_CONTROL to switch DSS clock sources to PRCM if not */
1838 sysc_write(ddata, 0x40, 0);
1839 }
1840
1841 /* 1-wire needs module's internal clocks enabled for reset */
sysc_pre_reset_quirk_hdq1w(struct sysc * ddata)1842 static void sysc_pre_reset_quirk_hdq1w(struct sysc *ddata)
1843 {
1844 int offset = 0x0c; /* HDQ_CTRL_STATUS */
1845 u16 val;
1846
1847 val = sysc_read(ddata, offset);
1848 val |= BIT(5);
1849 sysc_write(ddata, offset, val);
1850 }
1851
1852 /* AESS (Audio Engine SubSystem) needs autogating set after enable */
sysc_module_enable_quirk_aess(struct sysc * ddata)1853 static void sysc_module_enable_quirk_aess(struct sysc *ddata)
1854 {
1855 int offset = 0x7c; /* AESS_AUTO_GATING_ENABLE */
1856
1857 sysc_write(ddata, offset, 1);
1858 }
1859
1860 /* I2C needs to be disabled for reset */
sysc_clk_quirk_i2c(struct sysc * ddata,bool enable)1861 static void sysc_clk_quirk_i2c(struct sysc *ddata, bool enable)
1862 {
1863 int offset;
1864 u16 val;
1865
1866 /* I2C_CON, omap2/3 is different from omap4 and later */
1867 if ((ddata->revision & 0xffffff00) == 0x001f0000)
1868 offset = 0x24;
1869 else
1870 offset = 0xa4;
1871
1872 /* I2C_EN */
1873 val = sysc_read(ddata, offset);
1874 if (enable)
1875 val |= BIT(15);
1876 else
1877 val &= ~BIT(15);
1878 sysc_write(ddata, offset, val);
1879 }
1880
sysc_pre_reset_quirk_i2c(struct sysc * ddata)1881 static void sysc_pre_reset_quirk_i2c(struct sysc *ddata)
1882 {
1883 sysc_clk_quirk_i2c(ddata, false);
1884 }
1885
sysc_post_reset_quirk_i2c(struct sysc * ddata)1886 static void sysc_post_reset_quirk_i2c(struct sysc *ddata)
1887 {
1888 sysc_clk_quirk_i2c(ddata, true);
1889 }
1890
1891 /* RTC on am3 and 4 needs to be unlocked and locked for sysconfig */
sysc_quirk_rtc(struct sysc * ddata,bool lock)1892 static void sysc_quirk_rtc(struct sysc *ddata, bool lock)
1893 {
1894 u32 val, kick0_val = 0, kick1_val = 0;
1895 unsigned long flags;
1896 int error;
1897
1898 if (!lock) {
1899 kick0_val = 0x83e70b13;
1900 kick1_val = 0x95a4f1e0;
1901 }
1902
1903 local_irq_save(flags);
1904 /* RTC_STATUS BUSY bit may stay active for 1/32768 seconds (~30 usec) */
1905 error = readl_poll_timeout_atomic(ddata->module_va + 0x44, val,
1906 !(val & BIT(0)), 100, 50);
1907 if (error)
1908 dev_warn(ddata->dev, "rtc busy timeout\n");
1909 /* Now we have ~15 microseconds to read/write various registers */
1910 sysc_write(ddata, 0x6c, kick0_val);
1911 sysc_write(ddata, 0x70, kick1_val);
1912 local_irq_restore(flags);
1913 }
1914
sysc_module_unlock_quirk_rtc(struct sysc * ddata)1915 static void sysc_module_unlock_quirk_rtc(struct sysc *ddata)
1916 {
1917 sysc_quirk_rtc(ddata, false);
1918 }
1919
sysc_module_lock_quirk_rtc(struct sysc * ddata)1920 static void sysc_module_lock_quirk_rtc(struct sysc *ddata)
1921 {
1922 sysc_quirk_rtc(ddata, true);
1923 }
1924
1925 /* OTG omap2430 glue layer up to omap4 needs OTG_FORCESTDBY configured */
sysc_module_enable_quirk_otg(struct sysc * ddata)1926 static void sysc_module_enable_quirk_otg(struct sysc *ddata)
1927 {
1928 int offset = 0x414; /* OTG_FORCESTDBY */
1929
1930 sysc_write(ddata, offset, 0);
1931 }
1932
sysc_module_disable_quirk_otg(struct sysc * ddata)1933 static void sysc_module_disable_quirk_otg(struct sysc *ddata)
1934 {
1935 int offset = 0x414; /* OTG_FORCESTDBY */
1936 u32 val = BIT(0); /* ENABLEFORCE */
1937
1938 sysc_write(ddata, offset, val);
1939 }
1940
1941 /* 36xx SGX needs a quirk for to bypass OCP IPG interrupt logic */
sysc_module_enable_quirk_sgx(struct sysc * ddata)1942 static void sysc_module_enable_quirk_sgx(struct sysc *ddata)
1943 {
1944 int offset = 0xff08; /* OCP_DEBUG_CONFIG */
1945 u32 val = BIT(31); /* THALIA_INT_BYPASS */
1946
1947 sysc_write(ddata, offset, val);
1948 }
1949
1950 /* Watchdog timer needs a disable sequence after reset */
sysc_reset_done_quirk_wdt(struct sysc * ddata)1951 static void sysc_reset_done_quirk_wdt(struct sysc *ddata)
1952 {
1953 int wps, spr, error;
1954 u32 val;
1955
1956 wps = 0x34;
1957 spr = 0x48;
1958
1959 sysc_write(ddata, spr, 0xaaaa);
1960 error = readl_poll_timeout(ddata->module_va + wps, val,
1961 !(val & 0x10), 100,
1962 MAX_MODULE_SOFTRESET_WAIT);
1963 if (error)
1964 dev_warn(ddata->dev, "wdt disable step1 failed\n");
1965
1966 sysc_write(ddata, spr, 0x5555);
1967 error = readl_poll_timeout(ddata->module_va + wps, val,
1968 !(val & 0x10), 100,
1969 MAX_MODULE_SOFTRESET_WAIT);
1970 if (error)
1971 dev_warn(ddata->dev, "wdt disable step2 failed\n");
1972 }
1973
1974 /* PRUSS needs to set MSTANDBY_INIT inorder to idle properly */
sysc_module_disable_quirk_pruss(struct sysc * ddata)1975 static void sysc_module_disable_quirk_pruss(struct sysc *ddata)
1976 {
1977 u32 reg;
1978
1979 reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1980 reg |= SYSC_PRUSS_STANDBY_INIT;
1981 sysc_write(ddata, ddata->offsets[SYSC_SYSCONFIG], reg);
1982 }
1983
sysc_init_module_quirks(struct sysc * ddata)1984 static void sysc_init_module_quirks(struct sysc *ddata)
1985 {
1986 if (ddata->legacy_mode || !ddata->name)
1987 return;
1988
1989 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_HDQ1W) {
1990 ddata->pre_reset_quirk = sysc_pre_reset_quirk_hdq1w;
1991
1992 return;
1993 }
1994
1995 #ifdef CONFIG_OMAP_GPMC_DEBUG
1996 if (ddata->cfg.quirks & SYSC_QUIRK_GPMC_DEBUG) {
1997 ddata->cfg.quirks |= SYSC_QUIRK_NO_RESET_ON_INIT;
1998
1999 return;
2000 }
2001 #endif
2002
2003 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_I2C) {
2004 ddata->pre_reset_quirk = sysc_pre_reset_quirk_i2c;
2005 ddata->post_reset_quirk = sysc_post_reset_quirk_i2c;
2006
2007 return;
2008 }
2009
2010 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_AESS)
2011 ddata->module_enable_quirk = sysc_module_enable_quirk_aess;
2012
2013 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_DSS_RESET)
2014 ddata->pre_reset_quirk = sysc_pre_reset_quirk_dss;
2015
2016 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_RTC_UNLOCK) {
2017 ddata->module_unlock_quirk = sysc_module_unlock_quirk_rtc;
2018 ddata->module_lock_quirk = sysc_module_lock_quirk_rtc;
2019
2020 return;
2021 }
2022
2023 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_OTG) {
2024 ddata->module_enable_quirk = sysc_module_enable_quirk_otg;
2025 ddata->module_disable_quirk = sysc_module_disable_quirk_otg;
2026 }
2027
2028 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_SGX)
2029 ddata->module_enable_quirk = sysc_module_enable_quirk_sgx;
2030
2031 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_WDT) {
2032 ddata->reset_done_quirk = sysc_reset_done_quirk_wdt;
2033 ddata->module_disable_quirk = sysc_reset_done_quirk_wdt;
2034 }
2035
2036 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_PRUSS)
2037 ddata->module_disable_quirk = sysc_module_disable_quirk_pruss;
2038 }
2039
sysc_clockdomain_init(struct sysc * ddata)2040 static int sysc_clockdomain_init(struct sysc *ddata)
2041 {
2042 struct ti_sysc_platform_data *pdata = dev_get_platdata(ddata->dev);
2043 struct clk *fck = NULL, *ick = NULL;
2044 int error;
2045
2046 if (!pdata || !pdata->init_clockdomain)
2047 return 0;
2048
2049 switch (ddata->nr_clocks) {
2050 case 2:
2051 ick = ddata->clocks[SYSC_ICK];
2052 fallthrough;
2053 case 1:
2054 fck = ddata->clocks[SYSC_FCK];
2055 break;
2056 case 0:
2057 return 0;
2058 }
2059
2060 error = pdata->init_clockdomain(ddata->dev, fck, ick, &ddata->cookie);
2061 if (!error || error == -ENODEV)
2062 return 0;
2063
2064 return error;
2065 }
2066
2067 /*
2068 * Note that pdata->init_module() typically does a reset first. After
2069 * pdata->init_module() is done, PM runtime can be used for the interconnect
2070 * target module.
2071 */
sysc_legacy_init(struct sysc * ddata)2072 static int sysc_legacy_init(struct sysc *ddata)
2073 {
2074 struct ti_sysc_platform_data *pdata = dev_get_platdata(ddata->dev);
2075 int error;
2076
2077 if (!pdata || !pdata->init_module)
2078 return 0;
2079
2080 error = pdata->init_module(ddata->dev, ddata->mdata, &ddata->cookie);
2081 if (error == -EEXIST)
2082 error = 0;
2083
2084 return error;
2085 }
2086
2087 /*
2088 * Note that the caller must ensure the interconnect target module is enabled
2089 * before calling reset. Otherwise reset will not complete.
2090 */
sysc_reset(struct sysc * ddata)2091 static int sysc_reset(struct sysc *ddata)
2092 {
2093 int sysc_offset, sysc_val, error;
2094 u32 sysc_mask;
2095
2096 sysc_offset = ddata->offsets[SYSC_SYSCONFIG];
2097
2098 if (ddata->legacy_mode ||
2099 ddata->cap->regbits->srst_shift < 0 ||
2100 ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT)
2101 return 0;
2102
2103 sysc_mask = BIT(ddata->cap->regbits->srst_shift);
2104
2105 if (ddata->pre_reset_quirk)
2106 ddata->pre_reset_quirk(ddata);
2107
2108 if (sysc_offset >= 0) {
2109 sysc_val = sysc_read_sysconfig(ddata);
2110 sysc_val |= sysc_mask;
2111 sysc_write(ddata, sysc_offset, sysc_val);
2112
2113 /*
2114 * Some devices need a delay before reading registers
2115 * after reset. Presumably a srst_udelay is not needed
2116 * for devices that use a rstctrl register reset.
2117 */
2118 if (ddata->cfg.srst_udelay)
2119 fsleep(ddata->cfg.srst_udelay);
2120
2121 /*
2122 * Flush posted write. For devices needing srst_udelay
2123 * this should trigger an interconnect error if the
2124 * srst_udelay value is needed but not configured.
2125 */
2126 sysc_val = sysc_read_sysconfig(ddata);
2127 }
2128
2129 if (ddata->post_reset_quirk)
2130 ddata->post_reset_quirk(ddata);
2131
2132 error = sysc_wait_softreset(ddata);
2133 if (error)
2134 dev_warn(ddata->dev, "OCP softreset timed out\n");
2135
2136 if (ddata->reset_done_quirk)
2137 ddata->reset_done_quirk(ddata);
2138
2139 return error;
2140 }
2141
2142 /*
2143 * At this point the module is configured enough to read the revision but
2144 * module may not be completely configured yet to use PM runtime. Enable
2145 * all clocks directly during init to configure the quirks needed for PM
2146 * runtime based on the revision register.
2147 */
sysc_init_module(struct sysc * ddata)2148 static int sysc_init_module(struct sysc *ddata)
2149 {
2150 bool rstctrl_deasserted = false;
2151 int error = 0;
2152
2153 error = sysc_clockdomain_init(ddata);
2154 if (error)
2155 return error;
2156
2157 sysc_clkdm_deny_idle(ddata);
2158
2159 /*
2160 * Always enable clocks. The bootloader may or may not have enabled
2161 * the related clocks.
2162 */
2163 error = sysc_enable_opt_clocks(ddata);
2164 if (error)
2165 return error;
2166
2167 error = sysc_enable_main_clocks(ddata);
2168 if (error)
2169 goto err_opt_clocks;
2170
2171 if (!(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT)) {
2172 error = reset_control_deassert(ddata->rsts);
2173 if (error)
2174 goto err_main_clocks;
2175 rstctrl_deasserted = true;
2176 }
2177
2178 ddata->revision = sysc_read_revision(ddata);
2179 sysc_init_revision_quirks(ddata);
2180 sysc_init_module_quirks(ddata);
2181
2182 if (ddata->legacy_mode) {
2183 error = sysc_legacy_init(ddata);
2184 if (error)
2185 goto err_main_clocks;
2186 }
2187
2188 if (!ddata->legacy_mode) {
2189 error = sysc_enable_module(ddata->dev);
2190 if (error)
2191 goto err_main_clocks;
2192 }
2193
2194 error = sysc_reset(ddata);
2195 if (error)
2196 dev_err(ddata->dev, "Reset failed with %d\n", error);
2197
2198 if (error && !ddata->legacy_mode)
2199 sysc_disable_module(ddata->dev);
2200
2201 err_main_clocks:
2202 if (error)
2203 sysc_disable_main_clocks(ddata);
2204 err_opt_clocks:
2205 /* No re-enable of clockdomain autoidle to prevent module autoidle */
2206 if (error) {
2207 sysc_disable_opt_clocks(ddata);
2208 sysc_clkdm_allow_idle(ddata);
2209 }
2210
2211 if (error && rstctrl_deasserted &&
2212 !(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT))
2213 reset_control_assert(ddata->rsts);
2214
2215 return error;
2216 }
2217
sysc_init_sysc_mask(struct sysc * ddata)2218 static int sysc_init_sysc_mask(struct sysc *ddata)
2219 {
2220 struct device_node *np = ddata->dev->of_node;
2221 int error;
2222 u32 val;
2223
2224 error = of_property_read_u32(np, "ti,sysc-mask", &val);
2225 if (error)
2226 return 0;
2227
2228 ddata->cfg.sysc_val = val & ddata->cap->sysc_mask;
2229
2230 return 0;
2231 }
2232
sysc_init_idlemode(struct sysc * ddata,u8 * idlemodes,const char * name)2233 static int sysc_init_idlemode(struct sysc *ddata, u8 *idlemodes,
2234 const char *name)
2235 {
2236 struct device_node *np = ddata->dev->of_node;
2237 u32 val;
2238
2239 of_property_for_each_u32(np, name, val) {
2240 if (val >= SYSC_NR_IDLEMODES) {
2241 dev_err(ddata->dev, "invalid idlemode: %i\n", val);
2242 return -EINVAL;
2243 }
2244 *idlemodes |= (1 << val);
2245 }
2246
2247 return 0;
2248 }
2249
sysc_init_idlemodes(struct sysc * ddata)2250 static int sysc_init_idlemodes(struct sysc *ddata)
2251 {
2252 int error;
2253
2254 error = sysc_init_idlemode(ddata, &ddata->cfg.midlemodes,
2255 "ti,sysc-midle");
2256 if (error)
2257 return error;
2258
2259 error = sysc_init_idlemode(ddata, &ddata->cfg.sidlemodes,
2260 "ti,sysc-sidle");
2261 if (error)
2262 return error;
2263
2264 return 0;
2265 }
2266
2267 /*
2268 * Only some devices on omap4 and later have SYSCONFIG reset done
2269 * bit. We can detect this if there is no SYSSTATUS at all, or the
2270 * SYSTATUS bit 0 is not used. Note that some SYSSTATUS registers
2271 * have multiple bits for the child devices like OHCI and EHCI.
2272 * Depends on SYSC being parsed first.
2273 */
sysc_init_syss_mask(struct sysc * ddata)2274 static int sysc_init_syss_mask(struct sysc *ddata)
2275 {
2276 struct device_node *np = ddata->dev->of_node;
2277 int error;
2278 u32 val;
2279
2280 error = of_property_read_u32(np, "ti,syss-mask", &val);
2281 if (error) {
2282 if ((ddata->cap->type == TI_SYSC_OMAP4 ||
2283 ddata->cap->type == TI_SYSC_OMAP4_TIMER) &&
2284 (ddata->cfg.sysc_val & SYSC_OMAP4_SOFTRESET))
2285 ddata->cfg.quirks |= SYSC_QUIRK_RESET_STATUS;
2286
2287 return 0;
2288 }
2289
2290 if (!(val & 1) && (ddata->cfg.sysc_val & SYSC_OMAP4_SOFTRESET))
2291 ddata->cfg.quirks |= SYSC_QUIRK_RESET_STATUS;
2292
2293 ddata->cfg.syss_mask = val;
2294
2295 return 0;
2296 }
2297
2298 /*
2299 * Many child device drivers need to have fck and opt clocks available
2300 * to get the clock rate for device internal configuration etc.
2301 */
sysc_child_add_named_clock(struct sysc * ddata,struct device * child,const char * name)2302 static int sysc_child_add_named_clock(struct sysc *ddata,
2303 struct device *child,
2304 const char *name)
2305 {
2306 struct clk *clk;
2307 struct clk_lookup *l;
2308 int error = 0;
2309
2310 if (!name)
2311 return 0;
2312
2313 clk = clk_get(child, name);
2314 if (!IS_ERR(clk)) {
2315 error = -EEXIST;
2316 goto put_clk;
2317 }
2318
2319 clk = clk_get(ddata->dev, name);
2320 if (IS_ERR(clk))
2321 return -ENODEV;
2322
2323 l = clkdev_create(clk, name, dev_name(child));
2324 if (!l)
2325 error = -ENOMEM;
2326 put_clk:
2327 clk_put(clk);
2328
2329 return error;
2330 }
2331
sysc_child_add_clocks(struct sysc * ddata,struct device * child)2332 static int sysc_child_add_clocks(struct sysc *ddata,
2333 struct device *child)
2334 {
2335 int i, error;
2336
2337 for (i = 0; i < ddata->nr_clocks; i++) {
2338 error = sysc_child_add_named_clock(ddata,
2339 child,
2340 ddata->clock_roles[i]);
2341 if (error && error != -EEXIST) {
2342 dev_err(ddata->dev, "could not add child clock %s: %i\n",
2343 ddata->clock_roles[i], error);
2344
2345 return error;
2346 }
2347 }
2348
2349 return 0;
2350 }
2351
2352 static struct device_type sysc_device_type = {
2353 };
2354
sysc_child_to_parent(struct device * dev)2355 static struct sysc *sysc_child_to_parent(struct device *dev)
2356 {
2357 struct device *parent = dev->parent;
2358
2359 if (!parent || parent->type != &sysc_device_type)
2360 return NULL;
2361
2362 return dev_get_drvdata(parent);
2363 }
2364
sysc_child_runtime_suspend(struct device * dev)2365 static int __maybe_unused sysc_child_runtime_suspend(struct device *dev)
2366 {
2367 struct sysc *ddata;
2368 int error;
2369
2370 ddata = sysc_child_to_parent(dev);
2371
2372 error = pm_generic_runtime_suspend(dev);
2373 if (error)
2374 return error;
2375
2376 if (!ddata->enabled)
2377 return 0;
2378
2379 return sysc_runtime_suspend(ddata->dev);
2380 }
2381
sysc_child_runtime_resume(struct device * dev)2382 static int __maybe_unused sysc_child_runtime_resume(struct device *dev)
2383 {
2384 struct sysc *ddata;
2385 int error;
2386
2387 ddata = sysc_child_to_parent(dev);
2388
2389 if (!ddata->enabled) {
2390 error = sysc_runtime_resume(ddata->dev);
2391 if (error < 0)
2392 dev_err(ddata->dev,
2393 "%s error: %i\n", __func__, error);
2394 }
2395
2396 return pm_generic_runtime_resume(dev);
2397 }
2398
2399 #ifdef CONFIG_PM_SLEEP
sysc_child_suspend_noirq(struct device * dev)2400 static int sysc_child_suspend_noirq(struct device *dev)
2401 {
2402 struct sysc *ddata;
2403 int error;
2404
2405 ddata = sysc_child_to_parent(dev);
2406
2407 dev_dbg(ddata->dev, "%s %s\n", __func__,
2408 ddata->name ? ddata->name : "");
2409
2410 error = pm_generic_suspend_noirq(dev);
2411 if (error) {
2412 dev_err(dev, "%s error at %i: %i\n",
2413 __func__, __LINE__, error);
2414
2415 return error;
2416 }
2417
2418 if (!pm_runtime_status_suspended(dev)) {
2419 error = pm_generic_runtime_suspend(dev);
2420 if (error) {
2421 dev_dbg(dev, "%s busy at %i: %i\n",
2422 __func__, __LINE__, error);
2423
2424 return 0;
2425 }
2426
2427 error = sysc_runtime_suspend(ddata->dev);
2428 if (error) {
2429 dev_err(dev, "%s error at %i: %i\n",
2430 __func__, __LINE__, error);
2431
2432 return error;
2433 }
2434
2435 ddata->child_needs_resume = true;
2436 }
2437
2438 return 0;
2439 }
2440
sysc_child_resume_noirq(struct device * dev)2441 static int sysc_child_resume_noirq(struct device *dev)
2442 {
2443 struct sysc *ddata;
2444 int error;
2445
2446 ddata = sysc_child_to_parent(dev);
2447
2448 dev_dbg(ddata->dev, "%s %s\n", __func__,
2449 ddata->name ? ddata->name : "");
2450
2451 if (ddata->child_needs_resume) {
2452 ddata->child_needs_resume = false;
2453
2454 error = sysc_runtime_resume(ddata->dev);
2455 if (error)
2456 dev_err(ddata->dev,
2457 "%s runtime resume error: %i\n",
2458 __func__, error);
2459
2460 error = pm_generic_runtime_resume(dev);
2461 if (error)
2462 dev_err(ddata->dev,
2463 "%s generic runtime resume: %i\n",
2464 __func__, error);
2465 }
2466
2467 return pm_generic_resume_noirq(dev);
2468 }
2469 #endif
2470
2471 static struct dev_pm_domain sysc_child_pm_domain = {
2472 .ops = {
2473 SET_RUNTIME_PM_OPS(sysc_child_runtime_suspend,
2474 sysc_child_runtime_resume,
2475 NULL)
2476 USE_PLATFORM_PM_SLEEP_OPS
2477 SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(sysc_child_suspend_noirq,
2478 sysc_child_resume_noirq)
2479 }
2480 };
2481
2482 /* Caller needs to take list_lock if ever used outside of cpu_pm */
sysc_reinit_modules(struct sysc_soc_info * soc)2483 static void sysc_reinit_modules(struct sysc_soc_info *soc)
2484 {
2485 struct sysc_module *module;
2486 struct sysc *ddata;
2487
2488 list_for_each_entry(module, &sysc_soc->restored_modules, node) {
2489 ddata = module->ddata;
2490 sysc_reinit_module(ddata, ddata->enabled);
2491 }
2492 }
2493
2494 /**
2495 * sysc_context_notifier - optionally reset and restore module after idle
2496 * @nb: notifier block
2497 * @cmd: unused
2498 * @v: unused
2499 *
2500 * Some interconnect target modules need to be restored, or reset and restored
2501 * on CPU_PM CPU_PM_CLUSTER_EXIT notifier. This is needed at least for am335x
2502 * OTG and GPMC target modules even if the modules are unused.
2503 */
sysc_context_notifier(struct notifier_block * nb,unsigned long cmd,void * v)2504 static int sysc_context_notifier(struct notifier_block *nb, unsigned long cmd,
2505 void *v)
2506 {
2507 struct sysc_soc_info *soc;
2508
2509 soc = container_of(nb, struct sysc_soc_info, nb);
2510
2511 switch (cmd) {
2512 case CPU_CLUSTER_PM_ENTER:
2513 break;
2514 case CPU_CLUSTER_PM_ENTER_FAILED: /* No need to restore context */
2515 break;
2516 case CPU_CLUSTER_PM_EXIT:
2517 sysc_reinit_modules(soc);
2518 break;
2519 }
2520
2521 return NOTIFY_OK;
2522 }
2523
2524 /**
2525 * sysc_add_restored - optionally add reset and restore quirk hanlling
2526 * @ddata: device data
2527 */
sysc_add_restored(struct sysc * ddata)2528 static void sysc_add_restored(struct sysc *ddata)
2529 {
2530 struct sysc_module *restored_module;
2531
2532 restored_module = kzalloc(sizeof(*restored_module), GFP_KERNEL);
2533 if (!restored_module)
2534 return;
2535
2536 restored_module->ddata = ddata;
2537
2538 mutex_lock(&sysc_soc->list_lock);
2539
2540 list_add(&restored_module->node, &sysc_soc->restored_modules);
2541
2542 if (sysc_soc->nb.notifier_call)
2543 goto out_unlock;
2544
2545 sysc_soc->nb.notifier_call = sysc_context_notifier;
2546 cpu_pm_register_notifier(&sysc_soc->nb);
2547
2548 out_unlock:
2549 mutex_unlock(&sysc_soc->list_lock);
2550 }
2551
2552 /**
2553 * sysc_legacy_idle_quirk - handle children in omap_device compatible way
2554 * @ddata: device driver data
2555 * @child: child device driver
2556 *
2557 * Allow idle for child devices as done with _od_runtime_suspend().
2558 * Otherwise many child devices will not idle because of the permanent
2559 * parent usecount set in pm_runtime_irq_safe().
2560 *
2561 * Note that the long term solution is to just modify the child device
2562 * drivers to not set pm_runtime_irq_safe() and then this can be just
2563 * dropped.
2564 */
sysc_legacy_idle_quirk(struct sysc * ddata,struct device * child)2565 static void sysc_legacy_idle_quirk(struct sysc *ddata, struct device *child)
2566 {
2567 if (ddata->cfg.quirks & SYSC_QUIRK_LEGACY_IDLE)
2568 dev_pm_domain_set(child, &sysc_child_pm_domain);
2569 }
2570
sysc_notifier_call(struct notifier_block * nb,unsigned long event,void * device)2571 static int sysc_notifier_call(struct notifier_block *nb,
2572 unsigned long event, void *device)
2573 {
2574 struct device *dev = device;
2575 struct sysc *ddata;
2576 int error;
2577
2578 ddata = sysc_child_to_parent(dev);
2579 if (!ddata)
2580 return NOTIFY_DONE;
2581
2582 switch (event) {
2583 case BUS_NOTIFY_ADD_DEVICE:
2584 error = sysc_child_add_clocks(ddata, dev);
2585 if (error)
2586 return error;
2587 sysc_legacy_idle_quirk(ddata, dev);
2588 break;
2589 default:
2590 break;
2591 }
2592
2593 return NOTIFY_DONE;
2594 }
2595
2596 static struct notifier_block sysc_nb = {
2597 .notifier_call = sysc_notifier_call,
2598 };
2599
2600 /* Device tree configured quirks */
2601 struct sysc_dts_quirk {
2602 const char *name;
2603 u32 mask;
2604 };
2605
2606 static const struct sysc_dts_quirk sysc_dts_quirks[] = {
2607 { .name = "ti,no-idle-on-init",
2608 .mask = SYSC_QUIRK_NO_IDLE_ON_INIT, },
2609 { .name = "ti,no-reset-on-init",
2610 .mask = SYSC_QUIRK_NO_RESET_ON_INIT, },
2611 { .name = "ti,no-idle",
2612 .mask = SYSC_QUIRK_NO_IDLE, },
2613 };
2614
sysc_parse_dts_quirks(struct sysc * ddata,struct device_node * np,bool is_child)2615 static void sysc_parse_dts_quirks(struct sysc *ddata, struct device_node *np,
2616 bool is_child)
2617 {
2618 const struct property *prop;
2619 int i, len;
2620
2621 for (i = 0; i < ARRAY_SIZE(sysc_dts_quirks); i++) {
2622 const char *name = sysc_dts_quirks[i].name;
2623
2624 prop = of_get_property(np, name, &len);
2625 if (!prop)
2626 continue;
2627
2628 ddata->cfg.quirks |= sysc_dts_quirks[i].mask;
2629 if (is_child) {
2630 dev_warn(ddata->dev,
2631 "dts flag should be at module level for %s\n",
2632 name);
2633 }
2634 }
2635 }
2636
sysc_init_dts_quirks(struct sysc * ddata)2637 static int sysc_init_dts_quirks(struct sysc *ddata)
2638 {
2639 struct device_node *np = ddata->dev->of_node;
2640 int error;
2641 u32 val;
2642
2643 ddata->legacy_mode = of_get_property(np, "ti,hwmods", NULL);
2644
2645 sysc_parse_dts_quirks(ddata, np, false);
2646 error = of_property_read_u32(np, "ti,sysc-delay-us", &val);
2647 if (!error) {
2648 if (val > 255) {
2649 dev_warn(ddata->dev, "bad ti,sysc-delay-us: %i\n",
2650 val);
2651 }
2652
2653 ddata->cfg.srst_udelay = (u8)val;
2654 }
2655
2656 return 0;
2657 }
2658
sysc_unprepare(struct sysc * ddata)2659 static void sysc_unprepare(struct sysc *ddata)
2660 {
2661 int i;
2662
2663 if (!ddata->clocks)
2664 return;
2665
2666 for (i = 0; i < SYSC_MAX_CLOCKS; i++) {
2667 if (!IS_ERR_OR_NULL(ddata->clocks[i]))
2668 clk_unprepare(ddata->clocks[i]);
2669 }
2670 }
2671
2672 /*
2673 * Common sysc register bits found on omap2, also known as type1
2674 */
2675 static const struct sysc_regbits sysc_regbits_omap2 = {
2676 .dmadisable_shift = -ENODEV,
2677 .midle_shift = 12,
2678 .sidle_shift = 3,
2679 .clkact_shift = 8,
2680 .emufree_shift = 5,
2681 .enwkup_shift = 2,
2682 .srst_shift = 1,
2683 .autoidle_shift = 0,
2684 };
2685
2686 static const struct sysc_capabilities sysc_omap2 = {
2687 .type = TI_SYSC_OMAP2,
2688 .sysc_mask = SYSC_OMAP2_CLOCKACTIVITY | SYSC_OMAP2_EMUFREE |
2689 SYSC_OMAP2_ENAWAKEUP | SYSC_OMAP2_SOFTRESET |
2690 SYSC_OMAP2_AUTOIDLE,
2691 .regbits = &sysc_regbits_omap2,
2692 };
2693
2694 /* All omap2 and 3 timers, and timers 1, 2 & 10 on omap 4 and 5 */
2695 static const struct sysc_capabilities sysc_omap2_timer = {
2696 .type = TI_SYSC_OMAP2_TIMER,
2697 .sysc_mask = SYSC_OMAP2_CLOCKACTIVITY | SYSC_OMAP2_EMUFREE |
2698 SYSC_OMAP2_ENAWAKEUP | SYSC_OMAP2_SOFTRESET |
2699 SYSC_OMAP2_AUTOIDLE,
2700 .regbits = &sysc_regbits_omap2,
2701 .mod_quirks = SYSC_QUIRK_USE_CLOCKACT,
2702 };
2703
2704 /*
2705 * SHAM2 (SHA1/MD5) sysc found on omap3, a variant of sysc_regbits_omap2
2706 * with different sidle position
2707 */
2708 static const struct sysc_regbits sysc_regbits_omap3_sham = {
2709 .dmadisable_shift = -ENODEV,
2710 .midle_shift = -ENODEV,
2711 .sidle_shift = 4,
2712 .clkact_shift = -ENODEV,
2713 .enwkup_shift = -ENODEV,
2714 .srst_shift = 1,
2715 .autoidle_shift = 0,
2716 .emufree_shift = -ENODEV,
2717 };
2718
2719 static const struct sysc_capabilities sysc_omap3_sham = {
2720 .type = TI_SYSC_OMAP3_SHAM,
2721 .sysc_mask = SYSC_OMAP2_SOFTRESET | SYSC_OMAP2_AUTOIDLE,
2722 .regbits = &sysc_regbits_omap3_sham,
2723 };
2724
2725 /*
2726 * AES register bits found on omap3 and later, a variant of
2727 * sysc_regbits_omap2 with different sidle position
2728 */
2729 static const struct sysc_regbits sysc_regbits_omap3_aes = {
2730 .dmadisable_shift = -ENODEV,
2731 .midle_shift = -ENODEV,
2732 .sidle_shift = 6,
2733 .clkact_shift = -ENODEV,
2734 .enwkup_shift = -ENODEV,
2735 .srst_shift = 1,
2736 .autoidle_shift = 0,
2737 .emufree_shift = -ENODEV,
2738 };
2739
2740 static const struct sysc_capabilities sysc_omap3_aes = {
2741 .type = TI_SYSC_OMAP3_AES,
2742 .sysc_mask = SYSC_OMAP2_SOFTRESET | SYSC_OMAP2_AUTOIDLE,
2743 .regbits = &sysc_regbits_omap3_aes,
2744 };
2745
2746 /*
2747 * Common sysc register bits found on omap4, also known as type2
2748 */
2749 static const struct sysc_regbits sysc_regbits_omap4 = {
2750 .dmadisable_shift = 16,
2751 .midle_shift = 4,
2752 .sidle_shift = 2,
2753 .clkact_shift = -ENODEV,
2754 .enwkup_shift = -ENODEV,
2755 .emufree_shift = 1,
2756 .srst_shift = 0,
2757 .autoidle_shift = -ENODEV,
2758 };
2759
2760 static const struct sysc_capabilities sysc_omap4 = {
2761 .type = TI_SYSC_OMAP4,
2762 .sysc_mask = SYSC_OMAP4_DMADISABLE | SYSC_OMAP4_FREEEMU |
2763 SYSC_OMAP4_SOFTRESET,
2764 .regbits = &sysc_regbits_omap4,
2765 };
2766
2767 static const struct sysc_capabilities sysc_omap4_timer = {
2768 .type = TI_SYSC_OMAP4_TIMER,
2769 .sysc_mask = SYSC_OMAP4_DMADISABLE | SYSC_OMAP4_FREEEMU |
2770 SYSC_OMAP4_SOFTRESET,
2771 .regbits = &sysc_regbits_omap4,
2772 };
2773
2774 /*
2775 * Common sysc register bits found on omap4, also known as type3
2776 */
2777 static const struct sysc_regbits sysc_regbits_omap4_simple = {
2778 .dmadisable_shift = -ENODEV,
2779 .midle_shift = 2,
2780 .sidle_shift = 0,
2781 .clkact_shift = -ENODEV,
2782 .enwkup_shift = -ENODEV,
2783 .srst_shift = -ENODEV,
2784 .emufree_shift = -ENODEV,
2785 .autoidle_shift = -ENODEV,
2786 };
2787
2788 static const struct sysc_capabilities sysc_omap4_simple = {
2789 .type = TI_SYSC_OMAP4_SIMPLE,
2790 .regbits = &sysc_regbits_omap4_simple,
2791 };
2792
2793 /*
2794 * SmartReflex sysc found on omap34xx
2795 */
2796 static const struct sysc_regbits sysc_regbits_omap34xx_sr = {
2797 .dmadisable_shift = -ENODEV,
2798 .midle_shift = -ENODEV,
2799 .sidle_shift = -ENODEV,
2800 .clkact_shift = 20,
2801 .enwkup_shift = -ENODEV,
2802 .srst_shift = -ENODEV,
2803 .emufree_shift = -ENODEV,
2804 .autoidle_shift = -ENODEV,
2805 };
2806
2807 static const struct sysc_capabilities sysc_34xx_sr = {
2808 .type = TI_SYSC_OMAP34XX_SR,
2809 .sysc_mask = SYSC_OMAP2_CLOCKACTIVITY,
2810 .regbits = &sysc_regbits_omap34xx_sr,
2811 .mod_quirks = SYSC_QUIRK_USE_CLOCKACT | SYSC_QUIRK_UNCACHED |
2812 SYSC_QUIRK_LEGACY_IDLE,
2813 };
2814
2815 /*
2816 * SmartReflex sysc found on omap36xx and later
2817 */
2818 static const struct sysc_regbits sysc_regbits_omap36xx_sr = {
2819 .dmadisable_shift = -ENODEV,
2820 .midle_shift = -ENODEV,
2821 .sidle_shift = 24,
2822 .clkact_shift = -ENODEV,
2823 .enwkup_shift = 26,
2824 .srst_shift = -ENODEV,
2825 .emufree_shift = -ENODEV,
2826 .autoidle_shift = -ENODEV,
2827 };
2828
2829 static const struct sysc_capabilities sysc_36xx_sr = {
2830 .type = TI_SYSC_OMAP36XX_SR,
2831 .sysc_mask = SYSC_OMAP3_SR_ENAWAKEUP,
2832 .regbits = &sysc_regbits_omap36xx_sr,
2833 .mod_quirks = SYSC_QUIRK_UNCACHED | SYSC_QUIRK_LEGACY_IDLE,
2834 };
2835
2836 static const struct sysc_capabilities sysc_omap4_sr = {
2837 .type = TI_SYSC_OMAP4_SR,
2838 .regbits = &sysc_regbits_omap36xx_sr,
2839 .mod_quirks = SYSC_QUIRK_LEGACY_IDLE,
2840 };
2841
2842 /*
2843 * McASP register bits found on omap4 and later
2844 */
2845 static const struct sysc_regbits sysc_regbits_omap4_mcasp = {
2846 .dmadisable_shift = -ENODEV,
2847 .midle_shift = -ENODEV,
2848 .sidle_shift = 0,
2849 .clkact_shift = -ENODEV,
2850 .enwkup_shift = -ENODEV,
2851 .srst_shift = -ENODEV,
2852 .emufree_shift = -ENODEV,
2853 .autoidle_shift = -ENODEV,
2854 };
2855
2856 static const struct sysc_capabilities sysc_omap4_mcasp = {
2857 .type = TI_SYSC_OMAP4_MCASP,
2858 .regbits = &sysc_regbits_omap4_mcasp,
2859 .mod_quirks = SYSC_QUIRK_OPT_CLKS_NEEDED,
2860 };
2861
2862 /*
2863 * McASP found on dra7 and later
2864 */
2865 static const struct sysc_capabilities sysc_dra7_mcasp = {
2866 .type = TI_SYSC_OMAP4_SIMPLE,
2867 .regbits = &sysc_regbits_omap4_simple,
2868 .mod_quirks = SYSC_QUIRK_OPT_CLKS_NEEDED,
2869 };
2870
2871 /*
2872 * FS USB host found on omap4 and later
2873 */
2874 static const struct sysc_regbits sysc_regbits_omap4_usb_host_fs = {
2875 .dmadisable_shift = -ENODEV,
2876 .midle_shift = -ENODEV,
2877 .sidle_shift = 24,
2878 .clkact_shift = -ENODEV,
2879 .enwkup_shift = 26,
2880 .srst_shift = -ENODEV,
2881 .emufree_shift = -ENODEV,
2882 .autoidle_shift = -ENODEV,
2883 };
2884
2885 static const struct sysc_capabilities sysc_omap4_usb_host_fs = {
2886 .type = TI_SYSC_OMAP4_USB_HOST_FS,
2887 .sysc_mask = SYSC_OMAP2_ENAWAKEUP,
2888 .regbits = &sysc_regbits_omap4_usb_host_fs,
2889 };
2890
2891 static const struct sysc_regbits sysc_regbits_dra7_mcan = {
2892 .dmadisable_shift = -ENODEV,
2893 .midle_shift = -ENODEV,
2894 .sidle_shift = -ENODEV,
2895 .clkact_shift = -ENODEV,
2896 .enwkup_shift = 4,
2897 .srst_shift = 0,
2898 .emufree_shift = -ENODEV,
2899 .autoidle_shift = -ENODEV,
2900 };
2901
2902 static const struct sysc_capabilities sysc_dra7_mcan = {
2903 .type = TI_SYSC_DRA7_MCAN,
2904 .sysc_mask = SYSC_DRA7_MCAN_ENAWAKEUP | SYSC_OMAP4_SOFTRESET,
2905 .regbits = &sysc_regbits_dra7_mcan,
2906 .mod_quirks = SYSS_QUIRK_RESETDONE_INVERTED,
2907 };
2908
2909 /*
2910 * PRUSS found on some AM33xx, AM437x and AM57xx SoCs
2911 */
2912 static const struct sysc_capabilities sysc_pruss = {
2913 .type = TI_SYSC_PRUSS,
2914 .sysc_mask = SYSC_PRUSS_STANDBY_INIT | SYSC_PRUSS_SUB_MWAIT,
2915 .regbits = &sysc_regbits_omap4_simple,
2916 .mod_quirks = SYSC_MODULE_QUIRK_PRUSS,
2917 };
2918
sysc_init_pdata(struct sysc * ddata)2919 static int sysc_init_pdata(struct sysc *ddata)
2920 {
2921 struct ti_sysc_platform_data *pdata = dev_get_platdata(ddata->dev);
2922 struct ti_sysc_module_data *mdata;
2923
2924 if (!pdata)
2925 return 0;
2926
2927 mdata = devm_kzalloc(ddata->dev, sizeof(*mdata), GFP_KERNEL);
2928 if (!mdata)
2929 return -ENOMEM;
2930
2931 if (ddata->legacy_mode) {
2932 mdata->name = ddata->legacy_mode;
2933 mdata->module_pa = ddata->module_pa;
2934 mdata->module_size = ddata->module_size;
2935 mdata->offsets = ddata->offsets;
2936 mdata->nr_offsets = SYSC_MAX_REGS;
2937 mdata->cap = ddata->cap;
2938 mdata->cfg = &ddata->cfg;
2939 }
2940
2941 ddata->mdata = mdata;
2942
2943 return 0;
2944 }
2945
sysc_init_match(struct sysc * ddata)2946 static int sysc_init_match(struct sysc *ddata)
2947 {
2948 const struct sysc_capabilities *cap;
2949
2950 cap = of_device_get_match_data(ddata->dev);
2951 if (!cap)
2952 return -EINVAL;
2953
2954 ddata->cap = cap;
2955 if (ddata->cap)
2956 ddata->cfg.quirks |= ddata->cap->mod_quirks;
2957
2958 return 0;
2959 }
2960
ti_sysc_idle(struct work_struct * work)2961 static void ti_sysc_idle(struct work_struct *work)
2962 {
2963 struct sysc *ddata;
2964
2965 ddata = container_of(work, struct sysc, idle_work.work);
2966
2967 /*
2968 * One time decrement of clock usage counts if left on from init.
2969 * Note that we disable opt clocks unconditionally in this case
2970 * as they are enabled unconditionally during init without
2971 * considering sysc_opt_clks_needed() at that point.
2972 */
2973 if (ddata->cfg.quirks & (SYSC_QUIRK_NO_IDLE |
2974 SYSC_QUIRK_NO_IDLE_ON_INIT)) {
2975 sysc_disable_main_clocks(ddata);
2976 sysc_disable_opt_clocks(ddata);
2977 sysc_clkdm_allow_idle(ddata);
2978 }
2979
2980 /* Keep permanent PM runtime usage count for SYSC_QUIRK_NO_IDLE */
2981 if (ddata->cfg.quirks & SYSC_QUIRK_NO_IDLE)
2982 return;
2983
2984 /*
2985 * Decrement PM runtime usage count for SYSC_QUIRK_NO_IDLE_ON_INIT
2986 * and SYSC_QUIRK_NO_RESET_ON_INIT
2987 */
2988 if (pm_runtime_active(ddata->dev))
2989 pm_runtime_put_sync(ddata->dev);
2990 }
2991
2992 /*
2993 * SoC model and features detection. Only needed for SoCs that need
2994 * special handling for quirks, no need to list others.
2995 */
2996 static const struct soc_device_attribute sysc_soc_match[] = {
2997 SOC_FLAG("OMAP242*", SOC_2420),
2998 SOC_FLAG("OMAP243*", SOC_2430),
2999 SOC_FLAG("AM35*", SOC_AM35),
3000 SOC_FLAG("OMAP3[45]*", SOC_3430),
3001 SOC_FLAG("OMAP3[67]*", SOC_3630),
3002 SOC_FLAG("OMAP443*", SOC_4430),
3003 SOC_FLAG("OMAP446*", SOC_4460),
3004 SOC_FLAG("OMAP447*", SOC_4470),
3005 SOC_FLAG("OMAP54*", SOC_5430),
3006 SOC_FLAG("AM433", SOC_AM3),
3007 SOC_FLAG("AM43*", SOC_AM4),
3008 SOC_FLAG("DRA7*", SOC_DRA7),
3009
3010 { /* sentinel */ }
3011 };
3012
3013 /*
3014 * List of SoCs variants with disabled features. By default we assume all
3015 * devices in the device tree are available so no need to list those SoCs.
3016 */
3017 static const struct soc_device_attribute sysc_soc_feat_match[] = {
3018 /* OMAP3430/3530 and AM3517 variants with some accelerators disabled */
3019 SOC_FLAG("AM3505", DIS_SGX),
3020 SOC_FLAG("OMAP3525", DIS_SGX),
3021 SOC_FLAG("OMAP3515", DIS_IVA | DIS_SGX),
3022 SOC_FLAG("OMAP3503", DIS_ISP | DIS_IVA | DIS_SGX),
3023
3024 /* OMAP3630/DM3730 variants with some accelerators disabled */
3025 SOC_FLAG("AM3703", DIS_IVA | DIS_SGX),
3026 SOC_FLAG("DM3725", DIS_SGX),
3027 SOC_FLAG("OMAP3611", DIS_ISP | DIS_IVA | DIS_SGX),
3028 SOC_FLAG("OMAP3615/AM3715", DIS_IVA),
3029 SOC_FLAG("OMAP3621", DIS_ISP),
3030
3031 { /* sentinel */ }
3032 };
3033
sysc_add_disabled(unsigned long base)3034 static int sysc_add_disabled(unsigned long base)
3035 {
3036 struct sysc_address *disabled_module;
3037
3038 disabled_module = kzalloc(sizeof(*disabled_module), GFP_KERNEL);
3039 if (!disabled_module)
3040 return -ENOMEM;
3041
3042 disabled_module->base = base;
3043
3044 mutex_lock(&sysc_soc->list_lock);
3045 list_add(&disabled_module->node, &sysc_soc->disabled_modules);
3046 mutex_unlock(&sysc_soc->list_lock);
3047
3048 return 0;
3049 }
3050
3051 /*
3052 * One time init to detect the booted SoC, disable unavailable features
3053 * and initialize list for optional cpu_pm notifier.
3054 *
3055 * Note that we initialize static data shared across all ti-sysc instances
3056 * so ddata is only used for SoC type. This can be called from module_init
3057 * once we no longer need to rely on platform data.
3058 */
sysc_init_static_data(struct sysc * ddata)3059 static int sysc_init_static_data(struct sysc *ddata)
3060 {
3061 const struct soc_device_attribute *match;
3062 struct ti_sysc_platform_data *pdata;
3063 unsigned long features = 0;
3064 struct device_node *np;
3065
3066 if (sysc_soc)
3067 return 0;
3068
3069 sysc_soc = kzalloc(sizeof(*sysc_soc), GFP_KERNEL);
3070 if (!sysc_soc)
3071 return -ENOMEM;
3072
3073 mutex_init(&sysc_soc->list_lock);
3074 INIT_LIST_HEAD(&sysc_soc->disabled_modules);
3075 INIT_LIST_HEAD(&sysc_soc->restored_modules);
3076 sysc_soc->general_purpose = true;
3077
3078 pdata = dev_get_platdata(ddata->dev);
3079 if (pdata && pdata->soc_type_gp)
3080 sysc_soc->general_purpose = pdata->soc_type_gp();
3081
3082 match = soc_device_match(sysc_soc_match);
3083 if (match && match->data)
3084 sysc_soc->soc = (enum sysc_soc)(uintptr_t)match->data;
3085
3086 /*
3087 * Check and warn about possible old incomplete dtb. We now want to see
3088 * simple-pm-bus instead of simple-bus in the dtb for genpd using SoCs.
3089 */
3090 switch (sysc_soc->soc) {
3091 case SOC_AM3:
3092 case SOC_AM4:
3093 case SOC_4430 ... SOC_4470:
3094 case SOC_5430:
3095 case SOC_DRA7:
3096 np = of_find_node_by_path("/ocp");
3097 WARN_ONCE(np && of_device_is_compatible(np, "simple-bus"),
3098 "ti-sysc: Incomplete old dtb, please update\n");
3099 break;
3100 default:
3101 break;
3102 }
3103
3104 /* Ignore devices that are not available on HS and EMU SoCs */
3105 if (!sysc_soc->general_purpose) {
3106 switch (sysc_soc->soc) {
3107 case SOC_3430 ... SOC_3630:
3108 sysc_add_disabled(0x48304000); /* timer12 */
3109 break;
3110 case SOC_AM3:
3111 sysc_add_disabled(0x48310000); /* rng */
3112 break;
3113 default:
3114 break;
3115 }
3116 }
3117
3118 match = soc_device_match(sysc_soc_feat_match);
3119 if (!match)
3120 return 0;
3121
3122 if (match->data)
3123 features = (unsigned long)match->data;
3124
3125 /*
3126 * Add disabled devices to the list based on the module base.
3127 * Note that this must be done before we attempt to access the
3128 * device and have module revision checks working.
3129 */
3130 if (features & DIS_ISP)
3131 sysc_add_disabled(0x480bd400);
3132 if (features & DIS_IVA)
3133 sysc_add_disabled(0x5d000000);
3134 if (features & DIS_SGX)
3135 sysc_add_disabled(0x50000000);
3136
3137 return 0;
3138 }
3139
sysc_cleanup_static_data(void)3140 static void sysc_cleanup_static_data(void)
3141 {
3142 struct sysc_module *restored_module;
3143 struct sysc_address *disabled_module;
3144 struct list_head *pos, *tmp;
3145
3146 if (!sysc_soc)
3147 return;
3148
3149 if (sysc_soc->nb.notifier_call)
3150 cpu_pm_unregister_notifier(&sysc_soc->nb);
3151
3152 mutex_lock(&sysc_soc->list_lock);
3153 list_for_each_safe(pos, tmp, &sysc_soc->restored_modules) {
3154 restored_module = list_entry(pos, struct sysc_module, node);
3155 list_del(pos);
3156 kfree(restored_module);
3157 }
3158 list_for_each_safe(pos, tmp, &sysc_soc->disabled_modules) {
3159 disabled_module = list_entry(pos, struct sysc_address, node);
3160 list_del(pos);
3161 kfree(disabled_module);
3162 }
3163 mutex_unlock(&sysc_soc->list_lock);
3164 }
3165
sysc_check_disabled_devices(struct sysc * ddata)3166 static int sysc_check_disabled_devices(struct sysc *ddata)
3167 {
3168 struct sysc_address *disabled_module;
3169 int error = 0;
3170
3171 mutex_lock(&sysc_soc->list_lock);
3172 list_for_each_entry(disabled_module, &sysc_soc->disabled_modules, node) {
3173 if (ddata->module_pa == disabled_module->base) {
3174 dev_dbg(ddata->dev, "module disabled for this SoC\n");
3175 error = -ENODEV;
3176 break;
3177 }
3178 }
3179 mutex_unlock(&sysc_soc->list_lock);
3180
3181 return error;
3182 }
3183
3184 /*
3185 * Ignore timers tagged with no-reset and no-idle. These are likely in use,
3186 * for example by drivers/clocksource/timer-ti-dm-systimer.c. If more checks
3187 * are needed, we could also look at the timer register configuration.
3188 */
sysc_check_active_timer(struct sysc * ddata)3189 static int sysc_check_active_timer(struct sysc *ddata)
3190 {
3191 int error;
3192
3193 if (ddata->cap->type != TI_SYSC_OMAP2_TIMER &&
3194 ddata->cap->type != TI_SYSC_OMAP4_TIMER)
3195 return 0;
3196
3197 /*
3198 * Quirk for omap3 beagleboard revision A to B4 to use gpt12.
3199 * Revision C and later are fixed with commit 23885389dbbb ("ARM:
3200 * dts: Fix timer regression for beagleboard revision c"). This all
3201 * can be dropped if we stop supporting old beagleboard revisions
3202 * A to B4 at some point.
3203 */
3204 if (sysc_soc->soc == SOC_3430 || sysc_soc->soc == SOC_AM35)
3205 error = -ENXIO;
3206 else
3207 error = -EBUSY;
3208
3209 if ((ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT) &&
3210 (ddata->cfg.quirks & SYSC_QUIRK_NO_IDLE))
3211 return error;
3212
3213 return 0;
3214 }
3215
3216 static const struct of_device_id sysc_match_table[] = {
3217 { .compatible = "simple-bus", },
3218 { /* sentinel */ },
3219 };
3220
sysc_probe(struct platform_device * pdev)3221 static int sysc_probe(struct platform_device *pdev)
3222 {
3223 struct ti_sysc_platform_data *pdata = dev_get_platdata(&pdev->dev);
3224 struct sysc *ddata;
3225 int error;
3226
3227 ddata = devm_kzalloc(&pdev->dev, sizeof(*ddata), GFP_KERNEL);
3228 if (!ddata)
3229 return -ENOMEM;
3230
3231 ddata->offsets[SYSC_REVISION] = -ENODEV;
3232 ddata->offsets[SYSC_SYSCONFIG] = -ENODEV;
3233 ddata->offsets[SYSC_SYSSTATUS] = -ENODEV;
3234 ddata->dev = &pdev->dev;
3235 platform_set_drvdata(pdev, ddata);
3236
3237 error = sysc_init_static_data(ddata);
3238 if (error)
3239 return error;
3240
3241 error = sysc_init_match(ddata);
3242 if (error)
3243 return error;
3244
3245 error = sysc_init_dts_quirks(ddata);
3246 if (error)
3247 return error;
3248
3249 error = sysc_map_and_check_registers(ddata);
3250 if (error)
3251 return error;
3252
3253 error = sysc_init_sysc_mask(ddata);
3254 if (error)
3255 return error;
3256
3257 error = sysc_init_idlemodes(ddata);
3258 if (error)
3259 return error;
3260
3261 error = sysc_init_syss_mask(ddata);
3262 if (error)
3263 return error;
3264
3265 error = sysc_init_pdata(ddata);
3266 if (error)
3267 return error;
3268
3269 sysc_init_early_quirks(ddata);
3270
3271 error = sysc_check_disabled_devices(ddata);
3272 if (error)
3273 return error;
3274
3275 error = sysc_check_active_timer(ddata);
3276 if (error == -ENXIO)
3277 ddata->reserved = true;
3278 else if (error)
3279 return error;
3280
3281 error = sysc_get_clocks(ddata);
3282 if (error)
3283 return error;
3284
3285 error = sysc_init_resets(ddata);
3286 if (error)
3287 goto unprepare;
3288
3289 error = sysc_init_module(ddata);
3290 if (error)
3291 goto unprepare;
3292
3293 pm_runtime_enable(ddata->dev);
3294 error = pm_runtime_resume_and_get(ddata->dev);
3295 if (error < 0) {
3296 pm_runtime_disable(ddata->dev);
3297 goto unprepare;
3298 }
3299
3300 /* Balance use counts as PM runtime should have enabled these all */
3301 if (!(ddata->cfg.quirks &
3302 (SYSC_QUIRK_NO_IDLE | SYSC_QUIRK_NO_IDLE_ON_INIT))) {
3303 sysc_disable_main_clocks(ddata);
3304 sysc_disable_opt_clocks(ddata);
3305 sysc_clkdm_allow_idle(ddata);
3306 }
3307
3308 if (!(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT))
3309 reset_control_assert(ddata->rsts);
3310
3311 sysc_show_registers(ddata);
3312
3313 ddata->dev->type = &sysc_device_type;
3314
3315 if (!ddata->reserved) {
3316 error = of_platform_populate(ddata->dev->of_node,
3317 sysc_match_table,
3318 pdata ? pdata->auxdata : NULL,
3319 ddata->dev);
3320 if (error)
3321 goto err;
3322 }
3323
3324 INIT_DELAYED_WORK(&ddata->idle_work, ti_sysc_idle);
3325
3326 /* At least earlycon won't survive without deferred idle */
3327 if (ddata->cfg.quirks & (SYSC_QUIRK_NO_IDLE |
3328 SYSC_QUIRK_NO_IDLE_ON_INIT |
3329 SYSC_QUIRK_NO_RESET_ON_INIT)) {
3330 schedule_delayed_work(&ddata->idle_work, 3000);
3331 } else {
3332 pm_runtime_put(&pdev->dev);
3333 }
3334
3335 if (ddata->cfg.quirks & SYSC_QUIRK_REINIT_ON_CTX_LOST)
3336 sysc_add_restored(ddata);
3337
3338 return 0;
3339
3340 err:
3341 pm_runtime_put_sync(&pdev->dev);
3342 pm_runtime_disable(&pdev->dev);
3343 unprepare:
3344 sysc_unprepare(ddata);
3345
3346 return error;
3347 }
3348
sysc_remove(struct platform_device * pdev)3349 static int sysc_remove(struct platform_device *pdev)
3350 {
3351 struct sysc *ddata = platform_get_drvdata(pdev);
3352 int error;
3353
3354 /* Device can still be enabled, see deferred idle quirk in probe */
3355 if (cancel_delayed_work_sync(&ddata->idle_work))
3356 ti_sysc_idle(&ddata->idle_work.work);
3357
3358 error = pm_runtime_resume_and_get(ddata->dev);
3359 if (error < 0) {
3360 pm_runtime_disable(ddata->dev);
3361 goto unprepare;
3362 }
3363
3364 of_platform_depopulate(&pdev->dev);
3365
3366 pm_runtime_put_sync(&pdev->dev);
3367 pm_runtime_disable(&pdev->dev);
3368
3369 if (!reset_control_status(ddata->rsts))
3370 reset_control_assert(ddata->rsts);
3371
3372 unprepare:
3373 sysc_unprepare(ddata);
3374
3375 return 0;
3376 }
3377
3378 static const struct of_device_id sysc_match[] = {
3379 { .compatible = "ti,sysc-omap2", .data = &sysc_omap2, },
3380 { .compatible = "ti,sysc-omap2-timer", .data = &sysc_omap2_timer, },
3381 { .compatible = "ti,sysc-omap4", .data = &sysc_omap4, },
3382 { .compatible = "ti,sysc-omap4-timer", .data = &sysc_omap4_timer, },
3383 { .compatible = "ti,sysc-omap4-simple", .data = &sysc_omap4_simple, },
3384 { .compatible = "ti,sysc-omap3430-sr", .data = &sysc_34xx_sr, },
3385 { .compatible = "ti,sysc-omap3630-sr", .data = &sysc_36xx_sr, },
3386 { .compatible = "ti,sysc-omap4-sr", .data = &sysc_omap4_sr, },
3387 { .compatible = "ti,sysc-omap3-sham", .data = &sysc_omap3_sham, },
3388 { .compatible = "ti,sysc-omap-aes", .data = &sysc_omap3_aes, },
3389 { .compatible = "ti,sysc-mcasp", .data = &sysc_omap4_mcasp, },
3390 { .compatible = "ti,sysc-dra7-mcasp", .data = &sysc_dra7_mcasp, },
3391 { .compatible = "ti,sysc-usb-host-fs",
3392 .data = &sysc_omap4_usb_host_fs, },
3393 { .compatible = "ti,sysc-dra7-mcan", .data = &sysc_dra7_mcan, },
3394 { .compatible = "ti,sysc-pruss", .data = &sysc_pruss, },
3395 { },
3396 };
3397 MODULE_DEVICE_TABLE(of, sysc_match);
3398
3399 static struct platform_driver sysc_driver = {
3400 .probe = sysc_probe,
3401 .remove = sysc_remove,
3402 .driver = {
3403 .name = "ti-sysc",
3404 .of_match_table = sysc_match,
3405 .pm = &sysc_pm_ops,
3406 },
3407 };
3408
sysc_init(void)3409 static int __init sysc_init(void)
3410 {
3411 bus_register_notifier(&platform_bus_type, &sysc_nb);
3412
3413 return platform_driver_register(&sysc_driver);
3414 }
3415 module_init(sysc_init);
3416
sysc_exit(void)3417 static void __exit sysc_exit(void)
3418 {
3419 bus_unregister_notifier(&platform_bus_type, &sysc_nb);
3420 platform_driver_unregister(&sysc_driver);
3421 sysc_cleanup_static_data();
3422 }
3423 module_exit(sysc_exit);
3424
3425 MODULE_DESCRIPTION("TI sysc interconnect target driver");
3426 MODULE_LICENSE("GPL v2");
3427