1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Ingenic SoCs TCU IRQ driver
4  * Copyright (C) 2019 Paul Cercueil <paul@crapouillou.net>
5  * Copyright (C) 2020 周琰杰 (Zhou Yanjie) <zhouyanjie@wanyeetech.com>
6  */
7 
8 #include <linux/bitops.h>
9 #include <linux/clk.h>
10 #include <linux/clockchips.h>
11 #include <linux/clocksource.h>
12 #include <linux/interrupt.h>
13 #include <linux/mfd/ingenic-tcu.h>
14 #include <linux/mfd/syscon.h>
15 #include <linux/of.h>
16 #include <linux/of_address.h>
17 #include <linux/of_irq.h>
18 #include <linux/of_platform.h>
19 #include <linux/overflow.h>
20 #include <linux/platform_device.h>
21 #include <linux/regmap.h>
22 #include <linux/sched_clock.h>
23 
24 #include <dt-bindings/clock/ingenic,tcu.h>
25 
26 static DEFINE_PER_CPU(call_single_data_t, ingenic_cevt_csd);
27 
28 struct ingenic_soc_info {
29 	unsigned int num_channels;
30 };
31 
32 struct ingenic_tcu_timer {
33 	unsigned int cpu;
34 	unsigned int channel;
35 	struct clock_event_device cevt;
36 	struct clk *clk;
37 	char name[8];
38 };
39 
40 struct ingenic_tcu {
41 	struct regmap *map;
42 	struct device_node *np;
43 	struct clk *cs_clk;
44 	unsigned int cs_channel;
45 	struct clocksource cs;
46 	unsigned long pwm_channels_mask;
47 	struct ingenic_tcu_timer timers[];
48 };
49 
50 static struct ingenic_tcu *ingenic_tcu;
51 
52 static u64 notrace ingenic_tcu_timer_read(void)
53 {
54 	struct ingenic_tcu *tcu = ingenic_tcu;
55 	unsigned int count;
56 
57 	regmap_read(tcu->map, TCU_REG_TCNTc(tcu->cs_channel), &count);
58 
59 	return count;
60 }
61 
62 static u64 notrace ingenic_tcu_timer_cs_read(struct clocksource *cs)
63 {
64 	return ingenic_tcu_timer_read();
65 }
66 
67 static inline struct ingenic_tcu *
68 to_ingenic_tcu(struct ingenic_tcu_timer *timer)
69 {
70 	return container_of(timer, struct ingenic_tcu, timers[timer->cpu]);
71 }
72 
73 static inline struct ingenic_tcu_timer *
74 to_ingenic_tcu_timer(struct clock_event_device *evt)
75 {
76 	return container_of(evt, struct ingenic_tcu_timer, cevt);
77 }
78 
79 static int ingenic_tcu_cevt_set_state_shutdown(struct clock_event_device *evt)
80 {
81 	struct ingenic_tcu_timer *timer = to_ingenic_tcu_timer(evt);
82 	struct ingenic_tcu *tcu = to_ingenic_tcu(timer);
83 
84 	regmap_write(tcu->map, TCU_REG_TECR, BIT(timer->channel));
85 
86 	return 0;
87 }
88 
89 static int ingenic_tcu_cevt_set_next(unsigned long next,
90 				     struct clock_event_device *evt)
91 {
92 	struct ingenic_tcu_timer *timer = to_ingenic_tcu_timer(evt);
93 	struct ingenic_tcu *tcu = to_ingenic_tcu(timer);
94 
95 	if (next > 0xffff)
96 		return -EINVAL;
97 
98 	regmap_write(tcu->map, TCU_REG_TDFRc(timer->channel), next);
99 	regmap_write(tcu->map, TCU_REG_TCNTc(timer->channel), 0);
100 	regmap_write(tcu->map, TCU_REG_TESR, BIT(timer->channel));
101 
102 	return 0;
103 }
104 
105 static void ingenic_per_cpu_event_handler(void *info)
106 {
107 	struct clock_event_device *cevt = (struct clock_event_device *) info;
108 
109 	cevt->event_handler(cevt);
110 }
111 
112 static irqreturn_t ingenic_tcu_cevt_cb(int irq, void *dev_id)
113 {
114 	struct ingenic_tcu_timer *timer = dev_id;
115 	struct ingenic_tcu *tcu = to_ingenic_tcu(timer);
116 	call_single_data_t *csd;
117 
118 	regmap_write(tcu->map, TCU_REG_TECR, BIT(timer->channel));
119 
120 	if (timer->cevt.event_handler) {
121 		csd = &per_cpu(ingenic_cevt_csd, timer->cpu);
122 		csd->info = (void *) &timer->cevt;
123 		csd->func = ingenic_per_cpu_event_handler;
124 		smp_call_function_single_async(timer->cpu, csd);
125 	}
126 
127 	return IRQ_HANDLED;
128 }
129 
130 static struct clk *ingenic_tcu_get_clock(struct device_node *np, int id)
131 {
132 	struct of_phandle_args args;
133 
134 	args.np = np;
135 	args.args_count = 1;
136 	args.args[0] = id;
137 
138 	return of_clk_get_from_provider(&args);
139 }
140 
141 static int ingenic_tcu_setup_cevt(unsigned int cpu)
142 {
143 	struct ingenic_tcu *tcu = ingenic_tcu;
144 	struct ingenic_tcu_timer *timer = &tcu->timers[cpu];
145 	unsigned int timer_virq;
146 	struct irq_domain *domain;
147 	unsigned long rate;
148 	int err;
149 
150 	timer->clk = ingenic_tcu_get_clock(tcu->np, timer->channel);
151 	if (IS_ERR(timer->clk))
152 		return PTR_ERR(timer->clk);
153 
154 	err = clk_prepare_enable(timer->clk);
155 	if (err)
156 		goto err_clk_put;
157 
158 	rate = clk_get_rate(timer->clk);
159 	if (!rate) {
160 		err = -EINVAL;
161 		goto err_clk_disable;
162 	}
163 
164 	domain = irq_find_host(tcu->np);
165 	if (!domain) {
166 		err = -ENODEV;
167 		goto err_clk_disable;
168 	}
169 
170 	timer_virq = irq_create_mapping(domain, timer->channel);
171 	if (!timer_virq) {
172 		err = -EINVAL;
173 		goto err_clk_disable;
174 	}
175 
176 	snprintf(timer->name, sizeof(timer->name), "TCU%u", timer->channel);
177 
178 	err = request_irq(timer_virq, ingenic_tcu_cevt_cb, IRQF_TIMER,
179 			  timer->name, timer);
180 	if (err)
181 		goto err_irq_dispose_mapping;
182 
183 	timer->cpu = smp_processor_id();
184 	timer->cevt.cpumask = cpumask_of(smp_processor_id());
185 	timer->cevt.features = CLOCK_EVT_FEAT_ONESHOT;
186 	timer->cevt.name = timer->name;
187 	timer->cevt.rating = 200;
188 	timer->cevt.set_state_shutdown = ingenic_tcu_cevt_set_state_shutdown;
189 	timer->cevt.set_next_event = ingenic_tcu_cevt_set_next;
190 
191 	clockevents_config_and_register(&timer->cevt, rate, 10, 0xffff);
192 
193 	return 0;
194 
195 err_irq_dispose_mapping:
196 	irq_dispose_mapping(timer_virq);
197 err_clk_disable:
198 	clk_disable_unprepare(timer->clk);
199 err_clk_put:
200 	clk_put(timer->clk);
201 	return err;
202 }
203 
204 static int __init ingenic_tcu_clocksource_init(struct device_node *np,
205 					       struct ingenic_tcu *tcu)
206 {
207 	unsigned int channel = tcu->cs_channel;
208 	struct clocksource *cs = &tcu->cs;
209 	unsigned long rate;
210 	int err;
211 
212 	tcu->cs_clk = ingenic_tcu_get_clock(np, channel);
213 	if (IS_ERR(tcu->cs_clk))
214 		return PTR_ERR(tcu->cs_clk);
215 
216 	err = clk_prepare_enable(tcu->cs_clk);
217 	if (err)
218 		goto err_clk_put;
219 
220 	rate = clk_get_rate(tcu->cs_clk);
221 	if (!rate) {
222 		err = -EINVAL;
223 		goto err_clk_disable;
224 	}
225 
226 	/* Reset channel */
227 	regmap_update_bits(tcu->map, TCU_REG_TCSRc(channel),
228 			   0xffff & ~TCU_TCSR_RESERVED_BITS, 0);
229 
230 	/* Reset counter */
231 	regmap_write(tcu->map, TCU_REG_TDFRc(channel), 0xffff);
232 	regmap_write(tcu->map, TCU_REG_TCNTc(channel), 0);
233 
234 	/* Enable channel */
235 	regmap_write(tcu->map, TCU_REG_TESR, BIT(channel));
236 
237 	cs->name = "ingenic-timer";
238 	cs->rating = 200;
239 	cs->flags = CLOCK_SOURCE_IS_CONTINUOUS;
240 	cs->mask = CLOCKSOURCE_MASK(16);
241 	cs->read = ingenic_tcu_timer_cs_read;
242 
243 	err = clocksource_register_hz(cs, rate);
244 	if (err)
245 		goto err_clk_disable;
246 
247 	return 0;
248 
249 err_clk_disable:
250 	clk_disable_unprepare(tcu->cs_clk);
251 err_clk_put:
252 	clk_put(tcu->cs_clk);
253 	return err;
254 }
255 
256 static const struct ingenic_soc_info jz4740_soc_info = {
257 	.num_channels = 8,
258 };
259 
260 static const struct ingenic_soc_info jz4725b_soc_info = {
261 	.num_channels = 6,
262 };
263 
264 static const struct of_device_id ingenic_tcu_of_match[] = {
265 	{ .compatible = "ingenic,jz4740-tcu", .data = &jz4740_soc_info, },
266 	{ .compatible = "ingenic,jz4725b-tcu", .data = &jz4725b_soc_info, },
267 	{ .compatible = "ingenic,jz4770-tcu", .data = &jz4740_soc_info, },
268 	{ .compatible = "ingenic,x1000-tcu", .data = &jz4740_soc_info, },
269 	{ /* sentinel */ }
270 };
271 
272 static int __init ingenic_tcu_init(struct device_node *np)
273 {
274 	const struct of_device_id *id = of_match_node(ingenic_tcu_of_match, np);
275 	const struct ingenic_soc_info *soc_info = id->data;
276 	struct ingenic_tcu_timer *timer;
277 	struct ingenic_tcu *tcu;
278 	struct regmap *map;
279 	unsigned int cpu;
280 	int ret, last_bit = -1;
281 	long rate;
282 
283 	of_node_clear_flag(np, OF_POPULATED);
284 
285 	map = device_node_to_regmap(np);
286 	if (IS_ERR(map))
287 		return PTR_ERR(map);
288 
289 	tcu = kzalloc(struct_size(tcu, timers, num_possible_cpus()),
290 		      GFP_KERNEL);
291 	if (!tcu)
292 		return -ENOMEM;
293 
294 	/*
295 	 * Enable all TCU channels for PWM use by default except channels 0/1,
296 	 * and channel 2 if target CPU is JZ4780/X2000 and SMP is selected.
297 	 */
298 	tcu->pwm_channels_mask = GENMASK(soc_info->num_channels - 1,
299 					 num_possible_cpus() + 1);
300 	of_property_read_u32(np, "ingenic,pwm-channels-mask",
301 			     (u32 *)&tcu->pwm_channels_mask);
302 
303 	/* Verify that we have at least num_possible_cpus() + 1 free channels */
304 	if (hweight8(tcu->pwm_channels_mask) >
305 			soc_info->num_channels - num_possible_cpus() + 1) {
306 		pr_crit("%s: Invalid PWM channel mask: 0x%02lx\n", __func__,
307 			tcu->pwm_channels_mask);
308 		ret = -EINVAL;
309 		goto err_free_ingenic_tcu;
310 	}
311 
312 	tcu->map = map;
313 	tcu->np = np;
314 	ingenic_tcu = tcu;
315 
316 	for (cpu = 0; cpu < num_possible_cpus(); cpu++) {
317 		timer = &tcu->timers[cpu];
318 
319 		timer->cpu = cpu;
320 		timer->channel = find_next_zero_bit(&tcu->pwm_channels_mask,
321 						  soc_info->num_channels,
322 						  last_bit + 1);
323 		last_bit = timer->channel;
324 	}
325 
326 	tcu->cs_channel = find_next_zero_bit(&tcu->pwm_channels_mask,
327 					     soc_info->num_channels,
328 					     last_bit + 1);
329 
330 	ret = ingenic_tcu_clocksource_init(np, tcu);
331 	if (ret) {
332 		pr_crit("%s: Unable to init clocksource: %d\n", __func__, ret);
333 		goto err_free_ingenic_tcu;
334 	}
335 
336 	/* Setup clock events on each CPU core */
337 	ret = cpuhp_setup_state(CPUHP_AP_ONLINE_DYN, "Ingenic XBurst: online",
338 				ingenic_tcu_setup_cevt, NULL);
339 	if (ret < 0) {
340 		pr_crit("%s: Unable to start CPU timers: %d\n", __func__, ret);
341 		goto err_tcu_clocksource_cleanup;
342 	}
343 
344 	/* Register the sched_clock at the end as there's no way to undo it */
345 	rate = clk_get_rate(tcu->cs_clk);
346 	sched_clock_register(ingenic_tcu_timer_read, 16, rate);
347 
348 	return 0;
349 
350 err_tcu_clocksource_cleanup:
351 	clocksource_unregister(&tcu->cs);
352 	clk_disable_unprepare(tcu->cs_clk);
353 	clk_put(tcu->cs_clk);
354 err_free_ingenic_tcu:
355 	kfree(tcu);
356 	return ret;
357 }
358 
359 TIMER_OF_DECLARE(jz4740_tcu_intc,  "ingenic,jz4740-tcu",  ingenic_tcu_init);
360 TIMER_OF_DECLARE(jz4725b_tcu_intc, "ingenic,jz4725b-tcu", ingenic_tcu_init);
361 TIMER_OF_DECLARE(jz4770_tcu_intc,  "ingenic,jz4770-tcu",  ingenic_tcu_init);
362 TIMER_OF_DECLARE(x1000_tcu_intc,  "ingenic,x1000-tcu",  ingenic_tcu_init);
363 
364 static int __init ingenic_tcu_probe(struct platform_device *pdev)
365 {
366 	platform_set_drvdata(pdev, ingenic_tcu);
367 
368 	return 0;
369 }
370 
371 static int __maybe_unused ingenic_tcu_suspend(struct device *dev)
372 {
373 	struct ingenic_tcu *tcu = dev_get_drvdata(dev);
374 	unsigned int cpu;
375 
376 	clk_disable(tcu->cs_clk);
377 
378 	for (cpu = 0; cpu < num_online_cpus(); cpu++)
379 		clk_disable(tcu->timers[cpu].clk);
380 
381 	return 0;
382 }
383 
384 static int __maybe_unused ingenic_tcu_resume(struct device *dev)
385 {
386 	struct ingenic_tcu *tcu = dev_get_drvdata(dev);
387 	unsigned int cpu;
388 	int ret;
389 
390 	for (cpu = 0; cpu < num_online_cpus(); cpu++) {
391 		ret = clk_enable(tcu->timers[cpu].clk);
392 		if (ret)
393 			goto err_timer_clk_disable;
394 	}
395 
396 	ret = clk_enable(tcu->cs_clk);
397 	if (ret)
398 		goto err_timer_clk_disable;
399 
400 	return 0;
401 
402 err_timer_clk_disable:
403 	for (; cpu > 0; cpu--)
404 		clk_disable(tcu->timers[cpu - 1].clk);
405 	return ret;
406 }
407 
408 static const struct dev_pm_ops __maybe_unused ingenic_tcu_pm_ops = {
409 	/* _noirq: We want the TCU clocks to be gated last / ungated first */
410 	.suspend_noirq = ingenic_tcu_suspend,
411 	.resume_noirq  = ingenic_tcu_resume,
412 };
413 
414 static struct platform_driver ingenic_tcu_driver = {
415 	.driver = {
416 		.name	= "ingenic-tcu-timer",
417 #ifdef CONFIG_PM_SLEEP
418 		.pm	= &ingenic_tcu_pm_ops,
419 #endif
420 		.of_match_table = ingenic_tcu_of_match,
421 	},
422 };
423 builtin_platform_driver_probe(ingenic_tcu_driver, ingenic_tcu_probe);
424