xref: /openbmc/linux/drivers/clk/renesas/rzg2l-cpg.c (revision a5961bed)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * RZ/G2L Clock Pulse Generator
4  *
5  * Copyright (C) 2021 Renesas Electronics Corp.
6  *
7  * Based on renesas-cpg-mssr.c
8  *
9  * Copyright (C) 2015 Glider bvba
10  * Copyright (C) 2013 Ideas On Board SPRL
11  * Copyright (C) 2015 Renesas Electronics Corp.
12  */
13 
14 #include <linux/clk.h>
15 #include <linux/clk-provider.h>
16 #include <linux/clk/renesas.h>
17 #include <linux/delay.h>
18 #include <linux/device.h>
19 #include <linux/init.h>
20 #include <linux/iopoll.h>
21 #include <linux/mod_devicetable.h>
22 #include <linux/module.h>
23 #include <linux/of_address.h>
24 #include <linux/of_device.h>
25 #include <linux/platform_device.h>
26 #include <linux/pm_clock.h>
27 #include <linux/pm_domain.h>
28 #include <linux/reset-controller.h>
29 #include <linux/slab.h>
30 #include <linux/units.h>
31 
32 #include <dt-bindings/clock/renesas-cpg-mssr.h>
33 
34 #include "rzg2l-cpg.h"
35 
36 #ifdef DEBUG
37 #define WARN_DEBUG(x)	WARN_ON(x)
38 #else
39 #define WARN_DEBUG(x)	do { } while (0)
40 #endif
41 
42 #define DIV_RSMASK(v, s, m)	((v >> s) & m)
43 #define GET_SHIFT(val)		((val >> 12) & 0xff)
44 #define GET_WIDTH(val)		((val >> 8) & 0xf)
45 
46 #define KDIV(val)		DIV_RSMASK(val, 16, 0xffff)
47 #define MDIV(val)		DIV_RSMASK(val, 6, 0x3ff)
48 #define PDIV(val)		DIV_RSMASK(val, 0, 0x3f)
49 #define SDIV(val)		DIV_RSMASK(val, 0, 0x7)
50 
51 #define CLK_ON_R(reg)		(reg)
52 #define CLK_MON_R(reg)		(0x180 + (reg))
53 #define CLK_RST_R(reg)		(reg)
54 #define CLK_MRST_R(reg)		(0x180 + (reg))
55 
56 #define GET_REG_OFFSET(val)		((val >> 20) & 0xfff)
57 #define GET_REG_SAMPLL_CLK1(val)	((val >> 22) & 0xfff)
58 #define GET_REG_SAMPLL_CLK2(val)	((val >> 12) & 0xfff)
59 
60 #define MAX_VCLK_FREQ		(148500000)
61 
62 struct sd_hw_data {
63 	struct clk_hw hw;
64 	u32 conf;
65 	struct rzg2l_cpg_priv *priv;
66 };
67 
68 #define to_sd_hw_data(_hw)	container_of(_hw, struct sd_hw_data, hw)
69 
70 struct rzg2l_pll5_param {
71 	u32 pl5_fracin;
72 	u8 pl5_refdiv;
73 	u8 pl5_intin;
74 	u8 pl5_postdiv1;
75 	u8 pl5_postdiv2;
76 	u8 pl5_spread;
77 };
78 
79 struct rzg2l_pll5_mux_dsi_div_param {
80 	u8 clksrc;
81 	u8 dsi_div_a;
82 	u8 dsi_div_b;
83 };
84 
85 /**
86  * struct rzg2l_cpg_priv - Clock Pulse Generator Private Data
87  *
88  * @rcdev: Reset controller entity
89  * @dev: CPG device
90  * @base: CPG register block base address
91  * @rmw_lock: protects register accesses
92  * @clks: Array containing all Core and Module Clocks
93  * @num_core_clks: Number of Core Clocks in clks[]
94  * @num_mod_clks: Number of Module Clocks in clks[]
95  * @num_resets: Number of Module Resets in info->resets[]
96  * @last_dt_core_clk: ID of the last Core Clock exported to DT
97  * @info: Pointer to platform data
98  * @genpd: PM domain
99  * @mux_dsi_div_params: pll5 mux and dsi div parameters
100  */
101 struct rzg2l_cpg_priv {
102 	struct reset_controller_dev rcdev;
103 	struct device *dev;
104 	void __iomem *base;
105 	spinlock_t rmw_lock;
106 
107 	struct clk **clks;
108 	unsigned int num_core_clks;
109 	unsigned int num_mod_clks;
110 	unsigned int num_resets;
111 	unsigned int last_dt_core_clk;
112 
113 	const struct rzg2l_cpg_info *info;
114 
115 	struct generic_pm_domain genpd;
116 
117 	struct rzg2l_pll5_mux_dsi_div_param mux_dsi_div_params;
118 };
119 
120 static void rzg2l_cpg_del_clk_provider(void *data)
121 {
122 	of_clk_del_provider(data);
123 }
124 
125 static struct clk * __init
126 rzg2l_cpg_div_clk_register(const struct cpg_core_clk *core,
127 			   struct clk **clks,
128 			   void __iomem *base,
129 			   struct rzg2l_cpg_priv *priv)
130 {
131 	struct device *dev = priv->dev;
132 	const struct clk *parent;
133 	const char *parent_name;
134 	struct clk_hw *clk_hw;
135 
136 	parent = clks[core->parent & 0xffff];
137 	if (IS_ERR(parent))
138 		return ERR_CAST(parent);
139 
140 	parent_name = __clk_get_name(parent);
141 
142 	if (core->dtable)
143 		clk_hw = clk_hw_register_divider_table(dev, core->name,
144 						       parent_name, 0,
145 						       base + GET_REG_OFFSET(core->conf),
146 						       GET_SHIFT(core->conf),
147 						       GET_WIDTH(core->conf),
148 						       core->flag,
149 						       core->dtable,
150 						       &priv->rmw_lock);
151 	else
152 		clk_hw = clk_hw_register_divider(dev, core->name,
153 						 parent_name, 0,
154 						 base + GET_REG_OFFSET(core->conf),
155 						 GET_SHIFT(core->conf),
156 						 GET_WIDTH(core->conf),
157 						 core->flag, &priv->rmw_lock);
158 
159 	if (IS_ERR(clk_hw))
160 		return ERR_CAST(clk_hw);
161 
162 	return clk_hw->clk;
163 }
164 
165 static struct clk * __init
166 rzg2l_cpg_mux_clk_register(const struct cpg_core_clk *core,
167 			   void __iomem *base,
168 			   struct rzg2l_cpg_priv *priv)
169 {
170 	const struct clk_hw *clk_hw;
171 
172 	clk_hw = devm_clk_hw_register_mux(priv->dev, core->name,
173 					  core->parent_names, core->num_parents,
174 					  core->flag,
175 					  base + GET_REG_OFFSET(core->conf),
176 					  GET_SHIFT(core->conf),
177 					  GET_WIDTH(core->conf),
178 					  core->mux_flags, &priv->rmw_lock);
179 	if (IS_ERR(clk_hw))
180 		return ERR_CAST(clk_hw);
181 
182 	return clk_hw->clk;
183 }
184 
185 static int rzg2l_cpg_sd_clk_mux_determine_rate(struct clk_hw *hw,
186 					       struct clk_rate_request *req)
187 {
188 	return clk_mux_determine_rate_flags(hw, req, CLK_MUX_ROUND_CLOSEST);
189 }
190 
191 static int rzg2l_cpg_sd_clk_mux_set_parent(struct clk_hw *hw, u8 index)
192 {
193 	struct sd_hw_data *hwdata = to_sd_hw_data(hw);
194 	struct rzg2l_cpg_priv *priv = hwdata->priv;
195 	u32 off = GET_REG_OFFSET(hwdata->conf);
196 	u32 shift = GET_SHIFT(hwdata->conf);
197 	const u32 clk_src_266 = 2;
198 	u32 bitmask;
199 
200 	/*
201 	 * As per the HW manual, we should not directly switch from 533 MHz to
202 	 * 400 MHz and vice versa. To change the setting from 2’b01 (533 MHz)
203 	 * to 2’b10 (400 MHz) or vice versa, Switch to 2’b11 (266 MHz) first,
204 	 * and then switch to the target setting (2’b01 (533 MHz) or 2’b10
205 	 * (400 MHz)).
206 	 * Setting a value of '0' to the SEL_SDHI0_SET or SEL_SDHI1_SET clock
207 	 * switching register is prohibited.
208 	 * The clock mux has 3 input clocks(533 MHz, 400 MHz, and 266 MHz), and
209 	 * the index to value mapping is done by adding 1 to the index.
210 	 */
211 	bitmask = (GENMASK(GET_WIDTH(hwdata->conf) - 1, 0) << shift) << 16;
212 	if (index != clk_src_266) {
213 		u32 msk, val;
214 		int ret;
215 
216 		writel(bitmask | ((clk_src_266 + 1) << shift), priv->base + off);
217 
218 		msk = off ? CPG_CLKSTATUS_SELSDHI1_STS : CPG_CLKSTATUS_SELSDHI0_STS;
219 
220 		ret = readl_poll_timeout(priv->base + CPG_CLKSTATUS, val,
221 					 !(val & msk), 100,
222 					 CPG_SDHI_CLK_SWITCH_STATUS_TIMEOUT_US);
223 		if (ret) {
224 			dev_err(priv->dev, "failed to switch clk source\n");
225 			return ret;
226 		}
227 	}
228 
229 	writel(bitmask | ((index + 1) << shift), priv->base + off);
230 
231 	return 0;
232 }
233 
234 static u8 rzg2l_cpg_sd_clk_mux_get_parent(struct clk_hw *hw)
235 {
236 	struct sd_hw_data *hwdata = to_sd_hw_data(hw);
237 	struct rzg2l_cpg_priv *priv = hwdata->priv;
238 	u32 val = readl(priv->base + GET_REG_OFFSET(hwdata->conf));
239 
240 	val >>= GET_SHIFT(hwdata->conf);
241 	val &= GENMASK(GET_WIDTH(hwdata->conf) - 1, 0);
242 	if (val) {
243 		val--;
244 	} else {
245 		/* Prohibited clk source, change it to 533 MHz(reset value) */
246 		rzg2l_cpg_sd_clk_mux_set_parent(hw, 0);
247 	}
248 
249 	return val;
250 }
251 
252 static const struct clk_ops rzg2l_cpg_sd_clk_mux_ops = {
253 	.determine_rate = rzg2l_cpg_sd_clk_mux_determine_rate,
254 	.set_parent	= rzg2l_cpg_sd_clk_mux_set_parent,
255 	.get_parent	= rzg2l_cpg_sd_clk_mux_get_parent,
256 };
257 
258 static struct clk * __init
259 rzg2l_cpg_sd_mux_clk_register(const struct cpg_core_clk *core,
260 			      void __iomem *base,
261 			      struct rzg2l_cpg_priv *priv)
262 {
263 	struct sd_hw_data *clk_hw_data;
264 	struct clk_init_data init;
265 	struct clk_hw *clk_hw;
266 	int ret;
267 
268 	clk_hw_data = devm_kzalloc(priv->dev, sizeof(*clk_hw_data), GFP_KERNEL);
269 	if (!clk_hw_data)
270 		return ERR_PTR(-ENOMEM);
271 
272 	clk_hw_data->priv = priv;
273 	clk_hw_data->conf = core->conf;
274 
275 	init.name = GET_SHIFT(core->conf) ? "sd1" : "sd0";
276 	init.ops = &rzg2l_cpg_sd_clk_mux_ops;
277 	init.flags = 0;
278 	init.num_parents = core->num_parents;
279 	init.parent_names = core->parent_names;
280 
281 	clk_hw = &clk_hw_data->hw;
282 	clk_hw->init = &init;
283 
284 	ret = devm_clk_hw_register(priv->dev, clk_hw);
285 	if (ret)
286 		return ERR_PTR(ret);
287 
288 	return clk_hw->clk;
289 }
290 
291 static unsigned long
292 rzg2l_cpg_get_foutpostdiv_rate(struct rzg2l_pll5_param *params,
293 			       unsigned long rate)
294 {
295 	unsigned long foutpostdiv_rate;
296 
297 	params->pl5_intin = rate / MEGA;
298 	params->pl5_fracin = div_u64(((u64)rate % MEGA) << 24, MEGA);
299 	params->pl5_refdiv = 2;
300 	params->pl5_postdiv1 = 1;
301 	params->pl5_postdiv2 = 1;
302 	params->pl5_spread = 0x16;
303 
304 	foutpostdiv_rate =
305 		EXTAL_FREQ_IN_MEGA_HZ * MEGA / params->pl5_refdiv *
306 		((((params->pl5_intin << 24) + params->pl5_fracin)) >> 24) /
307 		(params->pl5_postdiv1 * params->pl5_postdiv2);
308 
309 	return foutpostdiv_rate;
310 }
311 
312 struct dsi_div_hw_data {
313 	struct clk_hw hw;
314 	u32 conf;
315 	unsigned long rate;
316 	struct rzg2l_cpg_priv *priv;
317 };
318 
319 #define to_dsi_div_hw_data(_hw)	container_of(_hw, struct dsi_div_hw_data, hw)
320 
321 static unsigned long rzg2l_cpg_dsi_div_recalc_rate(struct clk_hw *hw,
322 						   unsigned long parent_rate)
323 {
324 	struct dsi_div_hw_data *dsi_div = to_dsi_div_hw_data(hw);
325 	unsigned long rate = dsi_div->rate;
326 
327 	if (!rate)
328 		rate = parent_rate;
329 
330 	return rate;
331 }
332 
333 static unsigned long rzg2l_cpg_get_vclk_parent_rate(struct clk_hw *hw,
334 						    unsigned long rate)
335 {
336 	struct dsi_div_hw_data *dsi_div = to_dsi_div_hw_data(hw);
337 	struct rzg2l_cpg_priv *priv = dsi_div->priv;
338 	struct rzg2l_pll5_param params;
339 	unsigned long parent_rate;
340 
341 	parent_rate = rzg2l_cpg_get_foutpostdiv_rate(&params, rate);
342 
343 	if (priv->mux_dsi_div_params.clksrc)
344 		parent_rate /= 2;
345 
346 	return parent_rate;
347 }
348 
349 static int rzg2l_cpg_dsi_div_determine_rate(struct clk_hw *hw,
350 					    struct clk_rate_request *req)
351 {
352 	if (req->rate > MAX_VCLK_FREQ)
353 		req->rate = MAX_VCLK_FREQ;
354 
355 	req->best_parent_rate = rzg2l_cpg_get_vclk_parent_rate(hw, req->rate);
356 
357 	return 0;
358 }
359 
360 static int rzg2l_cpg_dsi_div_set_rate(struct clk_hw *hw,
361 				      unsigned long rate,
362 				      unsigned long parent_rate)
363 {
364 	struct dsi_div_hw_data *dsi_div = to_dsi_div_hw_data(hw);
365 	struct rzg2l_cpg_priv *priv = dsi_div->priv;
366 
367 	/*
368 	 * MUX -->DIV_DSI_{A,B} -->M3 -->VCLK
369 	 *
370 	 * Based on the dot clock, the DSI divider clock sets the divider value,
371 	 * calculates the pll parameters for generating FOUTPOSTDIV and the clk
372 	 * source for the MUX and propagates that info to the parents.
373 	 */
374 
375 	if (!rate || rate > MAX_VCLK_FREQ)
376 		return -EINVAL;
377 
378 	dsi_div->rate = rate;
379 	writel(CPG_PL5_SDIV_DIV_DSI_A_WEN | CPG_PL5_SDIV_DIV_DSI_B_WEN |
380 	       (priv->mux_dsi_div_params.dsi_div_a << 0) |
381 	       (priv->mux_dsi_div_params.dsi_div_b << 8),
382 	       priv->base + CPG_PL5_SDIV);
383 
384 	return 0;
385 }
386 
387 static const struct clk_ops rzg2l_cpg_dsi_div_ops = {
388 	.recalc_rate = rzg2l_cpg_dsi_div_recalc_rate,
389 	.determine_rate = rzg2l_cpg_dsi_div_determine_rate,
390 	.set_rate = rzg2l_cpg_dsi_div_set_rate,
391 };
392 
393 static struct clk * __init
394 rzg2l_cpg_dsi_div_clk_register(const struct cpg_core_clk *core,
395 			       struct clk **clks,
396 			       struct rzg2l_cpg_priv *priv)
397 {
398 	struct dsi_div_hw_data *clk_hw_data;
399 	const struct clk *parent;
400 	const char *parent_name;
401 	struct clk_init_data init;
402 	struct clk_hw *clk_hw;
403 	int ret;
404 
405 	parent = clks[core->parent & 0xffff];
406 	if (IS_ERR(parent))
407 		return ERR_CAST(parent);
408 
409 	clk_hw_data = devm_kzalloc(priv->dev, sizeof(*clk_hw_data), GFP_KERNEL);
410 	if (!clk_hw_data)
411 		return ERR_PTR(-ENOMEM);
412 
413 	clk_hw_data->priv = priv;
414 
415 	parent_name = __clk_get_name(parent);
416 	init.name = core->name;
417 	init.ops = &rzg2l_cpg_dsi_div_ops;
418 	init.flags = CLK_SET_RATE_PARENT;
419 	init.parent_names = &parent_name;
420 	init.num_parents = 1;
421 
422 	clk_hw = &clk_hw_data->hw;
423 	clk_hw->init = &init;
424 
425 	ret = devm_clk_hw_register(priv->dev, clk_hw);
426 	if (ret)
427 		return ERR_PTR(ret);
428 
429 	return clk_hw->clk;
430 }
431 
432 struct pll5_mux_hw_data {
433 	struct clk_hw hw;
434 	u32 conf;
435 	unsigned long rate;
436 	struct rzg2l_cpg_priv *priv;
437 };
438 
439 #define to_pll5_mux_hw_data(_hw)	container_of(_hw, struct pll5_mux_hw_data, hw)
440 
441 static int rzg2l_cpg_pll5_4_clk_mux_determine_rate(struct clk_hw *hw,
442 						   struct clk_rate_request *req)
443 {
444 	struct clk_hw *parent;
445 	struct pll5_mux_hw_data *hwdata = to_pll5_mux_hw_data(hw);
446 	struct rzg2l_cpg_priv *priv = hwdata->priv;
447 
448 	parent = clk_hw_get_parent_by_index(hw, priv->mux_dsi_div_params.clksrc);
449 	req->best_parent_hw = parent;
450 	req->best_parent_rate = req->rate;
451 
452 	return 0;
453 }
454 
455 static int rzg2l_cpg_pll5_4_clk_mux_set_parent(struct clk_hw *hw, u8 index)
456 {
457 	struct pll5_mux_hw_data *hwdata = to_pll5_mux_hw_data(hw);
458 	struct rzg2l_cpg_priv *priv = hwdata->priv;
459 
460 	/*
461 	 * FOUTPOSTDIV--->|
462 	 *  |             | -->MUX -->DIV_DSIA_B -->M3 -->VCLK
463 	 *  |--FOUT1PH0-->|
464 	 *
465 	 * Based on the dot clock, the DSI divider clock calculates the parent
466 	 * rate and clk source for the MUX. It propagates that info to
467 	 * pll5_4_clk_mux which sets the clock source for DSI divider clock.
468 	 */
469 
470 	writel(CPG_OTHERFUNC1_REG_RES0_ON_WEN | index,
471 	       priv->base + CPG_OTHERFUNC1_REG);
472 
473 	return 0;
474 }
475 
476 static u8 rzg2l_cpg_pll5_4_clk_mux_get_parent(struct clk_hw *hw)
477 {
478 	struct pll5_mux_hw_data *hwdata = to_pll5_mux_hw_data(hw);
479 	struct rzg2l_cpg_priv *priv = hwdata->priv;
480 
481 	return readl(priv->base + GET_REG_OFFSET(hwdata->conf));
482 }
483 
484 static const struct clk_ops rzg2l_cpg_pll5_4_clk_mux_ops = {
485 	.determine_rate = rzg2l_cpg_pll5_4_clk_mux_determine_rate,
486 	.set_parent	= rzg2l_cpg_pll5_4_clk_mux_set_parent,
487 	.get_parent	= rzg2l_cpg_pll5_4_clk_mux_get_parent,
488 };
489 
490 static struct clk * __init
491 rzg2l_cpg_pll5_4_mux_clk_register(const struct cpg_core_clk *core,
492 				  struct rzg2l_cpg_priv *priv)
493 {
494 	struct pll5_mux_hw_data *clk_hw_data;
495 	struct clk_init_data init;
496 	struct clk_hw *clk_hw;
497 	int ret;
498 
499 	clk_hw_data = devm_kzalloc(priv->dev, sizeof(*clk_hw_data), GFP_KERNEL);
500 	if (!clk_hw_data)
501 		return ERR_PTR(-ENOMEM);
502 
503 	clk_hw_data->priv = priv;
504 	clk_hw_data->conf = core->conf;
505 
506 	init.name = core->name;
507 	init.ops = &rzg2l_cpg_pll5_4_clk_mux_ops;
508 	init.flags = CLK_SET_RATE_PARENT;
509 	init.num_parents = core->num_parents;
510 	init.parent_names = core->parent_names;
511 
512 	clk_hw = &clk_hw_data->hw;
513 	clk_hw->init = &init;
514 
515 	ret = devm_clk_hw_register(priv->dev, clk_hw);
516 	if (ret)
517 		return ERR_PTR(ret);
518 
519 	return clk_hw->clk;
520 }
521 
522 struct sipll5 {
523 	struct clk_hw hw;
524 	u32 conf;
525 	unsigned long foutpostdiv_rate;
526 	struct rzg2l_cpg_priv *priv;
527 };
528 
529 #define to_sipll5(_hw)	container_of(_hw, struct sipll5, hw)
530 
531 static unsigned long rzg2l_cpg_get_vclk_rate(struct clk_hw *hw,
532 					     unsigned long rate)
533 {
534 	struct sipll5 *sipll5 = to_sipll5(hw);
535 	struct rzg2l_cpg_priv *priv = sipll5->priv;
536 	unsigned long vclk;
537 
538 	vclk = rate / ((1 << priv->mux_dsi_div_params.dsi_div_a) *
539 		       (priv->mux_dsi_div_params.dsi_div_b + 1));
540 
541 	if (priv->mux_dsi_div_params.clksrc)
542 		vclk /= 2;
543 
544 	return vclk;
545 }
546 
547 static unsigned long rzg2l_cpg_sipll5_recalc_rate(struct clk_hw *hw,
548 						  unsigned long parent_rate)
549 {
550 	struct sipll5 *sipll5 = to_sipll5(hw);
551 	unsigned long pll5_rate = sipll5->foutpostdiv_rate;
552 
553 	if (!pll5_rate)
554 		pll5_rate = parent_rate;
555 
556 	return pll5_rate;
557 }
558 
559 static long rzg2l_cpg_sipll5_round_rate(struct clk_hw *hw,
560 					unsigned long rate,
561 					unsigned long *parent_rate)
562 {
563 	return rate;
564 }
565 
566 static int rzg2l_cpg_sipll5_set_rate(struct clk_hw *hw,
567 				     unsigned long rate,
568 				     unsigned long parent_rate)
569 {
570 	struct sipll5 *sipll5 = to_sipll5(hw);
571 	struct rzg2l_cpg_priv *priv = sipll5->priv;
572 	struct rzg2l_pll5_param params;
573 	unsigned long vclk_rate;
574 	int ret;
575 	u32 val;
576 
577 	/*
578 	 *  OSC --> PLL5 --> FOUTPOSTDIV-->|
579 	 *                   |             | -->MUX -->DIV_DSIA_B -->M3 -->VCLK
580 	 *                   |--FOUT1PH0-->|
581 	 *
582 	 * Based on the dot clock, the DSI divider clock calculates the parent
583 	 * rate and the pll5 parameters for generating FOUTPOSTDIV. It propagates
584 	 * that info to sipll5 which sets parameters for generating FOUTPOSTDIV.
585 	 *
586 	 * OSC --> PLL5 --> FOUTPOSTDIV
587 	 */
588 
589 	if (!rate)
590 		return -EINVAL;
591 
592 	vclk_rate = rzg2l_cpg_get_vclk_rate(hw, rate);
593 	sipll5->foutpostdiv_rate =
594 		rzg2l_cpg_get_foutpostdiv_rate(&params, vclk_rate);
595 
596 	/* Put PLL5 into standby mode */
597 	writel(CPG_SIPLL5_STBY_RESETB_WEN, priv->base + CPG_SIPLL5_STBY);
598 	ret = readl_poll_timeout(priv->base + CPG_SIPLL5_MON, val,
599 				 !(val & CPG_SIPLL5_MON_PLL5_LOCK), 100, 250000);
600 	if (ret) {
601 		dev_err(priv->dev, "failed to release pll5 lock");
602 		return ret;
603 	}
604 
605 	/* Output clock setting 1 */
606 	writel(CPG_SIPLL5_CLK1_POSTDIV1_WEN | CPG_SIPLL5_CLK1_POSTDIV2_WEN |
607 	       CPG_SIPLL5_CLK1_REFDIV_WEN  | (params.pl5_postdiv1 << 0) |
608 	       (params.pl5_postdiv2 << 4) | (params.pl5_refdiv << 8),
609 	       priv->base + CPG_SIPLL5_CLK1);
610 
611 	/* Output clock setting, SSCG modulation value setting 3 */
612 	writel((params.pl5_fracin << 8), priv->base + CPG_SIPLL5_CLK3);
613 
614 	/* Output clock setting 4 */
615 	writel(CPG_SIPLL5_CLK4_RESV_LSB | (params.pl5_intin << 16),
616 	       priv->base + CPG_SIPLL5_CLK4);
617 
618 	/* Output clock setting 5 */
619 	writel(params.pl5_spread, priv->base + CPG_SIPLL5_CLK5);
620 
621 	/* PLL normal mode setting */
622 	writel(CPG_SIPLL5_STBY_DOWNSPREAD_WEN | CPG_SIPLL5_STBY_SSCG_EN_WEN |
623 	       CPG_SIPLL5_STBY_RESETB_WEN | CPG_SIPLL5_STBY_RESETB,
624 	       priv->base + CPG_SIPLL5_STBY);
625 
626 	/* PLL normal mode transition, output clock stability check */
627 	ret = readl_poll_timeout(priv->base + CPG_SIPLL5_MON, val,
628 				 (val & CPG_SIPLL5_MON_PLL5_LOCK), 100, 250000);
629 	if (ret) {
630 		dev_err(priv->dev, "failed to lock pll5");
631 		return ret;
632 	}
633 
634 	return 0;
635 }
636 
637 static const struct clk_ops rzg2l_cpg_sipll5_ops = {
638 	.recalc_rate = rzg2l_cpg_sipll5_recalc_rate,
639 	.round_rate = rzg2l_cpg_sipll5_round_rate,
640 	.set_rate = rzg2l_cpg_sipll5_set_rate,
641 };
642 
643 static struct clk * __init
644 rzg2l_cpg_sipll5_register(const struct cpg_core_clk *core,
645 			  struct clk **clks,
646 			  struct rzg2l_cpg_priv *priv)
647 {
648 	const struct clk *parent;
649 	struct clk_init_data init;
650 	const char *parent_name;
651 	struct sipll5 *sipll5;
652 	struct clk_hw *clk_hw;
653 	int ret;
654 
655 	parent = clks[core->parent & 0xffff];
656 	if (IS_ERR(parent))
657 		return ERR_CAST(parent);
658 
659 	sipll5 = devm_kzalloc(priv->dev, sizeof(*sipll5), GFP_KERNEL);
660 	if (!sipll5)
661 		return ERR_PTR(-ENOMEM);
662 
663 	init.name = core->name;
664 	parent_name = __clk_get_name(parent);
665 	init.ops = &rzg2l_cpg_sipll5_ops;
666 	init.flags = 0;
667 	init.parent_names = &parent_name;
668 	init.num_parents = 1;
669 
670 	sipll5->hw.init = &init;
671 	sipll5->conf = core->conf;
672 	sipll5->priv = priv;
673 
674 	writel(CPG_SIPLL5_STBY_SSCG_EN_WEN | CPG_SIPLL5_STBY_RESETB_WEN |
675 	       CPG_SIPLL5_STBY_RESETB, priv->base + CPG_SIPLL5_STBY);
676 
677 	clk_hw = &sipll5->hw;
678 	clk_hw->init = &init;
679 
680 	ret = devm_clk_hw_register(priv->dev, clk_hw);
681 	if (ret)
682 		return ERR_PTR(ret);
683 
684 	priv->mux_dsi_div_params.clksrc = 1; /* Use clk src 1 for DSI */
685 	priv->mux_dsi_div_params.dsi_div_a = 1; /* Divided by 2 */
686 	priv->mux_dsi_div_params.dsi_div_b = 2; /* Divided by 3 */
687 
688 	return clk_hw->clk;
689 }
690 
691 struct pll_clk {
692 	struct clk_hw hw;
693 	unsigned int conf;
694 	unsigned int type;
695 	void __iomem *base;
696 	struct rzg2l_cpg_priv *priv;
697 };
698 
699 #define to_pll(_hw)	container_of(_hw, struct pll_clk, hw)
700 
701 static unsigned long rzg2l_cpg_pll_clk_recalc_rate(struct clk_hw *hw,
702 						   unsigned long parent_rate)
703 {
704 	struct pll_clk *pll_clk = to_pll(hw);
705 	struct rzg2l_cpg_priv *priv = pll_clk->priv;
706 	unsigned int val1, val2;
707 	unsigned int mult = 1;
708 	unsigned int div = 1;
709 
710 	if (pll_clk->type != CLK_TYPE_SAM_PLL)
711 		return parent_rate;
712 
713 	val1 = readl(priv->base + GET_REG_SAMPLL_CLK1(pll_clk->conf));
714 	val2 = readl(priv->base + GET_REG_SAMPLL_CLK2(pll_clk->conf));
715 	mult = MDIV(val1) + KDIV(val1) / 65536;
716 	div = PDIV(val1) << SDIV(val2);
717 
718 	return DIV_ROUND_CLOSEST_ULL((u64)parent_rate * mult, div);
719 }
720 
721 static const struct clk_ops rzg2l_cpg_pll_ops = {
722 	.recalc_rate = rzg2l_cpg_pll_clk_recalc_rate,
723 };
724 
725 static struct clk * __init
726 rzg2l_cpg_pll_clk_register(const struct cpg_core_clk *core,
727 			   struct clk **clks,
728 			   void __iomem *base,
729 			   struct rzg2l_cpg_priv *priv)
730 {
731 	struct device *dev = priv->dev;
732 	const struct clk *parent;
733 	struct clk_init_data init;
734 	const char *parent_name;
735 	struct pll_clk *pll_clk;
736 
737 	parent = clks[core->parent & 0xffff];
738 	if (IS_ERR(parent))
739 		return ERR_CAST(parent);
740 
741 	pll_clk = devm_kzalloc(dev, sizeof(*pll_clk), GFP_KERNEL);
742 	if (!pll_clk)
743 		return ERR_PTR(-ENOMEM);
744 
745 	parent_name = __clk_get_name(parent);
746 	init.name = core->name;
747 	init.ops = &rzg2l_cpg_pll_ops;
748 	init.flags = 0;
749 	init.parent_names = &parent_name;
750 	init.num_parents = 1;
751 
752 	pll_clk->hw.init = &init;
753 	pll_clk->conf = core->conf;
754 	pll_clk->base = base;
755 	pll_clk->priv = priv;
756 	pll_clk->type = core->type;
757 
758 	return clk_register(NULL, &pll_clk->hw);
759 }
760 
761 static struct clk
762 *rzg2l_cpg_clk_src_twocell_get(struct of_phandle_args *clkspec,
763 			       void *data)
764 {
765 	unsigned int clkidx = clkspec->args[1];
766 	struct rzg2l_cpg_priv *priv = data;
767 	struct device *dev = priv->dev;
768 	const char *type;
769 	struct clk *clk;
770 
771 	switch (clkspec->args[0]) {
772 	case CPG_CORE:
773 		type = "core";
774 		if (clkidx > priv->last_dt_core_clk) {
775 			dev_err(dev, "Invalid %s clock index %u\n", type, clkidx);
776 			return ERR_PTR(-EINVAL);
777 		}
778 		clk = priv->clks[clkidx];
779 		break;
780 
781 	case CPG_MOD:
782 		type = "module";
783 		if (clkidx >= priv->num_mod_clks) {
784 			dev_err(dev, "Invalid %s clock index %u\n", type,
785 				clkidx);
786 			return ERR_PTR(-EINVAL);
787 		}
788 		clk = priv->clks[priv->num_core_clks + clkidx];
789 		break;
790 
791 	default:
792 		dev_err(dev, "Invalid CPG clock type %u\n", clkspec->args[0]);
793 		return ERR_PTR(-EINVAL);
794 	}
795 
796 	if (IS_ERR(clk))
797 		dev_err(dev, "Cannot get %s clock %u: %ld", type, clkidx,
798 			PTR_ERR(clk));
799 	else
800 		dev_dbg(dev, "clock (%u, %u) is %pC at %lu Hz\n",
801 			clkspec->args[0], clkspec->args[1], clk,
802 			clk_get_rate(clk));
803 	return clk;
804 }
805 
806 static void __init
807 rzg2l_cpg_register_core_clk(const struct cpg_core_clk *core,
808 			    const struct rzg2l_cpg_info *info,
809 			    struct rzg2l_cpg_priv *priv)
810 {
811 	struct clk *clk = ERR_PTR(-EOPNOTSUPP), *parent;
812 	struct device *dev = priv->dev;
813 	unsigned int id = core->id, div = core->div;
814 	const char *parent_name;
815 
816 	WARN_DEBUG(id >= priv->num_core_clks);
817 	WARN_DEBUG(PTR_ERR(priv->clks[id]) != -ENOENT);
818 
819 	if (!core->name) {
820 		/* Skip NULLified clock */
821 		return;
822 	}
823 
824 	switch (core->type) {
825 	case CLK_TYPE_IN:
826 		clk = of_clk_get_by_name(priv->dev->of_node, core->name);
827 		break;
828 	case CLK_TYPE_FF:
829 		WARN_DEBUG(core->parent >= priv->num_core_clks);
830 		parent = priv->clks[core->parent];
831 		if (IS_ERR(parent)) {
832 			clk = parent;
833 			goto fail;
834 		}
835 
836 		parent_name = __clk_get_name(parent);
837 		clk = clk_register_fixed_factor(NULL, core->name,
838 						parent_name, CLK_SET_RATE_PARENT,
839 						core->mult, div);
840 		break;
841 	case CLK_TYPE_SAM_PLL:
842 		clk = rzg2l_cpg_pll_clk_register(core, priv->clks,
843 						 priv->base, priv);
844 		break;
845 	case CLK_TYPE_SIPLL5:
846 		clk = rzg2l_cpg_sipll5_register(core, priv->clks, priv);
847 		break;
848 	case CLK_TYPE_DIV:
849 		clk = rzg2l_cpg_div_clk_register(core, priv->clks,
850 						 priv->base, priv);
851 		break;
852 	case CLK_TYPE_MUX:
853 		clk = rzg2l_cpg_mux_clk_register(core, priv->base, priv);
854 		break;
855 	case CLK_TYPE_SD_MUX:
856 		clk = rzg2l_cpg_sd_mux_clk_register(core, priv->base, priv);
857 		break;
858 	case CLK_TYPE_PLL5_4_MUX:
859 		clk = rzg2l_cpg_pll5_4_mux_clk_register(core, priv);
860 		break;
861 	case CLK_TYPE_DSI_DIV:
862 		clk = rzg2l_cpg_dsi_div_clk_register(core, priv->clks, priv);
863 		break;
864 	default:
865 		goto fail;
866 	}
867 
868 	if (IS_ERR_OR_NULL(clk))
869 		goto fail;
870 
871 	dev_dbg(dev, "Core clock %pC at %lu Hz\n", clk, clk_get_rate(clk));
872 	priv->clks[id] = clk;
873 	return;
874 
875 fail:
876 	dev_err(dev, "Failed to register %s clock %s: %ld\n", "core",
877 		core->name, PTR_ERR(clk));
878 }
879 
880 /**
881  * struct mstp_clock - MSTP gating clock
882  *
883  * @hw: handle between common and hardware-specific interfaces
884  * @off: register offset
885  * @bit: ON/MON bit
886  * @enabled: soft state of the clock, if it is coupled with another clock
887  * @priv: CPG/MSTP private data
888  * @sibling: pointer to the other coupled clock
889  */
890 struct mstp_clock {
891 	struct clk_hw hw;
892 	u16 off;
893 	u8 bit;
894 	bool enabled;
895 	struct rzg2l_cpg_priv *priv;
896 	struct mstp_clock *sibling;
897 };
898 
899 #define to_mod_clock(_hw) container_of(_hw, struct mstp_clock, hw)
900 
901 static int rzg2l_mod_clock_endisable(struct clk_hw *hw, bool enable)
902 {
903 	struct mstp_clock *clock = to_mod_clock(hw);
904 	struct rzg2l_cpg_priv *priv = clock->priv;
905 	unsigned int reg = clock->off;
906 	struct device *dev = priv->dev;
907 	unsigned long flags;
908 	unsigned int i;
909 	u32 bitmask = BIT(clock->bit);
910 	u32 value;
911 
912 	if (!clock->off) {
913 		dev_dbg(dev, "%pC does not support ON/OFF\n",  hw->clk);
914 		return 0;
915 	}
916 
917 	dev_dbg(dev, "CLK_ON %u/%pC %s\n", CLK_ON_R(reg), hw->clk,
918 		enable ? "ON" : "OFF");
919 	spin_lock_irqsave(&priv->rmw_lock, flags);
920 
921 	if (enable)
922 		value = (bitmask << 16) | bitmask;
923 	else
924 		value = bitmask << 16;
925 	writel(value, priv->base + CLK_ON_R(reg));
926 
927 	spin_unlock_irqrestore(&priv->rmw_lock, flags);
928 
929 	if (!enable)
930 		return 0;
931 
932 	if (!priv->info->has_clk_mon_regs)
933 		return 0;
934 
935 	for (i = 1000; i > 0; --i) {
936 		if (((readl(priv->base + CLK_MON_R(reg))) & bitmask))
937 			break;
938 		cpu_relax();
939 	}
940 
941 	if (!i) {
942 		dev_err(dev, "Failed to enable CLK_ON %p\n",
943 			priv->base + CLK_ON_R(reg));
944 		return -ETIMEDOUT;
945 	}
946 
947 	return 0;
948 }
949 
950 static int rzg2l_mod_clock_enable(struct clk_hw *hw)
951 {
952 	struct mstp_clock *clock = to_mod_clock(hw);
953 
954 	if (clock->sibling) {
955 		struct rzg2l_cpg_priv *priv = clock->priv;
956 		unsigned long flags;
957 		bool enabled;
958 
959 		spin_lock_irqsave(&priv->rmw_lock, flags);
960 		enabled = clock->sibling->enabled;
961 		clock->enabled = true;
962 		spin_unlock_irqrestore(&priv->rmw_lock, flags);
963 		if (enabled)
964 			return 0;
965 	}
966 
967 	return rzg2l_mod_clock_endisable(hw, true);
968 }
969 
970 static void rzg2l_mod_clock_disable(struct clk_hw *hw)
971 {
972 	struct mstp_clock *clock = to_mod_clock(hw);
973 
974 	if (clock->sibling) {
975 		struct rzg2l_cpg_priv *priv = clock->priv;
976 		unsigned long flags;
977 		bool enabled;
978 
979 		spin_lock_irqsave(&priv->rmw_lock, flags);
980 		enabled = clock->sibling->enabled;
981 		clock->enabled = false;
982 		spin_unlock_irqrestore(&priv->rmw_lock, flags);
983 		if (enabled)
984 			return;
985 	}
986 
987 	rzg2l_mod_clock_endisable(hw, false);
988 }
989 
990 static int rzg2l_mod_clock_is_enabled(struct clk_hw *hw)
991 {
992 	struct mstp_clock *clock = to_mod_clock(hw);
993 	struct rzg2l_cpg_priv *priv = clock->priv;
994 	u32 bitmask = BIT(clock->bit);
995 	u32 value;
996 
997 	if (!clock->off) {
998 		dev_dbg(priv->dev, "%pC does not support ON/OFF\n",  hw->clk);
999 		return 1;
1000 	}
1001 
1002 	if (clock->sibling)
1003 		return clock->enabled;
1004 
1005 	if (priv->info->has_clk_mon_regs)
1006 		value = readl(priv->base + CLK_MON_R(clock->off));
1007 	else
1008 		value = readl(priv->base + clock->off);
1009 
1010 	return value & bitmask;
1011 }
1012 
1013 static const struct clk_ops rzg2l_mod_clock_ops = {
1014 	.enable = rzg2l_mod_clock_enable,
1015 	.disable = rzg2l_mod_clock_disable,
1016 	.is_enabled = rzg2l_mod_clock_is_enabled,
1017 };
1018 
1019 static struct mstp_clock
1020 *rzg2l_mod_clock_get_sibling(struct mstp_clock *clock,
1021 			     struct rzg2l_cpg_priv *priv)
1022 {
1023 	struct clk_hw *hw;
1024 	unsigned int i;
1025 
1026 	for (i = 0; i < priv->num_mod_clks; i++) {
1027 		struct mstp_clock *clk;
1028 
1029 		if (priv->clks[priv->num_core_clks + i] == ERR_PTR(-ENOENT))
1030 			continue;
1031 
1032 		hw = __clk_get_hw(priv->clks[priv->num_core_clks + i]);
1033 		clk = to_mod_clock(hw);
1034 		if (clock->off == clk->off && clock->bit == clk->bit)
1035 			return clk;
1036 	}
1037 
1038 	return NULL;
1039 }
1040 
1041 static void __init
1042 rzg2l_cpg_register_mod_clk(const struct rzg2l_mod_clk *mod,
1043 			   const struct rzg2l_cpg_info *info,
1044 			   struct rzg2l_cpg_priv *priv)
1045 {
1046 	struct mstp_clock *clock = NULL;
1047 	struct device *dev = priv->dev;
1048 	unsigned int id = mod->id;
1049 	struct clk_init_data init;
1050 	struct clk *parent, *clk;
1051 	const char *parent_name;
1052 	unsigned int i;
1053 
1054 	WARN_DEBUG(id < priv->num_core_clks);
1055 	WARN_DEBUG(id >= priv->num_core_clks + priv->num_mod_clks);
1056 	WARN_DEBUG(mod->parent >= priv->num_core_clks + priv->num_mod_clks);
1057 	WARN_DEBUG(PTR_ERR(priv->clks[id]) != -ENOENT);
1058 
1059 	if (!mod->name) {
1060 		/* Skip NULLified clock */
1061 		return;
1062 	}
1063 
1064 	parent = priv->clks[mod->parent];
1065 	if (IS_ERR(parent)) {
1066 		clk = parent;
1067 		goto fail;
1068 	}
1069 
1070 	clock = devm_kzalloc(dev, sizeof(*clock), GFP_KERNEL);
1071 	if (!clock) {
1072 		clk = ERR_PTR(-ENOMEM);
1073 		goto fail;
1074 	}
1075 
1076 	init.name = mod->name;
1077 	init.ops = &rzg2l_mod_clock_ops;
1078 	init.flags = CLK_SET_RATE_PARENT;
1079 	for (i = 0; i < info->num_crit_mod_clks; i++)
1080 		if (id == info->crit_mod_clks[i]) {
1081 			dev_dbg(dev, "CPG %s setting CLK_IS_CRITICAL\n",
1082 				mod->name);
1083 			init.flags |= CLK_IS_CRITICAL;
1084 			break;
1085 		}
1086 
1087 	parent_name = __clk_get_name(parent);
1088 	init.parent_names = &parent_name;
1089 	init.num_parents = 1;
1090 
1091 	clock->off = mod->off;
1092 	clock->bit = mod->bit;
1093 	clock->priv = priv;
1094 	clock->hw.init = &init;
1095 
1096 	clk = clk_register(NULL, &clock->hw);
1097 	if (IS_ERR(clk))
1098 		goto fail;
1099 
1100 	dev_dbg(dev, "Module clock %pC at %lu Hz\n", clk, clk_get_rate(clk));
1101 	priv->clks[id] = clk;
1102 
1103 	if (mod->is_coupled) {
1104 		struct mstp_clock *sibling;
1105 
1106 		clock->enabled = rzg2l_mod_clock_is_enabled(&clock->hw);
1107 		sibling = rzg2l_mod_clock_get_sibling(clock, priv);
1108 		if (sibling) {
1109 			clock->sibling = sibling;
1110 			sibling->sibling = clock;
1111 		}
1112 	}
1113 
1114 	return;
1115 
1116 fail:
1117 	dev_err(dev, "Failed to register %s clock %s: %ld\n", "module",
1118 		mod->name, PTR_ERR(clk));
1119 }
1120 
1121 #define rcdev_to_priv(x)	container_of(x, struct rzg2l_cpg_priv, rcdev)
1122 
1123 static int rzg2l_cpg_reset(struct reset_controller_dev *rcdev,
1124 			   unsigned long id)
1125 {
1126 	struct rzg2l_cpg_priv *priv = rcdev_to_priv(rcdev);
1127 	const struct rzg2l_cpg_info *info = priv->info;
1128 	unsigned int reg = info->resets[id].off;
1129 	u32 dis = BIT(info->resets[id].bit);
1130 	u32 we = dis << 16;
1131 
1132 	dev_dbg(rcdev->dev, "reset id:%ld offset:0x%x\n", id, CLK_RST_R(reg));
1133 
1134 	/* Reset module */
1135 	writel(we, priv->base + CLK_RST_R(reg));
1136 
1137 	/* Wait for at least one cycle of the RCLK clock (@ ca. 32 kHz) */
1138 	udelay(35);
1139 
1140 	/* Release module from reset state */
1141 	writel(we | dis, priv->base + CLK_RST_R(reg));
1142 
1143 	return 0;
1144 }
1145 
1146 static int rzg2l_cpg_assert(struct reset_controller_dev *rcdev,
1147 			    unsigned long id)
1148 {
1149 	struct rzg2l_cpg_priv *priv = rcdev_to_priv(rcdev);
1150 	const struct rzg2l_cpg_info *info = priv->info;
1151 	unsigned int reg = info->resets[id].off;
1152 	u32 value = BIT(info->resets[id].bit) << 16;
1153 
1154 	dev_dbg(rcdev->dev, "assert id:%ld offset:0x%x\n", id, CLK_RST_R(reg));
1155 
1156 	writel(value, priv->base + CLK_RST_R(reg));
1157 	return 0;
1158 }
1159 
1160 static int rzg2l_cpg_deassert(struct reset_controller_dev *rcdev,
1161 			      unsigned long id)
1162 {
1163 	struct rzg2l_cpg_priv *priv = rcdev_to_priv(rcdev);
1164 	const struct rzg2l_cpg_info *info = priv->info;
1165 	unsigned int reg = info->resets[id].off;
1166 	u32 dis = BIT(info->resets[id].bit);
1167 	u32 value = (dis << 16) | dis;
1168 
1169 	dev_dbg(rcdev->dev, "deassert id:%ld offset:0x%x\n", id,
1170 		CLK_RST_R(reg));
1171 
1172 	writel(value, priv->base + CLK_RST_R(reg));
1173 	return 0;
1174 }
1175 
1176 static int rzg2l_cpg_status(struct reset_controller_dev *rcdev,
1177 			    unsigned long id)
1178 {
1179 	struct rzg2l_cpg_priv *priv = rcdev_to_priv(rcdev);
1180 	const struct rzg2l_cpg_info *info = priv->info;
1181 	unsigned int reg = info->resets[id].off;
1182 	u32 bitmask = BIT(info->resets[id].bit);
1183 	s8 monbit = info->resets[id].monbit;
1184 
1185 	if (info->has_clk_mon_regs) {
1186 		return !!(readl(priv->base + CLK_MRST_R(reg)) & bitmask);
1187 	} else if (monbit >= 0) {
1188 		u32 monbitmask = BIT(monbit);
1189 
1190 		return !!(readl(priv->base + CPG_RST_MON) & monbitmask);
1191 	}
1192 	return -ENOTSUPP;
1193 }
1194 
1195 static const struct reset_control_ops rzg2l_cpg_reset_ops = {
1196 	.reset = rzg2l_cpg_reset,
1197 	.assert = rzg2l_cpg_assert,
1198 	.deassert = rzg2l_cpg_deassert,
1199 	.status = rzg2l_cpg_status,
1200 };
1201 
1202 static int rzg2l_cpg_reset_xlate(struct reset_controller_dev *rcdev,
1203 				 const struct of_phandle_args *reset_spec)
1204 {
1205 	struct rzg2l_cpg_priv *priv = rcdev_to_priv(rcdev);
1206 	const struct rzg2l_cpg_info *info = priv->info;
1207 	unsigned int id = reset_spec->args[0];
1208 
1209 	if (id >= rcdev->nr_resets || !info->resets[id].off) {
1210 		dev_err(rcdev->dev, "Invalid reset index %u\n", id);
1211 		return -EINVAL;
1212 	}
1213 
1214 	return id;
1215 }
1216 
1217 static int rzg2l_cpg_reset_controller_register(struct rzg2l_cpg_priv *priv)
1218 {
1219 	priv->rcdev.ops = &rzg2l_cpg_reset_ops;
1220 	priv->rcdev.of_node = priv->dev->of_node;
1221 	priv->rcdev.dev = priv->dev;
1222 	priv->rcdev.of_reset_n_cells = 1;
1223 	priv->rcdev.of_xlate = rzg2l_cpg_reset_xlate;
1224 	priv->rcdev.nr_resets = priv->num_resets;
1225 
1226 	return devm_reset_controller_register(priv->dev, &priv->rcdev);
1227 }
1228 
1229 static bool rzg2l_cpg_is_pm_clk(struct rzg2l_cpg_priv *priv,
1230 				const struct of_phandle_args *clkspec)
1231 {
1232 	const struct rzg2l_cpg_info *info = priv->info;
1233 	unsigned int id;
1234 	unsigned int i;
1235 
1236 	if (clkspec->args_count != 2)
1237 		return false;
1238 
1239 	if (clkspec->args[0] != CPG_MOD)
1240 		return false;
1241 
1242 	id = clkspec->args[1] + info->num_total_core_clks;
1243 	for (i = 0; i < info->num_no_pm_mod_clks; i++) {
1244 		if (info->no_pm_mod_clks[i] == id)
1245 			return false;
1246 	}
1247 
1248 	return true;
1249 }
1250 
1251 static int rzg2l_cpg_attach_dev(struct generic_pm_domain *domain, struct device *dev)
1252 {
1253 	struct rzg2l_cpg_priv *priv = container_of(domain, struct rzg2l_cpg_priv, genpd);
1254 	struct device_node *np = dev->of_node;
1255 	struct of_phandle_args clkspec;
1256 	bool once = true;
1257 	struct clk *clk;
1258 	int error;
1259 	int i = 0;
1260 
1261 	while (!of_parse_phandle_with_args(np, "clocks", "#clock-cells", i,
1262 					   &clkspec)) {
1263 		if (rzg2l_cpg_is_pm_clk(priv, &clkspec)) {
1264 			if (once) {
1265 				once = false;
1266 				error = pm_clk_create(dev);
1267 				if (error) {
1268 					of_node_put(clkspec.np);
1269 					goto err;
1270 				}
1271 			}
1272 			clk = of_clk_get_from_provider(&clkspec);
1273 			of_node_put(clkspec.np);
1274 			if (IS_ERR(clk)) {
1275 				error = PTR_ERR(clk);
1276 				goto fail_destroy;
1277 			}
1278 
1279 			error = pm_clk_add_clk(dev, clk);
1280 			if (error) {
1281 				dev_err(dev, "pm_clk_add_clk failed %d\n",
1282 					error);
1283 				goto fail_put;
1284 			}
1285 		} else {
1286 			of_node_put(clkspec.np);
1287 		}
1288 		i++;
1289 	}
1290 
1291 	return 0;
1292 
1293 fail_put:
1294 	clk_put(clk);
1295 
1296 fail_destroy:
1297 	pm_clk_destroy(dev);
1298 err:
1299 	return error;
1300 }
1301 
1302 static void rzg2l_cpg_detach_dev(struct generic_pm_domain *unused, struct device *dev)
1303 {
1304 	if (!pm_clk_no_clocks(dev))
1305 		pm_clk_destroy(dev);
1306 }
1307 
1308 static void rzg2l_cpg_genpd_remove(void *data)
1309 {
1310 	pm_genpd_remove(data);
1311 }
1312 
1313 static int __init rzg2l_cpg_add_clk_domain(struct rzg2l_cpg_priv *priv)
1314 {
1315 	struct device *dev = priv->dev;
1316 	struct device_node *np = dev->of_node;
1317 	struct generic_pm_domain *genpd = &priv->genpd;
1318 	int ret;
1319 
1320 	genpd->name = np->name;
1321 	genpd->flags = GENPD_FLAG_PM_CLK | GENPD_FLAG_ALWAYS_ON |
1322 		       GENPD_FLAG_ACTIVE_WAKEUP;
1323 	genpd->attach_dev = rzg2l_cpg_attach_dev;
1324 	genpd->detach_dev = rzg2l_cpg_detach_dev;
1325 	ret = pm_genpd_init(genpd, &pm_domain_always_on_gov, false);
1326 	if (ret)
1327 		return ret;
1328 
1329 	ret = devm_add_action_or_reset(dev, rzg2l_cpg_genpd_remove, genpd);
1330 	if (ret)
1331 		return ret;
1332 
1333 	return of_genpd_add_provider_simple(np, genpd);
1334 }
1335 
1336 static int __init rzg2l_cpg_probe(struct platform_device *pdev)
1337 {
1338 	struct device *dev = &pdev->dev;
1339 	struct device_node *np = dev->of_node;
1340 	const struct rzg2l_cpg_info *info;
1341 	struct rzg2l_cpg_priv *priv;
1342 	unsigned int nclks, i;
1343 	struct clk **clks;
1344 	int error;
1345 
1346 	info = of_device_get_match_data(dev);
1347 
1348 	priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
1349 	if (!priv)
1350 		return -ENOMEM;
1351 
1352 	priv->dev = dev;
1353 	priv->info = info;
1354 	spin_lock_init(&priv->rmw_lock);
1355 
1356 	priv->base = devm_platform_ioremap_resource(pdev, 0);
1357 	if (IS_ERR(priv->base))
1358 		return PTR_ERR(priv->base);
1359 
1360 	nclks = info->num_total_core_clks + info->num_hw_mod_clks;
1361 	clks = devm_kmalloc_array(dev, nclks, sizeof(*clks), GFP_KERNEL);
1362 	if (!clks)
1363 		return -ENOMEM;
1364 
1365 	dev_set_drvdata(dev, priv);
1366 	priv->clks = clks;
1367 	priv->num_core_clks = info->num_total_core_clks;
1368 	priv->num_mod_clks = info->num_hw_mod_clks;
1369 	priv->num_resets = info->num_resets;
1370 	priv->last_dt_core_clk = info->last_dt_core_clk;
1371 
1372 	for (i = 0; i < nclks; i++)
1373 		clks[i] = ERR_PTR(-ENOENT);
1374 
1375 	for (i = 0; i < info->num_core_clks; i++)
1376 		rzg2l_cpg_register_core_clk(&info->core_clks[i], info, priv);
1377 
1378 	for (i = 0; i < info->num_mod_clks; i++)
1379 		rzg2l_cpg_register_mod_clk(&info->mod_clks[i], info, priv);
1380 
1381 	error = of_clk_add_provider(np, rzg2l_cpg_clk_src_twocell_get, priv);
1382 	if (error)
1383 		return error;
1384 
1385 	error = devm_add_action_or_reset(dev, rzg2l_cpg_del_clk_provider, np);
1386 	if (error)
1387 		return error;
1388 
1389 	error = rzg2l_cpg_add_clk_domain(priv);
1390 	if (error)
1391 		return error;
1392 
1393 	error = rzg2l_cpg_reset_controller_register(priv);
1394 	if (error)
1395 		return error;
1396 
1397 	return 0;
1398 }
1399 
1400 static const struct of_device_id rzg2l_cpg_match[] = {
1401 #ifdef CONFIG_CLK_R9A07G043
1402 	{
1403 		.compatible = "renesas,r9a07g043-cpg",
1404 		.data = &r9a07g043_cpg_info,
1405 	},
1406 #endif
1407 #ifdef CONFIG_CLK_R9A07G044
1408 	{
1409 		.compatible = "renesas,r9a07g044-cpg",
1410 		.data = &r9a07g044_cpg_info,
1411 	},
1412 #endif
1413 #ifdef CONFIG_CLK_R9A07G054
1414 	{
1415 		.compatible = "renesas,r9a07g054-cpg",
1416 		.data = &r9a07g054_cpg_info,
1417 	},
1418 #endif
1419 #ifdef CONFIG_CLK_R9A09G011
1420 	{
1421 		.compatible = "renesas,r9a09g011-cpg",
1422 		.data = &r9a09g011_cpg_info,
1423 	},
1424 #endif
1425 	{ /* sentinel */ }
1426 };
1427 
1428 static struct platform_driver rzg2l_cpg_driver = {
1429 	.driver		= {
1430 		.name	= "rzg2l-cpg",
1431 		.of_match_table = rzg2l_cpg_match,
1432 	},
1433 };
1434 
1435 static int __init rzg2l_cpg_init(void)
1436 {
1437 	return platform_driver_probe(&rzg2l_cpg_driver, rzg2l_cpg_probe);
1438 }
1439 
1440 subsys_initcall(rzg2l_cpg_init);
1441 
1442 MODULE_DESCRIPTION("Renesas RZ/G2L CPG Driver");
1443