xref: /openbmc/linux/drivers/clk/ti/dpll.c (revision 827634ad)
1 /*
2  * OMAP DPLL clock support
3  *
4  * Copyright (C) 2013 Texas Instruments, Inc.
5  *
6  * Tero Kristo <t-kristo@ti.com>
7  *
8  * This program is free software; you can redistribute it and/or modify
9  * it under the terms of the GNU General Public License version 2 as
10  * published by the Free Software Foundation.
11  *
12  * This program is distributed "as is" WITHOUT ANY WARRANTY of any
13  * kind, whether express or implied; without even the implied warranty
14  * of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
15  * GNU General Public License for more details.
16  */
17 
18 #include <linux/clk-provider.h>
19 #include <linux/slab.h>
20 #include <linux/err.h>
21 #include <linux/of.h>
22 #include <linux/of_address.h>
23 #include <linux/clk/ti.h>
24 #include "clock.h"
25 
26 #undef pr_fmt
27 #define pr_fmt(fmt) "%s: " fmt, __func__
28 
29 #if defined(CONFIG_ARCH_OMAP4) || defined(CONFIG_SOC_OMAP5) || \
30 	defined(CONFIG_SOC_DRA7XX)
31 static const struct clk_ops dpll_m4xen_ck_ops = {
32 	.enable		= &omap3_noncore_dpll_enable,
33 	.disable	= &omap3_noncore_dpll_disable,
34 	.recalc_rate	= &omap4_dpll_regm4xen_recalc,
35 	.round_rate	= &omap4_dpll_regm4xen_round_rate,
36 	.set_rate	= &omap3_noncore_dpll_set_rate,
37 	.set_parent	= &omap3_noncore_dpll_set_parent,
38 	.set_rate_and_parent	= &omap3_noncore_dpll_set_rate_and_parent,
39 	.determine_rate	= &omap4_dpll_regm4xen_determine_rate,
40 	.get_parent	= &omap2_init_dpll_parent,
41 };
42 #else
43 static const struct clk_ops dpll_m4xen_ck_ops = {};
44 #endif
45 
46 #if defined(CONFIG_ARCH_OMAP3) || defined(CONFIG_ARCH_OMAP4) || \
47 	defined(CONFIG_SOC_OMAP5) || defined(CONFIG_SOC_DRA7XX) || \
48 	defined(CONFIG_SOC_AM33XX) || defined(CONFIG_SOC_AM43XX)
49 static const struct clk_ops dpll_core_ck_ops = {
50 	.recalc_rate	= &omap3_dpll_recalc,
51 	.get_parent	= &omap2_init_dpll_parent,
52 };
53 
54 static const struct clk_ops dpll_ck_ops = {
55 	.enable		= &omap3_noncore_dpll_enable,
56 	.disable	= &omap3_noncore_dpll_disable,
57 	.recalc_rate	= &omap3_dpll_recalc,
58 	.round_rate	= &omap2_dpll_round_rate,
59 	.set_rate	= &omap3_noncore_dpll_set_rate,
60 	.set_parent	= &omap3_noncore_dpll_set_parent,
61 	.set_rate_and_parent	= &omap3_noncore_dpll_set_rate_and_parent,
62 	.determine_rate	= &omap3_noncore_dpll_determine_rate,
63 	.get_parent	= &omap2_init_dpll_parent,
64 };
65 
66 static const struct clk_ops dpll_no_gate_ck_ops = {
67 	.recalc_rate	= &omap3_dpll_recalc,
68 	.get_parent	= &omap2_init_dpll_parent,
69 	.round_rate	= &omap2_dpll_round_rate,
70 	.set_rate	= &omap3_noncore_dpll_set_rate,
71 	.set_parent	= &omap3_noncore_dpll_set_parent,
72 	.set_rate_and_parent	= &omap3_noncore_dpll_set_rate_and_parent,
73 	.determine_rate	= &omap3_noncore_dpll_determine_rate,
74 };
75 #else
76 static const struct clk_ops dpll_core_ck_ops = {};
77 static const struct clk_ops dpll_ck_ops = {};
78 static const struct clk_ops dpll_no_gate_ck_ops = {};
79 const struct clk_hw_omap_ops clkhwops_omap3_dpll = {};
80 #endif
81 
82 #ifdef CONFIG_ARCH_OMAP2
83 static const struct clk_ops omap2_dpll_core_ck_ops = {
84 	.get_parent	= &omap2_init_dpll_parent,
85 	.recalc_rate	= &omap2_dpllcore_recalc,
86 	.round_rate	= &omap2_dpll_round_rate,
87 	.set_rate	= &omap2_reprogram_dpllcore,
88 };
89 #else
90 static const struct clk_ops omap2_dpll_core_ck_ops = {};
91 #endif
92 
93 #ifdef CONFIG_ARCH_OMAP3
94 static const struct clk_ops omap3_dpll_core_ck_ops = {
95 	.get_parent	= &omap2_init_dpll_parent,
96 	.recalc_rate	= &omap3_dpll_recalc,
97 	.round_rate	= &omap2_dpll_round_rate,
98 };
99 #else
100 static const struct clk_ops omap3_dpll_core_ck_ops = {};
101 #endif
102 
103 #ifdef CONFIG_ARCH_OMAP3
104 static const struct clk_ops omap3_dpll_ck_ops = {
105 	.enable		= &omap3_noncore_dpll_enable,
106 	.disable	= &omap3_noncore_dpll_disable,
107 	.get_parent	= &omap2_init_dpll_parent,
108 	.recalc_rate	= &omap3_dpll_recalc,
109 	.set_rate	= &omap3_noncore_dpll_set_rate,
110 	.set_parent	= &omap3_noncore_dpll_set_parent,
111 	.set_rate_and_parent	= &omap3_noncore_dpll_set_rate_and_parent,
112 	.determine_rate	= &omap3_noncore_dpll_determine_rate,
113 	.round_rate	= &omap2_dpll_round_rate,
114 };
115 
116 static const struct clk_ops omap3_dpll_per_ck_ops = {
117 	.enable		= &omap3_noncore_dpll_enable,
118 	.disable	= &omap3_noncore_dpll_disable,
119 	.get_parent	= &omap2_init_dpll_parent,
120 	.recalc_rate	= &omap3_dpll_recalc,
121 	.set_rate	= &omap3_dpll4_set_rate,
122 	.set_parent	= &omap3_noncore_dpll_set_parent,
123 	.set_rate_and_parent	= &omap3_dpll4_set_rate_and_parent,
124 	.determine_rate	= &omap3_noncore_dpll_determine_rate,
125 	.round_rate	= &omap2_dpll_round_rate,
126 };
127 #endif
128 
129 static const struct clk_ops dpll_x2_ck_ops = {
130 	.recalc_rate	= &omap3_clkoutx2_recalc,
131 };
132 
133 /**
134  * _register_dpll - low level registration of a DPLL clock
135  * @hw: hardware clock definition for the clock
136  * @node: device node for the clock
137  *
138  * Finalizes DPLL registration process. In case a failure (clk-ref or
139  * clk-bypass is missing), the clock is added to retry list and
140  * the initialization is retried on later stage.
141  */
142 static void __init _register_dpll(struct clk_hw *hw,
143 				  struct device_node *node)
144 {
145 	struct clk_hw_omap *clk_hw = to_clk_hw_omap(hw);
146 	struct dpll_data *dd = clk_hw->dpll_data;
147 	struct clk *clk;
148 
149 	dd->clk_ref = of_clk_get(node, 0);
150 	dd->clk_bypass = of_clk_get(node, 1);
151 
152 	if (IS_ERR(dd->clk_ref) || IS_ERR(dd->clk_bypass)) {
153 		pr_debug("clk-ref or clk-bypass missing for %s, retry later\n",
154 			 node->name);
155 		if (!ti_clk_retry_init(node, hw, _register_dpll))
156 			return;
157 
158 		goto cleanup;
159 	}
160 
161 	/* register the clock */
162 	clk = clk_register(NULL, &clk_hw->hw);
163 
164 	if (!IS_ERR(clk)) {
165 		omap2_init_clk_hw_omap_clocks(clk);
166 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
167 		kfree(clk_hw->hw.init->parent_names);
168 		kfree(clk_hw->hw.init);
169 		return;
170 	}
171 
172 cleanup:
173 	kfree(clk_hw->dpll_data);
174 	kfree(clk_hw->hw.init->parent_names);
175 	kfree(clk_hw->hw.init);
176 	kfree(clk_hw);
177 }
178 
179 #if defined(CONFIG_ARCH_OMAP3) && defined(CONFIG_ATAGS)
180 void __iomem *_get_reg(u8 module, u16 offset)
181 {
182 	u32 reg;
183 	struct clk_omap_reg *reg_setup;
184 
185 	reg_setup = (struct clk_omap_reg *)&reg;
186 
187 	reg_setup->index = module;
188 	reg_setup->offset = offset;
189 
190 	return (void __iomem *)reg;
191 }
192 
193 struct clk *ti_clk_register_dpll(struct ti_clk *setup)
194 {
195 	struct clk_hw_omap *clk_hw;
196 	struct clk_init_data init = { NULL };
197 	struct dpll_data *dd;
198 	struct clk *clk;
199 	struct ti_clk_dpll *dpll;
200 	const struct clk_ops *ops = &omap3_dpll_ck_ops;
201 	struct clk *clk_ref;
202 	struct clk *clk_bypass;
203 
204 	dpll = setup->data;
205 
206 	if (dpll->num_parents < 2)
207 		return ERR_PTR(-EINVAL);
208 
209 	clk_ref = clk_get_sys(NULL, dpll->parents[0]);
210 	clk_bypass = clk_get_sys(NULL, dpll->parents[1]);
211 
212 	if (IS_ERR_OR_NULL(clk_ref) || IS_ERR_OR_NULL(clk_bypass))
213 		return ERR_PTR(-EAGAIN);
214 
215 	dd = kzalloc(sizeof(*dd), GFP_KERNEL);
216 	clk_hw = kzalloc(sizeof(*clk_hw), GFP_KERNEL);
217 	if (!dd || !clk_hw) {
218 		clk = ERR_PTR(-ENOMEM);
219 		goto cleanup;
220 	}
221 
222 	clk_hw->dpll_data = dd;
223 	clk_hw->ops = &clkhwops_omap3_dpll;
224 	clk_hw->hw.init = &init;
225 	clk_hw->flags = MEMMAP_ADDRESSING;
226 
227 	init.name = setup->name;
228 	init.ops = ops;
229 
230 	init.num_parents = dpll->num_parents;
231 	init.parent_names = dpll->parents;
232 
233 	dd->control_reg = _get_reg(dpll->module, dpll->control_reg);
234 	dd->idlest_reg = _get_reg(dpll->module, dpll->idlest_reg);
235 	dd->mult_div1_reg = _get_reg(dpll->module, dpll->mult_div1_reg);
236 	dd->autoidle_reg = _get_reg(dpll->module, dpll->autoidle_reg);
237 
238 	dd->modes = dpll->modes;
239 	dd->div1_mask = dpll->div1_mask;
240 	dd->idlest_mask = dpll->idlest_mask;
241 	dd->mult_mask = dpll->mult_mask;
242 	dd->autoidle_mask = dpll->autoidle_mask;
243 	dd->enable_mask = dpll->enable_mask;
244 	dd->sddiv_mask = dpll->sddiv_mask;
245 	dd->dco_mask = dpll->dco_mask;
246 	dd->max_divider = dpll->max_divider;
247 	dd->min_divider = dpll->min_divider;
248 	dd->max_multiplier = dpll->max_multiplier;
249 	dd->auto_recal_bit = dpll->auto_recal_bit;
250 	dd->recal_en_bit = dpll->recal_en_bit;
251 	dd->recal_st_bit = dpll->recal_st_bit;
252 
253 	dd->clk_ref = clk_ref;
254 	dd->clk_bypass = clk_bypass;
255 
256 	if (dpll->flags & CLKF_CORE)
257 		ops = &omap3_dpll_core_ck_ops;
258 
259 	if (dpll->flags & CLKF_PER)
260 		ops = &omap3_dpll_per_ck_ops;
261 
262 	if (dpll->flags & CLKF_J_TYPE)
263 		dd->flags |= DPLL_J_TYPE;
264 
265 	clk = clk_register(NULL, &clk_hw->hw);
266 
267 	if (!IS_ERR(clk))
268 		return clk;
269 
270 cleanup:
271 	kfree(dd);
272 	kfree(clk_hw);
273 	return clk;
274 }
275 #endif
276 
277 #if defined(CONFIG_ARCH_OMAP4) || defined(CONFIG_SOC_OMAP5) || \
278 	defined(CONFIG_SOC_DRA7XX) || defined(CONFIG_SOC_AM33XX) || \
279 	defined(CONFIG_SOC_AM43XX)
280 /**
281  * _register_dpll_x2 - Registers a DPLLx2 clock
282  * @node: device node for this clock
283  * @ops: clk_ops for this clock
284  * @hw_ops: clk_hw_ops for this clock
285  *
286  * Initializes a DPLL x 2 clock from device tree data.
287  */
288 static void _register_dpll_x2(struct device_node *node,
289 			      const struct clk_ops *ops,
290 			      const struct clk_hw_omap_ops *hw_ops)
291 {
292 	struct clk *clk;
293 	struct clk_init_data init = { NULL };
294 	struct clk_hw_omap *clk_hw;
295 	const char *name = node->name;
296 	const char *parent_name;
297 
298 	parent_name = of_clk_get_parent_name(node, 0);
299 	if (!parent_name) {
300 		pr_err("%s must have parent\n", node->name);
301 		return;
302 	}
303 
304 	clk_hw = kzalloc(sizeof(*clk_hw), GFP_KERNEL);
305 	if (!clk_hw)
306 		return;
307 
308 	clk_hw->ops = hw_ops;
309 	clk_hw->hw.init = &init;
310 
311 	init.name = name;
312 	init.ops = ops;
313 	init.parent_names = &parent_name;
314 	init.num_parents = 1;
315 
316 	/* register the clock */
317 	clk = clk_register(NULL, &clk_hw->hw);
318 
319 	if (IS_ERR(clk)) {
320 		kfree(clk_hw);
321 	} else {
322 		omap2_init_clk_hw_omap_clocks(clk);
323 		of_clk_add_provider(node, of_clk_src_simple_get, clk);
324 	}
325 }
326 #endif
327 
328 /**
329  * of_ti_dpll_setup - Setup function for OMAP DPLL clocks
330  * @node: device node containing the DPLL info
331  * @ops: ops for the DPLL
332  * @ddt: DPLL data template to use
333  *
334  * Initializes a DPLL clock from device tree data.
335  */
336 static void __init of_ti_dpll_setup(struct device_node *node,
337 				    const struct clk_ops *ops,
338 				    const struct dpll_data *ddt)
339 {
340 	struct clk_hw_omap *clk_hw = NULL;
341 	struct clk_init_data *init = NULL;
342 	const char **parent_names = NULL;
343 	struct dpll_data *dd = NULL;
344 	int i;
345 	u8 dpll_mode = 0;
346 
347 	dd = kzalloc(sizeof(*dd), GFP_KERNEL);
348 	clk_hw = kzalloc(sizeof(*clk_hw), GFP_KERNEL);
349 	init = kzalloc(sizeof(*init), GFP_KERNEL);
350 	if (!dd || !clk_hw || !init)
351 		goto cleanup;
352 
353 	memcpy(dd, ddt, sizeof(*dd));
354 
355 	clk_hw->dpll_data = dd;
356 	clk_hw->ops = &clkhwops_omap3_dpll;
357 	clk_hw->hw.init = init;
358 	clk_hw->flags = MEMMAP_ADDRESSING;
359 
360 	init->name = node->name;
361 	init->ops = ops;
362 
363 	init->num_parents = of_clk_get_parent_count(node);
364 	if (init->num_parents < 1) {
365 		pr_err("%s must have parent(s)\n", node->name);
366 		goto cleanup;
367 	}
368 
369 	parent_names = kzalloc(sizeof(char *) * init->num_parents, GFP_KERNEL);
370 	if (!parent_names)
371 		goto cleanup;
372 
373 	for (i = 0; i < init->num_parents; i++)
374 		parent_names[i] = of_clk_get_parent_name(node, i);
375 
376 	init->parent_names = parent_names;
377 
378 	dd->control_reg = ti_clk_get_reg_addr(node, 0);
379 
380 	/*
381 	 * Special case for OMAP2 DPLL, register order is different due to
382 	 * missing idlest_reg, also clkhwops is different. Detected from
383 	 * missing idlest_mask.
384 	 */
385 	if (!dd->idlest_mask) {
386 		dd->mult_div1_reg = ti_clk_get_reg_addr(node, 1);
387 #ifdef CONFIG_ARCH_OMAP2
388 		clk_hw->ops = &clkhwops_omap2xxx_dpll;
389 		omap2xxx_clkt_dpllcore_init(&clk_hw->hw);
390 #endif
391 	} else {
392 		dd->idlest_reg = ti_clk_get_reg_addr(node, 1);
393 		if (IS_ERR(dd->idlest_reg))
394 			goto cleanup;
395 
396 		dd->mult_div1_reg = ti_clk_get_reg_addr(node, 2);
397 	}
398 
399 	if (IS_ERR(dd->control_reg) || IS_ERR(dd->mult_div1_reg))
400 		goto cleanup;
401 
402 	if (dd->autoidle_mask) {
403 		dd->autoidle_reg = ti_clk_get_reg_addr(node, 3);
404 		if (IS_ERR(dd->autoidle_reg))
405 			goto cleanup;
406 	}
407 
408 	if (of_property_read_bool(node, "ti,low-power-stop"))
409 		dpll_mode |= 1 << DPLL_LOW_POWER_STOP;
410 
411 	if (of_property_read_bool(node, "ti,low-power-bypass"))
412 		dpll_mode |= 1 << DPLL_LOW_POWER_BYPASS;
413 
414 	if (of_property_read_bool(node, "ti,lock"))
415 		dpll_mode |= 1 << DPLL_LOCKED;
416 
417 	if (dpll_mode)
418 		dd->modes = dpll_mode;
419 
420 	_register_dpll(&clk_hw->hw, node);
421 	return;
422 
423 cleanup:
424 	kfree(dd);
425 	kfree(parent_names);
426 	kfree(init);
427 	kfree(clk_hw);
428 }
429 
430 #if defined(CONFIG_ARCH_OMAP4) || defined(CONFIG_SOC_OMAP5) || \
431 	defined(CONFIG_SOC_DRA7XX)
432 static void __init of_ti_omap4_dpll_x2_setup(struct device_node *node)
433 {
434 	_register_dpll_x2(node, &dpll_x2_ck_ops, &clkhwops_omap4_dpllmx);
435 }
436 CLK_OF_DECLARE(ti_omap4_dpll_x2_clock, "ti,omap4-dpll-x2-clock",
437 	       of_ti_omap4_dpll_x2_setup);
438 #endif
439 
440 #if defined(CONFIG_SOC_AM33XX) || defined(CONFIG_SOC_AM43XX)
441 static void __init of_ti_am3_dpll_x2_setup(struct device_node *node)
442 {
443 	_register_dpll_x2(node, &dpll_x2_ck_ops, NULL);
444 }
445 CLK_OF_DECLARE(ti_am3_dpll_x2_clock, "ti,am3-dpll-x2-clock",
446 	       of_ti_am3_dpll_x2_setup);
447 #endif
448 
449 #ifdef CONFIG_ARCH_OMAP3
450 static void __init of_ti_omap3_dpll_setup(struct device_node *node)
451 {
452 	const struct dpll_data dd = {
453 		.idlest_mask = 0x1,
454 		.enable_mask = 0x7,
455 		.autoidle_mask = 0x7,
456 		.mult_mask = 0x7ff << 8,
457 		.div1_mask = 0x7f,
458 		.max_multiplier = 2047,
459 		.max_divider = 128,
460 		.min_divider = 1,
461 		.freqsel_mask = 0xf0,
462 		.modes = (1 << DPLL_LOW_POWER_BYPASS) | (1 << DPLL_LOCKED),
463 	};
464 
465 	of_ti_dpll_setup(node, &omap3_dpll_ck_ops, &dd);
466 }
467 CLK_OF_DECLARE(ti_omap3_dpll_clock, "ti,omap3-dpll-clock",
468 	       of_ti_omap3_dpll_setup);
469 
470 static void __init of_ti_omap3_core_dpll_setup(struct device_node *node)
471 {
472 	const struct dpll_data dd = {
473 		.idlest_mask = 0x1,
474 		.enable_mask = 0x7,
475 		.autoidle_mask = 0x7,
476 		.mult_mask = 0x7ff << 16,
477 		.div1_mask = 0x7f << 8,
478 		.max_multiplier = 2047,
479 		.max_divider = 128,
480 		.min_divider = 1,
481 		.freqsel_mask = 0xf0,
482 	};
483 
484 	of_ti_dpll_setup(node, &omap3_dpll_core_ck_ops, &dd);
485 }
486 CLK_OF_DECLARE(ti_omap3_core_dpll_clock, "ti,omap3-dpll-core-clock",
487 	       of_ti_omap3_core_dpll_setup);
488 
489 static void __init of_ti_omap3_per_dpll_setup(struct device_node *node)
490 {
491 	const struct dpll_data dd = {
492 		.idlest_mask = 0x1 << 1,
493 		.enable_mask = 0x7 << 16,
494 		.autoidle_mask = 0x7 << 3,
495 		.mult_mask = 0x7ff << 8,
496 		.div1_mask = 0x7f,
497 		.max_multiplier = 2047,
498 		.max_divider = 128,
499 		.min_divider = 1,
500 		.freqsel_mask = 0xf00000,
501 		.modes = (1 << DPLL_LOW_POWER_STOP) | (1 << DPLL_LOCKED),
502 	};
503 
504 	of_ti_dpll_setup(node, &omap3_dpll_per_ck_ops, &dd);
505 }
506 CLK_OF_DECLARE(ti_omap3_per_dpll_clock, "ti,omap3-dpll-per-clock",
507 	       of_ti_omap3_per_dpll_setup);
508 
509 static void __init of_ti_omap3_per_jtype_dpll_setup(struct device_node *node)
510 {
511 	const struct dpll_data dd = {
512 		.idlest_mask = 0x1 << 1,
513 		.enable_mask = 0x7 << 16,
514 		.autoidle_mask = 0x7 << 3,
515 		.mult_mask = 0xfff << 8,
516 		.div1_mask = 0x7f,
517 		.max_multiplier = 4095,
518 		.max_divider = 128,
519 		.min_divider = 1,
520 		.sddiv_mask = 0xff << 24,
521 		.dco_mask = 0xe << 20,
522 		.flags = DPLL_J_TYPE,
523 		.modes = (1 << DPLL_LOW_POWER_STOP) | (1 << DPLL_LOCKED),
524 	};
525 
526 	of_ti_dpll_setup(node, &omap3_dpll_per_ck_ops, &dd);
527 }
528 CLK_OF_DECLARE(ti_omap3_per_jtype_dpll_clock, "ti,omap3-dpll-per-j-type-clock",
529 	       of_ti_omap3_per_jtype_dpll_setup);
530 #endif
531 
532 static void __init of_ti_omap4_dpll_setup(struct device_node *node)
533 {
534 	const struct dpll_data dd = {
535 		.idlest_mask = 0x1,
536 		.enable_mask = 0x7,
537 		.autoidle_mask = 0x7,
538 		.mult_mask = 0x7ff << 8,
539 		.div1_mask = 0x7f,
540 		.max_multiplier = 2047,
541 		.max_divider = 128,
542 		.min_divider = 1,
543 		.modes = (1 << DPLL_LOW_POWER_BYPASS) | (1 << DPLL_LOCKED),
544 	};
545 
546 	of_ti_dpll_setup(node, &dpll_ck_ops, &dd);
547 }
548 CLK_OF_DECLARE(ti_omap4_dpll_clock, "ti,omap4-dpll-clock",
549 	       of_ti_omap4_dpll_setup);
550 
551 static void __init of_ti_omap5_mpu_dpll_setup(struct device_node *node)
552 {
553 	const struct dpll_data dd = {
554 		.idlest_mask = 0x1,
555 		.enable_mask = 0x7,
556 		.autoidle_mask = 0x7,
557 		.mult_mask = 0x7ff << 8,
558 		.div1_mask = 0x7f,
559 		.max_multiplier = 2047,
560 		.max_divider = 128,
561 		.dcc_mask = BIT(22),
562 		.dcc_rate = 1400000000, /* DCC beyond 1.4GHz */
563 		.min_divider = 1,
564 		.modes = (1 << DPLL_LOW_POWER_BYPASS) | (1 << DPLL_LOCKED),
565 	};
566 
567 	of_ti_dpll_setup(node, &dpll_ck_ops, &dd);
568 }
569 CLK_OF_DECLARE(of_ti_omap5_mpu_dpll_clock, "ti,omap5-mpu-dpll-clock",
570 	       of_ti_omap5_mpu_dpll_setup);
571 
572 static void __init of_ti_omap4_core_dpll_setup(struct device_node *node)
573 {
574 	const struct dpll_data dd = {
575 		.idlest_mask = 0x1,
576 		.enable_mask = 0x7,
577 		.autoidle_mask = 0x7,
578 		.mult_mask = 0x7ff << 8,
579 		.div1_mask = 0x7f,
580 		.max_multiplier = 2047,
581 		.max_divider = 128,
582 		.min_divider = 1,
583 		.modes = (1 << DPLL_LOW_POWER_BYPASS) | (1 << DPLL_LOCKED),
584 	};
585 
586 	of_ti_dpll_setup(node, &dpll_core_ck_ops, &dd);
587 }
588 CLK_OF_DECLARE(ti_omap4_core_dpll_clock, "ti,omap4-dpll-core-clock",
589 	       of_ti_omap4_core_dpll_setup);
590 
591 #if defined(CONFIG_ARCH_OMAP4) || defined(CONFIG_SOC_OMAP5) || \
592 	defined(CONFIG_SOC_DRA7XX)
593 static void __init of_ti_omap4_m4xen_dpll_setup(struct device_node *node)
594 {
595 	const struct dpll_data dd = {
596 		.idlest_mask = 0x1,
597 		.enable_mask = 0x7,
598 		.autoidle_mask = 0x7,
599 		.mult_mask = 0x7ff << 8,
600 		.div1_mask = 0x7f,
601 		.max_multiplier = 2047,
602 		.max_divider = 128,
603 		.min_divider = 1,
604 		.m4xen_mask = 0x800,
605 		.lpmode_mask = 1 << 10,
606 		.modes = (1 << DPLL_LOW_POWER_BYPASS) | (1 << DPLL_LOCKED),
607 	};
608 
609 	of_ti_dpll_setup(node, &dpll_m4xen_ck_ops, &dd);
610 }
611 CLK_OF_DECLARE(ti_omap4_m4xen_dpll_clock, "ti,omap4-dpll-m4xen-clock",
612 	       of_ti_omap4_m4xen_dpll_setup);
613 
614 static void __init of_ti_omap4_jtype_dpll_setup(struct device_node *node)
615 {
616 	const struct dpll_data dd = {
617 		.idlest_mask = 0x1,
618 		.enable_mask = 0x7,
619 		.autoidle_mask = 0x7,
620 		.mult_mask = 0xfff << 8,
621 		.div1_mask = 0xff,
622 		.max_multiplier = 4095,
623 		.max_divider = 256,
624 		.min_divider = 1,
625 		.sddiv_mask = 0xff << 24,
626 		.flags = DPLL_J_TYPE,
627 		.modes = (1 << DPLL_LOW_POWER_BYPASS) | (1 << DPLL_LOCKED),
628 	};
629 
630 	of_ti_dpll_setup(node, &dpll_m4xen_ck_ops, &dd);
631 }
632 CLK_OF_DECLARE(ti_omap4_jtype_dpll_clock, "ti,omap4-dpll-j-type-clock",
633 	       of_ti_omap4_jtype_dpll_setup);
634 #endif
635 
636 static void __init of_ti_am3_no_gate_dpll_setup(struct device_node *node)
637 {
638 	const struct dpll_data dd = {
639 		.idlest_mask = 0x1,
640 		.enable_mask = 0x7,
641 		.mult_mask = 0x7ff << 8,
642 		.div1_mask = 0x7f,
643 		.max_multiplier = 2047,
644 		.max_divider = 128,
645 		.min_divider = 1,
646 		.modes = (1 << DPLL_LOW_POWER_BYPASS) | (1 << DPLL_LOCKED),
647 	};
648 
649 	of_ti_dpll_setup(node, &dpll_no_gate_ck_ops, &dd);
650 }
651 CLK_OF_DECLARE(ti_am3_no_gate_dpll_clock, "ti,am3-dpll-no-gate-clock",
652 	       of_ti_am3_no_gate_dpll_setup);
653 
654 static void __init of_ti_am3_jtype_dpll_setup(struct device_node *node)
655 {
656 	const struct dpll_data dd = {
657 		.idlest_mask = 0x1,
658 		.enable_mask = 0x7,
659 		.mult_mask = 0x7ff << 8,
660 		.div1_mask = 0x7f,
661 		.max_multiplier = 4095,
662 		.max_divider = 256,
663 		.min_divider = 2,
664 		.flags = DPLL_J_TYPE,
665 		.modes = (1 << DPLL_LOW_POWER_BYPASS) | (1 << DPLL_LOCKED),
666 	};
667 
668 	of_ti_dpll_setup(node, &dpll_ck_ops, &dd);
669 }
670 CLK_OF_DECLARE(ti_am3_jtype_dpll_clock, "ti,am3-dpll-j-type-clock",
671 	       of_ti_am3_jtype_dpll_setup);
672 
673 static void __init of_ti_am3_no_gate_jtype_dpll_setup(struct device_node *node)
674 {
675 	const struct dpll_data dd = {
676 		.idlest_mask = 0x1,
677 		.enable_mask = 0x7,
678 		.mult_mask = 0x7ff << 8,
679 		.div1_mask = 0x7f,
680 		.max_multiplier = 2047,
681 		.max_divider = 128,
682 		.min_divider = 1,
683 		.flags = DPLL_J_TYPE,
684 		.modes = (1 << DPLL_LOW_POWER_BYPASS) | (1 << DPLL_LOCKED),
685 	};
686 
687 	of_ti_dpll_setup(node, &dpll_no_gate_ck_ops, &dd);
688 }
689 CLK_OF_DECLARE(ti_am3_no_gate_jtype_dpll_clock,
690 	       "ti,am3-dpll-no-gate-j-type-clock",
691 	       of_ti_am3_no_gate_jtype_dpll_setup);
692 
693 static void __init of_ti_am3_dpll_setup(struct device_node *node)
694 {
695 	const struct dpll_data dd = {
696 		.idlest_mask = 0x1,
697 		.enable_mask = 0x7,
698 		.mult_mask = 0x7ff << 8,
699 		.div1_mask = 0x7f,
700 		.max_multiplier = 2047,
701 		.max_divider = 128,
702 		.min_divider = 1,
703 		.modes = (1 << DPLL_LOW_POWER_BYPASS) | (1 << DPLL_LOCKED),
704 	};
705 
706 	of_ti_dpll_setup(node, &dpll_ck_ops, &dd);
707 }
708 CLK_OF_DECLARE(ti_am3_dpll_clock, "ti,am3-dpll-clock", of_ti_am3_dpll_setup);
709 
710 static void __init of_ti_am3_core_dpll_setup(struct device_node *node)
711 {
712 	const struct dpll_data dd = {
713 		.idlest_mask = 0x1,
714 		.enable_mask = 0x7,
715 		.mult_mask = 0x7ff << 8,
716 		.div1_mask = 0x7f,
717 		.max_multiplier = 2047,
718 		.max_divider = 128,
719 		.min_divider = 1,
720 		.modes = (1 << DPLL_LOW_POWER_BYPASS) | (1 << DPLL_LOCKED),
721 	};
722 
723 	of_ti_dpll_setup(node, &dpll_core_ck_ops, &dd);
724 }
725 CLK_OF_DECLARE(ti_am3_core_dpll_clock, "ti,am3-dpll-core-clock",
726 	       of_ti_am3_core_dpll_setup);
727 
728 static void __init of_ti_omap2_core_dpll_setup(struct device_node *node)
729 {
730 	const struct dpll_data dd = {
731 		.enable_mask = 0x3,
732 		.mult_mask = 0x3ff << 12,
733 		.div1_mask = 0xf << 8,
734 		.max_divider = 16,
735 		.min_divider = 1,
736 	};
737 
738 	of_ti_dpll_setup(node, &omap2_dpll_core_ck_ops, &dd);
739 }
740 CLK_OF_DECLARE(ti_omap2_core_dpll_clock, "ti,omap2-dpll-core-clock",
741 	       of_ti_omap2_core_dpll_setup);
742