1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Alchemy clocks. 4 * 5 * Exposes all configurable internal clock sources to the clk framework. 6 * 7 * We have: 8 * - Root source, usually 12MHz supplied by an external crystal 9 * - 3 PLLs which generate multiples of root rate [AUX, CPU, AUX2] 10 * 11 * Dividers: 12 * - 6 clock dividers with: 13 * * selectable source [one of the PLLs], 14 * * output divided between [2 .. 512 in steps of 2] (!Au1300) 15 * or [1 .. 256 in steps of 1] (Au1300), 16 * * can be enabled individually. 17 * 18 * - up to 6 "internal" (fixed) consumers which: 19 * * take either AUXPLL or one of the above 6 dividers as input, 20 * * divide this input by 1, 2, or 4 (and 3 on Au1300). 21 * * can be disabled separately. 22 * 23 * Misc clocks: 24 * - sysbus clock: CPU core clock (CPUPLL) divided by 2, 3 or 4. 25 * depends on board design and should be set by bootloader, read-only. 26 * - peripheral clock: half the rate of sysbus clock, source for a lot 27 * of peripheral blocks, read-only. 28 * - memory clock: clk rate to main memory chips, depends on board 29 * design and is read-only, 30 * - lrclk: the static bus clock signal for synchronous operation. 31 * depends on board design, must be set by bootloader, 32 * but may be required to correctly configure devices attached to 33 * the static bus. The Au1000/1500/1100 manuals call it LCLK, on 34 * later models it's called RCLK. 35 */ 36 37 #include <linux/init.h> 38 #include <linux/io.h> 39 #include <linux/clk.h> 40 #include <linux/clk-provider.h> 41 #include <linux/clkdev.h> 42 #include <linux/slab.h> 43 #include <linux/spinlock.h> 44 #include <linux/types.h> 45 #include <asm/mach-au1x00/au1000.h> 46 47 /* Base clock: 12MHz is the default in all databooks, and I haven't 48 * found any board yet which uses a different rate. 49 */ 50 #define ALCHEMY_ROOTCLK_RATE 12000000 51 52 /* 53 * the internal sources which can be driven by the PLLs and dividers. 54 * Names taken from the databooks, refer to them for more information, 55 * especially which ones are share a clock line. 56 */ 57 static const char * const alchemy_au1300_intclknames[] = { 58 "lcd_intclk", "gpemgp_clk", "maempe_clk", "maebsa_clk", 59 "EXTCLK0", "EXTCLK1" 60 }; 61 62 static const char * const alchemy_au1200_intclknames[] = { 63 "lcd_intclk", NULL, NULL, NULL, "EXTCLK0", "EXTCLK1" 64 }; 65 66 static const char * const alchemy_au1550_intclknames[] = { 67 "usb_clk", "psc0_intclk", "psc1_intclk", "pci_clko", 68 "EXTCLK0", "EXTCLK1" 69 }; 70 71 static const char * const alchemy_au1100_intclknames[] = { 72 "usb_clk", "lcd_intclk", NULL, "i2s_clk", "EXTCLK0", "EXTCLK1" 73 }; 74 75 static const char * const alchemy_au1500_intclknames[] = { 76 NULL, "usbd_clk", "usbh_clk", "pci_clko", "EXTCLK0", "EXTCLK1" 77 }; 78 79 static const char * const alchemy_au1000_intclknames[] = { 80 "irda_clk", "usbd_clk", "usbh_clk", "i2s_clk", "EXTCLK0", 81 "EXTCLK1" 82 }; 83 84 /* aliases for a few on-chip sources which are either shared 85 * or have gone through name changes. 86 */ 87 static struct clk_aliastable { 88 char *alias; 89 char *base; 90 int cputype; 91 } alchemy_clk_aliases[] __initdata = { 92 { "usbh_clk", "usb_clk", ALCHEMY_CPU_AU1100 }, 93 { "usbd_clk", "usb_clk", ALCHEMY_CPU_AU1100 }, 94 { "irda_clk", "usb_clk", ALCHEMY_CPU_AU1100 }, 95 { "usbh_clk", "usb_clk", ALCHEMY_CPU_AU1550 }, 96 { "usbd_clk", "usb_clk", ALCHEMY_CPU_AU1550 }, 97 { "psc2_intclk", "usb_clk", ALCHEMY_CPU_AU1550 }, 98 { "psc3_intclk", "EXTCLK0", ALCHEMY_CPU_AU1550 }, 99 { "psc0_intclk", "EXTCLK0", ALCHEMY_CPU_AU1200 }, 100 { "psc1_intclk", "EXTCLK1", ALCHEMY_CPU_AU1200 }, 101 { "psc0_intclk", "EXTCLK0", ALCHEMY_CPU_AU1300 }, 102 { "psc2_intclk", "EXTCLK0", ALCHEMY_CPU_AU1300 }, 103 { "psc1_intclk", "EXTCLK1", ALCHEMY_CPU_AU1300 }, 104 { "psc3_intclk", "EXTCLK1", ALCHEMY_CPU_AU1300 }, 105 106 { NULL, NULL, 0 }, 107 }; 108 109 #define IOMEM(x) ((void __iomem *)(KSEG1ADDR(CPHYSADDR(x)))) 110 111 /* access locks to SYS_FREQCTRL0/1 and SYS_CLKSRC registers */ 112 static spinlock_t alchemy_clk_fg0_lock; 113 static spinlock_t alchemy_clk_fg1_lock; 114 static spinlock_t alchemy_clk_csrc_lock; 115 116 /* CPU Core clock *****************************************************/ 117 118 static unsigned long alchemy_clk_cpu_recalc(struct clk_hw *hw, 119 unsigned long parent_rate) 120 { 121 unsigned long t; 122 123 /* 124 * On early Au1000, sys_cpupll was write-only. Since these 125 * silicon versions of Au1000 are not sold, we don't bend 126 * over backwards trying to determine the frequency. 127 */ 128 if (unlikely(au1xxx_cpu_has_pll_wo())) 129 t = 396000000; 130 else { 131 t = alchemy_rdsys(AU1000_SYS_CPUPLL) & 0x7f; 132 if (alchemy_get_cputype() < ALCHEMY_CPU_AU1300) 133 t &= 0x3f; 134 t *= parent_rate; 135 } 136 137 return t; 138 } 139 140 void __init alchemy_set_lpj(void) 141 { 142 preset_lpj = alchemy_clk_cpu_recalc(NULL, ALCHEMY_ROOTCLK_RATE); 143 preset_lpj /= 2 * HZ; 144 } 145 146 static const struct clk_ops alchemy_clkops_cpu = { 147 .recalc_rate = alchemy_clk_cpu_recalc, 148 }; 149 150 static struct clk __init *alchemy_clk_setup_cpu(const char *parent_name, 151 int ctype) 152 { 153 struct clk_init_data id; 154 struct clk_hw *h; 155 struct clk *clk; 156 157 h = kzalloc(sizeof(*h), GFP_KERNEL); 158 if (!h) 159 return ERR_PTR(-ENOMEM); 160 161 id.name = ALCHEMY_CPU_CLK; 162 id.parent_names = &parent_name; 163 id.num_parents = 1; 164 id.flags = 0; 165 id.ops = &alchemy_clkops_cpu; 166 h->init = &id; 167 168 clk = clk_register(NULL, h); 169 if (IS_ERR(clk)) { 170 pr_err("failed to register clock\n"); 171 kfree(h); 172 } 173 174 return clk; 175 } 176 177 /* AUXPLLs ************************************************************/ 178 179 struct alchemy_auxpll_clk { 180 struct clk_hw hw; 181 unsigned long reg; /* au1300 has also AUXPLL2 */ 182 int maxmult; /* max multiplier */ 183 }; 184 #define to_auxpll_clk(x) container_of(x, struct alchemy_auxpll_clk, hw) 185 186 static unsigned long alchemy_clk_aux_recalc(struct clk_hw *hw, 187 unsigned long parent_rate) 188 { 189 struct alchemy_auxpll_clk *a = to_auxpll_clk(hw); 190 191 return (alchemy_rdsys(a->reg) & 0xff) * parent_rate; 192 } 193 194 static int alchemy_clk_aux_setr(struct clk_hw *hw, 195 unsigned long rate, 196 unsigned long parent_rate) 197 { 198 struct alchemy_auxpll_clk *a = to_auxpll_clk(hw); 199 unsigned long d = rate; 200 201 if (rate) 202 d /= parent_rate; 203 else 204 d = 0; 205 206 /* minimum is 84MHz, max is 756-1032 depending on variant */ 207 if (((d < 7) && (d != 0)) || (d > a->maxmult)) 208 return -EINVAL; 209 210 alchemy_wrsys(d, a->reg); 211 return 0; 212 } 213 214 static long alchemy_clk_aux_roundr(struct clk_hw *hw, 215 unsigned long rate, 216 unsigned long *parent_rate) 217 { 218 struct alchemy_auxpll_clk *a = to_auxpll_clk(hw); 219 unsigned long mult; 220 221 if (!rate || !*parent_rate) 222 return 0; 223 224 mult = rate / (*parent_rate); 225 226 if (mult && (mult < 7)) 227 mult = 7; 228 if (mult > a->maxmult) 229 mult = a->maxmult; 230 231 return (*parent_rate) * mult; 232 } 233 234 static const struct clk_ops alchemy_clkops_aux = { 235 .recalc_rate = alchemy_clk_aux_recalc, 236 .set_rate = alchemy_clk_aux_setr, 237 .round_rate = alchemy_clk_aux_roundr, 238 }; 239 240 static struct clk __init *alchemy_clk_setup_aux(const char *parent_name, 241 char *name, int maxmult, 242 unsigned long reg) 243 { 244 struct clk_init_data id; 245 struct clk *c; 246 struct alchemy_auxpll_clk *a; 247 248 a = kzalloc(sizeof(*a), GFP_KERNEL); 249 if (!a) 250 return ERR_PTR(-ENOMEM); 251 252 id.name = name; 253 id.parent_names = &parent_name; 254 id.num_parents = 1; 255 id.flags = CLK_GET_RATE_NOCACHE; 256 id.ops = &alchemy_clkops_aux; 257 258 a->reg = reg; 259 a->maxmult = maxmult; 260 a->hw.init = &id; 261 262 c = clk_register(NULL, &a->hw); 263 if (!IS_ERR(c)) 264 clk_register_clkdev(c, name, NULL); 265 else 266 kfree(a); 267 268 return c; 269 } 270 271 /* sysbus_clk *********************************************************/ 272 273 static struct clk __init *alchemy_clk_setup_sysbus(const char *pn) 274 { 275 unsigned long v = (alchemy_rdsys(AU1000_SYS_POWERCTRL) & 3) + 2; 276 struct clk *c; 277 278 c = clk_register_fixed_factor(NULL, ALCHEMY_SYSBUS_CLK, 279 pn, 0, 1, v); 280 if (!IS_ERR(c)) 281 clk_register_clkdev(c, ALCHEMY_SYSBUS_CLK, NULL); 282 return c; 283 } 284 285 /* Peripheral Clock ***************************************************/ 286 287 static struct clk __init *alchemy_clk_setup_periph(const char *pn) 288 { 289 /* Peripheral clock runs at half the rate of sysbus clk */ 290 struct clk *c; 291 292 c = clk_register_fixed_factor(NULL, ALCHEMY_PERIPH_CLK, 293 pn, 0, 1, 2); 294 if (!IS_ERR(c)) 295 clk_register_clkdev(c, ALCHEMY_PERIPH_CLK, NULL); 296 return c; 297 } 298 299 /* mem clock **********************************************************/ 300 301 static struct clk __init *alchemy_clk_setup_mem(const char *pn, int ct) 302 { 303 void __iomem *addr = IOMEM(AU1000_MEM_PHYS_ADDR); 304 unsigned long v; 305 struct clk *c; 306 int div; 307 308 switch (ct) { 309 case ALCHEMY_CPU_AU1550: 310 case ALCHEMY_CPU_AU1200: 311 v = __raw_readl(addr + AU1550_MEM_SDCONFIGB); 312 div = (v & (1 << 15)) ? 1 : 2; 313 break; 314 case ALCHEMY_CPU_AU1300: 315 v = __raw_readl(addr + AU1550_MEM_SDCONFIGB); 316 div = (v & (1 << 31)) ? 1 : 2; 317 break; 318 case ALCHEMY_CPU_AU1000: 319 case ALCHEMY_CPU_AU1500: 320 case ALCHEMY_CPU_AU1100: 321 default: 322 div = 2; 323 break; 324 } 325 326 c = clk_register_fixed_factor(NULL, ALCHEMY_MEM_CLK, pn, 327 0, 1, div); 328 if (!IS_ERR(c)) 329 clk_register_clkdev(c, ALCHEMY_MEM_CLK, NULL); 330 return c; 331 } 332 333 /* lrclk: external synchronous static bus clock ***********************/ 334 335 static struct clk __init *alchemy_clk_setup_lrclk(const char *pn, int t) 336 { 337 /* Au1000, Au1500: MEM_STCFG0[11]: If bit is set, lrclk=pclk/5, 338 * otherwise lrclk=pclk/4. 339 * All other variants: MEM_STCFG0[15:13] = divisor. 340 * L/RCLK = periph_clk / (divisor + 1) 341 * On Au1000, Au1500, Au1100 it's called LCLK, 342 * on later models it's called RCLK, but it's the same thing. 343 */ 344 struct clk *c; 345 unsigned long v = alchemy_rdsmem(AU1000_MEM_STCFG0); 346 347 switch (t) { 348 case ALCHEMY_CPU_AU1000: 349 case ALCHEMY_CPU_AU1500: 350 v = 4 + ((v >> 11) & 1); 351 break; 352 default: /* all other models */ 353 v = ((v >> 13) & 7) + 1; 354 } 355 c = clk_register_fixed_factor(NULL, ALCHEMY_LR_CLK, 356 pn, 0, 1, v); 357 if (!IS_ERR(c)) 358 clk_register_clkdev(c, ALCHEMY_LR_CLK, NULL); 359 return c; 360 } 361 362 /* Clock dividers and muxes *******************************************/ 363 364 /* data for fgen and csrc mux-dividers */ 365 struct alchemy_fgcs_clk { 366 struct clk_hw hw; 367 spinlock_t *reglock; /* register lock */ 368 unsigned long reg; /* SYS_FREQCTRL0/1 */ 369 int shift; /* offset in register */ 370 int parent; /* parent before disable [Au1300] */ 371 int isen; /* is it enabled? */ 372 int *dt; /* dividertable for csrc */ 373 }; 374 #define to_fgcs_clk(x) container_of(x, struct alchemy_fgcs_clk, hw) 375 376 static long alchemy_calc_div(unsigned long rate, unsigned long prate, 377 int scale, int maxdiv, unsigned long *rv) 378 { 379 long div1, div2; 380 381 div1 = prate / rate; 382 if ((prate / div1) > rate) 383 div1++; 384 385 if (scale == 2) { /* only div-by-multiple-of-2 possible */ 386 if (div1 & 1) 387 div1++; /* stay <=prate */ 388 } 389 390 div2 = (div1 / scale) - 1; /* value to write to register */ 391 392 if (div2 > maxdiv) 393 div2 = maxdiv; 394 if (rv) 395 *rv = div2; 396 397 div1 = ((div2 + 1) * scale); 398 return div1; 399 } 400 401 static int alchemy_clk_fgcs_detr(struct clk_hw *hw, 402 struct clk_rate_request *req, 403 int scale, int maxdiv) 404 { 405 struct clk_hw *pc, *bpc, *free; 406 long tdv, tpr, pr, nr, br, bpr, diff, lastdiff; 407 int j; 408 409 lastdiff = INT_MAX; 410 bpr = 0; 411 bpc = NULL; 412 br = -EINVAL; 413 free = NULL; 414 415 /* look at the rates each enabled parent supplies and select 416 * the one that gets closest to but not over the requested rate. 417 */ 418 for (j = 0; j < 7; j++) { 419 pc = clk_hw_get_parent_by_index(hw, j); 420 if (!pc) 421 break; 422 423 /* if this parent is currently unused, remember it. 424 * XXX: we would actually want clk_has_active_children() 425 * but this is a good-enough approximation for now. 426 */ 427 if (!clk_hw_is_prepared(pc)) { 428 if (!free) 429 free = pc; 430 } 431 432 pr = clk_hw_get_rate(pc); 433 if (pr < req->rate) 434 continue; 435 436 /* what can hardware actually provide */ 437 tdv = alchemy_calc_div(req->rate, pr, scale, maxdiv, NULL); 438 nr = pr / tdv; 439 diff = req->rate - nr; 440 if (nr > req->rate) 441 continue; 442 443 if (diff < lastdiff) { 444 lastdiff = diff; 445 bpr = pr; 446 bpc = pc; 447 br = nr; 448 } 449 if (diff == 0) 450 break; 451 } 452 453 /* if we couldn't get the exact rate we wanted from the enabled 454 * parents, maybe we can tell an available disabled/inactive one 455 * to give us a rate we can divide down to the requested rate. 456 */ 457 if (lastdiff && free) { 458 for (j = (maxdiv == 4) ? 1 : scale; j <= maxdiv; j += scale) { 459 tpr = req->rate * j; 460 if (tpr < 0) 461 break; 462 pr = clk_hw_round_rate(free, tpr); 463 464 tdv = alchemy_calc_div(req->rate, pr, scale, maxdiv, 465 NULL); 466 nr = pr / tdv; 467 diff = req->rate - nr; 468 if (nr > req->rate) 469 continue; 470 if (diff < lastdiff) { 471 lastdiff = diff; 472 bpr = pr; 473 bpc = free; 474 br = nr; 475 } 476 if (diff == 0) 477 break; 478 } 479 } 480 481 if (br < 0) 482 return br; 483 484 req->best_parent_rate = bpr; 485 req->best_parent_hw = bpc; 486 req->rate = br; 487 488 return 0; 489 } 490 491 static int alchemy_clk_fgv1_en(struct clk_hw *hw) 492 { 493 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 494 unsigned long v, flags; 495 496 spin_lock_irqsave(c->reglock, flags); 497 v = alchemy_rdsys(c->reg); 498 v |= (1 << 1) << c->shift; 499 alchemy_wrsys(v, c->reg); 500 spin_unlock_irqrestore(c->reglock, flags); 501 502 return 0; 503 } 504 505 static int alchemy_clk_fgv1_isen(struct clk_hw *hw) 506 { 507 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 508 unsigned long v = alchemy_rdsys(c->reg) >> (c->shift + 1); 509 510 return v & 1; 511 } 512 513 static void alchemy_clk_fgv1_dis(struct clk_hw *hw) 514 { 515 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 516 unsigned long v, flags; 517 518 spin_lock_irqsave(c->reglock, flags); 519 v = alchemy_rdsys(c->reg); 520 v &= ~((1 << 1) << c->shift); 521 alchemy_wrsys(v, c->reg); 522 spin_unlock_irqrestore(c->reglock, flags); 523 } 524 525 static int alchemy_clk_fgv1_setp(struct clk_hw *hw, u8 index) 526 { 527 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 528 unsigned long v, flags; 529 530 spin_lock_irqsave(c->reglock, flags); 531 v = alchemy_rdsys(c->reg); 532 if (index) 533 v |= (1 << c->shift); 534 else 535 v &= ~(1 << c->shift); 536 alchemy_wrsys(v, c->reg); 537 spin_unlock_irqrestore(c->reglock, flags); 538 539 return 0; 540 } 541 542 static u8 alchemy_clk_fgv1_getp(struct clk_hw *hw) 543 { 544 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 545 546 return (alchemy_rdsys(c->reg) >> c->shift) & 1; 547 } 548 549 static int alchemy_clk_fgv1_setr(struct clk_hw *hw, unsigned long rate, 550 unsigned long parent_rate) 551 { 552 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 553 unsigned long div, v, flags, ret; 554 int sh = c->shift + 2; 555 556 if (!rate || !parent_rate || rate > (parent_rate / 2)) 557 return -EINVAL; 558 ret = alchemy_calc_div(rate, parent_rate, 2, 512, &div); 559 spin_lock_irqsave(c->reglock, flags); 560 v = alchemy_rdsys(c->reg); 561 v &= ~(0xff << sh); 562 v |= div << sh; 563 alchemy_wrsys(v, c->reg); 564 spin_unlock_irqrestore(c->reglock, flags); 565 566 return 0; 567 } 568 569 static unsigned long alchemy_clk_fgv1_recalc(struct clk_hw *hw, 570 unsigned long parent_rate) 571 { 572 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 573 unsigned long v = alchemy_rdsys(c->reg) >> (c->shift + 2); 574 575 v = ((v & 0xff) + 1) * 2; 576 return parent_rate / v; 577 } 578 579 static int alchemy_clk_fgv1_detr(struct clk_hw *hw, 580 struct clk_rate_request *req) 581 { 582 return alchemy_clk_fgcs_detr(hw, req, 2, 512); 583 } 584 585 /* Au1000, Au1100, Au15x0, Au12x0 */ 586 static const struct clk_ops alchemy_clkops_fgenv1 = { 587 .recalc_rate = alchemy_clk_fgv1_recalc, 588 .determine_rate = alchemy_clk_fgv1_detr, 589 .set_rate = alchemy_clk_fgv1_setr, 590 .set_parent = alchemy_clk_fgv1_setp, 591 .get_parent = alchemy_clk_fgv1_getp, 592 .enable = alchemy_clk_fgv1_en, 593 .disable = alchemy_clk_fgv1_dis, 594 .is_enabled = alchemy_clk_fgv1_isen, 595 }; 596 597 static void __alchemy_clk_fgv2_en(struct alchemy_fgcs_clk *c) 598 { 599 unsigned long v = alchemy_rdsys(c->reg); 600 601 v &= ~(3 << c->shift); 602 v |= (c->parent & 3) << c->shift; 603 alchemy_wrsys(v, c->reg); 604 c->isen = 1; 605 } 606 607 static int alchemy_clk_fgv2_en(struct clk_hw *hw) 608 { 609 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 610 unsigned long flags; 611 612 /* enable by setting the previous parent clock */ 613 spin_lock_irqsave(c->reglock, flags); 614 __alchemy_clk_fgv2_en(c); 615 spin_unlock_irqrestore(c->reglock, flags); 616 617 return 0; 618 } 619 620 static int alchemy_clk_fgv2_isen(struct clk_hw *hw) 621 { 622 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 623 624 return ((alchemy_rdsys(c->reg) >> c->shift) & 3) != 0; 625 } 626 627 static void alchemy_clk_fgv2_dis(struct clk_hw *hw) 628 { 629 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 630 unsigned long v, flags; 631 632 spin_lock_irqsave(c->reglock, flags); 633 v = alchemy_rdsys(c->reg); 634 v &= ~(3 << c->shift); /* set input mux to "disabled" state */ 635 alchemy_wrsys(v, c->reg); 636 c->isen = 0; 637 spin_unlock_irqrestore(c->reglock, flags); 638 } 639 640 static int alchemy_clk_fgv2_setp(struct clk_hw *hw, u8 index) 641 { 642 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 643 unsigned long flags; 644 645 spin_lock_irqsave(c->reglock, flags); 646 c->parent = index + 1; /* value to write to register */ 647 if (c->isen) 648 __alchemy_clk_fgv2_en(c); 649 spin_unlock_irqrestore(c->reglock, flags); 650 651 return 0; 652 } 653 654 static u8 alchemy_clk_fgv2_getp(struct clk_hw *hw) 655 { 656 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 657 unsigned long flags, v; 658 659 spin_lock_irqsave(c->reglock, flags); 660 v = c->parent - 1; 661 spin_unlock_irqrestore(c->reglock, flags); 662 return v; 663 } 664 665 /* fg0-2 and fg4-6 share a "scale"-bit. With this bit cleared, the 666 * dividers behave exactly as on previous models (dividers are multiples 667 * of 2); with the bit set, dividers are multiples of 1, halving their 668 * range, but making them also much more flexible. 669 */ 670 static int alchemy_clk_fgv2_setr(struct clk_hw *hw, unsigned long rate, 671 unsigned long parent_rate) 672 { 673 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 674 int sh = c->shift + 2; 675 unsigned long div, v, flags, ret; 676 677 if (!rate || !parent_rate || rate > parent_rate) 678 return -EINVAL; 679 680 v = alchemy_rdsys(c->reg) & (1 << 30); /* test "scale" bit */ 681 ret = alchemy_calc_div(rate, parent_rate, v ? 1 : 2, 682 v ? 256 : 512, &div); 683 684 spin_lock_irqsave(c->reglock, flags); 685 v = alchemy_rdsys(c->reg); 686 v &= ~(0xff << sh); 687 v |= (div & 0xff) << sh; 688 alchemy_wrsys(v, c->reg); 689 spin_unlock_irqrestore(c->reglock, flags); 690 691 return 0; 692 } 693 694 static unsigned long alchemy_clk_fgv2_recalc(struct clk_hw *hw, 695 unsigned long parent_rate) 696 { 697 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 698 int sh = c->shift + 2; 699 unsigned long v, t; 700 701 v = alchemy_rdsys(c->reg); 702 t = parent_rate / (((v >> sh) & 0xff) + 1); 703 if ((v & (1 << 30)) == 0) /* test scale bit */ 704 t /= 2; 705 706 return t; 707 } 708 709 static int alchemy_clk_fgv2_detr(struct clk_hw *hw, 710 struct clk_rate_request *req) 711 { 712 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 713 int scale, maxdiv; 714 715 if (alchemy_rdsys(c->reg) & (1 << 30)) { 716 scale = 1; 717 maxdiv = 256; 718 } else { 719 scale = 2; 720 maxdiv = 512; 721 } 722 723 return alchemy_clk_fgcs_detr(hw, req, scale, maxdiv); 724 } 725 726 /* Au1300 larger input mux, no separate disable bit, flexible divider */ 727 static const struct clk_ops alchemy_clkops_fgenv2 = { 728 .recalc_rate = alchemy_clk_fgv2_recalc, 729 .determine_rate = alchemy_clk_fgv2_detr, 730 .set_rate = alchemy_clk_fgv2_setr, 731 .set_parent = alchemy_clk_fgv2_setp, 732 .get_parent = alchemy_clk_fgv2_getp, 733 .enable = alchemy_clk_fgv2_en, 734 .disable = alchemy_clk_fgv2_dis, 735 .is_enabled = alchemy_clk_fgv2_isen, 736 }; 737 738 static const char * const alchemy_clk_fgv1_parents[] = { 739 ALCHEMY_CPU_CLK, ALCHEMY_AUXPLL_CLK 740 }; 741 742 static const char * const alchemy_clk_fgv2_parents[] = { 743 ALCHEMY_AUXPLL2_CLK, ALCHEMY_CPU_CLK, ALCHEMY_AUXPLL_CLK 744 }; 745 746 static const char * const alchemy_clk_fgen_names[] = { 747 ALCHEMY_FG0_CLK, ALCHEMY_FG1_CLK, ALCHEMY_FG2_CLK, 748 ALCHEMY_FG3_CLK, ALCHEMY_FG4_CLK, ALCHEMY_FG5_CLK }; 749 750 static int __init alchemy_clk_init_fgens(int ctype) 751 { 752 struct clk *c; 753 struct clk_init_data id; 754 struct alchemy_fgcs_clk *a; 755 unsigned long v; 756 int i, ret; 757 758 switch (ctype) { 759 case ALCHEMY_CPU_AU1000...ALCHEMY_CPU_AU1200: 760 id.ops = &alchemy_clkops_fgenv1; 761 id.parent_names = alchemy_clk_fgv1_parents; 762 id.num_parents = 2; 763 break; 764 case ALCHEMY_CPU_AU1300: 765 id.ops = &alchemy_clkops_fgenv2; 766 id.parent_names = alchemy_clk_fgv2_parents; 767 id.num_parents = 3; 768 break; 769 default: 770 return -ENODEV; 771 } 772 id.flags = CLK_SET_RATE_PARENT | CLK_GET_RATE_NOCACHE; 773 774 a = kzalloc((sizeof(*a)) * 6, GFP_KERNEL); 775 if (!a) 776 return -ENOMEM; 777 778 spin_lock_init(&alchemy_clk_fg0_lock); 779 spin_lock_init(&alchemy_clk_fg1_lock); 780 ret = 0; 781 for (i = 0; i < 6; i++) { 782 id.name = alchemy_clk_fgen_names[i]; 783 a->shift = 10 * (i < 3 ? i : i - 3); 784 if (i > 2) { 785 a->reg = AU1000_SYS_FREQCTRL1; 786 a->reglock = &alchemy_clk_fg1_lock; 787 } else { 788 a->reg = AU1000_SYS_FREQCTRL0; 789 a->reglock = &alchemy_clk_fg0_lock; 790 } 791 792 /* default to first parent if bootloader has set 793 * the mux to disabled state. 794 */ 795 if (ctype == ALCHEMY_CPU_AU1300) { 796 v = alchemy_rdsys(a->reg); 797 a->parent = (v >> a->shift) & 3; 798 if (!a->parent) { 799 a->parent = 1; 800 a->isen = 0; 801 } else 802 a->isen = 1; 803 } 804 805 a->hw.init = &id; 806 c = clk_register(NULL, &a->hw); 807 if (IS_ERR(c)) 808 ret++; 809 else 810 clk_register_clkdev(c, id.name, NULL); 811 a++; 812 } 813 814 return ret; 815 } 816 817 /* internal sources muxes *********************************************/ 818 819 static int alchemy_clk_csrc_isen(struct clk_hw *hw) 820 { 821 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 822 unsigned long v = alchemy_rdsys(c->reg); 823 824 return (((v >> c->shift) >> 2) & 7) != 0; 825 } 826 827 static void __alchemy_clk_csrc_en(struct alchemy_fgcs_clk *c) 828 { 829 unsigned long v = alchemy_rdsys(c->reg); 830 831 v &= ~((7 << 2) << c->shift); 832 v |= ((c->parent & 7) << 2) << c->shift; 833 alchemy_wrsys(v, c->reg); 834 c->isen = 1; 835 } 836 837 static int alchemy_clk_csrc_en(struct clk_hw *hw) 838 { 839 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 840 unsigned long flags; 841 842 /* enable by setting the previous parent clock */ 843 spin_lock_irqsave(c->reglock, flags); 844 __alchemy_clk_csrc_en(c); 845 spin_unlock_irqrestore(c->reglock, flags); 846 847 return 0; 848 } 849 850 static void alchemy_clk_csrc_dis(struct clk_hw *hw) 851 { 852 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 853 unsigned long v, flags; 854 855 spin_lock_irqsave(c->reglock, flags); 856 v = alchemy_rdsys(c->reg); 857 v &= ~((3 << 2) << c->shift); /* mux to "disabled" state */ 858 alchemy_wrsys(v, c->reg); 859 c->isen = 0; 860 spin_unlock_irqrestore(c->reglock, flags); 861 } 862 863 static int alchemy_clk_csrc_setp(struct clk_hw *hw, u8 index) 864 { 865 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 866 unsigned long flags; 867 868 spin_lock_irqsave(c->reglock, flags); 869 c->parent = index + 1; /* value to write to register */ 870 if (c->isen) 871 __alchemy_clk_csrc_en(c); 872 spin_unlock_irqrestore(c->reglock, flags); 873 874 return 0; 875 } 876 877 static u8 alchemy_clk_csrc_getp(struct clk_hw *hw) 878 { 879 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 880 881 return c->parent - 1; 882 } 883 884 static unsigned long alchemy_clk_csrc_recalc(struct clk_hw *hw, 885 unsigned long parent_rate) 886 { 887 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 888 unsigned long v = (alchemy_rdsys(c->reg) >> c->shift) & 3; 889 890 return parent_rate / c->dt[v]; 891 } 892 893 static int alchemy_clk_csrc_setr(struct clk_hw *hw, unsigned long rate, 894 unsigned long parent_rate) 895 { 896 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 897 unsigned long d, v, flags; 898 int i; 899 900 if (!rate || !parent_rate || rate > parent_rate) 901 return -EINVAL; 902 903 d = (parent_rate + (rate / 2)) / rate; 904 if (d > 4) 905 return -EINVAL; 906 if ((d == 3) && (c->dt[2] != 3)) 907 d = 4; 908 909 for (i = 0; i < 4; i++) 910 if (c->dt[i] == d) 911 break; 912 913 if (i >= 4) 914 return -EINVAL; /* oops */ 915 916 spin_lock_irqsave(c->reglock, flags); 917 v = alchemy_rdsys(c->reg); 918 v &= ~(3 << c->shift); 919 v |= (i & 3) << c->shift; 920 alchemy_wrsys(v, c->reg); 921 spin_unlock_irqrestore(c->reglock, flags); 922 923 return 0; 924 } 925 926 static int alchemy_clk_csrc_detr(struct clk_hw *hw, 927 struct clk_rate_request *req) 928 { 929 struct alchemy_fgcs_clk *c = to_fgcs_clk(hw); 930 int scale = c->dt[2] == 3 ? 1 : 2; /* au1300 check */ 931 932 return alchemy_clk_fgcs_detr(hw, req, scale, 4); 933 } 934 935 static const struct clk_ops alchemy_clkops_csrc = { 936 .recalc_rate = alchemy_clk_csrc_recalc, 937 .determine_rate = alchemy_clk_csrc_detr, 938 .set_rate = alchemy_clk_csrc_setr, 939 .set_parent = alchemy_clk_csrc_setp, 940 .get_parent = alchemy_clk_csrc_getp, 941 .enable = alchemy_clk_csrc_en, 942 .disable = alchemy_clk_csrc_dis, 943 .is_enabled = alchemy_clk_csrc_isen, 944 }; 945 946 static const char * const alchemy_clk_csrc_parents[] = { 947 /* disabled at index 0 */ ALCHEMY_AUXPLL_CLK, 948 ALCHEMY_FG0_CLK, ALCHEMY_FG1_CLK, ALCHEMY_FG2_CLK, 949 ALCHEMY_FG3_CLK, ALCHEMY_FG4_CLK, ALCHEMY_FG5_CLK 950 }; 951 952 /* divider tables */ 953 static int alchemy_csrc_dt1[] = { 1, 4, 1, 2 }; /* rest */ 954 static int alchemy_csrc_dt2[] = { 1, 4, 3, 2 }; /* Au1300 */ 955 956 static int __init alchemy_clk_setup_imux(int ctype) 957 { 958 struct alchemy_fgcs_clk *a; 959 const char * const *names; 960 struct clk_init_data id; 961 unsigned long v; 962 int i, ret, *dt; 963 struct clk *c; 964 965 id.ops = &alchemy_clkops_csrc; 966 id.parent_names = alchemy_clk_csrc_parents; 967 id.num_parents = 7; 968 id.flags = CLK_SET_RATE_PARENT | CLK_GET_RATE_NOCACHE; 969 970 dt = alchemy_csrc_dt1; 971 switch (ctype) { 972 case ALCHEMY_CPU_AU1000: 973 names = alchemy_au1000_intclknames; 974 break; 975 case ALCHEMY_CPU_AU1500: 976 names = alchemy_au1500_intclknames; 977 break; 978 case ALCHEMY_CPU_AU1100: 979 names = alchemy_au1100_intclknames; 980 break; 981 case ALCHEMY_CPU_AU1550: 982 names = alchemy_au1550_intclknames; 983 break; 984 case ALCHEMY_CPU_AU1200: 985 names = alchemy_au1200_intclknames; 986 break; 987 case ALCHEMY_CPU_AU1300: 988 dt = alchemy_csrc_dt2; 989 names = alchemy_au1300_intclknames; 990 break; 991 default: 992 return -ENODEV; 993 } 994 995 a = kcalloc(6, sizeof(*a), GFP_KERNEL); 996 if (!a) 997 return -ENOMEM; 998 999 spin_lock_init(&alchemy_clk_csrc_lock); 1000 ret = 0; 1001 1002 for (i = 0; i < 6; i++) { 1003 id.name = names[i]; 1004 if (!id.name) 1005 goto next; 1006 1007 a->shift = i * 5; 1008 a->reg = AU1000_SYS_CLKSRC; 1009 a->reglock = &alchemy_clk_csrc_lock; 1010 a->dt = dt; 1011 1012 /* default to first parent clock if mux is initially 1013 * set to disabled state. 1014 */ 1015 v = alchemy_rdsys(a->reg); 1016 a->parent = ((v >> a->shift) >> 2) & 7; 1017 if (!a->parent) { 1018 a->parent = 1; 1019 a->isen = 0; 1020 } else 1021 a->isen = 1; 1022 1023 a->hw.init = &id; 1024 c = clk_register(NULL, &a->hw); 1025 if (IS_ERR(c)) 1026 ret++; 1027 else 1028 clk_register_clkdev(c, id.name, NULL); 1029 next: 1030 a++; 1031 } 1032 1033 return ret; 1034 } 1035 1036 1037 /**********************************************************************/ 1038 1039 1040 #define ERRCK(x) \ 1041 if (IS_ERR(x)) { \ 1042 ret = PTR_ERR(x); \ 1043 goto out; \ 1044 } 1045 1046 static int __init alchemy_clk_init(void) 1047 { 1048 int ctype = alchemy_get_cputype(), ret, i; 1049 struct clk_aliastable *t = alchemy_clk_aliases; 1050 struct clk *c; 1051 1052 /* Root of the Alchemy clock tree: external 12MHz crystal osc */ 1053 c = clk_register_fixed_rate(NULL, ALCHEMY_ROOT_CLK, NULL, 1054 0, ALCHEMY_ROOTCLK_RATE); 1055 ERRCK(c) 1056 1057 /* CPU core clock */ 1058 c = alchemy_clk_setup_cpu(ALCHEMY_ROOT_CLK, ctype); 1059 ERRCK(c) 1060 1061 /* AUXPLLs: max 1GHz on Au1300, 748MHz on older models */ 1062 i = (ctype == ALCHEMY_CPU_AU1300) ? 84 : 63; 1063 c = alchemy_clk_setup_aux(ALCHEMY_ROOT_CLK, ALCHEMY_AUXPLL_CLK, 1064 i, AU1000_SYS_AUXPLL); 1065 ERRCK(c) 1066 1067 if (ctype == ALCHEMY_CPU_AU1300) { 1068 c = alchemy_clk_setup_aux(ALCHEMY_ROOT_CLK, 1069 ALCHEMY_AUXPLL2_CLK, i, 1070 AU1300_SYS_AUXPLL2); 1071 ERRCK(c) 1072 } 1073 1074 /* sysbus clock: cpu core clock divided by 2, 3 or 4 */ 1075 c = alchemy_clk_setup_sysbus(ALCHEMY_CPU_CLK); 1076 ERRCK(c) 1077 1078 /* peripheral clock: runs at half rate of sysbus clk */ 1079 c = alchemy_clk_setup_periph(ALCHEMY_SYSBUS_CLK); 1080 ERRCK(c) 1081 1082 /* SDR/DDR memory clock */ 1083 c = alchemy_clk_setup_mem(ALCHEMY_SYSBUS_CLK, ctype); 1084 ERRCK(c) 1085 1086 /* L/RCLK: external static bus clock for synchronous mode */ 1087 c = alchemy_clk_setup_lrclk(ALCHEMY_PERIPH_CLK, ctype); 1088 ERRCK(c) 1089 1090 /* Frequency dividers 0-5 */ 1091 ret = alchemy_clk_init_fgens(ctype); 1092 if (ret) { 1093 ret = -ENODEV; 1094 goto out; 1095 } 1096 1097 /* diving muxes for internal sources */ 1098 ret = alchemy_clk_setup_imux(ctype); 1099 if (ret) { 1100 ret = -ENODEV; 1101 goto out; 1102 } 1103 1104 /* set up aliases drivers might look for */ 1105 while (t->base) { 1106 if (t->cputype == ctype) 1107 clk_add_alias(t->alias, NULL, t->base, NULL); 1108 t++; 1109 } 1110 1111 pr_info("Alchemy clocktree installed\n"); 1112 return 0; 1113 1114 out: 1115 return ret; 1116 } 1117 postcore_initcall(alchemy_clk_init); 1118