1 /* SPDX-License-Identifier: GPL-2.0-only */ 2 /* 3 * linux/include/linux/clk.h 4 * 5 * Copyright (C) 2004 ARM Limited. 6 * Written by Deep Blue Solutions Limited. 7 * Copyright (C) 2011-2012 Linaro Ltd <mturquette@linaro.org> 8 */ 9 #ifndef __LINUX_CLK_H 10 #define __LINUX_CLK_H 11 12 #include <linux/err.h> 13 #include <linux/kernel.h> 14 #include <linux/notifier.h> 15 16 struct device; 17 struct clk; 18 struct device_node; 19 struct of_phandle_args; 20 21 /** 22 * DOC: clk notifier callback types 23 * 24 * PRE_RATE_CHANGE - called immediately before the clk rate is changed, 25 * to indicate that the rate change will proceed. Drivers must 26 * immediately terminate any operations that will be affected by the 27 * rate change. Callbacks may either return NOTIFY_DONE, NOTIFY_OK, 28 * NOTIFY_STOP or NOTIFY_BAD. 29 * 30 * ABORT_RATE_CHANGE: called if the rate change failed for some reason 31 * after PRE_RATE_CHANGE. In this case, all registered notifiers on 32 * the clk will be called with ABORT_RATE_CHANGE. Callbacks must 33 * always return NOTIFY_DONE or NOTIFY_OK. 34 * 35 * POST_RATE_CHANGE - called after the clk rate change has successfully 36 * completed. Callbacks must always return NOTIFY_DONE or NOTIFY_OK. 37 * 38 */ 39 #define PRE_RATE_CHANGE BIT(0) 40 #define POST_RATE_CHANGE BIT(1) 41 #define ABORT_RATE_CHANGE BIT(2) 42 43 /** 44 * struct clk_notifier - associate a clk with a notifier 45 * @clk: struct clk * to associate the notifier with 46 * @notifier_head: a blocking_notifier_head for this clk 47 * @node: linked list pointers 48 * 49 * A list of struct clk_notifier is maintained by the notifier code. 50 * An entry is created whenever code registers the first notifier on a 51 * particular @clk. Future notifiers on that @clk are added to the 52 * @notifier_head. 53 */ 54 struct clk_notifier { 55 struct clk *clk; 56 struct srcu_notifier_head notifier_head; 57 struct list_head node; 58 }; 59 60 /** 61 * struct clk_notifier_data - rate data to pass to the notifier callback 62 * @clk: struct clk * being changed 63 * @old_rate: previous rate of this clk 64 * @new_rate: new rate of this clk 65 * 66 * For a pre-notifier, old_rate is the clk's rate before this rate 67 * change, and new_rate is what the rate will be in the future. For a 68 * post-notifier, old_rate and new_rate are both set to the clk's 69 * current rate (this was done to optimize the implementation). 70 */ 71 struct clk_notifier_data { 72 struct clk *clk; 73 unsigned long old_rate; 74 unsigned long new_rate; 75 }; 76 77 /** 78 * struct clk_bulk_data - Data used for bulk clk operations. 79 * 80 * @id: clock consumer ID 81 * @clk: struct clk * to store the associated clock 82 * 83 * The CLK APIs provide a series of clk_bulk_() API calls as 84 * a convenience to consumers which require multiple clks. This 85 * structure is used to manage data for these calls. 86 */ 87 struct clk_bulk_data { 88 const char *id; 89 struct clk *clk; 90 }; 91 92 #ifdef CONFIG_COMMON_CLK 93 94 /** 95 * clk_notifier_register - register a clock rate-change notifier callback 96 * @clk: clock whose rate we are interested in 97 * @nb: notifier block with callback function pointer 98 * 99 * ProTip: debugging across notifier chains can be frustrating. Make sure that 100 * your notifier callback function prints a nice big warning in case of 101 * failure. 102 */ 103 int clk_notifier_register(struct clk *clk, struct notifier_block *nb); 104 105 /** 106 * clk_notifier_unregister - unregister a clock rate-change notifier callback 107 * @clk: clock whose rate we are no longer interested in 108 * @nb: notifier block which will be unregistered 109 */ 110 int clk_notifier_unregister(struct clk *clk, struct notifier_block *nb); 111 112 /** 113 * devm_clk_notifier_register - register a managed rate-change notifier callback 114 * @dev: device for clock "consumer" 115 * @clk: clock whose rate we are interested in 116 * @nb: notifier block with callback function pointer 117 * 118 * Returns 0 on success, -EERROR otherwise 119 */ 120 int devm_clk_notifier_register(struct device *dev, struct clk *clk, 121 struct notifier_block *nb); 122 123 /** 124 * clk_get_accuracy - obtain the clock accuracy in ppb (parts per billion) 125 * for a clock source. 126 * @clk: clock source 127 * 128 * This gets the clock source accuracy expressed in ppb. 129 * A perfect clock returns 0. 130 */ 131 long clk_get_accuracy(struct clk *clk); 132 133 /** 134 * clk_set_phase - adjust the phase shift of a clock signal 135 * @clk: clock signal source 136 * @degrees: number of degrees the signal is shifted 137 * 138 * Shifts the phase of a clock signal by the specified degrees. Returns 0 on 139 * success, -EERROR otherwise. 140 */ 141 int clk_set_phase(struct clk *clk, int degrees); 142 143 /** 144 * clk_get_phase - return the phase shift of a clock signal 145 * @clk: clock signal source 146 * 147 * Returns the phase shift of a clock node in degrees, otherwise returns 148 * -EERROR. 149 */ 150 int clk_get_phase(struct clk *clk); 151 152 /** 153 * clk_set_duty_cycle - adjust the duty cycle ratio of a clock signal 154 * @clk: clock signal source 155 * @num: numerator of the duty cycle ratio to be applied 156 * @den: denominator of the duty cycle ratio to be applied 157 * 158 * Adjust the duty cycle of a clock signal by the specified ratio. Returns 0 on 159 * success, -EERROR otherwise. 160 */ 161 int clk_set_duty_cycle(struct clk *clk, unsigned int num, unsigned int den); 162 163 /** 164 * clk_get_scaled_duty_cycle - return the duty cycle ratio of a clock signal 165 * @clk: clock signal source 166 * @scale: scaling factor to be applied to represent the ratio as an integer 167 * 168 * Returns the duty cycle ratio multiplied by the scale provided, otherwise 169 * returns -EERROR. 170 */ 171 int clk_get_scaled_duty_cycle(struct clk *clk, unsigned int scale); 172 173 /** 174 * clk_is_match - check if two clk's point to the same hardware clock 175 * @p: clk compared against q 176 * @q: clk compared against p 177 * 178 * Returns true if the two struct clk pointers both point to the same hardware 179 * clock node. Put differently, returns true if @p and @q 180 * share the same &struct clk_core object. 181 * 182 * Returns false otherwise. Note that two NULL clks are treated as matching. 183 */ 184 bool clk_is_match(const struct clk *p, const struct clk *q); 185 186 #else 187 188 static inline int clk_notifier_register(struct clk *clk, 189 struct notifier_block *nb) 190 { 191 return -ENOTSUPP; 192 } 193 194 static inline int clk_notifier_unregister(struct clk *clk, 195 struct notifier_block *nb) 196 { 197 return -ENOTSUPP; 198 } 199 200 static inline int devm_clk_notifier_register(struct device *dev, 201 struct clk *clk, 202 struct notifier_block *nb) 203 { 204 return -ENOTSUPP; 205 } 206 207 static inline long clk_get_accuracy(struct clk *clk) 208 { 209 return -ENOTSUPP; 210 } 211 212 static inline long clk_set_phase(struct clk *clk, int phase) 213 { 214 return -ENOTSUPP; 215 } 216 217 static inline long clk_get_phase(struct clk *clk) 218 { 219 return -ENOTSUPP; 220 } 221 222 static inline int clk_set_duty_cycle(struct clk *clk, unsigned int num, 223 unsigned int den) 224 { 225 return -ENOTSUPP; 226 } 227 228 static inline unsigned int clk_get_scaled_duty_cycle(struct clk *clk, 229 unsigned int scale) 230 { 231 return 0; 232 } 233 234 static inline bool clk_is_match(const struct clk *p, const struct clk *q) 235 { 236 return p == q; 237 } 238 239 #endif 240 241 #ifdef CONFIG_HAVE_CLK_PREPARE 242 /** 243 * clk_prepare - prepare a clock source 244 * @clk: clock source 245 * 246 * This prepares the clock source for use. 247 * 248 * Must not be called from within atomic context. 249 */ 250 int clk_prepare(struct clk *clk); 251 int __must_check clk_bulk_prepare(int num_clks, 252 const struct clk_bulk_data *clks); 253 254 /** 255 * clk_is_enabled_when_prepared - indicate if preparing a clock also enables it. 256 * @clk: clock source 257 * 258 * Returns true if clk_prepare() implicitly enables the clock, effectively 259 * making clk_enable()/clk_disable() no-ops, false otherwise. 260 * 261 * This is of interest mainly to the power management code where actually 262 * disabling the clock also requires unpreparing it to have any material 263 * effect. 264 * 265 * Regardless of the value returned here, the caller must always invoke 266 * clk_enable() or clk_prepare_enable() and counterparts for usage counts 267 * to be right. 268 */ 269 bool clk_is_enabled_when_prepared(struct clk *clk); 270 #else 271 static inline int clk_prepare(struct clk *clk) 272 { 273 might_sleep(); 274 return 0; 275 } 276 277 static inline int __must_check 278 clk_bulk_prepare(int num_clks, const struct clk_bulk_data *clks) 279 { 280 might_sleep(); 281 return 0; 282 } 283 284 static inline bool clk_is_enabled_when_prepared(struct clk *clk) 285 { 286 return false; 287 } 288 #endif 289 290 /** 291 * clk_unprepare - undo preparation of a clock source 292 * @clk: clock source 293 * 294 * This undoes a previously prepared clock. The caller must balance 295 * the number of prepare and unprepare calls. 296 * 297 * Must not be called from within atomic context. 298 */ 299 #ifdef CONFIG_HAVE_CLK_PREPARE 300 void clk_unprepare(struct clk *clk); 301 void clk_bulk_unprepare(int num_clks, const struct clk_bulk_data *clks); 302 #else 303 static inline void clk_unprepare(struct clk *clk) 304 { 305 might_sleep(); 306 } 307 static inline void clk_bulk_unprepare(int num_clks, 308 const struct clk_bulk_data *clks) 309 { 310 might_sleep(); 311 } 312 #endif 313 314 #ifdef CONFIG_HAVE_CLK 315 /** 316 * clk_get - lookup and obtain a reference to a clock producer. 317 * @dev: device for clock "consumer" 318 * @id: clock consumer ID 319 * 320 * Returns a struct clk corresponding to the clock producer, or 321 * valid IS_ERR() condition containing errno. The implementation 322 * uses @dev and @id to determine the clock consumer, and thereby 323 * the clock producer. (IOW, @id may be identical strings, but 324 * clk_get may return different clock producers depending on @dev.) 325 * 326 * Drivers must assume that the clock source is not enabled. 327 * 328 * clk_get should not be called from within interrupt context. 329 */ 330 struct clk *clk_get(struct device *dev, const char *id); 331 332 /** 333 * clk_bulk_get - lookup and obtain a number of references to clock producer. 334 * @dev: device for clock "consumer" 335 * @num_clks: the number of clk_bulk_data 336 * @clks: the clk_bulk_data table of consumer 337 * 338 * This helper function allows drivers to get several clk consumers in one 339 * operation. If any of the clk cannot be acquired then any clks 340 * that were obtained will be freed before returning to the caller. 341 * 342 * Returns 0 if all clocks specified in clk_bulk_data table are obtained 343 * successfully, or valid IS_ERR() condition containing errno. 344 * The implementation uses @dev and @clk_bulk_data.id to determine the 345 * clock consumer, and thereby the clock producer. 346 * The clock returned is stored in each @clk_bulk_data.clk field. 347 * 348 * Drivers must assume that the clock source is not enabled. 349 * 350 * clk_bulk_get should not be called from within interrupt context. 351 */ 352 int __must_check clk_bulk_get(struct device *dev, int num_clks, 353 struct clk_bulk_data *clks); 354 /** 355 * clk_bulk_get_all - lookup and obtain all available references to clock 356 * producer. 357 * @dev: device for clock "consumer" 358 * @clks: pointer to the clk_bulk_data table of consumer 359 * 360 * This helper function allows drivers to get all clk consumers in one 361 * operation. If any of the clk cannot be acquired then any clks 362 * that were obtained will be freed before returning to the caller. 363 * 364 * Returns a positive value for the number of clocks obtained while the 365 * clock references are stored in the clk_bulk_data table in @clks field. 366 * Returns 0 if there're none and a negative value if something failed. 367 * 368 * Drivers must assume that the clock source is not enabled. 369 * 370 * clk_bulk_get should not be called from within interrupt context. 371 */ 372 int __must_check clk_bulk_get_all(struct device *dev, 373 struct clk_bulk_data **clks); 374 375 /** 376 * clk_bulk_get_optional - lookup and obtain a number of references to clock producer 377 * @dev: device for clock "consumer" 378 * @num_clks: the number of clk_bulk_data 379 * @clks: the clk_bulk_data table of consumer 380 * 381 * Behaves the same as clk_bulk_get() except where there is no clock producer. 382 * In this case, instead of returning -ENOENT, the function returns 0 and 383 * NULL for a clk for which a clock producer could not be determined. 384 */ 385 int __must_check clk_bulk_get_optional(struct device *dev, int num_clks, 386 struct clk_bulk_data *clks); 387 /** 388 * devm_clk_bulk_get - managed get multiple clk consumers 389 * @dev: device for clock "consumer" 390 * @num_clks: the number of clk_bulk_data 391 * @clks: the clk_bulk_data table of consumer 392 * 393 * Return 0 on success, an errno on failure. 394 * 395 * This helper function allows drivers to get several clk 396 * consumers in one operation with management, the clks will 397 * automatically be freed when the device is unbound. 398 */ 399 int __must_check devm_clk_bulk_get(struct device *dev, int num_clks, 400 struct clk_bulk_data *clks); 401 /** 402 * devm_clk_bulk_get_optional - managed get multiple optional consumer clocks 403 * @dev: device for clock "consumer" 404 * @num_clks: the number of clk_bulk_data 405 * @clks: pointer to the clk_bulk_data table of consumer 406 * 407 * Behaves the same as devm_clk_bulk_get() except where there is no clock 408 * producer. In this case, instead of returning -ENOENT, the function returns 409 * NULL for given clk. It is assumed all clocks in clk_bulk_data are optional. 410 * 411 * Returns 0 if all clocks specified in clk_bulk_data table are obtained 412 * successfully or for any clk there was no clk provider available, otherwise 413 * returns valid IS_ERR() condition containing errno. 414 * The implementation uses @dev and @clk_bulk_data.id to determine the 415 * clock consumer, and thereby the clock producer. 416 * The clock returned is stored in each @clk_bulk_data.clk field. 417 * 418 * Drivers must assume that the clock source is not enabled. 419 * 420 * clk_bulk_get should not be called from within interrupt context. 421 */ 422 int __must_check devm_clk_bulk_get_optional(struct device *dev, int num_clks, 423 struct clk_bulk_data *clks); 424 /** 425 * devm_clk_bulk_get_all - managed get multiple clk consumers 426 * @dev: device for clock "consumer" 427 * @clks: pointer to the clk_bulk_data table of consumer 428 * 429 * Returns a positive value for the number of clocks obtained while the 430 * clock references are stored in the clk_bulk_data table in @clks field. 431 * Returns 0 if there're none and a negative value if something failed. 432 * 433 * This helper function allows drivers to get several clk 434 * consumers in one operation with management, the clks will 435 * automatically be freed when the device is unbound. 436 */ 437 438 int __must_check devm_clk_bulk_get_all(struct device *dev, 439 struct clk_bulk_data **clks); 440 441 /** 442 * devm_clk_get - lookup and obtain a managed reference to a clock producer. 443 * @dev: device for clock "consumer" 444 * @id: clock consumer ID 445 * 446 * Context: May sleep. 447 * 448 * Return: a struct clk corresponding to the clock producer, or 449 * valid IS_ERR() condition containing errno. The implementation 450 * uses @dev and @id to determine the clock consumer, and thereby 451 * the clock producer. (IOW, @id may be identical strings, but 452 * clk_get may return different clock producers depending on @dev.) 453 * 454 * Drivers must assume that the clock source is neither prepared nor 455 * enabled. 456 * 457 * The clock will automatically be freed when the device is unbound 458 * from the bus. 459 */ 460 struct clk *devm_clk_get(struct device *dev, const char *id); 461 462 /** 463 * devm_clk_get_prepared - devm_clk_get() + clk_prepare() 464 * @dev: device for clock "consumer" 465 * @id: clock consumer ID 466 * 467 * Context: May sleep. 468 * 469 * Return: a struct clk corresponding to the clock producer, or 470 * valid IS_ERR() condition containing errno. The implementation 471 * uses @dev and @id to determine the clock consumer, and thereby 472 * the clock producer. (IOW, @id may be identical strings, but 473 * clk_get may return different clock producers depending on @dev.) 474 * 475 * The returned clk (if valid) is prepared. Drivers must however assume 476 * that the clock is not enabled. 477 * 478 * The clock will automatically be unprepared and freed when the device 479 * is unbound from the bus. 480 */ 481 struct clk *devm_clk_get_prepared(struct device *dev, const char *id); 482 483 /** 484 * devm_clk_get_enabled - devm_clk_get() + clk_prepare_enable() 485 * @dev: device for clock "consumer" 486 * @id: clock consumer ID 487 * 488 * Context: May sleep. 489 * 490 * Return: a struct clk corresponding to the clock producer, or 491 * valid IS_ERR() condition containing errno. The implementation 492 * uses @dev and @id to determine the clock consumer, and thereby 493 * the clock producer. (IOW, @id may be identical strings, but 494 * clk_get may return different clock producers depending on @dev.) 495 * 496 * The returned clk (if valid) is prepared and enabled. 497 * 498 * The clock will automatically be disabled, unprepared and freed 499 * when the device is unbound from the bus. 500 */ 501 struct clk *devm_clk_get_enabled(struct device *dev, const char *id); 502 503 /** 504 * devm_clk_get_optional - lookup and obtain a managed reference to an optional 505 * clock producer. 506 * @dev: device for clock "consumer" 507 * @id: clock consumer ID 508 * 509 * Context: May sleep. 510 * 511 * Return: a struct clk corresponding to the clock producer, or 512 * valid IS_ERR() condition containing errno. The implementation 513 * uses @dev and @id to determine the clock consumer, and thereby 514 * the clock producer. If no such clk is found, it returns NULL 515 * which serves as a dummy clk. That's the only difference compared 516 * to devm_clk_get(). 517 * 518 * Drivers must assume that the clock source is neither prepared nor 519 * enabled. 520 * 521 * The clock will automatically be freed when the device is unbound 522 * from the bus. 523 */ 524 struct clk *devm_clk_get_optional(struct device *dev, const char *id); 525 526 /** 527 * devm_clk_get_optional_prepared - devm_clk_get_optional() + clk_prepare() 528 * @dev: device for clock "consumer" 529 * @id: clock consumer ID 530 * 531 * Context: May sleep. 532 * 533 * Return: a struct clk corresponding to the clock producer, or 534 * valid IS_ERR() condition containing errno. The implementation 535 * uses @dev and @id to determine the clock consumer, and thereby 536 * the clock producer. If no such clk is found, it returns NULL 537 * which serves as a dummy clk. That's the only difference compared 538 * to devm_clk_get_prepared(). 539 * 540 * The returned clk (if valid) is prepared. Drivers must however 541 * assume that the clock is not enabled. 542 * 543 * The clock will automatically be unprepared and freed when the 544 * device is unbound from the bus. 545 */ 546 struct clk *devm_clk_get_optional_prepared(struct device *dev, const char *id); 547 548 /** 549 * devm_clk_get_optional_enabled - devm_clk_get_optional() + 550 * clk_prepare_enable() 551 * @dev: device for clock "consumer" 552 * @id: clock consumer ID 553 * 554 * Context: May sleep. 555 * 556 * Return: a struct clk corresponding to the clock producer, or 557 * valid IS_ERR() condition containing errno. The implementation 558 * uses @dev and @id to determine the clock consumer, and thereby 559 * the clock producer. If no such clk is found, it returns NULL 560 * which serves as a dummy clk. That's the only difference compared 561 * to devm_clk_get_enabled(). 562 * 563 * The returned clk (if valid) is prepared and enabled. 564 * 565 * The clock will automatically be disabled, unprepared and freed 566 * when the device is unbound from the bus. 567 */ 568 struct clk *devm_clk_get_optional_enabled(struct device *dev, const char *id); 569 570 /** 571 * devm_get_clk_from_child - lookup and obtain a managed reference to a 572 * clock producer from child node. 573 * @dev: device for clock "consumer" 574 * @np: pointer to clock consumer node 575 * @con_id: clock consumer ID 576 * 577 * This function parses the clocks, and uses them to look up the 578 * struct clk from the registered list of clock providers by using 579 * @np and @con_id 580 * 581 * The clock will automatically be freed when the device is unbound 582 * from the bus. 583 */ 584 struct clk *devm_get_clk_from_child(struct device *dev, 585 struct device_node *np, const char *con_id); 586 /** 587 * clk_rate_exclusive_get - get exclusivity over the rate control of a 588 * producer 589 * @clk: clock source 590 * 591 * This function allows drivers to get exclusive control over the rate of a 592 * provider. It prevents any other consumer to execute, even indirectly, 593 * opereation which could alter the rate of the provider or cause glitches 594 * 595 * If exlusivity is claimed more than once on clock, even by the same driver, 596 * the rate effectively gets locked as exclusivity can't be preempted. 597 * 598 * Must not be called from within atomic context. 599 * 600 * Returns success (0) or negative errno. 601 */ 602 int clk_rate_exclusive_get(struct clk *clk); 603 604 /** 605 * clk_rate_exclusive_put - release exclusivity over the rate control of a 606 * producer 607 * @clk: clock source 608 * 609 * This function allows drivers to release the exclusivity it previously got 610 * from clk_rate_exclusive_get() 611 * 612 * The caller must balance the number of clk_rate_exclusive_get() and 613 * clk_rate_exclusive_put() calls. 614 * 615 * Must not be called from within atomic context. 616 */ 617 void clk_rate_exclusive_put(struct clk *clk); 618 619 /** 620 * clk_enable - inform the system when the clock source should be running. 621 * @clk: clock source 622 * 623 * If the clock can not be enabled/disabled, this should return success. 624 * 625 * May be called from atomic contexts. 626 * 627 * Returns success (0) or negative errno. 628 */ 629 int clk_enable(struct clk *clk); 630 631 /** 632 * clk_bulk_enable - inform the system when the set of clks should be running. 633 * @num_clks: the number of clk_bulk_data 634 * @clks: the clk_bulk_data table of consumer 635 * 636 * May be called from atomic contexts. 637 * 638 * Returns success (0) or negative errno. 639 */ 640 int __must_check clk_bulk_enable(int num_clks, 641 const struct clk_bulk_data *clks); 642 643 /** 644 * clk_disable - inform the system when the clock source is no longer required. 645 * @clk: clock source 646 * 647 * Inform the system that a clock source is no longer required by 648 * a driver and may be shut down. 649 * 650 * May be called from atomic contexts. 651 * 652 * Implementation detail: if the clock source is shared between 653 * multiple drivers, clk_enable() calls must be balanced by the 654 * same number of clk_disable() calls for the clock source to be 655 * disabled. 656 */ 657 void clk_disable(struct clk *clk); 658 659 /** 660 * clk_bulk_disable - inform the system when the set of clks is no 661 * longer required. 662 * @num_clks: the number of clk_bulk_data 663 * @clks: the clk_bulk_data table of consumer 664 * 665 * Inform the system that a set of clks is no longer required by 666 * a driver and may be shut down. 667 * 668 * May be called from atomic contexts. 669 * 670 * Implementation detail: if the set of clks is shared between 671 * multiple drivers, clk_bulk_enable() calls must be balanced by the 672 * same number of clk_bulk_disable() calls for the clock source to be 673 * disabled. 674 */ 675 void clk_bulk_disable(int num_clks, const struct clk_bulk_data *clks); 676 677 /** 678 * clk_get_rate - obtain the current clock rate (in Hz) for a clock source. 679 * This is only valid once the clock source has been enabled. 680 * @clk: clock source 681 */ 682 unsigned long clk_get_rate(struct clk *clk); 683 684 /** 685 * clk_put - "free" the clock source 686 * @clk: clock source 687 * 688 * Note: drivers must ensure that all clk_enable calls made on this 689 * clock source are balanced by clk_disable calls prior to calling 690 * this function. 691 * 692 * clk_put should not be called from within interrupt context. 693 */ 694 void clk_put(struct clk *clk); 695 696 /** 697 * clk_bulk_put - "free" the clock source 698 * @num_clks: the number of clk_bulk_data 699 * @clks: the clk_bulk_data table of consumer 700 * 701 * Note: drivers must ensure that all clk_bulk_enable calls made on this 702 * clock source are balanced by clk_bulk_disable calls prior to calling 703 * this function. 704 * 705 * clk_bulk_put should not be called from within interrupt context. 706 */ 707 void clk_bulk_put(int num_clks, struct clk_bulk_data *clks); 708 709 /** 710 * clk_bulk_put_all - "free" all the clock source 711 * @num_clks: the number of clk_bulk_data 712 * @clks: the clk_bulk_data table of consumer 713 * 714 * Note: drivers must ensure that all clk_bulk_enable calls made on this 715 * clock source are balanced by clk_bulk_disable calls prior to calling 716 * this function. 717 * 718 * clk_bulk_put_all should not be called from within interrupt context. 719 */ 720 void clk_bulk_put_all(int num_clks, struct clk_bulk_data *clks); 721 722 /** 723 * devm_clk_put - "free" a managed clock source 724 * @dev: device used to acquire the clock 725 * @clk: clock source acquired with devm_clk_get() 726 * 727 * Note: drivers must ensure that all clk_enable calls made on this 728 * clock source are balanced by clk_disable calls prior to calling 729 * this function. 730 * 731 * clk_put should not be called from within interrupt context. 732 */ 733 void devm_clk_put(struct device *dev, struct clk *clk); 734 735 /* 736 * The remaining APIs are optional for machine class support. 737 */ 738 739 740 /** 741 * clk_round_rate - adjust a rate to the exact rate a clock can provide 742 * @clk: clock source 743 * @rate: desired clock rate in Hz 744 * 745 * This answers the question "if I were to pass @rate to clk_set_rate(), 746 * what clock rate would I end up with?" without changing the hardware 747 * in any way. In other words: 748 * 749 * rate = clk_round_rate(clk, r); 750 * 751 * and: 752 * 753 * clk_set_rate(clk, r); 754 * rate = clk_get_rate(clk); 755 * 756 * are equivalent except the former does not modify the clock hardware 757 * in any way. 758 * 759 * Returns rounded clock rate in Hz, or negative errno. 760 */ 761 long clk_round_rate(struct clk *clk, unsigned long rate); 762 763 /** 764 * clk_set_rate - set the clock rate for a clock source 765 * @clk: clock source 766 * @rate: desired clock rate in Hz 767 * 768 * Updating the rate starts at the top-most affected clock and then 769 * walks the tree down to the bottom-most clock that needs updating. 770 * 771 * Returns success (0) or negative errno. 772 */ 773 int clk_set_rate(struct clk *clk, unsigned long rate); 774 775 /** 776 * clk_set_rate_exclusive- set the clock rate and claim exclusivity over 777 * clock source 778 * @clk: clock source 779 * @rate: desired clock rate in Hz 780 * 781 * This helper function allows drivers to atomically set the rate of a producer 782 * and claim exclusivity over the rate control of the producer. 783 * 784 * It is essentially a combination of clk_set_rate() and 785 * clk_rate_exclusite_get(). Caller must balance this call with a call to 786 * clk_rate_exclusive_put() 787 * 788 * Returns success (0) or negative errno. 789 */ 790 int clk_set_rate_exclusive(struct clk *clk, unsigned long rate); 791 792 /** 793 * clk_has_parent - check if a clock is a possible parent for another 794 * @clk: clock source 795 * @parent: parent clock source 796 * 797 * This function can be used in drivers that need to check that a clock can be 798 * the parent of another without actually changing the parent. 799 * 800 * Returns true if @parent is a possible parent for @clk, false otherwise. 801 */ 802 bool clk_has_parent(const struct clk *clk, const struct clk *parent); 803 804 /** 805 * clk_set_rate_range - set a rate range for a clock source 806 * @clk: clock source 807 * @min: desired minimum clock rate in Hz, inclusive 808 * @max: desired maximum clock rate in Hz, inclusive 809 * 810 * Returns success (0) or negative errno. 811 */ 812 int clk_set_rate_range(struct clk *clk, unsigned long min, unsigned long max); 813 814 /** 815 * clk_set_min_rate - set a minimum clock rate for a clock source 816 * @clk: clock source 817 * @rate: desired minimum clock rate in Hz, inclusive 818 * 819 * Returns success (0) or negative errno. 820 */ 821 int clk_set_min_rate(struct clk *clk, unsigned long rate); 822 823 /** 824 * clk_set_max_rate - set a maximum clock rate for a clock source 825 * @clk: clock source 826 * @rate: desired maximum clock rate in Hz, inclusive 827 * 828 * Returns success (0) or negative errno. 829 */ 830 int clk_set_max_rate(struct clk *clk, unsigned long rate); 831 832 /** 833 * clk_set_parent - set the parent clock source for this clock 834 * @clk: clock source 835 * @parent: parent clock source 836 * 837 * Returns success (0) or negative errno. 838 */ 839 int clk_set_parent(struct clk *clk, struct clk *parent); 840 841 /** 842 * clk_get_parent - get the parent clock source for this clock 843 * @clk: clock source 844 * 845 * Returns struct clk corresponding to parent clock source, or 846 * valid IS_ERR() condition containing errno. 847 */ 848 struct clk *clk_get_parent(struct clk *clk); 849 850 /** 851 * clk_get_sys - get a clock based upon the device name 852 * @dev_id: device name 853 * @con_id: connection ID 854 * 855 * Returns a struct clk corresponding to the clock producer, or 856 * valid IS_ERR() condition containing errno. The implementation 857 * uses @dev_id and @con_id to determine the clock consumer, and 858 * thereby the clock producer. In contrast to clk_get() this function 859 * takes the device name instead of the device itself for identification. 860 * 861 * Drivers must assume that the clock source is not enabled. 862 * 863 * clk_get_sys should not be called from within interrupt context. 864 */ 865 struct clk *clk_get_sys(const char *dev_id, const char *con_id); 866 867 /** 868 * clk_save_context - save clock context for poweroff 869 * 870 * Saves the context of the clock register for powerstates in which the 871 * contents of the registers will be lost. Occurs deep within the suspend 872 * code so locking is not necessary. 873 */ 874 int clk_save_context(void); 875 876 /** 877 * clk_restore_context - restore clock context after poweroff 878 * 879 * This occurs with all clocks enabled. Occurs deep within the resume code 880 * so locking is not necessary. 881 */ 882 void clk_restore_context(void); 883 884 #else /* !CONFIG_HAVE_CLK */ 885 886 static inline struct clk *clk_get(struct device *dev, const char *id) 887 { 888 return NULL; 889 } 890 891 static inline int __must_check clk_bulk_get(struct device *dev, int num_clks, 892 struct clk_bulk_data *clks) 893 { 894 return 0; 895 } 896 897 static inline int __must_check clk_bulk_get_optional(struct device *dev, 898 int num_clks, struct clk_bulk_data *clks) 899 { 900 return 0; 901 } 902 903 static inline int __must_check clk_bulk_get_all(struct device *dev, 904 struct clk_bulk_data **clks) 905 { 906 return 0; 907 } 908 909 static inline struct clk *devm_clk_get(struct device *dev, const char *id) 910 { 911 return NULL; 912 } 913 914 static inline struct clk *devm_clk_get_prepared(struct device *dev, 915 const char *id) 916 { 917 return NULL; 918 } 919 920 static inline struct clk *devm_clk_get_enabled(struct device *dev, 921 const char *id) 922 { 923 return NULL; 924 } 925 926 static inline struct clk *devm_clk_get_optional(struct device *dev, 927 const char *id) 928 { 929 return NULL; 930 } 931 932 static inline struct clk *devm_clk_get_optional_prepared(struct device *dev, 933 const char *id) 934 { 935 return NULL; 936 } 937 938 static inline struct clk *devm_clk_get_optional_enabled(struct device *dev, 939 const char *id) 940 { 941 return NULL; 942 } 943 944 static inline int __must_check devm_clk_bulk_get(struct device *dev, int num_clks, 945 struct clk_bulk_data *clks) 946 { 947 return 0; 948 } 949 950 static inline int __must_check devm_clk_bulk_get_optional(struct device *dev, 951 int num_clks, struct clk_bulk_data *clks) 952 { 953 return 0; 954 } 955 956 static inline int __must_check devm_clk_bulk_get_all(struct device *dev, 957 struct clk_bulk_data **clks) 958 { 959 960 return 0; 961 } 962 963 static inline struct clk *devm_get_clk_from_child(struct device *dev, 964 struct device_node *np, const char *con_id) 965 { 966 return NULL; 967 } 968 969 static inline void clk_put(struct clk *clk) {} 970 971 static inline void clk_bulk_put(int num_clks, struct clk_bulk_data *clks) {} 972 973 static inline void clk_bulk_put_all(int num_clks, struct clk_bulk_data *clks) {} 974 975 static inline void devm_clk_put(struct device *dev, struct clk *clk) {} 976 977 978 static inline int clk_rate_exclusive_get(struct clk *clk) 979 { 980 return 0; 981 } 982 983 static inline void clk_rate_exclusive_put(struct clk *clk) {} 984 985 static inline int clk_enable(struct clk *clk) 986 { 987 return 0; 988 } 989 990 static inline int __must_check clk_bulk_enable(int num_clks, 991 const struct clk_bulk_data *clks) 992 { 993 return 0; 994 } 995 996 static inline void clk_disable(struct clk *clk) {} 997 998 999 static inline void clk_bulk_disable(int num_clks, 1000 const struct clk_bulk_data *clks) {} 1001 1002 static inline unsigned long clk_get_rate(struct clk *clk) 1003 { 1004 return 0; 1005 } 1006 1007 static inline int clk_set_rate(struct clk *clk, unsigned long rate) 1008 { 1009 return 0; 1010 } 1011 1012 static inline int clk_set_rate_exclusive(struct clk *clk, unsigned long rate) 1013 { 1014 return 0; 1015 } 1016 1017 static inline long clk_round_rate(struct clk *clk, unsigned long rate) 1018 { 1019 return 0; 1020 } 1021 1022 static inline bool clk_has_parent(struct clk *clk, struct clk *parent) 1023 { 1024 return true; 1025 } 1026 1027 static inline int clk_set_rate_range(struct clk *clk, unsigned long min, 1028 unsigned long max) 1029 { 1030 return 0; 1031 } 1032 1033 static inline int clk_set_min_rate(struct clk *clk, unsigned long rate) 1034 { 1035 return 0; 1036 } 1037 1038 static inline int clk_set_max_rate(struct clk *clk, unsigned long rate) 1039 { 1040 return 0; 1041 } 1042 1043 static inline int clk_set_parent(struct clk *clk, struct clk *parent) 1044 { 1045 return 0; 1046 } 1047 1048 static inline struct clk *clk_get_parent(struct clk *clk) 1049 { 1050 return NULL; 1051 } 1052 1053 static inline struct clk *clk_get_sys(const char *dev_id, const char *con_id) 1054 { 1055 return NULL; 1056 } 1057 1058 static inline int clk_save_context(void) 1059 { 1060 return 0; 1061 } 1062 1063 static inline void clk_restore_context(void) {} 1064 1065 #endif 1066 1067 /* clk_prepare_enable helps cases using clk_enable in non-atomic context. */ 1068 static inline int clk_prepare_enable(struct clk *clk) 1069 { 1070 int ret; 1071 1072 ret = clk_prepare(clk); 1073 if (ret) 1074 return ret; 1075 ret = clk_enable(clk); 1076 if (ret) 1077 clk_unprepare(clk); 1078 1079 return ret; 1080 } 1081 1082 /* clk_disable_unprepare helps cases using clk_disable in non-atomic context. */ 1083 static inline void clk_disable_unprepare(struct clk *clk) 1084 { 1085 clk_disable(clk); 1086 clk_unprepare(clk); 1087 } 1088 1089 static inline int __must_check 1090 clk_bulk_prepare_enable(int num_clks, const struct clk_bulk_data *clks) 1091 { 1092 int ret; 1093 1094 ret = clk_bulk_prepare(num_clks, clks); 1095 if (ret) 1096 return ret; 1097 ret = clk_bulk_enable(num_clks, clks); 1098 if (ret) 1099 clk_bulk_unprepare(num_clks, clks); 1100 1101 return ret; 1102 } 1103 1104 static inline void clk_bulk_disable_unprepare(int num_clks, 1105 const struct clk_bulk_data *clks) 1106 { 1107 clk_bulk_disable(num_clks, clks); 1108 clk_bulk_unprepare(num_clks, clks); 1109 } 1110 1111 /** 1112 * clk_drop_range - Reset any range set on that clock 1113 * @clk: clock source 1114 * 1115 * Returns success (0) or negative errno. 1116 */ 1117 static inline int clk_drop_range(struct clk *clk) 1118 { 1119 return clk_set_rate_range(clk, 0, ULONG_MAX); 1120 } 1121 1122 /** 1123 * clk_get_optional - lookup and obtain a reference to an optional clock 1124 * producer. 1125 * @dev: device for clock "consumer" 1126 * @id: clock consumer ID 1127 * 1128 * Behaves the same as clk_get() except where there is no clock producer. In 1129 * this case, instead of returning -ENOENT, the function returns NULL. 1130 */ 1131 static inline struct clk *clk_get_optional(struct device *dev, const char *id) 1132 { 1133 struct clk *clk = clk_get(dev, id); 1134 1135 if (clk == ERR_PTR(-ENOENT)) 1136 return NULL; 1137 1138 return clk; 1139 } 1140 1141 #if defined(CONFIG_OF) && defined(CONFIG_COMMON_CLK) 1142 struct clk *of_clk_get(struct device_node *np, int index); 1143 struct clk *of_clk_get_by_name(struct device_node *np, const char *name); 1144 struct clk *of_clk_get_from_provider(struct of_phandle_args *clkspec); 1145 #else 1146 static inline struct clk *of_clk_get(struct device_node *np, int index) 1147 { 1148 return ERR_PTR(-ENOENT); 1149 } 1150 static inline struct clk *of_clk_get_by_name(struct device_node *np, 1151 const char *name) 1152 { 1153 return ERR_PTR(-ENOENT); 1154 } 1155 static inline struct clk *of_clk_get_from_provider(struct of_phandle_args *clkspec) 1156 { 1157 return ERR_PTR(-ENOENT); 1158 } 1159 #endif 1160 1161 #endif 1162