1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Copyright (c) 2013 NVIDIA CORPORATION. All rights reserved. 4 */ 5 6 #include <linux/clk-provider.h> 7 #include <linux/device.h> 8 #include <linux/err.h> 9 #include <linux/slab.h> 10 11 static u8 clk_composite_get_parent(struct clk_hw *hw) 12 { 13 struct clk_composite *composite = to_clk_composite(hw); 14 const struct clk_ops *mux_ops = composite->mux_ops; 15 struct clk_hw *mux_hw = composite->mux_hw; 16 17 __clk_hw_set_clk(mux_hw, hw); 18 19 return mux_ops->get_parent(mux_hw); 20 } 21 22 static int clk_composite_set_parent(struct clk_hw *hw, u8 index) 23 { 24 struct clk_composite *composite = to_clk_composite(hw); 25 const struct clk_ops *mux_ops = composite->mux_ops; 26 struct clk_hw *mux_hw = composite->mux_hw; 27 28 __clk_hw_set_clk(mux_hw, hw); 29 30 return mux_ops->set_parent(mux_hw, index); 31 } 32 33 static unsigned long clk_composite_recalc_rate(struct clk_hw *hw, 34 unsigned long parent_rate) 35 { 36 struct clk_composite *composite = to_clk_composite(hw); 37 const struct clk_ops *rate_ops = composite->rate_ops; 38 struct clk_hw *rate_hw = composite->rate_hw; 39 40 __clk_hw_set_clk(rate_hw, hw); 41 42 return rate_ops->recalc_rate(rate_hw, parent_rate); 43 } 44 45 static int clk_composite_determine_rate_for_parent(struct clk_hw *rate_hw, 46 struct clk_rate_request *req, 47 struct clk_hw *parent_hw, 48 const struct clk_ops *rate_ops) 49 { 50 long rate; 51 52 req->best_parent_hw = parent_hw; 53 req->best_parent_rate = clk_hw_get_rate(parent_hw); 54 55 if (rate_ops->determine_rate) 56 return rate_ops->determine_rate(rate_hw, req); 57 58 rate = rate_ops->round_rate(rate_hw, req->rate, 59 &req->best_parent_rate); 60 if (rate < 0) 61 return rate; 62 63 req->rate = rate; 64 65 return 0; 66 } 67 68 static int clk_composite_determine_rate(struct clk_hw *hw, 69 struct clk_rate_request *req) 70 { 71 struct clk_composite *composite = to_clk_composite(hw); 72 const struct clk_ops *rate_ops = composite->rate_ops; 73 const struct clk_ops *mux_ops = composite->mux_ops; 74 struct clk_hw *rate_hw = composite->rate_hw; 75 struct clk_hw *mux_hw = composite->mux_hw; 76 struct clk_hw *parent; 77 unsigned long rate_diff; 78 unsigned long best_rate_diff = ULONG_MAX; 79 unsigned long best_rate = 0; 80 int i, ret; 81 82 if (rate_hw && rate_ops && 83 (rate_ops->determine_rate || rate_ops->round_rate) && 84 mux_hw && mux_ops && mux_ops->set_parent) { 85 req->best_parent_hw = NULL; 86 87 if (clk_hw_get_flags(hw) & CLK_SET_RATE_NO_REPARENT) { 88 struct clk_rate_request tmp_req = *req; 89 90 parent = clk_hw_get_parent(mux_hw); 91 92 ret = clk_composite_determine_rate_for_parent(rate_hw, 93 &tmp_req, 94 parent, 95 rate_ops); 96 if (ret) 97 return ret; 98 99 req->rate = tmp_req.rate; 100 req->best_parent_hw = tmp_req.best_parent_hw; 101 req->best_parent_rate = tmp_req.best_parent_rate; 102 103 return 0; 104 } 105 106 for (i = 0; i < clk_hw_get_num_parents(mux_hw); i++) { 107 struct clk_rate_request tmp_req = *req; 108 109 parent = clk_hw_get_parent_by_index(mux_hw, i); 110 if (!parent) 111 continue; 112 113 ret = clk_composite_determine_rate_for_parent(rate_hw, 114 &tmp_req, 115 parent, 116 rate_ops); 117 if (ret) 118 continue; 119 120 rate_diff = abs(req->rate - tmp_req.rate); 121 122 if (!rate_diff || !req->best_parent_hw 123 || best_rate_diff > rate_diff) { 124 req->best_parent_hw = parent; 125 req->best_parent_rate = tmp_req.best_parent_rate; 126 best_rate_diff = rate_diff; 127 best_rate = tmp_req.rate; 128 } 129 130 if (!rate_diff) 131 return 0; 132 } 133 134 req->rate = best_rate; 135 return 0; 136 } else if (rate_hw && rate_ops && rate_ops->determine_rate) { 137 __clk_hw_set_clk(rate_hw, hw); 138 return rate_ops->determine_rate(rate_hw, req); 139 } else if (mux_hw && mux_ops && mux_ops->determine_rate) { 140 __clk_hw_set_clk(mux_hw, hw); 141 return mux_ops->determine_rate(mux_hw, req); 142 } else { 143 pr_err("clk: clk_composite_determine_rate function called, but no mux or rate callback set!\n"); 144 return -EINVAL; 145 } 146 } 147 148 static long clk_composite_round_rate(struct clk_hw *hw, unsigned long rate, 149 unsigned long *prate) 150 { 151 struct clk_composite *composite = to_clk_composite(hw); 152 const struct clk_ops *rate_ops = composite->rate_ops; 153 struct clk_hw *rate_hw = composite->rate_hw; 154 155 __clk_hw_set_clk(rate_hw, hw); 156 157 return rate_ops->round_rate(rate_hw, rate, prate); 158 } 159 160 static int clk_composite_set_rate(struct clk_hw *hw, unsigned long rate, 161 unsigned long parent_rate) 162 { 163 struct clk_composite *composite = to_clk_composite(hw); 164 const struct clk_ops *rate_ops = composite->rate_ops; 165 struct clk_hw *rate_hw = composite->rate_hw; 166 167 __clk_hw_set_clk(rate_hw, hw); 168 169 return rate_ops->set_rate(rate_hw, rate, parent_rate); 170 } 171 172 static int clk_composite_set_rate_and_parent(struct clk_hw *hw, 173 unsigned long rate, 174 unsigned long parent_rate, 175 u8 index) 176 { 177 struct clk_composite *composite = to_clk_composite(hw); 178 const struct clk_ops *rate_ops = composite->rate_ops; 179 const struct clk_ops *mux_ops = composite->mux_ops; 180 struct clk_hw *rate_hw = composite->rate_hw; 181 struct clk_hw *mux_hw = composite->mux_hw; 182 unsigned long temp_rate; 183 184 __clk_hw_set_clk(rate_hw, hw); 185 __clk_hw_set_clk(mux_hw, hw); 186 187 temp_rate = rate_ops->recalc_rate(rate_hw, parent_rate); 188 if (temp_rate > rate) { 189 rate_ops->set_rate(rate_hw, rate, parent_rate); 190 mux_ops->set_parent(mux_hw, index); 191 } else { 192 mux_ops->set_parent(mux_hw, index); 193 rate_ops->set_rate(rate_hw, rate, parent_rate); 194 } 195 196 return 0; 197 } 198 199 static int clk_composite_is_enabled(struct clk_hw *hw) 200 { 201 struct clk_composite *composite = to_clk_composite(hw); 202 const struct clk_ops *gate_ops = composite->gate_ops; 203 struct clk_hw *gate_hw = composite->gate_hw; 204 205 __clk_hw_set_clk(gate_hw, hw); 206 207 return gate_ops->is_enabled(gate_hw); 208 } 209 210 static int clk_composite_enable(struct clk_hw *hw) 211 { 212 struct clk_composite *composite = to_clk_composite(hw); 213 const struct clk_ops *gate_ops = composite->gate_ops; 214 struct clk_hw *gate_hw = composite->gate_hw; 215 216 __clk_hw_set_clk(gate_hw, hw); 217 218 return gate_ops->enable(gate_hw); 219 } 220 221 static void clk_composite_disable(struct clk_hw *hw) 222 { 223 struct clk_composite *composite = to_clk_composite(hw); 224 const struct clk_ops *gate_ops = composite->gate_ops; 225 struct clk_hw *gate_hw = composite->gate_hw; 226 227 __clk_hw_set_clk(gate_hw, hw); 228 229 gate_ops->disable(gate_hw); 230 } 231 232 static struct clk_hw *__clk_hw_register_composite(struct device *dev, 233 const char *name, const char * const *parent_names, 234 const struct clk_parent_data *pdata, int num_parents, 235 struct clk_hw *mux_hw, const struct clk_ops *mux_ops, 236 struct clk_hw *rate_hw, const struct clk_ops *rate_ops, 237 struct clk_hw *gate_hw, const struct clk_ops *gate_ops, 238 unsigned long flags) 239 { 240 struct clk_hw *hw; 241 struct clk_init_data init = {}; 242 struct clk_composite *composite; 243 struct clk_ops *clk_composite_ops; 244 int ret; 245 246 composite = kzalloc(sizeof(*composite), GFP_KERNEL); 247 if (!composite) 248 return ERR_PTR(-ENOMEM); 249 250 init.name = name; 251 init.flags = flags; 252 if (parent_names) 253 init.parent_names = parent_names; 254 else 255 init.parent_data = pdata; 256 init.num_parents = num_parents; 257 hw = &composite->hw; 258 259 clk_composite_ops = &composite->ops; 260 261 if (mux_hw && mux_ops) { 262 if (!mux_ops->get_parent) { 263 hw = ERR_PTR(-EINVAL); 264 goto err; 265 } 266 267 composite->mux_hw = mux_hw; 268 composite->mux_ops = mux_ops; 269 clk_composite_ops->get_parent = clk_composite_get_parent; 270 if (mux_ops->set_parent) 271 clk_composite_ops->set_parent = clk_composite_set_parent; 272 if (mux_ops->determine_rate) 273 clk_composite_ops->determine_rate = clk_composite_determine_rate; 274 } 275 276 if (rate_hw && rate_ops) { 277 if (!rate_ops->recalc_rate) { 278 hw = ERR_PTR(-EINVAL); 279 goto err; 280 } 281 clk_composite_ops->recalc_rate = clk_composite_recalc_rate; 282 283 if (rate_ops->determine_rate) 284 clk_composite_ops->determine_rate = 285 clk_composite_determine_rate; 286 else if (rate_ops->round_rate) 287 clk_composite_ops->round_rate = 288 clk_composite_round_rate; 289 290 /* .set_rate requires either .round_rate or .determine_rate */ 291 if (rate_ops->set_rate) { 292 if (rate_ops->determine_rate || rate_ops->round_rate) 293 clk_composite_ops->set_rate = 294 clk_composite_set_rate; 295 else 296 WARN(1, "%s: missing round_rate op is required\n", 297 __func__); 298 } 299 300 composite->rate_hw = rate_hw; 301 composite->rate_ops = rate_ops; 302 } 303 304 if (mux_hw && mux_ops && rate_hw && rate_ops) { 305 if (mux_ops->set_parent && rate_ops->set_rate) 306 clk_composite_ops->set_rate_and_parent = 307 clk_composite_set_rate_and_parent; 308 } 309 310 if (gate_hw && gate_ops) { 311 if (!gate_ops->is_enabled || !gate_ops->enable || 312 !gate_ops->disable) { 313 hw = ERR_PTR(-EINVAL); 314 goto err; 315 } 316 317 composite->gate_hw = gate_hw; 318 composite->gate_ops = gate_ops; 319 clk_composite_ops->is_enabled = clk_composite_is_enabled; 320 clk_composite_ops->enable = clk_composite_enable; 321 clk_composite_ops->disable = clk_composite_disable; 322 } 323 324 init.ops = clk_composite_ops; 325 composite->hw.init = &init; 326 327 ret = clk_hw_register(dev, hw); 328 if (ret) { 329 hw = ERR_PTR(ret); 330 goto err; 331 } 332 333 if (composite->mux_hw) 334 composite->mux_hw->clk = hw->clk; 335 336 if (composite->rate_hw) 337 composite->rate_hw->clk = hw->clk; 338 339 if (composite->gate_hw) 340 composite->gate_hw->clk = hw->clk; 341 342 return hw; 343 344 err: 345 kfree(composite); 346 return hw; 347 } 348 349 struct clk_hw *clk_hw_register_composite(struct device *dev, const char *name, 350 const char * const *parent_names, int num_parents, 351 struct clk_hw *mux_hw, const struct clk_ops *mux_ops, 352 struct clk_hw *rate_hw, const struct clk_ops *rate_ops, 353 struct clk_hw *gate_hw, const struct clk_ops *gate_ops, 354 unsigned long flags) 355 { 356 return __clk_hw_register_composite(dev, name, parent_names, NULL, 357 num_parents, mux_hw, mux_ops, 358 rate_hw, rate_ops, gate_hw, 359 gate_ops, flags); 360 } 361 EXPORT_SYMBOL_GPL(clk_hw_register_composite); 362 363 struct clk_hw *clk_hw_register_composite_pdata(struct device *dev, 364 const char *name, 365 const struct clk_parent_data *parent_data, 366 int num_parents, 367 struct clk_hw *mux_hw, const struct clk_ops *mux_ops, 368 struct clk_hw *rate_hw, const struct clk_ops *rate_ops, 369 struct clk_hw *gate_hw, const struct clk_ops *gate_ops, 370 unsigned long flags) 371 { 372 return __clk_hw_register_composite(dev, name, NULL, parent_data, 373 num_parents, mux_hw, mux_ops, 374 rate_hw, rate_ops, gate_hw, 375 gate_ops, flags); 376 } 377 378 struct clk *clk_register_composite(struct device *dev, const char *name, 379 const char * const *parent_names, int num_parents, 380 struct clk_hw *mux_hw, const struct clk_ops *mux_ops, 381 struct clk_hw *rate_hw, const struct clk_ops *rate_ops, 382 struct clk_hw *gate_hw, const struct clk_ops *gate_ops, 383 unsigned long flags) 384 { 385 struct clk_hw *hw; 386 387 hw = clk_hw_register_composite(dev, name, parent_names, num_parents, 388 mux_hw, mux_ops, rate_hw, rate_ops, gate_hw, gate_ops, 389 flags); 390 if (IS_ERR(hw)) 391 return ERR_CAST(hw); 392 return hw->clk; 393 } 394 EXPORT_SYMBOL_GPL(clk_register_composite); 395 396 struct clk *clk_register_composite_pdata(struct device *dev, const char *name, 397 const struct clk_parent_data *parent_data, 398 int num_parents, 399 struct clk_hw *mux_hw, const struct clk_ops *mux_ops, 400 struct clk_hw *rate_hw, const struct clk_ops *rate_ops, 401 struct clk_hw *gate_hw, const struct clk_ops *gate_ops, 402 unsigned long flags) 403 { 404 struct clk_hw *hw; 405 406 hw = clk_hw_register_composite_pdata(dev, name, parent_data, 407 num_parents, mux_hw, mux_ops, rate_hw, rate_ops, 408 gate_hw, gate_ops, flags); 409 if (IS_ERR(hw)) 410 return ERR_CAST(hw); 411 return hw->clk; 412 } 413 414 void clk_unregister_composite(struct clk *clk) 415 { 416 struct clk_composite *composite; 417 struct clk_hw *hw; 418 419 hw = __clk_get_hw(clk); 420 if (!hw) 421 return; 422 423 composite = to_clk_composite(hw); 424 425 clk_unregister(clk); 426 kfree(composite); 427 } 428 429 void clk_hw_unregister_composite(struct clk_hw *hw) 430 { 431 struct clk_composite *composite; 432 433 composite = to_clk_composite(hw); 434 435 clk_hw_unregister(hw); 436 kfree(composite); 437 } 438 EXPORT_SYMBOL_GPL(clk_hw_unregister_composite); 439 440 static void devm_clk_hw_release_composite(struct device *dev, void *res) 441 { 442 clk_hw_unregister_composite(*(struct clk_hw **)res); 443 } 444 445 static struct clk_hw *__devm_clk_hw_register_composite(struct device *dev, 446 const char *name, const char * const *parent_names, 447 const struct clk_parent_data *pdata, int num_parents, 448 struct clk_hw *mux_hw, const struct clk_ops *mux_ops, 449 struct clk_hw *rate_hw, const struct clk_ops *rate_ops, 450 struct clk_hw *gate_hw, const struct clk_ops *gate_ops, 451 unsigned long flags) 452 { 453 struct clk_hw **ptr, *hw; 454 455 ptr = devres_alloc(devm_clk_hw_release_composite, sizeof(*ptr), 456 GFP_KERNEL); 457 if (!ptr) 458 return ERR_PTR(-ENOMEM); 459 460 hw = __clk_hw_register_composite(dev, name, parent_names, pdata, 461 num_parents, mux_hw, mux_ops, rate_hw, 462 rate_ops, gate_hw, gate_ops, flags); 463 464 if (!IS_ERR(hw)) { 465 *ptr = hw; 466 devres_add(dev, ptr); 467 } else { 468 devres_free(ptr); 469 } 470 471 return hw; 472 } 473 474 struct clk_hw *devm_clk_hw_register_composite_pdata(struct device *dev, 475 const char *name, 476 const struct clk_parent_data *parent_data, 477 int num_parents, 478 struct clk_hw *mux_hw, const struct clk_ops *mux_ops, 479 struct clk_hw *rate_hw, const struct clk_ops *rate_ops, 480 struct clk_hw *gate_hw, const struct clk_ops *gate_ops, 481 unsigned long flags) 482 { 483 return __devm_clk_hw_register_composite(dev, name, NULL, parent_data, 484 num_parents, mux_hw, mux_ops, 485 rate_hw, rate_ops, gate_hw, 486 gate_ops, flags); 487 } 488