1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * System Control and Management Interface (SCMI) Clock Protocol 4 * 5 * Copyright (C) 2018 ARM Ltd. 6 */ 7 8 #include "common.h" 9 10 enum scmi_clock_protocol_cmd { 11 CLOCK_ATTRIBUTES = 0x3, 12 CLOCK_DESCRIBE_RATES = 0x4, 13 CLOCK_RATE_SET = 0x5, 14 CLOCK_RATE_GET = 0x6, 15 CLOCK_CONFIG_SET = 0x7, 16 }; 17 18 struct scmi_msg_resp_clock_protocol_attributes { 19 __le16 num_clocks; 20 u8 max_async_req; 21 u8 reserved; 22 }; 23 24 struct scmi_msg_resp_clock_attributes { 25 __le32 attributes; 26 #define CLOCK_ENABLE BIT(0) 27 u8 name[SCMI_MAX_STR_SIZE]; 28 }; 29 30 struct scmi_clock_set_config { 31 __le32 id; 32 __le32 attributes; 33 }; 34 35 struct scmi_msg_clock_describe_rates { 36 __le32 id; 37 __le32 rate_index; 38 }; 39 40 struct scmi_msg_resp_clock_describe_rates { 41 __le32 num_rates_flags; 42 #define NUM_RETURNED(x) ((x) & 0xfff) 43 #define RATE_DISCRETE(x) !((x) & BIT(12)) 44 #define NUM_REMAINING(x) ((x) >> 16) 45 struct { 46 __le32 value_low; 47 __le32 value_high; 48 } rate[0]; 49 #define RATE_TO_U64(X) \ 50 ({ \ 51 typeof(X) x = (X); \ 52 le32_to_cpu((x).value_low) | (u64)le32_to_cpu((x).value_high) << 32; \ 53 }) 54 }; 55 56 struct scmi_clock_set_rate { 57 __le32 flags; 58 #define CLOCK_SET_ASYNC BIT(0) 59 #define CLOCK_SET_DELAYED BIT(1) 60 #define CLOCK_SET_ROUND_UP BIT(2) 61 #define CLOCK_SET_ROUND_AUTO BIT(3) 62 __le32 id; 63 __le32 value_low; 64 __le32 value_high; 65 }; 66 67 struct clock_info { 68 int num_clocks; 69 int max_async_req; 70 struct scmi_clock_info *clk; 71 }; 72 73 static int scmi_clock_protocol_attributes_get(const struct scmi_handle *handle, 74 struct clock_info *ci) 75 { 76 int ret; 77 struct scmi_xfer *t; 78 struct scmi_msg_resp_clock_protocol_attributes *attr; 79 80 ret = scmi_xfer_get_init(handle, PROTOCOL_ATTRIBUTES, 81 SCMI_PROTOCOL_CLOCK, 0, sizeof(*attr), &t); 82 if (ret) 83 return ret; 84 85 attr = t->rx.buf; 86 87 ret = scmi_do_xfer(handle, t); 88 if (!ret) { 89 ci->num_clocks = le16_to_cpu(attr->num_clocks); 90 ci->max_async_req = attr->max_async_req; 91 } 92 93 scmi_xfer_put(handle, t); 94 return ret; 95 } 96 97 static int scmi_clock_attributes_get(const struct scmi_handle *handle, 98 u32 clk_id, struct scmi_clock_info *clk) 99 { 100 int ret; 101 struct scmi_xfer *t; 102 struct scmi_msg_resp_clock_attributes *attr; 103 104 ret = scmi_xfer_get_init(handle, CLOCK_ATTRIBUTES, SCMI_PROTOCOL_CLOCK, 105 sizeof(clk_id), sizeof(*attr), &t); 106 if (ret) 107 return ret; 108 109 *(__le32 *)t->tx.buf = cpu_to_le32(clk_id); 110 attr = t->rx.buf; 111 112 ret = scmi_do_xfer(handle, t); 113 if (!ret) 114 strlcpy(clk->name, attr->name, SCMI_MAX_STR_SIZE); 115 else 116 clk->name[0] = '\0'; 117 118 scmi_xfer_put(handle, t); 119 return ret; 120 } 121 122 static int 123 scmi_clock_describe_rates_get(const struct scmi_handle *handle, u32 clk_id, 124 struct scmi_clock_info *clk) 125 { 126 u64 *rate; 127 int ret, cnt; 128 bool rate_discrete = false; 129 u32 tot_rate_cnt = 0, rates_flag; 130 u16 num_returned, num_remaining; 131 struct scmi_xfer *t; 132 struct scmi_msg_clock_describe_rates *clk_desc; 133 struct scmi_msg_resp_clock_describe_rates *rlist; 134 135 ret = scmi_xfer_get_init(handle, CLOCK_DESCRIBE_RATES, 136 SCMI_PROTOCOL_CLOCK, sizeof(*clk_desc), 0, &t); 137 if (ret) 138 return ret; 139 140 clk_desc = t->tx.buf; 141 rlist = t->rx.buf; 142 143 do { 144 clk_desc->id = cpu_to_le32(clk_id); 145 /* Set the number of rates to be skipped/already read */ 146 clk_desc->rate_index = cpu_to_le32(tot_rate_cnt); 147 148 ret = scmi_do_xfer(handle, t); 149 if (ret) 150 goto err; 151 152 rates_flag = le32_to_cpu(rlist->num_rates_flags); 153 num_remaining = NUM_REMAINING(rates_flag); 154 rate_discrete = RATE_DISCRETE(rates_flag); 155 num_returned = NUM_RETURNED(rates_flag); 156 157 if (tot_rate_cnt + num_returned > SCMI_MAX_NUM_RATES) { 158 dev_err(handle->dev, "No. of rates > MAX_NUM_RATES"); 159 break; 160 } 161 162 if (!rate_discrete) { 163 clk->range.min_rate = RATE_TO_U64(rlist->rate[0]); 164 clk->range.max_rate = RATE_TO_U64(rlist->rate[1]); 165 clk->range.step_size = RATE_TO_U64(rlist->rate[2]); 166 dev_dbg(handle->dev, "Min %llu Max %llu Step %llu Hz\n", 167 clk->range.min_rate, clk->range.max_rate, 168 clk->range.step_size); 169 break; 170 } 171 172 rate = &clk->list.rates[tot_rate_cnt]; 173 for (cnt = 0; cnt < num_returned; cnt++, rate++) { 174 *rate = RATE_TO_U64(rlist->rate[cnt]); 175 dev_dbg(handle->dev, "Rate %llu Hz\n", *rate); 176 } 177 178 tot_rate_cnt += num_returned; 179 /* 180 * check for both returned and remaining to avoid infinite 181 * loop due to buggy firmware 182 */ 183 } while (num_returned && num_remaining); 184 185 if (rate_discrete) 186 clk->list.num_rates = tot_rate_cnt; 187 188 err: 189 scmi_xfer_put(handle, t); 190 return ret; 191 } 192 193 static int 194 scmi_clock_rate_get(const struct scmi_handle *handle, u32 clk_id, u64 *value) 195 { 196 int ret; 197 struct scmi_xfer *t; 198 199 ret = scmi_xfer_get_init(handle, CLOCK_RATE_GET, SCMI_PROTOCOL_CLOCK, 200 sizeof(__le32), sizeof(u64), &t); 201 if (ret) 202 return ret; 203 204 *(__le32 *)t->tx.buf = cpu_to_le32(clk_id); 205 206 ret = scmi_do_xfer(handle, t); 207 if (!ret) { 208 __le32 *pval = t->rx.buf; 209 210 *value = le32_to_cpu(*pval); 211 *value |= (u64)le32_to_cpu(*(pval + 1)) << 32; 212 } 213 214 scmi_xfer_put(handle, t); 215 return ret; 216 } 217 218 static int scmi_clock_rate_set(const struct scmi_handle *handle, u32 clk_id, 219 u32 config, u64 rate) 220 { 221 int ret; 222 struct scmi_xfer *t; 223 struct scmi_clock_set_rate *cfg; 224 225 ret = scmi_xfer_get_init(handle, CLOCK_RATE_SET, SCMI_PROTOCOL_CLOCK, 226 sizeof(*cfg), 0, &t); 227 if (ret) 228 return ret; 229 230 cfg = t->tx.buf; 231 cfg->flags = cpu_to_le32(config); 232 cfg->id = cpu_to_le32(clk_id); 233 cfg->value_low = cpu_to_le32(rate & 0xffffffff); 234 cfg->value_high = cpu_to_le32(rate >> 32); 235 236 ret = scmi_do_xfer(handle, t); 237 238 scmi_xfer_put(handle, t); 239 return ret; 240 } 241 242 static int 243 scmi_clock_config_set(const struct scmi_handle *handle, u32 clk_id, u32 config) 244 { 245 int ret; 246 struct scmi_xfer *t; 247 struct scmi_clock_set_config *cfg; 248 249 ret = scmi_xfer_get_init(handle, CLOCK_CONFIG_SET, SCMI_PROTOCOL_CLOCK, 250 sizeof(*cfg), 0, &t); 251 if (ret) 252 return ret; 253 254 cfg = t->tx.buf; 255 cfg->id = cpu_to_le32(clk_id); 256 cfg->attributes = cpu_to_le32(config); 257 258 ret = scmi_do_xfer(handle, t); 259 260 scmi_xfer_put(handle, t); 261 return ret; 262 } 263 264 static int scmi_clock_enable(const struct scmi_handle *handle, u32 clk_id) 265 { 266 return scmi_clock_config_set(handle, clk_id, CLOCK_ENABLE); 267 } 268 269 static int scmi_clock_disable(const struct scmi_handle *handle, u32 clk_id) 270 { 271 return scmi_clock_config_set(handle, clk_id, 0); 272 } 273 274 static int scmi_clock_count_get(const struct scmi_handle *handle) 275 { 276 struct clock_info *ci = handle->clk_priv; 277 278 return ci->num_clocks; 279 } 280 281 static const struct scmi_clock_info * 282 scmi_clock_info_get(const struct scmi_handle *handle, u32 clk_id) 283 { 284 struct clock_info *ci = handle->clk_priv; 285 struct scmi_clock_info *clk = ci->clk + clk_id; 286 287 if (!clk->name[0]) 288 return NULL; 289 290 return clk; 291 } 292 293 static struct scmi_clk_ops clk_ops = { 294 .count_get = scmi_clock_count_get, 295 .info_get = scmi_clock_info_get, 296 .rate_get = scmi_clock_rate_get, 297 .rate_set = scmi_clock_rate_set, 298 .enable = scmi_clock_enable, 299 .disable = scmi_clock_disable, 300 }; 301 302 static int scmi_clock_protocol_init(struct scmi_handle *handle) 303 { 304 u32 version; 305 int clkid, ret; 306 struct clock_info *cinfo; 307 308 scmi_version_get(handle, SCMI_PROTOCOL_CLOCK, &version); 309 310 dev_dbg(handle->dev, "Clock Version %d.%d\n", 311 PROTOCOL_REV_MAJOR(version), PROTOCOL_REV_MINOR(version)); 312 313 cinfo = devm_kzalloc(handle->dev, sizeof(*cinfo), GFP_KERNEL); 314 if (!cinfo) 315 return -ENOMEM; 316 317 scmi_clock_protocol_attributes_get(handle, cinfo); 318 319 cinfo->clk = devm_kcalloc(handle->dev, cinfo->num_clocks, 320 sizeof(*cinfo->clk), GFP_KERNEL); 321 if (!cinfo->clk) 322 return -ENOMEM; 323 324 for (clkid = 0; clkid < cinfo->num_clocks; clkid++) { 325 struct scmi_clock_info *clk = cinfo->clk + clkid; 326 327 ret = scmi_clock_attributes_get(handle, clkid, clk); 328 if (!ret) 329 scmi_clock_describe_rates_get(handle, clkid, clk); 330 } 331 332 handle->clk_ops = &clk_ops; 333 handle->clk_priv = cinfo; 334 335 return 0; 336 } 337 338 static int __init scmi_clock_init(void) 339 { 340 return scmi_protocol_register(SCMI_PROTOCOL_CLOCK, 341 &scmi_clock_protocol_init); 342 } 343 subsys_initcall(scmi_clock_init); 344