1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) 2020 Linaro Ltd
4  */
5 
6 #include <linux/clk.h>
7 #include <linux/device.h>
8 #include <linux/interconnect-provider.h>
9 #include <linux/io.h>
10 #include <linux/module.h>
11 #include <linux/of_device.h>
12 #include <linux/of_platform.h>
13 #include <linux/platform_device.h>
14 #include <linux/pm_domain.h>
15 #include <linux/regmap.h>
16 #include <linux/slab.h>
17 
18 #include "smd-rpm.h"
19 #include "icc-common.h"
20 #include "icc-rpm.h"
21 
22 /* QNOC QoS */
23 #define QNOC_QOS_MCTL_LOWn_ADDR(n)	(0x8 + (n * 0x1000))
24 #define QNOC_QOS_MCTL_DFLT_PRIO_MASK	0x70
25 #define QNOC_QOS_MCTL_DFLT_PRIO_SHIFT	4
26 #define QNOC_QOS_MCTL_URGFWD_EN_MASK	0x8
27 #define QNOC_QOS_MCTL_URGFWD_EN_SHIFT	3
28 
29 /* BIMC QoS */
30 #define M_BKE_REG_BASE(n)		(0x300 + (0x4000 * n))
31 #define M_BKE_EN_ADDR(n)		(M_BKE_REG_BASE(n))
32 #define M_BKE_HEALTH_CFG_ADDR(i, n)	(M_BKE_REG_BASE(n) + 0x40 + (0x4 * i))
33 
34 #define M_BKE_HEALTH_CFG_LIMITCMDS_MASK	0x80000000
35 #define M_BKE_HEALTH_CFG_AREQPRIO_MASK	0x300
36 #define M_BKE_HEALTH_CFG_PRIOLVL_MASK	0x3
37 #define M_BKE_HEALTH_CFG_AREQPRIO_SHIFT	0x8
38 #define M_BKE_HEALTH_CFG_LIMITCMDS_SHIFT 0x1f
39 
40 #define M_BKE_EN_EN_BMASK		0x1
41 
42 /* NoC QoS */
43 #define NOC_QOS_PRIORITYn_ADDR(n)	(0x8 + (n * 0x1000))
44 #define NOC_QOS_PRIORITY_P1_MASK	0xc
45 #define NOC_QOS_PRIORITY_P0_MASK	0x3
46 #define NOC_QOS_PRIORITY_P1_SHIFT	0x2
47 
48 #define NOC_QOS_MODEn_ADDR(n)		(0xc + (n * 0x1000))
49 #define NOC_QOS_MODEn_MASK		0x3
50 
51 static int qcom_icc_set_qnoc_qos(struct icc_node *src, u64 max_bw)
52 {
53 	struct icc_provider *provider = src->provider;
54 	struct qcom_icc_provider *qp = to_qcom_provider(provider);
55 	struct qcom_icc_node *qn = src->data;
56 	struct qcom_icc_qos *qos = &qn->qos;
57 	int rc;
58 
59 	rc = regmap_update_bits(qp->regmap,
60 			qp->qos_offset + QNOC_QOS_MCTL_LOWn_ADDR(qos->qos_port),
61 			QNOC_QOS_MCTL_DFLT_PRIO_MASK,
62 			qos->areq_prio << QNOC_QOS_MCTL_DFLT_PRIO_SHIFT);
63 	if (rc)
64 		return rc;
65 
66 	return regmap_update_bits(qp->regmap,
67 			qp->qos_offset + QNOC_QOS_MCTL_LOWn_ADDR(qos->qos_port),
68 			QNOC_QOS_MCTL_URGFWD_EN_MASK,
69 			!!qos->urg_fwd_en << QNOC_QOS_MCTL_URGFWD_EN_SHIFT);
70 }
71 
72 static int qcom_icc_bimc_set_qos_health(struct qcom_icc_provider *qp,
73 					struct qcom_icc_qos *qos,
74 					int regnum)
75 {
76 	u32 val;
77 	u32 mask;
78 
79 	val = qos->prio_level;
80 	mask = M_BKE_HEALTH_CFG_PRIOLVL_MASK;
81 
82 	val |= qos->areq_prio << M_BKE_HEALTH_CFG_AREQPRIO_SHIFT;
83 	mask |= M_BKE_HEALTH_CFG_AREQPRIO_MASK;
84 
85 	/* LIMITCMDS is not present on M_BKE_HEALTH_3 */
86 	if (regnum != 3) {
87 		val |= qos->limit_commands << M_BKE_HEALTH_CFG_LIMITCMDS_SHIFT;
88 		mask |= M_BKE_HEALTH_CFG_LIMITCMDS_MASK;
89 	}
90 
91 	return regmap_update_bits(qp->regmap,
92 				  qp->qos_offset + M_BKE_HEALTH_CFG_ADDR(regnum, qos->qos_port),
93 				  mask, val);
94 }
95 
96 static int qcom_icc_set_bimc_qos(struct icc_node *src, u64 max_bw)
97 {
98 	struct qcom_icc_provider *qp;
99 	struct qcom_icc_node *qn;
100 	struct icc_provider *provider;
101 	u32 mode = NOC_QOS_MODE_BYPASS;
102 	u32 val = 0;
103 	int i, rc = 0;
104 
105 	qn = src->data;
106 	provider = src->provider;
107 	qp = to_qcom_provider(provider);
108 
109 	if (qn->qos.qos_mode != NOC_QOS_MODE_INVALID)
110 		mode = qn->qos.qos_mode;
111 
112 	/* QoS Priority: The QoS Health parameters are getting considered
113 	 * only if we are NOT in Bypass Mode.
114 	 */
115 	if (mode != NOC_QOS_MODE_BYPASS) {
116 		for (i = 3; i >= 0; i--) {
117 			rc = qcom_icc_bimc_set_qos_health(qp,
118 							  &qn->qos, i);
119 			if (rc)
120 				return rc;
121 		}
122 
123 		/* Set BKE_EN to 1 when Fixed, Regulator or Limiter Mode */
124 		val = 1;
125 	}
126 
127 	return regmap_update_bits(qp->regmap,
128 				  qp->qos_offset + M_BKE_EN_ADDR(qn->qos.qos_port),
129 				  M_BKE_EN_EN_BMASK, val);
130 }
131 
132 static int qcom_icc_noc_set_qos_priority(struct qcom_icc_provider *qp,
133 					 struct qcom_icc_qos *qos)
134 {
135 	u32 val;
136 	int rc;
137 
138 	/* Must be updated one at a time, P1 first, P0 last */
139 	val = qos->areq_prio << NOC_QOS_PRIORITY_P1_SHIFT;
140 	rc = regmap_update_bits(qp->regmap,
141 				qp->qos_offset + NOC_QOS_PRIORITYn_ADDR(qos->qos_port),
142 				NOC_QOS_PRIORITY_P1_MASK, val);
143 	if (rc)
144 		return rc;
145 
146 	return regmap_update_bits(qp->regmap,
147 				  qp->qos_offset + NOC_QOS_PRIORITYn_ADDR(qos->qos_port),
148 				  NOC_QOS_PRIORITY_P0_MASK, qos->prio_level);
149 }
150 
151 static int qcom_icc_set_noc_qos(struct icc_node *src, u64 max_bw)
152 {
153 	struct qcom_icc_provider *qp;
154 	struct qcom_icc_node *qn;
155 	struct icc_provider *provider;
156 	u32 mode = NOC_QOS_MODE_BYPASS;
157 	int rc = 0;
158 
159 	qn = src->data;
160 	provider = src->provider;
161 	qp = to_qcom_provider(provider);
162 
163 	if (qn->qos.qos_port < 0) {
164 		dev_dbg(src->provider->dev,
165 			"NoC QoS: Skipping %s: vote aggregated on parent.\n",
166 			qn->name);
167 		return 0;
168 	}
169 
170 	if (qn->qos.qos_mode != NOC_QOS_MODE_INVALID)
171 		mode = qn->qos.qos_mode;
172 
173 	if (mode == NOC_QOS_MODE_FIXED) {
174 		dev_dbg(src->provider->dev, "NoC QoS: %s: Set Fixed mode\n",
175 			qn->name);
176 		rc = qcom_icc_noc_set_qos_priority(qp, &qn->qos);
177 		if (rc)
178 			return rc;
179 	} else if (mode == NOC_QOS_MODE_BYPASS) {
180 		dev_dbg(src->provider->dev, "NoC QoS: %s: Set Bypass mode\n",
181 			qn->name);
182 	}
183 
184 	return regmap_update_bits(qp->regmap,
185 				  qp->qos_offset + NOC_QOS_MODEn_ADDR(qn->qos.qos_port),
186 				  NOC_QOS_MODEn_MASK, mode);
187 }
188 
189 static int qcom_icc_qos_set(struct icc_node *node, u64 sum_bw)
190 {
191 	struct qcom_icc_provider *qp = to_qcom_provider(node->provider);
192 	struct qcom_icc_node *qn = node->data;
193 
194 	dev_dbg(node->provider->dev, "Setting QoS for %s\n", qn->name);
195 
196 	switch (qp->type) {
197 	case QCOM_ICC_BIMC:
198 		return qcom_icc_set_bimc_qos(node, sum_bw);
199 	case QCOM_ICC_QNOC:
200 		return qcom_icc_set_qnoc_qos(node, sum_bw);
201 	default:
202 		return qcom_icc_set_noc_qos(node, sum_bw);
203 	}
204 }
205 
206 static int qcom_icc_rpm_set(int mas_rpm_id, int slv_rpm_id, u64 sum_bw)
207 {
208 	int ret = 0;
209 
210 	if (mas_rpm_id != -1) {
211 		ret = qcom_icc_rpm_smd_send(QCOM_SMD_RPM_ACTIVE_STATE,
212 					    RPM_BUS_MASTER_REQ,
213 					    mas_rpm_id,
214 					    sum_bw);
215 		if (ret) {
216 			pr_err("qcom_icc_rpm_smd_send mas %d error %d\n",
217 			       mas_rpm_id, ret);
218 			return ret;
219 		}
220 	}
221 
222 	if (slv_rpm_id != -1) {
223 		ret = qcom_icc_rpm_smd_send(QCOM_SMD_RPM_ACTIVE_STATE,
224 					    RPM_BUS_SLAVE_REQ,
225 					    slv_rpm_id,
226 					    sum_bw);
227 		if (ret) {
228 			pr_err("qcom_icc_rpm_smd_send slv %d error %d\n",
229 			       slv_rpm_id, ret);
230 			return ret;
231 		}
232 	}
233 
234 	return ret;
235 }
236 
237 static int __qcom_icc_set(struct icc_node *n, struct qcom_icc_node *qn,
238 			  u64 sum_bw)
239 {
240 	int ret;
241 
242 	if (!qn->qos.ap_owned) {
243 		/* send bandwidth request message to the RPM processor */
244 		ret = qcom_icc_rpm_set(qn->mas_rpm_id, qn->slv_rpm_id, sum_bw);
245 		if (ret)
246 			return ret;
247 	} else if (qn->qos.qos_mode != -1) {
248 		/* set bandwidth directly from the AP */
249 		ret = qcom_icc_qos_set(n, sum_bw);
250 		if (ret)
251 			return ret;
252 	}
253 
254 	return 0;
255 }
256 
257 /**
258  * qcom_icc_pre_bw_aggregate - cleans up values before re-aggregate requests
259  * @node: icc node to operate on
260  */
261 static void qcom_icc_pre_bw_aggregate(struct icc_node *node)
262 {
263 	struct qcom_icc_node *qn;
264 	size_t i;
265 
266 	qn = node->data;
267 	for (i = 0; i < QCOM_ICC_NUM_BUCKETS; i++) {
268 		qn->sum_avg[i] = 0;
269 		qn->max_peak[i] = 0;
270 	}
271 }
272 
273 /**
274  * qcom_icc_bw_aggregate - aggregate bw for buckets indicated by tag
275  * @node: node to aggregate
276  * @tag: tag to indicate which buckets to aggregate
277  * @avg_bw: new bw to sum aggregate
278  * @peak_bw: new bw to max aggregate
279  * @agg_avg: existing aggregate avg bw val
280  * @agg_peak: existing aggregate peak bw val
281  */
282 static int qcom_icc_bw_aggregate(struct icc_node *node, u32 tag, u32 avg_bw,
283 				 u32 peak_bw, u32 *agg_avg, u32 *agg_peak)
284 {
285 	size_t i;
286 	struct qcom_icc_node *qn;
287 
288 	qn = node->data;
289 
290 	if (!tag)
291 		tag = QCOM_ICC_TAG_ALWAYS;
292 
293 	for (i = 0; i < QCOM_ICC_NUM_BUCKETS; i++) {
294 		if (tag & BIT(i)) {
295 			qn->sum_avg[i] += avg_bw;
296 			qn->max_peak[i] = max_t(u32, qn->max_peak[i], peak_bw);
297 		}
298 	}
299 
300 	*agg_avg += avg_bw;
301 	*agg_peak = max_t(u32, *agg_peak, peak_bw);
302 	return 0;
303 }
304 
305 /**
306  * qcom_icc_bus_aggregate - aggregate bandwidth by traversing all nodes
307  * @provider: generic interconnect provider
308  * @agg_avg: an array for aggregated average bandwidth of buckets
309  * @agg_peak: an array for aggregated peak bandwidth of buckets
310  * @max_agg_avg: pointer to max value of aggregated average bandwidth
311  */
312 static void qcom_icc_bus_aggregate(struct icc_provider *provider,
313 				   u64 *agg_avg, u64 *agg_peak,
314 				   u64 *max_agg_avg)
315 {
316 	struct icc_node *node;
317 	struct qcom_icc_node *qn;
318 	int i;
319 
320 	/* Initialise aggregate values */
321 	for (i = 0; i < QCOM_ICC_NUM_BUCKETS; i++) {
322 		agg_avg[i] = 0;
323 		agg_peak[i] = 0;
324 	}
325 
326 	*max_agg_avg = 0;
327 
328 	/*
329 	 * Iterate nodes on the interconnect and aggregate bandwidth
330 	 * requests for every bucket.
331 	 */
332 	list_for_each_entry(node, &provider->nodes, node_list) {
333 		qn = node->data;
334 		for (i = 0; i < QCOM_ICC_NUM_BUCKETS; i++) {
335 			agg_avg[i] += qn->sum_avg[i];
336 			agg_peak[i] = max_t(u64, agg_peak[i], qn->max_peak[i]);
337 		}
338 	}
339 
340 	/* Find maximum values across all buckets */
341 	for (i = 0; i < QCOM_ICC_NUM_BUCKETS; i++)
342 		*max_agg_avg = max_t(u64, *max_agg_avg, agg_avg[i]);
343 }
344 
345 static int qcom_icc_set(struct icc_node *src, struct icc_node *dst)
346 {
347 	struct qcom_icc_provider *qp;
348 	struct qcom_icc_node *src_qn = NULL, *dst_qn = NULL;
349 	struct icc_provider *provider;
350 	u64 sum_bw;
351 	u64 rate;
352 	u64 agg_avg[QCOM_ICC_NUM_BUCKETS], agg_peak[QCOM_ICC_NUM_BUCKETS];
353 	u64 max_agg_avg;
354 	int ret, i;
355 	int bucket;
356 
357 	src_qn = src->data;
358 	if (dst)
359 		dst_qn = dst->data;
360 	provider = src->provider;
361 	qp = to_qcom_provider(provider);
362 
363 	qcom_icc_bus_aggregate(provider, agg_avg, agg_peak, &max_agg_avg);
364 
365 	sum_bw = icc_units_to_bps(max_agg_avg);
366 
367 	ret = __qcom_icc_set(src, src_qn, sum_bw);
368 	if (ret)
369 		return ret;
370 	if (dst_qn) {
371 		ret = __qcom_icc_set(dst, dst_qn, sum_bw);
372 		if (ret)
373 			return ret;
374 	}
375 
376 	for (i = 0; i < qp->num_clks; i++) {
377 		/*
378 		 * Use WAKE bucket for active clock, otherwise, use SLEEP bucket
379 		 * for other clocks.  If a platform doesn't set interconnect
380 		 * path tags, by default use sleep bucket for all clocks.
381 		 *
382 		 * Note, AMC bucket is not supported yet.
383 		 */
384 		if (!strcmp(qp->bus_clks[i].id, "bus_a"))
385 			bucket = QCOM_ICC_BUCKET_WAKE;
386 		else
387 			bucket = QCOM_ICC_BUCKET_SLEEP;
388 
389 		rate = icc_units_to_bps(max(agg_avg[bucket], agg_peak[bucket]));
390 		do_div(rate, src_qn->buswidth);
391 		rate = min_t(u64, rate, LONG_MAX);
392 
393 		if (qp->bus_clk_rate[i] == rate)
394 			continue;
395 
396 		ret = clk_set_rate(qp->bus_clks[i].clk, rate);
397 		if (ret) {
398 			pr_err("%s clk_set_rate error: %d\n",
399 			       qp->bus_clks[i].id, ret);
400 			return ret;
401 		}
402 		qp->bus_clk_rate[i] = rate;
403 	}
404 
405 	return 0;
406 }
407 
408 static const char * const bus_clocks[] = {
409 	"bus", "bus_a",
410 };
411 
412 int qnoc_probe(struct platform_device *pdev)
413 {
414 	struct device *dev = &pdev->dev;
415 	const struct qcom_icc_desc *desc;
416 	struct icc_onecell_data *data;
417 	struct icc_provider *provider;
418 	struct qcom_icc_node * const *qnodes;
419 	struct qcom_icc_provider *qp;
420 	struct icc_node *node;
421 	size_t num_nodes, i;
422 	const char * const *cds;
423 	int cd_num;
424 	int ret;
425 
426 	/* wait for the RPM proxy */
427 	if (!qcom_icc_rpm_smd_available())
428 		return -EPROBE_DEFER;
429 
430 	desc = of_device_get_match_data(dev);
431 	if (!desc)
432 		return -EINVAL;
433 
434 	qnodes = desc->nodes;
435 	num_nodes = desc->num_nodes;
436 
437 	if (desc->num_clocks) {
438 		cds = desc->clocks;
439 		cd_num = desc->num_clocks;
440 	} else {
441 		cds = bus_clocks;
442 		cd_num = ARRAY_SIZE(bus_clocks);
443 	}
444 
445 	qp = devm_kzalloc(dev, struct_size(qp, bus_clks, cd_num), GFP_KERNEL);
446 	if (!qp)
447 		return -ENOMEM;
448 
449 	qp->bus_clk_rate = devm_kcalloc(dev, cd_num, sizeof(*qp->bus_clk_rate),
450 					GFP_KERNEL);
451 	if (!qp->bus_clk_rate)
452 		return -ENOMEM;
453 
454 	data = devm_kzalloc(dev, struct_size(data, nodes, num_nodes),
455 			    GFP_KERNEL);
456 	if (!data)
457 		return -ENOMEM;
458 
459 	for (i = 0; i < cd_num; i++)
460 		qp->bus_clks[i].id = cds[i];
461 	qp->num_clks = cd_num;
462 
463 	qp->type = desc->type;
464 	qp->qos_offset = desc->qos_offset;
465 
466 	if (desc->regmap_cfg) {
467 		struct resource *res;
468 		void __iomem *mmio;
469 
470 		res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
471 		if (!res) {
472 			/* Try parent's regmap */
473 			qp->regmap = dev_get_regmap(dev->parent, NULL);
474 			if (qp->regmap)
475 				goto regmap_done;
476 			return -ENODEV;
477 		}
478 
479 		mmio = devm_ioremap_resource(dev, res);
480 		if (IS_ERR(mmio))
481 			return PTR_ERR(mmio);
482 
483 		qp->regmap = devm_regmap_init_mmio(dev, mmio, desc->regmap_cfg);
484 		if (IS_ERR(qp->regmap)) {
485 			dev_err(dev, "Cannot regmap interconnect bus resource\n");
486 			return PTR_ERR(qp->regmap);
487 		}
488 	}
489 
490 regmap_done:
491 	ret = devm_clk_bulk_get_optional(dev, qp->num_clks, qp->bus_clks);
492 	if (ret)
493 		return ret;
494 
495 	ret = clk_bulk_prepare_enable(qp->num_clks, qp->bus_clks);
496 	if (ret)
497 		return ret;
498 
499 	if (desc->has_bus_pd) {
500 		ret = dev_pm_domain_attach(dev, true);
501 		if (ret)
502 			return ret;
503 	}
504 
505 	provider = &qp->provider;
506 	provider->dev = dev;
507 	provider->set = qcom_icc_set;
508 	provider->pre_aggregate = qcom_icc_pre_bw_aggregate;
509 	provider->aggregate = qcom_icc_bw_aggregate;
510 	provider->xlate_extended = qcom_icc_xlate_extended;
511 	provider->data = data;
512 
513 	icc_provider_init(provider);
514 
515 	for (i = 0; i < num_nodes; i++) {
516 		size_t j;
517 
518 		node = icc_node_create(qnodes[i]->id);
519 		if (IS_ERR(node)) {
520 			ret = PTR_ERR(node);
521 			goto err_remove_nodes;
522 		}
523 
524 		node->name = qnodes[i]->name;
525 		node->data = qnodes[i];
526 		icc_node_add(node, provider);
527 
528 		for (j = 0; j < qnodes[i]->num_links; j++)
529 			icc_link_create(node, qnodes[i]->links[j]);
530 
531 		data->nodes[i] = node;
532 	}
533 	data->num_nodes = num_nodes;
534 
535 	ret = icc_provider_register(provider);
536 	if (ret)
537 		goto err_remove_nodes;
538 
539 	platform_set_drvdata(pdev, qp);
540 
541 	/* Populate child NoC devices if any */
542 	if (of_get_child_count(dev->of_node) > 0) {
543 		ret = of_platform_populate(dev->of_node, NULL, NULL, dev);
544 		if (ret)
545 			goto err_deregister_provider;
546 	}
547 
548 	return 0;
549 
550 err_deregister_provider:
551 	icc_provider_deregister(provider);
552 err_remove_nodes:
553 	icc_nodes_remove(provider);
554 	clk_bulk_disable_unprepare(qp->num_clks, qp->bus_clks);
555 
556 	return ret;
557 }
558 EXPORT_SYMBOL(qnoc_probe);
559 
560 int qnoc_remove(struct platform_device *pdev)
561 {
562 	struct qcom_icc_provider *qp = platform_get_drvdata(pdev);
563 
564 	icc_provider_deregister(&qp->provider);
565 	icc_nodes_remove(&qp->provider);
566 	clk_bulk_disable_unprepare(qp->num_clks, qp->bus_clks);
567 
568 	return 0;
569 }
570 EXPORT_SYMBOL(qnoc_remove);
571