1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * Interconnect framework core driver
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * Copyright (c) 2017-2019, Linaro Ltd.
6*4882a593Smuzhiyun * Author: Georgi Djakov <georgi.djakov@linaro.org>
7*4882a593Smuzhiyun */
8*4882a593Smuzhiyun
9*4882a593Smuzhiyun #include <linux/debugfs.h>
10*4882a593Smuzhiyun #include <linux/device.h>
11*4882a593Smuzhiyun #include <linux/idr.h>
12*4882a593Smuzhiyun #include <linux/init.h>
13*4882a593Smuzhiyun #include <linux/interconnect.h>
14*4882a593Smuzhiyun #include <linux/interconnect-provider.h>
15*4882a593Smuzhiyun #include <linux/list.h>
16*4882a593Smuzhiyun #include <linux/module.h>
17*4882a593Smuzhiyun #include <linux/mutex.h>
18*4882a593Smuzhiyun #include <linux/slab.h>
19*4882a593Smuzhiyun #include <linux/of.h>
20*4882a593Smuzhiyun #include <linux/overflow.h>
21*4882a593Smuzhiyun
22*4882a593Smuzhiyun #include "internal.h"
23*4882a593Smuzhiyun
24*4882a593Smuzhiyun #define CREATE_TRACE_POINTS
25*4882a593Smuzhiyun #include "trace.h"
26*4882a593Smuzhiyun
27*4882a593Smuzhiyun static DEFINE_IDR(icc_idr);
28*4882a593Smuzhiyun static LIST_HEAD(icc_providers);
29*4882a593Smuzhiyun static int providers_count;
30*4882a593Smuzhiyun static bool synced_state;
31*4882a593Smuzhiyun static DEFINE_MUTEX(icc_lock);
32*4882a593Smuzhiyun static struct dentry *icc_debugfs_dir;
33*4882a593Smuzhiyun
icc_summary_show_one(struct seq_file * s,struct icc_node * n)34*4882a593Smuzhiyun static void icc_summary_show_one(struct seq_file *s, struct icc_node *n)
35*4882a593Smuzhiyun {
36*4882a593Smuzhiyun if (!n)
37*4882a593Smuzhiyun return;
38*4882a593Smuzhiyun
39*4882a593Smuzhiyun seq_printf(s, "%-42s %12u %12u\n",
40*4882a593Smuzhiyun n->name, n->avg_bw, n->peak_bw);
41*4882a593Smuzhiyun }
42*4882a593Smuzhiyun
icc_summary_show(struct seq_file * s,void * data)43*4882a593Smuzhiyun static int icc_summary_show(struct seq_file *s, void *data)
44*4882a593Smuzhiyun {
45*4882a593Smuzhiyun struct icc_provider *provider;
46*4882a593Smuzhiyun
47*4882a593Smuzhiyun seq_puts(s, " node tag avg peak\n");
48*4882a593Smuzhiyun seq_puts(s, "--------------------------------------------------------------------\n");
49*4882a593Smuzhiyun
50*4882a593Smuzhiyun mutex_lock(&icc_lock);
51*4882a593Smuzhiyun
52*4882a593Smuzhiyun list_for_each_entry(provider, &icc_providers, provider_list) {
53*4882a593Smuzhiyun struct icc_node *n;
54*4882a593Smuzhiyun
55*4882a593Smuzhiyun list_for_each_entry(n, &provider->nodes, node_list) {
56*4882a593Smuzhiyun struct icc_req *r;
57*4882a593Smuzhiyun
58*4882a593Smuzhiyun icc_summary_show_one(s, n);
59*4882a593Smuzhiyun hlist_for_each_entry(r, &n->req_list, req_node) {
60*4882a593Smuzhiyun u32 avg_bw = 0, peak_bw = 0;
61*4882a593Smuzhiyun
62*4882a593Smuzhiyun if (!r->dev)
63*4882a593Smuzhiyun continue;
64*4882a593Smuzhiyun
65*4882a593Smuzhiyun if (r->enabled) {
66*4882a593Smuzhiyun avg_bw = r->avg_bw;
67*4882a593Smuzhiyun peak_bw = r->peak_bw;
68*4882a593Smuzhiyun }
69*4882a593Smuzhiyun
70*4882a593Smuzhiyun seq_printf(s, " %-27s %12u %12u %12u\n",
71*4882a593Smuzhiyun dev_name(r->dev), r->tag, avg_bw, peak_bw);
72*4882a593Smuzhiyun }
73*4882a593Smuzhiyun }
74*4882a593Smuzhiyun }
75*4882a593Smuzhiyun
76*4882a593Smuzhiyun mutex_unlock(&icc_lock);
77*4882a593Smuzhiyun
78*4882a593Smuzhiyun return 0;
79*4882a593Smuzhiyun }
80*4882a593Smuzhiyun DEFINE_SHOW_ATTRIBUTE(icc_summary);
81*4882a593Smuzhiyun
icc_graph_show_link(struct seq_file * s,int level,struct icc_node * n,struct icc_node * m)82*4882a593Smuzhiyun static void icc_graph_show_link(struct seq_file *s, int level,
83*4882a593Smuzhiyun struct icc_node *n, struct icc_node *m)
84*4882a593Smuzhiyun {
85*4882a593Smuzhiyun seq_printf(s, "%s\"%d:%s\" -> \"%d:%s\"\n",
86*4882a593Smuzhiyun level == 2 ? "\t\t" : "\t",
87*4882a593Smuzhiyun n->id, n->name, m->id, m->name);
88*4882a593Smuzhiyun }
89*4882a593Smuzhiyun
icc_graph_show_node(struct seq_file * s,struct icc_node * n)90*4882a593Smuzhiyun static void icc_graph_show_node(struct seq_file *s, struct icc_node *n)
91*4882a593Smuzhiyun {
92*4882a593Smuzhiyun seq_printf(s, "\t\t\"%d:%s\" [label=\"%d:%s",
93*4882a593Smuzhiyun n->id, n->name, n->id, n->name);
94*4882a593Smuzhiyun seq_printf(s, "\n\t\t\t|avg_bw=%ukBps", n->avg_bw);
95*4882a593Smuzhiyun seq_printf(s, "\n\t\t\t|peak_bw=%ukBps", n->peak_bw);
96*4882a593Smuzhiyun seq_puts(s, "\"]\n");
97*4882a593Smuzhiyun }
98*4882a593Smuzhiyun
icc_graph_show(struct seq_file * s,void * data)99*4882a593Smuzhiyun static int icc_graph_show(struct seq_file *s, void *data)
100*4882a593Smuzhiyun {
101*4882a593Smuzhiyun struct icc_provider *provider;
102*4882a593Smuzhiyun struct icc_node *n;
103*4882a593Smuzhiyun int cluster_index = 0;
104*4882a593Smuzhiyun int i;
105*4882a593Smuzhiyun
106*4882a593Smuzhiyun seq_puts(s, "digraph {\n\trankdir = LR\n\tnode [shape = record]\n");
107*4882a593Smuzhiyun mutex_lock(&icc_lock);
108*4882a593Smuzhiyun
109*4882a593Smuzhiyun /* draw providers as cluster subgraphs */
110*4882a593Smuzhiyun cluster_index = 0;
111*4882a593Smuzhiyun list_for_each_entry(provider, &icc_providers, provider_list) {
112*4882a593Smuzhiyun seq_printf(s, "\tsubgraph cluster_%d {\n", ++cluster_index);
113*4882a593Smuzhiyun if (provider->dev)
114*4882a593Smuzhiyun seq_printf(s, "\t\tlabel = \"%s\"\n",
115*4882a593Smuzhiyun dev_name(provider->dev));
116*4882a593Smuzhiyun
117*4882a593Smuzhiyun /* draw nodes */
118*4882a593Smuzhiyun list_for_each_entry(n, &provider->nodes, node_list)
119*4882a593Smuzhiyun icc_graph_show_node(s, n);
120*4882a593Smuzhiyun
121*4882a593Smuzhiyun /* draw internal links */
122*4882a593Smuzhiyun list_for_each_entry(n, &provider->nodes, node_list)
123*4882a593Smuzhiyun for (i = 0; i < n->num_links; ++i)
124*4882a593Smuzhiyun if (n->provider == n->links[i]->provider)
125*4882a593Smuzhiyun icc_graph_show_link(s, 2, n,
126*4882a593Smuzhiyun n->links[i]);
127*4882a593Smuzhiyun
128*4882a593Smuzhiyun seq_puts(s, "\t}\n");
129*4882a593Smuzhiyun }
130*4882a593Smuzhiyun
131*4882a593Smuzhiyun /* draw external links */
132*4882a593Smuzhiyun list_for_each_entry(provider, &icc_providers, provider_list)
133*4882a593Smuzhiyun list_for_each_entry(n, &provider->nodes, node_list)
134*4882a593Smuzhiyun for (i = 0; i < n->num_links; ++i)
135*4882a593Smuzhiyun if (n->provider != n->links[i]->provider)
136*4882a593Smuzhiyun icc_graph_show_link(s, 1, n,
137*4882a593Smuzhiyun n->links[i]);
138*4882a593Smuzhiyun
139*4882a593Smuzhiyun mutex_unlock(&icc_lock);
140*4882a593Smuzhiyun seq_puts(s, "}");
141*4882a593Smuzhiyun
142*4882a593Smuzhiyun return 0;
143*4882a593Smuzhiyun }
144*4882a593Smuzhiyun DEFINE_SHOW_ATTRIBUTE(icc_graph);
145*4882a593Smuzhiyun
node_find(const int id)146*4882a593Smuzhiyun static struct icc_node *node_find(const int id)
147*4882a593Smuzhiyun {
148*4882a593Smuzhiyun return idr_find(&icc_idr, id);
149*4882a593Smuzhiyun }
150*4882a593Smuzhiyun
path_init(struct device * dev,struct icc_node * dst,ssize_t num_nodes)151*4882a593Smuzhiyun static struct icc_path *path_init(struct device *dev, struct icc_node *dst,
152*4882a593Smuzhiyun ssize_t num_nodes)
153*4882a593Smuzhiyun {
154*4882a593Smuzhiyun struct icc_node *node = dst;
155*4882a593Smuzhiyun struct icc_path *path;
156*4882a593Smuzhiyun int i;
157*4882a593Smuzhiyun
158*4882a593Smuzhiyun path = kzalloc(struct_size(path, reqs, num_nodes), GFP_KERNEL);
159*4882a593Smuzhiyun if (!path)
160*4882a593Smuzhiyun return ERR_PTR(-ENOMEM);
161*4882a593Smuzhiyun
162*4882a593Smuzhiyun path->num_nodes = num_nodes;
163*4882a593Smuzhiyun
164*4882a593Smuzhiyun for (i = num_nodes - 1; i >= 0; i--) {
165*4882a593Smuzhiyun node->provider->users++;
166*4882a593Smuzhiyun hlist_add_head(&path->reqs[i].req_node, &node->req_list);
167*4882a593Smuzhiyun path->reqs[i].node = node;
168*4882a593Smuzhiyun path->reqs[i].dev = dev;
169*4882a593Smuzhiyun path->reqs[i].enabled = true;
170*4882a593Smuzhiyun /* reference to previous node was saved during path traversal */
171*4882a593Smuzhiyun node = node->reverse;
172*4882a593Smuzhiyun }
173*4882a593Smuzhiyun
174*4882a593Smuzhiyun return path;
175*4882a593Smuzhiyun }
176*4882a593Smuzhiyun
path_find(struct device * dev,struct icc_node * src,struct icc_node * dst)177*4882a593Smuzhiyun static struct icc_path *path_find(struct device *dev, struct icc_node *src,
178*4882a593Smuzhiyun struct icc_node *dst)
179*4882a593Smuzhiyun {
180*4882a593Smuzhiyun struct icc_path *path = ERR_PTR(-EPROBE_DEFER);
181*4882a593Smuzhiyun struct icc_node *n, *node = NULL;
182*4882a593Smuzhiyun struct list_head traverse_list;
183*4882a593Smuzhiyun struct list_head edge_list;
184*4882a593Smuzhiyun struct list_head visited_list;
185*4882a593Smuzhiyun size_t i, depth = 1;
186*4882a593Smuzhiyun bool found = false;
187*4882a593Smuzhiyun
188*4882a593Smuzhiyun INIT_LIST_HEAD(&traverse_list);
189*4882a593Smuzhiyun INIT_LIST_HEAD(&edge_list);
190*4882a593Smuzhiyun INIT_LIST_HEAD(&visited_list);
191*4882a593Smuzhiyun
192*4882a593Smuzhiyun list_add(&src->search_list, &traverse_list);
193*4882a593Smuzhiyun src->reverse = NULL;
194*4882a593Smuzhiyun
195*4882a593Smuzhiyun do {
196*4882a593Smuzhiyun list_for_each_entry_safe(node, n, &traverse_list, search_list) {
197*4882a593Smuzhiyun if (node == dst) {
198*4882a593Smuzhiyun found = true;
199*4882a593Smuzhiyun list_splice_init(&edge_list, &visited_list);
200*4882a593Smuzhiyun list_splice_init(&traverse_list, &visited_list);
201*4882a593Smuzhiyun break;
202*4882a593Smuzhiyun }
203*4882a593Smuzhiyun for (i = 0; i < node->num_links; i++) {
204*4882a593Smuzhiyun struct icc_node *tmp = node->links[i];
205*4882a593Smuzhiyun
206*4882a593Smuzhiyun if (!tmp) {
207*4882a593Smuzhiyun path = ERR_PTR(-ENOENT);
208*4882a593Smuzhiyun goto out;
209*4882a593Smuzhiyun }
210*4882a593Smuzhiyun
211*4882a593Smuzhiyun if (tmp->is_traversed)
212*4882a593Smuzhiyun continue;
213*4882a593Smuzhiyun
214*4882a593Smuzhiyun tmp->is_traversed = true;
215*4882a593Smuzhiyun tmp->reverse = node;
216*4882a593Smuzhiyun list_add_tail(&tmp->search_list, &edge_list);
217*4882a593Smuzhiyun }
218*4882a593Smuzhiyun }
219*4882a593Smuzhiyun
220*4882a593Smuzhiyun if (found)
221*4882a593Smuzhiyun break;
222*4882a593Smuzhiyun
223*4882a593Smuzhiyun list_splice_init(&traverse_list, &visited_list);
224*4882a593Smuzhiyun list_splice_init(&edge_list, &traverse_list);
225*4882a593Smuzhiyun
226*4882a593Smuzhiyun /* count the hops including the source */
227*4882a593Smuzhiyun depth++;
228*4882a593Smuzhiyun
229*4882a593Smuzhiyun } while (!list_empty(&traverse_list));
230*4882a593Smuzhiyun
231*4882a593Smuzhiyun out:
232*4882a593Smuzhiyun
233*4882a593Smuzhiyun /* reset the traversed state */
234*4882a593Smuzhiyun list_for_each_entry_reverse(n, &visited_list, search_list)
235*4882a593Smuzhiyun n->is_traversed = false;
236*4882a593Smuzhiyun
237*4882a593Smuzhiyun if (found)
238*4882a593Smuzhiyun path = path_init(dev, dst, depth);
239*4882a593Smuzhiyun
240*4882a593Smuzhiyun return path;
241*4882a593Smuzhiyun }
242*4882a593Smuzhiyun
243*4882a593Smuzhiyun /*
244*4882a593Smuzhiyun * We want the path to honor all bandwidth requests, so the average and peak
245*4882a593Smuzhiyun * bandwidth requirements from each consumer are aggregated at each node.
246*4882a593Smuzhiyun * The aggregation is platform specific, so each platform can customize it by
247*4882a593Smuzhiyun * implementing its own aggregate() function.
248*4882a593Smuzhiyun */
249*4882a593Smuzhiyun
aggregate_requests(struct icc_node * node)250*4882a593Smuzhiyun static int aggregate_requests(struct icc_node *node)
251*4882a593Smuzhiyun {
252*4882a593Smuzhiyun struct icc_provider *p = node->provider;
253*4882a593Smuzhiyun struct icc_req *r;
254*4882a593Smuzhiyun u32 avg_bw, peak_bw;
255*4882a593Smuzhiyun
256*4882a593Smuzhiyun node->avg_bw = 0;
257*4882a593Smuzhiyun node->peak_bw = 0;
258*4882a593Smuzhiyun
259*4882a593Smuzhiyun if (p->pre_aggregate)
260*4882a593Smuzhiyun p->pre_aggregate(node);
261*4882a593Smuzhiyun
262*4882a593Smuzhiyun hlist_for_each_entry(r, &node->req_list, req_node) {
263*4882a593Smuzhiyun if (r->enabled) {
264*4882a593Smuzhiyun avg_bw = r->avg_bw;
265*4882a593Smuzhiyun peak_bw = r->peak_bw;
266*4882a593Smuzhiyun } else {
267*4882a593Smuzhiyun avg_bw = 0;
268*4882a593Smuzhiyun peak_bw = 0;
269*4882a593Smuzhiyun }
270*4882a593Smuzhiyun p->aggregate(node, r->tag, avg_bw, peak_bw,
271*4882a593Smuzhiyun &node->avg_bw, &node->peak_bw);
272*4882a593Smuzhiyun
273*4882a593Smuzhiyun /* during boot use the initial bandwidth as a floor value */
274*4882a593Smuzhiyun if (!synced_state) {
275*4882a593Smuzhiyun node->avg_bw = max(node->avg_bw, node->init_avg);
276*4882a593Smuzhiyun node->peak_bw = max(node->peak_bw, node->init_peak);
277*4882a593Smuzhiyun }
278*4882a593Smuzhiyun }
279*4882a593Smuzhiyun
280*4882a593Smuzhiyun return 0;
281*4882a593Smuzhiyun }
282*4882a593Smuzhiyun
apply_constraints(struct icc_path * path)283*4882a593Smuzhiyun static int apply_constraints(struct icc_path *path)
284*4882a593Smuzhiyun {
285*4882a593Smuzhiyun struct icc_node *next, *prev = NULL;
286*4882a593Smuzhiyun struct icc_provider *p;
287*4882a593Smuzhiyun int ret = -EINVAL;
288*4882a593Smuzhiyun int i;
289*4882a593Smuzhiyun
290*4882a593Smuzhiyun for (i = 0; i < path->num_nodes; i++) {
291*4882a593Smuzhiyun next = path->reqs[i].node;
292*4882a593Smuzhiyun p = next->provider;
293*4882a593Smuzhiyun
294*4882a593Smuzhiyun /* both endpoints should be valid master-slave pairs */
295*4882a593Smuzhiyun if (!prev || (p != prev->provider && !p->inter_set)) {
296*4882a593Smuzhiyun prev = next;
297*4882a593Smuzhiyun continue;
298*4882a593Smuzhiyun }
299*4882a593Smuzhiyun
300*4882a593Smuzhiyun /* set the constraints */
301*4882a593Smuzhiyun ret = p->set(prev, next);
302*4882a593Smuzhiyun if (ret)
303*4882a593Smuzhiyun goto out;
304*4882a593Smuzhiyun
305*4882a593Smuzhiyun prev = next;
306*4882a593Smuzhiyun }
307*4882a593Smuzhiyun out:
308*4882a593Smuzhiyun return ret;
309*4882a593Smuzhiyun }
310*4882a593Smuzhiyun
icc_std_aggregate(struct icc_node * node,u32 tag,u32 avg_bw,u32 peak_bw,u32 * agg_avg,u32 * agg_peak)311*4882a593Smuzhiyun int icc_std_aggregate(struct icc_node *node, u32 tag, u32 avg_bw,
312*4882a593Smuzhiyun u32 peak_bw, u32 *agg_avg, u32 *agg_peak)
313*4882a593Smuzhiyun {
314*4882a593Smuzhiyun *agg_avg += avg_bw;
315*4882a593Smuzhiyun *agg_peak = max(*agg_peak, peak_bw);
316*4882a593Smuzhiyun
317*4882a593Smuzhiyun return 0;
318*4882a593Smuzhiyun }
319*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_std_aggregate);
320*4882a593Smuzhiyun
321*4882a593Smuzhiyun /* of_icc_xlate_onecell() - Translate function using a single index.
322*4882a593Smuzhiyun * @spec: OF phandle args to map into an interconnect node.
323*4882a593Smuzhiyun * @data: private data (pointer to struct icc_onecell_data)
324*4882a593Smuzhiyun *
325*4882a593Smuzhiyun * This is a generic translate function that can be used to model simple
326*4882a593Smuzhiyun * interconnect providers that have one device tree node and provide
327*4882a593Smuzhiyun * multiple interconnect nodes. A single cell is used as an index into
328*4882a593Smuzhiyun * an array of icc nodes specified in the icc_onecell_data struct when
329*4882a593Smuzhiyun * registering the provider.
330*4882a593Smuzhiyun */
of_icc_xlate_onecell(struct of_phandle_args * spec,void * data)331*4882a593Smuzhiyun struct icc_node *of_icc_xlate_onecell(struct of_phandle_args *spec,
332*4882a593Smuzhiyun void *data)
333*4882a593Smuzhiyun {
334*4882a593Smuzhiyun struct icc_onecell_data *icc_data = data;
335*4882a593Smuzhiyun unsigned int idx = spec->args[0];
336*4882a593Smuzhiyun
337*4882a593Smuzhiyun if (idx >= icc_data->num_nodes) {
338*4882a593Smuzhiyun pr_err("%s: invalid index %u\n", __func__, idx);
339*4882a593Smuzhiyun return ERR_PTR(-EINVAL);
340*4882a593Smuzhiyun }
341*4882a593Smuzhiyun
342*4882a593Smuzhiyun return icc_data->nodes[idx];
343*4882a593Smuzhiyun }
344*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(of_icc_xlate_onecell);
345*4882a593Smuzhiyun
346*4882a593Smuzhiyun /**
347*4882a593Smuzhiyun * of_icc_get_from_provider() - Look-up interconnect node
348*4882a593Smuzhiyun * @spec: OF phandle args to use for look-up
349*4882a593Smuzhiyun *
350*4882a593Smuzhiyun * Looks for interconnect provider under the node specified by @spec and if
351*4882a593Smuzhiyun * found, uses xlate function of the provider to map phandle args to node.
352*4882a593Smuzhiyun *
353*4882a593Smuzhiyun * Returns a valid pointer to struct icc_node_data on success or ERR_PTR()
354*4882a593Smuzhiyun * on failure.
355*4882a593Smuzhiyun */
of_icc_get_from_provider(struct of_phandle_args * spec)356*4882a593Smuzhiyun struct icc_node_data *of_icc_get_from_provider(struct of_phandle_args *spec)
357*4882a593Smuzhiyun {
358*4882a593Smuzhiyun struct icc_node *node = ERR_PTR(-EPROBE_DEFER);
359*4882a593Smuzhiyun struct icc_node_data *data = NULL;
360*4882a593Smuzhiyun struct icc_provider *provider;
361*4882a593Smuzhiyun
362*4882a593Smuzhiyun if (!spec)
363*4882a593Smuzhiyun return ERR_PTR(-EINVAL);
364*4882a593Smuzhiyun
365*4882a593Smuzhiyun mutex_lock(&icc_lock);
366*4882a593Smuzhiyun list_for_each_entry(provider, &icc_providers, provider_list) {
367*4882a593Smuzhiyun if (provider->dev->of_node == spec->np) {
368*4882a593Smuzhiyun if (provider->xlate_extended) {
369*4882a593Smuzhiyun data = provider->xlate_extended(spec, provider->data);
370*4882a593Smuzhiyun if (!IS_ERR(data)) {
371*4882a593Smuzhiyun node = data->node;
372*4882a593Smuzhiyun break;
373*4882a593Smuzhiyun }
374*4882a593Smuzhiyun } else {
375*4882a593Smuzhiyun node = provider->xlate(spec, provider->data);
376*4882a593Smuzhiyun if (!IS_ERR(node))
377*4882a593Smuzhiyun break;
378*4882a593Smuzhiyun }
379*4882a593Smuzhiyun }
380*4882a593Smuzhiyun }
381*4882a593Smuzhiyun mutex_unlock(&icc_lock);
382*4882a593Smuzhiyun
383*4882a593Smuzhiyun if (IS_ERR(node))
384*4882a593Smuzhiyun return ERR_CAST(node);
385*4882a593Smuzhiyun
386*4882a593Smuzhiyun if (!data) {
387*4882a593Smuzhiyun data = kzalloc(sizeof(*data), GFP_KERNEL);
388*4882a593Smuzhiyun if (!data)
389*4882a593Smuzhiyun return ERR_PTR(-ENOMEM);
390*4882a593Smuzhiyun data->node = node;
391*4882a593Smuzhiyun }
392*4882a593Smuzhiyun
393*4882a593Smuzhiyun return data;
394*4882a593Smuzhiyun }
395*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(of_icc_get_from_provider);
396*4882a593Smuzhiyun
devm_icc_release(struct device * dev,void * res)397*4882a593Smuzhiyun static void devm_icc_release(struct device *dev, void *res)
398*4882a593Smuzhiyun {
399*4882a593Smuzhiyun icc_put(*(struct icc_path **)res);
400*4882a593Smuzhiyun }
401*4882a593Smuzhiyun
devm_of_icc_get(struct device * dev,const char * name)402*4882a593Smuzhiyun struct icc_path *devm_of_icc_get(struct device *dev, const char *name)
403*4882a593Smuzhiyun {
404*4882a593Smuzhiyun struct icc_path **ptr, *path;
405*4882a593Smuzhiyun
406*4882a593Smuzhiyun ptr = devres_alloc(devm_icc_release, sizeof(*ptr), GFP_KERNEL);
407*4882a593Smuzhiyun if (!ptr)
408*4882a593Smuzhiyun return ERR_PTR(-ENOMEM);
409*4882a593Smuzhiyun
410*4882a593Smuzhiyun path = of_icc_get(dev, name);
411*4882a593Smuzhiyun if (!IS_ERR(path)) {
412*4882a593Smuzhiyun *ptr = path;
413*4882a593Smuzhiyun devres_add(dev, ptr);
414*4882a593Smuzhiyun } else {
415*4882a593Smuzhiyun devres_free(ptr);
416*4882a593Smuzhiyun }
417*4882a593Smuzhiyun
418*4882a593Smuzhiyun return path;
419*4882a593Smuzhiyun }
420*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(devm_of_icc_get);
421*4882a593Smuzhiyun
422*4882a593Smuzhiyun /**
423*4882a593Smuzhiyun * of_icc_get_by_index() - get a path handle from a DT node based on index
424*4882a593Smuzhiyun * @dev: device pointer for the consumer device
425*4882a593Smuzhiyun * @idx: interconnect path index
426*4882a593Smuzhiyun *
427*4882a593Smuzhiyun * This function will search for a path between two endpoints and return an
428*4882a593Smuzhiyun * icc_path handle on success. Use icc_put() to release constraints when they
429*4882a593Smuzhiyun * are not needed anymore.
430*4882a593Smuzhiyun * If the interconnect API is disabled, NULL is returned and the consumer
431*4882a593Smuzhiyun * drivers will still build. Drivers are free to handle this specifically,
432*4882a593Smuzhiyun * but they don't have to.
433*4882a593Smuzhiyun *
434*4882a593Smuzhiyun * Return: icc_path pointer on success or ERR_PTR() on error. NULL is returned
435*4882a593Smuzhiyun * when the API is disabled or the "interconnects" DT property is missing.
436*4882a593Smuzhiyun */
of_icc_get_by_index(struct device * dev,int idx)437*4882a593Smuzhiyun struct icc_path *of_icc_get_by_index(struct device *dev, int idx)
438*4882a593Smuzhiyun {
439*4882a593Smuzhiyun struct icc_path *path;
440*4882a593Smuzhiyun struct icc_node_data *src_data, *dst_data;
441*4882a593Smuzhiyun struct device_node *np;
442*4882a593Smuzhiyun struct of_phandle_args src_args, dst_args;
443*4882a593Smuzhiyun int ret;
444*4882a593Smuzhiyun
445*4882a593Smuzhiyun if (!dev || !dev->of_node)
446*4882a593Smuzhiyun return ERR_PTR(-ENODEV);
447*4882a593Smuzhiyun
448*4882a593Smuzhiyun np = dev->of_node;
449*4882a593Smuzhiyun
450*4882a593Smuzhiyun /*
451*4882a593Smuzhiyun * When the consumer DT node do not have "interconnects" property
452*4882a593Smuzhiyun * return a NULL path to skip setting constraints.
453*4882a593Smuzhiyun */
454*4882a593Smuzhiyun if (!of_find_property(np, "interconnects", NULL))
455*4882a593Smuzhiyun return NULL;
456*4882a593Smuzhiyun
457*4882a593Smuzhiyun /*
458*4882a593Smuzhiyun * We use a combination of phandle and specifier for endpoint. For now
459*4882a593Smuzhiyun * lets support only global ids and extend this in the future if needed
460*4882a593Smuzhiyun * without breaking DT compatibility.
461*4882a593Smuzhiyun */
462*4882a593Smuzhiyun ret = of_parse_phandle_with_args(np, "interconnects",
463*4882a593Smuzhiyun "#interconnect-cells", idx * 2,
464*4882a593Smuzhiyun &src_args);
465*4882a593Smuzhiyun if (ret)
466*4882a593Smuzhiyun return ERR_PTR(ret);
467*4882a593Smuzhiyun
468*4882a593Smuzhiyun of_node_put(src_args.np);
469*4882a593Smuzhiyun
470*4882a593Smuzhiyun ret = of_parse_phandle_with_args(np, "interconnects",
471*4882a593Smuzhiyun "#interconnect-cells", idx * 2 + 1,
472*4882a593Smuzhiyun &dst_args);
473*4882a593Smuzhiyun if (ret)
474*4882a593Smuzhiyun return ERR_PTR(ret);
475*4882a593Smuzhiyun
476*4882a593Smuzhiyun of_node_put(dst_args.np);
477*4882a593Smuzhiyun
478*4882a593Smuzhiyun src_data = of_icc_get_from_provider(&src_args);
479*4882a593Smuzhiyun
480*4882a593Smuzhiyun if (IS_ERR(src_data)) {
481*4882a593Smuzhiyun dev_err_probe(dev, PTR_ERR(src_data), "error finding src node\n");
482*4882a593Smuzhiyun return ERR_CAST(src_data);
483*4882a593Smuzhiyun }
484*4882a593Smuzhiyun
485*4882a593Smuzhiyun dst_data = of_icc_get_from_provider(&dst_args);
486*4882a593Smuzhiyun
487*4882a593Smuzhiyun if (IS_ERR(dst_data)) {
488*4882a593Smuzhiyun dev_err_probe(dev, PTR_ERR(dst_data), "error finding dst node\n");
489*4882a593Smuzhiyun kfree(src_data);
490*4882a593Smuzhiyun return ERR_CAST(dst_data);
491*4882a593Smuzhiyun }
492*4882a593Smuzhiyun
493*4882a593Smuzhiyun mutex_lock(&icc_lock);
494*4882a593Smuzhiyun path = path_find(dev, src_data->node, dst_data->node);
495*4882a593Smuzhiyun mutex_unlock(&icc_lock);
496*4882a593Smuzhiyun if (IS_ERR(path)) {
497*4882a593Smuzhiyun dev_err(dev, "%s: invalid path=%ld\n", __func__, PTR_ERR(path));
498*4882a593Smuzhiyun goto free_icc_data;
499*4882a593Smuzhiyun }
500*4882a593Smuzhiyun
501*4882a593Smuzhiyun if (src_data->tag && src_data->tag == dst_data->tag)
502*4882a593Smuzhiyun icc_set_tag(path, src_data->tag);
503*4882a593Smuzhiyun
504*4882a593Smuzhiyun path->name = kasprintf(GFP_KERNEL, "%s-%s",
505*4882a593Smuzhiyun src_data->node->name, dst_data->node->name);
506*4882a593Smuzhiyun if (!path->name) {
507*4882a593Smuzhiyun kfree(path);
508*4882a593Smuzhiyun path = ERR_PTR(-ENOMEM);
509*4882a593Smuzhiyun }
510*4882a593Smuzhiyun
511*4882a593Smuzhiyun free_icc_data:
512*4882a593Smuzhiyun kfree(src_data);
513*4882a593Smuzhiyun kfree(dst_data);
514*4882a593Smuzhiyun return path;
515*4882a593Smuzhiyun }
516*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(of_icc_get_by_index);
517*4882a593Smuzhiyun
518*4882a593Smuzhiyun /**
519*4882a593Smuzhiyun * of_icc_get() - get a path handle from a DT node based on name
520*4882a593Smuzhiyun * @dev: device pointer for the consumer device
521*4882a593Smuzhiyun * @name: interconnect path name
522*4882a593Smuzhiyun *
523*4882a593Smuzhiyun * This function will search for a path between two endpoints and return an
524*4882a593Smuzhiyun * icc_path handle on success. Use icc_put() to release constraints when they
525*4882a593Smuzhiyun * are not needed anymore.
526*4882a593Smuzhiyun * If the interconnect API is disabled, NULL is returned and the consumer
527*4882a593Smuzhiyun * drivers will still build. Drivers are free to handle this specifically,
528*4882a593Smuzhiyun * but they don't have to.
529*4882a593Smuzhiyun *
530*4882a593Smuzhiyun * Return: icc_path pointer on success or ERR_PTR() on error. NULL is returned
531*4882a593Smuzhiyun * when the API is disabled or the "interconnects" DT property is missing.
532*4882a593Smuzhiyun */
of_icc_get(struct device * dev,const char * name)533*4882a593Smuzhiyun struct icc_path *of_icc_get(struct device *dev, const char *name)
534*4882a593Smuzhiyun {
535*4882a593Smuzhiyun struct device_node *np;
536*4882a593Smuzhiyun int idx = 0;
537*4882a593Smuzhiyun
538*4882a593Smuzhiyun if (!dev || !dev->of_node)
539*4882a593Smuzhiyun return ERR_PTR(-ENODEV);
540*4882a593Smuzhiyun
541*4882a593Smuzhiyun np = dev->of_node;
542*4882a593Smuzhiyun
543*4882a593Smuzhiyun /*
544*4882a593Smuzhiyun * When the consumer DT node do not have "interconnects" property
545*4882a593Smuzhiyun * return a NULL path to skip setting constraints.
546*4882a593Smuzhiyun */
547*4882a593Smuzhiyun if (!of_find_property(np, "interconnects", NULL))
548*4882a593Smuzhiyun return NULL;
549*4882a593Smuzhiyun
550*4882a593Smuzhiyun /*
551*4882a593Smuzhiyun * We use a combination of phandle and specifier for endpoint. For now
552*4882a593Smuzhiyun * lets support only global ids and extend this in the future if needed
553*4882a593Smuzhiyun * without breaking DT compatibility.
554*4882a593Smuzhiyun */
555*4882a593Smuzhiyun if (name) {
556*4882a593Smuzhiyun idx = of_property_match_string(np, "interconnect-names", name);
557*4882a593Smuzhiyun if (idx < 0)
558*4882a593Smuzhiyun return ERR_PTR(idx);
559*4882a593Smuzhiyun }
560*4882a593Smuzhiyun
561*4882a593Smuzhiyun return of_icc_get_by_index(dev, idx);
562*4882a593Smuzhiyun }
563*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(of_icc_get);
564*4882a593Smuzhiyun
565*4882a593Smuzhiyun /**
566*4882a593Smuzhiyun * icc_set_tag() - set an optional tag on a path
567*4882a593Smuzhiyun * @path: the path we want to tag
568*4882a593Smuzhiyun * @tag: the tag value
569*4882a593Smuzhiyun *
570*4882a593Smuzhiyun * This function allows consumers to append a tag to the requests associated
571*4882a593Smuzhiyun * with a path, so that a different aggregation could be done based on this tag.
572*4882a593Smuzhiyun */
icc_set_tag(struct icc_path * path,u32 tag)573*4882a593Smuzhiyun void icc_set_tag(struct icc_path *path, u32 tag)
574*4882a593Smuzhiyun {
575*4882a593Smuzhiyun int i;
576*4882a593Smuzhiyun
577*4882a593Smuzhiyun if (!path)
578*4882a593Smuzhiyun return;
579*4882a593Smuzhiyun
580*4882a593Smuzhiyun mutex_lock(&icc_lock);
581*4882a593Smuzhiyun
582*4882a593Smuzhiyun for (i = 0; i < path->num_nodes; i++)
583*4882a593Smuzhiyun path->reqs[i].tag = tag;
584*4882a593Smuzhiyun
585*4882a593Smuzhiyun mutex_unlock(&icc_lock);
586*4882a593Smuzhiyun }
587*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_set_tag);
588*4882a593Smuzhiyun
589*4882a593Smuzhiyun /**
590*4882a593Smuzhiyun * icc_get_name() - Get name of the icc path
591*4882a593Smuzhiyun * @path: reference to the path returned by icc_get()
592*4882a593Smuzhiyun *
593*4882a593Smuzhiyun * This function is used by an interconnect consumer to get the name of the icc
594*4882a593Smuzhiyun * path.
595*4882a593Smuzhiyun *
596*4882a593Smuzhiyun * Returns a valid pointer on success, or NULL otherwise.
597*4882a593Smuzhiyun */
icc_get_name(struct icc_path * path)598*4882a593Smuzhiyun const char *icc_get_name(struct icc_path *path)
599*4882a593Smuzhiyun {
600*4882a593Smuzhiyun if (!path)
601*4882a593Smuzhiyun return NULL;
602*4882a593Smuzhiyun
603*4882a593Smuzhiyun return path->name;
604*4882a593Smuzhiyun }
605*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_get_name);
606*4882a593Smuzhiyun
607*4882a593Smuzhiyun /**
608*4882a593Smuzhiyun * icc_set_bw() - set bandwidth constraints on an interconnect path
609*4882a593Smuzhiyun * @path: reference to the path returned by icc_get()
610*4882a593Smuzhiyun * @avg_bw: average bandwidth in kilobytes per second
611*4882a593Smuzhiyun * @peak_bw: peak bandwidth in kilobytes per second
612*4882a593Smuzhiyun *
613*4882a593Smuzhiyun * This function is used by an interconnect consumer to express its own needs
614*4882a593Smuzhiyun * in terms of bandwidth for a previously requested path between two endpoints.
615*4882a593Smuzhiyun * The requests are aggregated and each node is updated accordingly. The entire
616*4882a593Smuzhiyun * path is locked by a mutex to ensure that the set() is completed.
617*4882a593Smuzhiyun * The @path can be NULL when the "interconnects" DT properties is missing,
618*4882a593Smuzhiyun * which will mean that no constraints will be set.
619*4882a593Smuzhiyun *
620*4882a593Smuzhiyun * Returns 0 on success, or an appropriate error code otherwise.
621*4882a593Smuzhiyun */
icc_set_bw(struct icc_path * path,u32 avg_bw,u32 peak_bw)622*4882a593Smuzhiyun int icc_set_bw(struct icc_path *path, u32 avg_bw, u32 peak_bw)
623*4882a593Smuzhiyun {
624*4882a593Smuzhiyun struct icc_node *node;
625*4882a593Smuzhiyun u32 old_avg, old_peak;
626*4882a593Smuzhiyun size_t i;
627*4882a593Smuzhiyun int ret;
628*4882a593Smuzhiyun
629*4882a593Smuzhiyun if (!path)
630*4882a593Smuzhiyun return 0;
631*4882a593Smuzhiyun
632*4882a593Smuzhiyun if (WARN_ON(IS_ERR(path) || !path->num_nodes))
633*4882a593Smuzhiyun return -EINVAL;
634*4882a593Smuzhiyun
635*4882a593Smuzhiyun mutex_lock(&icc_lock);
636*4882a593Smuzhiyun
637*4882a593Smuzhiyun old_avg = path->reqs[0].avg_bw;
638*4882a593Smuzhiyun old_peak = path->reqs[0].peak_bw;
639*4882a593Smuzhiyun
640*4882a593Smuzhiyun for (i = 0; i < path->num_nodes; i++) {
641*4882a593Smuzhiyun node = path->reqs[i].node;
642*4882a593Smuzhiyun
643*4882a593Smuzhiyun /* update the consumer request for this path */
644*4882a593Smuzhiyun path->reqs[i].avg_bw = avg_bw;
645*4882a593Smuzhiyun path->reqs[i].peak_bw = peak_bw;
646*4882a593Smuzhiyun
647*4882a593Smuzhiyun /* aggregate requests for this node */
648*4882a593Smuzhiyun aggregate_requests(node);
649*4882a593Smuzhiyun
650*4882a593Smuzhiyun trace_icc_set_bw(path, node, i, avg_bw, peak_bw);
651*4882a593Smuzhiyun }
652*4882a593Smuzhiyun
653*4882a593Smuzhiyun ret = apply_constraints(path);
654*4882a593Smuzhiyun if (ret) {
655*4882a593Smuzhiyun pr_debug("interconnect: error applying constraints (%d)\n",
656*4882a593Smuzhiyun ret);
657*4882a593Smuzhiyun
658*4882a593Smuzhiyun for (i = 0; i < path->num_nodes; i++) {
659*4882a593Smuzhiyun node = path->reqs[i].node;
660*4882a593Smuzhiyun path->reqs[i].avg_bw = old_avg;
661*4882a593Smuzhiyun path->reqs[i].peak_bw = old_peak;
662*4882a593Smuzhiyun aggregate_requests(node);
663*4882a593Smuzhiyun }
664*4882a593Smuzhiyun apply_constraints(path);
665*4882a593Smuzhiyun }
666*4882a593Smuzhiyun
667*4882a593Smuzhiyun mutex_unlock(&icc_lock);
668*4882a593Smuzhiyun
669*4882a593Smuzhiyun trace_icc_set_bw_end(path, ret);
670*4882a593Smuzhiyun
671*4882a593Smuzhiyun return ret;
672*4882a593Smuzhiyun }
673*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_set_bw);
674*4882a593Smuzhiyun
__icc_enable(struct icc_path * path,bool enable)675*4882a593Smuzhiyun static int __icc_enable(struct icc_path *path, bool enable)
676*4882a593Smuzhiyun {
677*4882a593Smuzhiyun int i;
678*4882a593Smuzhiyun
679*4882a593Smuzhiyun if (!path)
680*4882a593Smuzhiyun return 0;
681*4882a593Smuzhiyun
682*4882a593Smuzhiyun if (WARN_ON(IS_ERR(path) || !path->num_nodes))
683*4882a593Smuzhiyun return -EINVAL;
684*4882a593Smuzhiyun
685*4882a593Smuzhiyun mutex_lock(&icc_lock);
686*4882a593Smuzhiyun
687*4882a593Smuzhiyun for (i = 0; i < path->num_nodes; i++)
688*4882a593Smuzhiyun path->reqs[i].enabled = enable;
689*4882a593Smuzhiyun
690*4882a593Smuzhiyun mutex_unlock(&icc_lock);
691*4882a593Smuzhiyun
692*4882a593Smuzhiyun return icc_set_bw(path, path->reqs[0].avg_bw,
693*4882a593Smuzhiyun path->reqs[0].peak_bw);
694*4882a593Smuzhiyun }
695*4882a593Smuzhiyun
icc_enable(struct icc_path * path)696*4882a593Smuzhiyun int icc_enable(struct icc_path *path)
697*4882a593Smuzhiyun {
698*4882a593Smuzhiyun return __icc_enable(path, true);
699*4882a593Smuzhiyun }
700*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_enable);
701*4882a593Smuzhiyun
icc_disable(struct icc_path * path)702*4882a593Smuzhiyun int icc_disable(struct icc_path *path)
703*4882a593Smuzhiyun {
704*4882a593Smuzhiyun return __icc_enable(path, false);
705*4882a593Smuzhiyun }
706*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_disable);
707*4882a593Smuzhiyun
708*4882a593Smuzhiyun /**
709*4882a593Smuzhiyun * icc_get() - return a handle for path between two endpoints
710*4882a593Smuzhiyun * @dev: the device requesting the path
711*4882a593Smuzhiyun * @src_id: source device port id
712*4882a593Smuzhiyun * @dst_id: destination device port id
713*4882a593Smuzhiyun *
714*4882a593Smuzhiyun * This function will search for a path between two endpoints and return an
715*4882a593Smuzhiyun * icc_path handle on success. Use icc_put() to release
716*4882a593Smuzhiyun * constraints when they are not needed anymore.
717*4882a593Smuzhiyun * If the interconnect API is disabled, NULL is returned and the consumer
718*4882a593Smuzhiyun * drivers will still build. Drivers are free to handle this specifically,
719*4882a593Smuzhiyun * but they don't have to.
720*4882a593Smuzhiyun *
721*4882a593Smuzhiyun * Return: icc_path pointer on success, ERR_PTR() on error or NULL if the
722*4882a593Smuzhiyun * interconnect API is disabled.
723*4882a593Smuzhiyun */
icc_get(struct device * dev,const int src_id,const int dst_id)724*4882a593Smuzhiyun struct icc_path *icc_get(struct device *dev, const int src_id, const int dst_id)
725*4882a593Smuzhiyun {
726*4882a593Smuzhiyun struct icc_node *src, *dst;
727*4882a593Smuzhiyun struct icc_path *path = ERR_PTR(-EPROBE_DEFER);
728*4882a593Smuzhiyun
729*4882a593Smuzhiyun mutex_lock(&icc_lock);
730*4882a593Smuzhiyun
731*4882a593Smuzhiyun src = node_find(src_id);
732*4882a593Smuzhiyun if (!src)
733*4882a593Smuzhiyun goto out;
734*4882a593Smuzhiyun
735*4882a593Smuzhiyun dst = node_find(dst_id);
736*4882a593Smuzhiyun if (!dst)
737*4882a593Smuzhiyun goto out;
738*4882a593Smuzhiyun
739*4882a593Smuzhiyun path = path_find(dev, src, dst);
740*4882a593Smuzhiyun if (IS_ERR(path)) {
741*4882a593Smuzhiyun dev_err(dev, "%s: invalid path=%ld\n", __func__, PTR_ERR(path));
742*4882a593Smuzhiyun goto out;
743*4882a593Smuzhiyun }
744*4882a593Smuzhiyun
745*4882a593Smuzhiyun path->name = kasprintf(GFP_KERNEL, "%s-%s", src->name, dst->name);
746*4882a593Smuzhiyun if (!path->name) {
747*4882a593Smuzhiyun kfree(path);
748*4882a593Smuzhiyun path = ERR_PTR(-ENOMEM);
749*4882a593Smuzhiyun }
750*4882a593Smuzhiyun out:
751*4882a593Smuzhiyun mutex_unlock(&icc_lock);
752*4882a593Smuzhiyun return path;
753*4882a593Smuzhiyun }
754*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_get);
755*4882a593Smuzhiyun
756*4882a593Smuzhiyun /**
757*4882a593Smuzhiyun * icc_put() - release the reference to the icc_path
758*4882a593Smuzhiyun * @path: interconnect path
759*4882a593Smuzhiyun *
760*4882a593Smuzhiyun * Use this function to release the constraints on a path when the path is
761*4882a593Smuzhiyun * no longer needed. The constraints will be re-aggregated.
762*4882a593Smuzhiyun */
icc_put(struct icc_path * path)763*4882a593Smuzhiyun void icc_put(struct icc_path *path)
764*4882a593Smuzhiyun {
765*4882a593Smuzhiyun struct icc_node *node;
766*4882a593Smuzhiyun size_t i;
767*4882a593Smuzhiyun int ret;
768*4882a593Smuzhiyun
769*4882a593Smuzhiyun if (!path || WARN_ON(IS_ERR(path)))
770*4882a593Smuzhiyun return;
771*4882a593Smuzhiyun
772*4882a593Smuzhiyun ret = icc_set_bw(path, 0, 0);
773*4882a593Smuzhiyun if (ret)
774*4882a593Smuzhiyun pr_err("%s: error (%d)\n", __func__, ret);
775*4882a593Smuzhiyun
776*4882a593Smuzhiyun mutex_lock(&icc_lock);
777*4882a593Smuzhiyun for (i = 0; i < path->num_nodes; i++) {
778*4882a593Smuzhiyun node = path->reqs[i].node;
779*4882a593Smuzhiyun hlist_del(&path->reqs[i].req_node);
780*4882a593Smuzhiyun if (!WARN_ON(!node->provider->users))
781*4882a593Smuzhiyun node->provider->users--;
782*4882a593Smuzhiyun }
783*4882a593Smuzhiyun mutex_unlock(&icc_lock);
784*4882a593Smuzhiyun
785*4882a593Smuzhiyun kfree_const(path->name);
786*4882a593Smuzhiyun kfree(path);
787*4882a593Smuzhiyun }
788*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_put);
789*4882a593Smuzhiyun
icc_node_create_nolock(int id)790*4882a593Smuzhiyun static struct icc_node *icc_node_create_nolock(int id)
791*4882a593Smuzhiyun {
792*4882a593Smuzhiyun struct icc_node *node;
793*4882a593Smuzhiyun
794*4882a593Smuzhiyun /* check if node already exists */
795*4882a593Smuzhiyun node = node_find(id);
796*4882a593Smuzhiyun if (node)
797*4882a593Smuzhiyun return node;
798*4882a593Smuzhiyun
799*4882a593Smuzhiyun node = kzalloc(sizeof(*node), GFP_KERNEL);
800*4882a593Smuzhiyun if (!node)
801*4882a593Smuzhiyun return ERR_PTR(-ENOMEM);
802*4882a593Smuzhiyun
803*4882a593Smuzhiyun id = idr_alloc(&icc_idr, node, id, id + 1, GFP_KERNEL);
804*4882a593Smuzhiyun if (id < 0) {
805*4882a593Smuzhiyun WARN(1, "%s: couldn't get idr\n", __func__);
806*4882a593Smuzhiyun kfree(node);
807*4882a593Smuzhiyun return ERR_PTR(id);
808*4882a593Smuzhiyun }
809*4882a593Smuzhiyun
810*4882a593Smuzhiyun node->id = id;
811*4882a593Smuzhiyun
812*4882a593Smuzhiyun return node;
813*4882a593Smuzhiyun }
814*4882a593Smuzhiyun
815*4882a593Smuzhiyun /**
816*4882a593Smuzhiyun * icc_node_create() - create a node
817*4882a593Smuzhiyun * @id: node id
818*4882a593Smuzhiyun *
819*4882a593Smuzhiyun * Return: icc_node pointer on success, or ERR_PTR() on error
820*4882a593Smuzhiyun */
icc_node_create(int id)821*4882a593Smuzhiyun struct icc_node *icc_node_create(int id)
822*4882a593Smuzhiyun {
823*4882a593Smuzhiyun struct icc_node *node;
824*4882a593Smuzhiyun
825*4882a593Smuzhiyun mutex_lock(&icc_lock);
826*4882a593Smuzhiyun
827*4882a593Smuzhiyun node = icc_node_create_nolock(id);
828*4882a593Smuzhiyun
829*4882a593Smuzhiyun mutex_unlock(&icc_lock);
830*4882a593Smuzhiyun
831*4882a593Smuzhiyun return node;
832*4882a593Smuzhiyun }
833*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_node_create);
834*4882a593Smuzhiyun
835*4882a593Smuzhiyun /**
836*4882a593Smuzhiyun * icc_node_destroy() - destroy a node
837*4882a593Smuzhiyun * @id: node id
838*4882a593Smuzhiyun */
icc_node_destroy(int id)839*4882a593Smuzhiyun void icc_node_destroy(int id)
840*4882a593Smuzhiyun {
841*4882a593Smuzhiyun struct icc_node *node;
842*4882a593Smuzhiyun
843*4882a593Smuzhiyun mutex_lock(&icc_lock);
844*4882a593Smuzhiyun
845*4882a593Smuzhiyun node = node_find(id);
846*4882a593Smuzhiyun if (node) {
847*4882a593Smuzhiyun idr_remove(&icc_idr, node->id);
848*4882a593Smuzhiyun WARN_ON(!hlist_empty(&node->req_list));
849*4882a593Smuzhiyun }
850*4882a593Smuzhiyun
851*4882a593Smuzhiyun mutex_unlock(&icc_lock);
852*4882a593Smuzhiyun
853*4882a593Smuzhiyun kfree(node);
854*4882a593Smuzhiyun }
855*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_node_destroy);
856*4882a593Smuzhiyun
857*4882a593Smuzhiyun /**
858*4882a593Smuzhiyun * icc_link_create() - create a link between two nodes
859*4882a593Smuzhiyun * @node: source node id
860*4882a593Smuzhiyun * @dst_id: destination node id
861*4882a593Smuzhiyun *
862*4882a593Smuzhiyun * Create a link between two nodes. The nodes might belong to different
863*4882a593Smuzhiyun * interconnect providers and the @dst_id node might not exist (if the
864*4882a593Smuzhiyun * provider driver has not probed yet). So just create the @dst_id node
865*4882a593Smuzhiyun * and when the actual provider driver is probed, the rest of the node
866*4882a593Smuzhiyun * data is filled.
867*4882a593Smuzhiyun *
868*4882a593Smuzhiyun * Return: 0 on success, or an error code otherwise
869*4882a593Smuzhiyun */
icc_link_create(struct icc_node * node,const int dst_id)870*4882a593Smuzhiyun int icc_link_create(struct icc_node *node, const int dst_id)
871*4882a593Smuzhiyun {
872*4882a593Smuzhiyun struct icc_node *dst;
873*4882a593Smuzhiyun struct icc_node **new;
874*4882a593Smuzhiyun int ret = 0;
875*4882a593Smuzhiyun
876*4882a593Smuzhiyun if (!node->provider)
877*4882a593Smuzhiyun return -EINVAL;
878*4882a593Smuzhiyun
879*4882a593Smuzhiyun mutex_lock(&icc_lock);
880*4882a593Smuzhiyun
881*4882a593Smuzhiyun dst = node_find(dst_id);
882*4882a593Smuzhiyun if (!dst) {
883*4882a593Smuzhiyun dst = icc_node_create_nolock(dst_id);
884*4882a593Smuzhiyun
885*4882a593Smuzhiyun if (IS_ERR(dst)) {
886*4882a593Smuzhiyun ret = PTR_ERR(dst);
887*4882a593Smuzhiyun goto out;
888*4882a593Smuzhiyun }
889*4882a593Smuzhiyun }
890*4882a593Smuzhiyun
891*4882a593Smuzhiyun new = krealloc(node->links,
892*4882a593Smuzhiyun (node->num_links + 1) * sizeof(*node->links),
893*4882a593Smuzhiyun GFP_KERNEL);
894*4882a593Smuzhiyun if (!new) {
895*4882a593Smuzhiyun ret = -ENOMEM;
896*4882a593Smuzhiyun goto out;
897*4882a593Smuzhiyun }
898*4882a593Smuzhiyun
899*4882a593Smuzhiyun node->links = new;
900*4882a593Smuzhiyun node->links[node->num_links++] = dst;
901*4882a593Smuzhiyun
902*4882a593Smuzhiyun out:
903*4882a593Smuzhiyun mutex_unlock(&icc_lock);
904*4882a593Smuzhiyun
905*4882a593Smuzhiyun return ret;
906*4882a593Smuzhiyun }
907*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_link_create);
908*4882a593Smuzhiyun
909*4882a593Smuzhiyun /**
910*4882a593Smuzhiyun * icc_link_destroy() - destroy a link between two nodes
911*4882a593Smuzhiyun * @src: pointer to source node
912*4882a593Smuzhiyun * @dst: pointer to destination node
913*4882a593Smuzhiyun *
914*4882a593Smuzhiyun * Return: 0 on success, or an error code otherwise
915*4882a593Smuzhiyun */
icc_link_destroy(struct icc_node * src,struct icc_node * dst)916*4882a593Smuzhiyun int icc_link_destroy(struct icc_node *src, struct icc_node *dst)
917*4882a593Smuzhiyun {
918*4882a593Smuzhiyun struct icc_node **new;
919*4882a593Smuzhiyun size_t slot;
920*4882a593Smuzhiyun int ret = 0;
921*4882a593Smuzhiyun
922*4882a593Smuzhiyun if (IS_ERR_OR_NULL(src))
923*4882a593Smuzhiyun return -EINVAL;
924*4882a593Smuzhiyun
925*4882a593Smuzhiyun if (IS_ERR_OR_NULL(dst))
926*4882a593Smuzhiyun return -EINVAL;
927*4882a593Smuzhiyun
928*4882a593Smuzhiyun mutex_lock(&icc_lock);
929*4882a593Smuzhiyun
930*4882a593Smuzhiyun for (slot = 0; slot < src->num_links; slot++)
931*4882a593Smuzhiyun if (src->links[slot] == dst)
932*4882a593Smuzhiyun break;
933*4882a593Smuzhiyun
934*4882a593Smuzhiyun if (WARN_ON(slot == src->num_links)) {
935*4882a593Smuzhiyun ret = -ENXIO;
936*4882a593Smuzhiyun goto out;
937*4882a593Smuzhiyun }
938*4882a593Smuzhiyun
939*4882a593Smuzhiyun src->links[slot] = src->links[--src->num_links];
940*4882a593Smuzhiyun
941*4882a593Smuzhiyun new = krealloc(src->links, src->num_links * sizeof(*src->links),
942*4882a593Smuzhiyun GFP_KERNEL);
943*4882a593Smuzhiyun if (new)
944*4882a593Smuzhiyun src->links = new;
945*4882a593Smuzhiyun else
946*4882a593Smuzhiyun ret = -ENOMEM;
947*4882a593Smuzhiyun
948*4882a593Smuzhiyun out:
949*4882a593Smuzhiyun mutex_unlock(&icc_lock);
950*4882a593Smuzhiyun
951*4882a593Smuzhiyun return ret;
952*4882a593Smuzhiyun }
953*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_link_destroy);
954*4882a593Smuzhiyun
955*4882a593Smuzhiyun /**
956*4882a593Smuzhiyun * icc_node_add() - add interconnect node to interconnect provider
957*4882a593Smuzhiyun * @node: pointer to the interconnect node
958*4882a593Smuzhiyun * @provider: pointer to the interconnect provider
959*4882a593Smuzhiyun */
icc_node_add(struct icc_node * node,struct icc_provider * provider)960*4882a593Smuzhiyun void icc_node_add(struct icc_node *node, struct icc_provider *provider)
961*4882a593Smuzhiyun {
962*4882a593Smuzhiyun mutex_lock(&icc_lock);
963*4882a593Smuzhiyun
964*4882a593Smuzhiyun node->provider = provider;
965*4882a593Smuzhiyun list_add_tail(&node->node_list, &provider->nodes);
966*4882a593Smuzhiyun
967*4882a593Smuzhiyun /* get the initial bandwidth values and sync them with hardware */
968*4882a593Smuzhiyun if (provider->get_bw) {
969*4882a593Smuzhiyun provider->get_bw(node, &node->init_avg, &node->init_peak);
970*4882a593Smuzhiyun } else {
971*4882a593Smuzhiyun node->init_avg = INT_MAX;
972*4882a593Smuzhiyun node->init_peak = INT_MAX;
973*4882a593Smuzhiyun }
974*4882a593Smuzhiyun node->avg_bw = node->init_avg;
975*4882a593Smuzhiyun node->peak_bw = node->init_peak;
976*4882a593Smuzhiyun
977*4882a593Smuzhiyun if (provider->pre_aggregate)
978*4882a593Smuzhiyun provider->pre_aggregate(node);
979*4882a593Smuzhiyun
980*4882a593Smuzhiyun if (provider->aggregate)
981*4882a593Smuzhiyun provider->aggregate(node, 0, node->init_avg, node->init_peak,
982*4882a593Smuzhiyun &node->avg_bw, &node->peak_bw);
983*4882a593Smuzhiyun
984*4882a593Smuzhiyun provider->set(node, node);
985*4882a593Smuzhiyun node->avg_bw = 0;
986*4882a593Smuzhiyun node->peak_bw = 0;
987*4882a593Smuzhiyun
988*4882a593Smuzhiyun mutex_unlock(&icc_lock);
989*4882a593Smuzhiyun }
990*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_node_add);
991*4882a593Smuzhiyun
992*4882a593Smuzhiyun /**
993*4882a593Smuzhiyun * icc_node_del() - delete interconnect node from interconnect provider
994*4882a593Smuzhiyun * @node: pointer to the interconnect node
995*4882a593Smuzhiyun */
icc_node_del(struct icc_node * node)996*4882a593Smuzhiyun void icc_node_del(struct icc_node *node)
997*4882a593Smuzhiyun {
998*4882a593Smuzhiyun mutex_lock(&icc_lock);
999*4882a593Smuzhiyun
1000*4882a593Smuzhiyun list_del(&node->node_list);
1001*4882a593Smuzhiyun
1002*4882a593Smuzhiyun mutex_unlock(&icc_lock);
1003*4882a593Smuzhiyun }
1004*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_node_del);
1005*4882a593Smuzhiyun
1006*4882a593Smuzhiyun /**
1007*4882a593Smuzhiyun * icc_nodes_remove() - remove all previously added nodes from provider
1008*4882a593Smuzhiyun * @provider: the interconnect provider we are removing nodes from
1009*4882a593Smuzhiyun *
1010*4882a593Smuzhiyun * Return: 0 on success, or an error code otherwise
1011*4882a593Smuzhiyun */
icc_nodes_remove(struct icc_provider * provider)1012*4882a593Smuzhiyun int icc_nodes_remove(struct icc_provider *provider)
1013*4882a593Smuzhiyun {
1014*4882a593Smuzhiyun struct icc_node *n, *tmp;
1015*4882a593Smuzhiyun
1016*4882a593Smuzhiyun if (WARN_ON(IS_ERR_OR_NULL(provider)))
1017*4882a593Smuzhiyun return -EINVAL;
1018*4882a593Smuzhiyun
1019*4882a593Smuzhiyun list_for_each_entry_safe_reverse(n, tmp, &provider->nodes, node_list) {
1020*4882a593Smuzhiyun icc_node_del(n);
1021*4882a593Smuzhiyun icc_node_destroy(n->id);
1022*4882a593Smuzhiyun }
1023*4882a593Smuzhiyun
1024*4882a593Smuzhiyun return 0;
1025*4882a593Smuzhiyun }
1026*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_nodes_remove);
1027*4882a593Smuzhiyun
1028*4882a593Smuzhiyun /**
1029*4882a593Smuzhiyun * icc_provider_add() - add a new interconnect provider
1030*4882a593Smuzhiyun * @provider: the interconnect provider that will be added into topology
1031*4882a593Smuzhiyun *
1032*4882a593Smuzhiyun * Return: 0 on success, or an error code otherwise
1033*4882a593Smuzhiyun */
icc_provider_add(struct icc_provider * provider)1034*4882a593Smuzhiyun int icc_provider_add(struct icc_provider *provider)
1035*4882a593Smuzhiyun {
1036*4882a593Smuzhiyun if (WARN_ON(!provider->set))
1037*4882a593Smuzhiyun return -EINVAL;
1038*4882a593Smuzhiyun if (WARN_ON(!provider->xlate && !provider->xlate_extended))
1039*4882a593Smuzhiyun return -EINVAL;
1040*4882a593Smuzhiyun
1041*4882a593Smuzhiyun mutex_lock(&icc_lock);
1042*4882a593Smuzhiyun
1043*4882a593Smuzhiyun INIT_LIST_HEAD(&provider->nodes);
1044*4882a593Smuzhiyun list_add_tail(&provider->provider_list, &icc_providers);
1045*4882a593Smuzhiyun
1046*4882a593Smuzhiyun mutex_unlock(&icc_lock);
1047*4882a593Smuzhiyun
1048*4882a593Smuzhiyun dev_dbg(provider->dev, "interconnect provider added to topology\n");
1049*4882a593Smuzhiyun
1050*4882a593Smuzhiyun return 0;
1051*4882a593Smuzhiyun }
1052*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_provider_add);
1053*4882a593Smuzhiyun
1054*4882a593Smuzhiyun /**
1055*4882a593Smuzhiyun * icc_provider_del() - delete previously added interconnect provider
1056*4882a593Smuzhiyun * @provider: the interconnect provider that will be removed from topology
1057*4882a593Smuzhiyun *
1058*4882a593Smuzhiyun * Return: 0 on success, or an error code otherwise
1059*4882a593Smuzhiyun */
icc_provider_del(struct icc_provider * provider)1060*4882a593Smuzhiyun int icc_provider_del(struct icc_provider *provider)
1061*4882a593Smuzhiyun {
1062*4882a593Smuzhiyun mutex_lock(&icc_lock);
1063*4882a593Smuzhiyun if (provider->users) {
1064*4882a593Smuzhiyun pr_warn("interconnect provider still has %d users\n",
1065*4882a593Smuzhiyun provider->users);
1066*4882a593Smuzhiyun mutex_unlock(&icc_lock);
1067*4882a593Smuzhiyun return -EBUSY;
1068*4882a593Smuzhiyun }
1069*4882a593Smuzhiyun
1070*4882a593Smuzhiyun if (!list_empty(&provider->nodes)) {
1071*4882a593Smuzhiyun pr_warn("interconnect provider still has nodes\n");
1072*4882a593Smuzhiyun mutex_unlock(&icc_lock);
1073*4882a593Smuzhiyun return -EBUSY;
1074*4882a593Smuzhiyun }
1075*4882a593Smuzhiyun
1076*4882a593Smuzhiyun list_del(&provider->provider_list);
1077*4882a593Smuzhiyun mutex_unlock(&icc_lock);
1078*4882a593Smuzhiyun
1079*4882a593Smuzhiyun return 0;
1080*4882a593Smuzhiyun }
1081*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_provider_del);
1082*4882a593Smuzhiyun
of_count_icc_providers(struct device_node * np)1083*4882a593Smuzhiyun static int of_count_icc_providers(struct device_node *np)
1084*4882a593Smuzhiyun {
1085*4882a593Smuzhiyun struct device_node *child;
1086*4882a593Smuzhiyun int count = 0;
1087*4882a593Smuzhiyun const struct of_device_id __maybe_unused ignore_list[] = {
1088*4882a593Smuzhiyun { .compatible = "qcom,sc7180-ipa-virt" },
1089*4882a593Smuzhiyun {}
1090*4882a593Smuzhiyun };
1091*4882a593Smuzhiyun
1092*4882a593Smuzhiyun for_each_available_child_of_node(np, child) {
1093*4882a593Smuzhiyun if (of_property_read_bool(child, "#interconnect-cells") &&
1094*4882a593Smuzhiyun likely(!of_match_node(ignore_list, child)))
1095*4882a593Smuzhiyun count++;
1096*4882a593Smuzhiyun count += of_count_icc_providers(child);
1097*4882a593Smuzhiyun }
1098*4882a593Smuzhiyun
1099*4882a593Smuzhiyun return count;
1100*4882a593Smuzhiyun }
1101*4882a593Smuzhiyun
icc_sync_state(struct device * dev)1102*4882a593Smuzhiyun void icc_sync_state(struct device *dev)
1103*4882a593Smuzhiyun {
1104*4882a593Smuzhiyun struct icc_provider *p;
1105*4882a593Smuzhiyun struct icc_node *n;
1106*4882a593Smuzhiyun static int count;
1107*4882a593Smuzhiyun
1108*4882a593Smuzhiyun count++;
1109*4882a593Smuzhiyun
1110*4882a593Smuzhiyun if (count < providers_count)
1111*4882a593Smuzhiyun return;
1112*4882a593Smuzhiyun
1113*4882a593Smuzhiyun mutex_lock(&icc_lock);
1114*4882a593Smuzhiyun synced_state = true;
1115*4882a593Smuzhiyun list_for_each_entry(p, &icc_providers, provider_list) {
1116*4882a593Smuzhiyun dev_dbg(p->dev, "interconnect provider is in synced state\n");
1117*4882a593Smuzhiyun list_for_each_entry(n, &p->nodes, node_list) {
1118*4882a593Smuzhiyun if (n->init_avg || n->init_peak) {
1119*4882a593Smuzhiyun n->init_avg = 0;
1120*4882a593Smuzhiyun n->init_peak = 0;
1121*4882a593Smuzhiyun aggregate_requests(n);
1122*4882a593Smuzhiyun p->set(n, n);
1123*4882a593Smuzhiyun }
1124*4882a593Smuzhiyun }
1125*4882a593Smuzhiyun }
1126*4882a593Smuzhiyun mutex_unlock(&icc_lock);
1127*4882a593Smuzhiyun }
1128*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(icc_sync_state);
1129*4882a593Smuzhiyun
icc_init(void)1130*4882a593Smuzhiyun static int __init icc_init(void)
1131*4882a593Smuzhiyun {
1132*4882a593Smuzhiyun struct device_node *root = of_find_node_by_path("/");
1133*4882a593Smuzhiyun
1134*4882a593Smuzhiyun providers_count = of_count_icc_providers(root);
1135*4882a593Smuzhiyun of_node_put(root);
1136*4882a593Smuzhiyun
1137*4882a593Smuzhiyun icc_debugfs_dir = debugfs_create_dir("interconnect", NULL);
1138*4882a593Smuzhiyun debugfs_create_file("interconnect_summary", 0444,
1139*4882a593Smuzhiyun icc_debugfs_dir, NULL, &icc_summary_fops);
1140*4882a593Smuzhiyun debugfs_create_file("interconnect_graph", 0444,
1141*4882a593Smuzhiyun icc_debugfs_dir, NULL, &icc_graph_fops);
1142*4882a593Smuzhiyun return 0;
1143*4882a593Smuzhiyun }
1144*4882a593Smuzhiyun
1145*4882a593Smuzhiyun device_initcall(icc_init);
1146*4882a593Smuzhiyun
1147*4882a593Smuzhiyun MODULE_AUTHOR("Georgi Djakov <georgi.djakov@linaro.org>");
1148*4882a593Smuzhiyun MODULE_DESCRIPTION("Interconnect Driver Core");
1149*4882a593Smuzhiyun MODULE_LICENSE("GPL v2");
1150