xref: /OK3568_Linux_fs/u-boot/drivers/clk/clk-uclass.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun  * Copyright (C) 2015 Google, Inc
3*4882a593Smuzhiyun  * Written by Simon Glass <sjg@chromium.org>
4*4882a593Smuzhiyun  * Copyright (c) 2016, NVIDIA CORPORATION.
5*4882a593Smuzhiyun  * Copyright (c) 2018, Theobroma Systems Design und Consulting GmbH
6*4882a593Smuzhiyun  *
7*4882a593Smuzhiyun  * SPDX-License-Identifier:	GPL-2.0+
8*4882a593Smuzhiyun  */
9*4882a593Smuzhiyun 
10*4882a593Smuzhiyun #include <common.h>
11*4882a593Smuzhiyun #include <clk.h>
12*4882a593Smuzhiyun #include <clk-uclass.h>
13*4882a593Smuzhiyun #include <dm.h>
14*4882a593Smuzhiyun #include <dm/device-internal.h>
15*4882a593Smuzhiyun #include <dm/read.h>
16*4882a593Smuzhiyun #include <dt-structs.h>
17*4882a593Smuzhiyun #include <errno.h>
18*4882a593Smuzhiyun 
clk_dev_ops(struct udevice * dev)19*4882a593Smuzhiyun static inline const struct clk_ops *clk_dev_ops(struct udevice *dev)
20*4882a593Smuzhiyun {
21*4882a593Smuzhiyun 	return (const struct clk_ops *)dev->driver->ops;
22*4882a593Smuzhiyun }
23*4882a593Smuzhiyun 
24*4882a593Smuzhiyun #if CONFIG_IS_ENABLED(OF_CONTROL)
25*4882a593Smuzhiyun # if CONFIG_IS_ENABLED(OF_PLATDATA)
clk_get_by_index_platdata(struct udevice * dev,int index,struct phandle_1_arg * cells,struct clk * clk)26*4882a593Smuzhiyun int clk_get_by_index_platdata(struct udevice *dev, int index,
27*4882a593Smuzhiyun 			      struct phandle_1_arg *cells, struct clk *clk)
28*4882a593Smuzhiyun {
29*4882a593Smuzhiyun 	int ret;
30*4882a593Smuzhiyun 
31*4882a593Smuzhiyun 	if (index != 0)
32*4882a593Smuzhiyun 		return -ENOSYS;
33*4882a593Smuzhiyun 	ret = uclass_get_device(UCLASS_CLK, 0, &clk->dev);
34*4882a593Smuzhiyun 	if (ret)
35*4882a593Smuzhiyun 		return ret;
36*4882a593Smuzhiyun 	clk->id = cells[0].arg[0];
37*4882a593Smuzhiyun 
38*4882a593Smuzhiyun 	return 0;
39*4882a593Smuzhiyun }
40*4882a593Smuzhiyun # else
clk_of_xlate_default(struct clk * clk,struct ofnode_phandle_args * args)41*4882a593Smuzhiyun static int clk_of_xlate_default(struct clk *clk,
42*4882a593Smuzhiyun 				struct ofnode_phandle_args *args)
43*4882a593Smuzhiyun {
44*4882a593Smuzhiyun 	debug("%s(clk=%p)\n", __func__, clk);
45*4882a593Smuzhiyun 
46*4882a593Smuzhiyun 	if (args->args_count > 1) {
47*4882a593Smuzhiyun 		debug("Invaild args_count: %d\n", args->args_count);
48*4882a593Smuzhiyun 		return -EINVAL;
49*4882a593Smuzhiyun 	}
50*4882a593Smuzhiyun 
51*4882a593Smuzhiyun 	if (args->args_count)
52*4882a593Smuzhiyun 		clk->id = args->args[0];
53*4882a593Smuzhiyun 	else
54*4882a593Smuzhiyun 		clk->id = 0;
55*4882a593Smuzhiyun 
56*4882a593Smuzhiyun 	return 0;
57*4882a593Smuzhiyun }
58*4882a593Smuzhiyun 
clk_get_by_indexed_prop(struct udevice * dev,const char * prop_name,int index,struct clk * clk)59*4882a593Smuzhiyun static int clk_get_by_indexed_prop(struct udevice *dev, const char *prop_name,
60*4882a593Smuzhiyun 				   int index, struct clk *clk)
61*4882a593Smuzhiyun {
62*4882a593Smuzhiyun 	int ret;
63*4882a593Smuzhiyun 	struct ofnode_phandle_args args;
64*4882a593Smuzhiyun 	struct udevice *dev_clk;
65*4882a593Smuzhiyun 	const struct clk_ops *ops;
66*4882a593Smuzhiyun 
67*4882a593Smuzhiyun 	debug("%s(dev=%p, index=%d, clk=%p)\n", __func__, dev, index, clk);
68*4882a593Smuzhiyun 
69*4882a593Smuzhiyun 	assert(clk);
70*4882a593Smuzhiyun 	clk->dev = NULL;
71*4882a593Smuzhiyun 
72*4882a593Smuzhiyun 	ret = dev_read_phandle_with_args(dev, prop_name, "#clock-cells", 0,
73*4882a593Smuzhiyun 					 index, &args);
74*4882a593Smuzhiyun 	if (ret) {
75*4882a593Smuzhiyun 		debug("%s: fdtdec_parse_phandle_with_args failed: err=%d\n",
76*4882a593Smuzhiyun 		      __func__, ret);
77*4882a593Smuzhiyun 		return ret;
78*4882a593Smuzhiyun 	}
79*4882a593Smuzhiyun 
80*4882a593Smuzhiyun 	ret = uclass_get_device_by_ofnode(UCLASS_CLK, args.node, &dev_clk);
81*4882a593Smuzhiyun 	if (ret) {
82*4882a593Smuzhiyun 		debug("%s: uclass_get_device_by_of_offset failed: err=%d\n",
83*4882a593Smuzhiyun 		      __func__, ret);
84*4882a593Smuzhiyun 		return ret;
85*4882a593Smuzhiyun 	}
86*4882a593Smuzhiyun 
87*4882a593Smuzhiyun 	clk->dev = dev_clk;
88*4882a593Smuzhiyun 
89*4882a593Smuzhiyun 	ops = clk_dev_ops(dev_clk);
90*4882a593Smuzhiyun 
91*4882a593Smuzhiyun 	if (ops->of_xlate)
92*4882a593Smuzhiyun 		ret = ops->of_xlate(clk, &args);
93*4882a593Smuzhiyun 	else
94*4882a593Smuzhiyun 		ret = clk_of_xlate_default(clk, &args);
95*4882a593Smuzhiyun 	if (ret) {
96*4882a593Smuzhiyun 		debug("of_xlate() failed: %d\n", ret);
97*4882a593Smuzhiyun 		return ret;
98*4882a593Smuzhiyun 	}
99*4882a593Smuzhiyun 
100*4882a593Smuzhiyun 	return clk_request(dev_clk, clk);
101*4882a593Smuzhiyun }
102*4882a593Smuzhiyun 
clk_get_by_index(struct udevice * dev,int index,struct clk * clk)103*4882a593Smuzhiyun int clk_get_by_index(struct udevice *dev, int index, struct clk *clk)
104*4882a593Smuzhiyun {
105*4882a593Smuzhiyun 	return clk_get_by_indexed_prop(dev, "clocks", index, clk);
106*4882a593Smuzhiyun }
107*4882a593Smuzhiyun 
clk_get_bulk(struct udevice * dev,struct clk_bulk * bulk)108*4882a593Smuzhiyun int clk_get_bulk(struct udevice *dev, struct clk_bulk *bulk)
109*4882a593Smuzhiyun {
110*4882a593Smuzhiyun 	int i, ret, err, count;
111*4882a593Smuzhiyun 
112*4882a593Smuzhiyun 	bulk->count = 0;
113*4882a593Smuzhiyun 
114*4882a593Smuzhiyun 	count = dev_count_phandle_with_args(dev, "clocks", "#clock-cells");
115*4882a593Smuzhiyun 	if (!count)
116*4882a593Smuzhiyun 		return 0;
117*4882a593Smuzhiyun 
118*4882a593Smuzhiyun 	bulk->clks = devm_kcalloc(dev, count, sizeof(struct clk), GFP_KERNEL);
119*4882a593Smuzhiyun 	if (!bulk->clks)
120*4882a593Smuzhiyun 		return -ENOMEM;
121*4882a593Smuzhiyun 
122*4882a593Smuzhiyun 	for (i = 0; i < count; i++) {
123*4882a593Smuzhiyun 		ret = clk_get_by_index(dev, i, &bulk->clks[i]);
124*4882a593Smuzhiyun 		if (ret < 0)
125*4882a593Smuzhiyun 			goto bulk_get_err;
126*4882a593Smuzhiyun 
127*4882a593Smuzhiyun 		++bulk->count;
128*4882a593Smuzhiyun 	}
129*4882a593Smuzhiyun 
130*4882a593Smuzhiyun 	return 0;
131*4882a593Smuzhiyun 
132*4882a593Smuzhiyun bulk_get_err:
133*4882a593Smuzhiyun 	err = clk_release_all(bulk->clks, bulk->count);
134*4882a593Smuzhiyun 	if (err)
135*4882a593Smuzhiyun 		debug("%s: could release all clocks for %p\n",
136*4882a593Smuzhiyun 		      __func__, dev);
137*4882a593Smuzhiyun 
138*4882a593Smuzhiyun 	return ret;
139*4882a593Smuzhiyun }
140*4882a593Smuzhiyun 
clk_set_default_parents(struct udevice * dev)141*4882a593Smuzhiyun static int clk_set_default_parents(struct udevice *dev)
142*4882a593Smuzhiyun {
143*4882a593Smuzhiyun 	struct clk clk, parent_clk;
144*4882a593Smuzhiyun 	int index;
145*4882a593Smuzhiyun 	int num_parents;
146*4882a593Smuzhiyun 	int ret;
147*4882a593Smuzhiyun 
148*4882a593Smuzhiyun 	num_parents = dev_count_phandle_with_args(dev, "assigned-clock-parents",
149*4882a593Smuzhiyun 						  "#clock-cells");
150*4882a593Smuzhiyun 	if (num_parents < 0) {
151*4882a593Smuzhiyun 		debug("%s: could not read assigned-clock-parents for %p\n",
152*4882a593Smuzhiyun 		      __func__, dev);
153*4882a593Smuzhiyun 		return 0;
154*4882a593Smuzhiyun 	}
155*4882a593Smuzhiyun 
156*4882a593Smuzhiyun 	for (index = 0; index < num_parents; index++) {
157*4882a593Smuzhiyun 		ret = clk_get_by_indexed_prop(dev, "assigned-clock-parents",
158*4882a593Smuzhiyun 					      index, &parent_clk);
159*4882a593Smuzhiyun 		if (ret) {
160*4882a593Smuzhiyun 			debug("%s: could not get parent clock %d for %s\n",
161*4882a593Smuzhiyun 			      __func__, index, dev_read_name(dev));
162*4882a593Smuzhiyun 			return ret;
163*4882a593Smuzhiyun 		}
164*4882a593Smuzhiyun 
165*4882a593Smuzhiyun 		ret = clk_get_by_indexed_prop(dev, "assigned-clocks",
166*4882a593Smuzhiyun 					      index, &clk);
167*4882a593Smuzhiyun 		if (ret) {
168*4882a593Smuzhiyun 			debug("%s: could not get assigned clock %d for %s\n",
169*4882a593Smuzhiyun 			      __func__, index, dev_read_name(dev));
170*4882a593Smuzhiyun 			return ret;
171*4882a593Smuzhiyun 		}
172*4882a593Smuzhiyun 
173*4882a593Smuzhiyun 		ret = clk_set_parent(&clk, &parent_clk);
174*4882a593Smuzhiyun 
175*4882a593Smuzhiyun 		/*
176*4882a593Smuzhiyun 		 * Not all drivers may support clock-reparenting (as of now).
177*4882a593Smuzhiyun 		 * Ignore errors due to this.
178*4882a593Smuzhiyun 		 */
179*4882a593Smuzhiyun 		if (ret == -ENOSYS)
180*4882a593Smuzhiyun 			continue;
181*4882a593Smuzhiyun 
182*4882a593Smuzhiyun 		if (ret) {
183*4882a593Smuzhiyun 			debug("%s: failed to reparent clock %d for %s\n",
184*4882a593Smuzhiyun 			      __func__, index, dev_read_name(dev));
185*4882a593Smuzhiyun 			return ret;
186*4882a593Smuzhiyun 		}
187*4882a593Smuzhiyun 	}
188*4882a593Smuzhiyun 
189*4882a593Smuzhiyun 	return 0;
190*4882a593Smuzhiyun }
191*4882a593Smuzhiyun 
clk_set_default_rates(struct udevice * dev)192*4882a593Smuzhiyun static int clk_set_default_rates(struct udevice *dev)
193*4882a593Smuzhiyun {
194*4882a593Smuzhiyun 	struct clk clk;
195*4882a593Smuzhiyun 	int index;
196*4882a593Smuzhiyun 	int num_rates;
197*4882a593Smuzhiyun 	int size;
198*4882a593Smuzhiyun 	int ret = 0;
199*4882a593Smuzhiyun 	u32 *rates = NULL;
200*4882a593Smuzhiyun 
201*4882a593Smuzhiyun 	size = dev_read_size(dev, "assigned-clock-rates");
202*4882a593Smuzhiyun 	if (size < 0)
203*4882a593Smuzhiyun 		return 0;
204*4882a593Smuzhiyun 
205*4882a593Smuzhiyun 	num_rates = size / sizeof(u32);
206*4882a593Smuzhiyun 	rates = calloc(num_rates, sizeof(u32));
207*4882a593Smuzhiyun 	if (!rates)
208*4882a593Smuzhiyun 		return -ENOMEM;
209*4882a593Smuzhiyun 
210*4882a593Smuzhiyun 	ret = dev_read_u32_array(dev, "assigned-clock-rates", rates, num_rates);
211*4882a593Smuzhiyun 	if (ret)
212*4882a593Smuzhiyun 		goto fail;
213*4882a593Smuzhiyun 
214*4882a593Smuzhiyun 	for (index = 0; index < num_rates; index++) {
215*4882a593Smuzhiyun 		ret = clk_get_by_indexed_prop(dev, "assigned-clocks",
216*4882a593Smuzhiyun 					      index, &clk);
217*4882a593Smuzhiyun 		if (ret) {
218*4882a593Smuzhiyun 			debug("%s: could not get assigned clock %d for %s\n",
219*4882a593Smuzhiyun 			      __func__, index, dev_read_name(dev));
220*4882a593Smuzhiyun 			continue;
221*4882a593Smuzhiyun 		}
222*4882a593Smuzhiyun 
223*4882a593Smuzhiyun 		ret = clk_set_rate(&clk, rates[index]);
224*4882a593Smuzhiyun 		if (ret < 0) {
225*4882a593Smuzhiyun 			debug("%s: failed to set rate on clock %d for %s\n",
226*4882a593Smuzhiyun 			      __func__, index, dev_read_name(dev));
227*4882a593Smuzhiyun 			continue;
228*4882a593Smuzhiyun 		}
229*4882a593Smuzhiyun 	}
230*4882a593Smuzhiyun 
231*4882a593Smuzhiyun fail:
232*4882a593Smuzhiyun 	free(rates);
233*4882a593Smuzhiyun 	return ret;
234*4882a593Smuzhiyun }
235*4882a593Smuzhiyun 
clk_set_defaults(struct udevice * dev)236*4882a593Smuzhiyun int clk_set_defaults(struct udevice *dev)
237*4882a593Smuzhiyun {
238*4882a593Smuzhiyun 	int ret;
239*4882a593Smuzhiyun 
240*4882a593Smuzhiyun #ifndef CONFIG_SPL_BUILD
241*4882a593Smuzhiyun 	/* If this is running pre-reloc state, don't take any action. */
242*4882a593Smuzhiyun 	if (!(gd->flags & GD_FLG_RELOC))
243*4882a593Smuzhiyun 		return 0;
244*4882a593Smuzhiyun #endif
245*4882a593Smuzhiyun 	debug("%s(%s)\n", __func__, dev_read_name(dev));
246*4882a593Smuzhiyun 
247*4882a593Smuzhiyun 	ret = clk_set_default_parents(dev);
248*4882a593Smuzhiyun 	if (ret)
249*4882a593Smuzhiyun 		return ret;
250*4882a593Smuzhiyun 
251*4882a593Smuzhiyun 	ret = clk_set_default_rates(dev);
252*4882a593Smuzhiyun 	if (ret < 0)
253*4882a593Smuzhiyun 		return ret;
254*4882a593Smuzhiyun 
255*4882a593Smuzhiyun 	return 0;
256*4882a593Smuzhiyun }
257*4882a593Smuzhiyun # endif /* OF_PLATDATA */
258*4882a593Smuzhiyun 
clk_get_by_name(struct udevice * dev,const char * name,struct clk * clk)259*4882a593Smuzhiyun int clk_get_by_name(struct udevice *dev, const char *name, struct clk *clk)
260*4882a593Smuzhiyun {
261*4882a593Smuzhiyun 	int index;
262*4882a593Smuzhiyun 
263*4882a593Smuzhiyun 	debug("%s(dev=%p, name=%s, clk=%p)\n", __func__, dev, name, clk);
264*4882a593Smuzhiyun 	clk->dev = NULL;
265*4882a593Smuzhiyun 
266*4882a593Smuzhiyun 	index = dev_read_stringlist_search(dev, "clock-names", name);
267*4882a593Smuzhiyun 	if (index < 0) {
268*4882a593Smuzhiyun 		debug("fdt_stringlist_search() failed: %d\n", index);
269*4882a593Smuzhiyun 		return index;
270*4882a593Smuzhiyun 	}
271*4882a593Smuzhiyun 
272*4882a593Smuzhiyun 	return clk_get_by_index(dev, index, clk);
273*4882a593Smuzhiyun }
274*4882a593Smuzhiyun 
clk_release_all(struct clk * clk,int count)275*4882a593Smuzhiyun int clk_release_all(struct clk *clk, int count)
276*4882a593Smuzhiyun {
277*4882a593Smuzhiyun 	int i, ret;
278*4882a593Smuzhiyun 
279*4882a593Smuzhiyun 	for (i = 0; i < count; i++) {
280*4882a593Smuzhiyun 		debug("%s(clk[%d]=%p)\n", __func__, i, &clk[i]);
281*4882a593Smuzhiyun 
282*4882a593Smuzhiyun 		/* check if clock has been previously requested */
283*4882a593Smuzhiyun 		if (!clk[i].dev)
284*4882a593Smuzhiyun 			continue;
285*4882a593Smuzhiyun 
286*4882a593Smuzhiyun 		ret = clk_disable(&clk[i]);
287*4882a593Smuzhiyun 		if (ret && ret != -ENOSYS)
288*4882a593Smuzhiyun 			return ret;
289*4882a593Smuzhiyun 
290*4882a593Smuzhiyun 		ret = clk_free(&clk[i]);
291*4882a593Smuzhiyun 		if (ret && ret != -ENOSYS)
292*4882a593Smuzhiyun 			return ret;
293*4882a593Smuzhiyun 	}
294*4882a593Smuzhiyun 
295*4882a593Smuzhiyun 	return 0;
296*4882a593Smuzhiyun }
297*4882a593Smuzhiyun 
298*4882a593Smuzhiyun #endif /* OF_CONTROL */
299*4882a593Smuzhiyun 
clk_request(struct udevice * dev,struct clk * clk)300*4882a593Smuzhiyun int clk_request(struct udevice *dev, struct clk *clk)
301*4882a593Smuzhiyun {
302*4882a593Smuzhiyun 	const struct clk_ops *ops = clk_dev_ops(dev);
303*4882a593Smuzhiyun 
304*4882a593Smuzhiyun 	debug("%s(dev=%p, clk=%p)\n", __func__, dev, clk);
305*4882a593Smuzhiyun 
306*4882a593Smuzhiyun 	clk->dev = dev;
307*4882a593Smuzhiyun 
308*4882a593Smuzhiyun 	if (!ops->request)
309*4882a593Smuzhiyun 		return 0;
310*4882a593Smuzhiyun 
311*4882a593Smuzhiyun 	return ops->request(clk);
312*4882a593Smuzhiyun }
313*4882a593Smuzhiyun 
clk_free(struct clk * clk)314*4882a593Smuzhiyun int clk_free(struct clk *clk)
315*4882a593Smuzhiyun {
316*4882a593Smuzhiyun 	const struct clk_ops *ops = clk_dev_ops(clk->dev);
317*4882a593Smuzhiyun 
318*4882a593Smuzhiyun 	debug("%s(clk=%p)\n", __func__, clk);
319*4882a593Smuzhiyun 
320*4882a593Smuzhiyun 	if (!ops->free)
321*4882a593Smuzhiyun 		return 0;
322*4882a593Smuzhiyun 
323*4882a593Smuzhiyun 	return ops->free(clk);
324*4882a593Smuzhiyun }
325*4882a593Smuzhiyun 
clk_get_rate(struct clk * clk)326*4882a593Smuzhiyun ulong clk_get_rate(struct clk *clk)
327*4882a593Smuzhiyun {
328*4882a593Smuzhiyun 	const struct clk_ops *ops = clk_dev_ops(clk->dev);
329*4882a593Smuzhiyun 
330*4882a593Smuzhiyun 	debug("%s(clk=%p)\n", __func__, clk);
331*4882a593Smuzhiyun 
332*4882a593Smuzhiyun 	if (!ops->get_rate)
333*4882a593Smuzhiyun 		return -ENOSYS;
334*4882a593Smuzhiyun 
335*4882a593Smuzhiyun 	return ops->get_rate(clk);
336*4882a593Smuzhiyun }
337*4882a593Smuzhiyun 
clk_set_rate(struct clk * clk,ulong rate)338*4882a593Smuzhiyun ulong clk_set_rate(struct clk *clk, ulong rate)
339*4882a593Smuzhiyun {
340*4882a593Smuzhiyun 	const struct clk_ops *ops = clk_dev_ops(clk->dev);
341*4882a593Smuzhiyun 
342*4882a593Smuzhiyun 	debug("%s(clk=%p, rate=%lu)\n", __func__, clk, rate);
343*4882a593Smuzhiyun 
344*4882a593Smuzhiyun 	if (!ops->set_rate)
345*4882a593Smuzhiyun 		return -ENOSYS;
346*4882a593Smuzhiyun 
347*4882a593Smuzhiyun 	return ops->set_rate(clk, rate);
348*4882a593Smuzhiyun }
349*4882a593Smuzhiyun 
clk_get_phase(struct clk * clk)350*4882a593Smuzhiyun int clk_get_phase(struct clk *clk)
351*4882a593Smuzhiyun {
352*4882a593Smuzhiyun 	const struct clk_ops *ops = clk_dev_ops(clk->dev);
353*4882a593Smuzhiyun 
354*4882a593Smuzhiyun 	if (!ops->get_phase)
355*4882a593Smuzhiyun 		return -ENOSYS;
356*4882a593Smuzhiyun 
357*4882a593Smuzhiyun 	return ops->get_phase(clk);
358*4882a593Smuzhiyun }
359*4882a593Smuzhiyun 
clk_set_phase(struct clk * clk,int degrees)360*4882a593Smuzhiyun int clk_set_phase(struct clk *clk, int degrees)
361*4882a593Smuzhiyun {
362*4882a593Smuzhiyun 	const struct clk_ops *ops = clk_dev_ops(clk->dev);
363*4882a593Smuzhiyun 
364*4882a593Smuzhiyun 	if (!ops->set_phase)
365*4882a593Smuzhiyun 		return -ENOSYS;
366*4882a593Smuzhiyun 
367*4882a593Smuzhiyun 	return ops->set_phase(clk, degrees);
368*4882a593Smuzhiyun }
369*4882a593Smuzhiyun 
clk_set_parent(struct clk * clk,struct clk * parent)370*4882a593Smuzhiyun int clk_set_parent(struct clk *clk, struct clk *parent)
371*4882a593Smuzhiyun {
372*4882a593Smuzhiyun 	const struct clk_ops *ops = clk_dev_ops(clk->dev);
373*4882a593Smuzhiyun 
374*4882a593Smuzhiyun 	debug("%s(clk=%p, parent=%p)\n", __func__, clk, parent);
375*4882a593Smuzhiyun 
376*4882a593Smuzhiyun 	if (!ops->set_parent)
377*4882a593Smuzhiyun 		return -ENOSYS;
378*4882a593Smuzhiyun 
379*4882a593Smuzhiyun 	return ops->set_parent(clk, parent);
380*4882a593Smuzhiyun }
381*4882a593Smuzhiyun 
clk_enable(struct clk * clk)382*4882a593Smuzhiyun int clk_enable(struct clk *clk)
383*4882a593Smuzhiyun {
384*4882a593Smuzhiyun 	const struct clk_ops *ops = clk_dev_ops(clk->dev);
385*4882a593Smuzhiyun 
386*4882a593Smuzhiyun 	debug("%s(clk=%p)\n", __func__, clk);
387*4882a593Smuzhiyun 
388*4882a593Smuzhiyun 	if (!ops->enable)
389*4882a593Smuzhiyun 		return -ENOSYS;
390*4882a593Smuzhiyun 
391*4882a593Smuzhiyun 	return ops->enable(clk);
392*4882a593Smuzhiyun }
393*4882a593Smuzhiyun 
clk_enable_bulk(struct clk_bulk * bulk)394*4882a593Smuzhiyun int clk_enable_bulk(struct clk_bulk *bulk)
395*4882a593Smuzhiyun {
396*4882a593Smuzhiyun 	int i, ret;
397*4882a593Smuzhiyun 
398*4882a593Smuzhiyun 	for (i = 0; i < bulk->count; i++) {
399*4882a593Smuzhiyun 		ret = clk_enable(&bulk->clks[i]);
400*4882a593Smuzhiyun 		if (ret < 0 && ret != -ENOSYS)
401*4882a593Smuzhiyun 			return ret;
402*4882a593Smuzhiyun 	}
403*4882a593Smuzhiyun 
404*4882a593Smuzhiyun 	return 0;
405*4882a593Smuzhiyun }
406*4882a593Smuzhiyun 
clk_disable(struct clk * clk)407*4882a593Smuzhiyun int clk_disable(struct clk *clk)
408*4882a593Smuzhiyun {
409*4882a593Smuzhiyun 	const struct clk_ops *ops = clk_dev_ops(clk->dev);
410*4882a593Smuzhiyun 
411*4882a593Smuzhiyun 	debug("%s(clk=%p)\n", __func__, clk);
412*4882a593Smuzhiyun 
413*4882a593Smuzhiyun 	if (!ops->disable)
414*4882a593Smuzhiyun 		return -ENOSYS;
415*4882a593Smuzhiyun 
416*4882a593Smuzhiyun 	return ops->disable(clk);
417*4882a593Smuzhiyun }
418*4882a593Smuzhiyun 
clk_disable_bulk(struct clk_bulk * bulk)419*4882a593Smuzhiyun int clk_disable_bulk(struct clk_bulk *bulk)
420*4882a593Smuzhiyun {
421*4882a593Smuzhiyun 	int i, ret;
422*4882a593Smuzhiyun 
423*4882a593Smuzhiyun 	for (i = 0; i < bulk->count; i++) {
424*4882a593Smuzhiyun 		ret = clk_disable(&bulk->clks[i]);
425*4882a593Smuzhiyun 		if (ret < 0 && ret != -ENOSYS)
426*4882a593Smuzhiyun 			return ret;
427*4882a593Smuzhiyun 	}
428*4882a593Smuzhiyun 
429*4882a593Smuzhiyun 	return 0;
430*4882a593Smuzhiyun }
431*4882a593Smuzhiyun 
clks_probe(void)432*4882a593Smuzhiyun int clks_probe(void)
433*4882a593Smuzhiyun {
434*4882a593Smuzhiyun 	struct udevice *dev;
435*4882a593Smuzhiyun 	struct uclass *uc;
436*4882a593Smuzhiyun 	int ret;
437*4882a593Smuzhiyun 
438*4882a593Smuzhiyun 	ret = uclass_get(UCLASS_CLK, &uc);
439*4882a593Smuzhiyun 	if (ret)
440*4882a593Smuzhiyun 		return ret;
441*4882a593Smuzhiyun 
442*4882a593Smuzhiyun 	uclass_foreach_dev(dev, uc) {
443*4882a593Smuzhiyun 		ret = device_probe(dev);
444*4882a593Smuzhiyun 		if (ret)
445*4882a593Smuzhiyun 			printf("%s - probe failed: %d\n", dev->name, ret);
446*4882a593Smuzhiyun 	}
447*4882a593Smuzhiyun 
448*4882a593Smuzhiyun 	return 0;
449*4882a593Smuzhiyun }
450*4882a593Smuzhiyun 
451*4882a593Smuzhiyun UCLASS_DRIVER(clk) = {
452*4882a593Smuzhiyun 	.id		= UCLASS_CLK,
453*4882a593Smuzhiyun 	.name		= "clk",
454*4882a593Smuzhiyun };
455