1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-or-later
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * Copyright (c) 2014 MundoReader S.L.
4*4882a593Smuzhiyun * Author: Heiko Stuebner <heiko@sntech.de>
5*4882a593Smuzhiyun *
6*4882a593Smuzhiyun * Copyright (c) 2016 Rockchip Electronics Co. Ltd.
7*4882a593Smuzhiyun * Author: Xing Zheng <zhengxing@rock-chips.com>
8*4882a593Smuzhiyun *
9*4882a593Smuzhiyun * based on
10*4882a593Smuzhiyun *
11*4882a593Smuzhiyun * samsung/clk.c
12*4882a593Smuzhiyun * Copyright (c) 2013 Samsung Electronics Co., Ltd.
13*4882a593Smuzhiyun * Copyright (c) 2013 Linaro Ltd.
14*4882a593Smuzhiyun * Author: Thomas Abraham <thomas.ab@samsung.com>
15*4882a593Smuzhiyun */
16*4882a593Smuzhiyun
17*4882a593Smuzhiyun #include <linux/slab.h>
18*4882a593Smuzhiyun #include <linux/clk.h>
19*4882a593Smuzhiyun #include <linux/clk-provider.h>
20*4882a593Smuzhiyun #include <linux/io.h>
21*4882a593Smuzhiyun #include <linux/mfd/syscon.h>
22*4882a593Smuzhiyun #include <linux/regmap.h>
23*4882a593Smuzhiyun #include <linux/reboot.h>
24*4882a593Smuzhiyun #include <linux/rational.h>
25*4882a593Smuzhiyun #include "clk.h"
26*4882a593Smuzhiyun
27*4882a593Smuzhiyun /**
28*4882a593Smuzhiyun * Register a clock branch.
29*4882a593Smuzhiyun * Most clock branches have a form like
30*4882a593Smuzhiyun *
31*4882a593Smuzhiyun * src1 --|--\
32*4882a593Smuzhiyun * |M |--[GATE]-[DIV]-
33*4882a593Smuzhiyun * src2 --|--/
34*4882a593Smuzhiyun *
35*4882a593Smuzhiyun * sometimes without one of those components.
36*4882a593Smuzhiyun */
rockchip_clk_register_branch(const char * name,const char * const * parent_names,u8 num_parents,void __iomem * base,int muxdiv_offset,u8 mux_shift,u8 mux_width,u8 mux_flags,u32 * mux_table,int div_offset,u8 div_shift,u8 div_width,u8 div_flags,struct clk_div_table * div_table,int gate_offset,u8 gate_shift,u8 gate_flags,unsigned long flags,spinlock_t * lock)37*4882a593Smuzhiyun static struct clk *rockchip_clk_register_branch(const char *name,
38*4882a593Smuzhiyun const char *const *parent_names, u8 num_parents,
39*4882a593Smuzhiyun void __iomem *base,
40*4882a593Smuzhiyun int muxdiv_offset, u8 mux_shift, u8 mux_width, u8 mux_flags,
41*4882a593Smuzhiyun u32 *mux_table,
42*4882a593Smuzhiyun int div_offset, u8 div_shift, u8 div_width, u8 div_flags,
43*4882a593Smuzhiyun struct clk_div_table *div_table, int gate_offset,
44*4882a593Smuzhiyun u8 gate_shift, u8 gate_flags, unsigned long flags,
45*4882a593Smuzhiyun spinlock_t *lock)
46*4882a593Smuzhiyun {
47*4882a593Smuzhiyun struct clk_hw *hw;
48*4882a593Smuzhiyun struct clk_mux *mux = NULL;
49*4882a593Smuzhiyun struct clk_gate *gate = NULL;
50*4882a593Smuzhiyun struct clk_divider *div = NULL;
51*4882a593Smuzhiyun const struct clk_ops *mux_ops = NULL, *div_ops = NULL,
52*4882a593Smuzhiyun *gate_ops = NULL;
53*4882a593Smuzhiyun int ret;
54*4882a593Smuzhiyun
55*4882a593Smuzhiyun if (num_parents > 1) {
56*4882a593Smuzhiyun mux = kzalloc(sizeof(*mux), GFP_KERNEL);
57*4882a593Smuzhiyun if (!mux)
58*4882a593Smuzhiyun return ERR_PTR(-ENOMEM);
59*4882a593Smuzhiyun
60*4882a593Smuzhiyun mux->reg = base + muxdiv_offset;
61*4882a593Smuzhiyun mux->shift = mux_shift;
62*4882a593Smuzhiyun mux->mask = BIT(mux_width) - 1;
63*4882a593Smuzhiyun mux->flags = mux_flags;
64*4882a593Smuzhiyun mux->table = mux_table;
65*4882a593Smuzhiyun mux->lock = lock;
66*4882a593Smuzhiyun mux_ops = (mux_flags & CLK_MUX_READ_ONLY) ? &clk_mux_ro_ops
67*4882a593Smuzhiyun : &clk_mux_ops;
68*4882a593Smuzhiyun }
69*4882a593Smuzhiyun
70*4882a593Smuzhiyun if (gate_offset >= 0) {
71*4882a593Smuzhiyun gate = kzalloc(sizeof(*gate), GFP_KERNEL);
72*4882a593Smuzhiyun if (!gate) {
73*4882a593Smuzhiyun ret = -ENOMEM;
74*4882a593Smuzhiyun goto err_gate;
75*4882a593Smuzhiyun }
76*4882a593Smuzhiyun
77*4882a593Smuzhiyun gate->flags = gate_flags;
78*4882a593Smuzhiyun gate->reg = base + gate_offset;
79*4882a593Smuzhiyun gate->bit_idx = gate_shift;
80*4882a593Smuzhiyun gate->lock = lock;
81*4882a593Smuzhiyun gate_ops = &clk_gate_ops;
82*4882a593Smuzhiyun }
83*4882a593Smuzhiyun
84*4882a593Smuzhiyun if (div_width > 0) {
85*4882a593Smuzhiyun div = kzalloc(sizeof(*div), GFP_KERNEL);
86*4882a593Smuzhiyun if (!div) {
87*4882a593Smuzhiyun ret = -ENOMEM;
88*4882a593Smuzhiyun goto err_div;
89*4882a593Smuzhiyun }
90*4882a593Smuzhiyun
91*4882a593Smuzhiyun div->flags = div_flags;
92*4882a593Smuzhiyun if (div_offset)
93*4882a593Smuzhiyun div->reg = base + div_offset;
94*4882a593Smuzhiyun else
95*4882a593Smuzhiyun div->reg = base + muxdiv_offset;
96*4882a593Smuzhiyun div->shift = div_shift;
97*4882a593Smuzhiyun div->width = div_width;
98*4882a593Smuzhiyun div->lock = lock;
99*4882a593Smuzhiyun div->table = div_table;
100*4882a593Smuzhiyun div_ops = (div_flags & CLK_DIVIDER_READ_ONLY)
101*4882a593Smuzhiyun ? &clk_divider_ro_ops
102*4882a593Smuzhiyun : &clk_divider_ops;
103*4882a593Smuzhiyun }
104*4882a593Smuzhiyun
105*4882a593Smuzhiyun hw = clk_hw_register_composite(NULL, name, parent_names, num_parents,
106*4882a593Smuzhiyun mux ? &mux->hw : NULL, mux_ops,
107*4882a593Smuzhiyun div ? &div->hw : NULL, div_ops,
108*4882a593Smuzhiyun gate ? &gate->hw : NULL, gate_ops,
109*4882a593Smuzhiyun flags);
110*4882a593Smuzhiyun if (IS_ERR(hw)) {
111*4882a593Smuzhiyun kfree(div);
112*4882a593Smuzhiyun kfree(gate);
113*4882a593Smuzhiyun return ERR_CAST(hw);
114*4882a593Smuzhiyun }
115*4882a593Smuzhiyun
116*4882a593Smuzhiyun return hw->clk;
117*4882a593Smuzhiyun err_div:
118*4882a593Smuzhiyun kfree(gate);
119*4882a593Smuzhiyun err_gate:
120*4882a593Smuzhiyun kfree(mux);
121*4882a593Smuzhiyun return ERR_PTR(ret);
122*4882a593Smuzhiyun }
123*4882a593Smuzhiyun
124*4882a593Smuzhiyun struct rockchip_clk_frac {
125*4882a593Smuzhiyun struct notifier_block clk_nb;
126*4882a593Smuzhiyun struct clk_fractional_divider div;
127*4882a593Smuzhiyun struct clk_gate gate;
128*4882a593Smuzhiyun
129*4882a593Smuzhiyun struct clk_mux mux;
130*4882a593Smuzhiyun const struct clk_ops *mux_ops;
131*4882a593Smuzhiyun int mux_frac_idx;
132*4882a593Smuzhiyun
133*4882a593Smuzhiyun bool rate_change_remuxed;
134*4882a593Smuzhiyun int rate_change_idx;
135*4882a593Smuzhiyun };
136*4882a593Smuzhiyun
137*4882a593Smuzhiyun #define to_rockchip_clk_frac_nb(nb) \
138*4882a593Smuzhiyun container_of(nb, struct rockchip_clk_frac, clk_nb)
139*4882a593Smuzhiyun
rockchip_clk_frac_notifier_cb(struct notifier_block * nb,unsigned long event,void * data)140*4882a593Smuzhiyun static int rockchip_clk_frac_notifier_cb(struct notifier_block *nb,
141*4882a593Smuzhiyun unsigned long event, void *data)
142*4882a593Smuzhiyun {
143*4882a593Smuzhiyun struct clk_notifier_data *ndata = data;
144*4882a593Smuzhiyun struct rockchip_clk_frac *frac = to_rockchip_clk_frac_nb(nb);
145*4882a593Smuzhiyun struct clk_mux *frac_mux = &frac->mux;
146*4882a593Smuzhiyun int ret = 0;
147*4882a593Smuzhiyun
148*4882a593Smuzhiyun pr_debug("%s: event %lu, old_rate %lu, new_rate: %lu\n",
149*4882a593Smuzhiyun __func__, event, ndata->old_rate, ndata->new_rate);
150*4882a593Smuzhiyun if (event == PRE_RATE_CHANGE) {
151*4882a593Smuzhiyun frac->rate_change_idx =
152*4882a593Smuzhiyun frac->mux_ops->get_parent(&frac_mux->hw);
153*4882a593Smuzhiyun if (frac->rate_change_idx != frac->mux_frac_idx) {
154*4882a593Smuzhiyun frac->mux_ops->set_parent(&frac_mux->hw,
155*4882a593Smuzhiyun frac->mux_frac_idx);
156*4882a593Smuzhiyun frac->rate_change_remuxed = 1;
157*4882a593Smuzhiyun }
158*4882a593Smuzhiyun } else if (event == POST_RATE_CHANGE) {
159*4882a593Smuzhiyun /*
160*4882a593Smuzhiyun * The POST_RATE_CHANGE notifier runs directly after the
161*4882a593Smuzhiyun * divider clock is set in clk_change_rate, so we'll have
162*4882a593Smuzhiyun * remuxed back to the original parent before clk_change_rate
163*4882a593Smuzhiyun * reaches the mux itself.
164*4882a593Smuzhiyun */
165*4882a593Smuzhiyun if (frac->rate_change_remuxed) {
166*4882a593Smuzhiyun frac->mux_ops->set_parent(&frac_mux->hw,
167*4882a593Smuzhiyun frac->rate_change_idx);
168*4882a593Smuzhiyun frac->rate_change_remuxed = 0;
169*4882a593Smuzhiyun }
170*4882a593Smuzhiyun }
171*4882a593Smuzhiyun
172*4882a593Smuzhiyun return notifier_from_errno(ret);
173*4882a593Smuzhiyun }
174*4882a593Smuzhiyun
175*4882a593Smuzhiyun /**
176*4882a593Smuzhiyun * fractional divider must set that denominator is 20 times larger than
177*4882a593Smuzhiyun * numerator to generate precise clock frequency.
178*4882a593Smuzhiyun */
rockchip_fractional_approximation(struct clk_hw * hw,unsigned long rate,unsigned long * parent_rate,unsigned long * m,unsigned long * n)179*4882a593Smuzhiyun static void rockchip_fractional_approximation(struct clk_hw *hw,
180*4882a593Smuzhiyun unsigned long rate, unsigned long *parent_rate,
181*4882a593Smuzhiyun unsigned long *m, unsigned long *n)
182*4882a593Smuzhiyun {
183*4882a593Smuzhiyun struct clk_fractional_divider *fd = to_clk_fd(hw);
184*4882a593Smuzhiyun unsigned long p_rate, p_parent_rate;
185*4882a593Smuzhiyun struct clk_hw *p_parent;
186*4882a593Smuzhiyun unsigned long scale;
187*4882a593Smuzhiyun
188*4882a593Smuzhiyun p_rate = clk_hw_get_rate(clk_hw_get_parent(hw));
189*4882a593Smuzhiyun if ((rate * 20 > p_rate) && (p_rate % rate != 0)) {
190*4882a593Smuzhiyun p_parent = clk_hw_get_parent(clk_hw_get_parent(hw));
191*4882a593Smuzhiyun if (!p_parent) {
192*4882a593Smuzhiyun *parent_rate = p_rate;
193*4882a593Smuzhiyun } else {
194*4882a593Smuzhiyun p_parent_rate = clk_hw_get_rate(p_parent);
195*4882a593Smuzhiyun *parent_rate = p_parent_rate;
196*4882a593Smuzhiyun }
197*4882a593Smuzhiyun
198*4882a593Smuzhiyun if (*parent_rate < rate * 20) {
199*4882a593Smuzhiyun /*
200*4882a593Smuzhiyun * Fractional frequency divider to do
201*4882a593Smuzhiyun * integer frequency divider does not
202*4882a593Smuzhiyun * need 20 times the limit.
203*4882a593Smuzhiyun */
204*4882a593Smuzhiyun if (!(*parent_rate % rate)) {
205*4882a593Smuzhiyun *m = 1;
206*4882a593Smuzhiyun *n = *parent_rate / rate;
207*4882a593Smuzhiyun return;
208*4882a593Smuzhiyun } else if (!(fd->flags & CLK_FRAC_DIVIDER_NO_LIMIT)) {
209*4882a593Smuzhiyun pr_warn("%s p_rate(%ld) is low than rate(%ld)*20, use integer or half-div\n",
210*4882a593Smuzhiyun clk_hw_get_name(hw),
211*4882a593Smuzhiyun *parent_rate, rate);
212*4882a593Smuzhiyun *m = 0;
213*4882a593Smuzhiyun *n = 1;
214*4882a593Smuzhiyun return;
215*4882a593Smuzhiyun }
216*4882a593Smuzhiyun }
217*4882a593Smuzhiyun }
218*4882a593Smuzhiyun
219*4882a593Smuzhiyun /*
220*4882a593Smuzhiyun * Get rate closer to *parent_rate to guarantee there is no overflow
221*4882a593Smuzhiyun * for m and n. In the result it will be the nearest rate left shifted
222*4882a593Smuzhiyun * by (scale - fd->nwidth) bits.
223*4882a593Smuzhiyun */
224*4882a593Smuzhiyun scale = fls_long(*parent_rate / rate - 1);
225*4882a593Smuzhiyun if (scale > fd->nwidth)
226*4882a593Smuzhiyun rate <<= scale - fd->nwidth;
227*4882a593Smuzhiyun
228*4882a593Smuzhiyun rational_best_approximation(rate, *parent_rate,
229*4882a593Smuzhiyun GENMASK(fd->mwidth - 1, 0), GENMASK(fd->nwidth - 1, 0),
230*4882a593Smuzhiyun m, n);
231*4882a593Smuzhiyun }
232*4882a593Smuzhiyun
rockchip_clk_register_frac_branch(struct rockchip_clk_provider * ctx,const char * name,const char * const * parent_names,u8 num_parents,void __iomem * base,int muxdiv_offset,u8 div_flags,int gate_offset,u8 gate_shift,u8 gate_flags,unsigned long flags,struct rockchip_clk_branch * child,spinlock_t * lock)233*4882a593Smuzhiyun static struct clk *rockchip_clk_register_frac_branch(
234*4882a593Smuzhiyun struct rockchip_clk_provider *ctx, const char *name,
235*4882a593Smuzhiyun const char *const *parent_names, u8 num_parents,
236*4882a593Smuzhiyun void __iomem *base, int muxdiv_offset, u8 div_flags,
237*4882a593Smuzhiyun int gate_offset, u8 gate_shift, u8 gate_flags,
238*4882a593Smuzhiyun unsigned long flags, struct rockchip_clk_branch *child,
239*4882a593Smuzhiyun spinlock_t *lock)
240*4882a593Smuzhiyun {
241*4882a593Smuzhiyun struct clk_hw *hw;
242*4882a593Smuzhiyun struct rockchip_clk_frac *frac;
243*4882a593Smuzhiyun struct clk_gate *gate = NULL;
244*4882a593Smuzhiyun struct clk_fractional_divider *div = NULL;
245*4882a593Smuzhiyun const struct clk_ops *div_ops = NULL, *gate_ops = NULL;
246*4882a593Smuzhiyun
247*4882a593Smuzhiyun if (muxdiv_offset < 0)
248*4882a593Smuzhiyun return ERR_PTR(-EINVAL);
249*4882a593Smuzhiyun
250*4882a593Smuzhiyun if (child && child->branch_type != branch_mux) {
251*4882a593Smuzhiyun pr_err("%s: fractional child clock for %s can only be a mux\n",
252*4882a593Smuzhiyun __func__, name);
253*4882a593Smuzhiyun return ERR_PTR(-EINVAL);
254*4882a593Smuzhiyun }
255*4882a593Smuzhiyun
256*4882a593Smuzhiyun frac = kzalloc(sizeof(*frac), GFP_KERNEL);
257*4882a593Smuzhiyun if (!frac)
258*4882a593Smuzhiyun return ERR_PTR(-ENOMEM);
259*4882a593Smuzhiyun
260*4882a593Smuzhiyun if (gate_offset >= 0) {
261*4882a593Smuzhiyun gate = &frac->gate;
262*4882a593Smuzhiyun gate->flags = gate_flags;
263*4882a593Smuzhiyun gate->reg = base + gate_offset;
264*4882a593Smuzhiyun gate->bit_idx = gate_shift;
265*4882a593Smuzhiyun gate->lock = lock;
266*4882a593Smuzhiyun gate_ops = &clk_gate_ops;
267*4882a593Smuzhiyun }
268*4882a593Smuzhiyun
269*4882a593Smuzhiyun div = &frac->div;
270*4882a593Smuzhiyun div->flags = div_flags;
271*4882a593Smuzhiyun div->reg = base + muxdiv_offset;
272*4882a593Smuzhiyun div->mshift = 16;
273*4882a593Smuzhiyun div->mwidth = 16;
274*4882a593Smuzhiyun div->mmask = GENMASK(div->mwidth - 1, 0) << div->mshift;
275*4882a593Smuzhiyun div->nshift = 0;
276*4882a593Smuzhiyun div->nwidth = 16;
277*4882a593Smuzhiyun div->nmask = GENMASK(div->nwidth - 1, 0) << div->nshift;
278*4882a593Smuzhiyun div->lock = lock;
279*4882a593Smuzhiyun div->approximation = rockchip_fractional_approximation;
280*4882a593Smuzhiyun div_ops = &clk_fractional_divider_ops;
281*4882a593Smuzhiyun
282*4882a593Smuzhiyun hw = clk_hw_register_composite(NULL, name, parent_names, num_parents,
283*4882a593Smuzhiyun NULL, NULL,
284*4882a593Smuzhiyun &div->hw, div_ops,
285*4882a593Smuzhiyun gate ? &gate->hw : NULL, gate_ops,
286*4882a593Smuzhiyun flags | CLK_SET_RATE_UNGATE);
287*4882a593Smuzhiyun if (IS_ERR(hw)) {
288*4882a593Smuzhiyun kfree(frac);
289*4882a593Smuzhiyun return ERR_CAST(hw);
290*4882a593Smuzhiyun }
291*4882a593Smuzhiyun
292*4882a593Smuzhiyun if (child) {
293*4882a593Smuzhiyun struct clk_mux *frac_mux = &frac->mux;
294*4882a593Smuzhiyun struct clk_init_data init;
295*4882a593Smuzhiyun struct clk *mux_clk;
296*4882a593Smuzhiyun int ret;
297*4882a593Smuzhiyun
298*4882a593Smuzhiyun frac->mux_frac_idx = match_string(child->parent_names,
299*4882a593Smuzhiyun child->num_parents, name);
300*4882a593Smuzhiyun frac->mux_ops = &clk_mux_ops;
301*4882a593Smuzhiyun frac->clk_nb.notifier_call = rockchip_clk_frac_notifier_cb;
302*4882a593Smuzhiyun
303*4882a593Smuzhiyun frac_mux->reg = base + child->muxdiv_offset;
304*4882a593Smuzhiyun frac_mux->shift = child->mux_shift;
305*4882a593Smuzhiyun frac_mux->mask = BIT(child->mux_width) - 1;
306*4882a593Smuzhiyun frac_mux->flags = child->mux_flags;
307*4882a593Smuzhiyun if (child->mux_table)
308*4882a593Smuzhiyun frac_mux->table = child->mux_table;
309*4882a593Smuzhiyun frac_mux->lock = lock;
310*4882a593Smuzhiyun frac_mux->hw.init = &init;
311*4882a593Smuzhiyun
312*4882a593Smuzhiyun init.name = child->name;
313*4882a593Smuzhiyun init.flags = child->flags | CLK_SET_RATE_PARENT;
314*4882a593Smuzhiyun init.ops = frac->mux_ops;
315*4882a593Smuzhiyun init.parent_names = child->parent_names;
316*4882a593Smuzhiyun init.num_parents = child->num_parents;
317*4882a593Smuzhiyun
318*4882a593Smuzhiyun mux_clk = clk_register(NULL, &frac_mux->hw);
319*4882a593Smuzhiyun if (IS_ERR(mux_clk)) {
320*4882a593Smuzhiyun kfree(frac);
321*4882a593Smuzhiyun return mux_clk;
322*4882a593Smuzhiyun }
323*4882a593Smuzhiyun
324*4882a593Smuzhiyun rockchip_clk_add_lookup(ctx, mux_clk, child->id);
325*4882a593Smuzhiyun
326*4882a593Smuzhiyun /* notifier on the fraction divider to catch rate changes */
327*4882a593Smuzhiyun if (frac->mux_frac_idx >= 0) {
328*4882a593Smuzhiyun pr_debug("%s: found fractional parent in mux at pos %d\n",
329*4882a593Smuzhiyun __func__, frac->mux_frac_idx);
330*4882a593Smuzhiyun ret = clk_notifier_register(hw->clk, &frac->clk_nb);
331*4882a593Smuzhiyun if (ret)
332*4882a593Smuzhiyun pr_err("%s: failed to register clock notifier for %s\n",
333*4882a593Smuzhiyun __func__, name);
334*4882a593Smuzhiyun } else {
335*4882a593Smuzhiyun pr_warn("%s: could not find %s as parent of %s, rate changes may not work\n",
336*4882a593Smuzhiyun __func__, name, child->name);
337*4882a593Smuzhiyun }
338*4882a593Smuzhiyun }
339*4882a593Smuzhiyun
340*4882a593Smuzhiyun return hw->clk;
341*4882a593Smuzhiyun }
342*4882a593Smuzhiyun
rockchip_clk_register_factor_branch(const char * name,const char * const * parent_names,u8 num_parents,void __iomem * base,unsigned int mult,unsigned int div,int gate_offset,u8 gate_shift,u8 gate_flags,unsigned long flags,spinlock_t * lock)343*4882a593Smuzhiyun static struct clk *rockchip_clk_register_factor_branch(const char *name,
344*4882a593Smuzhiyun const char *const *parent_names, u8 num_parents,
345*4882a593Smuzhiyun void __iomem *base, unsigned int mult, unsigned int div,
346*4882a593Smuzhiyun int gate_offset, u8 gate_shift, u8 gate_flags,
347*4882a593Smuzhiyun unsigned long flags, spinlock_t *lock)
348*4882a593Smuzhiyun {
349*4882a593Smuzhiyun struct clk_hw *hw;
350*4882a593Smuzhiyun struct clk_gate *gate = NULL;
351*4882a593Smuzhiyun struct clk_fixed_factor *fix = NULL;
352*4882a593Smuzhiyun
353*4882a593Smuzhiyun /* without gate, register a simple factor clock */
354*4882a593Smuzhiyun if (gate_offset == 0) {
355*4882a593Smuzhiyun return clk_register_fixed_factor(NULL, name,
356*4882a593Smuzhiyun parent_names[0], flags, mult,
357*4882a593Smuzhiyun div);
358*4882a593Smuzhiyun }
359*4882a593Smuzhiyun
360*4882a593Smuzhiyun gate = kzalloc(sizeof(*gate), GFP_KERNEL);
361*4882a593Smuzhiyun if (!gate)
362*4882a593Smuzhiyun return ERR_PTR(-ENOMEM);
363*4882a593Smuzhiyun
364*4882a593Smuzhiyun gate->flags = gate_flags;
365*4882a593Smuzhiyun gate->reg = base + gate_offset;
366*4882a593Smuzhiyun gate->bit_idx = gate_shift;
367*4882a593Smuzhiyun gate->lock = lock;
368*4882a593Smuzhiyun
369*4882a593Smuzhiyun fix = kzalloc(sizeof(*fix), GFP_KERNEL);
370*4882a593Smuzhiyun if (!fix) {
371*4882a593Smuzhiyun kfree(gate);
372*4882a593Smuzhiyun return ERR_PTR(-ENOMEM);
373*4882a593Smuzhiyun }
374*4882a593Smuzhiyun
375*4882a593Smuzhiyun fix->mult = mult;
376*4882a593Smuzhiyun fix->div = div;
377*4882a593Smuzhiyun
378*4882a593Smuzhiyun hw = clk_hw_register_composite(NULL, name, parent_names, num_parents,
379*4882a593Smuzhiyun NULL, NULL,
380*4882a593Smuzhiyun &fix->hw, &clk_fixed_factor_ops,
381*4882a593Smuzhiyun &gate->hw, &clk_gate_ops, flags);
382*4882a593Smuzhiyun if (IS_ERR(hw)) {
383*4882a593Smuzhiyun kfree(fix);
384*4882a593Smuzhiyun kfree(gate);
385*4882a593Smuzhiyun return ERR_CAST(hw);
386*4882a593Smuzhiyun }
387*4882a593Smuzhiyun
388*4882a593Smuzhiyun return hw->clk;
389*4882a593Smuzhiyun }
390*4882a593Smuzhiyun
rockchip_clk_init(struct device_node * np,void __iomem * base,unsigned long nr_clks)391*4882a593Smuzhiyun struct rockchip_clk_provider *rockchip_clk_init(struct device_node *np,
392*4882a593Smuzhiyun void __iomem *base,
393*4882a593Smuzhiyun unsigned long nr_clks)
394*4882a593Smuzhiyun {
395*4882a593Smuzhiyun struct rockchip_clk_provider *ctx;
396*4882a593Smuzhiyun struct clk **clk_table;
397*4882a593Smuzhiyun int i;
398*4882a593Smuzhiyun
399*4882a593Smuzhiyun ctx = kzalloc(sizeof(struct rockchip_clk_provider), GFP_KERNEL);
400*4882a593Smuzhiyun if (!ctx)
401*4882a593Smuzhiyun return ERR_PTR(-ENOMEM);
402*4882a593Smuzhiyun
403*4882a593Smuzhiyun clk_table = kcalloc(nr_clks, sizeof(struct clk *), GFP_KERNEL);
404*4882a593Smuzhiyun if (!clk_table)
405*4882a593Smuzhiyun goto err_free;
406*4882a593Smuzhiyun
407*4882a593Smuzhiyun for (i = 0; i < nr_clks; ++i)
408*4882a593Smuzhiyun clk_table[i] = ERR_PTR(-ENOENT);
409*4882a593Smuzhiyun
410*4882a593Smuzhiyun ctx->reg_base = base;
411*4882a593Smuzhiyun ctx->clk_data.clks = clk_table;
412*4882a593Smuzhiyun ctx->clk_data.clk_num = nr_clks;
413*4882a593Smuzhiyun ctx->cru_node = np;
414*4882a593Smuzhiyun spin_lock_init(&ctx->lock);
415*4882a593Smuzhiyun
416*4882a593Smuzhiyun ctx->grf = syscon_regmap_lookup_by_phandle(ctx->cru_node,
417*4882a593Smuzhiyun "rockchip,grf");
418*4882a593Smuzhiyun ctx->pmugrf = syscon_regmap_lookup_by_phandle(ctx->cru_node,
419*4882a593Smuzhiyun "rockchip,pmugrf");
420*4882a593Smuzhiyun
421*4882a593Smuzhiyun return ctx;
422*4882a593Smuzhiyun
423*4882a593Smuzhiyun err_free:
424*4882a593Smuzhiyun kfree(ctx);
425*4882a593Smuzhiyun return ERR_PTR(-ENOMEM);
426*4882a593Smuzhiyun }
427*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rockchip_clk_init);
428*4882a593Smuzhiyun
rockchip_clk_of_add_provider(struct device_node * np,struct rockchip_clk_provider * ctx)429*4882a593Smuzhiyun void rockchip_clk_of_add_provider(struct device_node *np,
430*4882a593Smuzhiyun struct rockchip_clk_provider *ctx)
431*4882a593Smuzhiyun {
432*4882a593Smuzhiyun if (of_clk_add_provider(np, of_clk_src_onecell_get,
433*4882a593Smuzhiyun &ctx->clk_data))
434*4882a593Smuzhiyun pr_err("%s: could not register clk provider\n", __func__);
435*4882a593Smuzhiyun }
436*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rockchip_clk_of_add_provider);
437*4882a593Smuzhiyun
rockchip_clk_add_lookup(struct rockchip_clk_provider * ctx,struct clk * clk,unsigned int id)438*4882a593Smuzhiyun void rockchip_clk_add_lookup(struct rockchip_clk_provider *ctx,
439*4882a593Smuzhiyun struct clk *clk, unsigned int id)
440*4882a593Smuzhiyun {
441*4882a593Smuzhiyun if (ctx->clk_data.clks && id)
442*4882a593Smuzhiyun ctx->clk_data.clks[id] = clk;
443*4882a593Smuzhiyun }
444*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rockchip_clk_add_lookup);
445*4882a593Smuzhiyun
rockchip_clk_register_plls(struct rockchip_clk_provider * ctx,struct rockchip_pll_clock * list,unsigned int nr_pll,int grf_lock_offset)446*4882a593Smuzhiyun void rockchip_clk_register_plls(struct rockchip_clk_provider *ctx,
447*4882a593Smuzhiyun struct rockchip_pll_clock *list,
448*4882a593Smuzhiyun unsigned int nr_pll, int grf_lock_offset)
449*4882a593Smuzhiyun {
450*4882a593Smuzhiyun struct clk *clk;
451*4882a593Smuzhiyun int idx;
452*4882a593Smuzhiyun
453*4882a593Smuzhiyun for (idx = 0; idx < nr_pll; idx++, list++) {
454*4882a593Smuzhiyun clk = rockchip_clk_register_pll(ctx, list->type, list->name,
455*4882a593Smuzhiyun list->parent_names, list->num_parents,
456*4882a593Smuzhiyun list->con_offset, grf_lock_offset,
457*4882a593Smuzhiyun list->lock_shift, list->mode_offset,
458*4882a593Smuzhiyun list->mode_shift, list->rate_table,
459*4882a593Smuzhiyun list->flags, list->pll_flags);
460*4882a593Smuzhiyun if (IS_ERR(clk)) {
461*4882a593Smuzhiyun pr_err("%s: failed to register clock %s\n", __func__,
462*4882a593Smuzhiyun list->name);
463*4882a593Smuzhiyun continue;
464*4882a593Smuzhiyun }
465*4882a593Smuzhiyun
466*4882a593Smuzhiyun rockchip_clk_add_lookup(ctx, clk, list->id);
467*4882a593Smuzhiyun }
468*4882a593Smuzhiyun }
469*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rockchip_clk_register_plls);
470*4882a593Smuzhiyun
rockchip_clk_register_branches(struct rockchip_clk_provider * ctx,struct rockchip_clk_branch * list,unsigned int nr_clk)471*4882a593Smuzhiyun void rockchip_clk_register_branches(struct rockchip_clk_provider *ctx,
472*4882a593Smuzhiyun struct rockchip_clk_branch *list,
473*4882a593Smuzhiyun unsigned int nr_clk)
474*4882a593Smuzhiyun {
475*4882a593Smuzhiyun struct clk *clk = NULL;
476*4882a593Smuzhiyun unsigned int idx;
477*4882a593Smuzhiyun unsigned long flags;
478*4882a593Smuzhiyun
479*4882a593Smuzhiyun for (idx = 0; idx < nr_clk; idx++, list++) {
480*4882a593Smuzhiyun flags = list->flags;
481*4882a593Smuzhiyun
482*4882a593Smuzhiyun /* catch simple muxes */
483*4882a593Smuzhiyun switch (list->branch_type) {
484*4882a593Smuzhiyun case branch_mux:
485*4882a593Smuzhiyun if (list->mux_table)
486*4882a593Smuzhiyun clk = clk_register_mux_table(NULL, list->name,
487*4882a593Smuzhiyun list->parent_names, list->num_parents,
488*4882a593Smuzhiyun flags,
489*4882a593Smuzhiyun ctx->reg_base + list->muxdiv_offset,
490*4882a593Smuzhiyun list->mux_shift,
491*4882a593Smuzhiyun BIT(list->mux_width) - 1,
492*4882a593Smuzhiyun list->mux_flags, list->mux_table,
493*4882a593Smuzhiyun &ctx->lock);
494*4882a593Smuzhiyun else
495*4882a593Smuzhiyun clk = clk_register_mux(NULL, list->name,
496*4882a593Smuzhiyun list->parent_names, list->num_parents,
497*4882a593Smuzhiyun flags,
498*4882a593Smuzhiyun ctx->reg_base + list->muxdiv_offset,
499*4882a593Smuzhiyun list->mux_shift, list->mux_width,
500*4882a593Smuzhiyun list->mux_flags, &ctx->lock);
501*4882a593Smuzhiyun break;
502*4882a593Smuzhiyun case branch_muxgrf:
503*4882a593Smuzhiyun clk = rockchip_clk_register_muxgrf(list->name,
504*4882a593Smuzhiyun list->parent_names, list->num_parents,
505*4882a593Smuzhiyun flags, ctx->grf, list->muxdiv_offset,
506*4882a593Smuzhiyun list->mux_shift, list->mux_width,
507*4882a593Smuzhiyun list->mux_flags);
508*4882a593Smuzhiyun break;
509*4882a593Smuzhiyun case branch_muxpmugrf:
510*4882a593Smuzhiyun clk = rockchip_clk_register_muxgrf(list->name,
511*4882a593Smuzhiyun list->parent_names, list->num_parents,
512*4882a593Smuzhiyun flags, ctx->pmugrf, list->muxdiv_offset,
513*4882a593Smuzhiyun list->mux_shift, list->mux_width,
514*4882a593Smuzhiyun list->mux_flags);
515*4882a593Smuzhiyun break;
516*4882a593Smuzhiyun case branch_divider:
517*4882a593Smuzhiyun if (list->div_table)
518*4882a593Smuzhiyun clk = clk_register_divider_table(NULL,
519*4882a593Smuzhiyun list->name, list->parent_names[0],
520*4882a593Smuzhiyun flags,
521*4882a593Smuzhiyun ctx->reg_base + list->muxdiv_offset,
522*4882a593Smuzhiyun list->div_shift, list->div_width,
523*4882a593Smuzhiyun list->div_flags, list->div_table,
524*4882a593Smuzhiyun &ctx->lock);
525*4882a593Smuzhiyun else
526*4882a593Smuzhiyun clk = clk_register_divider(NULL, list->name,
527*4882a593Smuzhiyun list->parent_names[0], flags,
528*4882a593Smuzhiyun ctx->reg_base + list->muxdiv_offset,
529*4882a593Smuzhiyun list->div_shift, list->div_width,
530*4882a593Smuzhiyun list->div_flags, &ctx->lock);
531*4882a593Smuzhiyun break;
532*4882a593Smuzhiyun case branch_fraction_divider:
533*4882a593Smuzhiyun clk = rockchip_clk_register_frac_branch(ctx, list->name,
534*4882a593Smuzhiyun list->parent_names, list->num_parents,
535*4882a593Smuzhiyun ctx->reg_base, list->muxdiv_offset,
536*4882a593Smuzhiyun list->div_flags,
537*4882a593Smuzhiyun list->gate_offset, list->gate_shift,
538*4882a593Smuzhiyun list->gate_flags, flags, list->child,
539*4882a593Smuzhiyun &ctx->lock);
540*4882a593Smuzhiyun break;
541*4882a593Smuzhiyun case branch_half_divider:
542*4882a593Smuzhiyun clk = rockchip_clk_register_halfdiv(list->name,
543*4882a593Smuzhiyun list->parent_names, list->num_parents,
544*4882a593Smuzhiyun ctx->reg_base, list->muxdiv_offset,
545*4882a593Smuzhiyun list->mux_shift, list->mux_width,
546*4882a593Smuzhiyun list->mux_flags, list->div_offset,
547*4882a593Smuzhiyun list->div_shift, list->div_width,
548*4882a593Smuzhiyun list->div_flags, list->gate_offset,
549*4882a593Smuzhiyun list->gate_shift, list->gate_flags,
550*4882a593Smuzhiyun flags, &ctx->lock);
551*4882a593Smuzhiyun break;
552*4882a593Smuzhiyun case branch_gate:
553*4882a593Smuzhiyun flags |= CLK_SET_RATE_PARENT;
554*4882a593Smuzhiyun
555*4882a593Smuzhiyun clk = clk_register_gate(NULL, list->name,
556*4882a593Smuzhiyun list->parent_names[0], flags,
557*4882a593Smuzhiyun ctx->reg_base + list->gate_offset,
558*4882a593Smuzhiyun list->gate_shift, list->gate_flags, &ctx->lock);
559*4882a593Smuzhiyun break;
560*4882a593Smuzhiyun case branch_gate_no_set_rate:
561*4882a593Smuzhiyun flags &= ~CLK_SET_RATE_PARENT;
562*4882a593Smuzhiyun
563*4882a593Smuzhiyun clk = clk_register_gate(NULL, list->name,
564*4882a593Smuzhiyun list->parent_names[0], flags,
565*4882a593Smuzhiyun ctx->reg_base + list->gate_offset,
566*4882a593Smuzhiyun list->gate_shift, list->gate_flags, &ctx->lock);
567*4882a593Smuzhiyun break;
568*4882a593Smuzhiyun case branch_composite:
569*4882a593Smuzhiyun clk = rockchip_clk_register_branch(list->name,
570*4882a593Smuzhiyun list->parent_names, list->num_parents,
571*4882a593Smuzhiyun ctx->reg_base, list->muxdiv_offset,
572*4882a593Smuzhiyun list->mux_shift,
573*4882a593Smuzhiyun list->mux_width, list->mux_flags,
574*4882a593Smuzhiyun list->mux_table, list->div_offset,
575*4882a593Smuzhiyun list->div_shift, list->div_width,
576*4882a593Smuzhiyun list->div_flags, list->div_table,
577*4882a593Smuzhiyun list->gate_offset, list->gate_shift,
578*4882a593Smuzhiyun list->gate_flags, flags, &ctx->lock);
579*4882a593Smuzhiyun break;
580*4882a593Smuzhiyun case branch_mmc:
581*4882a593Smuzhiyun clk = rockchip_clk_register_mmc(
582*4882a593Smuzhiyun list->name,
583*4882a593Smuzhiyun list->parent_names, list->num_parents,
584*4882a593Smuzhiyun ctx->reg_base + list->muxdiv_offset,
585*4882a593Smuzhiyun list->div_shift
586*4882a593Smuzhiyun );
587*4882a593Smuzhiyun break;
588*4882a593Smuzhiyun case branch_inverter:
589*4882a593Smuzhiyun #ifdef CONFIG_ROCKCHIP_CLK_INV
590*4882a593Smuzhiyun clk = rockchip_clk_register_inverter(
591*4882a593Smuzhiyun list->name, list->parent_names,
592*4882a593Smuzhiyun list->num_parents,
593*4882a593Smuzhiyun ctx->reg_base + list->muxdiv_offset,
594*4882a593Smuzhiyun list->div_shift, list->div_flags, &ctx->lock);
595*4882a593Smuzhiyun #endif
596*4882a593Smuzhiyun break;
597*4882a593Smuzhiyun case branch_factor:
598*4882a593Smuzhiyun clk = rockchip_clk_register_factor_branch(
599*4882a593Smuzhiyun list->name, list->parent_names,
600*4882a593Smuzhiyun list->num_parents, ctx->reg_base,
601*4882a593Smuzhiyun list->div_shift, list->div_width,
602*4882a593Smuzhiyun list->gate_offset, list->gate_shift,
603*4882a593Smuzhiyun list->gate_flags, flags, &ctx->lock);
604*4882a593Smuzhiyun break;
605*4882a593Smuzhiyun case branch_ddrclk:
606*4882a593Smuzhiyun clk = rockchip_clk_register_ddrclk(
607*4882a593Smuzhiyun list->name, list->flags,
608*4882a593Smuzhiyun list->parent_names, list->num_parents,
609*4882a593Smuzhiyun list->muxdiv_offset, list->mux_shift,
610*4882a593Smuzhiyun list->mux_width, list->div_shift,
611*4882a593Smuzhiyun list->div_width, list->div_flags,
612*4882a593Smuzhiyun ctx->reg_base);
613*4882a593Smuzhiyun break;
614*4882a593Smuzhiyun }
615*4882a593Smuzhiyun
616*4882a593Smuzhiyun /* none of the cases above matched */
617*4882a593Smuzhiyun if (!clk) {
618*4882a593Smuzhiyun pr_err("%s: unknown clock type %d\n",
619*4882a593Smuzhiyun __func__, list->branch_type);
620*4882a593Smuzhiyun continue;
621*4882a593Smuzhiyun }
622*4882a593Smuzhiyun
623*4882a593Smuzhiyun if (IS_ERR(clk)) {
624*4882a593Smuzhiyun pr_err("%s: failed to register clock %s: %ld\n",
625*4882a593Smuzhiyun __func__, list->name, PTR_ERR(clk));
626*4882a593Smuzhiyun continue;
627*4882a593Smuzhiyun }
628*4882a593Smuzhiyun
629*4882a593Smuzhiyun rockchip_clk_add_lookup(ctx, clk, list->id);
630*4882a593Smuzhiyun }
631*4882a593Smuzhiyun }
632*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rockchip_clk_register_branches);
633*4882a593Smuzhiyun
rockchip_clk_register_armclk(struct rockchip_clk_provider * ctx,unsigned int lookup_id,const char * name,u8 num_parents,struct clk * parent,struct clk * alt_parent,const struct rockchip_cpuclk_reg_data * reg_data,const struct rockchip_cpuclk_rate_table * rates,int nrates)634*4882a593Smuzhiyun void rockchip_clk_register_armclk(struct rockchip_clk_provider *ctx,
635*4882a593Smuzhiyun unsigned int lookup_id,
636*4882a593Smuzhiyun const char *name,
637*4882a593Smuzhiyun u8 num_parents,
638*4882a593Smuzhiyun struct clk *parent, struct clk *alt_parent,
639*4882a593Smuzhiyun const struct rockchip_cpuclk_reg_data *reg_data,
640*4882a593Smuzhiyun const struct rockchip_cpuclk_rate_table *rates,
641*4882a593Smuzhiyun int nrates)
642*4882a593Smuzhiyun {
643*4882a593Smuzhiyun struct clk *clk;
644*4882a593Smuzhiyun
645*4882a593Smuzhiyun clk = rockchip_clk_register_cpuclk(name, num_parents,
646*4882a593Smuzhiyun parent, alt_parent,
647*4882a593Smuzhiyun reg_data, rates, nrates,
648*4882a593Smuzhiyun ctx->reg_base, &ctx->lock);
649*4882a593Smuzhiyun if (IS_ERR(clk)) {
650*4882a593Smuzhiyun pr_err("%s: failed to register clock %s: %ld\n",
651*4882a593Smuzhiyun __func__, name, PTR_ERR(clk));
652*4882a593Smuzhiyun return;
653*4882a593Smuzhiyun }
654*4882a593Smuzhiyun
655*4882a593Smuzhiyun rockchip_clk_add_lookup(ctx, clk, lookup_id);
656*4882a593Smuzhiyun }
657*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rockchip_clk_register_armclk);
658*4882a593Smuzhiyun
rockchip_clk_register_armclk_v2(struct rockchip_clk_provider * ctx,struct rockchip_clk_branch * list,const struct rockchip_cpuclk_rate_table * rates,int nrates)659*4882a593Smuzhiyun void rockchip_clk_register_armclk_v2(struct rockchip_clk_provider *ctx,
660*4882a593Smuzhiyun struct rockchip_clk_branch *list,
661*4882a593Smuzhiyun const struct rockchip_cpuclk_rate_table *rates,
662*4882a593Smuzhiyun int nrates)
663*4882a593Smuzhiyun {
664*4882a593Smuzhiyun struct clk *clk;
665*4882a593Smuzhiyun
666*4882a593Smuzhiyun clk = rockchip_clk_register_cpuclk_v2(list->name, list->parent_names,
667*4882a593Smuzhiyun list->num_parents, ctx->reg_base,
668*4882a593Smuzhiyun list->muxdiv_offset, list->mux_shift,
669*4882a593Smuzhiyun list->mux_width, list->mux_flags,
670*4882a593Smuzhiyun list->div_offset, list->div_shift,
671*4882a593Smuzhiyun list->div_width, list->div_flags,
672*4882a593Smuzhiyun list->flags, &ctx->lock, rates, nrates);
673*4882a593Smuzhiyun if (IS_ERR(clk)) {
674*4882a593Smuzhiyun pr_err("%s: failed to register clock %s: %ld\n",
675*4882a593Smuzhiyun __func__, list->name, PTR_ERR(clk));
676*4882a593Smuzhiyun return;
677*4882a593Smuzhiyun }
678*4882a593Smuzhiyun
679*4882a593Smuzhiyun rockchip_clk_add_lookup(ctx, clk, list->id);
680*4882a593Smuzhiyun }
681*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rockchip_clk_register_armclk_v2);
682*4882a593Smuzhiyun
683*4882a593Smuzhiyun void (*rk_dump_cru)(void);
684*4882a593Smuzhiyun EXPORT_SYMBOL(rk_dump_cru);
685*4882a593Smuzhiyun
rk_clk_panic(struct notifier_block * this,unsigned long ev,void * ptr)686*4882a593Smuzhiyun static int rk_clk_panic(struct notifier_block *this,
687*4882a593Smuzhiyun unsigned long ev, void *ptr)
688*4882a593Smuzhiyun {
689*4882a593Smuzhiyun if (rk_dump_cru)
690*4882a593Smuzhiyun rk_dump_cru();
691*4882a593Smuzhiyun return NOTIFY_DONE;
692*4882a593Smuzhiyun }
693*4882a593Smuzhiyun
694*4882a593Smuzhiyun static struct notifier_block rk_clk_panic_block = {
695*4882a593Smuzhiyun .notifier_call = rk_clk_panic,
696*4882a593Smuzhiyun };
697*4882a593Smuzhiyun
698*4882a593Smuzhiyun static void __iomem *rst_base;
699*4882a593Smuzhiyun static unsigned int reg_restart;
700*4882a593Smuzhiyun static void (*cb_restart)(void);
rockchip_restart_notify(struct notifier_block * this,unsigned long mode,void * cmd)701*4882a593Smuzhiyun static int rockchip_restart_notify(struct notifier_block *this,
702*4882a593Smuzhiyun unsigned long mode, void *cmd)
703*4882a593Smuzhiyun {
704*4882a593Smuzhiyun if (cb_restart)
705*4882a593Smuzhiyun cb_restart();
706*4882a593Smuzhiyun
707*4882a593Smuzhiyun writel(0xfdb9, rst_base + reg_restart);
708*4882a593Smuzhiyun return NOTIFY_DONE;
709*4882a593Smuzhiyun }
710*4882a593Smuzhiyun
711*4882a593Smuzhiyun static struct notifier_block rockchip_restart_handler = {
712*4882a593Smuzhiyun .notifier_call = rockchip_restart_notify,
713*4882a593Smuzhiyun .priority = 128,
714*4882a593Smuzhiyun };
715*4882a593Smuzhiyun
716*4882a593Smuzhiyun void
rockchip_register_restart_notifier(struct rockchip_clk_provider * ctx,unsigned int reg,void (* cb)(void))717*4882a593Smuzhiyun rockchip_register_restart_notifier(struct rockchip_clk_provider *ctx,
718*4882a593Smuzhiyun unsigned int reg,
719*4882a593Smuzhiyun void (*cb)(void))
720*4882a593Smuzhiyun {
721*4882a593Smuzhiyun int ret;
722*4882a593Smuzhiyun
723*4882a593Smuzhiyun rst_base = ctx->reg_base;
724*4882a593Smuzhiyun reg_restart = reg;
725*4882a593Smuzhiyun cb_restart = cb;
726*4882a593Smuzhiyun ret = register_restart_handler(&rockchip_restart_handler);
727*4882a593Smuzhiyun if (ret)
728*4882a593Smuzhiyun pr_err("%s: cannot register restart handler, %d\n",
729*4882a593Smuzhiyun __func__, ret);
730*4882a593Smuzhiyun atomic_notifier_chain_register(&panic_notifier_list,
731*4882a593Smuzhiyun &rk_clk_panic_block);
732*4882a593Smuzhiyun }
733*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rockchip_register_restart_notifier);
734*4882a593Smuzhiyun
735*4882a593Smuzhiyun #ifdef MODULE
736*4882a593Smuzhiyun static struct clk **protect_clocks;
737*4882a593Smuzhiyun static unsigned int protect_nclocks;
738*4882a593Smuzhiyun
rockchip_clk_protect(struct rockchip_clk_provider * ctx,unsigned int * clocks,unsigned int nclocks)739*4882a593Smuzhiyun int rockchip_clk_protect(struct rockchip_clk_provider *ctx,
740*4882a593Smuzhiyun unsigned int *clocks, unsigned int nclocks)
741*4882a593Smuzhiyun {
742*4882a593Smuzhiyun struct clk *clk = NULL;
743*4882a593Smuzhiyun int i = 0;
744*4882a593Smuzhiyun
745*4882a593Smuzhiyun if (protect_clocks || !ctx || !clocks || !ctx->clk_data.clks)
746*4882a593Smuzhiyun return 0;
747*4882a593Smuzhiyun
748*4882a593Smuzhiyun protect_clocks = kcalloc(nclocks, sizeof(void *), GFP_KERNEL);
749*4882a593Smuzhiyun if (!protect_clocks)
750*4882a593Smuzhiyun return -ENOMEM;
751*4882a593Smuzhiyun
752*4882a593Smuzhiyun for (i = 0; i < nclocks; i++) {
753*4882a593Smuzhiyun if (clocks[i] >= ctx->clk_data.clk_num) {
754*4882a593Smuzhiyun pr_err("%s: invalid clock id %u\n", __func__, clocks[i]);
755*4882a593Smuzhiyun continue;
756*4882a593Smuzhiyun }
757*4882a593Smuzhiyun clk = ctx->clk_data.clks[clocks[i]];
758*4882a593Smuzhiyun if (clk) {
759*4882a593Smuzhiyun clk_prepare_enable(clk);
760*4882a593Smuzhiyun protect_clocks[i] = clk;
761*4882a593Smuzhiyun }
762*4882a593Smuzhiyun }
763*4882a593Smuzhiyun protect_nclocks = nclocks;
764*4882a593Smuzhiyun
765*4882a593Smuzhiyun return 0;
766*4882a593Smuzhiyun }
767*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rockchip_clk_protect);
768*4882a593Smuzhiyun
rockchip_clk_unprotect(void)769*4882a593Smuzhiyun void rockchip_clk_unprotect(void)
770*4882a593Smuzhiyun {
771*4882a593Smuzhiyun int i = 0;
772*4882a593Smuzhiyun
773*4882a593Smuzhiyun if (!protect_clocks || !protect_nclocks)
774*4882a593Smuzhiyun return;
775*4882a593Smuzhiyun
776*4882a593Smuzhiyun for (i = 0; i < protect_nclocks; i++) {
777*4882a593Smuzhiyun if (protect_clocks[i])
778*4882a593Smuzhiyun clk_disable_unprepare(protect_clocks[i]);
779*4882a593Smuzhiyun }
780*4882a593Smuzhiyun protect_nclocks = 0;
781*4882a593Smuzhiyun kfree(protect_clocks);
782*4882a593Smuzhiyun protect_clocks = NULL;
783*4882a593Smuzhiyun
784*4882a593Smuzhiyun }
785*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rockchip_clk_unprotect);
786*4882a593Smuzhiyun #endif /* MODULE */
787