xref: /OK3568_Linux_fs/kernel/drivers/clk/x86/clk-cgu.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * Copyright (C) 2020 Intel Corporation.
4*4882a593Smuzhiyun  * Zhu YiXin <yixin.zhu@intel.com>
5*4882a593Smuzhiyun  * Rahul Tanwar <rahul.tanwar@intel.com>
6*4882a593Smuzhiyun  */
7*4882a593Smuzhiyun #include <linux/clk-provider.h>
8*4882a593Smuzhiyun #include <linux/device.h>
9*4882a593Smuzhiyun #include <linux/of.h>
10*4882a593Smuzhiyun 
11*4882a593Smuzhiyun #include "clk-cgu.h"
12*4882a593Smuzhiyun 
13*4882a593Smuzhiyun #define GATE_HW_REG_STAT(reg)	((reg) + 0x0)
14*4882a593Smuzhiyun #define GATE_HW_REG_EN(reg)	((reg) + 0x4)
15*4882a593Smuzhiyun #define GATE_HW_REG_DIS(reg)	((reg) + 0x8)
16*4882a593Smuzhiyun #define MAX_DDIV_REG	8
17*4882a593Smuzhiyun #define MAX_DIVIDER_VAL 64
18*4882a593Smuzhiyun 
19*4882a593Smuzhiyun #define to_lgm_clk_mux(_hw) container_of(_hw, struct lgm_clk_mux, hw)
20*4882a593Smuzhiyun #define to_lgm_clk_divider(_hw) container_of(_hw, struct lgm_clk_divider, hw)
21*4882a593Smuzhiyun #define to_lgm_clk_gate(_hw) container_of(_hw, struct lgm_clk_gate, hw)
22*4882a593Smuzhiyun #define to_lgm_clk_ddiv(_hw) container_of(_hw, struct lgm_clk_ddiv, hw)
23*4882a593Smuzhiyun 
lgm_clk_register_fixed(struct lgm_clk_provider * ctx,const struct lgm_clk_branch * list)24*4882a593Smuzhiyun static struct clk_hw *lgm_clk_register_fixed(struct lgm_clk_provider *ctx,
25*4882a593Smuzhiyun 					     const struct lgm_clk_branch *list)
26*4882a593Smuzhiyun {
27*4882a593Smuzhiyun 	unsigned long flags;
28*4882a593Smuzhiyun 
29*4882a593Smuzhiyun 	if (list->div_flags & CLOCK_FLAG_VAL_INIT) {
30*4882a593Smuzhiyun 		spin_lock_irqsave(&ctx->lock, flags);
31*4882a593Smuzhiyun 		lgm_set_clk_val(ctx->membase, list->div_off, list->div_shift,
32*4882a593Smuzhiyun 				list->div_width, list->div_val);
33*4882a593Smuzhiyun 		spin_unlock_irqrestore(&ctx->lock, flags);
34*4882a593Smuzhiyun 	}
35*4882a593Smuzhiyun 
36*4882a593Smuzhiyun 	return clk_hw_register_fixed_rate(NULL, list->name,
37*4882a593Smuzhiyun 					  list->parent_data[0].name,
38*4882a593Smuzhiyun 					  list->flags, list->mux_flags);
39*4882a593Smuzhiyun }
40*4882a593Smuzhiyun 
lgm_clk_mux_get_parent(struct clk_hw * hw)41*4882a593Smuzhiyun static u8 lgm_clk_mux_get_parent(struct clk_hw *hw)
42*4882a593Smuzhiyun {
43*4882a593Smuzhiyun 	struct lgm_clk_mux *mux = to_lgm_clk_mux(hw);
44*4882a593Smuzhiyun 	unsigned long flags;
45*4882a593Smuzhiyun 	u32 val;
46*4882a593Smuzhiyun 
47*4882a593Smuzhiyun 	spin_lock_irqsave(&mux->lock, flags);
48*4882a593Smuzhiyun 	if (mux->flags & MUX_CLK_SW)
49*4882a593Smuzhiyun 		val = mux->reg;
50*4882a593Smuzhiyun 	else
51*4882a593Smuzhiyun 		val = lgm_get_clk_val(mux->membase, mux->reg, mux->shift,
52*4882a593Smuzhiyun 				      mux->width);
53*4882a593Smuzhiyun 	spin_unlock_irqrestore(&mux->lock, flags);
54*4882a593Smuzhiyun 	return clk_mux_val_to_index(hw, NULL, mux->flags, val);
55*4882a593Smuzhiyun }
56*4882a593Smuzhiyun 
lgm_clk_mux_set_parent(struct clk_hw * hw,u8 index)57*4882a593Smuzhiyun static int lgm_clk_mux_set_parent(struct clk_hw *hw, u8 index)
58*4882a593Smuzhiyun {
59*4882a593Smuzhiyun 	struct lgm_clk_mux *mux = to_lgm_clk_mux(hw);
60*4882a593Smuzhiyun 	unsigned long flags;
61*4882a593Smuzhiyun 	u32 val;
62*4882a593Smuzhiyun 
63*4882a593Smuzhiyun 	val = clk_mux_index_to_val(NULL, mux->flags, index);
64*4882a593Smuzhiyun 	spin_lock_irqsave(&mux->lock, flags);
65*4882a593Smuzhiyun 	if (mux->flags & MUX_CLK_SW)
66*4882a593Smuzhiyun 		mux->reg = val;
67*4882a593Smuzhiyun 	else
68*4882a593Smuzhiyun 		lgm_set_clk_val(mux->membase, mux->reg, mux->shift,
69*4882a593Smuzhiyun 				mux->width, val);
70*4882a593Smuzhiyun 	spin_unlock_irqrestore(&mux->lock, flags);
71*4882a593Smuzhiyun 
72*4882a593Smuzhiyun 	return 0;
73*4882a593Smuzhiyun }
74*4882a593Smuzhiyun 
lgm_clk_mux_determine_rate(struct clk_hw * hw,struct clk_rate_request * req)75*4882a593Smuzhiyun static int lgm_clk_mux_determine_rate(struct clk_hw *hw,
76*4882a593Smuzhiyun 				      struct clk_rate_request *req)
77*4882a593Smuzhiyun {
78*4882a593Smuzhiyun 	struct lgm_clk_mux *mux = to_lgm_clk_mux(hw);
79*4882a593Smuzhiyun 
80*4882a593Smuzhiyun 	return clk_mux_determine_rate_flags(hw, req, mux->flags);
81*4882a593Smuzhiyun }
82*4882a593Smuzhiyun 
83*4882a593Smuzhiyun static const struct clk_ops lgm_clk_mux_ops = {
84*4882a593Smuzhiyun 	.get_parent = lgm_clk_mux_get_parent,
85*4882a593Smuzhiyun 	.set_parent = lgm_clk_mux_set_parent,
86*4882a593Smuzhiyun 	.determine_rate = lgm_clk_mux_determine_rate,
87*4882a593Smuzhiyun };
88*4882a593Smuzhiyun 
89*4882a593Smuzhiyun static struct clk_hw *
lgm_clk_register_mux(struct lgm_clk_provider * ctx,const struct lgm_clk_branch * list)90*4882a593Smuzhiyun lgm_clk_register_mux(struct lgm_clk_provider *ctx,
91*4882a593Smuzhiyun 		     const struct lgm_clk_branch *list)
92*4882a593Smuzhiyun {
93*4882a593Smuzhiyun 	unsigned long flags, cflags = list->mux_flags;
94*4882a593Smuzhiyun 	struct device *dev = ctx->dev;
95*4882a593Smuzhiyun 	u8 shift = list->mux_shift;
96*4882a593Smuzhiyun 	u8 width = list->mux_width;
97*4882a593Smuzhiyun 	struct clk_init_data init = {};
98*4882a593Smuzhiyun 	struct lgm_clk_mux *mux;
99*4882a593Smuzhiyun 	u32 reg = list->mux_off;
100*4882a593Smuzhiyun 	struct clk_hw *hw;
101*4882a593Smuzhiyun 	int ret;
102*4882a593Smuzhiyun 
103*4882a593Smuzhiyun 	mux = devm_kzalloc(dev, sizeof(*mux), GFP_KERNEL);
104*4882a593Smuzhiyun 	if (!mux)
105*4882a593Smuzhiyun 		return ERR_PTR(-ENOMEM);
106*4882a593Smuzhiyun 
107*4882a593Smuzhiyun 	init.name = list->name;
108*4882a593Smuzhiyun 	init.ops = &lgm_clk_mux_ops;
109*4882a593Smuzhiyun 	init.flags = list->flags;
110*4882a593Smuzhiyun 	init.parent_data = list->parent_data;
111*4882a593Smuzhiyun 	init.num_parents = list->num_parents;
112*4882a593Smuzhiyun 
113*4882a593Smuzhiyun 	mux->membase = ctx->membase;
114*4882a593Smuzhiyun 	mux->lock = ctx->lock;
115*4882a593Smuzhiyun 	mux->reg = reg;
116*4882a593Smuzhiyun 	mux->shift = shift;
117*4882a593Smuzhiyun 	mux->width = width;
118*4882a593Smuzhiyun 	mux->flags = cflags;
119*4882a593Smuzhiyun 	mux->hw.init = &init;
120*4882a593Smuzhiyun 
121*4882a593Smuzhiyun 	hw = &mux->hw;
122*4882a593Smuzhiyun 	ret = devm_clk_hw_register(dev, hw);
123*4882a593Smuzhiyun 	if (ret)
124*4882a593Smuzhiyun 		return ERR_PTR(ret);
125*4882a593Smuzhiyun 
126*4882a593Smuzhiyun 	if (cflags & CLOCK_FLAG_VAL_INIT) {
127*4882a593Smuzhiyun 		spin_lock_irqsave(&mux->lock, flags);
128*4882a593Smuzhiyun 		lgm_set_clk_val(mux->membase, reg, shift, width, list->mux_val);
129*4882a593Smuzhiyun 		spin_unlock_irqrestore(&mux->lock, flags);
130*4882a593Smuzhiyun 	}
131*4882a593Smuzhiyun 
132*4882a593Smuzhiyun 	return hw;
133*4882a593Smuzhiyun }
134*4882a593Smuzhiyun 
135*4882a593Smuzhiyun static unsigned long
lgm_clk_divider_recalc_rate(struct clk_hw * hw,unsigned long parent_rate)136*4882a593Smuzhiyun lgm_clk_divider_recalc_rate(struct clk_hw *hw, unsigned long parent_rate)
137*4882a593Smuzhiyun {
138*4882a593Smuzhiyun 	struct lgm_clk_divider *divider = to_lgm_clk_divider(hw);
139*4882a593Smuzhiyun 	unsigned long flags;
140*4882a593Smuzhiyun 	unsigned int val;
141*4882a593Smuzhiyun 
142*4882a593Smuzhiyun 	spin_lock_irqsave(&divider->lock, flags);
143*4882a593Smuzhiyun 	val = lgm_get_clk_val(divider->membase, divider->reg,
144*4882a593Smuzhiyun 			      divider->shift, divider->width);
145*4882a593Smuzhiyun 	spin_unlock_irqrestore(&divider->lock, flags);
146*4882a593Smuzhiyun 
147*4882a593Smuzhiyun 	return divider_recalc_rate(hw, parent_rate, val, divider->table,
148*4882a593Smuzhiyun 				   divider->flags, divider->width);
149*4882a593Smuzhiyun }
150*4882a593Smuzhiyun 
151*4882a593Smuzhiyun static long
lgm_clk_divider_round_rate(struct clk_hw * hw,unsigned long rate,unsigned long * prate)152*4882a593Smuzhiyun lgm_clk_divider_round_rate(struct clk_hw *hw, unsigned long rate,
153*4882a593Smuzhiyun 			   unsigned long *prate)
154*4882a593Smuzhiyun {
155*4882a593Smuzhiyun 	struct lgm_clk_divider *divider = to_lgm_clk_divider(hw);
156*4882a593Smuzhiyun 
157*4882a593Smuzhiyun 	return divider_round_rate(hw, rate, prate, divider->table,
158*4882a593Smuzhiyun 				  divider->width, divider->flags);
159*4882a593Smuzhiyun }
160*4882a593Smuzhiyun 
161*4882a593Smuzhiyun static int
lgm_clk_divider_set_rate(struct clk_hw * hw,unsigned long rate,unsigned long prate)162*4882a593Smuzhiyun lgm_clk_divider_set_rate(struct clk_hw *hw, unsigned long rate,
163*4882a593Smuzhiyun 			 unsigned long prate)
164*4882a593Smuzhiyun {
165*4882a593Smuzhiyun 	struct lgm_clk_divider *divider = to_lgm_clk_divider(hw);
166*4882a593Smuzhiyun 	unsigned long flags;
167*4882a593Smuzhiyun 	int value;
168*4882a593Smuzhiyun 
169*4882a593Smuzhiyun 	value = divider_get_val(rate, prate, divider->table,
170*4882a593Smuzhiyun 				divider->width, divider->flags);
171*4882a593Smuzhiyun 	if (value < 0)
172*4882a593Smuzhiyun 		return value;
173*4882a593Smuzhiyun 
174*4882a593Smuzhiyun 	spin_lock_irqsave(&divider->lock, flags);
175*4882a593Smuzhiyun 	lgm_set_clk_val(divider->membase, divider->reg,
176*4882a593Smuzhiyun 			divider->shift, divider->width, value);
177*4882a593Smuzhiyun 	spin_unlock_irqrestore(&divider->lock, flags);
178*4882a593Smuzhiyun 
179*4882a593Smuzhiyun 	return 0;
180*4882a593Smuzhiyun }
181*4882a593Smuzhiyun 
lgm_clk_divider_enable_disable(struct clk_hw * hw,int enable)182*4882a593Smuzhiyun static int lgm_clk_divider_enable_disable(struct clk_hw *hw, int enable)
183*4882a593Smuzhiyun {
184*4882a593Smuzhiyun 	struct lgm_clk_divider *div = to_lgm_clk_divider(hw);
185*4882a593Smuzhiyun 	unsigned long flags;
186*4882a593Smuzhiyun 
187*4882a593Smuzhiyun 	spin_lock_irqsave(&div->lock, flags);
188*4882a593Smuzhiyun 	lgm_set_clk_val(div->membase, div->reg, div->shift_gate,
189*4882a593Smuzhiyun 			div->width_gate, enable);
190*4882a593Smuzhiyun 	spin_unlock_irqrestore(&div->lock, flags);
191*4882a593Smuzhiyun 	return 0;
192*4882a593Smuzhiyun }
193*4882a593Smuzhiyun 
lgm_clk_divider_enable(struct clk_hw * hw)194*4882a593Smuzhiyun static int lgm_clk_divider_enable(struct clk_hw *hw)
195*4882a593Smuzhiyun {
196*4882a593Smuzhiyun 	return lgm_clk_divider_enable_disable(hw, 1);
197*4882a593Smuzhiyun }
198*4882a593Smuzhiyun 
lgm_clk_divider_disable(struct clk_hw * hw)199*4882a593Smuzhiyun static void lgm_clk_divider_disable(struct clk_hw *hw)
200*4882a593Smuzhiyun {
201*4882a593Smuzhiyun 	lgm_clk_divider_enable_disable(hw, 0);
202*4882a593Smuzhiyun }
203*4882a593Smuzhiyun 
204*4882a593Smuzhiyun static const struct clk_ops lgm_clk_divider_ops = {
205*4882a593Smuzhiyun 	.recalc_rate = lgm_clk_divider_recalc_rate,
206*4882a593Smuzhiyun 	.round_rate = lgm_clk_divider_round_rate,
207*4882a593Smuzhiyun 	.set_rate = lgm_clk_divider_set_rate,
208*4882a593Smuzhiyun 	.enable = lgm_clk_divider_enable,
209*4882a593Smuzhiyun 	.disable = lgm_clk_divider_disable,
210*4882a593Smuzhiyun };
211*4882a593Smuzhiyun 
212*4882a593Smuzhiyun static struct clk_hw *
lgm_clk_register_divider(struct lgm_clk_provider * ctx,const struct lgm_clk_branch * list)213*4882a593Smuzhiyun lgm_clk_register_divider(struct lgm_clk_provider *ctx,
214*4882a593Smuzhiyun 			 const struct lgm_clk_branch *list)
215*4882a593Smuzhiyun {
216*4882a593Smuzhiyun 	unsigned long flags, cflags = list->div_flags;
217*4882a593Smuzhiyun 	struct device *dev = ctx->dev;
218*4882a593Smuzhiyun 	struct lgm_clk_divider *div;
219*4882a593Smuzhiyun 	struct clk_init_data init = {};
220*4882a593Smuzhiyun 	u8 shift = list->div_shift;
221*4882a593Smuzhiyun 	u8 width = list->div_width;
222*4882a593Smuzhiyun 	u8 shift_gate = list->div_shift_gate;
223*4882a593Smuzhiyun 	u8 width_gate = list->div_width_gate;
224*4882a593Smuzhiyun 	u32 reg = list->div_off;
225*4882a593Smuzhiyun 	struct clk_hw *hw;
226*4882a593Smuzhiyun 	int ret;
227*4882a593Smuzhiyun 
228*4882a593Smuzhiyun 	div = devm_kzalloc(dev, sizeof(*div), GFP_KERNEL);
229*4882a593Smuzhiyun 	if (!div)
230*4882a593Smuzhiyun 		return ERR_PTR(-ENOMEM);
231*4882a593Smuzhiyun 
232*4882a593Smuzhiyun 	init.name = list->name;
233*4882a593Smuzhiyun 	init.ops = &lgm_clk_divider_ops;
234*4882a593Smuzhiyun 	init.flags = list->flags;
235*4882a593Smuzhiyun 	init.parent_data = list->parent_data;
236*4882a593Smuzhiyun 	init.num_parents = 1;
237*4882a593Smuzhiyun 
238*4882a593Smuzhiyun 	div->membase = ctx->membase;
239*4882a593Smuzhiyun 	div->lock = ctx->lock;
240*4882a593Smuzhiyun 	div->reg = reg;
241*4882a593Smuzhiyun 	div->shift = shift;
242*4882a593Smuzhiyun 	div->width = width;
243*4882a593Smuzhiyun 	div->shift_gate	= shift_gate;
244*4882a593Smuzhiyun 	div->width_gate	= width_gate;
245*4882a593Smuzhiyun 	div->flags = cflags;
246*4882a593Smuzhiyun 	div->table = list->div_table;
247*4882a593Smuzhiyun 	div->hw.init = &init;
248*4882a593Smuzhiyun 
249*4882a593Smuzhiyun 	hw = &div->hw;
250*4882a593Smuzhiyun 	ret = devm_clk_hw_register(dev, hw);
251*4882a593Smuzhiyun 	if (ret)
252*4882a593Smuzhiyun 		return ERR_PTR(ret);
253*4882a593Smuzhiyun 
254*4882a593Smuzhiyun 	if (cflags & CLOCK_FLAG_VAL_INIT) {
255*4882a593Smuzhiyun 		spin_lock_irqsave(&div->lock, flags);
256*4882a593Smuzhiyun 		lgm_set_clk_val(div->membase, reg, shift, width, list->div_val);
257*4882a593Smuzhiyun 		spin_unlock_irqrestore(&div->lock, flags);
258*4882a593Smuzhiyun 	}
259*4882a593Smuzhiyun 
260*4882a593Smuzhiyun 	return hw;
261*4882a593Smuzhiyun }
262*4882a593Smuzhiyun 
263*4882a593Smuzhiyun static struct clk_hw *
lgm_clk_register_fixed_factor(struct lgm_clk_provider * ctx,const struct lgm_clk_branch * list)264*4882a593Smuzhiyun lgm_clk_register_fixed_factor(struct lgm_clk_provider *ctx,
265*4882a593Smuzhiyun 			      const struct lgm_clk_branch *list)
266*4882a593Smuzhiyun {
267*4882a593Smuzhiyun 	unsigned long flags;
268*4882a593Smuzhiyun 	struct clk_hw *hw;
269*4882a593Smuzhiyun 
270*4882a593Smuzhiyun 	hw = clk_hw_register_fixed_factor(ctx->dev, list->name,
271*4882a593Smuzhiyun 					  list->parent_data[0].name, list->flags,
272*4882a593Smuzhiyun 					  list->mult, list->div);
273*4882a593Smuzhiyun 	if (IS_ERR(hw))
274*4882a593Smuzhiyun 		return ERR_CAST(hw);
275*4882a593Smuzhiyun 
276*4882a593Smuzhiyun 	if (list->div_flags & CLOCK_FLAG_VAL_INIT) {
277*4882a593Smuzhiyun 		spin_lock_irqsave(&ctx->lock, flags);
278*4882a593Smuzhiyun 		lgm_set_clk_val(ctx->membase, list->div_off, list->div_shift,
279*4882a593Smuzhiyun 				list->div_width, list->div_val);
280*4882a593Smuzhiyun 		spin_unlock_irqrestore(&ctx->lock, flags);
281*4882a593Smuzhiyun 	}
282*4882a593Smuzhiyun 
283*4882a593Smuzhiyun 	return hw;
284*4882a593Smuzhiyun }
285*4882a593Smuzhiyun 
lgm_clk_gate_enable(struct clk_hw * hw)286*4882a593Smuzhiyun static int lgm_clk_gate_enable(struct clk_hw *hw)
287*4882a593Smuzhiyun {
288*4882a593Smuzhiyun 	struct lgm_clk_gate *gate = to_lgm_clk_gate(hw);
289*4882a593Smuzhiyun 	unsigned long flags;
290*4882a593Smuzhiyun 	unsigned int reg;
291*4882a593Smuzhiyun 
292*4882a593Smuzhiyun 	spin_lock_irqsave(&gate->lock, flags);
293*4882a593Smuzhiyun 	reg = GATE_HW_REG_EN(gate->reg);
294*4882a593Smuzhiyun 	lgm_set_clk_val(gate->membase, reg, gate->shift, 1, 1);
295*4882a593Smuzhiyun 	spin_unlock_irqrestore(&gate->lock, flags);
296*4882a593Smuzhiyun 
297*4882a593Smuzhiyun 	return 0;
298*4882a593Smuzhiyun }
299*4882a593Smuzhiyun 
lgm_clk_gate_disable(struct clk_hw * hw)300*4882a593Smuzhiyun static void lgm_clk_gate_disable(struct clk_hw *hw)
301*4882a593Smuzhiyun {
302*4882a593Smuzhiyun 	struct lgm_clk_gate *gate = to_lgm_clk_gate(hw);
303*4882a593Smuzhiyun 	unsigned long flags;
304*4882a593Smuzhiyun 	unsigned int reg;
305*4882a593Smuzhiyun 
306*4882a593Smuzhiyun 	spin_lock_irqsave(&gate->lock, flags);
307*4882a593Smuzhiyun 	reg = GATE_HW_REG_DIS(gate->reg);
308*4882a593Smuzhiyun 	lgm_set_clk_val(gate->membase, reg, gate->shift, 1, 1);
309*4882a593Smuzhiyun 	spin_unlock_irqrestore(&gate->lock, flags);
310*4882a593Smuzhiyun }
311*4882a593Smuzhiyun 
lgm_clk_gate_is_enabled(struct clk_hw * hw)312*4882a593Smuzhiyun static int lgm_clk_gate_is_enabled(struct clk_hw *hw)
313*4882a593Smuzhiyun {
314*4882a593Smuzhiyun 	struct lgm_clk_gate *gate = to_lgm_clk_gate(hw);
315*4882a593Smuzhiyun 	unsigned int reg, ret;
316*4882a593Smuzhiyun 	unsigned long flags;
317*4882a593Smuzhiyun 
318*4882a593Smuzhiyun 	spin_lock_irqsave(&gate->lock, flags);
319*4882a593Smuzhiyun 	reg = GATE_HW_REG_STAT(gate->reg);
320*4882a593Smuzhiyun 	ret = lgm_get_clk_val(gate->membase, reg, gate->shift, 1);
321*4882a593Smuzhiyun 	spin_unlock_irqrestore(&gate->lock, flags);
322*4882a593Smuzhiyun 
323*4882a593Smuzhiyun 	return ret;
324*4882a593Smuzhiyun }
325*4882a593Smuzhiyun 
326*4882a593Smuzhiyun static const struct clk_ops lgm_clk_gate_ops = {
327*4882a593Smuzhiyun 	.enable = lgm_clk_gate_enable,
328*4882a593Smuzhiyun 	.disable = lgm_clk_gate_disable,
329*4882a593Smuzhiyun 	.is_enabled = lgm_clk_gate_is_enabled,
330*4882a593Smuzhiyun };
331*4882a593Smuzhiyun 
332*4882a593Smuzhiyun static struct clk_hw *
lgm_clk_register_gate(struct lgm_clk_provider * ctx,const struct lgm_clk_branch * list)333*4882a593Smuzhiyun lgm_clk_register_gate(struct lgm_clk_provider *ctx,
334*4882a593Smuzhiyun 		      const struct lgm_clk_branch *list)
335*4882a593Smuzhiyun {
336*4882a593Smuzhiyun 	unsigned long flags, cflags = list->gate_flags;
337*4882a593Smuzhiyun 	const char *pname = list->parent_data[0].name;
338*4882a593Smuzhiyun 	struct device *dev = ctx->dev;
339*4882a593Smuzhiyun 	u8 shift = list->gate_shift;
340*4882a593Smuzhiyun 	struct clk_init_data init = {};
341*4882a593Smuzhiyun 	struct lgm_clk_gate *gate;
342*4882a593Smuzhiyun 	u32 reg = list->gate_off;
343*4882a593Smuzhiyun 	struct clk_hw *hw;
344*4882a593Smuzhiyun 	int ret;
345*4882a593Smuzhiyun 
346*4882a593Smuzhiyun 	gate = devm_kzalloc(dev, sizeof(*gate), GFP_KERNEL);
347*4882a593Smuzhiyun 	if (!gate)
348*4882a593Smuzhiyun 		return ERR_PTR(-ENOMEM);
349*4882a593Smuzhiyun 
350*4882a593Smuzhiyun 	init.name = list->name;
351*4882a593Smuzhiyun 	init.ops = &lgm_clk_gate_ops;
352*4882a593Smuzhiyun 	init.flags = list->flags;
353*4882a593Smuzhiyun 	init.parent_names = pname ? &pname : NULL;
354*4882a593Smuzhiyun 	init.num_parents = pname ? 1 : 0;
355*4882a593Smuzhiyun 
356*4882a593Smuzhiyun 	gate->membase = ctx->membase;
357*4882a593Smuzhiyun 	gate->lock = ctx->lock;
358*4882a593Smuzhiyun 	gate->reg = reg;
359*4882a593Smuzhiyun 	gate->shift = shift;
360*4882a593Smuzhiyun 	gate->flags = cflags;
361*4882a593Smuzhiyun 	gate->hw.init = &init;
362*4882a593Smuzhiyun 
363*4882a593Smuzhiyun 	hw = &gate->hw;
364*4882a593Smuzhiyun 	ret = devm_clk_hw_register(dev, hw);
365*4882a593Smuzhiyun 	if (ret)
366*4882a593Smuzhiyun 		return ERR_PTR(ret);
367*4882a593Smuzhiyun 
368*4882a593Smuzhiyun 	if (cflags & CLOCK_FLAG_VAL_INIT) {
369*4882a593Smuzhiyun 		spin_lock_irqsave(&gate->lock, flags);
370*4882a593Smuzhiyun 		lgm_set_clk_val(gate->membase, reg, shift, 1, list->gate_val);
371*4882a593Smuzhiyun 		spin_unlock_irqrestore(&gate->lock, flags);
372*4882a593Smuzhiyun 	}
373*4882a593Smuzhiyun 
374*4882a593Smuzhiyun 	return hw;
375*4882a593Smuzhiyun }
376*4882a593Smuzhiyun 
lgm_clk_register_branches(struct lgm_clk_provider * ctx,const struct lgm_clk_branch * list,unsigned int nr_clk)377*4882a593Smuzhiyun int lgm_clk_register_branches(struct lgm_clk_provider *ctx,
378*4882a593Smuzhiyun 			      const struct lgm_clk_branch *list,
379*4882a593Smuzhiyun 			      unsigned int nr_clk)
380*4882a593Smuzhiyun {
381*4882a593Smuzhiyun 	struct clk_hw *hw;
382*4882a593Smuzhiyun 	unsigned int idx;
383*4882a593Smuzhiyun 
384*4882a593Smuzhiyun 	for (idx = 0; idx < nr_clk; idx++, list++) {
385*4882a593Smuzhiyun 		switch (list->type) {
386*4882a593Smuzhiyun 		case CLK_TYPE_FIXED:
387*4882a593Smuzhiyun 			hw = lgm_clk_register_fixed(ctx, list);
388*4882a593Smuzhiyun 			break;
389*4882a593Smuzhiyun 		case CLK_TYPE_MUX:
390*4882a593Smuzhiyun 			hw = lgm_clk_register_mux(ctx, list);
391*4882a593Smuzhiyun 			break;
392*4882a593Smuzhiyun 		case CLK_TYPE_DIVIDER:
393*4882a593Smuzhiyun 			hw = lgm_clk_register_divider(ctx, list);
394*4882a593Smuzhiyun 			break;
395*4882a593Smuzhiyun 		case CLK_TYPE_FIXED_FACTOR:
396*4882a593Smuzhiyun 			hw = lgm_clk_register_fixed_factor(ctx, list);
397*4882a593Smuzhiyun 			break;
398*4882a593Smuzhiyun 		case CLK_TYPE_GATE:
399*4882a593Smuzhiyun 			hw = lgm_clk_register_gate(ctx, list);
400*4882a593Smuzhiyun 			break;
401*4882a593Smuzhiyun 		default:
402*4882a593Smuzhiyun 			dev_err(ctx->dev, "invalid clk type\n");
403*4882a593Smuzhiyun 			return -EINVAL;
404*4882a593Smuzhiyun 		}
405*4882a593Smuzhiyun 
406*4882a593Smuzhiyun 		if (IS_ERR(hw)) {
407*4882a593Smuzhiyun 			dev_err(ctx->dev,
408*4882a593Smuzhiyun 				"register clk: %s, type: %u failed!\n",
409*4882a593Smuzhiyun 				list->name, list->type);
410*4882a593Smuzhiyun 			return -EIO;
411*4882a593Smuzhiyun 		}
412*4882a593Smuzhiyun 		ctx->clk_data.hws[list->id] = hw;
413*4882a593Smuzhiyun 	}
414*4882a593Smuzhiyun 
415*4882a593Smuzhiyun 	return 0;
416*4882a593Smuzhiyun }
417*4882a593Smuzhiyun 
418*4882a593Smuzhiyun static unsigned long
lgm_clk_ddiv_recalc_rate(struct clk_hw * hw,unsigned long parent_rate)419*4882a593Smuzhiyun lgm_clk_ddiv_recalc_rate(struct clk_hw *hw, unsigned long parent_rate)
420*4882a593Smuzhiyun {
421*4882a593Smuzhiyun 	struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
422*4882a593Smuzhiyun 	unsigned int div0, div1, exdiv;
423*4882a593Smuzhiyun 	u64 prate;
424*4882a593Smuzhiyun 
425*4882a593Smuzhiyun 	div0 = lgm_get_clk_val(ddiv->membase, ddiv->reg,
426*4882a593Smuzhiyun 			       ddiv->shift0, ddiv->width0) + 1;
427*4882a593Smuzhiyun 	div1 = lgm_get_clk_val(ddiv->membase, ddiv->reg,
428*4882a593Smuzhiyun 			       ddiv->shift1, ddiv->width1) + 1;
429*4882a593Smuzhiyun 	exdiv = lgm_get_clk_val(ddiv->membase, ddiv->reg,
430*4882a593Smuzhiyun 				ddiv->shift2, ddiv->width2);
431*4882a593Smuzhiyun 	prate = (u64)parent_rate;
432*4882a593Smuzhiyun 	do_div(prate, div0);
433*4882a593Smuzhiyun 	do_div(prate, div1);
434*4882a593Smuzhiyun 
435*4882a593Smuzhiyun 	if (exdiv) {
436*4882a593Smuzhiyun 		do_div(prate, ddiv->div);
437*4882a593Smuzhiyun 		prate *= ddiv->mult;
438*4882a593Smuzhiyun 	}
439*4882a593Smuzhiyun 
440*4882a593Smuzhiyun 	return prate;
441*4882a593Smuzhiyun }
442*4882a593Smuzhiyun 
lgm_clk_ddiv_enable(struct clk_hw * hw)443*4882a593Smuzhiyun static int lgm_clk_ddiv_enable(struct clk_hw *hw)
444*4882a593Smuzhiyun {
445*4882a593Smuzhiyun 	struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
446*4882a593Smuzhiyun 	unsigned long flags;
447*4882a593Smuzhiyun 
448*4882a593Smuzhiyun 	spin_lock_irqsave(&ddiv->lock, flags);
449*4882a593Smuzhiyun 	lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift_gate,
450*4882a593Smuzhiyun 			ddiv->width_gate, 1);
451*4882a593Smuzhiyun 	spin_unlock_irqrestore(&ddiv->lock, flags);
452*4882a593Smuzhiyun 	return 0;
453*4882a593Smuzhiyun }
454*4882a593Smuzhiyun 
lgm_clk_ddiv_disable(struct clk_hw * hw)455*4882a593Smuzhiyun static void lgm_clk_ddiv_disable(struct clk_hw *hw)
456*4882a593Smuzhiyun {
457*4882a593Smuzhiyun 	struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
458*4882a593Smuzhiyun 	unsigned long flags;
459*4882a593Smuzhiyun 
460*4882a593Smuzhiyun 	spin_lock_irqsave(&ddiv->lock, flags);
461*4882a593Smuzhiyun 	lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift_gate,
462*4882a593Smuzhiyun 			ddiv->width_gate, 0);
463*4882a593Smuzhiyun 	spin_unlock_irqrestore(&ddiv->lock, flags);
464*4882a593Smuzhiyun }
465*4882a593Smuzhiyun 
466*4882a593Smuzhiyun static int
lgm_clk_get_ddiv_val(u32 div,u32 * ddiv1,u32 * ddiv2)467*4882a593Smuzhiyun lgm_clk_get_ddiv_val(u32 div, u32 *ddiv1, u32 *ddiv2)
468*4882a593Smuzhiyun {
469*4882a593Smuzhiyun 	u32 idx, temp;
470*4882a593Smuzhiyun 
471*4882a593Smuzhiyun 	*ddiv1 = 1;
472*4882a593Smuzhiyun 	*ddiv2 = 1;
473*4882a593Smuzhiyun 
474*4882a593Smuzhiyun 	if (div > MAX_DIVIDER_VAL)
475*4882a593Smuzhiyun 		div = MAX_DIVIDER_VAL;
476*4882a593Smuzhiyun 
477*4882a593Smuzhiyun 	if (div > 1) {
478*4882a593Smuzhiyun 		for (idx = 2; idx <= MAX_DDIV_REG; idx++) {
479*4882a593Smuzhiyun 			temp = DIV_ROUND_UP_ULL((u64)div, idx);
480*4882a593Smuzhiyun 			if (div % idx == 0 && temp <= MAX_DDIV_REG)
481*4882a593Smuzhiyun 				break;
482*4882a593Smuzhiyun 		}
483*4882a593Smuzhiyun 
484*4882a593Smuzhiyun 		if (idx > MAX_DDIV_REG)
485*4882a593Smuzhiyun 			return -EINVAL;
486*4882a593Smuzhiyun 
487*4882a593Smuzhiyun 		*ddiv1 = temp;
488*4882a593Smuzhiyun 		*ddiv2 = idx;
489*4882a593Smuzhiyun 	}
490*4882a593Smuzhiyun 
491*4882a593Smuzhiyun 	return 0;
492*4882a593Smuzhiyun }
493*4882a593Smuzhiyun 
494*4882a593Smuzhiyun static int
lgm_clk_ddiv_set_rate(struct clk_hw * hw,unsigned long rate,unsigned long prate)495*4882a593Smuzhiyun lgm_clk_ddiv_set_rate(struct clk_hw *hw, unsigned long rate,
496*4882a593Smuzhiyun 		      unsigned long prate)
497*4882a593Smuzhiyun {
498*4882a593Smuzhiyun 	struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
499*4882a593Smuzhiyun 	u32 div, ddiv1, ddiv2;
500*4882a593Smuzhiyun 	unsigned long flags;
501*4882a593Smuzhiyun 
502*4882a593Smuzhiyun 	div = DIV_ROUND_CLOSEST_ULL((u64)prate, rate);
503*4882a593Smuzhiyun 
504*4882a593Smuzhiyun 	spin_lock_irqsave(&ddiv->lock, flags);
505*4882a593Smuzhiyun 	if (lgm_get_clk_val(ddiv->membase, ddiv->reg, ddiv->shift2, 1)) {
506*4882a593Smuzhiyun 		div = DIV_ROUND_CLOSEST_ULL((u64)div, 5);
507*4882a593Smuzhiyun 		div = div * 2;
508*4882a593Smuzhiyun 	}
509*4882a593Smuzhiyun 
510*4882a593Smuzhiyun 	if (div <= 0) {
511*4882a593Smuzhiyun 		spin_unlock_irqrestore(&ddiv->lock, flags);
512*4882a593Smuzhiyun 		return -EINVAL;
513*4882a593Smuzhiyun 	}
514*4882a593Smuzhiyun 
515*4882a593Smuzhiyun 	if (lgm_clk_get_ddiv_val(div, &ddiv1, &ddiv2)) {
516*4882a593Smuzhiyun 		spin_unlock_irqrestore(&ddiv->lock, flags);
517*4882a593Smuzhiyun 		return -EINVAL;
518*4882a593Smuzhiyun 	}
519*4882a593Smuzhiyun 
520*4882a593Smuzhiyun 	lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift0, ddiv->width0,
521*4882a593Smuzhiyun 			ddiv1 - 1);
522*4882a593Smuzhiyun 
523*4882a593Smuzhiyun 	lgm_set_clk_val(ddiv->membase, ddiv->reg,  ddiv->shift1, ddiv->width1,
524*4882a593Smuzhiyun 			ddiv2 - 1);
525*4882a593Smuzhiyun 	spin_unlock_irqrestore(&ddiv->lock, flags);
526*4882a593Smuzhiyun 
527*4882a593Smuzhiyun 	return 0;
528*4882a593Smuzhiyun }
529*4882a593Smuzhiyun 
530*4882a593Smuzhiyun static long
lgm_clk_ddiv_round_rate(struct clk_hw * hw,unsigned long rate,unsigned long * prate)531*4882a593Smuzhiyun lgm_clk_ddiv_round_rate(struct clk_hw *hw, unsigned long rate,
532*4882a593Smuzhiyun 			unsigned long *prate)
533*4882a593Smuzhiyun {
534*4882a593Smuzhiyun 	struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
535*4882a593Smuzhiyun 	u32 div, ddiv1, ddiv2;
536*4882a593Smuzhiyun 	unsigned long flags;
537*4882a593Smuzhiyun 	u64 rate64;
538*4882a593Smuzhiyun 
539*4882a593Smuzhiyun 	div = DIV_ROUND_CLOSEST_ULL((u64)*prate, rate);
540*4882a593Smuzhiyun 
541*4882a593Smuzhiyun 	/* if predivide bit is enabled, modify div by factor of 2.5 */
542*4882a593Smuzhiyun 	spin_lock_irqsave(&ddiv->lock, flags);
543*4882a593Smuzhiyun 	if (lgm_get_clk_val(ddiv->membase, ddiv->reg, ddiv->shift2, 1)) {
544*4882a593Smuzhiyun 		div = div * 2;
545*4882a593Smuzhiyun 		div = DIV_ROUND_CLOSEST_ULL((u64)div, 5);
546*4882a593Smuzhiyun 	}
547*4882a593Smuzhiyun 	spin_unlock_irqrestore(&ddiv->lock, flags);
548*4882a593Smuzhiyun 
549*4882a593Smuzhiyun 	if (div <= 0)
550*4882a593Smuzhiyun 		return *prate;
551*4882a593Smuzhiyun 
552*4882a593Smuzhiyun 	if (lgm_clk_get_ddiv_val(div, &ddiv1, &ddiv2) != 0)
553*4882a593Smuzhiyun 		if (lgm_clk_get_ddiv_val(div + 1, &ddiv1, &ddiv2) != 0)
554*4882a593Smuzhiyun 			return -EINVAL;
555*4882a593Smuzhiyun 
556*4882a593Smuzhiyun 	rate64 = *prate;
557*4882a593Smuzhiyun 	do_div(rate64, ddiv1);
558*4882a593Smuzhiyun 	do_div(rate64, ddiv2);
559*4882a593Smuzhiyun 
560*4882a593Smuzhiyun 	/* if predivide bit is enabled, modify rounded rate by factor of 2.5 */
561*4882a593Smuzhiyun 	spin_lock_irqsave(&ddiv->lock, flags);
562*4882a593Smuzhiyun 	if (lgm_get_clk_val(ddiv->membase, ddiv->reg, ddiv->shift2, 1)) {
563*4882a593Smuzhiyun 		rate64 = rate64 * 2;
564*4882a593Smuzhiyun 		rate64 = DIV_ROUND_CLOSEST_ULL(rate64, 5);
565*4882a593Smuzhiyun 	}
566*4882a593Smuzhiyun 	spin_unlock_irqrestore(&ddiv->lock, flags);
567*4882a593Smuzhiyun 
568*4882a593Smuzhiyun 	return rate64;
569*4882a593Smuzhiyun }
570*4882a593Smuzhiyun 
571*4882a593Smuzhiyun static const struct clk_ops lgm_clk_ddiv_ops = {
572*4882a593Smuzhiyun 	.recalc_rate = lgm_clk_ddiv_recalc_rate,
573*4882a593Smuzhiyun 	.enable	= lgm_clk_ddiv_enable,
574*4882a593Smuzhiyun 	.disable = lgm_clk_ddiv_disable,
575*4882a593Smuzhiyun 	.set_rate = lgm_clk_ddiv_set_rate,
576*4882a593Smuzhiyun 	.round_rate = lgm_clk_ddiv_round_rate,
577*4882a593Smuzhiyun };
578*4882a593Smuzhiyun 
lgm_clk_register_ddiv(struct lgm_clk_provider * ctx,const struct lgm_clk_ddiv_data * list,unsigned int nr_clk)579*4882a593Smuzhiyun int lgm_clk_register_ddiv(struct lgm_clk_provider *ctx,
580*4882a593Smuzhiyun 			  const struct lgm_clk_ddiv_data *list,
581*4882a593Smuzhiyun 			  unsigned int nr_clk)
582*4882a593Smuzhiyun {
583*4882a593Smuzhiyun 	struct device *dev = ctx->dev;
584*4882a593Smuzhiyun 	struct clk_hw *hw;
585*4882a593Smuzhiyun 	unsigned int idx;
586*4882a593Smuzhiyun 	int ret;
587*4882a593Smuzhiyun 
588*4882a593Smuzhiyun 	for (idx = 0; idx < nr_clk; idx++, list++) {
589*4882a593Smuzhiyun 		struct clk_init_data init = {};
590*4882a593Smuzhiyun 		struct lgm_clk_ddiv *ddiv;
591*4882a593Smuzhiyun 
592*4882a593Smuzhiyun 		ddiv = devm_kzalloc(dev, sizeof(*ddiv), GFP_KERNEL);
593*4882a593Smuzhiyun 		if (!ddiv)
594*4882a593Smuzhiyun 			return -ENOMEM;
595*4882a593Smuzhiyun 
596*4882a593Smuzhiyun 		init.name = list->name;
597*4882a593Smuzhiyun 		init.ops = &lgm_clk_ddiv_ops;
598*4882a593Smuzhiyun 		init.flags = list->flags;
599*4882a593Smuzhiyun 		init.parent_data = list->parent_data;
600*4882a593Smuzhiyun 		init.num_parents = 1;
601*4882a593Smuzhiyun 
602*4882a593Smuzhiyun 		ddiv->membase = ctx->membase;
603*4882a593Smuzhiyun 		ddiv->lock = ctx->lock;
604*4882a593Smuzhiyun 		ddiv->reg = list->reg;
605*4882a593Smuzhiyun 		ddiv->shift0 = list->shift0;
606*4882a593Smuzhiyun 		ddiv->width0 = list->width0;
607*4882a593Smuzhiyun 		ddiv->shift1 = list->shift1;
608*4882a593Smuzhiyun 		ddiv->width1 = list->width1;
609*4882a593Smuzhiyun 		ddiv->shift_gate = list->shift_gate;
610*4882a593Smuzhiyun 		ddiv->width_gate = list->width_gate;
611*4882a593Smuzhiyun 		ddiv->shift2 = list->ex_shift;
612*4882a593Smuzhiyun 		ddiv->width2 = list->ex_width;
613*4882a593Smuzhiyun 		ddiv->flags = list->div_flags;
614*4882a593Smuzhiyun 		ddiv->mult = 2;
615*4882a593Smuzhiyun 		ddiv->div = 5;
616*4882a593Smuzhiyun 		ddiv->hw.init = &init;
617*4882a593Smuzhiyun 
618*4882a593Smuzhiyun 		hw = &ddiv->hw;
619*4882a593Smuzhiyun 		ret = devm_clk_hw_register(dev, hw);
620*4882a593Smuzhiyun 		if (ret) {
621*4882a593Smuzhiyun 			dev_err(dev, "register clk: %s failed!\n", list->name);
622*4882a593Smuzhiyun 			return ret;
623*4882a593Smuzhiyun 		}
624*4882a593Smuzhiyun 		ctx->clk_data.hws[list->id] = hw;
625*4882a593Smuzhiyun 	}
626*4882a593Smuzhiyun 
627*4882a593Smuzhiyun 	return 0;
628*4882a593Smuzhiyun }
629