xref: /OK3568_Linux_fs/u-boot/arch/arm/mach-omap2/clocks-common.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun  *
3*4882a593Smuzhiyun  * Clock initialization for OMAP4
4*4882a593Smuzhiyun  *
5*4882a593Smuzhiyun  * (C) Copyright 2010
6*4882a593Smuzhiyun  * Texas Instruments, <www.ti.com>
7*4882a593Smuzhiyun  *
8*4882a593Smuzhiyun  * Aneesh V <aneesh@ti.com>
9*4882a593Smuzhiyun  *
10*4882a593Smuzhiyun  * Based on previous work by:
11*4882a593Smuzhiyun  *	Santosh Shilimkar <santosh.shilimkar@ti.com>
12*4882a593Smuzhiyun  *	Rajendra Nayak <rnayak@ti.com>
13*4882a593Smuzhiyun  *
14*4882a593Smuzhiyun  * SPDX-License-Identifier:	GPL-2.0+
15*4882a593Smuzhiyun  */
16*4882a593Smuzhiyun #include <common.h>
17*4882a593Smuzhiyun #include <i2c.h>
18*4882a593Smuzhiyun #include <asm/omap_common.h>
19*4882a593Smuzhiyun #include <asm/gpio.h>
20*4882a593Smuzhiyun #include <asm/arch/clock.h>
21*4882a593Smuzhiyun #include <asm/arch/sys_proto.h>
22*4882a593Smuzhiyun #include <asm/utils.h>
23*4882a593Smuzhiyun #include <asm/omap_gpio.h>
24*4882a593Smuzhiyun #include <asm/emif.h>
25*4882a593Smuzhiyun 
26*4882a593Smuzhiyun #ifndef CONFIG_SPL_BUILD
27*4882a593Smuzhiyun /*
28*4882a593Smuzhiyun  * printing to console doesn't work unless
29*4882a593Smuzhiyun  * this code is executed from SPL
30*4882a593Smuzhiyun  */
31*4882a593Smuzhiyun #define printf(fmt, args...)
32*4882a593Smuzhiyun #define puts(s)
33*4882a593Smuzhiyun #endif
34*4882a593Smuzhiyun 
35*4882a593Smuzhiyun const u32 sys_clk_array[8] = {
36*4882a593Smuzhiyun 	12000000,	       /* 12 MHz */
37*4882a593Smuzhiyun 	20000000,		/* 20 MHz */
38*4882a593Smuzhiyun 	16800000,	       /* 16.8 MHz */
39*4882a593Smuzhiyun 	19200000,	       /* 19.2 MHz */
40*4882a593Smuzhiyun 	26000000,	       /* 26 MHz */
41*4882a593Smuzhiyun 	27000000,	       /* 27 MHz */
42*4882a593Smuzhiyun 	38400000,	       /* 38.4 MHz */
43*4882a593Smuzhiyun };
44*4882a593Smuzhiyun 
__get_sys_clk_index(void)45*4882a593Smuzhiyun static inline u32 __get_sys_clk_index(void)
46*4882a593Smuzhiyun {
47*4882a593Smuzhiyun 	s8 ind;
48*4882a593Smuzhiyun 	/*
49*4882a593Smuzhiyun 	 * For ES1 the ROM code calibration of sys clock is not reliable
50*4882a593Smuzhiyun 	 * due to hw issue. So, use hard-coded value. If this value is not
51*4882a593Smuzhiyun 	 * correct for any board over-ride this function in board file
52*4882a593Smuzhiyun 	 * From ES2.0 onwards you will get this information from
53*4882a593Smuzhiyun 	 * CM_SYS_CLKSEL
54*4882a593Smuzhiyun 	 */
55*4882a593Smuzhiyun 	if (omap_revision() == OMAP4430_ES1_0)
56*4882a593Smuzhiyun 		ind = OMAP_SYS_CLK_IND_38_4_MHZ;
57*4882a593Smuzhiyun 	else {
58*4882a593Smuzhiyun 		/* SYS_CLKSEL - 1 to match the dpll param array indices */
59*4882a593Smuzhiyun 		ind = (readl((*prcm)->cm_sys_clksel) &
60*4882a593Smuzhiyun 			CM_SYS_CLKSEL_SYS_CLKSEL_MASK) - 1;
61*4882a593Smuzhiyun 	}
62*4882a593Smuzhiyun 	return ind;
63*4882a593Smuzhiyun }
64*4882a593Smuzhiyun 
65*4882a593Smuzhiyun u32 get_sys_clk_index(void)
66*4882a593Smuzhiyun 	__attribute__ ((weak, alias("__get_sys_clk_index")));
67*4882a593Smuzhiyun 
get_sys_clk_freq(void)68*4882a593Smuzhiyun u32 get_sys_clk_freq(void)
69*4882a593Smuzhiyun {
70*4882a593Smuzhiyun 	u8 index = get_sys_clk_index();
71*4882a593Smuzhiyun 	return sys_clk_array[index];
72*4882a593Smuzhiyun }
73*4882a593Smuzhiyun 
setup_post_dividers(u32 const base,const struct dpll_params * params)74*4882a593Smuzhiyun void setup_post_dividers(u32 const base, const struct dpll_params *params)
75*4882a593Smuzhiyun {
76*4882a593Smuzhiyun 	struct dpll_regs *const dpll_regs = (struct dpll_regs *)base;
77*4882a593Smuzhiyun 
78*4882a593Smuzhiyun 	/* Setup post-dividers */
79*4882a593Smuzhiyun 	if (params->m2 >= 0)
80*4882a593Smuzhiyun 		writel(params->m2, &dpll_regs->cm_div_m2_dpll);
81*4882a593Smuzhiyun 	if (params->m3 >= 0)
82*4882a593Smuzhiyun 		writel(params->m3, &dpll_regs->cm_div_m3_dpll);
83*4882a593Smuzhiyun 	if (params->m4_h11 >= 0)
84*4882a593Smuzhiyun 		writel(params->m4_h11, &dpll_regs->cm_div_m4_h11_dpll);
85*4882a593Smuzhiyun 	if (params->m5_h12 >= 0)
86*4882a593Smuzhiyun 		writel(params->m5_h12, &dpll_regs->cm_div_m5_h12_dpll);
87*4882a593Smuzhiyun 	if (params->m6_h13 >= 0)
88*4882a593Smuzhiyun 		writel(params->m6_h13, &dpll_regs->cm_div_m6_h13_dpll);
89*4882a593Smuzhiyun 	if (params->m7_h14 >= 0)
90*4882a593Smuzhiyun 		writel(params->m7_h14, &dpll_regs->cm_div_m7_h14_dpll);
91*4882a593Smuzhiyun 	if (params->h21 >= 0)
92*4882a593Smuzhiyun 		writel(params->h21, &dpll_regs->cm_div_h21_dpll);
93*4882a593Smuzhiyun 	if (params->h22 >= 0)
94*4882a593Smuzhiyun 		writel(params->h22, &dpll_regs->cm_div_h22_dpll);
95*4882a593Smuzhiyun 	if (params->h23 >= 0)
96*4882a593Smuzhiyun 		writel(params->h23, &dpll_regs->cm_div_h23_dpll);
97*4882a593Smuzhiyun 	if (params->h24 >= 0)
98*4882a593Smuzhiyun 		writel(params->h24, &dpll_regs->cm_div_h24_dpll);
99*4882a593Smuzhiyun }
100*4882a593Smuzhiyun 
do_bypass_dpll(u32 const base)101*4882a593Smuzhiyun static inline void do_bypass_dpll(u32 const base)
102*4882a593Smuzhiyun {
103*4882a593Smuzhiyun 	struct dpll_regs *dpll_regs = (struct dpll_regs *)base;
104*4882a593Smuzhiyun 
105*4882a593Smuzhiyun 	clrsetbits_le32(&dpll_regs->cm_clkmode_dpll,
106*4882a593Smuzhiyun 			CM_CLKMODE_DPLL_DPLL_EN_MASK,
107*4882a593Smuzhiyun 			DPLL_EN_FAST_RELOCK_BYPASS <<
108*4882a593Smuzhiyun 			CM_CLKMODE_DPLL_EN_SHIFT);
109*4882a593Smuzhiyun }
110*4882a593Smuzhiyun 
wait_for_bypass(u32 const base)111*4882a593Smuzhiyun static inline void wait_for_bypass(u32 const base)
112*4882a593Smuzhiyun {
113*4882a593Smuzhiyun 	struct dpll_regs *const dpll_regs = (struct dpll_regs *)base;
114*4882a593Smuzhiyun 
115*4882a593Smuzhiyun 	if (!wait_on_value(ST_DPLL_CLK_MASK, 0, &dpll_regs->cm_idlest_dpll,
116*4882a593Smuzhiyun 				LDELAY)) {
117*4882a593Smuzhiyun 		printf("Bypassing DPLL failed %x\n", base);
118*4882a593Smuzhiyun 	}
119*4882a593Smuzhiyun }
120*4882a593Smuzhiyun 
do_lock_dpll(u32 const base)121*4882a593Smuzhiyun static inline void do_lock_dpll(u32 const base)
122*4882a593Smuzhiyun {
123*4882a593Smuzhiyun 	struct dpll_regs *const dpll_regs = (struct dpll_regs *)base;
124*4882a593Smuzhiyun 
125*4882a593Smuzhiyun 	clrsetbits_le32(&dpll_regs->cm_clkmode_dpll,
126*4882a593Smuzhiyun 		      CM_CLKMODE_DPLL_DPLL_EN_MASK,
127*4882a593Smuzhiyun 		      DPLL_EN_LOCK << CM_CLKMODE_DPLL_EN_SHIFT);
128*4882a593Smuzhiyun }
129*4882a593Smuzhiyun 
wait_for_lock(u32 const base)130*4882a593Smuzhiyun static inline void wait_for_lock(u32 const base)
131*4882a593Smuzhiyun {
132*4882a593Smuzhiyun 	struct dpll_regs *const dpll_regs = (struct dpll_regs *)base;
133*4882a593Smuzhiyun 
134*4882a593Smuzhiyun 	if (!wait_on_value(ST_DPLL_CLK_MASK, ST_DPLL_CLK_MASK,
135*4882a593Smuzhiyun 		&dpll_regs->cm_idlest_dpll, LDELAY)) {
136*4882a593Smuzhiyun 		printf("DPLL locking failed for %x\n", base);
137*4882a593Smuzhiyun 		hang();
138*4882a593Smuzhiyun 	}
139*4882a593Smuzhiyun }
140*4882a593Smuzhiyun 
check_for_lock(u32 const base)141*4882a593Smuzhiyun inline u32 check_for_lock(u32 const base)
142*4882a593Smuzhiyun {
143*4882a593Smuzhiyun 	struct dpll_regs *const dpll_regs = (struct dpll_regs *)base;
144*4882a593Smuzhiyun 	u32 lock = readl(&dpll_regs->cm_idlest_dpll) & ST_DPLL_CLK_MASK;
145*4882a593Smuzhiyun 
146*4882a593Smuzhiyun 	return lock;
147*4882a593Smuzhiyun }
148*4882a593Smuzhiyun 
get_mpu_dpll_params(struct dplls const * dpll_data)149*4882a593Smuzhiyun const struct dpll_params *get_mpu_dpll_params(struct dplls const *dpll_data)
150*4882a593Smuzhiyun {
151*4882a593Smuzhiyun 	u32 sysclk_ind = get_sys_clk_index();
152*4882a593Smuzhiyun 	return &dpll_data->mpu[sysclk_ind];
153*4882a593Smuzhiyun }
154*4882a593Smuzhiyun 
get_core_dpll_params(struct dplls const * dpll_data)155*4882a593Smuzhiyun const struct dpll_params *get_core_dpll_params(struct dplls const *dpll_data)
156*4882a593Smuzhiyun {
157*4882a593Smuzhiyun 	u32 sysclk_ind = get_sys_clk_index();
158*4882a593Smuzhiyun 	return &dpll_data->core[sysclk_ind];
159*4882a593Smuzhiyun }
160*4882a593Smuzhiyun 
get_per_dpll_params(struct dplls const * dpll_data)161*4882a593Smuzhiyun const struct dpll_params *get_per_dpll_params(struct dplls const *dpll_data)
162*4882a593Smuzhiyun {
163*4882a593Smuzhiyun 	u32 sysclk_ind = get_sys_clk_index();
164*4882a593Smuzhiyun 	return &dpll_data->per[sysclk_ind];
165*4882a593Smuzhiyun }
166*4882a593Smuzhiyun 
get_iva_dpll_params(struct dplls const * dpll_data)167*4882a593Smuzhiyun const struct dpll_params *get_iva_dpll_params(struct dplls const *dpll_data)
168*4882a593Smuzhiyun {
169*4882a593Smuzhiyun 	u32 sysclk_ind = get_sys_clk_index();
170*4882a593Smuzhiyun 	return &dpll_data->iva[sysclk_ind];
171*4882a593Smuzhiyun }
172*4882a593Smuzhiyun 
get_usb_dpll_params(struct dplls const * dpll_data)173*4882a593Smuzhiyun const struct dpll_params *get_usb_dpll_params(struct dplls const *dpll_data)
174*4882a593Smuzhiyun {
175*4882a593Smuzhiyun 	u32 sysclk_ind = get_sys_clk_index();
176*4882a593Smuzhiyun 	return &dpll_data->usb[sysclk_ind];
177*4882a593Smuzhiyun }
178*4882a593Smuzhiyun 
get_abe_dpll_params(struct dplls const * dpll_data)179*4882a593Smuzhiyun const struct dpll_params *get_abe_dpll_params(struct dplls const *dpll_data)
180*4882a593Smuzhiyun {
181*4882a593Smuzhiyun #ifdef CONFIG_SYS_OMAP_ABE_SYSCK
182*4882a593Smuzhiyun 	u32 sysclk_ind = get_sys_clk_index();
183*4882a593Smuzhiyun 	return &dpll_data->abe[sysclk_ind];
184*4882a593Smuzhiyun #else
185*4882a593Smuzhiyun 	return dpll_data->abe;
186*4882a593Smuzhiyun #endif
187*4882a593Smuzhiyun }
188*4882a593Smuzhiyun 
get_ddr_dpll_params(struct dplls const * dpll_data)189*4882a593Smuzhiyun static const struct dpll_params *get_ddr_dpll_params
190*4882a593Smuzhiyun 			(struct dplls const *dpll_data)
191*4882a593Smuzhiyun {
192*4882a593Smuzhiyun 	u32 sysclk_ind = get_sys_clk_index();
193*4882a593Smuzhiyun 
194*4882a593Smuzhiyun 	if (!dpll_data->ddr)
195*4882a593Smuzhiyun 		return NULL;
196*4882a593Smuzhiyun 	return &dpll_data->ddr[sysclk_ind];
197*4882a593Smuzhiyun }
198*4882a593Smuzhiyun 
199*4882a593Smuzhiyun #ifdef CONFIG_DRIVER_TI_CPSW
get_gmac_dpll_params(struct dplls const * dpll_data)200*4882a593Smuzhiyun static const struct dpll_params *get_gmac_dpll_params
201*4882a593Smuzhiyun 			(struct dplls const *dpll_data)
202*4882a593Smuzhiyun {
203*4882a593Smuzhiyun 	u32 sysclk_ind = get_sys_clk_index();
204*4882a593Smuzhiyun 
205*4882a593Smuzhiyun 	if (!dpll_data->gmac)
206*4882a593Smuzhiyun 		return NULL;
207*4882a593Smuzhiyun 	return &dpll_data->gmac[sysclk_ind];
208*4882a593Smuzhiyun }
209*4882a593Smuzhiyun #endif
210*4882a593Smuzhiyun 
do_setup_dpll(u32 const base,const struct dpll_params * params,u8 lock,char * dpll)211*4882a593Smuzhiyun static void do_setup_dpll(u32 const base, const struct dpll_params *params,
212*4882a593Smuzhiyun 				u8 lock, char *dpll)
213*4882a593Smuzhiyun {
214*4882a593Smuzhiyun 	u32 temp, M, N;
215*4882a593Smuzhiyun 	struct dpll_regs *const dpll_regs = (struct dpll_regs *)base;
216*4882a593Smuzhiyun 
217*4882a593Smuzhiyun 	if (!params)
218*4882a593Smuzhiyun 		return;
219*4882a593Smuzhiyun 
220*4882a593Smuzhiyun 	temp = readl(&dpll_regs->cm_clksel_dpll);
221*4882a593Smuzhiyun 
222*4882a593Smuzhiyun 	if (check_for_lock(base)) {
223*4882a593Smuzhiyun 		/*
224*4882a593Smuzhiyun 		 * The Dpll has already been locked by rom code using CH.
225*4882a593Smuzhiyun 		 * Check if M,N are matching with Ideal nominal opp values.
226*4882a593Smuzhiyun 		 * If matches, skip the rest otherwise relock.
227*4882a593Smuzhiyun 		 */
228*4882a593Smuzhiyun 		M = (temp & CM_CLKSEL_DPLL_M_MASK) >> CM_CLKSEL_DPLL_M_SHIFT;
229*4882a593Smuzhiyun 		N = (temp & CM_CLKSEL_DPLL_N_MASK) >> CM_CLKSEL_DPLL_N_SHIFT;
230*4882a593Smuzhiyun 		if ((M != (params->m)) || (N != (params->n))) {
231*4882a593Smuzhiyun 			debug("\n %s Dpll locked, but not for ideal M = %d,"
232*4882a593Smuzhiyun 				"N = %d values, current values are M = %d,"
233*4882a593Smuzhiyun 				"N= %d" , dpll, params->m, params->n,
234*4882a593Smuzhiyun 				M, N);
235*4882a593Smuzhiyun 		} else {
236*4882a593Smuzhiyun 			/* Dpll locked with ideal values for nominal opps. */
237*4882a593Smuzhiyun 			debug("\n %s Dpll already locked with ideal"
238*4882a593Smuzhiyun 						"nominal opp values", dpll);
239*4882a593Smuzhiyun 
240*4882a593Smuzhiyun 			bypass_dpll(base);
241*4882a593Smuzhiyun 			goto setup_post_dividers;
242*4882a593Smuzhiyun 		}
243*4882a593Smuzhiyun 	}
244*4882a593Smuzhiyun 
245*4882a593Smuzhiyun 	bypass_dpll(base);
246*4882a593Smuzhiyun 
247*4882a593Smuzhiyun 	/* Set M & N */
248*4882a593Smuzhiyun 	temp &= ~CM_CLKSEL_DPLL_M_MASK;
249*4882a593Smuzhiyun 	temp |= (params->m << CM_CLKSEL_DPLL_M_SHIFT) & CM_CLKSEL_DPLL_M_MASK;
250*4882a593Smuzhiyun 
251*4882a593Smuzhiyun 	temp &= ~CM_CLKSEL_DPLL_N_MASK;
252*4882a593Smuzhiyun 	temp |= (params->n << CM_CLKSEL_DPLL_N_SHIFT) & CM_CLKSEL_DPLL_N_MASK;
253*4882a593Smuzhiyun 
254*4882a593Smuzhiyun 	writel(temp, &dpll_regs->cm_clksel_dpll);
255*4882a593Smuzhiyun 
256*4882a593Smuzhiyun setup_post_dividers:
257*4882a593Smuzhiyun 	setup_post_dividers(base, params);
258*4882a593Smuzhiyun 
259*4882a593Smuzhiyun 	/* Lock */
260*4882a593Smuzhiyun 	if (lock)
261*4882a593Smuzhiyun 		do_lock_dpll(base);
262*4882a593Smuzhiyun 
263*4882a593Smuzhiyun 	/* Wait till the DPLL locks */
264*4882a593Smuzhiyun 	if (lock)
265*4882a593Smuzhiyun 		wait_for_lock(base);
266*4882a593Smuzhiyun }
267*4882a593Smuzhiyun 
omap_ddr_clk(void)268*4882a593Smuzhiyun u32 omap_ddr_clk(void)
269*4882a593Smuzhiyun {
270*4882a593Smuzhiyun 	u32 ddr_clk, sys_clk_khz, omap_rev, divider;
271*4882a593Smuzhiyun 	const struct dpll_params *core_dpll_params;
272*4882a593Smuzhiyun 
273*4882a593Smuzhiyun 	omap_rev = omap_revision();
274*4882a593Smuzhiyun 	sys_clk_khz = get_sys_clk_freq() / 1000;
275*4882a593Smuzhiyun 
276*4882a593Smuzhiyun 	core_dpll_params = get_core_dpll_params(*dplls_data);
277*4882a593Smuzhiyun 
278*4882a593Smuzhiyun 	debug("sys_clk %d\n ", sys_clk_khz * 1000);
279*4882a593Smuzhiyun 
280*4882a593Smuzhiyun 	/* Find Core DPLL locked frequency first */
281*4882a593Smuzhiyun 	ddr_clk = sys_clk_khz * 2 * core_dpll_params->m /
282*4882a593Smuzhiyun 			(core_dpll_params->n + 1);
283*4882a593Smuzhiyun 
284*4882a593Smuzhiyun 	if (omap_rev < OMAP5430_ES1_0) {
285*4882a593Smuzhiyun 		/*
286*4882a593Smuzhiyun 		 * DDR frequency is PHY_ROOT_CLK/2
287*4882a593Smuzhiyun 		 * PHY_ROOT_CLK = Fdpll/2/M2
288*4882a593Smuzhiyun 		 */
289*4882a593Smuzhiyun 		divider = 4;
290*4882a593Smuzhiyun 	} else {
291*4882a593Smuzhiyun 		/*
292*4882a593Smuzhiyun 		 * DDR frequency is PHY_ROOT_CLK
293*4882a593Smuzhiyun 		 * PHY_ROOT_CLK = Fdpll/2/M2
294*4882a593Smuzhiyun 		 */
295*4882a593Smuzhiyun 		divider = 2;
296*4882a593Smuzhiyun 	}
297*4882a593Smuzhiyun 
298*4882a593Smuzhiyun 	ddr_clk = ddr_clk / divider / core_dpll_params->m2;
299*4882a593Smuzhiyun 	ddr_clk *= 1000;	/* convert to Hz */
300*4882a593Smuzhiyun 	debug("ddr_clk %d\n ", ddr_clk);
301*4882a593Smuzhiyun 
302*4882a593Smuzhiyun 	return ddr_clk;
303*4882a593Smuzhiyun }
304*4882a593Smuzhiyun 
305*4882a593Smuzhiyun /*
306*4882a593Smuzhiyun  * Lock MPU dpll
307*4882a593Smuzhiyun  *
308*4882a593Smuzhiyun  * Resulting MPU frequencies:
309*4882a593Smuzhiyun  * 4430 ES1.0	: 600 MHz
310*4882a593Smuzhiyun  * 4430 ES2.x	: 792 MHz (OPP Turbo)
311*4882a593Smuzhiyun  * 4460		: 920 MHz (OPP Turbo) - DCC disabled
312*4882a593Smuzhiyun  */
configure_mpu_dpll(void)313*4882a593Smuzhiyun void configure_mpu_dpll(void)
314*4882a593Smuzhiyun {
315*4882a593Smuzhiyun 	const struct dpll_params *params;
316*4882a593Smuzhiyun 	struct dpll_regs *mpu_dpll_regs;
317*4882a593Smuzhiyun 	u32 omap_rev;
318*4882a593Smuzhiyun 	omap_rev = omap_revision();
319*4882a593Smuzhiyun 
320*4882a593Smuzhiyun 	/*
321*4882a593Smuzhiyun 	 * DCC and clock divider settings for 4460.
322*4882a593Smuzhiyun 	 * DCC is required, if more than a certain frequency is required.
323*4882a593Smuzhiyun 	 * For, 4460 > 1GHZ.
324*4882a593Smuzhiyun 	 *     5430 > 1.4GHZ.
325*4882a593Smuzhiyun 	 */
326*4882a593Smuzhiyun 	if ((omap_rev >= OMAP4460_ES1_0) && (omap_rev < OMAP5430_ES1_0)) {
327*4882a593Smuzhiyun 		mpu_dpll_regs =
328*4882a593Smuzhiyun 			(struct dpll_regs *)((*prcm)->cm_clkmode_dpll_mpu);
329*4882a593Smuzhiyun 		bypass_dpll((*prcm)->cm_clkmode_dpll_mpu);
330*4882a593Smuzhiyun 		clrbits_le32((*prcm)->cm_mpu_mpu_clkctrl,
331*4882a593Smuzhiyun 			MPU_CLKCTRL_CLKSEL_EMIF_DIV_MODE_MASK);
332*4882a593Smuzhiyun 		setbits_le32((*prcm)->cm_mpu_mpu_clkctrl,
333*4882a593Smuzhiyun 			MPU_CLKCTRL_CLKSEL_ABE_DIV_MODE_MASK);
334*4882a593Smuzhiyun 		clrbits_le32(&mpu_dpll_regs->cm_clksel_dpll,
335*4882a593Smuzhiyun 			CM_CLKSEL_DCC_EN_MASK);
336*4882a593Smuzhiyun 	}
337*4882a593Smuzhiyun 
338*4882a593Smuzhiyun 	params = get_mpu_dpll_params(*dplls_data);
339*4882a593Smuzhiyun 
340*4882a593Smuzhiyun 	do_setup_dpll((*prcm)->cm_clkmode_dpll_mpu, params, DPLL_LOCK, "mpu");
341*4882a593Smuzhiyun 	debug("MPU DPLL locked\n");
342*4882a593Smuzhiyun }
343*4882a593Smuzhiyun 
344*4882a593Smuzhiyun #if defined(CONFIG_USB_EHCI_OMAP) || defined(CONFIG_USB_XHCI_OMAP) || \
345*4882a593Smuzhiyun 	defined(CONFIG_USB_MUSB_OMAP2PLUS)
setup_usb_dpll(void)346*4882a593Smuzhiyun static void setup_usb_dpll(void)
347*4882a593Smuzhiyun {
348*4882a593Smuzhiyun 	const struct dpll_params *params;
349*4882a593Smuzhiyun 	u32 sys_clk_khz, sd_div, num, den;
350*4882a593Smuzhiyun 
351*4882a593Smuzhiyun 	sys_clk_khz = get_sys_clk_freq() / 1000;
352*4882a593Smuzhiyun 	/*
353*4882a593Smuzhiyun 	 * USB:
354*4882a593Smuzhiyun 	 * USB dpll is J-type. Need to set DPLL_SD_DIV for jitter correction
355*4882a593Smuzhiyun 	 * DPLL_SD_DIV = CEILING ([DPLL_MULT/(DPLL_DIV+1)]* CLKINP / 250)
356*4882a593Smuzhiyun 	 *      - where CLKINP is sys_clk in MHz
357*4882a593Smuzhiyun 	 * Use CLKINP in KHz and adjust the denominator accordingly so
358*4882a593Smuzhiyun 	 * that we have enough accuracy and at the same time no overflow
359*4882a593Smuzhiyun 	 */
360*4882a593Smuzhiyun 	params = get_usb_dpll_params(*dplls_data);
361*4882a593Smuzhiyun 	num = params->m * sys_clk_khz;
362*4882a593Smuzhiyun 	den = (params->n + 1) * 250 * 1000;
363*4882a593Smuzhiyun 	num += den - 1;
364*4882a593Smuzhiyun 	sd_div = num / den;
365*4882a593Smuzhiyun 	clrsetbits_le32((*prcm)->cm_clksel_dpll_usb,
366*4882a593Smuzhiyun 			CM_CLKSEL_DPLL_DPLL_SD_DIV_MASK,
367*4882a593Smuzhiyun 			sd_div << CM_CLKSEL_DPLL_DPLL_SD_DIV_SHIFT);
368*4882a593Smuzhiyun 
369*4882a593Smuzhiyun 	/* Now setup the dpll with the regular function */
370*4882a593Smuzhiyun 	do_setup_dpll((*prcm)->cm_clkmode_dpll_usb, params, DPLL_LOCK, "usb");
371*4882a593Smuzhiyun }
372*4882a593Smuzhiyun #endif
373*4882a593Smuzhiyun 
setup_dplls(void)374*4882a593Smuzhiyun static void setup_dplls(void)
375*4882a593Smuzhiyun {
376*4882a593Smuzhiyun 	u32 temp;
377*4882a593Smuzhiyun 	const struct dpll_params *params;
378*4882a593Smuzhiyun 	struct emif_reg_struct *emif = (struct emif_reg_struct *)EMIF1_BASE;
379*4882a593Smuzhiyun 
380*4882a593Smuzhiyun 	debug("setup_dplls\n");
381*4882a593Smuzhiyun 
382*4882a593Smuzhiyun 	/* CORE dpll */
383*4882a593Smuzhiyun 	params = get_core_dpll_params(*dplls_data);	/* default - safest */
384*4882a593Smuzhiyun 	/*
385*4882a593Smuzhiyun 	 * Do not lock the core DPLL now. Just set it up.
386*4882a593Smuzhiyun 	 * Core DPLL will be locked after setting up EMIF
387*4882a593Smuzhiyun 	 * using the FREQ_UPDATE method(freq_update_core())
388*4882a593Smuzhiyun 	 */
389*4882a593Smuzhiyun 	if (emif_sdram_type(readl(&emif->emif_sdram_config)) ==
390*4882a593Smuzhiyun 	    EMIF_SDRAM_TYPE_LPDDR2)
391*4882a593Smuzhiyun 		do_setup_dpll((*prcm)->cm_clkmode_dpll_core, params,
392*4882a593Smuzhiyun 							DPLL_NO_LOCK, "core");
393*4882a593Smuzhiyun 	else
394*4882a593Smuzhiyun 		do_setup_dpll((*prcm)->cm_clkmode_dpll_core, params,
395*4882a593Smuzhiyun 							DPLL_LOCK, "core");
396*4882a593Smuzhiyun 	/* Set the ratios for CORE_CLK, L3_CLK, L4_CLK */
397*4882a593Smuzhiyun 	temp = (CLKSEL_CORE_X2_DIV_1 << CLKSEL_CORE_SHIFT) |
398*4882a593Smuzhiyun 	    (CLKSEL_L3_CORE_DIV_2 << CLKSEL_L3_SHIFT) |
399*4882a593Smuzhiyun 	    (CLKSEL_L4_L3_DIV_2 << CLKSEL_L4_SHIFT);
400*4882a593Smuzhiyun 	writel(temp, (*prcm)->cm_clksel_core);
401*4882a593Smuzhiyun 	debug("Core DPLL configured\n");
402*4882a593Smuzhiyun 
403*4882a593Smuzhiyun 	/* lock PER dpll */
404*4882a593Smuzhiyun 	params = get_per_dpll_params(*dplls_data);
405*4882a593Smuzhiyun 	do_setup_dpll((*prcm)->cm_clkmode_dpll_per,
406*4882a593Smuzhiyun 			params, DPLL_LOCK, "per");
407*4882a593Smuzhiyun 	debug("PER DPLL locked\n");
408*4882a593Smuzhiyun 
409*4882a593Smuzhiyun 	/* MPU dpll */
410*4882a593Smuzhiyun 	configure_mpu_dpll();
411*4882a593Smuzhiyun 
412*4882a593Smuzhiyun #if defined(CONFIG_USB_EHCI_OMAP) || defined(CONFIG_USB_XHCI_OMAP) || \
413*4882a593Smuzhiyun 	defined(CONFIG_USB_MUSB_OMAP2PLUS)
414*4882a593Smuzhiyun 	setup_usb_dpll();
415*4882a593Smuzhiyun #endif
416*4882a593Smuzhiyun 	params = get_ddr_dpll_params(*dplls_data);
417*4882a593Smuzhiyun 	do_setup_dpll((*prcm)->cm_clkmode_dpll_ddrphy,
418*4882a593Smuzhiyun 		      params, DPLL_LOCK, "ddr");
419*4882a593Smuzhiyun 
420*4882a593Smuzhiyun #ifdef CONFIG_DRIVER_TI_CPSW
421*4882a593Smuzhiyun 	params = get_gmac_dpll_params(*dplls_data);
422*4882a593Smuzhiyun 	do_setup_dpll((*prcm)->cm_clkmode_dpll_gmac, params,
423*4882a593Smuzhiyun 		      DPLL_LOCK, "gmac");
424*4882a593Smuzhiyun #endif
425*4882a593Smuzhiyun }
426*4882a593Smuzhiyun 
get_offset_code(u32 volt_offset,struct pmic_data * pmic)427*4882a593Smuzhiyun u32 get_offset_code(u32 volt_offset, struct pmic_data *pmic)
428*4882a593Smuzhiyun {
429*4882a593Smuzhiyun 	u32 offset_code;
430*4882a593Smuzhiyun 
431*4882a593Smuzhiyun 	volt_offset -= pmic->base_offset;
432*4882a593Smuzhiyun 
433*4882a593Smuzhiyun 	offset_code = (volt_offset + pmic->step - 1) / pmic->step;
434*4882a593Smuzhiyun 
435*4882a593Smuzhiyun 	/*
436*4882a593Smuzhiyun 	 * Offset codes 1-6 all give the base voltage in Palmas
437*4882a593Smuzhiyun 	 * Offset code 0 switches OFF the SMPS
438*4882a593Smuzhiyun 	 */
439*4882a593Smuzhiyun 	return offset_code + pmic->start_code;
440*4882a593Smuzhiyun }
441*4882a593Smuzhiyun 
do_scale_vcore(u32 vcore_reg,u32 volt_mv,struct pmic_data * pmic)442*4882a593Smuzhiyun void do_scale_vcore(u32 vcore_reg, u32 volt_mv, struct pmic_data *pmic)
443*4882a593Smuzhiyun {
444*4882a593Smuzhiyun 	u32 offset_code;
445*4882a593Smuzhiyun 	u32 offset = volt_mv;
446*4882a593Smuzhiyun 	int ret = 0;
447*4882a593Smuzhiyun 
448*4882a593Smuzhiyun 	if (!volt_mv)
449*4882a593Smuzhiyun 		return;
450*4882a593Smuzhiyun 
451*4882a593Smuzhiyun 	pmic->pmic_bus_init();
452*4882a593Smuzhiyun 	/* See if we can first get the GPIO if needed */
453*4882a593Smuzhiyun 	if (pmic->gpio_en)
454*4882a593Smuzhiyun 		ret = gpio_request(pmic->gpio, "PMIC_GPIO");
455*4882a593Smuzhiyun 
456*4882a593Smuzhiyun 	if (ret < 0) {
457*4882a593Smuzhiyun 		printf("%s: gpio %d request failed %d\n", __func__,
458*4882a593Smuzhiyun 							pmic->gpio, ret);
459*4882a593Smuzhiyun 		return;
460*4882a593Smuzhiyun 	}
461*4882a593Smuzhiyun 
462*4882a593Smuzhiyun 	/* Pull the GPIO low to select SET0 register, while we program SET1 */
463*4882a593Smuzhiyun 	if (pmic->gpio_en)
464*4882a593Smuzhiyun 		gpio_direction_output(pmic->gpio, 0);
465*4882a593Smuzhiyun 
466*4882a593Smuzhiyun 	/* convert to uV for better accuracy in the calculations */
467*4882a593Smuzhiyun 	offset *= 1000;
468*4882a593Smuzhiyun 
469*4882a593Smuzhiyun 	offset_code = get_offset_code(offset, pmic);
470*4882a593Smuzhiyun 
471*4882a593Smuzhiyun 	debug("do_scale_vcore: volt - %d offset_code - 0x%x\n", volt_mv,
472*4882a593Smuzhiyun 		offset_code);
473*4882a593Smuzhiyun 
474*4882a593Smuzhiyun 	if (pmic->pmic_write(pmic->i2c_slave_addr, vcore_reg, offset_code))
475*4882a593Smuzhiyun 		printf("Scaling voltage failed for 0x%x\n", vcore_reg);
476*4882a593Smuzhiyun 	if (pmic->gpio_en)
477*4882a593Smuzhiyun 		gpio_direction_output(pmic->gpio, 1);
478*4882a593Smuzhiyun }
479*4882a593Smuzhiyun 
get_voltrail_opp(int rail_offset)480*4882a593Smuzhiyun int __weak get_voltrail_opp(int rail_offset)
481*4882a593Smuzhiyun {
482*4882a593Smuzhiyun 	/*
483*4882a593Smuzhiyun 	 * By default return OPP_NOM for all voltage rails.
484*4882a593Smuzhiyun 	 */
485*4882a593Smuzhiyun 	return OPP_NOM;
486*4882a593Smuzhiyun }
487*4882a593Smuzhiyun 
optimize_vcore_voltage(struct volts const * v,int opp)488*4882a593Smuzhiyun static u32 optimize_vcore_voltage(struct volts const *v, int opp)
489*4882a593Smuzhiyun {
490*4882a593Smuzhiyun 	u32 val;
491*4882a593Smuzhiyun 
492*4882a593Smuzhiyun 	if (!v->value[opp])
493*4882a593Smuzhiyun 		return 0;
494*4882a593Smuzhiyun 	if (!v->efuse.reg[opp])
495*4882a593Smuzhiyun 		return v->value[opp];
496*4882a593Smuzhiyun 
497*4882a593Smuzhiyun 	switch (v->efuse.reg_bits) {
498*4882a593Smuzhiyun 	case 16:
499*4882a593Smuzhiyun 		val = readw(v->efuse.reg[opp]);
500*4882a593Smuzhiyun 		break;
501*4882a593Smuzhiyun 	case 32:
502*4882a593Smuzhiyun 		val = readl(v->efuse.reg[opp]);
503*4882a593Smuzhiyun 		break;
504*4882a593Smuzhiyun 	default:
505*4882a593Smuzhiyun 		printf("Error: efuse 0x%08x bits=%d unknown\n",
506*4882a593Smuzhiyun 		       v->efuse.reg[opp], v->efuse.reg_bits);
507*4882a593Smuzhiyun 		return v->value[opp];
508*4882a593Smuzhiyun 	}
509*4882a593Smuzhiyun 
510*4882a593Smuzhiyun 	if (!val) {
511*4882a593Smuzhiyun 		printf("Error: efuse 0x%08x bits=%d val=0, using %d\n",
512*4882a593Smuzhiyun 		       v->efuse.reg[opp], v->efuse.reg_bits, v->value[opp]);
513*4882a593Smuzhiyun 		return v->value[opp];
514*4882a593Smuzhiyun 	}
515*4882a593Smuzhiyun 
516*4882a593Smuzhiyun 	debug("%s:efuse 0x%08x bits=%d Vnom=%d, using efuse value %d\n",
517*4882a593Smuzhiyun 	      __func__, v->efuse.reg[opp], v->efuse.reg_bits, v->value[opp],
518*4882a593Smuzhiyun 	      val);
519*4882a593Smuzhiyun 	return val;
520*4882a593Smuzhiyun }
521*4882a593Smuzhiyun 
522*4882a593Smuzhiyun #ifdef CONFIG_IODELAY_RECALIBRATION
recalibrate_iodelay(void)523*4882a593Smuzhiyun void __weak recalibrate_iodelay(void)
524*4882a593Smuzhiyun {
525*4882a593Smuzhiyun }
526*4882a593Smuzhiyun #endif
527*4882a593Smuzhiyun 
528*4882a593Smuzhiyun /*
529*4882a593Smuzhiyun  * Setup the voltages for the main SoC core power domains.
530*4882a593Smuzhiyun  * We start with the maximum voltages allowed here, as set in the corresponding
531*4882a593Smuzhiyun  * vcores_data struct, and then scale (usually down) to the fused values that
532*4882a593Smuzhiyun  * are retrieved from the SoC. The scaling happens only if the efuse.reg fields
533*4882a593Smuzhiyun  * are initialised.
534*4882a593Smuzhiyun  * Rail grouping is supported for the DRA7xx SoCs only, therefore the code is
535*4882a593Smuzhiyun  * compiled conditionally. Note that the new code writes the scaled (or zeroed)
536*4882a593Smuzhiyun  * values back to the vcores_data struct for eventual reuse. Zero values mean
537*4882a593Smuzhiyun  * that the corresponding rails are not controlled separately, and are not sent
538*4882a593Smuzhiyun  * to the PMIC.
539*4882a593Smuzhiyun  */
scale_vcores(struct vcores_data const * vcores)540*4882a593Smuzhiyun void scale_vcores(struct vcores_data const *vcores)
541*4882a593Smuzhiyun {
542*4882a593Smuzhiyun 	int i, opp, j, ol;
543*4882a593Smuzhiyun 	struct volts *pv = (struct volts *)vcores;
544*4882a593Smuzhiyun 	struct volts *px;
545*4882a593Smuzhiyun 
546*4882a593Smuzhiyun 	for (i=0; i<(sizeof(struct vcores_data)/sizeof(struct volts)); i++) {
547*4882a593Smuzhiyun 		opp = get_voltrail_opp(i);
548*4882a593Smuzhiyun 		debug("%d -> ", pv->value[opp]);
549*4882a593Smuzhiyun 
550*4882a593Smuzhiyun 		if (pv->value[opp]) {
551*4882a593Smuzhiyun 			/* Handle non-empty members only */
552*4882a593Smuzhiyun 			pv->value[opp] = optimize_vcore_voltage(pv, opp);
553*4882a593Smuzhiyun      			px = (struct volts *)vcores;
554*4882a593Smuzhiyun 			j = 0;
555*4882a593Smuzhiyun 			while (px < pv) {
556*4882a593Smuzhiyun 				/*
557*4882a593Smuzhiyun 				 * Scan already handled non-empty members to see
558*4882a593Smuzhiyun 				 * if we have a group and find the max voltage,
559*4882a593Smuzhiyun 				 * which is set to the first occurance of the
560*4882a593Smuzhiyun 				 * particular SMPS; the other group voltages are
561*4882a593Smuzhiyun 				 * zeroed.
562*4882a593Smuzhiyun 				 */
563*4882a593Smuzhiyun 				ol = get_voltrail_opp(j);
564*4882a593Smuzhiyun 				if (px->value[ol] &&
565*4882a593Smuzhiyun 				    (pv->pmic->i2c_slave_addr ==
566*4882a593Smuzhiyun 				     px->pmic->i2c_slave_addr) &&
567*4882a593Smuzhiyun 				    (pv->addr == px->addr)) {
568*4882a593Smuzhiyun 					/* Same PMIC, same SMPS */
569*4882a593Smuzhiyun 					if (pv->value[opp] > px->value[ol])
570*4882a593Smuzhiyun 						px->value[ol] = pv->value[opp];
571*4882a593Smuzhiyun 
572*4882a593Smuzhiyun 					pv->value[opp] = 0;
573*4882a593Smuzhiyun 				}
574*4882a593Smuzhiyun 				px++;
575*4882a593Smuzhiyun 				j++;
576*4882a593Smuzhiyun 			}
577*4882a593Smuzhiyun 		}
578*4882a593Smuzhiyun 		debug("%d\n", pv->value[opp]);
579*4882a593Smuzhiyun 		pv++;
580*4882a593Smuzhiyun 	}
581*4882a593Smuzhiyun 
582*4882a593Smuzhiyun 	opp = get_voltrail_opp(VOLT_CORE);
583*4882a593Smuzhiyun 	debug("cor: %d\n", vcores->core.value[opp]);
584*4882a593Smuzhiyun 	do_scale_vcore(vcores->core.addr, vcores->core.value[opp],
585*4882a593Smuzhiyun 		       vcores->core.pmic);
586*4882a593Smuzhiyun 	/*
587*4882a593Smuzhiyun 	 * IO delay recalibration should be done immediately after
588*4882a593Smuzhiyun 	 * adjusting AVS voltages for VDD_CORE_L.
589*4882a593Smuzhiyun 	 * Respective boards should call __recalibrate_iodelay()
590*4882a593Smuzhiyun 	 * with proper mux, virtual and manual mode configurations.
591*4882a593Smuzhiyun 	 */
592*4882a593Smuzhiyun #ifdef CONFIG_IODELAY_RECALIBRATION
593*4882a593Smuzhiyun 	recalibrate_iodelay();
594*4882a593Smuzhiyun #endif
595*4882a593Smuzhiyun 
596*4882a593Smuzhiyun 	opp = get_voltrail_opp(VOLT_MPU);
597*4882a593Smuzhiyun 	debug("mpu: %d\n", vcores->mpu.value[opp]);
598*4882a593Smuzhiyun 	do_scale_vcore(vcores->mpu.addr, vcores->mpu.value[opp],
599*4882a593Smuzhiyun 		       vcores->mpu.pmic);
600*4882a593Smuzhiyun 	/* Configure MPU ABB LDO after scale */
601*4882a593Smuzhiyun 	abb_setup(vcores->mpu.efuse.reg[opp],
602*4882a593Smuzhiyun 		  (*ctrl)->control_wkup_ldovbb_mpu_voltage_ctrl,
603*4882a593Smuzhiyun 		  (*prcm)->prm_abbldo_mpu_setup,
604*4882a593Smuzhiyun 		  (*prcm)->prm_abbldo_mpu_ctrl,
605*4882a593Smuzhiyun 		  (*prcm)->prm_irqstatus_mpu_2,
606*4882a593Smuzhiyun 		  vcores->mpu.abb_tx_done_mask,
607*4882a593Smuzhiyun 		  OMAP_ABB_FAST_OPP);
608*4882a593Smuzhiyun 
609*4882a593Smuzhiyun 	opp = get_voltrail_opp(VOLT_MM);
610*4882a593Smuzhiyun 	debug("mm: %d\n", vcores->mm.value[opp]);
611*4882a593Smuzhiyun 	do_scale_vcore(vcores->mm.addr, vcores->mm.value[opp],
612*4882a593Smuzhiyun 		       vcores->mm.pmic);
613*4882a593Smuzhiyun 	/* Configure MM ABB LDO after scale */
614*4882a593Smuzhiyun 	abb_setup(vcores->mm.efuse.reg[opp],
615*4882a593Smuzhiyun 		  (*ctrl)->control_wkup_ldovbb_mm_voltage_ctrl,
616*4882a593Smuzhiyun 		  (*prcm)->prm_abbldo_mm_setup,
617*4882a593Smuzhiyun 		  (*prcm)->prm_abbldo_mm_ctrl,
618*4882a593Smuzhiyun 		  (*prcm)->prm_irqstatus_mpu,
619*4882a593Smuzhiyun 		  vcores->mm.abb_tx_done_mask,
620*4882a593Smuzhiyun 		  OMAP_ABB_FAST_OPP);
621*4882a593Smuzhiyun 
622*4882a593Smuzhiyun 	opp = get_voltrail_opp(VOLT_GPU);
623*4882a593Smuzhiyun 	debug("gpu: %d\n", vcores->gpu.value[opp]);
624*4882a593Smuzhiyun 	do_scale_vcore(vcores->gpu.addr, vcores->gpu.value[opp],
625*4882a593Smuzhiyun 		       vcores->gpu.pmic);
626*4882a593Smuzhiyun 	/* Configure GPU ABB LDO after scale */
627*4882a593Smuzhiyun 	abb_setup(vcores->gpu.efuse.reg[opp],
628*4882a593Smuzhiyun 		  (*ctrl)->control_wkup_ldovbb_gpu_voltage_ctrl,
629*4882a593Smuzhiyun 		  (*prcm)->prm_abbldo_gpu_setup,
630*4882a593Smuzhiyun 		  (*prcm)->prm_abbldo_gpu_ctrl,
631*4882a593Smuzhiyun 		  (*prcm)->prm_irqstatus_mpu,
632*4882a593Smuzhiyun 		  vcores->gpu.abb_tx_done_mask,
633*4882a593Smuzhiyun 		  OMAP_ABB_FAST_OPP);
634*4882a593Smuzhiyun 
635*4882a593Smuzhiyun 	opp = get_voltrail_opp(VOLT_EVE);
636*4882a593Smuzhiyun 	debug("eve: %d\n", vcores->eve.value[opp]);
637*4882a593Smuzhiyun 	do_scale_vcore(vcores->eve.addr, vcores->eve.value[opp],
638*4882a593Smuzhiyun 		       vcores->eve.pmic);
639*4882a593Smuzhiyun 	/* Configure EVE ABB LDO after scale */
640*4882a593Smuzhiyun 	abb_setup(vcores->eve.efuse.reg[opp],
641*4882a593Smuzhiyun 		  (*ctrl)->control_wkup_ldovbb_eve_voltage_ctrl,
642*4882a593Smuzhiyun 		  (*prcm)->prm_abbldo_eve_setup,
643*4882a593Smuzhiyun 		  (*prcm)->prm_abbldo_eve_ctrl,
644*4882a593Smuzhiyun 		  (*prcm)->prm_irqstatus_mpu,
645*4882a593Smuzhiyun 		  vcores->eve.abb_tx_done_mask,
646*4882a593Smuzhiyun 		  OMAP_ABB_FAST_OPP);
647*4882a593Smuzhiyun 
648*4882a593Smuzhiyun 	opp = get_voltrail_opp(VOLT_IVA);
649*4882a593Smuzhiyun 	debug("iva: %d\n", vcores->iva.value[opp]);
650*4882a593Smuzhiyun 	do_scale_vcore(vcores->iva.addr, vcores->iva.value[opp],
651*4882a593Smuzhiyun 		       vcores->iva.pmic);
652*4882a593Smuzhiyun 	/* Configure IVA ABB LDO after scale */
653*4882a593Smuzhiyun 	abb_setup(vcores->iva.efuse.reg[opp],
654*4882a593Smuzhiyun 		  (*ctrl)->control_wkup_ldovbb_iva_voltage_ctrl,
655*4882a593Smuzhiyun 		  (*prcm)->prm_abbldo_iva_setup,
656*4882a593Smuzhiyun 		  (*prcm)->prm_abbldo_iva_ctrl,
657*4882a593Smuzhiyun 		  (*prcm)->prm_irqstatus_mpu,
658*4882a593Smuzhiyun 		  vcores->iva.abb_tx_done_mask,
659*4882a593Smuzhiyun 		  OMAP_ABB_FAST_OPP);
660*4882a593Smuzhiyun }
661*4882a593Smuzhiyun 
enable_clock_domain(u32 const clkctrl_reg,u32 enable_mode)662*4882a593Smuzhiyun static inline void enable_clock_domain(u32 const clkctrl_reg, u32 enable_mode)
663*4882a593Smuzhiyun {
664*4882a593Smuzhiyun 	clrsetbits_le32(clkctrl_reg, CD_CLKCTRL_CLKTRCTRL_MASK,
665*4882a593Smuzhiyun 			enable_mode << CD_CLKCTRL_CLKTRCTRL_SHIFT);
666*4882a593Smuzhiyun 	debug("Enable clock domain - %x\n", clkctrl_reg);
667*4882a593Smuzhiyun }
668*4882a593Smuzhiyun 
disable_clock_domain(u32 const clkctrl_reg)669*4882a593Smuzhiyun static inline void disable_clock_domain(u32 const clkctrl_reg)
670*4882a593Smuzhiyun {
671*4882a593Smuzhiyun 	clrsetbits_le32(clkctrl_reg, CD_CLKCTRL_CLKTRCTRL_MASK,
672*4882a593Smuzhiyun 			CD_CLKCTRL_CLKTRCTRL_SW_SLEEP <<
673*4882a593Smuzhiyun 			CD_CLKCTRL_CLKTRCTRL_SHIFT);
674*4882a593Smuzhiyun 	debug("Disable clock domain - %x\n", clkctrl_reg);
675*4882a593Smuzhiyun }
676*4882a593Smuzhiyun 
wait_for_clk_enable(u32 clkctrl_addr)677*4882a593Smuzhiyun static inline void wait_for_clk_enable(u32 clkctrl_addr)
678*4882a593Smuzhiyun {
679*4882a593Smuzhiyun 	u32 clkctrl, idlest = MODULE_CLKCTRL_IDLEST_DISABLED;
680*4882a593Smuzhiyun 	u32 bound = LDELAY;
681*4882a593Smuzhiyun 
682*4882a593Smuzhiyun 	while ((idlest == MODULE_CLKCTRL_IDLEST_DISABLED) ||
683*4882a593Smuzhiyun 		(idlest == MODULE_CLKCTRL_IDLEST_TRANSITIONING)) {
684*4882a593Smuzhiyun 
685*4882a593Smuzhiyun 		clkctrl = readl(clkctrl_addr);
686*4882a593Smuzhiyun 		idlest = (clkctrl & MODULE_CLKCTRL_IDLEST_MASK) >>
687*4882a593Smuzhiyun 			 MODULE_CLKCTRL_IDLEST_SHIFT;
688*4882a593Smuzhiyun 		if (--bound == 0) {
689*4882a593Smuzhiyun 			printf("Clock enable failed for 0x%x idlest 0x%x\n",
690*4882a593Smuzhiyun 				clkctrl_addr, clkctrl);
691*4882a593Smuzhiyun 			return;
692*4882a593Smuzhiyun 		}
693*4882a593Smuzhiyun 	}
694*4882a593Smuzhiyun }
695*4882a593Smuzhiyun 
enable_clock_module(u32 const clkctrl_addr,u32 enable_mode,u32 wait_for_enable)696*4882a593Smuzhiyun static inline void enable_clock_module(u32 const clkctrl_addr, u32 enable_mode,
697*4882a593Smuzhiyun 				u32 wait_for_enable)
698*4882a593Smuzhiyun {
699*4882a593Smuzhiyun 	clrsetbits_le32(clkctrl_addr, MODULE_CLKCTRL_MODULEMODE_MASK,
700*4882a593Smuzhiyun 			enable_mode << MODULE_CLKCTRL_MODULEMODE_SHIFT);
701*4882a593Smuzhiyun 	debug("Enable clock module - %x\n", clkctrl_addr);
702*4882a593Smuzhiyun 	if (wait_for_enable)
703*4882a593Smuzhiyun 		wait_for_clk_enable(clkctrl_addr);
704*4882a593Smuzhiyun }
705*4882a593Smuzhiyun 
wait_for_clk_disable(u32 clkctrl_addr)706*4882a593Smuzhiyun static inline void wait_for_clk_disable(u32 clkctrl_addr)
707*4882a593Smuzhiyun {
708*4882a593Smuzhiyun 	u32 clkctrl, idlest = MODULE_CLKCTRL_IDLEST_FULLY_FUNCTIONAL;
709*4882a593Smuzhiyun 	u32 bound = LDELAY;
710*4882a593Smuzhiyun 
711*4882a593Smuzhiyun 	while ((idlest != MODULE_CLKCTRL_IDLEST_DISABLED)) {
712*4882a593Smuzhiyun 		clkctrl = readl(clkctrl_addr);
713*4882a593Smuzhiyun 		idlest = (clkctrl & MODULE_CLKCTRL_IDLEST_MASK) >>
714*4882a593Smuzhiyun 			 MODULE_CLKCTRL_IDLEST_SHIFT;
715*4882a593Smuzhiyun 		if (--bound == 0) {
716*4882a593Smuzhiyun 			printf("Clock disable failed for 0x%x idlest 0x%x\n",
717*4882a593Smuzhiyun 			       clkctrl_addr, clkctrl);
718*4882a593Smuzhiyun 			return;
719*4882a593Smuzhiyun 		}
720*4882a593Smuzhiyun 	}
721*4882a593Smuzhiyun }
722*4882a593Smuzhiyun 
disable_clock_module(u32 const clkctrl_addr,u32 wait_for_disable)723*4882a593Smuzhiyun static inline void disable_clock_module(u32 const clkctrl_addr,
724*4882a593Smuzhiyun 					u32 wait_for_disable)
725*4882a593Smuzhiyun {
726*4882a593Smuzhiyun 	clrsetbits_le32(clkctrl_addr, MODULE_CLKCTRL_MODULEMODE_MASK,
727*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_SW_DISABLE <<
728*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_SHIFT);
729*4882a593Smuzhiyun 	debug("Disable clock module - %x\n", clkctrl_addr);
730*4882a593Smuzhiyun 	if (wait_for_disable)
731*4882a593Smuzhiyun 		wait_for_clk_disable(clkctrl_addr);
732*4882a593Smuzhiyun }
733*4882a593Smuzhiyun 
freq_update_core(void)734*4882a593Smuzhiyun void freq_update_core(void)
735*4882a593Smuzhiyun {
736*4882a593Smuzhiyun 	u32 freq_config1 = 0;
737*4882a593Smuzhiyun 	const struct dpll_params *core_dpll_params;
738*4882a593Smuzhiyun 	u32 omap_rev = omap_revision();
739*4882a593Smuzhiyun 
740*4882a593Smuzhiyun 	core_dpll_params = get_core_dpll_params(*dplls_data);
741*4882a593Smuzhiyun 	/* Put EMIF clock domain in sw wakeup mode */
742*4882a593Smuzhiyun 	enable_clock_domain((*prcm)->cm_memif_clkstctrl,
743*4882a593Smuzhiyun 				CD_CLKCTRL_CLKTRCTRL_SW_WKUP);
744*4882a593Smuzhiyun 	wait_for_clk_enable((*prcm)->cm_memif_emif_1_clkctrl);
745*4882a593Smuzhiyun 	wait_for_clk_enable((*prcm)->cm_memif_emif_2_clkctrl);
746*4882a593Smuzhiyun 
747*4882a593Smuzhiyun 	freq_config1 = SHADOW_FREQ_CONFIG1_FREQ_UPDATE_MASK |
748*4882a593Smuzhiyun 	    SHADOW_FREQ_CONFIG1_DLL_RESET_MASK;
749*4882a593Smuzhiyun 
750*4882a593Smuzhiyun 	freq_config1 |= (DPLL_EN_LOCK << SHADOW_FREQ_CONFIG1_DPLL_EN_SHIFT) &
751*4882a593Smuzhiyun 				SHADOW_FREQ_CONFIG1_DPLL_EN_MASK;
752*4882a593Smuzhiyun 
753*4882a593Smuzhiyun 	freq_config1 |= (core_dpll_params->m2 <<
754*4882a593Smuzhiyun 			SHADOW_FREQ_CONFIG1_M2_DIV_SHIFT) &
755*4882a593Smuzhiyun 			SHADOW_FREQ_CONFIG1_M2_DIV_MASK;
756*4882a593Smuzhiyun 
757*4882a593Smuzhiyun 	writel(freq_config1, (*prcm)->cm_shadow_freq_config1);
758*4882a593Smuzhiyun 	if (!wait_on_value(SHADOW_FREQ_CONFIG1_FREQ_UPDATE_MASK, 0,
759*4882a593Smuzhiyun 			(u32 *) (*prcm)->cm_shadow_freq_config1, LDELAY)) {
760*4882a593Smuzhiyun 		puts("FREQ UPDATE procedure failed!!");
761*4882a593Smuzhiyun 		hang();
762*4882a593Smuzhiyun 	}
763*4882a593Smuzhiyun 
764*4882a593Smuzhiyun 	/*
765*4882a593Smuzhiyun 	 * Putting EMIF in HW_AUTO is seen to be causing issues with
766*4882a593Smuzhiyun 	 * EMIF clocks and the master DLL. Keep EMIF in SW_WKUP
767*4882a593Smuzhiyun 	 * in OMAP5430 ES1.0 silicon
768*4882a593Smuzhiyun 	 */
769*4882a593Smuzhiyun 	if (omap_rev != OMAP5430_ES1_0) {
770*4882a593Smuzhiyun 		/* Put EMIF clock domain back in hw auto mode */
771*4882a593Smuzhiyun 		enable_clock_domain((*prcm)->cm_memif_clkstctrl,
772*4882a593Smuzhiyun 					CD_CLKCTRL_CLKTRCTRL_HW_AUTO);
773*4882a593Smuzhiyun 		wait_for_clk_enable((*prcm)->cm_memif_emif_1_clkctrl);
774*4882a593Smuzhiyun 		wait_for_clk_enable((*prcm)->cm_memif_emif_2_clkctrl);
775*4882a593Smuzhiyun 	}
776*4882a593Smuzhiyun }
777*4882a593Smuzhiyun 
bypass_dpll(u32 const base)778*4882a593Smuzhiyun void bypass_dpll(u32 const base)
779*4882a593Smuzhiyun {
780*4882a593Smuzhiyun 	do_bypass_dpll(base);
781*4882a593Smuzhiyun 	wait_for_bypass(base);
782*4882a593Smuzhiyun }
783*4882a593Smuzhiyun 
lock_dpll(u32 const base)784*4882a593Smuzhiyun void lock_dpll(u32 const base)
785*4882a593Smuzhiyun {
786*4882a593Smuzhiyun 	do_lock_dpll(base);
787*4882a593Smuzhiyun 	wait_for_lock(base);
788*4882a593Smuzhiyun }
789*4882a593Smuzhiyun 
setup_clocks_for_console(void)790*4882a593Smuzhiyun static void setup_clocks_for_console(void)
791*4882a593Smuzhiyun {
792*4882a593Smuzhiyun 	/* Do not add any spl_debug prints in this function */
793*4882a593Smuzhiyun 	clrsetbits_le32((*prcm)->cm_l4per_clkstctrl, CD_CLKCTRL_CLKTRCTRL_MASK,
794*4882a593Smuzhiyun 			CD_CLKCTRL_CLKTRCTRL_SW_WKUP <<
795*4882a593Smuzhiyun 			CD_CLKCTRL_CLKTRCTRL_SHIFT);
796*4882a593Smuzhiyun 
797*4882a593Smuzhiyun 	/* Enable all UARTs - console will be on one of them */
798*4882a593Smuzhiyun 	clrsetbits_le32((*prcm)->cm_l4per_uart1_clkctrl,
799*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_MASK,
800*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_SW_EXPLICIT_EN <<
801*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_SHIFT);
802*4882a593Smuzhiyun 
803*4882a593Smuzhiyun 	clrsetbits_le32((*prcm)->cm_l4per_uart2_clkctrl,
804*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_MASK,
805*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_SW_EXPLICIT_EN <<
806*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_SHIFT);
807*4882a593Smuzhiyun 
808*4882a593Smuzhiyun 	clrsetbits_le32((*prcm)->cm_l4per_uart3_clkctrl,
809*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_MASK,
810*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_SW_EXPLICIT_EN <<
811*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_SHIFT);
812*4882a593Smuzhiyun 
813*4882a593Smuzhiyun 	clrsetbits_le32((*prcm)->cm_l4per_uart4_clkctrl,
814*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_MASK,
815*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_SW_EXPLICIT_EN <<
816*4882a593Smuzhiyun 			MODULE_CLKCTRL_MODULEMODE_SHIFT);
817*4882a593Smuzhiyun 
818*4882a593Smuzhiyun 	clrsetbits_le32((*prcm)->cm_l4per_clkstctrl, CD_CLKCTRL_CLKTRCTRL_MASK,
819*4882a593Smuzhiyun 			CD_CLKCTRL_CLKTRCTRL_HW_AUTO <<
820*4882a593Smuzhiyun 			CD_CLKCTRL_CLKTRCTRL_SHIFT);
821*4882a593Smuzhiyun }
822*4882a593Smuzhiyun 
do_enable_clocks(u32 const * clk_domains,u32 const * clk_modules_hw_auto,u32 const * clk_modules_explicit_en,u8 wait_for_enable)823*4882a593Smuzhiyun void do_enable_clocks(u32 const *clk_domains,
824*4882a593Smuzhiyun 			    u32 const *clk_modules_hw_auto,
825*4882a593Smuzhiyun 			    u32 const *clk_modules_explicit_en,
826*4882a593Smuzhiyun 			    u8 wait_for_enable)
827*4882a593Smuzhiyun {
828*4882a593Smuzhiyun 	u32 i, max = 100;
829*4882a593Smuzhiyun 
830*4882a593Smuzhiyun 	/* Put the clock domains in SW_WKUP mode */
831*4882a593Smuzhiyun 	for (i = 0; (i < max) && clk_domains && clk_domains[i]; i++) {
832*4882a593Smuzhiyun 		enable_clock_domain(clk_domains[i],
833*4882a593Smuzhiyun 				    CD_CLKCTRL_CLKTRCTRL_SW_WKUP);
834*4882a593Smuzhiyun 	}
835*4882a593Smuzhiyun 
836*4882a593Smuzhiyun 	/* Clock modules that need to be put in HW_AUTO */
837*4882a593Smuzhiyun 	for (i = 0; (i < max) && clk_modules_hw_auto &&
838*4882a593Smuzhiyun 		     clk_modules_hw_auto[i]; i++) {
839*4882a593Smuzhiyun 		enable_clock_module(clk_modules_hw_auto[i],
840*4882a593Smuzhiyun 				    MODULE_CLKCTRL_MODULEMODE_HW_AUTO,
841*4882a593Smuzhiyun 				    wait_for_enable);
842*4882a593Smuzhiyun 	};
843*4882a593Smuzhiyun 
844*4882a593Smuzhiyun 	/* Clock modules that need to be put in SW_EXPLICIT_EN mode */
845*4882a593Smuzhiyun 	for (i = 0; (i < max) && clk_modules_explicit_en &&
846*4882a593Smuzhiyun 		     clk_modules_explicit_en[i]; i++) {
847*4882a593Smuzhiyun 		enable_clock_module(clk_modules_explicit_en[i],
848*4882a593Smuzhiyun 				    MODULE_CLKCTRL_MODULEMODE_SW_EXPLICIT_EN,
849*4882a593Smuzhiyun 				    wait_for_enable);
850*4882a593Smuzhiyun 	};
851*4882a593Smuzhiyun 
852*4882a593Smuzhiyun 	/* Put the clock domains in HW_AUTO mode now */
853*4882a593Smuzhiyun 	for (i = 0; (i < max) && clk_domains && clk_domains[i]; i++) {
854*4882a593Smuzhiyun 		enable_clock_domain(clk_domains[i],
855*4882a593Smuzhiyun 				    CD_CLKCTRL_CLKTRCTRL_HW_AUTO);
856*4882a593Smuzhiyun 	}
857*4882a593Smuzhiyun }
858*4882a593Smuzhiyun 
do_disable_clocks(u32 const * clk_domains,u32 const * clk_modules_disable,u8 wait_for_disable)859*4882a593Smuzhiyun void do_disable_clocks(u32 const *clk_domains,
860*4882a593Smuzhiyun 			    u32 const *clk_modules_disable,
861*4882a593Smuzhiyun 			    u8 wait_for_disable)
862*4882a593Smuzhiyun {
863*4882a593Smuzhiyun 	u32 i, max = 100;
864*4882a593Smuzhiyun 
865*4882a593Smuzhiyun 
866*4882a593Smuzhiyun 	/* Clock modules that need to be put in SW_DISABLE */
867*4882a593Smuzhiyun 	for (i = 0; (i < max) && clk_modules_disable[i]; i++)
868*4882a593Smuzhiyun 		disable_clock_module(clk_modules_disable[i],
869*4882a593Smuzhiyun 				     wait_for_disable);
870*4882a593Smuzhiyun 
871*4882a593Smuzhiyun 	/* Put the clock domains in SW_SLEEP mode */
872*4882a593Smuzhiyun 	for (i = 0; (i < max) && clk_domains[i]; i++)
873*4882a593Smuzhiyun 		disable_clock_domain(clk_domains[i]);
874*4882a593Smuzhiyun }
875*4882a593Smuzhiyun 
876*4882a593Smuzhiyun /**
877*4882a593Smuzhiyun  * setup_early_clocks() - Setup early clocks needed for SoC
878*4882a593Smuzhiyun  *
879*4882a593Smuzhiyun  * Setup clocks for console, SPL basic initialization clocks and initialize
880*4882a593Smuzhiyun  * the timer. This is invoked prior prcm_init.
881*4882a593Smuzhiyun  */
setup_early_clocks(void)882*4882a593Smuzhiyun void setup_early_clocks(void)
883*4882a593Smuzhiyun {
884*4882a593Smuzhiyun 	switch (omap_hw_init_context()) {
885*4882a593Smuzhiyun 	case OMAP_INIT_CONTEXT_SPL:
886*4882a593Smuzhiyun 	case OMAP_INIT_CONTEXT_UBOOT_FROM_NOR:
887*4882a593Smuzhiyun 	case OMAP_INIT_CONTEXT_UBOOT_AFTER_CH:
888*4882a593Smuzhiyun 		setup_clocks_for_console();
889*4882a593Smuzhiyun 		enable_basic_clocks();
890*4882a593Smuzhiyun 		timer_init();
891*4882a593Smuzhiyun 		/* Fall through */
892*4882a593Smuzhiyun 	}
893*4882a593Smuzhiyun }
894*4882a593Smuzhiyun 
prcm_init(void)895*4882a593Smuzhiyun void prcm_init(void)
896*4882a593Smuzhiyun {
897*4882a593Smuzhiyun 	switch (omap_hw_init_context()) {
898*4882a593Smuzhiyun 	case OMAP_INIT_CONTEXT_SPL:
899*4882a593Smuzhiyun 	case OMAP_INIT_CONTEXT_UBOOT_FROM_NOR:
900*4882a593Smuzhiyun 	case OMAP_INIT_CONTEXT_UBOOT_AFTER_CH:
901*4882a593Smuzhiyun 		scale_vcores(*omap_vcores);
902*4882a593Smuzhiyun 		setup_dplls();
903*4882a593Smuzhiyun 		setup_warmreset_time();
904*4882a593Smuzhiyun 		break;
905*4882a593Smuzhiyun 	default:
906*4882a593Smuzhiyun 		break;
907*4882a593Smuzhiyun 	}
908*4882a593Smuzhiyun 
909*4882a593Smuzhiyun 	if (OMAP_INIT_CONTEXT_SPL != omap_hw_init_context())
910*4882a593Smuzhiyun 		enable_basic_uboot_clocks();
911*4882a593Smuzhiyun }
912*4882a593Smuzhiyun 
gpi2c_init(void)913*4882a593Smuzhiyun void gpi2c_init(void)
914*4882a593Smuzhiyun {
915*4882a593Smuzhiyun 	static int gpi2c = 1;
916*4882a593Smuzhiyun 
917*4882a593Smuzhiyun 	if (gpi2c) {
918*4882a593Smuzhiyun 		i2c_init(CONFIG_SYS_OMAP24_I2C_SPEED,
919*4882a593Smuzhiyun 			 CONFIG_SYS_OMAP24_I2C_SLAVE);
920*4882a593Smuzhiyun 		gpi2c = 0;
921*4882a593Smuzhiyun 	}
922*4882a593Smuzhiyun }
923