xref: /OK3568_Linux_fs/kernel/init/calibrate.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun /* calibrate.c: default delay calibration
3*4882a593Smuzhiyun  *
4*4882a593Smuzhiyun  * Excised from init/main.c
5*4882a593Smuzhiyun  *  Copyright (C) 1991, 1992  Linus Torvalds
6*4882a593Smuzhiyun  */
7*4882a593Smuzhiyun 
8*4882a593Smuzhiyun #include <linux/jiffies.h>
9*4882a593Smuzhiyun #include <linux/delay.h>
10*4882a593Smuzhiyun #include <linux/init.h>
11*4882a593Smuzhiyun #include <linux/timex.h>
12*4882a593Smuzhiyun #include <linux/smp.h>
13*4882a593Smuzhiyun #include <linux/percpu.h>
14*4882a593Smuzhiyun 
15*4882a593Smuzhiyun unsigned long lpj_fine;
16*4882a593Smuzhiyun unsigned long preset_lpj;
lpj_setup(char * str)17*4882a593Smuzhiyun static int __init lpj_setup(char *str)
18*4882a593Smuzhiyun {
19*4882a593Smuzhiyun 	preset_lpj = simple_strtoul(str,NULL,0);
20*4882a593Smuzhiyun 	return 1;
21*4882a593Smuzhiyun }
22*4882a593Smuzhiyun 
23*4882a593Smuzhiyun __setup("lpj=", lpj_setup);
24*4882a593Smuzhiyun 
25*4882a593Smuzhiyun #ifdef ARCH_HAS_READ_CURRENT_TIMER
26*4882a593Smuzhiyun 
27*4882a593Smuzhiyun /* This routine uses the read_current_timer() routine and gets the
28*4882a593Smuzhiyun  * loops per jiffy directly, instead of guessing it using delay().
29*4882a593Smuzhiyun  * Also, this code tries to handle non-maskable asynchronous events
30*4882a593Smuzhiyun  * (like SMIs)
31*4882a593Smuzhiyun  */
32*4882a593Smuzhiyun #define DELAY_CALIBRATION_TICKS			((HZ < 100) ? 1 : (HZ/100))
33*4882a593Smuzhiyun #define MAX_DIRECT_CALIBRATION_RETRIES		5
34*4882a593Smuzhiyun 
calibrate_delay_direct(void)35*4882a593Smuzhiyun static unsigned long calibrate_delay_direct(void)
36*4882a593Smuzhiyun {
37*4882a593Smuzhiyun 	unsigned long pre_start, start, post_start;
38*4882a593Smuzhiyun 	unsigned long pre_end, end, post_end;
39*4882a593Smuzhiyun 	unsigned long start_jiffies;
40*4882a593Smuzhiyun 	unsigned long timer_rate_min, timer_rate_max;
41*4882a593Smuzhiyun 	unsigned long good_timer_sum = 0;
42*4882a593Smuzhiyun 	unsigned long good_timer_count = 0;
43*4882a593Smuzhiyun 	unsigned long measured_times[MAX_DIRECT_CALIBRATION_RETRIES];
44*4882a593Smuzhiyun 	int max = -1; /* index of measured_times with max/min values or not set */
45*4882a593Smuzhiyun 	int min = -1;
46*4882a593Smuzhiyun 	int i;
47*4882a593Smuzhiyun 
48*4882a593Smuzhiyun 	if (read_current_timer(&pre_start) < 0 )
49*4882a593Smuzhiyun 		return 0;
50*4882a593Smuzhiyun 
51*4882a593Smuzhiyun 	/*
52*4882a593Smuzhiyun 	 * A simple loop like
53*4882a593Smuzhiyun 	 *	while ( jiffies < start_jiffies+1)
54*4882a593Smuzhiyun 	 *		start = read_current_timer();
55*4882a593Smuzhiyun 	 * will not do. As we don't really know whether jiffy switch
56*4882a593Smuzhiyun 	 * happened first or timer_value was read first. And some asynchronous
57*4882a593Smuzhiyun 	 * event can happen between these two events introducing errors in lpj.
58*4882a593Smuzhiyun 	 *
59*4882a593Smuzhiyun 	 * So, we do
60*4882a593Smuzhiyun 	 * 1. pre_start <- When we are sure that jiffy switch hasn't happened
61*4882a593Smuzhiyun 	 * 2. check jiffy switch
62*4882a593Smuzhiyun 	 * 3. start <- timer value before or after jiffy switch
63*4882a593Smuzhiyun 	 * 4. post_start <- When we are sure that jiffy switch has happened
64*4882a593Smuzhiyun 	 *
65*4882a593Smuzhiyun 	 * Note, we don't know anything about order of 2 and 3.
66*4882a593Smuzhiyun 	 * Now, by looking at post_start and pre_start difference, we can
67*4882a593Smuzhiyun 	 * check whether any asynchronous event happened or not
68*4882a593Smuzhiyun 	 */
69*4882a593Smuzhiyun 
70*4882a593Smuzhiyun 	for (i = 0; i < MAX_DIRECT_CALIBRATION_RETRIES; i++) {
71*4882a593Smuzhiyun 		pre_start = 0;
72*4882a593Smuzhiyun 		read_current_timer(&start);
73*4882a593Smuzhiyun 		start_jiffies = jiffies;
74*4882a593Smuzhiyun 		while (time_before_eq(jiffies, start_jiffies + 1)) {
75*4882a593Smuzhiyun 			pre_start = start;
76*4882a593Smuzhiyun 			read_current_timer(&start);
77*4882a593Smuzhiyun 		}
78*4882a593Smuzhiyun 		read_current_timer(&post_start);
79*4882a593Smuzhiyun 
80*4882a593Smuzhiyun 		pre_end = 0;
81*4882a593Smuzhiyun 		end = post_start;
82*4882a593Smuzhiyun 		while (time_before_eq(jiffies, start_jiffies + 1 +
83*4882a593Smuzhiyun 					       DELAY_CALIBRATION_TICKS)) {
84*4882a593Smuzhiyun 			pre_end = end;
85*4882a593Smuzhiyun 			read_current_timer(&end);
86*4882a593Smuzhiyun 		}
87*4882a593Smuzhiyun 		read_current_timer(&post_end);
88*4882a593Smuzhiyun 
89*4882a593Smuzhiyun 		timer_rate_max = (post_end - pre_start) /
90*4882a593Smuzhiyun 					DELAY_CALIBRATION_TICKS;
91*4882a593Smuzhiyun 		timer_rate_min = (pre_end - post_start) /
92*4882a593Smuzhiyun 					DELAY_CALIBRATION_TICKS;
93*4882a593Smuzhiyun 
94*4882a593Smuzhiyun 		/*
95*4882a593Smuzhiyun 		 * If the upper limit and lower limit of the timer_rate is
96*4882a593Smuzhiyun 		 * >= 12.5% apart, redo calibration.
97*4882a593Smuzhiyun 		 */
98*4882a593Smuzhiyun 		if (start >= post_end)
99*4882a593Smuzhiyun 			printk(KERN_NOTICE "calibrate_delay_direct() ignoring "
100*4882a593Smuzhiyun 					"timer_rate as we had a TSC wrap around"
101*4882a593Smuzhiyun 					" start=%lu >=post_end=%lu\n",
102*4882a593Smuzhiyun 				start, post_end);
103*4882a593Smuzhiyun 		if (start < post_end && pre_start != 0 && pre_end != 0 &&
104*4882a593Smuzhiyun 		    (timer_rate_max - timer_rate_min) < (timer_rate_max >> 3)) {
105*4882a593Smuzhiyun 			good_timer_count++;
106*4882a593Smuzhiyun 			good_timer_sum += timer_rate_max;
107*4882a593Smuzhiyun 			measured_times[i] = timer_rate_max;
108*4882a593Smuzhiyun 			if (max < 0 || timer_rate_max > measured_times[max])
109*4882a593Smuzhiyun 				max = i;
110*4882a593Smuzhiyun 			if (min < 0 || timer_rate_max < measured_times[min])
111*4882a593Smuzhiyun 				min = i;
112*4882a593Smuzhiyun 		} else
113*4882a593Smuzhiyun 			measured_times[i] = 0;
114*4882a593Smuzhiyun 
115*4882a593Smuzhiyun 	}
116*4882a593Smuzhiyun 
117*4882a593Smuzhiyun 	/*
118*4882a593Smuzhiyun 	 * Find the maximum & minimum - if they differ too much throw out the
119*4882a593Smuzhiyun 	 * one with the largest difference from the mean and try again...
120*4882a593Smuzhiyun 	 */
121*4882a593Smuzhiyun 	while (good_timer_count > 1) {
122*4882a593Smuzhiyun 		unsigned long estimate;
123*4882a593Smuzhiyun 		unsigned long maxdiff;
124*4882a593Smuzhiyun 
125*4882a593Smuzhiyun 		/* compute the estimate */
126*4882a593Smuzhiyun 		estimate = (good_timer_sum/good_timer_count);
127*4882a593Smuzhiyun 		maxdiff = estimate >> 3;
128*4882a593Smuzhiyun 
129*4882a593Smuzhiyun 		/* if range is within 12% let's take it */
130*4882a593Smuzhiyun 		if ((measured_times[max] - measured_times[min]) < maxdiff)
131*4882a593Smuzhiyun 			return estimate;
132*4882a593Smuzhiyun 
133*4882a593Smuzhiyun 		/* ok - drop the worse value and try again... */
134*4882a593Smuzhiyun 		good_timer_sum = 0;
135*4882a593Smuzhiyun 		good_timer_count = 0;
136*4882a593Smuzhiyun 		if ((measured_times[max] - estimate) <
137*4882a593Smuzhiyun 				(estimate - measured_times[min])) {
138*4882a593Smuzhiyun 			printk(KERN_NOTICE "calibrate_delay_direct() dropping "
139*4882a593Smuzhiyun 					"min bogoMips estimate %d = %lu\n",
140*4882a593Smuzhiyun 				min, measured_times[min]);
141*4882a593Smuzhiyun 			measured_times[min] = 0;
142*4882a593Smuzhiyun 			min = max;
143*4882a593Smuzhiyun 		} else {
144*4882a593Smuzhiyun 			printk(KERN_NOTICE "calibrate_delay_direct() dropping "
145*4882a593Smuzhiyun 					"max bogoMips estimate %d = %lu\n",
146*4882a593Smuzhiyun 				max, measured_times[max]);
147*4882a593Smuzhiyun 			measured_times[max] = 0;
148*4882a593Smuzhiyun 			max = min;
149*4882a593Smuzhiyun 		}
150*4882a593Smuzhiyun 
151*4882a593Smuzhiyun 		for (i = 0; i < MAX_DIRECT_CALIBRATION_RETRIES; i++) {
152*4882a593Smuzhiyun 			if (measured_times[i] == 0)
153*4882a593Smuzhiyun 				continue;
154*4882a593Smuzhiyun 			good_timer_count++;
155*4882a593Smuzhiyun 			good_timer_sum += measured_times[i];
156*4882a593Smuzhiyun 			if (measured_times[i] < measured_times[min])
157*4882a593Smuzhiyun 				min = i;
158*4882a593Smuzhiyun 			if (measured_times[i] > measured_times[max])
159*4882a593Smuzhiyun 				max = i;
160*4882a593Smuzhiyun 		}
161*4882a593Smuzhiyun 
162*4882a593Smuzhiyun 	}
163*4882a593Smuzhiyun 
164*4882a593Smuzhiyun 	printk(KERN_NOTICE "calibrate_delay_direct() failed to get a good "
165*4882a593Smuzhiyun 	       "estimate for loops_per_jiffy.\nProbably due to long platform "
166*4882a593Smuzhiyun 		"interrupts. Consider using \"lpj=\" boot option.\n");
167*4882a593Smuzhiyun 	return 0;
168*4882a593Smuzhiyun }
169*4882a593Smuzhiyun #else
calibrate_delay_direct(void)170*4882a593Smuzhiyun static unsigned long calibrate_delay_direct(void)
171*4882a593Smuzhiyun {
172*4882a593Smuzhiyun 	return 0;
173*4882a593Smuzhiyun }
174*4882a593Smuzhiyun #endif
175*4882a593Smuzhiyun 
176*4882a593Smuzhiyun /*
177*4882a593Smuzhiyun  * This is the number of bits of precision for the loops_per_jiffy.  Each
178*4882a593Smuzhiyun  * time we refine our estimate after the first takes 1.5/HZ seconds, so try
179*4882a593Smuzhiyun  * to start with a good estimate.
180*4882a593Smuzhiyun  * For the boot cpu we can skip the delay calibration and assign it a value
181*4882a593Smuzhiyun  * calculated based on the timer frequency.
182*4882a593Smuzhiyun  * For the rest of the CPUs we cannot assume that the timer frequency is same as
183*4882a593Smuzhiyun  * the cpu frequency, hence do the calibration for those.
184*4882a593Smuzhiyun  */
185*4882a593Smuzhiyun #define LPS_PREC 8
186*4882a593Smuzhiyun 
calibrate_delay_converge(void)187*4882a593Smuzhiyun static unsigned long calibrate_delay_converge(void)
188*4882a593Smuzhiyun {
189*4882a593Smuzhiyun 	/* First stage - slowly accelerate to find initial bounds */
190*4882a593Smuzhiyun 	unsigned long lpj, lpj_base, ticks, loopadd, loopadd_base, chop_limit;
191*4882a593Smuzhiyun 	int trials = 0, band = 0, trial_in_band = 0;
192*4882a593Smuzhiyun 
193*4882a593Smuzhiyun 	lpj = (1<<12);
194*4882a593Smuzhiyun 
195*4882a593Smuzhiyun 	/* wait for "start of" clock tick */
196*4882a593Smuzhiyun 	ticks = jiffies;
197*4882a593Smuzhiyun 	while (ticks == jiffies)
198*4882a593Smuzhiyun 		; /* nothing */
199*4882a593Smuzhiyun 	/* Go .. */
200*4882a593Smuzhiyun 	ticks = jiffies;
201*4882a593Smuzhiyun 	do {
202*4882a593Smuzhiyun 		if (++trial_in_band == (1<<band)) {
203*4882a593Smuzhiyun 			++band;
204*4882a593Smuzhiyun 			trial_in_band = 0;
205*4882a593Smuzhiyun 		}
206*4882a593Smuzhiyun 		__delay(lpj * band);
207*4882a593Smuzhiyun 		trials += band;
208*4882a593Smuzhiyun 	} while (ticks == jiffies);
209*4882a593Smuzhiyun 	/*
210*4882a593Smuzhiyun 	 * We overshot, so retreat to a clear underestimate. Then estimate
211*4882a593Smuzhiyun 	 * the largest likely undershoot. This defines our chop bounds.
212*4882a593Smuzhiyun 	 */
213*4882a593Smuzhiyun 	trials -= band;
214*4882a593Smuzhiyun 	loopadd_base = lpj * band;
215*4882a593Smuzhiyun 	lpj_base = lpj * trials;
216*4882a593Smuzhiyun 
217*4882a593Smuzhiyun recalibrate:
218*4882a593Smuzhiyun 	lpj = lpj_base;
219*4882a593Smuzhiyun 	loopadd = loopadd_base;
220*4882a593Smuzhiyun 
221*4882a593Smuzhiyun 	/*
222*4882a593Smuzhiyun 	 * Do a binary approximation to get lpj set to
223*4882a593Smuzhiyun 	 * equal one clock (up to LPS_PREC bits)
224*4882a593Smuzhiyun 	 */
225*4882a593Smuzhiyun 	chop_limit = lpj >> LPS_PREC;
226*4882a593Smuzhiyun 	while (loopadd > chop_limit) {
227*4882a593Smuzhiyun 		lpj += loopadd;
228*4882a593Smuzhiyun 		ticks = jiffies;
229*4882a593Smuzhiyun 		while (ticks == jiffies)
230*4882a593Smuzhiyun 			; /* nothing */
231*4882a593Smuzhiyun 		ticks = jiffies;
232*4882a593Smuzhiyun 		__delay(lpj);
233*4882a593Smuzhiyun 		if (jiffies != ticks)	/* longer than 1 tick */
234*4882a593Smuzhiyun 			lpj -= loopadd;
235*4882a593Smuzhiyun 		loopadd >>= 1;
236*4882a593Smuzhiyun 	}
237*4882a593Smuzhiyun 	/*
238*4882a593Smuzhiyun 	 * If we incremented every single time possible, presume we've
239*4882a593Smuzhiyun 	 * massively underestimated initially, and retry with a higher
240*4882a593Smuzhiyun 	 * start, and larger range. (Only seen on x86_64, due to SMIs)
241*4882a593Smuzhiyun 	 */
242*4882a593Smuzhiyun 	if (lpj + loopadd * 2 == lpj_base + loopadd_base * 2) {
243*4882a593Smuzhiyun 		lpj_base = lpj;
244*4882a593Smuzhiyun 		loopadd_base <<= 2;
245*4882a593Smuzhiyun 		goto recalibrate;
246*4882a593Smuzhiyun 	}
247*4882a593Smuzhiyun 
248*4882a593Smuzhiyun 	return lpj;
249*4882a593Smuzhiyun }
250*4882a593Smuzhiyun 
251*4882a593Smuzhiyun static DEFINE_PER_CPU(unsigned long, cpu_loops_per_jiffy) = { 0 };
252*4882a593Smuzhiyun 
253*4882a593Smuzhiyun /*
254*4882a593Smuzhiyun  * Check if cpu calibration delay is already known. For example,
255*4882a593Smuzhiyun  * some processors with multi-core sockets may have all cores
256*4882a593Smuzhiyun  * with the same calibration delay.
257*4882a593Smuzhiyun  *
258*4882a593Smuzhiyun  * Architectures should override this function if a faster calibration
259*4882a593Smuzhiyun  * method is available.
260*4882a593Smuzhiyun  */
calibrate_delay_is_known(void)261*4882a593Smuzhiyun unsigned long __attribute__((weak)) calibrate_delay_is_known(void)
262*4882a593Smuzhiyun {
263*4882a593Smuzhiyun 	return 0;
264*4882a593Smuzhiyun }
265*4882a593Smuzhiyun 
266*4882a593Smuzhiyun /*
267*4882a593Smuzhiyun  * Indicate the cpu delay calibration is done. This can be used by
268*4882a593Smuzhiyun  * architectures to stop accepting delay timer registrations after this point.
269*4882a593Smuzhiyun  */
270*4882a593Smuzhiyun 
calibration_delay_done(void)271*4882a593Smuzhiyun void __attribute__((weak)) calibration_delay_done(void)
272*4882a593Smuzhiyun {
273*4882a593Smuzhiyun }
274*4882a593Smuzhiyun 
calibrate_delay(void)275*4882a593Smuzhiyun void calibrate_delay(void)
276*4882a593Smuzhiyun {
277*4882a593Smuzhiyun 	unsigned long lpj;
278*4882a593Smuzhiyun 	static bool printed;
279*4882a593Smuzhiyun 	int this_cpu = smp_processor_id();
280*4882a593Smuzhiyun 
281*4882a593Smuzhiyun 	if (per_cpu(cpu_loops_per_jiffy, this_cpu)) {
282*4882a593Smuzhiyun 		lpj = per_cpu(cpu_loops_per_jiffy, this_cpu);
283*4882a593Smuzhiyun 		if (!printed)
284*4882a593Smuzhiyun 			pr_info("Calibrating delay loop (skipped) "
285*4882a593Smuzhiyun 				"already calibrated this CPU");
286*4882a593Smuzhiyun 	} else if (preset_lpj) {
287*4882a593Smuzhiyun 		lpj = preset_lpj;
288*4882a593Smuzhiyun 		if (!printed)
289*4882a593Smuzhiyun 			pr_info("Calibrating delay loop (skipped) "
290*4882a593Smuzhiyun 				"preset value.. ");
291*4882a593Smuzhiyun 	} else if ((!printed) && lpj_fine) {
292*4882a593Smuzhiyun 		lpj = lpj_fine;
293*4882a593Smuzhiyun 		pr_info("Calibrating delay loop (skipped), "
294*4882a593Smuzhiyun 			"value calculated using timer frequency.. ");
295*4882a593Smuzhiyun 	} else if ((lpj = calibrate_delay_is_known())) {
296*4882a593Smuzhiyun 		;
297*4882a593Smuzhiyun 	} else if ((lpj = calibrate_delay_direct()) != 0) {
298*4882a593Smuzhiyun 		if (!printed)
299*4882a593Smuzhiyun 			pr_info("Calibrating delay using timer "
300*4882a593Smuzhiyun 				"specific routine.. ");
301*4882a593Smuzhiyun 	} else {
302*4882a593Smuzhiyun 		if (!printed)
303*4882a593Smuzhiyun 			pr_info("Calibrating delay loop... ");
304*4882a593Smuzhiyun 		lpj = calibrate_delay_converge();
305*4882a593Smuzhiyun 	}
306*4882a593Smuzhiyun 	per_cpu(cpu_loops_per_jiffy, this_cpu) = lpj;
307*4882a593Smuzhiyun 	if (!printed)
308*4882a593Smuzhiyun 		pr_cont("%lu.%02lu BogoMIPS (lpj=%lu)\n",
309*4882a593Smuzhiyun 			lpj/(500000/HZ),
310*4882a593Smuzhiyun 			(lpj/(5000/HZ)) % 100, lpj);
311*4882a593Smuzhiyun 
312*4882a593Smuzhiyun 	loops_per_jiffy = lpj;
313*4882a593Smuzhiyun 	printed = true;
314*4882a593Smuzhiyun 
315*4882a593Smuzhiyun 	calibration_delay_done();
316*4882a593Smuzhiyun }
317