xref: /OK3568_Linux_fs/kernel/drivers/cpufreq/p4-clockmod.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-or-later
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  *	Pentium 4/Xeon CPU on demand clock modulation/speed scaling
4*4882a593Smuzhiyun  *	(C) 2002 - 2003 Dominik Brodowski <linux@brodo.de>
5*4882a593Smuzhiyun  *	(C) 2002 Zwane Mwaikambo <zwane@commfireservices.com>
6*4882a593Smuzhiyun  *	(C) 2002 Arjan van de Ven <arjanv@redhat.com>
7*4882a593Smuzhiyun  *	(C) 2002 Tora T. Engstad
8*4882a593Smuzhiyun  *	All Rights Reserved
9*4882a593Smuzhiyun  *
10*4882a593Smuzhiyun  *      The author(s) of this software shall not be held liable for damages
11*4882a593Smuzhiyun  *      of any nature resulting due to the use of this software. This
12*4882a593Smuzhiyun  *      software is provided AS-IS with no warranties.
13*4882a593Smuzhiyun  *
14*4882a593Smuzhiyun  *	Date		Errata			Description
15*4882a593Smuzhiyun  *	20020525	N44, O17	12.5% or 25% DC causes lockup
16*4882a593Smuzhiyun  */
17*4882a593Smuzhiyun 
18*4882a593Smuzhiyun #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
19*4882a593Smuzhiyun 
20*4882a593Smuzhiyun #include <linux/kernel.h>
21*4882a593Smuzhiyun #include <linux/module.h>
22*4882a593Smuzhiyun #include <linux/init.h>
23*4882a593Smuzhiyun #include <linux/smp.h>
24*4882a593Smuzhiyun #include <linux/cpufreq.h>
25*4882a593Smuzhiyun #include <linux/cpumask.h>
26*4882a593Smuzhiyun #include <linux/timex.h>
27*4882a593Smuzhiyun 
28*4882a593Smuzhiyun #include <asm/processor.h>
29*4882a593Smuzhiyun #include <asm/msr.h>
30*4882a593Smuzhiyun #include <asm/timer.h>
31*4882a593Smuzhiyun #include <asm/cpu_device_id.h>
32*4882a593Smuzhiyun 
33*4882a593Smuzhiyun #include "speedstep-lib.h"
34*4882a593Smuzhiyun 
35*4882a593Smuzhiyun /*
36*4882a593Smuzhiyun  * Duty Cycle (3bits), note DC_DISABLE is not specified in
37*4882a593Smuzhiyun  * intel docs i just use it to mean disable
38*4882a593Smuzhiyun  */
39*4882a593Smuzhiyun enum {
40*4882a593Smuzhiyun 	DC_RESV, DC_DFLT, DC_25PT, DC_38PT, DC_50PT,
41*4882a593Smuzhiyun 	DC_64PT, DC_75PT, DC_88PT, DC_DISABLE
42*4882a593Smuzhiyun };
43*4882a593Smuzhiyun 
44*4882a593Smuzhiyun #define DC_ENTRIES	8
45*4882a593Smuzhiyun 
46*4882a593Smuzhiyun 
47*4882a593Smuzhiyun static int has_N44_O17_errata[NR_CPUS];
48*4882a593Smuzhiyun static unsigned int stock_freq;
49*4882a593Smuzhiyun static struct cpufreq_driver p4clockmod_driver;
50*4882a593Smuzhiyun static unsigned int cpufreq_p4_get(unsigned int cpu);
51*4882a593Smuzhiyun 
cpufreq_p4_setdc(unsigned int cpu,unsigned int newstate)52*4882a593Smuzhiyun static int cpufreq_p4_setdc(unsigned int cpu, unsigned int newstate)
53*4882a593Smuzhiyun {
54*4882a593Smuzhiyun 	u32 l, h;
55*4882a593Smuzhiyun 
56*4882a593Smuzhiyun 	if ((newstate > DC_DISABLE) || (newstate == DC_RESV))
57*4882a593Smuzhiyun 		return -EINVAL;
58*4882a593Smuzhiyun 
59*4882a593Smuzhiyun 	rdmsr_on_cpu(cpu, MSR_IA32_THERM_STATUS, &l, &h);
60*4882a593Smuzhiyun 
61*4882a593Smuzhiyun 	if (l & 0x01)
62*4882a593Smuzhiyun 		pr_debug("CPU#%d currently thermal throttled\n", cpu);
63*4882a593Smuzhiyun 
64*4882a593Smuzhiyun 	if (has_N44_O17_errata[cpu] &&
65*4882a593Smuzhiyun 	    (newstate == DC_25PT || newstate == DC_DFLT))
66*4882a593Smuzhiyun 		newstate = DC_38PT;
67*4882a593Smuzhiyun 
68*4882a593Smuzhiyun 	rdmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, &l, &h);
69*4882a593Smuzhiyun 	if (newstate == DC_DISABLE) {
70*4882a593Smuzhiyun 		pr_debug("CPU#%d disabling modulation\n", cpu);
71*4882a593Smuzhiyun 		wrmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, l & ~(1<<4), h);
72*4882a593Smuzhiyun 	} else {
73*4882a593Smuzhiyun 		pr_debug("CPU#%d setting duty cycle to %d%%\n",
74*4882a593Smuzhiyun 			cpu, ((125 * newstate) / 10));
75*4882a593Smuzhiyun 		/* bits 63 - 5	: reserved
76*4882a593Smuzhiyun 		 * bit  4	: enable/disable
77*4882a593Smuzhiyun 		 * bits 3-1	: duty cycle
78*4882a593Smuzhiyun 		 * bit  0	: reserved
79*4882a593Smuzhiyun 		 */
80*4882a593Smuzhiyun 		l = (l & ~14);
81*4882a593Smuzhiyun 		l = l | (1<<4) | ((newstate & 0x7)<<1);
82*4882a593Smuzhiyun 		wrmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, l, h);
83*4882a593Smuzhiyun 	}
84*4882a593Smuzhiyun 
85*4882a593Smuzhiyun 	return 0;
86*4882a593Smuzhiyun }
87*4882a593Smuzhiyun 
88*4882a593Smuzhiyun 
89*4882a593Smuzhiyun static struct cpufreq_frequency_table p4clockmod_table[] = {
90*4882a593Smuzhiyun 	{0, DC_RESV, CPUFREQ_ENTRY_INVALID},
91*4882a593Smuzhiyun 	{0, DC_DFLT, 0},
92*4882a593Smuzhiyun 	{0, DC_25PT, 0},
93*4882a593Smuzhiyun 	{0, DC_38PT, 0},
94*4882a593Smuzhiyun 	{0, DC_50PT, 0},
95*4882a593Smuzhiyun 	{0, DC_64PT, 0},
96*4882a593Smuzhiyun 	{0, DC_75PT, 0},
97*4882a593Smuzhiyun 	{0, DC_88PT, 0},
98*4882a593Smuzhiyun 	{0, DC_DISABLE, 0},
99*4882a593Smuzhiyun 	{0, DC_RESV, CPUFREQ_TABLE_END},
100*4882a593Smuzhiyun };
101*4882a593Smuzhiyun 
102*4882a593Smuzhiyun 
cpufreq_p4_target(struct cpufreq_policy * policy,unsigned int index)103*4882a593Smuzhiyun static int cpufreq_p4_target(struct cpufreq_policy *policy, unsigned int index)
104*4882a593Smuzhiyun {
105*4882a593Smuzhiyun 	int i;
106*4882a593Smuzhiyun 
107*4882a593Smuzhiyun 	/* run on each logical CPU,
108*4882a593Smuzhiyun 	 * see section 13.15.3 of IA32 Intel Architecture Software
109*4882a593Smuzhiyun 	 * Developer's Manual, Volume 3
110*4882a593Smuzhiyun 	 */
111*4882a593Smuzhiyun 	for_each_cpu(i, policy->cpus)
112*4882a593Smuzhiyun 		cpufreq_p4_setdc(i, p4clockmod_table[index].driver_data);
113*4882a593Smuzhiyun 
114*4882a593Smuzhiyun 	return 0;
115*4882a593Smuzhiyun }
116*4882a593Smuzhiyun 
117*4882a593Smuzhiyun 
cpufreq_p4_get_frequency(struct cpuinfo_x86 * c)118*4882a593Smuzhiyun static unsigned int cpufreq_p4_get_frequency(struct cpuinfo_x86 *c)
119*4882a593Smuzhiyun {
120*4882a593Smuzhiyun 	if (c->x86 == 0x06) {
121*4882a593Smuzhiyun 		if (cpu_has(c, X86_FEATURE_EST))
122*4882a593Smuzhiyun 			pr_warn_once("Warning: EST-capable CPU detected. The acpi-cpufreq module offers voltage scaling in addition to frequency scaling. You should use that instead of p4-clockmod, if possible.\n");
123*4882a593Smuzhiyun 		switch (c->x86_model) {
124*4882a593Smuzhiyun 		case 0x0E: /* Core */
125*4882a593Smuzhiyun 		case 0x0F: /* Core Duo */
126*4882a593Smuzhiyun 		case 0x16: /* Celeron Core */
127*4882a593Smuzhiyun 		case 0x1C: /* Atom */
128*4882a593Smuzhiyun 			p4clockmod_driver.flags |= CPUFREQ_CONST_LOOPS;
129*4882a593Smuzhiyun 			return speedstep_get_frequency(SPEEDSTEP_CPU_PCORE);
130*4882a593Smuzhiyun 		case 0x0D: /* Pentium M (Dothan) */
131*4882a593Smuzhiyun 			p4clockmod_driver.flags |= CPUFREQ_CONST_LOOPS;
132*4882a593Smuzhiyun 			fallthrough;
133*4882a593Smuzhiyun 		case 0x09: /* Pentium M (Banias) */
134*4882a593Smuzhiyun 			return speedstep_get_frequency(SPEEDSTEP_CPU_PM);
135*4882a593Smuzhiyun 		}
136*4882a593Smuzhiyun 	}
137*4882a593Smuzhiyun 
138*4882a593Smuzhiyun 	if (c->x86 != 0xF)
139*4882a593Smuzhiyun 		return 0;
140*4882a593Smuzhiyun 
141*4882a593Smuzhiyun 	/* on P-4s, the TSC runs with constant frequency independent whether
142*4882a593Smuzhiyun 	 * throttling is active or not. */
143*4882a593Smuzhiyun 	p4clockmod_driver.flags |= CPUFREQ_CONST_LOOPS;
144*4882a593Smuzhiyun 
145*4882a593Smuzhiyun 	if (speedstep_detect_processor() == SPEEDSTEP_CPU_P4M) {
146*4882a593Smuzhiyun 		pr_warn("Warning: Pentium 4-M detected. The speedstep-ich or acpi cpufreq modules offer voltage scaling in addition of frequency scaling. You should use either one instead of p4-clockmod, if possible.\n");
147*4882a593Smuzhiyun 		return speedstep_get_frequency(SPEEDSTEP_CPU_P4M);
148*4882a593Smuzhiyun 	}
149*4882a593Smuzhiyun 
150*4882a593Smuzhiyun 	return speedstep_get_frequency(SPEEDSTEP_CPU_P4D);
151*4882a593Smuzhiyun }
152*4882a593Smuzhiyun 
153*4882a593Smuzhiyun 
154*4882a593Smuzhiyun 
cpufreq_p4_cpu_init(struct cpufreq_policy * policy)155*4882a593Smuzhiyun static int cpufreq_p4_cpu_init(struct cpufreq_policy *policy)
156*4882a593Smuzhiyun {
157*4882a593Smuzhiyun 	struct cpuinfo_x86 *c = &cpu_data(policy->cpu);
158*4882a593Smuzhiyun 	int cpuid = 0;
159*4882a593Smuzhiyun 	unsigned int i;
160*4882a593Smuzhiyun 
161*4882a593Smuzhiyun #ifdef CONFIG_SMP
162*4882a593Smuzhiyun 	cpumask_copy(policy->cpus, topology_sibling_cpumask(policy->cpu));
163*4882a593Smuzhiyun #endif
164*4882a593Smuzhiyun 
165*4882a593Smuzhiyun 	/* Errata workaround */
166*4882a593Smuzhiyun 	cpuid = (c->x86 << 8) | (c->x86_model << 4) | c->x86_stepping;
167*4882a593Smuzhiyun 	switch (cpuid) {
168*4882a593Smuzhiyun 	case 0x0f07:
169*4882a593Smuzhiyun 	case 0x0f0a:
170*4882a593Smuzhiyun 	case 0x0f11:
171*4882a593Smuzhiyun 	case 0x0f12:
172*4882a593Smuzhiyun 		has_N44_O17_errata[policy->cpu] = 1;
173*4882a593Smuzhiyun 		pr_debug("has errata -- disabling low frequencies\n");
174*4882a593Smuzhiyun 	}
175*4882a593Smuzhiyun 
176*4882a593Smuzhiyun 	if (speedstep_detect_processor() == SPEEDSTEP_CPU_P4D &&
177*4882a593Smuzhiyun 	    c->x86_model < 2) {
178*4882a593Smuzhiyun 		/* switch to maximum frequency and measure result */
179*4882a593Smuzhiyun 		cpufreq_p4_setdc(policy->cpu, DC_DISABLE);
180*4882a593Smuzhiyun 		recalibrate_cpu_khz();
181*4882a593Smuzhiyun 	}
182*4882a593Smuzhiyun 	/* get max frequency */
183*4882a593Smuzhiyun 	stock_freq = cpufreq_p4_get_frequency(c);
184*4882a593Smuzhiyun 	if (!stock_freq)
185*4882a593Smuzhiyun 		return -EINVAL;
186*4882a593Smuzhiyun 
187*4882a593Smuzhiyun 	/* table init */
188*4882a593Smuzhiyun 	for (i = 1; (p4clockmod_table[i].frequency != CPUFREQ_TABLE_END); i++) {
189*4882a593Smuzhiyun 		if ((i < 2) && (has_N44_O17_errata[policy->cpu]))
190*4882a593Smuzhiyun 			p4clockmod_table[i].frequency = CPUFREQ_ENTRY_INVALID;
191*4882a593Smuzhiyun 		else
192*4882a593Smuzhiyun 			p4clockmod_table[i].frequency = (stock_freq * i)/8;
193*4882a593Smuzhiyun 	}
194*4882a593Smuzhiyun 
195*4882a593Smuzhiyun 	/* cpuinfo and default policy values */
196*4882a593Smuzhiyun 
197*4882a593Smuzhiyun 	/* the transition latency is set to be 1 higher than the maximum
198*4882a593Smuzhiyun 	 * transition latency of the ondemand governor */
199*4882a593Smuzhiyun 	policy->cpuinfo.transition_latency = 10000001;
200*4882a593Smuzhiyun 	policy->freq_table = &p4clockmod_table[0];
201*4882a593Smuzhiyun 
202*4882a593Smuzhiyun 	return 0;
203*4882a593Smuzhiyun }
204*4882a593Smuzhiyun 
205*4882a593Smuzhiyun 
cpufreq_p4_get(unsigned int cpu)206*4882a593Smuzhiyun static unsigned int cpufreq_p4_get(unsigned int cpu)
207*4882a593Smuzhiyun {
208*4882a593Smuzhiyun 	u32 l, h;
209*4882a593Smuzhiyun 
210*4882a593Smuzhiyun 	rdmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, &l, &h);
211*4882a593Smuzhiyun 
212*4882a593Smuzhiyun 	if (l & 0x10) {
213*4882a593Smuzhiyun 		l = l >> 1;
214*4882a593Smuzhiyun 		l &= 0x7;
215*4882a593Smuzhiyun 	} else
216*4882a593Smuzhiyun 		l = DC_DISABLE;
217*4882a593Smuzhiyun 
218*4882a593Smuzhiyun 	if (l != DC_DISABLE)
219*4882a593Smuzhiyun 		return stock_freq * l / 8;
220*4882a593Smuzhiyun 
221*4882a593Smuzhiyun 	return stock_freq;
222*4882a593Smuzhiyun }
223*4882a593Smuzhiyun 
224*4882a593Smuzhiyun static struct cpufreq_driver p4clockmod_driver = {
225*4882a593Smuzhiyun 	.verify		= cpufreq_generic_frequency_table_verify,
226*4882a593Smuzhiyun 	.target_index	= cpufreq_p4_target,
227*4882a593Smuzhiyun 	.init		= cpufreq_p4_cpu_init,
228*4882a593Smuzhiyun 	.get		= cpufreq_p4_get,
229*4882a593Smuzhiyun 	.name		= "p4-clockmod",
230*4882a593Smuzhiyun 	.attr		= cpufreq_generic_attr,
231*4882a593Smuzhiyun };
232*4882a593Smuzhiyun 
233*4882a593Smuzhiyun static const struct x86_cpu_id cpufreq_p4_id[] = {
234*4882a593Smuzhiyun 	X86_MATCH_VENDOR_FEATURE(INTEL, X86_FEATURE_ACC, NULL),
235*4882a593Smuzhiyun 	{}
236*4882a593Smuzhiyun };
237*4882a593Smuzhiyun 
238*4882a593Smuzhiyun /*
239*4882a593Smuzhiyun  * Intentionally no MODULE_DEVICE_TABLE here: this driver should not
240*4882a593Smuzhiyun  * be auto loaded.  Please don't add one.
241*4882a593Smuzhiyun  */
242*4882a593Smuzhiyun 
cpufreq_p4_init(void)243*4882a593Smuzhiyun static int __init cpufreq_p4_init(void)
244*4882a593Smuzhiyun {
245*4882a593Smuzhiyun 	int ret;
246*4882a593Smuzhiyun 
247*4882a593Smuzhiyun 	/*
248*4882a593Smuzhiyun 	 * THERM_CONTROL is architectural for IA32 now, so
249*4882a593Smuzhiyun 	 * we can rely on the capability checks
250*4882a593Smuzhiyun 	 */
251*4882a593Smuzhiyun 	if (!x86_match_cpu(cpufreq_p4_id) || !boot_cpu_has(X86_FEATURE_ACPI))
252*4882a593Smuzhiyun 		return -ENODEV;
253*4882a593Smuzhiyun 
254*4882a593Smuzhiyun 	ret = cpufreq_register_driver(&p4clockmod_driver);
255*4882a593Smuzhiyun 	if (!ret)
256*4882a593Smuzhiyun 		pr_info("P4/Xeon(TM) CPU On-Demand Clock Modulation available\n");
257*4882a593Smuzhiyun 
258*4882a593Smuzhiyun 	return ret;
259*4882a593Smuzhiyun }
260*4882a593Smuzhiyun 
261*4882a593Smuzhiyun 
cpufreq_p4_exit(void)262*4882a593Smuzhiyun static void __exit cpufreq_p4_exit(void)
263*4882a593Smuzhiyun {
264*4882a593Smuzhiyun 	cpufreq_unregister_driver(&p4clockmod_driver);
265*4882a593Smuzhiyun }
266*4882a593Smuzhiyun 
267*4882a593Smuzhiyun 
268*4882a593Smuzhiyun MODULE_AUTHOR("Zwane Mwaikambo <zwane@commfireservices.com>");
269*4882a593Smuzhiyun MODULE_DESCRIPTION("cpufreq driver for Pentium(TM) 4/Xeon(TM)");
270*4882a593Smuzhiyun MODULE_LICENSE("GPL");
271*4882a593Smuzhiyun 
272*4882a593Smuzhiyun late_initcall(cpufreq_p4_init);
273*4882a593Smuzhiyun module_exit(cpufreq_p4_exit);
274