1*4882a593Smuzhiyun /* SPDX-License-Identifier: GPL-2.0 */
2*4882a593Smuzhiyun #ifndef _ASM_X86_CPUFEATURE_H
3*4882a593Smuzhiyun #define _ASM_X86_CPUFEATURE_H
4*4882a593Smuzhiyun
5*4882a593Smuzhiyun #include <asm/processor.h>
6*4882a593Smuzhiyun
7*4882a593Smuzhiyun #if defined(__KERNEL__) && !defined(__ASSEMBLY__)
8*4882a593Smuzhiyun
9*4882a593Smuzhiyun #include <asm/asm.h>
10*4882a593Smuzhiyun #include <linux/bitops.h>
11*4882a593Smuzhiyun #include <asm/alternative.h>
12*4882a593Smuzhiyun
13*4882a593Smuzhiyun enum cpuid_leafs
14*4882a593Smuzhiyun {
15*4882a593Smuzhiyun CPUID_1_EDX = 0,
16*4882a593Smuzhiyun CPUID_8000_0001_EDX,
17*4882a593Smuzhiyun CPUID_8086_0001_EDX,
18*4882a593Smuzhiyun CPUID_LNX_1,
19*4882a593Smuzhiyun CPUID_1_ECX,
20*4882a593Smuzhiyun CPUID_C000_0001_EDX,
21*4882a593Smuzhiyun CPUID_8000_0001_ECX,
22*4882a593Smuzhiyun CPUID_LNX_2,
23*4882a593Smuzhiyun CPUID_LNX_3,
24*4882a593Smuzhiyun CPUID_7_0_EBX,
25*4882a593Smuzhiyun CPUID_D_1_EAX,
26*4882a593Smuzhiyun CPUID_LNX_4,
27*4882a593Smuzhiyun CPUID_7_1_EAX,
28*4882a593Smuzhiyun CPUID_8000_0008_EBX,
29*4882a593Smuzhiyun CPUID_6_EAX,
30*4882a593Smuzhiyun CPUID_8000_000A_EDX,
31*4882a593Smuzhiyun CPUID_7_ECX,
32*4882a593Smuzhiyun CPUID_8000_0007_EBX,
33*4882a593Smuzhiyun CPUID_7_EDX,
34*4882a593Smuzhiyun };
35*4882a593Smuzhiyun
36*4882a593Smuzhiyun #ifdef CONFIG_X86_FEATURE_NAMES
37*4882a593Smuzhiyun extern const char * const x86_cap_flags[NCAPINTS*32];
38*4882a593Smuzhiyun extern const char * const x86_power_flags[32];
39*4882a593Smuzhiyun #define X86_CAP_FMT "%s"
40*4882a593Smuzhiyun #define x86_cap_flag(flag) x86_cap_flags[flag]
41*4882a593Smuzhiyun #else
42*4882a593Smuzhiyun #define X86_CAP_FMT "%d:%d"
43*4882a593Smuzhiyun #define x86_cap_flag(flag) ((flag) >> 5), ((flag) & 31)
44*4882a593Smuzhiyun #endif
45*4882a593Smuzhiyun
46*4882a593Smuzhiyun /*
47*4882a593Smuzhiyun * In order to save room, we index into this array by doing
48*4882a593Smuzhiyun * X86_BUG_<name> - NCAPINTS*32.
49*4882a593Smuzhiyun */
50*4882a593Smuzhiyun extern const char * const x86_bug_flags[NBUGINTS*32];
51*4882a593Smuzhiyun
52*4882a593Smuzhiyun #define test_cpu_cap(c, bit) \
53*4882a593Smuzhiyun arch_test_bit(bit, (unsigned long *)((c)->x86_capability))
54*4882a593Smuzhiyun
55*4882a593Smuzhiyun /*
56*4882a593Smuzhiyun * There are 32 bits/features in each mask word. The high bits
57*4882a593Smuzhiyun * (selected with (bit>>5) give us the word number and the low 5
58*4882a593Smuzhiyun * bits give us the bit/feature number inside the word.
59*4882a593Smuzhiyun * (1UL<<((bit)&31) gives us a mask for the feature_bit so we can
60*4882a593Smuzhiyun * see if it is set in the mask word.
61*4882a593Smuzhiyun */
62*4882a593Smuzhiyun #define CHECK_BIT_IN_MASK_WORD(maskname, word, bit) \
63*4882a593Smuzhiyun (((bit)>>5)==(word) && (1UL<<((bit)&31) & maskname##word ))
64*4882a593Smuzhiyun
65*4882a593Smuzhiyun /*
66*4882a593Smuzhiyun * {REQUIRED,DISABLED}_MASK_CHECK below may seem duplicated with the
67*4882a593Smuzhiyun * following BUILD_BUG_ON_ZERO() check but when NCAPINTS gets changed, all
68*4882a593Smuzhiyun * header macros which use NCAPINTS need to be changed. The duplicated macro
69*4882a593Smuzhiyun * use causes the compiler to issue errors for all headers so that all usage
70*4882a593Smuzhiyun * sites can be corrected.
71*4882a593Smuzhiyun */
72*4882a593Smuzhiyun #define REQUIRED_MASK_BIT_SET(feature_bit) \
73*4882a593Smuzhiyun ( CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 0, feature_bit) || \
74*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 1, feature_bit) || \
75*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 2, feature_bit) || \
76*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 3, feature_bit) || \
77*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 4, feature_bit) || \
78*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 5, feature_bit) || \
79*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 6, feature_bit) || \
80*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 7, feature_bit) || \
81*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 8, feature_bit) || \
82*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 9, feature_bit) || \
83*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 10, feature_bit) || \
84*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 11, feature_bit) || \
85*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 12, feature_bit) || \
86*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 13, feature_bit) || \
87*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 14, feature_bit) || \
88*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 15, feature_bit) || \
89*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 16, feature_bit) || \
90*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 17, feature_bit) || \
91*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(REQUIRED_MASK, 18, feature_bit) || \
92*4882a593Smuzhiyun REQUIRED_MASK_CHECK || \
93*4882a593Smuzhiyun BUILD_BUG_ON_ZERO(NCAPINTS != 19))
94*4882a593Smuzhiyun
95*4882a593Smuzhiyun #define DISABLED_MASK_BIT_SET(feature_bit) \
96*4882a593Smuzhiyun ( CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 0, feature_bit) || \
97*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 1, feature_bit) || \
98*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 2, feature_bit) || \
99*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 3, feature_bit) || \
100*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 4, feature_bit) || \
101*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 5, feature_bit) || \
102*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 6, feature_bit) || \
103*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 7, feature_bit) || \
104*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 8, feature_bit) || \
105*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 9, feature_bit) || \
106*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 10, feature_bit) || \
107*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 11, feature_bit) || \
108*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 12, feature_bit) || \
109*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 13, feature_bit) || \
110*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 14, feature_bit) || \
111*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 15, feature_bit) || \
112*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 16, feature_bit) || \
113*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 17, feature_bit) || \
114*4882a593Smuzhiyun CHECK_BIT_IN_MASK_WORD(DISABLED_MASK, 18, feature_bit) || \
115*4882a593Smuzhiyun DISABLED_MASK_CHECK || \
116*4882a593Smuzhiyun BUILD_BUG_ON_ZERO(NCAPINTS != 19))
117*4882a593Smuzhiyun
118*4882a593Smuzhiyun #define cpu_has(c, bit) \
119*4882a593Smuzhiyun (__builtin_constant_p(bit) && REQUIRED_MASK_BIT_SET(bit) ? 1 : \
120*4882a593Smuzhiyun test_cpu_cap(c, bit))
121*4882a593Smuzhiyun
122*4882a593Smuzhiyun #define this_cpu_has(bit) \
123*4882a593Smuzhiyun (__builtin_constant_p(bit) && REQUIRED_MASK_BIT_SET(bit) ? 1 : \
124*4882a593Smuzhiyun x86_this_cpu_test_bit(bit, \
125*4882a593Smuzhiyun (unsigned long __percpu *)&cpu_info.x86_capability))
126*4882a593Smuzhiyun
127*4882a593Smuzhiyun /*
128*4882a593Smuzhiyun * This macro is for detection of features which need kernel
129*4882a593Smuzhiyun * infrastructure to be used. It may *not* directly test the CPU
130*4882a593Smuzhiyun * itself. Use the cpu_has() family if you want true runtime
131*4882a593Smuzhiyun * testing of CPU features, like in hypervisor code where you are
132*4882a593Smuzhiyun * supporting a possible guest feature where host support for it
133*4882a593Smuzhiyun * is not relevant.
134*4882a593Smuzhiyun */
135*4882a593Smuzhiyun #define cpu_feature_enabled(bit) \
136*4882a593Smuzhiyun (__builtin_constant_p(bit) && DISABLED_MASK_BIT_SET(bit) ? 0 : static_cpu_has(bit))
137*4882a593Smuzhiyun
138*4882a593Smuzhiyun #define boot_cpu_has(bit) cpu_has(&boot_cpu_data, bit)
139*4882a593Smuzhiyun
140*4882a593Smuzhiyun #define set_cpu_cap(c, bit) set_bit(bit, (unsigned long *)((c)->x86_capability))
141*4882a593Smuzhiyun
142*4882a593Smuzhiyun extern void setup_clear_cpu_cap(unsigned int bit);
143*4882a593Smuzhiyun extern void clear_cpu_cap(struct cpuinfo_x86 *c, unsigned int bit);
144*4882a593Smuzhiyun
145*4882a593Smuzhiyun #define setup_force_cpu_cap(bit) do { \
146*4882a593Smuzhiyun set_cpu_cap(&boot_cpu_data, bit); \
147*4882a593Smuzhiyun set_bit(bit, (unsigned long *)cpu_caps_set); \
148*4882a593Smuzhiyun } while (0)
149*4882a593Smuzhiyun
150*4882a593Smuzhiyun #define setup_force_cpu_bug(bit) setup_force_cpu_cap(bit)
151*4882a593Smuzhiyun
152*4882a593Smuzhiyun #if defined(__clang__) && !defined(CONFIG_CC_HAS_ASM_GOTO)
153*4882a593Smuzhiyun
154*4882a593Smuzhiyun /*
155*4882a593Smuzhiyun * Workaround for the sake of BPF compilation which utilizes kernel
156*4882a593Smuzhiyun * headers, but clang does not support ASM GOTO and fails the build.
157*4882a593Smuzhiyun */
158*4882a593Smuzhiyun #ifndef __BPF_TRACING__
159*4882a593Smuzhiyun #warning "Compiler lacks ASM_GOTO support. Add -D __BPF_TRACING__ to your compiler arguments"
160*4882a593Smuzhiyun #endif
161*4882a593Smuzhiyun
162*4882a593Smuzhiyun #define static_cpu_has(bit) boot_cpu_has(bit)
163*4882a593Smuzhiyun
164*4882a593Smuzhiyun #else
165*4882a593Smuzhiyun
166*4882a593Smuzhiyun /*
167*4882a593Smuzhiyun * Static testing of CPU features. Used the same as boot_cpu_has(). It
168*4882a593Smuzhiyun * statically patches the target code for additional performance. Use
169*4882a593Smuzhiyun * static_cpu_has() only in fast paths, where every cycle counts. Which
170*4882a593Smuzhiyun * means that the boot_cpu_has() variant is already fast enough for the
171*4882a593Smuzhiyun * majority of cases and you should stick to using it as it is generally
172*4882a593Smuzhiyun * only two instructions: a RIP-relative MOV and a TEST.
173*4882a593Smuzhiyun */
_static_cpu_has(u16 bit)174*4882a593Smuzhiyun static __always_inline bool _static_cpu_has(u16 bit)
175*4882a593Smuzhiyun {
176*4882a593Smuzhiyun asm_volatile_goto(
177*4882a593Smuzhiyun ALTERNATIVE_TERNARY("jmp 6f", %P[feature], "", "jmp %l[t_no]")
178*4882a593Smuzhiyun ".section .altinstr_aux,\"ax\"\n"
179*4882a593Smuzhiyun "6:\n"
180*4882a593Smuzhiyun " testb %[bitnum],%[cap_byte]\n"
181*4882a593Smuzhiyun " jnz %l[t_yes]\n"
182*4882a593Smuzhiyun " jmp %l[t_no]\n"
183*4882a593Smuzhiyun ".previous\n"
184*4882a593Smuzhiyun : : [feature] "i" (bit),
185*4882a593Smuzhiyun [bitnum] "i" (1 << (bit & 7)),
186*4882a593Smuzhiyun [cap_byte] "m" (((const char *)boot_cpu_data.x86_capability)[bit >> 3])
187*4882a593Smuzhiyun : : t_yes, t_no);
188*4882a593Smuzhiyun t_yes:
189*4882a593Smuzhiyun return true;
190*4882a593Smuzhiyun t_no:
191*4882a593Smuzhiyun return false;
192*4882a593Smuzhiyun }
193*4882a593Smuzhiyun
194*4882a593Smuzhiyun #define static_cpu_has(bit) \
195*4882a593Smuzhiyun ( \
196*4882a593Smuzhiyun __builtin_constant_p(boot_cpu_has(bit)) ? \
197*4882a593Smuzhiyun boot_cpu_has(bit) : \
198*4882a593Smuzhiyun _static_cpu_has(bit) \
199*4882a593Smuzhiyun )
200*4882a593Smuzhiyun #endif
201*4882a593Smuzhiyun
202*4882a593Smuzhiyun #define cpu_has_bug(c, bit) cpu_has(c, (bit))
203*4882a593Smuzhiyun #define set_cpu_bug(c, bit) set_cpu_cap(c, (bit))
204*4882a593Smuzhiyun #define clear_cpu_bug(c, bit) clear_cpu_cap(c, (bit))
205*4882a593Smuzhiyun
206*4882a593Smuzhiyun #define static_cpu_has_bug(bit) static_cpu_has((bit))
207*4882a593Smuzhiyun #define boot_cpu_has_bug(bit) cpu_has_bug(&boot_cpu_data, (bit))
208*4882a593Smuzhiyun #define boot_cpu_set_bug(bit) set_cpu_cap(&boot_cpu_data, (bit))
209*4882a593Smuzhiyun
210*4882a593Smuzhiyun #define MAX_CPU_FEATURES (NCAPINTS * 32)
211*4882a593Smuzhiyun #define cpu_have_feature boot_cpu_has
212*4882a593Smuzhiyun
213*4882a593Smuzhiyun #define CPU_FEATURE_TYPEFMT "x86,ven%04Xfam%04Xmod%04X"
214*4882a593Smuzhiyun #define CPU_FEATURE_TYPEVAL boot_cpu_data.x86_vendor, boot_cpu_data.x86, \
215*4882a593Smuzhiyun boot_cpu_data.x86_model
216*4882a593Smuzhiyun
217*4882a593Smuzhiyun #endif /* defined(__KERNEL__) && !defined(__ASSEMBLY__) */
218*4882a593Smuzhiyun #endif /* _ASM_X86_CPUFEATURE_H */
219