xref: /OK3568_Linux_fs/kernel/arch/mips/include/asm/cpu-features.h (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun  * This file is subject to the terms and conditions of the GNU General Public
3*4882a593Smuzhiyun  * License.  See the file "COPYING" in the main directory of this archive
4*4882a593Smuzhiyun  * for more details.
5*4882a593Smuzhiyun  *
6*4882a593Smuzhiyun  * Copyright (C) 2003, 2004 Ralf Baechle
7*4882a593Smuzhiyun  * Copyright (C) 2004  Maciej W. Rozycki
8*4882a593Smuzhiyun  */
9*4882a593Smuzhiyun #ifndef __ASM_CPU_FEATURES_H
10*4882a593Smuzhiyun #define __ASM_CPU_FEATURES_H
11*4882a593Smuzhiyun 
12*4882a593Smuzhiyun #include <asm/cpu.h>
13*4882a593Smuzhiyun #include <asm/cpu-info.h>
14*4882a593Smuzhiyun #include <asm/isa-rev.h>
15*4882a593Smuzhiyun #include <cpu-feature-overrides.h>
16*4882a593Smuzhiyun 
17*4882a593Smuzhiyun #define __ase(ase)			(cpu_data[0].ases & (ase))
18*4882a593Smuzhiyun #define __isa(isa)			(cpu_data[0].isa_level & (isa))
19*4882a593Smuzhiyun #define __opt(opt)			(cpu_data[0].options & (opt))
20*4882a593Smuzhiyun 
21*4882a593Smuzhiyun /*
22*4882a593Smuzhiyun  * Check if MIPS_ISA_REV is >= isa *and* an option or ASE is detected during
23*4882a593Smuzhiyun  * boot (typically by cpu_probe()).
24*4882a593Smuzhiyun  *
25*4882a593Smuzhiyun  * Note that these should only be used in cases where a kernel built for an
26*4882a593Smuzhiyun  * older ISA *cannot* run on a CPU which supports the feature in question. For
27*4882a593Smuzhiyun  * example this may be used for features introduced with MIPSr6, since a kernel
28*4882a593Smuzhiyun  * built for an older ISA cannot run on a MIPSr6 CPU. This should not be used
29*4882a593Smuzhiyun  * for MIPSr2 features however, since a MIPSr1 or earlier kernel might run on a
30*4882a593Smuzhiyun  * MIPSr2 CPU.
31*4882a593Smuzhiyun  */
32*4882a593Smuzhiyun #define __isa_ge_and_ase(isa, ase)	((MIPS_ISA_REV >= (isa)) && __ase(ase))
33*4882a593Smuzhiyun #define __isa_ge_and_opt(isa, opt)	((MIPS_ISA_REV >= (isa)) && __opt(opt))
34*4882a593Smuzhiyun 
35*4882a593Smuzhiyun /*
36*4882a593Smuzhiyun  * Check if MIPS_ISA_REV is >= isa *or* an option or ASE is detected during
37*4882a593Smuzhiyun  * boot (typically by cpu_probe()).
38*4882a593Smuzhiyun  *
39*4882a593Smuzhiyun  * These are for use with features that are optional up until a particular ISA
40*4882a593Smuzhiyun  * revision & then become required.
41*4882a593Smuzhiyun  */
42*4882a593Smuzhiyun #define __isa_ge_or_ase(isa, ase)	((MIPS_ISA_REV >= (isa)) || __ase(ase))
43*4882a593Smuzhiyun #define __isa_ge_or_opt(isa, opt)	((MIPS_ISA_REV >= (isa)) || __opt(opt))
44*4882a593Smuzhiyun 
45*4882a593Smuzhiyun /*
46*4882a593Smuzhiyun  * Check if MIPS_ISA_REV is < isa *and* an option or ASE is detected during
47*4882a593Smuzhiyun  * boot (typically by cpu_probe()).
48*4882a593Smuzhiyun  *
49*4882a593Smuzhiyun  * These are for use with features that are optional up until a particular ISA
50*4882a593Smuzhiyun  * revision & are then removed - ie. no longer present in any CPU implementing
51*4882a593Smuzhiyun  * the given ISA revision.
52*4882a593Smuzhiyun  */
53*4882a593Smuzhiyun #define __isa_lt_and_ase(isa, ase)	((MIPS_ISA_REV < (isa)) && __ase(ase))
54*4882a593Smuzhiyun #define __isa_lt_and_opt(isa, opt)	((MIPS_ISA_REV < (isa)) && __opt(opt))
55*4882a593Smuzhiyun 
56*4882a593Smuzhiyun /*
57*4882a593Smuzhiyun  * Similarly allow for ISA level checks that take into account knowledge of the
58*4882a593Smuzhiyun  * ISA targeted by the kernel build, provided by MIPS_ISA_REV.
59*4882a593Smuzhiyun  */
60*4882a593Smuzhiyun #define __isa_ge_and_flag(isa, flag)	((MIPS_ISA_REV >= (isa)) && __isa(flag))
61*4882a593Smuzhiyun #define __isa_ge_or_flag(isa, flag)	((MIPS_ISA_REV >= (isa)) || __isa(flag))
62*4882a593Smuzhiyun #define __isa_lt_and_flag(isa, flag)	((MIPS_ISA_REV < (isa)) && __isa(flag))
63*4882a593Smuzhiyun #define __isa_range(ge, lt) \
64*4882a593Smuzhiyun 	((MIPS_ISA_REV >= (ge)) && (MIPS_ISA_REV < (lt)))
65*4882a593Smuzhiyun #define __isa_range_or_flag(ge, lt, flag) \
66*4882a593Smuzhiyun 	(__isa_range(ge, lt) || ((MIPS_ISA_REV < (lt)) && __isa(flag)))
67*4882a593Smuzhiyun #define __isa_range_and_ase(ge, lt, ase) \
68*4882a593Smuzhiyun 	(__isa_range(ge, lt) && __ase(ase))
69*4882a593Smuzhiyun 
70*4882a593Smuzhiyun /*
71*4882a593Smuzhiyun  * SMP assumption: Options of CPU 0 are a superset of all processors.
72*4882a593Smuzhiyun  * This is true for all known MIPS systems.
73*4882a593Smuzhiyun  */
74*4882a593Smuzhiyun #ifndef cpu_has_tlb
75*4882a593Smuzhiyun #define cpu_has_tlb		__opt(MIPS_CPU_TLB)
76*4882a593Smuzhiyun #endif
77*4882a593Smuzhiyun #ifndef cpu_has_ftlb
78*4882a593Smuzhiyun #define cpu_has_ftlb		__opt(MIPS_CPU_FTLB)
79*4882a593Smuzhiyun #endif
80*4882a593Smuzhiyun #ifndef cpu_has_tlbinv
81*4882a593Smuzhiyun #define cpu_has_tlbinv		__opt(MIPS_CPU_TLBINV)
82*4882a593Smuzhiyun #endif
83*4882a593Smuzhiyun #ifndef cpu_has_segments
84*4882a593Smuzhiyun #define cpu_has_segments	__opt(MIPS_CPU_SEGMENTS)
85*4882a593Smuzhiyun #endif
86*4882a593Smuzhiyun #ifndef cpu_has_eva
87*4882a593Smuzhiyun #define cpu_has_eva		__opt(MIPS_CPU_EVA)
88*4882a593Smuzhiyun #endif
89*4882a593Smuzhiyun #ifndef cpu_has_htw
90*4882a593Smuzhiyun #define cpu_has_htw		__opt(MIPS_CPU_HTW)
91*4882a593Smuzhiyun #endif
92*4882a593Smuzhiyun #ifndef cpu_has_ldpte
93*4882a593Smuzhiyun #define cpu_has_ldpte		__opt(MIPS_CPU_LDPTE)
94*4882a593Smuzhiyun #endif
95*4882a593Smuzhiyun #ifndef cpu_has_rixiex
96*4882a593Smuzhiyun #define cpu_has_rixiex		__isa_ge_or_opt(6, MIPS_CPU_RIXIEX)
97*4882a593Smuzhiyun #endif
98*4882a593Smuzhiyun #ifndef cpu_has_maar
99*4882a593Smuzhiyun #define cpu_has_maar		__opt(MIPS_CPU_MAAR)
100*4882a593Smuzhiyun #endif
101*4882a593Smuzhiyun #ifndef cpu_has_rw_llb
102*4882a593Smuzhiyun #define cpu_has_rw_llb		__isa_ge_or_opt(6, MIPS_CPU_RW_LLB)
103*4882a593Smuzhiyun #endif
104*4882a593Smuzhiyun 
105*4882a593Smuzhiyun /*
106*4882a593Smuzhiyun  * For the moment we don't consider R6000 and R8000 so we can assume that
107*4882a593Smuzhiyun  * anything that doesn't support R4000-style exceptions and interrupts is
108*4882a593Smuzhiyun  * R3000-like.  Users should still treat these two macro definitions as
109*4882a593Smuzhiyun  * opaque.
110*4882a593Smuzhiyun  */
111*4882a593Smuzhiyun #ifndef cpu_has_3kex
112*4882a593Smuzhiyun #define cpu_has_3kex		(!cpu_has_4kex)
113*4882a593Smuzhiyun #endif
114*4882a593Smuzhiyun #ifndef cpu_has_4kex
115*4882a593Smuzhiyun #define cpu_has_4kex		__isa_ge_or_opt(1, MIPS_CPU_4KEX)
116*4882a593Smuzhiyun #endif
117*4882a593Smuzhiyun #ifndef cpu_has_3k_cache
118*4882a593Smuzhiyun #define cpu_has_3k_cache	__isa_lt_and_opt(1, MIPS_CPU_3K_CACHE)
119*4882a593Smuzhiyun #endif
120*4882a593Smuzhiyun #define cpu_has_6k_cache	0
121*4882a593Smuzhiyun #define cpu_has_8k_cache	0
122*4882a593Smuzhiyun #ifndef cpu_has_4k_cache
123*4882a593Smuzhiyun #define cpu_has_4k_cache	__isa_ge_or_opt(1, MIPS_CPU_4K_CACHE)
124*4882a593Smuzhiyun #endif
125*4882a593Smuzhiyun #ifndef cpu_has_tx39_cache
126*4882a593Smuzhiyun #define cpu_has_tx39_cache	__opt(MIPS_CPU_TX39_CACHE)
127*4882a593Smuzhiyun #endif
128*4882a593Smuzhiyun #ifndef cpu_has_octeon_cache
129*4882a593Smuzhiyun #define cpu_has_octeon_cache	0
130*4882a593Smuzhiyun #endif
131*4882a593Smuzhiyun /* Don't override `cpu_has_fpu' to 1 or the "nofpu" option won't work.  */
132*4882a593Smuzhiyun #ifndef cpu_has_fpu
133*4882a593Smuzhiyun # ifdef CONFIG_MIPS_FP_SUPPORT
134*4882a593Smuzhiyun #  define cpu_has_fpu		(current_cpu_data.options & MIPS_CPU_FPU)
135*4882a593Smuzhiyun #  define raw_cpu_has_fpu	(raw_current_cpu_data.options & MIPS_CPU_FPU)
136*4882a593Smuzhiyun # else
137*4882a593Smuzhiyun #  define cpu_has_fpu		0
138*4882a593Smuzhiyun #  define raw_cpu_has_fpu	0
139*4882a593Smuzhiyun # endif
140*4882a593Smuzhiyun #else
141*4882a593Smuzhiyun # define raw_cpu_has_fpu	cpu_has_fpu
142*4882a593Smuzhiyun #endif
143*4882a593Smuzhiyun #ifndef cpu_has_32fpr
144*4882a593Smuzhiyun #define cpu_has_32fpr		__isa_ge_or_opt(1, MIPS_CPU_32FPR)
145*4882a593Smuzhiyun #endif
146*4882a593Smuzhiyun #ifndef cpu_has_counter
147*4882a593Smuzhiyun #define cpu_has_counter		__opt(MIPS_CPU_COUNTER)
148*4882a593Smuzhiyun #endif
149*4882a593Smuzhiyun #ifndef cpu_has_watch
150*4882a593Smuzhiyun #define cpu_has_watch		__opt(MIPS_CPU_WATCH)
151*4882a593Smuzhiyun #endif
152*4882a593Smuzhiyun #ifndef cpu_has_divec
153*4882a593Smuzhiyun #define cpu_has_divec		__isa_ge_or_opt(1, MIPS_CPU_DIVEC)
154*4882a593Smuzhiyun #endif
155*4882a593Smuzhiyun #ifndef cpu_has_vce
156*4882a593Smuzhiyun #define cpu_has_vce		__opt(MIPS_CPU_VCE)
157*4882a593Smuzhiyun #endif
158*4882a593Smuzhiyun #ifndef cpu_has_cache_cdex_p
159*4882a593Smuzhiyun #define cpu_has_cache_cdex_p	__opt(MIPS_CPU_CACHE_CDEX_P)
160*4882a593Smuzhiyun #endif
161*4882a593Smuzhiyun #ifndef cpu_has_cache_cdex_s
162*4882a593Smuzhiyun #define cpu_has_cache_cdex_s	__opt(MIPS_CPU_CACHE_CDEX_S)
163*4882a593Smuzhiyun #endif
164*4882a593Smuzhiyun #ifndef cpu_has_prefetch
165*4882a593Smuzhiyun #define cpu_has_prefetch	__isa_ge_or_opt(1, MIPS_CPU_PREFETCH)
166*4882a593Smuzhiyun #endif
167*4882a593Smuzhiyun #ifndef cpu_has_mcheck
168*4882a593Smuzhiyun #define cpu_has_mcheck		__isa_ge_or_opt(1, MIPS_CPU_MCHECK)
169*4882a593Smuzhiyun #endif
170*4882a593Smuzhiyun #ifndef cpu_has_ejtag
171*4882a593Smuzhiyun #define cpu_has_ejtag		__opt(MIPS_CPU_EJTAG)
172*4882a593Smuzhiyun #endif
173*4882a593Smuzhiyun #ifndef cpu_has_llsc
174*4882a593Smuzhiyun #define cpu_has_llsc		__isa_ge_or_opt(1, MIPS_CPU_LLSC)
175*4882a593Smuzhiyun #endif
176*4882a593Smuzhiyun #ifndef kernel_uses_llsc
177*4882a593Smuzhiyun #define kernel_uses_llsc	cpu_has_llsc
178*4882a593Smuzhiyun #endif
179*4882a593Smuzhiyun #ifndef cpu_has_guestctl0ext
180*4882a593Smuzhiyun #define cpu_has_guestctl0ext	__opt(MIPS_CPU_GUESTCTL0EXT)
181*4882a593Smuzhiyun #endif
182*4882a593Smuzhiyun #ifndef cpu_has_guestctl1
183*4882a593Smuzhiyun #define cpu_has_guestctl1	__opt(MIPS_CPU_GUESTCTL1)
184*4882a593Smuzhiyun #endif
185*4882a593Smuzhiyun #ifndef cpu_has_guestctl2
186*4882a593Smuzhiyun #define cpu_has_guestctl2	__opt(MIPS_CPU_GUESTCTL2)
187*4882a593Smuzhiyun #endif
188*4882a593Smuzhiyun #ifndef cpu_has_guestid
189*4882a593Smuzhiyun #define cpu_has_guestid		__opt(MIPS_CPU_GUESTID)
190*4882a593Smuzhiyun #endif
191*4882a593Smuzhiyun #ifndef cpu_has_drg
192*4882a593Smuzhiyun #define cpu_has_drg		__opt(MIPS_CPU_DRG)
193*4882a593Smuzhiyun #endif
194*4882a593Smuzhiyun #ifndef cpu_has_mips16
195*4882a593Smuzhiyun #define cpu_has_mips16		__isa_lt_and_ase(6, MIPS_ASE_MIPS16)
196*4882a593Smuzhiyun #endif
197*4882a593Smuzhiyun #ifndef cpu_has_mips16e2
198*4882a593Smuzhiyun #define cpu_has_mips16e2	__isa_lt_and_ase(6, MIPS_ASE_MIPS16E2)
199*4882a593Smuzhiyun #endif
200*4882a593Smuzhiyun #ifndef cpu_has_mdmx
201*4882a593Smuzhiyun #define cpu_has_mdmx		__isa_lt_and_ase(6, MIPS_ASE_MDMX)
202*4882a593Smuzhiyun #endif
203*4882a593Smuzhiyun #ifndef cpu_has_mips3d
204*4882a593Smuzhiyun #define cpu_has_mips3d		__isa_lt_and_ase(6, MIPS_ASE_MIPS3D)
205*4882a593Smuzhiyun #endif
206*4882a593Smuzhiyun #ifndef cpu_has_smartmips
207*4882a593Smuzhiyun #define cpu_has_smartmips	__isa_lt_and_ase(6, MIPS_ASE_SMARTMIPS)
208*4882a593Smuzhiyun #endif
209*4882a593Smuzhiyun 
210*4882a593Smuzhiyun #ifndef cpu_has_rixi
211*4882a593Smuzhiyun #define cpu_has_rixi		__isa_ge_or_opt(6, MIPS_CPU_RIXI)
212*4882a593Smuzhiyun #endif
213*4882a593Smuzhiyun 
214*4882a593Smuzhiyun #ifndef cpu_has_mmips
215*4882a593Smuzhiyun # if defined(__mips_micromips)
216*4882a593Smuzhiyun #  define cpu_has_mmips		1
217*4882a593Smuzhiyun # elif defined(CONFIG_SYS_SUPPORTS_MICROMIPS)
218*4882a593Smuzhiyun #  define cpu_has_mmips		__opt(MIPS_CPU_MICROMIPS)
219*4882a593Smuzhiyun # else
220*4882a593Smuzhiyun #  define cpu_has_mmips		0
221*4882a593Smuzhiyun # endif
222*4882a593Smuzhiyun #endif
223*4882a593Smuzhiyun 
224*4882a593Smuzhiyun #ifndef cpu_has_lpa
225*4882a593Smuzhiyun #define cpu_has_lpa		__opt(MIPS_CPU_LPA)
226*4882a593Smuzhiyun #endif
227*4882a593Smuzhiyun #ifndef cpu_has_mvh
228*4882a593Smuzhiyun #define cpu_has_mvh		__opt(MIPS_CPU_MVH)
229*4882a593Smuzhiyun #endif
230*4882a593Smuzhiyun #ifndef cpu_has_xpa
231*4882a593Smuzhiyun #define cpu_has_xpa		(cpu_has_lpa && cpu_has_mvh)
232*4882a593Smuzhiyun #endif
233*4882a593Smuzhiyun #ifndef cpu_has_vtag_icache
234*4882a593Smuzhiyun #define cpu_has_vtag_icache	(cpu_data[0].icache.flags & MIPS_CACHE_VTAG)
235*4882a593Smuzhiyun #endif
236*4882a593Smuzhiyun #ifndef cpu_has_dc_aliases
237*4882a593Smuzhiyun #define cpu_has_dc_aliases	(cpu_data[0].dcache.flags & MIPS_CACHE_ALIASES)
238*4882a593Smuzhiyun #endif
239*4882a593Smuzhiyun #ifndef cpu_has_ic_fills_f_dc
240*4882a593Smuzhiyun #define cpu_has_ic_fills_f_dc	(cpu_data[0].icache.flags & MIPS_CACHE_IC_F_DC)
241*4882a593Smuzhiyun #endif
242*4882a593Smuzhiyun #ifndef cpu_has_pindexed_dcache
243*4882a593Smuzhiyun #define cpu_has_pindexed_dcache	(cpu_data[0].dcache.flags & MIPS_CACHE_PINDEX)
244*4882a593Smuzhiyun #endif
245*4882a593Smuzhiyun 
246*4882a593Smuzhiyun /*
247*4882a593Smuzhiyun  * I-Cache snoops remote store.	 This only matters on SMP.  Some multiprocessors
248*4882a593Smuzhiyun  * such as the R10000 have I-Caches that snoop local stores; the embedded ones
249*4882a593Smuzhiyun  * don't.  For maintaining I-cache coherency this means we need to flush the
250*4882a593Smuzhiyun  * D-cache all the way back to whever the I-cache does refills from, so the
251*4882a593Smuzhiyun  * I-cache has a chance to see the new data at all.  Then we have to flush the
252*4882a593Smuzhiyun  * I-cache also.
253*4882a593Smuzhiyun  * Note we may have been rescheduled and may no longer be running on the CPU
254*4882a593Smuzhiyun  * that did the store so we can't optimize this into only doing the flush on
255*4882a593Smuzhiyun  * the local CPU.
256*4882a593Smuzhiyun  */
257*4882a593Smuzhiyun #ifndef cpu_icache_snoops_remote_store
258*4882a593Smuzhiyun #ifdef CONFIG_SMP
259*4882a593Smuzhiyun #define cpu_icache_snoops_remote_store	(cpu_data[0].icache.flags & MIPS_IC_SNOOPS_REMOTE)
260*4882a593Smuzhiyun #else
261*4882a593Smuzhiyun #define cpu_icache_snoops_remote_store	1
262*4882a593Smuzhiyun #endif
263*4882a593Smuzhiyun #endif
264*4882a593Smuzhiyun 
265*4882a593Smuzhiyun #ifndef cpu_has_mips_1
266*4882a593Smuzhiyun # define cpu_has_mips_1		(MIPS_ISA_REV < 6)
267*4882a593Smuzhiyun #endif
268*4882a593Smuzhiyun #ifndef cpu_has_mips_2
269*4882a593Smuzhiyun # define cpu_has_mips_2		__isa_lt_and_flag(6, MIPS_CPU_ISA_II)
270*4882a593Smuzhiyun #endif
271*4882a593Smuzhiyun #ifndef cpu_has_mips_3
272*4882a593Smuzhiyun # define cpu_has_mips_3		__isa_lt_and_flag(6, MIPS_CPU_ISA_III)
273*4882a593Smuzhiyun #endif
274*4882a593Smuzhiyun #ifndef cpu_has_mips_4
275*4882a593Smuzhiyun # define cpu_has_mips_4		__isa_lt_and_flag(6, MIPS_CPU_ISA_IV)
276*4882a593Smuzhiyun #endif
277*4882a593Smuzhiyun #ifndef cpu_has_mips_5
278*4882a593Smuzhiyun # define cpu_has_mips_5		__isa_lt_and_flag(6, MIPS_CPU_ISA_V)
279*4882a593Smuzhiyun #endif
280*4882a593Smuzhiyun #ifndef cpu_has_mips32r1
281*4882a593Smuzhiyun # define cpu_has_mips32r1	__isa_range_or_flag(1, 6, MIPS_CPU_ISA_M32R1)
282*4882a593Smuzhiyun #endif
283*4882a593Smuzhiyun #ifndef cpu_has_mips32r2
284*4882a593Smuzhiyun # define cpu_has_mips32r2	__isa_range_or_flag(2, 6, MIPS_CPU_ISA_M32R2)
285*4882a593Smuzhiyun #endif
286*4882a593Smuzhiyun #ifndef cpu_has_mips32r5
287*4882a593Smuzhiyun # define cpu_has_mips32r5	__isa_range_or_flag(5, 6, MIPS_CPU_ISA_M32R5)
288*4882a593Smuzhiyun #endif
289*4882a593Smuzhiyun #ifndef cpu_has_mips32r6
290*4882a593Smuzhiyun # define cpu_has_mips32r6	__isa_ge_or_flag(6, MIPS_CPU_ISA_M32R6)
291*4882a593Smuzhiyun #endif
292*4882a593Smuzhiyun #ifndef cpu_has_mips64r1
293*4882a593Smuzhiyun # define cpu_has_mips64r1	(cpu_has_64bits && \
294*4882a593Smuzhiyun 				 __isa_range_or_flag(1, 6, MIPS_CPU_ISA_M64R1))
295*4882a593Smuzhiyun #endif
296*4882a593Smuzhiyun #ifndef cpu_has_mips64r2
297*4882a593Smuzhiyun # define cpu_has_mips64r2	(cpu_has_64bits && \
298*4882a593Smuzhiyun 				 __isa_range_or_flag(2, 6, MIPS_CPU_ISA_M64R2))
299*4882a593Smuzhiyun #endif
300*4882a593Smuzhiyun #ifndef cpu_has_mips64r5
301*4882a593Smuzhiyun # define cpu_has_mips64r5	(cpu_has_64bits && \
302*4882a593Smuzhiyun 				 __isa_range_or_flag(5, 6, MIPS_CPU_ISA_M64R5))
303*4882a593Smuzhiyun #endif
304*4882a593Smuzhiyun #ifndef cpu_has_mips64r6
305*4882a593Smuzhiyun # define cpu_has_mips64r6	__isa_ge_and_flag(6, MIPS_CPU_ISA_M64R6)
306*4882a593Smuzhiyun #endif
307*4882a593Smuzhiyun 
308*4882a593Smuzhiyun /*
309*4882a593Smuzhiyun  * Shortcuts ...
310*4882a593Smuzhiyun  */
311*4882a593Smuzhiyun #define cpu_has_mips_2_3_4_5	(cpu_has_mips_2 | cpu_has_mips_3_4_5)
312*4882a593Smuzhiyun #define cpu_has_mips_3_4_5	(cpu_has_mips_3 | cpu_has_mips_4_5)
313*4882a593Smuzhiyun #define cpu_has_mips_4_5	(cpu_has_mips_4 | cpu_has_mips_5)
314*4882a593Smuzhiyun 
315*4882a593Smuzhiyun #define cpu_has_mips_2_3_4_5_r	(cpu_has_mips_2 | cpu_has_mips_3_4_5_r)
316*4882a593Smuzhiyun #define cpu_has_mips_3_4_5_r	(cpu_has_mips_3 | cpu_has_mips_4_5_r)
317*4882a593Smuzhiyun #define cpu_has_mips_4_5_r	(cpu_has_mips_4 | cpu_has_mips_5_r)
318*4882a593Smuzhiyun #define cpu_has_mips_5_r	(cpu_has_mips_5 | cpu_has_mips_r)
319*4882a593Smuzhiyun 
320*4882a593Smuzhiyun #define cpu_has_mips_3_4_5_64_r2_r6					\
321*4882a593Smuzhiyun 				(cpu_has_mips_3 | cpu_has_mips_4_5_64_r2_r6)
322*4882a593Smuzhiyun #define cpu_has_mips_4_5_64_r2_r6					\
323*4882a593Smuzhiyun 				(cpu_has_mips_4_5 | cpu_has_mips64r1 |	\
324*4882a593Smuzhiyun 				 cpu_has_mips_r2 | cpu_has_mips_r5 | \
325*4882a593Smuzhiyun 				 cpu_has_mips_r6)
326*4882a593Smuzhiyun 
327*4882a593Smuzhiyun #define cpu_has_mips32	(cpu_has_mips32r1 | cpu_has_mips32r2 | \
328*4882a593Smuzhiyun 			 cpu_has_mips32r5 | cpu_has_mips32r6)
329*4882a593Smuzhiyun #define cpu_has_mips64	(cpu_has_mips64r1 | cpu_has_mips64r2 | \
330*4882a593Smuzhiyun 			 cpu_has_mips64r5 | cpu_has_mips64r6)
331*4882a593Smuzhiyun #define cpu_has_mips_r1 (cpu_has_mips32r1 | cpu_has_mips64r1)
332*4882a593Smuzhiyun #define cpu_has_mips_r2 (cpu_has_mips32r2 | cpu_has_mips64r2)
333*4882a593Smuzhiyun #define cpu_has_mips_r5	(cpu_has_mips32r5 | cpu_has_mips64r5)
334*4882a593Smuzhiyun #define cpu_has_mips_r6	(cpu_has_mips32r6 | cpu_has_mips64r6)
335*4882a593Smuzhiyun #define cpu_has_mips_r	(cpu_has_mips32r1 | cpu_has_mips32r2 | \
336*4882a593Smuzhiyun 			 cpu_has_mips32r5 | cpu_has_mips32r6 | \
337*4882a593Smuzhiyun 			 cpu_has_mips64r1 | cpu_has_mips64r2 | \
338*4882a593Smuzhiyun 			 cpu_has_mips64r5 | cpu_has_mips64r6)
339*4882a593Smuzhiyun 
340*4882a593Smuzhiyun /* MIPSR2 - MIPSR6 have a lot of similarities */
341*4882a593Smuzhiyun #define cpu_has_mips_r2_r6	(cpu_has_mips_r2 | cpu_has_mips_r5 | \
342*4882a593Smuzhiyun 				 cpu_has_mips_r6)
343*4882a593Smuzhiyun 
344*4882a593Smuzhiyun /*
345*4882a593Smuzhiyun  * cpu_has_mips_r2_exec_hazard - return if IHB is required on current processor
346*4882a593Smuzhiyun  *
347*4882a593Smuzhiyun  * Returns non-zero value if the current processor implementation requires
348*4882a593Smuzhiyun  * an IHB instruction to deal with an instruction hazard as per MIPS R2
349*4882a593Smuzhiyun  * architecture specification, zero otherwise.
350*4882a593Smuzhiyun  */
351*4882a593Smuzhiyun #ifndef cpu_has_mips_r2_exec_hazard
352*4882a593Smuzhiyun #define cpu_has_mips_r2_exec_hazard					\
353*4882a593Smuzhiyun ({									\
354*4882a593Smuzhiyun 	int __res;							\
355*4882a593Smuzhiyun 									\
356*4882a593Smuzhiyun 	switch (current_cpu_type()) {					\
357*4882a593Smuzhiyun 	case CPU_M14KC:							\
358*4882a593Smuzhiyun 	case CPU_74K:							\
359*4882a593Smuzhiyun 	case CPU_1074K:							\
360*4882a593Smuzhiyun 	case CPU_PROAPTIV:						\
361*4882a593Smuzhiyun 	case CPU_P5600:							\
362*4882a593Smuzhiyun 	case CPU_M5150:							\
363*4882a593Smuzhiyun 	case CPU_QEMU_GENERIC:						\
364*4882a593Smuzhiyun 	case CPU_CAVIUM_OCTEON:						\
365*4882a593Smuzhiyun 	case CPU_CAVIUM_OCTEON_PLUS:					\
366*4882a593Smuzhiyun 	case CPU_CAVIUM_OCTEON2:					\
367*4882a593Smuzhiyun 	case CPU_CAVIUM_OCTEON3:					\
368*4882a593Smuzhiyun 		__res = 0;						\
369*4882a593Smuzhiyun 		break;							\
370*4882a593Smuzhiyun 									\
371*4882a593Smuzhiyun 	default:							\
372*4882a593Smuzhiyun 		__res = 1;						\
373*4882a593Smuzhiyun 	}								\
374*4882a593Smuzhiyun 									\
375*4882a593Smuzhiyun 	__res;								\
376*4882a593Smuzhiyun })
377*4882a593Smuzhiyun #endif
378*4882a593Smuzhiyun 
379*4882a593Smuzhiyun /*
380*4882a593Smuzhiyun  * MIPS32, MIPS64, VR5500, IDT32332, IDT32334 and maybe a few other
381*4882a593Smuzhiyun  * pre-MIPS32/MIPS64 processors have CLO, CLZ.	The IDT RC64574 is 64-bit and
382*4882a593Smuzhiyun  * has CLO and CLZ but not DCLO nor DCLZ.  For 64-bit kernels
383*4882a593Smuzhiyun  * cpu_has_clo_clz also indicates the availability of DCLO and DCLZ.
384*4882a593Smuzhiyun  */
385*4882a593Smuzhiyun #ifndef cpu_has_clo_clz
386*4882a593Smuzhiyun #define cpu_has_clo_clz	cpu_has_mips_r
387*4882a593Smuzhiyun #endif
388*4882a593Smuzhiyun 
389*4882a593Smuzhiyun /*
390*4882a593Smuzhiyun  * MIPS32 R2, MIPS64 R2, Loongson 3A and Octeon have WSBH.
391*4882a593Smuzhiyun  * MIPS64 R2, Loongson 3A and Octeon have WSBH, DSBH and DSHD.
392*4882a593Smuzhiyun  * This indicates the availability of WSBH and in case of 64 bit CPUs also
393*4882a593Smuzhiyun  * DSBH and DSHD.
394*4882a593Smuzhiyun  */
395*4882a593Smuzhiyun #ifndef cpu_has_wsbh
396*4882a593Smuzhiyun #define cpu_has_wsbh		cpu_has_mips_r2
397*4882a593Smuzhiyun #endif
398*4882a593Smuzhiyun 
399*4882a593Smuzhiyun #ifndef cpu_has_dsp
400*4882a593Smuzhiyun #define cpu_has_dsp		__ase(MIPS_ASE_DSP)
401*4882a593Smuzhiyun #endif
402*4882a593Smuzhiyun 
403*4882a593Smuzhiyun #ifndef cpu_has_dsp2
404*4882a593Smuzhiyun #define cpu_has_dsp2		__ase(MIPS_ASE_DSP2P)
405*4882a593Smuzhiyun #endif
406*4882a593Smuzhiyun 
407*4882a593Smuzhiyun #ifndef cpu_has_dsp3
408*4882a593Smuzhiyun #define cpu_has_dsp3		__ase(MIPS_ASE_DSP3)
409*4882a593Smuzhiyun #endif
410*4882a593Smuzhiyun 
411*4882a593Smuzhiyun #ifndef cpu_has_loongson_mmi
412*4882a593Smuzhiyun #define cpu_has_loongson_mmi		__ase(MIPS_ASE_LOONGSON_MMI)
413*4882a593Smuzhiyun #endif
414*4882a593Smuzhiyun 
415*4882a593Smuzhiyun #ifndef cpu_has_loongson_cam
416*4882a593Smuzhiyun #define cpu_has_loongson_cam		__ase(MIPS_ASE_LOONGSON_CAM)
417*4882a593Smuzhiyun #endif
418*4882a593Smuzhiyun 
419*4882a593Smuzhiyun #ifndef cpu_has_loongson_ext
420*4882a593Smuzhiyun #define cpu_has_loongson_ext		__ase(MIPS_ASE_LOONGSON_EXT)
421*4882a593Smuzhiyun #endif
422*4882a593Smuzhiyun 
423*4882a593Smuzhiyun #ifndef cpu_has_loongson_ext2
424*4882a593Smuzhiyun #define cpu_has_loongson_ext2		__ase(MIPS_ASE_LOONGSON_EXT2)
425*4882a593Smuzhiyun #endif
426*4882a593Smuzhiyun 
427*4882a593Smuzhiyun #ifndef cpu_has_mipsmt
428*4882a593Smuzhiyun #define cpu_has_mipsmt		__isa_range_and_ase(2, 6, MIPS_ASE_MIPSMT)
429*4882a593Smuzhiyun #endif
430*4882a593Smuzhiyun 
431*4882a593Smuzhiyun #ifndef cpu_has_vp
432*4882a593Smuzhiyun #define cpu_has_vp		__isa_ge_and_opt(6, MIPS_CPU_VP)
433*4882a593Smuzhiyun #endif
434*4882a593Smuzhiyun 
435*4882a593Smuzhiyun #ifndef cpu_has_userlocal
436*4882a593Smuzhiyun #define cpu_has_userlocal	__isa_ge_or_opt(6, MIPS_CPU_ULRI)
437*4882a593Smuzhiyun #endif
438*4882a593Smuzhiyun 
439*4882a593Smuzhiyun #ifdef CONFIG_32BIT
440*4882a593Smuzhiyun # ifndef cpu_has_nofpuex
441*4882a593Smuzhiyun # define cpu_has_nofpuex	__isa_lt_and_opt(1, MIPS_CPU_NOFPUEX)
442*4882a593Smuzhiyun # endif
443*4882a593Smuzhiyun # ifndef cpu_has_64bits
444*4882a593Smuzhiyun # define cpu_has_64bits		(cpu_data[0].isa_level & MIPS_CPU_ISA_64BIT)
445*4882a593Smuzhiyun # endif
446*4882a593Smuzhiyun # ifndef cpu_has_64bit_zero_reg
447*4882a593Smuzhiyun # define cpu_has_64bit_zero_reg	(cpu_data[0].isa_level & MIPS_CPU_ISA_64BIT)
448*4882a593Smuzhiyun # endif
449*4882a593Smuzhiyun # ifndef cpu_has_64bit_gp_regs
450*4882a593Smuzhiyun # define cpu_has_64bit_gp_regs		0
451*4882a593Smuzhiyun # endif
452*4882a593Smuzhiyun # ifndef cpu_vmbits
453*4882a593Smuzhiyun # define cpu_vmbits 31
454*4882a593Smuzhiyun # endif
455*4882a593Smuzhiyun #endif
456*4882a593Smuzhiyun 
457*4882a593Smuzhiyun #ifdef CONFIG_64BIT
458*4882a593Smuzhiyun # ifndef cpu_has_nofpuex
459*4882a593Smuzhiyun # define cpu_has_nofpuex		0
460*4882a593Smuzhiyun # endif
461*4882a593Smuzhiyun # ifndef cpu_has_64bits
462*4882a593Smuzhiyun # define cpu_has_64bits			1
463*4882a593Smuzhiyun # endif
464*4882a593Smuzhiyun # ifndef cpu_has_64bit_zero_reg
465*4882a593Smuzhiyun # define cpu_has_64bit_zero_reg		1
466*4882a593Smuzhiyun # endif
467*4882a593Smuzhiyun # ifndef cpu_has_64bit_gp_regs
468*4882a593Smuzhiyun # define cpu_has_64bit_gp_regs		1
469*4882a593Smuzhiyun # endif
470*4882a593Smuzhiyun # ifndef cpu_vmbits
471*4882a593Smuzhiyun # define cpu_vmbits cpu_data[0].vmbits
472*4882a593Smuzhiyun # define __NEED_VMBITS_PROBE
473*4882a593Smuzhiyun # endif
474*4882a593Smuzhiyun #endif
475*4882a593Smuzhiyun 
476*4882a593Smuzhiyun #if defined(CONFIG_CPU_MIPSR2_IRQ_VI) && !defined(cpu_has_vint)
477*4882a593Smuzhiyun # define cpu_has_vint		__opt(MIPS_CPU_VINT)
478*4882a593Smuzhiyun #elif !defined(cpu_has_vint)
479*4882a593Smuzhiyun # define cpu_has_vint			0
480*4882a593Smuzhiyun #endif
481*4882a593Smuzhiyun 
482*4882a593Smuzhiyun #if defined(CONFIG_CPU_MIPSR2_IRQ_EI) && !defined(cpu_has_veic)
483*4882a593Smuzhiyun # define cpu_has_veic		__opt(MIPS_CPU_VEIC)
484*4882a593Smuzhiyun #elif !defined(cpu_has_veic)
485*4882a593Smuzhiyun # define cpu_has_veic			0
486*4882a593Smuzhiyun #endif
487*4882a593Smuzhiyun 
488*4882a593Smuzhiyun #ifndef cpu_has_inclusive_pcaches
489*4882a593Smuzhiyun #define cpu_has_inclusive_pcaches	__opt(MIPS_CPU_INCLUSIVE_CACHES)
490*4882a593Smuzhiyun #endif
491*4882a593Smuzhiyun 
492*4882a593Smuzhiyun #ifndef cpu_dcache_line_size
493*4882a593Smuzhiyun #define cpu_dcache_line_size()	cpu_data[0].dcache.linesz
494*4882a593Smuzhiyun #endif
495*4882a593Smuzhiyun #ifndef cpu_icache_line_size
496*4882a593Smuzhiyun #define cpu_icache_line_size()	cpu_data[0].icache.linesz
497*4882a593Smuzhiyun #endif
498*4882a593Smuzhiyun #ifndef cpu_scache_line_size
499*4882a593Smuzhiyun #define cpu_scache_line_size()	cpu_data[0].scache.linesz
500*4882a593Smuzhiyun #endif
501*4882a593Smuzhiyun #ifndef cpu_tcache_line_size
502*4882a593Smuzhiyun #define cpu_tcache_line_size()	cpu_data[0].tcache.linesz
503*4882a593Smuzhiyun #endif
504*4882a593Smuzhiyun 
505*4882a593Smuzhiyun #ifndef cpu_hwrena_impl_bits
506*4882a593Smuzhiyun #define cpu_hwrena_impl_bits		0
507*4882a593Smuzhiyun #endif
508*4882a593Smuzhiyun 
509*4882a593Smuzhiyun #ifndef cpu_has_perf_cntr_intr_bit
510*4882a593Smuzhiyun #define cpu_has_perf_cntr_intr_bit	__opt(MIPS_CPU_PCI)
511*4882a593Smuzhiyun #endif
512*4882a593Smuzhiyun 
513*4882a593Smuzhiyun #ifndef cpu_has_vz
514*4882a593Smuzhiyun #define cpu_has_vz		__ase(MIPS_ASE_VZ)
515*4882a593Smuzhiyun #endif
516*4882a593Smuzhiyun 
517*4882a593Smuzhiyun #if defined(CONFIG_CPU_HAS_MSA) && !defined(cpu_has_msa)
518*4882a593Smuzhiyun # define cpu_has_msa		__ase(MIPS_ASE_MSA)
519*4882a593Smuzhiyun #elif !defined(cpu_has_msa)
520*4882a593Smuzhiyun # define cpu_has_msa		0
521*4882a593Smuzhiyun #endif
522*4882a593Smuzhiyun 
523*4882a593Smuzhiyun #ifndef cpu_has_ufr
524*4882a593Smuzhiyun # define cpu_has_ufr		__opt(MIPS_CPU_UFR)
525*4882a593Smuzhiyun #endif
526*4882a593Smuzhiyun 
527*4882a593Smuzhiyun #ifndef cpu_has_fre
528*4882a593Smuzhiyun # define cpu_has_fre		__opt(MIPS_CPU_FRE)
529*4882a593Smuzhiyun #endif
530*4882a593Smuzhiyun 
531*4882a593Smuzhiyun #ifndef cpu_has_cdmm
532*4882a593Smuzhiyun # define cpu_has_cdmm		__opt(MIPS_CPU_CDMM)
533*4882a593Smuzhiyun #endif
534*4882a593Smuzhiyun 
535*4882a593Smuzhiyun #ifndef cpu_has_small_pages
536*4882a593Smuzhiyun # define cpu_has_small_pages	__opt(MIPS_CPU_SP)
537*4882a593Smuzhiyun #endif
538*4882a593Smuzhiyun 
539*4882a593Smuzhiyun #ifndef cpu_has_nan_legacy
540*4882a593Smuzhiyun #define cpu_has_nan_legacy	__isa_lt_and_opt(6, MIPS_CPU_NAN_LEGACY)
541*4882a593Smuzhiyun #endif
542*4882a593Smuzhiyun #ifndef cpu_has_nan_2008
543*4882a593Smuzhiyun #define cpu_has_nan_2008	__isa_ge_or_opt(6, MIPS_CPU_NAN_2008)
544*4882a593Smuzhiyun #endif
545*4882a593Smuzhiyun 
546*4882a593Smuzhiyun #ifndef cpu_has_ebase_wg
547*4882a593Smuzhiyun # define cpu_has_ebase_wg	__opt(MIPS_CPU_EBASE_WG)
548*4882a593Smuzhiyun #endif
549*4882a593Smuzhiyun 
550*4882a593Smuzhiyun #ifndef cpu_has_badinstr
551*4882a593Smuzhiyun # define cpu_has_badinstr	__isa_ge_or_opt(6, MIPS_CPU_BADINSTR)
552*4882a593Smuzhiyun #endif
553*4882a593Smuzhiyun 
554*4882a593Smuzhiyun #ifndef cpu_has_badinstrp
555*4882a593Smuzhiyun # define cpu_has_badinstrp	__isa_ge_or_opt(6, MIPS_CPU_BADINSTRP)
556*4882a593Smuzhiyun #endif
557*4882a593Smuzhiyun 
558*4882a593Smuzhiyun #ifndef cpu_has_contextconfig
559*4882a593Smuzhiyun # define cpu_has_contextconfig	__opt(MIPS_CPU_CTXTC)
560*4882a593Smuzhiyun #endif
561*4882a593Smuzhiyun 
562*4882a593Smuzhiyun #ifndef cpu_has_perf
563*4882a593Smuzhiyun # define cpu_has_perf		__opt(MIPS_CPU_PERF)
564*4882a593Smuzhiyun #endif
565*4882a593Smuzhiyun 
566*4882a593Smuzhiyun #ifndef cpu_has_mac2008_only
567*4882a593Smuzhiyun # define cpu_has_mac2008_only	__opt(MIPS_CPU_MAC_2008_ONLY)
568*4882a593Smuzhiyun #endif
569*4882a593Smuzhiyun 
570*4882a593Smuzhiyun #ifndef cpu_has_ftlbparex
571*4882a593Smuzhiyun # define cpu_has_ftlbparex	__opt(MIPS_CPU_FTLBPAREX)
572*4882a593Smuzhiyun #endif
573*4882a593Smuzhiyun 
574*4882a593Smuzhiyun #ifndef cpu_has_gsexcex
575*4882a593Smuzhiyun # define cpu_has_gsexcex	__opt(MIPS_CPU_GSEXCEX)
576*4882a593Smuzhiyun #endif
577*4882a593Smuzhiyun 
578*4882a593Smuzhiyun #ifdef CONFIG_SMP
579*4882a593Smuzhiyun /*
580*4882a593Smuzhiyun  * Some systems share FTLB RAMs between threads within a core (siblings in
581*4882a593Smuzhiyun  * kernel parlance). This means that FTLB entries may become invalid at almost
582*4882a593Smuzhiyun  * any point when an entry is evicted due to a sibling thread writing an entry
583*4882a593Smuzhiyun  * to the shared FTLB RAM.
584*4882a593Smuzhiyun  *
585*4882a593Smuzhiyun  * This is only relevant to SMP systems, and the only systems that exhibit this
586*4882a593Smuzhiyun  * property implement MIPSr6 or higher so we constrain support for this to
587*4882a593Smuzhiyun  * kernels that will run on such systems.
588*4882a593Smuzhiyun  */
589*4882a593Smuzhiyun # ifndef cpu_has_shared_ftlb_ram
590*4882a593Smuzhiyun #  define cpu_has_shared_ftlb_ram \
591*4882a593Smuzhiyun 	__isa_ge_and_opt(6, MIPS_CPU_SHARED_FTLB_RAM)
592*4882a593Smuzhiyun # endif
593*4882a593Smuzhiyun 
594*4882a593Smuzhiyun /*
595*4882a593Smuzhiyun  * Some systems take this a step further & share FTLB entries between siblings.
596*4882a593Smuzhiyun  * This is implemented as TLB writes happening as usual, but if an entry
597*4882a593Smuzhiyun  * written by a sibling exists in the shared FTLB for a translation which would
598*4882a593Smuzhiyun  * otherwise cause a TLB refill exception then the CPU will use the entry
599*4882a593Smuzhiyun  * written by its sibling rather than triggering a refill & writing a matching
600*4882a593Smuzhiyun  * TLB entry for itself.
601*4882a593Smuzhiyun  *
602*4882a593Smuzhiyun  * This is naturally only valid if a TLB entry is known to be suitable for use
603*4882a593Smuzhiyun  * on all siblings in a CPU, and so it only takes effect when MMIDs are in use
604*4882a593Smuzhiyun  * rather than ASIDs or when a TLB entry is marked global.
605*4882a593Smuzhiyun  */
606*4882a593Smuzhiyun # ifndef cpu_has_shared_ftlb_entries
607*4882a593Smuzhiyun #  define cpu_has_shared_ftlb_entries \
608*4882a593Smuzhiyun 	__isa_ge_and_opt(6, MIPS_CPU_SHARED_FTLB_ENTRIES)
609*4882a593Smuzhiyun # endif
610*4882a593Smuzhiyun #endif /* SMP */
611*4882a593Smuzhiyun 
612*4882a593Smuzhiyun #ifndef cpu_has_shared_ftlb_ram
613*4882a593Smuzhiyun # define cpu_has_shared_ftlb_ram 0
614*4882a593Smuzhiyun #endif
615*4882a593Smuzhiyun #ifndef cpu_has_shared_ftlb_entries
616*4882a593Smuzhiyun # define cpu_has_shared_ftlb_entries 0
617*4882a593Smuzhiyun #endif
618*4882a593Smuzhiyun 
619*4882a593Smuzhiyun #ifdef CONFIG_MIPS_MT_SMP
620*4882a593Smuzhiyun # define cpu_has_mipsmt_pertccounters \
621*4882a593Smuzhiyun 	__isa_lt_and_opt(6, MIPS_CPU_MT_PER_TC_PERF_COUNTERS)
622*4882a593Smuzhiyun #else
623*4882a593Smuzhiyun # define cpu_has_mipsmt_pertccounters 0
624*4882a593Smuzhiyun #endif /* CONFIG_MIPS_MT_SMP */
625*4882a593Smuzhiyun 
626*4882a593Smuzhiyun /*
627*4882a593Smuzhiyun  * We only enable MMID support for configurations which natively support 64 bit
628*4882a593Smuzhiyun  * atomics because getting good performance from the allocator relies upon
629*4882a593Smuzhiyun  * efficient atomic64_*() functions.
630*4882a593Smuzhiyun  */
631*4882a593Smuzhiyun #ifndef cpu_has_mmid
632*4882a593Smuzhiyun # ifdef CONFIG_GENERIC_ATOMIC64
633*4882a593Smuzhiyun #  define cpu_has_mmid		0
634*4882a593Smuzhiyun # else
635*4882a593Smuzhiyun #  define cpu_has_mmid		__isa_ge_and_opt(6, MIPS_CPU_MMID)
636*4882a593Smuzhiyun # endif
637*4882a593Smuzhiyun #endif
638*4882a593Smuzhiyun 
639*4882a593Smuzhiyun #ifndef cpu_has_mm_sysad
640*4882a593Smuzhiyun # define cpu_has_mm_sysad	__opt(MIPS_CPU_MM_SYSAD)
641*4882a593Smuzhiyun #endif
642*4882a593Smuzhiyun 
643*4882a593Smuzhiyun #ifndef cpu_has_mm_full
644*4882a593Smuzhiyun # define cpu_has_mm_full	__opt(MIPS_CPU_MM_FULL)
645*4882a593Smuzhiyun #endif
646*4882a593Smuzhiyun 
647*4882a593Smuzhiyun /*
648*4882a593Smuzhiyun  * Guest capabilities
649*4882a593Smuzhiyun  */
650*4882a593Smuzhiyun #ifndef cpu_guest_has_conf1
651*4882a593Smuzhiyun #define cpu_guest_has_conf1	(cpu_data[0].guest.conf & (1 << 1))
652*4882a593Smuzhiyun #endif
653*4882a593Smuzhiyun #ifndef cpu_guest_has_conf2
654*4882a593Smuzhiyun #define cpu_guest_has_conf2	(cpu_data[0].guest.conf & (1 << 2))
655*4882a593Smuzhiyun #endif
656*4882a593Smuzhiyun #ifndef cpu_guest_has_conf3
657*4882a593Smuzhiyun #define cpu_guest_has_conf3	(cpu_data[0].guest.conf & (1 << 3))
658*4882a593Smuzhiyun #endif
659*4882a593Smuzhiyun #ifndef cpu_guest_has_conf4
660*4882a593Smuzhiyun #define cpu_guest_has_conf4	(cpu_data[0].guest.conf & (1 << 4))
661*4882a593Smuzhiyun #endif
662*4882a593Smuzhiyun #ifndef cpu_guest_has_conf5
663*4882a593Smuzhiyun #define cpu_guest_has_conf5	(cpu_data[0].guest.conf & (1 << 5))
664*4882a593Smuzhiyun #endif
665*4882a593Smuzhiyun #ifndef cpu_guest_has_conf6
666*4882a593Smuzhiyun #define cpu_guest_has_conf6	(cpu_data[0].guest.conf & (1 << 6))
667*4882a593Smuzhiyun #endif
668*4882a593Smuzhiyun #ifndef cpu_guest_has_conf7
669*4882a593Smuzhiyun #define cpu_guest_has_conf7	(cpu_data[0].guest.conf & (1 << 7))
670*4882a593Smuzhiyun #endif
671*4882a593Smuzhiyun #ifndef cpu_guest_has_fpu
672*4882a593Smuzhiyun #define cpu_guest_has_fpu	(cpu_data[0].guest.options & MIPS_CPU_FPU)
673*4882a593Smuzhiyun #endif
674*4882a593Smuzhiyun #ifndef cpu_guest_has_watch
675*4882a593Smuzhiyun #define cpu_guest_has_watch	(cpu_data[0].guest.options & MIPS_CPU_WATCH)
676*4882a593Smuzhiyun #endif
677*4882a593Smuzhiyun #ifndef cpu_guest_has_contextconfig
678*4882a593Smuzhiyun #define cpu_guest_has_contextconfig (cpu_data[0].guest.options & MIPS_CPU_CTXTC)
679*4882a593Smuzhiyun #endif
680*4882a593Smuzhiyun #ifndef cpu_guest_has_segments
681*4882a593Smuzhiyun #define cpu_guest_has_segments	(cpu_data[0].guest.options & MIPS_CPU_SEGMENTS)
682*4882a593Smuzhiyun #endif
683*4882a593Smuzhiyun #ifndef cpu_guest_has_badinstr
684*4882a593Smuzhiyun #define cpu_guest_has_badinstr	(cpu_data[0].guest.options & MIPS_CPU_BADINSTR)
685*4882a593Smuzhiyun #endif
686*4882a593Smuzhiyun #ifndef cpu_guest_has_badinstrp
687*4882a593Smuzhiyun #define cpu_guest_has_badinstrp	(cpu_data[0].guest.options & MIPS_CPU_BADINSTRP)
688*4882a593Smuzhiyun #endif
689*4882a593Smuzhiyun #ifndef cpu_guest_has_htw
690*4882a593Smuzhiyun #define cpu_guest_has_htw	(cpu_data[0].guest.options & MIPS_CPU_HTW)
691*4882a593Smuzhiyun #endif
692*4882a593Smuzhiyun #ifndef cpu_guest_has_ldpte
693*4882a593Smuzhiyun #define cpu_guest_has_ldpte	(cpu_data[0].guest.options & MIPS_CPU_LDPTE)
694*4882a593Smuzhiyun #endif
695*4882a593Smuzhiyun #ifndef cpu_guest_has_mvh
696*4882a593Smuzhiyun #define cpu_guest_has_mvh	(cpu_data[0].guest.options & MIPS_CPU_MVH)
697*4882a593Smuzhiyun #endif
698*4882a593Smuzhiyun #ifndef cpu_guest_has_msa
699*4882a593Smuzhiyun #define cpu_guest_has_msa	(cpu_data[0].guest.ases & MIPS_ASE_MSA)
700*4882a593Smuzhiyun #endif
701*4882a593Smuzhiyun #ifndef cpu_guest_has_kscr
702*4882a593Smuzhiyun #define cpu_guest_has_kscr(n)	(cpu_data[0].guest.kscratch_mask & (1u << (n)))
703*4882a593Smuzhiyun #endif
704*4882a593Smuzhiyun #ifndef cpu_guest_has_rw_llb
705*4882a593Smuzhiyun #define cpu_guest_has_rw_llb	(cpu_has_mips_r6 || (cpu_data[0].guest.options & MIPS_CPU_RW_LLB))
706*4882a593Smuzhiyun #endif
707*4882a593Smuzhiyun #ifndef cpu_guest_has_perf
708*4882a593Smuzhiyun #define cpu_guest_has_perf	(cpu_data[0].guest.options & MIPS_CPU_PERF)
709*4882a593Smuzhiyun #endif
710*4882a593Smuzhiyun #ifndef cpu_guest_has_maar
711*4882a593Smuzhiyun #define cpu_guest_has_maar	(cpu_data[0].guest.options & MIPS_CPU_MAAR)
712*4882a593Smuzhiyun #endif
713*4882a593Smuzhiyun #ifndef cpu_guest_has_userlocal
714*4882a593Smuzhiyun #define cpu_guest_has_userlocal	(cpu_data[0].guest.options & MIPS_CPU_ULRI)
715*4882a593Smuzhiyun #endif
716*4882a593Smuzhiyun 
717*4882a593Smuzhiyun /*
718*4882a593Smuzhiyun  * Guest dynamic capabilities
719*4882a593Smuzhiyun  */
720*4882a593Smuzhiyun #ifndef cpu_guest_has_dyn_fpu
721*4882a593Smuzhiyun #define cpu_guest_has_dyn_fpu	(cpu_data[0].guest.options_dyn & MIPS_CPU_FPU)
722*4882a593Smuzhiyun #endif
723*4882a593Smuzhiyun #ifndef cpu_guest_has_dyn_watch
724*4882a593Smuzhiyun #define cpu_guest_has_dyn_watch	(cpu_data[0].guest.options_dyn & MIPS_CPU_WATCH)
725*4882a593Smuzhiyun #endif
726*4882a593Smuzhiyun #ifndef cpu_guest_has_dyn_contextconfig
727*4882a593Smuzhiyun #define cpu_guest_has_dyn_contextconfig (cpu_data[0].guest.options_dyn & MIPS_CPU_CTXTC)
728*4882a593Smuzhiyun #endif
729*4882a593Smuzhiyun #ifndef cpu_guest_has_dyn_perf
730*4882a593Smuzhiyun #define cpu_guest_has_dyn_perf	(cpu_data[0].guest.options_dyn & MIPS_CPU_PERF)
731*4882a593Smuzhiyun #endif
732*4882a593Smuzhiyun #ifndef cpu_guest_has_dyn_msa
733*4882a593Smuzhiyun #define cpu_guest_has_dyn_msa	(cpu_data[0].guest.ases_dyn & MIPS_ASE_MSA)
734*4882a593Smuzhiyun #endif
735*4882a593Smuzhiyun #ifndef cpu_guest_has_dyn_maar
736*4882a593Smuzhiyun #define cpu_guest_has_dyn_maar	(cpu_data[0].guest.options_dyn & MIPS_CPU_MAAR)
737*4882a593Smuzhiyun #endif
738*4882a593Smuzhiyun 
739*4882a593Smuzhiyun #endif /* __ASM_CPU_FEATURES_H */
740