xref: /OK3568_Linux_fs/kernel/include/linux/sched/numa_balancing.h (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /* SPDX-License-Identifier: GPL-2.0 */
2*4882a593Smuzhiyun #ifndef _LINUX_SCHED_NUMA_BALANCING_H
3*4882a593Smuzhiyun #define _LINUX_SCHED_NUMA_BALANCING_H
4*4882a593Smuzhiyun 
5*4882a593Smuzhiyun /*
6*4882a593Smuzhiyun  * This is the interface between the scheduler and the MM that
7*4882a593Smuzhiyun  * implements memory access pattern based NUMA-balancing:
8*4882a593Smuzhiyun  */
9*4882a593Smuzhiyun 
10*4882a593Smuzhiyun #include <linux/sched.h>
11*4882a593Smuzhiyun 
12*4882a593Smuzhiyun #define TNF_MIGRATED	0x01
13*4882a593Smuzhiyun #define TNF_NO_GROUP	0x02
14*4882a593Smuzhiyun #define TNF_SHARED	0x04
15*4882a593Smuzhiyun #define TNF_FAULT_LOCAL	0x08
16*4882a593Smuzhiyun #define TNF_MIGRATE_FAIL 0x10
17*4882a593Smuzhiyun 
18*4882a593Smuzhiyun #ifdef CONFIG_NUMA_BALANCING
19*4882a593Smuzhiyun extern void task_numa_fault(int last_node, int node, int pages, int flags);
20*4882a593Smuzhiyun extern pid_t task_numa_group_id(struct task_struct *p);
21*4882a593Smuzhiyun extern void set_numabalancing_state(bool enabled);
22*4882a593Smuzhiyun extern void task_numa_free(struct task_struct *p, bool final);
23*4882a593Smuzhiyun extern bool should_numa_migrate_memory(struct task_struct *p, struct page *page,
24*4882a593Smuzhiyun 					int src_nid, int dst_cpu);
25*4882a593Smuzhiyun #else
task_numa_fault(int last_node,int node,int pages,int flags)26*4882a593Smuzhiyun static inline void task_numa_fault(int last_node, int node, int pages,
27*4882a593Smuzhiyun 				   int flags)
28*4882a593Smuzhiyun {
29*4882a593Smuzhiyun }
task_numa_group_id(struct task_struct * p)30*4882a593Smuzhiyun static inline pid_t task_numa_group_id(struct task_struct *p)
31*4882a593Smuzhiyun {
32*4882a593Smuzhiyun 	return 0;
33*4882a593Smuzhiyun }
set_numabalancing_state(bool enabled)34*4882a593Smuzhiyun static inline void set_numabalancing_state(bool enabled)
35*4882a593Smuzhiyun {
36*4882a593Smuzhiyun }
task_numa_free(struct task_struct * p,bool final)37*4882a593Smuzhiyun static inline void task_numa_free(struct task_struct *p, bool final)
38*4882a593Smuzhiyun {
39*4882a593Smuzhiyun }
should_numa_migrate_memory(struct task_struct * p,struct page * page,int src_nid,int dst_cpu)40*4882a593Smuzhiyun static inline bool should_numa_migrate_memory(struct task_struct *p,
41*4882a593Smuzhiyun 				struct page *page, int src_nid, int dst_cpu)
42*4882a593Smuzhiyun {
43*4882a593Smuzhiyun 	return true;
44*4882a593Smuzhiyun }
45*4882a593Smuzhiyun #endif
46*4882a593Smuzhiyun 
47*4882a593Smuzhiyun #endif /* _LINUX_SCHED_NUMA_BALANCING_H */
48