| /OK3568_Linux_fs/kernel/block/ |
| H A D | kyber-iosched.c | 211 unsigned int sched_domain, unsigned int type) in flush_latency_buckets() argument 213 unsigned int *buckets = kqd->latency_buckets[sched_domain][type]; in flush_latency_buckets() 214 atomic_t *cpu_buckets = cpu_latency->buckets[sched_domain][type]; in flush_latency_buckets() 226 unsigned int sched_domain, unsigned int type, in calculate_percentile() argument 229 unsigned int *buckets = kqd->latency_buckets[sched_domain][type]; in calculate_percentile() 242 if (!kqd->latency_timeout[sched_domain]) in calculate_percentile() 243 kqd->latency_timeout[sched_domain] = max(jiffies + HZ, 1UL); in calculate_percentile() 245 time_is_after_jiffies(kqd->latency_timeout[sched_domain])) { in calculate_percentile() 248 kqd->latency_timeout[sched_domain] = 0; in calculate_percentile() 256 memset(buckets, 0, sizeof(kqd->latency_buckets[sched_domain][type])); in calculate_percentile() [all …]
|
| /OK3568_Linux_fs/kernel/kernel/sched/ |
| H A D | topology.c | 39 static int sched_domain_debug_one(struct sched_domain *sd, int cpu, int level, in sched_domain_debug_one() 135 static void sched_domain_debug(struct sched_domain *sd, int cpu) in sched_domain_debug() 175 static int sd_degenerate(struct sched_domain *sd) in sd_degenerate() 193 sd_parent_degenerate(struct sched_domain *sd, struct sched_domain *parent) in sd_parent_degenerate() 583 static void destroy_sched_domain(struct sched_domain *sd) in destroy_sched_domain() 599 struct sched_domain *sd = container_of(rcu, struct sched_domain, rcu); in destroy_sched_domains_rcu() 602 struct sched_domain *parent = sd->parent; in destroy_sched_domains_rcu() 608 static void destroy_sched_domains(struct sched_domain *sd) in destroy_sched_domains() 623 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_llc); 627 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_numa); [all …]
|
| H A D | sched.h | 993 struct sched_domain __rcu *sd; 1478 static inline struct sched_domain *highest_flag_domain(int cpu, int flag) in highest_flag_domain() 1480 struct sched_domain *sd, *hsd = NULL; in highest_flag_domain() 1491 static inline struct sched_domain *lowest_flag_domain(int cpu, int flag) in lowest_flag_domain() 1493 struct sched_domain *sd; in lowest_flag_domain() 1503 DECLARE_PER_CPU(struct sched_domain __rcu *, sd_llc); 1507 DECLARE_PER_CPU(struct sched_domain __rcu *, sd_numa); 1508 DECLARE_PER_CPU(struct sched_domain __rcu *, sd_asym_packing); 1509 DECLARE_PER_CPU(struct sched_domain __rcu *, sd_asym_cpucapacity); 1953 extern void update_group_capacity(struct sched_domain *sd, int cpu);
|
| H A D | fair.c | 2002 struct sched_domain *sd; in task_numa_migrate() 5901 wake_affine_weight(struct sched_domain *sd, struct task_struct *p, in wake_affine_weight() 5943 static int wake_affine(struct sched_domain *sd, struct task_struct *p, in wake_affine() 5964 find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu); 6022 static inline int find_idlest_cpu(struct sched_domain *sd, struct task_struct *p, in find_idlest_cpu() 6039 struct sched_domain *tmp; in find_idlest_cpu() 6133 static int select_idle_core(struct task_struct *p, struct sched_domain *sd, int target) in select_idle_core() 6172 static int select_idle_smt(struct task_struct *p, struct sched_domain *sd, int target) in select_idle_smt() 6192 static inline int select_idle_core(struct task_struct *p, struct sched_domain *sd, int target) in select_idle_core() 6197 static inline int select_idle_smt(struct task_struct *p, struct sched_domain *sd, int target) in select_idle_smt() [all …]
|
| H A D | stats.c | 25 struct sched_domain *sd; in show_schedstat()
|
| H A D | debug.c | 303 sd_alloc_ctl_domain_table(struct sched_domain *sd) in sd_alloc_ctl_domain_table() 326 struct sched_domain *sd; in sd_alloc_ctl_cpu_table()
|
| H A D | rt.c | 1795 struct sched_domain *sd; in find_lowest_rq()
|
| H A D | deadline.c | 1960 struct sched_domain *sd; in find_later_rq()
|
| H A D | core.c | 676 struct sched_domain *sd; in get_nohz_timer_target() 2639 struct sched_domain *sd; in ttwu_stat()
|
| /OK3568_Linux_fs/kernel/include/linux/sched/ |
| H A D | topology.h | 83 struct sched_domain { struct 85 struct sched_domain __rcu *parent; /* top domain must be null terminated */ argument 86 struct sched_domain __rcu *child; /* bottom domain must be null terminated */ argument 164 static inline struct cpumask *sched_domain_span(struct sched_domain *sd) in sched_domain_span() argument 188 struct sched_domain *__percpu *sd;
|
| /OK3568_Linux_fs/kernel/Documentation/scheduler/ |
| H A D | sched-domains.rst | 5 Each CPU has a "base" scheduling domain (struct sched_domain). The domain 45 the parent sched_domain (if it exists), and the parent of the parent and so 69 struct sched_domain fields, SD_FLAG_*, SD_*_INIT to get an idea of
|
| H A D | sched-capacity.rst | 287 - The SD_ASYM_CPUCAPACITY flag will be set at the lowest sched_domain level that 316 sched_asym_cpucapacity static key will be enabled. However, the sched_domain 325 the sched_domain hierarchy (if relevant, i.e. the codepath targets a specific
|
| H A D | sched-energy.rst | 335 flag to be set in the sched_domain hierarchy.
|