| /OK3568_Linux_fs/kernel/mm/ |
| H A D | memcontrol.c | 136 struct mem_cgroup *memcg; member 150 int (*register_event)(struct mem_cgroup *memcg, 157 void (*unregister_event)(struct mem_cgroup *memcg, 169 static void mem_cgroup_threshold(struct mem_cgroup *memcg); 170 static void mem_cgroup_oom_notify(struct mem_cgroup *memcg); 241 struct vmpressure *memcg_to_vmpressure(struct mem_cgroup *memcg) in memcg_to_vmpressure() argument 243 if (!memcg) in memcg_to_vmpressure() 244 memcg = root_mem_cgroup; in memcg_to_vmpressure() 245 return &memcg->vmpressure; in memcg_to_vmpressure() 259 struct mem_cgroup *memcg; in obj_cgroup_release() local [all …]
|
| H A D | vmpressure.c | 80 struct mem_cgroup *memcg = mem_cgroup_from_css(css); in vmpressure_parent() local 82 memcg = parent_mem_cgroup(memcg); in vmpressure_parent() 83 if (!memcg) in vmpressure_parent() 85 return memcg_to_vmpressure(memcg); in vmpressure_parent() 242 void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, in vmpressure() argument 251 vmpr = memcg_to_vmpressure(memcg); in vmpressure() 253 trace_android_vh_vmpressure(memcg, &bypass); in vmpressure() 295 if (!memcg || mem_cgroup_is_root(memcg)) in vmpressure() 319 memcg->socket_pressure = jiffies + HZ; in vmpressure() 335 void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio) in vmpressure_prio() argument [all …]
|
| H A D | workingset.c | 290 struct mem_cgroup *memcg; in workingset_refault() local 348 memcg = page_memcg(page); in workingset_refault() 349 lruvec = mem_cgroup_lruvec(memcg, pgdat); in workingset_refault() 365 if (mem_cgroup_get_nr_swap_pages(memcg) > 0) { in workingset_refault() 399 struct mem_cgroup *memcg; in workingset_activation() local 410 memcg = page_memcg_rcu(page); in workingset_activation() 411 if (!mem_cgroup_disabled() && !memcg) in workingset_activation() 490 if (sc->memcg) { in count_shadow_nodes() 494 lruvec = mem_cgroup_lruvec(sc->memcg, NODE_DATA(sc->nid)); in count_shadow_nodes()
|
| H A D | list_lru.c | 65 struct mem_cgroup *memcg = NULL; in list_lru_from_kmem() local 70 memcg = mem_cgroup_from_obj(ptr); in list_lru_from_kmem() 71 if (!memcg) in list_lru_from_kmem() 74 l = list_lru_from_memcg_idx(nlru, memcg_cache_id(memcg)); in list_lru_from_kmem() 77 *memcg_ptr = memcg; in list_lru_from_kmem() 119 struct mem_cgroup *memcg; in list_lru_add() local 124 l = list_lru_from_kmem(nlru, item, &memcg); in list_lru_add() 128 memcg_set_shrinker_bit(memcg, nid, in list_lru_add() 175 int nid, struct mem_cgroup *memcg) in list_lru_count_one() argument 182 l = list_lru_from_memcg_idx(nlru, memcg_cache_id(memcg)); in list_lru_count_one() [all …]
|
| H A D | vmscan.c | 572 struct mem_cgroup *memcg, int priority) in shrink_slab_memcg() argument 578 if (!mem_cgroup_online(memcg)) in shrink_slab_memcg() 584 map = rcu_dereference_protected(memcg->nodeinfo[nid]->shrinker_map, in shrink_slab_memcg() 593 .memcg = memcg, in shrink_slab_memcg() 632 memcg_set_shrinker_bit(memcg, nid, i); in shrink_slab_memcg() 647 struct mem_cgroup *memcg, int priority) in shrink_slab_memcg() argument 674 struct mem_cgroup *memcg, in shrink_slab() argument 681 trace_android_vh_shrink_slab_bypass(gfp_mask, nid, memcg, priority, &bypass); in shrink_slab() 692 if (!mem_cgroup_disabled() && !mem_cgroup_is_root(memcg)) in shrink_slab() 693 return shrink_slab_memcg(gfp_mask, nid, memcg, priority); in shrink_slab() [all …]
|
| H A D | oom_kill.c | 76 return oc->memcg != NULL; in is_memcg_oom() 264 oc->totalpages = mem_cgroup_get_max(oc->memcg) ?: 1; in constrained_alloc() 395 mem_cgroup_scan_tasks(oc->memcg, oom_evaluate_task, oc); in select_bad_process() 470 mem_cgroup_scan_tasks(oc->memcg, dump_task, oc); in dump_tasks() 488 mem_cgroup_print_oom_context(oc->memcg, victim); in dump_oom_summary() 503 mem_cgroup_print_oom_meminfo(oc->memcg); in dump_header() 1034 oom_group = mem_cgroup_get_oom_group(victim, oc->memcg); in oom_kill_process()
|
| H A D | rmap.c | 775 struct mem_cgroup *memcg; member 847 struct mem_cgroup *memcg = pra->memcg; in invalid_page_referenced_vma() local 849 if (!mm_match_cgroup(vma->vm_mm, memcg)) in invalid_page_referenced_vma() 869 struct mem_cgroup *memcg, in page_referenced() argument 875 .memcg = memcg, in page_referenced() 902 if (memcg) { in page_referenced()
|
| H A D | backing-dev.c | 430 struct mem_cgroup *memcg; in cgwb_create() local 438 memcg = mem_cgroup_from_css(memcg_css); in cgwb_create() 441 memcg_cgwb_list = &memcg->cgwb_list; in cgwb_create() 643 void wb_memcg_offline(struct mem_cgroup *memcg) in wb_memcg_offline() argument 645 struct list_head *memcg_cgwb_list = &memcg->cgwb_list; in wb_memcg_offline()
|
| H A D | slab.h | 350 struct mem_cgroup *memcg; in mod_objcg_state() local 354 memcg = obj_cgroup_memcg(objcg); in mod_objcg_state() 355 lruvec = mem_cgroup_lruvec(memcg, pgdat); in mod_objcg_state()
|
| /OK3568_Linux_fs/kernel/include/linux/ |
| H A D | memcontrol.h | 127 struct mem_cgroup *memcg; /* Back pointer, we cannot */ member 197 struct mem_cgroup *memcg; member 358 static inline bool mem_cgroup_is_root(struct mem_cgroup *memcg) in mem_cgroup_is_root() argument 360 return (memcg == root_mem_cgroup); in mem_cgroup_is_root() 369 struct mem_cgroup *memcg, in mem_cgroup_protection() argument 411 if (root == memcg) in mem_cgroup_protection() 414 *min = READ_ONCE(memcg->memory.emin); in mem_cgroup_protection() 415 *low = READ_ONCE(memcg->memory.elow); in mem_cgroup_protection() 419 struct mem_cgroup *memcg); 421 static inline bool mem_cgroup_supports_protection(struct mem_cgroup *memcg) in mem_cgroup_supports_protection() argument [all …]
|
| H A D | vmpressure.h | 33 extern void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, 35 extern void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio); 39 extern struct vmpressure *memcg_to_vmpressure(struct mem_cgroup *memcg); 41 extern int vmpressure_register_event(struct mem_cgroup *memcg, 44 extern void vmpressure_unregister_event(struct mem_cgroup *memcg, 47 static inline void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, in vmpressure() argument 49 static inline void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, in vmpressure_prio() argument
|
| H A D | list_lru.h | 117 int nid, struct mem_cgroup *memcg); 123 return list_lru_count_one(lru, sc->nid, sc->memcg); in list_lru_shrink_count() 167 int nid, struct mem_cgroup *memcg, 184 int nid, struct mem_cgroup *memcg, 195 return list_lru_walk_one(lru, sc->nid, sc->memcg, isolate, cb_arg, in list_lru_shrink_walk() 203 return list_lru_walk_one_irq(lru, sc->nid, sc->memcg, isolate, cb_arg, in list_lru_shrink_walk_irq()
|
| H A D | swap.h | 373 extern unsigned long try_to_free_mem_cgroup_pages(struct mem_cgroup *memcg, 668 static inline int mem_cgroup_swappiness(struct mem_cgroup *memcg) in mem_cgroup_swappiness() argument 675 if (mem_cgroup_disabled() || mem_cgroup_is_root(memcg)) in mem_cgroup_swappiness() 678 return memcg->swappiness; in mem_cgroup_swappiness() 719 extern long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg); 737 static inline long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg) in mem_cgroup_get_nr_swap_pages() argument
|
| H A D | rmap.h | 218 struct mem_cgroup *memcg, unsigned long *vm_flags); 315 struct mem_cgroup *memcg, in page_referenced() argument
|
| H A D | shrinker.h | 35 struct mem_cgroup *memcg; member
|
| H A D | backing-dev.h | 177 void wb_memcg_offline(struct mem_cgroup *memcg); 380 static inline void wb_memcg_offline(struct mem_cgroup *memcg) in wb_memcg_offline() argument
|
| H A D | oom.h | 37 struct mem_cgroup *memcg; member
|
| /OK3568_Linux_fs/kernel/tools/testing/selftests/cgroup/ |
| H A D | test_memcontrol.c | 161 char *memcg; in test_memcg_current() local 163 memcg = cg_name(root, "memcg_test"); in test_memcg_current() 164 if (!memcg) in test_memcg_current() 167 if (cg_create(memcg)) in test_memcg_current() 170 current = cg_read_long(memcg, "memory.current"); in test_memcg_current() 174 if (cg_run(memcg, alloc_anon_50M_check, NULL)) in test_memcg_current() 177 if (cg_run(memcg, alloc_pagecache_50M_check, NULL)) in test_memcg_current() 183 cg_destroy(memcg); in test_memcg_current() 184 free(memcg); in test_memcg_current() 590 char *memcg; in test_memcg_high() local [all …]
|
| /OK3568_Linux_fs/kernel/include/trace/hooks/ |
| H A D | mm.h | 127 TP_PROTO(struct mem_cgroup *memcg, bool *bypass), 128 TP_ARGS(memcg, bypass)); 130 TP_PROTO(struct mem_cgroup *memcg), 131 TP_ARGS(memcg)); 133 TP_PROTO(struct mem_cgroup *memcg), 134 TP_ARGS(memcg)); 136 TP_PROTO(struct mem_cgroup *memcg), 137 TP_ARGS(memcg)); 139 TP_PROTO(struct cgroup_subsys_state *css, struct mem_cgroup *memcg), 140 TP_ARGS(css, memcg)); [all …]
|
| H A D | vmscan.h | 20 TP_PROTO(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg, int priority, bool *bypass), 21 TP_ARGS(gfp_mask, nid, memcg, priority, bypass)); 44 TP_PROTO(struct mem_cgroup *memcg, bool *skip), 45 TP_ARGS(memcg, skip));
|
| /OK3568_Linux_fs/kernel/Documentation/admin-guide/cgroup-v1/ |
| H A D | memcg_test.rst | 9 Because VM is getting complex (one of reasons is memcg...), memcg's behavior 10 is complex. This is a document for memcg's internal behavior. 61 At commit(), the page is associated with the memcg. 114 But brief explanation of the behavior of memcg around shmem will be 136 Each memcg has its own private LRU. Now, its handling is under global 138 Almost all routines around memcg's LRU is called by global LRU's 142 memcg's private LRU and call __isolate_lru_page() to extract a page 154 9.1 Small limit to memcg. 157 When you do test to do racy case, it's good test to set memcg's limit 167 Historically, memcg's shmem handling was poor and we saw some amount [all …]
|
| H A D | memory.rst | 19 see patch's title and function names tend to use "memcg". 47 - pages are linked to per-memcg LRU exclusively, and there is no global LRU. 205 Since page's memcg recorded into swap whatever memsw enabled, the page will 337 from inside the memcg. The creation is done lazily, so some objects can still be 339 belong to the same memcg. This only fails to hold when a task is migrated to a 340 different memcg during the page allocation by the cache. 359 This is the standard memcg limitation mechanism already present before kmem 521 Though rmdir() offlines memcg, but the memcg may still stay there due to 613 there is a swap storage available. This might lead to memcg OOM killer 641 This is similar to numa_maps but operates on a per-memcg basis. This is [all …]
|
| /OK3568_Linux_fs/kernel/tools/cgroup/ |
| H A D | memcg_slabinfo.py | 42 memcg = container_of(css, 'struct mem_cgroup', 'css') 43 MEMCGS[css.cgroup.kn.id.value_()] = memcg 171 memcg = MEMCGS[cgroup_id] 187 obj_cgroups.add(memcg.objcg.value_()) 189 memcg.objcg_list.address_of_(), 221 memcg.kmem_caches.address_of_(),
|
| /OK3568_Linux_fs/kernel/include/linux/sched/ |
| H A D | mm.h | 304 set_active_memcg(struct mem_cgroup *memcg) in set_active_memcg() argument 310 this_cpu_write(int_active_memcg, memcg); in set_active_memcg() 313 current->active_memcg = memcg; in set_active_memcg() 320 set_active_memcg(struct mem_cgroup *memcg) in set_active_memcg() argument
|
| /OK3568_Linux_fs/kernel/Documentation/vm/ |
| H A D | hwpoison.rst | 136 corrupt-filter-memcg 138 number of the memcg. 148 echo $memcg_ino > /debug/hwpoison/corrupt-filter-memcg
|