Lines Matching refs:lruvec
341 unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx) in lruvec_lru_size() argument
347 struct zone *zone = &lruvec_pgdat(lruvec)->node_zones[zid]; in lruvec_lru_size()
353 size += mem_cgroup_get_zone_lru_size(lruvec, lru, zid); in lruvec_lru_size()
1669 static __always_inline void update_lru_sizes(struct lruvec *lruvec, in update_lru_sizes() argument
1678 update_lru_size(lruvec, lru, zid, -nr_zone_taken[zid]); in update_lru_sizes()
1703 struct lruvec *lruvec, struct list_head *dst, in isolate_lru_pages() argument
1707 struct list_head *src = &lruvec->lists[lru]; in isolate_lru_pages()
1786 update_lru_sizes(lruvec, lru, nr_zone_taken); in isolate_lru_pages()
1825 struct lruvec *lruvec; in isolate_lru_page() local
1828 lruvec = mem_cgroup_page_lruvec(page, pgdat); in isolate_lru_page()
1833 del_page_from_lru_list(page, lruvec, lru); in isolate_lru_page()
1898 static unsigned noinline_for_stack move_pages_to_lru(struct lruvec *lruvec, in move_pages_to_lru() argument
1901 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in move_pages_to_lru()
1917 lruvec = mem_cgroup_page_lruvec(page, pgdat); in move_pages_to_lru()
1923 update_lru_size(lruvec, lru, page_zonenum(page), nr_pages); in move_pages_to_lru()
1924 list_move(&page->lru, &lruvec->lists[lru]); in move_pages_to_lru()
1930 del_page_from_lru_list(page, lruvec, lru); in move_pages_to_lru()
1941 workingset_age_nonresident(lruvec, nr_pages); in move_pages_to_lru()
1971 shrink_inactive_list(unsigned long nr_to_scan, struct lruvec *lruvec, in shrink_inactive_list() argument
1981 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in shrink_inactive_list()
2001 nr_taken = isolate_lru_pages(nr_to_scan, lruvec, &page_list, in shrink_inactive_list()
2008 __count_memcg_events(lruvec_memcg(lruvec), item, nr_scanned); in shrink_inactive_list()
2021 move_pages_to_lru(lruvec, &page_list); in shrink_inactive_list()
2024 lru_note_cost(lruvec, file, stat.nr_pageout); in shrink_inactive_list()
2028 __count_memcg_events(lruvec_memcg(lruvec), item, nr_reclaimed); in shrink_inactive_list()
2064 struct lruvec *lruvec, in shrink_active_list() argument
2078 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in shrink_active_list()
2086 nr_taken = isolate_lru_pages(nr_to_scan, lruvec, &l_hold, in shrink_active_list()
2093 __count_memcg_events(lruvec_memcg(lruvec), PGREFILL, nr_scanned); in shrink_active_list()
2157 nr_activate = move_pages_to_lru(lruvec, &l_active); in shrink_active_list()
2158 nr_deactivate = move_pages_to_lru(lruvec, &l_inactive); in shrink_active_list()
2163 __count_memcg_events(lruvec_memcg(lruvec), PGDEACTIVATE, nr_deactivate); in shrink_active_list()
2230 struct lruvec *lruvec, struct scan_control *sc) in shrink_list() argument
2234 shrink_active_list(nr_to_scan, lruvec, sc, lru); in shrink_list()
2240 return shrink_inactive_list(nr_to_scan, lruvec, sc, lru); in shrink_list()
2271 static bool inactive_is_low(struct lruvec *lruvec, enum lru_list inactive_lru) in inactive_is_low() argument
2279 inactive = lruvec_page_state(lruvec, NR_LRU_BASE + inactive_lru); in inactive_is_low()
2280 active = lruvec_page_state(lruvec, NR_LRU_BASE + active_lru); in inactive_is_low()
2314 static void get_scan_count(struct lruvec *lruvec, struct scan_control *sc, in get_scan_count() argument
2317 struct mem_cgroup *memcg = lruvec_memcg(lruvec); in get_scan_count()
2415 lruvec_size = lruvec_lru_size(lruvec, lru, sc->reclaim_idx); in get_scan_count()
2517 static void shrink_lruvec(struct lruvec *lruvec, struct scan_control *sc) in shrink_lruvec() argument
2528 get_scan_count(lruvec, sc, nr); in shrink_lruvec()
2559 lruvec, sc); in shrink_lruvec()
2624 if (total_swap_pages && inactive_is_low(lruvec, LRU_INACTIVE_ANON)) in shrink_lruvec()
2625 shrink_active_list(SWAP_CLUSTER_MAX, lruvec, in shrink_lruvec()
2707 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, pgdat); in shrink_node_memcgs() local
2749 shrink_lruvec(lruvec, sc); in shrink_node_memcgs()
2766 struct lruvec *target_lruvec; in shrink_node()
3081 struct lruvec *target_lruvec; in snapshot_refaults()
3151 struct lruvec *lruvec; in do_try_to_free_pages() local
3153 lruvec = mem_cgroup_lruvec(sc->target_mem_cgroup, in do_try_to_free_pages()
3155 clear_bit(LRUVEC_CONGESTED, &lruvec->flags); in do_try_to_free_pages()
3380 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, pgdat); in mem_cgroup_shrink_node() local
3405 shrink_lruvec(lruvec, &sc); in mem_cgroup_shrink_node()
3458 struct lruvec *lruvec; in age_active_anon() local
3463 lruvec = mem_cgroup_lruvec(NULL, pgdat); in age_active_anon()
3464 if (!inactive_is_low(lruvec, LRU_INACTIVE_ANON)) in age_active_anon()
3469 lruvec = mem_cgroup_lruvec(memcg, pgdat); in age_active_anon()
3470 shrink_active_list(SWAP_CLUSTER_MAX, lruvec, in age_active_anon()
3539 struct lruvec *lruvec = mem_cgroup_lruvec(NULL, pgdat); in clear_pgdat_congested() local
3541 clear_bit(LRUVEC_CONGESTED, &lruvec->flags); in clear_pgdat_congested()
4415 struct lruvec *lruvec; in check_move_unevictable_pages() local
4438 lruvec = mem_cgroup_page_lruvec(page, pgdat); in check_move_unevictable_pages()
4448 del_page_from_lru_list(page, lruvec, LRU_UNEVICTABLE); in check_move_unevictable_pages()
4449 add_page_to_lru_list(page, lruvec, lru); in check_move_unevictable_pages()