| /OK3568_Linux_fs/kernel/include/linux/ |
| H A D | oom.h | 13 struct zonelist; 31 struct zonelist *zonelist; member
|
| H A D | mmzone.h | 704 struct zonelist { struct 742 struct zonelist node_zonelists[MAX_ZONELISTS]; 1114 static inline struct zoneref *first_zones_zonelist(struct zonelist *zonelist, in first_zones_zonelist() argument 1118 return next_zones_zonelist(zonelist->_zonerefs, in first_zones_zonelist()
|
| H A D | swap.h | 370 extern unsigned long try_to_free_pages(struct zonelist *zonelist, int order,
|
| H A D | gfp.h | 504 static inline struct zonelist *node_zonelist(int nid, gfp_t flags) in node_zonelist()
|
| /OK3568_Linux_fs/kernel/mm/ |
| H A D | mm_init.c | 38 struct zonelist *zonelist; in mminit_verify_zonelist() local 47 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 58 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
|
| H A D | vmscan.c | 2994 static void shrink_zones(struct zonelist *zonelist, struct scan_control *sc) in shrink_zones() argument 3014 for_each_zone_zonelist_nodemask(zone, z, zonelist, in shrink_zones() 3108 static unsigned long do_try_to_free_pages(struct zonelist *zonelist, in do_try_to_free_pages() argument 3125 shrink_zones(zonelist, sc); in do_try_to_free_pages() 3142 for_each_zone_zonelist_nodemask(zone, z, zonelist, sc->reclaim_idx, in do_try_to_free_pages() 3245 static bool throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, in throttle_direct_reclaim() argument 3283 for_each_zone_zonelist_nodemask(zone, z, zonelist, in throttle_direct_reclaim() 3329 unsigned long try_to_free_pages(struct zonelist *zonelist, int order, in try_to_free_pages() argument 3358 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages() 3364 nr_reclaimed = do_try_to_free_pages(zonelist, &sc); in try_to_free_pages() [all …]
|
| H A D | internal.h | 157 struct zonelist *zonelist; member
|
| H A D | page_alloc.c | 2793 struct zonelist *zonelist = ac->zonelist; in unreserve_highatomic_pageblock() local 2801 for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->highest_zoneidx, in unreserve_highatomic_pageblock() 4181 .zonelist = ac->zonelist, in __alloc_pages_may_oom() 4430 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_compact_retry() 4530 progress = try_to_free_pages(ac->zonelist, order, gfp_mask, in __perform_reclaim() 4588 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, highest_zoneidx, in wake_all_kswapds() 4724 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_reclaim_retry() 4861 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 4948 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 5104 ac->zonelist = node_zonelist(preferred_nid, gfp_mask); in prepare_alloc_pages() [all …]
|
| H A D | oom_kill.c | 274 if (!oc->zonelist) in constrained_alloc() 298 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc()
|
| H A D | mempolicy.c | 1956 struct zonelist *zonelist; in mempolicy_slab_node() local 1958 zonelist = &NODE_DATA(node)->node_zonelists[ZONELIST_FALLBACK]; in mempolicy_slab_node() 1959 z = first_zones_zonelist(zonelist, highest_zoneidx, in mempolicy_slab_node()
|
| H A D | slab.c | 3097 struct zonelist *zonelist; in fallback_alloc() local 3111 zonelist = node_zonelist(mempolicy_slab_node(), flags); in fallback_alloc() 3118 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in fallback_alloc()
|
| H A D | compaction.c | 2210 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in compaction_zonelist_suitable() 2526 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in try_to_compact_pages()
|
| H A D | hugetlb.c | 1103 struct zonelist *zonelist; in dequeue_huge_page_nodemask() local 1108 zonelist = node_zonelist(nid, gfp_mask); in dequeue_huge_page_nodemask() 1112 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_huge_page_nodemask()
|
| H A D | slub.c | 2060 struct zonelist *zonelist; in get_any_partial() local 2091 zonelist = node_zonelist(mempolicy_slab_node(), flags); in get_any_partial() 2092 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in get_any_partial()
|
| H A D | memcontrol.c | 1747 .zonelist = NULL, in mem_cgroup_out_of_memory()
|
| /OK3568_Linux_fs/yocto/poky/meta/recipes-multimedia/libtiff/tiff/ |
| H A D | CVE-2022-2867.patch | 105 - seg = crop->zonelist[i].position; 106 - total = crop->zonelist[i].total; 107 + seg = crop->zonelist[j].position; 108 + total = crop->zonelist[j].total;
|
| H A D | 0001-tiffcrop-Fix-issue-330-and-some-more-from-320-to-349.patch | 384 seg = crop->zonelist[j].position; 385 total = crop->zonelist[j].total; 389 + if (crop->zonelist[j].position < 0 || crop->zonelist[j].total < 0) { 390 … zone values %d:%d are not allowed, thus skipped.", crop->zonelist[j].position, crop->zonelist[j].…
|
| /OK3568_Linux_fs/kernel/Documentation/vm/ |
| H A D | numa.rst | 76 an ordered "zonelist". A zonelist specifies the zones/nodes to visit when a 86 a default Node ordered zonelist. This means it tries to fallback to other zones 91 Linux will attempt to allocate from the first node in the appropriate zonelist 94 nodes' zones in the selected zonelist looking for the first zone in the list 122 zonelist--will not be the node itself. Rather, it will be the node that the
|
| /OK3568_Linux_fs/kernel/drivers/tty/ |
| H A D | sysrq.c | 386 .zonelist = node_zonelist(first_memory_node, gfp_mask), in moom_callback()
|
| /OK3568_Linux_fs/kernel/Documentation/admin-guide/sysctl/ |
| H A D | vm.rst | 634 In non-NUMA case, a zonelist for GFP_KERNEL is ordered as following. 640 Assume 2 node NUMA and below is zonelist of Node(0)'s GFP_KERNEL::
|
| /OK3568_Linux_fs/kernel/Documentation/admin-guide/mm/ |
| H A D | numa_memory_policy.rst | 236 node zonelist.
|
| /OK3568_Linux_fs/kernel/Documentation/admin-guide/ |
| H A D | kernel-parameters.txt | 3577 numa_zonelist_order= [KNL, BOOT] Select zonelist order for NUMA.
|