| /OK3568_Linux_fs/kernel/mm/ |
| H A D | memory_hotplug.c | 446 for (zone = pgdat->node_zones; in update_pgdat_span() 447 zone < pgdat->node_zones + MAX_NR_ZONES; zone++) { in update_pgdat_span() 734 struct zone *zone = &pgdat->node_zones[zid]; in default_kernel_zone_for_pfn() 740 return &pgdat->node_zones[ZONE_NORMAL]; in default_kernel_zone_for_pfn() 748 struct zone *movable_zone = &NODE_DATA(nid)->node_zones[ZONE_MOVABLE]; in default_zone_for_pfn() 774 return &NODE_DATA(nid)->node_zones[ZONE_MOVABLE]; in zone_for_pfn_range() 875 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_present_pages() 1447 present_pages += pgdat->node_zones[zt].present_pages; in node_states_check_changes_offline() 1460 present_pages += pgdat->node_zones[ZONE_HIGHMEM].present_pages; in node_states_check_changes_offline() 1475 present_pages += pgdat->node_zones[ZONE_MOVABLE].present_pages; in node_states_check_changes_offline()
|
| H A D | mmzone.c | 35 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone() 40 zone = pgdat->node_zones; in next_zone()
|
| H A D | shuffle.c | 158 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in __shuffle_free_memory()
|
| H A D | page_owner.c | 686 struct zone *node_zones = pgdat->node_zones; in init_zones_in_node() local 688 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in init_zones_in_node()
|
| H A D | vmstat.c | 300 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold() 976 struct zone *zones = NODE_DATA(node)->node_zones; in sum_zone_node_page_state() 993 struct zone *zones = NODE_DATA(node)->node_zones; in sum_zone_numa_state() 1399 struct zone *node_zones = pgdat->node_zones; in walk_zones_in_node() local 1402 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in walk_zones_in_node() 1611 struct zone *compare = &pgdat->node_zones[zid]; in is_zone_first_populated()
|
| H A D | vmscan.c | 347 struct zone *zone = &lruvec_pgdat(lruvec)->node_zones[zid]; in lruvec_lru_size() 2674 struct zone *zone = &pgdat->node_zones[z]; in should_continue_reclaim() 2846 struct zone *zone = &pgdat->node_zones[z]; in shrink_node() 3208 zone = &pgdat->node_zones[i]; in allow_direct_reclaim() 3489 zone = pgdat->node_zones + i; in pgdat_watermark_boosted() 3515 zone = pgdat->node_zones + i; in pgdat_balanced() 3600 zone = pgdat->node_zones + z; in kswapd_shrink_node() 3668 zone = pgdat->node_zones + i; in balance_pgdat() 3699 zone = pgdat->node_zones + i; in balance_pgdat() 3820 zone = pgdat->node_zones + i; in balance_pgdat()
|
| H A D | memremap.c | 278 zone = &NODE_DATA(nid)->node_zones[ZONE_DEVICE]; in pagemap_range() 292 memmap_init_zone_device(&NODE_DATA(nid)->node_zones[ZONE_DEVICE], in pagemap_range()
|
| H A D | mm_init.c | 48 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
|
| H A D | compaction.c | 391 struct zone *zone = &pgdat->node_zones[zoneid]; in reset_isolation_suitable() 1959 zone = &pgdat->node_zones[zoneid]; in fragmentation_score_node() 2598 zone = &pgdat->node_zones[zoneid]; in proactive_compact_node() 2628 zone = &pgdat->node_zones[zoneid]; in compact_node() 2742 zone = &pgdat->node_zones[zoneid]; in kcompactd_node_suitable() 2778 zone = &pgdat->node_zones[zoneid]; in kcompactd_do_work()
|
| H A D | page_alloc.c | 1568 struct zone *zone = &pgdat->node_zones[zid]; in init_reserved_page() 2042 zone = pgdat->node_zones + zid; in deferred_init_memmap() 5597 managed_pages += zone_managed_pages(&pgdat->node_zones[zone_type]); in si_meminfo_node() 5603 struct zone *zone = &pgdat->node_zones[zone_type]; in si_meminfo_node() 5893 zone = pgdat->node_zones + zone_type; in build_zonerefs_node() 6516 struct zone *zone = node->node_zones + j; in memmap_init() 6975 struct zone *zone = pgdat->node_zones + i; in calculate_node_totalpages() 7172 zone_init_internals(&pgdat->node_zones[z], z, nid, 0); in free_area_init_core_hotplug() 7194 struct zone *zone = pgdat->node_zones + j; in free_area_init_core() 7661 struct zone *zone = &pgdat->node_zones[zone_type]; in check_for_memory() [all …]
|
| H A D | memblock.c | 2030 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_managed_pages()
|
| H A D | page-writeback.c | 286 struct zone *zone = pgdat->node_zones + z; in node_dirtyable_memory() 322 z = &NODE_DATA(node)->node_zones[i]; in highmem_dirtyable_memory()
|
| H A D | migrate.c | 1982 struct zone *zone = pgdat->node_zones + z; in migrate_balanced_pgdat()
|
| /OK3568_Linux_fs/kernel/lib/ |
| H A D | show_mem.c | 23 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem()
|
| /OK3568_Linux_fs/kernel/include/linux/ |
| H A D | mmzone.h | 735 struct zone node_zones[MAX_NR_ZONES]; member 906 #define zone_idx(zone) ((zone) - (zone)->zone_pgdat->node_zones) 1046 for (zone = (first_online_pgdat())->node_zones; \ 1051 for (zone = (first_online_pgdat())->node_zones; \
|
| H A D | mm_inline.h | 38 __mod_zone_page_state(&pgdat->node_zones[zid], in __update_lru_size()
|
| H A D | dma-direct.h | 32 return zone_is_empty(&pgdat->node_zones[ZONE_DMA32]); in zone_dma32_is_empty()
|
| H A D | mm.h | 1527 return &NODE_DATA(page_to_nid(page))->node_zones[page_zonenum(page)]; in page_zone()
|
| /OK3568_Linux_fs/kernel/kernel/ |
| H A D | crash_core.c | 484 VMCOREINFO_OFFSET(pglist_data, node_zones); in crash_save_vmcoreinfo_init()
|
| /OK3568_Linux_fs/kernel/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_crat.c | 897 mem_in_bytes += zone_managed_pages(&pgdat->node_zones[zone_type]); in kfd_fill_mem_info_for_cpu()
|
| /OK3568_Linux_fs/kernel/Documentation/admin-guide/kdump/ |
| H A D | vmcoreinfo.rst | 144 (pglist_data, node_zones|nr_zones|node_mem_map|node_start_pfn|node_spanned_pages|node_id)
|