1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * mm_init.c - Memory initialisation verification and debugging
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * Copyright 2008 IBM Corporation, 2008
6*4882a593Smuzhiyun * Author Mel Gorman <mel@csn.ul.ie>
7*4882a593Smuzhiyun *
8*4882a593Smuzhiyun */
9*4882a593Smuzhiyun #include <linux/kernel.h>
10*4882a593Smuzhiyun #include <linux/init.h>
11*4882a593Smuzhiyun #include <linux/kobject.h>
12*4882a593Smuzhiyun #include <linux/export.h>
13*4882a593Smuzhiyun #include <linux/memory.h>
14*4882a593Smuzhiyun #include <linux/notifier.h>
15*4882a593Smuzhiyun #include <linux/sched.h>
16*4882a593Smuzhiyun #include <linux/mman.h>
17*4882a593Smuzhiyun #include "internal.h"
18*4882a593Smuzhiyun
19*4882a593Smuzhiyun #ifdef CONFIG_DEBUG_MEMORY_INIT
20*4882a593Smuzhiyun int __meminitdata mminit_loglevel;
21*4882a593Smuzhiyun
22*4882a593Smuzhiyun #ifndef SECTIONS_SHIFT
23*4882a593Smuzhiyun #define SECTIONS_SHIFT 0
24*4882a593Smuzhiyun #endif
25*4882a593Smuzhiyun
26*4882a593Smuzhiyun /* The zonelists are simply reported, validation is manual. */
mminit_verify_zonelist(void)27*4882a593Smuzhiyun void __init mminit_verify_zonelist(void)
28*4882a593Smuzhiyun {
29*4882a593Smuzhiyun int nid;
30*4882a593Smuzhiyun
31*4882a593Smuzhiyun if (mminit_loglevel < MMINIT_VERIFY)
32*4882a593Smuzhiyun return;
33*4882a593Smuzhiyun
34*4882a593Smuzhiyun for_each_online_node(nid) {
35*4882a593Smuzhiyun pg_data_t *pgdat = NODE_DATA(nid);
36*4882a593Smuzhiyun struct zone *zone;
37*4882a593Smuzhiyun struct zoneref *z;
38*4882a593Smuzhiyun struct zonelist *zonelist;
39*4882a593Smuzhiyun int i, listid, zoneid;
40*4882a593Smuzhiyun
41*4882a593Smuzhiyun BUILD_BUG_ON(MAX_ZONELISTS > 2);
42*4882a593Smuzhiyun for (i = 0; i < MAX_ZONELISTS * MAX_NR_ZONES; i++) {
43*4882a593Smuzhiyun
44*4882a593Smuzhiyun /* Identify the zone and nodelist */
45*4882a593Smuzhiyun zoneid = i % MAX_NR_ZONES;
46*4882a593Smuzhiyun listid = i / MAX_NR_ZONES;
47*4882a593Smuzhiyun zonelist = &pgdat->node_zonelists[listid];
48*4882a593Smuzhiyun zone = &pgdat->node_zones[zoneid];
49*4882a593Smuzhiyun if (!populated_zone(zone))
50*4882a593Smuzhiyun continue;
51*4882a593Smuzhiyun
52*4882a593Smuzhiyun /* Print information about the zonelist */
53*4882a593Smuzhiyun printk(KERN_DEBUG "mminit::zonelist %s %d:%s = ",
54*4882a593Smuzhiyun listid > 0 ? "thisnode" : "general", nid,
55*4882a593Smuzhiyun zone->name);
56*4882a593Smuzhiyun
57*4882a593Smuzhiyun /* Iterate the zonelist */
58*4882a593Smuzhiyun for_each_zone_zonelist(zone, z, zonelist, zoneid)
59*4882a593Smuzhiyun pr_cont("%d:%s ", zone_to_nid(zone), zone->name);
60*4882a593Smuzhiyun pr_cont("\n");
61*4882a593Smuzhiyun }
62*4882a593Smuzhiyun }
63*4882a593Smuzhiyun }
64*4882a593Smuzhiyun
mminit_verify_pageflags_layout(void)65*4882a593Smuzhiyun void __init mminit_verify_pageflags_layout(void)
66*4882a593Smuzhiyun {
67*4882a593Smuzhiyun int shift, width;
68*4882a593Smuzhiyun unsigned long or_mask, add_mask;
69*4882a593Smuzhiyun
70*4882a593Smuzhiyun shift = 8 * sizeof(unsigned long);
71*4882a593Smuzhiyun width = shift - SECTIONS_WIDTH - NODES_WIDTH - ZONES_WIDTH
72*4882a593Smuzhiyun - LAST_CPUPID_SHIFT - KASAN_TAG_WIDTH;
73*4882a593Smuzhiyun mminit_dprintk(MMINIT_TRACE, "pageflags_layout_widths",
74*4882a593Smuzhiyun "Section %d Node %d Zone %d Lastcpupid %d Kasantag %d Flags %d\n",
75*4882a593Smuzhiyun SECTIONS_WIDTH,
76*4882a593Smuzhiyun NODES_WIDTH,
77*4882a593Smuzhiyun ZONES_WIDTH,
78*4882a593Smuzhiyun LAST_CPUPID_WIDTH,
79*4882a593Smuzhiyun KASAN_TAG_WIDTH,
80*4882a593Smuzhiyun NR_PAGEFLAGS);
81*4882a593Smuzhiyun mminit_dprintk(MMINIT_TRACE, "pageflags_layout_shifts",
82*4882a593Smuzhiyun "Section %d Node %d Zone %d Lastcpupid %d Kasantag %d\n",
83*4882a593Smuzhiyun SECTIONS_SHIFT,
84*4882a593Smuzhiyun NODES_SHIFT,
85*4882a593Smuzhiyun ZONES_SHIFT,
86*4882a593Smuzhiyun LAST_CPUPID_SHIFT,
87*4882a593Smuzhiyun KASAN_TAG_WIDTH);
88*4882a593Smuzhiyun mminit_dprintk(MMINIT_TRACE, "pageflags_layout_pgshifts",
89*4882a593Smuzhiyun "Section %lu Node %lu Zone %lu Lastcpupid %lu Kasantag %lu\n",
90*4882a593Smuzhiyun (unsigned long)SECTIONS_PGSHIFT,
91*4882a593Smuzhiyun (unsigned long)NODES_PGSHIFT,
92*4882a593Smuzhiyun (unsigned long)ZONES_PGSHIFT,
93*4882a593Smuzhiyun (unsigned long)LAST_CPUPID_PGSHIFT,
94*4882a593Smuzhiyun (unsigned long)KASAN_TAG_PGSHIFT);
95*4882a593Smuzhiyun mminit_dprintk(MMINIT_TRACE, "pageflags_layout_nodezoneid",
96*4882a593Smuzhiyun "Node/Zone ID: %lu -> %lu\n",
97*4882a593Smuzhiyun (unsigned long)(ZONEID_PGOFF + ZONEID_SHIFT),
98*4882a593Smuzhiyun (unsigned long)ZONEID_PGOFF);
99*4882a593Smuzhiyun mminit_dprintk(MMINIT_TRACE, "pageflags_layout_usage",
100*4882a593Smuzhiyun "location: %d -> %d layout %d -> %d unused %d -> %d page-flags\n",
101*4882a593Smuzhiyun shift, width, width, NR_PAGEFLAGS, NR_PAGEFLAGS, 0);
102*4882a593Smuzhiyun #ifdef NODE_NOT_IN_PAGE_FLAGS
103*4882a593Smuzhiyun mminit_dprintk(MMINIT_TRACE, "pageflags_layout_nodeflags",
104*4882a593Smuzhiyun "Node not in page flags");
105*4882a593Smuzhiyun #endif
106*4882a593Smuzhiyun #ifdef LAST_CPUPID_NOT_IN_PAGE_FLAGS
107*4882a593Smuzhiyun mminit_dprintk(MMINIT_TRACE, "pageflags_layout_nodeflags",
108*4882a593Smuzhiyun "Last cpupid not in page flags");
109*4882a593Smuzhiyun #endif
110*4882a593Smuzhiyun
111*4882a593Smuzhiyun if (SECTIONS_WIDTH) {
112*4882a593Smuzhiyun shift -= SECTIONS_WIDTH;
113*4882a593Smuzhiyun BUG_ON(shift != SECTIONS_PGSHIFT);
114*4882a593Smuzhiyun }
115*4882a593Smuzhiyun if (NODES_WIDTH) {
116*4882a593Smuzhiyun shift -= NODES_WIDTH;
117*4882a593Smuzhiyun BUG_ON(shift != NODES_PGSHIFT);
118*4882a593Smuzhiyun }
119*4882a593Smuzhiyun if (ZONES_WIDTH) {
120*4882a593Smuzhiyun shift -= ZONES_WIDTH;
121*4882a593Smuzhiyun BUG_ON(shift != ZONES_PGSHIFT);
122*4882a593Smuzhiyun }
123*4882a593Smuzhiyun
124*4882a593Smuzhiyun /* Check for bitmask overlaps */
125*4882a593Smuzhiyun or_mask = (ZONES_MASK << ZONES_PGSHIFT) |
126*4882a593Smuzhiyun (NODES_MASK << NODES_PGSHIFT) |
127*4882a593Smuzhiyun (SECTIONS_MASK << SECTIONS_PGSHIFT);
128*4882a593Smuzhiyun add_mask = (ZONES_MASK << ZONES_PGSHIFT) +
129*4882a593Smuzhiyun (NODES_MASK << NODES_PGSHIFT) +
130*4882a593Smuzhiyun (SECTIONS_MASK << SECTIONS_PGSHIFT);
131*4882a593Smuzhiyun BUG_ON(or_mask != add_mask);
132*4882a593Smuzhiyun }
133*4882a593Smuzhiyun
set_mminit_loglevel(char * str)134*4882a593Smuzhiyun static __init int set_mminit_loglevel(char *str)
135*4882a593Smuzhiyun {
136*4882a593Smuzhiyun get_option(&str, &mminit_loglevel);
137*4882a593Smuzhiyun return 0;
138*4882a593Smuzhiyun }
139*4882a593Smuzhiyun early_param("mminit_loglevel", set_mminit_loglevel);
140*4882a593Smuzhiyun #endif /* CONFIG_DEBUG_MEMORY_INIT */
141*4882a593Smuzhiyun
142*4882a593Smuzhiyun struct kobject *mm_kobj;
143*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(mm_kobj);
144*4882a593Smuzhiyun
145*4882a593Smuzhiyun #ifdef CONFIG_SMP
146*4882a593Smuzhiyun s32 vm_committed_as_batch = 32;
147*4882a593Smuzhiyun
mm_compute_batch(int overcommit_policy)148*4882a593Smuzhiyun void mm_compute_batch(int overcommit_policy)
149*4882a593Smuzhiyun {
150*4882a593Smuzhiyun u64 memsized_batch;
151*4882a593Smuzhiyun s32 nr = num_present_cpus();
152*4882a593Smuzhiyun s32 batch = max_t(s32, nr*2, 32);
153*4882a593Smuzhiyun unsigned long ram_pages = totalram_pages();
154*4882a593Smuzhiyun
155*4882a593Smuzhiyun /*
156*4882a593Smuzhiyun * For policy OVERCOMMIT_NEVER, set batch size to 0.4% of
157*4882a593Smuzhiyun * (total memory/#cpus), and lift it to 25% for other policies
158*4882a593Smuzhiyun * to easy the possible lock contention for percpu_counter
159*4882a593Smuzhiyun * vm_committed_as, while the max limit is INT_MAX
160*4882a593Smuzhiyun */
161*4882a593Smuzhiyun if (overcommit_policy == OVERCOMMIT_NEVER)
162*4882a593Smuzhiyun memsized_batch = min_t(u64, ram_pages/nr/256, INT_MAX);
163*4882a593Smuzhiyun else
164*4882a593Smuzhiyun memsized_batch = min_t(u64, ram_pages/nr/4, INT_MAX);
165*4882a593Smuzhiyun
166*4882a593Smuzhiyun vm_committed_as_batch = max_t(s32, memsized_batch, batch);
167*4882a593Smuzhiyun }
168*4882a593Smuzhiyun
mm_compute_batch_notifier(struct notifier_block * self,unsigned long action,void * arg)169*4882a593Smuzhiyun static int __meminit mm_compute_batch_notifier(struct notifier_block *self,
170*4882a593Smuzhiyun unsigned long action, void *arg)
171*4882a593Smuzhiyun {
172*4882a593Smuzhiyun switch (action) {
173*4882a593Smuzhiyun case MEM_ONLINE:
174*4882a593Smuzhiyun case MEM_OFFLINE:
175*4882a593Smuzhiyun mm_compute_batch(sysctl_overcommit_memory);
176*4882a593Smuzhiyun default:
177*4882a593Smuzhiyun break;
178*4882a593Smuzhiyun }
179*4882a593Smuzhiyun return NOTIFY_OK;
180*4882a593Smuzhiyun }
181*4882a593Smuzhiyun
182*4882a593Smuzhiyun static struct notifier_block compute_batch_nb __meminitdata = {
183*4882a593Smuzhiyun .notifier_call = mm_compute_batch_notifier,
184*4882a593Smuzhiyun .priority = IPC_CALLBACK_PRI, /* use lowest priority */
185*4882a593Smuzhiyun };
186*4882a593Smuzhiyun
mm_compute_batch_init(void)187*4882a593Smuzhiyun static int __init mm_compute_batch_init(void)
188*4882a593Smuzhiyun {
189*4882a593Smuzhiyun mm_compute_batch(sysctl_overcommit_memory);
190*4882a593Smuzhiyun register_hotmemory_notifier(&compute_batch_nb);
191*4882a593Smuzhiyun
192*4882a593Smuzhiyun return 0;
193*4882a593Smuzhiyun }
194*4882a593Smuzhiyun
195*4882a593Smuzhiyun __initcall(mm_compute_batch_init);
196*4882a593Smuzhiyun
197*4882a593Smuzhiyun #endif
198*4882a593Smuzhiyun
mm_sysfs_init(void)199*4882a593Smuzhiyun static int __init mm_sysfs_init(void)
200*4882a593Smuzhiyun {
201*4882a593Smuzhiyun mm_kobj = kobject_create_and_add("mm", kernel_kobj);
202*4882a593Smuzhiyun if (!mm_kobj)
203*4882a593Smuzhiyun return -ENOMEM;
204*4882a593Smuzhiyun
205*4882a593Smuzhiyun return 0;
206*4882a593Smuzhiyun }
207*4882a593Smuzhiyun postcore_initcall(mm_sysfs_init);
208