xref: /OK3568_Linux_fs/kernel/samples/bpf/map_perf_test_user.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /* Copyright (c) 2016 Facebook
3*4882a593Smuzhiyun  */
4*4882a593Smuzhiyun #define _GNU_SOURCE
5*4882a593Smuzhiyun #include <sched.h>
6*4882a593Smuzhiyun #include <stdio.h>
7*4882a593Smuzhiyun #include <sys/types.h>
8*4882a593Smuzhiyun #include <asm/unistd.h>
9*4882a593Smuzhiyun #include <unistd.h>
10*4882a593Smuzhiyun #include <assert.h>
11*4882a593Smuzhiyun #include <sys/wait.h>
12*4882a593Smuzhiyun #include <stdlib.h>
13*4882a593Smuzhiyun #include <signal.h>
14*4882a593Smuzhiyun #include <string.h>
15*4882a593Smuzhiyun #include <time.h>
16*4882a593Smuzhiyun #include <sys/resource.h>
17*4882a593Smuzhiyun #include <arpa/inet.h>
18*4882a593Smuzhiyun #include <errno.h>
19*4882a593Smuzhiyun 
20*4882a593Smuzhiyun #include <bpf/bpf.h>
21*4882a593Smuzhiyun #include <bpf/libbpf.h>
22*4882a593Smuzhiyun 
23*4882a593Smuzhiyun #define TEST_BIT(t) (1U << (t))
24*4882a593Smuzhiyun #define MAX_NR_CPUS 1024
25*4882a593Smuzhiyun 
time_get_ns(void)26*4882a593Smuzhiyun static __u64 time_get_ns(void)
27*4882a593Smuzhiyun {
28*4882a593Smuzhiyun 	struct timespec ts;
29*4882a593Smuzhiyun 
30*4882a593Smuzhiyun 	clock_gettime(CLOCK_MONOTONIC, &ts);
31*4882a593Smuzhiyun 	return ts.tv_sec * 1000000000ull + ts.tv_nsec;
32*4882a593Smuzhiyun }
33*4882a593Smuzhiyun 
34*4882a593Smuzhiyun enum test_type {
35*4882a593Smuzhiyun 	HASH_PREALLOC,
36*4882a593Smuzhiyun 	PERCPU_HASH_PREALLOC,
37*4882a593Smuzhiyun 	HASH_KMALLOC,
38*4882a593Smuzhiyun 	PERCPU_HASH_KMALLOC,
39*4882a593Smuzhiyun 	LRU_HASH_PREALLOC,
40*4882a593Smuzhiyun 	NOCOMMON_LRU_HASH_PREALLOC,
41*4882a593Smuzhiyun 	LPM_KMALLOC,
42*4882a593Smuzhiyun 	HASH_LOOKUP,
43*4882a593Smuzhiyun 	ARRAY_LOOKUP,
44*4882a593Smuzhiyun 	INNER_LRU_HASH_PREALLOC,
45*4882a593Smuzhiyun 	LRU_HASH_LOOKUP,
46*4882a593Smuzhiyun 	NR_TESTS,
47*4882a593Smuzhiyun };
48*4882a593Smuzhiyun 
49*4882a593Smuzhiyun const char *test_map_names[NR_TESTS] = {
50*4882a593Smuzhiyun 	[HASH_PREALLOC] = "hash_map",
51*4882a593Smuzhiyun 	[PERCPU_HASH_PREALLOC] = "percpu_hash_map",
52*4882a593Smuzhiyun 	[HASH_KMALLOC] = "hash_map_alloc",
53*4882a593Smuzhiyun 	[PERCPU_HASH_KMALLOC] = "percpu_hash_map_alloc",
54*4882a593Smuzhiyun 	[LRU_HASH_PREALLOC] = "lru_hash_map",
55*4882a593Smuzhiyun 	[NOCOMMON_LRU_HASH_PREALLOC] = "nocommon_lru_hash_map",
56*4882a593Smuzhiyun 	[LPM_KMALLOC] = "lpm_trie_map_alloc",
57*4882a593Smuzhiyun 	[HASH_LOOKUP] = "hash_map",
58*4882a593Smuzhiyun 	[ARRAY_LOOKUP] = "array_map",
59*4882a593Smuzhiyun 	[INNER_LRU_HASH_PREALLOC] = "inner_lru_hash_map",
60*4882a593Smuzhiyun 	[LRU_HASH_LOOKUP] = "lru_hash_lookup_map",
61*4882a593Smuzhiyun };
62*4882a593Smuzhiyun 
63*4882a593Smuzhiyun enum map_idx {
64*4882a593Smuzhiyun 	array_of_lru_hashs_idx,
65*4882a593Smuzhiyun 	hash_map_alloc_idx,
66*4882a593Smuzhiyun 	lru_hash_lookup_idx,
67*4882a593Smuzhiyun 	NR_IDXES,
68*4882a593Smuzhiyun };
69*4882a593Smuzhiyun 
70*4882a593Smuzhiyun static int map_fd[NR_IDXES];
71*4882a593Smuzhiyun 
72*4882a593Smuzhiyun static int test_flags = ~0;
73*4882a593Smuzhiyun static uint32_t num_map_entries;
74*4882a593Smuzhiyun static uint32_t inner_lru_hash_size;
75*4882a593Smuzhiyun static int lru_hash_lookup_test_entries = 32;
76*4882a593Smuzhiyun static uint32_t max_cnt = 1000000;
77*4882a593Smuzhiyun 
check_test_flags(enum test_type t)78*4882a593Smuzhiyun static int check_test_flags(enum test_type t)
79*4882a593Smuzhiyun {
80*4882a593Smuzhiyun 	return test_flags & TEST_BIT(t);
81*4882a593Smuzhiyun }
82*4882a593Smuzhiyun 
test_hash_prealloc(int cpu)83*4882a593Smuzhiyun static void test_hash_prealloc(int cpu)
84*4882a593Smuzhiyun {
85*4882a593Smuzhiyun 	__u64 start_time;
86*4882a593Smuzhiyun 	int i;
87*4882a593Smuzhiyun 
88*4882a593Smuzhiyun 	start_time = time_get_ns();
89*4882a593Smuzhiyun 	for (i = 0; i < max_cnt; i++)
90*4882a593Smuzhiyun 		syscall(__NR_getuid);
91*4882a593Smuzhiyun 	printf("%d:hash_map_perf pre-alloc %lld events per sec\n",
92*4882a593Smuzhiyun 	       cpu, max_cnt * 1000000000ll / (time_get_ns() - start_time));
93*4882a593Smuzhiyun }
94*4882a593Smuzhiyun 
pre_test_lru_hash_lookup(int tasks)95*4882a593Smuzhiyun static int pre_test_lru_hash_lookup(int tasks)
96*4882a593Smuzhiyun {
97*4882a593Smuzhiyun 	int fd = map_fd[lru_hash_lookup_idx];
98*4882a593Smuzhiyun 	uint32_t key;
99*4882a593Smuzhiyun 	long val = 1;
100*4882a593Smuzhiyun 	int ret;
101*4882a593Smuzhiyun 
102*4882a593Smuzhiyun 	if (num_map_entries > lru_hash_lookup_test_entries)
103*4882a593Smuzhiyun 		lru_hash_lookup_test_entries = num_map_entries;
104*4882a593Smuzhiyun 
105*4882a593Smuzhiyun 	/* Populate the lru_hash_map for LRU_HASH_LOOKUP perf test.
106*4882a593Smuzhiyun 	 *
107*4882a593Smuzhiyun 	 * It is fine that the user requests for a map with
108*4882a593Smuzhiyun 	 * num_map_entries < 32 and some of the later lru hash lookup
109*4882a593Smuzhiyun 	 * may return not found.  For LRU map, we are not interested
110*4882a593Smuzhiyun 	 * in such small map performance.
111*4882a593Smuzhiyun 	 */
112*4882a593Smuzhiyun 	for (key = 0; key < lru_hash_lookup_test_entries; key++) {
113*4882a593Smuzhiyun 		ret = bpf_map_update_elem(fd, &key, &val, BPF_NOEXIST);
114*4882a593Smuzhiyun 		if (ret)
115*4882a593Smuzhiyun 			return ret;
116*4882a593Smuzhiyun 	}
117*4882a593Smuzhiyun 
118*4882a593Smuzhiyun 	return 0;
119*4882a593Smuzhiyun }
120*4882a593Smuzhiyun 
do_test_lru(enum test_type test,int cpu)121*4882a593Smuzhiyun static void do_test_lru(enum test_type test, int cpu)
122*4882a593Smuzhiyun {
123*4882a593Smuzhiyun 	static int inner_lru_map_fds[MAX_NR_CPUS];
124*4882a593Smuzhiyun 
125*4882a593Smuzhiyun 	struct sockaddr_in6 in6 = { .sin6_family = AF_INET6 };
126*4882a593Smuzhiyun 	const char *test_name;
127*4882a593Smuzhiyun 	__u64 start_time;
128*4882a593Smuzhiyun 	int i, ret;
129*4882a593Smuzhiyun 
130*4882a593Smuzhiyun 	if (test == INNER_LRU_HASH_PREALLOC && cpu) {
131*4882a593Smuzhiyun 		/* If CPU is not 0, create inner_lru hash map and insert the fd
132*4882a593Smuzhiyun 		 * value into the array_of_lru_hash map. In case of CPU 0,
133*4882a593Smuzhiyun 		 * 'inner_lru_hash_map' was statically inserted on the map init
134*4882a593Smuzhiyun 		 */
135*4882a593Smuzhiyun 		int outer_fd = map_fd[array_of_lru_hashs_idx];
136*4882a593Smuzhiyun 		unsigned int mycpu, mynode;
137*4882a593Smuzhiyun 
138*4882a593Smuzhiyun 		assert(cpu < MAX_NR_CPUS);
139*4882a593Smuzhiyun 
140*4882a593Smuzhiyun 		ret = syscall(__NR_getcpu, &mycpu, &mynode, NULL);
141*4882a593Smuzhiyun 		assert(!ret);
142*4882a593Smuzhiyun 
143*4882a593Smuzhiyun 		inner_lru_map_fds[cpu] =
144*4882a593Smuzhiyun 			bpf_create_map_node(BPF_MAP_TYPE_LRU_HASH,
145*4882a593Smuzhiyun 					    test_map_names[INNER_LRU_HASH_PREALLOC],
146*4882a593Smuzhiyun 					    sizeof(uint32_t),
147*4882a593Smuzhiyun 					    sizeof(long),
148*4882a593Smuzhiyun 					    inner_lru_hash_size, 0,
149*4882a593Smuzhiyun 					    mynode);
150*4882a593Smuzhiyun 		if (inner_lru_map_fds[cpu] == -1) {
151*4882a593Smuzhiyun 			printf("cannot create BPF_MAP_TYPE_LRU_HASH %s(%d)\n",
152*4882a593Smuzhiyun 			       strerror(errno), errno);
153*4882a593Smuzhiyun 			exit(1);
154*4882a593Smuzhiyun 		}
155*4882a593Smuzhiyun 
156*4882a593Smuzhiyun 		ret = bpf_map_update_elem(outer_fd, &cpu,
157*4882a593Smuzhiyun 					  &inner_lru_map_fds[cpu],
158*4882a593Smuzhiyun 					  BPF_ANY);
159*4882a593Smuzhiyun 		if (ret) {
160*4882a593Smuzhiyun 			printf("cannot update ARRAY_OF_LRU_HASHS with key:%u. %s(%d)\n",
161*4882a593Smuzhiyun 			       cpu, strerror(errno), errno);
162*4882a593Smuzhiyun 			exit(1);
163*4882a593Smuzhiyun 		}
164*4882a593Smuzhiyun 	}
165*4882a593Smuzhiyun 
166*4882a593Smuzhiyun 	in6.sin6_addr.s6_addr16[0] = 0xdead;
167*4882a593Smuzhiyun 	in6.sin6_addr.s6_addr16[1] = 0xbeef;
168*4882a593Smuzhiyun 
169*4882a593Smuzhiyun 	if (test == LRU_HASH_PREALLOC) {
170*4882a593Smuzhiyun 		test_name = "lru_hash_map_perf";
171*4882a593Smuzhiyun 		in6.sin6_addr.s6_addr16[2] = 0;
172*4882a593Smuzhiyun 	} else if (test == NOCOMMON_LRU_HASH_PREALLOC) {
173*4882a593Smuzhiyun 		test_name = "nocommon_lru_hash_map_perf";
174*4882a593Smuzhiyun 		in6.sin6_addr.s6_addr16[2] = 1;
175*4882a593Smuzhiyun 	} else if (test == INNER_LRU_HASH_PREALLOC) {
176*4882a593Smuzhiyun 		test_name = "inner_lru_hash_map_perf";
177*4882a593Smuzhiyun 		in6.sin6_addr.s6_addr16[2] = 2;
178*4882a593Smuzhiyun 	} else if (test == LRU_HASH_LOOKUP) {
179*4882a593Smuzhiyun 		test_name = "lru_hash_lookup_perf";
180*4882a593Smuzhiyun 		in6.sin6_addr.s6_addr16[2] = 3;
181*4882a593Smuzhiyun 		in6.sin6_addr.s6_addr32[3] = 0;
182*4882a593Smuzhiyun 	} else {
183*4882a593Smuzhiyun 		assert(0);
184*4882a593Smuzhiyun 	}
185*4882a593Smuzhiyun 
186*4882a593Smuzhiyun 	start_time = time_get_ns();
187*4882a593Smuzhiyun 	for (i = 0; i < max_cnt; i++) {
188*4882a593Smuzhiyun 		ret = connect(-1, (const struct sockaddr *)&in6, sizeof(in6));
189*4882a593Smuzhiyun 		assert(ret == -1 && errno == EBADF);
190*4882a593Smuzhiyun 		if (in6.sin6_addr.s6_addr32[3] <
191*4882a593Smuzhiyun 		    lru_hash_lookup_test_entries - 32)
192*4882a593Smuzhiyun 			in6.sin6_addr.s6_addr32[3] += 32;
193*4882a593Smuzhiyun 		else
194*4882a593Smuzhiyun 			in6.sin6_addr.s6_addr32[3] = 0;
195*4882a593Smuzhiyun 	}
196*4882a593Smuzhiyun 	printf("%d:%s pre-alloc %lld events per sec\n",
197*4882a593Smuzhiyun 	       cpu, test_name,
198*4882a593Smuzhiyun 	       max_cnt * 1000000000ll / (time_get_ns() - start_time));
199*4882a593Smuzhiyun }
200*4882a593Smuzhiyun 
test_lru_hash_prealloc(int cpu)201*4882a593Smuzhiyun static void test_lru_hash_prealloc(int cpu)
202*4882a593Smuzhiyun {
203*4882a593Smuzhiyun 	do_test_lru(LRU_HASH_PREALLOC, cpu);
204*4882a593Smuzhiyun }
205*4882a593Smuzhiyun 
test_nocommon_lru_hash_prealloc(int cpu)206*4882a593Smuzhiyun static void test_nocommon_lru_hash_prealloc(int cpu)
207*4882a593Smuzhiyun {
208*4882a593Smuzhiyun 	do_test_lru(NOCOMMON_LRU_HASH_PREALLOC, cpu);
209*4882a593Smuzhiyun }
210*4882a593Smuzhiyun 
test_inner_lru_hash_prealloc(int cpu)211*4882a593Smuzhiyun static void test_inner_lru_hash_prealloc(int cpu)
212*4882a593Smuzhiyun {
213*4882a593Smuzhiyun 	do_test_lru(INNER_LRU_HASH_PREALLOC, cpu);
214*4882a593Smuzhiyun }
215*4882a593Smuzhiyun 
test_lru_hash_lookup(int cpu)216*4882a593Smuzhiyun static void test_lru_hash_lookup(int cpu)
217*4882a593Smuzhiyun {
218*4882a593Smuzhiyun 	do_test_lru(LRU_HASH_LOOKUP, cpu);
219*4882a593Smuzhiyun }
220*4882a593Smuzhiyun 
test_percpu_hash_prealloc(int cpu)221*4882a593Smuzhiyun static void test_percpu_hash_prealloc(int cpu)
222*4882a593Smuzhiyun {
223*4882a593Smuzhiyun 	__u64 start_time;
224*4882a593Smuzhiyun 	int i;
225*4882a593Smuzhiyun 
226*4882a593Smuzhiyun 	start_time = time_get_ns();
227*4882a593Smuzhiyun 	for (i = 0; i < max_cnt; i++)
228*4882a593Smuzhiyun 		syscall(__NR_geteuid);
229*4882a593Smuzhiyun 	printf("%d:percpu_hash_map_perf pre-alloc %lld events per sec\n",
230*4882a593Smuzhiyun 	       cpu, max_cnt * 1000000000ll / (time_get_ns() - start_time));
231*4882a593Smuzhiyun }
232*4882a593Smuzhiyun 
test_hash_kmalloc(int cpu)233*4882a593Smuzhiyun static void test_hash_kmalloc(int cpu)
234*4882a593Smuzhiyun {
235*4882a593Smuzhiyun 	__u64 start_time;
236*4882a593Smuzhiyun 	int i;
237*4882a593Smuzhiyun 
238*4882a593Smuzhiyun 	start_time = time_get_ns();
239*4882a593Smuzhiyun 	for (i = 0; i < max_cnt; i++)
240*4882a593Smuzhiyun 		syscall(__NR_getgid);
241*4882a593Smuzhiyun 	printf("%d:hash_map_perf kmalloc %lld events per sec\n",
242*4882a593Smuzhiyun 	       cpu, max_cnt * 1000000000ll / (time_get_ns() - start_time));
243*4882a593Smuzhiyun }
244*4882a593Smuzhiyun 
test_percpu_hash_kmalloc(int cpu)245*4882a593Smuzhiyun static void test_percpu_hash_kmalloc(int cpu)
246*4882a593Smuzhiyun {
247*4882a593Smuzhiyun 	__u64 start_time;
248*4882a593Smuzhiyun 	int i;
249*4882a593Smuzhiyun 
250*4882a593Smuzhiyun 	start_time = time_get_ns();
251*4882a593Smuzhiyun 	for (i = 0; i < max_cnt; i++)
252*4882a593Smuzhiyun 		syscall(__NR_getegid);
253*4882a593Smuzhiyun 	printf("%d:percpu_hash_map_perf kmalloc %lld events per sec\n",
254*4882a593Smuzhiyun 	       cpu, max_cnt * 1000000000ll / (time_get_ns() - start_time));
255*4882a593Smuzhiyun }
256*4882a593Smuzhiyun 
test_lpm_kmalloc(int cpu)257*4882a593Smuzhiyun static void test_lpm_kmalloc(int cpu)
258*4882a593Smuzhiyun {
259*4882a593Smuzhiyun 	__u64 start_time;
260*4882a593Smuzhiyun 	int i;
261*4882a593Smuzhiyun 
262*4882a593Smuzhiyun 	start_time = time_get_ns();
263*4882a593Smuzhiyun 	for (i = 0; i < max_cnt; i++)
264*4882a593Smuzhiyun 		syscall(__NR_gettid);
265*4882a593Smuzhiyun 	printf("%d:lpm_perf kmalloc %lld events per sec\n",
266*4882a593Smuzhiyun 	       cpu, max_cnt * 1000000000ll / (time_get_ns() - start_time));
267*4882a593Smuzhiyun }
268*4882a593Smuzhiyun 
test_hash_lookup(int cpu)269*4882a593Smuzhiyun static void test_hash_lookup(int cpu)
270*4882a593Smuzhiyun {
271*4882a593Smuzhiyun 	__u64 start_time;
272*4882a593Smuzhiyun 	int i;
273*4882a593Smuzhiyun 
274*4882a593Smuzhiyun 	start_time = time_get_ns();
275*4882a593Smuzhiyun 	for (i = 0; i < max_cnt; i++)
276*4882a593Smuzhiyun 		syscall(__NR_getpgid, 0);
277*4882a593Smuzhiyun 	printf("%d:hash_lookup %lld lookups per sec\n",
278*4882a593Smuzhiyun 	       cpu, max_cnt * 1000000000ll * 64 / (time_get_ns() - start_time));
279*4882a593Smuzhiyun }
280*4882a593Smuzhiyun 
test_array_lookup(int cpu)281*4882a593Smuzhiyun static void test_array_lookup(int cpu)
282*4882a593Smuzhiyun {
283*4882a593Smuzhiyun 	__u64 start_time;
284*4882a593Smuzhiyun 	int i;
285*4882a593Smuzhiyun 
286*4882a593Smuzhiyun 	start_time = time_get_ns();
287*4882a593Smuzhiyun 	for (i = 0; i < max_cnt; i++)
288*4882a593Smuzhiyun 		syscall(__NR_getppid, 0);
289*4882a593Smuzhiyun 	printf("%d:array_lookup %lld lookups per sec\n",
290*4882a593Smuzhiyun 	       cpu, max_cnt * 1000000000ll * 64 / (time_get_ns() - start_time));
291*4882a593Smuzhiyun }
292*4882a593Smuzhiyun 
293*4882a593Smuzhiyun typedef int (*pre_test_func)(int tasks);
294*4882a593Smuzhiyun const pre_test_func pre_test_funcs[] = {
295*4882a593Smuzhiyun 	[LRU_HASH_LOOKUP] = pre_test_lru_hash_lookup,
296*4882a593Smuzhiyun };
297*4882a593Smuzhiyun 
298*4882a593Smuzhiyun typedef void (*test_func)(int cpu);
299*4882a593Smuzhiyun const test_func test_funcs[] = {
300*4882a593Smuzhiyun 	[HASH_PREALLOC] = test_hash_prealloc,
301*4882a593Smuzhiyun 	[PERCPU_HASH_PREALLOC] = test_percpu_hash_prealloc,
302*4882a593Smuzhiyun 	[HASH_KMALLOC] = test_hash_kmalloc,
303*4882a593Smuzhiyun 	[PERCPU_HASH_KMALLOC] = test_percpu_hash_kmalloc,
304*4882a593Smuzhiyun 	[LRU_HASH_PREALLOC] = test_lru_hash_prealloc,
305*4882a593Smuzhiyun 	[NOCOMMON_LRU_HASH_PREALLOC] = test_nocommon_lru_hash_prealloc,
306*4882a593Smuzhiyun 	[LPM_KMALLOC] = test_lpm_kmalloc,
307*4882a593Smuzhiyun 	[HASH_LOOKUP] = test_hash_lookup,
308*4882a593Smuzhiyun 	[ARRAY_LOOKUP] = test_array_lookup,
309*4882a593Smuzhiyun 	[INNER_LRU_HASH_PREALLOC] = test_inner_lru_hash_prealloc,
310*4882a593Smuzhiyun 	[LRU_HASH_LOOKUP] = test_lru_hash_lookup,
311*4882a593Smuzhiyun };
312*4882a593Smuzhiyun 
pre_test(int tasks)313*4882a593Smuzhiyun static int pre_test(int tasks)
314*4882a593Smuzhiyun {
315*4882a593Smuzhiyun 	int i;
316*4882a593Smuzhiyun 
317*4882a593Smuzhiyun 	for (i = 0; i < NR_TESTS; i++) {
318*4882a593Smuzhiyun 		if (pre_test_funcs[i] && check_test_flags(i)) {
319*4882a593Smuzhiyun 			int ret = pre_test_funcs[i](tasks);
320*4882a593Smuzhiyun 
321*4882a593Smuzhiyun 			if (ret)
322*4882a593Smuzhiyun 				return ret;
323*4882a593Smuzhiyun 		}
324*4882a593Smuzhiyun 	}
325*4882a593Smuzhiyun 
326*4882a593Smuzhiyun 	return 0;
327*4882a593Smuzhiyun }
328*4882a593Smuzhiyun 
loop(int cpu)329*4882a593Smuzhiyun static void loop(int cpu)
330*4882a593Smuzhiyun {
331*4882a593Smuzhiyun 	cpu_set_t cpuset;
332*4882a593Smuzhiyun 	int i;
333*4882a593Smuzhiyun 
334*4882a593Smuzhiyun 	CPU_ZERO(&cpuset);
335*4882a593Smuzhiyun 	CPU_SET(cpu, &cpuset);
336*4882a593Smuzhiyun 	sched_setaffinity(0, sizeof(cpuset), &cpuset);
337*4882a593Smuzhiyun 
338*4882a593Smuzhiyun 	for (i = 0; i < NR_TESTS; i++) {
339*4882a593Smuzhiyun 		if (check_test_flags(i))
340*4882a593Smuzhiyun 			test_funcs[i](cpu);
341*4882a593Smuzhiyun 	}
342*4882a593Smuzhiyun }
343*4882a593Smuzhiyun 
run_perf_test(int tasks)344*4882a593Smuzhiyun static void run_perf_test(int tasks)
345*4882a593Smuzhiyun {
346*4882a593Smuzhiyun 	pid_t pid[tasks];
347*4882a593Smuzhiyun 	int i;
348*4882a593Smuzhiyun 
349*4882a593Smuzhiyun 	assert(!pre_test(tasks));
350*4882a593Smuzhiyun 
351*4882a593Smuzhiyun 	for (i = 0; i < tasks; i++) {
352*4882a593Smuzhiyun 		pid[i] = fork();
353*4882a593Smuzhiyun 		if (pid[i] == 0) {
354*4882a593Smuzhiyun 			loop(i);
355*4882a593Smuzhiyun 			exit(0);
356*4882a593Smuzhiyun 		} else if (pid[i] == -1) {
357*4882a593Smuzhiyun 			printf("couldn't spawn #%d process\n", i);
358*4882a593Smuzhiyun 			exit(1);
359*4882a593Smuzhiyun 		}
360*4882a593Smuzhiyun 	}
361*4882a593Smuzhiyun 	for (i = 0; i < tasks; i++) {
362*4882a593Smuzhiyun 		int status;
363*4882a593Smuzhiyun 
364*4882a593Smuzhiyun 		assert(waitpid(pid[i], &status, 0) == pid[i]);
365*4882a593Smuzhiyun 		assert(status == 0);
366*4882a593Smuzhiyun 	}
367*4882a593Smuzhiyun }
368*4882a593Smuzhiyun 
fill_lpm_trie(void)369*4882a593Smuzhiyun static void fill_lpm_trie(void)
370*4882a593Smuzhiyun {
371*4882a593Smuzhiyun 	struct bpf_lpm_trie_key *key;
372*4882a593Smuzhiyun 	unsigned long value = 0;
373*4882a593Smuzhiyun 	unsigned int i;
374*4882a593Smuzhiyun 	int r;
375*4882a593Smuzhiyun 
376*4882a593Smuzhiyun 	key = alloca(sizeof(*key) + 4);
377*4882a593Smuzhiyun 	key->prefixlen = 32;
378*4882a593Smuzhiyun 
379*4882a593Smuzhiyun 	for (i = 0; i < 512; ++i) {
380*4882a593Smuzhiyun 		key->prefixlen = rand() % 33;
381*4882a593Smuzhiyun 		key->data[0] = rand() & 0xff;
382*4882a593Smuzhiyun 		key->data[1] = rand() & 0xff;
383*4882a593Smuzhiyun 		key->data[2] = rand() & 0xff;
384*4882a593Smuzhiyun 		key->data[3] = rand() & 0xff;
385*4882a593Smuzhiyun 		r = bpf_map_update_elem(map_fd[hash_map_alloc_idx],
386*4882a593Smuzhiyun 					key, &value, 0);
387*4882a593Smuzhiyun 		assert(!r);
388*4882a593Smuzhiyun 	}
389*4882a593Smuzhiyun 
390*4882a593Smuzhiyun 	key->prefixlen = 32;
391*4882a593Smuzhiyun 	key->data[0] = 192;
392*4882a593Smuzhiyun 	key->data[1] = 168;
393*4882a593Smuzhiyun 	key->data[2] = 0;
394*4882a593Smuzhiyun 	key->data[3] = 1;
395*4882a593Smuzhiyun 	value = 128;
396*4882a593Smuzhiyun 
397*4882a593Smuzhiyun 	r = bpf_map_update_elem(map_fd[hash_map_alloc_idx], key, &value, 0);
398*4882a593Smuzhiyun 	assert(!r);
399*4882a593Smuzhiyun }
400*4882a593Smuzhiyun 
fixup_map(struct bpf_object * obj)401*4882a593Smuzhiyun static void fixup_map(struct bpf_object *obj)
402*4882a593Smuzhiyun {
403*4882a593Smuzhiyun 	struct bpf_map *map;
404*4882a593Smuzhiyun 	int i;
405*4882a593Smuzhiyun 
406*4882a593Smuzhiyun 	bpf_object__for_each_map(map, obj) {
407*4882a593Smuzhiyun 		const char *name = bpf_map__name(map);
408*4882a593Smuzhiyun 
409*4882a593Smuzhiyun 		/* Only change the max_entries for the enabled test(s) */
410*4882a593Smuzhiyun 		for (i = 0; i < NR_TESTS; i++) {
411*4882a593Smuzhiyun 			if (!strcmp(test_map_names[i], name) &&
412*4882a593Smuzhiyun 			    (check_test_flags(i))) {
413*4882a593Smuzhiyun 				bpf_map__resize(map, num_map_entries);
414*4882a593Smuzhiyun 				continue;
415*4882a593Smuzhiyun 			}
416*4882a593Smuzhiyun 		}
417*4882a593Smuzhiyun 	}
418*4882a593Smuzhiyun 
419*4882a593Smuzhiyun 	inner_lru_hash_size = num_map_entries;
420*4882a593Smuzhiyun }
421*4882a593Smuzhiyun 
main(int argc,char ** argv)422*4882a593Smuzhiyun int main(int argc, char **argv)
423*4882a593Smuzhiyun {
424*4882a593Smuzhiyun 	struct rlimit r = {RLIM_INFINITY, RLIM_INFINITY};
425*4882a593Smuzhiyun 	int nr_cpus = sysconf(_SC_NPROCESSORS_ONLN);
426*4882a593Smuzhiyun 	struct bpf_link *links[8];
427*4882a593Smuzhiyun 	struct bpf_program *prog;
428*4882a593Smuzhiyun 	struct bpf_object *obj;
429*4882a593Smuzhiyun 	struct bpf_map *map;
430*4882a593Smuzhiyun 	char filename[256];
431*4882a593Smuzhiyun 	int i = 0;
432*4882a593Smuzhiyun 
433*4882a593Smuzhiyun 	if (setrlimit(RLIMIT_MEMLOCK, &r)) {
434*4882a593Smuzhiyun 		perror("setrlimit(RLIMIT_MEMLOCK)");
435*4882a593Smuzhiyun 		return 1;
436*4882a593Smuzhiyun 	}
437*4882a593Smuzhiyun 
438*4882a593Smuzhiyun 	if (argc > 1)
439*4882a593Smuzhiyun 		test_flags = atoi(argv[1]) ? : test_flags;
440*4882a593Smuzhiyun 
441*4882a593Smuzhiyun 	if (argc > 2)
442*4882a593Smuzhiyun 		nr_cpus = atoi(argv[2]) ? : nr_cpus;
443*4882a593Smuzhiyun 
444*4882a593Smuzhiyun 	if (argc > 3)
445*4882a593Smuzhiyun 		num_map_entries = atoi(argv[3]);
446*4882a593Smuzhiyun 
447*4882a593Smuzhiyun 	if (argc > 4)
448*4882a593Smuzhiyun 		max_cnt = atoi(argv[4]);
449*4882a593Smuzhiyun 
450*4882a593Smuzhiyun 	snprintf(filename, sizeof(filename), "%s_kern.o", argv[0]);
451*4882a593Smuzhiyun 	obj = bpf_object__open_file(filename, NULL);
452*4882a593Smuzhiyun 	if (libbpf_get_error(obj)) {
453*4882a593Smuzhiyun 		fprintf(stderr, "ERROR: opening BPF object file failed\n");
454*4882a593Smuzhiyun 		return 0;
455*4882a593Smuzhiyun 	}
456*4882a593Smuzhiyun 
457*4882a593Smuzhiyun 	map = bpf_object__find_map_by_name(obj, "inner_lru_hash_map");
458*4882a593Smuzhiyun 	if (libbpf_get_error(map)) {
459*4882a593Smuzhiyun 		fprintf(stderr, "ERROR: finding a map in obj file failed\n");
460*4882a593Smuzhiyun 		goto cleanup;
461*4882a593Smuzhiyun 	}
462*4882a593Smuzhiyun 
463*4882a593Smuzhiyun 	inner_lru_hash_size = bpf_map__max_entries(map);
464*4882a593Smuzhiyun 	if (!inner_lru_hash_size) {
465*4882a593Smuzhiyun 		fprintf(stderr, "ERROR: failed to get map attribute\n");
466*4882a593Smuzhiyun 		goto cleanup;
467*4882a593Smuzhiyun 	}
468*4882a593Smuzhiyun 
469*4882a593Smuzhiyun 	/* resize BPF map prior to loading */
470*4882a593Smuzhiyun 	if (num_map_entries > 0)
471*4882a593Smuzhiyun 		fixup_map(obj);
472*4882a593Smuzhiyun 
473*4882a593Smuzhiyun 	/* load BPF program */
474*4882a593Smuzhiyun 	if (bpf_object__load(obj)) {
475*4882a593Smuzhiyun 		fprintf(stderr, "ERROR: loading BPF object file failed\n");
476*4882a593Smuzhiyun 		goto cleanup;
477*4882a593Smuzhiyun 	}
478*4882a593Smuzhiyun 
479*4882a593Smuzhiyun 	map_fd[0] = bpf_object__find_map_fd_by_name(obj, "array_of_lru_hashs");
480*4882a593Smuzhiyun 	map_fd[1] = bpf_object__find_map_fd_by_name(obj, "hash_map_alloc");
481*4882a593Smuzhiyun 	map_fd[2] = bpf_object__find_map_fd_by_name(obj, "lru_hash_lookup_map");
482*4882a593Smuzhiyun 	if (map_fd[0] < 0 || map_fd[1] < 0 || map_fd[2] < 0) {
483*4882a593Smuzhiyun 		fprintf(stderr, "ERROR: finding a map in obj file failed\n");
484*4882a593Smuzhiyun 		goto cleanup;
485*4882a593Smuzhiyun 	}
486*4882a593Smuzhiyun 
487*4882a593Smuzhiyun 	bpf_object__for_each_program(prog, obj) {
488*4882a593Smuzhiyun 		links[i] = bpf_program__attach(prog);
489*4882a593Smuzhiyun 		if (libbpf_get_error(links[i])) {
490*4882a593Smuzhiyun 			fprintf(stderr, "ERROR: bpf_program__attach failed\n");
491*4882a593Smuzhiyun 			links[i] = NULL;
492*4882a593Smuzhiyun 			goto cleanup;
493*4882a593Smuzhiyun 		}
494*4882a593Smuzhiyun 		i++;
495*4882a593Smuzhiyun 	}
496*4882a593Smuzhiyun 
497*4882a593Smuzhiyun 	fill_lpm_trie();
498*4882a593Smuzhiyun 
499*4882a593Smuzhiyun 	run_perf_test(nr_cpus);
500*4882a593Smuzhiyun 
501*4882a593Smuzhiyun cleanup:
502*4882a593Smuzhiyun 	for (i--; i >= 0; i--)
503*4882a593Smuzhiyun 		bpf_link__destroy(links[i]);
504*4882a593Smuzhiyun 
505*4882a593Smuzhiyun 	bpf_object__close(obj);
506*4882a593Smuzhiyun 	return 0;
507*4882a593Smuzhiyun }
508