1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun #define _GNU_SOURCE
3*4882a593Smuzhiyun
4*4882a593Smuzhiyun #include <assert.h>
5*4882a593Smuzhiyun #include <fcntl.h>
6*4882a593Smuzhiyun #include <linux/perf_event.h>
7*4882a593Smuzhiyun #include <sched.h>
8*4882a593Smuzhiyun #include <stdio.h>
9*4882a593Smuzhiyun #include <stdlib.h>
10*4882a593Smuzhiyun #include <sys/ioctl.h>
11*4882a593Smuzhiyun #include <sys/resource.h>
12*4882a593Smuzhiyun #include <sys/time.h>
13*4882a593Smuzhiyun #include <sys/types.h>
14*4882a593Smuzhiyun #include <sys/wait.h>
15*4882a593Smuzhiyun #include <unistd.h>
16*4882a593Smuzhiyun
17*4882a593Smuzhiyun #include <bpf/bpf.h>
18*4882a593Smuzhiyun #include <bpf/libbpf.h>
19*4882a593Smuzhiyun #include "perf-sys.h"
20*4882a593Smuzhiyun
21*4882a593Smuzhiyun #define SAMPLE_PERIOD 0x7fffffffffffffffULL
22*4882a593Smuzhiyun
23*4882a593Smuzhiyun /* counters, values, values2 */
24*4882a593Smuzhiyun static int map_fd[3];
25*4882a593Smuzhiyun
check_on_cpu(int cpu,struct perf_event_attr * attr)26*4882a593Smuzhiyun static void check_on_cpu(int cpu, struct perf_event_attr *attr)
27*4882a593Smuzhiyun {
28*4882a593Smuzhiyun struct bpf_perf_event_value value2;
29*4882a593Smuzhiyun int pmu_fd, error = 0;
30*4882a593Smuzhiyun cpu_set_t set;
31*4882a593Smuzhiyun __u64 value;
32*4882a593Smuzhiyun
33*4882a593Smuzhiyun /* Move to target CPU */
34*4882a593Smuzhiyun CPU_ZERO(&set);
35*4882a593Smuzhiyun CPU_SET(cpu, &set);
36*4882a593Smuzhiyun assert(sched_setaffinity(0, sizeof(set), &set) == 0);
37*4882a593Smuzhiyun /* Open perf event and attach to the perf_event_array */
38*4882a593Smuzhiyun pmu_fd = sys_perf_event_open(attr, -1/*pid*/, cpu/*cpu*/, -1/*group_fd*/, 0);
39*4882a593Smuzhiyun if (pmu_fd < 0) {
40*4882a593Smuzhiyun fprintf(stderr, "sys_perf_event_open failed on CPU %d\n", cpu);
41*4882a593Smuzhiyun error = 1;
42*4882a593Smuzhiyun goto on_exit;
43*4882a593Smuzhiyun }
44*4882a593Smuzhiyun assert(bpf_map_update_elem(map_fd[0], &cpu, &pmu_fd, BPF_ANY) == 0);
45*4882a593Smuzhiyun assert(ioctl(pmu_fd, PERF_EVENT_IOC_ENABLE, 0) == 0);
46*4882a593Smuzhiyun /* Trigger the kprobe */
47*4882a593Smuzhiyun bpf_map_get_next_key(map_fd[1], &cpu, NULL);
48*4882a593Smuzhiyun /* Check the value */
49*4882a593Smuzhiyun if (bpf_map_lookup_elem(map_fd[1], &cpu, &value)) {
50*4882a593Smuzhiyun fprintf(stderr, "Value missing for CPU %d\n", cpu);
51*4882a593Smuzhiyun error = 1;
52*4882a593Smuzhiyun goto on_exit;
53*4882a593Smuzhiyun } else {
54*4882a593Smuzhiyun fprintf(stderr, "CPU %d: %llu\n", cpu, value);
55*4882a593Smuzhiyun }
56*4882a593Smuzhiyun /* The above bpf_map_lookup_elem should trigger the second kprobe */
57*4882a593Smuzhiyun if (bpf_map_lookup_elem(map_fd[2], &cpu, &value2)) {
58*4882a593Smuzhiyun fprintf(stderr, "Value2 missing for CPU %d\n", cpu);
59*4882a593Smuzhiyun error = 1;
60*4882a593Smuzhiyun goto on_exit;
61*4882a593Smuzhiyun } else {
62*4882a593Smuzhiyun fprintf(stderr, "CPU %d: counter: %llu, enabled: %llu, running: %llu\n", cpu,
63*4882a593Smuzhiyun value2.counter, value2.enabled, value2.running);
64*4882a593Smuzhiyun }
65*4882a593Smuzhiyun
66*4882a593Smuzhiyun on_exit:
67*4882a593Smuzhiyun assert(bpf_map_delete_elem(map_fd[0], &cpu) == 0 || error);
68*4882a593Smuzhiyun assert(ioctl(pmu_fd, PERF_EVENT_IOC_DISABLE, 0) == 0 || error);
69*4882a593Smuzhiyun assert(close(pmu_fd) == 0 || error);
70*4882a593Smuzhiyun assert(bpf_map_delete_elem(map_fd[1], &cpu) == 0 || error);
71*4882a593Smuzhiyun exit(error);
72*4882a593Smuzhiyun }
73*4882a593Smuzhiyun
test_perf_event_array(struct perf_event_attr * attr,const char * name)74*4882a593Smuzhiyun static void test_perf_event_array(struct perf_event_attr *attr,
75*4882a593Smuzhiyun const char *name)
76*4882a593Smuzhiyun {
77*4882a593Smuzhiyun int i, status, nr_cpus = sysconf(_SC_NPROCESSORS_CONF);
78*4882a593Smuzhiyun pid_t pid[nr_cpus];
79*4882a593Smuzhiyun int err = 0;
80*4882a593Smuzhiyun
81*4882a593Smuzhiyun printf("Test reading %s counters\n", name);
82*4882a593Smuzhiyun
83*4882a593Smuzhiyun for (i = 0; i < nr_cpus; i++) {
84*4882a593Smuzhiyun pid[i] = fork();
85*4882a593Smuzhiyun assert(pid[i] >= 0);
86*4882a593Smuzhiyun if (pid[i] == 0) {
87*4882a593Smuzhiyun check_on_cpu(i, attr);
88*4882a593Smuzhiyun exit(1);
89*4882a593Smuzhiyun }
90*4882a593Smuzhiyun }
91*4882a593Smuzhiyun
92*4882a593Smuzhiyun for (i = 0; i < nr_cpus; i++) {
93*4882a593Smuzhiyun assert(waitpid(pid[i], &status, 0) == pid[i]);
94*4882a593Smuzhiyun err |= status;
95*4882a593Smuzhiyun }
96*4882a593Smuzhiyun
97*4882a593Smuzhiyun if (err)
98*4882a593Smuzhiyun printf("Test: %s FAILED\n", name);
99*4882a593Smuzhiyun }
100*4882a593Smuzhiyun
test_bpf_perf_event(void)101*4882a593Smuzhiyun static void test_bpf_perf_event(void)
102*4882a593Smuzhiyun {
103*4882a593Smuzhiyun struct perf_event_attr attr_cycles = {
104*4882a593Smuzhiyun .freq = 0,
105*4882a593Smuzhiyun .sample_period = SAMPLE_PERIOD,
106*4882a593Smuzhiyun .inherit = 0,
107*4882a593Smuzhiyun .type = PERF_TYPE_HARDWARE,
108*4882a593Smuzhiyun .read_format = 0,
109*4882a593Smuzhiyun .sample_type = 0,
110*4882a593Smuzhiyun .config = PERF_COUNT_HW_CPU_CYCLES,
111*4882a593Smuzhiyun };
112*4882a593Smuzhiyun struct perf_event_attr attr_clock = {
113*4882a593Smuzhiyun .freq = 0,
114*4882a593Smuzhiyun .sample_period = SAMPLE_PERIOD,
115*4882a593Smuzhiyun .inherit = 0,
116*4882a593Smuzhiyun .type = PERF_TYPE_SOFTWARE,
117*4882a593Smuzhiyun .read_format = 0,
118*4882a593Smuzhiyun .sample_type = 0,
119*4882a593Smuzhiyun .config = PERF_COUNT_SW_CPU_CLOCK,
120*4882a593Smuzhiyun };
121*4882a593Smuzhiyun struct perf_event_attr attr_raw = {
122*4882a593Smuzhiyun .freq = 0,
123*4882a593Smuzhiyun .sample_period = SAMPLE_PERIOD,
124*4882a593Smuzhiyun .inherit = 0,
125*4882a593Smuzhiyun .type = PERF_TYPE_RAW,
126*4882a593Smuzhiyun .read_format = 0,
127*4882a593Smuzhiyun .sample_type = 0,
128*4882a593Smuzhiyun /* Intel Instruction Retired */
129*4882a593Smuzhiyun .config = 0xc0,
130*4882a593Smuzhiyun };
131*4882a593Smuzhiyun struct perf_event_attr attr_l1d_load = {
132*4882a593Smuzhiyun .freq = 0,
133*4882a593Smuzhiyun .sample_period = SAMPLE_PERIOD,
134*4882a593Smuzhiyun .inherit = 0,
135*4882a593Smuzhiyun .type = PERF_TYPE_HW_CACHE,
136*4882a593Smuzhiyun .read_format = 0,
137*4882a593Smuzhiyun .sample_type = 0,
138*4882a593Smuzhiyun .config =
139*4882a593Smuzhiyun PERF_COUNT_HW_CACHE_L1D |
140*4882a593Smuzhiyun (PERF_COUNT_HW_CACHE_OP_READ << 8) |
141*4882a593Smuzhiyun (PERF_COUNT_HW_CACHE_RESULT_ACCESS << 16),
142*4882a593Smuzhiyun };
143*4882a593Smuzhiyun struct perf_event_attr attr_llc_miss = {
144*4882a593Smuzhiyun .freq = 0,
145*4882a593Smuzhiyun .sample_period = SAMPLE_PERIOD,
146*4882a593Smuzhiyun .inherit = 0,
147*4882a593Smuzhiyun .type = PERF_TYPE_HW_CACHE,
148*4882a593Smuzhiyun .read_format = 0,
149*4882a593Smuzhiyun .sample_type = 0,
150*4882a593Smuzhiyun .config =
151*4882a593Smuzhiyun PERF_COUNT_HW_CACHE_LL |
152*4882a593Smuzhiyun (PERF_COUNT_HW_CACHE_OP_READ << 8) |
153*4882a593Smuzhiyun (PERF_COUNT_HW_CACHE_RESULT_MISS << 16),
154*4882a593Smuzhiyun };
155*4882a593Smuzhiyun struct perf_event_attr attr_msr_tsc = {
156*4882a593Smuzhiyun .freq = 0,
157*4882a593Smuzhiyun .sample_period = 0,
158*4882a593Smuzhiyun .inherit = 0,
159*4882a593Smuzhiyun /* From /sys/bus/event_source/devices/msr/ */
160*4882a593Smuzhiyun .type = 7,
161*4882a593Smuzhiyun .read_format = 0,
162*4882a593Smuzhiyun .sample_type = 0,
163*4882a593Smuzhiyun .config = 0,
164*4882a593Smuzhiyun };
165*4882a593Smuzhiyun
166*4882a593Smuzhiyun test_perf_event_array(&attr_cycles, "HARDWARE-cycles");
167*4882a593Smuzhiyun test_perf_event_array(&attr_clock, "SOFTWARE-clock");
168*4882a593Smuzhiyun test_perf_event_array(&attr_raw, "RAW-instruction-retired");
169*4882a593Smuzhiyun test_perf_event_array(&attr_l1d_load, "HW_CACHE-L1D-load");
170*4882a593Smuzhiyun
171*4882a593Smuzhiyun /* below tests may fail in qemu */
172*4882a593Smuzhiyun test_perf_event_array(&attr_llc_miss, "HW_CACHE-LLC-miss");
173*4882a593Smuzhiyun test_perf_event_array(&attr_msr_tsc, "Dynamic-msr-tsc");
174*4882a593Smuzhiyun }
175*4882a593Smuzhiyun
main(int argc,char ** argv)176*4882a593Smuzhiyun int main(int argc, char **argv)
177*4882a593Smuzhiyun {
178*4882a593Smuzhiyun struct rlimit r = {RLIM_INFINITY, RLIM_INFINITY};
179*4882a593Smuzhiyun struct bpf_link *links[2];
180*4882a593Smuzhiyun struct bpf_program *prog;
181*4882a593Smuzhiyun struct bpf_object *obj;
182*4882a593Smuzhiyun char filename[256];
183*4882a593Smuzhiyun int i = 0;
184*4882a593Smuzhiyun
185*4882a593Smuzhiyun setrlimit(RLIMIT_MEMLOCK, &r);
186*4882a593Smuzhiyun
187*4882a593Smuzhiyun snprintf(filename, sizeof(filename), "%s_kern.o", argv[0]);
188*4882a593Smuzhiyun obj = bpf_object__open_file(filename, NULL);
189*4882a593Smuzhiyun if (libbpf_get_error(obj)) {
190*4882a593Smuzhiyun fprintf(stderr, "ERROR: opening BPF object file failed\n");
191*4882a593Smuzhiyun return 0;
192*4882a593Smuzhiyun }
193*4882a593Smuzhiyun
194*4882a593Smuzhiyun /* load BPF program */
195*4882a593Smuzhiyun if (bpf_object__load(obj)) {
196*4882a593Smuzhiyun fprintf(stderr, "ERROR: loading BPF object file failed\n");
197*4882a593Smuzhiyun goto cleanup;
198*4882a593Smuzhiyun }
199*4882a593Smuzhiyun
200*4882a593Smuzhiyun map_fd[0] = bpf_object__find_map_fd_by_name(obj, "counters");
201*4882a593Smuzhiyun map_fd[1] = bpf_object__find_map_fd_by_name(obj, "values");
202*4882a593Smuzhiyun map_fd[2] = bpf_object__find_map_fd_by_name(obj, "values2");
203*4882a593Smuzhiyun if (map_fd[0] < 0 || map_fd[1] < 0 || map_fd[2] < 0) {
204*4882a593Smuzhiyun fprintf(stderr, "ERROR: finding a map in obj file failed\n");
205*4882a593Smuzhiyun goto cleanup;
206*4882a593Smuzhiyun }
207*4882a593Smuzhiyun
208*4882a593Smuzhiyun bpf_object__for_each_program(prog, obj) {
209*4882a593Smuzhiyun links[i] = bpf_program__attach(prog);
210*4882a593Smuzhiyun if (libbpf_get_error(links[i])) {
211*4882a593Smuzhiyun fprintf(stderr, "ERROR: bpf_program__attach failed\n");
212*4882a593Smuzhiyun links[i] = NULL;
213*4882a593Smuzhiyun goto cleanup;
214*4882a593Smuzhiyun }
215*4882a593Smuzhiyun i++;
216*4882a593Smuzhiyun }
217*4882a593Smuzhiyun
218*4882a593Smuzhiyun test_bpf_perf_event();
219*4882a593Smuzhiyun
220*4882a593Smuzhiyun cleanup:
221*4882a593Smuzhiyun for (i--; i >= 0; i--)
222*4882a593Smuzhiyun bpf_link__destroy(links[i]);
223*4882a593Smuzhiyun
224*4882a593Smuzhiyun bpf_object__close(obj);
225*4882a593Smuzhiyun return 0;
226*4882a593Smuzhiyun }
227