1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun #include <errno.h>
3*4882a593Smuzhiyun #include <linux/kernel.h>
4*4882a593Smuzhiyun #include <linux/types.h>
5*4882a593Smuzhiyun #include <inttypes.h>
6*4882a593Smuzhiyun #include <stdlib.h>
7*4882a593Smuzhiyun #include <unistd.h>
8*4882a593Smuzhiyun #include <stdio.h>
9*4882a593Smuzhiyun #include <string.h>
10*4882a593Smuzhiyun #include <sys/param.h>
11*4882a593Smuzhiyun #include <perf/cpumap.h>
12*4882a593Smuzhiyun #include <perf/evlist.h>
13*4882a593Smuzhiyun #include <perf/mmap.h>
14*4882a593Smuzhiyun
15*4882a593Smuzhiyun #include "debug.h"
16*4882a593Smuzhiyun #include "dso.h"
17*4882a593Smuzhiyun #include "env.h"
18*4882a593Smuzhiyun #include "parse-events.h"
19*4882a593Smuzhiyun #include "trace-event.h"
20*4882a593Smuzhiyun #include "evlist.h"
21*4882a593Smuzhiyun #include "evsel.h"
22*4882a593Smuzhiyun #include "thread_map.h"
23*4882a593Smuzhiyun #include "machine.h"
24*4882a593Smuzhiyun #include "map.h"
25*4882a593Smuzhiyun #include "symbol.h"
26*4882a593Smuzhiyun #include "event.h"
27*4882a593Smuzhiyun #include "record.h"
28*4882a593Smuzhiyun #include "util/mmap.h"
29*4882a593Smuzhiyun #include "util/synthetic-events.h"
30*4882a593Smuzhiyun #include "thread.h"
31*4882a593Smuzhiyun
32*4882a593Smuzhiyun #include "tests.h"
33*4882a593Smuzhiyun
34*4882a593Smuzhiyun #include <linux/ctype.h>
35*4882a593Smuzhiyun
36*4882a593Smuzhiyun #define BUFSZ 1024
37*4882a593Smuzhiyun #define READLEN 128
38*4882a593Smuzhiyun
39*4882a593Smuzhiyun struct state {
40*4882a593Smuzhiyun u64 done[1024];
41*4882a593Smuzhiyun size_t done_cnt;
42*4882a593Smuzhiyun };
43*4882a593Smuzhiyun
hex(char c)44*4882a593Smuzhiyun static unsigned int hex(char c)
45*4882a593Smuzhiyun {
46*4882a593Smuzhiyun if (c >= '0' && c <= '9')
47*4882a593Smuzhiyun return c - '0';
48*4882a593Smuzhiyun if (c >= 'a' && c <= 'f')
49*4882a593Smuzhiyun return c - 'a' + 10;
50*4882a593Smuzhiyun return c - 'A' + 10;
51*4882a593Smuzhiyun }
52*4882a593Smuzhiyun
read_objdump_chunk(const char ** line,unsigned char ** buf,size_t * buf_len)53*4882a593Smuzhiyun static size_t read_objdump_chunk(const char **line, unsigned char **buf,
54*4882a593Smuzhiyun size_t *buf_len)
55*4882a593Smuzhiyun {
56*4882a593Smuzhiyun size_t bytes_read = 0;
57*4882a593Smuzhiyun unsigned char *chunk_start = *buf;
58*4882a593Smuzhiyun
59*4882a593Smuzhiyun /* Read bytes */
60*4882a593Smuzhiyun while (*buf_len > 0) {
61*4882a593Smuzhiyun char c1, c2;
62*4882a593Smuzhiyun
63*4882a593Smuzhiyun /* Get 2 hex digits */
64*4882a593Smuzhiyun c1 = *(*line)++;
65*4882a593Smuzhiyun if (!isxdigit(c1))
66*4882a593Smuzhiyun break;
67*4882a593Smuzhiyun c2 = *(*line)++;
68*4882a593Smuzhiyun if (!isxdigit(c2))
69*4882a593Smuzhiyun break;
70*4882a593Smuzhiyun
71*4882a593Smuzhiyun /* Store byte and advance buf */
72*4882a593Smuzhiyun **buf = (hex(c1) << 4) | hex(c2);
73*4882a593Smuzhiyun (*buf)++;
74*4882a593Smuzhiyun (*buf_len)--;
75*4882a593Smuzhiyun bytes_read++;
76*4882a593Smuzhiyun
77*4882a593Smuzhiyun /* End of chunk? */
78*4882a593Smuzhiyun if (isspace(**line))
79*4882a593Smuzhiyun break;
80*4882a593Smuzhiyun }
81*4882a593Smuzhiyun
82*4882a593Smuzhiyun /*
83*4882a593Smuzhiyun * objdump will display raw insn as LE if code endian
84*4882a593Smuzhiyun * is LE and bytes_per_chunk > 1. In that case reverse
85*4882a593Smuzhiyun * the chunk we just read.
86*4882a593Smuzhiyun *
87*4882a593Smuzhiyun * see disassemble_bytes() at binutils/objdump.c for details
88*4882a593Smuzhiyun * how objdump chooses display endian)
89*4882a593Smuzhiyun */
90*4882a593Smuzhiyun if (bytes_read > 1 && !bigendian()) {
91*4882a593Smuzhiyun unsigned char *chunk_end = chunk_start + bytes_read - 1;
92*4882a593Smuzhiyun unsigned char tmp;
93*4882a593Smuzhiyun
94*4882a593Smuzhiyun while (chunk_start < chunk_end) {
95*4882a593Smuzhiyun tmp = *chunk_start;
96*4882a593Smuzhiyun *chunk_start = *chunk_end;
97*4882a593Smuzhiyun *chunk_end = tmp;
98*4882a593Smuzhiyun chunk_start++;
99*4882a593Smuzhiyun chunk_end--;
100*4882a593Smuzhiyun }
101*4882a593Smuzhiyun }
102*4882a593Smuzhiyun
103*4882a593Smuzhiyun return bytes_read;
104*4882a593Smuzhiyun }
105*4882a593Smuzhiyun
read_objdump_line(const char * line,unsigned char * buf,size_t buf_len)106*4882a593Smuzhiyun static size_t read_objdump_line(const char *line, unsigned char *buf,
107*4882a593Smuzhiyun size_t buf_len)
108*4882a593Smuzhiyun {
109*4882a593Smuzhiyun const char *p;
110*4882a593Smuzhiyun size_t ret, bytes_read = 0;
111*4882a593Smuzhiyun
112*4882a593Smuzhiyun /* Skip to a colon */
113*4882a593Smuzhiyun p = strchr(line, ':');
114*4882a593Smuzhiyun if (!p)
115*4882a593Smuzhiyun return 0;
116*4882a593Smuzhiyun p++;
117*4882a593Smuzhiyun
118*4882a593Smuzhiyun /* Skip initial spaces */
119*4882a593Smuzhiyun while (*p) {
120*4882a593Smuzhiyun if (!isspace(*p))
121*4882a593Smuzhiyun break;
122*4882a593Smuzhiyun p++;
123*4882a593Smuzhiyun }
124*4882a593Smuzhiyun
125*4882a593Smuzhiyun do {
126*4882a593Smuzhiyun ret = read_objdump_chunk(&p, &buf, &buf_len);
127*4882a593Smuzhiyun bytes_read += ret;
128*4882a593Smuzhiyun p++;
129*4882a593Smuzhiyun } while (ret > 0);
130*4882a593Smuzhiyun
131*4882a593Smuzhiyun /* return number of successfully read bytes */
132*4882a593Smuzhiyun return bytes_read;
133*4882a593Smuzhiyun }
134*4882a593Smuzhiyun
read_objdump_output(FILE * f,void * buf,size_t * len,u64 start_addr)135*4882a593Smuzhiyun static int read_objdump_output(FILE *f, void *buf, size_t *len, u64 start_addr)
136*4882a593Smuzhiyun {
137*4882a593Smuzhiyun char *line = NULL;
138*4882a593Smuzhiyun size_t line_len, off_last = 0;
139*4882a593Smuzhiyun ssize_t ret;
140*4882a593Smuzhiyun int err = 0;
141*4882a593Smuzhiyun u64 addr, last_addr = start_addr;
142*4882a593Smuzhiyun
143*4882a593Smuzhiyun while (off_last < *len) {
144*4882a593Smuzhiyun size_t off, read_bytes, written_bytes;
145*4882a593Smuzhiyun unsigned char tmp[BUFSZ];
146*4882a593Smuzhiyun
147*4882a593Smuzhiyun ret = getline(&line, &line_len, f);
148*4882a593Smuzhiyun if (feof(f))
149*4882a593Smuzhiyun break;
150*4882a593Smuzhiyun if (ret < 0) {
151*4882a593Smuzhiyun pr_debug("getline failed\n");
152*4882a593Smuzhiyun err = -1;
153*4882a593Smuzhiyun break;
154*4882a593Smuzhiyun }
155*4882a593Smuzhiyun
156*4882a593Smuzhiyun /* read objdump data into temporary buffer */
157*4882a593Smuzhiyun read_bytes = read_objdump_line(line, tmp, sizeof(tmp));
158*4882a593Smuzhiyun if (!read_bytes)
159*4882a593Smuzhiyun continue;
160*4882a593Smuzhiyun
161*4882a593Smuzhiyun if (sscanf(line, "%"PRIx64, &addr) != 1)
162*4882a593Smuzhiyun continue;
163*4882a593Smuzhiyun if (addr < last_addr) {
164*4882a593Smuzhiyun pr_debug("addr going backwards, read beyond section?\n");
165*4882a593Smuzhiyun break;
166*4882a593Smuzhiyun }
167*4882a593Smuzhiyun last_addr = addr;
168*4882a593Smuzhiyun
169*4882a593Smuzhiyun /* copy it from temporary buffer to 'buf' according
170*4882a593Smuzhiyun * to address on current objdump line */
171*4882a593Smuzhiyun off = addr - start_addr;
172*4882a593Smuzhiyun if (off >= *len)
173*4882a593Smuzhiyun break;
174*4882a593Smuzhiyun written_bytes = MIN(read_bytes, *len - off);
175*4882a593Smuzhiyun memcpy(buf + off, tmp, written_bytes);
176*4882a593Smuzhiyun off_last = off + written_bytes;
177*4882a593Smuzhiyun }
178*4882a593Smuzhiyun
179*4882a593Smuzhiyun /* len returns number of bytes that could not be read */
180*4882a593Smuzhiyun *len -= off_last;
181*4882a593Smuzhiyun
182*4882a593Smuzhiyun free(line);
183*4882a593Smuzhiyun
184*4882a593Smuzhiyun return err;
185*4882a593Smuzhiyun }
186*4882a593Smuzhiyun
read_via_objdump(const char * filename,u64 addr,void * buf,size_t len)187*4882a593Smuzhiyun static int read_via_objdump(const char *filename, u64 addr, void *buf,
188*4882a593Smuzhiyun size_t len)
189*4882a593Smuzhiyun {
190*4882a593Smuzhiyun char cmd[PATH_MAX * 2];
191*4882a593Smuzhiyun const char *fmt;
192*4882a593Smuzhiyun FILE *f;
193*4882a593Smuzhiyun int ret;
194*4882a593Smuzhiyun
195*4882a593Smuzhiyun fmt = "%s -z -d --start-address=0x%"PRIx64" --stop-address=0x%"PRIx64" %s";
196*4882a593Smuzhiyun ret = snprintf(cmd, sizeof(cmd), fmt, "objdump", addr, addr + len,
197*4882a593Smuzhiyun filename);
198*4882a593Smuzhiyun if (ret <= 0 || (size_t)ret >= sizeof(cmd))
199*4882a593Smuzhiyun return -1;
200*4882a593Smuzhiyun
201*4882a593Smuzhiyun pr_debug("Objdump command is: %s\n", cmd);
202*4882a593Smuzhiyun
203*4882a593Smuzhiyun /* Ignore objdump errors */
204*4882a593Smuzhiyun strcat(cmd, " 2>/dev/null");
205*4882a593Smuzhiyun
206*4882a593Smuzhiyun f = popen(cmd, "r");
207*4882a593Smuzhiyun if (!f) {
208*4882a593Smuzhiyun pr_debug("popen failed\n");
209*4882a593Smuzhiyun return -1;
210*4882a593Smuzhiyun }
211*4882a593Smuzhiyun
212*4882a593Smuzhiyun ret = read_objdump_output(f, buf, &len, addr);
213*4882a593Smuzhiyun if (len) {
214*4882a593Smuzhiyun pr_debug("objdump read too few bytes: %zd\n", len);
215*4882a593Smuzhiyun if (!ret)
216*4882a593Smuzhiyun ret = len;
217*4882a593Smuzhiyun }
218*4882a593Smuzhiyun
219*4882a593Smuzhiyun pclose(f);
220*4882a593Smuzhiyun
221*4882a593Smuzhiyun return ret;
222*4882a593Smuzhiyun }
223*4882a593Smuzhiyun
dump_buf(unsigned char * buf,size_t len)224*4882a593Smuzhiyun static void dump_buf(unsigned char *buf, size_t len)
225*4882a593Smuzhiyun {
226*4882a593Smuzhiyun size_t i;
227*4882a593Smuzhiyun
228*4882a593Smuzhiyun for (i = 0; i < len; i++) {
229*4882a593Smuzhiyun pr_debug("0x%02x ", buf[i]);
230*4882a593Smuzhiyun if (i % 16 == 15)
231*4882a593Smuzhiyun pr_debug("\n");
232*4882a593Smuzhiyun }
233*4882a593Smuzhiyun pr_debug("\n");
234*4882a593Smuzhiyun }
235*4882a593Smuzhiyun
read_object_code(u64 addr,size_t len,u8 cpumode,struct thread * thread,struct state * state)236*4882a593Smuzhiyun static int read_object_code(u64 addr, size_t len, u8 cpumode,
237*4882a593Smuzhiyun struct thread *thread, struct state *state)
238*4882a593Smuzhiyun {
239*4882a593Smuzhiyun struct addr_location al;
240*4882a593Smuzhiyun unsigned char buf1[BUFSZ];
241*4882a593Smuzhiyun unsigned char buf2[BUFSZ];
242*4882a593Smuzhiyun size_t ret_len;
243*4882a593Smuzhiyun u64 objdump_addr;
244*4882a593Smuzhiyun const char *objdump_name;
245*4882a593Smuzhiyun char decomp_name[KMOD_DECOMP_LEN];
246*4882a593Smuzhiyun bool decomp = false;
247*4882a593Smuzhiyun int ret;
248*4882a593Smuzhiyun
249*4882a593Smuzhiyun pr_debug("Reading object code for memory address: %#"PRIx64"\n", addr);
250*4882a593Smuzhiyun
251*4882a593Smuzhiyun if (!thread__find_map(thread, cpumode, addr, &al) || !al.map->dso) {
252*4882a593Smuzhiyun if (cpumode == PERF_RECORD_MISC_HYPERVISOR) {
253*4882a593Smuzhiyun pr_debug("Hypervisor address can not be resolved - skipping\n");
254*4882a593Smuzhiyun return 0;
255*4882a593Smuzhiyun }
256*4882a593Smuzhiyun
257*4882a593Smuzhiyun pr_debug("thread__find_map failed\n");
258*4882a593Smuzhiyun return -1;
259*4882a593Smuzhiyun }
260*4882a593Smuzhiyun
261*4882a593Smuzhiyun pr_debug("File is: %s\n", al.map->dso->long_name);
262*4882a593Smuzhiyun
263*4882a593Smuzhiyun if (al.map->dso->symtab_type == DSO_BINARY_TYPE__KALLSYMS &&
264*4882a593Smuzhiyun !dso__is_kcore(al.map->dso)) {
265*4882a593Smuzhiyun pr_debug("Unexpected kernel address - skipping\n");
266*4882a593Smuzhiyun return 0;
267*4882a593Smuzhiyun }
268*4882a593Smuzhiyun
269*4882a593Smuzhiyun pr_debug("On file address is: %#"PRIx64"\n", al.addr);
270*4882a593Smuzhiyun
271*4882a593Smuzhiyun if (len > BUFSZ)
272*4882a593Smuzhiyun len = BUFSZ;
273*4882a593Smuzhiyun
274*4882a593Smuzhiyun /* Do not go off the map */
275*4882a593Smuzhiyun if (addr + len > al.map->end)
276*4882a593Smuzhiyun len = al.map->end - addr;
277*4882a593Smuzhiyun
278*4882a593Smuzhiyun /* Read the object code using perf */
279*4882a593Smuzhiyun ret_len = dso__data_read_offset(al.map->dso, thread->maps->machine,
280*4882a593Smuzhiyun al.addr, buf1, len);
281*4882a593Smuzhiyun if (ret_len != len) {
282*4882a593Smuzhiyun pr_debug("dso__data_read_offset failed\n");
283*4882a593Smuzhiyun return -1;
284*4882a593Smuzhiyun }
285*4882a593Smuzhiyun
286*4882a593Smuzhiyun /*
287*4882a593Smuzhiyun * Converting addresses for use by objdump requires more information.
288*4882a593Smuzhiyun * map__load() does that. See map__rip_2objdump() for details.
289*4882a593Smuzhiyun */
290*4882a593Smuzhiyun if (map__load(al.map))
291*4882a593Smuzhiyun return -1;
292*4882a593Smuzhiyun
293*4882a593Smuzhiyun /* objdump struggles with kcore - try each map only once */
294*4882a593Smuzhiyun if (dso__is_kcore(al.map->dso)) {
295*4882a593Smuzhiyun size_t d;
296*4882a593Smuzhiyun
297*4882a593Smuzhiyun for (d = 0; d < state->done_cnt; d++) {
298*4882a593Smuzhiyun if (state->done[d] == al.map->start) {
299*4882a593Smuzhiyun pr_debug("kcore map tested already");
300*4882a593Smuzhiyun pr_debug(" - skipping\n");
301*4882a593Smuzhiyun return 0;
302*4882a593Smuzhiyun }
303*4882a593Smuzhiyun }
304*4882a593Smuzhiyun if (state->done_cnt >= ARRAY_SIZE(state->done)) {
305*4882a593Smuzhiyun pr_debug("Too many kcore maps - skipping\n");
306*4882a593Smuzhiyun return 0;
307*4882a593Smuzhiyun }
308*4882a593Smuzhiyun state->done[state->done_cnt++] = al.map->start;
309*4882a593Smuzhiyun }
310*4882a593Smuzhiyun
311*4882a593Smuzhiyun objdump_name = al.map->dso->long_name;
312*4882a593Smuzhiyun if (dso__needs_decompress(al.map->dso)) {
313*4882a593Smuzhiyun if (dso__decompress_kmodule_path(al.map->dso, objdump_name,
314*4882a593Smuzhiyun decomp_name,
315*4882a593Smuzhiyun sizeof(decomp_name)) < 0) {
316*4882a593Smuzhiyun pr_debug("decompression failed\n");
317*4882a593Smuzhiyun return -1;
318*4882a593Smuzhiyun }
319*4882a593Smuzhiyun
320*4882a593Smuzhiyun decomp = true;
321*4882a593Smuzhiyun objdump_name = decomp_name;
322*4882a593Smuzhiyun }
323*4882a593Smuzhiyun
324*4882a593Smuzhiyun /* Read the object code using objdump */
325*4882a593Smuzhiyun objdump_addr = map__rip_2objdump(al.map, al.addr);
326*4882a593Smuzhiyun ret = read_via_objdump(objdump_name, objdump_addr, buf2, len);
327*4882a593Smuzhiyun
328*4882a593Smuzhiyun if (decomp)
329*4882a593Smuzhiyun unlink(objdump_name);
330*4882a593Smuzhiyun
331*4882a593Smuzhiyun if (ret > 0) {
332*4882a593Smuzhiyun /*
333*4882a593Smuzhiyun * The kernel maps are inaccurate - assume objdump is right in
334*4882a593Smuzhiyun * that case.
335*4882a593Smuzhiyun */
336*4882a593Smuzhiyun if (cpumode == PERF_RECORD_MISC_KERNEL ||
337*4882a593Smuzhiyun cpumode == PERF_RECORD_MISC_GUEST_KERNEL) {
338*4882a593Smuzhiyun len -= ret;
339*4882a593Smuzhiyun if (len) {
340*4882a593Smuzhiyun pr_debug("Reducing len to %zu\n", len);
341*4882a593Smuzhiyun } else if (dso__is_kcore(al.map->dso)) {
342*4882a593Smuzhiyun /*
343*4882a593Smuzhiyun * objdump cannot handle very large segments
344*4882a593Smuzhiyun * that may be found in kcore.
345*4882a593Smuzhiyun */
346*4882a593Smuzhiyun pr_debug("objdump failed for kcore");
347*4882a593Smuzhiyun pr_debug(" - skipping\n");
348*4882a593Smuzhiyun return 0;
349*4882a593Smuzhiyun } else {
350*4882a593Smuzhiyun return -1;
351*4882a593Smuzhiyun }
352*4882a593Smuzhiyun }
353*4882a593Smuzhiyun }
354*4882a593Smuzhiyun if (ret < 0) {
355*4882a593Smuzhiyun pr_debug("read_via_objdump failed\n");
356*4882a593Smuzhiyun return -1;
357*4882a593Smuzhiyun }
358*4882a593Smuzhiyun
359*4882a593Smuzhiyun /* The results should be identical */
360*4882a593Smuzhiyun if (memcmp(buf1, buf2, len)) {
361*4882a593Smuzhiyun pr_debug("Bytes read differ from those read by objdump\n");
362*4882a593Smuzhiyun pr_debug("buf1 (dso):\n");
363*4882a593Smuzhiyun dump_buf(buf1, len);
364*4882a593Smuzhiyun pr_debug("buf2 (objdump):\n");
365*4882a593Smuzhiyun dump_buf(buf2, len);
366*4882a593Smuzhiyun return -1;
367*4882a593Smuzhiyun }
368*4882a593Smuzhiyun pr_debug("Bytes read match those read by objdump\n");
369*4882a593Smuzhiyun
370*4882a593Smuzhiyun return 0;
371*4882a593Smuzhiyun }
372*4882a593Smuzhiyun
process_sample_event(struct machine * machine,struct evlist * evlist,union perf_event * event,struct state * state)373*4882a593Smuzhiyun static int process_sample_event(struct machine *machine,
374*4882a593Smuzhiyun struct evlist *evlist,
375*4882a593Smuzhiyun union perf_event *event, struct state *state)
376*4882a593Smuzhiyun {
377*4882a593Smuzhiyun struct perf_sample sample;
378*4882a593Smuzhiyun struct thread *thread;
379*4882a593Smuzhiyun int ret;
380*4882a593Smuzhiyun
381*4882a593Smuzhiyun if (perf_evlist__parse_sample(evlist, event, &sample)) {
382*4882a593Smuzhiyun pr_debug("perf_evlist__parse_sample failed\n");
383*4882a593Smuzhiyun return -1;
384*4882a593Smuzhiyun }
385*4882a593Smuzhiyun
386*4882a593Smuzhiyun thread = machine__findnew_thread(machine, sample.pid, sample.tid);
387*4882a593Smuzhiyun if (!thread) {
388*4882a593Smuzhiyun pr_debug("machine__findnew_thread failed\n");
389*4882a593Smuzhiyun return -1;
390*4882a593Smuzhiyun }
391*4882a593Smuzhiyun
392*4882a593Smuzhiyun ret = read_object_code(sample.ip, READLEN, sample.cpumode, thread, state);
393*4882a593Smuzhiyun thread__put(thread);
394*4882a593Smuzhiyun return ret;
395*4882a593Smuzhiyun }
396*4882a593Smuzhiyun
process_event(struct machine * machine,struct evlist * evlist,union perf_event * event,struct state * state)397*4882a593Smuzhiyun static int process_event(struct machine *machine, struct evlist *evlist,
398*4882a593Smuzhiyun union perf_event *event, struct state *state)
399*4882a593Smuzhiyun {
400*4882a593Smuzhiyun if (event->header.type == PERF_RECORD_SAMPLE)
401*4882a593Smuzhiyun return process_sample_event(machine, evlist, event, state);
402*4882a593Smuzhiyun
403*4882a593Smuzhiyun if (event->header.type == PERF_RECORD_THROTTLE ||
404*4882a593Smuzhiyun event->header.type == PERF_RECORD_UNTHROTTLE)
405*4882a593Smuzhiyun return 0;
406*4882a593Smuzhiyun
407*4882a593Smuzhiyun if (event->header.type < PERF_RECORD_MAX) {
408*4882a593Smuzhiyun int ret;
409*4882a593Smuzhiyun
410*4882a593Smuzhiyun ret = machine__process_event(machine, event, NULL);
411*4882a593Smuzhiyun if (ret < 0)
412*4882a593Smuzhiyun pr_debug("machine__process_event failed, event type %u\n",
413*4882a593Smuzhiyun event->header.type);
414*4882a593Smuzhiyun return ret;
415*4882a593Smuzhiyun }
416*4882a593Smuzhiyun
417*4882a593Smuzhiyun return 0;
418*4882a593Smuzhiyun }
419*4882a593Smuzhiyun
process_events(struct machine * machine,struct evlist * evlist,struct state * state)420*4882a593Smuzhiyun static int process_events(struct machine *machine, struct evlist *evlist,
421*4882a593Smuzhiyun struct state *state)
422*4882a593Smuzhiyun {
423*4882a593Smuzhiyun union perf_event *event;
424*4882a593Smuzhiyun struct mmap *md;
425*4882a593Smuzhiyun int i, ret;
426*4882a593Smuzhiyun
427*4882a593Smuzhiyun for (i = 0; i < evlist->core.nr_mmaps; i++) {
428*4882a593Smuzhiyun md = &evlist->mmap[i];
429*4882a593Smuzhiyun if (perf_mmap__read_init(&md->core) < 0)
430*4882a593Smuzhiyun continue;
431*4882a593Smuzhiyun
432*4882a593Smuzhiyun while ((event = perf_mmap__read_event(&md->core)) != NULL) {
433*4882a593Smuzhiyun ret = process_event(machine, evlist, event, state);
434*4882a593Smuzhiyun perf_mmap__consume(&md->core);
435*4882a593Smuzhiyun if (ret < 0)
436*4882a593Smuzhiyun return ret;
437*4882a593Smuzhiyun }
438*4882a593Smuzhiyun perf_mmap__read_done(&md->core);
439*4882a593Smuzhiyun }
440*4882a593Smuzhiyun return 0;
441*4882a593Smuzhiyun }
442*4882a593Smuzhiyun
comp(const void * a,const void * b)443*4882a593Smuzhiyun static int comp(const void *a, const void *b)
444*4882a593Smuzhiyun {
445*4882a593Smuzhiyun return *(int *)a - *(int *)b;
446*4882a593Smuzhiyun }
447*4882a593Smuzhiyun
do_sort_something(void)448*4882a593Smuzhiyun static void do_sort_something(void)
449*4882a593Smuzhiyun {
450*4882a593Smuzhiyun int buf[40960], i;
451*4882a593Smuzhiyun
452*4882a593Smuzhiyun for (i = 0; i < (int)ARRAY_SIZE(buf); i++)
453*4882a593Smuzhiyun buf[i] = ARRAY_SIZE(buf) - i - 1;
454*4882a593Smuzhiyun
455*4882a593Smuzhiyun qsort(buf, ARRAY_SIZE(buf), sizeof(int), comp);
456*4882a593Smuzhiyun
457*4882a593Smuzhiyun for (i = 0; i < (int)ARRAY_SIZE(buf); i++) {
458*4882a593Smuzhiyun if (buf[i] != i) {
459*4882a593Smuzhiyun pr_debug("qsort failed\n");
460*4882a593Smuzhiyun break;
461*4882a593Smuzhiyun }
462*4882a593Smuzhiyun }
463*4882a593Smuzhiyun }
464*4882a593Smuzhiyun
sort_something(void)465*4882a593Smuzhiyun static void sort_something(void)
466*4882a593Smuzhiyun {
467*4882a593Smuzhiyun int i;
468*4882a593Smuzhiyun
469*4882a593Smuzhiyun for (i = 0; i < 10; i++)
470*4882a593Smuzhiyun do_sort_something();
471*4882a593Smuzhiyun }
472*4882a593Smuzhiyun
syscall_something(void)473*4882a593Smuzhiyun static void syscall_something(void)
474*4882a593Smuzhiyun {
475*4882a593Smuzhiyun int pipefd[2];
476*4882a593Smuzhiyun int i;
477*4882a593Smuzhiyun
478*4882a593Smuzhiyun for (i = 0; i < 1000; i++) {
479*4882a593Smuzhiyun if (pipe(pipefd) < 0) {
480*4882a593Smuzhiyun pr_debug("pipe failed\n");
481*4882a593Smuzhiyun break;
482*4882a593Smuzhiyun }
483*4882a593Smuzhiyun close(pipefd[1]);
484*4882a593Smuzhiyun close(pipefd[0]);
485*4882a593Smuzhiyun }
486*4882a593Smuzhiyun }
487*4882a593Smuzhiyun
fs_something(void)488*4882a593Smuzhiyun static void fs_something(void)
489*4882a593Smuzhiyun {
490*4882a593Smuzhiyun const char *test_file_name = "temp-perf-code-reading-test-file--";
491*4882a593Smuzhiyun FILE *f;
492*4882a593Smuzhiyun int i;
493*4882a593Smuzhiyun
494*4882a593Smuzhiyun for (i = 0; i < 1000; i++) {
495*4882a593Smuzhiyun f = fopen(test_file_name, "w+");
496*4882a593Smuzhiyun if (f) {
497*4882a593Smuzhiyun fclose(f);
498*4882a593Smuzhiyun unlink(test_file_name);
499*4882a593Smuzhiyun }
500*4882a593Smuzhiyun }
501*4882a593Smuzhiyun }
502*4882a593Smuzhiyun
503*4882a593Smuzhiyun #ifdef __s390x__
504*4882a593Smuzhiyun #include "header.h" // for get_cpuid()
505*4882a593Smuzhiyun #endif
506*4882a593Smuzhiyun
do_determine_event(bool excl_kernel)507*4882a593Smuzhiyun static const char *do_determine_event(bool excl_kernel)
508*4882a593Smuzhiyun {
509*4882a593Smuzhiyun const char *event = excl_kernel ? "cycles:u" : "cycles";
510*4882a593Smuzhiyun
511*4882a593Smuzhiyun #ifdef __s390x__
512*4882a593Smuzhiyun char cpuid[128], model[16], model_c[16], cpum_cf_v[16];
513*4882a593Smuzhiyun unsigned int family;
514*4882a593Smuzhiyun int ret, cpum_cf_a;
515*4882a593Smuzhiyun
516*4882a593Smuzhiyun if (get_cpuid(cpuid, sizeof(cpuid)))
517*4882a593Smuzhiyun goto out_clocks;
518*4882a593Smuzhiyun ret = sscanf(cpuid, "%*[^,],%u,%[^,],%[^,],%[^,],%x", &family, model_c,
519*4882a593Smuzhiyun model, cpum_cf_v, &cpum_cf_a);
520*4882a593Smuzhiyun if (ret != 5) /* Not available */
521*4882a593Smuzhiyun goto out_clocks;
522*4882a593Smuzhiyun if (excl_kernel && (cpum_cf_a & 4))
523*4882a593Smuzhiyun return event;
524*4882a593Smuzhiyun if (!excl_kernel && (cpum_cf_a & 2))
525*4882a593Smuzhiyun return event;
526*4882a593Smuzhiyun
527*4882a593Smuzhiyun /* Fall through: missing authorization */
528*4882a593Smuzhiyun out_clocks:
529*4882a593Smuzhiyun event = excl_kernel ? "cpu-clock:u" : "cpu-clock";
530*4882a593Smuzhiyun
531*4882a593Smuzhiyun #endif
532*4882a593Smuzhiyun return event;
533*4882a593Smuzhiyun }
534*4882a593Smuzhiyun
do_something(void)535*4882a593Smuzhiyun static void do_something(void)
536*4882a593Smuzhiyun {
537*4882a593Smuzhiyun fs_something();
538*4882a593Smuzhiyun
539*4882a593Smuzhiyun sort_something();
540*4882a593Smuzhiyun
541*4882a593Smuzhiyun syscall_something();
542*4882a593Smuzhiyun }
543*4882a593Smuzhiyun
544*4882a593Smuzhiyun enum {
545*4882a593Smuzhiyun TEST_CODE_READING_OK,
546*4882a593Smuzhiyun TEST_CODE_READING_NO_VMLINUX,
547*4882a593Smuzhiyun TEST_CODE_READING_NO_KCORE,
548*4882a593Smuzhiyun TEST_CODE_READING_NO_ACCESS,
549*4882a593Smuzhiyun TEST_CODE_READING_NO_KERNEL_OBJ,
550*4882a593Smuzhiyun };
551*4882a593Smuzhiyun
do_test_code_reading(bool try_kcore)552*4882a593Smuzhiyun static int do_test_code_reading(bool try_kcore)
553*4882a593Smuzhiyun {
554*4882a593Smuzhiyun struct machine *machine;
555*4882a593Smuzhiyun struct thread *thread;
556*4882a593Smuzhiyun struct record_opts opts = {
557*4882a593Smuzhiyun .mmap_pages = UINT_MAX,
558*4882a593Smuzhiyun .user_freq = UINT_MAX,
559*4882a593Smuzhiyun .user_interval = ULLONG_MAX,
560*4882a593Smuzhiyun .freq = 500,
561*4882a593Smuzhiyun .target = {
562*4882a593Smuzhiyun .uses_mmap = true,
563*4882a593Smuzhiyun },
564*4882a593Smuzhiyun };
565*4882a593Smuzhiyun struct state state = {
566*4882a593Smuzhiyun .done_cnt = 0,
567*4882a593Smuzhiyun };
568*4882a593Smuzhiyun struct perf_thread_map *threads = NULL;
569*4882a593Smuzhiyun struct perf_cpu_map *cpus = NULL;
570*4882a593Smuzhiyun struct evlist *evlist = NULL;
571*4882a593Smuzhiyun struct evsel *evsel = NULL;
572*4882a593Smuzhiyun int err = -1, ret;
573*4882a593Smuzhiyun pid_t pid;
574*4882a593Smuzhiyun struct map *map;
575*4882a593Smuzhiyun bool have_vmlinux, have_kcore, excl_kernel = false;
576*4882a593Smuzhiyun
577*4882a593Smuzhiyun pid = getpid();
578*4882a593Smuzhiyun
579*4882a593Smuzhiyun machine = machine__new_host();
580*4882a593Smuzhiyun machine->env = &perf_env;
581*4882a593Smuzhiyun
582*4882a593Smuzhiyun ret = machine__create_kernel_maps(machine);
583*4882a593Smuzhiyun if (ret < 0) {
584*4882a593Smuzhiyun pr_debug("machine__create_kernel_maps failed\n");
585*4882a593Smuzhiyun goto out_err;
586*4882a593Smuzhiyun }
587*4882a593Smuzhiyun
588*4882a593Smuzhiyun /* Force the use of kallsyms instead of vmlinux to try kcore */
589*4882a593Smuzhiyun if (try_kcore)
590*4882a593Smuzhiyun symbol_conf.kallsyms_name = "/proc/kallsyms";
591*4882a593Smuzhiyun
592*4882a593Smuzhiyun /* Load kernel map */
593*4882a593Smuzhiyun map = machine__kernel_map(machine);
594*4882a593Smuzhiyun ret = map__load(map);
595*4882a593Smuzhiyun if (ret < 0) {
596*4882a593Smuzhiyun pr_debug("map__load failed\n");
597*4882a593Smuzhiyun goto out_err;
598*4882a593Smuzhiyun }
599*4882a593Smuzhiyun have_vmlinux = dso__is_vmlinux(map->dso);
600*4882a593Smuzhiyun have_kcore = dso__is_kcore(map->dso);
601*4882a593Smuzhiyun
602*4882a593Smuzhiyun /* 2nd time through we just try kcore */
603*4882a593Smuzhiyun if (try_kcore && !have_kcore)
604*4882a593Smuzhiyun return TEST_CODE_READING_NO_KCORE;
605*4882a593Smuzhiyun
606*4882a593Smuzhiyun /* No point getting kernel events if there is no kernel object */
607*4882a593Smuzhiyun if (!have_vmlinux && !have_kcore)
608*4882a593Smuzhiyun excl_kernel = true;
609*4882a593Smuzhiyun
610*4882a593Smuzhiyun threads = thread_map__new_by_tid(pid);
611*4882a593Smuzhiyun if (!threads) {
612*4882a593Smuzhiyun pr_debug("thread_map__new_by_tid failed\n");
613*4882a593Smuzhiyun goto out_err;
614*4882a593Smuzhiyun }
615*4882a593Smuzhiyun
616*4882a593Smuzhiyun ret = perf_event__synthesize_thread_map(NULL, threads,
617*4882a593Smuzhiyun perf_event__process, machine, false);
618*4882a593Smuzhiyun if (ret < 0) {
619*4882a593Smuzhiyun pr_debug("perf_event__synthesize_thread_map failed\n");
620*4882a593Smuzhiyun goto out_err;
621*4882a593Smuzhiyun }
622*4882a593Smuzhiyun
623*4882a593Smuzhiyun thread = machine__findnew_thread(machine, pid, pid);
624*4882a593Smuzhiyun if (!thread) {
625*4882a593Smuzhiyun pr_debug("machine__findnew_thread failed\n");
626*4882a593Smuzhiyun goto out_put;
627*4882a593Smuzhiyun }
628*4882a593Smuzhiyun
629*4882a593Smuzhiyun cpus = perf_cpu_map__new(NULL);
630*4882a593Smuzhiyun if (!cpus) {
631*4882a593Smuzhiyun pr_debug("perf_cpu_map__new failed\n");
632*4882a593Smuzhiyun goto out_put;
633*4882a593Smuzhiyun }
634*4882a593Smuzhiyun
635*4882a593Smuzhiyun while (1) {
636*4882a593Smuzhiyun const char *str;
637*4882a593Smuzhiyun
638*4882a593Smuzhiyun evlist = evlist__new();
639*4882a593Smuzhiyun if (!evlist) {
640*4882a593Smuzhiyun pr_debug("perf_evlist__new failed\n");
641*4882a593Smuzhiyun goto out_put;
642*4882a593Smuzhiyun }
643*4882a593Smuzhiyun
644*4882a593Smuzhiyun perf_evlist__set_maps(&evlist->core, cpus, threads);
645*4882a593Smuzhiyun
646*4882a593Smuzhiyun str = do_determine_event(excl_kernel);
647*4882a593Smuzhiyun pr_debug("Parsing event '%s'\n", str);
648*4882a593Smuzhiyun ret = parse_events(evlist, str, NULL);
649*4882a593Smuzhiyun if (ret < 0) {
650*4882a593Smuzhiyun pr_debug("parse_events failed\n");
651*4882a593Smuzhiyun goto out_put;
652*4882a593Smuzhiyun }
653*4882a593Smuzhiyun
654*4882a593Smuzhiyun perf_evlist__config(evlist, &opts, NULL);
655*4882a593Smuzhiyun
656*4882a593Smuzhiyun evsel = evlist__first(evlist);
657*4882a593Smuzhiyun
658*4882a593Smuzhiyun evsel->core.attr.comm = 1;
659*4882a593Smuzhiyun evsel->core.attr.disabled = 1;
660*4882a593Smuzhiyun evsel->core.attr.enable_on_exec = 0;
661*4882a593Smuzhiyun
662*4882a593Smuzhiyun ret = evlist__open(evlist);
663*4882a593Smuzhiyun if (ret < 0) {
664*4882a593Smuzhiyun if (!excl_kernel) {
665*4882a593Smuzhiyun excl_kernel = true;
666*4882a593Smuzhiyun /*
667*4882a593Smuzhiyun * Both cpus and threads are now owned by evlist
668*4882a593Smuzhiyun * and will be freed by following perf_evlist__set_maps
669*4882a593Smuzhiyun * call. Getting refference to keep them alive.
670*4882a593Smuzhiyun */
671*4882a593Smuzhiyun perf_cpu_map__get(cpus);
672*4882a593Smuzhiyun perf_thread_map__get(threads);
673*4882a593Smuzhiyun perf_evlist__set_maps(&evlist->core, NULL, NULL);
674*4882a593Smuzhiyun evlist__delete(evlist);
675*4882a593Smuzhiyun evlist = NULL;
676*4882a593Smuzhiyun continue;
677*4882a593Smuzhiyun }
678*4882a593Smuzhiyun
679*4882a593Smuzhiyun if (verbose > 0) {
680*4882a593Smuzhiyun char errbuf[512];
681*4882a593Smuzhiyun evlist__strerror_open(evlist, errno, errbuf, sizeof(errbuf));
682*4882a593Smuzhiyun pr_debug("perf_evlist__open() failed!\n%s\n", errbuf);
683*4882a593Smuzhiyun }
684*4882a593Smuzhiyun
685*4882a593Smuzhiyun goto out_put;
686*4882a593Smuzhiyun }
687*4882a593Smuzhiyun break;
688*4882a593Smuzhiyun }
689*4882a593Smuzhiyun
690*4882a593Smuzhiyun ret = evlist__mmap(evlist, UINT_MAX);
691*4882a593Smuzhiyun if (ret < 0) {
692*4882a593Smuzhiyun pr_debug("evlist__mmap failed\n");
693*4882a593Smuzhiyun goto out_put;
694*4882a593Smuzhiyun }
695*4882a593Smuzhiyun
696*4882a593Smuzhiyun evlist__enable(evlist);
697*4882a593Smuzhiyun
698*4882a593Smuzhiyun do_something();
699*4882a593Smuzhiyun
700*4882a593Smuzhiyun evlist__disable(evlist);
701*4882a593Smuzhiyun
702*4882a593Smuzhiyun ret = process_events(machine, evlist, &state);
703*4882a593Smuzhiyun if (ret < 0)
704*4882a593Smuzhiyun goto out_put;
705*4882a593Smuzhiyun
706*4882a593Smuzhiyun if (!have_vmlinux && !have_kcore && !try_kcore)
707*4882a593Smuzhiyun err = TEST_CODE_READING_NO_KERNEL_OBJ;
708*4882a593Smuzhiyun else if (!have_vmlinux && !try_kcore)
709*4882a593Smuzhiyun err = TEST_CODE_READING_NO_VMLINUX;
710*4882a593Smuzhiyun else if (excl_kernel)
711*4882a593Smuzhiyun err = TEST_CODE_READING_NO_ACCESS;
712*4882a593Smuzhiyun else
713*4882a593Smuzhiyun err = TEST_CODE_READING_OK;
714*4882a593Smuzhiyun out_put:
715*4882a593Smuzhiyun thread__put(thread);
716*4882a593Smuzhiyun out_err:
717*4882a593Smuzhiyun
718*4882a593Smuzhiyun if (evlist) {
719*4882a593Smuzhiyun evlist__delete(evlist);
720*4882a593Smuzhiyun } else {
721*4882a593Smuzhiyun perf_cpu_map__put(cpus);
722*4882a593Smuzhiyun perf_thread_map__put(threads);
723*4882a593Smuzhiyun }
724*4882a593Smuzhiyun machine__delete_threads(machine);
725*4882a593Smuzhiyun machine__delete(machine);
726*4882a593Smuzhiyun
727*4882a593Smuzhiyun return err;
728*4882a593Smuzhiyun }
729*4882a593Smuzhiyun
test__code_reading(struct test * test __maybe_unused,int subtest __maybe_unused)730*4882a593Smuzhiyun int test__code_reading(struct test *test __maybe_unused, int subtest __maybe_unused)
731*4882a593Smuzhiyun {
732*4882a593Smuzhiyun int ret;
733*4882a593Smuzhiyun
734*4882a593Smuzhiyun ret = do_test_code_reading(false);
735*4882a593Smuzhiyun if (!ret)
736*4882a593Smuzhiyun ret = do_test_code_reading(true);
737*4882a593Smuzhiyun
738*4882a593Smuzhiyun switch (ret) {
739*4882a593Smuzhiyun case TEST_CODE_READING_OK:
740*4882a593Smuzhiyun return 0;
741*4882a593Smuzhiyun case TEST_CODE_READING_NO_VMLINUX:
742*4882a593Smuzhiyun pr_debug("no vmlinux\n");
743*4882a593Smuzhiyun return 0;
744*4882a593Smuzhiyun case TEST_CODE_READING_NO_KCORE:
745*4882a593Smuzhiyun pr_debug("no kcore\n");
746*4882a593Smuzhiyun return 0;
747*4882a593Smuzhiyun case TEST_CODE_READING_NO_ACCESS:
748*4882a593Smuzhiyun pr_debug("no access\n");
749*4882a593Smuzhiyun return 0;
750*4882a593Smuzhiyun case TEST_CODE_READING_NO_KERNEL_OBJ:
751*4882a593Smuzhiyun pr_debug("no kernel obj\n");
752*4882a593Smuzhiyun return 0;
753*4882a593Smuzhiyun default:
754*4882a593Smuzhiyun return -1;
755*4882a593Smuzhiyun };
756*4882a593Smuzhiyun }
757